Continual Learning Working Group: Kick Off
CEPSR 620 Schapiro 530 W. 120th StSpeaker: Mengye Ren Title: Lifelong and Human-like Learning in Foundation Models Abstract: Real-world agents, including humans, learn from online, lifelong experiences. However, today’s foundation models primarily acquire knowledge through offline, iid learning, while relying on in-context learning for most online adaptation. It is crucial to equip foundation models with lifelong and human-like learning abilities to enable more flexible…
ARNI NSF Site Visit
Innovation Hub Tang Family Hall - 2276 12TH AVENUE – FLOOR 02NSF Site Visit - The NSF team will evaluate the progress and achievements of ARNI’s projects to date and provide recommendations to steer future directions and funding for the project. If you are interested in learning more about ARNI over-all, join this Zoom link from 9am to 12pm or 2pm to 4:30pm.
CTN: Eva Dyer
Zuckerman Institute - L5-084 3227 Broadway, New York, NY, United StatesTitle: Large-scale pretraining on neural data allows for transfer across individuals, tasks and species Abstract: As neuroscience datasets grow in size and complexity, integrating diverse data sources to achieve a comprehensive understanding of brain function presents both an opportunity and a challenge. In this talk, I will introduce our approach to developing a multi-source foundation model for…
Continual Learning Working Group: Haozhe Shan
CEPSR 620 Schapiro 530 W. 120th StSpeaker: Haozhe Shan Title: A theory of continual learning in deep neural networks: task relations, network architecture and learning procedure Abstract: Imagine listening to this talk and afterwards forgetting everything else you’ve ever learned. This absurd scenario would be commonplace if the brain could not perform continual learning (CL) – acquiring new skills and knowledge without…
Multi-resource-cost Optimization for Neural Networks Models Working Group (NNMS): Tom Griffiths
Zuckerman Institute - L3-079 3227 Broadway, New York, NY, United StatesTitle: Bounded optimality: A cognitive perspective on neural computation with resource limitations
Multi-resource-cost Optimization for Neural Networks Models Working Group (NNMS): Simon Laughlin
Zuckerman Institute - L3-079 3227 Broadway, New York, NY, United StatesTitle: Neuronal energy consumption: basic measures and trade-offs, and their effects on efficiency Zoom: https://columbiauniversity.zoom.us/j/98299154214?pwd=1J3J0lEpF6XdqHkHy02c7LuD6xUWx2.1
Continual Learning Working Group: Amogh Inamdar
CSB 488Title: Taskonomy: Disentangling Task Transfer Learning Abstract: TBD Link: http://taskonomy.stanford.edu/taskonomy_CVPR2018.pdf
CTN: Brenden Lake
Zuckerman Institute - L3-079 3227 Broadway, New York, NY, United StatesTitle: Meta-learning for more powerful behavioral modeling Abstract: Two modeling paradigms have historically been in tension: Bayesian models provide an elegant way to incorporate prior knowledge, but they make simplifying and constraining assumptions; on the other hand, neural networks provide great modeling flexibility, but they make it difficult to incorporate prior knowledge. Here I describe how to…
Continual Learning Working Group: Lindsay Smith
CEPSR 620 Schapiro 530 W. 120th StTitle: A Practitioner’s Guide to Continual Multimodal Pretraining Reading: https://arxiv.org/pdf/2408.1447 Zoom: https://columbiauniversity.zoom.us/j/97176853843?pwd=VLZdh6yqHBcOQhdf816lkN5ByIpIsF.1
CTN: Benjamin Grewe
Zuckerman Institute - L5-084 3227 Broadway, New York, NY, United StatesTitle: Target Learning rather than Backpropagation Explains Learning in the Mammalian Neocortex Abstract: Modern computational neuroscience presents two competing hypotheses for hierarchical learning in the neocortex: (1) deep learning-inspired approximations of the backpropagation algorithm, where neurons adjust synapses to minimize error, and (2) target learning algorithms, where neurons reduce the feedback required to achieve a desired activity.…
Continual Learning Working Group: Yasaman Mahdaviyeh
CEPSR 620 Schapiro 530 W. 120th StTitle: Meta Continual Learning Revisited: Implicitly Enhancing Online Hessian Approximation via Variance Reduction Reading: https://openreview.net/pdf?id=TpD2aG1h0D Zoom: https://columbiauniversity.zoom.us/j/97176853843?pwd=VLZdh6yqHBcOQhdf816lkN5ByIpIsF.1
ARNI Annual Retreat
Faculty House 64 Morningside DrGeneral Agenda October 21st, Day 1 from 8:45am to 5pm Breakfast and Lunch Provided Opening 3 Keynote Speakers from ARNI Faculty Research Brainstorming and Discussions Project/Student Poster Session Education and Broader Impact Discussions October 22nd, Day 2 from 9am to 1pm Breakfast and Lunch Provided 1 Keynote Speaker Brainstorming and Discussion on Collaborations & Knowledge…