Universities today face many challenges. With today’s market shifts, universities must adapt. Today’s key issues include economic downtrends hitting endowments, enrollment challenges due to rising tuition, and competition from third-party online education and demographic shifts leading to a shrinking pool of new students. Meanwhile, brain drain…
MIT linguist Irene Heim shares Schock Prize in Logic and Philosophy
Linguist Irene Heim, professor emerita in MIT’s Department of Linguistics and Philosophy, has been named a co-recipient of the 2024 Rolf Schock Prize in Logic and Philosophy.
Heim shares the award with Hans Kamp, a professor of formal logics and philosophy of language at the University of Stuttgart in Germany. Heim and Kamp are being recognized for their independent work on the “conception and early development of dynamic semantics for natural language.”
The Schock Prize in Logic and Philosophy, sometimes referred to as the Nobel Prize of philosophy, is awarded every three years by the Schock Foundation to distinguished international recipients proposed by the Royal Swedish Academy of Sciences. A prize ceremony and symposium will be held at the Royal Academy of Fine Arts in Stockholm Nov. 11-12. MIT will host a separate event on campus celebrating Heim’s achievement on Dec. 7.
A press release from the Royal Swedish Academy of Sciences explains more about the research for which Heim and Kamp were recognized:
“Natural languages are highly context-dependent — how a sentence is interpreted often depends on the situation, but also on what has been uttered before. In one type of case, a pronoun depends on an earlier phrase in a separate clause. In the mid-1970s, some constructions of this type posed a hard problem for formal semantic theory.
“Around 1980, Hans Kamp and Irene Heim each separately developed similar solutions to this problem. Their theories brought far-reaching changes in the field. Both introduced a new level of representation between the linguistic expression and its worldly interpretation and, in both, this level has a new type of linguistic meaning. Instead of the traditional idea that a clause describes a worldly condition, meaning at this level consists in the way it contributes to updating information. Based on these fundamentally new ideas, the theories provide adequate interpretations of the problematic constructions.”
This is the first time the prize has been awarded for work done in linguistics. The work has had a transformative effect on three major subfields of linguistics: the study of linguistic mental representation (syntax), the study of their logical properties (semantics), and the study of the conditions on the use of linguistic expressions in conversation (pragmatics). Heim has published dozens of texts on semantics and syntax of language.
“I am struck again and again by how our field has progressed in the 50 years since I first entered it and the 40 years since my co-awardee and I contributed the work which won the award,” Heim said. “Those old contributions now look kind of simple-minded, in some spots even confused. But — like other influential ideas in this half-century of linguistics and philosophy of language — they have been influential not just because many people ran with them, but more so because many people picked them apart and explored ever more sophisticated and satisfying alternatives to them.”
Heim, a recognized leader in the fields of syntax and semantics, was born in Germany in 1954. She studied at the University of Konstanz and the Ludwig Maximilian University of Munich, where she earned an MA in philosophy while minoring in linguistics and mathematics. She later earned a PhD in linguistics at the University of Massachusetts at Amherst. She previously taught at the University of Texas at Austin and the University of California Los Angeles before joining MIT’s faculty in 1989.
“I am proud to think of myself as Irene’s student,” says Danny Fox, linguistics section head and the Anshen-Chomsky Professor of Language and Thought. “Irene’s work has served as the foundation of so many areas of our field, and she is rightfully famous for it. But her influence goes even deeper than that. She has taught generations of researchers, primarily by example, how to think anew about entrenched ideas (including her own contributions), how much there is to gain from careful analysis of theoretical proposals, and at the same time, how not to entirely neglect our ambitious aspirations to move beyond this careful work and think about when it might be appropriate to take substantive risks.”
UltiHash’s Sustainable Data Infrastructure Tackles AI Storage Challenges
UltiHash, a provider of high-performance and eco-friendly data storage solutions, has launched its object storage platform to address critical issues in AI data storage. This development aims to resolve mounting challenges in the AI industry related to infrastructure costs and environmental sustainability. The new platform enhances…
Bridging Knowledge Gaps in AI with RAG: Techniques and Strategies for Enhanced Performance
Artificial Intelligence (AI) has revolutionized how we interact with technology, leading to the rise of virtual assistants, chatbots, and other automated systems capable of handling complex tasks. Despite this progress, even the most advanced AI systems encounter significant limitations known as knowledge gaps. For instance, when…
Combining next-token prediction and video diffusion in computer vision and robotics
In the current AI zeitgeist, sequence models have skyrocketed in popularity for their ability to analyze data and predict what to do next. For instance, you’ve likely used next-token prediction models like ChatGPT, which anticipate each word (token) in a sequence to form answers to users’ queries. There are also full-sequence diffusion models like Sora, which convert words into dazzling, realistic visuals by successively “denoising” an entire video sequence.
Researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) have proposed a simple change to the diffusion training scheme that makes this sequence denoising considerably more flexible.
When applied to fields like computer vision and robotics, the next-token and full-sequence diffusion models have capability trade-offs. Next-token models can spit out sequences that vary in length. However, they make these generations while being unaware of desirable states in the far future — such as steering its sequence generation toward a certain goal 10 tokens away — and thus require additional mechanisms for long-horizon (long-term) planning. Diffusion models can perform such future-conditioned sampling, but lack the ability of next-token models to generate variable-length sequences.
Researchers from CSAIL want to combine the strengths of both models, so they created a sequence model training technique called “Diffusion Forcing.” The name comes from “Teacher Forcing,” the conventional training scheme that breaks down full sequence generation into the smaller, easier steps of next-token generation (much like a good teacher simplifying a complex concept).
Diffusion Forcing found common ground between diffusion models and teacher forcing: They both use training schemes that involve predicting masked (noisy) tokens from unmasked ones. In the case of diffusion models, they gradually add noise to data, which can be viewed as fractional masking. The MIT researchers’ Diffusion Forcing method trains neural networks to cleanse a collection of tokens, removing different amounts of noise within each one while simultaneously predicting the next few tokens. The result: a flexible, reliable sequence model that resulted in higher-quality artificial videos and more precise decision-making for robots and AI agents.
By sorting through noisy data and reliably predicting the next steps in a task, Diffusion Forcing can aid a robot in ignoring visual distractions to complete manipulation tasks. It can also generate stable and consistent video sequences and even guide an AI agent through digital mazes. This method could potentially enable household and factory robots to generalize to new tasks and improve AI-generated entertainment.
“Sequence models aim to condition on the known past and predict the unknown future, a type of binary masking. However, masking doesn’t need to be binary,” says lead author, MIT electrical engineering and computer science (EECS) PhD student, and CSAIL member Boyuan Chen. “With Diffusion Forcing, we add different levels of noise to each token, effectively serving as a type of fractional masking. At test time, our system can “unmask” a collection of tokens and diffuse a sequence in the near future at a lower noise level. It knows what to trust within its data to overcome out-of-distribution inputs.”
In several experiments, Diffusion Forcing thrived at ignoring misleading data to execute tasks while anticipating future actions.
When implemented into a robotic arm, for example, it helped swap two toy fruits across three circular mats, a minimal example of a family of long-horizon tasks that require memories. The researchers trained the robot by controlling it from a distance (or teleoperating it) in virtual reality. The robot is trained to mimic the user’s movements from its camera. Despite starting from random positions and seeing distractions like a shopping bag blocking the markers, it placed the objects into its target spots.
To generate videos, they trained Diffusion Forcing on “Minecraft” game play and colorful digital environments created within Google’s DeepMind Lab Simulator. When given a single frame of footage, the method produced more stable, higher-resolution videos than comparable baselines like a Sora-like full-sequence diffusion model and ChatGPT-like next-token models. These approaches created videos that appeared inconsistent, with the latter sometimes failing to generate working video past just 72 frames.
Diffusion Forcing not only generates fancy videos, but can also serve as a motion planner that steers toward desired outcomes or rewards. Thanks to its flexibility, Diffusion Forcing can uniquely generate plans with varying horizon, perform tree search, and incorporate the intuition that the distant future is more uncertain than the near future. In the task of solving a 2D maze, Diffusion Forcing outperformed six baselines by generating faster plans leading to the goal location, indicating that it could be an effective planner for robots in the future.
Across each demo, Diffusion Forcing acted as a full sequence model, a next-token prediction model, or both. According to Chen, this versatile approach could potentially serve as a powerful backbone for a “world model,” an AI system that can simulate the dynamics of the world by training on billions of internet videos. This would allow robots to perform novel tasks by imagining what they need to do based on their surroundings. For example, if you asked a robot to open a door without being trained on how to do it, the model could produce a video that’ll show the machine how to do it.
The team is currently looking to scale up their method to larger datasets and the latest transformer models to improve performance. They intend to broaden their work to build a ChatGPT-like robot brain that helps robots perform tasks in new environments without human demonstration.
“With Diffusion Forcing, we are taking a step to bringing video generation and robotics closer together,” says senior author Vincent Sitzmann, MIT assistant professor and member of CSAIL, where he leads the Scene Representation group. “In the end, we hope that we can use all the knowledge stored in videos on the internet to enable robots to help in everyday life. Many more exciting research challenges remain, like how robots can learn to imitate humans by watching them even when their own bodies are so different from our own!”
Chen and Sitzmann wrote the paper alongside recent MIT visiting researcher Diego Martí Monsó, and CSAIL affiliates: Yilun Du, a EECS graduate student; Max Simchowitz, former postdoc and incoming Carnegie Mellon University assistant professor; and Russ Tedrake, the Toyota Professor of EECS, Aeronautics and Astronautics, and Mechanical Engineering at MIT, vice president of robotics research at the Toyota Research Institute, and CSAIL member. Their work was supported, in part, by the U.S. National Science Foundation, the Singapore Defence Science and Technology Agency, Intelligence Advanced Research Projects Activity via the U.S. Department of the Interior, and the Amazon Science Hub. They will present their research at NeurIPS in December.
Your guide to LLMOps
Understanding the varied landscape of LLMOps is essential for harnessing the full potential of large language models in today’s digital world. …
Leveraging Human Attention Can Improve AI-Generated Images
New research from China has proposed a method for improving the quality of images generated by Latent Diffusion Models (LDMs) models such as Stable Diffusion. The method focuses on optimizing the salient regions of an image – areas most likely to attract human attention. Traditional methods,…
Scoring AI models: Endor Labs unveils evaluation tool
Endor Labs has begun scoring AI models based on their security, popularity, quality, and activity. Dubbed ‘Endor Scores for AI Models,’ this unique capability aims to simplify the process of identifying the most secure open-source AI models currently available on Hugging Face – a platform for…