HyperAI超神経
Back to Headlines

Meta’s Chief AI Scientist Argues That Scaling Alone Won’t Lead to Smarter AI Systems

12日前

Yann LeCun, Meta’s chief AI scientist, has challenged the prevailing notion that scaling AI systems will make them inherently smarter. For years, the AI community has subscribed to “scaling laws,” which posit that model performance primarily depends on the number of model parameters, the size of the training dataset, and the amount of computational power. This principle has driven significant investments in data centers and large-scale models, leading to impressive advances in areas like language understanding. However, recent critiques suggest that this approach may not be sufficient for achieving true intelligence. At a symposium at the National University of Singapore on Sunday, LeCun argued that scaling alone will not lead to smarter AI. He pointed out that while simple systems can perform well on straightforward tasks, assuming that they will naturally become more intelligent when scaled up is a misconception. Many current AI breakthroughs, he noted, are relatively easy to achieve because they involve processing large amounts of well-defined data. In contrast, real-world problems characterized by ambiguity and uncertainty require a different approach. LeCun explained that the visual cortex of a four-year-old contains about the same information as the largest language models today. This comparison highlights the gap between human-like intelligence and the capabilities of current AI systems. The effectiveness of scaling is partly due to the availability of public data, but this resource is becoming scarce, contributing to a slowdown in AI advancements. He stressed the need for AI systems to develop abilities beyond text and language, such as understanding the physical world, reasoning, planning, and having persistent memory. World models, according to LeCun, offer a promising direction. Unlike language models, which predict the next step based on patterns, world models can simulate and predict the consequences of actions, mimicking a higher level of cognition. This capability aligns more closely with the intelligence expected from humans or other complex beings. LeCun is not alone in his skepticism. Alexandr Wang, CEO of Scale AI, identified scaling as "the biggest question in the industry" last year at the Cerebral Valley conference. Aidan Gomez, CEO of Cohere, has referred to scaling as the "dumbest" way to improve AI models. These experts agree that focusing solely on scaling is shortsighted and that a more holistic approach, encompassing a broader range of cognitive tasks, is necessary. To address these challenges, LeCun advocates for a paradigm shift in AI training. He envisions systems that can quickly learn new tasks, possess common sense, and reason through complex scenarios. This involves moving away from the current practice of training models on massive datasets and instead developing more sophisticated algorithms that can adapt and learn efficiently from limited data. Such a change could bridge the gap between machine learning and true artificial intelligence, making AI systems more versatile and capable of handling realistic, unpredictable environments. Industry insiders and company leaders are beginning to acknowledge the limitations of scaling and are exploring new avenues to enhance AI capabilities. Companies like Scale AI and Cohere are investing in research to develop more intelligent algorithms, suggesting that the future of AI lies in a combination of better hardware, more efficient data use, and advanced training methods. Meta, under LeCun’s guidance, is likely to follow a similar path, focusing on creating AI systems that can better interact with and understand the real world. This shift in focus could redefine the AI landscape and pave the way for more significant advancements in the field.

Related Links