UHG
Search
Close this search box.

Yann LeCun Advices Students Getting Into AI Space to ‘Not Work on LLMs’

“I'm working on the next generation AI systems myself, not on LLMs,” said Yann LeCun.

Share

Yann LeCun LLM

At VivaTech, the annual technology conference for startups, under way in Paris, Meta AI chief Yann LeCun, advised students looking to work in the AI ecosystem, to not work on LLMs.

“If you are a student interested in building the next generation of AI systems, don’t work on LLMs. This is in the hands of large companies, there’s nothing you can bring to the table,” said LeCun at the conference.  

He also said that people should develop next-generation AI systems that overcome the limitations of large language models.

Moving Away from LLMs

Interestingly, the discussion on alternatives of LLM-based models has been ongoing for a while now. Recently, Mufeed VH, the young creator of Devika, a Devin alternative, spoke about how people should move away from Transformer models and start building new architectures. 

“Everyone’s doing the same thing, but if we focus on different architectures, like RMKV [an RNN architecture], it would be really good,” said Mufeed, who goes on to explain the unlimited context window and inference for that particular architecture. 

He also believes that with this approach, it is even possible to build something nearly as impressive as GPT-4. 

Moving away from LLMs is something LeCun has been prominently advocating. He believes in taking the control away from the hands of a few. This is another reason why he pushes for open-source too. 

“Eventually all our interactions with the digital world will be mediated by AI assistants,” he said, and urged platforms to not allow a small number of AI assistants to control the entire digital world. 

“This will be extremely dangerous for diversity of thought, for democracy, for just about everything”, he said.

But, LLMs are Only Advancing

While LeCun might be against LLMs, the Transformer training models are evolving. Dan Hou, an AI/ML advisor, spoke about GPT-4o and emphasised on its training model. 

When text was believed to be the basis for all sophisticated models, GPT-4o was designed to understand video and audio natively. This impacts the volume of data that future versions can be trained on. 

“How much smarter can AI get? With a natively multi-modal architecture, I suspect the answer is much, much better,” said Hou. 

Furthermore, Sam Altman, in a recent interview also spoke about how data wouldn’t be a problem anymore, thereby addressing the concerns of training LLMs. 

Source: X

📣 Want to advertise in AIM? Book here

Picture of Vandana Nair

Vandana Nair

As a rare blend of engineering, MBA, and journalism degree, Vandana Nair brings a unique combination of technical know-how, business acumen, and storytelling skills to the table. Her insatiable curiosity for all things startups, businesses, and AI technologies ensures that there's always a fresh and insightful perspective to her reporting.
Related Posts
19th - 23rd Aug 2024
Generative AI Crash Course for Non-Techies
Upcoming Large format Conference
Sep 25-27, 2024 | 📍 Bangalore, India
Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

Flagship Events

Rising 2024 | DE&I in Tech Summit
April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore
Data Engineering Summit 2024
May 30 and 31, 2024 | 📍 Bangalore, India
MachineCon USA 2024
26 July 2024 | 583 Park Avenue, New York
MachineCon GCC Summit 2024
June 28 2024 | 📍Bangalore, India
Cypher USA 2024
Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA
Cypher India 2024
September 25-27, 2024 | 📍Bangalore, India
discord-icon
AI Forum for India
Our Discord Community for AI Ecosystem, In collaboration with NVIDIA.