Sign up to save tools and stay up to date with the latest in AI
bg
bg
1

OpenAI and Others Seek New Path To Smarter AI as Current Methods Hit Limitations - Slashdot

Nov 11, 2024 - tech.slashdot.org
AI companies like OpenAI are developing new training techniques for language models to overcome challenges and delays. These techniques, used in OpenAI's o1 model, could reshape the AI industry and impact the resources these companies need. However, some AI scientists are questioning the "bigger is better" philosophy of scaling up models with more data and computing power, as results have plateaued. Ilya Sutskever, co-founder of AI labs Safe Superintelligence and OpenAI, is one of those questioning this approach.

OpenAI's newest model, Orion, has not shown significant improvement over its previous model, challenging the assumption that language models will continue to improve with more data and computing power. In response, the industry is shifting its focus to improving models after their initial training. Some CEOs, like Mark Zuckerberg of Meta Platforms, believe there is still potential to build products on top of the current technology, even if it doesn't improve.

Key takeaways:

  • AI companies like OpenAI are developing training techniques that use more human-like ways for algorithms to "think", which could reshape the AI arms race and have implications for the types of resources that AI companies demand.
  • Despite the previous belief that "scaling up" current models through adding more data and computing power will consistently lead to improved AI models, prominent AI scientists are now speaking out on the limitations of this philosophy.
  • Ilya Sutskever, co-founder of AI labs Safe Superintelligence (SSI) and OpenAI, stated that results from scaling up pre-training have plateaued, challenging the assumption that larger language models (LLMs) would continue to improve with more data and computing power.
  • In response to the recent challenge to training-based scaling laws posed by slowing GPT improvements, the industry appears to be shifting its effort to improving models after their initial training, potentially yielding a different type of scaling law.
View Full Article

Comments (0)

Be the first to comment!