Sign up to save tools and stay up to date with the latest in AI
bg
bg
1

Ask HN: Why is it taken for granted that LLM models will keep improving?

Dec 05, 2023 - news.ycombinator.com
The article is a discussion on Hacker News about the future of Large Language Models (LLMs) like ChatGPT. The original post questions why it is assumed that these models will continue to improve indefinitely, and whether there might be a plateau in performance or impractical compute requirements.

Various users respond with their perspectives, arguing that improvements can still be made in data, compute, and algorithms. Some suggest that the continuous investment of time, money, and effort by smart people in the field will lead to advancements. Others speculate about the potential for LLMs to be integrated into everyday tools or to hit a wall like any other technology. The discussion also touches on the techno-optimist belief in the potential of large computers and the possibility of creating a mechanical god.

Key takeaways:

  • The discussion revolves around the future of Large Language Models (LLMs) and whether their performance will continue to improve or plateau at some point.
  • One user argues that LLMs have room to improve due to advancements in data, compute, and algorithms. They mention the potential for more densely packed transistors, parallelism, and 3-D stacked transistors in computing, and improvements in data collection and algorithm design.
  • Another user suggests that the continuous investment of time, money, and effort by smart people into LLMs is a key reason for the expectation of continuous improvement. They also mention the potential for advancements in GPU capabilities and training techniques.
  • Some users express skepticism about the assumption of continuous improvement, suggesting that LLMs might hit a technological wall or that the performance gains might not be as significant beyond a certain point.
View Full Article

Comments (0)

Be the first to comment!