Article 6STCR Technical Challenges to Scale Beyond GPT4 to 100K H100s

Technical Challenges to Scale Beyond GPT4 to 100K H100s

by
Brian Wang
from NextBigFuture.com on (#6STCR)
Story ImageUp until late 2024, no one has been able to massively increase the amount of compute dedicated to a single model beyond the OpenAI GPT 4 model level. This information is from semianalysis and EIA. Google's Gemini Ultra, Nvidia Nemotron 340B, and Meta LLAMA 3 405B had similar or slightly more compute than GPT-4, but ...

Read more

External Content
Source RSS or Atom Feed
Feed Location http://feeds.feedburner.com/blogspot/advancednano
Feed Title NextBigFuture.com
Feed Link https://www.nextbigfuture.com/
Reply 0 comments