Article 6TWJH Is Deepseek Training Lying About Chips Used for Training its AI ?

Is Deepseek Training Lying About Chips Used for Training its AI ?

by
Brian Wang
from NextBigFuture.com on (#6TWJH)
Story ImageAltimeter Capital analyst and partner puts what Deepseek claims and results into numbers. $6M Training Costs = Plausible IMO Quick math: Training costs (active params * tokens). DeepSeek v3 (37B params; 14.8T tokens) vs. Llama3.1 (405B params; 15T tokens) = v3 theoretically should be 9% of Llama3.1's cost. And the disclosed actual figures aligned ...

Read more

External Content
Source RSS or Atom Feed
Feed Location http://feeds.feedburner.com/blogspot/advancednano
Feed Title NextBigFuture.com
Feed Link https://www.nextbigfuture.com/
Reply 0 comments