Article 6AA5B Cerebras-GPT for Open Source and More Efficient Large Language Models

Cerebras-GPT for Open Source and More Efficient Large Language Models

by
Brian Wang
from NextBigFuture.com on (#6AA5B)
Story ImageCerebras open sources seven GPT-3 models from 111 million to 13 billion parameters. Trained using the Chinchilla formula, these models set new benchmarks for accuracy and compute efficiency. Cerebras makes wafer scale computer chips. Cerebras-GPT has faster training times, lower training costs, and consumes less energy than any publicly available model to date. All models ...

Read more

External Content
Source RSS or Atom Feed
Feed Location http://feeds.feedburner.com/blogspot/advancednano
Feed Title NextBigFuture.com
Feed Link https://www.nextbigfuture.com/
Reply 0 comments