Ampere Announces 256-Core 3nm CPU, Unveils Partnership with Qualcomm
Arthur T Knackerbracket has processed the following story:
Ampere Computing todayintroducedits roadmap for the coming years, including new CPUs and collaborations with third parties. In particular, the company said it would launch its all-new 256-core AmpereOne processor next year, made on TSMC's N3 process technology. Also, Ampere is teaming up with Qualcomm to build AI inference servers with the company's accelerators. Apparently, Huawei is also looking at integrating third-party UCIe-compatible chiplets into its own platforms.
Ampere has begun shipping 192-core AmpereOne processors with an eight-channel DDR5 memory subsystem it introduced a year ago. Later this year, the company plans to introduce 192-core AmpereOne CPUs with a 12-channel DDR5 memory subsystem, requiring a brand-new platform.
Next year, the company will use this platform for its 256-core AmpereOne CPU, which will be made using one of TSMC's N3 fabrication processes. The company does not disclose whether the new processor will also feature a new microarchitecture, though it looks like it will continue to feature 2 MB of L2 cache per core.
"We are extending our product family to include a new 256-core product that delivers 40% more performance than any other CPU in the market," said Renee James, chief executive of Ampere. "It is not just about cores. It is about what you can do with the platform. We have several new features that enable efficient performance, memory, caching, and AI compute."
The company says that its 256-core CPU will use the same cooling system as its existing offerings, which implies that its thermal design power will remain in the 350-watt ballpark.
While Ampere can certainly address many general-purpose cloud instances, its capabilities for AI are fairly limited. The company itself says that its 128-core AmpereOne CPU with its two 128-bit vector units per core (and supporting INT8, INT16, FP16, and BFloat16 formats) can offer performance comparable to Nvidia's A10 GPU, albeit at lower power. [...] So, it teamed up with Qualcomm, and the two companies plan to build platforms for LLM inferencing based on Ampere's CPUs andQualcomm's Cloud AI 100 Ultra accelerators.
Read more of this story at SoylentNews.