• December 7, 2022

AI Shake-Up As Prominent AI Guru Proposes Mind-Bending “Mortal Computers” Which Also Gets AI Ethics And AI Law Dug In

Here’s something that you probably hadn’t been yet mulling over: Mortal computers. But maybe you should be. The heady topic came up at the recent and altogether quite prominent annual conference …

Piiano Releases Secure Database For Enterprises And Developers

Data protection company Piiano has released Piiano Vault, a secure database that’s designed to give enterprises the ability to store safely sensitive personal data in compliance with the EU’s GDPR, California’s …

How Creative Artists Agency (CAA) Is Employing Data, Analytics, And AI To Shape The Culture Of Our Times And Inspire The World

George Clooney. Cate Blanchett. Beyoncé. Lady Gaga. Brad Pitt. Tom Hanks. What do these artists have in common? Each of them, among others, in addition to a roster of professional athletes, …

Company stood up an AI Exaflop system in just three days!

At Supercomputing ‘22 in Dallas, Texas, startup Cerebras has announced the company has created its own supercomputer, much like NVIDIA has done with Selene. The system, called Andromeda, is composed of 16 CS-2 servers, with 13.5 million AI cores, far more than the TOP500 winner Frontier with 8.7 million cores. The cluster is driven by 284 Gen 3 EPYC processors, and will provide the community at large access to the unique Wafer-Scale Engine the company has invented to test AI models and determine the performance of a CS-2 cluster. To our knowledge this is the first demonstration of the company’s MemoryX and ScaleX technologies we covered here.

Cerebras is especially proud of three accomplishments with this system. First, it was super easy to set up, both physically and logically, to run AI jobs. 3 days! Holy cow. Second, it is simple to program; the Cerebras software stack and compiler take care of all the nitty gritty; you just enter a line of code, specifying on how many CS-2’s to run it on, and poof! You are done. Finally, this system demonstrates nearly perfect linear scaling; triple the number of nodes, and the job is done three times faster.

We are especially anxious to learn how well this supercomputer performs with very large language models. We know it scales well, and the initial results from work being done on Covid evolutionary dynamics is impressive as well.


The good news here is that Cerebras has demonstrated they can deliver on their promises: a super-fast AI machine that scales extremely well and is easy to use. The bad news is that they had to pay for it, instead of having a customer fund the system. Perhaps now that users can fully understand the potential of the WSE, MemoryX, and ScaleX we will begin to see more client deployments. Unlike several AI startups who have folded or initiated layoffs, Cerebras is still growing, and gaining a small but growing fan base that can see past the high initial dollar per server and realize the benefits of this one-of-a-kind AI machine.


Leave a Reply

Your email address will not be published.