Cerebras
   HOME

TheInfoList



OR:

Cerebras Systems is an American
artificial intelligence Artificial intelligence (AI) is intelligence—perceiving, synthesizing, and inferring information—demonstrated by machines, as opposed to intelligence displayed by animals and humans. Example tasks in which this is done include speech r ...
company with offices in Sunnyvale and
San Diego San Diego ( , ; ) is a city on the Pacific Ocean coast of Southern California located immediately adjacent to the Mexico–United States border. With a 2020 population of 1,386,932, it is the eighth most populous city in the United State ...
,
Toronto Toronto ( ; or ) is the capital city of the Canadian province of Ontario. With a recorded population of 2,794,356 in 2021, it is the most populous city in Canada and the fourth most populous city in North America. The city is the anch ...
,
Tokyo Tokyo (; ja, 東京, , ), officially the Tokyo Metropolis ( ja, 東京都, label=none, ), is the capital and List of cities in Japan, largest city of Japan. Formerly known as Edo, its metropolitan area () is the most populous in the world, ...
and
Bangalore Bangalore (), List of renamed places in India, officially Bengaluru (), is the Capital city, capital and largest city of the Indian state of Karnataka. It has a population of more than and a metropolitan area, metropolitan population of a ...
, India. Cerebras builds computer systems for complex artificial intelligence deep learning applications.


History

Cerebras was founded in 2015 by Andrew Feldman, Gary Lauterbach, Michael James, Sean Lie and Jean-Philippe Fricker. These five founders worked together at SeaMicro, which was started in 2007 by Feldman and Lauterbach and was later sold to AMD in 2012 for $334 million. In May 2016, Cerebras secured $27 million in
series A funding A series A round (also known as series A financing or series A investment) is the name typically given to a company's first significant round of venture capital financing. The name refers to the class of preferred stock sold to investors in exchan ...
led by
Benchmark Benchmark may refer to: Business and economics * Benchmarking, evaluating performance within organizations * Benchmark price * Benchmark (crude oil), oil-specific practices Science and technology * Benchmark (surveying), a point of known elevati ...
,
Foundation Capital Foundation Capital is a venture capital firm located in Silicon Valley. The firm was founded in 1995, and in 2012 managed more than $2.4 billion in investment capital. History Foundation Capital was founded in 1995. The firm raised its sevent ...
and Eclipse Ventures. In December 2016,
series B funding Venture capital financing is a type of funding by venture capital. It is private equity capital that can be provided at various stages or funding rounds. Common funding rounds include early-stage seed funding in high-potential, growth companies ( ...
was led by
Coatue Management Coatue is an American technology-focused investment manager led by founder and portfolio manager Philippe Laffont. Coatue invests in public and private markets with a focus on technology, media, telecommunications. the consumer and healthcare se ...
, followed in January 2017 with series C funding led by VY Capital. In November 2018, Cerebras closed its series D round with $88 million, making the company a
unicorn The unicorn is a legendary creature that has been described since antiquity as a beast with a single large, pointed, spiraling horn projecting from its forehead. In European literature and art, the unicorn has for the last thousand years o ...
. Investors in this round included
Altimeter An altimeter or an altitude meter is an instrument used to measure the altitude of an object above a fixed level. The measurement of altitude is called altimetry, which is related to the term bathymetry, the measurement of depth under water. The m ...
, VY Capital, Coatue, Foundation Capital, Benchmark, and Eclipse. On August 19, 2019, Cerebras announced its Wafer-Scale Engine (WSE). In November 2019, Cerebras closed its series E round with over $270 million for a valuation of $2.4 billion. In 2020, the company announced an office in Japan and partnership with Tokyo Electron Devices. In April 2021, Cerebras announced the CS-2 based on the company's Wafer Scale Engine Two (WSE-2), which has 850,000 cores. In August 2021, the company announced its brain-scale technology that can run a neural network with over 120 trillion connections. In November 2021, Cerebras announced that it had raised an additional $250 million in Series F funding, valuing the company at over $4 billion. The Series F financing round was led by Alpha Wave Ventures and Abu Dhabi Growth Fund (ADG). To date, the company has raised $720 million in financing. In August 2022, Cerebras was honored by the Computer History Museum in
Mountain View, California Mountain View is a city in Santa Clara County, California, United States. Named for its views of the Santa Cruz Mountains, it has a population of 82,376. Mountain View was integral to the early history and growth of Silicon Valley, and is the ...
. The museum added to its permanent collection and unveiled a new display featuring the WSE-2—the biggest computer chip made so far—marking an "epochal" achievement in the history of fabricating transistors as an integrated part. In August 2022, Cerebras announced the opening of a new office in Bangalore, India.


Technology

The Cerebras Wafer Scale Engine (WSE) is a single,
wafer A wafer is a crisp, often sweet, very thin, flat, light and dry biscuit, often used to decorate ice cream, and also used as a garnish on some sweet dishes. Wafers can also be made into cookies with cream flavoring sandwiched between them. They ...
-scale integrated circuit processor that includes
compute Computing is any goal-oriented activity requiring, benefiting from, or creating computing machinery. It includes the study and experimentation of algorithmic processes, and development of both hardware and software. Computing has scientific, e ...
,
memory Memory is the faculty of the mind by which data or information is encoded, stored, and retrieved when needed. It is the retention of information over time for the purpose of influencing future action. If past events could not be remembered ...
, and interconnect fabric.
Scheduling A schedule or a timetable, as a basic time-management tool, consists of a list of times at which possible tasks, events, or actions are intended to take place, or of a sequence of events in the chronological order in which such things are ...
uses a
dataflow architecture Dataflow architecture is a dataflow-based computer architecture that directly contrasts the traditional von Neumann architecture or control flow architecture. Dataflow architectures have no program counter, in concept: the executability and executi ...
. The WSE-1 powers the Cerebras CS-1, the firm's first-generation AI computer. It is a
19-inch rack A 19-inch rack is a standardized frame or enclosure for mounting multiple electronic equipment modules. Each module has a front panel that is wide. The 19 inch dimension includes the edges or "ears" that protrude from each side of the equ ...
-mounted appliance designed for AI training and inference workloads in a datacenter. The CS-1 includes one WSE primary processor with 400,000 processing cores, and twelve
100 Gigabit Ethernet 40 Gigabit Ethernet (40GbE) and 100 Gigabit Ethernet (100GbE) are groups of computer networking technologies for transmitting Ethernet frames at rates of 40 and 100 gigabits per second (Gbit/s), respectively. These technologies offer significantly ...
connections for data
input/output In computing, input/output (I/O, or informally io or IO) is the communication between an information processing system, such as a computer, and the outside world, possibly a human or another information processing system. Inputs are the signals ...
. The WSE-1 has 1.2 trillion transistors, 400,000 compute cores and 18 gigabytes of memory. In April 2021, Cerebras announced the CS-2 AI system based on the 2nd-generation Wafer Scale Engine (WSE-2), manufactured by the
7 nm process In semiconductor manufacturing, the International Technology Roadmap for Semiconductors defines the 7  nm process as the MOSFET technology node following the 10 nm node. It is based on FinFET (fin field-effect transistor) technology, ...
of Taiwan Semiconductor Manufacturing Company (
TSMC Taiwan Semiconductor Manufacturing Company Limited (TSMC; also called Taiwan Semiconductor) is a Taiwanese multinational semiconductor contract manufacturing and design company. It is the world's most valuable semiconductor company, the world' ...
). It is 26 inches tall and fits in one-third of a standard data center rack. The WSE-2 has 850,000 cores and 2.6 trillion transistors. The WSE-2 expanded on-chip
static random-access memory Static random-access memory (static RAM or SRAM) is a type of random-access memory (RAM) that uses latching circuitry (flip-flop) to store each bit. SRAM is volatile memory; data is lost when power is removed. The term ''static'' differe ...
(SRAM) to 40 gigabytes, memory bandwidth to 20 petabytes per second and total fabric bandwidth to 220 petabits per second. In August 2021, the company announced a system which connects multiple integrated circuits (commonly called ''chips'') into a neural network with many connections. It enables one system to support AI models with more than 120 trillion
parameters A parameter (), generally, is any characteristic that can help in defining or classifying a particular system (meaning an event, project, object, situation, etc.). That is, a parameter is an element of a system that is useful, or critical, when ...
. In June 2022, Cerebras set a record for the largest AI models ever trained on one device. Cerebras said that for the first time ever, one CS-2 system with one Cerebras wafer can train models with up to 20 billion parameters. The Cerebras CS-2 system can train multibillion-parameter natural language processing (NLP) models including GPT-3XL 1.3 billion models, GPT-J 6B, GPT-3 13B, and GPT-NeoX 20B with reduced software complexity and infrastructure. In August 2022, Cerebras announced that its customers can now train Transformer-style natural language AI models with 20x longer sequences than possible using traditional computer hardware, which is expected to lead to breakthroughs in natural language processing (NLP), especially for pharmaceuticals and life sciences. In September 2022, Cerebras announced that it can patch its chips together to create what would be the largest-ever computing cluster for AI computing. A Wafer-Scale Cluster can connect up to 192 CS-2 AI systems into a cluster, while a cluster of 16 CS-2 AI systems can create a computing system with 13.6 million cores for natural language processing. The key to the new Cerebras Wafer-Scale Cluster is the exclusive use of data parallelism to train, which is the preferred approach for all AI work. In November 2022, Cerebras unveiled its latest supercomputer, Andromeda, which combines 16 WSE-2 chips into one cluster with 13.5 million AI-optimized cores, delivering up to 1 exa FLOPS of AI computing power, or at least one quintillion (10) operations per second. The entire system uses 500 kilowatts, which is far less power than somewhat-comparable GPU-accelerated supercomputers. In November 2022, Cerebras announced its partnership with Cirrascale Cloud Services to provide a flat-rate "pay-per-model" compute time for its ''Cerebras AI Model Studio''. Pricing ranges from $2,500 for training "a 1.3-billion-parameter model of GPT-3 in 10 hours" to $2.5 million for training "70-billion-parameter version in 85 days". The service is said to reduce the cost—compared to the similar cloud services on the market—by half while increasing speed up to eight times faster.


Deployments

Customers are reportedly using Cerebras technologies in the pharmaceutical, life sciences, and energy sectors. In 2020, GlaxoSmithKline (GSK) began using the Cerebras CS-1 AI system in their London AI hub, for neural network models to accelerate genetic and genomic research and reduce the time taken in drug discovery. The GSK research team was able to increase the complexity of the encoder models they could generate, while reducing training time. Other pharmaceutical industry customers include
AstraZeneca AstraZeneca plc () is a British-Swedish multinational pharmaceutical and biotechnology company with its headquarters at the Cambridge Biomedical Campus in Cambridge, England. It has a portfolio of products for major diseases in areas includi ...
, who was able to reduce training time from two weeks on a cluster of GPUs to two days using the Cerebras CS-1 system. GSK and Cerebras recently co-publishe
research
in December 2021 on epigenomic language models. Argonne National Laboratory has been using the CS-1 since 2020 in COVID-19 research and cancer tumor research based on the world’s largest cancer treatment database. A series of models running on the CS-1 to predict cancer drug response to tumors achieved speed-ups of many hundreds of times on the CS-1 compared to their GPU baselines. Cerebras and the
National Energy Technology Laboratory The National Energy Technology Laboratory (NETL) is a U.S national laboratory under the Department of Energy Office of Fossil Energy. NETL focuses on applied research for the clean production and use of domestic energy resources. NETL perform ...
(NETL) demonstrated record-breaking performance of Cerebras' CS-1 system on a scientific compute workload in November 2020. The CS-1 was 200 times faster than the Joule Supercomputer on the key workload of Computational Fluid Dynamics. The Lawrence Livermore National Lab’s Lassen supercomputer incorporated the CS-1 in both classified and non-classified areas for physics simulations. The
Pittsburgh Supercomputing Center The Pittsburgh Supercomputing Center (PSC) is a high performance computing and networking center founded in 1986 and one of the original five NSF Supercomputing Centers.
(PSC) has also incorporated the CS-1 in their Neocortex supercomputer for dual HPC and AI workloads.
EPCC EPCC, formerly the Edinburgh Parallel Computing Centre, is a supercomputing centre based at the University of Edinburgh. Since its foundation in 1990, its stated mission has been to ''accelerate the effective exploitation of novel computing th ...
, the supercomputing center of the University of Edinburgh, has also deployed a CS-1 system for AI-based research. In August 2021, Cerebras announced a partnership wit
Peptilogics
on the development of AI for
peptide therapeutics Peptide therapeutics are peptides or polypeptides (oligomers or short polymers of amino acids) which are used to for the treatment of diseases. Naturally occurring peptides may serve as hormones, growth factors, neurotransmitters, ion channel ligan ...
. In March 2022, Cerebras announced that the Company deployed its CS-2 system in the Houston facilities of
TotalEnergies TotalEnergies SE is a French multinational integrated energy and petroleum company founded in 1924 and one of the seven supermajor oil companies. Its businesses cover the entire oil and gas chain, from crude oil and natural gas exploration and ...
, its first publicly disclosed customer in the energy sector. Cerebras also announced that it has deployed a CS-2 system a
nference
a startup that uses natural language processing to analyze massive amounts of biomedical data. The CS-2 will be used to train transformer models that are designed to process information from piles of unstructured medical data to provide fresh insights to doctors and improve patient recovery and treatment. In May 2022, Cerebras announced that the
National Center for Supercomputing Applications The National Center for Supercomputing Applications (NCSA) is a state-federal partnership to develop and deploy national-scale computer infrastructure that advances research, science and engineering based in the United States. NCSA operates as a ...
(NCSA) has deployed the Cerebras CS-2 system in their HOLL-I supercomputer. They also announced that the Leibniz Supercomputing Centre (LRZ) in Germany plans to deploy a new supercomputer featuring the CS-2 system along with the HPE Superdome Flex server. The new supercomputing system is expected to be delivered to LRZ this summer. This will be the first CS-2 system deployment in Europe. In October 2022, it was announced that the U.S. National Nuclear Security Administration would sponsor a study to investigate using Cerebras' CS-2 in nuclear stockpile stewardship computing. The multi-year contract will be executed through
Sandia National Laboratories Sandia National Laboratories (SNL), also known as Sandia, is one of three research and development laboratories of the United States Department of Energy's National Nuclear Security Administration (NNSA). Headquartered in Kirtland Air Force Bas ...
, Lawrence Livermore National Lab, and
Los Alamos National Laboratory Los Alamos National Laboratory (often shortened as Los Alamos and LANL) is one of the sixteen research and development laboratories of the United States Department of Energy (DOE), located a short distance northwest of Santa Fe, New Mexico, ...
. In November 2022, Cerebras and the
National Energy Technology Laboratory The National Energy Technology Laboratory (NETL) is a U.S national laboratory under the Department of Energy Office of Fossil Energy. NETL focuses on applied research for the clean production and use of domestic energy resources. NETL perform ...
(NETL) saw record-breaking performance on the scientific compute workload of forming and solving field equations. Cerebras demonstrated that its CS-2 system was as much as 470 times faster than NETL's Joule Supercomputer in field equation modeling. The 2022 Gordon Bell Special Prize Winner for HPC-Based COVID-19 Research, which honors outstanding research achievement towards the understanding of the COVID-19 pandemic through the use of high-performance computing, used Cerebras' CS-2 system to conduct this award-winning research to transform large language models to analyze COVID-19 variants. The paper was authored by a 34-person team from Argonne National Laboratory, California Institute of Technology, Harvard University, Northern Illinois University, Technical University of Munich, University of Chicago, University of Illinois Chicago, Nvidia, and Cerebras. ANL noted that using the CS-2 Wafer-Scale Engine cluster, the team was able to achieve convergence when training on the full SARS-CoV-2 genomes in less than a day.


See also

*
Wafer-scale integration Wafer-scale integration (WSI) is a rarely used system of building very-large integrated circuit (commonly called a "chip") networks from an entire silicon wafer to produce a single "super-chip". Combining large size and reduced packaging, WSI was ...
*
Wafer-level packaging Wafer-level packaging (WLP) is a process where packaging components are attached to an integrated circuit (IC) ''before'' the wafer – on which the IC is fabricated – is diced. In WSP, the top and bottom layers of the packaging and the sold ...
*
Semiconductor device fabrication Semiconductor device fabrication is the process used to manufacture semiconductor devices, typically integrated circuit (IC) chips such as modern computer processors, microcontrollers, and memory chips such as NAND flash and DRAM that are p ...
*
Transistor count The transistor count is the number of transistors in an electronic device (typically on a single substrate or "chip"). It is the most common measure of integrated circuit complexity (although the majority of transistors in modern microprocessors ...
* Deep learning processor


References


External links

* {{Official website, www.cerebras.net
Cerebras' presentation at Hot Chips 34 (2022)
Computer companies of the United States Companies based in California Companies based in Sunnyvale, California Companies based in Silicon Valley Semiconductor companies of the United States Fabless semiconductor companies Electronics companies established in 2016 Electronics companies of the United States