SambaNova Systems
Founded Year
2017Stage
Secondary Market | AliveTotal Raised
$1.132BMosaic Score The Mosaic Score is an algorithm that measures the overall financial health and market potential of private companies.
-16 points in the past 30 days
About SambaNova Systems
SambaNova Systems specializes in advanced artificial intelligence and machine learning with a focus on enterprise-scale AI platform development. The company offers a full stack AI solution, including hardware and software designed for generative AI, which enables the deployment of large complex foundation models that can transform business operations and unlock new insights from data. SambaNova's products cater to a variety of sectors, including financial services, healthcare, manufacturing, energy, and the public sector. It was founded in 2017 and is based in Palo Alto, California.
Loading...
SambaNova Systems's Product Videos
SambaNova Systems's Products & Differentiators
SambaNova Suite
SambaNova Suite delivers the most accurate pre-trained generative AI models, optimized for enterprise and government organizations, deployed on-premises or in the cloud, and adapted with an organization’s data for greater accuracy. Continuously updated with state-of-the-art open source models, customers retain ownership of models that have been adapted with their data and no customer data needs to be sent outside of their environment. SambaNova Suite empowers enterprises and government organizations to take advantage of the full potential of generative AI to solve their biggest business and operational challenges, while delivering the flexibility, privacy, and security required of modern technologies and tools.
Loading...
Expert Collections containing SambaNova Systems
Expert Collections are analyst-curated lists that highlight the companies you need to know in the most important technology spaces.
SambaNova Systems is included in 4 Expert Collections, including Unicorns- Billion Dollar Startups.
Unicorns- Billion Dollar Startups
1,244 items
Artificial Intelligence
14,767 items
Companies developing artificial intelligence solutions, including cross-industry applications, industry-specific products, and AI infrastructure solutions.
Conference Exhibitors
5,302 items
AI 100
100 items
SambaNova Systems Patents
SambaNova Systems has filed 215 patents.
The 3 most popular patent topics include:
- parallel computing
- instruction processing
- computer memory
Application Date | Grant Date | Title | Related Topics | Status |
---|---|---|---|---|
6/12/2023 | 9/17/2024 | Vacuum tube computers, Early computers, Instruction processing, Xbox, Telephone exchange equipment | Grant |
Application Date | 6/12/2023 |
---|---|
Grant Date | 9/17/2024 |
Title | |
Related Topics | Vacuum tube computers, Early computers, Instruction processing, Xbox, Telephone exchange equipment |
Status | Grant |
Latest SambaNova Systems News
Sep 19, 2024
“The reason they haven’t made Llama 405B available is that it would require 12 racks of hardware—12 wafers in total—making it costly for them.” Share Listen to this story The battle for token speed is intensifying as SambaNova, Cerebras, and Groq push the limits of inference performance. SambaNova recently made headlines by setting a new record for inference on Meta’s Llama 3.1 405B. The platform achieved 132 output tokens per second while running the model in native 16-bit precision. In an exclusive interview with AIM, SambaNova’s chief architect Sumti Jairath revealed how SambaNova is different from its competitors. Notably, among the three—Groq, Cerebras, and SambaNova—SambaNova is the only platform offering Llama 3.1 405B. The API inference on SambaNova cloud is powered by its SN40L custom AI chip, which features their Reconfigurable Dataflow Unit architecture. Manufactured on TSMC’s 5 nm process, the SN40L integrates DRAM, HBM3, and SRAM on each chip. Jairath explained that one of the key differences SambaNova has over Cerebras and Groq is its three levels of memory hierarchy. “If you look at Groq and Cerebras, they only have SRAM. There is no HPM, and there is no high-capacity memory.” He further explained that running the Llama 70B model on Groq requires nine racks of hardware, with each rack containing eight usable nodes and eight LPUs per node. The challenge, he explained, is having enough SRAM to accommodate 140 gigabytes of memory, which is necessary for the 70 billion parameters. He added that while Groq and Cerebras handle the 70B model, scaling up would be difficult if the parametres increase. “The reason Cerebras and Groq don’t have Llama 405B yet is that the required capacity would need to grow by another factor of five to ten. The amount of hardware needed becomes impractical,” he said. Speaking about SambaNova, he explained that they use the same rack for both Llama 70B and 405B models because of their large memory capacity. “We have 24 terabytes of DDR, one terabyte of HBM, and around eight gigabytes of SRAM. This combination allows us to manage up to 12 trillion parameters in a single box,” he said. Referring to Cerebras, he explained that they would need four racks of hardware, or wafers. “The reason they haven’t made Llama 405B available is that it would require 12 racks of hardware—12 wafers in total—making it costly for them,” he said. Would Love to Work With OpenAI SambaNova Systems recently launched a new demo on Hugging Face, offering a high-speed, open-source alternative to OpenAI’s o1 model. The demo uses Meta’s Llama 3.1 Instruct model and competes directly with OpenAI’s latest release. Jairath said that the company is willing to collaborate with any open-source model that matches the capabilities of OpenAI’s o1. Interestingly, he added that SambaNova is also willing to partner with OpenAI to run their models. “If OpenAI is interested in using our hardware and running their models behind their firewall, I see no reason why it wouldn’t deliver better performance,” he said. Jairath also pointed out that OpenAI’s models would likely benefit from SambaNova’s hardware, saying, “They are currently running on GPUs, and their model architectures are well-aligned with what SambaNova offers.” Better than Traditional GPUs Raghu Prabhakar, engineer at SambaNova Systems, explained that running a complex operation like a Transformer decoder on a GPU often involves multiple kernel launches, each with its own memory access patterns and synchronisation costs. In contrast, the dataflow architecture allows for the fusion of these operations into a single kernel call, drastically reducing overhead and increasing throughput. “On GPUs, each operator in a decoder might require multiple trips to high-performance memory (HPM), resulting in significant bandwidth and latency penalties. By grouping all operations into a single unit of execution, we eliminate these inefficiencies and achieve up to 3x to 4x better performance,” said Prabhkar. Speaking about the SN40L chip, he explained, “The chip features over half a gigabyte of SRAM on a single socket, approximately 64 gigabytes of HPM, and 1.5 terabytes of off-package DDR high-capacity memory. The system delivers 638 teraflops, which is roughly comparable to, though slightly below, the compute power of an air-cooled Hopper H100 for BF16 teraflops.” The Business Model of SambaNova Jairath said that SambaNova is a full-stack company. “We handle everything from the chip to the APIs, basically everything in between.” He explained that back in 2020, they started selling chip systems. “Our box was very much similar to what DGX boxes are from NVIDIA. Anybody can take it and then code and we still sell that. That’s a data scale product that we have,” he said. Jairath explained that their current business model caters to enterprises that want a ChatGPT-like service but behind their own firewall. “If someone wants the entire stack—chip system, all the APIs, and the model—but behind their firewall, where they can train and run the model without sending data out, that’s the business we are in,” he said. He said their goal is to offer enterprise AI solutions that allow customers to fine-tune and deploy models in the cloud. The company also provides an option for deploying these models securely behind the customer’s firewall. Moreover, he explained that in the near future with agentic AI coming in, companies won’t just deploy one model, instead there will be multiple models working together, which different departments within the company will be fine-tuning. “Enterprises will be running hundreds of these models. When they need to operate these many models, they can’t deploy a separate set of GPUs for each one, as the costs would grow exponentially,” concluded Jairath. 📣 Want to advertise in AIM? Book here
SambaNova Systems Frequently Asked Questions (FAQ)
When was SambaNova Systems founded?
SambaNova Systems was founded in 2017.
Where is SambaNova Systems's headquarters?
SambaNova Systems's headquarters is located at 2100 Geng Road, Palo Alto.
What is SambaNova Systems's latest funding round?
SambaNova Systems's latest funding round is Secondary Market.
How much did SambaNova Systems raise?
SambaNova Systems raised a total of $1.132B.
Who are the investors of SambaNova Systems?
Investors of SambaNova Systems include CrossWork, Google Ventures, Walden International, Intel Capital, BlackRock and 8 more.
Who are SambaNova Systems's competitors?
Competitors of SambaNova Systems include Groq and 6 more.
What products does SambaNova Systems offer?
SambaNova Systems's products include SambaNova Suite and 1 more.
Who are SambaNova Systems's customers?
Customers of SambaNova Systems include Riken Center for Computational Science, Argonne National Laboratory and Lawrence Livermore National Laboratory.
Loading...
Compare SambaNova Systems to Competitors
Cerebras focuses on artificial intelligence (AI) work in computer science and deep learning. The company offers a new class of computers, the CS-2, which is designed to train AI models efficiently, with applications in natural language processing (NLP), computer vision, and computing. Cerebras primarily serves sectors such as health and pharma, energy, government, scientific computing, financial services, and web and social media. It was founded in 2016 and is based in Sunnyvale, California.
Another Brain operates as a company focusing on the development of artificial intelligence (AI) technologies. It specializes in the development of Organic AI, a new generation of artificial intelligence technology within the AI industry. The company offers a vision quality control solution called Blue Phosphor that uses AI algorithms for intelligent defect detection in industrial supply chains. It was founded in 2017 and is based in Paris, France.
Groq specializes in AI inference technology within the artificial intelligence sector. The company offers a hardware and software platform that is designed to provide compute speeds and energy efficiency for AI applications. It was founded in 2016 and is based in Mountain View, California.
ChipIntelli is a leader in the intelligent voice chip industry, focusing on providing solutions for more natural, simple, and smart human-machine interactions. The company offers a range of intelligent voice chips and solutions that cater to various applications, including offline voice recognition and voice-enabled smart devices. ChipIntelli's products are primarily used in the smart home appliances, smart lighting, smart automotive, and smart education/entertainment sectors. It was founded in 2015 and is based in Chengdu, Sichuan.
BITMAIN manufactures the digital currency mining sector, specializing in mining servers. The company offers technology power efficiency and provides computational infrastructure solutions to the global blockchain network. It primarily serves the cryptocurrency mining industry. It was founded in 2013 and is based in Beijing, China.
Wave Computing is a company focused on revolutionizing artificial intelligence through its dataflow-based chips, systems, and software in the technology sector. The company's main offerings include high-performance, scalable, and power-efficient processors based on the RISC-V instruction set architecture, designed for high-end applications. These processors are primarily used in the automotive, high-performance computing and data center, and communications and networking sectors. It is based in Campbell, California.
Loading...