- New SN50 chip boasts a max speed of 5X faster than competitive chips [1]
- Run agentic AI at a 3X lower cost than GPUs โ slashing inference costs and maximizing margins [2]
- SoftBank Corp. will be the first customer to deploy SN50 within its nextโgeneration AI data centers in Japan
- SambaNova, Intel plan multi-year strategic collaboration to deliver cloud-scale AI inference to unlock multi-billion-dollar market opportunity
- $350 million in strategic Series E financing to expand manufacturing and cloud capacity; new investors include Vista Equity Partners, Cambium Capital, First Data, Intel Capital, Battery Ventures, and accounts advised by T. Rowe Price Associates, Inc.
DUBAI, United Arab Emirates–(BUSINESS WIRE)–#AI—SambaNova today introduced their SN50 AI chip, which boasts a max speed thatโs 5X faster than competitive chips. The company also announced a planned collaboration with Intel to deliver highโperformance, costโefficient AI inference solutions, and more than $350M in investment from new and existing investors.
Positioned as the most efficient chip for agentic AI, the SN50 chip offers enterprises a 3X lower total cost of ownership โ a powerful foundation to scale fast inference and bring autonomous AI agents into full production. The SN50 will be shipping to customers later this year.
To quickly scale and distribute SN50, SambaNova is collaborating with Intel, and has obtained $350 million in strategic Series E financing to expand manufacturing and cloud capacity.
โAI is no longer a contest to build the biggest model,โ said Rodrigo Liang, coโfounder and CEO of SambaNova. โWith the SN50 and our deep collaboration with Intel, the real race is about who can light up entire data centers with AI agents that answer instantly, never stall, and do it at a cost that turns AI from an experiment into the most profitable engine in the cloud.โ
โCustomers are asking for more choice and more efficient ways to scale AI,โ said Kevork Kechichian, EVP, General Manager, Data Center Group, Intel. โBy combining Intelโs leadership in compute, networking, and memory with SambaNovaโs full-stack AI systems and inference cloud platform, we are delivering a compelling option for organizations looking for GPU alternatives to deploy advanced AI at scale.โ
The SN50 delivers five times more compute per accelerator and four times more network bandwidth than the previous generation. It links up to 256 accelerators over a multiโterabyteโperโsecond interconnect, cutting timeโtoโfirstโtoken and supporting larger batch sizes. The result: Enterprises can deploy bigger, longerโcontext AI models with higher throughput and responsiveness โ while keeping performance high and costs and latency under control.
โAI is moving from a software story to an infrastructure story,โ said Landon Downs, co-founder and managing partner at Cambium Capital. โSN50 is engineered for the real-world latency and economic requirements that will determine who successfully deploys agentic AI at scale.โ
The news follows SambaNovaโs record bookings and revenue as they closed out 2025, reflecting accelerating demand for production-ready AI systems across financial services, telecommunications, energy, and sovereign deployments worldwide.
Built for Agentic Production
Built on SambaNovaโs Reconfigurable Data Unit (RDU) architecture, SN50 delivers:
- Instant AI Experiences โ Ultraโlow latency delivers realโtime responsiveness for nextโgen enterprise apps like voice assistants.
- Unmatched Scale and Concurrency โ Power thousands of simultaneous AI sessions with consistent high performance.
- Breakthrough Model Capacity โ Threeโtier memory architecture unlocks 10T+ parameter models and 10M+ context lengths for deeper reasoning and richer outputs.
- Maximum Efficiency at Scale โ Higher hardware utilization lowers costโperโtoken, driving greater performance and ROI.
- Smarter Memory, Smarter Efficiency โ Resident multiโmodel memory and agentic caching optimize the threeโtier architecture, cutting infrastructure costs for enterpriseโscale AI deployments.
โThe new SambaNova SN50 RDU changes the tokenomics of AI inference at scale. By delivering both high performance and high throughput with a chip that uses existing power and is air cooled, SambaNova is changing the game,โ says Peter Rutten, Research Vice-President Performance Intensive Computing at analyst firm IDC.
SoftBank Deploys SN50 within its AI Data Centers in Japan
SoftBank Corp. will be the first customer to deploy SN50 within its nextโgeneration AI data centers in Japan. The deployment will power lowโlatency inference services for sovereign and enterprise customers across AsiaโPacific, supporting both openโsource and proprietary frontier models with aggressive latency and throughput requirements.
โWith SN50, we are building an AI inference fabric for Japan that can serve our customers and partners with the speed, resiliency and sovereignty they expect from SoftBank,โ said Hironobu Tamba, Vice President and Head of the Data Platform Strategy Division of the Technology Unit at SoftBank Corp. โBy standardizing on SN50, we gain the ability to deliver worldโclass AI services on our own terms โ with the performance of the best GPU clusters, but with far better economics and control.โ
The SN50 deployment deepens SambaNovaโs existing relationship with SoftBank Corp., which already hosts SambaCloud to provide ultraโfast inference for developers in the region. By anchoring its newest clusters on SN50, SoftBank positions SambaNova as the inference backbone for its sovereign AI initiatives and future largeโscale agentic services.
SambaNova and Intel Plan MultiโYear Collaboration
SambaNova and Intel have entered into a planned multiโyear strategic collaboration to deliver highโperformance, costโefficient AI inference solutions for AIโnative companies, model providers, enterprises, and government organizations around the world. The collaboration will give customers a powerful alternative to GPUโcentric solutions, offering optimized performance for leading openโsource models with predictable throughput and total cost of ownership.
As part of the collaboration, Intel plans to make a strategic investment in SambaNova to accelerate the rollout of an Intelโpowered AI cloud. The collaboration is expected to span three key areas:
- AI Cloud Expansion โ Scaling SambaNovaโs vertically integrated AI cloud, built on Intel Xeonโbased infrastructure and optimized for large language and multimodal models. The platform will deliver lowโlatency, highโthroughput AI services, supported by reference architectures, deployment blueprints, and partnerships with system integrators and software vendors.
- Integrated AI Infrastructure โ Combining SambaNovaโs systems with Intelโs CPUs, accelerators, and networking technologies to power scalable, productionโready inference for reasoning, code generation, multimodal applications, and agentic workflows.
- GoโtoโMarket Execution โ Joint coโselling and coโmarketing through Intelโs global enterprise, cloud, and partner channels to accelerate adoption across the AI ecosystem.
Together, SambaNova and Intel aim to shape the next generation of heterogeneous AI data centers โ integrating Intel Xeon processors, Intel GPUs, Intel networking and storage, and SambaNova systems โ to unlock a multiโbillionโdollar inference market opportunity.
Raises $350M+, led by Vista and Cambium
The oversubscribed Series E round was led by Vista Equity Partners and Cambium Capital, with strong participation from Intel Capital.
New investors joining the round include: Assam Ventures, Battery Ventures, First Data, Gulf Development Public Company Limited, Mayfield Capital, Saudi First Data, Seligman Ventures, and accounts advised by T. Rowe Price Associates, Inc. Existing investors participating include: A&E, 8Square, Atlantic Bridge, BlackRock, GV, Nepenthe, Nuri Capital, and Redline Capital.
Mr. Sharaf Al Hariri, Chairman of First Data, stated that SambaNova forms a core part of First Dataโs investment strategy to bring advanced AI technologies to the Kingdom of Saudi Arabia and the wider Middle East region. He emphasized that First Data is investing in SambaNovaโs platforms to enable high-performance, low-latency, and sovereign-ready AI capabilities, while benefiting from low power requirements and the ability to operate efficiently within existing air-cooled data center environments. Within this strategy, SambaNova supports advanced inference and scalable AI services, strengthening First Dataโs ability to deliver world-class AI infrastructure and solutions. He further noted that this investment reflects First Dataโs broader, long-term commitment to diversifying its technology portfolio, building resilient innovation capabilities, and positioning the company to adapt to evolving global technology trends while delivering sustainable value to the region.
As agentic workloads expand, enterprises are discovering that infrastructure optimized for training struggles to meet production latency and cost requirements: โWeโre proud to be investing in SambaNova at such a pivotal time in the companyโs growth,โ said Monti Saroya, Partner at Vista Capital. โSN50 is engineered for agentic AI systems that orchestrate multiple models and process requests in near real-time more efficiently than traditional GPU-centric systems.โ
Proceeds will be used to expand SN50 production, scale SambaCloud, and deepen enterprise software integrations.
About SambaNova
SambaNova is a leader in nextโgeneration AI infrastructure, providing a full stack platform that powers the fastest, most efficient AI inference for enterprises, NeoClouds, AI labs and service providers, and sovereign AI initiatives worldwide. Founded in 2017 and headquartered in San Jose, Calif., SambaNova delivers chips, systems and cloud services that enable customers to deploy stateโofโtheโart models with superior performance, lower total cost of ownership and rapid time to value.
For more information, visit sambanova.ai or follow SambaNova on X and LinkedIn.
[1] SemiAnalysis InferenceX – Llama 3.3 70B max speed on Nvidia B200 at FP8 and 1K input/1k output – 184 tokens per second per user. Llama 3.3 70B max speed on SN50 at FP8 and 1K input/1k output – 895 tokens per second per user.
[2] SemiAnalysis InferenceX – Llama 3.3 70B throughput per chip on Nvidia B200 at FP8 and 1K input/1k output – across a range of configurations, total throughput per GPU versus total throughput per RDU moves from ~1X (at 33 tokens per second per user) to ~25X advantage for RDUs (at 184 tokens per second per user). 3X is derived as the average throughput advantage for SN50 across Llama 70B, GPT-OSS 120B and DeepSeek 671B, assuming a latency budget.
Contacts
Press Contact:
Emad Abdo
[email protected]
+971522038872




