Baseten Secures $150M to Power Next-Gen AI Inference

September 7, 2025, 3:31 pm
Greylock Partners
Greylock Partners
Location: United States, California, Menlo Park
Employees: 51-200
Founded date: 1965
Spark Capital
Spark Capital
PlatformFinTechSoftwareTechnologyHealthTechProductServiceDataITBusiness
Scribble Ventures
Scribble Ventures
PlatformSoftwareFinTechProductDataEdTechITCryptoManagementSaaS
Location: United States, California, Portola Valley
Employees: 1-10
Founded date: 2020
CapitalG
CapitalG
FinTechPlatformDataSecuritySoftwareITTechnologyBusinessServiceOnline
Location: Canada
Employees: 11-50
Baseten, an AI inference powerhouse, recently closed a $150 million Series D funding round. This investment values the company at $2.15 billion. BOND led the round, with participation from CapitalG and other key investors. The capital will fuel Baseten's expansion of its developer tools and research into AI model optimization. Baseten's platform provides critical infrastructure, enabling enterprises to deploy AI models up to 50% faster, enhancing performance, reliability, and cost-efficiency. It supports both inference and training, becoming a foundational layer for the rapidly growing AI economy and addressing intense market demand.

The artificial intelligence sector sees immense capital flow. Baseten Labs Inc. now stands at the forefront. The company recently announced a significant $150 million Series D investment. This latest funding round elevates Baseten’s valuation to a striking $2.15 billion.

BOND spearheaded the Series D round. Major investors joined, including Alphabet Inc.'s CapitalG fund, Premji Invest, and Scribble Ventures. Existing backers like Conviction, 01A, IVP, Spark, and Greylock reaffirmed their support. This influx of capital boosts Baseten's total funding to over $285 million. It signals a pivotal moment for the company. Baseten scales to meet the surging demand for robust AI infrastructure.

AI adoption accelerates globally. Applications reach millions of users swiftly. Industries transform daily. Healthcare, finance, education, and entertainment all benefit. At the core of this transformation lies inference. Inference runs trained AI models in real-world applications. It bridges AI with user interaction. Predictions are made. Decisions are supported. Value is created. Here, performance, cost, and scalability become paramount.

Baseten positions itself strategically within this shift. Its platform is purpose-built for AI inference. It manages the rigorous demands of modern AI applications. These applications require reliability, efficiency, and scale. Baseten merges applied AI research with flexible infrastructure. It also provides intuitive developer tooling. This combination empowers companies. They move from AI experimentation to full production with speed and confidence.

The company’s Inference Stack forms a foundational layer. It serves both hypergrowth AI startups and established enterprise teams. It helps them deploy cutting-edge models. Performance and economics remain uncompromised. Baseten claims its technology delivers up to 50% faster performance. This outperforms competing products.

Deployment offers flexibility. Customers can deploy Baseten's software on their own infrastructure. A cloud-based version is also available. Baseten manages this cloud option. It operates across ten different infrastructure-as-a-service platforms. When AI application traffic surges, Baseten responds. It automatically provisions additional graphics cards. It selects the most suitable platform. If a public cloud experiences issues, the software switches platforms. This ensures continuous operation.

Baseten employs advanced optimization techniques. Topology-aware parallelism is one such approach. When an AI model uses multiple graphics cards, they must coordinate. Data exchanges occur over the network. Topology-aware parallelism reduces this data traffic. It lowers hardware usage. The company optimizes not just hardware. It also optimizes the AI models themselves. Operator fusion combines calculations. These calculations are usually separate. Baseten merges them into a single computation. This saves significant time. A quantization tool shrinks neural networks. This reduces their memory requirements.

Core inference features integrate with developer tools. Software teams leverage these tools. They automate parts of the AI model release process. This streamlines production deployment. Baseten also simplifies dependency deployment. These are software modules vital for a model's operation.

Once a model deploys, developers gain insights. A built-in observability tool monitors performance. The software tracks key metrics. These include request counts, response times, and hardware usage. This comprehensive monitoring ensures optimal operation.

Baseten expanded its focus in May. It moved beyond inference. The company launched an AI training service. It provides access to infrastructure. Companies use this to build new AI models. The service also enables periodic saving of AI models. This creates backups during training. If the training process halts, developers can recover the most recent backup. This avoids starting from scratch.

The new funding targets specific growth areas. Baseten will expand its lineup of developer tools. It plans deeper investment in key research areas. New methods for speeding up AI models are a priority. The company will also build out its infrastructure. This supports even larger workloads. Refinement of developer tools continues. The goal is seamless model deployment and management. Customer-facing teams will also grow. This ensures clients receive expert support. They navigate the complex AI deployment landscape effectively.

Baseten has supported leading companies. Abridge, Captions, Clay, OpenEvidence, and Writer are among them. These customers push AI boundaries. They generate billions in revenue across various sectors. These partnerships highlight inference's critical role. It ensures successful AI products. Reliance on Baseten’s infrastructure grows. It delivers consistent, high-quality experiences to end-users.

Recent product milestones reinforce Baseten's market position. The launch of Baseten Model APIs expanded capabilities. Baseten Training offered new deployment methods. These tools provide avenues for tailoring models for specific uses. They combine with the Inference Stack. This reflects a broader vision for AI infrastructure. Multiple models can collaborate. They adapt to changing conditions. They deliver exceptional performance. Reliability and cost efficiency remain paramount.

The AI industry evolves rapidly. Its supporting infrastructure must evolve too. Baseten not only keeps pace. It actively shapes this evolution. Its dedicated focus on inference is key. Its commitment to the developer experience stands out. Its ability to scale with leading AI companies makes it critical. Baseten remains a pivotal player in the future of AI. The company truly powers the next generation of artificial intelligence.