Blog

Ceramic: Deepseek Infrastructure for Every AI-Native Company

by Lila Tretikov, Madison Faulkner and James KaplanMar 05, 2025

There are three components to building generative AI models: data, compute, and talent. Today, we’re entering a paradigm shift in the market dynamics and relative importance of these three components that we believe will spawn a new wave of companies training their own models.

The old paradigm, driven by OpenAI and the principles they learned (summarized in the Bitter Lesson), was that foundation models are primarily a game of scaling compute over building human knowledge into models. Compute was king, even over data. These beliefs had two underpinning assumptions: that a general-purpose model would outperform domain-specific models and that model training is highly capital intensive. This kicked off a race that was believed to have few winners at most and a multi-billion dollar entry ticket (which is why companies like OpenAI, Anthropic, Xai, and SSI have raised capital so aggressively). 

Still, just over two years since the launch of ChatGPT, this paradigm is showing cracks:

  • Pre-training scaling laws seem to not be holding, but we’re seeing the rise of a new scaling law for test-time compute / reasoning

  • The large model providers have seemingly tapped out the data that’s available on the internet

  • Model distillation democratized the most powerful models soon after launch, most famously by DeepSeek

  • DeepSeek also showed the potential of unsupervised reinforcement learning to take a well-crafted dataset and create a model very cheaply that outperformed on specific tasks

Source: Artificial Analysis as of 2/21/2025

The implication is that we’re in a new age of model training. We went from just a few research labs leveraging scale economies on compute to train very powerful general-purpose models to a potential democratization of models with any company being able to train their own task/domain-specific models if they have a unique dataset. 

The next-generation AI infrastructure is here and it is servicing both generative AI companies and those using generative AI. We believe this new wave of model users need optimization infrastructure across data, models, training and inference to achieve DeepSeek-like performance and efficiency, which is why we were so compelled by Ceramic Founder and CEO Anna Patterson’s vision. Today, Ceramic’s platform enables 2-3x performance over DeepSeek in training. 

Simply, Ceramic wants to give every AI-native company the ability to build and optimize their infrastructure with the acceleration power of DeepSeek through partnerships with core compute providers like Lambda and AMD. Ceramic is building an enterprise stack designed for simplifying the training and fine-tuning of foundation models. The software platform’s model can train with long contexts and any cluster size, enabling enterprises to develop, train, and scale their own AI models faster than traditional methods. Ceramic massively improves GPU utilization regardless of underlying hardware (technically, for a given dataset and a given target step count or val/loss, Ceramic will use fewer TFlops and thus take less time to train). 

Data

A new approach to data preprocessing, Ceramic reranks raining data, aligning micro-batches by topic. This enhances attention efficiency and prevents token masking issues found in other models.

Training

Ceramic is the only platform that can train large models on long-context data, providing unrivaled quality and performance. The company outperforms all reported benchmarks for long-context model training, maintaining high efficiency even for 70B+ parameter models. This is done through novel parallelizations, allowing model training to scale beyond 10k+ GPUs without experiencing the diminishing marginal benefits to more GPUs. 

Reasoning Models

In testing, Ceramic trained reasoning models with their platform that achieved an exact-match 92% Pass@1 score on GSM8K, outperforming Meta’s Llama70B (79%) and DeepSeek R1 (84%) according to their internal benchmarks. 

Inference 

Ceramic believes the inference ecosystem will rapidly innovate over the next couple years and will support the complexity of deployed inference optimization. This includes ease of transitioning between training and inference workloads. 

Talent

We are excited by Ceramic’s killer team. Anna is a pioneering computer scientist in AI, a PhD, former VP of Engineering in AI at Google, and one of the industry experts in search engine infrastructure. She started at Google in 2004 and left to start Cuil, a clustering-based search engine that was acquired back into Google. In addition to TeraGoogle and Cuil, she founded Recall (now part of the Wayback Machine). Anna has also built three of the biggest search engines of all time by pages indexed. During herGoogle tenure, she was the GM of Google Books, later becoming one of the founding members of Google's AI team. Later, she founded Gradient Ventures, Google's AI-focused early-stage venture fund. Anna’s view into scaling deep learning infrastructure training and inference over the past two decades gives her the unique lens to optimize the next decade of generative training and inference. 

***

Today, we are thrilled to announce our lead investment in Ceramic’s seed financing alongside various strategic investors such as IBM and Samsung Next, and look forward to partnering with Anna and the Ceramic team as they continue working to deliver fast and cost-effective AI model training for enterprises.

About the Authors

Lila Tretikov

Lila joined NEA in 2024 as Partner, Head of AI Strategy. She most recently served as Deputy CTO at Microsoft, driving large-scale AI transformation. Previously, Lila was SVP Engie, CEO & Vice Chair Terrawatt at Engie (EPA:ENGI) where she led the company's transition to renewable energy sources. Lila was also CEO at Wikipedia Foundation & Endowment, reversing its decline, introducing AI strategy, and growing 270%. Lila has been named a Forbes' Top 100 Most Powerful Women, a World Economic Forum Young Global Leader, and a distinguished alumna of the University of California, Berkeley. She completed graduate programs at University of Oxford’s Saïd Business School and Stanford University's Directors' College.
Lila joined NEA in 2024 as Partner, Head of AI Strategy. She most recently served as Deputy CTO at Microsoft, driving large-scale AI transformation. Previously, Lila was SVP Engie, CEO & Vice Chair Terrawatt at Engie (EPA:ENGI) where she led the company's transition to renewable energy sources. Lila was also CEO at Wikipedia Foundation & Endowment, reversing its decline, introducing AI strategy, and growing 270%. Lila has been named a Forbes' Top 100 Most Powerful Women, a World Economic Forum Young Global Leader, and a distinguished alumna of the University of California, Berkeley. She completed graduate programs at University of Oxford’s Saïd Business School and Stanford University's Directors' College.

Madison Faulkner

Madison joined NEA in 2024 as a Principal on the technology team focused on early-stage data, infrastructure, developer tools, data science, and AI/ML. Previously, she was a Vice President at Costanoa Ventures where she worked closely with Delphina.ai, Probabl.ai, Mindtrip.ai, Noteable.io (acq by Confluent), Rafay.co, and others. Prior to investing, Madison was Head of Data Science and Machine Learning at Thrasio, Head of Data Science at Greycroft, and held several data science positions at Facebook. Madison received a BS in Management Science and Engineering from Stanford University.
Madison joined NEA in 2024 as a Principal on the technology team focused on early-stage data, infrastructure, developer tools, data science, and AI/ML. Previously, she was a Vice President at Costanoa Ventures where she worked closely with Delphina.ai, Probabl.ai, Mindtrip.ai, Noteable.io (acq by Confluent), Rafay.co, and others. Prior to investing, Madison was Head of Data Science and Machine Learning at Thrasio, Head of Data Science at Greycroft, and held several data science positions at Facebook. Madison received a BS in Management Science and Engineering from Stanford University.

James Kaplan

James joined NEA in 2023 as an investor on the Technology team, focused on consumer and AI apps. Prior to NEA, James spent time at early-stage startups, including GlossGenius, a PLG vertical SaaS business, and consulting with Frost Giant Studios, a Starcraft spinout game studio building the next generation of real-time strategy (RTS) games. He also spent time at Credit Suisse in its technology group. James graduated from the University of Southern California.
James joined NEA in 2023 as an investor on the Technology team, focused on consumer and AI apps. Prior to NEA, James spent time at early-stage startups, including GlossGenius, a PLG vertical SaaS business, and consulting with Frost Giant Studios, a Starcraft spinout game studio building the next generation of real-time strategy (RTS) games. He also spent time at Credit Suisse in its technology group. James graduated from the University of Southern California.