o3 and o4-mini, Google Gemini on-prem and NVIDIA’s US chip manufacturing

Episode 51: o3 and o4-mini, Google Gemini on-prem and NVIDIA’s US chip manufacturing

OpenAI just dropped o3 and o4-mini. In episode 51 of Mixture of Experts, host Tim Hwang is joined by Chris Hay, Vyoma Gajjar and special guest John Willis, Owner of Botchagalupe Technologies. We analyze Sam Altman’s new AI models, o3 and o4-mini.

Next, Google announced that by Q3 you can run Gemini on-premises. What does this mean for enterprise AI adoption? Then, John takes us through AI evaluation tools and why we need them. Finally, NVIDIA is planning to move AI chip manufacturing to the US. Can they pull this off? 
 

Key takeaways:

  • 00:00 – Intro  
  • 00:56 – OpenAI o3 and o4 mini
  • 14:57 – Google Gemini on-prem
  • 23:43 – AI evaluation tools
  • 34:59 – NVIDIA's US chip manufacturing

The opinions expressed in this podcast are solely those of the participants and do not necessarily reflect the views of IBM or any other organization or entity.

Listen on Apple podcasts Spotify podcasts YouTube Casted

Learn more about AI

What is artificial intelligence (AI)?

Applications and devices equipped with AI can see and identify objects. They can understand and respond to human language. They can learn from new information and experience. But what is AI?

What is fine-tuning?

It has become a fundamental deep learning technique, particularly in the training process of foundation models used for generative AI. But what is fine-tuning and how does it work?

How to build an AI-powered multimodal RAG system with Docling and Granite?

In this tutorial, you will use IBM’s Docling and open-source IBM® Granite® vision, text-based embeddings and generative AI models to create a retrieval augmented generation (RAG) system.

Stay on top of the AI news with our experts

Follow us on Apple Podcasts and Spotify.

Subscribe to our playlist on YouTube