NVIDIA CEO Jensen Huang's Vision: GPU, AI, Robotics & Omniverse Complete Guide


YouTube Thumbnail
NVIDIA CEO Jensen Huang reveals the vision for GPU, deep learning, robotics & Omniverse—the future of computing for the next decade
Summary
In Cleo Abram’s ‘Huge Conversations,’ Jensen Huang traces the journey from GPUs and CUDA to deep learning revolution, Omniverse, Cosmos, robotics, digital biology, energy efficiency and AI safety, explaining where computing and robotics are heading in the next 10 years. A complete guide that unfolds a 30-year vision in one conversation.

1. The goal of the conversation

Host Cleo Abram frames the interview as “hearing what a key architect of the future is imagining,” declaring she will ask Jensen three things: how we got here, what is happening right now, and what he’s trying to build next.

The audience ranges from teenagers who might not know the difference between CPU and GPU to experts in various fields, so the format is designed as a “joint explainer that even complete beginners can understand about the future of computing.”

2. The birth of GPUs: observing parallel processing

Jensen Huang recalls that when founding NVIDIA in the early 1990s, they observed that “10% of software code accounts for over 90% of total computation, and that 90% can be processed in parallel.”

“The remaining 90% of code must be processed sequentially, so the ideal computer must strongly support both sequential and parallel processing. NVIDIA started with the goal of building computers that accelerate problems traditional CPUs couldn’t solve.”

3. Why start with video games?

3D graphics required massive amounts of identical or similar computations performed simultaneously—a typical parallel processing problem. In the 90s, game developers wanted more realistic graphics, but existing hardware performance couldn’t keep up.

Jensen says “games were applications we loved, and simulating virtual worlds perfectly matched parallel processing.” He believed the gaming market would become huge, which would sustain the massive R&D needed for complex GPU architectures.

4. Why GPUs are ‘time machines’

Jensen describes GPUs as “time machines that let you see the future faster,” sharing a story of a quantum chemistry researcher who said “thanks to your GPU, I can complete my life’s research within my lifetime.”

From weather prediction and autonomous driving simulation to urban traffic simulation, GPU acceleration enables calculations that would otherwise be impossible or take decades to perform, making it possible to “see the future in advance.”

5. CUDA: opening parallel computing to everyone

Early researchers had to trick GPUs into thinking scientific workloads were graphics problems or use complex workarounds. The Mass General research team using graphics processors for medical imaging CT reconstruction provided internal inspiration.

NVIDIA created the CUDA platform to let developers command GPUs using familiar languages like C. Jensen says he could bet the whole company on CUDA because “thanks to the video game market, our GPUs would become the world’s best-selling parallel processors.”

6. AlexNet and the deep learning transformation

In 2012, the University of Toronto’s AlexNet team used NVIDIA GeForce GTX 580 with CUDA to overwhelmingly improve image recognition competition performance, transforming GPUs from simple accelerators into engines of an entirely new computational paradigm.

“Inside NVIDIA, we were frustrated with computer vision too. When we saw a completely different approach—deep neural networks on GPUs outperforming existing algorithms—we started redesigning around ‘how far can this method go, how can it transform the entire computer industry.'”

7. Reinventing the entire computing stack

He explains, “If deep learning architectures can scale sufficiently, most machine learning problems can be expressed, and the range of problems we can solve becomes wide enough to completely reshape the computer industry.” This led to the decision to restructure hardware, systems, and software entirely.

The result was AI supercomputer platforms like DGX, building a new “accelerated computing“-centered paradigm after 60+ years of general-purpose computing models maintained since IBM System/360.

8. Why maintain the big bet for a decade

When Cleo asks “why did it take 10 years after 2012 for people to actually experience AI and NVIDIA,” he answers that throughout those 10 years, success and uncertainty were mixed, but “as long as fundamental assumptions, physical laws, and industrial insights don’t change, there’s no reason to change beliefs.”

Even when investors wanted short-term profits, they invested billions of dollars upfront because of “belief in our future.” He says he’s had the same belief in NVIDIA for over 30 years, so he continues forward.

9. Core beliefs: accelerated computing and scalable AI

Jensen’s first core belief is that “accelerated computing combining general-purpose computing + parallel accelerated computing is the basic model for the future,” and this principle remains unchanged today.

The second belief is that deep neural networks have amazing ability to learn patterns and relationships from data, and the scaling law showing they learn more knowledge as networks and data grow has been verified, with no physical or mathematical limits yet visible.

10. Multimodal AI: text, images, proteins

He explains that deep learning can now handle not only image recognition, speech recognition, and language understanding but also “text→text summarization, language translation, text→image generation, image→text captioning, amino acid sequence→protein structure prediction”—transforming various data types into each other.

In the future, “protein→text description,” “designing new proteins with desired properties,” “words→video,” “words→robot action tokens” will naturally become possible, opening AI application opportunities to almost every industry and job.

11. From science to ‘AI application science’ era

While the past decade was about advancing AI science itself, the next decade will be the “AI application science” era, he defines. Applying AI to real industries and problems like digital biology, climate/energy, agriculture/fisheries, robotics, education, logistics, and media will become central.

He adds, “Because the speed and scope of change we’re experiencing are growing, we’re entering an era where accurately predicting future use cases we see now is much more difficult.”

12. Omniverse, Cosmos and robots’ ‘world models’

Using ChatGPT as an example, Jensen explains that first-generation models generated impressive sentences but had ‘hallucination’ problems when lacking knowledge or answering at length, and later versions improved accuracy by conditioning models with ‘ground truth’ like PDFs or search results.

Physical world robots similarly need ‘world foundation models’ embodying physical common sense like gravity, friction, inertia, object permanence, and causality. NVIDIA is combining Cosmos as a world model with physics simulator Omniverse to generate and learn from infinite scenarios based on physical laws.

13. “Everything that moves will become robotic”

He says “everything that moves will eventually become robotic, and that time isn’t far off.” Almost everything that moves, like lawn mowers or cars, will be automated, and humanoid robot technology will soon reach practical levels.

“Robots will learn by experiencing countless virtual futures in Omniverse and Cosmos before being deployed to the real world. Individuals will have their own AI companion (like R2-D2) that grows with them throughout life, accessible anywhere—smart glasses, phones, PCs, cars, homes.”

Practical Q&A · CEO Interview Essentials

Q: How do we manage AI risks and safety?
“Beyond well-known risks like bias, toxic speech, hallucination, fake information, and digital impersonation, cases like autonomous vehicles where intentions are good but sensor failures or judgment errors could harm actual safety are important engineering challenges. Like aircraft having triple flight computers, two pilots, and air traffic control layers, AI must be designed with community architecture where models, systems, humans, regulations, and monitoring work together.”
Q: What are the limits? Energy and efficiency?
“Ultimately, everything comes down to how much work can be done within limited energy. The energy needed to flip and move bits is the physical limit. Since the DGX-1 delivered to OpenAI in 2016, NVIDIA has improved AI computing energy efficiency by about 10,000×, and today’s compact AI supercomputers deliver 6× more performance than past equipment with far less power.”
Q: Hardware design: generality vs. specialization?
“I don’t believe transformers will be the last AI architecture. Looking at software and algorithm history, no single idea has ever dominated forever. Because researchers and developers continue exploring new structures and attention mechanisms, it’s important to maintain sufficiently flexible acceleration architectures.”
Q: Jobs and personal preparation?
“While AI will greatly change how we work, people and companies that invest in themselves, learn new tools, and move to more creative and productive roles will have tremendous opportunities. People who use AI in coding, science, design, and content creation will gain greater influence like ‘superhumans.'”

5 Key CEO Insights

  • Accelerated computing (CPU + GPU) started from observations 30 years ago and remains an unchanged core principle today
  • Deep learning‘s scaling laws are verified, with no physical limits yet visible
  • Omniverse and Cosmos are key platforms for teaching robots physical world common sense
  • The future where everything that moves becomes robotic will be reality within 5-10 years
  • AI safety isn’t about single models but community architecture with layered systems, humans, and regulations

Conclusion—What Jensen Huang wants to leave behind

Finally, regarding how he wants to be remembered, he says not simply for company value or stock price, but as someone who “expanded human potential and created computing platforms that helped scientists and engineers complete their life’s work within their lifetime.”

This interview provides a complete guide that surveys the 30-year arc from GPU, CUDA, deep learning to Omniverse and Cosmos, plus Jensen Huang‘s long-term vision of how the physical and digital worlds will combine through AI over the next 10 years. It clearly shows NVIDIA isn’t just a GPU manufacturer but a company redesigning all future computing, providing essential insights for everyone preparing for the AI, robotics, and digital biology era.