Why GPUs Are Great for AI

Why GPUs Are Great for AI

GPUs have been called the rare Earth metals — even the gold — of artificial intelligence, because they’re foundational for today’s generative AI era.

Three technical reasons, and many stories, explain why that’s so. Each reason has multiple facets well worth exploring, but at a high level:

  • GPUs employ parallel processing.
  • GPU systems scale up to supercomputing heights.
  • The GPU software stack for AI is broad and deep.

The net result is GPUs perform technical calculations faster and with greater energy efficiency than CPUs. That means they deliver leading performance for AI training and inference as well as gains across a wide array of applications that use accelerated computing.

In its recent report on AI, Stanford’s Human-Centered AI group provided some context. GPU performance “has increased roughly 7,000 times” since 2003 and price per performance is “5,600 times greater,” it reported.

Stanford report on GPU performance increases
A 2023 report captured the steep rise in GPU performance and price/performance.

The report also cited analysis from Epoch, an independent research group that measures and forecasts AI advances.

“GPUs are the dominant computing platform for accelerating machine learning workloads, and most (if not all) of the biggest models over the last five years have been trained on GPUs … [they have] thereby centrally contributed to the recent progress in AI,” Epoch said on its site.

A 2020 study assessing AI technology for the U.S. government drew similar conclusions.

“We expect [leading-edge] AI chips are one to three orders of magnitude more cost-effective than leading-node CPUs when counting production and operating costs,” it said.

NVIDIA GPUs have increased performance on AI inference 1,000x in the last ten years, said Bill Dally, the company’s chief scientist in a keynote at Hot Chips, an annual gathering of semiconductor and systems engineers.

ChatGPT Spread the News

ChatGPT provided a powerful example of how GPUs are great for AI. The large language model (LLM), trained and run on thousands of NVIDIA GPUs, runs generative AI services used by more than 100 million people.

Since its 2018 launch, MLPerf, the industry-standard benchmark for AI, has provided numbers that detail the leading performance of NVIDIA GPUs on both AI training and inference.

For example, NVIDIA Grace Hopper Superchips swept the latest round of inference tests. NVIDIA TensorRT-LLM, inference software released since that test, delivers up to an 8x boost in performance and more than a 5x reduction in energy use and total cost of ownership. Indeed, NVIDIA GPUs have won every round of MLPerf training and inference tests since the benchmark was released in 2019.

In February, NVIDIA GPUs delivered leading results for inference, serving up thousands of inferences per second on the most demanding models in the STAC-ML Markets benchmark, a key technology performance gauge for the financial services industry.

A RedHat software engineering team put it succinctly in a blog: “GPUs have become the foundation of artificial intelligence.”

AI Under the Hood

A brief look under the hood shows why GPUs and AI make a powerful pairing.

An AI model, also called a neural network, is essentially a mathematical lasagna, made from layer upon layer of linear algebra equations. Each equation represents the likelihood that one piece of data is related to another.

For their part, GPUs pack thousands of cores, tiny calculators working in parallel to slice through the math that makes up an AI model. This, at a high level, is how AI computing works.

Highly Tuned Tensor Cores

Over time, NVIDIA’s engineers have tuned GPU cores to the evolving needs of AI models. The latest GPUs include Tensor Cores that are 60x more powerful than the first-generation designs for processing the matrix math neural networks use.

In addition, NVIDIA Hopper Tensor Core GPUs include a Transformer Engine that can automatically adjust to the optimal precision needed to process transformer models, the class of neural networks that spawned generative AI.

Along the way, each GPU generation has packed more memory and optimized techniques to store an entire AI model in a single GPU or set of GPUs.

Models Grow, Systems Expand

The complexity of AI models is expanding a whopping 10x a year.

The current state-of-the-art LLM, GPT4, packs more than a trillion parameters, a metric of its mathematical density. That’s up from less than 100 million parameters for a popular LLM in 2018.

Chart shows 1,000x performance improvement on AI inference over a decade for single GPUs
In a recent talk at Hot Chips, NVIDIA Chief Scientist Bill Dally described how single-GPU performance on AI inference expanded 1,000x in the last decade.

GPU systems have kept pace by ganging up on the challenge. They scale up to supercomputers, thanks to their fast NVLink interconnects and NVIDIA Quantum InfiniBand networks.

For example, the DGX GH200, a large-memory AI supercomputer, combines up to 256 NVIDIA GH200 Grace Hopper Superchips into a single data-center-sized GPU with 144 terabytes of shared memory.

Each GH200 superchip is a single server with 72 Arm Neoverse CPU cores and four petaflops of AI performance. A new four-way Grace Hopper systems configuration puts in a single compute node a whopping 288 Arm cores and 16 petaflops of AI performance with up to 2.3 terabytes of high-speed memory.

And NVIDIA H200 Tensor Core GPUs announced in November pack up to 288 gigabytes of the latest HBM3e memory technology.

Software Covers the Waterfront

An expanding ocean of GPU software has evolved since 2007 to enable every facet of AI, from deep-tech features to high-level applications.

The NVIDIA AI platform includes hundreds of software libraries and apps. The CUDA programming language and the cuDNN-X library for deep learning provide a base on top of which developers have created software like NVIDIA NeMo, a framework to let users build, customize and run inference on their own generative AI models.

Many of these elements are available as open-source software, the grab-and-go staple of software developers. More than a hundred of them are packaged into the NVIDIA AI Enterprise platform for companies that require full security and support. Increasingly, they’re also available from major cloud service providers as APIs and services on NVIDIA DGX Cloud.

SteerLM, one of the latest AI software updates for NVIDIA GPUs, lets users fine tune models during inference.

A 70x Speedup in 2008

Success stories date back to a 2008 paper from AI pioneer Andrew Ng, then a Stanford researcher. Using two NVIDIA GeForce GTX 280 GPUs, his three-person team achieved a 70x speedup over CPUs processing an AI model with 100 million parameters, finishing work that used to require several weeks in a single day.

“Modern graphics processors far surpass the computational capabilities of multicore CPUs, and have the potential to revolutionize the applicability of deep unsupervised learning methods,” they reported.

Picture of Andrew Ng showing slide in a talk on GPU performance for AI
Andrew Ng described his experiences using GPUs for AI in a GTC 2015 talk.

In a 2015 talk at NVIDIA GTC, Ng described how he continued using more GPUs to scale up his work, running larger models at Google Brain and Baidu. Later, he helped found Coursera, an online education platform where he taught hundreds of thousands of AI students.

Ng counts Geoff Hinton, one of the godfathers of modern AI, among the people he influenced. “I remember going to Geoff Hinton saying check out CUDA, I think it can help build bigger neural networks,” he said in the GTC talk.

The University of Toronto professor spread the word. “In 2009, I remember giving a talk at NIPS [now NeurIPS], where I told about 1,000 researchers they should all buy GPUs because GPUs are going to be the future of machine learning,” Hinton said in a press report.

Fast Forward With GPUs

AI’s gains are expected to ripple across the global economy.

A McKinsey report in June estimated that generative AI could add the equivalent of $2.6 trillion to $4.4 trillion annually across the 63 use cases it analyzed in industries like banking, healthcare and retail. So, it’s no surprise Stanford’s 2023 AI report said that a majority of business leaders expect to increase their investments in AI.

Today, more than 40,000 companies use NVIDIA GPUs for AI and accelerated computing, attracting a global community of 4 million developers. Together they’re advancing science, healthcare, finance and virtually every industry.

Among the latest achievements, NVIDIA described a whopping 700,000x speedup using AI to ease climate change by keeping carbon dioxide out of the atmosphere (see video below). It’s one of many ways NVIDIA is applying the performance of GPUs to AI and beyond.

Learn how GPUs put AI into production.

Read More

‘Call of Duty’ Comes to GeForce NOW

‘Call of Duty’ Comes to GeForce NOW

Let the games begin — this GFN Thursday brings the highly anticipated Call of Duty: Modern Warfare III to the cloud, the first Activision title on GeForce NOW as part of the NVIDIA and Microsoft partnership.

It’s joined by Call of Duty: Modern Warfare II and Call of Duty: Warzone — all three titles can be played from one central location via the Call of Duty logo on GeForce NOW.

And it’s the most wonderful time of the year — over 65 games are joining the GeForce NOW library in December, with 15 available to stream this week.

Plus, stream GeForce NOW on the go and get console-quality controls by simply snapping a mobile device into a Backbone One controller. For a limited time, Backbone is offering a 30% discount for premium GeForce NOW members starting today in the Rewards Portal. Free-level members can claim the discount starting Dec. 7.

The Lobby Awaits

Call of Duty on GeForce NOW
The war has changed.

Call of Duty: Modern Warfare III returns as a direct sequel to the record-breaking Call of Duty: Modern Warfare II and follows the story of Task Force 141 as they face off the ultimate threat.

Dig into the action-packed single-player campaign or head online to defeat the undead in an exciting open-world co-op experience that takes the Zombies mode that fans know and love to the next level. Those that prefer some multiplayer action can dip into a selection of Core Multiplayer maps from the 16 iconic launch maps of 2009’s Call of Duty: Modern Warfare 2 that are being brought over and modernized for Call of Duty: Modern Warfare III.

Plus, stay tuned to GFN Thursday for when other legacy Call of Duty titles as well as additional supported devices (Android, SHIELD TV and TV) will be added to the cloud. Check out the article for more details.

GeForce NOW Ultimate members can get the upper hand with NVIDIA DLSS 3 and Reflex to get the highest frame rates and lowest latencies for the smoothest gameplay by streaming from a GeForce RTX 4080 gaming rig in the cloud. Never worry about upgrading hardware or system specs again with GeForce NOW.

Presents, Galore

SteamWorld Build on GeForce NOW
Dig, dig, dig!

Break ground in SteamWorld Build from Thunderful Publishing. Dig deep and build wide to excavate long-lost spacefaring technology while ensuring everyone has the resources needed to survive and reach the final frontier. It launches Dec. 1 on Steam and PC Game Pass — check it out with the three free months of PC Game Pass included with the purchase of a six-month Ultimate membership, part of the GeForce NOW holiday bundle.

Members can start their adventures now with 15 newly supported titles in the cloud this week:

  • Last Train Home (New release on Steam, Nov. 28)
  • Gangs of Sherwood (New release on Steam, Nov. 30)
  • SteamWorld Build (New release on Steam, Xbox and available on PC Game Pass, Dec. 1)
  • Astrea: Six-Sided Oracles (Steam)
  • Call of Duty HQ, including Call of Duty: Modern Warfare III, Call of Duty: Modern Warfare II and Call of Duty: Warzone (Steam)
  • Galactic Civilizations IV (Steam)
  • Halls of Torment (Steam)
  • Kona II: Brume (Steam)
  • Laika: Aged Through Blood (Epic Games Store)
  • Pillars of Eternity (Xbox, available on PC Game Pass)
  • RESEARCH and DESTROY (Xbox, available on PC Game Pass)
  • Roboquest (Epic Games Store)
  • StrangerZ (Steam)

Then, check out the plentiful games for the rest of December:

  • Stargate: Timekeepers (New release on Steam, Dec. 12)
  • Pioneers of Pagonia (New release on Steam, Dec. 13)
  • House Flipper 2 (New release on Steam, Dec. 14)
  • Soulslinger: Envoy of Death (New release on Steam, Dec. 14)
  • Agatha Christie – Murder on the Orient Express (Steam)
  • Age of Wonders 4 (Xbox, available on the Microsoft Store)
  • AI: THE SOMNIUM FILES – nirvanA Initiative (Xbox, available on the Microsoft Store)
  • The Anacrusis (Xbox, available on the Microsoft Store)
  • BEAST (Steam)
  • Before We Leave (Xbox, available on the Microsoft Store)
  • Bloons TD Battles (Steam)
  • Control (Xbox, available on the Microsoft Store)
  • Dark Envoy (Steam)
  • Darksiders III (Xbox, available on the Microsoft Store)
  • The Day Before (Steam)
  • Destroy All Humans! (Xbox, available on the Microsoft Store)
  • Disgaea 4 Complete+ (Xbox, available on the Microsoft Store)
  • Escape the Backrooms (Steam)
  • Europa Universalis IV (Xbox, available on the Microsoft Store)
  • Evil Genius 2: World Domination (Xbox, available on the Microsoft Store)
  • Fae Tactics (Xbox, available on the Microsoft Store)
  • Figment 2: Creed Valley (Epic Games Store)
  • The Forgotten City (Xbox, available on the Microsoft Store)
  • Human Fall Flat (Xbox, available on PC Game Pass)
  • Ikonei Island: An Earthlock Adventure (Steam)
  • Immortal Realms: Vampire Wars (Xbox, available on the Microsoft Store)
  • Lethal League Blaze (Xbox, available on the Microsoft Store)
  • Loddlenaut (Steam)
  • Matchpoint – Tennis Championships (Xbox, available on the Microsoft Store)
  • Maneater (Xbox, available on the Microsoft Store)
  • The Medium (Xbox, available on the Microsoft Store)
  • Metro Exodus (Xbox, available on the Microsoft Store)
  • Mortal Shell (Xbox, available on the Microsoft Store)
  • MotoGP 20 (Xbox, available on the Microsoft Store)
  • Moving Out (Xbox, available on the Microsoft Store)
  • MUSYNX (Xbox, available on the Microsoft Store)
  • Nova-Life: Amboise (Steam)
  • Observer System Redux (Xbox, available on the Microsoft Store)
  • Pathologic 2 (Xbox, available on the Microsoft Store)
  • The Pedestrian (Xbox, available on the Microsoft Store)
  • Primal Carnage Extinction (Steam)
  • Recompile (Xbox, available on the Microsoft Store)
  • RESEARCH and DESTROY (Xbox, available on PC Game Pass)
  • RIDE 5 (Epic Games Store)
  • Sable (Xbox, available on the Microsoft Store)
  • The Smurfs 2 – The Prisoner of the Green Stone (Steam)
  • SpellForce 3: Soul Harvest (Xbox, available on the Microsoft Store)
  • Tainted Grail: Conquest (Xbox, available on the Microsoft Store)
  • Terminator: Dark Fate – Defiance (Steam)
  • Tintin Reporter – Cigars of the Pharaoh (Steam)
  • Universe Sandbox (Steam)
  • Warhammer 40,000: Rogue Trader (Steam)
  • World War Z: Aftermath (Xbox, available on the Microsoft Store)
  • Worms Rumble (Xbox, available on the Microsoft Store)
  • Worms W.M.D (Xbox, available on the Microsoft Store)

Nicely Done in November

On top of the 54 games announced in October, an additional 23 joined the cloud last month, including this week’s additions: Astrea: Six-Sided Oracles, Galactic Civilizations IV, Halls of Torment, Kona II: Brume, Laika: Aged Through Blood (Epic Games Store), Pillars of Eternity and SteamWorld Build.

  • Car Mechanic Simulator 2021 (Xbox, available on PC Game Pass)
  • Chivalry 2 (Xbox, available on PC Game Pass)
  • Disney Dreamlight Valley (Xbox, available on PC Game Pass)
  • Dungeons 4 (Epic Games Store)
  • Hello Neighbor 2 (Xbox, available on PC Game Pass)
  • The Invincible (Epic Games Store)
  • KarmaZoo (New release on Steam, Nov. 14)
  • Planet of Lana (Xbox, available on PC Game Pass)
  • Q.U.B.E. 10th Anniversary (Epic Games Store)
  • RoboCop: Rogue City (New release on Epic Games Store)
  • Roboquest (Xbox, available on PC Game Pass)
  • Rune Factory 4 Special (Xbox and available on PC Game Pass)
  • Saints Row IV: Re-Elected (Xbox, available on Microsoft Store)
  • State of Decay: Year-One Survival Edition (Steam)
  • Supraland: Six Inches Under (Xbox, available on PC Game Pass)
  • Turnip Boy Commits Tax Evasion (Epic Games Store)

Veiled Experts will no longer be coming to the service due to the closure of its live services, and Spirttea (PC Game Pass) didn’t make it to GeForce NOW in November due to technical issues. Stay tuned to GFN Thursday for future updates.

What are you planning to play this weekend? Let us know on Twitter or in the comments below.

Read More

Embracing Transformation: AWS and NVIDIA Forge Ahead in Generative AI and Cloud Innovation

Embracing Transformation: AWS and NVIDIA Forge Ahead in Generative AI and Cloud Innovation

Amazon Web Services and NVIDIA will bring the latest generative AI technologies to enterprises worldwide.

Combining AI and cloud computing, NVIDIA founder and CEO Jensen Huang joined AWS CEO Adam Selipsky Tuesday on stage at AWS re:Invent 2023 at the Venetian Expo Center in Las Vegas.

Selipsky said he was “thrilled” to announce the expansion of the partnership between AWS and NVIDIA with more offerings that will deliver advanced graphics, machine learning and generative AI infrastructure.

The two announced that AWS will be the first cloud provider to adopt the latest NVIDIA GH200 NVL32 Grace Hopper Superchip with new multi-node NVLink technology, that AWS is bringing NVIDIA DGX Cloud to AWS, and that AWS has integrated some of NVIDIA’s most popular software libraries.

Huang started the conversation by highlighting the integration of key NVIDIA libraries with AWS, encompassing a range from NVIDIA AI Enterprise to cuQuantum to BioNeMo, catering to domains like data processing, quantum computing and digital biology.

The partnership opens AWS to millions of developers and the nearly 40,000 companies who are using these libraries, Huang said, adding that it’s great to see AWS expand its cloud instance offerings to include NVIDIA’s new L4, L40S and, soon, H200 GPUs.

Selipsky then introduced the AWS debut of the NVIDIA GH200 Grace Hopper Superchip, a significant advancement in cloud computing, and prompted Huang for further details.

“Grace Hopper, which is GH200, connects two revolutionary processors together in a really unique way,” Huang said. He explained that the GH200 connects NVIDIA’s Grace Arm CPU with its H200 GPU using a chip-to-chip interconnect called NVLink, at an astonishing one terabyte per second.

Each processor has direct access to the high-performance HBM and efficient LPDDR5X memory. This configuration results in 4 petaflops of processing power and 600GB of memory for each superchip.

AWS and NVIDIA connect 32 Grace Hopper Superchips in each rack using a new NVLink switch. Each 32 GH200 NVLink-connected node can be a single Amazon EC2 instance. When these are integrated with AWS Nitro and EFA networking, customers can connect GH200 NVL32 instances to scale to thousands of GH200 Superchips

“With AWS Nitro, that becomes basically one giant virtual GPU instance,” Huang said.

The combination of AWS expertise in highly scalable cloud computing plus NVIDIA innovation with Grace Hopper will make this an amazing platform that delivers the highest performance for complex generative AI workloads, Huang said.

“It’s great to see the infrastructure, but it extends to the software, the services and all the other workflows that they have,” Selipsky said, introducing NVIDIA DGX Cloud on AWS.

This partnership will bring about the first DGX Cloud AI supercomputer powered by the GH200 Superchips, demonstrating the power of AWS’s cloud infrastructure and NVIDIA’s AI expertise.

Following up, Huang announced that this new DGX Cloud supercomputer design in AWS, codenamed Project Ceiba, will serve as NVIDIA’s newest AI supercomputer as well, for its own AI research and development.


Named after the majestic Amazonian Ceiba tree, the Project Ceiba DGX Cloud cluster incorporates 16,384 GH200 Superchips to achieve 65 exaflops of AI processing power, Huang said.

Ceiba will be the world’s first GH200 NVL32 AI supercomputer built and the newest AI supercomputer in NVIDIA DGX Cloud, Huang said.

Huang described Project Ceiba AI supercomputer as “utterly incredible,” saying it will be able to reduce the training time of the largest language models by half.

NVIDIA’s AI engineering teams will use this new supercomputer in DGX Cloud to advance AI for graphics, LLMs, image/video/3D generation, digital biology, robotics, self-driving cars, Earth-2 climate prediction and more, Huang said.

“DGX is NVIDIA’s cloud AI factory,” Huang said, noting that AI is now key to doing NVIDIA’s own work in everything from computer graphics to creating digital biology models to robotics to climate simulation and modeling.

“DGX Cloud is also our AI factory to work with enterprise customers to build custom AI models,” Huang said. “They bring data and domain expertise; we bring AI technology and infrastructure.”

In addition, Huang also announced that AWS will be bringing four Amazon EC2 instances based on the NVIDIA GH200 NVL, H200, L40S, L4 GPUs, coming to market early next year.

Selipsky wrapped up the conversation by announcing that GH200-based instances and DGX Cloud will be available on AWS in the coming year.

You can catch the discussion and Selipsky’s entire keynote on AWS’s YouTube channel. 

Read More

NVIDIA BioNeMo Enables Generative AI for Drug Discovery on AWS

NVIDIA BioNeMo Enables Generative AI for Drug Discovery on AWS

Researchers and developers at leading pharmaceutical and techbio companies can now easily deploy NVIDIA Clara software and services for accelerated healthcare through Amazon Web Services.

Announced today at AWS re:Invent, the initiative gives healthcare and life sciences developers using AWS cloud resources the flexibility to integrate NVIDIA-accelerated offerings such as NVIDIA BioNeMo — a generative AI platform for drug discovery — coming to NVIDIA DGX Cloud on AWS, and currently available via the AWS ParallelCluster cluster management tool for high performance computing and the Amazon SageMaker machine learning service.

Thousands of healthcare and life sciences companies globally use AWS. They will now be able to access BioNeMo to build or customize digital biology foundation models with proprietary data, scaling up model training and deployment using NVIDIA GPU-accelerated cloud servers on AWS.

Techbio innovators including Alchemab Therapeutics, Basecamp Research, Character Biosciences, Evozyne, Etcembly and LabGenius are among the AWS users already using BioNeMo for generative AI-accelerated drug discovery and development. This collaboration gives them more ways to rapidly scale up cloud computing resources for developing generative AI models trained on biomolecular data.

This announcement extends NVIDIA’s existing healthcare-focused offerings available on AWS — NVIDIA MONAI for medical imaging workflows and NVIDIA Parabricks for accelerated genomics.

New to AWS: NVIDIA BioNeMo Advances Generative AI for Drug Discovery

BioNeMo is a domain-specific framework for digital biology generative AI, including pretrained large language models (LLMs), data loaders and optimized training recipes that can help advance computer-aided drug discovery by speeding target identification, protein structure prediction and drug candidate screening.

Drug discovery teams can use their proprietary data to build or optimize models with BioNeMo and run them on cloud-based high performance computing clusters.

One of these models, ESM-2 — a powerful LLM that supports protein structure prediction —  achieves almost linear scaling on 256 NVIDIA H100 Tensor Core GPUs. Researchers can scale to 512 H100 GPUs to complete training in a few days instead of a month, the training time published in the original paper.

Developers can train ESM-2 at scale using checkpoints of 650 million or 3 billion parameters. Additional AI models supported in the BioNeMo training framework include small-molecule generative model MegaMolBART and protein sequence generation model ProtT5.

BioNeMo’s pretrained models and optimized training recipes — which are available using self-managed services like AWS ParallelCluster and Amazon ECS as well as integrated, managed services through NVIDIA DGX Cloud and Amazon SageMaker — can help R&D teams build foundation models that can explore more drug candidates, optimize wet lab experimentation and find promising clinical candidates faster.

Also Available on AWS: NVIDIA Clara for Medical Imaging and Genomics

Project MONAI, cofounded and enterprise-supported by NVIDIA to support medical imaging workflows, has been downloaded more than 1.8 million times and is available for deployment on AWS. Developers can harness their proprietary healthcare datasets already stored on AWS cloud resources to rapidly annotate and build AI models for medical imaging.

These models, trained on NVIDIA GPU-powered Amazon EC2 instances, can be used for interactive annotation and fine-tuning for segmentation, classification, registration and detection tasks in medical imaging. Developers can also harness MRI image synthesis models available in MONAI to augment training datasets.

To accelerate genomics pipelines, Parabricks enables variant calling on a whole human genome in around 15 minutes, compared to a day on a CPU-only system. On AWS, developers can quickly scale up to process large amounts of genomic data across multiple GPU nodes.

More than a dozen Parabricks workflows are available on AWS HealthOmics as Ready2Run workflows, which enable customers to easily run pre-built pipelines.

Get started with NVIDIA Clara on AWS to accelerate AI workflows for drug discovery, genomics and medical imaging.

Subscribe to NVIDIA healthcare news.

Read More

NVIDIA GPUs on AWS to Offer 2x Simulation Leap in Omniverse Isaac Sim, Accelerating Smarter Robots

NVIDIA GPUs on AWS to Offer 2x Simulation Leap in Omniverse Isaac Sim, Accelerating Smarter Robots

Developing more intelligent robots in the cloud is about to get a speed multiplier.

NVIDIA Isaac Sim and NVIDIA L40S GPUs are coming to Amazon Web Services, enabling developers to build and deploy accelerated robotics applications in the cloud. Isaac Sim, an extensible simulator for AI-enabled robots, is built on the NVIDIA Omniverse development platform for building and connecting OpenUSD applications.

Combining powerful AI compute with graphics and media acceleration, the L40S GPU is built to power the next generation of data center workloads. Based on the Ada Lovelace architecture, the L40S enables ultrafast real-time rendering delivering up to a 3.8x performance leap for Omniverse compared with the previous generation, boosting engineering and robotics teams.

The generational leap in acceleration results in 2x faster performance than the A40 GPU across a broad set of robotic simulations tasks when using Isaac Sim.

L40S GPUs can also be harnessed for generative AI workloads, from fine-tuning large language models within a matter of hours, to real-time inferencing for text-to-image and chat applications.

New Amazon Machine Images (AMIs) on the NVIDIA L40S in AWS Marketplace will enable roboticists to easily access preconfigured virtual machines to operate Isaac Sim workloads.

Robotics development in simulation is speeding the process of deploying applications, turbocharging industries such as retail, food processing, manufacturing, logistics and more.

Revenue from mobile robots in warehouses worldwide is expected to explode, more than tripling from $11.6 billion in 2023 to $42.2 billion by 2030, according to ABI Research.

Robotics systems have played an important role across fulfillment centers to help meet the demands of online shoppers and provide a better workplace for employees. Amazon Robotics has deployed more than 750,000 robots in its warehouses around the world to improve the experience for employees supporting package fulfillment and its customers.

“Simulation technology plays a critical role in how we develop, test and deploy our robots.” said Brian Basile, head of virtual systems at Amazon Robotics. “At Amazon Robotics we continue to increase the scale and complexity of our simulations. With the new AWS L40S offering we will push the boundaries of simulation, rendering and model training even further.”

Accelerated Robotics Development With Isaac Sim

Robotics systems can demand large datasets for precision operation in deployed applications. Gathering these datasets and testing them in the real world is time-consuming, costly and impractical.

Robotics simulation drives the training and testing of AI-based robotic applications. With synthetic data, simulations are enabling virtual advances like never before. Simulations can help verify, validate and optimize robot designs, systems and their algorithms before operation. It can also be used to optimize facility designs before construction or remodeling starts for maximum efficiencies, reducing costly manufacturing change orders.

Isaac Sim offers access to the latest robotics simulation tools and capabilities as well as cloud access, enabling teams to collaborate more effectively. Access to the Omniverse Replicator synthetic data generation engine in Isaac Sim allows machine learning engineers to build production-ready synthetic datasets for training robust deep learning perception models.

Customer Adoption of Isaac Sim on AWS

AWS early adopters tapping into the Isaac Sim platform include Amazon Robotics, Soft Robotics and Theory Studios.

Amazon Robotics has begun using Omniverse to build digital twins for automating, optimizing and planning its autonomous warehouses in virtual environments before deploying them into the real world.

Using Isaac Sim for sensor emulation, Amazon Robotics will accelerate development of its Proteus autonomous mobile robot, improving it to help the online retail giant efficiently manage fulfillment.

Learn more about Isaac Sim, powered by NVIDIA Omniverse.

 

Read More

NVIDIA Powers Training for Some of the Largest Amazon Titan Foundation Models

NVIDIA Powers Training for Some of the Largest Amazon Titan Foundation Models

Everything about large language models is big — giant models train on massive datasets across thousands of NVIDIA GPUs.

That can pose a lot of big challenges for companies pursuing generative AI. NVIDIA NeMo, a framework for building, customizing and running LLMs, helps overcome these challenges.

A team of experienced scientists and developers at Amazon Web Services creating Amazon Titan foundation models for Amazon Bedrock, a generative AI service for foundation models, has been using NVIDIA NeMo for over the past several months.

“One key reason for us to work with NeMo is that it is extensible, comes with optimizations that allow us to run with high GPU utilization while also enabling us to scale to larger clusters so we can train and deliver models to our customers faster,” said Leonard Lausen, a senior applied scientist at AWS.

Think Big, Really Big

Parallelism techniques in NeMo enable efficient LLM training at scale. When coupled with the Elastic Fabric Adapter from AWS, it allowed the team to spread its LLM across many GPUs to accelerate training.

EFA provides AWS customers with an UltraCluster Networking infrastructure that can directly connect more than 10,000 GPUs and bypass the operating system and CPU using NVIDIA GPUDirect.

The combination allowed the AWS scientists to deliver excellent model quality — something that’s not possible at scale when relying solely on data parallelism approaches.

Framework Fits All Sizes

“The flexibility of NeMo,” Lausen said, “allowed AWS to tailor the training software for the specifics of the new Titan model, datasets and infrastructure.”

AWS’s innovations include efficient streaming from Amazon Simple Storage Service (Amazon S3) to the GPU cluster. “It was easy to incorporate these improvements because NeMo builds upon popular libraries like PyTorch Lightning that standardize LLM training pipeline components,” Lausen said.

AWS and NVIDIA aim to infuse products like NVIDIA NeMo and services like Amazon Titan with lessons learned from their collaboration for the benefit of customers.

Read More

3D Artist Nourhan Ismail Brings Isometric Innovation ‘In the NVIDIA Studio’ With Adobe After Effects and Blender

3D Artist Nourhan Ismail Brings Isometric Innovation ‘In the NVIDIA Studio’ With Adobe After Effects and Blender

Editor’s note: This post is part of our weekly In the NVIDIA Studio series, which celebrates featured artists, offers creative tips and tricks, and demonstrates how NVIDIA Studio technology improves creative workflows. 

This week’s talented In the NVIDIA Studio artist, Nourhan Ismail, created a literal NVIDIA studio.

Her piece, called Creator by Day, Gamer by Night, was crafted with the isometric art style and impressive graphical fidelity Ismail’s known for, rich with vibrant colors and playful details. It also captures her “work hard, play hard” mentality as a 3D artist, interior designer and game level designer.

The same art style is featured in the NVIDIA Studio Sessions YouTube miniseries led by Ismail, which provides step-by-step tutorials on how to create a low-poly bedroom, from inception to final render.

Facial Animations Made Easier

Reallusion is the maker of Reallusion iClone, real-time 3D animation software built to produce professional animations for films and video games.

To expedite character animation workflows, the company recently launched its AccuFACE plug-in, which accurately captures facial expressions from webcams and conventional video files, without the need for expensive, specialized equipment.

The NVIDIA Maxine software development platform, the foundational technology behind the revolutionary NVIDIA Broadcast app, powers this incredible capability by weighing output and analyzing facial expressions and blendshapes to predict facial mesh animations.

From there, the AccuFACE plug-in converts this data into facial mesh assets for creators to apply seamlessly. It also fine-tunes lip and tongue articulation using proprietary AccuLIPS technology.

Download the plug-in today, available to creators with NVIDIA RTX GPUs.

Turning Pain Into Beauty

Ismail’s creative journey began at age four as a form of escape from the armed conflict occurring in Syria, her homeland. During that time, Ismail’s family faced many difficulties, including the loss of their home.

In the aftermath, she looked to her father, an accomplished artist and fashion designer, as a source of inspiration.

“His encouragement propelled me to showcase the pinnacle of my abilities, reminding me that art has the power to transform pain into beauty,” she said.

That encouragement has guided and fueled Ismail’s creative journey, eventually giving rise to her signature, single-room isometric style, an homage to the power of resilience and finding beauty in adversity.

Warm and homey.

“Starting with a single room, I delve into interior design, crafting spaces that reflect the comfort and joy I yearned for during challenging times,” she said. “To me, overcoming adversity proves that even from the harshest circumstances, beauty can emerge.”

Ismail started as a self-taught 3D artist, driven by a passion to learn the intricacies of creating digital masterpieces.

“Posting my works became a personal gauge of improvement — not for validation, but as a record of my learning curve,” she said.

Beautifully conceived, masterfully executed.

Each of Ismail’s pieces is a testament to her evolving skills, dedication and love for sharing her craft, especially with her father.

In fact, she dedicated her first isometric house to her father. “That was the happiest moment, to create something inspiring and make someone happy,” she said.

Isometric Art

Ismail first collects reference material on Adobe Behance to gain inspiration on ways to mix different art styles.

She then opens Blender and starts sketching in 3D. Blender Cycles’ RTX-accelerated OptiX ray tracing, powered by her GeForce RTX 3080 Ti GPU, ensured smooth viewport movement.

A gorgeous work in progress.

While the models are still fairly rudimentary, Ismail calculates the angles that light should be coming in from.

“Lighting is an emotional element,” she said. “The lighting of each piece evokes different emotions and a certain idiosyncratic introspectiveness, making the experience unique to each person.”

The NVIDIA Studio has NVIDIA Canvas!

Her trick is to regularly switch between rich, colorful scenes and plain color models to measure the emotional weight and visual impact. She either creates the custom textures herself or downloads premade ones online when on a time crunch.

Ismail’s incredible detail on full display.

Then, she plays with camera angles to analyze depth shadows and lighting, setting up animations and sequence shots in Blender. There, Blender Cycles’ RTX-accelerated OptiX ray tracing delivered seamless viewport movement.

 

Final touch-ups are done in post-production in Adobe After Effects. Over 30 GPU-accelerated effects sped the process, allowed Ismail to complete the project with time to spare.

“Creator by Day, Gamer by Night” in dark mode.

“There will always be hard times, so never give up and keep believing in yourself,” Ismail encourages content creators.

Digital 3D artist Nourhan Ismail.

Check out Ismail’s Instagram for more spectacular isometric art.

Follow NVIDIA Studio on Instagram, Twitter and Facebook. Access tutorials on the Studio YouTube channel and get updates directly in your inbox by subscribing to the Studio newsletter. 

Read More

Medical Imaging AI Made Easier: NVIDIA Offers MONAI as Hosted Cloud Service

Medical Imaging AI Made Easier: NVIDIA Offers MONAI as Hosted Cloud Service

NVIDIA today launched a cloud service for medical imaging AI to further streamline and accelerate the creation of ground-truth data and training of specialized AI models through fully managed, cloud-based application programming interfaces.

NVIDIA MONAI cloud APIs — announced at the annual meeting of RSNA, the Radiological Society of North America, taking place this week in Chicago — provide an expedited path for developers and platform providers to integrate AI into their medical imaging offerings using pretrained foundation models and AI workflows for enterprises. The APIs are built on the open-source MONAI project founded by NVIDIA and King’s College London.

Medical imaging is critical across healthcare, making up approximately 90% of healthcare data. It’s used by radiologists and clinicians to do screening, diagnosis and intervention, by biopharma researchers to evaluate how clinical trial patients respond to new drugs and by medical device makers to provide real-time decision support.

The scale of work across each of these areas requires a medical imaging-specific AI factory — an enterprise-grade platform that delivers large-scale data management, creates ground-truth annotations, accelerates model development and establishes seamless AI application deployment.

With NVIDIA MONAI cloud APIs, solution providers can more easily integrate AI into their medical imaging platforms, enabling them to provide supercharged tools for radiologists, researchers and clinical trial teams to build domain-specialized AI factories. The APIs are available in early access through the NVIDIA DGX Cloud AI supercomputing service.

The NVIDIA MONAI cloud API is integrated into Flywheel, a leading medical imaging data and AI platform that supports end-to-end workflows for AI development. Developers at medical image annotation companies including RedBrick AI and at machine learning operations (MLOps) platform providers including Dataiku are poised to integrate NVIDIA MONAI cloud APIs into their offerings.

Ready-to-Run Annotation and Training for Medical Imaging

Building efficient and cost-effective AI solutions requires a robust, domain-specialized development foundation that includes full-stack optimizations for software, scalable multi-node systems and state-of-the-art research. It also requires high-quality ground-truth data — which can be arduous and time-consuming to gather, particularly for 3D medical images that require a high level of expertise to annotate.

NVIDIA MONAI cloud APIs feature interactive annotation powered by the VISTA-3D (Vision Imaging Segmentation and Annotation) foundation model. It’s purpose-built for continuous learning, a capability that improves AI model performance based on user feedback and new data.

Trained on a dataset of annotated images from 3D CT scans from more than 4,000 patients, spanning various diseases and parts of the body, VISTA-3D accelerates the creation of 3D segmentation masks for medical image analysis. With continuous learning, the AI model’s annotation quality improves over time.

To further accelerate AI training, this release includes APIs that make it seamless to build custom models based on MONAI pretrained models. NVIDIA MONAI cloud APIs also include Auto3DSeg, which automates hyperparameter tuning and AI model selection for a given 3D segmentation task, simplifying the model development process.

NVIDIA researchers recently won four challenges at the MICCAI medical imaging conference using Auto3DSeg. These included AI models to analyze 3D CT scans of the kidneys and heart, brain MRIs and 3D ultrasounds of the heart.

Solutions Providers, Platform Builders Embrace NVIDIA MONAI Cloud APIs

Medical imaging solution providers and machine learning platforms are using NVIDIA MONAI cloud APIs to deliver critically valuable AI insights to accelerate their customers’ work.

Flywheel has integrated MONAI through NVIDIA AI Enterprise and is now offering NVIDIA MONAI cloud APIs to accelerate medical image curation, labeling analysis and training. The Minneapolis-based company’s centralized, cloud-based platform powers biopharma companies, life science organizations, healthcare providers and academic medical centers to identify, curate and train medical imaging data for the development of trustworthy AI.

“NVIDIA MONAI cloud APIs lower the cost of building high-quality AI models for radiology, disease research and the evaluation of clinical trial data,” said Dan Marcus, chief scientific officer at Flywheel. “With the addition of cloud APIs for interactive annotation and automated segmentation, customers of our medical imaging AI platform can accelerate AI model development to more quickly deliver innovative solutions.”

Annotation and viewer solution providers, including Redbrick AI, Radical Imaging, V7 Labs and Centaur Labs, will also use NVIDIA MONAI cloud APIs to bring AI-assisted annotation and training capabilities to market faster, without having to host and manage the AI infrastructure on their own.

RedBrick AI is integrating the VISTA-3D model available through NVIDIA MONAI cloud APIs to deliver interactive cloud annotation for its medical device customers that support distributed teams of clinicians.

“VISTA-3D allows our clients to rapidly build models across different modalities and conditions,” said Shivam Sharma, CEO of RedBrick AI. “The foundation model is generalizable, making it easy to fine-tune for various clinical applications with accurate, reliable segmentation results.”

To streamline enterprise AI model development, MLOps platform builders including Dataiku, ClearML and Weight & Biases are also investigating the use of NVIDIA MONAI cloud APIs.

Dataiku plans to integrate NVIDIA MONAI cloud APIs to further simplify AI model creation for medical imaging applications.

“With NVIDIA MONAI cloud APIs, Dataiku users would be able to easily use Auto3DSeg, a low-code option to accelerate the development of state-of-the-art segmentation models, through Dataiku’s web interface connected to an NVIDIA-hosted, GPU-accelerated service,” said Kelci Miclaus, global head of AI health and life sciences solutions at Dataiku. “This democratizes AI in biomedical imaging by extending the power to create and apply AI-driven workflows to both data and domain experts.”

Join the medical imaging innovators accelerating AI development with NVIDIA MONAI cloud APIs by signing up for early access.

Read More

Hungry for Gaming: 18 New Games to Join GeForce NOW

Hungry for Gaming: 18 New Games to Join GeForce NOW

GeForce NOW is bringing 18 new games to the cloud this week, part of a gratitude-filled GFN Thursday.

A collaboration between Chromebook Plus, CD PROJEKT RED and GeForce NOW brought an immersive 3D activation to Times Square over the weekend, containing a hidden Easter egg for Cyberpunk 2077 players.

Plus, this holiday season, give the gift of high-performance cloud gaming with a free, three-month PC Game Pass subscription included with a six-month GeForce NOW Ultimate membership.

Eye Spy

Chromebook Plus activation with GeForce NOW and Cyberpunk, Times Square
Crack the code now with Chromebook Plus, GeForce NOW and “Cyberpunk 2077.”

Times Square was transformed into a futuristic dystopia last weekend by a stunning 3D Cyberpunk 2077 activation. The immersive display showcased the power of GeForce NOW on a Chromebook Plus, enabling gamers to stream thousands of titles from popular digital gaming stores from the cloud, right out of the box.

Long-time Cyberpunk 2077 fans were also challenged to try their hand at spotting a code hidden in the activation — the latest hint in a long line of clues needed to solve a riddle that has had players guessing for the past three years.

Chromebook owners with an Ultimate or Priority membership can explore Cyberpunk 2077’s Night City with NVIDIA RTX ON for the most immersive gaming experience, including the “Phantom Liberty” expansion and 2.0 update. Google and NVIDIA are offering all Chromebook owners three free months of a GeForce NOW Priority membership to get gamers started.

And those interested in leveling up to the highest-performing tier can get three free months of a GeForce NOW Ultimate membership with the purchase of a Cloud Gaming Chromebook. Find more details on how to redeem the offer on the Chromebook Perks page.

Grateful for Great Games

Chivalry 2 on GeForce NOW
That’s one way to carve a turkey.

Get the weekend started with the new weekly games list:

  • Breathedge (Xbox, available on Microsoft Store)
  • Bridge Constructor: The Walking Dead (Xbox, available on Microsoft Store)
  • Bus Simulator 21 (Xbox, available on Microsoft Store)
  • Chivalry 2 (Xbox, available on PC Game Pass)
  • Dungeons 4 (Epic Games Store)
  • Hexarchy (Steam)
  • Hearts of Iron IV (Xbox, available on Microsoft Store)
  • I Am Future (Epic Games Store)
  • Imagine Earth (Xbox, available on Microsoft Store)
  • The Invincible (Epic Games Store)
  • Land of the Vikings (Steam)
  • Saints Row IV: Re-Elected (Xbox, available on Microsoft Store)
  • SHENZHEN I/O (Xbox, available on Microsoft Store)
  • Supraland: Six Inches Under (Xbox, available on PC Game Pass)
  • The Surge 2 (Xbox, available on Microsoft Store)
  • Thymesia (Xbox, available on Microsoft Store)
  • Tropico 6 (Xbox, available on Microsoft Store)
  • West of Dead (Xbox, available on Microsoft Store)

What are you looking forward to streaming? Let us know on Twitter or in the comments below.

Read More

Teenage Dream: Aspiring Computer Science Major Experiences NVIDIA Life With Make-A-Wish Visit

Teenage Dream: Aspiring Computer Science Major Experiences NVIDIA Life With Make-A-Wish Visit

A calendar packed with meetings, calls and lab visits may sound like a typical workday for many — but for Luca Lofranco, whose greatest wish was to experience what it’s like to work at NVIDIA, it was a dream come true.

Eighteen-year-old Lofranco recently traveled from his hometown near Toronto, Canada, to spend the day at our Santa Clara campus, supported by Make-A-Wish, a nonprofit that grants life-changing wishes for children with critical illnesses. The wish from Lofranco, who has Hodgkin’s lymphoma, was the fifth NVIDIA has been a part of in the last decade.

The NVIDIA team kept the day’s agenda a secret — surprising Lofranco with tours of the demo room and robotics lab, a chat with the University Recruiting team, a ride in a self-driving car and a video call with NVIDIA founder and CEO Jensen Huang. An aspiring computer science major, Lofranco was stoked for it all because, as his mom Cassandra shared, “NVIDIA is his Disneyland.”

NVIDIA’s auto garage

A Long-Time NVIDIA Fan 

After attending his first computer science summer camp when he was eight, Lofranco learned 3D modeling in Autodesk Maya, programming in Python, as well as 3D printing. His budding interest in tech grew and, soon enough, he was building his own gaming rigs.

NVIDIA quickly became Lofranco’s favorite tech company, he said, so much so that he carved the company logo out of a piece of wood using a computer numerical control machine.

For gaming, he enjoys using NVIDIA GeForce RTX 3070 and GeForce GTX 1080 Ti GPUs. But Lofranco’s ultimate draw to NVIDIA wasn’t its products but its culture.

“Everyone is driven to see the same outcome and comes together to make it happen,” he said. “Everything is designed for collaboration.”

Lofranco in NVIDIA gear

A VIP Experience

Ahead of Lofranco’s visit, the NVIDIA team sent him a box of swag — including a hoodie, a hat and a custom NVIDIA badge.

Once he arrived on campus, NVIDIA volunteers welcomed and whisked Lofranco off on a campus tour, followed by a meeting with the solutions architect team, which includes NVIDIANs focused on healthcare, auto, AI, cloud service providers and large language models.

Next, a visit to the robotics lab helped satisfy Lofranco’s “maker” curiosity. He saw an NVIDIA DGX Station as well as test robots for developing the NVIDIA Jetson edge AI platform, and was soon directing a robot arm to stack colored blocks.

After learning that Lofranco’s favorite foods include lobster, tiramisu and Kit Kat candy bars, the café team prepared a special menu for him and all employees in the office that day. Everyone enjoyed a lobster roll pop-up station in the campus park and tiramisu-flavored ice cream with assorted toppings, including Kit Kat pieces.

Lofranco checks out NVIDIA GPUs in the company’s demo room

Innovators

On a visit to the demo room at NVIDIA’s Santa Clara site, Lofranco and his father revealed that they tinker with innovations themselves. They programmed their water heater in the family hot tub to maintain a comfortable temperature and decrease the time needed to warm it — all thanks to Python code and Raspberry Pi experimentation.

With so much to soak in, Lofranco described his wish day at NVIDIA as “unfathomable” — and that was before his video call with Huang, which stretched from a planned quarter hour to 45 minutes.

After a conversation that spanned NVIDIA’s origins, many near failures and innovation, Huang gifted Lofranco a GeForce RTX 4090 Founders Edition GPU and shared some sound advice: “Keep playing video games — but make sure to prioritize your homework.”

Lofranco with the surprise gift from Huang following their chat

Capping a packed day of fun-filled support from nearly 50 NVIDIANs was a visit to the auto lab and a spin in one of NVIDIA’s self-driving test cars.

How was it all? “Breathtaking,” said Lofranco, who learned firsthand from Huang that while NVIDIA has evolved from being the underdog to a leading tech company, it still feels “like a family.”

Learn more about NVIDIA life, culture and careers

Read More