Unlimited Data, Unlimited Possibilities: UF Health and NVIDIA Build World’s Largest Clinical Language Generator

The University of Florida’s academic health center, UF Health, has teamed up with NVIDIA to develop a neural network that generates synthetic clinical data — a powerful resource that researchers can use to train other AI models in healthcare.

Trained on a decade of data representing more than 2 million patients, SynGatorTron is a language model that can create synthetic patient profiles that mimic the health records it’s learned from. The 5 billion-parameter model is the largest language generator in healthcare.

“Synthetic data isn’t actually linked to a real human being, but it has similar characteristics to real patients,” said Dr. Duane Mitchell, an assistant vice president for research and director of the UF Clinical and Translational Science Institute. “SynGatorTron can, for example, create health records of digital diabetes patients that have features just like a real population.”

Using this synthetic data, researchers can create tools, models and tasks without risks or privacy concerns. These can then be used on real data to ask clinical questions, look for associations and even explore patient outcomes.

Working with synthetic data also makes it easier for different research institutions to collaborate and share models. And since the amount of data that can be synthesized is virtually limitless, researchers can use SynGatorTron-generated data to augment small datasets of rare disease patients or minority populations to reduce model bias.

SynGatorTron was developed using the open-source NVIDIA Megatron-LM and NeMo frameworks. It’s based on UF Health’s GatorTron model, announced last year at NVIDIA GTC. The models were trained on HiPerGator-AI, the university’s in-house NVIDIA DGX SuperPOD system, which ranks among the world’s top 30 supercomputers.

GatorTron-S, a BERT-style transformer model trained on synthetic data generated by SynGatorTron, will be available for developers next month on the NGC software hub. 

SynGatorTron Opens Gate to Robust Training Data

To a doctor, an AI-generated doctor’s note can appear impractical at first glance — it doesn’t represent a real patient and won’t read as logical to an expert eye. So a clinician can’t make a direct analysis or diagnosis from it. But to an untrained AI, real and synthetic clinical data are both highly valuable.

“SynGatorTron’s generative capability is a great enabler of natural language processing for medicine,” said Dr. Mona Flores, global head of medical AI at NVIDIA. “Synthesizing different types of clinical records will democratize the ability to create all sorts of applications dependent on such data by addressing data sparsity and privacy.”

Once it’s available, research institutions outside UF Health could fine-tune the pretrained SynGatorTron model with their own localized data and apply it to their AI projects. For example, if a given condition or a patient population is underrepresented in a health system’s clinical data, SynGatorTron can be prompted to generate additional data with characteristics of that disease or population.

These AI-generated records could then be used to supplement and balance out real healthcare datasets used to train other neural networks, so that they better represent the population.

Since synthetic training datasets mimic real medical notes without being associated with specific patients, they can also be more readily shared across research institutions without raising privacy concerns.

“When you have the ability to mimic population characteristics without being tethered to real patients, it opens the imagination to see if we can generate realistic datasets that allow us to answer questions we couldn’t otherwise, due to constraints on access to data or limited information on patients of interest,” Mitchell said.

One potential application is in clinical trials, which often divide patients into treatment and control groups to measure the effectiveness of a new medication. An application derived from SynGatorTron-generated data could parse through real records and create a digital twin of patient records. These records could then be used as the control group in a clinical trial, instead of having a control group derived by giving real patients a placebo treatment.

Researchers developing a deep learning model to study a rare disease, or the effects of a treatment on a specific population, could also use SynGatorTron for data augmentation, generating more training data to supplement the limited amount of real medical records available.

Healthcare at GTC 

Register free for GTC, running online March 21-24, to discover the latest in AI and healthcare. Hear from SynGatorTron collaborators in the session “A Next-Generation Clinical Language Model,” taking place March 23 at 7 a.m. Pacific.

Watch the replay of NVIDIA founder and CEO Jensen Huang’s keynote address below:

The post Unlimited Data, Unlimited Possibilities: UF Health and NVIDIA Build World’s Largest Clinical Language Generator appeared first on NVIDIA Blog.

Read More

New NVIDIA RTX GPUs Tackle Demanding Professional Workflows and Hybrid Work, Enabling Creation From Anywhere

Remote work and hybrid workplaces are the new normal for professionals in many industries. Teams spread throughout the world are expected to create and collaborate while maintaining top productivity and performance.

Businesses use the NVIDIA RTX platform to enable their workers to keep up with the most demanding workloads, from anywhere. And today, NVIDIA is expanding its RTX offerings with seven new NVIDIA Ampere architecture GPUs for laptops and desktops.

The new NVIDIA RTX A500, RTX A1000, RTX A2000 8GB, RTX A3000 12GB, RTX A4500 and RTX A5500​ laptop GPUs expand access to AI and ray-tracing technology, delivering breakthrough performance no matter where you work. The laptops include the latest RTX and Max-Q technology, giving professionals the ability to take their workflows to the next level.

The new NVIDIA RTX A5500 desktop GPU combines the latest-generation RT Cores, Tensor Cores and CUDA cores with 24GB of memory for incredible rendering, AI, graphics and compute performance. Its ray-traced rendering is 2x faster than the previous generation and its motion blur rendering performance is up to 9x faster.

With the new NVIDIA RTX GPUs, artists can create photorealistic immersive digital experiences, scientists can make the latest groundbreaking discoveries, and engineers can develop innovative technology to advance us into the future.

Customer Adoption

Among the first to use the NVIDIA RTX A5500 is M4 Engineering, a leading aerospace engineering firm focused on conceptual aircraft design, analysis and development.

“The multi-app product development workflows we use at M4 are well-served by the NVIDIA RTX A5500 and its 24GB of memory,” said Brian Rotty, senior engineer at M4 Engineering. “My team can handle larger CAD and CAE datasets than before and, critically, we can interact with and iterate these larger datasets simultaneously by making use of the extra GPU memory headroom and compute capabilities of this new card.”

“The performance we get with the NVIDIA RTX A5500 is unprecedented,” said Hiram Rodriguez, design technology specialist at AS+GG Architecture. “Previously, our point cloud processing took too long and created a bottleneck for designing and analyzing current site conditions. Using the NVIDIA RTX A5500, in less than 20 minutes we can integrate a fully processed, geolocated, classified point cloud into our 3D models.”

“The new NVIDIA RTX A5500 gives us the ability to load highly detailed environments, while maintaining high frame rates and smooth camera motion to explore and develop the scenes,” said Yiotis Katsambas, executive director of Technology at Sony Pictures Animation. “By combining the A5500 with NVIDIA Omniverse Enterprise and our own FlixiVerse software, our artists and directors can immerse themselves in our virtual worlds and collaborate in real-time.”

Next-Generation RTX Technology

The new NVIDIA RTX GPUs feature the latest generation of NVIDIA RTX technology to accelerate graphics, AI and ray-tracing performance.

These GPUs are part of the comprehensive NVIDIA RTX platform, which includes NVIDIA GPU-accelerated software development kits, toolkits, frameworks and enterprise management tools.

These GPUs also take advantage of the accelerations of the NVIDIA Studio platform, including optimizations to leverage RTX hardware in 75 of the top creative apps, exclusive tools like NVIDIA Broadcast and Canvas, and the advanced real-time 3D design and collaboration platform, NVIDIA Omniverse.

NVIDIA RTX GPUs are certified by independent software vendors of over 50+ professional applications. Certification provides users with a reliable and dependable graphics and computing experience through testing and development.

NVIDIA RTX laptop GPUs include: 

  • The latest NVIDIA RTX technology: 2nd gen RT Cores, 3rd gen Tensor Cores and NVIDIA Ampere architecture streaming multiprocessor, which provide up to 2x the throughput of the previous-generation architecture to tackle demanding rendering, ray tracing and AI workflows from anywhere.
  • NVIDIA Max-Q technology: AI-based system optimizations make thin and light laptops perform quieter, faster and more efficiently with Dynamic Boost, CPU Optimizer, Rapid Core Scaling, WhisperMode, Battery Boost, Resizable BAR and NVIDIA DLSS technology.
  • Up to 16GB of GPU memory: For the largest models, scenes and assemblies. The RTX A2000 8GB, RTX A3000 12GB and RTX A4500 now feature 2x the memory of their previous-generation counterparts for working with larger models, datasets and multi-app workflows.
  • Rich suite of NVIDIA software technology: Users can tap into unique benefits ranging from tetherless VR to collaborative 3D design with a wide variety of software tools, including NVIDIA CloudXR, NVIDIA Omniverse, NVIDIA Canvas, NVIDIA Broadcast, NVIDIA NGC, NVIDIA RTX Experience and more.

The NVIDIA RTX A5500 features the latest technologies in the NVIDIA Ampere architecture:

  • Second-generation RT Cores: Up to 2x the throughput of the first generation with the ability to run concurrent ray tracing, shading and denoising tasks.
  • Third-generation Tensor Cores: Up to 12x the training throughput over the previous generation, with support for new TF32 and Bfloat16 data formats.
  • CUDA Cores: Up to 3x the single-precision floating point throughput over the previous generation.
  • Up to 48GB of GPU memory: The RTX A5500 features 24GB of GDDR6 memory with ECC (error correction code). The RTX A5500 is expandable up to 48GB of memory using NVIDIA NVLink to connect two GPUs.
  • Virtualization: The RTX A5500 supports NVIDIA RTX Virtual Workstation (vWS) software for multiple high-performance virtual workstation instances that enable remote users to share resources to drive high-end design, AI and compute workloads.
  • PCIe Gen 4: Doubles the bandwidth of the previous generation and speeds up data transfers for data-intensive tasks such as AI, data science and creating 3D models.

Availability 

The new NVIDIA RTX A5500 desktop GPUs are available today from channel partners and through global system builders starting in Q2.

The new NVIDIA RTX laptop GPUs will be available in mobile workstations from global OEM partners including Acer, ASUS, BOXX Technologies, Dell Technologies, HP, Lenovo, and MSI starting this spring. Contact these partners for specific system configuration details and availability.

To learn more about NVIDIA RTX, watch the GTC 2022 keynote from Jensen Huang. Register for GTC 2022 for free to attend sessions with NVIDIA and industry leaders.

The post New NVIDIA RTX GPUs Tackle Demanding Professional Workflows and Hybrid Work, Enabling Creation From Anywhere appeared first on NVIDIA Blog.

Read More

First Wave of Startups Harnesses UK’s Most Powerful Supercomputer to Power Digital Biology Breakthroughs

Four NVIDIA Inception members have been selected as the first cohort of startups to access Cambridge-1, the U.K.’s most powerful supercomputer.

The system will help British companies Alchemab Therapeutics, InstaDeep, Peptone and Relation Therapeutics enable breakthroughs in digital biology.

Officially launched in July, Cambridge-1 — an NVIDIA DGX SuperPOD cluster powered by NVIDIA DGX A100 systems, BlueField-2 DPUs and NVIDIA InfiniBand networking — brings together NVIDIA’s decades-long work in accelerated computing, AI and life sciences. Located between London and Cambridge, it ranks among the world’s top 50 fastest computers and is powered by 100 percent renewable energy.

The supercomputer’s five founding partners have already been using it to advance healthcare, using AI to research brain diseases like dementia, design new drugs and more.

Now, the four startups are preparing to use Cambridge-1 to accelerate drug discovery, genome sequencing and disease research.

Each is a member of NVIDIA Inception, a free program that nurtures startups revolutionizing industries with cutting-edge technology. Inception gives members a custom set of ongoing benefits, such as NVIDIA Deep Learning Institute credits, marketing support and technology assistance from experts.

Alchemab Therapeutics: Discovering Antibody Therapeutics

Alchemab Therapeutics is identifying novel drug targets and therapeutics, and building patient stratification tools, with an initial focus on neurodegenerative conditions and cancer.

The company’s antibody drug discovery engine is being built on “nature’s most effective search engine: adaptive immunity,” according to Jake Galson, head of technology at Alchemab. This is the type of immunity developed when a person’s immune system responds to a pathogenic protein, such as those produced by cancers, or a foreign microorganism, like after an infection.

Alchemab’s platform sequences B-cells, which produce antibodies that fight disease, and computationally analyzes antibody responses among individuals who are susceptible but resilient to certain diseases.

“Approximately 10 trillion human antibody variants are possible, and having access to Cambridge-1 gives us a unique opportunity to learn meaningful representations from such an enormous body of data,” Galson said. “This will increase our understanding of antibody structure and function, and ultimately contribute to the discovery and development of novel therapeutics.”

Attend Alchemab’s session on deciphering the language of antibodies at GTC, a global AI conference running through March 24.

InstaDeep: Creating Decision-Making Systems for Biology

InstaDeep, an Elite member of the NVIDIA Partner Network, delivers AI-powered decision-making systems for the development of next-generation vaccines and therapeutics.

The company is looking to train a large AI language model using genomics data — and share the model with healthcare researchers to use for protein design and molecular dynamics simulations.

“There are over 12 billion nucleotide sequences from 450,000 species that are publicly available,” said Karim Beguir, co-founder and CEO of InstaDeep. “Researchers and life science professionals could benefit tremendously from a large-scale model providing data-driven insights from genome sequencing.”

Access to Cambridge-1, Beguir said, will enable InstaDeep to significantly scale the startup’s “compute capabilities and ambitions, and tackle exciting challenges in the development of novel treatments for patients.”

Learn more about how AI language models are applied in biology at InstaDeep’s GTC session on revolutionizing protein research with high performance computing.

Peptone: Providing Insight About Disordered Proteins

Peptone, a startup that received early approval to access Cambridge-1 last fall, is developing a physics engine called Oppenheimer, which will help deliver precise structural insights about intrinsically disordered proteins (IDPs), or proteins that lack a fixed 3D structure.

Diseases that stem from IDPs are usually difficult to treat, but Cambridge-1 will give Peptone the power to potentially “transform a typically undruggable IDP into a plausible drugging target,” said Kamil Tamiola, founder and CEO of Peptone.

“The supercomputer will enable us to perform high-throughput inference on millions of proteins in parallel and in a matter of hours,” Tamiola said. “Oppenheimer integrates advanced atomistic biophysical experiments with a next-generation supercomputing stack built on NVIDIA DGX A100 systems.”

Ultimately, the company will use the calculations to develop a proprietary and first-in-class line of drugs targeting selected IDPs.

Relation Therapeutics: Mapping the Causes of Disease

Another startup, Relation Therapeutics, combines single-cell profiling, human genetics, functional genomics and machine learning to better understand human biology.

RelationTx uses graph-based recommender system technologies to reveal causal relationships in diseases. RelationTx’s platform can identify the areas of biology to focus on for drug discovery and accelerate research efforts for diseases that have not yet been widely studied.

The company aims to transform how drug discovery and development is conducted, leading to new treatments for disease, according to Lindsay Edwards, chief technology officer at RelationTx.

“Ultimately, our mission is to get new medicines to patients who need them, faster and more efficiently than the current paradigm,” Edwards said. “Access to Cambridge-1 opens up areas of biology that were almost impossible to understand before, such as how genetic variation affects gene expression in inaccessible complex tissues and organ systems.”

Learn More About AI in Healthcare

Groundbreaking work in digital biology is to come from these startups — and Cambridge-1’s founding companies are already harnessing its power.

Using the supercomputer, AstraZeneca and NVIDIA developed the latest iteration of MegaMolBART, a natural language processing model that reads the text format of chemical compounds and uses AI to generate new molecules. The transformer chemistry model is capable of training chemical language models with over 1 billion parameters using the NVIDIA NeMo Megatron framework.

Learn more about AI-based innovation at GTC, where Kimberly Powell, vice president of healthcare at NVIDIA, will discuss how researchers, developers and medical device makers use the NVIDIA Clara platform to create breakthroughs in healthcare and drug discovery.

Watch NVIDIA founder and CEO Jensen Huang’s GTC keynote address.

Subscribe to NVIDIA healthcare news.

The post First Wave of Startups Harnesses UK’s Most Powerful Supercomputer to Power Digital Biology Breakthroughs appeared first on NVIDIA Blog.

Read More

NVIDIA Omniverse Ecosystem Expands 10x, Amid New Features and Services for Developers, Enterprises and Creators

When it comes to creating and connecting virtual worlds, over 150,000 individuals have downloaded NVIDIA Omniverse to make huge leaps in transforming 3D design workflows and achieve new heights of real-time, physically accurate simulations.

At GTC, NVIDIA today announced new releases and updates for Omniverse — including the latest Omniverse Connectors and libraries — expanding the platform’s ecosystem by 10x, and making Omniverse even more accessible to creators, developers, designers, engineers and researchers worldwide.

NVIDIA Omniverse Enterprise is helping leading companies enhance their pipelines and creative workflows. New Omniverse Enterprise customers include Amazon, DB Netze, DNEG, Kroger, Lowe’s and PepsiCo, which are all using the platform to build physically accurate digital twins or develop realistic immersive experiences for customers.

Enhancing Content Creation With New Connections and Libraries

The Omniverse ecosystem is expanding beyond design and content creation. In one year, Omniverse connections, ways to connect or integrate with the Omniverse platform, have grown 10x — with 82 connections through the extended Omniverse ecosystem.

  • New Third-Party Connections for Adobe Substance 3D Material Extension and Painter Connector, Epic Games Unreal Engine Connector and Maxon Cinema 4D will enable live-sync workflows between third-party apps and Omniverse.
  • New CAD Importers: These convert 26 common CAD formats to Universal Scene Description (USD) to better enable manufacturing and product design workflows within Omniverse.
  • New Asset Library Integrations: TurboSquid by Shutterstock, Sketchfab and Reallusion ActorCore assets are now directly available within Omniverse Apps asset browsers so users can simply search, drag and drop from close to 1 million Omniverse-ready 3D assets. New Omniverse-ready 3D assets, materials, textures, avatars and animations are also now available from A23D.
  • New Hydra Render Delegate Support: Users can integrate and toggle between their favorite Hydra delegate-supported renderers and the Omniverse RTX Renderer directly within Omniverse Apps. Now available in beta for Chaos V-Ray, Maxon Redshift and OTOY Octane, with Blender Cycles, Autodesk Arnold coming soon.
Chaos V-Ray Hydra Render Delegate in NVIDIA Omniverse.

There are also new connections to industrial automation and digital twin software developers. Bentley Systems, the infrastructure engineering software company, announced the availability of LumenRT for NVIDIA Omniverse, powered by Bentley iTwin. It brings engineering-grade, industrial-scale real-time physically accurate visualization to nearly 39,000 Bentley System customers worldwide. Ipolog, a developer of factory, logistics and planning software, released three new connections to the platform. This, coupled with the growing Isaac Sim robotics ecosystem, allows customers such as BMW Group to better develop holistic digital twins.

LumenRT for NVIDIA Omniverse powered by Bentley iTwin.

Omniverse Enterprise Features and Availability Broadens

New updates are coming soon to Omniverse Enterprise, including the latest releases of Omniverse Kit 103, Omniverse Create and View 2022.1, Omniverse Farm, and DeepSearch.

Omniverse Enterprise on NVIDIA LaunchPad is now available across nine global regions. NVIDIA LaunchPad gives design practitioners and project reviewers instant, free turnkey access to hands-on Omniverse Enterprise labs, helping them make quicker, more confident software and infrastructure decisions.

Customers Drive Innovations With Omniverse Enterprise

Amazon has over 200 robotics facilities that handle millions of packages each day. It’s a complex operation that requires over half a million mobile drive robots to support warehouse logistics. Using Omniverse Enterprise and Isaac Sim, Amazon Robotics is building AI-enabled digital twins of its warehouses to better optimize warehouse design and flow, and train more intelligent robotic solutions.

PepsiCo is looking at using Omniverse Enterprise and Metropolis-powered digital twins to improve the efficiency and environmental sustainability of its supply chain of over 600 distribution centers in 200 regional markets.

“NVIDIA Omniverse will help us better streamline supply chain operations and reduce energy usage and waste, while advancing our mission toward sustainability,” said Qi Wang, vice president of Research and Development at PepsiCo. “When we’re looking at new products and processes, we will use digital twins to simulate and test models and environments in real time before applying changes to the physical distribution centers.”

Lowe’s Innovation Labs is exploring how Omniverse can help unlock the next generation of its stores. It is using the platform to push the boundaries of what’s possible in digital twins, simulation and advanced tools that remove friction for customers and associates.

Kroger plans to use Omniverse to optimize store efficiency and processes with digital twin store simulation.

Raising the Bar on Industrial Digital Twins

At GTC, NVIDIA announced NVIDIA OVX, a computing system architecture designed to power large-scale digital twins. NVIDIA OVX is built to operate complex simulations that will run within Omniverse, enabling designers, engineers and planners to create physically accurate digital twins and massive, true-to-reality simulation environments.

Latest Omniverse Technologies and Features

Major new releases and capabilities announced for Omniverse include:

  • New Developer Tools: Omniverse Code, an app that serves as an integrated development environment for developers and powers users to easily build their own Omniverse extensions, apps or microservices.
  • DeepSearch: a new AI-based search service that lets users quickly search through massive, untagged 3D asset libraries using natural language or images. DeepSearch is available for Omniverse Enterprise customers in early access.
  • Omniverse Replicator: a framework for generating physically accurate 3D synthetic data to accelerate training and accuracy of perception networks — now available within Omniverse Code so developers can build their own domain-specific synthetic data engines.
  • OmniGraph, ActionGraph and AnimGraph: major new releases controlling behavior and animation.
  • Omniverse Avatar: a platform that uses AI and simulation technology to enable developers to build custom, intelligent, realistic avatars.
  • Omniverse XR app: a VR-optimized configuration of Omniverse View that enables users to experience their full-fidelity 3D scenes with full RTX ray tracing, at 1:1 scale, coming soon.
  • New versions of Omniverse Kit, Create, View and Machinima.

Read about the full release of new platform features.

To learn more about NVIDIA Omniverse, watch the GTC 2022 keynote from Jensen Huang. Register for GTC 2022 for free to attend sessions with NVIDIA and industry leaders.

The post NVIDIA Omniverse Ecosystem Expands 10x, Amid New Features and Services for Developers, Enterprises and Creators appeared first on NVIDIA Blog.

Read More

NVIDIA Unveils Isaac Nova Orin to Accelerate Development of Autonomous Mobile Robots

Next time socks, cereal or sandpaper shows up in hours delivered to your doorstep, consider the behind-the-scenes logistics acrobatics that help get them there so fast.

Order fulfillment is a massive industry of moving parts. Heavily supported by autonomous mobile robots (AMRs), warehouses can span 1 million square feet, expanding and reconfiguring to meet demands. It’s an obstacle course of workers and bottlenecks for hospitals, retailers, airports, manufacturers and others.

To accelerate development of these AMRs, we’ve introduced Isaac Nova Orin, a state-of-the-art compute and sensor reference platform. It’s built on the powerful new NVIDIA Jetson AGX Orin edge AI system, available today. The platform includes the latest sensor technologies and high-performance AI compute capability.

New Isaac Software Arrives for AMR Ecosystem

In addition to Nova Orin, which will be available later this year, we’re delivering new software and simulation capabilities to accelerate AMR deployments — including hardware-accelerated modules, or Isaac ROS GEMs, that are essential for enabling robots to visually navigate. That’s key for mobile robots to better perceive their environment to safely avoid obstacles and efficiently plan paths.

New simulation capabilities, available in the NVIDIA Isaac Sim April release, will help save time when building virtual environments to test and train AMRs. Using 3D building blocks, developers can rapidly create realistic complex warehouse scenes and configurations to validate the robot’s performance on a breadth of logistics tasks.

Isaac Nova Orin Key Features 

Nova Orin comes with all of the compute and sensor hardware needed to design, build and test autonomy in AMRs.

Its two Jetson AGX Orin units provide upto 550 TOPS of AI compute for perception, navigation and human-machine interaction. These modules process data in real time from the AMR’s central nervous system — essentially the sensor suite comprising up to six cameras, three lidars and eight ultrasonic sensors.

Nova Orin includes tools necessary to simulate the robot in Isaac Sim on Omniverse, as well as support for numerous ROS software modules designed to accelerate perception and navigation tasks. Tools are also provided for accurately mapping the robots’ environment using NVIDIA DeepMap.

The entire platform is calibrated and tested to work out of the box and give developers valuable time to innovate on new features and capabilities.

3D sensor field of Nova Orin

Enabling the Future

Much is at stake in intralogistics for AMRs, a market expected to top $46 billion by 2030, up from under $8 billion in 2021, according to estimates from ABI Research.

The old method of designing the AMR compute and sensor stack from the ground up is too costly in time and effort. Tapping into an existing platform allows manufacturers to focus on building the right software stack for the right robot application.

Improving productivity for factories and warehouses will depend on AMRs working safely and efficiently side by side at scale. High levels of autonomy driven by 3D perception from  Nova Orin will help drive that revolution.

As AMRs evolve, the need for secure deployment and management of the critical AI software on board is paramount. Over-the-air software management support is already preintegrated in Nova Orin.

Learn more about Nova Orin and the complete Isaac for AMR platform.

 

The post NVIDIA Unveils Isaac Nova Orin to Accelerate Development of Autonomous Mobile Robots appeared first on NVIDIA Blog.

Read More

Driving on Air: Lucid Group Builds Intelligent EVs on NVIDIA DRIVE

Lucid Group may be a newcomer to the electric vehicle market, but its entrance has been grand.

The electric automaker announced at GTC that its current and future fleets are built on NVIDIA DRIVE Hyperion for programmable, intelligent capabilities. By developing on the scalable, software-defined platform, Lucid ensures its vehicles are always at the cutting edge, receiving continuous improvements over the air.

The EV maker recently launched its first vehicle, the Lucid Air, late last year to widespread acclaim. The luxury sedan won MotorTrend’s 2022 Car of the Year, with industry-leading battery range and fast charging.

And Lucid isn’t stopping there — the automaker recently announced Project Gravity, a long-range electric SUV slated for launch in 2024.

A defining feature of the Lucid Air is the DreamDrive Pro advanced driver assistance system — standard on Dream Edition and Grand Touring trims, optional on other models — which leverages the high-performance compute of NVIDIA DRIVE to provide a seamless automated driving experience.

Future-Ready Intelligence

DreamDrive Pro is designed to continuously improve via over-the-air software updates, with the scalable and high-performance AI compute of NVIDIA DRIVE at the center of the system.

It uses a rich suite of 14 cameras, one lidar, five radars and 12 ultrasonics for robust automated driving and intelligent cockpit features.

In addition to a diversity of sensors, Lucid’s dual-rail power system and proprietary Ethernet Ring offer a high degree of redundancy for key systems, such as braking and steering.

“The seamless integration of the software-defined NVIDIA DRIVE platform provides a powerful basis for Lucid to further enhance what DreamDrive can do in the future — all of which can be delivered to vehicles over the air,” said Mike Bell, Senior Vice President of Digital, Lucid Group.

Together, Lucid and NVIDIA will support these intelligent vehicles, enhancing the customer experience with new functions throughout the life of the car.

A DreamDrive Come True

Lucid plans to build on its success in deploying industry-leading electric vehicles, continuing to develop on NVIDIA DRIVE for future generations.

By starting with a programmable, high-performance compute architecture in the Lucid Air, the automaker can take advantage of the scalability of NVIDIA DRIVE and always incorporate the latest AI technology as it expands with more models.

The ability to continuously deliver innovative and exciting features to its vehicles will have Lucid customers driving on air.

The post Driving on Air: Lucid Group Builds Intelligent EVs on NVIDIA DRIVE appeared first on NVIDIA Blog.

Read More

NVIDIA DRIVE Continues Industry Momentum With $11 Billion Pipeline as DRIVE Orin Enters Production

NVIDIA DRIVE Hyperion and DRIVE Orin are gaining ground in the industry.

At NVIDIA GTC, BYD, the world’s second-largest electric vehicle maker, announced it is building its next-generation fleets on the DRIVE Hyperion architecture. This platform, based on DRIVE Orin, is now in production, and powering a wide ecosystem of 25 EV makers building software-defined vehicles on high-performance, energy-efficient AI compute.

A wave of innovative startups also joined the DRIVE Hyperion ecosystem this week, including DeepRoute, Pegasus, UPower and WeRide, while luxury EV maker Lucid Motors announced its automated driving system is built on NVIDIA DRIVE.

All together, this growing ecosystem makes up an automotive pipeline that exceeds $11 billion.

The open DRIVE Hyperion 8 platform allows these companies to individualize this programmable architecture to their needs, leveraging end-to-end solutions to accelerate autonomous driving development.

The NVIDIA DRIVE Orin system-on-a-chip achieves up to  254 trillions of operations per second (TOPS) and is designed to handle the large number of applications and deep neural networks (DNNs) that run simultaneously in autonomous vehicles, with the ability to achieve systematic safety standards such as ISO 26262 ASIL-D.

Together, DRIVE Hyperion and DRIVE Orin act as the nervous system and brain of the vehicle, processing massive amounts of sensor data in real time to safely perceive, plan and act.

The World Leader in NEV

New energy vehicles are disrupting the transportation industry. They’re introducing a novel architecture that is purpose-built for software-defined functionality, enabling continuous improvement and exciting business models.

BYD is an NEV pioneer, leveraging its heritage as a rechargeable battery maker to introduce the world’s first plug-in hybrid, the F3, in 2008.

The F3 became China’s best-selling sedan the following year, and since then BYD has continued to push the limits of what’s possible for alternative powertrains, with more than 780,000 BYD electric vehicles in operation.

And now, it’s adding software-defined to the BYD fleet resume, building its coming generation of EVs on DRIVE Hyperion 8.

These vehicles will feature a programmable compute platform based on DRIVE Orin for intelligent driving and parking.

Even More Intelligent Solutions

In addition to automakers, autonomous driving startups are developing on DRIVE Hyperion to deliver software-defined vehicles.

DeepRoute, a self-driving company building robotaxis, said it is integrating DRIVE Hyperion into its level 4 system. The automotive-grade platform is key to the company’s plans to bring its production-ready vehicles to market next year.

Self-driving startup Pegasus Technology is also developing intelligent driving solutions for taxis, trucks and buses on DRIVE Hyperion to seamlessly operate on complex city roads. Its autonomous system is designed to handle lane changes, busy intersections, roundabouts, highway entrances and exits, and more.

UPower is a startup dedicated to streamlining the EV development process with its Super Board skateboard chassis. This next-generation foundation for electric vehicles will include DRIVE Hyperion for automated and autonomous driving capabilities.

Autonomous driving technology company WeRide has been building self-driving platforms for urban transportation on NVIDIA DRIVE since 2017. During GTC, the startup announced it will develop its coming generation of intelligent driving solutions on DRIVE Hyperion.

As the DRIVE Hyperion ecosystem expands, software-defined transportation will become more prevalent around the world, delivering safer, more efficient driving experiences.

The post NVIDIA DRIVE Continues Industry Momentum With $11 Billion Pipeline as DRIVE Orin Enters Production appeared first on NVIDIA Blog.

Read More

Announcing NVIDIA DRIVE Map: Scalable, Multi-Modal Mapping Engine Accelerates Deployment of Level 3 and Level 4 Autonomy

With a detailed knowledge of the world and everything in it, maps provide the foresight AI uses to make advanced and safe driving decisions.

At his GTC keynote, NVIDIA founder and CEO Jensen Huang introduced NVIDIA DRIVE Map, a multimodal mapping platform designed to enable the highest levels of autonomy while improving safety. It combines the accuracy of DeepMap survey mapping with the freshness and scale of AI-based crowdsourced mapping.

With three localization layers — camera, lidar and radar — DRIVE Map provides the redundancy and versatility required by the most advanced AI drivers.

DRIVE Map will provide survey-level ground truth mapping coverage to 500,000 kilometers of roadway in North America, Europe and Asia by 2024. This map will be continuously updated and expanded with millions of passenger vehicles.

NVIDIA DRIVE Map is available to the entire autonomous vehicle industry.

Multi-Layered 

DRIVE Map contains multiple localization layers of data for use with camera, radar and lidar modalities. The AI driver can localize to each layer of the map independently, providing the diversity and redundancy required for the highest levels of autonomy.

The camera localization layer consists of map attributes such as lane dividers, road markings, road boundaries, traffic lights, signs and poles.

DRIVE Map semantic localization layer

The radar localization layer is an aggregate point cloud of radar returns. It’s particularly useful in poor lighting conditions, which are challenging for cameras, and in poor weather conditions, which are challenging for cameras and lidars.

DRIVE Map radar localization layer

Radar localization is also useful in suburban areas where typical map attributes aren’t available, enabling the AI driver to localize based on surrounding objects that generate a radar return.

The lidar voxel layer provides the most precise and reliable representation of the environment. It builds a 3D representation of the world at 5-centimeter resolution — accuracy impossible to achieve with camera and radar.

DRIVE Map lidar voxel localization layer

Once localized to the map, the AI can use the detailed semantic information provided by the map to plan ahead and safely perform driving decisions.

Best of Both Worlds 

DRIVE Map is built with two map engines — ground truth survey map engine and crowdsourced map engine — to gather and maintain a collective memory of an Earth-scale fleet.

This unique approach combines the best of both worlds, achieving centimeter-level accuracy with dedicated survey vehicles, as well as the freshness and scale that can only be achieved with millions of passenger vehicles continuously updating and expanding the map.

The ground truth engine is based on the DeepMap survey map engine — proven technology that has been developed and verified over the past six years.

The AI-based crowdsource engine gathers map updates from millions of cars, constantly uploading new data to the cloud as the vehicles drive. The data is then aggregated at full fidelity in NVIDIA Omniverse and used to update the map, providing the real-world fleet fresh over-the-air map updates within hours.

DRIVE Map also provides a data interface, DRIVE MapStream, to allow any passenger car that meets the DRIVE Map requirements to continuously update the map using camera, radar and lidar data.

Earth-Scale Digital Twin

In addition to assisting the AI to make the optimal driving decisions, DRIVE Map accelerates AV deployment, from generating ground-truth training data for deep neural network training, as well as for testing and validation.

These workflows are centered on Omniverse, where real-world map data is loaded and stored. Omniverse maintains an Earth-scale representation of the digital twin that is continuously updated and expanded by survey map vehicles and millions of passenger vehicles.

Using automated content generation tools built on Omniverse, the detailed map is converted into a drivable simulation environment that can be used with NVIDIA DRIVE Sim. Features such as road elevation, road markings, islands, traffic signals, signs and vertical posts are accurately replicated at centimeter-level accuracy.

With physically based sensor simulation and domain randomization, AV developers can use the simulated environment to generate training scenarios that aren’t available in real data.

AV developers can also apply scenario generation tools to test AV software on digital twin environments before deploying AV in the real world. Finally, the digital twin provides fleet operators a complete virtual view of where the vehicles are driving in the world, assisting remote operation when needed.

As a highly versatile and scalable platform, DRIVE Map equips the AI driver with the understanding of the world needed to continuously advance autonomous capabilities.

The post Announcing NVIDIA DRIVE Map: Scalable, Multi-Modal Mapping Engine Accelerates Deployment of Level 3 and Level 4 Autonomy appeared first on NVIDIA Blog.

Read More

Introducing NVIDIA DRIVE Hyperion 9: Next-Generation Platform for Software-Defined Autonomous Vehicle Fleets

NVIDIA DRIVE Hyperion is taking software-defined vehicle architectures to the next level.

At his GTC keynote, NVIDIA founder and CEO Jensen Huang announced DRIVE Hyperion 9, the next generation of the open platform for automated and autonomous vehicles. The programmable architecture, slated for 2026 production vehicles, is built on multiple DRIVE Atlan computers to achieve intelligent driving and in-cabin functionality.

DRIVE Hyperion is designed to be compatible across generations, with the same computer form factor and NVIDIA DriveWorks APIs. Partners can leverage current investments in the DRIVE Orin platform and seamlessly migrate to NVIDIA DRIVE Atlan and beyond.

The platform includes the computer architecture, sensor set and full NVIDIA DRIVE Chauffeur and Concierge applications. It is designed to be open and modular, so customers can select what they need. Current-generation systems scale from NCAP to level 3 driving and level 4 parking with advanced AI cockpit capabilities.

Core Compute

DRIVE Hyperion incorporates redundancy into the architecture’s compute.

With the DRIVE Atlan SoC, the next-generation platform will feature more than double the performance of the current DRIVE Orin-based architecture at the same power envelope. This compute is capable of handling level 4 autonomous driving, as well as the convenience and safety features provided by NVIDIA DRIVE Concierge.

DRIVE ​​Atlan is a technical marvel for safe and secure AI computing, fusing all of NVIDIA’s technologies in AI, automotive, robotics, safety and BlueField data centers.

Leveraging NVIDIA’s high-performance GPU architecture, Arm CPU cores and deep learning and computer vision accelerators, it provides ample compute horsepower for redundant and diverse deep neural networks and leaves headroom for developers to continue adding features and improvements.

DRIVE Hyperion is the nervous system of the vehicle, and DRIVE Atlan serves as the brain.

Heightened Sensing

With DRIVE Atlan’s compute performance, DRIVE Hyperion 9 can process even more sensor data as the car drives, improving redundancy and diversity.

This upgraded sensor suite includes surround imaging radar, enhanced cameras with higher frame rates, two additional side lidar and improved undercarriage sensing with better camera and ultrasonic placement.

In total, the DRIVE Hyperion 9 architecture includes 14 cameras, nine radars, three lidars and 20 ultrasonics for automated and autonomous driving, as well as three cameras and one radar for interior occupant sensing.

By incorporating a rich sensor set and high-performance compute, the entire system is architected to the highest levels of functional safety and cybersecurity.

DRIVE Hyperion 9 will begin production in 2026, giving the industry continuous access to the cutting edge in AI technology as it begins to roll out more intelligent transportation.

The post Introducing NVIDIA DRIVE Hyperion 9: Next-Generation Platform for Software-Defined Autonomous Vehicle Fleets appeared first on NVIDIA Blog.

Read More

Siemens Gamesa Taps NVIDIA Digital Twin Platform for Scientific Computing to Accelerate Clean Energy Transition

Siemens Gamesa Renewable Energy is working with NVIDIA to create physics-informed digital twins of wind farms — groups of wind turbines used to produce electricity.

The company has thousands of turbines around the globe that light up schools, homes, hospitals and factories with clean energy. In total they generate over 100 gigawatts of wind power, enough to power nearly 87 million households annually.

Virtual representations of Siemens Gamesa’s wind farms will be built using NVIDIA Omniverse and Modulus, which together comprise NVIDIA’s digital twin platform for scientific computing.

The platform will help Siemens Gamesa achieve quicker calculations to optimize wind farm layouts, which is expected to lead to farms capable of producing up to 20 percent more power than previous designs.

With the global level of annual wind power installations likely to quadruple between 2020 and 2025, it’s more important than ever to maximize the power produced by each turbine.

The global trillion-dollar renewable energy industry is turning to digital twins, like those of Siemens Gamesa’s wind farms — and one of Earth itself — to further climate research and accelerate the clean energy transition.

And the world’s rapid clean energy technology improvements mean that a dollar spent on wind and solar conversion systems today results in 4x more electricity than a dollar spent on the same systems a decade ago. This has tremendous bottom-line implications for the transition towards a greener Earth.

With NVIDIA Modulus, an AI framework for developing physics-informed machine learning models, and Omniverse, a 3D design collaboration and world simulation platform, researchers can now simulate computational fluid dynamics up to 4,000x faster than traditional methods — and view the simulations at high fidelity.

“The collaboration between Siemens Gamesa and NVIDIA has meant a great step forward in accelerating the computational speed and the deployment speed of our latest algorithms development in such a complex field as computational fluid dynamics,” said Sergio Dominguez,  onshore digital portfolio manager at Siemens Gamesa.

Maximizing Wind Power

Adding a turbine next to another on a farm can change the wind flow and create wake effects — that is, decreases in downstream wind speed — which lead to a reduction in the farm’s production of electricity.

Omniverse digital twins of wind farms will help Siemens Gamesa to accurately simulate the effect that a turbine might have on another when placed in close proximity.

Using NVIDIA Modulus and physics-ML models running on GPUs, researchers can now run computational fluid dynamics simulations orders of magnitude faster than with traditional methods, like those based on Reynolds-averaged Navier-Stokes equations or large eddy simulations, which can take over a month to run, even on a 100-CPU cluster.

This up to 4,000x speedup allows the rapid and accurate simulation of wake effects.

Analyzing and minimizing potential wake effects in real time, while simultaneously optimizing wind farms for a variety of other wind and weather scenarios, require hundreds or thousands of iterations and simulation runs, which were traditionally prohibited by time constraints and costs.

NVIDIA Omniverse and Modulus enable accurate simulations of the complex interactions between the turbines, using high-fidelity and high-resolution models that are based on low-resolution inputs.

Learn more about NVIDIA Omniverse and Modulus at GTC, running through March 24.

Watch NVIDIA founder and CEO Jensen Huang’s GTC keynote address.

The post Siemens Gamesa Taps NVIDIA Digital Twin Platform for Scientific Computing to Accelerate Clean Energy Transition appeared first on NVIDIA Blog.

Read More