Meta Works with NVIDIA to Build Massive AI Research Supercomputer

Meta Platforms gave a big thumbs up to NVIDIA, choosing our technologies for what it believes will be its most powerful research system to date.

The AI Research SuperCluster (RSC), announced today, is already training new models to advance AI.

Once fully deployed, Meta’s RSC is expected to be the largest customer installation of NVIDIA DGX A100 systems.

“We hope RSC will help us build entirely new AI systems that can, for example, power real-time voice translations to large groups of people, each speaking a different language, so they could seamlessly collaborate on a research project or play an AR game together,” the company said in a blog.

Training AI’s Largest Models

When RSC is fully built out, later this year, Meta aims to use it to train AI models with more than a trillion parameters. That could advance fields such as natural-language processing for jobs like identifying harmful content in real time.

In addition to performance at scale, Meta cited extreme reliability, security, privacy and the flexibility to handle “a wide range of AI models” as its key criteria for RSC.

Meta RSC system
Meta’s AI Research SuperCluster features hundreds of NVIDIA DGX systems linked on an NVIDIA Quantum InfiniBand network to accelerate the work of its AI research teams.

Under the Hood

The new AI supercomputer currently uses 760 NVIDIA DGX A100 systems as its compute nodes. They pack a total of 6,080 NVIDIA A100 GPUs linked on an NVIDIA Quantum 200Gb/s InfiniBand network to deliver 1,895 petaflops of TF32 performance.

Despite challenges from COVID-19, RSC took just 18 months to go from an idea on paper to a working AI supercomputer (shown in the video below) thanks in part to the NVIDIA DGX A100 technology at the foundation of Meta RSC.



20x Performance Gains

It’s the second time Meta has picked NVIDIA technologies as the base for its research infrastructure. In 2017, Meta built the first generation of this infrastructure for AI research with 22,000 NVIDIA V100 Tensor Core GPUs that handles 35,000 AI training jobs a day.

Meta’s early benchmarks showed RSC can train large NLP models 3x faster and run computer vision jobs 20x faster than the prior system.

In a second phase later this year, RSC will expand to 16,000 GPUs that Meta believes will deliver a whopping 5 exaflops of mixed precision AI performance. And Meta aims to expand RSC’s storage system to deliver up to an exabyte of data at 16 terabytes per second.

A Scalable Architecture

NVIDIA AI technologies are available to enterprises of any size.

NVIDIA DGX, which includes a full stack of NVIDIA AI software, scales easily from a single system to a DGX SuperPOD running on-premises or at a colocation provider. Customers can also rent DGX systems through NVIDIA DGX Foundry.

The post Meta Works with NVIDIA to Build Massive AI Research Supercomputer appeared first on The Official NVIDIA Blog.

Read More

How the Intelligent Supply Chain Broke and AI Is Fixing It

Let’s face it, the global supply chain may not be the most scintillating subject matter. Yet in homes and businesses around the world, it’s quickly become the topic du jour: empty shelves; record price increases; clogged ports and sick truckers leading to disruptions near and far.

The business of organizing resources to supply a product or service to its final user feels like it’s never been more challenged by so many variables. Shortages of raw materials, everything from resin and aluminum to paint and semiconductors, are nearing historic levels. Products that do get manufactured sit on cargo ships or in warehouses due to shortages of containers and workers and truck drivers that help deliver them to their final destinations. And consumer pocketbooks and paychecks are getting squeezed by rising prices.

The $9 trillion logistics industry is responding by investing in automation and using AI and big data to gain more insights throughout the supply chain. Big money is being poured into supply-chain technology startups, which raised $24.3 billion in venture funding in the first three quarters of 2021, 58 percent more than the full-year total for 2020, according to analytics firm PitchBook Data Inc.

Investing in AI

Behind these investments, businesses see technology and accelerated computing as key to finding firmer ground. At Manifest 2022, a logistics and supply chain conference taking place in Las Vegas, the industry is discussing how to refine supply chains and create cost efficiencies using AI and machine learning. Among their goals: address labor shortages, improve throughput in distribution centers, and route deliveries more efficiently.

Take a box of cereal. Getting it from the warehouse to a home has never been more expensive. Employee turnover rates of 30 percent to 46 percent in warehouses and distribution centers are just part of the problem.

To mitigate the challenge, Dematic, a global materials-handling company, is evaluating software from companies like Kinetic Vision, which has developed computer vision applications on the NVIDIA AI platform that add intelligence to automated warehouse systems.

Companies like Kinetic Vision and SF Technology use video data from cameras to optimize every step of the package lifecycle, accelerating throughput by up to 20 percent and reducing conveyor downtime, which can cost retailers $3,000 to $5,000 a minute.

Autonomous robot companies such as Gideon, 6 River Systems and Symbotic also use the NVIDIA AI platform to improve distribution center throughput with their autonomous guided vehicles that transport material efficiently within the warehouse or distribution centers.

And with NVIDIA Fleet Command, which securely deploys, manages and scales AI applications via the cloud across distributed edge infrastructure, these solutions can be remotely deployed and managed securely and at scale across hundreds of distribution centers.

Digital Twins and Simulation

Improving layouts of stores and distribution centers also has become key to achieving cost efficiencies. NVIDIA Omniverse, a virtual world simulation and 3D design collaboration platform, makes it possible to virtually design and simulate distribution centers at full fidelity. Users can improve workflows and throughput with photorealistic, physically accurate virtual environments.

Retailers could, for example, develop a solution on the Omniverse platform to design, test and simulate the flow of material and employee processes in digital twins of their distribution centers and then bring those optimizations into the real world.

Digital human simulations could test new workflows for employee ergonomics and productivity. And robots are trained and operated with the NVIDIA Isaac robotics platform, creating the most efficient layout and workflows.

Kinetic Vision is using NVIDIA Omniverse to deliver digital twins technology and simulation to optimize factories and retail and consumer packaged goods distribution centers.

Leaning In

While manufacturers, supply chain operators and retailers each will have their own approaches to solving challenges, they’re leaning in on AI as a key differentiator.

Successfully implementing AI-enabled supply-chain management has enabled early adopters to improve logistics costs by 15 percent, inventory levels by 35 percent and service levels by 65 percent, compared with slower-moving competitors, according to McKinsey.

With some experts predicting the global supply chain won’t return to a new normal until at least 2023, companies are moving to take measures that matter most to the bottom line.

For more on how NVIDIA AI is powering the most innovative AI solutions for the supply chain and logistics industry attend the following talks at Manifest:

  • A fireside chat, “Bringing Agility and Flexibility to Distribution Centers with AI,” on Wednesday, Jan. 26, at 2 p.m. Pacific, in Champagne 4 with Azita Martin, vice president and general manager of AI for retail at NVIDIA, and Michael Larsson, CEO of North America region at Dematic.
  • A presentation “The Next Frontier in Warehouse Intelligence” on the same date, at 11:30 a.m. Pacific, in Champagne 4 with Azita Martin and Omer Rashid, vice president of Solutions Designs at DHL Supply Chain, and Renato Bottiglieri, chief logistics officer at Eggo Kitchen & House.

The post How the Intelligent Supply Chain Broke and AI Is Fixing It appeared first on The Official NVIDIA Blog.

Read More

NVIDIA GPUs Enable Simulation of a Living Cell

Every living cell contains its own bustling microcosm, with thousands of components responsible for energy production, protein building, gene transcription and more.

Scientists at the University of Illinois at Urbana-Champaign have built a 3D simulation that replicates these physical and chemical characteristics at a particle scale — creating a fully dynamic model that mimics the behavior of a living cell.

Published in the journal Cell, the project simulates a living minimal cell, which contains a pared-down set of genes essential for the cell’s survival, function and replication. The model uses NVIDIA GPUs to simulate 7,000 genetic information processes over a 20-minute span of the cell cycle – making it what the scientists believe is the longest, most complex cell simulation to date.

Minimal cells are simpler than naturally occurring ones, making them easier to recreate digitally.

“Even a minimal cell requires 2 billion atoms,” said Zaida Luthey-Schulten, chemistry professor and co-director of the university’s Center for the Physics of Living Cells. “You cannot do a 3D model like this in a realistic human time scale without GPUs.”

Once further tested and refined, whole-cell models can help scientists predict how changes to the conditions or genomes of real-world cells will affect their function. But even at this stage, minimal cell simulation can give scientists insight into the physical and chemical processes that form the foundation of living cells.

“What we found is that fundamental behaviors emerge from the simulated cell — not because we programmed them in, but because we had the kinetic parameters and lipid mechanisms correct in our model,” she said.

Lattice Microbes, the GPU-accelerated software co-developed by Luthey-Schulten and used to simulate the 3D minimal cell, is available on the NVIDIA NGC software hub.

Minimal Cell With Maximum Realism

To build the living cell model, the Illinois researchers simulated one of the simplest living cells, a parasitic bacteria called mycoplasma. They based the model on a trimmed-down version of a mycoplasma cell synthesized by scientists at J. Craig Venter Institute in La Jolla, Calif., which had just under 500 genes to keep it viable.

For comparison, a single E. coli cell has around 5,000 genes. A human cell has more than 20,000.

Luthy-Schulten’s team then used known properties of the mycoplasma’s inner workings, including amino acids, nucleotides, lipids and small molecule metabolites to build out the model with DNA, RNA, proteins and membranes.

“We had enough of the reactions that we could reproduce everything known,” she said.

Using Lattice Microbes software on NVIDIA Tensor Core GPUs, the researchers ran a 20-minute 3D simulation of the cell’s life cycle, before it starts to substantially expand or replicate its DNA. The model showed that the cell dedicated most of its energy to transporting molecules across the cell membrane, which fits its profile as a parasitic cell.

“If you did these calculations serially, or at an all-atom level, it’d take years,” said graduate student and paper lead author Zane Thornburg. “But because they’re all independent processes, we could bring parallelization into the code and make use of GPUs.”

Thornburg is working on another GPU-accelerated project to simulate growth and cell division in 3D. The team has recently adopted NVIDIA DGX systems and RTX A5000 GPUs to further accelerate its work, and found that using A5000 GPUs sped up the benchmark simulation time by 40 percent compared to a development workstation with a previous-generation NVIDIA GPU.

Learn more about researchers using NVIDIA GPUs to accelerate science breakthroughs by registering free for NVIDIA GTC, running online March 21-24.

Main image is a snapshot from the 20-minute 3D spatial simulation, showing yellow and purple ribosomes, red and blue degradasomes, and smaller spheres representing DNA polymers and proteins.

The post NVIDIA GPUs Enable Simulation of a Living Cell appeared first on The Official NVIDIA Blog.

Read More

GFN Thursday: ‘Tom Clancy’s Rainbow Six Extraction’ Charges Into GeForce NOW

Hello, Operator.

This GFN Thursday brings the launch of Tom Clancy’s Rainbow Six Extraction to GeForce NOW.

Plus, four new games are joining the GeForce NOW library to let you start your weekend off right.

Your New Mission, Should You Choose to Accept It

Grab your gadgets and get ready to game. Tom Clancy’s Rainbow Six Extraction releases today and is available to stream on GeForce NOW with DLSS for higher frame rates and beautiful, sharp images.

Join millions of players in the Rainbow Six universe. Charge in on your own or battle with buddies in a squad of up to three in thrilling co-op gameplay.

Select from 18 different Operators with specialized skills and progression paths that sync with your strategy to take on different challenges. Play riveting PvE on detailed containment zones, collect critical information and fight an ever-evolving, highly lethal alien threat known as the Archaeans that’s reshaping the battlefield.

Playing With the Power of GeForce RTX 3080

Members can stream Tom Clancy’s Rainbow Six Extraction and the 1,100+ games on the GeForce NOW library, including nearly 100 free-to-play titles, with all of the perks that come with the new GeForce NOW RTX 3080 membership.

Rainbow Six Extraction on GeForce NOW
Build your team, pick your strategy and complete challenging missions in Tom Clancy’s Rainbow Six Extraction.

This new tier of service allows members to play across their devices – including underpowered PCs, Macs, Chromebooks, SHIELD TVs, Android devices, iPhones or iPads – with the power of GeForce RTX 3080. That means benefits like ultra-low latency and eight-hour gaming sessions — the longest available — for a maximized experience on the cloud.

Plus, RTX 3080 members have the ability to fully control and customize in-game graphics settings, with RTX ON rendering environments in cinematic quality for supported games.

For more information, check out our membership FAQ.

New Games Dropping This Week

Garfield Kart Furious Racing on GeForce NOW
It’s fast. It’s furry. It’s Garfield Kart – Furious Racing.

The fun doesn’t stop. Members can look for the following titles joining the GFN Thursday library this week:

  • Tom Clancy’s Rainbow Six Extraction (New release on Ubisoft Connect, Jan. 20)
  • Blacksmith Legends (Steam)
  • Fly Corp (Steam)
  • Garfield Kart – Furious Racing (Steam)

We make every effort to launch games on GeForce NOW as close to their release as possible, but, in some instances, games may not be available immediately.

Finally, we’ve got a question for you and your gaming crew this week. Talk to us on Twitter or in the comments below.

The post GFN Thursday: ‘Tom Clancy’s Rainbow Six Extraction’ Charges Into GeForce NOW appeared first on The Official NVIDIA Blog.

Read More

Van, Go: Pony.ai Unveils Next-Gen Robotaxi Fleet Built on NVIDIA DRIVE Orin

Robotaxis are on their way to delivering safer transportation, driving across various landscapes and through starry nights.

This week, Silicon Valley-based self-driving startup Pony.ai announced its next-generation autonomous computing platform, built on NVIDIA DRIVE Orin for high-performance and scalable compute. The centralized system will serve as the brain for a robotaxi fleet of Toyota Sienna multipurpose vehicles (MPVs), marking a major leap forward for the nearly six-year-old company.

The AI compute platform enables multiple configurations for scalable autonomous driving development, all the way to level 4 self-driving vehicles.

“By leveraging the world-class NVIDIA DRIVE Orin SoC, we’re demonstrating our design and industrialization capabilities and ability to develop and deliver a powerful mass-production platform at an unprecedented scale,” said James Peng, co-founder and CEO of Pony.ai, which is developing autonomous systems for both robotaxis and trucks.

The transition to DRIVE Orin has significantly accelerated the company’s plans to deploy safer, more efficient robotaxis, with road testing set to begin this year in China and commercial rollout planned for 2023.

State-of-the-Art Intelligence

DRIVE Orin serves as the brain of autonomous fleets, enabling them to perceive their environment and continuously improve over time.

Born out of the data center, DRIVE Orin achieves 254 trillions of operations per second, or TOPS. It’s designed to handle the large number of applications and deep neural networks that run simultaneously in autonomous trucks, while achieving systematic safety standards such as ISO 26262 ASIL-D.

Pony.ai’s DRIVE Orin-based autonomous computing unit features low latency, high performance and high reliability. It also incorporates a robust sensor solution that contains more than 23 sensors, including solid-state lidars, near-range lidars, radars and cameras.

The Pony.ai next-generation autonomous computing platform, built on NVIDIA DRIVE Orin.

This next-generation, automotive-grade system incorporates redundancy and diversity, maximizing safety while increasing performance and reducing weight and cost over previous iterations.

A Van for All Seasons

The Toyota Sienna MPV is a prime candidate for robotaxi services as it offers flexibility and ride comfort in a sleek package.

Toyota and Pony.ai began co-developing Sienna vehicles purpose-built for robotaxi services in 2019. The custom vehicles feature a dual-redundancy system and better control performance for level 4 autonomous driving capabilities.

The vehicles also debut new concept design cues, including rooftop signaling units that employ different colors and lighting configurations to communicate the robotaxi’s status and intentions.

This dedicated, future-forward design combined with the high-performance compute of NVIDIA DRIVE Orin lays a strong foundation for the coming generation of safer, more efficient robotaxi fleets.

The post Van, Go: Pony.ai Unveils Next-Gen Robotaxi Fleet Built on NVIDIA DRIVE Orin appeared first on The Official NVIDIA Blog.

Read More

New NVIDIA AI Enterprise Release Lights Up Data Centers

With a new year underway, NVIDIA is helping enterprises worldwide add modern workloads to their mainstream servers using the latest release of the NVIDIA AI Enterprise software suite.

NVIDIA AI Enterprise 1.1 is now generally available. Optimized, certified and supported by NVIDIA, the latest version of the software suite brings new updates including production support for containerized AI with the NVIDIA software on VMware vSphere with Tanzu, which was previously only available on a trial basis. Now, enterprises can run accelerated AI workloads on vSphere, running in both Kubernetes containers and virtual machines with NVIDIA AI Enterprise to support advanced AI development on mainstream IT infrastructure.

Enterprise AI Simplified with VMware vSphere with Tanzu, Coming Soon to NVIDIA LaunchPad

Among the top customer-requested features in NVIDIA AI Enterprise 1.1 is production support for running on VMware vSphere with Tanzu, which enables developers to run AI workloads on both containers and virtual machines within their vSphere environments. This new milestone in the AI-ready platform curated by NVIDIA and VMware provides an integrated, complete stack of containerized software and hardware optimized for AI, all fully managed by IT.

NVIDIA will soon add VMware vSphere with Tanzu support to the NVIDIA LaunchPad program for NVIDIA AI Enterprise, available at nine Equinix locations around the world. Qualified enterprises can test and prototype AI workloads at no charge through curated labs designed for the AI practitioner and IT admin. The labs showcase how to develop and manage common AI workloads like chatbots and recommendation systems, using NVIDIA AI Enterprise and VMware vSphere, and soon with Tanzu.

“Organizations are accelerating AI and ML development projects and VMware vSphere with Tanzu running NVIDIA AI Enterprise easily empowers AI development requirements with modern infrastructure services,” said Matt Morgan, vice president of Product Marketing, Cloud Infrastructure Business Group at VMware. “This announcement marks another key milestone for VMware and NVIDIA in our sustained efforts to help teams leverage AI across the enterprise.”

Growing Demand for Containerized AI Development

While enterprises are eager to use containerized development for AI, the complexity of these workloads requires orchestration across many layers of infrastructure. NVIDIA AI Enterprise 1.1 provides an ideal solution for these challenges as an AI-ready enterprise platform.

“AI is a very popular modern workload that is increasingly favoring deployment in containers. However, deploying AI capabilities at scale within the enterprise can be extremely complex, requiring enablement at multiple layers of the stack, from AI software frameworks, operating systems, containers, VMs, and down to the hardware,” said Gary Chen, research director, Software Defined Compute at IDC. “Turnkey, full-stack AI solutions can greatly simplify deployment and make AI more accessible within the enterprise.”

Domino Data Lab MLOps Validation Accelerates AI Research and Data Science Lifecycle

The 1.1 release of NVIDIA AI Enterprise also provides validation for the Domino Data Lab Enterprise MLOps Platform with VMware vSphere with Tanzu. This new integration enables more companies to cost-effectively scale data science by accelerating research, model development, and model deployment on mainstream accelerated servers.

“This new phase of our collaboration with NVIDIA further enables enterprises to solve the world’s most challenging problems by putting models at the heart of their businesses,” said Thomas Robinson, vice president of Strategic Partnerships at Domino Data Lab. “Together, we are providing every company the end-to-end platform to rapidly and cost-effectively deploy models enterprise-wide.”

NVIDIA AI Enterprise 1.1 stack diagram
NVIDIA AI Enterprise 1.1 features support for VMware vSphere with Tanzu and validation for the Domino Data Lab Enterprise MLOps Platform.

New OEMs and Integrators Offering NVIDIA-Certified Systems for NVIDIA AI Enterprise

Amidst the new release of NVIDIA AI Enterprise, the industry ecosystem is expanding with the first NVIDIA-Certified Systems from Cisco and Hitachi Vantara, as well as a growing roster of newly qualified system integrators offering solutions for the software suite.

The first Cisco system to be NVIDIA-Certified for NVIDIA AI Enterprise is the Cisco UCS C240 M6 rack server with NVIDIA A100 Tensor Core GPUs. The two-socket, 2RU form factor can power a wide range of storage and I/O-intensive applications, such as big data analytics, databases, collaboration, virtualization, consolidation and high-performance computing.

“At Cisco we are helping simplify customers’ hybrid cloud and cloud-native transformation. NVIDIA-Certified Cisco UCS servers, powered by Cisco Intersight, deliver the best-in-class AI workload experiences in the market,” said Siva Sivakumar, vice president of product management at Cisco. “The certification of the Cisco UCS C240 M6 rack server for NVIDIA AI Enterprise allows customers to add AI using the same infrastructure and management software deployed throughout their data center.”

The first NVIDIA-Certified System from Hitachi Vantara compatible with NVIDIA AI Enterprise is the Hitachi Advanced Server DS220 G2 with NVIDIA A100 Tensor Core GPUs. The general-purpose, dual-processor server is optimized for performance and capacity, and delivers a balance of compute and storage with the flexibility to power a wide range of solutions and applications.

“For many enterprises, cost is an important consideration when deploying new technologies like AI-powered quality control, recommender systems, chatbots and more,” said Dan McConnell, senior vice president, Product Management at Hitachi Vantara. “Accelerated with NVIDIA A100 GPUs and now certified for NVIDIA AI Enterprise, Hitachi Unified Compute Platform (UCP) solutions using the Hitachi Advanced Server DS220 G2 gives customers an ideal path for affordably integrating powerful AI-ready infrastructure to their data centers.”

A broad range of additional server manufacturers offer NVIDIA-Certified Systems for NVIDIA AI Enterprise. These include Atos, Dell Technologies, GIGABYTE, H3C, Hewlett Packard Enterprise, Inspur, Lenovo and Supermicro, all of whose systems feature NVIDIA A100, NVIDIA A30 or other NVIDIA GPUs. Customers can also choose to deploy NVIDIA AI Enterprise on their own servers or on as-a-service bare metal infrastructure from Equinix Metal across nine regions globally.

AMAX, Colfax International, Exxact Corporation and Lambda are the newest system integrators qualified for NVIDIA AI Enterprise, joining a global ecosystem of channel partners that includes Axians, Carahsoft Technology Corp., Computacenter, Insight Enterprises, NTT, Presidio, Sirius, SoftServe, SVA System Vertrieb Alexander GmbH, TD SYNNEX, Trace3 and World Wide Technology.

Enterprises interested in experiencing development with NVIDIA AI Enterprise can apply for instant access to no cost using curated labs via the NVIDIA LaunchPad program, which also features labs using NVIDIA Fleet Command for edge AI, as well as NVIDIA Base Command for demanding AI development workloads.

The post New NVIDIA AI Enterprise Release Lights Up Data Centers appeared first on The Official NVIDIA Blog.

Read More

Fusing Art and Tech: MORF Gallery CEO Scott Birnbaum on Digital Paintings, NFTs and More

Browse through MORF Gallery — virtually or at an in-person exhibition — and you’ll find robots that paint, digital dreamscape experiences, and fine art brought to life by visual effects.

The gallery showcases cutting-edge, one-of-a-kind artwork from award-winning artists who fuse their creative skills with AI, machine learning, robotics and neuroscience.

Scott Birnbaum, CEO and co-founder of MORF Gallery, a Silicon Valley startup, spoke with NVIDIA AI Podcast host Noah Kravitz about digital art, non-fungible tokens, as well as ArtStick, a plug-in device that turns any TV into a premium digital art gallery.

Key Points From This Episode:

  • Artists featured by MORF Gallery create fine art using cutting-edge technology. For example, robots help with mundane tasks like painting backgrounds. Visual effects add movement to still paintings. And machine learning can help make NeoMasters — paintings based on original works that were once lost but resurrected or recreated with AI’s help.
  • The digital art space offers new and expanding opportunities for artists, technologists, collectors and investors. For one, non-fungible tokens, Birnbaum says, have been gaining lots of attention recently. He gives an overview of NFTs and how they authenticate original pieces of digital art.

Tweetables:

Paintbrushes, cameras, computers and AI are all technologies that “move the art world forward … as extensions of human creativity.” — Scott Birnbaum [8:27]

“Technology is enabling creative artists to really push the boundaries of what their imaginations can allow.” — Scott Birnbaum [13:33]

You Might Also Like:

Art(ificial) Intelligence: Pindar Van Arman Builds Robots That Paint

Pindar Van Arman, an American artist and roboticist, designs painting robots that explore the differences between human and computational creativity. Since his first system in 2005, he has built multiple artificially creative robots. The most famous, Cloud Painter, was awarded first place at Robotart 2018.

Real or Not Real? Attorney Steven Frank Uses Deep Learning to Authenticate Art

Steven Frank is a partner at the law firm Morgan Lewis, specializing in intellectual property and commercial technology law. He’s also half of the husband-wife team that used convolutional neural networks to authenticate artistic masterpieces, including Da Vinci’s Salvador Mundi, with AI’s help.

Researchers Chris Downum and Leszek Pawlowicz Use Deep Learning to Accelerate Archaeology

Researchers in the Department of Anthropology at Northern Arizona University are using GPU-based deep learning algorithms to categorize sherds — tiny fragments of ancient pottery.

Subscribe to the AI Podcast

Get the AI Podcast through iTunes, Google Podcasts, Google Play, Castbox, DoggCatcher, Overcast, PlayerFM, Pocket Casts, Podbay, PodBean, PodCruncher, PodKicker, Soundcloud, Spotify, Stitcher and TuneIn. If your favorite isn’t listed here, drop us a note.

Make the AI Podcast Better

Have a few minutes to spare? Fill out this listener survey. Your answers will help us make a better podcast.

 

 

The post Fusing Art and Tech: MORF Gallery CEO Scott Birnbaum on Digital Paintings, NFTs and More appeared first on The Official NVIDIA Blog.

Read More

Billions Served: NVIDIA Merlin Helps Fuel Clicks for Online Giants

Online commerce has rocketed to trillions of dollars worldwide in the past decade, serving billions of consumers. Behind the scenes of this explosive growth in online sales is personalization driven by recommender engines.

Recommenders make shopping deeply personalized. While searching for products on e-commerce sites, they find you. Or suggestions can just appear. This wildly delightful corner of the internet is driven by ever more massive datasets and models.

NVIDIA Merlin is the rocket fuel of recommenders. Boosting training and inference, it enables businesses of all types to better harness data to build recommenders accelerated by NVIDIA GPUs.

The stakes are higher than ever for online businesses. Online sales in 2021 were expected to reach nearly $5 trillion worldwide, according to eMarketer, up nearly 17 percent from the prior year.

On some of the world’s largest online sites, even a 1 percent gain in relevance accuracy of recommendations can yield billions more sales.

Investment in recommender systems has become one of the biggest competitive advantages of internet giants today.

The market for recommenders is expected to reach $15.13 billion by 2026, up from $2.12 billion in 2020, according to Mordor Intelligence. The largest and fastest growing segment of the market for recommender engines is in the Asia Pacific region, according to the research firm.

But an industry challenge is that improved relevance requires more data and processing. This data consists of trillions of user-product interactions — clicks, views,  — on billions of products and consumer profiles.

Data of this scale can take days to train models. Yet the faster you can spin out new models informed by more data, the better your relevance.

The Merlin collection of models, methods, and libraries, includes tools for building deep learning-based systems capable of handling terabytes of data that can provide better predictions and increase clicks.

SNAP Taps Merlin and GPUs for Inference Upside

U.S. digital advertising is expected to reach $191.1 billion in 2021, up 25.5 percent from the year before, according to eMarketer.

Snap, parent company to social media app Snapchat, is based in Santa Monica, Calif., and has more than 300 million daily active users. It creates ad revenue from its social photo and video messaging service.

“We will continue to focus on delivering strong results for our advertising partners and innovating to expand the capabilities of our platform and better serve our community,” said Snap CEO Evan Spiegel in its third-quarter earnings statement.

The technical hurdle for Snap is that it seeks to continue to develop its workload’s higher-cost ranking models and expand into more complex models while reducing costs.

The company used NVIDIA GPUs and Merlin to boost its content ranking capabilities.

“Snap used NVIDIA GPUs and Merlin software to improve machine learning inference cost efficiency by 50 percent and decrease serving latency by 2x, providing the compute headroom to experiment and deploy heavier, more accurate ad and content ranking models,” said Nima Khajehnouri, VP of engineering at Snap.

Tencent Boosts Model Training With Merlin’s HugeCTR

Entertainment giant Tencent, which operates the enormously popular messaging service WeChat and payments platform WeChat Pay, is China’s largest company by market capitalization.

Its engineers need to rapidly iterate on models for its advertising recommendation system, putting increasing demands on its training performance.

“The advertising business is a relatively important business inside Tencent and the recommendation system is used to increase the overall advertising revenue,” said Xiangting Kong, expert engineer at Tencent.

The problem is that accuracy of advertising recommendation can only be improved by training more sample data, including more sample features, but this leads to longer training times that affect model update frequency.

“HugeCTR, as a recommendation training framework, is integrated into the advertising recommendation training system to make the update frequency of model training faster, and more samples can be trained to improve online effects,” he said.

After the training model performance is improved, more data can be trained to improve the accuracy of the model, increasing advertising revenue, he added.

Meituan Reduces Costs With NVIDIA A100 GPUs

Meituan’s business is at a crowded intersection of food, entertainment and on-demand services, among its 200 service categories. The Chinese internet giant has more than 667 million active users and 8.3 million active merchants.

Jun Huang, a senior technical expert at Meituan, said that if his team can greatly improve performance, it usually prefers to train more samples and more complex models.

The problem for Meituan was that as its models became more and more complex, it became difficult to optimize the training framework deeply, said Huang.

“We are working on integrating NVIDIA HugeCTR into our training system based on A100 GPUs. The cost is also greatly reduced. This is a preliminary optimization result, and there is still much room to optimize in the future,” he said.

Meituan recently reported its average number of transactions per transacting users increased to 32.8 for the trailing 12 months of the second quarter of 2021, compared with 25.7 for the trailing 12 months of the second quarter of 2020.

Learn more about NVIDIA Merlin. Learn more about NVIDIA Triton.

The post Billions Served: NVIDIA Merlin Helps Fuel Clicks for Online Giants appeared first on The Official NVIDIA Blog.

Read More

From Imagination to Animation, How an Omniverse Creator Makes Films Virtually

Editor’s note: This post is one in a series that features individual creators and developers who use NVIDIA Omniverse to boost their artistic processes.

A headshot of Jae Solina
Jae Solina

Growing up in the Philippines, award-winning filmmaker Jae Solina says he turned to movies for a reminder that the world was much larger than himself and his homeland.

He started the popular YouTube channel JSFILMZ a decade ago as a way to share home videos he made for fun.

Since then, he’s expanded the channel to showcase his computer graphics-based movies, which have won the Best Animation and Best Super Short Film awards at the Las Vegas Independent Film Festival.

He also posts tutorials for virtual filmmaking with tools, including NVIDIA Omniverse — a physically accurate 3D design collaboration platform exclusively available with NVIDIA RTX GPUs and part of the NVIDIA Studio suite of creator tools.

Making tutorials is a way of paying it forward for Solina, as he is self-taught, gaining his computer graphics skills from other artists’ YouTube videos.

Solina now lives in Las Vegas with his wife and two kids, balancing filmmaking with part-time school and a full-time job.

“The only thing stopping you from creating something is your effort and imagination,” he said. “There are so many free tools like Blender or Omniverse that are readily available, enabling us to create what we want.”

Virtual Film Production

Solina creates computer graphics-based animation films, which can usually take large amounts of time and money, he said. NVIDIA Omniverse eases this process.

“With Omniverse, I don’t have to wait a full week to render a 30-second animation,” Solina said. “The rendering speed in Omniverse is superb and saves me a lot of time, which is important when balancing my filmmaking, non-creative work and family.”

Solina uses an NVIDIA GeForce RTX 3060 GPU, as well as Omniverse apps like Audio2Face, Create and Machinima to create his films virtually.

He also uses Omniverse Connectors for 3D applications like Blender and Autodesk Maya, as well as Reallusion’s iClone and Character Creator, with which he edits motion-capture data.

As a solo filmmaker, Solina said his main challenge is finding virtual assets — like characters and environments — that are photorealistic enough to use for movies.

“My process can definitely be a bit backwards, since the ideal method would be to write a script and then find the assets to make the story come alive,” he said. “But when I’m limited in my resources, I have to think of a storyline that fits a character or an environment I find.”

New support for the Omniverse ecosystem provided by 3D marketplaces and digital asset libraries helps solve this challenge — with thousands of Omniverse-ready assets for creators, all based on Universal Scene Description format.

Looking forward, Solina plans to create a short film entirely inside Omniverse.

Explore the NVIDIA Omniverse Instagram, gallery, forums and Medium channel. Check out Omniverse tutorials on Twitter and YouTube, and join our Discord server and Twitch channel to chat with the community.

The post From Imagination to Animation, How an Omniverse Creator Makes Films Virtually appeared first on The Official NVIDIA Blog.

Read More

How Retailers Meet Tough Challenges Using NVIDIA AI 

At the National Retail Federation’s annual trade show, conversations tend to touch on recurring themes: “Will we be able to stock must-have products for next Christmas?,” “What incentives can I offer to loyal workers?” and “What happens to my margins if Susie Consumer purchases three of the same dresses online and returns two?”

The $26 trillion global retail industry is undergoing accelerated change, brought on by the pandemic and rapidly changing consumer habits. Now, it’s looking for accelerated problem solving using NVIDIA AI to address increasingly acute labor, logistics and supply chain challenges that are accompanying those changes.

Working with an ecosystem of more than 100 startups, equipment providers and software partners, NVIDIA offers an AI Enterprise platform for retailers and quick-service restaurants that helps speed the creation of intelligent stores, AI-driven forecasting, interactive chatbots, voice-enabled order taking and hyperpersonalized recommendations, and logistics and store optimization using digital twin technologies for simulation.

A Labor Crisis

Labor shortages have become a critical issue. In September and October, accommodation and food services businesses lost 1.6 million, or 6.2 percent, of their workforce, while 1.4 million people quit their retail jobs, according to the U.S. Bureau of Labor Statistics.

One way to address the problem is by creating autonomous shopping experiences. AiFi, AWM and Trigo’s autonomous shopping platforms, each shown at NRF, provide a seamless store checkout process. Customers can walk into a store, grab the items they want and pay with their mobile phone on their way out. Beyond addressing labor shortages, these autonomous stores provide live inventory management and prevent shrink.

Store associates are the face of retail organizations, so it makes sense to reduce the time they spend on tasks that aren’t customer facing, such as performing inventory counts or scanning for out-of-stock items. Spacee is using computer vision and AI to help retailers handle these basic, repetitive tasks.

NVIDIA partners Everseen and Graymatics provide asset protection applications at the point of sale to reduce shrinkage and provide customers a faster self-checkout experience. Deep North’s store analytics application is used for queue management, to optimize labor scheduling and store merchandising, resulting in increased sales.

All these startups are using the NVIDIA AI platform to deliver real-time recommendations in stores and distribution centers.

NVIDIA Tokkio conversational AI avatars and the NVIDIA Riva conversational AI framework, as well as recommendation engines based on the NVIDIA Merlin application framework, also help improve the customer experience and solve labor shortages by allowing for automated order taking and upsell based on customer shopping history.

Vistry.AI is delivering drive-thru automated order taking with a speech and recommendation engine, as well as computer vision applications for queue prediction, to predict when orders are ready, ensure food freshness and accelerate curbside pickup.

A Broken Supply Chain

The supply chain is the lifeblood of the retail industry; it was on life support for many in 2021 as attempts to recover from pandemic-related shutdowns around the world were stymied by trucker and dock worker shortages, inclement weather and persistent shortfalls in key food and electronics components.

According to a November report from Adobe Digital Insights, online shoppers in October were met with more than 2 billion out-of-stock messages — double the rate reported in October 2020.

With consumers more likely than not to go — and possibly stay with — a competitor, retailers are investing heavily in predictive analytics to gain real-time insights for forecasting and ordering from point of embarkation through to individual store shelves and distribution centers.

Dematic, a global materials handling company, and startups Kinetic Vision and Osaro are other key companies that use the NVIDIA AI platform to develop edge AI applications that add intelligence to automated warehouse systems. From computer vision AI applications to autonomous forklifts to pick-and-place robots, these AI applications improve distribution center throughput and reduce equipment downtime. And with NVIDIA Fleet Command, these solutions can be remotely deployed and managed securely and at scale in hundreds of distribution centers.

Improving Logistics

To help the $9 trillion logistics industry efficiently route goods from distribution centers to stores and from stores to homes, NVIDIA in November announced its NVIDIA ReOpt AI software.

NVIDIA ReOpt is an accelerated solver for machine learning that optimizes vehicle route planning and logistics in real time. Working with the NVIDIA ReOpt team, Domino’s Pizza implemented a real-time predictive system that helps it meet important delivery standards for customers eager for dinner.

Retail Goes AI 

The NVIDIA AI Enterprise platform is helping retailers weather challenges expected to continue well beyond 2022. With consumers increasingly demanding what the industry calls an omnichannel experience, one that lets them order online and pick up at curbside or have items delivered speedily to their homes, balancing supply with demand has increased the need for fast, actionable insights.

As consumers move from seeking goods and services to experiences, the depth and breadth of interaction between customers and retailers is requiring AI to complement human interaction. It’s a shift that has moved from wishlist to deployment.

The post How Retailers Meet Tough Challenges Using NVIDIA AI  appeared first on The Official NVIDIA Blog.

Read More