Computer vision has become so good that the days of general managers screaming at umpires in baseball games in disputes over pitches may become a thing of the past.
That’s because developments in image classification along with parallel processing make it possible for computers to see a baseball whizzing by at 95 miles per hour. Pair that with image detection to help geolocate balls, and you’ve got a potent umpire tool that’s hard to argue with.
But computer vision doesn’t stop at baseball.
What Is Computer Vision?
Computer vision is a broad term for the work done with deep neural networks to develop human-like vision capabilities for applications, most often run on NVIDIA GPUs. It can include specific training of neural nets for segmentation, classification and detection using images and videos for data.
Major League Baseball is testing AI-assisted calls at the plate using computer vision. Judging balls and strikes on baseballs that can take just .4 seconds to reach the plate isn’t easy for human eyes. It could be better handled by a camera feed run on image nets and NVIDIA GPUs that can process split-second decisions at a rate of more than 60 frames per second.
Hawk-Eye, based in London, is making this a reality in sports. Hawk-Eye’s NVIDIA GPU-powered ball tracking and SMART software is deployed in more than 20 sports, including baseball, basketball, tennis, soccer, cricket, hockey and NASCAR.
Yet computer vision can do much more than just make sports calls.
What Is Computer Vision Beyond Sports?
Computer vision can handle many more tasks. Developed with convolutional neural networks, computer vision can perform segmentation, classification and detection for a myriad of applications.
Segmentation: Image segmentation is about classifying pixels to belong to a certain category, such as a car, road or pedestrian. It’s widely used in self-driving vehicle applications, including the NVIDIA DRIVE software stack, to show roads, cars and people. Think of it as a sort of visualization technique that makes what computers do easier to understand for humans.
Classification: Image classification is used to determine what’s in an image. Neural networks can be trained to identify dogs or cats, for example, or many other things with a high degree of precision given sufficient data.
Detection: Image detection allows computers to localize where objects exist. It puts rectangular bounding boxes — like in the lower half of the image below — that fully contain the object. A detector might be trained to see where cars or people are within an image, for instance, as in the numbered boxes below.
What You Need to Know: Segmentation, Classification and Detection
AI-powered vehicles aren’t a future vision, they’re a reality today. And they’re only truly possible on NVIDIA Xavier, our system-on-a-chip for autonomous vehicles.
The key to these cutting-edge vehicles is inference — the process of running AI models in real time to extract insights from enormous amounts of data. And when it comes to in-vehicle inference, NVIDIA Xavier has been proven the best — and the only — platform capable of real-world AI processing, yet again.
NVIDIA GPUs smashed performance records across AI inference in data center and edge computing systems in the latest round of MLPerf benchmarks, the only consortium-based and peer-reviewed inference performance tests. NVIDIA Xavier extended its performance leadership demonstrated in the first AI inference tests, held last year, while supporting all new use cases added for energy-efficient, edge compute SoC.
Inferencing for intelligent vehicles is a full-stack problem. It requires the ability to process sensors and run the neural networks, operating system and applications all at once. This high level of complexity calls for a huge investment, which NVIDIA continues to make.
The new NVIDIA A100 GPU, based on the NVIDIA Ampere architecture, also rose above the competition, outperforming CPUs by up to 237x in data center inference. This level of performance in the data center is critical for training and validating the neural networks that will run in the car at the massive scale necessary for widespread deployment.
Achieving this performance isn’t easy. In fact, most of the companies that have proven the ability to run a full self-driving stack run it on NVIDIA.
The MLPerf tests demonstrate that AI processing capability lies beyond the pure number of trillions of operations per second (TOPS) a platform can achieve. It’s the architecture, flexibility and accompanying tools that define a compute platform’s AI proficiency.
Xavier Stands Alone
The inference tests represent a suite of benchmarks to assess the type of complex workload needed for software-defined vehicles. Many different benchmark tests across multiple scenarios, including edge computing, verify whether a solution can perform exceptionally at not just one task, but many, as would be required in a modern car.
In this year’s tests, NVIDIA Xavier dominated results for energy-efficient, edge compute SoCs — processors necessary for edge computing in vehicles and robots — in both single-stream and multi-stream inference tasks.
Xavier is the current generation SoC powering the brain of the NVIDIA DRIVE AGX computer for both self-driving and cockpit applications. It’s an AI supercomputer, incorporating six different types of processors, including CPU, GPU, deep learning accelerator, programmable vision accelerator, image signal processor and stereo/optical flow accelerator.
NVIDIA DRIVE AGX Xavier
Thanks to its architecture, Xavier stands alone when it comes to AI inference. Its programmable deep neural network accelerators optimally support the operations for high-throughput and low-latency DNN processing. Because these algorithms are still in their infancy, we built the Xavier compute platform to be flexible so it could handle new iterations.
Supporting new and diverse neural networks requires processing different types of data, through a wide range of neural nets. Xavier’s tremendous processing performance handles this inference load to deliver a safe automated or autonomous vehicle with an intelligent user interface.
Proven Effective with Industry Adoption
As the industry compares TOPS of performance to determine autonomous capabilities, it’s important to test how these platforms can handle actual AI workloads.
Xavier’s back-to-back leadership in the industry’s leading inference benchmarks demonstrates NVIDIA’s architectural advantage for AI application development. Our SoC really is the only proven platform up to this unprecedented challenge.
The vast majority of automakers, tier 1 suppliers and startups are developing on the DRIVE platform. NVIDIA has gained much experience running real-world AI applications on its partners’ platforms. All these learnings and improvements will further benefit the NVIDIA DRIVE ecosystem.
Raising the Bar Further
It doesn’t stop there. NVIDIA Orin, our next-generation SoC, is coming next year, delivering nearly 7x the performance of Xavier with incredible energy-efficiency.
NVIDIA Orin
Xavier is compatible with software tools such as CUDA and TensorRT to support the optimization of DNNs to target hardware. These same tools will be available on Orin, which means developers can seamlessly transfer past software development onto the latest hardware.
NVIDIA has shown time and again that it’s the only solution for real-world AI and will continue to drive transformational technology such as self-driving cars for a safer, more advanced future.
Inference, the work of using AI in applications, is moving into mainstream uses, and it’s running faster than ever.
NVIDIA GPUs won all tests of AI inference in data center and edge computing systems in the latest round of the industry’s only consortium-based and peer-reviewed benchmarks.
NVIDIA A100 and T4 GPUs swept all data center inference tests.
The A100, introduced in May, outperformed CPUs by up to 237x in data center inference, according to the MLPerf Inference 0.7 benchmarks. NVIDIA T4 small form factor, energy-efficient GPUs beat CPUs by up to 28x in the same tests.
To put this into perspective, a single NVIDIA DGX A100 system with eight A100 GPUs now provides the same performance as nearly 1,000 dual-socket CPU servers on some AI applications.
Leadership performance enables cost efficiency in taking AI from research to production.
This round of benchmarks also saw increased participation, with 23 organizations submitting — up from 12 in the last round — and with NVIDIA partners using the NVIDIA AI platform to power more than 85 percent of the total submissions.
A100 GPUs, Jetson AGX Xavier Take Performance to the Edge
While A100 is taking AI inference performance to new heights, the benchmarks show that T4 remains a solid inference platform for mainstream enterprise, edge servers and cost-effective cloud instances. In addition, the NVIDIA Jetson AGX Xavier builds on its leadership position in power constrained SoC-based edge devices by supporting all new use cases.
Jetson AGX Xavier joined the A100 and T4 GPUs in leadership performance at the edge.
The results also point to our vibrant, growing AI ecosystem, which submitted 1,029 results using NVIDIA solutions representing 85 percent of the total submissions in the data center and edge categories. The submissions demonstrated solid performance across systems from partners including Altos, Atos, Cisco, Dell EMC, Dividiti, Fujitsu, Gigabyte, Inspur, Lenovo, Nettrix and QCT.
Expanding Use Cases Bring AI to Daily Life
Backed by broad support from industry and academia, MLPerf benchmarks continue to evolve to represent industry use cases. Organizations that support MLPerf include Arm, Baidu, Facebook, Google, Harvard, Intel, Lenovo, Microsoft, Stanford, the University of Toronto and NVIDIA.
The latest benchmarks introduced four new tests, underscoring the expanding landscape for AI. The suite now scores performance in natural language processing, medical imaging, recommendation systems and speech recognition as well as AI use cases in computer vision.
You need go no further than a search engine to see the impact of natural language processing on daily life.
“The recent AI breakthroughs in natural language understanding are making a growing number of AI services like Bing more natural to interact with, delivering accurate and useful results, answers and recommendations in less than a second,” said Rangan Majumder, vice president of search and artificial intelligence at Microsoft.
“Industry-standard MLPerf benchmarks provide relevant performance data on widely used AI networks and help make informed AI platform buying decisions,” he said.
AI Helps Saves Lives in the Pandemic
The impact of AI in medical imaging is even more dramatic. For example, startup Caption Health uses AI to ease the job of taking echocardiograms, a capability that helped save lives in U.S. hospitals in the early days of the COVID-19 pandemic.
That’s why thought leaders in healthcare AI view models like 3D U-Net, used in the latest MLPerf benchmarks, as key enablers.
“We’ve worked closely with NVIDIA to bring innovations like 3D U-Net to the healthcare market,” said Klaus Maier-Hein, head of medical image computing at DKFZ, the German Cancer Research Center.
“Computer vision and imaging are at the core of AI research, driving scientific discovery and representing core components of medical care. And industry-standard MLPerf benchmarks provide relevant performance data that helps IT organizations and developers accelerate their specific projects and applications,” he added.
Commercially, AI use cases like recommendation systems, also part of the latest MLPerf tests, are already making a big impact. Alibaba used recommendation systems last November to transact $38 billion in online sales on Singles Day, its biggest shopping day of the year.
Adoption of NVIDIA AI Inference Passes Tipping Point
AI inference passed a major milestone this year.
NVIDIA GPUs delivered a total of more than 100 exaflops of AI inference performance in the public cloud over the last 12 months, overtaking inference on cloud CPUs for the first time. Total cloud AI Inference compute capacity on NVIDIA GPUs has been growing roughly tenfold every two years.
GPUs in major cloud services now account for more inference performance than CPUs.
With the high performance, usability and availability of NVIDIA GPU computing, a growing set of companies across industries such as automotive, cloud, robotics, healthcare, retail, financial services and manufacturing now rely on NVIDIA GPUs for AI inference. They include American Express, BMW, Capital One, Dominos, Ford, GE Healthcare, Kroger, Microsoft, Samsung and Toyota.
Companies across key industry sectors use NVIDIA’s AI platform for inference.
Why AI Inference Is Hard
Use cases for AI are clearly expanding, but AI inference is hard for many reasons.
New kinds of neural networks like generative adversarial networks are constantly being spawned for new use cases and the models are growing exponentially. The best language models for AI now encompass billions of parameters, and research in the field is still young.
These models need to run in the cloud, in enterprise data centers and at the edge of the network. That means the systems that run them must be highly programmable, executing with excellence across many dimensions.
NVIDIA founder and CEO Jensen Huang compressed the complexities in one word: PLASTER. Modern AI inference requires excellence in Programmability, Latency, Accuracy, Size of model, Throughput, Energy efficiency and Rate of learning.
To power excellence across every dimension, we’re focussed on constantly evolving our end-to-end AI platform to handle demanding inference jobs.
AI Requires Performance, Usability
An accelerator like the A100, with its third-generation Tensor Cores and the flexibility of its multi-instance GPU architecture, is just the beginning. Delivering leadership results requires a full software stack.
NVIDIA’s AI software begins with a variety of pretrained models ready to run AI inference. Our Transfer Learning Toolkit lets users optimize these models for their particular use cases and datasets.
NVIDIA TensorRT optimizes trained models for inference. With 2,000 optimizations, it’s been downloaded 1.3 million times by 16,000 organizations.
The NVIDIA Triton Inference Server provides a tuned environment to run these AI models supporting multiple GPUs and frameworks. Applications just send the query and the constraints — like the response time they need or throughput to scale to thousands of users — and Triton takes care of the rest.
These elements run on top of CUDA-X AI, a mature set of software libraries based on our popular accelerated computing platform.
Getting a Jump-Start with Applications Frameworks
Finally, our application frameworks jump-start adoption of enterprise AI across different industries and use cases.
3D artists and video editors have long used real-time AI features to improve their work and speed up how they turn inspiration into finished art. Now, those benefits are extending to Adobe Photoshop users with the introduction of GPU-accelerated neural filters.
These AI-powered tools, leveraging NVIDIA RTX GPUs with the Adobe creative applications, are being showcased at Adobe MAX, which is bringing together creators from around the world virtually through Oct. 22.
Neural filters are a new feature set for artists to try AI-powered tools that enable them to explore creative ideas and make amazing, complex adjustments to images in just seconds. Done manually, these adjustments would take artists hours of tedious work. AI allows artists to make these changes almost instantaneously.
NVIDIA GPUs accelerate nearly all these new filters. We’ll explain how to get the most out of them at a session at Adobe MAX.
Adobe and NVIDIA are closely collaborating on AI technology to improve creative tools in Creative Cloud and Photoshop. This collaboration includes the new Smart Portrait Filter, which is powered by NVIDIA StyleGAN2 technology and runs best on NVIDIA RTX GPUs.
With Smart Portrait in Photoshop, artists can easily experiment, making edits to facial characteristics, such as gaze direction and lighting angles, simply by dragging a slider. These types of complex corrections and adjustments would typically entail multiple manual steps. But Smart Portrait uses AI — based on a deep neural network developed by NVIDIA Research and trained on numerous portrait images — to achieve breathtaking results in seconds.
This gives artists greater flexibility with their images long after the photo shoot has ended. And they retain full control over their work with a non-destructive workflow, while the effects blend naturally into the original image.
Video editors in Adobe Premiere Pro also benefit from NVIDIA RTX GPUs with virtually all GPU-accelerated decoding offloaded to dedicated VRAM, resulting in smoother video playback and sharper responsiveness when scrubbing through footage, especially with ultra-high resolution and multistream footage. Advanced, AI-powered features such as Scene Edit Detection and Auto Reframe automate manual tasks, speeding up final exports and saving editors valuable time.
For the first time, Adobe Premiere Elements adds GPU acceleration to enable instant playback of popular video effects such as adding a lens flare or an animated overlay, cropping of videos, and overall playback in real-time, all without prerendering, rapidly speeding up the editing process.
AI and GPU-accelerated workflows are the result of the ongoing collaboration between teams at NVIDIA and Adobe. Over the years, we’ve developed tools and helped accelerate workflows in Adobe Photoshop, Lightroom, Premiere Pro, After Effects, Illustrator, Dimension, Substance Alchemist, Substance Painter and Substance Designer. As Adobe continues to build amazing software experiences, NVIDIA will be there to power and accelerate them, giving creators more time for creativity.
Working Smarter: Tapping into AI to Boost Creativity
Adobe is hosting more than 350 sessions across 10 tracks at this year’s MAX conference. Creators looking for new ways to improve their work while cutting down on the tasks that take away precious time can learn how to get the most out of new AI tools across Adobe creative apps.
NVIDIA is hosting an Adobe MAX session where attendees will discover new ways to tap into the power of AI. Whether a graphic artist, video editor, motion graphics professional, Photoshop professional, concept artist or other creator who needs computing speed, you’ll leave with valuable, time-saving tips.
Session attendees will discover:
How to improve creations with more precision, clarity and quality
How to let AI do the work under the hood, giving you more time to create
The NVIDIA Studio ecosystem of tools and products designed to supercharge creativity
Visit the session catalog to learn more and tune in on Wednesday, Oct. 21, from 11-11:30 a.m. Pacific time.
October Studio Driver Ready For Download
Alongside these updates to Adobe Photoshop, Adobe Premiere Pro and Adobe Premiere Elements, there are new releases of Adobe After Effects, Adobe Substance Alchemist, Notch and Daz 3D — all supported in the new October NVIDIA Studio Driver. Studio Drivers are built specifically for creators and tested extensively against top creative apps and workflows.
Learn more about NVIDIA Studio hardware and software for creators on the NVIDIA Studio website.
You can also stay up to date on the latest apps through NVIDIA’s Studio YouTube channel, featuring tutorials, tips and tricks by industry-leading artists.
Amid a pandemic that’s put much of the world’s work, learning, even family reunions online, two of the leaders who have made today’s virtual world possible met Thursday on, where else — Zoom — to talk about what’s next.
NVIDIA CEO Jensen Huang and Zoom CEO Eric Yuan spoke Thursday at the online video conference company’s Zoomtopia user event in a casual, wide-ranging conversation.
“If not for what Zoom has done, the recent pandemic would be unbearable,” Huang said. The present situation, Huang explained, “has accelerated the future, it has brought forward the urgency of a digital future.”
In front of a virtual audience from all over the globe, the two spoke about their entrepreneurial journeys, NVIDIA’s unique company culture, and how NVIDIA is knitting together the virtual and real worlds to help NVIDIA employees collaborate.
Huang’s appearance at Zoomtopia follows NVIDIA’s GPU Technology Conference last week, where Huang outlined NVIDIA’s view of data center computing and introduced new technologies in data centers, edge AI and healthcare.
Yuan playfully wore a leather jacket, matching Huang’s trademark attire—and briefly displayed a sleek virtual kitchen as his backdrop, paying tribute to the presentations Huang has given from his kitchen this year—began their conversation with Huang by asking about his early life.
“I was fortunate that my parents worked hard and all of the people I was surrounded by worked hard,” Huang said, adding that he was focused on on school and sports, especially table tennis. “To me working is living, working is breathing and, to me, it’s not work at all — I enjoy it too much.”
It’s NVIDIA’s mission, Huang said, that continues to motivate him, as the company has gone from inventing the GPU to pioneering new possibilities in robotics and AI.
The common thread: since the beginning, NVIDIA has had a singular focus on accelerated computing.
“We built a time machine,” Huang said, touching on NVIDIA’s work in drug discovery as an example. “So, instead of a particular drug taking 10 years to discover, we would like drugs and therapies and vaccines to be discovered in months.”
Zoom and NVIDIA, Huang said, share a “singular purpose and a sense of destiny,” Huang said, one that has made the world a better place.
“The fact that Zoom existed and your vision came to reality means we can be together even if we’re not together,” Huang said.
“You can look at your work and imagine the impact on society and the benefits it will bring and somehow it’s your job to do it,” Huang said. “If you don’t do it, no one else will — and that’s thrilling to me, I love that feeling.”
Yuan also asked about NVIDIA’s culture and the future of work, one which Huang believes will increasingly meld the physical and the virtual worlds.
Today, for example, we might report to your colleagues that we’ll be WFH, or working from home.
Office lingo, however, may change to reflect the new reality, where being at the office isn’t necessarily the norm.
“In the future we will say we’re ‘going to the office,’” Huang said. “Today we say ‘WFH,’ in the future we will say ‘GTO.’”
Tools such as Zoom enable colleagues to meet, face to face, from home, from an office, from anywhere in the world.
More and more, work will take place in a hybrid of office and home, physical and virtual reality.
NVIDIA, for example, has created a platform called NVIDIA Omniverse that lets colleagues working in different places and with different tools collaborate in real time.
“The Adobe world can connect to the Catia world and so on,” Huang said. “We can have different designers working with each other at their homes.”
The present moment has “brought forward the urgency of a digital future, it has made us aware that completely physical is not sufficient, that completely digital is not sufficient,” Huang said. “The future is a mixed reality world.”
Four new supercomputers backed by a pan-European initiative will use NVIDIA’s data center accelerators, networks and software to advance AI and high performance computing.
They include one system dubbed Leonardo, unveiled today at Italy’s CINECA research center, using NVIDIA technologies to deliver the world’s most powerful AI system. The four mark the first of eight systems to be announced this year targeting spots among the world’s 50 most powerful computers.
Together, they’ll form a regional network, “an engine to power Europe’s data economy,” said EuroHPC, the group driving the effort, in a white paper outlining its goals.
The systems will apply AI and data analytics across scientific and commercial applications that range from fighting COVID-19 and climate change to the design of advanced airplanes, cars, drugs and materials.
Joining Leonardo are a wave of new AI supercomputers planned for the Czech Republic, Luxembourg and Slovenia that will act as national centers of competence, expanding skills and creating jobs.
NVIDIA GPUs, InfiniBand Power Latest Systems
All four supercomputers announced use NVIDIA Ampere architecture GPUs and NVIDIA Mellanox HDR InfiniBand networks to tap an ecosystem of hundreds of HPC and AI applications. Atos, an NVIDIA systems partner headquartered in France, will build three of the four systems; Hewlett Packard Enterprise will construct the fourth.
The new systems join 333 of the world’s TOP500 supercomputers powered by NVIDIA GPUs, networking or both.
NVIDIA GPUs accelerate 1,800 HPC applications, nearly 800 of them available today in the GPU application catalog and NGC, NVIDIA’s hub for GPU-optimized software.
The new systems all use HDR 200Gb/s InfiniBand for low latency, high throughput and in-network computing. It’s the latest version of InfiniBand, already powering supercomputers across Europe.
A Brief Tour of Europe’s Latest Supercomputers
Leonardo will be the world’s fastest AI supercomputer. Atos is harnessing nearly 14,000 NVIDIA Ampere architecture GPUs and HDR 200Gb/s InfiniBand networking to deliver a system with 10 exaflops of AI performance. It will use the InfiniBand Dragonfly+ network topology to deliver both flexibility and scalable performance.
Researchers at CINECA will apply that power to advance science, simulating planetary forces behind climate change and molecular movements inside a coronavirus. The center is perhaps best known for its work on Quantum Espresso, a suite of open source codes for modeling nanoscale materials for jobs such as engineering better batteries.
A new supercomputer in Luxembourg called MeluXina, also part of the EuroHPC network, will connect 800 NVIDIA A100 GPUs on HDR 200Gb/s InfiniBand links. The system, to be built by Atos and powered by green energy from wood waste, will pack nearly 500 petaflops of AI performance.
MeluXina will address commercial applications and scientific research. It plans to offer access to users leveraging HPC and AI to advance work in financial services as well as manufacturing and healthcare.
Eastern Europe Powers Up
The new Vega supercomputer at the Institute of Information Science in Maribor, Slovenia, (IZUM) will be based on the Atos BullSequana XH2000 system. The supercomputer, named after Slovenian mathematician Jurij Vega, includes 240 A100 GPUs and 1,800 HDR 200Gb/s InfiniBand end points.
Vega will help “ensure a new generation of experts and developers, as well as the wider Slovenian community, can meet new challenges within our national consortium and contribute to regional and European HPC initiatives,” said Aleš Bošnjak, IZUM’s director in a statement issued by EuroHPC.
A total of 32 countries are participating in the EuroHPC effort.
The supercomputer will be based on the HPE Apollo 6500 systems from Hewlett Packard Enterprise (HPE). It will serve researchers at the VSB – Technical University of Ostrava, where it’s based, as well as an expanding set of external academic and industrial users employing a mix of simulations, data analytics and AI.
The story of Europe’s ambitions in HPC and AI is still being written.
EuroHPC has yet to announce its plans for systems in Bulgaria, Finland, Portugal and Spain. And beyond that work, the group has already sketched out plans that stretch to 2027.
Using AI and a supercomputer simulation, Ken Dill’s team drew the equivalent of wanted posters for a gang of proteins that make up COVID-19. With a little luck, one of their portraits could identify a way to arrest the coronavirus with a drug.
When the pandemic hit, “it was terrible for the world, and a big research challenge for us,” said Dill, who leads the Laufer Center for Physical & Quantitative Biology at Stony Brook University, in Long Island, New York.
For a decade, he helped the center assemble the researchers and tools needed to study the inner workings of proteins — complex molecules that are fundamental to cellular life. The center has a history of applying its knowledge to viral proteins, helping others identify drugs to disable them.
“So, when the pandemic came, our folks wanted to spring into action,” he said.
AI, Simulations Meet at the Summit
The team aimed to use a combination of physics and AI tools to predict the 3D structure of more than a dozen coronavirus proteins based on lists of the amino acid strings that define them. It won a grant for time on the IBM-built Summit supercomputer at Oak Ridge National Laboratory to crunch its complex calculations.
“We ran 30 very extensive simulations in parallel, one on each of 30 GPUs, and we ran them continuously for at least four days,” explained Emiliano Brini, a junior fellow at the Laufer Center. “Summit is a great machine because it has so many GPUs, so we can run many simulations in parallel,” he said.
“Our physics-based modeling eats a lot of compute cycles. We use GPUs almost exclusively for their speed,” said Dill.
Sharing Results to Help Accelerate Research
Thanks to the acceleration, the predictions are already in. The Laufer team quickly shared them with about a hundred researchers working on a dozen separate projects that conduct painstakingly slow experiments to determine the actual structure of the proteins.
“They indicated some experiments could be done faster if they had hunches from our work of what those 3D structures might be,” said Dill.
Now it’s a waiting game. If one of the predictions gives researchers a leg up in finding a weakness that drug makers can exploit, it would be a huge win. It could take science one step closer to putting a general antiviral drug on the shelf of your local pharmacy.
Melding Machine Learning and Physics
Dill’s team uses a molecular dynamics program called MELD. It blends physical simulations with insights from machine learning based on statistical models.
AI provides MELD key information to predict a protein’s 3D structure from its sequence of amino acids. It quickly finds patterns across a database of atomic-level information on 200,000 proteins gathered over the last 50 years.
MELD uses this information in compute-intensive physics simulations to determine the protein’s detailed structure. Further simulations then can predict, for example, what drug molecules will bind tightly to a specific viral protein.
“So, both these worlds — AI inference and physics simulations — are playing big roles in helping drug discovery,” said Dill. “We get the benefits of both methods, and that combination is where I think the future is.”
MELD runs on CUDA, NVIDIA’s accelerated computing platform for GPUs. “It would take prohibitively long to run its simulations on CPUs, so the majority of biological simulations are done on GPUs,” said Brini.
Playing a Waiting Game
The COVID-19 challenge gave Laufer researchers with a passion for chemistry a driving focus. Now they await feedback on their work on Summit.
“Once we get the results, we’ll publish what we learn from the mistakes. Many times, researchers have to go back to the drawing board,” he said.
And every once in a while, they celebrate, too.
Dill hosted a small, socially distanced gathering for a half-dozen colleagues in his backyard after the Summit work was complete. If those results turn up a win, there will be a much bigger celebration extending far beyond the Stony Brook campus.
In the battle against COVID-19, Greater Paris University Hospitals – Public Assistance Hospital of Paris (AP-HP is the French acronym) isn’t just on the medical front lines — it’s on the data front lines as well.
With a network of 39 hospitals treating 8.3 million patients each year, AP-HP is a major actor in the fight against COVID-19.
Along with its COVID-19 cases comes an awful lot of data, including now geodata that can potentially help lessen the impact of the pandemic. AP-HP, which partners with seven universities, already had the ability to analyze large amounts of medical data. It had previously created dashboards that combined cancer cases and geodata. So, it was logical to pursue and extend its role during the pandemic.
The expected volume of COVID-19 data and geodata would probably have tested AP-HP’s data crunching capacity. To mitigate this critical challenge, the hospital’s information systems administrators turned to Kinetica, a provider of streaming data warehouses and real-time analytics and a member of the NVIDIA Inception program for AI startups.
Kinetica’s offering harnesses the power of NVIDIA GPUs to quickly convert case location data into usable intelligence. And in the fight against COVID-19, speed is everything.
The project team also used NVIDIA RAPIDS to speed up the machine learning algorithms integrated into the platform. RAPIDS accelerates analytics and data science pipelines on NVIDIA GPUs by taking advantage of GPU parallelism and high memory bandwidth.
“Having the ability to perform this type of analysis in real time is really important during a pandemic,” said Hector Countouris, the project lead at AP-HP. “And more data is coming.”
Analyzing COVID Contact Data
What Countouris and his colleagues are most focused on is using COVID-related geodata to understand where virus “hot spots” are and the dynamic of the outbreak. Looking for cluster locations can help decision-making at the district or region level.
In addition, they’re looking at new signals to improve early detection of COVID patients. This includes working with data from other regional agencies.
If patients are diagnosed with COVID, they’ll be asked by the relevant agencies via a phone call about their recent whereabouts and contacts to help with contact tracing. This is the first time that a wide range of data from different partners in the Paris area will be integrated to allow for contact tracing and timely alerts about a potential exposure. The result will be a newfound ability to see how clusters of COVID-19 cases evolve.
“We hope that in the near future we will be able to follow how a cluster evolves in real time,” said Countouris.
The goal is to enable public health decision-makers to implement prevention and control measures and assess their effectiveness. The data can also be integrated with other demographic data to study the viral spread and its possible dependency on socio-economics and other factors.
Attacking Bottlenecks with GPUs
Prior to engaging with Kinetica, such data-intensive projects involved so much time for loading the data that they couldn’t be analyzed quickly enough to deliver real-time benefits.
“Now, I don’t feel like I have a bottleneck,” said Countouris. “We are continuously integrating data and delivering dashboards to decision makers within hours. And with robust real-time pipelines allowing for continuous data ingestion, we can now focus on building better dashboards.”
In the past, to get data in a specific and usable format, they would need to do a lot of pre-processing. With Kinetica’s Streaming Data Warehouse powered by NVIDIA V100 Tensor Core GPUs, that’s no longer the case. Users can access the much richer datasets they demand.
Kinetica’s platform is available on NVIDIA NGC, a catalog of GPU-optimized AI containers that let enterprises quickly operationalize extreme analytics, machine learning and data visualization. This eliminates complexity and lets organizations deploy cloud, on-premises or hybrid models for optimal business operations.
“I don’t think we could meet user expectations for geodata without GPU power,” he said. “There is just too much data and geodata to provide for too many users at the same time.”
AP-HP’s COVID-related work has already built a foundation upon which to do follow-up work related to emergency responses in general. The hospital information system’s interest for that kind of data is far from over.
“The fact that we helped the decision-making process and that officials are using our data is the measure of success,” said Countouris. “We have a lot to do. This is only the beginning.”
Not everyone needs to be a developer, but everyone will need to be an AI decision maker.
That was the message behind a panel discussion on Advancing Equitable AI, which took place at our GPU Technology Conference last week. It was one of several GTC events advancing the conversation on diversity, equity and ethics in AI.
This year, we strengthened our support for women and underrepresented developers and scientists at GTC by providing conference passes to members of professional organizations supporting women, Black and Latino developers. Professors at historically Black colleges and universities — including Prairie View A&M University, Hampton University and Jackson State University — as well as groups like Black in AI and LatinX in AI received complimentary access to training from the NVIDIA Deep Learning Institute.
A Forbes report last year named GTC as one of the U.S.’s top conferences for women to attend to further their careers in AI. At this month’s event, women made up better than one in five registered attendees — doubling last year’s count and an almost 4x increase since 2017 — and more than 100 of the speakers.
And in a collaboration with the National Society of Black Engineers that will extend beyond GTC, we created opportunities for the society’s collegiate and professional developers to engage with NVIDIA’s recruiting team, which provided guidance on navigating the new world of virtual interviewing and networking.
“We’re excited to be embarking on a partnership with NVIDIA,” said Johnnie Tangle, national finance chairman of NSBE Professionals. “Together, we are both on the mission of increasing the visibility of Blacks in development and showing why diversity in the space enhances the community as a whole.”
Panel Discussions: Paving Pathways for Equitable AI
Two power-packed, all-female panels at GTC focused on a roadmap for responsible and equitable AI.
In a live session that drew over 250 attendees, speakers from the University of Florida, the Boys and Girls Club of Western Pennsylvania and AI4All — a nonprofit working to increase diversity and inclusion in AI — discussed the importance of AI exposure and education for children and young adults from underrepresented groups.
When a broader group of young people has access to AI education, “we naturally see a way more diverse and interesting set of problems being addressed,” said Tess Posner, CEO of AI4All, “because young people and emerging leaders in the field are going to connect the technology to a problem they’ve seen in their own lives, in their own experience or in their communities.”
The conversation also covered the role parents and schools play in fostering awareness and exposure to STEM subjects in their children’s schools, as well as the need for everyone — developers or not — to have a foundational understanding of how AI works.
“We want students to be conscious consumers, and hopefully producers,” said Christina Gardner-McCune, associate professor and director of the Engaging Learning Lab at the University of Florida, and co-chair of the AI4K12 initiative. “Everybody is going to be making decisions about what AI technologies are used in their homes, what AI technologies their children interact with.”
The webinar featured representatives from the U.S. National Institute of Standards and Technology, Scotland-based innovation center The Data Lab, and C Minds, a think tank focused on AI initiatives in Latin America. Speakers shared their priorities for developing trustworthy AI, and defined what success would like to them five years in the future.
Dinner with Strangers: Developer Diversity in AI
In a virtual edition of the popular Dinner with Strangers networking events at GTC, experts from NVIDIA and NSBE partnered to moderate two conversations with GTC attendees. NVIDIA employees shared their experiences and tips with early-career attendees, offering advice on how to build a personal brand in a virtual world, craft a resume and prepare for interviews.
For more about GTC, watch NVIDIA founder and CEO Jensen Huang’s keynote below.
When large organizations require translation services, there’s no room for the amusing errors often produced by automated apps. That’s where Lilt, an AI-powered enterprise language translation company, comes in.
Lilt CEO Spence Green spoke with AI Podcast host Noah Kravitz about how the company is using a human-in-the-loop process to achieve fast, accurate and affordable translation.
Lilt does so with a predictive typing software, in which professional translators receive AI-based suggestions of how to translate content. By relying on machine assistance, Lilt’s translations are efficient while retaining accuracy.
However, including people in the company’s workflow also makes localization possible. Professional translators use cultural context to take direct translations and adjust phrases or words to reflect the local language and customs.
Lilt currently supports translations of 45 languages, and aims to continue improving its AI and make translation services more affordable.
Key Points From This Episode:
Green’s experience living in Abu Dhabi was part of the inspiration behind Lilt. While there, he met a man, an accountant, who had immigrated from Egypt. When asked why he no longer worked in accounting, the man explained that he didn’t speak English, and accountants who only spoke Arabic were paid less. Green didn’t want the difficulty of adult language learning to be a source of inequality in a business environment.
Lilt was founded in 2015, and evolved from a solely software company into a software and services business. Green explains the steps it took for the company to manage translators and act as a complete solution for enterprises.
Tweetables:
“We’re trying to provide technology that’s going to drive down the cost and increase the quality of this service, so that every organization can make all of its information available to anyone.” — Spence Green [2:53]
“One could argue that [machine translation systems] are getting better at a faster rate than at any point in the 70-year history of working on these systems.” — Spence Green [14:01]
Hugging Face is more than just an adorable emoji — it’s a company that’s demystifying AI by transforming the latest developments in deep learning into usable code for businesses and researchers, explains research engineer Sam Shleifer.
Capital One Senior Software Engineer Kyle Nicholson explains how modern machine learning techniques have become a key tool for financial and credit analysis.
Voice recognition is one thing, creating natural sounding artificial voices is quite another. Lyrebird co-founder Jose Solero speaks about how the startup is using deep learning to create a system that’s able to listen to human voices and generate speech mimicking the original human speaker.