Duplex is getting smarter and making life a little easier

Duplex is getting smarter and making life a little easier

In 2018, we introduced Duplex, our AI technology that uses natural conversation to get things done. Since then, we’ve been exploring how conversational technology can be both easy to interact with and  help people save more time. 

Today, during our Search On event, we shared an update on how Duplex and Google Assistant are helping people in their everyday lives. From providing more accurate business information in products like Google Maps, to booking appointments and reservations on your behalf, to waiting on hold for you, we’re continuing to bring Duplex to new places to make life a little easier.

Keeping local businesses information fresh 

This pandemic has shown us how critical up-to-date local information is, both for people trying to find services nearby and for businesses looking for ways to serve their customers. Whether you’re looking to grab dinner from your favorite restaurant or stop by your neighborhood florist, chances are you’ll check their hours of operation online first, and maybe find out if they offer things like dine-in or curbside pickup. 

To help people find accurate information about local businesses online, Duplex conversational technology is now calling businesses to automatically update business listings on Search and Maps with modified store hours and details like takeout or no-contact delivery. We began using Duplex to automatically update business information and add it to Search and Maps at scale in the U.S. last year. That means business owners don’t have to worry about manually updating these details, and potential customers get access to the most accurate information. When the pandemic started, we expanded business updates to eight countries, and have since made over 3 million updates to businesses like pharmacies, restaurants and grocery stores that have been seen over 20 billion times in Maps and Search. 

A personal assistant to save you time

From restaurant reservations to salon appointments, Duplex powers Google Assistant to help people save time, having completed more than a million bookings since its launch. So whenever you’re ready to dine out again, you can try asking your Assistant to book you a table at your favorite restaurant and let Duplex get it done. 

With Duplex on the web, Google Assistant can also complete tasks on the mobile web that would otherwise take up to 20 steps to complete, like booking a rental car or buying movie tickets. And we’re currently piloting the same experience with things like shopping and food ordering for a faster checkout.

Another way conversational AI helps people save time is with Call Screen, which lets Google Assistant answer unknown calls on Android phones to avoid spam calls. Every month, Call Screen helps save more than 2 million minutes on the phone. And now with Hold for Me, Duplex is powering Google Assistant to wait on hold for you and let you know when someone is on the line. 

More natural conversations

We still have a way to go towards having truly natural-feeling conversations with machines, so it’s exciting to see the great progress across the industry in neural speech recognition and synthesis, and in our own new language understanding models. For Duplex, these and many other advancements translate into significant improvements in quality. In fact, 99 percent of calls made by Duplex today are entirely automated. 

Our ability to interact with technology as naturally as we interact with each other remains a long-standing promise. As Duplex continues to take steps in this direction, we remain committed to developing our conversational technology in a responsible way, upholding the standards outlined in our AI principles and with transparency. For example, we always disclose that you’re speaking with an automated system when making a call. We’re excited by how far we’ve come, and more importantly, by how many people and businesses this technology can help in ways big and small.

Read More

Optimizing applications with EagleDream in Amazon CodeGuru Profiler

Optimizing applications with EagleDream in Amazon CodeGuru Profiler

This is a guest post by Dustin Potter at EagleDream Technologies. In their own words, “EagleDream Technologies educates, enables, and empowers the world’s greatest companies to use cloud-native technology to transform their business. With extensive experience architecting workloads on the cloud, as well as a full suite of skills in application modernization, data engineering, data lake design, and analytics, EagleDream has built a growing practice in helping businesses redefine what’s possible with technology.”

EagleDream Technologies is a trusted cloud-native transformation company and APN Premier Consulting Partner for businesses using AWS. EagleDream is unique in using its cloud-native software engineering and application modernization expertise to guide you through your journey to the cloud, optimize your operations, and transform how you do business using AWS. Our team of highly trained professionals helps accelerate projects at every stage of the cloud journey. This post shares our experience using Amazon CodeGuru Profiler to help one of our customers optimize their application under tight deadlines.

Project overview

Our team received a unique opportunity to work with one of the industry’s most disruptive airline technology leaders, who uses their expertise to build custom integrated airline booking, loyalty management, and ecommerce platforms. This customer reached out to our team to help optimize their new application. They already had a few clients using the system, but they recently signed a deal with a major airline that would represent a load increase to their platform five times in size. It was critical that they prepare for this significant increase in activity. The customer was running a traditional three-tier application written in Java that used Amazon Aurora for the data layer. They had already implemented autoscaling for the web servers and database but realized something was wrong when they started running load tests. During the first load test, the web tier expanded to over 80 servers and Aurora reached the max number of read replicas.

Our team knew we had to dive deep and investigate the application code. We had previously used other application profiling tools and realized how invaluable they can be when diagnosing these types of issues. Also, AWS recently announced Amazon CodeGuru and we were eager to try it out. On top of that, the price and ease of setup was a driving factor for us. We had looked at an existing commercial application performance monitoring tool, but it required more invasive changes to utilize. To automate the install of these tools, we would have needed to make changes to the customer’s deployment and infrastructure setup. We had to move quickly with as little disruption to their ongoing feature development as possible, which contributed to our final decision to use CodeGuru.

CodeGuru workflow

After we decided on CodeGuru, it was easy to get CodeGuru Profiler installed and start capturing metrics. There are two ways to profile an application. The first is to reference the profiler agent during the start of the application by using the standard -javaagent parameter. This is useful if the group performing the profiling isn’t the development team, for example in an organization with more traditional development and operation silos. This is easy to set up because all that’s needed is to download the .jar published in the documentation and alter any startup scripts to include the agent and the name of the profiling group to use.

The second way to profile the application is to include the profiler code via a dependency in your build system and instantiate a profiling thread somewhere at the entry point of the program. This option is great if the development team is handling the profiling. For this particular use case, we fell into the second group, so including it in the code was the quickest and easiest approach. We added the library as a Maven dependency and added a single line of application code. After the code was committed, we used the customer’s existing Jenkins setup to deploy the latest build to an integration environment. The final step of the pipeline was to run load tests against the new build. After the tests completed, we had a flame graph that we used to start identifying any issues.

The workflow includes the following steps:

  1. Developers check in code.
  2. The check-in triggers a Jenkins job.
  3. Maven compiles the code.
  4. Jenkins deploys the artifact to the development environment.
  5. Load tests run against the newly deployed code.
  6. CodeGuru Profiler monitors the environment and generates a flame graph and a recommendation report.

The following diagram illustrates the workflow.

Flame graphs group together stack traces and highlight which part of the code consumes the most resources. The following screenshot is a sample flame graph from an AWS demo application for reference.

After CodeGuru generated the flame graphs and recommendations report, we took an iterative approach and tackled the biggest offenders first. The flame graphs provided perceptive guidance for actionable recommendations that it discovers and made it easy to identify which execution paths were taking the longest to complete. By looking at the longest frames first, we identified that the customer faced challenges around thread safety, which was leading to locking issues. To resolve issues collaboratively with the client, we created a Slack channel to review the latest graphs and provide recommendations directly to the developers. After the developers implemented the suggested changes, we deployed a new build and had a corresponding graph in a few minutes.

Results

After just one week, our team successfully alleviated their scaling challenges at the web service layer. When we ran the load tests, we saw expected results of a few servers instead of the more than 80 servers previously. Additionally, because we optimized the code, we reduced the existing application footprint, which saved our customer 30% of compute load.

Cost savings aside, one of the most notable benefits of this project was developer education. With CodeGuru Profiler pinpointing where the bottlenecks were, the developers could recognize inefficient patterns in the code that might lead to severe performance hits down the road. This helped them better understand the features of the language they’re using and armed them with increased efficiency in future development and debugging.

Conclusion

With the web service layer better optimized, our next step is to use CodeGuru and other AWS tools like Performance Insights to tackle the database layer. Even if you aren’t experiencing extreme performance challenges, CodeGuru Profiler can provide valuable insights to the health of your application in any environment, from development all the way to production, with minimal CPU utilization. Integrating these results as part of the SDLC or DevOps process leads to better efficiency and gives you and your developers the tools you need to be successful. To learn more about how to get started with CodeGuru Profiler and CodeGuru Reviewer, check the documentation found here.


About the Author

Dustin Potter is a Principal Cloud Solutions Architect at EagleDream Technologies.

Read More

NVIDIA, Zoom CEOs Talk the Future of Work

NVIDIA, Zoom CEOs Talk the Future of Work

Amid a pandemic that’s put much of the world’s work, learning, even family reunions online, two of the leaders who have made today’s virtual world possible met Thursday on, where else — Zoom — to talk about what’s next.

NVIDIA CEO Jensen Huang and Zoom CEO Eric Yuan spoke Thursday at the online video conference company’s Zoomtopia user event in a casual, wide-ranging conversation.

“If not for what Zoom has done, the recent pandemic would be unbearable,” Huang said. The present situation, Huang explained, “has accelerated the future, it has brought forward the urgency of a digital future.”

In front of a virtual audience from all over the globe, the two spoke about their entrepreneurial journeys, NVIDIA’s unique company culture, and how NVIDIA is knitting together the virtual and real worlds to help NVIDIA employees collaborate.

Huang’s appearance at Zoomtopia follows NVIDIA’s GPU Technology Conference last week, where Huang outlined NVIDIA’s view of data center computing and introduced new technologies in data centers, edge AI and healthcare.

Yuan playfully wore a leather jacket, matching Huang’s trademark attire—and briefly displayed a sleek virtual kitchen as his backdrop, paying tribute to the presentations Huang has given from his kitchen this year—began their conversation with Huang by asking about his early life.

“I was fortunate that my parents worked hard and all of the people I was surrounded by worked hard,” Huang said, adding that he was focused on on school and sports, especially table tennis. “To me working is living, working is breathing and, to me, it’s not work at all — I enjoy it too much.”

It’s NVIDIA’s mission, Huang said, that continues to motivate him, as the company has gone from inventing the GPU to pioneering new possibilities in robotics and AI.

The common thread: since the beginning, NVIDIA has had a singular focus on accelerated computing.

“We built a time machine,” Huang said, touching on NVIDIA’s work in drug discovery as an example. “So, instead of a particular drug taking 10 years to discover, we would like drugs and therapies and vaccines to be discovered in months.”

Zoom and NVIDIA, Huang said, share a “singular purpose and a sense of destiny,” Huang said, one that has made the world a better place.

“The fact that Zoom existed and your vision came to reality means we can be together even if we’re not together,” Huang said.

“You can look at your work and imagine the impact on society and the benefits it will bring and somehow it’s your job to do it,” Huang said. “If you don’t do it, no one else will — and that’s thrilling to me, I love that feeling.”

Yuan also asked about NVIDIA’s culture and the future of work, one which Huang believes will increasingly meld the physical and the virtual worlds.

Today, for example, we might report to your colleagues that we’ll be WFH, or working from home.

Office lingo, however, may change to reflect the new reality, where being at the office isn’t necessarily the norm.

“In the future we will say we’re ‘going to the office,’” Huang said. “Today we say ‘WFH,’ in the future we will say ‘GTO.’”

Tools such as Zoom enable colleagues to meet, face to face, from home, from an office, from anywhere in the world.

More and more, work will take place in a hybrid of office and home, physical and virtual reality.

NVIDIA, for example, has created a platform called NVIDIA Omniverse that lets colleagues working in different places and with different tools collaborate in real time.

“The Adobe world can connect to the Catia world and so on,” Huang said. “We can have different designers working with each other at their homes.”

The present moment has “brought forward the urgency of a digital future, it has made us aware that completely physical is not sufficient, that completely digital is not sufficient,” Huang said. “The future is a mixed reality world.”

The post NVIDIA, Zoom CEOs Talk the Future of Work appeared first on The Official NVIDIA Blog.

Read More

Recreating Historical Streetscapes Using Deep Learning and Crowdsourcing

Recreating Historical Streetscapes Using Deep Learning and Crowdsourcing

Posted by Raimondas Kiveris, Software Engineer, Google Research

For many, gazing at an old photo of a city can evoke feelings of both nostalgia and wonder — what was it like to walk through Manhattan in the 1940s? How much has the street one grew up on changed? While Google Street View allows people to see what an area looks like in the present day, what if you want to explore how places looked in the past?

To create a rewarding “time travel” experience for both research and entertainment purposes, we are launching (pronounced as re”turn“), an open source, scalable system running on Google Cloud and Kubernetes that can reconstruct cities from historical maps and photos, representing an implementation of our suite of open source tools launched earlier this year. Referencing the common prefix meaning again or anew, is meant to represent the themes of reconstruction, research, recreation and remembering behind this crowdsourced research effort, and consists of three components:

  • A crowdsourcing platform, which allows users to upload historical maps of cities, georectify (i.e., match them to real world coordinates), and vectorize them
  • A temporal map server, which shows how maps of cities change over time
  • A 3D experience platform, which runs on top of the map server, creating the 3D experience by using deep learning to reconstruct buildings in 3D from limited historical images and maps data.

Our goal is for to become a compendium that allows history enthusiasts to virtually experience historical cities around the world, aids researchers, policy makers and educators, and provides a dose of nostalgia to everyday users.

Bird’s eye view of Chelsea, Manhattan with a time slider from 1890 to 1970, crafted from historical photos and maps and using ’s 3D reconstruction pipeline and colored with a preset Manhattan-inspired palette.

Crowdsourcing Data from Historical Maps
Reconstructing how cities used to look at scale is a challenge — historical image data is more difficult to work with than modern data, as there are far fewer images available and much less metadata captured from the images. To help with this difficulty, the maps module is a suite of open source tools that work together to create a map server with a time dimension, allowing users to jump back and forth between time periods using a slider. These tools allow users to upload scans of historical print maps, georectify them to match real world coordinates, and then convert them to vector format by tracing their geographic features. These vectorized maps are then served on a tile server and rendered as slippy maps, which lets the user zoom in and pan around.

Sub-modules of the suite of tools

The entry point of the maps module is Warper, a web app that allows users to upload historical images of maps and georectify them by finding control points on the historical map and corresponding points on a base map. The next app, Editor, allows users to load the georectified historical maps as the background and then trace their geographic features (e.g., building footprints, roads, etc.). This traced data is stored in an OpenStreetMap (OSM) vector format. They are then converted to vector tiles and served from the Server app, a vector tile server. Finally, our map renderer, Kartta, visualizes the spatiotemporal vector tiles allowing the users to navigate space and time on historical maps. These tools were built on top of numerous open source resources including OpenStreetMap, and we intend for our tools and data to be completely open source as well.

Warper and Editor work together to let users upload a map, anchor it to a base map using control points, and trace geographic features like building footprints and roads.

3D Experience
The 3D Models module aims to reconstruct the detailed full 3D structures of historical buildings using the associated images and maps data, organize these 3D models properly in one repository, and render them on the historical maps with a time dimension.

In many cases, there is only one historical image available for a building, which makes the 3D reconstruction an extremely challenging problem. To tackle this challenge, we developed a coarse-to-fine reconstruction-by-recognition algorithm.

High-level overview of ’s 3D reconstruction pipeline, which takes annotated images and maps and prepares them for 3D rendering.

Starting with footprints on maps and façade regions in historical images (both are annotated by crowdsourcing or detected by automatic algorithms), the footprint of one input building is extruded upwards to generate its coarse 3D structure. The height of this extrusion is set to the number of floors from the corresponding metadata in the maps database.

In parallel, instead of directly inferring the detailed 3D structures of each façade as one entity, the 3D reconstruction pipeline recognizes all individual constituent components (e.g., windows, entries, stairs, etc.) and reconstructs their 3D structures separately based on their categories. Then these detailed 3D structures are merged with the coarse one for the final 3D mesh. The results are stored in a 3D repository and ready for 3D rendering.

The key technology powering this feature is a number of state-of-art deep learning models:

  • Faster region-based convolutional neural networks (RCNN) were trained using the façade component annotations for each target semantic class (e.g., windows, entries, stairs, etc), which are used to localize bounding-box level instances in historical images.
  • DeepLab, a semantic segmentation model, was trained to provide pixel-level labels for each semantic class.
  • A specifically designed neural network was trained to enforce high-level regularities within the same semantic class. This ensured that windows generated on a façade were equally spaced and consistent in shape with each other. This also facilitated consistency across different semantic classes such as stairs to ensure they are placed at reasonable positions and have consistent dimensions relative to the associated entry ways.

Key Results

Street level view of 3D-reconstructed Chelsea, Manhattan

Conclusion
With , we have developed tools that facilitate crowdsourcing to tackle the main challenge of insufficient historical data when recreating virtual cities. The 3D experience is still a work-in-progress and we aim to improve it with future updates. We hope acts as a nexus for an active community of enthusiasts and casual users that not only utilizes our historical datasets and open source code, but actively contributes to both.

Acknowledgements
This effort has been successful thanks to the hard work of many people, including, but not limited to the following (in alphabetical order of last name): Yale Cong, Feng Han, Amol Kapoor, Raimondas Kiveris, Brandon Mayer, Mark Phillips, Sasan Tavakkol, and Tim Waters (Waters Geospatial Ltd).

Read More

Europe Launches New Era in HPC with World’s Fastest AI Supercomputer

Europe Launches New Era in HPC with World’s Fastest AI Supercomputer

Four new supercomputers backed by a pan-European initiative will use NVIDIA’s data center accelerators, networks and software to advance AI and high performance computing.

They include one system dubbed Leonardo, unveiled today at Italy’s CINECA research center, using NVIDIA technologies to deliver the world’s most powerful AI system. The four mark the first of eight systems to be announced this year targeting spots among the world’s 50 most powerful computers.

Together, they’ll form a regional network, “an engine to power Europe’s data economy,” said EuroHPC, the group driving the effort, in a white paper outlining its goals.

The systems will apply AI and data analytics across scientific and commercial applications that range from fighting COVID-19 and climate change to the design of advanced airplanes, cars, drugs and materials.

Joining Leonardo are a wave of new AI supercomputers planned for the Czech Republic, Luxembourg and Slovenia that will act as national centers of competence, expanding skills and creating jobs.

NVIDIA GPUs, InfiniBand Power Latest Systems

All four supercomputers announced use NVIDIA Ampere architecture GPUs and NVIDIA Mellanox HDR InfiniBand networks to tap an ecosystem of hundreds of HPC and AI applications. Atos, an NVIDIA systems partner headquartered in France, will build three of the four systems; Hewlett Packard Enterprise will construct the fourth.

The new systems join 333 of the world’s TOP500 supercomputers powered by NVIDIA GPUs, networking or both.

NVIDIA GPUs accelerate 1,800 HPC applications, nearly 800 of them available today in the GPU application catalog and NGC, NVIDIA’s hub for GPU-optimized software.

The new systems all use HDR 200Gb/s InfiniBand for low latency, high throughput and in-network computing. It’s the latest version of InfiniBand, already powering supercomputers across Europe.

A Brief Tour of Europe’s Latest Supercomputers

Leonardo will be the world’s fastest AI supercomputer. Atos is harnessing nearly 14,000 NVIDIA Ampere architecture GPUs and HDR 200Gb/s InfiniBand networking to deliver a system with 10 exaflops of AI performance. It will use the InfiniBand Dragonfly+ network topology to deliver both flexibility and scalable performance.

Researchers at CINECA will apply that power to advance science, simulating planetary forces behind climate change and molecular movements inside a coronavirus. The center is perhaps best known for its work on Quantum Espresso, a suite of open source codes for modeling nanoscale materials for jobs such as engineering better batteries.

A new supercomputer in Luxembourg called MeluXina, also part of the EuroHPC network, will connect 800 NVIDIA A100 GPUs on HDR 200Gb/s InfiniBand links. The system, to be built by Atos and powered by green energy from wood waste, will pack nearly 500 petaflops of AI performance.

MeluXina will address commercial applications and scientific research. It plans to offer access to users leveraging HPC and AI to advance work in financial services as well as manufacturing and healthcare.

Eastern Europe Powers Up

The new Vega supercomputer at the Institute of Information Science in Maribor, Slovenia, (IZUM) will be based on the Atos BullSequana XH2000 system. The supercomputer, named after Slovenian mathematician Jurij Vega, includes 240 A100 GPUs and 1,800 HDR 200Gb/s InfiniBand end points.

Vega will help “ensure a new generation of experts and developers, as well as the wider Slovenian community, can meet new challenges within our national consortium and contribute to regional and European HPC initiatives,” said Aleš Bošnjak, IZUM’s director in a statement issued by EuroHPC.

EuroHPC map
A total of 32 countries are participating in the EuroHPC effort.

The IT4Innovations National Supercomputing Center will host what’s expected to become the most powerful supercomputer in the Czech Republic. It will use 560 NVIDIA A100 GPUs to deliver nearly 350 petaflops of AI performance — 7x the capabilities of the center’s existing system.

The supercomputer will be based on the HPE Apollo 6500 systems from Hewlett Packard Enterprise (HPE). It will serve researchers at the VSB – Technical University of Ostrava, where it’s based, as well as an expanding set of external academic and industrial users employing a mix of simulations, data analytics and AI.

The story of Europe’s ambitions in HPC and AI is still being written.

EuroHPC has yet to announce its plans for systems in Bulgaria, Finland, Portugal and Spain. And beyond that work, the group has already sketched out plans that stretch to 2027.

The post Europe Launches New Era in HPC with World’s Fastest AI Supercomputer appeared first on The Official NVIDIA Blog.

Read More

Project Euphonia’s new step: 1,000 hours of speech recordings

Project Euphonia’s new step: 1,000 hours of speech recordings

Muratcan Cicek, a PhD candidate at UC Santa Cruz, worked as a summer intern on Google’s Project Euphonia, which aims to improve computers’ abilities to understand impaired speech. This work was especially relevant and important for Muratcan, who was born with cerebral palsy and has a severe speech impairment.

Before his internship, Muratcan recorded 2,000 phrases for Project Euphonia. These phrases, expressions like “Turn the lights on” and “Turn up thermostat to 74 degrees,” were used to build a personalized speech recognition model that could better recognize the unique sound of his voice and transcribe his speech. The prototype allowed Muratcan to share the transcription in a video call so others could better understand him. He used the prototype to converse with co-workers, give status updates during team meetings and connect with people in ways that were previously impossible. Muratcan says, “Euphonia transformed my communication skills in a way that I can leverage in my career as an engineer without feeling insecure about my condition.”

Muratcan, a Google intern

Muratcan, a summer research intern on the Euphonia team, uses the Euphonia prototype app

1,000 hours of speech samples

The phrases that Muratcan recorded were key to training custom machine learning models that could help him be more easily understood. To help other people that have impaired speech caused by ALS, Parkinson’s disease or Down Syndrome, we need to gather samples of their speech patterns. So we’ve worked with partners like CDSS, ALS TDI, ALSA, LSVT Global, Team Gleason and CureDuchenne to encourage people with speech impairments to record their voices and contribute to this research.

Since 2018, nearly 1,000 participants have recorded over 1,000 hours of speech samples. For many, it’s been a source of pride and purpose to shape the future of speech recognition, not only for themselves but also for others who struggle to be understood.

I contribute to this research so that I can help not only myself, but also a larger group of people with communication challenges that are often left out. Project Euphonia participant

While the technology is still under development, the speech samples we’ve collected helped us create personalized speech recognition models for individuals with speech impairments, like Muratcan. For more technical details about how these models work, see the Euphonia and Parrotron blog posts. We’re evaluating these personalized models with a group of early testers. The next phase of our research aims to improve speech recognition systems for many more people, but it requires many more speech samples from a broad range of speakers.

How you can contribute

To continue our research, we hope to collect speech samples from an additional 5,000 participants. If you have difficulty being understood by others and want to contribute to meaningful research to improve speech recognition technologies, learn more and consider signing up to record phrases. We look forward to hearing from more participants and experts— and together, helping everyone be understood.

Read More

AI Draws World’s Smallest Wanted Posters to Apprehend COVID

AI Draws World’s Smallest Wanted Posters to Apprehend COVID

Using AI and a supercomputer simulation, Ken Dill’s team drew the equivalent of wanted posters for a gang of proteins that make up COVID-19. With a little luck, one of their portraits could identify a way to arrest the coronavirus with a drug.

When the pandemic hit, “it was terrible for the world, and a big research challenge for us,” said Dill, who leads the Laufer Center for Physical & Quantitative Biology at Stony Brook University, in Long Island, New York.

For a decade, he helped the center assemble the researchers and tools needed to study the inner workings of proteins — complex molecules that are fundamental to cellular life. The center has a history of applying its knowledge to viral proteins, helping others identify drugs to disable them.

“So, when the pandemic came, our folks wanted to spring into action,” he said.

AI, Simulations Meet at the Summit

The team aimed to use a combination of physics and AI tools to predict the 3D structure of more than a dozen coronavirus proteins based on lists of the amino acid strings that define them. It won a grant for time on the IBM-built Summit supercomputer at Oak Ridge National Laboratory to crunch its complex calculations.

“We ran 30 very extensive simulations in parallel, one on each of 30 GPUs, and we ran them continuously for at least four days,” explained Emiliano Brini, a junior fellow at the Laufer Center. “Summit is a great machine because it has so many GPUs, so we can run many simulations in parallel,” he said.

“Our physics-based modeling eats a lot of compute cycles. We use GPUs almost exclusively for their speed,” said Dill.

Sharing Results to Help Accelerate Research

Thanks to the acceleration, the predictions are already in. The Laufer team quickly shared them with about a hundred researchers working on a dozen separate projects that conduct painstakingly slow experiments to determine the actual structure of the proteins.

“They indicated some experiments could be done faster if they had hunches from our work of what those 3D structures might be,” said Dill.

Now it’s a waiting game. If one of the predictions gives researchers a leg up in finding a weakness that drug makers can exploit, it would be a huge win. It could take science one step closer to putting a general antiviral drug on the shelf of your local pharmacy.

Melding Machine Learning and Physics

Dill’s team uses a molecular dynamics program called MELD. It blends physical simulations with insights from machine learning based on statistical models.

AI provides MELD key information to predict a protein’s 3D structure from its sequence of amino acids. It quickly finds patterns across a database of atomic-level information on 200,000 proteins gathered over the last 50 years.

MELD uses this information in compute-intensive physics simulations to determine the protein’s detailed structure. Further simulations then can predict, for example, what drug molecules will bind tightly to a specific viral protein.

“So, both these worlds — AI inference and physics simulations — are playing big roles in helping drug discovery,” said Dill. “We get the benefits of both methods, and that combination is where I think the future is.”

MELD runs on CUDA, NVIDIA’s accelerated computing platform for GPUs. “It would take prohibitively long to run its simulations on CPUs, so the majority of biological simulations are done on GPUs,” said Brini.

Playing a Waiting Game

The COVID-19 challenge gave Laufer researchers with a passion for chemistry a driving focus. Now they await feedback on their work on Summit.

“Once we get the results, we’ll publish what we learn from the mistakes. Many times, researchers have to go back to the drawing board,” he said.

And every once in a while, they celebrate, too.

Dill hosted a small, socially distanced gathering for a half-dozen colleagues in his backyard after the Summit work was complete. If those results turn up a win, there will be a much bigger celebration extending far beyond the Stony Brook campus.

The post AI Draws World’s Smallest Wanted Posters to Apprehend COVID appeared first on The Official NVIDIA Blog.

Read More

Machine learning uncovers potential new TB drugs

Machine learning uncovers potential new TB drugs

Machine learning is a computational tool used by many biologists to analyze huge amounts of data, helping them to identify potential new drugs. MIT researchers have now incorporated a new feature into these types of machine-learning algorithms, improving their prediction-making ability.

Using this new approach, which allows computer models to account for uncertainty in the data they’re analyzing, the MIT team identified several promising compounds that target a protein required by the bacteria that cause tuberculosis.

This method, which has previously been used by computer scientists but has not taken off in biology, could also prove useful in protein design and many other fields of biology, says Bonnie Berger, the Simons Professor of Mathematics and head of the Computation and Biology group in MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL).

“This technique is part of a known subfield of machine learning, but people have not brought it to biology,” Berger says. “This is a paradigm shift, and is absolutely how biological exploration should be done.”

Berger and Bryan Bryson, an assistant professor of biological engineering at MIT and a member of the Ragon Institute of MGH, MIT, and Harvard, are the senior authors of the study, which appears today in Cell Systems. MIT graduate student Brian Hie is the paper’s lead author.

Better predictions

Machine learning is a type of computer modeling in which an algorithm learns to make predictions based on data that it has already seen. In recent years, biologists have begun using machine learning to scour huge databases of potential drug compounds to find molecules that interact with particular targets.

One limitation of this method is that while the algorithms perform well when the data they’re analyzing are similar to the data they were trained on, they’re not very good at evaluating molecules that are very different from the ones they have already seen.

To overcome that, the researchers used a technique called Gaussian process to assign uncertainty values to the data that the algorithms are trained on. That way, when the models are analyzing the training data, they also take into account how reliable those predictions are.

For example, if the data going into the model predict how strongly a particular molecule binds to a target protein, as well as the uncertainty of those predictions, the model can use that information to make predictions for protein-target interactions that it hasn’t seen before. The model also estimates the certainty of its own predictions. When analyzing new data, the model’s predictions may have lower certainty for molecules that are very different from the training data. Researchers can use that information to help them decide which molecules to test experimentally.

Another advantage of this approach is that the algorithm requires only a small amount of training data. In this study, the MIT team trained the model with a dataset of 72 small molecules and their interactions with more than 400 proteins called protein kinases. They were then able to use this algorithm to analyze nearly 11,000 small molecules, which they took from the ZINC database, a publicly available repository that contains millions of chemical compounds. Many of these molecules were very different from those in the training data.

Using this approach, the researchers were able to identify molecules with very strong predicted binding affinities for the protein kinases they put into the model. These included three human kinases, as well as one kinase found in Mycobacterium tuberculosis. That kinase, PknB, is critical for the bacteria to survive, but is not targeted by any frontline TB antibiotics.

The researchers then experimentally tested some of their top hits to see how well they actually bind to their targets, and found that the model’s predictions were very accurate. Among the molecules that the model assigned the highest certainty, about 90 percent proved to be true hits — much higher than the 30 to 40 percent hit rate of existing machine learning models used for drug screens.

The researchers also used the same training data to train a traditional machine-learning algorithm, which does not incorporate uncertainty, and then had it analyze the same 11,000 molecule library. “Without uncertainty, the model just gets horribly confused and it proposes very weird chemical structures as interacting with the kinases,” Hie says.

The researchers then took some of their most promising PknB inhibitors and tested them against Mycobacterium tuberculosis grown in bacterial culture media, and found that they inhibited bacterial growth. The inhibitors also worked in human immune cells infected with the bacterium.

A good starting point

Another important element of this approach is that once the researchers get additional experimental data, they can add it to the model and retrain it, further improving the predictions. Even a small amount of data can help the model get better, the researchers say.

“You don’t really need very large data sets on each iteration,” Hie says. “You can just retrain the model with maybe 10 new examples, which is something that a biologist can easily generate.”

This study is the first in many years to propose new molecules that can target PknB, and should give drug developers a good starting point to try to develop drugs that target the kinase, Bryson says. “We’ve now provided them with some new leads beyond what has been already published,” he says.

The researchers also showed that they could use this same type of machine learning to boost the fluorescent output of a green fluorescent protein, which is commonly used to label molecules inside living cells. It could also be applied to many other types of biological studies, says Berger, who is now using it to analyze mutations that drive tumor development.

The research was funded by the U.S. Department of Defense through the National Defense Science and Engineering Graduate Fellowship; the National Institutes of Health; the Ragon Institute of MGH, MIT, and Harvard’ and MIT’s Department of Biological Engineering.

Read More

How GPUs Are Helping Paris’ Public Hospital System Combat the Spread of COVID-19

How GPUs Are Helping Paris’ Public Hospital System Combat the Spread of COVID-19

In the battle against COVID-19, Greater Paris University Hospitals – Public Assistance Hospital of Paris (AP-HP is the French acronym) isn’t just on the medical front lines — it’s on the data front lines as well.

With a network of 39 hospitals treating 8.3 million patients each year, AP-HP is a major actor in the fight against COVID-19.

Along with its COVID-19 cases comes an awful lot of data, including now geodata that can potentially help lessen the impact of the pandemic. AP-HP, which partners with seven universities, already had the ability to analyze large amounts of medical data. It had previously created dashboards that combined cancer cases and geodata. So, it was logical to pursue and extend its role during the pandemic.

The expected volume of COVID-19 data and geodata would probably have tested AP-HP’s data crunching capacity. To mitigate this critical challenge, the hospital’s information systems administrators turned to Kinetica, a provider of streaming data warehouses and real-time analytics and a member of the NVIDIA Inception program for AI startups.

Kinetica’s offering harnesses the power of NVIDIA GPUs to quickly convert case location data into usable intelligence. And in the fight against COVID-19, speed is everything.

The project team also used NVIDIA RAPIDS to speed up the machine learning algorithms integrated into the platform. RAPIDS accelerates analytics and data science pipelines on NVIDIA GPUs by taking advantage of GPU parallelism and high memory bandwidth.

“Having the ability to perform this type of analysis in real time is really important during a pandemic,” said Hector Countouris, the project lead at AP-HP. “And more data is coming.”

Analyzing COVID Contact Data

What Countouris and his colleagues are most focused on is using COVID-related geodata to understand where virus “hot spots” are and the dynamic of the outbreak. Looking for cluster locations can help decision-making at the district or region level.

In addition, they’re looking at new signals to improve early detection of COVID patients. This includes working with data from other regional agencies.

If patients are diagnosed with COVID, they’ll be asked by the relevant agencies via a phone call about their recent whereabouts and contacts to help with contact tracing. This is the first time that a wide range of data from different partners in the Paris area will be integrated to allow for contact tracing and timely alerts about a potential exposure. The result will be a newfound ability to see how clusters of COVID-19 cases evolve.

“We hope that in the near future we will be able to follow how a cluster evolves in real time,” said Countouris.

The goal is to enable public health decision-makers to implement prevention and control measures and assess their effectiveness. The data can also be integrated with other demographic data to study the viral spread and its possible dependency on socio-economics and other factors.

Attacking Bottlenecks with GPUs

Prior to engaging with Kinetica, such data-intensive projects involved so much time for loading the data that they couldn’t be analyzed quickly enough to deliver real-time benefits.

“Now, I don’t feel like I have a bottleneck,” said Countouris. “We are continuously integrating data and delivering dashboards to decision makers within hours. And with robust real-time pipelines allowing for continuous data ingestion, we can now focus on building better dashboards.”

In the past, to get data in a specific and usable format, they would need to do a lot of pre-processing. With Kinetica’s Streaming Data Warehouse powered by NVIDIA V100 Tensor Core GPUs, that’s no longer the case. Users can access the much richer datasets they demand.

Kinetica’s platform is available on NVIDIA NGC, a catalog of GPU-optimized AI containers that let enterprises quickly operationalize extreme analytics, machine learning and data visualization. This eliminates complexity and lets organizations deploy cloud, on-premises or hybrid models for optimal business operations.

“I don’t think we could meet user expectations for geodata without GPU power,” he said. “There is just too much data and geodata to provide for too many users at the same time.”

AP-HP’s COVID-related work has already built a foundation upon which to do follow-up work related to emergency responses in general. The hospital information system’s interest for that kind of data is far from over.

“The fact that we helped the decision-making process and that officials are using our data is the measure of success,” said Countouris. “We have a lot to do. This is only the beginning.”

Countouris presented the team’s work last week at the GPU Technology Conference. Registered GTC attendees can view the talk on demand. It will be available for replay to the general public early next month.

Kinetica will also be part of the NVIDIA Startup Village Booth at the HLTH conference, presenting on Oct. 16 at 2 p.m. Pacific time.

The post How GPUs Are Helping Paris’ Public Hospital System Combat the Spread of COVID-19 appeared first on The Official NVIDIA Blog.

Read More