MatterGen: Property-guided materials design

MatterGen: Property-guided materials design

MatterGen

Generative AI has revolutionized how we create text and images. How about designing novel materials? We at Microsoft Research AI4Science are thrilled to announce MatterGen, our generative model that enables broad property-guided materials design.

The central challenge in materials science is to discover materials with desired properties, e.g., high Li-ion conductivity for battery materials. Traditionally, this has been done by first finding novel materials and then filtering down based on the application. This is like trying to create the image of a cat by first generating a million different images and then searching for the one with a cat. In MatterGen, we directly generate novel materials with desired properties, similar to how DALL·E 3 tackles image generation.  

MatterGen is a diffusion model specifically designed for generating novel, stable materials. MatterGen also has adapter modules that can be fine-tuned to generate materials given a broad range of constraints, including chemistry, symmetry, and properties. MatterGen generates 2.9 times more stable (≤ 0.1 eV/atom of our training + test data convex hull), novel, unique structures than a SOTA model (CDVAE). It also generates structures 17.5 times closer to energy local minimum. MatterGen can directly generate materials satisfying desired magnetic, electronic, mechanical properties via classifier-free guidance. We verify generated materials with DFT-based workflows. 

Figure 1 (alt text) 

This figure displays six pairs of crystalline structures, two for each property constrain. The property constraints are, top to bottom and left to right, high space group symmetry, high bulk modulus, target chemical system, target band gap, high magnetic density, combined high magnetic density and low HHI index.
Figure 1: Stable and new materials generated by MatterGen while constrained on properties. 

Additionally, MatterGen can keep generating novel materials that satisfy a target property like high bulk modulus while screening methods instead saturate due to the exhaustion of materials in the database.

This is a line plot. The x axis indicates the number of DFT property calculations calls; the y axis reports the number of structures found. The title of the plot says
Figure 2: MatterGen discovers more novel stable high bulk modulus materials than the screening baseline, and does not plateau for increasing computational resources. MatterGen can find more than 250 materials with a bulk modulus > 400 GPa, while only 2 such materials are found in the reference dataset.

MatterGen can also generate materials given target chemical systems. It outperforms substitution and random structure search baselines equipped with MLFF filtering, especially in challenging 5-element systems. MatterGen also generates structures given target space groups. Finally, we tackle the multi-property materials design problem of finding low-supply-chain risk magnets. MatterGen proposes structures that have both high magnetic density and a low supply-chain risk chemical composition. 

We believe MatterGen is an important step forward in AI for materials design. Our results are currently verified via DFT, which has many known limitations. Experimental verification remains the ultimate test for real-word impact, and we hope to follow up with more results. 

None of this would be possible without the highly collaborative work between Andrew Fowler, Claudio Zeni, Daniel Zügner, Matthew Horton, Robert Pinsler, Ryota Tomioka, Tian Xie and our amazing interns Xiang Fu, Sasha Shysheya, and Jonathan Crabbé, as well as Jake Smith, Lixin Sun and the entire AI4Science Materials Design team.  

We are also grateful for all the help from Microsoft Research, AI4Science, and Azure Quantum.

The post MatterGen: Property-guided materials design appeared first on Microsoft Research.

Read More

500 Games and Apps Now Powered by RTX: A DLSS and Ray-Tracing Milestone

500 Games and Apps Now Powered by RTX: A DLSS and Ray-Tracing Milestone

We’re celebrating a milestone this week with 500 RTX games and applications utilizing NVIDIA DLSS, ray tracing or AI technologies. It’s an achievement anchored by NVIDIA’s revolutionary RTX technology, which has transformed gaming graphics and performance.

The journey began in 2018 at an electrifying event in Cologne. In a steel and concrete music venue amidst the city’s gritty industrial north side, over 1,200 gamers, breathless and giddy, erupted as NVIDIA founder and CEO Jensen Huang introduced NVIDIA RTX and declared, “This is a historic moment … Computer graphics has been reinvented.”

This groundbreaking launch, set against the backdrop of the world’s largest gaming expo, Gamescom, marked the introduction of the GeForce RTX 2080 Ti, 2080 and 2070 GPUs.

Launched in 2018, NVIDIA RTX has redefined visual fidelity and performance in modern gaming and creative applications.
Launched in 2018, NVIDIA RTX has redefined visual fidelity and performance in modern gaming and creative applications.

The most technically advanced games now rely on the techniques that RTX technologies have unlocked.

Ray tracing, enabled by dedicated RT Cores, delivers immersive, realistic lighting and reflections in games.

The technique has evolved from games with only a single graphics element executed in ray tracing to games such as Alan Wake 2, Cyberpunk 2077, Minecraft RTX and Portal RTX that use ray tracing for all the light in the game.

And NVIDIA DLSS, powered by Tensor Cores, accelerates AI graphics, now boosting performance with DLSS Frame Generation and improving RT effects with DLSS Ray Reconstruction in titles like Cyberpunk 2077: Phantom Liberty.

Beyond gaming, these technologies revolutionize creative workflows, enabling real-time, ray-traced previews in applications that once required extensive processing time.

Ray tracing, a technique first described in 1969 by Arthur Appel, mirrors how light interacts with objects to create lifelike images.

Ray tracing was once limited to high-end movie production. NVIDIA’s RTX graphics cards have made this cinematic quality accessible in real-time gaming, enhancing experiences with dynamic lighting, reflections and shadows.

High engagement rates in titles like Cyberpunk 2077, NARAKA: BLADEPOINT, Minecraft with RTX, Alan Wake 2 and Diablo IV, where 96% or higher of RTX 40 Series t gamers use RTX ON, underscore this success.

To commemorate this milestone, 20 $500 Green Man Gaming gift cards and exclusive #RTXON keyboard keycaps are up for grabs. Participants must follow GeForce’s social channels and comply with the sweepstakes rules.

Stay tuned for more RTX 500 giveaways.

NVIDIA’s advancement from the first RTX graphics card to powering 500 RTX games and applications with advanced technologies heralds a new gaming and creative tech era. And NVIDIA continues to lead, offering unparalleled experiences in gaming and creativity.

Stay tuned to GeForce News for more updates on RTX games and enhancements.

Read More

VALID: A perceptually validated virtual avatar library for inclusion and diversity

VALID: A perceptually validated virtual avatar library for inclusion and diversity

As virtual reality (VR) and augmented reality (AR) technologies continue to grow in popularity, virtual avatars are becoming an increasingly important part of our digital interactions. In particular, virtual avatars are at the center of many social VR and AR interactions, as they are key to representing remote participants and facilitating collaboration.

In the last decade, interdisciplinary scientists have dedicated a significant amount of effort to better understand the use of avatars, and have made many interesting observations, including the capacity of the users to embody their avatar (i.e., the illusion that the avatar body is their own) and the self-avatar follower effect, which creates a binding between the actions of the avatar and the user strong enough that the avatar can actually affect user behavior.

The use of avatars in experiments isn’t just about how users will interact and behave in VR spaces, but also about discovering the limits of human perception and neuroscience. In fact, some VR social experiments often rely on recreating scenarios that can’t be reproduced easily in the real world, such as bar crawls to explore ingroup vs. outgroup effects, or deception experiments, such as the Milgram obedience to authority inside virtual reality. Other studies try to explore deep neuroscientific phenomena, like the human mechanisms for motor control. This perhaps follows the trail of the rubber hand illusion on brain plasticity, where a person can start feeling as if they own a rubber hand while their real hand is hidden behind a curtain. There is also an increased number of possible therapies for psychiatric treatment using personalized avatars. In these cases, VR becomes an ecologically valid tool that allows scientists to explore or treat human behavior and perception.

None of these experiments and therapies could exist without good access to research tools and libraries that can enable easy experimentation. As such, multiple systems and open source tools have been released around avatar creation and animation over recent years. However, existing avatar libraries have not been validated systematically on the diversity spectrum. Societal bias and dynamics also transfer to VR/AR when interacting with avatars, which could lead to incomplete conclusions for studies on human behavior inside VR/AR.

To partially overcome this problem, we partnered with the University of Central Florida to create and release the open-source Virtual Avatar Library for Inclusion and Diversity (VALID). Described in our recent paper, published in Frontiers in Virtual Reality, this library of avatars is readily available for usage in VR/AR experiments and includes 210 avatars of seven different races and ethnicities recognized by the US Census Bureau. The avatars have been perceptually validated and designed to advance diversity and inclusion in virtual avatar research.

Headshots of all 42 base avatars available on the VALID library were created in extensive interaction with members of the 7 ethnic and racial groups from the Federal Register, which include (AIAN, Asian, Black, Hispanic, MENA, NHPI and White).

Creation and validation of the library

Our initial selection of races and ethnicities for the diverse avatar library follows the most recent guidelines of the US Census Bureau that as of 2023 recommended the use of 7 ethnic and racial groups representing a large demographic of the US society, which can also be extrapolated to the global population. These groups include Hispanic or Latino, American Indian or Alaska Native (AIAN), Asian, Black or African American, Native Hawaiian or Other Pacific Islander (NHPI), White, Middle East or North Africa (MENA). We envision the library will continue to evolve to bring even more diversity and representation with future additions of avatars.

The avatars were hand modeled and created using a process that combined average facial features with extensive collaboration with representative stakeholders from each racial group, where their feedback was used to artistically modify the facial mesh of the avatars. Then we conducted an online study with participants from 33 countries to determine whether the race and gender of each avatar in the library are recognizable. In addition to the avatars, we also provide labels statistically validated through observation of users for the race and gender of all 42 base avatars (see below).

Example of the headshots of a Black/African American avatar presented to participants during the validation of the library.

We found that all Asian, Black, and White avatars were universally identified as their modeled race by all participants, while our American Indian or Native Alaskan (AIAN), Hispanic, and Middle Eastern or North African (MENA) avatars were typically only identified by participants of the same race. This also indicates that participant race can improve identification of a virtual avatar of the same race. The paper accompanying the library release highlights how this ingroup familiarity should also be taken into account when studying avatar behavior in VR.

Confusion matrix heatmap of agreement rates for the 42 base avatars separated by other-race participants and same-race participants. One interesting aspect visible in this matrix, is that participants were significantly better at identifying the avatars of their own race than other races.

Dataset details

Our models are available in FBX format, are compatible with previous avatar libraries like the commonly used Rocketbox, and can be easily integrated into most game engines such as Unity and Unreal. Additionally, the avatars come with 69 bones and 65 facial blendshapes to enable researchers and developers to easily create and apply dynamic facial expressions and animations. The avatars were intentionally made to be partially cartoonish to avoid extreme look-a-like scenarios in which a person could be impersonated, but still representative enough to be able to run reliable user studies and social experiments.

Images of the skeleton rigging (bones that allow for animation) and some facial blend shapes included with the VALID avatars.

The avatars can be further combined with variations of casual attires and five professional attires, including medical, military, worker and business. This is an intentional improvement from prior libraries that in some cases reproduced stereotypical gender and racial bias into the avatar attires, and provided very limited diversity to certain professional avatars.

Images of some sample attire included with the VALID avatars.

Get started with VALID

We believe that the Virtual Avatar Library for Inclusion and Diversity (VALID) will be a valuable resource for researchers and developers working on VR/AR applications. We hope it will help to create more inclusive and equitable virtual experiences. To this end, we invite you to explore the avatar library, which we have released under the open source MIT license. You can download the avatars and use them in a variety of settings at no charge.

Acknowledgements

This library of avatars was born out of a collaboration with Tiffany D. Do, Steve Zelenty and Prof. Ryan P McMahan from the University of Central Florida.

Read More

LLMLingua: Innovating LLM efficiency with prompt compression

LLMLingua: Innovating LLM efficiency with prompt compression

This research paper was presented at the 2023 Conference on Empirical Methods in Natural Language Processing (opens in new tab) (EMNLP 2023), the premier conference on natural language processing and artificial intelligence.

EMNLP 2023 logo to the left of accepted paper

As large language models (LLMs) models advance and their potential becomes increasingly apparent, an understanding is emerging that the quality of their output is directly related to the nature of the prompt that is given to them. This has resulted in the rise of prompting technologies, such as chain-of-thought (CoT) and in-context-learning (ICL), which facilitate an increase in prompt length. In some instances, prompts now extend to tens of thousands of tokens, or units of text, and beyond. While longer prompts hold considerable potential, they also introduce a host of issues, such as the need to exceed the chat window’s maximum limit, a reduced capacity for retaining contextual information, and an increase in API costs, both in monetary terms and computational resources.

To address these challenges, we introduce a prompt-compression method in our paper, “LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models (opens in new tab),” presented at EMNLP 2023 (opens in new tab). Using a well-trained small language model, such as GPT2-small or LLaMA-7B, LLMLingua identifies and removes unimportant tokens from prompts. This compression technique enables closed LLMs to make inferences from the compressed prompt. Although the token-level compressed prompts may be difficult for humans to understand, they prove highly effective for LLMs. This is illustrated in Figure 1.

This is an illustration of the LLMLingua framework, which estimates the important tokens of a prompt based on a small language model. It consists of three modules: a budget controller, iterative token-level prompt compression, and distribution alignment. The framework can compress a complex prompt of 2,366 tokens down to 117 tokens, achieving a 20x compression while maintaining almost unchanged performance.
Figure 1. LLMLingua’s framework

LLMLingua’s method and evaluation

To develop LLMLingua’s framework, we employed a budget controller to balance the sensitivities of different modules in the prompt, preserving the language’s integrity. Our two-stage process involved course-grained prompt compression. We first streamlined the prompt by eliminating certain sentences and then individually compressed the remaining tokens. To preserve coherence, we employed an iterative token-level compression approach, refining the individual relationships between tokens. Additionally, we fine-tuned the smaller model to capture the distribution information from different closed LLMs by aligning it with the patterns in the LLMs’ generated data. We did this through instruction tuning.

To assess LLMLingua’s performance, we tested compressed prompts on four different datasets, GSM8K, BBH, ShareGPT, and Arxiv-March23, encompassing ICL, reasoning, summarization, and conversation. Our approach achieved impressive results, achieving up to 20x compression while preserving the original prompt’s capabilities, particularly in ICL and reasoning. LLMLingua also significantly reduced system latency.

During our test, we used LLaMA-7B as the small language model and GPT-3.5-Turbo-0301, one of OpenAI’s LLMs, as the closed LLM. The results show that LLMLingua maintains the original reasoning, summarization, and dialogue capabilities of the prompt, even at a maximum compression ratio of 20x, as reflected in the evaluation metric (EM) columns in Tables 1 and 2. At the same time, other compression methods failed to retain key semantic information in prompts, especially in logical reasoning details. For a more in-depth discussion of these results, refer to section 5.2 of the paper.

These are the experimental results on GSM8K and BBH using GPT-3.5-turbo, demonstrating the in-context learning and reasoning capabilities based on different methods and compression constraints. The results show that LLMLingua can achieve up to a 20x compression rate while only experiencing a 1.5-point performance loss.
Table 1. Performance of different methods at different target compression ratios on the GSM8K and BBH datasets.
These are the experimental results for ShareGPT (Conversation) and Arxiv-March23 (Summarization) using GPT-3.5-turbo, based on different methods and compression constraints. The results indicate that LLMLingua can effectively retain the semantic information from the original prompts while achieving a compression rate of 3x-9x.
Table 2. Performance of different methods at different target compression ratios for conversation and summarization tasks.

LLMLingua is robust, cost-effective, efficient, and recoverable

LLMLingua also showed impressive results across various small language models and different closed LLMs. When using GPT-2-small, LLMLingua achieved a strong performance score of 76.27 under the ¼-shot constraint, close to the LLaMA-7B’s result of 77.33 and surpassing the standard prompt results of 74.9. Similarly, even without aligning Claude-v1.3, one of the post powerful LLMs, LLMLingua’s score was 82.61 under the ½-shot constraint, outperforming the standard prompt result of 81.8.

LLMLingua also proved effective in reducing response length, leading to significant reductions in latency in the LLM’s generation process, with reductions ranging between 20 to 30 percent, as shown in Figure 2.

The figure demonstrates the relationship between the compression ratio and the number of response tokens. In different tasks, as the compression ratio increases, the response length decreases to varying extents, with a maximum reduction of 20%-30%.
Figure 2. The distribution of token lengths generated at varying compression ratios.

What makes LLMLingua even more impressive is its recoverability feature. When we used GPT-4 to restore the compressed prompts, it successfully recovered all key reasoning information from the full nine-step chain-of-thought (CoT) prompting, which enables LLMs to address problems through sequential intermediate steps. The recovered prompt was almost identical to the original, and its meaning was retained. This is shown in Tables 3 and 4.

This figure illustrates the original prompt, the compressed prompt, and the result of using GPT-4 to recover the compressed prompt. The original prompt consists of a 9-step Chain-of-Thought, and the compressed prompt is difficult for humans to understand. However, the recovered text includes all 9 steps of the Chain-of-Thought.
Table 3. Latency comparison on GSM8K. LLMLingua can accelerate LLMs’ end-to-end inference by a factor of 1.7–5.7x. 
This figure shows the end-to-end latency when using LLMLingua, without using LLMLingua, and the latency when compressing prompts. As the compression ratio increases, both the LLMLingua and end-to-end latency decrease, achieving up to a 5.7x acceleration with a 10x token compression rate.
Table 4. Recovering the compressed prompt from GSM8K using GPT-4.

Enhancing the user experience and looking ahead

LLMLingua is already proving its value through practical application. It has been integrated into LlamaIndex (opens in new tab), a widely adopted retrieval-augmented generation (RAG) framework. Currently, we are collaborating with product teams to reduce the number of tokens required in LLM calls, particularly for tasks like multi-document question-answering. Here, our goal is to significantly improve the user experience with LLMs. 

For the long-term, we have proposed LongLLMLingua, a prompt-compression technique designed for long-context scenarios, such as retrieval-augmented question-answering tasks in applications like chatbots, useful when information evolves dynamically over time. It’s also geared for tasks like summarizing online meetings. LongLLMLingua’s primary objective is to enhance LLMs’ ability to perceive key information, making it suitable for numerous real-world applications, notably information-based chatbots. We’re hopeful that this innovation paves the way for more sophisticated and user-friendly interactions with LLMs.

Learn more about our work on the LLMLingua (opens in new tab) page.

The post LLMLingua: Innovating LLM efficiency with prompt compression appeared first on Microsoft Research.

Read More

Getir end-to-end workforce management: Amazon Forecast and AWS Step Functions

Getir end-to-end workforce management: Amazon Forecast and AWS Step Functions

This is a guest post co-authored by Nafi Ahmet Turgut, Mehmet İkbal Özmen, Hasan Burak Yel, Fatma Nur Dumlupınar Keşir, Mutlu Polatcan and Emre Uzel from Getir.

Getir is the pioneer of ultrafast grocery delivery. The technology company has revolutionized last-mile delivery with its grocery in-minutes delivery proposition. Getir was founded in 2015 and operates in Turkey, the UK, the Netherlands, Germany, and the United States. Today, Getir is a conglomerate incorporating nine verticals under the same brand.

In this post, we describe the end-to-end workforce management system that begins with location-specific demand forecast, followed by courier workforce planning and shift assignment using Amazon Forecast and AWS Step Functions.

In the past, operational teams engaged in manual workforce management practices, which resulted in a significant waste of time and effort. However, with the implementation of our comprehensive end-to-end workforce management project, they are now able to efficiently generate the necessary courier plans for warehouses through a simplified, one-click process accessible via a web interface. Before the initiation of this project, business teams relied on more intuitive methods for demand forecasting, which required improvement in terms of precision.

Amazon Forecast is a fully managed service that uses machine learning (ML) algorithms to deliver highly accurate time series forecasts. In this post, we describe how we reduced the modelling time by 70% by doing the feature engineering and modelling using Amazon Forecast. We achieved a 90% reduction in elapsed time when running scheduling algorithms for all warehouses using AWS Step Functions, which is a fully managed service that makes it easier to coordinate the components of distributed applications and microservices using visual workflows. This solution also led to an 90% improvement in prediction accuracy across Turkey and several European countries.

Solution overview

The End-to-end Workforce Management Project (E2E Project) is a large-scale project and it can be described in three topics:

1. Calculating courier requirements

The first step is to estimate hourly demand for each warehouse, as explained in the Algorithm selection section. These predictions, produced with Amazon Forecast, help determine when and how many couriers each warehouse needs.

Based on the throughput ratio of the couriers in warehouses, the number of couriers required for each warehouse is calculated in hourly intervals. These calculations assist in determining the feasible courier counts considering legal working hours, which involves mathematical modeling.

2. Solving the shift Assignment problem

Once we have the courier needs and know the other constraints of the couriers and warehouses, we can solve the shift assignment problem. The problem is modelled with decision variables determining the couriers to be assigned and creating shift schedules, minimizing surplus and shortage that may cause missed orders. This is typically a mixed-integer programming (MIP) problem.

3. Utilizing AWS Step Functions

We use AWS Step Functions to coordinate and manage workflows with its capability to execute jobs in parallel. Each warehouse’s shift assignment process is defined as a separate workflow. AWS Step Functions automatically initiate and monitor these workflows by simplifying error handling.

Since this process requires extensive data and complex computations, services like AWS Step Functions offer a significant advantage in organizing and optimizing tasks. It allows for better control and efficient resource management.

In the solution architecture, we also take advantage of other AWS services by integrating them into AWS Step Functions:

The following diagrams show AWS Step Functions workflows and architecture of the shifting tool:

Figure 1 AWS Step Functions workflows

Figure 2 Shifting tool architecture

Algorithm selection

Forecasting locational demand constitutes the initial phase in the E2E project. The overarching goal of E2E is to determine the number of couriers to allocate to a specific warehouse, commencing with a forecast of the demand for that warehouse.

This forecasting component is pivotal within the E2E framework, as subsequent phases rely on these forecasting outcomes. Thus, any prediction inaccuracies can detrimentally impact the entire project’s efficacy.

The objective of the locational demand forecast phase is to generate predictions on a country-specific basis for every warehouse segmented hourly over the forthcoming two weeks. Initially, daily forecasts for each country are formulated through ML models. These daily predictions are subsequently broken down into hourly segments, as depicted in the following graph. Historic transactional demand data, location-based weather information, holiday dates, promotions and marketing campaign data are the features used in the model as shown in the graph below.

Figure 3 The architecture of location-specific forecasting

The team initially explored traditional forecasting techniques such as open-source SARIMA (Seasonal Auto-Regressive Integrated Moving Average), ARIMAX (Auto-Regressive Integrated Moving Average using exogenous variables), and Exponential Smoothing.

ARIMA (Auto-Regressive Integrated Moving Average) is a time series forecasting method that combines autoregressive (AR) and moving average (MA) components along with differencing to make the time series stationary.

SARIMA extends ARIMA by incorporating additional parameters to account for seasonality in the time series. It includes seasonal auto-regressive and seasonal moving average terms to capture repeating patterns over specific intervals, making it suitable for time series with a seasonal component.

ARIMAX builds upon ARIMA by introducing exogenous variables, which are external factors that can influence the time series. These additional variables are considered in the model to improve forecasting accuracy by accounting for external influences beyond the historical values of the time series.

Exponential Smoothing is another time series forecasting method that, unlike ARIMA, is based on weighted averages of past observations. It is particularly effective for capturing trends and seasonality in data. The method assigns exponentially decreasing weights to past observations, with more recent observations receiving higher weights.

The Amazon Forecast models were eventually selected for the algorithmic modeling segment. The vast array of models and the sophisticated feature engineering capabilities offered by AWS Forecast proved more advantageous and optimized our resource utilization.

Six algorithms available in Forecast were tested: Convolutional Neural Network – Quantile Regression (CNN-QR), DeepAR+, Prophet, Non-Parametric Time Series (NPTS), Autoregressive Integrated Moving Average (ARIMA), and Exponential Smoothing (ETS). Upon analysis of the forecast results, we determined that CNN-QR surpassed the others in efficacy. CNN-QR is a proprietary ML algorithm developed by Amazon for forecasting scalar (one-dimensional) time series using causal Convolutional Neural Networks (CNNs). Given the availability of diverse data sources at this juncture, employing the CNN-QR algorithm facilitated the integration of various features, operating within a supervised learning framework. This distinction separated it from univariate time-series forecasting models and markedly enhanced performance.

Utilizing Forecast proved effective due to the simplicity of providing the requisite data and specifying the forecast duration. Subsequently, Forecast employs the CNN-QR algorithm to generate predictions. This tool significantly expedited the process for our team, particularly in algorithmic modeling. Furthermore, utilizing Amazon Simple Storage Service (Amazon S3) buckets for input data repositories and Amazon Redshift for storing outcomes has facilitated centralized management of the entire procedure.

Conclusion

In this post, we showed you how Getir’s E2E project demonstrated how combining Amazon Forecast and AWS Step Functions services streamlines complex processes effectively. We achieved an impressive prediction accuracy of around 90% across countries in Europe and Turkey, and using Forecast reduced modeling time by 70% due to its efficient handling of feature engineering and modeling.

Using AWS Step Functions service has led to practical advantages, notably reducing scheduling time by 90% for all warehouses. Also, by considering field requirements, we improved compliance rates by 3%, helping allocate the workforce more efficiently. This, in turn, highlights the project’s success in optimizing operations and service delivery.

To access further details on commencing your journey with Forecast, please refer to the available Amazon Forecast resources. Additionally, for insights on constructing automated workflows and crafting machine learning pipelines, you can explore AWS Step Functions for comprehensive guidance.


About the Authors

Nafi Ahmet Turgut finished his master’s degree in electrical & Electronics Engineering and worked as graduate research scientist. His focus was building machine learning algorithms to simulate nervous network anomalies. He joined Getir in 2019 and currently works as a Senior Data Science & Analytics Manager. His team is responsible for designing, implementing, and maintaining end-to-end machine learning algorithms and data-driven solutions for Getir.

Mehmet İkbal Özmen received his Master’s Degree in Economics and worked as Graduate Research Assistant. His research area was mainly economic time series models, Markov simulations, and recession forecasting. He then joined Getir in 2019 and currently works as Data Science & Analytics Manager. His team is responsible for optimization and forecast algorithms to solve the complex problems experienced by the operation and supply chain businesses.

Hasan Burak Yel received his Bachelor’s Degree in Electrical & Electronics Engineering at Boğaziçi University. He worked at Turkcell, mainly focused on time series forecasting, data visualization, and network automation. He joined Getir in 2021 and currently works as a Data Science & Analytics Manager with the responsibility of Search, Recommendation, and Growth domains.

Fatma Nur Dumlupınar Keşir received her Bachelor’s Degree from Industrial Engineering Department at Boğaziçi University. She worked as a researcher at TUBITAK, focusing on time series forecasting & visualization. She then joined Getir in 2022 as a data scientist and has worked on Recommendation Engine projects, Mathematical Programming for Workforce Planning.

Emre Uzel received his Master’s Degree in Data Science from Koç University. He worked as a data science consultant at Eczacıbaşı Bilişim where he mainly focused on recommendation engine algorithms. He joined Getir in 2022 as a Data Scientist and started working on time-series forecasting and mathematical optimization projects.

Mutlu Polatcan is a Staff Data Engineer at Getir, specializing in designing and building cloud-native data platforms. He loves combining open-source projects with cloud services.

Esra Kayabalı is a Senior Solutions Architect at AWS, specializing in the analytics domain including data warehousing, data lakes, big data analytics, batch and real-time data streaming and data integration. She has 12 years of software development and architecture experience. She is passionate about learning and teaching cloud technologies.

Read More

Meet the Omnivore: SiBORG Lab Elevates Approach to Accessibility Using OpenUSD and NVIDIA Omniverse

Meet the Omnivore: SiBORG Lab Elevates Approach to Accessibility Using OpenUSD and NVIDIA Omniverse

Accessibility is a key element that all designers must consider before constructing a space or product — but the evaluation process has traditionally been tedious and time-consuming.

Mathew Schwartz, an assistant professor in architecture and design at the New Jersey Institute of Technology, is using the NVIDIA Omniverse platform and the Universal Scene Description framework, aka OpenUSD, to help architects, interior designers and industrial designers address this challenge.

Schwartz’s research and design lab SiBORG — which stands for simulation, biomechanics, robotics and graphics — focuses on understanding and improving design workflows, especially in relation to accessibility, human factors and automation. Schwartz and his team develop algorithms for research projects and turn them into usable products.

Using Omniverse  — a development platform that enables multi-app workflows and real-time collaboration — the team developed open-source, OpenUSD-based code that automatically generates a complex accessibility graph for building design. This code is based on Schwartz’s research paper, “Human centric accessibility graph for environment analysis.”

The graph provides feedback related to human movement, such as the estimated energy expenditure required for taking a certain path, the number of steps it takes to complete the path, or the angles of any inclines along it.

With Omniverse, teams can use Schwartz’s code to visualize the graph and the paths that it creates. This can help designers better evaluate building code and safety for occupants while providing important accessibility insights.


The Power of OpenUSD

Traditionally, feedback on accessibility and environmental conditions during the building design process has been limited to building code analysis. Schwartz’s work enables designers to overcome this obstacle by seamlessly integrating Omniverse and OpenUSD.

Previously, he had to switch between multiple applications to achieve different aspects of his simulation and modeling projects. His workflows were often split between tools such as Unity, which supports simulations with people, and McNeel Rhino3D, which offers 3D modeling features.

With OpenUSD, he can now combine his research, Python code, 3D environments and renders, and favorite tools into Omniverse.

“What got me hooked on Omniverse was how it allows me to combine the Python application programming interface with powerful physics, rendering and animation software,” he said. “My team took full advantage of the flexible Python APIs in Omniverse to develop almost the entire user interface.”

Schwartz’s team uses Omniverse to visualize and interact with existing open-source Python code in ways that don’t require external work, like seamlessly linking to a third-party app. The lab’s versatile data analysis tool can interact with any program that’s compatible with OpenUSD.

“With OpenUSD and Omniverse, we’ve been able to expand the scope of our research, as we can easily combine data analysis and visualization with the design process,” said Schwartz.

Running Realistic Renderings and Simulations

Schwartz also uses Omniverse to simulate crowd movement and interactions.

He accelerates large crowd simulations and animations using two NVIDIA RTX A4500 GPUs, which enable real-time visualization. These accelerated simulations can help designers gain valuable insights into how people with reduced mobility can navigate and interact in spaces.

“We can also show what locations will offer the best areas to place signage so that it’s most visible,” said Schwartz. “Our simulation work can be used to visualize paths taken in an early-stage design — this provides feedback on accessibility to prevent problems with building code, while allowing users to create designs that go beyond the minimum requirements.”

Schwartz also taps the feedback and assistance of many developers and researchers who actively engage on the Omniverse Discord channel. This collaborative environment has been instrumental in Schwartz’s journey, he said, as well as to the platform’s continuous improvement.

Schwartz’s open-source code is available for designers to use and enhance their design workflows. Learn more about his work and how NVIDIA Omniverse can revolutionize building design.

Join In on the Creation

Anyone can build their own Omniverse extension or Connector to enhance 3D workflows and tools.

Check out artwork from other “Omnivores” and submit projects in the Omniverse gallery. See how creators are using OpenUSD to accelerate a variety of 3D workflows in the latest OpenUSD All Stars.

Get started with NVIDIA Omniverse by downloading the standard license free, access OpenUSD resources, and learn how Omniverse Enterprise can connect your team. Stay up to date on Instagram, Medium and Twitter. For more, join the Omniverse community on the  forums, Discord server, Twitch and YouTube channels.

Read More

Good Fortunes: ‘The Day Before’ Leads 17 Games on GeForce NOW

Good Fortunes: ‘The Day Before’ Leads 17 Games on GeForce NOW

It’s a fortuitous GFN Thursday with 17 new games joining the GeForce NOW library, including The Day Before, Avatar: Frontiers of Pandora and the 100th PC Game Pass title to join the cloud Ori and the Will of the Wisps.

This week also marks a milestone: over 500 games and applications now support RTX ON. GeForce NOW Ultimate and Priority members can experience cinematic ray tracing on nearly any device thanks to NVIDIA RTX-powered gaming rigs in the cloud. Check out the RTX ON game row in the GeForce NOW app to play even more titles featuring this stunning graphics technology.

Stayin’ Alive

The Day Before on GeForce NOW
Take a trip to the big city and say hi to the locals.

The Day Before, Fntastic’s new open-world horror massively multiplayer online game, is a uniquely reimagined journey of survival set on the east coast of the present-day U.S. after the world has been overrun by zombies. Priority and Ultimate members can stream the game on nearly any device with support for RTX ON.

Explore the beautifully detailed New Fortune City, filled with skyscrapers, massive malls and grand stadiums, with a variety of vehicles. Fight against other players and those infected by a deadly virus. Survive by collecting loot, completing quests and building houses — all on a day-and-night cycle.

Help rebuild society from the comfort of the couch and across devices, streaming from the cloud. Priority members can build and survive at up to 1080p and 60 frames per second. Ultimate members can take advantage of longer session lengths, gain support for ultrawide resolutions and stream at up to 4K 120 fps. Both memberships offer support for real-time ray tracing, bringing cinematic lighting to every zombie encounter.

A New Adventure in the Clouds

You can fly, you can fly.

Fight for the future of the Na’vi in Ubisoft’s Avatar: Frontiers of Pandora. Expanding on the stories told in the hit Avatar films, the open-world action-adventure game explores a never-before-seen region of Pandora called the Western Frontier, with all-new environments, creatures and enemies.

Discover what it means to be Na’vi and join other clans to protect Pandora from the RDA, a corporation looking to exploit Pandora’s resources. Harness incredible strength and agility with character customization, craft new gear, and upgrade skills and weapons. Members can enjoy soaring across the skies with their Banshees, dragon-like creatures useful for exploring the vast Western Frontier and engaging the RDA in aerial combat.

Experience the epic adventure on nearly any device with a GeForce NOW Ultimate membership, streaming from GeForce RTX 4080-powered servers in the cloud. Ultimate members can save the Na’vi at up to 4K resolution or take in Pandora’s beautiful vistas at ultrawide resolution for the most cinematic, immersive gameplay.

A Journey of Courage

Ori and the Will of the Wisps on GeForce NOW
Ori-nge you glad you streamed it from the cloud?

Ori and the Blind Forest: Definitive Edition and Ori and the Will of the Wisps are the newest Xbox PC games to join GeForce NOW, which now includes 100 PC Game Pass titles. Developed by Moon Studios and published by Xbox Game Studios, the award-winning adventure series follows a spirit guardian named Ori as he explores beautiful, dangerous worlds.

In Ori and the Blind Forest, members must help Ori restore balance to the forest. Separated from his home during a storm and adopted by a creature called Naru, Ori must team up with a spirit named Sein to find his true destiny when calamity strikes the world of Nibel.

The sequel, Ori and the Will of the Wisps, brings Ori’s journey to a new world, Niwen, a hidden land of wonders and dangers. Ori must help a young, broken-winged owl named Ku and heal the land from dark corruption — all while encountering new friends, foes and mysteries that will test the spirit guardian’s courage and skills.

Members can take the adventure with them, streaming Ori’s adventures across nearly all of their devices thanks to the cloud. GeForce NOW Ultimate members can also light their journeys with high dynamic range on supported devices for an unparalleled visual experience.

Explore the Ori series and other games on GeForce NOW with PC Game Pass. Give the gift of cloud gaming with the latest membership bundle and get three months of PC Game Pass for free with the purchase of a six-month GeForce NOW Ultimate membership.

Building Blocks

LEGO Fortnite on GeForce NOW
Blast off to the cloud.

The magic of LEGOs and Fortnite collide in Epic Games’ LEGO Fortnite, launching in the cloud today. ​Get creative in building and ​customizing​​​ the ultimate home base​ using collected LEGO elements. Recruit villagers to gather materials and survive the night. Gear up and drop into deep caves to search for rare resources in hidden areas.

Don’t miss the 17 newly supported games joining the GeForce NOW library this week:

  • World War Z: Aftermath (New release on Xbox, available on PC Game Pass, Dec. 5)
  • Avatar: Frontiers of Pandora (New release on Ubisoft, Dec. 7)
  • Warhammer 40,000: Rogue Trader (New release on Steam, Dec. 7)
  • The Day Before (New release on Steam, Dec. 7)
  • Goat Simulator 3 (New release on Xbox, available on PC Game Pass, Dec. 7)
  • LEGO Fortnite (New release on Epic Games Store, Dec. 7)
  • Against the Storm (New release on Xbox, available on PC Game Pass, Dec. 8)
  • Rocket Racing (New release on Epic Games Store, Dec. 8)
  • Fortnite Festival (New release on Epic Games Store, Dec. 9)
  • Agatha Christie – Murder on the Orient Express (Steam)
  • BEAST (Steam)
  • Dungeons 4  (Xbox, available on PC Game Pass)
  • Farming Simulator 22 (Xbox, available on PC Game Pass)
  • Hollow Knight (Xbox, available on PC Game Pass)
  • Ori and the Will of the Wisps (Steam, Xbox and available on PC Game Pass)
  • Ori and the Blind Forest: Definitive Edition (Steam)
  • Spirittea (Xbox, available on PC Game Pass)

Halo Infinite was planned to join the cloud in September but encountered some technical issues. The GeForce NOW team is working with Microsoft and game developer 343 Industries to bring the game to the service in the coming months. Stay tuned to GFN Thursday for further updates.

What are you planning to play this weekend? Let us know on Twitter or in the comments below. Bonus points if it includes #RTXON.

Read More

HUGS: Human Gaussian Splats

Recent advances in neural rendering have improved both training and rendering times by orders of magnitude. While these methods demonstrate state-of-the-art quality and speed, they are designed for photogrammetry of static scenes and do not generalize well to freely moving humans in the environment. In this work, we introduce Human Gaussian Splats (HUGS) that represents an animatable human together with the scene using 3D Gaussian Splatting (3DGS). Our method takes only a monocular video with a small number of (50-100) frames, and it automatically learns to disentangle the static scene and a…Apple Machine Learning Research

DeepPCR: Parallelizing Sequential Operations in Neural Networks

Parallelization techniques have become ubiquitous for accelerating inference and training of deep neural networks. Despite this, several operations are still performed in a sequential manner. For instance, the forward and backward passes are executed layer-by-layer, and the output of diffusion models is produced by applying a sequence of denoising steps. This sequential approach results in a computational cost proportional to the number of steps involved, presenting a potential bottleneck as the number of steps increases. In this work, we introduce DeepPCR, a novel algorithm which parallelizes…Apple Machine Learning Research