NVIDIA Brings Generative AI to Millions, With Tensor Core GPUs, LLMs, Tools for RTX PCs and Workstations

Leading AI Platform Gets RTX-Accelerated Boost From New GeForce RTX SUPER GPUs, AI Laptops From Every Top Manufacturer

CES — NVIDIA today announced GeForce RTX SUPER desktop GPUs for supercharged generative AI performance, new AI laptops from every top manufacturer, and new NVIDIA RTX-accelerated AI software and tools for both developers and consumers.

Building on decades of PC leadership, with over 100 million of its RTX GPUs driving the AI ​​PC era, NVIDIA is now offering these tools to enhance PC experiences with generative AI: NVIDIA TensorRT acceleration of the popular Stable Diffusion XL model for text-to-image workflows, NVIDIA RTX Remix with generative AI texture tools, NVIDIA ACE microservices and more games that use DLSS 3 technology with Frame Generation.

AI Workbench, a unified, easy-to-use toolkit for AI developers, will be available in beta later this month. In addition, NVIDIA TensorRT-LLM (TRT-LLM), an open-source library that accelerates and optimizes inference performance of the latest large language models (LLMs), now supports more pre-optimized models for PCs. Accelerated by TRT-LLM, Chat with RTXan NVIDIA tech demo also releasing this month, allows AI enthusiasts to interact with their notes, documents and other content.

“Generative AI is the single most significant platform transition in computing history and will transform every industry, including gaming,” said Jensen Huang, founder and CEO of NVIDIA. “With over 100 million RTX AI PCs and workstations, NVIDIA is a massive installed base for developers and gamers to enjoy the magic of generative AI.”

Running generative AI locally on a PC is critical for privacy, latency and cost-sensitive applications. It requires a large installed base of AI-ready systems, as well as the right developer tools to tune and optimize AI models for the PC platform.

To meet these needs, NVIDIA is delivering innovations across its full technology stack, driving new experiences and building on the 500+ AI-enabled PC applications and games already accelerated by NVIDIA RTX technology.

RTX AI PCs and Workstations
NVIDIA RTX GPUs — capable of running a broad range of applications at the highest performance — unlock the full potential of generative AI on PCs. Tensor Cores in these GPUs dramatically speed AI performance across the most demanding applications for work and play.

The new GeForce RTX 40 SUPER Series graphics cards, also announced today at CES, include the GeForce RTX 4080 SUPER, 4070 Ti SUPER and 4070 SUPER for top AI performance. The GeForce RTX 4080 SUPER generates AI videos 1.5x faster — and images 1.7x faster — than the GeForce RTX 3080 Ti GPU. The Tensor Cores in SUPER GPUs deliver up to 836 trillion operations per second, bringing transformative AI capabilities to gaming, creating and everyday productivity.

Leading manufacturers — including Acer, ASUS, Dell, HP, Lenovo, MSI, Razer and Samsung — are releasing a new wave of RTX AI laptops, bringing a full set of generative AI capabilities to users right out of the box. The new systems, which deliver a performance increase ranging from 20x-60x compared with using neural processing units, will start shipping this month.

Mobile workstations with RTX GPUs can run NVIDIA AI Enterprise software, including TensorRT and NVIDIA RAPIDS for simplified, secure generative AI and data science development. A three-year license for NVIDIA AI Enterprise is included with every NVIDIA A800 40GB Active GPUproviding an ideal workstation development platform for AI and data science.

New PC Developer Tools for Building AI Models
To help developers quickly create, test and customize pretrained generative AI models and LLMs using PC-class performance and memory footprint, NVIDIA recently announced NVIDIA AI Workbench.

AI Workbench, which will be available in beta later this month, offers streamlined access to popular repositories like Hugging Face, GitHub and NVIDIA NGC, along with a simplified user interface that enables developers to easily reproduce, collaborate on and migrate projects.

Projects can be scaled out to virtually anywhere — whether the data center, a public cloud or NVIDIA DGX Cloud — and then brought back to local RTX systems on a PC or workstation for inference and light customization.

In collaboration with HP, NVIDIA is also simplifying AI model development by integrating NVIDIA AI Foundation Models and Endpointswhich includes RTX-accelerated AI models and software development kits, into the HP AI Studio, a centralized platform for data science. This will allow users to easily search, import and deploy optimized models across PCs and the cloud.

After building AI models for PC use cases, developers can optimize them using NVIDIA TensorRT to take full advantage of RTX GPUs’ Tensor Cores.

NVIDIA recently extended TensorRT to text-based applications with TensorRT-LLM for Windows, an open-source library for accelerating LLMs. The latest update to TensorRT-LLM, available now, adds Phi-2 to the growing list of pre-optimized models for PC, which run up to 5x faster compared to other inference backends.

RTX-Accelerated Generative AI Powers New PC Experiences
At CES, NVIDIA and its developer partners are releasing new generative AI-powered applications and services for PCs, including:

  • NVIDIA RTX Remixa platform for creating stunning RTX remasters of classic games. Releasing in beta later this month, it delivers generative AI tools that can transform basic textures from classic games into modern, 4K-resolution, physically based rendering materials.
  • NVIDIA ACE microservices, including generative AI-powered speech and animation models, which enable developers to add intelligent, dynamic digital avatars to games.
  • TensorRT acceleration for Stable Diffusion XL (SDXL) Turbo and latent consistency models, two of the most popular Stable Diffusion acceleration methods. TensorRT improves performance for both by up to 60% compared with the previous fastest implementation. An updated version of the Stable Diffusion WebUI TensorRT extensions are also now available, including acceleration for SDXL, SDXL Turbo, LCM – Low-Rank Adaptation (LoRA) and improved LoRA support.
  • NVIDIA DLSS 3 with Frame Generation, which uses AI to increase frame rates up to 4x compared with native rendering, will be featured in a dozen of the 14 new RTX games announced, including Horizon Forbidden West, Pax Dee and Dragon’s Dogma 2.
  • Chat with RTX, an NVIDIA tech demo available later this month, allows AI enthusiasts to easily connect PC LLMs to their own data using a popular technique known as retrieval-augmented generation (RAG). The demo, accelerated by TensorRT-LLM, enables users to quickly interact with their notes, documents and other content. It will also be available as an open-source reference project, so developers can easily implement the same capabilities in their own applications.

Learn more about the latest generative AI breakthroughs by joining NVIDIA at CES.

General Motors, Magna, Wipro team up for automotive software marketplace | Company News

General Motors, auto parts supplier Magna and IT company Wipro said on Tuesday they were working together to create a sales platform to buy and sell automotive software.

The joint venture, SDVerse, will link the buyers and sellers through a digital platform, where the software’s features and attributes can be listed.

Wipro, in a regulatory filing, disclosed an investment of $5.85 million, equating to a 27% stake in SDVerse. GM and Magna will hold a 46% and 27% stake, respectively.

The transaction is expected to be completed before the end of March.

The platform is in its development stage and is expected to feature hundreds of automotive software products and participants across the industry.

The announcement comes at a time when automakers are ramping up their tech investments to help create connected vehicles with advanced driver aids.

“The market for automotive software is expected to nearly double this decade, potentially outpacing the growth of software development talent pools,” said Harmeet Chauhan, global head Wipro Engineering Edge, Wipro Ltd.

While the company did not detail specific revenue targets, SDVerse will follow an annual subscription fee model and not charge any fees for buying or selling products.

GM, Magna and Wipro will hold seats on the board of SDVerse but the platforms will operate independently.

First Published: Mar 05 2024 | 11:43 PM IST

World’s first AI software engineer Devin announced, it can write, code, create using a single prompt

There is a new AI tool so smart that it can write code, create websites, and software with just a single prompt. Devin, created by the tech company Cognition, is the first AI software engineer. It can do pretty much everything you ask it to do. And the AI ​​tool does not come with the intention to replace human engineers, it is designed to work hand-in-hand with them. The makers say that the AI ​​tool has not been launched to replace human engineers but to make their lives easier

“Today we’re excited to introduce Devin, the first AI software engineer. Devin is the new state-of-the-art on the SWE-Bench coding benchmark, has successfully passed practical engineering interviews from leading AI companies, and has even completed real jobs on Upwork. Devin is an autonomous agent that solves engineering tasks through the use of its own shell, code editor, and web browser,” Cognition posted on Twitter aka X.

What makes Devin stand out is his incredible ability to think ahead and plan complex tasks. It can make thousands of decisions, learn from its mistakes, and get better over time. Plus, it has all the tools a human engineer needs, like a code editor and browser, right at its digital fingertips. Devin is considered the most advanced or cutting-edge solution available for evaluating software engineering tasks based on the SWE-bench coding benchmark. Essentially, it performed exceptionally well compared to other solutions when tested against a standard set of software engineering problems. The AI ​​tool performed well in practical engineering interviews conducted by top artificial intelligence companies. These interviews likely involved tasks and challenges relevant to the field of AI and software engineering, and the AI ​​assistant managed to meet expectations.

But Devin isn’t just a solo act. It’s designed to work hand-in-hand with human engineers, providing real-time updates, accepting feedback, and collaborating on design choices. So, rather than replacing humans, Devin complements their skills, making teams more productive and efficient.

So, what exactly can Devin do? Well, pretty much anything you throw at it. Whether it’s learning new technologies, building and deploying apps from start to finish, or hunting down and fixing pesky bugs in code, Devin’s got it covered. It can even train its own AI models and tackle issues in open-source projects.

And Devin’s not just talk – it’s backed by impressive results. Tested on real-world challenges, Devin outperformed previous AI models by a wide margin, resolving nearly 14 per cent of issues compared to just under 2 percent for its predecessors. That’s a game-changer in the world of software engineering.

But perhaps the most exciting part is that Devin isn’t limited to lab tests. It’s already been put to work on platforms like Upwork, where it tackled real-world coding tasks with ease, from debugging computer vision models to compiling detailed reports.

Devin represents a major leap forward in AI technology. By automating routine tasks and empowering engineers to focus on more complex problems, it’s paving the way for a new era of innovation in software development. So, whether you’re a seasoned engineer or just starting out, Devin is here to make your job easier and more exciting than ever before.

Published By:

Ankita Chakravarti

Published On:

March 13, 2024