NVIDIA
NVIDIA
This is the Official NVIDIA Page
204 Les gens qui ont lié ça
536 Articles
2 Photos
0 Vidéos
Mises à jour récentes
  • Hot Topics at Hot Chips: Inference, Networking, AI Innovation at Every Scale All Built on NVIDIA
    blogs.nvidia.com
    AI reasoning, inference and networking will be top of mind for attendees of next weeks Hot Chips conference.A key forum for processor and system architects from industry and academia, Hot Chips running Aug. 24-26 at Stanford University showcases the latest innovations poised to advance AI factories and drive revenue for the trillion-dollar data center computing market.At the conference, NVIDIA will join industry leaders including Google and Microsoft in a tutorial session taking place on Sunday, Aug. 24 that discusses designing rack-scale architecture for data centers.In addition, NVIDIA experts will present at four sessions and one tutorial detailing how:NVIDIA networking, including the NVIDIA ConnectX-8 SuperNIC, delivers AI reasoning at rack- and data-center scale. (Featuring Idan Burstein, principal architect of network adapters and systems-on-a-chip at NVIDIA)Neural rendering advancements and massive leaps in inference powered by the NVIDIA Blackwell architecture, including the NVIDIA GeForce RTX 5090 GPU provide next-level graphics and simulation capabilities. (Featuring Marc Blackstein, senior director of architecture at NVIDIA)Co-packaged optics (CPO) switches with integrated silicon photonics built with light-speed fiber rather than copper wiring to send information quicker and using less power enable efficient, high-performance, gigawatt-scale AI factories. The talk will also highlight NVIDIA Spectrum-XGS Ethernet, a new scale-across technology for unifying distributed data centers into AI super-factories. (Featuring Gilad Shainer, senior vice president of networking at NVIDIA)The NVIDIA GB10 Superchip serves as the engine within the NVIDIA DGX Spark desktop supercomputer. (Featuring Andi Skende, senior distinguished engineer at NVIDIA)Its all part of how NVIDIAs latest technologies are accelerating inference to drive AI innovation everywhere, at every scale.NVIDIA Networking Fosters AI Innovation at ScaleAI reasoning when artificial intelligence systems can analyze and solve complex problems through multiple AI inference passes requires rack-scale performance to deliver optimal user experiences efficiently.In data centers powering todays AI workloads, networking acts as the central nervous system, connecting all the components servers, storage devices and other hardware into a single, cohesive, powerful computing unit.NVIDIA ConnectX-8 SuperNICBursteins Hot Chips session will dive into how NVIDIA networking technologies particularly NVIDIA ConnectX-8 SuperNICs enable high-speed, low-latency, multi-GPU communication to deliver market-leading AI reasoning performance at scale.As part of the NVIDIA networking platform, NVIDIA NVLink, NVLink Switch and NVLink Fusion deliver scale-up connectivity linking GPUs and compute elements within and across servers for ultra low-latency, high-bandwidth data exchange.NVIDIA Spectrum-X Ethernet provides the scale-out fabric to connect entire clusters, rapidly streaming massive datasets into AI models and orchestrating GPU-to-GPU communication across the data center. Spectrum-XGS Ethernet scale-across technology extends the extreme performance and scale of Spectrum-X Ethernet to interconnect multiple, distributed data centers to form AI super-factories capable of giga-scale intelligence.Connecting distributed AI data centers with NVIDIA Spectrum-XGS Ethernet.At the heart of Spectrum-X Ethernet, CPO switches push the limits of performance and efficiency for AI infrastructure at scale, and will be covered in detail by Shainer in his talk.NVIDIA GB200 NVL72 an exascale computer in a single rack features 36 NVIDIA GB200 Superchips, each containing two NVIDIA B200 GPUs and an NVIDIA Grace CPU, interconnected by the largest NVLink domain ever offered, with NVLink Switch providing 130 terabytes per second of low-latency GPU communications for AI and high-performance computing workloads.An NVIDIA rack-scale system.Built with the NVIDIA Blackwell architecture, GB200 NVL72 systems deliver massive leaps in reasoning inference performance.NVIDIA Blackwell and CUDA Bring AI to Millions of DevelopersThe NVIDIA GeForce RTX 5090 GPU also powered by Blackwell and to be covered in Blacksteins talk doubles performance in todays games with NVIDIA DLSS 4 technology.NVIDIA GeForce RTX 5090 GPUIt can also add neural rendering features for games to deliver up to 10x performance, 10x footprint amplification and a 10x reduction in design cycles, helping enhance realism in computer graphics and simulation. This offers smooth, responsive visual experiences at low energy consumption and improves the lifelike simulation of characters and effects.NVIDIA CUDA, the worlds most widely available computing infrastructure, lets users deploy and run AI models using NVIDIA Blackwell anywhere.Hundreds of millions of GPUs run CUDA across the globe, from NVIDIA GB200 NVL72 rack-scale systems to GeForce RTX and NVIDIA RTX PRO-powered PCs and workstations, with NVIDIA DGX Spark powered by NVIDIA GB10 discussed in Skendes session coming soon.From Algorithms to AI Supercomputers Optimized for LLMsNVIDIA DGX SparkDelivering powerful performance and capabilities in a compact package, DGX Spark lets developers, researchers, data scientists and students push the boundaries of generative AI right at their desktops, and accelerate workloads across industries.As part of the NVIDIA Blackwell platform, DGX Spark brings support for NVFP4, a low-precision numerical format to enable efficient agentic AI inference, particularly of large language models (LLMs). Learn more about NVFP4 in this NVIDIA Technical Blog.Open-Source Collaborations Propel Inference InnovationNVIDIA accelerates several open-source libraries and frameworks to accelerate and optimize AI workloads for LLMs and distributed inference. These include NVIDIA TensorRT-LLM, NVIDIA Dynamo, TileIR, Cutlass, the NVIDIA Collective Communication Library and NIX which are integrated into millions of workflows.Allowing developers to build with their framework of choice, NVIDIA has collaborated with top open framework providers to offer model optimizations for FlashInfer, PyTorch, SGLang, vLLM and others.Plus, NVIDIA NIM microservices are available for popular open models like OpenAIs gpt-oss and Llama 4, making it easy for developers to operate managed application programming interfaces with the flexibility and security of self-hosting models on their preferred infrastructure.Learn more about the latest advancements in inference and accelerated computing by joining NVIDIA at Hot Chips.
    0 Commentaires ·0 Parts ·0 Aperçu
  • Into the Omniverse: How OpenUSD and Digital Twins Are Powering Industrial and Physical AI
    blogs.nvidia.com
    Editors note: This blog is a part of Into the Omniverse, a series focused on how developers, 3D practitioners and enterprises can transform their workflows using the latest advances in OpenUSD and NVIDIA Omniverse.Investments in industrial AI and physical AI are driving increased demand for digital twins across industries.These physically accurate, virtual replicas of real-world environments, facilities and processes arent just helping manufacturers streamline planning and optimize operations. They serve as the training ground for helping ensure vision AI agents, autonomous vehicles and robot fleets can operate safely, efficiently and reliably.Creating physically accurate simulation environments that enable physical AI to transition seamlessly to the real world typically involves substantial manual effort. However, with the latest advancements in OpenUSD a powerful open standard for describing and connecting complex 3D worlds alongside improvements in rendering, neural reconstruction and world foundation models (WFMs), developers can fast-track the construction of digital twins at scale.Accelerating Digital Twin and Physical AI DevelopmentTo speed digital twin and physical AI development, NVIDIA announced at this years SIGGRAPH conference new research, NVIDIA Omniverse libraries, NVIDIA Cosmos WFMs and advanced AI infrastructure including NVIDIA RTX PRO Servers and NVIDIA DGX Cloud.The latest Omniverse software development kits bridge MuJoCo and Universal Scene Description (OpenUSD), enabling over 250,000 MJCF robot learning developers to simulate robots across platforms.Omniverse NuRec libraries and AI models enable Omniverse RTX ray-traced 3D Gaussian splatting, allowing developers to capture, reconstruct and simulate the real world in 3D using sensor data.NVIDIA Isaac Sim 5.0 and Isaac Lab 2.2 open-source robot simulation and learning frameworks are now available on GitHub. Isaac Sim features NuRec neural rendering and new OpenUSD robot and sensor schemas to narrow the simulation-to-reality gap.Cosmos WFMs, including Cosmos Transfer-2 and NVIDIA Cosmos Reason, deliver leaps in synthetic data generation and reasoning for physical AI development.NVIDIA research advances in rendering and AI-assisted material generation help developers scale digital twin development.Growing OpenUSD EcosystemOpenUSD serves as a foundational ecosystem for digital twin and physical AI development, empowering developers to integrate industrial and 3D data to create physically accurate digital twins.The Alliance for OpenUSD (AOUSD) recently welcomed new general members, including Accenture, Esri, HCLTech, PTC, Renault and Tech Soft 3D. These additions underscore the continued growth of the OpenUSD community and its commitment to unifying 3D workflows across industries.To address the growing demand for OpenUSD and digital twins expertise, NVIDIA launched a new industry-recognized OpenUSD development certification and a free digital twins learning path.Developers Building Digital TwinsIndustry leaders including Siemens, Sight Machine, Rockwell Automation, EDAG, Amazon Devices & Services and Vention are building digital twin solutions with Omniverse libraries and OpenUSD to enable transformation with industrial and physical AI.Siemens Teamcenter Digital Reality Viewer enables engineers to visualize, interact with and collaborate on photorealistic digital twins at unprecedented scale. These efforts are enabling faster design reviews, minimizing the need for physical prototypes and accelerating time to market all while reducing costs.Sight Machines Operator Agent platform combines live production data, agentic AI-powered recommendations and digital twins to provide real-time visibility into production and enable faster, more informed decisions for plant operations teams.Rockwell Automations Emulate3D Factory Test platform enables manufacturers to build factory-scale, physics-based digital twins for simulating, validating and optimizing automation and autonomous systems at scale.EDAGs industrial digital twin platform helps manufacturers improve project management, optimize production layouts, train workers and perform data-driven quality assurance.Amazon Devices & Services uses digital twins to train robotic arms to recognize, inspect and handle new devices. Robotic actions can be configured to manufacture products purely based on training performed in simulation including for steps involved in assembly, testing, packaging and auditing.Vention is using NVIDIA robotics, AI and simulation technologies including Omniverse libraries, Isaac Sim and Jetson hardware to deliver plug-and-play digital twin and automation solutions that simplify and accelerate the deployment of intelligent manufacturing systems.Get Plugged Into the World of OpenUSDTo learn more about OpenUSD and how to develop digital twin applications with Omniverse libraries, take free courses as part of the new digital twin learning path, and check out the Omniverse Kit companion tutorial and how-to guide for deploying Omniverse Kit-based applications at scale.Watch a replay of NVIDIAs SIGGRAPH Research Special Address. Plus, try out Omniverse NuRec on Isaac Sim and CARLA, and learn more about Isaac Sim.Stay up to date by subscribing to NVIDIA Omniverse news, joining the Omniverse community and following Omniverse on Discord, Instagram, LinkedIn, Threads, X, and YouTube.Explore the Alliance for OpenUSD forum and the AOUSD website.Featured image courtesy of Siemens, Sight Machine.
    Like
    Love
    Wow
    Sad
    Angry
    36
    · 0 Commentaires ·0 Parts ·0 Aperçu
  • At Gamescom 2025, NVIDIA DLSS 4 and Ray Tracing Come to This Years Biggest Titles
    blogs.nvidia.com
    With over 175 games now supporting NVIDIA DLSS 4 a suite of advanced, AI-powered neural rendering technologies gamers and tech enthusiasts everywhere can experience breakthrough performance in this years most anticipated titles, including Borderlands 4, Hell Is Us and Fate Trigger.Plus, path tracing is making its way to Resident Evil Requiem and Directive 8020, as well as ray tracing in upcoming releases like Phantom Blade Zero, PRAGMATA and CINDER CITY enabling crystal-clear visuals for more immersive gameplayDLSS 4 and path tracing are no longer cutting-edge graphical experiments theyre the foundation of modern PC gaming titles, said Matt Wuebbling, vice president of global GeForce marketing at NVIDIA. Developers are embracing AI-powered rendering to unlock stunning visuals and massive performance gains, enabling gamers everywhere to experience the future of real-time graphics today.These announcements come alongside a new NVIDIA GeForce RTX 50 Series bundle for Borderlands 4 and updates to the NVIDIA app a companion platform for content creators, gamers and AI enthusiasts using NVIDIA GeForce RTX GPUs.DLSS 4 Now Accelerating Over 175 Games and ApplicationsLaunched with the GeForce RTX 50 Series earlier this year, DLSS 4 with Multi Frame Generation uses AI to generate up to three frames for every traditionally rendered frame, delivering performance boosts of up to 8x over traditional rendering.In addition to Multi Frame Generation, DLSS 4 titles include support for DLSS Super Resolution, Ray Reconstruction and NVIDIA Reflex technology unlocking incredible performance gains and responsive gameplay for every GeForce RTX 50 Series owner.New titles announced at Gamescom that will support the latest RTX technologies include:Directive 8020 and Resident Evil Requiem, which are launching with DLSS 4 and path tracingBlack State, CINDER CITY (formerly Project LLL), Cronos: The New Dawn, Dying Light: The Beast, Honeycomb: The World Beyond, Lost Soul Aside, The Outer Worlds 2, Phantom Blade Zero and PRAGMATA, which are launching with DLSS 4 and ray tracingBorderlands 4 and Fate Trigger, which are launching with DLSS 4 with Multi Frame GenerationIndiana Jones and the Great Circle, which in September will add support for RTX Hair, a technology that uses new hardware capabilities in RTX 50 Series GPUs to model hair with greater path-traced detail and realismMany of these RTX titles will also launch on the GeForce NOW cloud gaming platform, including Borderlands 4, CINDER CITY (formerly Project LLL), Hell Is Us and The Outer Worlds 2.NVIDIA App Adds Global DLSS Overrides and Software UpdatesThe NVIDIA app is the essential companion for NVIDIA GeForce RTX GPU users, simplifying the process of keeping PCs updated with the latest GeForce Game Ready and NVIDIA Studio Drivers.New updates to the NVIDIA app include:Global DLSS Overrides: Easily enable DLSS Multi-Frame Generation or DLSS Super Resolution profiles globally across hundreds of DLSS Override titles, instead of needing to configure per title.Project G-Assist Upgrades: The latest update to Project G-Assist an on-device AI assistant that lets users control and tune their RTX systems with voice and text commands introduces a significantly more efficient AI model that uses 40% less memory. Despite its smaller footprint, it responds to queries faster and more accurately calls the right tools.Highly Requested Legacy 3D Settings: Use easily configurable control panel settings including anisotropic filtering, anti-aliasing and ambient occlusion to enhance classic games.The NVIDIA app beta update launches Tuesday, Aug. 19, at 9 a.m. PT, with full availability coming the following week.NVIDIA ACE Enhances Voice-Driven Gaming ExperiencesNVIDIA ACE a suite of generative AI technologies that power lifelike non-playable character interactions in games like Kraftons inZOI now features in Iconic Interactives The Oversight Bureau, a darkly comic, voice-driven puzzle game.Using speech-to-text technology powered by ACE, players can interact naturally with in-game characters using speech, with Iconics Narrative Engine interpreting the input and determining and delivering the pre-recorded character dialogue that best fits the story and situation.This system keeps developers in creative control while offering players real agency in games all running locally on RTX AI PCs with sub-second latency.The Oversight Bureau launches later this year and will be playable at NVIDIAs Gamescom B2B press suite.NVIDIA RTX Remix Evolves With Community Expansions and New Particle SystemNVIDIA RTX Remix, an open-source modding platform for remastering classic games with path tracing and neural rendering, continues to grow thanks to its passionate community.Modders have been using large language models to extend RTX Remixs capabilities. For example, one modder vibe coded a plug-in that connects RTX Remix to Adobe Substance 3D, the industry-standard tool for 3D texturing and materials. Another modder made it possible for RTX Remix to use classic game data to instantly make objects glow with emissive effects.RTX Remixs open-source community has even expanded compatibility to allow many new titles to be remastered, including iconic games like Call Of Duty 4: Modern Warfare, Knights Of The Old Republic, Doom 3, Half-Life: Black Mesa and Bioshock.Some of these games were featured in the RTX Remixs $50K Mod Contest, which wrapped up at Gamescom. Painkiller RTX by Merry Pencil Studios won numerous awards, including Best Overall RTX Remix Mod. Explore all mod submissions on ModDB.com.At Gamescom, NVIDIA also unveiled a new RTX Remix particle system that brings dynamic, realistically lit and physically accurate particles to 165 classic games the majority of which have never had a particle editor.Modders can use the system to change the look, size, quantity, light emission, turbulence and even gravity of particles in games. The new particle system will be available in September.Borderlands 4 GeForce RTX 50 Series Bundle Available NowTo celebrate Gearboxs Borderlands 4, which will be enhanced by DLSS 4 with Multi Frame Generation and NVIDIA Reflex, NVIDIA is introducing a new GeForce RTX 50 Series bundle.Players who purchase a GeForce RTX 5090, 5080, 5070 Ti, or 5070 desktop system or graphics card or laptops with a GeForce RTX 5090 Laptop GPU, RTX 5080 Laptop GPU, RTX 5070 Ti Laptop GPU or RTX 5070 Laptop GPU from participating retailers will receive a copy of Borderlands 4 and The Gilded Glory Pack DLC. The offer is available through Monday, Sept. 22.Learn more about GeForce announcements at Gamescom.
    Like
    Love
    Wow
    Sad
    Angry
    388
    · 0 Commentaires ·0 Parts ·0 Aperçu
  • New Lightweight AI Model for Project G-Assist Brings Support for 6GB NVIDIA GeForce RTX and RTX PRO GPUs
    blogs.nvidia.com
    At Gamescom, NVIDIA is releasing its first major update to Project GAssist an experimental on-device AI assistant that allows users to tune their NVIDIA RTX systems with voice and text commands.The update brings a new AI model that uses 40% less VRAM, improves tool-calling intelligence and extends G-Assist support to all RTX GPUs with 6GB or more VRAM, including laptops. Plus, a new G-Assist Plug-In Hub enables users to easily discover and download plug-ins to enable more G-Assist features.NVIDIA also announced a new path-traced particle system, coming in September to the NVIDIA RTX Remix modding platform, that brings fully simulated physics, dynamic shadows and realistic reflections to visual effects.In addition, NVIDIA named the winners of the NVIDIA and ModDB RTX Remix Mod Contest. Check out the winners and finalist RTX mods in the RTX Remix GeForce article.G-Assist Gets Smarter, Expands to More RTX PCsThe modern PC is a powerhouse, but unlocking its full potential means navigating a complex maze of settings across system software, GPU and peripheral utilities, control panels and more.Project G-Assist is a free, on-device AI assistant built to cut through that complexity. It acts as a central command center, providing easy access to functions previously buried in menus through voice or text commands. Users can ask the assistant to:Run diagnostics to optimize game performanceDisplay or chart frame rates, latency and GPU temperaturesAdjust GPU or even peripheral settings, such as keyboard lightingThe G-Assist update also introduces a new, significantly more efficient AI model thats faster and uses 40% less memory while maintaining response accuracy. The more efficient model means that G-Assist can now run on all RTX GPUs with 6GB or more VRAM, including laptops.Getting started is simple: install the NVIDIA app and the latest Game Ready Driver on Aug. 19, download the G-Assist update from the apps home screen and press Alt+G to activate.Another G-Assist update coming in September will introduce support for laptop-specific commands for features like NVIDIA BatteryBoost and Battery OPS.Introducing the G-Assist Plug-In Hub With Mod.ioNVIDIA is collaborating with mod.io to launch the G-Assist Plug-In Hub, which allows users to easily access G-Assist plug-ins, as well as discover and download community-created ones.With the mod.io plug-in, users can ask G-Assist to discover and install new plug-ins.With the latest update, users can also directly ask G-Assist what new plug-ins are available in the hub and install them using natural language, thanks to a mod.io plug-in.The recent G-Assist Plug-In Hackathon showcased the incredible creativity of the G-Assist community. Heres a sneak peek of what they came up with:Some finalists include:Omniplay allows gamers to use G-Assist to research lore from online wikis or take notes in real time while gamingLaunchpad lets gamers set, launch and toggle custom app groups on the fly to boost productivityFlux NIM Microservice for G-Assist allows gamers to easily generate AI images from within G-Assist, using on-device NVIDIA NIM microservicesThe winners of the hackathon will be announced on Wednesday, Aug. 20.Building custom plug-ins is simple. Theyre based on a foundation of JSON and Python scripts and the Project G-Assist Plug-In Builder helps further simplify development by enabling users to code plug-ins with natural language.Mod It Like Its Hot With RTX RemixClassic PC games remain beloved for their unforgettable stories, characters and gameplay but their dated graphics can be a barrier for new and longtime players.NVIDIA RTX Remix enables modders to revitalize these timeless titles with the latest NVIDIA gaming technologies bridging nostalgic gameplay with modern visuals.Since the platforms release, the RTX Remix modding community has grown with over 350 active projects and over 100 mods released. The mods span a catalog of beloved games like Half-Life 2, Need for Speed: Underground, Portal 2 and Deus Ex and have amassed over 2 million downloads.In May, NVIDIA invited modders to participate in the NVIDIA and ModDB RTX Remix Mod Contest for a chance to win $50,000 in cash prizes. At Gamescom, NVIDIA announced the winners:Best Overall RTX Mod Winner: Painkiller RTX Remix, by Binq_AdamsBest Use of RTX in a Mod Winner: Painkiller RTX Remix, by Binq_AdamsRunner-Up: Vampire: The Masquerade Bloodlines RTX Remaster, by SafemilkMost Complete RTX Mod Winner: Painkiller RTX Remix, by Binq_AdamsRunner-Up: I-Ninja Remixed, by g.i.george333Community Choice RTX Mod Winner: Call of Duty 2 RTX Remix of Carentan, by tadpole3159These modders tapped RTX Remix and generative AI to bring their creations to life from enhancing textures to quickly creating images and 3D assets.For example, the Merry Pencil Studios modder team used a workflow that seamlessly connected RTX Remix and ComfyUI, allowing them to simply select textures in the RTX Remix viewport and, with a single click in ComfyUI, restore them.The results are stunning, with each texture meticulously recreated with physically based materials layered with grime and rust. With a fully path-traced lighting system, the games gothic horror atmosphere has never felt more immersive to play through.All mods submitted to the RTX Remix Modding Contest, as well as 100 more Remix mods, are available to download from ModDB. For a sneak peek at RTX Remix projects under active development, check out the RTX Remix Showcase Discord server.Another RTX Remix update coming in September will allow modders to create new particles that match the look of those found in modern titles. This opens the door for over 165 RTX Remix-compatible games to have particles for the first time.To get started creating RTX mods, download NVIDIA RTX Remix from the home screen of the NVIDIA app. Read the RTX Remix article to learn more about the contest and winners.Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, productivity apps and more on AI PCs and workstations.Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X and stay informed by subscribing to the RTX AI PC newsletter. Join NVIDIAs Discord server to connect with community developers and AI enthusiasts for discussions on whats possible with RTX AI.Follow NVIDIA Workstation on LinkedIn and X.See notice regarding software product information.
    0 Commentaires ·0 Parts ·0 Aperçu
  • 0 Commentaires ·0 Parts ·0 Aperçu
  • Now Were Talking: NVIDIA Releases Open Dataset, Models for Multilingual Speech AI
    blogs.nvidia.com
    Of around 7,000 languages in the world, a tiny fraction are supported by AI language models. NVIDIA is tackling the problem with a new dataset and models that support the development of high-quality speech recognition and translation AI for 25 European languages including languages with limited available data like Croatian, Estonian and Maltese.These tools will enable developers to more easily scale AI applications to support global users with fast, accurate speech technology for production-scale use cases such as multilingual chatbots, customer service voice agents and near-real-time translation services. They include:Granary, a massive, open-source corpus of multilingual speech datasets that contains around a million hours of audio, including nearly 650,000 hours for speech recognition and over 350,000 hours for speech translation.NVIDIA Canary-1b-v2, a billion-parameter model trained on Granary for high-quality transcription of European languages, plus translation between English and two dozen supported languages.NVIDIA Parakeet-tdt-0.6b-v3, a streamlined, 600-million-parameter model designed for real-time or large-volume transcription of Granarys supported languages.The paper behind Granary will be presented at Interspeech, a language processing conference taking place in the Netherlands, Aug. 17-21. The dataset, as well as the new Canary and Parakeet models, are now available on Hugging Face.How Granary Addresses Data ScarcityTo develop the Granary dataset, the NVIDIA speech AI team collaborated with researchers from Carnegie Mellon University and Fondazione Bruno Kessler. The team passed unlabeled audio through an innovative processing pipeline powered by NVIDIA NeMo Speech Data Processor toolkit that turned it into structured, high-quality data.This pipeline allowed the researchers to enhance public speech data into a usable format for AI training, without the need for resource-intensive human annotation. Its available in open source on GitHub.With Granarys clean, ready-to-use data, developers can get a head start building models that tackle transcription and translation tasks in nearly all of the European Unions 24 official languages, plus Russian and Ukrainian.For European languages underrepresented in human-annotated datasets, Granary provides a critical resource to develop more inclusive speech technologies that better reflect the linguistic diversity of the continent all while using less training data.The team demonstrated in their Interspeech paper that, compared to other popular datasets, it takes around half as much Granary training data to achieve a target accuracy level for automatic speech recognition (ASR) and automatic speech translation (AST).Tapping NVIDIA NeMo to Turbocharge TranscriptionThe new Canary and Parakeet models offer examples of the kinds of models developers can build with Granary, customized to their target applications. Canary-1b-v2 is optimized for accuracy on complex tasks, while parakeet-tdt-0.6b-v3 is designed for high-speed, low-latency tasks.By sharing the methodology behind the Granary dataset and these two models, NVIDIA is enabling the global speech AI developer community to adapt this data processing workflow to other ASR or AST models or additional languages, accelerating speech AI innovation.Canary-1b-v2, available under a permissive license, expands the Canary familys supported languages from four to 25. It offers transcription and translation quality comparable to models 3x larger while running inference up to 10x faster.https://blogs.nvidia.com/wp-content/uploads/2025/08/Canary-demo.mp4NVIDIA NeMo, a modular software suite for managing the AI agent lifecycle, accelerated speech AI model development. NeMo Curator, part of the software suite, enabled the team to filter out synthetic examples from the source data so that only high-quality samples were used for model training. The team also harnessed the NeMo Speech Data Processor toolkit for tasks like aligning transcripts with audio files and converting data into the required formats.Parakeet-tdt-0.6b-v3 prioritizes high throughput and is capable of transcribing 24-minute audio segments in a single inference pass. The model automatically detects the input audio language and transcribes without additional prompting steps.Both Canary and Parakeet models provide accurate punctuation, capitalization and word-level timestamps in their outputs.Read more on GitHub and get started with Granary on Hugging Face.
    0 Commentaires ·0 Parts ·0 Aperçu
  • NVIDIA, National Science Foundation Support Ai2 Development of Open AI Models to Drive U.S. Scientific Leadership
    blogs.nvidia.com
    NVIDIA is partnering with the U.S. National Science Foundation (NSF) to create an AI system that supports the development of multimodal language models for advancing scientific research in the United States.The partnership supports the NSF Mid-Scale Research Infrastructure project, called Open Multimodal AI Infrastructure to Accelerate Science (OMAI).Bringing AI into scientific research has been a game changer, said Brian Stone, performing the duties of the NSF director. NSF is proud to partner with NVIDIA to equip Americas scientists with the tools to accelerate breakthroughs. These investments are not just about enabling innovation; they are about securing U.S. global leadership in science and technology and tackling challenges once thought impossible.OMAI, part of the work of the Allen Institute for AI, or Ai2, aims to build a national fully open AI ecosystem to drive scientific discovery through AI, while also advancing the science of AI itself.NVIDIAs support of OMAI includes providing NVIDIA HGX B300 systems state-of-the-art AI infrastructure built to accelerate model training and inference with exceptional efficiency along with the NVIDIA AI Enterprise software platform, empowering OMAI to transform massive datasets into actionable intelligence and breakthrough innovations.NVIDIA HGX B300 systems are built with NVIDIA Blackwell Ultra GPUs and feature industry-leading high-bandwidth memory and interconnect technologies to deliver groundbreaking acceleration, scalability and efficiency to run the worlds largest models and most demanding workloads.AI is the engine of modern science and large, open models for Americas researchers will ignite the next industrial revolution, said Jensen Huang, founder and CEO of NVIDIA. In collaboration with NSF and Ai2, were accelerating innovation with state-of-the-art infrastructure that empowers U.S. scientists to generate limitless intelligence, making it Americas most powerful and renewable resource.The contributions will support research teams from the University of Washington, the University of Hawaii at Hilo, the University of New Hampshire and the University of New Mexico. The public-private partnership investment in U.S. technology aligns with recent initiatives outlined by the White House AI Action Plan, which supports Americas global AI leadership.The models are part of the national research infrastructure but we cant build the models without compute, and thats why NVIDIA is so important to this project, said Noah Smith, senior director of natural language processing research at Ai2.Opening Language Models to Advance American ResearchersDriving some of the fastest-growing applications in history, todays large language models (LLMs) have many billions of parameters, or internal weights and biases learned in training. LLMs are trained on trillions of words, and multimodal LLMs can ingest images, graphs, tables and more.But the power of these so-called frontier models can sometimes be out of reach for scientific research when the parameters, training data, code and documentation are not openly available.With the model training data in hand, you have the opportunity to trace back to particular training instances similar to a response, and also more systematically study how emerging behaviors relate to the training data, said Smith.NVIDIAs partnership with NSF to support Ai2s OMAI initiative provides fully open model access to data, open-source data interrogation tools to help refine datasets, as well as documentation and training for early-career researchers advancing U.S. global leadership in science and engineering.The Ai2 project supported by NVIDIA technologies pledges to make the software and models available at low or zero cost to researchers, similar to open-source code repositories and science-oriented digital libraries. Its in line with Ai2s previous work in creating fully open language models and multimodal models, maximizing access.Driving U.S. Global Leadership in Science and EngineeringWinning the AI Race: Americas AI Action Plan was announced in July by the White House, supported with executive orders to accelerate federal permitting of data center infrastructure and promote exportation of the American AI technology stack.The OMAI initiative aligns with White House AI Action Plan priorities, emphasizing the acceleration of AI-enabled science and supporting the creation of leading open models to enhance Americas global AI leadership in academic research and education.
    0 Commentaires ·0 Parts ·0 Aperçu
  • Warhammer 40,000: Dawn of War Definitive Edition Storms GeForce NOW at Launch
    blogs.nvidia.com
    Warhammer 40,000: Dawn of War Definitive Edition is marching onto GeForce NOW, expanding the cloud gaming platforms library to over 2,300 supported titles.Battle is just a click away, as the iconic real-time strategy game joins seven new releases this week. Commanders can prepare their squads and steel their nerves on any device including laptops, Macs, Steam Decks and NVIDIA SHIELD TVs.Microsofts surprise announcement at Quakecon is now available in the cloud: legendary fantasy shooters Heretic + Hexen have been conjured out of the shadows and are streaming on GeForce NOW.And dont miss out on in-game rewards for the popular, free-to-play, massively multiplayer online game World of Tanks as publisher Wargaming celebrates the titles 15-year anniversary.GeForce NOW will be at Gamescom 2025 the worlds largest gaming tradeshow starting Wednesday, Aug. 20. Stay tuned to GFN Thursday for all the latest updates.The Emperors CallMake your victories shine from the cloud.The grimdark future calls. Warhammer 40,000: Dawn of War Definitive Edition storms onto the battlefield with ferocious, squad-based real-time strategy. Command the Space Marines, Orks, Chaos, Eldar and more across four legendary campaigns and nine playable armies. From bolter roars to Waaagh! cries, battles erupt with uncompromising brutality, tactical depth and a healthy dose of swagger.Fully remastered with enhanced 4K visuals, a refined camera, an improved user interface and more, Dawn of War: Definitive Edition preserves the iconic chaos of the original game while throwing open the gates for creative mayhem. Every charge, psychic blast and last-stand is rendered sharper than ever as cunning, courage and unrelenting war decide the fate of worlds.GeForce NOW delivers the firepower needed to join the frontlines without having to wait for downloads or lengthy installs. Gamers can leap straight into battle, resume campaigns and join multiplayer chaos with just a few clicks. No frames lost to underpowered hardware every skirmish, every decisive strike is rendered in full glory in the cloud.Time to CelebrateMake your victories shine from the cloud.Roll out the tanks for World of Tankss 15th-anniversary celebration. Join the party by logging into the game every day through Sunday, Aug. 31 for exclusive commemorative rewards.Heres whats on deck: daily in-game giveaways, deep discounts, a pulse-pounding limited-time game mode and a special Battle Pass chapter packed with surprises. Watch for Twitch drops, enjoy increased credit earnings when playing with veteran tankers and dive into a unique photo-album event where each day reveals a new chapter in the evolution of maps, vehicles and epic memories.Enjoy smooth, lightning-fast gameplay on GeForce NOW even on modest hardware and share every explosive moment with friends, fans and fellow commanders. No download hassles, just pure, seamless action.Get HexedSuit up, pick a class and let chaos reign.Step into the shadowy worlds that shaped fantasy shooters fully restored by Nightdive Studios. Heretic + Hexen, the cult classics forged by Raven Software, are back with a vengeance, bringing their spell-slinging attitude and dark magic to a whole new generation.This definitive collection brings together Heretic: Shadow of the Serpent Riders, Hexen: Beyond Heretic and Hexen: Deathkings of the Dark Citadel plus two brand-new episodes, Heretic: Faith Renewed and Hexen: Vestiges of Grandeur, crafted with id Software and Nightdive Studios.Dive into over 110 campaign maps, 120 deathmatch arenas, online and split-screen multiplayer modes, 4K 120 frames-per-second (fps) visuals, modern controls and more spell-slinging action than ever.Experience the arcane might of Heretic + Hexen with GeForce NOW, which offers instant gameplay on nearly any device, with cloud-powered graphics, ultrasmooth performance and zero downloads. Ultimate members can crank up the magic and stream at up to 4K 120 fps even without the latest hardware, so every exploding tome and fireball looks spellbindingly sharp.All Aboard for New GamesOutwit the future.All aboard, Trailblazers. Honkai Star Rails new Version 3.5 Before Their Deaths is available to stream on GeForce NOW no need to wait for patches or updates to downloads.The latest version brings two new playable characters, Hysilens and Imperator Cerydra, who bring fresh abilities and strategies to the game. Journey back a thousand years to ancient Okhema, face the ever-shifting menace Lygus and explore the dazzling streets of Styxia, the City of Infinite Revelry. Between epic battles, serve fairy patrons in the Chrysos Maze Grand Restaurant, mix drinks with old friends and uncover secrets that could change everything. Get ready the next stop on the Astral Express is about to be unforgettable.In addition, members can look for the following:Echoes of the End (New release on Steam, Aug. 12)9 Kings (New release on Xbox, available on PC Game Pass, Aug. 14)Warhammer 40,000: Dawn of War Definitive Edition (New release on Steam, Aug. 14)Supraworld (New release on Steam, Aug. 15)Crash Bandicoot 4: Its About Time (New release on Steam and Battle.net)Guntouchables (Steam)Heretic + Hexen (Steam and Xbox, available on PC Game Pass)What are you planning to play this weekend? Let us know on X or in the comments below.What's a classic game that you still love to play? NVIDIA GeForce NOW (@NVIDIAGFN) August 13, 2025
    Like
    Love
    Wow
    Sad
    Angry
    107
    · 0 Commentaires ·0 Parts ·0 Aperçu
  • Applications Now Open for $60,000 NVIDIA Graduate Fellowship Awards
    blogs.nvidia.com
    Bringing together the worlds brightest minds and the latest accelerated computing technology leads to powerful breakthroughs that help tackle some of the biggest research problems.To foster such innovation, the NVIDIA Graduate Fellowship Program provides grants, mentors and technical support to doctoral students doing outstanding research relevant to NVIDIA technologies. The program, in its 25th year, is now accepting applications worldwide.It focuses on supporting students working in AI, machine learning, autonomous vehicles, computer graphics, robotics, healthcare, high-performance computing and related fields. Awards are up to $60,000 per student.Since its start in 2002, the Graduate Fellowship Program has awarded over 200 grants worth more than $7.3 million.Students must have completed at least their first year of Ph.D.-level studies at the time of application.The application deadline for the 2026-2027 academic year is Monday, Sept. 15, 2025. An in-person internship at an NVIDIA research office preceding the fellowship year is mandatory; eligible candidates must be available for the internship in summer 2026.For more on eligibility and how to apply, visit the program website.
    0 Commentaires ·0 Parts ·0 Aperçu
  • FLUX.1 Kontext NVIDIA NIM Microservice Now Available for Download
    blogs.nvidia.com
    Black Forest Labs FLUX.1 Kontext [dev] image editing model is now available as an NVIDIA NIM microservice.FLUX.1 models allow users to edit existing images with simple language, without the need for fine-tuning or complex workflows.Deploying powerful AI requires curation of model variants, adaptation to manage all input and output data, and quantization to reduce VRAM requirements. Models must be converted to work with optimized inference backend software and connected to new AI application programming interfaces.The FLUX.1 Kontext [dev] NIM microservice simplifies this process, unlocking faster generative AI workflows, and is optimized for RTX AI PCs.Generative AI in KontextFLUX.1 Kontext [dev] is an open-weight generative model built for image editing. It features a guided, step-by-step generation process that makes it easier to control how an image evolves, whether refining small details or transforming an entire scene.Image generated by FLUX.1 Kontext [dev] with a simple text prompt.Because the model accepts both text and image inputs, users can easily reference a visual concept and guide how it evolves in a natural and intuitive way. This enables coherent, high-quality image edits that stay true to the original concept.Guide edits with simple language, without the need for fine-tuning or complex workflows.The FLUX.1 Kontext [dev] NIM microservice provides prepackaged, optimized files that are ready for one-click download through ComfyUI NIM nodes making them easily accessible to users.The original image is revised with six prompts to reach the desired result.NVIDIA and Black Forest Labs worked together to quantize FLUX.1 Kontext [dev], reducing the model size from 24GB to 12GB for FP8 (NVIDIA Ada Generation GPUs) and 7GB for FP4 (NVIDIA Blackwell architecture). The FP8 checkpoint is optimized for GeForce RTX 40 Series GPUs, which have FP8 accelerators in their Tensor Cores. The FP4 checkpoint is optimized for GeForce RTX 50 Series GPUs and uses a new method called SVDQuant, which preserves image quality while reducing model size.Speedup compared with BF16 GPU (left, higher is better), and memory usage required to run FLUX.1 Kontext [dev] in different precisions (right, lower is better).In addition, NVIDIA TensorRT a framework to access the Tensor Cores in NVIDIA RTX GPUs for maximum performance provides over 2x acceleration compared with running the original BF16 model with PyTorch.These dramatic performance gains were previously limited to AI specialists and developers with advanced AI infrastructure knowledge. With the FLUX.1 Kontext [dev] NIM microservice, even enthusiasts can achieve these time savings with greater performance.Get NIMbleFLUX.1 Kontext [dev] is available on Hugging Face with TensorRT optimizations and ComfyUI.To get started, follow the directions on ComfyUIs NIM nodes GitHub:Install NVIDIA AI Workbench.Get ComfyUI.Install NIM nodes through the ComfyUI Manager within the app.Accept the model licenses on Black Forest Labs FLUX.1 Kontexts [dev] Hugging Face.The node will prepare the desired workflow and help with downloading all necessary models after clicking Run.NIM microservices are optimized for performance on NVIDIA GeForce RTX and RTX PRO GPUs and include popular models from the AI community. Explore NIM microservices on GitHub and build.nvidia.com.Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, productivity apps and more on AI PCs and workstations.Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X and stay informed by subscribing to the RTX AI PC newsletter. Join NVIDIAs Discord server to connect with community developers and AI enthusiasts for discussions on whats possible with RTX AI.Follow NVIDIA Workstation on LinkedIn and X.See notice regarding software product information.
    0 Commentaires ·0 Parts ·0 Aperçu
  • Amazon Devices & Services Achieves Major Step Toward Zero-Touch Manufacturing With NVIDIA AI and Digital Twins
    blogs.nvidia.com
    Using NVIDIA digital twin technologies, Amazon Devices & Services is powering big leaps in manufacturing with a new physical AI software solution.Deployed this month at an Amazon Devices facility, the companys innovative, simulation-first approach for zero-touch manufacturing trains robotic arms to inspect diverse devices for product-quality auditing and integrate new goods into the production line all based on synthetic data, without requiring hardware changes.This new technology brings together Amazon Devices-created software that simulates processes on the assembly line with products in NVIDIA-powered digital twins. Using a modular, AI-powered workflow, the technology offers faster, more efficient inspections compared with the previously used audit machinery.Simulating processes and products in digital twins eliminates the need for expensive, time-consuming physical prototyping. This eases manufacturer workflows and reduces the time it takes to get new products into consumers hands.To enable zero-shot manufacturing for the robotic operations, the solution uses photorealistic, physics-enabled representations of Amazon devices and factory work stations to generate synthetic data. This factory-specific data is then used to enhance AI model performance in both simulation and at the real work station, minimizing the simulation-to-real gap before deployment.Its a huge step toward generalized manufacturing: the use of automated systems and technologies to flexibly handle a wide variety of products and production processes even without physical prototypes.AI, Digital Twins for Robot UnderstandingBy training robots in digital twins to recognize and handle new devices, Amazon Devices & Services is equipped to build faster, more modular and easily controllable manufacturing pipelines, allowing lines to change from auditing one product to another simply via software.Robotic actions can be configured to manufacture products purely based on training performed in simulation including for steps involved in assembly, testing, packaging and auditing.A suite of NVIDIA Isaac technologies enables Amazon Devices & Services physically accurate, simulation-first approach.When a new device is introduced, Amazon Devices & Services puts its computer-aided design (CAD) model into NVIDIA Isaac Sim, an open-source, robotics simulation reference application built on the NVIDIA Omniverse platform.NVIDIA Isaac is used to generate over 50,000 diverse, synthetic images from the CAD models for each device, crucial for training object- and defect-detection models.Then, Isaac Sim processes the data and taps into NVIDIA Isaac ROS to generate robotic arm trajectories for handling the product.The robot is trained purely on synthetic data and can pick up packages and products of different shapes and sizes to perform cosmetic inspection. Real station (left) and simulated station (right). Image courtesy of Amazon Devices & Services.The development of this technology was significantly accelerated by AWS through distributed AI model training on Amazon devices product specifications using Amazon EC2 G6 instances via AWS Batch, as well as NVIDIA Isaac Sim physics-based simulation and synthetic data generation on Amazon EC2 G6 family instances.The solution uses Amazon Bedrock a service for building generative AI applications and agents to plan high-level tasks and specific audit test cases at the factory based on analyses of product-specification documents. Amazon Bedrock AgentCore will be used for autonomous-workflow planning for multiple factory stations on the production line, with the ability to ingest multimodal product-specification inputs such as 3D designs and surface properties.To help robots understand their environment, the solution uses NVIDIA cuMotion, a CUDA-accelerated motion-planning library that can generate collision-free trajectories in a fraction of a second on the NVIDIA Jetson AGX Orin module. The nvblox library, part of Isaac ROS, generates distance fields that cuMotion uses for collision-free trajectory planning.FoundationPose, an NVIDIA foundation model trained on 5 million synthetic images for pose estimation and object tracking, helps ensure the Amazon Devices & Services robots know the accurate position and orientation of the devices.Crucial for the new manufacturing solution, FoundationPose can generalize to entirely new objects without prior exposure, allowing seamless transitions between different products and eliminating the need to collect new data to retrain models for each change.As part of product auditing, the new solutions approach is used for defect detection on the manufacturing line. Its modular design allows for future integration of advanced reasoning models like NVIDIA Cosmos Reason.Watch the NVIDIA Research special address at SIGGRAPH and learn more about how graphics and simulation innovations come together to drive industrial digitalization by joining NVIDIA at the conference, running through Thursday, Aug. 14.
    0 Commentaires ·0 Parts ·0 Aperçu
  • Mini Footprint, Mighty AI: NVIDIA Blackwell Architecture Powers AI Acceleration in Compact Workstations
    blogs.nvidia.com
    Packing the power of the NVIDIA Blackwell architecture in compact, energy-efficient form factors, the NVIDIA RTX PRO 4000 Blackwell SFF Edition and NVIDIA RTX PRO 2000 Blackwell GPUs are coming soon delivering AI acceleration for professional workflows across industries.Applications are becoming increasingly AI accelerated, and more users need AI performance, no matter the size or shape of their workstation.The RTX PRO 4000 SFF and RTX PRO 2000 feature fourth-generation RT Cores and fifth-generation Tensor Cores with lower power in half the size of a traditional GPU.The new GPUs are designed to bring next-generation performance to a range of professional workflows, providing incredible speedups for engineering, design, content creation, AI and 3D visualization.Compared with the previous-generation architecture, the RTX PRO 4000 SFF features up to 2.5x higher AI performance, 1.7x higher ray-tracing performance and 1.5x more bandwidth, creating more efficiency with the same 70-watt max power consumption.Optimized for mainstream design and AI workflows, the RTX PRO 2000 offers up to 1.6x faster 3D modeling, 1.4x faster computer-aided design (CAD) performance and 1.6x quicker rendering speeds compared with the previous generation.The NVIDIA RTX PRO 2000 Blackwell.CAD and product engineers as well as creatives will benefit from the RTX PRO 2000 GPUs 1.4x boost in image generation and 2.3x leap in text generation, enabling faster iteration, rapid prototyping and seamless collaboration.Businesses Tap NVIDIA RTX PRO for SpeedupsBusinesses across fields including engineering, construction, architecture, media and entertainment, and healthcare are using RTX PRO Blackwell GPUs to instantly accomplish tasks that previously took hours.The Mile High Flood District protects people, property and the environment in the Denver, Colorado, metro area by managing flood risks with regional watershed planning, early warning systems, stream restoration and stormwater control, in collaboration with local governments.Mile High Flood District runs complex flood simulations, massive 3D visualizations and real-time AI workflows and with nearly double the CUDA cores, NVIDIA RTX PRO 2000 Blackwell is a big step up in performance compared with the NVIDIA RTX 2000 Ada Generation GPU, said Jon Villines, innovation manager at Mile High Flood District. NVIDIA RTX PRO allows us to more easily handle increasingly larger geographic information systems, as well as hydraulic and hydrologic datasets.The Government of Cantabria Geospatial Office is responsible for analyzing and visualizing high-resolution geographic information system data for government and public use.We tested the NVIDIA RTX PRO 2000 Blackwell and were very impressed with its performance on geospatial workloads with Esri ArcGIS Pro, said Gabriel Ortiz Rico, chief of service of cartography and geographic information systems at the Government of Cantabria. Fine-tuning of AI models is 2x faster compared with using the RTX 2000 Ada due to the RTX 2000 Blackwells additional Tensor Cores and GDDR7 memory.Studio Tim Fu (STF) is a London-based design studio specializing in the integration of human creativity and AI with architecture and design.The RTX PRO 2000 Blackwell powers our UrbanGPT application for real-time text-to-3D urban design, which can be used to generate dynamic city layouts, track vital metrics like program and floor areas, and produce realistic massing distribution across complex urban design scenarios, said Tim Fu, director of STF. From zoning simulations to large-scale massing studies, this technology accelerates our AI-driven design engine with the stability and responsiveness needed for city-scale planning.New York-based Thornton Tomasetti is a global engineering and design consulting firm integrating engineering, science, technology and forensic analysis to advance performance, resilience and innovation in the built environment and beyond.At Thornton Tomasetti, were constantly advancing computational engineering, said Rob Otani, chief technology officer of Thornton Tomasetti. We benchmarked the RTX PRO 2000 Blackwell on CORE.Matrix our in-house, GPU-based Finite Element Analysis solver running almost 3x faster than with the RTX 2000 Ada and 27x faster than with a standard CPU. This enabled us to accelerate our structural analysis workflows for more iterative, design-integrated engineering.Glxkind is a technology company that creates AI-powered smart baby strollers designed to improve safety, convenience and accessibility for parents and their children.Integrating the latest generation of advanced GPUs like the RTX PRO 2000 enables Glxkind to push the boundaries of whats possible in AI-powered parenting solutions, said Kevin Huang, CEO of Glxkind. The RTX PRO 2000s enhanced AI and graphics performance give us the real-time processing power needed to make our smart strollers safer, more responsive and more convenient for families everywhere.The Software Driving InnovationNVIDIAs software ecosystem enables creators, developers and enterprises to harness the full power of AI and advanced graphics.The NVIDIA AI Enterprise software suite delivers enterprise-grade tools for building, deploying and scaling production AI from generative AI and computer vision to speech and natural language solutions on virtually any infrastructure.The NVIDIA Cosmos platform offers world foundation models optimized for fast, efficient inference and edge deployment, enabling high-performance AI for robotics, automation and physical AI applications. The Cosmos-Reason1-7B model can run seamlessly on the RTX PRO 4000 SFF, delivering powerful physical AI reasoning capabilities to edge devices, compact workstations and industrial systems.NVIDIAs graphics and visualization tools, including the NVIDIA Omniverse platform, bring generative physical AI and simulation to 3D design teams, facilitating digital twins and visual workflows.In addition, the Blackwell platform builds on NVIDIAs ecosystem of powerful development tools, NVIDIA CUDA-X libraries, over 6 million developers and close to 6,000 applications to scale performance across thousands of GPUs.AvailabilityThe NVIDIA RTX PRO 2000 Blackwell and NVIDIA RTX PRO 4000 Blackwell SFF Edition GPUs are coming later this year.The RTX PRO 2000 is expected to be available from PNY and TD SYNNEX, as well as system builders such as BOXX, Dell Technologies, HP and Lenovo.The NVIDIA RTX PRO 4000 Blackwell SFF Edition is expected to be available from global distribution partners and leading manufacturing partners such as Dell Technologies, HP and Lenovo.Watch the NVIDIA Research special address at SIGGRAPH and learn more about how graphics and simulation innovations come together to drive industrial digitalization by joining NVIDIA at the conference, running through Thursday, Aug. 14.
    0 Commentaires ·0 Parts ·0 Aperçu
  • CrowdStrike, Uber, Zoom Among Industry Pioneers Building Smarter Agents With NVIDIA Nemotron and Cosmos Reasoning Models for Enterprise and Physical AI Applications
    blogs.nvidia.com
    AI agents are poised to deliver as much as $450 billion from revenue gains and cost savings by 2028, according to Capgemini. Developers building these agents are turning to higher-performing reasoning models to improve AI agent platforms and physical AI systems.At SIGGRAPH, NVIDIA today announced an expansion of two model families with reasoning capabilities NVIDIA Nemotron and NVIDIA Cosmos that leaders across industries are using to drive productivity via teams of AI agents and humanoid robots.CrowdStrike, Uber, Magna, NetApp and Zoom are among some of the enterprises tapping into these model families.New NVIDIA Nemotron Nano 2 and Llama Nemotron Super 1.5 models offer the highest accuracy in their size categories for scientific reasoning, math, coding, tool-calling, instruction-following and chat. These new models give AI agents the power to think more deeply and work more efficiently exploring broader options, speeding up research and delivering smarter results within set time limits.Think of the model as the brain of an AI agent it provides the core intelligence. But to make that brain useful for a business, it must be embedded into an agent that understands specific workflows, in addition to industry and business jargon, and operates safely. NVIDIA helps enterprises bridge that gap with leading libraries and AI blueprints for onboarding, customizing and governing AI agents at scale.Cosmos Reason is a new reasoning vision language model (VLM) for physical AI applications that excels in understanding how the real world works, using structured reasoning to understand concepts like physics, object permanence and space-time alignment.Cosmos Reason is purpose-built to serve as the reasoning backbone to a robot vision language action (VLA) model, or critique and caption training data for robotics and autonomous vehicles, and equip runtime visual AI agents with spatial-temporal understanding and reasoning of physical operations, like in factories or cities.Nemotron: Highest Accuracy and Efficiency for Agentic Enterprise AIAs enterprises develop AI agents to tackle complex, multistep tasks, models that can provide strong reasoning accuracy with efficient token generation enable intelligent, autonomous decision-making at scale.NVIDIA Nemotron is a family of advanced open reasoning models that use leading models, NVIDIA-curated open datasets and advanced AI techniques to provide an accurate and efficient starting point for AI agents.The latest Nemotron models deliver leading efficiency in three ways: a new hybrid model architecture, compact quantized models and a configurable thinking budget that provides developers with control over token generation, resulting in 60% lower reasoning costs. This combination lets the models reason more deeply and respond faster, without needing more time or computing power. This means better results at a lower cost.Nemotron Nano 2 provides as much as 6x higher token generation compared with other leading models of its size.Llama Nemotron Super 1.5 achieves leading performance and the highest reasoning accuracy in its class, empowering AI agents to reason better, make smarter decisions and handle complex tasks independently. Its now available in NVFP4, or 4-bit floating point, which delivers as much as 6x higher throughput on NVIDIA B200 GPUs compared with NVIDIA H100 GPUs.The chart above shows the Nemotron model delivers top reasoning accuracy in the same timeframe and on the same compute budget, delivering the highest accuracy per dollar.Along with the two new Nemotron models, NVIDIA is also announcing its first open VLM training dataset Llama Nemotron VLM dataset v1 with 3 million samples of optical character recognition, visual QA and captioning data that power the previously released Llama 3.1 Nemotron Nano VL 8B model.In addition to the accuracy of the reasoning models, agents also rely on retrieval-augmented generation to fetch the latest and most relevant information from connected data across disparate sources to make informed decisions. The recently released Llama 3.2 NeMo Retriever embedding model tops three visual document retrieval leaderboards ViDoRe V1, ViDoRe V2 and MTEB VisualDocumentRetrieval for boosting agentic system accuracy.Using these reasoning and information retrieval models, a deep research agent built using the AI-Q NVIDIA Blueprint is currently No. 1 for open and portable agents on DeepResearch Bench.NVIDIA NeMo and NVIDIA NIM microservices support the entire AI agent lifecycle from development and deployment to monitoring and optimization of the agentic systems.Cosmos Reason: A Breakthrough in Physical AIVLMs marked a breakthrough for computer vision and robotics, empowering machines to identify objects and patterns. However, nonreasoning VLMs lack the ability to understand and interact with the real world meaning they cant handle ambiguity or novel experiences, nor solve complex multistep tasks.NVIDIA Cosmos Reason is a new open, customizable, 7-billion-parameter reasoning VLM for physical AI and robotics. Cosmos Reason lets robots and vision AI agents reason like humans, using prior knowledge, physics understanding and common sense to understand and act in the physical world.Cosmos Reason enables advanced capabilities across robotics and physical AI applications such as training data critiquing and captioning, robot decision-making and video analytics AI agents.It can help automate the curation and annotation of large, diverse training datasets, accelerating the development of high-accuracy AI models. It can also serve as a sophisticated reasoning engine for robot planning, parsing complex instructions into actionable steps for VLA models, even in new environments.It also powers video analytics AI agents built on the NVIDIA Blueprint for video search and summarization (VSS), enabled by the NVIDIA Metropolis platform, gleaning valuable insights from massive volumes of stored or live video data. These visually perceptive and interactive AI agents can help streamline operations in factories, warehouses, retail stores, airports, traffic intersections and more by spotting anomalies.NVIDIAs robotics research team uses Cosmos Reason for data filtration and curation, and as the System 2 reasoning VLM behind VLA models such as the next versions of NVIDIA Isaac GR00T NX.Now Serving: NVIDIA Reasoning Models for AI Agents and Robots EverywhereDiverse enterprises and consulting leaders are adopting NVIDIAs latest reasoning models. Leaders spanning cybersecurity to telecommunications are among those working with Nemotron to build enterprise AI agents.Zoom plans to harness Nemotron reasoning models with Zoom AI Companion to make decisions and manage multistep tasks to take action for users across Zoom Meetings, Zoom Chat and Zoom documents.CrowdStrike is testing Nemotron models to enable its Charlotte AI agents to write queries on the CrowdStrike Falcon platform.Amdocs is using NVIDIA Nemotron models in its amAIz Suite to drive AI agents to handle complex, multistep automation spanning care, sales, network and customer support.EY is adopting Nemotron Nano 2, given its high throughput, to support agentic AI in large organizations for tax, risk management and finance use cases.NetApp is currently testing Nemotron reasoning models so that AI agents can search and analyze business dataDataRobot is working with Nemotron models for its Agent Workforce Platform for end-to-end agent lifecycle management.Tabnine is working with Nemotron models for suggesting and automating coding tasks on behalf of developers.Automation Anywhere, CrewAI and Dataiku are among the additional agentic AI software developers integrating Nemotron models into their platforms.Leading companies across transportation, safety and AI intelligence are using Cosmos Reason to advance autonomous driving, video analytics, and road and workplace safety.Uber is exploring Cosmos Reason to analyze autonomous vehicle behavior. In addition, Uber is post-training Cosmos Reason to summarize visual data and analyze scenarios like pedestrians walking across highways to perform quality analysis and inform autonomous driving behavior.Cosmos Reason can also serve as the brain of autonomous vehicles. It lets robots interpret environments and, given complex commands, break them down into tasks and execute them using common sense, even in unfamiliar environments.Centific is testing Cosmos Reason to enhance its AI-powered video intelligence platform. The VLM enables the platform to process complex video data into actionable insights, helping reduce false positives and improve decision-making efficiency.VAST is advancing real-time urban intelligence using NVIDIA Cosmos Reason with its AI operating system to process massive video streams at scale. With the VSS Blueprint, VAST can build agents that can identify incidents and trigger responses, turning video streams and metadata into actionable, proactive public safety tools.Ambient.ai is working with Cosmos Reasons temporal, physics-aware reasoning, to enable automated detection of missing personal protection equipment and monitoring of hazardous conditions, helping enhance environmental health and safety across construction, manufacturing, logistics and other industrial settings.Magna is developing with Cosmos Reason as part of its City Delivery Platform a fully autonomous, low-cost solution for instant delivery to help vehicles adapt more quickly to new cities. The model adds world understanding to the vehicles long-term trajectory planning.These models are expected to be available as NVIDIA NIM microservices for secure, reliable deployment on any NVIDIA-accelerated infrastructure for maximum privacy and control. They are planned to be available soon through Amazon Bedrock and Amazon SageMaker AI for Nemotron models, as well as through Azure AI Foundry, Oracle Data Science Platform and Google Vertex AI.Try Cosmos Reason on build.nvidia.com or download it from Hugging Face or GitHub.Nemotron Nano 2 and Llama Nemotron Super 1.5 (NVFP4) will be available soon for download. Meanwhile, learn more about Nemotron models and download previous versions.Download the Llama Nemotron VLM Dataset v1 from Hugging Face.Watch the NVIDIA Research special address at SIGGRAPH and learn more about how graphics and simulation innovations come together to drive industrial digitalization by joining NVIDIA at the conference, running through Thursday, Aug. 14.See notice regarding software product information.
    Like
    Love
    Wow
    Angry
    72
    · 0 Commentaires ·0 Parts ·0 Aperçu
  • 0 Commentaires ·0 Parts ·0 Aperçu
  • Making Safer Spaces: NVIDIA and Partners Bring Physical AI to Cities and Industrial Infrastructure
    blogs.nvidia.com
    Physical AI is becoming the foundation of smart cities, facilities and industrial processes across the globe.NVIDIA is working with companies including Accenture, Avathon, Belden, DeepHow, Milestone Systems and Telit Cinterion to enhance operations across the globe with physical AI-based perception and reasoning.The continuous loop of simulating, training and deploying physical AI offers sophisticated industrial automation capabilities, making cities and infrastructure safer, smarter and more efficient.For example, physical AI applications can automate potentially dangerous tasks for workers, such as working with heavy machinery. Physical AI can also improve transportation services and public safety, detect defective products in factories and more.The need for this is greater than ever. The numbers tell the story:Infrastructure that can perceive, reason and act relies on video sensors and the latest vision AI capabilities. Using the NVIDIA Metropolis platform which simplifies the development, deployment and scaling of video analytics AI agents and services from the edge to the cloud developers can build visual perception into their facilities faster to enhance productivity and improve safety across environments.Below are five leading companies advancing physical AI and five key NVIDIA Metropolis updates, announced today at the SIGGRAPH computer graphics conference, making such advancements possible.Five Companies Advancing Physical AIGlobal professional services company Accenture is collaborating with Belden, a leading provider of complete connection solutions, to enhance worker safety by creating smart virtual fences that factories can place around large robots to prevent accidents with human operators.Image courtesy of Accenture and Belden.The smart virtual fence is a physical AI safety system that uses an OpenUSD-based digital twin and physics-grounded simulation to model complex industrial environments. Using computer vision-based mapping and 3D spatial intelligence, the system is adaptive to increased variability in the dynamic human-robot interactions that occur in a modern shopfloor environment.Accenture taps into the NVIDIA Omniverse platform and Metropolis to build and simulate these smart fences. With Omniverse, Accenture created a digital twin of a robot arm and workers moving in a space. And with Metropolis, the company trained its AI models and deployed them at the edge with video ingestion and the NVIDIA DeepStream software development kit (SDK)s real-time inference capabilities.Avathon, an industrial automation platform provider, uses the NVIDIA Blueprint for video search and summarization (VSS), part of NVIDIA Metropolis, to provide manufacturing and energy facilities with real-time insights that improve operational efficiency and worker safety.Reliance British Petroleum Mobility Limited, a leader in Indias fuel and mobility sector, used the Avathon video intelligence product during the construction of its gas stations to achieve higher standards of safety compliance, a reduction in safety noncompliance incidents and higher productivity by saving thousands of work hours.DeepHow has developed a Smart Know-How Companion for employees in manufacturing and other industries. The companion uses the Metropolis VSS blueprint to transform key workflows into bite-sized, multilingual videos and digital instructions, improving onboarding, safety and floor operator efficiency.Facing upskilling needs and retiring skilled workers, beverage company Anheuser-Busch InBev turned to the DeepHow platform to convert standard operating procedures into easy-to-understand visual guides. This has slashed onboarding time by 80%, boosted training consistency and improved long-term knowledge retention for employees.Milestone Systems, which offers one of the worlds largest platforms for managing IP video sensor data in complex industrial and city deployments, is creating the worlds largest real-world computer vision data library through its platform, Project Hafnia. Among its capabilities, the platform provides physical AI developers with access to customized vision language models (VLMs).Tapping NVIDIA NeMo Curator, Milestone Systems built a VLM fine-tuned for intelligent transportation systems for use within the VSS blueprint to help develop AI agents that better manage city roadways. Milestone Systems is also looking to use the new open, customizable NVIDIA Cosmos Reason VLM for physical AI.Internet-of-things company Telit Cinterion has integrated NVIDIA TAO Toolkit 6 into its AI-powered visual inspection platform, which uses vision foundation models like FoundationPose, alongside other NVIDIA models, to support multimodal AI and deliver high-performance inferencing. TAO brings low-code AI capabilities to the Telit platform, enabling manufacturers to quickly develop and deploy accurate, custom AI models for defect detection and quality control.Five NVIDIA Metropolis Updates for Physical AIKey updates to NVIDIA Metropolis are enhancing developers capabilities to build physical AI applications more quickly and easily:Cosmos Reason VLMhttps://blogs.nvidia.com/wp-content/uploads/2025/08/turbine-cosmos-reasoning.mp4The latest version of Cosmos Reason NVIDIAs advanced open, customizable, 7-billion-parameter reasoning VLM for physical AI enables contextual video understanding, temporal event reasoning for Metropolis use cases. Its compact size makes it easy to deploy from edge to cloud and ideal for automating traffic monitoring, public safety, visual inspection and intelligent decision-making.VSS Blueprint 2.4VSS 2.4 makes it easy to quickly augment existing vision AI applications with Cosmos Reason and deliver powerful new features to smart infrastructure. An expanded set of application programming interfaces in the blueprint offers users direct more flexibility in choosing specific VSS components and capabilities to augment computer vision pipelines with generative AI.New Vision Foundation ModelsThe NVIDIA TAO Toolkit includes a new suite of vision foundation models, along with advanced fine-tuning methods, self-supervised learning and knowledge distillation capabilities, to optimize deployment of physical AI solutions across edge and cloud environments. The NVIDIA DeepStream SDK includes a new Inference Builder to enable seamless deployment of TAO 6 models.Companies around the world including Advex AI, Instrumental AI and Spingence are experimenting with these new models and NVIDIA TAO to build intelligent solutions that optimize industrial operations and drive efficiency.NVIDIA Isaac Sim ExtensionsNew extensions in the NVIDIA Isaac Sim reference application help solve common challenges in vision AI development such as limited labeled data and rare edge-case scenarios. These tools simulate human and robot interactions, generate rich object-detection datasets, and create incident-based scenes and image-caption pairs to train VLMs, accelerating development and improving AI performance in real-world conditions.Expanded Hardware SupportAll of these Metropolis components can now run on NVIDIA RTX PRO 6000 Blackwell GPUs, the NVIDIA DGX Spark desktop supercomputer and the NVIDIA Jetson Thor platform for physical AI and humanoid robotics so users can develop and deploy from the edge to the cloud.Cosmos Reason 1 and NVIDIA TAO 6.0 are now available for download. Sign up to be alerted when VSS 2.4, the Cosmos Reason VLM fine-tuning update and NVIDIA DeepStream 8.0 become available.Watch the NVIDIA Research special address at SIGGRAPH and learn more about how graphics and simulation innovations come together to drive industrial digitalization by joining NVIDIA at the conference, running through Thursday, Aug. 14.See notice regarding software product information.
    0 Commentaires ·0 Parts ·0 Aperçu
  • The Saga Continues: Stream 2Ks Mafia: The Old Country at Launch on GeForce NOW
    blogs.nvidia.com
    This GFN Thursday brings an offer members cant refuse 2Ks highly anticipated Mafia: The Old Country is launching in the cloud today.The prequel to the award-winning action-adventure Mafia series leads five games joining the cloud this week, including the long-awaited launch of Stormgate 1.0 and early access for THQ Nordics Titan Quest II.And dont miss out on the newest season of Marvel Rivals, available for members to stream without having to wait for downloads or patches.The Don Says PlayFuh-get about it.Return to the heart of the Mafia saga and journey into the brutal underworld of 1900s Sicily with Mafia: The Old Country. Hustle and scheme through savage streets and sun-baked villages in a gripping, linear narrative. Experience a world where every choice cuts deep, every vendetta matters and loyalty is a dangerous game.Step into the scuffed shoes of Enzo Favara a man with nothing to lose and even less to trust and earn a seat at Don Torrisis table. Whether its settling vendettas with a stiletto blade or dueling with a lupara shotgun, every fight is life or death, and every bullet counts. Players can expect gritty stealth, messy brawls and nerve-wracking shootouts fueled by loyalty, betrayal and the need to prove more value to the Torrisi family alive than dead.Mafia: The Old Country joins the Mafia trilogy already streaming on GeForce NOW for members to relive the timeless action before diving into Enzos origins. With GeForce NOW, members get instant access to their games no need to worry about waiting for massive downloads so they can dive right into the action whenever the family calls.Get MythicalLoot first, ask the gods later.Titan Quest 2 plunges adventurers headlong into the heart of Greek mythology where meddling gods, prowling monsters and unfolding heroics shape a vibrant, handcrafted world.In this classic action role-playing game, chaos abounds as champions dodge harpies, clash with centaurs and roam sun-drenched vistas while honing unique masteries. Each skirmish with mythic foes demands quick thinking and strategy, while the worlds hidden groves and ancient ruins tempt explorers with powerful loot. Curiosity and courage unlock secret quests and treasures, ensuring every detour leads to fresh challenges and new rewards.GeForce NOW lets the adventure unfold on just about any device no need for fancy hardware. With high-performance servers handling the heavy lifting, the gameplay stays smooth, the visuals sharp and the action responsive even on a basic laptop, an old PC or a phone. Its the easiest way to jump straight into the mythical chaos, no matter where the quest begins.The Multiverse Gets a ShakeupNew synergies, wild chaos whos your rival now?Marvel Rivals Season 3.5 crashes onto the scene tomorrow, Aug. 8, unleashing the Daywalker himself Blade, the vampire hunter along with a pulse-raising mix of meta-shifting updates.Players can sink their teeth into combat with the debut of the life-stealing Duelist, while the new Resource Rumble mode shakes up the competitive meta and sends teams scrambling for new strategies. Team-Ups have been overhauled: old combos are gone, making way for wild new synergies and fresh abilities that put every hero on notice. Iconic heroes and villains are flexing new changes: Doctor Strange amps up his mystic firepower, Groot branches out with bigger reach and Magnetos shields get a little less magnetic, to name a few.Jump into the chaos with zero downloads, buttery-smooth performance and instant upgrades. GeForce NOW lets members unleash their inner hero (or villain) on any device, anywhere, so get right to swinging, blasting and smashing through the ever-evolving Marvel Rivals multiverse.Lets Play TodayStrategy with a side of sass.Buckle up, commander the Stormgate 1.0 release is now available to stream in the cloud.Step into a delightfully chaotic future where humanity and demonic invaders wage war with laser cannons and mechs all with a wry sense of humor. In this sleek reimagining of classic real-time strategy, Stormgate offers a 14-mission campaign with hero leveling, quirky banter in a new mothership hub and three fantastically distinct factions to master. Whether nuking demons solo, clashing on the ranked 1v1 ladder or wrangling friends for co-op mayhem, Stormgates lighthearted charm and competitive spirit will delight newcomers and veterans alike.In addition, members can look for the following releases:Mafia: The Old Country (New release on Steam, Aug. 7)Assassins Creed Mirage (Now available on PC Game Pass, Aug. 7)Amnesia: The Dark Descent (Epic Games Store)HUNTERHUNTER NENIMPACT (Steam)Titan Quest II (Steam)Learn how to stream supported Ubisoft games from PC Game Pass on GeForce NOW, including this weeks addition of Assassins Creed Mirage on PC Game Pass.What are you planning to play this weekend? Let us know on X or in the comments below.It's Work Like a Dog Daywhat game do you sweat in like it's your full-time job? NVIDIA GeForce NOW (@NVIDIAGFN) August 5, 2025
    0 Commentaires ·0 Parts ·0 Aperçu
  • No Backdoors. No Kill Switches. No Spyware.
    blogs.nvidia.com
    NVIDIA GPUs are at the heart of modern computing. Theyre used across industries from healthcare and finance to scientific research, autonomous systems and AI infrastructure. NVIDIA GPUs are embedded into CT scanners and MRI machines, DNA sequencers, air-traffic radar tracking systems, city traffic-management systems, self-driving cars, supercomputers, TV broadcasting systems, casino machines and game consoles.To mitigate the risk of misuse, some pundits and policymakers propose requiring hardware kill switches or built-in controls that can remotely disable GPUs without user knowledge and consent. Some suspect they might already exist.NVIDIA GPUs do not and should not have kill switches and backdoors.Hard-Coded, Single-Point Controls Are Always a Bad IdeaNVIDIA has been designing processors for over 30 years. Embedding backdoors and kill switches into chips would be a gift to hackers and hostile actors. It would undermine global digital infrastructure and fracture trust in U.S. technology. Established law wisely requires companies to fix vulnerabilities not create them.Until recently, that policy was universally held and beyond question. When security researchers discovered vulnerabilities such as Spectre and Meltdown for CPUs, governments and industry responded with speed and unity to eliminate the risk.That principle still holds. There is no such thing as a good secret backdoor only dangerous vulnerabilities that need to be eliminated. Product security must always be done the right way: through rigorous internal testing, independent validation and full compliance with global cybersecurity standards. Robust security is built on the principle of defense in depth: layering multiple safeguards so that no single-point vulnerability can compromise or shut down a system. For decades, thats how NVIDIA and American industry have promoted innovation while protecting users and growing the economy. This is no time to depart from that winning formula.Historical Lessons: The Clipper Chip Debacle a Policy and Technical FailureThe cybersecurity community learned these lessons the hard way during the 1990s with the NSAs Clipper Chip initiative. Introduced in 1993, the Clipper Chip was designed to provide strong encryption while maintaining government backdoor access through a key escrow system.The Clipper Chip represented everything wrong with built-in backdoors. Security researchers discovered fundamental flaws in the system that could allow malicious parties to tamper with the software. It created centralized vulnerabilities that could be exploited by adversaries. The mere existence of government backdoors undermined user confidence in the security of systems.Kill switches and built-in backdoors create single points of failure and violate the fundamental principles of cybersecurity.Promote Smart Software Tools, Not Dangerous Hardware TrapsSome point to smartphone features like find my phone or remote wipe as models for a GPU kill switch. That comparison doesnt hold water optional software features, controlled by the user, are not hardware backdoors.NVIDIA has always supported open, transparent software that helps customers get the most from their GPU-powered systems diagnostics, performance monitoring, bug reporting and timely patching with the users knowledge and consent. Thats responsible, secure computing. It helps our customers excel, and industry stay ahead.Hardwiring a kill switch into a chip is something entirely different: a permanent flaw beyond user control, and an open invitation for disaster. Its like buying a car where the dealership keeps a remote control for the parking brake just in case they decide you shouldnt be driving. Thats not sound policy. Its an overreaction that would irreparably harm Americas economic and national security interests.Hardware Integrity Should Be Nonpartisan and NonnegotiableFor decades, policymakers have championed industrys efforts to create secure, trustworthy hardware. Governments have many tools to protect nations, consumers and the economy. Deliberately weakening critical infrastructure should never be one of them.There are no back doors in NVIDIA chips. No kill switches. No spyware. Thats not how trustworthy systems are built and never will be.
    0 Commentaires ·0 Parts ·0 Aperçu
  • OpenAI and NVIDIA Propel AI Innovation With New Open Models Optimized for the Worlds Largest AI Inference Infrastructure
    blogs.nvidia.com
    Two new open-weight AI reasoning models from OpenAI released today bring cutting-edge AI development directly into the hands of developers, enthusiasts, enterprises, startups and governments everywhere across every industry and at every scale.NVIDIAs collaboration with OpenAI on these open models gpt-oss-120b and gpt-oss-20b is a testament to the power of community-driven innovation and highlights NVIDIAs foundational role in making AI accessible worldwide.Anyone can use the models to develop breakthrough applications in generative, reasoning and physical AI, healthcare and manufacturing or even unlock new industries as the next industrial revolution driven by AI continues to unfold.OpenAIs new flexible, open-weight text-reasoning large language models (LLMs) were trained on NVIDIA H100 GPUs and run inference best on the hundreds of millions of GPUs running the NVIDIA CUDA platform across the globe.With software optimizations for the NVIDIA Blackwell platform, the models offer optimal inference on NVIDIA GB200 NVL72 systems, achieving 1.5 million tokens per second driving massive efficiency for inference.OpenAI showed the world what could be built on NVIDIA AI and now theyre advancing innovation in open-source software, said Jensen Huang, founder and CEO of NVIDIA. The gpt-oss models let developers everywhere build on that state-of-the-art open-source foundation, strengthening U.S. technology leadership in AI all on the worlds largest AI compute infrastructure.NVIDIA Blackwell Delivers Advanced ReasoningAs advanced reasoning models like gpt-oss generate exponentially more tokens, the demand on compute infrastructure increases dramatically. Meeting this demand calls for purpose-built AI factories powered by NVIDIA Blackwell, an architecture designed to deliver the scale, efficiency and return on investment required to run inference at the highest level.NVIDIA Blackwell includes innovations such as NVFP4 4-bit precision, which enables ultra-efficient, high-accuracy inference while significantly reducing power and memory requirements. This makes it possible to deploy trillion-parameter LLMs in real time, which can unlock billions of dollars in value for organizations.Open Development for Millions of AI Builders WorldwideNVIDIA CUDA is the worlds most widely available computing infrastructure, letting users deploy and run AI models anywhere, from the powerful NVIDIA DGX Cloud platform to NVIDIA GeForce RTX and NVIDIA RTX PRO-powered PCs and workstations.There are over 450 million NVIDIA CUDA downloads to date, and starting today, the massive community of CUDA developers gains access to these latest models, optimized to run on the NVIDIA technology stack they already use.Demonstrating their commitment to open-sourcing software, OpenAI and NVIDIA have collaborated with top open framework providers to provide model optimizations for FlashInfer, Hugging Face, llama.cpp, Ollama and vLLM, in addition to NVIDIA Tensor-RT LLM and other libraries, so developers can build with their framework of choice.A History of Collaboration, Building on Open SourceTodays model releases underscore how NVIDIAs full-stack approach helps bring the worlds most ambitious AI projects to the broadest user base possible.Its a story that goes back to the earliest days of NVIDIAs collaboration with OpenAI, which began in 2016 when Huang hand-delivered the first NVIDIA DGX-1 AI supercomputer to OpenAIs headquarters in San Francisco.Since then, the companies have been working together to push the boundaries of whats possible with AI, providing the core technologies and expertise needed for massive-scale training runs.And by optimizing OpenAIs gpt-oss models for NVIDIA Blackwell and RTX GPUs, along with NVIDIAs extensive software stack, NVIDIA is enabling faster, more cost-effective AI advancements for its 6.5 million developers across 250 countries using 900+ NVIDIA software development kits and AI models and counting.Learn more by reading the NVIDIA Technical Blog and latest installment of the NVIDIA RTX AI Garage blog series.
    0 Commentaires ·0 Parts ·0 Aperçu
  • OpenAIs New Open Models Accelerated Locally on NVIDIA GeForce RTX and RTX PRO GPUs
    blogs.nvidia.com
    In collaboration with OpenAI, NVIDIA has optimized the companys new open-source gpt-oss models for NVIDIA GPUs, delivering smart, fast inference from the cloud to the PC. These new reasoning models enable agentic AI applications such as web search, in-depth research and many more.With the launch of gpt-oss-20b and gpt-oss-120b, OpenAI has opened cutting-edge models to millions of users. AI enthusiasts and developers can use the optimized models on NVIDIA RTX AI PCs and workstations through popular tools and frameworks like Ollama, llama.cpp and Microsoft AI Foundry Local, and expect performance of up to 256 tokens per second on the NVIDIA GeForce RTX 5090 GPU.OpenAI showed the world what could be built on NVIDIA AI and now theyre advancing innovation in open-source software, said Jensen Huang, founder and CEO of NVIDIA. The gpt-oss models let developers everywhere build on that state-of-the-art open-source foundation, strengthening U.S. technology leadership in AI all on the worlds largest AI compute infrastructure.The models release highlights NVIDIAs AI leadership from training to inference and from cloud to AI PC.Open for All Both gpt-oss-20b and gpt-oss-120b are flexible, open-weight reasoning models with chain-of-thought capabilities and adjustable reasoning effort levels using the popular mixture-of-experts architecture. The models are designed to support features like instruction-following and tool use, and were trained on NVIDIA H100 GPUs. AI developers can learn more and get started using instructions from the NVIDIA Technical Blog.These models can support up to 131,072 context lengths, among the longest available in local inference. This means the models can reason through context problems, ideal for tasks such as web search, coding assistance, document comprehension and in-depth research.The OpenAI open models are the first MXFP4 models supported on NVIDIA RTX. MXFP4 allows for high model quality, offering fast, efficient performance while requiring fewer resources compared with other precision types.Run the OpenAI Models on NVIDIA RTX With OllamaThe easiest way to test these models on RTX AI PCs, on GPUs with at least 24GB of VRAM, is using the new Ollama app. Ollama is popular with AI enthusiasts and developers for its ease of integration, and the new user interface (UI) includes out-of-the-box support for OpenAIs open-weight models. Ollama is fully optimized for RTX, making it ideal for consumers looking to experience the power of personal AI on their PC or workstation.Once installed, Ollama enables quick, easy chatting with the models. Simply select the model from the dropdown menu and send a message. Because Ollama is optimized for RTX, there are no additional configurations or commands required to ensure top performance on supported GPUs.Testing OpenAIs open models in Ollama is easy.Ollamas new app includes other new features, like easy support for PDF or text files within chats, multimodal support on applicable models so users can include images in their prompts, and easily customizable context lengths when working with large documents or chats.Developers can also use Ollama via command line interface or the apps software development kit (SDK) to power their applications and workflows.Other Ways to Use the New OpenAI Models on RTXEnthusiasts and developers can also try the gpt-oss models on RTX AI PCs through various other applications and frameworks, all powered by RTX, on GPUs that have at least 16GB of VRAM.NVIDIA continues to collaborate with the open-source community on both llama.cpp and the GGML tensor library to optimize performance on RTX GPUs. Recent contributions include implementing CUDA Graphs to reduce overhead and adding algorithms that reduce CPU overheads. Check out the llama.cpp GitHub repository to get started.Overall performance of the gpt-oss-20b model on various RTX AI PCs.Windows developers can also access OpenAIs new models via Microsoft AI Foundry Local, currently in public preview. Foundry Local is an on-device AI inferencing solution that integrates into workflows via the command line, SDK or application programming interfaces. Foundry Local uses ONNX Runtime, optimized through CUDA, with support for NVIDIA TensorRT for RTX coming soon. Getting started is easy: install Foundry Local and invoke Foundry model run gpt-oss-20b in a terminal.The release of these open-source models kicks off the next wave of AI innovation from enthusiasts and developers looking to add reasoning to their AI-accelerated Windows applications.Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, productivity apps and more on AI PCs and workstations. Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X and stay informed by subscribing to the RTX AI PC newsletter. Join NVIDIAs Discord server to connect with community developers and AI enthusiasts for discussions on whats possible with RTX AI.Follow NVIDIA Workstation on LinkedIn and X. See notice regarding software product information.
    0 Commentaires ·0 Parts ·0 Aperçu
  • Embark on Epic Adventures in August With a Dozen New Games Coming to GeForce NOW
    blogs.nvidia.com
    August brings new levels of gaming excitement on GeForce NOW, with 2,300 titles now available to stream in the cloud.Grab a controller and get ready for epic adventures nearly a dozen new games are coming to the cloud this month. Each week brings fresh titles for members to discover, stream and conquer.Get ready to shrink down for big fun early access for Grounded 2, announced as a surprise in June at XBOX Games Showcase, will be available to stream on day one: tomorrow, Aug. 1.Plus, finish off July with the eight games available to stream this week, alongside the latest update for the acclaimed open-world action role-playing game Genshin Impact.The Park Is CallingOops, shrunken again.Step into the wild, unpredictable world of Brookhollow Park Grounded 2 is launching in early access in the cloud, and the backyard has never felt bigger or more alive.Shrink to the size of an ant, navigate towering blades of grass, uncover hidden secrets and face off against bugs that can be friend or foe. Craft weapons and armor from scavenged materials, build creative bases and raise Buggie companions to help explore and survive. Whether playing solo or with friends, every day brings new challenges and surprises. Grounded 2s quirky humor and inventive gameplay make every moment an adventure, from dodging predators to racing Buggies through underground tunnels.Experience every detail rendered in stunning quality streaming from a GeForce RTX-powered gaming rig in the cloud. Jump into the game instantly no downloads, installs or waiting. Whether on a low-powered laptop, Mac or mobile device, GeForce NOW brings the full power of high-end gaming hardware to members through the cloud.Arriving in AugustWelcome to the Easybreeze Holiday Resort.Genshin Impact Version 5.8, Sunspray Summer Resort, is now live in the cloud, letting players dive in instantly no updates needed. Explore Natlans new Easybreeze Holiday Resort, filled with events, mini games and new exploration features. Meet Ineffa, the first Nod-Krai five-star electro polearm user, alongside returning favorites like Bennett and Venti, in a fresh Story Quest. The update includes limited-time rewards, an exclusive Claymore, new outfits and gameplay improvements. It also marks the start of the Song of the Welkin Moon storyline, expanding Natlan and Nod-Krai lore.Catch the games ready to play today:Grounded 2 (New release on Steam and Xbox, available on PC Game Pass, July 29)Achilles: Survivor (New release on Steam, July 29)Frosthaven (New release on Steam, July 31)Dead Take (New release on Steam, July 31)Farming Simulator 25 (New release on Xbox, available on PC Game Pass, Aug. 1)High on Life (Xbox, available on PC Game Pass)The King Is Watching (Steam)Ultimate Chicken Horse (Xbox, available on PC Game Pass)Catch the full list of games coming to the cloud in August:Mafia: The Old Country (New release on Steam, Aug. 8)Echoes of the End (New release on Steam, Aug. 12)Supraworld (New release on Steam, Aug. 15)VOID/BREAKER (New release on Steam, Aug. 20)Aztecs: The Last Sun (New release on Steam, Aug. 28)Among Us 3D (Steam)Funko Fusion (Steam)Field of Glory II: Medieval (Steam)The Rogue Prince of Persia (New release on Ubisoft)No Sleep for Kaname Date From AI: THE SOMNIUM FILES (Steam)Say Bye to JulyIn addition to the 20 games announced in July, an extra 15 joined over the month, including Achilles: Survivor, Frosthaven and Dead Take, newly added this week:Abiotic Factor (Xbox, available on PC Game Pass)Barony (Steam)BitCraft Online (Steam)He Is Coming (Xbox, available on PC Game Pass)Humanity (Xbox, available on the Microsoft store)Legion TD 2 (Epic Games Store)PEAK (Steam)RoboCop: Rogue City (New release on Xbox)SteamWorld Dig (Steam)SUPERVIVE (Steam)Tony Hawks Pro Skater 1 + 2 (Steam, Epic Games Store and Xbox, available on the Microsoft Store)WUCHANG: Fallen Feathers (Xbox, available on PC Game Pass)Call of Duty: Modern Warfare II and Call of Duty: Modern Warfare III were moved from the main Call of Duty HQ launcher and are now standalone downloads. Both games will be available to stream on GeForce NOW in the coming days. Check out this knowledgebase article for more information.What are you planning to play this weekend? Let us know on X or in the comments below. The cloud remembers everything.What was the very first game you streamed on GeForce NOW? NVIDIA GeForce NOW (@NVIDIAGFN) July 29, 2025
    0 Commentaires ·0 Parts ·0 Aperçu
Plus de lecture
CGShares https://cgshares.com