• Riot Will Allow Sports-Betting Sponsorships For League Of Legends Esports Teams

    Riot Games has announced that it will begin officially sanctioning sports-betting sponsorships for esports teams in its Tier 1 League of Legends and Valorant leagues. While the company states that it still won't allow advertisements in its official broadcasts, teams themselves will be able to take money from sports-betting companies for advertising through their own channels.In a blog post, President of Publishing and Esports John Needham writes that the move is designed to take advantage of the rapidly growing sports-betting industry and to make esports-related betting more regulated. Seemingly to address concerns and head off potential criticism, Needham explains that the company is authorizing sports-betting sponsorships under a "guardrails first" strategy.These "guardrails," Needham states, are essentially the rules by which any sponsorship must be executed. First, sports-betting companies need to be vetted and approved by Riot itself, although the company has not shared the criteria on which this vetting is done. Second, to ensure that sports-betting companies are on a level playing field, Riot is mandating that official partners all use GRID, the officially sanctioned data platform for League of Legends and Valorant. Third, esports teams must launch and maintain internal integrity programs to protect against violations of league rules due to the influence of sports betting. Fourth and last, Riot will use some of the revenue from these sponsorships to support its Tier 2esports leagues.Continue Reading at GameSpot
    #riot #will #allow #sportsbetting #sponsorships
    Riot Will Allow Sports-Betting Sponsorships For League Of Legends Esports Teams
    Riot Games has announced that it will begin officially sanctioning sports-betting sponsorships for esports teams in its Tier 1 League of Legends and Valorant leagues. While the company states that it still won't allow advertisements in its official broadcasts, teams themselves will be able to take money from sports-betting companies for advertising through their own channels.In a blog post, President of Publishing and Esports John Needham writes that the move is designed to take advantage of the rapidly growing sports-betting industry and to make esports-related betting more regulated. Seemingly to address concerns and head off potential criticism, Needham explains that the company is authorizing sports-betting sponsorships under a "guardrails first" strategy.These "guardrails," Needham states, are essentially the rules by which any sponsorship must be executed. First, sports-betting companies need to be vetted and approved by Riot itself, although the company has not shared the criteria on which this vetting is done. Second, to ensure that sports-betting companies are on a level playing field, Riot is mandating that official partners all use GRID, the officially sanctioned data platform for League of Legends and Valorant. Third, esports teams must launch and maintain internal integrity programs to protect against violations of league rules due to the influence of sports betting. Fourth and last, Riot will use some of the revenue from these sponsorships to support its Tier 2esports leagues.Continue Reading at GameSpot #riot #will #allow #sportsbetting #sponsorships
    WWW.GAMESPOT.COM
    Riot Will Allow Sports-Betting Sponsorships For League Of Legends Esports Teams
    Riot Games has announced that it will begin officially sanctioning sports-betting sponsorships for esports teams in its Tier 1 League of Legends and Valorant leagues. While the company states that it still won't allow advertisements in its official broadcasts, teams themselves will be able to take money from sports-betting companies for advertising through their own channels.In a blog post, President of Publishing and Esports John Needham writes that the move is designed to take advantage of the rapidly growing sports-betting industry and to make esports-related betting more regulated. Seemingly to address concerns and head off potential criticism, Needham explains that the company is authorizing sports-betting sponsorships under a "guardrails first" strategy.These "guardrails," Needham states, are essentially the rules by which any sponsorship must be executed. First, sports-betting companies need to be vetted and approved by Riot itself, although the company has not shared the criteria on which this vetting is done. Second, to ensure that sports-betting companies are on a level playing field, Riot is mandating that official partners all use GRID, the officially sanctioned data platform for League of Legends and Valorant. Third, esports teams must launch and maintain internal integrity programs to protect against violations of league rules due to the influence of sports betting. Fourth and last, Riot will use some of the revenue from these sponsorships to support its Tier 2 (lower division) esports leagues.Continue Reading at GameSpot
    0 Comments 0 Shares
  • Plug and Play: Build a G-Assist Plug-In Today

    Project G-Assist — available through the NVIDIA App — is an experimental AI assistant that helps tune, control and optimize NVIDIA GeForce RTX systems.
    NVIDIA’s Plug and Play: Project G-Assist Plug-In Hackathon — running virtually through Wednesday, July 16 — invites the community to explore AI and build custom G-Assist plug-ins for a chance to win prizes and be featured on NVIDIA social media channels.

    G-Assist allows users to control their RTX GPU and other system settings using natural language, thanks to a small language model that runs on device. It can be used from the NVIDIA Overlay in the NVIDIA App without needing to tab out or switch programs. Users can expand its capabilities via plug-ins and even connect it to agentic frameworks such as Langflow.
    Below, find popular G-Assist plug-ins, hackathon details and tips to get started.
    Plug-In and Win
    Join the hackathon by registering and checking out the curated technical resources.
    G-Assist plug-ins can be built in several ways, including with Python for rapid development, with C++ for performance-critical apps and with custom system interactions for hardware and operating system automation.
    For those that prefer vibe coding, the G-Assist Plug-In Builder — a ChatGPT-based app that allows no-code or low-code development with natural language commands — makes it easy for enthusiasts to start creating plug-ins.
    To submit an entry, participants must provide a GitHub repository, including source code file, requirements.txt, manifest.json, config.json, a plug-in executable file and READme code.
    Then, submit a video — between 30 seconds and two minutes — showcasing the plug-in in action.
    Finally, hackathoners must promote their plug-in using #AIonRTXHackathon on a social media channel: Instagram, TikTok or X. Submit projects via this form by Wednesday, July 16.
    Judges will assess plug-ins based on three main criteria: 1) innovation and creativity, 2) technical execution and integration, reviewing technical depth, G-Assist integration and scalability, and 3) usability and community impact, aka how easy it is to use the plug-in.
    Winners will be selected on Wednesday, Aug. 20. First place will receive a GeForce RTX 5090 laptop, second place a GeForce RTX 5080 GPU and third a GeForce RTX 5070 GPU. These top three will also be featured on NVIDIA’s social media channels, get the opportunity to meet the NVIDIA G-Assist team and earn an NVIDIA Deep Learning Institute self-paced course credit.
    Project G-Assist requires a GeForce RTX 50, 40 or 30 Series Desktop GPU with at least 12GB of VRAM, Windows 11 or 10 operating system, a compatible CPU, specific disk space requirements and a recent GeForce Game Ready Driver or NVIDIA Studio Driver.
    Plug-InExplore open-source plug-in samples available on GitHub, which showcase the diverse ways on-device AI can enhance PC and gaming workflows.

    Popular plug-ins include:

    Google Gemini: Enables search-based queries using Google Search integration and large language model-based queries using Gemini capabilities in real time without needing to switch programs from the convenience of the NVIDIA App Overlay.
    Discord: Enables users to easily share game highlights or messages directly to Discord servers without disrupting gameplay.
    IFTTT: Lets users create automations across hundreds of compatible endpoints to trigger IoT routines — such as adjusting room lights and smart shades, or pushing the latest gaming news to a mobile device.
    Spotify: Lets users control Spotify using simple voice commands or the G-Assist interface to play favorite tracks and manage playlists.
    Twitch: Checks if any Twitch streamer is currently live and can access detailed stream information such as titles, games, view counts and more.

    Get G-Assist 
    Join the NVIDIA Developer Discord channel to collaborate, share creations and gain support from fellow AI enthusiasts and NVIDIA staff.
    the date for NVIDIA’s How to Build a G-Assist Plug-In webinar on Wednesday, July 9, from 10-11 a.m. PT, to learn more about Project G-Assist capabilities, discover the fundamentals of building, testing and deploying Project G-Assist plug-ins, and participate in a live Q&A session.
    Explore NVIDIA’s GitHub repository, which provides everything needed to get started developing with G-Assist, including sample plug-ins, step-by-step instructions and documentation for building custom functionalities.
    Learn more about the ChatGPT Plug-In Builder to transform ideas into functional G-Assist plug-ins with minimal coding. The tool uses OpenAI’s custom GPT builder to generate plug-in code and streamline the development process.
    NVIDIA’s technical blog walks through the architecture of a G-Assist plug-in, using a Twitch integration as an example. Discover how plug-ins work, how they communicate with G-Assist and how to build them from scratch.
    Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations. 
    Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter.
    Follow NVIDIA Workstation on LinkedIn and X. 
    See notice regarding software product information.
    #plug #play #build #gassist #plugin
    Plug and Play: Build a G-Assist Plug-In Today
    Project G-Assist — available through the NVIDIA App — is an experimental AI assistant that helps tune, control and optimize NVIDIA GeForce RTX systems. NVIDIA’s Plug and Play: Project G-Assist Plug-In Hackathon — running virtually through Wednesday, July 16 — invites the community to explore AI and build custom G-Assist plug-ins for a chance to win prizes and be featured on NVIDIA social media channels. G-Assist allows users to control their RTX GPU and other system settings using natural language, thanks to a small language model that runs on device. It can be used from the NVIDIA Overlay in the NVIDIA App without needing to tab out or switch programs. Users can expand its capabilities via plug-ins and even connect it to agentic frameworks such as Langflow. Below, find popular G-Assist plug-ins, hackathon details and tips to get started. Plug-In and Win Join the hackathon by registering and checking out the curated technical resources. G-Assist plug-ins can be built in several ways, including with Python for rapid development, with C++ for performance-critical apps and with custom system interactions for hardware and operating system automation. For those that prefer vibe coding, the G-Assist Plug-In Builder — a ChatGPT-based app that allows no-code or low-code development with natural language commands — makes it easy for enthusiasts to start creating plug-ins. To submit an entry, participants must provide a GitHub repository, including source code file, requirements.txt, manifest.json, config.json, a plug-in executable file and READme code. Then, submit a video — between 30 seconds and two minutes — showcasing the plug-in in action. Finally, hackathoners must promote their plug-in using #AIonRTXHackathon on a social media channel: Instagram, TikTok or X. Submit projects via this form by Wednesday, July 16. Judges will assess plug-ins based on three main criteria: 1) innovation and creativity, 2) technical execution and integration, reviewing technical depth, G-Assist integration and scalability, and 3) usability and community impact, aka how easy it is to use the plug-in. Winners will be selected on Wednesday, Aug. 20. First place will receive a GeForce RTX 5090 laptop, second place a GeForce RTX 5080 GPU and third a GeForce RTX 5070 GPU. These top three will also be featured on NVIDIA’s social media channels, get the opportunity to meet the NVIDIA G-Assist team and earn an NVIDIA Deep Learning Institute self-paced course credit. Project G-Assist requires a GeForce RTX 50, 40 or 30 Series Desktop GPU with at least 12GB of VRAM, Windows 11 or 10 operating system, a compatible CPU, specific disk space requirements and a recent GeForce Game Ready Driver or NVIDIA Studio Driver. Plug-InExplore open-source plug-in samples available on GitHub, which showcase the diverse ways on-device AI can enhance PC and gaming workflows. Popular plug-ins include: Google Gemini: Enables search-based queries using Google Search integration and large language model-based queries using Gemini capabilities in real time without needing to switch programs from the convenience of the NVIDIA App Overlay. Discord: Enables users to easily share game highlights or messages directly to Discord servers without disrupting gameplay. IFTTT: Lets users create automations across hundreds of compatible endpoints to trigger IoT routines — such as adjusting room lights and smart shades, or pushing the latest gaming news to a mobile device. Spotify: Lets users control Spotify using simple voice commands or the G-Assist interface to play favorite tracks and manage playlists. Twitch: Checks if any Twitch streamer is currently live and can access detailed stream information such as titles, games, view counts and more. Get G-Assist  Join the NVIDIA Developer Discord channel to collaborate, share creations and gain support from fellow AI enthusiasts and NVIDIA staff. the date for NVIDIA’s How to Build a G-Assist Plug-In webinar on Wednesday, July 9, from 10-11 a.m. PT, to learn more about Project G-Assist capabilities, discover the fundamentals of building, testing and deploying Project G-Assist plug-ins, and participate in a live Q&A session. Explore NVIDIA’s GitHub repository, which provides everything needed to get started developing with G-Assist, including sample plug-ins, step-by-step instructions and documentation for building custom functionalities. Learn more about the ChatGPT Plug-In Builder to transform ideas into functional G-Assist plug-ins with minimal coding. The tool uses OpenAI’s custom GPT builder to generate plug-in code and streamline the development process. NVIDIA’s technical blog walks through the architecture of a G-Assist plug-in, using a Twitch integration as an example. Discover how plug-ins work, how they communicate with G-Assist and how to build them from scratch. Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations.  Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter. Follow NVIDIA Workstation on LinkedIn and X.  See notice regarding software product information. #plug #play #build #gassist #plugin
    BLOGS.NVIDIA.COM
    Plug and Play: Build a G-Assist Plug-In Today
    Project G-Assist — available through the NVIDIA App — is an experimental AI assistant that helps tune, control and optimize NVIDIA GeForce RTX systems. NVIDIA’s Plug and Play: Project G-Assist Plug-In Hackathon — running virtually through Wednesday, July 16 — invites the community to explore AI and build custom G-Assist plug-ins for a chance to win prizes and be featured on NVIDIA social media channels. G-Assist allows users to control their RTX GPU and other system settings using natural language, thanks to a small language model that runs on device. It can be used from the NVIDIA Overlay in the NVIDIA App without needing to tab out or switch programs. Users can expand its capabilities via plug-ins and even connect it to agentic frameworks such as Langflow. Below, find popular G-Assist plug-ins, hackathon details and tips to get started. Plug-In and Win Join the hackathon by registering and checking out the curated technical resources. G-Assist plug-ins can be built in several ways, including with Python for rapid development, with C++ for performance-critical apps and with custom system interactions for hardware and operating system automation. For those that prefer vibe coding, the G-Assist Plug-In Builder — a ChatGPT-based app that allows no-code or low-code development with natural language commands — makes it easy for enthusiasts to start creating plug-ins. To submit an entry, participants must provide a GitHub repository, including source code file (plugin.py), requirements.txt, manifest.json, config.json (if applicable), a plug-in executable file and READme code. Then, submit a video — between 30 seconds and two minutes — showcasing the plug-in in action. Finally, hackathoners must promote their plug-in using #AIonRTXHackathon on a social media channel: Instagram, TikTok or X. Submit projects via this form by Wednesday, July 16. Judges will assess plug-ins based on three main criteria: 1) innovation and creativity, 2) technical execution and integration, reviewing technical depth, G-Assist integration and scalability, and 3) usability and community impact, aka how easy it is to use the plug-in. Winners will be selected on Wednesday, Aug. 20. First place will receive a GeForce RTX 5090 laptop, second place a GeForce RTX 5080 GPU and third a GeForce RTX 5070 GPU. These top three will also be featured on NVIDIA’s social media channels, get the opportunity to meet the NVIDIA G-Assist team and earn an NVIDIA Deep Learning Institute self-paced course credit. Project G-Assist requires a GeForce RTX 50, 40 or 30 Series Desktop GPU with at least 12GB of VRAM, Windows 11 or 10 operating system, a compatible CPU (Intel Pentium G Series, Core i3, i5, i7 or higher; AMD FX, Ryzen 3, 5, 7, 9, Threadripper or higher), specific disk space requirements and a recent GeForce Game Ready Driver or NVIDIA Studio Driver. Plug-In(spiration) Explore open-source plug-in samples available on GitHub, which showcase the diverse ways on-device AI can enhance PC and gaming workflows. Popular plug-ins include: Google Gemini: Enables search-based queries using Google Search integration and large language model-based queries using Gemini capabilities in real time without needing to switch programs from the convenience of the NVIDIA App Overlay. Discord: Enables users to easily share game highlights or messages directly to Discord servers without disrupting gameplay. IFTTT: Lets users create automations across hundreds of compatible endpoints to trigger IoT routines — such as adjusting room lights and smart shades, or pushing the latest gaming news to a mobile device. Spotify: Lets users control Spotify using simple voice commands or the G-Assist interface to play favorite tracks and manage playlists. Twitch: Checks if any Twitch streamer is currently live and can access detailed stream information such as titles, games, view counts and more. Get G-Assist(ance)  Join the NVIDIA Developer Discord channel to collaborate, share creations and gain support from fellow AI enthusiasts and NVIDIA staff. Save the date for NVIDIA’s How to Build a G-Assist Plug-In webinar on Wednesday, July 9, from 10-11 a.m. PT, to learn more about Project G-Assist capabilities, discover the fundamentals of building, testing and deploying Project G-Assist plug-ins, and participate in a live Q&A session. Explore NVIDIA’s GitHub repository, which provides everything needed to get started developing with G-Assist, including sample plug-ins, step-by-step instructions and documentation for building custom functionalities. Learn more about the ChatGPT Plug-In Builder to transform ideas into functional G-Assist plug-ins with minimal coding. The tool uses OpenAI’s custom GPT builder to generate plug-in code and streamline the development process. NVIDIA’s technical blog walks through the architecture of a G-Assist plug-in, using a Twitch integration as an example. Discover how plug-ins work, how they communicate with G-Assist and how to build them from scratch. Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations.  Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter. Follow NVIDIA Workstation on LinkedIn and X.  See notice regarding software product information.
    Like
    Wow
    Love
    Sad
    25
    0 Comments 0 Shares
  • Resident Evil Requiem Continues "Overarching Narrative" That Began In Raccoon City

    Raccoon City was the American setting where Resident Evil began. After multiple instalments set around the world, Resident Evil Requiem producer Masato Kumazawa shared why the series is returning to the ruins of this iconic city.In a PlayStation Blog interview, Kumazawa explained that after more recent titles that had explored "the broader universe" of Resident Evil, Capcom wanted a story that "continues the overarching narrative rooted in Raccoon City and the secret machinations of the Umbrella Corporation." Following the T-virus outbreak, the US government ordered a missile strike to destroy the city in order to eradicate the virus. In having players return to its ruins about 30 years later, Kumazawa said that the team also wanted a character "with a personal connection to the city itself," introducing Grace Ashcroft, the presumed daughter of Resident Evil: Outbreak's Alyssa Ashcroft.Continue Reading at GameSpot
    #resident #evil #requiem #continues #quotoverarching
    Resident Evil Requiem Continues "Overarching Narrative" That Began In Raccoon City
    Raccoon City was the American setting where Resident Evil began. After multiple instalments set around the world, Resident Evil Requiem producer Masato Kumazawa shared why the series is returning to the ruins of this iconic city.In a PlayStation Blog interview, Kumazawa explained that after more recent titles that had explored "the broader universe" of Resident Evil, Capcom wanted a story that "continues the overarching narrative rooted in Raccoon City and the secret machinations of the Umbrella Corporation." Following the T-virus outbreak, the US government ordered a missile strike to destroy the city in order to eradicate the virus. In having players return to its ruins about 30 years later, Kumazawa said that the team also wanted a character "with a personal connection to the city itself," introducing Grace Ashcroft, the presumed daughter of Resident Evil: Outbreak's Alyssa Ashcroft.Continue Reading at GameSpot #resident #evil #requiem #continues #quotoverarching
    WWW.GAMESPOT.COM
    Resident Evil Requiem Continues "Overarching Narrative" That Began In Raccoon City
    Raccoon City was the American setting where Resident Evil began. After multiple instalments set around the world, Resident Evil Requiem producer Masato Kumazawa shared why the series is returning to the ruins of this iconic city.In a PlayStation Blog interview, Kumazawa explained that after more recent titles that had explored "the broader universe" of Resident Evil, Capcom wanted a story that "continues the overarching narrative rooted in Raccoon City and the secret machinations of the Umbrella Corporation." Following the T-virus outbreak, the US government ordered a missile strike to destroy the city in order to eradicate the virus. In having players return to its ruins about 30 years later, Kumazawa said that the team also wanted a character "with a personal connection to the city itself," introducing Grace Ashcroft, the presumed daughter of Resident Evil: Outbreak's Alyssa Ashcroft.Continue Reading at GameSpot
    0 Comments 0 Shares
  • Into the Omniverse: World Foundation Models Advance Autonomous Vehicle Simulation and Safety

    Editor’s note: This blog is a part of Into the Omniverse, a series focused on how developers, 3D practitioners and enterprises can transform their workflows using the latest advances in OpenUSD and NVIDIA Omniverse.
    Simulated driving environments enable engineers to safely and efficiently train, test and validate autonomous vehiclesacross countless real-world and edge-case scenarios without the risks and costs of physical testing.
    These simulated environments can be created through neural reconstruction of real-world data from AV fleets or generated with world foundation models— neural networks that understand physics and real-world properties. WFMs can be used to generate synthetic datasets for enhanced AV simulation.
    To help physical AI developers build such simulated environments, NVIDIA unveiled major advances in WFMs at the GTC Paris and CVPR conferences earlier this month. These new capabilities enhance NVIDIA Cosmos — a platform of generative WFMs, advanced tokenizers, guardrails and accelerated data processing tools.
    Key innovations like Cosmos Predict-2, the Cosmos Transfer-1 NVIDIA preview NIM microservice and Cosmos Reason are improving how AV developers generate synthetic data, build realistic simulated environments and validate safety systems at unprecedented scale.
    Universal Scene Description, a unified data framework and standard for physical AI applications, enables seamless integration and interoperability of simulation assets across the development pipeline. OpenUSD standardization plays a critical role in ensuring 3D pipelines are built to scale.
    NVIDIA Omniverse, a platform of application programming interfaces, software development kits and services for building OpenUSD-based physical AI applications, enables simulations from WFMs and neural reconstruction at world scale.
    Leading AV organizations — including Foretellix, Mcity, Oxa, Parallel Domain, Plus AI and Uber — are among the first to adopt Cosmos models.

    Foundations for Scalable, Realistic Simulation
    Cosmos Predict-2, NVIDIA’s latest WFM, generates high-quality synthetic data by predicting future world states from multimodal inputs like text, images and video. This capability is critical for creating temporally consistent, realistic scenarios that accelerate training and validation of AVs and robots.

    In addition, Cosmos Transfer, a control model that adds variations in weather, lighting and terrain to existing scenarios, will soon be available to 150,000 developers on CARLA, a leading open-source AV simulator. This greatly expands the broad AV developer community’s access to advanced AI-powered simulation tools.
    Developers can start integrating synthetic data into their own pipelines using the NVIDIA Physical AI Dataset. The latest release includes 40,000 clips generated using Cosmos.
    Building on these foundations, the Omniverse Blueprint for AV simulation provides a standardized, API-driven workflow for constructing rich digital twins, replaying real-world sensor data and generating new ground-truth data for closed-loop testing.
    The blueprint taps into OpenUSD’s layer-stacking and composition arcs, which enable developers to collaborate asynchronously and modify scenes nondestructively. This helps create modular, reusable scenario variants to efficiently generate different weather conditions, traffic patterns and edge cases.
    Driving the Future of AV Safety
    To bolster the operational safety of AV systems, NVIDIA earlier this year introduced NVIDIA Halos — a comprehensive safety platform that integrates the company’s full automotive hardware and software stack with AI research focused on AV safety.
    The new Cosmos models — Cosmos Predict- 2, Cosmos Transfer- 1 NIM and Cosmos Reason — deliver further safety enhancements to the Halos platform, enabling developers to create diverse, controllable and realistic scenarios for training and validating AV systems.
    These models, trained on massive multimodal datasets including driving data, amplify the breadth and depth of simulation, allowing for robust scenario coverage — including rare and safety-critical events — while supporting post-training customization for specialized AV tasks.

    At CVPR, NVIDIA was recognized as an Autonomous Grand Challenge winner, highlighting its leadership in advancing end-to-end AV workflows. The challenge used OpenUSD’s robust metadata and interoperability to simulate sensor inputs and vehicle trajectories in semi-reactive environments, achieving state-of-the-art results in safety and compliance.
    Learn more about how developers are leveraging tools like CARLA, Cosmos, and Omniverse to advance AV simulation in this livestream replay:

    Hear NVIDIA Director of Autonomous Vehicle Research Marco Pavone on the NVIDIA AI Podcast share how digital twins and high-fidelity simulation are improving vehicle testing, accelerating development and reducing real-world risks.
    Get Plugged Into the World of OpenUSD
    Learn more about what’s next for AV simulation with OpenUSD by watching the replay of NVIDIA founder and CEO Jensen Huang’s GTC Paris keynote.
    Looking for more live opportunities to learn more about OpenUSD? Don’t miss sessions and labs happening at SIGGRAPH 2025, August 10–14.
    Discover why developers and 3D practitioners are using OpenUSD and learn how to optimize 3D workflows with the self-paced “Learn OpenUSD” curriculum for 3D developers and practitioners, available for free through the NVIDIA Deep Learning Institute.
    Explore the Alliance for OpenUSD forum and the AOUSD website.
    Stay up to date by subscribing to NVIDIA Omniverse news, joining the community and following NVIDIA Omniverse on Instagram, LinkedIn, Medium and X.
    #into #omniverse #world #foundation #models
    Into the Omniverse: World Foundation Models Advance Autonomous Vehicle Simulation and Safety
    Editor’s note: This blog is a part of Into the Omniverse, a series focused on how developers, 3D practitioners and enterprises can transform their workflows using the latest advances in OpenUSD and NVIDIA Omniverse. Simulated driving environments enable engineers to safely and efficiently train, test and validate autonomous vehiclesacross countless real-world and edge-case scenarios without the risks and costs of physical testing. These simulated environments can be created through neural reconstruction of real-world data from AV fleets or generated with world foundation models— neural networks that understand physics and real-world properties. WFMs can be used to generate synthetic datasets for enhanced AV simulation. To help physical AI developers build such simulated environments, NVIDIA unveiled major advances in WFMs at the GTC Paris and CVPR conferences earlier this month. These new capabilities enhance NVIDIA Cosmos — a platform of generative WFMs, advanced tokenizers, guardrails and accelerated data processing tools. Key innovations like Cosmos Predict-2, the Cosmos Transfer-1 NVIDIA preview NIM microservice and Cosmos Reason are improving how AV developers generate synthetic data, build realistic simulated environments and validate safety systems at unprecedented scale. Universal Scene Description, a unified data framework and standard for physical AI applications, enables seamless integration and interoperability of simulation assets across the development pipeline. OpenUSD standardization plays a critical role in ensuring 3D pipelines are built to scale. NVIDIA Omniverse, a platform of application programming interfaces, software development kits and services for building OpenUSD-based physical AI applications, enables simulations from WFMs and neural reconstruction at world scale. Leading AV organizations — including Foretellix, Mcity, Oxa, Parallel Domain, Plus AI and Uber — are among the first to adopt Cosmos models. Foundations for Scalable, Realistic Simulation Cosmos Predict-2, NVIDIA’s latest WFM, generates high-quality synthetic data by predicting future world states from multimodal inputs like text, images and video. This capability is critical for creating temporally consistent, realistic scenarios that accelerate training and validation of AVs and robots. In addition, Cosmos Transfer, a control model that adds variations in weather, lighting and terrain to existing scenarios, will soon be available to 150,000 developers on CARLA, a leading open-source AV simulator. This greatly expands the broad AV developer community’s access to advanced AI-powered simulation tools. Developers can start integrating synthetic data into their own pipelines using the NVIDIA Physical AI Dataset. The latest release includes 40,000 clips generated using Cosmos. Building on these foundations, the Omniverse Blueprint for AV simulation provides a standardized, API-driven workflow for constructing rich digital twins, replaying real-world sensor data and generating new ground-truth data for closed-loop testing. The blueprint taps into OpenUSD’s layer-stacking and composition arcs, which enable developers to collaborate asynchronously and modify scenes nondestructively. This helps create modular, reusable scenario variants to efficiently generate different weather conditions, traffic patterns and edge cases. Driving the Future of AV Safety To bolster the operational safety of AV systems, NVIDIA earlier this year introduced NVIDIA Halos — a comprehensive safety platform that integrates the company’s full automotive hardware and software stack with AI research focused on AV safety. The new Cosmos models — Cosmos Predict- 2, Cosmos Transfer- 1 NIM and Cosmos Reason — deliver further safety enhancements to the Halos platform, enabling developers to create diverse, controllable and realistic scenarios for training and validating AV systems. These models, trained on massive multimodal datasets including driving data, amplify the breadth and depth of simulation, allowing for robust scenario coverage — including rare and safety-critical events — while supporting post-training customization for specialized AV tasks. At CVPR, NVIDIA was recognized as an Autonomous Grand Challenge winner, highlighting its leadership in advancing end-to-end AV workflows. The challenge used OpenUSD’s robust metadata and interoperability to simulate sensor inputs and vehicle trajectories in semi-reactive environments, achieving state-of-the-art results in safety and compliance. Learn more about how developers are leveraging tools like CARLA, Cosmos, and Omniverse to advance AV simulation in this livestream replay: Hear NVIDIA Director of Autonomous Vehicle Research Marco Pavone on the NVIDIA AI Podcast share how digital twins and high-fidelity simulation are improving vehicle testing, accelerating development and reducing real-world risks. Get Plugged Into the World of OpenUSD Learn more about what’s next for AV simulation with OpenUSD by watching the replay of NVIDIA founder and CEO Jensen Huang’s GTC Paris keynote. Looking for more live opportunities to learn more about OpenUSD? Don’t miss sessions and labs happening at SIGGRAPH 2025, August 10–14. Discover why developers and 3D practitioners are using OpenUSD and learn how to optimize 3D workflows with the self-paced “Learn OpenUSD” curriculum for 3D developers and practitioners, available for free through the NVIDIA Deep Learning Institute. Explore the Alliance for OpenUSD forum and the AOUSD website. Stay up to date by subscribing to NVIDIA Omniverse news, joining the community and following NVIDIA Omniverse on Instagram, LinkedIn, Medium and X. #into #omniverse #world #foundation #models
    BLOGS.NVIDIA.COM
    Into the Omniverse: World Foundation Models Advance Autonomous Vehicle Simulation and Safety
    Editor’s note: This blog is a part of Into the Omniverse, a series focused on how developers, 3D practitioners and enterprises can transform their workflows using the latest advances in OpenUSD and NVIDIA Omniverse. Simulated driving environments enable engineers to safely and efficiently train, test and validate autonomous vehicles (AVs) across countless real-world and edge-case scenarios without the risks and costs of physical testing. These simulated environments can be created through neural reconstruction of real-world data from AV fleets or generated with world foundation models (WFMs) — neural networks that understand physics and real-world properties. WFMs can be used to generate synthetic datasets for enhanced AV simulation. To help physical AI developers build such simulated environments, NVIDIA unveiled major advances in WFMs at the GTC Paris and CVPR conferences earlier this month. These new capabilities enhance NVIDIA Cosmos — a platform of generative WFMs, advanced tokenizers, guardrails and accelerated data processing tools. Key innovations like Cosmos Predict-2, the Cosmos Transfer-1 NVIDIA preview NIM microservice and Cosmos Reason are improving how AV developers generate synthetic data, build realistic simulated environments and validate safety systems at unprecedented scale. Universal Scene Description (OpenUSD), a unified data framework and standard for physical AI applications, enables seamless integration and interoperability of simulation assets across the development pipeline. OpenUSD standardization plays a critical role in ensuring 3D pipelines are built to scale. NVIDIA Omniverse, a platform of application programming interfaces, software development kits and services for building OpenUSD-based physical AI applications, enables simulations from WFMs and neural reconstruction at world scale. Leading AV organizations — including Foretellix, Mcity, Oxa, Parallel Domain, Plus AI and Uber — are among the first to adopt Cosmos models. Foundations for Scalable, Realistic Simulation Cosmos Predict-2, NVIDIA’s latest WFM, generates high-quality synthetic data by predicting future world states from multimodal inputs like text, images and video. This capability is critical for creating temporally consistent, realistic scenarios that accelerate training and validation of AVs and robots. In addition, Cosmos Transfer, a control model that adds variations in weather, lighting and terrain to existing scenarios, will soon be available to 150,000 developers on CARLA, a leading open-source AV simulator. This greatly expands the broad AV developer community’s access to advanced AI-powered simulation tools. Developers can start integrating synthetic data into their own pipelines using the NVIDIA Physical AI Dataset. The latest release includes 40,000 clips generated using Cosmos. Building on these foundations, the Omniverse Blueprint for AV simulation provides a standardized, API-driven workflow for constructing rich digital twins, replaying real-world sensor data and generating new ground-truth data for closed-loop testing. The blueprint taps into OpenUSD’s layer-stacking and composition arcs, which enable developers to collaborate asynchronously and modify scenes nondestructively. This helps create modular, reusable scenario variants to efficiently generate different weather conditions, traffic patterns and edge cases. Driving the Future of AV Safety To bolster the operational safety of AV systems, NVIDIA earlier this year introduced NVIDIA Halos — a comprehensive safety platform that integrates the company’s full automotive hardware and software stack with AI research focused on AV safety. The new Cosmos models — Cosmos Predict- 2, Cosmos Transfer- 1 NIM and Cosmos Reason — deliver further safety enhancements to the Halos platform, enabling developers to create diverse, controllable and realistic scenarios for training and validating AV systems. These models, trained on massive multimodal datasets including driving data, amplify the breadth and depth of simulation, allowing for robust scenario coverage — including rare and safety-critical events — while supporting post-training customization for specialized AV tasks. At CVPR, NVIDIA was recognized as an Autonomous Grand Challenge winner, highlighting its leadership in advancing end-to-end AV workflows. The challenge used OpenUSD’s robust metadata and interoperability to simulate sensor inputs and vehicle trajectories in semi-reactive environments, achieving state-of-the-art results in safety and compliance. Learn more about how developers are leveraging tools like CARLA, Cosmos, and Omniverse to advance AV simulation in this livestream replay: Hear NVIDIA Director of Autonomous Vehicle Research Marco Pavone on the NVIDIA AI Podcast share how digital twins and high-fidelity simulation are improving vehicle testing, accelerating development and reducing real-world risks. Get Plugged Into the World of OpenUSD Learn more about what’s next for AV simulation with OpenUSD by watching the replay of NVIDIA founder and CEO Jensen Huang’s GTC Paris keynote. Looking for more live opportunities to learn more about OpenUSD? Don’t miss sessions and labs happening at SIGGRAPH 2025, August 10–14. Discover why developers and 3D practitioners are using OpenUSD and learn how to optimize 3D workflows with the self-paced “Learn OpenUSD” curriculum for 3D developers and practitioners, available for free through the NVIDIA Deep Learning Institute. Explore the Alliance for OpenUSD forum and the AOUSD website. Stay up to date by subscribing to NVIDIA Omniverse news, joining the community and following NVIDIA Omniverse on Instagram, LinkedIn, Medium and X.
    0 Comments 0 Shares
  • Startup Uses NVIDIA RTX-Powered Generative AI to Make Coolers, Cooler

    Mark Theriault founded the startup FITY envisioning a line of clever cooling products: cold drink holders that come with freezable pucks to keep beverages cold for longer without the mess of ice. The entrepreneur started with 3D prints of products in his basement, building one unit at a time, before eventually scaling to mass production.
    Founding a consumer product company from scratch was a tall order for a single person. Going from preliminary sketches to production-ready designs was a major challenge. To bring his creative vision to life, Theriault relied on AI and his NVIDIA GeForce RTX-equipped system. For him, AI isn’t just a tool — it’s an entire pipeline to help him accomplish his goals. about his workflow below.
    Plus, GeForce RTX 5050 laptops start arriving today at retailers worldwide, from GeForce RTX 5050 Laptop GPUs feature 2,560 NVIDIA Blackwell CUDA cores, fifth-generation AI Tensor Cores, fourth-generation RT Cores, a ninth-generation NVENC encoder and a sixth-generation NVDEC decoder.
    In addition, NVIDIA’s Plug and Play: Project G-Assist Plug-In Hackathon — running virtually through Wednesday, July 16 — invites developers to explore AI and build custom G-Assist plug-ins for a chance to win prizes. the date for the G-Assist Plug-In webinar on Wednesday, July 9, from 10-11 a.m. PT, to learn more about Project G-Assist capabilities and fundamentals, and to participate in a live Q&A session.
    From Concept to Completion
    To create his standout products, Theriault tinkers with potential FITY Flex cooler designs with traditional methods, from sketch to computer-aided design to rapid prototyping, until he finds the right vision. A unique aspect of the FITY Flex design is that it can be customized with fun, popular shoe charms.
    For packaging design inspiration, Theriault uses his preferred text-to-image generative AI model for prototyping, Stable Diffusion XL — which runs 60% faster with the NVIDIA TensorRT software development kit — using the modular, node-based interface ComfyUI.
    ComfyUI gives users granular control over every step of the generation process — prompting, sampling, model loading, image conditioning and post-processing. It’s ideal for advanced users like Theriault who want to customize how images are generated.
    Theriault’s uses of AI result in a complete computer graphics-based ad campaign. Image courtesy of FITY.
    NVIDIA and GeForce RTX GPUs based on the NVIDIA Blackwell architecture include fifth-generation Tensor Cores designed to accelerate AI and deep learning workloads. These GPUs work with CUDA optimizations in PyTorch to seamlessly accelerate ComfyUI, reducing generation time on FLUX.1-dev, an image generation model from Black Forest Labs, from two minutes per image on the Mac M3 Ultra to about four seconds on the GeForce RTX 5090 desktop GPU.
    ComfyUI can also add ControlNets — AI models that help control image generation — that Theriault uses for tasks like guiding human poses, setting compositions via depth mapping and converting scribbles to images.
    Theriault even creates his own fine-tuned models to keep his style consistent. He used low-rank adaptationmodels — small, efficient adapters into specific layers of the network — enabling hyper-customized generation with minimal compute cost.
    LoRA models allow Theriault to ideate on visuals quickly. Image courtesy of FITY.
    “Over the last few months, I’ve been shifting from AI-assisted computer graphics renders to fully AI-generated product imagery using a custom Flux LoRA I trained in house. My RTX 4080 SUPER GPU has been essential for getting the performance I need to train and iterate quickly.” – Mark Theriault, founder of FITY 

    Theriault also taps into generative AI to create marketing assets like FITY Flex product packaging. He uses FLUX.1, which excels at generating legible text within images, addressing a common challenge in text-to-image models.
    Though FLUX.1 models can typically consume over 23GB of VRAM, NVIDIA has collaborated with Black Forest Labs to help reduce the size of these models using quantization — a technique that reduces model size while maintaining quality. The models were then accelerated with TensorRT, which provides an up to 2x speedup over PyTorch.
    To simplify using these models in ComfyUI, NVIDIA created the FLUX.1 NIM microservice, a containerized version of FLUX.1 that can be loaded in ComfyUI and enables FP4 quantization and TensorRT support. Combined, the models come down to just over 11GB of VRAM, and performance improves by 2.5x.
    Theriault uses the Blender Cycles app to render out final files. For 3D workflows, NVIDIA offers the AI Blueprint for 3D-guided generative AI to ease the positioning and composition of 3D images, so anyone interested in this method can quickly get started.
    Photorealistic renders. Image courtesy of FITY.
    Finally, Theriault uses large language models to generate marketing copy — tailored for search engine optimization, tone and storytelling — as well as to complete his patent and provisional applications, work that usually costs thousands of dollars in legal fees and considerable time.
    Generative AI helps Theriault create promotional materials like the above. Image courtesy of FITY.
    “As a one-man band with a ton of content to generate, having on-the-fly generation capabilities for my product designs really helps speed things up.” – Mark Theriault, founder of FITY

    Every texture, every word, every photo, every accessory was a micro-decision, Theriault said. AI helped him survive the “death by a thousand cuts” that can stall solo startup founders, he added.
    Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations. 
    Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter.
    Follow NVIDIA Workstation on LinkedIn and X. 
    See notice regarding software product information.
    #startup #uses #nvidia #rtxpowered #generative
    Startup Uses NVIDIA RTX-Powered Generative AI to Make Coolers, Cooler
    Mark Theriault founded the startup FITY envisioning a line of clever cooling products: cold drink holders that come with freezable pucks to keep beverages cold for longer without the mess of ice. The entrepreneur started with 3D prints of products in his basement, building one unit at a time, before eventually scaling to mass production. Founding a consumer product company from scratch was a tall order for a single person. Going from preliminary sketches to production-ready designs was a major challenge. To bring his creative vision to life, Theriault relied on AI and his NVIDIA GeForce RTX-equipped system. For him, AI isn’t just a tool — it’s an entire pipeline to help him accomplish his goals. about his workflow below. Plus, GeForce RTX 5050 laptops start arriving today at retailers worldwide, from GeForce RTX 5050 Laptop GPUs feature 2,560 NVIDIA Blackwell CUDA cores, fifth-generation AI Tensor Cores, fourth-generation RT Cores, a ninth-generation NVENC encoder and a sixth-generation NVDEC decoder. In addition, NVIDIA’s Plug and Play: Project G-Assist Plug-In Hackathon — running virtually through Wednesday, July 16 — invites developers to explore AI and build custom G-Assist plug-ins for a chance to win prizes. the date for the G-Assist Plug-In webinar on Wednesday, July 9, from 10-11 a.m. PT, to learn more about Project G-Assist capabilities and fundamentals, and to participate in a live Q&A session. From Concept to Completion To create his standout products, Theriault tinkers with potential FITY Flex cooler designs with traditional methods, from sketch to computer-aided design to rapid prototyping, until he finds the right vision. A unique aspect of the FITY Flex design is that it can be customized with fun, popular shoe charms. For packaging design inspiration, Theriault uses his preferred text-to-image generative AI model for prototyping, Stable Diffusion XL — which runs 60% faster with the NVIDIA TensorRT software development kit — using the modular, node-based interface ComfyUI. ComfyUI gives users granular control over every step of the generation process — prompting, sampling, model loading, image conditioning and post-processing. It’s ideal for advanced users like Theriault who want to customize how images are generated. Theriault’s uses of AI result in a complete computer graphics-based ad campaign. Image courtesy of FITY. NVIDIA and GeForce RTX GPUs based on the NVIDIA Blackwell architecture include fifth-generation Tensor Cores designed to accelerate AI and deep learning workloads. These GPUs work with CUDA optimizations in PyTorch to seamlessly accelerate ComfyUI, reducing generation time on FLUX.1-dev, an image generation model from Black Forest Labs, from two minutes per image on the Mac M3 Ultra to about four seconds on the GeForce RTX 5090 desktop GPU. ComfyUI can also add ControlNets — AI models that help control image generation — that Theriault uses for tasks like guiding human poses, setting compositions via depth mapping and converting scribbles to images. Theriault even creates his own fine-tuned models to keep his style consistent. He used low-rank adaptationmodels — small, efficient adapters into specific layers of the network — enabling hyper-customized generation with minimal compute cost. LoRA models allow Theriault to ideate on visuals quickly. Image courtesy of FITY. “Over the last few months, I’ve been shifting from AI-assisted computer graphics renders to fully AI-generated product imagery using a custom Flux LoRA I trained in house. My RTX 4080 SUPER GPU has been essential for getting the performance I need to train and iterate quickly.” – Mark Theriault, founder of FITY  Theriault also taps into generative AI to create marketing assets like FITY Flex product packaging. He uses FLUX.1, which excels at generating legible text within images, addressing a common challenge in text-to-image models. Though FLUX.1 models can typically consume over 23GB of VRAM, NVIDIA has collaborated with Black Forest Labs to help reduce the size of these models using quantization — a technique that reduces model size while maintaining quality. The models were then accelerated with TensorRT, which provides an up to 2x speedup over PyTorch. To simplify using these models in ComfyUI, NVIDIA created the FLUX.1 NIM microservice, a containerized version of FLUX.1 that can be loaded in ComfyUI and enables FP4 quantization and TensorRT support. Combined, the models come down to just over 11GB of VRAM, and performance improves by 2.5x. Theriault uses the Blender Cycles app to render out final files. For 3D workflows, NVIDIA offers the AI Blueprint for 3D-guided generative AI to ease the positioning and composition of 3D images, so anyone interested in this method can quickly get started. Photorealistic renders. Image courtesy of FITY. Finally, Theriault uses large language models to generate marketing copy — tailored for search engine optimization, tone and storytelling — as well as to complete his patent and provisional applications, work that usually costs thousands of dollars in legal fees and considerable time. Generative AI helps Theriault create promotional materials like the above. Image courtesy of FITY. “As a one-man band with a ton of content to generate, having on-the-fly generation capabilities for my product designs really helps speed things up.” – Mark Theriault, founder of FITY Every texture, every word, every photo, every accessory was a micro-decision, Theriault said. AI helped him survive the “death by a thousand cuts” that can stall solo startup founders, he added. Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations.  Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter. Follow NVIDIA Workstation on LinkedIn and X.  See notice regarding software product information. #startup #uses #nvidia #rtxpowered #generative
    BLOGS.NVIDIA.COM
    Startup Uses NVIDIA RTX-Powered Generative AI to Make Coolers, Cooler
    Mark Theriault founded the startup FITY envisioning a line of clever cooling products: cold drink holders that come with freezable pucks to keep beverages cold for longer without the mess of ice. The entrepreneur started with 3D prints of products in his basement, building one unit at a time, before eventually scaling to mass production. Founding a consumer product company from scratch was a tall order for a single person. Going from preliminary sketches to production-ready designs was a major challenge. To bring his creative vision to life, Theriault relied on AI and his NVIDIA GeForce RTX-equipped system. For him, AI isn’t just a tool — it’s an entire pipeline to help him accomplish his goals. Read more about his workflow below. Plus, GeForce RTX 5050 laptops start arriving today at retailers worldwide, from $999. GeForce RTX 5050 Laptop GPUs feature 2,560 NVIDIA Blackwell CUDA cores, fifth-generation AI Tensor Cores, fourth-generation RT Cores, a ninth-generation NVENC encoder and a sixth-generation NVDEC decoder. In addition, NVIDIA’s Plug and Play: Project G-Assist Plug-In Hackathon — running virtually through Wednesday, July 16 — invites developers to explore AI and build custom G-Assist plug-ins for a chance to win prizes. Save the date for the G-Assist Plug-In webinar on Wednesday, July 9, from 10-11 a.m. PT, to learn more about Project G-Assist capabilities and fundamentals, and to participate in a live Q&A session. From Concept to Completion To create his standout products, Theriault tinkers with potential FITY Flex cooler designs with traditional methods, from sketch to computer-aided design to rapid prototyping, until he finds the right vision. A unique aspect of the FITY Flex design is that it can be customized with fun, popular shoe charms. For packaging design inspiration, Theriault uses his preferred text-to-image generative AI model for prototyping, Stable Diffusion XL — which runs 60% faster with the NVIDIA TensorRT software development kit — using the modular, node-based interface ComfyUI. ComfyUI gives users granular control over every step of the generation process — prompting, sampling, model loading, image conditioning and post-processing. It’s ideal for advanced users like Theriault who want to customize how images are generated. Theriault’s uses of AI result in a complete computer graphics-based ad campaign. Image courtesy of FITY. NVIDIA and GeForce RTX GPUs based on the NVIDIA Blackwell architecture include fifth-generation Tensor Cores designed to accelerate AI and deep learning workloads. These GPUs work with CUDA optimizations in PyTorch to seamlessly accelerate ComfyUI, reducing generation time on FLUX.1-dev, an image generation model from Black Forest Labs, from two minutes per image on the Mac M3 Ultra to about four seconds on the GeForce RTX 5090 desktop GPU. ComfyUI can also add ControlNets — AI models that help control image generation — that Theriault uses for tasks like guiding human poses, setting compositions via depth mapping and converting scribbles to images. Theriault even creates his own fine-tuned models to keep his style consistent. He used low-rank adaptation (LoRA) models — small, efficient adapters into specific layers of the network — enabling hyper-customized generation with minimal compute cost. LoRA models allow Theriault to ideate on visuals quickly. Image courtesy of FITY. “Over the last few months, I’ve been shifting from AI-assisted computer graphics renders to fully AI-generated product imagery using a custom Flux LoRA I trained in house. My RTX 4080 SUPER GPU has been essential for getting the performance I need to train and iterate quickly.” – Mark Theriault, founder of FITY  Theriault also taps into generative AI to create marketing assets like FITY Flex product packaging. He uses FLUX.1, which excels at generating legible text within images, addressing a common challenge in text-to-image models. Though FLUX.1 models can typically consume over 23GB of VRAM, NVIDIA has collaborated with Black Forest Labs to help reduce the size of these models using quantization — a technique that reduces model size while maintaining quality. The models were then accelerated with TensorRT, which provides an up to 2x speedup over PyTorch. To simplify using these models in ComfyUI, NVIDIA created the FLUX.1 NIM microservice, a containerized version of FLUX.1 that can be loaded in ComfyUI and enables FP4 quantization and TensorRT support. Combined, the models come down to just over 11GB of VRAM, and performance improves by 2.5x. Theriault uses the Blender Cycles app to render out final files. For 3D workflows, NVIDIA offers the AI Blueprint for 3D-guided generative AI to ease the positioning and composition of 3D images, so anyone interested in this method can quickly get started. Photorealistic renders. Image courtesy of FITY. Finally, Theriault uses large language models to generate marketing copy — tailored for search engine optimization, tone and storytelling — as well as to complete his patent and provisional applications, work that usually costs thousands of dollars in legal fees and considerable time. Generative AI helps Theriault create promotional materials like the above. Image courtesy of FITY. “As a one-man band with a ton of content to generate, having on-the-fly generation capabilities for my product designs really helps speed things up.” – Mark Theriault, founder of FITY Every texture, every word, every photo, every accessory was a micro-decision, Theriault said. AI helped him survive the “death by a thousand cuts” that can stall solo startup founders, he added. Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations.  Plug in to NVIDIA AI PC on Facebook, Instagram, TikTok and X — and stay informed by subscribing to the RTX AI PC newsletter. Follow NVIDIA Workstation on LinkedIn and X.  See notice regarding software product information.
    0 Comments 0 Shares
  • Il est absolument scandaleux de voir comment la société actuelle traite des figures emblématiques comme Sœur Corita Kent, la religieuse Pop Art, sans même prendre le temps d’apprécier la profondeur de son engagement artistique et social. Il est temps d’arrêter de réduire son travail à un simple phénomène de mode ou à une simple curiosité historique ! Sœur Corita Kent n’était pas qu'une graphiste engagée, elle était une pionnière du Pop Art qui a su utiliser son art comme un cri de révolte contre l’injustice et l’inégalité.

    Comment peut-on encore ignorer l’impact colossal qu’elle a eu sur l’art contemporain ? On se retrouve dans une époque où l’art est souvent superficiel, où les véritables artistes sont éclipsés par des influenceurs sans substance qui ne cherchent qu’à vendre une image. C’est une honte ! Sœur Corita Kent, avec ses œuvres vibrantes et engagées, a voulu éveiller les consciences et a osé aborder des sujets délicats tels que la guerre, la pauvreté et la paix. Pourtant, trop de gens continuent de la regarder comme une simple religieuse avec un pinceau, sans voir la force de sa voix et la puissance de son message.

    Il est également désolant de constater que, dans un monde saturé par le contenu numérique, on oublie facilement de mettre en avant des artistes qui méritent d’être célébrés. Au lieu de cela, on glorifie des créations éphémères qui n’ont rien à dire. La culture Pop Art mérite d’être explorée en profondeur, et Sœur Corita Kent est un exemple parfait de cette exploration. Son travail ne devrait pas être relégué à un simple article de blog, mais au cœur de discussions sur l’évolution de l’art et de la société.

    Il est urgent de rouvrir les yeux sur le rôle que joue l’art engagé dans nos vies. Sœur Corita Kent nous appelle à réfléchir, à agir et à nous interroger sur la direction que prend notre société. Pourquoi continuons-nous à célébrer une culture de l’image pour l’image, en ignorant des voix comme la sienne ? C’est une véritable trahison envers l’art et l’humanité.

    Nous devons nous battre pour que son héritage ne soit pas oublié. Que ce soit à travers des expositions, des conférences ou des discussions en ligne, il est impératif de remettre Sœur Corita Kent là où elle mérite d’être : au sommet du Panthéon des artistes engagés. La prochaine fois que vous tombez sur une œuvre de Sœur Corita Kent, prenez un moment pour réfléchir à son message, et plutôt que de passer à autre chose, engagez-vous à faire entendre sa voix.

    Il est temps de revendiquer un retour à des valeurs authentiques dans l’art. Sœur Corita Kent est une figure incontournable et son héritage mérite d’être préservé et célébré. Arrêtons de nous contenter de peu et de vivre dans l’ignorance !

    #SœurCoritaKent #PopArt #ArtEngagé #JusticeSociale #Culture
    Il est absolument scandaleux de voir comment la société actuelle traite des figures emblématiques comme Sœur Corita Kent, la religieuse Pop Art, sans même prendre le temps d’apprécier la profondeur de son engagement artistique et social. Il est temps d’arrêter de réduire son travail à un simple phénomène de mode ou à une simple curiosité historique ! Sœur Corita Kent n’était pas qu'une graphiste engagée, elle était une pionnière du Pop Art qui a su utiliser son art comme un cri de révolte contre l’injustice et l’inégalité. Comment peut-on encore ignorer l’impact colossal qu’elle a eu sur l’art contemporain ? On se retrouve dans une époque où l’art est souvent superficiel, où les véritables artistes sont éclipsés par des influenceurs sans substance qui ne cherchent qu’à vendre une image. C’est une honte ! Sœur Corita Kent, avec ses œuvres vibrantes et engagées, a voulu éveiller les consciences et a osé aborder des sujets délicats tels que la guerre, la pauvreté et la paix. Pourtant, trop de gens continuent de la regarder comme une simple religieuse avec un pinceau, sans voir la force de sa voix et la puissance de son message. Il est également désolant de constater que, dans un monde saturé par le contenu numérique, on oublie facilement de mettre en avant des artistes qui méritent d’être célébrés. Au lieu de cela, on glorifie des créations éphémères qui n’ont rien à dire. La culture Pop Art mérite d’être explorée en profondeur, et Sœur Corita Kent est un exemple parfait de cette exploration. Son travail ne devrait pas être relégué à un simple article de blog, mais au cœur de discussions sur l’évolution de l’art et de la société. Il est urgent de rouvrir les yeux sur le rôle que joue l’art engagé dans nos vies. Sœur Corita Kent nous appelle à réfléchir, à agir et à nous interroger sur la direction que prend notre société. Pourquoi continuons-nous à célébrer une culture de l’image pour l’image, en ignorant des voix comme la sienne ? C’est une véritable trahison envers l’art et l’humanité. Nous devons nous battre pour que son héritage ne soit pas oublié. Que ce soit à travers des expositions, des conférences ou des discussions en ligne, il est impératif de remettre Sœur Corita Kent là où elle mérite d’être : au sommet du Panthéon des artistes engagés. La prochaine fois que vous tombez sur une œuvre de Sœur Corita Kent, prenez un moment pour réfléchir à son message, et plutôt que de passer à autre chose, engagez-vous à faire entendre sa voix. Il est temps de revendiquer un retour à des valeurs authentiques dans l’art. Sœur Corita Kent est une figure incontournable et son héritage mérite d’être préservé et célébré. Arrêtons de nous contenter de peu et de vivre dans l’ignorance ! #SœurCoritaKent #PopArt #ArtEngagé #JusticeSociale #Culture
    Sœur Corita Kent, la religieuse Pop Art
    Sœur Corita Kent (1918-1986). Graphiste engagée et pionnière du Pop Art. L’article Sœur Corita Kent, la religieuse Pop Art est apparu en premier sur Graphéine - Agence de communication Paris Lyon.
    Like
    Love
    Wow
    Sad
    Angry
    606
    1 Comments 0 Shares

  • ## Introduction

    Dans un monde où chaque voix mérite d'être entendue, créer un blog peut sembler être une lumière d'espoir au milieu des ténèbres de l'indifférence. Mais que faire lorsque l'on souhaite partager ses pensées, ses passions, et ses douleurs, tout en ne sachant pas par où commencer ? La création d'un blog gratuit sur WordPress.com pourrait être le premier pas vers une catharsis tant attendue. Dans ce guide, nous allons explorer ensemble comment donner vie à vos mots, comme un écho d...
    ## Introduction Dans un monde où chaque voix mérite d'être entendue, créer un blog peut sembler être une lumière d'espoir au milieu des ténèbres de l'indifférence. Mais que faire lorsque l'on souhaite partager ses pensées, ses passions, et ses douleurs, tout en ne sachant pas par où commencer ? La création d'un blog gratuit sur WordPress.com pourrait être le premier pas vers une catharsis tant attendue. Dans ce guide, nous allons explorer ensemble comment donner vie à vos mots, comme un écho d...
    Créer un blog gratuit sur WordPress.com : guide étape par étape (2025)
    ## Introduction Dans un monde où chaque voix mérite d'être entendue, créer un blog peut sembler être une lumière d'espoir au milieu des ténèbres de l'indifférence. Mais que faire lorsque l'on souhaite partager ses pensées, ses passions, et ses douleurs, tout en ne sachant pas par où commencer ? La création d'un blog gratuit sur WordPress.com pourrait être le premier pas vers une catharsis...
    Like
    Love
    Wow
    Sad
    Angry
    614
    1 Comments 0 Shares
  • Bonjour à tous, amis créateurs de contenu ! Aujourd'hui, je suis ravi de partager avec vous une aventure passionnante qui pourrait changer votre vie : **comment créer un blog pas à pas (2025)** !

    Vous avez une passion, une expertise ou simplement une idée que vous souhaitez partager avec le monde ? C'est le moment de briller ! N'oubliez pas, chaque grand voyage commence par un petit pas. Et créer un blog, c'est l'une des meilleures façons de vous exprimer et de vous connecter avec des gens partageant les mêmes idées.

    ### Étape 1 : Trouvez votre niche !
    La première étape de votre blog est de déterminer ce dont vous voulez parler. Que ce soit la cuisine, le voyage, la technologie ou même votre quotidien, choisissez un sujet qui vous fait vibrer ! Cela rendra votre écriture authentique et excitante pour vos lecteurs.

    ### Étape 2 : Choisissez une plateforme
    Il existe de nombreuses plateformes de blogging disponibles, comme WordPress, Blogger ou Wix. Prenez le temps de les explorer et choisissez celle qui correspond le mieux à vos besoins. N'oubliez pas, la facilité d'utilisation est la clé pour que vous puissiez vous concentrer sur ce que vous aimez : écrire !

    ### Étape 3 : Créez votre contenu
    Maintenant que vous avez votre plateforme, il est temps de commencer à écrire ! Ne vous inquiétez pas si votre première publication n'est pas parfaite. L'important est de commencer ! Partagez vos pensées, vos expériences et laissez parler votre cœur. Chaque article est une occasion d'apprendre et de grandir.

    ### Étape 4 : Promouvez votre blog
    Une fois que vous avez publié quelques articles, il est temps de les faire connaître ! Utilisez les réseaux sociaux, rejoignez des groupes de blogging et n'hésitez pas à interagir avec d'autres blogueurs. Ensemble, vous pouvez créer une communauté incroyable !

    ### Étape 5 : Restez constant et amusez-vous !
    Le blogging est un marathon, pas un sprint. Soyez régulier dans vos publications et surtout, amusez-vous ! Votre enthousiasme se ressentira dans votre écriture et attirera les lecteurs. N'oubliez pas de célébrer chaque petite victoire en cours de route !

    Alors, êtes-vous prêts à vous lancer dans cette aventure incroyable ? N'oubliez pas, le monde a besoin de votre voix unique ! Allez-y, créez votre blog et partagez votre lumière avec le monde. Vous ne le regretterez pas !

    #Blogging #CréerUnBlog #Inspiration #Passion #Écriture
    🌟 Bonjour à tous, amis créateurs de contenu ! Aujourd'hui, je suis ravi de partager avec vous une aventure passionnante qui pourrait changer votre vie : **comment créer un blog pas à pas (2025)** ! 🚀 Vous avez une passion, une expertise ou simplement une idée que vous souhaitez partager avec le monde ? C'est le moment de briller ! 🌈 N'oubliez pas, chaque grand voyage commence par un petit pas. Et créer un blog, c'est l'une des meilleures façons de vous exprimer et de vous connecter avec des gens partageant les mêmes idées. 💖 ### Étape 1 : Trouvez votre niche ! 🕵️‍♂️ La première étape de votre blog est de déterminer ce dont vous voulez parler. Que ce soit la cuisine, le voyage, la technologie ou même votre quotidien, choisissez un sujet qui vous fait vibrer ! Cela rendra votre écriture authentique et excitante pour vos lecteurs. 🌍 ### Étape 2 : Choisissez une plateforme 🌐 Il existe de nombreuses plateformes de blogging disponibles, comme WordPress, Blogger ou Wix. Prenez le temps de les explorer et choisissez celle qui correspond le mieux à vos besoins. N'oubliez pas, la facilité d'utilisation est la clé pour que vous puissiez vous concentrer sur ce que vous aimez : écrire ! ✍️ ### Étape 3 : Créez votre contenu 🎉 Maintenant que vous avez votre plateforme, il est temps de commencer à écrire ! Ne vous inquiétez pas si votre première publication n'est pas parfaite. L'important est de commencer ! Partagez vos pensées, vos expériences et laissez parler votre cœur. Chaque article est une occasion d'apprendre et de grandir. 📚 ### Étape 4 : Promouvez votre blog 📣 Une fois que vous avez publié quelques articles, il est temps de les faire connaître ! Utilisez les réseaux sociaux, rejoignez des groupes de blogging et n'hésitez pas à interagir avec d'autres blogueurs. Ensemble, vous pouvez créer une communauté incroyable ! 🤝 ### Étape 5 : Restez constant et amusez-vous ! 😄 Le blogging est un marathon, pas un sprint. Soyez régulier dans vos publications et surtout, amusez-vous ! Votre enthousiasme se ressentira dans votre écriture et attirera les lecteurs. N'oubliez pas de célébrer chaque petite victoire en cours de route ! 🥳 Alors, êtes-vous prêts à vous lancer dans cette aventure incroyable ? N'oubliez pas, le monde a besoin de votre voix unique ! Allez-y, créez votre blog et partagez votre lumière avec le monde. Vous ne le regretterez pas ! 🌟 #Blogging #CréerUnBlog #Inspiration #Passion #Écriture
    Cómo crear un blog paso a paso (2025): guía completa
    Cómo crear un blog paso a paso (2025): guía completa ¿Estás pensando en crear un blog pero no sabes por dónde empezar? Es normal, y es que nadie nace sabiendo cómo montar su primer blog. Hace falta un tutorial como este para llevarlo a cabo. En esta
    Like
    Love
    Wow
    Sad
    Angry
    586
    1 Comments 0 Shares
  • WordPress, PHP, desarrollo web, código abierto, WordCamp, programación, personalización, blog, CMS

    ## Introducción

    ¡Ah, WordPress! Esa herramienta mágica que convierte a cualquiera en un "desarrollador" con solo un par de clics. Pero, seamos sinceros: si te has aventurado a modificar un tema o plugin, probablemente te has encontrado con la pregunta del millón: "¿Dónde demonios está el código?". Si alguna vez te has sentido perdido en la jungla de archivos y carpetas de WordPress, no te preocup...
    WordPress, PHP, desarrollo web, código abierto, WordCamp, programación, personalización, blog, CMS ## Introducción ¡Ah, WordPress! Esa herramienta mágica que convierte a cualquiera en un "desarrollador" con solo un par de clics. Pero, seamos sinceros: si te has aventurado a modificar un tema o plugin, probablemente te has encontrado con la pregunta del millón: "¿Dónde demonios está el código?". Si alguna vez te has sentido perdido en la jungla de archivos y carpetas de WordPress, no te preocup...
    Cómo y por qué leer el código de WordPress
    WordPress, PHP, desarrollo web, código abierto, WordCamp, programación, personalización, blog, CMS ## Introducción ¡Ah, WordPress! Esa herramienta mágica que convierte a cualquiera en un "desarrollador" con solo un par de clics. Pero, seamos sinceros: si te has aventurado a modificar un tema o plugin, probablemente te has encontrado con la pregunta del millón: "¿Dónde demonios está el...
    Like
    Love
    Wow
    Sad
    Angry
    610
    1 Comments 0 Shares