
Nvidias Vera Rubin represents a big bet on real-time AI reasoning
www.fastcompany.com
Welcome toAI Decoded,Fast Companys weekly newsletter that breaks down the most important news in the world of AI. You can sign up to receive this newsletter every weekhere.Nvidias Jensen Huang: The world got DeepSeek totally wrongAt Nvidias GTC developer event in San Jose this week, CEO Jensen Huang called it AIs Super Bowl. Indeed, Nvidia is the chip supplier of the AI boom, and arguably the most influential (and certainly most profitable) company in the growing AI industry. As it does every year, the company used Huangs keynote to announce its next-generation chips, which will power the training and operation of the AI models of the near future.Huang introduced a new platform called Vera Rubin, which includes Nvidias first custom-designed CPU, Vera, and two new Rubin GPUs. (The platform is named after astronomer Vera Rubin, who discovered evidence of dark matter.) Nvidia says Vera is twice as fast as the Arm-based CPU in last years Blackwell platform. Together, Vera and Rubin deliver more than double the inference performance of Blackwell.Huang also revealed Blackwell Ultra, a more powerful version of the current flagship GPU, with increased compute and memory. Its expected to launch in the second half of 2025.These platformsespecially Vera Rubinare designed to boost performance during inference, the real-time reasoning AI models perform to generate answers. Huang stressed the growing need for chips that can rapidly process and store massive amounts of data during inference. In a pointed comment, he said almost the entire world got it wrong about the DeepSeek phenomenon, referencing the Chinese startup that claimed it could train top-tier models with fewer, weaker GPUs. The amount of computation we need as a result of agentic AI, as a result of reasoning, is easily 100 times more than we thought we needed this time last year, he said. In other words, inference demand is set to soarand with it, demand for Nvidia hardware.Huang also updated Project Digits, a desktop AI platform for developers, researchers, and students. It will ship in two versions: a compact GTX Spark, and a larger GTX Station capable of running more complex models. Nvidia says both will reach the market this summer.As Nvidias role in generative AI deepens, its pushing beyond infrastructure and into the application layer. To that end, the company announced a new family of open-source, midsize reasoning models called Nemotron. Built on Metas Llama, these models can run on the Spark or Station hardware or in the cloud. While not the most advanced, theyre optimized for lighter enterprise tasks and run efficiently on Nvidia chipsoffering yet another incentive for companies to stay within the Nvidia ecosystem.Big AI gives the Trump administration pointers on its AI Action PlanAs the U.S. government rethinks its approach to artificial intelligence, the emerging consensus around this second Trump administration is clear: minimal oversight, maximum freedom for industry. The working theory is that President Donald Trump and Vice President JD Vance would take a hands-off approach to AI regulation, favoring rapid innovation over safety mandates.Both Trump and Vance have emphasized the need for U.S. AI companies to operate without transparency mandates or safety guidelines, arguing that this freedom is essential for global leadership. The AI future is not going to be won by hand-wringing about safety, Vance said during a February speech at the Artificial Intelligence Action Summit in Paris. On day one, the administration rescinded the Biden-era AI safety and transparency guidelines, which were largely voluntary.The Trump team is reportedly developing its own AI policy. As part of that effort, the Office of Science and Technology Policy (OSTP) solicited public input on an AI Action Plan. The plan, according to the OSTP request, aims to define priority policy actions needed to sustain and enhance Americas AI dominance while avoiding unnecessarily burdensome requirements on private-sector innovation. Comments were accepted through March 15, and several major AI companiesOpenAI, Anthropic, Meta, and Googleposted their submissions publicly.Across the board, these companies warned that overly aggressive safety rules could cause the U.S. to fall behind China in the AI race. Most supported maintaining export bans on advanced AI chips to adversarial nations like China. OpenAI, Anthropic, and Google also called for protecting the Copyright Acts fair use provision, which allows AI labs to train models on publicly available data.Still, some comments signaled a more nuanced view of risk. Anthropic urged the government to create mechanisms for assessing whether private-sector models could pose national security threats, and suggested opening dedicated communication channels between intelligence agencies and major AI labs.Interestingly, OpenAI seems to favor a ban on models from Chinese startup DeepSeek, citing national security concerns similar to those that led to the U.S. ban on Huawei equipment. DeepSeek faces requirements under Chinese law to comply with demands for user data and uses it to train more capable systems for the CCPs use, the company said.While an AI Action Plan may be written, companies arent expecting strict safety regulations to followespecially under the current administration and Congress. Yet as AI systems grow more powerful, so do the risks of unregulated misuse, potentially on a large scale.And as Wendy Gonzalez, CEO of the AI training data company Sama, points out, a lack of guardrails may actually be bad for business in the long run. The comparison to vehicle safety standards is a good example, she tells Fast Company. Just as we dont view seat belts as overregulation that stifles automotive innovation, thoughtful AI guardrails protect stakeholders while enabling progress.Superhuman is a case study in the right way to integrate AI into appsMuch of the post-ChatGPT AI boom has focused on models themselvesbut increasingly, the spotlight is shifting to how AI is applied and experienced in real-world apps. Some apps are entirely new, made possible only by generative AI. Others, like Superhuman, stand out for how thoughtfully theyve integrated the technology into existing products.Superhuman has been around since 2014, and the addition of AI features hasnt changed Superhumans spare and sleek design very much. Its creators have integrated AI features in a purposeful and understated way. Superhuman founder and CEO Rahul Vohra gave me a run-through of some of the newer AI features when we met at the HumanX conference last week.The AI works like a behind-the-scenes assistant, aware of the context and content of your inbox. For long or complex threads, it precomputes summaries. It can draft replies in your writing style, prompt you to respond to urgent or high-priority messages, and learn contacts communication patterns to time your replies for maximum visibility. It also helps organize your inbox by automatically routing less important emailslike marketing, social updates, or cold pitchesinto separate folders.Vohra claims Superhuman can save users up to four hours per week. One internal study by a major consulting firm found it saved partners 3.3 hours weekly, sped up response times by 3.6 hours, and increased email throughput by 60%.At $30 to $40 per month, Superhuman isnt cheap. But for professionals overwhelmed by email, its AI-powered productivity boost may be worth the price.More AI coverage from Fast Company:The most innovative companies in artificial intelligence for 2025AI data centers run hot. These lightweight motors keep them cool with less powerHollywood warns about AI industrys push to change copyright lawThe rise of the AI managerWant exclusive reporting and trend analysis on technology, business innovation, future of work, and design? Sign up for Fast Company Premium.
0 Comments
·0 Shares
·62 Views