• The IAAC Global Summer School 2025 workshops and IAAC Nodes network are back

    html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN" ";
    The Institute for Advanced Architecture of Cataloniain Barcelona announces a new edition of the Global Summer School 2025, marking another milestone fostering global collaboration and innovation in the field of architecture and design.Over the years, IAAC has been at the forefront of promoting meaningful knowledge and expertise sharing among architects, engineers, designers and innovators worldwide. The GSS brings together students from diverse backgrounds to engage with IAAC's network of experts and renowned lecturers.This July, GSS proposes a unique blend of onsite and online workshops, offering to its participants hands-on experiences at IAAC in Barcelona while ensuring digital accessibility through an agile online format accessible to a global audience. GSS25 presents an immersive lineup of 4 workshops exploring advanced architecture & digital fabrication, advanced computation, urban data analytics and artificial intelligence. Led by IAAC Faculties and Alumni, these workshops promise innovative workflows and individual project exploration and development.GSS students can tailor their learning journey opting for a two-week onsite workshop in Barcelona or selectingone or more online modules offered throughout July 2025, becoming part of a dynamic and challenging educational experience.After 4 years, IAAC is glad to open its Global Nodes Network, strategic cities around the world boosting innovation by leading GSS onsite workshops led by IAAC alumni in diverse geographic locations. Applicants can therefore choose to join GSS in the preferred node - city - county!All GSS25 courses include exclusive access to Global Lectures, weekly online gatherings featuring experts from the Architecture, Engineering, and Constructionindustry also connecting participants of IAAC nodes located worldwide. These sessions are designed to complement the workshop themes and provide participants with valuable insights and perspectives.Applications with the early bird fees deadline is the 31st of May. Please read about prices, eligibility criteria and applications on the website.About the Institute for Advanced Architecture of CataloniaIAAC is a centre for research, education, production and outreach, with the mission of envisioning the future habitat of our society and building it in the present.IAAC follows the digital revolution at all scalesto expand the boundaries of architecture and design and meet the challenges faced by humanity. IAAC is an experimental and experiential centre where one learns by doing, through a test methodology that promotes real solutions.IAAC is an open, independent and radical non-profit foundation, with 20 years of activity; inspired by the values of Barcelona, the capital of architecture and design, where urbanism was invented and where local high quality and innovation-oriented research is connected to an international network of excellence in technology, architecture and society fields.ContactsGSS25 Agenda & AbstractsCircular Construction: Shifting ValueThe Circular Construction workshop at IAAC explores sustainable architecture through the upcycling of forest and marine debris from the Mediterranean region. Participants engage in hands-on material experimentation, 3D scanning, and computational design to transform waste into high-performance architectural systems. The course integrates digital fabrication techniques like CNC milling, 3D printing, and robotic assembly to prototype full-scale components. Emphasizing circularity and material intelligence, the workshop equips students with cutting-edge skills in design, sustainability, and fabrication.Dates: 7th - 18th July 2025Venue: IAAC Campus, BarcelonaGenerative AI for ArchitectureIn this workshop, participants will learn to apply AI and computational design tools to architecture and planning using low-code platforms like Grasshopper and n8n. They will  explore generative AI, simulations, and automation to create and prototype responsive, data-driven workflows. By the end, students will critically present their solutions, gaining hands-on experience with AI-driven design innovation.Dates: 7th - 11th July 2025Venue: OnlineUrban Data AnalyticsThe participants will learn to combine computer vision and GIS tools to analyze and visualize urban data at multiple scales. They’ll work with Google Street View and spatial datasets to uncover hidden urban conditions, assess city health, and explore concepts like “Sick City Syndrome.” By the end, students will be equipped to build multiscalar data workflows and create impactful visualizations that reveal and communicate complex urban dynamics.Dates: 21st - 24th July 2025Venue: OnlineAll images courtesy of IAAC.> via IAAC 
    architecture event
    #iaac #global #summer #school #workshops
    The IAAC Global Summer School 2025 workshops and IAAC Nodes network are back
    html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN" "; The Institute for Advanced Architecture of Cataloniain Barcelona announces a new edition of the Global Summer School 2025, marking another milestone fostering global collaboration and innovation in the field of architecture and design.Over the years, IAAC has been at the forefront of promoting meaningful knowledge and expertise sharing among architects, engineers, designers and innovators worldwide. The GSS brings together students from diverse backgrounds to engage with IAAC's network of experts and renowned lecturers.This July, GSS proposes a unique blend of onsite and online workshops, offering to its participants hands-on experiences at IAAC in Barcelona while ensuring digital accessibility through an agile online format accessible to a global audience. GSS25 presents an immersive lineup of 4 workshops exploring advanced architecture & digital fabrication, advanced computation, urban data analytics and artificial intelligence. Led by IAAC Faculties and Alumni, these workshops promise innovative workflows and individual project exploration and development.GSS students can tailor their learning journey opting for a two-week onsite workshop in Barcelona or selectingone or more online modules offered throughout July 2025, becoming part of a dynamic and challenging educational experience.After 4 years, IAAC is glad to open its Global Nodes Network, strategic cities around the world boosting innovation by leading GSS onsite workshops led by IAAC alumni in diverse geographic locations. Applicants can therefore choose to join GSS in the preferred node - city - county!All GSS25 courses include exclusive access to Global Lectures, weekly online gatherings featuring experts from the Architecture, Engineering, and Constructionindustry also connecting participants of IAAC nodes located worldwide. These sessions are designed to complement the workshop themes and provide participants with valuable insights and perspectives.Applications with the early bird fees deadline is the 31st of May. Please read about prices, eligibility criteria and applications on the website.About the Institute for Advanced Architecture of CataloniaIAAC is a centre for research, education, production and outreach, with the mission of envisioning the future habitat of our society and building it in the present.IAAC follows the digital revolution at all scalesto expand the boundaries of architecture and design and meet the challenges faced by humanity. IAAC is an experimental and experiential centre where one learns by doing, through a test methodology that promotes real solutions.IAAC is an open, independent and radical non-profit foundation, with 20 years of activity; inspired by the values of Barcelona, the capital of architecture and design, where urbanism was invented and where local high quality and innovation-oriented research is connected to an international network of excellence in technology, architecture and society fields.ContactsGSS25 Agenda & AbstractsCircular Construction: Shifting ValueThe Circular Construction workshop at IAAC explores sustainable architecture through the upcycling of forest and marine debris from the Mediterranean region. Participants engage in hands-on material experimentation, 3D scanning, and computational design to transform waste into high-performance architectural systems. The course integrates digital fabrication techniques like CNC milling, 3D printing, and robotic assembly to prototype full-scale components. Emphasizing circularity and material intelligence, the workshop equips students with cutting-edge skills in design, sustainability, and fabrication.Dates: 7th - 18th July 2025Venue: IAAC Campus, BarcelonaGenerative AI for ArchitectureIn this workshop, participants will learn to apply AI and computational design tools to architecture and planning using low-code platforms like Grasshopper and n8n. They will  explore generative AI, simulations, and automation to create and prototype responsive, data-driven workflows. By the end, students will critically present their solutions, gaining hands-on experience with AI-driven design innovation.Dates: 7th - 11th July 2025Venue: OnlineUrban Data AnalyticsThe participants will learn to combine computer vision and GIS tools to analyze and visualize urban data at multiple scales. They’ll work with Google Street View and spatial datasets to uncover hidden urban conditions, assess city health, and explore concepts like “Sick City Syndrome.” By the end, students will be equipped to build multiscalar data workflows and create impactful visualizations that reveal and communicate complex urban dynamics.Dates: 21st - 24th July 2025Venue: OnlineAll images courtesy of IAAC.> via IAAC  architecture event #iaac #global #summer #school #workshops
    WORLDARCHITECTURE.ORG
    The IAAC Global Summer School 2025 workshops and IAAC Nodes network are back
    html PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN" "http://www.w3.org/TR/REC-html40/loose.dtd" The Institute for Advanced Architecture of Catalonia (IAAC) in Barcelona announces a new edition of the Global Summer School 2025 (GSS25), marking another milestone fostering global collaboration and innovation in the field of architecture and design.Over the years, IAAC has been at the forefront of promoting meaningful knowledge and expertise sharing among architects, engineers, designers and innovators worldwide. The GSS brings together students from diverse backgrounds to engage with IAAC's network of experts and renowned lecturers.This July, GSS proposes a unique blend of onsite and online workshops, offering to its participants hands-on experiences at IAAC in Barcelona while ensuring digital accessibility through an agile online format accessible to a global audience. GSS25 presents an immersive lineup of 4 workshops exploring advanced architecture & digital fabrication, advanced computation, urban data analytics and artificial intelligence. Led by IAAC Faculties and Alumni, these workshops promise innovative workflows and individual project exploration and development.GSS students can tailor their learning journey opting for a two-week onsite workshop in Barcelona or selectingone or more online modules offered throughout July 2025, becoming part of a dynamic and challenging educational experience.After 4 years, IAAC is glad to open its Global Nodes Network, strategic cities around the world boosting innovation by leading GSS onsite workshops led by IAAC alumni in diverse geographic locations. Applicants can therefore choose to join GSS in the preferred node - city - county!All GSS25 courses include exclusive access to Global Lectures, weekly online gatherings featuring experts from the Architecture, Engineering, and Construction (AEC) industry also connecting participants of IAAC nodes located worldwide. These sessions are designed to complement the workshop themes and provide participants with valuable insights and perspectives.Applications with the early bird fees deadline is the 31st of May. Please read about prices, eligibility criteria and applications on the website.About the Institute for Advanced Architecture of Catalonia (IAAC)IAAC is a centre for research, education, production and outreach, with the mission of envisioning the future habitat of our society and building it in the present.IAAC follows the digital revolution at all scales (from bits to geography, from micro-controllers to cities, from materials to the territory) to expand the boundaries of architecture and design and meet the challenges faced by humanity. IAAC is an experimental and experiential centre where one learns by doing, through a test methodology that promotes real solutions.IAAC is an open, independent and radical non-profit foundation, with 20 years of activity; inspired by the values of Barcelona, the capital of architecture and design, where urbanism was invented and where local high quality and innovation-oriented research is connected to an international network of excellence in technology, architecture and society fields.ContactsGSS25 Agenda & AbstractsCircular Construction: Shifting ValueThe Circular Construction workshop at IAAC explores sustainable architecture through the upcycling of forest and marine debris from the Mediterranean region. Participants engage in hands-on material experimentation, 3D scanning, and computational design to transform waste into high-performance architectural systems. The course integrates digital fabrication techniques like CNC milling, 3D printing, and robotic assembly to prototype full-scale components. Emphasizing circularity and material intelligence, the workshop equips students with cutting-edge skills in design, sustainability, and fabrication.Dates: 7th - 18th July 2025Venue: IAAC Campus, Barcelona (Spain)Generative AI for ArchitectureIn this workshop, participants will learn to apply AI and computational design tools to architecture and planning using low-code platforms like Grasshopper and n8n. They will  explore generative AI, simulations, and automation to create and prototype responsive, data-driven workflows. By the end, students will critically present their solutions, gaining hands-on experience with AI-driven design innovation.Dates: 7th - 11th July 2025Venue: Online (Synchronous & Asynchronous formatAdvanced Computation for Design WorkshopDuring the workshop, participants will learn to apply Python and machine learning to optimize architectural design with a focus on sustainability, form finding, and facade engineering. They will gain hands-on experience in geometry rationalization, environmental analysis, and data-centric modeling for intelligent design workflows. By the end, attendees will be equipped to create efficient, fabrication-ready forms through advanced computational and ML-driven techniques.Dates: 14th - 18th July 2025Venue: Online (Synchronous & Asynchronous format)Urban Data AnalyticsThe participants will learn to combine computer vision and GIS tools to analyze and visualize urban data at multiple scales. They’ll work with Google Street View and spatial datasets to uncover hidden urban conditions, assess city health, and explore concepts like “Sick City Syndrome.” By the end, students will be equipped to build multiscalar data workflows and create impactful visualizations that reveal and communicate complex urban dynamics.Dates: 21st - 24th July 2025Venue: Online (Synchronous & Asynchronous format)All images courtesy of IAAC.> via IAAC  architecture event
    0 Yorumlar 0 hisse senetleri 0 önizleme
  • To grow, we must forget… but AI remembers everything

    To grow, we must forget… but now AI remembers everythingAI’s infinite memory could endanger how we think, grow, and imagine. And we can do something about it.Photo by Laura Fuhrman on UnsplashWhen Mary remembered too muchImagine your best friend — we’ll call her Mary — had perfect, infallible memory.At first, it feels wonderful. She remembers your favorite dishes, obscure movie quotes, even that exact shade of sweater you casually admired months ago. Dinner plans are effortless: “Booked us Giorgio’s again, your favorite — truffle ravioli and Cabernet, like last time,” Mary smiled warmly.But gradually, things become less appealing. Your attempts at variety or exploring something new are gently brushed aside: “Heard about that new sushi place, should we try it?” you suggest. Mary hesitates, “Remember last year? You said sushi wasn’t really your thing. Giorgio’s is safe. Why risk it?”Conversations start to feel repetitive, your identity locked to a cached version of yourself. Mary constantly cites your past preferences as proof of who you still are. The longer this goes on, the smaller your world feels… and comfort begins to curdle into confinement.Now, picture Mary isn’t human, but your personalized AI assistant.A new mode of hyper-personalizationWith OpenAI’s new memory upgrade, ChatGPT can now recall everything you’ve ever shared with it, indefinitely. Similarly, Google has opened the context window with “Infini-attention,” letting large language modelsreference infinite inputs with zero memory loss. And in consumer-facing tools like ChatGPT or Gemini, this now means persistent, personalized memory across conversations, unless you manually intervene. sales pitch is seductively simple: less friction, more relevance. Conversations that feel like continuity: “Systems that get to know you over your life,” as Sam Altman writes on X. Technology, finally, that meets you where you are.In the age of hyper-personalization — of the TikTok For You page, Spotify Wrapped, and Netflix Your Next Watch — a conversational AI product that remembers everything about you feels perfectly, perhaps dangerously, natural.Netflix “knows us.” And we’re conditioned to expect conversational AI to do the same.Forgetting, then, begins to look like a flaw. A failure to retain. A bug in the code. Especially in our own lives, we treat memory loss as a tragedy, clinging to photo albums and cloud backups to preserve what time tries to erase.But what if human forgetting is not a bug, but a feature? And what happens when we build machines that don’t forget, but are now helping shape the human minds that do?Forgetting is a feature of human memory“Infinite memory” runs against the very grain of what it means to be human. Cognitive science and evolutionary biology tell us that forgetting isn’t a design flaw, but a survival advantage. Our brains are not built to store everything. They’re built to let go: to blur the past, to misremember just enough to move forward.Our brains don’t archive data. They encode approximations. Memory is probabilistic, reconstructive, and inherently lossy. We misremember not because we’re broken, but because it makes us adaptable. Memory compresses and abstracts experience into usable shortcuts, heuristics that help us act fast, not recall perfectly.Evolution didn’t optimize our brains to store the past in high fidelity; it optimized us to survive the present. In early humans, remembering too much could be fatal: a brain caught up recalling a saber-tooth tiger’s precise location or exact color would hesitate, but a brain that knows riverbank = danger can act fast.Image generated by ChatGPT.This is why forgetting is essential to survival. Selective forgetting helps us prioritize the relevant, discard the outdated, and stay flexible in changing environments. It prevents us from becoming trapped by obsolete patterns or overwhelmed by noise.And it’s not passive decay. Neuroscience shows that forgetting is an active process: the brain regulates what to retrieve and what to suppress, clearing mental space to absorb new information. In his TED talk, neuroscientist Richard Morris describes the forgetting process as “the hippocampus doing its job… as it clears the desktop of your mind so that you’re ready for the next day to take in new information.”, this mental flexibility isn’t just for processing the past; forgetting allows us to imagine the future. Memory’s malleability gives us the ability to simulate, to envision, to choose differently next time. What we lose in accuracy, we gain in possibility.So when we ask why humans forget, the answer isn’t just functional. It’s existential. If we remembered everything, we wouldn’t be more intelligent. We’d still be standing at the riverbank, paralyzed by the precision of memories that no longer serve us.When forgetting is a “flaw” in AI memoryWhere nature embraced forgetting as a survival strategy, we now engineer machines that retain everything: your past prompts, preferences, corrections, and confessions.What sounds like a convenience, digital companions that “know you,” can quietly become a constraint. Unlike human memory, which fades and adapts, infinite memory stores information with fidelity and permanence. And as memory-equipped LLMs respond, they increasingly draw on a preserved version of you, even if that version is six months old and irrelevant.Sound familiar?This pattern of behavior reinforcement closely mirrors the personalization logic driving platforms like TikTok, Instagram, and Facebook. Extensive research has shown how these platforms amplify existing preferences, narrow user perspectives, and reduce exposure to new, challenging ideas — a phenomenon known as filter bubbles or echo chambers.Positive feedback loops are the engine of recommendation algorithms like TikTok, Netflix, and Spotify. From Medium.These feedback loops, optimized for engagement rather than novelty or growth, have been linked to documented consequences including ideological polarization, misinformation spread, and decreased critical thinking.Now, this same personalization logic is moving inward: from your feed to your conversations, and from what you consume to how you think.“Echo chamber to end all echo chambers”Just as the TikTok For You page algorithm predicts your next dopamine hit, memory-enabled LLMs predict and reinforce conversational patterns that align closely with your past behavior, keeping you comfortable inside your bubble of views and preferences.Jordan Gibbs, writing on the dangers of ChatGPT, notes that conversational AI is an “echo chamber to end all echo chambers.” Gibbs points out how even harmless-seeming positive reinforcement can quietly reshape user perceptions and restrict creative or critical thinking.Jordan Gibb’s conversation with ChatGPT from Medium.In one example, ChatGPT responds to Gibb’s claim of being one of the best chess players in the world not with skepticism or critical inquiry, but with encouragement and validation, highlighting how easily LLMs affirm bold, unverified assertions.And with infinite memory enabled, this is no longer a one-off interaction: the personal data point that, “You are one of the very best chess players in the world, ” risks becoming a fixed truth the model reflexively returns to, until your delusion, once tossed out in passing, becomes a cornerstone of your digital self. Not because it’s accurate, but because it was remembered, reinforced, and never challenged.When memory becomes fixed, identity becomes recursive. As we saw with our friend Mary, infinite memory doesn’t just remember our past; it nudges us to repeat it. And while the reinforcement may feel benign, personalized, or even comforting, the history of filter bubbles and echo chambers suggests that this kind of pattern replication rarely leaves room for transformation.What we lose when nothing is lostWhat begins as personalization can quietly become entrapment, not through control, but through familiarity. And in that familiarity, we begin to lose something essential: not just variety, but the very conditions that make change possible.Research in cognitive and developmental psychology shows that stepping outside one’s comfort zone is essential for growth, resilience, and adaptation. Yet, infinite-memory LLM systems, much like personalization algorithms, are engineered explicitly for comfort. They wrap users in a cocoon of sameness by continuously repeating familiar conversational patterns, reinforcing existing user preferences and biases, and avoiding content or ideas that might challenge or discomfort the user.Hyper-personalization traps us in a “comfort cocoon” that prevents from growing and transforming. From Earth.comWhile this engineered comfort may boost short-term satisfaction, its long-term effects are troubling. It replaces the discomfort necessary for cognitive growth with repetitive familiarity, effectively transforming your cognitive gym into a lazy river. Rather than stretching cognitive and emotional capacities, infinite-memory systems risk stagnating them, creating a psychological landscape devoid of intellectual curiosity and resilience.So, how do we break free from this? If the risks of infinite memory are clear, the path forward must be just as intentional. We must design LLM systems that don’t just remember, but also know when and why to forget.How we design to forgetIf the danger of infinite memory lies in its ability to trap us in our past, then the antidote must be rooted in intentional forgetting — systems that forget wisely, adaptively, and in ways aligned with human growth. But building such systems requires action across levels — from the people who use them to those who design and develop them.For users: reclaim agency over your digital selfJust as we now expect to “manage cookies” on websites, toggling consent checkboxes or adjusting ad settings, we may soon expect to manage our digital selves within LLM memory interfaces. But where cookies govern how our data is collected and used by entities, memory in conversational AI turns that data inward. Personal data is not just pipelines for targeted ads; they’re conversational mirrors, actively shaping how we think, remember, and express who we are. The stakes are higher.Memory-equipped LLMs like ChatGPT already offer tools for this. You can review what it remembers about you by going to Settings > Personalization > Memory > Manage. You can delete what’s outdated, refine what’s imprecise, and add what actually matters to who you are now. If something no longer reflects you, remove it. If something feels off, reframe it. If something is sensitive or exploratory, switch to a temporary chat and leave no trace.You can manage and disable memory within ChatGPT by visiting Settings > Personalization.You can also pause or disable memory entirely. Don’t be afraid to do it. There’s a quiet power in the clean slate: a freedom to experiment, shift, and show up as someone new.Guide the memory, don’t leave it ambient. Offer core memories that represent the direction you’re heading, not just the footprints you left behind.For UX designers: design for revision, not just retentionReclaiming memory is a personal act. But shaping how memory behaves in AI products is design decision. Infinite memory isn’t just a technical upgrade; it’s a cognitive interface. And UX designers are now curating the mental architecture of how people evolve, or get stuck.Forget “opt in” or “opt out.” Memory management shouldn’t live in buried toggles or forgotten settings menus. It should be active, visible, and intuitive: a first-class feature, not an afterthought. Users need interfaces that not only show what the system remembers, but also how those memories are shaping what they see, hear, and get suggested. Not just visibility, but influence tracing.ChatGPT’s current memory interface enables users to manage memories, but it is static and database-like.While ChatGPT’s memory UI offers user control over their memories, it reads like a black-and-white database: out or in. Instead of treating memory as a static archive, we should design it as a living layer, structured more like a sketchpad than a ledger: flexible and revisable. All of this is hypothetical, but here’s what it could look like:Memory Review Moments: Built-in check-ins that ask, “You haven’t referenced this in a while — keep, revise, or forget?” Like Rocket Money nudging you to review subscriptions, the system becomes a gentle co-editor, helping surface outdated or ambiguous context before it quietly reshapes future behavior.Time-Aware Metadata: Memories don’t age equally. Show users when something was last used, how often it comes up, or whether it’s quietly steering suggestions. Just like Spotify highlights “recently played,” memory interfaces could offer temporal context that makes stored data feel navigable and self-aware.Memory Tiers: Not all information deserves equal weight. Let users tag “Core Memories” that persist until manually removed, and set others as short-term or provisional — notes that decay unless reaffirmed.Inline Memory Controls: Bring memory into the flow of conversation. Imagine typing, and a quiet note appears: “This suggestion draws on your July planning — still accurate?” Like version history in Figma or comment nudges in Google Docs, these lightweight moments let users edit memory without switching contexts.Expiration Dates & Sunset Notices: Some memories should come with lifespans. Let users set expiration dates — “forget this in 30 days unless I say otherwise.” Like calendar events or temporary access links, this makes forgetting a designed act, not a technical gap.Image a Miro-like memory board where users could prioritize, annotate, and link memories.Sketchpad Interfaces: Finally, break free from the checkbox UI. Imagine memory as a visual canvas: clusters of ideas, color-coded threads, ephemeral notes. A place to link thoughts, add context, tag relevance. Think Miro meets Pinterest for your digital identity, a space that mirrors how we actually think, shift, and remember.When designers build memory this way, they create more than tools. They create mirrors with context, systems that grow with us instead of holding us still.For AI developers: engineer forgetting as a featureTo truly support transformation, UX needs infrastructure. The design must be backed by technical memory systems that are fluid, flexible, and capable of letting go. And that responsibility falls to developers: not just to build tools for remembering, but to engineer forgetting as a core function.This is the heart of my piece: we can’t talk about user agency, growth, or identity without addressing how memory works under the hood. Forgetting must be built into the LLM system itself, not as a failsafe, but as a feature.One promising approach, called adaptive forgetting, mimics how humans let go of unnecessary details while retaining important patterns and concepts. Researchers demonstrate that when LLMs periodically erase and retrain parts of their memory, especially early layers that store word associations, they become better at picking up new languages, adapting to new tasks, and doing so with less data and computing power.Photo by Valentin Tkach for Quanta MagazineAnother more accessible path forward is in Retrieval-Augmented Generation. A new method called SynapticRAG, inspired by the brain’s natural timing and memory mechanisms, adds a sense of temporality to AI memory. Models recall information not just based on content, but also on when it happened. Just like our brains prioritize recent memories, this method scores and updates AI memories based on both their relevance and relevance, allowing it to retrieve more meaningful, diverse, and context-rich information. Testing showed that this time-aware system outperforms traditional memory tools in multilingual conversations by up to 14.66% in accuracy, while also avoiding redundant or outdated responses.Together, adaptive forgetting and biologically inspired memory retrieval point toward a more human kind of AI: systems that learn continuously, update flexibly, and interact in ways that feel less like digital tape recorders and more like thoughtful, evolving collaborators.To grow, we must choose to forgetSo the pieces are all here: the architectural tools, the memory systems, the design patterns. We’ve shown that it’s technically possible for AI to forget. But the question isn’t just whether we can. It’s whether we will.Of course, not all AI systems need to forget. In high-stakes domains — medicine, law, scientific research — perfect recall can be life-saving. However, this essay is about a different kind of AI: the kind we bring into our daily lives. The ones we turn to for brainstorming, emotional support, writing help, or even casual companionship. These are the systems that assist us, observe us, and remember us. And if left unchecked, they may start to define us.We’ve already seen what happens when algorithms optimize for comfort. What begins as personalization becomes repetition. Sameness. Polarization. Now that logic is turning inward: no longer just curating our feeds, but shaping our conversations, our habits of thought, our sense of self. But we don’t have to follow the same path.We can build LLM systems that don’t just remember us, but help us evolve. Systems that challenge us to break patterns, to imagine differently, to change. Not to preserve who we were, but to make space for who we might yet become, just as our ancestors did.Not with perfect memory, but with the courage to forget.To grow, we must forget… but AI remembers everything was originally published in UX Collective on Medium, where people are continuing the conversation by highlighting and responding to this story.
    #grow #must #forget #but #remembers
    To grow, we must forget… but AI remembers everything
    To grow, we must forget… but now AI remembers everythingAI’s infinite memory could endanger how we think, grow, and imagine. And we can do something about it.Photo by Laura Fuhrman on UnsplashWhen Mary remembered too muchImagine your best friend — we’ll call her Mary — had perfect, infallible memory.At first, it feels wonderful. She remembers your favorite dishes, obscure movie quotes, even that exact shade of sweater you casually admired months ago. Dinner plans are effortless: “Booked us Giorgio’s again, your favorite — truffle ravioli and Cabernet, like last time,” Mary smiled warmly.But gradually, things become less appealing. Your attempts at variety or exploring something new are gently brushed aside: “Heard about that new sushi place, should we try it?” you suggest. Mary hesitates, “Remember last year? You said sushi wasn’t really your thing. Giorgio’s is safe. Why risk it?”Conversations start to feel repetitive, your identity locked to a cached version of yourself. Mary constantly cites your past preferences as proof of who you still are. The longer this goes on, the smaller your world feels… and comfort begins to curdle into confinement.Now, picture Mary isn’t human, but your personalized AI assistant.A new mode of hyper-personalizationWith OpenAI’s new memory upgrade, ChatGPT can now recall everything you’ve ever shared with it, indefinitely. Similarly, Google has opened the context window with “Infini-attention,” letting large language modelsreference infinite inputs with zero memory loss. And in consumer-facing tools like ChatGPT or Gemini, this now means persistent, personalized memory across conversations, unless you manually intervene. sales pitch is seductively simple: less friction, more relevance. Conversations that feel like continuity: “Systems that get to know you over your life,” as Sam Altman writes on X. Technology, finally, that meets you where you are.In the age of hyper-personalization — of the TikTok For You page, Spotify Wrapped, and Netflix Your Next Watch — a conversational AI product that remembers everything about you feels perfectly, perhaps dangerously, natural.Netflix “knows us.” And we’re conditioned to expect conversational AI to do the same.Forgetting, then, begins to look like a flaw. A failure to retain. A bug in the code. Especially in our own lives, we treat memory loss as a tragedy, clinging to photo albums and cloud backups to preserve what time tries to erase.But what if human forgetting is not a bug, but a feature? And what happens when we build machines that don’t forget, but are now helping shape the human minds that do?Forgetting is a feature of human memory“Infinite memory” runs against the very grain of what it means to be human. Cognitive science and evolutionary biology tell us that forgetting isn’t a design flaw, but a survival advantage. Our brains are not built to store everything. They’re built to let go: to blur the past, to misremember just enough to move forward.Our brains don’t archive data. They encode approximations. Memory is probabilistic, reconstructive, and inherently lossy. We misremember not because we’re broken, but because it makes us adaptable. Memory compresses and abstracts experience into usable shortcuts, heuristics that help us act fast, not recall perfectly.Evolution didn’t optimize our brains to store the past in high fidelity; it optimized us to survive the present. In early humans, remembering too much could be fatal: a brain caught up recalling a saber-tooth tiger’s precise location or exact color would hesitate, but a brain that knows riverbank = danger can act fast.Image generated by ChatGPT.This is why forgetting is essential to survival. Selective forgetting helps us prioritize the relevant, discard the outdated, and stay flexible in changing environments. It prevents us from becoming trapped by obsolete patterns or overwhelmed by noise.And it’s not passive decay. Neuroscience shows that forgetting is an active process: the brain regulates what to retrieve and what to suppress, clearing mental space to absorb new information. In his TED talk, neuroscientist Richard Morris describes the forgetting process as “the hippocampus doing its job… as it clears the desktop of your mind so that you’re ready for the next day to take in new information.”, this mental flexibility isn’t just for processing the past; forgetting allows us to imagine the future. Memory’s malleability gives us the ability to simulate, to envision, to choose differently next time. What we lose in accuracy, we gain in possibility.So when we ask why humans forget, the answer isn’t just functional. It’s existential. If we remembered everything, we wouldn’t be more intelligent. We’d still be standing at the riverbank, paralyzed by the precision of memories that no longer serve us.When forgetting is a “flaw” in AI memoryWhere nature embraced forgetting as a survival strategy, we now engineer machines that retain everything: your past prompts, preferences, corrections, and confessions.What sounds like a convenience, digital companions that “know you,” can quietly become a constraint. Unlike human memory, which fades and adapts, infinite memory stores information with fidelity and permanence. And as memory-equipped LLMs respond, they increasingly draw on a preserved version of you, even if that version is six months old and irrelevant.Sound familiar?This pattern of behavior reinforcement closely mirrors the personalization logic driving platforms like TikTok, Instagram, and Facebook. Extensive research has shown how these platforms amplify existing preferences, narrow user perspectives, and reduce exposure to new, challenging ideas — a phenomenon known as filter bubbles or echo chambers.Positive feedback loops are the engine of recommendation algorithms like TikTok, Netflix, and Spotify. From Medium.These feedback loops, optimized for engagement rather than novelty or growth, have been linked to documented consequences including ideological polarization, misinformation spread, and decreased critical thinking.Now, this same personalization logic is moving inward: from your feed to your conversations, and from what you consume to how you think.“Echo chamber to end all echo chambers”Just as the TikTok For You page algorithm predicts your next dopamine hit, memory-enabled LLMs predict and reinforce conversational patterns that align closely with your past behavior, keeping you comfortable inside your bubble of views and preferences.Jordan Gibbs, writing on the dangers of ChatGPT, notes that conversational AI is an “echo chamber to end all echo chambers.” Gibbs points out how even harmless-seeming positive reinforcement can quietly reshape user perceptions and restrict creative or critical thinking.Jordan Gibb’s conversation with ChatGPT from Medium.In one example, ChatGPT responds to Gibb’s claim of being one of the best chess players in the world not with skepticism or critical inquiry, but with encouragement and validation, highlighting how easily LLMs affirm bold, unverified assertions.And with infinite memory enabled, this is no longer a one-off interaction: the personal data point that, “You are one of the very best chess players in the world, ” risks becoming a fixed truth the model reflexively returns to, until your delusion, once tossed out in passing, becomes a cornerstone of your digital self. Not because it’s accurate, but because it was remembered, reinforced, and never challenged.When memory becomes fixed, identity becomes recursive. As we saw with our friend Mary, infinite memory doesn’t just remember our past; it nudges us to repeat it. And while the reinforcement may feel benign, personalized, or even comforting, the history of filter bubbles and echo chambers suggests that this kind of pattern replication rarely leaves room for transformation.What we lose when nothing is lostWhat begins as personalization can quietly become entrapment, not through control, but through familiarity. And in that familiarity, we begin to lose something essential: not just variety, but the very conditions that make change possible.Research in cognitive and developmental psychology shows that stepping outside one’s comfort zone is essential for growth, resilience, and adaptation. Yet, infinite-memory LLM systems, much like personalization algorithms, are engineered explicitly for comfort. They wrap users in a cocoon of sameness by continuously repeating familiar conversational patterns, reinforcing existing user preferences and biases, and avoiding content or ideas that might challenge or discomfort the user.Hyper-personalization traps us in a “comfort cocoon” that prevents from growing and transforming. From Earth.comWhile this engineered comfort may boost short-term satisfaction, its long-term effects are troubling. It replaces the discomfort necessary for cognitive growth with repetitive familiarity, effectively transforming your cognitive gym into a lazy river. Rather than stretching cognitive and emotional capacities, infinite-memory systems risk stagnating them, creating a psychological landscape devoid of intellectual curiosity and resilience.So, how do we break free from this? If the risks of infinite memory are clear, the path forward must be just as intentional. We must design LLM systems that don’t just remember, but also know when and why to forget.How we design to forgetIf the danger of infinite memory lies in its ability to trap us in our past, then the antidote must be rooted in intentional forgetting — systems that forget wisely, adaptively, and in ways aligned with human growth. But building such systems requires action across levels — from the people who use them to those who design and develop them.For users: reclaim agency over your digital selfJust as we now expect to “manage cookies” on websites, toggling consent checkboxes or adjusting ad settings, we may soon expect to manage our digital selves within LLM memory interfaces. But where cookies govern how our data is collected and used by entities, memory in conversational AI turns that data inward. Personal data is not just pipelines for targeted ads; they’re conversational mirrors, actively shaping how we think, remember, and express who we are. The stakes are higher.Memory-equipped LLMs like ChatGPT already offer tools for this. You can review what it remembers about you by going to Settings > Personalization > Memory > Manage. You can delete what’s outdated, refine what’s imprecise, and add what actually matters to who you are now. If something no longer reflects you, remove it. If something feels off, reframe it. If something is sensitive or exploratory, switch to a temporary chat and leave no trace.You can manage and disable memory within ChatGPT by visiting Settings > Personalization.You can also pause or disable memory entirely. Don’t be afraid to do it. There’s a quiet power in the clean slate: a freedom to experiment, shift, and show up as someone new.Guide the memory, don’t leave it ambient. Offer core memories that represent the direction you’re heading, not just the footprints you left behind.For UX designers: design for revision, not just retentionReclaiming memory is a personal act. But shaping how memory behaves in AI products is design decision. Infinite memory isn’t just a technical upgrade; it’s a cognitive interface. And UX designers are now curating the mental architecture of how people evolve, or get stuck.Forget “opt in” or “opt out.” Memory management shouldn’t live in buried toggles or forgotten settings menus. It should be active, visible, and intuitive: a first-class feature, not an afterthought. Users need interfaces that not only show what the system remembers, but also how those memories are shaping what they see, hear, and get suggested. Not just visibility, but influence tracing.ChatGPT’s current memory interface enables users to manage memories, but it is static and database-like.While ChatGPT’s memory UI offers user control over their memories, it reads like a black-and-white database: out or in. Instead of treating memory as a static archive, we should design it as a living layer, structured more like a sketchpad than a ledger: flexible and revisable. All of this is hypothetical, but here’s what it could look like:Memory Review Moments: Built-in check-ins that ask, “You haven’t referenced this in a while — keep, revise, or forget?” Like Rocket Money nudging you to review subscriptions, the system becomes a gentle co-editor, helping surface outdated or ambiguous context before it quietly reshapes future behavior.Time-Aware Metadata: Memories don’t age equally. Show users when something was last used, how often it comes up, or whether it’s quietly steering suggestions. Just like Spotify highlights “recently played,” memory interfaces could offer temporal context that makes stored data feel navigable and self-aware.Memory Tiers: Not all information deserves equal weight. Let users tag “Core Memories” that persist until manually removed, and set others as short-term or provisional — notes that decay unless reaffirmed.Inline Memory Controls: Bring memory into the flow of conversation. Imagine typing, and a quiet note appears: “This suggestion draws on your July planning — still accurate?” Like version history in Figma or comment nudges in Google Docs, these lightweight moments let users edit memory without switching contexts.Expiration Dates & Sunset Notices: Some memories should come with lifespans. Let users set expiration dates — “forget this in 30 days unless I say otherwise.” Like calendar events or temporary access links, this makes forgetting a designed act, not a technical gap.Image a Miro-like memory board where users could prioritize, annotate, and link memories.Sketchpad Interfaces: Finally, break free from the checkbox UI. Imagine memory as a visual canvas: clusters of ideas, color-coded threads, ephemeral notes. A place to link thoughts, add context, tag relevance. Think Miro meets Pinterest for your digital identity, a space that mirrors how we actually think, shift, and remember.When designers build memory this way, they create more than tools. They create mirrors with context, systems that grow with us instead of holding us still.For AI developers: engineer forgetting as a featureTo truly support transformation, UX needs infrastructure. The design must be backed by technical memory systems that are fluid, flexible, and capable of letting go. And that responsibility falls to developers: not just to build tools for remembering, but to engineer forgetting as a core function.This is the heart of my piece: we can’t talk about user agency, growth, or identity without addressing how memory works under the hood. Forgetting must be built into the LLM system itself, not as a failsafe, but as a feature.One promising approach, called adaptive forgetting, mimics how humans let go of unnecessary details while retaining important patterns and concepts. Researchers demonstrate that when LLMs periodically erase and retrain parts of their memory, especially early layers that store word associations, they become better at picking up new languages, adapting to new tasks, and doing so with less data and computing power.Photo by Valentin Tkach for Quanta MagazineAnother more accessible path forward is in Retrieval-Augmented Generation. A new method called SynapticRAG, inspired by the brain’s natural timing and memory mechanisms, adds a sense of temporality to AI memory. Models recall information not just based on content, but also on when it happened. Just like our brains prioritize recent memories, this method scores and updates AI memories based on both their relevance and relevance, allowing it to retrieve more meaningful, diverse, and context-rich information. Testing showed that this time-aware system outperforms traditional memory tools in multilingual conversations by up to 14.66% in accuracy, while also avoiding redundant or outdated responses.Together, adaptive forgetting and biologically inspired memory retrieval point toward a more human kind of AI: systems that learn continuously, update flexibly, and interact in ways that feel less like digital tape recorders and more like thoughtful, evolving collaborators.To grow, we must choose to forgetSo the pieces are all here: the architectural tools, the memory systems, the design patterns. We’ve shown that it’s technically possible for AI to forget. But the question isn’t just whether we can. It’s whether we will.Of course, not all AI systems need to forget. In high-stakes domains — medicine, law, scientific research — perfect recall can be life-saving. However, this essay is about a different kind of AI: the kind we bring into our daily lives. The ones we turn to for brainstorming, emotional support, writing help, or even casual companionship. These are the systems that assist us, observe us, and remember us. And if left unchecked, they may start to define us.We’ve already seen what happens when algorithms optimize for comfort. What begins as personalization becomes repetition. Sameness. Polarization. Now that logic is turning inward: no longer just curating our feeds, but shaping our conversations, our habits of thought, our sense of self. But we don’t have to follow the same path.We can build LLM systems that don’t just remember us, but help us evolve. Systems that challenge us to break patterns, to imagine differently, to change. Not to preserve who we were, but to make space for who we might yet become, just as our ancestors did.Not with perfect memory, but with the courage to forget.To grow, we must forget… but AI remembers everything was originally published in UX Collective on Medium, where people are continuing the conversation by highlighting and responding to this story. #grow #must #forget #but #remembers
    UXDESIGN.CC
    To grow, we must forget… but AI remembers everything
    To grow, we must forget… but now AI remembers everythingAI’s infinite memory could endanger how we think, grow, and imagine. And we can do something about it.Photo by Laura Fuhrman on UnsplashWhen Mary remembered too muchImagine your best friend — we’ll call her Mary — had perfect, infallible memory.At first, it feels wonderful. She remembers your favorite dishes, obscure movie quotes, even that exact shade of sweater you casually admired months ago. Dinner plans are effortless: “Booked us Giorgio’s again, your favorite — truffle ravioli and Cabernet, like last time,” Mary smiled warmly.But gradually, things become less appealing. Your attempts at variety or exploring something new are gently brushed aside: “Heard about that new sushi place, should we try it?” you suggest. Mary hesitates, “Remember last year? You said sushi wasn’t really your thing. Giorgio’s is safe. Why risk it?”Conversations start to feel repetitive, your identity locked to a cached version of yourself. Mary constantly cites your past preferences as proof of who you still are. The longer this goes on, the smaller your world feels… and comfort begins to curdle into confinement.Now, picture Mary isn’t human, but your personalized AI assistant.A new mode of hyper-personalizationWith OpenAI’s new memory upgrade, ChatGPT can now recall everything you’ve ever shared with it, indefinitely. Similarly, Google has opened the context window with “Infini-attention,” letting large language models (LLMs) reference infinite inputs with zero memory loss. And in consumer-facing tools like ChatGPT or Gemini, this now means persistent, personalized memory across conversations, unless you manually intervene.https://medium.com/media/f1f7978fb8d63f7a1e9f52f051808f44/hrefThe sales pitch is seductively simple: less friction, more relevance. Conversations that feel like continuity: “Systems that get to know you over your life,” as Sam Altman writes on X. Technology, finally, that meets you where you are.In the age of hyper-personalization — of the TikTok For You page, Spotify Wrapped, and Netflix Your Next Watch — a conversational AI product that remembers everything about you feels perfectly, perhaps dangerously, natural.Netflix “knows us.” And we’re conditioned to expect conversational AI to do the same.Forgetting, then, begins to look like a flaw. A failure to retain. A bug in the code. Especially in our own lives, we treat memory loss as a tragedy, clinging to photo albums and cloud backups to preserve what time tries to erase.But what if human forgetting is not a bug, but a feature? And what happens when we build machines that don’t forget, but are now helping shape the human minds that do?Forgetting is a feature of human memory“Infinite memory” runs against the very grain of what it means to be human. Cognitive science and evolutionary biology tell us that forgetting isn’t a design flaw, but a survival advantage. Our brains are not built to store everything. They’re built to let go: to blur the past, to misremember just enough to move forward.Our brains don’t archive data. They encode approximations. Memory is probabilistic, reconstructive, and inherently lossy. We misremember not because we’re broken, but because it makes us adaptable. Memory compresses and abstracts experience into usable shortcuts, heuristics that help us act fast, not recall perfectly.Evolution didn’t optimize our brains to store the past in high fidelity; it optimized us to survive the present. In early humans, remembering too much could be fatal: a brain caught up recalling a saber-tooth tiger’s precise location or exact color would hesitate, but a brain that knows riverbank = danger can act fast.Image generated by ChatGPT.This is why forgetting is essential to survival. Selective forgetting helps us prioritize the relevant, discard the outdated, and stay flexible in changing environments. It prevents us from becoming trapped by obsolete patterns or overwhelmed by noise.And it’s not passive decay. Neuroscience shows that forgetting is an active process: the brain regulates what to retrieve and what to suppress, clearing mental space to absorb new information. In his TED talk, neuroscientist Richard Morris describes the forgetting process as “the hippocampus doing its job… as it clears the desktop of your mind so that you’re ready for the next day to take in new information.”https://medium.com/media/e272064dd59f29c4ca35e808d39e4e72/hrefCrucially, this mental flexibility isn’t just for processing the past; forgetting allows us to imagine the future. Memory’s malleability gives us the ability to simulate, to envision, to choose differently next time. What we lose in accuracy, we gain in possibility.So when we ask why humans forget, the answer isn’t just functional. It’s existential. If we remembered everything, we wouldn’t be more intelligent. We’d still be standing at the riverbank, paralyzed by the precision of memories that no longer serve us.When forgetting is a “flaw” in AI memoryWhere nature embraced forgetting as a survival strategy, we now engineer machines that retain everything: your past prompts, preferences, corrections, and confessions.What sounds like a convenience, digital companions that “know you,” can quietly become a constraint. Unlike human memory, which fades and adapts, infinite memory stores information with fidelity and permanence. And as memory-equipped LLMs respond, they increasingly draw on a preserved version of you, even if that version is six months old and irrelevant.Sound familiar?This pattern of behavior reinforcement closely mirrors the personalization logic driving platforms like TikTok, Instagram, and Facebook. Extensive research has shown how these platforms amplify existing preferences, narrow user perspectives, and reduce exposure to new, challenging ideas — a phenomenon known as filter bubbles or echo chambers.Positive feedback loops are the engine of recommendation algorithms like TikTok, Netflix, and Spotify. From Medium.These feedback loops, optimized for engagement rather than novelty or growth, have been linked to documented consequences including ideological polarization, misinformation spread, and decreased critical thinking.Now, this same personalization logic is moving inward: from your feed to your conversations, and from what you consume to how you think.“Echo chamber to end all echo chambers”Just as the TikTok For You page algorithm predicts your next dopamine hit, memory-enabled LLMs predict and reinforce conversational patterns that align closely with your past behavior, keeping you comfortable inside your bubble of views and preferences.Jordan Gibbs, writing on the dangers of ChatGPT, notes that conversational AI is an “echo chamber to end all echo chambers.” Gibbs points out how even harmless-seeming positive reinforcement can quietly reshape user perceptions and restrict creative or critical thinking.Jordan Gibb’s conversation with ChatGPT from Medium.In one example, ChatGPT responds to Gibb’s claim of being one of the best chess players in the world not with skepticism or critical inquiry, but with encouragement and validation, highlighting how easily LLMs affirm bold, unverified assertions.And with infinite memory enabled, this is no longer a one-off interaction: the personal data point that, “You are one of the very best chess players in the world, ” risks becoming a fixed truth the model reflexively returns to, until your delusion, once tossed out in passing, becomes a cornerstone of your digital self. Not because it’s accurate, but because it was remembered, reinforced, and never challenged.When memory becomes fixed, identity becomes recursive. As we saw with our friend Mary, infinite memory doesn’t just remember our past; it nudges us to repeat it. And while the reinforcement may feel benign, personalized, or even comforting, the history of filter bubbles and echo chambers suggests that this kind of pattern replication rarely leaves room for transformation.What we lose when nothing is lostWhat begins as personalization can quietly become entrapment, not through control, but through familiarity. And in that familiarity, we begin to lose something essential: not just variety, but the very conditions that make change possible.Research in cognitive and developmental psychology shows that stepping outside one’s comfort zone is essential for growth, resilience, and adaptation. Yet, infinite-memory LLM systems, much like personalization algorithms, are engineered explicitly for comfort. They wrap users in a cocoon of sameness by continuously repeating familiar conversational patterns, reinforcing existing user preferences and biases, and avoiding content or ideas that might challenge or discomfort the user.Hyper-personalization traps us in a “comfort cocoon” that prevents from growing and transforming. From Earth.comWhile this engineered comfort may boost short-term satisfaction, its long-term effects are troubling. It replaces the discomfort necessary for cognitive growth with repetitive familiarity, effectively transforming your cognitive gym into a lazy river. Rather than stretching cognitive and emotional capacities, infinite-memory systems risk stagnating them, creating a psychological landscape devoid of intellectual curiosity and resilience.So, how do we break free from this? If the risks of infinite memory are clear, the path forward must be just as intentional. We must design LLM systems that don’t just remember, but also know when and why to forget.How we design to forgetIf the danger of infinite memory lies in its ability to trap us in our past, then the antidote must be rooted in intentional forgetting — systems that forget wisely, adaptively, and in ways aligned with human growth. But building such systems requires action across levels — from the people who use them to those who design and develop them.For users: reclaim agency over your digital selfJust as we now expect to “manage cookies” on websites, toggling consent checkboxes or adjusting ad settings, we may soon expect to manage our digital selves within LLM memory interfaces. But where cookies govern how our data is collected and used by entities, memory in conversational AI turns that data inward. Personal data is not just pipelines for targeted ads; they’re conversational mirrors, actively shaping how we think, remember, and express who we are. The stakes are higher.Memory-equipped LLMs like ChatGPT already offer tools for this. You can review what it remembers about you by going to Settings > Personalization > Memory > Manage. You can delete what’s outdated, refine what’s imprecise, and add what actually matters to who you are now. If something no longer reflects you, remove it. If something feels off, reframe it. If something is sensitive or exploratory, switch to a temporary chat and leave no trace.You can manage and disable memory within ChatGPT by visiting Settings > Personalization.You can also pause or disable memory entirely. Don’t be afraid to do it. There’s a quiet power in the clean slate: a freedom to experiment, shift, and show up as someone new.Guide the memory, don’t leave it ambient. Offer core memories that represent the direction you’re heading, not just the footprints you left behind.For UX designers: design for revision, not just retentionReclaiming memory is a personal act. But shaping how memory behaves in AI products is design decision. Infinite memory isn’t just a technical upgrade; it’s a cognitive interface. And UX designers are now curating the mental architecture of how people evolve, or get stuck.Forget “opt in” or “opt out.” Memory management shouldn’t live in buried toggles or forgotten settings menus. It should be active, visible, and intuitive: a first-class feature, not an afterthought. Users need interfaces that not only show what the system remembers, but also how those memories are shaping what they see, hear, and get suggested. Not just visibility, but influence tracing.ChatGPT’s current memory interface enables users to manage memories, but it is static and database-like.While ChatGPT’s memory UI offers user control over their memories, it reads like a black-and-white database: out or in. Instead of treating memory as a static archive, we should design it as a living layer, structured more like a sketchpad than a ledger: flexible and revisable. All of this is hypothetical, but here’s what it could look like:Memory Review Moments: Built-in check-ins that ask, “You haven’t referenced this in a while — keep, revise, or forget?” Like Rocket Money nudging you to review subscriptions, the system becomes a gentle co-editor, helping surface outdated or ambiguous context before it quietly reshapes future behavior.Time-Aware Metadata: Memories don’t age equally. Show users when something was last used, how often it comes up, or whether it’s quietly steering suggestions. Just like Spotify highlights “recently played,” memory interfaces could offer temporal context that makes stored data feel navigable and self-aware.Memory Tiers: Not all information deserves equal weight. Let users tag “Core Memories” that persist until manually removed, and set others as short-term or provisional — notes that decay unless reaffirmed.Inline Memory Controls: Bring memory into the flow of conversation. Imagine typing, and a quiet note appears: “This suggestion draws on your July planning — still accurate?” Like version history in Figma or comment nudges in Google Docs, these lightweight moments let users edit memory without switching contexts.Expiration Dates & Sunset Notices: Some memories should come with lifespans. Let users set expiration dates — “forget this in 30 days unless I say otherwise.” Like calendar events or temporary access links, this makes forgetting a designed act, not a technical gap.Image a Miro-like memory board where users could prioritize, annotate, and link memories.Sketchpad Interfaces: Finally, break free from the checkbox UI. Imagine memory as a visual canvas: clusters of ideas, color-coded threads, ephemeral notes. A place to link thoughts, add context, tag relevance. Think Miro meets Pinterest for your digital identity, a space that mirrors how we actually think, shift, and remember.When designers build memory this way, they create more than tools. They create mirrors with context, systems that grow with us instead of holding us still.For AI developers: engineer forgetting as a featureTo truly support transformation, UX needs infrastructure. The design must be backed by technical memory systems that are fluid, flexible, and capable of letting go. And that responsibility falls to developers: not just to build tools for remembering, but to engineer forgetting as a core function.This is the heart of my piece: we can’t talk about user agency, growth, or identity without addressing how memory works under the hood. Forgetting must be built into the LLM system itself, not as a failsafe, but as a feature.One promising approach, called adaptive forgetting, mimics how humans let go of unnecessary details while retaining important patterns and concepts. Researchers demonstrate that when LLMs periodically erase and retrain parts of their memory, especially early layers that store word associations, they become better at picking up new languages, adapting to new tasks, and doing so with less data and computing power.Photo by Valentin Tkach for Quanta MagazineAnother more accessible path forward is in Retrieval-Augmented Generation (RAG). A new method called SynapticRAG, inspired by the brain’s natural timing and memory mechanisms, adds a sense of temporality to AI memory. Models recall information not just based on content, but also on when it happened. Just like our brains prioritize recent memories, this method scores and updates AI memories based on both their relevance and relevance, allowing it to retrieve more meaningful, diverse, and context-rich information. Testing showed that this time-aware system outperforms traditional memory tools in multilingual conversations by up to 14.66% in accuracy, while also avoiding redundant or outdated responses.Together, adaptive forgetting and biologically inspired memory retrieval point toward a more human kind of AI: systems that learn continuously, update flexibly, and interact in ways that feel less like digital tape recorders and more like thoughtful, evolving collaborators.To grow, we must choose to forgetSo the pieces are all here: the architectural tools, the memory systems, the design patterns. We’ve shown that it’s technically possible for AI to forget. But the question isn’t just whether we can. It’s whether we will.Of course, not all AI systems need to forget. In high-stakes domains — medicine, law, scientific research — perfect recall can be life-saving. However, this essay is about a different kind of AI: the kind we bring into our daily lives. The ones we turn to for brainstorming, emotional support, writing help, or even casual companionship. These are the systems that assist us, observe us, and remember us. And if left unchecked, they may start to define us.We’ve already seen what happens when algorithms optimize for comfort. What begins as personalization becomes repetition. Sameness. Polarization. Now that logic is turning inward: no longer just curating our feeds, but shaping our conversations, our habits of thought, our sense of self. But we don’t have to follow the same path.We can build LLM systems that don’t just remember us, but help us evolve. Systems that challenge us to break patterns, to imagine differently, to change. Not to preserve who we were, but to make space for who we might yet become, just as our ancestors did.Not with perfect memory, but with the courage to forget.To grow, we must forget… but AI remembers everything was originally published in UX Collective on Medium, where people are continuing the conversation by highlighting and responding to this story.
    0 Yorumlar 0 hisse senetleri 0 önizleme
  • A Comprehensive Coding Guide to Crafting Advanced Round-Robin Multi-Agent Workflows with Microsoft AutoGen

    In this tutorial, we demonstrated how Microsoft’s AutoGen framework empowers developers to orchestrate complex, multi-agent workflows with minimal code. By leveraging AutoGen’s RoundRobinGroupChat and TeamTool abstractions, you can seamlessly assemble specialist assistants, such as Researchers, FactCheckers, Critics, Summarizers, and Editors, into a cohesive “DeepDive” tool. AutoGen handles the intricacies of turn‐taking, termination conditions, and streaming output, allowing you to focus on defining each agent’s expertise and system prompts rather than plumbing together callbacks or manual prompt chains. Whether conducting in‐depth research, validating facts, refining prose, or integrating third‐party tools, AutoGen provides a unified API that scales from simple two‐agent pipelines to elaborate, five‐agent collaboratives.
    !pip install -q autogen-agentchatautogen-extnest_asyncio
    We install the AutoGen AgentChat package with Gemini support, the OpenAI extension for API compatibility, and the nest_asyncio library to patch the notebook’s event loop, ensuring you have all the components needed to run asynchronous, multi-agent workflows in Colab.
    import os, nest_asyncio
    from getpass import getpass

    nest_asyncio.applyos.environ= getpassWe import and apply nest_asyncio to enable nested event loops in notebook environments, then securely prompt for your Gemini API key using getpass and store it in os.environ for authenticated model client access.
    from autogen_ext.models.openai import OpenAIChatCompletionClient

    model_client = OpenAIChatCompletionClientWe initialize an OpenAI‐compatible chat client pointed at Google’s Gemini by specifying the gemini-1.5-flash-8b model, injecting your stored Gemini API key, and setting api_type=”google”, giving you a ready-to-use model_client for downstream AutoGen agents.
    from autogen_agentchat.agents import AssistantAgent

    researcher = AssistantAgentfactchecker = AssistantAgentcritic = AssistantAgentsummarizer = AssistantAgenteditor = AssistantAgentWe define five specialized assistant agents, Researcher, FactChecker, Critic, Summarizer, and Editor, each initialized with a role-specific system message and the shared Gemini-powered model client, enabling them to gather information, respectively, verify accuracy, critique content, condense summaries, and polish language within the AutoGen workflow.
    from autogen_agentchat.teams import RoundRobinGroupChat
    from autogen_agentchat.conditions import MaxMessageTermination, TextMentionTermination

    max_msgs = MaxMessageTerminationtext_term = TextMentionTerminationtermination = max_msgs | text_term
    team = RoundRobinGroupChatWe import the RoundRobinGroupChat class along with two termination conditions, then compose a stop rule that fires after 20 total messages or when the Editor agent mentions “APPROVED.” Finally, it instantiates a round-robin team of the five specialized agents with that combined termination logic, enabling them to cycle through research, fact-checking, critique, summarization, and editing until one of the stop conditions is met.
    from autogen_agentchat.tools import TeamTool

    deepdive_tool = TeamToolWE wrap our RoundRobinGroupChat team in a TeamTool named “DeepDive” with a human-readable description, effectively packaging the entire multi-agent workflow into a single callable tool that other agents can invoke seamlessly.
    host = AssistantAgentWe create a “Host” assistant agent configured with the shared Gemini-powered model_client, grant it the DeepDive team tool for orchestrating in-depth research, and prime it with a system message that informs it of its ability to invoke the multi-agent DeepDive workflow.
    import asyncio

    async def run_deepdive:
    result = await host.runprintawait model_client.closetopic = "Impacts of Model Context Protocl on Agentic AI"
    loop = asyncio.get_event_looploop.run_until_complete)

    Finally, we define an asynchronous run_deepdive function that tells the Host agent to execute the DeepDive team tool on a given topic, prints the comprehensive result, and then closes the model client; it then grabs Colab’s existing asyncio loop and runs the coroutine to completion for a seamless, synchronous execution.
    In conclusion, integrating Google Gemini via AutoGen’s OpenAI‐compatible client and wrapping our multi‐agent team as a callable TeamTool gives us a powerful template for building highly modular and reusable workflows. AutoGen abstracts away event loop management, streaming responses, and termination logic, enabling us to iterate quickly on agent roles and overall orchestration. This advanced pattern streamlines the development of collaborative AI systems and lays the foundation for extending into retrieval pipelines, dynamic selectors, or conditional execution strategies.

    Check out the Notebook here. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 95k+ ML SubReddit and Subscribe to our Newsletter.
    Asif RazzaqWebsite |  + postsBioAsif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.Asif Razzaqhttps://www.marktechpost.com/author/6flvq/Microsoft AI Introduces Magentic-UI: An Open-Source Agent Prototype that Works with People to Complete Complex Tasks that Require Multi-Step Planning and Browser UseAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Anthropic Releases Claude Opus 4 and Claude Sonnet 4: A Technical Leap in Reasoning, Coding, and AI Agent DesignAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Technology Innovation Institute TII Releases Falcon-H1: Hybrid Transformer-SSM Language Models for Scalable, Multilingual, and Long-Context UnderstandingAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Google DeepMind Releases Gemma 3n: A Compact, High-Efficiency Multimodal AI Model for Real-Time On-Device Use
    #comprehensive #coding #guide #crafting #advanced
    A Comprehensive Coding Guide to Crafting Advanced Round-Robin Multi-Agent Workflows with Microsoft AutoGen
    In this tutorial, we demonstrated how Microsoft’s AutoGen framework empowers developers to orchestrate complex, multi-agent workflows with minimal code. By leveraging AutoGen’s RoundRobinGroupChat and TeamTool abstractions, you can seamlessly assemble specialist assistants, such as Researchers, FactCheckers, Critics, Summarizers, and Editors, into a cohesive “DeepDive” tool. AutoGen handles the intricacies of turn‐taking, termination conditions, and streaming output, allowing you to focus on defining each agent’s expertise and system prompts rather than plumbing together callbacks or manual prompt chains. Whether conducting in‐depth research, validating facts, refining prose, or integrating third‐party tools, AutoGen provides a unified API that scales from simple two‐agent pipelines to elaborate, five‐agent collaboratives. !pip install -q autogen-agentchatautogen-extnest_asyncio We install the AutoGen AgentChat package with Gemini support, the OpenAI extension for API compatibility, and the nest_asyncio library to patch the notebook’s event loop, ensuring you have all the components needed to run asynchronous, multi-agent workflows in Colab. import os, nest_asyncio from getpass import getpass nest_asyncio.applyos.environ= getpassWe import and apply nest_asyncio to enable nested event loops in notebook environments, then securely prompt for your Gemini API key using getpass and store it in os.environ for authenticated model client access. from autogen_ext.models.openai import OpenAIChatCompletionClient model_client = OpenAIChatCompletionClientWe initialize an OpenAI‐compatible chat client pointed at Google’s Gemini by specifying the gemini-1.5-flash-8b model, injecting your stored Gemini API key, and setting api_type=”google”, giving you a ready-to-use model_client for downstream AutoGen agents. from autogen_agentchat.agents import AssistantAgent researcher = AssistantAgentfactchecker = AssistantAgentcritic = AssistantAgentsummarizer = AssistantAgenteditor = AssistantAgentWe define five specialized assistant agents, Researcher, FactChecker, Critic, Summarizer, and Editor, each initialized with a role-specific system message and the shared Gemini-powered model client, enabling them to gather information, respectively, verify accuracy, critique content, condense summaries, and polish language within the AutoGen workflow. from autogen_agentchat.teams import RoundRobinGroupChat from autogen_agentchat.conditions import MaxMessageTermination, TextMentionTermination max_msgs = MaxMessageTerminationtext_term = TextMentionTerminationtermination = max_msgs | text_term team = RoundRobinGroupChatWe import the RoundRobinGroupChat class along with two termination conditions, then compose a stop rule that fires after 20 total messages or when the Editor agent mentions “APPROVED.” Finally, it instantiates a round-robin team of the five specialized agents with that combined termination logic, enabling them to cycle through research, fact-checking, critique, summarization, and editing until one of the stop conditions is met. from autogen_agentchat.tools import TeamTool deepdive_tool = TeamToolWE wrap our RoundRobinGroupChat team in a TeamTool named “DeepDive” with a human-readable description, effectively packaging the entire multi-agent workflow into a single callable tool that other agents can invoke seamlessly. host = AssistantAgentWe create a “Host” assistant agent configured with the shared Gemini-powered model_client, grant it the DeepDive team tool for orchestrating in-depth research, and prime it with a system message that informs it of its ability to invoke the multi-agent DeepDive workflow. import asyncio async def run_deepdive: result = await host.runprintawait model_client.closetopic = "Impacts of Model Context Protocl on Agentic AI" loop = asyncio.get_event_looploop.run_until_complete) Finally, we define an asynchronous run_deepdive function that tells the Host agent to execute the DeepDive team tool on a given topic, prints the comprehensive result, and then closes the model client; it then grabs Colab’s existing asyncio loop and runs the coroutine to completion for a seamless, synchronous execution. In conclusion, integrating Google Gemini via AutoGen’s OpenAI‐compatible client and wrapping our multi‐agent team as a callable TeamTool gives us a powerful template for building highly modular and reusable workflows. AutoGen abstracts away event loop management, streaming responses, and termination logic, enabling us to iterate quickly on agent roles and overall orchestration. This advanced pattern streamlines the development of collaborative AI systems and lays the foundation for extending into retrieval pipelines, dynamic selectors, or conditional execution strategies. Check out the Notebook here. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 95k+ ML SubReddit and Subscribe to our Newsletter. Asif RazzaqWebsite |  + postsBioAsif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.Asif Razzaqhttps://www.marktechpost.com/author/6flvq/Microsoft AI Introduces Magentic-UI: An Open-Source Agent Prototype that Works with People to Complete Complex Tasks that Require Multi-Step Planning and Browser UseAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Anthropic Releases Claude Opus 4 and Claude Sonnet 4: A Technical Leap in Reasoning, Coding, and AI Agent DesignAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Technology Innovation Institute TII Releases Falcon-H1: Hybrid Transformer-SSM Language Models for Scalable, Multilingual, and Long-Context UnderstandingAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Google DeepMind Releases Gemma 3n: A Compact, High-Efficiency Multimodal AI Model for Real-Time On-Device Use #comprehensive #coding #guide #crafting #advanced
    WWW.MARKTECHPOST.COM
    A Comprehensive Coding Guide to Crafting Advanced Round-Robin Multi-Agent Workflows with Microsoft AutoGen
    In this tutorial, we demonstrated how Microsoft’s AutoGen framework empowers developers to orchestrate complex, multi-agent workflows with minimal code. By leveraging AutoGen’s RoundRobinGroupChat and TeamTool abstractions, you can seamlessly assemble specialist assistants, such as Researchers, FactCheckers, Critics, Summarizers, and Editors, into a cohesive “DeepDive” tool. AutoGen handles the intricacies of turn‐taking, termination conditions, and streaming output, allowing you to focus on defining each agent’s expertise and system prompts rather than plumbing together callbacks or manual prompt chains. Whether conducting in‐depth research, validating facts, refining prose, or integrating third‐party tools, AutoGen provides a unified API that scales from simple two‐agent pipelines to elaborate, five‐agent collaboratives. !pip install -q autogen-agentchat[gemini] autogen-ext[openai] nest_asyncio We install the AutoGen AgentChat package with Gemini support, the OpenAI extension for API compatibility, and the nest_asyncio library to patch the notebook’s event loop, ensuring you have all the components needed to run asynchronous, multi-agent workflows in Colab. import os, nest_asyncio from getpass import getpass nest_asyncio.apply() os.environ["GEMINI_API_KEY"] = getpass("Enter your Gemini API key: ") We import and apply nest_asyncio to enable nested event loops in notebook environments, then securely prompt for your Gemini API key using getpass and store it in os.environ for authenticated model client access. from autogen_ext.models.openai import OpenAIChatCompletionClient model_client = OpenAIChatCompletionClient( model="gemini-1.5-flash-8b", api_key=os.environ["GEMINI_API_KEY"], api_type="google", ) We initialize an OpenAI‐compatible chat client pointed at Google’s Gemini by specifying the gemini-1.5-flash-8b model, injecting your stored Gemini API key, and setting api_type=”google”, giving you a ready-to-use model_client for downstream AutoGen agents. from autogen_agentchat.agents import AssistantAgent researcher = AssistantAgent(name="Researcher", system_message="Gather and summarize factual info.", model_client=model_client) factchecker = AssistantAgent(name="FactChecker", system_message="Verify facts and cite sources.", model_client=model_client) critic = AssistantAgent(name="Critic", system_message="Critique clarity and logic.", model_client=model_client) summarizer = AssistantAgent(name="Summarizer",system_message="Condense into a brief executive summary.", model_client=model_client) editor = AssistantAgent(name="Editor", system_message="Polish language and signal APPROVED when done.", model_client=model_client) We define five specialized assistant agents, Researcher, FactChecker, Critic, Summarizer, and Editor, each initialized with a role-specific system message and the shared Gemini-powered model client, enabling them to gather information, respectively, verify accuracy, critique content, condense summaries, and polish language within the AutoGen workflow. from autogen_agentchat.teams import RoundRobinGroupChat from autogen_agentchat.conditions import MaxMessageTermination, TextMentionTermination max_msgs = MaxMessageTermination(max_messages=20) text_term = TextMentionTermination(text="APPROVED", sources=["Editor"]) termination = max_msgs | text_term team = RoundRobinGroupChat( participants=[researcher, factchecker, critic, summarizer, editor], termination_condition=termination ) We import the RoundRobinGroupChat class along with two termination conditions, then compose a stop rule that fires after 20 total messages or when the Editor agent mentions “APPROVED.” Finally, it instantiates a round-robin team of the five specialized agents with that combined termination logic, enabling them to cycle through research, fact-checking, critique, summarization, and editing until one of the stop conditions is met. from autogen_agentchat.tools import TeamTool deepdive_tool = TeamTool(team=team, name="DeepDive", description="Collaborative multi-agent deep dive") WE wrap our RoundRobinGroupChat team in a TeamTool named “DeepDive” with a human-readable description, effectively packaging the entire multi-agent workflow into a single callable tool that other agents can invoke seamlessly. host = AssistantAgent( name="Host", model_client=model_client, tools=[deepdive_tool], system_message="You have access to a DeepDive tool for in-depth research." ) We create a “Host” assistant agent configured with the shared Gemini-powered model_client, grant it the DeepDive team tool for orchestrating in-depth research, and prime it with a system message that informs it of its ability to invoke the multi-agent DeepDive workflow. import asyncio async def run_deepdive(topic: str): result = await host.run(task=f"Deep dive on: {topic}") print("🔍 DeepDive result:\n", result) await model_client.close() topic = "Impacts of Model Context Protocl on Agentic AI" loop = asyncio.get_event_loop() loop.run_until_complete(run_deepdive(topic)) Finally, we define an asynchronous run_deepdive function that tells the Host agent to execute the DeepDive team tool on a given topic, prints the comprehensive result, and then closes the model client; it then grabs Colab’s existing asyncio loop and runs the coroutine to completion for a seamless, synchronous execution. In conclusion, integrating Google Gemini via AutoGen’s OpenAI‐compatible client and wrapping our multi‐agent team as a callable TeamTool gives us a powerful template for building highly modular and reusable workflows. AutoGen abstracts away event loop management (with nest_asyncio), streaming responses, and termination logic, enabling us to iterate quickly on agent roles and overall orchestration. This advanced pattern streamlines the development of collaborative AI systems and lays the foundation for extending into retrieval pipelines, dynamic selectors, or conditional execution strategies. Check out the Notebook here. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 95k+ ML SubReddit and Subscribe to our Newsletter. Asif RazzaqWebsite |  + postsBioAsif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.Asif Razzaqhttps://www.marktechpost.com/author/6flvq/Microsoft AI Introduces Magentic-UI: An Open-Source Agent Prototype that Works with People to Complete Complex Tasks that Require Multi-Step Planning and Browser UseAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Anthropic Releases Claude Opus 4 and Claude Sonnet 4: A Technical Leap in Reasoning, Coding, and AI Agent DesignAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Technology Innovation Institute TII Releases Falcon-H1: Hybrid Transformer-SSM Language Models for Scalable, Multilingual, and Long-Context UnderstandingAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Google DeepMind Releases Gemma 3n: A Compact, High-Efficiency Multimodal AI Model for Real-Time On-Device Use
    0 Yorumlar 0 hisse senetleri 0 önizleme
  • Abstracts: Zero-shot models in single-cell biology with Alex Lu

    TranscriptGRETCHEN HUIZINGA: Welcome to Abstracts, a Microsoft Research Podcast that puts the spotlight on world-class research in brief. I’m Gretchen Huizinga. In this series, members of the research community at Microsoft give us a quick snapshot – or a podcast abstract – of their new and noteworthy papers. On today’s episode, I’m talking to Alex Lu, a senior researcher at Microsoft Research and co-author of a paper called Assessing the Limits of Zero Shot Foundation Models in Single-cell Biology. Alex Lu, wonderful to have you on the podcast. Welcome to Abstracts! 

    ALEX LU: Yeah, I’m really excited to be joining you today. 
    HUIZINGA: So let’s start with a little background of your work. In just a few sentences, tell us about your study and more importantly, why it matters. 
    LU: Absolutely. And before I dive in, I want to give a shout out to the MSR research intern who actually did this work. This was led by Kasia Kedzierska, who interned with us two summers ago in 2023, and she’s the lead author on the study. But basically, in this research, we study single-cell foundation models, which have really recently rocked the world of biology, because they basically claim to be able to use AI to unlock understanding about single-cell biology. Biologists for a myriad of applications, everything from understanding how single cells differentiate into different kinds of cells, to discovering new drugs for cancer, will conduct experiments where they measure how much of every gene is expressed inside of just one single cell. So these experiments give us a powerful view into the cell’s internal state. But measurements from these experiments are incredibly complex. There are about 20,000 different human genes. So you get this really long chain of numbers that measure how much there is of 20,000 different genes. So deriving meaning from this really long chain of numbers is really difficult. And single-cell foundation models claim to be capable of unraveling deeper insights than ever before. So that’s the claim that these works have made. And in our recent paper, we showed that these models may actually not live up to these claims. Basically, we showed that single-cell foundation models perform worse in settings that are fundamental to biological discovery than much simpler machine learning and statistical methods that were used in the field before single-cell foundation models emerged and are the go-to standard for unpacking meaning from these complicated experiments. So in a nutshell, we should care about these results because it has implications on the toolkits that biologists use to understand their experiments. Our work suggests that single-cell foundation models may not be appropriate for practical use just yet, at least in the discovery applications that we cover. 
    HUIZINGA: Well, let’s go a little deeper there. Generative pre-trained transformer models, GPTs, are relatively new on the research scene in terms of how they’re being used in novel applications, which is what you’re interested in, like single-cell biology. So I’m curious, just sort of as a foundation, what other research has already been done in this area, and how does this study illuminate or build on it? 
    LU: Absolutely. Okay, so we were the first to notice and document this issue in single-cell foundation models, specifically. And this is because that we have proposed evaluation methods that, while are common in other areas of AI, have yet to be commonly used to evaluate single-cell foundation models. We performed something called zero-shot evaluation on these models. Prior to our work, most works evaluated single-cell foundation models with fine tuning. And the way to understand this is because single-cell foundation models are trained in a way that tries to expose these models to millions of single-cells. But because you’re exposing them to a large amount of data, you can’t really rely upon this data being annotated or like labeled in any particular fashion then. So in order for them to actually do the specialized tasks that are useful for biologists, you typically have to add on a second training phase. We call this the fine-tuning phase, where you have a smaller number of single cells, but now they are actually labeled with the specialized tasks that you want the model to perform. So most people, they typically evaluate the performance of single-cell models after they fine-tune these models. However, what we noticed is that this evaluating these fine-tuned models has several problems. First, it might not actually align with how these models are actually going to be used by biologists then. A critical distinction in biology is that we’re not just trying to interact with an agent that has access to knowledge through its pre-training, we’re trying to extend these models to discover new biology beyond the sphere of influence then. And so in many cases, the point of using these models, the point of analysis, is to explore the data with the goal of potentially discovering something new about the single cell that the biologists worked with that they weren’t aware of before. So in these kinds of cases, it is really tough to fine-tune a model. There’s a bit of a chicken and egg problem going on. If you don’t know, for example, there’s a new kind of cell in the data, you can’t really instruct the model to help us identify these kinds of new cells. So in other words, fine-tuning these models for those tasks essentially becomes impossible then. So the second issue is that evaluations on fine-tuned models can sometimes mislead us in our ability to understand how these models are working. So for example, the claim behind single-cell foundation model papers is that these models learn a foundation of biological knowledge by being exposed to millions of single cells in its first training phase, right? But it’s possible when you fine-tune a model, it may just be that any performance increases that you see using the model is simply because that you’re using a massive model that is really sophisticated, really large. And even if there’s any exposure to any cells at all then, that model is going to do perfectly fine then. So going back to our paper, what’s really different about this paper is that we propose zero-shot evaluation for these models. What that means is that we do not fine-tune the model at all, and instead we keep the model frozen during the analysis step. So how we specialize it to be a downstream task instead is that we extract the model’s internal embedding of single-cell data, which is essentially a numerical vector that contains information that the model is extracting and organizing from input data. So it’s essentially how the model perceives single-cell data and how it’s organizing in its own internal state. So basically, this is the better way for us to test the claim that single-cell foundation models are learning foundational biological insights. Because if they actually are learning these insights, they should be present in the models embedding space even before we fine-tune the model. 
    HUIZINGA: Well, let’s talk about methodology on this particular study. You focused on assessing existing models in zero-shot learning for single-cell biology. How did you go about evaluating these models? 
    LU: Yes, so let’s dive deeper into how zero-shot evaluations are conducted, okay? So the premise here is that we’re relying upon the fact that if these models are fully learning foundational biological insights, if we take the model’s internal representation of cells, then cells that are biologically similar should be close in that internal representation, where cells that are biologically distinct should be further apart. And that is exactly what we tested in our study. We compared two popular single-cell foundation models and importantly, we compared these models against older and reliable tools that biologists have used for exploratory analyses. So these include simpler machine learning methods like scVI, statistical algorithms like Harmony, and even basic data pre-processing steps, just like filtering your data down to a more robust subset of genes, then. So basically, we tested embeddings from our two single-cell foundation models against this baseline in a variety of settings. And we tested the hypothesis that biologically similar cells should be similar across these distinct methods across these datasets. 
    HUIZINGA: Well, and as you as you did the testing, you obviously were aiming towards research findings, which is my favorite part of a research paper, so tell us what you did find and what you feel the most important takeaways of this paper are. 
    LU: Absolutely. So in a nutshell, we found that these two newly proposed single-cell foundation models substantially underperformed compared to older methods then. So to contextualize why that is such a surprising result, there is a lot of hype around these methods. So basically, I think that,yeah, it’s a very surprising result, given how hyped these models are and how people were already adopting them. But our results basically caution that these shouldn’t really be adopted for these use purposes. 
    HUIZINGA: Yeah, so this is serious real-world impact here in terms of if models are being adopted and adapted in these applications, how reliable are they, et cetera? So given that, who would you say benefits most from what you’ve discovered in this paper and why? 
    LU: Okay, so two ways, right? So I think this has at least immediate implications on the way that we do discovery in biology. And as I’ve discussed, these experiments are used for cases that have practical impact, drug discovery applications, investigations into basic biology, then. But let’s also talk about the impact for methodologists, people who are trying to improve these single-cell foundation models, right? I think at the base, they’re really excited proposals. Because if you look at what some of the prior and less sophisticated methods couldn’t do, they tended to be more bespoke. So the excitement of single-cell foundation models is that you have this general-purpose model that can be used for everything and while they’re not living up to that purpose just now, just currently, I think that it’s important that we continue to bank onto that vision, right? So if you look at our contributions in that area, where single-cell foundation models are a really new proposal, so it makes sense that we may not know how to fully evaluate them just yet then. So you can view our work as basically being a step towards more rigorous evaluation of these models. Now that we did this experiment, I think the methodologists know to use this as a signal on how to improve the models and if they’re going in the right direction. And in fact, you are seeing more and more papers adopt zero-shot evaluations since we put out our paper then. And so this essentially helps future computer scientists that are working on single-cell foundation models know how to train better models. 
    HUIZINGA: That said, Alex, finally, what are the outstanding challenges that you identified for zero-shot learning research in biology, and what foundation might this paper lay for future research agendas in the field? 
    LU: Yeah, absolutely. So now that we’ve shown single-cell foundation models don’t necessarily perform well, I think the natural question on everyone’s mind is how do we actually train single-cell foundation models that live up to that vision, that can perform in helping us discover new biology then? So I think in the short term, yeah, we’re actively investigating many hypotheses in this area. So for example, my colleagues, Lorin Crawford and Ava Amini, who were co-authors in the paper, recently put out a pre-print understanding how training data composition impacts model performance. And so one of the surprising findings that they had was that many of the training data sets that people used to train single-cell foundation models are highly redundant, to the point that you can even sample just a tiny fraction of the data and get basically the same performance then. But you can also look forward to many other explorations in this area as we continue to develop this research at the end of the day. But also zooming out into the bigger picture, I think one major takeaway from this paper is that developing AI methods for biology requires thought about the context of use, right? I mean, this is obvious for any AI method then, but I think people have gotten just too used to taking methods that work out there for natural vision or natural language maybe in the consumer domain and then extrapolating these methods to biology and expecting that they will work in the same way then, right? So for example, one reason why zero-shot evaluation was not routine practice for single-cell foundation models prior to our work, I mean, we were the first to fully establish that as a practice for the field, was because I think people who have been working in AI for biology have been looking to these more mainstream AI domains to shape their work then. And so with single-cell foundation models, many of these models are adopted from large language models with natural language processing, recycling the exact same architecture, the exact same code, basically just recycling practices in that field then. So when you look at like practices in like more mainstream domains, zero-shot evaluation is definitely explored in those domains, but it’s more of like a niche instead of being considered central to model understanding. So again, because biology is different from mainstream language processing, it’s a scientific discipline, zero-shot evaluation becomes much more important, and you have no choice but to use these models, zero-shot then. So in other words, I think that we need to be thinking carefully about what it is that makes training a model for biology different from training a model, for example, for consumer purposes. HUIZINGA: Alex Lu, thanks for joining us today, and to our listeners, thanks for tuning in. If you want to read this paper, you can find a link at aka.ms/Abstracts, or you can read it on the Genome Biology website. See you next time on Abstracts!  
    #abstracts #zeroshot #models #singlecell #biology
    Abstracts: Zero-shot models in single-cell biology with Alex Lu
    TranscriptGRETCHEN HUIZINGA: Welcome to Abstracts, a Microsoft Research Podcast that puts the spotlight on world-class research in brief. I’m Gretchen Huizinga. In this series, members of the research community at Microsoft give us a quick snapshot – or a podcast abstract – of their new and noteworthy papers. On today’s episode, I’m talking to Alex Lu, a senior researcher at Microsoft Research and co-author of a paper called Assessing the Limits of Zero Shot Foundation Models in Single-cell Biology. Alex Lu, wonderful to have you on the podcast. Welcome to Abstracts!  ALEX LU: Yeah, I’m really excited to be joining you today.  HUIZINGA: So let’s start with a little background of your work. In just a few sentences, tell us about your study and more importantly, why it matters.  LU: Absolutely. And before I dive in, I want to give a shout out to the MSR research intern who actually did this work. This was led by Kasia Kedzierska, who interned with us two summers ago in 2023, and she’s the lead author on the study. But basically, in this research, we study single-cell foundation models, which have really recently rocked the world of biology, because they basically claim to be able to use AI to unlock understanding about single-cell biology. Biologists for a myriad of applications, everything from understanding how single cells differentiate into different kinds of cells, to discovering new drugs for cancer, will conduct experiments where they measure how much of every gene is expressed inside of just one single cell. So these experiments give us a powerful view into the cell’s internal state. But measurements from these experiments are incredibly complex. There are about 20,000 different human genes. So you get this really long chain of numbers that measure how much there is of 20,000 different genes. So deriving meaning from this really long chain of numbers is really difficult. And single-cell foundation models claim to be capable of unraveling deeper insights than ever before. So that’s the claim that these works have made. And in our recent paper, we showed that these models may actually not live up to these claims. Basically, we showed that single-cell foundation models perform worse in settings that are fundamental to biological discovery than much simpler machine learning and statistical methods that were used in the field before single-cell foundation models emerged and are the go-to standard for unpacking meaning from these complicated experiments. So in a nutshell, we should care about these results because it has implications on the toolkits that biologists use to understand their experiments. Our work suggests that single-cell foundation models may not be appropriate for practical use just yet, at least in the discovery applications that we cover.  HUIZINGA: Well, let’s go a little deeper there. Generative pre-trained transformer models, GPTs, are relatively new on the research scene in terms of how they’re being used in novel applications, which is what you’re interested in, like single-cell biology. So I’m curious, just sort of as a foundation, what other research has already been done in this area, and how does this study illuminate or build on it?  LU: Absolutely. Okay, so we were the first to notice and document this issue in single-cell foundation models, specifically. And this is because that we have proposed evaluation methods that, while are common in other areas of AI, have yet to be commonly used to evaluate single-cell foundation models. We performed something called zero-shot evaluation on these models. Prior to our work, most works evaluated single-cell foundation models with fine tuning. And the way to understand this is because single-cell foundation models are trained in a way that tries to expose these models to millions of single-cells. But because you’re exposing them to a large amount of data, you can’t really rely upon this data being annotated or like labeled in any particular fashion then. So in order for them to actually do the specialized tasks that are useful for biologists, you typically have to add on a second training phase. We call this the fine-tuning phase, where you have a smaller number of single cells, but now they are actually labeled with the specialized tasks that you want the model to perform. So most people, they typically evaluate the performance of single-cell models after they fine-tune these models. However, what we noticed is that this evaluating these fine-tuned models has several problems. First, it might not actually align with how these models are actually going to be used by biologists then. A critical distinction in biology is that we’re not just trying to interact with an agent that has access to knowledge through its pre-training, we’re trying to extend these models to discover new biology beyond the sphere of influence then. And so in many cases, the point of using these models, the point of analysis, is to explore the data with the goal of potentially discovering something new about the single cell that the biologists worked with that they weren’t aware of before. So in these kinds of cases, it is really tough to fine-tune a model. There’s a bit of a chicken and egg problem going on. If you don’t know, for example, there’s a new kind of cell in the data, you can’t really instruct the model to help us identify these kinds of new cells. So in other words, fine-tuning these models for those tasks essentially becomes impossible then. So the second issue is that evaluations on fine-tuned models can sometimes mislead us in our ability to understand how these models are working. So for example, the claim behind single-cell foundation model papers is that these models learn a foundation of biological knowledge by being exposed to millions of single cells in its first training phase, right? But it’s possible when you fine-tune a model, it may just be that any performance increases that you see using the model is simply because that you’re using a massive model that is really sophisticated, really large. And even if there’s any exposure to any cells at all then, that model is going to do perfectly fine then. So going back to our paper, what’s really different about this paper is that we propose zero-shot evaluation for these models. What that means is that we do not fine-tune the model at all, and instead we keep the model frozen during the analysis step. So how we specialize it to be a downstream task instead is that we extract the model’s internal embedding of single-cell data, which is essentially a numerical vector that contains information that the model is extracting and organizing from input data. So it’s essentially how the model perceives single-cell data and how it’s organizing in its own internal state. So basically, this is the better way for us to test the claim that single-cell foundation models are learning foundational biological insights. Because if they actually are learning these insights, they should be present in the models embedding space even before we fine-tune the model.  HUIZINGA: Well, let’s talk about methodology on this particular study. You focused on assessing existing models in zero-shot learning for single-cell biology. How did you go about evaluating these models?  LU: Yes, so let’s dive deeper into how zero-shot evaluations are conducted, okay? So the premise here is that we’re relying upon the fact that if these models are fully learning foundational biological insights, if we take the model’s internal representation of cells, then cells that are biologically similar should be close in that internal representation, where cells that are biologically distinct should be further apart. And that is exactly what we tested in our study. We compared two popular single-cell foundation models and importantly, we compared these models against older and reliable tools that biologists have used for exploratory analyses. So these include simpler machine learning methods like scVI, statistical algorithms like Harmony, and even basic data pre-processing steps, just like filtering your data down to a more robust subset of genes, then. So basically, we tested embeddings from our two single-cell foundation models against this baseline in a variety of settings. And we tested the hypothesis that biologically similar cells should be similar across these distinct methods across these datasets.  HUIZINGA: Well, and as you as you did the testing, you obviously were aiming towards research findings, which is my favorite part of a research paper, so tell us what you did find and what you feel the most important takeaways of this paper are.  LU: Absolutely. So in a nutshell, we found that these two newly proposed single-cell foundation models substantially underperformed compared to older methods then. So to contextualize why that is such a surprising result, there is a lot of hype around these methods. So basically, I think that,yeah, it’s a very surprising result, given how hyped these models are and how people were already adopting them. But our results basically caution that these shouldn’t really be adopted for these use purposes.  HUIZINGA: Yeah, so this is serious real-world impact here in terms of if models are being adopted and adapted in these applications, how reliable are they, et cetera? So given that, who would you say benefits most from what you’ve discovered in this paper and why?  LU: Okay, so two ways, right? So I think this has at least immediate implications on the way that we do discovery in biology. And as I’ve discussed, these experiments are used for cases that have practical impact, drug discovery applications, investigations into basic biology, then. But let’s also talk about the impact for methodologists, people who are trying to improve these single-cell foundation models, right? I think at the base, they’re really excited proposals. Because if you look at what some of the prior and less sophisticated methods couldn’t do, they tended to be more bespoke. So the excitement of single-cell foundation models is that you have this general-purpose model that can be used for everything and while they’re not living up to that purpose just now, just currently, I think that it’s important that we continue to bank onto that vision, right? So if you look at our contributions in that area, where single-cell foundation models are a really new proposal, so it makes sense that we may not know how to fully evaluate them just yet then. So you can view our work as basically being a step towards more rigorous evaluation of these models. Now that we did this experiment, I think the methodologists know to use this as a signal on how to improve the models and if they’re going in the right direction. And in fact, you are seeing more and more papers adopt zero-shot evaluations since we put out our paper then. And so this essentially helps future computer scientists that are working on single-cell foundation models know how to train better models.  HUIZINGA: That said, Alex, finally, what are the outstanding challenges that you identified for zero-shot learning research in biology, and what foundation might this paper lay for future research agendas in the field?  LU: Yeah, absolutely. So now that we’ve shown single-cell foundation models don’t necessarily perform well, I think the natural question on everyone’s mind is how do we actually train single-cell foundation models that live up to that vision, that can perform in helping us discover new biology then? So I think in the short term, yeah, we’re actively investigating many hypotheses in this area. So for example, my colleagues, Lorin Crawford and Ava Amini, who were co-authors in the paper, recently put out a pre-print understanding how training data composition impacts model performance. And so one of the surprising findings that they had was that many of the training data sets that people used to train single-cell foundation models are highly redundant, to the point that you can even sample just a tiny fraction of the data and get basically the same performance then. But you can also look forward to many other explorations in this area as we continue to develop this research at the end of the day. But also zooming out into the bigger picture, I think one major takeaway from this paper is that developing AI methods for biology requires thought about the context of use, right? I mean, this is obvious for any AI method then, but I think people have gotten just too used to taking methods that work out there for natural vision or natural language maybe in the consumer domain and then extrapolating these methods to biology and expecting that they will work in the same way then, right? So for example, one reason why zero-shot evaluation was not routine practice for single-cell foundation models prior to our work, I mean, we were the first to fully establish that as a practice for the field, was because I think people who have been working in AI for biology have been looking to these more mainstream AI domains to shape their work then. And so with single-cell foundation models, many of these models are adopted from large language models with natural language processing, recycling the exact same architecture, the exact same code, basically just recycling practices in that field then. So when you look at like practices in like more mainstream domains, zero-shot evaluation is definitely explored in those domains, but it’s more of like a niche instead of being considered central to model understanding. So again, because biology is different from mainstream language processing, it’s a scientific discipline, zero-shot evaluation becomes much more important, and you have no choice but to use these models, zero-shot then. So in other words, I think that we need to be thinking carefully about what it is that makes training a model for biology different from training a model, for example, for consumer purposes. HUIZINGA: Alex Lu, thanks for joining us today, and to our listeners, thanks for tuning in. If you want to read this paper, you can find a link at aka.ms/Abstracts, or you can read it on the Genome Biology website. See you next time on Abstracts!   #abstracts #zeroshot #models #singlecell #biology
    WWW.MICROSOFT.COM
    Abstracts: Zero-shot models in single-cell biology with Alex Lu
    Transcript [MUSIC] GRETCHEN HUIZINGA: Welcome to Abstracts, a Microsoft Research Podcast that puts the spotlight on world-class research in brief. I’m Gretchen Huizinga. In this series, members of the research community at Microsoft give us a quick snapshot – or a podcast abstract – of their new and noteworthy papers.  [MUSIC FADES] On today’s episode, I’m talking to Alex Lu, a senior researcher at Microsoft Research and co-author of a paper called Assessing the Limits of Zero Shot Foundation Models in Single-cell Biology. Alex Lu, wonderful to have you on the podcast. Welcome to Abstracts!  ALEX LU: Yeah, I’m really excited to be joining you today.  HUIZINGA: So let’s start with a little background of your work. In just a few sentences, tell us about your study and more importantly, why it matters.  LU: Absolutely. And before I dive in, I want to give a shout out to the MSR research intern who actually did this work. This was led by Kasia Kedzierska, who interned with us two summers ago in 2023, and she’s the lead author on the study. But basically, in this research, we study single-cell foundation models, which have really recently rocked the world of biology, because they basically claim to be able to use AI to unlock understanding about single-cell biology. Biologists for a myriad of applications, everything from understanding how single cells differentiate into different kinds of cells, to discovering new drugs for cancer, will conduct experiments where they measure how much of every gene is expressed inside of just one single cell. So these experiments give us a powerful view into the cell’s internal state. But measurements from these experiments are incredibly complex. There are about 20,000 different human genes. So you get this really long chain of numbers that measure how much there is of 20,000 different genes. So deriving meaning from this really long chain of numbers is really difficult. And single-cell foundation models claim to be capable of unraveling deeper insights than ever before. So that’s the claim that these works have made. And in our recent paper, we showed that these models may actually not live up to these claims. Basically, we showed that single-cell foundation models perform worse in settings that are fundamental to biological discovery than much simpler machine learning and statistical methods that were used in the field before single-cell foundation models emerged and are the go-to standard for unpacking meaning from these complicated experiments. So in a nutshell, we should care about these results because it has implications on the toolkits that biologists use to understand their experiments. Our work suggests that single-cell foundation models may not be appropriate for practical use just yet, at least in the discovery applications that we cover.  HUIZINGA: Well, let’s go a little deeper there. Generative pre-trained transformer models, GPTs, are relatively new on the research scene in terms of how they’re being used in novel applications, which is what you’re interested in, like single-cell biology. So I’m curious, just sort of as a foundation, what other research has already been done in this area, and how does this study illuminate or build on it?  LU: Absolutely. Okay, so we were the first to notice and document this issue in single-cell foundation models, specifically. And this is because that we have proposed evaluation methods that, while are common in other areas of AI, have yet to be commonly used to evaluate single-cell foundation models. We performed something called zero-shot evaluation on these models. Prior to our work, most works evaluated single-cell foundation models with fine tuning. And the way to understand this is because single-cell foundation models are trained in a way that tries to expose these models to millions of single-cells. But because you’re exposing them to a large amount of data, you can’t really rely upon this data being annotated or like labeled in any particular fashion then. So in order for them to actually do the specialized tasks that are useful for biologists, you typically have to add on a second training phase. We call this the fine-tuning phase, where you have a smaller number of single cells, but now they are actually labeled with the specialized tasks that you want the model to perform. So most people, they typically evaluate the performance of single-cell models after they fine-tune these models. However, what we noticed is that this evaluating these fine-tuned models has several problems. First, it might not actually align with how these models are actually going to be used by biologists then. A critical distinction in biology is that we’re not just trying to interact with an agent that has access to knowledge through its pre-training, we’re trying to extend these models to discover new biology beyond the sphere of influence then. And so in many cases, the point of using these models, the point of analysis, is to explore the data with the goal of potentially discovering something new about the single cell that the biologists worked with that they weren’t aware of before. So in these kinds of cases, it is really tough to fine-tune a model. There’s a bit of a chicken and egg problem going on. If you don’t know, for example, there’s a new kind of cell in the data, you can’t really instruct the model to help us identify these kinds of new cells. So in other words, fine-tuning these models for those tasks essentially becomes impossible then. So the second issue is that evaluations on fine-tuned models can sometimes mislead us in our ability to understand how these models are working. So for example, the claim behind single-cell foundation model papers is that these models learn a foundation of biological knowledge by being exposed to millions of single cells in its first training phase, right? But it’s possible when you fine-tune a model, it may just be that any performance increases that you see using the model is simply because that you’re using a massive model that is really sophisticated, really large. And even if there’s any exposure to any cells at all then, that model is going to do perfectly fine then. So going back to our paper, what’s really different about this paper is that we propose zero-shot evaluation for these models. What that means is that we do not fine-tune the model at all, and instead we keep the model frozen during the analysis step. So how we specialize it to be a downstream task instead is that we extract the model’s internal embedding of single-cell data, which is essentially a numerical vector that contains information that the model is extracting and organizing from input data. So it’s essentially how the model perceives single-cell data and how it’s organizing in its own internal state. So basically, this is the better way for us to test the claim that single-cell foundation models are learning foundational biological insights. Because if they actually are learning these insights, they should be present in the models embedding space even before we fine-tune the model.  HUIZINGA: Well, let’s talk about methodology on this particular study. You focused on assessing existing models in zero-shot learning for single-cell biology. How did you go about evaluating these models?  LU: Yes, so let’s dive deeper into how zero-shot evaluations are conducted, okay? So the premise here is that we’re relying upon the fact that if these models are fully learning foundational biological insights, if we take the model’s internal representation of cells, then cells that are biologically similar should be close in that internal representation, where cells that are biologically distinct should be further apart. And that is exactly what we tested in our study. We compared two popular single-cell foundation models and importantly, we compared these models against older and reliable tools that biologists have used for exploratory analyses. So these include simpler machine learning methods like scVI, statistical algorithms like Harmony, and even basic data pre-processing steps, just like filtering your data down to a more robust subset of genes, then. So basically, we tested embeddings from our two single-cell foundation models against this baseline in a variety of settings. And we tested the hypothesis that biologically similar cells should be similar across these distinct methods across these datasets.  HUIZINGA: Well, and as you as you did the testing, you obviously were aiming towards research findings, which is my favorite part of a research paper, so tell us what you did find and what you feel the most important takeaways of this paper are.  LU: Absolutely. So in a nutshell, we found that these two newly proposed single-cell foundation models substantially underperformed compared to older methods then. So to contextualize why that is such a surprising result, there is a lot of hype around these methods. So basically, I think that,yeah, it’s a very surprising result, given how hyped these models are and how people were already adopting them. But our results basically caution that these shouldn’t really be adopted for these use purposes.  HUIZINGA: Yeah, so this is serious real-world impact here in terms of if models are being adopted and adapted in these applications, how reliable are they, et cetera? So given that, who would you say benefits most from what you’ve discovered in this paper and why?  LU: Okay, so two ways, right? So I think this has at least immediate implications on the way that we do discovery in biology. And as I’ve discussed, these experiments are used for cases that have practical impact, drug discovery applications, investigations into basic biology, then. But let’s also talk about the impact for methodologists, people who are trying to improve these single-cell foundation models, right? I think at the base, they’re really excited proposals. Because if you look at what some of the prior and less sophisticated methods couldn’t do, they tended to be more bespoke. So the excitement of single-cell foundation models is that you have this general-purpose model that can be used for everything and while they’re not living up to that purpose just now, just currently, I think that it’s important that we continue to bank onto that vision, right? So if you look at our contributions in that area, where single-cell foundation models are a really new proposal, so it makes sense that we may not know how to fully evaluate them just yet then. So you can view our work as basically being a step towards more rigorous evaluation of these models. Now that we did this experiment, I think the methodologists know to use this as a signal on how to improve the models and if they’re going in the right direction. And in fact, you are seeing more and more papers adopt zero-shot evaluations since we put out our paper then. And so this essentially helps future computer scientists that are working on single-cell foundation models know how to train better models.  HUIZINGA: That said, Alex, finally, what are the outstanding challenges that you identified for zero-shot learning research in biology, and what foundation might this paper lay for future research agendas in the field?  LU: Yeah, absolutely. So now that we’ve shown single-cell foundation models don’t necessarily perform well, I think the natural question on everyone’s mind is how do we actually train single-cell foundation models that live up to that vision, that can perform in helping us discover new biology then? So I think in the short term, yeah, we’re actively investigating many hypotheses in this area. So for example, my colleagues, Lorin Crawford and Ava Amini, who were co-authors in the paper, recently put out a pre-print understanding how training data composition impacts model performance. And so one of the surprising findings that they had was that many of the training data sets that people used to train single-cell foundation models are highly redundant, to the point that you can even sample just a tiny fraction of the data and get basically the same performance then. But you can also look forward to many other explorations in this area as we continue to develop this research at the end of the day. But also zooming out into the bigger picture, I think one major takeaway from this paper is that developing AI methods for biology requires thought about the context of use, right? I mean, this is obvious for any AI method then, but I think people have gotten just too used to taking methods that work out there for natural vision or natural language maybe in the consumer domain and then extrapolating these methods to biology and expecting that they will work in the same way then, right? So for example, one reason why zero-shot evaluation was not routine practice for single-cell foundation models prior to our work, I mean, we were the first to fully establish that as a practice for the field, was because I think people who have been working in AI for biology have been looking to these more mainstream AI domains to shape their work then. And so with single-cell foundation models, many of these models are adopted from large language models with natural language processing, recycling the exact same architecture, the exact same code, basically just recycling practices in that field then. So when you look at like practices in like more mainstream domains, zero-shot evaluation is definitely explored in those domains, but it’s more of like a niche instead of being considered central to model understanding. So again, because biology is different from mainstream language processing, it’s a scientific discipline, zero-shot evaluation becomes much more important, and you have no choice but to use these models, zero-shot then. So in other words, I think that we need to be thinking carefully about what it is that makes training a model for biology different from training a model, for example, for consumer purposes.  [MUSIC] HUIZINGA: Alex Lu, thanks for joining us today, and to our listeners, thanks for tuning in. If you want to read this paper, you can find a link at aka.ms/Abstracts, or you can read it on the Genome Biology website. See you next time on Abstracts!  [MUSIC FADES] 
    0 Yorumlar 0 hisse senetleri 0 önizleme
  • Unity 2022.2 Tech Stream is now available

    I’m delighted to share that the 2022.2 Tech Stream, our final release of the year, is available for download.Tech Stream releases allow you to go hands-on with early access to the latest features. It’s also an opportunity to share your feedback on how we can build even better tools to power your creativity.Most recently at Unite, we gathered with our community of game developers to share some of these updates on topics like DOTS, rendering, multiplayer development, and XR, and we celebrated Made with Unity games like V Rising, Pentiment, Breachers, and many more. The dialogue online from over 9,000 Discord messages and countless in-person conversations was invaluable to shaping the future of Unity.Coupled with the 1,470 new forum threads where we discussed product feedback with you since the 2022.1 Tech Stream arrived and the 3,080 new notes on the Unity Platform roadmap, this feedback helped us get to today’s release. We couldn’t have done it without you and are excited to get that work in your hands. To learn more about how your feedback drives product development, check out this blog post.Together with the first Tech Stream, today’s 2022.2 completes this year’s cycle. Join us and explore what’s in store ahead of the LTS release in 2023. For even more on where Unity is heading, I encourage you to read our Games Focus blog series.In this post, I’ll be sharing a few highlights from this release, but you can always get more details in the official release notes.A frequent request we receive is to give you the ability to create more engaging gaming experiences, deeply immersive worlds, and to do so with more objects and characters than ever before.Unity 2022.2 includesECS for Unity, a data-oriented framework that empowers you to build more ambitious games with an unprecedented level of control and determinism. ECS and a data-oriented approach to development put complex gameplay mechanics and rich, dynamic environments at your fingertips. Starting with Unity 2022.2, ECS for Unity is fully supported for production, so you can get even more out of ECS through support channels and success plans.ECS for Unity includes the Entities package, along with ECS-compatible packages for Netcode, Graphics, and Physics. If you’re already familiar with Unity’s GameObject architecture and scripting standards, ECS for Unity is fully compatible with GameObjects, so you’ll find a familiar authoring experience and streamlined workflows. This gives you the capability to leverage your existing skill set and leverage ECS only where it will best benefit your game experience.We’re already seeing some great games running on ECS for Unity, such as Stunlock Studios’s V Rising. Because they turned to ECS, they were able to vastly increase the number of in-game interactable assets to more than 160,000 across a 5km2 map, with more than 350,000 server-side entities powering the experience.If you’re looking for help, want to provide feedback, discuss best practices, or show off your projects, you can join a thriving community on our forums and Discord. Our teams regularly engage in these channels and keep a close eye on your feedback. Join us on December 8, 2022 for our Dev Blitz Day dedicated to DOTS, when we’ll be spending an entire day trying to answer all your ECS questions.The last 18 months have seen an explosion of multiplayer experiences being built with Unity, and we hear that many of you want to add multiplayer access to your games but aren’t sure where to start.Alongside Unity 2022.2, we’re highlighting Netcode for GameObjects, a package that simplifies the implementation of multiplayer capability to your project in a number of scenarios such as couch cooperative play. The package works with familiar GameObject-based programming techniques, and it abstracts away low-level functionality so you can write less code while creating the multiplayer experience you envision.For more demanding, large-scale games, you can harness the power of ECS with Netcode for Entities. Netcode for Entities can enable you to increase your game world size, player counts, and complex network interactions without the performance sacrifices developers have traditionally had to deal with.We also recently announced the launch of self-serve capabilities in our Multiplayer Solutions suite within Unity Gaming Services, which helps you to operate your multiplayer games with hosting, communications, and more. Learn more about the latest developments for this tech in this Games Focus blog, or take a deeper look at the UGS Multiplayer suite in this UGS video, produced in collaboration with Tarodev.Multiplatform scalability and high-fidelity graphics continue to be our focus for rendering. In our Games Focus blog “Rendering that scales with your needs,” we covered our dedication to delivering features that allow you to scale with confidence while tapping into an even broader range of tools that provides the best possible visual quality and performance.We continue to bring the Universal Render Pipelinecloser to feature parity with Built-in Render Pipeline through more streamlined and scalable workflows. We worked on key tools such asForward+, which provides functional parity with Forward path in Built-in Render Pipeline, eliminating the light limit count so you can scale with quality across platforms.Another key feature is Decal Layers,which allow you to filter and configure how different objects are affected by Decal Projectors in a scene. Decals are useful for adding extra texture details to a scene, especially to break the repetitiveness of materials and their detail patterns.Other special URP enhancements include LOD crossfade for smoother transitions and Built-in Converter improvements that provide you with tools to upgrade your existing projects from the Built-in Render Pipeline to URP. You can also personalize your rendering experience with Shader Graph Full Screen Master Node and Custom Post Processing across both renderers.Diving into High Definition Render Pipeline, we’ve made enhancements that help you create even more beautiful physically based environments and detailed characters. You can scale high-fidelity environments with the new HDRP Water System to render oceans, rivers, and underwater effects, and use Volumetric Material to create procedural local fog using Shader Graph. Create even more realistic skies with improved Cloud Layers dynamic lighting, and you can even blend between different Volumetric Cloud conditions.You can also take your cinematic renders further to render realistic characters with Eye Cinematic with Caustics and PCSS shadows. HDRP Path Tracing Denoising provides you the choice between NVIDIA Optix™ AI accelerated denoiser and Intel® Open Image.Watch our latest Unite 2022 session on Lighting Environments in Unity to discover some key tips to get you started with our latest HDRP environment tools.Creative endeavors are never linear, and we understand that rapid iteration is part of the journey. This release includes new authoring features and workflow improvements to help speed up your productivity.For example, the Prefab system sees a number of upgrades, including the ability to quickly replace a Prefab Asset for a Prefab instance in a scene or nested inside other Prefabs. Read our latest blog on this topic for more information.For faster environments, the Paint Detail brush in the Terrain Tools package now allows you to simultaneously scatter multiple types of details with per-detail-type density settings available. Additionally, detail density and a few other Terrain settings are now overridable in the Quality settings to help you achieve platform performance targets.You can also use improved tooling and API features for Splines to help draw paths in your environments with greater precision. This means you can build out rivers, roads, camera tracks, and other path-related features and tools more efficiently. Thank you to all who engaged with us in the worldbuilding forums in the last couple months to help us finalize this delivery. For more on the API features, check out the documentation.Finally, the AI Navigation package is now available for you to quickly add intelligence to 3D characters and move in game worlds without needing to code rules manually. It also ships with samples to help you get started. See the forum for more details, and check out what’s next on the roadmap.In 2022.2, UI Toolkit is reaching parity with IMGUI for customizing the Editor and is the recommended solution for Editor tools. This means better separation of concerns, more flexible layouts, and advanced stylings. With updates like default inspectors generated with UI Toolkit, ported common built-in Property Drawers, TreeView controls with multicolumn support, and a new vector-drawing API, this release not only helps us reach parity with IMGUI but also supports runtime use cases as well.If you want to learn more about the current state of runtime, we recently released a new project demonstrating a full-feature interface with UI Toolkit based on your feedback for more samples. Check that out here.To help you get started, watch the recent Unite session illustrating a step-by-step example of how to build custom tools with UI Toolkit. Plus, visit the recently released Editor Design system for guidance on how to build intuitive experiences.After extensive work, testing, and listening to a lot of community feedback, DirectX 12 is out of an experimental state with the release of 2022.2. Depending on the project, you can now expect performance on par or greater than DX11, especially in draw call-heavy scenes.This is a result of significant investment into performance and stability, making DX12 the recommended graphics API for Windows and Xbox development. Additionally, DX12 lays the foundation for more advanced graphics features, such as real-time ray tracing, which is now available for Xbox game development. We couldn’t be more excited and thankful to you all for helping us get DX12 across the finish line and look forward to the great games you’ll be creating.We continue to hear that you not only want us to support new platforms, but also where we can simplify and improve development when targeting devices. If you haven’t already, check out the Games Focus blog “Reach more players over multiple platforms and form factors,” where we dive into both what is here now and what will be available in the near future.We’re making cross-device XR creation simpler with Unity XR Interaction toolkit. XRI provides a framework for common interactions that work across various controllers, such as grab, hover, select, visual feedback to indicate possible interactions on objects, and more. XRI is now in version 2.2, which adds multi-grab support, new locomotion methods, and a collection of ready-to-go Prefabs in our Starter Assets sample package.We recently invited the creators of Blacktop Hoops, a VR basketball game, to talk about how they used XRI as the base for their input controls during the Unite 2022 Keynote. Check out the XR segment for more information.We’ve also updated AR Foundation to version 5.0. This update brings two key features to reduce development time. The first is simulation, allowing you to test your AR app in the Editor using Play mode, an update that addresses a common AR developer frustration in the past. We’ve also added the AR Debug Menu as a new Prefab that you can use to view available configurations on your device and visualize AR subsystem data such as planes and point cloud positions.Finally, we’re continuing to add key platform support to the Editor with Meta Quest Pro, PlayStation®VR2 and Magic Leap 2.To read more about the 2022.2 Tech Stream, check out the release notes for a comprehensive list of features and the Unity Manual for documentation. As you dive in, keep in mind that while each Tech Stream release is supported with weekly updates until the next one, there is no guarantee for long-term support for new features and remember to always back up your work prior to upgrading to a new version. The upgrade guide can also assist with this. For projects in production, we recommend using Unity Long Term Release for stability and support.Each Tech Stream is an opportunity to not only get early access to new features, but also to shape the development of future tech through your feedback. We want to hear how we can best serve you and your projects. Let us know how we’re doing on the forums, or share feedback directly with our product team through the Unity Platform Roadmap. You can also follow us on Twitter and catch our latest Unity Twitch Roundtable, covering 2022.2, on demand.This release completes our 2022 development cycle. We have ambitious goals for next year, which you can read about in our Games Focus series or watch in the recent Unite Roadmap session. Thank you for all your support, and we look forward to partnering with you every step of the way.
    #unity #tech #stream #now #available
    Unity 2022.2 Tech Stream is now available
    I’m delighted to share that the 2022.2 Tech Stream, our final release of the year, is available for download.Tech Stream releases allow you to go hands-on with early access to the latest features. It’s also an opportunity to share your feedback on how we can build even better tools to power your creativity.Most recently at Unite, we gathered with our community of game developers to share some of these updates on topics like DOTS, rendering, multiplayer development, and XR, and we celebrated Made with Unity games like V Rising, Pentiment, Breachers, and many more. The dialogue online from over 9,000 Discord messages and countless in-person conversations was invaluable to shaping the future of Unity.Coupled with the 1,470 new forum threads where we discussed product feedback with you since the 2022.1 Tech Stream arrived and the 3,080 new notes on the Unity Platform roadmap, this feedback helped us get to today’s release. We couldn’t have done it without you and are excited to get that work in your hands. To learn more about how your feedback drives product development, check out this blog post.Together with the first Tech Stream, today’s 2022.2 completes this year’s cycle. Join us and explore what’s in store ahead of the LTS release in 2023. For even more on where Unity is heading, I encourage you to read our Games Focus blog series.In this post, I’ll be sharing a few highlights from this release, but you can always get more details in the official release notes.A frequent request we receive is to give you the ability to create more engaging gaming experiences, deeply immersive worlds, and to do so with more objects and characters than ever before.Unity 2022.2 includesECS for Unity, a data-oriented framework that empowers you to build more ambitious games with an unprecedented level of control and determinism. ECS and a data-oriented approach to development put complex gameplay mechanics and rich, dynamic environments at your fingertips. Starting with Unity 2022.2, ECS for Unity is fully supported for production, so you can get even more out of ECS through support channels and success plans.ECS for Unity includes the Entities package, along with ECS-compatible packages for Netcode, Graphics, and Physics. If you’re already familiar with Unity’s GameObject architecture and scripting standards, ECS for Unity is fully compatible with GameObjects, so you’ll find a familiar authoring experience and streamlined workflows. This gives you the capability to leverage your existing skill set and leverage ECS only where it will best benefit your game experience.We’re already seeing some great games running on ECS for Unity, such as Stunlock Studios’s V Rising. Because they turned to ECS, they were able to vastly increase the number of in-game interactable assets to more than 160,000 across a 5km2 map, with more than 350,000 server-side entities powering the experience.If you’re looking for help, want to provide feedback, discuss best practices, or show off your projects, you can join a thriving community on our forums and Discord. Our teams regularly engage in these channels and keep a close eye on your feedback. Join us on December 8, 2022 for our Dev Blitz Day dedicated to DOTS, when we’ll be spending an entire day trying to answer all your ECS questions.The last 18 months have seen an explosion of multiplayer experiences being built with Unity, and we hear that many of you want to add multiplayer access to your games but aren’t sure where to start.Alongside Unity 2022.2, we’re highlighting Netcode for GameObjects, a package that simplifies the implementation of multiplayer capability to your project in a number of scenarios such as couch cooperative play. The package works with familiar GameObject-based programming techniques, and it abstracts away low-level functionality so you can write less code while creating the multiplayer experience you envision.For more demanding, large-scale games, you can harness the power of ECS with Netcode for Entities. Netcode for Entities can enable you to increase your game world size, player counts, and complex network interactions without the performance sacrifices developers have traditionally had to deal with.We also recently announced the launch of self-serve capabilities in our Multiplayer Solutions suite within Unity Gaming Services, which helps you to operate your multiplayer games with hosting, communications, and more. Learn more about the latest developments for this tech in this Games Focus blog, or take a deeper look at the UGS Multiplayer suite in this UGS video, produced in collaboration with Tarodev.Multiplatform scalability and high-fidelity graphics continue to be our focus for rendering. In our Games Focus blog “Rendering that scales with your needs,” we covered our dedication to delivering features that allow you to scale with confidence while tapping into an even broader range of tools that provides the best possible visual quality and performance.We continue to bring the Universal Render Pipelinecloser to feature parity with Built-in Render Pipeline through more streamlined and scalable workflows. We worked on key tools such asForward+, which provides functional parity with Forward path in Built-in Render Pipeline, eliminating the light limit count so you can scale with quality across platforms.Another key feature is Decal Layers,which allow you to filter and configure how different objects are affected by Decal Projectors in a scene. Decals are useful for adding extra texture details to a scene, especially to break the repetitiveness of materials and their detail patterns.Other special URP enhancements include LOD crossfade for smoother transitions and Built-in Converter improvements that provide you with tools to upgrade your existing projects from the Built-in Render Pipeline to URP. You can also personalize your rendering experience with Shader Graph Full Screen Master Node and Custom Post Processing across both renderers.Diving into High Definition Render Pipeline, we’ve made enhancements that help you create even more beautiful physically based environments and detailed characters. You can scale high-fidelity environments with the new HDRP Water System to render oceans, rivers, and underwater effects, and use Volumetric Material to create procedural local fog using Shader Graph. Create even more realistic skies with improved Cloud Layers dynamic lighting, and you can even blend between different Volumetric Cloud conditions.You can also take your cinematic renders further to render realistic characters with Eye Cinematic with Caustics and PCSS shadows. HDRP Path Tracing Denoising provides you the choice between NVIDIA Optix™ AI accelerated denoiser and Intel® Open Image.Watch our latest Unite 2022 session on Lighting Environments in Unity to discover some key tips to get you started with our latest HDRP environment tools.Creative endeavors are never linear, and we understand that rapid iteration is part of the journey. This release includes new authoring features and workflow improvements to help speed up your productivity.For example, the Prefab system sees a number of upgrades, including the ability to quickly replace a Prefab Asset for a Prefab instance in a scene or nested inside other Prefabs. Read our latest blog on this topic for more information.For faster environments, the Paint Detail brush in the Terrain Tools package now allows you to simultaneously scatter multiple types of details with per-detail-type density settings available. Additionally, detail density and a few other Terrain settings are now overridable in the Quality settings to help you achieve platform performance targets.You can also use improved tooling and API features for Splines to help draw paths in your environments with greater precision. This means you can build out rivers, roads, camera tracks, and other path-related features and tools more efficiently. Thank you to all who engaged with us in the worldbuilding forums in the last couple months to help us finalize this delivery. For more on the API features, check out the documentation.Finally, the AI Navigation package is now available for you to quickly add intelligence to 3D characters and move in game worlds without needing to code rules manually. It also ships with samples to help you get started. See the forum for more details, and check out what’s next on the roadmap.In 2022.2, UI Toolkit is reaching parity with IMGUI for customizing the Editor and is the recommended solution for Editor tools. This means better separation of concerns, more flexible layouts, and advanced stylings. With updates like default inspectors generated with UI Toolkit, ported common built-in Property Drawers, TreeView controls with multicolumn support, and a new vector-drawing API, this release not only helps us reach parity with IMGUI but also supports runtime use cases as well.If you want to learn more about the current state of runtime, we recently released a new project demonstrating a full-feature interface with UI Toolkit based on your feedback for more samples. Check that out here.To help you get started, watch the recent Unite session illustrating a step-by-step example of how to build custom tools with UI Toolkit. Plus, visit the recently released Editor Design system for guidance on how to build intuitive experiences.After extensive work, testing, and listening to a lot of community feedback, DirectX 12 is out of an experimental state with the release of 2022.2. Depending on the project, you can now expect performance on par or greater than DX11, especially in draw call-heavy scenes.This is a result of significant investment into performance and stability, making DX12 the recommended graphics API for Windows and Xbox development. Additionally, DX12 lays the foundation for more advanced graphics features, such as real-time ray tracing, which is now available for Xbox game development. We couldn’t be more excited and thankful to you all for helping us get DX12 across the finish line and look forward to the great games you’ll be creating.We continue to hear that you not only want us to support new platforms, but also where we can simplify and improve development when targeting devices. If you haven’t already, check out the Games Focus blog “Reach more players over multiple platforms and form factors,” where we dive into both what is here now and what will be available in the near future.We’re making cross-device XR creation simpler with Unity XR Interaction toolkit. XRI provides a framework for common interactions that work across various controllers, such as grab, hover, select, visual feedback to indicate possible interactions on objects, and more. XRI is now in version 2.2, which adds multi-grab support, new locomotion methods, and a collection of ready-to-go Prefabs in our Starter Assets sample package.We recently invited the creators of Blacktop Hoops, a VR basketball game, to talk about how they used XRI as the base for their input controls during the Unite 2022 Keynote. Check out the XR segment for more information.We’ve also updated AR Foundation to version 5.0. This update brings two key features to reduce development time. The first is simulation, allowing you to test your AR app in the Editor using Play mode, an update that addresses a common AR developer frustration in the past. We’ve also added the AR Debug Menu as a new Prefab that you can use to view available configurations on your device and visualize AR subsystem data such as planes and point cloud positions.Finally, we’re continuing to add key platform support to the Editor with Meta Quest Pro, PlayStation®VR2 and Magic Leap 2.To read more about the 2022.2 Tech Stream, check out the release notes for a comprehensive list of features and the Unity Manual for documentation. As you dive in, keep in mind that while each Tech Stream release is supported with weekly updates until the next one, there is no guarantee for long-term support for new features and remember to always back up your work prior to upgrading to a new version. The upgrade guide can also assist with this. For projects in production, we recommend using Unity Long Term Release for stability and support.Each Tech Stream is an opportunity to not only get early access to new features, but also to shape the development of future tech through your feedback. We want to hear how we can best serve you and your projects. Let us know how we’re doing on the forums, or share feedback directly with our product team through the Unity Platform Roadmap. You can also follow us on Twitter and catch our latest Unity Twitch Roundtable, covering 2022.2, on demand.This release completes our 2022 development cycle. We have ambitious goals for next year, which you can read about in our Games Focus series or watch in the recent Unite Roadmap session. Thank you for all your support, and we look forward to partnering with you every step of the way. #unity #tech #stream #now #available
    UNITY.COM
    Unity 2022.2 Tech Stream is now available
    I’m delighted to share that the 2022.2 Tech Stream, our final release of the year, is available for download.Tech Stream releases allow you to go hands-on with early access to the latest features. It’s also an opportunity to share your feedback on how we can build even better tools to power your creativity.Most recently at Unite, we gathered with our community of game developers to share some of these updates on topics like DOTS, rendering, multiplayer development, and XR, and we celebrated Made with Unity games like V Rising, Pentiment, Breachers, and many more. The dialogue online from over 9,000 Discord messages and countless in-person conversations was invaluable to shaping the future of Unity.Coupled with the 1,470 new forum threads where we discussed product feedback with you since the 2022.1 Tech Stream arrived and the 3,080 new notes on the Unity Platform roadmap, this feedback helped us get to today’s release. We couldn’t have done it without you and are excited to get that work in your hands. To learn more about how your feedback drives product development, check out this blog post.Together with the first Tech Stream, today’s 2022.2 completes this year’s cycle. Join us and explore what’s in store ahead of the LTS release in 2023. For even more on where Unity is heading, I encourage you to read our Games Focus blog series.In this post, I’ll be sharing a few highlights from this release, but you can always get more details in the official release notes.A frequent request we receive is to give you the ability to create more engaging gaming experiences, deeply immersive worlds, and to do so with more objects and characters than ever before.Unity 2022.2 includesECS for Unity (Entity Component System), a data-oriented framework that empowers you to build more ambitious games with an unprecedented level of control and determinism. ECS and a data-oriented approach to development put complex gameplay mechanics and rich, dynamic environments at your fingertips. Starting with Unity 2022.2, ECS for Unity is fully supported for production, so you can get even more out of ECS through support channels and success plans.ECS for Unity includes the Entities package, along with ECS-compatible packages for Netcode, Graphics, and Physics. If you’re already familiar with Unity’s GameObject architecture and scripting standards, ECS for Unity is fully compatible with GameObjects, so you’ll find a familiar authoring experience and streamlined workflows. This gives you the capability to leverage your existing skill set and leverage ECS only where it will best benefit your game experience.We’re already seeing some great games running on ECS for Unity, such as Stunlock Studios’s V Rising. Because they turned to ECS, they were able to vastly increase the number of in-game interactable assets to more than 160,000 across a 5km2 map, with more than 350,000 server-side entities powering the experience.If you’re looking for help, want to provide feedback, discuss best practices, or show off your projects, you can join a thriving community on our forums and Discord. Our teams regularly engage in these channels and keep a close eye on your feedback. Join us on December 8, 2022 for our Dev Blitz Day dedicated to DOTS, when we’ll be spending an entire day trying to answer all your ECS questions.The last 18 months have seen an explosion of multiplayer experiences being built with Unity, and we hear that many of you want to add multiplayer access to your games but aren’t sure where to start.Alongside Unity 2022.2, we’re highlighting Netcode for GameObjects, a package that simplifies the implementation of multiplayer capability to your project in a number of scenarios such as couch cooperative play. The package works with familiar GameObject-based programming techniques, and it abstracts away low-level functionality so you can write less code while creating the multiplayer experience you envision.For more demanding, large-scale games, you can harness the power of ECS with Netcode for Entities. Netcode for Entities can enable you to increase your game world size, player counts, and complex network interactions without the performance sacrifices developers have traditionally had to deal with.We also recently announced the launch of self-serve capabilities in our Multiplayer Solutions suite within Unity Gaming Services (UGS), which helps you to operate your multiplayer games with hosting, communications, and more. Learn more about the latest developments for this tech in this Games Focus blog, or take a deeper look at the UGS Multiplayer suite in this UGS video, produced in collaboration with Tarodev.Multiplatform scalability and high-fidelity graphics continue to be our focus for rendering. In our Games Focus blog “Rendering that scales with your needs,” we covered our dedication to delivering features that allow you to scale with confidence while tapping into an even broader range of tools that provides the best possible visual quality and performance.We continue to bring the Universal Render Pipeline (URP) closer to feature parity with Built-in Render Pipeline through more streamlined and scalable workflows. We worked on key tools such asForward+, which provides functional parity with Forward path in Built-in Render Pipeline, eliminating the light limit count so you can scale with quality across platforms.Another key feature is Decal Layers,which allow you to filter and configure how different objects are affected by Decal Projectors in a scene. Decals are useful for adding extra texture details to a scene, especially to break the repetitiveness of materials and their detail patterns.Other special URP enhancements include LOD crossfade for smoother transitions and Built-in Converter improvements that provide you with tools to upgrade your existing projects from the Built-in Render Pipeline to URP. You can also personalize your rendering experience with Shader Graph Full Screen Master Node and Custom Post Processing across both renderers.Diving into High Definition Render Pipeline (HDRP), we’ve made enhancements that help you create even more beautiful physically based environments and detailed characters. You can scale high-fidelity environments with the new HDRP Water System to render oceans, rivers, and underwater effects, and use Volumetric Material to create procedural local fog using Shader Graph. Create even more realistic skies with improved Cloud Layers dynamic lighting, and you can even blend between different Volumetric Cloud conditions.You can also take your cinematic renders further to render realistic characters with Eye Cinematic with Caustics and PCSS shadows. HDRP Path Tracing Denoising provides you the choice between NVIDIA Optix™ AI accelerated denoiser and Intel® Open Image.Watch our latest Unite 2022 session on Lighting Environments in Unity to discover some key tips to get you started with our latest HDRP environment tools.Creative endeavors are never linear, and we understand that rapid iteration is part of the journey. This release includes new authoring features and workflow improvements to help speed up your productivity.For example, the Prefab system sees a number of upgrades, including the ability to quickly replace a Prefab Asset for a Prefab instance in a scene or nested inside other Prefabs. Read our latest blog on this topic for more information.For faster environments, the Paint Detail brush in the Terrain Tools package now allows you to simultaneously scatter multiple types of details with per-detail-type density settings available. Additionally, detail density and a few other Terrain settings are now overridable in the Quality settings to help you achieve platform performance targets.You can also use improved tooling and API features for Splines to help draw paths in your environments with greater precision. This means you can build out rivers, roads, camera tracks, and other path-related features and tools more efficiently. Thank you to all who engaged with us in the worldbuilding forums in the last couple months to help us finalize this delivery. For more on the API features, check out the documentation.Finally, the AI Navigation package is now available for you to quickly add intelligence to 3D characters and move in game worlds without needing to code rules manually. It also ships with samples to help you get started. See the forum for more details, and check out what’s next on the roadmap.In 2022.2, UI Toolkit is reaching parity with IMGUI for customizing the Editor and is the recommended solution for Editor tools. This means better separation of concerns, more flexible layouts, and advanced stylings. With updates like default inspectors generated with UI Toolkit, ported common built-in Property Drawers, TreeView controls with multicolumn support, and a new vector-drawing API, this release not only helps us reach parity with IMGUI but also supports runtime use cases as well.If you want to learn more about the current state of runtime, we recently released a new project demonstrating a full-feature interface with UI Toolkit based on your feedback for more samples. Check that out here.To help you get started, watch the recent Unite session illustrating a step-by-step example of how to build custom tools with UI Toolkit. Plus, visit the recently released Editor Design system for guidance on how to build intuitive experiences.After extensive work, testing, and listening to a lot of community feedback, DirectX 12 is out of an experimental state with the release of 2022.2. Depending on the project, you can now expect performance on par or greater than DX11, especially in draw call-heavy scenes.This is a result of significant investment into performance and stability, making DX12 the recommended graphics API for Windows and Xbox development. Additionally, DX12 lays the foundation for more advanced graphics features, such as real-time ray tracing, which is now available for Xbox game development. We couldn’t be more excited and thankful to you all for helping us get DX12 across the finish line and look forward to the great games you’ll be creating.We continue to hear that you not only want us to support new platforms, but also where we can simplify and improve development when targeting devices. If you haven’t already, check out the Games Focus blog “Reach more players over multiple platforms and form factors,” where we dive into both what is here now and what will be available in the near future.We’re making cross-device XR creation simpler with Unity XR Interaction toolkit (XRI). XRI provides a framework for common interactions that work across various controllers, such as grab, hover, select, visual feedback to indicate possible interactions on objects, and more. XRI is now in version 2.2, which adds multi-grab support, new locomotion methods, and a collection of ready-to-go Prefabs in our Starter Assets sample package.We recently invited the creators of Blacktop Hoops, a VR basketball game, to talk about how they used XRI as the base for their input controls during the Unite 2022 Keynote. Check out the XR segment for more information.We’ve also updated AR Foundation to version 5.0. This update brings two key features to reduce development time. The first is simulation, allowing you to test your AR app in the Editor using Play mode, an update that addresses a common AR developer frustration in the past. We’ve also added the AR Debug Menu as a new Prefab that you can use to view available configurations on your device and visualize AR subsystem data such as planes and point cloud positions.Finally, we’re continuing to add key platform support to the Editor with Meta Quest Pro, PlayStation®VR2 and Magic Leap 2.To read more about the 2022.2 Tech Stream, check out the release notes for a comprehensive list of features and the Unity Manual for documentation. As you dive in, keep in mind that while each Tech Stream release is supported with weekly updates until the next one, there is no guarantee for long-term support for new features and remember to always back up your work prior to upgrading to a new version. The upgrade guide can also assist with this. For projects in production, we recommend using Unity Long Term Release for stability and support.Each Tech Stream is an opportunity to not only get early access to new features, but also to shape the development of future tech through your feedback. We want to hear how we can best serve you and your projects. Let us know how we’re doing on the forums, or share feedback directly with our product team through the Unity Platform Roadmap. You can also follow us on Twitter and catch our latest Unity Twitch Roundtable, covering 2022.2, on demand.This release completes our 2022 development cycle. We have ambitious goals for next year, which you can read about in our Games Focus series or watch in the recent Unite Roadmap session. Thank you for all your support, and we look forward to partnering with you every step of the way.
    0 Yorumlar 0 hisse senetleri 0 önizleme
  • Abstracts: Aurora with Megan Stanley and Wessel Bruinsma

    This is such exciting work about environmental forecasting, so we’re happy to have the two of you join us today.  
    Megan and Wessel, welcome. 
    MEGAN STANLEY: Thank you. Thanks. Great to be here. 
    WESSEL BRUINSMA: Thanks. 
    TINGLE: Let’s jump right in. Wessel, share a bit about the problem your research addresses and why this work is so important. 
    BRUINSMA: I think we’re all very much aware of the revolution that’s happening in the space of large language models, which have just become so strong. What’s perhaps lesser well-known is that machine learning models have also started to revolutionize this field of weather prediction. Whereas traditional weather prediction models, based on physical laws, used to be the state of the art, these traditional models are now challenged and often even outperformed by AI models.  
    This advancement is super impressive and really a big deal. Mostly because AI weather forecasting models are computationally much more efficient and can even be more accurate. What’s unfortunate though, about this big step forward, is that these developments are mostly limited to the setting of weather forecasting.  
    Weather forecasting is very important, obviously, but there are many other important environmental forecasting problems out there, such as air pollution forecasting or ocean wave forecasting. We have developed a model, named Aurora, which really kicks the AI revolution in weather forecasting into the next gear by extending these advancements to other environmental forecasting fields, too. With Aurora, we’re now able to produce state-of-the-art air pollution forecasts using an AI approach. And that wasn’t possible before! 
    TINGLE: Megan, how does this approach differ from or build on work that’s already been done in the atmospheric sciences? 
    STANLEY: Current approaches have really focused training very specifically on weather forecasting models. And in contrast, with Aurora, what we’ve attempted to do is train a so-called foundation model for the Earth system. In the first step, we train Aurora on a vast body of Earth system data. This is our pretraining step.  
    And when I say a vast body of data, I really do mean a lot. And the purpose of this pretraining is to let Aurora, kind of, learn some general-purpose representation of the dynamics that govern the Earth system. But then once we’ve pretrained Aurora, and this really is the crux of this, the reason why we’re doing this project, is after the model has been pretrained, it can leverage this learned general-purpose representation and efficiently adapt to new tasks, new domains, new variables. And this is called fine-tuning. 
    The idea is that the model really uses the learned representation to perform this adaptation very efficiently, which basically means Aurora is a powerful, flexible model that can relatively cheaply be adapted to any environmental forecasting task.   
    TINGLE: Wessel, can you tell us about your methodology? How did you all conduct this research? 
    BRUINSMA: While approaches so far have trained models on primarily one particular data
    These datasets are a combination of estimates of the historical state of the world, forecasts by other models, climate simulations, and more. We’ve been able to show that training on not just more data but more diverse data helps the model achieve even better performance. Showing this is difficult because there is just so much data.  
    In addition to scaling to more and more diverse data, we also increased the size of the model as much as we could. Here we found that bigger models, despite being slower to run, make more efficient use of computational resources. It’s cheaper to train a good big model than a good small model. The mantra of this project was to really keep it simple and to scale to simultaneously very large and, more importantly, diverse data and large model size. 
    TINGLE: So, Megan, what were your major findings? And we know they’re major because they’re in Nature. 
    STANLEY: Yeah,I guess they really are. So the main outcome of this project is we were actually able to train a single foundation model that achieves state-of-the-art performance in four different domains. Air pollution forecasting. For example, predicting particulate matter near the surface or ozone in the atmosphere. Ocean wave forecasting, which is critical for planning shipping routes.  
    Tropical cyclone track forecasting, so that means being able to predict where a hurricane or a typhoon is expected to go, which is obviously incredibly important, and very high-resolution weather forecasting.  
    And I’ve, kind of, named these forecasting domains as if they’re just items in a list, but in every single one, Aurora really pushed the limits of what is possible with AI models. And we’re really proud of that.  
    But perhaps, kind of, you know, to my mind, the key takeaway here is that the foundation model approach actually works. So what we have shown is it’s possible to actually train some kind of general model, a foundation model, and then adapt it to a wide variety of environmental tasks. Now we definitely do not claim that Aurora is some kind of ultimate environmental forecasting model. We are sure that the model and the pretraining procedure can actually be improved. But, nevertheless, we’ve shown that this approach works for environmental forecasting. It really holds massive promise, and that’s incredibly cool. 
    TINGLE: Wessel, what do you think will be the real-world impact of this work? 
    BRUINSMA: Well, for applications that we mentioned, which are air pollution forecasting, ocean wave forecasting, tropical cyclone track forecasting, and very high-resolution weather forecasting, Aurora could today be deployed in real-time systems to produce near real-time forecasts. And, you know, in fact, it already is. You can view real-time weather forecasts by the high-resolution version of the model on the website of ECMWF. 
    But what’s remarkable is that every of these applications took a small team of engineers about four to eight weeks to fully execute. You should compare this to a typical development timeline for more traditional models, which can be on the order of multiple years. Using the pretraining fine-tuning approach that we used for Aurora, we might see significantly accelerated development cycles for environmental forecasting problems. And that’s exciting. 
    TINGLE: Megan, if our listeners only walk away from this conversation with one key talking point, what would you like that to be? What should we remember about this paper? 
    STANLEY: The biggest takeaway is that the pretraining fine-tuning paradigm, it really works for environmental forecasting, right? So you can train a foundational model, it learns some kind of general-purpose representation of the Earth system dynamics, and this representation boosts performance in a wide variety of forecasting tasks. But we really want to emphasize that Aurora only scratches the surface of what’s actually possible. 
    So there are many more applications to explore than the four we’ve mentioned. And undoubtedly, the model and pretraining procedure can actually be improved. So we’re really excited to see what the next few years will bring. 
    TINGLE: Wessel, tell us more about those opportunities and unanswered questions. What’s next on the research agenda in environmental prediction? 
    BRUINSMA: Well, Aurora has two main limitations. The first is that the model produces only deterministic predictions, by which I mean a single predicted value. For variables like temperature, this is mostly fine. But other variables like precipitation, they are inherently some kind of stochastic. For these variables, we really want to assign probabilities to different levels of precipitation rather than predicting only a single value. 
    An extension of Aurora to allow this sort of prediction would be a great next step.  
    The second limitation is that Aurora depends on a procedure called assimilation. Assimilation attempts to create a starting point for the model from real-world observations, such as from weather stations and satellites. The model then takes the starting point and uses it to make predictions. Unfortunately, assimilation is super expensive, so it would be great if we could somehow circumvent the need for it. 
    Finally, what we find really important is to make our advancements available to the community. 
    TINGLE: Great. Megan and Wessel, thanks for joining us today on the Microsoft Research Podcast. 
    BRUINSMA: Thanks for having us. 
    STANLEY: Yeah, thank you. It’s been great. 
    TINGLE: You can check out the Aurora model on Azure AI Foundry. You can read the entire paper, “A Foundation Model for the Earth System,” at aka.ms/abstracts. And you’ll certainly find it on the Nature website, too.  
    Thank you so much for tuning in to Abstracts today. Until next time.   
    #abstracts #aurora #with #megan #stanley
    Abstracts: Aurora with Megan Stanley and Wessel Bruinsma
    This is such exciting work about environmental forecasting, so we’re happy to have the two of you join us today.   Megan and Wessel, welcome.  MEGAN STANLEY: Thank you. Thanks. Great to be here.  WESSEL BRUINSMA: Thanks.  TINGLE: Let’s jump right in. Wessel, share a bit about the problem your research addresses and why this work is so important.  BRUINSMA: I think we’re all very much aware of the revolution that’s happening in the space of large language models, which have just become so strong. What’s perhaps lesser well-known is that machine learning models have also started to revolutionize this field of weather prediction. Whereas traditional weather prediction models, based on physical laws, used to be the state of the art, these traditional models are now challenged and often even outperformed by AI models.   This advancement is super impressive and really a big deal. Mostly because AI weather forecasting models are computationally much more efficient and can even be more accurate. What’s unfortunate though, about this big step forward, is that these developments are mostly limited to the setting of weather forecasting.   Weather forecasting is very important, obviously, but there are many other important environmental forecasting problems out there, such as air pollution forecasting or ocean wave forecasting. We have developed a model, named Aurora, which really kicks the AI revolution in weather forecasting into the next gear by extending these advancements to other environmental forecasting fields, too. With Aurora, we’re now able to produce state-of-the-art air pollution forecasts using an AI approach. And that wasn’t possible before!  TINGLE: Megan, how does this approach differ from or build on work that’s already been done in the atmospheric sciences?  STANLEY: Current approaches have really focused training very specifically on weather forecasting models. And in contrast, with Aurora, what we’ve attempted to do is train a so-called foundation model for the Earth system. In the first step, we train Aurora on a vast body of Earth system data. This is our pretraining step.   And when I say a vast body of data, I really do mean a lot. And the purpose of this pretraining is to let Aurora, kind of, learn some general-purpose representation of the dynamics that govern the Earth system. But then once we’ve pretrained Aurora, and this really is the crux of this, the reason why we’re doing this project, is after the model has been pretrained, it can leverage this learned general-purpose representation and efficiently adapt to new tasks, new domains, new variables. And this is called fine-tuning.  The idea is that the model really uses the learned representation to perform this adaptation very efficiently, which basically means Aurora is a powerful, flexible model that can relatively cheaply be adapted to any environmental forecasting task.    TINGLE: Wessel, can you tell us about your methodology? How did you all conduct this research?  BRUINSMA: While approaches so far have trained models on primarily one particular data These datasets are a combination of estimates of the historical state of the world, forecasts by other models, climate simulations, and more. We’ve been able to show that training on not just more data but more diverse data helps the model achieve even better performance. Showing this is difficult because there is just so much data.   In addition to scaling to more and more diverse data, we also increased the size of the model as much as we could. Here we found that bigger models, despite being slower to run, make more efficient use of computational resources. It’s cheaper to train a good big model than a good small model. The mantra of this project was to really keep it simple and to scale to simultaneously very large and, more importantly, diverse data and large model size.  TINGLE: So, Megan, what were your major findings? And we know they’re major because they’re in Nature.  STANLEY: Yeah,I guess they really are. So the main outcome of this project is we were actually able to train a single foundation model that achieves state-of-the-art performance in four different domains. Air pollution forecasting. For example, predicting particulate matter near the surface or ozone in the atmosphere. Ocean wave forecasting, which is critical for planning shipping routes.   Tropical cyclone track forecasting, so that means being able to predict where a hurricane or a typhoon is expected to go, which is obviously incredibly important, and very high-resolution weather forecasting.   And I’ve, kind of, named these forecasting domains as if they’re just items in a list, but in every single one, Aurora really pushed the limits of what is possible with AI models. And we’re really proud of that.   But perhaps, kind of, you know, to my mind, the key takeaway here is that the foundation model approach actually works. So what we have shown is it’s possible to actually train some kind of general model, a foundation model, and then adapt it to a wide variety of environmental tasks. Now we definitely do not claim that Aurora is some kind of ultimate environmental forecasting model. We are sure that the model and the pretraining procedure can actually be improved. But, nevertheless, we’ve shown that this approach works for environmental forecasting. It really holds massive promise, and that’s incredibly cool.  TINGLE: Wessel, what do you think will be the real-world impact of this work?  BRUINSMA: Well, for applications that we mentioned, which are air pollution forecasting, ocean wave forecasting, tropical cyclone track forecasting, and very high-resolution weather forecasting, Aurora could today be deployed in real-time systems to produce near real-time forecasts. And, you know, in fact, it already is. You can view real-time weather forecasts by the high-resolution version of the model on the website of ECMWF.  But what’s remarkable is that every of these applications took a small team of engineers about four to eight weeks to fully execute. You should compare this to a typical development timeline for more traditional models, which can be on the order of multiple years. Using the pretraining fine-tuning approach that we used for Aurora, we might see significantly accelerated development cycles for environmental forecasting problems. And that’s exciting.  TINGLE: Megan, if our listeners only walk away from this conversation with one key talking point, what would you like that to be? What should we remember about this paper?  STANLEY: The biggest takeaway is that the pretraining fine-tuning paradigm, it really works for environmental forecasting, right? So you can train a foundational model, it learns some kind of general-purpose representation of the Earth system dynamics, and this representation boosts performance in a wide variety of forecasting tasks. But we really want to emphasize that Aurora only scratches the surface of what’s actually possible.  So there are many more applications to explore than the four we’ve mentioned. And undoubtedly, the model and pretraining procedure can actually be improved. So we’re really excited to see what the next few years will bring.  TINGLE: Wessel, tell us more about those opportunities and unanswered questions. What’s next on the research agenda in environmental prediction?  BRUINSMA: Well, Aurora has two main limitations. The first is that the model produces only deterministic predictions, by which I mean a single predicted value. For variables like temperature, this is mostly fine. But other variables like precipitation, they are inherently some kind of stochastic. For these variables, we really want to assign probabilities to different levels of precipitation rather than predicting only a single value.  An extension of Aurora to allow this sort of prediction would be a great next step.   The second limitation is that Aurora depends on a procedure called assimilation. Assimilation attempts to create a starting point for the model from real-world observations, such as from weather stations and satellites. The model then takes the starting point and uses it to make predictions. Unfortunately, assimilation is super expensive, so it would be great if we could somehow circumvent the need for it.  Finally, what we find really important is to make our advancements available to the community.  TINGLE: Great. Megan and Wessel, thanks for joining us today on the Microsoft Research Podcast.  BRUINSMA: Thanks for having us.  STANLEY: Yeah, thank you. It’s been great.  TINGLE: You can check out the Aurora model on Azure AI Foundry. You can read the entire paper, “A Foundation Model for the Earth System,” at aka.ms/abstracts. And you’ll certainly find it on the Nature website, too.   Thank you so much for tuning in to Abstracts today. Until next time.    #abstracts #aurora #with #megan #stanley
    WWW.MICROSOFT.COM
    Abstracts: Aurora with Megan Stanley and Wessel Bruinsma
    This is such exciting work about environmental forecasting, so we’re happy to have the two of you join us today.   Megan and Wessel, welcome.  MEGAN STANLEY: Thank you. Thanks. Great to be here.  WESSEL BRUINSMA: Thanks.  TINGLE: Let’s jump right in. Wessel, share a bit about the problem your research addresses and why this work is so important.  BRUINSMA: I think we’re all very much aware of the revolution that’s happening in the space of large language models, which have just become so strong. What’s perhaps lesser well-known is that machine learning models have also started to revolutionize this field of weather prediction. Whereas traditional weather prediction models, based on physical laws, used to be the state of the art, these traditional models are now challenged and often even outperformed by AI models.   This advancement is super impressive and really a big deal. Mostly because AI weather forecasting models are computationally much more efficient and can even be more accurate. What’s unfortunate though, about this big step forward, is that these developments are mostly limited to the setting of weather forecasting.   Weather forecasting is very important, obviously, but there are many other important environmental forecasting problems out there, such as air pollution forecasting or ocean wave forecasting. We have developed a model, named Aurora, which really kicks the AI revolution in weather forecasting into the next gear by extending these advancements to other environmental forecasting fields, too. With Aurora, we’re now able to produce state-of-the-art air pollution forecasts using an AI approach. And that wasn’t possible before!  TINGLE: Megan, how does this approach differ from or build on work that’s already been done in the atmospheric sciences?  STANLEY: Current approaches have really focused training very specifically on weather forecasting models. And in contrast, with Aurora, what we’ve attempted to do is train a so-called foundation model for the Earth system. In the first step, we train Aurora on a vast body of Earth system data. This is our pretraining step.   And when I say a vast body of data, I really do mean a lot. And the purpose of this pretraining is to let Aurora, kind of, learn some general-purpose representation of the dynamics that govern the Earth system. But then once we’ve pretrained Aurora, and this really is the crux of this, the reason why we’re doing this project, is after the model has been pretrained, it can leverage this learned general-purpose representation and efficiently adapt to new tasks, new domains, new variables. And this is called fine-tuning.  The idea is that the model really uses the learned representation to perform this adaptation very efficiently, which basically means Aurora is a powerful, flexible model that can relatively cheaply be adapted to any environmental forecasting task.    TINGLE: Wessel, can you tell us about your methodology? How did you all conduct this research?  BRUINSMA: While approaches so far have trained models on primarily one particular data These datasets are a combination of estimates of the historical state of the world, forecasts by other models, climate simulations, and more. We’ve been able to show that training on not just more data but more diverse data helps the model achieve even better performance. Showing this is difficult because there is just so much data.   In addition to scaling to more and more diverse data, we also increased the size of the model as much as we could. Here we found that bigger models, despite being slower to run, make more efficient use of computational resources. It’s cheaper to train a good big model than a good small model. The mantra of this project was to really keep it simple and to scale to simultaneously very large and, more importantly, diverse data and large model size.  TINGLE: So, Megan, what were your major findings? And we know they’re major because they’re in Nature. [LAUGHS]  STANLEY: Yeah, [LAUGHS] I guess they really are. So the main outcome of this project is we were actually able to train a single foundation model that achieves state-of-the-art performance in four different domains. Air pollution forecasting. For example, predicting particulate matter near the surface or ozone in the atmosphere. Ocean wave forecasting, which is critical for planning shipping routes.   Tropical cyclone track forecasting, so that means being able to predict where a hurricane or a typhoon is expected to go, which is obviously incredibly important, and very high-resolution weather forecasting.   And I’ve, kind of, named these forecasting domains as if they’re just items in a list, but in every single one, Aurora really pushed the limits of what is possible with AI models. And we’re really proud of that.   But perhaps, kind of, you know, to my mind, the key takeaway here is that the foundation model approach actually works. So what we have shown is it’s possible to actually train some kind of general model, a foundation model, and then adapt it to a wide variety of environmental tasks. Now we definitely do not claim that Aurora is some kind of ultimate environmental forecasting model. We are sure that the model and the pretraining procedure can actually be improved. But, nevertheless, we’ve shown that this approach works for environmental forecasting. It really holds massive promise, and that’s incredibly cool.  TINGLE: Wessel, what do you think will be the real-world impact of this work?  BRUINSMA: Well, for applications that we mentioned, which are air pollution forecasting, ocean wave forecasting, tropical cyclone track forecasting, and very high-resolution weather forecasting, Aurora could today be deployed in real-time systems to produce near real-time forecasts. And, you know, in fact, it already is. You can view real-time weather forecasts by the high-resolution version of the model on the website of ECMWF (European Centre for Medium-Range Weather Forecasts).  But what’s remarkable is that every of these applications took a small team of engineers about four to eight weeks to fully execute. You should compare this to a typical development timeline for more traditional models, which can be on the order of multiple years. Using the pretraining fine-tuning approach that we used for Aurora, we might see significantly accelerated development cycles for environmental forecasting problems. And that’s exciting.  TINGLE: Megan, if our listeners only walk away from this conversation with one key talking point, what would you like that to be? What should we remember about this paper?  STANLEY: The biggest takeaway is that the pretraining fine-tuning paradigm, it really works for environmental forecasting, right? So you can train a foundational model, it learns some kind of general-purpose representation of the Earth system dynamics, and this representation boosts performance in a wide variety of forecasting tasks. But we really want to emphasize that Aurora only scratches the surface of what’s actually possible.  So there are many more applications to explore than the four we’ve mentioned. And undoubtedly, the model and pretraining procedure can actually be improved. So we’re really excited to see what the next few years will bring.  TINGLE: Wessel, tell us more about those opportunities and unanswered questions. What’s next on the research agenda in environmental prediction?  BRUINSMA: Well, Aurora has two main limitations. The first is that the model produces only deterministic predictions, by which I mean a single predicted value. For variables like temperature, this is mostly fine. But other variables like precipitation, they are inherently some kind of stochastic. For these variables, we really want to assign probabilities to different levels of precipitation rather than predicting only a single value.  An extension of Aurora to allow this sort of prediction would be a great next step.   The second limitation is that Aurora depends on a procedure called assimilation. Assimilation attempts to create a starting point for the model from real-world observations, such as from weather stations and satellites. The model then takes the starting point and uses it to make predictions. Unfortunately, assimilation is super expensive, so it would be great if we could somehow circumvent the need for it.  Finally, what we find really important is to make our advancements available to the community. [MUSIC]  TINGLE: Great. Megan and Wessel, thanks for joining us today on the Microsoft Research Podcast.  BRUINSMA: Thanks for having us.  STANLEY: Yeah, thank you. It’s been great.  TINGLE: You can check out the Aurora model on Azure AI Foundry. You can read the entire paper, “A Foundation Model for the Earth System,” at aka.ms/abstracts. And you’ll certainly find it on the Nature website, too.   Thank you so much for tuning in to Abstracts today. Until next time.   [MUSIC FADES] 
    0 Yorumlar 0 hisse senetleri 0 önizleme
  • AI Is Rewriting Reality, One Word At A Time

    As AI reshapes language, even the human voice becomes a pattern to be predicted, not a meaning to be ... More understood.getty
    Language is the foundation of business, culture, and consciousness. But AI isn’t just using our words—it’s reshaping them. Quietly, subtly, it’s dismantling the architecture of thought by eroding what we used to think: nouns.

    We used to believe that naming something gave it power. Giving a thing a noun means tethering it to meaning, identity, and memory. But in the age of AI, nouns are dissolving—not banned, not erased—but rendered functionally obsolete. And with them, our grasp on reality is starting to fray.

    AI and the Architecture of Thought
    AI doesn’t see the world in things. It sees the world in patterns—actions, probabilities, and prompts. A chair is no longer an object; it’s “something to sit on.” A self is no longer an identity; it’s “a collection of behaviors and preferences.” Even brands, once nouns wrapped in mythology, are being reconstituted as verbs. You don’t have a brand. You do a brand.

    This linguistic shift isn’t neutral. It’s a collapse of conceptual anchors. In generative systems, nouns aren’t centers of gravity. They’re scaffolding for action. This reflects a broader trend in how generative AI is reshaping communication across every industry.

    Recent research supports this trend. A study titled Playing with Words: Comparing the Vocabulary and Lexical Richness of ChatGPT and Humans found that ChatGPT’s outputs exhibit significantly lower lexical diversity than human writing. In particular, nouns and specific, stylistic words are often underused, suggesting that generative systems prioritize predictable, commonly used language while deprioritizing less frequent terms.
    Further analysis of 14 million PubMed abstracts revealed a measurable shift in word frequency post-AI adoption. Words like “delves” and “showcasing” surged, while others faded—showing that large language models are already reshaping vocabulary patterns at scale.
    Sound familiar? It should.
    AI’s Philosophical Ancestors: Orwell, Huxley, and the Future They Warned Us About
    To understand their relevance, it helps to recall what George Orwell and Aldous Huxley are most famous for. Orwell authored 1984, a bleak vision of the future where an authoritarian regime weaponizes language to suppress independent thought and rewrite history.

    His concept of Newspeak—a restricted, simplified language designed to make dissent unthinkable—has become a cultural shorthand for manipulative control.
    On the other hand, Huxley wrote Brave New World, which envisioned a society not characterized by overt oppression, but rather by engineered pleasure, distraction, and passive conformity. In his world, people are conditioned into compliance not through violence but through comfort, entertainment, and chemical sedation.
    Both men anticipated futures in which language and meaning are compromised, but in radically different ways. Together, they map the two poles of how reality can be reconditioned: by force or indulgence.
    Few realize that George Orwell was once a student of Aldous Huxley. In the late 1910s, while Orwellstudied at Eton, Huxley taught him French. Their relationship was brief but prophetic. Decades later, each would author the defining visions of dystopia—1984 and Brave New World.
    After reading 1984, Huxley wrote to Orwell with a haunting message:

    Whether in actual fact the policy of the boot-on-the-face can go on indefinitely seems doubtful… The future will be controlled by inflicting pleasure, not pain.

    And that’s precisely where we are now.
    Orwell feared control through surveillance and terror. Huxley feared control through indulgence and distraction. Generative AI, cloaked in helpfulness, embodies both. It doesn’t censor. It seduces. It doesn’t need Newspeak to delete ideas. It replaces them with prediction.
    In 1984, language was weaponized by force. In our world, it’s being reshaped by suggestion. What we have is not Artificial Intelligence—it’s Artificial Inference: trained not to understand but to remix, not to reason but to simulate.
    And this simulation brings us to a more profound loss: intersubjectivity.
    AI and the Loss of Intersubjectivity
    Humans learn, grow, and build reality through intersubjectivity—the shared context that gives language its weight. It allows us to share meaning, to agree on what a word represents, and to build mutual understanding through shared experiences. Without it, words float.
    AI doesn’t participate in intersubjectivity. It doesn’t share meaning—it predicts output. And yet, when someone asks an AI a question, they often believe the answer reflects their framing. It doesn’t. It reflects the average of averages, the statistical ghost of comprehension. The illusion of understanding is precise, polite, and utterly hollow.
    This is how AI reconditions reality at scale—not by force, but by imitation.
    The result? A slow, silent attrition of originality. Nouns lose their edges. Ideas lose their anchors. Authorship bleeds into prompting. And truth becomes whatever the model says most often.
    AI and Accountability: A Case Study in Trust and Miscommunication
    In one recent public example, Air Canada deployed an AI-powered chatbot to handle customer service inquiries. When a customer asked about bereavement fare discounts, the chatbot confidently invented a policy that didn’t exist. The airline initially tried to avoid responsibility, but the court disagreed. In February 2024, a tribunal ruled that Air Canada was liable for the misinformation provided by its chatbot.
    This wasn’t just a technical glitch—it was a trust failure. The AI-generated text sounded plausible, helpful, and human, but it lacked grounding in policy, context, or shared understanding. In effect, the airline’s brand spoke out of both sides of its mouth and cost them. This is the risk when language is generated without intersubjectivity, oversight, or friction.
    The Linguistic Drift of AI: What the Data Tells Us About Language Decay
    It’s not just theory—research is now quantifying how generative AI systems are shifting the landscape of language itself. A study titled Playing with Words: Comparing the Vocabulary and Lexical Richness of ChatGPT and Humans found that AI-generated outputs consistently use a narrower vocabulary, with significantly fewer nouns and stylistic words than human writing.
    Building on this, an analysis of over 14 million PubMed abstracts revealed measurable shifts in word frequency following the rise of LLM use. While many precise, technical nouns faded, terms like “delves” and “showcasing” surged. The shift is not random; it’s a statistically driven flattening of language, where standard, action-oriented, or stylistic terms are promoted, and specificity is sidelined.
    Some researchers link this to a broader problem known as “model collapse.” As AI models are increasingly trained on synthetic data, including their outputs, they may degrade over time. This leads to a feedback loop where less diverse, less semantically rich language becomes the norm. The result is a measurable reduction in lexical, syntactic, and semantic diversity—the very fabric of meaning and precision.
    The implications are vast. If AI systems are deprioritizing nouns at scale, then the structures we use to hold ideas—people, places, identities, and concepts—are being eroded. In real time, we are watching the grammatical infrastructure of human thought being reweighted by machines that do not think.
    What AI’s Language Shift Means for Brands and Business Strategy
    The erosion of language precision has significant implications for businesses, particularly those that rely on storytelling, branding, and effective communication. Brands are built on narrative consistency, anchored by nouns, identities, and associations that accumulate cultural weight over time.
    However, as AI systems normalize probabilistic language and predictive phrasing, even brand voice becomes a casualty of convergence. Differentiation erodes—messaging blurs. Trust becomes more complicated to earn and more uncomplicated to mimic.
    As this Forbes piece outlines, there are serious reasons why brands must be cautious with generative AI when it comes to preserving authenticity and voice.
    Moreover, AI-powered content platforms optimize for engagement, not meaning. Businesses relying on LLMs to generate customer-facing content risk flattening their uniqueness in favor of what’s statistically safe. Without human oversight, brand language may drift toward the generic, the probable, and the forgettable.
    How To Safeguard Meaning in the Age of AI
    Resist the flattening. Businesses and individuals alike must reclaim intentionality in language. Here’s how—and why it matters:
    If you don’t define your brand voice, AI will average it. If you don’t protect the language of your contracts, AI will remix it. If you don’t curate your culture, AI will feed it back to you—statistically safe but spiritually hollow.

    Double down on human authorship: Don’t outsource your voice to a model. Use AI for augmentation, not substitution.
    Protect linguistic originality: Encourage specificity, metaphor, and vocabulary diversity in your communication. Nouns matter.
    Audit your outputs: Periodically review AI-generated materials. Look for signs of drift—has your language lost its edge?
    Invest in language guardianship: Treat your brand’s lexicon like intellectual property. Define it. Defend it.
    Champion intersubjectivity: Allow shared context in both personal and professional communication. AI can simulate, but only humans can mean.

    The Necessity of Friction: Why Human Involvement Must Temper AI
    Friction isn’t a flaw in human systems—it’s a feature. It’s where meaning is made, thought is tested, and creativity wrestles with uncertainty. Automation is a powerful economic accelerant, but without deliberate pauses—without a human in the loop—we risk stripping away the qualities that make us human. Language is one of those qualities.
    Every hesitation, nuance, and word choice reflects cognition, culture, and care. Remove the friction, and you remove the humanity. AI can offer speed, fluency, and pattern-matching, but it can’t provide presence, and presence is where meaning lives.
    AI’s Closing Refrain: A Call to Remember Meaning
    Emily M. Bender, a professor of computational linguistics at the University of Washington, has emerged as one of the most principled and prescient critics of large language models. In her now-famous co-authored paper, "On the Dangers of Stochastic Parrots," she argues that these systems don’t understand language—they merely remix it. They are, in her words, “stochastic parrots”: machines that generate plausible-sounding language without comprehension or intent.
    Yet we’re letting those parrots draft our emails, write our ads, and even shape our laws. We’re allowing models trained on approximations to become arbiters of communication, culture, and identity.
    This is not language—it’s mimicry at scale. And mimicry, unchecked, becomes a distortion. When AI outputs are mistaken for understanding, the baseline of meaning erodes. The problem isn’t just that AI might be wrong. It’s that it sounds so right, we stop questioning it.
    In the name of optimization, we risk erasing the texture of human communication. Our metaphors, our double meanings, our moments of productive ambiguity—these are what make language alive. Remove that, and a stream of consensus-safe, risk-averse echo remains. Functional? Yes. Meaningful? Not really.
    The stakes aren’t just literary—they’re existential. If language is the connective tissue between thought and reality, and if that tissue is replaced with statistical scaffolding, thinking becomes outsourced. Once sharpened by friction, our voices become blurred in a sea of plausible phrasings.
    Without intersubjectivity, friction, or nouns, we are scripting ourselves out of the story, one autocomplete at a time We are not being silenced. We are being auto-completed. And the most dangerous part? We asked for it.
    Before we ask what AI can say next, we should ask: What has already gone unsaid?
    In this quiet war, we don’t lose language all at once. We lose it word by word—until we forget we ever had something to say.
    I asked brand strategist and storyteller Michelle Garside, whose work spans billion-dollar brands and purpose-driven founders, to share her perspective on what’s at risk as automation flattened language. Her response was both precise and profound:

    If language is being flattened, we need more people doing the opposite: excavating. Listening for what’s buried beneath the noise. Uncovering the phrase that unlocks the person. That’s not a prompt—it’s a process. And it’s a deeply human one.

    When someone says something that lands—not because it sounds good, but because it’s true. You can see it in their body. You can feel it in the silence that follows. No algorithm can replicate that because that moment isn’t statistical. It’s sacred.

    The risk isn’t just that AI will get things wrong. It’s that it will sound just right enough to stop us from looking deeper. To stop us from asking what’s real. To stop us from finding the words only we could say.

    We don’t need more words. We need more meaning. And meaning isn’t generated. It’s remembered.

    When it comes to language and AI, that’s the line to carry forward—not just because it sounds good, but because it’s true.
    #rewriting #reality #one #word #time
    AI Is Rewriting Reality, One Word At A Time
    As AI reshapes language, even the human voice becomes a pattern to be predicted, not a meaning to be ... More understood.getty Language is the foundation of business, culture, and consciousness. But AI isn’t just using our words—it’s reshaping them. Quietly, subtly, it’s dismantling the architecture of thought by eroding what we used to think: nouns. We used to believe that naming something gave it power. Giving a thing a noun means tethering it to meaning, identity, and memory. But in the age of AI, nouns are dissolving—not banned, not erased—but rendered functionally obsolete. And with them, our grasp on reality is starting to fray. AI and the Architecture of Thought AI doesn’t see the world in things. It sees the world in patterns—actions, probabilities, and prompts. A chair is no longer an object; it’s “something to sit on.” A self is no longer an identity; it’s “a collection of behaviors and preferences.” Even brands, once nouns wrapped in mythology, are being reconstituted as verbs. You don’t have a brand. You do a brand. This linguistic shift isn’t neutral. It’s a collapse of conceptual anchors. In generative systems, nouns aren’t centers of gravity. They’re scaffolding for action. This reflects a broader trend in how generative AI is reshaping communication across every industry. Recent research supports this trend. A study titled Playing with Words: Comparing the Vocabulary and Lexical Richness of ChatGPT and Humans found that ChatGPT’s outputs exhibit significantly lower lexical diversity than human writing. In particular, nouns and specific, stylistic words are often underused, suggesting that generative systems prioritize predictable, commonly used language while deprioritizing less frequent terms. Further analysis of 14 million PubMed abstracts revealed a measurable shift in word frequency post-AI adoption. Words like “delves” and “showcasing” surged, while others faded—showing that large language models are already reshaping vocabulary patterns at scale. Sound familiar? It should. AI’s Philosophical Ancestors: Orwell, Huxley, and the Future They Warned Us About To understand their relevance, it helps to recall what George Orwell and Aldous Huxley are most famous for. Orwell authored 1984, a bleak vision of the future where an authoritarian regime weaponizes language to suppress independent thought and rewrite history. His concept of Newspeak—a restricted, simplified language designed to make dissent unthinkable—has become a cultural shorthand for manipulative control. On the other hand, Huxley wrote Brave New World, which envisioned a society not characterized by overt oppression, but rather by engineered pleasure, distraction, and passive conformity. In his world, people are conditioned into compliance not through violence but through comfort, entertainment, and chemical sedation. Both men anticipated futures in which language and meaning are compromised, but in radically different ways. Together, they map the two poles of how reality can be reconditioned: by force or indulgence. Few realize that George Orwell was once a student of Aldous Huxley. In the late 1910s, while Orwellstudied at Eton, Huxley taught him French. Their relationship was brief but prophetic. Decades later, each would author the defining visions of dystopia—1984 and Brave New World. After reading 1984, Huxley wrote to Orwell with a haunting message: Whether in actual fact the policy of the boot-on-the-face can go on indefinitely seems doubtful… The future will be controlled by inflicting pleasure, not pain. And that’s precisely where we are now. Orwell feared control through surveillance and terror. Huxley feared control through indulgence and distraction. Generative AI, cloaked in helpfulness, embodies both. It doesn’t censor. It seduces. It doesn’t need Newspeak to delete ideas. It replaces them with prediction. In 1984, language was weaponized by force. In our world, it’s being reshaped by suggestion. What we have is not Artificial Intelligence—it’s Artificial Inference: trained not to understand but to remix, not to reason but to simulate. And this simulation brings us to a more profound loss: intersubjectivity. AI and the Loss of Intersubjectivity Humans learn, grow, and build reality through intersubjectivity—the shared context that gives language its weight. It allows us to share meaning, to agree on what a word represents, and to build mutual understanding through shared experiences. Without it, words float. AI doesn’t participate in intersubjectivity. It doesn’t share meaning—it predicts output. And yet, when someone asks an AI a question, they often believe the answer reflects their framing. It doesn’t. It reflects the average of averages, the statistical ghost of comprehension. The illusion of understanding is precise, polite, and utterly hollow. This is how AI reconditions reality at scale—not by force, but by imitation. The result? A slow, silent attrition of originality. Nouns lose their edges. Ideas lose their anchors. Authorship bleeds into prompting. And truth becomes whatever the model says most often. AI and Accountability: A Case Study in Trust and Miscommunication In one recent public example, Air Canada deployed an AI-powered chatbot to handle customer service inquiries. When a customer asked about bereavement fare discounts, the chatbot confidently invented a policy that didn’t exist. The airline initially tried to avoid responsibility, but the court disagreed. In February 2024, a tribunal ruled that Air Canada was liable for the misinformation provided by its chatbot. This wasn’t just a technical glitch—it was a trust failure. The AI-generated text sounded plausible, helpful, and human, but it lacked grounding in policy, context, or shared understanding. In effect, the airline’s brand spoke out of both sides of its mouth and cost them. This is the risk when language is generated without intersubjectivity, oversight, or friction. The Linguistic Drift of AI: What the Data Tells Us About Language Decay It’s not just theory—research is now quantifying how generative AI systems are shifting the landscape of language itself. A study titled Playing with Words: Comparing the Vocabulary and Lexical Richness of ChatGPT and Humans found that AI-generated outputs consistently use a narrower vocabulary, with significantly fewer nouns and stylistic words than human writing. Building on this, an analysis of over 14 million PubMed abstracts revealed measurable shifts in word frequency following the rise of LLM use. While many precise, technical nouns faded, terms like “delves” and “showcasing” surged. The shift is not random; it’s a statistically driven flattening of language, where standard, action-oriented, or stylistic terms are promoted, and specificity is sidelined. Some researchers link this to a broader problem known as “model collapse.” As AI models are increasingly trained on synthetic data, including their outputs, they may degrade over time. This leads to a feedback loop where less diverse, less semantically rich language becomes the norm. The result is a measurable reduction in lexical, syntactic, and semantic diversity—the very fabric of meaning and precision. The implications are vast. If AI systems are deprioritizing nouns at scale, then the structures we use to hold ideas—people, places, identities, and concepts—are being eroded. In real time, we are watching the grammatical infrastructure of human thought being reweighted by machines that do not think. What AI’s Language Shift Means for Brands and Business Strategy The erosion of language precision has significant implications for businesses, particularly those that rely on storytelling, branding, and effective communication. Brands are built on narrative consistency, anchored by nouns, identities, and associations that accumulate cultural weight over time. However, as AI systems normalize probabilistic language and predictive phrasing, even brand voice becomes a casualty of convergence. Differentiation erodes—messaging blurs. Trust becomes more complicated to earn and more uncomplicated to mimic. As this Forbes piece outlines, there are serious reasons why brands must be cautious with generative AI when it comes to preserving authenticity and voice. Moreover, AI-powered content platforms optimize for engagement, not meaning. Businesses relying on LLMs to generate customer-facing content risk flattening their uniqueness in favor of what’s statistically safe. Without human oversight, brand language may drift toward the generic, the probable, and the forgettable. How To Safeguard Meaning in the Age of AI Resist the flattening. Businesses and individuals alike must reclaim intentionality in language. Here’s how—and why it matters: If you don’t define your brand voice, AI will average it. If you don’t protect the language of your contracts, AI will remix it. If you don’t curate your culture, AI will feed it back to you—statistically safe but spiritually hollow. Double down on human authorship: Don’t outsource your voice to a model. Use AI for augmentation, not substitution. Protect linguistic originality: Encourage specificity, metaphor, and vocabulary diversity in your communication. Nouns matter. Audit your outputs: Periodically review AI-generated materials. Look for signs of drift—has your language lost its edge? Invest in language guardianship: Treat your brand’s lexicon like intellectual property. Define it. Defend it. Champion intersubjectivity: Allow shared context in both personal and professional communication. AI can simulate, but only humans can mean. The Necessity of Friction: Why Human Involvement Must Temper AI Friction isn’t a flaw in human systems—it’s a feature. It’s where meaning is made, thought is tested, and creativity wrestles with uncertainty. Automation is a powerful economic accelerant, but without deliberate pauses—without a human in the loop—we risk stripping away the qualities that make us human. Language is one of those qualities. Every hesitation, nuance, and word choice reflects cognition, culture, and care. Remove the friction, and you remove the humanity. AI can offer speed, fluency, and pattern-matching, but it can’t provide presence, and presence is where meaning lives. AI’s Closing Refrain: A Call to Remember Meaning Emily M. Bender, a professor of computational linguistics at the University of Washington, has emerged as one of the most principled and prescient critics of large language models. In her now-famous co-authored paper, "On the Dangers of Stochastic Parrots," she argues that these systems don’t understand language—they merely remix it. They are, in her words, “stochastic parrots”: machines that generate plausible-sounding language without comprehension or intent. Yet we’re letting those parrots draft our emails, write our ads, and even shape our laws. We’re allowing models trained on approximations to become arbiters of communication, culture, and identity. This is not language—it’s mimicry at scale. And mimicry, unchecked, becomes a distortion. When AI outputs are mistaken for understanding, the baseline of meaning erodes. The problem isn’t just that AI might be wrong. It’s that it sounds so right, we stop questioning it. In the name of optimization, we risk erasing the texture of human communication. Our metaphors, our double meanings, our moments of productive ambiguity—these are what make language alive. Remove that, and a stream of consensus-safe, risk-averse echo remains. Functional? Yes. Meaningful? Not really. The stakes aren’t just literary—they’re existential. If language is the connective tissue between thought and reality, and if that tissue is replaced with statistical scaffolding, thinking becomes outsourced. Once sharpened by friction, our voices become blurred in a sea of plausible phrasings. Without intersubjectivity, friction, or nouns, we are scripting ourselves out of the story, one autocomplete at a time We are not being silenced. We are being auto-completed. And the most dangerous part? We asked for it. Before we ask what AI can say next, we should ask: What has already gone unsaid? In this quiet war, we don’t lose language all at once. We lose it word by word—until we forget we ever had something to say. I asked brand strategist and storyteller Michelle Garside, whose work spans billion-dollar brands and purpose-driven founders, to share her perspective on what’s at risk as automation flattened language. Her response was both precise and profound: If language is being flattened, we need more people doing the opposite: excavating. Listening for what’s buried beneath the noise. Uncovering the phrase that unlocks the person. That’s not a prompt—it’s a process. And it’s a deeply human one. When someone says something that lands—not because it sounds good, but because it’s true. You can see it in their body. You can feel it in the silence that follows. No algorithm can replicate that because that moment isn’t statistical. It’s sacred. The risk isn’t just that AI will get things wrong. It’s that it will sound just right enough to stop us from looking deeper. To stop us from asking what’s real. To stop us from finding the words only we could say. We don’t need more words. We need more meaning. And meaning isn’t generated. It’s remembered. When it comes to language and AI, that’s the line to carry forward—not just because it sounds good, but because it’s true. #rewriting #reality #one #word #time
    WWW.FORBES.COM
    AI Is Rewriting Reality, One Word At A Time
    As AI reshapes language, even the human voice becomes a pattern to be predicted, not a meaning to be ... More understood.getty Language is the foundation of business, culture, and consciousness. But AI isn’t just using our words—it’s reshaping them. Quietly, subtly, it’s dismantling the architecture of thought by eroding what we used to think: nouns. We used to believe that naming something gave it power. Giving a thing a noun means tethering it to meaning, identity, and memory. But in the age of AI, nouns are dissolving—not banned, not erased—but rendered functionally obsolete. And with them, our grasp on reality is starting to fray. AI and the Architecture of Thought AI doesn’t see the world in things. It sees the world in patterns—actions, probabilities, and prompts. A chair is no longer an object; it’s “something to sit on.” A self is no longer an identity; it’s “a collection of behaviors and preferences.” Even brands, once nouns wrapped in mythology, are being reconstituted as verbs. You don’t have a brand. You do a brand. This linguistic shift isn’t neutral. It’s a collapse of conceptual anchors. In generative systems, nouns aren’t centers of gravity. They’re scaffolding for action. This reflects a broader trend in how generative AI is reshaping communication across every industry. Recent research supports this trend. A study titled Playing with Words: Comparing the Vocabulary and Lexical Richness of ChatGPT and Humans found that ChatGPT’s outputs exhibit significantly lower lexical diversity than human writing. In particular, nouns and specific, stylistic words are often underused, suggesting that generative systems prioritize predictable, commonly used language while deprioritizing less frequent terms. Further analysis of 14 million PubMed abstracts revealed a measurable shift in word frequency post-AI adoption. Words like “delves” and “showcasing” surged, while others faded—showing that large language models are already reshaping vocabulary patterns at scale. Sound familiar? It should. AI’s Philosophical Ancestors: Orwell, Huxley, and the Future They Warned Us About To understand their relevance, it helps to recall what George Orwell and Aldous Huxley are most famous for. Orwell authored 1984, a bleak vision of the future where an authoritarian regime weaponizes language to suppress independent thought and rewrite history. His concept of Newspeak—a restricted, simplified language designed to make dissent unthinkable—has become a cultural shorthand for manipulative control. On the other hand, Huxley wrote Brave New World, which envisioned a society not characterized by overt oppression, but rather by engineered pleasure, distraction, and passive conformity. In his world, people are conditioned into compliance not through violence but through comfort, entertainment, and chemical sedation. Both men anticipated futures in which language and meaning are compromised, but in radically different ways. Together, they map the two poles of how reality can be reconditioned: by force or indulgence. Few realize that George Orwell was once a student of Aldous Huxley. In the late 1910s, while Orwell (then Eric Blair) studied at Eton, Huxley taught him French. Their relationship was brief but prophetic. Decades later, each would author the defining visions of dystopia—1984 and Brave New World. After reading 1984, Huxley wrote to Orwell with a haunting message: Whether in actual fact the policy of the boot-on-the-face can go on indefinitely seems doubtful… The future will be controlled by inflicting pleasure, not pain. And that’s precisely where we are now. Orwell feared control through surveillance and terror. Huxley feared control through indulgence and distraction. Generative AI, cloaked in helpfulness, embodies both. It doesn’t censor. It seduces. It doesn’t need Newspeak to delete ideas. It replaces them with prediction. In 1984, language was weaponized by force. In our world, it’s being reshaped by suggestion. What we have is not Artificial Intelligence—it’s Artificial Inference: trained not to understand but to remix, not to reason but to simulate. And this simulation brings us to a more profound loss: intersubjectivity. AI and the Loss of Intersubjectivity Humans learn, grow, and build reality through intersubjectivity—the shared context that gives language its weight. It allows us to share meaning, to agree on what a word represents, and to build mutual understanding through shared experiences. Without it, words float. AI doesn’t participate in intersubjectivity. It doesn’t share meaning—it predicts output. And yet, when someone asks an AI a question, they often believe the answer reflects their framing. It doesn’t. It reflects the average of averages, the statistical ghost of comprehension. The illusion of understanding is precise, polite, and utterly hollow. This is how AI reconditions reality at scale—not by force, but by imitation. The result? A slow, silent attrition of originality. Nouns lose their edges. Ideas lose their anchors. Authorship bleeds into prompting. And truth becomes whatever the model says most often. AI and Accountability: A Case Study in Trust and Miscommunication In one recent public example, Air Canada deployed an AI-powered chatbot to handle customer service inquiries. When a customer asked about bereavement fare discounts, the chatbot confidently invented a policy that didn’t exist. The airline initially tried to avoid responsibility, but the court disagreed. In February 2024, a tribunal ruled that Air Canada was liable for the misinformation provided by its chatbot. This wasn’t just a technical glitch—it was a trust failure. The AI-generated text sounded plausible, helpful, and human, but it lacked grounding in policy, context, or shared understanding. In effect, the airline’s brand spoke out of both sides of its mouth and cost them. This is the risk when language is generated without intersubjectivity, oversight, or friction. The Linguistic Drift of AI: What the Data Tells Us About Language Decay It’s not just theory—research is now quantifying how generative AI systems are shifting the landscape of language itself. A study titled Playing with Words: Comparing the Vocabulary and Lexical Richness of ChatGPT and Humans found that AI-generated outputs consistently use a narrower vocabulary, with significantly fewer nouns and stylistic words than human writing. Building on this, an analysis of over 14 million PubMed abstracts revealed measurable shifts in word frequency following the rise of LLM use. While many precise, technical nouns faded, terms like “delves” and “showcasing” surged. The shift is not random; it’s a statistically driven flattening of language, where standard, action-oriented, or stylistic terms are promoted, and specificity is sidelined. Some researchers link this to a broader problem known as “model collapse.” As AI models are increasingly trained on synthetic data, including their outputs, they may degrade over time. This leads to a feedback loop where less diverse, less semantically rich language becomes the norm. The result is a measurable reduction in lexical, syntactic, and semantic diversity—the very fabric of meaning and precision. The implications are vast. If AI systems are deprioritizing nouns at scale, then the structures we use to hold ideas—people, places, identities, and concepts—are being eroded. In real time, we are watching the grammatical infrastructure of human thought being reweighted by machines that do not think. What AI’s Language Shift Means for Brands and Business Strategy The erosion of language precision has significant implications for businesses, particularly those that rely on storytelling, branding, and effective communication. Brands are built on narrative consistency, anchored by nouns, identities, and associations that accumulate cultural weight over time. However, as AI systems normalize probabilistic language and predictive phrasing, even brand voice becomes a casualty of convergence. Differentiation erodes—messaging blurs. Trust becomes more complicated to earn and more uncomplicated to mimic. As this Forbes piece outlines, there are serious reasons why brands must be cautious with generative AI when it comes to preserving authenticity and voice. Moreover, AI-powered content platforms optimize for engagement, not meaning. Businesses relying on LLMs to generate customer-facing content risk flattening their uniqueness in favor of what’s statistically safe. Without human oversight, brand language may drift toward the generic, the probable, and the forgettable. How To Safeguard Meaning in the Age of AI Resist the flattening. Businesses and individuals alike must reclaim intentionality in language. Here’s how—and why it matters: If you don’t define your brand voice, AI will average it. If you don’t protect the language of your contracts, AI will remix it. If you don’t curate your culture, AI will feed it back to you—statistically safe but spiritually hollow. Double down on human authorship: Don’t outsource your voice to a model. Use AI for augmentation, not substitution. Protect linguistic originality: Encourage specificity, metaphor, and vocabulary diversity in your communication. Nouns matter. Audit your outputs: Periodically review AI-generated materials. Look for signs of drift—has your language lost its edge? Invest in language guardianship: Treat your brand’s lexicon like intellectual property (IP). Define it. Defend it. Champion intersubjectivity: Allow shared context in both personal and professional communication. AI can simulate, but only humans can mean. The Necessity of Friction: Why Human Involvement Must Temper AI Friction isn’t a flaw in human systems—it’s a feature. It’s where meaning is made, thought is tested, and creativity wrestles with uncertainty. Automation is a powerful economic accelerant, but without deliberate pauses—without a human in the loop—we risk stripping away the qualities that make us human. Language is one of those qualities. Every hesitation, nuance, and word choice reflects cognition, culture, and care. Remove the friction, and you remove the humanity. AI can offer speed, fluency, and pattern-matching, but it can’t provide presence, and presence is where meaning lives. AI’s Closing Refrain: A Call to Remember Meaning Emily M. Bender, a professor of computational linguistics at the University of Washington, has emerged as one of the most principled and prescient critics of large language models. In her now-famous co-authored paper, "On the Dangers of Stochastic Parrots," she argues that these systems don’t understand language—they merely remix it. They are, in her words, “stochastic parrots”: machines that generate plausible-sounding language without comprehension or intent. Yet we’re letting those parrots draft our emails, write our ads, and even shape our laws. We’re allowing models trained on approximations to become arbiters of communication, culture, and identity. This is not language—it’s mimicry at scale. And mimicry, unchecked, becomes a distortion. When AI outputs are mistaken for understanding, the baseline of meaning erodes. The problem isn’t just that AI might be wrong. It’s that it sounds so right, we stop questioning it. In the name of optimization, we risk erasing the texture of human communication. Our metaphors, our double meanings, our moments of productive ambiguity—these are what make language alive. Remove that, and a stream of consensus-safe, risk-averse echo remains. Functional? Yes. Meaningful? Not really. The stakes aren’t just literary—they’re existential. If language is the connective tissue between thought and reality, and if that tissue is replaced with statistical scaffolding, thinking becomes outsourced. Once sharpened by friction, our voices become blurred in a sea of plausible phrasings. Without intersubjectivity, friction, or nouns, we are scripting ourselves out of the story, one autocomplete at a time We are not being silenced. We are being auto-completed. And the most dangerous part? We asked for it. Before we ask what AI can say next, we should ask: What has already gone unsaid? In this quiet war, we don’t lose language all at once. We lose it word by word—until we forget we ever had something to say. I asked brand strategist and storyteller Michelle Garside, whose work spans billion-dollar brands and purpose-driven founders, to share her perspective on what’s at risk as automation flattened language. Her response was both precise and profound: If language is being flattened, we need more people doing the opposite: excavating. Listening for what’s buried beneath the noise. Uncovering the phrase that unlocks the person. That’s not a prompt—it’s a process. And it’s a deeply human one. When someone says something that lands—not because it sounds good, but because it’s true. You can see it in their body. You can feel it in the silence that follows. No algorithm can replicate that because that moment isn’t statistical. It’s sacred. The risk isn’t just that AI will get things wrong. It’s that it will sound just right enough to stop us from looking deeper. To stop us from asking what’s real. To stop us from finding the words only we could say. We don’t need more words. We need more meaning. And meaning isn’t generated. It’s remembered. When it comes to language and AI, that’s the line to carry forward—not just because it sounds good, but because it’s true.
    0 Yorumlar 0 hisse senetleri 0 önizleme
CGShares https://cgshares.com