• WWW.TECHRADAR.COM
    Leaked dummy unit image shows how thin the iPhone 17 Air may look against the iPhone 17 Pro
    We've now got a very good idea of how much the iPhone 17 Air will differ from the iPhone 17 Pro Max in terms of thickness.
    0 Commentarios 0 Acciones 22 Views
  • WWW.CNBC.COM
    For bitcoin bulls who self-custody crypto, the global risks are growing
    Many crypto believers don't trust brokers and exchanges to hold their bitcoin, but self-custody has become dangerous with climate and conflict risks.
    0 Commentarios 0 Acciones 30 Views
  • WWW.FASTCOMPANY.COM
    Why this iconic San Francisco pizza shop is going all in on QR codesdespite the skepticism
    To San Francisco chef and restaurateur Thomas McNaughton, QR codes are an efficient way to serve a crowd. Sure, the codesand restaurants that use themhave endured much loathing. And, yes, people still love to criticize them. But at the newest location of McNaughtons Flour + Water Pizzeria, set to open later this month, QR codes are the star.Theres good reason. The 1,800-square-foot restaurant sits a few blocks from Oracle Park, where the San Francisco Giants just opened the baseball season. It needs to handle serious spikes in business from game-day crowds and pump out pizzas fast.We envision a scenario where, for two hours, its completely gangbusters before the game. How can we help manage those crowds? McNaughton says.The team, with the help of a local design agency, got creative. Inside the restaurant, guests cant miss the large, stylized poster featuring a Giants player with a QR code for a face. Scan to place a takeout order, it reads, & pick up around the corner.The restaurant also printed baseball-style trading cards with the same image, for fast QR ordering (for pickup or delivery) from afar. Smaller codes dot a handful of tables. If youre lucky enough to snag one, orders placed digitally are delivered right to your seat.Research from the National Restaurant Association shows that about half of diners are keen to order via QR code at a quick-service restaurant like this onebut you might not know it based on very loud grumblings everywhere you look.The QR backlash has won, declared a headline in the Wall Street Journal last year. A year earlier, I was quoted as an industry expert in a New York Times piece titled, The QR-code menu is being shown the door. About twice a month, someone sends me an Instagram post where a celeb or influencer or random stranger complainsstillabout using the tech.[Source Photo: Flour + Water]For all the whining, the codes are . . . really useful. When deployed thoughtfully, theyre downright hospitable. The skepticism, McNaughton thinks, has to do with negative emotions associated with themoment that QR codes rocketed to restaurant infamy during the earliest days of the pandemic.I think the pushback that you heard was partly because everything was so different, he says. Every restaurant was just trying to stay afloat and trying to be accessible while still being safe.Its true. Clunky QR-code menus promised to keep shared surfaces touch-free, an almost quaint, if completely misguided effort from a challenging time. Since then, the codes have evolved with utility in mind. Major restaurant technology companies like point-of-sale and payments giant Toast built QR ordering into their products. Now, customer orders go straight into the system, bypassing human servers and their potential human mistakes.Digital orders also shorten the distance between diner and kitchen, McNaughton explains, a shortcut that allows the pizzeria to pump out orders much faster and keep crowds happier. (People placing digital orders get to skip the presumably long line.)Its a model specific to this location, which McNaughton calls a fast-casual offshoot of his restaurant groups original, much larger pizzeria. Plenty of fast-casual and fast-food restaurants (or quick service, in restaurant lingo) are chasing digital orders. Last month, Taco Bell parent company Yum Brands tapped AI powerhouse Nvidia to coax more orders online, in an effort to one day digitize every order.Flour + Water Pizzeria isnt planning to push all of its orders online; it employs humans who can take orders on countertop tablets, or, if necessary, flip them over to become self-service touch-screen ordering kiosks. The guest chooses based on how, and how fast, they want to get their order.You can talk to somebody, you can use a kiosk, or you can just skip that line completely, McNaughton says. Its about efficiency.
    0 Commentarios 0 Acciones 34 Views
  • WWW.YANKODESIGN.COM
    Next-Gen Xbox Handheld Leaked? NBC Today Show Accidentally Reveals Mysterious New Device
    AI Render VisualizationIn a blink-and-youll-miss-it moment that set gaming forums ablaze, the NBC Today Show may have accidentally pulled back the curtain on one of Microsofts most tightly kept secrets. While touring a confidential Xbox hardware lab, the camera briefly panned to Phil Spencers officean area known for housing Xbox relics and early design experiments. Among the memorabilia sat something far more curious: a handheld console with Xbox branding that doesnt match any known prototype, including the rumored Project Kennan.For years, whispers of a portable Xbox have floated through the gaming world, sometimes surfacing as patents or vague interviews. This time, the evidence was visual. The device shown had a familiar layoutyour standard face buttons, analog sticks, and directional padbut with an unusual pair of buttons placed below the right stick. That detail stood out because it didnt align with known handhelds like the Asus ROG Ally or even the Kishi V2 controller. If it were simply a third-party gadget, it would be odd to see it so prominently displayed in Spencers workspace.Designer: XboxSneak peek shown during interview by NBC TodayThe speculation machine kicked into high gear almost immediately. Notably, this handheld doesnt resemble the white Asus-made prototype recently teased online, which many believe is Project Kennan. That device, expected to run a Windows-based platform with an Xbox-specific UI, has been anticipated as Microsofts official answer to Steam Deck and ROG Ally dominance. But what Spencer had might suggest something either earlier in the dev cycleor a different approach entirely.AI Render VisualizationIt wouldnt be the first time Microsoft tested competing ideas under the same umbrella. During a 2024 podcast, Xbox insider Jez Corden mentioned several portable Xbox prototypes floating around internally. Some of those designs were shared privately, and at least one bears a strong resemblance to the mystery device shown in the NBC segment. Spencer himself has alluded to a desire for a handheld console, often referencing a hybrid design that borrows the best from both Windows and Xbox platforms.AI Render VisualizationWhat makes this sighting fascinating is its context. NBCs tour wasnt about announcing hardwareit was framed around showing the behind-the-scenes process of testing and validating Xbox peripherals. Seeing a potential next-gen product in that environment suggests its in active evaluation, not just some archival curiosity. Everything on Spencers shelves is deliberate, often foreshadowing what Microsoft is cooking up.AI Render VisualizationWhile we only caught the right half of the device, its physical presence raises real questions. Is this a shelved concept? A backup plan if Project Kennan falters? Or a step beyond it? Some fans think the layout implies a more gamepad-centric experience than the current PC-first handhelds. It might be optimized for streaming or even native gameplay, something more plug-and-play for casual users rather than the tinkering crowd drawn to SteamOS or Windows handhelds.AI Render VisualizationThe broader implication is clear: Xbox isnt content staying stationary. Handheld gaming is surging again (especially with the Switch 2 just getting officially announced last week), and Microsoft has the software library, infrastructure, and nowpotentiallythe hardware to make a serious play. Between Xbox Cloud Gaming, Game Pass, and a growing appetite for on-the-go experiences, a portable device could tighten the ecosystem while expanding Xboxs reach to new audiences.Image source: ThinkComputersWe may not see this particular handheld on shelves anytime soon. It could be a relic from a branch of development thats since pivoted. Or it might resurface, refined and ready, when the timing aligns with the next major Xbox cycle. Whats certain is that the door is wide openand the race for the perfect Xbox handheld is heating up behind it.The post Next-Gen Xbox Handheld Leaked? NBC Today Show Accidentally Reveals Mysterious New Device first appeared on Yanko Design.
    0 Commentarios 0 Acciones 24 Views
  • 0 Commentarios 0 Acciones 24 Views
  • WWW.WIRED.COM
    Scientists Are Mapping the Boundaries of What Is Knowable and Unknowable
    Math and computer science researchers have long known that some questions are fundamentally unanswerable. Now physicists are exploring how physical systems put hard limits on what we can predict.
    0 Commentarios 0 Acciones 26 Views
  • APPLEINSIDER.COM
    Apple may be able to delay price increases, but not for long
    In the face of crippling tariffs, Apple will have to raise prices of popular items like the iPhone in the US and elsewhere but is trying to delay increases until the next set of hardware upgrades.The iPhone and other Apple products will be getting more expensive due to tariffs.A new report suggests that Apple is pursuing several different initiatives in an effort to mitigate the cost increases of tariffs imposed by the United States and other countries on its flagship products. That said, US consumers should expect to see the price go up in the near future on iPhones, Macs, and other hardware products.While Apple has spent years diversifying its production facilities away from China, almost all of the countries it has set up shop in such as Brazil, India, and Vietnam also face withering import tariffs. According to Bloomberg, the company is said to be pursuing a diverse range of strategies to soften the blow as much as possible. Continue Reading on AppleInsider | Discuss on our Forums
    0 Commentarios 0 Acciones 27 Views
  • VENTUREBEAT.COM
    A Minecraft Movie hits $110M global opening day and its headed higher
    The Minecraft Movie is off to a pretty amazing start, with $58 million coming in on the first day in the U.S. and $110 million globally.Read More
    0 Commentarios 0 Acciones 29 Views
  • WWW.THEVERGE.COM
    How to use your phone as a hotspot
    If youre taking your laptop away from the safe environs of your home or office desk and still want to stay online, youve got a couple of choices (assuming it doesnt have cellular connectivity built in): hunt around for a Wi-Fi network you can connect to or run a Wi-Fi hotspot from your phone.Running a hotspot from your phone comes with advantages and disadvantages. Its more secure than a public Wi-Fi network, as youre in charge, and you may well get better upload and download speeds, too though this will, of course, depend on the 4G and 5G coverage in your part of the world. On the downside, you may be limited in terms of your data allowance, and battery life on your phone will take a hit.RelatedHow to set up Wi-Fi calling on Android and iPhonesIf you want to take the mobile hotspot route, heres how to do it.Set up a hotspot on a PixelWith Android devices, as always, the exact steps vary depending on the manufacturer. These are the steps for using a Pixel device with Android 15:Open Settings on Android.Choose Network & Internet > Hotspot & tethering.You can toggle Wi-Fi hotspot from here to enable it, but if youve never used the hotspot before, tap on it to set your options.Youll see options to set the Wi-Fi name and password, which other devices need to connect. You can also set the hotspot to turn itself off automatically if its no longer being used and use Instant Hotspot to automatically sign in other devices on your account.Enable the Use Wi-Fi hotspot toggle switch.Note: if you use Androids Data Saver, which stops some apps from sending or receiving data in the background, youll have to turn it off in order to use the hotspot.Set up a hotspot on a Samsung GalaxyThe steps are slightly different for a Samsung Galaxy device loaded with One UI 7:Open Settings.Select Connections > Mobile Hotspot and Tethering.Again, you can toggle Mobile Hotspot on from here, but you may want to tap on it instead to check the network name and password. Theres also an Auto Hotspot feature so you can automatically share with other devices signed in to the same Samsung account and info on how much mobile data has already been shared.When youre ready, toggle Mobile Hotspot on.1/3Hotspot options on a Pixel phone. Screenshot: GoogleSet up a hotspot on an iPhoneIf youre using an iPhone with iOS 18:Open Settings in iOS.Tap Personal Hotspot.Turn on the Allow Others to Join toggle switch.Enable Maximize Compatibility if you want to use 2.4GHz Wi-Fi (rather than 5GHz) for the hotspot. Its slower, but it ensures compatibility with older devices.The name of the Wi-Fi hotspot will be the name of your iPhone (set in General > About > Name in Settings). The password will be shown on the Personal Hotspot screen, and you can tap on the entry to change it.Apple also offers a feature called Instant Hotspot. If the hotspot is enabled on your iPhone, you can connect to it with a tap from iPads, Macs, and other Apple devices using the same Apple account you wont need to enter the password.The cost of hotspotsIts now the norm for carriers to offer unlimited or very generous amounts of data in their plans, but most of the time, hotspot data will be handled separately. If youre going to be using your phone as a Wi-Fi hotspot, you need to check whats included in your plan.For example, Verizons $40 per month Unlimited Plus plan gives you 30GB of high-speed hotspot data per month, while you get the same amount with the $41 per month Unlimited Extra EL plan from AT&T. In both cases, once you go over that limit, youll be restricted to much slower speeds for the rest of the month.There is another option, which is to buy a dedicated hotspot device, like these sold by T-Mobile. Youll need to pay for an extra SIM with its own data plan to use with the device, so it can be an expensive option but youll usually get better performance than you would from a phone hotspot (especially if youre connecting multiple gadgets), and you dont have to worry about draining your phones battery).Something like the Netgear Nighthawk M6, which varies in cost from about $500 to $600, will deliver Wi-Fi 6 speeds to up to 32 devices when supplied with a SIM card. Definitely not cheap, but itll offer better range and speeds than your phone, and its good enough to be used as a backup home router if your broadband goes out.See More:
    0 Commentarios 0 Acciones 30 Views
  • WWW.MARKTECHPOST.COM
    Transformer Meets Diffusion: How the Transfusion Architecture Empowers GPT-4os Creativity
    OpenAIs GPT-4o represents a new milestone in multimodal AI: a single model capable of generating fluent text and high-quality images in the same output sequence. Unlike previous systems (e.g., ChatGPT) that had to invoke an external image generator like DALL-E, GPT-4o produces images natively as part of its response. This advance is powered by a novel Transfusion architecture described in 2024 by researchers at Meta AI, Waymo, and USC. Transfusion marries the Transformer models used in language generation with the Diffusion models used in image synthesis, allowing one large model to handle text and images seamlessly. In GPT-4o, the language model can decide on the fly to generate an image, insert it into the output, and then continue generating text in one coherent sequence.Lets look into a detailed, technical exploration of GPT-4os image generation capabilities through the lens of the Transfusion architecture. First, we review how Transfusion works: a single Transformer-based model can output discrete text tokens and continuous image content by incorporating diffusion generation internally. We then contrast this with prior approaches, specifically, the tool-based method where a language model calls an external image API and the discrete token method exemplified by Metas earlier Chameleon (CM3Leon) model. We dissect the Transfusion design: special Begin-of-Image (BOI) and End-of-Image (EOI) tokens that bracket image content, the generation of image patches which are later refined in diffusion style, and the conversion of these patches into a final image via learned decoding layers (linear projections, U-Net upsamplers, and a variational autoencoder). We also compare empirical performance: Transfusion-based models (like GPT-4o) significantly outperform discretization-based models (Chameleon) in image quality and efficiency and match state-of-the-art diffusion models on image benchmarks. Finally, we situate this work in the context of 20232025 research on unified multimodal generation, highlighting how Transfusion and similar efforts unify language and image generation in a single forward pass or shared tokenization framework.Prior Tool-Based Approach: Before architectures like GPT-4o, if one wanted a conversational agent to produce images, a common approach was a pipeline or tool-invocation strategy. For example, ChatGPT could be augmented with a prompt to call an image generator (such as DALLE 3) when the user requests an image. In this two-model setup, the language model itself does not truly generate the image; it merely produces a textual description or API call, which an external diffusion model renders into an image. While effective, this approach has clear limitations: the image generation is not tightly integrated with the language models knowledge and context.Discrete Token Early-Fusion: An alternative line of research made image generation endogenously part of the sequence modeling by treating images as sequences of discrete tokens. Pioneered by models like DALLE (2021), which used a VQ-VAE to encode images into codebook indices, this approach allows a single transformer to generate text and image tokens from one vocabulary. For instance, Parti (Google, 2022) and Metas Chameleon (2024) extend language modeling to image synthesis by quantizing images into tokens and training the model to predict those tokens like words. The key idea of Chameleon was the early fusion of modalities: images and text are converted into a common token space from the start.However, this discretization approach introduces an information bottleneck. Converting an image into a sequence of discrete tokens necessarily throws away some detail. The VQ-VAE codebook has a fixed size, so it may not capture subtle color gradients or fine textures present in the original image. Moreover, to retain as much fidelity as possible, the image must be broken into many tokens, often hundreds or more for a single image. This makes generation slow and training costly. Despite these efforts, there is an inherent trade-off: using a larger codebook or more tokens improves image quality but increases sequence length and computation, whereas using a smaller codebook speeds up generation but loses detail. Empirically, models like Chameleon, while innovative, lag behind dedicated diffusion models in image fidelity.The Transfusion Architecture: Merging Transformers with DiffusionTransfusion takes a hybrid approach, directly integrating a continuous diffusion-based image generator into the transformers sequence modeling framework. The core of Transfusion is a single transformer model (decoder-only) trained on a mix of text and images but with different objectives for each. Text tokens use the standard next-token prediction loss. Image tokens, continuous embeddings of image patches, use a diffusion loss, the same kind of denoising objective used to train models like Stable Diffusion, except it is implemented within the transformer.Unified Sequence with BOI/EOI Markers: In Transfusion (and GPT-4o), text and image data are concatenated into one sequence during training. Special tokens mark the boundaries between modalities. A Begin-of-Image (BOI) token indicates that subsequent elements in the sequence are image content, and an End-of-Image (EOI) token signals that the image content has ended. Everything outside of BOIEOI is treated as normal text; everything inside is treated as a continuous image representation. The same transformer processes all sequences. Within an images BOIEOI block, the attention is bidirectional among image patch elements. This means the transformer can treat an image as a two-dimensional entity while treating the image as a whole as one step in an autoregressive sequence.Image Patches as Continuous Tokens: Transfusion represents an image as a small set of continuous vectors called latent patches rather than discrete codebook tokens. The image is first encoded by a variational autoencoder (VAE) into a lower-dimensional latent space. The latent image is then divided into a grid of patches, & each patch is flattened into a vector. These patch vectors are what the transformer sees and predicts for image regions. Since they are continuous-valued, the model cannot use a softmax over a fixed vocabulary to generate an image patch. Instead, image generation is learned via diffusion: The model is trained to output denoised patches from noised patches.Lightweight modality-specific layers project these patch vectors into the transformers input space. Two design options were explored: a simple linear layer or a small U-Net style encoder that further downsamples local patch content. The U-Net downsampler can capture more complex spatial structures from a larger patch. In practice, Transfusion found that using U-Net up/down blocks allowed them to compress an entire image into as few as 16 latent patches with minimal performance loss. Fewer patches mean shorter sequences and faster generation. In the best configuration, a Transfusion model at 7B scale represented an image with 22 latent patch vectors on average.Denoising Diffusion Integration: Training the model on images uses a diffusion objective embedded in the sequence. For each image, the latent patches are noised with a random noise level, as in a standard diffusion model. These noisy patches are given to the transformer (preceded by BOI). The transformer must predict the denoised version. The loss on image tokens is the usual diffusion loss (L2 error), while the loss on text tokens is cross-entropy. The two losses are simply added for joint training. Thus, depending on its current processing, the model learns to continue text or refine an image.At inference time, the generation procedure mirrors training. GPT-4o generates tokens autoregressively. If it generates a normal text token, it continues as usual. But if it generates the special BOI token, it transitions to image generation. Upon producing BOI, the model appends a block of latent image tokens initialized with pure random noise to the sequence. These serve as placeholders for the image. The model then enters diffusion decoding, repeatedly passing the sequence through the transformer to progressively denoise the image. Text tokens in the context act as conditioning. Once the image patches are fully generated, the model emits an EOI token to mark the end of the image block.Decoding Patches into an Image: The final latent patch vectors are converted into an actual image. This is done by inverting the earlier encoding: first, the patch vectors are mapped back to latent image tiles using either a linear projection or U-Net up blocks. After this, the VAE decoder decodes the latent image into the final RGB pixel image. The result is typically high quality and coherent because the image was generated through a diffusion process in latent space.Transfusion vs. Prior Methods: Key Differences and AdvantagesNative Integration vs. External Calls: The most immediate advantage of Transfusion is that image generation is native to the models forward pass, not a separate tool. This means the model can fluidly blend text and imagery. Moreover, the language models knowledge and reasoning abilities directly inform the image creation. GPT-4o excels at rendering text in images and handling multiple objects, likely due to this tighter integration.Continuous Diffusion vs. Discrete Tokens: Transfusions continuous patch diffusion approach retains much more information and yields higher-fidelity outputs. The transformer cannot choose from a limited palette by eliminating the quantization bottleneck. Instead, it predicts continuous values, allowing subtle variations. In benchmarks, a 7.3B-parameter Transfusion model achieved an FID of 6.78 on MS-COCO, compared to an FID of 26.7 for a similarly sized Chameleon model. Transfusion also had a higher CLIP score (0.63 vs 0.39), indicating better image-text alignment.Efficiency and Scaling: Transfusion can compress an image into as few as 1620 latent patches. Chameleon might require hundreds of tokens. This means that the transfusion transformer takes fewer steps per image. Transfusion matched Chameleons performance using only ~22% of the compute. The model reached the same language perplexity using roughly half the compute as Chameleon.Image Generation Quality: Transfusion generates photorealistic images comparable to state-of-the-art diffusion models. On the GenEval benchmark for text-to-image generation, a 7B Transfusion model outperformed DALL-E 2 and even SDXL 1.0. GPT-4o renders legible text in images and handles many distinct objects in a scene.Flexibility and Multi-turn Multimodality: GPT-4o can handle bimodal interactions, not just text-to-image but image-to-text and mixed tasks. For example, it can show an image and then continue generating text about it or edit it with further instructions. Transfusion enables these capabilities naturally within the same architecture.Limitations: While Transfusion outperforms discrete approaches, it still inherits some limitations from diffusion models. Image output is slower due to multiple iterative steps. The transformer must perform double duty, increasing training complexity. However, careful masking and normalization enable training to billions of parameters without collapse.Before Transfusion, most efforts fell into tool-augmented models and token-fusion models. HuggingGPT and Visual ChatGPT allowed an LLM to call various APIs for tasks like image generation. Token-fusion approaches include DALLE, CogView, and Parti, which treat images as sequences of tokens. Chameleon trained on interleaved image-text sequences. Kosmos-1 and Kosmos-2 were multimodal transformers aimed at understanding rather than generation.Transfusion bridges the gap by keeping the single-model elegance of token fusion but using continuous latent and iterative refinement like diffusion. Googles Muse and DeepFloyd IF introduced variations but used multiple stages or frozen language encoders. Transfusion integrates all capabilities into one transformer. Other examples include Metas Make-A-Scene and Paint-by-Example, Stability AIs DeepFloyd IF, and HuggingFaces IDEFICS.In conclusion, the Transfusion architecture demonstrates that unifying text and image generation in one transformer is possible. GPT-4o with Transfusion generates images natively, guided by context and knowledge, and produces high-quality visuals interleaved with text. Compared to prior models like Chameleon, it offers better image quality, more efficient training, and deeper integration.SourcesAlso,feel free to follow us onTwitterand dont forget to join our85k+ ML SubReddit. Asif RazzaqWebsite| + postsBioAsif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.Asif Razzaqhttps://www.marktechpost.com/author/6flvq/Meta AI Just Released Llama 4 Scout and Llama 4 Maverick: The First Set of Llama 4 ModelsAsif Razzaqhttps://www.marktechpost.com/author/6flvq/NVIDIA AI Released AgentIQ: An Open-Source Library for Efficiently Connecting and Optimizing Teams of AI AgentsAsif Razzaqhttps://www.marktechpost.com/author/6flvq/A Code Implementation to Building a Context-Aware AI Assistant in Google Colab Using LangChain, LangGraph, Gemini Pro, and Model Context Protocol (MCP) Principles with Tool Integration SupportAsif Razzaqhttps://www.marktechpost.com/author/6flvq/Augment Code Released Augment SWE-bench Verified Agent: An Open-Source Agent Combining Claude Sonnet 3.7 and OpenAI O1 to Excel in Complex Software Engineering Tasks
    0 Commentarios 0 Acciones 27 Views