0 Comments
0 Shares
126 Views
Directory
Directory
-
Please log in to like, share and comment!
-
GAMERANT.COMHalo: Master Chief Collection - In Which Order To Play All GamesAsk any Xbox owner what their favorite game is, and theres a good chance theyll say Halo. This isnt surprising once youve experienced the magic of the classic Halo titles.0 Comments 0 Shares 127 Views
-
GAMERANT.COMAdapting This Stephen King Novel Could Hold Answers to Mike Flanagan's The Dark TowerIn the grand pantheon of Stephen King's works, many have ties to his magnum opus, The Dark Tower series. These novels have one or more connections to the story of Roland the Gunslinger and his quest for The Dark Tower, the grand structure that binds all reality together. The most prominent that fans are aware of include Stephen Kings The Stand, The Shining, Salems Lot, and more, but there is one book in particular that not only has ties to the main antagonist of the Dark Tower series, but a figure that could hold the key to Rolands mission altogether, Insomnia.0 Comments 0 Shares 132 Views
-
GAMEDEV.NETDecember LogSo this will be a mostly unedited post because I haven't released any entries in a while. I just wanted to wish everyone a merry Christmas and a great New Year first! I also have to share the unfortunate news that, Harvest Fall's development has been signific0 Comments 0 Shares 146 Views
-
WWW.TECHRADAR.COMThis is the first rugged convertible Chrome tablet ever launched, but you can only buy it in Japan: Dynabook Chromebook C70 comes with a docked pen as well as a proper keyboardDynabook's Chromebook C70 is a rugged 10.1-inch convertible tablet with MIL-grade durability.0 Comments 0 Shares 126 Views
-
WWW.YANKODESIGN.COMThe Zen Home Offers The Perks & Minimalism Of A Tiny Home In A Truly Tiny PackageCalled the Zen, this tiny home is designed by the Australian tiny house building company Havenn Tiny Houses. It is a highly versatile home that is ideal for glamping enthusiasts. It isnt intended for full-time living, but it serves as a multi-purpose single-level dwelling for short-term stays. It isnt exactly a home, but an extra space for a home, quite like an ADU. Zen Tiny functions as an excellent additional space for a separate amenity. This unique and compact tiny abode is called an Essential Luxury, as it offers all the perks of a full-time home and the minimalism of a tiny home but in a truly tiny package.Designer: Havenn Tiny HousesZen is half the size of a typical home on wheels, as it features a length of 14.2 feet, a width of 7.5 feet, and a height of 9.8 feet. It features a simple single-level layout that is ideal for multi-purpose flexibility. It can be used as a lovely self-contained bedroom that is surrounded by nature, making for the perfect glamping escapade. It can also be utilized as a home office, as it has more than enough room for two office setups. It is equipped with high ceilings, large windows, and a cute French door. These features help create a space that feels spacious and free-flowing.It would be great as a guest bedroom, as it has sufficient space for a queen-sized bed and essential storage. It is made using natural materials, and equipped with solar panels for a sustainable touch. Sustainably sourced and recycled materials were used to construct the home. The Zen tiny home is also equipped with energy-efficient insulation and windows, eco-friendly toilets, and solar panels. Water conservation is also achieved via greywater systems or rainwater collection, paired up with efficient fixtures. The home is designed to minimize environmental impact while offering a comfortable living space.The house is built using termite-and rust-proof frames on a galvanized steel base. The R-value 2.0 insulation provides comfort throughout the year, while the premium hybrid flooring adds some luxury. A large skylight has been placed above the main section of the home, allowing more natural light to stream in. The home also includes a fully-equipped kitchen and laundry appliances. The bathroom is also well-designed, and the toilet has an eco-friendly version for off-the-grid situations. The Zen tiny house is compact and minimal, yet well-equipped making for an excellent tiny home. It is priced at $18,750, so isnt very expensive either. It lets you connect with the outdoors while reducing the impact on the environment.The post The Zen Home Offers The Perks & Minimalism Of A Tiny Home In A Truly Tiny Package first appeared on Yanko Design.0 Comments 0 Shares 124 Views
-
0 Comments 0 Shares 192 Views
-
WWW.MARKTECHPOST.COMMeet CoMERA: An Advanced Tensor Compression Framework Redefining AI Model Training with Speed and PrecisionTraining large-scale AI models such as transformers and language models have become an indispensable yet highly demanding process in AI. With billions of parameters, these models offer groundbreaking capabilities but come at a steep cost in terms of computational power, memory, and energy consumption. For example, OpenAIs GPT-3 comprises 175 billion parameters and requires weeks of GPU training. Such massive requirements limit these technologies to organizations with substantial computational resources, exacerbating concerns over energy efficiency and environmental impact. Addressing these challenges has become critical to ensuring the broader accessibility and sustainability of AI advancements.The inefficiencies in training large models stem primarily from their reliance on dense matrices, which demand significant memory and computing power. The limited support for optimized low-precision or low-rank operations in modern GPUs further compounds these requirements. While some methods, such as matrix factorization and heuristic rank reduction, have been proposed to alleviate these issues, their real-world applicability is constrained. For instance, GaLore enables training on single-batch settings but suffers from impractical runtime overhead. Similarly, LTE, which adopts low-rank adapters, struggles with convergence on large-scale tasks. The lack of a method that simultaneously reduces memory usage, computational cost, and training time without compromising performance has created an urgent need for innovative solutions.Researchers from the University at Albany SUNY, the University of California at Santa Barbara, Amazon Alexa AI, and Meta introduced Computing-and Memory-Efficient training method via Rank-Adaptive tensor optimization (CoMERA), a novel framework that combines memory efficiency with computational speed through rank-adaptive tensor compression. Unlike traditional methods focusing solely on compression, CoMERA adopts a multi-objective optimization approach to balance compression ratio and model accuracy. It utilizes tensorized embeddings and advanced tensor-network contractions to optimize GPU utilization, reducing runtime overhead while maintaining robust performance. The framework also introduces CUDA Graph to minimize kernel-launching delays during GPU operations, a significant bottleneck in traditional tensor compression approaches.CoMERAs foundation is based on adaptive tensor representations, which allow model layers to adjust their ranks dynamically based on resource constraints. By modifying tensor ranks, the framework achieves compression without compromising the integrity of neural network operations. This dynamic optimization is achieved through a two-stage training process:An early stage focused on stable convergenceA late stage that fine-tunes ranks to meet specific compression targetsIn a six-encoder transformer model, CoMERA achieved compression ratios ranging from 43x in its early stage to an impressive 361x in its late-stage optimizations. Also, it reduced memory consumption by 9x compared to GaLore, with 2-3x faster training per epoch.When applied to transformer models trained on the MNLI dataset, CoMERA reduced model sizes from 256 MB to as little as 3.2 MB while preserving accuracy. In large-scale recommendation systems like DLRM, CoMERA compressed models by 99x and achieved a 7x reduction in peak memory usage. The framework also excelled in pre-training CodeBERT, a domain-specific large language model, where it gained a 4.23x overall compression ratio and demonstrated a 2x speedup during certain training phases. These results underscore its ability to handle diverse tasks and architectures, extending its applicability across domains.The key takeaways from this research are as follows:CoMERA achieved compression ratios of up to 361x for specific layers and 99x for full models, drastically reducing storage and memory requirements.The framework delivered 2-3x faster training times per epoch for transformers and recommendation systems, saving computational resources and time.Using tensorized representations and CUDA Graph, CoMERA reduced peak memory consumption by 7x, enabling training on smaller GPUs.CoMERAs approach supports diverse architectures, including transformers and large language models, while maintaining or improving accuracy.By lowering the energy and resource demands of training, CoMERA contributes to more sustainable AI practices and makes cutting-edge models accessible to a broader audience.In conclusion, CoMERA addresses some of the most significant barriers to AI scalability and accessibility by enabling faster, memory-efficient training. Its adaptive optimization capabilities and compatibility with modern hardware make it a compelling choice for organizations seeking to train large models without incurring prohibitive costs. This studys results pave the way for further exploration of tensor-based optimizations in domains like distributed computing and resource-constrained edge devices.Check out the Paper. All credit for this research goes to the researchers of this project. Also,dont forget to follow us onTwitter and join ourTelegram Channel andLinkedIn Group. Dont Forget to join our60k+ ML SubReddit. Asif RazzaqAsif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences. [Download] Evaluation of Large Language Model Vulnerabilities Report (Promoted)0 Comments 0 Shares 127 Views
-
WWW.MARKTECHPOST.COMCoordTok: A Scalable Video Tokenizer that Learns a Mapping from Co-ordinate-based Representations to the Corresponding Patches of Input VideosBreaking down videos into smaller, meaningful parts for vision models remains challenging, particularly for long videos. Vision models rely on these smaller parts, called tokens, to process and understand video data, but creating these tokens efficiently is difficult. While recent tools achieve better video compression than older methods, they struggle to handle large video datasets effectively. A key issue is their inability to fully utilize temporal coherence, the natural pattern where video frames are often similar over short periods, which video codecs use for efficient compression. These tools are also computationally expensive to train and are limited to short clips, making them not very effective in capturing patterns and processing longer videos.Current video tokenization methods have high computational costs and struggle to handle long video sequences efficiently. Early approaches used image tokenizers to compress videos frame by frame but ignored the natural continuity between frames, reducing their effectiveness. Later methods introduced spatiotemporal layers, reduced redundancy, and used adaptive encoding, but they still required rebuilding entire video frames during training, which limited them to short clips. Video generation models like autoregressive methods, masked generative transformers, and diffusion models are also limited to short sequences.To solve this, researchers from KAIST and UC Berkeley proposed CoordTok, which learns a mapping from coordinate-based representations to the corresponding patches of input videos. Motivated by recent advances in 3D generative models, CoordTok encodes a video into factorized triplane representations and reconstructs patches corresponding to randomly sampled (x, y, t) coordinates. This approach allows large tokenizer models to be trained directly on long videos without requiring excessive resources. The video is divided into space-time patches and processed using transformer layers, with the decoder mapping sampled (x, y, t) coordinates to corresponding pixels. This reduces both memory and computational costs while preserving video quality.Based on this, researchers updated CoordTok to efficiently process a video by introducing a hierarchical architecture that grasped local and global features from the video. This architecture represented a factorized triplane to process patches of space and time, making long-duration video processing easier without excessively using computational resources. This approach greatly reduced the memory and computation requirements and maintained high video quality.Researchers improved the performance by adding a hierarchical structure that captured the local and global features of videos. This structure allowed the model to process space-time patches more efficiently using transformer layers, which helped generate factorized triplane representations. As a result, CoordTok handled longer videos without demanding excessive computational resources. For example, CoordTok encoded a 128-frame video with 128128 resolution into 1280 tokens, while baselines required 6144 or 8192 tokens to achieve similar reconstruction quality. The models reconstruction quality was further improved by fine-tuning with both 2 loss and LPIPS loss, enhancing the accuracy of the reconstructed frames. This combination of strategies reduced memory usage by up to 50% and computational costs while maintaining high-quality video reconstruction, with models like CoordTok-L achieving a PSNR of 26.9.In conclusion, the proposed framework by researchers, CoordTok, proves to be an efficient video tokenizer that uses coordinate-based representations to reduce computational costs and memory requirements while encoding long videos.It allows memory-efficient training for video generation models, making handling long videos with fewer tokens possible. However, it is not strong enough for dynamic videos and suggests further potential improvements, such as using multiple content planes or adaptive methods. This work can serve as a starting point for future research on scalable video tokenizers and generation, which can be beneficial for comprehending and generating long videos.Check out the Paper and Project. All credit for this research goes to the researchers of this project. Also,dont forget to follow us onTwitter and join ourTelegram Channel andLinkedIn Group. Dont Forget to join our60k+ ML SubReddit. Divyesh Vitthal Jawkhede+ postsDivyesh is a consulting intern at Marktechpost. He is pursuing a BTech in Agricultural and Food Engineering from the Indian Institute of Technology, Kharagpur. He is a Data Science and Machine learning enthusiast who wants to integrate these leading technologies into the agricultural domain and solve challenges. [Download] Evaluation of Large Language Model Vulnerabilities Report (Promoted)0 Comments 0 Shares 117 Views
-
WWW.CNET.COMBest Internet Providers in Sebring, FloridaLooking for the best internet provider in Sebring? CNETs broadband experts will guide you to the right plan for your needs.0 Comments 0 Shares 110 Views