0 Comments
0 Shares
27 Views
Directory
Directory
-
Please log in to like, share and comment!
-
WWW.TECHRADAR.COMSamsung introduces 2025 mini-LED TV lineup with A new era of Samsung AISamsung has announced its 2025 Neo QLED mini-LED TV lineup at CES.0 Comments 0 Shares 28 Views
-
WWW.CNBC.COMHow AI regulation could shake out in 2025This year will be a year of change for the U.S. political landscape and that comes with big implications for the direction of travel for global AI regulation.0 Comments 0 Shares 31 Views
-
-
WWW.FASTCOMPANY.COMLOrals new cosmetics counter device can predict your skins needsChoosing skincare products often comes down to a matter of trial and error.A tabletop device unveiled by LOral, called the LOral Cell BioPrint, is intended to change that. It will offer visitors to LOral-stocked beauty counters the ability to analyze their skin for signs of biological aging, responsiveness to particular active ingredients, and likelihood of conditions like dark spots or enlarged pores. As a result, the company claims, people will be able to choose the skincare products that are most effective for them.We have very good imaging systems that can tell you about what is on the surface of your skin, but what is on the surface of your skin has already happened, says Guive Balooch, global managing director of augmented beauty and open innovation at LOral Groupe. What matters for the future and for people to really get the right product is to understand their individual biology, and that has been something very challenging to do at scale.The new technology will let shoppers take a quick sample of their skin with a painless sticky tape and deposit it onto a test cartridge along with a drop of test solution, then insert it into the device for an analysis of various protein-based biomarkers. Within a few minutes, the user and an adviser at the beauty counter will be able to see a report on an in-store tablet they can use to guide their skincare product considerations, knowing what products are most likely to be effective and what skin issues may pop up in the future. That way, Balooch says, shoppers can shape their skincare routines to address issues like dark spots or wrinkles well before they actually crop up.[Photo: LOral]The goal is to let people be able to really, for the first time, get to know what they can do to correct the future of their skin health, says Balooch.LOrals experts have spent decades identifying biomarkers that can predict skin health, but it wasnt previously possible to do a real-time analysis or spot them in a sample without a traditional laboratory. The company worked with South Korean company NanoEntek, which has experience creating miniaturized biological analysis tools, to develop the technology, which Balooch says compares favorably to traditional lab-grade equipment. The analysis has been tested on roughly 1,000 different skin types and tones from across the world to ensure its helpful for a wide range of customers.Balooch says LOral intends to deploy the devices to thousands of beauty counters across the globe beginning in about a year. The exact business model, including any cost to the consumer, is yet to be determined. The company plans to offer training to help beauty advisers use the new equipment, though Balooch says its designed to be easy and quick to use. And as LOrals researchers continue to find new biomarkers that can inform skincare decisions, they can be added to the Cell BioPrint analysis, Balooch says.Advisers will also have a role in helping customers choose the right product based on their test results, since theyll also likely consider factors like how long a particular product lasts, thickness, and other factors beyond just the active ingredients, he says.Were making sure that the product recommendation is equally as strong as the assessment, Balooch says. Were working very hard on that so people get the right product.0 Comments 0 Shares 29 Views
-
WWW.CREATIVEBLOQ.COMCould a sleek rebrand make this the trendiest photo app for social media?Tezza is the photo-editing app that thinks it's a fashion brand.0 Comments 0 Shares 31 Views
-
VENTUREBEAT.COMLOreal Cell BioPrint analyzes your skin in five minutesLOral Groupe announced at CES 2025 theLOral Cell BioPrint, a hardware device that provides customized skin analysis in just five minutes.Read More0 Comments 0 Shares 28 Views
-
WWW.MARKTECHPOST.COMVITA-1.5: A Multimodal Large Language Model that Integrates Vision, Language, and Speech Through a Carefully Designed Three-Stage Training MethodologyThe development of multimodal large language models (MLLMs) has brought new opportunities in artificial intelligence. However, significant challenges persist in integrating visual, linguistic, and speech modalities. While many MLLMs perform well with vision and text, incorporating speech remains a hurdle. Speech, a natural medium for human interaction, plays an essential role in dialogue systems, yet the differences between modalitiesspatial versus temporal data representationscreate conflicts during training. Traditional systems relying on separate automatic speech recognition (ASR) and text-to-speech (TTS) modules are often slow and impractical for real-time applications.Researchers from NJU, Tencent Youtu Lab, XMU, and CASIA have introduced VITA-1.5, a multimodal large language model that integrates vision, language, and speech through a carefully designed three-stage training methodology. Unlike its predecessor, VITA-1.0, which depended on external TTS modules, VITA-1.5 employs an end-to-end framework, reducing latency and streamlining interaction. The model incorporates vision and speech encoders along with a speech decoder, enabling near real-time interactions. Through progressive multimodal training, it addresses conflicts between modalities while maintaining performance. The researchers have also made the training and inference code publicly available, fostering innovation in the field.Technical Details and BenefitsVITA-1.5 is built to balance efficiency and capability. It uses vision and audio encoders, employing dynamic patching for image inputs and downsampling techniques for audio. The speech decoder combines non-autoregressive (NAR) and autoregressive (AR) methods to ensure fluent and high-quality speech generation. The training process is divided into three stages:Vision-Language Training: This stage focuses on vision alignment and understanding, using descriptive captions and visual question answering (QA) tasks to establish a connection between visual and linguistic modalities.Audio Input Tuning: The audio encoder is aligned with the language model using speech-transcription data, enabling effective audio input processing.Audio Output Tuning: The speech decoder is trained with text-speech paired data, enabling coherent speech outputs and seamless speech-to-speech interactions.These strategies effectively address modality conflicts, allowing VITA-1.5 to handle image, video, and speech data seamlessly. The integrated approach enhances its real-time usability, eliminating common bottlenecks in traditional systems.Results and InsightsEvaluations of VITA-1.5 on various benchmarks demonstrate its robust capabilities. The model performs competitively in image and video understanding tasks, achieving results comparable to leading open-source models. For example, on benchmarks like MMBench and MMStar, VITA-1.5s vision-language capabilities are on par with proprietary models like GPT-4V. Additionally, it excels in speech tasks, achieving low character error rates (CER) in Mandarin and word error rates (WER) in English. Importantly, the inclusion of audio processing does not compromise its visual reasoning abilities. The models consistent performance across modalities highlights its potential for practical applications.ConclusionVITA-1.5 represents a thoughtful approach to resolving the challenges of multimodal integration. By addressing conflicts between vision, language, and speech modalities, it offers a coherent and efficient solution for real-time interactions. Its open-source availability ensures that researchers and developers can build upon its foundation, advancing the field of multimodal AI. VITA-1.5 not only enhances current capabilities but also points toward a more integrated and interactive future for AI systems.Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also,dont forget to follow us onTwitter and join ourTelegram Channel andLinkedIn Group. Dont Forget to join our60k+ ML SubReddit. FREE UPCOMING AI WEBINAR (JAN 15, 2025): Boost LLM Accuracy with Synthetic Data and Evaluation IntelligenceJoin this webinar to gain actionable insights into boosting LLM model performance and accuracy while safeguarding data privacy.The post VITA-1.5: A Multimodal Large Language Model that Integrates Vision, Language, and Speech Through a Carefully Designed Three-Stage Training Methodology appeared first on MarkTechPost.0 Comments 0 Shares 29 Views
-
WWW.MARKTECHPOST.COMAutoGraph: An Automatic Graph Construction Framework based on LLMs for RecommendationEnhancing user experiences and boosting retention using recommendation systems is an effective and ever-evolving strategy used by many industries, such as e-commerce, streaming services, social media, etc. These systems must analyze complex relationships between users, items, and contextual factors to suggest precisely what the user might want. However, the existing recommendation systems are static, relying on substantial historical data to build connections effectively. In cold start scenarios, which are heavily prevalent, mapping the relationships becomes impossible, weakening these systems even further. Researchers from the Shanghai Jiao Tong University and Huawei Noahs Ark Lab have introduced AutoGraph to address these issues. This framework automatically builds graphs incorporating dynamic adjustments and leverages LLMs for better contextual understanding.Commonly, graph-based recommendation systems are employed. Current systems, however, require people to set the features manually and their connections in a graph, consuming much time. Also, rules are set beforehand, limiting how these graphs could adapt. Incorporating unstructured data, which potentially has rich semantic information about user preferences, is also a significant issue. Therefore, there is a need for a new method that can resolve the data sparsity issues and the failure to capture nuanced relationships and adjust to user preferences in real-time.AutoGraph is an innovative framework to enhance recommendation systems leveraging Large Language Models (LLMs) and Knowledge Graphs (KGs). The methodology of AutoGraph is based on these features:Utilization of Pre-trained LLMs: The framework leverages pre-trained LLMs to analyze user input. It can draw relationships based on the analysis of natural language, even those that are apparently hidden.Knowledge Graph Construction: After the relationship extraction, LLMs generate graphs, which can be seen as structured representations of user preferences. Algorithms optimize such graphs to remove less relevant connections in an attempt to maximize the quality of the graph in its entirety.Integration with Graph Neural Networks (GNNs): The final step of the proposed method is to integrate the created knowledge graph with regular Graph Neural Networks. GNNs can provide more accurate recommendations by using both node features and graph structure, and they are sensitive to personal preferences and more significant trends among users.To evaluate the proposed frameworks efficacy, authors benchmarked against traditional recommendation techniques using e-commerce and streaming services datasets. There was a significant gain in the precision of recommendations, which shows that the framework is competent enough to give relevant recommendations. The proposed method had improved scalability for dealing with large datasets. The framework demonstrated reduced computational requirements compared to traditional graph construction approaches. Process automation, along with the use of advanced algorithms, helped in lowering resource usage without compromising the quality of the results.The Autograph framework represents a significant leap forward in recommendation systems. Automating graph construction with LLMs addresses long-standing scalability, adaptability, and contextual awareness challenges. The frameworks success demonstrates the transformative potential of integrating LLMs into graph-based systems, setting a new benchmark for future research and applications in personalized recommendations. AutoGraph opens new avenues for personalized user experiences in diverse domains by automating the construction of dynamic, context-aware recommendation graphs. This innovation highlights the growing role of LLMs in addressing real-world challenges, revolutionizing how we approach recommendation systems.Check out the Paper. All credit for this research goes to the researchers of this project. Also,dont forget to follow us onTwitter and join ourTelegram Channel andLinkedIn Group. Dont Forget to join our60k+ ML SubReddit. Afeerah Naseem+ postsAfeerah Naseem is a consulting intern at Marktechpost. She is pursuing her B.tech from the Indian Institute of Technology(IIT), Kharagpur. She is passionate about Data Science and fascinated by the role of artificial intelligence in solving real-world problems. She loves discovering new technologies and exploring how they can make everyday tasks easier and more efficient. Follow us on X (Twitter) to get regular AI Research and Dev Updates here...0 Comments 0 Shares 29 Views
-
WWW.IGN.COMCES 2025: Hyperkin Unveils DualSense-Style Xbox Controller Called The CompetitorAs part of CES 2025, peripheral manufacturer Hyperkin revealed a new version of The Competitor, which is its wired pro-style gamepad for Xbox consoles and PC. Its geared towards competitive-minded players (hence, the name) and it shares similarities to the PlayStation 5s DualSense, which we consider one of the best controllers for PC as well.PlayLike the previous model of The Competitor, a major design shift from traditional Xbox controllers is the stick placement the symmetrical positioning of the analog sticks mimics what you find on PlayStation controllers. Whether or not thats a good thing will boil down to preference, but a notable improvement over the stock Xbox gamepad is that the sticks are Hall Effect as opposed to regular analog. Hall Effect parts are magnetic and create a consistent resistance, but more importantly, they have better durability and aren't prone to stick drift (which can happen on the Switch Joy-Con and PlayStation 5 DualSense). The triggers are also Hall Effect to help with precision when applying pressure. These are features found on the previous iteration of The Competitor, so the big change is in the ergonomics which draws from the DualSense, especially in the white-black color scheme.Hyperkin Competitor - PhotosThe directional pad clearly mimics its PlayStation inspiration with each direction being separated and the buttons pointed inward. Both the triggers and the bumpers are shaped accordingly and there's a mic mute button at the lower-center as well. It also comes with two programmable back buttons, which is a major feature of third-party competitive controllers and high-end first-party offerings like the DualSense Edge and Xbox Elite Series 2.Hyperkin has historically been known for its catalog of retro gaming accessories and systems, but has also been making peripherals for current platforms. When it comes to Xbox, Hyperkin made a splash by bringing back The Duke (the original Xbox controller) and The DuchesS (the Xbox S controller) for modern systems. Currently, Hyperkin has yet to reveal a release date or price for the updated version of The Competitor.We'll be getting our hands on this new model of The Competitor from Hyperkin at CES 2025, so stay tuned for our impressions. If you're looking to upgrade your gamepad, be sure to check out our current roundup of the best controllers for Xbox, PC, and PS5. For all the important gaming news at the biggest tech convention of the year, be sure to check out our roundup of everything you need to know about CES 2025.0 Comments 0 Shares 30 Views