• O2 unleashes AI grandma on scammers
    www.computerworld.com
    Research by British telecommunications provider O2 has found that seven in ten Britons (71 percent) would like to take revenge on scammers who have tried to trick them or their loved ones. At the same time, however, one in two people does not want to waste their time on it.AI grandma against telephone scammersO2 now wants to remedy this with an artificial intelligence called Daisy. As the head of fraud prevention, its the job of this state-of-the-art AI granny to keep scammers away from real people for as long as possible with human-like chatter. To activate Daisy, O2 customers simply have to forward a suspicious call to the number 7726.Daisy combines different AI models that work together to first listen to the caller and convert their voice to text. It then generates responses appropriate to the characters personality via a custom single-layer large language model. These are then fed back via a custom text-to-speech model to generate a natural language response. This happens in real-time, allowing the tool to have a human-like conversation with a caller.Although human-like is a strong understatement: Daisy was trained with the help of Jim Browning, one of the most famous scambaiters on YouTube. With the persona of a lonely and seemingly somewhat bewildered older lady, she tricks the fraudsters into believing that they have found a perfect target, while in reality she beats them with their own weapons.
    0 Comments ·0 Shares ·110 Views
  • How this grassroots effort could make AI voices more diverse
    www.technologyreview.com
    We are on the cusp of a voice AI boom, with tech companies such as Apple and OpenAI rolling out the next generation of artificial-intelligence-powered assistants. But the default voices for these assistants are often white AmericanBritish, if youre luckyand most definitely speak English. They represent only a tiny proportion of the many dialects and accents in the English language, which spans many regions and cultures. And if youre one of the billions of people who dont speak English, bad luck: These tools dont sound nearly as good in other languages.This is because the data that has gone into training these models is limited. In AI research, most data used to train models is extracted from the English-language internet, which reflects Anglo-American culture. But there is a massive grassroots effort underway to change this status quo and bring more transparency and diversity to what AI sounds like: Mozillas Common Voice initiative.The data set Common Voice has created over the past seven years is one of the most useful resources for people wanting to build voice AI. It has seen a massive spike in downloads, partly thanks to the current AI boom; it recently hit the 5 million mark, up from 38,500 in 2020. Creating this data set has not been easy, mainly because the data collection relies on an army of volunteers. Their numbers have also jumped, from just under 500,000 in 2020 to over 900,000 in 2024. But by giving its data away, some members of this community argue, Mozilla is encouraging volunteers to effectively do free labor for Big Tech.Since 2017, volunteers for the Common Voice project have collected a total of 31,000 hours of voice data in around 180 languages as diverse as Russian, Catalan, and Marathi. If youve used a service that uses audio AI, its likely been trained at least partly on Common Voice.Mozillas cause is a noble one. As AI is integrated increasingly into our lives and the ways we communicate, it becomes more important that the tools we interact with sound like us. The technology could break down communication barriers and help convey information in a compelling way to, for example, people who cant read. But instead, an intense focus on English risks entrenching a new colonial world order and wiping out languages entirely.It would be such an own goal if, rather than finally creating truly multimodal, multilingual, high-performance translation models and making a more multilingual world, we actually ended up forcing everybody to operate in, like, English or French, says EM Lewis-Jong, a director for Common Voice.Common Voice is open source, which means anyone can see what has gone into the data set, and users can do whatever they want with it for free. This kind of transparency is unusual in AI data governance. Most large audio data sets simply arent publicly available, and many consist of data that has been scraped from sites like YouTube, according to research conducted by a team from the University of Washington, and Carnegie Mellon andNorthwestern universities.The vast majority of language data is collected by volunteers such as Blent zden, a researcher from Turkey. Since 2020, he has been not only donating his voice but also raising awareness around the project to get more people to donate. He recently spent two full-time months correcting data and checking for typos in Turkish. For him, improving AI models is not the only motivation to do this work.Im doing it to preserve cultures, especially low-resource [languages], zden says. He tells me he has recently started collecting samples of Turkeys smaller languages, such as Circassian and Zaza.However, as I dug into the data set, I noticed that the coverage of languages and accents is very uneven. There are only 22 hours of Finnish voices from 231 people. In comparison, the data set contains 3,554 hours of English from 94,665 speakers. Some languages, such as Korean and Punjabi, are even less well represented. Even though they have tens of millions of speakers, they account for only a couple of hours of recorded data.This imbalance has emerged because data collection efforts are started from the bottom up by language communities themselves, says Lewis-Jong.Were trying to give communities what they need to create their own AI training data sets. We have a particular focus on doing this for language communities where there isnt any data, or where maybe larger tech organizations might not be that interested in creating those data sets, Lewis-Jong says. They hope that with the help of volunteers and various bits of grant funding, the Common Voice data set will have close to 200 languages by the end of the year.Common Voices permissive license means that many companies rely on itfor example, the Swedish startup Mabel AI, which builds translation tools for health-care providers. One of the first languages the company used was Ukrainian; it built a translation tool to help Ukrainian refugees interact with Swedish social services, says Karolina Sjberg, Mabel AIs founder and CEO. The team has since expanded to other languages, such as Arabic and Russian.The problem with a lot of other audio data is that it consists of people reading from books or texts. The result is very different from how people really speak, especially when they are distressed or in pain, Sjberg says. Because anyone can submit sentences to Common Voice for others to read aloud, Mozillas data set also includes sentences that are more colloquial and feel more natural, she says.Not that it is perfectly representative. The Mabel AI team soon found out that most voice data in the languages it needed was donated by younger men, which is fairly typical for the data set.The refugees that we intended to use the app with were really anything but younger men, Sjberg says. So that meant that the voice data that we needed did not quite match the voice data that we had. The team started collecting its own voice data from Ukrainian women, as well as from elderly people.Unlike other data sets, Common Voice asks participants to share their gender and details about their accent. Making sure different genders are represented is important to fight bias in AI models, says Rebecca Ryakitimbo, a Common Voice fellow who created the projects gender action plan. More diversity leads not only to better representation but also to better models. Systems that are trained on narrow and homogenous data tend to spew stereotyped and harmful results.We dont want a case where we have a chatbot that is named after a woman but does not give the same response to a woman as it would a man, she says.Ryakitimbo has collected voice data in Kiswahili in Tanzania, Kenya, and the Democratic Republic of Congo. She tells me she wanted to collect voices from a socioeconomically diverse set of Kiswahili speakers and has reached out to women young and old living in rural areas, who might not always be literate or even have access to devices.This kind of data collection is challenging. The importance of collecting AI voice data can feel abstract to many people, especially if they arent familiar with the technologies. Ryakitimbo and volunteers would approach women in settings where they felt safe to begin with, such as presentations on menstrual hygiene, and explain how the technology could, for example, help disseminate information about menstruation. For women who did not know how to read, the team read out sentences that they would repeat for the recording.The Common Voice project is bolstered by the belief that languages form a really important part of identity. We think its not just about language, but about transmitting culture and heritage and treasuring peoples particular cultural context, says Lewis-Jong. There are all kinds of idioms and cultural catchphrases that just dont translate, they add.Common Voice is the only audio data set where English doesnt dominate, says Willie Agnew, a researcher at Carnegie Mellon University who has studied audio data sets. Im very impressed with how well theyve done that and how well theyve made this data set that is actually pretty diverse, Agnew says. It feels like theyre way far ahead of almost all the other projects we looked at.I spent some time verifying the recordings of other Finnish speakers on the Common Voice platform. As their voices echoed in my study, I felt surprisingly touched. We had all gathered around the same cause: making AI data more inclusive, and making sure our culture and language was properly represented in the next generation of AI tools.But I had some big questions about what would happen to my voice if I donated it. Once it was in the data set, I would have no control about how it might be used afterwards. The tech sector isnt exactly known for giving people proper credit, and the data is available for anyones use.As much as we want it to benefit the local communities, theres a possibility that also Big Tech could make use of the same data and build something that then comes out as the commercial product, says Ryakitimbo. Though Mozilla does not share who has downloaded Common Voice, Lewis-Jong tells me Meta and Nvidia have said that they have used it.Open access to this hard-won and rare language data is not something all minority groups want, says Harry H. Jiang, a researcher at Carnegie Mellon University, who was part of the team doing audit research. For example, Indigenous groups have raised concerns.Extractivism is something that Mozilla has been thinking about a lot over the past 18 months, says Lewis-Jong. Later this year the project will work with communities to pilot alternative licenses including Nwulite Obodo Open Data License, which was created by researchers at the University of Pretoria for sharing African data sets more equitably. For example, people who want to download the data might be asked to write a request with details on how they plan to use it, and they might be allowed to license it only for certain products or for a limited time. Users might also be asked to contribute to community projects that support poverty reduction, says Lewis-Jong.Lewis-Jong says the pilot is a learning exercise to explore whether people will want data with alternative licenses, and whether they are sustainable for communities managing them. The hope is that it could lead to something resembling open source 2.0.In the end, I decided to donate my voice. I received a list of phrases to say, sat in front of my computer, and hit Record. One day, I hope, my effort will help a company or researcher build voice AI that sounds less generic, and more like me.This story has been updated.
    0 Comments ·0 Shares ·160 Views
  • Google DeepMind has a new way to look inside an AIs mind
    www.technologyreview.com
    AI has led to breakthroughs in drug discovery and robotics and is in the process of entirely revolutionizing how we interact with machines and the web. The only problem is we dont know exactly how it works, or why it works so well. We have a fair idea, but the details are too complex to unpick. Thats a problem: It could lead us to deploy an AI system in a highly sensitive field like medicine without understanding that it could have critical flaws embedded in its workings.A team at Google DeepMind that studies something called mechanistic interpretability has been working on new ways to let us peer under the hood. At the end of July, it released Gemma Scope, a tool to help researchers understand what is happening when AI is generating an output. The hope is that if we have a better understanding of what is happening inside an AI model, well be able to control its outputs more effectively, leading to better AI systems in the future.I want to be able to look inside a model and see if its being deceptive, says Neel Nanda, who runs the mechanistic interpretability team at Google DeepMind. It seems like being able to read a models mind should help.Mechanistic interpretability, also known as mech interp, is a new research field that aims to understand how neural networks actually work. At the moment, very basically, we put inputs into a model in the form of a lot of data, and then we get a bunch of model weights at the end of training. These are the parameters that determine how a model makes decisions. We have some idea of whats happening between the inputs and the model weights: Essentially, the AI is finding patterns in the data and making conclusions from those patterns, but these patterns can be incredibly complex and often very hard for humans to interpret.Its like a teacher reviewing the answers to a complex math problem on a test. The studentthe AI, in this casewrote down the correct answer, but the work looks like a bunch of squiggly lines. This example assumes the AI is always getting the correct answer, but thats not always true; the AI student may have found an irrelevant pattern that its assuming is valid. For example, some current AI systems will give you the result that 9.11 is bigger than 9.8. Different methods developed in the field of mechanistic interpretability are beginning to shed a little bit of light on what may be happening, essentially making sense of the squiggly lines.A key goal of mechanistic interpretability is trying to reverse-engineer the algorithms inside these systems, says Nanda. We give the model a prompt, like Write a poem, and then it writes some rhyming lines. What is the algorithm by which it did this? Wed love to understand it.To find featuresor categories of data that represent a larger conceptin its AI model, Gemma, DeepMind ran a tool known as a sparse autoencoder on each of its layers. You can think of a sparse autoencoder as a microscope that zooms in on those layers and lets you look at their details. For example, if you prompt Gemma about a chihuahua, it will trigger the dogs feature, lighting up what the model knows about dogs. The reason it is considered sparse is that its limiting the number of neurons used, basically pushing for a more efficient and generalized representation of the data.The tricky part of sparse autoencoders is deciding how granular you want to get. Think again about the microscope. You can magnify something to an extreme degree, but it may make what youre looking at impossible for a human to interpret. But if you zoom too far out, you may be limiting what interesting things you can see and discover.DeepMinds solution was to run sparse autoencoders of different sizes, varying the number of features they want the autoencoder to find. The goal was not for DeepMinds researchers to thoroughly analyze the results on their own. Gemma and the autoencoders are open-source, so this project was aimed more at spurring interested researchers to look at what the sparse autoencoders found and hopefully make new insights into the models internal logic. Since DeepMind ran autoencoders on each layer of their model, a researcher could map the progression from input to output to a degree we havent seen before.This is really exciting for interpretability researchers, says Josh Batson, a researcher at Anthropic. If you have this model that youve open-sourced for people to study, it means that a bunch of interpretability research can now be done on the back of those sparse autoencoders. It lowers the barrier to entry to people learning from these methods.Neuronpedia, a platform for mechanistic interpretability, partnered with DeepMind in July to build a demo of Gemma Scope that you can play around with right now. In the demo, you can test out different prompts and see how the model breaks up your prompt and what activations your prompt lights up. You can also mess around with the model. For example, if you turn the feature about dogs way up and then ask the model a question about US presidents, Gemma will find some way to weave in random babble about dogs, or the model may just start barking at you.One interesting thing about sparse autoencoders is that they are unsupervised, meaning they find features on their own. That leads to surprising discoveries about how the models break down human concepts. My personal favorite feature is the cringe feature, says Joseph Bloom, science lead at Neuronpedia. It seems to appear in negative criticism of text and movies. Its just a great example of tracking things that are so human on some level.You can search for concepts on Neuronpedia and it will highlight what features are being activated on specific tokens, or words, and how strongly each one is activated. If you read the text and you see whats highlighted in green, thats when the model thinks the cringe concept is most relevant. The most active example for cringe is somebody preaching at someone else, says Bloom.Some features are proving easier to track than others. One of the most important features that you would want to find for a model is deception, says Johnny Lin, founder of Neuronpedia. Its not super easy to find: Oh, theres the feature that fires when its lying to us. From what Ive seen, it hasnt been the case that we can find deception and ban it.DeepMinds research is similar to what another AI company, Anthropic, did back in May with Golden Gate Claude. It used sparse autoencoders to find the parts of Claude, their model, that lit up when discussing the Golden Gate Bridge in San Francisco. It then amplified the activations related to the bridge to the point where Claude literally identified not as Claude, an AI model, but as the physical Golden Gate Bridge and would respond to prompts as the bridge.Although it may just seem quirky, mechanistic interpretability research may prove incredibly useful. As a tool for understanding how the model generalizes and what level of abstraction its working at, these features are really helpful, says Batson.For example, a team lead by Samuel Marks, now at Anthropic, used sparse autoencoders to find features that showed a particular model was associating certain professions with a specific gender. They then turned off these gender features to reduce bias in the model. This experiment was done on a very small model, so its unclear if the work will apply to a much larger model.Mechanistic interpretability research can also give us insights into why AI makes errors. In the case of the assertion that 9.11 is larger than 9.8, researchers from Transluce saw that the question was triggering the parts of an AI model related to Bible verses and September 11. The researchers concluded the AI could be interpreting the numbers as dates, asserting the later date, 9/11, as greater than 9/8. And in a lot of books like religious texts, section 9.11 comes after section 9.8, which may be why the AI thinks of it as greater. Once they knew why the AI made this error, the researchers tuned down the AIs activations on Bible verses and September 11, which led to the model giving the correct answer when prompted again on whether 9.11 is larger than 9.8.There are also other potential applications. Currently, a system-level prompt is built into LLMs to deal with situations like users who ask how to build a bomb. When you ask ChatGPT a question, the model is first secretly prompted by OpenAI to refrain from telling you how to make bombs or do other nefarious things. But its easy for users to jailbreak AI models with clever prompts, bypassing any restrictions.If the creators of the models are able to see where in an AI the bomb-building knowledge is, they can theoretically turn off those nodes permanently. Then even the most cleverly written prompt wouldnt elicit an answer about how to build a bomb, because the AI would literally have no information about how to build a bomb in its system.This type of granularity and precise control are easy to imagine but extremely hard to achieve with the current state of mechanistic interpretability.A limitation is the steering [influencing a model by adjusting its parameters] is just not working that well, and so when you steer to reduce violence in a model, it ends up completely lobotomizing its knowledge in martial arts. Theres a lot of refinement to be done in steering, says Lin. The knowledge of bomb making, for example, isnt just a simple on-and-off switch in an AI model. It most likely is woven into multiple parts of the model, and turning it off would probably involve hampering the AIs knowledge of chemistry. Any tinkering may have benefits but also significant trade-offs.That said, if we are able to dig deeper and peer more clearly into the mind of AI, DeepMind and others are hopeful that mechanistic interpretability could represent a plausible path to alignmentthe process of making sure AI is actually doing what we want it to do.
    0 Comments ·0 Shares ·158 Views
  • Save up to $250 on every M4 Mac mini, plus get M2 early Black Friday deals from $449
    appleinsider.com
    Apple's new Mac mini is eligible for promo code savings, with every M4 and M4 Pro spec up to $250 off. Plus, grab closeout deals on M2 models, with prices starting at $449.Coupon savings are in effect on every M4 Mac mini.The early Black Friday deals on Apple's M4 Mac mini are thanks to promo code APINSIDER at Apple Authorized Reseller Adorama. In business since 1974, Adorama has issued exclusive discounts of up to $250 off every model, including M4 Pro and 10 Gigabit Ethernet configurations. Continue Reading on AppleInsider | Discuss on our Forums
    0 Comments ·0 Shares ·112 Views
  • Black Friday mega sale: 36% off AI Photo Enhancer for 8x Bold Sight
    appleinsider.com
    Everyone has photos taken on earlier digital cameras that left much to be desired for detail and accuracy, but you can easily restore and upscale any photo up to 16K with Aiarty Image Enhancer.Aiarty Image Enhancer - Image credit: DigiartyAI photo enhancer tools are becoming more mainstream, but you'll want the best tools available to handle your precious memories with care. Whether you've got poorly scanned copies of old photographs, aging digital camera stills, or older iPhone images that don't hold up to today's standards Aiarty Image Enhancer can help.Access Black Friday Sale Continue Reading on AppleInsider
    0 Comments ·0 Shares ·106 Views
  • Immersive restaurant, Buenos Aires, Argentina - e-architect
    www.facebook.com
    The Immersive Restaurants, Argentina, project, designed by Carla Bechelli Arquitectos, offers a variety of spaces and experiences for its usershttps://www.e-architect.com/argentina/immersive-restaurant-buenos-aires-argentina#restaurantdesign #argentina #architecture #architectThe Immersive restaurant, Buenos Aires, Argentina, project, designed by Carla Bechelli Arquitectos, offers a variety of spaces
    0 Comments ·0 Shares ·112 Views
  • Nima Bay Development, Puerto Vallarta, Mexico - e-architect
    www.facebook.com
    Central de Arquitecturas design for the Nima Bay Resort Development on Mexicos Pacific Coast is a stylish destination for luxury travellershttps://www.e-architect.com/mexico/nima-bay-development-puerto-vallarta-mexico#resort #mexico #pacificcoast #luxurytravellers #architectureCentral de Arquitectura's design for the Nima Bay Resort Development in Mexico is a stylish destination for luxury travellers
    0 Comments ·0 Shares ·111 Views
  • Project Stress Index records slight improvements after September interest rate cuts
    archinect.com
    The new ConstructionConnect Project Stress Index (PSI) composite reading for October has shown a mild improvement to 104.2, rising 0.8% for the month after a revised September reading of 103.4.Compared to the same month one year ago, the composite is down 11.6%. This is the fourth consecutive month in a row for year-to-year contractions. However, ConstructionConnect's Chief Economist Michael Guckes tells us: "This year has been very good for the PSI, with year-to-date composite performance down 14.6%. If conditions continue at this level for the rest of the year, the PSI will arguably have its best year since 2021."Related on Archinect: July sees sharp improvement in project conditions as interest rates are expected to fallThe total number of project abandonments, along with the number ofon-hold projects, were both recorded as hitting new multi-year lows. Guckes says it is due largely to the preemptive anticipation of Federal Funds Rate (FFR) decreases, which have an impact of pr...
    0 Comments ·0 Shares ·135 Views
  • Home owned by late Sea Ranch architect George Homsey lists for $2.45 million
    archinect.com
    The home once owned by one of the leading late-modernist figures in Northern CaliforniatheEHDD co-founder George Homseys exemplary Third Bay Tradition style former personal residence in San Franciscois now for sale for $2.45 million viaSuperstructur. The online brokerage, which claims to focus "solely on architecture and design with integrity," has the listing in the Liberty HillHistoric District. They say: "The house is in very original condition and now requires a custodian who is sensitive to the building's architectural importance and to the Homsey family's wishes that it may be carefully restored and preserved for the enjoyment of future generations."Homsey passed away at the age of 93 in April of 2019. Last month in the region, we also highlighted the listing of Julia Morgan's original 1908 home in Berkeley for $1.9 million.
    0 Comments ·0 Shares ·126 Views
  • INC Architecture & Design uses fluted terra-cotta panels for The Treadwell, an art decoinspired residential tower
    www.facebook.com
    INC Architecture & Design uses fluted terra-cotta panels for The Treadwell, an art decoinspired residential towerNC Architecture & Design used fluted terra-cotta panels for The Treadwell, a residential Tower at the threshold of Midtown and the Upper East Side.
    0 Comments ·0 Shares ·135 Views