Productivity at Google I/O: 3D videoconferencing, real-time voice translation, AI agents
Google is implementing AI-driven enhancements to its software and collaboration tools that could translate to big productivity gains.
Google CEO Sundar Pichai started his keynote at the Google I/O developer conference this week with one of those features: real-time language translation in Google Meet, which was until now in research.
The feature uses Google’s AI technology to translate speech from one language to another in near real time, while matching the tone and expressions — such as “hmmm” — during delivery. The technology breaks down language barriers, Pichai said.
A video demonstrated the translation from spoken English to Spanish. A computer-generated voiceover spoke the translation after a one-second delay. Then the other participant’s response was translated from Spanish to English.
“We are even closer to having a natural and free-flowing conversation across languages,” Pichai said.
Translations to English and Spanish are now available in beta for Google AI Pro and Ultra subscribers, with more languages rolling out in the next few weeks.
“Real-time translations will be coming to enterprises later this year,” Pichai said.
Pichai, in the opening moments of his keynote, also mentioned a new product called Google Beam, a 3D video communications platform that transforms 2D video streams into a realistic experience. The product was in development for many years under a research effort called Project Starline.
Behind the scenes, an array of six cameras captures participants from different angles.
“With AI we can merge these video streams together and render you on a 3D light-field display with near perfect head tracking down to the millimeter and at 60 frames per second, all in real time,” Pichai said.
The result, Pichai said, was a much more natural and deeply immersive conversational experience.
The first Google Beam devices will be available for early customers later this year. Google is partnering with device maker HP, which will share more information about these devices a few weeks from now.
Traditional videoconferencing reduces many natural social cues that people experience in face-to-face interactions, which is where something like Google Beam fits in, said J.P. Gownder, vice president and principal analyst on Forrester’s Future of Work team.
“Those subtle cues contain a lot more information than people realize, so richer forms of videoconferencing that feel naturalistic could offer benefits,” said Gownder.
Since users don’t have to wear specialized equipment such as VR headgear, it will be that much more accessible.
“It’s probably a longer-term play, but if the experience is better, over time it could get traction,” Gownder said.
Google also shed more light on how it is implementing Workspace Flows, a feature that was introduced at last month’s Google Cloud Next as a way to automate work across Google Workspace apps.
Workspace Flows brings AI agents into the loop to get work done. Google can utilize AI agents called Gems that specialize in certain tasks such as customer service and work with other AI agents to complete tasks.
An onstage Gems demonstration showed how workers could use AI agents to automate processing customer service by sharing complaints with other employees via Google Chat, looking up product literature, referencing an internal genAI model for further answers, and automatically sending a possible resolution to customers.
Gems use Google’s Gemini AI model to analyze information, prioritize tasks, and generate feedback.
“These are AI experts you can create to solve particular tasks… you can actually have a team of Gems working together to solve these issues for you,” said Farhaz Karmali, Google’s product director for the Workspace ecosystem, during a keynote on the second day of Google I/O.
Karmali also shared some richer conversational features coming to Google Chat. Users will be able to subscribe to messages in a conversation, create groups, and manage memberships. These features will be helpful for ensuring targeted conversations and information reach the right agents.
“Imagine: you build a chat app that is agentic and you want to get all the information from a chat app. You can now summarize it, and this helps you take actions and so on,” Karmali said.
Other AI-powered features coming to Google Workspace include personalized smart replies, directed inbox cleanup, and fast appointment scheduling in Gmail; the ability to turn Google Slides decks into videos; and writing assistance in Google Docs that limits Gemini to specific sources designated by the user.
Google I/O overlapped with Microsoft’s Build developer conference, where the company introduced new Copilot agent features this week. Both tech giants are in the early stages of AI, and the shows focused on how they are still developing their unique agent ecosystems, Gownder said.
#productivity #google #videoconferencing #realtime #voice
Productivity at Google I/O: 3D videoconferencing, real-time voice translation, AI agents
Google is implementing AI-driven enhancements to its software and collaboration tools that could translate to big productivity gains.
Google CEO Sundar Pichai started his keynote at the Google I/O developer conference this week with one of those features: real-time language translation in Google Meet, which was until now in research.
The feature uses Google’s AI technology to translate speech from one language to another in near real time, while matching the tone and expressions — such as “hmmm” — during delivery. The technology breaks down language barriers, Pichai said.
A video demonstrated the translation from spoken English to Spanish. A computer-generated voiceover spoke the translation after a one-second delay. Then the other participant’s response was translated from Spanish to English.
“We are even closer to having a natural and free-flowing conversation across languages,” Pichai said.
Translations to English and Spanish are now available in beta for Google AI Pro and Ultra subscribers, with more languages rolling out in the next few weeks.
“Real-time translations will be coming to enterprises later this year,” Pichai said.
Pichai, in the opening moments of his keynote, also mentioned a new product called Google Beam, a 3D video communications platform that transforms 2D video streams into a realistic experience. The product was in development for many years under a research effort called Project Starline.
Behind the scenes, an array of six cameras captures participants from different angles.
“With AI we can merge these video streams together and render you on a 3D light-field display with near perfect head tracking down to the millimeter and at 60 frames per second, all in real time,” Pichai said.
The result, Pichai said, was a much more natural and deeply immersive conversational experience.
The first Google Beam devices will be available for early customers later this year. Google is partnering with device maker HP, which will share more information about these devices a few weeks from now.
Traditional videoconferencing reduces many natural social cues that people experience in face-to-face interactions, which is where something like Google Beam fits in, said J.P. Gownder, vice president and principal analyst on Forrester’s Future of Work team.
“Those subtle cues contain a lot more information than people realize, so richer forms of videoconferencing that feel naturalistic could offer benefits,” said Gownder.
Since users don’t have to wear specialized equipment such as VR headgear, it will be that much more accessible.
“It’s probably a longer-term play, but if the experience is better, over time it could get traction,” Gownder said.
Google also shed more light on how it is implementing Workspace Flows, a feature that was introduced at last month’s Google Cloud Next as a way to automate work across Google Workspace apps.
Workspace Flows brings AI agents into the loop to get work done. Google can utilize AI agents called Gems that specialize in certain tasks such as customer service and work with other AI agents to complete tasks.
An onstage Gems demonstration showed how workers could use AI agents to automate processing customer service by sharing complaints with other employees via Google Chat, looking up product literature, referencing an internal genAI model for further answers, and automatically sending a possible resolution to customers.
Gems use Google’s Gemini AI model to analyze information, prioritize tasks, and generate feedback.
“These are AI experts you can create to solve particular tasks… you can actually have a team of Gems working together to solve these issues for you,” said Farhaz Karmali, Google’s product director for the Workspace ecosystem, during a keynote on the second day of Google I/O.
Karmali also shared some richer conversational features coming to Google Chat. Users will be able to subscribe to messages in a conversation, create groups, and manage memberships. These features will be helpful for ensuring targeted conversations and information reach the right agents.
“Imagine: you build a chat app that is agentic and you want to get all the information from a chat app. You can now summarize it, and this helps you take actions and so on,” Karmali said.
Other AI-powered features coming to Google Workspace include personalized smart replies, directed inbox cleanup, and fast appointment scheduling in Gmail; the ability to turn Google Slides decks into videos; and writing assistance in Google Docs that limits Gemini to specific sources designated by the user.
Google I/O overlapped with Microsoft’s Build developer conference, where the company introduced new Copilot agent features this week. Both tech giants are in the early stages of AI, and the shows focused on how they are still developing their unique agent ecosystems, Gownder said.
#productivity #google #videoconferencing #realtime #voice
·67 مشاهدة