logo
#

Latest news with #GoogleIO

NotebookLM Is Still the Best AI Tool You're Missing Out On
NotebookLM Is Still the Best AI Tool You're Missing Out On

CNET

time4 days ago

  • CNET

NotebookLM Is Still the Best AI Tool You're Missing Out On

AI tools are everywhere, each promising to make your daily workflow easier in one way or another. More often than not, they overpromise, underdeliver or require a learning curve to get them to work for you reasonably well. NotebookLM is different. It's incredibly helpful at its most simple function and it just gets better from there. Plus, it's easy to navigate and see what else it can do with minimal poking around. I always appreciated the concept of what NotebookLM could do when it was first announced and enjoyed casually tinkering with it, but experiencing it in the real world -- and at a time when I needed a tool like it to exist -- solidified how awesome NotebookLM really is to me. It's now my favorite AI tool, ever, and I've never used an AI tool as much. Powered by Google's Gemini AI, NotebookLM breaks down complex subjects into an easy-to-understand format and helps you brainstorm new ideas. There's now a mobile app for iOS and Android, and new features were just announced during Google I/O last month. It keeps getting better without feeling like it's becoming overstuffed with features just for the sake of it. Everything Announced at Google I/O 2025 Everything Announced at Google I/O 2025 Click to unmute Video Player is loading. Play Video Pause Skip Backward Skip Forward Next playlist item Unmute Current Time 0:11 / Duration 0:30 Loaded : 0.00% 0:11 Stream Type LIVE Seek to live, currently behind live LIVE Remaining Time - 0:19 Share Fullscreen This is a modal window. This video is either unavailable or not supported in this browser Error Code: MEDIA_ERR_SRC_NOT_SUPPORTED The media could not be loaded, either because the server or network failed or because the format is not supported. Technical details : Session ID: 2025-06-24:5d5895938192cef4dcb7fa1 Player Element ID: vjs_video_3 OK Close Modal Dialog Beginning of dialog window. Escape will cancel and close the window. Text Color White Black Red Green Blue Yellow Magenta Cyan Opacity Opaque Semi-Transparent Text Background Color Black White Red Green Blue Yellow Magenta Cyan Opacity Opaque Semi-Transparent Transparent Caption Area Background Color Black White Red Green Blue Yellow Magenta Cyan Opacity Transparent Semi-Transparent Opaque Font Size 50% 75% 100% 125% 150% 175% 200% 300% 400% Text Edge Style None Raised Depressed Uniform Drop shadow Font Family Proportional Sans-Serif Monospace Sans-Serif Proportional Serif Monospace Serif Casual Script Small Caps Reset Done Close Modal Dialog End of dialog window. Close Modal Dialog This is a modal window. This modal can be closed by pressing the Escape key or activating the close button. Close Modal Dialog This is a modal window. This modal can be closed by pressing the Escape key or activating the close button. Everything Announced at Google I/O 2025 NotebookLM isn't just Google Keep stuffed with AI, nor is it just a chatbot that can take notes. It's both and neither. Instead of asking questions to Gemini, only for it to find an answer from the ether of the internet, NotebookLM will only search through the sources that you provide it. It's a dead simple concept that feels like one of the most practical uses of AI, giving way to the perfect study buddy for classes or work. And Google didn't stop there. Now it can do so much more, and will reward your poking around to see what it can do for you. And features like its impressive Audio Overviews have since trickled down into Gemini itself, allowing it to be used in a much wider set of Google's products. Below, I'll cover some of NotebookLM's most interesting features (including the newly announced ones) and how it became one of my favorite AI tools to use. For more, check out Google's smart glasses plans with AndroidXR. What is NotebookLM? NotebookLM is a Gemini-powered note-taking and research assistant tool that can be used in a multitude of ways. It all starts with the sources you feed it, whether they're webpage URLs, YouTube videos or audio clips, allowing you to pull multiple sources together into a cohesive package and bring some organization to your scattered thoughts or notes. The most obvious use case for NotebookLM is using it for school or work. Think of it -- you've kept up with countless classes and typed notes down for every one and even perhaps recorded some lectures. Sifting through everything individually can eventually get you to some semblance of understanding, but what if you could get them to work together? Once you've uploaded your sources, Gemini will get to work to create an overall summary of the material. From there, you can begin asking Gemini questions about specific topics on the sources and information from the sources will be displayed in an easy-to-understand format. This alone may be enough for some people just looking to get the most out of their notes, but that's really just scratching the surface. Available for desktop and mobile NotebookLM has a three-panel layout. Screenshot by Blake Stimac/CNET NotebookLM has been available for a while now on the desktop and is broken into a three-panel layout, consisting of Source, Chat and Studio panels. Both the Source and Studio panels are collapsible, so you can have a full-screen chat experience if you prefer. While the Source and Chat panels are pretty self-explanatory, the Studio panel is where magic can happen (though some of the features can also be created directly from the Chat panel). This is where you can get the most out of your NotebookLM experience. The NotebookLM app: The Philosopher's Stone for data, in your pocket The mobile app for Android and iOS launched the day before Google I/O 2025. Screenshots by Blake Stimac/CNET Those familiar with the desktop experience will feel right at home with the new mobile apps for iOS and Android. The streamlined app allows you to switch between the Source, Chat and Studio panels via a menu at the bottom. When you go to the view that shows all of your notebooks, you'll see tabs for Recent, Shared, Title and Downloaded. While not everything is on the app yet, it's likely just a matter of time before it matches the web version's full functionality. Audio Overviews If you didn't hear about NotebookLM when it was first announced, you likely did when Audio Overviews were released for it. Once you have at least one source uploaded, you can then opt to generate an Audio Overview, which will provide a "deep dive" on the source material. These overviews are created by none other than Gemini, and it's not just a quick summary of your material in audio format -- it's a full-blown podcast with two "hosts" that break down complex topics into easy-to-understand pieces of information. They're incredibly effective, too, often asking each other questions to dismantle certain topics. The default length of an Audio Overview will vary depending on how much material there is to go over and the complexity of the topic -- though I'm sure there are other factors at play. In my testing, a very short piece of text created a five-minute audio clip, whereas two lengthier and more dense Google Docs documents I uploaded created an 18-minute Overview. If you want a little more control on the length for your Audio Overview, you're in luck. Announced in a blog post during Google I/O earlier this month, users now have three options to choose from: shorter, default and longer. This is perfect if you either want to have a short and dense podcast-like experience of if you want to get into the nitty gritty on a subject with a longer Audio Overview. You can interact with your AI podcasters It gets even better. Last December, NotebookLM got a new design and new ways to interact with Audio Overviews. The customize button allows you to guide the conversation so that key points are covered. Type in your directive and then generate your Audio Overview. Now, if you want to make this feature even more interactive, you can choose the Interactive mode, which is still in beta, to join the conversation. The clip will play, and if you have a particular question in response to something that's said, you can click the join button. Once you do, the speakers will pause and acknowledge your presence and ask you to chime in with thoughts or questions, and you'll get a reply. I wanted to try something a little different, so I threw in the lyrics of a song as the only source, and the AI podcast duo began to dismantle the motivations and emotions behind the words. I used the join feature to point out a detail in the lyrics they didn't touch on, and the two began to dissect what my suggestion meant in the context of the writing. They then began linking the theme to other portions of the text. It was impressive to watch: They handled the emotional weight of the song so well, and tactfully at that. Mind Maps Generating a Mind Map is just one of several powerful features from NotebookLM. Google/Screenshot by Blake Stimac I'd heard interesting things about NotebookLM's Mind Map feature, but I wanted to go in blind when I tried it out, so I did a separate test. I took roughly 1,500 words of Homer's Odyssey and made that my only source. I then clicked the Mind Map button, and within seconds, an interactive and categorical breakdown of the text was displayed for me to poke around in. Many of the broken-down sections had subsections for deeper dives, some of which were dedicated to single lines for dissection. Clicking on a category or end-point of the map will open the chat with a prefilled prompt. I chose to dive into the line, "now without remedy," and once clicked, the chat portion of NotebookLM reopened with the prefilled prompt, "Discuss what these sources say about Now without remedy, in the larger context of [the subsection] Alternative (worse)." The full line was displayed, including who said it, what it was in response to and any motivations (or other references) for why the line was said in the text. Study guides and more If the combination of all that Audio Overviews and Mind Maps could do sounds like everything a student might need for the perfect study buddy, NotebookLM has a few other features that will solidify it in that place. Study guides After you've uploaded a source, you can create a quick study guide based on the material that will automatically provide a document with a quiz, potential essay questions, a glossary of key terms and answers for the quiz at the bottom. And if you want, you can even convert the study guide into a source for your notebook. FAQs Whether you're using it for school or want to create a FAQ page for your website, the NotebookLM button generates a series of potentially common questions based on your sources. Timeline If you're looking for a play-by-play sort of timeline, it's built right in. Creating a timeline for the Odyssey excerpt broke down main events in a bulleted list and placed them based on the times mentioned in the material. If an event takes place at an unspecified time, it will appear at the top of the timeline, stating this. A cast of characters for reference is also generated below the timeline of events. Briefing document The briefing document is just what it sounds like, giving you a quick snapshot of the key themes and important events to get someone up to speed. This will include specific quotes from the source and their location. A summary of the material is also created at the bottom of the document. How NotebookLM really 'sold' me I already really liked NotebookLM's concept and execution during its 1.0 days, and revisiting the new features only strengthened my appreciation for it. My testing was mostly for fun and to see how this tool can flex, but using it when I "needed" it helped me really get an idea of how powerful it can be, even for simple things. During a product briefing, I did my typical note-taking: Open a Google Doc, start typing in fragmented thoughts on key points, and hope I could translate what I meant when I needed to refer back to them. I knew I would also receive an official press release, so I wasn't (too) worried about it, but I wanted to put NotebookLM to the test in a real-world situation when I was using it for real -- and not just tinkering, when nearly anything seems impressive when it does what you tell it to. I decided to create a new notebook and make my crude notes (which looked like a series of bad haikus at first glance) the only source, just to see what came out on the other end. Not only did NotebookLM fill in the blanks, but the overall summary read almost as well as the press release I received the following day. I was impressed. It felt alchemical -- NotebookLM took some fairly unintelligible language and didn't just turn it into something passable, but rather, a pretty impressive description. Funny enough, I've since become a more thorough note-taker, but I'm relieved to know I have something that can save the day if I need it to. Video Overviews are on the way Another feature that was announced during Google I/O was Video Overviews, and it's exactly what it sounds like. There's currently no time frame outside of "coming soon" from the blog post, but it should be a good way to get a more visual experience from your notebooks. We'd previously heard that Video Overviews might be on the way, thanks to some sleuthing from Testing Catalog. The article also mentioned that the ability to make your notebooks publicly available and view an Editor's Picks list of notebooks will eventually make their way to NotebookLM. The Editors Picks feature has yet to rear its head, but you can indeed now share notebooks directly or make them publicly available for anyone to access. While we're waiting on View Overviews, here's a preview of one. If you need more from NotebookLM, consider upgrading Most people will likely never have the need to pay for NotebookLM, as the free version is robust enough. But if you're using it for work and need to be able to add more sources or the option to share your notebook with multiple people, NotebookLM Plus is worth considering. It gives you more of everything while introducing more customization, additional privacy and security features as well as analytics. It's worth noting that NotebookLM Plus will also be packaged in with Google's new AI subscriptions. For more, don't miss Google's going all-in on AI video with Flow and Veo 3.

Meet Google Martha, the company's Android XR smart glasses prototype
Meet Google Martha, the company's Android XR smart glasses prototype

Android Authority

time5 days ago

  • Android Authority

Meet Google Martha, the company's Android XR smart glasses prototype

TL;DR Google demoed its Android XR smart glasses prototype at Google I/O 2025. We now learn that this Android XR prototype is called 'Google Martha.' Its companion app handles connected features like notifications, settings access, video recording from the user's perspective, and more. After over a year of teasing with Project Astra, Google showed off its Android XR glasses on stage at Google I/O 2025. My colleague C. Scott Brown even got to try them on, and he was impressed with the demo. Since these are prototype glasses and not meant for retail sale, there's not a lot of information on them, but we've now spotted their codename. Meet Google Martha, Google's name for its smart glasses prototype. App developer Sayed Ali Alkamel shared a photo of the companion app of the Android XR prototype glasses (h/t SERoundtable), which shows off a few settings and features of the connected smart glasses. I've rotated the image and edited the perspective to give us a better look at what's on the phone: As we can see, the connected Android XR smart glasses prototype is called 'Google Martha.' The companion app has entries for Notifications and Settings, but unfortunately, we don't get to see the entries within. The app also has a Record your view entry, letting the wearer capture a video of their view and the glasses' UI. There are also entries for feedback and reporting a problem. From Google I/O 2025, we know these prototype smart glasses run on the Android XR platform, opening up several Gemini-oriented use cases, such as real-world identification and querying, live translation, and more. Google Martha has a screen in only the right lens by design, though other smart glasses can have a dual-lens screen, or even none at all, and rely only on audio. If you want to get your hands on Google Martha, you will likely be disappointed. A report from earlier in the year noted that Google and Samsung were jointly developing Android XR glasses that are seemingly scheduled for consumer release in 2026, but Google did not confirm or corroborate or confirm these plans at Google I/O when it showed off Google Martha. This pair of smart glasses is unlikely to reach consumers since it's just a prototype, but the door is open for future smart glasses based on Martha to become available for you and me eventually. Until then, you can look forward to XReal's Project Aura or even Samsung's Project Moohan. Got a tip? Talk to us! Email our staff at Email our staff at news@ . You can stay anonymous or get credit for the info, it's your choice.

Google Jules v2 Update Released : Free AI Coder Makes Creation Fun Again
Google Jules v2 Update Released : Free AI Coder Makes Creation Fun Again

Geeky Gadgets

time19-06-2025

  • Business
  • Geeky Gadgets

Google Jules v2 Update Released : Free AI Coder Makes Creation Fun Again

What if coding could feel less like a grind and more like a creative flow? Enter Google Jules v2, the latest evolution in AI-driven development tools that promises to redefine how programmers approach their craft. Unveiled at the prestigious Google I/O conference, this free asynchronous AI coder is more than just an upgrade—it's a fantastic option. Imagine automating tedious bug fixes, generating precise code snippets on demand, and managing multiple projects seamlessly, all while focusing on the creative aspects of your work. Built on the innovative Gemini 2.5 Pro model, Jules v2 doesn't just keep up with the demands of modern software development—it sets a new benchmark. Whether you're a seasoned developer or a coding newcomer, this tool offers something extraordinary: the freedom to innovate without being bogged down by repetitive tasks. In this video World of AI explain how Jules v2 is reshaping the coding landscape, from its autonomous bug-fixing capabilities to its intuitive task management tools. You'll discover how its asynchronous architecture allows developers to juggle up to 60 tasks daily, a staggering leap from its predecessor's capacity. But that's just the beginning—Jules v2 doesn't stop at efficiency; it's designed to empower creativity, offering features like customizable code panels and seamless GitHub integration. By the end, you'll see why this tool isn't just an assistant but a partner in innovation. Could this be the future of coding? Let's explore the possibilities. Key Features of Jules v2 Jules v2 is carefully designed to handle a wide range of coding tasks autonomously, allowing developers to focus on strategic and creative aspects of their projects. Its asynchronous architecture ensures seamless multitasking across multiple projects, effectively eliminating bottlenecks and delays. Some of its most notable features include: Automated bug fixing and code refactoring: Keep your codebase clean and functional with minimal manual effort, making sure higher code quality. Keep your codebase clean and functional with minimal manual effort, making sure higher code quality. Task management tools: Plan, organize, and execute projects efficiently with built-in tools that simplify workflow management. Plan, organize, and execute projects efficiently with built-in tools that simplify workflow management. Daily commit summaries: Stay informed with 'C codecast' podcasts that provide concise updates on project progress without requiring additional time investment. Stay informed with 'C codecast' podcasts that provide concise updates on project progress without requiring additional time investment. Autonomous application creation and deployment: Automate repetitive tasks to simplify and accelerate development pipelines. These features make Jules v2 an indispensable tool for developers aiming to optimize workflows, reduce time spent on routine tasks, and focus on innovation. What's New in Jules v2? The latest iteration of Jules introduces several significant upgrades designed to enhance productivity and user experience. Powered by the Gemini 2.5 Pro model, Jules v2 delivers a range of improvements that cater to the evolving needs of developers: Expanded task capacity: Manage up to 60 tasks daily, a substantial increase from the previous limit of 5, allowing greater efficiency. Manage up to 60 tasks daily, a substantial increase from the previous limit of 5, allowing greater efficiency. Direct code snippet management: Easily copy and download code snippets, improving accessibility and workflow integration. Easily copy and download code snippets, improving accessibility and workflow integration. Enhanced task management modal: Navigate multiple projects effortlessly with an intuitive and user-friendly interface. Navigate multiple projects effortlessly with an intuitive and user-friendly interface. Customizable code panel: Adapt the interface to suit specific project requirements, offering greater flexibility. Adapt the interface to suit specific project requirements, offering greater flexibility. Performance improvements: Experience faster, smoother interactions, making sure a more efficient and seamless coding experience. These enhancements position Jules v2 as a leading tool in the realm of AI-driven development, offering unparalleled flexibility and efficiency for developers of all experience levels. Google Jules v2 Update June 2025 Watch this video on YouTube. Explore further guides and articles from our vast library that you may find relevant to your interests in AI coding tools. The Gemini 2.5 Pro Model: Driving Innovation At the core of Jules v2 lies the Gemini 2.5 Pro model, a state-of-the-art AI specifically engineered for coding. This advanced model delivers exceptional performance, surpassing earlier versions and competing tools in handling complex development tasks. Its capabilities include: Advanced debugging and refactoring: Address intricate coding challenges with precision and efficiency. Address intricate coding challenges with precision and efficiency. Accurate code generation: Generate functional and contextually relevant code snippets tailored to diverse use cases. Generate functional and contextually relevant code snippets tailored to diverse use cases. Automated application development: Streamline the setup of backend and frontend components, significantly reducing project completion times. The Gemini 2.5 Pro model ensures that Jules v2 is not only reliable but also highly effective, making it an essential resource for developers seeking to enhance their productivity and coding accuracy. Real-World Applications Jules v2 excels in practical scenarios, addressing common development challenges with remarkable efficiency. For instance, it can autonomously set up backend and frontend components for a CRM dashboard, drastically reducing setup time. Its seamless integration with GitHub simplifies repository management, including branch creation, pull requests, and version control. Furthermore, its compatibility with popular databases like MongoDB and Superbase enables efficient data monitoring and management, making it a versatile tool for a variety of development needs. Accessibility and Onboarding One of the standout aspects of Jules v2 is its accessibility. The tool is completely free to use with a Google account, removing financial barriers and making it available to a wide range of developers. The onboarding process is straightforward, allowing users to quickly integrate Jules into their workflows. Whether you are a beginner exploring the fundamentals of coding or an experienced professional managing complex projects, Jules v2 is designed to accommodate diverse skill levels and requirements. Media Credit: WorldofAI Filed Under: AI, Guides Latest Geeky Gadgets Deals Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, Geeky Gadgets may earn an affiliate commission. Learn about our Disclosure Policy.

Everything you need to know from Google I/O 2025
Everything you need to know from Google I/O 2025

Yahoo

time02-06-2025

  • Business
  • Yahoo

Everything you need to know from Google I/O 2025

From the opening AI-influenced intro video set to "You Get What You Give" by New Radicals to CEO Sundar Pichai's sign-off, Google I/O 2025 was packed with news and updates for the tech giant and its products. And when we say packed, we mean it, as this year's Google I/O clocked in at nearly two hours long. During that time, Google shared some big wins for its AI products, such as Gemini topping various categories on the LMArena leaderboard. Another example that Google seemed really proud of was the fact that Gemini completed Pokémon Blue a few weeks ago. But, we know what you're really here for: Product updates and new product announcements. Aside from a few braggadocious moments, Google spent most of those 117 minutes talking about what's coming out next. Google I/O mixes consumer-facing product announcements with more developer-oriented ones, from the latest Gmail updates to Google's powerful new chip, Ironwood, coming to Google Cloud customers later this year. We're going to break down what product updates and announcements you need to know from the full two-hour event, so you can walk away with all the takeaways without spending the same time it takes to watch a major motion picture to learn about them. Before we dive in though, here's the most shocking news out of Google I/O: The subscription pricing that Google has for its Google AI Ultra plan. While Google provides a base subscription at $19.99 per month, the Ultra plan comes in at a whopping $249.99 per month for its entire suite of products with the highest rate limits available. Google tucked away what will easily be its most visible feature way too far back into the event, but we'll surface it to the top. At Google I/O, Google announced that the new AI Mode feature for Google Search is launching today to everyone in the United States. Basically, it will allow users to use Google's search feature but with longer, more complex queries. Using a "query fan-out technique," AI Mode will be able to break a search into multiple parts in order to process each part of the query, then pull all the information together to present to the user. Google says AI Mode "checks its work" too, but its unclear at this time exactly what that means. Google announces AI Mode in Google Search Credit: Google AI Mode is available now. Later in the summer, Google will launch Personal Context in AI Mode, which will make suggestions based on a user's past searches and other contextual information about the user from other Google products like Gmail. In addition, other new features will soon come to AI Mode, such as Deep Search, which can dive deeper into queries by searching through multiple websites, and data visualization features, which can take the search results and present them in a visual graph when applicable. According to Google, its AI overviews in search are viewed by 1.5 billion users every month, so AI Mode clearly has the largest potential user base out of all of Google's announcements today. Out of all the announcements at the event, these AI shopping features seemed to spark the biggest reaction from Google I/O live attendees. Connected to AI Mode, Google showed off its Shopping Graph, which includes more than 50 billion products globally. Users can just describe the type of product they are looking for – say a specific type of couch, and Google will present options that match that description. Google AI Shopping Credit: Google Google also had a significant presentation that showed its presenter upload a photo of themselves so that AI could create a visual of what she'd look like in a dress. This virtual try-on feature will be available in Google Labs, and it's the IRL version of Cher's Clueless closet. The presenter was then able to use an AI shopping agent to keep tabs on the item's availability and track its price. When the price dropped, the user received a notification of the pricing change. Google said users will be able to try on different looks via AI in Google Labs starting today. Google's long-awaited post-Google Glass AR/VR plans were finally presented at Google I/O. The company also unveiled a number of wearable products utilizing its AR/VR operating system, Android XR. One important part of the Android XR announcement is that Google seems to understand the different use cases for an immersive headset and an on-the-go pair of smartglasses and have built Android XR to accommodate that. While Samsung has previously teased its Project Moohan XR headset, Google I/O marked the first time that Google revealed the product, which is being built in partnership with the mobile giant and chipmaker Qualcomm. Google shared that the Project Moohan headset should be available later this year. Project Moohan Credit: Google In addition to the XR headset, Google announced Glasses with Android XR, smartglasses that incorporate a camera, speakers, and in-lens display that connect with a user's smartphone. Unlike Google Glass, these smart glasses will incorporate more fashionable looks thanks to partnerships with Gentle Monster and Warby Parker. Google shared that developers will be able to start developing for Glasses starting next year, so it's likely that a release date for the smartglasses will follow after that. Easily the star of Google I/O 2025 was the company's AI model, Gemini. Google announced a new updated Gemini 2.5 Pro, which it says is its most powerful model yet. The company showed Gemini 2.5 Pro being used to turn sketches into full applications in a demo. Along with that, Google introduced Gemini 2.5 Flash, which is a more affordable version of the powerful Pro model. The latter will be released in early June with the former coming out soon after. Google also revealed Gemini 2.5 Pro Deep Think for complex math and coding, which will only be available to "trusted testers" at first. Speaking of coding, Google shared its asynchronous coding agent Jules, which is currently in public beta. Developers will be able to utilize Jules in order to tackle codebase tasks and modify files. Jules coding agent Credit: Google Developers will also have access to a new Native Audio Output text-to-speech model which can replicate the same voice in different languages. The Gemini app will soon see a new Agent Mode, bringing users an AI agent who can research and complete tasks based on a user's prompts. Gemini will also be deeply integrated into Google products like Workspace with Personalized Smart Replies. Gemini will use personal context via documents, emails, and more from across a user's Google apps in order to match their tone, voice, and style in order to generate automatic replies. Workspace users will find the feature available in Gmail this summer. Other features announced for Gemini include Deep Research, which lets users upload their own files to guide the AI agent when asking questions, and Gemini in Chrome, an AI Assistant that answers queries using the context on the web page that a user is on. The latter feature is rolling out this week for Gemini subscribers in the U.S. Google intends to bring Gemini to all of its devices, including smartwatches, smart cars, and smart TVs. Gemini's AI assistant capabilities and language model updates were only a small piece of Google's broader AI puzzle. The company had a slew of generative AI announcements to make too. Google announced Imagen 4, its latest image generation model. According to Google, Imagen 4 provides richer details and better visuals. In addition, Imagen 4 is apparently much better at generating text and typography in its graphics. This is an area which AI models are notoriously bad at, so Imagen 4 appears to be a big step forward. Flow AI video tool Credit: Google A new video generation model, Veo 3, was also unveiled with a video generation tool called Flow. Google claims Veo 3 has a stronger understanding of physics when generating scenes and can also create accompanying sound effects, background noise, and dialogue. Both Veo 3 and Flow are available today alongside a new generative music model called Lyria 2. Google I/O also saw the debut of Gemini Canvas, which Google describes as a co-creation platform. Another big announcement out of Google I/O: Project Starline is no more. Google's immersive communication project will now be known as Google Beam, an AI-first communication platform. As part of Google Beam, Google announced Google Meet translations, which basically provides real-time speech translation during meetings on the platform. AI will be able to match a speaker's voice and tone, so it sounds like the translation is coming directly from them. Google Meet translations are available in English and Spanish starting today with more language on the way in the coming weeks. Google Meet translations Credit: Google Google also had another work-in-progress project to tease under Google Beam: A 3-D conferencing platform that uses multiple cameras to capture a user from different angles in order to render the individual on a 3-D light-field display. While Project Starline may have undergone a name change, it appears Project Astra is still kicking it at Google, at least for now. Project Astra is Google's real-world universal AI assistant and Google had plenty to announce as part of it. Gemini Live is a new AI assistant feature that can interact with a user's surroundings via their mobile device's camera and audio input. Users can ask Gemini Live questions about what they're capturing on camera and the AI assistant will be able to answer queries based on those visuals. According to Google, Gemini Live is rolling out today to Gemini users. Gemini Live Credit: Google It appears Google has plans to implement Project Astra's live AI capabilities into Google Search's AI mode as a Google Lens visual search enhancement. Google also highlighted some of its hopes for Gemini Live, such as being able to help as an accessibility tool for those with disabilities. Another one of Google's AI projects is an AI agent that can interact with the web in order to complete tasks for the user known as Project Mariner. While Project Mariner was previously announced late last year, Google had some updates such as a multi-tasking feature which would allow an AI agent to work on up to 10 different tasks simultaneously. Another new feature is Teach and Repeat, which would provide the AI agent with the ability to learn from previously completed tasks in order to complete similar ones without the need for the same detailed direction in the future. Google announced plans to bring these agentic AI capabilities to Chrome, Google Search via AI Mode, and the Gemini app.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store