logo
Gemma 3n AI model brings real-time multimodal power to mobiles

Gemma 3n AI model brings real-time multimodal power to mobiles

Techday NZ22-05-2025
Gemma 3n, a new artificial intelligence model architected for mobile and on-device computing, has been introduced as an early preview for developers.
Developed in partnership with mobile hardware manufacturers, Gemma 3n is designed to support real-time, multimodal AI experiences on phones, tablets, and laptops. The model extends the capabilities of the Gemma 3 family by focusing on performance and privacy in mobile scenarios.
The new architecture features collaboration with companies such as Qualcomm Technologies, MediaTek, and Samsung System LSI. The objective is to optimise the model for fast, responsive AI that can operate directly on device, rather than relying on cloud computing. This marks an extension of the Gemma initiative towards enabling AI applications in everyday devices, utilising a shared foundation that will underpin future releases across platforms like Android and Chrome.
According to information provided, Gemma 3n is also the core of the next generation of Gemini Nano, which is scheduled for broader release later in the year, bringing expanded AI features to Google apps and the wider on-device ecosystem. Developers can begin working with Gemma 3n today as part of the early preview, helping them to build and experiment with local AI functionalities ahead of general availability.
The model has performed strongly in chatbot benchmark rankings. One chart included in the announcement ranks AI models by Chatbot Arena Elo scores, with Gemma 3n noted as ranking highly amongst both popular proprietary and open models. Another chart demonstrates the model's mix-and-match performance with respect to model size.
Gemma 3n benefits from Google DeepMind's Per-Layer Embeddings (PLE) innovation, which leads to substantial reductions in RAM requirements. The model is available in 5 billion and 8 billion parameter versions, but, according to the release, it can operate with a memory footprint comparable to much smaller models—2 billion and 4 billion parameters—enabling operation with as little as 2GB to 3GB of dynamic memory. This allows the use of larger AI models on mobile devices or via cloud streaming, where memory overhead is often a constraint.
The company states, "Gemma 3n leverages a Google DeepMind innovation called Per-Layer Embeddings (PLE) that delivers a significant reduction in RAM usage. While the raw parameter count is 5B and 8B, this innovation allows you to run larger models on mobile devices or live-stream from the cloud, with a memory overhead comparable to a 2B and 4B model, meaning the models can operate with a dynamic memory footprint of just 2GB and 3GB."
Additional technical features of Gemma 3n include optimisations that allow the model to respond approximately 1.5 times faster on mobile devices compared to previous Gemma versions, with improved output quality and lower memory usage. The announcement highlights innovations such as Per Layer Embeddings, KVC sharing, and advanced activation quantisation as contributing to these improvements.
The model also supports what the company calls "many-in-1 flexibility." Utilizing a 4B active memory footprint, Gemma 3n incorporates a nested 2B active memory footprint submodel through the MatFormer training process. This design allows developers to balance performance and quality needs without operating separate models, composing submodels on the fly to match a specific application's requirements. Upcoming technical documentation is expected to elaborate on this mix-and-match capability.
Security and privacy are also prioritised. The development team states that local execution "enables features that respect user privacy and function reliably, even without an internet connection."
Gemma 3n brings enhanced multimodal comprehension, supporting the integration and understanding of audio, text, images, and video. Its audio functionality supports high-quality automatic speech recognition and multilingual translation. Furthermore, the model can accept inputs in multiple modalities simultaneously, enabling the parsing of complex multimodal interactions.
The company describes the expansion in audio capabilities: "Its audio capabilities enable the model to perform high-quality Automatic Speech Recognition (transcription) and Translation (speech to translated text). Additionally, the model accepts interleaved inputs across modalities, enabling understanding of complex multimodal interactions." A public release of these features is planned for the near future.
Gemma 3n features improved performance in multiple languages, with notable gains in Japanese, German, Korean, Spanish, and French. This is reflected in benchmark scores such as a 50.1% result on WMT24++ (ChrF), a multilingual evaluation metric.
The team behind Gemma 3n views the model as a catalyst for "intelligent, on-the-go applications." They note that developers will be able to "build live, interactive experiences that understand and respond to real-time visual and auditory cues from the user's environment," and design advanced applications capable of real-time speech transcription, translation, and multimodal contextual text generation, all executed privately on the device.
The company also outlined its commitment to responsible development. "Our commitment to responsible AI development is paramount. Gemma 3n, like all Gemma models, underwent rigorous safety evaluations, data governance, and fine-tuning alignment with our safety policies. We approach open models with careful risk assessment, continually refining our practices as the AI landscape evolves."
Developers have two initial routes for experimentation: exploring Gemma 3n via a cloud interface in Google AI Studio using browser-based access, or integrating the model locally through Google AI Edge's suite of developer tools. These options enable immediate testing of Gemma 3n's text and image processing capabilities.
The announcement states: "Gemma 3n marks the next step in democratizing access to cutting-edge, efficient AI. We're incredibly excited to see what you'll build as we make this technology progressively available, starting with today's preview."
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

AI YOUR WORKSPACE: Yealink launches full-scenario AI solutions for the hybrid era
AI YOUR WORKSPACE: Yealink launches full-scenario AI solutions for the hybrid era

Techday NZ

time25-06-2025

  • Techday NZ

AI YOUR WORKSPACE: Yealink launches full-scenario AI solutions for the hybrid era

Yealink has officially unveiled its latest innovations at the "AI YOUR WORKSPACE" global launch event, introducing a comprehensive lineup of AI-powered solutions across Android, Windows, ProAV, and Personal Collaboration. Backed by a 10-year strategic partnership with Microsoft, and joined by key ecosystem partners Intel and Qualcomm, the event underscored Yealink's commitment to building innovative, secure, and sustainable hybrid workplaces. All-in-one that goes bigger: MeetingBar A50 The first new solution introduced was the MeetingBar A50, Yealink's next-generation Android video bar. While all-in-one bars initially transformed small meeting spaces, today over 40% are deployed in medium to large rooms (Source: Frost & Sullivan), creating new demands for performance, audio quality, and camera intelligence. The A50 rises to meet those needs. Powered by the Qualcomm 8550 processor, it delivers 50% more CPU performance and over 10x AI processing power, enabling real-time framing, speaker tracking, and precision voice pickup. Its triple-camera system ensures everyone is clearly seen, while Microsoft Device Ecosystem Platform (MDEP) ensures enterprise-grade security meets the rigorous IT standards of large organizations. Smarter meetings for all: Upgraded MVC series Also unveiled was the upgraded MVC series for Microsoft Teams Rooms. Designed to tackle common hybrid meeting challenges—like inconsistent framing and limited visibility—MVC introduces meeting equity through flexible multi-camera and multi-lens configurations. Whether it's a compact huddle space or a fully integrated ProAV room, every participant gets a front-row seat. Powered by Intel Core Ultra processors with 34 TOPS of AI performance, the series is AI-ready and built to scale, unlocking capabilities like intelligent framing, content capture, and seamless integration. ProAV made simple: Yealink AV ONE Traditional AV deployments are often fragmented — requiring multiple vendors, software tools, and complex integrations across audio, video, and control. Yealink AV ONE changes that by delivering a fully integrated platform that brings together premium audio systems, AVoIP encoders and decoders, control systems, and AI-assisted room design — all into one unified AV ecosystem. Whether it's a huddle room, a hybrid divisible space, or a large training center, AV ONE covers the full AV stack and supports seamless integration with Microsoft Teams Rooms. With Yealink AV ONE, integrators can automatically generate system topologies, device layouts, signal paths, and control interfaces in just minutes —delivering professional-grade, scalable, and truly intelligent AV experiences over a single IP network, managed through one centralized platform. Personal collaboration reimagined For IP Phones, Yealink will roll out new deskphone SIP T7 and T8 series, designed with a new fashion look, AI-empowered audio, leading performance and security, redesigned UI, enhanced connectivity and more sustainability. From a traditional calling device to a modern, UC-driven companion device, Yealink sip phone, is still refreshing itself and showing capabilities of reinnovating, as a veteran in IP communications. Also introduced were new AI-driven devices — UH4X USB Wired Headsets, WH6X ANC DECT Wireless Headset, BH7X Bluetooth Headset, and SP9X USB Speakerphone — forming Yealink's fully AI-powered audio lineup that enhances AI tool performance and boosts workplace efficiency. From boardrooms to desktops, the AI YOUR WORKSPACE launch event showcases Yealink's vision of a one-stop, future-ready collaboration platform—driven by intelligence, simplicity, and seamless integration across all workspaces.

Lantronix & Aerora partner to advance AI-powered drone systems
Lantronix & Aerora partner to advance AI-powered drone systems

Techday NZ

time20-06-2025

  • Techday NZ

Lantronix & Aerora partner to advance AI-powered drone systems

Lantronix has entered into a new collaboration with Aerora to drive development in Edge AI-driven solutions for drones, robotics, and surveillance applications targeting multiple industry sectors. The relationship pairs Aerora's OEM platform for AI-powered visual navigation systems with Lantronix's Open-Q System-on-Module (SoM), which leverages Qualcomm Technologies chipsets. This aims to deliver processing power for situational awareness, computational imaging, and real-time decision-making in high-demand scenarios. Market opportunity The global drone market is anticipated to reach $163.6 billion by 2030, with a compound annual growth rate of 15% according to Grandview Research. Segments such as logistics, agriculture, infrastructure and public safety are identified as the principal drivers for this expansion. The U.S. Federal Government has affirmed the importance of unmanned aircraft systems for commercial and government uses, offering support for manufacturers in the field. Lantronix and Aerora are responding to increasing commercial and governmental demand for advanced drone solutions, especially where AI-driven features can offer new capabilities. The partnership is positioned to speed up technological advancement while helping original equipment manufacturers (OEMs) face the challenge of shortening development timelines and maintaining quality. Technology integration At the core of the solution is Aerora's platform, which incorporates pre-integration of key system components including the camera, gimbal, gimbal motors, housing, telemetry and interface. Developers can achieve simultaneous 4K video streaming alongside high-resolution thermal video, which expands operational capacity in surveillance, search and rescue, environmental monitoring, and industrial inspections. The integrated package also includes the Teledyne FLIR Hadron 640R module and Prism software, bringing thermal and RGB imaging to new levels of precision and reliability. Such technologies address the pressure for faster time-to-market and reduced engineering overhead in a competitive sector. "Lantronix's collaboration with Aerora promises to advance the development of AI-powered drones and other intelligent applications, equipping developers with cutting-edge tools from leading embedded compute technologies," said Saleel Awsare, CEO and president of Lantronix. "This breakthrough in advanced AI-driven solutions delivers a transformative impact, opening doors to new opportunities in both private and government sectors." Lantronix's Open-Q SoMs serve as the foundation for the AI systems, providing reliability for drone, robotics and surveillance application developers as they seek to deploy Edge AI while retaining access to embedded compute expertise. OEM focus and compliance Aerora's technology has an emphasis on achieving NDAA (National Defense Authorization Act) compliance, a critical standard for many government and defence customers in the United States. By merging capabilities from Lantronix, Qualcomm Technologies, and Teledyne FLIR, Aerora intends to offer a solution that is not only flexible but scalable for manufacturing at high volumes. "At Aerora, our core mission is to deliver rapid integration, flexible sensor solutions and fully NDAA-compliant manufacturing at scale. By collaborating closely with industry leaders like Lantronix and Qualcomm and integrating advanced imaging technologies such as Teledyne FLIR's Hadron 640R, we empower drone OEMs to significantly reduce development timelines, expand their operational capabilities and confidently meet demanding market requirements," said Ghel Ghedh, chief technology officer for Aerora. This ongoing incorporation of AI-powered navigation, imaging, and control systems is anticipated to support OEM drone manufacturers with faster and more reliable product development, meeting both technology and regulatory benchmarks. Through these efforts, both companies seek to support expansion in drone usage across civil, industrial, and public safety sectors, where high standards for imaging and rapid response are increasingly required.

Metlifecare Wins RVA Sustainability Award For 6 Green Star Care Homes
Metlifecare Wins RVA Sustainability Award For 6 Green Star Care Homes

Scoop

time18-06-2025

  • Scoop

Metlifecare Wins RVA Sustainability Award For 6 Green Star Care Homes

Press Release – Metlifecare Achieving 6 Green Star is the highest possible rating and a globally acknowledged benchmark for sustainable design and building practices. Metlifecare is thrilled to announce its win in the 2025 Retirement Village Association (RVA) Sustainability Awards for the Operator-Led Category. This accolade celebrates our ambitious project that has seen all six of our newest residential aged care homes achieve the globally recognised 6 Green Star 'Design' ratings by the New Zealand Green Building Council (NZGBC). Achieving 6 Green Star is the highest possible rating and a globally acknowledged benchmark for sustainable design and building practices. Our achievement in securing the initial 'design' rating for all six care homes sets a new benchmark in the industry and underscores our commitment to creating comfortable, sustainable environments for our residents and employees. 'We are over the moon to receive this industry award,' said Gemma James, Group Design Manager at Metlifecare. 'We've been on this Green Star journey since 2021, and the many lessons learned and insights gained, have become an integral part of how we design, build and operate moving forward.' What is a 6 Green Star care home? A 6 Green Star care home is designed and built to the highest standards of sustainability, focusing on energy efficiency, low carbon design, and the use of sustainable materials. Each building incorporates advanced features such as improved thermal performance, heat recovery ventilation, high levels of natural light, and superior acoustics. Operational benefits include water efficiency measures, solar panel electricity generation and the use of cross-laminated timber structures. Additionally, off-site manufacturing techniques are employed to reduce construction waste and minimise the environmental impact. These features not only reduce the environmental impact but also create a healthier and more comfortable living environment for residents and staff. The project has been a significant investment, with each building costing approximately $20 million, reflecting Metlifecare's commitment to long-term sustainability and operational efficiency. Metlifecare's 'Six x 6 Green Star' project not only sets a new standard for the industry but also highlights the importance of sustainable living options for New Zealand's aging population – now and into the future. 'Our team has worked tirelessly to ensure that our new care homes not only meet but exceed the highest standards of sustainability. 'To achieve one 6 Green Star design rating is a worthy celebration, but to achieve all six is an incredible accomplishment and testament to a total team effort. Thank you to each and every person who helped our vision become a reality,' says Gemma. Established in 1984, Metlifecare is a leading owner and operator of retirement villages, providing rewarding lifestyles and care to more around 7,000 New Zealanders. Metlifecare currently operates 37 villages around New Zealand and employs more than 2,000 staff.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store