
Baidu launches AI video generator, overhauls search features
AI-driven video generator
for businesses as well as a major upgrade to its search engine.
The
image-to-video model
, called
MuseSteamer
, is capable of generating videos up to 10 seconds long and comes in three versions - Turbo, Pro and Lite.
Over the past year, artificial intelligence heavyweights like OpenAI and big global tech companies have been expanding beyond chatbots to text-to-video or image-to-video generators. In China, ByteDance, Tencent and Alibaba have also launched models.
While many rival products, including OpenAI's Sora, target consumers with subscription plans, Baidu's MuseSteamer is aimed only at business users and a consumer app is not yet available.
The search engine overhaul includes a redesigned search box that accepts longer queries and supports voice and image-based searches. The platform also displays more targeted content using Baidu's AI technology.
Baidu has faced increasing pressure as AI-based chatbots such as ByteDance's Doubao and Tencent's Yuanbao become more popular.

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Time of India
an hour ago
- Time of India
Google's most advanced AI video model, Veo 3, is now available in the Middle East
Google's Veo 3 enables Middle East creators to generate hyper-realistic AI videos from text prompts on the Gemini platform./ Image: GoogleArabia/X Google has officially rolled out Veo 3, its latest AI-powered video creation tool, to Gemini users across the Middle East, marking a major expansion in the region for its generative media technology. Now available to paying users of Gemini, Veo 3 allows people to generate cinematic video clips from simple text prompts, complete with synchronized sound, music, dialogue, and realistic visuals. Veo 3: From Prompt to Production Initially unveiled at Google I/O 2024, the company's annual developer conference in May, Veo 3 has quickly drawn attention for its high realism, advanced physics simulation, and precise lip-syncing capabilities. 'From prompt to production, Veo 3 delivers best-in-class realism, physics, and lip syncing,' said Eli Collins, Vice President of Product at Google DeepMind, during the launch event at Google I/O. Users can input a short description of a scene, such as "a foggy street in old Tokyo with neon lights and light rain," and Veo 3 responds by generating an eight-second, 720p video that integrates ambient sound, spoken dialogue, realistic effects, and visual elements that closely mirror the input description. According to Collins, Veo 3 not only performs standard text-to-video generation, but also supports image prompting and introduces a new benchmark in responsive AI video design. He wrote in a blog post: 'Veo 3 excels from text and image prompting to real-world physics and accurate lip syncing,' highlighting its strengths in realism, responsiveness, and user control. by Taboola by Taboola Sponsored Links Sponsored Links Promoted Links Promoted Links You May Like Android tablet - brz, lagan, savršen za sve potrebe! Saznajte više Undo Competing with Sora and Raising the Bar With its ability to generate native audio, including background noise, soundtracks, and voiceovers, Veo 3 positions itself as a direct rival to OpenAI's Sora, which is also in the race to dominate the generative video space. What sets Veo 3 apart is the combination of multi-modal generation (text, image, sound) and its physics-aware rendering, making it capable of crafting scenes that feel lifelike and cinematic, whether it's for a dreamlike short film, a product concept, or even a viral meme. One example that gained popularity on social media involved a surreal AI-generated clip of Will Smith eating spaghetti, which drew attention across platforms like X (formerly Twitter). Transparency Through Watermarking To ensure content authenticity, all Veo 3 videos come with an embedded SynthID watermark, Google's invisible digital signature designed to label AI-generated content. This helps in tracing and verifying the origin of synthetic media, particularly in an era of increasing deepfake risks. In addition to the hidden watermark, Veo-generated videos, except those created by Ultra-tier members using Google's new Flow filmmaking platform, will also carry a visible watermark to clearly indicate that the video was AI-generated. Google is also testing a SynthID Detector tool to help individuals and platforms identify synthetic media with greater ease. Expanding Global Access Having made its debut in other markets earlier this year, Veo 3's launch in the Middle East opens the door for regional creators, filmmakers, marketers, and digital storytellers to explore high-end AI video creation without the need for traditional equipment or editing tools. The rollout is part of Google's broader strategy to integrate generative AI into everyday creative workflows, especially through its Gemini platform, which continues to gain traction among professionals and hobbyists alike.


Time of India
2 hours ago
- Time of India
Google's Veo 3 arrives in the Middle East, bringing hyper-real AI video generation to Gemini users
Google's Veo 3 enables Middle East creators to generate hyper-realistic AI videos from text prompts on the Gemini platform./ Image: GoogleArabia/X Google has officially rolled out Veo 3, its latest AI-powered video creation tool, to Gemini users across the Middle East, marking a major expansion in the region for its generative media technology. Now available to paying users of Gemini, Veo 3 allows people to generate cinematic video clips from simple text prompts, complete with synchronized sound, music, dialogue, and realistic visuals. Veo 3: From Prompt to Production Initially unveiled at Google I/O 2024, the company's annual developer conference in May, Veo 3 has quickly drawn attention for its high realism, advanced physics simulation, and precise lip-syncing capabilities. 'From prompt to production, Veo 3 delivers best-in-class realism, physics, and lip syncing,' said Eli Collins, Vice President of Product at Google DeepMind, during the launch event at Google I/O. Users can input a short description of a scene, such as "a foggy street in old Tokyo with neon lights and light rain," and Veo 3 responds by generating an eight-second, 720p video that integrates ambient sound, spoken dialogue, realistic effects, and visual elements that closely mirror the input description. According to Collins, Veo 3 not only performs standard text-to-video generation, but also supports image prompting and introduces a new benchmark in responsive AI video design. He wrote in a blog post: 'Veo 3 excels from text and image prompting to real-world physics and accurate lip syncing,' highlighting its strengths in realism, responsiveness, and user control. by Taboola by Taboola Sponsored Links Sponsored Links Promoted Links Promoted Links You May Like 21st Century Skills Start with Confident Communication Planet Spark Learn More Undo Competing with Sora and Raising the Bar With its ability to generate native audio, including background noise, soundtracks, and voiceovers, Veo 3 positions itself as a direct rival to OpenAI's Sora, which is also in the race to dominate the generative video space. What sets Veo 3 apart is the combination of multi-modal generation (text, image, sound) and its physics-aware rendering, making it capable of crafting scenes that feel lifelike and cinematic, whether it's for a dreamlike short film, a product concept, or even a viral meme. One example that gained popularity on social media involved a surreal AI-generated clip of Will Smith eating spaghetti, which drew attention across platforms like X (formerly Twitter). Transparency Through Watermarking To ensure content authenticity, all Veo 3 videos come with an embedded SynthID watermark, Google's invisible digital signature designed to label AI-generated content. This helps in tracing and verifying the origin of synthetic media, particularly in an era of increasing deepfake risks. In addition to the hidden watermark, Veo-generated videos, except those created by Ultra-tier members using Google's new Flow filmmaking platform, will also carry a visible watermark to clearly indicate that the video was AI-generated. Google is also testing a SynthID Detector tool to help individuals and platforms identify synthetic media with greater ease. Expanding Global Access Having made its debut in other markets earlier this year, Veo 3's launch in the Middle East opens the door for regional creators, filmmakers, marketers, and digital storytellers to explore high-end AI video creation without the need for traditional equipment or editing tools. The rollout is part of Google's broader strategy to integrate generative AI into everyday creative workflows, especially through its Gemini platform, which continues to gain traction among professionals and hobbyists alike.


Time of India
4 hours ago
- Time of India
What will learning look like in the age of superintelligence? Sam Altman says intelligence may soon cost no more than electricity
OpenAI CEO Sam Altman In his recent blog titled The Gentle Singularity , OpenAI CEO Sam Altman reflects on how the arrival of digital superintelligence may reshape every dimension of human learning. The post is not a speculative essay filled with distant hypotheticals. Instead, it reads like a quiet alert from someone at the very center of what he calls a "takeoff. " One of the most significant areas poised for transformation, according to Altman, is learning itself. As artificial intelligence systems surpass human capability in increasingly complex domains, the role of the learner is expected to evolve. In Altman's view, we are now past the hard part. The breakthroughs behind tools like ChatGPT have already laid the groundwork. What follows is a period where these tools begin to self-improve, causing knowledge creation, experimentation and implementation to accelerate at a pace the world has never seen before. "Already we live with incredible digital intelligence, and after some initial shock, most of us are pretty used to it," Altman writes. That shift in perception is critical, what was once astonishing has quickly become mundane. In education, this means that the bar will keep moving. by Taboola by Taboola Sponsored Links Sponsored Links Promoted Links Promoted Links You May Like Secure Your Child's Future with Strong English Fluency Planet Spark Learn More Undo Learners may no longer be evaluated on their ability to recall information or apply frameworks but rather on their ability to collaborate with machines, interpret insights and define new problems worth solving. Here are six radical shifts Altman's vision suggests we may see in how learning functions in an age of superintelligence: Cognitive agents will become co-learners Altman notes that 2025 marks the arrival of AI agents capable of performing real cognitive work. Writing software, solving novel problems and simulating thought are no longer limited to humans. This doesn't mean the end of learning but a reorientation of it. Students, professionals and educators alike may find themselves working alongside these agents, not as passive users but as active collaborators. The process of learning may increasingly center around guiding, auditing and amplifying the work of intelligent systems. The pace of scientific understanding will compress One of the most profound claims in Altman's blog is that the timeline for scientific discovery could collapse dramatically. "We may be able to discover new computing substrates, better algorithms, and who knows what else," he writes. "If we can do a decade's worth of research in a year, or a month, then the rate of progress will obviously be quite different." This will directly affect how educational systems operate, curricula may have to update monthly instead of yearly. Students might prepare not for known fields but for capabilities that do not yet exist. Personalisation will become the baseline Altman envisions AI systems that feel more like a global brain — "extremely personalized and easy for everyone to use." Such systems could radically alter how learning journeys are shaped. Education may shift away from standardisation and towards deep customisation, where each learner follows a uniquely adaptive path based on their goals, context and feedback loops with intelligent systems. This could also challenge long-held norms around grading, pacing and credentialing. Creativity will remain human, but enhanced Despite machines taking over many cognitive tasks, Altman emphasises that the need for art, storytelling and creative vision will remain. However, the way we express creativity is likely to change. Learners in creative fields will no longer be judged solely by their manual skill or originality but by how well they can prompt, guide and harness generative tools. Those who embrace this shift may open entirely new modes of thought and output. Intelligence will become infrastructural In Altman's projection, 'As datacenter production gets automated, the cost of intelligence should eventually converge to near the cost of electricity.' Once data centers can build other data centers and robots assist in manufacturing robots, the cost of deploying intelligence could plummet. This repositions knowledge from something rare and scarce to something ambient. Learning may become less about access and more about intent, what one chooses to do with the world's near-limitless cognitive resources. The meaning of expertise may change As systems outpace human ability in certain domains, the role of the expert will evolve. According to Altman, many of today's jobs might appear trivial or performative to future generations, just as subsistence farming seems primitive to us now. Yet meaning will remain rooted in context. Learners will continue to pursue mastery, not because the machine cannot do it but because the act of learning remains socially and personally meaningful. The human impulse to know and contribute will not vanish, it will be redirected. Throughout the blog, Altman remains clear-eyed about the challenges. "There will be very hard parts like whole classes of jobs going away," he admits, but he is equally optimistic that the world will become so much richer, so quickly, that new ways of structuring society, policy and education will follow. Learning may become less of a race to gain credentials and more of a lifelong dialogue with intelligent systems that expand what it means to know, to build and to belong. "From a relativistic perspective, the singularity happens bit by bit, and the merge happens slowly," Altman writes. The shift may not feel disruptive day to day but its long arc will redefine how we learn, what we teach and how intelligence itself is understood in the decades to come. Is your child ready for the careers of tomorrow? Enroll now and take advantage of our early bird offer! Spaces are limited.