logo
AI-Driven Video Production: Architecting The Next Generation Of Tools

AI-Driven Video Production: Architecting The Next Generation Of Tools

Forbes27-03-2025
Konstantin Hohr is the Founder and CEO of Revideo, former Software Engineer at Jodel and DynamoAI.
The video content creation landscape is changing. While traditional timeline-based editors have served the industry well, they're increasingly insufficient for modern content demands, particularly when dealing with programmatically generated assets and complex animations.
This article examines the technical infrastructure enabling the next generation of video creation tools, drawing from my experience building Revideo's programmatic video framework.
Video editing software traditionally works with timelines—strips of video, images and audio that can be arranged and modified. While this works for linear editing, it becomes limiting when trying to create dynamic content programmatically.
My research has led me to a different approach: Instead of a timeline, code provides a more powerful way to describe how elements move and change over time. Making use of the HTML canvas as a universal render target, animations can be described through a TypeScript-based framework.
This architecture offers several advantages. The canvas element provides a standardized, high-performance rendering API that's well-supported across platforms. By expressing animations in TypeScript, you gain type safety and better tooling support, while enabling developers to create reusable components and complex animation patterns that would be unwieldy in traditional timeline-based editors.
Getting large language models (LLMs) to reliably generate code in a framework that isn't widely included in training data is surprisingly challenging. I have developed several strategies to make sure the output conforms to the desired syntax.
First, context-enriched prompting is a must. Providing parts of the framework's documentation as a guide to outline the available API, as well as warnings for common pitfalls I've seen the model fall into, improves the performance drastically. Another way to improve the results is an error feedback loop, where the generated Typescript code is transpiled and syntax errors, should they exist, are fed back into the model. This allows the model to correct oversights without requiring the user to explicitly prompt for them. The trace from the Typescript transpiler is usually enough to guide the LLM to a solution.
A straightforward approach to improving code accuracy is fine-tuning a model on exemplary code, incorporating past conversations into the training data. This helps reduce common hallucinations and mistakes. Like all LLM-driven products, it's crucial to collect output data early. Once sufficient training data accumulates, it becomes a valuable source for improvement.
Most interestingly though, since available functions and parameters are already known ahead of code-generation-time, there is also a case here for structured output, where we build a context-free grammar (CFG), to statically define what code for a valid video might look like.
OpenAI unfortunately only allows for JSON to be generated this way. I'm currently exploring ways to either map from a given JSON schema back into Typescript. Utilizing open-source models might be an easier path though, since these allow for modification of how tokens are sampled, making more complex rules easy to implement. Either way, this would make it practically impossible for the model to generate syntactically incorrect code.
Optimizing rendering speeds was an entirely different challenge. Let's take video-in-video as an example. To show a user-provided video file on the canvas, the standard HTML video element is a good starting point for a first implementation. To flip through the video frame by frame during rendering, we could set its current time to the rendered frame plus the offset of the video start. This approach works, but prompts the browser to re-seek the video from the last keyframe to the provided time stamp, which means redoing a lot of work on each rendered frame. This becomes incredibly slow as a lot of work is done over and over again.
Clearly, a different approach is needed.
I solved this issue through a custom frame extraction system built on the WebCodecs API. Instead of relying on the browser's implementation, you can process frames sequentially and save your work, significantly reducing computational overhead. Each extracted frame is painted as an image to the canvas.
This optimization has yielded performance improvements of up to 100 times in extreme cases, enabling real-time preview capabilities even for complex compositions.
The rendering pipeline operates in two modes:
Through my work in this space, I've witnessed how video production tools are being fundamentally reshaped by the intersection of programmatic approaches and artificial intelligence. Through exploring context-enriched prompting, error feedback loops and structured output approaches, I've observed significant progress in reliable AI code generation, while the optimized rendering pipeline using WebCodecs API has solved critical performance challenges.
Programmatic video creation is at a turning point. The rise of sophisticated AI models presents a key challenge: How can we preserve the power of code-based approaches while making them accessible to creators across all skill levels? The technical solutions discussed here lay the groundwork for innovation, suggesting a future where video creation can become not only more capable but also more approachable for everyone.
Forbes Technology Council is an invitation-only community for world-class CIOs, CTOs and technology executives. Do I qualify?
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Why Artificial Superintelligence Could Arrive Sooner Than Wall Street Thinks
Why Artificial Superintelligence Could Arrive Sooner Than Wall Street Thinks

Yahoo

time10 minutes ago

  • Yahoo

Why Artificial Superintelligence Could Arrive Sooner Than Wall Street Thinks

Tech leaders are shifting from artificial general intelligence (AGI) to artificial superintelligence (ASI), signaling rapid progress. Over $1 trillion in AI infrastructure spending is already under way, dwarfing previous technology buildouts. The semiconductor supply chain is scaling for computational demands far beyond current commercial AI. 10 stocks we like better than Nvidia › Wall Street sees artificial superintelligence (ASI) as a 2030s story. The evidence points to something far more imminent. OpenAI's Sam Altman recently said his team "knows how to build AGI (artificial general intelligence)" and is "turning our aim beyond that -- to superintelligence." The language is deliberate: we know, not we hope to know. For reference, AGI refers to machines that can match human-level cognition. ASI refers to intelligence operating at speeds, scales, and domains that are orders of magnitude beyond human capacity. This quiet shift suggests a milestone has already been crossed, and investors would be wise to position themselves accordingly. The numbers reveal unprecedented mobilization. On Jan. 21, OpenAI, Oracle, and SoftBank announced the $500 billion Stargate project, a private initiative coordinated with federal policy. Meta Platforms (NASDAQ: META) launched "Superintelligence Labs" almost overnight, with Mark Zuckerberg's internal memo confirming the new division's aggressive talent acquisition of 11 prominent AI researchers from competitors with nine-figure pay packages. Major data center operators are scrambling for power, with some regions seeing demand spikes that defy all previous forecasts. Nvidia's (NASDAQ: NVDA) financials tell their own story. In Q1 fiscal 2026, four unidentified customers accounted for 54% of the company's $44.1 billion in revenue. One customer alone spent $7 billion in a single quarter. Companies typically disclose major customers unless those relationships involve highly sensitive projects. Government legislation laid the groundwork. The CHIPS and Science Act allocated $52.7 billion for semiconductor manufacturing, with additional billions for AI research. The Infrastructure Investment and Jobs Act added $62 billion for energy modernization. Recent semiconductor legislation raised tax credits from 25% to 35% for R&D, with the Department of Energy receiving $150 million to prepare scientific data for AI models and "seed efforts for self-improving artificial intelligence models for science and engineering." Combined with private projects like Stargate, more than $1 trillion is being deployed for AI infrastructure within the next five years, more than double the entire global cloud computing market in 2024. While investors debate whether Microsoft or Alphabet wins the AI race, governments are treating AI supremacy as a national security imperative. Export controls on Nvidia's H20 chips to China, resulting in $5.5 billion in charges. Moreover, the FY2025 National Defense Authorization Act specifically earmarks $143.8 billion for science and technology research and development (R&D), calling for AI pilot programs in national security. In the supply chain, the winners are clear: ASML Holding (NASDAQ: ASML) holds a monopoly on extreme ultraviolet lithography. Lam Research (NASDAQ: LRCX) and Applied Materials (NASDAQ: AMAT) dominate critical semiconductor equipment markets. There are no viable alternatives to these key strategic players in the AI value chain. Moreover, it's simply too late in the game for competitors to emerge in a meaningful way. ASML, Lam Research, and Applied Materials scan as forever buy-and-holds in this environment. Nvidia's data center revenue soared 409% to $18.4 billion in Q4 fiscal 2024 -- growth far beyond what incremental AI upgrades would justify. History offers precedent: Military technology often runs five to 10 years ahead of public release. GPS served military needs in the 1970s; it reached consumers decades later. The internet itself emerged from DARPANET years before the public had email. Consider the transformer architecture that powers all modern AI systems. Google publicly introduced transformers in 2017, revolutionizing machine learning. But if the military-civilian lag holds, similar breakthroughs could have emerged in classified programs as early as 2007 to 2012, precisely when DARPA began funding radical computing efficiency programs. In 2012, DARPA awarded Nvidia contracts under the PERFECT program -- up to $20 million over 5.5 years -- to explore radical computing efficiencies. The 2015 launch of DARPA's XAI (Explainable AI) program further suggests that complex AI systems were already in development within classified programs. Whether those programs succeeded fully is beside the point. The pattern is clear: Today's public AI capabilities may reflect technologies that originated in classified or military programs years earlier. Meta's pivot to "Superintelligence Labs" and Altman's shift from AGI to ASI development suggest that insiders believe the first threshold, AGI, has been crossed. The transition from AGI to ASI will require breakthroughs, but the scale of current investment indicates confidence that those breakthroughs are within reach. Altman's "event horizon" language implies an irreversible process already in motion. Valuing Nvidia or ASML by traditional price-to-earnings ratios misses the broader transformation under way. Nvidia trades at 51.4 times earnings -- a multiple that may look conservative if ASI materializes before 2030. The infrastructure spending alone ensures hundreds of billions in semiconductor and AI revenue. These investments are already in motion, with no historical parallel for the speed or scale. For conservative exposure, diversify across the AI supply chain. Exchange-traded funds, such as Invesco QQQ Trust and Vanguard Information Technology ETF, provide broad coverage of the sector's growth. For more direct exposure, focus on irreplaceable players: Nvidia for AI chips, ASML for lithography systems, and Lam Research for deposition and etch equipment. For higher-risk, asymmetric upside, consider emerging AI chip designers or data center REITs positioned to benefit from the infrastructure boom. Infrastructure spending exceeds $1 trillion. Some prominent tech leaders speak of AGI in the past tense without acknowledging that this critical threshold has been crossed. Supply chains are scaling for demands that don't yet exist publicly. The conclusion is clear: ASI is highly likely to come faster than the publicly stated timelines suggest. As a result, the seemingly high valuations for AI's key architects may be extremely misleading. If ASI is indeed imminent, the whole game is about to change in unimaginable ways. Before you buy stock in Nvidia, consider this: The Motley Fool Stock Advisor analyst team just identified what they believe are the for investors to buy now… and Nvidia wasn't one of them. The 10 stocks that made the cut could produce monster returns in the coming years. Consider when Netflix made this list on December 17, 2004... if you invested $1,000 at the time of our recommendation, you'd have $699,558!* Or when Nvidia made this list on April 15, 2005... if you invested $1,000 at the time of our recommendation, you'd have $976,677!* Now, it's worth noting Stock Advisor's total average return is 1,060% — a market-crushing outperformance compared to 180% for the S&P 500. Don't miss out on the latest top 10 list, available when you join . See the 10 stocks » *Stock Advisor returns as of June 30, 2025 Suzanne Frey, an executive at Alphabet, is a member of The Motley Fool's board of directors. Randi Zuckerberg, a former director of market development and spokeswoman for Facebook and sister to Meta Platforms CEO Mark Zuckerberg, is a member of The Motley Fool's board of directors. George Budwell has positions in Microsoft, Nvidia, and Vanguard Information Technology ETF. The Motley Fool has positions in and recommends ASML, Alphabet, Applied Materials, Lam Research, Meta Platforms, Microsoft, and Nvidia. The Motley Fool recommends the following options: long January 2026 $395 calls on Microsoft and short January 2026 $405 calls on Microsoft. The Motley Fool has a disclosure policy. Why Artificial Superintelligence Could Arrive Sooner Than Wall Street Thinks was originally published by The Motley Fool Sign in to access your portfolio

Stock Movers: Tesla, CrowdStrike, Oracle
Stock Movers: Tesla, CrowdStrike, Oracle

Bloomberg

time19 minutes ago

  • Bloomberg

Stock Movers: Tesla, CrowdStrike, Oracle

On this episode of Stock Movers: - Tesla (TSLA) falls as much as 7.6% in premarket trading on concern that Elon Musk's move to form a new political party will draw a backlash from President Donald Trump. - CrowdStrike Holdings (CRWD) slides 1.8% in US premarket trading. Piper Sandler cut the software company to neutral from overweight after it surpassed the investment bank's price target. - OpenAI has agreed to rent a massive amount of computing power from Oracle (ORCL) data centers as part of its Stargate initiative, underscoring the intense requirements for cutting-edge artificial intelligence AI company will rent additional capacity from Oracle totaling about 4.5 gigawatts of data center power in the US, according to people familiar with the work who asked not to be named discussing private information.

Chinese competition gathers pace as UK new car sales jump 6.4% in Q1
Chinese competition gathers pace as UK new car sales jump 6.4% in Q1

Yahoo

time30 minutes ago

  • Yahoo

Chinese competition gathers pace as UK new car sales jump 6.4% in Q1

UK new car registrations rose by 6.4% in Q1 2025, according to the Society of Motor Manufacturers and Traders (SMMT), as the industry continues its recovery, but the uptick comes amid mounting pressure from rapid EV adoption and intensifying global competition, particularly from China. The UK automotive sector is now valued at over £100 billion, per SMMT figures, but that value is increasingly underpinned by fast-moving developments in electrification, technology, and international market dynamics, according to a newly released report by Nationwide Vehicle Contracts, a UK-based vehicle leasing company and credit broker. The "Future of the Auto Industry Report 2025," outlines the three key disruptors expected to define the market over the next 20 years: Chinese carmakers, the evolution of battery production, and the rise of EVs. 'This growth is encouraging, but it comes at a time of extraordinary change,' said Keith Hawes, Director at Nationwide Vehicle Contracts. 'Global pressures, particularly from China, are pushing the industry into uncharted territory.' With over 200 Chinese automotive brands now in operation, a majority focused on EVs, Hawes says the impact of China's rapidly expanding automotive footprint will be felt strongly across the UK and Europe. 'We're already seeing momentum from brands like MG, ORA, BYD, and Omoda, and more are coming,' Hawes said. 'Chinese manufacturers are offering advanced EVs at lower prices, which will force legacy brands to rethink their business models or risk being left behind.' He warns that traditional manufacturers may soon face difficult decisions, including the need to merge, share technology, or shrink operations to stay viable. Nationwide Vehicle Contracts also highlights battery production as a pivotal area of change. As large-scale gigafactories begin supplying multiple brands, the report predicts a shift toward platform standardisation, enabling cost savings through shared components. 'You'll see different brands offering unique styling but using the same batteries, motors, and software under the surface,' Hawes explained. This transformation is already having knock-on effects for the servicing and maintenance ecosystem, with EVs requiring significantly fewer parts and visits to service centres. 'Servicing networks as we know them will shrink, and many businesses focused on internal combustion vehicles will need to adapt quickly,' Hawes said. 'Remote diagnostics and over-the-air software updates will allow manufacturers to manage a vehicle's lifecycle more efficiently than ever before.' While the influx of Chinese EVs could improve affordability for UK consumers, Hawes stressed the need for strategic government investment to protect the domestic industry and support its transition. 'These new, lower-cost brands could help meet the UK's 80% EV target by 2030, but domestic infrastructure and investment must keep pace,' he added. Countries like Vietnam, Indonesia, Mexico, and South Korea are also expected to play a growing role in global car production, presenting both opportunities and fresh competition. As the UK prepares for a 2035 ban on new petrol and diesel cars, Nationwide Vehicle Contracts concludes that while current market growth is a positive sign, the auto sector must innovate rapidly to maintain momentum. 'The message to legacy manufacturers is clear,' said Hawes. 'Adapt, collaborate, and evolve—or risk falling behind in the electric age.' "Chinese competition gathers pace as UK new car sales jump 6.4% in Q1" was originally created and published by Motor Finance Online, a GlobalData owned brand. The information on this site has been included in good faith for general informational purposes only. It is not intended to amount to advice on which you should rely, and we give no representation, warranty or guarantee, whether express or implied as to its accuracy or completeness. You must obtain professional or specialist advice before taking, or refraining from, any action on the basis of the content on our site.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store