logo
Ian Buck built Nvidia's secret weapon. He may spend the rest of his career defending it.

Ian Buck built Nvidia's secret weapon. He may spend the rest of his career defending it.

Ian Buck, Nvidia's vice president of hyperscale and high-performance computing, felt a twinge of nostalgia during CEO Jensen Huang's keynote presentation at their GTC conference in March.
Huang spent nearly eight minutes on a single slide listing software products.
"This slide is genuinely my favorite," said Huang onstage in front of 17,000 people. "A long time ago — 20 years ago — this slide was all we had," the CEO continued.
Buck said he was instantly transported back to 2004 when he started building the company's breakthrough software, called Compute Unified Device Architecture.
Back then, the team had two people and two libraries. Today, CUDA supports more than 900 libraries and artificial intelligence models. Each library corresponds to an industry using Nvidia technology.
"It is a passionate and very personal slide for me," Buck told reporters the next day.
The 48-year-old's contribution to Nvidia is hard-coded into the company's history. But his influence is just beginning. CUDA is the platform from which Nvidia jumped to, at one point, 90% market share in AI computing. CUDA is how the company defends its moat.
One architecture to rule them all
Dion Harris, Nvidia's senior director of high-performance computing and AI factory solutions, sometimes forgets that he's in the room with the Dr. Ian Buck. Then it hits him that his boss, and friend, is a computing legend.
Since Buck's undergrad days at Princeton in the late 1990s, he had been focused on graphics — a particularly punishing field within computer science with no obvious connection to AI at the time.
"Computer graphics was such a dweebie field," said Stephen Witt, the author of " The Thinking Machine," which details Nvidia's rise from obscurity to the most valuable company in the world.
"There was a stigma to working in computer graphics — you were maybe some kind of man-child if this was your focus," Witt said.
While getting his Ph.D. at Stanford, Buck connected multiple graphics processing units with the aim of stretching them to their limits. He had interned at Nvidia before pursuing his Ph.D., so he was familiar with the GPU. Initially, he used it for graphics like everyone else.
Buck has said that he and his cohort would use the chips to play video games such as "Quake" and "Doom," but eventually, he started asking himself what else his gaming setup could do.
He became fixated on proving that you could use GPUs for anything and everything. He received funding from Nvidia and the Defense Advanced Research Projects Agency, among others, to develop tools to turn a GPU into a general-purpose supercomputing machine. When the company saw Brook—Buck's attempt at a programming language that applied the power of GPUs beyond graphics, Nvidia hired him.
He wasn't alone. John Nickolls, a hardware expert and then director of architecture for GPU computing, was also instrumental in building CUDA. Buck may have been forever paired with Nickolls if the latter had not died from cancer in 2011.
"Both Nickolls and Buck had this obsession with making computers go faster in the way that a Formula 1 mechanic would have an obsession with making the race car go faster," Witt told BI. (The author said Huang expressed frustration that Nickolls doesn't get the recognition he deserves since his passing.)
Buck, Nickolls, and a small team of experts built a framework that allowed developers to use an existing coding language, C, to harness the GPU's ability to run immense calculations simultaneously rather than one at a time and apply it to any field. The result was CUDA, a vehicle to bring parallel computing to the masses.
The rise of CUDA as an essential element in the world of AI wasn't inevitable. Huang insisted on making every chip compatible with the software, though hardly anyone was using it despite being free. In fact, Nvidia lost millions of dollars for more than a decade because of CUDA.
The rest is lore. When ChatGPT launched, Nvidia was already powering the AI computing revolution that is now the focus of $7 trillion in infrastructure spending, much of which eventually goes to Nvidia.
King of the nerds
Buck's smarts do have limits. He joked to an eager audience at GTC that quantum chromodynamics, a field of particle physics, just won't stick in his brain. But he thrives in Nvidia's notoriously rigorous environment.
The Santa Clara, California, company has an intense culture that eschews one-on-one meetings and airs missteps and disagreements in public. It might sound terrifying, but for those with the brains to keep up, the directness and the rigor are ideal.
For this report, Business Insider spoke with four people who have either worked directly with Buck at Stanford, Nvidia, or both.
Those who know Buck personally describe him as gregarious and easygoing but capable of intensity when goals are on the line. He's focused on results rather than theory.
In his public remarks, at panels and in interviews on behalf of Nvidia, Buck volleys from rapid, technical lines of thought to slower, simple descriptions in layman's terms.
At a GTC press conference, he detailed the latest development in convolutional neural networks and then described proteins as "complicated 3D squigglies in your body." He describes the tiny, sensitive interconnects between parts of an Nvidia chipset like the Geek Squad explaining the back of a TV from memory — it's all in his head.
Harris said storytelling ability is particularly important in the upper ranks at Nvidia. Since the company essentially had a promising technology waiting for a market for years, Huang still sees being early as a business strategy. He has branded it "going after zero billion-dollar markets." The potential of AI, "AI factories," and the infrastructure spending that goes with them is a story Nvidia can't stop telling.
Buck's shilling skills have improved over the years. But even in 15-year-old footage, he's most animated when explaining the inner workings of Nvidia's technology.
"A lot of developers are amazing, but they say, 'Leave me alone. I'm going to write my code in the mountains somewhere," Paul Bloch, president of Nvidia partner DDN, told BI. Nvidia's leaders aren't like that, he said. Much of Nvidia's upper echelon may have the skills to match the reclusive set, but they don't choose between showmanship and code.
Ian Buck's next act
Ian Buck's work at Nvidia began with a simple mandate: make the GPU work for every industry. That mission is very nearly accomplished. There are hundreds of CUDA libraries targeting industries from weather forecasting to medical imaging.
"The libraries are really there to connect the dots so that every business doesn't have to learn CUDA," Harris said.
CUDA draws its strength from millions of developers, amassed over decades, who constantly innovate and improve on the platform. So far, no one has caught Nvidia's heels, but the competition is coming faster than ever.
Even as Buck spoke at GTC, developers across the world were trying to break through CUDA's dominance. The first night of the conference, a cast of competitors convened by TensorWave, an AI cloud company exclusively using chips from Nvidia's only US rival, AMD, held an event entitled "Beyond CUDA."
Tensorwave cofounder Jeff Tatarchuck said it included more than "24 presenters talking about what they're doing to overcome the CUDA moat."
AMD, which also presented at the event, is making an explicit effort to catch up on the software side of AI computing, but industry analysts say they're not there yet.
Harris told BI Buck's team spends a lot of time speaking with researchers to stay on top. That's always been true, but the nature of the work has changed. A decade ago, Buck was convincing researchers to apply accelerated computing to their problems; now the tables have turned.
"One of the most challenging parts of my job often is to try to predict the future, but AI is always surprising us," Buck said at a Bank of America conference this month. Understanding what the smartest minds in AI need from Nvidia is paramount.
Many saw DeepSeek, the company that spooked markets with its ostensibly cheap reasoning AI model, as a threat to Nvidia since the team bypassed CUDA to squeeze out the performance gains that allowed it to achieve competitive results with less compute.
But Buck recently described the Chinese team as "one of the best CUDA developers out there."
AI is entering a new phase as more companies commercialize their tools. Even with Nvidia's enormous head start, in part built by Buck, the pace is intense.
For example, one of the products Nvidia debuted at GTC, Dynamo, is an inference computing platform designed to adapt to the rise of reasoning models. Nvidia launched Dynamo a couple of months after the DeepSeek earthquake, but some users had already built their own versions. That's how fast AI is evolving.
"Inference is really hard. It's wickedly hard," said Buck at GTC.
Talent is also a big part of how Nvidia is going to try to maintain its dominance, and another place where, Witt says, Buck has value beyond his technical skills. He's not exactly a household name, even at Stanford.
But for a certain type of developer, the ones who can play in Buck's extremely complex sandbox, he is a draw.
"Everyone's trying to hire these guys, especially after DeepSeek," said Witt. "This was not a sexy domain in computer science for a long time. Now it is white hot."
"Now these guys are going for big money. So, I think Ian Buck has to be out there advertising what his group is doing," Witt continued.
Nvidia declined to make Ian Buck available for an interview with Business Insider and declined to comment on this report.
Who's Nvidia's next CEO?
Buck is more than a decade younger than Huang, who is 62, and doesn't plan on going anywhere anytime soon. Yet, questions about succession are inevitable.
Lip-Bu Tan, a semiconductor industry legend who recently became Intel's CEO, told BI that Buck is one of a handful of true collaborators for Huang, who has more than 60 direct reports.
"Jensen has three right-hand men," Tan told BI before he took over at Intel. Buck is one. Vice President of GPU Engineering Jonah Alben is another. And CFO Colette Kress, though clearly not a man, is the third, Tan said.
Jay Puri, Nvidia's executive vice president for worldwide field operations, and Sanja Fidler, vice president of AI research, are also names that come up in such conversations.
"I don't think Ian spends a lot of time doing business strategy. He's more like the world's best mechanic," Witt said.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Cutting Costs, Driving Progress: Evolving Your Organization With Agentic AI
Cutting Costs, Driving Progress: Evolving Your Organization With Agentic AI

Forbes

time22 minutes ago

  • Forbes

Cutting Costs, Driving Progress: Evolving Your Organization With Agentic AI

President of TELUS Digital Solutions at TELUS Digital, a company dedicated to crafting unique and enduring digital customer experiences. Artificial intelligence (AI) is rapidly moving past generative AI (GenAI) tools like ChatGPT, DALL-E and Gemini toward genuine autonomy. Google, OpenAI and Nvidia are among major technology companies that have recently announced expansions of agentic AI capability. These moves are prompting enterprise leaders across all sectors to consider—and prepare for—a future with agentic AI. By definition, agentic AI has agency or autonomy to make decisions, plan and execute complex, end-to-end tasks. It takes the gains of GenAI, which generates human-like content, into new territory. By automating process optimization and resource allocation, agentic AI's autonomous decision-making capabilities present significant opportunities for businesses to reduce costs through efficiency gains. Future-forward brands looking to harness the true power of AI to achieve business goals and put themselves on the leading edge are already taking note: • A Statista survey commissioned by my firm, TELUS Digital, found more than a third (36%) of businesses plan to allocate more than $4 million apiece to GenAI initiatives in 2025. • Deloitte predicts that 25% of companies currently using GenAI will launch agentic AI pilots or proofs of concept this year, growing to 50% by 2027. • KPMG's survey of 252 business leaders found more than half (57%) plan to invest in or adopt agentic AI in the next six months and a third (34%) plan to within the next 12 months. Even if a brand isn't equipped to take this next step, those looking to remain competitive in the long term should be readying their organizations to meet this transformative moment by laying the groundwork now. The AI Ecosystem: Combining GenAI And Agentic Capabilities GenAI assistants and similar foundational applications have allowed enterprises to create content and software at scale—as the name implies, these tools are focused on generation, namely through creating text, images and code. We've also seen GenAI streamline work processes, boost productivity, enhance search tools across large, complex data repositories and augment multimodal voice AI applications. Many of the models behind these advancements are essentially supercharged chat interfaces that rely on human prompts, instructions and back-and-forth conversations. A core difference of agentic AI is that humans set the goals and the AI assistant decides how to approach the task. In the finance industry, for example: • A GenAI assistant may respond to a customer's or agent's request to change the payment due date on a credit card with step-by-step instructions and a link. • An agentic AI assistant, however, could autonomously analyze a customer's financial profile (spending patterns and account history) and identify and enact cross-selling opportunities. It could proactively initiate a personalized offer, such as a travel rewards credit card with no foreign transaction fees based on observed international spending patterns, complete with pre-calculated savings and a pre-approval application. This autonomous approach could significantly reduce customer acquisition costs while increasing conversion rates through more efficient, targeted engagement. Taking The Next Steps With Agentic AI As enterprises look to expand their AI capabilities from GenAI to include agentic AI, leaders must assess their readiness in terms of both organizational structure and technological infrastructure. Preparing your organization for agentic AI requires both cultural transformation and structured implementation planning. Success depends on fostering a culture of continuous learning, where employees develop new technical and soft skills through accessible training opportunities. Building on this foundation, organizational readiness may include: • Cross-departmental alignment: Establish clear goals and success metrics for your agentic AI strategy across all departments, ensuring measurements reflect both departmental needs and broader organizational objectives. • Strategic prototyping: Identify an initial agentic AI use case with clean, accessible data for proof of concept. Define clear boundaries and limitations to manage expectations and maintain focused development efforts. • Prioritizing use cases: Create an inventory of potential use cases across departments based on business impact, technical feasibility and implementation complexity to develop a phased implementation road map for scaling agentic AI across the organization. • Documenting and assessing workflows: Nothing is more counterproductive than automating a poor workflow, so use this time to document and reassess the workflows themselves. • Establishing governance frameworks: Implement clear ethical guidelines and oversight mechanisms, including a human-in-the-loop approach and regular audits, to ensure agentic AI aligns with organizational values throughout its deployment and scaling. Assessing your organization's AI capabilities is another key stage for agentic AI readiness. At TELUS Digital, we use the AI maturity curve as a framework to understand how advanced or integrated the use of AI is, from early experimentation to full-scale strategic development. Organizations should aim for a point on the curve where AI is being deployed across business functions before proceeding. To assess your technological readiness: 1. List every AI or machine learning project underway—even those still in the early concepting stages—and the data that flows into these systems. 2. Review your tech stack, data integration systems and processing capabilities to ensure they can support agentic AI requirements. 3. Assess your data governance frameworks and quality control processes to confirm they can support autonomous decision-making. 4. Identify and prioritize gaps in your current technical infrastructure that could impact agentic AI implementation, focusing on your highest-priority use cases. This dual focus on organizational and technological readiness ensures a strategic approach to agentic AI adoption that balances innovation with practical implementation. Understand The Future To Prepare For It As enterprises explore how agentic AI can complement their GenAI capabilities, the potential for cost reduction becomes increasingly clear. While GenAI has already demonstrated efficiency gains, agentic AI's capabilities promise even greater savings by independently executing complex tasks and proactively identifying optimization opportunities. However, it's crucial to recognize that agentic AI is one piece of a broader, interconnected AI ecosystem, and its full impact on various industries is still unfolding. The key to realizing agentic AI's potential lies in thoughtful preparation and implementation. Forbes Technology Council is an invitation-only community for world-class CIOs, CTOs and technology executives. Do I qualify?

Foundations Rethink Research Dollars as Funding Is Pulled
Foundations Rethink Research Dollars as Funding Is Pulled

Medscape

timean hour ago

  • Medscape

Foundations Rethink Research Dollars as Funding Is Pulled

This spring the notices came without warning: Federal funding for thousands of approved research projects, many already under way, had been delayed or canceled. Among those left in the lurch were grant recipients seeking to do things like reduce vaccine hesitancy, improve access to healthcare for minority populations, or treat an aggressive form of brain cancer. All told, the National Institutes of Health (NIH) and the National Science Foundation withdrew or placed on hold almost $4 billion in funding. And that money may be gone for the foreseeable future because the White House's proposed budget for the next fiscal year slashes support for those agencies themselves by 40% and 56%, respectively. 'So much has been cut by the wrecking-ball, bludgeoning approach that went on at NIH,' said Alonzo Plough, PhD, chief science officer at the Robert Wood Johnson Foundation (RWJF), Princeton, New Jersey. 'They're not just cutting bureaucrats and funding for projects that are not efficient. These are the pipelines that have put American bioscience at the forefront of the world, provided treatments that saved your family members' lives.' Foundations are trying to meet the moment, doling out emergency support to the extent they can. RWJF, for its part, has several such initiatives, including grants for information gathering on climate change and health, tracking AIDS/HIV, and science-based assessments of childhood vaccines. Still, foundations can't replace the government. 'The billions and billions that have been cut are orders of magnitude bigger than anything philanthropy can patch,' Plough said. 'We can keep certain things on life support, but life support is not a good thing to be on.' Private-Public Partnerships Drive Discovery It won't surprise you to hear that medical and health research requires a lot of money — $245.1 billion in 2020 alone. Industry investment accounts for more than half of that, but corporations generally don't step up until a concept is pretty far along, when they're confident it will pay off. Most early-stage research doesn't qualify. 'Basic, fundamental research may not have an obvious application. Some of it may never lead to one,' said Cynthia Friend, PhD, chief executive officer of the Kavli Foundation, Los Angeles, which helps fund early research. 'When those applications do come forth, the time scale is on average 20 years to end up at something useful.' Support for the long haul, beginning at a stage when researchers may not know the usefulness of their findings, is where the federal government has made a big difference. It normally supplies 25.1% of all medical and health research funding. In comparison, foundations, associations, and societies provide just 1.2%. 'All of science philanthropy together, if you add it up, doesn't come close to the amount of support in the federal budget,' Friend said. The government also plays a large role in building and maintaining complex machinery and large-scale facilities where significant discoveries may be made. Case in point: the synchrotron, a sophisticated x-ray machine used, for example, with CRISPR gene editing technology. The one at Brookhaven National Laboratory costs nearly $1 billion. 'An individual could never make these things or have them,' Friend said. 'But individuals can go use them because they've been built for the overall community.' There's a compelling upside to such spending: Economists estimate that for every dollar invested in research, the US gets $5-$20 in benefits. Foundations Seek New Strategies Although they can't match the government's financial clout, philanthropic organizations are looking for ways to address the extraordinary, nationwide funding gap. When the Council on Foundations surveyed its members in March, 80% of respondents said they were making or considering at least one change to their approach. Many indicated that they were open to ideas like collaborating with other funders, reexamining priorities to address the gaps, and increasing their flexibility in grant-making. 'Philanthropies that care about a particular area have an opportunity,' said Shaady Salehi, co-executive director of the Trust-Based Philanthropy Project. 'What's the range of things they fund as an institution — and what is being defunded? They can step in and support the necessary research.' While many science-focused foundations have yet to comment on how they're dealing with the new reality, some have gone public: The Kavli Foundation is offering bridge funding to early-career scientists who've lost federal support, Friend said. Its relatively small, individual grants are good for up to 2 years. The Spencer Foundation — along with the Kapor Foundation, the William T. Grant Foundation, and the Alfred P. Sloan Foundation — has also made bridge funding available: $25,000 grants to 'address immediate needs following grant cancellations.' The Prebys Foundation has designated $7 million in emergency support for biomedical research in San Diego. The American Association for Cancer Research (AACR) has created new AACR Trailblazer Cancer Research Grants: $15 million to support early-stage and mid-career researchers. The Breast Cancer Research Foundation has announced new grants for early-career investigators facing delays or disruptions, emergency funding for affected projects, and nine new research grants. Funders Emphasize Coordinated Efforts Even before the current crisis, the Kavli Foundation had invited fellow funders to preliminary discussions. That conversation will continue as the foundation works with scientists to shape the projects that receive its support. 'The idea is that if you have more resources to focus in a particular area that you think is important, that will accelerate progress,' Friend said. 'And it will also accelerate if something doesn't work out. We have to be prepared for things not working.' Other foundations are also looking for fresh ways to work together. More than 170 philanthropic organizations of all kinds have signed on to a pledge from the Trust-Based Philanthropy Project. It calls for them to go beyond business as usual — to collaborate, to pool funds, and to be responsive to grantees in a coordinated way. As Salehi sees it, this is essential. 'I think the next level up is going to be a higher level of coordination among private funders, comparing notes on who they're funding, who's not funding, who's being left behind,' she said. 'Where are the gaps?'

CoreWeave stock climbs after company announces $1.5 billion bond sale
CoreWeave stock climbs after company announces $1.5 billion bond sale

CNBC

timean hour ago

  • CNBC

CoreWeave stock climbs after company announces $1.5 billion bond sale

CoreWeave stock rose more than 1% after the renter of artificial intelligence data centers said it plans to sell $1.5 billion worth of bonds. The company said in a release that the notes, due in 2031, will use the capital for general purposes, such as paying off debt. In May, the company announced a $2 billion debt offering plan that sent shares soaring 19%. At the time, CNBC confirmed that the debt was five times oversubscribed. Last week, shares rallied after the company announced a $6 billion AI data center project in Pennsylvania. CoreWeave, whose biggest clients include Nvidia and Microsoft, has more than tripled in share price since its March debut on the Nasdaq. In its IPO prospectus filing, CoreWeave said that it was "one of the largest private debt financings in history and signals the confidence that debt investors have in funding our company to build and scale the next generation AI cloud." Some investors have raised concerns about the company's debt and the sustainability of demand for its products. In May, CEO Michael Intrator defended CoreWeave's spending plans and said it is meeting major client "demand signals."

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store