Best deal of the year: Get 2 years of exclusive tech scoops and reporting for just $299/year. Deal ends Friday.Lock in and Save 25%

The Information
Sign inSubscribe

    Data Tools

    • About Pro
    • The Next GPs 2025
    • The Rising Stars of AI Research
    • Leaders of the AI Shopping Revolution
    • Enterprise Software Startup Takeover List
    • Org Charts
    • Sports Tech Owners Database
    • The Information 50 2025
    • Generative AI Takeover List
    • Generative AI Database
    • AI Chip Database
    • AI Data Center Database
    • Cloud Database
    • Creator Economy Database
    • Tech IPO Tracker
    • Tech Sentiment Tracker
    • Sports Rights Database
    • Tesla Diaspora Database
    • Gigafactory Database
    • Pro Newsletter

    Special Projects

    • The Information 50 Database
    • VC Diversity Index
    • Enterprise Tech Powerlist
  • Org Charts
  • Tech
  • Finance
  • Weekend
  • Events
  • TITV
    • Directory

      Search, find and engage with others who are serious about tech and business.

    • Forum

      Follow and be a part of discussions about tech, finance and media.

    • Brand Partnerships

      Premium advertising opportunities for brands

    • Group Subscriptions

      Team access to our exclusive tech news

    • Newsletters

      Journalists who break and shape the news, in your inbox

    • Video

      Catch up on conversations with global leaders in tech, media and finance

    • Partner Content

      Explore our recent partner collaborations

      XFacebookLinkedInThreadsInstagram
    • Help & Support
    • RSS Feed
    • Careers
  • About Pro
  • The Next GPs 2025
  • The Rising Stars of AI Research
  • Leaders of the AI Shopping Revolution
  • Enterprise Software Startup Takeover List
  • Org Charts
  • Sports Tech Owners Database
  • The Information 50 2025
  • Generative AI Takeover List
  • Generative AI Database
  • AI Chip Database
  • AI Data Center Database
  • Cloud Database
  • Creator Economy Database
  • Tech IPO Tracker
  • Tech Sentiment Tracker
  • Sports Rights Database
  • Tesla Diaspora Database
  • Gigafactory Database
  • Pro Newsletter

SPECIAL PROJECTS

  • The Information 50 Database
  • VC Diversity Index
  • Enterprise Tech Powerlist
Deep Research
TITV
Tech
Finance
Weekend
Events
Newsletters
  • Directory

    Search, find and engage with others who are serious about tech and business.

  • Forum

    Follow and be a part of discussions about tech, finance and media.

  • Brand Partnerships

    Premium advertising opportunities for brands

  • Group Subscriptions

    Team access to our exclusive tech news

  • Newsletters

    Journalists who break and shape the news, in your inbox

  • Video

    Catch up on conversations with global leaders in tech, media and finance

  • Partner Content

    Explore our recent partner collaborations

Subscribe
  • Sign in
  • Search
  • Opinion
  • Venture Capital
  • Artificial Intelligence
  • Startups
  • Market Research
    XFacebookLinkedInThreadsInstagram
  • Help & Support
  • RSS Feed
  • Careers

Answer tough business questions, faster than ever. Ask

Partner Content

AI Summit Recap: IBM’s Gil on How Enterprises Can Get AI Right

AI Summit Recap: IBM’s Gil on How Enterprises Can Get AI RightPhoto Credit: Erin Beach
By
The Information Partnerships
[email protected]Profile and archive

As director of IBM Research and co-chair of the MIT-IBM Watson AI Lab, Dr. Dario Gil oversees one of the world’s largest research institutes. He shapes IBM’s AI vision and, as chair of the National Science Board, has a rare vantage point into the technology’s many evolutions and uses.

As part of The Information’s recent San Francisco AI Summit, Amir Efrati, executive editor of The Information, sat down with Gil to discuss the future of AI, why open AI is better, and what organizations can do to make customizing their models more affordable.

AI’s Small (Model) Future

Not long ago, the narrative around AI was that bigger was better. After OpenAI’s ChatGPT exploded on the scene, tech companies competed for who could pack their large language models with the greatest number of parameters. Gil said that since then the tech world has realized the value of smaller models.

“I’m not of the camp that you can continue to scale compute and data and then nirvana is going to hit. It’s not a simple story of building larger and larger clusters, and that will solve for all of AI’s challenges,” said Gil.

Not only do enterprises not necessarily need the firepower of the largest LLM in the market, Gil noted that inference costs can be prohibitive in the long term. Instead of aiming to train models with 1 trillion to 2 trillion parameters, he said, the new thinking is to train small models with more tokens for specific tasks, which makes AI more affordable to operate.

Why AI Should Be Open

Gil has long been a proponent of open-source AI, arguing that transparency is key in a technology this transformative. He said this is especially true for enterprises looking to train a model on their own proprietary data.

“If you’re going to commingle your crown-jewel data with the data that is used in the AI model, it better be open source, it better be transparent and it better be indemnified to you,” he said.

When Efrati pressed him on the competitive pricing of ChatGPT-4o mini, a closed small model that slashed development costs overnight, Gil noted that training is only part of the AI cost equation, and that enterprises need to think through inference when weighing long-term value.

“The cost structure of inference is a function of the size of the model and how well optimized the stack is to do inference at the lowest possible cost. On both those metrics, open models are doing phenomenally well,” he said.

The Importance of Data Representation

Proprietary data is often the foundation of a company’s competitive advantage. Yet too often, only a tiny portion of that data—1%, by Gil’s estimates—makes its way into the foundational AI model a company uses. If businesses want AI to deliver a true competitive advantage, it is vital that their own data become the backbone of that model.

“I tell my enterprise clients, [if] you take your valuable assets, your data, and in a well-engineered fashion, with the right legal constraints, make sure more of it is represented in foundational models that you will own, it will turn out to be one of your most valuable assets,” said Gil. “The tech world totally gets it, but enterprises are like, ‘How do I do it?’”

Lowering the Cost of Customized Models

So how can enterprises affordably customize their models to increase data representation? Pre-training models from scratch is too time-consuming and cost-prohibitive for most organizations. That leaves two options: retrieval-augmented generation and fine-tuning.

With RAG, organizations create a vector database that stores their private data and lives outside the LLM. Before an LLM returns an answer to a user’s question, it consults this database, grounding the answers in enterprise data. However, RAG isn’t as adaptable as fine-tuning, which allows organizations to adjust the parameters of the model. But because fine-tuning is more specialized, it is also more expensive.

Gil said that thanks to InstructLab, an open-source project launched by IBM and Red Hat, there is a third solution. By allowing people to collaboratively and incrementally add new knowledge and skills to any model, InstructLab lowers the cost of fine-tuning. When enterprises use InstructLab to train smaller models with larger data sets, Gil said they often have better results than even with GPT-4 and Llama.

“You can outperform larger models at a fraction of the cost and build an asset inside your business,” he said.

Most Popular

  • AI AgendaGoogle Unseats Anthropic With Gemini 3
  • AI AgendaWhy OpenAI Should Worry About Google’s Pretraining Prowess
  • DealmakerIn Las Vegas, Kalshi Is King
  • The ElectricThe Electric: Look for Gasoline Cars to be Crowding Roads for Decades Longer

Recommended