The last few years—even the last few months—have seen artificial intelligence (AI) breakthroughs come at a dizzying pace. AI that can generate paragraphs of text as well as a human, create realistic imagery and video from text, or perform hundreds of different tasks has captured the public’s attention. People see AI’s high level of performance, creative potential and, in some cases, the ability for anyone to use them with little to no technical expertise. This wave of AI is attributable to what are known as foundation models.

What are foundation models?

As the name suggests, foundation models can be the foundation for many kinds of AI systems. Using machine learning techniques, these models apply information learned about one situation to another situation. While the amount of data required is considerably more than the average person needs to transfer understanding from one task to another, the result is relatively similar. For example, once you spend enough time learning how to cook, without too much effort you can figure out how to cook almost any dish, and even invent new ones.

This wave of AI looks to replace the task-specific models that have dominated the landscape. And the potential benefits of foundation models to the economy and society are vast. For example, identifying candidate molecules for novel drugs or identifying suitable materials for new battery technologies requires sophisticated knowledge about chemistry and time-intensive screening and evaluation of different molecules. IBM’s MoLFormer-XL, a foundation model trained on data about 1.1 billion molecules, helps scientists rapidly predict the 3D structure of molecules and infer their physical properties, such as their ability to cross the blood-brain barrier. IBM recently announced a partnership with Moderna to use MoLFormer models to help design better mRNA medicines. IBM also partners with NASA to analyze geospatial satellite data—to better inform efforts to fight climate change—using foundation models.

However, there are also concerns about their potential to cause harm in new or unforeseen ways. Some risks of using foundation models are like those of other kinds of AI, like risks related to bias. But they can also pose new risks and amplify existing risks, such as hallucination, the capability of generation of false yet plausible-seeming content. These concerns are prompting the public and policymakers to question whether existing regulatory frameworks can protect against these potential harms.

What should policymakers do?

Policymakers should take productive steps to address these concerns, recognizing that a risk and context-based approach to AI regulation remains the most effective strategy to minimize the risks of all AI, including those posed by foundation models.

The best way policymakers can meaningfully address concerns related to foundation models is to ensure any AI policy framework is risk-based and appropriately focused on the deployers of AI systems. Read the IBM Policy Lab’s A Policymaker’s Guide to Foundation Models—a new white paper from us, IBM’s Chief Privacy & Trust Officer Christina Montgomery, AI Ethics Global Leader Francesca Rossi, and IBM Policy Lab Senior Fellow Joshua New—to understand why IBM is asking policymakers to:

  1. Promote transparency
  2. Leverage flexible approaches
  3. Differentiate between different kinds of business models
  4. Carefully study emerging risks

Given the incredible benefits of foundation models, effectively protecting the economy and society from its potential risks will help to ensure that the technology is a force for good. Policymakers should swiftly act to better understand and mitigate the risks of foundation models while still ensuring the approach to governing AI remains risk-based and technology neutral. 

Read “A Policymaker’s Guide to Foundation Models”
Was this article helpful?
YesNo

More from Artificial intelligence

AI Bundle for IBM Z and LinuxONE

5 min read - IT leaders have long faced a need to add compute capacity to meet the increased demands from their business. Adoption of mobile technologies and ongoing digital transformation has added to these capacity demands, and IT leaders have been forced to plan for the increasing need for compute infrastructure. We have seen that the explosion in interest and adoption of AI has led IT leaders to revisit their capacity plans. They are seeing the need for increasing compute resources at a scale…

Unlock the value of your Informix data for advanced analytics and AI with watsonx.data

3 min read - Every conversation that starts with AI ends in data. There's an urgent need for businesses to harness their data for advanced analytics and AI for competitive edge. But it’s not as simple as it sounds. Data is exploding, both in volume and in variety. According to IDC, by 2025, stored data will grow 250% across on-premises and cloud storages. With growth comes complexity—multiple data applications, formats and data silos make it harder for organizations to utilize all their data while managing costs. To unlock…

How to prevent prompt injection attacks

8 min read - Large language models (LLMs) may be the biggest technological breakthrough of the decade. They are also vulnerable to prompt injections, a significant security flaw with no apparent fix. As generative AI applications become increasingly ingrained in enterprise IT environments, organizations must find ways to combat this pernicious cyberattack. While researchers have not yet found a way to completely prevent prompt injections, there are ways of mitigating the risk.  What are prompt injection attacks, and why are they a problem? Prompt…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters