Defining fairness: How IBM is tackling AI governance

The dangers of AI
(Image credit: Shutterstock)

Enterprises are hesitant to adopt AI solutions due to the difficulty in balancing the cost of governance with the behaviours of large language models (LLM), such as hallucinations, data privacy violations, and the potential for the models to output harmful content.

One of the most difficult challenges facing the adoption of LLM is in specifying to the model what a harmful answer is, but IBM believes it can help improve the situation for firms everywhere.

Speaking at an event in Zurich, Elizabeth Daly, STSM, Research Manager, Interactive AI Group of IBM Research Europe, highlighted that the company is looking to develop AI that developers can trust, noting, “It's easy to measure and quantify clicks, it's not so easy to measure and quantify what is harmful content.”

Detect, Control, Audit

Generic governance policies are not enough to control LLMs, therefore IBM is looking to develop LLMS to use the law, corporate standards and the internal governance of each individual enterprise as a control mechanism - allowing governance to go beyond corporate standards and incorporate the individual ethics and social norms of the country, region or industry it is used in.

These documents can provide context to a LLM, and can be used to ‘reward’ an LLM for remaining relevant to its current task. This allows an innovative level of fine tuning in determining when AI is outputting harmful content that may violate the social norms of a region, and can even allow an AI to detect if it’s own outputs could be identified as harmful.

Moreover, IBM has been meticulous in developing its LLMs on data that is trustworthy, and detects, controls and audits for potential biases at each level, and has implemented detection mechanisms at each stage of the pipeline. This is in stark contrast to off-the-shelf foundation models which are typically trained on biassed data and even if this data is later removed, the biases can still resurface.

The proposed EU AI Act will link the governance of AI with the intentions of its users, and IBM states that usage is a fundamental part of how it will govern its model, as some users may use it’s AI for summarization tasks, and others may use it for classification tasks. Daly states that usage is therefore a “first class citizen” in IBM’s model of governance.

More from TechRadar Pro

Benedict Collins
Staff Writer (Security)

Benedict has been writing about security issues for over 7 years, first focusing on geopolitics and international relations while at the University of Buckingham. During this time he studied BA Politics with Journalism, for which he received a second-class honours (upper division), then continuing his studies at a postgraduate level, achieving a distinction in MA Security, Intelligence and Diplomacy. Upon joining TechRadar Pro as a Staff Writer, Benedict transitioned his focus towards cybersecurity, exploring state-sponsored threat actors, malware, social engineering, and national security. Benedict is also an expert on B2B security products, including firewalls, antivirus, endpoint security, and password management.

Read more
A representative abstraction of artificial intelligence
Enterprises aren’t aligning AI governance and AI security. That’s a real problem
Avast cybersecurity
How to beat ‘shadow AI’ across your organization
An AI face in profile against a digital background.
Navigating transparency, bias, and the human imperative in the age of democratized AI
A scale with AI on one side and a brain on the other
What is AI bias? Almost everything you should know about bias in AI results
Ai tech, businessman show virtual graphic Global Internet connect Chatgpt Chat with AI, Artificial Intelligence.
What companies can learn from the gold rush for the AI boom
A hand reaching out to touch a futuristic rendering of an AI processor.
Balancing innovation and security in an era of intensifying global competition
Latest in Pro
cybersecurity
What's the right type of web hosting for me?
Security padlock and circuit board to protect data
Trust in digital services around the world sees a massive drop as security worries continue
Hacker silhouette working on a laptop with North Korean flag on the background
North Korea unveils new military unit targeting AI attacks
An image of network security icons for a network encircling a digital blue earth.
US government warns agencies to make sure their backups are safe from NAKIVO security issue
Laptop computer displaying logo of WordPress, a free and open-source content management system (CMS)
This top WordPress plugin could be hiding a worrying security flaw, so be on your guard
construction
Building in the digital age: why construction’s future depends on scaling jobsite intelligence
Latest in News
Ray-Ban Meta Smart Glasses
Samsung's rumored smart specs may be launching before the end of 2025
Apple iPhone 16 Review
The latest iPhone 18 leak hints at a major chipset upgrade for all four models
Quordle on a smartphone held in a hand
Quordle hints and answers for Monday, March 24 (game #1155)
NYT Strands homescreen on a mobile phone screen, on a light blue background
NYT Strands hints and answers for Monday, March 24 (game #386)
NYT Connections homescreen on a phone, on a purple background
NYT Connections hints and answers for Monday, March 24 (game #652)
Quordle on a smartphone held in a hand
Quordle hints and answers for Sunday, March 23 (game #1154)