IBM has recently introduced a set of artificial intelligence tools aimed at preventing large language models from producing inappropriate or offensive content. Called WatsonX, this AI software platform was unveiled earlier this year to enable the creation of generative AI applications. As a part of the suite, IBM announced the inclusion of governance tools that would ensure the responsible use of these models. Excitingly, now these tools are ready to be implemented.
To be more specific, IBM has announced that its WatsonX.governance software tool will be available for general use starting from December 5th. This platform will assist companies in utilizing large AI models by ensuring their outcomes remain unbiased, factually accurate, and transparent. Through WatsonX.governance, businesses can gain a deeper understanding of the data driving the AI models and ultimately eliminate any ambiguity in the results.
Initially, this software will work with IBM's own LLMs (large language models) and also those stored in the repository maintained by HuggingFace, an AI startup. Additionally, it will be compatible with Meta's Llama 2 model. In the near future, IBM plans to expand its compatibility with other models as well.
With the introduction of WatsonX.governance, IBM aims to provide businesses with the tools necessary to maintain the integrity of their AI applications. By shedding light on the inner workings of these models and ensuring responsible usage, companies can build trust and confidence in their AI-powered solutions like never before.
IBM Introduces new Governance Software for AI Models
IBM has recently unveiled new governance software aimed at tackling the concerns raised by a recent executive order from the White House. The software is designed to help customers manage risk, enhance transparency, and anticipate compliance with future AI-focused regulations.
According to IBM, one of the challenges faced by companies utilizing AI models is the reliance on data from unreliable sources on the internet. This lack of validation often leads to concerns over fairness and accuracy. The new software seeks to address this issue by providing tools to automate AI governance processes, monitor models, and take corrective action when necessary.
In response to the White House's executive order, which aims to enhance government scrutiny of AI systems, IBM believes that its governance tools can help meet the requirements laid out in the order. The company sees this as a significant opportunity for software that can monitor the output from AI models and contribute to overall AI safety.
Kareem Yusuf, a Senior Vice President at IBM, emphasizes the growing interest among company boards and CEOs in leveraging more powerful AI models. However, concerns about transparency and the ability to govern these models have been holding them back. The new governance software aims to address these concerns by providing increased visibility and the ability to translate regulations into enforceable policies.
As new AI regulations continue to emerge globally, IBM believes that its software's capabilities will become even more critical for organizations seeking to navigate the regulatory landscape effectively.
Overall, IBM's new governance software offers promising solutions for businesses aiming to harness the power of AI models while mitigating risks and complying with evolving regulations.
Eric J. Savitz contributed to this report.