Bias and discrimination in AI are serious issues with wide-ranging societal impacts, from perpetuating inequalities to reducing trust in AI systems. The causes are often complex, involving:-
As models scale, they often develop unexpected abilities, such as performing tasks like translation or coding, even without explicit training. These emergent behaviours which may contain biases or discriminate are difficult to explain, as they weren’t an-ticipated by the model's developers, making it nearly impossible to understand how or why the model acquired them.
Biased AI systems can:-
If an AI system is perceived as biased, people will lose trust in its fairness and accuracy.
The challenges in explainability of generative AI systems in particular creates a trust gap. Without clear explanations for the behaviour of the generative AI systems, it becomes difficult to identify and mitigate biases in outputs; ensure accountability, especially in high-stakes domains like healthcare, law, or finance; or satisfy regulatory demands, such as those from GDPR or emerging AI regulations.
This can lead to a reluctance to adopt AI technologies in fields where they could be beneficial, such as healthcare or education.
Tackling these issues for AI systems requires a combination of technical, ethical, and regulatory measures, including:-
AI systems should be transparent and explainable, helping stakeholders understand decisions and detect bias. New methods to make generative AI more explainable are required and are becoming available, such as:
• Post-hoc analysis: Trying to provide explanations after a model has generated its output, though this is still a developing area.
• Model distillation: Simplifying generative models to create more interpretable versions.
• Human-in-the-loop review: Involving (i) bias audits and annotation, where hu-man reviewers analyse the AI system's interactions to identify and correct biased or offensive responses, allowing developers to retrain the system with improved data; and, (ii) Feedback Loops where user feedback on biased responses helps re-fine the AI system’s behaviour and improve its responses over time.
However, the fundamental nature of how generative AI models function makes explainability a highly challenging, if not impossible, goal to fully achieve with current methods.
This innovative AI company is developing AI models that are:-
Generative AI, especially large language models (LLMs), operates as a "black box," where inputs and outputs are visible, but internal processes are opaque. For example, an AI system trained on CVs of predominantly male employees might replicate gender bias, and addressing one bias could inadvertently introduce others. Transparency is cruc
Generative AI, especially large language models (LLMs), operates as a "black box," where inputs and outputs are visible, but internal processes are opaque. For example, an AI system trained on CVs of predominantly male employees might replicate gender bias, and addressing one bias could inadvertently introduce others. Transparency is crucial to understanding which factors influence the output and how. Capisco is transparent and able to explain its reasoning thus allowing flaws to be visible and so addressed.
Where an unwanted bias is detected it should be possible to remove it without having to rebuild the entire system. In the example above Capisco allows the straightforward removal of sex as a factor without rebuilding the entire system.
Corrigibility is the ability to correct both the reasoning and the actions of an AI system. Capisco is corrigible in detail without the need to tear it down and start from scratch
Due to the low data volumes used in training Capisco and the much more efficient mathematics used there are six orders of magnitude difference in terms of energy consumption.
Because it is based on a well founded structured knowledge model Capisco doesn't have any subtle errors built in and so does not suffer from the hallucinations that other methods have built in.
If you are working on, or know of, other pioneering solutions please let us know!
Copyright © 2024 Fit AI - All Rights Reserved.
They are 'essential' cookies required for the website to work and cannot be switched off. You may choose to accept our optional cookies - they analyse website traffic and optimise your website experience and aggregate your data with that of all other users. Please see our Privacy Notice for more information.