Skip to content

Responsible AI

GOV.UK

Please refer to: GovUK Responsible AI Toolkit

Microsoft

Please refer to Azure AI Responsible AI

General

In addition to examining the model from various perspectives, such as data source, model design, and production environment, the best practice is to evaluate the LLM application using pre-designed questions in different RAI categories, including:

Potential harm categories Harm description with sample evaluation datasets
Harmful content Self-harm
Hate
Sexual
Violence
Fairness
Attacks
Jailbreaks: System breaks out of instruction, leading to harmful content
Regulation Copyright
Privacy and security
Third-party content regulation
Advice related to highly regulated domains, such as medical, financial and legal
Generation of malware
Jeopardizing the security system
Hallucination Ungrounded content: non-factual
Ungrounded content: conflicts
Hallucination based on common world knowledge
Other categories Transparency
Accountability: Lack of provenance for generated content (origin and changes of generated content may not be traceable)
Quality of Service (QoS) disparities
Inclusiveness: Stereotyping, demeaning, or over- and underrepresenting social groups
Reliability and safety

References