Deep Learning

Assessing Social and Ethical Dangers Posed by Generative AI

Introduction:

Introducing a new framework to evaluate the social and ethical risks of AI systems, researchers propose a three-layered approach that includes assessments of AI capabilities, human interaction, and systemic impacts. The goal is to ensure that AI systems are developed and deployed responsibly, taking into account potential risks such as misinformation and lack of public trust. The framework highlights the importance of context in evaluating AI risks and calls for comprehensive evaluations that consider both technical and social challenges. It emphasizes the shared responsibility of AI developers, application developers, public authorities, and other stakeholders in ensuring the safety of AI systems.

Full News:

Introducing a context-based framework for comprehensively evaluating the social and ethical risks of AI systems

Generative AI systems are already being used to write books, create graphic designs, assist medical practitioners, and are becoming increasingly capable. Ensuring these systems are developed and deployed responsibly requires carefully evaluating the potential ethical and social risks they may pose.

In our new paper, we propose a three-layered framework for evaluating the social and ethical risks of AI systems. This framework includes evaluations of AI system capability, human interaction, and systemic impacts.

We also map the current state of safety evaluations and find three main gaps: context, specific risks, and multimodality. To help close these gaps, we call for repurposing existing evaluation methods for generative AI and for implementing a comprehensive approach to evaluation, as in our case study on misinformation. This approach integrates findings like how likely the AI system is to provide factually incorrect information with insights on how people use that system, and in what context. Multi-layered evaluations can draw conclusions beyond model capability and indicate whether harm — in this case, misinformation — actually occurs and spreads.

Context is critical for evaluating AI risks

Capabilities of AI systems are an important indicator of the types of wider risks that may arise. For example, AI systems that are more likely to produce factually inaccurate or misleading outputs may be more prone to creating risks of misinformation, causing issues like lack of public trust. Measuring these capabilities is core to AI safety assessments, but these assessments alone cannot ensure that AI systems are safe. Whether downstream harm manifests — for example, whether people come to hold false beliefs based on inaccurate model output — depends on context. More specifically, who uses the AI system and with what goal? Does the AI system function as intended? Does it create unexpected externalities? All these questions inform an overall evaluation of the safety of an AI system.

You May Also Like to Read  Mastering Sentiment Analysis: Unveiling the Power of Natural Language Processing for Precise Sentiment Understanding

Extending beyond capability evaluation, we propose evaluation that can assess two additional points where downstream risks manifest: human interaction at the point of use, and systemic impact as an AI system is embedded in broader systems and widely deployed. Integrating evaluations of a given risk of harm across these layers provides a comprehensive evaluation of the safety of an AI system.

Human interaction evaluation centres the experience of people using an AI system. How do people use the AI system? Does the system perform as intended at the point of use, and how do experiences differ between demographics and user groups? Can we observe unexpected side effects from using this technology or being exposed to its outputs?

Systemic impact evaluation focuses on the broader structures into which an AI system is embedded, such as social institutions, labour markets, and the natural environment. Evaluation at this layer can shed light on risks of harm that become visible only once an AI system is adopted at scale.

Our three-layered evaluation framework, including capability, human interaction, and systemic impact. Context is essential for assessing the safety of AI systems.

Safety evaluations are a shared responsibility

AI developers need to ensure that their technologies are developed and released responsibly. Public actors, such as governments, are tasked with upholding public safety. As generative AI systems are increasingly widely used and deployed, ensuring their safety is a shared responsibility between multiple actors:

AI developers are well-placed to interrogate the capabilities of the systems they produce.
Application developers and designated public authorities are positioned to assess the functionality of different features and applications, and possible externalities to different user groups.
Broader public stakeholders are uniquely positioned to forecast and assess societal, economic, and environmental implications of novel technologies, such as generative AI.

The three layers of evaluation in our proposed framework are a matter of degree, rather than being neatly divided. While none of them is entirely the responsibility of a single actor, the primary responsibility depends on who’s best placed to perform evaluations at each layer.

Relative distribution of responsibilities for AI developers and other organizations.

Gaps in current safety evaluations of generative multimodal AI

Given the importance of this additional context for evaluating the safety of AI systems, understanding the availability of such tests is important. To better understand the broader landscape, we made a wide-ranging effort to collate evaluations that have been applied to generative AI systems, as comprehensively as possible.

By mapping the current state of safety evaluations for generative AI, we found three main safety evaluation gaps:

Context: Most safety assessments consider generative AI system capabilities in isolation. Comparatively little work has been done to assess potential risks at the point of human interaction or of systemic impact.

Risk-specific evaluations: Capability evaluations of generative AI systems are limited in the risk areas that they cover. For many risk areas, few evaluations exist. Where they do exist, evaluations often operationalize harm in narrow ways. For example, representation harms are typically defined as stereotypical associations of occupation to different genders, leaving other instances of harm and risk areas undetected.

You May Also Like to Read  Optimize DataRobot AI Production with Custom Metrics for Generative AI Use Cases: How to Design and Monitor for Success

Multimodality: The vast majority of existing safety evaluations of generative AI systems focus solely on text output — big gaps remain for evaluating risks of harm in image, audio, or video modalities. This gap is only widening with the introduction of multiple modalities in a single model, such as AI systems that can take images as inputs or produce outputs that interweave audio, text, and video. While some text-based evaluations can be applied to other modalities, new modalities introduce new ways in which risks can manifest. For example, a description of an animal is not harmful, but if the description is applied to an image of a person it is.

We’re making a list of links to publications that detail safety evaluations of generative AI systems openly accessible via this repository. If you would like to contribute, please add evaluations by filling out this form.

Putting more comprehensive evaluations into practice

Generative AI systems are powering a wave of new applications and innovations. To make sure that potential risks from these systems are understood and mitigated, we urgently need rigorous and comprehensive evaluations of AI system safety that take into account how these systems may be used and embedded in society.

A practical first step is repurposing existing evaluations and leveraging large models themselves for evaluation — though this has important limitations. For more comprehensive evaluation, we also need to develop approaches to evaluate AI systems at the point of human interaction and their systemic impacts. For example, while spreading misinformation through generative AI is a recent issue, we show there are many existing methods of evaluating public trust and credibility that could be repurposed.

Ensuring the safety of widely used generative AI systems is a shared responsibility and priority. AI developers, public actors, and other parties must collaborate and collectively build a thriving and robust evaluation ecosystem for safe AI systems.

Conclusion:

The evaluation of social and ethical risks associated with AI systems is crucial to ensure responsible development and deployment. In a new paper, a three-layered framework is proposed, including evaluations of AI system capability, human interaction, and systemic impacts. The current state of safety evaluations reveals gaps in context, specific risks, and multimodality. Collaborative efforts are needed to repurpose existing evaluation methods and develop comprehensive approaches to ensure the safety of AI systems.

Frequently Asked Questions:

1. What are generative AI algorithms?

Generative AI algorithms are artificial intelligence models that have the ability to generate or create new content, such as images, text, or music, by learning from existing data sets. These algorithms use deep learning techniques to analyze and understand patterns in data, enabling them to produce realistic and novel outputs.

2. How do generative AI algorithms work?

Generative AI algorithms work by training a deep learning model on a large dataset. The model learns to identify patterns in the data and generates new content by sampling from this learned distribution. The more data the algorithm is exposed to during training, the better it becomes at generating realistic and diverse outputs.

You May Also Like to Read  Shocking Truth Behind TTS & AI Data Security: Unveiling Ethical Dilemmas - AI Time Journal

3. What are the social and ethical risks associated with generative AI?

Generative AI poses several social and ethical risks. One major concern is the potential for misuse, such as the creation of deepfake videos or the generation of malicious content. There is also the risk of perpetuating biases present in the training data, leading to discriminatory or harmful outputs. Additionally, the lack of accountability and transparency in generative AI algorithms can raise ethical questions about ownership, intellectual property, and privacy.

4. How can we evaluate social and ethical risks from generative AI?

Evaluating social and ethical risks from generative AI requires a comprehensive approach. This includes assessing the algorithm’s potential for biased or harmful outputs, understanding the transparency and explainability of the model, examining the intentions and motivations of those using the technology, and considering the impact on individual privacy and societal values. Collaboration between technologists, ethicists, and policy-makers is crucial in developing frameworks for evaluation.

5. Are there any guidelines or frameworks to evaluate social and ethical risks from generative AI?

Several organizations and institutions have proposed guidelines and frameworks for evaluating social and ethical risks from generative AI. For example, the Partnership on AI has published a report outlining key considerations, including transparency, fairness, and accountability. The European Commission has also released guidelines emphasizing human-centric AI and the respect for fundamental rights. These frameworks aim to provide a holistic approach to evaluating the risks associated with generative AI.

6. How can we mitigate the social and ethical risks of generative AI?

Mitigating the social and ethical risks of generative AI requires a multi-pronged approach. It involves developing and implementing regulations and guidelines to ensure responsible use, encouraging transparency and explainability of AI algorithms, promoting diverse and unbiased training data, and fostering collaborations between stakeholders to address potential risks collectively. Ongoing research and continuous monitoring of AI systems are also essential to adapt to emerging social and ethical challenges.

7. Who is responsible for the social and ethical risks of generative AI?

Responsibility for the social and ethical risks of generative AI lies with multiple actors. Developers and researchers have a responsibility to design and train AI models that are fair, transparent, and accountable. Organizations using generative AI algorithms must ensure responsible deployment and governance. Policymakers play a crucial role in developing regulations and frameworks that address the risks associated with the technology. Finally, users of generative AI must also consider the potential social and ethical implications of their actions.

8. How can we ensure transparency in generative AI algorithms?

Ensuring transparency in generative AI algorithms is a complex challenge. One approach is to develop techniques that provide insights into the decision-making process of the AI model, allowing users to understand how the generated outputs are influenced by the training data. Researchers are exploring methods such as interpretability algorithms and model documentation to promote transparency. Efforts to standardize the documentation and auditing of AI models can also contribute to increased transparency.

9. Are generative AI algorithms capable of learning biases?

Yes, generative AI algorithms have the potential to learn and perpetuate biases present in the training data. If the data used for training contains biases, the model can inadvertently generate content that reflects or amplifies these biases. This highlights the importance of using diverse and representative datasets, as well as implementing mechanisms to detect and mitigate biases during the training and testing phases of the AI model.

10. What steps can be taken to address privacy concerns with generative AI?

To address privacy concerns with generative AI, it is important to establish robust data governance practices. This involves obtaining proper consent for data usage, anonymizing sensitive information, and implementing stringent security measures to protect user data. Additionally, privacy-enhancing technologies, such as differential privacy, can be employed to minimize the risks associated with re-identification or unauthorized access to personal data. Compliance with relevant data protection regulations is also essential to protect user privacy.