International Scientific Report on the Safety of Advanced AI: Key Information

post by Aryeh Englander (alenglander) · 2024-05-18T01:45:10.194Z · LW · GW · 0 comments

Contents

  1 Introduction
  2 Capabilities
    2.1 How does General-Purpose AI gain its capabilities?
    2.2 What current general-purpose AI systems are capable of
    2.3 Recent trends in capabilities and their drivers
    2.4 Capability progress in coming years
  3 Methodology to assess and understand general-purpose AI systems
  4 Risks
    4.1 Malicious use risks
      4.1.1 Harm to individuals through fake content
      4.1.2 Disinformation and manipulation of public opinion
      4.1.3 Cyber offence
      4.1.4 Dual use science risks
    4.2 Risks from malfunctions
      4.2.1 Risks from product functionality issues
      4.2.2 Risks from bias and underrepresentation
      4.2.3 Loss of control
    4.3 Systemic risks
      4.3.1 Labour market risks
      4.3.2 Global AI divide
      4.3.3 Market concentration risks and single points of failure
      4.3.4 Risks to the environment
      4.3.5 Risks to privacy
      4.3.6 Copyright infringement
    4.4 Cross-cutting risk factors
      4.4.1 Cross-cutting technical risk factors
      4.4.2 Cross-cutting societal risk factors
  5 Technical approaches to mitigate risks
    5.1 Risk management and safety engineering
    5.2 Training more trustworthy models
    5.3 Monitoring and intervention
    5.4 Technical approaches to fairness and representation in general-purpose AI systems
    5.5 Privacy methods for general-purpose AI systems
None
No comments

I thought that the recently released International Scientific Report on the Safety of Advanced AI seemed like a pretty good summary of the state of the field on AI risks, in addition to being about as close to a statement of expert consensus as we're likely to get at this point. I noticed that each section of the report has a useful "Key Information" bit with a bunch of bullet points summarizing that section.

So for my own use as well as perhaps the use of others, and because I like bullet-point summaries, I've copy-pasted all the "Key Information" lists here.

1 Introduction

[Bullet points taken from the “About this report” part of the Executive Summary]

2 Capabilities

2.1 How does General-Purpose AI gain its capabilities?


 

There are various types of general-purpose AI. Examples of general-purpose AI models include:

2.2 What current general-purpose AI systems are capable of

2.4 Capability progress in coming years

3 Methodology to assess and understand general-purpose AI systems

4 Risks

4.1 Malicious use risks

4.1.1 Harm to individuals through fake content

4.1.2 Disinformation and manipulation of public opinion

4.1.3 Cyber offence

4.1.4 Dual use science risks

4.2 Risks from malfunctions

4.2.1 Risks from product functionality issues

4.2.2 Risks from bias and underrepresentation

4.2.3 Loss of control

4.3 Systemic risks

4.3.1 Labour market risks

4.3.2 Global AI divide

4.3.3 Market concentration risks and single points of failure

4.3.4 Risks to the environment

4.3.5 Risks to privacy

4.4 Cross-cutting risk factors

4.4.1 Cross-cutting technical risk factors

This section covers seven cross-cutting technical risk factors – technical factors that each contribute to many general-purpose AI risks.

  1. General-purpose AI systems can be applied in many ways and contexts, making it hard to test and assure their trustworthiness across all realistic use-cases.
  2. General-purpose AI developers have a highly limited understanding of how general-purpose AI models and systems function internally to achieve the capabilities they output.
  3. General-purpose AI systems can act in accordance with unintended goals, leading to potentially harmful outputs, despite testing and mitigation efforts by AI developers.
  4. A general-purpose AI system can be rapidly deployed to very large numbers of users, so if a faulty system is deployed at scale, resulting harm could be rapid and global.
  5. Currently, risk assessment and evaluation methods for general-purpose AI systems are immature and can require significant effort, time, resources, and expertise.
  6. Despite attempting to debug and diagnose, developers are not able to prevent overtly harmful behaviours across all circumstances in which general-purpose AI systems are used.
  7. Some developers are working to create general-purpose AI systems that can act with increasing autonomy, which could increase the risks by enabling more widespread applications of general-purpose AI systems with less human oversight.

4.4.2 Cross-cutting societal risk factors

This section covers four cross-cutting societal risk factors – non-technical aspects of generalpurpose AI development and deployment that each contribute to many risks from generalpurpose AI:

  1. AI developers competing for market share may have limited incentives to invest in mitigating risks.
  2. As general-purpose AI advances rapidly, regulatory or enforcement efforts can struggle to keep pace.
  3. Lack of transparency makes liability harder to determine, potentially hindering governance and enforcement.
  4. It is very difficult to track how general-purpose AI models and systems are trained, deployed and used.

5 Technical approaches to mitigate risks

5.1 Risk management and safety engineering

5.2 Training more trustworthy models

5.3 Monitoring and intervention

5.4 Technical approaches to fairness and representation in general-purpose AI systems

5.5 Privacy methods for general-purpose AI systems

0 comments

Comments sorted by top scores.