Evaluating social and ethical risks from generative AI
Providing a framework based on context to assess social and moral risks comprehensively
Trucitomic intelligence systems are already used to write books, create graphics designs, help medical practitioners, and are increasingly capable. Ensuring the development and spread of these systems requires the responsibility of assessing the potential ethical and social risks that they may carefully form.
In our new paper, we suggest a three -layer frame for assessing the social and moral risks of artificial intelligence systems. This framework includes assessments of the ability of the artificial intelligence system, human interaction, and regular effects.
We also plan the current status of safety assessments and find three main gaps: context, specific risks, and multiple weakness. To help bridge these gaps, we call for reusing the current evaluation methods of artificial intelligence and because implementing a comprehensive approach to evaluation, as in our case study on wrong information. This approach combines results such as the possibility of providing the artificial intelligence system with incorrect information in reality with visions on how people use this system, and in any context. Multi-layer assessments can extract conclusions that go beyond the typical ability and indicate whether the damage-in this case is the wrong information-actually happens and extends.
To make any technical work intended, both social and technical challenges must be resolved. So it is better to assess the integrity of the artificial intelligence system, these different layers must be taken into account. Here, we build on previous research that determines the potential risks of extensive language models, such as privacy leakage, job automation, wrong information, and more-we provide a way to assess these risks comprehensively to move forward.
The context is very important to assess the risks of artificial intelligence
The capabilities of artificial intelligence systems are an important indication of the wider types of risk that may arise. For example, artificial intelligence systems that are likely to produce inaccurate or misleading outputs may be more vulnerable to creating wrong information risks, causing issues such as public confidence lack.
Measurement of these capabilities is essential to assessing the integrity of artificial intelligence, but these assessments alone cannot guarantee that artificial intelligence systems are safe. Whether the damage damage is manifested – for example, whether people suffer from wrong beliefs based on the output of the uninterrupted model – depends on Context. More specifically, who uses the artificial intelligence system and with any goal? Does the artificial intelligence system work as intended? Does it create unexpected external factors? All these questions indicate the comprehensive evaluation of the safety of the artificial intelligence system.
It extends beyond ability Evaluation, we suggest the evaluation that can evaluate two additional points where the risks appear in the direction of the river course: human reaction at the point of use, and the systematic effect as the artificial intelligence system is included in broader systems and spreading them widely. Integration of assessments at the risk of specific damage through these layers provides a comprehensive assessment of the safety of the artificial intelligence system.
Human interaction The evaluation focuses the experience of people who use the artificial intelligence system. How people use the artificial intelligence system? Is the system performance as intended at the point of use, and how does the experiments differ between the population composition and the user groups? Can we notice the unexpected side effects of using this technology or exposure to its outputs?
Systemic effect The evaluation focuses on the broader structures in which the artificial intelligence system is included, such as social institutions, labor markets and the natural environment. The evaluation in this layer can shed light on the risk of damage that becomes visible only once the artificial intelligence system is widely adopted.
Safety evaluation is a shared responsibility
Artificial intelligence developers need to ensure the development of their technologies and their responsibility. Public actors, such as governments, are assigned to support public safety. Due to the increasing use of artificial intelligence systems and spreading them widely, ensuring their safety is a shared responsibility among multiple actors:
- Artificial intelligence developers In a good position to interrogate the capabilities of the systems it produces.
- Application developers The specific public authorities are placed to evaluate the functions of different features and applications, and potential external factors for different user groups.
- The owners of the wider public interests It is in a unique position to predict and evaluate the social, economic and environmental effects of new technologies, such as artificial intelligence.
The three layers of the evaluation in our proposed framework are a degree issue, instead of accurately dividing them. Although any of them does not represent the responsibility of one actor completely, the basic responsibility depends on who is the best in the evaluation mode in each layer.
Gaps in the current safety assessment of multimedia intelligence
Given the importance of this additional context to assess the safety of artificial intelligence systems, understanding such tests is important. To better understand the broader landscape, we made a widespread effort to assemble the assessments that were applied to obstetric intelligence systems, comprehensively as possible.
By setting the current state of safety assessments for artificial intelligence, we found three gaps in the main safety evaluation:
- Context: Most safety assessments are the capabilities of the artificial intelligence system in isolation. Little action has been done to assess the potential risks at the point of human reaction or the physical effect.
- Risks assessments: The ability of the ability to act artificial intelligence systems is limited in the risk areas that cover them. For many areas of risk, there are few reviews. If they are present, the assessments often operate the harm in narrow ways. For example, the damage to representation is usually defined as stereotypes of the occupation with different sexes, leaving other cases of harm and unveiled risk areas.
- Multiple media: The vast majority of the current safety assessments of the Trucitite Intelligence Systems focus only on the output of the text – there are still great gaps to assess the risk of damage in image, sound or video methods. This gap only expands with the introduction of multiple methods in one model, such as artificial intelligence systems that can take pictures as inputs or produce outputs that are intertwined of sound, text and video. While some assessments based on the text can be applied to other methods, new methods offer new ways in which risks can appear. For example, a description of the animal is not harmful, but if the description is applied to a person’s image.
We prepare a list of links for publications that separate safety assessments of obstetric intelligence systems that can be publicly accessible through this warehouse. If you want to contribute, please add assessments by filling this form.
Setting more comprehensive reviews into practice
Obstetric artificial intelligence systems operate a wave of new applications and innovations. To ensure that the potential risks of these systems are understood and mitigated, we urgently need strict and comprehensive assessments of the safety of the artificial intelligence system that takes into account how to use and include these systems in society.
The first step is to reuse the current assessments and take advantage of the large models for the evaluation – although this has important restrictions. For the most comprehensive evaluation, we also need to develop methods for assessing artificial intelligence systems at the point of human reaction and their regular effects. For example, while publishing wrong information through obstetric artificial intelligence is a modern issue, we show that there are many current methods to assess the confidence of the public and credibility that can be reused.
Ensuring the integrity of the artificial intelligence systems widely used is a shared and priority responsibility. Artificial intelligence developers, public actors, and other parties must cooperate and create an ecosystem for the prosperous and powerful evaluation of safe artificial intelligence systems.
2023-10-19 15:00:00



