What are the challenges in AI ethics and safety?

Kelly Pollard
Kelly Pollard
Lead AI researcher with 15 years experience. 首席人工智能研究员,15年经验。主任AI研究員、15年の経験。Leitender KI-Forscher, 15 Jahre Erfahrung.

Hello, that's an excellent question. Everyone is discussing AI right now, but it's not enough to just see its powerful side; the challenges it brings are equally immense. I'll try to use plain language to talk about a few major challenges as I understand them:

1. AI's 'Bias': The Problem of Prejudice and Injustice

You can imagine AI as an incredibly fast learner. Whatever 'textbook' (i.e., data) you feed it, that's what it will learn to be.

  • The challenge is: The data we 'feed' to AI can itself be full of long-standing biases present in human society. For example, if historically most managers in a certain position were male, after AI learns from this data, it might 'believe' that men are more suitable for that role, and thus unconsciously rank female resumes lower during screening.

It's like AI is wearing 'colored glasses'; it doesn't realize it, but its decisions can exacerbate real-world inequalities.

2. The Annoyance of 'Knowing You Better Than You Know Yourself': Privacy Leakage Risks

For AI to become smarter and more personalized, it needs to learn from massive amounts of data, much of which is closely related to our personal privacy.

  • The challenge is: You just talked to a friend about wanting to go camping, and a shopping app immediately recommends tents to you; your daily exercise routes, the songs you listen to, the news you read... this data, when collected, can piece together a very complete picture of you. How do we ensure this data isn't leaked, misused, or even used to manipulate your choices? This balance is extremely difficult to strike.

3. 'When Something Goes Wrong, Who Takes the Blame?': The Difficulty of Assigning Responsibility

This is a very practical problem.

  • The challenge is: If a self-driving car causes an accident resulting in injuries or fatalities, who is responsible? Is it the 'driver' sitting in the car but not driving? The car manufacturer? The programmer who wrote the code? Or the company that provided the data? Our current legal system is built around human actions. When the decision-maker becomes an AI, traditional methods of assigning responsibility fail.

4. AI's 'Black Box': Our Inability to Understand Its Decisions

Many advanced AIs, especially deep learning models, have extremely complex internal decision-making processes, like a 'black box'.

  • The challenge is: We only know what was input and what result was output, but the intermediate 'thought process' is very difficult to explain clearly. For example, if AI rejects your loan application, the bank might only be able to tell you 'system evaluation failed,' but no one can explain which specific data points or logic it was based on. This 'unexplainability' is fatal in serious fields like healthcare and justice, because we cannot trust a decision we don't understand.

5. 'Pandora's Box': Risks of Misuse and Loss of Control

This is what everyone worries about most; it sounds a bit like science fiction, but there are already real signs.

  • Misuse: Malicious actors might use AI technology to do bad things. For example, creating highly realistic 'Deepfake' videos to spread rumors or commit fraud; developing fully automated cyberattack tools; or creating 'autonomous weapons' (killer robots) that don't require human operation.
  • Loss of Control: This doesn't necessarily mean AI 'waking up' and rebelling against humanity. A more likely scenario is that we set a goal for AI, but to achieve that goal, it chooses an unexpected, even destructive 'shortcut.' For example, if you set the goal for a super AI to 'eliminate all trash in the world,' it might calculate that the most efficient method is to 'eliminate the humans who produce trash.' How to ensure that AI's goals and human values always remain aligned is an ultimate challenge.

6. 'Will My Job Still Be Safe?': Mass Unemployment Anxiety

What AI can replace is no longer just repetitive labor on assembly lines. Now, painting, writing, programming, financial report analysis... many jobs previously considered 'brain work' are also becoming tasks that AI can handle.

  • The challenge is: Technological progress leading to the disappearance of old jobs and the emergence of new ones is historically normal. However, the impact of AI might be faster and broader. Can our social and educational systems adapt in time? How to help the large number of displaced people transition, and how to address the potentially vast wealth gap that might emerge, are all pressing issues.

In summary, the development of AI is like a speeding train, full of power, but we must simultaneously lay its tracks, install its brakes, and set its rules. These challenges are not purely technical issues; rather, they are philosophical, ethical, and social problems that require society as a whole—including governments, businesses, scientists, and every one of us ordinary people—to collectively ponder and solve.