FAQ
< All Topics

How do some applications of AI harm humanity?

Six areas of harm

Some applications of AI impact what it means to be a human being by diminishing our minds, relationships, freedom and privacy, moral responsibility, our sense of what is real and the dignity of work.

Loss of cognitive acuity

Cognitive acuity is diminished through reliance on decision support systems (e.g. in financial, judicial and medical areas). In addition, these systems incorporate bias and lack transparency in respect of any decision reached. Such systems only deliver a probability. It is naïve to think that such algorithms can be made transparent nor free of bias, given that they are usually stochastic processes, not rule based, and data will always be biased because humans are biased. Decisions that impact individuals and groups must always have human oversight and a right of appeal with such process involving human assessment and judgement only.

Damage to relationships

Relationships are damaged through over engagement and reliance on digital assistants and a drive to create ever more realistic simulations of humanness. Whilst we do not propose an outright ban on such devices, we believe that it should be a requirement that users should always know that they are interacting with an artefact, not a human. There is a need for more empirical research in this area regarding harms. Research should be conducted on methods to ensure that such artefacts do not appear human (e.g. using non-human voices). The evaluation of a user’s emotions, personality and character by AI based artefacts, simulating a dialogue should be banned (e.g. Interviewing systems).

Loss of freedom & privacy

Freedom and privacy is lost through the use of private data and surveillance of citizens, whether by the state or private companies. The use of AI to monitor, track, and identify citizens from facial or other personal attributes is unprecedented in any civilisation and is quite different from the use of other biometrics such as fingerprints. We believe that the European Commission and other governments are well aware of these dangers and action is required urgently to avoid mass surveillance being normalised.

Although not involving AI, the deployment of Covid-19 tracking apps has brought this prospect even closer. We urge an outright ban on the states use of AI based surveillance technologies. An even greater level of surveillance has already been established in the private sector through Big Tech’s use of a user’s browsing data, shopping activity and a host of other data gatherers such as FitBit health monitors. Much of humanity has already lost its freedom and autonomy! GDPR legislation in Europe needs strengthening to avoid the extraction and use of personal data. We believe that the practice of companies providing free services or products in exchange for data should be banned without explicit and informed consent.

Loss of Moral Autonomy

In assigning moral agency to artefacts such as autonomous weapons and self-drive vehicles, humans are loosing their moral agency. This should be banned and it should be a requirement that human decision making is required where life is at risk. In other words, humans should be responsible for moral decisions.

Loss of reality & addiction

We are in danger of losing a sense of what is real and embodied through the over use of Augmented and Virtual Reality systems. Research is needed in this area to provide more empirical evidence and to inform potential health warnings for the use of such devices. There should also be strong human oversight and a cautious approach to application development to protect the harms to humanity that can occur from users losing touch with the real world and real relationships.

Loss of the dignity of work

AI systems, including robotics is already changing the workplace and displacing jobs. There is human dignity in work and that alternative work should be a condition of job replacement by AI systems and robotics, except where such systems preserve life in carrying out hazardous tasks.

Table of Contents