
This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (CC BY).
OPINION
Ethical aspects of artificial intelligence
1 Demidov PG Yaroslavl State University (YarSU), Yaroslavl, Russia
2 Yaroslavl Diocese, Yaroslavl, Russia
Correspondence should be addressed: Yuri Yurievich Ierusalimskiy
Svobody St., 29, Yaroslavl, 150003, Russia; ur.ca.rayinu@rayinso
Author contribution: the authors have made an equal contribution to the paper.
Nowadays, artificial intelligence (AI) is mainly developed separately from ethical issues. Ethical risks that can negate the practical usefulness of AI are often not taken into account. It is necessary to analyze the main problems of artificial intelligence development, which, in our opinion, are related not to the technical, but to the humanitarian aspect. It is worth noting that the problems of AI transformations today are a matter of national security of the Russian Federation. Artificial intelligence is primarily high-end technologies. But AI is secondary to values, hence the importance of forming its value-target foundations [1].
WHY THE TERM ‘ARTIFICIAL INTELLIGENCE’ IS CONVENTIONAL
The title of the article “Ethical aspects of artificial intelligence” is conventional, as is the term “artificial intelligence” itself. The point is that the role of a scientist or specialist is still crucial in the course of research or other activities, since AI cannot independently generate new knowledge. From the point of view of specialists in technical sciences, philology, linguistics and just common sense, there is no AI, it’s just algorithms that work according to a specific program. Marketers present a set of algorithms as intelligence. Real intelligence is obtained, among other things, under the influence of perception, but what kind of perception do algorithms, a webcam and a microphone have? The term “artificial intelligence” has a right to exist only because it is now widely used, but it must be remembered that it is so nothing more than a machine algorithm for processing routine operations [2–8].
In an interview with The Argumenty Nedeli in 2024, a leading Russian specialist in the field of multiprocessor computing and control systems, Kalyaev IA shared his thoughts on the nature of AI. He claimed that AI should not be considered as a “superbrain,” since in reality it is only a complex computer program designed to expand people’s cognitive abilities. Kalyaev IA compared AI to a tool, for example, a hammer, which serves rather to enhance human mental capabilities than to be an independent thinking unit [9].
‘If a hammer enhances a person’s physical capabilities, such computer programs enhance a person’s mental capabilities. It mainly happens due to the huge speed of the computer. At the same time, artificial intelligence is not able to create something new, different from what it was trained to do. Artificial intelligence can replace humans where you have not to think, but rather follow instructions. He can’t replace a thinking person who is capable of creating something new. At least now’, noted the scientist [9].
Kalyaev IA also stressed that, despite the apparent ability of computers to “think,” in fact, they do not have their own intelligence and perform only those tasks that were put into them by programmers. He pointed out that computers can’t go beyond the hard-coded instructions. However, according to the academician, the so-called “emergent intelligence” may arise, which will look like behavior of a swarm of insects or an anthill, where collective interaction leads to the emergence of new properties of the system [9].
It is worth emphasizing that there is no intelligence in the term “artificial intelligence” because intelligence is inherent in humans. Intelligence goes hand in hand with a soul and creative thinking. AI is a creation of humans. AI was not created by itself, it did not appear as a result of Divine creation or as a result of evolutionary processes in nature. If God created a man, then a man created AI. Artificial intelligence is primarily a system of algorithms and software. Algorithms are written by a person, according to these algorithms, a person writes a program.
AI solves a question or a problem that is posed by a person. AI is neither cold nor hot, it has no worries, it does not get sick, it does not feel pain, it is not afraid of death. It has no fear at all. AI cannot love. It has no concepts of greed and extravagance. AI lacks all those concepts that are typical of humans. A person sets a goal according to what he needs in life. And he solves it. AI does not set goals for itself due to the fact that it does not need to do that. Since AI is a program created by a person and an algorithm, it solves a problem set by a person and within the framework determined by the person.
Therefore, there are hackers, as well as various groups of programmers who try to beat each other. They display their intelligence, creativity, and implement their programs into existing software. But software created by one person destructs the actions of others, because that person is more skilled and more professional.
AI AND NATIONAL SECURITY ISSUES
While solving problems related to the security of the country, including nuclear safety, AI will do it within the framework set by humans, but it must be borne in mind that AI has no fear of death. If AI receives a message that a rocket launch has occurred, it will respond according to the program. It has neither worries nor doubts. You can include an error element in the program, but AI will evaluate the situation again according to the software that the specialists have assigned to it. If AI decides that there is a high probability of a nuclear strike on the country, it will give the command, and the missiles will fly in response. Could it be a false alarm? In the modern history of mankind, there was already a case in 1983 when an officer of the Soviet Army prevented a possible nuclear catastrophe by questioning the instrument readings on the combat launch of NATO missiles [10]. Modern military experts have no confidence that AI would have acted similarly in a similar case.
At the same time, AI was good at processing data obtained from reconnaissance satellites and drones, analyzing the operational situation, etc. It is difficult to imagine a modern army, intelligence and counterintelligence without the successful and effective use of AI.
AI AND MEDICINE
It is quite possible that in the near future AI will perform the most complex surgeries, which are now performed by outstanding surgeons with a worldwide reputation only. So, according to a media report, GigaChat neural network model developed by Sber that had been trained on 42 GB specialized information database, passed a medical exam to a commission of professors from the National Medical Research Center named after Almazov VA. The neural network underwent the same tests that a student who has completed six courses at a medical university goes through. GigaChat passed the exam in general medicine, after which the graduate is awarded a professional graduate degree (General Practitioner). The exam was taken by a committee of professors of therapy, surgery, obstetrics and gynecology. The oral examination paper contained three situational tasks: therapy, surgery, obstetrics and gynecology. Several questions on each topic were attached to the tasks. The neural network also passed a test of 100 questions and scored 82% with a threshold of 70%. “In the future, the model can become the basis for creating a doctor’s and patient’s assistant,” said Sergei Zhdanov, director of Sberbank’s Center for the Health Industry, assessing the future of neural network development” [11, 12].
But ethical issues arise here as well. This is due to the fact that there is no ethical side to AI. The program of a conditional “artificial surgeon” can include various restrictions, an algorithm of actions that should be followed in a certain case. But an AI surgeon still does not have an inner creative, ethical, or moral principle. The “machine surgeon” will operate without the slightest doubt. It will not worry about the death of a patient. But a doctor’s motto is “Do no harm!” AI can do harm, it has a program for how and what to do. It will perform according to the program, and as a result, a person may die. Thus, AI can become a good assistant, but the main decision will still be made by a doctor.
AI AND THE HUMAN BRAIN
It should be understood that AI is not intelligence, but software that can malfunction from the point of view of ethics, morality, and professionalism. Therefore it must be constantly monitored by a specialist. The software system definitely calculates a large amount of data well and quickly. At the same time, AI calculated that the human body processes 400 billion pieces of information within a second. As each cell sends a signal about itself, the brain reacts. Humans have a huge number of cells, nervous system, venous system, etc. They all send signals, and the brain calculates them all. There is no computer in the world that can calculate as much as our brain calculates. So far, no AI will be able to process such a volume of information.
COMPUTER PROGRAMS AND CHESS
Why do the best chess players lose to the computer? The point is that they lose to the computer in terms of calculating the options. It’s an external action, it’s a game. AI calculates what is outside. And those 400 billion bits are an internal calculation, an internal processing of information. Every second our brain receives 400 billion pieces of information through our senses, whereas consciousness processes only 2,000 pieces of it. But all the other information is also available to our brain. When we start thinking the information over, we just use the available 2000 units and that’s it [13]. But perhaps intuition processes a lot more information. How many times did the first correct decision come to us instantly? Thinking and intuition are all inside a person, and the game, including chess, is outside of a person.
Garry Kasparov, the 13th world chess champion (declared a foreign agent in 2022 and included in the list of extremists and terrorists in 2024), has been playing chess with a computer since 1996. That computer could evaluate up to 200 million positions per second, but Kasparov won the first series of games [14]. Then they created more advanced computers. Other talented chess players appeared, and they lost too. Chess players say that we simply cannot keep in mind as many options as the computer does, so we lose. Chess players are not competing with a computer now. It’s useless as any of the best grandmasters will lose to the computer. But this is an external calculation, it is done purely in a certain direction, and our brain and spinal cord calculate everything related to human vital cells. Therefore, the brain is more powerful than AI. Another thing is that it is designed for life, not for play.
In the mid-1990s, the 13th world chess champion accused IBM of cheating, pointing out that a human helped the computer [15]. A few years later, any current world chess champion began to lose to the computer. But, as noted above, chess is a game, an external action that is not related to a human life. If a person creates a more advanced program or a powerful computer, then this computer will beat another computer. So what is the greatness of AI? It appeared due to the creative activity of a person. A computer can’t create itself. Computers and AI have no motivation.
SELF-DEVELOPMENT OF AI IS A MARKETING PLOY
A number of researchers and specialists who describe AI note its ability to self-develop [16]. In fact, these are just beautiful words and a marketing ploy. There are self-learning systems, but they are trained according to a program laid down by a person. That is, to ensure self-education of a machine, it is necessary to include such a program, a system of self-learning, self-education, and self-development, acting in a certain direction. She can’t embrace everything as a human being. The field of application for artificial intelligence is indicated by a person.
PROBLEMS OF AI DEVELOPMENT
Can AI eventually get out of human control? No, when experts create AI, they program its activities with certain limitations. However, all the nuances are difficult to take into account and anticipate. Let’s consider an example with a conditional program for preservation of the human population. For example, out of 10 people, 8 must die in order for two to remain alive and continue the human race. AI will complete the task and save a man and a woman of reproductive age. If the conditions are changed and it is necessary to keep alive only males or only females, AI will solve this problem. But what if AI thinks that the best representatives of humanity are gay or lesbian? Then it will exterminate all the others.
AI AND LEGAL PROCEEDINGS
It is believed in the society that artificial Intelligence is impartial and that it will judge according to the law. It will not take into account the defendant’s motivations and will simply sort out the crime committed. The main thing is that AI law enforcement complies with the principles of the European Ethical Charter, revised in accordance with the Russian legal tradition [17, 18]. AI will find the necessary article of the Criminal Code and make a decision: the first criminal will be sentenced to three years in prison, the second to seventy-five years in prison, the third to death, the fourth to probation. And everything will be without subjectivity and corruption. AI will judge impartially. Most likely, the program will take into account relapses. But it is much more difficult to consider whether illegal acts were forced or not, who influenced the defendant, whether he repented or not, etc. Is it possible to completely exclude the influence of relatives or accomplices of the defendants on the programmer serving the “judicial program”? It is quite natural that Chucha SY, the lawyer, emphasized: “In the professional judicial environment, introduction of AI, which resolves disputes instead of humans, is officially … categorically discouraged” [17].
CONCLUSION
The crisis of values is the main problem of humanity today. You can get out of crisis through values only. We need axiological messages that would meet the traditions and modernity, and, at the same time, lead to the future. The Russian Empire and the Soviet Union repeatedly saved the world from destruction. In the beginning of the second quarter of the 21st century, Russia can give humanity a new model of development again. Namely, it can combine artificial intelligence with value-based development. The Russian project should combine the technologies of the future, primarily AI, and traditional Russian spiritual and moral values, which can be offered to the whole world as a guideline for breaking the trend of society atomization, and, consequently, its disintegration [1].
When considering the issues of AI, taking into account the ethical side is the most important thing. AI has neither bad nor good intentions, it acts the way programmers write. But if conditions change, bad intentions can seem good and vice versa. For example, if one person kills the other person without reason in peacetime, it is, of course, evil. But what if he is at the frontline, defending his country or saving himself and his family in peacetime? Will AI be able to take this into account? What decision will it make? Will AI be able to take into account that staying at the frontline modifies psychotic behavior of fighters? They can be aggressive, not always adequate, and they need many months of rehabilitation. Some demobilized soldiers have been in the habit of squatting for a long time after hearing a sharp, loud sound, because such a habit can save lives at the front.
Humanity can disappear if a living creature will be replaced by a dead one. AI is not alive. A human personality is destructed when they can’t love and be loved any longer. When only AI and machines remain on Earth, a human life will cease, and there will be no people left. And this is the main ethical issue of AI.