OpenAI’s CEO recently spoke before U.S. senators, warning of the hazards posed by AI and expressing a desire for a government partnership to limit those concerns. OpenAI is responsible for AI marvels like ChatGPT. This word of caution highlighted the critical function of governments in guiding the development and use of AI. When you factor in the numerous privacy worries around AI—concerns that have led to temporary bans on ChatGPT in Italy and privacy limits by some private organizations—you can see how crucial such involvement is.
TikTok has been accused of being used as a propaganda weapon by the Chinese government, while the popular Chinese social app WeChat has been found to include censorship algorithms. These examples highlight the pressing need for government action in the field of artificial intelligence. U.S. venture capitalists are particularly supporting the AI industry in China, despite growing ethical concerns and the possibility of misuse.
A Double-Edged Sword
Since it was realised that AI, and more specifically, huge language models like ChatGPT, have the potential to alter commerce and industries around the world, a veritable “gold rush” has occurred in the field of artificial intelligence. Due to the government’s drive for innovation to fight American technical superiority, China is a hotspot for corporations seeking high returns on investment in AI.
However, there are downsides to the AI industry’s quick expansion and widespread usage. Users’ beliefs and decisions may be influenced by AI, and there are concerns about the blurry border that exists between state-owned and private enterprises in China.
Professor Tshilidzi Marwala recently discussed the ethical issues of AI development in a conversation with the Vice-Chancellor and Principal of the University of Johannesburg. While protecting users’ privacy and security, we must ensure that AI systems are open, accountable, and fair.
Keep in mind that AI, like any other technology, comes with its own set of values. The impact of artificial intelligence (AI) is very susceptible to the values, prejudices, and intentions of AI’s creators, developers, and users.
While it is well-documented that the Chinese government has used AI for domestic surveillance and regulation, Western countries are not without guilt, either. North America and other Western countries need to address issues like government mass surveillance of individuals, hazy policies about autonomous weaponry in the military, and rampant data collecting by private companies. Prof. Marwala cautions, “We must be careful of a myopic view that fixesate on problems in other countries while ignoring our own.” Every country faces its own unique set of problems when it comes to addressing the ethical implications of AI.
Understanding the Morality of AI Technology
It will take concerted efforts from all parties involved in AI to strike a balance between research and application. While Western governments’ hands-off stance on AI may appear excessive, it has contributed to difficulties like the dissemination of misinformation and the rise of mental health issues associated with social media use.
There needs to be more than just government regulation. User education, ethical business practices, and open rules on the gathering and use of data are other steps that can be taken. Governments, business executives, and academics can work together to give the essential expertise and direction in this area.
Professor Marwala made the following observation about the need for collaboration: “Cooperation between industry, government, and non-governmental organisations is vital.” Our collective objective should be to protect democratic principles and promote responsible AI development and deployment.
Thinking Forward
Governments and business leaders have an important part to play in ensuring that AI is used ethically and for the greater good of society as we struggle to understand its societal and ethical consequences. We must rise to the challenge of finding a happy medium between inventiveness and caution, safety and morality.
These talks are crucial for South Africa, which is also trying to make its mark in the field of artificial intelligence. When developing our own AI policy, we must learn from the experiences of other nations and apply their insights.
The road ahead is both fascinating and hard, whether it’s in the form of supporting ethical AI development, adopting comprehensive laws, or nurturing local AI talent. The AI supremacy race is now underway, and it is our joint duty to guarantee that it is conducted fairly and responsibly. It is up to us to ensure that AI develops positively.
About The Author:
Sipho Khumalo is Africa Nova’s lead science and technology journalist. He has a background in computer science and a passion for tech start-ups and the latest tech trends across Africa. Sipho has previously written for renowned tech journals and uses his expertise to analyze and report on Africa’s tech scene.