Why the pope has the ears of G7 leaders on the ethics of AI

Pope Francis is leaning on thinking of Paolo Benanti, a friar adept at explaining how technology can change world

The Guardian has some good analysis on Why the pope has the ears of G7 leaders on the ethics of AI | Artificial intelligence (AI). The PM of Italy, Meloni, invited the pope to address the G7 leaders on the issue of AI. I blogged about this here. It is worth pointing out that this is not the first time the Vatican has intervened on the issue of AI ethics. Here is a short timeline:

  • In 2020 a bunch of Catholic organizations and industry heavyweights sign the Rome Call (Call for AI Ethics). The Archbishop of Canterbury just signed.
  • In 2021 they create the RenAIssance Foundation building on the Rome Call. It’s scientific director is Paolo Benati, a charismatic Franciscan monk, professor, and writer on religion and technology. He is apparently advising both Meloni and the pope and he coined the term “algo-ethics”. Most of his publications are in Italian, but there is an interview in English. He is also apparently on the OECD’s expert panel now.
  • 2022 Benati publishes Human in the Loop: Decisioni umane e intelligenze artificiali (Human in the Loop: Human Decisions and Artificial Intelligences) which is about the importance of ethics to AI and the human in ethics.
  • 2024 Meloni invites the Pope to address the G7 leaders gathered in Italy on AI.

UN launches recommendations for urgent action to curb harm from spread of mis and disinformation and hate speech Global Principles for Information Integrity address risks posed by advances in AI

United Nations, New York, 24 June 2024 – The world must respond to the harm caused by the spread of online hate and lies while robustly upholding human rights, United Nations Secretary- General António Guterres said today at the launch of the United Nations Global Principles for Information Integrity.

The UN has issued a press release announcing that the UN launches recommendations for urgent action to curb harm from spread of mis and disinformation and hate speech Global Principles for Information Integrity address risks posed by advances in AI. This press release marks the launch of the United Nations Global Principles for Information Integrity.

The recommendations in the press release include:

Tech companies should ensure safety and privacy by design in all products, alongside consistent application of policies and resources across countries and languages, with particular attention to the needs of those groups often targeted online. They should elevate crisis response and take measures to support information integrity around elections.

Tech companies should scope business models that do not rely on programmatic advertising and do not prioritize engagement above human rights, privacy, and safety, allowing users greater choice and control over their online experience and personal data.

Advertisers should demand transparency in digital advertising processes from the tech sector to help ensure that ad budgets do not inadvertently fund disinformation or hate or undermine human rights.

Tech companies and AI developers should ensure meaningful transparency and allow researchers and academics access to data while respecting user privacy, commission publicly available independent audits and co-develop industry accountability frameworks.

 

Decker – A HyperCard for the Web

I’m at the CSDH-SCHN conference which is in Montreal. We have relocated to U de Montreal from McGill where Congress is taking place. Jason Boyd gave a paper about the Centre for Digital Humanities at TMU that he directs. He mentioned an authoring environment called Decker that recreates a deck/card based environment similar to what HyperCard was like.

Decker can be used to create visual novels, interactive texts, hypertexts, educational apps, and small games. It has a programming language related to Lua. It has simple graphics tools.

Decker looks really neat and seems to work within a browser as a HTML page. This mean that you can Save As a page and get the development environment locally. All the code and data in a page that can be forked or passed around.

As a lover of HyperCard I am thrilled to see something that replicates its spirit!

Rebind | Read Like Never Before With Experts, AI, & Original Content

Experience the next chapter in reading with Rebind: the first AI-reading platform. Embark on expert-guided journeys through timeless classics.

From a NYTimes story I learned about John Kaag’s new initiative Rebind | Read Like Never Before With Experts, AI, & Original Content. The philosophers Kaag and Clancy Martin have teamed up with an investor to start a company that create AI enhanced “rebindings” of classics. They work with out of copyright book and then pay someone to interpret or comment on the book. The commentary is then used to train an AI with whom you can dialogue as you go through the book. The end result (which I am on the waitlist to try) will be a reading experience enhanced by interpretative videos and chances to interact. It answers Plato’s old critique of text that you can ask questions of it. Now you can.

This reminds me of an experiment by Schwitzgebel, Strasser, and Crosby who created a Daniel Dennett chatbot. Here you can see SChwitzgebel’s reflections on the project.

This project raised ethical issues like whether it was ethical to simulate a living person. In this case they asked for Dennett’s permission and didn’t give people direct access to the chatbot. With the announcements about Apple Intelligence it looks like Apple may provide an AI that is part of the system that will have access to your combined files so as to help with search and to help you talk with yourself. Internal dialogue, of course, is the paradigmatic manifestation of consciousness. Could one import one or two thinkers to have a multi-party dialogue about ones thinking over time … “What do you think Plato; should I write another paper about ethics and technology?”

Surgeon General: Social Media Platforms Need a Health Warning

It’s time for decisive action to protect our young people.

The New York Times is carrying an opinion piece by Vivek H. Murthy, the Surgeon General of the USA arguing that  Social Media Platforms Need a Health Warning. He argues that we have a youth mental health crisis and “social media has emerged as an important contributor.” For this reason he wants social media platforms to carry a warning label similar to cigarettes, something that would take congressional action.

He has more advice in a Social Media and Youth Mental Health advisory including protecting youth from harassment and problematic content. The rhetoric is to give parents support:

There is no seatbelt for parents to click, no helmet to snap in place, no assurance that trusted experts have investigated and ensured that these platforms are safe for our kids. There are just parents and their children, trying to figure it out on their own, pitted against some of the best product engineers and most well-resourced companies in the world.

Social media has gone from a tool of democracy (remember Tahir Square?) to a info-plague in a little over ten years. Just as it is easy to seek salvation in technology, and the platforms encourage such hype, it is also easy to blame it. The Surgeon General’s sort of advice will get broad support, but will anything happen? How long will it take regulation and civil society to box the platforms into civil business? The Surgeon General calls how well we protect our children a “moral test.” Indeed.

Pope to G7: AI is ‘neither objective nor neutral’

Vatican News has a full report on the Pope’s address to the G7 leaders on Artificial Intelligence. In the address the Pope called AI “a true cognitive-industrial revolution” that could lead to “complex epochal transformations”. The full address is available (in various translations) here.

After all, we cannot doubt that the advent of artificial intelligence represents a true cognitive-industrial revolution, which will contribute to the creation of a new social system characterised by complex epochal transformations. For example, artificial intelligence could enable a democratization of access to knowledge, the exponential advancement of scientific research and the possibility of giving demanding and arduous work to machines. Yet at the same time, it could bring with it a greater injustice between advanced and developing nations or between dominant and oppressed social classes, raising the dangerous possibility that a “throwaway culture” be preferred to a “culture of encounter”.

Partecipazione del Santo Padre Francesco al G7 a Borgo Egnazia, 14.06.2024

The Pope makes a number of interesting points starting with a point about how tool making is a looking outward to the environment – a techno-human condition that is part of being human. It is a particular form of openness to the environment that can lead to good or be corrupted which is why ethics are important. “To speak of technology is to speak of what it means to be human and thus of our singular status as beings who possess both freedom and responsibility. This means speaking about ethics.”

The Pope also makes a point that I think Lyotard made in The Postmodern Condition, namely that datafication is limiting our ideas about what knowledge could be. AI could go further and limit our ideas about what it is to think at all. As the Pope says, “We cannot, therefore, conceal the concrete risk, inherent in its fundamental design, that artificial intelligence might limit our worldview to realities expressible in numbers and enclosed in predetermined categories, thereby excluding the contribution of other forms of truth and imposing uniform anthropological, socio-economic and cultural models.”

The Pope concludes by reminding us that we cannot avoid politics and that what we need is a healthy politics capable of creating “the conditions for such good use [of AI] to be possible and fruitful.”

Media Monitoring of the Past · impresso

Leveraging an unprecedented corpus of newspaper and radio archives, **Impresso – Media Monitoring of the Past** is an interdisciplinary research project that uses machine learning to pursue a paradigm shift in the processing, semantic enrichment, representation, exploration, and study of historical media across modalities, time, languages, and national borders.

I just learned about the Swiss project  Impresso: Media Monitoring of the Past. This project has an impressive Web application that lets you search across 76 newspapers in two languages from two countries.

Key to the larger project is using machine learning to handle multiple modalities like:

  • News text and radio broadcasts
  • Text and Images
  • French and German
  • Different countries

A Data Lab that uses IPython is coming soon. They also have documentation about a Topic Modelling tool, but I couldn’t find the actual tool.

Anyway, this strikes me as an example of an advanced multi-modal news research environment.

 

Canadian AI 2024 Conference

I’m at the Canadian AI 2024 Conference where I will be on a panel about “The Future of Responsible AI and AI for Social Good in Canada” on Thursday. This panel is timely given that we seem to be seeing a sea-change in AI regulation. If initially there was a lot of talk about the dangers (to innovation) of regulation, we now have large players like China, the US and the EU introducing regulations.

  • President Biden has issued an “Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence.” It is unlikely that Biden can get legislation through Congress so he is issuing executive orders.
  • The EU recently passed their AI Act which is risk-based.
  • The Canadian AI and Data Act (AIDA) is coming and is similarly risk-based.

In light of AIDA I would imagine that the short-term future for Responsible AI in Canada might include the following:

  • Debate about AIDA and amendments to align it with other jurisdictions and to respond to industry concerns. Will there be a more inclusive consultation?
  • Attempts to better define what are high-impact AIs so as to better anticipate what will need onerous documentation and assessment.
  • Elaboration of how to best run an impact assessment.
  • Discussions around responsibility and how to assign it in different situations.

I hope there will also be a critical exploration of the assumptions and limits of responsible AI.

OpenAI Board Forms Safety and Security Committee

OpenAI has announced the formation of a Safety and Security Committee. See OpenAI Board Forms Safety and Security Committee. This is after Ilya Sutskever and Jan Leike, who were the co-leads of the SuperAlignment project left.

What is notable is that this Committee is an offshoot of the board and has 4 board members on it, including Sam Altman. It sounds to me that the board will keep it on a short leash. I doubt it will have the independence to stand up to the board.

I also note that OpenAI’s ethics is no longer about “superalignment”, but is now about safety and security. With Sutskever and Leike gone they are no longer trying to build an AI to align superintelligences. OpenAI claims that their mission is “to ensure that artificial general intelligence benefits all of humanity” but how will they ensure such beneficence? They no longer have any like an open ethics strategy. Its now just safety and security.

I should add that OpenAI shared a safety update as part of the AI Seoul Summit.

Securing Canada’s AI advantage | Prime Minister of Canada

AI is already unlocking massive growth in industries across the economy. Many Canadians are already feeling the benefits of using AI to work smarter and faster.

The Prime Minister’s office has just announced a large investment in AI. See Securing Canada’s AI advantage | Prime Minister of Canada. This is a pre-budget announcement of $2.4 billion going to AI related things including:

  • 2 billion “to build and provide access to computing capabilities and technological infrastructure for Canada’s world-leading AI researchers, start-ups, and scale-ups”
  • Setting up a “Canadian AI Safety Institute” with $50 million “to further the safe development and deployment of AI”. This sounds like a security rather than ethics institute as it will “help Canada better understand and protect against the risks of advanced or nefarious AI systems, including to specific communities.”
  • Funding for the “enforcement of the Artificial Intelligence and Data Act, with $5.1 million for the Office of the AI and Data Commissioner.”

There are also funds for startups, workers, and businesses.

The massive funding for infrastructure follows a weekend opinion piece in the Globe and Mail (March 21, 2024) on Canada’s AI infrastructure does not compute. The article suggests we have a lot of talent, but don’t have the metal. Well … now we are getting some metal.