Back Resolutions
Back Current Initiatives
Back Donate

 

Resolution Details

Company:

Alphabet, Inc.

Year:

2024

Issue Area:

Human Rights & Worker Rights

Focus Area:

AI / Artificial Intelligence, Algorithmic Harm

Status:

Filed

Resolution Text

Whereas: Generative Artificial Intelligence (gAI) threatens to amplify misinformation and disinformation, as exemplified by reports about Bard, Gemini, and other Alphabet AI-driven products, including targeted ads, compromising human rights and democratic processes. This is of particular concern as 2024 will feature critical elections in the United States, India, Mexico, and Russia.

Eurasia Group ranked gAI the third highest political risk confronting the world, warning new technologies “will be a gift to autocrats bent on undermining democracy abroad and stifling dissent at home.”1 Some threats from gAI stem from its generation of inaccurate and invented information in text and images and its ability to accelerate their spread.2 Other threats come from gAI tools that enable precise ad targeting that could propagate disinformation among voters.3

Sam Altman, leading AI executive, said he is “particularly worried that these models could be used for large-scale disinformation.”4 The Information has noted that gAI drops “the cost of generating believable misinformation by several

orders of magnitude.”5 Environmental advocates warn that AI “threatens to amplify the types of climate disinformation that have plagued the social media era.”6 One study found Google’s Palm chat technology created misinformation “hallucinations” at a rate of 27 percent, the highest among AI systems tested.7 Members of the team developing Bard “openly debate the AI tool’s effectiveness and utility, with some questioning whether the enormous resources going into development are worth it.”8 Alphabet has invested an estimated $200 billion in AI over the last decade.9

While Alphabet publicly acknowledges the risks of AI and the need for reliable guardrails,10 it continues to “supercharge”11 gAI product development without addressing the existential threats posed by the technology, undermining Google’s established human rights commitments.12 Researchers at Princeton, Virginia Tech, and Stanford have found that the

guardrails many companies, including Alphabet, rely on to mitigate the risks “aren’t as sturdy as A.I. developers seem to believe.”13 Further, legal experts believe content generated by Alphabet’s own technology is unlikely to be shielded by Section 230 (Communications Decency Act), which has historically provided legal protection when third-party content is posted.

Shareholders are concerned that Alphabet incurs significant legal, financial, and reputational risks because of its rapid development and deployment of gAI products, absent parallel assessments of the threats they pose to the Company and society.

Resolved: Shareholders request the Board issue a report, at reasonable cost, omitting proprietary or legally privileged information, to be published within one year of the Annual Meeting and updated annually thereafter, assessing the risks to the Company’s operations and finances, and to public welfare, presented by the Company’s role in facilitating misinformation and disinformation generated, disseminated, and/or amplified via generative Artificial Intelligence; what steps the Company plans to take to remediate those harms; and how it will measure the effectiveness of such efforts.

1 https://www.eurasiagroup.net/issues/top-risks-2023 

2 https://www.cureus.com/articles/176775-artificial-hallucinations-by-google-bard-think-before-you-leap#!/ 

3 https://www.brookings.edu/articles/how-ai-will-transform-the-2024-elections/ 

4 https://fortune.com/2023/06/08/sam-altman-openai-chatgpt-worries-15-quotes/ 

5 https://www.theinformation.com/articles/what-to-do-about-misinformation-in-the-upcoming-election-cycle 

6 https://epic.org/wp-content/uploads/2023/09/Final-Letter-to-Sen.-Schumer-on-Climate-AI-1.pdf 

7 https://www.nytimes.com/2023/11/06/technology/chatbots-hallucination-rates.html 

8 https://www.bloomberg.com/news/articles/2023-10-11/google-insiders-question-usefulness-of-bard-ai-chatbot

9 https://www.reuters.com/technology/ai-lesson-microsoft-google-spend-money-make-money-2023-07-25 

10 https://blog.google/technology/ai/our-responsible-approach-to-building-guardrails-for-generative-ai/ 11 https://blog.google/products/search/generative-ai-search/ 

12 https://about.google/intl/ALL_us/human-rights/ 

13 https://www.nytimes.com/2023/10/19/technology/guardrails-artificial-intelligence-open-source.html