Hostname: page-component-cd9895bd7-fscjk Total loading time: 0 Render date: 2024-12-23T13:40:37.718Z Has data issue: false hasContentIssue false

The United States Unveils Political Declaration on Responsible Military Use of Artificial Intelligence and Autonomy

Published online by Cambridge University Press:  30 October 2023

Rights & Permissions [Opens in a new window]

Abstract

Type
Use of Force, Arms Control, and Non-proliferation
Copyright
Copyright © The Author(s), 2023. Published by Cambridge University Press for The American Society of International Law

For more than a decade, as the technology has become increasingly real, the substantial advantages and real dangers of artificial intelligence (AI)—“the ability of machines to perform tasks that would otherwise require human intelligence,” such as “recognizing patterns, learning from experience, drawing conclusions, making predictions, or taking action”—have preoccupied governments, industry, academics, and non-governmental organizations.Footnote 1 For militaries, the promise of AI is substantial.Footnote 2 As noted by Assistant Secretary of State for Arms Control, Verification, and Compliance Mallory Stewart, “AI-enhanced data analysis could optimize logistics processes, improve decision support, and provide commanders with enhanced situational awareness that enables them to avoid unintended engagements and minimize civilian casualties.”Footnote 3 AI “could increase accuracy and precision in the use of force which can also help strengthen implementation of international humanitarian law's protections for civilians and civilian objects.”Footnote 4 It could also “advance arms control by helping . . . solve complex verification challenges and increasing confidence in states’ adherence to their commitments.”Footnote 5 With these advancements, though, come serious risks. AI generates information, but that information can be “inaccurate, lack[] context, or [be] completely made up.”Footnote 6 Systems can be “poorly designed, inadequately tested, or [have] users [who] do not possess an adequate understanding of the contexts and limitations of those systems.”Footnote 7 Military use of AI could therefore result in errors such as improperly identifying civilians as lawful targets and deploying disproportionate force. Concerns have thus arisen about AI reliability, autonomy, human control, biases, monitoring, and accountability. Accordingly, with the rapid pace of the technology's advancement, the increasing commitment of significant resources to weapons development as militaries race to gain advantage, and the deployment in Ukraine and elsewhere of robotic combat vehicles, unmanned aerial vehicles, and unmanned surface vessels,Footnote 8 the regulation of the military use of AI has gained increased urgency.Footnote 9

The content and form of that regulation are contested, however. The United States has sought to establish agreed-upon non-binding international understandings of AI's proper military application through diplomacy and the articulation and publication of its own positions under the “responsible AI” (RAI) framework. In early 2023, building on prior policies, the U.S. Department of Defense issued an updated directive on Autonomy in Weapons SystemsFootnote 10 and the U.S. Department of State proposed a Political Declaration on Responsible Military Use of Artificial Intelligence and Autonomy.Footnote 11 Many states and non-governmental organizations disagree with the RAI approach taken by the United States and other military powers and have instead called for the negotiation of a legally binding international agreement that would apply and develop the international law of autonomous weapons systems. These debates are part of broader efforts, national and international, to regulate AI.Footnote 12

To date, multilateral consideration of military AI has focused on lethal autonomous weapons systems (LAWS) and has taken place in the meetings of the Group of Governmental Experts on Emerging Technologies in the Area of Lethal Autonomous Weapons Systems (GGE) established in 2016Footnote 13 by the Fifth Review Conference of the High Contracting Parties to the Convention on Certain Conventional Weapons (CCW).Footnote 14 In its early years, the GGE drafted the Guiding Principles on Emerging Technologies in the Area of Lethal Autonomous Weapons Systems,Footnote 15 which were endorsed by the 2019 CCW meeting of the parties.Footnote 16 Since then, the GGE's mandate has centered on “the consideration of proposals and [the] elaborat[ion], by consensus, [of] possible measures . . . and other options related to the normative and operational framework on emerging technologies in the area of lethal autonomous weapon systems.”Footnote 17 Many working papers have been tabled, but no consensus has emerged on fundamental substantive issues, including definitions (even of the term “lethal autonomous weapons systems”). Nor has there been agreement on the form of the outcome—whether to work toward a non-binding text or a binding instrument. While a clear majority of states favor the latter, for years the GGE's consensus requirement has led to modest outcomes.

The United States is among those that oppose a binding agreement,Footnote 18 preferring instead the issuance of statements that apply existing international law to LAWS.Footnote 19 It views the role of GGE as “developing guidance for States—measures that would strengthen the implementation of international humanitarian law (IHL) and promote responsible behavior by States.”Footnote 20 To this end, at the most recent GGE meeting, in May 2023, the United States, together with six other states, proposed “draft articles on the development, deployment, and use of autonomous weapon systems” (covering prohibited weapons, distinction, proportionality, precaution, and accountability).Footnote 21 The proposal was based on a compilation of consensus conclusions and recommendations previously endorsed by the GGE that “seek to clarify the requirements imposed by existing IHL and specify measures to effectively satisfy these requirements.”Footnote 22 The draft articles were not adopted, nor were proposals made by other states and the chair.Footnote 23 Instead, the session endorsed four sets of non-binding conclusions regarding compliance with existing international law, in particular IHL.Footnote 24

Inside and outside the GGE, numerous statesFootnote 25 and non-governmental organizations (many through the Campaign to Stop Killer Robots)Footnote 26 have called for a legally binding instrument to prohibit fully autonomous weapons and to otherwise regulate LAWS. To this end, some are pushing to move negotiations out of the GGE, where that outcome is unlikely, to either a standalone diplomatic conference, as was done with the Anti-Personnel Mine Ban Convention and the Convention on Cluster Munitions, or the UN General Assembly, where the Treaty on the Prohibition of Nuclear Weapons was drafted.Footnote 27 UN human rights rapporteurs have also called for a global prohibition on LAWS.Footnote 28 And in July 2023, UN Secretary-General António Guterres, in his most forward-leaning statement to date, called on states to “conclude, by 2026, a legally binding instrument to prohibit lethal autonomous weapon systems that function without human control or oversight, and which cannot be used in compliance with international humanitarian law, and to regulate all other types of autonomous weapons systems.”Footnote 29

The United States and other countries promoted the contrasting “responsible AI” approach to military AI at the February 2023 summit on Responsible AI in the Military Domain (REAIM) that was co-organized by The Netherlands and South Korea.Footnote 30 Military RAI's origins stem from the U.S. Department of Defense's (DoD) Artificial Intelligence Strategy in 2018.Footnote 31 In 2020, DoD adopted five Ethical Principles for Artificial Intelligence to “ensure the responsible use of AI by the department,” including: responsibility, equity, traceability, reliability, and governability.Footnote 32 Subsequently, DoD implemented the Ethical Principles in a memorandum articulating “foundational tenets”Footnote 33 and in a pathway setting out RAI's operationalization within the department.Footnote 34 Updated DoD Directive 3000.09Footnote 35 established requirements “to minimize the probability and consequences of failures in autonomous and semi-autonomous weapon systems that could lead to unintended engagements.”Footnote 36 Other countries have adopted similar documents.Footnote 37 The North Atlantic Treaty Organization too has promulgated its own Principles of Responsible Use of Artificial Intelligence in Defence, which are substantially similar to DoD's Ethical Principles.Footnote 38

Military RAI has evolved into a set of best practices and techniques for AI's cautious and safe use. Consistent with the U.S. position in the GGE, it attempts to facilitate, not prohibit, military AI's development and use through the establishment of guardrails that minimize risk in accordance with law. RAI's non-binding character provides flexibility for militaries and implies that existing international law regulates AI adequately. As Under Secretary of State for Arms Control and International Security Bonnie Denise Jenkins explained at the REAIM summit, “[t]he United States approaches using artificial intelligence for military purposes from the perspective of responsible speed, meaning our attempts harness the benefits of AI have to be accompanied by a focus on safe and responsible behavior that is consistent with the law of war and international humanitarian law.”Footnote 39

Fifty-seven government representatives at the summit, including all permanent members of the Security Council except Russia (which was not invited to participate), agreed on a “joint call to action on the responsible development, deployment and use of artificial intelligence (AI) in the military domain.”Footnote 40 Among other commitments, the signatories promised “to continu[e] the global dialogue on responsible AI in the military domain in a multi-stakeholder and inclusive manner and call[ed] on all stakeholders to take their responsibility in contributing to international security and stability in accordance with international law.”Footnote 41 The call also “invite[d] states to develop national frameworks, strategies and principles on responsible AI in the military domain.”Footnote 42 There were some notable absences from the list of countries endorsing the call, including Brazil, India, Israel, and South Africa.

Also at the REAIM summit, the State Department unveiled a Political Declaration on Responsible Military Use of Artificial Intelligence and Autonomy.Footnote 43 Internationalizing the Defense Department's and NATO's RAI texts, the Declaration comprises twelve sets of “best practices that the endorsing States believe should be implemented in the development, deployment, and use of military AI capabilities, including those enabling autonomous systems.”Footnote 44 These include: “subjecting systems to rigorous testing and assurance, taking steps to avoid unintended consequences, minimizing unintended biases, and . . . ensuring AI is used in accordance with States’ obligations under international law.”Footnote 45 “The Political Declaration,” a State Department official said in August, “will serve as a launch pad for sustained dialogue among endorsing states.”Footnote 46 To date, however, no endorsing states have been announced. The Declaration did not satisfy those who advocate for an international agreement to govern AI in the military domain.Footnote 47 A Stop Killer Roberts official described the Political Declaration as “the most backwards position seen from any state, in years” and “an attempt to radically undermine global effort towards establishing a new Treaty on Autonomous Weapons Systems.”Footnote 48

References

1 The definition of AI here is from the U.S. Department of State's Political Declaration on Responsible Military Use of Artificial Intelligence and Autonomy. See U.S. Dep't of State Bureau of Arms Control, Verification and Compliance, Political Declaration on Responsible Military Use of Artificial Intelligence and Autonomy (Feb. 16, 2023), at https://www.state.gov/political-declaration-on-responsible-military-use-of-artificial-intelligence-and-autonomy [https://perma.cc/SJW3-3HL2] [hereinafter Political Declaration].

2 See U.S. Dep't of Defense, Summary of the 2018 Department of Defense Artificial Intelligence Strategy: Harnessing AI to Advance Our Security and Prosperity (Feb. 2019), at https://media.defense.gov/2019/Feb/12/2002088963/-1/-1/1/SUMMARY-OF-%20DOD-AI-STRATEGY.PDF [https://perma.cc/WRX5-NZNH] [hereinafter DoD AI Strategy]; Tate Nurkin & Julia Siegel, Battlefield Applications for Human-Machine Teaming: Demonstrating Value, Experimenting with New Capabilities, and Accelerating Adoption, Atlantic Council (Aug. 2023), at https://www.atlanticcouncil.org/wp-content/uploads/2023/08/Battlefield-Applications-for-HMT.pdf [https://perma.cc/TR4D-ERDL].

3 U.S. Mission to International Organizations in Geneva Press Release, Assistant Secretary Stewart Remarks to the CD: New Types of WMD and New Systems of Such Weapons (May 25, 2023), at https://geneva.usmission.gov/2023/05/25/remarks-to-the-cd-thematic-discussion-agenda-item-5 [https://perma.cc/9FHT-YJKT].

4 Id.

5 Id.

6 Id.

7 Id.

8 See, e.g., Eric Lipton, A.I. Brings the Robot Wingman to Aerial Combat, N.Y. Times (Aug. 27, 2023), at https://www.nytimes.com/2023/08/27/us/politics/ai-air-force.html; Lauren Kahn, How Ukraine Is Remaking War, For. Aff. (Aug. 29, 2022), at https://www.foreignaffairs.com/ukraine/how-ukraine-remaking-war; Michael T. Klare, When AI and Robotics Combine, Nation (Aug. 9, 2023), at https://www.thenation.com/article/world/when-ai-and-robotics-combine; Nancy A. Youssef & Michael R. Gordon, Pentagon Plans Vast AI Fleet to Counter China Threat, Wall St. J. (Sept. 6, 2023), at https://www.wsj.com/politics/national-security/pentagon-plans-vast-ai-fleet-to-counter-china-threat-4186a186.

9 See, e.g., Joint Statement on Lethal Autonomous Weapons Systems First Committee, 77th United Nations General Assembly Thematic Debate – Conventional Weapons (Oct. 21 2022), at https://estatements.unmeetings.org/estatements/11.0010/20221021/A1jJ8bNfWGlL/KLw9WYcSnnAm_en.pdf [https://perma.cc/42UP-6FAU].

10 See Autonomy in Weapons Systems, DoD Directive 3000.09 (2012; updated Jan. 25, 2023), at https://www.esd.whs.mil/portals/54/documents/dd/issuances/dodd/300009p.pdf [https://perma.cc/4JXB-T7NN] [hereinafter Directive 3000.09].

11 See Political Declaration, supra note 1.

12 In July, for example, at a Security Council held a meeting on “Artificial intelligence: Opportunities and Risks for International Peace and Security.” See UN Doc. S/PV.9381 (July 18, 2023). The European Union is considering legislation and the Council of Europe is drafting a Convention on Artificial Intelligence, Human Rights, Democracy and the Rule of Law. See Adam Satariano, Europeans Take a Major Step Toward Regulating A.I., N.Y. Times (June 14, 2023), at https://www.nytimes.com/2023/06/14/technology/europe-ai-regulation.html; Delegation of the European Union to the Council of Europe Press Release, Council of Europe Committee on AI Completes a First Reading of the Draft AI Convention (June 29, 2023), at https://www.eeas.europa.eu/delegations/council-europe/council-europe-committee-ai-completes-first-reading-draft-ai-convention_en [https://perma.cc/T3S7-H6GN].

13 Informal meetings had taken place since 2014.

14 See Final Document of the Fifth Review Conference, Decision 1, CCW/CONF.V/10 (Dec. 23, 2016).

15 See Guiding Principles Affirmed by the Group of Governmental Experts on Emerging Technologies in the Area of Lethal Autonomous Weapons System (Nov. 2019), CCW/MSP/2019/9, Annex III (2019); UN Office for Disarmament Affairs, Background on LAWS in the CCW, at https://perma.cc/VY3R-CN98.

16 See Final Report, CCW/MSP/2019/9, para. 31 (2019).

17 Report of the 2023 Session of the Group of Governmental Experts on Emerging Technologies in the Area of Lethal Autonomous Weapons Systems, CCW/GGE.1/2023/2, para. 13 (May 24, 2023) (advance version) [hereinafter GGE 2023 Session Report].

18 See Damien Gayle, UK, US and Russia Among Those Opposing Killer Robot Ban, Guardian (Mar. 29, 2019), at https://www.theguardian.com/science/2019/mar/29/uk-us-russia-opposing-killer-robot-ban-un-ai.

19 See, e.g., U.S. Mission to International Organizations in Geneva, Convention on CCW Group of Governmental Experts on Emerging Technologies in the Area of LAWS: Opening Statement (Dec. 2, 2021), at https://geneva.usmission.gov/2021/12/03/convention-on-ccw-group-of-governmental-experts-on-emerging-technologies-in-the-area-of-laws [https://perma.cc/3YQ7-GHKQ] [hereinafter USG 2021 Opening Statement].

20 U.S. Mission to International Organizations in Geneva Press Release, Second Session in 2023 of the Group of Governmental Experts (GGE) on Emerging Technologies in the Area of Lethal Autonomous Weapons Systems (LAWS) – Opening Statement – U.S. Delegation – Delivered by Joshua Dorosin (May 15, 2023), at https://geneva.usmission.gov/2023/05/15/second-session-in-2023-of-the-gge-on-emerging-technologies-in-the-area-of-laws [https://perma.cc/HX3B-P55G].

21 See Draft Articles on Autonomous Weapon Systems – Prohibitions and Other Regulatory Measures on the Basis of International Humanitarian Law (“IHL”) – Submitted by Australia, Canada, Japan, Poland, the Republic of Korea, the United Kingdom, and the United States, CCW/GGE.1/2023/WP.4/Rev.2 (May 15, 2023); see also Principles and Good Practices on Emerging Technologies in the Area of Lethal Autonomous Weapons Systems – Proposed by: Australia, Canada, Japan, the Republic of Korea, the United Kingdom, and the United States (Mar. 7, 2022). In 2021, the United States had recommended the development of a non-binding code of conduct. See USG 2021 Opening Statement, supra note 19.

22 See Draft Articles, supra note 21, Draft Art. 7.

23 See, e.g., Report of the 2023 Session of the Group of Governmental Experts on Emerging Technologies in the Area of Lethal Autonomous Weapons Systems, CCW/GGE.1/2023/CRP.2 (May 6, 2023) (chair's draft report).

24 See GGE 2023 Session Report, supra note 17, paras. 20–23.

25 See, e.g., XXVIII Cumbre Iberoamericana de Jefas y Jefes de Estado y de Gobierno, Comunicado Especial Sobre el Impacto Social y Humanitario de las Armas Autónomas (Mar. 25, 2023), at https://www.segib.org/wp-content/uploads/11.-Comunicado-especial-sobre-el-impacto-social-y-humanitario-de-las-armas-autonomas_Es.pdf [https://perma.cc/QR5K-J2N8]; Proposal for an International Legal Instrument on Lethal Autonomous Weapons Systems (LAWS). Submitted by Pakistan, CCW/GGE.1/2023/WP.3/Rev.1 (Mar. 8, 2023); Communiqué of the Latin American and the Caribbean Conference of Social and Humanitarian Impact of Autonomous Weapons (Feb. 23–24, 2023), at https://conferenciaawscostarica2023.com/wp-content/uploads/2023/02/EN-Communique-of-La-Ribera-de-Belen-Costa-Rica-February-23-24-2023..pdf [https://perma.cc/JU99-2F7E]; Working Paper to Be Submitted by the Bolivarian Republic of Venezuela on Behalf of the Non-Aligned Movement (NAM) and Other States Parties to the Convention on Certain Conventional Weapons (CCW), para. 17 (June 2021), at https://documents.unoda.org/wp-content/uploads/2021/06/NAM.pdf [https://perma.cc/P6M3-8TVR]; Proposal for a Mandate to Negotiate a Legally Binding Instrument that Addresses the Legal, Humanitarian and Ethical Concerns Posed by Emerging Technologies in the Area of Lethal Autonomous Weapons Systems (LAWS) – Submitted by Austria, Brazil and Chile, CCW/GGE.2/2018/WP.7 (Aug. 29, 2018).

26 Stop Killer Robots, at https://www.stopkillerrobots.org; International Committee of the Red Cross, Autonomous Weapons: ICRC Urges States to Launch Negotiations for Legally Binding Rules (June 5, 2023), at https://www.icrc.org/en/document/statement-international-committee-red-cross-icrc-following-meeting-group-governmental [https://perma.cc/U79T-6L5D].

27 See, e.g., Human Rights Watch, An Agenda for Action: Alternative Processes for Negotiating a Killer Robots Treaty (2022), at https://www.hrw.org/sites/default/files/media_2022/12/arms1122web.pdf.

28 See, e.g., Human Rights Implications of the Development, Use and Transfer of New Technologies in the Context of Counter-terrorism and Countering and Preventing Violent Extremism – Report of the Special Rapporteur on the Promotion and Protection of Human Rights and Fundamental Freedoms While Countering Terrorism, Fionnuala Ní Aoláin, para. 29, UN Doc. A/HRC/52/39 (2023) (advanced edited version); see also Report of the Special Rapporteur on Extrajudicial, Summary or Arbitrary Executions (Apr. 9, 2013), para.78. The Human Rights Council has begun consideration of military AI not limited to LAWS. It recently requested the “Human Rights Council Advisory Committee to prepare a study examining the human rights implications of new and emerging technologies in the military domain, while taking into account ongoing discussions within the United Nations system.” Human Rights Council, Res. 51/22, UN Doc. A/HRC/RES/51/22 (Oct. 7, 2022).

29 United Nations, Our Common Agenda Policy Brief 9: A New Agenda for Peace (July 2023), at https://www.un.org/sites/un2.un.org/files/our-common-agenda-policy-brief-new-agenda-for-peace-en.pdf [https://perma.cc/HKF6-KBVL]. The secretary-general had previously called for “internationally agreed limits on lethal autonomous weapons systems.” UN Press Release, Secretary-General's Briefing to the General Assembly on Priorities for 2023 (Feb. 6, 2023), at https://www.un.org/sg/en/content/sg/statement/2023-02-06/secretary-generals-briefing-the-general-assembly-priorities-for-2023-scroll-down-for-bilingual-delivered-all-english-and-all-french-versions [https://perma.cc/Z395-E9BD]; see also UN Press Release, Address to the Paris Peace Forum (Nov. 11, 2018), at https://www.un.org/sg/en/content/sg/speeches/2018-11-11/address-paris-peace-forum [https://perma.cc/TPX5-W9WC].

30 In addition to weapons systems, military AI also pertains to communications, decision making, intelligence, logistics, planning, risk assessment, situational awareness, surveillance, training, and other issues.

31 See DoD AI Strategy, supra note 2, at 5.

32 U.S. Dep't of Defense Press Release, DoD Adopts Ethical Principles for Artificial Intelligence (Feb. 24, 2020), at https://www.defense.gov/News/Releases/Release/Article/2091996/dod-adopts-ethical-principles-for-artificial-intelligence [https://perma.cc/D2YH-ECN6].

33 See U.S. Dep't of Defense, Implementing Responsible Artificial Intelligence in the Department of Defense (May 26, 2021), at https://media.defense.gov/2021/May/27/2002730593/-1/-1/0/IMPLEMENTING-RESPONSIBLE-ARTIFICIAL-INTELLIGENCE-IN-THE-DEPARTMENT-OF-DEFENSE.PDF [https://perma.cc/T77K-RE9C].

34 See U.S. Dep't of Defense, Responsible Artificial Intelligence Strategy and Implementation Pathway (June 2022), at https://www.ai.mil/docs/RAI_Strategy_and_Implementation_Pathway_6-21-22.pdf [https://perma.cc/GZ8G-R4PA].

35 Directive 3000.09, supra note 10.

36 U.S. Dep't of Defense Press Release, DoD Announces Update to DoD Directive 3000.09, “Autonomy in Weapon Systems” (Jan. 25, 2023), at https://www.defense.gov/News/Releases/Release/Article/3278076/dod-announces-update-to-dod-directive-300009-autonomy-in-weapon-systems [https://perma.cc/SMZ5-KW9F].

38 See Summary of the NATO Artificial Intelligence Strategy (Oct. 22, 2021), at https://www.nato.int/cps/en/natohq/official_texts_187617.htm [https://perma.cc/2HER-DQC8].

39 U.S. Dep't of State Press Release, Keynote Remarks by U/S Jenkins (T) to the Summit on Responsible Artificial Intelligence in the Military Domain (REAIM) Ministerial Segment (Feb. 16, 2023), at https://www.state.gov/keynote-remarks-by-u-s-jenkins-t-to-the-summit-on-responsible-artificial-intelligence-in-the-military-domain-reaim-ministerial-segment [https://perma.cc/X7KD-VBNB].

40 See Government of The Netherlands Press Release, Call to Action on Responsible Use of AI in the Military Domain (Feb. 16, 2023), at https://www.government.nl/latest/news/2023/02/16/reaim-2023-call-to-action [https://perma.cc/6X5F-3SS5]; see also REAIM 2023 Endorsing Countries and Territories (Feb. 21, 2023), at https://www.government.nl/binaries/government/documenten/publications/2023/02/16/reaim-2023-endorsing-countries/Endorsing+Countries+and+Territories+REAIM+Call+to+Action.pdf [https://perma.cc/V4PL-RE8N].

42 Id., para. 19.

43 See Political Declaration, supra note 1. The one significant innovation was the inclusion of a statement that: “States should maintain human control and involvement for all actions critical to informing and executing sovereign decisions concerning nuclear weapons employment.” Id. The United States, United Kingdom, and France jointly made this commitment at the 2022 Non-Proliferation Treaty Review Conference. See Principles and Responsible Practices for Nuclear Weapon States – Working Paper Submitted by France, the United Kingdom of Great Britain and Northern Ireland and the United States of America, para. 5(vii), NPT/CONF.2020/WP.70 (July 29, 2022).

44 Id.

45 U.S. Mission to International Organizations in Geneva Press Release, U.S. Deputy Permanent Representative Aud-Frances McKernan's Remarks to the Conference on Disarmament on Artificial Intelligence in the Military Domain (Aug. 3, 2023), at https://geneva.usmission.gov/2023/08/03/remarks-to-the-conference-on-disarmament-on-artificial-intelligence-in-the-military-domain [https://perma.cc/CV7C-69X5].

46 Id.

47 See Mary Wareham (@marywareham), Twitter (Feb. 16, 2023, 11:00 a.m.), at https://twitter.com/marywareham/status/1626250076241956864.

48 Ousman Noor (@ousmannoor), Twitter (Feb. 16, 2023, 10:07 a.m.), at https://twitter.com/ousmannoor/status/1626236879535489024; Ousman Noor (@ousmannoor), Twitter (Feb. 16, 2023, 10:07 a.m.), at https://twitter.com/ousmannoor/status/1626236901006114818.