Biden meets experts on the dangers of AI during a visit to San Francisco
California politics
Queenie WongJune 20, 2023
President
joe
Biden will meet with researchers and lawyers with expertise in artificial intelligence in San Francisco on Tuesday
as his government is trying
to address potential dangers of a technology that could
lead to disinformation, job losses, discrimination and privacy violations.
The meeting comes as Biden rampages
fundraising efforts
for his re-election bid in 2024, including from tech billionaires. While visiting Silicon Valley on Monday, he attended two fundraisers, including one co-hosted by an entrepreneur
Reid Hoffman, who has numerous ties to AI companies. The venture capitalist was an early investor in Open AI
,
which
built
the popular ChatGPT app, and sits on the boards of technology companies, including Microsoft, which invest heavily in AI.
E
Experts Biden is expected to meet
of
Tuesday are some of Big Tech’s loudest critics. The list includes
S
child advocate Jim Steyer, founder and leader of Common Sense Media; Tristan Harris, executive director and co-founder of the Center for
human human
Technologies; Joy Buolamwini, founder of the Algorithmic Justice League; and Fei-Fei Li, co-director of Stanford’s Human-Centered AI Institute.
Some experts have first-hand experience at large technology companies. Harris, a former Google product manager and design ethicist, has spoken out about how social media companies like Facebook and Twitter can harm people’s mental health and amplify misinformation.
Biden’s encounters with
both
AI researchers and tech staffers underline how the president plays on both sides as his campaign tries to attract wealthy donors as his administration examines the risks of the burgeoning technology. While Biden has been critical of tech giants, executives and employees of companies like Apple, Microsoft, Google and
MetaFacebook
also contributed millions of dollars to his campaign in the 2020 election cycle.
“AI is a top priority for the president and his team. Generative AI tools have increased significantly in recent months and we do not want to solve yesterday’s problem,” a White House official said in a statement.
The Biden administration has focused on the risks of AI. Last year, the administration released a blueprint for an AI Bill of Rights, outlining five principles developers should keep in mind before releasing new AI-powered tools. The administration also met with tech CEOs, announced steps the federal government had taken to address AI risks, and continued other efforts to promote responsible US innovation.
Lina Khan, the Biden-appointed chair of the Federal Trade Commission, said in a May op-ed published in the New York Times that the rise of technology platforms such as Facebook and Google is costing users their privacy and security.
As the use of AI becomes more widespread, government officials have a responsibility to ensure this hard-won history doesn’t repeat itself, Khan said.
Tech giants are using AI in various products to recommend videos, control virtual assistants and transcribe audio. While AI has been around for decades, the popularity of an AI chatbot known as ChatGPT has intensified a race between big tech players like Microsoft, Google, and Facebook’s parent company Meta. Launched in 2022 by OpenAI, ChatGPT can answer questions, generate text and perform various tasks.
The rush to advance AI technology has left tech workers, researchers, legislators and regulators worried about whether new products will be released before they are safe. In March Tesla, SpaceX and Twitter
General manager
Elon Musk, Apple co-founder Steve Wozniak and other technology leaders called on AI labs to pause training of advanced AI systems and urged developers to work with policy makers. AI pioneer Geoffrey Hinton, 75, resigned from Google so he could speak more openly about the risks of AI.
As technology advances rapidly, legislators and regulators struggle to keep up. Gavin Newsom indicated that he wants to be careful with the state
–
level AI regulation. Newsom said at a conference in Los Angeles in May that the biggest mistake politicians can make is to assert themselves without first trying to understand. California legislators have put forward several ideas, including legislation that would combat algorithmic discrimination, create an artificial intelligence agency, and set up a task force that would provide the legislature with an AI report.
Writers and artists are also concerned that companies could use AI to replace workers. Using AI to generate text and art raises ethical questions, including about plagiarism and copyright infringement. The Writers Guild of America, still on strike, proposed rules in March
round up
how Hollywood studios can use AI. Text generated by, for example, AI chatbots cannot be counted in determining writing credits according to the proposed rules.
the potential
misuse of AI to spread political propaganda and conspiracy theories, an issue that has plagued social media, is another major concern of disinformation researchers. They are afraid of
That
AI tools that can spit out text and images will make it easier and cheaper for bad actors to spread misleading information.
AI is already being deployed in some mainstream political advertising.
The Republican National Committee posted an AI-generated video ad depicting a dystopian future if Biden wins his 2024 re-election bid. AI tools have also been used to create fake audio clips of politicians and celebrities making comments they didn’t actually say.
in real life.
The campaign of the GOP presidential candidate and Florida governor. Ron DeSantis shared a video of what appeared to be AI-generated images of the former president
Donald
hugging Trump
Dr
Anthony Fauci.
Tech companies are not against guardrails. They welcome regulations, but also try to shape them. In May, Microsoft released a 42-page report on governing AI, which noted that no company is above the law. The report includes a blueprint for AI public governance that outlines five points, including creating safety barriers for AI systems that control the power grid, water systems and other critical infrastructure.
That same month, Sam Altman, CEO of OpenAI, tested for Congress
and called for AI regulation.
My biggest fear is that we, the technology industry, will do significant harm to the world, Altman told lawmakers. If this technology goes wrong, it can go very wrong. Altman, who met with
world
leaders
included
in Europe, Asia, Africa and the Middle East also signed a one-sentence letter in May with scientists and other leaders warning of the risk of extinction posed by AI.