AI, politics and the battle against misinformation

0

The Financial Times held a Future of AI summit in London on November 6-7. Below we publish an edited transcript of a conversation about AI, politics and the battle against misinformation between Javier Espinoza, the FT’s EU correspondent covering competition and digital policy, and Elizabeth Dubois, professor at the Centre for Law, Technology and Society at the University of Ottawa.

Elizabeth Dubois: I research the political uses of new technology, so that includes AI, but also social media and search engines and all kinds of other communication technologies that have been infiltrating our electoral systems. And so, recently, I wrote a report looking at AI uses in Canadian politics. But, as you’ve mentioned, I’ve obviously been following the US election to see what’s on the cusp — and what we can expect in others coming forward.

Javier Espinoza: And it’s the perfect timing, I think, for our conversation. Today, we have fresh news and, apart from Trump being re-elected . . . Elon Musk has emerged, in my view, as the other big personality, character, player that has influence. I don’t know to what degree, but he has been a player in this election like we have not seen before — also, with X [formerly Twitter] as a platform to help. I don’t know how many of you are on X, I know that the numbers are dwindling, but the algorithm in my X profile is as if I’m a Republican. What are your thoughts on this?

ED: It’s a really interesting example because people with a lot of money have had a lot of influence in US politics for a long time. That’s not new. And social media companies being this controller of information, they decide what to incentivise . . . what to prioritise in your feed. That’s also not new. But the combination of those things has really played out in a way that I don’t think we were fully prepared for or fully expected.

You mentioned your feed looks like you are a Republican. And we know that that’s the case in a lot of people’s feeds — even though there are roughly equal numbers of Democrats and Republicans that we’re reporting through surveys to be using those tools. So it’s really interesting to see the power of those algorithms. It really shows that our information environment is controlled by these systems, and sometimes by particular people when they take over a large company, for example.

JE: And do you think we’re just learning about the effects that X and Musk have had on this election? Do you think that it’s about the number of people that he managed to reach through the platform, or is it about mobilising the ones that are already converted? What are your thoughts on . . . the use of this platform and also misinformation?

ED: Yeah, I think that X is an example of the larger kind of misinformation conversation, where a lot of the most effective misinformation and disinformation right now is actually about mobilising particular communities and convincing people in specific groups to believe one thing or another. It’s not, at this point, as much about creating a mass misunderstanding of reality, but rather getting certain people who are highly active, who have a loud voice, to be sharing this information and resharing it over and over. 

A person standing at a podium speaks to a large rally crowd, while another individual jumps joyfully next to him, wearing a cap and casual clothes. The scene includes American flags and a festive atmosphere
Elon Musk joins Donald Trump on stage at a Pennsylvania rally in October. The president-elect has since offered the owner of X a role in government © Jim Watson/AFP via Getty Images

JE: I was having a discussion just earlier this morning with someone from the UK regulatory part of the government who was saying that, in their research, they have noticed that people might be seeking out exposure to deep fakes or to misinformation on purpose and excluding the verifiable . . . highly sophisticated information that the Financial Times and other media outlets are producing every day. Have you picked this up in your research? 

ED: Yeah. So one of the things about disinformation research is we really think everyone’s going to want true content, right? If we just have enough high quality content, it’ll be fine. But the reality is people like to be entertained. People like to feel community. People like to have their ideas supported and reinforced. So there’s a bunch of reasons that people are going to intentionally choose or just not kind of question disinformation. 

JE: And I guess this sort of amplifies the use and the efficiency that we’re talking about in terms of X. But, moving beyond X and the elections today, can you . . . flesh out some of the ways that you’ve identified people, agents or countries are using misinformation, and playing with algorithms?

ED: Yeah. So, when we’re thinking about AI and disinformation, the immediate idea is ‘Oh, it’s the deepfakes’. And, absolutely, deepfakes are happening. We have seen examples, even in this US election. But AI use goes beyond that. The thing that we’ve seen emerging is people making use of generative AI tools like ChatGPT as a search engine, and we know that those tools often hallucinate, often produce information that has inaccuracies in it or that lacks contextual information.

So you end up with often ‘true’ misinformation. There isn’t an intent to harm necessarily, but nevertheless, people are being sent to polling stations on the wrong day, as an example.

JE: Wow. That’s quite shocking to hear. In your research, what counts as AI in elections? Give us a little bit more . . . 

ED: There’s a lot of political softwares that embed AI technologies into their systems to help a campaign better target, or better profile, potential voters — and people to not pay attention to.

We also have examples of AI being used for translation or creating robocalls to make it easier to reach greater [numbers of] people, which could be good — we could see that is a very helpful democratic thing if you’re engaging communities that speak particular languages that maybe no candidates speak. But it can also be very deceptive and confusing for people.

And, then, we have this whole group of conversational agents or spoken bots that we’re starting to see emerge. In the last Mexican presidential election, for example, there was a presidential candidate who created an AI-powered “spokesbot” to literally be a spokesperson for her and her campaign. And that really changes the landscape of information. 

JE: How effective was that? Was the candidate elected as a result? 

ED: The candidate was not elected. With all of these kinds of tools, it’s going to be really hard to say that AI was the thing that gets someone elected or doesn’t get someone elected. There are so many different versions of these kinds of tools, and they’re embedded into really complex campaign structures.

JE: But what we do know is it is changing the way that we can interact with political candidates. Do you think – and I know that we have to do the research but let’s speculate a little bit … so we can … at least think about these things … I mean, we’re talking about AI now but … the Brexit vote arguably was also influenced by social media, and the outcome. Do you think that . . . we are seeing outcomes in elections that we wouldn’t have, if we didn’t have these new emerging technologies or ways of disseminating misinformation?

A group of people on stage holding signs supporting Brexit with slogans such as ‘Leave Means Leave,’ ‘No Deal? No Problem!,’ ‘Believe in Britain,’ and ‘Stop Brexit Betrayal.’ The backdrop also displays similar slogans
A 2018 ‘Leave Means Leave’ rally. Anti-Brexit campaigners have claimed that misinformation on social media affected the vote © Daniel Leal/AFP via Getty Images

ED: There’s absolutely no denying the fact that technology impacts elections and the way campaigns are run. It’s really hard to tease apart what is the main thing that changed the result of an election. But it is very clear that the way these new technologies are being integrated into campaigns changes how campaigns run. It changes how journalists report on campaigns, and it changes how the public interacts with the information that comes out of those systems.

So absolutely, we’re seeing impacts now. Does mis- or disinformation impact elections to the point where we can’t trust the results or question the integrity? I think that, so far, what we’ve seen, particularly in the recent US election, is that AI, in terms of its deceptive ability and the way it’s being used is not having the kind of impacts or the kind of disordering effects that we initially expected.

But that doesn’t mean that it’s not going to change as these tools evolve and are integrated into our daily lives.

Audience question: What more do you think the platform owners themselves can do to combat misinformation and disinformation? 

ED: I’ll start first with AI and particularly generative AI tools. I think there needs to be very clear safeguards built into the system so that it is not able to hallucinate or offer inaccurate information or, contextualised information, particularly when it’s relating to how people can vote, and when they should vote, and who’s running in their elections. Those are really essential pieces of information that will undermine the integrity of an election.

Then we go to the larger question of mis- and disinformation being spread across all kinds of social media and search, and that’s a much more tricky one. I think one option is to have increased transparency and clarity on how prioritisation and deprioritisation algorithms work. Let’s make sure there are trust and safety teams that these companies support and make use of, to make sure that when potentially harmful information is being spread across those platforms, they’re actually responding to it.

I also think that, at some point, there may need to be more substantial governmental regulation coming in, because we know that these platforms each make individual choices, and that the change in leadership in one of these organisations can drastically shift the information environment very quickly, which — in the context of an actual election campaign — could be really risky.

Audience question: As we’ve seen in the US election, in states like Georgia, we have seen genuine disruption . . . people being discouraged to vote. And that has had a serious effect. Have you got any general comments about how we can go further to minimise and mitigate that?

ED: In terms of . . . Georgia, I would take it even broader and say, looking at the US election and potential impacts on people’s ability to vote, we know that it extends beyond the uses of AI. And there are electoral systems which make it difficult for people living abroad to get ballots in time, as an example. There’s a variety of things that I think come into play when we’re thinking about whether or not people were informed enough to get their vote cast. That’s not something that’s brand new. But it is something that . . . mis- and disinformation can be used to exacerbate because information can travel so quickly. With AI in so many different formats online, it can be really hard to track what is true and know how to execute the next steps. 

Audience question: The algorithms in social media, they reward engagement, right? And, a lot of times, there’s a bias in humans where they are more drawn to negative, hateful, provocative content. Therefore, the algorithms, like social networks, can just say: ‘Oh, we’re just we’re just repeating what people want to see’. And so politicians that kind of exploit negativity and hate benefit from that. Do you think that the actual algorithms themselves need to be regulated to protect against that?

ED: It’s a great point. It makes me think about the research that’s been done on attack ads. Political attack ads are known to be very effective at making people not trust or not want to vote for whoever is being attacked. But there’s a rebound effect where whoever created the attack also takes a hit in the polls. So that has been a bit of a natural deterrent to using too much of that sort of negative campaigning.

But what we see in online systems is that it’s a lot easier for political campaigns to distance themselves from those attacks and from those kinds of fear-based approaches. So you end up being able to undermine your opponent without necessarily taking the hit yourself.

It can be exacerbated again by different kinds of generative AI tools, the social media algorithms that are amplifying it. So, should the algorithms themselves be regulated? I don’t know that we necessarily want a situation where every platform has to comply with a particular kind of prioritisation, for both business reasons and kind of access to information reasons. But I do think we need a lot more transparency in how those work, and we need additional options so that people can choose different kinds of curators of their information.

This transcript has been edited for brevity and clarity.

#politics #battle #misinformation

Leave a Reply

Your email address will not be published. Required fields are marked *