.

  • Written by Anjana Susarla, Associate Professor of Information Systems, Michigan State University
How artificial intelligence can detect – and create – fake news

When Mark Zuckerberg told Congress Facebook would use artificial intelligence to detect fake news[1] posted on the social media site, he wasn’t particularly specific about what that meant. Given my own work using image and video analytics[2], I suggest the company should be careful. Despite some basic potential flaws, AI can be a useful tool for spotting online propaganda – but it can also be startlingly good at creating misleading material[3].

This sure looks like Barack Obama saying some things he probably would never say.

Researchers already know that online fake news spreads much more quickly and more widely[4] than real news. My research has similarly found that online posts with fake medical information get more views[5], comments and likes than those with accurate medical content. In an online world where viewers have limited attention and are saturated with content choices, it often appears as though fake information is more appealing or engaging to viewers.

The problem is getting worse: By 2022, people in developed economies could be encountering more fake news than real information[6]. This could bring about a phenomenon researchers have dubbed “reality vertigo[7]” – in which computers can generate such convincing content that regular people may have a hard time figuring out what’s true anymore.

Detecting falsehood

Machine learning algorithms, one type of AI, have been successful for decades fighting spam email, by analyzing messages’ text[8] and determining how likely it is that a particular message is a real communication from an actual person – or a mass-distributed solicitation for pharmaceuticals[9] or claim of a long-lost fortune[10].

Building on this type of text analysis in spam-fighting, AI systems can evaluate how well a post’s text, or a headline, compares with the actual content[11] of an article someone is sharing online. Another method could examine similar articles to see whether other news media have differing facts[12]. Similar systems can identify specific accounts and source websites[13] that spread fake news.

An endless cycle

However, those methods assume the people who spread fake news don’t change their approaches. They often shift tactics, manipulating the content of fake posts[14] in efforts to make them look more authentic.

Using AI to evaluate information can also expose – and amplify – certain biases[15] in society. This can relate to gender, racial background[16] or neighborhood stereotypes[17]. It can even have political consequences, potentially restricting expression of particular viewpoints[18]. For example, YouTube has cut off advertising[19] from certain types of video channels, costing their creators money.

Context is also key. Words’ meanings can change over time. And the same word can mean different things on liberal sites and conservative ones. For example, a post with the terms “WikiLeaks” and “DNC” on a more liberal site could be more likely to be news, while on a conservative site it could refer to a particular set of conspiracy theories[20].

Using AI to make fake news

The biggest challenge, however, of using AI to detect fake news is that it puts technology in an arms race with itself[21]. Machine learning systems are already proving spookily capable at creating what are being called “deepfakes” – photos and videos that realistically replace one person’s face with another, to make it appear that, for example, a celebrity was photographed in a revealing pose[22] or a public figure is saying things he’d never actually say[23]. Even smartphone apps[24] are capable of this sort of substitution – which makes this technology available to just about anyone, even without Hollywood-level video editing skills.

Researchers are already preparing to use AI to identify these AI-created fakes. For example, techniques for video magnification can detect changes in human pulse[25] that would establish whether a person in a video is real or computer-generated. But both fakers and fake-detectors will get better[26]. Some fakes could become so sophisticated that they become very hard to rebut or dismiss – unlike earlier generations of fakes[27], which used simple language and made easily refuted claims.

Human intelligence is the real key

The best way to combat the spread of fake news may be to depend on people[28]. The societal consequences of fake news[29] – greater political polarization, increased partisanship, and eroded trust in mainstream media and government – are significant. If more people knew the stakes were that high, they might be more wary of information, particularly if it is more emotionally based, because that’s an effective way to get people’s attention[30].

When someone sees an enraging post, that person would do better to investigate the information[31], rather than sharing it immediately. The act of sharing also lends credibility to a post: When other people see it, they register that it was shared by someone they know and presumably trust at least a bit, and are less likely to notice whether the original source is questionable.

Social media sites like YouTube and Facebook could voluntarily decide to label their content, showing clearly whether an item purporting to be news is verified by a reputable source[32]. Zuckerberg told Congress he wants to mobilize the “community” of Facebook users[33] to direct his company’s algorithms. Facebook could crowd-source[34] verification efforts[35]. Wikipedia also offers a model, of dedicated volunteers[36] who track and verify information.

Facebook could use its partnerships with news organizations and volunteers to train AI, continually tweaking the system to respond to propagandists’ changes in topics and tactics. This won’t catch every piece of news posted online, but it would make it easier for large numbers of people to tell fact from fake. That could reduce the chances that fictional and misleading stories would become popular online.

Reassuringly, people who have some exposure to accurate news are better at distinguishing between real and fake information[37]. The key is to make sure that at least some of what people see online is, in fact, true.

References

  1. ^ artificial intelligence to detect fake news (www.washingtonpost.com)
  2. ^ image and video analytics (scholar.google.com)
  3. ^ creating misleading material (www.theverge.com)
  4. ^ much more quickly and more widely (doi.org)
  5. ^ fake medical information get more views (dx.doi.org)
  6. ^ more fake news than real information (www.gartner.com)
  7. ^ reality vertigo (www.nature.com)
  8. ^ analyzing messages’ text (blog.malwarebytes.com)
  9. ^ solicitation for pharmaceuticals (www.incapsula.com)
  10. ^ claim of a long-lost fortune (www.fbi.gov)
  11. ^ compares with the actual content (dx.doi.org)
  12. ^ have differing facts (towcenter.org)
  13. ^ identify specific accounts and source websites (doi.org)
  14. ^ manipulating the content of fake posts (doi.org)
  15. ^ expose – and amplify – certain biases (theconversation.com)
  16. ^ racial background (theconversation.com)
  17. ^ neighborhood stereotypes (theconversation.com)
  18. ^ restricting expression of particular viewpoints (thehill.com)
  19. ^ cut off advertising (www.theverge.com)
  20. ^ particular set of conspiracy theories (snap.stanford.edu)
  21. ^ an arms race with itself (motherboard.vice.com)
  22. ^ photographed in a revealing pose (www.theverge.com)
  23. ^ saying things he’d never actually say (www.theverge.com)
  24. ^ smartphone apps (itunes.apple.com)
  25. ^ detect changes in human pulse (people.csail.mit.edu)
  26. ^ fakers and fake-detectors will get better (www.nature.com)
  27. ^ earlier generations of fakes (www.factcheck.org)
  28. ^ depend on people (www.nytimes.com)
  29. ^ societal consequences of fake news (theconversation.com)
  30. ^ effective way to get people’s attention (www.degruyter.com)
  31. ^ investigate the information (gweb-newslab-qa.appspot.com)
  32. ^ verified by a reputable source (towcenter.org)
  33. ^ mobilize the “community” of Facebook users (www.vox.com)
  34. ^ crowd-source (doi.org)
  35. ^ verification efforts (theconversation.com)
  36. ^ dedicated volunteers (doi.org)
  37. ^ better at distinguishing between real and fake information (dx.doi.org)

Authors: Anjana Susarla, Associate Professor of Information Systems, Michigan State University

Read more http://theconversation.com/how-artificial-intelligence-can-detect-and-create-fake-news-95404

Metropolitan republishes selected articles from The Conversation USA with permission

Visit The Conversation to see more