AI Distinguishes between Fake News and Satire

Researchers at George Washington University, Amazon AWS AI, and startup AdVerifai investigated a machine learning approach to classifying misleading speech. They say the AI model they developed, which outperformed the baseline, lays the groundwork for the study of additional linguistic features, according to Venturebeat
Their work follows that of MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL), which earlier this year architected an AI model that could determine whether a news source is accurate or politically prejudiced. In subsequent work, MIT CSAIL used one of the world’s largest fact-checking data sets to develop automated systems that could detect false statements.

The paper’s coauthors note that efforts to reduce the spread of misinformation have occasionally resulted in the flagging of legitimate satire, particularly on social media. Complicating matters, some fake news purveyors have begun masquerading as satire sites. These developments of course threaten the business of legitimate publishers, which might struggle to monetize their satire, but also they affect the experience of consumers, who could miss out on miscategorized content.

The researchers hypothesized that metrics of text coherence might be useful in capturing semantic relatedness between sentences of a story. To this end, they used a set of indices related to text statistics implemented by Coh-Metrix, a tool for producing linguistic and discourse representations. There were 108 in total, including (but not limited to) the number of words and sentences; referential cohesion, which refers to overlap in content words between sentences; various text readability formulas; and different types of connective words.

The AI researchers leveraged a statistical technique called principal component analysis to convert potentially correlated metrics into uncorrelated variables (or principal components), which they used in two logistic regression models (functions that model the probability of certain classes) with the fake and satire labels their dependent variables, Next, they evaluated the models’ performance on a corpus containing 283 fake news stories and 203 satirical stories that had been verified by hand.

the “significant” indices outperformed the baseline F1 score, a measure of the frequency of false positives and negatives. The top-performing algorithm achieved a 0.78 score, where 1 is perfect, while revealing that satirical articles tended to be more sophisticated (and less easy to read) than fake news articles.

In future work, the researchers plan to study linguistic cues such as absurdity, incongruity, and other humor-related features.

0 0 votes
Article Rating

Related posts

AIJRF Releases the 2nd Edition of the AI Journalism Professional Ethics and Codes of Conduct (AIJEC)

AIJRF Launches AI and the Media and Academic Content Creation Challenge (AIMAC)

  In cooperation with UNICEF and AIJRF: Egyptian Ministry of Social Solidarity Unveils First Government Ecosystem for AI-Generated Content

Subscribe
Notify of
2 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments