Checkthat2020

My personal take of CheckThat! 2020

Disclaimer. I am one of the organisers of CheckThat! 2020 . This is my very personal overview of what happened during the three parallel sessions. During all three sessions I was either chairing, hosting, and live-tweeting (often at the same time). I might have missed important details. I might show biased impressions

Session 1. Overview and Tasks 1 and 5

Session 2. Tasks 2, 3, and 4

Session 3. Invited speakers and a round table

The closing session of CheckThat! 2020 features three guest speakers:

We targetted to bring them together to know about the fact-checking practices from the perspective of leading organisations in the field of fact-checking in differtent markets.

D. Corney. Making fact checks work harder: claim matching at Full Fact

David started with a very clear statement about what is important for a journalist in Full Fact:

  1. What is the improtant claim to check today?
  2. Is this new claim by that person already checked?
  3. I want to check claims fast!

He explains that, at FullFact, they receive about 100k sentences from the UK alone every day and they usually have 100 fact checks to keep track of. That means 10M pairwise comparisons out of which very few actually match.

In order to find matches, triggering interesting findings, they fuse multiple signals, including sentence BERT similarities, BM25 scores , and the amoount of shared entities, among others.

Whereas based in the UK, Full Fact collaborates with organisations such as Africa Check and the Argentinian Chequeado .

Questions from the audience

(Not all questions were actually asked during the session, due to time concerns. I do not include the answers in most cases, anyway.)

  1. From Tamer E. Are you focused more on political claims? No. For instance, not it is mostly COVID
  2. From Dietrich T. Are you using previously fact-checked claims in future decisions about which claims to check?
  3. From Gautam S. How do you verify if the claim has been already checked? I still see many duplicated fact-checked articles.

R. Míguez Newtral. ClaimHunter - an unattended claim detection service for fact-checkers

Different from other organisations (amon them the other two invited to this event), Newtral is a for profit one. Fact-checking is only one of their activities. They are an independent TV producer with 80+ employees and TV shows in Spanish television.

As for their fact-checking efforts, an interesing aspect is that they do not focus on the Web only. They record and transcribe videos from the Spanish television and pass such transcriptions through a selection process. Interestingly, punctuation is a key aspect for them, because they fact-check sentences.

According to the numbers, only 1.39% of the sentences are relevant and check-worthy. The automatic filtering performed saves the time of the journalists by an estimated 50%.

They have another channel to reach the audience: whatspapp. They are able to match claims multilingually (e.g., Spanish, Catalan, Gallegan) and, if they believe they have fact-checked the user claim already, they just answer to it. When asked about the high risk of answering automatically, Ruben argues that not answering is worst. They indeed shoot the answer, but with a warning, telling the user that this is the result of an automatic process and openind the door for user feedback.

Among the technology that enables their fact-checking process, they have Elastic Search and language-agnostic BERT .

Their ClaimHuneter model has an R=0.82 and P=0.75

Questions from the audience

(Not all questions were actually asked during the session, due to time concerns. I do not include the answers in most cases, anyway.)

  1. How do you identify the verified claims to reply on Whatsapp?
  2. From Tamer E. Isn’t it too risky to auto-reply to the suspicious claims people are sending on Whasapp?
  3. Do you have a strategy to scale?
  4. From Alberto B. You are a for-profit company. Are you profitable?
  5. From Alberto B. Do your journalists feel their job is in danger because of this automation?

M. Althaher. The philosophy of Fake News - The Big Questions!

Moath started with an important (and truth) claim: MENA is not one single category. It is not one single market. People are quite different and have different interests in the Gulf, Levant, Egypt, Magreb. The region with the highest number of fake news is Egypt. One cause is simply the amount of web users.

He explains an interesting example of fake news that remained around for long time: a claim that people in Italy were massively killing each other during the beginning of the COVID-19 crisis! As worrisome as it sounds, there was another even more harmful: that muslims were not at risk with COVID.

Questions from the audience

(Not all questions were actually asked during the session, due to time concerns. I do not include the answers in most cases, anyway.)

  1. From Dietrich T. Regarding the split bu country, are these claims about regional aspects? Otherwise, how do they differ?
  2. Are COVID-19 claims only medical or about health issues?
  3. From Dietrich T. How do you get your claims data? From FB directly? People messaging you? Web scrapping?

Round Table

The round table was an interesting discussion and it would be hard to overview it all. Still, there are some highlights:

  • According to David, social media is important, but it is not everything. Plenty of important information is still on TV. *Indeed, Newtral is permanently listening to the TV *
  • An issue was raied about the typical “who checks fact-checkers?”. How do they deal with bias? all speakers mentioned that they are IFCN signers and that gives them a label of credibility and the commitment to follow strict protocols.
  • In terms of bias, according to Rubén the best way to realise you are doing your job is when everybody is angry with you, regardless of their colour
  • Regarding whether fact-checking should be automated, Moath believes the answer is yes. But the final decision should be made by an expert.
Questions from the audience

(Not all questions were actually asked during the session, due to time concerns. I do not include the answers in most cases, anyway.)

  1. How are you categorising the fake news? Manuually or automatically? I found some times it’s hard to put a fake news in one category like covid+politics. Do you have opinions on that?

They all

could be other reasons.

. **rgewho do fact checking as a Other than int of view of

Avatar
Alberto Barrón-Cedeño
Associate Professor

My research interests include .