Key Scientific Paper Redux: How humans impair automated deception detection performance

How Humans Impair Automated Deception Detection Performance

Authored by Jason Apollo Voss

Jason Apollo Voss is a: conscious capitalist, believer in human potential, pursuer of wisdom & knowledge, and your advocate. He shares his wisdom, intelligence, knowledge, and humility through books, whitepapers, scientific research, articles, workshops, and executive coaching.

28/02/2023

At Deception And Truth Analysis (D.A.T.A.) we use Natural Language Processing (NLP) to look for the known behavioral differences between deceivers and truth-tellers in communications. Frequently we encounter an objection from our prospects that they already are excellent at uncovering deception in documents. Importantly, the findings of “How humans impair automated deception detection performance”[i]calls that claim into question scientifically. Understanding this key scientific paper’s findings is important for due-diligence pros reliant on the representations of people in their work; i.e. most of us.

How humans impair automated deception detection performance: Study Details

For decades the academic community has researched how to better identify if someone is lying or telling the truth. More recently researchers have begun to focus on developing pro-active approaches that attempt to discern whether someone is currently a threat.

Techniques that rely on trained experts unfortunately are hard to scale up. But one promising avenue is linguistic analyses that show a statistically significant higher deception detection accuracy than the classic and disappointing 54% success found in the scientific research when people rely on audiovisual cues (i.e. body language).

Because NLP has been shown capable of analyzing verbal content, it seems only natural to combine the success of people in linguistic analyses with NLP to see if a hybrid approach is beneficial. For example, in social media content moderation machine learning algorithms flag certain content as problematic and then rely on a person to formally judge it.

In “How humans impair automated deception detection performance” people were asked to assess for deceptiveness and truthfulness the statements made by other people about their most significant “non work-related” activity for the forthcoming week. Those providing the statements to be examined were asked to either tell the truth or to deceive. Researchers knew the actual biographical details of these statements so that ground-truth was established. In total, 2,027 people provided a statement for evaluation.

Furthermore, the participants were asked to evaluate on a scale of 1 to 10 how motivated they were to succeed in convincing others of their deception. This addressed one of the criticisms of some deceptions science research. Namely, that deceivers in a laboratory setting may not be as motivated as those who seek to deceive in real life.

The deception scientists in this paper decided to test several different hypotheses in evaluation of the statements they collected:

1. Have both people and a machine learning linguistic algorithm, classify x as being either deceptive or truthful.

2. Examine the success of the above under four conditions:

     a. Pure machine learning NLP-based algorithm.

     b. Pure human judgement.

     c. Hybrid Condition 1, people are fully allowed to overrule the assessment of the algorithm.

     d. Hybrid Condition 2, people are constrained in how much they may override the assessment of the algorithm.

How humans impair automated deception detection performance: Major Findings

  1. Success rates for the above four conditions were:

     a. 69% = Pure machine learning NLP-based algorithm.

     b. 50% = Pure human judgement.

     c. 51% = Hybrid Condition 1, people may overrule the algorithm’s judgment.

     d. 67% = Hybrid Condition 2, people may tweak the algorithm’s judgment.

2. Accuracy at detecting deception vs. truth:

     a. 76% deception & 60% truth = Pure machine learning NLP-based algorithm.

     b. 24% deception & 76% truth = Pure human judgment.

     c. 25% deception & 76% truth = Hybrid Condition 1, people may overrule the algorithm’s judgment.

     d. 60% deception & 74% truth = Hybrid Condition 2, people may tweak the algorithm’s judgment.

3. People in the hybrid conditions tended to adjust the judgments of the algorithm more toward truthful. In other words, similar to other work done in deception science, this indicates that people have a truth bias, rather than a deception bias. That is, when evaluating veracity they tend to judge statements as being more truthful than they should.

4. Regarding the motivation of those in the deceive condition, they scored 8.45 on the 10 point scale. In other words, they seem motivated.

Quotes of Note

  • “Automated systems struggle to extract the implausibility and falsehood of such a statement. To that end, human judges could help since they can interpret context but lack the cognitive capacity to make inferences from high-dimensional data.”
  • “[I]n the majority of cases, human judgment tended to adjust the rating towards “more truthful.”
  • “[I]n none of the human conditions does the judgment of humans improve the automated judgment or exceed the random classification performance.”

    

[i]Kleinberg, Bennett & Bruno Verschuere. “How humans impair automated deception detection performance.” Acta Psychologica. 12 January 2021

You may also like…

0 Comments

Submit a Comment

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.