Say we have a test that predicts HIV in a patient. The test successfully detects all infections. However, the test returns positive result for 1% of healthy people. Only .1% of people are believed to have HIV in the population. Now, say you just took the test and received a positive result. What’s the probability that you have HIV?
Let’s set the problem up into a Bayesian formulation.
\[ Pr(HIV = 1 | Test = 1) = \frac{Pr(Test = 1 | HIV = 1)Pr(HIV=1)}{Pr(Test = 1)}\]
Now, let’s plug in the information that we know.
We don’t know the probability of a positive test, but we can calculate it from the information provided.
\[ Pr(Test = 1) = Pr(Test = 1 | HIV = 1 ) Pr(HIV = 1) + Pr(Test = 1 | HIV = 0 )Pr(HIV=0) \]
\[ Pr(Test = 1) = 1 (.001) + .01 (1 - .001) \]
\[ Pr(Test = 1) = .01099 \]
Plug everything back into the equation.
\[ Pr(HIV = 1 | Test = 1) = \frac{1 (.001)}{.01099} = .091\]
What’s the probability that you have HIV given you tested positive? 9.1% – really low!
Suppose Facebook is concerned about hate speech on its platform. It developed an algorithm that correctly detects posts with hate speech 80% of the time \(Pr(Flagged = 1 | Hate = 1) = .8\). However, it makes mistakes: 14% of the time it incorrectly classifies a normal post as hate speech \(Pr(Flagged = 1 | Hate = 0) = .14\). Across it’s platform, Facebook estimates that hate speech is rare: only .3% of all posts are considered hate speech. Finally, Facebook wants to rid itself of hate speech entirely, so it blocks the accounts of all people who posted any hate speech content.
Your friend’s account was just block for using hate speech. What’s the probability that he/she actually posted hate speech?
\[ Pr(Hate = 1 | Flagged = 1) = \frac{Pr(Flagged = 1|Hate =1)Pr(Hate =1)}{Pr(Flagged=1)} \]
pr_flagged_given_hs = .8 # Positive Classification given hate speech post
pr_flagged_given_n = .14 # Positive Classification given normal post
pr_hs = .003 # Prop. of hate speech content on FB
pr_n = 1 - pr_hs # Prop. of hate speech
# Denominator
pr_flagged = (pr_flagged_given_hs * pr_hs) + (pr_flagged_given_n*pr_n)
# Calculate the probability.
(pr_flagged_given_hs*pr_hs)/pr_flagged # only 1.7%
## 0.016903789266093816
Let’s think about the same set up, but a little differently. Rather than providing probabilities, let’s use frequencies.
With this information, what’s the probability that your friend — who was just kicked off the platform — actually posted hate content?
Here all we need to do is divide the true positives by all posts with a positive classifications (true positives + false positives).
## 0.016903789266093816
The following materials were generated for students enrolled in PPOL564. Please do not distribute without permission.
ed769@georgetown.edu | www.ericdunford.com