--- title: "Chapter 2 Notes" author: "Emanuel Rodriguez" execute: message: false warning: false format: html: monofont: "Cascadia Mono" highlight-style: gruvbox-dark css: styles.css callout-icon: false callout-apperance: simple --- In this chapter we step through an example of "fake" vs "real" news to build a framework to determine the probability of real vs fake of a new news article titled "The President has a secret!" ```{r} #| message: false #| warning: false # libraries library(bayesrules) library(dplyr) library(tidyr) library(gt) data(fake_news) fake_news <- tibble::as_tibble(fake_news) ``` What is the proportion of news articles that were labeled fake vs real. ```{r} fake_news |> glimpse() fake_news |> group_by(type) |> summarise( total = n(), prop = total / nrow(fake_news) ) ``` If we let $B$ be the event that a news article is "fake" news, and $B^c$ be the event that a news article is "real", we can write the following: $$P(B) = .4$$ $$P(B^c) = .6$$ This is the first "clue" or set of data that we have to build into our framework. Namely, majority of articles are "real", therefore we could simply predict that the new article is "real". This updated sense or reality now becomes our priors. Getting additional data, and updating our priors, based on additional data. The new observation we make is the use of exclamation marks "!". We note that the use of "!" is more frequent in news articles labeled as "fake". We will want to incorporate this into our framework to decide whether the new incoming should be labelled as real or fake. ### Likelihood :::{.callout-note} ## Probability and Likelihood When the event $B$ is known, then we can evaluate the uncertainy of events $A$ and $A^c$ given $B$ $$P(A|B) \text{ vs } P(A^c|B)$$ If on the other hand, we know event $A$ then we can evaluate the relative compatability of data $A$ with $B$ and $B^c$ using likelihood functions $$L(B|A) \text{ vs } L(B^c|A)$$ $$=P(A|B) \text{ vs } P(A|B^c)$$ ::: So in our case, we don't know whether this new incoming article is real or not, but we do know that the title has an exclamation mark. This means we can evaluate how likely this article is real or not given that it contains an "!" in the title using likelihood functions. We can formualte this as: $$L(B|A) \text{ vs } L(B^c|A)$$ And perform the computation in R as follows: ```{r} # if fake, what are the proprotions of ! vs no-! prop_of_excl_within_type <- fake_news |> group_by(type, title_has_excl) |> summarise( total = n() ) |> ungroup() |> group_by(type) |> summarise( has_excl = title_has_excl, prop_within_type = total / sum(total) ) ``` ```{r} prop_of_excl_within_type |> pivot_wider(names_from = "type", values_from = prop_within_type) |> gt() |> gt::cols_label( has_excl = "Contains Exclamtion", fake = "Fake", real = "Real") |> gt::fmt_number(columns=c("fake", "real"), decimals = 3) |> gt::cols_width(everything() ~ px(100)) ``` The table above also shows the likelihoods for the case when an article does not contain exclamation point in the title as well. It's really important to note that these are likelihoods, and its not the case that $L(B|A) + L(B^c|A) = 1$ as a matter of fact this value evaluates to a number less than one. However, since we have that $L(B|A) = .267$ and $L(B^c|A) = .022$ then we have gained additional knowledge in knowing the use of "!" in a title is more compatible with a fake news article than a real one. Up to this point we can summarize our framework as follows | event | $B$ | $B^c$ | Total | |------- |-----|-------|------| | prior | .4 | .6 | 1 | | likelihood |.267 | .022 | .289 | Our next goal is come up with normalizing factors in order to build our probability table: | | $B$| $B^c$| Total | |------|----|------|-------| |$A$ | (1)| (2) | | |$A^c$ | (3)| (4) | | |Total | .4 | .6 | 1 | A couple things to note about our table (1) + (3) = .4 and (2) + (4) = .6. (1) + (2) + (3) + (4) = 1. (1) $P(A \cap B) = P(A|B)P(B)$ we know the likelihood of $L(B|A) = P(A|B)$ and we also know the prior so we insert these to get $$ P(A \cap B) = P(A|B)P(B) = .267 \times .4 = .1068$$ (3) $P(A^c \cap B) = P(A^c|B)P(B)$ in this case we do know the prior $P(B) = .4$, but we don't directly know the value of $P(A^c|B)$, however, we note that $P(A|B) + P(A^c|B) = 1$, therefore we compute $P(A^c|B) = 1 - P(A|B) = 1 - .267 = .733$ $$ P(A^c \cap B) = P(A^c|B)P(B) = .733 \times .4 = .2932$$ we now can confirm that $.1068 + .2932 = .4$