MBFC Data Deep Dive

In my previous post about the rise of misinformation, I talked about Media Bias Fact Check (MBFC) as source for determining the integrity of a media outlet. Intrigued by the amount of data they had (almost 9,000 media outlets rated in the world) I decided to do a meta analysis of those in the United States.

The United States enjoys a high degree of press freedom, protected by the First Amendment to the Constitution, which guarantees freedom of speech and the press. This freedom fosters a diverse media landscape, encompassing everything from major national outlets to local newspapers, independent blogs, and social media platforms. As a result, Americans are exposed to a broad spectrum of political views and biases. This diversity has both strengths and challenges. On one hand, it encourages vibrant public discourse, enabling individuals to access a variety of perspectives and make informed decisions. On the other hand, the multiplicity of viewpoints can lead to polarization, as some media outlets emphasize sensationalism or partisan narratives to attract audiences and generate revenue. The rise of digital media and algorithms further amplifies this effect, often creating “echo chambers” where individuals are exposed primarily to views that align with their own.

This is why media literacy, fact checking, and self introspection are so important. If we maintain the status quo and take things at face value then we are doomed to misinformation. We must always ask ourselves if we can trust the information.

MBFC uses a sophisticated methodology to determine the credibility of an outlet using four primary categories: bias, factual reporting, traffic/longevity, and press freedom. Given the relatively high press freedom in the United States, the latter does not apply. Within each of these they assign a score and the total score determines high, medium, and low credibility ratings.

Number of Media Outlets by Bias

---
config:
  xyChart:
    width: 1200
    height: 600
    xAxis:
      labelPadding: 0
      labelFontSize: 12
  themeVariables:
    xyChart:
      backgroundColor: "#161618"
---
xychart-beta
   title "Count of Media Outlets by Bias"
	x-axis "Bias" ["Left", "Left-Center", "Least Biased", "Right-Center", "Right", "Pro-Science", "Conspiracy", "Questionable"]
	y-axis "Count" 0 --> 1700
	bar [487, 1182, 1479, 517, 363, 283, 380, 1540]

Insights

  • Conspiracy-Pseudoscience and Questionable Sources accounts for a shocking third of the MBFC dataset (1,920 out of 6,231). This simply cannot be ignored as it lends credence to the systemic and parasitic nature of misinformation.
  • Least Biased outlets account for about a quarter of the dataset (1,479 out of 6,231) indicating that there is still a good amount of unbiased information out there.
  • Left leaning publications outnumber right by almost 2:1.

Average Credibility Rating by Bias

---
config:
  xyChart:
    width: 1200
    height: 600
    xAxis:
      labelPadding: 0
      labelFontSize: 12
  themeVariables:
    xyChart:
      backgroundColor: "#161618"
---
xychart-beta
	title "Average Credibility Rating of Media Outlet by Bias"
	x-axis "Bias" ["Left", "Left-Center", "Least Biased", "Right-Center", "Right", "Pro-Science", "Conspiracy", "Questionable"]
	y-axis "Average Credibility" 0 --> 10
	bar [4.69, 6.45, 7.64, 6.23, 3.61, 7.72, 1.08, 1.90]

Insights

  • Not surprisingly, Conspiracy-Pseudoscience and Questionable Sources are the least credible and should ultimately be avoided while Pro-Science are the most credible and should be sought out whenever possible.
  • Least Biased publications have second highest average credibility rating, indicating that they are good source of information.
  • On average, right leaning publications have lower credibility than left leaning.

Now, it is easy to fall into the trap of looking at numbers and drawing broad conclusions without any sort of statistical analysis that can help weed out randomness. I’m curious to know if media bias, to some extent, determines trust. In order to objectively do this, we need to know if the differences between the average credibility rating of a bias group is due to chance or not. The Analysis of Variance (ANOVA) statistical method is a good choice for this.

Hypothesis

There is a statistically significant difference between the average credibility rating of media bias groups.

Analysis of Variance (ANOVA) Method

Given the number of groups, independence, and slight skewness of the data, using ANOVA, seems to be practical. Generally speaking, you want the data to follow a normal distribution, however, the one-way ANOVA can tolerate data that is non-normal (skewed or kurtotic distributions) with only a small effect on the Type I error rate (the probability of incorrectly stating that there is a difference between the averages of each group)

ANOVA

A statistical method used to examine and compare the means of two or more data groups, ANOVA determines whether the observed differences between group means are statistically significant or merely attributable to random variation.

Parameters

To calculate the result of the ANOVA test, we need to calculate the following for each media bias group:

  • number of samples
  • sum of the crediting ratings
  • average of the credibility ratings
  • standard deviation of the credibility ratings

Once we have these we can then calculate the ANOVA which consists of:

  • degrees of freedom between and within groups
  • sum of squares between and within groups
  • mean of squares between and within groups
  • the F statistic which represents the ratio of the variance between the groups and the variance inside the groups.

Data Tables

BiasSizeSumAverageStandard Deviation
Left48720154.6860465121.029422948
Left-Center118274706.4452113890.8038502572
Least Biased1479111157.6391752580.7237221392
Right-Center51731386.2261904760.8462456976
Right36311123.610389610.7641110166
Pro-Science28321307.7173913041.095372949
Conspiracy-Pseudoscience3803541.0821.081.085688071.06357806
Questionable Sources154026731.901137980.5888998001
Total6231300075.116283035
SourceDegrees of FreedomSum of SquaresMean of SquaresF-StatP-Value
Between Groups737069.7829561.997583889< .00001
Within Groups62233933.9830810.6321682599
Total6230

Calculation

For those not interested in the specifics of the math, you are welcome to skip on down to the conclusion

Given the somewhat complex calculation for ANOVA, I’ve decided to break down each step in the table above for clarity.

Degrees of freedom (DF) between groups is simply the number of bias groups (g) minus one

Degrees of freedom (DF) within groups is the total number of observations (n) across all bias groups minus the number of bias groups (g)

The sum of squares (SS) between groups is the sum of observations within each bias group (n) multiplied by the difference between the mean of the bias group and the mean of the entire data set squared

The sum of squares (SS) within groups is the sum of observations within each bias group (n) minus one, multiplied by the standard deviation of the bias group (s) squared

The mean of squares (MS) between groups is the sum of squares (SS) between groups divided by the number of groups (g) minus one

The mean of squares (MS) within groups is the sum of squares (SS) within groups divided by the total number of samples (n) minus the number of bias groups (g)

The F-Statistic is the mean of squares between groups divided by the mean of squares within groups

You can determine the p-value by doing a lookup in an F-distribution table with the degrees of freedom between and within. Most tables will not go up to very large values such as the case here, however, it can be calculated with statistical software. The resulting p-value is less than 0.00001. In other words, there is less than a 1 in 100,000 chance that the average credibility rating differences are due to chance.

Result

The bias of a media publication can be an indicator of trust given that the difference between the average credibility ratings of each bias group are extremely statistically significant, meaning it is very unlikely due to chance.

Conclusion

Yes, media bias does determine trust, to an extent. Unsurprisingly, you should seek out Pro-Science and Least Biased publications if you want the most credible sources and avoid like the plague anything classified as Conspiracy-Pseudoscience and Questionable Sources

Interestingly, on average, left leaning publications produce more trustworthy content. This is more pronounced when looking at Left vs Right with 4.69 and 3.61 credibility ratings. Given the rise of MAGA and Trumpism, which are notorious for producing divisive propaganda, this may be influencing the lower credibility on the right. The right also tends to be more susceptible to misinformation and is less-informed which could encourage Right bias outlets to produce lower quality content. In a sense, this susceptibility is self-serving.

People tend to seek out information that aligns with their ideals. This doesn’t seem to be a great way to exclusively source your information given the lower credibility of polarized content relative to those that are Least Biased or Pro Science.

Just some thoughts.

When possible, all supporting articles or references I linked to were fact checked and credibility verified. Below is the data from MBFC

SourceBiasReporting AccuracyCredibility Rating
Media MattersLeftHighHigh
Science.orgPro-ScienceHighVery High