This is a severely truncated article from “Vox”… I am only interested in giving you Bayes’ Method, that will help you go to reality with more certainty than you have now, and spend less time, a whole lot less wondering about your feelings… Whatever they may be. 1
To my dismay, this is what you have been doing, emoting. You make your decisions based on emotions… Life spent in utter uselessness.
I must admit: I didn’t know. I have never done that. I have never been thoroughly seduced to go down THAT rabbit hole, and spend my life in utter uselessness there.
Pondering about your feelings is the result of planet-wide conspiracy to take you out of the game of life.
But if you have been doing that all your life, it is going to take a lot of work for you to abandon that habit, and come out on the other end a human ready to start your journey to human being.
Bayes’ Theorem, Bayes’ Methodology of deciding what is likely the truth, what is likely be happening can help you look and make decisions, more accurate decisions about things that matter…
- For people who are in the Playground course, the most important question is: is this thing happening, is this real? What is the reality? What is happening in reality? You’ll be shocked!
- And for empaths and sensitives the question that if asked and answered accurately can make them productive members of society, thus a lot happier is: is this feeling mine? Do I have a reason to feel this way?
Emotions, emotional decision making is the opposite of Rationality where you use the thinking, rational part of your brain… facts, reality…
The most powerful antidote to irrationality comes from a surprising source: an 18th century English priest named Reverend Thomas Bayes. His pioneering work in statistic (the science of probability) uncovered an immensely powerful mental tool that, if properly used, can drastically improve the way we reason about the world.
Bayes’ theorem, explained
Our modern world is notoriously unpredictable and complex. Should I buy bitcoin? Is that news headline reliable? Is my crush actually into me, or just stringing me along?
Whether it’s our finances or our careers or our love lives, we have to tackle tricky decisions on a daily basis. Additionally, our smartphones bombard us around the clock with a never-ending stream of news and information. Some of that information is reliable, some is noise, and some is intentionally created to mislead. So how do we decide what to believe?
Reverend Bayes made enormous steps toward solving this age-old problem. He was a statistician by training, and his work on the nature of probability and chance laid the groundwork of what is now known as Bayes’ theorem. While its formal definition appears as a rather intimidating mathematical equation, it essentially boils down to this:
In other words, whenever we receive a new piece of evidence, how much should it affect what we currently believe to be true? Does the information support that belief, dispute it, or not affect it at all?
This line of questioning is known as Bayesian reasoning, and chances are you have been using this method of belief-building all your life without realizing it has a formal name.
For example, imagine a co-worker comes to you with a shocking piece of news: He suspects that your boss has been siphoning money from the company. You’ve always respected your boss, and if you had been asked to estimate the likelihood of him being a thief prior to hearing any gossip (the “prior odds”), you would think it extremely unlikely. Meanwhile, your colleague has been known to exaggerate and dramatize situations, especially about people in managerial positions. As such, their word alone carries little evidential weight — and you don’t take their accusation too seriously. Statistically speaking, your “posterior odds” stay pretty much the same.
Now, take the same scenario but instead of verbal information, your colleague produces a paper trail of company money going into a bank account in your boss’s name. In this case, the weight of evidence against him is much stronger, and so the likelihood of “boss = thief” should increase proportionally. The stronger the evidence, the stronger your level of belief. And if the evidence is compelling enough, it should make you change your mind about him entirely.
If this feels obvious and intuitive, it should. The human brain is, to some extent, a natural Bayesian reasoning machine through a process known as predictive processing. The trouble is, almost all our intuitions evolved out of simpler times for savannah-type survival situations. The complexity of more modern-day decisions can sometimes cause our Bayesian reasoning to malfunction, especially when something we really care about is on the line.
What if, instead of respecting your boss, you’re annoyed at him because you feel he’d been unfairly promoted to his current position instead of you? Objectively speaking, your “prior” belief that he is an actual account-skimming thief should be almost as unlikely as in the previous example.
However, because you dislike him for another reason, you now have extra motivation to believe the gossip from your co-worker. This can result in you excessively shifting your “posterior” likelihood despite the lack of hard evidence … and perhaps even doing or saying something unwise.
The phenomenon of being swayed from accurate belief-building by our personal desires or emotions is known as motivated reasoning, and it affects every one of us, no matter how rational we think we are. I’ve lost count of how many times I’ve made an objectively stupid play at the poker table thanks to an excessive emotional attachment to a particular outcome — from chasing lost chips with reckless bluffs after an unlucky run of cards, to foolhardy heroics against opponents who’ve gotten under my skin.
When we identify too strongly with a deeply held belief, idea, or outcome, a plethora of cognitive biases can rear their ugly heads.
Take confirmation bias, for example. This is our inclination to eagerly accept any information that confirms our opinion, and undervalue anything that contradicts it. It’s remarkably easy to spot in other people (especially those you don’t agree with politically), but extremely hard to spot in ourselves because the biasing happens unconsciously. But it’s always there.
And this kind of Bayesian error can have very real and tragic consequences: Criminal cases where jurors unconsciously ignore exonerating evidence and send an innocent person to jail because of a bad experience with someone of the defendant’s demographic. The growing inability to hear alternative arguments in good faith from other parts of the political spectrum. Conspiracy theorists swallowing any unconventional belief they can get their hands on until they think the Earth is flat, or movie stars are lizards, or that a random pizza shop is the base for a sex slavery ring because of a comment thread they read on the internet.
So how do we overcome this deeply ingrained part of human nature? How can we become better Bayesians?
Extraordinary claims require extraordinary evidence
For motivated reasoning, the solution is somewhat obvious: self-awareness.
While confirmation bias is usually invisible to us in the moment, its physiological triggers are more detectable. Is there someone who makes your jaw clench and blood boil the moment they’re mentioned? A societal or religious belief you hold so dear that you think anyone is ridiculous to even want to discuss it?
We all have some deeply held belief that immediately puts us on the defensive. Defensiveness doesn’t mean that belief is actually incorrect. But it does mean we’re vulnerable to bad reasoning around it. And if you can learn to identify the emotional warning signs in yourself, you stand a better chance of evaluating the other side’s evidence or arguments more objectively.
The question is: how likely? how unlikely?
…When dealing with statistics, it is so easy to focus on fear-mongering headlines, like “thousands of people died from terrorism last year,” and forget about the other equally relevant part of the equation: the number of people last year who didn’t die from it.
Occasionally, “red pill” or conspiracy enthusiasts fall into a similar statistical trap. On its face, questioning mainstream belief is a good scientific practice — it can uncover injustice and prevent systemic mistakes from repeating in society. But for some, proving the mainstream wrong becomes an all-consuming mission. And this is especially dangerous in the internet era, where a Google search will always spit out something that fits a chosen narrative. Bayes’ rule teaches you that extraordinary claims require extraordinary evidence.
Yet for some people, the less likely an explanation, the more likely they are to believe it. Take flat-Earth believers. Their claim rests on the idea that all the pilots, astronomers, geologists, physicists, and GPS engineers in the world are intentionally coordinating to mislead the public about the shape of the planet. From a prior odds perspective, the likelihood of a plot so enormous and intricate coming together out of all other conceivable possibilities is vanishingly small. But bizarrely, any demonstration of counterevidence, no matter how strong, just seems to cement their worldview further.
If there is one thing Bayes can teach us to be certain of, however, it is that there is no such thing as absolute certainty of belief. Like a spaceship trying to reach the speed of light, a posterior likelihood can only ever approach 100 percent (or 0 percent). It can never exactly reach it.
And so, anytime we say or think, “I’m absolutely 100 percent certain!” — even for something as probable as our globe-shaped Earth — we’re not only being foolish, we’re being factually wrong. By that statement, we’re effectively saying there is no further evidence in the world, no matter how strong, that could change our minds. And that is as ridiculous as claiming, “I know everything about everything that could ever possibly happen in the universe, ever,” because there are always some unknown unknowns we cannot conceive of, no matter how knowledgeable and wise we think we are.
Which is why science never officially “proves” anything — it just seeks evidence to improve or weaken current theories until they approach 0 percent or 100 percent. This should serve as a reminder that we should always remain open to the possibility of changing our minds if strong enough evidence emerges. And most importantly, we must remember to see our deepest beliefs for what they ultimately are: just another prior probability, floating in a sea of uncertainty.
The original article on Bayes’ Theorem was written by Liv Boeree, a science communicator and TV host specializing in astrophysics, rationality, and poker