No one can tell us what is wrong and right, but its great lesson to hear opinions from other people.
The best way to never make a mistake is to either never try anything new, or never have opinion of anything.
SO, to be right or wrong we’ve got to have opinions, whatever they might be.
The following cognitive biases were my personal revelation and made me fix some of the lies I’ve been telling myself.
I hope they help you too!
1. Anchoring bias.
People are over-reliant on the first piece of information they hear. In a salary negotiation, whoever makes the first offer establishes a range of reasonable possibilities in each person’s mind.
2. Availability heuristic.
People overestimate the importance of information that is available to them. A persons might argue that smoking is not unhealthy because they know someone who lived to 100 and smoked three packs a day.
3. Bandwagon effect.
The probability of one person adopting a belief increases based on the number of people who hold that belief. This is a powerful form of groupthink and its reason why meetings are often unproductive.
4. Blind-spot bias.
Failing to recognize your own cognitive biases is a bias in itself. People notice cognitive and motivational biases much more in others than in themselves.
5. Choice-supportive bias.
When you choose something, you tend to feel positive about it, even if that choice has flaws. Like how you think your dog is awesome – even if it bites people every once in a while.
6. Clustering illusion.
This is the tendency to see patterns in random events. It is key to various gambling fallacies, like the idea that red is more or less likely to turn up on a roulette table after a string of reds.
7. Confirmation bias.
We tend to listen only to information that confirms our preconceptions – one of the many reasons it’s so hard to have an intelligent conversations about climate change.
8. Conservatism bias.
Where people favor prior evidence over new evidence or information that has emerged. People were slow to accept that the Earth was round because they maintained their earlier understanding that the planet was flat.
9. Information bias.
The tendency to seek information when it does not affect action. More information is not always better. With less information, people can often make more accurate prediction.
10. Ostrich effect.
The decision to ignore dangerous or negative information by “burying” one’s head in the sand, like an ostrich. Research suggests that investors check the value of their holdings significantly less often during bad markets.
11. Outcome bias.
Judging a decision based on the outcome – rather than how exactly the decision was made in the moment. Just because you won a lot in Vegas doesn’t mean gambling your money was a smart decision.
Some of us are too confident about our abilities, and this causes us to take greater risks in our daily lives. Experts are more prone to this bias than laypeople, since they are more convinced that they are right.
13. Placebo effect.
When simply believing that something will have a certain effect on you causes it to have that effect. In medicine, people given fake pills often experience the same physiological effect as people given the real thing.
14. Pro-innovation bias.
When a proponent of an innovation tends to overvalue its usefulness and undervalue its limitations. Sounds familiar, Silicon Valley?
The tendency to weight the latest information more heavily than older data. Investors often think the market will always look the way it looks today and make unwise decisions.
Our tendency to focus on the most easily recognizable features of a person or concept. When you think about dying, you might worry about being mauled by a lion, as opposed to what is statistically more likely, like dying in a car accident.
17. Selective perception.
Allowing our expectations to influence how we perceive the world. An experiment involving a football game between students from two universities showed that one team saw the opposing team commit more infractions.
Expecting a group or person to have certain qualities without having real information about the person. It allows us to quickly identify strangers as friends or enemies, but people tend to overuse and abuse it.
19. Survivorship bias.
An error that comes from focusing only on surviving examples, causing us to misjudge a situation. For instance, we might think that being an entrepreneur is easy because we haven’t heard of all those who failed.
20. Zero-risk bias.
Sociologists have found that we love certainty – even if it’s counterproductive. Eliminating risk entirely means there is no chance of harm being caused.
From: Business Insider