Science Scientists Explain Why ‘Doing Your Own Research’ Leads to Believing Conspiracies - Just trust us bro

1703195766130.png

If you've ever found it puzzling that "do your own research" is a slogan for conspiracy theorists of all stripes, new research may have some answers.

While conventional wisdom holds that researching the veracity of fake news would reduce belief in misinformation, a study published on Wednesday in Nature has found that using online search engines to vet conspiracies can actually increase the chance that someone will believe it. The researchers point to a known problem in search called "data voids." Sometimes, there's not a lot of high-quality information to counter misleading headlines or surrounding fringe theories. So, when someone sees an article online about an “engineered famine” due to COVID-19 lockdowns and vaccines, and conducts an unsophisticated search based on those keywords, they may find articles that reaffirm their bias.

“The question here was what happens when people encounter an article online, they’re not sure if it’s true or false, and so they go look for more information about it using a search engine,” Joshua Tucker, co-author and co-director of NYU's Center for Social Media and Politics, told Motherboard. “You see exactly this kind of suggestion in a lot of digital literacy guides.”

In particular, Tucker explained, their research team was interested in knowing how people verify news that’s just happened and hasn’t yet had a chance to be verified by fact-checkers like Snopes or PolitiFact.

In the first experiment of their study, which began in late 2019, some 3,000 people across the US evaluated the accuracy of news articles that had been published in a 48 hour period about topics like COVID-19 vaccines, the Trump impeachment proceedings, and climate events. Some articles were collected from reputable sources, while others were intentionally misleading. Half of the participants were encouraged to search online to help them vet the articles. At the same time, all of the articles were given a ‘true,’ ‘false or misleading,’ or ‘could not determine’ label by professional fact checkers.

People who had been nudged to look for more information online were 19 percent more likely to rate a false or misleading article as fact, compared to those who weren’t encouraged.

“What we find over and over again is that people are over-relying on these search engines or their social connections. They put this blind faith in them,” said Chirag Shah, a professor of information science at the University of Washington who wasn’t involved in the studies. “They think they’ve done their due diligence and checked but it makes it worse than not checking.”

In four other experiments, which ran between 2019 and 2021, researchers found that even if people had initially rated an article as misleading, roughly 18 percent of them changed their mind and said the article was true after searching online (compared to just shy of 6 percent changing from true to false). This held even if the articles were months, instead of hours, old or if the news was well-covered, like the COVID-19 pandemic.

“It was incredible to us how remarkably consistent this effect was across multiple different studies that we ran,” said Tucker. “That’s a real strength of this work. This isn’t just ‘Oh we ran one study’. We’re very, very confident this is happening.”

The researchers showed this effect arises because of the quality of information churned out by Google’s search engine. Partly, this was because of what are called data voids or, as Tucker put it, “the internet is full of junk theory."

“There may be false information out there but not the corresponding true information to correct it,” he said.

These are all issues we’ll continue to grapple with, particularly as large language models and generative AI flood the internet with even more misinformation. These data voids will only grow.

But it also partly came down to how people were searching, added co-author Zeve Sanderson. Seventy-seven percent of people who used the false article’s headline or even the URL in their search got misinformation in their top results. For example, searching for “engineered famine”—based on the false headline “U.S. faces engineered famine as COVID lockdowns and vax mandates could lead to widespread hunger, unrest this winter"—turned up false results more than half the time, as opposed to just “famine” which surfaced zero misinformation.

“That’s not what we’d consider super sophisticated searching strategies but are strategies that we saw people use,” the founding executive director of NYU's Center for Social Media and Politics, told Motherboard. Sanderson added that low quality news publishers tend all to use the same low quality terms, exacerbating this effect.

A spokesperson for Google told Motherboard in an email that it’s not surprising people would find these kinds of results when they search in this way because, although the search engine focuses on quality, if people are looking for something specific, like a headline, they will find it and similar items.

“The four most dangerous words are ‘do your own research’,” said Shah. “It seems counterintuitive because I’m an educator and we encourage students to do this. The problem is people don’t know how to do this.”

Digital literacy curricula shouldn’t just say to search, but instead offer advice on how to search, Kevin Aslett, an assistant professor at the University of Central Florida and co-author, told Motherboard. He and his fellow researchers suggest paying attention to the source of the information you’re searching, not just the content.

Aslett added that more resources need to be pumped into fact-checking organizations so they can at least start to fill the data void that exists.

Google’s spokesperson said that the lack of quality information on particular topics is a known challenge for search engines, but that they’ve built solutions to try and tackle this.

For example, the About This Result feature allows people to see more context around a result by clicking the three dots next to it. Google also provides content advisories when a situation is rapidly evolving, or when they know there isn’t a lot of reliable information.

The spokesperson also emphasized that several independent studies attest that Google surfaces significantly higher quality top results, with less harmful misinformation, than other search engines. “At Google, we design our ranking systems to emphasize quality, and not to expose people to harmful or misleading information that they are not looking for. And we also provide people tools that help them evaluate the credibility of sources, and see a wide range of perspectives, even if they already have an idea in mind.”

Shah said that it’s the responsibility of tech companies like Google to offer tools to help people parse fact from fiction. “We should be equipping them with the right tools. Those tools could come and should come from tech companies and search service providers.” Adding that it’s not up to them or governments to police content. “It’s not only technically infeasible but morally and socially wrong to suppress everything.”

“First we need to have that awareness of ‘Just because you’re doing your research, that doesn’t mean that’s enough.' The more awareness people have, the more chance we have of having people think twice about the information they’re reading," said Shah.

https://www.vice.com/en/article/v7b...our-own-research-leads-to-buying-conspiracies (Archive)
 
Isn't watching the MSM news also "doing your own thing" ?
 
I just assumed whoever built the WTC went with the cheapest contractors and lied about the metalurgical makeup of the structure to avoid getting their pants sued off of them 🤷
I vaguely remember that it was originally built with asbestos cladding in case of fire, however in the 90's they did a major asbestos abatement, replacing it with something else that was easily knocked off by the physical shock of the mass of an airliner slamming into the building.

Effectively, if it still had the asbestos, some people 50 years from then might've died from mesothelioma, but the towers might not have fallen.
 
Another Sam Hyde prediction is slowly coming true. Looking for information online is going to be such a mess that asking your neighbor about the Taliban will be a faster and more effective affair.

That said, the MSM nuked their credibility a long time ago. Misinformation was always a thing but they really tanked it during 2016 with lies blatant enough that even the proles are taking notice. Then 2020s hit. This was when the MSM really showed their fangs and threw everything they got as people were paying attention to them again thanks to the Coof. Here's some highlights.

-They decried anyone trying to defend their homes from BLM during the summer of love as "white nationalists" (despite David Dorn being a black man who died defending his property.)
-On that same vein, praised the heroics of BLM during the firey but peaceful protests
-They got the niggercattle to take the clotshot and enforced the narrative that you will be a jobless wreck if you don't take the totally safe and effective shot.
-Happily swept under the rug the deaths caused by the shot all the while showcasing those who fall sick and die from the coof
-They were instrumental in getting Biden in power as they flubbed the results shown and declared victory even while the count was still going.
-Ran a media blitz to help the public forget about the coof and focus on Ukraine

The media has proven itself as an enemy. And should be treated as such.
 
Back