Are Google and Facebook Responsible for the Medical Quackery They Host?

Social media algorithms help charlatans spread autism cures, vaccine disinformation, and AIDS denialism through online videos. Who’s really to blame?

Biswaroop Roy Chowdhury is an Indian engineer with, he says, an honorary Ph.D. in diabetes science from Alliance International University, a school in Zambia that bears many of the hallmarks of an online scam. He runs a small nutrition clinic near Delhi. Two months ago, Chowdhury posted a brief video on YouTube arguing that HIV is not real, and that anti-retroviral medication actually causes AIDS. He offered to inject himself with the blood of someone who had tested positive. Within weeks, the video had more than 380,000 views on YouTube. Tens of thousands more people watched on Facebook. Most of the viewers appear to be in India, where some 60,000 people die of HIV-related causes each year.

MATTERS OF FACT:
Exploring the culture of science.

Facebook and Youtube are not just hosting medical quackery. Their algorithms are helping people to find it, and to view more of it.

After the March video, Chowdhury kept on posting. Follow-up videos on HIV racked up hundreds of thousands more hits. He also distributed copies of an ebook titled “HIV-AIDS: The Greatest Lie of 21st Century.”

When I spoke with Chowdhury by phone last month, he claimed that 700 people had gotten in touch to say they had gone off their HIV medications. The actual number, he added, might be even higher. “We don’t know what people are doing on their own. I can only tell you about the people who report to us,” he said.

Chowdhury’s figures are impossible to verify, but his skills with digital media are apparent — as are the troubling questions they raise about the role of Silicon Valley platforms in spreading misinformation. Such concerns, of course, aren’t new: Over the past two years, consumers, lawmakers, and media integrity advocates in the United States and Europe have become increasingly alarmed at the speed with which incendiary, inaccurate, and often deliberately false content spreads on sites like Facebook and YouTube — the latter a Google subsidiary.

Much of that concern has focused on hate speech, conspiracy theories, and fake political news. But the same questions apply to public health threats, too. Dubious — and potentially deadly — cures for autism have found sanctuary on social media for years, for example. Desperate cancer patients have been lured online to baseless treatments peddled by shady “experts.” And opponents of childhood vaccination thrive in an expanding and self-reinforcing internet bubble that researchers describe as “more real and more true” for its inhabitants than anything coming from the outside.

The question is, when content on social media and similar platforms nudges people toward dangerous medical decisions, do those websites bear any responsibility? And if so, how should they regulate such reckless speech? These questions grow particularly stark with AIDS denialism. After all, convincing someone that the 2001 terrorist attacks on the World Trade Center and other U.S. sites was an inside job can make them sound a little whacky at parties. Convincing someone that AIDS is a hoax can kill them.


AIDS denialism, in the United States at least, emerged in at least the early 1990s, at a time of genuine uncertainty about human immunodeficiency virus and the causes of AIDS. But even as the scientific evidence became clearer and treatments more effective, the notion that the disease is an elaborate hoax has persisted in some corners. Believers in this theory have launched nonprofits, held conferences, and produced documentaries. In South Africa, where the country’s president in the late 1990s and 2000s came to believe many denialist theories, researchers have estimated that the resulting policies led to more than 330,000 deaths.

“What denialists are doing is basically interfering with the application of some highly effective, science-based treatments.”

“What denialists are doing is basically interfering with the application of some highly effective, science-based treatments that have been developed over the last 30 some-odd years,” said David Gorski, a surgical oncologist and the managing editor of Science-Based Medicine, a site that often challenges quackery. “So obviously there’s a problem.”

To some extent, the AIDS denialist movement had been on the wane in recent years — not least because many of its most prominent figures have died. Advances in modern medicine, too, have helped to blunt the movement’s impact. Seth Kalichman, a psychologist at the University of Connecticut and an expert on AIDS denialism, noted that treatment for HIV has made it possible to live an active life, for decades, with the disease — a testament to efficacy of treatment. “The facts of the common day person are becoming harder and harder to deny,” Kalichman told me.

Still, AIDS denialists have adapted readily to the internet age. “With each expansion of their technological soapbox, they have more reach,” Kalichman said over the phone. In an email to me, he was blunter. “Their visibility is increasing even as they become less and less relevant,” he wrote. “I think they are fizzling out and if not for social media, they’d probably have become a thing of the past.”

On YouTube, an account called Question Everything has racked up nearly 1.8 million views by posting documentaries, interviews with AIDS denialists, and other content. A series of AIDS-skeptical videos produced in 2010 by RT, a Kremlin-funded, English-language news service, are also popular on YouTube. And then there are people like Chowdhury, whose low-budget content can quickly reach hundreds of thousands of viewers.

Peter Meylakhs, a researcher at the National Research University’s Higher School of Economics in St. Petersburg, Russia, has studied how AIDS denialists organize on VK, Russia’s equivalent to Facebook. Whether in the U.S. or Russia, he told me, social media offers “a very low transactional cost of getting into the group of people with similar views.”


Love Undark? Sign up for our newsletter!


More worrying, though, is that the digital platforms that host such material and conversations aren’t always passive participants in the recruitment process. Their algorithms, after all, are trained to give visitors more of the kind of content that they like — whatever that might be. While researching this column, for example, I started watching a lot of AIDS denialism videos on YouTube. Immediately, the site began suggesting other denialist videos that I might want to watch, essentially serving up content to keep me on the site longer. Had I been an HIV-positive YouTube user looking for answers about a troubling diagnosis, the effect, perhaps, would have been powerful. (Google did not respond to repeated requests for comment and chose not to answer a list of questions submitted by Undark).

YouTube also uses AIDS denialist videos to sell advertising space. Here’s a screenshot of a commercial for the new Toyota Corolla, appearing before a video that claims that HIV tests are “extremely inaccurate” and medical therapies don’t work:

And here’s a screenshot of a Mercedes commercial that preceded an RT segment titled “The AIDS myth unraveled”:

YouTube’s questionable medical content isn’t limited to AIDS denialism. YouTube videos are really a big deal,” said Gorski, adding that “it seems like every quack has his own series of YouTube videos.”

Had I been an HIV-positive YouTube user looking for answers about a troubling diagnosis, the effect would have been powerful.

Denialist content spreads on Facebook, too. When Undark sent Facebook examples of AIDS denialist content on its site, including Chowdhury’s most popular HIV video, the company said that the content was not in violation of its policies. In an email, Lauren Svensson, a Facebook spokesperson, told Undark that the company recognizes that “contentious perspectives exist” on its platform. “Counter-speech in the form of accurate information and alternative viewpoints can help create a safer and more respectful environment, and we have long believed that simply removing provocative thinking does little to build awareness around facts and approaches to health,” she wrote.

Under Section 230 of the Communications Decency Act, internet platforms are immune from legal liability for most of the content they publish. At the same time, companies like YouTube are allowed to police content on their sites as they deem appropriate, and videos like these could be seen as violations of YouTube’s own terms and conditions, which prohibit content “that intends to incite violence or encourage dangerous or illegal activities that have an inherent risk of serious physical harm or death.” In at least one high-profile case, the site has taken down a channel that peddled questionable medical advice.

Do platforms, or even regulators, have a responsibility to police this sort of misleading or dangerous content more closely?


Share this story!

Lyrissa Lidsky, dean of the University of Missouri Law School and a scholar of free speech online, cautioned against taking an approach that was overly paternalistic, or that gave too much decision-making power to a handful of companies. While AIDS denialism seems clear-cut, she pointed out, many other debates about public truth are murkier. “That’s really the problem: Who gets to the be arbiter of what’s true and what’s false?” Lidsky said. “Some questions are easy, but most of them aren’t easy, even in the realm of scientific fact. And how paternalistic do you want the government to be, and then how paternalistic do you want YouTube or Facebook or Google to be? How comfortable are you with ceding the power to them to determine what’s good for us?”

Lidsky advocates for more education to help citizens become savvier consumers of media, while recognizing the limits of oversight policies. “Respect for individual autonomy says that individuals get to make choices,” she said, “which includes them having a right to make bad choices that cause them harm.”


So far, European countries have been more willing than the United States to suppress certain kinds of misleading or even dangerous health information online. And in Russia, lawmakers have moved to outright ban AIDS denialist content. The country is currently dealing with an HIV epidemic, and there are documented cases of HIV-positive children dying after their parents, skeptical of medical science, refuse to give them medication. But Meylakhs, the Russian scholar, worries that this kind of censoring strategy could backfire. “This will make them more martyr-like, and may, at least, increase their credibility,” he said.

“When there are clear examples of false … information that endanger individuals’ lives, that is when the platforms’ responsibilities are at their apex.”

Part of what makes this issue so complicated, perhaps, is the way that digital platforms bend some of our perceptions of both free speech and public space. Many Americans, steeped in First Amendment principles, for example, would defend a constitutional right for neo-Nazis to march down a public street, or for people to host a rally in a public park in order to promote a dangerous medical conspiracy theory. In one view, YouTube or Facebook simply offer a more high-tech version of that sidewalk or park.

But Facebook isn’t just offering a space for people to meet; its algorithms and search functions are helping them find each other. And YouTube isn’t just a neutral platform for incendiary content: It actively organizes it, and then algorithmically nudges people to view more of it. If a city government actually helped fringe groups put together mailing lists and distribute leaflets, citizens might, understandably, grow more suspicious.

With these and other nuances in mind, some scholars and advocates have begun arguing for more aggressive accountability for online platforms. “When there are clear examples of false … information that endanger individuals’ lives, that is when the platforms’ responsibilities are at their apex, where they really have to start thinking deeply about their role and their responsibility in highlighting this stuff,” said Frank Pasquale, professor at the University of Maryland School of Law and a prominent critic of Silicon Valley policies.

Pasquale is skeptical of concerns that Silicon Valley platforms will wield too much power if they start declaring certain kinds of content off-limits. “They have the power already. The question is whether they’re going to exercise that power responsibly, or are they going to hide behind algorithmic ordering, and just say, ‘Well, it’s all algorithmic, it’s all done by some computer program in the sky, and we don’t really have responsibility for that,’” Pasquale said.

“But the amount of money they have — they clearly have enough money to support this type of intervention,” he added.

For now, platforms like YouTube fall into a gray space, somewhere between a public utility and a media company. The consequences of that ambiguity can unfold in places far from the American public eye. Chowdhury, the nutritionist in India, began posting YouTube videos in 2014. At the time, he told me, he was speaking to auditoriums about his cures, but he wanted to reach a larger audience, so he began filming his talks.

Today, his videos have been viewed close to 12 million times. One of them, about a purported cure for diabetes that does not require medicine, has more than three million views today on YouTube alone. He told me the site has never taken one of his videos down.

“I tried to go to the public through conventional media like television and newspapers,” Chowdhury said. “And I was shocked to know that they are really — even if I pay money — they are not ready … to cover my views, because my method goes opposite of that of conventional medical science.”

Chowdhury found the rules online to be far more forgiving.

“I prefer to go to YouTube,” he said. “I am happy.”


Michael Schulson is an American freelance writer covering science, religion, technology, and ethics. His work has been published by Pacific Standard magazine, Aeon, New York magazine, and The Washington Post, among other outlets, and he writes the Matters of Fact and Tracker columns for Undark.

Top visual: Heritage Images via Getty
See What Others Are Saying

10 comments / Join the Discussion

    As responsible as the government of the town allowing the scene depicted atop this essay.

    Reply

    HIV has never been isolated in the lab. The test tests for a so called immune response as opposed to the existence of the ‘virus’ itself. The inventor of the test even acknowledges its flaws. The ‘screening’ is based on a lifestyle and habits Q&A session that varies wildly from country to country and people have gotten positive results in one country but negative results in another or retested 6 months later in the same country and come back negative as well. All typically after lifestyle changes (sleep & nutrition, stress reduction etc)

    To the sheep among us the above revelations result in a blank stare. To the critical thinkers a process of research and discovery begins.

    To be denied the information we wish to consume and essentially force fed a diet of curated and nutritionless information from one source (the censor) is akin to being given only GMO foods that are poisoned with cancer causing glyphosate (as per the WHO’s findings ) and told to trust our censors that the food/information is perfectly safe.

    No thanks. I’ll source my own info and buy what I can’t grow from producers I have personally vetted.

    Reply

    /* where they become critical thinkers, i.e., skeptics */

    Except that critical thinker shouldn’t become an ignorant skeptic, who denies new facts blindly. Critical thinker should be without bias both toward acceptation, both dismissal.

    Reply

    The answer, as always is MORE free speech, not less. You counter bad ideas with good ideas. Because the simple fact is that there is no one, anywhere, ever, who can be trusted to be the censor. Humans are imperfect and biased. No human, or group of humans, or algorithm programmed by humans, can be blindly trusted to distill truth and fact.

    It is as “TruSkeptik” said above, people must be critical thinkers. It is the only option. Those who will not think critically will always be sheep, easily deceived and misled and ultimately killed by the lies they refuse to think critically about. And because humans are imperfect, there will always be some who mislead, and always some who follow them, and always some who die for it. That is not correctable, and never will be. It is simply part of the human condition.

    Reply

    How can any company with an acknowledged political bias be expected to determine the veracity of articles from charlatans? There are enough articles reproduced in allegedly respected scientific journals which are later proven to be erroneous or patently false. It is preposterous to imagine that the social media would be cabals if, or even interested in, separating the wheat from the chaff

    Reply

    No, youtube is not responsible. Youtube facilitates conversations and communication among a variety of different communities. It is not youtube’s responsibility to police the content of its users submissions. The only exception to this is that youtube is a privately held company and can set whatever rules or guidelines for its users it sees fit. Even charlatans have the right to free speech. The responsibility to determine if the information being presented is factual or beneficial lies with the consumer.

    Reply

    What a wonderful idea. Have the quacks police the quacks. The ONLY thing that will ever work is to educate people to the point where they become critical thinkers, i.e., skeptics. Nothing else will have any effect whatsoever on our perverse and diseased society.

    Reply

    The answer to this article is simple. Who decides what is “quackery”? Youtube and Facebook are qualified to make that assessment? The question is absurd. Perhaps a government agency of completely unbiased bureaucrats could decide!! Yeah, government is completely competent enough to do this! We should have them censoring individuals who lie too!!! What could possibly go wrong?

    Reply

    Medicinal marijuana is likely a similar contentious issue, Given the lack of reliable research

    Reply
Join the discussion

Your email address will not be published. Required fields are marked *

Top

Whistleblowers
& Tipsters

Corruption in science?
Academic discrimination?
Research censorship?
Government cover-ups?

Undark wants to hear about it.

Email us at tips@undark.org, or visit our contact page for more secure options.

Flip
Pocket
Share
Tweet
Stumble
Reddit