Fast Company

advertisement
  • 9:00 am

Non-English Wikipedia has a misinformation problem

The crowdsourced encyclopedia is accurate only if the crowd is big enough and diverse enough to catch factual problems reliably—and that often isn’t the case for Wikipedia’s hundreds of non-English sites.

Non-English Wikipedia has a misinformation problem
[Photo: zodebala/iStock]
advertisement
advertisement
advertisement

Wikipedia doesn’t exactly enjoy a reputation as the most reliable source on the internet. But a report released in June by the Wikimedia Foundation, which operates Wikipedia, suggests that the digital encyclopedia’s misinformation woes may run even deeper than many of its English-speaking users realize.

advertisement

In the report’s summary, the foundation acknowledged that a small network of volunteer administrators of the Croatian-language version of Wikipedia have been abusing their powers and distorting articles “in a way that matched the narratives of political organizations and groups that can broadly be defined as the Croatian radical right.” For almost a decade, the rogue administrators have been altering pages to whitewash crimes committed by Croatia’s Nazi-allied Ustashe regime during World War II and to promote a fascist worldview. For example, it was reported in 2018 that Auschwitz—which English Wikipedia unambiguously deems a concentration camp—was referred to on Croatian Wikipedia as a collection camp, a term that carries fewer negative connotations. The Jasenovac concentration camp, known as Croatia’s Auschwitz, was also referred to as a collection camp.

Croatian Wikipedia users have been calling attention to similar discrepancies since at least 2013, but the Wikimedia Foundation began taking action only last year. That the disinformation campaign went unchecked for so long speaks to a fundamental weakness of Wikipedia’s crowdsourced approach to quality control: It works only if the crowd is large, diverse, and independent enough to reliably weed out assertions that run counter to fact.

By and large, the English version of Wikipedia meets these criteria. As of August 2021, it has more than 120,000 editors who, thanks to the language’s status as a lingua franca, come from a diversity of geographic and cultural backgrounds. English Wikipedia is considered by many researchers to be almost, but still not quite as accurate as traditional encyclopedias. But Wikipedia exists in more than 300 languages, half of which have fewer than 10 active contributors. These non-English versions of Wikipedia can be especially vulnerable to manipulation by ideologically motivated networks.

advertisement

I saw this for myself earlier this year, when I investigated disinformation on the Japanese edition of Wikipedia. Although the Japanese edition is second in popularity only to the English Wikipedia, it receives fewer than one-sixth as many page views and is run by only a few dozen administrators. (The English-language site has nearly 1,100 administrators.) I discovered that on the Japanese Wikipedia, similar to the Croatian version, politically motivated users were abusing their power and whitewashing war crimes committed by the Japanese military during World War II.

For instance, in 2010 the title of the page “The Nanjing Massacre” was changed to “The Nanjing Incident,” an edit that downplayed the atrocity. (Since then, the term Nanjing Incident has become mainstream in Japan.) When I spoke with Wikimedia Foundation representatives about this historical revisionism, they told me that while they’d had periodic contact with the Japanese Wikipedia community over the years, they weren’t aware of the problems I’d raised in an article I wrote for Slate. In one email, a representative wrote that with more than 300 different languages on Wikipedia, it can be difficult to discover these issues.

The fact is, many non-English editions of Wikipedia—particularly those with small, homogenous editing communities—need to be monitored to safeguard the quality and accuracy of their articles. But the Wikimedia Foundation has provided little support on that front, and not for lack of funds. The Daily Dot reported this year that the foundation’s total funds have risen to $300 million. Instead, the foundation noted in an email that because Wikipedia’s model is to uphold the editorial independence of each community, it “does not often get involved in issues related to the creation and maintenance of content on the site.” (While the foundation now says that their trust and safety team is working with a native Japanese speaker to evaluate the issues with Japanese Wikipedia, when I spoke to them in March they told me that Japanese Wikipedia wasn’t a priority.)

If the Wikimedia Foundation can’t ensure the quality of all of its various language versions, perhaps it should make just one Wikipedia.

The idea came to me recently while I watched an interview with the foundation’s former CEO, Katherine Maher, speaking about Wikipedia’s efforts to fight misinformation. During the interview, Maher seemed to imply that for any given topic, there is just one page that’s viewed by “absolutely everyone across the globe.” But that’s not correct. Different versions of the same page can vary substantially from one language to the next.

But what if we could ensure that everyone across the globe saw the same page? What if we could create one universal Wikipedia—one shared, authoritative volume of pages that users from around the world could all read and edit in the language of their choice?

This would be a technological feat, but probably not an impossible one. Machine translation has been quickly improving in recent years and has become a part of everyday life in many non-English-speaking countries. Some Japanese users, rather than read the Japanese version of Wikipedia, choose to translate the English Wikipedia using Google Translate because they know the English version will be more comprehensive and less biased. As translation technology continues to improve, it’s possible to imagine people from all over the world will want to do the same.

Maher has previously stated that “higher-trafficked articles tend to be the highest quality”—that the more relevant a piece of information is to the largest number of people, the higher the quality of its Wikipedia entry. It’s reasonable to expect, then, that if Wikipedia were to merge its various language editions into one global encyclopedia, each entry would see increased traffic and, as a result, improved quality.

In a prepared emailed statement, the Wikimedia Foundation said that a global Wikipedia would change Wikipedia’s current model significantly and pose many challenges. It also said, “with this option, it is also likely that language communities that are bigger and more established will dominate the narrative.”

But the Wikimedia Foundation has already launched one project aimed at consolidating information from around the globe: Wikidata, a collaborative, multilingual, and machine-readable database. The various language editions of Wikipedia all pull information from the same version of Wikidata, creating consistency of content across the platform. But Mark Graham, a professor at the Oxford Internet Institute, has expressed concerns about the project. He cautioned that a universal database might disregard the opinions of marginalized groups, and he predicted that “worldviews of the dominant cultures in the Wikipedia community will win out.”

It’s a valid concern. But in many ways, the current Wikipedia already feels colonial. The site—operated by a foundation whose board of trustees are mostly American and European—has dominated the global internet ecosystem ever since Google began putting it at the top of its search engine results.

Although Wikipedia may have managed, belatedly, to remove abusive editors from its Japanese and Croatian sites, the same thing could happen again; the problem is built into the system.

Creating a global Wikipedia would be challenging, but it would bring further transparency, accuracy, and accountability to a resource that has become one of the world’s go-to repositories of information. If the Wikimedia Foundation is to achieve its stated mission to “help everyone share in the sum of all knowledge,” it might first need to create the sum of all Wikipedias.


Yumiko Sato (@YumikoSatoMTBC) is an author, music therapist, and UX designer.

This article was originally published on Undark. Read the original article.

Volume 0%
Next Up
Pour a cold one for your dog with Anheuser-Busch’s latest brew
06:09
 
This video file cannot be played.(Error Code: 224003)

Fast Company

advertisement
  • 9:00 am

Non-English Wikipedia has a misinformation problem

The crowdsourced encyclopedia is accurate only if the crowd is big enough and diverse enough to catch factual problems reliably—and that often isn’t the case for Wikipedia’s hundreds of non-English sites.

Non-English Wikipedia has a misinformation problem
[Photo: zodebala/iStock]

Wikipedia doesn’t exactly enjoy a reputation as the most reliable source on the internet. But a report released in June by the Wikimedia Foundation, which operates Wikipedia, suggests that the digital encyclopedia’s misinformation woes may run even deeper than many of its English-speaking users realize.

In the report’s summary, the foundation acknowledged that a small network of volunteer administrators of the Croatian-language version of Wikipedia have been abusing their powers and distorting articles “in a way that matched the narratives of political organizations and groups that can broadly be defined as the Croatian radical right.” For almost a decade, the rogue administrators have been altering pages to whitewash crimes committed by Croatia’s Nazi-allied Ustashe regime during World War II and to promote a fascist worldview. For example, it was reported in 2018 that Auschwitz—which English Wikipedia unambiguously deems a concentration camp—was referred to on Croatian Wikipedia as a collection camp, a term that carries fewer negative connotations. The Jasenovac concentration camp, known as Croatia’s Auschwitz, was also referred to as a collection camp.

Croatian Wikipedia users have been calling attention to similar discrepancies since at least 2013, but the Wikimedia Foundation began taking action only last year. That the disinformation campaign went unchecked for so long speaks to a fundamental weakness of Wikipedia’s crowdsourced approach to quality control: It works only if the crowd is large, diverse, and independent enough to reliably weed out assertions that run counter to fact.

By and large, the English version of Wikipedia meets these criteria. As of August 2021, it has more than 120,000 editors who, thanks to the language’s status as a lingua franca, come from a diversity of geographic and cultural backgrounds. English Wikipedia is considered by many researchers to be almost, but still not quite as accurate as traditional encyclopedias. But Wikipedia exists in more than 300 languages, half of which have fewer than 10 active contributors. These non-English versions of Wikipedia can be especially vulnerable to manipulation by ideologically motivated networks.

advertisement

I saw this for myself earlier this year, when I investigated disinformation on the Japanese edition of Wikipedia. Although the Japanese edition is second in popularity only to the English Wikipedia, it receives fewer than one-sixth as many page views and is run by only a few dozen administrators. (The English-language site has nearly 1,100 administrators.) I discovered that on the Japanese Wikipedia, similar to the Croatian version, politically motivated users were abusing their power and whitewashing war crimes committed by the Japanese military during World War II.

For instance, in 2010 the title of the page “The Nanjing Massacre” was changed to “The Nanjing Incident,” an edit that downplayed the atrocity. (Since then, the term Nanjing Incident has become mainstream in Japan.) When I spoke with Wikimedia Foundation representatives about this historical revisionism, they told me that while they’d had periodic contact with the Japanese Wikipedia community over the years, they weren’t aware of the problems I’d raised in an article I wrote for Slate. In one email, a representative wrote that with more than 300 different languages on Wikipedia, it can be difficult to discover these issues.

The fact is, many non-English editions of Wikipedia—particularly those with small, homogenous editing communities—need to be monitored to safeguard the quality and accuracy of their articles. But the Wikimedia Foundation has provided little support on that front, and not for lack of funds. The Daily Dot reported this year that the foundation’s total funds have risen to $300 million. Instead, the foundation noted in an email that because Wikipedia’s model is to uphold the editorial independence of each community, it “does not often get involved in issues related to the creation and maintenance of content on the site.” (While the foundation now says that their trust and safety team is working with a native Japanese speaker to evaluate the issues with Japanese Wikipedia, when I spoke to them in March they told me that Japanese Wikipedia wasn’t a priority.)

If the Wikimedia Foundation can’t ensure the quality of all of its various language versions, perhaps it should make just one Wikipedia.

The idea came to me recently while I watched an interview with the foundation’s former CEO, Katherine Maher, speaking about Wikipedia’s efforts to fight misinformation. During the interview, Maher seemed to imply that for any given topic, there is just one page that’s viewed by “absolutely everyone across the globe.” But that’s not correct. Different versions of the same page can vary substantially from one language to the next.

But what if we could ensure that everyone across the globe saw the same page? What if we could create one universal Wikipedia—one shared, authoritative volume of pages that users from around the world could all read and edit in the language of their choice?

This would be a technological feat, but probably not an impossible one. Machine translation has been quickly improving in recent years and has become a part of everyday life in many non-English-speaking countries. Some Japanese users, rather than read the Japanese version of Wikipedia, choose to translate the English Wikipedia using Google Translate because they know the English version will be more comprehensive and less biased. As translation technology continues to improve, it’s possible to imagine people from all over the world will want to do the same.

Maher has previously stated that “higher-trafficked articles tend to be the highest quality”—that the more relevant a piece of information is to the largest number of people, the higher the quality of its Wikipedia entry. It’s reasonable to expect, then, that if Wikipedia were to merge its various language editions into one global encyclopedia, each entry would see increased traffic and, as a result, improved quality.

In a prepared emailed statement, the Wikimedia Foundation said that a global Wikipedia would change Wikipedia’s current model significantly and pose many challenges. It also said, “with this option, it is also likely that language communities that are bigger and more established will dominate the narrative.”

But the Wikimedia Foundation has already launched one project aimed at consolidating information from around the globe: Wikidata, a collaborative, multilingual, and machine-readable database. The various language editions of Wikipedia all pull information from the same version of Wikidata, creating consistency of content across the platform. But Mark Graham, a professor at the Oxford Internet Institute, has expressed concerns about the project. He cautioned that a universal database might disregard the opinions of marginalized groups, and he predicted that “worldviews of the dominant cultures in the Wikipedia community will win out.”

It’s a valid concern. But in many ways, the current Wikipedia already feels colonial. The site—operated by a foundation whose board of trustees are mostly American and European—has dominated the global internet ecosystem ever since Google began putting it at the top of its search engine results.

Although Wikipedia may have managed, belatedly, to remove abusive editors from its Japanese and Croatian sites, the same thing could happen again; the problem is built into the system.

Creating a global Wikipedia would be challenging, but it would bring further transparency, accuracy, and accountability to a resource that has become one of the world’s go-to repositories of information. If the Wikimedia Foundation is to achieve its stated mission to “help everyone share in the sum of all knowledge,” it might first need to create the sum of all Wikipedias.


Yumiko Sato (@YumikoSatoMTBC) is an author, music therapist, and UX designer.

This article was originally published on Undark. Read the original article.

advertisement
  • 6:00 am

These 3 remarkable Zoom apps make your meetings way better—for free

Learn a bit about who’s on your Zoom calls, ditch the Brady Bunch grid, and eliminate distracting background noise.

These 3 remarkable Zoom apps make your meetings way better—for free
[Source photo: fizkes/iStock]

Let’s not beat around the bush: Nobody really likes video meetings. We all get that same sense of dread every time we see the phrase “schedule a call,” and we all feel ourselves die a teensy bit inside whenever a new Zoom invitation invades our inbox.

It’s not Zoom’s fault, either. It’s mostly just the inherently annoying nature of the medium and all of the awkwardness that comes with it.

Well, we may not be able to avoid video meetings entirely, but we can take some steps to make them a little less unpleasant. The Zoom apps outlined below will bring clever, thoughtful, and broadly transformative enhancements into your next virtual gatherings. They don’t just integrate Zoom with some other service, which is what the vast majority of Zoom apps seem to do, such as those that let you manage meetings from Slack, Calendly, or Google Calendar. These tools will genuinely change the way you experience online meetings.

And you’d better believe they’ll do it for the better.

1. Your personal context genie

Our first exceptional Zoom app is a freshly launched add-on called Warmly. If you remember the old Gmail add-on Rapportive—or if you’ve heard much about the current Gmail app Superhuman—Warmly’s concept will be familiar.

In short, the app gives you contextual info about anyone you’re meeting with, kind of like an invisible genie whispering impressively relevant info into your ear. Once you have it installed and configured, Warmly appears in a panel at the right side of your Zoom window. And it automatically pulls in all sorts of details about the people in your current call and presents them in an easily scannable profile.

Warmly empowers you with useful info about the other people in your Zoom meetings (even if they look suspiciously like Mario).

Warmly uses email addresses from your meeting invitations to identify other attendees and then grab publicly available bio and contact info from the web. That means every single tidbit might not always be accurate—I don’t actually live in New York, for instance—but it gives you enough of an overview to save face and look like you know at least a little bit about the people you’re talking to. No more awkward pauses while you hop over to another window to search LinkedIn, in other words, especially if you’re someone who meets with lots of clients, contractors, or prospective customers. (And as for the ever-important question of privacy, the app doesn’t share any of the information with anyone else, or use it for any other purpose.)

Warmly works with Google Meet and other platforms as well, and it’s completely free to use for now. Its creator tells me the plan is to eventually add in a paid team-level option and possibly also a premium subscription for certain advanced features. But anyone who signs up within this year, he says, will be grandfathered in and able to use all of the app’s current features for free indefinitely.

2. The window-shattering meeting reshaper

Next up is a Zoom app that completely reimagines the way you interact with Zoom. Circles breaks you out of the typical Zoom window and puts all of your meeting participants into floating, resizable circles (get it?!) that can be placed anywhere on your regular desktop.

That means instead of staring at a grid of faces, you can look at whatever you want—your notes, a document, a website, or even your inbox—while still seeing the other people from your meeting and staying in sync with what’s going on.

Somewhat paradoxically, that arrangement makes video meetings feel more like they’re about people. Having faces appear in their own individual bubbles within the more intimate environment of your desktop gives a meeting a much more personal vibe than the standard, and rather sterile, full-screen video grid. And since Circles keeps meetings from completely taking over your computer, it also makes them feel less like they’re taking over your life.

Circles is available for MacOS, while a Windows version is in the works and accessible only via a waitlist as of now. The app is currently free, though the specifics of how that’ll remain viable over the long term aren’t entirely clear.

3. The Zoom audio magician

The third and final spectacular Zoom app in our collection is all about audio, and my goodness, is it something to see—er, hear. Krisp runs quietly in the background on your computer and enhances both the audio you’re sending out to other people through your microphone and the audio you’re receiving during your meetings.

The software uses artificial intelligence to selectively remove background noise, echoing, and other unwanted sounds from the mix and then sends only the sound of actual humans talking to everyone on the other end. And if all of that sounds too good to be true, let me tell you: It isn’t.

I tested Krisp in a variety of conditions, including with someone mowing the lawn right outside my home office window and even with a white noise app making all sorts of racket right into my microphone. I typed on my most clackety mechanical keyboard while talking, too, and had my trusty, professional cacophony-making consultants (aka my 4-year-old and 6-year-old) do their best to disrupt me (a skill at which they’re impossibly well-practiced).

No matter what I threw its way, Krisp was able to separate the ruckus from the rhetoric and send a crisp, clean, clear-sounding feed that made it seem like everyone in the meeting was sitting in a professional audio booth. The second I turned off the app, all of the commotion came crashing back in. The difference is astounding.

Krisp is free for up to 240 minutes of use per week. You can bump up to unlimited filtering for 60 bucks a year, either individually or per person with a team. The app is available for both Mac and Windows.

For even more next-level productivity knowledge, check out my Android Intelligence newsletter.

advertisement

How the cofounders of Ephemeral reinvented tattoos to make them last just a year

Cofounders Brennal Pierre and Vandan Shah came up with a new kind of biodegradable tattoo ink.

How the cofounders of Ephemeral reinvented tattoos to make them last just a year
[Illustration: Julian Ardilla]

Tattoos are a lifetime commitment.

Brennal Pierre

Or they used to be. New York–based startup Ephemeral opened a shop last March offering tattoos that promise to fade away after one year. The tattoo itself is applied by traditional, mechanical needles. What’s different is the ink. Typical tattoo inks clump up in the skin’s dermis, where the pigments resist anti- body proteins that try to break them down. Ephemeral cofounders Brennal Pierre and Vandan Shah developed an ink that’s laced with the same FDA-approved plastics used in pills and medical implants, so it still clumps in the skin but biodegrades as the immune system breaks it down. Once the ink has been broken down, it can pass through the skin, gradually lifting over a period of 9 to 15 months.

Pierre and Shah, who got their PhDs in the same chemical engineering lab at New York University, were inspired to create Ephemeral after their friend (and co-founder) David Seung Shin, now chief of staff at GetLabs, went through the painful and expensive process of having a tattoo laser-removed. The duo began studying how tattoos functioned, and how the ink might be engineered to self-destruct. After developing their first formula, in 2016, the team tested it on pig skin with great success. But when Shah enthusiastically turned the tattoo gun on himself, the results weren’t as good. “It didn’t transfer into my skin at all,” Shah says with a laugh.

Vandan Shah

After five years of development and safety testing—”We’re still the guinea pigs,” Shah jokes—the team found a formula that works (and applied for two pat- ents) and opened a shop in Williamsburg, Brooklyn. It’s currently booked eight months out, and business is growing 100% month over month. The company will open a second studio, in Los Angeles, this fall.

advertisement

Read more about Fast Company’s Most Creative People in Business 2021

About the author

Mark Wilson is a senior writer at Fast Company who has written about design, technology, and culture for almost 15 years. His work has appeared at Gizmodo, Kotaku, PopMech, PopSci, Esquire, American Photo and Lucky Peach

More