Nieman Foundation at Harvard
HOME
          
LATEST STORY
How journalists can avoid amplifying misinformation in their stories
ABOUT                    SUBSCRIBE
July 16, 2013, 1:17 p.m.
LINK: plus.google.com  ➚   |   Posted by: Caroline O'Donovan   |   July 16, 2013

Ushahidi — the crowdsourced mapping technology — has grown from one project in Kenya to more than 100 countries around the world. Today, they hosted a Google Hangout with three organizations utilizing Ushahidi’s technology to track and document the crisis in Syria — Women Under Siege, Syria Tracker, and Syria Deeply.

Lauren Wolfe is the director of the Women Under Siege project, which has been tracking the use of sexualized violence as a tactic in Syria. Speaking to the presenter from Syria Tracker about keeping data sources confidential and safety, she said:

I think both our projects are an interesting mix of human rights research and journalism. I know you guys aren’t journalists, but I am. I think that, because both projects are so public facing, the media does interface with each one a lot — reporting on our data and methods. And at our project, I do a lot of reporting on our work. So it’s an interesting mix of what you make public and what you don’t. The safety of the people reporting and your team members is the number one issue, but there is this funny back and forth issue about what to share and what not.

Here’s the recording of their conversation — check out Twitter and the Hangouts page for more conversation on the topic.

Show tags Show comments / Leave a comment
 
Join the 50,000 who get the freshest future-of-journalism news in our daily email.
How journalists can avoid amplifying misinformation in their stories
We need new tools to ensure visual media travels in secure ways that keep us safer online. Overlays are among these tools.
How China used the media to spread its Covid narrative — and win friends around the world
China’s image plummeted in North America, but over half of 50 nations surveyed at the end of 2020 reported coverage of China had become more positive in their national media since the onset of the pandemic.
From deepfakes to TikTok filters: How do you label AI content?
How should we label AI media in ways that people understand? And how might the labels backfire?