Category Archives: Social Computing

MicroMappers: Towards Next Generation Humanitarian Technology

The MicroMappers platform has come a long way and still has a ways to go. Our vision for MicroMappers is simple: combine human computing (smart crowd-sourcing) with machine computing (artificial intelligence) to filter, fuse and map a variety of different data types such as text, photo, video and satellite/aerial imagery. To do this, we have created a collection of “Clickers” for MicroMappers. Clickers are simply web-based crowdsourcing apps used to make sense of “Big Data”. The “Text Cicker” is used to filter tweets & SMS’s; “Photo Clicker” to filter photos; “Video Clicker” to filter videos and yes the Satellite & Aerial Clickers to filter both satellite and aerial imagery. These are the Data Clickers. We also have a collection of Geo Clickers that digital volunteers use to geo-tag tweets, photos and videos filtered by the”Data Clickers. Note that these Geo Clickers auto-matically display the results of the crowdsourced geo-tagging on our MicroMaps like the one below.

MM Ruby Tweet Map

Thanks to our Artificial Intelligence (AI) engine AIDR, the MicroMappers “Text Clicker” already combines human and machine computing. This means that tweets and text messages can be automatically filtered (classified) after some initial crowdsourced filtering. The filtered tweets are then pushed to the Geo Clickers for geo-tagging purposes. We want to do the same (semi-automation) for photos posted to social media as well as videos; although this is still a very active area of research and development in the field of computer vision.

So we are prioritizing our next hybrid human-machine computing efforts on aerial imagery instead. Just like the “Text Clicker” above, we want to semi-automate feature detection in aerial imagery by adding an AI engine to the “Aerial Clicker”. We’ve just starting to explore this with computer vision experts in Switzerland and Canada. Another development we’re eyeing vis-a-vis UAVs is live video streaming. To be sure, UAVs will increasingly be transmitting live video feeds directly to the web. This means we may eventually need to develop a “Streaming Clicker”, which would in some respects resemble our existing “Video Clicker” except that the video would be broadcasting live rather than play back from YouTube, for example. The “Streaming Clicker” is for later, however, or at least until a prospective partner organization approaches us with an immediate and compelling social innovation use-case.

In the meantime, my team & I at QCRI will continue to improve our maps (data visualizations) along with the human computing component of the Clickers. The MicroMappers smartphone apps, for example, need more work. We also need to find partners to help us develop apps for tablets like the iPad. In addition, we’re hoping to create a “Translate Clicker” with Translators Without Borders (TWB). The purpose of this Clicker would be to rapidly crowdsource the translation of tweets, text messages, etc. This could open up rather interesting possibilities for machine translation, which is certainly an exciting prospect.

MM All Map

Ultimately, we want to have one and only one map to display the data filtered via the Data and Geo Clickers. This map, using (Humanitarian) OpenStreetMap as a base layer, would display filtered tweets, SMS’s, photos, videos and relevant features from satellite and UAV imagery. Each data type would simply be a different layer on this fused “Meta-Data Crisis Map”; and end-users would simply turn individual layers on and off as needed. Note also the mainstream news feeds (CNN and BBC) depicted in the above image. We’re working with our partners at UN/OCHA, GDELT & SBTF to create a “3W Clicker” to complement our MicroMap. As noted in my forthcoming book, GDELT is the ultimate source of data for the world’s digitized news media. The 3Ws refers to Who, What, Where; an important spreadsheet that OCHA puts together and maintains in the aftermath of major disasters to support coordination efforts.

In response to Typhoon Ruby in the Philippines, Andrej Verity (OCHA) and I collaborated with Kalev Leetaru from GDELT to explore how the MicroMappers “3W Clicker” might work. The result is the Google Spreadsheet below (click to enlarge) that is automatically updated every 15 minutes with the latest news reports that refer to one or more humanitarian organizations in the Philippines. GDELT includes the original URL of the news article as well as the list of humanitarian organizations referenced in the article. In addition, GDELT automatically identifies the locations referred to in the articles, key words (tags) and the date of the news article. The spreadsheet below is already live and working. So all we need now is the “3W Clicker” to crowdsource the “What”.

MM GDELT output

The first version of the mock-up we’ve created for the “3W Clicker” is displayed below. Digital volunteers are presented with an interface that includes an news article with the names of humanitarian organizations highlighted in red for easy reference. GDELT auto-populates the URL, the organization name (or names if there are more than one) and the location. Note that both the “Who” & “Where” information can be edited directly by the volunteer incase GDELT’s automated algorithm gets those wrong. The main role of digital volunteers, however, would simply be to identify the “What” by quickly skimming the article.

MM 3W Clicker v2

The output of the “3W Clicker” would simply be another MicroMap layer. As per Andrej’s suggestion, the resulting data could also be automatically pushed to another Google Spreadsheet in HXL format. We’re excited about the possibilities and plan to move forward on this sooner rather than later. In addition to GDELT, pulling in feeds from CrisisNET may be worth exploring. I’m also really keen on exploring ways to link up with the Global Disaster Alert & Coordination System (GDACS) as well as GeoFeedia.

In the meantime, we’re hoping to pilot our “Satellite Clicker” thanks to recent conversations with Planet Labs and SkyBox Imaging. Overlaying user-generated content such as tweets and images on top of both satellite and aerial imagery can go a long way to helping verify (“ground truth”) social media during disasters and other events. This is evidenced by recent empirical studies such as this one in Germany and this one in the US. On this note, as my QCRI colleague Heather Leson recently pointed out, the above vision for MicroMappers is still missing one important data feed; namely sensors—the Internet of Things. She is absolutely spot on, so we’ll be sure to look for potential pilot projects that would allow us to explore this new data source within MicroMappers.

The above vision is a tad ambitious (understatement). We really can’t do this alone. To this end, please do get in touch if you’re interested in joining the team and getting MicroMappers to the next level. Note that MicroMappers is free and open source and in no way limited to disaster response applications. Indeed, we recently used the Aerial Clicker for this wildlife protection project in Namibia. This explains why our friends over at National Geographic have also expressed an interest in potentially piloting the MicroMappers platform for some of their projects. And of course, one need not use all the Clickers for a project, simply the one(s) that make sense. Another advantage of MicroMappers is that the Clickers (and maps) can be deployed very rapidly (since the platform was initially developed for rapid disaster response purposes). In any event, if you’d like to pilot the platform, then do get in touch.

bio

See also: Digital Humanitarians – The Book

Digital Jedis Complete Response to Typhoon Ruby

Thank you, Digital Jedis!

Every Click you made on MicroMappers was a gift. Typhoon Ruby (Hagupit) disrupted the lives of many and caused damage in regions already affected by previous disasters. As MicroMappers, you gave your time, clicks and skills to make a difference. Catherine, the Head of the UN’s Information Management Unit in the Philippines had this to say: “I would like to thank all the volunteers […] for their invaluable contribution over the past few days. We are lucky that Hagupit [Ruby] made less damages than expected and that the emergency quickly scaled down.”

MM Ruby Tweet Map

MicroMappers and our partners at the Standby Task Force (SBTF) were activated by the United Nations Office for the Coordination of Humanitarian Affairs (OCHA). The Mission?

To augment the situational awareness of humanitarian actors on the ground by making sense of social media generated following the Typhoon.

Over the course of 72 hours, these Digital Jedis united to MicroMap one Click at a time. By reviewing tweets and image, each MicroMapper built collective intelligence and insights that were used to build a comprehensive situational awareness reports and maps for the UN. Many hands, and in this case, Clicks, make light work.

As Catherine rightly notes, there was thankfully less damage than many feared. This explains why our MicroMaps (above and below) are thankfully not riddled with hundreds of markers. In addition, we prioritize quality over quantity at MicroMappers. Our UN partners had specifically asked for tweets related to:

(1) Requests for Help / Needs
(2) Infrastructure Damage
(3) Humanitarian Aid Provided

Together, these tweets—which are mapped above—represented less than 5% of the Ruby-related tweets that were collected during the first 72 hours of the Typhoon making landfall. This doesn’t mean that only 5% of the information on Twitter was relevant for emergency response, however. Indeed, we also tagged tweets that were not related to the above 3 categories but that were still informative. These constituted more than 20% of all tweets collected (which are not included in the map above). In the analysis provided to UN partners, we did include a review of those other relevant tweets.

MM Ruby Tweet Clicker

Some 700 Digital Jedis joined the response online, a new record for MicroMappers! An astounding 50,394 Clicks were made using the Text Clicker pictured above (each tweet was reviewed by at least 3 digital volunteers for quality assurance purposes). And a further 3,555 Clicks were carefully made by the SBTF to geo-locate (map) relevant tweets. In other words, close to 55,000 Clicks went into making the high quality map displayed above! That’s over 12 Clicks per minute non-stop for more than 4,300 consecutive minutes!

MM Ruby Image Map

The United Nations also asked Digital Jedis to identify pictures posted on Twitter that showed disaster damage. Over 30,000 Clicks went into this operation with a further 7,413 Clicks made by the SBTF to map images that showed severe and mild damage. In sum, over 40,000 Clicks went into the MicroMap above. Overall, the entire MicroMappers response was powered by close to 100,000 Clicks!

Screen Shot 2014-12-10 at 8.36.04 AMMM Infographic 2MM Infographic 3

Digital Jedis have yet again shown that together, we can help people get positively involved in their world, even when half-a-globe and many timezones away. Yes, we can and should donate $$ to support relief efforts and good causes around the world but we can also get directly involved by donating our time, or what we call M&M’s, Minutes and Mouse clicks. This year MicroMappers have mobilized to support wildlife protection in Namibia, food security efforts in the Philippines and of course this most recent response to Typhoon Ruby. On that note, thanks again to all volunteers who supported the MicroMappers response to the Typhoon in partnership with the United Nations. You truly are Digital Jedis! And the UK Guardian certainly agrees, check out their article on our digital response.

So what’s next? We will continue to solicit your feedback on how to improve the Clickers and will get started right away. (Add your MicroMappers feedback here). In the meantime, we will leave the Clickers online for newcomers who wish to practice. We are also in touch with the UN and UAV partners in the Philippines as they may soon fly their small, remote-control planes to take aerial photographs over disaster affected areas. If they do, they will send us the photographs for analysis via MicroMappers, so stay tuned.

In closing, MicroMappers was developed by QCRI in partnership SBTF/OCHA. So a million thanks to the QCRI team and SBTF for deploying MicroMappers in support of these digital humanitarian efforts. Special thanks go to Ji Lucas, Jus Mackinnon, ChaTo Castillo, Muhammad Imran, Heather Leson, Sarah Vieweg and last but certainly not least Peter Mosur.

(Ed. note: Blog post was cross-posted from MicroMappers.org. Infrographic uses Infogr.am software)

Social Media Generates Social Capital: Implications for City Resilience and Disaster Response

A new empirical and peer-reviewed study provides “the first evidence that online networks are able to produce social capital. In the case of bonding social capital, online ties are more effective in forming close networks than theory predicts.” Entitled, “Tweeting Alone? An Analysis of Bridging and Bonding Social Capital in Online Networks,” the study analyzes Twitter data generated during three large events: “the Occupy movement in 2011, the IF Campaign in 2013, and the Chilean Presidential Election of the same year.”

cityres

What is the relationship between social media and social capital formation? More specifically, how do connections established via social media—in this case Twitter—lead to the formation of two specific forms of social capital, bridging and bonding capital? Does the interplay between bridging and bonding capital online differ to what we see in face-to-face world interactions?

“Bonding social capital exists in the strong ties occurring within, often homogeneous, groups—families, friendship circles, work teams, choirs, criminal gangs, and bowling clubs, for example. Bonding social capital acts as a social glue, building trust and norms within groups, but also potentially increasing intolerance and distrust of out-group members. Bridging social capital exists in the ties that link otherwise separate, often heterogeneous, groups—so for example, individuals with ties to other groups, messengers, or more generically the notion of brokers. Bridging social capital allows different groups to share and exchange information, resources, and help coordinate action across diverse interests.” The authors emphasize that “these are not either/or categories, but that in well-functioning societies the two types or dimensions develop together.”

The study uses social network analysis to measure bonding and bridging social capital. More specifically, they use two associated metrics as indicators of social capital: closure and brokerage. “Closure refers to the level of connectedness between particular groups of members within a broader network and encourages the formation of trust and collaboration. Brokerage refers to the existence of structural holes within a network that are ’bridged’ by a particular member of the network. Brokerage permits the transmission of information across the entire network. Social capital, then, is comprised of the combination of these two elements, which interact over time.”

The authors thus analyze the “observed values for closure and brokerage over time and compare them with different simulations based on theoretical network models to show how they compare to what we would expect offline. From this, [they provide an evaluation of the existence and formation of social capital in online networks.”

The results demonstrate that “online networks show evidence of social capital and these networks exhibit higher levels of closure than what would be expected based on theoretical models. However, the presence of organizations and professional brokers is key to the formation of bridging social capital. Similar to traditional (offline) conditions, bridging social capital in online networks does not exist organically and requires the purposive efforts of network members to connect across different groups. Finally, the data show interaction between closure and brokerage goes in the right direction, moving and growing together.”

These conclusions suggest that the same metrics—closure and brokerage—can be used to monitor “City Resilience” before, during and after major disasters. This is of particular interest to me since my team and I at QCRI are collaborating with the Rockefeller Foundation’s 100 Resilient Cities initiative to determine whether social media can indeed help monitor (proxy indicators of) resilience. Recent studies have shown that changes in employment, economic activity and mobility—each of which is are drivers of resilience—can be gleamed from social media.

While more research is needed, the above findings are compelling enough for us to move forward with Rockefeller on our joint project. So we’ll be launching AIRS in early 2015. AIRS, which stands for “Artificial Intelligence for Resilient Societies” is a free and open source platform specifically designed to enable Rockefeller’s partners cities to monitor proxy indicators of resilience on Twitter.

Bio

See also:

  • Using Social Media to Predict Disaster Resilience [link]
  • Social Media = Social Capital = Disaster Resilience? [link]
  • Does Social Capital Drive Disaster Resilience? [link]
  • Digital Social Capital Matters for Resilience & Response [link]

Using Social Media to Anticipate Human Mobility and Resilience During Disasters

The analysis of cell phone data can already be used to predict mobility patterns after major natural disasters. Now, a new peer-reviewed scientific study suggests that travel patterns may also be predictable using tweets generated following large disasters. In “Quantifying Human Mobility Perturbation and Resilience in Hurricane Sandy,” co-authors Qi Wang and John Taylor analyze some 700,000 geo-tagged tweets posted by ~53,000 individuals as they moved around over the course of 12 days. Results of the analysis confirm that “Sandy did impact the mobility patterns of individuals in New York City,” but this “perturbation was surprisingly brief and the mobility patterns encouragingly resilient. This resilience occurred even in the large-scale absence of mobility infrastructure.”

Twitter Mobility

In sum, this new study suggests that “Human mobility appears to possess an inherent resilience—even in perturbed states—such that movement deviations, in aggregate, follow predictable patterns in hurricanes. Therefore, it may be possible to use human mobility data collected in steady states to predict perturbation states during extreme events and, as a result, develop strategies to improve evacuation effectiveness & speed critical disaster response to minimize loss of life and human suffering.”

Authors Wang and Taylor are now turning their attention to “10 other storms and typhoons that they’ve collected data on.” They hope to further demonstrate that quantifying mobility patterns before and after disasters will eventually help cities “predict mobility in the face of a future disaster, and thereby protect and serve residents better.” They also want to “understand where the ‘upper limit’ of resilience lies. ‘After Haiyan,’—the deadliest-ever Philippine Typhoon that struck last November—’there was a total breakdown in mobility patterns,’ says Taylor.”

Of course, Twitter data comes with well-known limitations such as demographic bias, for example. This explains why said data must be interpreted carefully and why the results simply augment rather than replace the analysis of traditional data sources used for damage after needs assessments after disasters.

bio

See also:

  • Social Media & Emergency Management: Supply and Demand [link]
  • Using AIDR to Automatically Classify Disaster Tweets [link]
  • Visualization of Photos Posted to Instagram During Sandy [link]
  • Using Twitter to Map Blackouts During Hurricane Sandy [link]
  • Analyzing Foursquare Check-Ins During Hurricane Sandy [link]

Digital Jedis: There Has Been An Awakening…

Computing Research Institutes as an Innovation Pathway for Humanitarian Technology

The World Humanitarian Summit (WHS) is an initiative by United Nations Secretary-General Ban Ki-moon to improve humanitarian action. The Summit, which is to be held in 2016, stands to be one of the most important humanitarian conferences in a decade. One key pillar of WHS is humanitarian innovation. “Transformation through Innovation” is the WHS Working Group dedicated to transforming humanitarian action by focusing explicitly on innovation. I have the pleasure of being a member of this working group where my contribution focuses on the role of new technologies, data science and advanced computing. As such, I’m working on an applied study to explore the role of computing research institutes as an innovation pathway for humanitarian technology. The purpose of this blog post is to invite feedback on the ideas presented below.

WHS_Logo_0

I first realized that the humanitarian community faced a “Big Data” challenge in 2010, just months after I had joined Ushahidi as Director of Crisis Mapping, and just months after co-founding CrisisMappers: The Humanitarian Technology Network. The devastating Haiti Earthquake resulted in a massive overflow of information generated via mainstream news, social media, text messages and satellite imagery. I launched and spearheaded the Haiti Crisis Map at the time and together with hundreds of digital volunteers from all around the world went head-to head with Big Data. As noted in my forthcoming book, we realized there and then that crowdsourcing and mapping software alone were no match for Big (Crisis) Data.

Digital Humanitarians: The Book

This explains why I decided to join an advanced computing research institute, namely QCRI. It was clear to me after Haiti that humanitarian organizations had to partner directly with advanced computing experts to manage the new Big Data challenge in disaster response. So I “embedded” myself in an institute with leading experts in Big Data Analytics, Data Science and Social Computing. I believe that computing research institutes (CRI’s) can & must play an important role in fostering innovation in next generation humanitarian technology by partnering with humanitarian organizations on research & development (R&D).

There is already some evidence to support this proposition. We (QCRI) teamed up with the UN Office for the Coordination of Humanitarian Affairs (OCHA) to create the Artificial Intelligence for Disaster Response platform, AIDR as well as MicroMappers. We are now extending AIDR to analyze text messages (SMS) in partnership with UNICEF. We are also spearheading efforts around the use and analysis of aerial imagery (captured via UAVs) for disaster response (see the Humanitarian UAV Network: UAViators). On the subject of UAVs, I believe that this new technology presents us (in the WHS Innovation team) with an ideal opportunity to analyze in “real time” how a new, disruptive technology gets adopted within the humanitarian system. In addition to UAVs, we catalyzed a partnership with Planet Labs and teamed up with Zooniverse to take satellite imagery analysis to the next level with large scale crowd computing. To this end, we are working with humanitarian organizations to enable them to make sense of Big Data generated via social media, SMS, aerial imagery & satellite imagery.

The incentives for humanitarian organizations to collaborate with CRI’s are obvious, especially if the latter (like QCRI) commits to making the resulting prototypes freely accessible and open source. But why should CRI’s collaborate with humanitarian organizations in the first place? Because the latter come with real-world challenges and unique research questions that many computer scientists are very interested in for several reasons. First, carrying out scientific research on real-world problems is of interest to the vast majority of computer scientists I collaborate with, both within QCRI and beyond. These scientists want to apply their skills to make the world a better place. Second, the research questions that humanitarian organizations bring enable computer scientists to differentiate themselves in the publishing world. Third, the resulting research can help advanced the field of computer science and advanced computing.

So why are we see not seeing more collaboration between CRI’s & humanitarian organizations? Because of this cognitive surplus mismatch. It takes a Director of Social Innovation (or related full-time position) to serve as a translational leader between CRI’s and humanitarian organizations. It takes someone (ideally a team) to match the problem owners and problem solvers; to facilitate and manage the collaboration between these two very different types of expertise and organizations. In sum, CRI’s can serve as an innovation pathway if the following three ingredients are in place: 1) Translation Leader; 2) Committed CRI; and 3) Committed Humanitarian Organization. These are necessary but not sufficient conditions for success.

While research institutes have a comparative advantage in R&D, they are not the best place to scale humanitarian technology prototypes. In order to take these prototypes to the next level, make them sustainable and have them develop into enterprise level software, they need to be taken up by for-profit companies. The majority of CRI’s (QCRI included) actually do have a mandate to incubate start-up companies. As such, we plan to spin-off some of the above platforms as independent companies in order to scale the technologies in a robust manner. Note that the software will remain free to use for humanitarian applications; other uses of the platform will require a paid license. Therein lies the end-to-end innovation path that computing research institutes can offer humanitarian organization vis-a-vis next generation humanitarian technologies.

As noted above, part of my involvement with the WHS Innovation Team entails working on an applied study to document and replicate this innovation pathway. As such, I am looking for feedback on the above as well as on the research methodology described below.

I plan to interview Microsoft Research, IBM Research, Yahoo Research, QCRI and other institutes as part of this research. More specifically, the interview questions will include:

  • Have you already partnered with humanitarian organizations? Why/why not?
  • If you have partnered with humanitarian organizations, what was the outcome? What were the biggest challenges? Was the partnership successful? If so, why? If not, why not?
  • If you have not yet partnered with humanitarian organizations, why not? What factors would be conducive to such partnerships and what factors serve as hurdles?
  • What are your biggest concerns vis-a-vis working with humanitarian groups?
  • What funding models did you explore if any?

I also plan to interview humanitarian organizations to better understand the prospects for this potential innovation pathway. More specifically, I plan to interview ICRC, UNHCR, UNICEF and OCHA using the following questions:

  • Have you already partnered with computing research groups? Why/why not?
  • If you have partnered with computing research groups, what was the outcome? What were the biggest challenges? Was the partnership successful? If so, why? If not, why not?
  • If you have not yet partnered with computing research groups, why not? What factors would be conducive to such partnerships and what factors serve as hurdles?
  • What are your biggest concerns vis-a-vis working with computing research groups?
  • What funding models did you explore if any?

My plan is to carry out the above semi-structured interviews in February-March 2015 along with secondary research. My ultimate aim with this deliverable is to develop a model to facilitate greater collaboration between computing research institutes and humanitarian organizations. To this end, I welcome feedback on all of the above (feel free to email me and/or add comments below). Thank you.

Bio

See also:

  • Research Framework for Next Generation Humanitarian Technology and Innovation [link]
  • From Gunfire at Sea to Maps of War: Implications for Humanitarian Innovation [link]

Establishing Social Media Hashtag Standards for Disaster Response

The UN Office for the Coordination of Humanitarian Affairs (OCHA) has just published an important, must-read report on the use of social media for disaster response. As noted by OCHA, this document was inspired by conversations with my team and I at QCRI. We jointly recognize that innovation in humanitarian technology is not enough. What is needed—and often lacking—is innovation in policymaking. Only then can humanitarian technology have widespread impact. This new think piece by OCHA seeks to catalyze enlightened policymaking.

ebolatags

I was pleased to provide feedback on earlier drafts of this new study and look forward to discussing the report’s recommendations with policymakers across the humanitarian space. In the meantime, many thanks to Roxanne Moore and Andrej Verity for making this report a reality. As Andrej notes in his blog post on this new study, the Filipino Government has just announced that “twitter will become another source of information for the Philippines official emergency response mechanism,” which will lead to an even more pressing Big (Crisis) Data challenge. The use of standardized hashtags will thus be essential.

hashtags-cartoon

The overflow of information generated during disasters can be as paralyzing to disaster response as the absence of information. While information scarcity has long characterized our information landscapes, today’s information-scapes are increasingly marked by an overflow of information—Big Data. To this end, encouraging the proactive standardization of hashtags may be one way to reduce this Big Data challenge. Indeed, standardized hashtags—i.e., more structured information—would enable paid emergency responders (as well as affected communities) to “better leverage crowdsourced information for operational planning and response.” At present, the Government of the Philippines seems to be the few actors that actually endorse the use of specific hashtags during major disasters as evidenced by their official crisis hashtags strategy.

The OCHA report thus proposes three hashtag standards and also encourages social media users to geo-tag their content during disasters. The latter can be done by enabling auto-GPS tagging or by using What3Words. Users should of course be informed of data-privacy considerations when geo-tagging their reports. As for the three hashtag standards:

  1. Early standardization of hashtags designating a specific disaster
  2. Standard, non-changing hashtag for reporting non-emergency needs
  3. Standard, non-changing hashtags for reporting emergency needs

1. As the OCHA think piece rightly notes, “News stations have been remarkably successful in encouraging early standardization of hashtags, especially during political events.” OCHA thus proposes that humanitarian organizations take a “similar approach for emergency response reporting and develop partnerships with Twitter as well as weather and news teams to publicly encourage such standardization. Storm cycles that create hurricanes and cyclones are named prior to the storm. For these events, an official hashtag should be released at the same time as the storm announcement.” For other hazards, “emergency response agencies should monitor the popular hashtag identifying a disaster, while trying to encourage a standard name.”

2. OCHA advocates for the use of #iSee, #iReport or #PublicRep for members of the public to designate tweets that refer to non-emergency needs such as “power lines, road closures, destroyed bridges, large-scale housing damage, population displacement or geographic spread (e.g., fire or flood).” When these hashtags are accompanied with GPS information, “responders can more easily identify and verify the information, therefore supporting more timely response & facilitating recovery.” In addition, responders can more easily create live crisis maps on the fly thanks to this structured, geo-tagged information.

3. As for standard hashtags for emergency reports, OCHA notes emergency calls are starting to give way to emergency SMS’s. Indeed, “Cell phone users will soon be able to send an SMS to a toll-free phone number. For emergency reporting, this new technology could dramatically alter the way the public interacts with nation-based emergency response call centers. It does not take a large imaginary leap to see the potential move from SMS emergency calls to social media emergency calls. Hashtags could be one way to begin reporting emergencies through social media.”

Most if not all countries have national emergency phone numbers already. So OCHA suggests using these existing, well-known numbers as the basis for social media hashtags. More specifically, an emergency hashtag would be composed of the country’s emergency number (such as 911 in the US, 999 in the UK, 133 in Austria, etc) followed by the country’s two-letter code (US, UK, AT respectively). In other words: #911US, #999UK, #133AT. Some countries, like Austria, have different emergency phone numbers for different types of emergencies. So these could also be used accordingly. OCHA recognizes that many “federal agencies fear that such a system would result in people reporting through social media outside of designated monitoring times. This is a valid concern. However, as with the implementation of any new technology in the public service, it will take time and extensive promotion to ensure effective use.”

Digital Humanitarians: The Book

Of course, “no monitoring system will be perfect in terms of low-cost, real-time analysis and high accuracy.” OCHA knows very well that there are a number of important limitations to the system they propose above. To be sure, “significant steps need to be taken to ensure that information flows from the public to response agencies and back to the public through improved efforts.” This is an important theme in my forthcoming book “Digital Humanitarians.”

bio

See also:

  • Social Media & Emergency Management: Supply and Demand [link]
  • Using AIDR to Automatically Classify Disaster Tweets [link]