Could Big Data Help End Hunger in Africa?


Lenny Ruvaga at VOA News: “Computer algorithms power much of modern life from our Facebook feeds to international stock exchanges. Could they help end malnutrition and hunger in Africa? The International Center for Tropical Agriculture thinks so.

The International Center for Tropical Agriculture has spent the past four years developing the Nutrition Early Warning System, or NEWS.

The goal is to catch the subtle signs of a hunger crisis brewing in Africa as much as a year in advance.

CIAT says the system uses machine learning. As more information is fed into the system, the algorithms will get better at identifying patterns and trends. The system will get smarter.

Information Technology expert Andy Jarvis leads the project.

“The cutting edge side of this is really about bringing in streams of information from multiple sources and making sense of it. … But it is a huge volume of information and what it does, the novelty then, is making sense of that using things like artificial intelligence, machine learning, and condensing it into simple messages,” he said.

Other nutrition surveillance systems exist, like FEWSnet, the Famine Early Warning System Network which was created in the mid-1980s.

But CIAT says NEWS will be able to draw insights from a massive amount of diverse data enabling it to identify hunger risks faster than traditional methods.

“What is different about NEWS is that it pays attention to malnutrition, not just drought or famine, but the nutrition outcome that really matters, malnutrition especially in women and children. For the first time, we are saying these are the options way ahead of time. That gives policy makers an opportunity to really do what they intend to do which is make the lives of women and children better in Africa,” said Dr. Mercy Lung’aho, a CIAT nutrition expert.

While food emergencies like famine and drought grab headlines, the International Center for Tropical Agriculture says chronic malnutrition affects one in four people in Africa, taking a serious toll on economic growth and leaving them especially vulnerable in times of crisis….(More)”.

Our path to better science in less time using open data science tools


Julia S. Stewart Lowndes et al in Nature: “Reproducibility has long been a tenet of science but has been challenging to achieve—we learned this the hard way when our old approaches proved inadequate to efficiently reproduce our own work. Here we describe how several free software tools have fundamentally upgraded our approach to collaborative research, making our entire workflow more transparent and streamlined. By describing specific tools and how we incrementally began using them for the Ocean Health Index project, we hope to encourage others in the scientific community to do the same—so we can all produce better science in less time.

Figure 1: Better science in less time, illustrated by the Ocean Health Index project.
Figure 1

Every year since 2012 we have repeated Ocean Health Index (OHI) methods to track change in global ocean health36,37. Increased reproducibility and collaboration has reduced the amount of time required to repeat methods (size of bubbles) with updated data annually, allowing us to focus on improving methods each year (text labels show the biggest innovations). The original assessment in 2012 focused solely on scientific methods (for example, obtaining and analysing data, developing models, calculating, and presenting results; dark shading). In 2013, by necessity we gave more focus to data science (for example, data organization and wrangling, coding, versioning, and documentation; light shading), using open data science tools. We established R as the main language for all data preparation and modelling (using RStudio), which drastically decreased the time involved to complete the assessment. In 2014, we adopted Git and GitHub for version control, project management, and collaboration. This further decreased the time required to repeat the assessment. We also created the OHI Toolbox, which includes our R package ohicore for core analytical operations used in all OHI assessments. In subsequent years we have continued (and plan to continue) this trajectory towards better science in less time by improving code with principles of tidy data33; standardizing file and data structure; and focusing more on communication, in part by creating websites with the same open data science tools and workflow. See text and Table 1 for more details….(More)”

Information for the People: Tunisia Embraces Open Government, 2011–2016


Case study by Tristan Dreisback at Innovations for Successful Societies: “In January 2011, mass demonstrations in Tunisia ousted a regime that had tolerated little popular participation, opening the door to a new era of transparency. The protesters demanded an end to the secrecy that had protected elite privilege. Five months later, the president issued a decree that increased citizen access to government data and formed a steering committee to guide changes in information practices, building on small projects already in development. Advocates in the legislature and the public service joined with civil society leaders to support a strong access-to-information policy, to change the culture of public administration, and to secure the necessary financial and technical resources to publish large quantities of data online in user-friendly formats. Several government agencies launched their own open-data websites. External pressure, coupled with growing interest from civil society and legislators, helped keep transparency reforms on the cabinet office agenda despite frequent changes in top leadership. In 2016, Tunisia adopted one of the world’s strongest laws regarding access to information. Although members of the public did not put all of the resources to use immediately, the country moved much closer to having the data needed to improve access to services, enhance government performance, and support the evidence-based deliberation on which a healthy democracy depended…(More)”

How can we study disguised propaganda on social media? Some methodological reflections


Jannick Schou and Johan Farkas at DataDrivenJournalism: ’Fake news’ has recently become a seemingly ubiquitous concept among journalists, researchers, and citizens alike. With the rise of platforms such as Facebook and Twitter, it has become possible to spread deliberate forms of misinformation in hitherto unforeseen ways. This has also spilled over into the political domain, where new forms of (disguised) propaganda and false information have recently begun to emerge. These new forms of propaganda have very real effects: they serve to obstruct political decision-making processes, instil false narratives within the general public, and add fuel to already heated sites of political conflict. They represent a genuine democratic problem.

Yet, so far, both critical researchers and journalists have faced a number of issues and challenges when attempting to understand these new forms of political propaganda. Simply put: when it comes to disguised propaganda and social media, we know very little about the actual mechanisms through which such content is produced, disseminated, and negotiated. One of the key explanations for this might be that fake profiles and disguised political agendas are incredibly difficult to study. They present a serious methodological challenge. This is not only due to their highly ephemeral nature, with Facebook pages being able to vanish after only a few days or hours, but also because of the anonymity of its producers. Often, we simply do not know who is disseminating what and with what purpose. This makes it difficult for us to understand and research exactly what is going on.

This post takes its point of departure from a new article published in the international academic journal New Media & Society. Based on the research done for this article, we want to offer some methodological reflections as to how disguised propaganda might be investigated. How can we research fake and disguised political agendas? And what methodological tools do we have at our disposal?…

two main methodological advices spring to mind. First of all: collect as much data as you can in as many ways as possible. Make screenshots, take detailed written observations, use data scraping, and (if possible) participate in citizen groups. One of the most valuable resources we had at our disposal was the set of heterogeneous data we collected from each page. Using this allowed us to carefully dissect and retrace the complex set of practices involved in each page long after they were gone. While we certainly tried to be as systematic in our data collection as possible, we also had to use every tool at our disposal. And we had to constantly be on our toes. As soon as a page emerged, we were there: ready to write down notes and collect data.

Second: be willing to participate and collaborate. Our research showcases the immense potential in researchers (and journalists) actively collaborating with citizen groups and grassroots movements. Using the collective insights and attention of this group allowed us to quickly find and track down pages. It gave us renewed methodological strength. Collaborating across otherwise closed boundaries between research and journalism opens up new avenues for deeper and more detailed insights….(More)”

Big Data: A New Empiricism and its Epistemic and Socio-Political Consequences


Chapter by Gernot Rieder and Judith Simon in by Berechenbarkeit der Welt? Philosophie und Wissenschaft im Zeitalter von Big Data: “…paper investigates the rise of Big Data in contemporary society. It examines the most prominent epistemological claims made by Big Data proponents, calls attention to the potential socio-political consequences of blind data trust, and proposes a possible way forward. The paper’s main focus is on the interplay between an emerging new empiricism and an increasingly opaque algorithmic environment that challenges democratic demands for transparency and accountability. It concludes that a responsible culture of quantification requires epistemic vigilance as well as a greater awareness of the potential dangers and pitfalls of an ever more data-driven society….(More)”.

Data Collaboratives: exchanging data to create public value across Latin America and the Caribbean


Stefaan Verhulst, Andrew Young and Prianka Srinivasan at IADB’s Abierto al Publico: “Data is playing an ever-increasing role in bolstering businesses across Latin America – and the rest of the word. In Brazil, Mexico and Colombia alone, the revenue from Big Data is calculated at more than US$603.7 million, a market that is only set to increase as more companies across Latin America and the Caribbean embrace data-driven strategies to enhance their bottom-line. Brazilian banking giant Itau plans to create six data centers across the country, and already uses data collected from consumers online to improve cross-selling techniques and streamline their investments. Data from web-clicks, social media profiles, and telecommunication services is fueling a new generation of entrepreneurs keen to make big dollars from big data.

What if this same data could be used not just to improve business, but to improve the collective well-being of our communities, public spaces, and cities? Analysis of social media data can offer powerful insights to city officials into public trends and movements to better plan infrastructure and policies. Public health officials and humanitarian workers can use mobile phone data to, for instance, map human mobility and better target their interventions. By repurposing the data collected by companies for their business interests, governments, international organizations and NGOs can leverage big data insights for the greater public good.

Key question is thus: How to unlock useful data collected by corporations in a responsible manner and ensure its vast potential does not go to waste?

Data Collaboratives” are emerging as a possible answer. Data collaboratives are a new type of public-private partnerships aimed at creating public value by exchanging data across sectors.

Research conducted by the GovLab finds that Data Collaboratives offer several potential benefits across a number of sectors, including humanitarian and anti-poverty efforts, urban planning, natural resource stewardship, health, and disaster management. As a greater number of companies in Latin America look to data to spur business interests, our research suggests that some companies are also sharing and collaborating around data to confront some of society’s most pressing problems.

Consider the following Data Collaboratives that seek to enhance…(More)”

Digital platforms and democracy


Ricard Espelt and Monica Garriga  at Open Democracy: “The impact of digital platforms in recent years affects all areas and all sorts of organizations: from production to consumption, from political parties to social movements, from business to public administration, trade unions, universities or the mass media. The disruption they generate is cross-section and intergenerational. Undoubtedly, their outstanding assets – at least from a discursive point of view –, are self-management and disintermediation. Today, through technology, people can participate actively in processes related to any particular activity. This is why we often talk about digital platforms as tools for democratizing participation, overcoming as they do the traditional tyranny of space and time. If we analyze them in detail, however, and look at the organizations that promote them, we realize that the improvement in citizen involvement tends to vary, sometimes considerably, as does the logic behind their approach…..

La Teixidora, a democratic digital platform

Being aware now of the risks of partial evaluation of the impact of technology and the key elements to be considered in analyzing it, let us return to our starting point: democratizing participation. Given the importance of local assessment of global digital tools, let us now see the case of the multimedia platform La Teixidora, which allows us to synthesize the aspects which, in our opinion, shape democratic participation.

Platform cooperativism or open cooperativism, whether it focuses on the social strength of cooperative values or on the need to reappropriate common goods, calls for a detailed critical review of the local activity of its digital platforms.

This initiative, launched in 2016 in Barcelona, organizes in real time a collaborative structure with the aim of mapping distributed knowledge generated in different parts of the city during conferences, meetings, workshops and other offline meeting formats related to technopolitics and the commons. To do this, it appropriates several open source tools (collaborative editor, wiki, content storage spaces) and uses a Creative Commons license which, while recognizing authorship, allows anyone to adapt the contents and even use them commercially. Two significant apps illustrate the value of its functionalities in relation to democratizing participation:

  1. In March 2016 La Teixidora covered, with a team of some twenty people, a debate on Collaborative Economy (Economies Col·laboratives Procomuns). The classified data were then transferred to the Decidim Barcelona platform, which has helped to define, through a broad participatory process, the Municipal Action Plan of the Barcelona City Council.
  2. At the same time, the tool has been used to monitor the fifteen teams which have been following the economic development program La Comunificadora, whose aim is the promotion of social transformation projects and the advancement of entrepreneurship. Through La Teixidora, the participants have been able to establish a space for exchanging knowledge among them, with the mentors, with the city service managers and with citizens in general. All its contents are open and reusable.

In short, through this platform, both processes have been able not only to contribute proposals, but also to form an open learning space. And by mapping participation, which makes these processes – both of which are promoted by the Public Administration – transparent and accountable, thus improving their democratic quality. At the same time, the information and the learning from their use are helping to redesign the technological platform itself and adapt it to the needs of the communities involved….(More)”.

Twitter as a data source: An overview of tools for journalists


Wasim Ahmed at Data Driven Journalism: “Journalists may wish to use data from social media platforms in order to provide greater insight and context to a news story. For example, journalists may wish to examine the contagion of hashtags and whether they are capable of achieving political or social change. Moreover, newsrooms may also wish to tap into social media posts during unfolding crisis events. For example, to find out who tweeted about a crisis event first, and to empirically examine the impact of social media.

Furthermore, Twitter users and accounts such as WikiLeaks may operate outside the constraints of traditional journalism, and therefore it becomes important to have tools and mechanisms in place in order to examine these kinds of influential users. For example, it was found that those who were backing Marine Le Pen on Twitter could have been users who had an affinity to Donald Trump.

There remains a number of different methods for analysing social media data. Take text analytics, for example, which can include using sentiment analysis to place bulk social media posts into categories of a particular feeling, such as positive, negative, or neutral. Or machine learning, which can automatically assign social media posts to a number of different topics.

There are other methods such as social network analysis, which examines online communities and the relationships between them. A number of qualitative methodologies also exist, such as content analysis and thematic analysis, which can be used to manually label social media posts. From a journalistic perspective, network analysis may be of importance initially via tools such as NodeXL. This is because it can quickly provide an overview of influential Twitter users alongside a topic overview.

From an industry standpoint, there has been much focus on gaining insight into users’ personalities, through services such as IBM Watson’s Personality Insights service. This uses linguistic analytics to derive intrinsic personality insights, such as emotions like anxiety, self-consciousness, and depression. This information can then be used by marketers to target certain products; for example, anti-anxiety medication to users who are more anxious…(An overview of tools for 2017).”

UK government watchdog examining political use of data analytics


“Given the big data revolution, it is understandable that political campaigns are exploring the potential of advanced data analysis tools to help win votes,” Elizabeth Denham, the information commissioner, writes on the ICO’s blog. However, “the public have the right to expect” that this takes place in accordance with existing data protection laws, she adds.

Political parties are able to use Facebook to target voters with different messages, tailoring the advert to recipients based on their demographic. In the 2015 UK general election, the Conservative party spent £1.2 million on Facebook campaigns and the Labour party £16,000. It is expected that Labour will vastly increase that spend for the general election on 8 June….

Political parties and third-party companies are allowed to collect data from sites like Facebook and Twitter that lets them tailor these ads to broadly target different demographics. However, if those ads target identifiable individuals, it runs afoul of the law….(More)”