Training Students to Extract Value from Big Data


New report by the National Research Council: “As the availability of high-throughput data-collection technologies, such as information-sensing mobile devices, remote sensing, internet log records, and wireless sensor networks has grown, science, engineering, and business have rapidly transitioned from striving to develop information from scant data to a situation in which the challenge is now that the amount of information exceeds a human’s ability to examine, let alone absorb, it. Data sets are increasingly complex, and this potentially increases the problems associated with such concerns as missing information and other quality concerns, data heterogeneity, and differing data formats.
The nation’s ability to make use of data depends heavily on the availability of a workforce that is properly trained and ready to tackle high-need areas. Training students to be capable in exploiting big data requires experience with statistical analysis, machine learning, and computational infrastructure that permits the real problems associated with massive data to be revealed and, ultimately, addressed. Analysis of big data requires cross-disciplinary skills, including the ability to make modeling decisions while balancing trade-offs between optimization and approximation, all while being attentive to useful metrics and system robustness. To develop those skills in students, it is important to identify whom to teach, that is, the educational background, experience, and characteristics of a prospective data-science student; what to teach, that is, the technical and practical content that should be taught to the student; and how to teach, that is, the structure and organization of a data-science program.
Training Students to Extract Value from Big Data summarizes a workshop convened in April 2014 by the National Research Council’s Committee on Applied and Theoretical Statistics to explore how best to train students to use big data. The workshop explored the need for training and curricula and coursework that should be included. One impetus for the workshop was the current fragmented view of what is meant by analysis of big data, data analytics, or data science. New graduate programs are introduced regularly, and they have their own notions of what is meant by those terms and, most important, of what students need to know to be proficient in data-intensive work. This report provides a variety of perspectives about those elements and about their integration into courses and curricula…”

3D printed maps could help the blind navigate their city


Springwise: “Modern technology has turned many of the things we consume from physical objects into pixels on a screen. While this has benefited the majority of us, those with sight difficulties don’t get along well with visual stimuli or touchscreen devices. In the past, we’ve seen Yahoo! Japan develop Hands On Search, a project that lets blind kids carry out web searches with 3D printed results. Now the country’s governmental department GSI is creating software that will enable those with visual impairments to print out 3D versions of online maps.
The official mapping body for Japan — much like the US Geological Survey — GSI already has paper maps for the blind, using embossed surfaces to mark out roads. It’s now developing a program that is able to do the same thing for digital maps.
The software first differentiates the highways, railway lines and walkways from the rest of the landscape. It then creates a 3D relief model that uses different textures to distinguish the features so that anyone running their finger along them will be able to determine what it is. The program also takes into account contour lines, creating accurate topographical representations of a particular area….
Website: www.gsi.go.jp

Welcome to The Open Standard


Welcome to The Open Standard.

From the beginning, Mozilla has dedicated itself to advocating for an open Web in wholehearted belief that open systems create more opportunity for everyone.
From its advocacy work to web literacy programs, to the creation of the Firefox browser, Mozilla has exemplified the journalism adage, “show, don’t tell.” It’s in that tradition that we’re excited to bring you The Open Standard, an original news site dedicated to covering the ideas and opinions that support the open, transparent and collaborative systems at work in our daily lives.
We advocate that open systems create healthier communities and more successful societies overall. We will cover everything from open source to open government and the need for transparency; privacy and security, the “Internet of Things” vs. “pervasive computing”, to education and if it’s keeping up with the technological changes. The bottom line? Open is better.
This is just the beginning. Over the next few months, The Open Standard will open itself to collaboration with you, our readers; everything from contributing to the site, to drawing our attention to uncovered issues, to crowdsourcing the news…”

Ebola: Can big data analytics help contain its spread?


at BBC News: “While emergency response teams, medical charities and non-governmental organisations struggle to contain the virus, could big data analytics help?
A growing number of data scientists believe so….
Mobile phones, widely owned in even the poorest countries in Africa, are proving to be a rich source of data in a region where other reliable sources are sorely lacking.
Orange Telecom in Senegal handed over anonymised voice and text data from 150,000 mobile phones to Flowminder, a Swedish non-profit organisation, which was then able to draw up detailed maps of typical population movements in the region.
Authorities could then see where the best places were to set up treatment centres, and more controversially, the most effective ways to restrict travel in an attempt to contain the disease.
The drawback with this data was that it was historic, when authorities really need to be able to map movements in real time. People’s movements tend to change during an epidemic.
This is why the US Centers for Disease Control and Prevention (CDC) is also collecting mobile phone mast activity data from mobile operators and mapping where calls to helplines are mostly coming from.

Population movement map of West AfricaMobile phone data from West Africa is being used to map population movements and predict how the Ebola virus might spread

A sharp increase in calls to a helpline from one particular area would suggest an outbreak and alert authorities to direct more resources there.
Mapping software company Esri is helping CDC to visualise this data and overlay other existing sources of data from censuses to build up a richer picture.
The level of activity at each mobile phone mast also gives a kind of heatmap of where people are and crucially, where and how far they are moving.

“We’ve never had this large-scale, anonymised mobile phone data before as a species,” says Nuria Oliver, a scientific director at mobile phone company Telefonica.

“The most positive impact we can have is to help emergency relief organisations and governments anticipate how a disease is likely to spread.
“Until now they had to rely on anecdotal information, on-the-ground surveys, police and hospital reports.”…

This Headline Is One of Many Experiments on You


Will Knight at MIT Technology Review: “On your way to this article, you probably took part in several experiments. You may have helped a search engine test a new way of displaying its results or an online retailer fine-tune an algorithm for recommending products. You may even have helped a news website decide which of two headlines readers are most likely to click on.
In other words, whether you realize it or not, the Web is already a gigantic, nonstop user-testing laboratory. Experimentation offers companies a powerful way to understand what customers want and how they are likely to behave, but it also seems that few people realize quite how much of it is going on.

This became clear in June, when Facebook experienced a backlash after publishing a study on the way negative emotions can spread across its network. The study, conducted by a team of internal researchers and academics, involved showing some people more negative posts than they would otherwise have seen, and then measuring how this affected their behavior. They in fact tended to post more negative content themselves, revealing a kind of “emotional contagion” (see “Facebook’s Emotion Study Is Just Its Latest Effort to Prod Users”).
Businesses have performed market research and other small experiments for years, but the practice has reached new levels of sophistication and complexity, largely because it is so easy to control the user experience on the Web, and then track how people’s behavior changes (see “What Facebook Knows”).
So companies with large numbers of users routinely tweak the information some of them see, and measure the resulting effect on their behavior—a practice known in the industry as A/B testing. Next time you see a credit card offer, for example, you might be one of a small group of users selected at random to see a new design. Or when you log onto Gmail, you may one of a chosen subset that gets to use a new feature developed by Google’s engineers.
“When doing things online, there’s a very large probability you’re going to be involved in multiple experiments every day,” Sinan Aral, a professor at MIT’s Sloan School of Management, said during a break at a conference for practitioners of large-scale user experiments last weekend in Cambridge, Massachusetts. “Look at Google, Amazon, eBay, Airbnb, Facebook—all of these businesses run hundreds of experiments, and they also account for a large proportion of Web traffic.”
At the Sloan conference, Ron Kohavi, general manager of the analysis and experimentation team at Microsoft, said each time someone uses the company’s search engine, Bing, he or she is probably involved in around 300 experiments. The insights that designers, engineers, and product managers can glean from these experiments can be worth millions of dollars in advertising revenue, Kohavi said…”

Putting Government Data to Work


U.S. Department of Commerce Press Release: “The Governance Lab (GovLab) at New York University today released “Realizing The Potential of Open Government Data: A Roundtable with the U.S. Department of Commerce,” a report on findings and recommendations for ways the U.S. Commerce Department can improve its data management, dissemination and use. The report summarizes a June 2014 Open Data Roundtable, co-hosted by The GovLab and the White House Office of Science and Technology Policy with the Commerce Department, which brought together Commerce data providers and 25 representatives from the private sector and nonprofit organizations for an action-oriented dialogue on data issues and potential solutions. The GovLab is convening a series of other Open Data Roundtables in its mission to help make government more effective and connected to the public through technology.

“We were honored to work with the White House and the Department of Commerce to convene this event,” said Joel Gurin, senior advisor at The GovLab and project director of the Open Data 500 and the Roundtable Series. “The Department’s commitment to engaging with its data customers opens up great opportunities for public-private collaboration.”
Under Secretary of Commerce for Economic Affairs Mark Doms said, “At the Commerce Department, we are only at the beginning of our open data effort. We share the goals and objectives embodied by the call of the Open Data 500: to deliver data that is valuable to industry and that provides greater economic opportunity for millions of Americans.” …”

Big Thinkers. Big Data. Big Opportunity: Announcing The LinkedIn Economic Graph Challeng


at Linkedin Official Blog: “LinkedIn’s vision is to create economic opportunity for every member of the global workforce. Facilitating economic empowerment is a big task that will require bold thinking by smart, passionate individuals and groups. Today, we’re kicking off an initiative that aims to encourage this type of big thinking: the LinkedIn Economic Graph Challenge.
The LinkedIn Economic Graph Challenge is an idea that emerged from the development of the Economic Graph, a digital mapping of the global economy, comprised of a profile for every professional, company, job opportunity, the skills required to obtain those opportunities, every higher education organization, and all the professionally relevant knowledge associated with each of these entities. With these elements in place, we can connect talent with opportunity at massive scale.
We are launching the LinkedIn Economic Graph Challenge to encourage researchers, academics, and data-driven thinkers to propose how they would use data from LinkedIn to solve some of the most challenging economic problems of our times. We invite anyone who is interested to submit your most innovative, ambitious ideas. In return, we will recognize the three strongest proposals for using data from LinkedIn to generate a positive impact on the global economy, and present the team and/or individual with a $25,000 (USD) research award and the resources to complete their proposed research, with the potential to have it published….
We look forward to your submissions! For more information, please visit the LinkedIn Economic Graph Challenge website….”

Canada's Action Plan on Open Government 2014-2016


Draft action plan: “Canada’s second Action Plan on Open Government consists of twelve commitments that will advance open government principles in Canada over the next two years and beyond. The Directive on Open Government, new policy direction to federal departments and agencies on open government, will provide foundational support for each of the additional commitments which fall under three streams: Open Data, Open Information, and Open Dialogue.
Figure 1: Our Commitments
Open Government Directive Diagram

 

More:

Table of Contents

 
 

Data revolution: How the UN is visualizing the future


Kate Krukiel at Microsoft Government: “…world leaders met in New York for the 69th session of the United Nations (UN) General Assembly. Progress toward achieving the eight Millennium Development Goals (MDGs) by the December 2015 target date—just 454 days away—was top of mind. So was the post-2015 agenda, which will pick up where the MDGs leave off. Ahead of the meetings, the UN Millennium Campaign asked Microsoft to build real-time visualizations of the progress on each goal—based on data spanning 21 targets, 60 indicators, and about 190 member countries. With the data visualizations we created (see them at http://www.mdgleaders.org/), UN and global leaders can decide where to focus in the next 15 months and, more importantly, where change needs to happen post-2015. Their experience offers three lessons for governments:

1. Data has a shelf life.

Since the MDGs were launched in 2000, the UN has relied on annual reports to assess its progress. But in August, UN Secretary-General Ban Ki-moon called for a “data revolution for sustainable development”, which in effect makes real-time data visualization a requirement, not just for tracking the MDGs, but for everything from Ebola to climate change….

2.Governments need visualization tools.

Just as the UN is using data visualization to track its progress and plan for the future, you can use the technology to better understand the massive amounts of data you collect—on everything from water supply and food prices to child mortality and traffic jams. Data visualization technology makes it possible to pull insights from historical data, develop forecasts, and spot gaps in your data far easier than you can with raw data. As they say, a picture is worth a thousand words. To get a better idea of what’s possible, check out the MDG visualizations Microsoft created for the UN using our Power BI tool.

3.The private sector can help.

The UN called on the private sector to assist in determining the exact MDG progress and inspire ongoing global efforts. …

Follow the UN’s lead and join the #datarevolution now, if you haven’t already. It’s an opportunity to work across silos and political boundaries to address the world’s most pressing problems. It takes citizens’ points of view into account through What People Want. And it extends to the private sector, where expertise in using technology to create a sustainable future already exists. I encourage all government leaders to engage. To follow where the UN takes its revolution, watch for updates on the Data Revolution Group website or follow them on Twitter @data_rev….”