Twitter releasing trove of user data to scientists for research


Joe Silver at ArsTechnica: “Twitter has a 200-million-strong and ever-growing user base that broadcasts 500 million updates daily. It has been lauded for its ability to unsettle repressive political regimes, bring much-needed accountability to corporations that mistreat their customers, and combat other societal ills (whether such characterizations are, in fact, accurate). Now, the company has taken aim at disrupting another important sphere of human society: the scientific research community.
Back in February, the site announced its plan—in collaboration with Gnip—to provide a handful of research institutions with free access to its data sets from 2006 to the present. It’s a pilot program called “Twitter Data Grants,” with the hashtag #DataGrants. At the time, Twitter’s engineering blog explained the plan to enlist grant applications to access its treasure trove of user data:

Twitter has an expansive set of data from which we can glean insights and learn about a variety of topics, from health-related information such as when and where the flu may hit to global events like ringing in the new year. To date, it has been challenging for researchers outside the company who are tackling big questions to collaborate with us to access our public, historical data. Our Data Grants program aims to change that by connecting research institutions and academics with the data they need.

In April, Twitter announced that, after reviewing the more than 1,300 proposals submitted from more than 60 different countries, it had selected six institutions to provide with data access. Projects approved included a study of foodborne gastrointestinal illnesses, a study measuring happiness levels in cities based on images shared on Twitter, and a study using geosocial intelligence to model urban flooding in Jakarta, Indonesia. There’s even a project exploring the relationship between tweets and sports team performance.
Twitter did not directly respond to our questions on Tuesday afternoon regarding the specific amount and types of data the company is providing to the six institutions. But in its privacy policy, Twitter explains that most user information is intended to be broadcast widely. As a result, the company likely believes that sharing such information with scientific researchers is well within its rights, as its services “are primarily designed to help you share information with the world,” Twitter says. “Most of the information you provide us is information you are asking us to make public.”
While mining such data sets will undoubtedly aid scientists in conducting experiments for which similar data was previously either unavailable or quite limited, these applications raise some legal and ethical questions. For example, Scientific American has asked whether Twitter will be able to retain any legal rights to scientific findings and whether mining tweets (many of which are not publicly accessible) for scientific research when Twitter users have not agreed to such uses is ethically sound.
In response, computational epidemiologists Caitlin Rivers and Bryan Lewis have proposed guidelines for ethical research practices when using social media data, such as avoiding personally identifiable information and making all the results publicly available….”

HHS releases new data and tools to increase transparency on hospital utilization and other trends


Pressrelease: “With more than 2,000 entrepreneurs, investors, data scientists, researchers, policy experts, government employees and more in attendance, the Department of Health and Human Services (HHS) is releasing new data and launching new initiatives at the annual Health Datapalooza conference in Washington, D.C.
Today, the Centers for Medicare & Medicaid Services (CMS) is releasing its first annual update to the Medicare hospital charge data, or information comparing the average amount a hospital bills for services that may be provided in connection with a similar inpatient stay or outpatient visit. CMS is also releasing a suite of other data products and tools aimed to increase transparency about Medicare payments. The data trove on CMS’s website now includes inpatient and outpatient hospital charge data for 2012, and new interactive dashboards for the CMS Chronic Conditions Data Warehouse and geographic variation data. Also today, the Food and Drug Administration (FDA) will launch a new open data initiative. And before the end of the conference, the Office of the National Coordinator for Health Information Technology (ONC) will announce the winners of two data challenges.
“The release of these data sets furthers the administration’s efforts to increase transparency and support data-driven decision making which is essential for health care transformation,” said HHS Secretary Kathleen Sebelius.
“These public data resources provide a better understanding of Medicare utilization, the burden of chronic conditions among beneficiaries and the implications for our health care system and how this varies by where beneficiaries are located,” said Bryan Sivak, HHS chief technology officer. “This information can be used to improve care coordination and health outcomes for Medicare beneficiaries nationwide, and we are looking forward to seeing what the community will do with these releases. Additionally, the openFDA initiative being launched today will for the first time enable a new generation of consumer facing and research applications to embed relevant and timely data in machine-readable, API-based formats.”
2012 Inpatient and Outpatient Hospital Charge Data
The data posted today on the CMS website provide the first annual update of the hospital inpatient and outpatient data released by the agency last spring. The data include information comparing the average charges for services that may be provided in connection with the 100 most common Medicare inpatient stays at over 3,000 hospitals in all 50 states and Washington, D.C. Hospitals determine what they will charge for items and services provided to patients and these “charges” are the amount the hospital generally bills for those items or services.
With two years of data now available, researchers can begin to look at trends in hospital charges. For example, average charges for medical back problems increased nine percent from $23,000 to $25,000, but the total number of discharges decreased by nearly 7,000 from 2011 to 2012.
In April, ONC launched a challenge – the Code-a-Palooza challenge – calling on developers to create tools that will help patients use the Medicare data to make health care choices. Fifty-six innovators submitted proposals and 10 finalists are presenting their applications during Datapalooza. The winning products will be announced before the end of the conference.
Chronic Conditions Warehouse and Dashboard
CMS recently released new and updated information on chronic conditions among Medicare fee-for-service beneficiaries, including:

  • Geographic data summarized to national, state, county, and hospital referral regions levels for the years 2008-2012;
  • Data for examining disparities among specific Medicare populations, such as beneficiaries with disabilities, dual-eligible beneficiaries, and race/ethnic groups;
  • Data on prevalence, utilization of select Medicare services, and Medicare spending;
  • Interactive dashboards that provide customizable information about Medicare beneficiaries with chronic conditions at state, county, and hospital referral regions levels for 2012; and
  • Chartbooks and maps.

These public data resources support the HHS Initiative on Multiple Chronic Conditions by providing researchers and policymakers a better understanding of the burden of chronic conditions among beneficiaries and the implications for our health care system.
Geographic Variation Dashboard
The Geographic Variation Dashboards present Medicare fee-for-service per-capita spending at the state and county levels in interactive formats. CMS calculated the spending figures in these dashboards using standardized dollars that remove the effects of the geographic adjustments that Medicare makes for many of its payment rates. The dashboards include total standardized per capita spending, as well as standardized per capita spending by type of service. Users can select the indicator and year they want to display. Users can also compare data for a given state or county to the national average. All of the information presented in the dashboards is also available for download from the Geographic Variation Public Use File.
Research Cohort Estimate Tool
CMS also released a new tool that will help researchers and other stakeholders estimate the number of Medicare beneficiaries with certain demographic profiles or health conditions. This tool can assist a variety of stakeholders interested in specific figures on Medicare enrollment. Researchers can also use this tool to estimate the size of their proposed research cohort and the cost of requesting CMS data to support their study.
Digital Privacy Notice Challenge
ONC, with the HHS Office of Civil Rights, will be awarding the winner of the Digital Privacy Notice Challenge during the conference. The winning products will help consumers get notices of privacy practices from their health care providers or health plans directly in their personal health records or from their providers’ patient portals.
OpenFDA
The FDA’s new initiative, openFDA, is designed to facilitate easier access to large, important public health datasets collected by the agency. OpenFDA will make FDA’s publicly available data accessible in a structured, computer readable format that will make it possible for technology specialists, such as mobile application creators, web developers, data visualization artists and researchers to quickly search, query, or pull massive amounts of information on an as needed basis. The initiative is the result of extensive research to identify FDA’s publicly available datasets that are often in demand, but traditionally difficult to use. Based on this research, openFDA is beginning with a pilot program involving millions of reports of drug adverse events and medication errors submitted to the FDA from 2004 to 2013. The pilot will later be expanded to include the FDA’s databases on product recalls and product labeling.
For more information about CMS data products, please visit http://www.cms.gov/Research-Statistics-Data-and-Systems/Research-Statistics-Data-and-Systems.html.
For more information about today’s FDA announcement visit: http://www.fda.gov/NewsEvents/Newsroom/PressAnnouncements/UCM399335 or http://open.fda.gov/

Closing the Feedback Loop: Can Technology Bridge the Accountability Gap


(WorldBank) Book edited by Björn-Sören Gigler and Savita Bailur:  “This book is a collection of articles, written by both academics and practitioners as an evidence base for citizen engagement through information and communication technologies (ICTs). In it, the authors ask: how do ICTs empower through participation, transparency and accountability? Specifically, the authors examine two principal questions: Are technologies an accelerator to closing the “accountability gap” – the space between the supply (governments, service providers) and demand (citizens, communities, civil society organizations or CSOs) that requires bridging for open and collaborative governance? And under what conditions does this occur? The introductory chapters lay the theoretical groundwork for understanding the potential of technologies to achieving intended goals. Chapter 1 takes us through the theoretical linkages between empowerment, participation, transparency and accountability. In Chapter 2, the authors devise an informational capability framework, relating human abilities and well-being to the use of ICTs. The chapters to follow highlight practical examples that operationalize ICT-led initiatives. Chapter 3 reviews a sample of projects targeting the goals of transparency and accountability in governance to make preliminary conclusions around what evidence exists to date, and where to go from here. In chapter 4, the author reviews the process of interactive community mapping (ICM) with examples that support general local development and others that mitigate natural disasters. Chapter 5 examines crowdsourcing in fragile states to track aid flows, report on incitement or organize grassroots movements. In chapter 6, the author reviews Check My School (CMS), a community monitoring project in the Philippines designed to track the provision of services in public schools. Chapter 7 introduces four key ICT-led, citizen-governance initiatives in primary health care in Karnataka, India. Chapter 8 analyzes the World Bank Institute’s use of ICTs in expanding citizen project input to understand the extent to which technologies can either engender a new “feedback loop” or ameliorate a “broken loop”. The authors’ analysis of the evidence signals ICTs as an accelerator to closing the “accountability gap”. In Chapter 9, the authors conclude with the Loch Ness model to illustrate how technologies contribute to shrinking the gap, why the gap remains open in many cases, and what can be done to help close it. This collection is a critical addition to existing literature on ICTs and citizen engagement for two main reasons: first, it is expansive, covering initiatives that leverage a wide range of technology tools, from mobile phone reporting to crowdsourcing to interactive mapping; second, it is the first of its kind to offer concrete recommendations on how to close feedback loops.”

Open government: getting beyond impenetrable online data


Jed Miller in The Guardian: “Mathematician Blaise Pascal famously closed a long letter by apologising that he hadn’t had time to make it shorter. Unfortunately, his pithy point about “download time” is regularly attributed to Mark Twain and Henry David Thoreau, probably because the public loves writers more than it loves statisticians. Scientists may make things provable, but writers make them memorable.
The World Bank confronted a similar reality of data journalism earlier this month when it revealed that, of the 1,600 bank reports posted online on from 2008 to 2012, 32% had never been downloaded at all and another 40% were downloaded under 100 times each.
Taken together, these cobwebbed documents represent millions of dollars in World Bank funds and hundreds of thousands of person-hours, spent by professionals who themselves represent millions of dollars in university degrees. It’s difficult to see the return on investment in producing expert research and organising it into searchable web libraries when almost three quarters of the output goes largely unseen.
The World Bank works at a scale unheard of by most organisations, but expert groups everywhere face the same challenges. Too much knowledge gets trapped in multi-page pdf files that are slow to download (especially in low-bandwidth areas), costly to print, and unavailable for computer analysis until someone manually or automatically extracts the raw data.
Even those who brave the progress bar find too often that urgent, incisive findings about poverty, health, discrimination, conflict or social change are presented in prose written by and for high-level experts, rendering it impenetrable to almost everyone else. Information isn’t just trapped in pdfs; it’s trapped in PhDs.
Governments and NGOs are beginning to realise that digital strategy means more than posting a document online, but what will it take for these groups to change not just their tools, but their thinking? It won’t be enough to partner with WhatsApp or hire GrumpyCat.
I asked strategists from the development, communications and social media fields to offer simple, “Tweetable” suggestions for how the policy community can become better communicators.

For nonprofits and governments that still publish 100-page pdfs on their websites and do not optimise the content to share in other channels such as social: it is a huge waste of time and ineffective. Stop it now.

– Beth Kanter, author and speaker. Beth’s Blog: How Nonprofits Can Use Social Media

Treat text as #opendata so infomediaries can mash it up and make it more accessible (see, for example federalregister.gov) and don’t just post and blast: distribute information in a targeted way to those most likely to be interested.

– Beth Noveck, director at the Governance Lab and former director at White House Open Government Initiative

Don’t be boring. Sounds easy, actually quite hard, super-important.

– Eli Pariser, CEO of Upworthy

Surprise me. Uncover the key finding that inspired you, rather than trying to tell it all at once and show me how the world could change because of it.

– Jay Golden, co-founder of Wakingstar Storyworks

For the Bank or anyone who is generating policy information they actually want people to use, they must actually write it for the user, not for themselves. As Steve Jobs said, ‘Simple can be harder than complex’.

– Kristen Grimm, founder and president at Spitfire Strategies

The way to reach the widest audience is to think beyond content format and focus on content strategy.

– Laura Silber, director of public affairs at Open Society Foundations

Open the door to policy work with short, accessible pieces – a blog post, a video take, infographics – that deliver the ‘so what’ succinctly.

– Robert McMahon, editor at Council on Foreign Relations

Policy information is more usable if it’s linked to corresponding actions one can take, or if it helps stir debate.  Also, whichever way you slice it, there will always be a narrow market for raw policy reports … that’s why explainer sites, listicles and talking heads exist.

– Ory Okolloh, director of investments at Omidyar Network and former public policy and government relations manager at Google Africa
Ms Okolloh, who helped found the citizen reporting platform Ushahidi, also offered a simple reminder about policy reports: “‘Never gets downloaded’ doesn’t mean ‘never gets read’.” Just as we shouldn’t mistake posting for dissemination, we shouldn’t confuse popularity with influence….”

Collaborative approaches to public sector innovation: A scoping study


Paper by K Szkuta, R Pizzicannella, D Osimo in Telecommunications Policy: “In the last 15 years, European countries have invested considerable resources to provide e-government services. Despite of its increasing availability, its level of adoption has not been satisfying. On the other hand, over the last years, coinciding with the web 2.0 trend, the e-government services co-produced by citizens start to appear, often without the support, acknowledgement and even awareness of the government. This trend stems from a well-established tradition of offline co- production of public services, i.e. services provided by the voluntary sector, but brought to an unprecedented scale thanks to the advent of web 2.0. Still, the concept remains not well-defined and its impact is not yet well studied. The paper explores on a limited sets of cases what does it mean to collaboratively deliver online public services; what are the success factors based on the cases under study and what are the incentives for service providers (other than public administration), citizens as users and public administration. The authors propose an ostensive definition of the collaborative delivery of public services: collaborative public services are created and run by government, civil society or by private sector building on the re-use of government data or citizens data. Those services are focused on public goods delivery (e.g. health, education, public transport) and are meant to change the traditional government services by engaging in an open dialogue with public administration about the best way to deliver those services. The analysis of six case studies of innovative collaborative online public services suggests that the online collaborative public service delivery increases its quality with the users׳ growth contrary to the traditional offline service delivery. The study results indicate that the current developers interest lies in delivering complementary services to the government run services rather than substitutive services. The authors propose also the initial list of success factors, enabling conditions, and benefits for all main stakeholders (users, innovators and public administration).”

Global democracy and the democratic minimum: Why a procedural account alone is insufficient


Paper by Klaus Dingwerth in the European Journal of International Relations: “In this critical comment on the global democracy debate, I take stock of contemporary proposals for democratizing global governance. In the first part of the article, I show that, empirically, many international institutions are now evaluated in terms of their democratic credentials. At the same time, the notions of democracy that underpin such evaluations are often very formalistic. They focus on granting access to civil society organizations, making policy-relevant documents available online or establishing global parliamentary assemblies to give citizens a voice in the decision-making of international organizations. In the second part, I challenge this focus on formal procedures and argue that a normatively persuasive conception of global democracy would shift our focus to areas such as health, education and subsistence. Contrary to much contemporary thinking about global democracy, I thus defend the view that the institutions we have are sufficiently democratic. What is needed are not better procedures, but investments that help the weaker members of global society to make effective use of the democracy-relevant institutions that exist in contemporary international politics”

The rise of open data driven businesses in emerging markets


Alla Morrison at the Worldbank blog:

Key findings —

  • Many new data companies have emerged around the world in the last few years. Of these companies, the majority use some form of government data.
  • There are a large number of data companies in sectors with high social impact and tremendous development opportunities.
  • An actionable pipeline of data-driven companies exists in Latin America and in Asia. The most desired type of financing is equity, followed by quasi-equity in the amounts ranging from $100,000 to $5 million, with averages of between $2 and $3 million depending on the region. The total estimated need for financing may exceed $400 million.

“The economic value of open data is no longer a hypothesis
How can one make money with open data which is akin to air – free and open to everyone? Should the World Bank Group be in the catalyzer role for a sector that is just emerging?  And if so, what set of interventions would be the most effective? Can promoting open data-driven businesses contribute to the World Bank Group’s twin goals of fighting poverty and boosting shared prosperity?
These questions have been top of the mind since the World Bank Open Finances team convened a group of open data entrepreneurs from across Latin America to share their business models, success stories and challenges at the Open Data Business Models workshop in Uruguay in June 2013. We were in Uruguay to find out whether open data could lead to the creation of sustainable new businesses and jobs. To do so, we tested a couple of hypotheses: open data has economic value, beyond the benefits of increased transparency and accountability; and open data companies with sustainable business models already exist in emerging economies.
Encouraged by our findings in Uruguay we set out to further explore the economic development potential of open data, with a focus on:

  • Contribution of open data to countries’ GDP;
  • Innovative solutions to tackle social problems in key sectors like agriculture, health, education, transportation, climate change, financial services, especially those benefiting low income populations;
  • Economic benefits of governments’ buy-in into the commercial value of open data and resulting release of new datasets, which in turn would lead to increased transparency in public resource management (reductions in misallocations, a more level playing field in procurement) and better service delivery; and
  • Creation of data-related private sector jobs, especially suited for the tech savvy young generation.

We proposed a joint IFC/World Bank approach (From open data to development impact – the crucial role of private sector) that envisages providing financing to data-driven companies through a dedicated investment fund, as well as loans and grants to governments to create a favorable enabling environment. The concept was received enthusiastically for the most part by a wide group of peers at the Bank, the IFC, as well as NGOs, foundations, DFIs and private sector investors.
Thanks also in part to a McKinsey report last fall stating that open data could help unlock more than $3 trillion in value every year, the potential value of open data is now better understood. The acquisition of Climate Corporation (whose business model holds enormous potential for agriculture and food security, if governments open up the right data) for close to a billion dollars last November and the findings of the Open Data 500 project led by GovLab of the NYU further substantiated the hypothesis. These days no one asks whether open data has economic value; the focus has shifted to finding ways for companies, both startups and large corporations, and governments to unlock it. The first question though is – is it still too early to plan a significant intervention to spur open data driven economic growth in emerging markets?”

The merits of participatory budgeting


at Aljazeera America: “For many Americans, government just isn’t working. In 2013, government dysfunction surpassed the economy as the top identified U.S. problem. A recent survey found that nearly 6 out of 10 Americans rate the health of our democracy as weak — and unlikely to get better anytime soon. But in small corners throughout the United States, democratic innovations are creating new opportunities for citizens to be a part of governance. Collectively known as open government or civic innovation, these projects are engaging policymakers, citizens and civil society and proving the skeptics wrong.
One particularly promising innovation in participatory budgeting, or PB — a process to directly empower citizens to make spending decisions on a defined public budget. PB was first attempted in Porto Alegre, Brazil, in 1989. Its success led to the World Bank calling PB a “best practice” in democratic innovation. Since then, PB has expanded to over 1,500 cities worldwide, including several in the U.S. Starting in 2009 in Chicago’s 49th Ward with a budget of just $1 million, PB in the United States has expanded to a $27 million-a-year experiment. Municipal leaders from Vallejo, California, to New York City have turned over a portion of their discretionary funds to neighborhood residents. Boston recently launched the first youth-driven PB. Nearly half of New York’s City Council members are slated to participate this fall, after newly elected Mayor Bill de Blasio made it a cornerstone of his campaign. Chicago Mayor Rahm Emanuel created a new manager of participatory budgeting who will help coordinate Council districts that want to participate. The White House recently included federally supported participatory budgeting as part of its international Open Government Partnership commitments.

Wants and needs

In PB, citizens are empowered to identify community needs, work with elected officials to craft budget proposals and vote upon where to spend public funds. The decisions are binding. And that’s important: Making democracy work is not just about making better citizens or changing policies. It is also about creating structures that create the conditions that make the effective exercise of democratic citizenship possible, and PB is uniquely structured to do that.

Chicago has been a particularly insightful petri dish to study PB in the U.S., mainly because the city is an unlikely candidate for democratic innovations. For decades its Democratic machine retained a strong and continuous hold over city government. The Daley family held the mayoralty for a combined 12 terms. While discretionary funds (known as “menu money”) are allocated equally — but not equitably, given different needs — to all 50 wards, the process of spending this money is at the discretion of locally elected aldermen. From 1972 to 2009, 30 Chicago aldermen were indicted and convicted of federal crimes ranging from income tax evasion to extortion, embezzlement and conspiracy. Clearly, Chicago has not always been a model of good governance.
Against this backdrop, PB has continued to expand in Chicago. This year three districts participated. The Fifth Ward, home to the University of Chicago, decided not to continue the process again this year. Instead, this year the ward had four groups of residents each allocate $250,000. The alderwoman noted that this enabled the transparency and engagement aspect of PB with fewer process resources — they had only 100 people come out to vote.
Different versions of PB are aimed to lower the current barriers to civic engagement. I have seen PB bring out people who have never before engaged in politics. Many longtime civic participants often cite PB as the single most meaningful civic engagement of their lives — far above, say, jury duty. Suddenly, citizens are empowered with real decision-making authority and leave with new relationships with their peers, community and elected officials.
However, PB is not a stand-alone endeavor. It must be part of a larger effort to improve governance. This must include greater transparency in public decision making and empowering citizens to hold their elected officials more accountable. The process provides an enormous education that can be translated into civic activity beyond PB. Ideally after engaging in PB, a citizen will be better equipped to volunteer in the community, vote or push for policy reform. What other infrastructure, both online and off, is needed to support citizens who want to further engage in more collaborative governance?  …”

Health plan giants to make payment data accessible to public


Paul Demko in ModernHealthCare: “A new initiative by three of the country’s largest health plans has the potential to transform the accessibility of claims payment data, according to healthcare finance experts. UnitedHealthcare, Aetna and Humana announced a partnership on Wednesday with the Health Care Cost Institute to create a payment database that will be available to the public for free. …The database will be created by HCCI, a not-for-profit group established in 2011, from information provided by the insurers. HCCI expects it to be available in 2015 and that more health plans will join the initiative prior to its launch.
UnitedHealthcare is the largest insurer in the country in terms of the number of individuals covered through its products. All three participating plans are publicly traded, for-profit companies.
Stephen Parente, chair of HCCI’s board, said the organization was approached by the insurance companies about the initiative. “I’m not quite sure what the magic trigger was,” said Parente, who is a professor at the University of Minnesota and advised John McCain’s 2008 presidential campaign on healthcare issues. “We’ve kind of proven as a nonprofit and an independent group that we can be trustworthy in working with their data.”
Experts say cost transparency is being spurred by a number of developments in the healthcare sector. The trend towards high-deductible plans is giving consumers a greater incentive to understand how much healthcare costs and to utilize it more efficiently. In addition, the launch of the exchanges under the Patient Protection and Affordable Care Act has brought unprecedented attention to the difficulties faced by individuals in shopping for insurance coverage.
“There’s so many things that are kind of pushing the industry toward this more transparent state,” Hempstead said. “There’s just this drumbeat that people want to have this information.”
Insurers may also be realizing they aren’t likely to have a choice about sharing payment information. In recent years, more and more states have passed laws requiring the creation of claims databases. Currently, 11 states have all payer claims databases, and six other states are in the process of creating such a resource, according to the All-Payer Claims Database Council….”

Continued Progress and Plans for Open Government Data


Steve VanRoekel, and Todd Park at the White House:  “One year ago today, President Obama signed an executive order that made open and machine-readable data the new default for government information. This historic step is helping to make government-held data more accessible to the public and to entrepreneurs while appropriately safeguarding sensitive information and rigorously protecting privacy.
Freely available data from the U.S. government is an important national resource, serving as fuel for entrepreneurship, innovation, scientific discovery, and economic growth. Making information about government operations more readily available and useful is also core to the promise of a more efficient and transparent government. This initiative is a key component of the President’s Management Agenda and our efforts to ensure the government is acting as an engine to expand economic growth and opportunity for all Americans. The Administration is committed to driving further progress in this area, including by designating Open Data as one of our key Cross-Agency Priority Goals.
Over the past few years, the Administration has launched a number of Open Data Initiatives aimed at scaling up open data efforts across the Health, Energy, Climate, Education, Finance, Public Safety, and Global Development sectors. The White House has also launched Project Open Data, designed to share best practices, examples, and software code to assist federal agencies with opening data. These efforts have helped unlock troves of valuable data—that taxpayers have already paid for—and are making these resources more open and accessible to innovators and the public.
Other countries are also opening up their data. In June 2013, President Obama and other G7 leaders endorsed the Open Data Charter, in which the United States committed to publish a roadmap for our nation’s approach to releasing and improving government data for the public.
Building upon the Administration’s Open Data progress, and in fulfillment of the Open Data Charter, today we are excited to release the U.S. Open Data Action Plan. The plan includes a number of exciting enhancements and new data releases planned in 2014 and 2015, including:

  • Small Business Data: The Small Business Administration’s (SBA) database of small business suppliers will be enhanced so that software developers can create tools to help manufacturers more easily find qualified U.S. suppliers, ultimately reducing the transaction costs to source products and manufacture domestically.
  • Smithsonian American Art Museum Collection: The Smithsonian American Art Museum’s entire digitized collection will be opened to software developers to make educational apps and tools. Today, even museum curators do not have easily accessible information about their art collections. This information will soon be available to everyone.
  • FDA Adverse Drug Event Data: Each year, healthcare professionals and consumers submit millions of individual reports on drug safety to the Food and Drug Administration (FDA). These anonymous reports are a critical tool to support drug safety surveillance. Today, this data is only available through limited quarterly reports. But the Administration will soon be making these reports available in their entirety so that software developers can build tools to help pull potentially dangerous drugs off shelves faster than ever before.

We look forward to implementing the U.S. Open Data Action Plan, and to continuing to work with our partner countries in the G7 to take the open data movement global”.