Mashable: “We reached out to a few organizations using information, both hand- and algorithm-collected, to create helpful tools for their communities. This is only a small sample of what’s out there — plenty more pop up each day, and as more information becomes public, the trend will only grow….
1. Transit Time NYC
Transit Time NYC, an interactive map developed by WNYC, lets New Yorkers click a spot in any of the city’s five boroughs for an estimate of subway or train travel times. To create it, WNYC lead developer Steve Melendez broke the city into 2,930 hexagons, then pulled data from open source itinerary platform OpenTripPlanner — the Wikipedia of mapping software — and coupled it with the MTA’s publicly downloadable subway schedule….
2. Twitter’s ‘Topography of Tweets
In a blog post, Twitter unveiled a new data visualization map that displays billions of geotagged tweets in a 3D landscape format. The purpose is to display, topographically, which parts of certain cities most people are tweeting from…
3. Homicide Watch D.C.
Homicide Watch D.C. is a community-driven data site that aims to cover every murder in the District of Columbia. It’s sorted by “suspect” and “victim” profiles, where it breaks down each person’s name, age, gender and race, as well as original articles reported by Homicide Watch staff…
4. Falling Fruit
Can you find a hidden apple tree along your daily bike commute? Falling Fruit can.
The website highlights overlooked or hidden edibles in urban areas across the world. By collecting public information from the U.S. Department of Agriculture, municipal tree inventories, foraging maps and street tree databases, the site has created a network of 615 types of edibles in more than 570,000 locations. The purpose is to remind urban dwellers that agriculture does exist within city boundaries — it’s just more difficult to find….
5. AIDSvu
AIDSVu is an interactive map that illustrates the prevalence of HIV in the United States. The data is pulled from the U.S. Center for Disease Control’s national HIV surveillance reports, which are collected at both state and county levels each year…”
9 models to scale open data – past, present and future
Ones that are working now
1) Form a community to enter in new data. Open Street Map and MusicBrainz are two big examples. It works as the community is the originator of the data. That said, neither has dominated its industry as much as I thought they would have by now.
2) Sell tools to an upstream generator of open data. This is what CKAN does for central Governments (and the new ScraperWiki CKAN tool helps with). It’s what mySociety does, when selling FixMyStreet installs to local councils, thereby publishing their potholes as RSS feeds.
3) Use open data (quietly). Every organisation does this and never talks about it. It’s key to quite old data resellers like Bloomberg. It is what most of ScraperWiki’s professional services customers ask us to do. The value to society is enormous and invisible. The big flaw is that it doesn’t help scale supply of open data.
4) Sell tools to downstream users. This isn’t necessarily open data specific – existing software like spreadsheets and Business Intelligence can be used with open or closed data. Lots of open data is on the web, so tools like the new ScraperWiki which work well with web data are particularly suited to it.
Ones that haven’t worked
5) Collaborative curation ScraperWiki started as an audacious attempt to create an open data curation community, based on editing scraping code in a wiki. In its original form (now called ScraperWiki Classic) this didn’t scale. …With a few exceptions, notably OpenCorporates, there aren’t yet open data curation projects.
6) General purpose data marketplaces, particularly ones that are mainly reusing open data, haven’t taken off. They might do one day, however I think they need well-adopted higher level standards for data formatting and syncing first (perhaps something like dat, perhaps something based on CSV files).
Ones I expect more of in the future
These are quite exciting models which I expect to see a lot more of.
7) Give labour/money to upstream to help them create better data. This is quite new. The only, and most excellent, example of it is the UK’s National Archive curating the Statute Law Database. They do the work with the help of staff seconded from commercial legal publishers and other parts of Government.
It’s clever because it generates money for upstream, which people trust the most, and which has the most ability to improve data quality.
8) Viral open data licensing. MySQL made lots of money this way, offering proprietary dual licenses of GPLd software to embedded systems makers. In data this could use OKFN’s Open Database License, and organisations would pay when they wanted to mix the open data with their own closed data. I don’t know anyone actively using it, although Chris Taggart from OpenCorporates mentioned this model to me years ago.
9) Corporations release data for strategic advantage. Companies are starting to release their own data for strategic gain. This is very new. Expect more of it.”
The Little Data Book on Information and Communication Technology 2013
The World Bank: “This new addition to the Little Data Book series presents at-a-glance tables for more than 200 economies showing the most recent national data on key indicators of information and communications technology (ICT), including access, quality, affordability, efficiency, sustainability and applications.”
Power of open data reveals global corporate networks
Open Data Institute: “The ODI today welcomed the move by OpenCorporates to release open data visualisations which show the global corporate networks of millions of businesses and the power of open data.
See the Maps
OpenCorporates, a company based at the ODI, has produced visuals using several sources, which it has published as open data for the first time:
- Filings made by large domestic and foreign companies to the U.S. Securities and Exchange Commission
- Banking data held by the National Information Center of the Federal Reserve System in the U.S.
- Information about individual shareholders published by the official New Zealand corporate registry
Launched today, the visualisations are available through the main OpenCorporates website.”
The 20 Basics of Open Government
Watching what is going on around the world in national, state, and local governments, we think opengov is maturing and that the time has come for a basics resource for newbies. Our goal was to include the full expanse of open government and show how it all ties together so that when you, the astute reader, meet up with one of the various opengov cliques that uses the terminology in a narrowly defined way, you can see how they fit into the bigger picture. You should also be able to determine how opengov can best be applied to benefit whatever you’re up to, while keeping in mind the need to provide both access for citizens to engage with government and access to information.
Have a read through it, and let us know what you think! When you find a typo – or something you disagree with – or something we missed, let us know that as well. The easiest way to do it is right there in the comments (we’re not afraid to be called out in public!), but we’re open to email and twitter as well. We’re looking forward to hearing what you think!.”
The Real-Time City? Big Data and Smart Urbanism
New paper by Rob Kitchin from the National University of Ireland, Maynooth (NUI Maynooth) – NIRSA: “‘Smart cities’ is a term that has gained traction in academia, business and government to describe cities that, on the one hand, are increasingly composed of and monitored by pervasive and ubiquitous computing and, on the other, whose economy and governance is being driven by innovation, creativity and entrepreneurship, enacted by smart people. This paper focuses on the former and how cities are being instrumented with digital devices and infrastructure that produce ‘big data’ which enable real-time analysis of city life, new modes of technocratic urban governance, and a re-imagining of cities. The paper details a number of projects that seek to produce a real-time analysis of the city and provides a critical reflection on the implications of big data and smart urbanism”
Open Government is an Open Conversation
Lisa Ellman and Hollie Russon Gilman at the White House Blog: “President Obama launched the first U.S. Open Government National Action Plan in September 2011, as part of the Nation’s commitment to the principles of the global Open Government Partnership. The Plan laid out twenty-six concrete steps the United States would take to promote public participation in government, increase transparency in government, and manage public resources more effectively.
A year and a half later, we have fulfilled twenty-four of the Plan’s prescribed commitments—including launching the online We the People petition platform, which has been used by more than 9.6 million people, and unleashing thousands of government data resources as part of the Administration’s Open Data Initiatives.
We are proud of this progress, but recognize that there is always more work to be done to build a more efficient, effective, and transparent government. In that spirit, as part of our ongoing commitment to the international Open Government Partnership, the Obama Administration has committed to develop a second National Action Plan on Open Government.
To accomplish this task effectively, we’ll need all-hands-on-deck. That’s why we plan to solicit and incorporate your input as we develop the National Action Plan “2.0.”…
Over the next few months, we will continue to gather your thoughts. We will leverage online platforms such as Quora, Google+, and Twitter to communicate with the public and collect feedback. We will meet with members of open government civil society organizations and other experts, to ensure all voices are brought to the table. We will solicit input from Federal agencies on lessons learned from their unique experiences, and gather information about successful initiatives that could potentially be scaled across government. And finally, we will canvass the international community for their diverse insights and innovative ideas.”
Frontiers in Massive Data Analysis
New Report from the National Research Council: “From Facebook to Google searches to bookmarking a webpage in our browsers, today’s society has become one with an enormous amount of data. Some internet-based companies such as Yahoo! are even storing exabytes (10 to the 18 bytes) of data. Like these companies and the rest of the world, scientific communities are also generating large amounts of data-—mostly terabytes and in some cases near petabytes—from experiments, observations, and numerical simulation. However, the scientific community, along with defense enterprise, has been a leader in generating and using large data sets for many years. The issue that arises with this new type of large data is how to handle it—this includes sharing the data, enabling data security, working with different data formats and structures, dealing with the highly distributed data sources, and more.
Frontiers in Massive Data Analysis presents the Committee on the Analysis of Massive Data’s work to make sense of the current state of data analysis for mining of massive sets of data, to identify gaps in the current practice and to develop methods to fill these gaps. The committee thus examines the frontiers of research that is enabling the analysis of massive data which includes data representation and methods for including humans in the data-analysis loop. The report includes the committee’s recommendations, details concerning types of data that build into massive data, and information on the seven computational giants of massive data analysis.”
E-petition systems and political participation: About institutional challenges and democratic opportunities
New paper by Knud Böhle and Ulrich Riehm in First Monday: “The implementation of e–petition systems holds the promise to increase the participative and deliberative potential of petitions. The most ambitious e–petition systems allow for electronic submission, make publicly available the petition text, related documents and the final decision, allow supporting a petition by electronically co–signing it, and provide electronic discussion forums. Based on a comprehensive survey (2010/2011) of parliamentary petition bodies at the national level covering the 27 member states of the European Union (EU) plus Norway and Switzerland, the state of public e–petitioning in the EU is presented, and the relevance of e–petition systems as a means of political participation is discussed….
The most interesting finding is that some petition systems — by leveraging the potential of the Internet — further the involvement of the public considerably. This happens in two ways: first by nudging e–petition systems in the direction of lightweight instruments of direct democracy and second by making the institution itself more open, transparent, accountable, effective, and responsive through the involvement of the public. Both development paths might also lead to expectations that eventually cannot be complied with by the petition body without more substantial transformations of the institution. This or that might happen. Empirically, we ain’t seen almost nothing yet.”
Immersion: Using E-Mail Data to Connect the Dots of Your Life
Brian Chen in The New York Times: “The Obama administration for over two years allowed the National Security Agency to collect enormous amounts of metadata on e-mail usage by Americans, according to one of the latest leaks of government documents by the now-famous whistle-blower Edward J. Snowden.
But what is e-mail metadata anyway? It’s information about the people you’re sending e-mails to and receiving e-mails from, and the times that the messages were sent — as opposed to the contents of the messages. It’s the digital equivalent of a postal service worker looking at your mail envelope instead of opening it up and reading what’s inside.
That sounds harmless, but it turns out your e-mail metadata can be used to connect the dots of your life story. I learned this from participating in Immersion, a project by M.I.T.’s Media Laboratory, earlier reported by my colleague Juliet Lapidos. Immersion is a tool that mines your e-mail metadata and automatically stitches it all together into an interactive graphic. The result is a creepy spider web showing all the people you’ve corresponded with, how they know each other, and who your closest friends and professional partners are.
After entering my Google mail credentials, Immersion took five minutes to stitch together metadata from e-mails going back eight years. A quick glimpse at my results gives an accurate description of my life.”
Sign up here: https://immersion.media.mit.edu/