A report by SDSN TReNDS and DataReady Limited on behalf of Contracts4DataCollaboration: “Building upon issues discussed in the C4DC report, “Laying the Foundation for Effective Partnerships: An Examination of Data Sharing Agreements,” this brief examines the potential of sunset clauses or sunset provisions to be a legally binding, enforceable, and accountable way of ensuring COVID-19 related data sharing agreements are wound down responsibly at the end of the pandemic. The brief is divided into four substantive parts: Part I introduces sunset clauses as legislative tools, highlighting a number of examples of how they have been used in both COVID-19 related and other contexts; Part II discusses sunset provisions in the context of data sharing agreements and attempts to explain the complex interrelationship between data ownership, intellectual property, and sunset provisions; Part III identifies some key issues policymakers should consider when assessing the utility and viability of sunset provisions within their data sharing agreements and arrangements; and Part IV highlights the value of a memorandum of understanding (MoU) as a viable vehicle for sunset provisions in contexts where data sharing agreements are either non-existent or not regularly used….(More)“.(Contracts 4 Data Collaboration Framework).
NIH Releases New Policy for Data Management and Sharing
NIH Blogpost by Carrie Wolinetz: “Today, nearly twenty years after the publication of the Final NIH Statement on Sharing Research Data in 2003, we have released a Final NIH Policy for Data Management and Sharing. This represents the agency’s continued commitment to share and make broadly available the results of publicly funded biomedical research. We hope it will be a critical step in moving towards a culture change, in which data management and sharing is seen as integral to the conduct of research. Responsible data management and sharing is good for science; it maximizes availability of data to the best and brightest minds, underlies reproducibility, honors the participation of human participants by ensuring their data is both protected and fully utilized, and provides an element of transparency to ensure public trust and accountability.
This policy has been years in the making and has benefited enormously from feedback and input from stakeholders throughout the process. We are grateful to all those who took the time to comment on Request for Information, the Draft policy, or to participate in workshops or Tribal consultations. That thoughtful feedback has helped shape the Final policy, which we believe strikes a balance between reasonable expectations for data sharing and flexibility to allow for a diversity of data types and circumstances. How we incorporated public comments and decision points that led to the Final policy are detailed in the Preamble to the DMS policy.
The Final policy applies to all research funded or conducted by NIH that results in the generation of scientific data. The Final Policy has two main requirements (1) the submission of a Data Management and Sharing Plan (Plan); and (2) compliance with the approved Plan. We are asking for Plans at the time of submission of the application, because we believe planning and budgeting for data management and sharing needs to occur hand in hand with planning the research itself. NIH recognizes that science evolves throughout the research process, which is why we have built in the ability to update DMS Plans, but at the end of the day, we are expecting investigators and institutions to be accountable to the Plans they have laid out for themselves….
Anticipating that variation in readiness, and in recognition of the cultural change we are trying to seed, there is a two-year implementation period. This time will be spent developing the information, support, and tools that the biomedical enterprise will need to comply with this new policy. NIH has already provided additional supplementary information – on (1) elements of a data management and sharing plan; (2) allowable costs; and (3) selecting a data repository – in concert with the policy release….(More)”
Your phone already tracks your location. Now that data could fight voter suppression
Article by Seth Rosenblatt: “Smartphone location data is a dream for marketers who want to know where you go and how long you spend there—and a privacy nightmare. But this kind of geolocation data could also be used to protect people’s voting rights on Election Day.
The newly founded nonprofit Center for New Data is now tracking voters at the polls using smartphone location data to help researchers understand how easy—or difficult—it is for people to vote in different places. Called the Observing Democracy project, the nonpartisan effort is making data on how far people have to travel to vote and how long they have to wait in line available in a privacy-friendly way so it can be used to craft election policies that ensure voting is accessible for everyone.
Election data has already fueled changes in various municipalities and states. A 66-page lawsuit filed by Fair Fight Action against the state of Georgia in the wake of Stacey Abrams’s narrow loss to Brian Kemp in the 2018 gubernatorial race relies heavily on data to back its assertions of unconstitutionally delayed and deferred voter registration, unfair challenges to absentee and provisional ballots, and unjustified purges of voter rolls—all hallmarks of voter suppression.
The promise of Observing Democracy is to make this type of impactful data available much more rapidly than ever before. Barely a month old, Observing Democracy isn’t wasting any time: Its all-volunteer staffers will be receiving data potentially as soon as Nov. 4 on voter wait times at polling locations, travel times to polling stations, and how frequently ballot drop-off boxes are visited, courtesy of location-data mining companies X-Mode Social and Veraset, which was spun off from SafeGraph….(More)”.
To mitigate the costs of future pandemics, establish a common data space
Article by Stephanie Chin and Caitlin Chin: “To improve data sharing during global public health crises, it is time to explore the establishment of a common data space for highly infectious diseases. Common data spaces integrate multiple data sources, enabling a more comprehensive analysis of data based on greater volume, range, and access. At its essence, a common data space is like a public library system, which has collections of different types of resources from books to video games; processes to integrate new resources and to borrow resources from other libraries; a catalog system to organize, sort, and search through resources; a library card system to manage users and authorization; and even curated collections or displays that highlight themes among resources.
Even before the COVID-19 pandemic, there was significant momentum to make critical data more widely accessible. In the United States, Title II of the Foundations for Evidence-Based Policymaking Act of 2018, or the OPEN Government Data Act, requires federal agencies to publish their information online as open data, using standardized, machine-readable data formats. This information is now available on the federal data.gov catalog and includes 50 state- or regional-level data hubs and 47 city- or county-level data hubs. In Europe, the European Commission released a data strategy in February 2020 that calls for common data spaces in nine sectors, including healthcare, shared by EU businesses and governments.
Going further, a common data space could help identify outbreaks and accelerate the development of new treatments by compiling line list incidence data, epidemiological information and models, genome and protein sequencing, testing protocols, results of clinical trials, passive environmental monitoring data, and more.
Moreover, it could foster a common understanding and consensus around the facts—a prerequisite to reach international buy-in on policies to address situations unique to COVID-19 or future pandemics, such as the distribution of medical equipment and PPE, disruption to the tourism industry and global supply chains, social distancing or quarantine, and mass closures of businesses….(More). See also Call for Action for a Data Infrastructure to tackle Pandemics and other Dynamic Threats.
Third Wave of Open Data
Paper (and site) by Stefaan G. Verhulst, Andrew Young, Andrew J. Zahuranec, Susan Ariel Aaronson, Ania Calderon, and Matt Gee on “How To Accelerate the Re-Use of Data for Public Interest Purposes While Ensuring Data Rights and Community Flourishing”: “The paper begins with a description of earlier waves of open data. Emerging from freedom of information laws adopted over the last half century, the First Wave of Open Data brought about newfound transparency, albeit one only available on request to an audience largely composed of journalists, lawyers, and activists.
The Second Wave of Open Data, seeking to go beyond access to public records and inspired by the open source movement, called upon national governments to make their data open by default. Yet, this approach too had its limitations, leaving many data silos at the subnational level and in the private sector untouched..
The Third Wave of Open Data seeks to build on earlier successes and take into account lessons learned to help open data realize its transformative potential. Incorporating insights from various data experts, the paper describes the emergence of a Third Wave driven by the following goals:
- Publishing with Purpose by matching the supply of data with the demand for it, providing assets that match public interests;
- Fostering Partnerships and Data Collaboration by forging relationships with community-based organizations, NGOs, small businesses, local governments, and others who understand how data can be translated into meaningful real-world action;
- Advancing Open Data at the Subnational Level by providing resources to cities, municipalities, states, and provinces to address the lack of subnational information in many regions.
- Prioritizing Data Responsibility and Data Rights by understanding the risks of using (and not using) data to promote and preserve the public’s general welfare.
Riding the Wave
Achieving these goals will not be an easy task and will require investments and interventions across the data ecosystem. The paper highlights eight actions that decision and policy makers can take to foster more equitable, impactful benefits… (More) (PDF) “
Data to Go: The Value of Data Portability as a Means to Data Liquidity
Juliet McMurren and Stefaan G. Verhulst at Data & Policy: “If data is the “new oil,” why isn’t it flowing? For almost two decades, data management in fields such as government, healthcare, finance, and research has aspired to achieve a state of data liquidity, in which data can be reused where and when it is needed. For the most part, however, this aspiration remains unrealized. The majority of the world’s data continues to stagnate in silos, controlled by data holders and inaccessible to both its subjects and others who could use it to create or improve services, for research, or to solve pressing public problems.
Efforts to increase liquidity have focused on forms of voluntary institutional data sharing such as data pools or other forms of data collaboratives. Although useful, these arrangements can only advance liquidity so far. Because they vest responsibility and control over liquidity in the hands of data holders, their success depends on data holders’ willingness and ability to provide access to their data for the greater good. While that willingness exists in some fields, particularly medical research, a willingness to share data is much less likely where data holders are commercial competitors and data is the source of their competitive advantage. And even where willingness exists, the ability of data holders to share data safely, securely, and interoperably may not. Without a common set of secure, standardized, and interoperable tools and practices, the best that such bottom-up collaboration can achieve is a disconnected patchwork of initiatives, rather than the data liquidity proponents are seeking.
Data portability is one potential solution to this problem. As enacted in the EU General Data Protection Regulation (2018) and the California Consumer Privacy Act (2018), the right to data portability asserts that individuals have a right to obtain, copy, and reuse their personal data and transfer it between platforms or services. In so doing, it shifts control over data liquidity to data subjects, obliging data holders to release data whether or not it is in their commercial interests to do so. Proponents of data portability argue that, once data is unlocked and free to move between platforms, it can be combined and reused in novel ways and in contexts well beyond those in which it was originally collected, all while enabling greater individual control.
To date, however, arguments for the benefits of the right to data portability have typically failed to connect this rights-based approach with the larger goal of data liquidity and how portability might advance it. This failure to connect these principles and to demonstrate their collective benefits to data subjects, data holders, and society has real-world consequences. Without a clear view of what can be achieved, policymakers are unlikely to develop interventions and incentives to advance liquidity and portability, individuals will not exercise their rights to data portability, and industry will not experiment with use cases and develop the tools and standards needed to make portability and liquidity a reality.
Toward these ends, we have been exploring the current literature on data portability and liquidity, searching for lessons and insights into the benefits that can be unlocked when data liquidity is enabled through the right to data portability. Below we identify some of the greatest potential benefits for society, individuals, and data-holding organizations. These benefits are sometimes in conflict with one another, making the field a contentious one that demands further research on the trade-offs and empirical evidence of impact. In the final section, we also discuss some barriers and challenges to achieving greater data liquidity….(More)”.
Open data and data sharing: An economic analysis
Paper by Alevtina Krotova, Armin Mertens, Marc Scheufen: “Data is an important business resource. It forms the basis for various digital technologies such as artificial intelligence or smart services. However, access to data is unequally distributed in the market. Hence, some business ideas fail due to a lack of data sources. Although many governments have recognised the importance of open data and already make administrative data available to the public on a large scale, many companies are still reluctant to share their data among other firms and competitors. As a result, the economic potential of data is far from being fully exploited. Against this background, we analyse current developments in the area of open data. We compare the characteristics of open governmental and open company data in order to define the necessary framework conditions for data sharing. Subsequently, we examine the status quo of data sharing among firms. We use a qualitative analysis of survey data of European companies to derive the sufficient conditions to strengthen data sharing. Our analysis shows that governmental data is a public good, while company data can be seen as a club or private good. Latter frequently build the core for companies’ business models and hence are less suitable for data sharing. Finally, we find that promoting legal certainty and the economic impact present important policy steps for fostering data sharing….(More)”
Airbnb’s Data ‘Portal’ Promises a Better Relationship With Cities
Article by Patrick Sisson: “When startups go public, a big part of the process is opening up their books and being more transparent about their business model. With global short-term rental giant Airbnb moving towards its own IPO, the company has introduced a new product that seeks to address recent safety concerns and answer the data-sharing requests that critics have long claimed make the company a less-than-perfect partner for local leaders.
The Airbnb City Portal, which launched on Wednesday as a pilot program with 15 global cities and tourism agencies, aims to provide municipal staff with more efficient access to data about listings, including whether or not they’re complying with local laws. Each city, including Buffalo, San Francisco and Seattle, will have access to a new data dashboard as well as a dedicated staffer at Airbnb. Like so many of its sharing economy and Silicon Valley peers, Airbnb has had a contentious, and evolving, relationship with municipalities and local government ever since launching (an especially fraught situation in Europe, as an EU court just ruled in favor of city regulations of the site).
At a time when so many tech platforms are wrestling, often unsuccessfully, with the need to moderate the behavior of bad actors who use the site, Airbnb’s City Portal is an attempt to “productize” how the home-sharing site works with local government, says Chris Lehane, Airbnb’s senior vice president for global policy and communications. It’s a more useful framework to access information and report violations, he says. And it delivers on the platform’s long-term goals around sharing data, paying taxes and working with cities on regulation. He frames the move as part of a balancing act around the security and safety responsibilities of local governments and a private global company.
The dashboard will also be useful for local tourism officials: It will provide visitor information, including city of origin and demographic information, that helps bureaus better target their advertising and marketing campaigns….(More)”
Announcing the New Data4COVID19 Repository
Blog by Andrew Zahuranec: “It’s been a long year. Back in March, The GovLab released a Call for Action to build the data infrastructure and ecosystem we need to tackle pandemics and other dynamic societal and environmental threats. As part of that work, we launched a Data4COVID19 repository to monitor progress and curate projects that reused data to address the pandemic. At the time, it was hard to say how long it would remain relevant. We did not know how long the pandemic would last nor how many organizations would publish dashboards, visualizations, mobile apps, user tools, and other resources directed at the crisis’s worst consequences.
Seven months later, the COVID-19 pandemic is still with us. Over one million people around the world are dead and many countries face ever-worsening social and economic costs. Though the frequency with which data reuse projects are announced has slowed since the crisis’s early days, they have not stopped. For months, The GovLab has posted dozens of additions to an increasingly unwieldy GoogleDoc.
Today, we are making a change. Given the pandemic’s continued urgency and relevance into 2021 and beyond, The GovLab is pleased to release the new Data4COVID19 Living Repository. The upgraded platform allows people to more easily find and understand projects related to the COVID-19 pandemic and data reuse.
On the platform, visitors will notice a few improvements that distinguish the repository from its earlier iteration. In addition to a main page with short descriptions of each example, we’ve added improved search and filtering functionality. Visitors can sort through any of the projects by:
- Scope: the size of the target community;
- Region: the geographic area in which the project takes place;
- Topic: the aspect of the crisis the project seeks to address; and
- Pandemic Phase: the stage of pandemic response the project aims to address….(More)”.
A New Normal for Data Collection: Using the Power of Community to Tackle Gender Violence Amid COVID-19
Claudia Wells at SDG Knowledge Hub: “A shocking increase in violence against women and girls has been reported in many countries during the COVID-19 pandemic, amounting to what UN Women calls a “shadow pandemic.”
The jarring facts are:
- Globally 243 million women and girls have been subjected to sexual and/or physical violence by an intimate partner in the past 12 months.
- The UNFPA estimates that the pandemic will cause a one-third reduction in progress towards ending gender-based violence by 2030;
- UNFPA predicts an additional 15 million cases of gender-based violence for every three months of lockdown.
- Official data captures only a fraction of the true prevalence and nature of gender-based violence.
The response to these new challenges were discussed at a meeting in July with a community-led response delivered through local actors highlighted as key. This means that timely, disaggregated, community-level data on the nature and prevalence of gender-based violence has never been more important. Data collected within communities can play a vital role to fill the gaps and ensure that data-informed policies reflect the lived experiences of the most marginalized women and girls.
Community Scorecards: Example from Nepal
Collecting and using community-level data can be challenging, particularly under the restrictions of the pandemic. Working in partnerships is therefore vital if we are to respond quickly and flexibly to new and old challenges.
A great example of this is the Leave No One Behind Partnership, which responds to these challenges while delivering on crucial data and evidence at the community level. This important partnership brings together international civil society organizations with national NGOs, civic platforms and community-based organizations to monitor progress towards the SDGs….
While COVID-19 has highlighted the need for local, community-driven data, public health restrictions have also made it more challenging to collect such data. For example the usual focus group approach to creating a community scorecard is no longer possible.
The coalition in Nepal therefore faces an increased demand for community-driven data while needing to develop a “new normal for data collection.”. Partners must: make data collection more targeted; consider how data on gender-based violence are included in secondary sources; and map online resources and other forms of data collection.
Addressing these new challenges may include using more blended collection approaches such as mobile phones or web-based platforms. However, while these may help to facilitate data collection, they come with increased privacy and safeguarding risks that have to be carefully considered to ensure that participants, particularly women and girls, are not at increased risk of violence or have their privacy and confidentiality exposed….(More)”.