France Bans Judge Analytics, 5 Years In Prison For Rule Breakers


Artificial Lawyer: “In a startling intervention that seeks to limit the emerging litigation analytics and prediction sector, the French Government has banned the publication of statistical information about judges’ decisions – with a five year prison sentence set as the maximum punishment for anyone who breaks the new law.

Owners of legal tech companies focused on litigation analytics are the most likely to suffer from this new measure.

The new law, encoded in Article 33 of the Justice Reform Act, is aimed at preventing anyone – but especially legal tech companies focused on litigation prediction and analytics – from publicly revealing the pattern of judges’ behaviour in relation to court decisions.

A key passage of the new law states:

‘The identity data of magistrates and members of the judiciary cannot be reused with the purpose or effect of evaluating, analysing, comparing or predicting their actual or alleged professional practices.’ *

As far as Artificial Lawyer understands, this is the very first example of such a ban anywhere in the world.

Insiders in France told Artificial Lawyer that the new law is a direct result of an earlier effort to make all case law easily accessible to the general public, which was seen at the time as improving access to justice and a big step forward for transparency in the justice sector.

However, judges in France had not reckoned on NLP and machine learning companies taking the public data and using it to model how certain judges behave in relation to particular types of legal matter or argument, or how they compare to other judges.

In short, they didn’t like how the pattern of their decisions – now relatively easy to model – were potentially open for all to see.

Unlike in the US and the UK, where judges appear to have accepted the fait accompli of legal AI companies analysing their decisions in extreme detail and then creating models as to how they may behave in the future, French judges have decided to stamp it out….(More)”.

The Landscape of Open Data Policies


Apograf: “Open Access (OA) publishing has a long history, going back to the early 1990s, and was born with the explicit intention of improving access to scholarly literature. The internet has played a pivotal role in garnering support for free and reusable research publications, as well as stronger and more democratic peer-review systems — ones are not bogged down by the restrictions of influential publishing platforms….

Looking back, looking forward

Launched in 1991, ArXiv.org was a pioneering platform in this regard, a telling example of how researchers could cooperate to publish academic papers for free and in full view for the public. Though it has limitations — papers are curated by moderators and are not peer-reviewed — arXiv is a demonstration of how technology can be used to overcome some of the incentive and distribution problems that scientific research had long been subjected to.

The scientific community has itself assumed the mantle to this end: the Budapest Open Access Initiative (BOAI) and the Berlin Declaration on Open Access Initiative, launched in 2002 and 2003 respectively, are considered landmark movements in the push for unrestricted access to scientific research. While mostly symbolic, the effort highlighted the growing desire to solve the problems plaguing the space through technology.

The BOAI manifesto begins with a statement that is an encapsulation of the movement’s purpose,

“An old tradition and a new technology have converged to make possible an unprecedented public good. The old tradition is the willingness of scientists and scholars to publish the fruits of their research in scholarly journals without payment, for the sake of inquiry and knowledge. The new technology is the internet. The public good they make possible is the world-wide electronic distribution of the peer-reviewed journal literature and completely free and unrestricted access to it by all scientists, scholars, teachers, students, and other curious minds.”

Plan S is a more recent attempt to make publicly funded research available to all. Launched by Science Europe in September 2018, Plan S — short for ‘Shock’ — has energized the research community with its resolution to make access to publicly funded knowledge a right to everyone and dissolve the profit-driven ecosystem of research publication. Members of the European Union have vowed to achieve this by 2020.

Plan S has been supported by governments outside Europe as well. China has thrown itself behind it, and the state of California has enacted a law that requires open access to research one year after publishing. It is, of course, not without its challenges: advocacy and ensuring that publishing is not restricted a few venues are two such obstacles. However, the organization behind forming the guidelines, cOAlition S, has agreed to make the guidelines more flexible.

The emergence of this trend is not without its difficulties, however, and numerous obstacles continue to hinder the dissemination of information in a manner that is truly transparent and public. Chief among these are the many gates that continue to keep research as somewhat of exclusive property, besides the fact that the infrastructure and development for such systems are short on funding and staff…..(More)”.

Opening Data for Global Health


Chapter by Matt Laessig, Bryon Jacob and Carla AbouZahr in The Palgrave Handbook of Global Health Data Methods for Policy and Practice: “…provide best practices for organizations to adopt to disseminate data openly for others to use. They describe development of the open data movement and its rapid adoption by governments, non-governmental organizations, and research groups. The authors provide examples from the health sector—an early adopter—but acknowledge concerns specific to health relating to informed consent, intellectual property, and ownership of personal data. Drawing on their considerable contributions to the open data movement, Laessig and Jacob share their Open Data Progression Model. They describe six stages to make data open: from data collection, documentation of the data, opening the data, engaging the community of users, making the data interoperable, to finally linking the data….(More)”

Come to Finland if you want to glimpse the future of health data!


Jukka Vahti at Sitra: “The Finnish tradition of establishing, maintaining and developing data registers goes back to the 1600s, when parish records were first kept.

When this old custom is combined with the opportunities afforded by digitisation, the positive approach Finns have towards research and technology, and the recently updated legislation enabling the data economy, Finland and the Finnish people can lead the way as Europe gradually, or even suddenly, switches to a fair data economy.

The foundations for a fair data economy already exist

The fair data economy is a natural continuation of the former projects promoting e-services that were undertaken in Finland.

For example, the Data Exchange Layer is already speeding up the transfer of data from one system to another in Finland and in Estonia, the country where the system originated, and a system unique to just these two countries.

In May 2019 Finland also saw the entry into force of the Act on the Secondary Use of Health and Social Data, according to which the information on social welfare and healthcare held in registers may be used for purposes of statistics, research, education, knowledge management, control and supervision conducted by authorities, and development and innovation activity.

The new law will make the work of researchers and service developers more effective, as the business of acquiring a permit will take place through a one-stop-shop principle and it will be possible to use data from more than one source more readily than before….(More)”.

Open Data and the Private Sector


Chapter by Joel Gurin, Carla Bonini and Stefaan Verhulst in State of Open Data: “The open data movement launched a decade ago with a focus on transparency, good governance, and citizen participation. As other chapters in this collection have documented in detail, those critical uses of open data have remained paramount and are continuing to grow in importance at a time of fake news and increased secrecy. But the value of open data extends beyond transparency and accountability – open data is also an important resource for business and economic growth.

The past several years have seen an increased focus on the value of open data to the private sector. In 2012, the Open Data Institute (ODI) was founded in the United Kingdom (UK) and backed with GBP 10 million by the UK government to maximise the value of open data in business and government. A year later, McKinsey released a report suggesting open data could help unlock USD 3 to 5 trillion in economic value annually. At around the same time, Monsanto acquired the Climate Corporation, a digital agriculture company that leverages open data to inform farmers for approximately USD 1.1 billion. In 2014, the GovLab launched the Open Data 500,2the first national study of businesses using open government data (now in six countries), and, in 2015, Open Data for Development (OD4D) launched the Open Data Impact Map, which today contains more than 1 100 examples of private sector companies using open data. The potential business applications of open data continue to be a priority for many governments around the world as they plan and develop their data programmes.

The use of open data has become part of the broader business practice of using data and data science to inform business decisions, ranging from launching new products and services to optimising processes and outsmarting the competition. In this chapter, we take stock of the state of open data and the private sector by analysing how the private sector both leverages and contributes to the open data ecosystem….(More)”.

Re-Use Of Public Sector Open Data: Characterising The Phenomena


Paper by Josefin Lassinantti at the International Journal of Public Information Systems: “Despite the growing number of open data, re-use of this data is not reaching the expected levels and now this phenomenon seems hampered in its evolvement. Therefore, this study sets out to characterize the re-use of open data from public sector in order to increase our elaborate understanding of this practice, and does so by performing a literature review inspired by the processes for defining concepts, and contextualized within the historical evolvement of European open data policies. Apart from the identification of three main research approaches towards open data re-use and an elaborated definition of re-use, the findings led to the creation of a framework enabling us to see open data re-use as an iterative value-creating process in two different contexts, the public task context and the non-public task context. This process builds on three categories of meta-activities for reuse practice: 1) gaining access to and understanding data, 2) handling and re-purposing the data, and 3) creating broader value of data, as well as indications of value for whom. Lastly, implications of this re-use process and framework was discussed, along with implications of an identified practice-policy mismatch that risk hampering the future evolvement of open data re-use….(More)”.

Microsoft’s Open Notre Dame initiative calls for sharing of open data in restoration effort


Hamza Jawad at Neowin: “On April 15, a disastrous fire ravaged the famous Notre-Dame cathedral in France. In the wake of the episode, tech companies, such as Apple, announced that they would be donating to help in rebuilding efforts. On the other hand, some companies, like Ubisoft, took a different approach to support the restorations that followed.

A few days ago, Microsoft and Iconem announced the “Open Notre Dame” initiative to contribute towards the restoration of the ‘Lady of Paris’. The open data project is said to help gather and analyze existing documents on the monument, while simultaneously producing and sharing its 3D models. Today, the company has once again detailed the workings of this initiative, along with a call for the sharing of open data to help quicken the restoration efforts….

GitHub will host temporal models of the building, which can then be easily shared to and accessed by various other initiatives in a concerted effort to maintain accuracy as much as possible. Many companies, including Ubisoft, have already provided data that will help form the foundation for these open source models. More details regarding the project can be obtained on the original blog post….(More)”.

Open data could have helped us learn from another mining dam disaster


Paulo A. de Souza Jr. at Nature: “The recent Brumadinho dam disaster in Brazil is an example of infrastructure failure with catastrophic consequences. Over 300 people were reported dead or missing, and nearly 400 more were rescued alive. The environmental impact is massive and difficult to quantify. The frequency of these disasters demonstrates that the current assets for monitoring integrity and generating alerting managers, authorities and the public to ongoing change in tailings are, in many cases, not working as they should. There is also the need for adequate prevention procedures. Monitoring can be perfect, but without timely and appropriate action, it will be useless. Good management therefore requires quality data. Undisputedly, management practices of industrial sites, including audit procedures, must improve, and data and metadata available from preceding accidents should be better used. There is a rich literature available about design, construction, operation, maintenance and decommissioning of tailing facilities. These include guidelines, standards, case studies, technical reports, consultancy and audit practices, and scientific papers. Regulation varies from country to country and in some cases, like Australia and Canada, it is controlled by individual state agencies. There are, however, few datasets available that are shared with the technical and scientific community more globally; particularly for prior incidents. Conspicuously lacking are comprehensive data related to monitoring of large infrastructures such as mining dams.

Today, Scientific Data published a Data Descriptor presenting a dataset obtained from 54 laboratory experiments on the breaching of fluvial dikes because of flow overtopping. (Re)use of such data can help improve our understanding of fundamental processes underpinning industrial infrastructure collapse (e.g., fluvial dike breaching, mining dam failure), and assess the accuracy of numerical models for the prediction of such incidents. This is absolutely essential for better management of floods, mitigation of dam collapses, and similar accidents. The authors propose a framework that could exemplify how data involving similar infrastructure can be stored, shared, published, and reused…(More)”.

The State of Open Data


Open Access Book edited by Tim Davies, Stephen B. Walker, Mor Rubinstein and Fernando Perini: “It’s been ten years since open data first broke onto the global stage. Over the past decade, thousands of programmes and projects around the world have worked to open data and use it to address a myriad of social and economic challenges. Meanwhile, issues related to data rights and privacy have moved to the centre of public and political discourse. As the open data movement enters a new phase in its evolution, shifting to target real-world problems and embed open data thinking into other existing or emerging communities of practice, big questions still remain. How will open data initiatives respond to new concerns about privacy, inclusion, and artificial intelligence? And what can we learn from the last decade in order to deliver impact where it is most needed? 

The State of Open Data brings together over 60 authors from around the world to address these questions and to take stock of the real progress made to date across sectors and around the world, uncovering the issues that will shape the future of open data in the years to come….(More)”.

Missing Numbers


Introduction by Anna Powell-Smith of a new “blog on the data the government should collect, but doesn’t”: “…Over time, I started to notice a pattern. Across lots of different policy areas, it was impossible for governments to make good decisions because of a basic lack of data. There was always critical data that the state either didn’t collect at all, or collected so badly that it made change impossible.

Eventually, I decided that the power to not collect data is one of the most important and little-understood sources of power that governments have. This is why I’m writing Missing Numbers: to encourage others to ask “is this lack of data a deliberate ploy to get away with something”?

By refusing to amass knowledge in the first place, decision-makers exert power over over the rest of us. It’s time that this power was revealed, so we can have better conversations about what we need to know to run this country successfully.

A typical example

The government records and publishes data on how often each NHS hospital receives formal complaints. This is very helpful, because it means patients and the people who care for them can spot hospitals whose performance is worrying.

But the government simply doesn’t record data, even internally, on how often formal complaints are made about each Jobcentre. (That FOI response is from 2015, but I’ve confirmed it’s still true in 2019.) So it is impossible for it to know if some Jobcentres are being seriously mismanaged….(More)”.