Complex Algorithm Auto-Writes Books, Could Transform Science


Mashable: “Could a sophisticated algorithm be the future of science? One innovative economist thinks so.
Phil Parker, who holds a doctorate in business economics from the Wharton School, has built  an algorithm that auto-writes books. Now he’s taking that model and applying it to loftier goals than simply penning periodicals: namely, medicine and forensics. Working with professors and researchers at NYU, Parker is trying to decode complex genetic structures and find cures for diseases. And he’s doing it with the help of man’s real best friend: technology.
Parker’s recipe is a complex computer program that mimics formulaic writing….
Parker’s been at this for years. His formula, originally used for printing, is able to churn out entire books in minutes. It’s similar to the work being done by Narrative Science and StatSheet, except those companies are known for short form auto-writing for newspapers. Parker’s work is much longer, focusing on obscure non-fiction and even poetry.
It’s not creative writing, though, and Parker isn’t interested in introspection, exploring emotion or storytelling. He’s interested in exploiting reproducible patterns — that’s how his algorithm can find, collect and “write” so quickly. And how he can apply that model to other disciplines, like science.
Parker’s method seems to be a success; indeed, his ICON Group International, Inc., has auto-written so many books that Parker has lost count. But this isn’t the holy grail of literature, he insists. Instead, he says, his work is a play on mechanizing processes to create a simple formula. And he thinks that “finding new knowledge structures within data” stretches far beyond print.”

Empowering Consumers through the Smart Disclosure of Data


OSTP: “Today, the Administration’s interagency National Science and Technology Council released Smart Disclosure and Consumer Decision Making: Report of the Task Force on Smart Disclosurethe first comprehensive description of the Federal Government’s efforts to promote the smart disclosure of information that can help consumers make wise decisions in the marketplace.
Whether they are searching for colleges, health insurance, credit cards, airline flights, or energy providers, consumers can find it difficult to identify the specific product or service that best suits their particular needs. In some cases, the effort required to sift through all of the available information is so large that consumers make decisions using inadequate information. As a result, they may overpay, miss out on a product that would better meet their needs, or be surprised by fees.
The report released today outlines ways in which Federal agencies and other governmental and non-governmental organizations can use—and in many cases are already using—smart disclosure approaches that increase market transparency and empower consumers facing complex choices in domains such as health, education, energy and personal finance.”

Mary Meeker’s Internet Trends Report


AllThingsD: For the second year in a row, Mary Meeker is unveiling her now famed Internet Trends report at the D11 Conference.

Meeker, the Kleiner Perkins Caufield & Byers partner, highlights growth of Internet usage and other activities on mobile devices and updates that now infamous gap between mobile internet usage and mobile monetization.
But there are many new additions. Among them are the rise of wearable tech as perhaps the next big tech cycle of the coming decade and a look at how Americans’ online sharing habits compare to the rest of the world.
Here’s Meeker’s full presentation:

KPCB Internet Trends 2013 from Kleiner Perkins Caufield & Byers

The Performativity of Data: Re-conceptualizing the Web of Data


New Paper by several authors of the Rensselaer Polytechnic Institute, including Jim Hendler, Marie Joan Kristine Gloria, Dominic DiFranzo and Marco Fernando Navarro: “As the discipline of Web Science matures, its interdisciplinary claim has many researchers unsure about its core theory and methodology. Instead, we often see results that are more multi-disciplinary than interdisciplinary. The following contribution attempts to recast our understanding of the current methodologies and tools leveraged by the Web Science community.  Specifically, we review the Semantic Web and Linked Data technologies not just from a technical perspective; but, through a critical reading of key social theories such as Goffman’s theory of performance. Our goal is to re- conceptualize the performativity of semantic web tools their boundaries, and any potential avenues for future research”

The Art of Data Visualization (video)


PBS Off Book: “Humans have a powerful capacity to process visual information, skills that date far back in our evolutionary lineage. And since the advent of science, we have employed intricate visual strategies to communicate data, often utilizing design principles that draw on these basic cognitive skills. In a modern world where we have far more data than we can process, the practice of data visualization has gained even more importance. From scientific visualization to pop infographics, designers are increasingly tasked with incorporating data into the media experience. Data has emerged as such a critical part of modern life that it has entered into the realm of art, where data-driven visual experiences challenge viewers to find personal meaning from a sea of information, a task that is increasingly present in every aspect of our information-infused lives.”

How We Imagined the Internet Before the Internet Even Existed


Matt Novak in Paleofuture : “In a few years, men will be able to communicate more effectively through a machine than face to face. Sounds obvious today. But in 1968, a full year before ARPANET made its first connection? It was downright clairvoyant…
The paper was written by J.C.R. Licklider and Robert Taylor, illustrated by Rowland B. Wilson, and appeared in the April 1968 issue of Science and Technology. The article includes some of the most amazingly accurate predictions for what networked computing would eventually allow….

The article rather boldly predicts that the computerized networks of the future will be even more important for communication than the “printing press and the picture tube”—another idea not taken for granted in 1968:

Creative, interactive communication requires a plastic or moldable medium that can be modeled, a dynamic medium in which premises will flow into consequences, and above all a common medium that can be contributed to and experimented with by all.
Such a medium is at hand—the programmed digital computer. Its presence can change the nature and value of communication even more profoundly than did the printing press and the picture tube, for, as we shall show, a well-programmed computer can provide direct access both to informational resources and to the processes for making use of the resources.

The paper predicts that the person-to-person interaction that a networked computer system allows for will not only build relationships between individuals, but will build communities.

What will on-line interactive communities be like? In most fields they will consist of geographically separated members, sometimes grouped in small clusters and sometimes working individually. They will be communities not of common location, but of common interest. In each field, the overall community of interest will be large enough to support a comprehensive system of field-oriented programs and data.

…In the end, Licklider and Taylor predict that all of this interconnectedness will make us happier and even make unemployment a thing of the past. Their vision of everyone sitting at a console, working “through the network” is stunningly accurate for an information-driven society that fifty years ago would’ve looked far less tech-obsessed.

When people do their informational work “at the console” and “through the network,” telecommunication will be as natural an extension of individual work as face-to-face communication is now. The impact of that fact, and of the marked facilitation of the communicative process, will be very great—both on the individual and on society.
First, life will be happier for the on-line individual because the people with whom one interacts most strongly will be selected more by commonality of interests and goals than by accidents of proximity. Second, communication will be more effective and productive, and therefore more enjoyable. Third, much communication and interaction will be with programs and programmed models, which will be (a) highly responsive, (b) supplementary to one’s own capabilities, rather than competitive, and (c) capable of representing progressively more complex ideas without necessarily displaying all the levels of their structure at the same time-and which will therefore be both challenging and rewarding. And, fourth, there will be plenty of opportunity for everyone (who can afford a console) to find his calling, for the whole world of information, with all its fields and disciplines, will be open to him—with programs ready to guide him or to help him explore.

(You can read the entire paper online [pdf]. )”

What Our Words Tell Us


David Brooks in the New York Times: “About two years ago, the folks at Google released a database of 5.2 million books published between 1500 and 2008. You can type a search word into the database and find out how frequently different words were used at different epochs….

I’d like to tell a story about the last half-century, based on studies done with this search engine. The first element in this story is rising individualism. A study by Jean M. Twenge, W. Keith Campbell and Brittany Gentile found that between 1960 and 2008 individualistic words and phrases increasingly overshadowed communal words and phrases. That is to say, over those 48 years, words and phrases like “personalized,” “self,” “standout,” “unique,” “I come first” and “I can do it myself” were used more frequently. Communal words and phrases like “community,” “collective,” “tribe,” “share,” “united,” “band together” and “common good” receded.

The second element of the story is demoralization. A study by Pelin Kesebir and Selin Kesebir found that general moral terms like “virtue,” “decency” and “conscience” were used less frequently over the course of the 20th century. Words associated with moral excellence, like “honesty,” “patience” and “compassion” were used much less frequently. The Kesebirs identified 50 words associated with moral virtue and found that 74 percent were used less frequently as the century progressed. Certain types of virtues were especially hard hit. Usage of courage words like “bravery” and “fortitude” fell by 66 percent. Usage of gratitude words like “thankfulness” and “appreciation” dropped by 49 percent. Usage of humility words like “modesty” and “humbleness” dropped by 52 percent. Usage of compassion words like “kindness” and “helpfulness” dropped by 56 percent. Meanwhile, usage of words associated with the ability to deliver, like “discipline” and “dependability” rose over the century, as did the usage of words associated with fairness. The Kesebirs point out that these sorts of virtues are most relevant to economic production and exchange.

Daniel Klein of George Mason University has conducted one of the broadest studies with the Google search engine. He found further evidence of the two elements I’ve mentioned. On the subject of individualization, he found that the word “preferences” was barely used until about 1930, but usage has surged since. On the general subject of demoralization, he finds a long decline of usage in terms like “faith,” “wisdom,” “ought,” “evil” and “prudence,” and a sharp rise in what you might call social science terms like “subjectivity,” “normative,” “psychology” and “information.” Klein adds the third element to our story, which he calls “governmentalization.” Words having to do with experts have shown a steady rise. So have phrases like “run the country,” “economic justice,” “nationalism,” “priorities,” “right-wing” and “left-wing.” The implication is that politics and government have become more prevalent.

So the story I’d like to tell is this: Over the past half-century, society has become more individualistic. As it has become more individualistic, it has also become less morally aware, because social and moral fabrics are inextricably linked. The atomization and demoralization of society have led to certain forms of social breakdown, which government has tried to address, sometimes successfully and often impotently.”

NSF-funded Research on Elections and Cooperation


John Sides @ the Monkey Cage: “Much of politics is about collective action, whereby groups of people need to cooperate in order to produce an outcome.  One of the biggest challenges is getting people to cooperate in providing a public good, which by its nature can be shared by everyone regardless of whether they’ve cooperated in the first place.
One way to enforce cooperation is via some central authority that’s external to the group (like a government).  Another way, prominent in Elinor Ostrom’s work, is via internal policing by peers within the group.
In this NSF-funded study by Guy Grossman and Delia Baldassarri show that a third way can work as well: developing a leadership or authority structure within the group itself.  More importantly, they show that the success of such an authority depends on politics itself.  Leaders need to be elected to induce cooperation.The study was conducted among Ugandans who are members of farmer organizations and experience directly the challenges of cooperating to produce public goods.  Grossman and Baldassarri not only examined how these people behaved when asked to play a simple “public goods game” in a quasi-laboratory setting, but how they actually behaved within their farmer organization in real life.  In both contexts, members cooperated significantly more when leaders were democratically elected—as was true in one experimental condition of the public goods game—or when they perceived the leadership of their farmer organization as more legitimate.
For more in this week’s presentation of NSF-funded research recently published in the American Journal of Political Science, see here, here, and here.]”

Data Edge


Steven Weber, professor in the School of Information and Political Science department at UC Berkeley, in Policy by the Numbers“It’s commonly said that most people overestimate the impact of technology in the short term, and underestimate its impact over the longer term.
Where is Big Data in 2013? Starting to get very real, in our view, and right on the cusp of underestimation in the long term. The short term hype cycle is (thankfully) burning itself out, and the profound changes that data science can and will bring to human life are just now coming into focus. It may be that Data Science is right now about where the Internet itself was in 1993 or so. That’s roughly when it became clear that the World Wide Web was a wind that would blow across just about every sector of the modern economy while transforming foundational things we thought were locked in about human relationships, politics, and social change. It’s becoming a reasonable bet that Data Science is set to do the same—again, and perhaps even more profoundly—over the next decade. Just possibly, more quickly than that….
Can data, no matter how big, change the world for the better? It may be the case that in some fields of human endeavor and behavior, the scientific analysis of big data by itself will create such powerful insights that change will simply have to happen, that businesses will deftly re-organize, that health care will remake itself for efficiency and better outcomes, that people will adopt new behaviors that make them happier, healthier, more prosperous and peaceful. Maybe. But almost everything we know about technology and society across human history argues that it won’t be so straightforward.
…join senior industry and academic leaders at DataEDGE at UC Berkeley on May 30-31 to engage in what will be a lively and important conversation aimed at answering today’s questions about the data science revolution—and formulating tomorrow’s.