Review by New York Times Review
IN "ON WHAT WE CAN NOT DO," a short and pungent essay published a few years ago, the Italian philosopher Giorgio Agamben outlined two ways in which power operates today. There's the conventional type that seeks to limit our potential for self-development by restricting material resources and banning certain behaviors. But there's also a subtler, more insidious type, which limits not what we can do but what we can not do. What's at stake here is not so much our ability to do things but our capacity not to make use of that very ability. While each of us can still choose not to be on Facebook, have a credit history or build a presence online, can we really afford not to do any of those things today? It was acceptable not to have a cellphone when most people didn't have them; today, when almost everybody does and when our phone habits can even be used to assess whether we qualify for a loan, such acts of refusal border on the impossible. For Agamben, it's this double power "to be and to not be, to do and to not do" that makes us human. This active necessity to choose (and err) contributes to the development of individual faculties that shape our subjectivity. The tragedy of modern man, then, is that "he has become blind not to his capacities but to his incapacities, not to what he can do but to what he cannot, or can not, do." This blindness to the question of incapacities mars most popular books on recent advances in our ability to store, analyze and profit from vast amounts of data generated by our gadgets. (Our wherewithal not to call this phenomenon by the ugly, jargony name of Big Data seems itself to be under threat.) The two books under review, alas, are no exception. In "The Naked Future," Patrick Tucker, an editor at large for The Futurist magazine, surveys how this influx of readily available data will transform every domain of our existence, from improving our ability to predict earthquakes (thanks to the proliferation of sensors) to producing highly customized education courses that would tailor their content and teaching style, in real time, to the needs of individual students. His verdict: It's all for the better. Since most of us lead rather structured, regular lives - work, home, weekend - even a handful of data points (our location, how often we call our friends) proves useful in predicting what we may be doing a day or a year from now. "A flat tire on a Monday at 10 a.m. isn't actually random. ... We just don't yet know how to model it," Tucker writes. Seeking to integrate data streams from multiple sources - our inboxes, our phones, our cars and, with its recent acquisition of a company that makes thermostats and smoke detectors, our bedrooms - a company like Google is well positioned not just to predict our future but also to detect just how much risk we take on every day, be it fire, a flat tire or a default on a loan. (Banks and insurance companies beware: You will be disrupted next!) With so much predictive power, we may soon know the exact price of "preferring not to," as a modern-day Bartleby might put it. Would you skip the gym tonight if your smartphone told you this would (a) increase your risk of heart attack by 5 percent or (b) result in higher health insurance payments? Tucker doesn't appear too concerned. To his own prediction that there will come a day when his gadgets will send him a similar note, he can only complain that for this to work, we need data from other people, not just him - and that "our outmoded ideas of privacy begin to get in the way of progress and better health." The predictive models Tucker celebrates are good at telling us what could happen, but they cannot tell us why. As Tucker himself acknowledges, we can learn that some people are more prone to having flat tires and, by analyzing heaps of data, we can even identify who they are - which might be enough to prevent an accident - but the exact reasons defy us. Such aversion to understanding causality has a political cost. To apply such logic to more consequential problems - health, education, crime - could bias us into thinking that our problems stem from our own poor choices. This is not very surprising, given that the self-tracking gadget in our hands can only nudge us to change our behavior, not reform society at large. But surely many of the problems that plague our health and educational systems stem from the failures of institutions, not just individuals. IN HIS NEW BOOK, "Social Physics," Alex Pentland, a prominent data scientist at M.I.T., shows as much uncritical enthusiasm for prediction as Tucker, while making a case that we need a new science - social physics - that can make sense of all the digital bread crumbs, from call records to credit card transactions, that we leave as we navigate our daily life. (That the idea of social physics was once promoted by the positivist Auguste Comte, one scholar who would have warmed to the idea of Big Data, goes unmentioned.) What is social physics good for? It would allow us to detect and improve "idea flow" - the way ideas and behaviors travel through social networks. For example, Pentland wants to arm employers with sophisticated gadgets that would allow them to monitor the communicative activities of their employees and coax them toward more productive behaviors so their cognitive activity isn't wasted on trifles. That this might lead to a new form of intellectual Taylorism, with managers optimizing the efficiency of the brainstorming session (rather than the time spent at the conveyor belt), seems of little concern to Pentland, who dryly remarks, "What isn't measured can't be managed." Employers would certainly love this, but why should employees acquiesce to ubiquitous surveillance? Pentland rarely pauses to discuss the political implications of his agenda, arguing that we must make our social systems more dynamic, automated and data-dependent, as if data, by itself, can settle all political conflicts once and for all. Both books reveal - mostly through their flaws - that the Big Data debate needs grounding in philosophy. When Big Data allows us to automate decisionmaking, or at least contextualize every decision with a trove of data about its likely consequences, we need to grapple with the question of just how much we want to leave to chance and to those simple, low-tech, unautomated options of democratic contestation and deliberation. As we gain the capacity to predict and even pre-empt crises, we risk eliminating the very kinds of experimental behaviors that have been conducive to social innovation. Occasionally, someone needs to break the law, engage in an act of civil disobedience or simply refuse to do something the rest of us find useful. The temptation of Big Data lies precisely in allowing us to identify and make such loopholes unavailable to deviants, who might actually be dissidents in disguise. It may be that the first kind of power identified by Agamben is actually less pernicious, for, in barring us from doing certain things, it at least preserves, even nurtures, our capacity to resist. But as we lose our ability not to do - here Agamben is absolutely right - our capacity to resist goes away with it. Perhaps it's easier to resist the power that bars us from using our smartphones than the one that bars us from not using them. Big Data does not a free society make, at least not without basic political judgment. EVGENY MOROZOV is the author, most recently, of "To Save Everything, Click Here" and a senior editor at The New Republic.
Copyright (c) The New York Times Company [June 5, 2014]
Review by Publisher's Weekly Review
Every time we swipe a debit card or a subway card, activate our GPS, or post on Facebook from our phones, we leave an electronic trail for others to follow. According to the Futurist magazine's deputy editor Tucker, each of us now creates 1.8 megabytes of data a year using our devices in such ways. Thanks to the wonders of telemetry-the transmission of measurements of data-we are now on the edge of a world in which individuals and collective agencies will be able to use our data to predict many aspects of our lives. In this fascinating and gripping book, Tucker illustrates how such predictive powers will tell us about our personal health before we know it and how that health will affect others, where and when a crime might happen and who might become a victim of a crime, and when you might fall in love. As much as we wish to retain our privacy, even when using these devices, the future in which we stand naked to the world is closer than we imagine, according to Tucker. (Apr.) (c) Copyright PWxyz, LLC. All rights reserved.
(c) Copyright PWxyz, LLC. All rights reserved
Review by Kirkus Book Review
An upbeat view of big data as an empowering means for predicting the future. Futurist magazine deputy editor Tucker provides an anecdote-filled account of the many ways in which massive sets of datathe same digital information often used by governments and large corporations for privacy-invading tracking and surveillancecan be used by individuals to "live much more healthily, realize more of your own goals in less time [and] avoid inconvenience and danger." Based on interviews with hackers, entrepreneurs, scientists and others, the author argues that a "thrilling and historic transformation" lies ahead in our ability to predict the future using continuously sourced streams of information accessed via smartphones. Such information, distributed from the site of a fire or disaster as a live-stream video by anyone with a cellphone, can prepare emergency workers. In the same way, individuals bent on improving their personal health can track signals, physical states and other data to assess upcoming issues. Acting in groups, individuals can share highly personal health data and make it possible to predict strokes based on correlations among thousands of patients. Also, with better and faster reporting on new flu strains, it becomes possible to predict more accurately where a flu outbreak will go next. Tucker's exploration of computer-aided forecasting shows the growing role of big data in aspects of American life, including education, online dating, predictive policing and customer loyalty programs. He urges readers to become familiar with existing technologies that make it possible to collect big data (systems, networks and communities) and put it to work (apps, programs and platforms) and to understand how the data can be used, or abused, as many fear, by consumers, activists and regular people. A well-written consideration of how, "in the next two decades, we will be able to predict huge areas of the future with far greater accuracy than ever before in human history, including events long thought to be beyond the realm of human interference."]] Copyright Kirkus Reviews, used with permission.
Copyright (c) Kirkus Reviews, used with permission.
Review by New York Times Review
Review by Publisher's Weekly Review
Review by Kirkus Book Review