Buy new:
-55% $12.13$12.13
Delivery Friday, December 13
Ships from: Amazon Sold by: Petek Books Store
Save with Used - Very Good
$7.28$7.28
Delivery Friday, December 6
Ships from: Amazon Sold by: PRIME-BOOKS
Download the free Kindle app and start reading Kindle books instantly on your smartphone, tablet, or computer - no Kindle device required.
Read instantly on your browser with Kindle for Web.
Using your mobile phone camera - scan the code below and download the Kindle app.
Follow the authors
OK
Big Data: A Revolution That Will Transform How We Live, Work, and Think Hardcover – March 5, 2013
Purchase options and add-ons
A revelatory exploration of the hottest trend in technology and the dramatic impact it will have on the economy, science, and society at large.
Which paint color is most likely to tell you that a used car is in good shape? How can officials identify the most dangerous New York City manholes before they explode? And how did Google searches predict the spread of the H1N1 flu outbreak?
The key to answering these questions, and many more, is big data. “Big data” refers to our burgeoning ability to crunch vast collections of information, analyze it instantly, and draw sometimes profoundly surprising conclusions from it. This emerging science can translate myriad phenomena—from the price of airline tickets to the text of millions of books—into searchable form, and uses our increasing computing power to unearth epiphanies that we never could have seen before. A revolution on par with the Internet or perhaps even the printing press, big data will change the way we think about business, health, politics, education, and innovation in the years to come. It also poses fresh threats, from the inevitable end of privacy as we know it to the prospect of being penalized for things we haven’t even done yet, based on big data’s ability to predict our future behavior.
In this brilliantly clear, often surprising work, two leading experts explain what big data is, how it will change our lives, and what we can do to protect ourselves from its hazards. Big Data is the first big book about the next big thing.
www.big-data-book.com
- Print length256 pages
- LanguageEnglish
- PublisherEamon Dolan/Houghton Mifflin Harcourt
- Publication dateMarch 5, 2013
- Dimensions6.5 x 1 x 9.5 inches
- ISBN-100544002695
- ISBN-13978-0544002692
Book recommendations, author interviews, editors' picks, and more. Read it now
Frequently bought together

Similar items that may deliver to you quickly
Data Science for Business: What You Need to Know about Data Mining and Data-Analytic ThinkingPaperback$14.39 shippingGet it as soon as Friday, Dec 13Only 1 left in stock - order soon.
Big Data: A Revolution That Will Transform How We Live, Work, and ThinkAudio CD$12.88 shippingOnly 2 left in stock - order soon.
Editorial Reviews
Amazon.com Review
Amazon Exclusive: Q&A with Kenneth Cukier and Viktor Mayer-Schonberger
Q. What did it take to write Big Data?
A. Kenn has written about technology and business from Europe, Asia, and the US for The Economist, and is well-connected to the data community. Viktor had researched the information economy as a professor at Harvard and now at Oxford, and his book Delete had been well received. So we thought we had a good basis to make a contribution in the area. As we wrote the book, we had to dig deep to find unheard stories about big data pioneers and interview them. We wanted Big Data to be about a big idea, but also to be full of examples and success stories -- and be engrossing to read.
Q. Are you big data’s cheerleaders?
A. Absolutely not. We are the messengers of big data, not its evangelists. The big data age is happening, and in the book we take a look at the drivers, and big data’s likely trajectory: how it will change how we work and live. We emphasize that the fundamental shift is not in the machines that calculate data, but in the data itself and how we use it.
Q. In discovering big data applications, what was your biggest surprise?
A. It is tempting to say that it was predicting exploding manholes, tracking inflation in real time, or how big data saves the lives of premature babies. But the biggest surprise for us perhaps was the very diversity of the uses of big data, and how it already is changing people’s everyday world. Many people see big data through the lens of the Internet economy, since Google and Facebook have so much data. But that misses the point: big data is everywhere.
Q. Is Big Data then primarily a story about economic efficiency?
A. Big data improves economic efficiency, but that’s only a very small part of the story. We realized when talking to dozens and dozens of big data pioneers that it improves health care, advances better education, and helps predict societal change—from urban sprawl to the spread of the flu. Big data is roaring through all sectors of the economy and all areas of life.
Q. So big data offers only “upside”?
A. Not at all. We are very concerned about what we call in our book “the dark side of big data.” However the real challenge is that the problem is not necessarily where we initially tend to think it is, such as surveillance and privacy. After looking into the potential misuses of big data, we became much more troubled by “propensity” -- that is, big data predictions being used to police and punish. And by the “fetishization” of data that may occur, whereby organizations may blindly defer to what the data says without understanding its limitations.
Q. What can we do about this “dark side”?
A. Knowing about it is the first step. We thought hard to suggest concrete steps that can be taken to minimize and mitigate big data’s risk, and came up with a few ways to ensure transparency, guarantee human free will, and strike a better balance on privacy and the use of personal information. These are deeply serious issues. If we do not take action soon, it might be too late.
From Booklist
Review
—Lawrence Lessig, Roy L. Furman Professor of Law, Harvard Law School, and author of Remix and Free Culture
"This brilliant book cuts through the mystery and the hype surrounding big data.
A must-read for anyone in business, information technology, public policy, intelligence, and medicine. And anyone else who is just plain curious about the future."
—John Seely Brown, former Chief Scientist, Xerox Corp., and head of Xerox Palo Alto Research Center
—Joi Ito, Director of the MIT Media Lab"Big Data is a must-read for anyone who wants to stay ahead of one of the key trends defining the future of business."
—Marc Benioff, Chairman and CEO, salesforce.com"An optimistic and practical look at the Big Data revolution — just the thing to get your head around the big changes already underway and the bigger changes to come."
—Cory Doctorow, boingboing.com"Just as water is wet in a way that individual water molecules aren’t, big data can reveal information in a way that individual bits of data can’t. The authors show us the surprising ways that enormous, complex, and messy collections of data can be used to predict everything from shopping patterns to flu outbreaks."
—Clay Shirky, author of Cognitive Surplus and Here Comes Everybody"The book teems with great insights on the new ways of harnessing information, and offers a convincing vision of the future. It is essential reading for anyone who uses — or is affected by — big data."
—Jeff Jonas, IBM Fellow & Chief Scientist, IBM Entity Analytics“What I’m certain about is that Big Data will be the defining text in the discussion for some time to come.”
—Forbes.com
“The authors make clear that ‘big data’ is much more than a Silicon Valley buzzword… No other book offers such an accessible and balanced tour of the many benefits and downsides of our continuing infatuation with data.”
—Wall Street Journal
"Plenty of books extol the technical marvels of our information society, but this is an original analysis of the information itself—trillions of searches, calls, clicks, queries and purchases....A fascinating, enthusiastic view of the possibilities of vast computer correlations and the entrepreneurs who are taking advantage of them."
—STARRED Kirkus Reviews
"This book offers important insights and information"
—Booklist
"'big data' [is] one of the buzzwords of corporate executives, tech-savvy politicians, and worried civil libertarians. If you want to know what they’re all talking about, then Big Data is the book for you, a comprehensive and entertaining introduction to a very large topic....Mayer-Schönberger and Cukier offer up some sensible suggestions on how we can have the blessings of big data and our freedoms, too. Just as well; their lively book leaves no doubt that big data’s growth spurt is just beginning."
—Boston Globe
From the Inside Flap
A revelatory exploration of the hottest trend in technology and the dramatic impact it will have on the economy, science, and society at large.
Which paint color is most likely to tell you that a used car is in good shape? How can officials identify the most dangerous New York City manholes before they explode? And how did Google searches predict the spread of the H1N1 flu outbreak?
The key to answering these questions, and many more, is big data. Big data refers to our burgeoning ability to crunch vast collections of information, analyze it instantly, and draw sometimes profoundly surprising conclusions from it. This emerging science can translate myriad phenomena from the price of airline tickets to the text of millions of books into searchable form, and uses our burgeoning computing power to unearth epiphanies that we never could have seen before. A revolution on par with the Internet or perhaps even the printing press, big data will change the way we think about business, health, politics, education, and innovation in the years to come. It also poses fresh threats, from the inevitable end of privacy as we know it to the prospect of being penalized for things we haven t even done yet, based on big data s ability to predict our future behavior.
In this brilliantly clear, often surprising work, two leading experts explain what big data is, how it will change our lives, and what we can do to protect ourselves from its hazards. Big Data is the first big book about the next big thing.
From the Back Cover
[Quotes TK]
About the Author
KENNETH CUKIER is the Data Editor of the Economist and co-author of Big Data: A Revolution That Will Transform How We Live, Work, and Think. His writings on business and economics have appeared in Foreign Affairs, the New York Times, the Financial Times, and elsewhere.
Excerpt. © Reprinted by permission. All rights reserved.
NOW
IN 2009 A NEW FLU virus was discovered. Combining elements of the viruses that cause bird flu and swine flu, this new strain, dubbed H1N1, spread quickly. Within weeks, public health agencies around the world feared a terrible pandemic was under way. Some commentators warned of an outbreak on the scale of the 1918 Spanish flu that had infected half a billion people and killed tens of millions. Worse, no vaccine against the new virus was readily available. The only hope public health authorities had was to slow its spread. But to do that, they needed to know where it already was.
In the United States, the Centers for Disease Control and Prevention (CDC) requested that doctors inform them of new flu cases. Yet the picture of the pandemic that emerged was always a week or two out of date. People might feel sick for days but wait before consulting a doctor. Relaying the information back to the central organizations took time, and the CDC only tabulated the numbers once a week. With a rapidly spreading disease, a two-week lag is an eternity. This delay completely blinded public health agencies at the most crucial moments.
As it happened, a few weeks before the H1N1 virus made headlines, engineers at the Internet giant Google published a remarkable paper in the scientific journal Nature. It created a splash among health officials and computer scientists but was otherwise overlooked. The authors explained how Google could “predict” the spread of the winter flu in the United States, not just nationally, but down to specific regions and even states. The company could achieve this by looking at what people were searching for on the Internet. Since Google receives more than three billion search queries every day and saves them all, it had plenty of data to work with.
Google took the 50 million most common search terms that Americans type and compared the list with CDC data on the spread of seasonal flu between 2003 and 2008. The idea was to identify people infected by the flu virus by what they searched for on the Internet. Others had tried to do this with Internet search terms, but no one else had as much data, processing power, and statistical know-how as Google.
While the Googlers guessed that the searches might be aimed at getting flu information — typing phrases like “medicine for cough and fever” — that wasn’t the point: they didn’t know, and they designed a system that didn’t care. All their system did was look for correlations between the frequency of certain search queries and the spread of the flu over time and space. In total, they processed a staggering 450 million different mathematical models in order to test the search terms, comparing its predictions against actual flu cases from the CDC in 2007 and 2008. And they struck gold: their software found a combination of 45 search terms that, when used together in a mathematical model, had a strong correlation between their prediction and the official figures nationwide. Like the CDC, they could tell where the flu had spread, but unlike the CDC they could tell it in near real-time, not a week or two after the fact.
Thus when the H1N1 crisis struck in 2009, Google’s system proved to be a more useful and timely indicator than government statistics with their natural reporting lags. Public health officials were armed with valuable information.
Strikingly, Google’s method does not involve distributing mouth swabs or contacting physicians’ offices. Instead, it is built on “big data” — the ability of society to harness information in novel ways to produce useful insights or goods and services of significant value. With it, by the time the next pandemic comes around, the world will have a better tool at its disposal to predict and thus prevent its spread. Public health is only one area where big data is making a big difference. Entire business sectors are being reshaped by big data as well. Buying airplane tickets is a good example.
In 2003 Oren Etzioni needed to fly from Seattle to Los Angeles for his younger brother’s wedding. Months before the big day, he went online and bought a plane ticket, believing that the earlier you book, the less you pay. On the flight, curiosity got the better of him and he asked the fellow in the next seat how much his ticket had cost and when he had bought it. The man turned out to have paid considerably less than Etzioni, even though he had purchased the ticket much more recently. Infuriated, Etzioni asked another passenger and then another. Most had paid less.
For most of us, the sense of economic betrayal would have dissipated by the time we closed our tray tables and put our seats in the full, upright, and locked position. But Etzioni is one of America’s foremost computer scientists. He sees the world as a series of big-data problems — ones that he can solve. And he has been mastering them since he graduated from Harvard in 1986 as its first undergrad to major in computer science.
From his perch at the University of Washington, he started a slew of big-data companies before the term “big data” became known. He helped build one of the Web’s first search engines, MetaCrawler, which was launched in 1994 and snapped up by InfoSpace, then a major online property. He co-founded Netbot, the first major comparison-shopping website, which he sold to Excite. His startup for extracting meaning from text documents, called ClearForest, was later acquired by Reuters.
Back on terra firma, Etzioni was determined to figure out a way for people to know if a ticket price they see online is a good deal or not. An airplane seat is a commodity: each one is basically indistinguishable from others on the same flight. Yet the prices vary wildly, being based on a myriad of factors that are mostly known only by the airlines themselves.
Etzioni concluded that he didn’t need to decrypt the rhyme or reason for the price differences. Instead, he simply had to predict whether the price being shown was likely to increase or decrease in the future. That is possible, if not easy, to do. All it requires is analyzing all the ticket sales for a given route and examining the prices paid relative to the number of days before the departure.
If the average price of a ticket tended to decrease, it would make sense to wait and buy the ticket later. If the average price usually increased, the system would recommend buying the ticket right away at the price shown. In other words, what was needed was a souped-up version of the informal survey Etzioni conducted at 30,000 feet. To be sure, it was yet another massive computer science problem. But again, it was one he could solve. So he set to work.
Using a sample of 12,000 price observations that was obtained by “scraping” information from a travel website over a 41-day period, Etzioni created a predictive model that handed its simulated passengers a tidy savings. The model had no understanding of why, only what. That is, it didn’t know any of the variables that go into airline pricing decisions, such as number of seats that remained unsold, seasonality, or whether some sort of magical Saturday-night-stay might reduce the fare. It based its prediction on what it did know: probabilities gleaned from the data about other flights. “To buy or not to buy, that is the question,” Etzioni mused. Fittingly, he named the research project Hamlet.
The little project evolved into a venture capital-backed startup called Farecast. By predicting whether the price of an airline ticket was likely to go up or down, and by how much, Farecast empowered consumers to choose when to click the “buy” button. It armed them with information to which they had never had access before. Upholding the virtue of transparency against itself, Farecast even scored the degree of confidence it had in own predictions and presented that information to users too.
To work, the system needed lots of data. To improve its performance, Etzioni got his hands on one of the industry’s flight reservation databases. With that information, the system could make predictions based on every seat on every flight for most routes in American commercial aviation over the course of a year. Farecast was now crunching nearly 200 billion flight-price records to make its predictions. In so doing, it was saving consumers a bundle.
With his sandy brown hair, toothy grin, and cherubic good looks, Etzioni hardly seemed like the sort of person who would deny the airline industry millions of dollars of potential revenue. In fact, he set his sights on doing even more than that. By 2008 he was planning to apply the method to other goods like hotel rooms, concert tickets, and used cars: anything with little product differentiation, a high degree of price variation, and tons of data. But before he could hatch his plans, Microsoft came knocking on his door, snapped up Farecast for around $110 million, and integrated it into the Bing search engine. By 2012 the system was making the correct call 75 percent of the time and saving travelers, on average, $50 per ticket.
Farecast is the epitome of a big-data company and an example of where the world is headed. Etzioni couldn’t have built the company five or ten years earlier. “It would have been impossible,” he says. The amount of computing power and storage he needed was too expensive. But although changes in technology have been a critical factor making it possible, something more important changed too, something subtle. There was a shift in mindset about how data could be used.
Data was no longer regarded as static or stale, whose usefulness was finished once the purpose for which it was collected was achieved, such as after the plane landed (or in Google’s case, once a search query had been processed). Rather, data became a raw material of business, a vital economic input, used to create a new form of economic value. In fact, with the right mindset, data can be cleverly reused to become a fountain of innovation and new services. The data can reveal secrets to those with the humility, the willingness, and the tools to listen.
Letting the data speak
The fruits of the information society are easy to see, with a cellphone in every pocket, a computer in every backpack, and big information technology systems in back offices everywhere. But less noticeable is the information itself. Half a century after computers entered mainstream society, the data has begun to accumulate to the point where something new and special is taking place. Not only is the world awash with more information than ever before, but that information is growing faster. The change of scale has led to a change of state. The quantitative change has led to a qualitative one. The sciences like astronomy and genomics, which first experienced the explosion in the 2000s, coined the term “big data.” The concept is now migrating to all areas of human endeavor.
There is no rigorous definition of big data. Initially the idea was that the volume of information had grown so large that the quantity being examined no longer fit into the memory that computers use for processing, so engineers needed to revamp the tools they used for analyzing it all. That is the origin of new processing technologies like Google’s MapReduce and its open-source equivalent, Hadoop, which came out of Yahoo. These let one manage far larger quantities of data than before, and the data — importantly — need not be placed in tidy rows or classic database tables. Other data-crunching technologies that dispense with the rigid hierarchies and homogeneity of yore are also on the horizon. At the same time, because Internet companies could collect vast troves of data and had a burning financial incentive to make sense of them, they became the leading users of the latest processing technologies, superseding offline companies that had, in some cases, decades more experience.
One way to think about the issue today — and the way we do in the book — is this: big data refers to things one can do at a large scale that cannot be done at a smaller one, to extract new insights or create new forms of value, in ways that change markets, organizations, the relationship between citizens and governments, and more.
But this is just the start. The era of big data challenges the way we live and interact with the world. Most strikingly, society will need to shed some of its obsession for causality in exchange for simple correlations: not knowing why but only what. This overturns centuries of established practices and challenges our most basic understanding of how to make decisions and comprehend reality.
Big data marks the beginning of a major transformation. Like so many new technologies, big data will surely become a victim of Silicon Valley’s notorious hype cycle: after being feted on the cover of magazines and at industry conferences, the trend will be dismissed and many of the data-smitten startups will flounder. But both the infatuation and the damnation profoundly misunderstand the importance of what is taking place. Just as the telescope enabled us to comprehend the universe and the microscope allowed us to understand germs, the new techniques for collecting and analyzing huge bodies of data will help us make sense of our world in ways we are just starting to appreciate. In this book we are not so much big data’s evangelists, but merely its messengers. And, again, the real revolution is not in the machines that calculate data but in data itself and how we
use it. To appreciate the degree to which an information revolution is already under way, consider trends from across the spectrum of society. Our digital universe is constantly expanding. Take astronomy. When the Sloan Digital Sky Survey began in 2000, its telescope in New Mexico collected more data in its first few weeks than had been amassed in the entire history of astronomy. By 2010 the survey’s archive teemed with a whopping 140 terabytes of information. But a successor, the Large Synoptic Survey Telescope in Chile, due to come on stream in 2016, will acquire that quantity of data every five days.
Such astronomical quantities are found closer to home as well. When scientists first decoded the human genome in 2003, it took them a decade of intensive work to sequence the three billion base pairs. Now, a decade later, a single facility can sequence that much DNA in a day. In finance, about seven billion shares exchange hands every day on U.S. equity markets, of which around two-thirds is traded by computer algorithms based on mathematical models that crunch mountains of data to predict gains while trying to reduce risk.
Internet companies have been particularly swamped. Google processes more than 24 petabyte of data per day, a volume that is thousands of times the quantity of all printed material in the U.S. Library of Congress. Facebook, a company that didn’t exist a decade ago, gets more than 10 million new photos uploaded every hour. Facebook members click a “like” button or leave a comment nearly three billion times per day, creating a digital trail that the company can mine to learn about users’ preferences. Meanwhile, the 800 million monthly users of Google’s YouTube service upload over an hour of video every second. The number of messages on Twitter grows at around 200 percent a year and by 2012 had exceeded 400 million tweets a day.
From the sciences to healthcare, from banking to the Internet, the sectors may be diverse yet together they tell a similar story: the amount of data in the world is growing fast, outstripping not just our machines but our imaginations.
Many people have tried to put an actual figure on the quantity of information that surrounds us and to calculate how fast it grows. They’ve had varying degrees of success because they’ve measured different things. One of the more comprehensive studies was done by Martin Hilbert of the University of Southern California’s Annenberg School for Communication and Journalism. He has striven to put a figure on everything that has been produced, stored, and communicated. That would include not only books, paintings, emails, photographs, music, and video (analog and digital), but video games, phone calls, even car navigation systems and letters sent through the mail. He also included broadcast media like television and radio, based on audience reach.
By Hilbert’s reckoning, more than 300 exabytes of stored data existed in 2007. To understand what this means in slightly more human terms, think of it like this. A full-length feature film in digital form can be compressed into a one gigabyte file. An exabyte is one billion gigabytes. In short, it’s a lot. Interestingly, in 2007 only about 7 percent of the data was analog (paper, books, photographic prints, and so on). The rest was digital. But not long ago the picture looked very different. Though the ideas of the “information revolution” and “digital age” have been around since the 1960s, they have only just become a reality by some measures. As recently as the year 2000, only one-fourth of the stored information in the world was digital. The other three-quarters were on paper, film, vinyl LP records, magnetic cassette tapes, and the like.
The mass of digital information then was not much — a humbling thought for those who have been surfing the Web and buying books online for a long time. (In fact, in 1986 around 40 percent of the world’s general-purpose computing power took the form of pocket calculators, which represented more processing power than all personal computers at the time.) But because digital data expands so quickly — a doubling a little more than every three years, according to Hilbert — the situation quickly inverted itself. Analog information, in contrast, hardly grows at all. So in 2013 the amount of stored information in the world is estimated to be around 1,200 exabytes, of which less than 2 percent is non-digital.
There is no good way to think about what this size of data means. If it were all printed in books, they would cover the entire surface of the United States some 52 layers thick. If it were placed on CD-ROMs and stacked up, they would stretch to the moon in five separate piles. In the third century B.C., as Ptolemy II of Egypt strove to store a copy of every written work, the great Library of Alexandria represented the sum of all knowledge in the world. The digital deluge now sweeping the globe is the equivalent of giving every person living on Earth today 320 times as much information as is estimated to have been stored in the Library of Alexandria.
Product details
- Publisher : Eamon Dolan/Houghton Mifflin Harcourt; 1st edition (March 5, 2013)
- Language : English
- Hardcover : 256 pages
- ISBN-10 : 0544002695
- ISBN-13 : 978-0544002692
- Item Weight : 1 pounds
- Dimensions : 6.5 x 1 x 9.5 inches
- Best Sellers Rank: #955,555 in Books (See Top 100 in Books)
- #430 in Management Science
- #496 in Information Management (Books)
- #666 in Computers & Technology Industry
- Customer Reviews:
About the authors

Viktor Mayer-Schönberger (born 1966) is Professor of Internet Governance and Regulation at the Oxford Internet Institute, University of Oxford. He is also a faculty affiliate at Harvard's Belfer Center. Mayer-Schönberger is the co-author of "Framers" (with Kenneth Cukier and Francis de Vericourt), the acclaimed "Reinventing Capitalism" (with Thomas Ramge), the international bestseller "Big Data" (with Kenneth Cukier), and the awards-winning 'Delete".

Kenneth Cukier is an award-winning journalist and bestselling author. He is the Deputy Executive Editor at The Economist and host of its weekly tech podcast. His book "Big Data" with Viktor Mayer-Schönberger was a NYT bestseller and translated into over 20 languages. From 2002-04 Kenn was a research fellow at Harvard's Kennedy School of Government. He is on the board of directors of Chatham House, a British foreign-policy think-tank and is an associate fellow at Saïd Business School at the University of Oxford. He is a member of the Council on Foreign Relations.
Customer reviews
Customer Reviews, including Product Star Ratings help customers to learn more about the product and decide whether it is the right product for them.
To calculate the overall star rating and percentage breakdown by star, we don’t use a simple average. Instead, our system considers things like how recent a review is and if the reviewer bought the item on Amazon. It also analyzed reviews to verify trustworthiness.
Learn more how customers reviews work on AmazonCustomers say
Customers find the writing quality thought-provoking and solid. They describe the book as interesting, wonderful, and an important quick read. Readers also say it provides an excellent introduction to big data with nice examples and stories.
AI-generated from the text of customer reviews
Customers find the writing quality of the book thought-provoking and solid. They say the examples are well-articulated, the notes and bibliography are worth reading. Readers also mention the book provides a good overview of big data, what it is, and how it works.
"...There's a lot of useful information in this book, especially for someone just trying to learn about the concept of Big Data...." Read more
"...The chapters are set out fantastically by how we are now collecting and using data, how we can obtain more data and how (and why) this will be useful..." Read more
"...Drilling down is easier and more informative. Etc...." Read more
"...To the initiated in Big Data, there are some fantastic arguments and well thought out opinions on how the industry should proceed as a whole...." Read more
Customers find the book worth reading, interesting, and a quick read. They appreciate the wonderful, descriptive prose. Readers also mention the book does a good job of explaining a new phenomenon.
"...In any case the book is interesting and highly readable despite those shortcomings...." Read more
"...& quite frankly the popular film examples were less useful - but well worth a read." Read more
"...It's worth a read I think...." Read more
"The book is great. Truthfully, I would recommend to anyone who is interested in any type of data...." Read more
Customers find the book excellent and helpful for understanding today and the future. They say the authors do an excellent job highlighting the fundamental changes that big data has made. Readers also mention the book helps review and rethink some traditional mindsets we have developed about data analysis. They say it's a good introduction to the scientific method of probe first and formulate. Overall, they describe it as an extremely important and interesting book that leads readers to see an exciting future.
""Big Data: A Revolution..." was often engaging and included some interesting examples, but it was a disappointment...." Read more
"...Mayer-Schonberger does a great job on identifying the key issues around Big Data and offering his opinion and insights on how we should move..." Read more
"...Written to a general audience, the authors do an excellent job highlighting the fundamental changes that big data has made on society, law, and the..." Read more
"...The book also does a great job of covering numerous examples and benefits, along with the limitations and dangers of Big Data...." Read more
-
Top reviews
Top reviews from the United States
There was a problem filtering reviews right now. Please try again later.
The authors start by discussing how Google using its analysis of people's queries is more predictive about flu epidemics than medical experts have been. The human genome can be codified in a fraction of the time that was required when it was being decoded for the first time. They discuss how big data has enabled entrepreneurs to inform customers about the optimal time to buy flight tickets given that airlines vary their prices according to hidden methods that big data statistics has helped to make more sense of. The examples are a good starting point to start the discussion with the reader. The authors start by discussing how we have always been trying to come up with data about our populations, desires to do census analysis has been with us for a long time. We made progress through sampling techniques and statistics helped to enable data gathering about the population at large using smaller and less time consuming samples. The authors discuss how big data is messy, it is imprecise and is helpful for overviews but not for model building with respect to figuring out the mechanics of what is being observed. When you try to get all of the data about something there will inevitably be noise and looking for correlations can sometimes be the most fruitful way to use the data to figure out empirical relationships rather than search for underlying dynamics. The authors discuss datification which means the consolidation of data into a larger database that can then be used to give much more useful guidance to the population at large about phenomenon that required a look from above at all the data together. Matthew Maury is used to reinforce the usefulness of this approach, he was a naval officer who aggregated ships logs to help inform ship captains about most useful routes and more efficient transiting. The authors move on to the more concrete and start to discuss the value of big data. They give the obvious background on the value of traditional data and then give food for thought on how having data for everything can lead to new ideas and utility that was unimaginable in the past. Big data analytics will be required for document translation, smart device coordination, smart cities and social network analysis. The value in big data is of course, the data, but the utility of that data might be further midstream or downstream that others are better placed to harvest. The authors move on to discuss the data value chain and how to think about it. The authors discuss the implication of the big data revolution and how it is enabling consumers to get the best deals and how statisticians are a highly desirable skill set. The authors move on to the risks of big data which are numerous of course. Much discussed are the privacy of the data that is generated. The ownership of that data and the licensing of it are topics which will continue to surface and the legal framework to analyze disputes will need to be further developed. Misunderstanding correlation and causation will also be a risk in big data analytics and hypotheticals like the government quarantining those who search for flu on google are used as hyperbolized examples. The authors finally leave the reader with a view on the future. They use an example of how big data statistics was used to substantially improve the ability to find overcrowded illegal slum housing as a concrete example of how we can use data to enhance our cities and improve governance and efficiency.
Big data is a subject which continues to step into more and more categories as our ability to measure continues to improve. How big data can be used will be a continued subject that both academics and practitioners will continue to be thought about and experimented on. It will give rise to a new consumer culture and potentially to new ways of organizing people and infrastructure. Big Data is an excellent readable overview of how data has always been used to guide policy, how big data is being used today, what the value chain of the data industry looks like, what the risks are of big data and how big data can enhance the future. Its easy to read and illuminating.
I encountered two huge issues in the text. First, the authors repeatedly argue that it's OK if Big Data contains "messy" data, because they assert that when "n=all" then the statistical rules about sampling don't apply. This argument fails two ways: first, if n=all but if the data contains "messy" (erroneous) data points in critical places, then it will be misleading and perhaps even completely wrong. Second, when using past data where "n=all" to project future events, then it's no longer true that "n=all." Instead, we have data for "n=all(where(time=past))" and we're using that data to try to predict events in a completely separate data set ("time=future"), and it's entirely possible that there are critical differences demarcated by "time=now."
The second huge issue, for me, was the authors' focus on the concept that Big Data brings with it a huge risk that we will use data to predict future behavior -- and that we will then use those predictions to punish people for acts they have not committed (e.g., the "Minority Report" problem). They distort this argument in two ways: first, by assuming that society would actually do this, and second, by asserting that any action taken based on these predictions (such as increasing scrutiny or assigning social workers to visit at-risk juveniles) is "punishment."
I was also skeptical of the authors' general reverence of, and deference to, data scientists as professionals and experts. The author believe that it's plausible to expect a new profession of internal and external "algorithmists" to arise, to protect consumers' privacy interests and society's interests against the potential abuses by Big Data users.
The book also failed to provide real-world "how-to" examples, instead providing only "end result" examples and conclusions that often seem incomplete and sometimes implausible. Their many useful examples of useful information extracted from Big Data all doubtless represent the end-point of many, many explorations of Big Data; they probably also represent a subset of correlations derived, after many misleading correlations were removed.
Finally, note that the book's lengthy end notes, bibliography, and index represent a full one-third of the book's length.
There's a lot of useful information in this book, especially for someone just trying to learn about the concept of Big Data. But there's also a lot of hype, and a lot of repetition of ideas without meaningful factual support.
If you are looking to understand what the revolution is all about this book explains it very well without going into too detail about the tools that are used to get there.






