Enjoy fast, free delivery, exclusive deals, and award-winning movies & TV shows with Prime
Try Prime
and start saving today with fast, free delivery
Amazon Prime includes:
Fast, FREE Delivery is available to Prime members. To join, select "Try Amazon Prime and start saving today with Fast, FREE Delivery" below the Add to Cart button.
Amazon Prime members enjoy:- Cardmembers earn 5% Back at Amazon.com with a Prime Credit Card.
- Unlimited Free Two-Day Delivery
- Streaming of thousands of movies and TV shows with limited ads on Prime Video.
- A Kindle book to borrow for free each month - with no due dates
- Listen to over 2 million songs and hundreds of playlists
- Unlimited photo storage with anywhere access
Important: Your credit card will NOT be charged when you start your free trial or if you cancel during the trial period. If you're happy with Amazon Prime, do nothing. At the end of the free trial, your membership will automatically upgrade to a monthly membership.
Buy new:
$21.49$21.49
FREE delivery: Tuesday, Feb 13 on orders over $35.00 shipped by Amazon.
Ships from: Amazon.com Sold by: Amazon.com
Buy used: $15.49
Other Sellers on Amazon
+ $3.99 shipping
96% positive over last 12 months
Download the free Kindle app and start reading Kindle books instantly on your smartphone, tablet, or computer - no Kindle device required.
Read instantly on your browser with Kindle for Web.
Using your mobile phone camera - scan the code below and download the Kindle app.
Follow the authors
OK
Algorithms to Live By: The Computer Science of Human Decisions Hardcover – April 19, 2016
Explore your book, then jump right back to where you left off with Page Flip.
View high quality images that let you zoom in to take a closer look.
Enjoy features only possible in digital – start reading right away, carry your library with you, adjust the font, create shareable notes and highlights, and more.
Discover additional details about the events, people, and places in your book, with Wikipedia integration.
Purchase options and add-ons
An exploration of how computer algorithms can be applied to our everyday lives to solve common decision-making problems and illuminate the workings of the human mind.
What should we do, or leave undone, in a day or a lifetime? How much messiness should we accept? What balance of the new and familiar is the most fulfilling? These may seem like uniquely human quandaries, but they are not. Computers, like us, confront limited space and time, so computer scientists have been grappling with similar problems for decades. And the solutions they’ve found have much to teach us.
In a dazzlingly interdisciplinary work, Brian Christian and Tom Griffiths show how algorithms developed for computers also untangle very human questions. They explain how to have better hunches and when to leave things to chance, how to deal with overwhelming choices and how best to connect with others. From finding a spouse to finding a parking spot, from organizing one’s inbox to peering into the future, Algorithms to Live By transforms the wisdom of computer science into strategies for human living.
Review
“A remarkable book... A solid, research-based book that’s applicable to real life. The algorithms the authors discuss are, in fact, more applicable to real-life problems than I’d have ever predicted.... It’s well worth the time to find a copy of Algorithms to Live By and dig deeper.”
―Forbes
“By the end of the book, I was convinced. Not because I endorse the idea of living like some hyper-rational Vulcan, but because computing algorithms could be a surprisingly useful way to embrace the messy compromises of real, non-Vulcan life.”
―The Guardian (UK)
“I absolutely reveled in this book... It's the perfect antidote to the argument you often hear from young math students: ‘What's the point? I'll never use this in real life!’... The whole business, whether it's the relative simplicity of the 37% rule or the mind-twisting possibilities of game theory, is both potentially practical and highly enjoyable as presented here. Recommended.”
―Popular Science (UK)
“An entertaining, intelligently presented book... Craftily programmed to build from one good idea to the next... The value of being aware of algorithmic thinking―of the thornier details of ‘human algorithm design,’ as Christian and Griffiths put it―is not just better problem solving, but also greater insight into the human mind. And who doesn’t want to know how we tick?”
―Kirkus Reviews
“Compelling and entertaining, Algorithms to Live By is packed with practical advice about how to use time, space, and effort more efficiently. And it’s a fascinating exploration of the workings of computer science and the human mind. Whether you want to optimize your to-do list, organize your closet, or understand human memory, this is a great read.”
―Charles Duhigg, author of The Power of Habit
“In this remarkably lucid, fascinating, and compulsively readable book, Christian and Griffiths show how much we can learn from computers. We’ve all heard about the power of algorithms―but Algorithms to Live By actually explains, brilliantly, how they work, and how we can take advantage of them to make better decisions in our own lives.”
―Alison Gopnik, coauthor of The Scientist in the Crib
“I’ve been waiting for a book to come along that merges computational models with human psychology―and Christian and Griffiths have succeeded beyond all expectations. This is a wonderful book, written so that anyone can understand the computer science that runs our world―and more importantly, what it means to our lives.”
―David Eagleman, author of Incognito: The Secret Lives of the Brain
About the Author
Tom Griffiths is a professor of psychology and cognitive science at UC Berkeley, where he directs the Computational Cognitive Science Lab. He has received widespread recognition for his scientific work, including awards from the American Psychological Association and the Sloan Foundation.
Excerpt. © Reprinted by permission. All rights reserved.
Algorithms to Live By
The Computer Science of Human Decisions
By Brian Christian, Tom GriffithsHenry Holt and Company
Copyright © 2016 Brian Christian and Tom GriffithsAll rights reserved.
ISBN: 978-1-62779-036-9
Contents
Title Page,Copyright Notice,
Dedication,
Introduction,
Algorithms to Live By,
1 Optimal Stopping Optimal Stopping When to Stop Looking,
2 Explore/Exploit The Latest vs. the Greatest,
3 Sorting Making Order,
4 Caching Forget About It,
5 Scheduling First Things First,
6 Bayes's Rule Predicting the Future,
7 Overfitting When to Think Less,
8 Relaxation Let It Slide,
9 Randomness When to Leave It to Chance,
10 Networking How We Connect,
11 Game Theory The Minds of Others,
Conclusion,
Computational Kindness,
Notes,
Bibliography,
Index,
Acknowledgments,
Also by Brian Christian,
About the Authors,
Copyright,
CHAPTER 1
Optimal Stopping
When to Stop Looking
Though all Christians start a wedding invitation by solemnly declaring their marriage is due to special Divine arrangement, I, as a philosopher, would like to talk in greater detail about this ... — JOHANNES KEPLER
If you prefer Mr. Martin to every other person; if you think him the most agreeable man you have ever been in company with, why should you hesitate? — JANE AUSTEN, EMMA
It's such a common phenomenon that college guidance counselors even have a slang term for it: the "turkey drop." High-school sweethearts come home for Thanksgiving of their freshman year of college and, four days later, return to campus single.
An angst-ridden Brian went to his own college guidance counselor his freshman year. His high-school girlfriend had gone to a different college several states away, and they struggled with the distance. They also struggled with a stranger and more philosophical question: how good a relationship did they have? They had no real benchmark of other relationships by which to judge it. Brian's counselor recognized theirs as a classic freshman-year dilemma, and was surprisingly nonchalant in her advice: "Gather data."
The nature of serial monogamy, writ large, is that its practitioners are confronted with a fundamental, unavoidable problem. When have you met enough people to know who your best match is? And what if acquiring the data costs you that very match? It seems the ultimate Catch-22 of the heart.
As we have seen, this Catch-22, this angsty freshman cri de coeur, is what mathematicians call an "optimal stopping" problem, and it may actually have an answer: 37%.
Of course, it all depends on the assumptions you're willing to make about love.
The Secretary Problem
In any optimal stopping problem, the crucial dilemma is not which option to pick, but how many options to even consider. These problems turn out to have implications not only for lovers and renters, but also for drivers, homeowners, burglars, and beyond.
The 37% Rule derives from optimal stopping's most famous puzzle, which has come to be known as the "secretary problem." Its setup is much like the apartment hunter's dilemma that we considered earlier. Imagine you're interviewing a set of applicants for a position as a secretary, and your goal is to maximize the chance of hiring the single best applicant in the pool. While you have no idea how to assign scores to individual applicants, you can easily judge which one you prefer. (A mathematician might say you have access only to the ordinal numbers — the relative ranks of the applicants compared to each other — but not to the cardinal numbers, their ratings on some kind of general scale.) You interview the applicants in random order, one at a time. You can decide to offer the job to an applicant at any point and they are guaranteed to accept, terminating the search. But if you pass over an applicant, deciding not to hire them, they are gone forever.
The secretary problem is widely considered to have made its first appearance in print — sans explicit mention of secretaries — in the February 1960 issue of Scientific American, as one of several puzzles posed in Martin Gardner's beloved column on recreational mathematics. But the origins of the problem are surprisingly mysterious. Our own initial search yielded little but speculation, before turning into unexpectedly physical detective work: a road trip down to the archive of Gardner's papers at Stanford, to haul out boxes of his midcentury correspondence. Reading paper correspondence is a bit like eavesdropping on someone who's on the phone: you're only hearing one side of the exchange, and must infer the other. In our case, we only had the replies to what was apparently Gardner's own search for the problem's origins fiftysome years ago. The more we read, the more tangled and unclear the story became.
Harvard mathematician Frederick Mosteller recalled hearing about the problem in 1955 from his colleague Andrew Gleason, who had heard about it from somebody else. Leo Moser wrote from the University of Alberta to say that he read about the problem in "some notes" by R. E. Gaskell of Boeing, who himself credited a colleague. Roger Pinkham of Rutgers wrote that he first heard of the problem in 1955 from Duke University mathematician J. Shoenfield, "and I believe he said that he had heard the problem from someone at Michigan."
"Someone at Michigan" was almost certainly someone named Merrill Flood. Though he is largely unheard of outside mathematics, Flood's influence on computer science is almost impossible to avoid. He's credited with popularizing the traveling salesman problem (which we discuss in more detail in chapter 8), devising the prisoner's dilemma (which we discuss in chapter 11), and even with possibly coining the term "software." It's Flood who made the first known discovery of the 37% Rule, in 1958, and he claims to have been considering the problem since 1949 — but he himself points back to several other mathematicians.
Suffice it to say that wherever it came from, the secretary problem proved to be a near-perfect mathematical puzzle: simple to explain, devilish to solve, succinct in its answer, and intriguing in its implications. As a result, it moved like wildfire through the mathematical circles of the 1950s, spreading by word of mouth, and thanks to Gardner's column in 1960 came to grip the imagination of the public at large. By the 1980s the problem and its variations had produced so much analysis that it had come to be discussed in papers as a subfield unto itself.
As for secretaries — it's charming to watch each culture put its own anthropological spin on formal systems. We think of chess, for instance, as medieval European in its imagery, but in fact its origins are in eighth-century India; it was heavy-handedly "Europeanized" in the fifteenth century, as its shahs became kings, its viziers turned to queens, and its elephants became bishops. Likewise, optimal stopping problems have had a number of incarnations, each reflecting the predominating concerns of its time. In the nineteenth century such problems were typified by baroque lotteries and by women choosing male suitors; in the early twentieth century by holidaying motorists searching for hotels and by male suitors choosing women; and in the paper-pushing, male-dominated mid-twentieth century, by male bosses choosing female assistants. The first explicit mention of it by name as the "secretary problem" appears to be in a 1964 paper, and somewhere along the way the name stuck.
Whence 37%?
In your search for a secretary, there are two ways you can fail: stopping early and stopping late. When you stop too early, you leave the best applicant undiscovered. When you stop too late, you hold out for a better applicant who doesn't exist. The optimal strategy will clearly require finding the right balance between the two, walking the tightrope between looking too much and not enough.
If your aim is finding the very best applicant, settling for nothing less, it's clear that as you go through the interview process you shouldn't even consider hiring somebody who isn't the best you've seen so far. However, simply being the best yet isn't enough for an offer; the very first applicant, for example, will of course be the best yet by definition. More generally, it stands to reason that the rate at which we encounter "best yet" applicants will go down as we proceed in our interviews. For instance, the second applicant has a 50/50 chance of being the best we've yet seen, but the fifth applicant only has a 1-in-5 chance of being the best so far, the sixth has a 1-in-6 chance, and so on. As a result, best-yet applicants will become steadily more impressive as the search continues (by definition, again, they're better than all those who came before) — but they will also become more and more infrequent.
Okay, so we know that taking the first best-yet applicant we encounter (a.k.a. the first applicant, period) is rash. If there are a hundred applicants, it also seems hasty to make an offer to the next one who's best-yet, just because she was better than the first. So how do we proceed?
Intuitively, there are a few potential strategies. For instance, making an offer the third time an applicant trumps everyone seen so far — or maybe the fourth time. Or perhaps taking the next best-yet applicant to come along after a long "drought" — a long streak of poor ones.
But as it happens, neither of these relatively sensible strategies comes out on top. Instead, the optimal solution takes the form of what we'll call the Look-Then-Leap Rule: You set a predetermined amount of time for "looking" — that is, exploring your options, gathering data — in which you categorically don't choose anyone, no matter how impressive. After that point, you enter the "leap" phase, prepared to instantly commit to anyone who outshines the best applicant you saw in the look phase.
We can see how the Look-Then-Leap Rule emerges by considering how the secretary problem plays out in the smallest applicant pools. With just one applicant the problem is easy to solve — hire her! With two applicants, you have a 50/50 chance of success no matter what you do. You can hire the first applicant (who'll turn out to be the best half the time), or dismiss the first and by default hire the second (who is also best half the time).
Add a third applicant, and all of a sudden things get interesting. The odds if we hire at random are one-third, or 33%. With two applicants we could do no better than chance; with three, can we? It turns out we can, and it all comes down to what we do with the second interviewee. When we see the first applicant, we have no information — she'll always appear to be the best yet. When we see the third applicant, we have no agency — we have to make an offer to the final applicant, since we've dismissed the others. But when we see the second applicant, we have a little bit of both: we know whether she's better or worse than the first, and we have the freedom to either hire or dismiss her. What happens when we just hire her if she's better than the first applicant, and dismiss her if she's not? This turns out to be the best possible strategy when facing three applicants; using this approach it's possible, surprisingly, to do just as well in the three-applicant problem as with two, choosing the best applicant exactly half the time.
Enumerating these scenarios for four applicants tells us that we should still begin to leap as soon as the second applicant; with five applicants in the pool, we shouldn't leap before the third.
As the applicant pool grows, the exact place to draw the line between looking and leaping settles to 37% of the pool, yielding the 37% Rule: look at the first 37% of the applicants, choosing none, then be ready to leap for anyone better than all those you've seen so far.
As it turns out, following this optimal strategy ultimately gives us a 37% chance of hiring the best applicant; it's one of the problem's curious mathematical symmetries that the strategy itself and its chance of success work out to the very same number. The table above shows the optimal strategy for the secretary problem with different numbers of applicants, demonstrating how the chance of success — like the point to switch from looking to leaping — converges on 37% as the number of applicants increases.
A 63% failure rate, when following the best possible strategy, is a sobering fact. Even when we act optimally in the secretary problem, we will still fail most of the time — that is, we won't end up with the single best applicant in the pool. This is bad news for those of us who would frame romance as a search for "the one." But here's the silver lining. Intuition would suggest that our chances of picking the single best applicant should steadily decrease as the applicant pool grows. If we were hiring at random, for instance, then in a pool of a hundred applicants we'd have a 1% chance of success, and in a pool of a million applicants we'd have a 0.0001% chance. Yet remarkably, the math of the secretary problem doesn't change. If you're stopping optimally, your chance of finding the single best applicant in a pool of a hundred is 37%. And in a pool of a million, believe it or not, your chance is still 37%. Thus the bigger the applicant pool gets, the more valuable knowing the optimal algorithm becomes. It's true that you're unlikely to find the needle the majority of the time, but optimal stopping is your best defense against the haystack, no matter how large.
Lover's Leap
The passion between the sexes has appeared in every age to be so nearly the same that it may always be considered, in algebraic language, as a given quantity. — THOMAS MALTHUS
I married the first man I ever kissed. When I tell this to my children they just about throw up. — BARBARA BUSH
Before he became a professor of operations research at Carnegie Mellon, Michael Trick was a graduate student, looking for love. "It hit me that the problem has been studied: it is the Secretary Problem! I had a position to fill [and] a series of applicants, and my goal was to pick the best applicant for the position." So he ran the numbers. He didn't know how many women he could expect to meet in his lifetime, but there's a certain flexibility in the 37% Rule: it can be applied to either the number of applicants or the time over which one is searching. Assuming that his search would run from ages eighteen to forty, the 37% Rule gave age 26.1 years as the point at which to switch from looking to leaping. A number that, as it happened, was exactly Trick's age at the time. So when he found a woman who was a better match than all those he had dated so far, he knew exactly what to do. He leapt. "I didn't know if she was Perfect (the assumptions of the model don't allow me to determine that), but there was no doubt that she met the qualifications for this step of the algorithm. So I proposed," he writes.
"And she turned me down."
Mathematicians have been having trouble with love since at least the seventeenth century. The legendary astronomer Johannes Kepler is today perhaps best remembered for discovering that planetary orbits are elliptical and for being a crucial part of the "Copernican Revolution" that included Galileo and Newton and upended humanity's sense of its place in the heavens. But Kepler had terrestrial concerns, too. After the death of his first wife in 1611, Kepler embarked on a long and arduous quest to remarry, ultimately courting a total of eleven women. Of the first four, Kepler liked the fourth the best ("because of her tall build and athletic body") but did not cease his search. "It would have been settled," Kepler wrote, "had not both love and reason forced a fifth woman on me. This one won me over with love, humble loyalty, economy of household, diligence, and the love she gave the stepchildren."
"However," he wrote, "I continued."
Kepler's friends and relations went on making introductions for him, and he kept on looking, but halfheartedly. His thoughts remained with number five. After eleven courtships in total, he decided he would search no further. "While preparing to travel to Regensburg, I returned to the fifth woman, declared myself, and was accepted." Kepler and Susanna Reuttinger were wed and had six children together, along with the children from Kepler's first marriage. Biographies describe the rest of Kepler's domestic life as a particularly peaceful and joyous time.
Both Kepler and Trick — in opposite ways — experienced firsthand some of the ways that the secretary problem oversimplifies the search for love. In the classical secretary problem, applicants always accept the position, preventing the rejection experienced by Trick. And they cannot be "recalled" once passed over, contrary to the strategy followed by Kepler.
In the decades since the secretary problem was first introduced, a wide range of variants on the scenario have been studied, with strategies for optimal stopping worked out under a number of different conditions. The possibility of rejection, for instance, has a straightforward mathematical solution: propose early and often. If you have, say, a 50/50 chance of being rejected, then the same kind of mathematical analysis that yielded the 37% Rule says you should start making offers after just a quarter of your search. If turned down, keep making offers to every best-yet person you see until somebody accepts. With such a strategy, your chance of overall success — that is, proposing and being accepted by the best applicant in the pool — will also be 25%. Not such terrible odds, perhaps, for a scenario that combines the obstacle of rejection with the general difficulty of establishing one's standards in the first place.
(Continues...)Excerpted from Algorithms to Live By by Brian Christian, Tom Griffiths. Copyright © 2016 Brian Christian and Tom Griffiths. Excerpted by permission of Henry Holt and Company.
All rights reserved. No part of this excerpt may be reproduced or reprinted without permission in writing from the publisher.
Excerpts are provided by Dial-A-Book Inc. solely for the personal use of visitors to this web site.
- Print length368 pages
- LanguageEnglish
- PublisherHenry Holt and Co.
- Publication dateApril 19, 2016
- Dimensions6.55 x 1.15 x 9.55 inches
- ISBN-101627790365
- ISBN-13978-1627790369
Frequently bought together

Similar items that may ship from close to you
Exploration in itself has value, since trying new things increases our chances of finding the best. So taking the future into account, rather than focusing just on the present, drives us toward novelty.Highlighted by 3,053 Kindle readers
“To try and fail is at least to learn; to fail to try is to suffer the inestimable loss of what might have been.”Highlighted by 2,734 Kindle readers
This is the first and most fundamental insight of sorting theory. Scale hurts.Highlighted by 1,936 Kindle readers
Product details
- Publisher : Henry Holt and Co.; 1st edition (April 19, 2016)
- Language : English
- Hardcover : 368 pages
- ISBN-10 : 1627790365
- ISBN-13 : 978-1627790369
- Item Weight : 2.31 pounds
- Dimensions : 6.55 x 1.15 x 9.55 inches
- Best Sellers Rank: #77,967 in Books (See Top 100 in Books)
- #132 in Business Decision Making
- #259 in Decision-Making & Problem Solving
- #285 in Cognitive Psychology (Books)
- Customer Reviews:
Important information
To report an issue with this product or seller, click here.
About the authors

Brian Christian is the author of the acclaimed bestsellers "The Most Human Human," a New York Times editors’ choice and a New Yorker favorite book of the year, and "Algorithms to Live By" (with Tom Griffiths), a #1 Audible bestseller, Amazon best science book of the year and MIT Technology Review best book of the year.
Christian’s writing has appeared in The New Yorker, The Atlantic, Wired, and The Wall Street Journal, as well as peer-reviewed journals such as Cognitive Science. He has been featured on The Daily Show and Radiolab, and has lectured at Google, Facebook, Microsoft, the Santa Fe Institute, and the London School of Economics. His work has won several awards, including publication in Best American Science & Nature Writing, and has been translated into nineteen languages.
Christian holds degrees in computer science, philosophy, and poetry from Brown University and the University of Washington. A Visiting Scholar at the University of California, Berkeley, he lives in San Francisco.

Tom Griffiths is a professor of psychology and computer science at Princeton, where he directs the Computational Cognitive Science Lab. He has published scientific papers on topics ranging from cognitive psychology to cultural evolution, and has received awards from the National Academy of Sciences, the Sloan Foundation, the American Psychological Association, and the Psychonomic Society, among others. He lives in Princeton, New Jersey.
Customer reviews
Customer Reviews, including Product Star Ratings help customers to learn more about the product and decide whether it is the right product for them.
To calculate the overall star rating and percentage breakdown by star, we don’t use a simple average. Instead, our system considers things like how recent a review is and if the reviewer bought the item on Amazon. It also analyzed reviews to verify trustworthiness.
Learn more how customers reviews work on AmazonReviews with images
-
Top reviews
Top reviews from the United States
There was a problem filtering reviews right now. Please try again later.
Optimal stopping - how many people out of 100 possible candidates should one interview for a given position (including that of spouse)? 37%, Why? Read the book.
The Explore/Exploit dichotomy - Should one ask the question "What's new" or "What's best"? Your answer may depend on your time horizon. As your time horizon shortens, "what's best" may be the better question. The book explains why. The book also looks at the multi-armed bandit as an example of the explore/exploit dichotomy. What's a multi-armed bandit? Think of the one-armed bandit in Vegas and multiply its arms. Mathematicians do so. Their conclusions may be useful. The trials of music critics also fit into the explore/exploit dichotomy. The authors explain why music critics find exploration a chore.
Sorting - libraries are the metaphor for computer sorting. Human memory also requires sorting. Maybe the decline in memory as humans age may be due to the amount of information through which it must sort and not due to declining faculties. A five-year old has a lot less information to go through than a seventy-five year old. The authors consider sorting techniques with email, Yelp, and other common uses. There is much useful information.
Caching - when is forgetting necessary? According to the authors, the first computer cache was developed for a supercomputer in 1962 ub Manchester, England. I wonder how "super" that computer was? Caching allows some information to be stored for repetitive use and uncached information to be kept in the background.
Scheduling - many scheduling problems have "intractable" solutions. The authors suggest different solutions based on algorithms such as precedence constraints, earliest due date (one I personally use frequently, which I couple with a personal likely to get me in the most trouble the quickest test) and shortest processing time. The scheduling problem has received substantial effort from mathematicians.
Bayes's Rule - how to use statistical inference to make useful predictions. Couple a well-defined problem with a range of prior outcomes and one can make accurate guesses. A .300 hitter comes to the plate against the same pitcher who has already struck the batter out twice and it may be a fair guess that the hitter is due for a hit.
Overfitting - don't overthink and over complicate a problem. The authors advise against practicing the idolatry of data. A more complex theorem may well lead to less accuracy rather than more. On the level of incentive compensation, the authors quote Steve Jobs for being careful that you include only those elements in your incentive package that matter; you will get what you measure.
Relaxatrion - the perfect is the enemy of the good. To get any useful answer from your mathematical model, it may be necessary to relax some of your constraints (insisting that your model never allow the traveling salesman to re-enter the same city twice may preclude any answer at all in a time period of less than the remaining life of the universe).
Randomness - mathematicians sometimes realize that the best answer comes from sampling and not from strict calculations. This may explain why I get so many survey requests. Algorithms for prime numbers use this technique. And, apparently, thousands of years ago the Greeks were already looking for prime numbers.
Networking - here the authors examine the "Byzantine generals" problem, which plays a part in explaining how computers communicate with each other.
Game Theory - Alan Turing investigated the "halting problem" in the 1930s. What if you give your computer a problem and it just keeps going? Rock, paper, scissors is a game with which most are familiar. It, too, is part of game theory. When a game seems to have no satisfactory answer, maybe it's time to change the game. What happens when you have an "information cascade"?
If any ot this interests you, I believe that you will enjoy the book. I recommend it highly.
As aforementioned, the book explores how people use algorithms in their day-to-day to accomplish tasks. They focus on several elements: explore/exploit, or when it is best to continue to look for something better or make a choice from what one already knows; sorting and tradeoffs; and scheduling being among the subjects of focus. What makes these sections interesting is that they often talk about tradeoffs that one would seem counterintuitive. An example of this is in the scheduling section. The authors mention how the placement of a task on a schedule may be influenced by how much one knows about the task: by its duration or difficulty. This may increase the difficulty of scheduling if one were to know every detail of every task that must be done for the day. They also mention that while some may be tempted to schedule tasks based on how easy they are, this may also come with downsides. Especially if one decides to prioritize harder tasks before easier ones, only to realize that its completion requires completing an easier task. They give an example of a NASA Mars rover being frozen due to this fact. The rover was programmed to prioritize high priority tasks first in its queue over low priority tasks. However, one of the low-priority tasks kept being pulled from the bottom of the queue to the top. This caused the rover to freeze. Thus, even well-thought-out systems can lead to problems.
The above example with NASA shows another aspect of the book I like; the use of real world examples. The authors tell stories involving real world mathematicians and scientists struggling with these issues in their personal lives. This helps make the subjects feel personal and applicable to one's own life. In fact, I would argue that the only issue with the book is that these anecdotes seem to be an afterthought. This is due to the fact that the anecdotes become more prominent as the book progresses towards the end. Thus, the first few chapters can be somewhat dry in its presentation which may turn off a lay reader. Furthermore, the use of hypothetical scenarios in the earlier chapters feel like a pale imitation of the personal anecdotes of later chapters.
All in all, this book was fairly enjoyable. While having some rough patches, the authors did try and succeed in making an accessible book.
















