Random the Book

Random the Book: Matt Ballantine and Nick Drage's experiment in serendipity and chance.


How could you get honest answers to sensitive questions?

Questions for you:

  • When collecting sensitive information (employee satisfaction with management, ethical violations, competitive intelligence), how could you tell if your current methods are failing because respondents fear identification and consequences?
  • Looking at survey data on controversial topics, do you account for social desirability bias, where people give “acceptable” answers rather than honest ones, thereby systematically skewing the results?
  • When privacy concerns prevent the collection of necessary information, could introducing calibrated randomness (e.g., randomised response) enable honest answers by creating plausible deniability?
  • In designing data collection systems, do you consider the trade-off between individual privacy and aggregate accuracy, or assume these goals are incompatible?

Organisational applications:

Randomised response for sensitive employee feedback: Traditional surveys about management effectiveness, ethical concerns, or workplace issues fail because employees fear retaliation despite claims of “anonymity”. Randomised responses creates mathematical privacy: instruct respondents to flip coin privately – heads means answer “yes” regardless of truth, tails means answer honestly. Anyone answering “yes” might be truthful or might have flipped heads, creating plausible deniability. Statisticians calculate true proportions knowing ~50% of “yes” answers are random. This enables honest feedback on harassment, discrimination, policy violations, or management problems whilst openly protecting individuals from exposure.

Differential privacy for user data analysis: Companies need aggregate insights about user behaviour without compromising individual privacy. Differential privacy (used by Apple, Google, Microsoft) adds calibrated random noise to individual responses before aggregation. This creates a mathematical guarantee that specific user data cannot be determined whilst preserving statistical properties for useful analysis. You could learn that 15% of users have a particular app installed without identifying which specific users. Apply this technique to product usage analytics, customer behaviour patterns, employee productivity metrics, and any situation requiring population-level insights without individual surveillance.

Plausible deniability in competitive intelligence: Gathering information about competitor strategies, employee dissatisfaction, or market problems faces resistance because sources fear exposure. Randomised techniques protect sources: when collecting sensitive market intelligence through surveys or interviews, introduce randomness to create uncertainty about whether a specific response came from a particular individual. This enables industry-wide data collection (salary benchmarks, strategic priorities, operational challenges) whilst protecting participants from competitive disadvantage or retaliation.

Calibrating privacy-accuracy trade-offs: Traditional thinking assumes privacy and accuracy are incompatible – more privacy means worse data. Randomised response techniques demonstrate this is a false dichotomy. Carefully calibrated randomness simultaneously protects individuals whilst enabling valid aggregate analysis. The key is mathematical precision: too much randomness destroys statistical accuracy, too little fails to protect privacy. Build systems with explicit privacy budgets, quantified noise levels, and validated statistical properties rather than vague “anonymous” claims that often prove false.

Further reading

Randomised response and survey methodology

Survey Sampling by Leslie Kish – foundational survey methodology text including discussion of randomised response techniques for sensitive questions, demonstrating how introducing randomness enables honest answers whilst protecting respondent privacy.

The Encyclopedia of Survey Research Methods by Paul Lavrakas (Editor) – comprehensive survey research reference including chapters on privacy-preserving techniques and methods for reducing response bias in sensitive domains.

Differential privacy and privacy-preserving data analysis

The Algorithmic Foundations of Differential Privacy by Cynthia Dwork and Aaron Roth – a technical introduction to differential privacy showing how adding calibrated noise to data enables useful aggregate analysis whilst providing mathematical privacy guarantees (available free online).

Privacy’s Blueprint by Woodrow Hartzog – examines privacy-preserving technologies including differential privacy, showing how design choices can protect individuals whilst enabling beneficial data use.

Weapons of Math Destruction by Cathy O’Neil – whilst primarily critiquing algorithmic harm, discusses importance of privacy-preserving techniques and consequences of inadequate privacy protection in data systems.

Privacy, anonymity, and data ethics

Obfuscation by Finn Brunton and Helen Nissenbaum – examines how adding noise and randomness to data protects privacy, showing randomised response as example of deliberate obfuscation serving beneficial purpose.

Data and Goliath by Bruce Schneier – examines surveillance and privacy showing why “anonymous” data often isn’t, discussing technical approaches like differential privacy that provide genuine mathematical privacy guarantees.

Understanding Privacy by Daniel J. Solove – explores privacy theory and practice including discussion of how technical mechanisms like randomisation can protect privacy whilst enabling necessary information flows.

About the image

Who knew that clipboards came in families?

Photo montage by Matt Ballantine, Photo by Dee Mamora , 2026