Rubber-hose cryptanalysis
In cryptography, rubber-hose cryptanalysis is a euphemism for the extraction of cryptographic secrets (e.g. the password to an encrypted file) from a person by coercion or torture[1][2]—such as beating that person with a rubber hose, thus the name—in contrast to a mathematical or technical cryptanalytic attack.
Details
According to Amnesty International and the UN, many countries in the world routinely torture people.[3][4][5][6] It is therefore logical to assume that at least some of those countries use (or would be willing to use) some form of rubber-hose cryptanalysis.[1] In practice, psychological coercion can prove as effective as physical torture. Not physically violent but highly intimidating methods include such tactics as the threat of harsh legal penalties. The incentive to cooperate may be some form of plea bargain, such as an offer to drop or reduce criminal charges against a suspect in return for full co-operation with investigators. Alternatively, in some countries threats may be made to prosecute as co-conspirators (or inflict violence upon) close relatives (e.g. spouse, children, or parents) of the person being questioned unless they co-operate.[4][7]
In some contexts, rubber-hose cryptanalysis may not be a viable attack because of a need to decrypt data covertly; information such as a password may lose its value if it is known to have been compromised. It has been argued that one of the purposes of strong cryptography is to force adversaries to resort to less covert attacks.[8]
The earliest known use of the term was on the sci.crypt newsgroup, in a message posted 16 October 1990 by Marcus J. Ranum, alluding to corporal punishment:
...the rubber-hose technique of cryptanalysis. (in which a rubber hose is applied forcefully and frequently to the soles of the feet until the key to the cryptosystem is discovered, a process that can take a surprisingly short time and is quite computationally inexpensive).— [9]
Although the term is used tongue-in-cheek, its implications are serious: in modern cryptosystems, the weakest link is often the human user.[10] A direct attack on a cipher algorithm, or the cryptographic protocols used, is likely to be much more expensive and difficult than targeting the people who use or manage the system. Thus, many cryptosystems and security systems are designed with special emphasis on keeping human vulnerability to a minimum. For example, in public-key cryptography, the defender may hold the key to encrypt the message, but not the decryption key needed to decipher it. The problem here is that the defender may be unable to convince the attacker to stop coercion. In plausibly deniable encryption, a second key is created which unlocks a second convincing but relatively harmless message (for example, apparently personal writings expressing "deviant" thoughts or desires of some type that are lawful but taboo), so the defender can prove to have handed over the keys whilst the attacker remains unaware of the primary hidden message. In this case, the designer's expectation is that the attacker will not realize this, and forego threats or actual torture. See also steganography. The risk, however, is that the attacker may be aware of deniable encryption and will assume the defender knows more than one key, meaning the attacker may refuse to stop coercing the defender even if one or more keys are revealed: on the assumption the defender is still withholding additional keys which hold additional information.
In law
In some jurisdictions, statutes assume the opposite—that human operators know (or have access to) such things as session keys, an assumption which parallels that made by rubber-hose practitioners. An example is the United Kingdom's Regulation of Investigatory Powers Act,[11][12] which makes it a crime to not surrender encryption keys on demand from a government official authorized by the act.
According to the Home Office, the burden of proof that an accused person is in possession of a key rests on the prosecution; moreover, the act contains a defence for operators who have lost or forgotten a key, and they are not liable if they are judged to have done what they can to recover a key.[11][12]
See also
|
References
- 1 2 Soghoian, Chris (October 24, 2008). "Turkish police may have beaten encryption key out of TJ Maxx suspect". Surveillance State. CNET Networks. Retrieved August 29, 2009.
- ↑ Schneier, Bruce (October 27, 2008). "Rubber-Hose Cryptanalysis". Schneier on Security. Retrieved August 29, 2009.
- ↑ Pincock, Stephen (November 1, 2003). "Exposing the horror of torture". The Lancet. 362 (9394): 1462–1463. doi:10.1016/S0140-6736(03)14730-7. PMID 14603923. Retrieved August 29, 2009.
- 1 2 "Many countries still appear willing to use torture, warns UN human rights official" (Press release). UN News Service. October 27, 2004. Retrieved August 28, 2009.
- ↑ Modvig, J.; Pagaduan-Lopez, J.; Rodenburg, J.; Salud, CMD; Cabigon, RV; Panelo, CIA (November 18, 2000). "Torture and trauma in post-conflict East Timor". The Lancet. 356 (9243): 1763. doi:10.1016/S0140-6736(00)03218-9. PMID 11095275. Archived from the original on August 27, 2005. Retrieved August 29, 2009.
- ↑ Iacopino, Vincent (November 30, 1996). "Turkish physicians coerced to conceal systematic torture". The Lancet. 348 (9040): 1500. doi:10.1016/S0140-6736(05)65892-8. PMID 11654536. Retrieved August 29, 2009.
- ↑ Hoffman, Russell D. (February 2, 1996). "Interview with author of PGP (Pretty Good Privacy)". High Tech Today. Retrieved August 29, 2009.
- ↑ Percival, Colin (May 13, 2010). "Everything you need to know about cryptography in 1 hour (conference slides)" (PDF). Retrieved December 29, 2011.
- ↑ Ranum, Marcus J. (October 16, 1990). "Re: Cryptography and the Law...". Newsgroup: sci.crypt. Usenet: 1990Oct16.050000.4965@decuac.dec.com. Retrieved October 11, 2013.
- ↑ "The Weakest Link: The Human Factor Lessons Learned from the German WWII Enigma Cryptosystem". SANS. Retrieved 6 June 2013.
- 1 2 "The RIP Act". The Guardian. London. October 25, 2001.
- 1 2 "Regulation of Investigatory Powers Bill; in Session 1999-2000, Internet Publications, Other Bills before Parliament". House of Lords. 9 May 2000. Retrieved 5 Jan 2011.