AustLII Home | Databases | WorldLII | Search | Feedback

University of New South Wales Faculty of Law Research Series

Faculty of Law, UNSW
You are here:  AustLII >> Databases >> University of New South Wales Faculty of Law Research Series >> 2009 >> [2009] UNSWLRS 34

Database Search | Name Search | Recent Articles | Noteup | LawCite | Author Info | Download | Help

de Villiers, Meiring --- "Information Security Standards" [2009] UNSWLRS 34

Last Updated: 4 October 2009


Information Security Standards


Meiring de Villiers, University of New South Wales


Citation

This paper has been accepted for publication by the Journal of Internet Law.


Abstract

Businesses, non-profit organizations and government agencies may be held liable for failure to safeguard sensitive information in their possession. The threat of liability creates incentives to improve security standards, but uncertainty about the required standard and its judicial application may result in under- or overcompliance. Perfect security is neither possible nor the goal of tort law, but where does the law draw the line? This article analyzes the legal standard of information security that must be achieved to avoid liability. A numerical example illustrates its implementation.

1. INTRODUCTION

Information is the lifeblood of modern society. Businesses, non-profit organizations and government agencies regularly compile and maintain electronic databases of information about individuals who interact with these institutions. Computerized data include contact information, personal histories, financial records and official identifiers such as social security numbers. This wealth of information allows business and government to operate more efficiently, but also exposes the persons to whom the information relates to risks such as identity theft, monetary losses, loss of intellectual property, loss of privacy and reputation, stalking and blackmail.[1]
Database owners may be held liable for failure to safeguard the confidentiality, integrity, and availability of sensitive information in their possession. The threat of liability creates incentives to improve security standards, but uncertainty about the required standard and its judicial application may result in under- or overcompliance.[2] Perfect security is neither possible[3] nor the goal of tort law,[4] but where does the law draw the line? This article analyzes the legal standard of care in information security.
A victim of an information security breach may pursue a civil action under a negligence theory, the most widely used theory of liability in the law of torts. Negligence is generally defined as the failure to take reasonable precautions against a foreseeable risk of harm to another. Negligence may consist of an act, such as careless disposal of confidential customer information in a public place, or an omission, such as failing to remedy a known security vulnerability in a computer system.
The plaintiff defines the standard of care in negligence law, by identifying and pleading an untaken precaution. The plaintiff must then prove that the precaution is cost-effective, and that the defendant's failure to take the precaution was the actual and proximate cause of real harm to the plaintiff. The untaken precaution forms the basis of the plaintiff's case, and defines the standard of care that the defendant, the court hearing the case, and perhaps a subsequent appellate court will use.
The due care standard in information security is an issue of law, economics and technology. The untaken precaution must have been technically capable of preventing the plaintiff's harm, if taken. The plaintiff typically selects among precautions such as a better firewall, an intrusion detection system or a virus scanner. The cost-effectiveness of the selected precaution is determined by balancing the benefits it provides in risk reduction against its cost. The sociology of human-machine interaction also plays a role in the liability analysis. The article argues that a defendant may be held liable for negligently enabling a cyber crime if the criminal can be characterized as a free radical. Free radicals, in this context, are individuals who are not deterred by the threat of tort or criminal liability, because they are shielded from liability by factors such as anonymity, insufficient assets, lack of mental capacity or lack of good judgment. Research shows that cyber wrongdoers generally fit the profile of a free radical.
The article is organized as follows. Section 2 introduces the principles of computer viruses and worms, the weapons of choice of cyber wrongdoers. It also discusses the most common virus detection technologies as potential untaken precautions. Section 3 analyzes the standard of due care in a negligence cause of action, and shows how it is determined in an information security environment. A numerical example illustrates the cost-benefit calculus of an untaken precaution. A final section concludes.


2. PRINCIPLES OF MALEVOLENT SOFTWARE

Malevolent software is a term for computer code that is designed to disrupt the operation of a computer system. The most common of these rogue programs are the computer virus and its common variant, the worm. Other forms of malicious software include so-called logic bombs,[5] Trojan horses,[6] and trap doors.[7] Viruses and worms can be programmed to corrupt, delete or steal sensitive information. Malevolent code can also be exploited by cyber terrorists to disrupt elements of the national critical information infrastructure, such as banking, transportation, communications and energy provision systems, by attacking the computer systems at the heart of these structures.


COMPUTER VIRUSES

The term "virus," Latin for "poison," was first formally defined by Dr. Fred Cohen in 1983,[8] even though the concept originated in John von Neumann's studies of self-replicating mathematical automata in the 1940s.[9] A computer virus is a series of instructions (a program) that (i) infects a host program by attaching itself to the host, (ii) executes when the host is executed, and (iii) spreads by cloning itself, or part of itself, and attaching the new versions to other host programs. In addition, many viruses have a so-called payload capable of harmful side-effects, such as deleting, stealing or modifying digital information.[10] As the definition suggests, a typical computer virus consists of three basic modules or mechanisms, namely an infection module, payload trigger, and payload.


Infection module

The infection mechanism is the most salient technical property of a computer virus.[11] It enables a virus to reproduce and spread, by locating a prospective host program, and installing a copy of the virus onto it.[12] When the host program runs, control is eventually passed to the resident virus code, allowing it to execute. The executing virus repeats the infection cycle by automatically replicating itself and copying the newly-created clones to other executable host files on the system or network, and even across networks.[13]
A virus may infect a computer or a network through several possible points of entry, including via a downloaded infected file, through web browsing, through removable media such as writable compact disks and DVDs, via infected files in shared directories, via an infected e-mail attachment, and even hidden in infected commercial shrinkwrapped software.[14]
E-mail is the most widely used medium of exchanging files and sharing information, but it has also become a convenient and efficient vehicle for virus and worm propagation.[15] Fast-spreading viruses such as ExploreZip and Melissa, exploit automatic mailing programs to spread within and across networks.[16] Melissa typically arrived in the e-mail Inbox of its victim, disguised as an e-mail message with a Microsoft Word attachment. When the recipient opened the attachment, Melissa executed. First, it verified whether the recipient had the Microsoft Outlook e-mail program on its computer. If Outlook were present, Melissa would mail a copy of itself to the first fifty names in Outlook's address book, creating the appearance to the fifty new recipients that the user of the infected system had sent them a personal e-mail message. Melissa would then repeat the process with each of the fifty recipients of the infected e-mail message (provided they had Outlook), by automatically transmitting clones of itself to fifty more people. Melissa attacks frequently escalated and resulted in clogged e-mail servers and system crashes.[17]

Payload

In addition to replicating and spreading, viruses may be programmed to perform specific harmful actions. The module that implements this functionality is known as the payload.[18] A payload can perform a wide range of functions, depending on the aims and objectives of the virus author.[19] A payload can be programmed to perform destructive operations such as corrupting, deleting and stealing information.[20] A payload may also create intrusion-enabling devices, such as a backdoor[21] that allows unauthorized access to the infected machine.[22] Some payload effects are immediately obvious, such as a system crash, while others are subtle, such as transposition of numbers and alteration of decimal places.[23] Subtle effects tend to be dangerous because their presence may not be detected until substantial harm has been done. Payloads are often relatively harmless and do no more than entertain the user with a humorous message, musical tune, or graphical display.[24]
The payload is triggered when a specific condition is satisfied. Triggering conditions come in a variety of forms, such as a specified number of infections, a certain date, or specific time. The Friday-the-13th virus, for instance, only activated its payload on dates with the cursed designation.[25] More recently, the first CodeRed worm alternated between continuing its infection cycle, remaining dormant, and attacking the official White House Web page, depending on the day of the month.[26] In the simplest case, a payload executes whenever the virus executes, without a triggering event. Viruses do not always have a payload module, but even viruses without a payload may harm their environment by consuming valuable computing resources, by for instance, filling up available memory space and slowing the execution of important programs.


Computer worms

Worms are similar to viruses, but differ in two important respects. Worms propagate autonomously across networks without human intervention, and they replicate and spread without infecting a host program.[27] The CodeRed worm, for instance, propagated by injecting copies of itself into the memory of a remote system by exploiting a security vulnerability in the target system. It located potential targets by scanning the Internet for vulnerable systems, to which it propagated automatically.[28] The typical virus in contrast, needs to attach itself to an executable file, and then relies on human interaction to propagate across networks. Like viruses, worms may carry destructive payloads, but even without a destructive payload a fast-spreading worm can do significant harm by slowing down a system through the prolific network traffic it generates.[29]
The original worm was implemented by scientists at Xerox PARC in 1978,[30] but the so-called Morris Worm, created by Cornell University graduate student, Robert T. Morris, was the first to become a household name.[31] The 1989 Morris worm used a security flaw in a UNIX program to invade and shut down much of the Internet. By some accounts, this event first woke the world up to the dangers of computer security vulnerabilities, such as the buffer overflow flaw that enabled the Morris worm to paralyze the Internet.[32]


VIRUS DETECTION

Technical anti-virus defenses come in four varieties, namely signature scanners, activity monitors, integrity checkers, and heuristic techniques.[33] Scanners detect specific, known viruses by indentifying patterns that are unique to a particular virus strain. Activity monitors look out for virus-like activity in a computer. Integrity checkers sound an alarm when detecting suspicious modifications to computer files. Heuristic techniques combine virus-specific scanning with generic techniques, to provide a significantly broadened range of detection.


Scanners

Scanners are the most widely used anti-virus defense. A scanner reads executable programs and searches for the presence of virus patterns, known as "signatures." A virus signature consists of patterns of hexadecimal digits embedded in viral code that are unique to a particular virus strain.[34] These signatures are created by human experts at institutions such as IBM's High Integrity Computing Laboratory, who scrutinize viral code and extract sections of code with unusual patterns.[35] The selected byte patterns are collected in a signature database and used in anti-virus scanners. A scanner detects a virus in a program by comparing the program to its database of signatures, and announcing a match as a possible virus.[36]
An ideal virus signature would issue neither false negatives nor false positives.[37] Although this ideal is unachievable in practice, anti-virus researchers pursue optimal solutions within practical constraints. The IBM High Integrity Computing Laboratory, for instance, has developed an optimal statistical signature extraction technique that examines all sections of code in a virus, and selects the byte strings that optimize the tradeoff between false positives and negatives.[38]
Scanners are easy to use, but they are limited to detecting known signatures. Furthermore, a scanner's signature database has to be continually updated as new viruses are discovered and their signatures catalogued, a burdensome requirement in an environment where new viruses appear daily. Modern antivirus vendors have attempted to ligten the burden on users by distributing signature updates directly to their customers via the Internet.[39] As the number of known viruses grows, the scanning process will inevitably slow down as an ever-increasing set of possibilities has to be evaluated.

Activity monitors

Activity monitors are resident programs that monitor activities in a computer for behavior commonly associated with viruses. Suspicious activities include operations such as attempts by a program to delete information and mass mail copies of itself. When suspicious activity is detected, the monitor may simply halt execution and alert the user, or take definite action to neutralize the activity.[40] Activity monitors, unlike scanners, do not need to know the signature of a virus to detect it. Its function is to recognize generic suspicious behavior, not the precise identity of the culprit.
The greatest strength of activity monitors is their ability to detect unknown virus strains, but they also have significant weaknesses. They can only detect viruses that are actually executing, possibly after substantial harm has been done. A virus may, furthermore, execute before the monitor code does, and do harm before it is detected. A virus may also be programmed to alter monitor code on machines that do not have protection against such modification.
The effectiveness of activity monitors is limited by the lack of unambiguous rules defining "suspicious" activity. This ambiguity may result in false alarms when an activity monitor picks up legitimate activities which resemble virus-like behavior.[41] False negatives may result when an activity monitor fails to recognize viral activity which does not fit the monitor's programmed definitions.[42] Recurrent false alarms may ultimately lead users to ignore warnings from the monitor.


Integrity verification

An integrity verifier applies the electronic equivalent of a tamper-proof seal to protected programs, and issues an alert when the seal has been broken, presumably by a virus intrusion. An integrity verification program generates a code, known as a "checksum," for protected files. A checksum may be an arithmetic calculation based on variables such as the total number of bytes in a file, the numerical value of the file size and its creation date. A checksum is periodically recomputed and compared to the original. When a virus infects a file, it usually modifies the contents, resulting in a change in the checksum. When the recomputed value does not match the original, the file is presumed to have been modified since the previous inspection, and an alert is issued.[43]
The advantage of integrity checking is that it detects most instances of viral infection, as infection usually alters the target file. Its main drawback is its tendency to generate false alarms, as a file can change for "legitimate" reasons unrelated to virus infection.[44] Integrity checking software therefore presents a high likelihood of false positives, because of the general difficulty of determining whether a program modification is legitimate or due to a virus.[45] Integrity checking works best on static files, such as system utilities, but it is, of course, an inappropriate technique for files that naturally change frequently, such as Word documents.


Heuristic detection

Modern virus detectors increasingly use heuristic methods. Heuristic rules solve complex problems "fairly well" and "fairly quickly," but less than perfectly. Virus detection is an example of a complex problem that is amenable to heuristic solution. It has been proven that it is mathematically impossible to write a virus detection program that is capable of consistent perfect detection.[46] Heuristic virus detection methods accept such limitations and attempt to achieve a heuristic solution, namely a detection rate that is below the (unachievable) perfect rate, but representing an optimal tradeoff between detection accuracy, speed and computational expense.[47]
Heuristics detect novel viruses by examining the structure and logic of executable code for evidence of virus-like behavior. The heuristic program then assesses the likelihood that a scrutinized program constitutes a virus by calculating a score based on the number and type of virus-like characteristics detected. If the score exceeds a certain threshold, the scanner classifies the program as malevolent code, and notifies the user. Instructions to send an e-mail message with an attachment to every listing in an address book, for instance, would add significantly to the score. Other high-scoring routines include capabilities to replicate, hide from detection, and execute some kind of payload.[48]
A heuristic scanner typically operates in two phases. The scanning algorithm first narrows the search by identifying the location most likely to contain a virus. It then analyzes code from that location to determine its likely behavior upon execution. A static heuristic scanner compares the code from the "most likely" location to a database of byte sequences commonly associated with virus-like behavior, and decides whether to classify the code as viral.[49] A dynamic heuristic scanner uses CPU[50] emulation.[51] It loads suspect code into a virtual computer, emulates its execution and monitors its behavior. Because it is only a virtual computer, virus-like behavior can be safely observed in what is essentially a laboratory setting, with no need to be concerned about real damage.[52] Although dynamic heuristics can be time-consuming due to the computationally intensive CPU emulation process, they are sometimes superior to static heuristics. This will be the case when the suspect code is obscure and not easily recognizable as viral in its static state, but clearly reveals its viral nature in a dynamic state.
A heuristic assessment is less than perfect by design, and will inevitably provide false positives and negatives. A low scanner threshold will result in false positives. A scanner with a threshold that is set too high, on the other hand, will fail to detect viruses that are malicious but that do not exactly match the unrealistically tight specifications, resulting in false negatives.[53] As in the case of activity monitors, the term "suspicious" is ambiguous. Many legitimate programs, including even some anti-virus programs, perform operations that resemble virus-like behavior.[54] Nevertheless, state-of-the-art heuristic scanners achieve a 70-80 percent success rate at detecting unknown viruses.[55]
A major advantage of heuristic scanning over generic anti-virus technologies such as behavior monitoring and integrity checking, is its ability to detect viruses before they execute and cause harm. Its advantage over conventional signature scanners lies in its capability to detect novel virus strains whose signatures have not yet been catalogued. Heuristic scanners are also capable of detecting complex virus families, such as polymorphic viruses which complicate detection by changing their signatures from infection to infection.[56]
The explosive growth in new virus strains has made reliable detection and identification of individual strains very difficult and costly, making heuristics more important and increasingly prevalent.[57] Commercial heuristic scanners include IBM's AntiVirus boot scanner and Symantec's Bloodhound technology.


3. LIABILITY ANALYSIS

Cyber intruders may be subject to criminal,[58] as well as civil liability.[59] However, cyber rogues are often judgment-proof and may be difficult to identify or subject to jurisdiction.[60] A deep-pocketed defendant, such as an organization whose lax security negligently enabled a cyber crime, is therefore usually the preferred target for a civil lawsuit.[61] This section discusses liability issues related to information security failures. It focuses on the paradigmatic case where a defendant negligently enabled a crime by a third party against the plaintiff.
A plaintiff may pursue a civil action under a negligence theory, the most widely used theory of liability in the law of torts.[62] Negligence is generally defined as a breach of the duty not to impose an unreasonable risk on society.[63] It applies to any risk that can be characterized as unreasonable, including risks associated with information security failures. A victim of an information security breach may therefore bring legal action under a negligence theory against anyone who contributed to the risks associated with the breach, including those who failed in their duty to reduce or eliminate the risk.[64]
The plaintiff in a negligence action has to prove the following elements to establish her claim.

  1. A legal duty on the part of the defendant not to expose the plaintiff to unreasonable risks.
  2. A breach of the duty, namely a failure on the part of the defendant to conform to the norm of reasonableness.
  3. A causal connection between the defendant's conduct and the plaintiff's harm. This element includes actual as well as proximate cause. A defendant's negligence is the actual cause of the plaintiff's harm if, but for the negligence, the harm would not have occurred.[65] The proximate causation element requires the defendant's conduct to be reasonably related to the plaintiff's harm.[66]
  4. Actual damage resulting from the defendant's negligence.[67]

Duty

Negligence liability of a defendant depends first and foremost on the existence of a legal duty of care to the plaintiff. A duty of care may be imposed by common law tort principles.[68] A duty may also be imposed by statute, either expressly,[69] or by legal precedent, if the statute does not expressly provide for civil liability.[70] California's Security Breach Information Act (SBIA),[71] for instance, expressly creates a civil cause of action for a business' failure to protect customers' personal information, and provides that aggrieved customers may recover damages for breach of that duty. The relevant provision states: "A business that owns or licenses personal information about a California resident shall implement and maintain reasonable security procedures and practices appropriate to the nature of the information, to protect the personal information from unauthorized access, destruction, use, modification, or disclosure."[72] The legislation further provides, "Any customer injured by a violation of this title may institute a civil action to recover damages."[73]
The civil cause of action created by the SBIA is rooted in negligence, because the duty it imposes is based on reasonableness, the essence of the standard of care imposed by negligence law.[74] The statute requires, for instance, implementation of "reasonable security procedures and practices"[75] that are "appropriate to the nature of the information."[76]


Breach of duty

"Breach of duty" refers to a violation of the duty to avoid unreasonable risks of harm to others. This element defines the standard of due care to which a defendant is held, and determines whether the standard has been violated. The traditional law and economics approach to negligence analysis defines due care as a level of precaution that minimizes a global cost function.[77] In practice, however, courts do not follow this approach. Instead, the common law requires the plaintiff to identify and plead an untaken precaution that would have prevented the accident, if taken. The defendant will then be considered to have breached her duty if the benefits of risk reduction provided by the pleaded precaution exceeded its cost.[78]
The breach calculus weighs the cost of the untaken precaution against the value of the reduction in all foreseeable risks that the precaution would have provided, not just the risk that actually materialized.[79] For instance, the common computer security vulnerability known as a buffer overflow,[80] may enable a variety of cyber crimes, including identity theft, denial of service, and data corruption. A victim of identity theft enabled by a buffer overflow may argue that the defendant should have patched the vulnerability in a timely fashion, and plead this as an untaken precaution. To show breach, the plaintiff must balance the cost of eliminating the buffer overflow against the aggregate benefit of all risks so reduced, including risks related to denial of service and data corruption, not just the risk of identity theft.
The role of the untaken precaution in negligence law is illustrated in Cooley v. Public Service Co.[81] In Cooley, the plaintiff sustained injuries from a loud noise over a telephone wire. She suggested two untaken precautions that would have prevented the harm, namely (i) a strategically positioned wire mesh basket, and (ii) insulating the wires. The court ruled that neither untaken precaution constituted a breach of duty. Both precautions would have increased the risk of electrocution to passersby sufficiently to outweigh the benefits in harm reduction.
The Cooley court noted that, although the suggested precautions did not succeed, there may exist a cost-effective precaution other than the ones pleaded, that would have satisfied the breach requirement. It is however, the plaintiff's burden to identify and plead such a precaution, if indeed it exists. In a negligence case, more than one untaken precaution may have greater benefits than costs, and the plaintiff may allege several precautions in the alternative.[82] A court may base a finding of negligence on one or more of the pleaded precautions.[83]
The following stylized numerical example illustrates the cost-benefit calculus needed to prove a breach of duty that enabled a cyber attack.

Numerical example

In this hypothetical, we assume that an organization uses a signature scanner to detect viruses in its computer network. An attacker launches a virus which evades detection and compromises the confidential information of a customer. The culprit virus is a novel strain that has been documented recently for the first time. It was not detected because its signature had not been included in the database of the organization's scanner at the time of the attack.
The victimized customer contemplates a negligence lawsuit against the organization. The plaintiff must prove the defendant's breach of duty by identifying an alternative cost-effective precaution that would have avoided the virus. The plaintiff has several pleading options. Potential untaken precautions include more frequent updating of the signature database, or perhaps use of a generic scanner that does not depend on an updated database. Each option has its own set of costs and benefits which must be balanced to determine its cost-effectiveness.[84]
Suppose the plaintiff selects a policy of updating the signature database more frequently. The numbers in Table 1 suggest that this precaution is efficient.[85] The precaution would add 3 cents to the defendant's average cost of production, but would reduce its expected information security-related loss by 8 cents. The first column lists the alternative precautions at issue, namely continuing scanning at the current rate and scanning at the proposed increased rate, respectively. The second column lists the defendant's average production cost for each precaution. The third column lists the probabilities of virus transmission corresponding to the respective precautions, the fifth the expected losses from a virus attack, and the final column lists the full cost per unit of production, namely production cost plus expected losses due to a virus attack. We assume that a virus attack will result in expected damages of $10,000.


2009_3400.png

With the defendant's precaution at its current level, the production cost per unit is 40 cents, the probability of a successful virus infection is 1/100,000, and the loss if an infection occurs is $10,000. The expected loss per unit due to a virus attack therefore equals 10 cents (1/100,000 X $10,000), and the total cost per unit is 50 cents.
If the defendant implemented the proposed precaution pleaded by the plaintiff, its production cost would increase to 43 cents, the probability of virus infection would decline to 1/500,000, and the expected loss due to a virus attack would be 2 cents, giving a total cost per unit of 45 cents. The plaintiff should therefore prevail on the issue of breach. Although the pleaded precaution would increase the defendant's cost by 3 cents per unit, it would lower expected information security-related losses by 8 cents.
The cost-benefit analysis draws on the law, economics and technology of information security, and it requires data on the types, nature and frequencies of cyber attacks. Such an analysis is feasible in principle, and it is aided by the appearance of cost-benefit models of viruses and anti-virus defenses in the computer security literature,[86] and the increasing availability of empirical data on cyber attacks and their economic impact.[87]


Proximate cause

The plaintiff must prove that the defendant's failure to take the untaken precaution was the actual and proximate cause of real harm to the plaintiff. There are two doctrines of proximate cause in the common law, namely the direct consequences doctrine and the reasonable foresight doctrine.[88] The reasonable foresight doctrine considers whether the type of harm sustained by the plaintiff was a reasonably foreseeable consequence of the defendant's wrongdoing.[89] The direct consequences doctrine applies where the wrongful acts of multiple tortfeasors were all necessary causes of the plaintiff's injury. The doctrine examines whether the subsequent tortfeasors have cut off the liability of the original tortfeasor.[90]
The Encourage Free Radicals (EFR) doctrine,[91] a paradigm within the direct consequences doctrine, preserves the liability of an original tortfeasor who has encouraged the opportunistic behavior of so-called "free radicals." Free radicals are individuals who are not deterred by the threat of tort or criminal liability, because they lack mental capacity and good judgment, and are shielded from liability by anonymity and insufficient assets. Such trouble-prone individuals are termed "free radicals" because of their tendency to bond with trouble. Examples of free radicals include children, anonymous crowds, criminals, mentally incompetent individuals, terrorists blinded by ideological or religious motivations, and cyber wrongdoers.[92] The EFR doctrine recognizes that the deterrence rationale of negligence law would be defeated if responsible people who foreseeably encouraged free radicals to be negligent were allowed to escape judgment by shifting liability to the latter. Common law negligence rules therefore preserve the liability of the responsible individuals. The EFR doctrine imposes liability on the encourager, even when intentional or criminal behavior by a free radical intervened.[93]
Guille v Swan[94] was possibly the original EFR case in the United States.[95] In Guille, the defendant descended in a balloon over New York City into the plaintiff's garden in a manner that attracted a crowd. The defendant's balloon dragged over the plaintiff's garden, but the crowd did much more damage to the garden. The defendant argued that he should be responsible only for his share of the damages, and not for that caused by the crowd, but the court nevertheless held him responsible for all the damages. The crowd were free radicals in that particular situation, and the defendant foreseeably encouraged their behavior. People who are otherwise perfectly rational often behave differently when they are shielded by the anonymity and diminished accountability of a crowd. Chief Justice Spencer stated that the defendant's manner of descent would foreseeably draw a crowd, with predictable consequences for which he should be held responsible, a classic description of the EFR doctrine.[96]
Under the EFR doctrine, the liability of a defendant who enabled a cyber crime by, for instance, failing to correct a security vulnerability, may be preserved if it can be ascertained that the security vulnerability had foreseeably provided encouragement to free radicals. This will often be the case, as research shows that cyber wrongdoers have properties commonly associated with free radicals. They are often judgment-proof and shielded by the anonymity of the Internet architecture.[97] They appear undeterred by the threat of liability,[98] and unconcerned about the problems caused by their actions.[99] Furthermore, security breaches are under-reported, under-prosecuted and the probability of apprehending a cyber attacker is comparatively low.[100] Studies show that most virus authors would either be unaffected or, perversely, actually encouraged by stricter legislation against computer fraud and abuse.[101] These factors are consistent with a free radical profile.
A defendant, such as an organization whose negligent security policies enabled a cyber crime, may therefore be held liable to victims of the crime, even though the criminal act of the attacker intervened between the defendant's wrongdoing and the plaintiff's harm. Such a result would be helpful to a plaintiff in cases where the primary tortfeasor has sufficient assets to pay a judgment, but the actual perpetrator is judgment-proof or otherwise immune to liability.

4. CONCLUSION

This article analyzes the legal standard of due care in information security, as an issue of law, economics and technology. The court deciding the case does not determine the standard of care. The plaintiff in a negligence action defines the standard of care by identifying and pleading an untaken precaution, and proving that the precaution is cost-effective. The plaintiff must further prove that failure to take the precaution was the actual and proximate cause of the plaintiff's harm. The plaintiff's selection of untaken precaution therefore defines every aspect of the negligence analysis that will ultimately determine the defendant's liability.


[1] See e.g. Remarks of J. Howard Beales, III, Director, Bureau of Consumer Protection, Federal Trade Commission, Before the 2003 Symposium on the Patriot Act, Consumer Privacy, and Cybercrime, 5 N.C. J.L. & TECH. 1, 2.

[2] See e.g. John E. Calfee & Richard Craswell, Some Effects of Uncertainty on Compliance With Legal Standards, 70 VA. L. REV. 965, 965.

[3] See e.g. Nicholas Weaver et al., A Taxonomy of Computer Worms, 2003 ACM Workshop on Rapid Malcode, Wash. D.C., p. 16 ("It may be tempting to say that we could build secure systems which will not have exploitable vulnerabilities. However, even highly secure software systems with reputations for robustness and which have received considerable security scrutiny including multiple code reviews ... have contained major security holes.")

[4] See e.g. Michael L. Rustad and Thomas H. Koenig, The Tort of Negligent Enablement of Cybercrime, 20 BERKELEY TECH. L. J. 1553, 1598 ("[T]he overarching goal of tort law is to control the costs of accidents rather than to eliminate them.")

[5] A logic bomb is "a section of code, preprogrammed into a larger program, that waits for a trigger event to perform a harmful function. Logic bombs do not reproduce and are therefore not viral, but a virus may contain a logic bomb as a payload." Peter Szor, THE ART OF COMPUTER VIRUS RESEARCH AND DEFENSE (2005), at 30.

[6] A Trojan horse is a program that appears to be beneficial, but contains a harmful payload. Peter Szor, THE ART OF COMPUTER VIRUS RESEARCH AND DEFENSE (2005), at 663.

[7] A trapdoor, or backdoor, is a function built into a program or system to allow unauthorized access to the system. Peter Szor, THE ART OF COMPUTER VIRUS RESEARCH AND DEFENSE (2005), at 643.

[8] FRED COHEN, COMPUTER VIRUSES. PhD dissertation, University of Southern California (1985).

[9] See e.g. R. Lehtinen et al., COMPUTER SECURITY BASICS (O'Reilly, 2006), at 83.

[10] See e.g. FREDERICK B. COHEN, A SHORT COURSE ON COMPUTER VIRUSES (Wiley, 1994, 2d ed.), at 1-2.

[11] LANCE J. HOFFMAN (ed.), ROGUE PROGRAMS: VIRUSES, WORMS, TROJAN HORSES (Van Nostrand Reinhold, 1990), at 247.

[12] See e.g. DOROTHY E. DENNING and PETER J. DENNING, INTERNET BESIEGED (ACM Press, New York, 1998), at 81.

[13] See Ed Skoudis, MALWARE FIGHTING MALICIOUS CODE (Prentice Hall, 2004), at 31-37.

[14] See e.g. DOROTHY E. DENNING and PETER J. DENNING, INTERNET BESIEGED (ACM Press, New York, 1998), at 81.

[15] ICSA Labs 10th Annual Computer Virus Prevalence Survey 2004, Table 5 and Fig. 10, p. 15 (Survey demonstrating popularity of e-mail as medium of virus attack.)

[16] A. Bisset and G. Shipton, Some Human Dimensions of Computer Virus Creation and Infection, 52 INTERNATIONAL J. HUM. COMPUTER STUD. (2000), 899; R. Ford, No Surprises in Melissa Land, 18 COMPUTERS AND SECURITY, 300-302.

[17] David Harley et al., VIRUSES REVEALED UNDERSTAND AND COUNTER MALICIOUS SOFTWARE (Osborne/McGraw-Hill, 2001), 406-410.

[18] JAN HRUSKA, COMPUTER VIRUSES AND ANTI-VIRUS WARFARE, (Ellis Horwood Ltd., 1990), at 17, 18.

[19] See e.g. Nicholas Weaver et al., A Taxonomy of Computer Worms, 2003 ACM Workshop on Rapid Malcode, Wash. D.C., p. 4 ("The payload is limited only by the imagination of the attacker.")

[20] See e.g. Meiring de Villiers, Computer Viruses and Civil Liability: A Conceptual Framework, TORT TRIAL AND INSURANCE PRACTICE LAW JOURNAL, Fall 2004 (40:1), 123, 172 (Discussion of damage due to virus infection.)

[21] A backdoor is a method of gaining remote access to a computer without passing through normal security controls on a system. See Robert Slade, DICTIONARY OF INFORMATION SECURITY (2006), at 19.

[22] Ed Skoudis, MALWARE FIGHTING MALICIOUS CODE (Prentice Hall, 2004), at 27.

[23] See e.g. Peter Szor, THE ART OF COMPUTER VIRUS RESEARCH AND DEFENSE (2005), at 302-303 (Describing "data diddlers" as viruses that "do not destroy data all of a sudden in avery evident form, ... but slowly manipulate the data, such as the content of the hard disk.")

[24] E.J. Sinrod and W.P. Reilly, Cyber-Crimes: A Practical Approach to the Application of Federal Computer Crimes Laws, 16 SANTA CLARA COMPUTER & HIGH TECH. L.J. 117, 218 (describing the W95.LoveSong.998 virus, designed to trigger a love song on a particular date.)

[25] See, e.g., Eric J. Sinrod and William P. Reilly, Cyber Crimes A Practical Approach to the Application of Federal Computer Crime Laws, 16 SANTA CLARA COMPUTER & HIGH TECH. L. J. 177 (2000), at 217, n. 176.

[26] See Meiring de Villiers, Free Radicals in Cyberspace: Complex Liability Issues in Information Warfare, 4 NORTHWESTERN TECH & IP L. J. 13, 13-14 (Fall, 2005).

[27] See United States v. Robert Tappan Morris, 928 F.2d 504 (1991); Nicholas Weaver et al., A Taxonomy of Computer Worms, 2003 ACM Workshop on Rapid Malcode, Wash. D.C., pp. 11-18. http://doi.acm.org/10.1145/948187.948190.

[28] See Peter Szor, THE ART OF COMPUTER VIRUS RESEARCH AND DEFENSE (2005), at 398-401.

[29] See, generally, John F. Schoch and Jon A. Hupp, The "Worm" Programs - Early Experience with a Distributed Computation, COMM. ACM, Vol 25, No 3, March 1982, 172.

[30] http://www.parc.xerox.com/about/history/default.html.

[31] See David Harley, Robert Slade, and Urs E. Gattiker, VIRUSES REVEALED (2001), at 347-52.

[32] Takanen et al., Running Malicious Code By Buffer Overflows: A Survey of Publicly Available Exploits, 162. EICAR 2000 Best Paper Proceedings. Available at http://www.papers.weburb.dk.

[33] See Peter Szor, THE ART OF COMPUTER VIRUS RESEARCH AND DEFENSE (2005), p. Ch. 11.

[34] JAN HRUSKA, COMPUTER VIRUSES AND ANTI-VIRUS WARFARE (Ellis Horwood, Ltd., 1990), at 42.

[35] See Ed Skoudis, MALWARE FIGHTING MALICIOUS CODE (Prentice Hall, 2004), 53, 54.

[36] JEFFREY O. KEPHART ET AL., Automatic Extraction of Computer Virus Signatures, Proceedings of the 4th Virus Bulletin International Conference, R. Ford, ed., Virus Bulletin Ltd., Abingdon, England, 1994, pp. 179-194, at 2.

[37] See e.g. Jeffrey O. Kephart et al., Blueprint for a Computer Immune System, IBM Thomas J. Watson Research Center Report, at 11.

[38] See e.g. Jeffrey O. Kephart et al., Automatic Extraction of Computer Virus Signatures, Proceedings of the 4th Virus Bulletin International Conference, R. Ford, ed., Virus Bulletin Ltd., Abingdon, England, 1994, pp. 179-194.

[39] See e.g. Ed Skoudis, MALWARE FIGHTING MALICIOUS CODE (Prentice Hall, 2004), 54.

[40] Sandeep Kumar and Eugene H. Spafford, A Generic Virus Scanner in C++, Technical report CSD-TR-92-062, Dept. of Computer Science, Indiana University, at 3-4.

[41] See e.g. ROBERT SLADE, ROBERT SLADE'S GUIDE TO COMPUTER VIRUSES (Springer, 2d ed., 1996), at 40-41.

[42] JAN HRUSKA, COMPUTER VIRUSES AND ANTI-VIRUS WARFARE, (Ellis Horwood Ltd., 1990), at 75.

[43] See Ed Skoudis, MALWARE FIGHTING MALICIOUS CODE (Prentice Hall, 2004), 58.

[44] PHILIP FRITES, PETER JOHNSTON AND MARTIN KRATZ, THE COMPUTER VIRUS CRISIS (Van Nostrand Reinhold, New York, 2d ed., 1992), at 125; Ed Skoudis, MALWARE FIGHTING MALICIOUS CODE (Prentice Hall, 2004), 58.

[45] ROBERT SLADE, ROBERT SLADE'S GUIDE TO COMPUTER VIRUSES (Springer, 2d ed., 1996) 157.

[46] Diomidis Spinellis, Reliable Identification of Bounded-Length Viruses is NP-Complete, IEEE TRANSACTIONS ON INFORMATION THEORY, 49(1), 280, 282 (January 2003) (Stating that theoretically perfect detection is in the general case undecidable, and for known viruses, NP-complete.); Chess & White, Undetectable Computer Virus, IBM Research Paper,
available at http://www.research.ibm.com/antivirus/SciPapers/VB2000DC.htm.

[47] See e.g. Carey Nachenberg, Future Imperfect, VIRUS BULLETIN, August 1997, 6.

[48] See e.g. Peter Szor, THE ART OF COMPUTER VIRUS RESEARCH AND DEFENSE (2005), at 472-74.

[49] Sandeep Kumar and Eugene H. Spafford, A Generic Virus Scanner in C++, Technical report CSD-TR-92-062, Dept. of Computer Science, Indiana University, at 4-5.

[50] The CPU, or central processing unit, of a computer is responsible for data processing and computation.

[51] See e.g. JAN HRUSKA, COMPUTER VIRUSES AND ANTI-VIRUS WARFARE, (Ellis Horwood Ltd., 1990), at 115; D. BENDER, COMPUTER LAW: EVIDENCE AND PROCEDURE (1982), §2.02, at 2-7, -9.

[52] Sandeep Kumar and Eugene H. Spafford, A Generic Virus Scanner in C++, Technical report CSD-TR-92-062, Dept. of Computer Science, Indiana University, at 4.

[53] Ed Skoudis, MALWARE FIGHTING MALICIOUS CODE (Prentice Hall, 2004), 56.

[54] Francisco Fernandez, Heuristic Engines, Proc. 11th Intl. Virus Bulletin Conference, September 2001, Virus Bulletin Ltd., Abingdon, England, 1994, at 409.

[55] Carey Nachenberg, Future Imperfect, VIRUS BULLETIN, August 1997, at 7; Understanding Heuristics: Symantec's Bloodhound Technology, Symantec White Paper Series, v. XXXIV, at 9.

[56] Polymorphic viruses have the ability to "mutate" by varying the code sequences written to target files. To detect such viruses requires a more complex algorithm than simple pattern matching. See, e.g., DOROTHY E. DENNING and PETER J. DENNING, INTERNET BESIEGED (ACM Press, New York, 1998), at 89.

[57] Carey Nachenberg, Future Imperfect, VIRUS BULLETIN, August 1997, at 9.

[58] See S.D. Personick, ed., Critical Information Infrastructure Protection and the Law: An Overview of Key Issues, 37-39; Brent Wible, A Site Where Hackers are Welcome: Using Hack-In Contests to Shape Preferences and Deter Computer Crime[2003] YaleLawJl 8; , 112 YALE L. J. 1577, 1581-85 (2003).

[59] See Michael L. Rustad, Private Enforcement of Cybercrime on the Electronic Frontier, 11 S. CAL. INTERDISC. L.J. 63, 66; Robin A. Brooks, Deterring the Spread of Viruses Online: Can Tort Law Tighten the Net?, 17 REV. LIT. 343 (1998); 18 U.S.C. 1030(g) (2000) (Provision in Computer Fraud and Abuse Act allowing civil action against wrongdoer "to obtain compensatory damages and injunctive relief or other equitable relief.")

[60] See Meiring de Villiers, Free Radicals in Cyberspace: Complex Liability Issues in Information Warfare, 4 NORTHWESTERN TECH & IP L. J. 13 (Fall, 2005), § 4.

[61] See Meiring de Villiers, Free Radicals in Cyberspace: Complex Liability Issues in Information Warfare, 4 NORTHWESTERN TECH & IP L. J. 13 (Fall, 2005), § 3.3 (Analysis of civil liability for enablement of cyber crime and tort.)

[62] See James A. Henderson, Why Negligence Law Dominates Tort, 50 UCLA L. REV. 377 (2003).

[63] PROSSER AND KEETON ON THE LAW OF TORTS (5th ed., West Publ. Co., 1984), § 31. Second Restatement of Torts, § 282.

[64] Dan B. Dobbs, The Law of Torts, at 258 (The plaintiff can assert that any conduct counts as negligence.)

[65] See e.g. Meiring de Villiers, Computer Viruses and Civil Liability: A Conceptual Framework, TORT TRIAL AND INSURANCE PRACTICE LAW JOURNAL, Fall 2004 (40:1) 123, 141-42 (Analysis of causality in context of virus attack.)

[66] See e.g. David G. Owen, Idea: The Five Elements of Negligence, 35 HOFSTRA L. REV. 1671, 1681 (Defining proximate cause as "a reasonably close connection between a defendant's wrong and the plaintiff's injury, a connection that is not remote.")

[67] See e.g. Meiring de Villiers, Computer Viruses and Civil Liability: A Conceptual Framework, TORT TRIAL AND INSURANCE PRACTICE LAW JOURNAL, Fall 2004 (40:1) 123, 172-74 (Discussion of damage due to virus infection.)

[68] See e.g. Vincent R. Johnson, Cybersecurity, Identity theft, and the Limits of Tort Liability, 57 S.C. L. REV. 255, 272-82; Michael J. Rustad & Thomas H. Koening, Extending Learned Hand's Negligence Formula to Information Security Breaches, 3 ISJLP 237, 239-40 (Arguing that "companies have a duty to provide reasonable information security practices under the common law of torts.")

[69] See Restatement (Third) of Torts: Liability for Physical Harm 14 cmt. b (Proposed Final Draft No. 1, 2005) (Discussing express and implied statutory causes of action.)

[70] See e.g. RESTATEMENT (THIRD) OF TORTS§ 12 (Discussion Draft 1999) ("An actor is negligent if, without excuse, the actor violates a statute that is designed to protect against the type of accident the actor's conduct causes, and if the accident victim is within the class of persons the statute is designed to protect."); Vincent R. Johnson and Alan Gunn, STUDIES IN AMERICAN TORT LAW, 305-06 (3d ed., 2005).

[71] Act effective July 1, 2003, ch. 915, 2002 Cal. Legis. Serv. (West), available at CA LEGIS 915 (2002) (Westlaw).

[72] Cal. Civ. Code 1798.81.5 (West Supp. 2005).

[73] Cal. Civ. Code 1798.81(b) (West Supp. 2005).

[74] RESTATEMENT (SECOND) OF TORTS § 283 (1965.); R.W. Wright, The Standards of Care in Negligence Law, in D.G. Owen (ed.), PHILOSOPHICAL FOUNDATIONS OF TORT LAW (Clarendon Press, 1995), at 249; Second Restatement of Torts, § 282.

[75] Cal. Civ. Code 1798.81.5 (b) (West Supp. 2005).

[76] Cal. Civ. Code 1798.81.5(c) (West Supp. 2005).

[77] See e.g. John Prather Brown, Toward an Economic Theory of Liability, 2 J. LEGAL STUD. 323 (1973); William M. Landes & Richard A. Posner, THE ECONOMIC STRUCTURE OF TORT LAW 63 (1987); Oliver Wendell Holmes, Jr., THE COMMON LAW 111 (1881).

[78] See Mark F. Grady, UNTAKEN PRECAUTIONS, 18 J. LEGAL STUD. 139, 143 (1989); Delisi v. St. Luke's Episcopal-Presbyterian Hosp., Inc., 701 S.W.2d 170 (Mo. App. 1985) (Plaintiff had to prove physician's breach of duty by specifying the treatment that should have been given, but was not.)

[79] See e.g. Restatement (Second) of Torts § 281(b), comment e (1965) ("Conduct is negligent because it tends to subject the interests of another to an unreasonable risk of harm. Such a risk may be made up of a number of different hazards, which frequently are of a more or less definite character. The actor's negligence lies in subjecting the other to the aggregate of such hazards."); Mark F. Grady, UNTAKEN PRECAUTIONS, 18 J. LEGAL STUD. 139, 146 (1989).

[80] See e.g. Meiring de Villiers, Reasonable Foreseeability in Information Security Law: A Forensic Analysis, 30 HASTINGS COMM. ENT. L.J. 419, 437-39 (Description of the buffer overflow vulnerability.)

[81] 90 N.H. 460, 10 A.2d 673 (1940).

[82] MARK F. GRADY, Untaken Precautions, 18 J. LEGAL STUD. 139, 144 (1989).

[83] See MARK F. GRADY, Untaken Precautions, 18 J. LEGAL STUD. 139, 144-45 (1989).

[84] See e.g. Fred Cohen, A Cost Analysis of Typical Computer Viruses and Defenses, COMPUTERS & SECURITY, 10 (1991).

[85] Based on an example in A.M. Polinsky, INTRODUCTION TO LAW AND ECONOMICS 98 (table 11) (1983).

[86] See, e.g., Fred Cohen, A Cost Analysis of Typical Computer Viruses and Defenses, COMPUTERS & SECURITY, 10 (1991); Robert W. Hahn & Anne Layne-Farra, The Law and Economics of Software Security, Working Paper 06-08, AEI Brookings Joint Center for Regulatory Studies; Huaqiang Wei et al., Cost-Benefit Analysis for Network Intrusion Detection Systems, CSI 28th Annual Computer Security Conefrence, October 2001; Meiring de Villiers, Computer Viruses and Civil Liability: A Conceptual Framework, TORT TRIAL AND INSURANCE PRACTICE LAW JOURNAL, Fall 2004 (40:1), 123, 161-63; TECH404, http://www.tech-404.com/calculator.html (Online cost calculator enabling estimation of financial impact of information security breaches.)

[87] See e.g. Robert Richardson, 2008 CSI COMPUTER CRIME & SECURITY SURVEY; ICSA Labs 10th Annual Computer Virus Prevalence Survey 2004; Anat Hovav & John D'Arcy, Capital Market Reaction to Defective IT Products: The Case of Computer Viruses, 24 COMPUTERS & SECURITY 409 (2005); Brian Cashell et al., The Economic Impact of Cyber Attacks, CRS Report for Congress, April 1, 2004; Shane Coursen, The Financial Impact of Viruses, INFORMATION SYSTEMS SECURITY 64 (Spring 1997). See however, Michael L. Rustad and Thomas H. Koenig, The Tort of Negligent Enablement of Cybercrime, 20 BERKELEY TECH. L. J. 1553, 1599 ("[N]o reliable data exists on the probability of cybercrime, its severity, or the costs of minimizing intrusions, so it is difficult to determine the most efficient level of precaution.")

[88] MARK F. GRADY, Untaken Precautions, 18 J. LEGAL STUD. 139, 151 (1989).

[89] See e.g. W. Jonathan Cardi, Purging Foreseeability: The New Vision of Duty and Judicial Power in the Proposed Restatement (Third) of Torts, 58 VAND. L. REV. 739, 749 ("[A] plaintiff may fail to survive the proximate cause inquiry where the defendant's actions resulted in (1) an unforeseeable type of injury, (2) an injury occurring in an unforeseeable manner, or (3) injury to an unforeseeable plaintiff.")

[90] Mark F. Grady, Proximate Cause Decoded, 50 UCLA L. REV. 299 (2002).

[91] See Mark F. Grady, The Free Radicals of Tort, SUPREME COURT ECONOMIC REVIEW 189 (2004).

[92] Mark F. Grady, Proximate Cause Decoded, 50 UCLA L. REV. 293, 306-312 (2002).

[93] Mark F. Grady, Proximate Cause Decoded, 50 UCLA L. REV. 293, 308 (2002).

[94] 19 Johns. 381 (N.Y. 1822).

[95] See Mark F. Grady, The Free Radicals of Tort, SUPREME COURT ECONOMIC REVIEW, 2004, 189, 201, 201.

[96] Mark F. Grady, The Free Radicals of Tort, at 113.

[97] See Meiring de Villiers, Free Radicals in Cyberspace: Complex Liability Issues in Information Warfare, 4 NORTHWESTERN TECH & IP L. J. 13, 39-42 (Fall, 2005).

[98] See e.g. Sarah Gordon, Virus Writers: The End of Innocence..; R. Lemos (1999), 'Tis the Season for Computer Viruses. www..zdnet.co.uk/news/1999/49/ns-12098.html.

[99] See Meiring de Villiers, Free Radicals in Cyberspace: Complex Liability Issues in Information Warfare, 4 NORTHWESTERN TECH & IP L. J. 13, 42-44. (Fall, 2005).

[100] See e.g. Jelena Mirkovic et al., INTERNET DENIAL-OF-SERVICE: ATTACK AND DEFENSE MECHANISMS 14 (2005).

[101] See e.g. Sarah Gordon, Virus Writers: The End of Innocence. IBM White Paper, http://www.research.ibm.com/antivirus/SciPapers/VB2000SG.htm.


AustLII: Copyright Policy | Disclaimers | Privacy Policy | Feedback
URL: http://www.austlii.edu.au/au/journals/UNSWLRS/2009/34.html