scispace - formally typeset
Search or ask a question

Showing papers on "Intrusion detection system published in 2000"


Journal ArticleDOI
TL;DR: The purpose of this article is to attempt to identify the shortcomings of the Lincoln Lab effort in the hope that future efforts of this kind will be placed on a sounder footing.
Abstract: In 1998 and again in 1999, the Lincoln Laboratory of MIT conducted a comparative evaluation of intrusion detection systems (IDSs) developed under DARPA funding. While this evaluation represents a significant and monumental undertaking, there are a number of issues associated with its design and execution that remain unsettled. Some methodologies used in the evaluation are questionable and may have biased its results. One problem is that the evaluators have published relatively little concerning some of the more critical aspects of their work, such as validation of their test data. The appropriateness of the evaluation techniques used needs further investigation. The purpose of this article is to attempt to identify the shortcomings of the Lincoln Lab effort in the hope that future efforts of this kind will be placed on a sounder footing. Some of the problems that the article points out might well be resolved if the evaluators were to publish a detailed description of their procedures and the rationale that led to their adoption, but other problems would clearly remain./par>

1,346 citations


Proceedings ArticleDOI
01 Aug 2000
TL;DR: This paper examines the vulnerabilities of a wireless ad-hoc network, the reason why intrusion detection is needed, and the reasons why the current methods cannot be applied directly, and describes the new intrusion detection and response mechanisms that are developing for wirelessAd-Hoc networks.
Abstract: As the recent denial-of-service attacks on several major Internet sites have shown us, no open computer network is immune from intrusions. The wireless ad-hoc network is particularly vulnerable due to its features of open medium, dynamic changing topology, cooperative algorithms, lack of centralized monitoring and management point, and lack of a clear line of defense. Many of the intrusion detection techniques developed on a fixed wired network are not applicable in this new environment. How to do it differently and effectively is a challenging research problem. In this paper, we first examine the vulnerabilities of a wireless ad-hoc network, the reason why we need intrusion detection, and the reason why the current methods cannot be applied directly. We then describe the new intrusion detection and response mechanisms that we are developing for wireless ad-hoc networks.

1,126 citations


Journal ArticleDOI
TL;DR: A novel framework, MADAM ID, for Mining Audit Data for Automated Models for Instrusion Detection, which uses data mining algorithms to compute activity patterns from system audit data and extracts predictive features from the patterns.
Abstract: Intrusion detection (ID) is an important component of infrastructure protection mechanisms. Intrusion detection systems (IDSs) need to be accurate, adaptive, and extensible. Given these requirements and the complexities of today's network environments, we need a more systematic and automated IDS development process rather that the pure knowledge encoding and engineering approaches. This article describes a novel framework, MADAM ID, for Mining Audit Data for Automated Models for Instrusion Detection. This framework uses data mining algorithms to compute activity patterns from system audit data and extracts predictive features from the patterns. It then applies machine learning algorithms to the audit records taht are processed according to the feature definitions to generate intrusion detection rules. Results from the 1998 DARPA Intrusion Detection Evaluation showed that our ID model was one of the best performing of all the participating systems. We also briefly discuss our experience in converting the detection models produced by off-line data mining programs to real-time modules of existing IDSs.

1,033 citations


Journal ArticleDOI
01 Oct 2000
TL;DR: This report describes new and known approaches and strategies that were used to make attacks stealthy for the 1999 DARPA Intrusion Detection Evaluation, and includes many examples of stealthy scripts that can be use to implement stealthy procedures.
Abstract: Eight sites participated in the second Defense Advanced Research Projects Agency (DARPA) off-line intrusion detection evaluation in 1999. A test bed generated live background traffic similar to that on a government site containing hundreds of users on thousands of hosts. More than 200 instances of 58 attack types were launched against victim UNIX and Windows NT hosts in three weeks of training data and two weeks of test data. False-alarm rates were low (less than 10 per day). The best detection was provided by network-based systems for old probe and old denial-of-service (DoS) attacks and by host-based systems for Solaris user-to-root (U2R) attacks. The best overall performance would have been provided by a combined system that used both host- and network-based intrusion detection. Detection accuracy was poor for previously unseen, new, stealthy and Windows NT attacks. Ten of the 58 attack types were completely missed by all systems. Systems missed attacks because signatures for old attacks did not generalize to new attacks, auditing was not available on all hosts, and protocols and TCP services were not analyzed at all or to the depth required. Promising capabilities were demonstrated by host-based systems, anomaly detection systems and a system that performs forensic analysis on file system data.

893 citations


DOI
01 Jan 2000
TL;DR: “¦e—žŒ4—&ŒŽ“2¦2“n™n¤2– ‹UŒ ¥ —S‹e¦§¯4e– ̈©“SšS–œ‹™– ‘¬Œ’¦ e-S«—S«
Abstract: ŠG‹UŒŽe‘’”“n‹e‘Y• “n‘Ž– —|‘Ž– ’”“n`‘Y‘4– ˜ ™’”Œ›šœŒ4™– —žŒ ’Ÿ‹2— ‹™– Œd “S4¡&– ‹U¢£’”“n‹e¤2– ‹UŒ ¥ —S‹e¦§Œ4e– 4– ̈©“S–œ‹™– – ¦WŒŽ“[a:–œ•e4“n¤2•`Œ4«”š[¦e– ŒŽ– ˜¬ŒŽ– ¦+—S‹e¦\¦e– —S«”Œ8 8’ŸŒ4;­8® – ̄’°‹A± ŒŽe‘4’”“n‹Œdš£• – ‘ ¥™“S ̈L e’Ÿ˜2¦™– ŒŽ– ˜¬Œ4’Ÿ“n‹ ‘ŽšA‘ŽŒŽ– ¤3‘ ¤3—)š[‹™“SŒ – ¢S– ‹ a – ́— —ž4–S¥`—ž4– Œ4™–§¤2“n‘ŽŒ¦e’”μ"˜ e«ŸŒJŒŽ“ ¦™– ŒŽ– ˜¬Œ ­·¶ ™44– ‹UŒ3‘4’” ̧n‹`—žŒ4™4– a`—S‘4– ¦1¤2– Œ4™“A¦e‘3—S‹e¦ «”– —ž‹e’Ÿ‹e ̧œ—S«” ̧S“S’ŸŒ4e¤3‘ 8e’Ÿ˜ "– «”š3“n‹ «Ÿ—ža – «”– ¦[¦e—žŒ4—&ŒŽ“2ŒŽ—S’Ÿ‹@¥U ̧S– ‹™–  —S«Ÿ«”š"˜ —S‹ ‹™“SŒ2¦™– ŒŽ– ˜¬Œ2Œ4™– ‘Ž– ‹™– o’Ÿ‹UŒŽe‘4’Ÿ“n‹e‘ ­(»+–2•e4– ‘Ž– ‹UŒ2—\ ̈1⁄4—S¤2– “S4¡ ̈©“S3—S™ŒŽ“ž± ¤3—žŒ4’Ÿ˜ —S«Ÿ«”š ¦™– ŒŽ– ˜¬Œ4’°‹™ ̧[’°‹mŒŽ`‘4’”“n‹e‘ ¥P‹™– 1⁄2“S ́“SŒ4™– 4 ’Ÿ‘Ž–S¥`– ¢S– ‹(’” ̈]Œ4™– š\—ž4–œšS– Œ e‹™¡A‹™“) 8‹3⁄4ŒŽ“ Œ4e–W‘ŽšA‘ŽŒŽ– ¤ ­^ŠG‹3⁄4“n™2‘Žš£‘4ŒŽ– ¤ ¥8‹™“%¤3—S‹£e—S«Ÿ«”š%“S[“SŒ4™– 4 8’Ÿ‘4– ˜ «Ÿ—S‘4‘4’”¿`– ¦ ¦e—žŒ4—N’Ÿ‘&‹™– ˜¬– ‘4‘4—ž4š ̈1⁄4“S2ŒŽ—S’Ÿ‹e’Ÿ‹e ̧™­ À ́™2¤2– Œ4™“A¦Á’Ÿ‘&—ža`«”–2ŒŽ“N¦e–¬± ŒŽ– ˜¬ŒÂ¤3—S‹Uš§¦e’”Ã:– 4– ‹UŒ8Œ›šU• – ‘ “S ̈/’Ÿ‹UŒŽe‘4’Ÿ“n‹e‘ ¥™ 8e’°«”–$¤3—S’Ÿ‹UŒ4—S’Ÿ‹e’°‹™ ̧3—«Ÿ“) C ̈1⁄4—S«Ÿ‘4– •:“n‘’”Œ4’”¢S– ́—žŒŽ–S­

880 citations


Proceedings ArticleDOI
01 Jan 2000
TL;DR: In this paper, an intrusion detection evaluation test bed was developed which generated normal traffic similar to that on a government site containing 100's of users on 1000's of hosts, and more than 300 instances of 38 different automated attacks were launched against victim UNIX hosts in seven weeks of training data and two weeks of test data.
Abstract: An intrusion detection evaluation test bed was developed which generated normal traffic similar to that on a government site containing 100's of users on 1000's of hosts. More than 300 instances of 38 different automated attacks were launched against victim UNIX hosts in seven weeks of training data and two weeks of test data. Six research groups participated in a blind evaluation and results were analyzed for probe, denial-of-service (DoS) remote-to-local (R2L), and user to root (U2R) attacks. The best systems detected old attacks included in the training data, at moderate detection rates ranging from 63% to 93% at a false alarm rate of 10 false alarms per day. Detection rates were much worse for new and novel R2L and DoS attacks included only in the test data. The best systems failed to detect roughly half these new attacks which included damaging access to root-level privileges by remote users. These results suggest that further research should focus on developing techniques to find new attacks instead of extending existing rule-based approaches.

747 citations


Journal ArticleDOI
Stefan Axelsson1
TL;DR: There are indications that at least some types of intrusion detection have far to go before they can attain such low false alarm rates, due to the base-rate fallacy phenomenon.
Abstract: Many different demands can be made of intrusion detection systems. An important requirement is that an intrusion detection system be effective; that is, it should detect a substantial percentage of intrusions into the supervised system, while still keeping the false alarm rate at an acceptable level. This article demonstrates that, for a reasonable set of assumptions, the false alarm rate is the limiting factor for the performance of an intrusion detection system. This is due to the base-rate fallacy phenomenon, that in order to achieve substantial values of the Bayesian detection rate P(Intrusion***Alarm), we have to achieve a (perhaps in some cases unattainably) low false alarm rate. A selection of reports of intrusion detection performance are reviewed, and the conclusion is reached that there are indications that at least some types of intrusion detection have far to go before they can attain such low false alarm rates.

627 citations


Journal ArticleDOI
TL;DR: The vast majority of security professionals would agree that real-time ID systems are not technically advanced enough to detect sophisticated cyberattacks by trained professionals, but these systems have not matured to a level where sophisticated attacks are reliably detected, verified, and assessed.
Abstract: The vast majority of security professionals would agree that real-time ID systems are not technically advanced enough to detect sophisticated cyberattacks by trained professionals. For example, during the Langley cyberattack the ID systems failed to detect substantial volumes of email bombs that crashed critical email servers. Coordinated efforts from various international locations were observed as hackers worked to understand the rules-based filter used in counterinformation operations against massive email bomb attacks [1]. At the other end of the technical spectrum, false alarms from ID systems are problematic, persistent, and preponderant. Numerous systems administrators have been the subject of an ID system reporting normal work activities as hostile actions. These types of false alarms result in financial losses to organizations when technical resources are denied access to computer systems or security resources are misdirected to investigate nonintrusion events. In addition, when systems are prone to false alarms, user confidence is marginalized and misused systems are poorly maintained and underutilized. ID systems that examine operating system audit trails, or network traffic [3, 8] and other similar detection systems, have not matured to a level where sophisticated attacks are reliably detected, verified, and assessed. Comprehensive and reliable systems are complex and the technological designs of these advanced

623 citations


Proceedings ArticleDOI
01 Jan 2000
TL;DR: There is clear evidence that state-of-the-art commercial fraud detection systems can be substantially improved in stopping losses due to fraud by combining multiple models of fraudulent transaction shared among banks.
Abstract: We describe the results achieved using the JAM distributed data mining system for the real world problem of fraud detection in financial information systems. For this domain we provide clear evidence that state-of-the-art commercial fraud detection systems can be substantially improved in stopping losses due to fraud by combining multiple models of fraudulent transaction shared among banks. We demonstrate that the traditional statistical metrics used to train and evaluate the performance of learning systems (ie. statistical accuracy or ROC analysis) are misleading and perhaps inappropriate for this application. Cost-based metrics are more relevant in certain domains, and defining such metrics poses significant and interesting research questions both in evaluating systems and alternative models, and in formalizing the problems to which one may wish to apply data mining technologies. This paper also demonstrates how the techniques developed for fraud detection can be generalized and applied to the important area of intrusion detection in networked information systems. We report the outcome of recent evaluations of our system applied to tcpdump network intrusion data specifically with respect to statistical accuracy. This work involved building additional components of JAM that we have come to call, MADAM ID (Mining Audit Data for Automated Models for Intrusion Detection). However, taking the next step to define cost-based models for intrusion detection poses interesting new research questions. We describe our initial ideas about how to evaluate intrusion detection systems using cost models learned during our work on fraud detection.

604 citations


ReportDOI
01 Jan 2000
TL;DR: A goal of this report is to provide an unbiased assessment of publicly available ID technology and it is hoped this will help those who purchase and use ID technology to gain a realistic understanding of its capabilities and limitations.
Abstract: : Attacks on the nation's computer infrastructures are a serious problem. Over the past 12 years, the growing number of computer security incidents on the Internet has reflected the growth of the Internet itself. Because most deployed computer systems are vulnerable to attack, intrusion detection (ID) is a rapidly developing field. Intrusion detection is an important technology business sector as well as an active area of research. Vendors make many claims for their products in the commercial marketplace so separating hype from reality can be a major challenge. A goal of this report is to provide an unbiased assessment of publicly available ID technology. We hope this will help those who purchase and use ID technology to gain a realistic understanding of its capabilities and limitations. The report raises issues that we believe are important for ID system (IDS) developers to address as they formulate product strategies. The report also points out relevant issues for the research community as they formulate research directions and allocate funds.

431 citations


Journal ArticleDOI
01 Oct 2000
TL;DR: The AAFID architecture and the existing prototype, as well as some design and implementation experiences and future research issues are described, which constitute a useful framework for the research and testing of intrusion detection algorithms and mechanisms.
Abstract: AAFID is a distributed intrusion detection architecture and system, developed in CERIAS at Purdue University. AAFID was the first architecture that proposed the use of autonomous agents for doing intrusion detection. With its prototype implementation, it constitutes a useful framework for the research and testing of intrusion detection algorithms and mechanisms. We describe the AAFID architecture and the existing prototype, as well as some design and implementation experiences and future research issues. ” 2000 Elsevier Science B.V. All rights reserved.

Journal ArticleDOI
TL;DR: A data mining framework for constructing intrusion detection models that uses meta-learning as a mechanism to makeintrusion detection models more effective and adaptive and uses an iterative level-wise approximation mining procedure to uncover the low frequency but important patterns.
Abstract: In this paper we describe a data mining framework for constructing intrusion detection models. The first key idea is to mine system audit data for consistent and useful patterns of program and user behavior. The other is to use the set of relevant system features presented in the patterns to compute inductively learned classifiers that can recognize anomalies and known intrusions. In order for the classifiers to be effective intrusion detection models, we need to have sufficient audit data for training and also select a set of predictive system features. We propose to use the association rules and frequent episodes computed from audit data as the basis for guiding the audit data gathering and feature selection processes. We modify these two basic algorithms to use axis attribute(s) and reference attribute(s) as forms of item constraints to compute only the relevant patterns. In addition, we use an iterative level-wise approximate mining procedure to uncover the low frequency but important patterns. We use meta-learning as a mechanism to make intrusion detection models more effective and adaptive. We report our extensive experiments in using our framework on real-world audit data.

Journal ArticleDOI
Hervé Debar1, Marc Dacier1, Andreas Wespi1
TL;DR: This paper extends the taxonomy beyond real- time intrusion detection to include additional aspects of security monitoring, such as vulnerability assessment, and introduces a taxonomy of intrusion- detection systems that highlights the various aspects of this area.
Abstract: Intrusion-detection systems aim at detecting attacks against computer systems and networks, or in general against information systems Indeed, it is difficult to provide provably secure information systems and to maintain them in such a secure state during their lifetime and utilization Sometimes, legacy or operational constraints do not even allow the definition of a fully secure information system Therefore, intrusion- detection systems have the task of monitoring the usage of such systems to detect apparition of insecure states They detect attempts and active misuse, either by legitimate users of the information systems or by external parties, to abuse their privileges or exploit security vulnerabilities In a previous paper [Computer networks 31, 805–822 (1999)], we introduced a taxonomy of intrusion- detection systems that highlights the various aspects of this area This paper extends the taxonomy beyond real- time intrusion detection to include additional aspects of security monitoring, such as vulnerability assessment

Book ChapterDOI
02 Oct 2000
TL;DR: In this paper, a high-performance, adaptive, model-based technique for attack detection, using Bayes net technology to analyze bursts of traffic, is presented, which has the attractive features of both signature based and statistical techniques.
Abstract: Inference methods for detecting attacks on information resources typically use signature analysis or statistical anomaly detection methods. The former have the advantage of attack specificity, but may not be able to generalize. The latter detect attacks probabilistically, allowing for generalization potential. However, they lack attack models and can potentially "learn" to consider an attack normal. Herein, we present a high-performance, adaptive, model-based technique for attack detection, using Bayes net technology to analyze bursts of traffic. Attack classes are embodied as model hypotheses, which are adaptively reinforced. This approach has the attractive features of both signature based and statistical techniques: model specificity, adaptability, and generalization potential. Our initial prototype sensor examines TCP headers and communicates in IDIP, delivering a complementary inference technique to an IDS sensor suite. The inference technique is itself suitable for sensor correlation.

Book ChapterDOI
02 Oct 2000
TL;DR: This article presents an attack description language that is based on logic and uses a declarative approach and the various steps of the attack process are associated to events, which may be combined using specific algebraic operators.
Abstract: This article presents an attack description language. This language is based on logic and uses a declarative approach. In the language, the conditions and effects of an attack are described with logical formulas related to the state of the target computer system. The various steps of the attack process are associated to events, which may be combined using specific algebraic operators. These elements provide a description of the attack from the point of view of the attacker. They are complemented with additional elements corresponding to the point of view of intrusion detection systems and audit programs. These detection and verification aspects provide the language user with means to tailor the description of the attack to the needs of a specific intrusion detection system or a specific environment.

Journal ArticleDOI
TL;DR: The role of IDSs in an organization's overall defensive posture is considered and guidelines for IDS deployment, operation, and maintenance are provided.
Abstract: Intrusion detection systems are an important component of defensive measures protecting computer systems and networks from abuse. This article considers the role of IDSs in an organization's overall defensive posture and provides guidelines for IDS deployment, operation, and maintenance.

Journal ArticleDOI
01 Oct 2000
TL;DR: This approach was used to improve the baseline keyword intrusion detection system used to detect user-to-root attacks in the 1998 DARPA Intrusion Detection Evaluation, reducing the false-alarm rate required to obtain 80% correct detections by two orders of magnitude.
Abstract: The most common computer intrusion detection systems detect signatures of known attacks by searching for attack-specific keywords in network traffic. Many of these systems suffer from high false-alarm rates (often hundreds of false alarms per day) and poor detection of new attacks. Poor performance can be improved using a combination of discriminative training and generic keywords. Generic keywords are selected to detect attack preparations, the actual break-in, and actions after the break-in. Discriminative training weights keyword counts to discriminate between the few attack sessions where keywords are known to occur and the many normal sessions where keywords may occur in other contexts. This approach was used to improve the baseline keyword intrusion detection system used to detect user-to-root attacks in the 1998 DARPA Intrusion Detection Evaluation. It reduced the false-alarm rate required to obtain 80% correct detections by two orders of magnitude to roughly one false alarm per day. The improved keyword system detects new as well as old attacks in this database and has roughly the same computation requirements as the original baseline system. Both generic keywords and discriminant training were required to obtain this large performance improvement.

Proceedings ArticleDOI
13 Jul 2000
TL;DR: This paper describes the components in the FIRE architecture and explains their roles, with particular attention given to explaining the benefits of data mining and how this can improve the meaningfulness of the fuzzy sets.
Abstract: The Fuzzy Intrusion Recognition Engine (FIRE) is an anomaly-based intrusion detection system that uses fuzzy logic to assess whether malicious activity is taking place on a network. It uses simple data mining techniques to process the network input data and help expose metrics that are particularly significant to anomaly detection. These metrics are then evaluated as fuzzy sets. FIRE uses a fuzzy analysis engine to evaluate the fuzzy inputs and trigger alert levels for the security administrator. This paper describes the components in the FIRE architecture and explains their roles. Particular attention is given to explaining the benefits of data mining and how this can improve the meaningfulness of the fuzzy sets. Fuzzy rules are developed for some common intrusion detection scenarios. The results of tests with actual network data and actual malicious attacks are described. The FIRE IDS can detect a wide-range of common attack types.

Book
22 Sep 2000
TL;DR: With detailed explanations and illustrative examples from his own career, Northcutt covers the topic completely, from detect evaluation, analysis, and situation handling, through the theories involved in understanding hackers, intelligence gathering, and coordinated attacks.
Abstract: From the Publisher: Written to be both a training aid and a technical reference for intrusion detection analysts, Northcutt's book contains practical experience that can't be found anywhere else. With detailed explanations and illustrative examples from his own career, Northcutt covers the topic completely, from detect evaluation, analysis, and situation handling, through the theories involved in understanding hackers, intelligence gathering, and coordinated attacks, to an arsenal of preventive and aggressive security measures.

Patent
Douglas B. Moran1
30 Aug 2000
TL;DR: In this paper, a system and method for detecting intrusions in a host system on a network is described, which comprises an analysis engine configured to use continuations and apply forward and backward chaining using rules.
Abstract: A system and method are disclosed for detecting intrusions in a host system on a network. The intrusion detection system comprises an analysis engine configured to use continuations and apply forward- and backward-chaining using rules. Also provided are sensors, which communicate with the analysis engine using a meta-protocol in which the data packet comprises a 4-tuple. A configuration discovery mechanism locates host system files and communicates the locations to the analysis engine. A file processing mechanism matches contents of a deleted file to a directory or filename, and a directory processing mechanism extracts deallocated directory entries from a directory, creating a partial ordering of the entries. A signature checking mechanism computes the signature of a file and compares it to previously computed signatures. A buffer overflow attack detector compares access times of commands and their associated files. The intrusion detection system further includes a mechanism for checking timestamps to identify and analyze forward and backward time steps in a log file.

Book ChapterDOI
02 Oct 2000
TL;DR: Eight sites participated in the second DARPA off-line intrusion detection evaluation in 1999 and best detection was provided by network-based systems for old probe and old denial-of-service (DoS) attacks and by host- based systems for Solaris user-to-root (U2R) attacks.
Abstract: Eight sites participated in the second DARPA off-line intrusion detection evaluation in 1999. Three weeks of training and two weeks of test data were generated on a test bed that emulates a small government site. More than 200 instances of 58 attack types were launched against victim UNIX and Windows NT hosts. False alarm rates were low (less than 10 per day). Best detection was provided by network-based systems for old probe and old denial-of-service (DoS) attacks and by host-based systems for Solaris user-to-root (U2R) attacks. Best overall performance would have been provided by a combined system that used both host- and network-based intrusion detection. Detection accuracy was poor for previously unseen new, stealthy, and Windows NT attacks. Ten of the 58 attack types were completely missed by all systems. Systems missed attacks because protocols and TCP services were not analyzed at all or to the depth required, because signatures for old attacks did not generalize to new attacks, and because auditing was not available on all hosts.

12 May 2000
TL;DR: A theoretical result is presented which shows that a precise analysis of the transformed program, in the general case, is NP-hard and the applicability of the techniques is demonstrated with empirical results.
Abstract: Reliable execution of software on untrustworthy platforms is a difficult problem. On the one hand, the underlying system services cannot be relied upon to provide execution assurance, while on the other hand, the effect of a tampered execution can be disastrous -- consider intrusion detection programs. What is needed, in this case, is tamper resistant software. Code obfuscation has been an area of development, in part, to enhance software tamper resistance. However, most obfuscation techniques are ad hoc, without the support of sound theoretical basis or provable results. In this paper, we address one aspect of software protection by obstructing static analysis of programs. Our techniques are based, fundamentally, on the difficulty of resolving aliases in programs. The presence of aliases has been proven to greatly restrict the precision of static data-flow analysis. Meanwhile, effective alias detection has been shown to be NP-Hard. While this represents a significant hurdle for code optimization, it provides a theoretical basis for structuring tamper-resistant programs -- systematic introduction of nontrivial aliases transforms programs to a form that yields data flow information very slowly and/or with little precision. Precise alias analysis relies on the collection of static control flow information. We further hinder the analysis by a systematic "break-down" of the program control-flow; transforming high level control transfers to indirect addressing through aliased pointers. By doing so, we have made the basic control-flow analysis into a general alias analysis problem, and the data-flow analysis and control-flow analysis are made co-dependent. We present a theoretical result which shows that a precise analysis of the transformed program, in the general case, is NP-hard and demonstrate the applicability of our techniques with empirical results.

Proceedings ArticleDOI
01 Aug 2000
TL;DR: An experimental application to network intrusion detection shows that SmartSifter was able to identify data with high scores that corresponded to attacks, with low computational costs.
Abstract: Outlier detection is a fundamental issue in data mining, specifically in fraud detection, network intrusion detection, network monitoring, etc. SmartSifter is an outlier detection engine addressing this problem from the viewpoint of statistical learning theory. This paper provides a theoretical basis for SmartSifter and empirically demonstrates its effectiveness. SmartSifter detects outliers in an on-line process through the on-line unsupervised learning of a probabilistic model (using a finite mixture model) of the information source. Each time a datum is input SmartSifter employs an on-line discounting learning algorithm to learn the probabilistic model. A score is given to the datum based on the learned model with a high score indicating a high possibility of being a statistical outlier. The novel features of SmartSifter are: (1) it is adaptive to non-stationary sources of data; (2) a score has a clear statistical/information-theoretic meaning; (3) it is computationally inexpensive; and (4) it can handle both categorical and continuous variables. An experimental application to network intrusion detection shows that SmartSifter was able to identify data with high scores that corresponded to attacks, with low computational costs. Further experimental application has identified a number of meaningful rare cases in actual health insurance pathology data from Australia's Health Insurance Commission.

01 Jan 2000
TL;DR: The history of research in intrusion detection as performed in software in the context of operating systems for a single computer, a distributed system, or a network of computers is reviewed.
Abstract: The ability to detect intruders in computer systems increases in importance as computers are increasingly integrated into the systems that we rely on for the correct functioning of society. This paper reviews the history of research in intrusion detection as performed in software in the context of operating systems for a single computer, a distributed system, or a network of computers. There are two basic approaches: anomaly detection and misuse detection. Both have been practiced since the 1980s. Both have naturally scaled to use in distributed systems and networks.

Journal ArticleDOI
TL;DR: A modification of a previously reported algorithm for mining fuzzy association rules, the concept of fuzzy frequency episodes is defined, and an original algorithm formining fuzzy frequencyepisode is presented.
Abstract: Lee, Stolfo, and Mok 1 previously reported the use of association rules and frequency episodes for mining audit data to gain knowledge for intrusion detection. The integration of association rules and frequency episodes with fuzzy logic can produce more abstract and flexible patterns for intrusion detection, since many quantitative features are involved in intrusion detection and security itself is fuzzy. We present a modification of a previously reported algorithm for mining fuzzy association rules, define the concept of fuzzy frequency episodes, and present an original algorithm for mining fuzzy frequency episodes. We add a normalization step to the procedure for mining fuzzy association rules in order to prevent one data instance from contributing more than others. We also modify the procedure for mining frequency episodes to learn fuzzy frequency episodes. Experimental results show the utility of fuzzy association rules and fuzzy frequency episodes for intrusion detection.

Proceedings ArticleDOI
25 Jan 2000
TL;DR: An overview of the intruder detection and isolation protocol (IDIP) architecture and how it supports the need for an intrusion detection and response infrastructure is provided.
Abstract: Automated response to intrusions has become a major issue in defending critical systems. Because the adversary can take actions at computer speeds, systems need the capability to react without human intervention. An infrastructure that supports development of automated response systems is critically needed. This infrastructure must allow easy integration of detection and response components to enable experimentation with automated response strategies. This paper provides an overview of the intruder detection and isolation protocol (IDIP) architecture and how it supports the need for an intrusion detection and response infrastructure.

Patent
29 Nov 2000
TL;DR: In this article, a method and a system for providing security to a network by at least identifying and unauthorized user (20) who is attempting to gain access to a node (16) on the network (12), and preferably by actively blocking that unauthorized user(20) from further activities.
Abstract: A method and a system for providing security to a network (12) by at least identifying and unauthorized user (20) who is attempting to gain access to a node (16) on the network (12), and preferably by then actively blocking that unauthorized user (20) from further activities. Detection is facilitated by the unauthorized user (20) providing 'earmark', or specially crafted false data, which the unauthorized user (20) gathers during the information collection stage performed before an attack. The earmark is designed such that any attempt by the unauthorized user (20) to use such false data results in the immediate identification of the unauthorized user (20) as hostile, and indicates that an intrusion of the network (12) is being attempted. Preferably, further access to the network (12) is then blocked by diverting traffic from the unauthorized user (20) to a secure zone (32), where the activities of the unauthorized user can be contained without damage to the network (12).

Journal ArticleDOI
TL;DR: Experimental results suggest that the partial memory learner notably reduced memory requirements at the slight expense of predictive accuracy, and tracked concept drift as well as other learners designed for this task.
Abstract: This paper describes a method for selecting training examples for a partial memory learning system. The method selects extreme examples that lie at the boundaries of concept descriptions and uses these examples with new training examples to induce new concept descriptions. Forgetting mechanisms also may be active to remove examples from partial memory that are irrelevant or outdated for the learning task. Using an implementation of the method, we conducted a lesion study and a direct comparison to examine the effects of partial memory learning on predictive accuracy and on the number of training examples maintained during learning. These experiments involved the STAGGER Concepts, a synthetic problem, and two real-world problems: a blasting cap detection problem and a computer intrusion detection problem. Experimental results suggest that the partial memory learner notably reduced memory requirements at the slight expense of predictive accuracy, and tracked concept drift as well as other learners designed for this task.

01 Jan 2000
TL;DR: After discussing the design of a network monitoring system which would maximize the potential of the self-organizing map, the experimental results in which a simpler system resoundingly detected two different exploits which the authors perpetrated against one of their servers are described.
Abstract: The Kohonen self-organizing map is an extremely powerful mechanism for automatic mathematical characterization of acceptable system activity. Because it spontaneously develops a sophisticated characterization of the system whose behaviors it is trained to recognize, it could detect intrusions which it has never observed simply by noting the degree to which they differ from normal activity. After discussing the design of a network monitoring system which would maximize the potential of the self-organizing map, we describe briefly our experimental results in which a simpler system resoundingly detected two different exploits which we perpetrated against one of our servers.

Book ChapterDOI
02 Oct 2000
TL;DR: In this paper, the authors describe an experimental system based on the Common Intrusion Detection Framework (CIDF), where multiple IDSs can exchange attack information to detect distributed intrusions.
Abstract: As the recent distributed Denial-of-Service (DDOS) attacks on several major Internet sites have shown us, no open computer network is immune from intrusions. Furthermore, intrusion detection systems (IDSs) need to be updated timely whenever a novel intrusion surfaces; and geographically distributed IDSs need to cooperate to detect distributed and coordinated intrusions. In this paper, we describe an experimental system, based on the Common Intrusion Detection Framework (CIDF), where multiple IDSs can exchange attack information to detect distributed intrusions. The system also includes an ID model builder, where a data mining engine can receive audit data of a novel attack from an IDS, compute a new detection model, and then distribute it to other IDSs. We describe our experiences in implementing such system and the preliminary results of deploying the system in an experimental network.