Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to...

15
Generating Test Data for Insider Threat Detectors * Brian Lindauer 1† , Joshua Glasser 2 , Mitch Rosen 2 , and Kurt Wallnau 1 1 Software Engineering Institute Carnegie Mellon University Pittsburgh, Pennsylvania, U.S.A. {lindauer, kcw}@sei.cmu.edu 2 ExactData, LLC Rochester, New York, U.S.A. {joshua.glasser, mitch.rosen}@exactdata.net Abstract The threat of malicious insider activity continues to be of paramount concern in both the public and private sectors. Though there is great interest in advancing the state of the art in predicting and stopping these threats, the difficulty of obtaining suitable data for research, development, and testing remains a significant hindrance. We outline the use of a synthetic data generator to enable research progress, while discussing the benefits and limitations of synthetic insider threat data, the meaning of realism in this context, comparisons to a hybrid real/synthetic data approach, and future research directions. Keywords: insider threat, synthetic data, modeling and simulation 1 Introduction Malicious insiders are current or former employees, or trusted partners of an organization who abuse their authorized access to an organization’s networks, systems, or data. Insider threats, the malicious acts carried out by these trusted insiders, include, but are not limited to, theft of intellectual property or national security information, fraud, and sabotage. Many government, academic, and industry groups seek to improve the dependability of their overall socio-technical systems by discovering and developing solutions to detect and protect against these insider threats. A significant impediment to these programs is the lack of data to analyze. To be useful, this data must contain a detailed account of human behavior within the monitored environment. Insider threats are, after all, about the actions of humans and not machines, and detection techniques will inevitably incorporate methods from the social sciences. But such data sets are difficult to come by. Researchers in this domain have two families of options: those requiring use or collection of real user data, and those using only synthetic data. Malicious insiders are, first and foremost, insiders, and to collect real data, an organization must directly monitor and record the behavior and actions of its own employees. Confidentiality and privacy concerns create barriers to the collection and use of such data for research purposes. Thus, it is sometimes preferable to proceed with synthetic data. With a mature synthetic data generation framework like the one used here, a user can flexibly control and rapidly and economically generate data sets with desired characteristics, size, and quality relative to measurable properties. Because they are not real, the data sets are fully intact, with no need for de- identification, and are free of privacy restrictions or limitations. Because they are generated, there is, Journal of Wireless Mobile Networks, Ubiquitous Computing, and Dependable Applications, volume: 5, number: 2, pp. 80-94 * A preliminary version of this paper appeared in the Proceedings of the IEEE Security and Privacy Workshops 2013 [1]. Corresponding author: Software Engineering Institute 4500 Fifth Avenue, Pittsburgh, PA 15213, USA, Tel: +1-512-666- 5438 80

Transcript of Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to...

Page 1: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors∗

Brian Lindauer1†, Joshua Glasser2, Mitch Rosen2, and Kurt Wallnau1

1 Software Engineering InstituteCarnegie Mellon University

Pittsburgh, Pennsylvania, U.S.A.{lindauer, kcw}@sei.cmu.edu

2 ExactData, LLCRochester, New York, U.S.A.

{joshua.glasser, mitch.rosen}@exactdata.net

Abstract

The threat of malicious insider activity continues to be of paramount concern in both the public andprivate sectors. Though there is great interest in advancing the state of the art in predicting andstopping these threats, the difficulty of obtaining suitable data for research, development, and testingremains a significant hindrance. We outline the use of a synthetic data generator to enable researchprogress, while discussing the benefits and limitations of synthetic insider threat data, the meaningof realism in this context, comparisons to a hybrid real/synthetic data approach, and future researchdirections.

Keywords: insider threat, synthetic data, modeling and simulation

1 Introduction

Malicious insiders are current or former employees, or trusted partners of an organization who abusetheir authorized access to an organization’s networks, systems, or data. Insider threats, the maliciousacts carried out by these trusted insiders, include, but are not limited to, theft of intellectual property ornational security information, fraud, and sabotage. Many government, academic, and industry groupsseek to improve the dependability of their overall socio-technical systems by discovering and developingsolutions to detect and protect against these insider threats.

A significant impediment to these programs is the lack of data to analyze. To be useful, this datamust contain a detailed account of human behavior within the monitored environment. Insider threatsare, after all, about the actions of humans and not machines, and detection techniques will inevitablyincorporate methods from the social sciences. But such data sets are difficult to come by. Researchersin this domain have two families of options: those requiring use or collection of real user data, andthose using only synthetic data. Malicious insiders are, first and foremost, insiders, and to collect realdata, an organization must directly monitor and record the behavior and actions of its own employees.Confidentiality and privacy concerns create barriers to the collection and use of such data for researchpurposes. Thus, it is sometimes preferable to proceed with synthetic data.

With a mature synthetic data generation framework like the one used here, a user can flexibly controland rapidly and economically generate data sets with desired characteristics, size, and quality relativeto measurable properties. Because they are not real, the data sets are fully intact, with no need for de-identification, and are free of privacy restrictions or limitations. Because they are generated, there is,

Journal of Wireless Mobile Networks, Ubiquitous Computing, and Dependable Applications, volume: 5, number: 2, pp. 80-94∗A preliminary version of this paper appeared in the Proceedings of the IEEE Security and Privacy Workshops 2013 [1].†Corresponding author: Software Engineering Institute 4500 Fifth Avenue, Pittsburgh, PA 15213, USA, Tel: +1-512-666-

5438

80

Page 2: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

theoretically, no limit on the length of time or number of individuals represented. Such data can bepublished, which allows other researchers to repeat experiments and compare algorithms. Also, becausetrue positives in the data set are labeled, synthetic data can enable development, quality assurance, andperformance testing in ways that may be difficult or impossible with real data.

Because of obvious sensitivity reasons, we refer to the insider threat research that motivated our datasynthesis project only as “the research program.” In light of the variety of legal, ethical, and businessissues associated with using real—even de-identified—corporate data, the research program turned toproxy data sets and synthetic data. Our task was to generate data to simulate the aggregated collectionof logs from host-based sensors distributed across all the computer workstations within a large businessor government organization over a 500-day period. Despite widespread use of synthetic data to testclassification systems [2], producing synthetic data that achieves a high level of human realism is a muchmore difficult problem [3].

In the process of performing this work, we made pragmatic choices to achieve sufficient fidelityand learned important lessons about the benefits and limitations of synthetic data in this domain. Ourpurpose in this paper is to give a simplified overview of our general approach; to highlight some of thechallenges and lessons learned about the uses and misuses of synthetic data, especially regarding the roleand meaning of realism in synthetic data; and to mention opportunities for future research.

2 Related Work

Most systems for generating cybersecurity test data focus on network traffic. These include network traf-fic generation appliances, such as BreakingPoint, and software solutions such as Swing [4] or Harpoon[5]. BreakingPoint is delivered with preloaded profiles for various types of traffic, while Swing and Har-poon are designed to learn traffic profiles by example. All three systems directly output network trafficto simulate a large number of machines. An alternate approach, used by Lincoln Labs’ LARIAT [6] andSkaion’s TGS, programatically drives applications on a virtual machine to more naturally generate thetraffic. This approach is more interesting for our problem domain, since something like these systemscould drive machines to generate other host-based sensor data. In fact, we see this in later work forthe research program where Skaion’s ConsoleUser, a relative of TGS, is used to generate activity logsfrom a host-based sensor. This later work overlays synthetic data on real data, using a strategy similarto previous visualization data generation work at Pacific Northwest National Labs [7]. We discuss therelative merits of this related approach in the Comparison to Hybrid Test Data section of this report.Unlike these systems, the generator used for the effort discussed here simulates the sensor internally anddirectly outputs the activity logs.

Regardless of whether the final data is output directly by the generator or is generated as a side effectof activity, LARIAT, TGS, and ConsoleUser follow an approach that calls for individual models foreach “agent” being simulated. This model may be a set sequence of steps or a probabilistic model fromwhich decisions are made. In the network domain, models and configurations based on real networksare already available. But the insider threat domain concerns itself with higher level user activity, and inthis case, host-based activity, and such models are not readily available. Our work required us to choosereasonable dimensions and parameters for those dimensions to generate plausible background data.

While methods for generating large graphs that correctly model real-world network properties doexist ([8] [9]), they have not yet been integrated with a low-level agent-based system such as the oneemployed here in order to generate both the desired macro-level graph properties and the desired micro-level agent actions.

As an alternative to synthetic data, some data sets collected from real user studies are available.These were, however, unsuitable to support the breadth of detection techniques for which the data was

81

Page 3: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

intended. The Schonlau data set contains command line histories of 50 users and while useful for study-ing algorithms using that particular feature, it is missing important data types for insider threat detectionincluding social structure, file accesses, content, and sentiment [10]. Ben Salem and Stolfo’s RUU dataset contains process, registry, and file access data from 18 users, along with those same data types frommasqueraders [11]. Although we did make some use of this data, and it contains more data types than theSchonlau data, it is still missing several very important dimensions. Because of the difficulty of runningsuch a study, both data sets are also based on a relatively small population of users. Previous work alsoaddresses the structure and use of, as well as measures of realism for, synthetic data. Wright discusesthe merits of synthetic data in running cybersecurity experiments that are controllable and repeatable [6].Berk examines metrics for comparing real network traffic to traffic generated by LARIAT and concludesthat simulated traffic has “substantial shortcomings” [12]. Choosing measures appropriate to our domain,we found that our synthetic data, though useful, also exhibited these substantial shortcomings.

3 Our Approach to Realism

Synthetic data will only ever be realistic in some limited number of dimensions, those being the onesthat were called out as requirements, designed in, and tested for. It will rarely be coincidentally realisticin any other dimensions. These dimensions of realism are normally defined by the System Under Test(SUT). By SUT, we refer to any procedure, process, or system that will use the data as an input. Thus,anything from a spreadsheet to an insider threat detection system that uses the data will contribute bothcriteria for realism and a judgment as to the level of realism of the resulting data. In the absence of sucha clear context, ”realism” as an acceptance criteria cannot be measured, and the effort to achieve greaterrealism becomes open ended and unfocused. While consumers of synthetic test data often wish at theoutset for realism in a large number of dimensions, the SUT’s needs usually turn out to be limited to amuch narrower set.

To illustrate the wide range of possible definitions for “realistic,” consider a hypothetical example ofsynthetic respondent data for the United States Census. Say that the Census mails out forms to people,who fill them in and send them back. Then the forms go through data capture, and the captured data getsprocessed in a variety of ways. If the data were intended to test handwriting recognition and data capture,then to be considered realistic, the data might only need to appear on forms in the correct locations. Ifthe only criterion was that the captured data matched the data to be captured, the language content couldbe nonsensical and the data would still be sufficiently realistic.

If the SUT were to integrate data type validation rules, then data would need to not only be in theright position but also be semantically meaningful at least to the extent of having the correct data type,and so on. The more business logic the data capture system integrates, the more sophisticated the dataneeds to be in order to be considered sufficiently realistic to be usable. Once past data capture, the datamight be processed in a variety of ways, each with additional filters and business logic. A single pieceof data that may be valid on its own may be inconsistent in relation to other pieces of data. Makingsynthetic data consistent in many dimensions and at many layers of abstraction is extremely difficult todo without making the consistency innate.

In our Census example, we see that if you want to test the system as a whole, you must createextremely rich and realistic data. But if you focus on some particular part of the process (e.g., onlyhandwriting recognition), much of that richness and realism will be of no value to you.

Providing data for insider threat research presents an unusual challenge for many reasons. Oneprinciple reason is that the threat domain is highly dependent on characteristics of human behavior asopposed to the more purely technical domains of network data generation or handwriting recognition.To provide a rich enough test bed for development and testing of techniques to identify those threats,

82

Page 4: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

the data needs to model the micro and macro effects of human agency. Furthermore, if consumers areto use the data not just for engineering, but for research, there is no SUT to dictate what constitutessufficient realism. In this use case, it would be desirable, though probably not realistic, to use the datafor exploration rather than simply confirmation (see Example 1). Put more concretely, synthetic data canbe useful to confirm that a system detects a particular type of anomaly when that anomaly can be definedand measured. But without a much more sophisticated model of the natural world, it cannot be used toconfirm that the given anomaly or collection of anomalies correlates to malicious insider behavior. Atbest, it can be used to confirm that the SUT can detect threats consistent with the data producers’ modelsof malicious and normal user behavior. By way of analogy to the Census example above, data can clearlybe generated to test whether the SUT can accurately recognize handwritten letters and numbers. But itwould be difficult or impossible to generate data that could be used to sufficiently prove or disprove aSUT’s ability to discern the census-taker’s state of mind based on their writing.

Confirmatory Hypothesis: The SUT detects users who rank in the top 0.1% by frequency of removableUSB drive use.

Exploratory Hypothesis: Users who rank in the top 0.1% by frequency of removable USB driveuse are more likely to be malicious insiders.Example 1: Confirmation vs. Exploration. Synthetic data is better suited to test confirmatory hypothesesthan exploratory hypotheses.

While we may wish to be able to point to real data and say ”just like that,” one must articulate mea-surable and concrete requirements for realism. In this case, there were several different data consumers,each with different metrics and requirements for realism. There was, furthermore, a tension between theexploratory use of the data, where knowledge of the SUT could further invalidate any exploratory results,and the confirmatory use of the data, where knowledge of the SUT is absolutely necessary to provideuseful test data. We resolved this tension by soliciting feedback from the detection teams about dimen-sions of realism that they would like to see added, or about unrealistic artifacts found in the synthetic datasets. Because the generator is parameterized and extensible, we were able to iteratively assess, extend,and tune its behavior.

When simulating human agents, as we do here, one must make a choice between generating ob-servable variables directly or generating those observations implicitly by modeling latent variables. Re-stricting ourselves strictly to variables observable in our context is the simplest approach, and the onemost easily provable and measurable against real data. For example, we might try to match the averagenumber of times per day an average real user copies a file to a removable drive. Driving a system such asours with latent variables means modeling the motivations, preferences, affinities, and other hidden statesof the simulated users. Generating data from a latent model requires the generator designer to assert atheory of human behavior that goes beyond a simple count of observations, and the provability of thesetheories is limited by the availability of human subject research studies. We treaded a middle ground bycombining models based on observable behaviors, models based on studies of human behavior, and mod-els based on our own unvalidated but plausible theories of human behavior. By creating causal effectsbetween these individually simple models, we are able to create a virtual world that is both consistentand seemingly complex. See Figure 3 for an example of these interdependencies. In so doing, we aimedto provide synthetic data with sufficient realism and complexity to reasonably approximate real data fordevelopment and testing purposes.

83

Page 5: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

4 Implementation

Our implementation reflects these practical constraints and tradeoffs while yielding synthetic social net-work topologies and artifacts that are rich with a variety of types of links connecting sets of individualsand entities. It employs a number of different model types including graph models of the population’ssocial structure, topic models for generating content, psychometric models for dictating behavior andpreferences, models of workplace behaviors, and others that do not fit neatly into these categories. Thefollowing are descriptions of some of the most important components of our implementation.

Dynamic Data GeneratorTM: ExactData’s Dynamic Data GeneratorTM(DDG) provides a genericcapability for creating large data record collections consisting of multiple record types, with complexconstraints and relationships within and across records, and with randomness (within the constraints)reflecting underlying population demographics. The current technology has been deliberately developedto provide a clear and well-defined mechanism to add to and extend our existing capabilities as limitationsof the current capability are encountered. Data is rarely of interest simply in isolation; there is generallya greater context from which it derives significance and meaning, and without which interpretation isextremely limited. For each test data set being generated, DDG instantiates a model universe we callGAMUT (Great Automated Model Universe for Test), illustrated in Figure 1.

Figure 1: GAMUT - Great Automated Model Universe for Test

GAMUT contains first-principal models relevant to the data set being generated. It also supportsnumerous configuration controls so as to enable customization of the data generation logic according tothe customer requirements. It gives the data the greater context the SUT requires for the data to be correctand meaningful, and to achieve the level of realism demanded. This approach also enables the creationof consistent disparate data stores—numerous output files that reflect the data from GAMUT in differentformats. Also, it is precisely because of this design that DDG is capable of producing subtle engineerederrors which are truly “needles in the haystack.” Using our GAMUT approach, the synthetic threatbehavior along with the artifacts generated by the behavior (the “needles”), within our insider threat datasets, are emergent from the initial properties assigned to the users. They are fully and naturally embedded

84

Page 6: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

within the behavior and trace artifacts from the general population of users (the “haystack”), without anynecessity for a second pass over the generated data.

Once the generative models are established within DDG, the system exposes the supported configu-ration points and makes them accessible to the operator through a graphic user interface (see Figure 2).Examples of data set characteristics that can be configured include the following:

• including or excluding malicious insiders

• demographics

• personality and psychological factors

• social network topology

• organizational structures

• work processes

• file system artifacts

• the frequency and complexity of synthetic communications such as instant messaging, email, anddocuments

Data is generated by simulating these users’ actions and interactions as they move forward in time.

Figure 2: Sample DDG Configuration Screens

Relationship Graph Model: A graph representation of a real-world relational network within or-ganizations. This graph is constructed from a variety of typed/named interconnections representingrelationships induced by organizational structure, subsets within the organization (departments), supportrelationships, and affinities derived from demographic elements. In early versions of the generator, thegraph was limited to those affinities (a “friendship graph”) and in its simplicity, we were able to matchseveral known power laws for social graphs [9]. But by adding these other elements to the constructionof the relationship model, we lost the ability to implicitly induce those power law properties.

85

Page 7: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

Asset Graph Model: A model of non-human assets such as computers, removable media, files, andprinters associated with various individuals in the organization.

Behavior Model: A model dictating how employees use assets, given their relationships, interests,affinities, and psychological profiles.

Communications Model: A probabilistic expression of the current state of organization relation-ships through a variety of electronic communications artifacts.

Topic Model: A user’s topical interests both for data consumption (i.e., web browsing) and dataproduction (i.e., email communication). In earlier versions of the generator, content was generated as abag of words. The algorithm for choosing the mix of words is inspired by Latent Dirichlet Allocation(LDA) [13], which we postulated would be one of the analysis methods the data would be targeting.Since we were treating each synthetic web page or email as a collection of words about some topic ortopics, we needed a labeled source for words associated with named topics. We chose a collection ofWikipedia articles as that source. Treating each article as a topic, we performed a TF.IDF [14] analysisto determine an approximate weighting of term importance for each topic. For each generated document,we first chose the mix of topics and the desired length of the document. Then we randomly selected thecollection of words for each topic giving more weight to the terms with the highest TF.IDF scores. As theprogram progressed, it became more important to provide grammatically coherent text, so we partiallyabandoned this method of choosing individual words for a method of choosing complete sentences, inproportion to the desired mix of topics. It should be noted that these methods of generating text couldonly be useful for testing topic analysis algorithms. Other approaches, such as sentiment analysis, wouldrequire a different strategy.

Psychometric Model: Each user is endowed at the outset with a set of static personality characteris-tics as well as a dynamic job satisfaction variable. We chose these personality variables and their affectsaccording to [15] and [16].

Decoy Model: A set of decoy files that are based on Ben Salem and Stolfo’s masquerade detectionwork [11].

Threat Scenarios: Each data set produced contains a small number of insider threat incidents, or“scenarios”. These scenarios were developed in consultation with counter-intelligence experts. Eachscenario is instantiated as synthetic data in the same form and scope as the “normal” background datagenerated for all users in the data set. For the high-level description of one of these scenarios, seeExample 2.

“A member of a group decimated by layoffs suffers a drop in job satisfaction. Angry at the company, theemployee uploads documents to Dropbox, planning to use them for personal gain.”

This scenario would result in a number of observables in the generated data:

• Data streams end for laid-off co-workers, and they disappear from the LDAP directory.

• As evidenced by logon and logoff times, subject becomes less punctual because of a drop in jobsatisfaction.

• HTTP logs show document uploads by subject to Dropbox.

Example 2: Sample Threat Scenario

Variable Interdependency: Causal relationships are established between various observable andlatent variables, creating the appearance of a complex, natural system. For example, job satisfactioninfluences punctuality, graph relationships and communications affect topic affinities, and relationships

86

Page 8: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

Figure 3: Example Variable Interdependencies. Read A→B as “A influences B.”

to laid off employees affect job satisfaction. A simplified example appears in Figure 3.Every decision point within each component is controllable, allowing both the tuning of each com-

ponent (for instance, to establish the base rate for the presence of features or the actions of entities) andestablishing triggers for, and injecting of, specific patterns or occurrences of features or the actions ofentities.

The system collects information about both observable and latent state as data is produced. Both dataand meta-data are automatically captured for ground truth analysis.

5 Successes, Failures, and Lessons Learned

The data sets we produced were shared within a large community of researchers engaged in a researchprogram to develop techniques for insider threat detection. Their analysis of our data confirmed ourbelief that simple models of social network topology, together with abstract models of user activity, cansuccessfully aid in the development of those technologies, at least to the point of integration testing andconfirmatory hypothesis testing. But their experience also showed that exploratory uses of this syntheticdata are not fruitful, and this could motivate the creation of more valid synthetic data in the future. Thefollowing are some of the most most important successes, failures, and lessons learned in this effort.

5.1 Development and Integration Testing

We confirmed that synthetic data serves well as a stub during development and integration testing. It isan excellent fit for confirming that well-defined functions of the SUT work as expected.

87

Page 9: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

5.2 No Privacy Restrictions

We confirmed the high value of data that is absolutely free of the confidentiality and privacy concernsthat come with real data. Synthetic data provides the basis for meaningful experimental results and canbe used freely by researchers for demonstrations. Research can be performed anywhere, even by thosewho might otherwise be ineligible to access the real data. There is no exposure to possible privacyviolations when using the data. Neither is there any need to de-identify the data or worry that the datacan be un-de-identified since the data is not, and never was, real.

5.3 Enhanced Collaboration

Another important benefit of synthetic data to the insider threat research community is that it enablesbroad collaboration and increases the velocity of innovation. Synthetic data provides a common basisfor all researchers, allowing them to share a common set of threat scenarios. With synthetic data, theground truth can be perfectly known; therefore, researchers can readily explore data transforms andtest algorithms. Enhanced collaboration enables design of common representations and interfaces thatpromote and enable integration of algorithms within and between groups and areas of research.

5.4 Controllable Social Network Topology

Research has shown that real-world social network graphs exhibit similar properties even when thegraphs seemingly represent very different relationships (e.g., blog posts, campaign donations, and userratings of movies). Academics interested in understanding these graphs have begun compiling ”laws”that real-world graphs tend to follow [17, 9].

While the research community found that our generated data demonstrated many important charac-teristics of realism, in our early data sets, they also found that the time-varying social network extractedfrom our data sets lacked many of these expected laws found in real social networks.

While the laws of real-world networks do seem to be surprisingly consistent across real-world do-mains, the work used to derive the laws did not study the kinds of network we are simulating (e.g., asocial network of friends within an office environment). Consequently, it may not be the case that oursimulated networks should follow the laws of other networks. However, the recurrence of these lawsacross many different networks makes a compelling suggestion that our social network would be morerealistic if it also followed these laws.

This shortcoming also illustrates another important benefit of synthetic data for insider threat re-search: controllable social network topology. With synthetic data, the social network topology can beboth controlled and perfectly known. After further tuning, our later data sets were able to exhibit threeof the social graph power laws that one would expect to find in a real social graph (see Figure 4). Asnoted in Section 4, we achieved this result only with the friendship graph in isolation. In this network offriendships, we explicitly chose node degrees that would result in the expected properties. But the ob-served communication graph was composed of both this friendship graph and a corporate organizationalgraph, reflecting the corporate reporting structure and project assignments. Imposing those additionalconstraints created a composite graph that obscured the power law properties of the friendship graph.

Where graph topology is important, future work in this area should consider approaching the problemstarting with an unlabeled graph exhibiting the desired properties, such as one produced by a Kroneckergraph generator [8]. Semantics such as friendships and project assignments would then be assignedaccording to the topology, rather than attempting to produce a macro-level topology based on agent-leveleffects. While this approach presents its own difficulties, based on our experience, it is an avenue worthexploring.

88

Page 10: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

Figure 4: The friendship graph, in isolation only, conforms to the (top to bottom) node degree, eigenvaluerank, and triangle power law properties of social graphs.

5.5 Text Mining

We had hoped to provide email, file, and web page content that would be valuable in some way toresearchers developing topic and sentiment analysis algorithms. Achieving that goal turned out to beeven more difficult than initially anticipated. To test simpler analysis algorithms, our LDA-based bag ofwords, and later bag of sentences, approach might have been sufficient. But the systems under test wishedto perform more sophisticated analyses and would have benefited from more realistic content. Well-known methods such as those built upon context-free grammars [18] and n-gram models [19] exist, butthese may still have been insufficiently realistic, and language generation systems are generally limitedto specific problem domains. We believe improved generalized artificial text generation is an importantarea for future research if truly realistic insider threat data is to be generated.

5.6 Data “Cleanliness”

Though some inconsistencies were deliberately introduced, the synthetic data produced here turned outto be far more consistent and “clean” than the real data that the program eventually targeted. Our designchoices about where to intentionally introduce inconsistencies, not surprisingly, failed to match the actualsources of inconsistency in the real data. We do not, however, regret our choice to insert intentionalinconsistencies into the data, as it prepared the data consumers for the class of problems they wouldencounter with real data. Synthetic data can be unrealistic because it presents states that would impossiblein real life, but it can also be unrealistic because it fails to take into account the inevitable flaws in thesensors collecting the data.

5.7 Inability to Verify Theories

Because we did not have access to analogous real, de-identified data, our choices for parameter valueswere largely ad hoc rather than derived from such real data. (One notable exception is the behaviorsurrounding file decoys, which was based on statistics from a published user study [11].) As a result, thedetection research teams were unable to use the data to confirm theories of how benign and malicioususers behave. Instead, the data provided a plausible development and demonstration platform for theoperation of the detection systems as a whole. To improve upon this clear deficiency in validity, future

89

Page 11: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

synthetic data sets for insider threats should strive to include both (1) normal background data basedon models and parameters derived from real data and (2) threat activity based on technical logs of realinsider threat activity. This goal is not easily attainable. Sources of (2) are rare or non-existent, but (1)combined with high confidence synthetic threat activity can still provide value in proving certain theoriesof insider threat detection.

6 Comparison to Hybrid Test Data

The data sets described above consist of synthetic background data integrated with synthetic incidentdata. This is only one of the four possible permutations of real/synthetic background and real/syntheticincident data that can form a test data set. For comparison, the Software Engineering Institute (SEI)provides “red team” data for test and evaluation of insider threat research. This red team data is a hybridof real background data together with synthetic incident data. The real data is a de-identified copy oflogs from host-based sensors and is augmented in place with synthetic incident data. In producing thesedata sets, we leverage a metaphor of drama and narrative. The real users are regarded as a pool of “user-actors” who are characterized and abstracted by a “central casting” service; user-actors are subsequently“cast” as characters in dramatic stories, which we refer to as “scenarios.” The “real world” actions ofuser-actors recorded in a data corpus are then augmented by the synthetic actions of virtual charactersrecorded performing their dramatic actions. The resulting hybrid data is presented as a performance to anaudience or analysts, as seen through the medium of the detectors under test. This approach combines theadvantages of realism of monitored user behavior and experimental control of synthetic data to delivertest data that is high in realistic social complexity, low in technical and social artifacts, and targeted tothe evaluation concerns at hand. The process is similar in many ways to a scenario-based data generationprocess used in previous work on visualization data synthesis [7].

Although hybrid data provides a more realistic environment for development and integration testing,the fact that access to it is highly restricted—in a closed lab in our case—makes using it in this fashionless practical than the fully synthetic alternative. As a result, unrestricted fully synthetic data is usuallya more appropriate choice for this type of activity. However, as discussed above, fully synthetic data isnot generally an appropriate basis for testing scientific theories. Hybrid data’s real strength over purelysynthetic data lies in the ability to verify such theories. Because all of the non-threat data comes from realusers, researchers can draw scientifically valid conclusions about the base rates of human activity andsocial network topological properties in the monitored environment. A data set, in some sense, providesonly a single point of reference for these properties, but this is no small advantage, especially in the caseof insider threat detection, where approaches seek to differentiate a small number of malicious actorsfrom a large and diverse population of benign users.

Confounds to scientific validity do exist even in hybrid data, though they are less obvious than in thefully synthetic case. As in fully synthetic data, the data producers exert full control over the actions ofthe simulated malicious insiders. The generalizability of a tool’s ability to detect the inserted threats islargely dependent on the correspondence between the inserted insider threat scenarios and future real-world instances. We mitigate this potential confound through separation of roles. Counter-intelligenceexperts on a separate team in the CERT Division of the SEI provide scenarios, which are realistic, butfictional, accounts of insider threat cases. Frequently, elements of the stories are inspired by real cases,though for sensitivity reasons, real cases are never used directly. The experts writing these scenariosdo so without detailed knowledge of the monitoring environment or properties of the users who will be“cast” into the roles in their story. Other team members take responsibility for editing scenarios—asminimally as possible—to fit the environment from which the real background data is collected. Thisgives us data that contains naturally occurring background activity, along with synthesized activity that

90

Page 12: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

leading insider threat experts have chosen to represent the types of insider threats that the tools wouldencounter in an operational deployment.

7 Obtaining the Data

The fully synthetic data produced for this project can be downloaded on the CERT website(http://www.cert.org/insider threat/datasets).

8 Conclusions and Future Work

Access to data remains a significant impediment to advancing the science of insider threat research.Manually created data sets do not provide the coverage and consistency needed to test complex systemsand come with a very high price and error rate. Production data sets cannot be disseminated withoutjeopardizing personally identifiable, proprietary, or even classified information. They are static and gen-erally in limited supply, so it is very hard to get multiple data sets with matching known characteristics.De-identified data has elements masked that make it less effective for testing while not fully eliminatingprivacy/confidentiality risks. Random synthetic data generators create absurd data sets, useless for thetype of sophisticated systems we contemplate here. Fully synthetic data cannot replace real data, but—along with other benefits—it can significantly lower the barriers to entry into research requiring suchdata and provide the type of experimental control necessary to help establish a solid scientific foundationfor such research.

However, the user of synthetic data must keep its limitations in mind. Though the data can be madequite complex, as we have shown here, it will only be realistic in those dimensions where realism isexplicitly defined and measured. Unless those dimensions of realism are made to match rigorous modelsof the natural world, the ability of researchers to use synthetic data to prove theories of human behaviorwill be limited.

We see a number of areas for future research, some of which we are already pursuing:

• Efficient Generation of Threat Scenarios. Since the threat scenarios were authored individuallyand the generator manually extended for each one, a relatively small number of truly differentthreats were produced. In ongoing work, we seek to automate the process of generating threats toprovide a greater number and variety of test cases, as well as to implement an API to incorporatethese automatically authored scenarios into the generator.

• Better Models of Realism. Continued study of models representing normal user behavior oncomputer systems as well as collection of baseline parameters from real users to drive and tunethose models in synthetic data generators. By improving the quality of the innocuous backgroundbehavior, it becomes possible to more effectively hide threat signals. Having more of these formalmodels of realism also leads to metrics for realism, and we believe measurability is key to syntheticdata progress and experimental validity.

• Model More End-User Applications. Enhance synthetic data by injecting applications and theirartifacts. Allow malicious behavior to exploit the presence of new applications.

• Improved Social Graph Generation. Further exposure and tuning of the social graph generatorto achieve both micro- and macro-level realism, and exploring the integration of other social graphgenerators (such as [8]) to extend capabilities.

91

Page 13: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

• Improved Content Creation. Further improvement of natural language generation especiallyincorporating the expression of sentiment as part of content.

References[1] J. Glasser and B. Lindauer, “Bridging the gap: A pragmatic approach to generating insider threat data,” in

Proc. of the 2013 IEEE Security and Privacy Workshops (SPW’13), San Francisco, California, USA. IEEE,May 2013, pp. 98–104.

[2] K. B. Paxton and T. Hager, “Use of synthetic data in testing administrative records systems,” in Proc. of theFederal Committee on Statistical Methodology (FCSM), Washington, D.C., USA, January 2012.

[3] P. Christen, Data Matching. Springer, 2012.[4] K. V. Vishwanath and A. Vahdat, “Swing: Realistic and responsive network traffic generation,” IEEE/ACM

Transactions on Networking, vol. 17, no. 3, pp. 712–725, June 2009.[5] J. Sommers and P. Barford, “Self-configuring network traffic generation,” in Proc. of the 4th ACM SIGCOMM

Conference on Internet Measurement (IMC 04), Taormina, Sicily, Italy. ACM, October 2004, pp. 68–81.[6] C. V. Wright, C. Connelly, T. Braje, J. C. Rabek, L. M. Rossey, and R. K. Cunningham, “Generating client

workloads and high-fidelity network traffic for controllable, repeatable experiments in computer security,” inProc. of the 13th International Conference on Recent Advances in Intrusion Detection (RAID’10), Ottawa,Ontario, Canada, LNCS, vol. 6307. Springer-Verlag, September 2010, pp. 218–237.

[7] M. A. Whiting, J. Haack, and C. Varley, “Creating realistic, scenario-based synthetic data for test and eval-uation of information analytics software,” in Proc. of the 2008 Workshop on BEyond time and errors: novelevaLuation methods for Information Visualization (BELIV’08), Florence, Italy. ACM, April 2008, p. 8.

[8] J. Leskovec, D. Chakrabarti, J. Kleinberg, C. Faloutsos, and Z. Ghahramani, “Kronecker graphs: An approachto modeling networks,” The Journal of Machine Learning Research, vol. 11, pp. 985–1042, 2010.

[9] L. Akoglu and C. Faloutsos, “RTG: a recursive realistic graph generator using random typing,” Data Miningand Knowledge Discovery, vol. 19, no. 2, pp. 194–209, October 2009.

[10] M. Schonlau, “Masquerading user data,” http://www.schonlau.net/intrusion.html, 1998, accessed: 2014-06-12.

[11] M. B. Salem and S. J. Stolfo, “Modeling user search behavior for masquerade detection,” in Proc. of the 14thInternational Conference on Recent Advances in Intrusion Detection (RAID’11), Menlo Park, California,USA, LNCS, vol. 6961. Springer-Verlag, September 2011, pp. 181–200.

[12] V. H. Berk, I. G. de Souza, and J. P. Murphy, “Generating realistic environments for cyber operationsdevelopment, testing, and training,” in SPIE Proc. of Sensors, and Command, Control, Communications, andIntelligence (C3I) Technologies for Homeland Security and Homeland Defense XI, Baltimore, Maryland,USA, vol. 8359, May 2012. [Online]. Available: http://dx.doi.org/10.1117/12.924762

[13] D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent dirichlet allocation,” The Journal of Machine LearningResearch, vol. 3, pp. 993–1022, March 2003.

[14] K. S. Jones, “A statistical interpretation of term specificity and its application in retrieval,” Journal of Docu-mentation, vol. 28, no. 1, pp. 11–21, 1972.

[15] M. Mount, R. Ilies, and E. Johnson, “Relationship of personality traits and counterproductive work behaviors:The mediating effects of job satisfaction,” Personnel Psychology, vol. 59, no. 3, pp. 591–622, Autumn 2006.

[16] M. Cullen, S. Russell, M. Bosshardt, S. Juraska, A. Stellmack, E. Duehr, and K. Jeansonne, “Five-factormodel of personality and counterproductive cyber behaviors,” 2011, poster presented at the annual conferenceof the Society for Industrial and Organizational Psychology, Chicago, Illinois, USA.

[17] M. McGlohon, L. Akoglu, and C. Faloutsos, “Weighted graphs and disconnected components: patterns anda generator,” in Proc. of the 14th ACM SIGKDD International Conference on Knowledge Discovery andData Mining (KDD’08), Las Vegas, Nevada, USA. ACM, August 2008, pp. 524–532. [Online]. Available:http://dx.doi.org/10.1145/1401890.1401955

[18] J. Stribling, M. Krohn, and D. Aguayo, “Scigen–an automatic cs paper generator,” 2005, accessed: 2014-06-12.

92

Page 14: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

[19] A. H. Oh and A. I. Rudnicky, “Stochastic natural language generation for spoken dialog systems,” ComputerSpeech & Language, vol. 16, no. 3, pp. 387–407, July–October 2002.

——————————————————————————

Author Biography

Brian Lindauer is a Research Scientist in CERT’s Science of Cybersecurity group.His research focus is on applications of machine learning and data mining to cyberse-curity problems. Prior to CERT, Lindauer worked was a senior software engineer forthe development of a the CounterStorm network anomaly detection system. Much ofhis work at CERT has centered around developing rigorous tests for anomaly detec-tion systems.

Joshua Glasser is Director of Research at ExactData, where his focus is on the cre-ation of specialized test data for BIG DATA systems. He has extensive experiencedeveloping tools and techniques for the creation of dynamic synthetic data. Prior toExactData, he has worked for many different Defense Research and Imaging compa-nies, including Honeywell, Kodak, and Xerox. He is the founder of the consultingfirm Robo Imaging Research, and he has held various teaching positions, including atClarkson University and the Rochester Institute of Technology.

Mitch Rosen is a Senior Imaging Engineer at ExactData, where his work focuses ondesign and implementation of algorithms supporting the derivation of domain specificsynthetic data. Previously he was a Research Professor at the Rochester Instituteof Technology, researching issues in high quality color imaging and non-traditionaldisplay and capture technologies. He earned his Ph.D. from RIT.

Kurt Wallnau joined the SEI in 1993 to advance the theory and practice of softwareengineering. His current interest is in the use of theories of drama and narrative tospecify threat data for testing insider threat analytics. From 2010-12 Dr. Wallnaucontributed to NSF’s XSEDE program, first in defining its software and systems en-gineering processes and later as manager of software development. Dr. Wallnau ledSEI research that combined software model checking, real-time analysis, and programgeneration to create software with predictable runtime behavior backed by verifiable

evidence. He has also led SEI advanced technology research in automated production of proof-carryingcode, and strategy-proof auctions for allocating tactical network bandwidth for data fusion.

Copyright 2014 Carnegie Mellon University

This material is based upon work funded and supported by the Department of Defense under Contract No. FA8721-05-C-0003 with Carnegie Mellon University for the operation of the Software Engineering Institute, a federally funded research anddevelopment center.

93

Page 15: Generating Test Data for Insider Threat Detectorsisyou.info/jowua/papers/jowua-v5n2-5.pdfseek to improve the dependability of their overall socio-technical systems by discovering and

Generating Test Data for Insider Threat Detectors Lindauer, Glasser, Rosen, and Wallnau

References herein to any specific commercial product, process, or service by trade name, trade mark, manufacturer, or other-wise, does not necessarily constitute or imply its endorsement, recommendation, or favoring by Carnegie Mellon University orits Software Engineering Institute.

NO WARRANTY. THIS CARNEGIE MELLON UNIVERSITY AND SOFTWARE ENGINEERING INSTITUTE MATE-RIAL IS FURNISHED ON AN “AS-IS” BASIS. CARNEGIE MELLON UNIVERSITY MAKES NO WARRANTIES OFANY KIND, EITHER EXPRESSED OR IMPLIED, AS TO ANY MATTER INCLUDING, BUT NOT LIMITED TO, WAR-RANTY OF FITNESS FOR PURPOSE OR MERCHANTABILITY, EXCLUSIVITY, OR RESULTS OBTAINED FROM USEOF THE MATERIAL. CARNEGIE MELLON UNIVERSITY DOES NOT MAKE ANY WARRANTY OF ANY KIND WITHRESPECT TO FREEDOM FROM PATENT, TRADEMARK, OR COPYRIGHT INFRINGEMENT.

This material has been approved for public release and unlimited distribution. (DM-0001351)

94