chevron_left All Data Sets

Keystroke Dynamics - Benchmark Data Set

Accompaniment to "Comparing Anomaly-Detection Algorithms for Keystroke Dynamics" (DSN-2009)

Author: Kevin Killourhy and Roy Maxion (Contact Me)

This webpage is a benchmark data set for keystroke dynamics. It is a supplement to the paper "Comparing Anomaly-Detection Algorithms for Keystroke Dynamics," by Kevin Killourhy and Roy Maxion, published in the proceedings of the DSN 2009 conference [1]. The webpage is organized as follows:

Sections 1 – 4 each consist of a brief explanation of their contents, followed by a list of common questions that provide more detail about the material. Click on a question to show the answer, or display all answers by clicking on:

1. Introduction

On this webpage, we share the data, scripts, and results of our evaluation so that other researchers can use the data, reproduce our results, and extend them; or, use the data for investigations of related topics, such as intrusion, masquerader or insider detection. We hope these resources will be useful to the research community.

Common Questions

Click to expand or collapse the answers.

2. The Data

The data consist of keystroke-timing information from 51 subjects (typists), each typing a password (.tie5Roanl) 400 times.

get_app
DSL-StrongPasswordData.txt(Fixed-width format)
MD5 hash = e5b72954c2e093a0a4ec7ca1485f9d05
get_app
DSL-StrongPasswordData.csv(Comma-separated-value format)
MD5 hash = 470235f96568f28f9ea0da62234ec857
get_app
DSL-StrongPasswordData.xls(Excel format)
MD5 hash = e1a69b03315664d5dcaefd52583d6ad9

Common Questions

3. Evaluation Scripts

The following procedure—written in the R language for statistical computing (www.r-project.org) — demonstrates how to use the data to evaluate three anomaly detectors (called Euclidean, Manhattan, and Mahalanobis).

get_app
evaluation-script.R(R script)

Note that this script depends on the R package ROCR for generating ROC curves [2].

Common questions

4. Table of Results

The following table ranks 14 anomaly detectors based on their average equal-error rates. The evaluation procedure described in the script above was used to obtain the equal-error rates for each anomaly detector. For example, the average equal-error rate for the scaled Manhattan detector (across all subjects) was 9.62%, and the standard deviation was 0.0694.

Detector Average Equal-Error Rate (stddev)
Manhattan (scaled) 0.0962 (0.0694)
Nearest Neighbor (Mahalanobis) 0.0996 (0.0642)
Outlier Count (z-score) 0.1022 (0.0767)
SVM (one-class) 0.1025 (0.0650)
Mahalanobis 0.1101 (0.0645)
Mahalanobis (normed) 0.1101 (0.0645)
Manhattan (filter) 0.1360 (0.0828)
Manhattan 0.1529 (0.0925)
Neural Network (auto-assoc) 0.1614 (0.0797)
Euclidean 0.1706 (0.0952)
Euclidean (normed) 0.2153 (0.1187)
Fuzzy Logic 0.2213 (0.1051)
k Means 0.3722 (0.1391)
Neural Network (standard) 0.8283 (0.1483)

Note that these are results are fractional rates between 0.0 and 1.0 (not percentages between 0% and 100%).

Common Questions

5. References

1

Kevin S. Killourhy and Roy A. Maxion. "Comparing Anomaly Detectors for Keystroke Dynamics," in Proceedings of the 39th Annual International Conference on Dependable Systems and Networks (DSN-2009), pages 125-134, Estoril, Lisbon, Portugal, June 29-July 2, 2009. IEEE Computer Society Press, Los Alamitos, California, 2009. (pdf)


2

T. Sing, O. Sander, N. Beerenwinkel, T. Lengauer. "ROCR: visualizing classifier performance in R," Bioinformatics 21(20):3940-3941 (2005). (link)

arrow_upward Top