We use cookies to improve your experience with our site.
Shuzheng Liu, Shuai Ma, Hanching Chen, Lizhen Cui, Jie Ding. Combining KNN with AutoEncoder for Outlier Detection[J]. Journal of Computer Science and Technology. DOI: 10.1007/s11390-023-2403-y
Citation: Shuzheng Liu, Shuai Ma, Hanching Chen, Lizhen Cui, Jie Ding. Combining KNN with AutoEncoder for Outlier Detection[J]. Journal of Computer Science and Technology. DOI: 10.1007/s11390-023-2403-y

Combining KNN with AutoEncoder for Outlier Detection

  • K-nearest neighbor (KNN) is one of the most fundamental methods for unsupervised outlier detection because of its various advantages, e.g., ease of use and relatively high accuracy. Currently, most data analytic tasks need to deal with high dimensional data, and the KNN based methods often fail due to “the curse of dimensionality”. AutoEncoder based methods have recently been introduced to use reconstruction errors for outlier detection on high dimensional data, but the direct use of AutoEncoder typically does not preserve well the data proximity relationships for outlier detection. In this study, we propose to combine KNN with AutoEncoder for outlier detection. First, we propose the nearest neighbor AutoEncoder (NNAE) by persevering the original data proximity in a much lower dimension that is more suitable for performing KNN. Second, we propose the k-nearest reconstruction neighbors (KNRN) by incorporating the reconstruction errors of NNAE with the k-distances of KNN to detect outliers. Third, we develop a method to automatically choose better parameters for optimizing the structure of NNAE. Finally, using five real-life datasets, we experimentally show that our proposed approach NNAE+KNRN is much better than existing methods, i.e., KNN, Isolation Forest, AutoEncoder-RE and Robust AutoEncoder.
  • loading

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return