{ "id": "1806.08340", "version": "v1", "published": "2018-06-21T17:30:26.000Z", "updated": "2018-06-21T17:30:26.000Z", "title": "Interpretable Discovery in Large Image Data Sets", "authors": [ "Kiri L. Wagstaff", "Jake Lee" ], "comment": "Presented at the 2018 ICML Workshop on Human Interpretability in Machine Learning (WHI 2018), Stockholm, Sweden", "categories": [ "cs.LG", "cs.AI", "stat.ML" ], "abstract": "Automated detection of new, interesting, unusual, or anomalous images within large data sets has great value for applications from surveillance (e.g., airport security) to science (observations that don't fit a given theory can lead to new discoveries). Many image data analysis systems are turning to convolutional neural networks (CNNs) to represent image content due to their success in achieving high classification accuracy rates. However, CNN representations are notoriously difficult for humans to interpret. We describe a new strategy that combines novelty detection with CNN image features to achieve rapid discovery with interpretable explanations of novel image content. We applied this technique to familiar images from ImageNet as well as to a scientific image collection from planetary science.", "revisions": [ { "version": "v1", "updated": "2018-06-21T17:30:26.000Z" } ], "analyses": { "keywords": [ "large image data sets", "interpretable discovery", "achieving high classification accuracy rates", "image data analysis systems", "novel image content" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }