Residential College | false |
Status | 即將出版Forthcoming |
CriSp: Leveraging Tread Depth Maps for Enhanced Crime-Scene Shoeprint Matching | |
Shafique, Samia1; Kong, Shu2,3,4; Fowlkes, Charless1 | |
2025 | |
Conference Name | 18th European Conference on Computer Vision, ECCV 2024 |
Source Publication | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
![]() |
Volume | 15121 LNCS |
Pages | 217-235 |
Conference Date | 29 September 2024 to 4 October 2024 |
Conference Place | Milan; Italy |
Publisher | Springer Science and Business Media Deutschland GmbH |
Abstract | Shoeprints are a common type of evidence found at crime scenes and are used regularly in forensic investigations. However, existing methods cannot effectively employ deep learning techniques to match noisy and occluded crime-scene shoeprints to a shoe database due to a lack of training data. Moreover, all existing methods match crime-scene shoeprints to clean reference prints, yet our analysis shows matching to more informative tread depth maps yields better retrieval results. The matching task is further complicated by the necessity to identify similarities only in corresponding regions (heels, toes, etc.) of prints and shoe treads. To overcome these challenges, we leverage shoe tread images from online retailers and utilize an off-the-shelf predictor to estimate depth maps and clean prints. Our method, named CriSp, matches crime-scene shoeprints to tread depth maps by training on this data. CriSp incorporates data augmentation to simulate crime-scene shoeprints, an encoder to learn spatially-aware features, and a masking module to ensure only visible regions of crime-scene prints affect retrieval results. To validate our approach, we introduce two validation sets by reprocessing existing datasets of crime-scene shoeprints and establish a benchmarking protocol for comparison. On this benchmark, CriSp significantly outperforms state-of-the-art methods in both automated shoeprint matching and image retrieval tailored to this task. (code and dataset at https://github.com/Samia067/CriSp. |
Keyword | Forensics Image Retrieval Shoeprint Matching |
DOI | 10.1007/978-3-031-73036-8_13 |
URL | View the original |
Language | 英語English |
Scopus ID | 2-s2.0-85210807373 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | University of Macau |
Affiliation | 1.University of California, Irvine, United States 2.University of Macau, Macau, China 3.Institute of Collaborative Innovation, Macau, China 4.Texas A&M University, College Station, United States |
Recommended Citation GB/T 7714 | Shafique, Samia,Kong, Shu,Fowlkes, Charless. CriSp: Leveraging Tread Depth Maps for Enhanced Crime-Scene Shoeprint Matching[C]:Springer Science and Business Media Deutschland GmbH, 2025, 217-235. |
APA | Shafique, Samia., Kong, Shu., & Fowlkes, Charless (2025). CriSp: Leveraging Tread Depth Maps for Enhanced Crime-Scene Shoeprint Matching. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 15121 LNCS, 217-235. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment