This publication discusses the forensic interpretation and processing of user generated audio recordings (UGRs).
In this paper, researchers study analytical and practical constraints for combining and synchronizing user-generated recordings (UGRs) and develop forensic recommendations. For audio forensic analysis, it is increasingly likely that multiple UGRs may be presented as evidence in a criminal investigation. Audio evidence may come from handheld smartphones, private surveillance systems, police body cameras, and other unsynchronized recording devices. When multiple UGRs are available, the recordings could provide spatial and temporal information about the location and orientation of sound sources, and potentially a means to increase intelligibility of spoken utterances. However, UGRs generally start and stop at different times, differ in technical format specifications, and seldom have sufficiently reliable time stamp information for exact time and position synchronization. (Published Abstract Provided)
Downloads
Similar Publications
- Collaborations Between Community Corrections and Community Organizations: Understanding their Potential for System-Involved Women and Gender Responsive Programming
- Assessment of Sexual Assault Kit (SAK) Evidence Selection Leading to Development of SAK Evidence Machine-Learning Model (SAK-ML Model)
- Collaborative Exercise: Analysis of Age Estimation Using a QIAGEN Protocol and the PyroMark Q48 Platform