Most source recording device identification models for Web media forensics are based on a single feature to complete the identification task and often have the disadvantages of long time and poor accuracy. The purpose of this paper is to propose a new method for end-to-end network source identification of multi-feature fusion devices.
This paper proposes an efficient multi-feature fusion source recording device identification method based on end-to-end and attention mechanism, so as to achieve efficient and convenient identification of recording devices of Web media forensics.
The authors conducted sufficient experiments to prove the effectiveness of the models that they have proposed. The experiments show that the end-to-end system is improved by 7.1% compared to the baseline i-vector system, compared to the authors’ previous system, the accuracy is improved by 0.4%, and the training time is reduced by 50%.
With the development of Web media forensics and internet technology, the use of Web media as evidence is increasing. Among them, it is particularly important to study the authenticity and accuracy of Web media audio.
This paper aims to promote the development of source recording device identification and provide effective technology for Web media forensics and judicial record evidence that need to apply device source identification technology.
This research was supported by National Natural Science Foundation of China (No.61901165, 61501199), Science and Technology Research Project of Hubei Education Department (No. Q20191406), Hubei Natural Science Foundation (No. 2017CFB683) and self-determined research funds of CCNU from the colleges basic research and operation of MOE (No. CCNU20ZT010).
Zeng, C., Zhu, D., Wang, Z., Wang, Z., Zhao, N. and He, L. (2020), "An end-to-end deep source recording device identification system for Web media forensics", International Journal of Web Information Systems, Vol. 16 No. 4, pp. 413-425. https://doi.org/10.1108/IJWIS-06-2020-0038Download as .RIS
Emerald Publishing Limited
Copyright © 2020, Emerald Publishing Limited