Verifikasi Suara menggunakan Jaringan Syaraf Tiruan dan Ekstraksi Ciri Mel Frequency Cepstral Coefficient

Voice recording is an important part of the evidence for the suspect, so it is necessary to verify the voice suspects to prove the allegations of the suspect. The research aims to develop a voice verification system using artificial neural networks and extraction characteristics mel frequency cepstral coefficient. As the input data analyzed is the data of the unrecognized voice recorder of the owner and the recorded data of the sound that the owner has known as the comparison data. Data input is processed by feature extraction consisting of framing, windowing, fast Fourier transform, mel frequency wrapping, discrete cosine transform resulting in mel-frequency wrapping coefficient. The mel frequency wrapping coefficient of each frame in each input voice, is used as input on pattern recognition using artificial neural networks. The results of artificial neural networks are analyzed using decision logic to get a decision whether these two voices are the same or not. The output of the system is a decision that the tested sound is the same as or not with a voice comparison. Based on the level of compatibility of the test data produces a voice verification system with mel-frequency wrapping and artificial neural networks have a rate of 96% accuracy. The accuracy of the voice verification system can be an option to help resolve the issues in verification of voice recordings.
Article Metrics:
- Ahmed, N., Natarajan, T., Rao, K. R., 1974. Discrete Cosine Transform. Computers, IEEE Transactions on, C-23(1), 90–93
- Boë, L. J., 2000. Forensic voice identification in France, Speech Communication, 31(2), 205–224
- Dede, G., Sazlı, M. H., 2010. Speech recognition with artificial neural networks, Digital Signal Processing, 20(3), 763–768
- Ganchev, T., Ganchev, T., Fakotakis, N., Fakotakis, N., Kokkinakis, G., & Kokkinakis, G., 2005. Comparative evaluation of various MFCC implementations on the speaker verification task, In Proc. of the SPECOM-2005, 191–194
- Kusumadewi, S., 2003. Artificial Intelligence, Penerbit Graha Ilmu, Yogyakarta
- Ooi, C. S., Seng, K. P., Ang, L. M., Chew, L. W., 2014. A new approach of audio emotion recognition, Expert Systems with Applications, 41(13), 5858–5869
- Pal, M., Saha, G., 2015. On robustness of speech based biometric systems against voice conversion attack, Applied Soft Computing, 30, 214–228
- Riyanto, E., 2013. Speaker Recognition System with MFCC Feature Extraction and Neural Network Backpropagation, ICISBC, 62–66
- Sahidullah, M., Saha, G., 2012. Design, analysis and experimental evaluation of block based transformation in MFCC computation for speaker recognition, Speech Communication, 54(4), 543–565
- Shahamiri, S. R., Binti Salim, S. S., 2014a. Artificial neural networks as speech recognisers for dysarthric speech: Identifying the best-performing set of MFCC parameters and studying a speaker-independent approach, Advanced Engineering Informatics, 28(1), 102–110
- Shahamiri, S. R., Binti Salim, S. S., 2014b. Real-time frequency-based noise-robust Automatic Speech Recognition using Multi-Nets Artificial Neural Networks: A multi-views multi-learners approach, Neurocomputing, 129, 199–207
- Siniscalchi, S. M., Svendsen, T., Lee, C.-H., 2014. An artificial neural network approach to automatic speech processing, Neurocomputing, 140, 326–338
- Zhang, J., Ji, N., Liu, J., Pan, J., Meng, D., 2015. Enhancing performance of the backpropagation algorithm via sparse response regularization, Neurocomputing, 153, 20–40
Last update: 2021-03-01 22:47:29
Last update: 2021-03-01 22:47:31
Penulis yang mengirimkan naskah harus memahami dan menyetujui bahwa jika diterima untuk dipublikasikan, hak cipta dari artikel adalah milik JSINBIS dan Universitas Diponegoro sebagai penerbit jurnal.
Hak cipta (copyright) meliputi hak eksklusif untuk mereproduksi dan memberikan artikel dalam semua bentuk dan media, termasuk cetak ulang, foto, mikrofilm dan setiap reproduksi lain yang sejenis, serta terjemahan. Penulis mempunyai hak untuk hal-hal berikut:
- menggandakan seluruh atau sebagian materi yang dipublikasikan untuk digunakan oleh penulis sendiri sebagai bahan pengajaran di kelas atau bahan presentasi lisan dalam berbagai forum;
- menggunakan kembali sebagian atau keseluruhan materi sebagai bahan kompilasi bagi karya tulis penulis;
- membuat salinan dari bahan yang dipublikasikan untuk didistribusikan di lingkungan institusi tempat penulis bekerja.
JSINBIS dan Universitas Diponegoro serta Editor melakukan segala upaya untuk memastikan bahwa tidak ada data, pendapat atau pernyataan yang salah atau menyesatkan yang dipublikasikan di jurnal ini. Isi artikel yang diterbitkan di JSINBIS adalah tanggung jawab tunggal dan eksklusif dari masing-masing penulis.