Please use this identifier to cite or link to this item: https://elib.vku.udn.vn/handle/123456789/6223
Full metadata record
DC FieldValueLanguage
dc.contributor.authorHa, Minh Tan-
dc.contributor.authorLe, Dinh Nguyen-
dc.contributor.authorDang, An-
dc.date.accessioned2026-01-20T03:11:34Z-
dc.date.available2026-01-20T03:11:34Z-
dc.date.issued2026-01-
dc.identifier.isbn978-3-032-00971-5 (p)-
dc.identifier.isbn978-3-032-00972-2 (e)-
dc.identifier.urihttps://doi.org/10.1007/978-3-032-00972-2_12-
dc.identifier.urihttps://elib.vku.udn.vn/handle/123456789/6223-
dc.descriptionLecture Notes in Networks and Systems (LNNS,volume 1581); The 14th Conference on Information Technology and Its Applications (CITA 2025) ; pp: 147-158vi_VN
dc.description.abstractSpeaker extraction addresses isolating the specific speaker’s voice from a bend of other speakers using supplementary information. This paper proposes a time-domain speaker extraction using a parallel intra- and inter-framework (TSEPII). An efficient intra- and inter-architecture converts mixed utterance into multi-scale embedding coefficients. Additionally, we incorporate parallel architectures to achieve more stability than previous single architectures. This architecture includes the main components such as the auxiliary encoder (the talker encoding block), the extraction encoder (utterance encoding block), the talker extraction block, and the extraction decoder (the utterance decoding block). In particular, the time domain-based raw voice processing system keeps important information. The utterance encoding block transforms the mixed voice into multiple-scale embedding values, while the talker encoding block learns the target talker by the talker embedding feature. The talker extraction block plays an important role and uses multiple-scale embedding values and the talker embedding feature as the input features. It estimates the time-domain mask for the system. Finally, the utterance decoding block recreates the utterance of the target talker. Experiments show that the TSEPII achieves state-of-the-art performance and competes with current methods.vi_VN
dc.language.isoenvi_VN
dc.publisherSpringer Naturevi_VN
dc.subjectTarget speaker extractionvi_VN
dc.subjectInformed talker extractionvi_VN
dc.subjectTime-domain talker extractionvi_VN
dc.subjectParallel intra- and inter-frameworkvi_VN
dc.subjectDeep learningvi_VN
dc.subjectEnd-to-end deep neural networkvi_VN
dc.titleTime-Domain Target Speaker Extraction with Parallel Intra and Inter-frameworkvi_VN
dc.typeWorking Papervi_VN
Appears in Collections:CITA 2025 (International)

Files in This Item:

 Sign in to read



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.