Improving End-to-end Mixed-case ASR with Knowledge Distillation and Integration of Voice Activity CuesSashi NovitasariTakashi Fukudaet al.2025INTERSPEECH 2025
Voice Activity-based Text Segmentation for ASR Text DenormalizationSashi NovitasariTakashi Fukudaet al.2025INTERSPEECH 2025
Knowledge Distillation Based Training of Unified Conformer CTC Models for Multi-form ASRTakashi FukudaGakuto Kurataet al.2025ICASSP 2025
Effective Training of RNN Transducer Models on Diverse Sources of Speech and Text DataTakashi FukudaSamuel Thomas2023ICASSP 2023
Global RNN Transducer Models For Multi-dialect Speech RecognitionTakashi FukudaSamuel Thomaset al.2022INTERSPEECH 2022
Improving Generalization of Deep Neural Network Acoustic Models with Length Perturbation and N-best Based Label SmoothingXiaodong CuiGeorge Saonet al.2022INTERSPEECH 2022
Improving ASR Robustness in Noisy Condition Through VAD IntegrationSashi NovitasariTakashi Fukudaet al.2022INTERSPEECH 2022
Knowledge distillation based training of universal ASR source models for cross-lingual transferTakashi FukudaSamuel Thomas2021INTERSPEECH 2021
Generalized Knowledge Distillation from An Ensemble of Specialized Teachers Leveraging Unsupervised Neural ClusteringTakashi FukudaGakuto Kurata2021ICASSP 2021
Implicit transfer of privileged acoustic information in a generalized knowledge distillation frameworkTakashi FukudaSamuel Thomas2020INTERSPEECH 2020