or: Token-Level Ensemble Distillation for Grapheme-to-Phoneme Conversion

Generates a phoneme sequence from text. Uses a large ensemble teacher model trained on a medium dataset, then uses the teacher’s inferences over unlabeled large-scale data to train a smaller student model.

(Decoding the abbreviation: grapheme-to-phoneme with knowledge distillation)