Today, transformer and conformer models are commonly used in end-to-end speech recognition. Generally, conformer models are more efficient than transformers, but both suffer from large sizes, and expensive computing cost making their use environmentally unfriendly. In this paper, we propose compressing these models using quantization and pruning, evaluating size and computing time improvements while monitoring […]