mT5-small based Azerbaijani News Summarization

In this model, Google Multilingual T5-small is fine-tuned on Azerbaijani News Summary Dataset for Summarization downstream task. The model is trained with 3 epochs, 64 batch size and 10e-4 learning rate. It took almost 12 hours on GPU instance with Ubuntu Server 20.04 LTS image in Microsoft Azure. The max news length is kept as 2048 and max summary length is determined as 128.

Deployments

© 2023 Nijat Zeynalov