GitXplorerGitXplorer
y

mt5-small-Arabic-Summarization

public
0 stars
0 forks
0 issues

Commits

List of commits on branch main.
Verified
e29e427615edca5a716df6e1a2d9d813b92fb10f

Update README.md

yyalsaffar committed a year ago
Unverified
44a6321ed1c244012d481e8b86ca00c008372f21

adding requirements

yyalsaffar committed a year ago
Unverified
ac4045d599ded74f123ba000dac63a9dc67f754c

Update README.md

yyalsaffar committed a year ago
Unverified
7cd74350417a3b2edcfd3bcdaa25841cc0f39ebd

adding files

yyalsaffar committed a year ago
Unverified
b8a704e37edfbbca169424a1257bd065a35a0da6

first commit

yyalsaffar committed a year ago

README

The README file for this repository.

Arabic News Article Summarization with MT5

This project fine-tunes the google/mt5-small model on the BBC Arabic news dataset for the task of summarization. It involves summarizing news articles into concise summaries, utilizing a Transformer-based model for its state-of-the-art performance in natural language understanding and generation tasks.

Introduction

Leveraging the power of the google/mt5-small model, this project aims at harnessing its multilingual processing capabilities to address the unique linguistic nuances of Arabic. The project utilizes the Transformers library for fine-tuning the model on the BBC Arabic news dataset, focusing on summarization tasks. Through custom training configurations and optimizations, the model is enhanced to generate accurate and concise summaries of Arabic news articles. The Seq2Seq training framework supports an efficient training loop, and by employing evaluation metrics such as ROUGE scores, the project ensures the summaries' quality closely matches that of human experts.

Installation

To set up this project for use or development, clone the repository and install the required packages:

git clone https://github.com/yalsaffar/mt5-small-Arabic-Summarization.git
cd mt5-small-Arabic-Summarization
pip install -r requirements.txt

Hugging Face Implementation

The fine-tuned model is available on Hugging Face, providing an easy way to integrate and use the summarization model in various applications. The model can be accessed and utilized directly through the Hugging Face platform.

Usage

There are two ways to replicate the model fine-tuning:

  1. Using the Command Line:
python train_mt5.py --path_to_data [path_to_data] --model_name google/mt5-small --output_dir [output_directory] --wandb_key [your_wandb_key]
  1. Using the Jupyter Notebook: Open the training_mt5.ipynb notebook and follow the instructions within to fine-tune the model interactively.

Dataset

The dataset used for training is the BBC Arabic news dataset, comprising news articles and their corresponding summaries. The dataset is split into 32,000 training rows, 4,000 testing, and 4,000 validation rows.

Model

google/mt5-small represents a significant advancement in the domain of multilingual natural language processing. Originating from the T5 (Text-to-Text Transfer Transformer) family, mT5 extends the capabilities of T5 to a global scale by being pre-trained on a diverse dataset covering 101 languages, including Arabic​​​​.

License

This project is licensed under the MIT License - see the LICENSE file for details.