Text Summarization
Collection
5 items
•
Updated
This model is a fine-tuned version of google/pegasus-multi_news.
This is a text summarization model of news articles.
For more information on how it was created, check out the following link: https://github.com/DunnBC22/NLP_Projects/blob/main/Text%20Summarization/Text_Summarization_BBC_News-Pegasus.ipynb
This model is intended to demonstrate my ability to solve a complex problem using technology.
Dataset Source: https://www.kaggle.com/datasets/pariza/bbc-news-summary
The following hyperparameters were used during training:
Unfortunately, I did not set the metrics to automatically upload here. They are as follows:
Training Loss | Epoch | Step | rouge1 | rouge2 | rougeL | rougeLsum |
---|---|---|---|---|---|---|
6.41979 | 2.0 | 214 | 0.584474 | 0.463574 | 0.408729 | 0.408431 |