--- language: en tags: - bert - sequence-classification - mrpc - paraphrase license: mit --- # Model description Fine-tuned version of `bert-base-uncased` on the Microsoft Research Paraphrase Corpus (MRPC) dataset for paraphrase detection using the MRPC dataset. ## Intended uses & limitations This model is intended for paraphrase detection tasks, particularly those similar to the MRPC dataset. It may not perform well on substantially different datasets or tasks. ## Training and evaluation data The model was trained on the MRPC dataset, which contains 5,801 sentence pairs extracted from news sources on the web. 3,900 pairs were labeled as paraphrases by human annotators. ## Training procedure The model was fine-tuned using the Hugging Face Transformers library. We used a batch size of 16, learning rate of 2e-5, and trained for 3 epochs. ## Evaluation results The model achieved the following results on the MRPC validation set: - Accuracy: 0.8480 - F1 Score: 0.8927