Edit model card

This is a ruRoberta-large model trained on the RuCoLa dataset. It can be used to classify Russian sentences into fluent or non-fluent ones, where fluency is understood as linguistic acceptability.

Training notebook: task_oriented_TST/fluency/rucola_classifier_v1.ipynb (in a private repo).

Training parameters:

  • optimizer: Adam
  • lr=2e-6
  • batch_size=32
  • epochs=10
  • clip_grad_norm=1.0

Test accuracy (on the leaderboard this model is submitted as ruroberta-base-cased-rucola-v1): 0.81.

Downloads last month
781
Safetensors
Model size
355M params
Tensor type
I64
·
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.