Edit model card

xlm-roberta-longformer-base-16384

xlm-roberta-longformer is a multilingual Longformer initialized with XLM-RoBERTa's weights without further pretraining. It is intended to be fine-tuned on a downstream task.

Model attention_window hidden_size num_hidden_layers model_max_length
base 256 768 12 16384
large 512 1024 24 16384

Framework versions

  • Transformers 4.26.0
  • TensorFlow 2.11.0
  • Tokenizers 0.13.2
Downloads last month
82
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.