Edit model card

This model is a fine-tuned version of BAAI/bge-large-en-v1.5 designed for the following use case:

academic research data search

How to Use

This model can be easily integrated into your NLP pipeline for tasks such as text classification, sentiment analysis, entity recognition, and more. Here's a simple example to get you started:

from sentence_transformers import SentenceTransformer
from sentence_transformers.util import cos_sim

model = SentenceTransformer(
    'fine-tuned/ArguAna-256-24-gpt-4o-2024-05-13-413991',
    trust_remote_code=True
)

embeddings = model.encode([
    'first text to embed',
    'second text to embed'
])
print(cos_sim(embeddings[0], embeddings[1]))
Downloads last month
12
Safetensors
Model size
335M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train fine-tuned/ArguAna-256-24-gpt-4o-2024-05-13-413991

Spaces using fine-tuned/ArguAna-256-24-gpt-4o-2024-05-13-413991 2