--- base_model: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 datasets: [] language: [] library_name: sentence-transformers pipeline_tag: sentence-similarity tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:521 - loss:MultipleNegativesRankingLoss widget: - source_sentence: 'Marketing digital: estrategias para redes sociales y SEO.' sentences: - AI developer with reinforcement learning skills. - Ingeniero civil con experiencia en diseño de estructuras. - Especialista en marketing digital con experiencia en campañas de Google Ads y Facebook Ads. - source_sentence: AI for speech recognition and synthesis. sentences: - Ingeniero de machine learning con habilidades en PyTorch - AI developer with speech recognition skills. - Teacher with classroom management skills. - source_sentence: Advanced CSS and responsive design. sentences: - Sort, query, and structure data in Pandas, the Python library. Describe how to model and interpret data using Python. Create basic data visualizations with Python libraries - Engineer with circuit design experience. - Front-end developer with advanced CSS and responsive web design skills. - source_sentence: PostgreSQL Database Administration Course. sentences: - Nutritionist with clinical dietetics skills. - Community manager with experience in managing social networks and creating viral content. - Database administrator with PostgreSQL experience. - source_sentence: Búsqueda, reconocimiento y captación de potenciales clientes nuevos en el sector público.Exploración de tendencias y competidores en el mercado, ajustando estrategias de comercialización.Elaborar y presentar propuestas personalizadas resaltando las ventajas de los servicios en la nube.Negociar condiciones, términos y precios con posibles clientes para garantizar la concreción de acuerdos de venta.Ofrecer asistencia posterior a la venta, resolver problemas y asegurar la satisfacción del cliente.Fomentar relaciones con clientes ya existentes, comprendiendo sus necesidades a largo plazo.Detectar oportunidades adicionales en cuentas existentes mediante la presentación de nuevas soluciones y servicios que beneficien a los clientes. sentences: - Demonstrate mastery of skills and knowledge acquired in the IBM Full Stack Software Developer Professional Certificate.. Apply understanding of common technologies related to full-stack, front-end, and back-end application development.. Explain concepts in cloud computing, web development, HTML, CSS, JavaScript, GitHub, Python and Django programming, microservices, and containers.. Analyze and troubleshoot issues in software design, development, deployment, and operations. - Digital Marketing, Media Production, Social Media, Marketing - Orador público con habilidades en presentaciones efectivas y comunicación en público --- # SentenceTransformer based on sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2) on the dataset dataset. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2) - **Maximum Sequence Length:** 128 tokens - **Output Dimensionality:** 384 tokens - **Similarity Function:** Cosine Similarity - **Training Dataset:** - dataset ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("saraleivam/GURU-trained-model1") # Run inference sentences = [ 'Búsqueda, reconocimiento y captación de potenciales clientes nuevos en el sector público.Exploración de tendencias y competidores en el mercado, ajustando estrategias de comercialización.Elaborar y presentar propuestas personalizadas resaltando las ventajas de los servicios en la nube.Negociar condiciones, términos y precios con posibles clientes para garantizar la concreción de acuerdos de venta.Ofrecer asistencia posterior a la venta, resolver problemas y asegurar la satisfacción del cliente.Fomentar relaciones con clientes ya existentes, comprendiendo sus necesidades a largo plazo.Detectar oportunidades adicionales en cuentas existentes mediante la presentación de nuevas soluciones y servicios que beneficien a los clientes.', 'Digital Marketing, Media Production, Social Media, Marketing', 'Demonstrate mastery of skills and knowledge acquired in the IBM Full Stack Software Developer Professional Certificate.. Apply understanding of common technologies related to full-stack, front-end, and back-end application development.. Explain concepts in cloud computing, web development, HTML, CSS, JavaScript, GitHub, Python and Django programming, microservices, and containers.. Analyze and troubleshoot issues in software design, development, deployment, and operations.', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 384] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` ## Training Details ### Training Dataset #### dataset * Dataset: dataset * Size: 521 training samples * Columns: anchor, positive, and negative * Approximate statistics based on the first 1000 samples: | | anchor | positive | negative | |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| | type | string | string | string | | details | | | | * Samples: | anchor | positive | negative | |:-------------------------------------------------------------|:-----------------------------------------------------------------------------|:-----------------------------------------------------------| | Introduction to Docker and containerization. | DevOps engineer with Docker and container orchestration skills. | Biologist with field research experience. | | Curso de desarrollo de aplicaciones con Vue.js. | Desarrollador web con habilidades en Vue.js. | Médico con habilidades en cardiología. | | Desarrollo de videojuegos con Godot | Desarrollador de videojuegos con experiencia en Godot | Profesor de arte con experiencia en escultura | * Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: ```json { "scale": 20.0, "similarity_fct": "cos_sim" } ``` ### Evaluation Dataset #### dataset * Dataset: dataset * Size: 131 evaluation samples * Columns: anchor, positive, and negative * Approximate statistics based on the first 1000 samples: | | anchor | positive | negative | |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| | type | string | string | string | | details | | | | * Samples: | anchor | positive | negative | |:-----------------------------------------------------------|:----------------------------------------------------------------------|:------------------------------------------------------------------------| | Swift Mobile Application Development. | iOS developer with experience in Swift and Xcode. | Psychologist with trauma therapy experience. | | Diseño de interfaces de usuario con Figma | Diseñador UX/UI con habilidades en Figma y prototipado | Ingeniero eléctrico con experiencia en sistemas de energía | | Principles of natural language understanding. | NLP engineer with natural language understanding skills. | Chef with traditional cuisine skills. | * Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: ```json { "scale": 20.0, "similarity_fct": "cos_sim" } ``` ### Training Hyperparameters #### All Hyperparameters
Click to expand - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: no - `prediction_loss_only`: True - `per_device_train_batch_size`: 8 - `per_device_eval_batch_size`: 8 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `learning_rate`: 5e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1.0 - `num_train_epochs`: 3.0 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: False - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: proportional
### Training Logs | Epoch | Step | dataset loss | |:-----:|:----:|:------------:| | 3.0 | 198 | 0.0467 | ### Framework Versions - Python: 3.10.12 - Sentence Transformers: 3.0.1 - Transformers: 4.41.2 - PyTorch: 2.3.0+cu121 - Accelerate: 0.31.0 - Datasets: 2.20.0 - Tokenizers: 0.19.1 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```