Model Card: AdemGPT
General Information Model Name: AdemGPT Description: AdemGPT is a pre-trained generative language model that seeks to generate coherent and relevant text based on a wide spectrum of linguistic tasks.
Authors and Affiliations Authors: [Trat80] Affiliations: [N/A]
Model Functionality Supported Tasks: Text generation, Answering questions, Text to text, etc. Supported Languages: Mainly Spanish. Examples of Use: Generation of summaries, creative writing, informal conversation, among others.
Dataset and Training Dataset Origin: Created from multiple sources of text in Spanish (books, online articles, conversations, etc.). Dataset Size: Contains millions of examples of text in Spanish. Training Procedures: The GPT-3 architecture was used and trained for several weeks in a high-performance environment.
Model Performance Evaluation Metrics: Text coherence, precision in questions and answers, language fluency, etc. Results: Achieved high scores on text generation tests and language processing tasks.
Ethical Considerations Bias Considerations: Efforts have been made to mitigate bias, but there may be some inherent biases in the training data. Privacy and Security: The model does not store user information and caution should be taken when using it with sensitive data.
Limitations of the Model Known Limitations: Cannot provide information in other languages and may have difficulty with very specialized or technical concepts.
License and Conditions of Use License: [cc-by-nc-sa4.0] Conditions of Use: The model is available for non-commercial and educational use. It is recommended to review the license terms.
Example
Install request:
pip install requests
After that, put that in you python:
import requests import json
model_name = 'Trat80/AdemGPT'
api_token = 'tu_api_token' # You token api
input_text = "Hi! My Name Is AdemGPT!"
headers = { 'Authorization': f'Bearer {api_token}', 'Content-Type': 'application/json' }
data = { 'inputs': input_text, 'parameters': { 'max_new_tokens': 100 } }
response = requests.post(f'https://api-inference.huggingface.co/models/{model_name}', headers=headers, data=json.dumps(data))
if response.status_code == 200: generated_text = response.json().get('generated_text') print(generated_text) else: print("Error en la solicitud:", response.status_code, response.text)
- Downloads last month
- 3