--- license: apache-2.0 datasets: - THUdyh/Oryx-Image-Data base_model: - Qwen/Qwen2-7B-Instruct pipeline_tag: text-generation --- # Oryx-7B-Image ## Model Summary The Oryx-Image models are 7/34B parameter models trained on [Oryx-Image-Data](https://huggingface.co/datasets/THUdyh/Oryx-Image-Data), based on Qwen2 language model with a context window of 32K tokens. Oryx offers an on-demand solution to seamlessly and efficiently process visual inputs with arbitrary spatial sizes and temporal lengths. - **Repository:** https://github.com/Oryx-mllm/Oryx - **Languages:** English, Chinese - **Paper:** https://arxiv.org/abs/2409.12961 ### Model Architecture - **Architecture:** Pre-trained [Oryx-ViT](https://huggingface.co/THUdyh/Oryx-ViT) + Qwen2-7B - **Data:** a mixture of 4M image data - **Precision:** BFloat16 #### Hardware & Software - **Hardware:** 64 * NVIDIA Tesla A100 - **Orchestration:** HuggingFace Trainer - **Code:** Pytorch ## Citation