Update README.md
Browse files
README.md
CHANGED
@@ -4,11 +4,9 @@
|
|
4 |
|
5 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/645364cbf666f76551f93111/ZviQjj2NvCvl0R7IZiRai.png)
|
6 |
|
7 |
-
#### Welcome to the FeynModel repository, a Vision Language model with the
|
8 |
|
9 |
-
####
|
10 |
-
|
11 |
-
#### It use a S6 block to wire context memory for Q* TS (experimental)
|
12 |
|
13 |
# how to use
|
14 |
|
|
|
4 |
|
5 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/645364cbf666f76551f93111/ZviQjj2NvCvl0R7IZiRai.png)
|
6 |
|
7 |
+
#### Welcome to the FeynModel repository, a Vision Language model with the reasoning capabilities of an LLM (Large Language Model). It aims to explore the combined power of vision and language for scientific reasoning tasks. This model is fine-tuned using the LoRA (Low-Rank Adaptation) method, optimizing it for enhanced performance in a variety of vision and language tasks.
|
8 |
|
9 |
+
#### Version 0.1 utilizes pretrained layers from the DaVit Vision Tower of Florence2-base (Microsoft) and Gemma2-2B (Google), and was fine-tuned on M3IT, COCO, and ScienceQA datasets. It employs an S6 block to integrate context memory for Q*TS (experimental).
|
|
|
|
|
10 |
|
11 |
# how to use
|
12 |
|