AdamyaG commited on
Commit
ae4a903
1 Parent(s): 06f46b9

Update chatbot.py

Browse files
Files changed (1) hide show
  1. chatbot.py +6 -6
chatbot.py CHANGED
@@ -5,11 +5,11 @@ import random
5
  from threading import Thread
6
  from typing import List, Dict, Union
7
  import subprocess
8
- subprocess.run(
9
- "pip install flash-attn --no-build-isolation",
10
- env={"FLASH_ATTENTION_SKIP_CUDA_BUILD": "TRUE"},
11
- shell=True,
12
- )
13
  import torch
14
  import gradio as gr
15
  from bs4 import BeautifulSoup
@@ -31,7 +31,7 @@ model_id = "llava-hf/llava-interleave-qwen-7b-hf"
31
  # model_id = "llava-hf/llava-interleave-qwen-7b-dpo-hf"
32
  processor = LlavaProcessor.from_pretrained(model_id)
33
  model = LlavaForConditionalGeneration.from_pretrained(model_id,torch_dtype=torch.float16, use_flash_attention_2=True)
34
- model.to("cuda")
35
  # Credit to merve for code of llava interleave qwen
36
 
37
  GROQ_API_KEY = os.environ.get("GROQ_API_KEY", None)
 
5
  from threading import Thread
6
  from typing import List, Dict, Union
7
  import subprocess
8
+ # subprocess.run(
9
+ # "pip install flash-attn --no-build-isolation",
10
+ # env={"FLASH_ATTENTION_SKIP_CUDA_BUILD": "TRUE"},
11
+ # shell=True,
12
+ # )
13
  import torch
14
  import gradio as gr
15
  from bs4 import BeautifulSoup
 
31
  # model_id = "llava-hf/llava-interleave-qwen-7b-dpo-hf"
32
  processor = LlavaProcessor.from_pretrained(model_id)
33
  model = LlavaForConditionalGeneration.from_pretrained(model_id,torch_dtype=torch.float16, use_flash_attention_2=True)
34
+ model.to("cpu")
35
  # Credit to merve for code of llava interleave qwen
36
 
37
  GROQ_API_KEY = os.environ.get("GROQ_API_KEY", None)