doberst commited on
Commit
c80968e
1 Parent(s): 11c462d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -23,7 +23,9 @@ Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://
23
  --Summarization Quality (1-5): 4 (Above Average)
24
  --Hallucinations: No hallucinations observed in test runs.
25
 
26
- For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
 
 
27
 
28
  ### Model Description
29
 
@@ -99,15 +101,15 @@ If you are using a HuggingFace generation script:
99
  inputs = tokenizer(new_prompt, return_tensors="pt")
100
  start_of_output = len(inputs.input_ids[0])
101
 
102
- # temperature: set at 0.3 for consistency of output
103
  # max_new_tokens: set at 100 - may prematurely stop a few of the summaries
104
 
105
  outputs = model.generate(
106
  inputs.input_ids.to(device),
107
  eos_token_id=tokenizer.eos_token_id,
108
  pad_token_id=tokenizer.eos_token_id,
109
- do_sample=True,
110
- temperature=0.3,
111
  max_new_tokens=100,
112
  )
113
 
 
23
  --Summarization Quality (1-5): 4 (Above Average)
24
  --Hallucinations: No hallucinations observed in test runs.
25
 
26
+ For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
27
+
28
+ Note: the Pytorch version answered 1 question with "Not Found" while the quantized version answered it correctly, hence the small difference in scores.
29
 
30
  ### Model Description
31
 
 
101
  inputs = tokenizer(new_prompt, return_tensors="pt")
102
  start_of_output = len(inputs.input_ids[0])
103
 
104
+ # temperature: set at 0.0 with do_sample=False for consistency of output
105
  # max_new_tokens: set at 100 - may prematurely stop a few of the summaries
106
 
107
  outputs = model.generate(
108
  inputs.input_ids.to(device),
109
  eos_token_id=tokenizer.eos_token_id,
110
  pad_token_id=tokenizer.eos_token_id,
111
+ do_sample=False,
112
+ temperature=0.0,
113
  max_new_tokens=100,
114
  )
115