isayahc commited on
Commit
2ccbf4d
·
1 Parent(s): 463e62a

attempt to fix memory error

Browse files
Files changed (1) hide show
  1. app.py +12 -7
app.py CHANGED
@@ -32,18 +32,23 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
32
 
33
  embeddings = HuggingFaceHubEmbeddings()
34
 
35
- # model_id = "HuggingFaceH4/zephyr-7b-beta"
36
- model_id = "meta-llama/Llama-2-7b-chat-hf"
37
 
38
- model = AutoModelForCausalLM.from_pretrained(
39
- model_id,
40
- device_map="auto",
41
- low_cpu_mem_usage=True
42
- )
43
 
44
  print( "initalized model")
45
 
 
 
 
 
46
  tokenizer = AutoTokenizer.from_pretrained(model_id)
 
47
 
48
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, max_new_tokens=10)
49
  hf = HuggingFacePipeline(pipeline=pipe)
 
32
 
33
  embeddings = HuggingFaceHubEmbeddings()
34
 
35
+ model_id = "HuggingFaceH4/zephyr-7b-beta"
36
+ # model_id = "meta-llama/Llama-2-7b-chat-hf"
37
 
38
+ # model = AutoModelForCausalLM.from_pretrained(
39
+ # model_id,
40
+ # device_map="auto",
41
+ # low_cpu_mem_usage=True
42
+ # )
43
 
44
  print( "initalized model")
45
 
46
+ # tokenizer = AutoTokenizer.from_pretrained(model_id)
47
+
48
+
49
+
50
  tokenizer = AutoTokenizer.from_pretrained(model_id)
51
+ model = AutoModelForCausalLM.from_pretrained(model_id)
52
 
53
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, max_new_tokens=10)
54
  hf = HuggingFacePipeline(pipeline=pipe)