songs1 commited on
Commit
6c4b07c
·
1 Parent(s): 0f9c95c
Files changed (3) hide show
  1. app.py +2 -13
  2. mymodule/__init__.py +0 -0
  3. mymodule/utils.py +17 -0
app.py CHANGED
@@ -1,21 +1,10 @@
1
- import os
2
-
3
  import gradio as gr
4
  import spaces
5
  from guidance import gen
6
  from guidance.models import Transformers
7
- from transformers import AutoModelForCausalLM, AutoTokenizer
8
 
9
- AUTH_TOKEN = os.environ.get("HF_TOKEN", False)
10
- model = AutoModelForCausalLM.from_pretrained(
11
- "meta-llama/Llama-3.2-3B-Instruct",
12
- token=AUTH_TOKEN,
13
- )
14
- tok = AutoTokenizer.from_pretrained(
15
- "meta-llama/Llama-3.2-3B-Instruct",
16
- token=AUTH_TOKEN,
17
- )
18
- model = model.to("cuda").eval()
19
 
20
 
21
  @spaces.GPU(duration=10)
 
 
 
1
  import gradio as gr
2
  import spaces
3
  from guidance import gen
4
  from guidance.models import Transformers
5
+ from mymodule.utils import load_model_and_tok
6
 
7
+ model, tok = load_model_and_tok()
 
 
 
 
 
 
 
 
 
8
 
9
 
10
  @spaces.GPU(duration=10)
mymodule/__init__.py ADDED
File without changes
mymodule/utils.py ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+
3
+ from transformers import AutoModelForCausalLM, AutoTokenizer
4
+
5
+
6
+ def load_model_and_tok():
7
+ AUTH_TOKEN = os.environ.get("HF_TOKEN", False)
8
+ model = AutoModelForCausalLM.from_pretrained(
9
+ "meta-llama/Llama-3.2-3B-Instruct",
10
+ token=AUTH_TOKEN,
11
+ )
12
+ tok = AutoTokenizer.from_pretrained(
13
+ "meta-llama/Llama-3.2-3B-Instruct",
14
+ token=AUTH_TOKEN,
15
+ )
16
+ model = model.to("cuda").eval()
17
+ return model, tok