We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f1dbfdb commit eb60ab5Copy full SHA for eb60ab5
experimental/serverless-fleets/tutorials/inferencing/src/app.py
@@ -5,18 +5,15 @@
5
from transformers import AutoTokenizer, AutoModelForCausalLM
6
7
model_name = 'tiiuae/falcon-7b-instruct'
8
-cache_dir = '/mnt/ce_shared_data'
9
10
tokenizer = AutoTokenizer.from_pretrained(
11
model_name,
12
- cache_dir=cache_dir,
13
)
14
15
model = AutoModelForCausalLM.from_pretrained(
16
17
device_map='auto',
18
low_cpu_mem_usage=False,
19
20
21
22
generator = transformers.pipeline(
0 commit comments