mohamedmostafa259 commited on
Commit
e02181b
·
1 Parent(s): de49fb6

removed `@gr.cache` decorator

Browse files
Files changed (1) hide show
  1. app.py +13 -19
app.py CHANGED
@@ -14,26 +14,20 @@ CREATIVITY_SETTINGS = {
14
  2: {"top_p": 0.6, "temperature": 1.5}, # Creative
15
  }
16
 
17
- # Load model and tokenizer
18
- @gr.cache
19
- def load_model():
20
- print("Loading model...")
21
- base_model = BartForConditionalGeneration.from_pretrained(MODEL_NAME)
22
- model = PeftModel.from_pretrained(base_model, PEFT_MODEL_ID)
23
- tokenizer = BartTokenizer.from_pretrained(PEFT_MODEL_ID)
24
-
25
- # Set to eval mode
26
- model.eval()
27
-
28
- # Move to GPU if available
29
- device = "cuda" if torch.cuda.is_available() else "cpu"
30
- model.to(device)
31
-
32
- print(f"Model loaded on {device}")
33
- return model, tokenizer, device
34
 
35
- # Initialize model
36
- model, tokenizer, device = load_model()
37
 
38
  # Translation function
39
  def translate_to_emoji(text, creativity_level=1):
 
14
  2: {"top_p": 0.6, "temperature": 1.5}, # Creative
15
  }
16
 
17
+ # Load model and tokenizer at startup (automatically cached)
18
+ print("Loading model...")
19
+ base_model = BartForConditionalGeneration.from_pretrained(MODEL_NAME)
20
+ model = PeftModel.from_pretrained(base_model, PEFT_MODEL_ID)
21
+ tokenizer = BartTokenizer.from_pretrained(PEFT_MODEL_ID)
22
+
23
+ # Set to eval mode
24
+ model.eval()
25
+
26
+ # Move to GPU if available
27
+ device = "cuda" if torch.cuda.is_available() else "cpu"
28
+ model.to(device)
 
 
 
 
 
29
 
30
+ print(f"Model loaded on {device}")
 
31
 
32
  # Translation function
33
  def translate_to_emoji(text, creativity_level=1):