scp0097 commited on
Commit
c62b0e6
·
verified ·
1 Parent(s): 47bec6f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +23 -8
app.py CHANGED
@@ -1,24 +1,39 @@
1
 
2
  import gradio as gr
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
4
  import torch
5
 
6
- model_name = "muhtasham/tajik-qwen2.5-7b_finetuned"
 
7
 
8
- tokenizer = AutoTokenizer.from_pretrained(model_name)
9
- model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
 
 
 
 
 
 
 
 
 
 
10
 
11
  def chat(prompt):
12
- inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
13
- outputs = model.generate(**inputs, max_new_tokens=150, temperature=0.8, top_p=0.9)
14
- return tokenizer.decode(outputs[0], skip_special_tokens=True)
 
 
 
15
 
16
  demo = gr.Interface(
17
  fn=chat,
18
  inputs=gr.Textbox(lines=3, label="Введите запрос (на таджикском, русском или английском)"),
19
  outputs="text",
20
- title="Tajik-Qwen2.5-7B Demo",
21
- description="Нейросеть, обученная на таджикском языке — muhtasham/tajik-qwen2.5-7b_finetuned"
22
  )
23
 
24
  if __name__ == "__main__":
 
1
 
2
  import gradio as gr
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
4
+ from peft import PeftModel
5
  import torch
6
 
7
+ base_model = "Qwen/Qwen2.5-7B" # базовая модель
8
+ adapter_model = "muhtasham/tajik-qwen2.5-7b_finetuned" # таджикский fine-tune
9
 
10
+ tokenizer = AutoTokenizer.from_pretrained(base_model)
11
+
12
+ # Загружаем базовую модель
13
+ model = AutoModelForCausalLM.from_pretrained(
14
+ base_model,
15
+ torch_dtype="auto",
16
+ low_cpu_mem_usage=True,
17
+ device_map="cpu"
18
+ )
19
+
20
+ # Подключаем LoRA адаптер
21
+ model = PeftModel.from_pretrained(model, adapter_model)
22
 
23
  def chat(prompt):
24
+ try:
25
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
26
+ outputs = model.generate(**inputs, max_new_tokens=150, temperature=0.8, top_p=0.9)
27
+ return tokenizer.decode(outputs[0], skip_special_tokens=True)
28
+ except Exception as e:
29
+ return f"⚠️ Error: {str(e)}"
30
 
31
  demo = gr.Interface(
32
  fn=chat,
33
  inputs=gr.Textbox(lines=3, label="Введите запрос (на таджикском, русском или английском)"),
34
  outputs="text",
35
+ title="Tajik-Qwen2.5-7B Demo (LoRA)",
36
+ description="Файнтюнинг модели Qwen2.5-7B на таджикском языке."
37
  )
38
 
39
  if __name__ == "__main__":