Spaces:
Runtime error
Runtime error
| import torch | |
| from transformers import LlamaTokenizer, LlamaForCausalLM, GenerationConfig | |
| import gradio as gr | |
| import warnings | |
| import os | |
| # Remove command-line arguments parsing and use hardcoded defaults for simplicity in Spaces | |
| MODEL_PATH = "/model/13B_hf" | |
| LORA_PATH = "checkpoint-3000" | |
| USE_TYPEWRITER = 1 | |
| USE_LOCAL = 1 | |
| tokenizer = LlamaTokenizer.from_pretrained(MODEL_PATH) | |
| LOAD_8BIT = True | |
| BASE_MODEL = MODEL_PATH | |
| LORA_WEIGHTS = LORA_PATH | |
| lora_bin_path = os.path.join(LORA_PATH, "adapter_model.bin") | |
| if not os.path.exists(lora_bin_path) and USE_LOCAL: | |
| # ... [rest of the path fixing logic] | |
| # ... [rest of the device and model loading logic] | |
| def generate_prompt(instruction, input=None): | |
| # ... [rest of the generate_prompt function] | |
| def evaluate( | |
| input, | |
| temperature=0.1, | |
| top_p=0.75, | |
| top_k=40, | |
| num_beams=4, | |
| max_new_tokens=128, | |
| min_new_tokens=1, | |
| repetition_penalty=2.0, | |
| **kwargs, | |
| ): | |
| # ... [rest of the evaluate function] | |
| gr.Interface( | |
| fn=evaluate, | |
| inputs=[ | |
| gr.components.Textbox(lines=2, label="Input", placeholder="Tell me about alpacas."), | |
| # ... [rest of the inputs] | |
| ], | |
| outputs=[ | |
| gr.inputs.Textbox(lines=25, label="Output"), | |
| ], | |
| title="Chinese-Vicuna 中文小羊驼", | |
| description="Chatlaw app trained on HK law data", | |
| ).launch() | |