developer-lunark commited on
Commit
9e7a09d
·
verified ·
1 Parent(s): e042c39

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +95 -0
README.md ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: Qwen/Qwen3-30B-A3B-Instruct-2507
4
+ tags:
5
+ - korean
6
+ - roleplay
7
+ - character-ai
8
+ - finetuned
9
+ - lora
10
+ - qwen3
11
+ - moe
12
+ language:
13
+ - ko
14
+ - en
15
+ pipeline_tag: text-generation
16
+ library_name: transformers
17
+ ---
18
+
19
+ # Qwen3-30B-A3B-Kaidol-v4
20
+
21
+ Korean character roleplay model fine-tuned from Qwen3-30B-A3B-Instruct-2507.
22
+
23
+ ## Model Description
24
+
25
+ This model is optimized for Korean character roleplay conversations, trained with custom character datasets featuring distinct personalities, speech patterns, and emotional expressions.
26
+
27
+ ### Training Details
28
+
29
+ | Parameter | Value |
30
+ |-----------|-------|
31
+ | Base Model | Qwen/Qwen3-30B-A3B-Instruct-2507 |
32
+ | Method | LoRA (merged) |
33
+ | LoRA Rank | 32 |
34
+ | LoRA Alpha | 64 |
35
+ | Target Modules | q_proj, k_proj, v_proj, o_proj |
36
+ | Training Epochs | 3 |
37
+ | Learning Rate | 2e-5 |
38
+ | Max Sequence Length | 2048 |
39
+
40
+ ### Target Modules
41
+ - **Attention only**: q_proj, k_proj, v_proj, o_proj
42
+
43
+ ## Intended Use
44
+
45
+ This model is designed for:
46
+ - Korean character roleplay conversations
47
+ - Interactive storytelling
48
+ - Character-based chat applications
49
+
50
+ ## Usage
51
+
52
+ ```python
53
+ from transformers import AutoModelForCausalLM, AutoTokenizer
54
+
55
+ model_name = "developer-lunark/Qwen3-30B-A3B-Kaidol-v4"
56
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
57
+ model = AutoModelForCausalLM.from_pretrained(
58
+ model_name,
59
+ torch_dtype="auto",
60
+ device_map="auto"
61
+ )
62
+
63
+ messages = [
64
+ {"role": "system", "content": "You are a helpful assistant."},
65
+ {"role": "user", "content": "안녕하세요!"}
66
+ ]
67
+
68
+ text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
69
+ inputs = tokenizer(text, return_tensors="pt").to(model.device)
70
+ outputs = model.generate(**inputs, max_new_tokens=512)
71
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
72
+ ```
73
+
74
+ ## vLLM Serving
75
+
76
+ ```bash
77
+ vllm serve developer-lunark/Qwen3-30B-A3B-Kaidol-v4 \
78
+ --tensor-parallel-size 2 \
79
+ --max-model-len 8192
80
+ ```
81
+
82
+ ## Limitations
83
+
84
+ - Optimized for Korean language; performance in other languages may vary
85
+ - Character roleplay focused; may not be optimal for factual Q&A
86
+ - Inherits limitations from the base Qwen3 model
87
+
88
+ ## License
89
+
90
+ Apache 2.0 (following the base model license)
91
+
92
+ ## Acknowledgments
93
+
94
+ - Base model: [Qwen/Qwen3-30B-A3B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507)
95
+ - Fine-tuned by [developer-lunark](https://huggingface.co/developer-lunark)