|
|
|
|
|
|
|
|
import argparse |
|
|
import json |
|
|
import random |
|
|
from ltp import LTP |
|
|
from tqdm import tqdm |
|
|
|
|
|
def parse_args(): |
|
|
parser = argparse.ArgumentParser() |
|
|
parser.add_argument("--input", type=str, required=True) |
|
|
parser.add_argument("--output", type=str, required=True) |
|
|
parser.add_argument("--ltp_model", type=str, required=True) |
|
|
parser.add_argument("--basic_hanzi", type=str, default="confusion/basic_hanzi_2500.txt") |
|
|
parser.add_argument("--sound_confusion", type=str, default="confusion/sound_confusion.txt") |
|
|
parser.add_argument("--shape_confusion", type=str, default="confusion/shape_confusion.txt") |
|
|
parser.add_argument("--same_ratio", type=float, default=0.1) |
|
|
parser.add_argument("--repeat_ratio", type=float, default=0.15) |
|
|
parser.add_argument("--delete_ratio", type=float, default=0.15) |
|
|
parser.add_argument("--sound_ratio", type=float, default=0.5) |
|
|
parser.add_argument("--shape_ratio", type=float, default=0.1) |
|
|
parser.add_argument("--whitelist", type=str, default="一二三四五六七八九十") |
|
|
parser.add_argument("--seed", type=int, default=42) |
|
|
args = parser.parse_args() |
|
|
return args |
|
|
|
|
|
def isChinese(word): |
|
|
for ch in word: |
|
|
cp = ord(ch) |
|
|
if cp >= 0x4E00 and cp <= 0x9FA5: |
|
|
continue |
|
|
return False |
|
|
return True |
|
|
|
|
|
def load_hanzi(path): |
|
|
hanzi = set() |
|
|
with open(path, mode="r", encoding="utf-8") as handle: |
|
|
for line in handle: |
|
|
line = line.strip() |
|
|
assert len(line) == 1 |
|
|
hanzi.update(line) |
|
|
return hanzi |
|
|
|
|
|
def load_confusion_set(path, hanzi): |
|
|
confusion_set = {} |
|
|
with open(path, mode="r", encoding="utf-8") as handle: |
|
|
for line in handle: |
|
|
line = line.strip().split() |
|
|
if len(line) < 2: continue |
|
|
key, val = line[0], [] |
|
|
for c in line[1]: |
|
|
if c in hanzi and c not in val and c != key: |
|
|
val.append(c) |
|
|
if val: |
|
|
confusion_set[key] = val |
|
|
return confusion_set |
|
|
|
|
|
def do_mask(sent, args): |
|
|
|
|
|
cws, pos = args.ltp.pipeline(sent, tasks=["cws", "pos"], return_dict=False) |
|
|
|
|
|
n = len(cws) |
|
|
i = random.choice(range(n)) |
|
|
word = cws[i] |
|
|
if not isChinese(word): |
|
|
i = random.choice(range(n)) |
|
|
word = cws[i] |
|
|
if not isChinese(word): |
|
|
return sent |
|
|
|
|
|
p = random.random() |
|
|
p1 = args.same_ratio |
|
|
p2 = p1 + args.repeat_ratio |
|
|
p3 = p2 + args.delete_ratio |
|
|
p4 = p3 + args.sound_ratio |
|
|
p5 = p4 + args.shape_ratio |
|
|
assert abs(p5 - 1) < 0.001 |
|
|
if p < p1: |
|
|
return sent |
|
|
if p < p2: |
|
|
|
|
|
cws[i] += word |
|
|
return ''.join(cws) |
|
|
if pos[i] in ['nh', 'ns']: |
|
|
|
|
|
return sent |
|
|
chars = list(word) |
|
|
k = random.choice(range(len(word))) |
|
|
c = chars[k] |
|
|
if c in args.whitelist: |
|
|
return sent |
|
|
if p < p3: |
|
|
if len(word) < 2: |
|
|
return sent |
|
|
chars[k] = '' |
|
|
cws[i] = ''.join(chars) |
|
|
return ''.join(cws) |
|
|
if p < p4: |
|
|
if c in args.sound_set: |
|
|
chars[k] = random.choice(args.sound_set[c]) |
|
|
else: |
|
|
if c in args.shape_set: |
|
|
chars[k] = random.choice(args.shape_set[c]) |
|
|
cws[i] = ''.join(chars) |
|
|
return ''.join(cws) |
|
|
|
|
|
if __name__ == "__main__": |
|
|
args = parse_args() |
|
|
random.seed(args.seed) |
|
|
|
|
|
hanzi = load_hanzi(args.basic_hanzi) |
|
|
|
|
|
|
|
|
args.sound_set = load_confusion_set(args.sound_confusion, hanzi) |
|
|
args.shape_set = load_confusion_set(args.shape_confusion, hanzi) |
|
|
args.hanzi = list(hanzi) |
|
|
|
|
|
|
|
|
args.ltp = LTP(args.ltp_model) |
|
|
|
|
|
output = open(args.output, mode="w") |
|
|
with open(args.input, mode="r", encoding="utf-8") as handle: |
|
|
for line in tqdm(handle): |
|
|
sent = line.strip() |
|
|
if len(sent) < 4: |
|
|
continue |
|
|
source = do_mask(sent, args) |
|
|
label = int(source != sent) |
|
|
output.write( |
|
|
json.dumps({"source": source, "target": sent, "label": label}, ensure_ascii=False) |
|
|
) |
|
|
output.write("\n") |
|
|
|
|
|
|