| import os |
| import json |
| import time |
| import random |
| from typing import Dict, Any, List, Optional, Tuple |
|
|
| from openai import OpenAI |
| import multiprocessing as mp |
|
|
| |
| |
| |
| def load_json(file_path: str): |
| with open(file_path, "r", encoding="utf-8") as f: |
| return json.load(f) |
|
|
| def append_jsonl(path: str, obj: Dict[str, Any]): |
| os.makedirs(os.path.dirname(path), exist_ok=True) |
| with open(path, "a", encoding="utf-8") as f: |
| f.write(json.dumps(obj, ensure_ascii=False) + "\n") |
|
|
| def load_done_keys(jsonl_path: str): |
| done = set() |
| if not os.path.exists(jsonl_path): |
| return done |
| with open(jsonl_path, "r", encoding="utf-8") as f: |
| for line in f: |
| line = line.strip() |
| if not line: |
| continue |
| try: |
| obj = json.loads(line) |
| k = obj.get("key") |
| if k is not None: |
| done.add(k) |
| except Exception: |
| continue |
| return done |
|
|
| |
| |
| |
| def safe_json_parse(text: str) -> Optional[Dict[str, Any]]: |
| text = (text or "").strip() |
| if text.startswith("```"): |
| text = text.strip("`").strip() |
| if text.lower().startswith("json"): |
| text = text[4:].strip() |
| try: |
| return json.loads(text) |
| except Exception: |
| return None |
|
|
| def call_chat_json( |
| client: OpenAI, |
| model: str, |
| system: str, |
| user: str, |
| max_tokens: int, |
| temperature: float, |
| top_p: float = 0.9, |
| max_retries: int = 4, |
| ) -> Dict[str, Any]: |
| last_err = None |
| for attempt in range(1, max_retries + 1): |
| try: |
| resp = client.chat.completions.create( |
| model=model, |
| messages=[ |
| {"role": "system", "content": system}, |
| {"role": "user", "content": user}, |
| ], |
| temperature=temperature, |
| top_p=top_p, |
| max_tokens=max_tokens, |
| ) |
| text = resp.choices[0].message.content |
| obj = safe_json_parse(text) |
| if obj is None: |
| raise ValueError(f"JSON parse failed. Raw: {text[:200]}...") |
| return obj |
| except Exception as e: |
| last_err = str(e) |
| time.sleep(1.0 * attempt) |
| raise RuntimeError(f"call_chat_json failed after {max_retries} retries. Last error: {last_err}") |
|
|
| def normalize_text(s: str) -> str: |
| return " ".join((s or "").lower().split()) |
|
|
| |
| |
| |
| _BINARY_ANS = {"yes", "no", "true", "false"} |
| _AUX_START = { |
| "is", "are", "was", "were", |
| "do", "does", "did", |
| "can", "could", |
| "will", "would", "should", |
| "has", "have", "had", |
| "may", "might", "must", |
| } |
|
|
| def is_binary_qa(question: str, gt_answer: str) -> bool: |
| a = normalize_text(gt_answer) |
| if a in _BINARY_ANS: |
| return True |
| q = (question or "").strip().lower() |
| if q.endswith("?"): |
| first = q.split()[0] if q.split() else "" |
| if first in _AUX_START: |
| return True |
| return False |
|
|
| |
| |
| |
| SYSTEM_STATEMENT = """Convert a VQA (question + ground-truth answer) into ONE declarative sentence. |
| Return STRICT JSON: {"statement": "..."} only. |
| |
| Rules: |
| - Must be accurate given the answer. |
| - Natural, concise, factual. |
| - Preserve referenced entity labels if any (e.g., "chair A"). |
| - JSON only. |
| """ |
|
|
| USER_STATEMENT = """Question: {question} |
| Ground-truth answer: {answer} |
| |
| Return JSON only. |
| """ |
|
|
| def gen_statement(client: OpenAI, model: str, question: str, answer: str) -> str: |
| obj = call_chat_json( |
| client=client, |
| model=model, |
| system=SYSTEM_STATEMENT, |
| user=USER_STATEMENT.format(question=question, answer=answer), |
| max_tokens=128, |
| temperature=0.3, |
| ) |
| st = obj.get("statement", "") |
| if not isinstance(st, str) or not st.strip(): |
| raise ValueError(f"Bad statement: {obj}") |
| return st.strip() |
|
|
| |
| |
| |
| SYSTEM_WRONG_9 = """ |
| You generate distractors for a ground-truth statement. |
| |
| OUTPUT MUST BE STRICT JSON ONLY: |
| {{"distractors": ["<s1>", "<s2>", "<s3>", "<s4>", "<s5>", "<s6>", "<s7>", "<s8>", "<s9>"]}} |
| |
| Rules: |
| - Exactly 9 strings, all unique. |
| - Each is a single declarative sentence. |
| - Each must be WRONG given the ground-truth answer. |
| - Preserve the same main subject/entity (same instance label if present). |
| - The distractors should be diverse, non-ambiguous, realistic. The difference between ground-truth and distractors should be VERY clear. |
| - No markdown, no extra keys. |
| """ |
|
|
| USER_WRONG_9 = """INPUT |
| Question: {question} |
| Ground-truth answer: {answer} |
| Ground-truth statement: {statement} |
| |
| TASK |
| Generate 9 WRONG but plausible alternative declarative statements. |
| |
| REMINDER |
| Output JSON only with the exact schema: |
| {{"distractors": ["...","...","...","...","...","...","...","...","..."]}} |
| """ |
|
|
| SYSTEM_WRONG_1 = """ |
| You generate distractors for a ground-truth statement. |
| |
| OUTPUT MUST BE STRICT JSON ONLY: |
| {{"distractors": ["<s1>"]}} |
| |
| Rules: |
| - Exactly 1 string. |
| - Must be a single declarative sentence. |
| - Must be WRONG given the ground-truth answer. |
| - Preserve the same main subject/entity (same instance label if present). |
| - Make the wrong answer VERY clear (not ambiguous). |
| - No markdown, no extra keys. |
| """ |
|
|
| USER_WRONG_1 = """INPUT |
| Question: {question} |
| Ground-truth answer: {answer} |
| Ground-truth statement: {statement} |
| |
| TASK |
| Generate 1 WRONG but plausible alternative declarative statement. |
| |
| REMINDER |
| Output JSON only with the exact schema: |
| {{"distractors": ["..."]}} |
| """ |
|
|
| def gen_wrong_sentences( |
| client: OpenAI, |
| model: str, |
| question: str, |
| answer: str, |
| statement: str, |
| n: int = 9, |
| max_rounds: int = 5, |
| ) -> List[str]: |
| target = n |
| collected: List[str] = [] |
| seen = set([normalize_text(statement)]) |
|
|
| if n == 1: |
| system = SYSTEM_WRONG_1 |
| user_tmpl = USER_WRONG_1 |
| max_tokens = 256 |
| temperature = 0.7 |
| else: |
| system = SYSTEM_WRONG_9 |
| user_tmpl = USER_WRONG_9 |
| max_tokens = 512 |
| temperature = 0.85 |
|
|
| for _round in range(max_rounds): |
| obj = call_chat_json( |
| client=client, |
| model=model, |
| system=system, |
| user=user_tmpl.format(question=question, answer=answer, statement=statement), |
| max_tokens=max_tokens, |
| temperature=temperature, |
| ) |
|
|
| ds = obj.get("distractors", None) |
| if not isinstance(ds, list): |
| continue |
|
|
| for x in ds: |
| if not isinstance(x, str): |
| continue |
| x = x.strip() |
| if not x: |
| continue |
| nx = normalize_text(x) |
| if nx in seen: |
| continue |
| seen.add(nx) |
| collected.append(x) |
| if len(collected) >= target: |
| return collected[:target] |
|
|
| raise ValueError(f"Could not collect {target} unique distractors; got {len(collected)}") |
|
|
| |
| |
| |
| def build_mcq(correct: str, wrongs: List[str], seed: Optional[int] = None): |
| options = [correct] + list(wrongs) |
| if seed is not None: |
| rnd = random.Random(seed) |
| rnd.shuffle(options) |
| else: |
| random.shuffle(options) |
| label = options.index(correct) |
| return options, label |
|
|
| |
| |
| |
| _WORKER_CLIENT = None |
| _WORKER_MODEL = None |
|
|
| def _init_worker(base_url: str, model_name: str, timeout: int = 3600): |
| """Initializer runs once per process.""" |
| global _WORKER_CLIENT, _WORKER_MODEL |
| _WORKER_MODEL = model_name |
| _WORKER_CLIENT = OpenAI(api_key="EMPTY", base_url=base_url, timeout=timeout) |
|
|
| def _process_one(args: Tuple[str, int, str, str, str]) -> Dict[str, Any]: |
| """ |
| Args: |
| key, idx, scan_id, question, gt_answer |
| Returns: |
| {"ok": True, "out": {...}} or {"ok": False, "err": {...}} |
| """ |
| key, idx, scan_id, question, gt_answer = args |
|
|
| try: |
| correct_stmt = gen_statement(_WORKER_CLIENT, _WORKER_MODEL, question, gt_answer) |
|
|
| n_wrong = 1 if is_binary_qa(question, gt_answer) else 9 |
| wrongs = gen_wrong_sentences( |
| _WORKER_CLIENT, |
| _WORKER_MODEL, |
| question, |
| gt_answer, |
| correct_stmt, |
| n=n_wrong, |
| max_rounds=6 if n_wrong == 9 else 4, |
| ) |
|
|
| seed = abs(hash(key)) % (2**32) |
| options, label = build_mcq(correct_stmt, wrongs, seed=seed) |
|
|
| out = { |
| "key": key, |
| "scan_id": scan_id, |
| "question": question, |
| "gt_answer": gt_answer, |
| "correct": correct_stmt, |
| "options": options, |
| "label": label, |
| "is_binary": (n_wrong == 1), |
| } |
| return {"ok": True, "out": out} |
|
|
| except Exception as e: |
| return {"ok": False, "err": { |
| "key": key, |
| "scan_id": scan_id, |
| "question": question, |
| "gt_answer": gt_answer, |
| "error": str(e), |
| }} |
|
|
| |
| |
| |
| def main(): |
| base_url = "http://lrc-alpha-sg-gpu06:22001/v1" |
| model_name = "Qwen/Qwen3-VL-8B-Instruct" |
|
|
| in_path = "/home/m50048399/transfered/ye_project/PointMapVerse/existing_datasets/ScanNet/annotations/qa/ScanQA_v1.0_val.json" |
| out_jsonl = "/home/m50048399/transfered/ye_project/PointMapVerse/derived/scanqa_sentence_mcq_val.jsonl" |
| err_jsonl = out_jsonl + ".errors.jsonl" |
|
|
| data = load_json(in_path) |
| done = load_done_keys(out_jsonl) |
| print(f"Loaded {len(data)} items. Already done: {len(done)}") |
|
|
| |
| tasks = [] |
| for idx, item in enumerate(data): |
| scan_id = item.get("scene_id", "") |
| question = (item.get("question") or "").strip() |
| answers = item.get("answers") or [] |
| gt_answer = (answers[0] if answers else "").strip() |
|
|
| if not question or not gt_answer: |
| continue |
|
|
| key = f"{scan_id}::{idx}" |
| if key in done: |
| continue |
|
|
| tasks.append((key, idx, scan_id, question, gt_answer)) |
|
|
| print(f"To process: {len(tasks)}") |
|
|
| |
| ctx = mp.get_context("spawn") |
| with ctx.Pool( |
| processes=8, |
| initializer=_init_worker, |
| initargs=(base_url, model_name, 3600), |
| maxtasksperchild=50, |
| ) as pool: |
| processed = 0 |
| ok_cnt = 0 |
| err_cnt = 0 |
|
|
| for res in pool.imap_unordered(_process_one, tasks, chunksize=4): |
| processed += 1 |
| if res["ok"]: |
| append_jsonl(out_jsonl, res["out"]) |
| ok_cnt += 1 |
| else: |
| append_jsonl(err_jsonl, res["err"]) |
| err_cnt += 1 |
|
|
| if processed % 100 == 0: |
| print(f"Finished {processed}/{len(tasks)} | ok={ok_cnt} err={err_cnt}") |
|
|
| print("Done.") |
|
|
| if __name__ == "__main__": |
| main() |
|
|