Dataset Viewer
Auto-converted to Parquet Duplicate
The dataset viewer is not available for this split.
Parquet error: Scan size limit exceeded: attempted to read 24065557441 bytes, limit is 300000000 bytes Make sure that 1. the Parquet files contain a page index to enable random access without loading entire row groups2. otherwise use smaller row-group sizes when serializing the Parquet files
Error code:   TooBigContentError

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

RLVR-Env-Retrieval-Source-code-search-net-python

RLVR-ready retrieval environment derived from Nan-Do/code-search-net-python.

Author: Aman Priyanshu

What Is This

A 100k-row retrieval QA dataset where each row contains a question, ground-truth chunks, and pre-mined distractor chunks (random + semantically similar). Designed for training and evaluating retrieval agents in an RLVR (Reinforcement Learning with Verifiable Rewards) setup — the agent searches through distractors to find the correct chunk(s).

Domain: Python open-source functions from GitHub (CodeSearchNet)

Source

Derived from Nan-Do/code-search-net-python (455,243 unique functions). Original license: Apache 2.0 — retained here.

Schema

qa.parquet (100,000 rows)

Column Type Description
qa_id string Unique ID (search_py_0, search_py_1, ...)
question string The retrieval query
gt_chunks JSON string List of ground-truth chunk texts. 1 target code chunk per question (the function matching the summary)
random_chunks JSON string List of random distractor texts. ~500 random code chunks (>=20 chars, deduplicated against gt and similar)
similar_chunks JSON string List of hard-negative distractor texts. ~178 similar chunks via MiniLM cosine (<0.97) + char trigram edit-distance (<0.97 seq ratio), deduplicated

metadata.parquet (100,000 rows)

Column Type Description
qa_id string Matches qa.parquet
... ... chunk_idx, func_name, repo, char_count

chunks.parquet

455,243 code chunks with MiniLM embeddings. Kept for reference — not needed at inference time.

Deduplication

Within each row: gt > similar > random priority. No chunk text appears in more than one column per row. Similar chunks are internally deduplicated. Random chunks are filtered against both gt and similar.

How To Use

import json
import pyarrow.parquet as pq

t = pq.read_table("qa.parquet")
row = {col: t.column(col)[0].as_py() for col in t.column_names}
gt = json.loads(row["gt_chunks"])
distractors = json.loads(row["random_chunks"]) + json.loads(row["similar_chunks"])

License

Apache 2.0 (inherited from source dataset).

Downloads last month
1

Collection including AmanPriyanshu/RLVR-Env-Retrieval-Source-code-search-net-python