| | """Module for testing streaming dataset sequence packing""" |
| | import pytest |
| | from datasets import concatenate_datasets, load_dataset |
| | from torch.utils.data import DataLoader, RandomSampler |
| | from transformers import AutoTokenizer |
| |
|
| | from axolotl.datasets import TokenizedPromptDataset |
| | from axolotl.prompt_strategies.completion import load |
| | from axolotl.utils.collators import V2BatchSamplerDataCollatorForSeq2Seq |
| | from axolotl.utils.dict import DictDefault |
| | from axolotl.utils.samplers import MultipackBatchSampler, get_dataset_lengths |
| |
|
| |
|
| | @pytest.fixture(name="tokenizer") |
| | def fixture_tokenizer(): |
| | tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b") |
| | tokenizer.pad_token = "</s>" |
| | return tokenizer |
| |
|
| |
|
| | @pytest.fixture(name="max_seq_length") |
| | def fixture_max_seq_length(): |
| | return 4096 |
| |
|
| |
|
| | class TestBatchedSamplerPacking: |
| | """ |
| | Test class for packing streaming dataset sequences |
| | """ |
| |
|
| | @pytest.mark.parametrize( |
| | "batch_size, num_workers", |
| | [ |
| | (1, 0), |
| | (2, 0), |
| | (1, 2), |
| | (2, 2), |
| | ], |
| | ) |
| | def test_packing(self, batch_size, num_workers, tokenizer, max_seq_length): |
| | import axolotl.monkeypatch.data.batch_dataset_fetcher |
| |
|
| | dataset = load_dataset( |
| | "Trelis/tiny-shakespeare", |
| | split="train", |
| | ) |
| |
|
| | cfg = DictDefault( |
| | { |
| | "train_on_inputs": True, |
| | "sequence_len": max_seq_length, |
| | } |
| | ) |
| | ds_cfg = DictDefault( |
| | { |
| | "field": "Text", |
| | } |
| | ) |
| | completion_strategy = load(tokenizer, cfg, ds_cfg) |
| | dataset_wrapper = TokenizedPromptDataset( |
| | completion_strategy, |
| | dataset, |
| | ) |
| | train_dataset = concatenate_datasets([dataset_wrapper]) |
| | batch_sampler = MultipackBatchSampler( |
| | sampler=RandomSampler(train_dataset), |
| | batch_size=batch_size, |
| | drop_last=True, |
| | batch_max_len=max_seq_length, |
| | lengths=get_dataset_lengths(train_dataset), |
| | ) |
| |
|
| | loader = DataLoader( |
| | train_dataset, |
| | batch_sampler=batch_sampler, |
| | collate_fn=V2BatchSamplerDataCollatorForSeq2Seq( |
| | tokenizer=tokenizer, |
| | padding=True, |
| | pad_to_multiple_of=max_seq_length, |
| | return_tensors="pt", |
| | ), |
| | num_workers=num_workers, |
| | ) |
| | inputs = next(iter(loader)) |
| |
|
| | assert inputs["input_ids"].shape == (batch_size, max_seq_length) |
| | assert inputs["labels"].shape == (batch_size, max_seq_length) |
| | assert inputs["attention_mask"].shape == (batch_size, max_seq_length) |
| |
|
| | assert inputs["input_ids"].tolist()[0][0] == 2 |
| | assert inputs["labels"].tolist()[0][0] == -100 |
| | assert inputs["attention_mask"].tolist()[0][0] == 0 |
| | assert inputs["attention_mask"].tolist()[0][-1] > 1 |
| |
|
| | if batch_size >= 2: |
| | assert inputs["input_ids"].tolist()[1][0] == 2 |
| | assert inputs["labels"].tolist()[1][0] == -100 |
| | assert inputs["attention_mask"].tolist()[1][0] == 0 |
| | assert inputs["attention_mask"].tolist()[1][-1] > 1 |
| |
|