--- dataset_info: features: - name: text dtype: string - name: pos dtype: float64 splits: - name: train num_bytes: 5335090828.0 num_examples: 1002630 download_size: 3227201658 dataset_size: 5335090828.0 configs: - config_name: default data_files: - split: train path: data/train-* --- 1,378,234,368 tokens (using the Llama tokenizer, ~1.18b gpt4 tokens) from a deduped pile raw shard, filter len<896, ask-llm ([“How to Train Data-Efficient LLMs”](https://arxiv.org/abs/2402.09668)) w/ [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2), keep top 1/4 ``` { "text": "Once upon a time...", "pos": -5.654354325 } ```