| | --- |
| | license: mit |
| | task_categories: |
| | - text-classification |
| | - question-answering |
| | - text-generation |
| | language: |
| | - en |
| | size_categories: |
| | - 10M<n<100M |
| | --- |
| | |
| | # 📚 TinyWay-Gutenberg-Clean (Compressed Shards) |
| |
|
| | A large-scale, high-quality English text dataset derived from Project Gutenberg. |
| | The corpus has been cleaned, normalized, deduplicated, segmented into fixed-length samples, and stored as compressed JSONL shards for efficient large-scale language model training. |
| |
|
| | This dataset is intended for pretraining and experimentation with small and medium language models such as **TinyWay**, tokenizer training, and large-scale NLP research. |
| |
|
| | --- |
| |
|
| | ## 📦 Dataset Overview |
| |
|
| | * **Name:** TinyWay-Gutenberg-Clean |
| | * **Current Release:** ~19 compressed shards (`.jsonl.gz`) |
| | * **Estimated Samples:** Tens of millions of text segments |
| | * **Language:** English |
| | * **Format:** Gzip-compressed JSON Lines (`.jsonl.gz`) |
| | * **Source:** Project Gutenberg (public domain books) |
| | * **License:** Public Domain |
| | * **Maintainer:** Shivam (NNEngine / ITM AIR Lab) |
| |
|
| | Each record contains a clean text segment between **30 and 60 words**. |
| |
|
| | Future releases will scale this dataset further (e.g., 100M+ samples). |
| |
|
| | --- |
| |
|
| | ## Data Format |
| |
|
| | Each line is a JSON object: |
| |
|
| | ```json |
| | { |
| | "id": "twg_000000012345", |
| | "text": "Cleaned natural English text segment between thirty and sixty words.", |
| | "word_count": 42, |
| | "source": "gutenberg" |
| | } |
| | ``` |
| |
|
| | ### Fields |
| |
|
| | | Field | Description | |
| | | ------------ | ------------------------------ | |
| | | `id` | Unique sample identifier | |
| | | `text` | Clean English text segment | |
| | | `word_count` | Number of words in the segment | |
| | | `source` | Data source identifier | |
| |
|
| | --- |
| |
|
| | ## Data Processing Pipeline |
| |
|
| | The dataset was generated using a fully streaming pipeline to ensure scalability and low memory usage. |
| |
|
| | ### Processing Steps |
| |
|
| | 1. **Streaming Input** |
| |
|
| | * Text streamed from a Project Gutenberg mirror on Hugging Face. |
| |
|
| | 2. **Text Cleaning** |
| |
|
| | * Removed Gutenberg headers and footers. |
| | * Removed chapter titles, page numbers, and boilerplate text. |
| | * Normalized whitespace and line breaks. |
| | * Removed non-ASCII and control characters. |
| | * Filtered malformed or extremely short segments. |
| |
|
| | 3. **Segmentation** |
| |
|
| | * Text segmented into chunks of **30–60 words**. |
| |
|
| | 4. **Validation** |
| |
|
| | * Enforced word count limits. |
| | * Filtered invalid or noisy segments. |
| |
|
| | 5. **Deduplication** |
| |
|
| | * Exact hash-based deduplication applied during generation. |
| |
|
| | 6. **Compression & Sharding** |
| |
|
| | * Data stored as `.jsonl.gz` shards for efficient disk usage and streaming. |
| |
|
| | --- |
| |
|
| | ## How to Load the Dataset |
| |
|
| | ### Using Hugging Face Datasets (Streaming) |
| |
|
| | ```python |
| | from datasets import load_dataset |
| | |
| | dataset = load_dataset( |
| | "NNEngine/TinyWay-Gutenberg-Clean", |
| | split="train", |
| | streaming=True |
| | ) |
| | |
| | for i, sample in enumerate(dataset): |
| | print(sample) |
| | if i == 3: |
| | break |
| | ``` |
| |
|
| | --- |
| |
|
| | ### Reading a Shard Manually |
| |
|
| | ```python |
| | import gzip |
| | import json |
| | |
| | with gzip.open("train-00000.jsonl.gz", "rt", encoding="utf-8") as f: |
| | for _ in range(3): |
| | print(json.loads(next(f))) |
| | ``` |
| |
|
| | --- |
| |
|
| | ## Dataset Characteristics (Approximate) |
| |
|
| | * **Average words per sample:** ~45 |
| | * **Style:** Literary and narrative English |
| | * **Domain:** Fiction, non-fiction, historical texts |
| | * **Vocabulary:** Large natural English vocabulary |
| | * **Compression:** ~60–70% size reduction vs raw JSONL |
| |
|
| | Exact statistics may vary per shard and will be expanded in future releases. |
| |
|
| | --- |
| |
|
| | ## Limitations |
| |
|
| | * Primarily literary and historical language. |
| | * No conversational chat data. |
| | * No code or structured technical documentation. |
| | * Some archaic vocabulary and sentence structures may appear. |
| | * Deduplication is hash-based (near-duplicates may remain). |
| |
|
| | For conversational or web-style language modeling, this dataset should be mixed with complementary corpora. |
| |
|
| | --- |
| |
|
| | ## License |
| |
|
| | All source texts originate from Project Gutenberg and are in the **public domain**. |
| | This processed dataset is released for unrestricted research and commercial use. |
| |
|
| | --- |
| |
|
| | ## Versioning & Roadmap |
| |
|
| | Planned future updates: |
| |
|
| | - Larger releases (target: 100M+ samples) |
| | - Improved deduplication (near-duplicate filtering) |
| | - Dataset statistics and analytics |
| | - Additional language normalization |
| |
|
| | Each major release will be versioned clearly. |
| |
|
| | --- |
| |
|
| | ## Citation |
| |
|
| | If you use this dataset in research or publications, please cite: |
| |
|
| | ``` |
| | TinyWay-Gutenberg-Clean |
| | Shivam (NNEngine), 2026 |
| | ``` |