| --- |
| dataset_info: |
| features: |
| - name: text |
| dtype: string |
| splits: |
| - name: train |
| num_bytes: 1784778472 |
| num_examples: 2005712 |
| download_size: 1106679567 |
| dataset_size: 1784778472 |
| tags: |
| - turkish |
| - pretraining |
| - masked-language-modeling |
| - diffusion |
| - wikipedia |
| - oscar |
| - news |
| configs: |
| - config_name: default |
| data_files: |
| - split: train |
| path: data/train-* |
| task_categories: |
| - text-generation |
| language: |
| - tr |
| --- |
| |
| # DiffutronLM-Pretraining-Corpus |
|
|
| **DiffutronLM-Pretraining-Corpus** is the comprehensive, filtered Turkish text dataset used during the Continual Pre-training (CPT) phase of the [Diffutron](https://huggingface.co/collections/diffutron/diffutronlm) language models. |
|
|
| The primary goal of this dataset was to align the cross-lingual representations of a multilingual base encoder (`jhu-clsp/mmBERT-base`) with the agglutinative complexity and morphological nuances of the Turkish language, without inducing catastrophic forgetting. |
|
|
| ## 📊 Dataset Composition |
|
|
| To ensure a balance between structured encyclopedic knowledge and natural, diverse web/news usage, the corpus is a composite of three primary open-source collections. It contains a total of **approximately 2 million sequences**. |
|
|
| * **Turkish Wikipedia (~406,000 sequences):** Sourced from the standard encyclopedic subset from the Wikimedia Foundation. It provides high-quality, factual, and structurally sound Turkish text. |
| * **Havadis & Temiz-OSCAR (~1,600,000 sequences):** * *Havadis:* A robust dataset of Turkish news articles providing formal and contemporary language usage. |
| * *Temiz-OSCAR:* A heavily filtered and cleaned version of the Common Crawl-based Turkish OSCAR corpus, representing diverse internet text. |
| * These two sources were merged, filtered, and uniformly sampled to extract 1.6 million high-quality sequences. |
|
|
| ## ⚙️ Preprocessing & Curation Strategy |
|
|
| The data was strictly curated to match the architectural constraints of the base Masked Diffusion Language Model (MDLM): |
|
|
| 1. **Length Filtering:** To ensure compatibility and training stability, a strict length constraint was applied across all data sources. Any sequences exceeding a **maximum token length of 512** were filtered out. |
| 2. **Tokenization Alignment:** The text was tokenized using the `jhu-clsp/mmBERT-base` tokenizer. This was a crucial step to maintain absolute alignment with the pre-trained embedding space of the frozen backbone. |
| 3. **Shuffling & Distribution:** The web and news subsets were thoroughly shuffled prior to sampling to ensure distributional uniformity during the training process. |
|
|
| ## 🚀 Intended Use |
|
|
| This corpus is optimized for: |
| * **Continual Pre-Training (CPT):** Adapting existing multilingual or general-purpose encoders to the Turkish language. |
| * **Masked Language Modeling (MLM):** Training models to predict masked or corrupted tokens (the foundational mechanism of discrete diffusion models). |
| * **Domain Adaptation:** Serving as a baseline corpus for general Turkish language modeling before task-specific instruction tuning. |
|
|
| ## ⚠️ Limitations |
|
|
| * **Length Constraint:** The dataset inherently lacks long-form document structures, as all sequences are hard-capped at 512 tokens. It is not suitable for training long-context models without additional data. |
| * **Tokenization:** While provided as text, researchers should be aware that the length filters were applied based on the specific subword tokenization of `mmBERT`. Re-tokenizing with a different tokenizer (like LLaMA's or a custom BPE) may yield different sequence lengths. |
|
|
| ## 📝 Citation |
|
|
| If you use this dataset in your research, please cite the Diffutron paper: |
|
|
| ```bibtex |
| @misc{diffutron2026, |
| title={Diffutron: A Masked Diffusion Language Model for Turkish Language}, |
| author={Şuayp Talha Kocabay and Talha Rüzgar Akkuş}, |
| year={2026}, |
| eprint={2603.20466}, |
| archivePrefix={arXiv}, |
| primaryClass={cs.CL}, |
| url={https://arxiv.org/abs/2603.20466}, |
| } |
| ``` |