[D] Released a 100k-sample dataset on Hugging Face

Posted by AdhesivenessSea9511@reddit | LocalLLaMA | View on Reddit | 6 comments

We’ve released a 100,000-sample Chain-of-Thought (CoT) dataset for fine-tuning local reasoning models.

Each sample includes explicit intermediate reasoning traces, rather than answer-only supervision. The goal is to improve reasoning consistency during supervised fine-tuning, especially for smaller local models.

We’re sharing it here to gather feedback from people working on local LLM fine-tuning and reasoning distillation.

I’d especially love feedback on:

- CoT length

- consistency of reasoning style

- whether full reasoning traces help or hurt smaller local models

Hugging Face:

https://huggingface.co/datasets/Kamisori-daijin/email-datasets-v2-100k