Skip to content
#

llm-training-data

Here are 8 public repositories matching this topic...

Language: All
Filter by language
Cre4T3Tiv3

Stratified LLM Subsets delivers diverse training data at 100K-1M scales across pre-training (FineWeb-Edu, Proof-Pile-2), instruction-following (Tulu-3, Orca AgentInstruct), and reasoning distillation (Llama-Nemotron). Embedding-based k-means clustering ensures maximum diversity across 5 high-quality open datasets.

  • Updated Oct 4, 2025
  • HTML

Improve this page

Add a description, image, and links to the llm-training-data topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the llm-training-data topic, visit your repo's landing page and select "manage topics."

Learn more