building a cognitive core 1: synthetic pretraining for very small reasoning models.



SUMMARY

A cognitive core is a very small reasoning model with high context engineering, very high reasoning skills, and little static knowledge (see e.g. here). I want to build this.

To this goal, this project asks a first practical question: can synthetic pretraining data make very small language models better at reasoning?

We pretrain dense 0.8B Qwen-style models on a math corpus, then compare the original data against three synthetic augmentations. The answer is yes: synthetic pretraining improves few-shot reasoning on GSM8K and MATH500, strengthens in-context learning, and reaches the original model's final performance with substantially fewer training tokens.

1 - Why synthetic pretraining?

Synthetic data is becoming a central part of modern language-model pretraining. Beyond simply adding more text, it gives us control over the training signal: we can rewrite examples to expose hidden reasoning steps, add missing context, and make dependencies easier for an autoregressive learner to recover.

The hypothesis here is simple. If math-heavy data is rewritten into more explicit, reasoning-friendly trajectories, then each token should carry more useful supervision. This should matter especially for very small models, where every token and every parameter has to work harder.

We test three data transformations:
1. TPT, which adds explicit thought-process structure.
2. First principles, which turns source documents into learning-note style explanations.
3. Rephrasing, which rewrites documents to maximize learnability for a small student model.

All synthetic data is generated with a same-scale 0.8B Qwen model in non-thinking mode. This is deliberately constrained: the goal is not to distill from a huge teacher, but to test whether synthetic formatting and reasoning structure can improve pretraining even when the generator is small.

2 - Main result: better few-shot reasoning

We evaluate the pretrained models on GSM8K and MATH500 using chain-of-thought prompting across different numbers of demonstrations. The key metric is the gain over zero-shot accuracy: $$ \text{gain}(k) = \text{acc}(k) - \text{acc}(0). $$ This isolates how much the model benefits from in-context demonstrations, rather than only measuring its baseline zero-shot ability.
Few-shot gain over zero-shot accuracy for original and synthetic pretrained models on GSM8K and MATH500.
Figure 1. Synthetic pretraining improves few-shot gains on GSM8K and MATH500. The gap grows with more demonstrations, suggesting that synthetic data strengthens in-context learning rather than only improving zero-shot performance.
Across both benchmarks, the synthetic variants improve over the original corpus. The gains are especially clear from 0-shot to 2-shot and remain stable as more examples are added.

3 - Robustness to prompt examples

Few-shot evaluation can be fragile: results depend on which examples are shown, in what order, and how closely the prompt format matches the model's training distribution. To check that the effect is not just a lucky set of demonstrations, we repeat the evaluation with random chain-of-thought examples sampled per question.
Few-shot gain with randomly sampled chain-of-thought demonstrations across random seeds.
Figure 2. The gains remain under randomly sampled demonstrations. Error bars show variation across seeds, and synthetic pretraining continues to outperform the original dataset.
The randomized setting keeps the same overall conclusion and makes differences between synthetic prompts easier to see. This supports the view that the models are genuinely learning better reasoning behavior, not just benefiting from a fixed demonstration template.

4 - Is it reasoning, or just formatting?

Math benchmarks can reward formatting as well as reasoning. For example, a model may know the answer but fail because it does not place it in the expected format. To separate these effects, we compare strict parsing with more flexible answer extraction.
Breakdown of cases where the original model fails under strict and flexible parsing.
Figure 3. Most improvements are not just formatting fixes. The synthetic models often solve examples that the original model fails even under flexible parsing.
The dominant pattern is that the original model fails even when answers are parsed flexibly. That means the synthetic models are usually gaining real problem-solving ability, not merely learning to wrap answers in the correct syntax.
Supplementary breakdown of questions solved by only the original model, only the synthetic model, or both.
Supplementary Figure 1. A complementary view of strict and flexible correctness, separating examples solved only by the original model, only by the synthetic model, or by both.


5 - Token efficiency during training

Another useful question is whether the benefit appears only at the end of training, or whether synthetic data changes the training trajectory itself. The checkpoint results suggest that the effect emerges early.
Strict accuracy as a function of training tokens seen for original and synthetic models.
Figure 4. Synthetic models reach the original model's final performance with fewer tokens. The effect is strongest for the rephrasing and TPT prompts, which show more stable gains across training.
In practice, synthetic runs match the original model's final performance using roughly 3-6x fewer training tokens on GSM8K and about 2.5x fewer tokens on MATH500. This is the main practical reason the result is interesting: synthetic data improves the utility of each pretraining token.

6 - What changes in the data?

The three prompts do not produce the same kind of augmentation. TPT and first-principles generations are much longer, while rephrasing stays closer to the original row length. Despite these differences, all three variants improve downstream reasoning.
Distribution of row lengths for the original and synthetic datasets.
Figure 5. Synthetic prompts change the length distribution in different ways. Rephrasing is comparatively compact, while TPT and first-principles augmentations are longer and more explanatory.
This suggests that there is more than one useful mechanism. Longer synthetic examples may add explicit reasoning traces, while compact rephrasing may make the same content easier to learn. The common ingredient is not simply more tokens, but a better training signal.

7 - Takeaways

The main result is that synthetic pretraining can improve reasoning in very small models, even when the generator is itself very small. This challenges two common assumptions: that synthetic data only helps when produced by much larger teachers, and that it mainly helps low-quality corpora. Here, the source corpus is already heavily curated, yet synthetic augmentation still improves reasoning.

I think this is a useful first step toward building cognitive-core style models: small systems with less static knowledge, but stronger reasoning over the context they are given. The open question is how to separate gains from synthetic data generation, data reformatting, and implicit distillation. More details, methodology, and citation information are available in the full Tufa Labs post.


refs:

[1] Synthetic pretraining for very small reasoning models. Saponati, M. Tufa Labs Research, Apr 2026.