Self-Foveate: Enhancing Diversity and Difficulty of Synthesized Instructions from Unsupervised Text via Multi-Level Foveation

1Harbin Institute of Technology
#Corresponding Author
TL;DR

An LLM-driven Multi-Level Foveation method
for synthesizing high-quality instruction data from unsupervised text
with enhanced diversity and difficulty.

Abstract

Synthesizing high-quality instruction data from unsupervised text is a promising paradigm for training large language models (LLMs), yet automated methods for this task still exhibit significant limitations in the diversity and difficulty of synthesized instructions. To address these challenges, we propose Self-Foveate, an LLM-driven method for instruction synthesis. Inspired by hierarchical human visual perception, Self-Foveate introduces a "Micro-Scatter-Macro" multi-level foveation methodology that guides the extraction of textual information at three complementary granularities, from fine-grained details through cross-region connections to holistic patterns, thereby enhancing both the diversity and difficulty of synthesized instructions. Furthermore, a re-synthesis module is incorporated to improve the fidelity of instructions to source text and their overall quality. Comprehensive experiments across multiple unsupervised corpora and diverse model architectures demonstrate that Self-Foveate consistently outperforms existing methods. We publicly release our code at https://github.com/Mubuky/Self-Foveate.

Multi-Level Foveation Methodology

Inspired by hierarchical human visual perception, Self-Foveate introduces a "Micro-Scatter-Macro" methodology that extracts textual information at three complementary granularities:

  • Micro Level (Word): Fine-grained entity/attribute extraction focusing on individual words for detailed features.
  • Scatter Level (Multi-keyword): Cross-entity relationship grouping that combines 1-3 keywords into diverse feature groups.
  • Macro Level (Sentence): Rhetorical/figurative device extraction capturing complete sentences as contextual features.

Framework Overview

Compared to baseline methods like Self-QA that employ single-step generation producing simple and monotonous instruction candidates, Self-Foveate leverages multi-level foveation to:

  • Extract diverse details: Multi-level foveation enables the LLM to extract details (highlighted in distinct colors) of the text.
  • Synthesize with diversity: Different synthesis paradigms generate instructions with enhanced diversity and difficulty.
  • Ensure quality through re-synthesis: A re-synthesis module improves the fidelity of instructions to source text.

Experimental Results

Comprehensive experiments demonstrate that Self-Foveate consistently outperforms existing methods across multiple unsupervised corpora and diverse model architectures.

Accuracy trend across different settings

Recall trend across different settings

Downstream Task Performance

Recall (Rec.) and LLM Accuracy (Acc.) on downstream tasks: Self-Foveate vs. baselines.

Settings GPT-4o mini DeepSeek-V3
SQuAD HotpotQA FilmWiki SQuAD HotpotQA FilmWiki
Rec.Acc. Rec.Acc. Rec.Acc. Rec.Acc. Rec.Acc. Rec.Acc.
Llama-3.1-8B
None* 0.3090.202 0.2440.160 0.2120.082 0.3090.202 0.2440.160 0.2120.082
Self-QA 0.3670.384 0.3720.358 0.3280.201 0.3890.412 0.3990.378 0.3700.239
Wiki2023 0.3270.361 0.3380.322 0.3330.235 0.3420.370 0.3400.328 0.3490.244
Bonito* 0.3860.405 0.3600.372 0.2190.153 0.3860.405 0.3600.372 0.2190.153
Self-Foveate 0.4840.490 0.5070.486 0.5120.367 0.4810.491 0.5250.501 0.5480.397
Qwen2.5-7B
None* 0.2510.300 0.2660.234 0.1390.032 0.2510.300 0.2660.234 0.1390.032
Self-QA 0.2490.232 0.2760.246 0.2060.082 0.1190.125 0.1020.106 0.1110.056
Wiki2023 0.2150.221 0.1350.112 0.1920.093 0.1700.083 0.1970.203 0.2020.136
Bonito* 0.1430.109 0.2120.199 0.1680.098 0.1430.109 0.2120.199 0.1680.098
Self-Foveate 0.4080.414 0.3720.329 0.2830.140 0.3880.389 0.3420.331 0.2610.140
Gemma-2-9B
None* 0.2240.121 0.1750.078 0.2110.099 0.2240.121 0.1750.078 0.2210.099
Self-QA 0.3830.409 0.4080.389 0.4290.315 0.4020.435 0.4240.408 0.5090.386
Wiki2023 0.3360.378 0.3610.352 0.4780.384 0.3640.399 0.3730.365 0.4940.401
Bonito* 0.4110.457 0.3660.373 0.2550.196 0.4110.457 0.3660.373 0.2550.196
Self-Foveate 0.5070.525 0.5370.520 0.6720.528 0.4990.514 0.5520.525 0.6970.581

* Indicates that the base model was not fine-tuned using instructions synthesized by GPT-4o mini or DeepSeek-V3.

Citation

@inproceedings{li2025self,
  title={Self-Foveate: Enhancing Diversity and Difficulty of Synthesized Instructions from Unsupervised Text via Multi-Level Foveation},
  author={Li, Mingzhe and Lu, Xin and Zhao, Yanyan},
  booktitle={Findings of the Association for Computational Linguistics: ACL 2025},
  pages={7274--7289},
  year={2025}
}