The quest for optimal neural network architectures has been hampered by the immense computational burden of traditional Neural Architecture Search (NAS). Existing methods, which rely on Large Language Models (LLMs) to synthesize complete model implementations from scratch, are prohibitively expensive and generate verbose code. This paper introduces a paradigm shift with Delta-Code Generation, a novel pipeline that leverages fine-tuned LLMs to produce compact unified diffs (deltas) that refine existing baseline architectures.
Efficiency Through Incremental Refinement
The core innovation lies in shifting from full model synthesis to incremental modification. By fine-tuning LLMs, specifically three 7B-class models (DeepSeek-Coder-7B, Qwen2.5-Coder-7B, and Mistral-7B), on curated architectures from the LEMUR dataset and employing MinHash-Jaccard novelty filtering, the approach generates concise deltas. This method drastically reduces output lengths, achieving a 75-85% reduction compared to full generation (30-50 lines versus 200+ lines). This token-efficient strategy makes LLM-driven NAS far more accessible.