Research Statement

Agentic Small Language Model

Empowering small language models through distillation, tool integration, and context optimization.

sLM
+
Tools
Agent
Minki Kang
Minki Kang Ph.D. Student, KAIST AI
1

Small models are the deployment path for everyday agents.

My industry experience repeatedly pointed to the same constraint: useful agents must often run close to users and products.

Cost Always-on interaction

Agent workloads scale with tasks, users, and repeated tool calls.

Privacy Local personal context

Private memories and user data should not always leave controlled environments.

Edge Low-latency products

Games, medical AI, and copilots need responsive agents under real deployment limits.

2

Why are small models weak?

Small models are bottlenecked by parametric memory.

They are not simply unable to reason. Many reasoning tasks secretly require closed-book memory: facts, calculations, intermediate checks, and verification signals.

Knowledge-intensive reasoning Facts must be recalled before reasoning can help.

KARD shows that external knowledge reduces what a small model must memorize internally.

Self-verification Checking answers can require exact computation and fact checking.

T1 shows that tools turn memorization-heavy verification into tool-use and interpretation.

The goal is to convert memory-heavy problems into agentic problems.

3

Small models need tools more, but use tools worse.

Tool need sLM > LLM

Small models need external memory, computation, and verification to overcome limited capacity.

Agentic capability sLM < LLM

Without training, they are worse at retrieving, computing, checking, and recovering.

My research trains small models to use agentic scaffolding reliably enough that external memory and computation compensate for limited parametric capacity.

01

Retrieve missing knowledge

02

Compute and verify exact steps

03

Distill reliable agentic behavior

4

Research Program

Building blocks for agentic small language models.

External knowledge

KARD: Knowledge-Augmented Reasoning

Retrieve knowledge for rationale generation.

NeurIPS 2023
Tool verification

T1: Tool-integrated Self-verification

Use tools for calculation and fact-checking.

ICLR 2026
Agent distillation

Distilling LLM Agent into Small Models

Distill retrieval, code, observation, and recovery traces.

NeurIPS 2025 Spotlight
Long-horizon context

ACON: Optimizing Context Compression

Compress context for long-horizon productive work.

ICML 2026
5

Future Vision

Agentic scaffolding can turn small models into self-improving agents.

01 Master agentic scaffolding

Use tools, memory, verification, and planning reliably.

02 Do model-training research

Design data, objectives, evaluations, and training recipes.

03 Improve itself

Turn research feedback into better agentic behavior.

Takeaway: once an SLM masters agentic scaffolding, it can conduct research on model training and use that feedback to improve its own capabilities.

6
1 / 6