LLMs · arXiv · Score 9.8
Embarrassingly Simple Self-Distillation Improves Code Generation
Ruixiang Zhang, Yifan Song, Apple
Origin & Signal · 7.4/10
Apple-authored arXiv preprint with linked project code.
Credible industrial authorship and clear practical relevance, but still a fresh preprint.
A simple post-training method that lifts coding performance is exactly the kind of high-leverage paper people can try quickly in real agentic systems.
A minimal post-training recipe improves pass@1 on code-generation benchmarks without requiring a verifier, reward model, or reinforcement learning.
Cyber · arXiv · Score 9.1
AudioHijack: Context-Agnostic Auditory Prompt Injection Against Large Audio-Language Models
Meng Chen, ZJU MUSLAB
Origin & Signal · 6.8/10
Academic preprint with reproducibility-focused artifact release.
Strong security relevance and artifact support, though venue outcome is not yet known.
Multimodal prompt injection research is becoming directly relevant to real voice and agent interfaces, so strong evaluations here matter a lot.
We study prompt injection in multimodal audio-language systems and evaluate how adversarial auditory prompts transfer across model families and target behaviors.
Robotics · OpenAlex · Score 8.7
Embodied Planning with Vision-Language-Action Models in Cluttered Homes
Jordan Lee, Mina Park, Robotics Lab
Origin & Signal · 6.5/10
University robotics lab preprint with sim-to-real experiments.
Solid embodied-AI fit and real-world evaluation, but limited external prestige signal.
It is a good example of robotics work that earns a spot in the digest because it connects foundation-model planning to real-world control.
A vision-language-action system improves manipulation planning in cluttered home environments through grounded subgoal generation and sim-to-real transfer.