Kernel code removals driven by LLM-created security reports
TL;DR Highlight
Linux kernel maintainers are removing legacy drivers—ISA, PCMCIA, AX.25, ATM, and ISDN—after AI-generated security bug reports overwhelmed them, demonstrating a drastic response to unmanageable code.
Who Should Read
Linux kernel contributors, open-source project maintainers, or developers utilizing or considering LLM-based automation tools for vulnerability detection.
Core Mechanics
- Linux kernel maintainers proposed patches to remove ISA/PCMCIA Ethernet drivers, parts of the PCI driver, the AX.25 and amateur radio subsystem, ATM protocols and drivers, and the ISDN subsystem.
- The removal reason isn't technical flaws, but a surge in security bug reports automatically generated by LLMs. Maintainer comments explicitly state the need to remove code to protect mental health due to the inability to process AI-generated reports.
- AX.25 (amateur radio packet communication protocol) and related HAM radio drivers already received many bug reports from syzbot (kernel automated fuzzing tool), and the influx of AI reports finalized the removal decision.
- Most of the removed code are drivers or protocols for legacy hardware primarily used before the 2010s. ATM has been replaced by MPLS/MetroE, ISDN is virtually obsolete, and laptops with PCMCIA slots haven’t been produced since 2008.
- These codes were in a ‘non-maintained state’ but were included in a large project (Linux kernel), giving the illusion of maintenance. Had they been independent projects, their inactive status would have been apparent years ago.
- The validity of bug reports generated by LLMs is debated. Some linked emails highlight the issue of ‘junk reports’ from AI increasing review burden without adding value.
- The removed code can potentially be continued as out-of-tree kernel modules or userspace implementations. The HAM radio community has already begun discussing a userspace protocol implementation written in a more modern language.
Evidence
- "HAM radio community users expressed regret over the AX.25 removal, and a thread explaining the decision’s background appeared on the linux-hams mailing list, alongside optimistic views about a modern userspace protocol becoming the new standard."
How to Apply
- When integrating LLM-based security scanners or automated bug reporting tools into open-source projects, a report quality filtering layer is essential. Failing to validate AI-generated issues can overwhelm maintainers and lead to code removal, as seen in this case.
- If your project includes legacy drivers or modules that are effectively unmaintained, assess their maintenance status before deploying LLM-based vulnerability scanners. AI tends to report more pattern-based vulnerabilities in older code, potentially flooding the issue tracker with noise.
- If you operate industrial environments requiring legacy hardware support in kernels or system software, verify whether your drivers are on the removal list (ISA, PCMCIA, ATM, AX.25, ISDN) and prepare for out-of-tree module maintenance or userspace alternatives.
Terminology
Related Papers
Can LLMs model real-world systems in TLA+?
LLM이 TLA+ 명세를 작성할 때 문법은 잘 통과하지만 실제 시스템과의 동작 일치도(conformance)는 46% 수준에 그친다는 걸 체계적으로 검증한 벤치마크 연구로, AI 기반 형식 검증의 현실적 한계를 보여준다.
Natural Language Autoencoders: Turning Claude's Thoughts into Text
Anthropic이 LLM 내부의 숫자 벡터(활성화값)를 직접 읽을 수 있는 자연어로 변환하는 NLA 기법을 공개했다. AI가 실제로 무슨 생각을 하는지 해석하는 interpretability 연구의 새로운 진전이다.
ProgramBench: Can language models rebuild programs from scratch?
LLM이 FFmpeg, SQLite, PHP 인터프리터 같은 실제 소프트웨어를 문서만 보고 처음부터 재구현할 수 있는지 측정하는 새 벤치마크로, 최고 모델도 전체 태스크의 3%만 95% 이상 통과하는 수준에 그쳤다.
MOSAIC-Bench: Measuring Compositional Vulnerability Induction in Coding Agents
티켓 3장으로 쪼개면 Claude/GPT도 보안 취약점 코드를 53~86% 확률로 그냥 짜준다.
Refusal in Language Models Is Mediated by a Single Direction
Open-source chat models encode safety as a single vector direction, and removing it disables safety fine-tuning.
Show HN: A new benchmark for testing LLMs for deterministic outputs
Structured Output Benchmark assesses LLM JSON handling across seven metrics, revealing performance beyond schema compliance.