Large Language Models (LLMs) are revolutionizing f ields previously dominated by human effort. This work presents the first systematic investigation of how LLMs can team with analysts during software reverse engineering (SRE). To accomplish this, we first document the state of LLMs in SRE with an online survey of 153 practitioners, and then we design a finegrained human study on two Capture-The-Flag-style binaries representative of real-world software. In our human study, we instrumented the SRE workflow of 48 participants (split between 24 novices and 24 experts), observing over 109 hours of SRE. Through 18 findings, we found various benefits and harms of LLMs in SRE. Remarkably, we found that LLM assistance narrows the expertise gap: novices’ comprehension rate rises by approximately 98%, matching that of experts, whereas experts gain little; however, they also had harmful hallucinations, unhelpful suggestions, and ineffective results. Known-algorithm functions are triaged up to 2.4× faster, and artifact recovery (symbols, comments, types) increases by at least 66%. Overall, our findings identify powerful synergies of humans and LLMs in SRE, but also emphasize the significant shortcomings of LLMs in their current integration.
Decompiling the synergy: An empirical study of human–LLM teaming in software reverse engineering
NDSS 2026, 33rd Network and Distributed System Security Symposium, 23-27 February 2026, San Diego, USA
Type:
Conférence
City:
San Diego
Date:
2026-02-23
Department:
Sécurité numérique
Eurecom Ref:
8548
Copyright:
Copyright Usenix. Personal use of this material is permitted. The definitive version of this paper was published in NDSS 2026, 33rd Network and Distributed System Security Symposium, 23-27 February 2026, San Diego, USA and is available at :
See also:
PERMALINK : https://www.eurecom.fr/publication/8548