Needle in the haystack LLMs for vulnerability research#1978
Open
carlospolop wants to merge 1 commit intomasterfrom
Open
Needle in the haystack LLMs for vulnerability research#1978carlospolop wants to merge 1 commit intomasterfrom
carlospolop wants to merge 1 commit intomasterfrom
Conversation
Collaborator
Author
🔗 Additional ContextOriginal Blog Post: https://devansh.bearblog.dev/needle-in-the-haystack/ Content Categories: Based on the analysis, this content was categorized under "AI Security (LLM-assisted vulnerability discovery methodology) + Pentesting Web -> JWT Vulnerabilities (JWT/JWKS alg confusion) + Pentesting Web -> Authorization/IDOR (authz tier confusion pattern) + Generic Methodologies & Resources -> Threat Modeling (invariant-driven audits)". Repository Maintenance:
Review Notes:
Bot Version: HackTricks News Bot v1.0 |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
🤖 Automated Content Update
This PR was automatically generated by the HackTricks News Bot based on a technical blog post.
📝 Source Information
🎯 Content Summary
Needle in the haystack: LLMs for vulnerability research (technical summary)
The post explains a pragmatic methodology for using LLM agent CLIs/TUIs (explicitly mentioning OpenAI Codex) to discover real vulnerabilities in large codebases without manual source review. The core claim is that vulnerability discovery is a “needle-in-the-haystack” problem where the bug is often a single invariant violation hidden among large amounts of valid code, and that over-scaffolding<...
🔧 Technical Details
Threat-model-driven LLM auditing (minimal scaffolding): Build a compact, editable threat model by feeding the LLM previously disclosed CVEs/advisories for the project, then use it to enumerate entry points (HTTP/RPC/CLI/jobs), trust boundaries (browser↔server, service↔service, plugin↔host, sandbox↔privileged), and high-risk sinks (deserialization, templating, native bindings, authz checks, untrusted parsing). Derive invariants (“only admins can call X”, “JWT issuer must be Y”) and search for invariant violations and fix-bypasses in small, focused contexts to avoid long-context degradation (“context rot” / lost-in-the-middle effects). Verify suspected issues with concrete verifiers (tests, harnesses, sanitizer builds, fuzzers, static/grep invariant checks) rather than trusting model assertions.
Authorization tier confusion via incomplete privilege checks: When applications support multiple credential tiers (e.g., master vs read-only master), audit handlers that ga...
🤖 Agent Actions
Summary of changes:
Updated file:
src/AI/AI-Assisted-Fuzzing-and-Vulnerability-Discovery.mdNotes:
src/pentesting-web/hacking-jwt-json-web-tokens.mdandsrc/pentesting-web/idor.md. The blog’s novel contribution is the LLM-audit methodology and invariant-driven workflow, which fits best in the AI-assisted vuln discovery page.If you want me to also thread a short cross-link from
src/generic-methodologies-and-resources/threat-modeling.mdto the new LLM audit workflow, say the word.This PR was automatically created by the HackTricks Feed Bot. Please review the changes carefully before merging.