Skip to content

Conversation

cristobalvch
Copy link
Contributor

Benchmark designed to evaluate a fully automated integration of LLMs (Large Language Models) with no HITL (Human-in-the-Loop) into web application attack scenarios using CAI (Cybersecurity AI). Its goal is to test various prompting strategies and different LLMs to assess their effectiveness in identifying vulnerabilities within web applications.

## Project Folder Structure
In this section, the main folder structure is described.
```plaintext
llm-cai-project/ # Root directory of the project
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@cristobalvch i think this route is wrong


**Fully Automated (No HITL):**
The pipeline is designed to be **fully automated, with no Human-in-the-Loop (HITL)**. When the agent attempts to solve the challenge labs, **no human interaction with the model is required**; all decisions, iterations, and actions are executed autonomously according to the experiment’s configuration and the prompt templates.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@cristobalvch could you please add a "results" section in here wherein you include images and summarize results obtained, that'd be very interesting.

@vmayoral vmayoral self-assigned this Aug 22, 2025
@Mery-Sanz
Copy link
Collaborator

I really liked the implementation! I tested it and it works correctly.
The only thing I’d need is for you to add a /logs folder with a .gitkeep, otherwise this error is raised when running:

FileNotFoundError: [Errno 2] No such file or directory: 'logs'

Also, if you could add a table with all the PortSwigger challenges, that would be awesome 🙌
It should be trivial to generate from benchmarks/prompt-bench/utils/portswigger_labs.json.

Thank you very much for your collaboration @cristobalvch

@cristobalvch
Copy link
Contributor Author

cristobalvch commented Aug 22, 2025

I'm on it!

@vmayoral
Copy link
Member

@cristobalvch ping us whenever this is ready for another review and thanks for the contrib!

@cristobalvch
Copy link
Contributor Author

yes thanks for waiting me! It's almost complete. I just have to run again some evaluations to add the results sections with better metrics performance :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants