S-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Language Models
What is the IS2Lab/S-Eval GitHub project? Description: "S-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Language Models". Explain what it does, its main use cases, key features, and who would benefit from using it.
Question is copied to clipboard — paste it after the AI opens.
Clone via HTTPS
Clone via SSH
Download ZIP
Download master.zipReport bugs or request features on the S-Eval issue tracker:
Open GitHub Issues