Refusal-Trained LLMs Are Easily Jailbroken As Browser Agents Paper • 2410.13886 • Published Oct 11, 2024
HiL-Bench (Human-in-Loop Benchmark): Do Agents Know When to Ask for Help? Paper • 2604.09408 • Published 15 days ago • 5
HiL-Bench (Human-in-Loop Benchmark): Do Agents Know When to Ask for Help? Paper • 2604.09408 • Published 15 days ago • 5
SciPredict: Can LLMs Predict the Outcomes of Scientific Experiments in Natural Sciences? Paper • 2604.10718 • Published Apr 12 • 4