First-hand information for everyone
Researchers introduce AVISE (AI Vulnerability Identification and Security Evaluation), a modular open-source framework to find security prob
This paper introduces AVISE (AI Vulnerability Identification and Security Evaluation), a modular, open‑source framework meant to help resear
AI safety problems sometimes hide across many short interactions. A single conversation or log file can look harmless, but when a small set
This paper presents ThinkJEPA, a method that combines two ways of understanding video to predict future states for tasks like hand-manipulat
This paper studies how large language models (LLMs) organize moral thinking across multiple intermediate steps. The authors introduce the id
This paper asks a simple question: how much do large language models (LLMs) already know about sounds from text-only training, and does that
This paper introduces FinTradeBench, a benchmark that tests whether Large Language Models (LLMs) can reason about both company fundamentals
This paper looks at how we check the checkers for long-form question-answering systems. The authors focus on ScholarQA-CS2, a benchmark for
Large language models can reason in impressive ways. But they also make systematic reasoning mistakes that are hard to fix with broad retrai
This paper introduces LieCraft, a new evaluation framework and sandbox for measuring deception in large language models (LLMs). In plain ter