arXiv News

First-hand information for everyone

Legal

Privacy PolicyTerms of Use

© 2026 arXiv News

arXiv News

EnglishJapanese

Switch language

EnglishJapanese
Loading account…

First-hand information for everyone

Latest
AVISE: an open framework that automates finding jailbreaks in language modelsAVISE: an open‑source framework that automates tests for AI security and jailbreaksMeerkat: a new way to find safety failures that only appear across many AI tracesThinkJEPA blends a dense video predictor with a vision–language “thinker” to forecast longer-range hand movementsHow large language models switch moral frameworks while they reasonStudy measures how much sound knowledge lives inside language models and how that affects audio AIFinTradeBench: a new test that asks language models to combine company filings and market price signalsStudy shows simple human preferences can hide problems when judging long scientific answersREdit: reshaping neural circuits to edit specific reasoning errors in large language modelsLieCraft: a sandbox game that tests whether large language models will lie to meet goalsAVISE: an open framework that automates finding jailbreaks in language modelsAVISE: an open‑source framework that automates tests for AI security and jailbreaksMeerkat: a new way to find safety failures that only appear across many AI tracesThinkJEPA blends a dense video predictor with a vision–language “thinker” to forecast longer-range hand movementsHow large language models switch moral frameworks while they reasonStudy measures how much sound knowledge lives inside language models and how that affects audio AIFinTradeBench: a new test that asks language models to combine company filings and market price signalsStudy shows simple human preferences can hide problems when judging long scientific answersREdit: reshaping neural circuits to edit specific reasoning errors in large language modelsLieCraft: a sandbox game that tests whether large language models will lie to meet goals

Today's Briefing

Friday, May 1, 2026
AllArtificial IntelligenceMachine LearningNatural Language ProcessingComputer VisionRoboticsCryptographyPhysicsMathematics
Artificial IntelligenceFeatured briefing

AVISE: an open framework that automates finding jailbreaks in language models

Researchers introduce AVISE (AI Vulnerability Identification and Security Evaluation), a modular open-source framework to find security prob

April 24, 2026EN2 min read
Read full article

Latest Research

Artificial Intelligence
April 23, 2026

AVISE: an open‑source framework that automates tests for AI security and jailbreaks

This paper introduces AVISE (AI Vulnerability Identification and Security Evaluation), a modular, open‑source framework meant to help resear

EN
2 min read
Artificial Intelligence
April 14, 2026

Meerkat: a new way to find safety failures that only appear across many AI traces

AI safety problems sometimes hide across many short interactions. A single conversation or log file can look harmless, but when a small set

EN
2 min read
Artificial Intelligence
March 24, 2026

ThinkJEPA blends a dense video predictor with a vision–language “thinker” to forecast longer-range hand movements

This paper presents ThinkJEPA, a method that combines two ways of understanding video to predict future states for tasks like hand-manipulat

EN
2 min read
Advertisement
Artificial Intelligence
March 23, 2026

How large language models switch moral frameworks while they reason

This paper studies how large language models (LLMs) organize moral thinking across multiple intermediate steps. The authors introduce the id

EN
2 min read
Natural Language Processing
March 20, 2026

Study measures how much sound knowledge lives inside language models and how that affects audio AI

This paper asks a simple question: how much do large language models (LLMs) already know about sounds from text-only training, and does that

EN
2 min read
Artificial Intelligence
March 20, 2026

FinTradeBench: a new test that asks language models to combine company filings and market price signals

This paper introduces FinTradeBench, a benchmark that tests whether Large Language Models (LLMs) can reason about both company fundamentals

EN
2 min read
Natural Language Processing
March 14, 2026

Study shows simple human preferences can hide problems when judging long scientific answers

This paper looks at how we check the checkers for long-form question-answering systems. The authors focus on ScholarQA-CS2, a benchmark for

EN
2 min read
Natural Language Processing
March 14, 2026

REdit: reshaping neural circuits to edit specific reasoning errors in large language models

Large language models can reason in impressive ways. But they also make systematic reasoning mistakes that are hard to fix with broad retrai

EN
2 min read
Artificial Intelligence
March 14, 2026

LieCraft: a sandbox game that tests whether large language models will lie to meet goals

This paper introduces LieCraft, a new evaluation framework and sandbox for measuring deception in large language models (LLMs). In plain ter

EN
2 min read
Next page of briefings