
A new study using graph-based tasks reveals that large language models employ two parallel mechanisms for in-context learning rather than relying on a single approach. Through PCA analysis and causal intervention techniques, researchers show that LLMs simultaneously encode global topology information and local transition patterns, with late-layer circuits responsible for transferring structural preferences between different graph configurations.
Researchers propose CoCoDA, a new framework that organizes tool libraries as compositional code DAGs (directed acyclic graphs) to help smaller language models efficiently use external skills without exceeding context budgets. The system enables an 8B parameter model to match or exceed performance of a 32B model on mathematical reasoning benchmarks like GSM8K and MATH by co-evolving the tool library and planner in tandem, using typed retrieval and composition-aware rewards.
Ilya Sutskever, OpenAI's former chief scientist, testified Monday that his involvement in Sam Altman's removal from the company was motivated by concerns about its direction, not malice. Despite his current estrangement from OpenAI, Sutskever publicly defended the company during the legal proceedings.
OpenAI has unveiled Daybreak, a new security initiative that uses its Codex Security AI agent to identify and patch software vulnerabilities before attackers can exploit them. The launch comes weeks after rival Anthropic released Claude Mythos, a security-focused model designed to uncover dangerous code flaws through AI-driven threat modeling and automated risk detection.
Google has identified the first known instance of criminal hackers leveraging artificial intelligence to discover and exploit a previously unknown software flaw, marking a significant escalation in attack sophistication. Security experts characterize the incident as a harbinger of emerging threats, warning that AI-powered vulnerability discovery will likely become more prevalent among threat actors.
Google has disclosed that threat actors used artificial intelligence to develop a significant security flaw, marking a notable escalation in how attackers are leveraging AI capabilities. The revelation highlights growing concerns about the dual-use nature of AI tools and their potential to accelerate vulnerability discovery and exploitation.
Anthropic has introduced Claude Platform as a native offering on Amazon Web Services, allowing users to access the AI model directly through their AWS accounts. This integration strengthens the partnership between Anthropic and AWS, providing enterprises with streamlined deployment and management of Claude within their existing cloud infrastructure.
A developer has created a from-scratch ML compiler that lowers language models like TinyLlama and Qwen2.5-7B to optimized CUDA kernels through six intermediate representations. The compiler achieves 1.11× speedup over PyTorch eager execution and 1.20× over torch.compile on RTX 5090, with selective wins reaching 4.7× on operations like attention and KV projections.
Thinking Machines, founded by former OpenAI CTO Mira Murati, announced a new approach called 'interaction models' that process audio, video, and text simultaneously to enable real-time AI collaboration. Unlike current models that wait passively for users to finish typing or speaking, interaction models will perceive and respond to user actions continuously, mimicking natural human-to-human collaboration.
Google has stated it likely prevented a coordinated effort by hackers to use artificial intelligence for a large-scale exploitation campaign targeting software vulnerabilities. The incident underscores growing concerns that threat actors are rapidly adopting AI tools to discover previously unknown security flaws and scale attacks more efficiently.
Researchers mechanistically analyzed three popular vision-language models (LLaVA-1.5, PaliGemma, Qwen2-VL) and found that sharp attention maps—long assumed to signal model confidence—are nearly useless predictors of correctness, with near-zero correlation. Instead, model reliability is encoded in hidden-state geometry and sparse late-layer circuits, with hidden-state probes achieving >0.95 AUROC and self-consistency emerging as the strongest behavioral predictor. The study also reveals critical architectural differences: late-fusion models like LLaVA concentrate reliability in a fragile bottleneck, while early-fusion models distribute it robustly.