Here's your daily roundup of the most relevant AI and ML news for January 09, 2026. We're also covering 8 research developments. Click through to read the full articles from our curated sources.
Research & Papers
1. Defense Against Indirect Prompt Injection via Tool Result Parsing
arXiv:2601.04795v1 Announce Type: new Abstract: As LLM agents transition from digital assistants to physical controllers in autonomous systems and robotics, they face an escalating threat from indirect prompt injection. By embedding adversarial instructions into the results of tool calls, attack...
Source: arXiv - AI | 1 hours ago
2. Latent Fusion Jailbreak: Blending Harmful and Harmless Representations to Elicit Unsafe LLM Outputs
arXiv:2508.10029v2 Announce Type: replace-cross Abstract: While Large Language Models (LLMs) have achieved remarkable progress, they remain vulnerable to jailbreak attacks. Existing methods, primarily relying on discrete input optimization (e.g., GCG), often suffer from high computational costs ...
Source: arXiv - AI | 1 hours ago
3. Crafting Adversarial Inputs for Large Vision-Language Models Using Black-Box Optimization
arXiv:2601.01747v2 Announce Type: replace-cross Abstract: Recent advancements in Large Vision-Language Models (LVLMs) have shown groundbreaking capabilities across diverse multimodal tasks. However, these models remain vulnerable to adversarial jailbreak attacks, where adversaries craft subtle p...
Source: arXiv - Machine Learning | 1 hours ago
4. Know Thy Enemy: Securing LLMs Against Prompt Injection via Diverse Data Synthesis and Instruction-Level Chain-of-Thought Learning
arXiv:2601.04666v1 Announce Type: new Abstract: Large language model (LLM)-integrated applications have become increasingly prevalent, yet face critical security vulnerabilities from prompt injection (PI) attacks. Defending against PI attacks faces two major issues: malicious instructions can be...
Source: arXiv - AI | 1 hours ago
5. Measuring and Fostering Peace through Machine Learning and Artificial Intelligence
arXiv:2601.05232v1 Announce Type: cross Abstract: We used machine learning and artificial intelligence: 1) to measure levels of peace in countries from news and social media and 2) to develop on-line tools that promote peace by helping users better understand their own media diet. For news media...
Source: arXiv - Machine Learning | 1 hours ago
6. Quaternion-Hadamard Network: A Novel Defense Against Adversarial Attacks with a New Dataset
arXiv:2502.10452v3 Announce Type: replace Abstract: Adverse-weather image restoration (e.g., rain, snow, haze) models remain highly vulnerable to gradient-based white-box adversarial attacks, wherein minimal loss-aligned perturbations cause substantial degradation in the restored output. This pa...
Source: arXiv - Machine Learning | 1 hours ago
7. Prior-Informed Zeroth-Order Optimization with Adaptive Direction Alignment for Memory-Efficient LLM Fine-Tuning
arXiv:2601.04710v1 Announce Type: cross Abstract: Fine-tuning large language models (LLMs) has achieved remarkable success across various NLP tasks, but the substantial memory overhead during backpropagation remains a critical bottleneck, especially as model scales grow. Zeroth-order (ZO) optimi...
Source: arXiv - Machine Learning | 1 hours ago
8. ResMAS: Resilience Optimization in LLM-based Multi-agent Systems
arXiv:2601.04694v1 Announce Type: new Abstract: Large Language Model-based Multi-Agent Systems (LLM-based MAS), where multiple LLM agents collaborate to solve complex tasks, have shown impressive performance in many areas. However, MAS are typically distributed across different devices or enviro...
Source: arXiv - AI | 1 hours ago
About This Digest
This digest is automatically curated from leading AI and tech news sources, filtered for relevance to AI security and the ML ecosystem. Stories are scored and ranked based on their relevance to model security, supply chain safety, and the broader AI landscape.
Want to see how your favorite models score on security? Check our model dashboard for trust scores on the top 500 HuggingFace models.