Here's your daily roundup of the most relevant AI and ML news for April 16, 2026. Today's digest includes 2 security-focused stories. We're also covering 6 research developments. Click through to read the full articles from our curated sources.
Security & Safety
1. Show HN: HEOR Agent. AI for health economics research from Claude
Article URL: https://github.com/neptun2000/heor-agent-mcp Comments URL: https://news.ycombinator.com/item?id=47792860 Points: 2
Comments: 0
Source: Hacker News - ML Security | just now
2. OpenAI Launches GPT-5.4-Cyber with Expanded Access for Security Teams
OpenAI on Tuesday unveiled GPT-5.4-Cyber, a variant of its latest flagship model, GPT‑5.4, that's specifically optimized for defensive cybersecurity use cases, days after rival Anthropic unveiled its own frontier model, Mythos. "The progressive use of AI accelerates defenders – tho...
Source: The Hacker News (Security) | 1 day ago
Research & Papers
3. Robust Adversarial Policy Optimization Under Dynamics Uncertainty
arXiv:2604.10974v2 Announce Type: replace Abstract: Reinforcement learning (RL) policies often fail under dynamics that differ from training, a gap not fully addressed by domain randomization or existing adversarial RL methods. Distributionally robust RL provides a formal remedy but still relies...
Source: arXiv - Machine Learning | 10 hours ago
4. The Mirror Design Pattern: Strict Data Geometry over Model Scale for Prompt Injection Detection
arXiv:2603.11875v2 Announce Type: replace-cross Abstract: Prompt injection defenses are often framed as semantic understanding problems and delegated to increasingly large neural detectors. For the first screening layer, however, the requirements are different: the detector runs on every request...
Source: arXiv - AI | 10 hours ago
5. SelfGrader: Stable Jailbreak Detection for Large Language Models using Token-Level Logits
arXiv:2604.01473v2 Announce Type: replace-cross Abstract: Large Language Models (LLMs) are powerful tools for answering user queries, yet they remain highly vulnerable to jailbreak attacks. Existing guardrail methods typically rely on internal features or textual responses to detect malicious qu...
Source: arXiv - AI | 10 hours ago
6. Artificial intelligence application in lymphoma diagnosis with Vision Transformer using weakly supervised training
arXiv:2604.13795v1 Announce Type: cross Abstract: Vision transformers (ViT) have been shown to allow for more flexible feature detection and can outperform convolutional neural network (CNN) when pre-trained on sufficient data. Due to their promising feature detection capabilities, we deployed V...
Source: arXiv - Machine Learning | 10 hours ago
7. TREX: Automating LLM Fine-tuning via Agent-Driven Tree-based Exploration
arXiv:2604.14116v1 Announce Type: new Abstract: While Large Language Models (LLMs) have empowered AI research agents to perform isolated scientific tasks, automating complex, real-world workflows, such as LLM training, remains a significant challenge. In this paper, we introduce TREX, a multi-ag...
Source: arXiv - AI | 10 hours ago
8. Multi-Task LLM with LoRA Fine-Tuning for Automated Cancer Staging and Biomarker Extraction
arXiv:2604.13328v1 Announce Type: new Abstract: Pathology reports serve as the definitive record for breast cancer staging, yet their unstructured format impedes large-scale data curation. While Large Language Models (LLMs) offer semantic reasoning, their deployment is often limited by high comp...
Source: arXiv - Machine Learning | 10 hours ago
About This Digest
This digest is automatically curated from leading AI and tech news sources, filtered for relevance to AI security and the ML ecosystem. Stories are scored and ranked based on their relevance to model security, supply chain safety, and the broader AI landscape.
Want to see how your favorite models score on security? Check our model dashboard for trust scores on the top 500 HuggingFace models.