r/accelerate Singularity by 2026 4d ago

News Daily AI Archive 8/27/2025

  • Anthropic paper | Detecting and countering misuse of AI: August 2025 - Agentic LMs now execute full-spectrum intrusion and fraud: a vibe hacking crew ran Claude Code with a persistent CLAUDE.md to encode TTPs, automate OSINT targeting, scan VPNs, enumerate AD, steal creds, move laterally, build evasion malware (obfuscated Chisel, new TCP proxies masked as MSBuild.exe), exfiltrate data, price ransoms, and drop boot-embedded HTML notes; NK operators simulate competence to pass interviews and ship daily work; a UK no-code RaaS ships ChaCha20+RSA with FreshyCalls/RecycledGate and shadow copy wipes; a China actor spans 12 ATT&CK tactics; AI now powers MCP stealer-log profiling, carding stores, romance bots, and synthetic IDs. Mitigations include bans, tailored classifiers, malware-gen detection, and IOC sharing, but the skill curve is collapsing to zero, so defense must field autonomous, continuously learning counter-agents at internet scale. https://www.anthropic.com/news/detecting-countering-misuse-aug-2025; https://www-cdn.anthropic.com/b2a76c6f6992465c09a6f2fce282f6c0cea8c200.pdf
  • Anthropic launched a National Security Advisory Council with 11 senior U.S. natsec leaders to shape AI use in defense, intelligence, and science, tied to Claude Gov models, a $200M DoD deal, 10k LLNL users, NNSA safeguards, $1 gov access, and joint model stress-testing for bio, cyber, and R&D risks. https://www.anthropic.com/news/introducing-the-anthropic-national-security-and-public-sector-advisory-council
  • Google has integrated Gemini CLI into the Zed code editor, allowing developers to generate, refactor, and review code with AI directly in their IDE while maintaining full control. https://developers.googleblog.com/en/gemini-cli-is-now-integrated-into-zed/
  • OpenAI + Anthropic ran cross-lab safety tests on each other’s public models. Claude 4 excelled at instruction hierarchy + prompt-extraction but was weaker on jailbreaks and often refused answers in hallucination tests; OpenAI o3/o4-mini resisted jailbreaks better, answered more, but hallucinated more; GPT-4o/4.1 were more jailbreak-prone yet sometimes best at person-hallucination accuracy. Scheming results were mixed across labs; reasoning sometimes helped, sometimes worsened. OpenAI says GPT-5 improved sycophancy, hallucinations, and misuse resistance; cross-lab testing surfaced useful gaps, showing value of ongoing joint safety evals. https://openai.com/index/openai-anthropic-safety-evaluation/
  • You will soon be able to branch conversations in ChatGPT allowing branching of a conversation to a new conversation after a response https://x.com/btibor91/status/1960623245956411548
  • OpenAI has open sourced their benchmark called HeathBench under MIT license on huggingaface today https://huggingface.co/datasets/openai/healthbench
  • PixVerse has released PixVerse V5 of their video gen model and it scores 2nd place on I2V and 3rd place on T2V on Artificial Analysis above Veo3 in both cases but slightly worse than SeeDance 1.0 but the upside is its significantly cheaper than Veo 3 and its even cheaper than SeeDance Which makes it an amazing price to performance ratio video model https://x.com/PixVerse_/status/1960730919993799024
  • OpenAI released big Codex updates: https://help.openai.com/en/articles/6825453-chatgpt-release-notes#h_dcaac4ec67
    • IDE Extension: The new extension brings codex into VS Code, Cursor, and other VS Code forks, so that you can seamlessly preview local changes and edit code
    • Sign in with ChatGPT: Available in both the IDE and CLI, eliminating API key setup and providing access directly through your existing ChatGPT plan
    • Seamless Local ↔ Cloud Handoff: Developers can pair with Codex locally and then delegate tasks to the cloud to execute asynchronously without losing state
    • Upgraded Codex CLI: Refreshed UI, new commands, and bug fixes
    • Code reviews in GitHub: Set up Codex to automatically review new PRs in a repo, or mention u/codex in PRs to get reviews and suggested fixes
  • Prime Intellect launched the Environments Hub, an open community platform for creating, sharing, and scaling RL environments to advance open-source AGI. The hub, along with their open-source RL infrastructure (prime-rl), aims to lower barriers to training and serving large agentic models by providing accessible compute, tools, and RFT. They also released SYNTHETIC-2, a planetary-scale dataset of four million verified reasoning traces, and introduced the Prime Collective Communications Library (PCCL) for decentralized global training. https://www.primeintellect.ai/blog/environments
  • Kimi released a new feature text to slides pretty self explanatory but cool for free of course https://x.com/crystalsssup/status/1960912750068273186
  • Tencent released HunyuanVideo-Foley which builds a TV2A stack that fixes data scarcity, modality imbalance, and mediocre audio by scaling a 100k-hour pipeline (8 s chunking, silence/SNR/bandwidth filters, AudioBox-aesthetics gating, ImageBind/AV-align checks, GenAU captions), then training a flow-matching hybrid with N1 dual-stream MMDiT blocks and N2 audio-only DiT blocks modulated by Synchformer sync features and interleaved RoPE for frame-level A/V coupling; text enters later via cross-attention to prevent text dominance. A REPA loss aligns mid-layer DiT states to ATST-Frame features through cosine similarity, stabilizing training and boosting fidelity; an enhanced DAC-VAE swaps RVQ for continuous 128-dim, 50 Hz latents at 48 kHz to improve reconstruction. Trained at scale (18 MMDiT + 36 DiT, d=1536, 12 heads, CFG 0.1), it lands SoTA on audio quality, visual-semantic alignment, and sync on Kling-Audio-Eval and MovieGen-Audio-Bench, with VGGSound distribution gaps likely due to its low-grade audio. Ablations show joint A/V self-attention followed by text cross-attention, interleaved RoPE, and shallow-layer REPA on the unimodal branch (ATST > EAT, EAT+ATST harmful) drive the gains. If reproducibility holds, this is a serious step toward fully automatic, pro-grade Foley for any video stream, compressing human post-production into a programmable primitive. https://huggingface.co/tencent/HunyuanVideo-Foley; paper; https://arxiv.org/abs/2508.16930: code: https://github.com/Tencent-Hunyuan/HunyuanVideo-Foley

let me know if I missed anything

9 Upvotes

3 comments sorted by

3

u/stealthispost Acceleration Advocate 3d ago

Branching conversations is dope

2

u/pigeon57434 Singularity by 2026 3d ago

it might still be a couple weeks before it actually ships but its in the works

2

u/stealthispost Acceleration Advocate 3d ago

it's kinda sad it took this long to make what is a pretty crucial feature