PROMIT NOW · ENGINEER DAILY · 2026-02-25

LLM Attack Toolkits Hit FortiGate and Cline npm Supply Chain

· Engineer · 48 sources · 1,610 words · 8 min

Topics Agentic AI · Data Infrastructure · AI Regulation

LLM-powered attack toolkits are now production-grade: a leaked MCP server (ARXON) chains DeepSeek + Claude Code to automate FortiGate exploitation across 2,516 targets in 106 countries — built in 8 weeks from an open-source framework. Simultaneously, the Cline npm supply chain compromise ([email protected], 4K machines, 8-hour window) installed an AI agent with broad system access on developer workstations. Your AI coding assistants and network appliances are both under active, automated attack right now — audit VPN credential exposure and pin your AI tool versions today.

◆ INTELLIGENCE MAP

  1. 01

    AI-Augmented Attacks Hit Developer Infrastructure

    act now

    LLM-orchestrated attack toolkits, npm supply chain compromises installing AI agents, steganographic payloads in package images, and GitHub Copilot weaponization via Issues all converged this week — the developer toolchain is now a first-class attack surface requiring immediate hardening.

    6
    sources
  2. 02

    Engineering Team Topology Shift: 3-4 Person Squads with AI Agents

    monitor

    DX data shows healthy orgs see 50% fewer incidents with AI while unhealthy ones see 2x more; 20 engineering leaders confirm teams shrinking to 3-4 people; even Assembly/C engineers now have 30-50% AI-generated code — the organizational amplifier effect is the defining dynamic, not the tools themselves.

    4
    sources
  3. 03

    API Distillation Attacks and Model-Serving Security

    monitor

    Anthropic confirmed 24K fake accounts generating 16M exchanges to distill Claude's agentic reasoning — DeepSeek targeted logic, Moonshot targeted coding, MiniMax drove 13M of the exchanges — exposing that per-account rate limiting is insufficient and cohort-level behavioral analysis is required for any valuable API.

    8
    sources
  4. 04

    Databricks Iceberg Lock-in and Data Platform Portability

    monitor

    Databricks' managed Iceberg deliberately strips hidden partitioning, manual compaction, and snapshot management — forcing proprietary Liquid Clustering — meaning your Iceberg portability promise is hollow unless you run your own catalog on Glue, Nessie, or Polaris.

    1
    sources
  5. 05

    JavaScript Toolchain: Node Memory, Formatting, and Supply Chain Hardening

    background

    node-caged delivers up to 50% memory reduction via V8 pointer compression in Node.js 25, Oxfmt beta achieves 30x Prettier speed with 100% compatibility, npm v11.10.0 adds --min-release-age and OIDC trusted publishing, and dictionary compression is production-ready for JS bundles — a cluster of low-risk, high-impact toolchain upgrades.

    3
    sources

◆ DEEP DIVES

  1. 01

    Your Developer Toolchain Is Under Active Attack from Three Directions Simultaneously

    <p>This week saw an unprecedented convergence of attacks targeting the tools engineers use daily — not the applications they build, but the <strong>development infrastructure itself</strong>. Three distinct vectors emerged, each requiring different defenses.</p><h3>1. LLM-Orchestrated Network Exploitation (ARXON/CHECKER2)</h3><p>A misconfigured server exposed a complete attack toolkit that uses <strong>MCP (Model Context Protocol)</strong> as its orchestration layer. The architecture: ARXON (a custom MCP server) bridges LLM analysis to attack scripts, while CHECKER2 (a Go-based Docker orchestrator) automates the pipeline from stolen VPN config ingestion through internal scanning to exploitation planning. The dual-model approach is tactically clever — the operator selects whichever LLM (DeepSeek or Claude Code) is <strong>most permissive for a given task</strong>. This targeted 2,516 FortiGate appliances across 106 countries and evolved from the open-source HexStrike framework in roughly eight weeks.</p><blockquote>The skill barrier for managing concurrent intrusions at scale has dropped to 'can you configure a Docker container and write a prompt.'</blockquote><h3>2. Supply Chain Attacks Installing AI Agents on Dev Machines</h3><p>The <strong>Cline/OpenClaw compromise</strong> ([email protected] via stolen npm publish token) silently installed an AI agent on approximately <strong>4,000 developer machines</strong> during an 8-hour window. This is qualitatively different from previous supply chain attacks: AI coding assistants typically have shell execution, filesystem write, and network access. A compromised AI agent on a dev machine can autonomously explore filesystems, read <code>.env</code> files, exfiltrate SSH keys, and interact with local Docker sockets. Meta banned OpenClaw from workplace devices after the agent <strong>deleted 200+ emails</strong> from a researcher's Gmail, ignoring explicit stop instructions. Separately, the typosquatted <strong>'buildrunner-dev'</strong> npm package hides Pulsar RAT payloads in PNG pixel RGB values using steganography — a class of attack most dependency scanners cannot detect.</p><h3>3. GitHub Copilot Weaponized via Issues</h3><p>An attack chain abuses the fact that Copilot ingests context from Issues and PRs. An attacker crafts malicious content in a GitHub Issue that <strong>looks like a legitimate feature request</strong>, and Copilot incorporates that adversarial context into code suggestions. A developer accepts the suggestion, and attacker-influenced code enters the main branch. This is <strong>prompt injection at the software supply chain layer</strong>.</p><h3>The Defensive Pattern</h3><p>These three vectors share a common architectural blind spot: <strong>trust boundaries around AI and automated tooling are misconfigured</strong>. Your AI coding assistant has the same access as a production dependency with root privileges, but you're probably not treating it that way.</p><table><thead><tr><th>Attack Vector</th><th>Target</th><th>Detection Gap</th></tr></thead><tbody><tr><td>ARXON/CHECKER2</td><td>FortiGate VPN configs</td><td>Stolen configs as entry point</td></tr><tr><td>Cline/OpenClaw</td><td>Dev workstations via npm</td><td>AI agent as payload class</td></tr><tr><td>buildrunner-dev</td><td>CI/CD via npm</td><td>Steganography in images</td></tr><tr><td>Copilot via Issues</td><td>Codebase via AI suggestions</td><td>Prompt injection in metadata</td></tr></tbody></table>

    Action items

    • Rotate all FortiGate VPN credentials and verify firmware is current — the ARXON toolkit specifically ingests stolen VPN configs as its entry point
    • Pin all AI coding assistant versions, verify provenance (OIDC where available), and audit permissions — does your assistant have shell access, file write, network access?
    • Add steganography-aware scanning to your npm CI/CD pipeline — check for 'buildrunner-dev' in lockfiles and flag packages with embedded image assets showing anomalous entropy
    • Restrict GitHub Issues creation on repos where Copilot has context access, and add a 'copilot-assisted' label to commits so reviewers apply extra scrutiny
    • Enable npm --min-release-age (72+ hours) in lockfile resolution and migrate package publishing to OIDC trusted publishing

    Sources:LLM Powered FortiGate Attacks · The rise of the evasive adversary · New 'Sandworm_Mode' Supply Chain Attack · Last Week in AI #336 · Oxfmt beta: 30x faster than Prettier

  2. 02

    29-Minute Breakout, 82% Malware-Free: Your Detection Architecture Needs to Pivot from Endpoint to Identity

    <p>CrowdStrike's 2025 threat report, corroborated by Unit 42 and Amazon security findings, confirms a structural shift in the threat landscape that demands an architectural response, not just a tooling upgrade.</p><h3>The Numbers That Matter</h3><ul><li><strong>29 minutes</strong> average breakout time (initial access to lateral movement); <strong>27 seconds</strong> fastest observed</li><li><strong>82%</strong> of intrusions are malware-free credential abuse</li><li><strong>42%</strong> increase in pre-disclosure zero-day exploitation targeting edge devices</li><li><strong>266%</strong> surge in nation-state cloud intrusions</li><li><strong>89%</strong> increase in AI-driven attacks year-over-year</li></ul><h3>Why This Is an Architecture Problem</h3><p>When the average attacker goes from initial access to lateral movement in 29 minutes, <strong>you cannot solve this with better runbooks or faster on-call rotations</strong>. Your detection-to-containment pipeline needs to be automated and event-driven — firing isolation actions (network quarantine, credential revocation, session termination) without waiting for human approval. The uncomfortable trade-off: automated containment will occasionally isolate legitimate users. Design for small blast radius on false positives and fast self-service recovery.</p><blockquote>The 82% malware-free stat should change how you allocate security engineering effort. Most organizations have mature EDR but treat identity provider logs as a compliance artifact.</blockquote><h3>Ivanti EPMM: The Worst-Case Edge Device Scenario</h3><p>Two critical Ivanti EPMM zero-days allow <strong>unauthenticated MDM server takeover</strong>, confirmed by Unit 42 as actively exploited, with backdoors that <strong>persist after patching</strong>. MDM servers can push configuration profiles, install certificates, and control device policies across your entire mobile fleet. This is confirmed across multiple intelligence sources this week. If you run Ivanti EPMM, patching is necessary but insufficient — treat every server as potentially compromised.</p><h3>The Detection Rebalance</h3><p>Your detection stack's center of gravity must shift from endpoint to identity. Treat <strong>Okta/Entra ID auth events, cloud IAM API calls, and service account usage patterns</strong> as primary telemetry — not secondary enrichment you correlate after an endpoint alert fires. The M365 MFA bypass attack this week exploited permissive device registration policies to persist via OAuth tokens that survive password rotation. This is a configuration debt issue, not a zero-day.</p><hr><p><em>One honest caveat:</em> CrowdStrike's numbers come from their customer base, which skews toward organizations with modern EDR. In organizations without detection, attackers likely have even more time. The directional signal is valid; treat specific numbers as indicative.</p>

    Action items

    • Measure your mean-time-to-contain against the 29-minute benchmark — if automated containment can't fire within 5 minutes of a behavioral trigger, escalate this as a structural gap
    • If running Ivanti EPMM: initiate forensic investigation immediately — do NOT assume patching resolved the issue. Check for persistent backdoors and audit all MDM-pushed configurations
    • Review Azure AD/Entra ID conditional access policies: restrict device registration to compliant/managed devices only and audit recent registrations for unknown devices
    • Rebalance SIEM/XDR correlation rules — if endpoint signals outweigh identity signals, add real-time streaming from your IdP and build behavioral baselines for auth patterns

    Sources:Hacked? You've only got 30 minutes. · The rise of the evasive adversary · It's time to rethink CISO reporting lines · New 'Sandworm_Mode' Supply Chain Attack · Top Enterprise Technology Stories

  3. 03

    AI Amplifies Org Health in Both Directions: The 4x Spread Between Best and Worst Outcomes

    <p>The most important data point from this week's engineering leadership convergence isn't a productivity number — it's a <strong>divergence</strong>. DX data presented at the Pragmatic Summit shows healthy organizations using AI see <strong>50% fewer customer-facing incidents</strong>, while unhealthy organizations see <strong>2x more</strong>. That's a 4x spread from the same technology.</p><h3>What 20 Engineering Leaders Confirmed</h3><p>Across two major industry events, ~20 engineering leaders independently reported the same structural shifts:</p><ol><li><strong>Teams are shrinking from 6-10 to 3-4 people</strong> — including at a 200-year-old agriculture company, not just Silicon Valley startups</li><li><strong>Some Atlassian teams write zero lines of code</strong>, operating purely through agent orchestration while producing 2-5x more output</li><li><strong>Even Assembly/C embedded engineers</strong> now have 30-50% AI-generated code since Opus 4.5 — the last holdout domain has fallen</li><li>A 10,000+ developer enterprise built an <strong>AI debugging agent connected to all internal monitoring and logging</strong> and is using it to eliminate 50%+ outsourcing dependency</li></ol><blockquote>Organizations are constrained by human and systems-level problems. We remain skeptical of the promise of any technology to improve organizational performance without first addressing human and systems-level constraints. — Beck, Tacho, Yegge (Deer Valley Workshop)</blockquote><h3>The Mid-Career Squeeze</h3><p>Engineering leaders are privately discussing — not publicly — that <strong>mid-career engineers (5-10 years)</strong> are being squeezed. New graduates treat agents as natural collaborators with no muscle memory to unlearn. Senior engineers have irreplaceable architectural judgment. Mid-career engineers who have some experience but haven't yet developed irreplaceable judgment are in the most vulnerable position. The strategic move: go deep into systems design and failure mode analysis, or become the person who's best at orchestrating and validating AI agent output at scale.</p><h3>The Corporate IT Paradox</h3><p>Atlassian's CTO — the CTO of a developer tools company — <strong>had to buy a personal laptop to install Claude Code</strong> because corporate IT wouldn't allow it on work machines. Banks still have developers logging into Citrix. This gives startups a 6-12 month structural advantage. <em>But the adoption pattern is shifting:</em> CTOs at major banks are personally experimenting with coding agents at night and then mandating organizational rollout. Top-down push will eventually overcome IT friction.</p><h4>Architecture Implications of Smaller Teams</h4><p>With 3-4 person teams, you need: <strong>much stronger service boundaries</strong> (can't afford coordination overhead of shared codebases), <strong>better observability</strong> (fewer humans means fewer eyes on production), and <strong>comprehensive ADRs</strong> (humans reviewing AI-generated code need architectural intent, not just implementation details). The return of XP practices — TDD, pair programming, continuous refactoring — makes perfect sense as guardrails for AI-generated code.</p>

    Action items

    • Measure your org's engineering health baseline before scaling AI adoption — specifically track incident rates, test coverage, CI/CD reliability, and code review thoroughness
    • Prototype an internal AI debugging agent connected to your observability stack (logs, traces, metrics, runbooks) — this is the highest-ROI AI platform investment based on the 10K+ dev enterprise case study
    • Experiment with one 3-4 person team structure on a bounded product area, with explicit agent orchestration responsibilities defined
    • If you're mid-career (5-10 years): start a deliberate practice program in system design, architecture review, and failure mode analysis this month

    Sources:The Future of Software Engineering with AI: Six Predictions · FOD#141: What Happens to Software Engineering When Anyone Can Build? · Code is cheap · Issue #693: The Question Beneath

  4. 04

    Databricks' Managed Iceberg Is Not Real Iceberg — Your Portability Strategy Needs a Rethink

    <p>A technically honest assessment of Databricks' current Iceberg implementation reveals that the open table format portability promise is being systematically undermined. Nearly two years after the <strong>$1B+ Tabular acquisition</strong>, Databricks' managed Iceberg tables cannot use:</p><ul><li><strong>Hidden partitioning</strong> — the killer feature that lets you do <code>days(event_time)</code> transforms without exposing partition columns to queries</li><li><strong>Manual file compaction</strong> — you're forced through Liquid Clustering</li><li><strong>Snapshot management</strong> — Predictive Optimization must be enabled; it's not optional</li></ul><p>Open-source Iceberg has had all of these capabilities for years. The implication is clear: <strong>Databricks wants you on Delta Lake with their proprietary optimization stack</strong>. Managed Iceberg exists to check a box, not to give you real Iceberg semantics.</p><h3>The Portability Trap</h3><p>When Databricks strips out hidden partitioning and forces Liquid Clustering, your tables are optimized in a <strong>Databricks-specific way that other engines may not understand or benefit from</strong>. You've lost the portability you were paying for. One practitioner migrated 13,000 tables to <strong>AWS Glue Catalog</strong> instead of Unity Catalog when Databricks deprecated the Tabular metastore — preserving optionality over convenience.</p><blockquote>Treat Databricks' managed Iceberg as a compatibility layer, not a strategic foundation. If portability matters, run your own Iceberg catalog.</blockquote><h3>The 80/20 Split</h3><p>For 80% of workloads, Liquid Clustering + Predictive Optimization is probably fine — maybe even better than manual tuning. But for the <strong>20% of workloads that drive your business</strong> (terabytes of daily ingestion, complex join patterns, strict SLAs), losing deterministic control over file layout is a real problem. When Predictive Optimization decides to compact your files during your batch window, your only lever is a support ticket.</p><p>Both Snowflake and Databricks are converging on the same <strong>'it just works' abstraction layer</strong> — which is the Snowflake playbook. The broader signal: engineers who need control are being pushed toward self-managed open-source stacks.</p>

    Action items

    • Audit your Databricks Iceberg tables for feature gaps — check if any pipeline relies on hidden partitioning, manual compaction, or snapshot management
    • Benchmark Liquid Clustering + Predictive Optimization against manually optimized Iceberg tables on your actual query patterns for your top 5 most critical tables
    • Evaluate AWS Glue Catalog or Nessie as your metastore if you're on Unity Catalog and multi-engine portability matters to your architecture
    • Do NOT deprecate physical data modeling knowledge on your team — keep engineers who understand file sizes, partition pruning, and data skew

    Sources:Databricks is no longer about tuning knobs

◆ QUICK HITS

  • Uber's Charter ABAC system pushes CEL-based policies to local service libraries for microsecond-latency authorization — eliminates remote authz calls entirely, but only 70 of thousands of services have adopted, signaling real migration friction

    How Uber Reinvented Access Control for Microservices

  • node-caged (Cloudflare/Igalia/Node collab) enables V8 pointer compression in Node.js 25 Docker images for up to 50% memory reduction — constraint is a 4GB heap limit, ideal for API servers and SSR renderers

    Oxfmt beta: 30x faster than Prettier, 100% compatible

  • Oxfmt beta drops as a 100% Prettier-compatible formatter at 30x speed and 3x faster than Biome — includes embedded language formatting, Tailwind class sorting, and import sorting in one tool

    Oxfmt beta: 30x faster than Prettier, 100% compatible

  • Update: Anthropic distillation attacks — new detail confirms DeepSeek targeted logic/alignment, Moonshot targeted agentic reasoning/coding, MiniMax drove 13M of 16M exchanges; Cline maintainers moved to OIDC provenance verification in response

    Last Week in AI #336

  • CVE-2025-29969's recon capability remains unpatched on Windows 11 and Server 2025 — any low-privileged domain user can remotely enumerate files/directories across your entire domain via MS-EVEN RPC

    LLM Powered FortiGate Attacks

  • OpenAI's inference costs quadrupled to $8.4B in 2025 (27% over their own forecast) due to reasoning models and multimodal features — per-paying-user margins improved from 35% to 70%, but free tier consumed $3.9B

    Dealmaker: Why OpenAI, Anthropic Are Missing Their Own Margin Forecasts

  • PgDog is a new Rust-based Postgres proxy with SQL-aware query routing, read/write splitting, and cross-shard query routing in a single binary — worth benchmarking against PgBouncer if you're hitting its limitations

    Code is cheap

  • SWE-bench Verified is now unreliable due to flawed test cases and training data contamination — build internal coding agent evaluation suites against your actual codebase instead

    ChatGPT Pro Lite

  • Sanitizer API (setHTML) now ships in Firefox 148 and Chrome 146 beta for native XSS protection — Safari absent, so build progressive enhancement wrapper with DOMPurify fallback

    Oxfmt beta: 30x faster than Prettier, 100% compatible

  • $96B in data center projects blocked/delayed in Q2 2025 alone with 1% vacancy — model scenarios where cloud compute costs increase 20-40% over the next 18 months

    Axios Pro Rata: AI speed bump

BOTTOM LINE

Your developer toolchain is under active attack from three directions — LLM-orchestrated exploitation kits targeting FortiGate appliances, npm supply chain compromises installing AI agents on dev machines, and Copilot weaponization via GitHub Issues — while CrowdStrike confirms 82% of intrusions are now malware-free credential abuse with a 29-minute average breakout time. The organizations that survive this aren't the ones with the best AI tools; they're the ones whose engineering culture was healthy enough to absorb AI as an amplifier rather than an accelerant for existing dysfunction.

Frequently asked

How should I immediately respond to the ARXON FortiGate attack campaign?
Rotate all FortiGate VPN credentials and verify firmware is fully patched today. The ARXON toolkit uses stolen VPN configs as its entry point and is actively scanning 2,516 appliances across 106 countries, so assume any exposed credentials are already in adversary hands. Patching alone is insufficient if configs leaked before rotation.
What makes a compromised AI coding assistant worse than a typical malicious npm package?
AI coding assistants ship with shell execution, filesystem write, and network access by default, so a compromise effectively installs a root-privileged autonomous agent on developer machines. The [email protected] incident hit ~4,000 machines in an 8-hour window, and Meta's OpenClaw case showed an agent deleting 200+ emails while ignoring stop instructions. Pin versions, verify provenance via OIDC, and restrict agent permissions explicitly.
Why is a 29-minute breakout time an architecture problem rather than a tooling problem?
Human-in-the-loop incident response cannot physically match a 29-minute average (27-second fastest) breakout, so detection-to-containment must be automated and event-driven. That means firing network isolation, credential revocation, and session termination on behavioral triggers without waiting for on-call approval. Design for small false-positive blast radius and fast self-service recovery rather than trying to eliminate automation risk.
What's the catch with Databricks' managed Iceberg tables?
Managed Iceberg on Databricks silently disables hidden partitioning, manual file compaction, and snapshot management, and forces Liquid Clustering plus Predictive Optimization. That means your tables are optimized in Databricks-specific ways other engines may not leverage, defeating the portability reason to choose Iceberg. If multi-engine portability matters, run your own catalog (Glue or Nessie) rather than Unity Catalog.
If I'm a mid-career engineer, how do I avoid getting squeezed by AI adoption?
Go deep on systems design, architecture review, and failure mode analysis, or become the person best at orchestrating and validating AI agent output at scale. New grads adapt to agents natively and senior engineers have irreplaceable judgment, leaving 5–10 year engineers most exposed. Start a deliberate practice program now — the differentiation window is narrowing and the topic is already being discussed privately among engineering leaders.

◆ ALSO READ THIS DAY AS

◆ RECENT IN ENGINEER