AI company ethical restrictions are contractually penetrable through multi-tier deployment chains because Anthropic's autonomous weapons restrictions did not prevent Claude's use in combat targeting via Palantir's separate contract
Claude is being used for AI-assisted combat targeting in the Iran war via Palantir's Maven integration, generating target lists and ranking them by strategic importance, while Anthropic simultaneously argues in court that it should be allowed to restrict autonomous weapons use. Hunton & Williams notes that 'Claude remains on classified networks via Palantir's existing contract (Palantir is not designated a supply chain risk). The supply chain designation targets direct Anthropic contracts, not Palantir reselling Claude.' This reveals a structural loophole: Anthropic's ethical restrictions on autonomous weapons use do not apply when Claude is deployed through Palantir's separate government contract. The multi-tier deployment chain—Anthropic to Palantir to DoD Maven—means voluntary safety commitments are contractually penetrable. Anthropic's restrictions bind only its direct contracts, not downstream use by intermediaries. This is not a technical failure but an architectural one: voluntary ethical constraints cannot survive multi-party deployment chains where each tier operates under separate agreements. The most consequential use case (combat targeting) occurs through the exact channel that Anthropic's restrictions do not cover. This demonstrates that AI company safety pledges are structurally insufficient when deployment architectures involve intermediary contractors with independent government relationships.