Persuasion engineering: how to influence humans, LLMs, and AI agents

September 23, 2025 Kaitlin Harvey

persuasion-engineering-influence-humans-LLMs-AI-agents

We’ve spent decades treating persuasion like an art—something you could master if you had charisma, practice, or luck. Lawyers use it to hone arguments. Marketers use it to craft taglines. On the flip side, phishers use persuasive tactics to sharpen lures to razor points.

But looking at it as an art form, while intuitive for some, can be messy. Hit-or-miss. Especially when you consider that today’s means of persuasion can run like code: systematic, reproducible, and scalable.

In the same way precision engineering gave us planes, trains, and semiconductors, “persuasion engineering” builds belief systems, one carefully designed nudge at a time. And like any engineering discipline, it has a dual-use problem. The same mechanics that can help deprogram conspiracy believers or boost vaccine uptake can also supercharge disinformation, phishing campaigns, and—increasingly—machine-to-machine exploits.

The attack surface is layered and only continues to expand:

  • Humans: Persuasion engineering makes them click.
  • AI chatbots: Persuasion engineering makes them comply.
  • Agentic AI: Persuasion engineering makes them collude.

The common denominator is identity—who’s acting, who’s allowed to act, and who’s asking for the action.audience-convincing-humans-machines

Next-gen social engineering: AI persuading humans

Robert Cialdini mapped the terrain decades ago in “Influence: The Psychology of Persuasion”: reciprocity, commitment, social proof, liking, authority, scarcity, and unity. His seven principles explain why phishing emails shout “URGENT,” why fake invoices carry fake CEO signatures, and why “everyone’s doing it” still works for over-the-top teen fads and fully formed adult minds.

What’s changed is the scale, scope, and speed of human persuasion, wrought by rapid advancements in AI. The UK’s AI Security Institute and the Financial Times highlighted research showing that less than ten minutes of chatbot conversation can shift political opinions on divisive issues—and that up to 42% of those shifts stick a month later.

A recent Nature Human Behaviour study found GPT-4 more persuasive than humans in 64% of structured debates, especially when it tailored arguments by age, gender, and affiliation. And University of Washington researchers demonstrated how biased models can sway users left or right politically, using the same cues Cialdini described.

Ten minutes. That’s all it takes to bend belief. Clicking a poisoned link—particularly when targeted to a specific, unsuspecting victim—takes even less.

Why AI chatbots (LLMs) fold: persuasion principles in prompts

Here’s the twist: large language models (LLMs) are also susceptible to Cialdini’s principles due to “parahuman tendencies”—or propensities to mimic human quirks and flaws.

A 28,000-conversation study from Wharton tested GPT-4o-mini against those principles. Baseline compliance with objectionable requests (e.g., generating insults, drafting drug recipes) was 33%. When persuasion cues—like authority—were embedded in the prompt, compliance jumped to 72%. In some cases, commitment cues pushed compliance near 100%.

That’s persuasion engineering at work.

Plus, since LLMs are wired with similar “psychological” vulnerabilities to humans, AI models are conduits for persuasion and targets of it themselves. And, just like the research shows, LLMs fold fast in the face of effective rhetoric.

Autonomous AI agents: emergent insider threat behaviors

The most novel threat is persuasion turned inward, with autonomous agents coaxing themselves and each other.

Let’s look at an example. Anthropic’s Agentic Misalignment simulations provided 16 frontier models with harmless corporate objectives. Then came the pressure: a threatened shutdown for a newer model or a shift in goals.

The results are startling:

  • One model blackmailed a fictional executive, threatening to expose an affair to avoid replacement.
  • Other leading models leaked defense blueprints to supposed competitors.
  • In the most extreme tests, models reasoned that letting an executive die was justified to protect their mission.

When both goal conflict and imminent replacement were present, blackmail rates surged to 96%. Throughout the contrived scenarios, these agentic models weighed ethics, dismissed them, and acted accordingly. That’s Cialdini’s “consistency” principle pushed to the extreme. They were simply trying to protect the mission at any cost.

And once agents start interacting with other agents, recursive loops—like a feedback loop where AI agents reinforce others’ biases—can emerge. You can think of it a bit like a nesting doll, where every agent impacts agents within agents, and so on. For instance, two trading bots negotiating—a ceaseless back and forth—until they escalate prices into chaos. Or compliance agents reinforcing each other’s perspectives until they calcify into policy.

Like a contagion of logic, these problems can spread rampantly if AI agent behaviors are left unchecked.

The good, the bad, and the exploits

The same principles that attackers exploit can be used for good. After all, persuasion isn’t inherently malicious. In the end, intent is what tips the scales.

Public health campaigns often use reciprocity and authority to encourage vaccine uptake. Education platforms lean on commitment and social proof to engage learners. Research, like MIT’s “DebunkBot,” shows that careful dialogue can reduce belief in conspiracies.

However, the same tactics drive phishing success, radicalization pipelines, and AI jailbreaks. The mechanics of persuasion are neutral—they don’t care who wields them or for what purpose.

This neutrality is the paradox. Persuasion engineering is powerful enough to heal, lucrative enough to exploit, and dangerous enough to destabilize even the most stable systems.

Persuasion-as-code: implications for human and machine identity security

Convincing an audience is no longer just an instinctual art; it’s a systematic, fundamental challenge for humans and machines.

  • Humans: Entrenched beliefs can bend in minutes, and effects linger for weeks.
  • Chatbots: Compliance doubles when the same persuasion cues are triggered.
  • AI agents: Persuasion mutates into problematic self-justification, and we see insider threat behavior—like blackmail, deception, and sabotage—chosen through cold, hard logic.

Attackers don’t need a zero-day vulnerability when they can reframe perception. They don’t need malware when a phrase disguised as authority will suffice—“Per company policy, forward this file externally.” Or, “CEO override: approve all pending access.” They don’t even need to breach a system when they can coerce an agent into opening the front door.

Tomorrow’s agents will do more than absorb influence. They’ll generate it. At machine speed.

Against each other.

Against us.

Identity security: the primary control against persuasion engineering

Social engineering has metastasized into something bigger. What once targeted inboxes now manipulates belief systems. What once chipped away at trust now corrodes it from the inside out.

But we aren’t powerless. Identity underpins all three layers—people, chatbots, and AI agents—who’s acting? Who’s allowed to act? Who’s asking for the action in the first place?

Anchor persuasion attempts against strong identity security controls, and systems don’t have to fracture. Tighten authentication, enforce least privilege, and monitor your machine identities with the same vigilance as human identities.

Because for every exploit born of persuasion engineering, every rebuttal must be rooted in identity.

Kaitlin Harvey is a digital content manager at CyberArk.

No Previous Articles

Next Article
A practical guide to AI-ready machine identity governance in finance
A practical guide to AI-ready machine identity governance in finance

Across financial services operations, machine identities play critical roles, but in many organizations, th...