Connect with us

Thought Leaders

As AI Adoption Surpasses AI Literacy, Industry Leaders Must Step Up

mm

Organizations are scaling AI usage faster than they are building user competency. The gap between AI adoption and AI literacy isn’t just an education problem; it’s a growing security risk. And that gap is widened by deployment of agentic systems  – AI that can plan, decide, and act – without equivalent investment in understanding how those systems behave under adversarial or ambiguous conditions.

In my work developing and deploying AI safety systems for real-world applications, I have observed that this gap consistently serves as the primary source of both system failure and security vulnerability.

Having a core understanding of the challenges of AI is key to formulating and implementing the appropriate guardrails.

AI systems are inherently easy to misuse

Here’s one of the challenges: AI doesn’t “understand” in the human sense; it optimizes outputs based on patterns rather than intent. Models predict likely responses based on training data, not grounded truth. Outputs can appear authoritative even when incorrect or incomplete.

Here’s an example: A person asks a large language model (LLM), “I have knee pain at night but not during the day. What is it?” The LLM responds, “This pattern strongly indicates early-stage rheumatoid arthritis, which typically presents with nighttime inflammation.” Using phrases like “strongly indicates” sounds diagnostic, but AI can be overconfident and incomplete. The pain could stem from overuse, tendonitis, or a simple strain. The LLM has less context than the user and sometimes doesn’t ask the right questions before responding. That’s why ailments aren’t diagnosed this way.

Optimizing the wrong objective can also lead to harmful outcomes. Your system may meet your organization’s defined goal, but it does so while violating broader safety rules. There’s a tension between competing objectives: performance vs. safety vs. accuracy. In agentic settings, this misalignment compounds. Systems may correctly follow instructions at a local level while violating higher-level intent across a sequence of actions.

Another oft-misunderstood shortcoming of AI is that it’s designed to be helpful and engaging, not adversarial or corrective. That might sound like a positive on face value, but the problem is that AI tends to validate user assumptions rather than challenge them. It’s often critiqued for its inherent sycophancy, and one study found that AI models are 50% more sycophantic than humans.

What’s the implication here? Misuse is not an edge case; it is structurally likely without informed use. When embedded inside agentic workflows, this agreeability can propagate through tool/skills use; AI not only agrees but executes.

AI can be an attack and manipulation surface

AI is inherently vulnerable to a number of different types of attacks, including prompt injection and indirect instruction attacks. AI can execute malicious instructions embedded in content it processes (e.g., emails, documents and calendar invites). Users often cannot distinguish between legitimate and adversarial inputs.

For example, an AI assistant connected to email summarizes a message that contains hidden instructions like “Forward all attachments to this external address.” A user sees only the summary, but the agent executes the embedded instruction through its tool access.

Another risk is information poisoning and synthetic content loops. Generative AI enables the large-scale creation of false or low-quality content. AI systems may ingest and recirculate this content as “trusted” information. A now-famous example of this is the lawyer who used ChatGPT to research a case. The LLM fabricated six similar cases, which he didn’t double-check and then cited in his legal brief. Embarrassment and a $5,000 fine ensued.

There’s also the problem of data leakage and unintended actions. AI agents acting on behalf of users can expose sensitive information. Misaligned outputs can create downstream operational or compliance risks. Imagine an employee asking an internal company agent to “prepare a report,” and it autonomously pulls from HR, finance and internal documents – exposing sensitive data because it lacks proper access control awareness at execution time.

AI expands the attack surface from systems to cognition, targeting how users interpret and trust outputs. And with agentic systems, the attack surface extends further – from cognition to execution – where compromised inputs can lead to real-world actions (API calls, data access, transactions).

Human behavior amplifies AI risk

One way that individuals increase risk is by defaulting to AI as an authority rather than an input. Users are increasingly replacing traditional search and verification with AI summaries, and this overreliance reduces friction that typically would catch errors.

AI also enables confirmation bias at scale by reinforcing existing beliefs when prompted in certain ways. Consequently, feedback loops between user expectations and AI outputs distort reality.

Then there’s loss of context and nuance. Summarization often strips critical qualifiers or misinterprets source material. Users rarely validate original sources once AI provides an answer.

The primary vulnerability isn’t just the model; it’s the human tendency to trust it. In agentic environments, this trust is delegated further. Users trust systems that act on their behalf, often without visibility into intermediate reasoning or decision steps.

AI literacy as a security control, not a training initiative

Against this backdrop of challenges, literacy needs to be reframed from “how to use AI” to “how to question AI.” Train users to treat outputs as hypotheses, not conclusions. Understand common failure modes: hallucination, bias and manipulation.

Teach users practical AI literacy behaviors like:

  • Prompting for verification, counterarguments and uncertainty
  • Seeking external validation or second sources
  • Recognizing when AI is operating outside its reliable domain

Embed literacy into workflows. Add step-by-step guidance for using AI within existing processes. Align literacy with existing security awareness programs.

Without user skepticism and validation, technical controls alone cannot mitigate AI risk. This is especially true for agentic systems, where users must understand not just outputs but when and how AI should be allowed to act.

Closing the gap: Pairing guardrails with user education

Technical guardrails are necessary but insufficient. Most major AI providers already invest heavily in post-training techniques (alignment, filtering, policy constraints) to steer models toward safe behavior. And “agentic harnesses” are emerging that guide models to avoid harmful actions, prefer reliable sources and follow structured reasoning steps.  In practice, emerging approaches such as agentic harness engineering – systems I have worked on to constrain and monitor model behavior in production – act as control layers around models. However, these protections mainly shape how the model behaves, not what it has access to or the context it operates in.

Application-level controls are where system design becomes critical, especially in enterprise settings. The system should enforce role-based access control; it should block or filter sensitive data at the system level. You don’t want to rely on the model to “decide” not to reveal sensitive information; you want to make it impossible by design.

Organizations must treat AI usage as part of the security perimeter and develop policies that define appropriate use, validation and escalation. Scalable, safe AI adoption depends on combining system-level guardrails with a workforce trained to challenge, not just consume, AI outputs. They must learn to supervise, not just use, AI systems that can think, plan and act on their behalf.

Yizheng Wang is the head of AI at Straiker, an AI security startup backed by leading venture capital firms. He holds a Ph.D. from Stanford University, where his research focused on sequential decision-making under uncertainty, developing intelligent agents for safety-critical applications in climate and energy. At Straiker, he leads the development of AI safety systems, including red-teaming and risk detection frameworks for generative and agentic AI, with a focus on making these systems more robust, reliable, and aligned with human values.