Large language models (LLMs) have moved from the experimental edge into the operational core of modern enterprises. They are powering customer service chatbots, enabling knowledge search across internal systems, and accelerating software development. For many organizations, LLMs are no longer “nice to have”; they’re essential infrastructure.
But with this rapid adoption comes a growing challenge: how do we secure them?
The 2025 SANS Security Awareness Report makes clear that human risk remains the primary attack vector. Social engineering continues to top the list of organizational concerns, while “incorrect use of AI at work” has now emerged as a leading human risk category. These findings highlight a new reality: securing LLMs isn’t just about technical guardrails. It’s about equipping people with the knowledge, behaviors, and culture to use AI responsibly.
LLMs amplify both the capabilities of organizations and the capabilities of attackers. On one hand, they automate tasks, accelerate decision-making, and improve productivity. On the other, they can be weaponized to craft highly personalized phishing messages, generate malicious code, or impersonate voices and identities with alarming realism.
But the more immediate risk for most enterprises is internal misuse. Employees may:
These aren’t malicious acts. They’re human ones, and that’s what makes them so dangerous. Attackers know the weakest link isn’t the model itself, it’s the workforce using it.
Enterprises are investing heavily in technical controls to secure AI platforms: AI governance, access management, auditing, and red-teaming. These are necessary steps. But as the SANS report points out, technology alone cannot solve human risk.
Consider the parallels to phishing. Despite decades of investment in email security gateways, phishing remains the top human risk because attackers exploit human judgment, not just technical vulnerabilities. LLMs are no different. Even with the strongest guardrails in place, an uninformed user can undermine security in a single prompt.
That’s why training is no longer optional. It’s essential.
Securing LLMs is not a single skillset. It requires training on two distinct but interconnected fronts.
First, IT and security teams that implement LLMs must understand how to properly secure them against AI-specific threats. This includes hardening models against prompt injection, ensuring strong access controls, monitoring for misuse, and designing guardrails that prevent data leakage or malicious code generation. Technical teams need a clear framework for identifying where LLMs create new attack surfaces and how to mitigate them with sound engineering practices.
Second, security teams themselves must undergo a cultural shift. To be effective defenders, security professionals must recognize that the same AI-driven manipulation techniques targeting the broader workforce (deepfakes, voice cloning, AI-crafted phishing) can just as easily be aimed at them. Training isn’t just about configuring models correctly; it’s about preparing people inside security functions to question what they see and hear, validate information, and model the resilient behaviors they expect from the wider workforce.
Without both dimensions, technical mastery and cultural adaptation, LLM security programs will remain incomplete. Organizations that invest in this two-fold approach equip their teams not only to secure the technology but also to withstand the AI + human threats that will define the next decade of cyber risk.
The risks of insecure LLM use are not hypothetical. For security teams tasked with building and securing enterprise LLMs, these scenarios illustrate the stakes:
Each of these risks presents financial, legal, and reputational consequences. Regulators are also beginning to scrutinize AI governance, meaning organizations that lack clear training and policy frameworks may soon face compliance exposure.
The return on investment for LLM security training is simple: one prevented breach, one avoided compliance penalty, or one averted reputational crisis more than pays for the program.
Ultimately, securing LLMs is about securing trust: in the data, in the outputs, and in the workforce. Organizations that treat training as an afterthought will find themselves constantly playing catch-up against increasingly AI-powered attackers.
Those that embed LLM-specific training into their culture will not only reduce risk but also unlock AI’s true potential. When employees know how to use these tools safely, they become innovators, not liabilities.
GuidePoint Security’s AI Security for Large Language Models training course was built with this exact challenge in mind. It combines theoretical knowledge with practical applications. Participants will gain in-depth understanding of AI fundamentals, security implications and the unique challenges posed by LLMs.
The convergence of human and AI risk is the defining security challenge of the decade. Training your people to secure the LLMs your organization uses is one of the fastest, most effective ways to reduce risk today while preparing for an AI-driven future.
David Bressler
Principal Security Consultant - Application Security
David is a Principal Security Consultant at GuidePoint Security within the Application Security Team. David has broad-based, hands-on experience with application security assessments, source code review, architecture review, penetration testing, digital and physical social-engineering assessments dating back to 2006. Before joining GuidePoint Security, David worked within Boston Children’s Hospital’s internal security team, and was the technical lead for the application security, vulnerability management and incident response programs throughout the hospital.
David’s experience includes developing numerous open-source security tools and Paterva Maltego open-source intelligence integrations, including NWMaltego, CuckooforCanari, Bitcoin-Explorer and Nextego. He also has been a speaker at Bsides Boston, MassHackers and RSA’s Security Analytics Summit events. David holds the Offensive Security Certified Professional (OSCP) and Microsoft Certified Systems Administrator (MCSA) certifications, as well as several COMPTIA certifications, including the Security+, Network+, and A+.