Brian D. Colwell

Menu
  • Home
  • Blog
  • Contact
Menu

Category: Prompt Injection & Jailbreaking

A Brief Introduction To AI Prompt Injection Attacks

Posted on June 8, 2025 by Brian Colwell

The Open Worldwide Application Security Project (OWASP), a nonprofit organization focused on education “about the potential security risks when deploying and managing Large Language Models (LLMs) and Generative AI applications”, initiated its…

Defining The Token-level AI Jailbreaking Techniques

Posted on June 8, 2025June 8, 2025 by Brian Colwell

Token-level Jailbreaking optimizes the raw sequence of tokens fed into the LLM to elicit responses that violate the model’s intended behavior. Unlike prompt-level attacks that rely on semantic manipulation, token-level methods treat…

Defining The Prompt-Level AI Jailbreaking Techniques

Posted on June 8, 2025June 8, 2025 by Brian Colwell

Prompt-level attacks are considered social-engineering-based, semantically meaningful prompts which elicit objectionable content from LLMs, distinguishing them from token-level attacks that use mathematical optimization of raw token sequences. Now, let’s consider specific prompt-level…

A Brief Introduction To AI Jailbreaking Attacks

Posted on June 8, 2025June 8, 2025 by Brian Colwell

System prompts for LLMs don’t just specify what the model should do – they also include safeguards that establish boundaries for what the model should not do. “Jailbreaking,” a conventional concept in software systems…

The Big List Of AI Jailbreaking References And Resources

Posted on June 8, 2025June 8, 2025 by Brian Colwell

Note that the below are in alphabetical order by title. Please let me know if there are any sources you would like to see added to this list. Enjoy! Thanks for reading!

The Big List Of AI Prompt Injection References And Resources

Posted on June 8, 2025June 8, 2025 by Brian Colwell

Note that the below are in alphabetical order by title. Please let me know if there are any sources you would like to see added to this list. Enjoy! Thanks for reading!

A History Of AI Jailbreaking Attacks

Posted on June 7, 2025June 7, 2025 by Brian Colwell

The last couple years have seen an explosion in research into jailbreaking attack methods and jailbreaking has emerged as the primary attack vector for bypassing Large Language Model (LLM) safeguards. To date,…

Browse Topics

  • Artificial Intelligence
    • Adversarial Examples
    • Alignment & Ethics
    • Backdoor & Trojan Attacks
    • Data Poisoning
    • Federated Learning
    • Model Extraction
    • Model Inversion
    • Prompt Injection & Jailbreaking
    • Sensitive Information Disclosure
    • Supply Chain
    • Training Data Extraction
    • Watermarking
  • Biotech & Agtech
  • Commodities
    • Agricultural
    • Energies & Energy Metals
    • Gases
    • Gold
    • Industrial Metals
    • Minerals & Metalloids
  • Economics & Game Theory
  • Management
  • Marketing
  • Philosophy
  • Robotics
  • Sociology
    • Group Dynamics
    • Political Science
    • Religious Sociology
    • Sociological Theory
  • Web3 Studies
    • Bitcoin & Cryptocurrencies
    • Blockchain & Cryptography
    • DAOs & Decentralized Organizations
    • NFTs & Digital Identity

Recent Posts

  • The Bitcoin Whitepaper – Satoshi Nakamoto

    The Bitcoin Whitepaper – Satoshi Nakamoto

    June 13, 2025
  • The Big List Of AI Supply Chain Attack Resources

    The Big List Of AI Supply Chain Attack Resources

    June 11, 2025
  • AI Supply Chain Attacks Are A Pervasive Threat

    AI Supply Chain Attacks Are A Pervasive Threat

    June 11, 2025
©2025 Brian D. Colwell | Theme by SuperbThemes