Red Team Brainstorming With GPTs Accelerates Threat Modeling

AI Summary

The article discusses how GPT hallucinations can be used as untested ideas in threat modeling. Erica Burgess views these hallucinations as potential threats that need further testing.

Texas Relevance

No explicit mention of Texas, TX, TCUD, or any Texas-specific entities.

Original Content

Large language models have a well-earned reputation for making things up. But for AI cybersecurity architect Erica Burgess, rather than being a bug, GPT hallucinations can be a threat-modeling feature. "I like to think of the hallucinations as just ideas that haven't been tested yet," she said.