Non-members are welcome to access the full story here
Press enter or click to view image in full size
This is my write-up for the TryHackMe room on AI Models & Data. Written in 2026, I hope this write-up helps others learn and practice cybersecurity.
This section introduces the foundational concept that an AI model’s security risks begin with its training data. It highlights how invisible, poorly documented data supply chains can embed vulnerabilities like PII, credentials, and compromised safety mechanisms long before the model is actually deployed.
I understand the learning objectives and am ready to learn about AI models and data!
No answer needed
This section explores the origins of AI training data, emphasizing the heavy reliance on web scraping and the security risks associated with poor data provenance. It explains how sensitive information, such as PII and API keys, can become permanently baked into model weights, highlighting the need for an ML-BOM (Machine Learning Bill of Materials) to track and verify data sources.
What term describes the ability to answer where data came from, when it was collected, and whether it has been modified?