Execute adversarial testing of LLMs, agentic AI systems, recommendation models, and fraud detection tools using techniques such as prompt injection, jailbreaking, data poisoning, model inversion, and membership inference attacks
Develop synthetic attack datasets tailored to fundraising and trust & safety scenarios
Build automated adversarial testing pipelines integrated into CI/CD and reusable robustness evaluation libraries
Deploy real-time detection for prompt injection and model evasion; implement input validation, output filtering, adversarial training, and differential privacy mechanisms
Lead red team operations and strengthen AI/ML systems powering fraud detection, content moderation, and Trust & Safety at scale
Partner with Trust & Safety, Product Security, and Data Science to mitigate algorithmic bias and implement security controls
Establish AI security policies, training, deployment review processes aligned with NIST AI RMF
Build monitoring and incident response systems for AI security
Stay current with emerging attack vectors; contribute to open-source adversarial tools and publish externally to advance GoFundMe’s AI security leadership
Requirements
6–8 years in cybersecurity with a focus on AI/ML security or adversarial ML