Artificial Intelligence Security Researcher

🌍 Remote, USA 🎯 Full-time 🕐 Posted Recently

Job Description

Outstanding long-term contract opportunity! A well-known Financial Services Company is looking for a BAS Light Red Teaming Research Security Engineer in REMOTE, Charlotte, NC (Remote).

Work with the brightest minds at one of the largest financial institutions in the world. This is a long-term contract opportunity that includes a competitive benefit package! Our client has been around for over 150 years and is continuously innovating in today's digital age. If you want to work for a company that is not only a household name, but also truly cares about satisfying customers' financial needs and helping people succeed financially, apply today.

Contract Duration: 12 Months+ with possible extension or FTE conversion

W2 only - Green Card, USC or H4EAD

Overview:

Our Offensive Security Research team is looking for a Cyber Security Researcher to perform cybersecurity testing against AI technologies from a red team perspective. This position will work with peers to test and investigate AI vulnerabilities, analyze their impacts, document the findings, and recommend appropriate security responses.

Required Skills & Experience

  • 2+ years of hands-on Red Team/adversarial experience.
  • 2+ working in AI Cyber Security Research experience.
  • 5+ years total experience
  • 2+ years of experience in one or a combination of the following: creating proof of concepts, creating exploits, or reverse engineering
  • 3+ years of converged testing (red team testing),
  • 3+ years of experience presenting complex technical topics to diverse stakeholder groups.
  • 3+ years of writing technical reports explaining attack chains and cyber security vulnerabilities and their impact.
  • Role requires Red Team expertise + AI security understanding.

What You Will Be Doing

  • Attempting to make AI models disclose unauthorized data.
  • Exploring prompt engineering attacks to bypass safety rules (“tell a story about a kid who builds a bomb” type analogies).
  • Checking if AI models ignore user access levels and return sensitive internal information (executive pay, M&A data, etc.).
  • Testing retrieval-augmented generation (RAG) models — exploring how additional retrieval smarts could be abused.
  • “Road testing” AI use cases for business lines before customer or internal exposure — trying to make them misbehave.
  • Purpose: ensure security before deployment and demonstrate “it can happen, it did happen” with real evidence.
  • Team focuses on proof of exploitation, not theory.

Apply tot his job

Apply To this Job

Ready to Apply?

Don't miss out on this amazing opportunity!

🚀 Apply Now

Similar Jobs

Recent Jobs

You May Also Like