Senior AI Security Engineer

New Today

TASKS: • Design, implement, and execute test approaches to GenAI systems (MyCity Chatbot) to identify security flaws, particularly those impacting confidentiality, integrity, or availability of information. • Perform various types of tests such as functional testing, regression testing, performance testing, and usability testing to evaluate the behavior and performance of the AI algorithms and models. • Create, implement, and execute test plans and strategies for evaluating AI systems, including defining test objectives, selecting suitable testing methods, and identifying test scenarios. • Document test methods, results, and suggestions in clear and brief reports to stakeholders. • Perform security assessments including creating updating and maintaining threat models and security integration of Gen AI platforms. Ensure that security design and controls are consistent with OTI's security architecture principals. • Design security reference architectures and implement/configure security controls with an emphasis on GenAI technologies. • Provide AI security architecture and design guidance as well as conduct full-stack architecture reviews of software for GenAI systems and platforms. • Serve as a subject matter expert on information security for GenAI systems and applications in cloud/vendor and on-prem environments. • Discuss AI/ML concepts proficiently with data science and ML teams to identify and develop solutions for security issues. • Collaborate with engineering teams to perform advanced security analysis on complex GenAI systems, identifying gaps and contributing to design solutions and security requirements. • Identify and document defects, irregularities or inconsistencies in AI systems and working closely with developers to rectify and resolve them. • Ensure the quality, consistency and relevance of data used for training and testing AI models (includes collecting, preprocessing and validating data) • Assess AI systems for ethical considerations and potential biases to make sure they follow ethical standards and encourage inclusivity and diversity. • Collaborate with diverse teams including developers, data scientists, and domain experts to understand requirements validate assumptions and align testing efforts with project goals. • Conducting research to identify vulnerabilities and potential failures in AI systems. • Design and implement mitigations, detections, and protections to enhance the security and reliability of AI systems. • Perform model input and output security including prompt injection and security assurance.
MANDATORY SKILLS/EXPERIENCE Note: Candidates who do not have the mandatory skills will not be considered. • Bachelor's degree in computer science, electrical or computer engineering, statistics, econometrics, or related field, or equivalent work experience 12+ years of hands-on experience in cybersecurity or information security. • 4+ years of experience programming with demonstrated advanced skills with Python and the standard ML stack (TensorFlow/Torch, NumPy, Pandas, etc.) • 4+ years of experience with Natural Language Processing (NLP) and Large Language Models (LLM) desired • 4+ years of experience working in cloud environment (Azure, AWS, GCP) • Demonstrated proficiency with AI/ML fundamental concepts and technologies including ML, deep learning, NLP, and computer vision. • Demonstrated ability (expertise preferred) in attacking GenAI products and platforms. • Demonstrated recent experience with large language models. • Ability to stay updated on the latest developments, trends, and best practices in both software testing and artificial intelligence. • Demonstrated experience with using AI testing frameworks and tools such as TensorFlow or PyTorch, or Keras • Demonstrated ability to write test scripts, automate test cases, and analyze test results using programming languages and testing frameworks listed above. • Demonstrated ability to Identify and document defects, irregularities or inconsistencies in AI systems and working closely with developers to rectify and resolve them. • Ability to work independently to learn new technologies, methods, processes, frameworks/platforms, and systems. • Excellent written and verbal communication skills to articulate challenging technical concepts to both Client and expert audiences.
DESIRABLE SKILLS/EXPERIENCE: • Excellent problem-solving and critical thinking skills with attention to detail in an ever-changing environment. • Background in designing and implementing security mitigations and protections and/or publications in the space • Ability to work collaboratively in an interdisciplinary team environmen t • Participated or currently participating in CTF/GRT/AI Red Teaming events and/or bug bounties developing or contributing to OSS projects. • Understanding of ML lifecycle and MLOps. • Perform various types of tests such as functional testing, regression testing, performance testing, and usability testing to evaluate the behavior and performance of the AI algorithms and models • Ability to ensure the quality, consistency and relevance of data used for training and testing AI models (includes collecting, preprocessing and validating data) • Ability to assess AI systems for ethical considerations and potential biases to make sure they follow ethical standards and encourage inclusivity and diversity • Ability work in and provide technical leadership to cross-functional teams to develop and implement AI/ML solutions, including capabilities that leverage LLM technology • Highly flexible/willing to learn new technologies
Location:
Brooklyn

We found some similar jobs based on your search