What are the responsibilities and job description for the Lead AI Safety and Enablement Engineer position at Carex Consulting Group?
W2 only, no C2C, must be authorized to work in the US (both now and in the future).
Lead AI Safety and Enablement Engineer (Direct Hire) - Hybrid in Madison, Wisconsin. Apply now for this opportunity at Carex!
Carex is partnering with a BioTech industry partner to hire a Lead AI Safety and Enablement Engineer based in Madison, WI.
Summary
The Lead AI Safety and Enablement Engineer ensures the safe, reliable, and scalable use of AI and machine learning across the organization. This role focuses on developing and implementing systems, tools, and frameworks that embed responsible AI principles into the technology ecosystem. The position combines software engineering expertise with a strong understanding of AI risk management, compliance, and observability.
What You'll Do
Lead AI Safety and Enablement Engineer (Direct Hire) - Hybrid in Madison, Wisconsin. Apply now for this opportunity at Carex!
Carex is partnering with a BioTech industry partner to hire a Lead AI Safety and Enablement Engineer based in Madison, WI.
Summary
The Lead AI Safety and Enablement Engineer ensures the safe, reliable, and scalable use of AI and machine learning across the organization. This role focuses on developing and implementing systems, tools, and frameworks that embed responsible AI principles into the technology ecosystem. The position combines software engineering expertise with a strong understanding of AI risk management, compliance, and observability.
What You'll Do
- Support implementation of enterprise standards for AI safety, transparency, and reliability
- Develop and maintain shared AI safety tools such as model catalogs, metadata registries, and monitoring systems for bias, drift, and performance
- Build APIs, templates, and SDKs that integrate governance and validation into AI/ML development pipelines
- Contribute to the design of observability and telemetry solutions for continuous monitoring of model and data quality
- Collaborate with Legal, Privacy, Compliance, and Information Security teams to translate AI policies into automated controls and technical safeguards
- Work with ML and GenAI teams to embed validation and safety checkpoints in AI workflows
- Participate in cross-functional reviews and discussions supporting AI governance and responsible AI practices
- Promote awareness and adoption of responsible AI principles through tools, documentation, and knowledge sharing
- Support and comply with Quality Management System policies and procedures
- Uphold organizational values through accountability, innovation, integrity, quality, and teamwork
- Bachelor’s degree in Computer Science, Software Engineering, Artificial Intelligence, or a related field
- 8 years of experience in software engineering, ML systems, or platform engineering
- Hands-on experience with AI lifecycle tools such as MLflow, Arize, WhyLabs, or Label Studio
- Proficiency in Python, CI/CD processes, and cloud platforms (AWS, Azure, or GCP)
- Experience developing scalable and compliant ML systems and tools
- Strong understanding of AI risk management, governance, and observability concepts
- Ability to collaborate cross-functionally and translate policy into technical implementation
- Preferred:
- Master’s degree in Computer Science, Data Engineering, or Artificial Intelligence
- Experience implementing AI assurance, observability, or risk management frameworks
- Knowledge of GenAI, LLM evaluation, and prompt safety practices
- Familiarity with FDA, HIPAA, or GxP compliance standards