
AI Governance & Risk Management - Warsaw - Vice President- Software Engineering
at Goldman Sachs
Posted 18 hours ago
No clicks
- Compensation
- Not specified
- City
- Warsaw
- Country
- Poland
Currency: Not specified
As a Senior AI Guardrail Evaluation Expert & Project Manager, you will safeguard the integrity and trustworthiness of AI systems by evaluating and refining guardrails to prevent harmful, biased, or unfaithful content. You will lead planning, execution, and documentation of complex guardrail evaluation projects, ensuring timely delivery and quality results. You will onboard, train, and motivate human annotators, providing clear guidance to maintain high-performance feedback loops. You will collaborate cross-functionally with AI researchers, engineers, product managers, and legal teams to integrate ethical considerations throughout the AI lifecycle in the financial industry.
As a Senior AI Guardrail Evaluation Expert & Project Manager, you will play a pivotal role in safeguarding the integrity and trustworthiness of our AI systems. You will lead the charge in evaluating and refining the guardrails that protect our models from generating harmful, biased, or unfaithful content. This is a unique chance to combine your deep technical expertise with your leadership skills, working within a dynamic, hybrid environment to define best practices in AI ethics and safety for the financial industry.
You’ll do
- Lead the comprehensive evaluation and continuous refinement of AI guardrails to ensure robust detection and mitigation of harmful content and behaviours.
- Develop and improve sophisticated strategies for evaluating guardrails against:
- Profane, Toxic, and Adult Content
- Bias detection and mitigation in AI outputs
- Illegal and Harmful Content generation
- Misuse of Personal Data (PII) and other sensitive information
- Hallucination and the generation of unfaithful or factually incorrect content
- Drive the planning, execution, and meticulous documentation of complex AI guardrail evaluation projects, ensuring adherence to timelines, resources, and quality deliverables.
- Identify, onboard, train, and motivate a team of skilled human annotators, providing clear guidance and fostering a high-performance environment.
- Leverage strong linguistic understanding to guide annotation efforts, refine guardrail performance, and ensure high-quality human feedback.
- Collaborate cross-functionally with AI researchers, engineers, product managers, and legal teams to integrate ethical considerations throughout the AI development lifecycle.
You'll bring
- A proven track record in AI/ML ethics, content moderation, linguistic analysis, or a related field, with significant project leadership experience.
- Deep practical expertise in evaluating and improving AI guardrails on an evaluation platform like Braintrust, specifically addressing the categories mentioned above.
- Proficiency with version control systems (e.g., GitHub), collaboration platforms (e.g., Confluence), and data science environments (e.g., JupyterHub).
- Scripting and automation skills, particularly with Python, for data analysis and workflow optimization.
- Demonstrated ability to plan, execute, and meticulously document complex projects, managing timelines, resources, and deliverables effectively.
- Exceptional leadership skills, with experience in building, motivating, and managing high-performing teams, especially human annotators.
- Superior linguistic understanding and analytical capabilities to guide complex annotation tasks and interpret nuanced AI behaviours.
- Excellent communication, interpersonal, and problem-solving skills, with the ability to articulate complex technical concepts.









