BNY
Overview

Engineering AI Upskilling Program

This program is built for practitioners who are already building on Eliza, already using Windsurf, Claude, and Codex, and already operating at the frontier. BNY's top 500 engineers don't need foundational AI training. They need the architecture, evaluation, and governance skills that separate engineers who use AI from engineers who build and govern production AI systems at scale.

Why this program

Your engineers are already at the frontier. This is what comes next.

This is not a product management program repurposed for engineers. It's advanced AI engineering built from the ground up for practitioners already operating at the frontier.

BNY's top engineers are highly resourceful and self-motivated. Many are already consuming frontier content independently from sources like Hugging Face on their own time. They're not waiting to be taught. What they need is live instruction from practitioners solving the same problems, in a collaborative environment where engineers push each other and produce real artifacts against real BNY use cases.

The curriculum goes directly into advanced agent architecture, multi-agent system design, evaluation frameworks, and AI governance in regulated environments, structured around the specific capability gap between Builder and Pioneer. Every session produces something deployable.

The one day of Product for Engineers is designed specifically for engineers, giving them the context to build across the product boundary: validating use cases, designing for real user outcomes, and handing off cleanly from prototype to production.

Program Scope
Engineers 500 — top 10% across 35 platforms
Format Live, instructor-led · 20 cohorts
Cohort size 25–35 engineers · pods of 4–5
Entry level Builder — AI tools in production
Options 2-day · 4-day + Buildathon · 5-day + Buildathon
What We're Proposing

A live, instructor-led program that turns engineers into builders of production AI systems — not just users of them.

What it is

Two to four days of advanced AI engineering instruction, an optional Product for Engineers day, and a 3-day Buildathon where engineers build and evaluate real AI systems against BNY's own use cases. Every session produces a tangible artifact.

Who delivers it

Senior practitioners from OpenAI, Google, Meta, Intuit, and Amazon: people who have built and shipped production AI systems. Not academic curriculum. Not generic content. Instruction from people doing this work today.

What it produces

Engineers who can architect multi-agent systems, govern agent ecosystems at scale, and engineer organizational knowledge layers, operating at BNY's Level 3 Pioneer across the engineering organization.

Program Design

Four days of advanced AI engineering. One day of product thinking. Three days of building.

This is not a foundational AI training program. The curriculum moves directly into advanced agent design, system architecture, evaluation, and governance, structured around the specific capability gap BNY's top-tier engineers need to close. The Buildathon is where they prove it on real BNY use cases.

Program Specifications
Audience500 tier-1 software engineers + select product owners
FormatOption A: 2 days · Option B: 4 days + Buildathon · Option C: 5 days + Buildathon + On-Demand
DeliveryLive, instructor-led (hybrid options available)
Cohort size25–30 engineers per cohort
Cohorts20 cohorts across all options
Level of entryBuilder — engineers already using AI tools in production
Target outcomePioneer — architects of production AI systems
CertificationProduct School AI Engineering
Program Structure by Option
Option A · 2 Days Live
AI System Development Foundations

Build agent-based systems (RAG, agentic patterns, tool use) · Design multi-step workflows with orchestration · Integrate LLMs with APIs and internal data · Debug failure modes · Define production readiness · Apply systems to real BNY use cases.

Option B · 4 Days Live + 3-Day Buildathon
AI Systems + Production Readiness & Control

Everything in Option A, plus: build evaluation pipelines (LLM-as-judge, programmatic evals) · Define performance metrics (accuracy, latency, cost, hallucination) · Implement monitoring and feedback loops · Establish governance and deployment controls · 3-day Buildathon where teams build and evaluate AI systems on real internal BNY use cases with live demos.

Option C · 5 Days Live + Buildathon + On-Demand
End-to-End AI Systems + Product Integration at Scale

Everything in Option B, plus: align engineering with product and design using shared frameworks · Define validation briefs, PRDs, and sprint-ready plans · Connect AI systems to real user and business outcomes · Accelerate prototype-to-production handoff · On-demand core library covering product and execution fundamentals for engineers.

Program Governance

Single-point accountability across all 20 cohorts

Product School owns end-to-end delivery across all cohorts running in parallel, not just curriculum design. BNY has one point of contact throughout.

Program Accountability
Single PS Lead

Dedicated Product School lead owns end-to-end delivery. Defined escalation paths, weekly program health reporting, and single point of contact for BNY stakeholders throughout.

Cohort Orchestration
~20 Cohorts

Full cohort schedule owned and managed by PS with BNY. Instructor allocation across parallel cohorts, fallback protocols for last-minute changes, and completion tracking across all ~500 engineers.

Continuous Improvement
Pulse → Revise

Regular course pulse surveys, monthly program reviews with BNY stakeholders, curriculum revision cycles built into the roadmap, and outcome metrics tied to engineering KPIs.

Curriculum

11 modules. Every one produces a real artifact.

The program is structured around production outputs, not passive learning. Every module ends with an engineer holding something they built: a functional system, a governance spec, an eval suite. Content is developed for engineers already operating at the Builder level in BNY's environment.

Advanced AI Engineering: Frontier — Days 1–4

22 hours of live, instructor-led technical training delivered across four days. Builds from agent architecture through multi-agent governance, production deployment, and evaluation infrastructure.

M1 · Agent Architecture & Autonomous Systems

From Passive Models to Autonomous Orchestrators

The agentic shift: from AI-assisted coding to agents that plan, act, and verify. Covers the Agent Spectrum (Levels 0–3), autonomy decision frameworks, and the Agent Workflow Spec: actors, plan, memory, and tool boundaries. Engineers build their first functional agent.

Functional agent + Agent Workflow Spec
M2 · Scalable Agentic Design Patterns

Matching Architecture to Problem Type

Core architectural building blocks: state & memory, tooling fabric, multi-agent runtimes, governance hooks. Six agentic design patterns: sequential, parallel, critic-executor, delegation tree, negotiation, hierarchical. Engineers develop pattern-matching judgment and integrate a tooling fabric into their existing agent.

Pattern-matched agent design + tooling fabric integrated into existing agent
M3 · Multi-Agent Collaboration & Workflow Architecture

From Lone Agents to Governed Agent Ecosystems

Six multi-agent archetypes, PM blueprint for agent teams, collaboration spec design, and governing and scaling multi-agent workflows in a regulated financial environment. Key reliability metrics for agent-to-agent coordination at BNY scale.

Multi-Agent Collaboration Spec + high-level multi-agent workflow diagram
M4 · Failure Modes & Risk Diagnosis

How Agent Failures Differ from Traditional Bugs

Agent failures are silent, confident, and wrong. This module covers the failure modes unique to production agents — tool misuse, reasoning loops, memory drift, permission escalation, overconfidence — and the Agent Debugging Stack. Four eval types as quality gates.

Failure mode map + debugged agent with eval gates and detection metrics defined
M5 · Production Deployment & Launch Readiness

Agents as Business Value Drivers

Five strategic value drivers and trade-offs, four launch-readiness questions (value, ownership, trust, scale), risk-gated rollout strategies, and post-launch governance and feedback loops, including the PRD handoff spec for engineering-to-product handoff at BNY.

Production-ready PRD with logic spec, failure plan, and autonomy level + agent demo
M6 · Building Trusted AI Products on Modern Eval Metrics

Why Evaluation is the Most Critical Discipline

Three eval types: code-based, human, and LLM-as-Judge. Five trust-centric metrics: latency, hallucination, fairness, robustness, UX. The 95% accuracy trap. Engineers run their first LLM-as-Judge eval against their own agent.

Evaluation strategy canvas + first LLM-as-Judge eval run with trust metric trade-offs defined
M7 · Failure Mode Discovery & Error Analysis

A Failure Discovery Mindset vs. Reactive Bug Fixing

Four-step error analysis framework, Failure Taxonomy Canvas, and P0–P3 prioritization using a severity-frequency matrix. Engineers translate failure patterns into business risk language appropriate for BNY's regulatory and audit context.

Failure Taxonomy Canvas + prioritized risk map in business risk language
M8 · Systematic Eval Suites & Evaluator Design

From Deterministic to Probabilistic Evaluators

Eval Suite Pyramid for full-risk coverage, gold dataset construction, TPR/TNR analysis, and evaluator evaluation (hallucination in LLM-as-Judge). Eval Spec anatomy and stakeholder framing for engineering, UX, and exec audiences.

Simulated eval suite + Eval Spec with defined thresholds for engineering, UX, and exec audiences
M9 · Deploy Safely with Eval Gates

Earned Autonomy: The Tuned Trust Loop

Evaluators vs. eval gates. Advisory, soft, and hard gate types. Enforcement across hallucination, latency, bias, tone, and robustness. Codifying gates and thresholds in PRDs. PM as legislator, engineering as enforcer. Directly relevant to BNY's AI governance pipeline.

Eval gate mapping + PRD section with release criteria, thresholds, and mitigation plans
M10 · Scaling Evals Across Product Lines

Eval as Infrastructure

From siloed eval tasks to a centralized eval platform. The eval maturity curve, centralized vs. decentralized models, and continuous evaluation stack: code evals, drift monitors, periodic audits. Portfolio coverage matrix and cost vs. coverage vs. velocity trade-offs at BNY's scale.

Portfolio Coverage Matrix + eval budget allocation with justified trade-offs
M11 · Championing Responsible AI Culture

From Tactical QA to Org-Wide AI Governance

Three models of trust governance, ownership vs. execution, embedding responsible AI bottom-up and top-down. Eval Playbook and KPI dashboard. Pyramid Principle for high-stakes executive communication. Engineers deliver a ship/hold decision with eval-backed rationale to a stakeholder panel.

Deploy/hold decision memo + presentation delivered to stakeholders with eval-backed rationale
Product for Engineers — Day 5 (included in Option C)

As AI dissolves the boundary between product and engineering, engineers who can translate technical capability into product decisions have an outsized impact. This day covers UX fundamentals, validation, PRD design, and high-velocity prototyping, taught from an engineering perspective.

M1 · Understand Customers & Problems

User research, Design Thinking, Customer Journey Maps, and the Value Proposition Canvas as an engineering scope-framing tool.

Annotated Customer Journey Map + Value Proposition Canvas
M2 · Validating an Opportunity Hypothesis

Effort vs. User Value matrix, A/B testing mechanics, MVP types, and customer interview frameworks for hypothesis-testing without confirmation bias.

Validation brief with A/B test hypothesis + MVP type selection with rationale
M3 · Define Requirements & Work with PMs

PRD anatomy and lifecycle, roadmap types, product prioritization methods, and how to run effective engineering ↔ product relationships.

PRD with user scenarios, requirements, and sprint-ready roadmap
M4 · Execute Prototyping Velocity

High-velocity prototyping cycle, Skill Markdowns for injecting PRD context into build prompts, and The Confidence Line: criteria for knowing when a prototype is worth shipping.

Working prototype grounded in the validation brief + reusable Skill Markdown
Buildathon — 3 Days (included in Options B and C)

Module deliverables aren't just homework. They feed directly into the Buildathon. Teams build, evaluate, and defend real AI systems against BNY's own use cases.

Foundation Sprint
Agent Workflow Spec + architecture blueprint define team's project scope. Teams commit to use case and autonomy level by end of Day 1. → Project charter locked: use case, agent type, tooling stack, autonomy level declared.
Resilience & Deployment Sprint
Failure taxonomy and production PRD harden the prototype. Teams apply the debugging stack to their own agent's failure modes. Launch-readiness checklist gates the build before evaluation phase. → Hardened prototype with failure modes, mitigation plan, and PRD ready.
Eval Build Sprint
Evaluation strategy canvas, eval suite, and Eval Spec are built directly against each team's own agent. → Live eval suite running against the team's agent with P0 thresholds and Eval Spec complete.
Buildathon Showcase
Teams present their fully evaluated agent to a panel. Deploy/hold decision defended using real eval data, failure taxonomy, and trust metrics. → Final agent demo + deploy/hold memo.
Curriculum Walkthrough

Video overview of the full program

Shewit walks through the complete curriculum module by module, covering delivery approach, the level of technical depth, and how each session connects to the Pioneer capability target.

Instructors

Senior practitioners from the organizations defining AI today

Every instructor is currently working in AI, building production systems, shipping models, and solving the same types of problems BNY's engineers are being asked to solve. No academics. No retired executives. People doing this work now.

The instructors below are a representative sample. Product School draws from a global network of 200+ practitioner instructors, matched to each cohort's stack and domain. View the full instructor network →

Akshay Goel
Akshay Goel
AI Systems Builder
Meta · Ex-Standard Cognition

Leads AI product systems at Meta across Facebook Ads — engineering ML optimization pipelines that serve billions of impressions. Co-founded Explorer, an AI-powered autonomous vehicle mapping platform, and led product at Standard Cognition building computer vision systems for autonomous retail at scale. Previously Head of Product Innovation at Tata Group. Advisor at Carnegie Mellon University.

Ed Bayes
Ed Bayes
AI Experience Architect
OpenAI

Leads design strategy for generative AI interfaces at OpenAI — responsible for the interaction paradigms used by hundreds of millions of users. Pioneers the patterns for conversational agents, multimodal interfaces, and AI-native development environments.

Tulsee Doshi
Tulsee Doshi
Responsible AI Lead
Google · Ex-YouTube · Ex-Apple

Leads Google's ML Fairness and Responsible AI effort — developing the governance standards used across Google's AI products. BS in Symbolic Systems and MS in Computer Science from Stanford. Specializes in production AI governance at regulated scale.

Babak Zandi
Babak Zandi
AI Product Engineer
Meta · Ex-Klarna · Ex-omni:us

12+ years building AI products at scale, including Instagram and Messenger at Meta. Former Lead PM for AI at omni:us and Senior PM for Fraud & AI at Klarna — direct experience with AI governance in regulated financial services. Degree in Mathematics and Computer Science.

JE
Jamal Eason
AI Platform Lead
Google · Ex-Intel

Launched Android Studio — Google's official developer environment — and the Android Developer Preview program. Deep background in enterprise developer tooling and go-to-market for AI platforms. MS in IT from University of Maryland, MBA from Harvard Business School.

Alex Shih
Alex Shih
AI Engineering Lead
Slack · Ex-Google, Twitter, Airbnb

AI systems lead at Slack with a career spanning Google, Twitter, Airbnb, and Facebook — shipping complex AI-integrated systems at each. MIT master's in Technology & Policy and Engineering & Management; Cornell BS in Mechanical Engineering. Brings the systems-engineering discipline that turns AI prototypes into production infrastructure at FAANG scale.

Outcomes

Builder to Pioneer — the capability gap this program closes

BNY's engineers are not starting from the beginning. They are already at the Builder level, using AI tools daily, integrating APIs, and working within Eliza. This program is designed for the specific gap between where they are and what Level 3 Pioneer requires.

Level 2 · Builder
Where BNY engineers are now
  • Integrates AI APIs and tools into existing applications
  • Works within Eliza and established AI tooling (Windsurf, Claude, Codex)
  • Writes effective prompts for well-defined engineering tasks
  • Understands retrieval patterns and basic RAG
  • Builds single-agent features; can deploy simple AI-augmented components
  • Aware of AI evaluation and safety, but not yet systematic
The Gap
What this program addresses
  • Advanced agent architecture: building agents that act, not just respond
  • Agentic design pattern selection for complex real-world problems
  • Multi-agent system design and governance at scale
  • Organizational knowledge engineering: context that AI systems can reliably use
  • Production-grade evaluation and governance in a regulated environment
Level 3 · Pioneer
Where this program takes them
  • Architects and deploys multi-agent systems in production
  • Governs teams of agents operating at organizational scale
  • Engineers organizational context layers that AI systems can access reliably
  • Selects and applies the right agentic design pattern for each problem type
  • Runs systematic AI evaluation; owns governance in a regulated environment
BNY's committed target: 100% of engineers operating at Level 3 Pioneer: building agents, working in agent-human systems, and operating in AI-native development environments. The Global CIO has committed to this outcome across the full 10,000-engineer organization. This program is the mechanism for reaching it.
Investment

Three options. Same 500 engineers. Different depth.

All three options cover BNY's top-tier 500 engineers across 20 cohorts. The difference is program depth, from AI system fundamentals through full production readiness, product integration, and a 3-day Buildathon.

Option A
AI System Development Foundations
Two days of live AI engineering instruction. Engineers build, deploy, and debug production-ready AI systems integrated with real BNY data and workflows.
$750K USD
500 engineers · 20 cohorts · 2 days live
  • All 11 Advanced AI Engineering modules (M1–M11)
  • Agent architecture, multi-agent systems, and governance
  • Production deployment and launch readiness
  • Eval foundations and failure mode diagnosis
  • Artifact production every session
  • Product School AI Engineering certification
Option B
AI Systems + Production Readiness & Control
Everything in Option A, plus a 3-day Buildathon where teams build and evaluate AI systems against real BNY use cases, with live demos and a deploy/hold decision defended to a panel.
$1.25M USD
500 engineers · 20 cohorts · 4 days + 3-day Buildathon
  • Everything in Option A
  • 3-day Buildathon: build, evaluate, and demo on BNY use cases
  • Full eval suite running against team's own agent
  • Deploy/hold decision defended with eval-backed rationale
  • Defined metrics, risk thresholds, and deployment criteria
  • Governance and deployment controls established

All options cover 500 engineers across 20 cohorts.

About Product School
2M+
Product professionals in global network
12+
Years delivering enterprise programs
200+
Practitioner instructors worldwide
500+
Enterprise clients trained
Product School has transformed product organizations at
Next Steps

Ready to move forward.

This proposal reflects the full scope of what we've discussed. Three steps from internal alignment to cohort launch.

Step 01

Review with Jeremy & Team

The proposal and program overview have been shared. This is an opportunity to walk Jeremy and his team through the curriculum directly, answer any outstanding questions, and ensure full alignment before moving forward.

Schedule a Call →
Step 02

Confirm Scope

Based on our last conversation, BNY is leaning toward Option C, the full program including AI Engineering Foundations, Advanced AI Systems, the 3-day Buildathon, and Product for Engineers. Once the internal review is complete, we confirm scope and begin curriculum customization.

Step 03

Cohort Kick-Off

Product School owns cohort scheduling end-to-end across all 20 cohorts. Once scope is locked, we move directly into scheduling and mobilize quickly. Parallel cohort delivery begins on BNY's timeline.

Let's Schedule a Call
Doug Hanson · doug@productschool.com · (206) 713-4843