Tallo logoTallo logo

Agentic AI Data Lead Software Engineer

Job

JP Morgan Chase Company

Columbus, OH (In Person)

Full-Time

Posted 4 days ago (Updated 12 hours ago) • Actively hiring

Expires 6/13/2026

Apply for this opportunity

This job application is on an outside website. Be sure to review the job posting there to verify it's the same.

Review key factors to help you decide if the role fits your goals.
Pay Growth
?
out of 5
Not enough data
Not enough info to score pay or growth
Job Security
?
out of 5
Not enough data
Calculating job security score...
Total Score
99
out of 100
Average of individual scores

Were these scores useful?

Skill Insights

Compare your current skills to what this opportunity needs—we'll show you what you already have and what could strengthen your application.

Job Description

Join our team as a Lead Data Products Software Engineer responsible for architecting, building, and scaling the Data Products Framework — a next-generation platform that enables users to discover, design, build, and productionize governed data products at enterprise scale. You will lead a team of engineers, driving the technical strategy and execution of a platform that orchestrates the end-to-end data product lifecycle leveraging AI/Agentic AI, policy-based governance, and cloud-native architectures on AWS. As a Lead Software Engineer at JPMorganChase within the Consumer & Community Banking Marketing Process Automation Team, you are an integral part of an agile team that works to enhance, build, and deliver trusted market-leading technology products in a secure, stable, and scalable way. As a core technical contributor, you are responsible for conducting critical technology solutions across multiple technical areas within various business functions in support of the firm's business objectives. Job responsibilities Lead, mentor, and grow a high-performing team of 5 - 7 engineers across multiple workstreams, fostering a culture of innovation, ownership, and technical excellence. Set the technical vision and engineering roadmap for the Data Products platform, aligning with firmwide priorities. Drive cross-functional collaboration with platform teams, domain Data Product Owners, AI/ML teams and governance teams. Architect and own the end-to-end technical design of the Data Products Studio — a scalable, enterprise-grade platform that orchestrates the discovery, design, build, and productionization of data products from the CCB Data Lake and Snowflake. Design the platform's AI/Agentic AI layer, leveraging intent agents, NLP Text-to-SQL, Knowledge Graphs (KAG), RAG, Vector Databases, and Agent-to-Agent (A2A) communication to enable intelligent, automated data product creation and natural language interaction with the data estate. Define the platform's integration architecture with various firmwide systems as appropriate. Establish and enforce architectural standards, design patterns, and engineering best practices across the team — ensuring scalability, security, resilience, and maintainability. Lead the design and development of Agentic AI capabilities that power the Data Products Framework — including autonomous discovery agents that profile and recommend data product candidates, design agents that auto-generate data contracts and schema recommendations, build agents that generate and optimize data pipelines, governance agents that auto-apply entitlements based on data classification, and quality agents that detect anomalies, drift, and trigger self-healing remediation. Architect the Agent-to-Agent communication layer enabling multi-agent orchestration across the data product lifecycle — from discovery through productionization. Leverage RAG (Retrieval Augmented Generation) and Vector Databases to enable contextual, knowledge-grounded AI interactions with metadata, lineage, and data catalog information. Implement NLP Text-to-SQL capabilities allowing business users to explore the CCB Data Lake and Snowflake using natural language, lowering the barrier to data product discovery. Required qualifications, capabilities, and skills Formal training or certification on software engineering concepts and 5+ years applied experience 10+ years of progressive experience in software engineering, data engineering, or platform engineering Strong leadership experience in guiding and mentoring varying levels of Software Engineers Proven track record of architecting and delivering large-scale, enterprise-grade data platforms or frameworks from concept through production in a large corporate environment Deep hands-on expertise in Python, SQL, and at least one additional language (preferably Java 17+, Spring, Boot), with strong system design and distributed systems knowledge Extensive experience designing, building, and optimizing ETL/ELT pipelines at scale, including batch and real-time data processing. Strong proficiency in PySpark for distributed data processing, including DataFrame and Dataset APIs and Spark SQL. Experience working with UI frameworks (React, Angular) will be an added advantage. Extensive experience with AWS cloud services including S3, Athena, Glue, Lambda, Step Functions, IAM, KMS, and Terraform. Basic knowledge of Snowflake (architecture, performance optimization, Tasks, Streams, Stored Procedures, Materialized Views, security model) is preferred, but not mandatory. Deep understanding of data governance principles including metadata management, data lineage, access control (RBAC/ABAC), data classification, and policy enforcement. Preferred qualifications, capabilities, and skills Experience with Grafana or equivalent observability platforms for custom dashboards, APM, SLA monitoring and alerting is a plus Experience working with UI frameworks (React, Angular) will be an added advantage

Similar remote jobs

Similar jobs in Columbus, OH

Similar jobs in Ohio