Tallo logoTallo logo

Data Engineer

Job

Qode

San Diego, CA (In Person)

Full-Time

Posted 5 days ago (Updated 2 days ago) • Actively hiring

Expires 6/7/2026

Apply for this opportunity

This job application is on an outside website. Be sure to review the job posting there to verify it's the same.

Review key factors to help you decide if the role fits your goals.
Pay Growth
?
out of 5
Not enough data
Not enough info to score pay or growth
Job Security
?
out of 5
Not enough data
Calculating job security score...
Total Score
78
out of 100
Average of individual scores

Were these scores useful?

Skill Insights

Compare your current skills to what this opportunity needs—we'll show you what you already have and what could strengthen your application.

Job Description

Data Engineer Qode San Diego, CA Job Details Full-time 22 hours ago Qualifications Data model design Data lake Data modeling projects SQL Distributed computing Real-time data processing implementation Database software proficiency Full Job Description This role focuses on enabling front-office, advisor, and trading operations through low-latency data pipelines, scalable architectures, and governed data platforms. You will work closely with trading desks, portfolio management, and digital platforms to deliver reliable, compliant, and high-throughput data solutions. Key Responsibilities Trading Data Platform Engineering
  • Design and build real-time and batch data pipelines supporting trading workflows (orders, executions, positions, market data)
  • Develop low-latency data processing systems for near real-time decisioning
  • Build scalable data architectures for high-volume transaction data
  • Enable event-driven architectures using streaming platforms (Kafka, Kinesis) Wealth Management & Trading Integration
  • Integrate with trading platforms (OMS/EMS), portfolio systems, and advisor platforms
  • Support use cases such as:
  • Trade lifecycle tracking (order execution settlement)
  • Portfolio performance and analytics
  • Advisor dashboards and client reporting
  • Ensure data consistency across front-, middle-, and back-office systems Data Engineering & Architecture
  • Build and manage data lakes / lakehouse architectures (Delta Lake, Iceberg, etc.)
  • Develop ETL/ELT pipelines using modern frameworks
  • Design data models optimized for trading and analytics workloads
  • Implement API-driven data access layers for downstream consumption Performance, Scalability & Reliability
  • Optimize pipelines for low latency, high throughput, and fault tolerance
  • Implement data quality, reconciliation, and observability frameworks
  • Ensure high availability and disaster recovery for critical trading data systems Governance, Risk & Compliance
  • Implement data governance, lineage, and auditability
  • Ensure compliance with regulatory requirements (SEC, FINRA, etc.)
  • Enable data security, entitlements, and access controls
  • Support trade surveillance and reporting requirements Collaboration & Delivery
  • Partner with trading desks, product teams, and architects to translate requirements into scalable data solutions
  • Work closely with AI/analytics teams to enable downstream insights and models
  • Mentor junior engineers and contribute to data engineering best practices Required Qualifications
  • 7-12+ years of experience in data engineering or backend engineering
  • Strong expertise in:
  • Python / Scala / Java
  • SQL and distributed data processing (Spark, Flink, etc.)
  • Hands-on experience with:
  • Streaming platforms (Kafka, Kinesis, Pulsar)
  • Data lake / warehouse technologies (Snowflake, Databricks, Redshift)
  • Experience building real-time or near real-time data pipelines
  • Strong understanding of data modeling and large-scale distributed systems Preferred Qualifications
  • Experience in Wealth Management or Capital Markets trading systems
  • Familiarity with OMS/EMS platforms (e.g., Charles River Development, Aladdin, FIS)
  • Knowledge of market data (equities, fixed income, derivatives) and trade lifecycle / post-trade processing
  • Experience with cloud-native data platforms (AWS, Azure, GCP)
  • Exposure to real-time analytics and risk systems

Similar remote jobs

Similar jobs in San Diego, CA

Similar jobs in California