Tallo logoTallo logo

Data Engineer with Ab initio

Job

Applied Thought Auditors & Consultants Inc.

Berkeley Heights, NJ (In Person)

Full-Time

Posted 2 days ago (Updated 5 hours ago) • Actively hiring

Expires 6/14/2026

Apply for this opportunity

This job application is on an outside website. Be sure to review the job posting there to verify it's the same.

Review key factors to help you decide if the role fits your goals.
Pay Growth
?
out of 5
Not enough data
Not enough info to score pay or growth
Job Security
?
out of 5
Not enough data
Calculating job security score...
Total Score
82
out of 100
Average of individual scores

Were these scores useful?

Skill Insights

Compare your current skills to what this opportunity needs—we'll show you what you already have and what could strengthen your application.

Job Description

Title:
Data Engineer (Ab Initio, OzoneCH, Flink)
Location:
Berkeley Heights, NJ (5 Days Onsite)
FULLTIME
Role:
Streaming Data Engineer -
Ab Initio, Ozone CH, Flink Job Description:
We are seeking a highly skilled resource to design and implement high-performance, event-driven data pipelines, ensuring low-latency data processing and high availability system for the large credit card processing system. The ideal candidate will work with the Ab Initio ecosystem (GDE, EME, Conduct>It) to build stateful services that ingest, filter, and transform data from sources like Kafka or message queues, pushing updates to dashboards or downstream databases in near-real-time.
Key Responsibilities:
Create complex Ab Initio continuous flow graphs, including stateful joins, sliding time windows, and aggregations. Implement event-driven data pipelines using Kafka, MQ, and file streams. Ensure the resilience of continuous flows, including checkpointing and recovery, to guarantee "exactly-once" processing. Apply advanced Ab Initio components (e.g., Reformat, Rollup, Join, Partition) to ensure low-latency performance. Proactively monitor live production streams to ensure 24/7 reliability and troubleshooting data issues Develop ETL pipelines for batch and real-time data ingestion and transformation. Implement and ensure data validation, data security, integrity, and compliance across big data platforms. Monitor and troubleshoot performance issues in large-scale clusters. Collaborate with data scientists, analysts, and application teams to deliver high-quality data solutions. Automate workflows and improve operational efficiency using scripting and orchestration tools.
Required Skills & Qualifications:
Deep understanding of credit card process system Deep knowledge of GDE (Graphical Development Environment), EME (Enterprise Meta>Environment), Conduct>It, and Continuous Flows. Understanding of Kafka, message queues, and real-time stateful services. Proficiency in Unix/Linux shell scripting, SQL, and database technologies (e.g., Oracle, Teradata). Experience in Java , Scala, Python or Kafka is plus. Familiarity with Linux/Unix environments and shell scripting. Understanding of data security, governance, and compliance standards. Experience with cloud-based big data platforms Exposure to containerization (Docker, Kubernetes) for big data workloads. Knowledge of CI/CD pipelines for data engineering projects.
Behavioral Skills:
Good Communication skills 5 days Work from Office at Berkley Heights, NJ Team Player Ability to work in a changing environment Strong problem solving and analytical skills Ability to work independently or within a team Manage day-to-day challenges and communicate developmental risks with the technical team
Qualifications:
Bachelor s degree in computer science, Software Engineering, or a related field. Proficiency in business process modeling and documentation tools. Product implementation experience is preferred

Similar remote jobs

Similar jobs in Berkeley Heights, NJ

Similar jobs in New Jersey