Tallo logoTallo logo

Data Engineer with EKS

Job

Techno Talent Inc.

Tysons, VA (In Person)

Full-Time

Posted 4 days ago (Updated 1 day ago) • Actively hiring

Expires 6/12/2026

Apply for this opportunity

This job application is on an outside website. Be sure to review the job posting there to verify it's the same.

Review key factors to help you decide if the role fits your goals.
Pay Growth
?
out of 5
Not enough data
Not enough info to score pay or growth
Job Security
?
out of 5
Not enough data
Calculating job security score...
Total Score
79
out of 100
Average of individual scores

Were these scores useful?

Skill Insights

Compare your current skills to what this opportunity needs—we'll show you what you already have and what could strengthen your application.

Job Description

Big Data Engineer with
EKS:
Positions x2
Location:
Tysons Corner, VA Data Engineer Job Description Summary We are seeking a highly skilled and experienced Big Data Engineer to design, develop, and optimize large-scale data processing systems. In this role, you will work closely with cross-functional teams to architect data pipelines, implement data integration solutions, and ensure the performance, scalability, and reliability of big data platforms. The ideal candidate will have deep expertise in distributed systems, cloud platforms, and modern big data technologies such as Hadoop, Spark, and Kubernetes-based orchestration.
Responsibilities:
Design, develop, and maintain large-scale data processing pipelines using Big Data technologies (e.g., Hadoop, Spark, Python, Scala). Architect and deploy containerized big data workloads on Amazon EMR on EKS (Elastic Kubernetes Service). Design and implement Kubernetes-based infrastructure for running Spark applications at scale. Implement data ingestion, storage, transformation, and analysis solutions that are scalable, efficient, and reliable. Stay current with industry trends and emerging Big Data technologies to continuously improve the data architecture. Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions. Optimize and enhance existing data pipelines for performance, scalability, and reliability. Develop automated testing frameworks and implement continuous testing for data quality assurance. Conduct unit, integration, and system testing to ensure the robustness and accuracy of data pipelines. Work with data scientists and analysts to support data-driven decision-making across the organization. Ability to write and maintain automated unit, integration, and end-to-end tests. Monitor and troubleshoot data pipelines in production environments to identify and resolve issues. Manage Kubernetes clusters, pods, services, and deployments for big data workloads.
Essential Technical Skills:
AI Tool Proficiency:
Hands-on experience with AI development tools (GitHub Copilot, Q Developer, ChatGPT, Claude, etc.)
Big Data Technologies:
Experience with Big data technologies such as Hadoop, Spark, Hive & Trino Understanding of common issues like data skew and strategies to mitigate it, working with massive data volumes in PetaBytes, and troubleshooting job failures due to resource limitations, bad data, and scalability challenges. Real-world experience with debugging and mitigation strategies.
Container Orchestration & Kubernetes:
Strong experience with Kubernetes architecture, concepts, and operations (pods, services, deployments, namespaces, ConfigMaps, Secrets) Hands-on experience with Amazon EMR on EKS (Kubernetes) for running Apache Spark workloads Experience with Kubernetes resource management, scheduling, and auto-scaling Knowledge of Helm charts for deploying and managing applications on Kubernetes Understanding of Kubernetes networking, storage (PVs, PVCs), and security best practices Experience with kubectl and Kubernetes YAML manifests Ability to troubleshoot Kubernetes cluster issues, pod failures, and resource constraints Experience integrating Spark with Kubernetes operators and dynamic allocation
Cloud Technologies:
Experience with AWS services like
S3, EMR, EMR
on EKS, Glue, Lambda, Athena, etc. Hands-on experience using S3 with Spark (e.g., dealing with file formats, consistency issues) Strong experience with Amazon EKS (Elastic Kubernetes Service) architecture and best practices Experience with AWS IAM roles for service accounts (IRSA) for Kubernetes workloads Knowledge of AWS networking for EKS (VPC, subnets, security groups) Experience with AWS monitoring and logging tools (CloudWatch, CloudTrail) for Kubernetes workloads Serverless knowledge (Lambda, Fargate) Programming -
Python or Scala:
Ability to write clean, modular, and performant code Experience with functional programming concepts (e.g., immutability, higher-order functions) Real-world use cases where scalable data processing code was implemented Strong understanding of collections, concurrency, and memory management SQL Skills (Window Functions, Joins, Complex Queries): Proficiency with SQL window functions, multi-table joins, and aggregations Ability to write and optimize complex SQL queries Experience handling edge cases like NULLs, duplicates, and ordering

Similar remote jobs

Similar jobs in Tysons, VA

Similar jobs in Virginia