Tallo logoTallo logo

REMOTE Staff Data Engineer

Job

Insight Global

Remote

Full-Time

Posted 6 weeks ago (Updated 6 weeks ago) • Actively hiring

Expires 5/27/2026

Apply for this opportunity

This job application is on an outside website. Be sure to review the job posting there to verify it's the same.

Review key factors to help you decide if the role fits your goals.
Pay Growth
?
out of 5
Not enough data
Not enough info to score pay or growth
Job Security
?
out of 5
Not enough data
Calculating job security score...
Total Score
81
out of 100
Average of individual scores

Were these scores useful?

Skill Insights

Compare your current skills to what this opportunity needs—we'll show you what you already have and what could strengthen your application.

Job Description

Job Description Insight Global is seeking a Staff Data Engineer to join our actuarial consultant customer 100% remotely. In this position as a Staff Data Engineer of our client's Data Platform, you will be responsible for designing and implementing robust Data Platform solutions that meet business objectives while ensuring compliance with industry-leading data privacy standards. You will collaborate closely with cross-functional agile teams to drive data architecture decisions, implement best practices, and contribute to the success of our client's projects.
What you will be doing:
  •  Acts as a subject matter expert and thought leader within the Data Platform Domain
Data Strategy:
Serves as a thought leader in data processing design and implementation, defining advanced structure for moving, storing, and maintaining high-quality data.
Team Leadership:
Leads projects by managing timelines, coordinating teams, and communicating project statuses. Influences organizational direction through effective leadership and strategic collaboration
Data Governance and Security:
Serves as a subject matter expert on governance standards, continuously aligning data practices with evolving industry best practices and requirements
Project Management and Scope of Work:
Contributes to defining the overall vision and strategy for data engineering within the organization, ensuring alignment with organizational goals and long-term objectives
Results Orientation:
Establishes visionary goals, advises on strategic plans, employs advanced monitoring, influences high-level stakeholders, and delivers transformative results
Data Platform:
Expansion of our Data Warehouse(s) and Lakehouse solutions for a healthcare data focused enterprise
Data Governance:
Configuring and maintaining unity catalog to enable enterprise data lineage, data quality, auditability and data stewardship
Data Security:
Building out Data Security protocols and best practices including the management of identified and de-identified (PHI/PII) solutions
Access Management:
Always ensure a policy of least privilege is followed for anything being implemented
External Data Products:
Building data solutions for clients while upholding the best standards for reliability, quality, and performance
ETL:
Building solutions within Delta Live Tables and automation of transformations
Medallion Architecture:
Building out performant enterprise-level medallion architecture(s)
Streaming and Batch Processing:
Building fit-for-purpose near real-time streaming and batch solutions
Large Data Management:
Building out performant and efficient enterprise solutions for internal and external users for both structured and unstructured healthcare data
Platform Engineering:
Building out Infrastructure as Code using Terraform and Asset Bundles
Costs:
Working with the business to build cost effective and cost transparent Data solutions
Pipeline/ETL Management:
You will help architect, build, and maintain robust and scalable data pipelines, monitoring, and optimizing performance
  •  Experience working with Migration tools i.e. Fivetran, AWS technologies and custom solutions
  •  Identify and implement improvements to enhance data processing efficiency
  •  Design and implement reliable and resilient Event Driven data processing
  •  Experience with building out effective pipeline monitoring solutions
  •  Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL, Delta Live Tables, Python, Scala, and cloud based 'big data' technologies
API Development:
Drive our design and implementation of internal APIs for integrating data between different systems and applications
  •  Integration with external systems utilizing API driven processes to ingest data
  •  Develop APIs built on top of datasets for internal systems to consume data from Databricks
  •  Experience integrating with external APIs including but not limited to Salesforce, Financial systems, HR systems and other external systems
Data Modeling:
Lead design, implementation, and maintenance of standards based (FHIR, OMOP, etc.) and efficient data models for both structured and unstructured data
  •  Assemble large, complex data sets that meet functional and non-functional business requirements
  •  Develop and maintain data models, ensuring they align with business objectives and data privacy regulations
Collaboration:
Partner internally and externally with key stakeholders to ensure we are providing meaningful, functional, and valuable data
  •  Effectively work with Data, Development, Analysts, Data Science, and Business team members to gather requirements, propose, and build solutions.
  •  Communicate complex technical concepts to non-technical stakeholders and provide guidance on best practices.
  •  Ensure that technology execution aligns with business strategy and provides efficient, secure solutions and systems
  •  Gather requirements and build out project plans to implement those requirements with forecasted efforts to implement
Processes and Tools:
Identify, design, and implement internal process improvements: o automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  •  Build analytics tools that utilize the data pipeline to provide actionable insights into operational efficiency and other key business performance metrics.
  •  Create data tools for clinical, analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
  •  Lead investigation of new tooling, develop implementation plans, and deployment of necessary tooling We are a company committed to creating diverse and inclusive environments where people can bring their full, authentic selves to work every day.
We are an equal opportunity/affirmative action employer that believes everyone matters. Qualified candidates will receive consideration for employment regardless of their race, color, ethnicity, religion, sex (including pregnancy), sexual orientation, gender identity and expression, marital status, national origin, ancestry, genetic factors, age, disability, protected veteran status, military or uniformed service member status, or any other status or characteristic protected by applicable laws, regulations, and ordinances. If you need assistance and/or a reasonable accommodation due to a disability during the application or recruiting process, please send a request to HR@insightglobal.com.

To learn more about how we collect, keep, and process your private information, please review
Insight Global's Workforce Privacy Policy:
https://insightglobal.com/workforce-privacy-policy/. Skills and Requirements
  •  15+ years of relevant experience in design, development, and testing of Data Platform solutions, such as Data Warehouses, Data Lakes, and Data Products
  •  Expert level experience working in Databricks and AWS
  •  Expert level experience working in both relational and non-relational databases such as SQL Server, PostgreSQL, DynamoDB, DocumentDB
  •  Experience building and managing solutions on AWS
  •  Expert in building out data models, data warehouses, designing of data lakes for enterprise and product use cases
  •  Familiarity with designing and building APIs, ETL and data ingestion processes and utilization of tools to support enterprise solutions
  •  Experience in performance tuning, query optimization, security, monitoring, and release management
  •  Experience working with and managing large, disparate, identified and de-identified data sets from multiple data sources
  •  Experience with building and deploying IAC using terraform, asset bundles and GitHub
  •  Experience collaborating with Data Science teams and building AI based solutions to drive efficiencies and business value
  •  Bachelor's degree or master's degree in computer science, data engineering or related field
  •  Experience managing and standardizing clinical data from structured and unstructured sources
  •  Health and Life Insurance business experience
  •  Knowledge in healthcare standards including FHIR, C-CDA, and traditional HL7
  •  Knowledge in clinical standards/ontologies including ICD10/SNOMED/NDC/LOINC/Rx Norm
  •  Associate or Professional level solution architecture certification in Azure and/or AWS
  •  Experience in Snowflake
  •  Experience in Spark
  •  Experience with Salesforce Integration

Similar remote jobs

Similar jobs in Brookfield, WI

Similar jobs in Wisconsin