About the job 888 - Data Engineer (DBT + Spark + Argo) · Senior · Remote · LATAM
Location: LATAM (Remote supporting US-based teams)
Job Type: Contractor (Full-Time, Remote)
Project: Data Platform Modernization Healthcare Sector
Time Zone: Aligned with GMT-3 (Argentina)
English Level: B2/C1
Get to Know Us
At Darwoft, we build software that drives real change. But were more than just tech were people first. With a remote-first culture and a highly collaborative team spread across LATAM, we partner with global companies to co-create reliable, scalable, and impactful digital products.
Were currently working with a leading US-based healthtech platform, in a major transformation of their data pipeline ecosystem migrating legacy SQL logic into modern, scalable cloud-based infrastructure using DBT, Spark, Argo, and AWS.
Were Looking For a Senior Data Engineer (DBT + Spark + Argo)
In this role, you will be at the core of a strategic data transformation initiative: converting monolithic SQL Server logic into a modular, testable DBT architecture, while integrating Spark for performance and Argo for orchestration. You will work with cutting-edge lakehouse formats like Apache Hudi, Parquet, and Iceberg, and enable real-time analytics through ElasticSearch integration.
If you're passionate about modern data engineering and want to work in a data-driven, cloud-native, healthcare-focused environment, this is the role for you.
What Youll Be Doing
- Translate legacy T-SQL logic into modular, scalable DBT models powered by Spark SQL
- Build reusable and performant data transformation pipelines
- Develop testing frameworks to ensure data accuracy and integrity in DBT workflows
- Design and orchestrate workflows using Argo Workflows and CI/CD pipelines with Argo CD
- Manage mock data and reference datasets (e.g., ICD-10, CPT), ensuring version control and governance
- Implement efficient storage/query strategies using Apache Hudi, Parquet, and Iceberg
- Integrate ElasticSearch for analytics by building APIs and pipelines to support indexing and querying
- Collaborate with DevOps teams to optimize S3 usage, enforce data security, and ensure compliance
- Work in Agile squads and participate in planning, estimation, and sprint reviews
What You Bring
- Strong experience with DBT for data modeling, testing, and deployment
- Hands-on proficiency in Spark SQL, including performance tuning
- Solid programming skills in Python for automation and data manipulation
- Familiarity with Jinja templating for building reusable DBT components
- Practical experience with data lake formats: Apache Hudi, Parquet, Iceberg
- Expertise in Argo Workflows and CI/CD integration with Argo CD
- Deep understanding of AWS S3 data storage, performance tuning, and cost optimization
- Strong command of ElasticSearch for indexing structured/unstructured data
- Knowledge of ICD-10, CPT, and other healthcare data standards
- Ability to work cross-functionally in Agile environments
Nice to Have
- Experience with Docker, Kubernetes, and container orchestration
- Familiarity with cloud-native data tools: AWS Glue, Databricks, EMR, or GCP equivalents
- Prior work on CI/CD automation for data engineering workflows
- Knowledge of data compliance standards: HIPAA, SOC2, etc.
- Contributions to open-source projects in DBT, Spark, or data engineering frameworks
Perks & Benefits
- Contractor agreement with payment in USD
- 100% remote work
- Argentinas public holidays
- English classes
- Referral program
- Access to learning platforms
Explore this and other opportunities at:
www.darwoft.com/careers