Job Openings Data Engineer (Snowflake & dbt Specialist)

About the job Data Engineer (Snowflake & dbt Specialist)

Job Title: Data Engineer (Snowflake & dbt Specialist)

Location: Remote Latin America Preferred

Type of Contract: Contractor, Full-Time

Salary Range: 4K to 6K USD/month

Language Requirements: English Advanced (Written & Spoken)

We are seeking an experienced Data Engineer with strong expertise in Snowflake and dbt (Data Build Tool) to join our dynamic data team. You will play a key role in building and optimizing a next-generation cloud data warehouse that powers analytics, reporting, and data science initiatives. Your work will directly enhance data quality, scalability, and accessibility across the organization.

Key Responsibilities

  • Design, develop, and maintain robust ELT/ETL data pipelines using modern methodologies within the Snowflake ecosystem.
  • Manage and optimize data transformation workflows using dbt, including modular SQL modeling, data testing, and documentation.
  • Implement efficient data ingestion solutions using Snowpipe, Kafka, or third-party tools to process structured and semi-structured data.
  • Optimize Snowflake performance through warehouse tuning, query optimization, clustering, and materialized views.
  • Design and implement dimensional data models (Star and Snowflake schemas) to support reporting and analytical needs.
  • Configure and manage Snowflake security features, including RBAC, data masking, and compliance with governance standards.
  • Collaborate with Data Analysts, Data Scientists, and stakeholders to deliver scalable, data-driven solutions.
  • Implement and maintain CI/CD pipelines and Infrastructure as Code (IaC) for dbt and Snowflake using Git, Airflow, or Azure DevOps/AWS Glue.
  • Monitor pipeline health, ensuring data integrity, reliability, and performance through proactive alerting and logging.

Must-Have Qualifications

  • 4 years of professional experience in Data Engineering, Business Intelligence, or related roles.
  • Proven, hands-on expertise with Snowflake data warehouse architecture, design, and optimization.
  • Advanced proficiency with dbt, including modeling, testing, macros, and package management.
  • Strong SQL skills for query design, optimization, and stored procedures (Snowflake Scripting or PL/SQL).
  • Proficiency in Python for automation, orchestration, and data transformation tasks.
  • Experience working with cloud platforms such as AWS, Azure, or GCP, and related services (e.g., S3, ADLS).

Preferred Qualifications

  • Experience with workflow orchestration tools such as Apache Airflow, Prefect, or dbt Cloud.
  • Familiarity with Data Vault 2.0 or advanced data modeling techniques.
  • Experience with real-time data streaming technologies like Kafka or Kinesis.
  • Bachelors degree in Computer Science, Engineering, or a related field.
  • Excellent communication and collaboration skills with a focus on data quality and operational excellence.