About the job Databricks Data Warehouse Architect
Data Architect Databricks & AWS
About the Role
One of the largest companies in the global gaming industry is seeking a hands-on Data Architect with Data Warehouse Engineering expertise in Databricks (DBX) and AWS-native data services.
Youll lead the design and implementation of a new data warehouse instance for a major product linebuilding from the ground up, architecting scalable pipelines, optimizing lakehouse performance, and integrating with real-time and batch data sources across AWS.
The ideal candidate is passionate about data architecture, thrives in fast-paced environments, and has a proven record of setting up high-performance lakehouse platforms on Databricks, with a strong foundation in data warehousing principles.
⸻
Key Responsibilities
- Design and deploy a new Databricks Lakehouse instance tailored to product-level data needs.
- Architect and implement robust data ingestion pipelines using Spark (PySpark/Scala) and Delta Lake.
- Integrate AWS-native services (S3, Glue, Athena, Redshift, Lambda) with Databricks for optimized performance and scalability.
- Define data models, optimize query performance, and establish warehouse governance best practices.
- Collaborate cross-functionally with product teams, data scientists, and DevOps to streamline workflows.
- Maintain CI/CD pipelines for data workflows using GitOps and Infrastructure-as-Code practices (e.g., DBX).
- Monitor data jobs and resolve performance bottlenecks or failures across environments.
⸻
Required Skills & Experience
- Databricks / Lakehouse Architecture
- End-to-end setup of Databricks workspaces and Unity Catalog.
- Expertise in Delta Lake internals, file compaction, and schema enforcement.
- Advanced PySpark/SQL skills for ETL and transformations.
- AWS Native Integration
- Deep experience with AWS Glue, S3, Redshift Spectrum, Lambda, and Athena.
- Knowledge of IAM and VPC configuration for secure cloud integrations.
- Data Warehousing & Modeling
- Strong grasp of modern dimensional modeling (star/snowflake schemas).
- Experience with lakehouse design patterns for mixed workloads.
- Automation & DevOps
- Familiarity with CI/CD for data engineering using DBX, Terraform, GitHub Actions, or Azure DevOps.
- Proficient in monitoring tools like CloudWatch, Datadog, or New Relic for data pipelines.
⸻
Bonus / Nice to Have
- Experience supporting gaming or real-time analytics workloads.
- Familiarity with Airflow, Kafka, or EventBridge.
- Exposure to data privacy and compliance frameworks (GDPR, CCPA).
⸻
Other Details
- Location: Latin America (LATAM) or USA Remote
- Contract Length: 1+ year
- Client: Global gaming company
⸻
Unique Skill Keywords
- Databricks Workspace Orchestration
- Data Warehouse Design
- Data Modeling
- Delta Lake Optimization
- Unity Catalog Implementation
- PySpark
- AWS Glue Job Orchestration
- Airflow
- S3-to-DBX Streaming Pipelines
- Lakehouse Schema Design
- Cross-region Data Mesh (AWS + DBX)
- GitOps-Driven ETL
- Product-Centric Data Layering
- Multitenant DBX Setup
- Redshift + Databricks Dual-query Optimization