Job Openings Data Engineer (Architect)

About the job Data Engineer (Architect)

Job Description

  • Architect, administer, and optimize Databricks workspaces, clusters, jobs, and workflows.
  • Design and develop scalable Azure-based data engineering solutions, including Data Lake, Data Factory, Synapse, Key Vault, etc.
  • Build high-performance data pipelines using Python and PySpark.
  • Work with Hadoop-based platforms and distributed storage systems.
  • Implement and manage real-time data streaming solutions and ingestion pipelines.
  • Ensure proper configuration, monitoring, governance, and performance tuning of data platforms.
  • Integrate CI/CD pipelines on Azure and work with containerization technologies (Docker/K8s).
  • Develop and optimize SQL queries, stored procedures, and data models using MS SQL.
  • Collaborate with architects, data scientists, and engineering teams to deliver enterprise-grade solutions.
  • Define best practices, coding standards, and data engineering frameworks across the organization.
  • Troubleshoot complex data issues and ensure high availability of mission-critical data systems.

Requirements

  • Minimum 10+ years of total IT experience.
  • Minimum 6+ years strong experience in the following areas: Databricks Administration , Azure Cloud , Python & PySpark , Hadoop ecosystem , Data streaming & data pipeline engineering , MS SQL
  • Proven experience architecting large-scale distributed data systems.
  • Strong understanding of data lakes, data warehousing, and big data architecture patterns.

Nice-to-Have Skills

  • Experience with Apache Spark optimization and tuning.
  • Hands-on experience with CI/CD pipelines on Azure (GitHub Actions, Azure DevOps, etc.).
  • Knowledge of containers (Docker, Kubernetes).
  • Understanding of MLOps or advanced analytics integrations.