Job Openings Software Engineer - Snowflake

About the job Software Engineer - Snowflake

We are looking for a talented and motivated Software Engineer with experience in Confluent, Snowflake and QLIK to join our team. This role involves designing, developing, and maintaining real-time data streaming and cloud-based data warehousing solutions using Confluent Kafka, Snowflake. The ideal candidate will work closely with cross-functional teams to build scalable, efficient, and secure data pipelines that drive business intelligence and analytics.

Responsibilities:

  • Data Pipeline Development: Design, develop, and optimize real-time and batch data pipelines using Confluent Kafka and Snowflake.

  • Streaming Data Processing: Implement event-driven architectures leveraging Kafka Streams, ksqlDB, and Kafka Connect.

  • Data Integration: Integrate structured and semi-structured data from multiple sources into Snowflake for analytics and reporting.

  • Database Management: Develop, optimize, and maintain Snowflake-based data warehouses, ensuring high availability and performance.

  • Security & Compliance: Implement data security best practices, including encryption, role-based access control (RBAC), and compliance with GDPR, HIPAA, or other relevant regulations.

  • Optimization: Optimize query performance and storage efficiency in Snowflake using clustering, partitioning, and caching techniques.

  • Testing: Conduct unit testing and data validation to ensure accuracy and reliability of data pipelines.

  • Troubleshooting: Diagnose and resolve issues related to Kafka messaging, Snowflake storage, and data transformations.

  • Documentation: Prepare and maintain technical documentation, including architecture diagrams, data flow processes, and best practices.

  • Collaboration: Work with Data Scientists, BI Teams, and Cloud Engineers to provide real-time and batch data solutions.

  • Maintenance: Monitor and maintain the stability of streaming pipelines and data warehouse operations.

Qualifications:

Must have

  • Strong understanding of real-time data streaming and event-driven architectures.

  • Hands-on experience with Confluent Kafka (Kafka Streams, Kafka Connect, ksqlDB).

  • Experience in Snowflake data warehouse development and administration.

  • Proficiency in SQL, Python, or Scala for data processing and transformation.

  • Familiarity with cloud platforms (AWS, Azure, GCP) for data pipeline orchestration.

  • Experience with ETL/ELT processes and data pipeline automation.

  • Knowledge of CI/CD for data pipelines, version control (Git), and DevOps tools.

  • Strong problem-solving and analytical skills.

Nice to have

  • Confluent Kafka Certification or Snowflake Certification.

  • Experience with Apache Airflow, dbt, or other data orchestration tools.

  • Knowledge of Big Data Technologies (Spark, Flink, or Druid).

  • Understanding of GraphQL, REST APIs, and WebSockets for data integration.

  • Experience working in finance, healthcare, or other regulated industries.