Software Engineer - Snowflake
Salary undisclosed
Checking job availability...
Original
Simplified
We are looking for a talented and motivated Software Engineer with experience in Confluent, Snowflake and QLIK to join our team. This role involves designing, developing, and maintaining real-time data streaming and cloud-based data warehousing solutions using Confluent Kafka, Snowflake. The ideal candidate will work closely with cross-functional teams to build scalable, efficient, and secure data pipelines that drive business intelligence and analytics.
Responsibilities:
• Data Pipeline Development: Design, develop, and optimize real-time and batch data pipelines using Confluent Kafka and Snowflake.
• Streaming Data Processing: Implement event-driven architectures leveraging Kafka Streams, ksqlDB, and Kafka Connect.
• Data Integration: Integrate structured and semi-structured data from multiple sources into Snowflake for analytics and reporting.
• Database Management: Develop, optimize, and maintain Snowflake-based data warehouses, ensuring high availability and performance.
• Security & Compliance: Implement data security best practices, including encryption, role-based access control (RBAC), and compliance with GDPR, HIPAA, or other relevant regulations.
• Optimization: Optimize query performance and storage efficiency in Snowflake using clustering, partitioning, and caching techniques.
• Testing: Conduct unit testing and data validation to ensure accuracy and reliability of data pipelines.
• Troubleshooting: Diagnose and resolve issues related to Kafka messaging, Snowflake storage, and data transformations.
• Documentation: Prepare and maintain technical documentation, including architecture diagrams, data flow processes, and best practices.
• Collaboration: Work with Data Scientists, BI Teams, and Cloud Engineers to provide real-time and batch data solutions.
• Maintenance: Monitor and maintain the stability of streaming pipelines and data warehouse operations.
Qualifications:
Must have
• Strong understanding of real-time data streaming and event-driven architectures.
• Hands-on experience with Confluent Kafka (Kafka Streams, Kafka Connect, ksqlDB).
• Experience in Snowflake data warehouse development and administration.
• Proficiency in SQL, Python, or Scala for data processing and transformation.
• Familiarity with cloud platforms (AWS, Azure, GCP) for data pipeline orchestration.
• Experience with ETL/ELT processes and data pipeline automation.
• Knowledge of CI/CD for data pipelines, version control (Git), and DevOps tools.
• Strong problem-solving and analytical skills.
Nice to have
• Confluent Kafka Certification or Snowflake Certification.
• Experience with Apache Airflow, dbt, or other data orchestration tools.
• Knowledge of Big Data Technologies (Spark, Flink, or Druid).
• Understanding of GraphQL, REST APIs, and WebSockets for data integration.
• Experience working in finance, healthcare, or other regulated industries.
Responsibilities:
• Data Pipeline Development: Design, develop, and optimize real-time and batch data pipelines using Confluent Kafka and Snowflake.
• Streaming Data Processing: Implement event-driven architectures leveraging Kafka Streams, ksqlDB, and Kafka Connect.
• Data Integration: Integrate structured and semi-structured data from multiple sources into Snowflake for analytics and reporting.
• Database Management: Develop, optimize, and maintain Snowflake-based data warehouses, ensuring high availability and performance.
• Security & Compliance: Implement data security best practices, including encryption, role-based access control (RBAC), and compliance with GDPR, HIPAA, or other relevant regulations.
• Optimization: Optimize query performance and storage efficiency in Snowflake using clustering, partitioning, and caching techniques.
• Testing: Conduct unit testing and data validation to ensure accuracy and reliability of data pipelines.
• Troubleshooting: Diagnose and resolve issues related to Kafka messaging, Snowflake storage, and data transformations.
• Documentation: Prepare and maintain technical documentation, including architecture diagrams, data flow processes, and best practices.
• Collaboration: Work with Data Scientists, BI Teams, and Cloud Engineers to provide real-time and batch data solutions.
• Maintenance: Monitor and maintain the stability of streaming pipelines and data warehouse operations.
Qualifications:
Must have
• Strong understanding of real-time data streaming and event-driven architectures.
• Hands-on experience with Confluent Kafka (Kafka Streams, Kafka Connect, ksqlDB).
• Experience in Snowflake data warehouse development and administration.
• Proficiency in SQL, Python, or Scala for data processing and transformation.
• Familiarity with cloud platforms (AWS, Azure, GCP) for data pipeline orchestration.
• Experience with ETL/ELT processes and data pipeline automation.
• Knowledge of CI/CD for data pipelines, version control (Git), and DevOps tools.
• Strong problem-solving and analytical skills.
Nice to have
• Confluent Kafka Certification or Snowflake Certification.
• Experience with Apache Airflow, dbt, or other data orchestration tools.
• Knowledge of Big Data Technologies (Spark, Flink, or Druid).
• Understanding of GraphQL, REST APIs, and WebSockets for data integration.
• Experience working in finance, healthcare, or other regulated industries.