About the Role:We are seeking an experienced Senior Data Engineer to join our dynamic data team. The ideal candidate will be responsible for designing, building, and maintaining scalable data pipelines and infrastructure, ensuring the availability, reliability, and quality of our data. This role requires strong technical expertise, problem-solving skills, and the ability to collaborate across teams to deliver data-driven solutions.Key Responsibilities:
Design, implement, and maintain robust, scalable, and high-performance data pipelines and ETL processes.
Develop and optimize data models, schemas, and storage solutions to support analytics and machine learning initiatives.
Collaborate with software engineers and product managers to understand data requirements and deliver high-quality solutions.
Ensure data quality, integrity, and governance across multiple sources and systems.
Monitor and troubleshoot data workflows, resolving performance and reliability issues.
Evaluate and implement new data technologies and frameworks to improve the data platform.
Document processes, best practices, and data architecture.
Mentor junior data engineers and contribute to team knowledge sharing.
Requirements:
Bachelors or Masters degree in Computer Science, Engineering, or a related field.
5+ years of experience in data engineering, ETL development, or a similar role.
Strong proficiency in SQL and experience with relational and NoSQL databases.
Experience with data pipeline frameworks and tools such as: Apache Spark, Airflow & Kafka. – MUST
Familiarity with cloud platforms (AWS, GCP, or Azure) and their data services.
Solid programming skills in Python, Java, or Scala.
Strong problem-solving, analytical, and communication skills.
Knowledge of data governance, security, and compliance standards.
Experience with data warehousing, big data technologies, and data modeling best practices such as ClickHouse, SingleStore, StarRocks.
Bachelors or Masters degree in Computer Science, Engineering, or a related field.
5+ years of experience in data engineering, ETL development, or a similar role.
Strong proficiency in SQL and experience with relational and NoSQL databases.
Experience with data pipeline frameworks and tools such as: Apache Spark, Airflow & Kafka. – MUST
Familiarity with cloud platforms (AWS, GCP, or Azure) and their data services.
Solid programming skills in Python, Java, or Scala.
Strong problem-solving, analytical, and communication skills.
Knowledge of data governance, security, and compliance standards.
Experience with data warehousing, big data technologies, and data modeling best practices such as ClickHouse, SingleStore, StarRocks.
This position is open to all candidates.










