abra R&D is looking for a Data engineer! We are looking for a Data Engineer to join the team and contribute to AI-related projects. The role involves handling large volumes of incoming data, performing deep analysis, and collaborating closely with Data Scientists. You will be responsible for designing and developing critical, diverse, and large-scale data pipelines in both cloud and on-premise environments.
Requirements:
Minimum 5 years of experience as a Data Engineer mandatory 5 years of experience working with Object-Oriented Programming (OOP) languages mandatory 5 years of hands-on experience with Python mandatory Hands-on experience with Spark for large-scale data processing mandatory At least 2 years of practical experience with AWS , including services such as Athena, Glue, Step Functions, EMR, Redshift, and RDS strong advantage
* Deep understanding of design, development, and optimization of complex solutions handling or processing large-scale data
* Familiarity with optimization techniques and working with data partitioning and formats such as Parquet, Avro, HDF5, Delta Lake
* Experience working with Docker, Linux, CI/CD tools, and Kubernetes
* Experience with data pipeline orchestration tools like Airflow or Kubeflow Bachelor’s degree in Computer Science, Engineering, Mathematics, or Statistics – mandatory
* Understanding of machine learning concepts and workflows
* Familiarity with GenAI solutions or prompt engineering advantage
Minimum 5 years of experience as a Data Engineer mandatory 5 years of experience working with Object-Oriented Programming (OOP) languages mandatory 5 years of hands-on experience with Python mandatory Hands-on experience with Spark for large-scale data processing mandatory At least 2 years of practical experience with AWS , including services such as Athena, Glue, Step Functions, EMR, Redshift, and RDS strong advantage
* Deep understanding of design, development, and optimization of complex solutions handling or processing large-scale data
* Familiarity with optimization techniques and working with data partitioning and formats such as Parquet, Avro, HDF5, Delta Lake
* Experience working with Docker, Linux, CI/CD tools, and Kubernetes
* Experience with data pipeline orchestration tools like Airflow or Kubeflow Bachelor’s degree in Computer Science, Engineering, Mathematics, or Statistics – mandatory
* Understanding of machine learning concepts and workflows
* Familiarity with GenAI solutions or prompt engineering advantage
This position is open to all candidates.