Job Opening
Data Engineering Specialist
Job Type: Full-time
Location: Geneva, Switzerland/Remote
Job Summary:
We are seeking a highly skilled Data Engineering Specialist with more than 5 years of experience to join our team. The ideal candidate will have expert knowledge in developing and implementing ETL jobs for data warehouses using SQL, and good knowledge and experience with AWS Big Data tools such as Glue, Athena, Redshift, and Kinesis. The successful candidate should have a deep understanding of data engineering tools and methods, as well as a knowledge of CI/CD concepts in general and AWS CodePipeline/CodeBuild/CodeDeploy in particular. The candidate must also have experience in the provisioning of data APIs and knowledge of information security concepts and terminology.
Responsibilities:
· Identify and document how data is structured and organized in existing data sources.
· Design and implement data pipelines in Python and SQL using Apache Spark and AWS Glue extensions.
· Implement data drift monitoring.
· Connect existing data visualization and reporting tools to final datasets.
· Assist data science specialists in the definition and extraction of relevant features for data analysis.
· Collaborate with business areas to improve the performance of central collation of big datasets.
· Assist the project team to define and implement suitable tools and processes to automate manual data extractions.
Skills and Qualifications:
· More than 5 years of experience in data engineering.
· Expert knowledge and experience developing and implementing ETL jobs for data warehouses using SQL.
· Good knowledge and experience with AWS Big Data tools: Glue, Athena, Redshift, Kinesis.
· Expert knowledge and experience of data engineering tools and methods.
· Good knowledge and experience with AWS CloudFormation.
· Knowledge of CI/CD concepts in general and AWS CodePipeline/CodeBuild/CodeDeploy in particular.
· Knowledge of provisioning of data APIs.
· Knowledge of information security concepts and terminology.
· Strong communication and teamwork skills with an advanced level of written and spoken English.
· Proficient in SQL, Python, PySpark, Athena, Glue, Spark, Kinesis, and Redshift.
Deliverables:
· Design and implementation of data pipelines.
· Final datasets connected to data visualization and reporting tools.
· Definition and extraction of relevant features for data analysis.
· Suitable tools and processes to automate manual data extractions.
If you meet the qualifications and are interested in this exciting opportunity, please submit your resume and cover letter with the Job Description title for consideration to woludhe@mdosisolutions.com .