|Job type:||Full Time Job|
|Work hours:||8h per day / 40h a week|
|Preferred Office hours:||08:00 - 16:00|
|Possibility of remote work:|
1 day a week
|Project team size:||5|
Ideal profile: A passionate data engineer who loves software development and distributed data processing and is already experienced with data processing and analysis in Python and AWS, Familiar with principles and common design patterns regarding distributed systems based on microservices and/ or serverless architecture running on AWS, Familiar with AWS data processing stack as AWS Kinesis, Athena, S3, Lambdas, Firehose, Experienced in deploying production-grade software with fully automated CICD pipeline and configuring fully automated logging and monitoring of a distributed system.
Must have skills: Extensive experience in using Java or Python, Working knowledge of DevOps and IaC tooling, Openness to new technologies and languages, Understanding the problems of distributed systems design, Knowledge of distributed data storage systems/formats using, parallel processes and/or columnar data stores such as AWS, Redshift, Elastic Search, SOLID, Clean Code, Nice to have: Experience with online and offline data processing tools and technologies (Storm, Kafka, Spark, Flink), Knowledge of Cloud Computing on multiple Amazon Web Services or any other public cloud offering, Background in Data Collection + Ingestion, Data Visualization, ETL, Data Processing Pipeline and Data Modelling,