Position
Within the Data Lake team, you provide the data to all teams of the company, from business teams (CEO included) to Data Science team.
What you’ll do
- Build or update ETL (AWS Data Pipeline) to bring data into the Data Warehouse (AWS Redshift)
- Optimise the ETL processing as Redshift is a distibuted DWH containing around 100 To
- Maintain the Data Lake (dozens of big external tables) accessible by AWS Athena and Spectrum
- Learn quickly new technologies, whether already used in the team or not
- Maintain a good level of service for everyone
- Share good practices
- Participate to the Agile/Scrum methodology (stand-up, weekly, …)
- Work in a highly-growing technical startup environment
Who you are
- >3 years of experience
- >2 years of experience with SQL (with query optimization)
- >2 years of experience with ETL, DWH
- AWS knowledges (Redshift, Glue, other tools) are a plus
- Scripting knowledges (Python, Batch) are a plus
- Big Data knowledges (Spark, Scala, Hadoop) are a plus
- Github, Jira, Confluence, Asana knowledges are a plus
- Startup experience is a plus