You will be taking part in exciting projects that cover the entire data lifecycle – from integrating raw data with primary and third-party systems, through advanced data modelling, to state-of-the-art data visualisation and the development of innovative data products.
You will have the chance to learn how to construct and operate both batch and real-time data processing pipelines. You will work in a contemporary cloud-based data warehousing environment alongside a team of diverse, passionate, and fascinating colleagues. You will coordinate with other departments – such as product & tech, the core business units, trust & safety, finance, and others – to enable them to succeed.
Your responsibilities
- Design, implement and support data warehousing;
- Integrating raw data with primary and third-party systems
- Data warehouse modelling for operational & application data layers
- Developing in Amazon Redshift cluster
- SQL development as part of an agile team workflow
- ETL design and implementation in Matillion ETL
- Real-time data pipelines and applications using serverless and managed AWS services such as Lambda, Kinesis, API Gateway, etc.
- Designing and implementing data products that enable data-driven features or business solutions
- Building data dashboards and advanced visualisations in Sisense for data cloud teams (formerly Periscope Data) with a focus on UX, simplicity, and usability
- Working with other departments on data products – i.e. product & technology, marketing & growth, finance, core business, advertising, and others
- Being part of and contributing towards a strong team culture with the ambition to be on the cutting edge of big data
- Evaluating and improving data quality by implementing test cases, alerts, and data quality safeguards
- Living the team values: Simpler. Better. Faster.
- Strong desire to learn
Required minimum experience (necessary)
- 3 - 5 years of experience in data processing, analysis, and problem-solving with large quantities of data;
- Strong SQL skills across various relational data warehousing technologies especially in cloud data warehousing (e.g. Amazon Redshift, Google BigQuery, Snowflake, Vertica, etc.)
- 3+ years of experience with one or more programming languages, particularly Python
- Ability to communicate insights and findings to a non-technical audience
- Written and verbal proficiency in English
- Entrepreneurial spirit and the ability to think creatively; highly-driven and self-motivated; strong curiosity and a desire for continuous learning
- Graduating at the top of your class from a University technical degree such as computer science, engineering, maths, physics.
Additional experience (strong advantage)
- Experience working with customer-centric data on a big data scale, preferably in an online/e-commerce context
- Experience with modern big data ETL tools (like Matillion)
- Experience with the AWS data ecosystem (or other cloud providers)
- Track record in business intelligence solutions, building and scaling data warehouses, and data modelling
- Tagging, Tracking, and reporting with Google Analytics 360
- Knowledge of modern real-time data pipelines (like serverless framework, lambda, kinesis, etc.)
- Experience with modern data visualisation platforms such as Periscope, Looker, Tableau, Google Data Studio, etc.
- Linux, bash scripting, Javascript, HTML, XML
- Docker Containers and Kubernetes
#LI-TM1