Analytics Engineer

  • Full Time
Job expired!

Job Purpose

The Intermediate Analytics Engineer is a rising data professional who will begin their careers as part of the analytics engineering team to create strong, interconnected, and effective data products that provide first-rate, use-case led analytics across the organization. The role involves collaborating with senior or leading analytics engineers to construct data products and pipelines for high-impact projects that offer scalability and automation, and improve data availability and quality. The Intermediate Analytics Engineer has a zeal for honing their competencies in data-, cloud-, and software engineering to deliver analytic use-cases that generate business value and drive data as a competitive edge.

Role Description

  • Formulate and sustain data pipelines using SQL and Python to produce reliable, scalable, and fit-for-purpose data products in a cloud environment.
  • Convert technical requirements into reliable, scalable data products or pipelines that cater to the organization's requirements.
  • Cooperate with data scientists and analysts to comprehend data needs and furnish data products for analytics use-cases.
  • Execute monitoring, testing, and automation practices for data products.
  • Engage in code reviews, ensuring observance of coding standards and best practices.
  • Collaborate with team members to identify and resolve data-related problems and propose solutions.
  • Provide primary support for data pipelines.
  • Contribute to the formulation and maintenance of documentation for datasets and analytical processes, ensuring that consistent terminology and definitions are used to facilitate seamless teamwork.
  • Contribute to the creation of a library of reusable software engineering artifacts aimed at accelerating the creation of data products.
  • Preserve technical documentation related to data products and pipelines.
  • Streamline and implement components within a data product.
  • Support DataOps initiatives within the team.
  • Jointly work with the team to carry out standard testing procedures and conduct regular monitoring of datasets, with a focus on data accuracy and quality as a significant contributor to our analytics schemes.
  • Participate in the incorporation and adoption of software engineering best practices within the data team, contributing to the enforcement of coding standards, version control, and cooperative workflows.

Experience and Skills

  • A Degree or Diploma in Computer Science, Software Development, Engineering, or a related field.
  • A minimum of 1 year of hands-on experience in a data team, functioning as a data engineer or data-centric software engineer.
  • A knowledge of modern data processing tools and technologies, contributing to the development, refinement, and productionization of data products.
  • Expertise in Python and SQL, capable of undertaking development tasks.
  • Familiarity with the basic concepts of using Apache Spark for distributed computing, acquired through coursework or initial projects.
  • Familiarity with cloud platforms such as AWS, Azure, or Google Cloud through coursework or basic projects, with a basic understanding of cloud concepts related to computation and storage.
  • Basics of Infrastructure as Code and its importance in data engineering processes.
  • Possessing an exposure to handling voluminous data sets and understanding business models.
  • Basic understanding of software development best practices and version control systems (e.g., Git).
  • A commitment to coding standards, encompassing code readability, effective commenting, and consistent naming conventions.
  • Understanding of core testing methodologies to ensure code quality.
  • Willingness to learn and adopt established coding patterns and best practices within the team.
  • Experience of collaboration within and across multifunctional teams.
  • Experience of working within an Agile environment.