Job Details
Job Description
Job Purpose
The ideal candidate will use their passion for big data and analytics to provide insights to the business covering a range of topics. They will be responsible for conducting both recurring and ad hoc analysis for business users. As a Data Engineer at Pepkor Lifestyle, you will play a critical role in the development and maintenance of our data infrastructure. You will work closely with cross-functional teams to ensure data availability, quality, and accessibility for analysis. The ideal candidate will use their passion for big data and analytics to provide insights to the business covering a range of topics. They will be responsible for conducting both recurring and ad hoc analysis for business users.
Position outputs/competencies
- Collaborate with data scientists, analysts, and business stakeholders to understand data requirements.
- Design, develop, and maintain data pipelines and ETL processes.
- Implement and maintain data warehousing and data storage solutions.
- Optimize data pipelines for performance, scalability, and reliability.
- Ensure data quality and integrity through data validation and cleansing processes.
- Monitor and troubleshoot data infrastructure issues.
- Stay current with emerging technologies and best practices in data engineering.
- Systematic solution design of the ETL and data pipeline inline with business user specifications
- Develop and implement ETL pipelines aligned to the approved solution design
- Ensure data governance and data quality assurance standards are upheld
- Deal with customers in a customer centric manner
- Effective Self-Management and Team work
Minimum qualification and Experience
- Bachelor's degree in Computer Science, Information Technology, or a related field.
- Proven experience as a Data Engineer in a professional setting.
- Proficiency in data engineering technologies and programming languages (e.g., SQL, Python, Scala, Java).
- Strong knowledge of data storage, database design, and data modelling concepts
- Experience with ETL tools, data integration, and data pipeline orchestration.
- Familiarity with data warehousing solutions (e.g., Snowflake, Redshift).
- Excellent problem-solving and troubleshooting skills.
- Strong communication and collaboration skills.
- 5-10 years’ Experience and understanding in designing and developing data warehouses according to the Kimball methodology.
- Adept at design and development of ETL processes.
- SQL development experience, preferably SAS data studio and AWS experience The ability to ingest/output CSV, JSON and other flat file types and any related data sources.
- Proficient in Python or R or willingness to learn. Experience within Retail, Financial Services and Logistics environments.
- Redshift Technologies
- Understanding of data security and compliance best practices.
- Relevant certifications (e.g., AWS Certified Data Analytics, Google Cloud Professional Data Engineer).