PBT Group is currently offering an opportunity for an AWS Data Engineer with 4+ years of relevant experience.
The role of a Data Engineer involves constructing and maintaining data pipelines and datamarts, emphasizing scalability, repeatability, and security. Data Engineers play a pivotal role in facilitating the acquisition of data from diverse sources, ensuring its conformity to data quality standards, and enabling downstream users to access data promptly. This position is an integral part of an agile team.
These professionals are entrusted with the responsibility of establishing the infrastructure required to derive insights from raw data, integrating data from various sources seamlessly. They empower solutions by efficiently managing substantial volumes of data, both in batch and real-time, utilizing cutting-edge technologies from the realms of big data and cloud computing. Additional responsibilities encompass the development of proof-of-concepts and the implementation of intricate big data solutions, with a primary focus on collecting, parsing, managing, analyzing, and visualizing extensive datasets. They are adept at employing technologies to resolve challenges associated with handling vast amounts of data in diverse formats, thereby delivering innovative solutions.
Data Engineering is a technically demanding role that necessitates a broad spectrum of expertise in software development and programming. These professionals possess knowledge in data analysis, understanding end-user and business requirements, and have the ability to translate these needs into technical solutions. They exhibit a strong grasp of physical database design and the systems development lifecycle. Collaboration within a team environment is essential for success in this role.
Key Responsibilities:
- Architecting Data analytics framework.
- Translating complex functional and technical requirements into detailed architecture, design, and high-performance software.
- Leading the development of data and batch/real-time analytical solutions by leveraging transformative technologies.
- Engaging in multiple projects as a technical lead, overseeing user story analysis, design, software development, testing, and automation tool creation.
Duties: Primary Job Objectives:
- Development and Operations
- Database Development and Operations
- Establishment and Adherence to Policies, Standards, and Procedures
- Communication
- Business Continuity and Disaster Recovery Planning
- Research and Evaluation
- Coaching and Mentoring
Required Skills, Knowledge, and Experience:
- A minimum of 5 years of experience in Data Engineering or Software Engineering.
- Demonstrated leadership experience, managing teams of engineers for 3-5 years.
- A minimum of 2 years of experience in Big Data.
- At least 5 years of experience with Extract, Transform, and Load (ETL) processes.
- A minimum of 2 years of experience with AWS (Amazon Web Services).
- Demonstrated experience with agile or other rapid application development methodologies for at least 2 years (e.g., Agile, Kanban, Scrum).
- 5 years of proven expertise in object-oriented design, coding, testing patterns, and working with commercial or open-source software platforms and large-scale data infrastructures.
- Proficiency in creating data feeds from on-premise to AWS Cloud (2 years).
- Support experience for data feeds in production on a break-fix basis (2 years).
- A minimum of 4 years of experience in creating data marts using Talend or similar ETL development tools.
- Proficiency in data manipulation using Python and PySpark (2 years).
- Experience in processing data using the Hadoop paradigm, particularly with EMR, AWS's distribution of Hadoop (2 years).
- DevOps experience in Big Data and Business Intelligence, including automated testing and deployment (2 years).
- Extensive knowledge of various programming or scripting languages.
- Expertise in data modeling and an understanding of different data structures and their suitability for specific use cases.
Additional Technical Skills Required:
- The ability to design highly scalable distributed systems using various open-source tools.
- Proficiency in both batch and streaming Big Data tools.
- Experience with Talend for at least 1 year.
- Familiarity with AWS services such as EMR, EC2, and S3 for at least 1 year.
- Proficiency in Python for at least 1 year.
- Familiarity with PySpark or Spark (desirable for at least 1 year).
- Experience in Business Intelligence data modeling for 3 years.
- Proficiency in SQL for 3 years.
Qualifications/Certifications:
- A Bachelor's degree in computer science, computer engineering, or equivalent work experience for a minimum of 4 years.
- AWS Certification, at least at the associate level.
* In order to comply with the POPI Act, for future career opportunities, we require your permission to maintain your personal details on our database. By completing and returning this form you give PBT your consent