Primary Responsibilities:
- Design, build, and maintain robust data pipelines.
- Enhance and optimize data processing workflows to support scalability.
- Identify and resolve performance issues in data queries and pipelines.
- Work closely with other teams to support and improve data storage and transformation processes.
- Contribute to the strategic planning and technology decisions for future developments.
Required Qualifications:
- Over 5 years of relevant experience in data development.
- Proficient in Python, with a strong ability to develop and manage object-oriented applications.
- Experience in batch data processing and handling large-scale automated data operations.
- Skilled in ETL pipelines and managing Data Lakes.
- Extensive knowledge of AWS, Docker, and version control systems, including GitLab or GitHub.
- Competence in Linux, Bash, and an understanding of Linux server operations and architecture.
- Ability to utilize profilers and debuggers for effective troubleshooting.
- Experience in analyzing and enhancing the performance of various data queries, jobs, and pipelines.