We are tech transformation specialists, uniting human expertise with AI to create scalable tech solutions.
With over 8,000 CI&Ters around the world, we’ve built partnerships with more than 1,000 clients during our 30 years of history. Artificial Intelligence is our reality.
Responsibilities:
- Design and Develop Data Pipelines: Create robust, scalable data pipelines using Databricks, Apache Spark, and SQL to transform and process large datasets efficiently.
- Performance Optimization: Monitor and optimize the performance of existing data pipelines and workflows to ensure high throughput and low latency.
- Collaboration with Stakeholders: Work closely with data scientists, analysts, and business stakeholders to understand data requirements and translate them into technical specifications.
- Data Quality and Governance: Implement data quality checks and governance practices to ensure data consistency, accuracy, and compliance.
- Documentation and Best Practices: Maintain comprehensive documentation for data pipelines and processes, and contribute to the establishment of best practices within the team.
Requirements for this Challenge:
- Sólid experience in data engineering or a related field, with a focus on data pipelines
- Soft Skills: Excellent communication in English and Portuguese, with the ability to work effectively in a team-oriented environment and engage with clients.
- Technical Skills: Proficiency in Databricks, Apache Spark, SQL, and Python (or Scala).
- Cloud Technologies: Experience with cloud Azure
- Problem-Solving: Excellent analytical and problem-solving skills, with the ability to troubleshoot complex data issues.
- Education: Bachelor’s degree in Computer Science, Information Technology, or a related field.