We are looking for a highly qualified Data Engineer to join our innovative team. The ideal candidate will be responsible for developing Python-based solutions to deploy and manage cloud data services (Azure Databricks and Azure Data Factory).Key ResponsibilitiesCode Development and AdaptationRefactor and adapt existing Python code to handle new data schemas and transformations.Extend current code by incorporating new data fields and new data sources.Develop new modules to process, transform, and load data using PySpark or Python-based frameworks.Implement data validation and error-handling mechanisms.Databricks Platform Support (Nice to Have)Support the initial configuration of Databricks, including library installation and resource optimization.Integrate Databricks with cloud storage and data sources (Azure and/or Databricks).Automate jobs in Databricks or Azure Data Factory.CI/CD, Testing, and Quality AssuranceSet up CI/CD pipelines for code and configuration deployment.Write and execute unit, integration, and performance tests for data pipelines.Debug and troubleshoot issues in distributed environments.Collaboration and DocumentationWork closely with data engineers, analysts, and cloud architects to ensure seamless integration and operation.Document code, workflows, and platform configurations for transparency and future reference.Minimum RequirementsSkills and Competencies:Hands-on mindset with a proactive approach to problem-solving.Strong interest in automation, test-first principles, and continuous delivery.Excellent communication and interpersonal skills, with the ability to mentor and support others.Creative thinking and adaptability in dynamic environments.Advanced English proficiency (reading, writing, and speaking).#Infosernt New Tecnológicas
Responsibilities
Job Requirements
Apply now