NTT DATA Romania

Senior Data Engineer Job

23 December 2024
Apply Now
Deadline date:
£83000 - £155000 / year

Job Description

Who we are

 

We are seeking an experienced and motivated Data Engineer to join our team. The role focuses on the development, optimization, and maintenance of Python-based ETL pipelines within an Azure Synapse environment. You will also be responsible for ensuring the reliability, scalability, and performance of our data systems. The ideal candidate will have a strong technical background, excellent communication skills, and the ability to work collaboratively with diverse stakeholders.

 

What you’ll be doing

 

  • Develop and optimize Python-based ETL pipelines using Azure Synapse
  • Maintain and improve existing pipelines for reliability and performance
  • Plan and manage capacity and sizing of cloud service components
  • Oversee operations and maintenance of software, infrastructure, and databases
  • Manage interactions with PowerBI APIs
  • Monitor system performance and ensure SLA compliance
  • Create and maintain service-related documentation
  • Optimize processes and configurations across environments
  • Coordinate with development, architecture, rollout, operations teams, and customers for releases and upgrades
  • Support risk assessments and audits for service-related components

 

What you’ll bring along

 

  • Minimum 5-7 years of experience in data engineering
  • Minimum 2 years of experience with Python for data processing (Polars, Pandas, PySpark, etc.)
  • At least 1 year of hands-on experience with Microsoft Azure
  • Solid understanding of cloud architecture and ITIL principles
  • Familiarity with the software-as-a-service (SaaS) model
  • Strong analytical and organizational skills
  • Proven ability to rapidly adapt and respond to changing environments and priorities
  • Advanced proficiency in English (written and spoken)
  • Excellent communication skills with a high degree of customer focus.
  • Nice to have:
  • Experience with Azure Synapse, Delta Lake, and/or PowerBI
  • Exposure to Business Intelligence (BI) projects
  • Familiarity with DevOps practices and tools
  • Knowledge of DataVault 2.0 methodologies