Tkxel
Principal Data Engineer
Job Description
Job Description:
We are seeking an experienced Data Engineer to the
design, development, and optimization of our client data infrastructure. This
role requires deep expertise in cloud technologies (primarily Azure, AWS is a plus) and data engineering best practices, with additional
experience in Apache Spark and Databricks for large-scale data
processing. The Data Engineer will work closely with data scientists, analysts,
and other stakeholders to create scalable and efficient data systems that
support advanced analytics and business intelligence. Additionally, this role involves
mentoring junior engineers and driving technical innovation within the data
engineering team.
Key Responsibilities:
- Collaborate
with Solution Architects: Work with Big Data Solution Architects to
design, prototype, implement, and optimize data ingestion pipelines,
ensuring effective data sharing across business systems. - ETL/ELT
Pipeline Development: Build and optimize ETL/ELT pipelines and
analytics solutions using a combination of cloud-based technologies, with
an emphasis on Apache Spark and Databricks for large-scale
data processing. - Data
Processing with Spark: Leverage Apache Spark for distributed
data processing, data transformation, and analytics at scale. Experience
with Databricks for optimized Spark execution is highly desirable. - Production-Ready
Solutions: Ensure data architecture, code, and processes meet
operational, security, and compliance standards, making solutions
production-ready in cloud environments. - Project
Support & Delivery: Actively participate in project and production
delivery meetings, providing technical expertise to resolve issues quickly
and ensure successful project execution. - Database
Management: Manage both SQL (e.g., PostgreSQL, MySQL) and NoSQL (e.g., DynamoDB, MongoDB) databases, ensuring data is efficiently stored,
retrieved, and queried. - Real-Time
Data Processing: Implement and maintain real-time data streaming
solutions using tools such as Apache Kafka, AWS Kinesis, or
other technologies for low-latency data processing. - Cloud
Monitoring & Automation: Use monitoring and automation tools
(e.g., AWS CloudWatch, Azure Monitor) to ensure efficient
use of cloud resources and optimize data pipelines. - Data
Governance & Security: Implement best practices for data
governance, security, and compliance, including data
encryption, access controls, and audit trails to meet regulatory
standards. - Collaboration
with Stakeholders: Work closely with data scientists, analysts, and
business teams to align data infrastructure with strategic business
objectives and goals. - Documentation:
Maintain clear and detailed documentation of data models, pipeline
processes, and system architectures to support collaboration and
troubleshooting.
Requirements
- 5+
years of experience as a Data Engineer, with strong expertise in
cloud-based data warehousing, ETL pipelines, and large-scale data
processing. - Proficiency
with cloud technologies, with experience in platforms like Azure or AWS. - Hands-on
experience with Apache Spark for distributed data processing and
transformation. Experience with Databricks is highly desirable. - Strong SQL skills and experience with relational databases (e.g., PostgreSQL, MySQL) as well as NoSQL databases (e.g., MongoDB, DynamoDB).
- Proficient
in Python for data processing, automation tasks, and building data
workflows. - Experience
with PySpark for large-scale data engineering, particularly in Spark
clusters or Databricks. - Experience
in designing and optimizing data warehouse architectures, ensuring
optimal query performance in large-scale environments. - A
strong understanding of data governance, security, and compliance best practices, including encryption, access control, and data privacy.
Preferred Qualifications:
- Bachelor’s
degree in Computer Science, Engineering, or a related field. - Certifications in Data Engineering from cloud providers (e.g., AWS Certified
Big Data – Specialty, Microsoft Certified: Azure Data Engineer
Associate) are a plus. - Experience
with advanced data engineering tools and platforms such as Databricks, Apache Spark, or similar distributed computing technologies