Job Description:
Company Name: Zeta
Position: Data Engineer I
Qualifications: Bachelor’s/ Master’s Degree
Experience: 1-3 (Years)
Location: Bangalore, India
Key Responsibilities:
Database Design and Management: Design, implement, and maintain robust database systems while ensuring optimal performance, data integrity, and timely resolution of database issues.
ETL (Extract, Transform, Load) Processes: Develop, manage, and optimize ETL workflows to efficiently move and transform data between systems, ensuring the reliability of data pipelines.
Data Modeling: Design and update logical and physical data models to accurately represent and organize data structures.
Data Warehousing: Build and maintain scalable data warehouses to support the storage, retrieval, and analysis of large datasets.
Data Integration: Integrate data from diverse sources, including APIs, internal databases, and external datasets, to enable unified data access and reporting.
Data Quality and Governance: Implement data quality checks, enforce data governance standards, and contribute to the development of governance policies and best practices.
Scripting and Programming: Automate and streamline data processes using programming languages such as Python, Java, and SQL, including the development of validation scripts and error-handling mechanisms.
Version Control: Utilize version control systems like Git to manage and track changes in the data pipeline codebase effectively.
Monitoring and Optimization: Establish monitoring frameworks to assess system performance and health, and optimize data processes for scalability, speed, and efficiency.
Cloud Platforms: Deploy, manage, and optimize data infrastructure on cloud platforms such as AWS, Azure, or Google Cloud Platform (GCP), leveraging cloud-based tools for data storage, processing, and analytics.
Security: Implement and maintain data security measures in compliance with organizational and regulatory standards, ensuring the protection of sensitive information.
Troubleshooting and Support: Provide technical support for data-related issues, conduct root cause analyses, and implement preventive measures to ensure system reliability.
Eligibility Criteria:
Bachelor’s or Master’s degree in Computer Science, Information Systems, or a related engineering field, with 1-3 years of experience in data engineering, Bl engineering, or data warehouse development.
Proficient in one or more programming languages, preferably Python or Java.
Strong expertise in writing complex SQL queries for data extraction and manipulation.
Familiarity with data workflow and orchestration tools such as Apache Flink and Airflow.
Working knowledge of DBT for data transformation and pipeline development.
Experience with container orchestration using Kubernetes.
In-depth understanding of Apache Spark architecture and internals, with hands-on experience in large-scale data processing.
Experience working with distributed SQL engines such as Amazon Athena or Presto.
Proven experience in building and managing ETL data pipelines.
Ability to evaluate and select appropriate technologies based on principles of reliability, scalability, and maintainability, avoiding unnecessary complexity or buzzwords.
Skills:
Data Modeling and Architecture: Design and implement scalable and efficient data models; develop and maintain conceptual, logical, and physical data architectures.
ETL Development: Build, optimize, and maintain ETL pipelines to enable efficient data movement across systems; implement data transformation and cleansing processes to ensure accuracy and integrity.
Data Warehouse Management: Contribute to the design, development, and maintenance of data warehouses to support analytical and reporting needs.
Data Integration: Collaborate with cross-functional teams to integrate data from multiple sources; design and implement both real-time and batch data integration solutions.
Data Quality and Governance: Establish, monitor, and enforce data quality standards in alignment with governance policies and best practices.
Performance Tuning: Monitor and optimize database performance for large-scale datasets; identify, troubleshoot, and resolve issues related to data processing, storage, and scalability.
Apply Link: https://www.zeta.tech/in/careers/job-description?cae42b40-38c4-4a55-afe7-95027b44d42b
Additionally you can follow the Linkedin group- jobs and referrals would be posted!
LinkedIn Group :👇
https://www.linkedin.com/groups/14489007
👉WhatsApp Channel : 👇
https://whatsapp.com/channel/0029VaFEYqW4Y9lfZmlQsP44
👉Telegram Channel :👇
https://t.me/panindjobs