UIX Labs Logo

Data Engineer

UIX Labs

All India, Navi Mumbai • 1 month ago

Experience: 3 to 7 Yrs

PREMIUM
Deal of the Day --:--:--

15 Days Free Trial

After Free Trial → Flat 50% OFF

Upgrade to CVX24 Premium

Offer Announcement Banner
  • Free Resume Writing
  • Get a Verified Blue tick
  • See who viewed your profile
  • Unlimited chat with recruiters
  • Rank higher in recruiter searches
  • Get up to 10× more recruiter visibility
  • Auto-forward profile to 10 top recruiters
  • Receive verified recruiter messages directly
  • Unlock hidden jobs, not visible to free users

A small token amount will be charged to verify. Get Refund in 48 Hours.
Free Earplugs Delivery Only after Payment of Rs. 99 for Five Consecutive Months.
After free-trial 6 Months subscription will be auto Activated @ $ 1 (Cancel Anytime). Quoted price includes 50% discount.

Job Description

As a Data Engineer at our company, you will be responsible for designing, developing, and maintaining real-time and batch data pipelines in Databricks. Your key responsibilities will include: - Integrating Apache Kafka with Databricks using Structured Streaming. - Implementing robust data ingestion frameworks using Databricks Autoloader. - Building and maintaining Medallion Architecture pipelines across Bronze, Silver, and Gold layers. - Implementing checkpointing, output modes, and appropriate processing modes in structured streaming jobs. - Designing and implementing Change Data Capture (CDC) workflows and Slowly Changing Dimensions (SCD) Type 1 and Type 2 logic. - Developing reusable components for merge/upsert operations and window function-based transformations. - Handling large volumes of data efficiently through proper partitioning, caching, and cluster tuning techniques. - Collaborating with cross-functional teams to ensure data availability, reliability, and consistency. Qualifications required for this role include: - Strong hands-on experience with Apache Kafka, including integration, topic management, and schema registry (Avro/JSON). - Expertise in Databricks and Spark Structured Streaming. - Proficiency in writing efficient queries and transformations using SQL & Spark SQL. - Knowledge of data governance practices, including schema enforcement, data quality checks, and monitoring. Additionally, it would be beneficial if you have: - Strong coding skills in Python and PySpark. - Experience working in CI/CD environments for data pipelines. - Exposure to cloud platforms such as AWS, Azure, or GCP. - Understanding of Delta Lake, time travel, and data versioning. - Familiarity with orchestration tools like Airflow or Azure Data Factory. This position is based in Mumbai and requires 3-6 years of experience. If you are passionate about data engineering and have a proven track record of success in building and optimizing data pipelines, we encourage you to apply for this exciting opportunity. As a Data Engineer at our company, you will be responsible for designing, developing, and maintaining real-time and batch data pipelines in Databricks. Your key responsibilities will include: - Integrating Apache Kafka with Databricks using Structured Streaming. - Implementing robust data ingestion frameworks using Databricks Autoloader. - Building and maintaining Medallion Architecture pipelines across Bronze, Silver, and Gold layers. - Implementing checkpointing, output modes, and appropriate processing modes in structured streaming jobs. - Designing and implementing Change Data Capture (CDC) workflows and Slowly Changing Dimensions (SCD) Type 1 and Type 2 logic. - Developing reusable components for merge/upsert operations and window function-based transformations. - Handling large volumes of data efficiently through proper partitioning, caching, and cluster tuning techniques. - Collaborating with cross-functional teams to ensure data availability, reliability, and consistency. Qualifications required for this role include: - Strong hands-on experience with Apache Kafka, including integration, topic management, and schema registry (Avro/JSON). - Expertise in Databricks and Spark Structured Streaming. - Proficiency in writing efficient queries and transformations using SQL & Spark SQL. - Knowledge of data governance practices, including schema enforcement, data quality checks, and monitoring. Additionally, it would be beneficial if you have: - Strong coding skills in Python and PySpark. - Experience working in CI/CD environments for data pipelines. - Exposure to cloud platforms such as AWS, Azure, or GCP. - Understanding of Delta Lake, time travel, and data versioning. - Familiarity with orchestration tools like Airflow or Azure Data Factory. This position is based in Mumbai and requires 3-6 years of experience. If you are passionate about data engineering and have a proven track record of success in building and optimizing data pipelines, we encourage you to apply for this exciting opportunity.

Posted on: March 30, 2026

Relevant Jobs

Cloud Data Architecture Database Administrator

Accenture

All India

View Job →

Cloud Data Architecture Database Administrator

Accenture

All India

View Job →

Data Engineer

UIX Labs

All India, Navi Mumbai

View Job →

Data Engineer

UIX Labs

All India, Navi Mumbai

View Job →

Data Engineer

UIX Labs

All India, Navi Mumbai

View Job →

Data Engineer

UIX Labs

All India, Navi Mumbai

View Job →

Data Engineer

UIX Labs

All India, Navi Mumbai

View Job →

Data Engineer

UIX Labs

All India, Navi Mumbai

View Job →

Data Engineer

UIX Labs

All India, Navi Mumbai

View Job →

Data Engineer

UIX Labs

All India, Navi Mumbai

View Job →