Role- DEVELOPER
RelevantExperience (in Yrs) 4-6 YRS
MustHave Technical / Functional Skills 3years of experience in Databricks engineering solutions on AWS Cloud platformsusing PySpark, Databricks SQL, Data pipelines using Delta Lake.
- 3 years of experience years of experience inETL, Big Data / Hadoop and data warehouse architecture & delivery.
- Minimum 3 year of Experience in one or moreprogramming languages Python, Java, Scala.
- 2 years of experience developing CICDpipelines using GIT, Jenkins, Docker, Kubernetes, Shell Scripting, Terraform
- Good knowledge on Data Warehousing
ExperienceRequired 4-6 YRS
Roles& Responsibilities Work on clientprojects to deliver AWS, PySpark, Databricks based Data engineering &Analytics solutions.
Build and operate very large data warehousesor data lakes.ETL optimization, designing, coding, &tuning big data processes using Apache Spark.Show efficiency in handling data - trackingdata lineage, ensuring data quality, and improving discoverability of data.