Job Role : Data Engineer
Location : Phoenix, AZ (Onsite)
Type : Long Term Contract
Experience : 8+ Years
Skills : Pyspark, Spark, Big Data / Data Warehousing, Python, Hive, CI / CD, GitHub, Tableau, SQL
About the Role
We are seeking a highly skilled Data Engineer with strong expertise in PySpark, Python, and Hive, along with experience in Tableau integration development. The ideal candidate will have hands-on experience in accessing and working with Cornerstone tables and be able to quickly learn our existing data systems and reporting processes.
This role requires a solid development background with CI / CD pipelines, version control using GitHub, and a commitment to Engineering Excellence.
Key Responsibilities
- Design, develop, and maintain data pipelines using PySpark and Hive.
- Access, query, and manage data from Cornerstone tables for reporting and analytics purposes.
- Develop and enhance Tableau dashboards and integrate them with backend data sources.
- Collaborate with stakeholders to understand reporting requirements and translate them into technical solutions.
- Implement CI / CD pipelines to automate deployment processes.
- Maintain code repositories in GitHub following best practices in code quality and version control.
- Ensure adherence to Engineering Excellence standards, including code reviews, documentation, and testing.
- Troubleshoot and optimize data workflows for performance and scalability.
- Adapt quickly to existing systems, tools, and reporting processes.
Required Skills & Qualifications
Strong experience in PySpark and Python development.Proficiency in Hive and SQL-based data querying.Experience with Cornerstone tables and large-scale data environments.Proven skills in Tableau dashboard development and backend integration.Familiarity with CI / CD tools and practices.Proficiency in GitHub for version control.Ability to learn new systems and processes quickly.Strong problem-solving, communication, and collaboration skills.