
Databricks, PySpark Data Engineer
Doran Jones Inc.
full-time
Posted on:
Location Type: Hybrid
Location: Dallas • Texas • United States
Visit company websiteExplore more
About the role
- We are seeking a hands-on Data Engineer with strong Databricks and PySpark experience to build scalable data pipelines and analytics applications within a modern data platform.
- This role focuses on modernizing legacy ETL and reporting systems (Teradata, Informatica, Tableau) into Databricks-native pipelines, dashboards, and Python-based data applications. The ideal candidate operates across both data engineering and lightweight application development.
Requirements
- 5+ years of hands-on experience building data pipelines using PySpark in production environments
- Strong experience with the Databricks platform (workspaces, clusters, Jobs & Workflows, Unity Catalog)
- Experience building analytics dashboards within Databricks (Databricks SQL)
- Proven experience designing and building scalable ETL/ELT data pipelines
- Strong Python development skills, including building REST APIs or data services
- Experience building or supporting data-driven applications (not just traditional ETL pipelines)
- Solid understanding of data modeling, including dimensional modeling and transformation patterns
- Experience using AI-assisted development tools (e.g., Copilot, ChatGPT) in engineering workflows
- Exposure to LLM integration or an AI-powered data application
- Familiarity with cloud platforms (AWS, Azure, or GCP)
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard Skills & Tools
DatabricksPySparkETLELTPythonREST APIsdata modelingdimensional modelingAI-assisted development toolsLLM integration