ABOUT THE MISSION
Relitix is currently seeking a hands-on Senior Data Engineer to lead a critical transformation of our data platform. We are moving away from legacy structures and re-engineering our ecosystem into a modern, Python-scripted Lakehouse architecture. This is not a maintenance role; it is a builder role for an engineer who values code-first engineering over GUI tools.
If you are energized by owning systems end-to-end and want to build a platform that will scale significantly, we want to meet you. You will own the mission of migrating ADF-based pipelines to Databricks, ensuring internal BI data is accurate, and establishing secure governance via Unity Catalog.
WHO YOU ARE
We don’t just list responsibilities; we define clear outcomes. To achieve them, we look for specific traits in our engineers. These are the principles that will guide your success here:
Code-First Mindset | Deep Ownership | Automation First | Governance Minded | Builder Energy
The purpose of the Senior Data Engineer at Relitix is to modernize core architecture and secure the platform while ensuring reliable downstream BI. You don’t just spot problems; you fix them.
Job Responsibilities
• Modernize Core Architecture: Lead the full migration of priority pipelines into modular Python/Spark jobs running in Databricks, moving away from legacy ADF dependencies.
• Optimize Performance: Re-architect pipelines to significantly reduce runtime and compute costs through partitioning, caching, and optimized orchestration.
• Stabilize Data Ingestion: Architect and deploy robust Python connectors for API integrations (MLS, CRM, operational systems) with consistent logging and error handling.
• Elevate Data Quality: Take ownership of internal BI datasets, implementing automated validation tests to ensure dashboards and operational metrics remain consistent and reliable.
• Secure the Platform: Establish a governance foundation by implementing RBAC roles and Unity Catalog controls across bronze, silver, and gold data layers.
Job Requirements
• Strong Python Engineering: Strong Python engineering skills for ETL/ELT with a preference for writing clean, maintainable code over using drag-and-drop GUI tools.
• Spark/Databricks Expertise: Experience with PySpark and SparkSQL inside Databricks, comfortable designing and building out Delta Lake and medallion architectures.
• Governance Knowledge: Understanding of the importance of governance with exposure to Unity Catalog or similar frameworks for managing data permissions.
• Automation Mindset: Approach problems with an automation-first mindset, utilizing Git-based workflows and CI/CD principles.
• Accountability: High sense of ownership and accountability to ensure reliable downstream BI.
Preferred Qualifications
• Experience with data quality tools (e.g., Great Expectations).
• Familiarity with MLflow or AI-ready pipeline structures.
• Prior experience in multi-tenant SaaS data environments.
• Madison, WI, Milwaukee, WI or Chicago IL area preferred
Job Types: Full-time, Permanent
Pay: $90,000.00 - $115,000.00 per year
Benefits:
• 401(k)
• Dental insurance
• Flexible schedule
• Health insurance
• Health savings account
• Paid time off
• Vision insurance
Work Location: Remote
Apply Now
Apply Now