Responsibilities
- Help design, build, and maintain scalable data pipelines in Databricks using PySpark, Spark SQL, Delta Lake, Auto Loader, and Databricks Workflows for data ingestion, transformation, and analytics delivery.
- Support the analysis of data-related challenges and implement solutions aligned with business and technical requirements.
- Collaborate with data analysts and engineers to understand data needs and deliver clean, analytics-ready datasets.
- Assist in developing and maintaining data models, transformations, and workflows in Databricks, following Delta Lake best practices.
- Monitor data pipelines and troubleshoot issues related to data quality, failures, and performance bottlenecks.
- Support data governance and access management in Unity Catalog, ensuring proper organization and handling of datasets.
- Use Auto Loader and Databricks Workflows to automate data ingestion and orchestration from multiple sources.
- Follow software engineering best practices such as testing, documentation, version control, and peer review.
- Work with IT, platform, engineering, and product teams to deliver reliable data solutions to production.
- Ensure compliance with data security, governance, and regulatory standards.
- Communicate effectively with stakeholders and team members to align on priorities, share progress, and support delivery.
- Operate within an Agile development environment using source control, CI/CD, automated testing, and structured release processes.
Requirements
- Bachelor’s or Master’s degree in Computer Science or a related field.
- Experience designing robust technical solutions for complex software problems, particularly in data systems.
- Familiarity with modern software development practices and architectural patterns.
- Proven ability to communicate with technical and non-technical stakeholders to align requirements.
- Hands-on experience in software prototyping, development, and testing.
- Experience building ETL/ELT pipelines supporting both batch and real-time data processing.
- Proficiency in Python and Apache Spark technologies such as PySpark, Spark SQL, and Spark Streaming.
- Experience using data manipulation libraries like Pandas and NumPy.
- Strong skills in SQL for querying and transforming data.
- Knowledge of Databricks and associated tools including Delta Lake and Delta Live Tables.
- Understanding of both relational databases (e.g., SQL Server) and NoSQL systems (e.g., MongoDB).
- Demonstrated ownership, attention to detail, and commitment to high-quality deliverables.
- Strong written and spoken communication skills in English.
Nice to Have
- Familiarity with message queues and streaming platforms such as Azure Event Hubs or Kafka.
- Experience with cloud-native technologies, particularly Microsoft Azure.
Benefits
- competitive pay
- generous paid time off
- flexible working options including XFlex programme and a 'work from anywhere' allowance
- enhanced leave for all new parents, regardless of gender, family structure, or path to parenthood
- comprehensive healthcare coverage
- fully funded insurance and income protection
- access to 24/7 virtual care, mental health support, legal advice, and financial counselling through employee assistance programmes
- contributory retirement plans and savings support
- subscriptions to Calm and Blinkist, fitness and lifestyle credits, global travel assistance, and various discounts
Compensation
competitive pay
Work Arrangement
Hybrid
Team
300+ team members in a global award-winning SaaS provider with strong revenue growth, ambitious expansion goals, and a culture focused on inclusion and results.
Other
If you require any reasonable adjustments or support during the application or hiring process, please email Work@Exclaimer.com confidentially to notify the team.