Consultant Data Engineer (Python/Databricks)
About the company:
Softermii (https://www.softermii.com/) is an outsourcing/outstaffing company with more than 8 years of experience in web and mobile application development with a team of more than 120 specialists from 6 countries.
We have withstood multiple challenges, such as COVID-19 pandemic and war aggression against Ukraine, we have secured sustainable development processes, saved all our projects and jobs. Even more – we’re constantly growing.
Every day we add new project to our portfolio in the following areas of expertise:
- Healthcare
- Real estate
- Livestream E-commerce
- Fintech (WealthTech, Capital Trading)
- Real Time Communications (RTC)
In addition, we develop 2 of our own solutions that help our customers get faster applications for video or audio calls (https://www.softermii.com/solutions/vidrtc) or an application for real estate with 3D visualization
(https://www.softermii.com/solutions/apartmii)
Company’s mission:
Softermii: Succeed Without (L) imitations
We will turn any idea into a valuable product.
Our plans as a company:
We want to move from a regular outsourcing business to an ecosystem of outsourcing, education and product development companies.
Softermii is looking for a part-time Data Engineering Consultant / Tech Lead who will do technical interviews, assist with upcoming projects, and occasionally be hands-on with complex development tasks — including data pipeline design and solution optimization on Databricks.
Type of cooperation: Part-time
⚡️Your responsibilities on the project will be:
Interview and hire Data Engineers
- Supervise work of other Engineers and have hands on for the most complicated tasks from backlog, focus on unblocking other data Engineers in case of technical difficulties
- Develop and maintain scalable data pipelines using Databricks (Apache Spark) for batch and streaming use cases.
- Work with data scientists and analysts to provide reliable, performant, and well-modeled data sets for analytics and machine learning.
- Optimize and manage data workflows using Databricks Workflows and orchestrate jobs for complex data transformation tasks.
- Design and implement data ingestion frameworks to bring data from various sources (files, APIs, databases) into Delta Lake.
- Ensure data quality, lineage, and governance using tools such as Unity Catalog, Delta Live Tables, and built-in monitoring features.
- Collaborate with cross-functional teams to understand data needs and support production-grade machine learning workflows.
- Apply data engineering best practices: versioning, testing (e.g., with pytest or dbx), documentation, and CI/CD pipelines
🕹Tools we use: Jira, Confluence, Git, Figma
🗞Our requirements to you:
- 5+ years of experience in data engineering or big data development, with production-level work.
- Architect and develop scalable data solutions** on the Databricks platform, leveraging Apache Spark, Delta Lake, and the lakehouse architecture to support advanced analytics and machine learning initiatives.
- Design, build, and maintain production-grade data pipelines** using Python (or Scala) and SQL, ensuring efficient data ingestion, transformation, and delivery across distributed systems.
- Lead the implementation of Databricks features** such as Delta Live Tables, Unity Catalog, and Workflows to ensure secure, reliable, and automated data operations.
- Optimize Spark performance and resource utilization**, applying best practices in distributed computing, caching, and tuning for large-scale data processing.
- Integrate data from cloud-based sources** (e.g., AWS S3), ensuring data quality, lineage, and consistency throughout the pipeline lifecycle.
- Manage orchestration and automation of data workflows** using tools like Airflow or Databricks Jobs, while implementing robust CI/CD pipelines for code deployment and testing.
- Collaborate cross-functionally with data scientists, analysts, and business stakeholders** to understand data needs and deliver actionable insights through robust data infrastructure.
- Mentor and guide junior engineers**, promoting engineering best practices, code quality, and continuous learning within the team.
- Ensure adherence to data governance and security policies**, utilizing tools such as Unity Catalog for access control and compliance.
- Continuously evaluate new technologies and practices**, driving innovation and improvements in data engineering strategy and execution.
- Experience in designing, building, and maintaining data pipelines using Apache Airflow, including DAG creation, task orchestration, and workflow optimization for scalable data processing.
- Upper-Intermediate English level.
👨💻Who will you have the opportunity to meet during the hiring process (stages):
Call, HR, Tech interview, PM interview.
What we can offer you:
- We have stable and highly-functioning processes – everyone has their own role and clear responsibilities, so decisions are made quickly and without unnecessary approvals.
- You will have enough independence to make decisions that can affect not only the project but also the work of the company.
- We are a team of like-minded experts who create interesting products during working hours, and enjoy spending free time together.
- Do you like to learn something new in your profession or do you want to improve your English? We will be happy to pay 50% of the cost of courses/conferences/speaking clubs.
- Do you want an individual development plan? We will form one especially for you + you can count on mentoring from our seniors and leaders.
- Do you have a friend who is currently looking for new job opportunities? Recommend them to us and get a bonus.
- And what if you want to relax? Then we have 21 working days off.
- What if you are feeling bad? You can take 5 sick leaves a year.
- Do you want to volunteer? We will add you to a chat, where we can get a bulletproof vest, buy a pickup truck or send children's drawings to the front.
- And we have the most empathetic HRs (who also volunteers!). So we are ready to support your well-being in various ways.
A little more information that you may find useful:
- our adaptation period lasts 3 months, this period of time is enough for us to understand each other better;
- there is a performance review after each year of our collaboration where we use a skills map to track your growth;
- we really have no boundaries in the truest sense of the word – we have flexible working day, remote or office is up to you.
Of course, we have a referral bonus system. So if you have a friend who likes our job opening, you can give us their telegram contact, and we in turn will thank you with a bonus if we start working with them!
Apply for this job
*
indicates a required field