Become a leader of our data engineering team and help us build the foundation of our data-driven success! As a Data Architect , you will be at the forefront of designing, implementing, and maintaining our data infrastructure. You will lead a small team of data engineers, working collaboratively to build data lakes, data warehouses, pipelines, and monitoring systems. Your expertise will be crucial in ensuring data quality, accessibility, and reliability across the organization.
This is a fully in-office role based in Kraków
Key Responsibilities
- Data Infrastructure Development : Lead the design and implementation of data lakes, data warehouses, pipelines and DataOps.
- Data Modeling : Develop and maintain data models to support data warehousing and analytical use cases.
- Data Orchestration and DataOps : Oversee the creation of robust data pipelines, ensuring efficient data flow and (near) real-time processing.
- Data Acquisition : Collaborate with various teams to acquire and integrate data from multiple sources, including large industry-secific databases and primary data provided by partners.
- Monitoring Systems : Establish monitoring systems to ensure data quality, accuracy, and availability.
- Leadership : Manage and mentor a team of data engineers, fostering a collaborative and innovative environment.
- International Collaboration : Work effectively within a diverse, international team spread across Switzerland, Serbia, and China.
Required Hard Skills
University degree, ideally a PhD or Masters in Computer Science or related science, technology or engineering field.8+ years of relevant work experience in data engineering roles.Expertise in designing larger scale data systems.Expertise in data acquisition, laking, warehousing, modeling, and orchestration.Strong knowledge of common columnar file formats used in data applications.Proficiency in data partitioning and incremental scalability.Proficiency in data quality and data governance.Proficiency in SQL (including window functions and CTE).Hands-on experience with RDBMS MySQL, PostgreSQL).Proficiency in Python (with libraries like Polars, optionally Arrow / PyArrow API).Proficiency with OLAP query engines Clickhouse, DuckDB, Apache Spark).Proficiency with Apache Airflow (or similar orchestration tools like Dagster or Prefect).Experience with Linux (Ubuntu / Debian).Experience with dbt.Familiarity with orchestration and containerization technologies Docker, Kubernetes).Good to Have
Experience in entity disambiguation.Experience with Git and Atlassian tools (Jira, Confluence).Understanding of graph databases neo4j).Required Soft Skills
Strong teamwork and communication abilities.Self-motivated with the ability to work independently.Why Join Us?
Be part of a dynamic and innovative team at the forefront of data technology.Opportunity to lead and shape a new team, driving impactful projects.Work in a collaborative environment that values creativity and diversity.Competitive salary.Private medical care (fully covered).MultiSport card (partially covered).Team building activities.Li-Onsite