Work type : Office / Remote Technical Level : Middle Job Category : Software Development
We are seeking a Middle Data Engineer with proven expertise in AWS, Snowflake, and dbt to design and build scalable data pipelines and modern data infrastructure. You'll play a key role in shaping the data ecosystem, ensuring data availability, quality, and performance across business units.
Requirements :
- 4+ years of experience in Data Engineering roles.
- Experience with the AWS cloud platform.
- Proven experience with Snowflake in production environments.
- Hands-on experience building data pipelines using dbt .
- Python skills for data processing and orchestration.
- Deep understanding of data modeling and ELT best practices.
- Experience with CI / CD and version control systems (e.g., Git).
- Strong communication and collaboration skills.
Must-Have :
Strong experience with Snowflake (e.g., performance tuning, storage layers, cost management)Production-level proficiency with dbt (modular development, testing, deployment)..Experience developing Python data pipelines.Proficiency in SQL (analytical queries, performance optimization).Nice-to-Have :
Experience with orchestration tools like Airflow, Prefect, or Dagster.Familiarity with cloud platforms (e.g., GCP, or Azure).Knowledge of data governance, lineage, and catalog tools.Experience in working in Agile teams and CI / CD deployment pipelines.Exposure to BI tools like Tableau or Power BI.We offer
Flexible working format - remote, office-based or flexibleA competitive salary and good compensation packagePersonalized career growthProfessional development tools (mentorship program, tech talks and trainings, centers of excellence, and more)Active tech communities with regular knowledge sharingEducation reimbursementMemorable anniversary presentsCorporate events and team buildingsOther location-specific benefitsnot applicable for freelancers