In this role, you will collaborate with project teams to design and develop a cutting-edge data platform in the AWS cloud for a global pharmaceutical leader. Working in a dynamic, international environment, you will focus on cloud-native Big Data architecture, leveraging the latest technologies and best practices.
Warsaw Hybrid / Remote
Requirements
Minimum 5 years of experience in designing and building Business Intelligence, ETL / ELT, Data Warehouse, Data Lake, and Big Data solutions
Hands-on experience with various relational database engines (e.g., SQL Server / SQL Database, Oracle, PostgreSQL), non-relational databases (e.g., MongoDB, CosmosDB, DynamoDB, Neo4j), data warehouse solutions (e.g., Redshift, Synapse, BigQuery, Databricks, Snowflake), and MapReduce frameworks (Spark, Hadoop)
Very good knowledge of SQL and Python (minimum 5 years of experience)
Good understanding of Spark / PySpark
Ability to design solutions for processing large volumes of data (several TBs per day), including data processing optimization techniques (partitioning / distribution, broadcasting / replication, caching, etc.)
Ability to design data models within multi-layered architectures (e.g., medallion architecture) and using multidimensional modeling (Kimball methodology)
Understanding of data governance, data quality, and batch / streaming data processing concepts
Advanced proficiency with Git
Practical experience with data-related services in the AWS (or another cloud provider with a willingness to gain deep expertise in AWS services)
Experience conducting code reviews and defining standards around data services
Independence, effectiveness in action, and accountability for assigned tasks
Very good knowledge of English (min. C1)
Responsibilities
Taking initiative to improve existing solutions and designing new ones within the data platform — for both business and technical requests
Coordinating architectural decisions with client Architects and other vendors
Coordinating work of Data Engineers (responsible for the development of the data platform and ETL / ELT processes) and Data Analysts (responsible for the data model and report development) on the BitPeak side, across multiple parallel projects for the same client; close collaboration with Project Managers
Optimizing the delivery process by introducing both procedural and tooling improvements
Ensuring adherence to existing platform development standards, as well as defining and improving those standards where gaps are identified
Ensuring the quality of delivered solutions, including conducting code reviews
Hands-on work as a Data Engineer and Data Analyst (to stay in touch with the technology)
We offer
Global projects in cloud - we work with clients from all over the world based on modern cloud technologies
Certification reimbursement - we fund exams, certifications from Microsoft, AWS, Databricks, Snowflake
Time to learn - 60 paid hours per year
Flexible approach - you can choose to work from home or our office
Personalized benefits - medical care, subsidized sports packages, language tuition, new employee referral bonus (up to PLN 15,000) as well as annual and media bonus
If you are interested in the offer, send your CV
I believe that clear and honest communication is the determinant of successful cooperation. Through achieving those, we build a strong and cohesive team. If you are a candidate who values open and direct communication, then we would love to hear your questions about the company!
As a leader of Data Engineering department I emphasize an honest, effective communication, pragmatic approach to projects and initiative to improve your knowledge about technology. I focus on direct cooperation and actively using strenghts of my team.
If you have any question about our future cooperation - let's talk!
#J-18808-Ljbffr
Solution Architect • Warszawa, Województwo mazowieckie, Polska