Data Engineer
Apply NowJob details
6 months contract/45 pounds an hour Inside it 35 PAYE/Fully remote Who you'll work with You’ll be a member of a global team working on GenAI initiative. McKinsey’s Tech Ecosystem function is responsible for developing and delivering all technology solutions for the firm’s internal use. We are taking a cloud-first approach to transform our data platforms and analytical applications across the firm. We are seeking an experienced data engineer to shape and accelerate the delivery of target state data platform enabling GenAI use-cases. We want a passionate specialist who loves to build data solutions as part of multi-disciplinary team, working closely with digital product professionals, data scientists, cloud engineers and others. Your impact within our firm We are looking for a Data Engineer with expertise in Python development, who is passionate about cloud-based data engineering using AWS services and loves to build data solutions as part of a multi-disciplinary team. You would be working closely with digital product professionals, data scientists, cloud engineers and others. You’ll be a member of a global team working on the GenAI initiative, based in one of our European offices. Our company's Tech Ecosystem function is responsible for developing and delivering all technology solutions for the firm’s internal use. You will work in a team of data engineers to develop data ingestion pipelines, create and mature data processing capabilities that ingest data into a data system used by GenAI applications. Your work would include but won't be limited to creation of the python code, tests, creation and modification of GitHub Action CICD pipelines, working with AWS-based infrastructure and docker containers. What you'll do You will work in a team of data engineers to develop data ingestion pipelines, create and mature data processing capabilities that ingest data into a data system used by GenAI applications. Work includes but not limited to creation of the python code, tests, creation and modification of GitHub Action CICD pipelines, working with AWS-based infrastructure and docker containers Skills: [Must] Experience in building production-grade data pipelines in Python, preferably on AWS [Must] Practicing high coding standards with clean code, modularity, error handling, testing automation and more [Must] Strong experience with relational databases, understanding of vector stores and its specifics [Must] Very Driven, superstrong on execution and output orientation, likes to get stuff done attitude and ability to figure things out independently. Able to work in complex and very fast paced environment Hands-on experience with Docker as big plus Solid and demonstrable background in pipeline performance and diagnostics Interest in Generative AI and other ML topics Kedro experience as a plus Holds their ground, opinionated, not afraid to speak up at any level Familiarity with agile principles and product development Willing to be flexible in hours: Half of team is based in US East Coast, the other half is located in Europe Education: Bachelor's degree in quantitative field like Computer Science, Engineering, Statistics, Mathematics or related field required. Advanced degree is a strong plus
Apply Now