···
Log in / Register
Data Engineer
Negotiable Salary
Workable
Full-time
Onsite
No experience limit
No degree limit
Cambridge, UK
Favourites
Share
Description

Position: Data Engineer Location: Cambridge / Luton, UK (Hybrid 2-3 days onsite in a week) Duration: Long Term B2B Contract Job Description: The ideal candidate with a minimum of 5 +years of experience having strong experience working with Snowflake, DBT, Python, and AWS to deliver ETL/ELT Pipelines using different resources. • Proficiency in Snowflake data warehouse architecture Design, build, and optimize ETL/ELT pipelines using DBT (Data Build Tool) and Snowflake. • Experience with DBT (Data Build Tool) for data transformation and modelling. Implement data transformation workflows using DBT (core/cloud). • Strong Python programming skills for automation and data processing. Leverage Python to create automation scripts and optimize data processing tasks. • Proficiency in SQL performance tuning, and query optimization techniques using snowflake. • Troubleshoot and optimize DBT models, and Snowflake performance. • Knowledge of CI/CD, version control (Git) tools. Experience with orchestration tools such as Airflow, • Strong analytical and problem-solving skills with an ability to work in agile development environment independently. • Ensure data quality, reliability, and consistency across different environments. • Collaborate with other data engineers, data analysts, and business stakeholders to understand data needs and translate them into engineering solutions. • Certification in AWS, Snowflake, or DBT is a plus. Requirements Position: Data Engineer Location: Cambridge / Luton, UK (Hybrid 2-3 days onsite in a week) Duration: Long Term B2B Contract Job Description: The ideal candidate with a minimum of 5 +years of experience having strong experience working with Snowflake, DBT, Python, and AWS to deliver ETL/ELT Pipelines using different resources. • Proficiency in Snowflake data warehouse architecture Design, build, and optimize ETL/ELT pipelines using DBT (Data Build Tool) and Snowflake. • Experience with DBT (Data Build Tool) for data transformation and modelling. Implement data transformation workflows using DBT (core/cloud). • Strong Python programming skills for automation and data processing. Leverage Python to create automation scripts and optimize data processing tasks. • Proficiency in SQL performance tuning, and query optimization techniques using snowflake. • Troubleshoot and optimize DBT models, and Snowflake performance. • Knowledge of CI/CD, version control (Git) tools. Experience with orchestration tools such as Airflow, • Strong analytical and problem-solving skills with an ability to work in agile development environment independently. • Ensure data quality, reliability, and consistency across different environments. • Collaborate with other data engineers, data analysts, and business stakeholders to understand data needs and translate them into engineering solutions. • Certification in AWS, Snowflake, or DBT is a plus.

Source:  workable View original post
Workable · HR

Company

Workable
Cookie
Cookie Settings
Our Apps
Download
Download on the
APP Store
Download
Get it on
Google Play
© 2025 Servanan International Pte. Ltd.