Data Engineer
Hays plc
Watford, United Kingdom
2 days ago
Role details
Contract type
Temporary to permanent Employment type
Full-time (> 32 hours) Working hours
Regular working hours Languages
English Experience level
Senior Compensation
£ 143KJob location
Remote
Watford, United Kingdom
Tech stack
Java
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Data analysis
Big Data
Cloud Computing
Databases
Continuous Integration
Information Engineering
ETL
Data Systems
Data Warehousing
DevOps
Amazon DynamoDB
Hadoop
Python
Open Source Technology
Scala
Amazon Web Services (AWS)
Unstructured Data
Data Storage Technologies
State Machines
Build Management
Amazon Web Services (AWS)
Data Lake
PySpark
Infrastructure Automation Frameworks
Cloudwatch
Data Pipelines
Redshift
Job description
This role sits within our client's rapidly growing Cloud Data Platforms team, part of the Insights and Data Global Practice. You will join a multidisciplinary group of data and platform specialists who deliver modern cloud-based transformation for clients across a range of sectors. In this role, you will design and build data pipelines, develop ETL/ELT processes, and create innovative data solutions using the latest cloud technologies and frameworks across AWS.
Some responsibilities include:
- Build data pipelines to ingest, process and transform data for analytics and reporting.
- Develop ETL/ELT workflows to move data efficiently into data warehouses, data lakes and lake houses using open-source and AWS tooling.
- Apply DevOps practices, including CI/CD, infrastructure as code and automation, to improve and streamline data engineering processes.
- Design effective data solutions that meet complex business needs and support informed decision-making., Senior Data Visualisation Engineer | TypeScript | JavaScript | SVG | Canvas | London Want to make an application Make sure your CV is up to date, then read the following job specs carefully before applying. London | to £200,000 depending on exp. I'm hiring a Senior...
Requirements
- Strong AWS expertise, including tools such as Glue, Lambda, Kinesis, EMR, Athena, DynamoDB, CloudWatch, SNS and Step Functions.
- Skilled in modern programming, particularly Python, Java, Scala and PySpark.
- Solid knowledge of data storage and big data technologies, including data warehouses, databases, Redshift, RDS and Hadoop.
- Experience building and managing AWS data lakes on S3 for both structured and unstructured data.
About the company
Job Description Make your application after reading the following skill and qualification requirements for this position. Your new companyJoin a leading building services consultancy with a reputation for delivering mission-critical data centre projects across the UK and...