Job Description
Data / DevOps Engineer
Perm
Located in Stretford, Trafford Park and comprises of Hybrid working.
Up to £65,000pa
You will be working in the Data Engineering team whose main function is developing, maintaining and improving the end-to-end data pipeline. This includes real-time data processing; extract, transform, load (ETL) jobs; artificial intelligence; and data analytics on a complex and large dataset.
Your role will primarily be to perform DevOps, backend and cloud development on the data infrastructure to develop innovative solutions to effectively scale and maintain the data platform. You will be working on complex data problems in a challenging and fun environment, using some of the latest Big Data open-source technologies like Apache Spark, as well as Amazon Web Service technologies including Elastic MapReduce, Athena and Lambda to develop scalable data solutions.
Key Responsibilities:
· Adhering to Company Policies and Procedures with respect to Security, Quality and Health & Safety.
· Writing application code and tests that conform to standards.
· Developing infrastructure automation and scheduling scripts for reliable data processing.
· Continually evaluating and contribute towards using cutting-edge tools and technologies to improve the design, architecture, and performance of the data platform.
· Supporting the production systems running the deployed data software.
· Regularly reviewing colleagues’ work and providing helpful feedback.
· Working with stakeholders to fully understand requirements.
· Be the subject matter expert for the data platform and supporting processes and be able to present to others to knowledge share.
About You:
Here’s what we’re looking for:
· The ability to problem-solve.
· Knowledge of AWS or equivalent cloud technologies.
· Knowledge of Serverless technologies, frameworks and best practices.
· Apache Spark (Scala or Pyspark)
· Experience using AWS CloudFormation or Terraform for infrastructure automation.
· Knowledge of Scala or OO language such as Java or C#.
· SQL or Python development experience.
· High-quality coding and testing practices.
· Willingness to learn new technologies and methodologies.
· Knowledge of agile software development practices including continuous integration, automated testing and working with software engineering requirements and specifications.
· Good interpersonal skills, positive attitude, willing to help other members of the team.
· Experience debugging and dealing with failures on business-critical systems.
Preferred:
· Exposure to Apache Spark, Apache Trino, or another big data processing system.
· Knowledge of streaming data principles and best practices.
· Understanding of database technologies and standards.
· Experience working on large and complex datasets.
· Exposure to Data Engineering practices used in Machine Learning training and inference.
· Experience using Git, Jenkins and other CI/CD tools