Senior Data Engineer Job Description Template
Our company is looking for a Senior Data Engineer to join our team.
Responsibilities:
- Coach and mentor both peers and junior members of Engineering in to develop a data-driven culture with supporting practises;
- Investigate, test and introduce new ways of working and technologies where appropriate;
- Assemble large, complex data sets that meet functional / non-functional business requirements;
- Constantly seek new and creative ways to automate workflows and iterate on system improvements;
- As part of our Platform crew, collaborate with Platform (DevOps) and Test Engineers on release and test processes in an incremental and agile manner;
- Work with data and analytics experts to strive for greater functionality in our data systems;
- Build efficient and reliable data pipelines, ingest and transform data sets using SQL, Python & AWS;
- Collaborate with cross-functional peers (e.g. Data Architect and Data Scientist – TBH) to development data governance and management processes;
- Writing technical specifications for projects as required;
- Collaborate with Business Analysts on business insight and new product requirements;
- Provide logical and physical data design and database modelling;
- Transfer knowledge to DataOps team, ensuring the design and build are documented and hence supportable;
- Convert algorithms, models and features into production solutions;
- Set the standards for software engineering practices within the team;
- There will be a mix of maintenance work and redesigning and optimising old code as well as creating new pipelines.
Requirements:
- Experience with persistence and caching technologies (Cassandra, AWS DynamoDB, RDBMS, Redis, Memcached, Hazelcast, etc.);
- AWS cloud services: EC2, EMR, RDS, Redshift, Athena, Lambda;
- Data pipeline and workflow management tools: Airflow, AWS Glue, Step functions etc;
- Hands on experience of building streaming pipelines using Kafka/AWS Kinesis;
- Have proven experience working in a cloud environment with GCP being strongly preferred;
- Familiarity with Linux/UNIX and tools including Git and SVN;
- Proven experience developing large scale data solutions with technology such as Python, Scala, and Spark, preferably for telco, digital and app data;
- Familiarity with software engineering principles and product delivery;
- Educated to a Bachelor’s degree in Computer Science or related field;
- Knowledge of at least one scripting language i.e. Python;
- Educated to a bachelor’s degree level in Computer Science or a related field;
- Performance tuning queries;
- Ability to communicate with diverse stakeholders;
- Strong knowledge and aptitude at building and maintaining data warehouses in Redshift/Snowflake;
- Post-graduate in Computer Science, Computer Engineering, Physics or a related subject.