About the Role
Create and maintain optimal data pipeline architecture.
Build processes supporting data transformation, data structures, metadata, dependency and workload management.
Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
Assemble large, complex data sets that meet functional / non-functional business requirements.
Reformulating existing frameworks to optimize their functions.
Testing such structures to ensure that they are fit for use.
Preparing raw data for manipulation by Data Scientists.
Remaining up-to-date with industry standards and technological advancements that will improve the quality of your outputs.
Bachelor degree in Computer Science, Computer Engineering or a related technical field.
2+ years experience in ETL, data processing, database programming and data analytics or similar.
Expert proficiency in Python, Java and SQL.
Familiarity with Apache Spark, Hadoop or suitable equivalent.
Experience in developing cloud applications running on AWS (Amazon Web Services) and GCP (Google Cloud Platform).
Experience in designing and implementing systems using emerging technology.
Strong analytic skills related to working with unstructured datasets.