1. Create and maintain end-to-end data pipeline for optimal data quality and reliability.
2. Assemble large, complex data sets that meet functional / non-functional business requirements.
3. Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
4. Work with data analytics/scientist team to strive for greater functionality in our data systems.
1. Min Bachelor Degree with IPK 3.00.
2. Proficient in writing SQL queries
3. Familiar with Linux Environment and Shell Scripting.
4. Familiar with Big Data tools: Hadoop, Spark, Kafka, etc.
5. Familiar with ETL tools: Informatica, Talend, Pentaho, etc.
6. Familiar with NoSQL databases: HBase, etc.
7. Familiar with object-oriented/object function scripting languages: Python, Java.
8. Technical experience with designing data models