Maintain and improve Redshift data warehousing system: Databases, ETL/ELT, data streaming system
Monitoring data integrity, performance, advising and implementing necessary infrastructure changes
Selecting and integrating any Big Data tools and frameworks (EMR Spark, AWS Athena, etc.) required to provide requested capabilities
Participating in data product development, with a focus on:
The implementation of practical machine learning solutions
Bringing data solutions in production (REST API)
2-3 years experience in software engineering/ data engineering / ops
Hands-on working experience with large-scale datasets
Databases: Practical knowledge with SQL and no-SQL databases. You’re comfortable with querying and writing to databases.
Very proficient in Python.
Linux sys-admin skills
Self-starter, natural planner who looks ahead, raises issues, resolves them and meet deadlines
Good to have (but not a must)
Hands-on experience with Machine Learning (classification, clustering)
Proficient in a compiled language would be a plus.
Familiarity with AWS (DynamoDB, Redshift, S3, EC2, RDS)
Understanding of some BI Tools (Tableau, Qlikview, etc.)
Experience in creating a REST API that can handle a production load (code + deploy)
The above role is a permanent position and a critical hire. The headcount has been approved.
We regret that only candidate based in Singapore will be considered due to Singapore Covid19 Travel and Quarantine restrictions.