This job has been added to your Saved jobs.
You have reached the limit of 20 Saved Jobs. If you want to create a new one, please manage your Saved Jobs.
Job Expertise:
Job Domain:
IT Services and IT Consulting
Software Development Outsourcing
Software Products and Web Services
AI Software & Services
Top 3 reasons to join us
- Disruptive innovations
- People-oriented philosophy of doing business
- Your footprint on regional on-demand market
Job description
You will lead Ahamove’s Data Engineering team, responsible for building and scaling our terabyte-scale data warehouse. Our systems handle nearly 200,000 daily orders across both real-time streaming and batch processing pipelines. Your mission is to ensure the reliability, scalability, and performance of our data infrastructure, empowering:
- Real-time dashboards for operational visibility
- Machine Learning services to power intelligent decision-making
- Robust query experiences for internal teams and external stakeholders
Main Duties:
- Build, maintain and optimize in-house data infrastructure including database, data warehouse, orchestration system, data streaming and batching pipelines.
- Work with multiple Cloud Platforms such as GCP, AWS, Databricks.
- Work with big and complex datasets and multi databases for multiple departments.
- Create a benchmark, alert, and audit log for the data system to ensure stability and scalability of data and system.
- Lead data engineer team and responsible for tech-stack and tech-cost of data platforms
- Communicate with stakeholders including Product Owner, Software Engineers, Business Users, Data Analysts and Machine Learning Engineer to solve data-related problems.
Your skills and experience
- Bachelor degree in Computer Science or Software Engineer or Information System
- Specializing in data science or a higher degree is a big plus.
- At least 06-year-experience in data engineer role and building data platforms and pipelines for analytics.
- Excellence in Python and one programming language Java, Scala, Go, Javascript, Typescript, R is a big plus.
- Excellence at SQL in DBMS.
- Excellence with cloud services such as GCP or AWS.
- Experience with Hadoop, Spark, Databricks is a big plus.
- Excellence with various OLTP and OLAP databases and data warehouses: MongoDB, PostgreSQL, BigQuery, ClickHouse, MotherDuck, etc.
- Excellence with streaming process platforms and streaming concepts such as Redpanda, Kafka, RabbitMQ, CDC or any open-sources.
- Experience with data pipeline and workflow management tools: Airflow, dbt, Airbyte, etc.
- Knowledge of data visualization tools like Metabase, PowerBI, Looker Studio, etc.
- Exposure to emerging open-source or technologies to apply.
- Excellence with using source version control such as gitlab, github.
- Experience with Kubernetes or DevOps skills (Linux, Networking) is a big plus.
Why you'll love working here
- Physical Wellbeing Benefit: General Insurance, Medical check-up, Accident Insurance, Healthcare Insurance
- Emotional Wellbeing Benefit: Company Trip, Year End Party, Aha Hour Activities, Special Day Gifts, Aha Club (Badminton, Soccer)
- Financial Wellbeing Benefit: Grab/Be For Work (Tech/Lead Level), Workplace Relocation, 13th Month Salary, PP Appreciate, Annual Leave Remain
More jobs for you
Get similar jobs by email
SUPER HOT
Posted
7 days ago
Head of Data (Azure, Data Engineer, Data Analysis, ETL)
At office
Ho Chi Minh
Feedback