Holidu is one of the world’s fastest-growing vacation rental technology companies. Our mission is to make booking and hosting holiday homes free of doubt and full of joy, by helping hosts generate more bookings with less work and helping guests find a holiday home they truly enjoy. Our team of 700 colleagues from 60+ nations shares a passion for tech, an ambition for constant improvement, and a relentless drive to bring the best experience to more than 40k vacation rental hosts and 4 million annual guests.
Your future team
You will be part of the Data Engineering team in the Business Intelligence department. This team is responsible for creating, maintaining, and refining the core data pipeline tools, as well as supporting several end-user pipelines. Your team will assist the organization in addressing all its data needs and, simultaneously, enhance the data platform with new capabilities and refactoring.
Your team will consist of 5 data engineers (3 Seniors, 1 Tech Lead manager and 1 Staff Engineer). You will work closely with the data analytics and data science teams to enable them to leverage further the tools you have built, as well as develop and automate best practices for handling data on our platform.
Our Tech Stack
Rich Data Environment:
- We useAirflowas a scheduler andData Build Tool(dbt-core) to craft efficient data pipelines with an extensively customized setup, leveraging the flexibility ofPython.
- UtilizingRedshift,Athena,andDuckDB, we interact with data seamlessly.
- Our Redshift setup utilizes a decentralized structure through data sharing, enhancing flexibility and scalability.
- The everyday tasks associated with data warehousing are effectively automated through the proficient application ofthe Spring BootCLI, thereby streamlining operations and ensuring operational efficiency.
- We utilize AI (LLM) tools extensively to enhance productivity. (Claude, Copilot, Codex, MCP Servers, Agentic systems, or tool of your choice).
Comprehensive Cloud and DevOps Integration:
- Our tech arsenal includes Terraform, Docker, and Jenkins, forming a robust foundation for cloud-based development and seamless DevOps integration.
Detailed Monitoring Capabilities:
- Monitoring is a top priority, and we employ ELK, Grafana, Looker, OpsGenie, and internally developed technologies to ensure real-time visibility into system performance and data workflows.
Data Ingestion Platform:
- Build and maintain Java microservices that collect user tracking data from our websites and mobile applications via REST APIs, using Kafka as the central message broker.
- Deploy applications on AWS EKS (Kubernetes), ensuring high availability, scalability, and efficient resource management.
- Develop Java/Kotlin SDKs that enable backend development teams to send user event data to Kafka and AWS Firehose seamlessly.
- Architect event-driven data collection systems processing millions of user interactions daily from across our digital properties on systems handling millions of events daily with sub-second latency requirements.
- Configure and optimize third-party ingestion tools (Airbyte, Fivetran) to extract data from external APIs, databases, and SaaS platforms.
Your role in this journey
- Play a key role in a high-performing cross-functional team with a strong focus on data products, velocity, quality and engineering culture.
- Engage comprehensively in software development - from ideation to release and monitoring.
- Architect, design, and optimize robust data pipelines to extract, transform, and load diverse datasets efficiently and reliably.
- Devote efforts to crafting practical development toolchains to empower other teams in the organization to build high-quality datasets.
- Shape engineering objectives in tandem with engineering managers.
- Passionate about the latest data technologies? You'll research new solutions, optimize our existing stack, and pioneer innovations that push our platform forward.
- Seek cost-efficient solutions while maintaining high standards of quality (SLA).
- Proactively identify opportunities anddrive initiatives to enhance data engineering processes, monitor pipeline performance, and identify opportunities for optimization and efficiency gains.
- Focus on the growth of your team by offering consistent and valuable feedback and recruiting new engineers.
- Apply best practices and continuously learn and share your thoughts with your peers.
- Support Data Analysts in migrating from Airflow PostgresOperator (Redshift) to DBT using Athena.
- Evaluating new implementations constantly with cost, productivity, and velocity in mind.
- Challenge the status quo, proposing solutions that will have a company impact.
Your backpack is filled with
- A bachelor's degree in Computer Science, a related technical field, or equivalent practical experience.
- Experience building and implementing Lakehouse architectures in AWS or other similar setups.
- Effectively build batch and streaming data pipelines using technologies such as Airflow, DBT, Redshift, Athena/Presto, Firehose, Spark, SQL databases, and similar technologies.
- Strong knowledge of how distributed systems work.
- DataOps knowledge (e.g., Infrastructure as Code, CI/CD for data pipelines, automated testing, monitoring
#J-18808-Ljbffr