LeethubLeethub
JobsCompaniesBlog
Go to dashboard

Leethub

Curated tech jobs from FAANG and top companies worldwide.

Top Companies

  • Google Jobs
  • Meta Jobs
  • Amazon Jobs
  • Apple Jobs
  • Netflix Jobs
  • All Companies →

Job Categories

  • Software Engineering
  • Data, AI & Machine Learning
  • Product Management
  • Design & User Experience
  • Operations & Strategy
  • Remote Jobs
  • All Categories →

Browse by Type

  • Remote Jobs
  • Hybrid Jobs
  • Senior Positions
  • Entry Level
  • All Jobs →

Resources

  • Google Interview Guide
  • Salary Guide 2025
  • Salary Negotiation
  • LeetCode Study Plan
  • All Articles →

Company

  • Dashboard
  • Privacy Policy
  • Contact Us
© 2026 Leethub LLC. All rights reserved.
Home›Jobs›Qloo›Data Engineer
Qloo

About Qloo

AI-driven insights for personalized cultural recommendations

👥 21-100 employees📍 Bowery, New York, NY
B2BArtificial IntelligenceMachine LearningSaaSAPI

Key Highlights

  • Headquartered in Bowery, New York, NY
  • Serves clients like Starbucks with tailored solutions
  • Funding rounds in 2023 and 2024 for expansion
  • Offers a self-service interface for SMEs

Qloo is an AI-powered recommendation API based in Bowery, New York, that helps businesses understand cultural tastes globally. By providing AI-generated correlation data, Qloo enables companies like Starbucks to enhance customer experiences, such as creating tailored in-store music playlists. With a...

🎁 Benefits

Employees enjoy comprehensive health insurance, a retirement plan, and generous paid time off. Qloo also offers flexible work arrangements to promote ...

🌟 Culture

Qloo fosters a culture centered around leveraging AI to break down data silos and enhance personalization in consumer experiences. The company values ...

🌐 WebsiteAll 2 jobs →
Qloo

Data Engineer

Qloo • New York City

Posted 2w ago🏛️ On-SiteMid-LevelData engineer📍 New york
Apply Now →

Skills & Technologies

AWSAirflowApache SparkPythonAws glueEmrAthenaHiveElasticsearch

Job Description

About Us
At Qloo, we harness large-scale behavioral and catalog data to power recommendations and insights across entertainment, dining, travel, retail, and more. Our platform is built on a modern AWS data stack and supports analytics, APIs, and machine-learning models used by leading brands. We are looking for an experienced Data Engineer to help evolve and scale this platform.

Role Overview
As a Data Engineer at Qloo, you will design, build, and operate the pipelines that move data from external vendors, internal systems, and public sources into our S3-based data lake and downstream services. You’ll work across AWS Glue, EMR (Spark), Athena/Hive, and Airflow (MWAA) to ensure that our data is accurate, well-modeled, and efficiently accessible for analytics, indexing, and machine-learning workloads.

You should be comfortable owning end-to-end data flows, from ingestion and transformation to quality checks, monitoring, and performance tuning.

Responsibilities
- Design, develop, and maintain batch data pipelines using Python, Spark (EMR), and AWS Glue, loading data from S3, RDS, and external sources into Hive/Athena tables.
- Model datasets in our S3/Hive data lake to support analytics (Hex), API use cases, Elasticsearch indexes, and ML models.
- Implement and operate workflows in Airflow (MWAA), including dependency management, scheduling, retries, and alerting via Slack.
- Build robust data quality and validation checks (schema validation, freshness/volume checks, anomaly detection) and ensure issues are surfaced quickly with monitoring and alerts.
- Optimize jobs for cost and performance (partitioning, file formats, join strategies, proper use of EMR/Glue resources).
- Collaborate closely with data scientists, ML engineers, and application engineers to understand data requirements and design schemas and pipelines that serve multiple use cases.
- Contribute to internal tooling and shared libraries that make working with our data platform faster, safer, and more consistent.
- Document pipelines, datasets, and best practices so the broader team can easily understand and work with our data.

Qualifications
- Bachelor’s degree in Computer Science, Software Engineering, or a related field, or equivalent practical experience.
- Experience with Python and distributed data processing using Spark (PySpark) on EMR or a similar environment.
- Hands-on experience with core AWS data services, ideally including:
- S3 (data lake, partitioning, lifecycle management)
- AWS Glue (jobs, crawlers, catalogs)
- EMR or other managed Spark platforms
- Athena/Hive and SQL for querying large datasets
- Relational databases such as RDS (PostgreSQL/MySQL or similar)
- Experience building and operating workflows in Airflow (MWAA experience is a plus).
- Strong SQL skills and familiarity with data modeling concepts for analytics and APIs.
- Solid understanding of data quality practices (testing, validation frameworks, monitoring/observability).
- Comfortable working in a collaborative environment, managing multiple projects, and owning systems end-to-end.

We Offer
- Competitive salary and benefits package, including health insurance, retirement plan, and paid time off.
- The opportunity to shape a modern cloud-based data platform that powers real products and ML experiences.
- A collaborative, low-ego work environment where your ideas are valued and your contributions are visible.
- Flexible work arrangements (remote and hybrid options) and a healthy respect for work-life balance.

Interested in this role?

Apply now or save it for later. Get alerts for similar jobs at Qloo.

Apply Now →Get Job Alerts