Remote Data Jobs · Scala

Job listings

$190,800–$267,100/yr

  • Be the Analytics Engineering lead within the Sales and Marketing organization.
  • Be the data steward for Sales and Marketing: architect and improve the data collection.
  • Develop and maintain robust data pipelines and workflows for data ingestion and transformation.

Reddit is a community-driven platform built on shared interests and trust, fostering open and authentic conversations. With over 100,000 active communities and approximately 116 million daily active unique visitors, it serves as a major source of information on the internet.

  • Design and implement large-scale distributed data processing systems.
  • Build robust data pipelines and infrastructure that transform complex data into actionable insights.
  • Architect data lakes, warehouses, and real-time streaming platforms.

Cloudbeds is transforming hospitality with an intelligently designed platform that powers properties across 150 countries. With a completely remote team of 650+ employees across 40+ countries, they are solving challenges and redefining what's possible in hospitality.

  • Design, build, and maintain infrastructure for data ingestion, processing, and analysis.
  • Translate business requirements into technical solutions in collaboration with stakeholders.
  • Ensure data quality, integrity, and security throughout the data lifecycle.

Jobgether is a platform that connects job seekers with companies. They use AI-powered matching to ensure applications are reviewed quickly and fairly.

$110,572–$145,000/yr
US Unlimited PTO

  • Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and design data models and schemas that facilitate data analysis and reporting
  • Design, develop, and maintain scalable and efficient data pipelines and ETL processes to ingest, process, and transform large volumes of data from various sources into usable formats
  • Build and optimize data storage and processing systems, including data warehouses, data lakes, and big data platforms, using AWS services such as Amazon Redshift, AWS Glue, AWS EMR, AWS S3, and AWS Lambda, to enable efficient data retrieval and analysis

ATPCO is the world's primary source for air fare content. They hold over 200 million fares across 160 countries and the travel industry relies on their technology and data solutions. ATPCO believes in flexibility, trust, and a culture where your wellbeing comes first.

  • Design, build, and evolve scalable data pipelines and systems that ensure financial accuracy and integrity at scale.
  • Explore and apply Spotify’s Data and AI ecosystem to solve engineering problems and improve developer efficiency.
  • Partner closely with Finance and Audiobooks stakeholders to understand their needs and deliver financially impactful features.

Spotify transformed music listening forever when we launched in 2008. Today, they are the world’s most popular audio streaming subscription service.

US 4w PTO

  • Architect, design, and oversee delivery of end-to-end AI/ML solutions.
  • Lead cross-functional teams to implement robust ML platforms, pipelines, and applications.
  • Communicate the business value and ROI of AI/ML solutions to stakeholders.

Jobgether is using an AI-powered matching process to ensure your application is reviewed quickly, objectively, and fairly against the role's core requirements. The system identifies the top-fitting candidates, and this shortlist is then shared directly with the hiring company.

  • Lead and mentor a team of data engineers, fostering innovation, collaboration, and continuous improvement.
  • Design, implement, and optimize scalable data pipelines and ETL processes to meet evolving business needs.
  • Ensure data quality, governance, security, and compliance with industry standards and best practices.

Jobgether is a platform that connects job seekers with companies. They use an AI-powered matching process to ensure applications are reviewed quickly, objectively, and fairly against the role's core requirements.

$145,290–$185,000/yr
Unlimited PTO

  • Partner with data scientists and stakeholders to translate business and ML/AI use cases into scalable data architectures.
  • Design, develop, and maintain scalable and efficient data pipelines and ETL processes to ingest, process, and transform large data.
  • Build and optimize data storage and processing systems using AWS services to enable efficient data retrieval and analysis.

ATPCO is the world's primary source for air fare content, holding over 200 million fares across 160 countries. They provide technology and data solutions to the travel industry, helping millions of travelers reach their destinations efficiently. ATPCO believes in flexibility, trust, and a culture where your wellbeing comes first.

$85,000–$90,000/yr
US 4w PTO

  • Write and deploy crawling scripts to collect source data from the web
  • Write and run data transformers in Scala Spark to standardize bulk data sets
  • Write and run modules in Python to parse entity references and relationships from source data

Sayari is a risk intelligence provider equipping sectors with visibility into commercial relationships, delivering corporate and trade data from over 250 jurisdictions. Headquartered in Washington, D.C., its solutions are trusted globally and recognized for growth and workplace culture.

  • Design and engineer robust data pipelines using technologies like Databricks, Azure Data Factory, Apache Spark, and Delta Lake.
  • Craft healthcare data solutions - processing massive healthcare datasets, optimizing performance, and ensuring data is accurate and secure.
  • Communicate technical concepts to non-technical stakeholders, manage multiple priorities, and meet deadlines.

Gentiva offers compassionate care in the comfort of patients' homes as a national leader in hospice, palliative, home health care, and advanced illness management. They have nearly 600 locations and thousands of clinicians across 38 states, offering rewarding careers in a collaborative environment.