Source Job

$15–$40/hr
US

  • Delivering high-quality data and annotations for scenarios involving MACROHARD, and testing Computer Use Agents in digital environments.
  • Identifying subtle bugs, failure modes, and unexpected agent behaviors during testing sessions to help improve Computer Use models.
  • Assisting in designing and improving annotation tools tailored for MACROHARD data, agent evaluation, and QA workflows.

Data Collection Analytical Communication

20 jobs similar to MACROHARD Tutor

Jobs ranked by similarity.

Mexico

  • Set client QA strategies and adapt to scope/volume changes.
  • Run root-cause analyses; drive CAPA plans with owners, timelines, and effectiveness checks.
  • Plan training & certification for raters/annotators and coordinators; track completion and impact.

Welo Data provides high-quality, ethically sourced, relevant, diverse, and scalable datasets to technology companies to supercharge their AI models. As a Welocalize brand, WeloData leverages over 25 years of experience and brings together a curated global community of over 500,000 AI training and domain experts.

US

  • Help build the next generation of fraud-detection AI.
  • Fast and easy task using the Appen Mobile app.
  • Flexible participation — record between 1 to 10 short videos.

Mannequins Video is working on Project Orpheus. They are seeking contributors to participate in a short video data collection task supporting fraud-detection AI development.

Kenya

  • Record short videos of unique mannequins.
  • Complete up to 10 short mannequin clips per participant.
  • Use the Appen Mobile app to submit recordings.

CrowdGen is part of Appen and focuses on providing data solutions. The company utilizes a community of independent contractors to contribute to various AI development projects, offering flexible, project-based opportunities.

Global

  • Completing AI training tasks such as analyzing, editing, and writing in Mandarin
  • Judging the performance of AI in performing Mandarin prompts
  • Improving cutting-edge AI models

Prolific is building the biggest pool of quality human data in the world and is not just another player in the AI space. Over 35,000 AI developers, researchers, and organizations use Prolific to gather data from paid study participants with a wide variety of experiences, knowledge, and skills.

Global

  • Completing AI training tasks such as analyzing, editing, and writing computer science–related content
  • Judging the performance of AI on programming, algorithms, data structures, and systems prompts
  • Improving cutting-edge AI models using your understanding of software engineering and computational thinking

Prolific is building the biggest pool of quality human data in the world. Over 35,000 AI developers, researchers, and organizations use Prolific to gather data from paid study participants with a wide variety of experiences, knowledge, and skills.

$30–$75/hr
US

  • Train and refine Grok for voice interactions across diverse languages.
  • Curate and annotate high-quality audio data to enhance Grok's global accessibility.
  • Collaborate with technical staff to improve AI's handling of multilingual audio nuances.

xAI aims to create AI systems that understand the universe and aid humanity. The team is small, motivated, and focused on engineering excellence with a flat organizational structure, expecting all employees to be hands-on.

Global

  • Evaluate AI model outputs related to your field.
  • Assess content relevant to your area of expertise.
  • Deliver clear feedback to improve the model's comprehension.

Handshake is recruiting College Career/Technical Education Professors to contribute to an hourly, temporary AI research project. In this program, you’ll leverage your professional experience to evaluate what AI models produce in your field.

$22–$22/hr
US

  • Annotate surgical workflows and objects within operating room videos using Apella’s internal tools.
  • Build your understanding of machine learning systems and develop a knowledge of surgical environments.
  • Refine workflows and provide feedback on ways to improve Apella's tools and processes.

Apella is applying computer vision, machine learning and natural language processing to improve surgery. They build applications to enable surgeons, nurses, and hospital administrators to deliver the highest quality care.

Europe

  • Evaluate AI-generated responses for accuracy, grammar, and cultural relevance.
  • Identify issues and provide refined, high-quality rewritten responses.
  • Create natural prompts and responses in Spanish to improve conversational datasets.

Welo Data, part of Welocalize, is a global AI data company with 500,000+ contributors delivering high-quality, ethical data to train the world’s most advanced AI systems. They're building smarter, more human AI with a diverse community in 100+ countries.

  • Leverage professional experience to evaluate AI models' output in your field.
  • Assess content and deliver feedback to strengthen the model’s understanding.
  • Work independently from anywhere, with flexible hours and no minimum commitment.

Handshake is a recruiting platform. They connect students and recent graduates with employers.

Global

  • Evaluate AI model outputs related to your field.
  • Assess content and provide feedback to strengthen the model’s understanding.
  • Develop prompts for AI models reflecting your field and evaluate responses.

Handshake is recruiting Medical Secretaries and Administrative Assistant Professionals to contribute to an hourly, temporary AI research project. The Handshake AI opportunity runs year-round, with project opportunities opening periodically across different areas of expertise.

$120,000–$150,000/yr
US

  • Lead and manage AI program execution and data creation projects.
  • Translate complex objectives into clear milestones and measurable impact.
  • Conduct regular spot checks and quality reviews of output to ensure data meets client standards.

SuperAnnotate is a fast-growing, Series B startup revolutionizing the field of AI-data Infrastructure. They specialize in providing cutting-edge data pipeline solutions for Machine Learning, LLM, and GenAI solutions to large enterprise clients, helping them leverage the power of AI to transform their businesses.

Global

  • Evaluate AI model outputs in your field.
  • Assess content related to your field of work.
  • Deliver feedback to strengthen AI understanding.

Handshake is connecting students, new grads, and young professionals with job opportunities. They aim to close the opportunity gap and ensure everyone has equal access to meaningful employment.

  • Design, build, and optimize high-performance systems in Python supporting AI data pipelines and evaluation workflows.
  • Develop full-stack tooling and backend services for large-scale data annotation, validation, and quality control.
  • Improve reliability, performance, and safety across existing Python codebases.

Alignerr connects top technical experts with leading AI labs to build, evaluate, and improve next-generation models. They work on real production systems and high-impact research workflows across data, tooling, and infrastructure.

Indonesia

  • Review short, pre-segmented datasets.
  • Evaluate model-generated replies based on Tone or Fluency .
  • Read a user prompt and two model replies, then rate each using a five-point scale.

CrowdGen, by Appen, focuses on AI response evaluation. They are looking for native Javanese speakers to contribute to a multilingual AI response evaluation project where you review large language model outputs.

Global

  • Native or near-native fluency in Central Khmer.
  • Based in: Cambodia, Thailand.
  • Comfortable with digital tools.

Welo Data, part of Welocalize, is a global AI data company with 500,000+ contributors delivering high-quality, ethical data to train the world’s most advanced AI systems. They’re building smarter, more human AI with a diverse community in 100+ countries.

Global

  • Evaluate AI models' output in occupational therapy.
  • Assess content related to the occupational therapy field.
  • Provide clear feedback to improve AI understanding.

Handshake connects students with early talent recruiting. They provide opportunity to evaluate what AI models produce and deliver feedback that strengthens the model’s understanding of workplace tasks and language.