• Apple Inc. (San Francisco, CA)
    …Learning Research Engineer to design and develop safe AI benchmarking methodologies. This role involves collaboration with various teams to implement responsible ... research background, proficiency in Python, and experience in AI/ ML model evaluation. The position offers competitive compensation, stock option opportunities,… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Apple Inc. (San Francisco, CA)
    …models. Description Our team, part of Apple Services Engineering, is looking for an ML Research Engineer to lead the design and continuous development of ... automated safety benchmarking methodologies. In this role, you...research background in empirical evaluation, experimental design, or benchmarking Strong proficiency in Python (pandas, NumPy, Jupyter, PyTorch,… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Rust codebases. Collaborate with data, research , and engineering teams to support model training and evaluation workflows. Identify ... with data annotation, data quality, or evaluation systems. Familiarity with AI/ ML workflows, model training, or benchmarking pipelines. Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Python codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Scale AI, Inc. (San Francisco, CA)
    …learning, especially in generative AI, evaluation, or oversight. Significant experience leading ML research in academia or industry. Strong written and verbal ... research team is shaping the next generation of safety science for frontier AI models and works at...research at top‑tier venues and contribute to open‑source benchmarking initiatives. Remain deeply engaged with the research more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Scale AI, Inc. (San Francisco, CA)
    …of frontier AI research , product, and go-to-market. You'll partner closely with ML teams in high-stakes meetings, scope and pitch solutions to top AI labs, and ... technical execution across accounts. You have Deep technical background in applied AI/ ML : 5-10+ years in research , engineering, solutions engineering, or… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …supporting AI training and benchmarking Improve performance, scalability, and safety across existing codebases Collaborate with data, research , and ... Experience with data annotation, data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing C++ codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Python codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing C++ codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Rust codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing C++ codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing C# codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... experience with data annotation, data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Rust codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing C++ codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... experience with data annotation, data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Rust codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... experience with data annotation, data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing C++ codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... experience with data annotation, data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Rust codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... experience with data annotation, data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Rust codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source
  • Labelbox (San Francisco, CA)
    …and safety across existing Python codebases Collaborate with data, research , and engineering teams to support model training and evaluation workflows Identify ... with data annotation , data quality, or evaluation systems Familiarity with AI/ ML workflows, model training, or benchmarking pipelines Experience with… more
    job goal (01/13/26)
    - Save Job - Related Jobs - Block Source