Nejo Logo
Jobs finden
nach Anstellungsart

Finde Jobs nach Arbeitszeit

  • Geringfügige Jobs
  • Teilzeit Jobs
  • Lehrstellen
  • Praktikumsplätze
nach Stadt

Jobs in deiner Nähe finden

  • Jobs in Wien
  • Jobs in Graz
  • Jobs in Linz
  • Jobs in Salzburg
  • Jobs in Innsbruck
  • weitere Städte
nach Beruf

Erkunde Jobs nach Berufsfeld

  • Fahrer Jobs
  • IT Jobs
  • Feuerwehr Jobs
  • Hausmeister Jobs
  • Vertrieb Jobs
  • weitere Berufe
nach Erfahrungslevel

Jobs passend zu deiner Erfahrung

  • Quereinsteiger Jobs
  • Berufseinsteiger Jobs
  • Manager Jobs
nach Arbeitsweise

Wähle deine bevorzugte Arbeitsweise

  • Remote Jobs
  • Home Office Jobs
Studenten
Schüler
Blog
Jobs finden
nach Anstellungsart

Finde Jobs nach Arbeitszeit

  • Geringfügige Jobs
  • Teilzeit Jobs
  • Lehrstellen
  • Praktikumsplätze
nach Stadt

Jobs in deiner Nähe finden

  • Jobs in Wien
  • Jobs in Graz
  • Jobs in Linz
  • Jobs in Salzburg
  • Jobs in Innsbruck
  • weitere Städte
nach Beruf

Erkunde Jobs nach Berufsfeld

  • Fahrer Jobs
  • IT Jobs
  • Feuerwehr Jobs
  • Hausmeister Jobs
  • Vertrieb Jobs
  • weitere Berufe
nach Erfahrungslevel

Jobs passend zu deiner Erfahrung

  • Quereinsteiger Jobs
  • Berufseinsteiger Jobs
  • Manager Jobs
nach Arbeitsweise

Wähle deine bevorzugte Arbeitsweise

  • Remote Jobs
  • Home Office Jobs
StudentenSchülerBlogNejo LinkedIn

AI Research Engineer (Model Evaluation)(m/w/x)

Tether Operations Limited
Bern

You focus on developing evaluation frameworks and benchmarks for AI models, ensuring they align with business goals while analyzing data to improve performance and reliability throughout the model lifecycle.

Anforderungen

  • •Degree in Computer Science or related field
  • •PhD in NLP or Machine Learning preferred
  • •Experience in designing AI models
  • •Strong programming skills required
  • •Hands-on expertise in evaluation benchmarks
  • •Ability to conduct iterative experiments
  • •Experience collaborating with diverse teams

Deine Aufgaben

  • •Develop and deploy integrated evaluation frameworks.
  • •Define and track key performance indicators.
  • •Curate high-quality evaluation datasets.
  • •Design standardized benchmarks for model quality.
  • •Align evaluation metrics with business objectives.
  • •Present findings through dashboards and reports.
  • •Analyze evaluation data to identify bottlenecks.
  • •Propose optimizations for model performance.
  • •Conduct experiments to refine evaluation methodologies.
  • •Stay updated on emerging techniques and trends.

Original Beschreibung

## AI Research Engineer (Model Evaluation) ***About the job:*** As a member of our AI model team, you will drive innovation across the entire AI lifecycle by developing and implementing rigorous evaluation frameworks and benchmark methodologies for pre-training, post-training, and inference. Your work will focus on designing metrics and assessment strategies that ensure our models are highly responsive, efficient, and reliable across real-world applications. You will work on a wide spectrum of systems, from resource-efficient models designed for limited hardware environments to complex, multi-modal architectures that integrate text, images, and audio. We expect you to have deep expertise in advanced model architectures, pre-training and post-training practices, and inference evaluation frameworks. Adopting a hands-on, research-driven approach, you will develop, test, and implement novel evaluation strategies that rigorously track key performance indicators such as accuracy, latency, throughput, and memory footprint. Your evaluations will not only benchmark model performance at each stage, from the foundational pre-training phase to targeted post-training refinements and final inference but will also provide actionable insights. A key element of this role is collaborating with cross-functional teams including product management, engineering, and operations to share your evaluation findings and integrate stakeholder feedback. You will engineer robust evaluation pipelines and performance dashboards that serve as a common reference point for all stakeholders, ensuring that the insights drive continuous improvement in model deployment strategies. The ultimate goal is to set industry-leading standards for AI model quality and reliability, delivering scalable performance and tangible value in dynamic, real-world scenarios. **Responsibilities**: * Develop, test, and deploy integrated frameworks that rigorously assess models during pre-training, post-training, and inference. Define and track key performance indicators such as accuracy, loss metrics, latency, throughput, and memory footprint across diverse deployment scenarios. * Curate high-quality evaluation datasets and design standardized benchmarks to reliably measure model quality and robustness. Ensure that these benchmarks accurately reflect improvements achieved through both pre-training and post-training processes, and drive consistency in evaluation practices. * Engage with product management, engineering, data science, and operations teams to align evaluation metrics with business objectives. Present evaluation findings, actionable insights, and recommendations through comprehensive dashboards and reports that support decision-making across functions. * Systematically analyze evaluation data to identify and resolve bottlenecks across the model lifecycle. Propose and implement optimizations that enhance model performance, scalability, and resource utilization on resource-constrained platforms, ensuring efficient pre-training, post-training, and inference. * Conduct iterative experiments and empirical research to refine evaluation methodologies, staying abreast of emerging techniques and trends. Leverage insights to continuously enhance benchmarking practices and improve overall model reliability, ensuring that all stages of the model lifecycle deliver measurable value in real-world applications. ## Requirements * A degree in Computer Science or related field. Ideally PhD in NLP, Machine Learning, or a related field, complemented by a solid track record in AI R&D (with good publications in A\* conferences). * Demonstrated experience in designing and evaluating AI models at multiple stages from pre-training, post-training, and inference. You should be proficient in developing evaluation frameworks that rigorously assess accuracy, convergence, loss improvements, and overall model robustness, ensuring each stage of the AI lifecycle delivers measurable real-world value. * Strong programming skills and hands-on expertise in evaluation benchmarks and frameworks are essential. Familiarity with building, automating, and scaling complex evaluation and benchmarking pipelines, and experience with performance metrics: latency, throughput, and memory footprint. * Proven ability to conduct iterative experiments and empirical research that drive the continuous refinement of evaluation methodologies. You should be adept at staying abreast of emerging trends and techniques, leveraging insights to enhance benchmarking practices and model reliability. * Demonstrated experience collaborating with diverse teams such as product, engineering, and operations in order to align evaluation strategies with organizational goals. You must be skilled at translating technical findings into actionable insights for stakeholders and driving process improvements across the model development lifecycle.
Lade Jobdetails..
Über UnsProdukteKontaktImpressumDatenschutzNutzungsbedingungenCookie-Einstellungen
© 2025 Nejo
© 2025 nejo jobs