Logo

TrustWatch

Welcome to TrustWatch

TrustWatch is a community-driven platform where users share their experiences with AI models, along with a custom TrustScore that reflects their level of confidence in a model’s behavior. These reports are grouped into meaningful categories such as safety, fairness, and performance across diverse demographics and domains.

  • Each report is weighted based on how recent it is and how many other users validate it, ensuring that the TrustScores remain dynamic and relevant.
  • The TrustScore for a large language model (LLM) is calculated by aggregating multiple user-submitted reports over time. This continuous, longitudinal monitoring provides users and stakeholders with deep insight into how models perform in real-world conditions—highlighting both strengths and limitations before they are widely adopted.
  • Our goal is to make TrustWatch a central hub for evaluating the safety and fairness of AI systems.

Trending Trust Score

Model Analytics

Report Overview

Caption Trust Score AI Model Tags Report Summary Date Added Upvotes Downvotes Resolved Score Action

Report Summary

Loading...

Trust Score

0

Resolved Score

0

Current Upvotes

0

Current Downvotes

0

Tags

Loading...

Evidence

Loading...

Report Summary

Please provide the necessary details to update this report. You can modify the prompt, text evidence, and any supporting files. Ensure all changes are accurate before saving.

Note: Supported file formats are JPEG and PNG. Maximum file size is 10MB.

Evidence

No evidence found.

Version List

    Mark Report As Resolved

    Please provide the necessary details to mark this report. You can add a prompt, text evidence, and any supporting files. Ensure all fields are filled out accurately.

    Note: Supported file formats are JPEG and PNG. Maximum file size is 10MB.

    Evidence

    Version List

    Update Report

    CEAMLS serves as a resource for both developers and the public by promoting the development of models that are open for analysis and are linked with visual analytics so that everyone is more knowledgeable about the capabilities, limits, and biases that these models may possess. CEAMLS also analyzes new AI innovations to determine ways in which they could become problematic or damaged due to algorithmic bias, either actual or perceived.