Skip to end of banner
Go to start of banner

MLFlow Tracking

Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 2 Current »

MLFlow server tracking is utilized to measure different data and statistics to analyze how generative AI is performing. Here’re a few things currently being used.

Statistics currently measured:

Sure! Here are simple definitions for F1 Score, Precision, and Recall:

  1. Precision:

    • Definition: Precision is the ratio of correctly predicted positive observations to the total predicted positives. It tells us how many of the predicted positive cases were actually correct.

    • Formula:

      image-20240614-144211.png

    • Example: If a model predicts 10 positive cases and 7 of them are actually positive, the precision is

      image-20240614-144435.png

  2. Recall (also known as Sensitivity or True Positive Rate):

    • Definition: Recall is the ratio of correctly predicted positive observations to all the observations in the actual class. It measures how well the model can identify all relevant cases.

    • Formula:

      image-20240614-144305.png

    • Example: If there are 20 actual positive cases and the model correctly identifies 15 of them, the recall is

      image-20240614-144508.png

  3. F1 Score:

    • Definition: The F1 Score is the harmonic mean of precision and recall. It provides a single metric that balances both the precision and recall of the model.

    • Formula:

      image-20240614-144337.png

    • image-20240614-144541.png

These metrics are often used to evaluate the performance of classification models, especially in cases where the classes are imbalanced.

  • No labels