import numpy as np import pandas as pd import plotly.express as px from plotly.graph_objs import Figure from src.display.utils import BENCHMARK_COLS, AutoEvalColumn, Task, Tasks # from src.display.utils import human_baseline_row as HUMAN_BASELINE from src.leaderboard.filter_models import FLAGGED_MODELS def create_scores_df(results_df: list[dict]) -> pd.DataFrame: """ Generates a DataFrame containing the maximum scores until each date. :param results_df: A DataFrame containing result information including metric scores and dates. :return: A new DataFrame containing the maximum scores until each date for every metric. """ # Step 1: Ensure 'date' is in datetime format and sort the DataFrame by it results_df["date"] = pd.to_datetime(results_df["date"], format="mixed", utc=True) results_df.sort_values(by="date", inplace=True) # Step 2: Initialize the scores dictionary scores = {k: [] for k in BENCHMARK_COLS + [AutoEvalColumn.average.name]} # Step 3: Iterate over the rows of the DataFrame and update the scores dictionary for task in [t.value for t in Tasks] + [Task("Average", "avg", AutoEvalColumn.average.name)]: current_max = 0 last_date = "" column = task.col_name for _, row in results_df.iterrows(): current_model = row[AutoEvalColumn.fullname.name] # We ignore models that are flagged/no longer on the hub/not finished to_ignore = ( not row[AutoEvalColumn.still_on_hub.name] or not row[AutoEvalColumn.not_flagged.name] or current_model in FLAGGED_MODELS ) if to_ignore: continue current_date = row[AutoEvalColumn.date.name] current_score = row[task.col_name] if current_score > current_max: if current_date == last_date and len(scores[column]) > 0: scores[column][-1] = {"model": current_model, "date": current_date, "score": current_score} else: scores[column].append({"model": current_model, "date": current_date, "score": current_score}) current_max = current_score last_date = current_date # Step 4: Return all dictionaries as DataFrames return {k: pd.DataFrame(v) for k, v in scores.items()} def create_plot_df(scores_df: dict[str : pd.DataFrame]) -> pd.DataFrame: """ Transforms the scores DataFrame into a new format suitable for plotting. :param scores_df: A DataFrame containing metric scores and dates. :return: A new DataFrame reshaped for plotting purposes. """ # Initialize the list to store DataFrames dfs = [] # Iterate over the cols and create a new DataFrame for each column for col in BENCHMARK_COLS + [AutoEvalColumn.average.name]: d = scores_df[col].reset_index(drop=True) d["task"] = col dfs.append(d) # Concatenate all the created DataFrames concat_df = pd.concat(dfs, ignore_index=True) # # Sort values by 'date' # concat_df.sort_values(by="date", inplace=True) # concat_df.reset_index(drop=True, inplace=True) # return concat_df def create_metric_plot_obj(df: pd.DataFrame, metrics: list[str], title: str) -> Figure: """ Create a Plotly figure object with lines representing different metrics and horizontal dotted lines representing human baselines. :param df: The DataFrame containing the metric values, names, and dates. :param metrics: A list of strings representing the names of the metrics to be included in the plot. :param title: A string representing the title of the plot. :return: A Plotly figure object with lines representing metrics and horizontal dotted lines representing human baselines. """ # Filter the DataFrame based on the specified metrics df = df[df["task"].isin(metrics)] # Filter the human baselines based on the specified metrics filtered_human_baselines = {k: v for k, v in HUMAN_BASELINE.items() if k in metrics} # Create a line figure using plotly express with specified markers and custom data fig = px.line( df, x="date", y="score", color="task", markers=True, custom_data=["task", "score", "model"], title=title, ) # Update hovertemplate for better hover interaction experience fig.update_traces( hovertemplate="
".join( [ "Model Name: %{customdata[2]}", "Metric Name: %{customdata[0]}", "Date: %{x}", "Metric Value: %{y}", ] ) ) # Update the range of the y-axis fig.update_layout(yaxis_range=[0, 100]) # Create a dictionary to hold the color mapping for each metric metric_color_mapping = {} # Map each metric name to its color in the figure for trace in fig.data: metric_color_mapping[trace.name] = trace.line.color # Iterate over filtered human baselines and add horizontal lines to the figure for metric, value in filtered_human_baselines.items(): color = metric_color_mapping.get(metric, "blue") # Retrieve color from mapping; default to blue if not found location = "top left" if metric == "HellaSwag" else "bottom left" # Set annotation position # Add horizontal line with matched color and positioned annotation fig.add_hline( y=value, line_dash="dot", annotation_text=f"{metric} human baseline", annotation_position=location, annotation_font_size=10, annotation_font_color=color, line_color=color, ) return fig # Example Usage: # human_baselines dictionary is defined. # chart = create_metric_plot_obj(scores_df, ["ARC", "HellaSwag", "MMLU", "TruthfulQA"], human_baselines, "Graph Title")