A different - worse - dataframe is presented after the dataframe is rebuilt to gain a consensus across labellers. Something is quite broken in the function compute_consensual_labels_and_sample_weights() - currently commented out but central to the training side of the tool.
I am unsure if the issue is "new", but previously models were training to high scores (~0.95), and now the score drops to ~0.6 if the function compute_consensual_labels_and_sample_weights is used. For now, this is just commented out.
A different - worse - dataframe is presented after the dataframe is rebuilt to gain a consensus across labellers. Something is quite broken in the function compute_consensual_labels_and_sample_weights() - currently commented out but central to the training side of the tool.
I am unsure if the issue is "new", but previously models were training to high scores (~0.95), and now the score drops to ~0.6 if the function compute_consensual_labels_and_sample_weights is used. For now, this is just commented out.