You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+7-6Lines changed: 7 additions & 6 deletions
Original file line number
Diff line number
Diff line change
@@ -18,12 +18,13 @@ To quickly learn how to run cleanlab on your own data, first check out the [quic
18
18
| 8 |[huggingface_keras_imdb](huggingface_keras_imdb/huggingface_keras_imdb.ipynb)| CleanLearning for text classification with Keras Model + pretrained BERT backbone and Tensorflow Dataset. |
19
19
| 9 |[fasttext_amazon_reviews](fasttext_amazon_reviews/fasttext_amazon_reviews.ipynb)| Finding label errors in Amazon Reviews text dataset using a cleanlab-compatible [FastText model](https://github.com/cleanlab/cleanlab/blob/master/cleanlab/models/fasttext.py). |
20
20
| 10 |[multiannotator_cifar10](multiannotator_cifar10/multiannotator_cifar10.ipynb)| Iteratively improve consensus labels and trained classifier from data labeled by multiple annotators. |
21
-
| 11 |[active_learning_multiannotator](active_learning_multiannotator/active_learning.ipynb)| Improve model performance by iteratively collecting additional labels from annotators. This active learning pipeline allows for examples labeled in batches by multiple annotators. |
22
-
| 12 |[outlier_detection_cifar10](outlier_detection_cifar10/outlier_detection_cifar10.ipynb)| Train AutoML for image classification and use it to detect out-of-distribution images. |
23
-
| 13 |[multilabel_classification](multilabel_classification/image_tagging.ipynb)| Find label errors in an image tagging dataset ([CelebA](https://mmlab.ie.cuhk.edu.hk/projects/CelebA.html)) using a [Pytorch model](multilabel_classification/pytorch_network_training.ipynb) you can easily train for multi-label classification. |
24
-
| 14 |[entity_recognition](entity_recognition/entity_recognition_training.ipynb)| Train Transformer model for Named Entity Recognition and produce out-of-sample `pred_probs` for **cleanlab.token_classification**. |
25
-
| 15 |[transformer_sklearn](transformer_sklearn/transformer_sklearn.ipynb)| How to use `KerasWrapperModel` to make any Keras model sklearn-compatible, demonstrated here for a BERT Transformer. |
26
-
| 16 |[cnn_coteaching_cifar10](cnn_coteaching_cifar10)| Train a [Convolutional Neural Network](https://github.com/cleanlab/cleanlab/blob/master/cleanlab/experimental/cifar_cnn.py) on noisily labeled Cifar10 image data using cleanlab with [coteaching](https://github.com/cleanlab/cleanlab/blob/master/cleanlab/experimental/coteaching.py). |
21
+
| 11 |[active_learning_multiannotator](active_learning_multiannotator/active_learning.ipynb)| Improve a classifier model by iteratively collecting additional labels from data annotators. This active learning pipeline considers data labeled in batches by multiple (imperfect) annotators. |
22
+
| 12 |[active_learning_single_annotator](active_learning_single_annotator/active_learning_single_annotator.ipynb)| Improve a classifier model by iteratively labeling batches of currently-unlabeled data. This demonstrates a standard active learning pipeline with *at most one label* collected for each example (unlike our multi-annotator active learning notebook which allows re-labeling). |
23
+
| 13 |[outlier_detection_cifar10](outlier_detection_cifar10/outlier_detection_cifar10.ipynb)| Train AutoML for image classification and use it to detect out-of-distribution images. |
24
+
| 14 |[multilabel_classification](multilabel_classification/image_tagging.ipynb)| Find label errors in an image tagging dataset ([CelebA](https://mmlab.ie.cuhk.edu.hk/projects/CelebA.html)) using a [Pytorch model](multilabel_classification/pytorch_network_training.ipynb) you can easily train for multi-label classification. |
25
+
| 15 |[entity_recognition](entity_recognition/entity_recognition_training.ipynb)| Train Transformer model for Named Entity Recognition and produce out-of-sample `pred_probs` for **cleanlab.token_classification**. |
26
+
| 16 |[transformer_sklearn](transformer_sklearn/transformer_sklearn.ipynb)| How to use `KerasWrapperModel` to make any Keras model sklearn-compatible, demonstrated here for a BERT Transformer. |
27
+
| 17 |[cnn_coteaching_cifar10](cnn_coteaching_cifar10)| Train a [Convolutional Neural Network](https://github.com/cleanlab/cleanlab/blob/master/cleanlab/experimental/cifar_cnn.py) on noisily labeled Cifar10 image data using cleanlab with [coteaching](https://github.com/cleanlab/cleanlab/blob/master/cleanlab/experimental/coteaching.py). |
0 commit comments