You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -17,7 +17,7 @@ The VideoFrameDataset class (an implementation of `torch.utils.data.Dataset`) se
17
17
video dataset to have a certain structure on disk and expects a .txt annotation file that enumerates each video sample. Details on this
18
18
can be found below. Pre-made annotation files and preparation scripts are also provided for [Kinetics 400](https://github.com/cvdfoundation/kinetics-dataset), [Something Something V2](https://20bn.com/datasets/something-something) and [Epic Kitchens 100](https://epic-kitchens.github.io/2021).
19
19
2) Efficiently because the video loading pipeline that this class implements is very fast. This minimizes GPU waiting time during training by eliminating CPU input bottlenecks that can slow down training time by several folds.
20
-
3) Effectively because the implemented sampling strategy for video frames is very practical. Video training using the entire sequence of
20
+
3) Effectively because the implemented sampling strategy for video frames is very representative. Video training using the entire sequence of
21
21
video frames (often several hundred) is too memory and compute intense. Therefore, this implementation samples frames evenly from the video (sparse temporal sampling)
22
22
so that the loaded frames represent every part of the video, with support for arbitrary and differing video lengths within the same dataset.
23
23
This approach has shown to be very effective and is taken from
0 commit comments