-
Notifications
You must be signed in to change notification settings - Fork 30
Open
Labels
Description
after discussion, look like epochs are not really needed, we can directly use batches. so going from "round -> epoch -> batch" to have "round -> batch". that would give more direct control on
- rework datasets to generate samples (ie batches) of the loaded data
- randomized over the whole loaded data as to avoid skewing training
- remove limitation in gpt-tfjs of running at most five batches
- superseed by number of batch per round
- remove
TrainingInformation.epochs
&EpochLogs
- in
Task
, userounds
as the top level count of run, thenbatchesPerRound
(renamed fromroundDuration
) - flatten generators from
{Trainer,Model}.fit