Multi-Dataset Validation (LM-Loss/Perplexity)#178
Conversation
|
@jlamypoirier What do you think of such approach? |
|
I have implemented it so that the config will look something like this: data:
datasets:
Training: # training dataset, hardcoded name
type: memmap
path: some_path1
Test: # test dataset, hardcoded name
type: memmap
path: some_path2
validation_dataset_name1: # validation dataset, any name
type: memmap
path: some_path3
validation_dataset_name2: # validation dataset, any name
type: memmap
path: some_path4
training:
training_iters: 2
test_iters: 2
validation:
validation_dataset_name1:
interval: 2
iterations: 1
validation_dataset_name2:
interval: 2
iterations: 1@jlamypoirier, have I got it right? |
|
Maybe to add |
Yes
I don't think it's worth it, but feel free to fix the capitalization if you find a way to do it without breaking backward comparibility |
…heir usage config to EvaluationConfig
|
I have changed |
jlamypoirier
left a comment
There was a problem hiding this comment.
Almost ready to merge, some minor issues and suggestions
|
Guys can we get to a conclusion here please? I’d like us to merge this by tomorrow end of day, at the latest. Thanks |
jlamypoirier
left a comment
There was a problem hiding this comment.
@bigximik Adjusted the code, ready to merge if it works for you.
tscholak
left a comment
There was a problem hiding this comment.
thank you both, looks good to me!
|
I have updated the instruction fine-tuning documentation that came from the merge but created a separate issue for moving the check of used dataset definitions to |

✨ Description
Closes #65
🔍 Type of change
Select all that apply:
📝 Changes
List the key changes introduced in this PR:
training.validationconfig field is a dictionary of validation dataset names and their application parameters.✅ Checklist
Make sure the following tasks are completed before submitting the PR:
General
Testing