forked from EleutherAI/gpt-neox
-
Notifications
You must be signed in to change notification settings - Fork 3
/
Copy pathmagma_setup.yml
54 lines (48 loc) · 1.66 KB
/
magma_setup.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
# Suggested data paths when using GPT-NeoX locally
{
"train-data-paths": [
"/gpfs/alpine/csc499/proj-shared/LAION-400m-webdataset/data/{00000..40000}.tar",
],
"train-data-weights": [
1.0,
],
"valid-data-paths": ["/gpfs/alpine/csc499/proj-shared/LAION-400m-webdataset/data/{40001..41000}.tar",],
"valid-data-weights": [
1.0,
],
"test-data-paths": ["/gpfs/alpine/csc499/proj-shared/LAION-400m-webdataset/data/{41000..41455}.tar",],
"test-data-weights": [
1.0,
],
"tokenizer-type": "HFTokenizer",
"vocab-file": "/gpfs/alpine/csc499/scratch/$whoami/20B_tokenizer.json",
"save": "/gpfs/alpine/csc499/scratch/$whoami/robin/checkpoints/magma_410M_testrun5",
"load": "/gpfs/alpine/csc499/proj-shared/pythia_magma_checkpoints/mp1_pp1/pythia",
# "load": "/gpfs/alpine/csc499/scratch/kshitijkg/magma/gpt-neox/checkpoints/magma_410M_testrun5",
"load_clip": "/gpfs/alpine/csc499/proj-shared/magma/models/clip",
"checkpoint_validation_with_forward_pass": False,
"tensorboard-dir": "/gpfs/alpine/csc499/scratch/$whoami/robin/tensorboard/magma_410M_testrun5",
"log-dir": "/gpfs/alpine/csc499/scratch/$whoami/robin/logs/magma_410M_testrun5",
"use_wandb": False,
"wandb_host": "https://api.wandb.ai",
"wandb_project": "magma_410M_testrun5",
"launcher": "jsrun",
"deepspeed_jsrun": true,
"finetune": true,
"num_workers": 0,
"optimizer": {
"type": "Adam",
"params": {
"lr": 0.0008,
"betas": [0.9, 0.95],
"eps": 1.0e-8,
}
},
"min_lr": 0.00008,
# LLAMA Config
# batch / data settings
"train_batch_size": 24,
"train_micro_batch_size_per_gpu": 4,
"data-impl": "mmap",
"split": "949,50,1",
}