|
37 | 37 | "python": "python3", |
38 | 38 | "module": "MaxText.decode", |
39 | 39 | "args": ["src/MaxText/configs/base.yml", |
40 | | - "run_name=runner_$(date +%Y-%m-%d-%H-%M)", |
| 40 | + "run_name=runner_$(date +%Y-%m-%d-%H-%M)", |
41 | 41 | "base_output_directory=gs://test-maxtext-output", |
42 | 42 | "dataset_path=gs://test-maxtext-dataset", |
43 | 43 | "steps=2", |
|
53 | 53 | "python": "python3", |
54 | 54 | "module": "MaxText.train", |
55 | 55 | "args": ["src/MaxText/configs/base.yml", |
56 | | - "run_name=runner_$(date +%Y-%m-%d-%H-%M)", |
| 56 | + "run_name=runner_$(date +%Y-%m-%d-%H-%M)", |
57 | 57 | "base_output_directory=gs://test-maxtext-output", |
58 | 58 | "dataset_path=gs://test-maxtext-dataset", |
59 | 59 | "steps=2", |
|
66 | 66 | "console": "integratedTerminal", |
67 | 67 | "justMyCode": false, |
68 | 68 | "python": "python3", |
69 | | - "module": "MaxText.inference_microbenchmark", |
| 69 | + "module": "maxtext.inference_microbenchmark", |
70 | 70 | "args": [ |
71 | 71 | "src/MaxText/configs/base.yml", |
72 | 72 | "model_name=llama2-7b", |
|
82 | 82 | "inference_microbenchmark_prefill_lengths=32,64,128,256,512,1024", |
83 | 83 | "inference_microbenchmark_stages=generate", |
84 | 84 | "inference_microbenchmark_loop_iters=1", |
85 | | - "run_name=runner_$(date +%Y-%m-%d-%H-%M)", |
| 85 | + "run_name=runner_$(date +%Y-%m-%d-%H-%M)", |
86 | 86 | "base_output_directory=gs://test-maxtext-output", |
87 | 87 | "prefill_cache_axis_order=0,2,1,3", |
88 | 88 | "ar_cache_axis_order=0,2,1,3", |
|
0 commit comments