diff --git a/docs/cli.md b/docs/cli.md index b35595cc..0a73e1f3 100644 --- a/docs/cli.md +++ b/docs/cli.md @@ -111,7 +111,7 @@ sb deploy [--docker-image] | `--docker-username` | `None` | Docker registry username if authentication is needed. | | `--host-file` `-f` | `None` | Path to Ansible inventory host file. | | `--host-list` `-l` | `None` | Comma separated host list. | -| `--host-password` | `None` | Host password or key passphase if needed. | +| `--host-password` | `None` | Host password or key passphrase if needed. | | `--host-username` | `None` | Host username if needed. | | `--no-image-pull` | `False` | Skip pull and use local Docker image. | | `--output-dir` | `None` | Path to output directory, outputs/{datetime} will be used if not specified. | @@ -373,7 +373,7 @@ sb run [--config-file] | `--get-info` | `False` | Collect system info. | | `--host-file` `-f` | `None` | Path to Ansible inventory host file. | | `--host-list` `-l` | `None` | Comma separated host list. | -| `--host-password` | `None` | Host password or key passphase if needed. | +| `--host-password` | `None` | Host password or key passphrase if needed. | | `--host-username` | `None` | Host username if needed. | | `--no-docker` | `False` | Run on host directly without Docker. | | `--output-dir` | `None` | Path to output directory, outputs/{datetime} will be used if not specified. | diff --git a/docs/superbench-config.mdx b/docs/superbench-config.mdx index 340c7d61..b8ad058f 100644 --- a/docs/superbench-config.mdx +++ b/docs/superbench-config.mdx @@ -295,7 +295,7 @@ Enable current benchmark or not, can be overwritten by [`superbench.enable`](#su ### `timeout` -Set the timeout value in seconds, the benchmarking will stop early if timeout is triggerred. +Set the timeout value in seconds, the benchmarking will stop early if timeout is triggered. * default value: none @@ -336,16 +336,16 @@ A list of models to run, only supported in model-benchmark. Parameters for benchmark to use, varying for different benchmarks. -There have four common parameters for all benchmarks: -* run_count: how many times do user want to run this benchmark, default value is 1. +There are four common parameters for all benchmarks: +* run_count: how many times does user want to run this benchmark, default value is 1. * duration: the elapsed time of benchmark in seconds. It can work for all model-benchmark. But for micro-benchmark, benchmark authors should consume it by themselves. * log_raw_data: log raw data into file instead of saving it into result object, default value is `False`. Benchmarks who have large raw output may want to set it as `True`, such as `nccl-bw`/`rccl-bw`. * log_flushing: real-time log flushing, default value is `False`. -For Model-Benchmark, there have some parameters that can control the elapsed time. +For Model-Benchmark, there are some parameters that can control the elapsed time. * duration: the elapsed time of benchmark in seconds. -* num_warmup: the number of warmup step, should be positive integer. -* num_steps: the number of test step. +* num_warmup: the number of warmup steps, should be positive integer. +* num_steps: the number of test steps. If `duration > 0` and `num_steps > 0`, then benchmark will take the least as the elapsed time. Otherwise only one of them will take effect. @@ -429,7 +429,7 @@ while `proc_num: 8, node_num: null` will run 32-GPU distributed training on all Command prefix to use in the mode, in Python formatted string. -Available variables in formatted string includes: +Available variables in formatted string include: + `proc_rank` + `proc_num`