062bcb8751
This pr is auto merged as it contains a mandatory file and is opened for more than 10 days. |
||
---|---|---|
.vscode | ||
bench | ||
.gitignore | ||
BuildLinux.cmd | ||
BuildWindows.cmd | ||
Dockerfile | ||
LICENSE | ||
NOTICE.md | ||
README.md | ||
SECURITY.md | ||
full_config.yaml | ||
go.mod | ||
go.sum | ||
labench.yaml | ||
main.go | ||
web_requester.go |
README.md
Introduction
LaBench (for LAtency BENCHmark) is a tool that measures latency percentiles of HTTP GET or POST requests under very even and steady load.
The main feature and distinction of this tool is that (unlike many other benchmarking tools) it dictates request rate to the server and tries to maintain that rate very evenly even when server is experiencing slowdowns and hiccups. While other tools would usually back off and let the server to recover (see Coordinated Omission Problem for more details).
The main difference from wrk2 tool is very even load generated by LaBench.
Quick-Start Guide
- Copy or compile LaBench binary (there are both Windows and Linux executables). Windows version has more precise clock.
- Modify
labench.yaml
to meet your needs, most basic params should be self-explanatory. For the full list of supported parameters look atfull_config.yaml
. - Run the benchmark by simply running labench (you can also specify .yaml file on command line, but labench.yaml is used by default).
- BEFORE looking at the latency results check the following things in the tool output:
- TimelyTicks percentage. If it's less than say 99.9% then you need to increase number of Clients in yaml config. It's very realistic to keep it at 100%.
- TimelySends percentage. If it's less than say 99.9% then you need a beefier machine to run the test. It's very realistic to keep it at 100%.
- Number of errors returned by the server (non-200 responses). Some small percentage is OK, but they are not accounted for in latency results.
- Throughput reported in last line. If should be close to the value RequestRatePerSec in your .yaml config.
- If ANY of the above is not satisfied then the run was not valid and there is no point in looking at the latency results produced, so fix and re-run.
- The measurement results (latency percentiles) are placed in
out\res.hgrm
file. You can open it in Excel or go to http://hdrhistogram.github.io/HdrHistogram/plotFiles.html to plot it. - Note that plotted results have logarithmic X axis (i.e. the distance between 99% and 99.9% is the same as the distance between 99.9% and 99.99%).
Contributing
This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.microsoft.com.
When you submit a pull request, a CLA-bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., label, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.