2009-02-02 15:02:31 +03:00
|
|
|
Queue sysfs files
|
|
|
|
=================
|
|
|
|
|
|
|
|
This text file will detail the queue files that are located in the sysfs tree
|
|
|
|
for each block device. Note that stacked devices typically do not export
|
|
|
|
any settings, since their queue merely functions are a remapping target.
|
|
|
|
These files are the ones found in the /sys/block/xxx/queue/ directory.
|
|
|
|
|
|
|
|
Files denoted with a RO postfix are readonly and the RW postfix means
|
|
|
|
read-write.
|
|
|
|
|
2012-08-09 17:28:05 +04:00
|
|
|
add_random (RW)
|
|
|
|
----------------
|
2014-08-26 14:33:20 +04:00
|
|
|
This file allows to turn off the disk entropy contribution. Default
|
2012-08-09 17:28:05 +04:00
|
|
|
value of this file is '1'(on).
|
|
|
|
|
2016-08-09 21:01:30 +03:00
|
|
|
dax (RO)
|
|
|
|
--------
|
|
|
|
This file indicates whether the device supports Direct Access (DAX),
|
|
|
|
used by CPU-addressable storage to bypass the pagecache. It shows '1'
|
|
|
|
if true, '0' if not.
|
|
|
|
|
2012-08-09 17:28:05 +04:00
|
|
|
discard_granularity (RO)
|
|
|
|
-----------------------
|
|
|
|
This shows the size of internal allocation of the device in bytes, if
|
|
|
|
reported by the device. A value of '0' means device does not support
|
|
|
|
the discard functionality.
|
|
|
|
|
2015-07-16 18:14:26 +03:00
|
|
|
discard_max_hw_bytes (RO)
|
2012-08-09 17:28:05 +04:00
|
|
|
----------------------
|
|
|
|
Devices that support discard functionality may have internal limits on
|
|
|
|
the number of bytes that can be trimmed or unmapped in a single operation.
|
|
|
|
The discard_max_bytes parameter is set by the device driver to the maximum
|
|
|
|
number of bytes that can be discarded in a single operation. Discard
|
|
|
|
requests issued to the device must not exceed this limit. A discard_max_bytes
|
|
|
|
value of 0 means that the device does not support discard functionality.
|
|
|
|
|
2015-07-16 18:14:26 +03:00
|
|
|
discard_max_bytes (RW)
|
|
|
|
----------------------
|
|
|
|
While discard_max_hw_bytes is the hardware limit for the device, this
|
|
|
|
setting is the software limit. Some devices exhibit large latencies when
|
|
|
|
large discards are issued, setting this value lower will make Linux issue
|
|
|
|
smaller discards and potentially help reduce latencies induced by large
|
|
|
|
discard operations.
|
|
|
|
|
2012-08-09 17:28:05 +04:00
|
|
|
discard_zeroes_data (RO)
|
|
|
|
------------------------
|
|
|
|
When read, this file will show if the discarded block are zeroed by the
|
|
|
|
device or not. If its value is '1' the blocks are zeroed otherwise not.
|
|
|
|
|
2009-02-02 15:02:31 +03:00
|
|
|
hw_sector_size (RO)
|
|
|
|
-------------------
|
|
|
|
This is the hardware sector size of the device, in bytes.
|
|
|
|
|
2016-08-09 21:01:30 +03:00
|
|
|
io_poll (RW)
|
|
|
|
------------
|
|
|
|
When read, this file shows the total number of block IO polls and how
|
|
|
|
many returned success. Writing '0' to this file will disable polling
|
|
|
|
for this device. Writing any non-zero value will enable this feature.
|
|
|
|
|
2012-08-09 17:28:05 +04:00
|
|
|
iostats (RW)
|
|
|
|
-------------
|
|
|
|
This file is used to control (on/off) the iostats accounting of the
|
|
|
|
disk.
|
|
|
|
|
|
|
|
logical_block_size (RO)
|
|
|
|
-----------------------
|
2016-06-28 23:10:57 +03:00
|
|
|
This is the logical block size of the device, in bytes.
|
2012-08-09 17:28:05 +04:00
|
|
|
|
2009-02-02 15:02:31 +03:00
|
|
|
max_hw_sectors_kb (RO)
|
|
|
|
----------------------
|
|
|
|
This is the maximum number of kilobytes supported in a single data transfer.
|
|
|
|
|
2012-08-09 17:28:05 +04:00
|
|
|
max_integrity_segments (RO)
|
|
|
|
---------------------------
|
|
|
|
When read, this file shows the max limit of integrity segments as
|
|
|
|
set by block layer which a hardware controller can handle.
|
|
|
|
|
2009-02-02 15:02:31 +03:00
|
|
|
max_sectors_kb (RW)
|
|
|
|
-------------------
|
|
|
|
This is the maximum number of kilobytes that the block layer will allow
|
|
|
|
for a filesystem request. Must be smaller than or equal to the maximum
|
|
|
|
size allowed by the hardware.
|
|
|
|
|
2012-08-09 17:28:05 +04:00
|
|
|
max_segments (RO)
|
|
|
|
-----------------
|
|
|
|
Maximum number of segments of the device.
|
|
|
|
|
|
|
|
max_segment_size (RO)
|
|
|
|
---------------------
|
|
|
|
Maximum segment size of the device.
|
|
|
|
|
|
|
|
minimum_io_size (RO)
|
|
|
|
--------------------
|
2014-08-26 14:33:20 +04:00
|
|
|
This is the smallest preferred IO size reported by the device.
|
2012-08-09 17:28:05 +04:00
|
|
|
|
2009-02-02 15:02:31 +03:00
|
|
|
nomerges (RW)
|
|
|
|
-------------
|
2010-01-29 11:04:08 +03:00
|
|
|
This enables the user to disable the lookup logic involved with IO
|
|
|
|
merging requests in the block layer. By default (0) all merges are
|
|
|
|
enabled. When set to 1 only simple one-hit merges will be tried. When
|
|
|
|
set to 2 no merge algorithms will be tried (including one-hit or more
|
|
|
|
complex tree/hash lookups).
|
2009-02-02 15:02:31 +03:00
|
|
|
|
|
|
|
nr_requests (RW)
|
|
|
|
----------------
|
|
|
|
This controls how many requests may be allocated in the block layer for
|
|
|
|
read or write requests. Note that the total allocated number may be twice
|
|
|
|
this amount, since it applies only to reads or writes (not the accumulated
|
|
|
|
sum).
|
|
|
|
|
blkcg: implement per-blkg request allocation
Currently, request_queue has one request_list to allocate requests
from regardless of blkcg of the IO being issued. When the unified
request pool is used up, cfq proportional IO limits become meaningless
- whoever grabs the next request being freed wins the race regardless
of the configured weights.
This can be easily demonstrated by creating a blkio cgroup w/ very low
weight, put a program which can issue a lot of random direct IOs there
and running a sequential IO from a different cgroup. As soon as the
request pool is used up, the sequential IO bandwidth crashes.
This patch implements per-blkg request_list. Each blkg has its own
request_list and any IO allocates its request from the matching blkg
making blkcgs completely isolated in terms of request allocation.
* Root blkcg uses the request_list embedded in each request_queue,
which was renamed to @q->root_rl from @q->rq. While making blkcg rl
handling a bit harier, this enables avoiding most overhead for root
blkcg.
* Queue fullness is properly per request_list but bdi isn't blkcg
aware yet, so congestion state currently just follows the root
blkcg. As writeback isn't aware of blkcg yet, this works okay for
async congestion but readahead may get the wrong signals. It's
better than blkcg completely collapsing with shared request_list but
needs to be improved with future changes.
* After this change, each block cgroup gets a full request pool making
resource consumption of each cgroup higher. This makes allowing
non-root users to create cgroups less desirable; however, note that
allowing non-root users to directly manage cgroups is already
severely broken regardless of this patch - each block cgroup
consumes kernel memory and skews IO weight (IO weights are not
hierarchical).
v2: queue-sysfs.txt updated and patch description udpated as suggested
by Vivek.
v3: blk_get_rl() wasn't checking error return from
blkg_lookup_create() and may cause oops on lookup failure. Fix it
by falling back to root_rl on blkg lookup failures. This problem
was spotted by Rakesh Iyer <rni@google.com>.
v4: Updated to accomodate 458f27a982 "block: Avoid missed wakeup in
request waitqueue". blk_drain_queue() now wakes up waiters on all
blkg->rl on the target queue.
Signed-off-by: Tejun Heo <tj@kernel.org>
Acked-by: Vivek Goyal <vgoyal@redhat.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2012-06-27 02:05:44 +04:00
|
|
|
To avoid priority inversion through request starvation, a request
|
|
|
|
queue maintains a separate request pool per each cgroup when
|
|
|
|
CONFIG_BLK_CGROUP is enabled, and this parameter applies to each such
|
|
|
|
per-block-cgroup request pool. IOW, if there are N block cgroups,
|
2013-05-09 03:56:16 +04:00
|
|
|
each request queue may have up to N request pools, each independently
|
blkcg: implement per-blkg request allocation
Currently, request_queue has one request_list to allocate requests
from regardless of blkcg of the IO being issued. When the unified
request pool is used up, cfq proportional IO limits become meaningless
- whoever grabs the next request being freed wins the race regardless
of the configured weights.
This can be easily demonstrated by creating a blkio cgroup w/ very low
weight, put a program which can issue a lot of random direct IOs there
and running a sequential IO from a different cgroup. As soon as the
request pool is used up, the sequential IO bandwidth crashes.
This patch implements per-blkg request_list. Each blkg has its own
request_list and any IO allocates its request from the matching blkg
making blkcgs completely isolated in terms of request allocation.
* Root blkcg uses the request_list embedded in each request_queue,
which was renamed to @q->root_rl from @q->rq. While making blkcg rl
handling a bit harier, this enables avoiding most overhead for root
blkcg.
* Queue fullness is properly per request_list but bdi isn't blkcg
aware yet, so congestion state currently just follows the root
blkcg. As writeback isn't aware of blkcg yet, this works okay for
async congestion but readahead may get the wrong signals. It's
better than blkcg completely collapsing with shared request_list but
needs to be improved with future changes.
* After this change, each block cgroup gets a full request pool making
resource consumption of each cgroup higher. This makes allowing
non-root users to create cgroups less desirable; however, note that
allowing non-root users to directly manage cgroups is already
severely broken regardless of this patch - each block cgroup
consumes kernel memory and skews IO weight (IO weights are not
hierarchical).
v2: queue-sysfs.txt updated and patch description udpated as suggested
by Vivek.
v3: blk_get_rl() wasn't checking error return from
blkg_lookup_create() and may cause oops on lookup failure. Fix it
by falling back to root_rl on blkg lookup failures. This problem
was spotted by Rakesh Iyer <rni@google.com>.
v4: Updated to accomodate 458f27a982 "block: Avoid missed wakeup in
request waitqueue". blk_drain_queue() now wakes up waiters on all
blkg->rl on the target queue.
Signed-off-by: Tejun Heo <tj@kernel.org>
Acked-by: Vivek Goyal <vgoyal@redhat.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2012-06-27 02:05:44 +04:00
|
|
|
regulated by nr_requests.
|
|
|
|
|
2012-08-09 17:28:05 +04:00
|
|
|
optimal_io_size (RO)
|
|
|
|
--------------------
|
2014-08-26 14:33:20 +04:00
|
|
|
This is the optimal IO size reported by the device.
|
2012-08-09 17:28:05 +04:00
|
|
|
|
|
|
|
physical_block_size (RO)
|
|
|
|
------------------------
|
|
|
|
This is the physical block size of device, in bytes.
|
|
|
|
|
2009-02-02 15:02:31 +03:00
|
|
|
read_ahead_kb (RW)
|
|
|
|
------------------
|
|
|
|
Maximum number of kilobytes to read-ahead for filesystems on this block
|
|
|
|
device.
|
|
|
|
|
2012-08-09 17:28:05 +04:00
|
|
|
rotational (RW)
|
|
|
|
---------------
|
|
|
|
This file is used to stat if the device is of rotational type or
|
|
|
|
non-rotational type.
|
|
|
|
|
2009-02-02 15:02:31 +03:00
|
|
|
rq_affinity (RW)
|
|
|
|
----------------
|
2011-07-23 22:44:25 +04:00
|
|
|
If this option is '1', the block layer will migrate request completions to the
|
|
|
|
cpu "group" that originally submitted the request. For some workloads this
|
|
|
|
provides a significant reduction in CPU cycles due to caching effects.
|
|
|
|
|
|
|
|
For storage configurations that need to maximize distribution of completion
|
|
|
|
processing setting this option to '2' forces the completion to run on the
|
|
|
|
requesting cpu (bypassing the "group" aggregation logic).
|
2009-02-02 15:02:31 +03:00
|
|
|
|
|
|
|
scheduler (RW)
|
|
|
|
--------------
|
|
|
|
When read, this file will display the current and available IO schedulers
|
|
|
|
for this block device. The currently active IO scheduler will be enclosed
|
|
|
|
in [] brackets. Writing an IO scheduler name to this file will switch
|
|
|
|
control of this block device to that new IO scheduler. Note that writing
|
|
|
|
an IO scheduler name to this file will attempt to load that IO scheduler
|
|
|
|
module, if it isn't already present in the system.
|
|
|
|
|
2016-04-12 21:32:46 +03:00
|
|
|
write_cache (RW)
|
|
|
|
----------------
|
|
|
|
When read, this file will display whether the device has write back
|
|
|
|
caching enabled or not. It will return "write back" for the former
|
|
|
|
case, and "write through" for the latter. Writing to this file can
|
|
|
|
change the kernels view of the device, but it doesn't alter the
|
|
|
|
device state. This means that it might not be safe to toggle the
|
|
|
|
setting from "write back" to "write through", since that will also
|
|
|
|
eliminate cache flushes issued by the kernel.
|
2009-02-02 15:02:31 +03:00
|
|
|
|
2016-08-09 21:01:30 +03:00
|
|
|
write_same_max_bytes (RO)
|
|
|
|
-------------------------
|
|
|
|
This is the number of bytes the device can write in a single write-same
|
|
|
|
command. A value of '0' means write-same is not supported by this
|
|
|
|
device.
|
|
|
|
|
block: hook up writeback throttling
Enable throttling of buffered writeback to make it a lot
more smooth, and has way less impact on other system activity.
Background writeback should be, by definition, background
activity. The fact that we flush huge bundles of it at the time
means that it potentially has heavy impacts on foreground workloads,
which isn't ideal. We can't easily limit the sizes of writes that
we do, since that would impact file system layout in the presence
of delayed allocation. So just throttle back buffered writeback,
unless someone is waiting for it.
The algorithm for when to throttle takes its inspiration in the
CoDel networking scheduling algorithm. Like CoDel, blk-wb monitors
the minimum latencies of requests over a window of time. In that
window of time, if the minimum latency of any request exceeds a
given target, then a scale count is incremented and the queue depth
is shrunk. The next monitoring window is shrunk accordingly. Unlike
CoDel, if we hit a window that exhibits good behavior, then we
simply increment the scale count and re-calculate the limits for that
scale value. This prevents us from oscillating between a
close-to-ideal value and max all the time, instead remaining in the
windows where we get good behavior.
Unlike CoDel, blk-wb allows the scale count to to negative. This
happens if we primarily have writes going on. Unlike positive
scale counts, this doesn't change the size of the monitoring window.
When the heavy writers finish, blk-bw quickly snaps back to it's
stable state of a zero scale count.
The patch registers a sysfs entry, 'wb_lat_usec'. This sets the latency
target to me met. It defaults to 2 msec for non-rotational storage, and
75 msec for rotational storage. Setting this value to '0' disables
blk-wb. Generally, a user would not have to touch this setting.
We don't enable WBT on devices that are managed with CFQ, and have
a non-root block cgroup attached. If we have a proportional share setup
on this particular disk, then the wbt throttling will interfere with
that. We don't have a strong need for wbt for that case, since we will
rely on CFQ doing that for us.
Signed-off-by: Jens Axboe <axboe@fb.com>
2016-11-09 22:38:14 +03:00
|
|
|
wb_lat_usec (RW)
|
|
|
|
----------------
|
|
|
|
If the device is registered for writeback throttling, then this file shows
|
|
|
|
the target minimum read latency. If this latency is exceeded in a given
|
|
|
|
window of time (see wb_window_usec), then the writeback throttling will start
|
|
|
|
scaling back writes.
|
|
|
|
|
2009-02-02 15:02:31 +03:00
|
|
|
|
|
|
|
Jens Axboe <jens.axboe@oracle.com>, February 2009
|