]> git.proxmox.com Git - mirror_qemu.git/blame - block/throttle-groups.c
qemu-iotests: Test removing a throttle group member with a pending timer
[mirror_qemu.git] / block / throttle-groups.c
CommitLineData
2ff1f2e3
AG
1/*
2 * QEMU block throttling group infrastructure
3 *
4 * Copyright (C) Nodalink, EURL. 2014
5 * Copyright (C) Igalia, S.L. 2015
6 *
7 * Authors:
8 * BenoƮt Canet <benoit.canet@nodalink.com>
9 * Alberto Garcia <berto@igalia.com>
10 *
11 * This program is free software; you can redistribute it and/or
12 * modify it under the terms of the GNU General Public License as
13 * published by the Free Software Foundation; either version 2 or
14 * (at your option) version 3 of the License.
15 *
16 * This program is distributed in the hope that it will be useful,
17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
19 * GNU General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, see <http://www.gnu.org/licenses/>.
23 */
24
80c71a24 25#include "qemu/osdep.h"
31dce3cc 26#include "sysemu/block-backend.h"
2ff1f2e3 27#include "block/throttle-groups.h"
432d889e 28#include "qemu/throttle-options.h"
76f4afb4
AG
29#include "qemu/queue.h"
30#include "qemu/thread.h"
31#include "sysemu/qtest.h"
432d889e 32#include "qapi/error.h"
9af23989 33#include "qapi/qapi-visit-block-core.h"
432d889e
MP
34#include "qom/object.h"
35#include "qom/object_interfaces.h"
36
37static void throttle_group_obj_init(Object *obj);
38static void throttle_group_obj_complete(UserCreatable *obj, Error **errp);
2ff1f2e3
AG
39
40/* The ThrottleGroup structure (with its ThrottleState) is shared
022cdc9f 41 * among different ThrottleGroupMembers and it's independent from
2ff1f2e3
AG
42 * AioContext, so in order to use it from different threads it needs
43 * its own locking.
44 *
45 * This locking is however handled internally in this file, so it's
d87d01e1 46 * transparent to outside users.
2ff1f2e3
AG
47 *
48 * The whole ThrottleGroup structure is private and invisible to
49 * outside users, that only use it through its ThrottleState.
50 *
022cdc9f 51 * In addition to the ThrottleGroup structure, ThrottleGroupMember has
2ff1f2e3 52 * fields that need to be accessed by other members of the group and
27ccdd52 53 * therefore also need to be protected by this lock. Once a
022cdc9f 54 * ThrottleGroupMember is registered in a group those fields can be accessed
27ccdd52 55 * by other threads any time.
2ff1f2e3
AG
56 *
57 * Again, all this is handled internally and is mostly transparent to
58 * the outside. The 'throttle_timers' field however has an additional
59 * constraint because it may be temporarily invalid (see for example
0d2fac8e 60 * blk_set_aio_context()). Therefore in this file a thread will
022cdc9f
MP
61 * access some other ThrottleGroupMember's timers only after verifying that
62 * that ThrottleGroupMember has throttled requests in the queue.
2ff1f2e3
AG
63 */
64typedef struct ThrottleGroup {
432d889e
MP
65 Object parent_obj;
66
67 /* refuse individual property change if initialization is complete */
68 bool is_initialized;
2ff1f2e3
AG
69 char *name; /* This is constant during the lifetime of the group */
70
71 QemuMutex lock; /* This lock protects the following four fields */
72 ThrottleState ts;
022cdc9f
MP
73 QLIST_HEAD(, ThrottleGroupMember) head;
74 ThrottleGroupMember *tokens[2];
2ff1f2e3 75 bool any_timer_armed[2];
dbe824cc 76 QEMUClockType clock_type;
2ff1f2e3 77
432d889e 78 /* This field is protected by the global QEMU mutex */
2ff1f2e3
AG
79 QTAILQ_ENTRY(ThrottleGroup) list;
80} ThrottleGroup;
81
432d889e 82/* This is protected by the global QEMU mutex */
2ff1f2e3
AG
83static QTAILQ_HEAD(, ThrottleGroup) throttle_groups =
84 QTAILQ_HEAD_INITIALIZER(throttle_groups);
85
432d889e
MP
86
87/* This function reads throttle_groups and must be called under the global
88 * mutex.
89 */
90static ThrottleGroup *throttle_group_by_name(const char *name)
91{
92 ThrottleGroup *iter;
93
94 /* Look for an existing group with that name */
95 QTAILQ_FOREACH(iter, &throttle_groups, list) {
96 if (!g_strcmp0(name, iter->name)) {
97 return iter;
98 }
99 }
100
101 return NULL;
102}
103
d8e7d87e
MP
104/* This function reads throttle_groups and must be called under the global
105 * mutex.
106 */
107bool throttle_group_exists(const char *name)
108{
109 return throttle_group_by_name(name) != NULL;
110}
111
2ff1f2e3
AG
112/* Increments the reference count of a ThrottleGroup given its name.
113 *
114 * If no ThrottleGroup is found with the given name a new one is
115 * created.
116 *
432d889e
MP
117 * This function edits throttle_groups and must be called under the global
118 * mutex.
119 *
2ff1f2e3 120 * @name: the name of the ThrottleGroup
973f2ddf 121 * @ret: the ThrottleState member of the ThrottleGroup
2ff1f2e3 122 */
973f2ddf 123ThrottleState *throttle_group_incref(const char *name)
2ff1f2e3
AG
124{
125 ThrottleGroup *tg = NULL;
2ff1f2e3
AG
126
127 /* Look for an existing group with that name */
432d889e
MP
128 tg = throttle_group_by_name(name);
129
130 if (tg) {
131 object_ref(OBJECT(tg));
132 } else {
133 /* Create a new one if not found */
134 /* new ThrottleGroup obj will have a refcnt = 1 */
135 tg = THROTTLE_GROUP(object_new(TYPE_THROTTLE_GROUP));
2ff1f2e3 136 tg->name = g_strdup(name);
432d889e 137 throttle_group_obj_complete(USER_CREATABLE(tg), &error_abort);
2ff1f2e3
AG
138 }
139
973f2ddf 140 return &tg->ts;
2ff1f2e3
AG
141}
142
143/* Decrease the reference count of a ThrottleGroup.
144 *
145 * When the reference count reaches zero the ThrottleGroup is
146 * destroyed.
147 *
432d889e
MP
148 * This function edits throttle_groups and must be called under the global
149 * mutex.
150 *
973f2ddf 151 * @ts: The ThrottleGroup to unref, given by its ThrottleState member
2ff1f2e3 152 */
973f2ddf 153void throttle_group_unref(ThrottleState *ts)
2ff1f2e3 154{
973f2ddf 155 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
432d889e 156 object_unref(OBJECT(tg));
2ff1f2e3
AG
157}
158
022cdc9f 159/* Get the name from a ThrottleGroupMember's group. The name (and the pointer)
49d2165d 160 * is guaranteed to remain constant during the lifetime of the group.
2ff1f2e3 161 *
022cdc9f 162 * @tgm: a ThrottleGroupMember
2ff1f2e3
AG
163 * @ret: the name of the group.
164 */
022cdc9f 165const char *throttle_group_get_name(ThrottleGroupMember *tgm)
2ff1f2e3 166{
022cdc9f 167 ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts);
2ff1f2e3
AG
168 return tg->name;
169}
170
022cdc9f
MP
171/* Return the next ThrottleGroupMember in the round-robin sequence, simulating
172 * a circular list.
2ff1f2e3
AG
173 *
174 * This assumes that tg->lock is held.
175 *
022cdc9f
MP
176 * @tgm: the current ThrottleGroupMember
177 * @ret: the next ThrottleGroupMember in the sequence
2ff1f2e3 178 */
022cdc9f 179static ThrottleGroupMember *throttle_group_next_tgm(ThrottleGroupMember *tgm)
2ff1f2e3 180{
022cdc9f 181 ThrottleState *ts = tgm->throttle_state;
2ff1f2e3 182 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
022cdc9f 183 ThrottleGroupMember *next = QLIST_NEXT(tgm, round_robin);
2ff1f2e3
AG
184
185 if (!next) {
31dce3cc 186 next = QLIST_FIRST(&tg->head);
2ff1f2e3
AG
187 }
188
022cdc9f 189 return next;
2ff1f2e3
AG
190}
191
6bf77e1c 192/*
022cdc9f 193 * Return whether a ThrottleGroupMember has pending requests.
6bf77e1c
AG
194 *
195 * This assumes that tg->lock is held.
196 *
022cdc9f
MP
197 * @tgm: the ThrottleGroupMember
198 * @is_write: the type of operation (read/write)
199 * @ret: whether the ThrottleGroupMember has pending requests.
6bf77e1c 200 */
022cdc9f 201static inline bool tgm_has_pending_reqs(ThrottleGroupMember *tgm,
6bf77e1c
AG
202 bool is_write)
203{
022cdc9f 204 return tgm->pending_reqs[is_write];
6bf77e1c
AG
205}
206
022cdc9f
MP
207/* Return the next ThrottleGroupMember in the round-robin sequence with pending
208 * I/O requests.
76f4afb4
AG
209 *
210 * This assumes that tg->lock is held.
211 *
022cdc9f 212 * @tgm: the current ThrottleGroupMember
76f4afb4 213 * @is_write: the type of operation (read/write)
022cdc9f
MP
214 * @ret: the next ThrottleGroupMember with pending requests, or tgm if
215 * there is none.
76f4afb4 216 */
022cdc9f
MP
217static ThrottleGroupMember *next_throttle_token(ThrottleGroupMember *tgm,
218 bool is_write)
76f4afb4 219{
022cdc9f
MP
220 ThrottleState *ts = tgm->throttle_state;
221 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
222 ThrottleGroupMember *token, *start;
76f4afb4
AG
223
224 start = token = tg->tokens[is_write];
225
226 /* get next bs round in round robin style */
022cdc9f
MP
227 token = throttle_group_next_tgm(token);
228 while (token != start && !tgm_has_pending_reqs(token, is_write)) {
229 token = throttle_group_next_tgm(token);
76f4afb4
AG
230 }
231
232 /* If no IO are queued for scheduling on the next round robin token
022cdc9f
MP
233 * then decide the token is the current tgm because chances are
234 * the current tgm got the current request queued.
76f4afb4 235 */
022cdc9f
MP
236 if (token == start && !tgm_has_pending_reqs(token, is_write)) {
237 token = tgm;
76f4afb4
AG
238 }
239
022cdc9f
MP
240 /* Either we return the original TGM, or one with pending requests */
241 assert(token == tgm || tgm_has_pending_reqs(token, is_write));
6bf77e1c 242
76f4afb4
AG
243 return token;
244}
245
022cdc9f
MP
246/* Check if the next I/O request for a ThrottleGroupMember needs to be
247 * throttled or not. If there's no timer set in this group, set one and update
248 * the token accordingly.
76f4afb4
AG
249 *
250 * This assumes that tg->lock is held.
251 *
022cdc9f 252 * @tgm: the current ThrottleGroupMember
76f4afb4
AG
253 * @is_write: the type of operation (read/write)
254 * @ret: whether the I/O request needs to be throttled or not
255 */
022cdc9f
MP
256static bool throttle_group_schedule_timer(ThrottleGroupMember *tgm,
257 bool is_write)
76f4afb4 258{
022cdc9f 259 ThrottleState *ts = tgm->throttle_state;
76f4afb4 260 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
022cdc9f 261 ThrottleTimers *tt = &tgm->throttle_timers;
76f4afb4
AG
262 bool must_wait;
263
022cdc9f 264 if (atomic_read(&tgm->io_limits_disabled)) {
ce0f1412
PB
265 return false;
266 }
267
76f4afb4
AG
268 /* Check if any of the timers in this group is already armed */
269 if (tg->any_timer_armed[is_write]) {
270 return true;
271 }
272
273 must_wait = throttle_schedule_timer(ts, tt, is_write);
274
022cdc9f 275 /* If a timer just got armed, set tgm as the current token */
76f4afb4 276 if (must_wait) {
022cdc9f 277 tg->tokens[is_write] = tgm;
76f4afb4
AG
278 tg->any_timer_armed[is_write] = true;
279 }
280
281 return must_wait;
282}
283
022cdc9f 284/* Start the next pending I/O request for a ThrottleGroupMember. Return whether
3b170dc8
PB
285 * any request was actually pending.
286 *
022cdc9f 287 * @tgm: the current ThrottleGroupMember
3b170dc8
PB
288 * @is_write: the type of operation (read/write)
289 */
022cdc9f 290static bool coroutine_fn throttle_group_co_restart_queue(ThrottleGroupMember *tgm,
3b170dc8
PB
291 bool is_write)
292{
93001e9d 293 bool ret;
3b170dc8 294
022cdc9f
MP
295 qemu_co_mutex_lock(&tgm->throttled_reqs_lock);
296 ret = qemu_co_queue_next(&tgm->throttled_reqs[is_write]);
297 qemu_co_mutex_unlock(&tgm->throttled_reqs_lock);
93001e9d
PB
298
299 return ret;
3b170dc8
PB
300}
301
76f4afb4
AG
302/* Look for the next pending I/O request and schedule it.
303 *
304 * This assumes that tg->lock is held.
305 *
022cdc9f 306 * @tgm: the current ThrottleGroupMember
76f4afb4
AG
307 * @is_write: the type of operation (read/write)
308 */
022cdc9f 309static void schedule_next_request(ThrottleGroupMember *tgm, bool is_write)
76f4afb4 310{
022cdc9f
MP
311 ThrottleState *ts = tgm->throttle_state;
312 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
76f4afb4 313 bool must_wait;
022cdc9f 314 ThrottleGroupMember *token;
76f4afb4
AG
315
316 /* Check if there's any pending request to schedule next */
022cdc9f
MP
317 token = next_throttle_token(tgm, is_write);
318 if (!tgm_has_pending_reqs(token, is_write)) {
76f4afb4
AG
319 return;
320 }
321
322 /* Set a timer for the request if it needs to be throttled */
323 must_wait = throttle_group_schedule_timer(token, is_write);
324
325 /* If it doesn't have to wait, queue it for immediate execution */
326 if (!must_wait) {
022cdc9f 327 /* Give preference to requests from the current tgm */
76f4afb4 328 if (qemu_in_coroutine() &&
022cdc9f
MP
329 throttle_group_co_restart_queue(tgm, is_write)) {
330 token = tgm;
76f4afb4 331 } else {
022cdc9f 332 ThrottleTimers *tt = &token->throttle_timers;
dbe824cc 333 int64_t now = qemu_clock_get_ns(tg->clock_type);
7258ed93 334 timer_mod(tt->timers[is_write], now);
76f4afb4
AG
335 tg->any_timer_armed[is_write] = true;
336 }
337 tg->tokens[is_write] = token;
338 }
339}
340
341/* Check if an I/O request needs to be throttled, wait and set a timer
342 * if necessary, and schedule the next request using a round robin
343 * algorithm.
344 *
022cdc9f 345 * @tgm: the current ThrottleGroupMember
76f4afb4
AG
346 * @bytes: the number of bytes for this I/O
347 * @is_write: the type of operation (read/write)
348 */
022cdc9f 349void coroutine_fn throttle_group_co_io_limits_intercept(ThrottleGroupMember *tgm,
76f4afb4
AG
350 unsigned int bytes,
351 bool is_write)
352{
353 bool must_wait;
022cdc9f
MP
354 ThrottleGroupMember *token;
355 ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts);
76f4afb4
AG
356 qemu_mutex_lock(&tg->lock);
357
358 /* First we check if this I/O has to be throttled. */
022cdc9f 359 token = next_throttle_token(tgm, is_write);
76f4afb4
AG
360 must_wait = throttle_group_schedule_timer(token, is_write);
361
362 /* Wait if there's a timer set or queued requests of this type */
022cdc9f
MP
363 if (must_wait || tgm->pending_reqs[is_write]) {
364 tgm->pending_reqs[is_write]++;
76f4afb4 365 qemu_mutex_unlock(&tg->lock);
022cdc9f
MP
366 qemu_co_mutex_lock(&tgm->throttled_reqs_lock);
367 qemu_co_queue_wait(&tgm->throttled_reqs[is_write],
368 &tgm->throttled_reqs_lock);
369 qemu_co_mutex_unlock(&tgm->throttled_reqs_lock);
76f4afb4 370 qemu_mutex_lock(&tg->lock);
022cdc9f 371 tgm->pending_reqs[is_write]--;
76f4afb4
AG
372 }
373
374 /* The I/O will be executed, so do the accounting */
022cdc9f 375 throttle_account(tgm->throttle_state, is_write, bytes);
76f4afb4
AG
376
377 /* Schedule the next request */
022cdc9f 378 schedule_next_request(tgm, is_write);
76f4afb4
AG
379
380 qemu_mutex_unlock(&tg->lock);
381}
382
3b170dc8 383typedef struct {
022cdc9f 384 ThrottleGroupMember *tgm;
3b170dc8
PB
385 bool is_write;
386} RestartData;
387
388static void coroutine_fn throttle_group_restart_queue_entry(void *opaque)
7258ed93 389{
3b170dc8 390 RestartData *data = opaque;
022cdc9f
MP
391 ThrottleGroupMember *tgm = data->tgm;
392 ThrottleState *ts = tgm->throttle_state;
393 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
3b170dc8 394 bool is_write = data->is_write;
7258ed93
PB
395 bool empty_queue;
396
022cdc9f 397 empty_queue = !throttle_group_co_restart_queue(tgm, is_write);
7258ed93
PB
398
399 /* If the request queue was empty then we have to take care of
400 * scheduling the next one */
401 if (empty_queue) {
402 qemu_mutex_lock(&tg->lock);
022cdc9f 403 schedule_next_request(tgm, is_write);
7258ed93
PB
404 qemu_mutex_unlock(&tg->lock);
405 }
43a5dc02
MP
406
407 g_free(data);
7258ed93
PB
408}
409
022cdc9f 410static void throttle_group_restart_queue(ThrottleGroupMember *tgm, bool is_write)
3b170dc8
PB
411{
412 Coroutine *co;
43a5dc02
MP
413 RestartData *rd = g_new0(RestartData, 1);
414
415 rd->tgm = tgm;
416 rd->is_write = is_write;
3b170dc8 417
43a5dc02 418 co = qemu_coroutine_create(throttle_group_restart_queue_entry, rd);
c61791fc 419 aio_co_enter(tgm->aio_context, co);
3b170dc8
PB
420}
421
022cdc9f 422void throttle_group_restart_tgm(ThrottleGroupMember *tgm)
a72f6414 423{
022cdc9f
MP
424 if (tgm->throttle_state) {
425 throttle_group_restart_queue(tgm, 0);
426 throttle_group_restart_queue(tgm, 1);
a72f6414
PB
427 }
428}
429
2ff1f2e3
AG
430/* Update the throttle configuration for a particular group. Similar
431 * to throttle_config(), but guarantees atomicity within the
432 * throttling group.
433 *
022cdc9f 434 * @tgm: a ThrottleGroupMember that is a member of the group
2ff1f2e3
AG
435 * @cfg: the configuration to set
436 */
022cdc9f 437void throttle_group_config(ThrottleGroupMember *tgm, ThrottleConfig *cfg)
2ff1f2e3 438{
022cdc9f 439 ThrottleState *ts = tgm->throttle_state;
2ff1f2e3
AG
440 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
441 qemu_mutex_lock(&tg->lock);
27e4cf13 442 throttle_config(ts, tg->clock_type, cfg);
2ff1f2e3 443 qemu_mutex_unlock(&tg->lock);
a72f6414 444
022cdc9f 445 throttle_group_restart_tgm(tgm);
2ff1f2e3
AG
446}
447
448/* Get the throttle configuration from a particular group. Similar to
449 * throttle_get_config(), but guarantees atomicity within the
450 * throttling group.
451 *
022cdc9f 452 * @tgm: a ThrottleGroupMember that is a member of the group
2ff1f2e3
AG
453 * @cfg: the configuration will be written here
454 */
022cdc9f 455void throttle_group_get_config(ThrottleGroupMember *tgm, ThrottleConfig *cfg)
2ff1f2e3 456{
022cdc9f 457 ThrottleState *ts = tgm->throttle_state;
2ff1f2e3
AG
458 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
459 qemu_mutex_lock(&tg->lock);
460 throttle_get_config(ts, cfg);
461 qemu_mutex_unlock(&tg->lock);
462}
463
76f4afb4
AG
464/* ThrottleTimers callback. This wakes up a request that was waiting
465 * because it had been throttled.
466 *
c61791fc 467 * @tgm: the ThrottleGroupMember whose request had been throttled
76f4afb4
AG
468 * @is_write: the type of operation (read/write)
469 */
c61791fc 470static void timer_cb(ThrottleGroupMember *tgm, bool is_write)
76f4afb4 471{
022cdc9f 472 ThrottleState *ts = tgm->throttle_state;
76f4afb4 473 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
76f4afb4
AG
474
475 /* The timer has just been fired, so we can update the flag */
476 qemu_mutex_lock(&tg->lock);
477 tg->any_timer_armed[is_write] = false;
478 qemu_mutex_unlock(&tg->lock);
479
480 /* Run the request that was waiting for this timer */
022cdc9f 481 throttle_group_restart_queue(tgm, is_write);
76f4afb4
AG
482}
483
484static void read_timer_cb(void *opaque)
485{
486 timer_cb(opaque, false);
487}
488
489static void write_timer_cb(void *opaque)
490{
491 timer_cb(opaque, true);
492}
493
022cdc9f
MP
494/* Register a ThrottleGroupMember from the throttling group, also initializing
495 * its timers and updating its throttle_state pointer to point to it. If a
31dce3cc 496 * throttling group with that name does not exist yet, it will be created.
2ff1f2e3 497 *
432d889e
MP
498 * This function edits throttle_groups and must be called under the global
499 * mutex.
500 *
022cdc9f 501 * @tgm: the ThrottleGroupMember to insert
2ff1f2e3 502 * @groupname: the name of the group
c61791fc 503 * @ctx: the AioContext to use
2ff1f2e3 504 */
022cdc9f 505void throttle_group_register_tgm(ThrottleGroupMember *tgm,
c61791fc
MP
506 const char *groupname,
507 AioContext *ctx)
2ff1f2e3
AG
508{
509 int i;
973f2ddf
HR
510 ThrottleState *ts = throttle_group_incref(groupname);
511 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
022cdc9f
MP
512
513 tgm->throttle_state = ts;
c61791fc 514 tgm->aio_context = ctx;
2ff1f2e3
AG
515
516 qemu_mutex_lock(&tg->lock);
022cdc9f 517 /* If the ThrottleGroup is new set this ThrottleGroupMember as the token */
2ff1f2e3
AG
518 for (i = 0; i < 2; i++) {
519 if (!tg->tokens[i]) {
022cdc9f 520 tg->tokens[i] = tgm;
2ff1f2e3
AG
521 }
522 }
523
022cdc9f 524 QLIST_INSERT_HEAD(&tg->head, tgm, round_robin);
76f4afb4 525
022cdc9f 526 throttle_timers_init(&tgm->throttle_timers,
c61791fc 527 tgm->aio_context,
dbe824cc 528 tg->clock_type,
76f4afb4
AG
529 read_timer_cb,
530 write_timer_cb,
c61791fc 531 tgm);
f738cfc8
MP
532 qemu_co_mutex_init(&tgm->throttled_reqs_lock);
533 qemu_co_queue_init(&tgm->throttled_reqs[0]);
534 qemu_co_queue_init(&tgm->throttled_reqs[1]);
76f4afb4 535
2ff1f2e3
AG
536 qemu_mutex_unlock(&tg->lock);
537}
538
022cdc9f 539/* Unregister a ThrottleGroupMember from its group, removing it from the list,
31dce3cc 540 * destroying the timers and setting the throttle_state pointer to NULL.
2ff1f2e3 541 *
022cdc9f
MP
542 * The ThrottleGroupMember must not have pending throttled requests, so the
543 * caller has to drain them first.
5ac72418 544 *
2ff1f2e3
AG
545 * The group will be destroyed if it's empty after this operation.
546 *
022cdc9f 547 * @tgm the ThrottleGroupMember to remove
2ff1f2e3 548 */
022cdc9f 549void throttle_group_unregister_tgm(ThrottleGroupMember *tgm)
2ff1f2e3 550{
022cdc9f
MP
551 ThrottleState *ts = tgm->throttle_state;
552 ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
553 ThrottleGroupMember *token;
2ff1f2e3
AG
554 int i;
555
d8e7d87e
MP
556 if (!ts) {
557 /* Discard already unregistered tgm */
558 return;
559 }
560
022cdc9f
MP
561 assert(tgm->pending_reqs[0] == 0 && tgm->pending_reqs[1] == 0);
562 assert(qemu_co_queue_empty(&tgm->throttled_reqs[0]));
563 assert(qemu_co_queue_empty(&tgm->throttled_reqs[1]));
5ac72418 564
2ff1f2e3
AG
565 qemu_mutex_lock(&tg->lock);
566 for (i = 0; i < 2; i++) {
6fccbb47
SH
567 if (timer_pending(tgm->throttle_timers.timers[i])) {
568 tg->any_timer_armed[i] = false;
569 schedule_next_request(tgm, i);
570 }
022cdc9f
MP
571 if (tg->tokens[i] == tgm) {
572 token = throttle_group_next_tgm(tgm);
573 /* Take care of the case where this is the last tgm in the group */
574 if (token == tgm) {
2ff1f2e3
AG
575 token = NULL;
576 }
577 tg->tokens[i] = token;
578 }
579 }
580
022cdc9f
MP
581 /* remove the current tgm from the list */
582 QLIST_REMOVE(tgm, round_robin);
583 throttle_timers_destroy(&tgm->throttle_timers);
2ff1f2e3
AG
584 qemu_mutex_unlock(&tg->lock);
585
973f2ddf 586 throttle_group_unref(&tg->ts);
022cdc9f 587 tgm->throttle_state = NULL;
2ff1f2e3
AG
588}
589
c61791fc
MP
590void throttle_group_attach_aio_context(ThrottleGroupMember *tgm,
591 AioContext *new_context)
592{
593 ThrottleTimers *tt = &tgm->throttle_timers;
594 throttle_timers_attach_aio_context(tt, new_context);
595 tgm->aio_context = new_context;
596}
597
598void throttle_group_detach_aio_context(ThrottleGroupMember *tgm)
599{
341e0b56 600 ThrottleGroup *tg = container_of(tgm->throttle_state, ThrottleGroup, ts);
c61791fc 601 ThrottleTimers *tt = &tgm->throttle_timers;
341e0b56 602 int i;
dc868fb0
SH
603
604 /* Requests must have been drained */
605 assert(tgm->pending_reqs[0] == 0 && tgm->pending_reqs[1] == 0);
606 assert(qemu_co_queue_empty(&tgm->throttled_reqs[0]));
607 assert(qemu_co_queue_empty(&tgm->throttled_reqs[1]));
608
341e0b56
SH
609 /* Kick off next ThrottleGroupMember, if necessary */
610 qemu_mutex_lock(&tg->lock);
611 for (i = 0; i < 2; i++) {
612 if (timer_pending(tt->timers[i])) {
613 tg->any_timer_armed[i] = false;
614 schedule_next_request(tgm, i);
615 }
616 }
617 qemu_mutex_unlock(&tg->lock);
618
c61791fc
MP
619 throttle_timers_detach_aio_context(tt);
620 tgm->aio_context = NULL;
621}
622
432d889e
MP
623#undef THROTTLE_OPT_PREFIX
624#define THROTTLE_OPT_PREFIX "x-"
625
626/* Helper struct and array for QOM property setter/getter */
627typedef struct {
628 const char *name;
629 BucketType type;
630 enum {
631 AVG,
632 MAX,
633 BURST_LENGTH,
634 IOPS_SIZE,
635 } category;
636} ThrottleParamInfo;
637
638static ThrottleParamInfo properties[] = {
639 {
640 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL,
641 THROTTLE_OPS_TOTAL, AVG,
642 },
643 {
644 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL_MAX,
645 THROTTLE_OPS_TOTAL, MAX,
646 },
647 {
648 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL_MAX_LENGTH,
649 THROTTLE_OPS_TOTAL, BURST_LENGTH,
650 },
651 {
652 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ,
653 THROTTLE_OPS_READ, AVG,
654 },
655 {
656 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ_MAX,
657 THROTTLE_OPS_READ, MAX,
658 },
659 {
660 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ_MAX_LENGTH,
661 THROTTLE_OPS_READ, BURST_LENGTH,
662 },
663 {
664 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE,
665 THROTTLE_OPS_WRITE, AVG,
666 },
667 {
668 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE_MAX,
669 THROTTLE_OPS_WRITE, MAX,
670 },
671 {
672 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE_MAX_LENGTH,
673 THROTTLE_OPS_WRITE, BURST_LENGTH,
674 },
675 {
676 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL,
677 THROTTLE_BPS_TOTAL, AVG,
678 },
679 {
680 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL_MAX,
681 THROTTLE_BPS_TOTAL, MAX,
682 },
683 {
684 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL_MAX_LENGTH,
685 THROTTLE_BPS_TOTAL, BURST_LENGTH,
686 },
687 {
688 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ,
689 THROTTLE_BPS_READ, AVG,
690 },
691 {
692 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ_MAX,
693 THROTTLE_BPS_READ, MAX,
694 },
695 {
696 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ_MAX_LENGTH,
697 THROTTLE_BPS_READ, BURST_LENGTH,
698 },
699 {
700 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE,
701 THROTTLE_BPS_WRITE, AVG,
702 },
703 {
704 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE_MAX,
705 THROTTLE_BPS_WRITE, MAX,
706 },
707 {
708 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE_MAX_LENGTH,
709 THROTTLE_BPS_WRITE, BURST_LENGTH,
710 },
711 {
712 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_SIZE,
713 0, IOPS_SIZE,
714 }
715};
716
717/* This function edits throttle_groups and must be called under the global
718 * mutex */
719static void throttle_group_obj_init(Object *obj)
720{
721 ThrottleGroup *tg = THROTTLE_GROUP(obj);
722
723 tg->clock_type = QEMU_CLOCK_REALTIME;
724 if (qtest_enabled()) {
725 /* For testing block IO throttling only */
726 tg->clock_type = QEMU_CLOCK_VIRTUAL;
727 }
728 tg->is_initialized = false;
729 qemu_mutex_init(&tg->lock);
730 throttle_init(&tg->ts);
731 QLIST_INIT(&tg->head);
732}
733
734/* This function edits throttle_groups and must be called under the global
735 * mutex */
736static void throttle_group_obj_complete(UserCreatable *obj, Error **errp)
737{
738 ThrottleGroup *tg = THROTTLE_GROUP(obj);
739 ThrottleConfig cfg;
740
741 /* set group name to object id if it exists */
742 if (!tg->name && tg->parent_obj.parent) {
743 tg->name = object_get_canonical_path_component(OBJECT(obj));
744 }
745 /* We must have a group name at this point */
746 assert(tg->name);
747
748 /* error if name is duplicate */
d8e7d87e 749 if (throttle_group_exists(tg->name)) {
432d889e
MP
750 error_setg(errp, "A group with this name already exists");
751 return;
752 }
753
754 /* check validity */
755 throttle_get_config(&tg->ts, &cfg);
756 if (!throttle_is_valid(&cfg, errp)) {
757 return;
758 }
759 throttle_config(&tg->ts, tg->clock_type, &cfg);
760 QTAILQ_INSERT_TAIL(&throttle_groups, tg, list);
761 tg->is_initialized = true;
762}
763
764/* This function edits throttle_groups and must be called under the global
765 * mutex */
766static void throttle_group_obj_finalize(Object *obj)
767{
768 ThrottleGroup *tg = THROTTLE_GROUP(obj);
769 if (tg->is_initialized) {
770 QTAILQ_REMOVE(&throttle_groups, tg, list);
771 }
772 qemu_mutex_destroy(&tg->lock);
773 g_free(tg->name);
774}
775
776static void throttle_group_set(Object *obj, Visitor *v, const char * name,
777 void *opaque, Error **errp)
778
779{
780 ThrottleGroup *tg = THROTTLE_GROUP(obj);
781 ThrottleConfig *cfg;
782 ThrottleParamInfo *info = opaque;
783 Error *local_err = NULL;
784 int64_t value;
785
786 /* If we have finished initialization, don't accept individual property
787 * changes through QOM. Throttle configuration limits must be set in one
788 * transaction, as certain combinations are invalid.
789 */
790 if (tg->is_initialized) {
791 error_setg(&local_err, "Property cannot be set after initialization");
792 goto ret;
793 }
794
795 visit_type_int64(v, name, &value, &local_err);
796 if (local_err) {
797 goto ret;
798 }
799 if (value < 0) {
800 error_setg(&local_err, "Property values cannot be negative");
801 goto ret;
802 }
803
804 cfg = &tg->ts.cfg;
805 switch (info->category) {
806 case AVG:
807 cfg->buckets[info->type].avg = value;
808 break;
809 case MAX:
810 cfg->buckets[info->type].max = value;
811 break;
812 case BURST_LENGTH:
813 if (value > UINT_MAX) {
814 error_setg(&local_err, "%s value must be in the"
815 "range [0, %u]", info->name, UINT_MAX);
816 goto ret;
817 }
818 cfg->buckets[info->type].burst_length = value;
819 break;
820 case IOPS_SIZE:
821 cfg->op_size = value;
822 break;
823 }
824
825ret:
826 error_propagate(errp, local_err);
827 return;
828
829}
830
831static void throttle_group_get(Object *obj, Visitor *v, const char *name,
832 void *opaque, Error **errp)
833{
834 ThrottleGroup *tg = THROTTLE_GROUP(obj);
835 ThrottleConfig cfg;
836 ThrottleParamInfo *info = opaque;
837 int64_t value;
838
839 throttle_get_config(&tg->ts, &cfg);
840 switch (info->category) {
841 case AVG:
842 value = cfg.buckets[info->type].avg;
843 break;
844 case MAX:
845 value = cfg.buckets[info->type].max;
846 break;
847 case BURST_LENGTH:
848 value = cfg.buckets[info->type].burst_length;
849 break;
850 case IOPS_SIZE:
851 value = cfg.op_size;
852 break;
853 }
854
855 visit_type_int64(v, name, &value, errp);
856}
857
858static void throttle_group_set_limits(Object *obj, Visitor *v,
859 const char *name, void *opaque,
860 Error **errp)
861
862{
863 ThrottleGroup *tg = THROTTLE_GROUP(obj);
864 ThrottleConfig cfg;
865 ThrottleLimits arg = { 0 };
866 ThrottleLimits *argp = &arg;
867 Error *local_err = NULL;
868
869 visit_type_ThrottleLimits(v, name, &argp, &local_err);
870 if (local_err) {
871 goto ret;
872 }
873 qemu_mutex_lock(&tg->lock);
874 throttle_get_config(&tg->ts, &cfg);
875 throttle_limits_to_config(argp, &cfg, &local_err);
876 if (local_err) {
877 goto unlock;
878 }
879 throttle_config(&tg->ts, tg->clock_type, &cfg);
880
881unlock:
882 qemu_mutex_unlock(&tg->lock);
883ret:
884 error_propagate(errp, local_err);
885 return;
886}
887
888static void throttle_group_get_limits(Object *obj, Visitor *v,
889 const char *name, void *opaque,
890 Error **errp)
891{
892 ThrottleGroup *tg = THROTTLE_GROUP(obj);
893 ThrottleConfig cfg;
894 ThrottleLimits arg = { 0 };
895 ThrottleLimits *argp = &arg;
896
897 qemu_mutex_lock(&tg->lock);
898 throttle_get_config(&tg->ts, &cfg);
899 qemu_mutex_unlock(&tg->lock);
900
901 throttle_config_to_limits(&cfg, argp);
902
903 visit_type_ThrottleLimits(v, name, &argp, errp);
904}
905
906static bool throttle_group_can_be_deleted(UserCreatable *uc)
907{
908 return OBJECT(uc)->ref == 1;
909}
910
911static void throttle_group_obj_class_init(ObjectClass *klass, void *class_data)
912{
913 size_t i = 0;
914 UserCreatableClass *ucc = USER_CREATABLE_CLASS(klass);
915
916 ucc->complete = throttle_group_obj_complete;
917 ucc->can_be_deleted = throttle_group_can_be_deleted;
918
919 /* individual properties */
920 for (i = 0; i < sizeof(properties) / sizeof(ThrottleParamInfo); i++) {
921 object_class_property_add(klass,
922 properties[i].name,
923 "int",
924 throttle_group_get,
925 throttle_group_set,
926 NULL, &properties[i],
927 &error_abort);
928 }
929
930 /* ThrottleLimits */
931 object_class_property_add(klass,
932 "limits", "ThrottleLimits",
933 throttle_group_get_limits,
934 throttle_group_set_limits,
935 NULL, NULL,
936 &error_abort);
937}
938
939static const TypeInfo throttle_group_info = {
940 .name = TYPE_THROTTLE_GROUP,
941 .parent = TYPE_OBJECT,
942 .class_init = throttle_group_obj_class_init,
943 .instance_size = sizeof(ThrottleGroup),
944 .instance_init = throttle_group_obj_init,
945 .instance_finalize = throttle_group_obj_finalize,
946 .interfaces = (InterfaceInfo[]) {
947 { TYPE_USER_CREATABLE },
948 { }
949 },
950};
951
2ff1f2e3
AG
952static void throttle_groups_init(void)
953{
432d889e 954 type_register_static(&throttle_group_info);
2ff1f2e3
AG
955}
956
432d889e 957type_init(throttle_groups_init);