#define THROTL_SERVICE_QUEUE_INITIALIZER \
(struct throtl_service_queue){ .pending_tree = RB_ROOT }
+enum tg_state_flags {
+ THROTL_TG_PENDING = 1 << 0, /* on parent's pending tree */
+};
+
#define rb_entry_tg(node) rb_entry((node), struct throtl_grp, rb_node)
/* Per-cpu group stats */
return blkg_to_tg(td->queue->root_blkg);
}
-enum tg_state_flags {
- THROTL_TG_FLAG_on_rr = 0, /* on round-robin busy list */
-};
-
-#define THROTL_TG_FNS(name) \
-static inline void throtl_mark_tg_##name(struct throtl_grp *tg) \
-{ \
- (tg)->flags |= (1 << THROTL_TG_FLAG_##name); \
-} \
-static inline void throtl_clear_tg_##name(struct throtl_grp *tg) \
-{ \
- (tg)->flags &= ~(1 << THROTL_TG_FLAG_##name); \
-} \
-static inline int throtl_tg_##name(const struct throtl_grp *tg) \
-{ \
- return ((tg)->flags & (1 << THROTL_TG_FLAG_##name)) != 0; \
-}
-
-THROTL_TG_FNS(on_rr);
-
#define throtl_log_tg(td, tg, fmt, args...) do { \
char __pbuf[128]; \
\
struct throtl_service_queue *sq = &td->service_queue;
tg_service_queue_add(sq, tg);
- throtl_mark_tg_on_rr(tg);
+ tg->flags |= THROTL_TG_PENDING;
sq->nr_pending++;
}
static void throtl_enqueue_tg(struct throtl_data *td, struct throtl_grp *tg)
{
- if (!throtl_tg_on_rr(tg))
+ if (!(tg->flags & THROTL_TG_PENDING))
__throtl_enqueue_tg(td, tg);
}
static void __throtl_dequeue_tg(struct throtl_data *td, struct throtl_grp *tg)
{
throtl_rb_erase(&tg->rb_node, &td->service_queue);
- throtl_clear_tg_on_rr(tg);
+ tg->flags &= ~THROTL_TG_PENDING;
}
static void throtl_dequeue_tg(struct throtl_data *td, struct throtl_grp *tg)
{
- if (throtl_tg_on_rr(tg))
+ if (tg->flags & THROTL_TG_PENDING)
__throtl_dequeue_tg(td, tg);
}
throtl_start_new_slice(td, tg, 0);
throtl_start_new_slice(td, tg, 1);
- if (throtl_tg_on_rr(tg)) {
+ if (tg->flags & THROTL_TG_PENDING) {
tg_update_disptime(td, tg);
throtl_schedule_next_dispatch(td);
}