#include <linux/rcu_segcblist.h>
-/* Return number of callbacks in simple callback list. */
-static inline long rcu_cblist_n_cbs(struct rcu_cblist *rclp)
-{
- return rclp->len;
-}
-
/* Return number of lazy callbacks in simple callback list. */
static inline long rcu_cblist_n_lazy_cbs(struct rcu_cblist *rclp)
{
return;
/* Do the accounting first. */
- rdp->n_cbs_adopted += rcu_cblist_n_cbs(&rsp->orphan_done);
- if (rcu_cblist_n_lazy_cbs(&rsp->orphan_done) !=
- rcu_cblist_n_cbs(&rsp->orphan_done))
+ rdp->n_cbs_adopted += rsp->orphan_done.len;
+ if (rcu_cblist_n_lazy_cbs(&rsp->orphan_done) != rsp->orphan_done.len)
rcu_idle_count_callbacks_posted();
rcu_segcblist_insert_count(&rdp->cblist, &rsp->orphan_done);
* Stop only if limit reached and CPU has something to do.
* Note: The rcl structure counts down from zero.
*/
- if (-rcu_cblist_n_cbs(&rcl) >= bl &&
+ if (-rcl.len >= bl &&
(need_resched() ||
(!is_idle_task(current) && !rcu_is_callbacks_kthread())))
break;
}
local_irq_save(flags);
- count = -rcu_cblist_n_cbs(&rcl);
+ count = -rcl.len;
trace_rcu_batch_end(rsp->name, count, !!rcl.head, need_resched(),
is_idle_task(current), rcu_is_callbacks_kthread());
struct rcu_data *rdp,
unsigned long flags)
{
- long ql = rcu_cblist_n_cbs(&rsp->orphan_done);
+ long ql = rsp->orphan_done.len;
long qll = rcu_cblist_n_lazy_cbs(&rsp->orphan_done);
/* If this is not a no-CBs CPU, tell the caller to do it the old way. */
rsp->n_force_qs - rsp->n_force_qs_ngp,
READ_ONCE(rsp->n_force_qs_lh),
rcu_cblist_n_lazy_cbs(&rsp->orphan_done),
- rcu_cblist_n_cbs(&rsp->orphan_done));
+ rsp->orphan_done.len);
for (rnp = &rsp->node[0]; rnp - &rsp->node[0] < rcu_num_nodes; rnp++) {
if (rnp->level != level) {
seq_puts(m, "\n");