Lines Matching refs:blkcg
488 static struct bfq_group_data *blkcg_to_bfqgd(struct blkcg *blkcg)
490 return cpd_to_bfqgd(blkcg_to_cpd(blkcg, &blkcg_policy_bfq));
511 struct blkcg *blkcg, gfp_t gfp)
535 struct bfq_group_data *d = blkcg_to_bfqgd(blkg->blkcg);
611 bio_associate_blkg_from_css(bio, &blkg->blkcg->css);
617 &bfqg_to_blkg(bfqd->root_group)->blkcg->css);
754 * Move bic to blkcg, assuming that bfqd->lock is held; which makes
785 serial_nr = bfqg_to_blkg(bfqg)->blkcg->css.serial_nr;
788 * Check whether blkcg has changed. The condition may trigger
1006 struct blkcg *blkcg = css_to_blkcg(seq_css(sf));
1007 struct bfq_group_data *bfqgd = blkcg_to_bfqgd(blkcg);
1030 struct blkcg *blkcg = css_to_blkcg(seq_css(sf));
1031 struct bfq_group_data *bfqgd = blkcg_to_bfqgd(blkcg);
1034 blkcg_print_blkgs(sf, blkcg, bfqg_prfill_weight_device,
1076 struct blkcg *blkcg = css_to_blkcg(css);
1077 struct bfq_group_data *bfqgd = blkcg_to_bfqgd(blkcg);
1085 spin_lock_irq(&blkcg->lock);
1087 hlist_for_each_entry(blkg, &blkcg->blkg_list, blkcg_node) {
1093 spin_unlock_irq(&blkcg->lock);
1104 struct blkcg *blkcg = css_to_blkcg(of_css(of));
1110 ret = blkg_conf_prep(blkcg, &blkcg_policy_bfq, &ctx);