Lines Matching refs:b_dev_info

14 static void balloon_page_enqueue_one(struct balloon_dev_info *b_dev_info,
24 balloon_page_insert(b_dev_info, page);
32 * @b_dev_info: balloon device descriptor where we will insert a new page to
40 size_t balloon_page_list_enqueue(struct balloon_dev_info *b_dev_info,
47 spin_lock_irqsave(&b_dev_info->pages_lock, flags);
50 balloon_page_enqueue_one(b_dev_info, page);
53 spin_unlock_irqrestore(&b_dev_info->pages_lock, flags);
61 * @b_dev_info: balloon device descriptor where we will grab a page from.
76 size_t balloon_page_list_dequeue(struct balloon_dev_info *b_dev_info,
83 spin_lock_irqsave(&b_dev_info->pages_lock, flags);
84 list_for_each_entry_safe(page, tmp, &b_dev_info->pages, lru) {
108 spin_unlock_irqrestore(&b_dev_info->pages_lock, flags);
136 * @b_dev_info: balloon device descriptor where we will insert a new page
146 void balloon_page_enqueue(struct balloon_dev_info *b_dev_info,
151 spin_lock_irqsave(&b_dev_info->pages_lock, flags);
152 balloon_page_enqueue_one(b_dev_info, page);
153 spin_unlock_irqrestore(&b_dev_info->pages_lock, flags);
160 * @b_dev_info: balloon device descriptor where we will grab a page from.
177 struct page *balloon_page_dequeue(struct balloon_dev_info *b_dev_info)
183 n_pages = balloon_page_list_dequeue(b_dev_info, &pages, 1);
193 spin_lock_irqsave(&b_dev_info->pages_lock, flags);
194 if (unlikely(list_empty(&b_dev_info->pages) &&
195 !b_dev_info->isolated_pages))
197 spin_unlock_irqrestore(&b_dev_info->pages_lock, flags);
209 struct balloon_dev_info *b_dev_info = balloon_page_device(page);
212 spin_lock_irqsave(&b_dev_info->pages_lock, flags);
214 b_dev_info->isolated_pages++;
215 spin_unlock_irqrestore(&b_dev_info->pages_lock, flags);
222 struct balloon_dev_info *b_dev_info = balloon_page_device(page);
225 spin_lock_irqsave(&b_dev_info->pages_lock, flags);
226 list_add(&page->lru, &b_dev_info->pages);
227 b_dev_info->isolated_pages--;
228 spin_unlock_irqrestore(&b_dev_info->pages_lock, flags);