Lines Matching refs:paths

266 		bch2_btree_path_verify(trans, &trans->paths[iter->update_path]);
662 i->old_v = bch2_btree_path_peek_slot(trans->paths + i->path, &i->old_k).v;
1024 if (trans->paths[idx].uptodate) {
1025 __btree_path_get(&trans->paths[idx], false);
1027 __btree_path_put(&trans->paths[idx], false);
1040 * We used to assert that all paths had been traversed here
1136 struct btree_path *path = &trans->paths[path_idx];
1160 path = &trans->paths[path_idx];
1239 btree_path_copy(trans, trans->paths + new, trans->paths + src);
1240 __btree_path_get(trans->paths + new, intent);
1242 trans->paths[new].ip_allocated = ip;
1251 __btree_path_put(trans->paths + path, intent);
1253 trans->paths[path].preserve = false;
1262 int cmp = bpos_cmp(new_pos, trans->paths[path_idx].pos);
1265 EBUG_ON(!trans->paths[path_idx].ref);
1269 struct btree_path *path = trans->paths + path_idx;
1347 __bch2_btree_path_unlock(trans, trans->paths + path);
1348 btree_path_list_remove(trans, trans->paths + path);
1374 struct btree_path *path = trans->paths + path_idx, *dup;
1407 if (!__btree_path_put(trans->paths + path, intent))
1477 struct btree_path *path = trans->paths + path_idx;
1513 struct btree_path *path = trans->paths + path_idx;
1633 struct btree_path *paths = p;
1634 *trans_paths_nr(paths) = nr;
1635 memcpy(paths, trans->paths, trans->nr_paths * sizeof(struct btree_path));
1648 rcu_assign_pointer(trans->paths, paths);
1681 struct btree_path *path = &trans->paths[idx];
1720 trans->paths[path_pos].cached == cached &&
1721 trans->paths[path_pos].btree_id == btree_id &&
1722 trans->paths[path_pos].level == level) {
1723 __btree_path_get(trans->paths + path_pos, intent);
1725 path = trans->paths + path_idx;
1728 path = trans->paths + path_idx;
1778 struct btree_path *path = trans->paths + path_idx;
1877 EBUG_ON(trans->paths[iter->path].cached);
1925 EBUG_ON(trans->paths[iter->path].cached);
2157 btree_path_set_should_be_locked(trans->paths + iter->key_cache_path);
2159 k = bch2_btree_path_peek_slot(trans->paths + iter->key_cache_path, &u);
2305 !bkey_eq(trans->paths[iter->update_path].pos, k.k->p)) {
2328 __btree_path_get(trans->paths + iter->path, iter->flags & BTREE_ITER_intent);
2387 ret = bch2_btree_path_relock(trans, trans->paths + iter->update_path, _THIS_IP_);
2391 btree_path_set_should_be_locked(trans->paths + iter->update_path);
2630 k = bch2_btree_path_peek_slot(trans->paths + iter->path, &iter->k);
2745 BUG_ON(trans->paths[idx].sorted_idx != i);
2760 panic("trans paths out of order!\n");
2790 if (btree_path_cmp(trans->paths + trans->sorted[i],
2791 trans->paths + trans->sorted[i + 1]) > 0) {
2793 trans->paths[trans->sorted[i]].sorted_idx = i;
2794 trans->paths[trans->sorted[i + 1]].sorted_idx = i + 1;
2825 trans->paths[trans->sorted[i]].sorted_idx = i;
2832 struct btree_path *path = trans->paths + path_idx;
2834 path->sorted_idx = pos ? trans->paths[pos].sorted_idx + 1 : trans->nr_sorted;
2848 trans->paths[trans->sorted[i]].sorted_idx = i;
2913 __btree_path_get(trans->paths + src->path, src->flags & BTREE_ITER_intent);
2915 __btree_path_get(trans->paths + src->update_path, src->flags & BTREE_ITER_intent);
3178 trans->paths = trans->_paths;
3181 *trans_paths_nr(trans->paths) = BTREE_ITER_INITIAL;
3220 bch_err(c, "btree paths leaked from %s!", trans->fn);
3239 __btree_path_put(trans->paths + i->path, true);
3264 trans->paths = NULL;
3327 /* trans->paths is rcu protected vs. freeing */
3331 struct btree_path *paths = rcu_dereference(trans->paths);
3332 if (!paths)
3335 unsigned long *paths_allocated = trans_paths_allocated(paths);
3337 trans_for_each_path_idx_from(paths_allocated, *trans_paths_nr(paths), idx, 1) {
3338 struct btree_path *path = paths + idx;