/linux-master/drivers/md/bcache/ |
H A D | btree.c | 107 static inline struct bset *write_block(struct btree *b) 139 static uint64_t btree_csum_set(struct btree *b, struct bset *i) 151 struct bset *i = btree_bset_first(b); 173 err = "unsupported bset version"; 341 struct bset *i = btree_bset_last(b); 405 struct bset *i = btree_bset_last(b); 477 struct bset *i = btree_bset_last(b); 1407 struct bset *n1 = btree_bset_first(new_nodes[i]); 1408 struct bset *n2 = btree_bset_first(new_nodes[i - 1]);
|
H A D | bset.h | 40 * A bset is an array of bkeys laid out contiguously in memory in sorted order, 48 * (described below) for searching an individial bset, and on top of that we 53 * Most of the code in bcache doesn't care about an individual bset - it needs 62 * Since keys are variable length, we can't use a binary search on a bset - we 161 * better: see comments in bset.c at cacheline_to_bkey() for 176 * The nodes in the bset tree point to specific keys - this 186 struct bset *data; 250 struct bset *i) 256 struct bset *i) 284 static inline struct bset *bset_next_se [all...] |
H A D | bset.c | 12 #include "bset.h" 21 void bch_dump_bset(struct btree_keys *b, struct bset *i, unsigned int set) 255 * It definites the number of bytes (in struct bset) per struct bkey_float in 260 * we're touching one fewer cacheline in the bset tree in exchange for one more 666 void bch_bset_init_next(struct btree_keys *b, struct bset *i, uint64_t magic) 880 struct bset *i = bset_tree_last(b)->data; 1145 WARN_ONCE(1, "bset was corrupt!\n"); 1194 static void btree_mergesort(struct btree_keys *b, struct bset *out, 1241 struct bset *out = (void *) __get_free_pages(__GFP_NOWARN|GFP_NOWAIT,
|
H A D | bcache.h | 194 #include "bset.h" 674 struct bset *verify_ondisk; 848 * This is used for various on disk data structures - cache_sb, prio_set, bset,
|
H A D | Makefile | 5 bcache-y := alloc.o bset.o btree.o debug.o extents.o io.o\
|
/linux-master/block/ |
H A D | bsg-lib.c | 278 struct bsg_set *bset = local 291 ret = bset->job_fn(blk_mq_rq_to_pdu(req)); 323 struct bsg_set *bset = local 326 bsg_unregister_queue(bset->bd); 329 blk_mq_free_tag_set(&bset->tag_set); 330 kfree(bset); 337 struct bsg_set *bset = local 340 if (!bset->timeout_fn) 342 return bset->timeout_fn(rq); 366 struct bsg_set *bset; local [all...] |
/linux-master/fs/bcachefs/ |
H A D | bcachefs_format.h | 1582 struct bset { 1586 * Highest journal entry this bset contains keys for. 1587 * If on recovery we don't see that journal entry, this bset is ignored: 1602 LE32_BITMASK(BSET_CSUM_TYPE, struct bset, flags, 0, 4); 1604 LE32_BITMASK(BSET_BIG_ENDIAN, struct bset, flags, 4, 5); 1606 struct bset, flags, 5, 6); 1609 LE32_BITMASK(BSET_OFFSET, struct bset, flags, 16, 32); 1615 /* this flags field is encrypted, unlike bset->flags: */ 1625 struct bset keys; 1658 struct bset key 1574 struct bset { struct [all...] |
H A D | btree_update_interior.c | 788 struct bset *last = btree_bset_last(b); 1450 struct bset *bsets[2];
|
H A D | btree_types.h | 41 * better: see comments in bset.c at cacheline_to_bkey() for 638 static inline struct bset *bset(const struct btree *b, 647 __btree_node_ptr_to_offset(b, vstruct_last(bset(b, t))); 651 const struct bset *i) 657 static inline struct bset *btree_bset_first(struct btree *b) 659 return bset(b, b->set); 662 static inline struct bset *btree_bset_last(struct btree *b) 664 return bset(b, bset_tree_last(b)); 681 return t->data_offset + offsetof(struct bset, _dat 571 static inline struct bset *bset(const struct btree *b, function [all...] |
H A D | btree_io.c | 104 static void set_needs_whiteout(struct bset *i, int v) 222 (compacting && !bset_written(b, bset(b, t))); 235 struct bset *i = bset(b, t); 307 struct bset *start_bset = bset(b, &b->set[start_idx]); 319 u64s += le16_to_cpu(bset(b, t)->u64s); 343 /* Make sure we preserve bset journal_seq: */ 345 seq = max(seq, le64_to_cpu(bset(b, t)->journal_seq)); 424 * We're about to add another bset t [all...] |
H A D | btree_trans_commit.c | 96 * If the last bset has been written, or if it's gotten too big - start 97 * a new bset to insert into: 293 struct bset *i = bset(b, t);
|
H A D | debug.c | 39 struct bset *sorted, *inmemory = &b->data->keys; 74 struct bset *i; 130 struct bset *inmemory = &b->data->keys; 224 struct bset *i;
|
H A D | btree_update_interior.h | 229 static inline bool bset_written(struct btree *b, struct bset *i) 283 if (unlikely(bset_written(b, bset(b, t)))) { 316 * write lock must be held on @b (else the dirty bset that we were going to
|
H A D | Makefile | 12 bset.o \
|
H A D | bset.c | 11 #include "bset.h" 56 struct bset *i, unsigned set) 108 bch2_dump_bset(c, b, bset(b, t), t - b->set); 187 k2->_data - bset(b, t)->_data); 683 /* bytes remaining - only valid for last bset: */ 814 void bch2_bset_init_first(struct btree *b, struct bset *i) 830 struct bset *i = &bne->keys; 846 * find _some_ key in the same bset as @k that precedes @k - not necessarily the 1024 EBUG_ON((int) le16_to_cpu(bset(b, t)->u64s) < 1028 le16_add_cpu(&bset( [all...] |
H A D | bset.h | 40 * A bset is an array of bkeys laid out contiguously in memory in sorted order, 48 * (described below) for searching an individial bset, and on top of that we 53 * Most of the code in bcache doesn't care about an individual bset - it needs 62 * Since keys are variable length, we can't use a binary search on a bset - we 181 * It definites the number of bytes (in struct bset) per struct bkey_float in 186 * we're touching one fewer cacheline in the bset tree in exchange for one more 257 static inline struct bset *bset_next_set(struct btree *b, 260 struct bset *i = btree_bset_last(b); 269 void bch2_bset_init_first(struct btree *, struct bset *); 466 unsigned bset, [all...] |
H A D | btree_io.h | 6 #include "bset.h" 91 static inline struct nonce btree_nonce(struct bset *i, unsigned offset) 101 static inline int bset_encrypt(struct bch_fs *c, struct bset *i, unsigned offset)
|
H A D | bkey_sort.c | 6 #include "bset.h" 95 bch2_key_sort_fix_overlapping(struct bch_fs *c, struct bset *dst, 123 bch2_sort_repack(struct bset *dst, struct btree *src,
|
H A D | bkey_sort.h | 43 bch2_key_sort_fix_overlapping(struct bch_fs *, struct bset *, 47 bch2_sort_repack(struct bset *, struct btree *,
|
/linux-master/arch/arc/kernel/ |
H A D | head.S | 32 bset r5, r5, 0 ; I$ exists, but is not used 44 bset r5, r5, 0 ; Disable (+Inv) 56 bset r5, r5, STATUS_AD_BIT
|
H A D | entry-arcv2.S | 210 bset.nz r11, r11, AUX_IRQ_ACT_BIT_U ; NZ means U
|
/linux-master/sound/soc/sh/ |
H A D | rz-ssi.c | 148 u32 bclr, u32 bset) 153 val = (val & ~bclr) | bset; 147 rz_ssi_reg_mask_setl(struct rz_ssi_priv *priv, uint reg, u32 bclr, u32 bset) argument
|
/linux-master/drivers/staging/rtl8712/ |
H A D | rtl871x_ioctl_linux.c | 1814 u8 *pparmbuf, bset; local 1822 bset = (u8)(p->flags & 0xFFFF); 1849 if (bset) { 1867 if (bset == 0x00) { /* query info */
|
/linux-master/arch/arc/include/asm/ |
H A D | entry-arcv2.h | 278 bset r9, r9, STATUS_IE_BIT
|
/linux-master/arch/m68k/kernel/ |
H A D | head.S | 674 bset #CPUTYPE_060,%d1 675 bset #CPUTYPE_0460,%d1 683 bset #CPUTYPE_040,%d1 684 bset #CPUTYPE_0460,%d1 692 bset #CPUTYPE_020,%d1 2872 bset #5,%d0 2901 bset #1,%a1@(LMFP_TSR) 3055 bset #5,%d0
|