/linux-master/net/ipv4/ |
H A D | tcp_output.c | 3817 const struct tcp_congestion_ops *ca; local 3824 ca = tcp_ca_find_key(ca_key); 3825 if (likely(ca && bpf_try_module_get(ca, ca->owner))) { 3828 icsk->icsk_ca_ops = ca;
|
/linux-master/fs/bcachefs/ |
H A D | super.c | 186 for_each_member_device_rcu(c, ca, NULL) 187 if (ca->disk_sb.bdev && ca->disk_sb.bdev->bd_dev == dev) { 233 for_each_member_device_rcu(c, ca, NULL) 301 for_each_member_device(c, ca) 302 bch2_dev_allocator_remove(c, ca); 472 for_each_rw_member(c, ca) 473 bch2_dev_allocator_add(c, ca); 625 for_each_member_device(c, ca) 626 if (ca 665 struct bch_dev *ca = rcu_dereference_protected(c->devs[i], true); local 1187 struct bch_dev *ca = container_of(kobj, struct bch_dev, kobj); local 1192 bch2_dev_free(struct bch_dev *ca) argument 1219 __bch2_dev_offline(struct bch_fs *c, struct bch_dev *ca) argument 1244 struct bch_dev *ca = container_of(ref, struct bch_dev, ref); local 1251 struct bch_dev *ca = container_of(ref, struct bch_dev, io_ref); local 1256 bch2_dev_sysfs_online(struct bch_fs *c, struct bch_dev *ca) argument 1288 struct bch_dev *ca; local 1333 bch2_dev_attach(struct bch_fs *c, struct bch_dev *ca, unsigned dev_idx) argument 1350 struct bch_dev *ca = NULL; local 1370 __bch2_dev_attach_bdev(struct bch_dev *ca, struct bch_sb_handle *sb) argument 1405 struct bch_dev *ca; local 1449 bch2_dev_state_allowed(struct bch_fs *c, struct bch_dev *ca, enum bch_member_state new_state, int flags) argument 1495 struct bch_dev *ca; local 1527 __bch2_dev_read_only(struct bch_fs *c, struct bch_dev *ca) argument 1536 __bch2_dev_read_write(struct bch_fs *c, struct bch_dev *ca) argument 1546 __bch2_dev_set_state(struct bch_fs *c, struct bch_dev *ca, enum bch_member_state new_state, int flags) argument 1577 bch2_dev_set_state(struct bch_fs *c, struct bch_dev *ca, enum bch_member_state new_state, int flags) argument 1591 bch2_dev_remove_alloc(struct bch_fs *c, struct bch_dev *ca) argument 1617 bch2_dev_remove(struct bch_fs *c, struct bch_dev *ca, int flags) argument 1725 struct bch_dev *ca = NULL; local 1864 struct bch_dev *ca; local 1925 bch2_dev_offline(struct bch_fs *c, struct bch_dev *ca, int flags) argument 1947 bch2_dev_resize(struct bch_fs *c, struct bch_dev *ca, u64 nbuckets) argument [all...] |
H A D | super-io.c | 231 for_each_online_member(c, ca) { 232 struct bch_sb_handle *dev_sb = &ca->disk_sb; 235 percpu_ref_put(&ca->io_ref); 536 for_each_member_device(c, ca) { 537 struct bch_member m = bch2_sb_member_get(src, ca->dev_idx); 538 ca->mi = bch2_mi_to_cpu(&m); 616 int bch2_sb_from_fs(struct bch_fs *c, struct bch_dev *ca) argument 618 return __copy_super(&ca->disk_sb, c->disk_sb.sb); 857 struct bch_dev *ca = bio->bi_private; local 861 if (bch2_dev_io_err_on(bio->bi_status, ca, 874 read_back_super(struct bch_fs *c, struct bch_dev *ca) argument 892 write_one_super(struct bch_fs *c, struct bch_dev *ca, unsigned idx) argument 1017 struct bch_dev *ca = *cap; local 1046 struct bch_dev *ca = *cap; local 1058 struct bch_dev *ca = *cap; local [all...] |
H A D | recovery.c | 364 struct bch_dev *ca = bch_dev_bkey_exists(c, le32_to_cpu(u->dev)); local 368 ca->usage_base->d[i].buckets = le64_to_cpu(u->d[i].buckets); 369 ca->usage_base->d[i].sectors = le64_to_cpu(u->d[i].sectors); 370 ca->usage_base->d[i].fragmented = le64_to_cpu(u->d[i].fragmented); 952 for_each_member_device(c, ca) 953 bch2_dev_usage_init(ca); 980 for_each_online_member(c, ca) 981 ca->new_fs_bucket_idx = 0;
|
H A D | sb-members.c | 369 for_each_member_device_rcu(c, ca, NULL) { 370 struct bch_member *m = __bch2_members_v2_get_mut(mi, ca->dev_idx); 373 m->errors[e] = cpu_to_le64(atomic64_read(&ca->errors[e])); 378 void bch2_dev_io_errors_to_text(struct printbuf *out, struct bch_dev *ca) argument 380 struct bch_fs *c = ca->fs; 383 mutex_lock(&ca->fs->sb_lock); 384 m = bch2_sb_member_get(c->disk_sb.sb, ca->dev_idx); 385 mutex_unlock(&ca->fs->sb_lock); 396 prt_u64(out, atomic64_read(&ca->errors[i])); 410 prt_u64(out, atomic64_read(&ca 416 bch2_dev_errors_reset(struct bch_dev *ca) argument [all...] |
H A D | sb-clean.c | 239 for_each_member_device(c, ca) { 247 u->dev = cpu_to_le32(ca->dev_idx); 250 u->d[i].buckets = cpu_to_le64(ca->usage_base->d[i].buckets); 251 u->d[i].sectors = cpu_to_le64(ca->usage_base->d[i].sectors); 252 u->d[i].fragmented = cpu_to_le64(ca->usage_base->d[i].fragmented);
|
H A D | sb-members.h | 21 static inline bool bch2_dev_is_online(struct bch_dev *ca) argument 23 return !percpu_ref_is_zero(&ca->io_ref); 26 static inline bool bch2_dev_is_readable(struct bch_dev *ca) argument 28 return bch2_dev_is_online(ca) && 29 ca->mi.state != BCH_MEMBER_STATE_failed; 32 static inline bool bch2_dev_get_ioref(struct bch_dev *ca, int rw) argument 34 if (!percpu_ref_tryget(&ca->io_ref)) 37 if (ca->mi.state == BCH_MEMBER_STATE_rw || 38 (ca->mi.state == BCH_MEMBER_STATE_ro && rw == READ)) 41 percpu_ref_put(&ca 86 struct bch_dev *ca = NULL; local 98 __bch2_next_dev(struct bch_fs *c, struct bch_dev *ca, const struct bch_devs_mask *mask) argument 108 bch2_get_next_dev(struct bch_fs *c, struct bch_dev *ca) argument 131 bch2_get_next_online_dev(struct bch_fs *c, struct bch_dev *ca, unsigned state_mask) argument 234 bch2_dev_btree_bitmap_marked_sectors(struct bch_dev *ca, u64 start, unsigned sectors) argument [all...] |
H A D | journal.c | 921 static int __bch2_set_nr_journal_buckets(struct bch_dev *ca, unsigned nr, argument 924 struct bch_fs *c = ca->fs; 925 struct journal_device *ja = &ca->journal; 945 bu[nr_got] = bch2_bucket_alloc_new_fs(ca); 951 ob[nr_got] = bch2_bucket_alloc(c, ca, BCH_WATERMARK_normal, cl); 957 bch2_trans_mark_metadata_bucket(trans, ca, 959 ca->mi.bucket_size)); 1003 ret = bch2_journal_buckets_to_sb(c, ca, new_buckets, nr); 1038 bch2_trans_mark_metadata_bucket(trans, ca, 1056 int bch2_set_nr_journal_buckets(struct bch_fs *c, struct bch_dev *ca, argument 1106 bch2_dev_journal_alloc(struct bch_dev *ca) argument 1171 bch2_dev_journal_stop(struct journal *j, struct bch_dev *ca) argument 1285 bch2_dev_journal_exit(struct bch_dev *ca) argument 1300 bch2_dev_journal_init(struct bch_dev *ca, struct bch_sb *sb) argument [all...] |
H A D | io_write.c | 39 static inline void bch2_congested_acct(struct bch_dev *ca, u64 io_latency, argument 43 ca->io_latency[rw].quantiles.entries[QUANTILE_IDX(1)].m; 54 if (atomic_read(&ca->congested) < CONGESTED_MAX) 57 &ca->congested); 59 ca->congested_last = now; 60 } else if (atomic_read(&ca->congested) > 0) { 61 atomic_dec(&ca->congested); 65 void bch2_latency_acct(struct bch_dev *ca, u64 submit_time, int rw) argument 67 atomic64_t *latency = &ca->cur_latency[rw]; 89 bch2_congested_acct(ca, io_latenc 412 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local 653 struct bch_dev *ca = bch_dev_bkey_exists(c, wbio->dev); local 1296 struct bch_dev *ca = bch_dev_bkey_exists(c, i->b.inode); local [all...] |
H A D | fs.c | 1744 for_each_online_member(c, ca) { 1748 seq_puts(seq, ca->disk_sb.sb_name); 1966 for_each_online_member(c, ca) { 1967 struct block_device *bdev = ca->disk_sb.bdev; 1972 percpu_ref_put(&ca->io_ref);
|
H A D | backpointers.h | 42 struct bch_dev *ca = bch_dev_bkey_exists(c, bp_pos.inode); local 45 return POS(bp_pos.inode, sector_to_bucket(ca, bucket_sector)); 48 static inline struct bpos bucket_pos_to_bp_noerror(const struct bch_dev *ca, argument 53 (bucket_to_sector(ca, bucket.offset) << 64 struct bch_dev *ca = bch_dev_bkey_exists(c, bucket.inode); local 65 struct bpos ret = bucket_pos_to_bp_noerror(ca, bucket, bucket_offset);
|
H A D | backpointers.c | 52 struct bch_dev *ca = bch_dev_bkey_exists(c, bp.k->p.inode); local 56 bkey_fsck_err_on((bp.v->bucket_offset >> MAX_EXTENT_COMPRESS_RATIO_SHIFT) >= ca->mi.bucket_size || 57 !bpos_eq(bp.k->p, bucket_pos_to_bp_noerror(ca, bucket, bp.v->bucket_offset)), 463 struct bch_dev *ca = bch_dev_bkey_exists(c, dev); local 464 if (!bch2_dev_get_ioref(ca, READ)) 473 bio = bio_alloc(ca->disk_sb.bdev, buf_pages(data_buf, bytes), REQ_OP_READ, GFP_KERNEL); 497 percpu_ref_put(&ca->io_ref);
|
H A D | alloc_background.h | 18 struct bch_dev *ca; local 23 ca = bch_dev_bkey_exists(c, pos.inode); 24 return pos.offset >= ca->mi.first_bucket && 25 pos.offset < ca->mi.nbuckets; 84 static inline unsigned bch2_bucket_sectors_fragmented(struct bch_dev *ca, argument 89 return d ? max(0, ca->mi.bucket_size - d) : 0; 108 struct bch_dev *ca) 111 !bch2_bucket_sectors_fragmented(ca, a)) 115 return div_u64(d * (1ULL << 31), ca->mi.bucket_size); 241 static inline u64 should_invalidate_buckets(struct bch_dev *ca, argument 107 alloc_lru_idx_fragmentation(struct bch_alloc_v4 a, struct bch_dev *ca) argument [all...] |
H A D | alloc_background.c | 592 struct bch_dev *ca = bch_dev_bkey_exists(c, k.k->p.inode); 594 for (u64 b = max_t(u64, ca->mi.first_bucket, start); 595 b < min_t(u64, ca->mi.nbuckets, end); 597 *bucket_gen(ca, b) = g->gens[b & KEY_TYPE_BUCKET_GENS_MASK]; 610 struct bch_dev *ca = bch_dev_bkey_exists(c, k.k->p.inode); 613 *bucket_gen(ca, k.k->p.offset) = bch2_alloc_to_v4(k, &a)->gen; 633 struct bch_dev *ca = bch_dev_bkey_exists(c, alloc_k.k->p.inode); local 675 if (ca->mi.freespace_initialized && 746 struct bch_dev *ca = bch_dev_bkey_exists(c, new.k->p.inode); local 863 *bucket_gen(ca, ne 963 struct bch_dev *ca; local 1010 struct bch_dev *ca = bch_dev_bkey_exists(c, bucket.inode); local 1028 struct bch_dev *ca; local 1157 struct bch_dev *ca; local 1340 struct bch_dev *ca; local 1646 struct bch_dev *ca; member in struct:discard_buckets_state 1650 discard_buckets_next_dev(struct bch_fs *c, struct discard_buckets_state *s, struct bch_dev *ca) argument 1676 struct bch_dev *ca; local 1858 struct bch_dev *ca; local 1906 struct bch_dev *ca = bch_dev_bkey_exists(c, bucket.inode); local 2037 bch2_dev_freespace_init(struct bch_fs *c, struct bch_dev *ca, u64 bucket_start, u64 bucket_end) argument 2284 bch2_dev_has_open_write_point(struct bch_fs *c, struct bch_dev *ca) argument 2303 bch2_dev_allocator_remove(struct bch_fs *c, struct bch_dev *ca) argument 2338 bch2_dev_allocator_add(struct bch_fs *c, struct bch_dev *ca) argument [all...] |
H A D | buckets.c | 66 for_each_member_device(c, ca) { 67 struct bch_dev_usage dev = bch2_dev_usage_read(ca); 71 ca->mi.bucket_size; 77 static inline struct bch_dev_usage *dev_usage_ptr(struct bch_dev *ca, argument 84 ? ca->usage_gc 85 : ca->usage[journal_seq & JOURNAL_BUF_MASK]); 88 void bch2_dev_usage_read_fast(struct bch_dev *ca, struct bch_dev_usage *usage) argument 90 struct bch_fs *c = ca->fs; 95 memcpy(usage, ca->usage_base, u64s * sizeof(u64)); 96 for (i = 0; i < ARRAY_SIZE(ca 270 bch2_dev_usage_init(struct bch_dev *ca) argument 299 bch2_dev_usage_update(struct bch_fs *c, struct bch_dev *ca, const struct bch_alloc_v4 *old, const struct bch_alloc_v4 *new, u64 journal_seq, bool gc) argument 343 bch2_dev_usage_update_m(struct bch_fs *c, struct bch_dev *ca, struct bucket *old, struct bucket *new) argument 499 bch2_mark_metadata_bucket(struct bch_fs *c, struct bch_dev *ca, size_t b, enum bch_data_type data_type, unsigned sectors, struct gc_pos pos, unsigned flags) argument 560 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local 853 struct bch_dev *ca = bch_dev_bkey_exists(c, p.ptr.dev); local 1120 __bch2_trans_mark_metadata_bucket(struct btree_trans *trans, struct bch_dev *ca, size_t b, enum bch_data_type type, unsigned sectors) argument 1164 bch2_trans_mark_metadata_bucket(struct btree_trans *trans, struct bch_dev *ca, size_t b, enum bch_data_type type, unsigned sectors) argument 1173 bch2_trans_mark_metadata_sectors(struct btree_trans *trans, struct bch_dev *ca, u64 start, u64 end, enum bch_data_type type, u64 *bucket, unsigned *bucket_sectors) argument 1201 __bch2_trans_mark_dev_sb(struct btree_trans *trans, struct bch_dev *ca) argument 1245 bch2_trans_mark_dev_sb(struct bch_fs *c, struct bch_dev *ca) argument 1342 bch2_dev_buckets_resize(struct bch_fs *c, struct bch_dev *ca, u64 nbuckets) argument 1408 bch2_dev_buckets_free(struct bch_dev *ca) argument 1418 bch2_dev_buckets_alloc(struct bch_fs *c, struct bch_dev *ca) argument [all...] |
H A D | btree_node_scan.c | 20 struct bch_dev *ca; member in struct:find_btree_nodes_worker 121 static void try_read_btree_node(struct find_btree_nodes *f, struct bch_dev *ca, argument 126 bio_reset(bio, ca->disk_sb.bdev, REQ_OP_READ); 131 if (bch2_dev_io_err_on(bio->bi_status, ca, BCH_MEMBER_ERROR_read, 163 .ptrs[0].dev = ca->dev_idx, 164 .ptrs[0].gen = *bucket_gen(ca, sector_to_bucket(ca, offset)), 187 struct bch_dev *ca = w->ca; local 198 for (u64 bucket = ca [all...] |
H A D | journal_io.c | 24 struct bch_dev *ca = bch_dev_bkey_exists(c, i->dev); local 27 div64_u64_rem(i->sector, ca->mi.bucket_size, &offset); 114 static int journal_entry_add(struct bch_fs *c, struct bch_dev *ca, argument 176 if (ptr->dev == ca->dev_idx) 864 struct bch_dev *ca, 879 ca ? ca->name : c->name, 891 ca ? ca->name : c->name, 914 struct bch_dev *ca, 863 jset_validate(struct bch_fs *c, struct bch_dev *ca, struct jset *jset, u64 sector, enum bkey_invalid_flags flags) argument 913 jset_validate_early(struct bch_fs *c, struct bch_dev *ca, struct jset *jset, u64 sector, unsigned bucket_sectors_left, unsigned sectors_read) argument 981 journal_read_bucket(struct bch_dev *ca, struct journal_read_buf *buf, struct journal_list *jlist, unsigned bucket) argument 1124 struct bch_dev *ca = container_of(ja, struct bch_dev, journal); local 1393 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local 1446 struct bch_dev *ca; local 1504 struct bch_dev *ca; local 1706 struct bch_dev *ca = jbio->ca; local 1734 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local [all...] |
H A D | chardev.c | 25 /* returns with ref on ca->ref */ 29 struct bch_dev *ca; local 36 ca = rcu_dereference(c->devs[dev]); 37 if (ca) 38 percpu_ref_get(&ca->ref); 41 if (!ca) 51 ca = bch2_dev_lookup(c, path); 55 return ca; 334 struct bch_dev *ca; local 346 ca 376 struct bch_dev *ca; local 401 struct bch_dev *ca; local 587 struct bch_dev *ca; local 628 struct bch_dev *ca; local 677 struct bch_dev *ca = NULL; local 740 struct bch_dev *ca; local 763 struct bch_dev *ca; local [all...] |
H A D | btree_io.c | 514 struct bch_dev *ca, 522 if (ca) 523 prt_printf(out, "on %s ", ca->name); 537 struct bch_dev *ca, 548 btree_err_msg(&out, c, ca, b, i, b->written, write); 599 #define btree_err(type, c, ca, b, i, _err_type, msg, ...) \ 601 int _ret = __btree_err(type, c, ca, b, i, write, have_retry, \ 668 static int validate_bset(struct bch_fs *c, struct bch_dev *ca, argument 680 c, ca, b, i, 713 c, ca, local 513 btree_err_msg(struct printbuf *out, struct bch_fs *c, struct bch_dev *ca, struct btree *b, struct bset *i, unsigned offset, int write) argument 535 __btree_err(int ret, struct bch_fs *c, struct bch_dev *ca, struct btree *b, struct bset *i, int write, bool have_retry, enum bch_sb_error_id err_type, const char *fmt, ...) argument 735 c, ca, b, i, local 758 c, ca, b, i, local 764 c, ca, b, i, local 806 c, ca, b, i, local 975 bch2_btree_node_read_done(struct bch_fs *c, struct bch_dev *ca, struct btree *b, bool have_retry, bool *saw_error) argument 1007 c, ca, b, NULL, local 1283 struct bch_dev *ca = bch_dev_bkey_exists(c, rb->pick.ptr.dev); local 1366 struct bch_dev *ca = bch_dev_bkey_exists(c, rb->pick.ptr.dev); local 1563 struct bch_dev *ca = bch_dev_bkey_exists(c, rb->pick.ptr.dev); local 1605 struct bch_dev *ca = bch_dev_bkey_exists(c, pick.ptr.dev); local 1650 struct bch_dev *ca; local 1899 struct bch_dev *ca = bch_dev_bkey_exists(c, wbio->dev); local [all...] |
H A D | btree_gc.c | 606 struct bch_dev *ca = bch_dev_bkey_exists(c, p.ptr.dev); local 607 struct bucket *g = PTR_GC_BUCKET(ca, &p.ptr); 614 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr), 631 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr), 652 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr), g->gen, 663 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr), 679 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr), 739 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local 740 struct bucket *g = PTR_GC_BUCKET(ca, ptr); 750 struct bch_dev *ca local 1117 mark_metadata_sectors(struct bch_fs *c, struct bch_dev *ca, u64 start, u64 end, enum bch_data_type type, unsigned flags) argument 1135 bch2_mark_dev_superblock(struct bch_fs *c, struct bch_dev *ca, unsigned flags) argument 1210 struct bch_dev *ca = NULL; local 1368 struct bch_dev *ca = bch_dev_bkey_exists(c, iter->pos.inode); local 1857 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local 1866 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local 1887 struct bch_dev *ca = bch_dev_bkey_exists(trans->c, iter->pos.inode); local [all...] |
H A D | sysfs.c | 795 static void dev_alloc_debug_to_text(struct printbuf *out, struct bch_dev *ca) argument 797 struct bch_fs *c = ca->fs; 798 struct bch_dev_usage stats = bch2_dev_usage_read(ca); 821 prt_u64(out, bch2_dev_buckets_reserved(ca, i)); 843 prt_u64(out, ca->nr_open_buckets); 868 prt_u64(out, should_invalidate_buckets(ca, stats)); 883 static void dev_io_done_to_text(struct printbuf *out, struct bch_dev *ca) argument 893 percpu_u64_get(&ca->io_done->sectors[rw][i]) << 9); 899 struct bch_dev *ca = container_of(kobj, struct bch_dev, kobj); local 900 struct bch_fs *c = ca 953 struct bch_dev *ca = container_of(kobj, struct bch_dev, kobj); local [all...] |
H A D | fs-io.c | 41 struct bch_dev *ca; member in struct:nocow_flush 51 percpu_ref_put(&bio->ca->io_ref); 60 struct bch_dev *ca; local 73 ca = rcu_dereference(c->devs[dev]); 74 if (ca && !percpu_ref_tryget(&ca->io_ref)) 75 ca = NULL; 78 if (!ca) 81 bio = container_of(bio_alloc_bioset(ca->disk_sb.bdev, 0, 87 bio->ca [all...] |
H A D | ec.c | 101 struct bch_dev *ca; member in struct:ec_bio 256 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local 266 g = PTR_GC_BUCKET(ca, ptr); 272 ptr->dev, PTR_BUCKET_NR(ca, ptr), g->gen, 296 bch2_dev_usage_update_m(c, ca, &old, &new); 612 struct bch_dev *ca = bch_dev_bkey_exists(c, v->ptrs[i].dev); local 618 bch_err_ratelimited(ca, "%s", err.buf); 623 bch2_io_error(ca, BCH_MEMBER_ERROR_checksum); 678 struct bch_dev *ca = ec_bio->ca; local 708 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local 1324 struct bch_dev *ca = bch_dev_bkey_exists(c, ob->dev); local 1522 struct bch_dev *ca; local 2067 __bch2_ec_stop(struct bch_fs *c, struct bch_dev *ca) argument 2100 bch2_ec_stop_dev(struct bch_fs *c, struct bch_dev *ca) argument [all...] |
/linux-master/drivers/gpu/drm/panel/ |
H A D | panel-ilitek-ili9341.c | 122 /* ca: TODO: need comments for this register */ 123 u8 ca[ILI9341_CA_LEN]; member in struct:ili9341_config 209 .ca = {0xc3, 0x08, 0x50}, 308 mipi_dbi_command_stackbuf(dbi, 0xca, cfg->ca, ILI9341_CA_LEN);
|
/linux-master/fs/nfsd/ |
H A D | nfs4xdr.c | 1663 struct nfsd4_channel_attrs *ca) 1673 ca->maxreq_sz = be32_to_cpup(p++); 1674 ca->maxresp_sz = be32_to_cpup(p++); 1675 ca->maxresp_cached = be32_to_cpup(p++); 1676 ca->maxops = be32_to_cpup(p++); 1677 ca->maxreqs = be32_to_cpup(p++); 1678 ca->nr_rdma_attrs = be32_to_cpup(p); 1679 switch (ca->nr_rdma_attrs) { 1683 if (xdr_stream_decode_u32(argp->xdr, &ca->rdma_attrs) < 0) 1662 nfsd4_decode_channel_attrs4(struct nfsd4_compoundargs *argp, struct nfsd4_channel_attrs *ca) argument
|