Searched refs:ca (Results 1 - 25 of 174) sorted by last modified time

1234567

/linux-master/net/ipv4/
H A Dtcp_output.c3817 const struct tcp_congestion_ops *ca; local
3824 ca = tcp_ca_find_key(ca_key);
3825 if (likely(ca && bpf_try_module_get(ca, ca->owner))) {
3828 icsk->icsk_ca_ops = ca;
/linux-master/fs/bcachefs/
H A Dsuper.c186 for_each_member_device_rcu(c, ca, NULL)
187 if (ca->disk_sb.bdev && ca->disk_sb.bdev->bd_dev == dev) {
233 for_each_member_device_rcu(c, ca, NULL)
301 for_each_member_device(c, ca)
302 bch2_dev_allocator_remove(c, ca);
472 for_each_rw_member(c, ca)
473 bch2_dev_allocator_add(c, ca);
625 for_each_member_device(c, ca)
626 if (ca
665 struct bch_dev *ca = rcu_dereference_protected(c->devs[i], true); local
1187 struct bch_dev *ca = container_of(kobj, struct bch_dev, kobj); local
1192 bch2_dev_free(struct bch_dev *ca) argument
1219 __bch2_dev_offline(struct bch_fs *c, struct bch_dev *ca) argument
1244 struct bch_dev *ca = container_of(ref, struct bch_dev, ref); local
1251 struct bch_dev *ca = container_of(ref, struct bch_dev, io_ref); local
1256 bch2_dev_sysfs_online(struct bch_fs *c, struct bch_dev *ca) argument
1288 struct bch_dev *ca; local
1333 bch2_dev_attach(struct bch_fs *c, struct bch_dev *ca, unsigned dev_idx) argument
1350 struct bch_dev *ca = NULL; local
1370 __bch2_dev_attach_bdev(struct bch_dev *ca, struct bch_sb_handle *sb) argument
1405 struct bch_dev *ca; local
1449 bch2_dev_state_allowed(struct bch_fs *c, struct bch_dev *ca, enum bch_member_state new_state, int flags) argument
1495 struct bch_dev *ca; local
1527 __bch2_dev_read_only(struct bch_fs *c, struct bch_dev *ca) argument
1536 __bch2_dev_read_write(struct bch_fs *c, struct bch_dev *ca) argument
1546 __bch2_dev_set_state(struct bch_fs *c, struct bch_dev *ca, enum bch_member_state new_state, int flags) argument
1577 bch2_dev_set_state(struct bch_fs *c, struct bch_dev *ca, enum bch_member_state new_state, int flags) argument
1591 bch2_dev_remove_alloc(struct bch_fs *c, struct bch_dev *ca) argument
1617 bch2_dev_remove(struct bch_fs *c, struct bch_dev *ca, int flags) argument
1725 struct bch_dev *ca = NULL; local
1864 struct bch_dev *ca; local
1925 bch2_dev_offline(struct bch_fs *c, struct bch_dev *ca, int flags) argument
1947 bch2_dev_resize(struct bch_fs *c, struct bch_dev *ca, u64 nbuckets) argument
[all...]
H A Dsuper-io.c231 for_each_online_member(c, ca) {
232 struct bch_sb_handle *dev_sb = &ca->disk_sb;
235 percpu_ref_put(&ca->io_ref);
536 for_each_member_device(c, ca) {
537 struct bch_member m = bch2_sb_member_get(src, ca->dev_idx);
538 ca->mi = bch2_mi_to_cpu(&m);
616 int bch2_sb_from_fs(struct bch_fs *c, struct bch_dev *ca) argument
618 return __copy_super(&ca->disk_sb, c->disk_sb.sb);
857 struct bch_dev *ca = bio->bi_private; local
861 if (bch2_dev_io_err_on(bio->bi_status, ca,
874 read_back_super(struct bch_fs *c, struct bch_dev *ca) argument
892 write_one_super(struct bch_fs *c, struct bch_dev *ca, unsigned idx) argument
1017 struct bch_dev *ca = *cap; local
1046 struct bch_dev *ca = *cap; local
1058 struct bch_dev *ca = *cap; local
[all...]
H A Drecovery.c364 struct bch_dev *ca = bch_dev_bkey_exists(c, le32_to_cpu(u->dev)); local
368 ca->usage_base->d[i].buckets = le64_to_cpu(u->d[i].buckets);
369 ca->usage_base->d[i].sectors = le64_to_cpu(u->d[i].sectors);
370 ca->usage_base->d[i].fragmented = le64_to_cpu(u->d[i].fragmented);
952 for_each_member_device(c, ca)
953 bch2_dev_usage_init(ca);
980 for_each_online_member(c, ca)
981 ca->new_fs_bucket_idx = 0;
H A Dsb-members.c369 for_each_member_device_rcu(c, ca, NULL) {
370 struct bch_member *m = __bch2_members_v2_get_mut(mi, ca->dev_idx);
373 m->errors[e] = cpu_to_le64(atomic64_read(&ca->errors[e]));
378 void bch2_dev_io_errors_to_text(struct printbuf *out, struct bch_dev *ca) argument
380 struct bch_fs *c = ca->fs;
383 mutex_lock(&ca->fs->sb_lock);
384 m = bch2_sb_member_get(c->disk_sb.sb, ca->dev_idx);
385 mutex_unlock(&ca->fs->sb_lock);
396 prt_u64(out, atomic64_read(&ca->errors[i]));
410 prt_u64(out, atomic64_read(&ca
416 bch2_dev_errors_reset(struct bch_dev *ca) argument
[all...]
H A Dsb-clean.c239 for_each_member_device(c, ca) {
247 u->dev = cpu_to_le32(ca->dev_idx);
250 u->d[i].buckets = cpu_to_le64(ca->usage_base->d[i].buckets);
251 u->d[i].sectors = cpu_to_le64(ca->usage_base->d[i].sectors);
252 u->d[i].fragmented = cpu_to_le64(ca->usage_base->d[i].fragmented);
H A Dsb-members.h21 static inline bool bch2_dev_is_online(struct bch_dev *ca) argument
23 return !percpu_ref_is_zero(&ca->io_ref);
26 static inline bool bch2_dev_is_readable(struct bch_dev *ca) argument
28 return bch2_dev_is_online(ca) &&
29 ca->mi.state != BCH_MEMBER_STATE_failed;
32 static inline bool bch2_dev_get_ioref(struct bch_dev *ca, int rw) argument
34 if (!percpu_ref_tryget(&ca->io_ref))
37 if (ca->mi.state == BCH_MEMBER_STATE_rw ||
38 (ca->mi.state == BCH_MEMBER_STATE_ro && rw == READ))
41 percpu_ref_put(&ca
86 struct bch_dev *ca = NULL; local
98 __bch2_next_dev(struct bch_fs *c, struct bch_dev *ca, const struct bch_devs_mask *mask) argument
108 bch2_get_next_dev(struct bch_fs *c, struct bch_dev *ca) argument
131 bch2_get_next_online_dev(struct bch_fs *c, struct bch_dev *ca, unsigned state_mask) argument
234 bch2_dev_btree_bitmap_marked_sectors(struct bch_dev *ca, u64 start, unsigned sectors) argument
[all...]
H A Djournal.c921 static int __bch2_set_nr_journal_buckets(struct bch_dev *ca, unsigned nr, argument
924 struct bch_fs *c = ca->fs;
925 struct journal_device *ja = &ca->journal;
945 bu[nr_got] = bch2_bucket_alloc_new_fs(ca);
951 ob[nr_got] = bch2_bucket_alloc(c, ca, BCH_WATERMARK_normal, cl);
957 bch2_trans_mark_metadata_bucket(trans, ca,
959 ca->mi.bucket_size));
1003 ret = bch2_journal_buckets_to_sb(c, ca, new_buckets, nr);
1038 bch2_trans_mark_metadata_bucket(trans, ca,
1056 int bch2_set_nr_journal_buckets(struct bch_fs *c, struct bch_dev *ca, argument
1106 bch2_dev_journal_alloc(struct bch_dev *ca) argument
1171 bch2_dev_journal_stop(struct journal *j, struct bch_dev *ca) argument
1285 bch2_dev_journal_exit(struct bch_dev *ca) argument
1300 bch2_dev_journal_init(struct bch_dev *ca, struct bch_sb *sb) argument
[all...]
H A Dio_write.c39 static inline void bch2_congested_acct(struct bch_dev *ca, u64 io_latency, argument
43 ca->io_latency[rw].quantiles.entries[QUANTILE_IDX(1)].m;
54 if (atomic_read(&ca->congested) < CONGESTED_MAX)
57 &ca->congested);
59 ca->congested_last = now;
60 } else if (atomic_read(&ca->congested) > 0) {
61 atomic_dec(&ca->congested);
65 void bch2_latency_acct(struct bch_dev *ca, u64 submit_time, int rw) argument
67 atomic64_t *latency = &ca->cur_latency[rw];
89 bch2_congested_acct(ca, io_latenc
412 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
653 struct bch_dev *ca = bch_dev_bkey_exists(c, wbio->dev); local
1296 struct bch_dev *ca = bch_dev_bkey_exists(c, i->b.inode); local
[all...]
H A Dfs.c1744 for_each_online_member(c, ca) {
1748 seq_puts(seq, ca->disk_sb.sb_name);
1966 for_each_online_member(c, ca) {
1967 struct block_device *bdev = ca->disk_sb.bdev;
1972 percpu_ref_put(&ca->io_ref);
H A Dbackpointers.h42 struct bch_dev *ca = bch_dev_bkey_exists(c, bp_pos.inode); local
45 return POS(bp_pos.inode, sector_to_bucket(ca, bucket_sector));
48 static inline struct bpos bucket_pos_to_bp_noerror(const struct bch_dev *ca, argument
53 (bucket_to_sector(ca, bucket.offset) <<
64 struct bch_dev *ca = bch_dev_bkey_exists(c, bucket.inode); local
65 struct bpos ret = bucket_pos_to_bp_noerror(ca, bucket, bucket_offset);
H A Dbackpointers.c52 struct bch_dev *ca = bch_dev_bkey_exists(c, bp.k->p.inode); local
56 bkey_fsck_err_on((bp.v->bucket_offset >> MAX_EXTENT_COMPRESS_RATIO_SHIFT) >= ca->mi.bucket_size ||
57 !bpos_eq(bp.k->p, bucket_pos_to_bp_noerror(ca, bucket, bp.v->bucket_offset)),
463 struct bch_dev *ca = bch_dev_bkey_exists(c, dev); local
464 if (!bch2_dev_get_ioref(ca, READ))
473 bio = bio_alloc(ca->disk_sb.bdev, buf_pages(data_buf, bytes), REQ_OP_READ, GFP_KERNEL);
497 percpu_ref_put(&ca->io_ref);
H A Dalloc_background.h18 struct bch_dev *ca; local
23 ca = bch_dev_bkey_exists(c, pos.inode);
24 return pos.offset >= ca->mi.first_bucket &&
25 pos.offset < ca->mi.nbuckets;
84 static inline unsigned bch2_bucket_sectors_fragmented(struct bch_dev *ca, argument
89 return d ? max(0, ca->mi.bucket_size - d) : 0;
108 struct bch_dev *ca)
111 !bch2_bucket_sectors_fragmented(ca, a))
115 return div_u64(d * (1ULL << 31), ca->mi.bucket_size);
241 static inline u64 should_invalidate_buckets(struct bch_dev *ca, argument
107 alloc_lru_idx_fragmentation(struct bch_alloc_v4 a, struct bch_dev *ca) argument
[all...]
H A Dalloc_background.c592 struct bch_dev *ca = bch_dev_bkey_exists(c, k.k->p.inode);
594 for (u64 b = max_t(u64, ca->mi.first_bucket, start);
595 b < min_t(u64, ca->mi.nbuckets, end);
597 *bucket_gen(ca, b) = g->gens[b & KEY_TYPE_BUCKET_GENS_MASK];
610 struct bch_dev *ca = bch_dev_bkey_exists(c, k.k->p.inode);
613 *bucket_gen(ca, k.k->p.offset) = bch2_alloc_to_v4(k, &a)->gen;
633 struct bch_dev *ca = bch_dev_bkey_exists(c, alloc_k.k->p.inode); local
675 if (ca->mi.freespace_initialized &&
746 struct bch_dev *ca = bch_dev_bkey_exists(c, new.k->p.inode); local
863 *bucket_gen(ca, ne
963 struct bch_dev *ca; local
1010 struct bch_dev *ca = bch_dev_bkey_exists(c, bucket.inode); local
1028 struct bch_dev *ca; local
1157 struct bch_dev *ca; local
1340 struct bch_dev *ca; local
1646 struct bch_dev *ca; member in struct:discard_buckets_state
1650 discard_buckets_next_dev(struct bch_fs *c, struct discard_buckets_state *s, struct bch_dev *ca) argument
1676 struct bch_dev *ca; local
1858 struct bch_dev *ca; local
1906 struct bch_dev *ca = bch_dev_bkey_exists(c, bucket.inode); local
2037 bch2_dev_freespace_init(struct bch_fs *c, struct bch_dev *ca, u64 bucket_start, u64 bucket_end) argument
2284 bch2_dev_has_open_write_point(struct bch_fs *c, struct bch_dev *ca) argument
2303 bch2_dev_allocator_remove(struct bch_fs *c, struct bch_dev *ca) argument
2338 bch2_dev_allocator_add(struct bch_fs *c, struct bch_dev *ca) argument
[all...]
H A Dbuckets.c66 for_each_member_device(c, ca) {
67 struct bch_dev_usage dev = bch2_dev_usage_read(ca);
71 ca->mi.bucket_size;
77 static inline struct bch_dev_usage *dev_usage_ptr(struct bch_dev *ca, argument
84 ? ca->usage_gc
85 : ca->usage[journal_seq & JOURNAL_BUF_MASK]);
88 void bch2_dev_usage_read_fast(struct bch_dev *ca, struct bch_dev_usage *usage) argument
90 struct bch_fs *c = ca->fs;
95 memcpy(usage, ca->usage_base, u64s * sizeof(u64));
96 for (i = 0; i < ARRAY_SIZE(ca
270 bch2_dev_usage_init(struct bch_dev *ca) argument
299 bch2_dev_usage_update(struct bch_fs *c, struct bch_dev *ca, const struct bch_alloc_v4 *old, const struct bch_alloc_v4 *new, u64 journal_seq, bool gc) argument
343 bch2_dev_usage_update_m(struct bch_fs *c, struct bch_dev *ca, struct bucket *old, struct bucket *new) argument
499 bch2_mark_metadata_bucket(struct bch_fs *c, struct bch_dev *ca, size_t b, enum bch_data_type data_type, unsigned sectors, struct gc_pos pos, unsigned flags) argument
560 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
853 struct bch_dev *ca = bch_dev_bkey_exists(c, p.ptr.dev); local
1120 __bch2_trans_mark_metadata_bucket(struct btree_trans *trans, struct bch_dev *ca, size_t b, enum bch_data_type type, unsigned sectors) argument
1164 bch2_trans_mark_metadata_bucket(struct btree_trans *trans, struct bch_dev *ca, size_t b, enum bch_data_type type, unsigned sectors) argument
1173 bch2_trans_mark_metadata_sectors(struct btree_trans *trans, struct bch_dev *ca, u64 start, u64 end, enum bch_data_type type, u64 *bucket, unsigned *bucket_sectors) argument
1201 __bch2_trans_mark_dev_sb(struct btree_trans *trans, struct bch_dev *ca) argument
1245 bch2_trans_mark_dev_sb(struct bch_fs *c, struct bch_dev *ca) argument
1342 bch2_dev_buckets_resize(struct bch_fs *c, struct bch_dev *ca, u64 nbuckets) argument
1408 bch2_dev_buckets_free(struct bch_dev *ca) argument
1418 bch2_dev_buckets_alloc(struct bch_fs *c, struct bch_dev *ca) argument
[all...]
H A Dbtree_node_scan.c20 struct bch_dev *ca; member in struct:find_btree_nodes_worker
121 static void try_read_btree_node(struct find_btree_nodes *f, struct bch_dev *ca, argument
126 bio_reset(bio, ca->disk_sb.bdev, REQ_OP_READ);
131 if (bch2_dev_io_err_on(bio->bi_status, ca, BCH_MEMBER_ERROR_read,
163 .ptrs[0].dev = ca->dev_idx,
164 .ptrs[0].gen = *bucket_gen(ca, sector_to_bucket(ca, offset)),
187 struct bch_dev *ca = w->ca; local
198 for (u64 bucket = ca
[all...]
H A Djournal_io.c24 struct bch_dev *ca = bch_dev_bkey_exists(c, i->dev); local
27 div64_u64_rem(i->sector, ca->mi.bucket_size, &offset);
114 static int journal_entry_add(struct bch_fs *c, struct bch_dev *ca, argument
176 if (ptr->dev == ca->dev_idx)
864 struct bch_dev *ca,
879 ca ? ca->name : c->name,
891 ca ? ca->name : c->name,
914 struct bch_dev *ca,
863 jset_validate(struct bch_fs *c, struct bch_dev *ca, struct jset *jset, u64 sector, enum bkey_invalid_flags flags) argument
913 jset_validate_early(struct bch_fs *c, struct bch_dev *ca, struct jset *jset, u64 sector, unsigned bucket_sectors_left, unsigned sectors_read) argument
981 journal_read_bucket(struct bch_dev *ca, struct journal_read_buf *buf, struct journal_list *jlist, unsigned bucket) argument
1124 struct bch_dev *ca = container_of(ja, struct bch_dev, journal); local
1393 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
1446 struct bch_dev *ca; local
1504 struct bch_dev *ca; local
1706 struct bch_dev *ca = jbio->ca; local
1734 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
[all...]
H A Dchardev.c25 /* returns with ref on ca->ref */
29 struct bch_dev *ca; local
36 ca = rcu_dereference(c->devs[dev]);
37 if (ca)
38 percpu_ref_get(&ca->ref);
41 if (!ca)
51 ca = bch2_dev_lookup(c, path);
55 return ca;
334 struct bch_dev *ca; local
346 ca
376 struct bch_dev *ca; local
401 struct bch_dev *ca; local
587 struct bch_dev *ca; local
628 struct bch_dev *ca; local
677 struct bch_dev *ca = NULL; local
740 struct bch_dev *ca; local
763 struct bch_dev *ca; local
[all...]
H A Dbtree_io.c514 struct bch_dev *ca,
522 if (ca)
523 prt_printf(out, "on %s ", ca->name);
537 struct bch_dev *ca,
548 btree_err_msg(&out, c, ca, b, i, b->written, write);
599 #define btree_err(type, c, ca, b, i, _err_type, msg, ...) \
601 int _ret = __btree_err(type, c, ca, b, i, write, have_retry, \
668 static int validate_bset(struct bch_fs *c, struct bch_dev *ca, argument
680 c, ca, b, i,
713 c, ca, local
513 btree_err_msg(struct printbuf *out, struct bch_fs *c, struct bch_dev *ca, struct btree *b, struct bset *i, unsigned offset, int write) argument
535 __btree_err(int ret, struct bch_fs *c, struct bch_dev *ca, struct btree *b, struct bset *i, int write, bool have_retry, enum bch_sb_error_id err_type, const char *fmt, ...) argument
735 c, ca, b, i, local
758 c, ca, b, i, local
764 c, ca, b, i, local
806 c, ca, b, i, local
975 bch2_btree_node_read_done(struct bch_fs *c, struct bch_dev *ca, struct btree *b, bool have_retry, bool *saw_error) argument
1007 c, ca, b, NULL, local
1283 struct bch_dev *ca = bch_dev_bkey_exists(c, rb->pick.ptr.dev); local
1366 struct bch_dev *ca = bch_dev_bkey_exists(c, rb->pick.ptr.dev); local
1563 struct bch_dev *ca = bch_dev_bkey_exists(c, rb->pick.ptr.dev); local
1605 struct bch_dev *ca = bch_dev_bkey_exists(c, pick.ptr.dev); local
1650 struct bch_dev *ca; local
1899 struct bch_dev *ca = bch_dev_bkey_exists(c, wbio->dev); local
[all...]
H A Dbtree_gc.c606 struct bch_dev *ca = bch_dev_bkey_exists(c, p.ptr.dev); local
607 struct bucket *g = PTR_GC_BUCKET(ca, &p.ptr);
614 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr),
631 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr),
652 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr), g->gen,
663 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr),
679 p.ptr.dev, PTR_BUCKET_NR(ca, &p.ptr),
739 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
740 struct bucket *g = PTR_GC_BUCKET(ca, ptr);
750 struct bch_dev *ca local
1117 mark_metadata_sectors(struct bch_fs *c, struct bch_dev *ca, u64 start, u64 end, enum bch_data_type type, unsigned flags) argument
1135 bch2_mark_dev_superblock(struct bch_fs *c, struct bch_dev *ca, unsigned flags) argument
1210 struct bch_dev *ca = NULL; local
1368 struct bch_dev *ca = bch_dev_bkey_exists(c, iter->pos.inode); local
1857 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
1866 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
1887 struct bch_dev *ca = bch_dev_bkey_exists(trans->c, iter->pos.inode); local
[all...]
H A Dsysfs.c795 static void dev_alloc_debug_to_text(struct printbuf *out, struct bch_dev *ca) argument
797 struct bch_fs *c = ca->fs;
798 struct bch_dev_usage stats = bch2_dev_usage_read(ca);
821 prt_u64(out, bch2_dev_buckets_reserved(ca, i));
843 prt_u64(out, ca->nr_open_buckets);
868 prt_u64(out, should_invalidate_buckets(ca, stats));
883 static void dev_io_done_to_text(struct printbuf *out, struct bch_dev *ca) argument
893 percpu_u64_get(&ca->io_done->sectors[rw][i]) << 9);
899 struct bch_dev *ca = container_of(kobj, struct bch_dev, kobj); local
900 struct bch_fs *c = ca
953 struct bch_dev *ca = container_of(kobj, struct bch_dev, kobj); local
[all...]
H A Dfs-io.c41 struct bch_dev *ca; member in struct:nocow_flush
51 percpu_ref_put(&bio->ca->io_ref);
60 struct bch_dev *ca; local
73 ca = rcu_dereference(c->devs[dev]);
74 if (ca && !percpu_ref_tryget(&ca->io_ref))
75 ca = NULL;
78 if (!ca)
81 bio = container_of(bio_alloc_bioset(ca->disk_sb.bdev, 0,
87 bio->ca
[all...]
H A Dec.c101 struct bch_dev *ca; member in struct:ec_bio
256 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
266 g = PTR_GC_BUCKET(ca, ptr);
272 ptr->dev, PTR_BUCKET_NR(ca, ptr), g->gen,
296 bch2_dev_usage_update_m(c, ca, &old, &new);
612 struct bch_dev *ca = bch_dev_bkey_exists(c, v->ptrs[i].dev); local
618 bch_err_ratelimited(ca, "%s", err.buf);
623 bch2_io_error(ca, BCH_MEMBER_ERROR_checksum);
678 struct bch_dev *ca = ec_bio->ca; local
708 struct bch_dev *ca = bch_dev_bkey_exists(c, ptr->dev); local
1324 struct bch_dev *ca = bch_dev_bkey_exists(c, ob->dev); local
1522 struct bch_dev *ca; local
2067 __bch2_ec_stop(struct bch_fs *c, struct bch_dev *ca) argument
2100 bch2_ec_stop_dev(struct bch_fs *c, struct bch_dev *ca) argument
[all...]
/linux-master/drivers/gpu/drm/panel/
H A Dpanel-ilitek-ili9341.c122 /* ca: TODO: need comments for this register */
123 u8 ca[ILI9341_CA_LEN]; member in struct:ili9341_config
209 .ca = {0xc3, 0x08, 0x50},
308 mipi_dbi_command_stackbuf(dbi, 0xca, cfg->ca, ILI9341_CA_LEN);
/linux-master/fs/nfsd/
H A Dnfs4xdr.c1663 struct nfsd4_channel_attrs *ca)
1673 ca->maxreq_sz = be32_to_cpup(p++);
1674 ca->maxresp_sz = be32_to_cpup(p++);
1675 ca->maxresp_cached = be32_to_cpup(p++);
1676 ca->maxops = be32_to_cpup(p++);
1677 ca->maxreqs = be32_to_cpup(p++);
1678 ca->nr_rdma_attrs = be32_to_cpup(p);
1679 switch (ca->nr_rdma_attrs) {
1683 if (xdr_stream_decode_u32(argp->xdr, &ca->rdma_attrs) < 0)
1662 nfsd4_decode_channel_attrs4(struct nfsd4_compoundargs *argp, struct nfsd4_channel_attrs *ca) argument

Completed in 319 milliseconds

1234567