/linux-master/drivers/gpu/drm/nouveau/include/nvif/ |
H A D | if0004.h | 7 } vn; member in union:nv04_nvsw_event_args
|
H A D | if000e.h | 19 } vn; member in union:nvif_event_allow_args 24 } vn; member in union:nvif_event_block_args
|
H A D | clb069.h | 13 } vn; member in union:nvif_clb069_event_args
|
H A D | if0013.h | 15 } vn; member in union:nvif_head_event_args
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | memnv04.c | 35 struct nv04_mem_map_vn vn; member in union:__anon3249 41 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) 55 struct nv04_mem_vn vn; member in union:__anon3250 59 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn)))
|
H A D | memgf100.c | 39 struct gf100_mem_map_vn vn; member in union:__anon3247 50 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { 73 struct gf100_mem_vn vn; member in union:__anon3248 82 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
|
H A D | memnv50.c | 39 struct nv50_mem_map_vn vn; member in union:__anon3251 52 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { 70 struct nv50_mem_vn vn; member in union:__anon3252 80 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
|
H A D | vmmnv04.c | 84 struct nv04_vmm_map_vn vn; member in union:__anon3278 87 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) 108 struct nv04_vmm_vn vn; member in union:__anon3279 117 return nvif_unvers(-ENOSYS, &argv, &argc, args->vn);
|
H A D | vmmgm200.c | 150 struct gm200_vmm_vn vn; member in union:__anon3273 163 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
|
H A D | vmmgp100.c | 411 struct gp100_vmm_map_vn vn; member in union:__anon796 429 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { 524 struct gp100_vmm_fault_replay_vn vn; member in union:__anon798 528 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { 606 struct gp100_vmm_vn vn; member in union:__anon799 615 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
|
/linux-master/net/8021q/ |
H A D | vlanproc.c | 94 struct vlan_net *vn = net_generic(net, vlan_net_id); local 96 if (vn->proc_vlan_conf) 97 remove_proc_entry(name_conf, vn->proc_vlan_dir); 99 if (vn->proc_vlan_dir) 113 struct vlan_net *vn = net_generic(net, vlan_net_id); local 115 vn->proc_vlan_dir = proc_net_mkdir(net, name_root, net->proc_net); 116 if (!vn->proc_vlan_dir) 119 vn->proc_vlan_conf = proc_create_net(name_conf, S_IFREG | 0600, 120 vn->proc_vlan_dir, &vlan_seq_ops, 122 if (!vn 139 struct vlan_net *vn = net_generic(dev_net(vlandev), vlan_net_id); local 205 struct vlan_net *vn = net_generic(net, vlan_net_id); local [all...] |
H A D | vlan.c | 219 struct vlan_net *vn = net_generic(net, vlan_net_id); local 232 switch (vn->name_type) { 607 struct vlan_net *vn; local 609 vn = net_generic(net, vlan_net_id); 610 vn->name_type = args.u.name_type; 660 struct vlan_net *vn = net_generic(net, vlan_net_id); local 663 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
|
/linux-master/fs/reiserfs/ |
H A D | fix_node.c | 54 struct virtual_node *vn = tb->tb_vn; local 61 vn->vn_size = 66 vn->vn_nr_item = (vn->vn_size - DC_SIZE) / (DC_SIZE + KEY_SIZE); 71 vn->vn_nr_item = 72 B_NR_ITEMS(Sh) + ((vn->vn_mode == M_INSERT) ? 1 : 0) - 73 ((vn->vn_mode == M_DELETE) ? 1 : 0); 76 vn->vn_vi = (struct virtual_item *)(tb->tb_vn + 1); 77 memset(vn->vn_vi, 0, vn 197 struct virtual_node *vn = tb->tb_vn; local 283 struct virtual_node *vn = tb->tb_vn; local 380 struct virtual_node *vn = tb->tb_vn; local 654 struct virtual_node *vn = tb->tb_vn; local 698 struct virtual_node *vn = tb->tb_vn; local 1326 struct virtual_node *vn = tb->tb_vn; local 1728 struct virtual_node *vn = tb->tb_vn; local 1927 struct virtual_node *vn = tb->tb_vn; local 2068 struct virtual_node *vn; local [all...] |
H A D | item_ops.c | 60 static int sd_create_vi(struct virtual_node *vn, argument 150 static int direct_create_vi(struct virtual_node *vn, argument 288 static int indirect_create_vi(struct virtual_node *vn, argument 458 static int direntry_create_vi(struct virtual_node *vn, argument 479 ((is_affected) ? ((vn->vn_mode == M_CUT) ? -1 : 480 (vn->vn_mode == M_PASTE ? 1 : 0)) : 0); 483 j = old_entry_num(is_affected, i, vn->vn_pos_in_item, 484 vn->vn_mode); 493 if (is_affected && vn->vn_mode == M_PASTE) 494 dir_u->entry_sizes[vn 669 errcatch_create_vi(struct virtual_node *vn, struct virtual_item *vi, int is_affected, int insert_size) argument [all...] |
/linux-master/mm/ |
H A D | vmalloc.c | 1050 struct vmap_node *vn; local 1055 vn = &vmap_nodes[i]; 1057 spin_lock(&vn->busy.lock); 1058 *va = __find_vmap_area_exceed_addr(addr, &vn->busy.root); 1063 spin_unlock(&vn->busy.lock); 1072 vn = addr_to_node(va_start_lowest); 1074 spin_lock(&vn->busy.lock); 1075 *va = __find_vmap_area(va_start_lowest, &vn->busy.root); 1078 return vn; 1080 spin_unlock(&vn 1791 struct vmap_node *vn = addr_to_node(va->va_start); local 1832 size_to_va_pool(struct vmap_node *vn, unsigned long size) argument 1860 node_pool_del_va(struct vmap_node *vn, unsigned long size, unsigned long align, unsigned long vstart, unsigned long vend) argument 1939 struct vmap_node *vn; local 2107 decay_va_pool_node(struct vmap_node *vn, bool full_decay) argument 2167 struct vmap_node *vn = container_of(work, local 2209 struct vmap_node *vn; local 2313 struct vmap_node *vn; local 2355 struct vmap_node *vn; local 2392 struct vmap_node *vn; local 2615 struct vmap_node *vn; local 3050 struct vmap_node *vn = addr_to_node(va->va_start); local 4242 struct vmap_node *vn; local 4671 struct vmap_node *vn = addr_to_node(vas[area]->va_start); local 4790 struct vmap_node *vn; local 4846 struct vmap_node *vn; local 4865 struct vmap_node *vn; local 4992 struct vmap_node *vn; local 5048 struct vmap_node *vn; local 5076 struct vmap_node *vn; local [all...] |
/linux-master/scripts/gdb/linux/ |
H A D | vmalloc.py | 34 vn = gdb.parse_and_eval('&vmap_nodes[%d]' % i) 35 for vmap_area in lists.list_for_each_entry(vn['busy']['head'], vmap_area_ptr_type, "list"):
|
/linux-master/drivers/net/vxlan/ |
H A D | vxlan_multicast.c | 125 bool vxlan_group_used(struct vxlan_net *vn, struct vxlan_dev *dev, argument 151 list_for_each_entry(vxlan, &vn->vxlan_list, next) { 218 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local 225 !vxlan_group_used(vn, vxlan, v->vni, &v->remote_ip, 257 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local 261 !vxlan_group_used(vn, vxlan, 0, NULL, 0)) {
|
H A D | vxlan_vnifilter.c | 43 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local 47 spin_lock(&vn->sock_lock); 71 spin_unlock(&vn->sock_lock); 78 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local 86 spin_lock(&vn->sock_lock); 97 spin_unlock(&vn->sock_lock); 103 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local 109 spin_lock(&vn->sock_lock); 116 spin_unlock(&vn->sock_lock); 521 struct vxlan_net *vn local 618 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local [all...] |
H A D | vxlan_private.h | 54 struct vxlan_net *vn = net_generic(net, vxlan_net_id); local 56 return &vn->sock_list[hash_32(ntohs(port), PORT_HASH_BITS)]; 224 bool vxlan_group_used(struct vxlan_net *vn, struct vxlan_dev *dev,
|
H A D | vxlan_core.c | 1502 struct vxlan_net *vn; local 1509 vn = net_generic(sock_net(vs->sock->sk), vxlan_net_id); 1510 spin_lock(&vn->sock_lock); 1516 spin_unlock(&vn->sock_lock); 2813 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local 2815 spin_lock(&vn->sock_lock); 2820 spin_unlock(&vn->sock_lock); 2826 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local 2830 spin_lock(&vn->sock_lock); 2832 spin_unlock(&vn 3257 struct vxlan_net *vn = net_generic(net, vxlan_net_id); local 3507 struct vxlan_net *vn = net_generic(net, vxlan_net_id); local 3560 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); local 3632 struct vxlan_net *vn = net_generic(src_net, vxlan_net_id); local 3890 struct vxlan_net *vn = net_generic(net, vxlan_net_id); local 4597 vxlan_handle_lowerdev_unregister(struct vxlan_net *vn, struct net_device *dev) argument 4623 struct vxlan_net *vn = net_generic(dev_net(dev), vxlan_net_id); local 4805 struct vxlan_net *vn = net_generic(net, vxlan_net_id); local 4819 vxlan_destroy_tunnels(struct vxlan_net *vn, struct list_head *dev_to_kill) argument 4835 struct vxlan_net *vn = net_generic(net, vxlan_net_id); local 4845 struct vxlan_net *vn = net_generic(net, vxlan_net_id); local [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/sw/ |
H A D | nvsw.c | 36 if (argc != sizeof(args->vn))
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | uevent.c | 41 if (argc != sizeof(args->vn)) 52 if (argc != sizeof(args->vn))
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem_evict.c | 478 struct i915_vma *vma, *vn; local 515 list_for_each_entry_safe(vma, vn, &locked_eviction_list, evict_link) { 527 list_for_each_entry_safe(vma, vn, &eviction_list, evict_link) {
|
/linux-master/drivers/usb/gadget/function/ |
H A D | f_mass_storage.h | 137 void fsg_common_set_inquiry_string(struct fsg_common *common, const char *vn,
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/fault/ |
H A D | user.c | 39 if (argc != sizeof(args->vn))
|