/linux-master/fs/jfs/ |
H A D | jfs_discard.c | 20 * FUNCTION: TRIM the specified block range on device, if supported 59 * range - the range, given by user space 65 int jfs_ioc_trim(struct inode *ip, struct fstrim_range *range) argument 80 start = range->start >> sb->s_blocksize_bits; 81 end = start + (range->len >> sb->s_blocksize_bits) - 1; 82 minlen = range->minlen >> sb->s_blocksize_bits; 88 range->len < sb->s_blocksize) 95 * we trim all ag's within the range 103 range [all...] |
H A D | ioctl.c | 113 struct fstrim_range range; local 124 if (copy_from_user(&range, (struct fstrim_range __user *)arg, 125 sizeof(range))) 128 range.minlen = max_t(unsigned int, range.minlen, 131 ret = jfs_ioc_trim(inode, &range); 135 if (copy_to_user((struct fstrim_range __user *)arg, &range, 136 sizeof(range)))
|
/linux-master/drivers/base/ |
H A D | map.c | 23 unsigned long range; member in struct:kobj_map::probe 32 int kobj_map(struct kobj_map *domain, dev_t dev, unsigned long range, argument 36 unsigned int n = MAJOR(dev + range - 1) - MAJOR(dev) + 1; 53 p->range = range; 59 while (*s && (*s)->range < range) 68 void kobj_unmap(struct kobj_map *domain, dev_t dev, unsigned long range) argument 70 unsigned int n = MAJOR(dev + range - 1) - MAJOR(dev) + 1; 83 if (p->dev == dev && p->range [all...] |
/linux-master/mm/ |
H A D | hmm.c | 32 struct hmm_range *range; member in struct:hmm_vma_walk 43 struct hmm_range *range, unsigned long cpu_flags) 45 unsigned long i = (addr - range->start) >> PAGE_SHIFT; 48 range->hmm_pfns[i] = cpu_flags; 53 * hmm_vma_fault() - fault in a range lacking valid pmd or pte(s) 54 * @addr: range virtual start address (inclusive) 55 * @end: range virtual end address (exclusive) 61 * or whenever there is no page directory covering the virtual address range. 90 struct hmm_range *range = hmm_vma_walk->range; local 42 hmm_pfns_fill(unsigned long addr, unsigned long end, struct hmm_range *range, unsigned long cpu_flags) argument 125 struct hmm_range *range = hmm_vma_walk->range; local 151 struct hmm_range *range = hmm_vma_walk->range; local 176 pmd_to_hmm_pfn_flags(struct hmm_range *range, pmd_t pmd) argument 192 struct hmm_range *range = hmm_vma_walk->range; local 215 pte_to_hmm_pfn_flags(struct hmm_range *range, pte_t pte) argument 228 struct hmm_range *range = hmm_vma_walk->range; local 326 struct hmm_range *range = hmm_vma_walk->range; local 401 pud_to_hmm_pfn_flags(struct hmm_range *range, pud_t pud) argument 415 struct hmm_range *range = hmm_vma_walk->range; local 479 struct hmm_range *range = hmm_vma_walk->range; local 529 struct hmm_range *range = hmm_vma_walk->range; local 587 hmm_range_fault(struct hmm_range *range) argument [all...] |
/linux-master/drivers/gpu/drm/logicvc/ |
H A D | logicvc_of.c | 42 .range = { 50 .range = { 57 .range = { 8, 24 }, 79 .range = { 8, 24 }, 84 .range = { 92 .range = { 158 if (property->range[0] || property->range[1]) 159 if (value < property->range[0] || value > property->range[ [all...] |
/linux-master/net/netfilter/ |
H A D | xt_nat.c | 55 struct nf_nat_range2 range; local 64 xt_nat_convert_range(&range, &mr->range[0]); 65 return nf_nat_setup_info(ct, &range, NF_NAT_MANIP_SRC); 72 struct nf_nat_range2 range; local 80 xt_nat_convert_range(&range, &mr->range[0]); 81 return nf_nat_setup_info(ct, &range, NF_NAT_MANIP_DST); 88 struct nf_nat_range2 range; local 97 memcpy(&range, range_v 107 struct nf_nat_range2 range; local 124 const struct nf_nat_range2 *range = par->targinfo; local 139 const struct nf_nat_range2 *range = par->targinfo; local [all...] |
H A D | xt_REDIRECT.c | 36 const struct nf_nat_range2 *range = par->targinfo; local 38 if (range->flags & NF_NAT_RANGE_MAP_IPS) 53 if (mr->range[0].flags & NF_NAT_RANGE_MAP_IPS) { 68 struct nf_nat_range2 range = { local 69 .flags = mr->range[0].flags, 70 .min_proto = mr->range[0].min, 71 .max_proto = mr->range[0].max, 74 return nf_nat_redirect_ipv4(skb, &range, xt_hooknum(par));
|
H A D | nf_nat_redirect.c | 29 nf_nat_redirect(struct sk_buff *skb, const struct nf_nat_range2 *range, argument 40 newrange.flags = range->flags | NF_NAT_RANGE_MAP_IPS; 43 newrange.min_proto = range->min_proto; 44 newrange.max_proto = range->max_proto; 50 nf_nat_redirect_ipv4(struct sk_buff *skb, const struct nf_nat_range2 *range, argument 77 return nf_nat_redirect(skb, range, &newdst); 104 nf_nat_redirect_ipv6(struct sk_buff *skb, const struct nf_nat_range2 *range, argument 136 return nf_nat_redirect(skb, range, &newdst);
|
H A D | xt_NETMAP.c | 21 const struct nf_nat_range2 *range = par->targinfo; local 29 for (i = 0; i < ARRAY_SIZE(range->min_addr.ip6); i++) 30 netmask.ip6[i] = ~(range->min_addr.ip6[i] ^ 31 range->max_addr.ip6[i]); 41 new_addr.ip6[i] |= range->min_addr.ip6[i] & 45 newrange.flags = range->flags | NF_NAT_RANGE_MAP_IPS; 48 newrange.min_proto = range->min_proto; 49 newrange.max_proto = range->max_proto; 56 const struct nf_nat_range2 *range = par->targinfo; local 58 if (!(range [all...] |
H A D | nf_nat_core.c | 286 const struct nf_nat_range2 *range) 289 return ntohl(t->src.u3.ip) >= ntohl(range->min_addr.ip) && 290 ntohl(t->src.u3.ip) <= ntohl(range->max_addr.ip); 292 return ipv6_addr_cmp(&t->src.u3.in6, &range->min_addr.in6) >= 0 && 293 ipv6_addr_cmp(&t->src.u3.in6, &range->max_addr.in6) <= 0; 328 * that meet the constraints of range. 331 const struct nf_nat_range2 *range) 334 * range specified, otherwise let this drag us onto a new src IP. 336 if (range->flags & NF_NAT_RANGE_MAP_IPS && 337 !nf_nat_inet_in_range(tuple, range)) 285 nf_nat_inet_in_range(const struct nf_conntrack_tuple *t, const struct nf_nat_range2 *range) argument 330 nf_in_range(const struct nf_conntrack_tuple *tuple, const struct nf_nat_range2 *range) argument 361 find_appropriate_src(struct net *net, const struct nf_conntrack_zone *zone, const struct nf_conntrack_tuple *tuple, struct nf_conntrack_tuple *result, const struct nf_nat_range2 *range) argument 393 find_best_ips_proto(const struct nf_conntrack_zone *zone, struct nf_conntrack_tuple *tuple, const struct nf_nat_range2 *range, const struct nf_conn *ct, enum nf_nat_manip_type maniptype) argument 465 nf_nat_l4proto_unique_tuple(struct nf_conntrack_tuple *tuple, const struct nf_nat_range2 *range, enum nf_nat_manip_type maniptype, const struct nf_conn *ct) argument 591 get_unique_tuple(struct nf_conntrack_tuple *tuple, const struct nf_conntrack_tuple *orig_tuple, const struct nf_nat_range2 *range, struct nf_conn *ct, enum nf_nat_manip_type maniptype) argument 667 nf_nat_setup_info(struct nf_conn *ct, const struct nf_nat_range2 *range, enum nf_nat_manip_type maniptype) argument 746 struct nf_nat_range2 range = { local 931 nf_nat_l4proto_nlattr_to_range(struct nlattr *tb[], struct nf_nat_range2 *range) argument 946 nfnetlink_parse_nat_proto(struct nlattr *attr, const struct nf_conn *ct, struct nf_nat_range2 *range) argument 969 nf_nat_ipv4_nlattr_to_range(struct nlattr *tb[], struct nf_nat_range2 *range) argument 985 nf_nat_ipv6_nlattr_to_range(struct nlattr *tb[], struct nf_nat_range2 *range) argument 1004 nfnetlink_parse_nat(const struct nlattr *nat, const struct nf_conn *ct, struct nf_nat_range2 *range) argument 1044 struct nf_nat_range2 range; local [all...] |
H A D | nf_nat_ovs.c | 7 * range is only used for new, uninitialized NAT state. 12 const struct nf_nat_range2 *range, 57 err = (range && range->flags & NF_NAT_RANGE_MAP_IPS) 61 ? nf_nat_setup_info(ct, range, maniptype) 87 const struct nf_nat_range2 *range, bool commit) 119 err = nf_ct_nat_execute(skb, ct, ctinfo, action, range, maniptype); 127 err = nf_ct_nat_execute(skb, ct, ctinfo, action, range, 10 nf_ct_nat_execute(struct sk_buff *skb, struct nf_conn *ct, enum ip_conntrack_info ctinfo, int *action, const struct nf_nat_range2 *range, enum nf_nat_manip_type maniptype) argument 85 nf_ct_nat(struct sk_buff *skb, struct nf_conn *ct, enum ip_conntrack_info ctinfo, int *action, const struct nf_nat_range2 *range, bool commit) argument
|
/linux-master/tools/testing/selftests/net/ |
H A D | ip_local_port_range.c | 28 static void unpack_port_range(__u32 range, __u16 *lo, __u16 *hi) argument 30 *lo = range & 0xffff; 31 *hi = range >> 16; 104 static int get_ip_local_port_range(int fd, __u32 *range) argument 115 *range = val; 199 /* Empty range: low port > high port */ 221 { 30000, 39999 }, /* socket range below netns range */ 222 { 50000, 59999 }, /* socket range above netns range */ 238 __u32 range; local 281 __u32 range; local 308 __u32 range; local 365 __u32 range; local 420 __u32 range; local [all...] |
/linux-master/arch/s390/boot/ |
H A D | physmem_info.c | 38 * are merged together into single memory range. 42 struct physmem_range *range; local 45 range = __get_physmem_range_ptr(physmem_info.range_count - 1); 46 if (range->end == start) { 47 range->end = end; 52 range = __get_physmem_range_ptr(physmem_info.range_count); 53 range->start = start; 54 range->end = end; 143 unsigned long range = 1 << (MAX_PHYSMEM_BITS - 20); /* in 1MB blocks */ local 147 while (range > 189 struct reserved_range *range; local 298 struct reserved_range *range = &physmem_info.reserved[type]; local [all...] |
/linux-master/drivers/pci/hotplug/ |
H A D | ibmphp_res.c | 116 /* need to insert our range */ 164 * 2. If cannot allocate out of PFMem range, allocate from Mem ranges. PFmemFromMem 294 * assign a -1 and then update once the range 336 * range actually appears... 361 * This function adds a range into a sorted list of ranges per bus for a particular 362 * range type, it then calls another routine to update the range numbers on the 365 * Input: type of the resource, range to add, current bus 366 * Output: 0 or -1, bus and range ptrs 368 static int add_bus_range(int type, struct range_node *range, struc argument 487 fix_me(struct resource_node *res, struct bus_node *bus_cur, struct range_node *range) argument 544 struct range_node *range; local 919 struct range_node *range = NULL; local 955 struct range_node *range = NULL; local 1757 struct range_node *range; local 1865 range_exists_already(struct range_node *range, struct bus_node *bus_cur, u8 type) argument 1916 struct range_node *range; local [all...] |
/linux-master/drivers/of/ |
H A D | address.c | 45 u64 (*map)(__be32 *addr, const __be32 *range, 65 static u64 of_bus_default_map(__be32 *addr, const __be32 *range, argument 70 cp = of_read_number(range + fna, na - fna); 71 s = of_read_number(range + na + pna, ns); 103 static u64 of_bus_default_flags_map(__be32 *addr, const __be32 *range, int na, argument 107 if (*addr != *range) 110 return of_bus_default_map(addr, range, na, ns, pna, fna); 183 static u64 of_bus_pci_map(__be32 *addr, const __be32 *range, int na, int ns, argument 189 rf = of_bus_pci_get_flags(range); 195 return of_bus_default_map(addr, range, n 215 of_pci_range_to_resource(struct of_pci_range *range, struct device_node *np, struct resource *res) argument 268 struct of_range range; local 300 of_bus_isa_map(__be32 *addr, const __be32 *range, int na, int ns, int pna, int fna) argument 791 of_pci_range_parser_one(struct of_pci_range_parser *parser, struct of_pci_range *range) argument 895 struct of_range range; local 980 struct of_range range; local [all...] |
/linux-master/drivers/gpu/drm/sprd/ |
H A D | megacores_pll.c | 221 u32 range[2], constant; local 236 range[L] = 50 * scale; 237 range[H] = INFINITY; 238 val[CLK] = DIV_ROUND_UP(range[L] * (factor << 1), t_byteck) - 2; 243 range[L] = 38 * scale; 244 range[H] = 95 * scale; 245 tmp = AVERAGE(range[L], range[H]); 246 val[CLK] = DIV_ROUND_UP(AVERAGE(range[L], range[ [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_port_range.c | 13 struct mlxsw_sp_port_range range; member in struct:mlxsw_sp_port_range_reg 36 mlxsw_reg_pprr_src_set(pprr_pl, prr->range.source); 37 mlxsw_reg_pprr_dst_set(pprr_pl, !prr->range.source); 40 mlxsw_reg_pprr_port_range_min_set(pprr_pl, prr->range.min); 41 mlxsw_reg_pprr_port_range_max_set(pprr_pl, prr->range.max); 48 const struct mlxsw_sp_port_range *range, 59 prr->range = *range; 66 NL_SET_ERR_MSG_MOD(extack, "Exceeded number of port range registers"); 72 NL_SET_ERR_MSG_MOD(extack, "Failed to configure port range registe 47 mlxsw_sp_port_range_reg_create(struct mlxsw_sp *mlxsw_sp, const struct mlxsw_sp_port_range *range, struct netlink_ext_ack *extack) argument 98 mlxsw_sp_port_range_reg_find(struct mlxsw_sp *mlxsw_sp, const struct mlxsw_sp_port_range *range) argument 115 mlxsw_sp_port_range_reg_get(struct mlxsw_sp *mlxsw_sp, const struct mlxsw_sp_port_range *range, struct netlink_ext_ack *extack, u8 *p_prr_index) argument [all...] |
/linux-master/tools/arch/x86/kcpuid/ |
H A D | kcpuid.c | 74 * basic: basic functions range: [0... ] 75 * ext: extended functions range: [0x80000000... ] 139 static bool cpuid_store(struct cpuid_range *range, u32 f, int subleaf, argument 153 func = &range->funcs[f & 0xffff]; 182 static void raw_dump_range(struct cpuid_range *range) argument 187 printf("%s Leafs :\n", range->is_ext ? "Extended" : "Basic"); 190 for (f = 0; (int)f < range->nr; f++) { 191 struct cpuid_func *func = &range->funcs[f]; 194 if (range->is_ext) 212 struct cpuid_range *range; local 295 struct cpuid_range *range; local 509 show_range(struct cpuid_range *range) argument 519 struct cpuid_range *range; local [all...] |
/linux-master/arch/x86/platform/efi/ |
H A D | fake_mem.c | 10 * specific memory range by updating original (firmware provided) EFI 33 if (m1->range.start < m2->range.start) 35 if (m1->range.start > m2->range.start) 49 new_nr_map += efi_memmap_split_count(md, &efi_range->range); 107 efi_fake_mems[nr_fake_mem].range.start = start; 108 efi_fake_mems[nr_fake_mem].range.end = start + mem_size - 1; 121 efi_fake_mems[i].attribute, efi_fake_mems[i].range.start, 122 efi_fake_mems[i].range [all...] |
/linux-master/security/selinux/ss/ |
H A D | mls.c | 44 u32 index_sens = context->range.level[l].sens; 50 e = &context->range.level[l].cat; 70 if (mls_level_eq(&context->range.level[0], 71 &context->range.level[1])) 104 context->range.level[l].sens - 1)); 110 e = &context->range.level[l].cat; 147 if (mls_level_eq(&context->range.level[0], 148 &context->range.level[1])) 196 if (!mls_range_isvalid(p, &c->range)) 203 * User must be authorized for the MLS range 379 mls_range_set(struct context *context, struct mls_range *range) argument [all...] |
H A D | context.c | 30 hash = mls_range_hash(&c->range, hash);
|
/linux-master/arch/x86/pci/ |
H A D | amd_bus.c | 6 #include <linux/range.h> 81 struct range range[RANGE_NUM]; local 128 /* Check if that register is enabled for bus range */ 157 memset(range, 0, sizeof(range)); 158 add_range(range, RANGE_NUM, 0, 0, 0xffff + 1); 182 subtract_range(range, RANGE_NUM, start, end + 1); 184 /* add left over io port range to def node/link, [0, 0xffff] */ 189 if (!range[ [all...] |
/linux-master/include/linux/ |
H A D | bma150.h | 36 unsigned char range; /* one of BMA150_RANGE_xxx */ member in struct:bma150_cfg
|
/linux-master/drivers/nvdimm/ |
H A D | badrange.c | 167 dev_dbg(bb->dev, "Found a bad range (0x%llx, 0x%llx)\n", 176 * __add_badblock_range() - Convert a physical address range to bad sectors 178 * @ns_offset: namespace offset where the error range begins (in bytes) 181 * This assumes that the range provided with (ns_offset, len) is within 214 struct badblocks *bb, const struct range *range) 225 if (bre_end < range->start) 227 if (bre->start > range->end) 230 if (bre->start >= range->start) { 234 if (bre_end <= range 213 badblocks_populate(struct badrange *badrange, struct badblocks *bb, const struct range *range) argument 269 nvdimm_badblocks_populate(struct nd_region *nd_region, struct badblocks *bb, const struct range *range) argument [all...] |
/linux-master/drivers/soc/ti/ |
H A D | knav_qmss_queue.c | 114 static int knav_queue_setup_irq(struct knav_range_info *range, argument 117 unsigned queue = inst->id - range->queue_base; 120 if (range->flags & RANGE_HAS_IRQ) { 121 irq = range->irqs[queue].irq; 127 if (range->irqs[queue].cpu_mask) { 128 ret = irq_set_affinity_hint(irq, range->irqs[queue].cpu_mask); 130 dev_warn(range->kdev->dev, 141 struct knav_range_info *range = inst->range; local 142 unsigned queue = inst->id - inst->range 243 struct knav_range_info *range = inst->range; local 319 struct knav_range_info *range = inst->range; local 385 knav_gp_set_notify(struct knav_range_info *range, struct knav_queue_inst *inst, bool enabled) argument 401 knav_gp_open_queue(struct knav_range_info *range, struct knav_queue_inst *inst, unsigned flags) argument 407 knav_gp_close_queue(struct knav_range_info *range, struct knav_queue_inst *inst) argument 565 struct knav_range_info *range = inst->range; local 1215 struct knav_range_info *range; local 1311 struct device_node *type, *range; local 1328 knav_free_queue_range(struct knav_device *kdev, struct knav_range_info *range) argument 1339 struct knav_range_info *range; local 1683 knav_queue_init_queue(struct knav_device *kdev, struct knav_range_info *range, struct knav_queue_inst *inst, unsigned id) argument 1709 struct knav_range_info *range; local [all...] |