/linux-master/include/linux/ |
H A D | part_stat.h | 20 * {disk|part|all}_stat_{add|sub|inc|dec}() modify the stat counters and should 28 #define part_stat_get_cpu(part, field, cpu) \ 29 (per_cpu_ptr((part)->bd_stats, (cpu))->field) 31 #define part_stat_get(part, field) \ 32 part_stat_get_cpu(part, field, smp_processor_id()) 34 #define part_stat_read(part, field) \ 36 typeof((part)->bd_stats->field) res = 0; \ 39 res += per_cpu_ptr((part)->bd_stats, _cpu)->field; \ 43 static inline void part_stat_set_all(struct block_device *part, int value) argument 48 memset(per_cpu_ptr(part [all...] |
H A D | bcm963xx_nvram.h | 51 #define BCM963XX_NVRAM_NAND_PART_OFFSET(nvram, part) \ 52 bcm963xx_nvram_nand_part_offset(nvram, BCM963XX_NVRAM_NAND_PART_ ##part) 56 enum bcm963xx_nvram_nand_part part) 58 return nvram->nand_part_offset[part] * SZ_1K; 61 #define BCM963XX_NVRAM_NAND_PART_SIZE(nvram, part) \ 62 bcm963xx_nvram_nand_part_size(nvram, BCM963XX_NVRAM_NAND_PART_ ##part) 66 enum bcm963xx_nvram_nand_part part) 68 return nvram->nand_part_size[part] * SZ_1K; 54 bcm963xx_nvram_nand_part_offset( const struct bcm963xx_nvram *nvram, enum bcm963xx_nvram_nand_part part) argument 64 bcm963xx_nvram_nand_part_size( const struct bcm963xx_nvram *nvram, enum bcm963xx_nvram_nand_part part) argument
|
/linux-master/drivers/net/ethernet/sfc/ |
H A D | mtd.c | 30 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 37 part->name, part->dev_type_name, rc); 40 static void efx_mtd_remove_partition(struct efx_mtd_partition *part) argument 45 rc = mtd_device_unregister(&part->mtd); 51 list_del(&part->node); 57 struct efx_mtd_partition *part; local 61 part = (struct efx_mtd_partition *)((char *)parts + 64 part->mtd.writesize = 1; 66 if (!(part 100 struct efx_mtd_partition *parts, *part, *next; local 118 struct efx_mtd_partition *part; local [all...] |
/linux-master/drivers/net/ethernet/sfc/siena/ |
H A D | mtd.c | 30 struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); local 37 part->name, part->dev_type_name, rc); 40 static void efx_siena_mtd_remove_partition(struct efx_mtd_partition *part) argument 45 rc = mtd_device_unregister(&part->mtd); 51 list_del(&part->node); 57 struct efx_mtd_partition *part; local 61 part = (struct efx_mtd_partition *)((char *)parts + 64 part->mtd.writesize = 1; 66 if (!(part 100 struct efx_mtd_partition *parts, *part, *next; local 118 struct efx_mtd_partition *part; local [all...] |
/linux-master/drivers/net/ethernet/sfc/falcon/ |
H A D | mtd.c | 30 struct ef4_mtd_partition *part = to_ef4_mtd_partition(mtd); local 37 part->name, part->dev_type_name, rc); 40 static void ef4_mtd_remove_partition(struct ef4_mtd_partition *part) argument 45 rc = mtd_device_unregister(&part->mtd); 51 list_del(&part->node); 57 struct ef4_mtd_partition *part; local 61 part = (struct ef4_mtd_partition *)((char *)parts + 64 part->mtd.writesize = 1; 66 part 97 struct ef4_mtd_partition *parts, *part, *next; local 115 struct ef4_mtd_partition *part; local [all...] |
/linux-master/drivers/mtd/ |
H A D | rfd_ftl.c | 93 static int build_block_map(struct partition *part, int block_no) argument 95 struct block *block = &part->blocks[block_no]; 98 block->offset = part->block_size * block_no; 100 if (le16_to_cpu(part->header_cache[0]) != RFD_MAGIC) { 107 for (i=0; i<part->data_sectors_per_block; i++) { 110 entry = le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]); 123 if (entry >= part->sector_count) { 127 part->mbd.mtd->name, block_no, i, entry); 131 if (part->sector_map[entry] != -1) { 134 part 151 scan_header(struct partition *part) argument 242 struct partition *part = container_of(dev, struct partition, mbd); local 268 erase_block(struct partition *part, int block) argument 319 move_block_contents(struct partition *part, int block_no, u_long *old_sector) argument 407 reclaim_block(struct partition *part, u_long *old_sector) argument 479 find_free_block(struct partition *part) argument 503 find_writable_block(struct partition *part, u_long *old_sector) argument 545 mark_sector_deleted(struct partition *part, u_long old_addr) argument 582 find_free_sector(const struct partition *part, const struct block *block) argument 603 struct partition *part = container_of(dev, struct partition, mbd); local 669 struct partition *part = container_of(dev, struct partition, mbd); local 711 struct partition *part = container_of(dev, struct partition, mbd); local 738 struct partition *part = container_of(dev, struct partition, mbd); local 749 struct partition *part; local 795 struct partition *part = container_of(dev, struct partition, mbd); local [all...] |
H A D | ftl.c | 151 static int scan_header(partition_t *part) argument 157 part->header.FormattedSize = 0; 158 max_offset = (0x100000<part->mbd.mtd->size)?0x100000:part->mbd.mtd->size; 162 offset += part->mbd.mtd->erasesize ? : 0x2000) { 164 err = mtd_read(part->mbd.mtd, offset, sizeof(header), &ret, 183 if ((1 << header.EraseUnitSize) != part->mbd.mtd->erasesize) { 185 1 << header.EraseUnitSize,part->mbd.mtd->erasesize); 188 part->header = header; 192 static int build_maps(partition_t *part) argument 328 erase_xfer(partition_t *part, uint16_t xfernum) argument 370 prepare_xfer(partition_t *part, int i) argument 428 copy_erase_unit(partition_t *part, uint16_t srcunit, uint16_t xferunit) argument 568 reclaim_block(partition_t *part) argument 675 dump_lists(partition_t *part) argument 687 find_free(partition_t *part) argument 748 ftl_read(partition_t *part, caddr_t buffer, u_long sector, u_long nblocks) argument 794 set_bam_entry(partition_t *part, uint32_t log_addr, uint32_t virt_addr) argument 862 ftl_write(partition_t *part, caddr_t buffer, u_long sector, u_long nblocks) argument 944 partition_t *part = container_of(dev, struct partition_t, mbd); local 972 partition_t *part = container_of(dev, struct partition_t, mbd); local 994 ftl_freepart(partition_t *part) argument [all...] |
H A D | mtdpart.c | 37 WARN_ON(!list_empty(&mtd->part.node)); 42 const struct mtd_partition *part, 49 parent->part.size : parent->size; 57 name = kstrdup(part->name, GFP_KERNEL); 68 child->part.flags = parent->flags & ~part->mask_flags; 69 child->part.flags |= part->add_flags; 70 child->flags = child->part.flags; 71 child->part 41 allocate_partition(struct mtd_info *parent, const struct mtd_partition *part, int partno, uint64_t cur_offset) argument 251 struct mtd_partition part; local [all...] |
/linux-master/block/partitions/ |
H A D | mac.c | 16 extern void note_bootable_part(dev_t dev, int part, int goodness); 43 struct mac_partition *part; local 61 if (partoffset + sizeof(*part) > datasize) 63 part = (struct mac_partition *) (data + partoffset); 64 if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) { 68 blocks_in_map = be32_to_cpu(part->map_count); 84 part = (struct mac_partition *) (data + pos%512); 85 if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) 88 be32_to_cpu(part->start_block) * (secsize/512), 89 be32_to_cpu(part [all...] |
/linux-master/arch/arm/mach-omap2/ |
H A D | prminst44xx.c | 58 u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx) argument 60 BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS || 61 part == OMAP4430_INVALID_PRCM_PARTITION || 62 !_prm_bases[part].va); 63 return readl_relaxed(_prm_bases[part].va + inst + idx); 67 void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx) argument 69 BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS || 70 part == OMAP4430_INVALID_PRCM_PARTITION || 71 !_prm_bases[part].va); 72 writel_relaxed(val, _prm_bases[part] 76 omap4_prminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, s16 inst, u16 idx) argument 99 omap4_prminst_is_hardreset_asserted(u8 shift, u8 part, s16 inst, u16 rstctrl_offs) argument 123 omap4_prminst_assert_hardreset(u8 shift, u8 part, s16 inst, u16 rstctrl_offs) argument 152 omap4_prminst_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 inst, u16 rstctrl_offs, u16 rstst_offs) argument [all...] |
H A D | cm.h | 56 int (*wait_module_ready)(u8 part, s16 prcm_mod, u16 idlest_reg, 58 int (*wait_module_idle)(u8 part, s16 prcm_mod, u16 idlest_reg, 60 void (*module_enable)(u8 mode, u8 part, u16 inst, u16 clkctrl_offs); 61 void (*module_disable)(u8 part, u16 inst, u16 clkctrl_offs); 62 u32 (*xlate_clkctrl)(u8 part, u16 inst, u16 clkctrl_offs); 67 int omap_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_reg, 69 int omap_cm_wait_module_idle(u8 part, s16 prcm_mod, u16 idlest_reg, 71 int omap_cm_module_enable(u8 mode, u8 part, u16 inst, u16 clkctrl_offs); 72 int omap_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs); 73 u32 omap_cm_xlate_clkctrl(u8 part, u1 [all...] |
H A D | prminst44xx.h | 20 extern u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx); 21 extern void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx); 22 extern u32 omap4_prminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, 27 extern int omap4_prminst_is_hardreset_asserted(u8 shift, u8 part, s16 inst, 29 extern int omap4_prminst_assert_hardreset(u8 shift, u8 part, s16 inst, 31 int omap4_prminst_deassert_hardreset(u8 shift, u8 st_shift, u8 part,
|
H A D | cminst44xx.c | 45 * 0x2 idle: Module is in Idle mode (only OCP part). It is functional if 74 static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx); 78 * @part: PRCM partition ID that the CM_CLKCTRL register exists in 85 static u32 _clkctrl_idlest(u8 part, u16 inst, u16 clkctrl_offs) argument 87 u32 v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs); 95 * @part: PRCM partition ID that the CM_CLKCTRL register exists in 102 static bool _is_module_ready(u8 part, u16 inst, u16 clkctrl_offs) argument 106 v = _clkctrl_idlest(part, inst, clkctrl_offs); 113 static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx) argument 115 BUG_ON(part > 122 omap4_cminst_write_inst_reg(u32 val, u8 part, u16 inst, u16 idx) argument 131 omap4_cminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, u16 inst, s16 idx) argument 144 omap4_cminst_set_inst_reg_bits(u32 bits, u8 part, u16 inst, s16 idx) argument 149 omap4_cminst_clear_inst_reg_bits(u32 bits, u8 part, u16 inst, s16 idx) argument 155 omap4_cminst_read_inst_reg_bits(u8 part, u16 inst, s16 idx, u32 mask) argument 180 _clktrctrl_write(u8 c, u8 part, u16 inst, u16 cdoffs) argument 199 omap4_cminst_is_clkdm_in_hwsup(u8 part, u16 inst, u16 cdoffs) argument 219 omap4_cminst_clkdm_enable_hwsup(u8 part, u16 inst, u16 cdoffs) argument 234 omap4_cminst_clkdm_disable_hwsup(u8 part, u16 inst, u16 cdoffs) argument 248 omap4_cminst_clkdm_force_wakeup(u8 part, u16 inst, u16 cdoffs) argument 257 omap4_cminst_clkdm_force_sleep(u8 part, u16 inst, u16 cdoffs) argument 274 omap4_cminst_wait_module_ready(u8 part, s16 inst, u16 clkctrl_offs, u8 bit_shift) argument 297 omap4_cminst_wait_module_idle(u8 part, s16 inst, u16 clkctrl_offs, u8 bit_shift) argument 318 omap4_cminst_module_enable(u8 mode, u8 part, u16 inst, u16 clkctrl_offs) argument 337 omap4_cminst_module_disable(u8 part, u16 inst, u16 clkctrl_offs) argument 476 omap4_cminst_xlate_clkctrl(u8 part, u16 inst, u16 offset) argument [all...] |
/linux-master/drivers/acpi/acpica/ |
H A D | utmath.c | 23 struct uint64_struct part; member in union:uint64_overlay 62 ACPI_MUL_64_BY_32(0, multiplicand_ovl.part.hi, multiplier, 63 product.part.hi, carry32); 65 ACPI_MUL_64_BY_32(0, multiplicand_ovl.part.lo, multiplier, 66 product.part.lo, carry32); 68 product.part.hi += carry32; 100 operand_ovl.part.hi = operand_ovl.part.lo; 101 operand_ovl.part.lo = 0; 104 ACPI_SHIFT_LEFT_64_BY_32(operand_ovl.part [all...] |
/linux-master/drivers/misc/sgi-xp/ |
H A D | xpc_partition.c | 13 * This is the part of XPC that detects the presence/absence of 265 static int __xpc_partition_disengaged(struct xpc_partition *part, argument 268 short partid = XPC_PARTID(part); 272 if (part->disengage_timeout) { 274 if (time_is_after_jiffies(part->disengage_timeout)) { 290 part->disengage_timeout = 0; 294 del_timer_sync(&part->disengage_timer); 296 DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING && 297 part->act_state != XPC_P_AS_INACTIVE); 298 if (part 306 xpc_partition_disengaged(struct xpc_partition *part) argument 311 xpc_partition_disengaged_from_timer(struct xpc_partition *part) argument 320 xpc_mark_partition_active(struct xpc_partition *part) argument 344 xpc_deactivate_partition(const int line, struct xpc_partition *part, enum xp_retval reason) argument 392 xpc_mark_partition_inactive(struct xpc_partition *part) argument 532 struct xpc_partition *part; local [all...] |
H A D | xpc_main.c | 16 * partition This part detects the presence/absence of other 20 * channel This part manages the channels and sends/receives 68 .init_name = "", /* set to "part" at xpc_init() time */ 167 struct xpc_partition *part = from_timer(part, t, disengage_timer); local 169 DBUG_ON(time_is_after_jiffies(part->disengage_timeout)); 171 xpc_partition_disengaged_from_timer(part); 173 DBUG_ON(part->disengage_timeout != 0); 174 DBUG_ON(xpc_arch_ops.partition_engaged(XPC_PARTID(part))); 216 struct xpc_partition *part; local 329 xpc_channel_mgr(struct xpc_partition *part) argument 390 xpc_setup_ch_structures(struct xpc_partition *part) argument 482 xpc_teardown_ch_structures(struct xpc_partition *part) argument 524 struct xpc_partition *part = &xpc_partitions[partid]; local 574 xpc_activate_partition(struct xpc_partition *part) argument 638 xpc_kthread_waitmsgs(struct xpc_partition *part, struct xpc_channel *ch) argument 675 struct xpc_partition *part = &xpc_partitions[partid]; local 767 struct xpc_partition *part = &xpc_partitions[ch->partid]; local 841 struct xpc_partition *part; local 890 struct xpc_partition *part; local 941 struct xpc_partition *part; local 1078 struct xpc_partition *part; local [all...] |
/linux-master/drivers/irqchip/ |
H A D | irq-partition-percpu.c | 25 static bool partition_check_cpu(struct partition_desc *part, argument 28 return cpumask_test_cpu(cpu, &part->parts[hwirq].mask); 33 struct partition_desc *part = irq_data_get_irq_chip_data(d); local 34 struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); 35 struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); 37 if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && 44 struct partition_desc *part = irq_data_get_irq_chip_data(d); local 45 struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); 46 struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); 48 if (partition_check_cpu(part, smp_processor_i 57 struct partition_desc *part = irq_data_get_irq_chip_data(d); local 72 struct partition_desc *part = irq_data_get_irq_chip_data(d); local 85 struct partition_desc *part = irq_data_get_irq_chip_data(d); local 97 struct partition_desc *part = irq_data_get_irq_chip_data(d); local 115 struct partition_desc *part = irq_desc_get_handler_data(desc); local 142 struct partition_desc *part; local 176 struct partition_affinity *part = NULL; local [all...] |
/linux-master/drivers/net/wireless/ti/wlcore/ |
H A D | io.c | 3 * This file is part of wl1271 57 struct wlcore_partition_set *part = &wl->curr_part; local 69 if ((addr >= part->mem.start) && 70 (addr < part->mem.start + part->mem.size)) 71 return addr - part->mem.start; 72 else if ((addr >= part->reg.start) && 73 (addr < part->reg.start + part->reg.size)) 74 return addr - part [all...] |
/linux-master/fs/orangefs/ |
H A D | dir.c | 17 struct orangefs_dir_part *part; member in struct:orangefs_dir 40 * part and is stored in a linked list in case more than one part is 43 * The position pointer (ctx->pos) encodes the part and offset on which 44 * to begin reading at. Bits above PART_SHIFT encode the part and bits 51 * part list. Data is parsed from the current position as it is needed. 58 * first part of the part list. 134 struct orangefs_dir_part *part, *new; local 138 part 188 fill_from_part(struct orangefs_dir_part *part, struct dir_context *ctx) argument 242 struct orangefs_dir_part *part; local 288 struct orangefs_dir_part *part = od->part; local 380 struct orangefs_dir_part *part = od->part; local [all...] |
/linux-master/drivers/misc/ |
H A D | sram-exec.c | 23 struct sram_partition *part) 25 unsigned long base = (unsigned long)part->base; 37 int sram_add_protect_exec(struct sram_partition *part) argument 40 list_add_tail(&part->list, &exec_pool_list); 49 * @pool: struct gen_pool retrieved that is part of this sram 76 struct sram_partition *part = NULL, *p; local 85 part = p; 89 if (!part) 95 base = (unsigned long)part->base; 98 mutex_lock(&part 22 sram_check_protect_exec(struct sram_dev *sram, struct sram_reserve *block, struct sram_partition *part) argument [all...] |
H A D | sram.h | 48 struct sram_partition *part); 49 int sram_add_protect_exec(struct sram_partition *part); 53 struct sram_partition *part) 58 static inline int sram_add_protect_exec(struct sram_partition *part) argument 51 sram_check_protect_exec(struct sram_dev *sram, struct sram_reserve *block, struct sram_partition *part) argument
|
H A D | sram.c | 29 struct sram_partition *part; local 31 part = container_of(attr, struct sram_partition, battr); 33 mutex_lock(&part->lock); 34 memcpy_fromio(buf, part->base + pos, count); 35 mutex_unlock(&part->lock); 44 struct sram_partition *part; local 46 part = container_of(attr, struct sram_partition, battr); 48 mutex_lock(&part->lock); 49 memcpy_toio(part->base + pos, buf, count); 50 mutex_unlock(&part 55 sram_add_pool(struct sram_dev *sram, struct sram_reserve *block, phys_addr_t start, struct sram_partition *part) argument 75 sram_add_export(struct sram_dev *sram, struct sram_reserve *block, phys_addr_t start, struct sram_partition *part) argument 97 struct sram_partition *part = &sram->partition[sram->partitions]; local 148 struct sram_partition *part; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum1_kvdl.c | 74 struct mlxsw_sp1_kvdl_part *part, *min_part = NULL; local 78 part = kvdl->parts[i]; 79 if (alloc_size <= part->info.alloc_size && 81 part->info.alloc_size <= min_part->info.alloc_size)) 82 min_part = part; 91 struct mlxsw_sp1_kvdl_part *part; local 95 part = kvdl->parts[i]; 96 if (kvdl_index >= part->info.start_index && 97 kvdl_index <= part->info.end_index) 98 return part; 118 mlxsw_sp1_kvdl_part_alloc(struct mlxsw_sp1_kvdl_part *part, u32 *p_kvdl_index) argument 136 mlxsw_sp1_kvdl_part_free(struct mlxsw_sp1_kvdl_part *part, u32 kvdl_index) argument 152 struct mlxsw_sp1_kvdl_part *part; local 169 struct mlxsw_sp1_kvdl_part *part; local 184 struct mlxsw_sp1_kvdl_part *part; local 195 mlxsw_sp1_kvdl_part_update(struct mlxsw_sp1_kvdl_part *part, struct mlxsw_sp1_kvdl_part *part_prev, unsigned int size) argument 213 struct mlxsw_sp1_kvdl_part *part; local 239 mlxsw_sp1_kvdl_part_fini(struct mlxsw_sp1_kvdl_part *part) argument 277 mlxsw_sp1_kvdl_part_occ(struct mlxsw_sp1_kvdl_part *part) argument 308 struct mlxsw_sp1_kvdl_part *part; local 317 struct mlxsw_sp1_kvdl_part *part; local 326 struct mlxsw_sp1_kvdl_part *part; local [all...] |
H A D | spectrum2_kvdl.c | 56 static int mlxsw_sp2_kvdl_part_find_zero_bits(struct mlxsw_sp2_kvdl_part *part, argument 65 start_bit = part->last_allocated_bit + 1; 66 if (start_bit == part->usage_bit_count) 70 bit = find_next_zero_bit(part->usage, part->usage_bit_count, bit); 71 if (!wrap && bit + bit_count >= part->usage_bit_count) { 79 if (test_bit(bit + i, part->usage)) { 88 static int mlxsw_sp2_kvdl_part_alloc(struct mlxsw_sp2_kvdl_part *part, argument 97 bit_count = DIV_ROUND_UP(size, part->indexes_per_usage_bit); 98 err = mlxsw_sp2_kvdl_part_find_zero_bits(part, bit_coun 124 mlxsw_sp2_kvdl_part_free(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp2_kvdl_part *part, unsigned int size, u32 kvdl_index) argument 152 struct mlxsw_sp2_kvdl_part *part = kvdl->parts[type]; local 164 struct mlxsw_sp2_kvdl_part *part = kvdl->parts[type]; local 184 struct mlxsw_sp2_kvdl_part *part; local 216 mlxsw_sp2_kvdl_part_fini(struct mlxsw_sp2_kvdl_part *part) argument [all...] |
/linux-master/drivers/mtd/parsers/ |
H A D | cmdlinepart.c | 16 * if specified or truncated size is 0 the part is skipped 18 * if omitted the part will immediately follow the previous part 19 * or 0 if the first part 240 * make sure that part-names with ":" will not be handled as 241 * part of the mtd-id with an ":" 252 * as an <mtd-id>/<part-definition> separator. 339 struct cmdline_mtd_partition *part; local 353 for (part = partitions; part; par [all...] |