/linux-master/drivers/gpu/drm/amd/display/dc/gpio/dcn30/ |
H A D | hw_factory_dcn30.c | 56 #define block HPD macro 72 #define REGI(reg_name, block, id)\ 73 BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 74 mm ## block ## id ## _ ## reg_name
|
/linux-master/drivers/gpu/drm/amd/display/dc/gpio/dcn32/ |
H A D | hw_factory_dcn32.c | 49 #define block HPD macro 65 #define REGI(reg_name, block, id)\ 66 BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \ 67 reg ## block ## id ## _ ## reg_name
|
/linux-master/tools/testing/selftests/drivers/net/mlxsw/ |
H A D | sch_red_root.sh | 71 install_qdisc qevent early_drop block 10
|
/linux-master/fs/squashfs/ |
H A D | squashfs_fs_sb.h | 30 u64 block; member in struct:squashfs_cache_entry
|
H A D | Makefile | 7 squashfs-y += block.o cache.o dir.o export.o file.o fragment.o id.o inode.o
|
/linux-master/fs/pstore/ |
H A D | ram_core.c | 127 uint8_t *block; local 136 block = buffer->data + (start & ~(ecc_block_size - 1)); 140 if (block + ecc_block_size > buffer_end) 141 size = buffer_end - block; 142 persistent_ram_encode_rs8(prz, block, size, par); 143 block += ecc_block_size; 145 } while (block < buffer->data + start + count); 162 uint8_t *block; local 168 block = buffer->data; 170 while (block < buffe [all...] |
/linux-master/arch/powerpc/mm/kasan/ |
H A D | init_32.c | 61 void *block; local 68 block = memblock_alloc(k_end - k_start, PAGE_SIZE); 69 if (!block) 74 void *va = block + k_cur - k_start;
|
/linux-master/fs/nilfs2/ |
H A D | mdt.c | 30 nilfs_mdt_insert_new_block(struct inode *inode, unsigned long block, argument 44 ret = nilfs_bmap_insert(ii->i_bmap, block, (unsigned long)bh); 61 trace_nilfs2_mdt_insert_new_block(inode, inode->i_ino, block); 66 static int nilfs_mdt_create_block(struct inode *inode, unsigned long block, argument 80 bh = nilfs_grab_buffer(inode, inode->i_mapping, block, 0); 93 err = nilfs_mdt_insert_new_block(inode, block, bh, init_block); 168 static int nilfs_mdt_read_block(struct inode *inode, unsigned long block, argument 176 err = nilfs_mdt_submit_block(inode, block, REQ_OP_READ, &first_bh); 184 blkoff = block + 1; 204 "I/O error reading meta-data file (ino=%lu, block 329 nilfs_mdt_delete_block(struct inode *inode, unsigned long block) argument 357 nilfs_mdt_forget_block(struct inode *inode, unsigned long block) argument [all...] |
/linux-master/arch/mips/alchemy/common/ |
H A D | usb.c | 3 * USB block power/access management abstraction. 5 * Au1000+: The OHCI block control register is at the far end of the OHCI memory 27 /* Au1000 USB block config bits */ 29 #define USBHEN_CE (1 << 3) /* OHCI block clock enable */ 30 #define USBHEN_E (1 << 2) /* OHCI block enable */ 31 #define USBHEN_C (1 << 1) /* OHCI block coherency bit */ 115 /* no USB block active, do disable all PHYs */ 131 r = __raw_readl(base + USB_DWC_CTRL3); /* enable OHCI block */ 265 static inline int au1300_usb_control(int block, int enable) argument 271 switch (block) { 360 au1200_usb_control(int block, int enable) argument 458 au1000_usb_control(int block, int enable, unsigned long rb, int creg) argument 478 alchemy_usb_control(int block, int enable) argument [all...] |
/linux-master/drivers/i2c/busses/ |
H A D | i2c-sis630.c | 172 /* check if block transmitted */ 236 len = data->block[0]; 244 "set data 0x%02x\n", data->block[i]); 246 sis630_write(SMB_BYTE + (i - 1) % 8, data->block[i]); 280 data->block[0] = len = 0; 293 data->block[0] = sis630_read(SMB_COUNT); 296 if (data->block[0] > 32) 297 data->block[0] = 32; 300 "block data read len=0x%x\n", data->block[ [all...] |
H A D | i2c-ali1535.c | 111 #define ALI1535_BLOCK_CLR 0x04 /* reset block data index */ 402 len = data->block[0]; 405 data->block[0] = len; 409 data->block[0] = len; 415 outb_p(data->block[i], SMBBLKDAT); 447 data->block[0] = len; 450 for (i = 1; i <= data->block[0]; i++) { 451 data->block[i] = inb_p(SMBBLKDAT); 453 len, i, data->block[i]);
|
H A D | i2c-ali15x3.c | 389 len = data->block[0]; 392 data->block[0] = len; 396 data->block[0] = len; 402 outb_p(data->block[i], SMBBLKDAT); 435 data->block[0] = len; 438 for (i = 1; i <= data->block[0]; i++) { 439 data->block[i] = inb_p(SMBBLKDAT); 441 len, i, data->block[i]);
|
/linux-master/fs/qnx4/ |
H A D | inode.c | 69 // logical block is before EOF 93 u32 block = try_extent(&qnx4_inode->di_first_xtnt, &offset); local 95 if (block) { 103 // read next xtnt block. 106 QNX4DEBUG((KERN_ERR "qnx4: I/O error reading xtnt block [%ld])\n", i_xblk - 1)); 111 QNX4DEBUG((KERN_ERR "qnx4: block at %ld is not a valid xtnt\n", qnx4_inode->i_xblk)); 115 block = try_extent(&xblk->xblk_xtnts[ix], &offset); 116 if (block) { 131 QNX4DEBUG((KERN_INFO "qnx4: mapping block %ld of inode %ld = %ld\n",iblock,inode->i_ino,block)); 272 qnx4_bmap(struct address_space *mapping, sector_t block) argument 286 int block; local [all...] |
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_btree_staging.c | 169 * single root block). When the creation is complete, the new root can be 175 * then construct a staged btree cursor. Refer to the block comments about 201 * The number of items placed in each btree block is computed via the following 205 * level, the desired number of items per block is defined as: 215 * btree block is defined as: 220 * needed to handle uneven division. If the number of records in any block 232 * Put a btree block that we're loading onto the ordered list and release it. 272 * Allocate and initialize one btree block for bulk loading. 274 * The new btree block will have its level and numrecs fields set to the values 278 * sibling of the new block, i 355 xfs_btree_bload_leaf( struct xfs_btree_cur *cur, unsigned int recs_this_block, xfs_btree_bload_get_records_fn get_records, struct xfs_btree_block *block, void *priv) argument 385 xfs_btree_bload_node( struct xfs_btree_cur *cur, unsigned int recs_this_block, union xfs_btree_ptr *child_ptr, struct xfs_btree_block *block) argument 679 struct xfs_btree_block *block = NULL; local [all...] |
/linux-master/arch/mips/cavium-octeon/executive/ |
H A D | cvmx-interrupt-decodes.c | 51 * @block: interrupt register block_id 53 void __cvmx_interrupt_gmxx_rxx_int_en_enable(int index, int block) argument 56 cvmx_write_csr(CVMX_GMXX_RXX_INT_REG(index, block), 57 cvmx_read_csr(CVMX_GMXX_RXX_INT_REG(index, block))); 229 cvmx_write_csr(CVMX_GMXX_RXX_INT_EN(index, block), gmx_rx_int_en.u64); 234 * @block: interrupt register block_id 236 void __cvmx_interrupt_pcsx_intx_en_reg_enable(int index, int block) argument 239 cvmx_write_csr(CVMX_PCSX_INTX_REG(index, block), 240 cvmx_read_csr(CVMX_PCSX_INTX_REG(index, block))); 272 cvmx_write_csr(CVMX_PCSX_INTX_EN_REG(index, block), pcs_int_en_re [all...] |
/linux-master/fs/freevxfs/ |
H A D | vxfs_olt.c | 33 vxfs_oblock(struct super_block *sbp, daddr_t block, u_long bsize) argument 36 return (block * (sbp->s_blocksize / bsize));
|
/linux-master/arch/um/drivers/ |
H A D | random.c | 34 static int rng_dev_read(struct hwrng *rng, void *buf, size_t max, bool block) argument 40 if (block && ret == -EAGAIN) {
|
/linux-master/drivers/misc/ |
H A D | sram-exec.c | 22 int sram_check_protect_exec(struct sram_dev *sram, struct sram_reserve *block, argument 26 unsigned long end = base + block->size;
|
/linux-master/drivers/md/ |
H A D | dm-verity-fec.h | 22 /* we need buffers for at most 1 << block size RS blocks */ 72 enum verity_block_type type, sector_t block, 103 sector_t block, u8 *dest, 100 verity_fec_decode(struct dm_verity *v, struct dm_verity_io *io, enum verity_block_type type, sector_t block, u8 *dest, struct bvec_iter *iter) argument
|
/linux-master/fs/udf/ |
H A D | misc.c | 182 * Read the first block of a tagged descriptor. 188 struct buffer_head *udf_read_tagged(struct super_block *sb, uint32_t block, argument 195 /* Read the block */ 196 if (block == 0xFFFFFFFF) 199 bh = sb_bread(sb, block); 201 udf_err(sb, "read failed, block=%u, location=%u\n", 202 block, location); 211 udf_debug("location mismatch block %u, tag %u != %u\n", 212 block, le32_to_cpu(tag_p->tagLocation), location); 219 udf_err(sb, "tag checksum failed, block [all...] |
/linux-master/block/ |
H A D | disk-events.c | 16 int block; /* event blocking depth */ member in struct:disk_events 59 * disk_block_events - block and flush disk event checking 60 * @disk: disk to block events for 67 * Note that this intentionally does not block event checking from 89 cancel = !ev->block++; 106 if (WARN_ON_ONCE(ev->block <= 0)) 109 if (--ev->block) 127 * Undo disk_block_events(). When the block count reaches zero, it 160 if (!ev->block) 203 if (!ev->block [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mc/ |
H A D | gp100.c | 78 intr->func->block(intr, i, 0xffffffff); 86 .block = gp100_mc_intr_block,
|
/linux-master/fs/adfs/ |
H A D | inode.c | 13 * Lookup/Create a block at offset 'block' into 'inode'. We currently do 17 adfs_get_block(struct inode *inode, sector_t block, struct buffer_head *bh, argument 21 if (block >= inode->i_blocks) 24 block = __adfs_block_map(inode->i_sb, ADFS_I(inode)->indaddr, 25 block); 26 if (block) 27 map_bh(bh, inode->i_sb, block); 72 static sector_t _adfs_bmap(struct address_space *mapping, sector_t block) argument 74 return generic_block_bmap(mapping, block, adfs_get_bloc [all...] |
/linux-master/drivers/gpu/drm/omapdrm/ |
H A D | omap_gem.c | 42 /** protects pin_cnt, block, pages, dma_addrs and vaddr */ 80 * tiler block used when buffer is remapped in DMM/TILER. 82 struct tiler_block *block; member in struct:omap_gem_object 114 struct tiler_block *block; /* the reserved tiler block */ member in struct:omap_drm_usergart_entry 457 err = tiler_pin(entry->block, pages, ARRAY_SIZE(pages), 0, true); 653 if (omap_obj->block) { 658 ret = tiler_pin(omap_obj->block, omap_obj->pages, npages, 751 struct tiler_block *block; local 754 BUG_ON(omap_obj->block); 1501 struct tiler_block *block; local [all...] |
/linux-master/include/linux/iio/ |
H A D | buffer-dma.h | 35 * struct iio_dma_buffer_block - IIO buffer block 37 * @size: Total size of the block in bytes 42 * @kref: kref used to manage the lifetime of block 43 * @state: Current state of the block 46 /* May only be accessed by the owner of the block */ 52 * by anybody holding a reference to the block. 63 * queue->list_lock if the block is not owned by the core. 72 * @pos: Read offset in the active block 73 * @block_size: Size of each block 74 * @next_dequeue: index of next block tha [all...] |