/netgear-R7800-V1.0.2.28/target/linux/generic/files/fs/yaffs2/ |
H A D | yaffs_checkptrw.c | 17 static int yaffs2_checkpt_space_ok(yaffs_dev_t *dev) argument 19 int blocksAvailable = dev->n_erased_blocks - dev->param.n_reserved_blocks; 29 static int yaffs_checkpt_erase(yaffs_dev_t *dev) argument 33 if (!dev->param.erase_fn) 36 dev->internal_start_block, dev->internal_end_block)); 38 for (i = dev->internal_start_block; i <= dev->internal_end_block; i++) { 39 yaffs_block_info_t *bi = yaffs_get_block_info(dev, 62 yaffs2_checkpt_find_erased_block(yaffs_dev_t *dev) argument 90 yaffs2_checkpt_find_block(yaffs_dev_t *dev) argument 126 yaffs2_checkpt_open(yaffs_dev_t *dev, int forWriting) argument 180 yaffs2_get_checkpt_sum(yaffs_dev_t *dev, __u32 *sum) argument 188 yaffs2_checkpt_flush_buffer(yaffs_dev_t *dev) argument 241 yaffs2_checkpt_wr(yaffs_dev_t *dev, const void *data, int n_bytes) argument 276 yaffs2_checkpt_rd(yaffs_dev_t *dev, void *data, int n_bytes) argument 351 yaffs_checkpt_close(yaffs_dev_t *dev) argument 390 yaffs2_checkpt_invalidate_stream(yaffs_dev_t *dev) argument [all...] |
H A D | yaffs_yaffs2.h | 19 void yaffs_calc_oldest_dirty_seq(yaffs_dev_t *dev); 20 void yaffs2_find_oldest_dirty_seq(yaffs_dev_t *dev); 21 void yaffs2_clear_oldest_dirty_seq(yaffs_dev_t *dev, yaffs_block_info_t *bi); 22 void yaffs2_update_oldest_dirty_seq(yaffs_dev_t *dev, unsigned block_no, yaffs_block_info_t *bi); 23 int yaffs_block_ok_for_gc(yaffs_dev_t *dev, yaffs_block_info_t *bi); 24 __u32 yaffs2_find_refresh_block(yaffs_dev_t *dev); 25 int yaffs2_checkpt_required(yaffs_dev_t *dev); 26 int yaffs_calc_checkpt_blocks_required(yaffs_dev_t *dev); 29 void yaffs2_checkpt_invalidate(yaffs_dev_t *dev); 30 int yaffs2_checkpt_save(yaffs_dev_t *dev); [all...] |
H A D | yaffs_yaffs1.h | 20 int yaffs1_scan(yaffs_dev_t *dev);
|
H A D | yaffs_allocator.h | 21 void yaffs_init_raw_tnodes_and_objs(yaffs_dev_t *dev); 22 void yaffs_deinit_raw_tnodes_and_objs(yaffs_dev_t *dev); 24 yaffs_tnode_t *yaffs_alloc_raw_tnode(yaffs_dev_t *dev); 25 void yaffs_free_raw_tnode(yaffs_dev_t *dev, yaffs_tnode_t *tn); 27 yaffs_obj_t *yaffs_alloc_raw_obj(yaffs_dev_t *dev); 28 void yaffs_free_raw_obj(yaffs_dev_t *dev, yaffs_obj_t *obj);
|
H A D | yaffs_checkptrw.h | 21 int yaffs2_checkpt_open(yaffs_dev_t *dev, int forWriting); 23 int yaffs2_checkpt_wr(yaffs_dev_t *dev, const void *data, int n_bytes); 25 int yaffs2_checkpt_rd(yaffs_dev_t *dev, void *data, int n_bytes); 27 int yaffs2_get_checkpt_sum(yaffs_dev_t *dev, __u32 *sum); 29 int yaffs_checkpt_close(yaffs_dev_t *dev); 31 int yaffs2_checkpt_invalidate_stream(yaffs_dev_t *dev);
|
H A D | yaffs_nand.c | 20 int yaffs_rd_chunk_tags_nand(yaffs_dev_t *dev, int nand_chunk, argument 27 int realignedChunkInNAND = nand_chunk - dev->chunk_offset; 29 dev->n_page_reads++; 35 if (dev->param.read_chunk_tags_fn) 36 result = dev->param.read_chunk_tags_fn(dev, realignedChunkInNAND, buffer, 39 result = yaffs_tags_compat_rd(dev, 47 bi = yaffs_get_block_info(dev, nand_chunk/dev->param.chunks_per_block); 48 yaffs_handle_chunk_error(dev, b 54 yaffs_wr_chunk_tags_nand(yaffs_dev_t *dev, int nand_chunk, const __u8 *buffer, yaffs_ext_tags *tags) argument 91 yaffs_mark_bad(yaffs_dev_t *dev, int block_no) argument 102 yaffs_query_init_block_state(yaffs_dev_t *dev, int block_no, yaffs_block_state_t *state, __u32 *seq_number) argument 118 yaffs_erase_block(struct yaffs_dev_s *dev, int flash_block) argument 132 yaffs_init_nand(struct yaffs_dev_s *dev) argument [all...] |
H A D | yaffs_bitmap.c | 20 static Y_INLINE __u8 *yaffs_BlockBits(yaffs_dev_t *dev, int blk) argument 22 if (blk < dev->internal_start_block || blk > dev->internal_end_block) { 28 return dev->chunk_bits + 29 (dev->chunk_bit_stride * (blk - dev->internal_start_block)); 32 void yaffs_verify_chunk_bit_id(yaffs_dev_t *dev, int blk, int chunk) argument 34 if (blk < dev->internal_start_block || blk > dev->internal_end_block || 35 chunk < 0 || chunk >= dev 43 yaffs_clear_chunk_bits(yaffs_dev_t *dev, int blk) argument 50 yaffs_clear_chunk_bit(yaffs_dev_t *dev, int blk, int chunk) argument 59 yaffs_set_chunk_bit(yaffs_dev_t *dev, int blk, int chunk) argument 68 yaffs_check_chunk_bit(yaffs_dev_t *dev, int blk, int chunk) argument 76 yaffs_still_some_chunks(yaffs_dev_t *dev, int blk) argument 88 yaffs_count_chunk_bits(yaffs_dev_t *dev, int blk) argument [all...] |
H A D | yaffs_bitmap.h | 23 void yaffs_verify_chunk_bit_id(yaffs_dev_t *dev, int blk, int chunk); 24 void yaffs_clear_chunk_bits(yaffs_dev_t *dev, int blk); 25 void yaffs_clear_chunk_bit(yaffs_dev_t *dev, int blk, int chunk); 26 void yaffs_set_chunk_bit(yaffs_dev_t *dev, int blk, int chunk); 27 int yaffs_check_chunk_bit(yaffs_dev_t *dev, int blk, int chunk); 28 int yaffs_still_some_chunks(yaffs_dev_t *dev, int blk); 29 int yaffs_count_chunk_bits(yaffs_dev_t *dev, int blk);
|
H A D | yaffs_getblockinfo.h | 23 static Y_INLINE yaffs_block_info_t *yaffs_get_block_info(yaffs_dev_t * dev, int blk) argument 25 if (blk < dev->internal_start_block || blk > dev->internal_end_block) { 32 return &dev->block_info[blk - dev->internal_start_block];
|
H A D | yaffs_nand.h | 22 int yaffs_rd_chunk_tags_nand(yaffs_dev_t *dev, int nand_chunk, 26 int yaffs_wr_chunk_tags_nand(yaffs_dev_t *dev, 31 int yaffs_mark_bad(yaffs_dev_t *dev, int block_no); 33 int yaffs_query_init_block_state(yaffs_dev_t *dev, 38 int yaffs_erase_block(struct yaffs_dev_s *dev, 41 int yaffs_init_nand(struct yaffs_dev_s *dev);
|
H A D | yaffs_nandemul2k.h | 23 int nandemul2k_WriteChunkWithTagsToNAND(struct yaffs_dev_s *dev, 26 int nandemul2k_ReadChunkWithTagsFromNAND(struct yaffs_dev_s *dev, 29 int nandemul2k_MarkNANDBlockBad(struct yaffs_dev_s *dev, int block_no); 30 int nandemul2k_QueryNANDBlock(struct yaffs_dev_s *dev, int block_no, 32 int nandemul2k_EraseBlockInNAND(struct yaffs_dev_s *dev, 34 int nandemul2k_InitialiseNAND(struct yaffs_dev_s *dev);
|
H A D | yaffs_mtdif1.h | 17 int nandmtd1_WriteChunkWithTagsToNAND(yaffs_dev_t *dev, int nand_chunk, 20 int nandmtd1_ReadChunkWithTagsFromNAND(yaffs_dev_t *dev, int nand_chunk, 23 int nandmtd1_MarkNANDBlockBad(struct yaffs_dev_s *dev, int block_no); 25 int nandmtd1_QueryNANDBlock(struct yaffs_dev_s *dev, int block_no,
|
H A D | yaffs_mtdif2.h | 20 int nandmtd2_WriteChunkWithTagsToNAND(yaffs_dev_t *dev, int nand_chunk, 23 int nandmtd2_ReadChunkWithTagsFromNAND(yaffs_dev_t *dev, int nand_chunk, 25 int nandmtd2_MarkNANDBlockBad(struct yaffs_dev_s *dev, int block_no); 26 int nandmtd2_QueryNANDBlock(struct yaffs_dev_s *dev, int block_no,
|
/netgear-R7800-V1.0.2.28/package/iproute2/files/ |
H A D | tc.debug | 3 for dev in $(ls -1 /sys/class/net/); do 4 [ -d /sys/class/net/${dev} ] || continue 5 log tc -s qdisc show dev ${dev}
|
/netgear-R7800-V1.0.2.28/package/qca-hostap/files/ |
H A D | wps-reset-hotplug.sh | 20 echo "" > /dev/console 21 echo "RESET TO FACTORY SETTING EVENT DETECTED" > /dev/console 22 echo "PLEASE WAIT WHILE REBOOTING THE DEVICE..." > /dev/console
|
/netgear-R7800-V1.0.2.28/package/qca-ssdk/src/include/ref/ |
H A D | ref_vlan.h | 23 qca_ar8327_sw_set_vlan(struct switch_dev *dev, 28 qca_ar8327_sw_get_vlan(struct switch_dev *dev, 33 qca_ar8327_sw_set_vid(struct switch_dev *dev, 38 qca_ar8327_sw_get_vid(struct switch_dev *dev, 43 qca_ar8327_sw_get_pvid(struct switch_dev *dev, int port, int *vlan); 46 qca_ar8327_sw_set_pvid(struct switch_dev *dev, int port, int vlan); 49 qca_ar8327_sw_get_ports(struct switch_dev *dev, struct switch_val *val); 52 qca_ar8327_sw_set_ports(struct switch_dev *dev, struct switch_val *val); 55 qca_ar8327_sw_hw_apply(struct switch_dev *dev);
|
H A D | ref_port_ctrl.h | 23 qca_ar8327_sw_get_port_link(struct switch_dev *dev, int port, 27 qca_ar8327_sw_mac_polling_task(struct switch_dev *dev);
|
H A D | ref_mib.h | 23 qca_ar8327_sw_set_reset_mibs(struct switch_dev *dev, 28 qca_ar8327_sw_set_port_reset_mib(struct switch_dev *dev, 34 qca_ar8327_sw_get_port_mib(struct switch_dev *dev, 39 qca_ar8327_sw_mib_task(struct switch_dev *dev);
|
/netgear-R7800-V1.0.2.28/target/linux/generic/files/crypto/ocf/kirkwood/mvHal/kw_family/ctrlEnv/sys/ |
H A D | mvSysUsb.h | 83 #define MV_USB_BRIDGE_CTRL_REG(dev) (USB_REG_BASE(dev) + 0x300) 85 #define MV_USB_WIN_CTRL_REG(dev, win) (USB_REG_BASE(dev) + 0x320 + ((win)<<4)) 86 #define MV_USB_WIN_BASE_REG(dev, win) (USB_REG_BASE(dev) + 0x324 + ((win)<<4)) 110 #define MV_USB_BRIDGE_IPG_REG(dev) (USB_REG_BASE(dev) + 0x360) 113 MV_STATUS mvUsbInit(int dev, MV_BOOL isHost); 115 MV_STATUS mvUsbWinInit(int dev); [all...] |
/netgear-R7800-V1.0.2.28/target/linux/omap35xx/gumstix/base-files/lib/preinit/ |
H A D | 95_ttyS1_noecho | 5 stty -echo < /dev/ttyO1
|
/netgear-R7800-V1.0.2.28/target/linux/ubicom32/files/arch/ubicom32/kernel/ |
H A D | timer_broadcast.c | 58 struct clock_event_device *dev = &__get_cpu_var(local_clockevent); local 60 dev->event_handler(dev); 67 static int dummy_timer_set_next_event(unsigned long cycles, struct clock_event_device *dev) argument 87 struct clock_event_device *dev = &per_cpu(local_clockevent, cpu); local 89 dev->name = "timer-dummy"; 90 dev->features = CLOCK_EVT_FEAT_DUMMY; 91 dev->rating = 200; 92 dev->mult = 1; 93 dev [all...] |
H A D | timer_device.c | 60 static int timer_device_set_next_event(unsigned long cycles, struct clock_event_device *dev) argument 62 timer_set(dev->irq, cycles); 70 static void timer_device_set_mode(enum clock_event_mode mode, struct clock_event_device *dev) argument 78 printk(KERN_NOTICE "timer[%d]: shutdown\n", dev->irq); 79 ldsr_disable_vector(dev->irq); 87 printk(KERN_NOTICE "timer[%d]: oneshot\n", dev->irq); 88 ldsr_disable_vector(dev->irq); 96 dev->irq, frequency / CONFIG_HZ); 97 timer_set(dev->irq, frequency / CONFIG_HZ); 103 dev 117 struct clock_event_device *dev = (struct clock_event_device *)dev_id; local 174 struct clock_event_device *dev; local [all...] |
/netgear-R7800-V1.0.2.28/package/telnetenable/files/ |
H A D | RtDebug.sh | 3 mkdir /dev/pts 4 mknod -m 666 /dev/ptmx c 5 2 5 mknod -m 666 /dev/pts/0 c 136 0 6 mknod -m 666 /dev/pts/1 c 136 1
|
/netgear-R7800-V1.0.2.28/target/linux/ar71xx/base-files/etc/uci-defaults/ |
H A D | vlan-migration | 6 local dev="$(uci -q get network.@switch_vlan[0].device)" 9 if [ "$dev" = "rtl8366s" ] && [ "$vlan" = 0 ]; then
|
/netgear-R7800-V1.0.2.28/target/linux/ubicom32/files/arch/ubicom32/include/asm/ |
H A D | dma-mapping.h | 41 dma_supported(struct device *dev, u64 mask) argument 43 BUG_ON(dev->bus != &pci_bus_type); 45 return pci_dma_supported(to_pci_dev(dev), mask); 49 dma_set_mask(struct device *dev, u64 dma_mask) argument 51 BUG_ON(dev->bus != &pci_bus_type); 53 return pci_set_dma_mask(to_pci_dev(dev), dma_mask); 57 dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, argument 60 BUG_ON(dev->bus != &pci_bus_type); 62 return pci_alloc_consistent(to_pci_dev(dev), size, dma_handle); 66 dma_free_coherent(struct device *dev, size_ argument 75 dma_map_single(struct device *dev, void *cpu_addr, size_t size, enum dma_data_direction direction) argument 84 dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, enum dma_data_direction direction) argument 93 dma_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction direction) argument 103 dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, enum dma_data_direction direction) argument 112 dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction) argument 121 dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, enum dma_data_direction direction) argument 130 dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument 140 dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument 150 dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction) argument 159 dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction) argument 168 dma_mapping_error(struct device *dev, dma_addr_t dma_addr) argument 177 dma_supported(struct device *dev, u64 mask) argument 183 dma_set_mask(struct device *dev, u64 dma_mask) argument 190 dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t flag) argument 198 dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr_t dma_handle) argument 205 dma_map_single(struct device *dev, void *cpu_addr, size_t size, enum dma_data_direction direction) argument 213 dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, enum dma_data_direction direction) argument 220 dma_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction direction) argument 229 dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, enum dma_data_direction direction) argument 236 dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction) argument 244 dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, enum dma_data_direction direction) argument 251 dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument 258 dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument 265 dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction) argument 272 dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction) argument 279 dma_mapping_error(struct device *dev, dma_addr_t dma_addr) argument 301 dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, unsigned long offset, size_t size, enum dma_data_direction direction) argument 310 dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, unsigned long offset, size_t size, enum dma_data_direction direction) argument 319 dma_cache_sync(struct device *dev, void *vaddr, size_t size, enum dma_data_direction direction) argument [all...] |