Searched refs:dev (Results 1 - 25 of 700) sorted by relevance

1234567891011>>

/netgear-R7800-V1.0.2.28/target/linux/generic/files/fs/yaffs2/
H A Dyaffs_checkptrw.c17 static int yaffs2_checkpt_space_ok(yaffs_dev_t *dev) argument
19 int blocksAvailable = dev->n_erased_blocks - dev->param.n_reserved_blocks;
29 static int yaffs_checkpt_erase(yaffs_dev_t *dev) argument
33 if (!dev->param.erase_fn)
36 dev->internal_start_block, dev->internal_end_block));
38 for (i = dev->internal_start_block; i <= dev->internal_end_block; i++) {
39 yaffs_block_info_t *bi = yaffs_get_block_info(dev,
62 yaffs2_checkpt_find_erased_block(yaffs_dev_t *dev) argument
90 yaffs2_checkpt_find_block(yaffs_dev_t *dev) argument
126 yaffs2_checkpt_open(yaffs_dev_t *dev, int forWriting) argument
180 yaffs2_get_checkpt_sum(yaffs_dev_t *dev, __u32 *sum) argument
188 yaffs2_checkpt_flush_buffer(yaffs_dev_t *dev) argument
241 yaffs2_checkpt_wr(yaffs_dev_t *dev, const void *data, int n_bytes) argument
276 yaffs2_checkpt_rd(yaffs_dev_t *dev, void *data, int n_bytes) argument
351 yaffs_checkpt_close(yaffs_dev_t *dev) argument
390 yaffs2_checkpt_invalidate_stream(yaffs_dev_t *dev) argument
[all...]
H A Dyaffs_yaffs2.h19 void yaffs_calc_oldest_dirty_seq(yaffs_dev_t *dev);
20 void yaffs2_find_oldest_dirty_seq(yaffs_dev_t *dev);
21 void yaffs2_clear_oldest_dirty_seq(yaffs_dev_t *dev, yaffs_block_info_t *bi);
22 void yaffs2_update_oldest_dirty_seq(yaffs_dev_t *dev, unsigned block_no, yaffs_block_info_t *bi);
23 int yaffs_block_ok_for_gc(yaffs_dev_t *dev, yaffs_block_info_t *bi);
24 __u32 yaffs2_find_refresh_block(yaffs_dev_t *dev);
25 int yaffs2_checkpt_required(yaffs_dev_t *dev);
26 int yaffs_calc_checkpt_blocks_required(yaffs_dev_t *dev);
29 void yaffs2_checkpt_invalidate(yaffs_dev_t *dev);
30 int yaffs2_checkpt_save(yaffs_dev_t *dev);
[all...]
H A Dyaffs_yaffs1.h20 int yaffs1_scan(yaffs_dev_t *dev);
H A Dyaffs_allocator.h21 void yaffs_init_raw_tnodes_and_objs(yaffs_dev_t *dev);
22 void yaffs_deinit_raw_tnodes_and_objs(yaffs_dev_t *dev);
24 yaffs_tnode_t *yaffs_alloc_raw_tnode(yaffs_dev_t *dev);
25 void yaffs_free_raw_tnode(yaffs_dev_t *dev, yaffs_tnode_t *tn);
27 yaffs_obj_t *yaffs_alloc_raw_obj(yaffs_dev_t *dev);
28 void yaffs_free_raw_obj(yaffs_dev_t *dev, yaffs_obj_t *obj);
H A Dyaffs_checkptrw.h21 int yaffs2_checkpt_open(yaffs_dev_t *dev, int forWriting);
23 int yaffs2_checkpt_wr(yaffs_dev_t *dev, const void *data, int n_bytes);
25 int yaffs2_checkpt_rd(yaffs_dev_t *dev, void *data, int n_bytes);
27 int yaffs2_get_checkpt_sum(yaffs_dev_t *dev, __u32 *sum);
29 int yaffs_checkpt_close(yaffs_dev_t *dev);
31 int yaffs2_checkpt_invalidate_stream(yaffs_dev_t *dev);
H A Dyaffs_nand.c20 int yaffs_rd_chunk_tags_nand(yaffs_dev_t *dev, int nand_chunk, argument
27 int realignedChunkInNAND = nand_chunk - dev->chunk_offset;
29 dev->n_page_reads++;
35 if (dev->param.read_chunk_tags_fn)
36 result = dev->param.read_chunk_tags_fn(dev, realignedChunkInNAND, buffer,
39 result = yaffs_tags_compat_rd(dev,
47 bi = yaffs_get_block_info(dev, nand_chunk/dev->param.chunks_per_block);
48 yaffs_handle_chunk_error(dev, b
54 yaffs_wr_chunk_tags_nand(yaffs_dev_t *dev, int nand_chunk, const __u8 *buffer, yaffs_ext_tags *tags) argument
91 yaffs_mark_bad(yaffs_dev_t *dev, int block_no) argument
102 yaffs_query_init_block_state(yaffs_dev_t *dev, int block_no, yaffs_block_state_t *state, __u32 *seq_number) argument
118 yaffs_erase_block(struct yaffs_dev_s *dev, int flash_block) argument
132 yaffs_init_nand(struct yaffs_dev_s *dev) argument
[all...]
H A Dyaffs_bitmap.c20 static Y_INLINE __u8 *yaffs_BlockBits(yaffs_dev_t *dev, int blk) argument
22 if (blk < dev->internal_start_block || blk > dev->internal_end_block) {
28 return dev->chunk_bits +
29 (dev->chunk_bit_stride * (blk - dev->internal_start_block));
32 void yaffs_verify_chunk_bit_id(yaffs_dev_t *dev, int blk, int chunk) argument
34 if (blk < dev->internal_start_block || blk > dev->internal_end_block ||
35 chunk < 0 || chunk >= dev
43 yaffs_clear_chunk_bits(yaffs_dev_t *dev, int blk) argument
50 yaffs_clear_chunk_bit(yaffs_dev_t *dev, int blk, int chunk) argument
59 yaffs_set_chunk_bit(yaffs_dev_t *dev, int blk, int chunk) argument
68 yaffs_check_chunk_bit(yaffs_dev_t *dev, int blk, int chunk) argument
76 yaffs_still_some_chunks(yaffs_dev_t *dev, int blk) argument
88 yaffs_count_chunk_bits(yaffs_dev_t *dev, int blk) argument
[all...]
H A Dyaffs_bitmap.h23 void yaffs_verify_chunk_bit_id(yaffs_dev_t *dev, int blk, int chunk);
24 void yaffs_clear_chunk_bits(yaffs_dev_t *dev, int blk);
25 void yaffs_clear_chunk_bit(yaffs_dev_t *dev, int blk, int chunk);
26 void yaffs_set_chunk_bit(yaffs_dev_t *dev, int blk, int chunk);
27 int yaffs_check_chunk_bit(yaffs_dev_t *dev, int blk, int chunk);
28 int yaffs_still_some_chunks(yaffs_dev_t *dev, int blk);
29 int yaffs_count_chunk_bits(yaffs_dev_t *dev, int blk);
H A Dyaffs_getblockinfo.h23 static Y_INLINE yaffs_block_info_t *yaffs_get_block_info(yaffs_dev_t * dev, int blk) argument
25 if (blk < dev->internal_start_block || blk > dev->internal_end_block) {
32 return &dev->block_info[blk - dev->internal_start_block];
H A Dyaffs_nand.h22 int yaffs_rd_chunk_tags_nand(yaffs_dev_t *dev, int nand_chunk,
26 int yaffs_wr_chunk_tags_nand(yaffs_dev_t *dev,
31 int yaffs_mark_bad(yaffs_dev_t *dev, int block_no);
33 int yaffs_query_init_block_state(yaffs_dev_t *dev,
38 int yaffs_erase_block(struct yaffs_dev_s *dev,
41 int yaffs_init_nand(struct yaffs_dev_s *dev);
H A Dyaffs_nandemul2k.h23 int nandemul2k_WriteChunkWithTagsToNAND(struct yaffs_dev_s *dev,
26 int nandemul2k_ReadChunkWithTagsFromNAND(struct yaffs_dev_s *dev,
29 int nandemul2k_MarkNANDBlockBad(struct yaffs_dev_s *dev, int block_no);
30 int nandemul2k_QueryNANDBlock(struct yaffs_dev_s *dev, int block_no,
32 int nandemul2k_EraseBlockInNAND(struct yaffs_dev_s *dev,
34 int nandemul2k_InitialiseNAND(struct yaffs_dev_s *dev);
H A Dyaffs_mtdif1.h17 int nandmtd1_WriteChunkWithTagsToNAND(yaffs_dev_t *dev, int nand_chunk,
20 int nandmtd1_ReadChunkWithTagsFromNAND(yaffs_dev_t *dev, int nand_chunk,
23 int nandmtd1_MarkNANDBlockBad(struct yaffs_dev_s *dev, int block_no);
25 int nandmtd1_QueryNANDBlock(struct yaffs_dev_s *dev, int block_no,
H A Dyaffs_mtdif2.h20 int nandmtd2_WriteChunkWithTagsToNAND(yaffs_dev_t *dev, int nand_chunk,
23 int nandmtd2_ReadChunkWithTagsFromNAND(yaffs_dev_t *dev, int nand_chunk,
25 int nandmtd2_MarkNANDBlockBad(struct yaffs_dev_s *dev, int block_no);
26 int nandmtd2_QueryNANDBlock(struct yaffs_dev_s *dev, int block_no,
/netgear-R7800-V1.0.2.28/package/iproute2/files/
H A Dtc.debug3 for dev in $(ls -1 /sys/class/net/); do
4 [ -d /sys/class/net/${dev} ] || continue
5 log tc -s qdisc show dev ${dev}
/netgear-R7800-V1.0.2.28/package/qca-hostap/files/
H A Dwps-reset-hotplug.sh20 echo "" > /dev/console
21 echo "RESET TO FACTORY SETTING EVENT DETECTED" > /dev/console
22 echo "PLEASE WAIT WHILE REBOOTING THE DEVICE..." > /dev/console
/netgear-R7800-V1.0.2.28/package/qca-ssdk/src/include/ref/
H A Dref_vlan.h23 qca_ar8327_sw_set_vlan(struct switch_dev *dev,
28 qca_ar8327_sw_get_vlan(struct switch_dev *dev,
33 qca_ar8327_sw_set_vid(struct switch_dev *dev,
38 qca_ar8327_sw_get_vid(struct switch_dev *dev,
43 qca_ar8327_sw_get_pvid(struct switch_dev *dev, int port, int *vlan);
46 qca_ar8327_sw_set_pvid(struct switch_dev *dev, int port, int vlan);
49 qca_ar8327_sw_get_ports(struct switch_dev *dev, struct switch_val *val);
52 qca_ar8327_sw_set_ports(struct switch_dev *dev, struct switch_val *val);
55 qca_ar8327_sw_hw_apply(struct switch_dev *dev);
H A Dref_port_ctrl.h23 qca_ar8327_sw_get_port_link(struct switch_dev *dev, int port,
27 qca_ar8327_sw_mac_polling_task(struct switch_dev *dev);
H A Dref_mib.h23 qca_ar8327_sw_set_reset_mibs(struct switch_dev *dev,
28 qca_ar8327_sw_set_port_reset_mib(struct switch_dev *dev,
34 qca_ar8327_sw_get_port_mib(struct switch_dev *dev,
39 qca_ar8327_sw_mib_task(struct switch_dev *dev);
/netgear-R7800-V1.0.2.28/target/linux/generic/files/crypto/ocf/kirkwood/mvHal/kw_family/ctrlEnv/sys/
H A DmvSysUsb.h83 #define MV_USB_BRIDGE_CTRL_REG(dev) (USB_REG_BASE(dev) + 0x300)
85 #define MV_USB_WIN_CTRL_REG(dev, win) (USB_REG_BASE(dev) + 0x320 + ((win)<<4))
86 #define MV_USB_WIN_BASE_REG(dev, win) (USB_REG_BASE(dev) + 0x324 + ((win)<<4))
110 #define MV_USB_BRIDGE_IPG_REG(dev) (USB_REG_BASE(dev) + 0x360)
113 MV_STATUS mvUsbInit(int dev, MV_BOOL isHost);
115 MV_STATUS mvUsbWinInit(int dev);
[all...]
/netgear-R7800-V1.0.2.28/target/linux/omap35xx/gumstix/base-files/lib/preinit/
H A D95_ttyS1_noecho5 stty -echo < /dev/ttyO1
/netgear-R7800-V1.0.2.28/target/linux/ubicom32/files/arch/ubicom32/kernel/
H A Dtimer_broadcast.c58 struct clock_event_device *dev = &__get_cpu_var(local_clockevent); local
60 dev->event_handler(dev);
67 static int dummy_timer_set_next_event(unsigned long cycles, struct clock_event_device *dev) argument
87 struct clock_event_device *dev = &per_cpu(local_clockevent, cpu); local
89 dev->name = "timer-dummy";
90 dev->features = CLOCK_EVT_FEAT_DUMMY;
91 dev->rating = 200;
92 dev->mult = 1;
93 dev
[all...]
H A Dtimer_device.c60 static int timer_device_set_next_event(unsigned long cycles, struct clock_event_device *dev) argument
62 timer_set(dev->irq, cycles);
70 static void timer_device_set_mode(enum clock_event_mode mode, struct clock_event_device *dev) argument
78 printk(KERN_NOTICE "timer[%d]: shutdown\n", dev->irq);
79 ldsr_disable_vector(dev->irq);
87 printk(KERN_NOTICE "timer[%d]: oneshot\n", dev->irq);
88 ldsr_disable_vector(dev->irq);
96 dev->irq, frequency / CONFIG_HZ);
97 timer_set(dev->irq, frequency / CONFIG_HZ);
103 dev
117 struct clock_event_device *dev = (struct clock_event_device *)dev_id; local
174 struct clock_event_device *dev; local
[all...]
/netgear-R7800-V1.0.2.28/package/telnetenable/files/
H A DRtDebug.sh3 mkdir /dev/pts
4 mknod -m 666 /dev/ptmx c 5 2
5 mknod -m 666 /dev/pts/0 c 136 0
6 mknod -m 666 /dev/pts/1 c 136 1
/netgear-R7800-V1.0.2.28/target/linux/ar71xx/base-files/etc/uci-defaults/
H A Dvlan-migration6 local dev="$(uci -q get network.@switch_vlan[0].device)"
9 if [ "$dev" = "rtl8366s" ] && [ "$vlan" = 0 ]; then
/netgear-R7800-V1.0.2.28/target/linux/ubicom32/files/arch/ubicom32/include/asm/
H A Ddma-mapping.h41 dma_supported(struct device *dev, u64 mask) argument
43 BUG_ON(dev->bus != &pci_bus_type);
45 return pci_dma_supported(to_pci_dev(dev), mask);
49 dma_set_mask(struct device *dev, u64 dma_mask) argument
51 BUG_ON(dev->bus != &pci_bus_type);
53 return pci_set_dma_mask(to_pci_dev(dev), dma_mask);
57 dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, argument
60 BUG_ON(dev->bus != &pci_bus_type);
62 return pci_alloc_consistent(to_pci_dev(dev), size, dma_handle);
66 dma_free_coherent(struct device *dev, size_ argument
75 dma_map_single(struct device *dev, void *cpu_addr, size_t size, enum dma_data_direction direction) argument
84 dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, enum dma_data_direction direction) argument
93 dma_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction direction) argument
103 dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, enum dma_data_direction direction) argument
112 dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction) argument
121 dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, enum dma_data_direction direction) argument
130 dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument
140 dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument
150 dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction) argument
159 dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction) argument
168 dma_mapping_error(struct device *dev, dma_addr_t dma_addr) argument
177 dma_supported(struct device *dev, u64 mask) argument
183 dma_set_mask(struct device *dev, u64 dma_mask) argument
190 dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t flag) argument
198 dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr_t dma_handle) argument
205 dma_map_single(struct device *dev, void *cpu_addr, size_t size, enum dma_data_direction direction) argument
213 dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, enum dma_data_direction direction) argument
220 dma_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction direction) argument
229 dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, enum dma_data_direction direction) argument
236 dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction direction) argument
244 dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, enum dma_data_direction direction) argument
251 dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument
258 dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument
265 dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction) argument
272 dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, enum dma_data_direction direction) argument
279 dma_mapping_error(struct device *dev, dma_addr_t dma_addr) argument
301 dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, unsigned long offset, size_t size, enum dma_data_direction direction) argument
310 dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, unsigned long offset, size_t size, enum dma_data_direction direction) argument
319 dma_cache_sync(struct device *dev, void *vaddr, size_t size, enum dma_data_direction direction) argument
[all...]

Completed in 81 milliseconds

1234567891011>>