/linux-master/arch/x86/kvm/ |
H A D | x86.c | 2744 * a small delta (1 second) of virtual cycle time against the 9918 u64 cycle; local 9931 if (!kvm_get_walltime_and_clockread(&ts, &cycle)) 9936 clock_pairing.tsc = kvm_read_l1_tsc(vcpu, cycle); 10486 * exception and start the cycle all over. 12445 * loose cycle time. This isn't too big a deal, since the loss will be
|
/linux-master/drivers/ata/ |
H A D | libata-core.c | 1741 /* This is cycle times not frequency - watch the logic! */ 1742 if (pio > 240) /* PIO2 is 240nS per cycle */ 3328 * ata_timing_cycle2mode - find xfer mode for the specified cycle duration 3330 * @cycle: cycle duration in ns 3332 * Return matching xfer mode for @cycle. The returned mode is of 3333 * the transfer type specified by @xfer_shift. If @cycle is too 3334 * slow for @xfer_shift, 0xff is returned. If @cycle is faster 3343 u8 ata_timing_cycle2mode(unsigned int xfer_shift, int cycle) argument 3360 this_cycle = t->cycle; [all...] |
/linux-master/fs/bcachefs/ |
H A D | btree_locking.c | 86 prt_printf(out, "Found lock cycle (%u entries):", g->nr); 193 static noinline int break_cycle(struct lock_graph *g, struct printbuf *cycle) argument 203 if (cycle) { 204 print_cycle(cycle, g); 220 prt_printf(&buf, bch2_fmt(g->g->trans->c, "cycle of nofail locks")); 249 struct printbuf *cycle) 257 return break_cycle(g, cycle); 269 if (cycle) 285 int bch2_check_for_deadlock(struct btree_trans *trans, struct printbuf *cycle) argument 296 if (cycle) 248 lock_graph_descend(struct lock_graph *g, struct btree_trans *trans, struct printbuf *cycle) argument [all...] |
H A D | trace.h | 1239 const char *cycle), 1240 TP_ARGS(trans, cycle)
|
/linux-master/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 3057 * TCTT unit is cycle_ns * 64 cycle 3129 unsigned int cycle; local 3144 cycle = fec_enet_us_to_itr_clock(ndev, ec->rx_coalesce_usecs); 3145 if (cycle > 0xFFFF) { 3150 cycle = fec_enet_us_to_itr_clock(ndev, ec->tx_coalesce_usecs); 3151 if (cycle > 0xFFFF) {
|
/linux-master/drivers/firewire/ |
H A D | ohci.c | 384 ", nonatomic cycle timer = " __stringify(QUIRK_CYCLE_TIMER) 510 [0x8] = "cycle start", [0x9] = "Lk req", 1752 * iso cycle timer register: 1801 * the most significant bit of the cycle timer in bit 6 so that we can detect 1988 * If the invalid data looks like a cycle start packet, 1989 * it's likely to be the result of the cycle master 2192 ohci_notice(ohci, "isochronous cycle too long\n"); 2205 ohci_notice(ohci, "isochronous cycle inconsistent\n"); 2700 * We might be called just after the cycle timer has wrapped 3091 s32 cycle, u3 3090 ohci_start_iso(struct fw_iso_context *base, s32 cycle, u32 sync, u32 tags) argument [all...] |
/linux-master/include/linux/ |
H A D | libata.h | 1011 unsigned short cycle; /* t0 */ member in struct:ata_timing 1264 extern u8 ata_timing_cycle2mode(unsigned int xfer_shift, int cycle);
|
H A D | ptp_kvm.h | 19 int kvm_arch_ptp_get_crosststamp(u64 *cycle,
|
/linux-master/net/sched/ |
H A D | sch_taprio.c | 218 /* Get how much time has been already elapsed in the current cycle. */ 328 ktime_t earliest_txtime = KTIME_MAX, txtime, cycle, transmit_end_time; local 345 cycle = sched->cycle_time; 348 cycle_end = ktime_add_ns(curr_intv_end, cycle); 375 * first available interval in the next cycle. 379 *interval_start = ktime_add_ns(curr_intv_start, cycle); 380 *interval_end = ktime_add_ns(curr_intv_end, cycle); 386 n = div_s64(ktime_sub(txtime, curr_intv_start), cycle); 387 *interval_start = ktime_add(curr_intv_start, n * cycle); 388 *interval_end = ktime_add(curr_intv_end, n * cycle); 1149 ktime_t cycle = 0; local 1208 ktime_t now, base, cycle; local 1243 ktime_t cycle; local [all...] |
H A D | act_gate.c | 38 ktime_t now, base, cycle; local 49 cycle = param->tcfg_cycletime; 51 n = div64_u64(ktime_sub_ns(now, base), cycle); 52 *start = ktime_add_ns(base, (n + 1) * cycle); 80 /* cycle start, clear pending bit, clear total octets */ 99 ktime_t cycle, base; local 102 cycle = p->tcfg_cycletime; 104 n = div64_u64(ktime_sub_ns(now, base), cycle); 105 close_time = ktime_add_ns(base, (n + 1) * cycle); 406 ktime_t cycle local [all...] |
/linux-master/drivers/usb/host/ |
H A D | xhci-dbgcap.c | 268 u32 length, control, cycle; local 277 cycle = ring->cycle_state; 281 if (cycle) 295 * the cycle bit: 299 if (cycle) 864 * Add a barrier between reading the cycle flag and any
|
/linux-master/drivers/usb/gadget/udc/ |
H A D | tegra-xudc.c | 356 BUILD_TRB_RW(cycle, control, 0, 0x1) 1344 * and must correct trb cycle bit to the last un-enqueued state. 1359 * Retrieve the correct cycle bit state from the first trb of
|
/linux-master/drivers/staging/vme_user/ |
H A D | vme.c | 137 u32 aspace, cycle, dwidth; local 142 &aspace, &cycle, &dwidth); 149 &buf_base, &aspace, &cycle); 244 * @cycle: Required VME data transfer cycle type. 247 * address space and data transfer cycle. 252 u32 cycle) 276 ((slave_image->cycle_attr & cycle) == cycle) && 318 * @cycle 251 vme_slave_request(struct vme_dev *vdev, u32 address, u32 cycle) argument 327 vme_slave_set(struct vme_resource *resource, int enabled, unsigned long long vme_base, unsigned long long size, dma_addr_t buf_base, u32 aspace, u32 cycle) argument 377 vme_slave_get(struct vme_resource *resource, int *enabled, unsigned long long *vme_base, unsigned long long *size, dma_addr_t *buf_base, u32 *aspace, u32 *cycle) argument 449 vme_master_request(struct vme_dev *vdev, u32 address, u32 cycle, u32 dwidth) argument 528 vme_master_set(struct vme_resource *resource, int enabled, unsigned long long vme_base, unsigned long long size, u32 aspace, u32 cycle, u32 dwidth) argument 579 vme_master_get(struct vme_resource *resource, int *enabled, unsigned long long *vme_base, unsigned long long *size, u32 *aspace, u32 *cycle, u32 *dwidth) argument 1010 vme_dma_vme_attribute(unsigned long long address, u32 aspace, u32 cycle, u32 dwidth) argument 1530 vme_lm_set(struct vme_resource *resource, unsigned long long lm_base, u32 aspace, u32 cycle) argument 1566 vme_lm_get(struct vme_resource *resource, unsigned long long *lm_base, u32 *aspace, u32 *cycle) argument [all...] |
/linux-master/sound/firewire/ |
H A D | amdtp-stream.h | 105 u32 cycle; member in struct:pkt_desc 198 // To start processing content of packets at the same cycle in several contexts for
|
H A D | amdtp-stream.c | 225 // of syt interval. This comes from the interval of isoc cycle. As 1394 477 static unsigned int compute_syt_offset(unsigned int syt, unsigned int cycle, argument 480 unsigned int cycle_lo = (cycle % CYCLES_PER_SECOND) & 0x0f; 527 dst->syt_offset = compute_syt_offset(src->syt, src->cycle, transfer_delay); 680 static void build_it_pkt_header(struct amdtp_stream *s, unsigned int cycle, argument 700 trace_amdtp_packet(s, cycle, cip_header, payload_length + header_length, data_blocks, 803 static int parse_ir_ctx_header(struct amdtp_stream *s, unsigned int cycle, argument 838 // Handle the cycle so that empty packet arrives. 852 trace_amdtp_packet(s, cycle, cip_header, payload_length, *data_blocks, 872 static inline u32 increment_ohci_cycle_count(u32 cycle, unsigne argument 905 u32 cycle = compute_ohci_cycle_count(ctx_header_tstamp); local 926 unsigned int cycle; local 994 compute_syt(unsigned int syt_offset, unsigned int cycle, unsigned int transfer_delay) argument 1220 unsigned int cycle; local 1265 unsigned int cycle = compute_ohci_it_cycle(ctx_header[offset], queue_size); local 1355 unsigned int cycle; local 1394 unsigned int cycle = compute_ohci_cycle_count(ctx_header[1]); local 1490 unsigned int cycle = UINT_MAX; local 1601 unsigned int cycle = s->next_cycle; local [all...] |
/linux-master/fs/xfs/ |
H A D | xfs_sysfs.c | 407 int cycle; local 412 cycle = log->l_curr_cycle; 416 return sysfs_emit(buf, "%d:%d\n", cycle, block); 425 int cycle; local 429 xlog_crack_atomic_lsn(&log->l_tail_lsn, &cycle, &block); 430 return sysfs_emit(buf, "%d:%d\n", cycle, block); 440 int cycle; local 444 xlog_crack_grant_head(&log->l_reserve_head.grant, &cycle, &bytes); 445 return sysfs_emit(buf, "%d:%d\n", cycle, bytes); 454 int cycle; local [all...] |
H A D | xfs_log_recover.c | 264 * log which contains the given cycle. It uses a binary search algorithm. 274 uint cycle) 289 if (mid_cycle == cycle) 320 uint cycle; local 351 cycle = xlog_get_cycle(buf); 352 if (cycle == stop_on_cycle_no) { 468 * last_blk will be the 1st block # with a given cycle #. We may end 488 * LR have complete transactions. We only know that a cycle number of 489 * current cycle number -1 won't be present in the log if we start writing 493 * cycle numbe 269 xlog_find_cycle_start( struct xlog *log, char *buffer, xfs_daddr_t first_blk, xfs_daddr_t *last_blk, uint cycle) argument 1486 xlog_add_record( struct xlog *log, char *buf, int cycle, int block, int tail_cycle, int tail_block) argument 1508 xlog_write_log_records( struct xlog *log, int cycle, int start_block, int blocks, int tail_cycle, int tail_block) argument [all...] |
H A D | xfs_log.c | 154 int cycle, space; local 156 xlog_crack_grant_head_val(head_val, &cycle, &space); 161 cycle--; 165 new = xlog_assign_grant_head_val(cycle, space); 181 int cycle, space; local 183 xlog_crack_grant_head_val(head_val, &cycle, &space); 190 cycle++; 194 new = xlog_assign_grant_head_val(cycle, space); 1318 * is passed in the cycle/bytes formal parms. In the special case where 1367 xfs_alert(log->l_mp, " GH cycle 1983 uint32_t cycle = get_unaligned_be32(data + i); local 3557 int cycle, space; local [all...] |
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_log_format.h | 26 #define XLOG_HEADER_MAGIC_NUM 0xFEEDbabe /* Invalid cycle number */ 33 #define XLOG_HEADER_CYCLE_SIZE (32*1024) /* cycle data in header */ 57 static inline xfs_lsn_t xlog_assign_lsn(uint cycle, uint block) argument 59 return ((xfs_lsn_t)cycle << 32) | block; 165 __be32 h_cycle; /* write cycle of log : 4 */ 181 __be32 xh_cycle; /* write cycle of log : 4 */
|
/linux-master/drivers/pwm/ |
H A D | pwm-sl28cpld.c | 25 * - The hardware cannot generate a 100% duty cycle if the prescaler is 0. 30 * - The duty cycle will switch immediately and not after a complete cycle. 60 * We calculate the duty cycle like this: 129 unsigned int cycle, prescaler; local 152 cycle = SL28CPLD_PWM_FROM_DUTY_CYCLE(state->duty_cycle); 153 cycle = min_t(unsigned int, cycle, SL28CPLD_PWM_MAX_DUTY_CYCLE(prescaler)); 157 * cycle if the prescaler is 0. Set prescaler to 1 instead. We don't 163 if (cycle [all...] |
/linux-master/drivers/perf/ |
H A D | xgene_pmu.c | 289 XGENE_PMU_EVENT_ATTR(cycle-count, 0x00), 290 XGENE_PMU_EVENT_ATTR(cycle-count-div-64, 0x01), 307 XGENE_PMU_EVENT_ATTR(cycle-count, 0x00), 308 XGENE_PMU_EVENT_ATTR(cycle-count-div-64, 0x01), 324 XGENE_PMU_EVENT_ATTR(cycle-count, 0x00), 325 XGENE_PMU_EVENT_ATTR(cycle-count-div-64, 0x01), 334 XGENE_PMU_EVENT_ATTR(cycle-count, 0x00), 335 XGENE_PMU_EVENT_ATTR(cycle-count-div-64, 0x01), 387 XGENE_PMU_EVENT_ATTR(cycle-count, 0x00), 430 XGENE_PMU_EVENT_ATTR(cycle [all...] |
H A D | alibaba_uncore_drw_pmu.c | 199 ALI_DRW_PMU_EVENT_ATTR(cycle, 0x80),
|
/linux-master/drivers/perf/hisilicon/ |
H A D | hisi_uncore_pa_pmu.c | 324 HISI_PMU_EVENT_ATTR(cycle, 0x78),
|
/linux-master/drivers/ptp/ |
H A D | ptp_kvm_common.c | 33 u64 cycle; local 39 ret = kvm_arch_ptp_get_crosststamp(&cycle, &tspec, &cs_id); 48 system_counter->cycles = cycle;
|
H A D | ptp_kvm_x86.c | 95 int kvm_arch_ptp_get_crosststamp(u64 *cycle, struct timespec64 *tspec, argument 123 *cycle = __pvclock_read_cycles(src, clock_pair->tsc);
|