/linux-master/arch/mips/kernel/ |
H A D | rtlx.c | 22 #include <asm/vpe.h> 68 void rtlx_starting(int vpe) argument 81 void rtlx_stopping(int vpe) argument
|
H A D | Makefile | 65 obj-$(CONFIG_MIPS_VPE_LOADER) += vpe.o 66 obj-$(CONFIG_MIPS_VPE_LOADER_MT) += vpe-mt.o
|
H A D | cpu-probe.c | 2033 void cpu_set_vpe_id(struct cpuinfo_mips *cpuinfo, unsigned int vpe) argument 2036 WARN_ON(vpe > (MIPS_GLOBALNUMBER_VP >> MIPS_GLOBALNUMBER_VP_SHF)); 2039 WARN_ON(vpe && !IS_ENABLED(CONFIG_MIPS_MT_SMP) && 2043 cpuinfo->globalnumber |= vpe << MIPS_GLOBALNUMBER_VP_SHF;
|
H A D | perf_event_mipsxx.c | 98 #define M_PERFCTL_VPEID(vpe) ((vpe) << MIPS_PERFCTRL_VPEID_S)
|
H A D | rtlx-mt.c | 18 #include <asm/vpe.h>
|
H A D | vpe-mt.c | 18 #include <asm/vpe.h> 26 int vpe_run(struct vpe *v) 176 /* give me a vpe */ 180 struct vpe *v; 182 /* find a vpe */ 195 int vpe_start(void *vpe, unsigned long start) argument 197 struct vpe *v = vpe; 205 int vpe_stop(void *vpe) argument 207 struct vpe * 226 vpe_free(void *vpe) argument 263 struct vpe *vpe = get_vpe(aprp_cpu_index()); local 281 struct vpe *vpe = get_vpe(aprp_cpu_index()); local 289 struct vpe *vpe = get_vpe(aprp_cpu_index()); local 311 ATTRIBUTE_GROUPS(vpe); variable [all...] |
H A D | vpe.c | 36 #include <asm/vpe.h> 52 /* get the vpe associated with this minor */ 53 struct vpe *get_vpe(int minor) 55 struct vpe *res, *v; 73 /* get the vpe associated with this minor */ 91 /* allocate a vpe and associate it with this minor (or index) */ 92 struct vpe *alloc_vpe(int minor) 94 struct vpe *v; 96 v = kzalloc(sizeof(struct vpe), GFP_KERNEL); 133 void release_vpe(struct vpe * [all...] |
/linux-master/arch/arm64/kvm/vgic/ |
H A D | vgic-its.c | 382 if (map.vpe) 383 atomic_dec(&map.vpe->vlpi_count); 384 map.vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe; 385 atomic_inc(&map.vpe->vlpi_count);
|
H A D | vgic-v4.c | 108 static void vgic_v4_sync_sgi_config(struct its_vpe *vpe, struct vgic_irq *irq) argument 110 vpe->sgi_config[irq->intid].enabled = irq->enabled; 111 vpe->sgi_config[irq->intid].group = irq->group; 112 vpe->sgi_config[irq->intid].priority = irq->priority; 117 struct its_vpe *vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe; local 137 irq->host_irq = irq_find_mapping(vpe->sgi_domain, i); 140 vgic_v4_sync_sgi_config(vpe, irq); 215 struct its_vpe *vpe = &irq->target_vcpu->arch.vgic_cpu.vgic_v3.its_vpe; local 220 va = page_address(vpe->vpt_page); 341 struct its_vpe *vpe local 351 struct its_vpe *vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe; local 387 struct its_vpe *vpe = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe; local [all...] |
/linux-master/arch/mips/include/asm/ |
H A D | cpu-info.h | 179 extern void cpu_set_vpe_id(struct cpuinfo_mips *cpuinfo, unsigned int vpe);
|
H A D | rtlx.h | 30 void rtlx_starting(int vpe); 31 void rtlx_stopping(int vpe);
|
H A D | smp-cps.h | 30 extern void mips_cps_boot_vpes(struct core_boot_config *cfg, unsigned vpe);
|
H A D | vpe.h | 17 #define VPE_MODULE_NAME "vpe" 49 struct vpe { struct 52 /* (device) minor associated with this vpe */ 63 /* tc's associated with this vpe */ 66 /* The list of vpe's */ 82 struct vpe *pvpe; /* parent VPE */ 88 void (*start)(int vpe); 89 void (*stop)(int vpe); 108 struct vpe *get_vpe(int minor); 110 struct vpe *alloc_vp [all...] |
/linux-master/arch/mips/lantiq/ |
H A D | irq.c | 50 #define ltq_icu_w32(vpe, m, x, y) \ 51 ltq_w32((x), ltq_icu_membase[vpe] + m*LTQ_ICU_IM_SIZE + (y)) 53 #define ltq_icu_r32(vpe, m, x) \ 54 ltq_r32(ltq_icu_membase[vpe] + m*LTQ_ICU_IM_SIZE + (x)) 83 int vpe; local 88 for_each_present_cpu(vpe) { 89 ltq_icu_w32(vpe, im, 90 ltq_icu_r32(vpe, im, LTQ_ICU_IER) & ~BIT(offset), 101 int vpe; local 106 for_each_present_cpu(vpe) { 120 int vpe; local 136 int vpe; local 291 int vpe = smp_processor_id(); local 342 int i, ret, vpe; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu.h | 1009 /* vpe */ 1010 struct amdgpu_vpe vpe; member in struct:amdgpu_device
|
H A D | amdgpu_discovery.c | 1324 if (adev->vpe.num_instances < AMDGPU_MAX_VPE_INSTANCES) 1325 adev->vpe.num_instances++; 1328 adev->vpe.num_instances + 1,
|
H A D | amdgpu_kms.c | 394 fw_info->ver = adev->vpe.fw_version; 395 fw_info->feature = adev->vpe.feature_version; 516 if (adev->vpe.ring.sched.ready)
|
H A D | amdgpu_umsch_mm.c | 192 if (adev->vpe.collaborate_mode) 293 if (adev->vpe.collaborate_mode) 304 dev_err(adev->dev, "vpe queue submission timeout\n");
|
H A D | amdgpu_vpe.c | 121 int amdgpu_vpe_configure_dpm(struct amdgpu_vpe *vpe) argument 123 struct amdgpu_device *adev = vpe->ring.adev; 134 dpm_ctl = RREG32(vpe_get_reg_offset(vpe, 0, vpe->regs.dpm_enable)); 136 WREG32(vpe_get_reg_offset(vpe, 0, vpe->regs.dpm_enable), dpm_ctl); 147 /* vpe dpm only cares 4 levels. */ 190 WREG32(vpe_get_reg_offset(vpe, 0, vpe->regs.dpm_pratio), pratio_ctl); /* PRatio */ 191 WREG32(vpe_get_reg_offset(vpe, 222 amdgpu_vpe_init_microcode(struct amdgpu_vpe *vpe) argument 264 amdgpu_vpe_ring_init(struct amdgpu_vpe *vpe) argument 284 amdgpu_vpe_ring_fini(struct amdgpu_vpe *vpe) argument 294 struct amdgpu_vpe *vpe = &adev->vpe; local 330 vpe_common_init(struct amdgpu_vpe *vpe) argument 354 struct amdgpu_vpe *vpe = &adev->vpe; local 379 struct amdgpu_vpe *vpe = &adev->vpe; local 396 struct amdgpu_vpe *vpe = &adev->vpe; local 419 struct amdgpu_vpe *vpe = &adev->vpe; local 597 struct amdgpu_vpe *vpe = &adev->vpe; local 645 struct amdgpu_vpe *vpe = &adev->vpe; local 665 struct amdgpu_vpe *vpe = &adev->vpe; local 684 struct amdgpu_vpe *vpe = &adev->vpe; local 703 struct amdgpu_vpe *vpe = &adev->vpe; local 833 struct amdgpu_vpe *vpe = &adev->vpe; local [all...] |
H A D | amdgpu_vpe.h | 35 uint32_t (*get_reg_offset)(struct amdgpu_vpe *vpe, uint32_t inst, uint32_t offset); 36 int (*set_regs)(struct amdgpu_vpe *vpe); 37 int (*irq_init)(struct amdgpu_vpe *vpe); 38 int (*init_microcode)(struct amdgpu_vpe *vpe); 39 int (*load_microcode)(struct amdgpu_vpe *vpe); 40 int (*ring_init)(struct amdgpu_vpe *vpe); 41 int (*ring_start)(struct amdgpu_vpe *vpe); 42 int (*ring_stop)(struct amdgpu_vpe *vpe); 43 int (*ring_fini)(struct amdgpu_vpe *vpe); 85 int amdgpu_vpe_init_microcode(struct amdgpu_vpe *vpe); [all...] |
H A D | umsch_mm_v4_0.c | 277 adev->vpe.collaborate_mode ? 0x3 : 0x0; 353 add_queue.collaboration_mode = adev->vpe.collaborate_mode ? 1 : 0;
|
H A D | vpe_v6_1.c | 31 #include "ivsrcid/vpe/irqsrcs_vpe_6_1.h" 32 #include "vpe/vpe_6_1_0_offset.h" 33 #include "vpe/vpe_6_1_0_sh_mask.h" 63 static uint32_t vpe_v6_1_get_reg_offset(struct amdgpu_vpe *vpe, uint32_t inst, uint32_t offset) argument 67 base = vpe->ring.adev->reg_offset[VPE_HWIP][inst][0]; 72 static void vpe_v6_1_halt(struct amdgpu_vpe *vpe, bool halt) argument 74 struct amdgpu_device *adev = vpe->ring.adev; 77 for (i = 0; i < vpe->num_instances; i++) { 78 f32_cntl = RREG32(vpe_get_reg_offset(vpe, i, regVPEC_F32_CNTL)); 81 WREG32(vpe_get_reg_offset(vpe, 85 vpe_v6_1_irq_init(struct amdgpu_vpe *vpe) argument 99 vpe_v6_1_set_collaborate_mode(struct amdgpu_vpe *vpe, bool enable) argument 122 vpe_v6_1_load_microcode(struct amdgpu_vpe *vpe) argument 205 vpe_v6_1_ring_start(struct amdgpu_vpe *vpe) argument 275 vpe_v_6_1_ring_stop(struct amdgpu_vpe *vpe) argument 313 struct amdgpu_vpe *vpe = &adev->vpe; local 350 vpe_v6_1_set_regs(struct amdgpu_vpe *vpe) argument 393 vpe_v6_1_set_funcs(struct amdgpu_vpe *vpe) argument [all...] |
H A D | vpe_v6_1.h | 27 void vpe_v6_1_set_funcs(struct amdgpu_vpe *vpe);
|
/linux-master/drivers/gpu/drm/ingenic/ |
H A D | ingenic-drm-drv.c | 284 unsigned int vpe, vds, vde, vt, hpe, hds, hde, ht; local 286 vpe = mode->crtc_vsync_end - mode->crtc_vsync_start; 298 vpe << JZ_LCD_VSYNC_VPE_OFFSET); 332 (ht * vpe / 3) << JZ_LCD_IPUR_IPUR_LSB);
|
/linux-master/drivers/irqchip/ |
H A D | irq-gic-v3-its.c | 266 static int vpe_to_cpuid_lock(struct its_vpe *vpe, unsigned long *flags) argument 268 raw_spin_lock_irqsave(&vpe->vpe_lock, *flags); 269 return vpe->col_idx; 272 static void vpe_to_cpuid_unlock(struct its_vpe *vpe, unsigned long flags) argument 274 raw_spin_unlock_irqrestore(&vpe->vpe_lock, flags); 281 struct its_vpe *vpe = NULL; local 285 vpe = irq_data_get_irq_chip_data(d); 289 vpe = map->vpe; 292 if (vpe) { 307 struct its_vpe *vpe = NULL; local 329 valid_vpe(struct its_node *its, struct its_vpe *vpe) argument 390 struct its_vpe *vpe; member in struct:its_cmd_desc::__anon333::__anon343 394 struct its_vpe *vpe; member in struct:its_cmd_desc::__anon333::__anon344 400 struct its_vpe *vpe; member in struct:its_cmd_desc::__anon333::__anon345 408 struct its_vpe *vpe; member in struct:its_cmd_desc::__anon333::__anon346 415 struct its_vpe *vpe; member in struct:its_cmd_desc::__anon333::__anon347 422 struct its_vpe *vpe; member in struct:its_cmd_desc::__anon333::__anon348 426 struct its_vpe *vpe; member in struct:its_cmd_desc::__anon333::__anon349 789 struct its_vpe *vpe = valid_vpe(its, desc->its_vmapp_cmd.vpe); local 1304 its_send_vmapp(struct its_node *its, struct its_vpe *vpe, bool valid) argument 1316 its_send_vmovp(struct its_vpe *vpe) argument 1360 its_send_vinvall(struct its_node *its, struct its_vpe *vpe) argument 1410 its_send_invdb(struct its_node *its, struct its_vpe *vpe) argument 1811 struct its_vpe *vpe = vm->vpes[i]; local 3736 its_vpe_db_proxy_unmap_locked(struct its_vpe *vpe) argument 3762 its_vpe_db_proxy_unmap(struct its_vpe *vpe) argument 3777 its_vpe_db_proxy_map_locked(struct its_vpe *vpe) argument 3800 its_vpe_db_proxy_move(struct its_vpe *vpe, int from, int to) argument 3834 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 3898 its_vpe_schedule(struct its_vpe *vpe) argument 3934 its_vpe_deschedule(struct its_vpe *vpe) argument 3945 its_vpe_invall(struct its_vpe *vpe) argument 3967 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 3992 its_vpe_send_cmd(struct its_vpe *vpe, void (*cmd)(struct its_device *, u32)) argument 4007 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4038 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4098 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4123 its_vpe_4_1_schedule(struct its_vpe *vpe, struct its_cmd_info *info) argument 4138 its_vpe_4_1_deschedule(struct its_vpe *vpe, struct its_cmd_info *info) argument 4175 its_vpe_4_1_invall(struct its_vpe *vpe) argument 4198 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4234 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4254 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4262 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4289 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4306 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4358 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4387 struct its_vpe *vpe = args; local 4424 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4458 its_vpe_init(struct its_vpe *vpe) argument 4492 its_vpe_teardown(struct its_vpe *vpe) argument 4511 struct its_vpe *vpe = irq_data_get_irq_chip_data(data); local 4582 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local 4612 struct its_vpe *vpe = irq_data_get_irq_chip_data(d); local [all...] |