Searched refs:rings (Results 1 - 14 of 14) sorted by relevance

/netbsd-current/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
H A Damdgpu_test.c54 if (adev->rings[i])
55 n -= adev->rings[i]->ring_size;
H A Damdgpu_ring.c98 * This is the generic insert_nop function for rings except SDMA
113 * This is the generic pad_ib function for rings except SDMA
268 adev->rings[ring->idx] = ring;
345 DRM_ERROR("Failed to register debugfs file for rings !\n");
364 if (!(ring->adev) || !(ring->adev->rings[ring->idx]))
383 ring->adev->rings[ring->idx] = NULL;
397 * Helper for rings that don't support write and wait in a
H A Damdgpu_fence.c397 * Not all asics have all rings, so each asic will only
398 * start the fence driver on the rings it has.
501 * for all possible rings.
505 * Init the fence driver for all possible rings (all asics).
506 * Not all asics have all rings, so each asic will only
507 * start the fence driver on the rings it has using
521 * for all possible rings.
525 * Tear down the fence driver for all possible rings (all asics).
533 struct amdgpu_ring *ring = adev->rings[i];
559 * for all possible rings
[all...]
H A Damdgpu_ib.c326 * amdgpu_ib_ring_tests - test IBs on the rings
363 struct amdgpu_ring *ring = adev->rings[i];
366 /* KIQ rings don't have an IB test because we never submit IBs
H A Damdgpu_gmc.c381 ring = adev->rings[i];
H A Damdgpu_job.c81 (*job)->base.sched = &adev->rings[0]->sched;
H A Damdgpu_debugfs.c970 struct amdgpu_ring *ring = adev->rings[i];
986 struct amdgpu_ring *ring = adev->rings[i];
1154 ring = adev->rings[val];
H A Damdgpu.h424 * CP & rings.
526 #define AMDGPU_MAX_WB 128 /* Reserve at most 128 WB slots for amdgpu-owned rings. */
868 /* rings */
871 struct amdgpu_ring *rings[AMDGPU_MAX_RINGS]; member in struct:amdgpu_device
965 /* keep an lru list of rings by HW IP */
H A Damdgpu_drv.c566 * It is used to enable gfx rings that could be configured with different prioritites or equal priorities
569 "Asynchronous GFX rings that could be configured with either different priorities (HP3D ring and LP3D ring), or equal priorities (0 = disabled, 1 = enabled (default))");
1222 /* wait for all rings to drain before suspending */
1224 struct amdgpu_ring *ring = adev->rings[i];
H A Damdgpu_device.c4013 struct amdgpu_ring *ring = adev->rings[i];
4318 struct amdgpu_ring *ring = tmp_adev->rings[i];
4394 struct amdgpu_ring *ring = tmp_adev->rings[i];
H A Damdgpu_vm.c1020 ring = adev->rings[i];
1022 /* only compute rings */
H A Damdgpu_pm.c3498 struct amdgpu_ring *ring = adev->rings[i];
/netbsd-current/sys/dev/ic/
H A Ddwc_gmac.c277 * Allocate Tx and Rx rings
280 aprint_error_dev(sc->sc_dev, "could not allocate DMA rings\n");
587 void *rings; local
607 descsize, &rings, BUS_DMA_NOWAIT | BUS_DMA_COHERENT);
614 error = bus_dmamap_load(sc->sc_dmat, sc->sc_dma_ring_map, rings,
623 sc->sc_rxq.r_desc = rings;
626 /* and next rings to the TX side */
/netbsd-current/sys/external/bsd/drm2/dist/drm/i915/gvt/
H A Dcmd_parser.c424 /* rings that support this cmd: BLT/RCS/VCS/VECS */
425 u16 rings; member in struct:cmd_info
660 if (opcode == e->info->opcode && e->info->rings & BIT(ring_id))
3038 unsigned int opcode, unsigned long rings)
3043 for_each_set_bit(ring, &rings, I915_NUM_ENGINES) {
3070 e->info->opcode, e->info->rings);
3082 gvt_dbg_cmd("add %-30s op %04x flag %x devs %02x rings %02x\n",
3084 e->info->devices, e->info->rings);
3037 find_cmd_entry_any_ring(struct intel_gvt *gvt, unsigned int opcode, unsigned long rings) argument

Completed in 160 milliseconds