/linux-master/drivers/gpu/drm/radeon/ |
H A D | ni_dma.c | 319 unsigned ndw; local 322 ndw = count * 2; 323 if (ndw > 0xFFFFE) 324 ndw = 0xFFFFE; 327 0, 0, ndw); 333 pe += ndw * 4; 334 src += ndw * 4; 335 count -= ndw / 2; 359 unsigned ndw; local 362 ndw 407 unsigned ndw; local [all...] |
H A D | si_dma.c | 111 unsigned ndw; local 114 ndw = count * 2; 115 if (ndw > 0xFFFFE) 116 ndw = 0xFFFFE; 119 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, ndw); 122 for (; ndw > 0; ndw -= 2, --count, pe += 8) { 158 unsigned ndw; local 161 ndw = count * 2; 162 if (ndw > [all...] |
H A D | radeon_ring.c | 102 * @ndw: number of dwords to allocate in the ring buffer 104 * Allocate @ndw dwords in the ring buffer (all asics). 107 int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) argument 112 if (ndw > (ring->ring_size / 4)) 117 ndw = (ndw + ring->align_mask) & ~ring->align_mask; 118 while (ndw > (ring->ring_free_dw - 1)) { 120 if (ndw < ring->ring_free_dw) { 127 ring->count_dw = ndw; 137 * @ndw 143 radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) argument [all...] |
H A D | radeon_vm.c | 647 unsigned count = 0, pt_idx, ndw; local 652 ndw = 64; 655 ndw += vm->max_pde_used * 6; 658 if (ndw > 0xfffff) 661 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, ndw * 4); 705 WARN_ON(ib.length_dw > ndw); 917 unsigned nptes, ncmds, ndw; local 973 ndw = 64; 978 ndw += ncmds * 7; 982 ndw [all...] |
H A D | cik_sdma.c | 847 unsigned ndw; local 850 ndw = count * 2; 851 if (ndw > 0xFFFFE) 852 ndw = 0xFFFFE; 859 ib->ptr[ib->length_dw++] = ndw; 860 for (; ndw > 0; ndw -= 2, --count, pe += 8) { 896 unsigned ndw; local 899 ndw = count; 900 if (ndw > [all...] |
H A D | r100.c | 912 unsigned ndw; local 924 ndw = 64 + (10 * num_loops); 925 r = radeon_ring_lock(rdev, ring, ndw); 927 DRM_ERROR("radeon: moving bo (%d) asking for %u dw.\n", r, ndw);
|
H A D | radeon.h | 979 int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *cp, unsigned ndw); 980 int radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *cp, unsigned ndw);
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm_sdma.c | 58 unsigned int ndw; local 62 ndw = AMDGPU_VM_SDMA_MIN_NUM_DW; 64 ndw += count * 2; 65 ndw = min(ndw, AMDGPU_VM_SDMA_MAX_NUM_DW); 68 ndw * 4, pool, &p->job); 72 p->num_dw_left = ndw; 233 unsigned int i, ndw, nptes; local 252 ndw = p->num_dw_left; 253 ndw [all...] |
H A D | si_dma.c | 335 unsigned ndw = count * 2; local 337 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, ndw); 340 for (; ndw > 0; ndw -= 2) { 365 unsigned ndw; local 368 ndw = count * 2; 369 if (ndw > 0xFFFFE) 370 ndw = 0xFFFFE; 378 ib->ptr[ib->length_dw++] = DMA_PTE_PDE_PACKET(ndw); 387 pe += ndw * [all...] |
H A D | amdgpu_ring.c | 76 * @ndw: number of dwords to allocate in the ring buffer 78 * Allocate @ndw dwords in the ring buffer (all asics). 81 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigned int ndw) argument 85 ndw = (ndw + ring->funcs->align_mask) & ~ring->funcs->align_mask; 90 if (WARN_ON_ONCE(ndw > ring->max_dw)) 93 ring->count_dw = ndw;
|
H A D | amdgpu_gmc.c | 686 unsigned int ndw; local 710 ndw = kiq->pmf->invalidate_tlbs_size + 8; 713 ndw += kiq->pmf->invalidate_tlbs_size; 716 ndw += kiq->pmf->invalidate_tlbs_size; 719 amdgpu_ring_alloc(ring, ndw);
|
H A D | cik_sdma.c | 746 unsigned ndw = count * 2; local 752 ib->ptr[ib->length_dw++] = ndw; 753 for (; ndw > 0; ndw -= 2) {
|
H A D | sdma_v2_4.c | 681 unsigned ndw = count * 2; local 687 ib->ptr[ib->length_dw++] = ndw; 688 for (; ndw > 0; ndw -= 2) {
|
H A D | sdma_v3_0.c | 954 unsigned ndw = count * 2; local 960 ib->ptr[ib->length_dw++] = ndw; 961 for (; ndw > 0; ndw -= 2) {
|
H A D | mes_v10_1.c | 93 int ndw = size / 4; local 104 if (amdgpu_ring_alloc(ring, ndw)) { 113 amdgpu_ring_write_multiple(ring, pkt, ndw);
|
H A D | mes_v11_0.c | 107 int ndw = size / 4; local 125 if (amdgpu_ring_alloc(ring, ndw)) { 134 amdgpu_ring_write_multiple(ring, pkt, ndw);
|
H A D | sdma_v5_0.c | 1175 unsigned ndw = count * 2; local 1181 ib->ptr[ib->length_dw++] = ndw - 1; 1182 for (; ndw > 0; ndw -= 2) {
|
H A D | sdma_v6_0.c | 1029 unsigned ndw = count * 2; local 1035 ib->ptr[ib->length_dw++] = ndw - 1; 1036 for (; ndw > 0; ndw -= 2) {
|
H A D | sdma_v4_4_2.c | 1109 unsigned ndw = count * 2; local 1115 ib->ptr[ib->length_dw++] = ndw - 1; 1116 for (; ndw > 0; ndw -= 2) {
|
H A D | sdma_v5_2.c | 1014 unsigned ndw = count * 2; local 1020 ib->ptr[ib->length_dw++] = ndw - 1; 1021 for (; ndw > 0; ndw -= 2) {
|
H A D | amdgpu_ring.h | 339 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigned ndw);
|
H A D | sdma_v4_0.c | 1571 unsigned ndw = count * 2; local 1577 ib->ptr[ib->length_dw++] = ndw - 1; 1578 for (; ndw > 0; ndw -= 2) {
|
/linux-master/drivers/crypto/aspeed/ |
H A D | aspeed-acry.c | 253 int nbits, ndw; local 273 ndw = DIV_ROUND_UP(nbytes, BYTES_PER_DWORD); 279 for (j = ndw; j > 0; j--) {
|
/linux-master/kernel/rcu/ |
H A D | tree_nocb.h | 1036 int ndw; variable 1044 ndw = rdp_gp->nocb_defer_wakeup; 1045 ret = __wake_nocb_gp(rdp_gp, rdp, ndw == RCU_NOCB_WAKE_FORCE, flags);
|