Lines Matching refs:tbl

48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl,
52 tbl->it_blocksize = 16;
53 tbl->it_base = (unsigned long)tce_mem;
54 tbl->it_page_shift = page_shift;
55 tbl->it_offset = dma_offset >> tbl->it_page_shift;
56 tbl->it_index = 0;
57 tbl->it_size = tce_size >> 3;
58 tbl->it_busno = 0;
59 tbl->it_type = TCE_PCI;
83 static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc)
85 __be64 *tmp = user ? tbl->it_userspace : (__be64 *) tbl->it_base;
86 int level = tbl->it_indirect_levels;
87 const long shift = ilog2(tbl->it_level_size);
88 unsigned long mask = (tbl->it_level_size - 1) << (level * shift);
100 tmp2 = pnv_alloc_tce_level(tbl->it_nid,
101 ilog2(tbl->it_level_size) + 3);
110 ilog2(tbl->it_level_size) + 3, 1);
124 int pnv_tce_build(struct iommu_table *tbl, long index, long npages,
129 u64 rpn = __pa(uaddr) >> tbl->it_page_shift;
137 ((rpn + i) << tbl->it_page_shift);
138 unsigned long idx = index - tbl->it_offset + i;
140 *(pnv_tce(tbl, false, idx, true)) = cpu_to_be64(newtce);
147 int pnv_tce_xchg(struct iommu_table *tbl, long index,
152 unsigned long idx = index - tbl->it_offset;
155 BUG_ON(*hpa & ~IOMMU_PAGE_MASK(tbl));
158 ptce = pnv_tce(tbl, false, idx, false);
166 ptce = pnv_tce(tbl, false, idx, true);
181 __be64 *pnv_tce_useraddrptr(struct iommu_table *tbl, long index, bool alloc)
183 if (WARN_ON_ONCE(!tbl->it_userspace))
186 return pnv_tce(tbl, true, index - tbl->it_offset, alloc);
190 void pnv_tce_free(struct iommu_table *tbl, long index, long npages)
195 unsigned long idx = index - tbl->it_offset + i;
196 __be64 *ptce = pnv_tce(tbl, false, idx, false);
202 i |= tbl->it_level_size - 1;
206 unsigned long pnv_tce_get(struct iommu_table *tbl, long index)
208 __be64 *ptce = pnv_tce(tbl, false, index - tbl->it_offset, false);
240 void pnv_pci_ioda2_table_free_pages(struct iommu_table *tbl)
242 const unsigned long size = tbl->it_indirect_levels ?
243 tbl->it_level_size : tbl->it_size;
245 if (!tbl->it_size)
248 pnv_pci_ioda2_table_do_free_pages((__be64 *)tbl->it_base, size,
249 tbl->it_indirect_levels);
250 if (tbl->it_userspace) {
251 pnv_pci_ioda2_table_do_free_pages(tbl->it_userspace, size,
252 tbl->it_indirect_levels);
292 bool alloc_userspace_copy, struct iommu_table *tbl)
347 pnv_pci_setup_iommu_table(tbl, addr, tce_table_size, bus_offset,
349 tbl->it_level_size = 1ULL << (level_shift - 3);
350 tbl->it_indirect_levels = levels - 1;
351 tbl->it_userspace = uas;
352 tbl->it_nid = nid;
355 window_size, tce_table_size, bus_offset, tbl->it_base,
356 tbl->it_userspace, 1, levels);
370 void pnv_pci_unlink_table_and_group(struct iommu_table *tbl,
377 if (!tbl || !table_group)
384 list_for_each_entry_rcu(tgl, &tbl->it_group_list, next) {
400 if (table_group->tables[i] == tbl) {
401 iommu_tce_table_put(tbl);
411 struct iommu_table *tbl,
416 if (WARN_ON(!tbl || !table_group))
425 list_add_rcu(&tgl->next, &tbl->it_group_list);
427 table_group->tables[num] = iommu_tce_table_get(tbl);