Page MenuHomeFreeBSD

D39184.diff
No OneTemporary

D39184.diff

diff --git a/sys/arm64/iommu/iommu_pmap.h b/sys/arm64/iommu/iommu_pmap.h
--- a/sys/arm64/iommu/iommu_pmap.h
+++ b/sys/arm64/iommu/iommu_pmap.h
@@ -33,19 +33,28 @@
#ifndef _ARM64_IOMMU_IOMMU_PMAP_H_
#define _ARM64_IOMMU_IOMMU_PMAP_H_
+struct smmu_pmap {
+ struct mtx sp_mtx;
+ vm_paddr_t sp_l0_paddr;
+ pd_entry_t *sp_l0;
+#ifdef INVARIANTS
+ long sp_resident_count;
+#endif
+};
+
/* System MMU (SMMU). */
-int smmu_pmap_enter(pmap_t pmap, vm_offset_t va, vm_paddr_t pa, vm_prot_t prot,
- u_int flags);
-int smmu_pmap_remove(pmap_t pmap, vm_offset_t va);
+int smmu_pmap_enter(struct smmu_pmap *pmap, vm_offset_t va, vm_paddr_t pa,
+ vm_prot_t prot, u_int flags);
+int smmu_pmap_remove(struct smmu_pmap *pmap, vm_offset_t va);
/* Mali GPU */
-int pmap_gpu_enter(pmap_t pmap, vm_offset_t va, vm_paddr_t pa,
+int pmap_gpu_enter(struct smmu_pmap *pmap, vm_offset_t va, vm_paddr_t pa,
vm_prot_t prot, u_int flags);
-int pmap_gpu_remove(pmap_t pmap, vm_offset_t va);
+int pmap_gpu_remove(struct smmu_pmap *pmap, vm_offset_t va);
/* Common */
-void smmu_pmap_remove_pages(pmap_t pmap);
-void smmu_pmap_release(pmap_t pmap);
-int smmu_pmap_pinit(pmap_t);
+void smmu_pmap_remove_pages(struct smmu_pmap *pmap);
+void smmu_pmap_release(struct smmu_pmap *pmap);
+int smmu_pmap_pinit(struct smmu_pmap *pmap);
#endif /* !_ARM64_IOMMU_IOMMU_PMAP_H_ */
diff --git a/sys/arm64/iommu/iommu_pmap.c b/sys/arm64/iommu/iommu_pmap.c
--- a/sys/arm64/iommu/iommu_pmap.c
+++ b/sys/arm64/iommu/iommu_pmap.c
@@ -62,6 +62,11 @@
#define IOMMU_PAGE_SIZE 4096
+#define SMMU_PMAP_LOCK(pmap) mtx_lock(&(pmap)->sp_mtx)
+#define SMMU_PMAP_UNLOCK(pmap) mtx_unlock(&(pmap)->sp_mtx)
+#define SMMU_PMAP_LOCK_ASSERT(pmap, type) \
+ mtx_assert(&(pmap)->sp_mtx, (type))
+
#define NL0PG (IOMMU_PAGE_SIZE/(sizeof (pd_entry_t)))
#define NL1PG (IOMMU_PAGE_SIZE/(sizeof (pd_entry_t)))
#define NL2PG (IOMMU_PAGE_SIZE/(sizeof (pd_entry_t)))
@@ -80,9 +85,9 @@
#define smmu_l2_index(va) (((va) >> IOMMU_L2_SHIFT) & IOMMU_Ln_ADDR_MASK)
#define smmu_l3_index(va) (((va) >> IOMMU_L3_SHIFT) & IOMMU_Ln_ADDR_MASK)
-static vm_page_t _pmap_alloc_l3(pmap_t pmap, vm_pindex_t ptepindex);
-static void _smmu_pmap_unwire_l3(pmap_t pmap, vm_offset_t va, vm_page_t m,
- struct spglist *free);
+static vm_page_t _pmap_alloc_l3(struct smmu_pmap *pmap, vm_pindex_t ptepindex);
+static void _smmu_pmap_unwire_l3(struct smmu_pmap *pmap, vm_offset_t va,
+ vm_page_t m, struct spglist *free);
/*
* These load the old table data and store the new value.
@@ -98,10 +103,10 @@
/********************/
static __inline pd_entry_t *
-smmu_pmap_l0(pmap_t pmap, vm_offset_t va)
+smmu_pmap_l0(struct smmu_pmap *pmap, vm_offset_t va)
{
- return (&pmap->pm_l0[smmu_l0_index(va)]);
+ return (&pmap->sp_l0[smmu_l0_index(va)]);
}
static __inline pd_entry_t *
@@ -114,7 +119,7 @@
}
static __inline pd_entry_t *
-smmu_pmap_l1(pmap_t pmap, vm_offset_t va)
+smmu_pmap_l1(struct smmu_pmap *pmap, vm_offset_t va)
{
pd_entry_t *l0;
@@ -145,7 +150,7 @@
}
static __inline pd_entry_t *
-smmu_pmap_l2(pmap_t pmap, vm_offset_t va)
+smmu_pmap_l2(struct smmu_pmap *pmap, vm_offset_t va)
{
pd_entry_t *l1;
@@ -181,7 +186,7 @@
* The next level may or may not point to a valid page or block.
*/
static __inline pd_entry_t *
-smmu_pmap_pde(pmap_t pmap, vm_offset_t va, int *level)
+smmu_pmap_pde(struct smmu_pmap *pmap, vm_offset_t va, int *level)
{
pd_entry_t *l0, *l1, *l2, desc;
@@ -216,7 +221,7 @@
* the first invalid level.
*/
static __inline pt_entry_t *
-smmu_pmap_pte(pmap_t pmap, vm_offset_t va, int *level)
+smmu_pmap_pte(struct smmu_pmap *pmap, vm_offset_t va, int *level)
{
pd_entry_t *l1, *l2, desc;
pt_entry_t *l3;
@@ -266,25 +271,37 @@
CTASSERT(IOMMU_L1_BLOCK == IOMMU_L2_BLOCK);
+#ifdef INVARIANTS
static __inline void
-smmu_pmap_resident_count_inc(pmap_t pmap, int count)
+smmu_pmap_resident_count_inc(struct smmu_pmap *pmap, int count)
{
- PMAP_LOCK_ASSERT(pmap, MA_OWNED);
- pmap->pm_stats.resident_count += count;
+ SMMU_PMAP_LOCK_ASSERT(pmap, MA_OWNED);
+ pmap->sp_resident_count += count;
}
static __inline void
-smmu_pmap_resident_count_dec(pmap_t pmap, int count)
+smmu_pmap_resident_count_dec(struct smmu_pmap *pmap, int count)
{
- PMAP_LOCK_ASSERT(pmap, MA_OWNED);
- KASSERT(pmap->pm_stats.resident_count >= count,
+ SMMU_PMAP_LOCK_ASSERT(pmap, MA_OWNED);
+ KASSERT(pmap->sp_resident_count >= count,
("pmap %p resident count underflow %ld %d", pmap,
- pmap->pm_stats.resident_count, count));
- pmap->pm_stats.resident_count -= count;
+ pmap->sp_resident_count, count));
+ pmap->sp_resident_count -= count;
+}
+#else
+static __inline void
+smmu_pmap_resident_count_inc(struct smmu_pmap *pmap, int count)
+{
}
+static __inline void
+smmu_pmap_resident_count_dec(struct smmu_pmap *pmap, int count)
+{
+}
+#endif
+
/***************************************************
* Page table page management routines.....
***************************************************/
@@ -316,7 +333,7 @@
* page table page was unmapped and FALSE otherwise.
*/
static inline boolean_t
-smmu_pmap_unwire_l3(pmap_t pmap, vm_offset_t va, vm_page_t m,
+smmu_pmap_unwire_l3(struct smmu_pmap *pmap, vm_offset_t va, vm_page_t m,
struct spglist *free)
{
@@ -329,11 +346,11 @@
}
static void
-_smmu_pmap_unwire_l3(pmap_t pmap, vm_offset_t va, vm_page_t m,
+_smmu_pmap_unwire_l3(struct smmu_pmap *pmap, vm_offset_t va, vm_page_t m,
struct spglist *free)
{
- PMAP_LOCK_ASSERT(pmap, MA_OWNED);
+ SMMU_PMAP_LOCK_ASSERT(pmap, MA_OWNED);
/*
* unmap the page table page
*/
@@ -385,7 +402,7 @@
}
int
-smmu_pmap_pinit(pmap_t pmap)
+smmu_pmap_pinit(struct smmu_pmap *pmap)
{
vm_page_t m;
@@ -394,14 +411,11 @@
*/
m = vm_page_alloc_noobj(VM_ALLOC_WAITOK | VM_ALLOC_WIRED |
VM_ALLOC_ZERO);
- pmap->pm_l0_paddr = VM_PAGE_TO_PHYS(m);
- pmap->pm_l0 = (pd_entry_t *)PHYS_TO_DMAP(pmap->pm_l0_paddr);
-
- vm_radix_init(&pmap->pm_root);
- bzero(&pmap->pm_stats, sizeof(pmap->pm_stats));
+ pmap->sp_l0_paddr = VM_PAGE_TO_PHYS(m);
+ pmap->sp_l0 = (pd_entry_t *)PHYS_TO_DMAP(pmap->sp_l0_paddr);
- pmap->pm_levels = 4;
- pmap->pm_ttbr = VM_PAGE_TO_PHYS(m);
+ pmap->sp_resident_count = 0;
+ mtx_init(&pmap->sp_mtx, "smmu pmap", NULL, MTX_DEF);
return (1);
}
@@ -418,11 +432,11 @@
* race conditions.
*/
static vm_page_t
-_pmap_alloc_l3(pmap_t pmap, vm_pindex_t ptepindex)
+_pmap_alloc_l3(struct smmu_pmap *pmap, vm_pindex_t ptepindex)
{
vm_page_t m, l1pg, l2pg;
- PMAP_LOCK_ASSERT(pmap, MA_OWNED);
+ SMMU_PMAP_LOCK_ASSERT(pmap, MA_OWNED);
/*
* Allocate a page table page.
@@ -456,7 +470,7 @@
vm_pindex_t l0index;
l0index = ptepindex - (NUL2E + NUL1E);
- l0 = &pmap->pm_l0[l0index];
+ l0 = &pmap->sp_l0[l0index];
smmu_pmap_store(l0, VM_PAGE_TO_PHYS(m) | IOMMU_L0_TABLE);
} else if (ptepindex >= NUL2E) {
vm_pindex_t l0index, l1index;
@@ -466,7 +480,7 @@
l1index = ptepindex - NUL2E;
l0index = l1index >> IOMMU_L0_ENTRIES_SHIFT;
- l0 = &pmap->pm_l0[l0index];
+ l0 = &pmap->sp_l0[l0index];
tl0 = smmu_pmap_load(l0);
if (tl0 == 0) {
/* recurse for allocating page dir */
@@ -492,7 +506,7 @@
l1index = ptepindex >> Ln_ENTRIES_SHIFT;
l0index = l1index >> IOMMU_L0_ENTRIES_SHIFT;
- l0 = &pmap->pm_l0[l0index];
+ l0 = &pmap->sp_l0[l0index];
tl0 = smmu_pmap_load(l0);
if (tl0 == 0) {
/* recurse for allocating page dir */
@@ -542,19 +556,18 @@
* Should only be called if the map contains no valid mappings.
*/
void
-smmu_pmap_release(pmap_t pmap)
+smmu_pmap_release(struct smmu_pmap *pmap)
{
vm_page_t m;
- KASSERT(pmap->pm_stats.resident_count == 0,
+ KASSERT(pmap->sp_resident_count == 0,
("pmap_release: pmap resident count %ld != 0",
- pmap->pm_stats.resident_count));
- KASSERT(vm_radix_is_empty(&pmap->pm_root),
- ("pmap_release: pmap has reserved page table page(s)"));
+ pmap->sp_resident_count));
- m = PHYS_TO_VM_PAGE(pmap->pm_l0_paddr);
+ m = PHYS_TO_VM_PAGE(pmap->sp_l0_paddr);
vm_page_unwire_noq(m);
vm_page_free_zero(m);
+ mtx_destroy(&pmap->sp_mtx);
}
/***************************************************
@@ -565,7 +578,7 @@
* Add a single Mali GPU entry. This function does not sleep.
*/
int
-pmap_gpu_enter(pmap_t pmap, vm_offset_t va, vm_paddr_t pa,
+pmap_gpu_enter(struct smmu_pmap *pmap, vm_offset_t va, vm_paddr_t pa,
vm_prot_t prot, u_int flags)
{
pd_entry_t *pde;
@@ -578,7 +591,6 @@
int lvl;
int rv;
- KASSERT(pmap != kernel_pmap, ("kernel pmap used for GPU"));
KASSERT(va < VM_MAXUSER_ADDRESS, ("wrong address space"));
KASSERT((va & PAGE_MASK) == 0, ("va is misaligned"));
KASSERT((pa & PAGE_MASK) == 0, ("pa is misaligned"));
@@ -594,7 +606,7 @@
CTR2(KTR_PMAP, "pmap_gpu_enter: %.16lx -> %.16lx", va, pa);
- PMAP_LOCK(pmap);
+ SMMU_PMAP_LOCK(pmap);
/*
* In the case that a page table page is not
@@ -639,7 +651,7 @@
rv = KERN_SUCCESS;
out:
- PMAP_UNLOCK(pmap);
+ SMMU_PMAP_UNLOCK(pmap);
return (rv);
}
@@ -648,7 +660,7 @@
* Remove a single Mali GPU entry.
*/
int
-pmap_gpu_remove(pmap_t pmap, vm_offset_t va)
+pmap_gpu_remove(struct smmu_pmap *pmap, vm_offset_t va)
{
pd_entry_t *pde;
pt_entry_t *pte;
@@ -656,9 +668,8 @@
int rc;
KASSERT((va & PAGE_MASK) == 0, ("va is misaligned"));
- KASSERT(pmap != kernel_pmap, ("kernel pmap used for GPU"));
- PMAP_LOCK(pmap);
+ SMMU_PMAP_LOCK(pmap);
pde = smmu_pmap_pde(pmap, va, &lvl);
if (pde == NULL || lvl != 2) {
@@ -674,7 +685,7 @@
rc = KERN_SUCCESS;
out:
- PMAP_UNLOCK(pmap);
+ SMMU_PMAP_UNLOCK(pmap);
return (rc);
}
@@ -683,7 +694,7 @@
* Add a single SMMU entry. This function does not sleep.
*/
int
-smmu_pmap_enter(pmap_t pmap, vm_offset_t va, vm_paddr_t pa,
+smmu_pmap_enter(struct smmu_pmap *pmap, vm_offset_t va, vm_paddr_t pa,
vm_prot_t prot, u_int flags)
{
pd_entry_t *pde;
@@ -707,7 +718,7 @@
CTR2(KTR_PMAP, "pmap_senter: %.16lx -> %.16lx", va, pa);
- PMAP_LOCK(pmap);
+ SMMU_PMAP_LOCK(pmap);
/*
* In the case that a page table page is not
@@ -737,7 +748,7 @@
rv = KERN_SUCCESS;
out:
- PMAP_UNLOCK(pmap);
+ SMMU_PMAP_UNLOCK(pmap);
return (rv);
}
@@ -746,13 +757,13 @@
* Remove a single SMMU entry.
*/
int
-smmu_pmap_remove(pmap_t pmap, vm_offset_t va)
+smmu_pmap_remove(struct smmu_pmap *pmap, vm_offset_t va)
{
pt_entry_t *pte;
int lvl;
int rc;
- PMAP_LOCK(pmap);
+ SMMU_PMAP_LOCK(pmap);
pte = smmu_pmap_pte(pmap, va, &lvl);
KASSERT(lvl == 3,
@@ -765,7 +776,7 @@
} else
rc = KERN_FAILURE;
- PMAP_UNLOCK(pmap);
+ SMMU_PMAP_UNLOCK(pmap);
return (rc);
}
@@ -776,7 +787,7 @@
* this function panics.
*/
void
-smmu_pmap_remove_pages(pmap_t pmap)
+smmu_pmap_remove_pages(struct smmu_pmap *pmap)
{
pd_entry_t l0e, *l1, l1e, *l2, l2e;
pt_entry_t *l3, l3e;
@@ -787,11 +798,11 @@
vm_paddr_t pa1;
int i, j, k, l;
- PMAP_LOCK(pmap);
+ SMMU_PMAP_LOCK(pmap);
for (sva = VM_MINUSER_ADDRESS, i = smmu_l0_index(sva);
(i < Ln_ENTRIES && sva < VM_MAXUSER_ADDRESS); i++) {
- l0e = pmap->pm_l0[i];
+ l0e = pmap->sp_l0[i];
if ((l0e & ATTR_DESCR_VALID) == 0) {
sva += IOMMU_L0_SIZE;
continue;
@@ -848,11 +859,11 @@
smmu_pmap_resident_count_dec(pmap, 1);
vm_page_free(m0);
- smmu_pmap_clear(&pmap->pm_l0[i]);
+ smmu_pmap_clear(&pmap->sp_l0[i]);
}
- KASSERT(pmap->pm_stats.resident_count == 0,
- ("Invalid resident count %jd", pmap->pm_stats.resident_count));
+ KASSERT(pmap->sp_resident_count == 0,
+ ("Invalid resident count %jd", pmap->sp_resident_count));
- PMAP_UNLOCK(pmap);
+ SMMU_PMAP_UNLOCK(pmap);
}
diff --git a/sys/arm64/iommu/smmu.c b/sys/arm64/iommu/smmu.c
--- a/sys/arm64/iommu/smmu.c
+++ b/sys/arm64/iommu/smmu.c
@@ -840,7 +840,7 @@
uint64_t val;
vm_size_t size;
struct smmu_cd *cd;
- pmap_t p;
+ struct smmu_pmap *p;
size = 1 * (CD_DWORDS << 3);
@@ -875,8 +875,8 @@
val |= ((64 - sc->ias) << CD0_T0SZ_S);
val |= CD0_IPS_48BITS;
- paddr = p->pm_l0_paddr & CD1_TTB0_M;
- KASSERT(paddr == p->pm_l0_paddr, ("bad allocation 1"));
+ paddr = p->sp_l0_paddr & CD1_TTB0_M;
+ KASSERT(paddr == p->sp_l0_paddr, ("bad allocation 1"));
ptr[1] = paddr;
ptr[2] = 0;
@@ -1729,7 +1729,6 @@
domain->asid = (uint16_t)new_asid;
smmu_pmap_pinit(&domain->p);
- PMAP_LOCK_INIT(&domain->p);
error = smmu_init_cd(sc, domain);
if (error) {
diff --git a/sys/arm64/iommu/smmuvar.h b/sys/arm64/iommu/smmuvar.h
--- a/sys/arm64/iommu/smmuvar.h
+++ b/sys/arm64/iommu/smmuvar.h
@@ -35,6 +35,8 @@
#ifndef _ARM64_IOMMU_SMMUVAR_H_
#define _ARM64_IOMMU_SMMUVAR_H_
+#include <arm64/iommu/iommu_pmap.h>
+
#define SMMU_DEVSTR "ARM System Memory Management Unit"
#define SMMU_LOCK(_sc) mtx_lock(&(_sc)->sc_mtx)
#define SMMU_UNLOCK(_sc) mtx_unlock(&(_sc)->sc_mtx)
@@ -55,7 +57,7 @@
LIST_ENTRY(smmu_domain) next;
u_int entries_cnt;
struct smmu_cd *cd;
- struct pmap p;
+ struct smmu_pmap p;
uint16_t asid;
};

File Metadata

Mime Type
text/plain
Expires
Fri, Jan 10, 7:55 PM (15 h, 17 m)
Storage Engine
blob
Storage Format
Raw Data
Storage Handle
15745946
Default Alt Text
D39184.diff (12 KB)

Event Timeline