/* $NetBSD: pmap_pvt.c,v 1.15 2022/05/08 22:03:02 rin Exp $ */
/*-
* Copyright (c) 2014, 2020 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Taylor R. Campbell.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
#include <sys/cdefs.h>
__RCSID("$NetBSD: pmap_pvt.c,v 1.15 2022/05/08 22:03:02 rin Exp $");
#include <sys/param.h>
#include <sys/atomic.h>
#include <sys/kmem.h>
#include <sys/pserialize.h>
#include <uvm/uvm.h>
#include <uvm/pmap/pmap_pvt.h>
#if !defined(PMAP_PV_TRACK_ONLY_STUBS)
/*
* unmanaged pv-tracked ranges
*
* This is a linear list for now because the only user are the DRM
* graphics drivers, with a single tracked range per device, for the
* graphics aperture, so there are expected to be few of them.
*
* This is used only after the VM system is initialized well enough
* that we can use kmem_alloc.
*/
struct pv_track {
paddr_t pvt_start;
psize_t pvt_size;
struct pv_track *pvt_next;
struct pmap_page pvt_pages[];
};
static struct {
kmutex_t lock;
pserialize_t psz;
struct pv_track *list;
} pv_unmanaged __cacheline_aligned;
void
pmap_pv_init(void)
{
mutex_init(&pv_unmanaged.lock, MUTEX_DEFAULT, IPL_NONE);
pv_unmanaged.psz = pserialize_create();
pv_unmanaged.list = NULL;
}
void
pmap_pv_track(paddr_t start, psize_t size)
{
struct pv_track *pvt;
size_t npages;
KASSERT(start == trunc_page(start));
KASSERT(size == trunc_page(size));
/* We may sleep for allocation. */
ASSERT_SLEEPABLE();
npages = size >> PAGE_SHIFT;
pvt = kmem_zalloc(offsetof(struct pv_track, pvt_pages[npages]),
KM_SLEEP);
pvt->pvt_start = start;
pvt->pvt_size = size;
#ifdef PMAP_PAGE_INIT
for (size_t i = 0; i < npages; i++)
PMAP_PAGE_INIT(&pvt->pvt_pages[i]);
#endif
mutex_enter(&pv_unmanaged.lock);
pvt->pvt_next = pv_unmanaged.list;
atomic_store_release(&pv_unmanaged.list, pvt);
mutex_exit(&pv_unmanaged.lock);
}
void
pmap_pv_untrack(paddr_t start, psize_t size)
{
struct pv_track **pvtp, *pvt;
size_t npages;
KASSERT(start == trunc_page(start));
KASSERT(size == trunc_page(size));
/* We may sleep for pserialize_perform. */
ASSERT_SLEEPABLE();
mutex_enter(&pv_unmanaged.lock);
for (pvtp = &pv_unmanaged.list;
(pvt = *pvtp) != NULL;
pvtp = &pvt->pvt_next) {
if (pvt->pvt_start != start)
continue;
if (pvt->pvt_size != size)
panic("pmap_pv_untrack: pv-tracking at 0x%"PRIxPADDR
": 0x%"PRIxPSIZE" bytes, not 0x%"PRIxPSIZE" bytes",
pvt->pvt_start, pvt->pvt_size, size);
/*
* Remove from list. Readers can safely see the old
* and new states of the list.
*/
atomic_store_relaxed(pvtp, pvt->pvt_next);
/* Wait for readers who can see the old state to finish. */
pserialize_perform(pv_unmanaged.psz);
/*
* We now have exclusive access to pvt and can destroy
* it. Poison it to catch bugs.
*/
explicit_memset(&pvt->pvt_next, 0x1a, sizeof pvt->pvt_next);
goto out;
}
panic("pmap_pv_untrack: pages not pv-tracked at 0x%"PRIxPADDR
" (0x%"PRIxPSIZE" bytes)",
start, size);
out: mutex_exit(&pv_unmanaged.lock);
npages = size >> PAGE_SHIFT;
kmem_free(pvt, offsetof(struct pv_track, pvt_pages[npages]));
}
struct pmap_page *
pmap_pv_tracked(paddr_t pa)
{
struct pv_track *pvt;
size_t pgno;
int s;
KASSERT(pa == trunc_page(pa));
s = pserialize_read_enter();
for (pvt = atomic_load_consume(&pv_unmanaged.list);
pvt != NULL;
pvt = pvt->pvt_next) {
if ((pvt->pvt_start <= pa) &&
((pa - pvt->pvt_start) < pvt->pvt_size))
break;
}
pserialize_read_exit(s);
if (pvt == NULL)
return NULL;
KASSERT(pvt->pvt_start <= pa);
KASSERT((pa - pvt->pvt_start) < pvt->pvt_size);
pgno = (pa - pvt->pvt_start) >> PAGE_SHIFT;
return &pvt->pvt_pages[pgno];
}
#else /* PMAP_PV_TRACK_ONLY_STUBS */
/*
* Provide empty stubs just for MODULAR kernels.
*/
void
pmap_pv_init(void)
{
}
struct pmap_page *
pmap_pv_tracked(paddr_t pa)
{
return NULL;
}
#if notdef
/*
* pmap_pv_{,un}track() are intentionally commented out. If modules
* call these functions, the result should be an inconsistent state.
*
* Such modules require real PV-tracking support. Let us make the
* two symbols undefined, and prevent these modules from loaded.
*/
void
pmap_pv_track(paddr_t start, psize_t size)
{
panic("PV-tracking not supported");
}
void
pmap_pv_untrack(paddr_t start, psize_t size)
{
panic("PV-tracking not supported");
}
#endif /* notdef */
#endif /* PMAP_PV_TRACK_ONLY_STUBS */