405 lines
		
	
	
		
			9.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			405 lines
		
	
	
		
			9.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Copyright 2017 Red Hat Inc.
 | 
						|
 *
 | 
						|
 * Permission is hereby granted, free of charge, to any person obtaining a
 | 
						|
 * copy of this software and associated documentation files (the "Software"),
 | 
						|
 * to deal in the Software without restriction, including without limitation
 | 
						|
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 | 
						|
 * and/or sell copies of the Software, and to permit persons to whom the
 | 
						|
 * Software is furnished to do so, subject to the following conditions:
 | 
						|
 *
 | 
						|
 * The above copyright notice and this permission notice shall be included in
 | 
						|
 * all copies or substantial portions of the Software.
 | 
						|
 *
 | 
						|
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | 
						|
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | 
						|
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 | 
						|
 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
 | 
						|
 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
 | 
						|
 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
 | 
						|
 * OTHER DEALINGS IN THE SOFTWARE.
 | 
						|
 */
 | 
						|
#include "uvmm.h"
 | 
						|
#include "umem.h"
 | 
						|
#include "ummu.h"
 | 
						|
 | 
						|
#include <core/client.h>
 | 
						|
#include <core/memory.h>
 | 
						|
 | 
						|
#include <nvif/if000c.h>
 | 
						|
#include <nvif/unpack.h>
 | 
						|
 | 
						|
static const struct nvkm_object_func nvkm_uvmm;
 | 
						|
struct nvkm_vmm *
 | 
						|
nvkm_uvmm_search(struct nvkm_client *client, u64 handle)
 | 
						|
{
 | 
						|
	struct nvkm_object *object;
 | 
						|
 | 
						|
	object = nvkm_object_search(client, handle, &nvkm_uvmm);
 | 
						|
	if (IS_ERR(object))
 | 
						|
		return (void *)object;
 | 
						|
 | 
						|
	return nvkm_uvmm(object)->vmm;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
nvkm_uvmm_mthd_pfnclr(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
 | 
						|
{
 | 
						|
	union {
 | 
						|
		struct nvif_vmm_pfnclr_v0 v0;
 | 
						|
	} *args = argv;
 | 
						|
	struct nvkm_vmm *vmm = uvmm->vmm;
 | 
						|
	int ret = -ENOSYS;
 | 
						|
	u64 addr, size;
 | 
						|
 | 
						|
	if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
 | 
						|
		addr = args->v0.addr;
 | 
						|
		size = args->v0.size;
 | 
						|
	} else
 | 
						|
		return ret;
 | 
						|
 | 
						|
	if (size) {
 | 
						|
		mutex_lock(&vmm->mutex);
 | 
						|
		ret = nvkm_vmm_pfn_unmap(vmm, addr, size);
 | 
						|
		mutex_unlock(&vmm->mutex);
 | 
						|
	}
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
nvkm_uvmm_mthd_pfnmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
 | 
						|
{
 | 
						|
	union {
 | 
						|
		struct nvif_vmm_pfnmap_v0 v0;
 | 
						|
	} *args = argv;
 | 
						|
	struct nvkm_vmm *vmm = uvmm->vmm;
 | 
						|
	int ret = -ENOSYS;
 | 
						|
	u64 addr, size, *phys;
 | 
						|
	u8  page;
 | 
						|
 | 
						|
	if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) {
 | 
						|
		page = args->v0.page;
 | 
						|
		addr = args->v0.addr;
 | 
						|
		size = args->v0.size;
 | 
						|
		phys = args->v0.phys;
 | 
						|
		if (argc != (size >> page) * sizeof(args->v0.phys[0]))
 | 
						|
			return -EINVAL;
 | 
						|
	} else
 | 
						|
		return ret;
 | 
						|
 | 
						|
	if (size) {
 | 
						|
		mutex_lock(&vmm->mutex);
 | 
						|
		ret = nvkm_vmm_pfn_map(vmm, page, addr, size, phys);
 | 
						|
		mutex_unlock(&vmm->mutex);
 | 
						|
	}
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
nvkm_uvmm_mthd_unmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
 | 
						|
{
 | 
						|
	union {
 | 
						|
		struct nvif_vmm_unmap_v0 v0;
 | 
						|
	} *args = argv;
 | 
						|
	struct nvkm_vmm *vmm = uvmm->vmm;
 | 
						|
	struct nvkm_vma *vma;
 | 
						|
	int ret = -ENOSYS;
 | 
						|
	u64 addr;
 | 
						|
 | 
						|
	if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
 | 
						|
		addr = args->v0.addr;
 | 
						|
	} else
 | 
						|
		return ret;
 | 
						|
 | 
						|
	mutex_lock(&vmm->mutex);
 | 
						|
	vma = nvkm_vmm_node_search(vmm, addr);
 | 
						|
	if (ret = -ENOENT, !vma || vma->addr != addr) {
 | 
						|
		VMM_DEBUG(vmm, "lookup %016llx: %016llx",
 | 
						|
			  addr, vma ? vma->addr : ~0ULL);
 | 
						|
		goto done;
 | 
						|
	}
 | 
						|
 | 
						|
	if (ret = -ENOENT, vma->busy) {
 | 
						|
		VMM_DEBUG(vmm, "denied %016llx: %d", addr, vma->busy);
 | 
						|
		goto done;
 | 
						|
	}
 | 
						|
 | 
						|
	if (ret = -EINVAL, !vma->memory) {
 | 
						|
		VMM_DEBUG(vmm, "unmapped");
 | 
						|
		goto done;
 | 
						|
	}
 | 
						|
 | 
						|
	nvkm_vmm_unmap_locked(vmm, vma, false);
 | 
						|
	ret = 0;
 | 
						|
done:
 | 
						|
	mutex_unlock(&vmm->mutex);
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
nvkm_uvmm_mthd_map(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
 | 
						|
{
 | 
						|
	struct nvkm_client *client = uvmm->object.client;
 | 
						|
	union {
 | 
						|
		struct nvif_vmm_map_v0 v0;
 | 
						|
	} *args = argv;
 | 
						|
	u64 addr, size, handle, offset;
 | 
						|
	struct nvkm_vmm *vmm = uvmm->vmm;
 | 
						|
	struct nvkm_vma *vma;
 | 
						|
	struct nvkm_memory *memory;
 | 
						|
	int ret = -ENOSYS;
 | 
						|
 | 
						|
	if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) {
 | 
						|
		addr = args->v0.addr;
 | 
						|
		size = args->v0.size;
 | 
						|
		handle = args->v0.memory;
 | 
						|
		offset = args->v0.offset;
 | 
						|
	} else
 | 
						|
		return ret;
 | 
						|
 | 
						|
	memory = nvkm_umem_search(client, handle);
 | 
						|
	if (IS_ERR(memory)) {
 | 
						|
		VMM_DEBUG(vmm, "memory %016llx %ld\n", handle, PTR_ERR(memory));
 | 
						|
		return PTR_ERR(memory);
 | 
						|
	}
 | 
						|
 | 
						|
	mutex_lock(&vmm->mutex);
 | 
						|
	if (ret = -ENOENT, !(vma = nvkm_vmm_node_search(vmm, addr))) {
 | 
						|
		VMM_DEBUG(vmm, "lookup %016llx", addr);
 | 
						|
		goto fail;
 | 
						|
	}
 | 
						|
 | 
						|
	if (ret = -ENOENT, vma->busy) {
 | 
						|
		VMM_DEBUG(vmm, "denied %016llx: %d", addr, vma->busy);
 | 
						|
		goto fail;
 | 
						|
	}
 | 
						|
 | 
						|
	if (ret = -EINVAL, vma->mapped && !vma->memory) {
 | 
						|
		VMM_DEBUG(vmm, "pfnmap %016llx", addr);
 | 
						|
		goto fail;
 | 
						|
	}
 | 
						|
 | 
						|
	if (ret = -EINVAL, vma->addr != addr || vma->size != size) {
 | 
						|
		if (addr + size > vma->addr + vma->size || vma->memory ||
 | 
						|
		    (vma->refd == NVKM_VMA_PAGE_NONE && !vma->mapref)) {
 | 
						|
			VMM_DEBUG(vmm, "split %d %d %d "
 | 
						|
				       "%016llx %016llx %016llx %016llx",
 | 
						|
				  !!vma->memory, vma->refd, vma->mapref,
 | 
						|
				  addr, size, vma->addr, (u64)vma->size);
 | 
						|
			goto fail;
 | 
						|
		}
 | 
						|
 | 
						|
		vma = nvkm_vmm_node_split(vmm, vma, addr, size);
 | 
						|
		if (!vma) {
 | 
						|
			ret = -ENOMEM;
 | 
						|
			goto fail;
 | 
						|
		}
 | 
						|
	}
 | 
						|
	vma->busy = true;
 | 
						|
	mutex_unlock(&vmm->mutex);
 | 
						|
 | 
						|
	ret = nvkm_memory_map(memory, offset, vmm, vma, argv, argc);
 | 
						|
	if (ret == 0) {
 | 
						|
		/* Successful map will clear vma->busy. */
 | 
						|
		nvkm_memory_unref(&memory);
 | 
						|
		return 0;
 | 
						|
	}
 | 
						|
 | 
						|
	mutex_lock(&vmm->mutex);
 | 
						|
	vma->busy = false;
 | 
						|
	nvkm_vmm_unmap_region(vmm, vma);
 | 
						|
fail:
 | 
						|
	mutex_unlock(&vmm->mutex);
 | 
						|
	nvkm_memory_unref(&memory);
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
nvkm_uvmm_mthd_put(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
 | 
						|
{
 | 
						|
	union {
 | 
						|
		struct nvif_vmm_put_v0 v0;
 | 
						|
	} *args = argv;
 | 
						|
	struct nvkm_vmm *vmm = uvmm->vmm;
 | 
						|
	struct nvkm_vma *vma;
 | 
						|
	int ret = -ENOSYS;
 | 
						|
	u64 addr;
 | 
						|
 | 
						|
	if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
 | 
						|
		addr = args->v0.addr;
 | 
						|
	} else
 | 
						|
		return ret;
 | 
						|
 | 
						|
	mutex_lock(&vmm->mutex);
 | 
						|
	vma = nvkm_vmm_node_search(vmm, args->v0.addr);
 | 
						|
	if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) {
 | 
						|
		VMM_DEBUG(vmm, "lookup %016llx: %016llx %d", addr,
 | 
						|
			  vma ? vma->addr : ~0ULL, vma ? vma->part : 0);
 | 
						|
		goto done;
 | 
						|
	}
 | 
						|
 | 
						|
	if (ret = -ENOENT, vma->busy) {
 | 
						|
		VMM_DEBUG(vmm, "denied %016llx: %d", addr, vma->busy);
 | 
						|
		goto done;
 | 
						|
	}
 | 
						|
 | 
						|
	nvkm_vmm_put_locked(vmm, vma);
 | 
						|
	ret = 0;
 | 
						|
done:
 | 
						|
	mutex_unlock(&vmm->mutex);
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
nvkm_uvmm_mthd_get(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
 | 
						|
{
 | 
						|
	union {
 | 
						|
		struct nvif_vmm_get_v0 v0;
 | 
						|
	} *args = argv;
 | 
						|
	struct nvkm_vmm *vmm = uvmm->vmm;
 | 
						|
	struct nvkm_vma *vma;
 | 
						|
	int ret = -ENOSYS;
 | 
						|
	bool getref, mapref, sparse;
 | 
						|
	u8 page, align;
 | 
						|
	u64 size;
 | 
						|
 | 
						|
	if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
 | 
						|
		getref = args->v0.type == NVIF_VMM_GET_V0_PTES;
 | 
						|
		mapref = args->v0.type == NVIF_VMM_GET_V0_ADDR;
 | 
						|
		sparse = args->v0.sparse;
 | 
						|
		page = args->v0.page;
 | 
						|
		align = args->v0.align;
 | 
						|
		size = args->v0.size;
 | 
						|
	} else
 | 
						|
		return ret;
 | 
						|
 | 
						|
	mutex_lock(&vmm->mutex);
 | 
						|
	ret = nvkm_vmm_get_locked(vmm, getref, mapref, sparse,
 | 
						|
				  page, align, size, &vma);
 | 
						|
	mutex_unlock(&vmm->mutex);
 | 
						|
	if (ret)
 | 
						|
		return ret;
 | 
						|
 | 
						|
	args->v0.addr = vma->addr;
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
nvkm_uvmm_mthd_page(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
 | 
						|
{
 | 
						|
	union {
 | 
						|
		struct nvif_vmm_page_v0 v0;
 | 
						|
	} *args = argv;
 | 
						|
	const struct nvkm_vmm_page *page;
 | 
						|
	int ret = -ENOSYS;
 | 
						|
	u8 type, index, nr;
 | 
						|
 | 
						|
	page = uvmm->vmm->func->page;
 | 
						|
	for (nr = 0; page[nr].shift; nr++);
 | 
						|
 | 
						|
	if (!(nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
 | 
						|
		if ((index = args->v0.index) >= nr)
 | 
						|
			return -EINVAL;
 | 
						|
		type = page[index].type;
 | 
						|
		args->v0.shift = page[index].shift;
 | 
						|
		args->v0.sparse = !!(type & NVKM_VMM_PAGE_SPARSE);
 | 
						|
		args->v0.vram = !!(type & NVKM_VMM_PAGE_VRAM);
 | 
						|
		args->v0.host = !!(type & NVKM_VMM_PAGE_HOST);
 | 
						|
		args->v0.comp = !!(type & NVKM_VMM_PAGE_COMP);
 | 
						|
	} else
 | 
						|
		return -ENOSYS;
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
nvkm_uvmm_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc)
 | 
						|
{
 | 
						|
	struct nvkm_uvmm *uvmm = nvkm_uvmm(object);
 | 
						|
	switch (mthd) {
 | 
						|
	case NVIF_VMM_V0_PAGE  : return nvkm_uvmm_mthd_page  (uvmm, argv, argc);
 | 
						|
	case NVIF_VMM_V0_GET   : return nvkm_uvmm_mthd_get   (uvmm, argv, argc);
 | 
						|
	case NVIF_VMM_V0_PUT   : return nvkm_uvmm_mthd_put   (uvmm, argv, argc);
 | 
						|
	case NVIF_VMM_V0_MAP   : return nvkm_uvmm_mthd_map   (uvmm, argv, argc);
 | 
						|
	case NVIF_VMM_V0_UNMAP : return nvkm_uvmm_mthd_unmap (uvmm, argv, argc);
 | 
						|
	case NVIF_VMM_V0_PFNMAP: return nvkm_uvmm_mthd_pfnmap(uvmm, argv, argc);
 | 
						|
	case NVIF_VMM_V0_PFNCLR: return nvkm_uvmm_mthd_pfnclr(uvmm, argv, argc);
 | 
						|
	case NVIF_VMM_V0_MTHD(0x00) ... NVIF_VMM_V0_MTHD(0x7f):
 | 
						|
		if (uvmm->vmm->func->mthd) {
 | 
						|
			return uvmm->vmm->func->mthd(uvmm->vmm,
 | 
						|
						     uvmm->object.client,
 | 
						|
						     mthd, argv, argc);
 | 
						|
		}
 | 
						|
		break;
 | 
						|
	default:
 | 
						|
		break;
 | 
						|
	}
 | 
						|
	return -EINVAL;
 | 
						|
}
 | 
						|
 | 
						|
static void *
 | 
						|
nvkm_uvmm_dtor(struct nvkm_object *object)
 | 
						|
{
 | 
						|
	struct nvkm_uvmm *uvmm = nvkm_uvmm(object);
 | 
						|
	nvkm_vmm_unref(&uvmm->vmm);
 | 
						|
	return uvmm;
 | 
						|
}
 | 
						|
 | 
						|
static const struct nvkm_object_func
 | 
						|
nvkm_uvmm = {
 | 
						|
	.dtor = nvkm_uvmm_dtor,
 | 
						|
	.mthd = nvkm_uvmm_mthd,
 | 
						|
};
 | 
						|
 | 
						|
int
 | 
						|
nvkm_uvmm_new(const struct nvkm_oclass *oclass, void *argv, u32 argc,
 | 
						|
	      struct nvkm_object **pobject)
 | 
						|
{
 | 
						|
	struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu;
 | 
						|
	const bool more = oclass->base.maxver >= 0;
 | 
						|
	union {
 | 
						|
		struct nvif_vmm_v0 v0;
 | 
						|
	} *args = argv;
 | 
						|
	const struct nvkm_vmm_page *page;
 | 
						|
	struct nvkm_uvmm *uvmm;
 | 
						|
	int ret = -ENOSYS;
 | 
						|
	u64 addr, size;
 | 
						|
	bool managed;
 | 
						|
 | 
						|
	if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, more))) {
 | 
						|
		managed = args->v0.managed != 0;
 | 
						|
		addr = args->v0.addr;
 | 
						|
		size = args->v0.size;
 | 
						|
	} else
 | 
						|
		return ret;
 | 
						|
 | 
						|
	if (!(uvmm = kzalloc(sizeof(*uvmm), GFP_KERNEL)))
 | 
						|
		return -ENOMEM;
 | 
						|
	nvkm_object_ctor(&nvkm_uvmm, oclass, &uvmm->object);
 | 
						|
	*pobject = &uvmm->object;
 | 
						|
 | 
						|
	if (!mmu->vmm) {
 | 
						|
		ret = mmu->func->vmm.ctor(mmu, managed, addr, size, argv, argc,
 | 
						|
					  NULL, "user", &uvmm->vmm);
 | 
						|
		if (ret)
 | 
						|
			return ret;
 | 
						|
 | 
						|
		uvmm->vmm->debug = max(uvmm->vmm->debug, oclass->client->debug);
 | 
						|
	} else {
 | 
						|
		if (size)
 | 
						|
			return -EINVAL;
 | 
						|
 | 
						|
		uvmm->vmm = nvkm_vmm_ref(mmu->vmm);
 | 
						|
	}
 | 
						|
 | 
						|
	page = uvmm->vmm->func->page;
 | 
						|
	args->v0.page_nr = 0;
 | 
						|
	while (page && (page++)->shift)
 | 
						|
		args->v0.page_nr++;
 | 
						|
	args->v0.addr = uvmm->vmm->start;
 | 
						|
	args->v0.size = uvmm->vmm->limit;
 | 
						|
	return 0;
 | 
						|
}
 |