125 lines
3.4 KiB
C
125 lines
3.4 KiB
C
/*
|
|
* drivers/gpu/drm/omapdrm/omap_gem_helpers.c
|
|
*
|
|
* Copyright (C) 2011 Texas Instruments
|
|
* Author: Rob Clark <rob.clark@linaro.org>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 as published by
|
|
* the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with
|
|
* this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
/* temporary copy of drm_gem_{get,put}_pages() until the
|
|
* "drm/gem: add functions to get/put pages" patch is merged..
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/types.h>
|
|
#include <linux/shmem_fs.h>
|
|
|
|
#include <drm/drmP.h>
|
|
|
|
/**
|
|
* drm_gem_get_pages - helper to allocate backing pages for a GEM object
|
|
* @obj: obj in question
|
|
* @gfpmask: gfp mask of requested pages
|
|
*/
|
|
struct page **_drm_gem_get_pages(struct drm_gem_object *obj, gfp_t gfpmask)
|
|
{
|
|
struct inode *inode;
|
|
struct address_space *mapping;
|
|
struct page *p, **pages;
|
|
int i, npages;
|
|
|
|
/* This is the shared memory object that backs the GEM resource */
|
|
inode = file_inode(obj->filp);
|
|
mapping = inode->i_mapping;
|
|
|
|
npages = obj->size >> PAGE_SHIFT;
|
|
|
|
pages = drm_malloc_ab(npages, sizeof(struct page *));
|
|
if (pages == NULL)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
gfpmask |= mapping_gfp_mask(mapping);
|
|
|
|
for (i = 0; i < npages; i++) {
|
|
p = shmem_read_mapping_page_gfp(mapping, i, gfpmask);
|
|
if (IS_ERR(p))
|
|
goto fail;
|
|
pages[i] = p;
|
|
|
|
/* There is a hypothetical issue w/ drivers that require
|
|
* buffer memory in the low 4GB.. if the pages are un-
|
|
* pinned, and swapped out, they can end up swapped back
|
|
* in above 4GB. If pages are already in memory, then
|
|
* shmem_read_mapping_page_gfp will ignore the gfpmask,
|
|
* even if the already in-memory page disobeys the mask.
|
|
*
|
|
* It is only a theoretical issue today, because none of
|
|
* the devices with this limitation can be populated with
|
|
* enough memory to trigger the issue. But this BUG_ON()
|
|
* is here as a reminder in case the problem with
|
|
* shmem_read_mapping_page_gfp() isn't solved by the time
|
|
* it does become a real issue.
|
|
*
|
|
* See this thread: http://lkml.org/lkml/2011/7/11/238
|
|
*/
|
|
BUG_ON((gfpmask & __GFP_DMA32) &&
|
|
(page_to_pfn(p) >= 0x00100000UL));
|
|
}
|
|
|
|
return pages;
|
|
|
|
fail:
|
|
while (i--)
|
|
page_cache_release(pages[i]);
|
|
|
|
drm_free_large(pages);
|
|
return ERR_CAST(p);
|
|
}
|
|
|
|
/**
|
|
* drm_gem_put_pages - helper to free backing pages for a GEM object
|
|
* @obj: obj in question
|
|
* @pages: pages to free
|
|
*/
|
|
void _drm_gem_put_pages(struct drm_gem_object *obj, struct page **pages,
|
|
bool dirty, bool accessed)
|
|
{
|
|
int i, npages;
|
|
|
|
npages = obj->size >> PAGE_SHIFT;
|
|
|
|
for (i = 0; i < npages; i++) {
|
|
if (dirty)
|
|
set_page_dirty(pages[i]);
|
|
|
|
if (accessed)
|
|
mark_page_accessed(pages[i]);
|
|
|
|
/* Undo the reference we took when populating the table */
|
|
page_cache_release(pages[i]);
|
|
}
|
|
|
|
drm_free_large(pages);
|
|
}
|
|
|
|
int
|
|
_drm_gem_create_mmap_offset_size(struct drm_gem_object *obj, size_t size)
|
|
{
|
|
struct drm_device *dev = obj->dev;
|
|
struct drm_gem_mm *mm = dev->mm_private;
|
|
|
|
return drm_vma_offset_add(&mm->vma_manager, &obj->vma_node,
|
|
size / PAGE_SIZE);
|
|
}
|