mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-23 19:14:30 +08:00
f3b2bbdc8a
Port over the mgag200 fix to cirrus as it suffers the same issue. On F19 testing, it was noticed we get a lot of errors in dmesg about being unable to reserve the buffer when plymouth starts, this is due to the buffer being in the process of migrating, so it makes sense we can't reserve it. In order to deal with it, this adds delayed updates for the dirty updates, when the bo is unreservable, in the normal console case this shouldn't ever happen, its just when plymouth or X is pushing the console bo to system memory. Cc: stable@vger.kernel.org Signed-off-by: Dave Airlie <airlied@redhat.com>
332 lines
7.9 KiB
C
332 lines
7.9 KiB
C
/*
|
|
* Copyright 2012 Red Hat
|
|
*
|
|
* This file is subject to the terms and conditions of the GNU General
|
|
* Public License version 2. See the file COPYING in the main
|
|
* directory of this archive for more details.
|
|
*
|
|
* Authors: Matthew Garrett
|
|
* Dave Airlie
|
|
*/
|
|
#include <linux/module.h>
|
|
#include <drm/drmP.h>
|
|
#include <drm/drm_fb_helper.h>
|
|
#include <drm/drm_crtc_helper.h>
|
|
|
|
#include <linux/fb.h>
|
|
|
|
#include "cirrus_drv.h"
|
|
|
|
static void cirrus_dirty_update(struct cirrus_fbdev *afbdev,
|
|
int x, int y, int width, int height)
|
|
{
|
|
int i;
|
|
struct drm_gem_object *obj;
|
|
struct cirrus_bo *bo;
|
|
int src_offset, dst_offset;
|
|
int bpp = (afbdev->gfb.base.bits_per_pixel + 7)/8;
|
|
int ret;
|
|
bool unmap = false;
|
|
bool store_for_later = false;
|
|
int x2, y2;
|
|
unsigned long flags;
|
|
|
|
obj = afbdev->gfb.obj;
|
|
bo = gem_to_cirrus_bo(obj);
|
|
|
|
/*
|
|
* try and reserve the BO, if we fail with busy
|
|
* then the BO is being moved and we should
|
|
* store up the damage until later.
|
|
*/
|
|
ret = cirrus_bo_reserve(bo, true);
|
|
if (ret) {
|
|
if (ret != -EBUSY)
|
|
return;
|
|
store_for_later = true;
|
|
}
|
|
|
|
x2 = x + width - 1;
|
|
y2 = y + height - 1;
|
|
spin_lock_irqsave(&afbdev->dirty_lock, flags);
|
|
|
|
if (afbdev->y1 < y)
|
|
y = afbdev->y1;
|
|
if (afbdev->y2 > y2)
|
|
y2 = afbdev->y2;
|
|
if (afbdev->x1 < x)
|
|
x = afbdev->x1;
|
|
if (afbdev->x2 > x2)
|
|
x2 = afbdev->x2;
|
|
|
|
if (store_for_later) {
|
|
afbdev->x1 = x;
|
|
afbdev->x2 = x2;
|
|
afbdev->y1 = y;
|
|
afbdev->y2 = y2;
|
|
spin_unlock_irqrestore(&afbdev->dirty_lock, flags);
|
|
return;
|
|
}
|
|
|
|
afbdev->x1 = afbdev->y1 = INT_MAX;
|
|
afbdev->x2 = afbdev->y2 = 0;
|
|
spin_unlock_irqrestore(&afbdev->dirty_lock, flags);
|
|
|
|
if (!bo->kmap.virtual) {
|
|
ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap);
|
|
if (ret) {
|
|
DRM_ERROR("failed to kmap fb updates\n");
|
|
cirrus_bo_unreserve(bo);
|
|
return;
|
|
}
|
|
unmap = true;
|
|
}
|
|
for (i = y; i < y + height; i++) {
|
|
/* assume equal stride for now */
|
|
src_offset = dst_offset = i * afbdev->gfb.base.pitches[0] + (x * bpp);
|
|
memcpy_toio(bo->kmap.virtual + src_offset, afbdev->sysram + src_offset, width * bpp);
|
|
|
|
}
|
|
if (unmap)
|
|
ttm_bo_kunmap(&bo->kmap);
|
|
|
|
cirrus_bo_unreserve(bo);
|
|
}
|
|
|
|
static void cirrus_fillrect(struct fb_info *info,
|
|
const struct fb_fillrect *rect)
|
|
{
|
|
struct cirrus_fbdev *afbdev = info->par;
|
|
sys_fillrect(info, rect);
|
|
cirrus_dirty_update(afbdev, rect->dx, rect->dy, rect->width,
|
|
rect->height);
|
|
}
|
|
|
|
static void cirrus_copyarea(struct fb_info *info,
|
|
const struct fb_copyarea *area)
|
|
{
|
|
struct cirrus_fbdev *afbdev = info->par;
|
|
sys_copyarea(info, area);
|
|
cirrus_dirty_update(afbdev, area->dx, area->dy, area->width,
|
|
area->height);
|
|
}
|
|
|
|
static void cirrus_imageblit(struct fb_info *info,
|
|
const struct fb_image *image)
|
|
{
|
|
struct cirrus_fbdev *afbdev = info->par;
|
|
sys_imageblit(info, image);
|
|
cirrus_dirty_update(afbdev, image->dx, image->dy, image->width,
|
|
image->height);
|
|
}
|
|
|
|
|
|
static struct fb_ops cirrusfb_ops = {
|
|
.owner = THIS_MODULE,
|
|
.fb_check_var = drm_fb_helper_check_var,
|
|
.fb_set_par = drm_fb_helper_set_par,
|
|
.fb_fillrect = cirrus_fillrect,
|
|
.fb_copyarea = cirrus_copyarea,
|
|
.fb_imageblit = cirrus_imageblit,
|
|
.fb_pan_display = drm_fb_helper_pan_display,
|
|
.fb_blank = drm_fb_helper_blank,
|
|
.fb_setcmap = drm_fb_helper_setcmap,
|
|
};
|
|
|
|
static int cirrusfb_create_object(struct cirrus_fbdev *afbdev,
|
|
struct drm_mode_fb_cmd2 *mode_cmd,
|
|
struct drm_gem_object **gobj_p)
|
|
{
|
|
struct drm_device *dev = afbdev->helper.dev;
|
|
u32 bpp, depth;
|
|
u32 size;
|
|
struct drm_gem_object *gobj;
|
|
|
|
int ret = 0;
|
|
drm_fb_get_bpp_depth(mode_cmd->pixel_format, &depth, &bpp);
|
|
|
|
if (bpp > 24)
|
|
return -EINVAL;
|
|
size = mode_cmd->pitches[0] * mode_cmd->height;
|
|
ret = cirrus_gem_create(dev, size, true, &gobj);
|
|
if (ret)
|
|
return ret;
|
|
|
|
*gobj_p = gobj;
|
|
return ret;
|
|
}
|
|
|
|
static int cirrusfb_create(struct drm_fb_helper *helper,
|
|
struct drm_fb_helper_surface_size *sizes)
|
|
{
|
|
struct cirrus_fbdev *gfbdev = (struct cirrus_fbdev *)helper;
|
|
struct drm_device *dev = gfbdev->helper.dev;
|
|
struct cirrus_device *cdev = gfbdev->helper.dev->dev_private;
|
|
struct fb_info *info;
|
|
struct drm_framebuffer *fb;
|
|
struct drm_mode_fb_cmd2 mode_cmd;
|
|
struct device *device = &dev->pdev->dev;
|
|
void *sysram;
|
|
struct drm_gem_object *gobj = NULL;
|
|
struct cirrus_bo *bo = NULL;
|
|
int size, ret;
|
|
|
|
mode_cmd.width = sizes->surface_width;
|
|
mode_cmd.height = sizes->surface_height;
|
|
mode_cmd.pitches[0] = mode_cmd.width * ((sizes->surface_bpp + 7) / 8);
|
|
mode_cmd.pixel_format = drm_mode_legacy_fb_format(sizes->surface_bpp,
|
|
sizes->surface_depth);
|
|
size = mode_cmd.pitches[0] * mode_cmd.height;
|
|
|
|
ret = cirrusfb_create_object(gfbdev, &mode_cmd, &gobj);
|
|
if (ret) {
|
|
DRM_ERROR("failed to create fbcon backing object %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
bo = gem_to_cirrus_bo(gobj);
|
|
|
|
sysram = vmalloc(size);
|
|
if (!sysram)
|
|
return -ENOMEM;
|
|
|
|
info = framebuffer_alloc(0, device);
|
|
if (info == NULL)
|
|
return -ENOMEM;
|
|
|
|
info->par = gfbdev;
|
|
|
|
ret = cirrus_framebuffer_init(cdev->dev, &gfbdev->gfb, &mode_cmd, gobj);
|
|
if (ret)
|
|
return ret;
|
|
|
|
gfbdev->sysram = sysram;
|
|
gfbdev->size = size;
|
|
|
|
fb = &gfbdev->gfb.base;
|
|
if (!fb) {
|
|
DRM_INFO("fb is NULL\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* setup helper */
|
|
gfbdev->helper.fb = fb;
|
|
gfbdev->helper.fbdev = info;
|
|
|
|
strcpy(info->fix.id, "cirrusdrmfb");
|
|
|
|
|
|
info->flags = FBINFO_DEFAULT;
|
|
info->fbops = &cirrusfb_ops;
|
|
|
|
drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
|
|
drm_fb_helper_fill_var(info, &gfbdev->helper, sizes->fb_width,
|
|
sizes->fb_height);
|
|
|
|
/* setup aperture base/size for vesafb takeover */
|
|
info->apertures = alloc_apertures(1);
|
|
if (!info->apertures) {
|
|
ret = -ENOMEM;
|
|
goto out_iounmap;
|
|
}
|
|
info->apertures->ranges[0].base = cdev->dev->mode_config.fb_base;
|
|
info->apertures->ranges[0].size = cdev->mc.vram_size;
|
|
|
|
info->screen_base = sysram;
|
|
info->screen_size = size;
|
|
|
|
info->fix.mmio_start = 0;
|
|
info->fix.mmio_len = 0;
|
|
|
|
ret = fb_alloc_cmap(&info->cmap, 256, 0);
|
|
if (ret) {
|
|
DRM_ERROR("%s: can't allocate color map\n", info->fix.id);
|
|
ret = -ENOMEM;
|
|
goto out_iounmap;
|
|
}
|
|
|
|
DRM_INFO("fb mappable at 0x%lX\n", info->fix.smem_start);
|
|
DRM_INFO("vram aper at 0x%lX\n", (unsigned long)info->fix.smem_start);
|
|
DRM_INFO("size %lu\n", (unsigned long)info->fix.smem_len);
|
|
DRM_INFO("fb depth is %d\n", fb->depth);
|
|
DRM_INFO(" pitch is %d\n", fb->pitches[0]);
|
|
|
|
return 0;
|
|
out_iounmap:
|
|
return ret;
|
|
}
|
|
|
|
static int cirrus_fbdev_destroy(struct drm_device *dev,
|
|
struct cirrus_fbdev *gfbdev)
|
|
{
|
|
struct fb_info *info;
|
|
struct cirrus_framebuffer *gfb = &gfbdev->gfb;
|
|
|
|
if (gfbdev->helper.fbdev) {
|
|
info = gfbdev->helper.fbdev;
|
|
|
|
unregister_framebuffer(info);
|
|
if (info->cmap.len)
|
|
fb_dealloc_cmap(&info->cmap);
|
|
framebuffer_release(info);
|
|
}
|
|
|
|
if (gfb->obj) {
|
|
drm_gem_object_unreference_unlocked(gfb->obj);
|
|
gfb->obj = NULL;
|
|
}
|
|
|
|
vfree(gfbdev->sysram);
|
|
drm_fb_helper_fini(&gfbdev->helper);
|
|
drm_framebuffer_unregister_private(&gfb->base);
|
|
drm_framebuffer_cleanup(&gfb->base);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct drm_fb_helper_funcs cirrus_fb_helper_funcs = {
|
|
.gamma_set = cirrus_crtc_fb_gamma_set,
|
|
.gamma_get = cirrus_crtc_fb_gamma_get,
|
|
.fb_probe = cirrusfb_create,
|
|
};
|
|
|
|
int cirrus_fbdev_init(struct cirrus_device *cdev)
|
|
{
|
|
struct cirrus_fbdev *gfbdev;
|
|
int ret;
|
|
int bpp_sel = 24;
|
|
|
|
/*bpp_sel = 8;*/
|
|
gfbdev = kzalloc(sizeof(struct cirrus_fbdev), GFP_KERNEL);
|
|
if (!gfbdev)
|
|
return -ENOMEM;
|
|
|
|
cdev->mode_info.gfbdev = gfbdev;
|
|
gfbdev->helper.funcs = &cirrus_fb_helper_funcs;
|
|
spin_lock_init(&gfbdev->dirty_lock);
|
|
|
|
ret = drm_fb_helper_init(cdev->dev, &gfbdev->helper,
|
|
cdev->num_crtc, CIRRUSFB_CONN_LIMIT);
|
|
if (ret) {
|
|
kfree(gfbdev);
|
|
return ret;
|
|
}
|
|
drm_fb_helper_single_add_all_connectors(&gfbdev->helper);
|
|
|
|
/* disable all the possible outputs/crtcs before entering KMS mode */
|
|
drm_helper_disable_unused_functions(cdev->dev);
|
|
drm_fb_helper_initial_config(&gfbdev->helper, bpp_sel);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void cirrus_fbdev_fini(struct cirrus_device *cdev)
|
|
{
|
|
if (!cdev->mode_info.gfbdev)
|
|
return;
|
|
|
|
cirrus_fbdev_destroy(cdev->dev, cdev->mode_info.gfbdev);
|
|
kfree(cdev->mode_info.gfbdev);
|
|
cdev->mode_info.gfbdev = NULL;
|
|
}
|