| /* |
| * Copyright 2012 Red Hat |
| * |
| * This file is subject to the terms and conditions of the GNU General |
| * Public License version 2. See the file COPYING in the main |
| * directory of this archive for more details. |
| * |
| * Authors: Matthew Garrett |
| * Dave Airlie |
| */ |
| #include <linux/module.h> |
| #include <drm/drmP.h> |
| #include <drm/drm_fb_helper.h> |
| #include <drm/drm_crtc_helper.h> |
| |
| #include <linux/fb.h> |
| |
| #include "cirrus_drv.h" |
| |
| static void cirrus_dirty_update(struct cirrus_fbdev *afbdev, |
| int x, int y, int width, int height) |
| { |
| int i; |
| struct drm_gem_object *obj; |
| struct cirrus_bo *bo; |
| int src_offset, dst_offset; |
| int bpp = (afbdev->gfb.base.bits_per_pixel + 7)/8; |
| int ret = -EBUSY; |
| bool unmap = false; |
| bool store_for_later = false; |
| int x2, y2; |
| unsigned long flags; |
| |
| obj = afbdev->gfb.obj; |
| bo = gem_to_cirrus_bo(obj); |
| |
| /* |
| * try and reserve the BO, if we fail with busy |
| * then the BO is being moved and we should |
| * store up the damage until later. |
| */ |
| if (!drm_can_sleep()) |
| ret = cirrus_bo_reserve(bo, true); |
| if (ret) { |
| if (ret != -EBUSY) |
| return; |
| store_for_later = true; |
| } |
| |
| x2 = x + width - 1; |
| y2 = y + height - 1; |
| spin_lock_irqsave(&afbdev->dirty_lock, flags); |
| |
| if (afbdev->y1 < y) |
| y = afbdev->y1; |
| if (afbdev->y2 > y2) |
| y2 = afbdev->y2; |
| if (afbdev->x1 < x) |
| x = afbdev->x1; |
| if (afbdev->x2 > x2) |
| x2 = afbdev->x2; |
| |
| if (store_for_later) { |
| afbdev->x1 = x; |
| afbdev->x2 = x2; |
| afbdev->y1 = y; |
| afbdev->y2 = y2; |
| spin_unlock_irqrestore(&afbdev->dirty_lock, flags); |
| return; |
| } |
| |
| afbdev->x1 = afbdev->y1 = INT_MAX; |
| afbdev->x2 = afbdev->y2 = 0; |
| spin_unlock_irqrestore(&afbdev->dirty_lock, flags); |
| |
| if (!bo->kmap.virtual) { |
| ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap); |
| if (ret) { |
| DRM_ERROR("failed to kmap fb updates\n"); |
| cirrus_bo_unreserve(bo); |
| return; |
| } |
| unmap = true; |
| } |
| for (i = y; i < y + height; i++) { |
| /* assume equal stride for now */ |
| src_offset = dst_offset = i * afbdev->gfb.base.pitches[0] + (x * bpp); |
| memcpy_toio(bo->kmap.virtual + src_offset, afbdev->sysram + src_offset, width * bpp); |
| |
| } |
| if (unmap) |
| ttm_bo_kunmap(&bo->kmap); |
| |
| cirrus_bo_unreserve(bo); |
| } |
| |
| static void cirrus_fillrect(struct fb_info *info, |
| const struct fb_fillrect *rect) |
| { |
| struct cirrus_fbdev *afbdev = info->par; |
| sys_fillrect(info, rect); |
| cirrus_dirty_update(afbdev, rect->dx, rect->dy, rect->width, |
| rect->height); |
| } |
| |
| static void cirrus_copyarea(struct fb_info *info, |
| const struct fb_copyarea *area) |
| { |
| struct cirrus_fbdev *afbdev = info->par; |
| sys_copyarea(info, area); |
| cirrus_dirty_update(afbdev, area->dx, area->dy, area->width, |
| area->height); |
| } |
| |
| static void cirrus_imageblit(struct fb_info *info, |
| const struct fb_image *image) |
| { |
| struct cirrus_fbdev *afbdev = info->par; |
| sys_imageblit(info, image); |
| cirrus_dirty_update(afbdev, image->dx, image->dy, image->width, |
| image->height); |
| } |
| |
| |
| static struct fb_ops cirrusfb_ops = { |
| .owner = THIS_MODULE, |
| .fb_check_var = drm_fb_helper_check_var, |
| .fb_set_par = drm_fb_helper_set_par, |
| .fb_fillrect = cirrus_fillrect, |
| .fb_copyarea = cirrus_copyarea, |
| .fb_imageblit = cirrus_imageblit, |
| .fb_pan_display = drm_fb_helper_pan_display, |
| .fb_blank = drm_fb_helper_blank, |
| .fb_setcmap = drm_fb_helper_setcmap, |
| }; |
| |
| static int cirrusfb_create_object(struct cirrus_fbdev *afbdev, |
| struct drm_mode_fb_cmd2 *mode_cmd, |
| struct drm_gem_object **gobj_p) |
| { |
| struct drm_device *dev = afbdev->helper.dev; |
| u32 bpp, depth; |
| u32 size; |
| struct drm_gem_object *gobj; |
| |
| int ret = 0; |
| drm_fb_get_bpp_depth(mode_cmd->pixel_format, &depth, &bpp); |
| |
| if (bpp > 24) |
| return -EINVAL; |
| size = mode_cmd->pitches[0] * mode_cmd->height; |
| ret = cirrus_gem_create(dev, size, true, &gobj); |
| if (ret) |
| return ret; |
| |
| *gobj_p = gobj; |
| return ret; |
| } |
| |
| static int cirrusfb_create(struct drm_fb_helper *helper, |
| struct drm_fb_helper_surface_size *sizes) |
| { |
| struct cirrus_fbdev *gfbdev = (struct cirrus_fbdev *)helper; |
| struct drm_device *dev = gfbdev->helper.dev; |
| struct cirrus_device *cdev = gfbdev->helper.dev->dev_private; |
| struct fb_info *info; |
| struct drm_framebuffer *fb; |
| struct drm_mode_fb_cmd2 mode_cmd; |
| struct device *device = &dev->pdev->dev; |
| void *sysram; |
| struct drm_gem_object *gobj = NULL; |
| struct cirrus_bo *bo = NULL; |
| int size, ret; |
| |
| mode_cmd.width = sizes->surface_width; |
| mode_cmd.height = sizes->surface_height; |
| mode_cmd.pitches[0] = mode_cmd.width * ((sizes->surface_bpp + 7) / 8); |
| mode_cmd.pixel_format = drm_mode_legacy_fb_format(sizes->surface_bpp, |
| sizes->surface_depth); |
| size = mode_cmd.pitches[0] * mode_cmd.height; |
| |
| ret = cirrusfb_create_object(gfbdev, &mode_cmd, &gobj); |
| if (ret) { |
| DRM_ERROR("failed to create fbcon backing object %d\n", ret); |
| return ret; |
| } |
| |
| bo = gem_to_cirrus_bo(gobj); |
| |
| sysram = vmalloc(size); |
| if (!sysram) |
| return -ENOMEM; |
| |
| info = framebuffer_alloc(0, device); |
| if (info == NULL) |
| return -ENOMEM; |
| |
| info->par = gfbdev; |
| |
| ret = cirrus_framebuffer_init(cdev->dev, &gfbdev->gfb, &mode_cmd, gobj); |
| if (ret) |
| return ret; |
| |
| gfbdev->sysram = sysram; |
| gfbdev->size = size; |
| |
| fb = &gfbdev->gfb.base; |
| if (!fb) { |
| DRM_INFO("fb is NULL\n"); |
| return -EINVAL; |
| } |
| |
| /* setup helper */ |
| gfbdev->helper.fb = fb; |
| gfbdev->helper.fbdev = info; |
| |
| strcpy(info->fix.id, "cirrusdrmfb"); |
| |
| |
| info->flags = FBINFO_DEFAULT; |
| info->fbops = &cirrusfb_ops; |
| |
| drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth); |
| drm_fb_helper_fill_var(info, &gfbdev->helper, sizes->fb_width, |
| sizes->fb_height); |
| |
| /* setup aperture base/size for vesafb takeover */ |
| info->apertures = alloc_apertures(1); |
| if (!info->apertures) { |
| ret = -ENOMEM; |
| goto out_iounmap; |
| } |
| info->apertures->ranges[0].base = cdev->dev->mode_config.fb_base; |
| info->apertures->ranges[0].size = cdev->mc.vram_size; |
| |
| info->fix.smem_start = cdev->dev->mode_config.fb_base; |
| info->fix.smem_len = cdev->mc.vram_size; |
| |
| info->screen_base = sysram; |
| info->screen_size = size; |
| |
| info->fix.mmio_start = 0; |
| info->fix.mmio_len = 0; |
| |
| ret = fb_alloc_cmap(&info->cmap, 256, 0); |
| if (ret) { |
| DRM_ERROR("%s: can't allocate color map\n", info->fix.id); |
| ret = -ENOMEM; |
| goto out_iounmap; |
| } |
| |
| DRM_INFO("fb mappable at 0x%lX\n", info->fix.smem_start); |
| DRM_INFO("vram aper at 0x%lX\n", (unsigned long)info->fix.smem_start); |
| DRM_INFO("size %lu\n", (unsigned long)info->fix.smem_len); |
| DRM_INFO("fb depth is %d\n", fb->depth); |
| DRM_INFO(" pitch is %d\n", fb->pitches[0]); |
| |
| return 0; |
| out_iounmap: |
| return ret; |
| } |
| |
| static int cirrus_fbdev_destroy(struct drm_device *dev, |
| struct cirrus_fbdev *gfbdev) |
| { |
| struct fb_info *info; |
| struct cirrus_framebuffer *gfb = &gfbdev->gfb; |
| |
| if (gfbdev->helper.fbdev) { |
| info = gfbdev->helper.fbdev; |
| |
| unregister_framebuffer(info); |
| if (info->cmap.len) |
| fb_dealloc_cmap(&info->cmap); |
| framebuffer_release(info); |
| } |
| |
| if (gfb->obj) { |
| drm_gem_object_unreference_unlocked(gfb->obj); |
| gfb->obj = NULL; |
| } |
| |
| vfree(gfbdev->sysram); |
| drm_fb_helper_fini(&gfbdev->helper); |
| drm_framebuffer_unregister_private(&gfb->base); |
| drm_framebuffer_cleanup(&gfb->base); |
| |
| return 0; |
| } |
| |
| static struct drm_fb_helper_funcs cirrus_fb_helper_funcs = { |
| .gamma_set = cirrus_crtc_fb_gamma_set, |
| .gamma_get = cirrus_crtc_fb_gamma_get, |
| .fb_probe = cirrusfb_create, |
| }; |
| |
| int cirrus_fbdev_init(struct cirrus_device *cdev) |
| { |
| struct cirrus_fbdev *gfbdev; |
| int ret; |
| int bpp_sel = 24; |
| |
| /*bpp_sel = 8;*/ |
| gfbdev = kzalloc(sizeof(struct cirrus_fbdev), GFP_KERNEL); |
| if (!gfbdev) |
| return -ENOMEM; |
| |
| cdev->mode_info.gfbdev = gfbdev; |
| gfbdev->helper.funcs = &cirrus_fb_helper_funcs; |
| spin_lock_init(&gfbdev->dirty_lock); |
| |
| ret = drm_fb_helper_init(cdev->dev, &gfbdev->helper, |
| cdev->num_crtc, CIRRUSFB_CONN_LIMIT); |
| if (ret) { |
| kfree(gfbdev); |
| return ret; |
| } |
| drm_fb_helper_single_add_all_connectors(&gfbdev->helper); |
| |
| /* disable all the possible outputs/crtcs before entering KMS mode */ |
| drm_helper_disable_unused_functions(cdev->dev); |
| drm_fb_helper_initial_config(&gfbdev->helper, bpp_sel); |
| |
| return 0; |
| } |
| |
| void cirrus_fbdev_fini(struct cirrus_device *cdev) |
| { |
| if (!cdev->mode_info.gfbdev) |
| return; |
| |
| cirrus_fbdev_destroy(cdev->dev, cdev->mode_info.gfbdev); |
| kfree(cdev->mode_info.gfbdev); |
| cdev->mode_info.gfbdev = NULL; |
| } |