[vbox-dev] [PATCH 3/7] Additions: linux/drm: Change vbox_fb.c to kernel coding style
Hans de Goede
hdegoede at redhat.com
Thu Jun 8 19:08:10 UTC 2017
This is the result of running linux/scripts/Lindent + manual cleanups.
After this the file passes linux/scripts/checkpatch -f
except for the LINUX_VERSION_CODE checks.
This patch contains no functional changes, only coding style fixes,
including changing uintXX_t types to uXX.
Signed-off-by: Hans de Goede <hdegoede at redhat.com>
---
src/VBox/Additions/linux/drm/vbox_fb.c | 618 +++++++++++++++++----------------
1 file changed, 315 insertions(+), 303 deletions(-)
diff --git a/src/VBox/Additions/linux/drm/vbox_fb.c b/src/VBox/Additions/linux/drm/vbox_fb.c
index cf538c1a..0519bd61 100644
--- a/src/VBox/Additions/linux/drm/vbox_fb.c
+++ b/src/VBox/Additions/linux/drm/vbox_fb.c
@@ -46,7 +46,6 @@
#include <linux/fb.h>
#include <linux/init.h>
-
#include <drm/drmP.h>
#include <drm/drm_crtc.h>
#include <drm/drm_fb_helper.h>
@@ -58,407 +57,420 @@
* Tell the host about dirty rectangles to update.
*/
static void vbox_dirty_update(struct vbox_fbdev *fbdev,
- int x, int y, int width, int height)
+ int x, int y, int width, int height)
{
- struct drm_gem_object *obj;
- struct vbox_bo *bo;
- int ret = -EBUSY;
- bool store_for_later = false;
- int x2, y2;
- unsigned long flags;
- struct drm_clip_rect rect;
-
- obj = fbdev->afb.obj;
- bo = gem_to_vbox_bo(obj);
-
- /*
- * try and reserve the BO, if we fail with busy
- * then the BO is being moved and we should
- * store up the damage until later.
- */
- if (drm_can_sleep())
- ret = vbox_bo_reserve(bo, true);
- if (ret) {
- if (ret != -EBUSY)
- return;
-
- store_for_later = true;
- }
-
- x2 = x + width - 1;
- y2 = y + height - 1;
- spin_lock_irqsave(&fbdev->dirty_lock, flags);
-
- if (fbdev->y1 < y)
- y = fbdev->y1;
- if (fbdev->y2 > y2)
- y2 = fbdev->y2;
- if (fbdev->x1 < x)
- x = fbdev->x1;
- if (fbdev->x2 > x2)
- x2 = fbdev->x2;
-
- if (store_for_later) {
- fbdev->x1 = x;
- fbdev->x2 = x2;
- fbdev->y1 = y;
- fbdev->y2 = y2;
- spin_unlock_irqrestore(&fbdev->dirty_lock, flags);
- return;
- }
-
- fbdev->x1 = fbdev->y1 = INT_MAX;
- fbdev->x2 = fbdev->y2 = 0;
- spin_unlock_irqrestore(&fbdev->dirty_lock, flags);
-
- /* Not sure why the original code subtracted 1 here, but I will keep it that
- * way to avoid unnecessary differences. */
- rect.x1 = x;
- rect.x2 = x2 + 1;
- rect.y1 = y;
- rect.y2 = y2 + 1;
- vbox_framebuffer_dirty_rectangles(&fbdev->afb.base, &rect, 1);
-
- vbox_bo_unreserve(bo);
+ struct drm_gem_object *obj;
+ struct vbox_bo *bo;
+ int ret = -EBUSY;
+ bool store_for_later = false;
+ int x2, y2;
+ unsigned long flags;
+ struct drm_clip_rect rect;
+
+ obj = fbdev->afb.obj;
+ bo = gem_to_vbox_bo(obj);
+
+ /*
+ * try and reserve the BO, if we fail with busy
+ * then the BO is being moved and we should
+ * store up the damage until later.
+ */
+ if (drm_can_sleep())
+ ret = vbox_bo_reserve(bo, true);
+ if (ret) {
+ if (ret != -EBUSY)
+ return;
+
+ store_for_later = true;
+ }
+
+ x2 = x + width - 1;
+ y2 = y + height - 1;
+ spin_lock_irqsave(&fbdev->dirty_lock, flags);
+
+ if (fbdev->y1 < y)
+ y = fbdev->y1;
+ if (fbdev->y2 > y2)
+ y2 = fbdev->y2;
+ if (fbdev->x1 < x)
+ x = fbdev->x1;
+ if (fbdev->x2 > x2)
+ x2 = fbdev->x2;
+
+ if (store_for_later) {
+ fbdev->x1 = x;
+ fbdev->x2 = x2;
+ fbdev->y1 = y;
+ fbdev->y2 = y2;
+ spin_unlock_irqrestore(&fbdev->dirty_lock, flags);
+ return;
+ }
+
+ fbdev->x1 = INT_MAX;
+ fbdev->y1 = INT_MAX;
+ fbdev->x2 = 0;
+ fbdev->y2 = 0;
+
+ spin_unlock_irqrestore(&fbdev->dirty_lock, flags);
+
+ /*
+ * Not sure why the original code subtracted 1 here, but I will keep
+ * it that way to avoid unnecessary differences.
+ */
+ rect.x1 = x;
+ rect.x2 = x2 + 1;
+ rect.y1 = y;
+ rect.y2 = y2 + 1;
+ vbox_framebuffer_dirty_rectangles(&fbdev->afb.base, &rect, 1);
+
+ vbox_bo_unreserve(bo);
}
#ifdef CONFIG_FB_DEFERRED_IO
-static void vbox_deferred_io(struct fb_info *info,
- struct list_head *pagelist)
+static void vbox_deferred_io(struct fb_info *info, struct list_head *pagelist)
{
- struct vbox_fbdev *fbdev = info->par;
- unsigned long start, end, min, max;
- struct page *page;
- int y1, y2;
-
- min = ULONG_MAX;
- max = 0;
- list_for_each_entry(page, pagelist, lru) {
- start = page->index << PAGE_SHIFT;
- end = start + PAGE_SIZE - 1;
- min = min(min, start);
- max = max(max, end);
- }
-
- if (min < max) {
- y1 = min / info->fix.line_length;
- y2 = (max / info->fix.line_length) + 1;
- printk(KERN_INFO "%s: Calling dirty update: 0, %d, %d, %d\n",
- __func__, y1, info->var.xres, y2 - y1 - 1);
- vbox_dirty_update(fbdev, 0, y1, info->var.xres, y2 - y1 - 1);
- }
+ struct vbox_fbdev *fbdev = info->par;
+ unsigned long start, end, min, max;
+ struct page *page;
+ int y1, y2;
+
+ min = ULONG_MAX;
+ max = 0;
+ list_for_each_entry(page, pagelist, lru) {
+ start = page->index << PAGE_SHIFT;
+ end = start + PAGE_SIZE - 1;
+ min = min(min, start);
+ max = max(max, end);
+ }
+
+ if (min < max) {
+ y1 = min / info->fix.line_length;
+ y2 = (max / info->fix.line_length) + 1;
+ DRM_INFO("%s: Calling dirty update: 0, %d, %d, %d\n",
+ __func__, y1, info->var.xres, y2 - y1 - 1);
+ vbox_dirty_update(fbdev, 0, y1, info->var.xres, y2 - y1 - 1);
+ }
}
-static struct fb_deferred_io vbox_defio =
-{
- .delay = VBOX_DIRTY_DELAY,
- .deferred_io = vbox_deferred_io,
+static struct fb_deferred_io vbox_defio = {
+ .delay = VBOX_DIRTY_DELAY,
+ .deferred_io = vbox_deferred_io,
};
#endif
-static void vbox_fillrect(struct fb_info *info,
- const struct fb_fillrect *rect)
+static void vbox_fillrect(struct fb_info *info, const struct fb_fillrect *rect)
{
- struct vbox_fbdev *fbdev = info->par;
- sys_fillrect(info, rect);
- vbox_dirty_update(fbdev, rect->dx, rect->dy, rect->width,
- rect->height);
+ struct vbox_fbdev *fbdev = info->par;
+
+ sys_fillrect(info, rect);
+ vbox_dirty_update(fbdev, rect->dx, rect->dy, rect->width, rect->height);
}
-static void vbox_copyarea(struct fb_info *info,
- const struct fb_copyarea *area)
+static void vbox_copyarea(struct fb_info *info, const struct fb_copyarea *area)
{
- struct vbox_fbdev *fbdev = info->par;
- sys_copyarea(info, area);
- vbox_dirty_update(fbdev, area->dx, area->dy, area->width,
- area->height);
+ struct vbox_fbdev *fbdev = info->par;
+
+ sys_copyarea(info, area);
+ vbox_dirty_update(fbdev, area->dx, area->dy, area->width, area->height);
}
-static void vbox_imageblit(struct fb_info *info,
- const struct fb_image *image)
+static void vbox_imageblit(struct fb_info *info, const struct fb_image *image)
{
- struct vbox_fbdev *fbdev = info->par;
- sys_imageblit(info, image);
- vbox_dirty_update(fbdev, image->dx, image->dy, image->width,
- image->height);
+ struct vbox_fbdev *fbdev = info->par;
+
+ sys_imageblit(info, image);
+ vbox_dirty_update(fbdev, image->dx, image->dy, image->width,
+ image->height);
}
static struct fb_ops vboxfb_ops = {
- .owner = THIS_MODULE,
- .fb_check_var = drm_fb_helper_check_var,
- .fb_set_par = drm_fb_helper_set_par,
- .fb_fillrect = vbox_fillrect,
- .fb_copyarea = vbox_copyarea,
- .fb_imageblit = vbox_imageblit,
- .fb_pan_display = drm_fb_helper_pan_display,
- .fb_blank = drm_fb_helper_blank,
- .fb_setcmap = drm_fb_helper_setcmap,
- .fb_debug_enter = drm_fb_helper_debug_enter,
- .fb_debug_leave = drm_fb_helper_debug_leave,
+ .owner = THIS_MODULE,
+ .fb_check_var = drm_fb_helper_check_var,
+ .fb_set_par = drm_fb_helper_set_par,
+ .fb_fillrect = vbox_fillrect,
+ .fb_copyarea = vbox_copyarea,
+ .fb_imageblit = vbox_imageblit,
+ .fb_pan_display = drm_fb_helper_pan_display,
+ .fb_blank = drm_fb_helper_blank,
+ .fb_setcmap = drm_fb_helper_setcmap,
+ .fb_debug_enter = drm_fb_helper_debug_enter,
+ .fb_debug_leave = drm_fb_helper_debug_leave,
};
static int vboxfb_create_object(struct vbox_fbdev *fbdev,
- struct DRM_MODE_FB_CMD *mode_cmd,
- struct drm_gem_object **gobj_p)
+ struct DRM_MODE_FB_CMD *mode_cmd,
+ struct drm_gem_object **gobj_p)
{
- struct drm_device *dev = fbdev->helper.dev;
- u32 size;
- struct drm_gem_object *gobj;
+ struct drm_device *dev = fbdev->helper.dev;
+ u32 size;
+ struct drm_gem_object *gobj;
#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 3, 0)
- __u32 pitch = mode_cmd->pitch;
+ u32 pitch = mode_cmd->pitch;
#else
- __u32 pitch = mode_cmd->pitches[0];
+ u32 pitch = mode_cmd->pitches[0];
#endif
- int ret = 0;
+ int ret = 0;
- size = pitch * mode_cmd->height;
- ret = vbox_gem_create(dev, size, true, &gobj);
- if (ret)
- return ret;
+ size = pitch * mode_cmd->height;
+ ret = vbox_gem_create(dev, size, true, &gobj);
+ if (ret)
+ return ret;
- *gobj_p = gobj;
- return ret;
+ *gobj_p = gobj;
+ return ret;
}
static int vboxfb_create(struct drm_fb_helper *helper,
- struct drm_fb_helper_surface_size *sizes)
+ struct drm_fb_helper_surface_size *sizes)
{
- struct vbox_fbdev *fbdev =
- container_of(helper, struct vbox_fbdev, helper);
- struct drm_device *dev = fbdev->helper.dev;
- struct DRM_MODE_FB_CMD mode_cmd;
- struct drm_framebuffer *fb;
- struct fb_info *info;
- __u32 pitch;
- int size, ret;
- struct device *device = &dev->pdev->dev;
- struct drm_gem_object *gobj = NULL;
- struct vbox_bo *bo = NULL;
- mode_cmd.width = sizes->surface_width;
- mode_cmd.height = sizes->surface_height;
- pitch = mode_cmd.width * ((sizes->surface_bpp + 7) / 8);
+ struct vbox_fbdev *fbdev =
+ container_of(helper, struct vbox_fbdev, helper);
+ struct drm_device *dev = fbdev->helper.dev;
+ struct DRM_MODE_FB_CMD mode_cmd;
+ struct drm_framebuffer *fb;
+ struct fb_info *info;
+ struct device *device = &dev->pdev->dev;
+ struct drm_gem_object *gobj = NULL;
+ struct vbox_bo *bo = NULL;
+ int size, ret;
+ u32 pitch;
+
+ mode_cmd.width = sizes->surface_width;
+ mode_cmd.height = sizes->surface_height;
+ pitch = mode_cmd.width * ((sizes->surface_bpp + 7) / 8);
#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 3, 0)
- mode_cmd.bpp = sizes->surface_bpp;
- mode_cmd.depth = sizes->surface_depth;
- mode_cmd.pitch = pitch;
+ mode_cmd.bpp = sizes->surface_bpp;
+ mode_cmd.depth = sizes->surface_depth;
+ mode_cmd.pitch = pitch;
#else
- mode_cmd.pixel_format = drm_mode_legacy_fb_format(sizes->surface_bpp,
- sizes->surface_depth);
- mode_cmd.pitches[0] = pitch;
+ mode_cmd.pixel_format = drm_mode_legacy_fb_format(sizes->surface_bpp,
+ sizes->surface_depth);
+ mode_cmd.pitches[0] = pitch;
#endif
- size = pitch * mode_cmd.height;
-
- ret = vboxfb_create_object(fbdev, &mode_cmd, &gobj);
- if (ret) {
- DRM_ERROR("failed to create fbcon backing object %d\n", ret);
- return ret;
- }
-
- ret = vbox_framebuffer_init(dev, &fbdev->afb, &mode_cmd, gobj);
- if (ret)
- return ret;
-
- bo = gem_to_vbox_bo(gobj);
-
- ret = vbox_bo_reserve(bo, false);
- if (ret)
- return ret;
-
- ret = vbox_bo_pin(bo, TTM_PL_FLAG_VRAM, NULL);
- if (ret) {
- vbox_bo_unreserve(bo);
- return ret;
- }
-
- ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap);
- vbox_bo_unreserve(bo);
- if (ret) {
- DRM_ERROR("failed to kmap fbcon\n");
- return ret;
- }
-
- info = framebuffer_alloc(0, device);
- if (!info)
- return -ENOMEM;
- info->par = fbdev;
-
- fbdev->size = size;
-
- fb = &fbdev->afb.base;
- fbdev->helper.fb = fb;
- fbdev->helper.fbdev = info;
-
- strcpy(info->fix.id, "vboxdrmfb");
-
- /* The last flag forces a mode set on VT switches even if the kernel does
- * not think it is needed. */
- info->flags = FBINFO_DEFAULT | FBINFO_CAN_FORCE_OUTPUT
- | FBINFO_MISC_ALWAYS_SETPAR;
- info->fbops = &vboxfb_ops;
-
- ret = fb_alloc_cmap(&info->cmap, 256, 0);
- if (ret)
- return -ENOMEM;
-
- /* This seems to be done for safety checking that the framebuffer is not
- * registered twice by different drivers. */
- info->apertures = alloc_apertures(1);
- if (!info->apertures)
- return -ENOMEM;
- info->apertures->ranges[0].base = pci_resource_start(dev->pdev, 0);
- info->apertures->ranges[0].size = pci_resource_len(dev->pdev, 0);
+ size = pitch * mode_cmd.height;
+
+ ret = vboxfb_create_object(fbdev, &mode_cmd, &gobj);
+ if (ret) {
+ DRM_ERROR("failed to create fbcon backing object %d\n", ret);
+ return ret;
+ }
+
+ ret = vbox_framebuffer_init(dev, &fbdev->afb, &mode_cmd, gobj);
+ if (ret)
+ return ret;
+
+ bo = gem_to_vbox_bo(gobj);
+
+ ret = vbox_bo_reserve(bo, false);
+ if (ret)
+ return ret;
+
+ ret = vbox_bo_pin(bo, TTM_PL_FLAG_VRAM, NULL);
+ if (ret) {
+ vbox_bo_unreserve(bo);
+ return ret;
+ }
+
+ ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap);
+ vbox_bo_unreserve(bo);
+ if (ret) {
+ DRM_ERROR("failed to kmap fbcon\n");
+ return ret;
+ }
+
+ info = framebuffer_alloc(0, device);
+ if (!info)
+ return -ENOMEM;
+ info->par = fbdev;
+
+ fbdev->size = size;
+
+ fb = &fbdev->afb.base;
+ fbdev->helper.fb = fb;
+ fbdev->helper.fbdev = info;
+
+ strcpy(info->fix.id, "vboxdrmfb");
+
+ /*
+ * The last flag forces a mode set on VT switches even if the kernel
+ * does not think it is needed.
+ */
+ info->flags = FBINFO_DEFAULT | FBINFO_CAN_FORCE_OUTPUT |
+ FBINFO_MISC_ALWAYS_SETPAR;
+ info->fbops = &vboxfb_ops;
+
+ ret = fb_alloc_cmap(&info->cmap, 256, 0);
+ if (ret)
+ return -ENOMEM;
+
+ /*
+ * This seems to be done for safety checking that the framebuffer
+ * is not registered twice by different drivers.
+ */
+ info->apertures = alloc_apertures(1);
+ if (!info->apertures)
+ return -ENOMEM;
+ info->apertures->ranges[0].base = pci_resource_start(dev->pdev, 0);
+ info->apertures->ranges[0].size = pci_resource_len(dev->pdev, 0);
#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 11, 0)
- drm_fb_helper_fill_fix(info, fb->pitches[0], fb->format->depth);
+ drm_fb_helper_fill_fix(info, fb->pitches[0], fb->format->depth);
#else
- drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
+ drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
#endif
- drm_fb_helper_fill_var(info, &fbdev->helper, sizes->fb_width, sizes->fb_height);
+ drm_fb_helper_fill_var(info, &fbdev->helper, sizes->fb_width,
+ sizes->fb_height);
- info->screen_base = bo->kmap.virtual;
- info->screen_size = size;
+ info->screen_base = bo->kmap.virtual;
+ info->screen_size = size;
#ifdef CONFIG_FB_DEFERRED_IO
- info->fbdefio = &vbox_defio;
- fb_deferred_io_init(info);
+ info->fbdefio = &vbox_defio;
+ fb_deferred_io_init(info);
#endif
- info->pixmap.flags = FB_PIXMAP_SYSTEM;
+ info->pixmap.flags = FB_PIXMAP_SYSTEM;
- DRM_DEBUG_KMS("allocated %dx%d\n",
- fb->width, fb->height);
+ DRM_DEBUG_KMS("allocated %dx%d\n", fb->width, fb->height);
- return 0;
+ return 0;
}
static void vbox_fb_gamma_set(struct drm_crtc *crtc, u16 red, u16 green,
- u16 blue, int regno)
+ u16 blue, int regno)
{
-
}
static void vbox_fb_gamma_get(struct drm_crtc *crtc, u16 *red, u16 *green,
- u16 *blue, int regno)
+ u16 *blue, int regno)
{
- *red = regno;
- *green = regno;
- *blue = regno;
+ *red = regno;
+ *green = regno;
+ *blue = regno;
}
static struct drm_fb_helper_funcs vbox_fb_helper_funcs = {
- .gamma_set = vbox_fb_gamma_set,
- .gamma_get = vbox_fb_gamma_get,
- .fb_probe = vboxfb_create,
+ .gamma_set = vbox_fb_gamma_set,
+ .gamma_get = vbox_fb_gamma_get,
+ .fb_probe = vboxfb_create,
};
-static void vbox_fbdev_destroy(struct drm_device *dev,
- struct vbox_fbdev *fbdev)
+static void vbox_fbdev_destroy(struct drm_device *dev, struct vbox_fbdev *fbdev)
{
- struct fb_info *info;
- struct vbox_framebuffer *afb = &fbdev->afb;
- if (fbdev->helper.fbdev) {
- info = fbdev->helper.fbdev;
- unregister_framebuffer(info);
- if (info->cmap.len)
- fb_dealloc_cmap(&info->cmap);
- framebuffer_release(info);
- }
-
- if (afb->obj) {
- struct vbox_bo *bo = gem_to_vbox_bo(afb->obj);
- if (!vbox_bo_reserve(bo, false)) {
- if (bo->kmap.virtual)
- ttm_bo_kunmap(&bo->kmap);
- /* QXL does this, but is it really needed before freeing? */
- if (bo->pin_count)
- vbox_bo_unpin(bo);
- vbox_bo_unreserve(bo);
- }
- drm_gem_object_unreference_unlocked(afb->obj);
- afb->obj = NULL;
- }
- drm_fb_helper_fini(&fbdev->helper);
+ struct fb_info *info;
+ struct vbox_framebuffer *afb = &fbdev->afb;
+
+ if (fbdev->helper.fbdev) {
+ info = fbdev->helper.fbdev;
+ unregister_framebuffer(info);
+ if (info->cmap.len)
+ fb_dealloc_cmap(&info->cmap);
+ framebuffer_release(info);
+ }
+
+ if (afb->obj) {
+ struct vbox_bo *bo = gem_to_vbox_bo(afb->obj);
+
+ if (!vbox_bo_reserve(bo, false)) {
+ if (bo->kmap.virtual)
+ ttm_bo_kunmap(&bo->kmap);
+ /*
+ * QXL does this, but is it really needed before
+ * freeing?
+ */
+ if (bo->pin_count)
+ vbox_bo_unpin(bo);
+ vbox_bo_unreserve(bo);
+ }
+ drm_gem_object_unreference_unlocked(afb->obj);
+ afb->obj = NULL;
+ }
+ drm_fb_helper_fini(&fbdev->helper);
#if LINUX_VERSION_CODE >= KERNEL_VERSION(3, 9, 0)
- drm_framebuffer_unregister_private(&afb->base);
+ drm_framebuffer_unregister_private(&afb->base);
#endif
- drm_framebuffer_cleanup(&afb->base);
+ drm_framebuffer_cleanup(&afb->base);
}
int vbox_fbdev_init(struct drm_device *dev)
{
- struct vbox_private *vbox = dev->dev_private;
- struct vbox_fbdev *fbdev;
- int ret;
+ struct vbox_private *vbox = dev->dev_private;
+ struct vbox_fbdev *fbdev;
+ int ret;
- fbdev = kzalloc(sizeof(struct vbox_fbdev), GFP_KERNEL);
- if (!fbdev)
- return -ENOMEM;
+ fbdev = kzalloc(sizeof(*fbdev), GFP_KERNEL);
+ if (!fbdev)
+ return -ENOMEM;
- vbox->fbdev = fbdev;
- spin_lock_init(&fbdev->dirty_lock);
+ vbox->fbdev = fbdev;
+ spin_lock_init(&fbdev->dirty_lock);
#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 17, 0)
- fbdev->helper.funcs = &vbox_fb_helper_funcs;
+ fbdev->helper.funcs = &vbox_fb_helper_funcs;
#else
- drm_fb_helper_prepare(dev, &fbdev->helper, &vbox_fb_helper_funcs);
+ drm_fb_helper_prepare(dev, &fbdev->helper, &vbox_fb_helper_funcs);
#endif
#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 11, 0)
- ret = drm_fb_helper_init(dev, &fbdev->helper, vbox->num_crtcs);
+ ret = drm_fb_helper_init(dev, &fbdev->helper, vbox->num_crtcs);
#else
- ret = drm_fb_helper_init(dev, &fbdev->helper, vbox->num_crtcs, vbox->num_crtcs);
+ ret =
+ drm_fb_helper_init(dev, &fbdev->helper, vbox->num_crtcs,
+ vbox->num_crtcs);
#endif
- if (ret)
- goto free;
+ if (ret)
+ goto free;
- ret = drm_fb_helper_single_add_all_connectors(&fbdev->helper);
- if (ret)
- goto fini;
+ ret = drm_fb_helper_single_add_all_connectors(&fbdev->helper);
+ if (ret)
+ goto fini;
- /* disable all the possible outputs/crtcs before entering KMS mode */
- drm_helper_disable_unused_functions(dev);
+ /* disable all the possible outputs/crtcs before entering KMS mode */
+ drm_helper_disable_unused_functions(dev);
- ret = drm_fb_helper_initial_config(&fbdev->helper, 32);
- if (ret)
- goto fini;
+ ret = drm_fb_helper_initial_config(&fbdev->helper, 32);
+ if (ret)
+ goto fini;
+
+ return 0;
- return 0;
fini:
- drm_fb_helper_fini(&fbdev->helper);
+ drm_fb_helper_fini(&fbdev->helper);
free:
- kfree(fbdev);
- vbox->fbdev = NULL;
- return ret;
+ kfree(fbdev);
+ vbox->fbdev = NULL;
+
+ return ret;
}
void vbox_fbdev_fini(struct drm_device *dev)
{
- struct vbox_private *vbox = dev->dev_private;
+ struct vbox_private *vbox = dev->dev_private;
- if (!vbox->fbdev)
- return;
+ if (!vbox->fbdev)
+ return;
- vbox_fbdev_destroy(dev, vbox->fbdev);
- kfree(vbox->fbdev);
- vbox->fbdev = NULL;
+ vbox_fbdev_destroy(dev, vbox->fbdev);
+ kfree(vbox->fbdev);
+ vbox->fbdev = NULL;
}
void vbox_fbdev_set_suspend(struct drm_device *dev, int state)
{
- struct vbox_private *vbox = dev->dev_private;
+ struct vbox_private *vbox = dev->dev_private;
- if (!vbox->fbdev)
- return;
+ if (!vbox->fbdev)
+ return;
- fb_set_suspend(vbox->fbdev->helper.fbdev, state);
+ fb_set_suspend(vbox->fbdev->helper.fbdev, state);
}
void vbox_fbdev_set_base(struct vbox_private *vbox, unsigned long gpu_addr)
{
- vbox->fbdev->helper.fbdev->fix.smem_start =
- vbox->fbdev->helper.fbdev->apertures->ranges[0].base +
- gpu_addr;
- vbox->fbdev->helper.fbdev->fix.smem_len = vbox->available_vram_size - gpu_addr;
+ vbox->fbdev->helper.fbdev->fix.smem_start =
+ vbox->fbdev->helper.fbdev->apertures->ranges[0].base + gpu_addr;
+ vbox->fbdev->helper.fbdev->fix.smem_len =
+ vbox->available_vram_size - gpu_addr;
}
--
2.13.0
More information about the vbox-dev
mailing list