mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-30 23:26:45 +07:00
22ee861c81
The host may be touching this part of VRAM at modesetting, even if we never use it ourselves, since we blit screen updates from 3D surfaces. Make sure no DMA buffers are placed in this part of VRAM. V2: Fix an error check in vmw_surface_dmabuf_pin(). Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com> Signed-off-by: Jakob Bornecrantz <jakob@vmware.com> Signed-off-by: Dave Airlie <airlied@redhat.com>
903 lines
24 KiB
C
903 lines
24 KiB
C
/**************************************************************************
|
|
*
|
|
* Copyright © 2009 VMware, Inc., Palo Alto, CA., USA
|
|
* All Rights Reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the
|
|
* "Software"), to deal in the Software without restriction, including
|
|
* without limitation the rights to use, copy, modify, merge, publish,
|
|
* distribute, sub license, and/or sell copies of the Software, and to
|
|
* permit persons to whom the Software is furnished to do so, subject to
|
|
* the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the
|
|
* next paragraph) shall be included in all copies or substantial portions
|
|
* of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
|
|
* THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
|
|
* DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
|
|
* OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
|
|
* USE OR OTHER DEALINGS IN THE SOFTWARE.
|
|
*
|
|
**************************************************************************/
|
|
|
|
#include "vmwgfx_kms.h"
|
|
|
|
/* Might need a hrtimer here? */
|
|
#define VMWGFX_PRESENT_RATE ((HZ / 60 > 0) ? HZ / 60 : 1)
|
|
|
|
static int vmw_surface_dmabuf_pin(struct vmw_framebuffer *vfb);
|
|
static int vmw_surface_dmabuf_unpin(struct vmw_framebuffer *vfb);
|
|
|
|
void vmw_display_unit_cleanup(struct vmw_display_unit *du)
|
|
{
|
|
if (du->cursor_surface)
|
|
vmw_surface_unreference(&du->cursor_surface);
|
|
if (du->cursor_dmabuf)
|
|
vmw_dmabuf_unreference(&du->cursor_dmabuf);
|
|
drm_crtc_cleanup(&du->crtc);
|
|
drm_encoder_cleanup(&du->encoder);
|
|
drm_connector_cleanup(&du->connector);
|
|
}
|
|
|
|
/*
|
|
* Display Unit Cursor functions
|
|
*/
|
|
|
|
int vmw_cursor_update_image(struct vmw_private *dev_priv,
|
|
u32 *image, u32 width, u32 height,
|
|
u32 hotspotX, u32 hotspotY)
|
|
{
|
|
struct {
|
|
u32 cmd;
|
|
SVGAFifoCmdDefineAlphaCursor cursor;
|
|
} *cmd;
|
|
u32 image_size = width * height * 4;
|
|
u32 cmd_size = sizeof(*cmd) + image_size;
|
|
|
|
if (!image)
|
|
return -EINVAL;
|
|
|
|
cmd = vmw_fifo_reserve(dev_priv, cmd_size);
|
|
if (unlikely(cmd == NULL)) {
|
|
DRM_ERROR("Fifo reserve failed.\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
memset(cmd, 0, sizeof(*cmd));
|
|
|
|
memcpy(&cmd[1], image, image_size);
|
|
|
|
cmd->cmd = cpu_to_le32(SVGA_CMD_DEFINE_ALPHA_CURSOR);
|
|
cmd->cursor.id = cpu_to_le32(0);
|
|
cmd->cursor.width = cpu_to_le32(width);
|
|
cmd->cursor.height = cpu_to_le32(height);
|
|
cmd->cursor.hotspotX = cpu_to_le32(hotspotX);
|
|
cmd->cursor.hotspotY = cpu_to_le32(hotspotY);
|
|
|
|
vmw_fifo_commit(dev_priv, cmd_size);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void vmw_cursor_update_position(struct vmw_private *dev_priv,
|
|
bool show, int x, int y)
|
|
{
|
|
__le32 __iomem *fifo_mem = dev_priv->mmio_virt;
|
|
uint32_t count;
|
|
|
|
iowrite32(show ? 1 : 0, fifo_mem + SVGA_FIFO_CURSOR_ON);
|
|
iowrite32(x, fifo_mem + SVGA_FIFO_CURSOR_X);
|
|
iowrite32(y, fifo_mem + SVGA_FIFO_CURSOR_Y);
|
|
count = ioread32(fifo_mem + SVGA_FIFO_CURSOR_COUNT);
|
|
iowrite32(++count, fifo_mem + SVGA_FIFO_CURSOR_COUNT);
|
|
}
|
|
|
|
int vmw_du_crtc_cursor_set(struct drm_crtc *crtc, struct drm_file *file_priv,
|
|
uint32_t handle, uint32_t width, uint32_t height)
|
|
{
|
|
struct vmw_private *dev_priv = vmw_priv(crtc->dev);
|
|
struct ttm_object_file *tfile = vmw_fpriv(file_priv)->tfile;
|
|
struct vmw_display_unit *du = vmw_crtc_to_du(crtc);
|
|
struct vmw_surface *surface = NULL;
|
|
struct vmw_dma_buffer *dmabuf = NULL;
|
|
int ret;
|
|
|
|
if (handle) {
|
|
ret = vmw_user_surface_lookup_handle(dev_priv, tfile,
|
|
handle, &surface);
|
|
if (!ret) {
|
|
if (!surface->snooper.image) {
|
|
DRM_ERROR("surface not suitable for cursor\n");
|
|
return -EINVAL;
|
|
}
|
|
} else {
|
|
ret = vmw_user_dmabuf_lookup(tfile,
|
|
handle, &dmabuf);
|
|
if (ret) {
|
|
DRM_ERROR("failed to find surface or dmabuf: %i\n", ret);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* takedown old cursor */
|
|
if (du->cursor_surface) {
|
|
du->cursor_surface->snooper.crtc = NULL;
|
|
vmw_surface_unreference(&du->cursor_surface);
|
|
}
|
|
if (du->cursor_dmabuf)
|
|
vmw_dmabuf_unreference(&du->cursor_dmabuf);
|
|
|
|
/* setup new image */
|
|
if (surface) {
|
|
/* vmw_user_surface_lookup takes one reference */
|
|
du->cursor_surface = surface;
|
|
|
|
du->cursor_surface->snooper.crtc = crtc;
|
|
du->cursor_age = du->cursor_surface->snooper.age;
|
|
vmw_cursor_update_image(dev_priv, surface->snooper.image,
|
|
64, 64, du->hotspot_x, du->hotspot_y);
|
|
} else if (dmabuf) {
|
|
struct ttm_bo_kmap_obj map;
|
|
unsigned long kmap_offset;
|
|
unsigned long kmap_num;
|
|
void *virtual;
|
|
bool dummy;
|
|
|
|
/* vmw_user_surface_lookup takes one reference */
|
|
du->cursor_dmabuf = dmabuf;
|
|
|
|
kmap_offset = 0;
|
|
kmap_num = (64*64*4) >> PAGE_SHIFT;
|
|
|
|
ret = ttm_bo_reserve(&dmabuf->base, true, false, false, 0);
|
|
if (unlikely(ret != 0)) {
|
|
DRM_ERROR("reserve failed\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ret = ttm_bo_kmap(&dmabuf->base, kmap_offset, kmap_num, &map);
|
|
if (unlikely(ret != 0))
|
|
goto err_unreserve;
|
|
|
|
virtual = ttm_kmap_obj_virtual(&map, &dummy);
|
|
vmw_cursor_update_image(dev_priv, virtual, 64, 64,
|
|
du->hotspot_x, du->hotspot_y);
|
|
|
|
ttm_bo_kunmap(&map);
|
|
err_unreserve:
|
|
ttm_bo_unreserve(&dmabuf->base);
|
|
|
|
} else {
|
|
vmw_cursor_update_position(dev_priv, false, 0, 0);
|
|
return 0;
|
|
}
|
|
|
|
vmw_cursor_update_position(dev_priv, true, du->cursor_x, du->cursor_y);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int vmw_du_crtc_cursor_move(struct drm_crtc *crtc, int x, int y)
|
|
{
|
|
struct vmw_private *dev_priv = vmw_priv(crtc->dev);
|
|
struct vmw_display_unit *du = vmw_crtc_to_du(crtc);
|
|
bool shown = du->cursor_surface || du->cursor_dmabuf ? true : false;
|
|
|
|
du->cursor_x = x + crtc->x;
|
|
du->cursor_y = y + crtc->y;
|
|
|
|
vmw_cursor_update_position(dev_priv, shown,
|
|
du->cursor_x, du->cursor_y);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void vmw_kms_cursor_snoop(struct vmw_surface *srf,
|
|
struct ttm_object_file *tfile,
|
|
struct ttm_buffer_object *bo,
|
|
SVGA3dCmdHeader *header)
|
|
{
|
|
struct ttm_bo_kmap_obj map;
|
|
unsigned long kmap_offset;
|
|
unsigned long kmap_num;
|
|
SVGA3dCopyBox *box;
|
|
unsigned box_count;
|
|
void *virtual;
|
|
bool dummy;
|
|
struct vmw_dma_cmd {
|
|
SVGA3dCmdHeader header;
|
|
SVGA3dCmdSurfaceDMA dma;
|
|
} *cmd;
|
|
int ret;
|
|
|
|
cmd = container_of(header, struct vmw_dma_cmd, header);
|
|
|
|
/* No snooper installed */
|
|
if (!srf->snooper.image)
|
|
return;
|
|
|
|
if (cmd->dma.host.face != 0 || cmd->dma.host.mipmap != 0) {
|
|
DRM_ERROR("face and mipmap for cursors should never != 0\n");
|
|
return;
|
|
}
|
|
|
|
if (cmd->header.size < 64) {
|
|
DRM_ERROR("at least one full copy box must be given\n");
|
|
return;
|
|
}
|
|
|
|
box = (SVGA3dCopyBox *)&cmd[1];
|
|
box_count = (cmd->header.size - sizeof(SVGA3dCmdSurfaceDMA)) /
|
|
sizeof(SVGA3dCopyBox);
|
|
|
|
if (cmd->dma.guest.pitch != (64 * 4) ||
|
|
cmd->dma.guest.ptr.offset % PAGE_SIZE ||
|
|
box->x != 0 || box->y != 0 || box->z != 0 ||
|
|
box->srcx != 0 || box->srcy != 0 || box->srcz != 0 ||
|
|
box->w != 64 || box->h != 64 || box->d != 1 ||
|
|
box_count != 1) {
|
|
/* TODO handle none page aligned offsets */
|
|
/* TODO handle partial uploads and pitch != 256 */
|
|
/* TODO handle more then one copy (size != 64) */
|
|
DRM_ERROR("lazy programer, cant handle wierd stuff\n");
|
|
return;
|
|
}
|
|
|
|
kmap_offset = cmd->dma.guest.ptr.offset >> PAGE_SHIFT;
|
|
kmap_num = (64*64*4) >> PAGE_SHIFT;
|
|
|
|
ret = ttm_bo_reserve(bo, true, false, false, 0);
|
|
if (unlikely(ret != 0)) {
|
|
DRM_ERROR("reserve failed\n");
|
|
return;
|
|
}
|
|
|
|
ret = ttm_bo_kmap(bo, kmap_offset, kmap_num, &map);
|
|
if (unlikely(ret != 0))
|
|
goto err_unreserve;
|
|
|
|
virtual = ttm_kmap_obj_virtual(&map, &dummy);
|
|
|
|
memcpy(srf->snooper.image, virtual, 64*64*4);
|
|
srf->snooper.age++;
|
|
|
|
/* we can't call this function from this function since execbuf has
|
|
* reserved fifo space.
|
|
*
|
|
* if (srf->snooper.crtc)
|
|
* vmw_ldu_crtc_cursor_update_image(dev_priv,
|
|
* srf->snooper.image, 64, 64,
|
|
* du->hotspot_x, du->hotspot_y);
|
|
*/
|
|
|
|
ttm_bo_kunmap(&map);
|
|
err_unreserve:
|
|
ttm_bo_unreserve(bo);
|
|
}
|
|
|
|
void vmw_kms_cursor_post_execbuf(struct vmw_private *dev_priv)
|
|
{
|
|
struct drm_device *dev = dev_priv->dev;
|
|
struct vmw_display_unit *du;
|
|
struct drm_crtc *crtc;
|
|
|
|
mutex_lock(&dev->mode_config.mutex);
|
|
|
|
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
|
|
du = vmw_crtc_to_du(crtc);
|
|
if (!du->cursor_surface ||
|
|
du->cursor_age == du->cursor_surface->snooper.age)
|
|
continue;
|
|
|
|
du->cursor_age = du->cursor_surface->snooper.age;
|
|
vmw_cursor_update_image(dev_priv,
|
|
du->cursor_surface->snooper.image,
|
|
64, 64, du->hotspot_x, du->hotspot_y);
|
|
}
|
|
|
|
mutex_unlock(&dev->mode_config.mutex);
|
|
}
|
|
|
|
/*
|
|
* Generic framebuffer code
|
|
*/
|
|
|
|
int vmw_framebuffer_create_handle(struct drm_framebuffer *fb,
|
|
struct drm_file *file_priv,
|
|
unsigned int *handle)
|
|
{
|
|
if (handle)
|
|
handle = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Surface framebuffer code
|
|
*/
|
|
|
|
#define vmw_framebuffer_to_vfbs(x) \
|
|
container_of(x, struct vmw_framebuffer_surface, base.base)
|
|
|
|
struct vmw_framebuffer_surface {
|
|
struct vmw_framebuffer base;
|
|
struct vmw_surface *surface;
|
|
struct vmw_dma_buffer *buffer;
|
|
struct delayed_work d_work;
|
|
struct mutex work_lock;
|
|
bool present_fs;
|
|
};
|
|
|
|
void vmw_framebuffer_surface_destroy(struct drm_framebuffer *framebuffer)
|
|
{
|
|
struct vmw_framebuffer_surface *vfb =
|
|
vmw_framebuffer_to_vfbs(framebuffer);
|
|
|
|
cancel_delayed_work_sync(&vfb->d_work);
|
|
drm_framebuffer_cleanup(framebuffer);
|
|
vmw_surface_unreference(&vfb->surface);
|
|
|
|
kfree(framebuffer);
|
|
}
|
|
|
|
static void vmw_framebuffer_present_fs_callback(struct work_struct *work)
|
|
{
|
|
struct delayed_work *d_work =
|
|
container_of(work, struct delayed_work, work);
|
|
struct vmw_framebuffer_surface *vfbs =
|
|
container_of(d_work, struct vmw_framebuffer_surface, d_work);
|
|
struct vmw_surface *surf = vfbs->surface;
|
|
struct drm_framebuffer *framebuffer = &vfbs->base.base;
|
|
struct vmw_private *dev_priv = vmw_priv(framebuffer->dev);
|
|
|
|
struct {
|
|
SVGA3dCmdHeader header;
|
|
SVGA3dCmdPresent body;
|
|
SVGA3dCopyRect cr;
|
|
} *cmd;
|
|
|
|
mutex_lock(&vfbs->work_lock);
|
|
if (!vfbs->present_fs)
|
|
goto out_unlock;
|
|
|
|
cmd = vmw_fifo_reserve(dev_priv, sizeof(*cmd));
|
|
if (unlikely(cmd == NULL))
|
|
goto out_resched;
|
|
|
|
cmd->header.id = cpu_to_le32(SVGA_3D_CMD_PRESENT);
|
|
cmd->header.size = cpu_to_le32(sizeof(cmd->body) + sizeof(cmd->cr));
|
|
cmd->body.sid = cpu_to_le32(surf->res.id);
|
|
cmd->cr.x = cpu_to_le32(0);
|
|
cmd->cr.y = cpu_to_le32(0);
|
|
cmd->cr.srcx = cmd->cr.x;
|
|
cmd->cr.srcy = cmd->cr.y;
|
|
cmd->cr.w = cpu_to_le32(framebuffer->width);
|
|
cmd->cr.h = cpu_to_le32(framebuffer->height);
|
|
vfbs->present_fs = false;
|
|
vmw_fifo_commit(dev_priv, sizeof(*cmd));
|
|
out_resched:
|
|
/**
|
|
* Will not re-add if already pending.
|
|
*/
|
|
schedule_delayed_work(&vfbs->d_work, VMWGFX_PRESENT_RATE);
|
|
out_unlock:
|
|
mutex_unlock(&vfbs->work_lock);
|
|
}
|
|
|
|
|
|
int vmw_framebuffer_surface_dirty(struct drm_framebuffer *framebuffer,
|
|
unsigned flags, unsigned color,
|
|
struct drm_clip_rect *clips,
|
|
unsigned num_clips)
|
|
{
|
|
struct vmw_private *dev_priv = vmw_priv(framebuffer->dev);
|
|
struct vmw_framebuffer_surface *vfbs =
|
|
vmw_framebuffer_to_vfbs(framebuffer);
|
|
struct vmw_surface *surf = vfbs->surface;
|
|
struct drm_clip_rect norect;
|
|
SVGA3dCopyRect *cr;
|
|
int i, inc = 1;
|
|
|
|
struct {
|
|
SVGA3dCmdHeader header;
|
|
SVGA3dCmdPresent body;
|
|
SVGA3dCopyRect cr;
|
|
} *cmd;
|
|
|
|
if (!num_clips ||
|
|
!(dev_priv->fifo.capabilities &
|
|
SVGA_FIFO_CAP_SCREEN_OBJECT)) {
|
|
int ret;
|
|
|
|
mutex_lock(&vfbs->work_lock);
|
|
vfbs->present_fs = true;
|
|
ret = schedule_delayed_work(&vfbs->d_work, VMWGFX_PRESENT_RATE);
|
|
mutex_unlock(&vfbs->work_lock);
|
|
if (ret) {
|
|
/**
|
|
* No work pending, Force immediate present.
|
|
*/
|
|
vmw_framebuffer_present_fs_callback(&vfbs->d_work.work);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
if (!num_clips) {
|
|
num_clips = 1;
|
|
clips = &norect;
|
|
norect.x1 = norect.y1 = 0;
|
|
norect.x2 = framebuffer->width;
|
|
norect.y2 = framebuffer->height;
|
|
} else if (flags & DRM_MODE_FB_DIRTY_ANNOTATE_COPY) {
|
|
num_clips /= 2;
|
|
inc = 2; /* skip source rects */
|
|
}
|
|
|
|
cmd = vmw_fifo_reserve(dev_priv, sizeof(*cmd) + (num_clips - 1) * sizeof(cmd->cr));
|
|
if (unlikely(cmd == NULL)) {
|
|
DRM_ERROR("Fifo reserve failed.\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
memset(cmd, 0, sizeof(*cmd));
|
|
|
|
cmd->header.id = cpu_to_le32(SVGA_3D_CMD_PRESENT);
|
|
cmd->header.size = cpu_to_le32(sizeof(cmd->body) + num_clips * sizeof(cmd->cr));
|
|
cmd->body.sid = cpu_to_le32(surf->res.id);
|
|
|
|
for (i = 0, cr = &cmd->cr; i < num_clips; i++, cr++, clips += inc) {
|
|
cr->x = cpu_to_le16(clips->x1);
|
|
cr->y = cpu_to_le16(clips->y1);
|
|
cr->srcx = cr->x;
|
|
cr->srcy = cr->y;
|
|
cr->w = cpu_to_le16(clips->x2 - clips->x1);
|
|
cr->h = cpu_to_le16(clips->y2 - clips->y1);
|
|
}
|
|
|
|
vmw_fifo_commit(dev_priv, sizeof(*cmd) + (num_clips - 1) * sizeof(cmd->cr));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct drm_framebuffer_funcs vmw_framebuffer_surface_funcs = {
|
|
.destroy = vmw_framebuffer_surface_destroy,
|
|
.dirty = vmw_framebuffer_surface_dirty,
|
|
.create_handle = vmw_framebuffer_create_handle,
|
|
};
|
|
|
|
int vmw_kms_new_framebuffer_surface(struct vmw_private *dev_priv,
|
|
struct vmw_surface *surface,
|
|
struct vmw_framebuffer **out,
|
|
unsigned width, unsigned height)
|
|
|
|
{
|
|
struct drm_device *dev = dev_priv->dev;
|
|
struct vmw_framebuffer_surface *vfbs;
|
|
int ret;
|
|
|
|
vfbs = kzalloc(sizeof(*vfbs), GFP_KERNEL);
|
|
if (!vfbs) {
|
|
ret = -ENOMEM;
|
|
goto out_err1;
|
|
}
|
|
|
|
ret = drm_framebuffer_init(dev, &vfbs->base.base,
|
|
&vmw_framebuffer_surface_funcs);
|
|
if (ret)
|
|
goto out_err2;
|
|
|
|
if (!vmw_surface_reference(surface)) {
|
|
DRM_ERROR("failed to reference surface %p\n", surface);
|
|
goto out_err3;
|
|
}
|
|
|
|
/* XXX get the first 3 from the surface info */
|
|
vfbs->base.base.bits_per_pixel = 32;
|
|
vfbs->base.base.pitch = width * 32 / 4;
|
|
vfbs->base.base.depth = 24;
|
|
vfbs->base.base.width = width;
|
|
vfbs->base.base.height = height;
|
|
vfbs->base.pin = &vmw_surface_dmabuf_pin;
|
|
vfbs->base.unpin = &vmw_surface_dmabuf_unpin;
|
|
vfbs->surface = surface;
|
|
mutex_init(&vfbs->work_lock);
|
|
INIT_DELAYED_WORK(&vfbs->d_work, &vmw_framebuffer_present_fs_callback);
|
|
*out = &vfbs->base;
|
|
|
|
return 0;
|
|
|
|
out_err3:
|
|
drm_framebuffer_cleanup(&vfbs->base.base);
|
|
out_err2:
|
|
kfree(vfbs);
|
|
out_err1:
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Dmabuf framebuffer code
|
|
*/
|
|
|
|
#define vmw_framebuffer_to_vfbd(x) \
|
|
container_of(x, struct vmw_framebuffer_dmabuf, base.base)
|
|
|
|
struct vmw_framebuffer_dmabuf {
|
|
struct vmw_framebuffer base;
|
|
struct vmw_dma_buffer *buffer;
|
|
};
|
|
|
|
void vmw_framebuffer_dmabuf_destroy(struct drm_framebuffer *framebuffer)
|
|
{
|
|
struct vmw_framebuffer_dmabuf *vfbd =
|
|
vmw_framebuffer_to_vfbd(framebuffer);
|
|
|
|
drm_framebuffer_cleanup(framebuffer);
|
|
vmw_dmabuf_unreference(&vfbd->buffer);
|
|
|
|
kfree(vfbd);
|
|
}
|
|
|
|
int vmw_framebuffer_dmabuf_dirty(struct drm_framebuffer *framebuffer,
|
|
unsigned flags, unsigned color,
|
|
struct drm_clip_rect *clips,
|
|
unsigned num_clips)
|
|
{
|
|
struct vmw_private *dev_priv = vmw_priv(framebuffer->dev);
|
|
struct drm_clip_rect norect;
|
|
struct {
|
|
uint32_t header;
|
|
SVGAFifoCmdUpdate body;
|
|
} *cmd;
|
|
int i, increment = 1;
|
|
|
|
if (!num_clips) {
|
|
num_clips = 1;
|
|
clips = &norect;
|
|
norect.x1 = norect.y1 = 0;
|
|
norect.x2 = framebuffer->width;
|
|
norect.y2 = framebuffer->height;
|
|
} else if (flags & DRM_MODE_FB_DIRTY_ANNOTATE_COPY) {
|
|
num_clips /= 2;
|
|
increment = 2;
|
|
}
|
|
|
|
cmd = vmw_fifo_reserve(dev_priv, sizeof(*cmd) * num_clips);
|
|
if (unlikely(cmd == NULL)) {
|
|
DRM_ERROR("Fifo reserve failed.\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
for (i = 0; i < num_clips; i++, clips += increment) {
|
|
cmd[i].header = cpu_to_le32(SVGA_CMD_UPDATE);
|
|
cmd[i].body.x = cpu_to_le32(clips->x1);
|
|
cmd[i].body.y = cpu_to_le32(clips->y1);
|
|
cmd[i].body.width = cpu_to_le32(clips->x2 - clips->x1);
|
|
cmd[i].body.height = cpu_to_le32(clips->y2 - clips->y1);
|
|
}
|
|
|
|
vmw_fifo_commit(dev_priv, sizeof(*cmd) * num_clips);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct drm_framebuffer_funcs vmw_framebuffer_dmabuf_funcs = {
|
|
.destroy = vmw_framebuffer_dmabuf_destroy,
|
|
.dirty = vmw_framebuffer_dmabuf_dirty,
|
|
.create_handle = vmw_framebuffer_create_handle,
|
|
};
|
|
|
|
static int vmw_surface_dmabuf_pin(struct vmw_framebuffer *vfb)
|
|
{
|
|
struct vmw_private *dev_priv = vmw_priv(vfb->base.dev);
|
|
struct vmw_framebuffer_surface *vfbs =
|
|
vmw_framebuffer_to_vfbs(&vfb->base);
|
|
unsigned long size = vfbs->base.base.pitch * vfbs->base.base.height;
|
|
int ret;
|
|
|
|
vfbs->buffer = kzalloc(sizeof(*vfbs->buffer), GFP_KERNEL);
|
|
if (unlikely(vfbs->buffer == NULL))
|
|
return -ENOMEM;
|
|
|
|
vmw_overlay_pause_all(dev_priv);
|
|
ret = vmw_dmabuf_init(dev_priv, vfbs->buffer, size,
|
|
&vmw_vram_ne_placement,
|
|
false, &vmw_dmabuf_bo_free);
|
|
vmw_overlay_resume_all(dev_priv);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int vmw_surface_dmabuf_unpin(struct vmw_framebuffer *vfb)
|
|
{
|
|
struct ttm_buffer_object *bo;
|
|
struct vmw_framebuffer_surface *vfbs =
|
|
vmw_framebuffer_to_vfbs(&vfb->base);
|
|
|
|
bo = &vfbs->buffer->base;
|
|
ttm_bo_unref(&bo);
|
|
vfbs->buffer = NULL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vmw_framebuffer_dmabuf_pin(struct vmw_framebuffer *vfb)
|
|
{
|
|
struct vmw_private *dev_priv = vmw_priv(vfb->base.dev);
|
|
struct vmw_framebuffer_dmabuf *vfbd =
|
|
vmw_framebuffer_to_vfbd(&vfb->base);
|
|
int ret;
|
|
|
|
|
|
vmw_overlay_pause_all(dev_priv);
|
|
|
|
ret = vmw_dmabuf_to_start_of_vram(dev_priv, vfbd->buffer);
|
|
|
|
vmw_overlay_resume_all(dev_priv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vmw_framebuffer_dmabuf_unpin(struct vmw_framebuffer *vfb)
|
|
{
|
|
struct vmw_private *dev_priv = vmw_priv(vfb->base.dev);
|
|
struct vmw_framebuffer_dmabuf *vfbd =
|
|
vmw_framebuffer_to_vfbd(&vfb->base);
|
|
|
|
if (!vfbd->buffer) {
|
|
WARN_ON(!vfbd->buffer);
|
|
return 0;
|
|
}
|
|
|
|
return vmw_dmabuf_from_vram(dev_priv, vfbd->buffer);
|
|
}
|
|
|
|
int vmw_kms_new_framebuffer_dmabuf(struct vmw_private *dev_priv,
|
|
struct vmw_dma_buffer *dmabuf,
|
|
struct vmw_framebuffer **out,
|
|
unsigned width, unsigned height)
|
|
|
|
{
|
|
struct drm_device *dev = dev_priv->dev;
|
|
struct vmw_framebuffer_dmabuf *vfbd;
|
|
int ret;
|
|
|
|
vfbd = kzalloc(sizeof(*vfbd), GFP_KERNEL);
|
|
if (!vfbd) {
|
|
ret = -ENOMEM;
|
|
goto out_err1;
|
|
}
|
|
|
|
ret = drm_framebuffer_init(dev, &vfbd->base.base,
|
|
&vmw_framebuffer_dmabuf_funcs);
|
|
if (ret)
|
|
goto out_err2;
|
|
|
|
if (!vmw_dmabuf_reference(dmabuf)) {
|
|
DRM_ERROR("failed to reference dmabuf %p\n", dmabuf);
|
|
goto out_err3;
|
|
}
|
|
|
|
/* XXX get the first 3 from the surface info */
|
|
vfbd->base.base.bits_per_pixel = 32;
|
|
vfbd->base.base.pitch = width * vfbd->base.base.bits_per_pixel / 8;
|
|
vfbd->base.base.depth = 24;
|
|
vfbd->base.base.width = width;
|
|
vfbd->base.base.height = height;
|
|
vfbd->base.pin = vmw_framebuffer_dmabuf_pin;
|
|
vfbd->base.unpin = vmw_framebuffer_dmabuf_unpin;
|
|
vfbd->buffer = dmabuf;
|
|
*out = &vfbd->base;
|
|
|
|
return 0;
|
|
|
|
out_err3:
|
|
drm_framebuffer_cleanup(&vfbd->base.base);
|
|
out_err2:
|
|
kfree(vfbd);
|
|
out_err1:
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Generic Kernel modesetting functions
|
|
*/
|
|
|
|
static struct drm_framebuffer *vmw_kms_fb_create(struct drm_device *dev,
|
|
struct drm_file *file_priv,
|
|
struct drm_mode_fb_cmd *mode_cmd)
|
|
{
|
|
struct vmw_private *dev_priv = vmw_priv(dev);
|
|
struct ttm_object_file *tfile = vmw_fpriv(file_priv)->tfile;
|
|
struct vmw_framebuffer *vfb = NULL;
|
|
struct vmw_surface *surface = NULL;
|
|
struct vmw_dma_buffer *bo = NULL;
|
|
int ret;
|
|
|
|
ret = vmw_user_surface_lookup_handle(dev_priv, tfile,
|
|
mode_cmd->handle, &surface);
|
|
if (ret)
|
|
goto try_dmabuf;
|
|
|
|
if (!surface->scanout)
|
|
goto err_not_scanout;
|
|
|
|
ret = vmw_kms_new_framebuffer_surface(dev_priv, surface, &vfb,
|
|
mode_cmd->width, mode_cmd->height);
|
|
|
|
/* vmw_user_surface_lookup takes one ref so does new_fb */
|
|
vmw_surface_unreference(&surface);
|
|
|
|
if (ret) {
|
|
DRM_ERROR("failed to create vmw_framebuffer: %i\n", ret);
|
|
return NULL;
|
|
}
|
|
return &vfb->base;
|
|
|
|
try_dmabuf:
|
|
DRM_INFO("%s: trying buffer\n", __func__);
|
|
|
|
ret = vmw_user_dmabuf_lookup(tfile, mode_cmd->handle, &bo);
|
|
if (ret) {
|
|
DRM_ERROR("failed to find buffer: %i\n", ret);
|
|
return NULL;
|
|
}
|
|
|
|
ret = vmw_kms_new_framebuffer_dmabuf(dev_priv, bo, &vfb,
|
|
mode_cmd->width, mode_cmd->height);
|
|
|
|
/* vmw_user_dmabuf_lookup takes one ref so does new_fb */
|
|
vmw_dmabuf_unreference(&bo);
|
|
|
|
if (ret) {
|
|
DRM_ERROR("failed to create vmw_framebuffer: %i\n", ret);
|
|
return NULL;
|
|
}
|
|
|
|
return &vfb->base;
|
|
|
|
err_not_scanout:
|
|
DRM_ERROR("surface not marked as scanout\n");
|
|
/* vmw_user_surface_lookup takes one ref */
|
|
vmw_surface_unreference(&surface);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static struct drm_mode_config_funcs vmw_kms_funcs = {
|
|
.fb_create = vmw_kms_fb_create,
|
|
};
|
|
|
|
int vmw_kms_init(struct vmw_private *dev_priv)
|
|
{
|
|
struct drm_device *dev = dev_priv->dev;
|
|
int ret;
|
|
|
|
drm_mode_config_init(dev);
|
|
dev->mode_config.funcs = &vmw_kms_funcs;
|
|
dev->mode_config.min_width = 1;
|
|
dev->mode_config.min_height = 1;
|
|
/* assumed largest fb size */
|
|
dev->mode_config.max_width = 8192;
|
|
dev->mode_config.max_height = 8192;
|
|
|
|
ret = vmw_kms_init_legacy_display_system(dev_priv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int vmw_kms_close(struct vmw_private *dev_priv)
|
|
{
|
|
/*
|
|
* Docs says we should take the lock before calling this function
|
|
* but since it destroys encoders and our destructor calls
|
|
* drm_encoder_cleanup which takes the lock we deadlock.
|
|
*/
|
|
drm_mode_config_cleanup(dev_priv->dev);
|
|
vmw_kms_close_legacy_display_system(dev_priv);
|
|
return 0;
|
|
}
|
|
|
|
int vmw_kms_cursor_bypass_ioctl(struct drm_device *dev, void *data,
|
|
struct drm_file *file_priv)
|
|
{
|
|
struct drm_vmw_cursor_bypass_arg *arg = data;
|
|
struct vmw_display_unit *du;
|
|
struct drm_mode_object *obj;
|
|
struct drm_crtc *crtc;
|
|
int ret = 0;
|
|
|
|
|
|
mutex_lock(&dev->mode_config.mutex);
|
|
if (arg->flags & DRM_VMW_CURSOR_BYPASS_ALL) {
|
|
|
|
list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
|
|
du = vmw_crtc_to_du(crtc);
|
|
du->hotspot_x = arg->xhot;
|
|
du->hotspot_y = arg->yhot;
|
|
}
|
|
|
|
mutex_unlock(&dev->mode_config.mutex);
|
|
return 0;
|
|
}
|
|
|
|
obj = drm_mode_object_find(dev, arg->crtc_id, DRM_MODE_OBJECT_CRTC);
|
|
if (!obj) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
crtc = obj_to_crtc(obj);
|
|
du = vmw_crtc_to_du(crtc);
|
|
|
|
du->hotspot_x = arg->xhot;
|
|
du->hotspot_y = arg->yhot;
|
|
|
|
out:
|
|
mutex_unlock(&dev->mode_config.mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
void vmw_kms_write_svga(struct vmw_private *vmw_priv,
|
|
unsigned width, unsigned height, unsigned pitch,
|
|
unsigned bbp, unsigned depth)
|
|
{
|
|
if (vmw_priv->capabilities & SVGA_CAP_PITCHLOCK)
|
|
vmw_write(vmw_priv, SVGA_REG_PITCHLOCK, pitch);
|
|
else if (vmw_fifo_have_pitchlock(vmw_priv))
|
|
iowrite32(pitch, vmw_priv->mmio_virt + SVGA_FIFO_PITCHLOCK);
|
|
vmw_write(vmw_priv, SVGA_REG_WIDTH, width);
|
|
vmw_write(vmw_priv, SVGA_REG_HEIGHT, height);
|
|
vmw_write(vmw_priv, SVGA_REG_BITS_PER_PIXEL, bbp);
|
|
vmw_write(vmw_priv, SVGA_REG_DEPTH, depth);
|
|
vmw_write(vmw_priv, SVGA_REG_RED_MASK, 0x00ff0000);
|
|
vmw_write(vmw_priv, SVGA_REG_GREEN_MASK, 0x0000ff00);
|
|
vmw_write(vmw_priv, SVGA_REG_BLUE_MASK, 0x000000ff);
|
|
}
|
|
|
|
int vmw_kms_save_vga(struct vmw_private *vmw_priv)
|
|
{
|
|
vmw_priv->vga_width = vmw_read(vmw_priv, SVGA_REG_WIDTH);
|
|
vmw_priv->vga_height = vmw_read(vmw_priv, SVGA_REG_HEIGHT);
|
|
vmw_priv->vga_bpp = vmw_read(vmw_priv, SVGA_REG_BITS_PER_PIXEL);
|
|
vmw_priv->vga_depth = vmw_read(vmw_priv, SVGA_REG_DEPTH);
|
|
vmw_priv->vga_pseudo = vmw_read(vmw_priv, SVGA_REG_PSEUDOCOLOR);
|
|
vmw_priv->vga_red_mask = vmw_read(vmw_priv, SVGA_REG_RED_MASK);
|
|
vmw_priv->vga_green_mask = vmw_read(vmw_priv, SVGA_REG_GREEN_MASK);
|
|
vmw_priv->vga_blue_mask = vmw_read(vmw_priv, SVGA_REG_BLUE_MASK);
|
|
if (vmw_priv->capabilities & SVGA_CAP_PITCHLOCK)
|
|
vmw_priv->vga_pitchlock =
|
|
vmw_read(vmw_priv, SVGA_REG_PITCHLOCK);
|
|
else if (vmw_fifo_have_pitchlock(vmw_priv))
|
|
vmw_priv->vga_pitchlock =
|
|
ioread32(vmw_priv->mmio_virt + SVGA_FIFO_PITCHLOCK);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int vmw_kms_restore_vga(struct vmw_private *vmw_priv)
|
|
{
|
|
vmw_write(vmw_priv, SVGA_REG_WIDTH, vmw_priv->vga_width);
|
|
vmw_write(vmw_priv, SVGA_REG_HEIGHT, vmw_priv->vga_height);
|
|
vmw_write(vmw_priv, SVGA_REG_BITS_PER_PIXEL, vmw_priv->vga_bpp);
|
|
vmw_write(vmw_priv, SVGA_REG_DEPTH, vmw_priv->vga_depth);
|
|
vmw_write(vmw_priv, SVGA_REG_PSEUDOCOLOR, vmw_priv->vga_pseudo);
|
|
vmw_write(vmw_priv, SVGA_REG_RED_MASK, vmw_priv->vga_red_mask);
|
|
vmw_write(vmw_priv, SVGA_REG_GREEN_MASK, vmw_priv->vga_green_mask);
|
|
vmw_write(vmw_priv, SVGA_REG_BLUE_MASK, vmw_priv->vga_blue_mask);
|
|
if (vmw_priv->capabilities & SVGA_CAP_PITCHLOCK)
|
|
vmw_write(vmw_priv, SVGA_REG_PITCHLOCK,
|
|
vmw_priv->vga_pitchlock);
|
|
else if (vmw_fifo_have_pitchlock(vmw_priv))
|
|
iowrite32(vmw_priv->vga_pitchlock,
|
|
vmw_priv->mmio_virt + SVGA_FIFO_PITCHLOCK);
|
|
|
|
return 0;
|
|
}
|