#include #include #include #include #include #include #include "backend/drm/drm.h" #include "backend/drm/fb.h" #include "backend/drm/iface.h" #include "backend/drm/util.h" static char *atomic_commit_flags_str(uint32_t flags) { const char *const l[] = { (flags & DRM_MODE_PAGE_FLIP_EVENT) ? "PAGE_FLIP_EVENT" : NULL, (flags & DRM_MODE_PAGE_FLIP_ASYNC) ? "PAGE_FLIP_ASYNC" : NULL, (flags & DRM_MODE_ATOMIC_TEST_ONLY) ? "ATOMIC_TEST_ONLY" : NULL, (flags & DRM_MODE_ATOMIC_NONBLOCK) ? "ATOMIC_NONBLOCK" : NULL, (flags & DRM_MODE_ATOMIC_ALLOW_MODESET) ? "ATOMIC_ALLOW_MODESET" : NULL, }; char *buf = NULL; size_t size = 0; FILE *f = open_memstream(&buf, &size); if (f == NULL) { return NULL; } for (size_t i = 0; i < sizeof(l) / sizeof(l[0]); i++) { if (l[i] == NULL) { continue; } if (ftell(f) > 0) { fprintf(f, " | "); } fprintf(f, "%s", l[i]); } if (ftell(f) == 0) { fprintf(f, "none"); } fclose(f); return buf; } struct atomic { drmModeAtomicReq *req; bool failed; }; static void atomic_begin(struct atomic *atom) { *atom = (struct atomic){0}; atom->req = drmModeAtomicAlloc(); if (!atom->req) { wlr_log_errno(WLR_ERROR, "Allocation failed"); atom->failed = true; return; } } static bool atomic_commit(struct atomic *atom, struct wlr_drm_backend *drm, const struct wlr_drm_device_state *state, struct wlr_drm_page_flip *page_flip, uint32_t flags) { if (atom->failed) { return false; } int ret = drmModeAtomicCommit(drm->fd, atom->req, flags, page_flip); if (ret != 0) { enum wlr_log_importance log_level = WLR_ERROR; if (flags & DRM_MODE_ATOMIC_TEST_ONLY) { log_level = WLR_DEBUG; } if (state->connectors_len == 1) { struct wlr_drm_connector *conn = state->connectors[0].connector; wlr_drm_conn_log_errno(conn, log_level, "Atomic commit failed"); } else { wlr_log_errno(log_level, "Atomic commit failed"); } char *flags_str = atomic_commit_flags_str(flags); wlr_log(WLR_DEBUG, "(Atomic commit flags: %s)", flags_str ? flags_str : ""); free(flags_str); return false; } return true; } static void atomic_finish(struct atomic *atom) { drmModeAtomicFree(atom->req); } static void atomic_add(struct atomic *atom, uint32_t id, uint32_t prop, uint64_t val) { if (!atom->failed && drmModeAtomicAddProperty(atom->req, id, prop, val) < 0) { wlr_log_errno(WLR_ERROR, "Failed to add atomic DRM property"); atom->failed = true; } } static bool create_mode_blob(struct wlr_drm_connector *conn, const struct wlr_drm_connector_state *state, uint32_t *blob_id) { if (!state->active) { *blob_id = 0; return true; } if (drmModeCreatePropertyBlob(conn->backend->fd, &state->mode, sizeof(drmModeModeInfo), blob_id)) { wlr_log_errno(WLR_ERROR, "Unable to create mode property blob"); return false; } return true; } static bool create_gamma_lut_blob(struct wlr_drm_backend *drm, size_t size, const uint16_t *lut, uint32_t *blob_id) { if (size == 0) { *blob_id = 0; return true; } struct drm_color_lut *gamma = malloc(size * sizeof(*gamma)); if (gamma == NULL) { wlr_log(WLR_ERROR, "Failed to allocate gamma table"); return false; } const uint16_t *r = lut; const uint16_t *g = lut + size; const uint16_t *b = lut + 2 * size; for (size_t i = 0; i < size; i++) { gamma[i].red = r[i]; gamma[i].green = g[i]; gamma[i].blue = b[i]; } if (drmModeCreatePropertyBlob(drm->fd, gamma, size * sizeof(*gamma), blob_id) != 0) { wlr_log_errno(WLR_ERROR, "Unable to create gamma LUT property blob"); free(gamma); return false; } free(gamma); return true; } bool create_fb_damage_clips_blob(struct wlr_drm_backend *drm, int width, int height, const pixman_region32_t *damage, uint32_t *blob_id) { if (!pixman_region32_not_empty(damage)) { *blob_id = 0; return true; } pixman_region32_t clipped; pixman_region32_init(&clipped); pixman_region32_intersect_rect(&clipped, damage, 0, 0, width, height); int rects_len; const pixman_box32_t *rects = pixman_region32_rectangles(&clipped, &rects_len); int ret = drmModeCreatePropertyBlob(drm->fd, rects, sizeof(*rects) * rects_len, blob_id); pixman_region32_fini(&clipped); if (ret != 0) { wlr_log_errno(WLR_ERROR, "Failed to create FB_DAMAGE_CLIPS property blob"); return false; } return true; } static uint64_t max_bpc_for_format(uint32_t format) { switch (format) { case DRM_FORMAT_XRGB2101010: case DRM_FORMAT_ARGB2101010: case DRM_FORMAT_XBGR2101010: case DRM_FORMAT_ABGR2101010: return 10; case DRM_FORMAT_XBGR16161616F: case DRM_FORMAT_ABGR16161616F: case DRM_FORMAT_XBGR16161616: case DRM_FORMAT_ABGR16161616: return 16; default: return 8; } } static uint64_t pick_max_bpc(struct wlr_drm_connector *conn, struct wlr_drm_fb *fb) { uint32_t format = DRM_FORMAT_INVALID; struct wlr_dmabuf_attributes attribs = {0}; if (wlr_buffer_get_dmabuf(fb->wlr_buf, &attribs)) { format = attribs.format; } uint64_t target_bpc = max_bpc_for_format(format); if (target_bpc < conn->max_bpc_bounds[0]) { target_bpc = conn->max_bpc_bounds[0]; } if (target_bpc > conn->max_bpc_bounds[1]) { target_bpc = conn->max_bpc_bounds[1]; } return target_bpc; } static void destroy_blob(struct wlr_drm_backend *drm, uint32_t id) { if (id == 0) { return; } if (drmModeDestroyPropertyBlob(drm->fd, id) != 0) { wlr_log_errno(WLR_ERROR, "Failed to destroy blob"); } } static void commit_blob(struct wlr_drm_backend *drm, uint32_t *current, uint32_t next) { if (*current == next) { return; } destroy_blob(drm, *current); *current = next; } static void rollback_blob(struct wlr_drm_backend *drm, uint32_t *current, uint32_t next) { if (*current == next) { return; } destroy_blob(drm, next); } bool drm_atomic_connector_prepare(struct wlr_drm_connector_state *state, bool modeset) { struct wlr_drm_connector *conn = state->connector; struct wlr_drm_backend *drm = conn->backend; struct wlr_output *output = &conn->output; struct wlr_drm_crtc *crtc = conn->crtc; uint32_t mode_id = crtc->mode_id; if (modeset) { if (!create_mode_blob(conn, state, &mode_id)) { return false; } } uint32_t gamma_lut = crtc->gamma_lut; if (state->base->committed & WLR_OUTPUT_STATE_GAMMA_LUT) { // Fallback to legacy gamma interface when gamma properties are not // available (can happen on older Intel GPUs that support gamma but not // degamma). if (crtc->props.gamma_lut == 0) { if (!drm_legacy_crtc_set_gamma(drm, crtc, state->base->gamma_lut_size, state->base->gamma_lut)) { return false; } } else { if (!create_gamma_lut_blob(drm, state->base->gamma_lut_size, state->base->gamma_lut, &gamma_lut)) { return false; } } } uint32_t fb_damage_clips = 0; if ((state->base->committed & WLR_OUTPUT_STATE_DAMAGE) && crtc->primary->props.fb_damage_clips != 0) { create_fb_damage_clips_blob(drm, state->primary_fb->wlr_buf->width, state->primary_fb->wlr_buf->height, &state->base->damage, &fb_damage_clips); } bool prev_vrr_enabled = output->adaptive_sync_status == WLR_OUTPUT_ADAPTIVE_SYNC_ENABLED; bool vrr_enabled = prev_vrr_enabled; if ((state->base->committed & WLR_OUTPUT_STATE_ADAPTIVE_SYNC_ENABLED)) { if (state->base->adaptive_sync_enabled && !output->adaptive_sync_supported) { return false; } vrr_enabled = state->base->adaptive_sync_enabled; } state->mode_id = mode_id; state->gamma_lut = gamma_lut; state->fb_damage_clips = fb_damage_clips; state->vrr_enabled = vrr_enabled; return true; } void drm_atomic_connector_apply_commit(struct wlr_drm_connector_state *state) { struct wlr_drm_connector *conn = state->connector; struct wlr_drm_crtc *crtc = conn->crtc; struct wlr_drm_backend *drm = conn->backend; if (!crtc->own_mode_id) { crtc->mode_id = 0; // don't try to delete previous master's blobs } crtc->own_mode_id = true; commit_blob(drm, &crtc->mode_id, state->mode_id); commit_blob(drm, &crtc->gamma_lut, state->gamma_lut); conn->output.adaptive_sync_status = state->vrr_enabled ? WLR_OUTPUT_ADAPTIVE_SYNC_ENABLED : WLR_OUTPUT_ADAPTIVE_SYNC_DISABLED; destroy_blob(drm, state->fb_damage_clips); } void drm_atomic_connector_rollback_commit(struct wlr_drm_connector_state *state) { struct wlr_drm_connector *conn = state->connector; struct wlr_drm_crtc *crtc = conn->crtc; struct wlr_drm_backend *drm = conn->backend; rollback_blob(drm, &crtc->mode_id, state->mode_id); rollback_blob(drm, &crtc->gamma_lut, state->gamma_lut); destroy_blob(drm, state->fb_damage_clips); } static void plane_disable(struct atomic *atom, struct wlr_drm_plane *plane) { uint32_t id = plane->id; const struct wlr_drm_plane_props *props = &plane->props; atomic_add(atom, id, props->fb_id, 0); atomic_add(atom, id, props->crtc_id, 0); } static void set_plane_props(struct atomic *atom, struct wlr_drm_backend *drm, struct wlr_drm_plane *plane, struct wlr_drm_fb *fb, uint32_t crtc_id, int32_t x, int32_t y) { uint32_t id = plane->id; const struct wlr_drm_plane_props *props = &plane->props; if (fb == NULL) { wlr_log(WLR_ERROR, "Failed to acquire FB for plane %"PRIu32, plane->id); atom->failed = true; return; } uint32_t width = fb->wlr_buf->width; uint32_t height = fb->wlr_buf->height; // The src_* properties are in 16.16 fixed point atomic_add(atom, id, props->src_x, 0); atomic_add(atom, id, props->src_y, 0); atomic_add(atom, id, props->src_w, (uint64_t)width << 16); atomic_add(atom, id, props->src_h, (uint64_t)height << 16); atomic_add(atom, id, props->crtc_w, width); atomic_add(atom, id, props->crtc_h, height); atomic_add(atom, id, props->fb_id, fb->id); atomic_add(atom, id, props->crtc_id, crtc_id); atomic_add(atom, id, props->crtc_x, (uint64_t)x); atomic_add(atom, id, props->crtc_y, (uint64_t)y); } static bool supports_cursor_hotspots(const struct wlr_drm_plane* plane) { return plane->props.hotspot_x && plane->props.hotspot_y; } static void atomic_connector_add(struct atomic *atom, const struct wlr_drm_connector_state *state, bool modeset) { struct wlr_drm_connector *conn = state->connector; struct wlr_drm_backend *drm = conn->backend; struct wlr_drm_crtc *crtc = conn->crtc; bool active = state->active; atomic_add(atom, conn->id, conn->props.crtc_id, active ? crtc->id : 0); if (modeset && active && conn->props.link_status != 0) { atomic_add(atom, conn->id, conn->props.link_status, DRM_MODE_LINK_STATUS_GOOD); } if (active && conn->props.content_type != 0) { atomic_add(atom, conn->id, conn->props.content_type, DRM_MODE_CONTENT_TYPE_GRAPHICS); } if (modeset && active && conn->props.max_bpc != 0 && conn->max_bpc_bounds[1] != 0) { atomic_add(atom, conn->id, conn->props.max_bpc, pick_max_bpc(conn, state->primary_fb)); } atomic_add(atom, crtc->id, crtc->props.mode_id, state->mode_id); atomic_add(atom, crtc->id, crtc->props.active, active); if (active) { if (crtc->props.gamma_lut != 0) { atomic_add(atom, crtc->id, crtc->props.gamma_lut, state->gamma_lut); } if (crtc->props.vrr_enabled != 0) { atomic_add(atom, crtc->id, crtc->props.vrr_enabled, state->vrr_enabled); } set_plane_props(atom, drm, crtc->primary, state->primary_fb, crtc->id, 0, 0); if (crtc->primary->props.fb_damage_clips != 0) { atomic_add(atom, crtc->primary->id, crtc->primary->props.fb_damage_clips, state->fb_damage_clips); } if (crtc->cursor) { if (drm_connector_is_cursor_visible(conn)) { set_plane_props(atom, drm, crtc->cursor, state->cursor_fb, crtc->id, conn->cursor_x, conn->cursor_y); if (supports_cursor_hotspots(crtc->cursor)) { atomic_add(atom, crtc->cursor->id, crtc->cursor->props.hotspot_x, conn->cursor_hotspot_x); atomic_add(atom, crtc->cursor->id, crtc->cursor->props.hotspot_y, conn->cursor_hotspot_y); } } else { plane_disable(atom, crtc->cursor); } } } else { plane_disable(atom, crtc->primary); if (crtc->cursor) { plane_disable(atom, crtc->cursor); } } } static bool atomic_device_commit(struct wlr_drm_backend *drm, const struct wlr_drm_device_state *state, struct wlr_drm_page_flip *page_flip, uint32_t flags, bool test_only) { bool ok = false; for (size_t i = 0; i < state->connectors_len; i++) { if (!drm_atomic_connector_prepare(&state->connectors[i], state->modeset)) { goto out; } } struct atomic atom; atomic_begin(&atom); for (size_t i = 0; i < state->connectors_len; i++) { atomic_connector_add(&atom, &state->connectors[i], state->modeset); } if (test_only) { flags |= DRM_MODE_ATOMIC_TEST_ONLY; } if (state->modeset) { flags |= DRM_MODE_ATOMIC_ALLOW_MODESET; } if (!test_only && state->nonblock) { flags |= DRM_MODE_ATOMIC_NONBLOCK; } ok = atomic_commit(&atom, drm, state, page_flip, flags); atomic_finish(&atom); out: for (size_t i = 0; i < state->connectors_len; i++) { struct wlr_drm_connector_state *conn_state = &state->connectors[i]; if (ok && !test_only) { drm_atomic_connector_apply_commit(conn_state); } else { drm_atomic_connector_rollback_commit(conn_state); } } return ok; } bool drm_atomic_reset(struct wlr_drm_backend *drm) { struct atomic atom; atomic_begin(&atom); for (size_t i = 0; i < drm->num_crtcs; i++) { struct wlr_drm_crtc *crtc = &drm->crtcs[i]; atomic_add(&atom, crtc->id, crtc->props.mode_id, 0); atomic_add(&atom, crtc->id, crtc->props.active, 0); } struct wlr_drm_connector *conn; wl_list_for_each(conn, &drm->connectors, link) { atomic_add(&atom, conn->id, conn->props.crtc_id, 0); } for (size_t i = 0; i < drm->num_planes; i++) { plane_disable(&atom, &drm->planes[i]); } uint32_t flags = DRM_MODE_ATOMIC_ALLOW_MODESET; bool ok = atomic_commit(&atom, drm, NULL, NULL, flags); atomic_finish(&atom); return ok; } const struct wlr_drm_interface atomic_iface = { .commit = atomic_device_commit, .reset = drm_atomic_reset, };