/* * drivers/video/tegra/dc/window.c * * Copyright (C) 2010 Google, Inc. * * Copyright (c) 2010-2013, NVIDIA CORPORATION, All rights reserved. * * This software is licensed under the terms of the GNU General Public * License version 2, as published by the Free Software Foundation, and * may be copied, distributed, and modified under those terms. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * */ #include #include #include #include #include #include "dc_reg.h" #include "dc_config.h" #include "dc_priv.h" static int no_vsync; module_param_named(no_vsync, no_vsync, int, S_IRUGO | S_IWUSR); static bool tegra_dc_windows_are_clean(struct tegra_dc_win *windows[], int n) { int i; struct tegra_dc *dc = windows[0]->dc; mutex_lock(&dc->lock); for (i = 0; i < n; i++) { if (windows[i]->dirty) { mutex_unlock(&dc->lock); return false; } } mutex_unlock(&dc->lock); return true; } int tegra_dc_config_frame_end_intr(struct tegra_dc *dc, bool enable) { mutex_lock(&dc->lock); tegra_dc_io_start(dc); if (enable) { atomic_inc(&dc->frame_end_ref); tegra_dc_unmask_interrupt(dc, FRAME_END_INT); } else if (!atomic_dec_return(&dc->frame_end_ref)) tegra_dc_mask_interrupt(dc, FRAME_END_INT); tegra_dc_io_end(dc); mutex_unlock(&dc->lock); return 0; } static int get_topmost_window(u32 *depths, unsigned long *wins, int win_num) { int idx, best = -1; for_each_set_bit(idx, wins, win_num) { if (best == -1 || depths[idx] < depths[best]) best = idx; } clear_bit(best, wins); return best; } static u32 blend_topwin(u32 flags) { if (flags & TEGRA_WIN_FLAG_BLEND_COVERAGE) return BLEND(NOKEY, ALPHA, 0xff, 0xff); else if (flags & TEGRA_WIN_FLAG_BLEND_PREMULT) return BLEND(NOKEY, PREMULT, 0xff, 0xff); else return BLEND(NOKEY, FIX, 0xff, 0xff); } static u32 blend_2win(int idx, unsigned long behind_mask, u32 *flags, int xy, int win_num) { int other; for (other = 0; other < win_num; other++) { if (other != idx && (xy-- == 0)) break; } if (BIT(other) & behind_mask) return blend_topwin(flags[idx]); else if (flags[other]) return BLEND(NOKEY, DEPENDANT, 0x00, 0x00); else return BLEND(NOKEY, FIX, 0x00, 0x00); } static u32 blend_3win(int idx, unsigned long behind_mask, u32 *flags, int win_num) { unsigned long infront_mask; int first, second; infront_mask = ~(behind_mask | BIT(idx)); infront_mask &= (BIT(win_num) - 1); first = ffs(infront_mask) - 1; if (win_num != 3) { if (!infront_mask) return blend_topwin(flags[idx]); else if (behind_mask && first != -1 && flags[first]) return BLEND(NOKEY, DEPENDANT, 0x00, 0x00); else return BLEND(NOKEY, FIX, 0x0, 0x0); } else { if (!infront_mask) { return blend_topwin(flags[idx]); } else if (behind_mask) { /* If the first (top) window is not opaque, check if * the current (middle) window is not opaque. If yes * then we need to enable appropriate blending mode * otherwise current window is dependent on the first * window. If the first window is opaque then the * current window has no contribution and we set its * weight to 0. */ if (flags[first]) { if (flags[idx]) return blend_topwin(flags[idx]); else return BLEND(NOKEY, DEPENDANT, 0x00, 0x00); } else { return BLEND(NOKEY, FIX, 0x00, 0x00); } } else { infront_mask &= ~(BIT(first)); second = ffs(infront_mask) - 1; /* If both windows above the bottom window are not * opaque then the bottom window is dependent otherwise * the bottom window has no contribution and we set * its weight to 0. */ if (flags[first] && flags[second]) return BLEND(NOKEY, DEPENDANT, 0x00, 0x00); else return BLEND(NOKEY, FIX, 0x00, 0x00); } } } static void tegra_dc_blend_parallel(struct tegra_dc *dc, struct tegra_dc_blend *blend) { int win_num = dc->gen1_blend_num; unsigned long mask = BIT(win_num) - 1; tegra_dc_io_start(dc); while (mask) { int idx = get_topmost_window(blend->z, &mask, win_num); tegra_dc_writel(dc, WINDOW_A_SELECT << idx, DC_CMD_DISPLAY_WINDOW_HEADER); tegra_dc_writel(dc, BLEND(NOKEY, FIX, 0xff, 0xff), DC_WIN_BLEND_NOKEY); tegra_dc_writel(dc, blend_2win(idx, mask, blend->flags, 0, win_num), DC_WIN_BLEND_2WIN_X); tegra_dc_writel(dc, blend_2win(idx, mask, blend->flags, 1, win_num), DC_WIN_BLEND_2WIN_Y); tegra_dc_writel(dc, blend_3win(idx, mask, blend->flags, win_num), DC_WIN_BLEND_3WIN_XY); } tegra_dc_io_end(dc); } static void tegra_dc_blend_sequential(struct tegra_dc *dc, struct tegra_dc_blend *blend) { int idx; unsigned long mask = dc->valid_windows; tegra_dc_io_start(dc); for_each_set_bit(idx, &mask, dc->n_windows) { if (!tegra_dc_feature_is_gen2_blender(dc, idx)) continue; tegra_dc_writel(dc, WINDOW_A_SELECT << idx, DC_CMD_DISPLAY_WINDOW_HEADER); if (blend->flags[idx] & TEGRA_WIN_FLAG_BLEND_COVERAGE) { #if defined(CONFIG_TEGRA_DC_BLENDER_DEPTH) tegra_dc_writel(dc, WIN_K1(blend->alpha[idx]) | WIN_K2(0xff) | WIN_BLEND_ENABLE | WIN_DEPTH(dc->blend.z[idx]), DC_WINBUF_BLEND_LAYER_CONTROL); #else tegra_dc_writel(dc, WIN_K1(blend->alpha[idx]) | WIN_K2(0xff) | WIN_BLEND_ENABLE, DC_WINBUF_BLEND_LAYER_CONTROL); #endif tegra_dc_writel(dc, WIN_BLEND_FACT_SRC_COLOR_MATCH_SEL_K1_TIMES_SRC | WIN_BLEND_FACT_DST_COLOR_MATCH_SEL_NEG_K1_TIMES_SRC | WIN_BLEND_FACT_SRC_ALPHA_MATCH_SEL_K2 | WIN_BLEND_FACT_DST_ALPHA_MATCH_SEL_ZERO, DC_WINBUF_BLEND_MATCH_SELECT); tegra_dc_writel(dc, WIN_BLEND_FACT_SRC_COLOR_NOMATCH_SEL_K1_TIMES_SRC | WIN_BLEND_FACT_DST_COLOR_NOMATCH_SEL_NEG_K1_TIMES_SRC | WIN_BLEND_FACT_SRC_ALPHA_NOMATCH_SEL_K2 | WIN_BLEND_FACT_DST_ALPHA_NOMATCH_SEL_ZERO, DC_WINBUF_BLEND_NOMATCH_SELECT); tegra_dc_writel(dc, WIN_ALPHA_1BIT_WEIGHT0(0) | WIN_ALPHA_1BIT_WEIGHT1(0xff), DC_WINBUF_BLEND_ALPHA_1BIT); } else if (blend->flags[idx] & TEGRA_WIN_FLAG_BLEND_PREMULT) { #if defined(CONFIG_TEGRA_DC_BLENDER_DEPTH) tegra_dc_writel(dc, WIN_K1(blend->alpha[idx]) | WIN_K2(0xff) | WIN_BLEND_ENABLE | WIN_DEPTH(dc->blend.z[idx]), DC_WINBUF_BLEND_LAYER_CONTROL); #else tegra_dc_writel(dc, WIN_K1(blend->alpha[idx]) | WIN_K2(0xff) | WIN_BLEND_ENABLE, DC_WINBUF_BLEND_LAYER_CONTROL); #endif tegra_dc_writel(dc, WIN_BLEND_FACT_SRC_COLOR_MATCH_SEL_K1 | WIN_BLEND_FACT_DST_COLOR_MATCH_SEL_NEG_K1_TIMES_SRC | WIN_BLEND_FACT_SRC_ALPHA_MATCH_SEL_K2 | WIN_BLEND_FACT_DST_ALPHA_MATCH_SEL_ZERO, DC_WINBUF_BLEND_MATCH_SELECT); tegra_dc_writel(dc, WIN_BLEND_FACT_SRC_COLOR_NOMATCH_SEL_NEG_K1_TIMES_DST | WIN_BLEND_FACT_DST_COLOR_NOMATCH_SEL_K1 | WIN_BLEND_FACT_SRC_ALPHA_NOMATCH_SEL_K2 | WIN_BLEND_FACT_DST_ALPHA_NOMATCH_SEL_ZERO, DC_WINBUF_BLEND_NOMATCH_SELECT); tegra_dc_writel(dc, WIN_ALPHA_1BIT_WEIGHT0(0) | WIN_ALPHA_1BIT_WEIGHT1(0xff), DC_WINBUF_BLEND_ALPHA_1BIT); } else { #if defined(CONFIG_TEGRA_DC_BLENDER_DEPTH) tegra_dc_writel(dc, WIN_BLEND_BYPASS | WIN_DEPTH(dc->blend.z[idx]), DC_WINBUF_BLEND_LAYER_CONTROL); #else tegra_dc_writel(dc, WIN_BLEND_BYPASS, DC_WINBUF_BLEND_LAYER_CONTROL); #endif } } tegra_dc_io_end(dc); } /* does not support syncing windows on multiple dcs in one call */ int tegra_dc_sync_windows(struct tegra_dc_win *windows[], int n) { int ret; if (n < 1 || n > windows[0]->dc->n_windows) return -EINVAL; if (!windows[0]->dc->enabled) return -EFAULT; trace_sync_windows(windows[0]->dc); ret = wait_event_interruptible_timeout(windows[0]->dc->wq, tegra_dc_windows_are_clean(windows, n), HZ); /* tegra_dc_hold_dc_out() called in update_windows */ tegra_dc_release_dc_out(windows[0]->dc); return ret; } EXPORT_SYMBOL(tegra_dc_sync_windows); static inline u32 compute_dda_inc(fixed20_12 in, unsigned out_int, bool v, unsigned bpp) { /* * min(round((prescaled_size_in_pixels - 1) * 0x1000 / * (post_scaled_size_in_pixels - 1)), MAX) * Where the value of MAX is as follows: * For V_DDA_INCREMENT: 15.0 (0xF000) * For H_DDA_INCREMENT: 4.0 (0x4000) for 4 Bytes/pix formats. * 8.0 (0x8000) for 2 Bytes/pix formats. */ fixed20_12 out = dfixed_init(out_int); u32 dda_inc; int max; if (v) { max = 15; } else { switch (bpp) { default: WARN_ON_ONCE(1); /* fallthrough */ case 4: max = 4; break; case 2: max = 8; break; } } out.full = max_t(u32, out.full - dfixed_const(1), dfixed_const(1)); in.full -= dfixed_const(1); dda_inc = dfixed_div(in, out); dda_inc = min_t(u32, dda_inc, dfixed_const(max)); return dda_inc; } static inline u32 compute_initial_dda(fixed20_12 in) { return dfixed_frac(in); } static inline void tegra_dc_update_scaling(struct tegra_dc *dc, struct tegra_dc_win *win, unsigned bpp, unsigned bpp_bw, bool scan_column) { u32 h_dda, h_dda_init; u32 v_dda, v_dda_init; h_dda_init = compute_initial_dda(win->x); v_dda_init = compute_initial_dda(win->y); if (scan_column) { if (tegra_dc_feature_has_interlace(dc, win->idx) && (dc->mode.vmode == FB_VMODE_INTERLACED)) { if (WIN_IS_INTERLACE(win)) tegra_dc_writel(dc, V_PRESCALED_SIZE(dfixed_trunc(win->w)) | H_PRESCALED_SIZE(dfixed_trunc(win->h) * bpp), DC_WIN_PRESCALED_SIZE); else tegra_dc_writel(dc, V_PRESCALED_SIZE(dfixed_trunc(win->w)) | H_PRESCALED_SIZE(dfixed_trunc(win->h) * bpp), DC_WIN_PRESCALED_SIZE); } else { tegra_dc_writel(dc, V_PRESCALED_SIZE(dfixed_trunc(win->w)) | H_PRESCALED_SIZE(dfixed_trunc(win->h) * bpp), DC_WIN_PRESCALED_SIZE); } tegra_dc_writel(dc, v_dda_init, DC_WIN_H_INITIAL_DDA); tegra_dc_writel(dc, h_dda_init, DC_WIN_V_INITIAL_DDA); h_dda = compute_dda_inc(win->h, win->out_w, false, bpp_bw); v_dda = compute_dda_inc(win->w, win->out_h, true, bpp_bw); } else { if (tegra_dc_feature_has_interlace(dc, win->idx) && (dc->mode.vmode == FB_VMODE_INTERLACED)) { if (WIN_IS_INTERLACE(win)) tegra_dc_writel(dc, V_PRESCALED_SIZE(dfixed_trunc(win->h) >> 1) | H_PRESCALED_SIZE(dfixed_trunc(win->w) * bpp), DC_WIN_PRESCALED_SIZE); else tegra_dc_writel(dc, V_PRESCALED_SIZE(dfixed_trunc(win->h)) | H_PRESCALED_SIZE(dfixed_trunc(win->w) * bpp), DC_WIN_PRESCALED_SIZE); } else { tegra_dc_writel(dc, V_PRESCALED_SIZE(dfixed_trunc(win->h)) | H_PRESCALED_SIZE(dfixed_trunc(win->w) * bpp), DC_WIN_PRESCALED_SIZE); } tegra_dc_writel(dc, h_dda_init, DC_WIN_H_INITIAL_DDA); tegra_dc_writel(dc, v_dda_init, DC_WIN_V_INITIAL_DDA); h_dda = compute_dda_inc(win->w, win->out_w, false, bpp_bw); v_dda = compute_dda_inc(win->h, win->out_h, true, bpp_bw); } if (tegra_dc_feature_has_interlace(dc, win->idx) && (dc->mode.vmode == FB_VMODE_INTERLACED)) { if (WIN_IS_INTERLACE(win)) tegra_dc_writel(dc, V_DDA_INC(v_dda) | H_DDA_INC(h_dda), DC_WIN_DDA_INCREMENT); else tegra_dc_writel(dc, V_DDA_INC(v_dda << 1) | H_DDA_INC(h_dda), DC_WIN_DDA_INCREMENT); } else { tegra_dc_writel(dc, V_DDA_INC(v_dda) | H_DDA_INC(h_dda), DC_WIN_DDA_INCREMENT); } } /* Does not support updating windows on multiple dcs in one call. * Requires a matching sync_windows to avoid leaking ref-count on clocks. */ int tegra_dc_update_windows(struct tegra_dc_win *windows[], int n) { struct tegra_dc *dc; unsigned long update_mask = GENERAL_ACT_REQ; unsigned long win_options; bool update_blend_par = false; bool update_blend_seq = false; int i; dc = windows[0]->dc; trace_update_windows(dc); if (dc->out->flags & TEGRA_DC_OUT_ONE_SHOT_MODE) { /* Acquire one_shot_lock to avoid race condition between * cancellation of old delayed work and schedule of new * delayed work. */ mutex_lock(&dc->one_shot_lock); cancel_delayed_work_sync(&dc->one_shot_work); } mutex_lock(&dc->lock); if (!dc->enabled) { mutex_unlock(&dc->lock); if (dc->out->flags & TEGRA_DC_OUT_ONE_SHOT_MODE) mutex_unlock(&dc->one_shot_lock); return -EFAULT; } tegra_dc_hold_dc_out(dc); if (no_vsync) tegra_dc_writel(dc, WRITE_MUX_ACTIVE | READ_MUX_ACTIVE, DC_CMD_STATE_ACCESS); else tegra_dc_writel(dc, WRITE_MUX_ASSEMBLY | READ_MUX_ASSEMBLY, DC_CMD_STATE_ACCESS); for (i = 0; i < n; i++) { struct tegra_dc_win *win = windows[i]; struct tegra_dc_win *dc_win = tegra_dc_get_window(dc, win->idx); bool scan_column = 0; fixed20_12 h_offset, v_offset; bool invert_h = (win->flags & TEGRA_WIN_FLAG_INVERT_H) != 0; bool invert_v = (win->flags & TEGRA_WIN_FLAG_INVERT_V) != 0; bool yuv = tegra_dc_is_yuv(win->fmt); bool yuvp = tegra_dc_is_yuv_planar(win->fmt); bool yuvsp = tegra_dc_is_yuv_semi_planar(win->fmt); unsigned bpp = tegra_dc_fmt_bpp(win->fmt) / 8; /* Bytes per pixel of bandwidth, used for dda_inc calculation */ unsigned bpp_bw = bpp * ((yuvp || yuvsp) ? 2 : 1); const bool filter_h = win_use_h_filter(dc, win); const bool filter_v = win_use_v_filter(dc, win); if (dc->out->type == TEGRA_DC_OUT_HDMI && win->phys_addr == dc->fb->phys_start) { dev_info(&dc->ndev->dev, "Ignore update windows for HDMI FB.\n"); continue; } #if defined(CONFIG_TEGRA_DC_SCAN_COLUMN) scan_column = (win->flags & TEGRA_WIN_FLAG_SCAN_COLUMN); #endif /* Update blender */ if ((win->z != dc->blend.z[win->idx]) || ((win->flags & TEGRA_WIN_BLEND_FLAGS_MASK) != dc->blend.flags[win->idx])) { dc->blend.z[win->idx] = win->z; dc->blend.flags[win->idx] = win->flags & TEGRA_WIN_BLEND_FLAGS_MASK; if (tegra_dc_feature_is_gen2_blender(dc, win->idx)) update_blend_seq = true; else update_blend_par = true; } if ((win->flags & TEGRA_WIN_BLEND_FLAGS_MASK) != dc->blend.flags[win->idx]) { dc->blend.flags[win->idx] = win->flags & TEGRA_WIN_BLEND_FLAGS_MASK; if (tegra_dc_feature_is_gen2_blender(dc, win->idx)) update_blend_seq = true; else update_blend_par = true; } tegra_dc_writel(dc, WINDOW_A_SELECT << win->idx, DC_CMD_DISPLAY_WINDOW_HEADER); if (!no_vsync) update_mask |= WIN_A_ACT_REQ << win->idx; if (!WIN_IS_ENABLED(win)) { dc_win->dirty = no_vsync ? 0 : 1; tegra_dc_writel(dc, 0, DC_WIN_WIN_OPTIONS); continue; } tegra_dc_writel(dc, tegra_dc_fmt(win->fmt), DC_WIN_COLOR_DEPTH); tegra_dc_writel(dc, tegra_dc_fmt_byteorder(win->fmt), DC_WIN_BYTE_SWAP); tegra_dc_writel(dc, V_POSITION(win->out_y) | H_POSITION(win->out_x), DC_WIN_POSITION); if (tegra_dc_feature_has_interlace(dc, win->idx) && (dc->mode.vmode == FB_VMODE_INTERLACED)) { tegra_dc_writel(dc, V_SIZE((win->out_h) >> 1) | H_SIZE(win->out_w), DC_WIN_SIZE); } else { tegra_dc_writel(dc, V_SIZE(win->out_h) | H_SIZE(win->out_w), DC_WIN_SIZE); } /* Check scan_column flag to set window size and scaling. */ win_options = WIN_ENABLE; if (scan_column) { win_options |= WIN_SCAN_COLUMN; win_options |= H_FILTER_ENABLE(filter_v); win_options |= V_FILTER_ENABLE(filter_h); } else { win_options |= H_FILTER_ENABLE(filter_h); win_options |= V_FILTER_ENABLE(filter_v); } /* Update scaling registers if window supports scaling. */ if (likely(tegra_dc_feature_has_scaling(dc, win->idx))) tegra_dc_update_scaling(dc, win, bpp, bpp_bw, scan_column); tegra_dc_writel(dc, tegra_dc_reg_l32(win->phys_addr), DC_WINBUF_START_ADDR); if (tegra_dc_feature_has_hi_regs(dc, win->idx, HAS_HI_REGS)) tegra_dc_writel(dc, tegra_dc_reg_h32(win->phys_addr), DC_WINBUF_START_ADDR_HI); if (!yuvp && !yuvsp) { tegra_dc_writel(dc, win->stride, DC_WIN_LINE_STRIDE); } else if (yuvp) { tegra_dc_writel(dc, tegra_dc_reg_l32(win->phys_addr_u), DC_WINBUF_START_ADDR_U); tegra_dc_writel(dc, tegra_dc_reg_l32(win->phys_addr_v), DC_WINBUF_START_ADDR_V); if (tegra_dc_feature_has_hi_regs(dc, win->idx, HAS_HI_UV_REGS)) { tegra_dc_writel(dc, tegra_dc_reg_h32(win->phys_addr_u), DC_WINBUF_START_ADDR_HI_U); tegra_dc_writel(dc, tegra_dc_reg_h32(win->phys_addr_v), DC_WINBUF_START_ADDR_HI_V); } tegra_dc_writel(dc, LINE_STRIDE(win->stride) | UV_LINE_STRIDE(win->stride_uv), DC_WIN_LINE_STRIDE); } else { tegra_dc_writel(dc, tegra_dc_reg_l32(win->phys_addr_u), DC_WINBUF_START_ADDR_U); if (tegra_dc_feature_has_hi_regs(dc, win->idx, HAS_HI_UV_REGS)) tegra_dc_writel(dc, tegra_dc_reg_h32(win->phys_addr_u), DC_WINBUF_START_ADDR_HI_U); tegra_dc_writel(dc, LINE_STRIDE(win->stride) | UV_LINE_STRIDE(win->stride_uv), DC_WIN_LINE_STRIDE); } if (invert_h) { h_offset.full = win->x.full + win->w.full; h_offset.full = dfixed_floor(h_offset) * bpp; h_offset.full -= dfixed_const(1); } else { h_offset.full = dfixed_floor(win->x) * bpp; } v_offset = win->y; if (invert_v) v_offset.full += win->h.full - dfixed_const(1); tegra_dc_writel(dc, dfixed_trunc(h_offset), DC_WINBUF_ADDR_H_OFFSET); tegra_dc_writel(dc, dfixed_trunc(v_offset), DC_WINBUF_ADDR_V_OFFSET); #if defined(CONFIG_TEGRA_DC_INTERLACE) if ((dc->mode.vmode == FB_VMODE_INTERLACED) && WIN_IS_FB(win)) { if (!WIN_IS_INTERLACE(win)) win->phys_addr2 = win->phys_addr; } if (tegra_dc_feature_has_interlace(dc, win->idx) && (dc->mode.vmode == FB_VMODE_INTERLACED)) { tegra_dc_writel(dc, tegra_dc_reg_l32(win->phys_addr2), DC_WINBUF_START_ADDR_FIELD2); if (tegra_dc_feature_has_hi_regs(dc, win->idx, HAS_HI_FIELD2_REGS)) tegra_dc_writel(dc, tegra_dc_reg_h32(win->phys_addr2), DC_WINBUF_START_ADDR_FIELD2_HI); if (yuvp) { tegra_dc_writel(dc, tegra_dc_reg_l32(win->phys_addr_u2), DC_WINBUF_START_ADDR_FIELD2_U); tegra_dc_writel(dc, tegra_dc_reg_l32(win->phys_addr_v2), DC_WINBUF_START_ADDR_FIELD2_V); if (tegra_dc_feature_has_hi_regs(dc, win->idx, HAS_HI_FIELD2_UV_REGS)) { tegra_dc_writel(dc, tegra_dc_reg_h32(win->phys_addr_u2), DC_WINBUF_START_ADDR_FIELD2_HI_U); tegra_dc_writel(dc, tegra_dc_reg_h32(win->phys_addr_v2), DC_WINBUF_START_ADDR_FIELD2_HI_V); } } else if (yuvsp) { tegra_dc_writel(dc, tegra_dc_reg_l32(win->phys_addr_u2), DC_WINBUF_START_ADDR_FIELD2_U); if (tegra_dc_feature_has_hi_regs(dc, win->idx, HAS_HI_FIELD2_UV_REGS)) tegra_dc_writel(dc, tegra_dc_reg_h32(win->phys_addr_u2), DC_WINBUF_START_ADDR_FIELD2_HI_U); } tegra_dc_writel(dc, dfixed_trunc(h_offset), DC_WINBUF_ADDR_H_OFFSET_FIELD2); if (WIN_IS_INTERLACE(win)) { tegra_dc_writel(dc, dfixed_trunc(v_offset), DC_WINBUF_ADDR_V_OFFSET_FIELD2); } else { v_offset.full += dfixed_const(1); tegra_dc_writel(dc, dfixed_trunc(v_offset), DC_WINBUF_ADDR_V_OFFSET_FIELD2); } } #endif if (tegra_dc_feature_has_tiling(dc, win->idx)) { if (WIN_IS_TILED(win)) tegra_dc_writel(dc, DC_WIN_BUFFER_ADDR_MODE_TILE | DC_WIN_BUFFER_ADDR_MODE_TILE_UV, DC_WIN_BUFFER_ADDR_MODE); else tegra_dc_writel(dc, DC_WIN_BUFFER_ADDR_MODE_LINEAR | DC_WIN_BUFFER_ADDR_MODE_LINEAR_UV, DC_WIN_BUFFER_ADDR_MODE); } #if defined(CONFIG_TEGRA_DC_BLOCK_LINEAR) if (tegra_dc_feature_has_blocklinear(dc, win->idx) || tegra_dc_feature_has_tiling(dc, win->idx)) { if (WIN_IS_BLOCKLINEAR(win)) { tegra_dc_writel(dc, DC_WIN_BUFFER_SURFACE_BL_16B2 | (win->block_height_log2 << BLOCK_HEIGHT_SHIFT), DC_WIN_BUFFER_SURFACE_KIND); } else if (WIN_IS_TILED(win)) { tegra_dc_writel(dc, DC_WIN_BUFFER_SURFACE_TILED, DC_WIN_BUFFER_SURFACE_KIND); } else { tegra_dc_writel(dc, DC_WIN_BUFFER_SURFACE_PITCH, DC_WIN_BUFFER_SURFACE_KIND); } } #endif if (yuv) win_options |= CSC_ENABLE; else if (tegra_dc_fmt_bpp(win->fmt) < 24) win_options |= COLOR_EXPAND; /* * For gen2 blending, change in the global alpha needs rewrite * to blending regs. */ if ((dc->blend.alpha[win->idx] != win->global_alpha) && (tegra_dc_feature_is_gen2_blender(dc, win->idx))) update_blend_seq = true; /* Cache the global alpha of each window here. It is necessary * for in-order blending settings. */ dc->blend.alpha[win->idx] = win->global_alpha; if (!tegra_dc_feature_is_gen2_blender(dc, win->idx)) { /* Update global alpha if blender is gen1. */ if (win->global_alpha == 255) { tegra_dc_writel(dc, 0, DC_WIN_GLOBAL_ALPHA); } else { tegra_dc_writel(dc, GLOBAL_ALPHA_ENABLE | win->global_alpha, DC_WIN_GLOBAL_ALPHA); win_options |= CP_ENABLE; } if (win->flags & TEGRA_WIN_FLAG_BLEND_COVERAGE) { tegra_dc_writel(dc, BLEND(NOKEY, ALPHA, 0xff, 0xff), DC_WIN_BLEND_1WIN); } else { /* global_alpha is 255 if not in use */ tegra_dc_writel(dc, BLEND(NOKEY, FIX, win->global_alpha, win->global_alpha), DC_WIN_BLEND_1WIN); } } if (win->ppflags & TEGRA_WIN_PPFLAG_CP_ENABLE) win_options |= CP_ENABLE; win_options |= H_DIRECTION_DECREMENT(invert_h); win_options |= V_DIRECTION_DECREMENT(invert_v); #if defined(CONFIG_TEGRA_DC_INTERLACE) if (tegra_dc_feature_has_interlace(dc, win->idx)) { if (dc->mode.vmode == FB_VMODE_INTERLACED) win_options |= INTERLACE_ENABLE; } #endif tegra_dc_writel(dc, win_options, DC_WIN_WIN_OPTIONS); dc_win->dirty = no_vsync ? 0 : 1; trace_window_update(dc, win); } if (update_blend_par || update_blend_seq) { if (update_blend_par) tegra_dc_blend_parallel(dc, &dc->blend); if (update_blend_seq) tegra_dc_blend_sequential(dc, &dc->blend); for (i = 0; i < dc->n_windows; i++) { if (!no_vsync) dc->windows[i].dirty = 1; update_mask |= WIN_A_ACT_REQ << i; } } tegra_dc_set_dynamic_emc(windows, n); tegra_dc_writel(dc, update_mask << 8, DC_CMD_STATE_CONTROL); if (!no_vsync) { set_bit(V_BLANK_FLIP, &dc->vblank_ref_count); tegra_dc_unmask_interrupt(dc, FRAME_END_INT | V_BLANK_INT | ALL_UF_INT()); set_bit(V_PULSE2_FLIP, &dc->vpulse2_ref_count); tegra_dc_unmask_interrupt(dc, V_PULSE2_INT); } else { clear_bit(V_BLANK_FLIP, &dc->vblank_ref_count); tegra_dc_mask_interrupt(dc, V_BLANK_INT | ALL_UF_INT()); clear_bit(V_PULSE2_FLIP, &dc->vpulse2_ref_count); tegra_dc_mask_interrupt(dc, V_PULSE2_INT); if (!atomic_read(&dc->frame_end_ref)) tegra_dc_mask_interrupt(dc, FRAME_END_INT); } if (dc->out->flags & TEGRA_DC_OUT_ONE_SHOT_MODE) { schedule_delayed_work(&dc->one_shot_work, msecs_to_jiffies(dc->one_shot_delay_ms)); } dc->crc_pending = true; /* update EMC clock if calculated bandwidth has changed */ tegra_dc_program_bandwidth(dc, false); if (dc->out->flags & TEGRA_DC_OUT_ONE_SHOT_MODE) update_mask |= NC_HOST_TRIG; tegra_dc_writel(dc, update_mask, DC_CMD_STATE_CONTROL); /* tegra_dc_release_dc_out() is called in tegra_dc_sync_windows() */ mutex_unlock(&dc->lock); if (dc->out->flags & TEGRA_DC_OUT_ONE_SHOT_MODE) mutex_unlock(&dc->one_shot_lock); return 0; } EXPORT_SYMBOL(tegra_dc_update_windows); void tegra_dc_trigger_windows(struct tegra_dc *dc) { u32 val, i; u32 completed = 0; u32 dirty = 0; bool interlace_done = true; #if defined(CONFIG_TEGRA_DC_INTERLACE) if (dc->mode.vmode == FB_VMODE_INTERLACED) { val = tegra_dc_readl(dc, DC_DISP_INTERLACE_CONTROL); interlace_done = (val & INTERLACE_MODE_ENABLE) && (val & INTERLACE_STATUS_FIELD_2); } #endif val = tegra_dc_readl(dc, DC_CMD_STATE_CONTROL); for (i = 0; i < dc->n_windows; i++) { if (!(val & (WIN_A_ACT_REQ << i)) && interlace_done) { dc->windows[i].dirty = 0; completed = 1; } else dirty = 1; } if (!dirty) { if (!(dc->out->flags & TEGRA_DC_OUT_ONE_SHOT_MODE) && !atomic_read(&dc->frame_end_ref)) tegra_dc_mask_interrupt(dc, FRAME_END_INT); } if (completed) wake_up(&dc->wq); }