2 * Copyright 2016 Advanced Micro Devices, Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
26 #include <linux/slab.h>
28 #include "dm_services.h"
32 #include "include/irq_service_interface.h"
33 #include "dcn20/dcn20_resource.h"
35 #include "dcn10/dcn10_hubp.h"
36 #include "dcn10/dcn10_ipp.h"
37 #include "dcn20_hubbub.h"
38 #include "dcn20_mpc.h"
39 #include "dcn20_hubp.h"
40 #include "irq/dcn20/irq_service_dcn20.h"
41 #include "dcn20_dpp.h"
42 #include "dcn20_optc.h"
43 #include "dcn20_hwseq.h"
44 #include "dce110/dce110_hw_sequencer.h"
45 #include "dcn10/dcn10_resource.h"
46 #include "dcn20_opp.h"
48 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
49 #include "dcn20_dsc.h"
52 #include "dcn20_link_encoder.h"
53 #include "dcn20_stream_encoder.h"
54 #include "dce/dce_clock_source.h"
55 #include "dce/dce_audio.h"
56 #include "dce/dce_hwseq.h"
57 #include "virtual/virtual_stream_encoder.h"
58 #include "dce110/dce110_resource.h"
59 #include "dml/display_mode_vba.h"
60 #include "dcn20_dccg.h"
61 #include "dcn20_vmid.h"
63 #include "navi10_ip_offset.h"
65 #include "dcn/dcn_2_0_0_offset.h"
66 #include "dcn/dcn_2_0_0_sh_mask.h"
68 #include "nbio/nbio_2_3_offset.h"
70 #include "dcn20/dcn20_dwb.h"
71 #include "dcn20/dcn20_mmhubbub.h"
73 #include "mmhub/mmhub_2_0_0_offset.h"
74 #include "mmhub/mmhub_2_0_0_sh_mask.h"
76 #include "reg_helper.h"
77 #include "dce/dce_abm.h"
78 #include "dce/dce_dmcu.h"
79 #include "dce/dce_aux.h"
80 #include "dce/dce_i2c.h"
81 #include "vm_helper.h"
83 #include "amdgpu_socbb.h"
85 #define SOC_BOUNDING_BOX_VALID true
86 #define DC_LOGGER_INIT(logger)
88 struct _vcs_dpi_ip_params_st dcn2_0_ip = {
92 .gpuvm_max_page_table_levels = 4,
93 .hostvm_max_page_table_levels = 4,
94 .hostvm_cached_page_table_levels = 0,
95 .pte_group_size_bytes = 2048,
96 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
101 .rob_buffer_size_kbytes = 168,
102 .det_buffer_size_kbytes = 164,
103 .dpte_buffer_size_in_pte_reqs_luma = 84,
104 .pde_proc_buffer_size_64k_reqs = 48,
105 .dpp_output_buffer_pixels = 2560,
106 .opp_output_buffer_lines = 1,
107 .pixel_chunk_size_kbytes = 8,
108 .pte_chunk_size_kbytes = 2,
109 .meta_chunk_size_kbytes = 2,
110 .writeback_chunk_size_kbytes = 2,
111 .line_buffer_size_bits = 789504,
112 .is_line_buffer_bpp_fixed = 0,
113 .line_buffer_fixed_bpp = 0,
114 .dcc_supported = true,
115 .max_line_buffer_lines = 12,
116 .writeback_luma_buffer_size_kbytes = 12,
117 .writeback_chroma_buffer_size_kbytes = 8,
118 .writeback_chroma_line_buffer_width_pixels = 4,
119 .writeback_max_hscl_ratio = 1,
120 .writeback_max_vscl_ratio = 1,
121 .writeback_min_hscl_ratio = 1,
122 .writeback_min_vscl_ratio = 1,
123 .writeback_max_hscl_taps = 12,
124 .writeback_max_vscl_taps = 12,
125 .writeback_line_buffer_luma_buffer_size = 0,
126 .writeback_line_buffer_chroma_buffer_size = 14643,
127 .cursor_buffer_size = 8,
128 .cursor_chunk_size = 2,
132 .max_dchub_pscl_bw_pix_per_clk = 4,
133 .max_pscl_lb_bw_pix_per_clk = 2,
134 .max_lb_vscl_bw_pix_per_clk = 4,
135 .max_vscl_hscl_bw_pix_per_clk = 4,
142 .dispclk_ramp_margin_percent = 1,
143 .underscan_factor = 1.10,
144 .min_vblank_lines = 32, //
145 .dppclk_delay_subtotal = 77, //
146 .dppclk_delay_scl_lb_only = 16,
147 .dppclk_delay_scl = 50,
148 .dppclk_delay_cnvc_formatter = 8,
149 .dppclk_delay_cnvc_cursor = 6,
150 .dispclk_delay_subtotal = 87, //
151 .dcfclk_cstate_latency = 10, // SRExitTime
152 .max_inter_dcn_tile_repeaters = 8,
154 .xfc_supported = true,
155 .xfc_fill_bw_overhead_percent = 10.0,
156 .xfc_fill_constant_bytes = 0,
159 struct _vcs_dpi_soc_bounding_box_st dcn2_0_soc = {
160 /* Defaults that get patched on driver load from firmware. */
165 .fabricclk_mhz = 560.0,
166 .dispclk_mhz = 513.0,
171 .dram_speed_mts = 8960.0,
176 .fabricclk_mhz = 694.0,
177 .dispclk_mhz = 642.0,
182 .dram_speed_mts = 11104.0,
187 .fabricclk_mhz = 875.0,
188 .dispclk_mhz = 734.0,
193 .dram_speed_mts = 14000.0,
197 .dcfclk_mhz = 1000.0,
198 .fabricclk_mhz = 1000.0,
199 .dispclk_mhz = 1100.0,
200 .dppclk_mhz = 1100.0,
202 .socclk_mhz = 1000.0,
204 .dram_speed_mts = 16000.0,
208 .dcfclk_mhz = 1200.0,
209 .fabricclk_mhz = 1200.0,
210 .dispclk_mhz = 1284.0,
211 .dppclk_mhz = 1284.0,
213 .socclk_mhz = 1200.0,
215 .dram_speed_mts = 16000.0,
217 /*Extra state, no dispclk ramping*/
220 .dcfclk_mhz = 1200.0,
221 .fabricclk_mhz = 1200.0,
222 .dispclk_mhz = 1284.0,
223 .dppclk_mhz = 1284.0,
225 .socclk_mhz = 1200.0,
227 .dram_speed_mts = 16000.0,
231 .sr_exit_time_us = 8.6,
232 .sr_enter_plus_exit_time_us = 10.9,
233 .urgent_latency_us = 4.0,
234 .urgent_latency_pixel_data_only_us = 4.0,
235 .urgent_latency_pixel_mixed_with_vm_data_us = 4.0,
236 .urgent_latency_vm_data_only_us = 4.0,
237 .urgent_out_of_order_return_per_channel_pixel_only_bytes = 4096,
238 .urgent_out_of_order_return_per_channel_pixel_and_vm_bytes = 4096,
239 .urgent_out_of_order_return_per_channel_vm_only_bytes = 4096,
240 .pct_ideal_dram_sdp_bw_after_urgent_pixel_only = 40.0,
241 .pct_ideal_dram_sdp_bw_after_urgent_pixel_and_vm = 40.0,
242 .pct_ideal_dram_sdp_bw_after_urgent_vm_only = 40.0,
243 .max_avg_sdp_bw_use_normal_percent = 40.0,
244 .max_avg_dram_bw_use_normal_percent = 40.0,
245 .writeback_latency_us = 12.0,
246 .ideal_dram_bw_after_urgent_percent = 40.0,
247 .max_request_size_bytes = 256,
248 .dram_channel_width_bytes = 2,
249 .fabric_datapath_to_dcn_data_return_bytes = 64,
250 .dcn_downspread_percent = 0.5,
251 .downspread_percent = 0.38,
252 .dram_page_open_time_ns = 50.0,
253 .dram_rw_turnaround_time_ns = 17.5,
254 .dram_return_buffer_per_channel_bytes = 8192,
255 .round_trip_ping_latency_dcfclk_cycles = 131,
256 .urgent_out_of_order_return_per_channel_bytes = 256,
257 .channel_interleave_bytes = 256,
260 .vmm_page_size_bytes = 4096,
261 .dram_clock_change_latency_us = 404.0,
262 .dummy_pstate_latency_us = 5.0,
263 .writeback_dram_clock_change_latency_us = 23.0,
264 .return_bus_width_bytes = 64,
265 .dispclk_dppclk_vco_speed_mhz = 3850,
266 .xfc_bus_transport_time_us = 20,
267 .xfc_xbuf_latency_tolerance_us = 4,
268 .use_urgent_burst_bw = 0
272 #ifndef mmDP0_DP_DPHY_INTERNAL_CTRL
273 #define mmDP0_DP_DPHY_INTERNAL_CTRL 0x210f
274 #define mmDP0_DP_DPHY_INTERNAL_CTRL_BASE_IDX 2
275 #define mmDP1_DP_DPHY_INTERNAL_CTRL 0x220f
276 #define mmDP1_DP_DPHY_INTERNAL_CTRL_BASE_IDX 2
277 #define mmDP2_DP_DPHY_INTERNAL_CTRL 0x230f
278 #define mmDP2_DP_DPHY_INTERNAL_CTRL_BASE_IDX 2
279 #define mmDP3_DP_DPHY_INTERNAL_CTRL 0x240f
280 #define mmDP3_DP_DPHY_INTERNAL_CTRL_BASE_IDX 2
281 #define mmDP4_DP_DPHY_INTERNAL_CTRL 0x250f
282 #define mmDP4_DP_DPHY_INTERNAL_CTRL_BASE_IDX 2
283 #define mmDP5_DP_DPHY_INTERNAL_CTRL 0x260f
284 #define mmDP5_DP_DPHY_INTERNAL_CTRL_BASE_IDX 2
285 #define mmDP6_DP_DPHY_INTERNAL_CTRL 0x270f
286 #define mmDP6_DP_DPHY_INTERNAL_CTRL_BASE_IDX 2
290 enum dcn20_clk_src_array_id {
300 /* begin *********************
301 * macros to expend register list macro defined in HW object header file */
304 /* TODO awful hack. fixup dcn20_dwb.h */
306 #define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
308 #define BASE(seg) BASE_INNER(seg)
310 #define SR(reg_name)\
311 .reg_name = BASE(mm ## reg_name ## _BASE_IDX) + \
314 #define SRI(reg_name, block, id)\
315 .reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
316 mm ## block ## id ## _ ## reg_name
318 #define SRIR(var_name, reg_name, block, id)\
319 .var_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
320 mm ## block ## id ## _ ## reg_name
322 #define SRII(reg_name, block, id)\
323 .reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
324 mm ## block ## id ## _ ## reg_name
326 #define DCCG_SRII(reg_name, block, id)\
327 .block ## _ ## reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
328 mm ## block ## id ## _ ## reg_name
331 #define NBIO_BASE_INNER(seg) \
332 NBIO_BASE__INST0_SEG ## seg
334 #define NBIO_BASE(seg) \
337 #define NBIO_SR(reg_name)\
338 .reg_name = NBIO_BASE(mm ## reg_name ## _BASE_IDX) + \
342 #define MMHUB_BASE_INNER(seg) \
343 MMHUB_BASE__INST0_SEG ## seg
345 #define MMHUB_BASE(seg) \
346 MMHUB_BASE_INNER(seg)
348 #define MMHUB_SR(reg_name)\
349 .reg_name = MMHUB_BASE(mmMM ## reg_name ## _BASE_IDX) + \
352 static const struct bios_registers bios_regs = {
353 NBIO_SR(BIOS_SCRATCH_3),
354 NBIO_SR(BIOS_SCRATCH_6)
357 #define clk_src_regs(index, pllid)\
359 CS_COMMON_REG_LIST_DCN2_0(index, pllid),\
362 static const struct dce110_clk_src_regs clk_src_regs[] = {
371 static const struct dce110_clk_src_shift cs_shift = {
372 CS_COMMON_MASK_SH_LIST_DCN2_0(__SHIFT)
375 static const struct dce110_clk_src_mask cs_mask = {
376 CS_COMMON_MASK_SH_LIST_DCN2_0(_MASK)
379 static const struct dce_dmcu_registers dmcu_regs = {
380 DMCU_DCN10_REG_LIST()
383 static const struct dce_dmcu_shift dmcu_shift = {
384 DMCU_MASK_SH_LIST_DCN10(__SHIFT)
387 static const struct dce_dmcu_mask dmcu_mask = {
388 DMCU_MASK_SH_LIST_DCN10(_MASK)
391 static const struct dce_abm_registers abm_regs = {
395 static const struct dce_abm_shift abm_shift = {
396 ABM_MASK_SH_LIST_DCN20(__SHIFT)
399 static const struct dce_abm_mask abm_mask = {
400 ABM_MASK_SH_LIST_DCN20(_MASK)
403 #define audio_regs(id)\
405 AUD_COMMON_REG_LIST(id)\
408 static const struct dce_audio_registers audio_regs[] = {
418 #define DCE120_AUD_COMMON_MASK_SH_LIST(mask_sh)\
419 SF(AZF0ENDPOINT0_AZALIA_F0_CODEC_ENDPOINT_INDEX, AZALIA_ENDPOINT_REG_INDEX, mask_sh),\
420 SF(AZF0ENDPOINT0_AZALIA_F0_CODEC_ENDPOINT_DATA, AZALIA_ENDPOINT_REG_DATA, mask_sh),\
421 AUD_COMMON_MASK_SH_LIST_BASE(mask_sh)
423 static const struct dce_audio_shift audio_shift = {
424 DCE120_AUD_COMMON_MASK_SH_LIST(__SHIFT)
427 static const struct dce_audio_mask audio_mask = {
428 DCE120_AUD_COMMON_MASK_SH_LIST(_MASK)
431 #define stream_enc_regs(id)\
433 SE_DCN2_REG_LIST(id)\
436 static const struct dcn10_stream_enc_registers stream_enc_regs[] = {
445 static const struct dcn10_stream_encoder_shift se_shift = {
446 SE_COMMON_MASK_SH_LIST_DCN20(__SHIFT)
449 static const struct dcn10_stream_encoder_mask se_mask = {
450 SE_COMMON_MASK_SH_LIST_DCN20(_MASK)
454 #define aux_regs(id)\
456 DCN2_AUX_REG_LIST(id)\
459 static const struct dcn10_link_enc_aux_registers link_enc_aux_regs[] = {
468 #define hpd_regs(id)\
473 static const struct dcn10_link_enc_hpd_registers link_enc_hpd_regs[] = {
482 #define link_regs(id, phyid)\
484 LE_DCN10_REG_LIST(id), \
485 UNIPHY_DCN2_REG_LIST(phyid), \
486 SRI(DP_DPHY_INTERNAL_CTRL, DP, id) \
489 static const struct dcn10_link_enc_registers link_enc_regs[] = {
498 static const struct dcn10_link_enc_shift le_shift = {
499 LINK_ENCODER_MASK_SH_LIST_DCN20(__SHIFT)
502 static const struct dcn10_link_enc_mask le_mask = {
503 LINK_ENCODER_MASK_SH_LIST_DCN20(_MASK)
506 #define ipp_regs(id)\
508 IPP_REG_LIST_DCN20(id),\
511 static const struct dcn10_ipp_registers ipp_regs[] = {
520 static const struct dcn10_ipp_shift ipp_shift = {
521 IPP_MASK_SH_LIST_DCN20(__SHIFT)
524 static const struct dcn10_ipp_mask ipp_mask = {
525 IPP_MASK_SH_LIST_DCN20(_MASK),
528 #define opp_regs(id)\
530 OPP_REG_LIST_DCN20(id),\
533 static const struct dcn20_opp_registers opp_regs[] = {
542 static const struct dcn20_opp_shift opp_shift = {
543 OPP_MASK_SH_LIST_DCN20(__SHIFT)
546 static const struct dcn20_opp_mask opp_mask = {
547 OPP_MASK_SH_LIST_DCN20(_MASK)
550 #define aux_engine_regs(id)\
552 AUX_COMMON_REG_LIST0(id), \
555 .AUX_RESET_MASK = DP_AUX0_AUX_CONTROL__AUX_RESET_MASK, \
558 static const struct dce110_aux_registers aux_engine_regs[] = {
569 TF_REG_LIST_DCN20(id),\
572 static const struct dcn2_dpp_registers tf_regs[] = {
581 static const struct dcn2_dpp_shift tf_shift = {
582 TF_REG_LIST_SH_MASK_DCN20(__SHIFT)
585 static const struct dcn2_dpp_mask tf_mask = {
586 TF_REG_LIST_SH_MASK_DCN20(_MASK)
589 #define dwbc_regs_dcn2(id)\
591 DWBC_COMMON_REG_LIST_DCN2_0(id),\
594 static const struct dcn20_dwbc_registers dwbc20_regs[] = {
598 static const struct dcn20_dwbc_shift dwbc20_shift = {
599 DWBC_COMMON_MASK_SH_LIST_DCN2_0(__SHIFT)
602 static const struct dcn20_dwbc_mask dwbc20_mask = {
603 DWBC_COMMON_MASK_SH_LIST_DCN2_0(_MASK)
606 #define mcif_wb_regs_dcn2(id)\
608 MCIF_WB_COMMON_REG_LIST_DCN2_0(id),\
611 static const struct dcn20_mmhubbub_registers mcif_wb20_regs[] = {
612 mcif_wb_regs_dcn2(0),
615 static const struct dcn20_mmhubbub_shift mcif_wb20_shift = {
616 MCIF_WB_COMMON_MASK_SH_LIST_DCN2_0(__SHIFT)
619 static const struct dcn20_mmhubbub_mask mcif_wb20_mask = {
620 MCIF_WB_COMMON_MASK_SH_LIST_DCN2_0(_MASK)
623 static const struct dcn20_mpc_registers mpc_regs = {
624 MPC_REG_LIST_DCN2_0(0),
625 MPC_REG_LIST_DCN2_0(1),
626 MPC_REG_LIST_DCN2_0(2),
627 MPC_REG_LIST_DCN2_0(3),
628 MPC_REG_LIST_DCN2_0(4),
629 MPC_REG_LIST_DCN2_0(5),
630 MPC_OUT_MUX_REG_LIST_DCN2_0(0),
631 MPC_OUT_MUX_REG_LIST_DCN2_0(1),
632 MPC_OUT_MUX_REG_LIST_DCN2_0(2),
633 MPC_OUT_MUX_REG_LIST_DCN2_0(3),
634 MPC_OUT_MUX_REG_LIST_DCN2_0(4),
635 MPC_OUT_MUX_REG_LIST_DCN2_0(5),
638 static const struct dcn20_mpc_shift mpc_shift = {
639 MPC_COMMON_MASK_SH_LIST_DCN2_0(__SHIFT)
642 static const struct dcn20_mpc_mask mpc_mask = {
643 MPC_COMMON_MASK_SH_LIST_DCN2_0(_MASK)
647 [id] = {TG_COMMON_REG_LIST_DCN2_0(id)}
650 static const struct dcn_optc_registers tg_regs[] = {
659 static const struct dcn_optc_shift tg_shift = {
660 TG_COMMON_MASK_SH_LIST_DCN2_0(__SHIFT)
663 static const struct dcn_optc_mask tg_mask = {
664 TG_COMMON_MASK_SH_LIST_DCN2_0(_MASK)
667 #define hubp_regs(id)\
669 HUBP_REG_LIST_DCN20(id)\
672 static const struct dcn_hubp2_registers hubp_regs[] = {
681 static const struct dcn_hubp2_shift hubp_shift = {
682 HUBP_MASK_SH_LIST_DCN20(__SHIFT)
685 static const struct dcn_hubp2_mask hubp_mask = {
686 HUBP_MASK_SH_LIST_DCN20(_MASK)
689 static const struct dcn_hubbub_registers hubbub_reg = {
690 HUBBUB_REG_LIST_DCN20(0)
693 static const struct dcn_hubbub_shift hubbub_shift = {
694 HUBBUB_MASK_SH_LIST_DCN20(__SHIFT)
697 static const struct dcn_hubbub_mask hubbub_mask = {
698 HUBBUB_MASK_SH_LIST_DCN20(_MASK)
701 #define vmid_regs(id)\
703 DCN20_VMID_REG_LIST(id)\
706 static const struct dcn_vmid_registers vmid_regs[] = {
725 static const struct dcn20_vmid_shift vmid_shifts = {
726 DCN20_VMID_MASK_SH_LIST(__SHIFT)
729 static const struct dcn20_vmid_mask vmid_masks = {
730 DCN20_VMID_MASK_SH_LIST(_MASK)
733 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
734 #define dsc_regsDCN20(id)\
736 DSC_REG_LIST_DCN20(id)\
739 static const struct dcn20_dsc_registers dsc_regs[] = {
748 static const struct dcn20_dsc_shift dsc_shift = {
749 DSC_REG_LIST_SH_MASK_DCN20(__SHIFT)
752 static const struct dcn20_dsc_mask dsc_mask = {
753 DSC_REG_LIST_SH_MASK_DCN20(_MASK)
757 static const struct dccg_registers dccg_regs = {
761 static const struct dccg_shift dccg_shift = {
762 DCCG_MASK_SH_LIST_DCN2(__SHIFT)
765 static const struct dccg_mask dccg_mask = {
766 DCCG_MASK_SH_LIST_DCN2(_MASK)
769 static const struct resource_caps res_cap_nv10 = {
770 .num_timing_generator = 6,
772 .num_video_plane = 6,
774 .num_stream_encoder = 6,
779 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
784 static const struct dc_plane_cap plane_cap = {
785 .type = DC_PLANE_TYPE_DCN_UNIVERSAL,
786 .blends_with_above = true,
787 .blends_with_below = true,
788 .per_pixel_alpha = true,
790 .pixel_format_support = {
796 .max_upscale_factor = {
802 .max_downscale_factor = {
808 static const struct resource_caps res_cap_nv14 = {
809 .num_timing_generator = 5,
811 .num_video_plane = 5,
813 .num_stream_encoder = 5,
819 static const struct dc_debug_options debug_defaults_drv = {
820 .disable_dmcu = true,
821 .force_abm_enable = false,
822 .timing_trace = false,
824 .disable_pplib_clock_request = true,
825 .pipe_split_policy = MPC_SPLIT_DYNAMIC,
826 .force_single_disp_pipe_split = true,
827 .disable_dcc = DCC_ENABLE,
829 .performance_trace = false,
830 .max_downscale_src_width = 5120,/*upto 5K*/
831 .disable_pplib_wm_range = false,
832 .scl_reset_length10 = true,
833 .sanity_checks = false,
834 .disable_tri_buf = true,
835 .underflow_assert_delay_us = 0xFFFFFFFF,
838 static const struct dc_debug_options debug_defaults_diags = {
839 .disable_dmcu = true,
840 .force_abm_enable = false,
841 .timing_trace = true,
843 .disable_dpp_power_gate = true,
844 .disable_hubp_power_gate = true,
845 .disable_clock_gate = true,
846 .disable_pplib_clock_request = true,
847 .disable_pplib_wm_range = true,
848 .disable_stutter = true,
849 .scl_reset_length10 = true,
850 .underflow_assert_delay_us = 0xFFFFFFFF,
853 void dcn20_dpp_destroy(struct dpp **dpp)
855 kfree(TO_DCN20_DPP(*dpp));
859 struct dpp *dcn20_dpp_create(
860 struct dc_context *ctx,
863 struct dcn20_dpp *dpp =
864 kzalloc(sizeof(struct dcn20_dpp), GFP_KERNEL);
869 if (dpp2_construct(dpp, ctx, inst,
870 &tf_regs[inst], &tf_shift, &tf_mask))
878 struct input_pixel_processor *dcn20_ipp_create(
879 struct dc_context *ctx, uint32_t inst)
881 struct dcn10_ipp *ipp =
882 kzalloc(sizeof(struct dcn10_ipp), GFP_KERNEL);
889 dcn20_ipp_construct(ipp, ctx, inst,
890 &ipp_regs[inst], &ipp_shift, &ipp_mask);
895 struct output_pixel_processor *dcn20_opp_create(
896 struct dc_context *ctx, uint32_t inst)
898 struct dcn20_opp *opp =
899 kzalloc(sizeof(struct dcn20_opp), GFP_KERNEL);
906 dcn20_opp_construct(opp, ctx, inst,
907 &opp_regs[inst], &opp_shift, &opp_mask);
911 struct dce_aux *dcn20_aux_engine_create(
912 struct dc_context *ctx,
915 struct aux_engine_dce110 *aux_engine =
916 kzalloc(sizeof(struct aux_engine_dce110), GFP_KERNEL);
921 dce110_aux_engine_construct(aux_engine, ctx, inst,
922 SW_AUX_TIMEOUT_PERIOD_MULTIPLIER * AUX_TIMEOUT_PERIOD,
923 &aux_engine_regs[inst]);
925 return &aux_engine->base;
927 #define i2c_inst_regs(id) { I2C_HW_ENGINE_COMMON_REG_LIST(id) }
929 static const struct dce_i2c_registers i2c_hw_regs[] = {
938 static const struct dce_i2c_shift i2c_shifts = {
939 I2C_COMMON_MASK_SH_LIST_DCN2(__SHIFT)
942 static const struct dce_i2c_mask i2c_masks = {
943 I2C_COMMON_MASK_SH_LIST_DCN2(_MASK)
946 struct dce_i2c_hw *dcn20_i2c_hw_create(
947 struct dc_context *ctx,
950 struct dce_i2c_hw *dce_i2c_hw =
951 kzalloc(sizeof(struct dce_i2c_hw), GFP_KERNEL);
956 dcn2_i2c_hw_construct(dce_i2c_hw, ctx, inst,
957 &i2c_hw_regs[inst], &i2c_shifts, &i2c_masks);
961 struct mpc *dcn20_mpc_create(struct dc_context *ctx)
963 struct dcn20_mpc *mpc20 = kzalloc(sizeof(struct dcn20_mpc),
969 dcn20_mpc_construct(mpc20, ctx,
978 struct hubbub *dcn20_hubbub_create(struct dc_context *ctx)
981 struct dcn20_hubbub *hubbub = kzalloc(sizeof(struct dcn20_hubbub),
987 hubbub2_construct(hubbub, ctx,
992 for (i = 0; i < res_cap_nv10.num_vmid; i++) {
993 struct dcn20_vmid *vmid = &hubbub->vmid[i];
997 vmid->regs = &vmid_regs[i];
998 vmid->shifts = &vmid_shifts;
999 vmid->masks = &vmid_masks;
1002 return &hubbub->base;
1005 struct timing_generator *dcn20_timing_generator_create(
1006 struct dc_context *ctx,
1009 struct optc *tgn10 =
1010 kzalloc(sizeof(struct optc), GFP_KERNEL);
1015 tgn10->base.inst = instance;
1016 tgn10->base.ctx = ctx;
1018 tgn10->tg_regs = &tg_regs[instance];
1019 tgn10->tg_shift = &tg_shift;
1020 tgn10->tg_mask = &tg_mask;
1022 dcn20_timing_generator_init(tgn10);
1024 return &tgn10->base;
1027 static const struct encoder_feature_support link_enc_feature = {
1028 .max_hdmi_deep_color = COLOR_DEPTH_121212,
1029 .max_hdmi_pixel_clock = 600000,
1030 .hdmi_ycbcr420_supported = true,
1031 .dp_ycbcr420_supported = true,
1032 .flags.bits.IS_HBR2_CAPABLE = true,
1033 .flags.bits.IS_HBR3_CAPABLE = true,
1034 .flags.bits.IS_TPS3_CAPABLE = true,
1035 .flags.bits.IS_TPS4_CAPABLE = true
1038 struct link_encoder *dcn20_link_encoder_create(
1039 const struct encoder_init_data *enc_init_data)
1041 struct dcn20_link_encoder *enc20 =
1042 kzalloc(sizeof(struct dcn20_link_encoder), GFP_KERNEL);
1047 dcn20_link_encoder_construct(enc20,
1050 &link_enc_regs[enc_init_data->transmitter],
1051 &link_enc_aux_regs[enc_init_data->channel - 1],
1052 &link_enc_hpd_regs[enc_init_data->hpd_source],
1056 return &enc20->enc10.base;
1059 struct clock_source *dcn20_clock_source_create(
1060 struct dc_context *ctx,
1061 struct dc_bios *bios,
1062 enum clock_source_id id,
1063 const struct dce110_clk_src_regs *regs,
1066 struct dce110_clk_src *clk_src =
1067 kzalloc(sizeof(struct dce110_clk_src), GFP_KERNEL);
1072 if (dcn20_clk_src_construct(clk_src, ctx, bios, id,
1073 regs, &cs_shift, &cs_mask)) {
1074 clk_src->base.dp_clk_src = dp_clk_src;
1075 return &clk_src->base;
1078 BREAK_TO_DEBUGGER();
1082 static void read_dce_straps(
1083 struct dc_context *ctx,
1084 struct resource_straps *straps)
1086 generic_reg_get(ctx, mmDC_PINSTRAPS + BASE(mmDC_PINSTRAPS_BASE_IDX),
1087 FN(DC_PINSTRAPS, DC_PINSTRAPS_AUDIO), &straps->dc_pinstraps_audio);
1090 static struct audio *dcn20_create_audio(
1091 struct dc_context *ctx, unsigned int inst)
1093 return dce_audio_create(ctx, inst,
1094 &audio_regs[inst], &audio_shift, &audio_mask);
1097 struct stream_encoder *dcn20_stream_encoder_create(
1098 enum engine_id eng_id,
1099 struct dc_context *ctx)
1101 struct dcn10_stream_encoder *enc1 =
1102 kzalloc(sizeof(struct dcn10_stream_encoder), GFP_KERNEL);
1107 dcn20_stream_encoder_construct(enc1, ctx, ctx->dc_bios, eng_id,
1108 &stream_enc_regs[eng_id],
1109 &se_shift, &se_mask);
1114 static const struct dce_hwseq_registers hwseq_reg = {
1115 HWSEQ_DCN2_REG_LIST()
1118 static const struct dce_hwseq_shift hwseq_shift = {
1119 HWSEQ_DCN2_MASK_SH_LIST(__SHIFT)
1122 static const struct dce_hwseq_mask hwseq_mask = {
1123 HWSEQ_DCN2_MASK_SH_LIST(_MASK)
1126 struct dce_hwseq *dcn20_hwseq_create(
1127 struct dc_context *ctx)
1129 struct dce_hwseq *hws = kzalloc(sizeof(struct dce_hwseq), GFP_KERNEL);
1133 hws->regs = &hwseq_reg;
1134 hws->shifts = &hwseq_shift;
1135 hws->masks = &hwseq_mask;
1140 static const struct resource_create_funcs res_create_funcs = {
1141 .read_dce_straps = read_dce_straps,
1142 .create_audio = dcn20_create_audio,
1143 .create_stream_encoder = dcn20_stream_encoder_create,
1144 .create_hwseq = dcn20_hwseq_create,
1147 static const struct resource_create_funcs res_create_maximus_funcs = {
1148 .read_dce_straps = NULL,
1149 .create_audio = NULL,
1150 .create_stream_encoder = NULL,
1151 .create_hwseq = dcn20_hwseq_create,
1154 void dcn20_clock_source_destroy(struct clock_source **clk_src)
1156 kfree(TO_DCE110_CLK_SRC(*clk_src));
1160 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1162 struct display_stream_compressor *dcn20_dsc_create(
1163 struct dc_context *ctx, uint32_t inst)
1165 struct dcn20_dsc *dsc =
1166 kzalloc(sizeof(struct dcn20_dsc), GFP_KERNEL);
1169 BREAK_TO_DEBUGGER();
1173 dsc2_construct(dsc, ctx, inst, &dsc_regs[inst], &dsc_shift, &dsc_mask);
1177 void dcn20_dsc_destroy(struct display_stream_compressor **dsc)
1179 kfree(container_of(*dsc, struct dcn20_dsc, base));
1185 static void destruct(struct dcn20_resource_pool *pool)
1189 for (i = 0; i < pool->base.stream_enc_count; i++) {
1190 if (pool->base.stream_enc[i] != NULL) {
1191 kfree(DCN10STRENC_FROM_STRENC(pool->base.stream_enc[i]));
1192 pool->base.stream_enc[i] = NULL;
1196 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1197 for (i = 0; i < pool->base.res_cap->num_dsc; i++) {
1198 if (pool->base.dscs[i] != NULL)
1199 dcn20_dsc_destroy(&pool->base.dscs[i]);
1203 if (pool->base.mpc != NULL) {
1204 kfree(TO_DCN20_MPC(pool->base.mpc));
1205 pool->base.mpc = NULL;
1207 if (pool->base.hubbub != NULL) {
1208 kfree(pool->base.hubbub);
1209 pool->base.hubbub = NULL;
1211 for (i = 0; i < pool->base.pipe_count; i++) {
1212 if (pool->base.dpps[i] != NULL)
1213 dcn20_dpp_destroy(&pool->base.dpps[i]);
1215 if (pool->base.ipps[i] != NULL)
1216 pool->base.ipps[i]->funcs->ipp_destroy(&pool->base.ipps[i]);
1218 if (pool->base.hubps[i] != NULL) {
1219 kfree(TO_DCN20_HUBP(pool->base.hubps[i]));
1220 pool->base.hubps[i] = NULL;
1223 if (pool->base.irqs != NULL) {
1224 dal_irq_service_destroy(&pool->base.irqs);
1228 for (i = 0; i < pool->base.res_cap->num_ddc; i++) {
1229 if (pool->base.engines[i] != NULL)
1230 dce110_engine_destroy(&pool->base.engines[i]);
1231 if (pool->base.hw_i2cs[i] != NULL) {
1232 kfree(pool->base.hw_i2cs[i]);
1233 pool->base.hw_i2cs[i] = NULL;
1235 if (pool->base.sw_i2cs[i] != NULL) {
1236 kfree(pool->base.sw_i2cs[i]);
1237 pool->base.sw_i2cs[i] = NULL;
1241 for (i = 0; i < pool->base.res_cap->num_opp; i++) {
1242 if (pool->base.opps[i] != NULL)
1243 pool->base.opps[i]->funcs->opp_destroy(&pool->base.opps[i]);
1246 for (i = 0; i < pool->base.res_cap->num_timing_generator; i++) {
1247 if (pool->base.timing_generators[i] != NULL) {
1248 kfree(DCN10TG_FROM_TG(pool->base.timing_generators[i]));
1249 pool->base.timing_generators[i] = NULL;
1253 for (i = 0; i < pool->base.res_cap->num_dwb; i++) {
1254 if (pool->base.dwbc[i] != NULL) {
1255 kfree(TO_DCN20_DWBC(pool->base.dwbc[i]));
1256 pool->base.dwbc[i] = NULL;
1258 if (pool->base.mcif_wb[i] != NULL) {
1259 kfree(TO_DCN20_MMHUBBUB(pool->base.mcif_wb[i]));
1260 pool->base.mcif_wb[i] = NULL;
1264 for (i = 0; i < pool->base.audio_count; i++) {
1265 if (pool->base.audios[i])
1266 dce_aud_destroy(&pool->base.audios[i]);
1269 for (i = 0; i < pool->base.clk_src_count; i++) {
1270 if (pool->base.clock_sources[i] != NULL) {
1271 dcn20_clock_source_destroy(&pool->base.clock_sources[i]);
1272 pool->base.clock_sources[i] = NULL;
1276 if (pool->base.dp_clock_source != NULL) {
1277 dcn20_clock_source_destroy(&pool->base.dp_clock_source);
1278 pool->base.dp_clock_source = NULL;
1282 if (pool->base.abm != NULL)
1283 dce_abm_destroy(&pool->base.abm);
1285 if (pool->base.dmcu != NULL)
1286 dce_dmcu_destroy(&pool->base.dmcu);
1288 if (pool->base.dccg != NULL)
1289 dcn_dccg_destroy(&pool->base.dccg);
1291 if (pool->base.pp_smu != NULL)
1292 dcn20_pp_smu_destroy(&pool->base.pp_smu);
1296 struct hubp *dcn20_hubp_create(
1297 struct dc_context *ctx,
1300 struct dcn20_hubp *hubp2 =
1301 kzalloc(sizeof(struct dcn20_hubp), GFP_KERNEL);
1306 if (hubp2_construct(hubp2, ctx, inst,
1307 &hubp_regs[inst], &hubp_shift, &hubp_mask))
1308 return &hubp2->base;
1310 BREAK_TO_DEBUGGER();
1315 static void get_pixel_clock_parameters(
1316 struct pipe_ctx *pipe_ctx,
1317 struct pixel_clk_params *pixel_clk_params)
1319 const struct dc_stream_state *stream = pipe_ctx->stream;
1320 bool odm_combine = dc_res_get_odm_bottom_pipe(pipe_ctx) != NULL;
1322 pixel_clk_params->requested_pix_clk_100hz = stream->timing.pix_clk_100hz;
1323 pixel_clk_params->encoder_object_id = stream->link->link_enc->id;
1324 pixel_clk_params->signal_type = pipe_ctx->stream->signal;
1325 pixel_clk_params->controller_id = pipe_ctx->stream_res.tg->inst + 1;
1326 /* TODO: un-hardcode*/
1327 pixel_clk_params->requested_sym_clk = LINK_RATE_LOW *
1328 LINK_RATE_REF_FREQ_IN_KHZ;
1329 pixel_clk_params->flags.ENABLE_SS = 0;
1330 pixel_clk_params->color_depth =
1331 stream->timing.display_color_depth;
1332 pixel_clk_params->flags.DISPLAY_BLANKED = 1;
1333 pixel_clk_params->pixel_encoding = stream->timing.pixel_encoding;
1335 if (stream->timing.pixel_encoding == PIXEL_ENCODING_YCBCR422)
1336 pixel_clk_params->color_depth = COLOR_DEPTH_888;
1338 if (optc1_is_two_pixels_per_containter(&stream->timing) || odm_combine)
1339 pixel_clk_params->requested_pix_clk_100hz /= 2;
1341 if (stream->timing.timing_3d_format == TIMING_3D_FORMAT_HW_FRAME_PACKING)
1342 pixel_clk_params->requested_pix_clk_100hz *= 2;
1346 static void build_clamping_params(struct dc_stream_state *stream)
1348 stream->clamping.clamping_level = CLAMPING_FULL_RANGE;
1349 stream->clamping.c_depth = stream->timing.display_color_depth;
1350 stream->clamping.pixel_encoding = stream->timing.pixel_encoding;
1353 static enum dc_status build_pipe_hw_param(struct pipe_ctx *pipe_ctx)
1356 get_pixel_clock_parameters(pipe_ctx, &pipe_ctx->stream_res.pix_clk_params);
1358 pipe_ctx->clock_source->funcs->get_pix_clk_dividers(
1359 pipe_ctx->clock_source,
1360 &pipe_ctx->stream_res.pix_clk_params,
1361 &pipe_ctx->pll_settings);
1363 pipe_ctx->stream->clamping.pixel_encoding = pipe_ctx->stream->timing.pixel_encoding;
1365 resource_build_bit_depth_reduction_params(pipe_ctx->stream,
1366 &pipe_ctx->stream->bit_depth_params);
1367 build_clamping_params(pipe_ctx->stream);
1372 enum dc_status dcn20_build_mapped_resource(const struct dc *dc, struct dc_state *context, struct dc_stream_state *stream)
1374 enum dc_status status = DC_OK;
1375 struct pipe_ctx *pipe_ctx = resource_get_head_pipe_for_stream(&context->res_ctx, stream);
1377 /*TODO Seems unneeded anymore */
1378 /* if (old_context && resource_is_stream_unchanged(old_context, stream)) {
1379 if (stream != NULL && old_context->streams[i] != NULL) {
1380 todo: shouldn't have to copy missing parameter here
1381 resource_build_bit_depth_reduction_params(stream,
1382 &stream->bit_depth_params);
1383 stream->clamping.pixel_encoding =
1384 stream->timing.pixel_encoding;
1386 resource_build_bit_depth_reduction_params(stream,
1387 &stream->bit_depth_params);
1388 build_clamping_params(stream);
1396 return DC_ERROR_UNEXPECTED;
1399 status = build_pipe_hw_param(pipe_ctx);
1404 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1406 static void acquire_dsc(struct resource_context *res_ctx,
1407 const struct resource_pool *pool,
1408 struct display_stream_compressor **dsc)
1412 ASSERT(*dsc == NULL);
1415 /* Find first free DSC */
1416 for (i = 0; i < pool->res_cap->num_dsc; i++)
1417 if (!res_ctx->is_dsc_acquired[i]) {
1418 *dsc = pool->dscs[i];
1419 res_ctx->is_dsc_acquired[i] = true;
1424 static void release_dsc(struct resource_context *res_ctx,
1425 const struct resource_pool *pool,
1426 struct display_stream_compressor **dsc)
1430 for (i = 0; i < pool->res_cap->num_dsc; i++)
1431 if (pool->dscs[i] == *dsc) {
1432 res_ctx->is_dsc_acquired[i] = false;
1441 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1442 static enum dc_status add_dsc_to_stream_resource(struct dc *dc,
1443 struct dc_state *dc_ctx,
1444 struct dc_stream_state *dc_stream)
1446 enum dc_status result = DC_OK;
1448 const struct resource_pool *pool = dc->res_pool;
1450 /* Get a DSC if required and available */
1451 for (i = 0; i < dc->res_pool->pipe_count; i++) {
1452 struct pipe_ctx *pipe_ctx = &dc_ctx->res_ctx.pipe_ctx[i];
1454 if (pipe_ctx->stream != dc_stream)
1457 acquire_dsc(&dc_ctx->res_ctx, pool, &pipe_ctx->stream_res.dsc);
1459 /* The number of DSCs can be less than the number of pipes */
1460 if (!pipe_ctx->stream_res.dsc) {
1461 dm_output_to_console("No DSCs available\n");
1462 result = DC_NO_DSC_RESOURCE;
1472 static enum dc_status remove_dsc_from_stream_resource(struct dc *dc,
1473 struct dc_state *new_ctx,
1474 struct dc_stream_state *dc_stream)
1476 struct pipe_ctx *pipe_ctx = NULL;
1479 for (i = 0; i < MAX_PIPES; i++) {
1480 if (new_ctx->res_ctx.pipe_ctx[i].stream == dc_stream && !new_ctx->res_ctx.pipe_ctx[i].top_pipe) {
1481 pipe_ctx = &new_ctx->res_ctx.pipe_ctx[i];
1487 return DC_ERROR_UNEXPECTED;
1489 if (pipe_ctx->stream_res.dsc) {
1490 struct pipe_ctx *odm_pipe = dc_res_get_odm_bottom_pipe(pipe_ctx);
1492 release_dsc(&new_ctx->res_ctx, dc->res_pool, &pipe_ctx->stream_res.dsc);
1494 release_dsc(&new_ctx->res_ctx, dc->res_pool, &odm_pipe->stream_res.dsc);
1502 enum dc_status dcn20_add_stream_to_ctx(struct dc *dc, struct dc_state *new_ctx, struct dc_stream_state *dc_stream)
1504 enum dc_status result = DC_ERROR_UNEXPECTED;
1506 result = resource_map_pool_resources(dc, new_ctx, dc_stream);
1508 if (result == DC_OK)
1509 result = resource_map_phy_clock_resources(dc, new_ctx, dc_stream);
1511 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1512 /* Get a DSC if required and available */
1513 if (result == DC_OK && dc_stream->timing.flags.DSC)
1514 result = add_dsc_to_stream_resource(dc, new_ctx, dc_stream);
1517 if (result == DC_OK)
1518 result = dcn20_build_mapped_resource(dc, new_ctx, dc_stream);
1524 enum dc_status dcn20_remove_stream_from_ctx(struct dc *dc, struct dc_state *new_ctx, struct dc_stream_state *dc_stream)
1526 enum dc_status result = DC_OK;
1528 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1529 result = remove_dsc_from_stream_resource(dc, new_ctx, dc_stream);
1536 static void swizzle_to_dml_params(
1537 enum swizzle_mode_values swizzle,
1538 unsigned int *sw_mode)
1542 *sw_mode = dm_sw_linear;
1545 *sw_mode = dm_sw_4kb_s;
1548 *sw_mode = dm_sw_4kb_s_x;
1551 *sw_mode = dm_sw_4kb_d;
1554 *sw_mode = dm_sw_4kb_d_x;
1557 *sw_mode = dm_sw_64kb_s;
1559 case DC_SW_64KB_S_X:
1560 *sw_mode = dm_sw_64kb_s_x;
1562 case DC_SW_64KB_S_T:
1563 *sw_mode = dm_sw_64kb_s_t;
1566 *sw_mode = dm_sw_64kb_d;
1568 case DC_SW_64KB_D_X:
1569 *sw_mode = dm_sw_64kb_d_x;
1571 case DC_SW_64KB_D_T:
1572 *sw_mode = dm_sw_64kb_d_t;
1574 case DC_SW_64KB_R_X:
1575 *sw_mode = dm_sw_64kb_r_x;
1578 *sw_mode = dm_sw_var_s;
1581 *sw_mode = dm_sw_var_s_x;
1584 *sw_mode = dm_sw_var_d;
1587 *sw_mode = dm_sw_var_d_x;
1591 ASSERT(0); /* Not supported */
1596 static bool dcn20_split_stream_for_combine(
1597 struct resource_context *res_ctx,
1598 const struct resource_pool *pool,
1599 struct pipe_ctx *primary_pipe,
1600 struct pipe_ctx *secondary_pipe,
1601 bool is_odm_combine)
1603 int pipe_idx = secondary_pipe->pipe_idx;
1604 struct scaler_data *sd = &primary_pipe->plane_res.scl_data;
1605 struct pipe_ctx *sec_bot_pipe = secondary_pipe->bottom_pipe;
1608 *secondary_pipe = *primary_pipe;
1609 secondary_pipe->bottom_pipe = sec_bot_pipe;
1611 secondary_pipe->pipe_idx = pipe_idx;
1612 secondary_pipe->plane_res.mi = pool->mis[secondary_pipe->pipe_idx];
1613 secondary_pipe->plane_res.hubp = pool->hubps[secondary_pipe->pipe_idx];
1614 secondary_pipe->plane_res.ipp = pool->ipps[secondary_pipe->pipe_idx];
1615 secondary_pipe->plane_res.xfm = pool->transforms[secondary_pipe->pipe_idx];
1616 secondary_pipe->plane_res.dpp = pool->dpps[secondary_pipe->pipe_idx];
1617 secondary_pipe->plane_res.mpcc_inst = pool->dpps[secondary_pipe->pipe_idx]->inst;
1618 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1619 secondary_pipe->stream_res.dsc = NULL;
1621 if (primary_pipe->bottom_pipe && primary_pipe->bottom_pipe != secondary_pipe) {
1622 ASSERT(!secondary_pipe->bottom_pipe);
1623 secondary_pipe->bottom_pipe = primary_pipe->bottom_pipe;
1624 secondary_pipe->bottom_pipe->top_pipe = secondary_pipe;
1626 primary_pipe->bottom_pipe = secondary_pipe;
1627 secondary_pipe->top_pipe = primary_pipe;
1629 if (is_odm_combine) {
1630 if (primary_pipe->plane_state) {
1631 /* HACTIVE halved for odm combine */
1633 /* Copy scl_data to secondary pipe */
1634 secondary_pipe->plane_res.scl_data = *sd;
1636 /* Calculate new vp and recout for left pipe */
1637 /* Need at least 16 pixels width per side */
1638 if (sd->recout.x + 16 >= sd->h_active)
1640 new_width = sd->h_active - sd->recout.x;
1641 sd->viewport.width -= dc_fixpt_floor(dc_fixpt_mul_int(
1642 sd->ratios.horz, sd->recout.width - new_width));
1643 sd->viewport_c.width -= dc_fixpt_floor(dc_fixpt_mul_int(
1644 sd->ratios.horz_c, sd->recout.width - new_width));
1645 sd->recout.width = new_width;
1647 /* Calculate new vp and recout for right pipe */
1648 sd = &secondary_pipe->plane_res.scl_data;
1649 new_width = sd->recout.width + sd->recout.x - sd->h_active;
1650 /* Need at least 16 pixels width per side */
1651 if (new_width <= 16)
1653 sd->viewport.width -= dc_fixpt_floor(dc_fixpt_mul_int(
1654 sd->ratios.horz, sd->recout.width - new_width));
1655 sd->viewport_c.width -= dc_fixpt_floor(dc_fixpt_mul_int(
1656 sd->ratios.horz_c, sd->recout.width - new_width));
1657 sd->recout.width = new_width;
1658 sd->viewport.x += dc_fixpt_floor(dc_fixpt_mul_int(
1659 sd->ratios.horz, sd->h_active - sd->recout.x));
1660 sd->viewport_c.x += dc_fixpt_floor(dc_fixpt_mul_int(
1661 sd->ratios.horz_c, sd->h_active - sd->recout.x));
1664 secondary_pipe->stream_res.opp = pool->opps[secondary_pipe->pipe_idx];
1665 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1666 if (secondary_pipe->stream->timing.flags.DSC == 1) {
1667 acquire_dsc(res_ctx, pool, &secondary_pipe->stream_res.dsc);
1668 ASSERT(secondary_pipe->stream_res.dsc);
1669 if (secondary_pipe->stream_res.dsc == NULL)
1674 ASSERT(primary_pipe->plane_state);
1675 resource_build_scaling_params(primary_pipe);
1676 resource_build_scaling_params(secondary_pipe);
1682 void dcn20_populate_dml_writeback_from_context(
1683 struct dc *dc, struct resource_context *res_ctx, display_e2e_pipe_params_st *pipes)
1687 for (i = 0, pipe_cnt = 0; i < dc->res_pool->pipe_count; i++) {
1688 struct dc_writeback_info *wb_info = &res_ctx->pipe_ctx[i].stream->writeback_info[0];
1690 if (!res_ctx->pipe_ctx[i].stream)
1693 /* Set writeback information */
1694 pipes[pipe_cnt].dout.wb_enable = (wb_info->wb_enabled == true) ? 1 : 0;
1695 pipes[pipe_cnt].dout.num_active_wb++;
1696 pipes[pipe_cnt].dout.wb.wb_src_height = wb_info->dwb_params.cnv_params.crop_height;
1697 pipes[pipe_cnt].dout.wb.wb_src_width = wb_info->dwb_params.cnv_params.crop_width;
1698 pipes[pipe_cnt].dout.wb.wb_dst_width = wb_info->dwb_params.dest_width;
1699 pipes[pipe_cnt].dout.wb.wb_dst_height = wb_info->dwb_params.dest_height;
1700 pipes[pipe_cnt].dout.wb.wb_htaps_luma = 1;
1701 pipes[pipe_cnt].dout.wb.wb_vtaps_luma = 1;
1702 pipes[pipe_cnt].dout.wb.wb_htaps_chroma = wb_info->dwb_params.scaler_taps.h_taps_c;
1703 pipes[pipe_cnt].dout.wb.wb_vtaps_chroma = wb_info->dwb_params.scaler_taps.v_taps_c;
1704 pipes[pipe_cnt].dout.wb.wb_hratio = 1.0;
1705 pipes[pipe_cnt].dout.wb.wb_vratio = 1.0;
1706 if (wb_info->dwb_params.out_format == dwb_scaler_mode_yuv420) {
1707 if (wb_info->dwb_params.output_depth == DWB_OUTPUT_PIXEL_DEPTH_8BPC)
1708 pipes[pipe_cnt].dout.wb.wb_pixel_format = dm_420_8;
1710 pipes[pipe_cnt].dout.wb.wb_pixel_format = dm_420_10;
1712 pipes[pipe_cnt].dout.wb.wb_pixel_format = dm_444_32;
1719 int dcn20_populate_dml_pipes_from_context(
1720 struct dc *dc, struct resource_context *res_ctx, display_e2e_pipe_params_st *pipes)
1723 bool synchronized_vblank = true;
1725 for (i = 0, pipe_cnt = -1; i < dc->res_pool->pipe_count; i++) {
1726 if (!res_ctx->pipe_ctx[i].stream)
1733 if (!resource_are_streams_timing_synchronizable(
1734 res_ctx->pipe_ctx[pipe_cnt].stream,
1735 res_ctx->pipe_ctx[i].stream)) {
1736 synchronized_vblank = false;
1741 for (i = 0, pipe_cnt = 0; i < dc->res_pool->pipe_count; i++) {
1742 struct dc_crtc_timing *timing = &res_ctx->pipe_ctx[i].stream->timing;
1745 if (!res_ctx->pipe_ctx[i].stream)
1748 pipes[pipe_cnt].pipe.src.dynamic_metadata_enable = 0;
1749 pipes[pipe_cnt].pipe.src.dcc = 0;
1750 pipes[pipe_cnt].pipe.src.vm = 0;*/
1752 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
1753 pipes[pipe_cnt].dout.dsc_enable = res_ctx->pipe_ctx[i].stream->timing.flags.DSC;
1754 /* todo: rotation?*/
1755 pipes[pipe_cnt].dout.dsc_slices = res_ctx->pipe_ctx[i].stream->timing.dsc_cfg.num_slices_h;
1757 if (res_ctx->pipe_ctx[i].stream->use_dynamic_meta) {
1758 pipes[pipe_cnt].pipe.src.dynamic_metadata_enable = true;
1760 pipes[pipe_cnt].pipe.src.dynamic_metadata_lines_before_active =
1761 (timing->v_total - timing->v_addressable
1762 - timing->v_border_top - timing->v_border_bottom) / 2;
1763 /* 36 bytes dp, 32 hdmi */
1764 pipes[pipe_cnt].pipe.src.dynamic_metadata_xmit_bytes =
1765 dc_is_dp_signal(res_ctx->pipe_ctx[i].stream->signal) ? 36 : 32;
1767 pipes[pipe_cnt].pipe.src.dcc = false;
1768 pipes[pipe_cnt].pipe.src.dcc_rate = 1;
1769 pipes[pipe_cnt].pipe.dest.synchronized_vblank_all_planes = synchronized_vblank;
1770 pipes[pipe_cnt].pipe.dest.hblank_start = timing->h_total - timing->h_front_porch;
1771 pipes[pipe_cnt].pipe.dest.hblank_end = pipes[pipe_cnt].pipe.dest.hblank_start
1772 - timing->h_addressable
1773 - timing->h_border_left
1774 - timing->h_border_right;
1775 pipes[pipe_cnt].pipe.dest.vblank_start = timing->v_total - timing->v_front_porch;
1776 pipes[pipe_cnt].pipe.dest.vblank_end = pipes[pipe_cnt].pipe.dest.vblank_start
1777 - timing->v_addressable
1778 - timing->v_border_top
1779 - timing->v_border_bottom;
1780 pipes[pipe_cnt].pipe.dest.htotal = timing->h_total;
1781 pipes[pipe_cnt].pipe.dest.vtotal = timing->v_total;
1782 pipes[pipe_cnt].pipe.dest.hactive = timing->h_addressable;
1783 pipes[pipe_cnt].pipe.dest.vactive = timing->v_addressable;
1784 pipes[pipe_cnt].pipe.dest.interlaced = timing->flags.INTERLACE;
1785 pipes[pipe_cnt].pipe.dest.pixel_rate_mhz = timing->pix_clk_100hz/10000.0;
1786 if (timing->timing_3d_format == TIMING_3D_FORMAT_HW_FRAME_PACKING)
1787 pipes[pipe_cnt].pipe.dest.pixel_rate_mhz *= 2;
1788 pipes[pipe_cnt].pipe.dest.otg_inst = res_ctx->pipe_ctx[i].stream_res.tg->inst;
1789 pipes[pipe_cnt].dout.dp_lanes = 4;
1790 pipes[pipe_cnt].pipe.dest.vtotal_min = res_ctx->pipe_ctx[i].stream->adjust.v_total_min;
1791 pipes[pipe_cnt].pipe.dest.vtotal_max = res_ctx->pipe_ctx[i].stream->adjust.v_total_max;
1793 switch (res_ctx->pipe_ctx[i].stream->signal) {
1794 case SIGNAL_TYPE_DISPLAY_PORT_MST:
1795 case SIGNAL_TYPE_DISPLAY_PORT:
1796 pipes[pipe_cnt].dout.output_type = dm_dp;
1798 case SIGNAL_TYPE_EDP:
1799 pipes[pipe_cnt].dout.output_type = dm_edp;
1801 case SIGNAL_TYPE_HDMI_TYPE_A:
1802 case SIGNAL_TYPE_DVI_SINGLE_LINK:
1803 case SIGNAL_TYPE_DVI_DUAL_LINK:
1804 pipes[pipe_cnt].dout.output_type = dm_hdmi;
1807 /* In case there is no signal, set dp with 4 lanes to allow max config */
1808 pipes[pipe_cnt].dout.output_type = dm_dp;
1809 pipes[pipe_cnt].dout.dp_lanes = 4;
1812 switch (res_ctx->pipe_ctx[i].stream->timing.display_color_depth) {
1813 case COLOR_DEPTH_666:
1816 case COLOR_DEPTH_888:
1819 case COLOR_DEPTH_101010:
1822 case COLOR_DEPTH_121212:
1825 case COLOR_DEPTH_141414:
1828 case COLOR_DEPTH_161616:
1831 #ifdef CONFIG_DRM_AMD_DC_DCN2_0
1832 case COLOR_DEPTH_999:
1835 case COLOR_DEPTH_111111:
1845 switch (res_ctx->pipe_ctx[i].stream->timing.pixel_encoding) {
1846 case PIXEL_ENCODING_RGB:
1847 case PIXEL_ENCODING_YCBCR444:
1848 pipes[pipe_cnt].dout.output_format = dm_444;
1849 pipes[pipe_cnt].dout.output_bpp = output_bpc * 3;
1851 case PIXEL_ENCODING_YCBCR420:
1852 pipes[pipe_cnt].dout.output_format = dm_420;
1853 pipes[pipe_cnt].dout.output_bpp = (output_bpc * 3) / 2;
1855 case PIXEL_ENCODING_YCBCR422:
1856 if (true) /* todo */
1857 pipes[pipe_cnt].dout.output_format = dm_s422;
1859 pipes[pipe_cnt].dout.output_format = dm_n422;
1860 pipes[pipe_cnt].dout.output_bpp = output_bpc * 2;
1863 pipes[pipe_cnt].dout.output_format = dm_444;
1864 pipes[pipe_cnt].dout.output_bpp = output_bpc * 3;
1866 pipes[pipe_cnt].pipe.src.hsplit_grp = res_ctx->pipe_ctx[i].pipe_idx;
1867 if (res_ctx->pipe_ctx[i].top_pipe && res_ctx->pipe_ctx[i].top_pipe->plane_state
1868 == res_ctx->pipe_ctx[i].plane_state)
1869 pipes[pipe_cnt].pipe.src.hsplit_grp = res_ctx->pipe_ctx[i].top_pipe->pipe_idx;
1871 /* todo: default max for now, until there is logic reflecting this in dc*/
1872 pipes[pipe_cnt].dout.output_bpc = 12;
1874 * Use max cursor settings for calculations to minimize
1875 * bw calculations due to cursor on/off
1877 pipes[pipe_cnt].pipe.src.num_cursors = 2;
1878 pipes[pipe_cnt].pipe.src.cur0_src_width = 256;
1879 pipes[pipe_cnt].pipe.src.cur0_bpp = dm_cur_32bit;
1880 pipes[pipe_cnt].pipe.src.cur1_src_width = 256;
1881 pipes[pipe_cnt].pipe.src.cur1_bpp = dm_cur_32bit;
1883 if (!res_ctx->pipe_ctx[i].plane_state) {
1884 pipes[pipe_cnt].pipe.src.source_scan = dm_horz;
1885 pipes[pipe_cnt].pipe.src.sw_mode = dm_sw_linear;
1886 pipes[pipe_cnt].pipe.src.macro_tile_size = dm_64k_tile;
1887 pipes[pipe_cnt].pipe.src.viewport_width = timing->h_addressable;
1888 if (pipes[pipe_cnt].pipe.src.viewport_width > 1920)
1889 pipes[pipe_cnt].pipe.src.viewport_width = 1920;
1890 pipes[pipe_cnt].pipe.src.viewport_height = timing->v_addressable;
1891 if (pipes[pipe_cnt].pipe.src.viewport_height > 1080)
1892 pipes[pipe_cnt].pipe.src.viewport_height = 1080;
1893 pipes[pipe_cnt].pipe.src.data_pitch = ((pipes[pipe_cnt].pipe.src.viewport_width + 63) / 64) * 64; /* linear sw only */
1894 pipes[pipe_cnt].pipe.src.source_format = dm_444_32;
1895 pipes[pipe_cnt].pipe.dest.recout_width = pipes[pipe_cnt].pipe.src.viewport_width; /*vp_width/hratio*/
1896 pipes[pipe_cnt].pipe.dest.recout_height = pipes[pipe_cnt].pipe.src.viewport_height; /*vp_height/vratio*/
1897 pipes[pipe_cnt].pipe.dest.full_recout_width = pipes[pipe_cnt].pipe.dest.recout_width; /*when is_hsplit != 1*/
1898 pipes[pipe_cnt].pipe.dest.full_recout_height = pipes[pipe_cnt].pipe.dest.recout_height; /*when is_hsplit != 1*/
1899 pipes[pipe_cnt].pipe.scale_ratio_depth.lb_depth = dm_lb_16;
1900 pipes[pipe_cnt].pipe.scale_ratio_depth.hscl_ratio = 1.0;
1901 pipes[pipe_cnt].pipe.scale_ratio_depth.vscl_ratio = 1.0;
1902 pipes[pipe_cnt].pipe.scale_ratio_depth.scl_enable = 0; /*Lb only or Full scl*/
1903 pipes[pipe_cnt].pipe.scale_taps.htaps = 1;
1904 pipes[pipe_cnt].pipe.scale_taps.vtaps = 1;
1905 pipes[pipe_cnt].pipe.src.is_hsplit = 0;
1906 pipes[pipe_cnt].pipe.dest.odm_combine = 0;
1907 pipes[pipe_cnt].pipe.dest.vtotal_min = timing->v_total;
1908 pipes[pipe_cnt].pipe.dest.vtotal_max = timing->v_total;
1910 struct dc_plane_state *pln = res_ctx->pipe_ctx[i].plane_state;
1911 struct scaler_data *scl = &res_ctx->pipe_ctx[i].plane_res.scl_data;
1913 pipes[pipe_cnt].pipe.src.immediate_flip = pln->flip_immediate;
1914 pipes[pipe_cnt].pipe.src.is_hsplit = (res_ctx->pipe_ctx[i].bottom_pipe
1915 && res_ctx->pipe_ctx[i].bottom_pipe->plane_state == pln)
1916 || (res_ctx->pipe_ctx[i].top_pipe
1917 && res_ctx->pipe_ctx[i].top_pipe->plane_state == pln);
1918 pipes[pipe_cnt].pipe.dest.odm_combine = (res_ctx->pipe_ctx[i].bottom_pipe
1919 && res_ctx->pipe_ctx[i].bottom_pipe->plane_state == pln
1920 && res_ctx->pipe_ctx[i].bottom_pipe->stream_res.opp
1921 != res_ctx->pipe_ctx[i].stream_res.opp)
1922 || (res_ctx->pipe_ctx[i].top_pipe
1923 && res_ctx->pipe_ctx[i].top_pipe->plane_state == pln
1924 && res_ctx->pipe_ctx[i].top_pipe->stream_res.opp
1925 != res_ctx->pipe_ctx[i].stream_res.opp);
1926 pipes[pipe_cnt].pipe.src.source_scan = pln->rotation == ROTATION_ANGLE_90
1927 || pln->rotation == ROTATION_ANGLE_270 ? dm_vert : dm_horz;
1928 pipes[pipe_cnt].pipe.src.viewport_y_y = scl->viewport.y;
1929 pipes[pipe_cnt].pipe.src.viewport_y_c = scl->viewport_c.y;
1930 pipes[pipe_cnt].pipe.src.viewport_width = scl->viewport.width;
1931 pipes[pipe_cnt].pipe.src.viewport_width_c = scl->viewport_c.width;
1932 pipes[pipe_cnt].pipe.src.viewport_height = scl->viewport.height;
1933 pipes[pipe_cnt].pipe.src.viewport_height_c = scl->viewport_c.height;
1934 if (pln->format >= SURFACE_PIXEL_FORMAT_VIDEO_BEGIN) {
1935 pipes[pipe_cnt].pipe.src.data_pitch = pln->plane_size.surface_pitch;
1936 pipes[pipe_cnt].pipe.src.data_pitch_c = pln->plane_size.chroma_pitch;
1937 pipes[pipe_cnt].pipe.src.meta_pitch = pln->dcc.meta_pitch;
1938 pipes[pipe_cnt].pipe.src.meta_pitch_c = pln->dcc.meta_pitch_c;
1940 pipes[pipe_cnt].pipe.src.data_pitch = pln->plane_size.surface_pitch;
1941 pipes[pipe_cnt].pipe.src.meta_pitch = pln->dcc.meta_pitch;
1943 pipes[pipe_cnt].pipe.src.dcc = pln->dcc.enable;
1944 pipes[pipe_cnt].pipe.dest.recout_width = scl->recout.width;
1945 pipes[pipe_cnt].pipe.dest.recout_height = scl->recout.height;
1946 pipes[pipe_cnt].pipe.dest.full_recout_width = scl->recout.width;
1947 pipes[pipe_cnt].pipe.dest.full_recout_height = scl->recout.height;
1948 if (res_ctx->pipe_ctx[i].bottom_pipe && res_ctx->pipe_ctx[i].bottom_pipe->plane_state == pln) {
1949 pipes[pipe_cnt].pipe.dest.full_recout_width +=
1950 res_ctx->pipe_ctx[i].bottom_pipe->plane_res.scl_data.recout.width;
1951 pipes[pipe_cnt].pipe.dest.full_recout_height +=
1952 res_ctx->pipe_ctx[i].bottom_pipe->plane_res.scl_data.recout.height;
1953 } else if (res_ctx->pipe_ctx[i].top_pipe && res_ctx->pipe_ctx[i].top_pipe->plane_state == pln) {
1954 pipes[pipe_cnt].pipe.dest.full_recout_width +=
1955 res_ctx->pipe_ctx[i].top_pipe->plane_res.scl_data.recout.width;
1956 pipes[pipe_cnt].pipe.dest.full_recout_height +=
1957 res_ctx->pipe_ctx[i].top_pipe->plane_res.scl_data.recout.height;
1960 pipes[pipe_cnt].pipe.scale_ratio_depth.lb_depth = dm_lb_16;
1961 pipes[pipe_cnt].pipe.scale_ratio_depth.hscl_ratio = (double) scl->ratios.horz.value / (1ULL<<32);
1962 pipes[pipe_cnt].pipe.scale_ratio_depth.hscl_ratio_c = (double) scl->ratios.horz_c.value / (1ULL<<32);
1963 pipes[pipe_cnt].pipe.scale_ratio_depth.vscl_ratio = (double) scl->ratios.vert.value / (1ULL<<32);
1964 pipes[pipe_cnt].pipe.scale_ratio_depth.vscl_ratio_c = (double) scl->ratios.vert_c.value / (1ULL<<32);
1965 pipes[pipe_cnt].pipe.scale_ratio_depth.scl_enable =
1966 scl->ratios.vert.value != dc_fixpt_one.value
1967 || scl->ratios.horz.value != dc_fixpt_one.value
1968 || scl->ratios.vert_c.value != dc_fixpt_one.value
1969 || scl->ratios.horz_c.value != dc_fixpt_one.value /*Lb only or Full scl*/
1970 || dc->debug.always_scale; /*support always scale*/
1971 pipes[pipe_cnt].pipe.scale_taps.htaps = scl->taps.h_taps;
1972 pipes[pipe_cnt].pipe.scale_taps.htaps_c = scl->taps.h_taps_c;
1973 pipes[pipe_cnt].pipe.scale_taps.vtaps = scl->taps.v_taps;
1974 pipes[pipe_cnt].pipe.scale_taps.vtaps_c = scl->taps.v_taps_c;
1976 pipes[pipe_cnt].pipe.src.macro_tile_size =
1977 swizzle_mode_to_macro_tile_size(pln->tiling_info.gfx9.swizzle);
1978 swizzle_to_dml_params(pln->tiling_info.gfx9.swizzle,
1979 &pipes[pipe_cnt].pipe.src.sw_mode);
1981 switch (pln->format) {
1982 case SURFACE_PIXEL_FORMAT_VIDEO_420_YCbCr:
1983 case SURFACE_PIXEL_FORMAT_VIDEO_420_YCrCb:
1984 pipes[pipe_cnt].pipe.src.source_format = dm_420_8;
1986 case SURFACE_PIXEL_FORMAT_VIDEO_420_10bpc_YCbCr:
1987 case SURFACE_PIXEL_FORMAT_VIDEO_420_10bpc_YCrCb:
1988 pipes[pipe_cnt].pipe.src.source_format = dm_420_10;
1990 case SURFACE_PIXEL_FORMAT_GRPH_ARGB16161616:
1991 case SURFACE_PIXEL_FORMAT_GRPH_ARGB16161616F:
1992 case SURFACE_PIXEL_FORMAT_GRPH_ABGR16161616F:
1993 pipes[pipe_cnt].pipe.src.source_format = dm_444_64;
1995 case SURFACE_PIXEL_FORMAT_GRPH_ARGB1555:
1996 case SURFACE_PIXEL_FORMAT_GRPH_RGB565:
1997 pipes[pipe_cnt].pipe.src.source_format = dm_444_16;
1999 case SURFACE_PIXEL_FORMAT_GRPH_PALETA_256_COLORS:
2000 pipes[pipe_cnt].pipe.src.source_format = dm_444_8;
2003 pipes[pipe_cnt].pipe.src.source_format = dm_444_32;
2011 /* populate writeback information */
2012 dc->res_pool->funcs->populate_dml_writeback_from_context(dc, res_ctx, pipes);
2017 unsigned int dcn20_calc_max_scaled_time(
2018 unsigned int time_per_pixel,
2019 enum mmhubbub_wbif_mode mode,
2020 unsigned int urgent_watermark)
2022 unsigned int time_per_byte = 0;
2023 unsigned int total_y_free_entry = 0x200; /* two memory piece for luma */
2024 unsigned int total_c_free_entry = 0x140; /* two memory piece for chroma */
2025 unsigned int small_free_entry, max_free_entry;
2026 unsigned int buf_lh_capability;
2027 unsigned int max_scaled_time;
2029 if (mode == PACKED_444) /* packed mode */
2030 time_per_byte = time_per_pixel/4;
2031 else if (mode == PLANAR_420_8BPC)
2032 time_per_byte = time_per_pixel;
2033 else if (mode == PLANAR_420_10BPC) /* p010 */
2034 time_per_byte = time_per_pixel * 819/1024;
2036 if (time_per_byte == 0)
2039 small_free_entry = (total_y_free_entry > total_c_free_entry) ? total_c_free_entry : total_y_free_entry;
2040 max_free_entry = (mode == PACKED_444) ? total_y_free_entry + total_c_free_entry : small_free_entry;
2041 buf_lh_capability = max_free_entry*time_per_byte*32/16; /* there is 4bit fraction */
2042 max_scaled_time = buf_lh_capability - urgent_watermark;
2043 return max_scaled_time;
2046 void dcn20_set_mcif_arb_params(
2048 struct dc_state *context,
2049 display_e2e_pipe_params_st *pipes,
2052 enum mmhubbub_wbif_mode wbif_mode;
2053 struct mcif_arb_params *wb_arb_params;
2054 int i, j, k, dwb_pipe;
2056 /* Writeback MCIF_WB arbitration parameters */
2058 for (i = 0; i < dc->res_pool->pipe_count; i++) {
2060 if (!context->res_ctx.pipe_ctx[i].stream)
2063 for (j = 0; j < MAX_DWB_PIPES; j++) {
2064 if (context->res_ctx.pipe_ctx[i].stream->writeback_info[j].wb_enabled == false)
2067 //wb_arb_params = &context->res_ctx.pipe_ctx[i].stream->writeback_info[j].mcif_arb_params;
2068 wb_arb_params = &context->bw_ctx.bw.dcn.bw_writeback.mcif_wb_arb[dwb_pipe];
2070 if (context->res_ctx.pipe_ctx[i].stream->writeback_info[j].dwb_params.out_format == dwb_scaler_mode_yuv420) {
2071 if (context->res_ctx.pipe_ctx[i].stream->writeback_info[j].dwb_params.output_depth == DWB_OUTPUT_PIXEL_DEPTH_8BPC)
2072 wbif_mode = PLANAR_420_8BPC;
2074 wbif_mode = PLANAR_420_10BPC;
2076 wbif_mode = PACKED_444;
2078 for (k = 0; k < sizeof(wb_arb_params->cli_watermark)/sizeof(wb_arb_params->cli_watermark[0]); k++) {
2079 wb_arb_params->cli_watermark[k] = get_wm_writeback_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2080 wb_arb_params->pstate_watermark[k] = get_wm_writeback_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2082 wb_arb_params->time_per_pixel = 16.0 / context->res_ctx.pipe_ctx[i].stream->phy_pix_clk; /* 4 bit fraction, ms */
2083 wb_arb_params->slice_lines = 32;
2084 wb_arb_params->arbitration_slice = 2;
2085 wb_arb_params->max_scaled_time = dcn20_calc_max_scaled_time(wb_arb_params->time_per_pixel,
2087 wb_arb_params->cli_watermark[0]); /* assume 4 watermark sets have the same value */
2091 if (dwb_pipe >= MAX_DWB_PIPES)
2094 if (dwb_pipe >= MAX_DWB_PIPES)
2099 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
2100 static bool dcn20_validate_dsc(struct dc *dc, struct dc_state *new_ctx)
2104 /* Validate DSC config, dsc count validation is already done */
2105 for (i = 0; i < dc->res_pool->pipe_count; i++) {
2106 struct pipe_ctx *pipe_ctx = &new_ctx->res_ctx.pipe_ctx[i];
2107 struct dc_stream_state *stream = pipe_ctx->stream;
2108 struct dsc_config dsc_cfg;
2110 /* Only need to validate top pipe */
2111 if (pipe_ctx->top_pipe || !stream || !stream->timing.flags.DSC)
2114 dsc_cfg.pic_width = stream->timing.h_addressable + stream->timing.h_border_left
2115 + stream->timing.h_border_right;
2116 dsc_cfg.pic_height = stream->timing.v_addressable + stream->timing.v_border_top
2117 + stream->timing.v_border_bottom;
2118 if (dc_res_get_odm_bottom_pipe(pipe_ctx))
2119 dsc_cfg.pic_width /= 2;
2120 dsc_cfg.pixel_encoding = stream->timing.pixel_encoding;
2121 dsc_cfg.color_depth = stream->timing.display_color_depth;
2122 dsc_cfg.dc_dsc_cfg = stream->timing.dsc_cfg;
2124 if (!pipe_ctx->stream_res.dsc->funcs->dsc_validate_stream(pipe_ctx->stream_res.dsc, &dsc_cfg))
2131 bool dcn20_fast_validate_bw(
2133 struct dc_state *context,
2134 display_e2e_pipe_params_st *pipes,
2136 int *pipe_split_from,
2141 int pipe_cnt, i, pipe_idx, vlevel, vlevel_unsplit;
2142 bool odm_capable = context->bw_ctx.dml.ip.odm_capable;
2143 bool force_split = false;
2144 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
2145 bool failed_non_odm_dsc = false;
2147 int split_threshold = dc->res_pool->pipe_count / 2;
2148 bool avoid_split = dc->debug.pipe_split_policy != MPC_SPLIT_DYNAMIC;
2155 for (i = 0; i < dc->res_pool->pipe_count; i++) {
2156 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i];
2157 struct pipe_ctx *hsplit_pipe = pipe->bottom_pipe;
2159 if (!hsplit_pipe || hsplit_pipe->plane_state != pipe->plane_state)
2162 /* merge previously split pipe since mode support needs to make the decision */
2163 pipe->bottom_pipe = hsplit_pipe->bottom_pipe;
2164 if (hsplit_pipe->bottom_pipe)
2165 hsplit_pipe->bottom_pipe->top_pipe = pipe;
2166 hsplit_pipe->plane_state = NULL;
2167 hsplit_pipe->stream = NULL;
2168 hsplit_pipe->top_pipe = NULL;
2169 hsplit_pipe->bottom_pipe = NULL;
2170 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
2171 if (hsplit_pipe->stream_res.dsc && hsplit_pipe->stream_res.dsc != pipe->stream_res.dsc)
2172 release_dsc(&context->res_ctx, dc->res_pool, &hsplit_pipe->stream_res.dsc);
2174 /* Clear plane_res and stream_res */
2175 memset(&hsplit_pipe->plane_res, 0, sizeof(hsplit_pipe->plane_res));
2176 memset(&hsplit_pipe->stream_res, 0, sizeof(hsplit_pipe->stream_res));
2177 if (pipe->plane_state)
2178 resource_build_scaling_params(pipe);
2181 if (dc->res_pool->funcs->populate_dml_pipes)
2182 pipe_cnt = dc->res_pool->funcs->populate_dml_pipes(dc,
2183 &context->res_ctx, pipes);
2185 pipe_cnt = dcn20_populate_dml_pipes_from_context(dc,
2186 &context->res_ctx, pipes);
2188 *pipe_cnt_out = pipe_cnt;
2195 context->bw_ctx.dml.ip.odm_capable = 0;
2197 vlevel = dml_get_voltage_level(&context->bw_ctx.dml, pipes, pipe_cnt);
2199 context->bw_ctx.dml.ip.odm_capable = odm_capable;
2201 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
2202 /* 1 dsc per stream dsc validation */
2203 if (vlevel <= context->bw_ctx.dml.soc.num_states)
2204 if (!dcn20_validate_dsc(dc, context)) {
2205 failed_non_odm_dsc = true;
2206 vlevel = context->bw_ctx.dml.soc.num_states + 1;
2210 if (vlevel > context->bw_ctx.dml.soc.num_states && odm_capable)
2211 vlevel = dml_get_voltage_level(&context->bw_ctx.dml, pipes, pipe_cnt);
2213 if (vlevel > context->bw_ctx.dml.soc.num_states)
2216 if ((context->stream_count > split_threshold && dc->current_state->stream_count <= split_threshold)
2217 || (context->stream_count <= split_threshold && dc->current_state->stream_count > split_threshold))
2218 context->commit_hints.full_update_needed = true;
2220 /*initialize pipe_just_split_from to invalid idx*/
2221 for (i = 0; i < MAX_PIPES; i++)
2222 pipe_split_from[i] = -1;
2224 /* Single display only conditionals get set here */
2225 for (i = 0; i < dc->res_pool->pipe_count; i++) {
2226 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i];
2227 bool exit_loop = false;
2229 if (!pipe->stream || pipe->top_pipe)
2232 if (dc->debug.force_single_disp_pipe_split) {
2236 force_split = false;
2240 if (dc->debug.pipe_split_policy == MPC_SPLIT_AVOID_MULT_DISP) {
2242 avoid_split = false;
2252 if (context->stream_count > split_threshold)
2255 vlevel_unsplit = vlevel;
2256 for (i = 0, pipe_idx = 0; i < dc->res_pool->pipe_count; i++) {
2257 if (!context->res_ctx.pipe_ctx[i].stream)
2259 for (; vlevel_unsplit <= context->bw_ctx.dml.soc.num_states; vlevel_unsplit++)
2260 if (context->bw_ctx.dml.vba.NoOfDPP[vlevel_unsplit][0][pipe_idx] == 1)
2265 for (i = 0, pipe_idx = -1; i < dc->res_pool->pipe_count; i++) {
2266 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i];
2267 struct pipe_ctx *hsplit_pipe = pipe->bottom_pipe;
2268 bool need_split = true;
2271 if (!pipe->stream || pipe_split_from[i] >= 0)
2276 if (dc->debug.force_odm_combine & (1 << pipe->stream_res.tg->inst)) {
2278 context->bw_ctx.dml.vba.ODMCombineEnabled[pipe_idx] = true;
2279 context->bw_ctx.dml.vba.ODMCombineEnablePerState[vlevel][pipe_idx] = true;
2281 if (force_split && context->bw_ctx.dml.vba.NoOfDPP[vlevel][context->bw_ctx.dml.vba.maxMpcComb][pipe_idx] == 1)
2282 context->bw_ctx.dml.vba.RequiredDPPCLK[vlevel][context->bw_ctx.dml.vba.maxMpcComb][pipe_idx] /= 2;
2283 if (!pipe->top_pipe && !pipe->plane_state && context->bw_ctx.dml.vba.ODMCombineEnabled[pipe_idx]) {
2284 hsplit_pipe = find_idle_secondary_pipe(&context->res_ctx, dc->res_pool, pipe);
2285 ASSERT(hsplit_pipe);
2286 if (!dcn20_split_stream_for_combine(
2287 &context->res_ctx, dc->res_pool,
2291 pipe_split_from[hsplit_pipe->pipe_idx] = pipe_idx;
2292 dcn20_build_mapped_resource(dc, context, pipe->stream);
2295 if (!pipe->plane_state)
2297 /* Skip 2nd half of already split pipe */
2298 if (pipe->top_pipe && pipe->plane_state == pipe->top_pipe->plane_state)
2301 need_split3d = ((pipe->stream->view_format ==
2302 VIEW_3D_FORMAT_SIDE_BY_SIDE ||
2303 pipe->stream->view_format ==
2304 VIEW_3D_FORMAT_TOP_AND_BOTTOM) &&
2305 (pipe->stream->timing.timing_3d_format ==
2306 TIMING_3D_FORMAT_TOP_AND_BOTTOM ||
2307 pipe->stream->timing.timing_3d_format ==
2308 TIMING_3D_FORMAT_SIDE_BY_SIDE));
2310 if (avoid_split && vlevel_unsplit <= context->bw_ctx.dml.soc.num_states && !force_split && !need_split3d) {
2312 vlevel = vlevel_unsplit;
2313 context->bw_ctx.dml.vba.maxMpcComb = 0;
2315 need_split = context->bw_ctx.dml.vba.NoOfDPP[vlevel][context->bw_ctx.dml.vba.maxMpcComb][pipe_idx] == 2;
2317 /* We do not support mpo + odm at the moment */
2318 if (hsplit_pipe && hsplit_pipe->plane_state != pipe->plane_state
2319 && context->bw_ctx.dml.vba.ODMCombineEnabled[pipe_idx])
2322 if (need_split3d || need_split || force_split) {
2323 if (!hsplit_pipe || hsplit_pipe->plane_state != pipe->plane_state) {
2324 /* pipe not split previously needs split */
2325 hsplit_pipe = find_idle_secondary_pipe(&context->res_ctx, dc->res_pool, pipe);
2326 ASSERT(hsplit_pipe || force_split);
2330 if (!dcn20_split_stream_for_combine(
2331 &context->res_ctx, dc->res_pool,
2333 context->bw_ctx.dml.vba.ODMCombineEnabled[pipe_idx]))
2335 pipe_split_from[hsplit_pipe->pipe_idx] = pipe_idx;
2337 } else if (hsplit_pipe && hsplit_pipe->plane_state == pipe->plane_state) {
2338 /* merge should already have been done */
2342 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
2343 /* Actual dsc count per stream dsc validation*/
2344 if (failed_non_odm_dsc && !dcn20_validate_dsc(dc, context)) {
2345 context->bw_ctx.dml.vba.ValidationStatus[context->bw_ctx.dml.vba.soc.num_states] =
2346 DML_FAIL_DSC_VALIDATION_FAILURE;
2351 *vlevel_out = vlevel;
2363 void dcn20_calculate_wm(
2364 struct dc *dc, struct dc_state *context,
2365 display_e2e_pipe_params_st *pipes,
2367 int *pipe_split_from,
2370 int pipe_cnt, i, pipe_idx;
2372 for (i = 0, pipe_idx = 0, pipe_cnt = 0; i < dc->res_pool->pipe_count; i++) {
2373 if (!context->res_ctx.pipe_ctx[i].stream)
2376 pipes[pipe_cnt].clks_cfg.refclk_mhz = dc->res_pool->ref_clocks.dchub_ref_clock_inKhz / 1000.0;
2377 pipes[pipe_cnt].clks_cfg.dispclk_mhz = context->bw_ctx.dml.vba.RequiredDISPCLK[vlevel][context->bw_ctx.dml.vba.maxMpcComb];
2379 if (pipe_split_from[i] < 0) {
2380 pipes[pipe_cnt].clks_cfg.dppclk_mhz =
2381 context->bw_ctx.dml.vba.RequiredDPPCLK[vlevel][context->bw_ctx.dml.vba.maxMpcComb][pipe_idx];
2382 if (context->bw_ctx.dml.vba.BlendingAndTiming[pipe_idx] == pipe_idx)
2383 pipes[pipe_cnt].pipe.dest.odm_combine =
2384 context->bw_ctx.dml.vba.ODMCombineEnablePerState[vlevel][pipe_idx];
2386 pipes[pipe_cnt].pipe.dest.odm_combine = 0;
2389 pipes[pipe_cnt].clks_cfg.dppclk_mhz =
2390 context->bw_ctx.dml.vba.RequiredDPPCLK[vlevel][context->bw_ctx.dml.vba.maxMpcComb][pipe_split_from[i]];
2391 if (context->bw_ctx.dml.vba.BlendingAndTiming[pipe_split_from[i]] == pipe_split_from[i])
2392 pipes[pipe_cnt].pipe.dest.odm_combine =
2393 context->bw_ctx.dml.vba.ODMCombineEnablePerState[vlevel][pipe_split_from[i]];
2395 pipes[pipe_cnt].pipe.dest.odm_combine = 0;
2398 if (dc->config.forced_clocks) {
2399 pipes[pipe_cnt].clks_cfg.dispclk_mhz = context->bw_ctx.dml.soc.clock_limits[0].dispclk_mhz;
2400 pipes[pipe_cnt].clks_cfg.dppclk_mhz = context->bw_ctx.dml.soc.clock_limits[0].dppclk_mhz;
2402 if (dc->debug.min_disp_clk_khz > pipes[pipe_cnt].clks_cfg.dispclk_mhz * 1000)
2403 pipes[pipe_cnt].clks_cfg.dispclk_mhz = dc->debug.min_disp_clk_khz / 1000.0;
2404 if (dc->debug.min_dpp_clk_khz > pipes[pipe_cnt].clks_cfg.dppclk_mhz * 1000)
2405 pipes[pipe_cnt].clks_cfg.dppclk_mhz = dc->debug.min_dpp_clk_khz / 1000.0;
2410 if (pipe_cnt != pipe_idx) {
2411 if (dc->res_pool->funcs->populate_dml_pipes)
2412 pipe_cnt = dc->res_pool->funcs->populate_dml_pipes(dc,
2413 &context->res_ctx, pipes);
2415 pipe_cnt = dcn20_populate_dml_pipes_from_context(dc,
2416 &context->res_ctx, pipes);
2419 *out_pipe_cnt = pipe_cnt;
2421 pipes[0].clks_cfg.voltage = vlevel;
2422 pipes[0].clks_cfg.dcfclk_mhz = context->bw_ctx.dml.soc.clock_limits[vlevel].dcfclk_mhz;
2423 pipes[0].clks_cfg.socclk_mhz = context->bw_ctx.dml.soc.clock_limits[vlevel].socclk_mhz;
2425 /* only pipe 0 is read for voltage and dcf/soc clocks */
2427 pipes[0].clks_cfg.voltage = 1;
2428 pipes[0].clks_cfg.dcfclk_mhz = context->bw_ctx.dml.soc.clock_limits[1].dcfclk_mhz;
2429 pipes[0].clks_cfg.socclk_mhz = context->bw_ctx.dml.soc.clock_limits[1].socclk_mhz;
2431 context->bw_ctx.bw.dcn.watermarks.b.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2432 context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2433 context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2434 context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2435 context->bw_ctx.bw.dcn.watermarks.b.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2438 pipes[0].clks_cfg.voltage = 2;
2439 pipes[0].clks_cfg.dcfclk_mhz = context->bw_ctx.dml.soc.clock_limits[2].dcfclk_mhz;
2440 pipes[0].clks_cfg.socclk_mhz = context->bw_ctx.dml.soc.clock_limits[2].socclk_mhz;
2442 context->bw_ctx.bw.dcn.watermarks.c.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2443 context->bw_ctx.bw.dcn.watermarks.c.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2444 context->bw_ctx.bw.dcn.watermarks.c.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2445 context->bw_ctx.bw.dcn.watermarks.c.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2446 context->bw_ctx.bw.dcn.watermarks.c.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2449 pipes[0].clks_cfg.voltage = 3;
2450 pipes[0].clks_cfg.dcfclk_mhz = context->bw_ctx.dml.soc.clock_limits[2].dcfclk_mhz;
2451 pipes[0].clks_cfg.socclk_mhz = context->bw_ctx.dml.soc.clock_limits[2].socclk_mhz;
2453 context->bw_ctx.bw.dcn.watermarks.d.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2454 context->bw_ctx.bw.dcn.watermarks.d.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2455 context->bw_ctx.bw.dcn.watermarks.d.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2456 context->bw_ctx.bw.dcn.watermarks.d.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2457 context->bw_ctx.bw.dcn.watermarks.d.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2459 pipes[0].clks_cfg.voltage = vlevel;
2460 pipes[0].clks_cfg.dcfclk_mhz = context->bw_ctx.dml.soc.clock_limits[vlevel].dcfclk_mhz;
2461 pipes[0].clks_cfg.socclk_mhz = context->bw_ctx.dml.soc.clock_limits[vlevel].socclk_mhz;
2462 context->bw_ctx.bw.dcn.watermarks.a.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2463 context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2464 context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2465 context->bw_ctx.bw.dcn.watermarks.a.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2466 context->bw_ctx.bw.dcn.watermarks.a.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000;
2469 void dcn20_calculate_dlg_params(
2470 struct dc *dc, struct dc_state *context,
2471 display_e2e_pipe_params_st *pipes,
2477 /* Writeback MCIF_WB arbitration parameters */
2478 dc->res_pool->funcs->set_mcif_arb_params(dc, context, pipes, pipe_cnt);
2480 context->bw_ctx.bw.dcn.clk.dispclk_khz = context->bw_ctx.dml.vba.DISPCLK * 1000;
2481 context->bw_ctx.bw.dcn.clk.dcfclk_khz = context->bw_ctx.dml.vba.DCFCLK * 1000;
2482 context->bw_ctx.bw.dcn.clk.socclk_khz = context->bw_ctx.dml.vba.SOCCLK * 1000;
2483 context->bw_ctx.bw.dcn.clk.dramclk_khz = context->bw_ctx.dml.vba.DRAMSpeed * 1000 / 16;
2484 context->bw_ctx.bw.dcn.clk.dcfclk_deep_sleep_khz = context->bw_ctx.dml.vba.DCFCLKDeepSleep * 1000;
2485 context->bw_ctx.bw.dcn.clk.fclk_khz = 0;
2486 context->bw_ctx.bw.dcn.clk.p_state_change_support =
2487 context->bw_ctx.dml.vba.DRAMClockChangeSupport[vlevel][context->bw_ctx.dml.vba.maxMpcComb]
2488 != dm_dram_clock_change_unsupported;
2489 context->bw_ctx.bw.dcn.clk.dppclk_khz = 0;
2493 for (i = 0, pipe_idx = 0; i < dc->res_pool->pipe_count; i++) {
2494 if (!context->res_ctx.pipe_ctx[i].stream)
2496 pipes[pipe_idx].pipe.dest.vstartup_start = context->bw_ctx.dml.vba.VStartup[pipe_idx];
2497 pipes[pipe_idx].pipe.dest.vupdate_offset = context->bw_ctx.dml.vba.VUpdateOffsetPix[pipe_idx];
2498 pipes[pipe_idx].pipe.dest.vupdate_width = context->bw_ctx.dml.vba.VUpdateWidthPix[pipe_idx];
2499 pipes[pipe_idx].pipe.dest.vready_offset = context->bw_ctx.dml.vba.VReadyOffsetPix[pipe_idx];
2500 if (context->bw_ctx.bw.dcn.clk.dppclk_khz < pipes[pipe_idx].clks_cfg.dppclk_mhz * 1000)
2501 context->bw_ctx.bw.dcn.clk.dppclk_khz = pipes[pipe_idx].clks_cfg.dppclk_mhz * 1000;
2502 context->res_ctx.pipe_ctx[i].plane_res.bw.dppclk_khz =
2503 pipes[pipe_idx].clks_cfg.dppclk_mhz * 1000;
2504 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
2505 context->res_ctx.pipe_ctx[i].stream_res.dscclk_khz =
2506 context->bw_ctx.dml.vba.DSCCLK_calculated[pipe_idx] * 1000;
2508 context->res_ctx.pipe_ctx[i].pipe_dlg_param = pipes[pipe_idx].pipe.dest;
2511 /*save a original dppclock copy*/
2512 context->bw_ctx.bw.dcn.clk.bw_dppclk_khz = context->bw_ctx.bw.dcn.clk.dppclk_khz;
2513 context->bw_ctx.bw.dcn.clk.bw_dispclk_khz = context->bw_ctx.bw.dcn.clk.dispclk_khz;
2514 context->bw_ctx.bw.dcn.clk.max_supported_dppclk_khz = context->bw_ctx.dml.soc.clock_limits[vlevel].dppclk_mhz*1000;
2515 context->bw_ctx.bw.dcn.clk.max_supported_dispclk_khz = context->bw_ctx.dml.soc.clock_limits[vlevel].dispclk_mhz*1000;
2517 for (i = 0, pipe_idx = 0; i < dc->res_pool->pipe_count; i++) {
2518 bool cstate_en = context->bw_ctx.dml.vba.PrefetchMode[vlevel][context->bw_ctx.dml.vba.maxMpcComb] != 2;
2520 if (!context->res_ctx.pipe_ctx[i].stream)
2523 context->bw_ctx.dml.funcs.rq_dlg_get_dlg_reg(&context->bw_ctx.dml,
2524 &context->res_ctx.pipe_ctx[i].dlg_regs,
2525 &context->res_ctx.pipe_ctx[i].ttu_regs,
2530 context->bw_ctx.bw.dcn.clk.p_state_change_support,
2531 false, false, false);
2533 context->bw_ctx.dml.funcs.rq_dlg_get_rq_reg(&context->bw_ctx.dml,
2534 &context->res_ctx.pipe_ctx[i].rq_regs,
2535 pipes[pipe_idx].pipe);
2540 static bool dcn20_validate_bandwidth_internal(struct dc *dc, struct dc_state *context,
2545 BW_VAL_TRACE_SETUP();
2548 int pipe_split_from[MAX_PIPES];
2550 display_e2e_pipe_params_st *pipes = kzalloc(dc->res_pool->pipe_count * sizeof(display_e2e_pipe_params_st), GFP_KERNEL);
2551 DC_LOGGER_INIT(dc->ctx->logger);
2553 BW_VAL_TRACE_COUNT();
2555 out = dcn20_fast_validate_bw(dc, context, pipes, &pipe_cnt, pipe_split_from, &vlevel);
2563 BW_VAL_TRACE_END_VOLTAGE_LEVEL();
2565 if (fast_validate) {
2566 BW_VAL_TRACE_SKIP(fast);
2570 dcn20_calculate_wm(dc, context, pipes, &pipe_cnt, pipe_split_from, vlevel);
2571 dcn20_calculate_dlg_params(dc, context, pipes, pipe_cnt, vlevel);
2573 BW_VAL_TRACE_END_WATERMARKS();
2578 DC_LOG_WARNING("Mode Validation Warning: %s failed validation.\n",
2579 dml_get_status_message(context->bw_ctx.dml.vba.ValidationStatus[context->bw_ctx.dml.vba.soc.num_states]));
2581 BW_VAL_TRACE_SKIP(fail);
2587 BW_VAL_TRACE_FINISH();
2593 bool dcn20_validate_bandwidth(struct dc *dc, struct dc_state *context,
2596 bool voltage_supported = false;
2597 bool full_pstate_supported = false;
2598 bool dummy_pstate_supported = false;
2599 double p_state_latency_us = context->bw_ctx.dml.soc.dram_clock_change_latency_us;
2602 return dcn20_validate_bandwidth_internal(dc, context, true);
2605 // Best case, we support full UCLK switch latency
2606 voltage_supported = dcn20_validate_bandwidth_internal(dc, context, false);
2607 full_pstate_supported = context->bw_ctx.bw.dcn.clk.p_state_change_support;
2609 if (context->bw_ctx.dml.soc.dummy_pstate_latency_us == 0 ||
2610 (voltage_supported && full_pstate_supported)) {
2611 context->bw_ctx.bw.dcn.clk.p_state_change_support = true;
2612 goto restore_dml_state;
2615 // Fallback #1: Try to only support G6 temperature read latency
2616 context->bw_ctx.dml.soc.dram_clock_change_latency_us = context->bw_ctx.dml.soc.dummy_pstate_latency_us;
2618 voltage_supported = dcn20_validate_bandwidth_internal(dc, context, false);
2619 dummy_pstate_supported = context->bw_ctx.bw.dcn.clk.p_state_change_support;
2621 if (voltage_supported && dummy_pstate_supported) {
2622 context->bw_ctx.bw.dcn.clk.p_state_change_support = false;
2623 goto restore_dml_state;
2626 // Fallback #2: Retry with "new" DCN20 to support G6 temperature read latency
2627 memcpy (&context->bw_ctx.dml, &dc->work_arounds.alternate_dml, sizeof (struct display_mode_lib));
2628 context->bw_ctx.dml.soc.dram_clock_change_latency_us = context->bw_ctx.dml.soc.dummy_pstate_latency_us;
2630 voltage_supported = dcn20_validate_bandwidth_internal(dc, context, false);
2631 dummy_pstate_supported = context->bw_ctx.bw.dcn.clk.p_state_change_support;
2633 if (voltage_supported && dummy_pstate_supported) {
2634 context->bw_ctx.bw.dcn.clk.p_state_change_support = false;
2635 goto restore_dml_state;
2638 // ERROR: fallback #2 is supposed to always work.
2642 memcpy(&context->bw_ctx.dml, &dc->dml, sizeof(struct display_mode_lib));
2643 context->bw_ctx.dml.soc.dram_clock_change_latency_us = p_state_latency_us;
2645 return voltage_supported;
2648 struct pipe_ctx *dcn20_acquire_idle_pipe_for_layer(
2649 struct dc_state *state,
2650 const struct resource_pool *pool,
2651 struct dc_stream_state *stream)
2653 struct resource_context *res_ctx = &state->res_ctx;
2654 struct pipe_ctx *head_pipe = resource_get_head_pipe_for_stream(res_ctx, stream);
2655 struct pipe_ctx *idle_pipe = find_idle_secondary_pipe(res_ctx, pool, head_pipe);
2663 idle_pipe->stream = head_pipe->stream;
2664 idle_pipe->stream_res.tg = head_pipe->stream_res.tg;
2665 idle_pipe->stream_res.opp = head_pipe->stream_res.opp;
2667 idle_pipe->plane_res.hubp = pool->hubps[idle_pipe->pipe_idx];
2668 idle_pipe->plane_res.ipp = pool->ipps[idle_pipe->pipe_idx];
2669 idle_pipe->plane_res.dpp = pool->dpps[idle_pipe->pipe_idx];
2670 idle_pipe->plane_res.mpcc_inst = pool->dpps[idle_pipe->pipe_idx]->inst;
2675 bool dcn20_get_dcc_compression_cap(const struct dc *dc,
2676 const struct dc_dcc_surface_param *input,
2677 struct dc_surface_dcc_cap *output)
2679 return dc->res_pool->hubbub->funcs->get_dcc_compression_cap(
2680 dc->res_pool->hubbub,
2685 static void dcn20_destroy_resource_pool(struct resource_pool **pool)
2687 struct dcn20_resource_pool *dcn20_pool = TO_DCN20_RES_POOL(*pool);
2689 destruct(dcn20_pool);
2695 static struct dc_cap_funcs cap_funcs = {
2696 .get_dcc_compression_cap = dcn20_get_dcc_compression_cap
2700 enum dc_status dcn20_get_default_swizzle_mode(struct dc_plane_state *plane_state)
2702 enum dc_status result = DC_OK;
2704 enum surface_pixel_format surf_pix_format = plane_state->format;
2705 unsigned int bpp = resource_pixel_format_to_bpp(surf_pix_format);
2707 enum swizzle_mode_values swizzle = DC_SW_LINEAR;
2710 swizzle = DC_SW_64KB_D;
2712 swizzle = DC_SW_64KB_S;
2714 plane_state->tiling_info.gfx9.swizzle = swizzle;
2718 static struct resource_funcs dcn20_res_pool_funcs = {
2719 .destroy = dcn20_destroy_resource_pool,
2720 .link_enc_create = dcn20_link_encoder_create,
2721 .validate_bandwidth = dcn20_validate_bandwidth,
2722 .acquire_idle_pipe_for_layer = dcn20_acquire_idle_pipe_for_layer,
2723 .add_stream_to_ctx = dcn20_add_stream_to_ctx,
2724 .remove_stream_from_ctx = dcn20_remove_stream_from_ctx,
2725 .populate_dml_writeback_from_context = dcn20_populate_dml_writeback_from_context,
2726 .get_default_swizzle_mode = dcn20_get_default_swizzle_mode,
2727 .set_mcif_arb_params = dcn20_set_mcif_arb_params,
2728 .find_first_free_match_stream_enc_for_link = dcn10_find_first_free_match_stream_enc_for_link
2731 bool dcn20_dwbc_create(struct dc_context *ctx, struct resource_pool *pool)
2734 uint32_t pipe_count = pool->res_cap->num_dwb;
2736 ASSERT(pipe_count > 0);
2738 for (i = 0; i < pipe_count; i++) {
2739 struct dcn20_dwbc *dwbc20 = kzalloc(sizeof(struct dcn20_dwbc),
2743 dm_error("DC: failed to create dwbc20!\n");
2746 dcn20_dwbc_construct(dwbc20, ctx,
2751 pool->dwbc[i] = &dwbc20->base;
2756 bool dcn20_mmhubbub_create(struct dc_context *ctx, struct resource_pool *pool)
2759 uint32_t pipe_count = pool->res_cap->num_dwb;
2761 ASSERT(pipe_count > 0);
2763 for (i = 0; i < pipe_count; i++) {
2764 struct dcn20_mmhubbub *mcif_wb20 = kzalloc(sizeof(struct dcn20_mmhubbub),
2768 dm_error("DC: failed to create mcif_wb20!\n");
2772 dcn20_mmhubbub_construct(mcif_wb20, ctx,
2778 pool->mcif_wb[i] = &mcif_wb20->base;
2783 struct pp_smu_funcs *dcn20_pp_smu_create(struct dc_context *ctx)
2785 struct pp_smu_funcs *pp_smu = kzalloc(sizeof(*pp_smu), GFP_KERNEL);
2790 dm_pp_get_funcs(ctx, pp_smu);
2792 if (pp_smu->ctx.ver != PP_SMU_VER_NV)
2793 pp_smu = memset(pp_smu, 0, sizeof(struct pp_smu_funcs));
2798 void dcn20_pp_smu_destroy(struct pp_smu_funcs **pp_smu)
2800 if (pp_smu && *pp_smu) {
2806 static void cap_soc_clocks(
2807 struct _vcs_dpi_soc_bounding_box_st *bb,
2808 struct pp_smu_nv_clock_table max_clocks)
2812 // First pass - cap all clocks higher than the reported max
2813 for (i = 0; i < bb->num_states; i++) {
2814 if ((bb->clock_limits[i].dcfclk_mhz > (max_clocks.dcfClockInKhz / 1000))
2815 && max_clocks.dcfClockInKhz != 0)
2816 bb->clock_limits[i].dcfclk_mhz = (max_clocks.dcfClockInKhz / 1000);
2818 if ((bb->clock_limits[i].dram_speed_mts > (max_clocks.uClockInKhz / 1000) * 16)
2819 && max_clocks.uClockInKhz != 0)
2820 bb->clock_limits[i].dram_speed_mts = (max_clocks.uClockInKhz / 1000) * 16;
2822 if ((bb->clock_limits[i].fabricclk_mhz > (max_clocks.fabricClockInKhz / 1000))
2823 && max_clocks.fabricClockInKhz != 0)
2824 bb->clock_limits[i].fabricclk_mhz = (max_clocks.fabricClockInKhz / 1000);
2826 if ((bb->clock_limits[i].dispclk_mhz > (max_clocks.displayClockInKhz / 1000))
2827 && max_clocks.displayClockInKhz != 0)
2828 bb->clock_limits[i].dispclk_mhz = (max_clocks.displayClockInKhz / 1000);
2830 if ((bb->clock_limits[i].dppclk_mhz > (max_clocks.dppClockInKhz / 1000))
2831 && max_clocks.dppClockInKhz != 0)
2832 bb->clock_limits[i].dppclk_mhz = (max_clocks.dppClockInKhz / 1000);
2834 if ((bb->clock_limits[i].phyclk_mhz > (max_clocks.phyClockInKhz / 1000))
2835 && max_clocks.phyClockInKhz != 0)
2836 bb->clock_limits[i].phyclk_mhz = (max_clocks.phyClockInKhz / 1000);
2838 if ((bb->clock_limits[i].socclk_mhz > (max_clocks.socClockInKhz / 1000))
2839 && max_clocks.socClockInKhz != 0)
2840 bb->clock_limits[i].socclk_mhz = (max_clocks.socClockInKhz / 1000);
2842 if ((bb->clock_limits[i].dscclk_mhz > (max_clocks.dscClockInKhz / 1000))
2843 && max_clocks.dscClockInKhz != 0)
2844 bb->clock_limits[i].dscclk_mhz = (max_clocks.dscClockInKhz / 1000);
2847 // Second pass - remove all duplicate clock states
2848 for (i = bb->num_states - 1; i > 1; i--) {
2849 bool duplicate = true;
2851 if (bb->clock_limits[i-1].dcfclk_mhz != bb->clock_limits[i].dcfclk_mhz)
2853 if (bb->clock_limits[i-1].dispclk_mhz != bb->clock_limits[i].dispclk_mhz)
2855 if (bb->clock_limits[i-1].dppclk_mhz != bb->clock_limits[i].dppclk_mhz)
2857 if (bb->clock_limits[i-1].dram_speed_mts != bb->clock_limits[i].dram_speed_mts)
2859 if (bb->clock_limits[i-1].dscclk_mhz != bb->clock_limits[i].dscclk_mhz)
2861 if (bb->clock_limits[i-1].fabricclk_mhz != bb->clock_limits[i].fabricclk_mhz)
2863 if (bb->clock_limits[i-1].phyclk_mhz != bb->clock_limits[i].phyclk_mhz)
2865 if (bb->clock_limits[i-1].socclk_mhz != bb->clock_limits[i].socclk_mhz)
2873 static void update_bounding_box(struct dc *dc, struct _vcs_dpi_soc_bounding_box_st *bb,
2874 struct pp_smu_nv_clock_table *max_clocks, unsigned int *uclk_states, unsigned int num_states)
2876 struct _vcs_dpi_voltage_scaling_st calculated_states[MAX_CLOCK_LIMIT_STATES] = {0};
2878 int num_calculated_states = 0;
2881 if (num_states == 0)
2884 if (dc->bb_overrides.min_dcfclk_mhz > 0)
2885 min_dcfclk = dc->bb_overrides.min_dcfclk_mhz;
2887 // Accounting for SOC/DCF relationship, we can go as high as
2888 // 506Mhz in Vmin. We need to code 507 since SMU will round down to 506.
2891 for (i = 0; i < num_states; i++) {
2892 int min_fclk_required_by_uclk;
2893 calculated_states[i].state = i;
2894 calculated_states[i].dram_speed_mts = uclk_states[i] * 16 / 1000;
2896 // FCLK:UCLK ratio is 1.08
2897 min_fclk_required_by_uclk = mul_u64_u32_shr(BIT_ULL(32) * 1080 / 1000000, uclk_states[i], 32);
2899 calculated_states[i].fabricclk_mhz = (min_fclk_required_by_uclk < min_dcfclk) ?
2900 min_dcfclk : min_fclk_required_by_uclk;
2902 calculated_states[i].socclk_mhz = (calculated_states[i].fabricclk_mhz > max_clocks->socClockInKhz / 1000) ?
2903 max_clocks->socClockInKhz / 1000 : calculated_states[i].fabricclk_mhz;
2905 calculated_states[i].dcfclk_mhz = (calculated_states[i].fabricclk_mhz > max_clocks->dcfClockInKhz / 1000) ?
2906 max_clocks->dcfClockInKhz / 1000 : calculated_states[i].fabricclk_mhz;
2908 calculated_states[i].dispclk_mhz = max_clocks->displayClockInKhz / 1000;
2909 calculated_states[i].dppclk_mhz = max_clocks->displayClockInKhz / 1000;
2910 calculated_states[i].dscclk_mhz = max_clocks->displayClockInKhz / (1000 * 3);
2912 calculated_states[i].phyclk_mhz = max_clocks->phyClockInKhz / 1000;
2914 num_calculated_states++;
2917 calculated_states[num_calculated_states - 1].socclk_mhz = max_clocks->socClockInKhz / 1000;
2918 calculated_states[num_calculated_states - 1].fabricclk_mhz = max_clocks->socClockInKhz / 1000;
2919 calculated_states[num_calculated_states - 1].dcfclk_mhz = max_clocks->dcfClockInKhz / 1000;
2921 memcpy(bb->clock_limits, calculated_states, sizeof(bb->clock_limits));
2922 bb->num_states = num_calculated_states;
2924 // Duplicate the last state, DML always an extra state identical to max state to work
2925 memcpy(&bb->clock_limits[num_calculated_states], &bb->clock_limits[num_calculated_states - 1], sizeof(struct _vcs_dpi_voltage_scaling_st));
2926 bb->clock_limits[num_calculated_states].state = bb->num_states;
2929 static void patch_bounding_box(struct dc *dc, struct _vcs_dpi_soc_bounding_box_st *bb)
2932 if ((int)(bb->sr_exit_time_us * 1000) != dc->bb_overrides.sr_exit_time_ns
2933 && dc->bb_overrides.sr_exit_time_ns) {
2934 bb->sr_exit_time_us = dc->bb_overrides.sr_exit_time_ns / 1000.0;
2937 if ((int)(bb->sr_enter_plus_exit_time_us * 1000)
2938 != dc->bb_overrides.sr_enter_plus_exit_time_ns
2939 && dc->bb_overrides.sr_enter_plus_exit_time_ns) {
2940 bb->sr_enter_plus_exit_time_us =
2941 dc->bb_overrides.sr_enter_plus_exit_time_ns / 1000.0;
2944 if ((int)(bb->urgent_latency_us * 1000) != dc->bb_overrides.urgent_latency_ns
2945 && dc->bb_overrides.urgent_latency_ns) {
2946 bb->urgent_latency_us = dc->bb_overrides.urgent_latency_ns / 1000.0;
2949 if ((int)(bb->dram_clock_change_latency_us * 1000)
2950 != dc->bb_overrides.dram_clock_change_latency_ns
2951 && dc->bb_overrides.dram_clock_change_latency_ns) {
2952 bb->dram_clock_change_latency_us =
2953 dc->bb_overrides.dram_clock_change_latency_ns / 1000.0;
2958 #define fixed16_to_double(x) (((double) x) / ((double) (1 << 16)))
2959 #define fixed16_to_double_to_cpu(x) fixed16_to_double(le32_to_cpu(x))
2961 static bool init_soc_bounding_box(struct dc *dc,
2962 struct dcn20_resource_pool *pool)
2964 const struct gpu_info_soc_bounding_box_v1_0 *bb = dc->soc_bounding_box;
2965 DC_LOGGER_INIT(dc->ctx->logger);
2967 if (!bb && !SOC_BOUNDING_BOX_VALID) {
2968 DC_LOG_ERROR("%s: not valid soc bounding box/n", __func__);
2972 if (bb && !SOC_BOUNDING_BOX_VALID) {
2975 dcn2_0_soc.sr_exit_time_us =
2976 fixed16_to_double_to_cpu(bb->sr_exit_time_us);
2977 dcn2_0_soc.sr_enter_plus_exit_time_us =
2978 fixed16_to_double_to_cpu(bb->sr_enter_plus_exit_time_us);
2979 dcn2_0_soc.urgent_latency_us =
2980 fixed16_to_double_to_cpu(bb->urgent_latency_us);
2981 dcn2_0_soc.urgent_latency_pixel_data_only_us =
2982 fixed16_to_double_to_cpu(bb->urgent_latency_pixel_data_only_us);
2983 dcn2_0_soc.urgent_latency_pixel_mixed_with_vm_data_us =
2984 fixed16_to_double_to_cpu(bb->urgent_latency_pixel_mixed_with_vm_data_us);
2985 dcn2_0_soc.urgent_latency_vm_data_only_us =
2986 fixed16_to_double_to_cpu(bb->urgent_latency_vm_data_only_us);
2987 dcn2_0_soc.urgent_out_of_order_return_per_channel_pixel_only_bytes =
2988 le32_to_cpu(bb->urgent_out_of_order_return_per_channel_pixel_only_bytes);
2989 dcn2_0_soc.urgent_out_of_order_return_per_channel_pixel_and_vm_bytes =
2990 le32_to_cpu(bb->urgent_out_of_order_return_per_channel_pixel_and_vm_bytes);
2991 dcn2_0_soc.urgent_out_of_order_return_per_channel_vm_only_bytes =
2992 le32_to_cpu(bb->urgent_out_of_order_return_per_channel_vm_only_bytes);
2993 dcn2_0_soc.pct_ideal_dram_sdp_bw_after_urgent_pixel_only =
2994 fixed16_to_double_to_cpu(bb->pct_ideal_dram_sdp_bw_after_urgent_pixel_only);
2995 dcn2_0_soc.pct_ideal_dram_sdp_bw_after_urgent_pixel_and_vm =
2996 fixed16_to_double_to_cpu(bb->pct_ideal_dram_sdp_bw_after_urgent_pixel_and_vm);
2997 dcn2_0_soc.pct_ideal_dram_sdp_bw_after_urgent_vm_only =
2998 fixed16_to_double_to_cpu(bb->pct_ideal_dram_sdp_bw_after_urgent_vm_only);
2999 dcn2_0_soc.max_avg_sdp_bw_use_normal_percent =
3000 fixed16_to_double_to_cpu(bb->max_avg_sdp_bw_use_normal_percent);
3001 dcn2_0_soc.max_avg_dram_bw_use_normal_percent =
3002 fixed16_to_double_to_cpu(bb->max_avg_dram_bw_use_normal_percent);
3003 dcn2_0_soc.writeback_latency_us =
3004 fixed16_to_double_to_cpu(bb->writeback_latency_us);
3005 dcn2_0_soc.ideal_dram_bw_after_urgent_percent =
3006 fixed16_to_double_to_cpu(bb->ideal_dram_bw_after_urgent_percent);
3007 dcn2_0_soc.max_request_size_bytes =
3008 le32_to_cpu(bb->max_request_size_bytes);
3009 dcn2_0_soc.dram_channel_width_bytes =
3010 le32_to_cpu(bb->dram_channel_width_bytes);
3011 dcn2_0_soc.fabric_datapath_to_dcn_data_return_bytes =
3012 le32_to_cpu(bb->fabric_datapath_to_dcn_data_return_bytes);
3013 dcn2_0_soc.dcn_downspread_percent =
3014 fixed16_to_double_to_cpu(bb->dcn_downspread_percent);
3015 dcn2_0_soc.downspread_percent =
3016 fixed16_to_double_to_cpu(bb->downspread_percent);
3017 dcn2_0_soc.dram_page_open_time_ns =
3018 fixed16_to_double_to_cpu(bb->dram_page_open_time_ns);
3019 dcn2_0_soc.dram_rw_turnaround_time_ns =
3020 fixed16_to_double_to_cpu(bb->dram_rw_turnaround_time_ns);
3021 dcn2_0_soc.dram_return_buffer_per_channel_bytes =
3022 le32_to_cpu(bb->dram_return_buffer_per_channel_bytes);
3023 dcn2_0_soc.round_trip_ping_latency_dcfclk_cycles =
3024 le32_to_cpu(bb->round_trip_ping_latency_dcfclk_cycles);
3025 dcn2_0_soc.urgent_out_of_order_return_per_channel_bytes =
3026 le32_to_cpu(bb->urgent_out_of_order_return_per_channel_bytes);
3027 dcn2_0_soc.channel_interleave_bytes =
3028 le32_to_cpu(bb->channel_interleave_bytes);
3029 dcn2_0_soc.num_banks =
3030 le32_to_cpu(bb->num_banks);
3031 dcn2_0_soc.num_chans =
3032 le32_to_cpu(bb->num_chans);
3033 dcn2_0_soc.vmm_page_size_bytes =
3034 le32_to_cpu(bb->vmm_page_size_bytes);
3035 dcn2_0_soc.dram_clock_change_latency_us =
3036 fixed16_to_double_to_cpu(bb->dram_clock_change_latency_us);
3037 dcn2_0_soc.writeback_dram_clock_change_latency_us =
3038 fixed16_to_double_to_cpu(bb->writeback_dram_clock_change_latency_us);
3039 dcn2_0_soc.return_bus_width_bytes =
3040 le32_to_cpu(bb->return_bus_width_bytes);
3041 dcn2_0_soc.dispclk_dppclk_vco_speed_mhz =
3042 le32_to_cpu(bb->dispclk_dppclk_vco_speed_mhz);
3043 dcn2_0_soc.xfc_bus_transport_time_us =
3044 le32_to_cpu(bb->xfc_bus_transport_time_us);
3045 dcn2_0_soc.xfc_xbuf_latency_tolerance_us =
3046 le32_to_cpu(bb->xfc_xbuf_latency_tolerance_us);
3047 dcn2_0_soc.use_urgent_burst_bw =
3048 le32_to_cpu(bb->use_urgent_burst_bw);
3049 dcn2_0_soc.num_states =
3050 le32_to_cpu(bb->num_states);
3052 for (i = 0; i < dcn2_0_soc.num_states; i++) {
3053 dcn2_0_soc.clock_limits[i].state =
3054 le32_to_cpu(bb->clock_limits[i].state);
3055 dcn2_0_soc.clock_limits[i].dcfclk_mhz =
3056 fixed16_to_double_to_cpu(bb->clock_limits[i].dcfclk_mhz);
3057 dcn2_0_soc.clock_limits[i].fabricclk_mhz =
3058 fixed16_to_double_to_cpu(bb->clock_limits[i].fabricclk_mhz);
3059 dcn2_0_soc.clock_limits[i].dispclk_mhz =
3060 fixed16_to_double_to_cpu(bb->clock_limits[i].dispclk_mhz);
3061 dcn2_0_soc.clock_limits[i].dppclk_mhz =
3062 fixed16_to_double_to_cpu(bb->clock_limits[i].dppclk_mhz);
3063 dcn2_0_soc.clock_limits[i].phyclk_mhz =
3064 fixed16_to_double_to_cpu(bb->clock_limits[i].phyclk_mhz);
3065 dcn2_0_soc.clock_limits[i].socclk_mhz =
3066 fixed16_to_double_to_cpu(bb->clock_limits[i].socclk_mhz);
3067 dcn2_0_soc.clock_limits[i].dscclk_mhz =
3068 fixed16_to_double_to_cpu(bb->clock_limits[i].dscclk_mhz);
3069 dcn2_0_soc.clock_limits[i].dram_speed_mts =
3070 fixed16_to_double_to_cpu(bb->clock_limits[i].dram_speed_mts);
3074 if (pool->base.pp_smu) {
3075 struct pp_smu_nv_clock_table max_clocks = {0};
3076 unsigned int uclk_states[8] = {0};
3077 unsigned int num_states = 0;
3078 enum pp_smu_status status;
3079 bool clock_limits_available = false;
3080 bool uclk_states_available = false;
3082 if (pool->base.pp_smu->nv_funcs.get_uclk_dpm_states) {
3083 status = (pool->base.pp_smu->nv_funcs.get_uclk_dpm_states)
3084 (&pool->base.pp_smu->nv_funcs.pp_smu, uclk_states, &num_states);
3086 uclk_states_available = (status == PP_SMU_RESULT_OK);
3089 if (pool->base.pp_smu->nv_funcs.get_maximum_sustainable_clocks) {
3090 status = (*pool->base.pp_smu->nv_funcs.get_maximum_sustainable_clocks)
3091 (&pool->base.pp_smu->nv_funcs.pp_smu, &max_clocks);
3092 /* SMU cannot set DCF clock to anything equal to or higher than SOC clock
3094 if (max_clocks.dcfClockInKhz >= max_clocks.socClockInKhz)
3095 max_clocks.dcfClockInKhz = max_clocks.socClockInKhz - 1000;
3096 clock_limits_available = (status == PP_SMU_RESULT_OK);
3099 if (clock_limits_available && uclk_states_available && num_states)
3100 update_bounding_box(dc, &dcn2_0_soc, &max_clocks, uclk_states, num_states);
3101 else if (clock_limits_available)
3102 cap_soc_clocks(&dcn2_0_soc, max_clocks);
3105 dcn2_0_ip.max_num_otg = pool->base.res_cap->num_timing_generator;
3106 dcn2_0_ip.max_num_dpp = pool->base.pipe_count;
3107 patch_bounding_box(dc, &dcn2_0_soc);
3112 static bool construct(
3113 uint8_t num_virtual_links,
3115 struct dcn20_resource_pool *pool)
3118 struct dc_context *ctx = dc->ctx;
3119 struct irq_service_init_data init_data;
3121 ctx->dc_bios->regs = &bios_regs;
3122 pool->base.funcs = &dcn20_res_pool_funcs;
3124 if (ASICREV_IS_NAVI14_M(ctx->asic_id.hw_internal_rev)) {
3125 pool->base.res_cap = &res_cap_nv14;
3126 pool->base.pipe_count = 5;
3127 pool->base.mpcc_count = 5;
3129 pool->base.res_cap = &res_cap_nv10;
3130 pool->base.pipe_count = 6;
3131 pool->base.mpcc_count = 6;
3133 /*************************************************
3134 * Resource + asic cap harcoding *
3135 *************************************************/
3136 pool->base.underlay_pipe_index = NO_UNDERLAY_PIPE;
3138 dc->caps.max_downscale_ratio = 200;
3139 dc->caps.i2c_speed_in_khz = 100;
3140 dc->caps.max_cursor_size = 256;
3141 dc->caps.dmdata_alloc_size = 2048;
3143 dc->caps.max_slave_planes = 1;
3144 dc->caps.post_blend_color_processing = true;
3145 dc->caps.force_dp_tps4_for_cp2520 = true;
3146 dc->caps.hw_3d_lut = true;
3148 if (dc->ctx->dce_environment == DCE_ENV_PRODUCTION_DRV) {
3149 dc->debug = debug_defaults_drv;
3150 } else if (dc->ctx->dce_environment == DCE_ENV_FPGA_MAXIMUS) {
3151 pool->base.pipe_count = 4;
3152 pool->base.mpcc_count = pool->base.pipe_count;
3153 dc->debug = debug_defaults_diags;
3155 dc->debug = debug_defaults_diags;
3158 dc->work_arounds.dedcn20_305_wa = true;
3160 // Init the vm_helper
3162 vm_helper_init(dc->vm_helper, 16);
3164 /*************************************************
3165 * Create resources *
3166 *************************************************/
3168 pool->base.clock_sources[DCN20_CLK_SRC_PLL0] =
3169 dcn20_clock_source_create(ctx, ctx->dc_bios,
3170 CLOCK_SOURCE_COMBO_PHY_PLL0,
3171 &clk_src_regs[0], false);
3172 pool->base.clock_sources[DCN20_CLK_SRC_PLL1] =
3173 dcn20_clock_source_create(ctx, ctx->dc_bios,
3174 CLOCK_SOURCE_COMBO_PHY_PLL1,
3175 &clk_src_regs[1], false);
3176 pool->base.clock_sources[DCN20_CLK_SRC_PLL2] =
3177 dcn20_clock_source_create(ctx, ctx->dc_bios,
3178 CLOCK_SOURCE_COMBO_PHY_PLL2,
3179 &clk_src_regs[2], false);
3180 pool->base.clock_sources[DCN20_CLK_SRC_PLL3] =
3181 dcn20_clock_source_create(ctx, ctx->dc_bios,
3182 CLOCK_SOURCE_COMBO_PHY_PLL3,
3183 &clk_src_regs[3], false);
3184 pool->base.clock_sources[DCN20_CLK_SRC_PLL4] =
3185 dcn20_clock_source_create(ctx, ctx->dc_bios,
3186 CLOCK_SOURCE_COMBO_PHY_PLL4,
3187 &clk_src_regs[4], false);
3188 pool->base.clock_sources[DCN20_CLK_SRC_PLL5] =
3189 dcn20_clock_source_create(ctx, ctx->dc_bios,
3190 CLOCK_SOURCE_COMBO_PHY_PLL5,
3191 &clk_src_regs[5], false);
3192 pool->base.clk_src_count = DCN20_CLK_SRC_TOTAL;
3193 /* todo: not reuse phy_pll registers */
3194 pool->base.dp_clock_source =
3195 dcn20_clock_source_create(ctx, ctx->dc_bios,
3196 CLOCK_SOURCE_ID_DP_DTO,
3197 &clk_src_regs[0], true);
3199 for (i = 0; i < pool->base.clk_src_count; i++) {
3200 if (pool->base.clock_sources[i] == NULL) {
3201 dm_error("DC: failed to create clock sources!\n");
3202 BREAK_TO_DEBUGGER();
3207 pool->base.dccg = dccg2_create(ctx, &dccg_regs, &dccg_shift, &dccg_mask);
3208 if (pool->base.dccg == NULL) {
3209 dm_error("DC: failed to create dccg!\n");
3210 BREAK_TO_DEBUGGER();
3214 pool->base.dmcu = dcn20_dmcu_create(ctx,
3218 if (pool->base.dmcu == NULL) {
3219 dm_error("DC: failed to create dmcu!\n");
3220 BREAK_TO_DEBUGGER();
3224 pool->base.abm = dce_abm_create(ctx,
3228 if (pool->base.abm == NULL) {
3229 dm_error("DC: failed to create abm!\n");
3230 BREAK_TO_DEBUGGER();
3234 pool->base.pp_smu = dcn20_pp_smu_create(ctx);
3237 if (!init_soc_bounding_box(dc, pool)) {
3238 dm_error("DC: failed to initialize soc bounding box!\n");
3239 BREAK_TO_DEBUGGER();
3243 dml_init_instance(&dc->dml, &dcn2_0_soc, &dcn2_0_ip, DML_PROJECT_NAVI10);
3244 dml_init_instance(&dc->work_arounds.alternate_dml, &dcn2_0_soc, &dcn2_0_ip, DML_PROJECT_NAVI10v2);
3246 if (!dc->debug.disable_pplib_wm_range) {
3247 struct pp_smu_wm_range_sets ranges = {0};
3250 ranges.num_reader_wm_sets = 0;
3252 if (dcn2_0_soc.num_states == 1) {
3253 ranges.reader_wm_sets[0].wm_inst = i;
3254 ranges.reader_wm_sets[0].min_drain_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MIN;
3255 ranges.reader_wm_sets[0].max_drain_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MAX;
3256 ranges.reader_wm_sets[0].min_fill_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MIN;
3257 ranges.reader_wm_sets[0].max_fill_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MAX;
3259 ranges.num_reader_wm_sets = 1;
3260 } else if (dcn2_0_soc.num_states > 1) {
3261 for (i = 0; i < 4 && i < dcn2_0_soc.num_states; i++) {
3262 ranges.reader_wm_sets[i].wm_inst = i;
3263 ranges.reader_wm_sets[i].min_drain_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MIN;
3264 ranges.reader_wm_sets[i].max_drain_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MAX;
3265 ranges.reader_wm_sets[i].min_fill_clk_mhz = (i > 0) ? (dcn2_0_soc.clock_limits[i - 1].dram_speed_mts / 16) + 1 : 0;
3266 ranges.reader_wm_sets[i].max_fill_clk_mhz = dcn2_0_soc.clock_limits[i].dram_speed_mts / 16;
3268 ranges.num_reader_wm_sets = i + 1;
3271 ranges.reader_wm_sets[0].min_fill_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MIN;
3272 ranges.reader_wm_sets[ranges.num_reader_wm_sets - 1].max_fill_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MAX;
3275 ranges.num_writer_wm_sets = 1;
3277 ranges.writer_wm_sets[0].wm_inst = 0;
3278 ranges.writer_wm_sets[0].min_fill_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MIN;
3279 ranges.writer_wm_sets[0].max_fill_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MAX;
3280 ranges.writer_wm_sets[0].min_drain_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MIN;
3281 ranges.writer_wm_sets[0].max_drain_clk_mhz = PP_SMU_WM_SET_RANGE_CLK_UNCONSTRAINED_MAX;
3283 /* Notify PP Lib/SMU which Watermarks to use for which clock ranges */
3284 if (pool->base.pp_smu->nv_funcs.set_wm_ranges)
3285 pool->base.pp_smu->nv_funcs.set_wm_ranges(&pool->base.pp_smu->nv_funcs.pp_smu, &ranges);
3288 init_data.ctx = dc->ctx;
3289 pool->base.irqs = dal_irq_service_dcn20_create(&init_data);
3290 if (!pool->base.irqs)
3293 /* mem input -> ipp -> dpp -> opp -> TG */
3294 for (i = 0; i < pool->base.pipe_count; i++) {
3295 pool->base.hubps[i] = dcn20_hubp_create(ctx, i);
3296 if (pool->base.hubps[i] == NULL) {
3297 BREAK_TO_DEBUGGER();
3299 "DC: failed to create memory input!\n");
3303 pool->base.ipps[i] = dcn20_ipp_create(ctx, i);
3304 if (pool->base.ipps[i] == NULL) {
3305 BREAK_TO_DEBUGGER();
3307 "DC: failed to create input pixel processor!\n");
3311 pool->base.dpps[i] = dcn20_dpp_create(ctx, i);
3312 if (pool->base.dpps[i] == NULL) {
3313 BREAK_TO_DEBUGGER();
3315 "DC: failed to create dpps!\n");
3319 for (i = 0; i < pool->base.res_cap->num_ddc; i++) {
3320 pool->base.engines[i] = dcn20_aux_engine_create(ctx, i);
3321 if (pool->base.engines[i] == NULL) {
3322 BREAK_TO_DEBUGGER();
3324 "DC:failed to create aux engine!!\n");
3327 pool->base.hw_i2cs[i] = dcn20_i2c_hw_create(ctx, i);
3328 if (pool->base.hw_i2cs[i] == NULL) {
3329 BREAK_TO_DEBUGGER();
3331 "DC:failed to create hw i2c!!\n");
3334 pool->base.sw_i2cs[i] = NULL;
3337 for (i = 0; i < pool->base.res_cap->num_opp; i++) {
3338 pool->base.opps[i] = dcn20_opp_create(ctx, i);
3339 if (pool->base.opps[i] == NULL) {
3340 BREAK_TO_DEBUGGER();
3342 "DC: failed to create output pixel processor!\n");
3347 for (i = 0; i < pool->base.res_cap->num_timing_generator; i++) {
3348 pool->base.timing_generators[i] = dcn20_timing_generator_create(
3350 if (pool->base.timing_generators[i] == NULL) {
3351 BREAK_TO_DEBUGGER();
3352 dm_error("DC: failed to create tg!\n");
3357 pool->base.timing_generator_count = i;
3359 pool->base.mpc = dcn20_mpc_create(ctx);
3360 if (pool->base.mpc == NULL) {
3361 BREAK_TO_DEBUGGER();
3362 dm_error("DC: failed to create mpc!\n");
3366 pool->base.hubbub = dcn20_hubbub_create(ctx);
3367 if (pool->base.hubbub == NULL) {
3368 BREAK_TO_DEBUGGER();
3369 dm_error("DC: failed to create hubbub!\n");
3373 #ifdef CONFIG_DRM_AMD_DC_DSC_SUPPORT
3374 for (i = 0; i < pool->base.res_cap->num_dsc; i++) {
3375 pool->base.dscs[i] = dcn20_dsc_create(ctx, i);
3376 if (pool->base.dscs[i] == NULL) {
3377 BREAK_TO_DEBUGGER();
3378 dm_error("DC: failed to create display stream compressor %d!\n", i);
3384 if (!dcn20_dwbc_create(ctx, &pool->base)) {
3385 BREAK_TO_DEBUGGER();
3386 dm_error("DC: failed to create dwbc!\n");
3389 if (!dcn20_mmhubbub_create(ctx, &pool->base)) {
3390 BREAK_TO_DEBUGGER();
3391 dm_error("DC: failed to create mcif_wb!\n");
3395 if (!resource_construct(num_virtual_links, dc, &pool->base,
3396 (!IS_FPGA_MAXIMUS_DC(dc->ctx->dce_environment) ?
3397 &res_create_funcs : &res_create_maximus_funcs)))
3400 dcn20_hw_sequencer_construct(dc);
3402 dc->caps.max_planes = pool->base.pipe_count;
3404 for (i = 0; i < dc->caps.max_planes; ++i)
3405 dc->caps.planes[i] = plane_cap;
3407 dc->cap_funcs = cap_funcs;
3418 struct resource_pool *dcn20_create_resource_pool(
3419 const struct dc_init_data *init_data,
3422 struct dcn20_resource_pool *pool =
3423 kzalloc(sizeof(struct dcn20_resource_pool), GFP_KERNEL);
3428 if (construct(init_data->num_virtual_links, dc, pool))
3431 BREAK_TO_DEBUGGER();