Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at master 922 lines 25 kB view raw
1// SPDX-License-Identifier: GPL-2.0-only OR MIT 2/* Copyright (c) 2023 Imagination Technologies Ltd. */ 3 4#include "pvr_device.h" 5#include "pvr_device_info.h" 6 7#include "pvr_fw.h" 8#include "pvr_power.h" 9#include "pvr_queue.h" 10#include "pvr_rogue_cr_defs.h" 11#include "pvr_stream.h" 12#include "pvr_vm.h" 13 14#include <drm/drm_print.h> 15 16#include <linux/bitfield.h> 17#include <linux/clk.h> 18#include <linux/compiler_attributes.h> 19#include <linux/compiler_types.h> 20#include <linux/dma-mapping.h> 21#include <linux/err.h> 22#include <linux/firmware.h> 23#include <linux/gfp.h> 24#include <linux/interrupt.h> 25#include <linux/of.h> 26#include <linux/platform_device.h> 27#include <linux/pm_runtime.h> 28#include <linux/reset.h> 29#include <linux/slab.h> 30#include <linux/stddef.h> 31#include <linux/types.h> 32#include <linux/workqueue.h> 33 34#include <kunit/visibility.h> 35 36/* Major number for the supported version of the firmware. */ 37#define PVR_FW_VERSION_MAJOR 1 38 39/** 40 * pvr_device_reg_init() - Initialize kernel access to a PowerVR device's 41 * control registers. 42 * @pvr_dev: Target PowerVR device. 43 * 44 * Sets struct pvr_device->regs. 45 * 46 * This method of mapping the device control registers into memory ensures that 47 * they are unmapped when the driver is detached (i.e. no explicit cleanup is 48 * required). 49 * 50 * Return: 51 * * 0 on success, or 52 * * Any error returned by devm_platform_get_and_ioremap_resource(). 53 */ 54static int 55pvr_device_reg_init(struct pvr_device *pvr_dev) 56{ 57 struct drm_device *drm_dev = from_pvr_device(pvr_dev); 58 struct platform_device *plat_dev = to_platform_device(drm_dev->dev); 59 struct resource *regs_resource; 60 void __iomem *regs; 61 62 pvr_dev->regs_resource = NULL; 63 pvr_dev->regs = NULL; 64 65 regs = devm_platform_get_and_ioremap_resource(plat_dev, 0, &regs_resource); 66 if (IS_ERR(regs)) 67 return dev_err_probe(drm_dev->dev, PTR_ERR(regs), 68 "failed to ioremap gpu registers\n"); 69 70 pvr_dev->regs = regs; 71 pvr_dev->regs_resource = regs_resource; 72 73 return 0; 74} 75 76/** 77 * pvr_device_clk_init() - Initialize clocks required by a PowerVR device 78 * @pvr_dev: Target PowerVR device. 79 * 80 * Sets struct pvr_device->core_clk, struct pvr_device->sys_clk and 81 * struct pvr_device->mem_clk. 82 * 83 * Three clocks are required by the PowerVR device: core, sys and mem. On 84 * return, this function guarantees that the clocks are in one of the following 85 * states: 86 * 87 * * All successfully initialized, 88 * * Core errored, sys and mem uninitialized, 89 * * Core deinitialized, sys errored, mem uninitialized, or 90 * * Core and sys deinitialized, mem errored. 91 * 92 * Return: 93 * * 0 on success, 94 * * Any error returned by devm_clk_get(), or 95 * * Any error returned by devm_clk_get_optional(). 96 */ 97static int pvr_device_clk_init(struct pvr_device *pvr_dev) 98{ 99 struct drm_device *drm_dev = from_pvr_device(pvr_dev); 100 struct clk *core_clk; 101 struct clk *sys_clk; 102 struct clk *mem_clk; 103 104 core_clk = devm_clk_get(drm_dev->dev, "core"); 105 if (IS_ERR(core_clk)) 106 return dev_err_probe(drm_dev->dev, PTR_ERR(core_clk), 107 "failed to get core clock\n"); 108 109 sys_clk = devm_clk_get_optional(drm_dev->dev, "sys"); 110 if (IS_ERR(sys_clk)) 111 return dev_err_probe(drm_dev->dev, PTR_ERR(sys_clk), 112 "failed to get sys clock\n"); 113 114 mem_clk = devm_clk_get_optional(drm_dev->dev, "mem"); 115 if (IS_ERR(mem_clk)) 116 return dev_err_probe(drm_dev->dev, PTR_ERR(mem_clk), 117 "failed to get mem clock\n"); 118 119 pvr_dev->core_clk = core_clk; 120 pvr_dev->sys_clk = sys_clk; 121 pvr_dev->mem_clk = mem_clk; 122 123 return 0; 124} 125 126/** 127 * pvr_device_process_active_queues() - Process all queue related events. 128 * @pvr_dev: PowerVR device to check 129 * 130 * This is called any time we receive a FW event. It iterates over all 131 * active queues and calls pvr_queue_process() on them. 132 */ 133static void pvr_device_process_active_queues(struct pvr_device *pvr_dev) 134{ 135 struct pvr_queue *queue, *tmp_queue; 136 LIST_HEAD(active_queues); 137 138 mutex_lock(&pvr_dev->queues.lock); 139 140 /* Move all active queues to a temporary list. Queues that remain 141 * active after we're done processing them are re-inserted to 142 * the queues.active list by pvr_queue_process(). 143 */ 144 list_splice_init(&pvr_dev->queues.active, &active_queues); 145 146 list_for_each_entry_safe(queue, tmp_queue, &active_queues, node) 147 pvr_queue_process(queue); 148 149 mutex_unlock(&pvr_dev->queues.lock); 150} 151 152static bool pvr_device_safety_irq_pending(struct pvr_device *pvr_dev) 153{ 154 u32 events; 155 156 WARN_ON_ONCE(!pvr_dev->has_safety_events); 157 158 events = pvr_cr_read32(pvr_dev, ROGUE_CR_EVENT_STATUS); 159 160 return (events & ROGUE_CR_EVENT_STATUS_SAFETY_EN) != 0; 161} 162 163static void pvr_device_safety_irq_clear(struct pvr_device *pvr_dev) 164{ 165 WARN_ON_ONCE(!pvr_dev->has_safety_events); 166 167 pvr_cr_write32(pvr_dev, ROGUE_CR_EVENT_CLEAR, 168 ROGUE_CR_EVENT_CLEAR_SAFETY_EN); 169} 170 171static void pvr_device_handle_safety_events(struct pvr_device *pvr_dev) 172{ 173 struct drm_device *drm_dev = from_pvr_device(pvr_dev); 174 u32 events; 175 176 WARN_ON_ONCE(!pvr_dev->has_safety_events); 177 178 events = pvr_cr_read32(pvr_dev, ROGUE_CR_SAFETY_EVENT_STATUS__ROGUEXE); 179 180 /* Handle only these events on the host and leave the rest to the FW. */ 181 events &= ROGUE_CR_SAFETY_EVENT_STATUS__ROGUEXE__FAULT_FW_EN | 182 ROGUE_CR_SAFETY_EVENT_STATUS__ROGUEXE__WATCHDOG_TIMEOUT_EN; 183 184 pvr_cr_write32(pvr_dev, ROGUE_CR_SAFETY_EVENT_CLEAR__ROGUEXE, events); 185 186 if (events & ROGUE_CR_SAFETY_EVENT_STATUS__ROGUEXE__FAULT_FW_EN) { 187 u32 fault_fw = pvr_cr_read32(pvr_dev, ROGUE_CR_FAULT_FW_STATUS); 188 189 pvr_cr_write32(pvr_dev, ROGUE_CR_FAULT_FW_CLEAR, fault_fw); 190 191 drm_info(drm_dev, "Safety event: FW fault (mask=0x%08x)\n", fault_fw); 192 } 193 194 if (events & ROGUE_CR_SAFETY_EVENT_STATUS__ROGUEXE__WATCHDOG_TIMEOUT_EN) { 195 /* 196 * The watchdog timer is disabled by the driver so this event 197 * should never be fired. 198 */ 199 drm_info(drm_dev, "Safety event: Watchdog timeout\n"); 200 } 201} 202 203static irqreturn_t pvr_device_irq_thread_handler(int irq, void *data) 204{ 205 struct pvr_device *pvr_dev = data; 206 struct drm_device *drm_dev = from_pvr_device(pvr_dev); 207 irqreturn_t ret = IRQ_NONE; 208 209 /* We are in the threaded handler, we can keep dequeuing events until we 210 * don't see any. This should allow us to reduce the number of interrupts 211 * when the GPU is receiving a massive amount of short jobs. 212 */ 213 while (pvr_fw_irq_pending(pvr_dev)) { 214 pvr_fw_irq_clear(pvr_dev); 215 216 if (pvr_dev->fw_dev.booted) { 217 pvr_fwccb_process(pvr_dev); 218 pvr_kccb_wake_up_waiters(pvr_dev); 219 pvr_device_process_active_queues(pvr_dev); 220 } 221 222 pm_runtime_mark_last_busy(drm_dev->dev); 223 224 ret = IRQ_HANDLED; 225 } 226 227 if (pvr_dev->has_safety_events) { 228 while (pvr_device_safety_irq_pending(pvr_dev)) { 229 pvr_device_safety_irq_clear(pvr_dev); 230 pvr_device_handle_safety_events(pvr_dev); 231 232 ret = IRQ_HANDLED; 233 } 234 } 235 236 return ret; 237} 238 239static irqreturn_t pvr_device_irq_handler(int irq, void *data) 240{ 241 struct pvr_device *pvr_dev = data; 242 bool safety_irq_pending = false; 243 244 if (pvr_dev->has_safety_events) 245 safety_irq_pending = pvr_device_safety_irq_pending(pvr_dev); 246 247 if (!pvr_fw_irq_pending(pvr_dev) && !safety_irq_pending) 248 return IRQ_NONE; /* Spurious IRQ - ignore. */ 249 250 return IRQ_WAKE_THREAD; 251} 252 253static void pvr_device_safety_irq_init(struct pvr_device *pvr_dev) 254{ 255 u32 num_ecc_rams = 0; 256 257 /* 258 * Safety events are an optional feature of the RogueXE platform. They 259 * are only enabled if at least one of ECC memory or the watchdog timer 260 * are present in HW. While safety events can be generated by other 261 * systems, that will never happen if the above mentioned hardware is 262 * not present. 263 */ 264 if (!PVR_HAS_FEATURE(pvr_dev, roguexe)) { 265 pvr_dev->has_safety_events = false; 266 return; 267 } 268 269 PVR_FEATURE_VALUE(pvr_dev, ecc_rams, &num_ecc_rams); 270 271 pvr_dev->has_safety_events = 272 num_ecc_rams > 0 || PVR_HAS_FEATURE(pvr_dev, watchdog_timer); 273} 274 275/** 276 * pvr_device_irq_init() - Initialise IRQ required by a PowerVR device 277 * @pvr_dev: Target PowerVR device. 278 * 279 * Returns: 280 * * 0 on success, 281 * * Any error returned by platform_get_irq_byname(), or 282 * * Any error returned by request_irq(). 283 */ 284static int 285pvr_device_irq_init(struct pvr_device *pvr_dev) 286{ 287 struct drm_device *drm_dev = from_pvr_device(pvr_dev); 288 struct platform_device *plat_dev = to_platform_device(drm_dev->dev); 289 290 init_waitqueue_head(&pvr_dev->kccb.rtn_q); 291 292 pvr_device_safety_irq_init(pvr_dev); 293 294 pvr_dev->irq = platform_get_irq(plat_dev, 0); 295 if (pvr_dev->irq < 0) 296 return pvr_dev->irq; 297 298 /* Clear any pending events before requesting the IRQ line. */ 299 pvr_fw_irq_clear(pvr_dev); 300 301 if (pvr_dev->has_safety_events) 302 pvr_device_safety_irq_clear(pvr_dev); 303 304 /* 305 * The ONESHOT flag ensures IRQs are masked while the thread handler is 306 * running. 307 */ 308 return request_threaded_irq(pvr_dev->irq, pvr_device_irq_handler, 309 pvr_device_irq_thread_handler, 310 IRQF_SHARED | IRQF_ONESHOT, "gpu", pvr_dev); 311} 312 313/** 314 * pvr_device_irq_fini() - Deinitialise IRQ required by a PowerVR device 315 * @pvr_dev: Target PowerVR device. 316 */ 317static void 318pvr_device_irq_fini(struct pvr_device *pvr_dev) 319{ 320 free_irq(pvr_dev->irq, pvr_dev); 321} 322 323/** 324 * pvr_build_firmware_filename() - Construct a PowerVR firmware filename 325 * @pvr_dev: Target PowerVR device. 326 * @base: First part of the filename. 327 * @major: Major version number. 328 * 329 * A PowerVR firmware filename consists of three parts separated by underscores 330 * (``'_'``) along with a '.fw' file suffix. The first part is the exact value 331 * of @base, the second part is the hardware version string derived from @pvr_fw 332 * and the final part is the firmware version number constructed from @major with 333 * a 'v' prefix, e.g. powervr/rogue_4.40.2.51_v1.fw. 334 * 335 * The returned string will have been slab allocated and must be freed with 336 * kfree(). 337 * 338 * Return: 339 * * The constructed filename on success, or 340 * * Any error returned by kasprintf(). 341 */ 342static char * 343pvr_build_firmware_filename(struct pvr_device *pvr_dev, const char *base, 344 u8 major) 345{ 346 struct pvr_gpu_id *gpu_id = &pvr_dev->gpu_id; 347 348 return kasprintf(GFP_KERNEL, "%s_%d.%d.%d.%d_v%d.fw", base, gpu_id->b, 349 gpu_id->v, gpu_id->n, gpu_id->c, major); 350} 351 352static void 353pvr_release_firmware(void *data) 354{ 355 struct pvr_device *pvr_dev = data; 356 357 release_firmware(pvr_dev->fw_dev.firmware); 358} 359 360/** 361 * pvr_request_firmware() - Load firmware for a PowerVR device 362 * @pvr_dev: Target PowerVR device. 363 * 364 * See pvr_build_firmware_filename() for details on firmware file naming. 365 * 366 * Return: 367 * * 0 on success, 368 * * Any error returned by pvr_build_firmware_filename(), or 369 * * Any error returned by request_firmware(). 370 */ 371static int 372pvr_request_firmware(struct pvr_device *pvr_dev) 373{ 374 struct drm_device *drm_dev = &pvr_dev->base; 375 char *filename; 376 const struct firmware *fw; 377 int err; 378 379 filename = pvr_build_firmware_filename(pvr_dev, "powervr/rogue", 380 PVR_FW_VERSION_MAJOR); 381 if (!filename) 382 return -ENOMEM; 383 384 /* 385 * This function takes a copy of &filename, meaning we can free our 386 * instance before returning. 387 */ 388 err = request_firmware(&fw, filename, pvr_dev->base.dev); 389 if (err) { 390 drm_err(drm_dev, "failed to load firmware %s (err=%d)\n", 391 filename, err); 392 goto err_free_filename; 393 } 394 395 drm_info(drm_dev, "loaded firmware %s\n", filename); 396 kfree(filename); 397 398 pvr_dev->fw_dev.firmware = fw; 399 400 return devm_add_action_or_reset(drm_dev->dev, pvr_release_firmware, pvr_dev); 401 402err_free_filename: 403 kfree(filename); 404 405 return err; 406} 407 408/** 409 * pvr_gpuid_decode_reg() - Decode the GPU ID from GPU register 410 * 411 * Sets the b, v, n, c fields of struct pvr_dev.gpu_id. 412 * 413 * @pvr_dev: Target PowerVR device. 414 * @gpu_id: Output to be updated with the GPU ID. 415 */ 416static void 417pvr_gpuid_decode_reg(const struct pvr_device *pvr_dev, struct pvr_gpu_id *gpu_id) 418{ 419 /* 420 * Try reading the BVNC using the newer (cleaner) method first. If the 421 * B value is zero, fall back to the older method. 422 */ 423 u64 bvnc = pvr_cr_read64(pvr_dev, ROGUE_CR_CORE_ID__PBVNC); 424 425 gpu_id->b = PVR_CR_FIELD_GET(bvnc, CORE_ID__PBVNC__BRANCH_ID); 426 if (gpu_id->b != 0) { 427 gpu_id->v = PVR_CR_FIELD_GET(bvnc, CORE_ID__PBVNC__VERSION_ID); 428 gpu_id->n = PVR_CR_FIELD_GET(bvnc, CORE_ID__PBVNC__NUMBER_OF_SCALABLE_UNITS); 429 gpu_id->c = PVR_CR_FIELD_GET(bvnc, CORE_ID__PBVNC__CONFIG_ID); 430 } else { 431 u32 core_rev = pvr_cr_read32(pvr_dev, ROGUE_CR_CORE_REVISION); 432 u32 core_id = pvr_cr_read32(pvr_dev, ROGUE_CR_CORE_ID); 433 u16 core_id_config = PVR_CR_FIELD_GET(core_id, CORE_ID_CONFIG); 434 435 gpu_id->b = PVR_CR_FIELD_GET(core_rev, CORE_REVISION_MAJOR); 436 gpu_id->v = PVR_CR_FIELD_GET(core_rev, CORE_REVISION_MINOR); 437 gpu_id->n = FIELD_GET(0xFF00, core_id_config); 438 gpu_id->c = FIELD_GET(0x00FF, core_id_config); 439 } 440} 441 442/** 443 * pvr_gpuid_decode_string() - Decode the GPU ID from a module input string 444 * 445 * Sets the b, v, n, c fields of struct pvr_dev.gpu_id. 446 * 447 * @pvr_dev: Target PowerVR device. 448 * @param_bvnc: GPU ID (BVNC) module parameter. 449 * @gpu_id: Output to be updated with the GPU ID. 450 */ 451VISIBLE_IF_KUNIT int 452pvr_gpuid_decode_string(const struct pvr_device *pvr_dev, 453 const char *param_bvnc, struct pvr_gpu_id *gpu_id) 454{ 455 const struct drm_device *drm_dev = &pvr_dev->base; 456 char str_cpy[PVR_GPUID_STRING_MAX_LENGTH]; 457 char *pos, *tkn; 458 int ret, idx = 0; 459 u16 user_bvnc_u16[4]; 460 u8 dot_cnt = 0; 461 462 ret = strscpy(str_cpy, param_bvnc); 463 464 /* 465 * strscpy() should return at least a size 7 for the input to be valid. 466 * Returns -E2BIG for the case when the string is empty or too long. 467 */ 468 if (ret < PVR_GPUID_STRING_MIN_LENGTH) { 469 drm_info(drm_dev, 470 "Invalid size of the input GPU ID (BVNC): %s", 471 str_cpy); 472 return -EINVAL; 473 } 474 475 while (*param_bvnc) { 476 if (*param_bvnc == '.') 477 dot_cnt++; 478 param_bvnc++; 479 } 480 481 if (dot_cnt != 3) { 482 drm_info(drm_dev, 483 "Invalid format of the input GPU ID (BVNC): %s", 484 str_cpy); 485 return -EINVAL; 486 } 487 488 pos = str_cpy; 489 490 while ((tkn = strsep(&pos, ".")) != NULL && idx < 4) { 491 /* kstrtou16() will also handle the case of consecutive dots */ 492 ret = kstrtou16(tkn, 10, &user_bvnc_u16[idx]); 493 if (ret) { 494 drm_info(drm_dev, 495 "Invalid format of the input GPU ID (BVNC): %s", 496 str_cpy); 497 return -EINVAL; 498 } 499 idx++; 500 } 501 502 gpu_id->b = user_bvnc_u16[0]; 503 gpu_id->v = user_bvnc_u16[1]; 504 gpu_id->n = user_bvnc_u16[2]; 505 gpu_id->c = user_bvnc_u16[3]; 506 507 return 0; 508} 509EXPORT_SYMBOL_IF_KUNIT(pvr_gpuid_decode_string); 510 511static bool pvr_exp_hw_support; 512module_param_named(exp_hw_support, pvr_exp_hw_support, bool, 0600); 513MODULE_PARM_DESC(exp_hw_support, "Bypass runtime checks for fully supported GPU cores. WARNING: enabling this option may result in a buggy, insecure, or otherwise unusable driver."); 514 515/** 516 * enum pvr_gpu_support_level - The level of support for a gpu_id in the current 517 * version of the driver. 518 * 519 * @PVR_GPU_UNKNOWN: Cores that are unknown to the driver. These may not even exist. 520 * @PVR_GPU_EXPERIMENTAL: Cores that have experimental support. 521 * @PVR_GPU_SUPPORTED: Cores that are supported and maintained. 522 */ 523enum pvr_gpu_support_level { 524 PVR_GPU_UNKNOWN, 525 PVR_GPU_EXPERIMENTAL, 526 PVR_GPU_SUPPORTED, 527}; 528 529static enum pvr_gpu_support_level 530pvr_gpu_support_level(const struct pvr_gpu_id *gpu_id) 531{ 532 switch (pvr_gpu_id_to_packed_bvnc(gpu_id)) { 533 case PVR_PACKED_BVNC(33, 15, 11, 3): 534 case PVR_PACKED_BVNC(36, 53, 104, 796): 535 return PVR_GPU_SUPPORTED; 536 537 case PVR_PACKED_BVNC(36, 52, 104, 182): 538 return PVR_GPU_EXPERIMENTAL; 539 540 default: 541 return PVR_GPU_UNKNOWN; 542 } 543} 544 545static int 546pvr_check_gpu_supported(struct pvr_device *pvr_dev, 547 const struct pvr_gpu_id *gpu_id) 548{ 549 struct drm_device *drm_dev = from_pvr_device(pvr_dev); 550 551 switch (pvr_gpu_support_level(gpu_id)) { 552 case PVR_GPU_SUPPORTED: 553 if (pvr_exp_hw_support) 554 drm_info(drm_dev, "Module parameter 'exp_hw_support' was set, but this hardware is fully supported by the current driver."); 555 556 break; 557 558 case PVR_GPU_EXPERIMENTAL: 559 if (!pvr_exp_hw_support) { 560 drm_err(drm_dev, "Unsupported GPU! Set 'exp_hw_support' to bypass this check."); 561 return -ENODEV; 562 } 563 564 drm_warn(drm_dev, "Running on unsupported hardware; you may encounter bugs!"); 565 break; 566 567 /* NOTE: This code path may indicate misbehaving hardware. */ 568 case PVR_GPU_UNKNOWN: 569 default: 570 if (!pvr_exp_hw_support) { 571 drm_err(drm_dev, "Unknown GPU! Set 'exp_hw_support' to bypass this check."); 572 return -ENODEV; 573 } 574 575 drm_warn(drm_dev, "Running on unknown hardware; expect issues."); 576 break; 577 } 578 579 return 0; 580} 581 582static char *pvr_gpuid_override; 583module_param_named(gpuid, pvr_gpuid_override, charp, 0400); 584MODULE_PARM_DESC(gpuid, "GPU ID (BVNC) to be used instead of the value read from hardware."); 585 586/** 587 * pvr_load_gpu_id() - Load a PowerVR device's GPU ID (BVNC) from control 588 * registers or input parameter. The input parameter is processed instead 589 * of the GPU register if provided. 590 * 591 * Sets the arch field of struct pvr_dev.gpu_id. 592 * 593 * @pvr_dev: Target PowerVR device. 594 */ 595static int 596pvr_load_gpu_id(struct pvr_device *pvr_dev) 597{ 598 struct pvr_gpu_id *gpu_id = &pvr_dev->gpu_id; 599 600 if (!pvr_gpuid_override || !pvr_gpuid_override[0]) { 601 pvr_gpuid_decode_reg(pvr_dev, gpu_id); 602 } else { 603 drm_warn(from_pvr_device(pvr_dev), 604 "Using custom GPU ID (BVNC) provided by the user!"); 605 606 int err = pvr_gpuid_decode_string(pvr_dev, pvr_gpuid_override, 607 gpu_id); 608 if (err) 609 return err; 610 } 611 612 return pvr_check_gpu_supported(pvr_dev, gpu_id); 613} 614 615/** 616 * pvr_set_dma_info() - Set PowerVR device DMA information 617 * @pvr_dev: Target PowerVR device. 618 * 619 * Sets the DMA mask and max segment size for the PowerVR device. 620 * 621 * Return: 622 * * 0 on success, 623 * * Any error returned by PVR_FEATURE_VALUE(), or 624 * * Any error returned by dma_set_mask(). 625 */ 626 627static int 628pvr_set_dma_info(struct pvr_device *pvr_dev) 629{ 630 struct drm_device *drm_dev = from_pvr_device(pvr_dev); 631 u16 phys_bus_width; 632 int err; 633 634 err = PVR_FEATURE_VALUE(pvr_dev, phys_bus_width, &phys_bus_width); 635 if (err) { 636 drm_err(drm_dev, "Failed to get device physical bus width\n"); 637 return err; 638 } 639 640 err = dma_set_mask(drm_dev->dev, DMA_BIT_MASK(phys_bus_width)); 641 if (err) { 642 drm_err(drm_dev, "Failed to set DMA mask (err=%d)\n", err); 643 return err; 644 } 645 646 dma_set_max_seg_size(drm_dev->dev, UINT_MAX); 647 648 return 0; 649} 650 651/** 652 * pvr_device_gpu_init() - GPU-specific initialization for a PowerVR device 653 * @pvr_dev: Target PowerVR device. 654 * 655 * The following steps are taken to ensure the device is ready: 656 * 657 * 1. Read the hardware version information from control registers, 658 * 2. Initialise the hardware feature information, 659 * 3. Setup the device DMA information, 660 * 4. Setup the device-scoped memory context, and 661 * 5. Load firmware into the device. 662 * 663 * Return: 664 * * 0 on success, 665 * * -%ENODEV if the GPU is not supported, 666 * * Any error returned by pvr_set_dma_info(), 667 * * Any error returned by pvr_memory_context_init(), or 668 * * Any error returned by pvr_request_firmware(). 669 */ 670static int 671pvr_device_gpu_init(struct pvr_device *pvr_dev) 672{ 673 int err; 674 675 err = pvr_load_gpu_id(pvr_dev); 676 if (err) 677 return err; 678 679 err = pvr_request_firmware(pvr_dev); 680 if (err) 681 return err; 682 683 err = pvr_fw_validate_init_device_info(pvr_dev); 684 if (err) 685 return err; 686 687 if (PVR_HAS_FEATURE(pvr_dev, meta)) 688 pvr_dev->fw_dev.processor_type = PVR_FW_PROCESSOR_TYPE_META; 689 else if (PVR_HAS_FEATURE(pvr_dev, mips)) 690 pvr_dev->fw_dev.processor_type = PVR_FW_PROCESSOR_TYPE_MIPS; 691 else if (PVR_HAS_FEATURE(pvr_dev, riscv_fw_processor)) 692 pvr_dev->fw_dev.processor_type = PVR_FW_PROCESSOR_TYPE_RISCV; 693 else 694 return -EINVAL; 695 696 pvr_stream_create_musthave_masks(pvr_dev); 697 698 err = pvr_set_dma_info(pvr_dev); 699 if (err) 700 return err; 701 702 if (pvr_dev->fw_dev.processor_type != PVR_FW_PROCESSOR_TYPE_MIPS) { 703 pvr_dev->kernel_vm_ctx = pvr_vm_create_context(pvr_dev, false); 704 if (IS_ERR(pvr_dev->kernel_vm_ctx)) 705 return PTR_ERR(pvr_dev->kernel_vm_ctx); 706 } 707 708 err = pvr_fw_init(pvr_dev); 709 if (err) 710 goto err_vm_ctx_put; 711 712 return 0; 713 714err_vm_ctx_put: 715 if (pvr_dev->fw_dev.processor_type != PVR_FW_PROCESSOR_TYPE_MIPS) { 716 pvr_vm_context_put(pvr_dev->kernel_vm_ctx); 717 pvr_dev->kernel_vm_ctx = NULL; 718 } 719 720 return err; 721} 722 723/** 724 * pvr_device_gpu_fini() - GPU-specific deinitialization for a PowerVR device 725 * @pvr_dev: Target PowerVR device. 726 */ 727static void 728pvr_device_gpu_fini(struct pvr_device *pvr_dev) 729{ 730 pvr_fw_fini(pvr_dev); 731 732 if (pvr_dev->fw_dev.processor_type != PVR_FW_PROCESSOR_TYPE_MIPS) { 733 WARN_ON(!pvr_vm_context_put(pvr_dev->kernel_vm_ctx)); 734 pvr_dev->kernel_vm_ctx = NULL; 735 } 736} 737 738/** 739 * pvr_device_init() - Initialize a PowerVR device 740 * @pvr_dev: Target PowerVR device. 741 * 742 * If this function returns successfully, the device will have been fully 743 * initialized. Otherwise, any parts of the device initialized before an error 744 * occurs will be de-initialized before returning. 745 * 746 * NOTE: The initialization steps currently taken are the bare minimum required 747 * to read from the control registers. The device is unlikely to function 748 * until further initialization steps are added. [This note should be 749 * removed when that happens.] 750 * 751 * Return: 752 * * 0 on success, 753 * * Any error returned by pvr_device_reg_init(), 754 * * Any error returned by pvr_device_clk_init(), or 755 * * Any error returned by pvr_device_gpu_init(). 756 */ 757int 758pvr_device_init(struct pvr_device *pvr_dev) 759{ 760 struct drm_device *drm_dev = from_pvr_device(pvr_dev); 761 struct device *dev = drm_dev->dev; 762 int err; 763 764 /* Get the platform-specific data based on the compatible string. */ 765 pvr_dev->device_data = of_device_get_match_data(dev); 766 767 /* Enable and initialize clocks required for the device to operate. */ 768 err = pvr_device_clk_init(pvr_dev); 769 if (err) 770 return err; 771 772 err = pvr_dev->device_data->pwr_ops->init(pvr_dev); 773 if (err) 774 return err; 775 776 /* Explicitly power the GPU so we can access control registers before the FW is booted. */ 777 err = pm_runtime_resume_and_get(dev); 778 if (err) 779 return err; 780 781 /* Map the control registers into memory. */ 782 err = pvr_device_reg_init(pvr_dev); 783 if (err) 784 goto err_pm_runtime_put; 785 786 /* Perform GPU-specific initialization steps. */ 787 err = pvr_device_gpu_init(pvr_dev); 788 if (err) 789 goto err_pm_runtime_put; 790 791 err = pvr_device_irq_init(pvr_dev); 792 if (err) 793 goto err_device_gpu_fini; 794 795 pm_runtime_put(dev); 796 797 return 0; 798 799err_device_gpu_fini: 800 pvr_device_gpu_fini(pvr_dev); 801 802err_pm_runtime_put: 803 pm_runtime_put_sync_suspend(dev); 804 805 return err; 806} 807 808/** 809 * pvr_device_fini() - Deinitialize a PowerVR device 810 * @pvr_dev: Target PowerVR device. 811 */ 812void 813pvr_device_fini(struct pvr_device *pvr_dev) 814{ 815 /* 816 * Deinitialization stages are performed in reverse order compared to 817 * the initialization stages in pvr_device_init(). 818 */ 819 pvr_device_irq_fini(pvr_dev); 820 pvr_device_gpu_fini(pvr_dev); 821} 822 823bool 824pvr_device_has_uapi_quirk(struct pvr_device *pvr_dev, u32 quirk) 825{ 826 switch (quirk) { 827 case 47217: 828 return PVR_HAS_QUIRK(pvr_dev, 47217); 829 case 48545: 830 return PVR_HAS_QUIRK(pvr_dev, 48545); 831 case 49927: 832 return PVR_HAS_QUIRK(pvr_dev, 49927); 833 case 51764: 834 return PVR_HAS_QUIRK(pvr_dev, 51764); 835 case 62269: 836 return PVR_HAS_QUIRK(pvr_dev, 62269); 837 default: 838 return false; 839 }; 840} 841 842bool 843pvr_device_has_uapi_enhancement(struct pvr_device *pvr_dev, u32 enhancement) 844{ 845 switch (enhancement) { 846 case 35421: 847 return PVR_HAS_ENHANCEMENT(pvr_dev, 35421); 848 case 42064: 849 return PVR_HAS_ENHANCEMENT(pvr_dev, 42064); 850 default: 851 return false; 852 }; 853} 854 855/** 856 * pvr_device_has_feature() - Look up device feature based on feature definition 857 * @pvr_dev: Device pointer. 858 * @feature: Feature to look up. Should be one of %PVR_FEATURE_*. 859 * 860 * Returns: 861 * * %true if feature is present on device, or 862 * * %false if feature is not present on device. 863 */ 864bool 865pvr_device_has_feature(struct pvr_device *pvr_dev, u32 feature) 866{ 867 switch (feature) { 868 case PVR_FEATURE_CLUSTER_GROUPING: 869 return PVR_HAS_FEATURE(pvr_dev, cluster_grouping); 870 871 case PVR_FEATURE_COMPUTE_MORTON_CAPABLE: 872 return PVR_HAS_FEATURE(pvr_dev, compute_morton_capable); 873 874 case PVR_FEATURE_FB_CDC_V4: 875 return PVR_HAS_FEATURE(pvr_dev, fb_cdc_v4); 876 877 case PVR_FEATURE_GPU_MULTICORE_SUPPORT: 878 return PVR_HAS_FEATURE(pvr_dev, gpu_multicore_support); 879 880 case PVR_FEATURE_ISP_ZLS_D24_S8_PACKING_OGL_MODE: 881 return PVR_HAS_FEATURE(pvr_dev, isp_zls_d24_s8_packing_ogl_mode); 882 883 case PVR_FEATURE_S7_TOP_INFRASTRUCTURE: 884 return PVR_HAS_FEATURE(pvr_dev, s7_top_infrastructure); 885 886 case PVR_FEATURE_TESSELLATION: 887 return PVR_HAS_FEATURE(pvr_dev, tessellation); 888 889 case PVR_FEATURE_TPU_DM_GLOBAL_REGISTERS: 890 return PVR_HAS_FEATURE(pvr_dev, tpu_dm_global_registers); 891 892 case PVR_FEATURE_VDM_DRAWINDIRECT: 893 return PVR_HAS_FEATURE(pvr_dev, vdm_drawindirect); 894 895 case PVR_FEATURE_VDM_OBJECT_LEVEL_LLS: 896 return PVR_HAS_FEATURE(pvr_dev, vdm_object_level_lls); 897 898 case PVR_FEATURE_ZLS_SUBTILE: 899 return PVR_HAS_FEATURE(pvr_dev, zls_subtile); 900 901 /* Derived features. */ 902 case PVR_FEATURE_CDM_USER_MODE_QUEUE: { 903 u8 cdm_control_stream_format = 0; 904 905 PVR_FEATURE_VALUE(pvr_dev, cdm_control_stream_format, &cdm_control_stream_format); 906 return (cdm_control_stream_format >= 2 && cdm_control_stream_format <= 4); 907 } 908 909 case PVR_FEATURE_REQUIRES_FB_CDC_ZLS_SETUP: 910 if (PVR_HAS_FEATURE(pvr_dev, fbcdc_algorithm)) { 911 u8 fbcdc_algorithm = 0; 912 913 PVR_FEATURE_VALUE(pvr_dev, fbcdc_algorithm, &fbcdc_algorithm); 914 return (fbcdc_algorithm < 3 || PVR_HAS_FEATURE(pvr_dev, fb_cdc_v4)); 915 } 916 return false; 917 918 default: 919 WARN(true, "Looking up undefined feature %u\n", feature); 920 return false; 921 } 922}