a6xx_state 83 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c void *state_kcalloc(struct a6xx_gpu_state *a6xx_state, int nr, size_t objsize) a6xx_state 91 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c list_add_tail(&obj->node, &a6xx_state->objs); a6xx_state 95 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c void *state_kmemdup(struct a6xx_gpu_state *a6xx_state, void *src, a6xx_state 98 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c void *dst = state_kcalloc(a6xx_state, 1, size); a6xx_state 229 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 235 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kcalloc(a6xx_state, VBIF_DEBUGBUS_BLOCK_SIZE, a6xx_state 283 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 290 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kcalloc(a6xx_state, block->count, sizeof(u64)); a6xx_state 301 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 308 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kcalloc(a6xx_state, block->count, sizeof(u64)); a6xx_state 319 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state) a6xx_state 377 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->debugbus = state_kcalloc(a6xx_state, a6xx_state 379 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c sizeof(*a6xx_state->debugbus)); a6xx_state 381 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (a6xx_state->debugbus) { a6xx_state 386 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state, a6xx_state 388 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->debugbus[i]); a6xx_state 390 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->nr_debugbus = ARRAY_SIZE(a6xx_debugbus_blocks); a6xx_state 393 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->vbif_debugbus = a6xx_state 394 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c state_kcalloc(a6xx_state, 1, a6xx_state 395 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c sizeof(*a6xx_state->vbif_debugbus)); a6xx_state 397 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (a6xx_state->vbif_debugbus) a6xx_state 398 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_vbif_debugbus_block(gpu, a6xx_state, a6xx_state 399 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->vbif_debugbus); a6xx_state 402 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->cx_debugbus = a6xx_state 403 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c state_kcalloc(a6xx_state, a6xx_state 405 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c sizeof(*a6xx_state->cx_debugbus)); a6xx_state 407 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (a6xx_state->cx_debugbus) { a6xx_state 412 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state, a6xx_state 414 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->cx_debugbus[i]); a6xx_state 416 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->nr_cx_debugbus = a6xx_state 428 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 469 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kmemdup(a6xx_state, dumper->ptr + A6XX_CD_DATA_OFFSET, a6xx_state 474 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 479 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->dbgahb_clusters = state_kcalloc(a6xx_state, a6xx_state 481 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c sizeof(*a6xx_state->dbgahb_clusters)); a6xx_state 483 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (!a6xx_state->dbgahb_clusters) a6xx_state 486 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->nr_dbgahb_clusters = ARRAY_SIZE(a6xx_dbgahb_clusters); a6xx_state 489 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_dbgahb_cluster(gpu, a6xx_state, a6xx_state 491 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->dbgahb_clusters[i], dumper); a6xx_state 496 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 540 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kmemdup(a6xx_state, dumper->ptr + A6XX_CD_DATA_OFFSET, a6xx_state 545 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 550 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->clusters = state_kcalloc(a6xx_state, a6xx_state 551 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c ARRAY_SIZE(a6xx_clusters), sizeof(*a6xx_state->clusters)); a6xx_state 553 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (!a6xx_state->clusters) a6xx_state 556 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->nr_clusters = ARRAY_SIZE(a6xx_clusters); a6xx_state 559 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_cluster(gpu, a6xx_state, &a6xx_clusters[i], a6xx_state 560 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->clusters[i], dumper); a6xx_state 565 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 591 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kmemdup(a6xx_state, dumper->ptr + A6XX_CD_DATA_OFFSET, a6xx_state 596 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 601 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->shaders = state_kcalloc(a6xx_state, a6xx_state 602 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c ARRAY_SIZE(a6xx_shader_blocks), sizeof(*a6xx_state->shaders)); a6xx_state 604 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (!a6xx_state->shaders) a6xx_state 607 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->nr_shaders = ARRAY_SIZE(a6xx_shader_blocks); a6xx_state 610 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_shader_block(gpu, a6xx_state, &a6xx_shader_blocks[i], a6xx_state 611 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->shaders[i], dumper); a6xx_state 616 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 648 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kmemdup(a6xx_state, dumper->ptr + A6XX_CD_DATA_OFFSET, a6xx_state 654 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 686 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kmemdup(a6xx_state, dumper->ptr + A6XX_CD_DATA_OFFSET, a6xx_state 692 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 702 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kcalloc(a6xx_state, regcount, sizeof(u32)); a6xx_state 718 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 731 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kcalloc(a6xx_state, regcount, sizeof(u32)); a6xx_state 746 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state) a6xx_state 751 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->gmu_registers = state_kcalloc(a6xx_state, a6xx_state 752 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c 2, sizeof(*a6xx_state->gmu_registers)); a6xx_state 754 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (!a6xx_state->gmu_registers) a6xx_state 757 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->nr_gmu_registers = 2; a6xx_state 760 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c _a6xx_get_gmu_registers(gpu, a6xx_state, &a6xx_gmu_reglist[0], a6xx_state 761 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->gmu_registers[0]); a6xx_state 769 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c _a6xx_get_gmu_registers(gpu, a6xx_state, &a6xx_gmu_reglist[1], a6xx_state 770 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->gmu_registers[1]); a6xx_state 774 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 782 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->registers = state_kcalloc(a6xx_state, a6xx_state 783 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c count, sizeof(*a6xx_state->registers)); a6xx_state 785 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (!a6xx_state->registers) a6xx_state 788 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->nr_registers = count; a6xx_state 792 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state, &a6xx_ahb_reglist[i], a6xx_state 793 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->registers[index++]); a6xx_state 797 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state, &a6xx_reglist[i], a6xx_state 798 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->registers[index++], a6xx_state 803 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state, &a6xx_hlsq_reglist[i], a6xx_state 804 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->registers[index++], a6xx_state 810 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state, a6xx_state 817 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c obj->data = state_kcalloc(a6xx_state, indexed->count, sizeof(u32)); a6xx_state 830 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state) a6xx_state 836 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->indexed_regs = state_kcalloc(a6xx_state, count, a6xx_state 837 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c sizeof(a6xx_state->indexed_regs)); a6xx_state 838 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (!a6xx_state->indexed_regs) a6xx_state 842 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_indexed_regs(gpu, a6xx_state, &a6xx_indexed_reglist[i], a6xx_state 843 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->indexed_regs[i]); a6xx_state 850 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_indexed_regs(gpu, a6xx_state, &a6xx_cp_mempool_indexed, a6xx_state 851 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c &a6xx_state->indexed_regs[i]); a6xx_state 857 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->indexed_regs[i].data[0x2000] = mempool_size; a6xx_state 862 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_state->nr_indexed_regs = count; a6xx_state 870 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state = kzalloc(sizeof(*a6xx_state), a6xx_state 873 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (!a6xx_state) a6xx_state 876 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c INIT_LIST_HEAD(&a6xx_state->objs); a6xx_state 879 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c adreno_gpu_state_get(gpu, &a6xx_state->base); a6xx_state 881 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_gmu_registers(gpu, a6xx_state); a6xx_state 885 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c return &a6xx_state->base; a6xx_state 888 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_indexed_registers(gpu, a6xx_state); a6xx_state 892 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_registers(gpu, a6xx_state, &dumper); a6xx_state 893 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_shaders(gpu, a6xx_state, &dumper); a6xx_state 894 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_clusters(gpu, a6xx_state, &dumper); a6xx_state 895 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_dbgahb_clusters(gpu, a6xx_state, &dumper); a6xx_state 900 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_get_debugbus(gpu, a6xx_state); a6xx_state 902 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c return &a6xx_state->base; a6xx_state 910 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state = container_of(state, a6xx_state 913 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c list_for_each_entry_safe(obj, tmp, &a6xx_state->objs, node) a6xx_state 917 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c kfree(a6xx_state); a6xx_state 1086 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c static void a6xx_show_debugbus(struct a6xx_gpu_state *a6xx_state, a6xx_state 1091 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c for (i = 0; i < a6xx_state->nr_debugbus; i++) { a6xx_state 1092 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state_obj *obj = &a6xx_state->debugbus[i]; a6xx_state 1097 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c if (a6xx_state->vbif_debugbus) { a6xx_state 1098 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state_obj *obj = a6xx_state->vbif_debugbus; a6xx_state 1107 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c for (i = 0; i < a6xx_state->nr_cx_debugbus; i++) { a6xx_state 1108 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state_obj *obj = &a6xx_state->cx_debugbus[i]; a6xx_state 1117 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state *a6xx_state = container_of(state, a6xx_state 1127 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c for (i = 0; i < a6xx_state->nr_registers; i++) { a6xx_state 1128 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state_obj *obj = &a6xx_state->registers[i]; a6xx_state 1138 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c for (i = 0; i < a6xx_state->nr_gmu_registers; i++) { a6xx_state 1139 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c struct a6xx_gpu_state_obj *obj = &a6xx_state->gmu_registers[i]; a6xx_state 1149 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c for (i = 0; i < a6xx_state->nr_indexed_regs; i++) a6xx_state 1150 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_show_indexed_regs(&a6xx_state->indexed_regs[i], p); a6xx_state 1153 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c for (i = 0; i < a6xx_state->nr_shaders; i++) a6xx_state 1154 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_show_shader(&a6xx_state->shaders[i], p); a6xx_state 1157 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c for (i = 0; i < a6xx_state->nr_clusters; i++) a6xx_state 1158 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_show_cluster(&a6xx_state->clusters[i], p); a6xx_state 1160 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c for (i = 0; i < a6xx_state->nr_dbgahb_clusters; i++) a6xx_state 1161 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_show_dbgahb_cluster(&a6xx_state->dbgahb_clusters[i], p); a6xx_state 1164 drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c a6xx_show_debugbus(a6xx_state, p);