ptc 415 drivers/acpi/processor_throttling.c union acpi_object *ptc = NULL; ptc 427 drivers/acpi/processor_throttling.c ptc = (union acpi_object *)buffer.pointer; ptc 428 drivers/acpi/processor_throttling.c if (!ptc || (ptc->type != ACPI_TYPE_PACKAGE) ptc 429 drivers/acpi/processor_throttling.c || (ptc->package.count != 2)) { ptc 439 drivers/acpi/processor_throttling.c obj = ptc->package.elements[0]; ptc 456 drivers/acpi/processor_throttling.c obj = ptc->package.elements[1]; ptc 118 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h } ptc, ptp; ptc 122 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c struct nvkm_mmu_ptc *ptc; ptc 124 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c list_for_each_entry(ptc, &mmu->ptc.list, head) { ptc 125 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c if (ptc->size == size) ptc 126 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c return ptc; ptc 129 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c ptc = kmalloc(sizeof(*ptc), GFP_KERNEL); ptc 130 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c if (ptc) { ptc 131 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c INIT_LIST_HEAD(&ptc->item); ptc 132 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c ptc->size = size; ptc 133 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c ptc->refs = 0; ptc 134 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c list_add(&ptc->head, &mmu->ptc.list); ptc 137 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c return ptc; ptc 154 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c mutex_lock(&mmu->ptc.mutex); ptc 155 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c if (pt->ptc->refs < 8 /* Heuristic. */ && !force) { ptc 156 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c list_add_tail(&pt->head, &pt->ptc->item); ptc 157 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c pt->ptc->refs++; ptc 162 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c mutex_unlock(&mmu->ptc.mutex); ptc 169 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c struct nvkm_mmu_ptc *ptc; ptc 182 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c mutex_lock(&mmu->ptc.mutex); ptc 183 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c ptc = nvkm_mmu_ptc_find(mmu, size); ptc 184 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c if (!ptc) { ptc 185 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c mutex_unlock(&mmu->ptc.mutex); ptc 190 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c pt = list_first_entry_or_null(&ptc->item, typeof(*pt), head); ptc 195 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c ptc->refs--; ptc 196 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c mutex_unlock(&mmu->ptc.mutex); ptc 199 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c mutex_unlock(&mmu->ptc.mutex); ptc 204 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c pt->ptc = ptc; ptc 222 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c struct nvkm_mmu_ptc *ptc; ptc 223 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c list_for_each_entry(ptc, &mmu->ptc.list, head) { ptc 225 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c list_for_each_entry_safe(pt, tt, &ptc->item, head) { ptc 236 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c struct nvkm_mmu_ptc *ptc, *ptct; ptc 238 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c list_for_each_entry_safe(ptc, ptct, &mmu->ptc.list, head) { ptc 239 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c WARN_ON(!list_empty(&ptc->item)); ptc 240 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c list_del(&ptc->head); ptc 241 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c kfree(ptc); ptc 248 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c mutex_init(&mmu->ptc.mutex); ptc 249 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c INIT_LIST_HEAD(&mmu->ptc.list); ptc 52 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/priv.h struct nvkm_mmu_ptc *ptc; ptc 1428 drivers/usb/gadget/udc/fsl_udc_core.c u16 ptc = 0; ptc 1451 drivers/usb/gadget/udc/fsl_udc_core.c ptc = wIndex >> 8; ptc 1471 drivers/usb/gadget/udc/fsl_udc_core.c if (ptc) { ptc 1475 drivers/usb/gadget/udc/fsl_udc_core.c tmp = fsl_readl(&dr_regs->portsc1) | (ptc << 16); ptc 1477 drivers/usb/gadget/udc/fsl_udc_core.c printk(KERN_INFO "udc: switch to test mode %d.\n", ptc);