uf 210 arch/powerpc/kernel/rtas_flash.c struct rtas_update_flash_t *const uf = &rtas_update_flash_data; uf 214 arch/powerpc/kernel/rtas_flash.c if (uf->flist) { uf 222 arch/powerpc/kernel/rtas_flash.c if (uf->status != FLASH_AUTH) uf 223 arch/powerpc/kernel/rtas_flash.c uf->status = flash_list_valid(uf->flist); uf 225 arch/powerpc/kernel/rtas_flash.c if (uf->status == FLASH_IMG_READY) uf 226 arch/powerpc/kernel/rtas_flash.c rtas_firmware_flash_list = uf->flist; uf 228 arch/powerpc/kernel/rtas_flash.c free_flash_list(uf->flist); uf 230 arch/powerpc/kernel/rtas_flash.c uf->flist = NULL; uf 275 arch/powerpc/kernel/rtas_flash.c struct rtas_update_flash_t *const uf = &rtas_update_flash_data; uf 281 arch/powerpc/kernel/rtas_flash.c status = uf->status; uf 292 arch/powerpc/kernel/rtas_flash.c struct rtas_update_flash_t *const uf = &rtas_update_flash_data; uf 297 arch/powerpc/kernel/rtas_flash.c status = uf->status; uf 313 arch/powerpc/kernel/rtas_flash.c struct rtas_update_flash_t *const uf = &rtas_update_flash_data; uf 320 arch/powerpc/kernel/rtas_flash.c if (uf->status == FLASH_AUTH || count == 0) uf 327 arch/powerpc/kernel/rtas_flash.c if (uf->flist == NULL) { uf 328 arch/powerpc/kernel/rtas_flash.c uf->flist = kmem_cache_zalloc(flash_block_cache, GFP_KERNEL); uf 329 arch/powerpc/kernel/rtas_flash.c if (!uf->flist) uf 333 arch/powerpc/kernel/rtas_flash.c fl = uf->flist; uf 4777 arch/x86/kvm/mmu.c u8 uf = (pfec & PFERR_USER_MASK) ? (u8)~u : 0; uf 4821 arch/x86/kvm/mmu.c mmu->permissions[byte] = ff | uf | wf | smepf | smapf; uf 4870 arch/x86/kvm/mmu.c bool check_pkey, check_write, ff, uf, wf, pte_user; uf 4874 arch/x86/kvm/mmu.c uf = pfec & PFERR_USER_MASK; uf 4889 arch/x86/kvm/mmu.c check_write = check_pkey && wf && (uf || wp); uf 709 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct amdgpu_bo *uf = ttm_to_amdgpu_bo(p->uf_entry.tv.bo); uf 711 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = amdgpu_ttm_alloc_gart(&uf->tbo); uf 712 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c p->job->uf_addr += amdgpu_bo_gpu_offset(uf); uf 778 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct amdgpu_bo *uf = ttm_to_amdgpu_bo(parser->uf_entry.tv.bo); uf 780 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c amdgpu_bo_unref(&uf); uf 917 drivers/net/tun.c struct tun_filter uf; uf 920 drivers/net/tun.c if (copy_from_user(&uf, arg, sizeof(uf))) uf 923 drivers/net/tun.c if (!uf.count) { uf 929 drivers/net/tun.c alen = ETH_ALEN * uf.count; uf 930 drivers/net/tun.c addr = memdup_user(arg + sizeof(uf), alen); uf 941 drivers/net/tun.c for (n = 0; n < uf.count && n < FLT_EXACT_COUNT; n++) uf 949 drivers/net/tun.c for (; n < uf.count; n++) { uf 959 drivers/net/tun.c if ((uf.flags & TUN_FLT_ALLMULTI)) uf 1522 drivers/scsi/bfa/bfa_fcs.c bfa_fcs_uf_recv(void *cbarg, struct bfa_uf_s *uf) uf 1525 drivers/scsi/bfa/bfa_fcs.c struct fchs_s *fchs = bfa_uf_get_frmbuf(uf); uf 1526 drivers/scsi/bfa/bfa_fcs.c u16 len = bfa_uf_get_frmlen(uf); uf 1535 drivers/scsi/bfa/bfa_fcs.c bfa_stats(fcs, uf.tagged); uf 1536 drivers/scsi/bfa/bfa_fcs.c vft = bfa_uf_get_frmbuf(uf); uf 1547 drivers/scsi/bfa/bfa_fcs.c bfa_stats(fcs, uf.vfid_unknown); uf 1548 drivers/scsi/bfa/bfa_fcs.c bfa_uf_free(uf); uf 1560 drivers/scsi/bfa/bfa_fcs.c bfa_stats(fcs, uf.untagged); uf 1573 drivers/scsi/bfa/bfa_fcs.c bfa_uf_free(uf); uf 677 drivers/scsi/bfa/bfa_fcs.h } uf; uf 5352 drivers/scsi/bfa/bfa_svc.c struct bfa_uf_s *uf = cbarg; uf 5353 drivers/scsi/bfa/bfa_svc.c struct bfa_uf_mod_s *ufm = BFA_UF_MOD(uf->bfa); uf 5356 drivers/scsi/bfa/bfa_svc.c ufm->ufrecv(ufm->cbarg, uf); uf 5391 drivers/scsi/bfa/bfa_svc.c struct bfa_uf_s *uf; uf 5401 drivers/scsi/bfa/bfa_svc.c for (i = 0, uf = ufm->uf_list; i < ufm->num_ufs; i++, uf++) { uf 5402 drivers/scsi/bfa/bfa_svc.c memset(uf, 0, sizeof(struct bfa_uf_s)); uf 5403 drivers/scsi/bfa/bfa_svc.c uf->bfa = ufm->bfa; uf 5404 drivers/scsi/bfa/bfa_svc.c uf->uf_tag = i; uf 5405 drivers/scsi/bfa/bfa_svc.c uf->pb_len = BFA_PER_UF_DMA_SZ; uf 5406 drivers/scsi/bfa/bfa_svc.c uf->buf_kva = bfa_mem_get_dmabuf_kva(ufm, i, BFA_PER_UF_DMA_SZ); uf 5407 drivers/scsi/bfa/bfa_svc.c uf->buf_pa = ufm_pbs_pa(ufm, i); uf 5408 drivers/scsi/bfa/bfa_svc.c list_add_tail(&uf->qe, &ufm->uf_free_q); uf 5414 drivers/scsi/bfa/bfa_svc.c bfa_mem_kva_curp(ufm) = (u8 *) uf; uf 5470 drivers/scsi/bfa/bfa_svc.c struct bfa_uf_s *uf; uf 5472 drivers/scsi/bfa/bfa_svc.c bfa_q_deq(&uf_mod->uf_free_q, &uf); uf 5473 drivers/scsi/bfa/bfa_svc.c return uf; uf 5477 drivers/scsi/bfa/bfa_svc.c bfa_uf_put(struct bfa_uf_mod_s *uf_mod, struct bfa_uf_s *uf) uf 5479 drivers/scsi/bfa/bfa_svc.c list_add_tail(&uf->qe, &uf_mod->uf_free_q); uf 5483 drivers/scsi/bfa/bfa_svc.c bfa_uf_post(struct bfa_uf_mod_s *ufm, struct bfa_uf_s *uf) uf 5491 drivers/scsi/bfa/bfa_svc.c memcpy(uf_post_msg, &ufm->uf_buf_posts[uf->uf_tag], uf 5495 drivers/scsi/bfa/bfa_svc.c bfa_trc(ufm->bfa, uf->uf_tag); uf 5497 drivers/scsi/bfa/bfa_svc.c list_add_tail(&uf->qe, &ufm->uf_posted_q); uf 5504 drivers/scsi/bfa/bfa_svc.c struct bfa_uf_s *uf; uf 5506 drivers/scsi/bfa/bfa_svc.c while ((uf = bfa_uf_get(uf_mod)) != NULL) { uf 5507 drivers/scsi/bfa/bfa_svc.c if (bfa_uf_post(uf_mod, uf) != BFA_STATUS_OK) uf 5517 drivers/scsi/bfa/bfa_svc.c struct bfa_uf_s *uf = &ufm->uf_list[uf_tag]; uf 5523 drivers/scsi/bfa/bfa_svc.c bfa_mem_get_dmabuf_kva(ufm, uf_tag, uf->pb_len); uf 5531 drivers/scsi/bfa/bfa_svc.c list_del(&uf->qe); /* dequeue from posted queue */ uf 5533 drivers/scsi/bfa/bfa_svc.c uf->data_ptr = buf; uf 5534 drivers/scsi/bfa/bfa_svc.c uf->data_len = m->xfr_len; uf 5536 drivers/scsi/bfa/bfa_svc.c WARN_ON(uf->data_len < sizeof(struct fchs_s)); uf 5538 drivers/scsi/bfa/bfa_svc.c if (uf->data_len == sizeof(struct fchs_s)) { uf 5540 drivers/scsi/bfa/bfa_svc.c uf->data_len, (struct fchs_s *)buf); uf 5544 drivers/scsi/bfa/bfa_svc.c BFA_PL_EID_RX, uf->data_len, uf 5549 drivers/scsi/bfa/bfa_svc.c __bfa_cb_uf_recv(uf, BFA_TRUE); uf 5551 drivers/scsi/bfa/bfa_svc.c bfa_cb_queue(bfa, &uf->hcb_qe, __bfa_cb_uf_recv, uf); uf 5558 drivers/scsi/bfa/bfa_svc.c struct bfa_uf_s *uf; uf 5565 drivers/scsi/bfa/bfa_svc.c uf = (struct bfa_uf_s *) qe; uf 5566 drivers/scsi/bfa/bfa_svc.c list_del(&uf->qe); uf 5567 drivers/scsi/bfa/bfa_svc.c bfa_uf_put(ufm, uf); uf 5601 drivers/scsi/bfa/bfa_svc.c bfa_uf_free(struct bfa_uf_s *uf) uf 5603 drivers/scsi/bfa/bfa_svc.c bfa_uf_put(BFA_UF_MOD(uf->bfa), uf); uf 5604 drivers/scsi/bfa/bfa_svc.c bfa_uf_post_all(BFA_UF_MOD(uf->bfa)); uf 342 drivers/scsi/bfa/bfa_svc.h typedef void (*bfa_cb_uf_recv_t) (void *cbarg, struct bfa_uf_s *uf); uf 640 drivers/scsi/bfa/bfa_svc.h bfa_uf_get_frmbuf(struct bfa_uf_s *uf) uf 642 drivers/scsi/bfa/bfa_svc.h return uf->data_ptr; uf 646 drivers/scsi/bfa/bfa_svc.h bfa_uf_get_frmlen(struct bfa_uf_s *uf) uf 648 drivers/scsi/bfa/bfa_svc.h return uf->data_len; uf 656 drivers/scsi/bfa/bfa_svc.h void bfa_uf_free(struct bfa_uf_s *uf); uf 63 drivers/scsi/isci/unsolicited_frame_control.c struct sci_unsolicited_frame *uf; uf 110 drivers/scsi/isci/unsolicited_frame_control.c uf = &uf_control->buffers.array[i]; uf 114 drivers/scsi/isci/unsolicited_frame_control.c uf->buffer = virt; uf 115 drivers/scsi/isci/unsolicited_frame_control.c uf->header = &uf_control->headers.array[i]; uf 116 drivers/scsi/isci/unsolicited_frame_control.c uf->state = UNSOLICITED_FRAME_EMPTY; uf 111 drivers/usb/host/ehci-hcd.c unsigned uf; uf 113 drivers/usb/host/ehci-hcd.c uf = ehci_readl(ehci, &ehci->regs->frame_index); uf 114 drivers/usb/host/ehci-hcd.c if (unlikely((uf & 7) == 0)) uf 115 drivers/usb/host/ehci-hcd.c uf = ehci_readl(ehci, &ehci->regs->frame_index); uf 116 drivers/usb/host/ehci-hcd.c return uf; uf 265 drivers/usb/host/ehci-sched.c unsigned uframe, uf, x; uf 280 drivers/usb/host/ehci-sched.c for (uf = ps->phase_uf; uf < 8; ++uf) { uf 281 drivers/usb/host/ehci-sched.c x += budget_line[uf]; uf 285 drivers/usb/host/ehci-sched.c budget_line[uf] = x; uf 288 drivers/usb/host/ehci-sched.c budget_line[uf] = 125; uf 383 drivers/usb/host/ehci-sched.c unsigned i, uf; uf 389 drivers/usb/host/ehci-sched.c uf = frame << 3; uf 390 drivers/usb/host/ehci-sched.c for (i = 0; i < 8; (++i, ++uf)) uf 391 drivers/usb/host/ehci-sched.c tt_usecs[i] = ehci->tt_budget[uf]; uf 1402 drivers/usb/host/ehci-sched.c unsigned frame, uf; uf 1422 drivers/usb/host/ehci-sched.c uf = uframe & 7; uf 1423 drivers/usb/host/ehci-sched.c if (!tt_available(ehci, &stream->ps, tt, frame, uf)) uf 1439 drivers/usb/host/ehci-sched.c uf = uframe; uf 1441 drivers/usb/host/ehci-sched.c for (tmp = stream->ps.cs_mask & 0xff; tmp; tmp >>= 1, uf++) { uf 1442 drivers/usb/host/ehci-sched.c if (ehci->bandwidth[uf] > max_used) uf 1450 drivers/usb/host/ehci-sched.c uf = uframe & ~7; uf 1455 drivers/usb/host/ehci-sched.c if (ehci->bandwidth[uf+i] > max_used) uf 1700 drivers/usb/host/ehci-sched.c struct ehci_iso_packet *uf = &iso_sched->packet[index]; uf 1708 drivers/usb/host/ehci-sched.c itd->hw_transaction[uframe] = uf->transaction; uf 1710 drivers/usb/host/ehci-sched.c itd->hw_bufp[pg] |= cpu_to_hc32(ehci, uf->bufp & ~(u32)0); uf 1711 drivers/usb/host/ehci-sched.c itd->hw_bufp_hi[pg] |= cpu_to_hc32(ehci, (u32)(uf->bufp >> 32)); uf 1714 drivers/usb/host/ehci-sched.c if (unlikely(uf->cross)) { uf 1715 drivers/usb/host/ehci-sched.c u64 bufp = uf->bufp + 4096; uf 2118 drivers/usb/host/ehci-sched.c struct ehci_iso_packet *uf = &iso_sched->packet[index]; uf 2124 drivers/usb/host/ehci-sched.c sitd->hw_results = uf->transaction; uf 2127 drivers/usb/host/ehci-sched.c bufp = uf->bufp; uf 2131 drivers/usb/host/ehci-sched.c sitd->hw_buf[1] = cpu_to_hc32(ehci, uf->buf1); uf 2132 drivers/usb/host/ehci-sched.c if (uf->cross) uf 2365 drivers/usb/host/ehci-sched.c unsigned uf, now_frame, frame; uf 2377 drivers/usb/host/ehci-sched.c uf = ehci_read_frame_index(ehci); uf 2378 drivers/usb/host/ehci-sched.c now_frame = (uf >> 3) & fmask; uf 2407 drivers/usb/host/ehci-sched.c for (uf = 0; uf < 8; uf++) { uf 2408 drivers/usb/host/ehci-sched.c if (q.itd->hw_transaction[uf] & uf 2412 drivers/usb/host/ehci-sched.c if (uf < 8) { uf 4307 drivers/usb/host/fotg210-hcd.c struct fotg210_iso_packet *uf = &iso_sched->packet[index]; uf 4313 drivers/usb/host/fotg210-hcd.c itd->hw_transaction[uframe] = uf->transaction; uf 4315 drivers/usb/host/fotg210-hcd.c itd->hw_bufp[pg] |= cpu_to_hc32(fotg210, uf->bufp & ~(u32)0); uf 4316 drivers/usb/host/fotg210-hcd.c itd->hw_bufp_hi[pg] |= cpu_to_hc32(fotg210, (u32)(uf->bufp >> 32)); uf 4319 drivers/usb/host/fotg210-hcd.c if (unlikely(uf->cross)) { uf 4320 drivers/usb/host/fotg210-hcd.c u64 bufp = uf->bufp + 4096; uf 4585 drivers/usb/host/fotg210-hcd.c unsigned uf; uf 4607 drivers/usb/host/fotg210-hcd.c for (uf = 0; uf < 8; uf++) { uf 4608 drivers/usb/host/fotg210-hcd.c if (q.itd->hw_transaction[uf] & uf 4612 drivers/usb/host/fotg210-hcd.c if (uf < 8) { uf 4654 drivers/usb/host/fotg210-hcd.c unsigned uf, now_frame, frame, ret; uf 4664 drivers/usb/host/fotg210-hcd.c uf = fotg210_read_frame_index(fotg210); uf 4665 drivers/usb/host/fotg210-hcd.c now_frame = (uf >> 3) & fmask; uf 856 fs/userfaultfd.c void userfaultfd_unmap_complete(struct mm_struct *mm, struct list_head *uf) uf 861 fs/userfaultfd.c list_for_each_entry_safe(ctx, n, uf, list) { uf 2344 include/linux/mm.h struct list_head *uf); uf 2348 include/linux/mm.h struct list_head *uf); uf 2350 include/linux/mm.h struct list_head *uf, bool downgrade); uf 2352 include/linux/mm.h struct list_head *uf); uf 2358 include/linux/mm.h struct list_head *uf) uf 2360 include/linux/mm.h return do_mmap(file, addr, len, prot, flags, 0, pgoff, populate, uf); uf 75 include/linux/userfaultfd_k.h struct list_head *uf); uf 77 include/linux/userfaultfd_k.h struct list_head *uf); uf 135 include/linux/userfaultfd_k.h struct list_head *uf) uf 141 include/linux/userfaultfd_k.h struct list_head *uf) uf 485 kernel/fork.c LIST_HEAD(uf); uf 548 kernel/fork.c retval = dup_userfaultfd(tmp, &uf); uf 615 kernel/fork.c dup_userfaultfd_complete(&uf); uf 186 mm/mmap.c struct list_head *uf); uf 196 mm/mmap.c LIST_HEAD(uf); uf 249 mm/mmap.c ret = __do_munmap(mm, newbrk, oldbrk-newbrk, &uf, true); uf 265 mm/mmap.c if (do_brk_flags(oldbrk, newbrk-oldbrk, 0, &uf) < 0) uf 275 mm/mmap.c userfaultfd_unmap_complete(mm, &uf); uf 1393 mm/mmap.c struct list_head *uf) uf 1569 mm/mmap.c addr = mmap_region(file, addr, len, vm_flags, pgoff, uf); uf 1713 mm/mmap.c struct list_head *uf) uf 1739 mm/mmap.c if (do_munmap(mm, addr, len, uf)) uf 2735 mm/mmap.c struct list_head *uf, bool downgrade) uf 2799 mm/mmap.c if (unlikely(uf)) { uf 2809 mm/mmap.c int error = userfaultfd_unmap_prep(vma, start, end, uf); uf 2844 mm/mmap.c struct list_head *uf) uf 2846 mm/mmap.c return __do_munmap(mm, start, len, uf, false); uf 2853 mm/mmap.c LIST_HEAD(uf); uf 2858 mm/mmap.c ret = __do_munmap(mm, start, len, &uf, downgrade); uf 2870 mm/mmap.c userfaultfd_unmap_complete(mm, &uf); uf 2992 mm/mmap.c static int do_brk_flags(unsigned long addr, unsigned long len, unsigned long flags, struct list_head *uf) uf 3018 mm/mmap.c if (do_munmap(mm, addr, len, uf)) uf 3070 mm/mmap.c LIST_HEAD(uf); uf 3081 mm/mmap.c ret = do_brk_flags(addr, len, flags, &uf); uf 3084 mm/mmap.c userfaultfd_unmap_complete(mm, &uf); uf 321 mm/mremap.c bool *locked, struct vm_userfaultfd_ctx *uf, uf 381 mm/mremap.c mremap_userfaultfd_prep(new_vma, uf); uf 500 mm/mremap.c struct vm_userfaultfd_ctx *uf, uf 564 mm/mremap.c ret = move_vma(vma, addr, old_len, new_len, new_addr, locked, uf, uf 605 mm/mremap.c struct vm_userfaultfd_ctx uf = NULL_VM_UFFD_CTX; uf 646 mm/mremap.c &locked, &uf, &uf_unmap_early, &uf_unmap); uf 724 mm/mremap.c &locked, &uf, &uf_unmap); uf 738 mm/mremap.c mremap_userfaultfd_complete(&uf, addr, new_addr, old_len); uf 1106 mm/nommu.c struct list_head *uf) uf 1485 mm/nommu.c int do_munmap(struct mm_struct *mm, unsigned long start, size_t len, struct list_head *uf) uf 490 mm/util.c LIST_HEAD(uf); uf 497 mm/util.c &populate, &uf); uf 499 mm/util.c userfaultfd_unmap_complete(mm, &uf); uf 1819 net/bluetooth/hci_sock.c struct hci_ufilter uf = { .opcode = 0 }; uf 1864 net/bluetooth/hci_sock.c uf.type_mask = f->type_mask; uf 1865 net/bluetooth/hci_sock.c uf.opcode = f->opcode; uf 1866 net/bluetooth/hci_sock.c uf.event_mask[0] = *((u32 *) f->event_mask + 0); uf 1867 net/bluetooth/hci_sock.c uf.event_mask[1] = *((u32 *) f->event_mask + 1); uf 1870 net/bluetooth/hci_sock.c len = min_t(unsigned int, len, sizeof(uf)); uf 1871 net/bluetooth/hci_sock.c if (copy_from_user(&uf, optval, len)) { uf 1877 net/bluetooth/hci_sock.c uf.type_mask &= hci_sec_filter.type_mask; uf 1878 net/bluetooth/hci_sock.c uf.event_mask[0] &= *((u32 *) hci_sec_filter.event_mask + 0); uf 1879 net/bluetooth/hci_sock.c uf.event_mask[1] &= *((u32 *) hci_sec_filter.event_mask + 1); uf 1885 net/bluetooth/hci_sock.c f->type_mask = uf.type_mask; uf 1886 net/bluetooth/hci_sock.c f->opcode = uf.opcode; uf 1887 net/bluetooth/hci_sock.c *((u32 *) f->event_mask + 0) = uf.event_mask[0]; uf 1888 net/bluetooth/hci_sock.c *((u32 *) f->event_mask + 1) = uf.event_mask[1]; uf 1905 net/bluetooth/hci_sock.c struct hci_ufilter uf; uf 1949 net/bluetooth/hci_sock.c memset(&uf, 0, sizeof(uf)); uf 1950 net/bluetooth/hci_sock.c uf.type_mask = f->type_mask; uf 1951 net/bluetooth/hci_sock.c uf.opcode = f->opcode; uf 1952 net/bluetooth/hci_sock.c uf.event_mask[0] = *((u32 *) f->event_mask + 0); uf 1953 net/bluetooth/hci_sock.c uf.event_mask[1] = *((u32 *) f->event_mask + 1); uf 1956 net/bluetooth/hci_sock.c len = min_t(unsigned int, len, sizeof(uf)); uf 1957 net/bluetooth/hci_sock.c if (copy_to_user(optval, &uf, len))