/linux-4.4.14/fs/nfsd/ |
H A D | nfs4acl.c | 189 summarize_posix_acl(struct posix_acl *acl, struct posix_acl_summary *pas) summarize_posix_acl() argument 194 * Only pas.users and pas.groups need initialization; previous summarize_posix_acl() 198 memset(pas, 0, sizeof(*pas)); summarize_posix_acl() 199 pas->mask = 07; summarize_posix_acl() 206 pas->owner = pa->e_perm; FOREACH_ACL_ENTRY() 209 pas->group = pa->e_perm; FOREACH_ACL_ENTRY() 212 pas->users |= pa->e_perm; FOREACH_ACL_ENTRY() 215 pas->groups |= pa->e_perm; FOREACH_ACL_ENTRY() 218 pas->other = pa->e_perm; FOREACH_ACL_ENTRY() 221 pas->mask = pa->e_perm; FOREACH_ACL_ENTRY() 226 pas->users &= pas->mask; 227 pas->group &= pas->mask; 228 pas->groups &= pas->mask; 238 struct posix_acl_summary pas; _posix_to_nfsv4_one() local 244 summarize_posix_acl(pacl, &pas); _posix_to_nfsv4_one() 250 deny = ~pas.owner; _posix_to_nfsv4_one() 255 deny &= pas.users | pas.group | pas.groups | pas.other; _posix_to_nfsv4_one() 274 deny = ~(pa->e_perm & pas.mask); _posix_to_nfsv4_one() 275 deny &= pas.groups | pas.group | pas.other; _posix_to_nfsv4_one() 287 ace->access_mask = mask_from_posix(pa->e_perm & pas.mask, _posix_to_nfsv4_one() 305 ace->access_mask = mask_from_posix(pas.group, flags); _posix_to_nfsv4_one() 314 ace->access_mask = mask_from_posix(pa->e_perm & pas.mask, _posix_to_nfsv4_one() 327 deny = ~pas.group & pas.other; _posix_to_nfsv4_one() 339 deny = ~(pa->e_perm & pas.mask); _posix_to_nfsv4_one() 340 deny &= pas.other; _posix_to_nfsv4_one()
|
/linux-4.4.14/mm/ |
H A D | highmem.c | 416 struct page_address_slot *pas; page_address() local 421 pas = page_slot(page); page_address() 423 spin_lock_irqsave(&pas->lock, flags); page_address() 424 if (!list_empty(&pas->lh)) { page_address() 427 list_for_each_entry(pam, &pas->lh, list) { page_address() 435 spin_unlock_irqrestore(&pas->lock, flags); page_address() 449 struct page_address_slot *pas; set_page_address() local 454 pas = page_slot(page); set_page_address() 460 spin_lock_irqsave(&pas->lock, flags); set_page_address() 461 list_add_tail(&pam->list, &pas->lh); set_page_address() 462 spin_unlock_irqrestore(&pas->lock, flags); set_page_address() 464 spin_lock_irqsave(&pas->lock, flags); set_page_address() 465 list_for_each_entry(pam, &pas->lh, list) { set_page_address() 468 spin_unlock_irqrestore(&pas->lock, flags); set_page_address() 472 spin_unlock_irqrestore(&pas->lock, flags); set_page_address()
|
/linux-4.4.14/drivers/infiniband/hw/mlx5/ |
H A D | mem.c | 146 * pas - bus addresses array to fill 152 __be64 *pas, int access_flags) __mlx5_ib_populate_pas() 173 pas[i] = cpu_to_be64(umem_dma_to_mtt(pa)); __mlx5_ib_populate_pas() 188 pas[i >> shift] = cpu_to_be64(cur); __mlx5_ib_populate_pas() 189 mlx5_ib_dbg(dev, "pas[%d] 0x%llx\n", __mlx5_ib_populate_pas() 190 i >> shift, be64_to_cpu(pas[i >> shift])); __mlx5_ib_populate_pas() 200 int page_shift, __be64 *pas, int access_flags) mlx5_ib_populate_pas() 203 ib_umem_num_pages(umem), pas, mlx5_ib_populate_pas() 150 __mlx5_ib_populate_pas(struct mlx5_ib_dev *dev, struct ib_umem *umem, int page_shift, size_t offset, size_t num_pages, __be64 *pas, int access_flags) __mlx5_ib_populate_pas() argument 199 mlx5_ib_populate_pas(struct mlx5_ib_dev *dev, struct ib_umem *umem, int page_shift, __be64 *pas, int access_flags) mlx5_ib_populate_pas() argument
|
H A D | mr.c | 773 __be64 *pas; reg_umr() local 794 * To avoid copying garbage after the pas array, we allocate reg_umr() 803 pas = PTR_ALIGN(mr_pas, MLX5_UMR_ALIGN); reg_umr() 804 mlx5_ib_populate_pas(dev, umem, page_shift, pas, MLX5_IB_MTT_PRESENT); reg_umr() 806 memset(pas + npages, 0, size - npages * sizeof(u64)); reg_umr() 808 dma = dma_map_single(ddev, pas, size, DMA_TO_DEVICE); reg_umr() 865 __be64 *pas; mlx5_ib_update_mtt() local 893 * code, when we are called from an invalidation. The pas buffer must mlx5_ib_update_mtt() 895 pas = (__be64 *)get_zeroed_page(GFP_ATOMIC); mlx5_ib_update_mtt() 896 if (!pas) { mlx5_ib_update_mtt() 898 pas = mlx5_ib_update_mtt_emergency_buffer; mlx5_ib_update_mtt() 902 memset(pas, 0, size); mlx5_ib_update_mtt() 905 dma = dma_map_single(ddev, pas, size, DMA_TO_DEVICE); mlx5_ib_update_mtt() 923 start_page_index, npages, pas, mlx5_ib_update_mtt() 927 memset(pas + npages, 0, size - npages * sizeof(u64)); mlx5_ib_update_mtt() 969 free_page((unsigned long)pas); mlx5_ib_update_mtt() 993 inlen = sizeof(*in) + sizeof(*in->pas) * ((npages + 1) / 2) * 2; reg_create() 999 mlx5_ib_populate_pas(dev, umem, page_shift, in->pas, reg_create()
|
H A D | srq.c | 122 *inlen = sizeof(**in) + sizeof(*(*in)->pas) * ncont; create_srq_user() 129 mlx5_ib_populate_pas(dev, srq->umem, page_shift, (*in)->pas, 0); create_srq_user() 188 *inlen = sizeof(**in) + sizeof(*(*in)->pas) * npages; create_srq_kernel() 194 mlx5_fill_page_array(&srq->buf, (*in)->pas); create_srq_kernel()
|
H A D | cq.c | 658 *inlen = sizeof(**cqb) + sizeof(*(*cqb)->pas) * ncont; create_cq_user() 664 mlx5_ib_populate_pas(dev, cq->buf.umem, page_shift, (*cqb)->pas, 0); create_cq_user() 719 *inlen = sizeof(**cqb) + sizeof(*(*cqb)->pas) * cq->buf.buf.npages; create_cq_kernel() 725 mlx5_fill_page_array(&cq->buf.buf, (*cqb)->pas); create_cq_kernel() 1123 inlen = sizeof(*in) + npas * sizeof(in->pas[0]); mlx5_ib_resize_cq() 1132 in->pas, 0); mlx5_ib_resize_cq() 1134 mlx5_fill_page_array(&cq->resize_buf->buf, in->pas); mlx5_ib_resize_cq()
|
H A D | mlx5_ib.h | 596 __be64 *pas, int access_flags); 598 int page_shift, __be64 *pas, int access_flags);
|
H A D | qp.c | 669 *inlen = sizeof(**in) + sizeof(*(*in)->pas) * ncont; create_user_qp() 676 mlx5_ib_populate_pas(dev, qp->umem, page_shift, (*in)->pas, 0); create_user_qp() 770 *inlen = sizeof(**in) + sizeof(*(*in)->pas) * qp->buf.npages; create_kernel_qp() 783 mlx5_fill_page_array(&qp->buf, (*in)->pas); create_kernel_qp()
|
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | pagealloc.c | 86 __be64 pas[0]; member in struct:mlx5_manage_pages_inbox 93 __be64 pas[0]; member in struct:mlx5_manage_pages_outbox 313 inlen = sizeof(*in) + npages * sizeof(in->pas[0]); give_pages() 333 in->pas[i] = cpu_to_be64(addr); give_pages() 362 free_4k(dev, be64_to_cpu(in->pas[i])); give_pages() 385 outlen = sizeof(*out) + npages * sizeof(out->pas[0]); reclaim_pages() 412 addr = be64_to_cpu(out->pas[i]); reclaim_pages() 485 FIELD_SIZEOF(struct mlx5_manage_pages_outbox, pas[0]); optimal_reclaimed_pages()
|
H A D | srq.c | 226 void *pas; create_xrc_srq_cmd() local 240 pas = MLX5_ADDR_OF(create_xrc_srq_in, create_in, pas); create_xrc_srq_cmd() 243 memcpy(pas, in->pas, pas_size); create_xrc_srq_cmd() 349 memcpy(MLX5_ADDR_OF(rmpc, rmpc, wq.pas), in->pas, pas_size); create_rmp_cmd()
|
H A D | alloc.c | 200 void mlx5_fill_page_array(struct mlx5_buf *buf, __be64 *pas) mlx5_fill_page_array() argument 208 pas[i] = cpu_to_be64(addr); mlx5_fill_page_array()
|
H A D | eq.c | 356 inlen = sizeof(*in) + sizeof(in->pas[0]) * eq->buf.npages; mlx5_create_map_eq() 364 mlx5_fill_page_array(&eq->buf, in->pas); mlx5_create_map_eq()
|
H A D | en_main.c | 401 (__be64 *)MLX5_ADDR_OF(wq, wq, pas)); mlx5e_enable_rq() 634 (__be64 *)MLX5_ADDR_OF(wq, wq, pas)); mlx5e_enable_sq() 818 (__be64 *)MLX5_ADDR_OF(create_cq_in, in, pas)); mlx5e_enable_cq()
|
/linux-4.4.14/include/linux/mlx5/ |
H A D | device.h | 703 __be64 pas[0]; member in struct:mlx5_create_srq_mbox_in 734 __be64 pas[0]; member in struct:mlx5_query_srq_mbox_out 777 __be64 pas[0]; member in struct:mlx5_create_cq_mbox_in 808 __be64 pas[0]; member in struct:mlx5_query_cq_mbox_out 817 __be64 pas[0]; member in struct:mlx5_modify_cq_mbox_in 870 __be64 pas[0]; member in struct:mlx5_create_eq_mbox_in 965 __be64 pas[0]; member in struct:mlx5_create_mkey_mbox_in 992 __be64 pas[0]; member in struct:mlx5_query_mkey_mbox_out 998 __be64 pas[0]; member in struct:mlx5_modify_mkey_mbox_in
|
H A D | qp.h | 514 __be64 pas[0]; member in struct:mlx5_create_qp_mbox_in 561 __be64 pas[0]; member in struct:mlx5_query_qp_mbox_out
|
H A D | mlx5_ifc.h | 888 struct mlx5_ifc_cmd_pas_bits pas[0]; member in struct:mlx5_ifc_wq_bits 2907 u8 pas[0][0x40]; member in struct:mlx5_ifc_query_xrc_srq_out_bits 3077 u8 pas[0][0x40]; member in struct:mlx5_ifc_query_srq_out_bits 3250 u8 pas[0][0x40]; member in struct:mlx5_ifc_query_qp_out_bits 3719 u8 pas[0][0x40]; member in struct:mlx5_ifc_query_eq_out_bits 3773 u8 pas[0][0x40]; member in struct:mlx5_ifc_query_cq_out_bits 4312 u8 pas[0][0x40]; member in struct:mlx5_ifc_modify_cq_in_bits 4376 u8 pas[0][0x40]; member in struct:mlx5_ifc_manage_pages_out_bits 4397 u8 pas[0][0x40]; member in struct:mlx5_ifc_manage_pages_in_bits 5219 u8 pas[0][0x40]; member in struct:mlx5_ifc_create_xrc_srq_in_bits 5295 u8 pas[0][0x40]; member in struct:mlx5_ifc_create_srq_in_bits 5423 u8 pas[0][0x40]; member in struct:mlx5_ifc_create_qp_in_bits 5610 u8 pas[0][0x40]; member in struct:mlx5_ifc_create_eq_in_bits 5664 u8 pas[0][0x40]; member in struct:mlx5_ifc_create_cq_in_bits
|
H A D | driver.h | 750 void mlx5_fill_page_array(struct mlx5_buf *buf, __be64 *pas);
|
/linux-4.4.14/drivers/cpufreq/ |
H A D | pasemi-cpufreq.c | 260 .name = "pas-cpufreq",
|
/linux-4.4.14/drivers/misc/vmw_vmci/ |
H A D | vmci_queue_pair.c | 157 dma_addr_t *pas; member in struct:vmci_queue_kern_if::__anon6358::__anon6359 282 queue->kernel_if->u.g.pas[i]); qp_free_queue() 305 (sizeof(*queue->kernel_if->u.g.pas) + qp_alloc_queue() 309 pas_size = num_pages * sizeof(*queue->kernel_if->u.g.pas); qp_alloc_queue() 322 queue->kernel_if->u.g.pas = (dma_addr_t *)(queue->kernel_if + 1); qp_alloc_queue() 324 (void **)((u8 *)queue->kernel_if->u.g.pas + pas_size); qp_alloc_queue() 330 &queue->kernel_if->u.g.pas[i], qp_alloc_queue() 505 produce_q->kernel_if->u.g.pas[i] >> PAGE_SHIFT; qp_alloc_ppn_set() 518 consume_q->kernel_if->u.g.pas[i] >> PAGE_SHIFT; qp_alloc_ppn_set()
|
/linux-4.4.14/drivers/char/ |
H A D | applicom.c | 10 /* LoopCount n'etait pas initialise a 0. */
|
/linux-4.4.14/sound/pci/lx6464es/ |
H A D | lx6464es.c | 607 /* le xilinx ES peut ne pas etre encore pret, on attend. */ lx_init_xilinx_reset()
|