/linux-4.1.27/tools/testing/selftests/powerpc/pmu/ebb/ |
H A D | cpu_event_pinned_vs_ebb_test.c | 18 * Tests a pinned cpu event vs an EBB - in that order. The pinned cpu event 26 event->attr.pinned = 1; setup_cpu_event()
|
H A D | task_event_pinned_vs_ebb_test.c | 18 * Tests a pinned per-task event vs an EBB - in that order. The pinned per-task 26 event->attr.pinned = 1; setup_child_event()
|
H A D | event_attributes_test.c | 115 leader.attr.pinned = 0; event_attributes() 116 /* Expected to fail, leader isn't pinned */ event_attributes()
|
H A D | ebb.c | 337 e->attr.pinned = 1; event_leader_ebb_init()
|
/linux-4.1.27/kernel/events/ |
H A D | hw_breakpoint.c | 53 /* Number of pinned cpu breakpoints in a cpu */ 57 /* Number of non-pinned cpu/task breakpoints in a cpu */ 74 /* Gather the number of total pinned and un-pinned bp in a cpuset */ 76 unsigned int pinned; member in struct:bp_busy_slots 97 * Report the maximum number of pinned breakpoints a task 141 * Report the number of pinned/un-pinned breakpoints we have in 161 if (nr > slots->pinned) for_each_cpu() 162 slots->pinned = nr; for_each_cpu() 171 * For now, continue to consider flexible as pinned, until we can 172 * ensure no flexible event can ever be scheduled before a pinned event 178 slots->pinned += weight; fetch_this_slot() 182 * Add a pinned breakpoint for the given task in our constraint table 242 * == Non-pinned counter == (Considered as pinned for now) 249 * -> If there are already non-pinned counters in this cpu, it means 306 if (slots.pinned + (!!slots.flexible) > nr_slots[type]) __reserve_bp_slot()
|
H A D | core.c | 430 * when we know the cgroup is pinned (css_get) update_cgrp_time_from_event() 453 * unless we have the cgroup pinned (css_get) perf_cgroup_set_timestamp() 1218 if (event->attr.pinned) ctx_group_list() 2260 if (leader->attr.pinned) { __perf_event_enable() 2717 * If this pinned group hasn't been scheduled, ctx_pinned_sched_in() 2773 * First go through the list and put on any pinned groups ctx_sched_in() 2806 * cpu pinned (that don't need to move), task pinned, perf_event_context_sched_in() 3039 * Rotate the first entry last of non-pinned groups. Rotation might be rotate_ctx() 3882 * error state (i.e. because it was pinned but it couldn't be perf_read_hw() 4750 * Since pinned accounting is per vm we cannot allow fork() to copy our perf_mmap() 8106 * Only a group leader can be exclusive or pinned SYSCALL_DEFINE5() 8108 if (attr.exclusive || attr.pinned) SYSCALL_DEFINE5()
|
/linux-4.1.27/include/trace/events/ |
H A D | xen.h | 328 TP_PROTO(struct mm_struct *mm, unsigned long pfn, unsigned level, bool pinned), 329 TP_ARGS(mm, pfn, level, pinned), 334 __field(bool, pinned) 339 __entry->pinned = pinned), 342 __entry->pinned ? "" : "un") 346 TP_PROTO(unsigned long pfn, unsigned level, bool pinned), 347 TP_ARGS(pfn, level, pinned), 351 __field(bool, pinned) 355 __entry->pinned = pinned), 358 __entry->pinned ? "" : "un")
|
/linux-4.1.27/fs/btrfs/ |
H A D | inode-map.c | 212 struct btrfs_free_space_ctl *pinned = root->free_ino_pinned; btrfs_return_ino() local 218 __btrfs_add_free_space(pinned, objectid, 1); btrfs_return_ino() 231 __btrfs_add_free_space(pinned, objectid, 1); btrfs_return_ino() 239 * smaller than root->ino_cache_progress from pinned tree to free_ino tree, and 352 * - The pinned tree is only used during the process of caching pinned_use_bitmap() 367 struct btrfs_free_space_ctl *pinned = root->free_ino_pinned; btrfs_init_free_ino_ctl() local 384 spin_lock_init(&pinned->tree_lock); btrfs_init_free_ino_ctl() 385 pinned->unit = 1; btrfs_init_free_ino_ctl() 386 pinned->start = 0; btrfs_init_free_ino_ctl() 387 pinned->private = NULL; btrfs_init_free_ino_ctl() 388 pinned->extents_thresh = 0; btrfs_init_free_ino_ctl() 389 pinned->op = &pinned_free_ino_op; btrfs_init_free_ino_ctl()
|
H A D | extent-tree.c | 137 WARN_ON(cache->pinned > 0); btrfs_put_block_group() 3320 * Keep in mind this has to include any pinned space which could end up cache_save_setup() 3963 * If we don't have enough pinned space to deal with this btrfs_check_data_free_space() 4445 /* See if there is enough pinned space to make this reservation */ may_commit_transaction() 5679 cache->pinned += num_bytes; update_block_group() 5748 cache->pinned += num_bytes; pin_down_extent() 5798 * the pinned extents. btrfs_pin_extent_for_log_replay() 6011 cache->pinned -= len; unpin_extent_range() 6531 /* unlocks the pinned mutex */ btrfs_free_extent() 7144 printk(KERN_INFO "BTRFS: space_info total=%llu, used=%llu, pinned=%llu, " dump_space_info() 7160 "%llu used %llu pinned %llu reserved %s\n", dump_space_info() 7162 btrfs_block_group_used(&cache->item), cache->pinned, dump_space_info() 8788 num_bytes = cache->key.offset - cache->reserved - cache->pinned - set_block_group_ro() 8938 num_bytes = cache->key.offset - cache->reserved - cache->pinned - btrfs_set_block_group_rw() 9979 * We could have pending pinned extents for this block group, btrfs_delete_unused_bgs() 10012 /* Reset pinned so btrfs_put_block_group doesn't complain */ btrfs_delete_unused_bgs() 10016 space_info->bytes_pinned -= block_group->pinned; btrfs_delete_unused_bgs() 10017 space_info->bytes_readonly += block_group->pinned; btrfs_delete_unused_bgs() 10019 -block_group->pinned); btrfs_delete_unused_bgs() 10020 block_group->pinned = 0; btrfs_delete_unused_bgs()
|
H A D | free-space-cache.c | 633 * Since we attach pinned extents after the fact we can have contiguous sections 636 * entries since we would have merged the entries when adding the pinned extents 1050 * We want to add any pinned extents to our free space cache write_pinned_extent_entries() 1053 * We shouldn't have switched the pinned extents yet so this is the write_pinned_extent_entries() 1067 /* This pinned extent is out of our range */ write_pinned_extent_entries() 1280 * Some spaces that are freed in the current transaction are pinned, __btrfs_write_out_cache()
|
H A D | ctree.h | 1149 u64 bytes_pinned; /* total bytes pinned, will be freed when the 1173 * bytes_pinned is kept in line with what is actually pinned, as in 1176 * bytes_pinned does not reflect the bytes that will be pinned once the 1285 u64 pinned; member in struct:btrfs_block_group_cache
|
H A D | relocation.c | 4080 /* get rid of pinned extents */ relocate_block_group() 4283 WARN_ON(rc->block_group->pinned > 0); btrfs_relocate_block_group()
|
/linux-4.1.27/drivers/staging/android/uapi/ |
H A D | ashmem.h | 25 /* Return values from ASHMEM_GET_PIN_STATUS: Is the mapping pinned? */
|
/linux-4.1.27/drivers/gpu/drm/i915/ |
H A D | i915_gem_shrinker.c | 260 unsigned long pinned, bound, unbound, freed_pages; i915_gem_shrinker_oom() local 282 * assert that there are no objects with pinned pages that are not i915_gem_shrinker_oom() 285 unbound = bound = pinned = 0; i915_gem_shrinker_oom() 291 pinned += obj->base.size; i915_gem_shrinker_oom() 300 pinned += obj->base.size; i915_gem_shrinker_oom() 309 pr_info("Purging GPU memory, %lu bytes freed, %lu bytes still pinned.\n", i915_gem_shrinker_oom() 310 freed_pages << PAGE_SHIFT, pinned); i915_gem_shrinker_oom()
|
H A D | i915_gem_userptr.c | 553 int pinned, ret; __i915_gem_userptr_get_pages_worker() local 556 pinned = 0; __i915_gem_userptr_get_pages_worker() 566 while (pinned < num_pages) { __i915_gem_userptr_get_pages_worker() 568 obj->userptr.ptr + pinned * PAGE_SIZE, __i915_gem_userptr_get_pages_worker() 569 num_pages - pinned, __i915_gem_userptr_get_pages_worker() 571 pvec + pinned, NULL); __i915_gem_userptr_get_pages_worker() 575 pinned += ret; __i915_gem_userptr_get_pages_worker() 583 } else if (pinned == num_pages) { __i915_gem_userptr_get_pages_worker() 587 pinned = 0; __i915_gem_userptr_get_pages_worker() 596 release_pages(pvec, pinned, 0); __i915_gem_userptr_get_pages_worker() 608 int pinned, ret; i915_gem_userptr_get_pages() local 628 pinned = 0; i915_gem_userptr_get_pages() 638 pinned = __get_user_pages_fast(obj->userptr.ptr, num_pages, i915_gem_userptr_get_pages() 641 if (pinned < num_pages) { i915_gem_userptr_get_pages() 642 if (pinned < 0) { i915_gem_userptr_get_pages() 643 ret = pinned; i915_gem_userptr_get_pages() 644 pinned = 0; i915_gem_userptr_get_pages() 696 pinned = 0; i915_gem_userptr_get_pages() 700 release_pages(pvec, pinned, 0); i915_gem_userptr_get_pages() 780 * restriction is that we do not allow userptr surfaces to be pinned to the
|
H A D | i915_gem_evict.c | 67 * ignores pinned vmas, and not object where the backing storage itself is 68 * pinned. Hence obj->pages_pin_count does not protect against eviction.
|
H A D | i915_gpu_error.c | 50 static const char *pin_flag(int pinned) pin_flag() argument 52 if (pinned > 0) pin_flag() 54 else if (pinned < 0) pin_flag() 204 err_puts(m, pin_flag(err->pinned)); print_error_buffers() 691 err->pinned = 0; capture_bo() 693 err->pinned = 1; capture_bo()
|
H A D | i915_gem_gtt.h | 189 * How many users have pinned this object in GTT space. The following 193 * framebuffer code has at most two buffers pinned per crtc.
|
H A D | i915_gem.c | 153 size_t pinned; i915_gem_get_aperture_ioctl() local 155 pinned = 0; i915_gem_get_aperture_ioctl() 159 pinned += i915_gem_obj_ggtt_size(obj); i915_gem_get_aperture_ioctl() 163 args->aper_available_size = args->aper_size - pinned; i915_gem_get_aperture_ioctl() 2064 /* Get the list of pages out of our struct file. They'll be pinned i915_gem_object_get_pages_gtt() 2153 * and pinned into our object. i915_gem_object_get_pages() may be called 2573 * pinned in place. i915_gem_reset_ring_cleanup() 3557 /* If we don't have a page list set up, then we're not pinned i915_gem_clflush_object() 3726 DRM_DEBUG("can not change the cache level of pinned objects\n"); i915_gem_object_set_cache_level() 3915 /* As the user may map the buffer once pinned in the display plane i915_gem_object_pin_to_display_plane() 4141 "bo is already pinned in %s with incorrect alignment:" i915_gem_object_do_pin()
|
H A D | i915_drv.h | 508 s32 pinned:2; member in struct:drm_i915_error_state::drm_i915_error_buffer 2371 /* Early gen2 have a totally busted CS tlb and require pinned batches. */
|
H A D | i915_gem_execbuffer.c | 1589 * - The batch is already pinned into the relevant ppgtt, so we i915_gem_do_execbuffer()
|
H A D | intel_lrc.c | 1777 /* Ring buffer start address is not known until the buffer is pinned. populate_lr_context()
|
H A D | i915_debugfs.c | 146 seq_printf(m, " (pinned x %d)", pin_count); describe_obj() 463 seq_printf(m, "%u pinned mappable objects, %zu bytes\n", i915_gem_object_info()
|
/linux-4.1.27/drivers/gpu/drm/qxl/ |
H A D | qxl_object.c | 53 void qxl_ttm_placement_from_domain(struct qxl_bo *qbo, u32 domain, bool pinned) qxl_ttm_placement_from_domain() argument 56 u32 pflag = pinned ? TTM_PL_FLAG_NO_EVICT : 0; qxl_ttm_placement_from_domain() 79 unsigned long size, bool kernel, bool pinned, u32 domain, qxl_bo_create() 102 bo->pin_count = pinned ? 1 : 0; qxl_bo_create() 109 qxl_ttm_placement_from_domain(bo, domain, pinned); qxl_bo_create() 78 qxl_bo_create(struct qxl_device *qdev, unsigned long size, bool kernel, bool pinned, u32 domain, struct qxl_surface *surf, struct qxl_bo **bo_ptr) qxl_bo_create() argument
|
H A D | qxl_object.h | 89 bool kernel, bool pinned, u32 domain, 100 extern void qxl_ttm_placement_from_domain(struct qxl_bo *qbo, u32 domain, bool pinned);
|
H A D | qxl_release.c | 263 since these objects are pinned no need to reserve */ qxl_release_reserve_list() 287 since these objects are pinned no need to reserve */ qxl_release_backoff_reserve_list() 444 since these objects are pinned no need to reserve */ qxl_release_fence_buffer_objects()
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | cdmm.h | 37 * Any CPU pinned threads/timers should be disabled. 39 * CPU pinned threads/timers can be restarted.
|
/linux-4.1.27/include/linux/ |
H A D | hw_breakpoint.h | 18 * As it's for in-kernel or ptrace use, we want it to be pinned hw_breakpoint_init() 21 attr->pinned = 1; hw_breakpoint_init()
|
H A D | host1x.h | 200 /* Array of handles to be pinned & unpinned */
|
H A D | page-flags.h | 220 PAGEFLAG(Pinned, pinned) TESTSCFLAG(Pinned, pinned) /* Xen */
|
H A D | pagemap.h | 156 * found a page in the radix tree here, we have pinned its refcount by page_cache_get_speculative() 445 * lock_page may only be called if we have the page's inode pinned.
|
H A D | dcache.h | 553 * dentry or it may be on a lower dentry pinned by the upper. 570 * dentry or it may be a lower dentry pinned by the upper.
|
H A D | fscache-cache.h | 151 void *context; /* netfs read context (pinned) */ 264 /* discard the resources pinned by an object and effect retirement if
|
H A D | mm_types.h | 224 * conditions. These are held in a global tree and are pinned by the VMAs that
|
H A D | perf_event.h | 600 * Must ensure cgroup is pinned (css_get) before calling
|
H A D | fscache.h | 413 * Permit data-storage cache objects to be pinned in the cache.
|
H A D | list.h | 581 * exception to this is if the cursor element (pos) is pinned in the list,
|
H A D | sched.h | 338 extern int get_nohz_timer_target(int pinned); 342 static inline int get_nohz_timer_target(int pinned) get_nohz_timer_target() argument
|
/linux-4.1.27/drivers/gpu/drm/msm/ |
H A D | msm_gem_prime.c | 26 BUG_ON(!msm_obj->sgt); /* should have already pinned! */ msm_gem_prime_get_sg_table()
|
/linux-4.1.27/arch/frv/kernel/ |
H A D | sysctl.c | 121 * permit the mm_struct the nominated process is using have its MMU context ID pinned 154 /* read the currently pinned CXN */ procctl_frv_pin_cxnr()
|
/linux-4.1.27/drivers/oprofile/ |
H A D | nmi_timer_int.c | 24 .pinned = 1,
|
H A D | oprofile_perf.c | 54 * settings in counter_config. Attributes are created as `pinned' events and 70 attr->pinned = 1; op_perf_setup()
|
/linux-4.1.27/arch/microblaze/kernel/ |
H A D | misc.S | 26 * We avoid flushing the pinned 0, 1 and possibly 2 entries.
|
/linux-4.1.27/arch/sh/mm/ |
H A D | gup.c | 206 * @pages: array that receives pointers to the pages pinned. 213 * Returns number of pages pinned. This may be fewer than the number 215 * were pinned, returns -errno.
|
/linux-4.1.27/arch/s390/mm/ |
H A D | gup.c | 209 * @pages: array that receives pointers to the pages pinned. 216 * Returns number of pages pinned. This may be fewer than the number 218 * were pinned, returns -errno.
|
/linux-4.1.27/arch/mips/mm/ |
H A D | gup.c | 252 * @pages: array that receives pointers to the pages pinned. 259 * Returns number of pages pinned. This may be fewer than the number 261 * were pinned, returns -errno.
|
/linux-4.1.27/tools/testing/selftests/memfd/ |
H A D | fuse_test.c | 10 * ADD_SEALS must fail if the memfd pages are still pinned. Note that we use 267 * This guarantees that the receive-buffer is pinned for 1s until the main() 269 * the pages are still pinned. */ main()
|
/linux-4.1.27/arch/x86/xen/ |
H A D | mmu.c | 274 /* If page is not pinned, we can just update the entry xen_set_pmd() 467 /* If page is not pinned, we can just update the entry xen_set_pud() 549 * 2. It is always pinned 571 /* If page is not pinned, we can just update the entry xen_set_pgd() 582 /* If it's pinned, then we can at least batch the kernel and xen_set_pgd() 745 flush = 0; /* already pinned */ xen_pin_page() 764 * pinned, so Xen won't do the trap'n'emulate. xen_pin_page() 800 read-only, and can be pinned. */ __xen_pgd_pin() 871 * The init_mm pagetable is really pinned as soon as its created, but 903 * partially-pinned state. xen_unpin_page() 966 * kernel doesn't see any unexpected pinned pagetables. 1087 /* pgd may not be pinned in the error exit path of execve */ xen_exit_mmap() 1478 * this point it is not possible to tell if a page is pinned or not, 1502 everything is pinned. */ xen_alloc_pte_init() 1521 /* Early release_pte assumes that all pts are pinned, since there's 1522 only init_mm and anything attached to that is pinned. */ xen_release_pte_init() 1557 /* This needs to make sure the new pte page is pinned iff its being 1558 attached to a pinned pagetable. */ xen_alloc_ptpage() 1562 bool pinned = PagePinned(virt_to_page(mm->pgd)); xen_alloc_ptpage() local 1564 trace_xen_mmu_alloc_ptpage(mm, pfn, level, pinned); xen_alloc_ptpage() 1566 if (pinned) { xen_alloc_ptpage() 1602 bool pinned = PagePinned(page); xen_release_ptpage() local 1604 trace_xen_mmu_release_ptpage(pfn, level, pinned); xen_release_ptpage() 1606 if (pinned) { xen_release_ptpage()
|
/linux-4.1.27/mm/ |
H A D | mlock.c | 96 * Assumes lru_lock already held and page already pinned. 350 * a non-TPH page already pinned and in the @pvec, and that it belongs to @zone. 354 * pages also get pinned. 367 * Initialize pte walk starting at the already pinned page where we __munlock_pagevec_fill() 368 * are sure that there is a pte, as it was pinned under the same __munlock_pagevec_fill() 377 /* The page next to the pinned page is the first we will try to get */ __munlock_pagevec_fill()
|
H A D | util.c | 244 * page pinned 258 * @pages: array that receives pointers to the pages pinned. 261 * Returns number of pages pinned. This may be fewer than the number 263 * were pinned, returns -errno.
|
H A D | gup.c | 367 * @pages: array that receives pointers to the pages pinned. 374 * Returns number of pages pinned. This may be fewer than the number 376 * were pinned, returns -errno. Each page returned must be released 767 * @pages: array that receives pointers to the pages pinned. 773 * Returns number of pages pinned. This may be fewer than the number 775 * were pinned, returns -errno. Each page returned must be released 943 * Returns struct page pointer of user page pinned for dump, 1338 * @pages: array that receives pointers to the pages pinned. 1345 * Returns number of pages pinned. This may be fewer than the number 1347 * were pinned, returns -errno.
|
H A D | swap.c | 111 * released while the head is still pinned by the slab code 371 * @pages: array that receives pointers to the pages pinned. 374 * Returns number of pages pinned. This may be fewer than the number 376 * were pinned, returns -errno. Each page returned must be released 400 * @pages: array that receives pointer to the page pinned. 403 * Returns 1 if page is pinned. If the page was not pinned, returns
|
H A D | mmu_notifier.c | 289 * so mm has to be current->mm or the mm should be pinned safely such 293 * unregister the notifier. mm_count is automatically pinned to allow
|
H A D | madvise.c | 441 * typically, to avoid COWing pages pinned by get_user_pages().
|
H A D | rmap.c | 347 * process it belongs to. The root anon_vma needs to be pinned until anon_vma_fork() 536 /* we pinned the anon_vma, its safe to sleep */ page_lock_anon_vma_read()
|
H A D | page_alloc.c | 708 * If the zone was previously in an "all pages pinned" state then look to 712 * pinned" detection logic. 1393 * Note that this function must be called with the thread pinned to 1416 * thread pinned to the current processor or a processor that 1440 * thread pinned to the current processor or a processor that 1455 * The CPU has to be pinned. When zone parameter is non-NULL, spill just 2576 * pages are pinned on the per-cpu lists. Drain them and try again __alloc_pages_direct_reclaim()
|
H A D | vmscan.c | 474 * that page is locked, the mapping is pinned. 2448 * If a zone is deemed to be full of pinned pages then just give it a light 3076 * There is special handling here for zones which are full of pinned pages.
|
H A D | compaction.c | 751 * Migration will fail if an anonymous page is pinned in memory, isolate_migratepages_block()
|
H A D | memory-failure.c | 1013 * We pinned the head page for hwpoison handling, hwpoison_user_mappings()
|
H A D | huge_memory.c | 1745 * to be pinned by the caller. __split_huge_page_refcount()
|
H A D | shmem.c | 1919 * found pages pinned. shmem_wait_for_pins()
|
H A D | memcontrol.c | 2112 * a thread which is pinned to local cpu. 4567 * but the memcg is pinned through dependency. The memcg should stop
|
/linux-4.1.27/tools/perf/tests/ |
H A D | parse-events.c | 486 TEST_ASSERT_VAL("wrong pinned", !evsel->attr.pinned); test__checkevent_pmu_events() 504 TEST_ASSERT_VAL("wrong pinned", !evsel->attr.pinned); test__checkevent_pmu_events_mix() 516 TEST_ASSERT_VAL("wrong pinned", !evsel->attr.pinned); test__checkevent_pmu_events_mix() 1143 TEST_ASSERT_VAL("wrong pinned", evsel->attr.pinned); test__checkevent_pinned_modifier() 1161 TEST_ASSERT_VAL("wrong pinned", evsel->attr.pinned); test__pinned_group() 1163 /* cache-misses - can not be pinned, but will go on with the leader */ test__pinned_group() 1168 TEST_ASSERT_VAL("wrong pinned", !evsel->attr.pinned); test__pinned_group() 1174 TEST_ASSERT_VAL("wrong pinned", !evsel->attr.pinned); test__pinned_group()
|
H A D | attr.c | 98 WRITE_ASS(pinned, "d"); store_event()
|
H A D | attr.py | 37 'pinned',
|
/linux-4.1.27/drivers/gpu/drm/omapdrm/ |
H A D | omap_fbdev.c | 164 /* note: this keeps the bo pinned.. which is perhaps not ideal, omap_fbdev_create() 168 * (non-tiled buffer doesn't need to be pinned for fbcon to write omap_fbdev_create()
|
H A D | omap_plane.c | 53 /* last fb that we pinned: */ 84 /* update which fb (if any) is pinned for scanout */ omap_plane_update_pin()
|
H A D | omap_gem.c | 126 * not necessarily pinned in TILER all the time, and (b) when they are 140 struct drm_gem_object *obj; /* the current pinned obj */ 840 /* Get rotated scanout address (only valid if already pinned), at the 923 * least they would no longer need to be pinned if everyone has omap_gem_put_pages() 1289 /* this means the object is still pinned.. which really should omap_gem_free_object()
|
/linux-4.1.27/net/rds/ |
H A D | info.c | 49 * are pinned for the duration of the copy. 61 * - info source copy must be pinned, may be "large"
|
H A D | ib_rdma.c | 79 atomic_t free_pinned; /* memory pinned by free MRs */ 501 unsigned int pinned = ibmr->sg_len; rds_ib_teardown_mr() local 504 if (pinned) { rds_ib_teardown_mr() 508 atomic_sub(pinned, &pool->free_pinned); rds_ib_teardown_mr() 711 /* If we've pinned too many pages, request a flush */ rds_ib_free_mr()
|
H A D | page.c | 52 * with private pages. If, say, we were to do directed receive to pinned
|
H A D | iw_rdma.c | 70 atomic_t free_pinned; /* memory pinned by free MRs */ 563 /* If we've pinned too many pages, request a flush */ rds_iw_free_mr()
|
/linux-4.1.27/tools/perf/util/ |
H A D | parse-events.c | 716 int pinned; member in struct:event_modifier 730 int pinned = evsel ? evsel->attr.pinned : 0; get_event_modifier() local 768 pinned = 1; get_event_modifier() 797 mod->pinned = pinned; get_event_modifier() 852 evsel->attr.pinned = mod.pinned; __evlist__for_each()
|
H A D | python.c | 524 "pinned", pyrf_evsel__init() 549 pinned = 0, pyrf_evsel__init() local 572 &pinned, &exclusive, &exclude_user, pyrf_evsel__init() 591 attr.pinned = pinned; pyrf_evsel__init()
|
H A D | evsel.c | 1092 PRINT_ATTRf(pinned, p_unsigned); perf_event_attr__fprintf()
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
H A D | radeon_ttm.c | 534 unsigned pinned = 0, nents; radeon_ttm_tt_pin_userptr() local 555 unsigned num_pages = ttm->num_pages - pinned; radeon_ttm_tt_pin_userptr() 556 uint64_t userptr = gtt->userptr + pinned * PAGE_SIZE; radeon_ttm_tt_pin_userptr() 557 struct page **pages = ttm->pages + pinned; radeon_ttm_tt_pin_userptr() 564 pinned += r; radeon_ttm_tt_pin_userptr() 566 } while (pinned < ttm->num_pages); radeon_ttm_tt_pin_userptr() 588 release_pages(ttm->pages, pinned, 0); radeon_ttm_tt_pin_userptr()
|
H A D | radeon_object.h | 88 * Note: object should either be pinned or reserved when calling this
|
H A D | atombios_crtc.c | 1169 /* If atomic, assume fb object is pinned & idle & fenced and dce4_crtc_do_set_base() 1491 /* If atomic, assume fb object is pinned & idle & fenced and avivo_crtc_do_set_base()
|
H A D | radeon.h | 2455 /* tracking pinned memory */
|
/linux-4.1.27/arch/x86/mm/ |
H A D | gup.c | 314 * @pages: array that receives pointers to the pages pinned. 321 * Returns number of pages pinned. This may be fewer than the number 323 * were pinned, returns -errno.
|
/linux-4.1.27/drivers/gpu/drm/ |
H A D | drm_gem_cma_helper.c | 401 * drm_gem_cma_prime_get_sg_table - provide a scatter/gather table of pinned 410 * A pointer to the scatter/gather table of pinned pages or NULL on failure. 437 * driver's scatter/gather table of pinned pages 440 * @sgt: scatter/gather table of pinned pages
|
H A D | drm_prime.c | 318 * - @gem_prime_get_sg_table: provide a scatter/gather table of pinned pages 745 * @sg: the sg-table which was pinned at import time
|
H A D | drm_gem.c | 449 * whole object is covered by the page-array and pinned in memory.
|
/linux-4.1.27/drivers/gpu/host1x/hw/ |
H A D | channel_hw.c | 157 /* end CDMA submit & stash pinned hMems into sync queue */ channel_submit()
|
/linux-4.1.27/arch/arm64/include/asm/ |
H A D | processor.h | 58 /* Hardware breakpoints pinned to this task. */
|
/linux-4.1.27/kernel/ |
H A D | kcmp.c | 55 /* The caller must have pinned the task */
|
H A D | watchdog.c | 269 .pinned = 1,
|
H A D | cgroup.c | 1833 * With the sb pinned, tryget_live can reliably indicate for_each_root() 2942 * doesn't need to be pinned. The RCU locking is not necessary cgroup_file_write() 5424 * or can't be pinned, an ERR_PTR value is returned.
|
H A D | sys.c | 2002 * since it's pinned to mm_struct, but allow prctl_set_mm()
|
/linux-4.1.27/drivers/infiniband/core/ |
H A D | umem.c | 80 * @access: IB_ACCESS_xxx flags for memory being pinned 257 * ib_umem_release - release memory pinned with ib_umem_get
|
/linux-4.1.27/fs/xfs/ |
H A D | xfs_buf.c | 933 * If we come across a stale, pinned, locked buffer, we know that we are 935 * pinned, we know that the log has not been pushed to disk and hence it 958 * If we come across a stale, pinned, locked buffer, we know that we 960 * it is pinned, we know that the log has not been pushed to disk and 1794 int pinned = 0; __xfs_buf_delwri_submit() local 1799 pinned++; list_for_each_entry_safe() 1845 return pinned; 1851 * This will take the @buffer_list, write all non-locked and non-pinned buffers
|
H A D | xfs_extfree_item.c | 131 * While EFIs cannot really be pinned, the unpin operation is the last place at 158 * is very similar to being pinned. Return XFS_ITEM_PINNED so that the caller
|
H A D | xfs_buf_item.c | 376 * is held while the item is pinned in memory. This means that we can 399 * item which was previously pinned with a call to xfs_buf_item_pin(). 523 * If we have just raced with a buffer being pinned and it has xfs_buf_item_push() 526 * race condition here so xfsaild recognizes the buffer is pinned xfs_buf_item_push() 561 * item was logged, then another reference was taken when it was pinned, so we
|
H A D | xfs_icache.c | 863 * clean, pinned(*) 0 requeue 864 * stale, pinned EAGAIN requeue 868 * (*) dgc: I don't think the clean, pinned state is possible but it gets 885 * pinned, async => requeue 886 * pinned, sync => unpin
|
H A D | xfs_dquot_item.c | 94 * dquot must have been previously pinned with a call to
|
H A D | xfs_mount.c | 979 * that has been freed has its buffer still pinned in memory because xfs_unmountfs() 996 * inodes and none should be pinned or locked, but use synchronous xfs_unmountfs()
|
H A D | xfs_trans_ail.c | 343 * If we encountered pinned items or did not finish writing out all xfsaild_push() 477 * spinning on the same items, and if they are pinned will all xfsaild_push()
|
H A D | xfs_inode_item.c | 389 * item which was previously pinned with a call to xfs_inode_item_pin().
|
H A D | xfs_dquot.c | 1033 * If the buffer is pinned then push on the log so we won't xfs_qm_dqflush()
|
H A D | xfs_trans_buf.c | 565 * pinned by another transaction. If that is the case, then we'll wait until
|
H A D | xfs_inode.c | 2459 * in at least shared mode so that the buffer cannot be subsequently pinned 3257 * Try to get locks. If any are unavailable or it is pinned, xfs_iflush_cluster() 3429 * If the buffer is pinned then push on the log now so we won't xfs_iflush()
|
H A D | xfs_log.c | 1023 * tail of the log is pinned by an item that is modified in the CIL. Hence 3888 * c. those who're sleeping on log reservations, pinned objects and
|
/linux-4.1.27/drivers/misc/genwqe/ |
H A D | card_base.h | 185 struct list_head pin_list; /* list of pinned memory for dev */ 413 spinlock_t pin_lock; /* lock for pinned memory */ 414 struct list_head pin_list; /* list of pinned memory */
|
H A D | card_dev.c | 322 spin_lock_init(&cfile->pin_lock); /* list of user pinned memory */ genwqe_open()
|
H A D | card_utils.c | 505 * free_user_pages() - Give pinned pages back
|
/linux-4.1.27/drivers/gpu/drm/gma500/ |
H A D | gtt.c | 158 * Roll an existing pinned mapping by moving the pages through the GTT. 176 the right position when it gets pinned */ psb_gtt_roll() 376 * object has been pinned by mmap users we clean this up here currently.
|
H A D | gem.c | 194 /* For now the mmap pins the object and it stays pinned. As things psb_gem_fault()
|
/linux-4.1.27/drivers/bus/ |
H A D | mips_cdmm.c | 281 * pre-emption or by running from a pinned kernel thread. 340 * pre-emption or by running from a pinned kernel thread. 426 * pre-emption or by running from a pinned kernel thread.
|
/linux-4.1.27/fs/configfs/ |
H A D | dir.c | 1117 * If we succeeded, the fs is pinned via other methods. If not, configfs_depend_item() 1129 * pinned, thus the subsystem is pinned, and therefore configfs is pinned. 1137 * Since we can trust everything is pinned, we just need configfs_undepend_item() 1321 * so the parent is safely pinned. We can drop our working configfs_mkdir()
|
/linux-4.1.27/fs/cachefiles/ |
H A D | interface.c | 248 * discard the resources pinned by an object and effect retirement if 371 /* make sure all pages pinned by operations on behalf of the netfs are cachefiles_sync_cache()
|
/linux-4.1.27/fs/exportfs/ |
H A D | expfs.c | 400 * it's still pinned by and still positive. exportfs_encode_fh() 520 * At this point we are done with the parent, but it's pinned exportfs_decode_fh()
|
/linux-4.1.27/fs/jfs/ |
H A D | jfs_dtree.c | 577 * return: 0 - search result on stack, leaf page pinned; 839 * dtSearch() returns (leaf page pinned, index at which to insert). dtInsert() 1070 * return <rp> pinned and its extent descriptor <rpxd> dtSplitUp() 1122 * There are a maximum of 4 pages pinned at any time: dtSplitUp() 1124 * keep the child pages pinned while working on the parent. dtSplitUp() 1130 /* keep current child pages (<lp>, <rp>) pinned */ dtSplitUp() 1245 * return <rp> pinned and its block number <rbn>. dtSplitUp() 1255 /* smp and rmp are pinned */ dtSplitUp() 1328 * return split and new page pinned; 1640 * return extended page pinned; 1873 * return new page pinned; 2104 * dtSearch() returns (leaf page pinned, index at which to delete). dtDelete() 2488 /* release the pinned parent page */ dtRelocate() 2637 * NOTE: Search result on stack, the found page is pinned at exit. 3435 * get leftmost leaf page pinned dtReadNext() 3572 /* return target leaf page pinned */ dtReadNext() 4540 * dtSearch() returns (leaf page pinned, index at which to modify). dtModify()
|
H A D | jfs_xtree.c | 237 * the page containing the entry is pinned at exit. 561 * pinned, index at which to insert). xtInsert() 791 * return <rp> pinned and its block number <rpbn>. xtSplitUp() 817 * There are a maximum of 3 pages pinned at any time: xtSplitUp() 819 * to keep the child page pinned while working on the parent. xtSplitUp() 825 /* keep current child pages <rcp> pinned */ xtSplitUp() 867 * return <rp> pinned and its block number <rpbn>. xtSplitUp() 878 /* keep new child page <rp> pinned */ xtSplitUp() 2200 * pinned, index at which to insert). xtAppend() 2889 * the page containing the entry is pinned at exit.
|
H A D | namei.c | 93 * (dtSearch() returns parent directory page pinned) jfs_create() 225 * (dtSearch() returns parent directory page pinned) jfs_mkdir() 901 * (dtSearch() returns parent directory page pinned) jfs_symlink()
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | spu_csa.h | 239 * pinned or mapped to/from application memory, as
|
/linux-4.1.27/arch/powerpc/mm/ |
H A D | 44x_mmu.c | 226 /* 44x has a 256M TLB entry pinned at boot */ setup_initial_memory_limit()
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | perf_cpum_sf.c | 706 /* Event is pinned to a particular CPU, retrieve the per-CPU __hw_perf_event_init() 779 * from the event. If the event is not pinned to a particular __hw_perf_event_init() 784 /* Event is pinned to a particular CPU */ __hw_perf_event_init() 787 /* Event is not pinned, allocate sampling buffer on __hw_perf_event_init() 830 /* Check online status of the CPU to which the event is pinned */ cpumsf_pmu_event_init()
|
/linux-4.1.27/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_execbuf.c | 742 * vmw_query_bo_switch_prepare - Prepare to switch pinned buffer for queries. 749 * query results, and if another buffer currently is pinned for query 751 * switching pinned buffers after successful submission of the current 795 * vmw_query_bo_switch_commit - Finalize switching pinned query buffer 2235 * Don't validate pinned buffers. vmw_validate_single_buffer() 2640 * __vmw_execbuf_release_pinned_bo - Flush queries and unpin the pinned 2648 * This function should be used to unpin the pinned query bo, or 2731 * vmw_execbuf_release_pinned_bo - Flush queries and unpin the pinned 2736 * This function should be used to unpin the pinned query bo, or
|
H A D | vmwgfx_buffer.c | 498 * Note that the buffer object must be either pinned or reserved before 536 * either be reserved or pinned.
|
H A D | vmwgfx_dmabuf.c | 277 * @bo: Pointer to a struct ttm_buffer_object. Must be pinned or reserved.
|
H A D | vmwgfx_fifo.c | 638 * must also be either reserved or pinned when this function is called.
|
H A D | vmwgfx_scrn.c | 144 /* Ok to assume that buffer is pinned in vram */ vmw_sou_fifo_create()
|
H A D | vmwgfx_drv.c | 306 * Create the bo as pinned, so that a tryreserve will vmw_dummy_query_bo_create() 380 * the pinned bo. vmw_release_device()
|
H A D | vmwgfx_resource.c | 368 * vmw_dmabuf_acc_size - Calculate the pinned memory usage of buffers
|
H A D | svga_reg.h | 1521 * are assumed to be pinned and valid for DMA operations.
|
/linux-4.1.27/drivers/media/v4l2-core/ |
H A D | videobuf2-dma-contig.c | 495 /* negative error means that no page was pinned */ vb2_dc_get_user_pages() 744 pr_err("dmabuf buffer is already pinned\n"); vb2_dc_map_dmabuf()
|
H A D | videobuf2-dma-sg.c | 621 pr_err("dmabuf buffer is already pinned\n"); vb2_dma_sg_map_dmabuf()
|
/linux-4.1.27/fs/gfs2/ |
H A D | incore.h | 136 BUFFER_FNS(Pinned, pinned) 137 TAS_BUFFER_FNS(Pinned, pinned)
|
H A D | lops.c | 37 * @bh: The buffer to be pinned 169 * that is pinned in the pagecache.
|
H A D | meta_io.c | 298 * gfs2_meta_wipe - make inode's buffers so they aren't dirty/pinned anymore
|
H A D | glops.c | 49 * None of the buffers should be dirty, locked, or pinned.
|
H A D | log.c | 845 * We wake up gfs2_logd if the number of pinned blocks exceed thresh1 846 * or the total number of used blocks (pinned blocks plus AIL blocks)
|
/linux-4.1.27/net/9p/ |
H A D | trans_virtio.c | 330 * We allow only p9_max_pages pinned. We wait for the p9_get_mapped_pages() 499 * Non kernel buffers are pinned, unpin them p9_virtio_zc_request()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_hmc.c | 126 * 1. The memory for pd should be pinned down, physically contiguous and
|
/linux-4.1.27/arch/arm/mm/ |
H A D | flush.c | 333 * assume that the page is pinned via kmap.
|
/linux-4.1.27/fs/fscache/ |
H A D | cache.c | 383 /* make sure all pages pinned by operations on behalf of the netfs are fscache_withdraw_cache()
|
/linux-4.1.27/include/drm/ttm/ |
H A D | ttm_bo_api.h | 168 * pinned in physical memory. If this behaviour is not desired, this member 458 * pinned in physical memory. If this behaviour is not desired, this member 505 * pinned in physical memory. If this behaviour is not desired, this member
|
/linux-4.1.27/drivers/gpu/drm/nouveau/ |
H A D | nouveau_bo.c | 339 NV_ERROR(drm, "bo %p pinned elsewhere: " nouveau_bo_pin() 360 * move a pinned buffer nouveau_bo_pin() 1299 NV_WARN(drm, "Moving pinned object %p!\n", nvbo); nouveau_bo_move()
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | head_44x.S | 841 * permanently pinned (necessary since Book E 880 * Configure and load pinned entry into TLB slot 63. 1108 * Configure and load pinned entry into TLB for the kernel core
|
H A D | entry_32.S | 982 * assume the instructions here are mapped by a pinned TLB entry */ \
|
/linux-4.1.27/arch/x86/kvm/ |
H A D | pmu.c | 173 .pinned = true, reprogram_counter()
|
H A D | vmx.c | 411 * we must keep them pinned while L2 runs. 7863 * APIC access page of its own; that page is still pinned. vmx_set_apic_access_page_addr() 8672 * address for vmcs02. Keep the page pinned, so this nested_get_vmcs12_pages()
|
/linux-4.1.27/drivers/gpu/host1x/ |
H A D | cdma.c | 459 * from the pushbuffer. The handles for a submit must all be pinned at the same
|
/linux-4.1.27/fs/ocfs2/ |
H A D | dcache.c | 272 * since we have it pinned, so our reference is safe. ocfs2_dentry_attach_lock()
|
H A D | stackglue.c | 47 * the module is pinned, and the locking protocol cannot be changed.
|
/linux-4.1.27/arch/s390/kvm/ |
H A D | gaccess.h | 73 * At a later point guest vcpu lowcore access should happen via pinned
|
/linux-4.1.27/include/uapi/drm/ |
H A D | i915_drm.h | 736 /** Inform the kernel that the batch is and will always be pinned. This 772 /** Handle of the buffer to be pinned. */
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
H A D | ipath_user_sdma.c | 274 /* return number of pages pinned... */ ipath_user_sdma_pin_pages()
|
/linux-4.1.27/arch/microblaze/include/asm/ |
H A D | pgtable.h | 90 /* Make sure to map the vmalloc area above the pinned kernel memory area pte_mkspecial()
|
/linux-4.1.27/fs/xfs/libxfs/ |
H A D | xfs_bmap_btree.c | 849 * transaction. For modification when the buffers are already pinned in memory,
|
/linux-4.1.27/fs/jbd/ |
H A D | checkpoint.c | 500 /* If the oldest pinned transaction is at the tail of the log cleanup_journal_tail()
|
H A D | transaction.c | 1839 * concerned remains pinned as an anonymous buffer belonging to an older
|
/linux-4.1.27/include/uapi/linux/ |
H A D | perf_event.h | 296 pinned : 1, /* must always be on PMU */ member in struct:perf_event_attr
|
/linux-4.1.27/arch/powerpc/perf/ |
H A D | power8-pmu.c | 389 * EBB events are pinned & exclusive, so this should never actually power8_get_constraint()
|
H A D | core-book3s.c | 523 if (!leader->attr.pinned || !leader->attr.exclusive) ebb_event_check()
|
/linux-4.1.27/fs/ext4/ |
H A D | move_extent.c | 418 /* Buffer was busy because probably is pinned to journal transaction, move_extent_per_page()
|
H A D | mballoc.c | 1050 * would have pinned buddy page to page cache. ext4_mb_init_group() 1645 * take the page reference. We want the page to be pinned ext4_mb_use_best_found() 2791 /* we expect to find existing buddy because it's pinned */ ext4_free_data_callback()
|
/linux-4.1.27/security/keys/ |
H A D | process_keys.c | 71 * pinned by a session, but the user_struct pointing to it install_user_keyrings()
|
H A D | key.c | 904 * - we can drop the locks first as we have the key pinned key_create_or_update()
|
/linux-4.1.27/kernel/time/ |
H A D | hrtimer.c | 198 int pinned) switch_hrtimer_base() 203 int cpu = get_nohz_timer_target(pinned); switch_hrtimer_base() 197 switch_hrtimer_base(struct hrtimer *timer, struct hrtimer_clock_base *base, int pinned) switch_hrtimer_base() argument
|
H A D | timer.c | 773 bool pending_only, int pinned) __mod_timer() 790 cpu = get_nohz_timer_target(pinned); __mod_timer() 772 __mod_timer(struct timer_list *timer, unsigned long expires, bool pending_only, int pinned) __mod_timer() argument
|
/linux-4.1.27/fs/ |
H A D | namei.c | 817 /* Special files should not get pinned to the filesystem. */ safe_hardlink_source() 821 /* Setuid files should not get pinned to the filesystem. */ safe_hardlink_source() 825 /* Executable setgid files should not get pinned to the filesystem. */ safe_hardlink_source() 1185 * because a mount-point is always pinned. __follow_mount_rcu() 2284 * mountpoints are always pinned in the dcache, their ancestors are too. Thus, 2331 * No cached dentry. Mounted dentries are pinned in the cache, mountpoint_last()
|
H A D | pipe.c | 852 /* OK, we have a pipe and it's pinned down */ fifo_open()
|
H A D | super.c | 92 * prune the dcache first as the icache is pinned by it, then super_cache_scan()
|
H A D | direct-io.c | 504 * This also helps to limit the peak amount of pinned userspace memory.
|
H A D | fs-writeback.c | 386 * Wait for writeback on an inode to complete. Caller must have inode pinned.
|
H A D | inode.c | 665 * Any inodes which are pinned purely because of attached pagecache have their
|
/linux-4.1.27/fs/ntfs/ |
H A D | mft.c | 185 * count on the page thus releasing it from the pinned state. 856 * inode is pinned. The locked ntfs inode is then returned in @locked_ni. The 2175 * allocated mft record. *@mrec is then set to the allocated, mapped, pinned, 2223 * - returned the ntfs_inode as well as the allocated mapped, pinned, and 2734 * well as the mapped, pinned, and locked mft record. ntfs_mft_record_alloc()
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/socklnd/ |
H A D | socklnd_cb.c | 460 /* ZC_REQ is going to be pinned to the peer */ ksocknal_check_zc_req() 905 /* the message is going to be pinned to the peer */ ksocknal_launch_packet()
|
/linux-4.1.27/drivers/vhost/ |
H A D | scsi.c | 814 /* No pages were pinned */ vhost_scsi_map_to_sgl() 817 /* Less pages pinned than wanted */ vhost_scsi_map_to_sgl()
|
/linux-4.1.27/fs/ocfs2/cluster/ |
H A D | heartbeat.c | 106 #define O2HB_DEBUG_REGION_PINNED "pinned" 147 * In global heartbeat mode, all regions are pinned if there are one or more
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
H A D | c2_wr.h | 1397 u64 reply_buf; /* pinned host buf for reply */
|
/linux-4.1.27/drivers/misc/sgi-xp/ |
H A D | xpc.h | 35 /* define the process name of HB checker and the CPU it is pinned to */
|
/linux-4.1.27/drivers/staging/android/ |
H A D | ashmem.c | 658 * or partially pinned. We handle those two cases here. ashmem_unpin()
|
/linux-4.1.27/drivers/staging/lustre/lustre/osc/ |
H A D | osc_page.c | 630 * having already been removed from LRU and pinned discard_pagevec()
|
H A D | osc_lock.c | 648 * sleeping for lock mutex, but olck is pinned in memory. osc_dlm_blocking_ast0()
|
/linux-4.1.27/drivers/gpu/drm/ast/ |
H A D | ast_mode.c | 937 DRM_DEBUG_KMS("pinned cursor cache at %llx\n", ast->cursor_cache_gpu_addr); ast_cursor_init()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib_common.h | 388 #define QIB_CMD_CPUS_LIST 36 /* list of cpus allocated, for pinned
|
H A D | qib_user_sdma.c | 661 /* return number of pages pinned... */ qib_user_sdma_pin_pages()
|
/linux-4.1.27/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 22 * userpsace pages pinned into memory. We also assume devices and IOMMU
|
/linux-4.1.27/include/xen/interface/ |
H A D | xen.h | 282 * mfn: Machine frame number to be (un)pinned as a p.t. page.
|
/linux-4.1.27/ipc/ |
H A D | mqueue.c | 267 * this is pinned memory, we need to assume worst case, so mqueue_get_inode()
|
/linux-4.1.27/net/rxrpc/ |
H A D | rxkad.c | 52 * this holds a pinned cipher so that keventd doesn't get called by the cipher
|
/linux-4.1.27/drivers/ide/ |
H A D | hpt366.c | 965 * been already "pinned" (via function 1) in init_setup_hpt374() init_chipset_hpt366()
|
/linux-4.1.27/drivers/staging/lustre/lustre/obdclass/ |
H A D | cl_page.c | 1261 * As page->cp_obj is pinned by a reference from page->cp_req, it is cl_page_completion()
|
/linux-4.1.27/fs/kernfs/ |
H A D | dir.c | 524 * owns the cache, the memcg will get pinned forever. So do not account __kernfs_new_node()
|
/linux-4.1.27/lib/ |
H A D | radix-tree.c | 177 * that the caller has pinned this thread of control to the current CPU.
|
/linux-4.1.27/drivers/pci/ |
H A D | pci.c | 1367 unsigned int pinned:1; member in struct:pci_devres 1391 if (this->enabled && !this->pinned) pcim_release() 1457 dr->pinned = 1; pcim_pin_device()
|
/linux-4.1.27/block/ |
H A D | blk-throttle.c | 50 * tree pinned while bios are in flight.
|
/linux-4.1.27/fs/ubifs/ |
H A D | recovery.c | 1239 * @inode: inode if pinned in memory awaiting rw mode to fix it
|
/linux-4.1.27/fs/nfsd/ |
H A D | vfs.c | 117 * original mnt is pinned down by exp, so we should nfsd_cross_mnt()
|
/linux-4.1.27/arch/powerpc/platforms/cell/spufs/ |
H A D | switch.c | 2196 * and may need to be pinned or otherwise specially aligned.
|
/linux-4.1.27/arch/ia64/kernel/ |
H A D | entry.S | 941 srlz.d // ensure that inter. collection is off (VHPT is don't care, since text is pinned)
|
H A D | perfmon.c | 1023 * system-wide session are pinned and self-monitoring pfm_restore_monitoring() 4247 * be pinned on a CPU throughout the session. pfm_context_load()
|
/linux-4.1.27/fs/reiserfs/ |
H A D | inode.c | 2173 * On success, page_result is set to a locked, pinned page, and bh_result 3138 * very conservative, leave the buffer pinned if invalidatepage_can_drop()
|
/linux-4.1.27/kernel/sched/ |
H A D | core.c | 596 int get_nohz_timer_target(int pinned) get_nohz_timer_target() argument 602 if (pinned || !get_sysctl_timer_migration() || !idle_cpu(cpu)) get_nohz_timer_target() 1945 * and the cgroup is pinned to this child due to cgroup_fork() sched_fork()
|
H A D | fair.c | 6832 * Max backoff if we encounter pinned tasks. Pretty arbitrary value, but 7059 /* All tasks on this runqueue were pinned by CPU affinity */ load_balance() 7154 * We reach balance because all tasks are pinned at this level so load_balance()
|
/linux-4.1.27/drivers/misc/vmw_vmci/ |
H A D | vmci_queue_pair.c | 828 * Pages are pinned.
|
/linux-4.1.27/fs/jbd2/ |
H A D | transaction.c | 1992 * concerned remains pinned as an anonymous buffer belonging to an older
|
/linux-4.1.27/arch/powerpc/platforms/powernv/ |
H A D | pci-ioda.c | 1055 /* The PE number isn't pinned by M64 */ pnv_ioda_setup_bus_PE()
|