entry 13 arch/alpha/include/asm/a.out.h set_personality (((BFPM->taso || EX.ah.entry < 0x100000000L \ entry 32 arch/alpha/include/uapi/asm/a.out.h __u64 entry; entry 69 arch/alpha/include/uapi/asm/a.out.h #define a_entry ah.entry entry 34 arch/alpha/kernel/binfmt_loader.c bprm->taso = eh->ah.entry < 0x100000000UL; entry 134 arch/alpha/kernel/srm_env.c srm_env_t *entry; entry 182 arch/alpha/kernel/srm_env.c entry = srm_named_entries; entry 183 arch/alpha/kernel/srm_env.c while (entry->name && entry->id) { entry 184 arch/alpha/kernel/srm_env.c if (!proc_create_data(entry->name, 0644, named_dir, entry 185 arch/alpha/kernel/srm_env.c &srm_env_proc_fops, (void *)entry->id)) entry 187 arch/alpha/kernel/srm_env.c entry++; entry 148 arch/arc/include/asm/entry-arcv2.h ; 2. Upon entry SP is always saved (for any inspection, unwinding etc), entry 180 arch/arc/include/asm/entry-arcv2.h ; - Z flag set on K is inverse of what hardware does on interrupt entry entry 65 arch/arc/kernel/perf_event.c struct perf_callchain_entry_ctx *entry = ctrl->perf_stuff; entry 67 arch/arc/kernel/perf_event.c perf_callchain_store(entry, addr); entry 75 arch/arc/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 80 arch/arc/kernel/perf_event.c .perf_stuff = entry, entry 86 arch/arc/kernel/perf_event.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry, entry 93 arch/arc/kernel/perf_event.c perf_callchain_store(entry, instruction_pointer(regs)); entry 254 arch/arc/mm/tlb.c unsigned int entry; entry 267 arch/arc/mm/tlb.c for (entry = 0; entry < num_tlb; entry++) { entry 269 arch/arc/mm/tlb.c write_aux_reg(ARC_REG_TLBINDEX, entry); entry 279 arch/arc/mm/tlb.c for (entry = stlb_idx; entry < stlb_idx + 16; entry++) { entry 280 arch/arc/mm/tlb.c write_aux_reg(ARC_REG_TLBINDEX, entry); entry 489 arch/arm/include/asm/assembler.h #define _ASM_NOKPROBE(entry) \ entry 492 arch/arm/include/asm/assembler.h .long entry; \ entry 495 arch/arm/include/asm/assembler.h #define _ASM_NOKPROBE(entry) entry 68 arch/arm/include/asm/kexec.h static inline phys_addr_t boot_phys_to_phys(unsigned long entry) entry 70 arch/arm/include/asm/kexec.h return idmap_to_phys(entry); entry 41 arch/arm/include/asm/tlb.h #define tlb_remove_table(tlb, entry) tlb_remove_page(tlb, entry) entry 7 arch/arm/kernel/jump_label.c static void __arch_jump_label_transform(struct jump_entry *entry, entry 11 arch/arm/kernel/jump_label.c void *addr = (void *)entry->code; entry 15 arch/arm/kernel/jump_label.c insn = arm_gen_branch(entry->code, entry->target); entry 25 arch/arm/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry, entry 28 arch/arm/kernel/jump_label.c __arch_jump_label_transform(entry, type, false); entry 31 arch/arm/kernel/jump_label.c void arch_jump_label_transform_static(struct jump_entry *entry, entry 34 arch/arm/kernel/jump_label.c __arch_jump_label_transform(entry, type, true); entry 35 arch/arm/kernel/perf_callchain.c struct perf_callchain_entry_ctx *entry) entry 50 arch/arm/kernel/perf_callchain.c perf_callchain_store(entry, buftail.lr); entry 63 arch/arm/kernel/perf_callchain.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) entry 72 arch/arm/kernel/perf_callchain.c perf_callchain_store(entry, regs->ARM_pc); entry 79 arch/arm/kernel/perf_callchain.c while ((entry->nr < entry->max_stack) && entry 81 arch/arm/kernel/perf_callchain.c tail = user_backtrace(tail, entry); entry 93 arch/arm/kernel/perf_callchain.c struct perf_callchain_entry_ctx *entry = data; entry 94 arch/arm/kernel/perf_callchain.c perf_callchain_store(entry, fr->pc); entry 99 arch/arm/kernel/perf_callchain.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) entry 109 arch/arm/kernel/perf_callchain.c walk_stackframe(&fr, callchain_trace, entry); entry 67 arch/arm/kernel/vmlinux.lds.h *(.entry.text) \ entry 40 arch/arm/mm/fault-armv.c pte_t entry = *ptep; entry 46 arch/arm/mm/fault-armv.c ret = pte_present(entry); entry 52 arch/arm/mm/fault-armv.c if (ret && (pte_val(entry) & L_PTE_MT_MASK) != shared_pte_mask) { entry 56 arch/arm/mm/fault-armv.c pte_val(entry) &= ~L_PTE_MT_MASK; entry 57 arch/arm/mm/fault-armv.c pte_val(entry) |= shared_pte_mask; entry 58 arch/arm/mm/fault-armv.c set_pte_at(vma->vm_mm, address, ptep, entry); entry 738 arch/arm/probes/kprobes/test-core.c struct coverage_entry *entry = coverage->base + coverage->num_entries; entry 747 arch/arm/probes/kprobes/test-core.c entry->header = h; entry 748 arch/arm/probes/kprobes/test-core.c entry->regs = coverage_start_registers(h); entry 749 arch/arm/probes/kprobes/test-core.c entry->nesting = coverage->nesting; entry 750 arch/arm/probes/kprobes/test-core.c entry->matched = false; entry 775 arch/arm/probes/kprobes/test-core.c coverage_add_registers(struct coverage_entry *entry, kprobe_opcode_t insn) entry 777 arch/arm/probes/kprobes/test-core.c int regs = entry->header->type_regs.bits >> DECODE_TYPE_BITS; entry 793 arch/arm/probes/kprobes/test-core.c entry->regs &= ~(flag << i); entry 827 arch/arm/probes/kprobes/test-core.c entry->regs &= ~(COVERAGE_PCWB << i); entry 844 arch/arm/probes/kprobes/test-core.c struct coverage_entry *entry = coverage.base; entry 849 arch/arm/probes/kprobes/test-core.c for (; entry < end; ++entry) { entry 850 arch/arm/probes/kprobes/test-core.c const struct decode_header *h = entry->header; entry 853 arch/arm/probes/kprobes/test-core.c if (entry->nesting > nesting) entry 856 arch/arm/probes/kprobes/test-core.c if (entry->nesting < nesting) entry 862 arch/arm/probes/kprobes/test-core.c entry->matched = true; entry 874 arch/arm/probes/kprobes/test-core.c coverage_add_registers(entry, insn); entry 891 arch/arm/probes/kprobes/test-core.c struct coverage_entry *entry = coverage.base; entry 894 arch/arm/probes/kprobes/test-core.c for (; entry < end; ++entry) { entry 895 arch/arm/probes/kprobes/test-core.c u32 mask = entry->header->mask.bits; entry 896 arch/arm/probes/kprobes/test-core.c u32 value = entry->header->value.bits; entry 898 arch/arm/probes/kprobes/test-core.c if (entry->regs) { entry 900 arch/arm/probes/kprobes/test-core.c mask, value, entry->regs); entry 903 arch/arm/probes/kprobes/test-core.c if (!entry->matched) { entry 36 arch/arm/xen/p2m.c struct xen_p2m_entry *entry; entry 41 arch/arm/xen/p2m.c entry = rb_entry(parent, struct xen_p2m_entry, rbnode_phys); entry 43 arch/arm/xen/p2m.c if (new->pfn == entry->pfn) entry 46 arch/arm/xen/p2m.c if (new->pfn < entry->pfn) entry 58 arch/arm/xen/p2m.c __func__, &new->pfn, &new->mfn, &entry->pfn, &entry->mfn); entry 66 arch/arm/xen/p2m.c struct xen_p2m_entry *entry; entry 71 arch/arm/xen/p2m.c entry = rb_entry(n, struct xen_p2m_entry, rbnode_phys); entry 72 arch/arm/xen/p2m.c if (entry->pfn <= pfn && entry 73 arch/arm/xen/p2m.c entry->pfn + entry->nr_pages > pfn) { entry 74 arch/arm/xen/p2m.c unsigned long mfn = entry->mfn + (pfn - entry->pfn); entry 78 arch/arm/xen/p2m.c if (pfn < entry->pfn) entry 38 arch/arm64/include/asm/acpi.h #define BAD_MADT_GICC_ENTRY(entry, end) \ entry 39 arch/arm64/include/asm/acpi.h (!(entry) || (entry)->header.length < ACPI_MADT_GICC_MIN_LENGTH || \ entry 40 arch/arm64/include/asm/acpi.h (unsigned long)(entry) + (entry)->header.length > (end)) entry 361 arch/arm64/include/asm/cpufeature.h cpucap_multi_entry_cap_matches(const struct arm64_cpu_capabilities *entry, entry 366 arch/arm64/include/asm/cpufeature.h for (caps = entry->match_list; caps->matches; caps++) entry 37 arch/arm64/include/asm/hugetlb.h extern pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma, entry 677 arch/arm64/include/asm/pgtable.h pte_t entry, int dirty); entry 683 arch/arm64/include/asm/pgtable.h pmd_t entry, int dirty) entry 685 arch/arm64/include/asm/pgtable.h return ptep_set_access_flags(vma, address, (pte_t *)pmdp, pmd_pte(entry), dirty); entry 13 arch/arm64/kernel/cpu-reset.h void __cpu_soft_restart(unsigned long el2_switch, unsigned long entry, entry 16 arch/arm64/kernel/cpu-reset.h static inline void __noreturn cpu_soft_restart(unsigned long entry, entry 28 arch/arm64/kernel/cpu-reset.h restart(el2_switch, entry, arg0, arg1, arg2); entry 18 arch/arm64/kernel/cpu_errata.c is_affected_midr_range(const struct arm64_cpu_capabilities *entry, int scope) entry 24 arch/arm64/kernel/cpu_errata.c if (!is_midr_in_range(midr, &entry->midr_range)) entry 29 arch/arm64/kernel/cpu_errata.c for (fix = entry->fixed_revs; fix && fix->revidr_mask; fix++) entry 37 arch/arm64/kernel/cpu_errata.c is_affected_midr_range_list(const struct arm64_cpu_capabilities *entry, entry 41 arch/arm64/kernel/cpu_errata.c return is_midr_in_range_list(read_cpuid_id(), entry->midr_range_list); entry 45 arch/arm64/kernel/cpu_errata.c is_kryo_midr(const struct arm64_cpu_capabilities *entry, int scope) entry 55 arch/arm64/kernel/cpu_errata.c return model == entry->midr_range.model; entry 59 arch/arm64/kernel/cpu_errata.c has_mismatched_cache_type(const struct arm64_cpu_capabilities *entry, entry 378 arch/arm64/kernel/cpu_errata.c static bool has_ssbd_mitigation(const struct arm64_cpu_capabilities *entry, entry 392 arch/arm64/kernel/cpu_errata.c if (is_midr_in_range_list(read_cpuid_id(), entry->midr_range_list)) entry 438 arch/arm64/kernel/cpu_errata.c pr_info_once("%s mitigation not required\n", entry->desc); entry 484 arch/arm64/kernel/cpu_errata.c pr_info_once("%s disabled from command-line\n", entry->desc); entry 488 arch/arm64/kernel/cpu_errata.c pr_info_once("%s forced from command-line\n", entry->desc); entry 508 arch/arm64/kernel/cpu_errata.c has_cortex_a76_erratum_1463225(const struct arm64_cpu_capabilities *entry, entry 595 arch/arm64/kernel/cpu_errata.c check_branch_predictor(const struct arm64_cpu_capabilities *entry, int scope) entry 645 arch/arm64/kernel/cpu_errata.c needs_tx2_tvm_workaround(const struct arm64_cpu_capabilities *entry, entry 650 arch/arm64/kernel/cpu_errata.c if (!is_affected_midr_range_list(entry, scope) || entry 663 arch/arm64/kernel/cpu_errata.c has_neoverse_n1_erratum_1542419(const struct arm64_cpu_capabilities *entry, entry 111 arch/arm64/kernel/cpufeature.c cpufeature_pan_not_uao(const struct arm64_cpu_capabilities *entry, int __unused); entry 860 arch/arm64/kernel/cpufeature.c feature_matches(u64 reg, const struct arm64_cpu_capabilities *entry) entry 862 arch/arm64/kernel/cpufeature.c int val = cpuid_feature_extract_field(reg, entry->field_pos, entry->sign); entry 864 arch/arm64/kernel/cpufeature.c return val >= entry->min_field_value; entry 868 arch/arm64/kernel/cpufeature.c has_cpuid_feature(const struct arm64_cpu_capabilities *entry, int scope) entry 874 arch/arm64/kernel/cpufeature.c val = read_sanitised_ftr_reg(entry->sys_reg); entry 876 arch/arm64/kernel/cpufeature.c val = __read_sysreg_by_encoding(entry->sys_reg); entry 878 arch/arm64/kernel/cpufeature.c return feature_matches(val, entry); entry 881 arch/arm64/kernel/cpufeature.c static bool has_useable_gicv3_cpuif(const struct arm64_cpu_capabilities *entry, int scope) entry 885 arch/arm64/kernel/cpufeature.c if (!has_cpuid_feature(entry, scope)) entry 891 arch/arm64/kernel/cpufeature.c entry->desc); entry 896 arch/arm64/kernel/cpufeature.c static bool has_no_hw_prefetch(const struct arm64_cpu_capabilities *entry, int __unused) entry 906 arch/arm64/kernel/cpufeature.c static bool has_no_fpsimd(const struct arm64_cpu_capabilities *entry, int __unused) entry 914 arch/arm64/kernel/cpufeature.c static bool has_cache_idc(const struct arm64_cpu_capabilities *entry, entry 939 arch/arm64/kernel/cpufeature.c static bool has_cache_dic(const struct arm64_cpu_capabilities *entry, entry 953 arch/arm64/kernel/cpufeature.c has_useable_cnp(const struct arm64_cpu_capabilities *entry, int scope) entry 963 arch/arm64/kernel/cpufeature.c return has_cpuid_feature(entry, scope); entry 969 arch/arm64/kernel/cpufeature.c static bool unmap_kernel_at_el0(const struct arm64_cpu_capabilities *entry, entry 991 arch/arm64/kernel/cpufeature.c if (has_cpuid_feature(entry, scope)) entry 1151 arch/arm64/kernel/cpufeature.c static bool runs_at_el2(const struct arm64_cpu_capabilities *entry, int __unused) entry 1260 arch/arm64/kernel/cpufeature.c static bool can_use_gic_priorities(const struct arm64_cpu_capabilities *entry, entry 1263 arch/arm64/kernel/cpufeature.c return enable_pseudo_nmi && has_useable_gicv3_cpuif(entry, scope); entry 2112 arch/arm64/kernel/cpufeature.c cpufeature_pan_not_uao(const struct arm64_cpu_capabilities *entry, int __unused) entry 12 arch/arm64/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry, entry 15 arch/arm64/kernel/jump_label.c void *addr = (void *)jump_entry_code(entry); entry 19 arch/arm64/kernel/jump_label.c insn = aarch64_insn_gen_branch_imm(jump_entry_code(entry), entry 20 arch/arm64/kernel/jump_label.c jump_entry_target(entry), entry 29 arch/arm64/kernel/jump_label.c void arch_jump_label_transform_static(struct jump_entry *entry, entry 85 arch/arm64/kernel/machine_kexec.c kimage_entry_t *entry; entry 87 arch/arm64/kernel/machine_kexec.c for (entry = &kimage->head; ; entry++) { entry 92 arch/arm64/kernel/machine_kexec.c __flush_dcache_area(entry, sizeof(kimage_entry_t)); entry 94 arch/arm64/kernel/machine_kexec.c flag = *entry & IND_FLAGS; entry 98 arch/arm64/kernel/machine_kexec.c addr = phys_to_virt(*entry & PAGE_MASK); entry 103 arch/arm64/kernel/machine_kexec.c entry = (kimage_entry_t *)addr - 1; entry 99 arch/arm64/kernel/pci.c struct resource_entry *entry, *tmp; entry 103 arch/arm64/kernel/pci.c resource_list_for_each_entry_safe(entry, tmp, &ci->resources) { entry 104 arch/arm64/kernel/pci.c if (!(entry->res->flags & IORESOURCE_WINDOW)) entry 105 arch/arm64/kernel/pci.c resource_list_destroy_entry(entry); entry 24 arch/arm64/kernel/perf_callchain.c struct perf_callchain_entry_ctx *entry) entry 43 arch/arm64/kernel/perf_callchain.c perf_callchain_store(entry, lr); entry 72 arch/arm64/kernel/perf_callchain.c struct perf_callchain_entry_ctx *entry) entry 88 arch/arm64/kernel/perf_callchain.c perf_callchain_store(entry, buftail.lr); entry 102 arch/arm64/kernel/perf_callchain.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry, entry 110 arch/arm64/kernel/perf_callchain.c perf_callchain_store(entry, regs->pc); entry 118 arch/arm64/kernel/perf_callchain.c while (entry->nr < entry->max_stack && entry 120 arch/arm64/kernel/perf_callchain.c tail = user_backtrace(tail, entry); entry 128 arch/arm64/kernel/perf_callchain.c while ((entry->nr < entry->max_stack) && entry 130 arch/arm64/kernel/perf_callchain.c tail = compat_user_backtrace(tail, entry); entry 142 arch/arm64/kernel/perf_callchain.c struct perf_callchain_entry_ctx *entry = data; entry 143 arch/arm64/kernel/perf_callchain.c perf_callchain_store(entry, frame->pc); entry 147 arch/arm64/kernel/perf_callchain.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 158 arch/arm64/kernel/perf_callchain.c walk_stackframe(current, &frame, callchain_trace, entry); entry 199 arch/arm64/mm/fault.c pte_t entry, int dirty) entry 204 arch/arm64/mm/fault.c if (pte_same(pte, entry)) entry 208 arch/arm64/mm/fault.c pte_val(entry) &= PTE_RDONLY | PTE_AF | PTE_WRITE | PTE_DIRTY; entry 216 arch/arm64/mm/fault.c pte_val(entry) ^= PTE_RDONLY; entry 221 arch/arm64/mm/fault.c pteval |= pte_val(entry); entry 297 arch/arm64/mm/hugetlbpage.c pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma, entry 303 arch/arm64/mm/hugetlbpage.c entry = pte_mkcont(entry); entry 305 arch/arm64/mm/hugetlbpage.c entry = pmd_pte(pmd_mkcont(pte_pmd(entry))); entry 310 arch/arm64/mm/hugetlbpage.c return entry; entry 36 arch/csky/kernel/perf_callchain.c struct perf_callchain_entry_ctx *entry) entry 39 arch/csky/kernel/perf_callchain.c perf_callchain_store(entry, fr->lr); entry 47 arch/csky/kernel/perf_callchain.c static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry, entry 67 arch/csky/kernel/perf_callchain.c perf_callchain_store(entry, lr); entry 86 arch/csky/kernel/perf_callchain.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry, entry 96 arch/csky/kernel/perf_callchain.c perf_callchain_store(entry, regs->pc); entry 104 arch/csky/kernel/perf_callchain.c fp = user_backtrace(entry, fp, regs->lr); entry 106 arch/csky/kernel/perf_callchain.c while (fp && !(fp & 0x3) && entry->nr < entry->max_stack) entry 107 arch/csky/kernel/perf_callchain.c fp = user_backtrace(entry, fp, 0); entry 110 arch/csky/kernel/perf_callchain.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 123 arch/csky/kernel/perf_callchain.c walk_stackframe(&fr, entry); entry 64 arch/ia64/hp/common/aml_nfw.c ia64_sal_handler entry; entry 69 arch/ia64/hp/common/aml_nfw.c entry = (ia64_sal_handler) &virt_entry; entry 71 arch/ia64/hp/common/aml_nfw.c IA64_FW_CALL(entry, c->ret, entry 476 arch/ia64/include/asm/pgtable.h #define __swp_type(entry) (((entry).val >> 1) & 0x7f) entry 477 arch/ia64/include/asm/pgtable.h #define __swp_offset(entry) (((entry).val << 1) >> 9) entry 50 arch/ia64/include/asm/sal.h #define __IA64_FW_CALL(entry,result,a0,a1,a2,a3,a4,a5,a6,a7) \ entry 51 arch/ia64/include/asm/sal.h result = (*entry)(a0,a1,a2,a3,a4,a5,a6,a7) entry 53 arch/ia64/include/asm/sal.h # define IA64_FW_CALL(entry,result,args...) do { \ entry 58 arch/ia64/include/asm/sal.h __IA64_FW_CALL(entry, result, args); \ entry 507 arch/ia64/kernel/acpi.c slit_table->entry[i * slit_table->locality_count + j]; entry 982 arch/ia64/kernel/kprobes.c unsigned long arch_deref_entry_point(void *entry) entry 984 arch/ia64/kernel/kprobes.c return ((struct fnptr *)entry)->ip; entry 51 arch/ia64/kernel/palinfo.c struct proc_dir_entry *entry; /* registered entry (removal) */ entry 1205 arch/ia64/kernel/perfmon.c pfm_buffer_fmt_t * entry; entry 1208 arch/ia64/kernel/perfmon.c entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list); entry 1209 arch/ia64/kernel/perfmon.c if (pfm_uuid_cmp(uuid, entry->fmt_uuid) == 0) entry 1210 arch/ia64/kernel/perfmon.c return entry; entry 5544 arch/ia64/kernel/perfmon.c pfm_buffer_fmt_t * entry; entry 5578 arch/ia64/kernel/perfmon.c entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list); entry 5580 arch/ia64/kernel/perfmon.c entry->fmt_uuid, entry->fmt_name); entry 61 arch/ia64/kernel/salinfo.c struct proc_dir_entry *entry; /* registered entry (removal) */ entry 596 arch/ia64/kernel/salinfo.c struct proc_dir_entry *dir, *entry; entry 619 arch/ia64/kernel/salinfo.c entry = proc_create_data("event", S_IRUSR, dir, entry 621 arch/ia64/kernel/salinfo.c if (!entry) entry 623 arch/ia64/kernel/salinfo.c *sdir++ = entry; entry 625 arch/ia64/kernel/salinfo.c entry = proc_create_data("data", S_IRUSR | S_IWUSR, dir, entry 627 arch/ia64/kernel/salinfo.c if (!entry) entry 629 arch/ia64/kernel/salinfo.c *sdir++ = entry; entry 2189 arch/ia64/kernel/unwind.c const struct unw_table_entry *entry, *start, *end; entry 2213 arch/ia64/kernel/unwind.c for (entry = start; entry < end; ++entry) entry 2214 arch/ia64/kernel/unwind.c size += 3*8 + 8 + 8*UNW_LENGTH(*(u64 *) (segbase + entry->info_offset)); entry 2228 arch/ia64/kernel/unwind.c for (entry = start; entry < end; ++entry, lp += 3) { entry 2229 arch/ia64/kernel/unwind.c info_size = 8 + 8*UNW_LENGTH(*(u64 *) (segbase + entry->info_offset)); entry 2231 arch/ia64/kernel/unwind.c memcpy(info, (char *) segbase + entry->info_offset, info_size); entry 2233 arch/ia64/kernel/unwind.c lp[0] = segbase + entry->start_offset; /* start */ entry 2234 arch/ia64/kernel/unwind.c lp[1] = segbase + entry->end_offset; /* end */ entry 69 arch/ia64/mm/hugetlbpage.c #define mk_pte_huge(entry) { pte_val(entry) |= _PAGE_P; } entry 152 arch/ia64/pci/pci.c struct resource_entry *entry) entry 155 arch/ia64/pci/pci.c struct resource *resource, *res = entry->res; entry 170 arch/ia64/pci/pci.c space_nr = new_space(entry->offset, sparse); entry 175 arch/ia64/pci/pci.c min = res->start - entry->offset; entry 176 arch/ia64/pci/pci.c max = res->end - entry->offset; entry 202 arch/ia64/pci/pci.c entry->offset = base_port; entry 239 arch/ia64/pci/pci.c struct resource_entry *entry, *tmp; entry 245 arch/ia64/pci/pci.c resource_list_for_each_entry_safe(entry, tmp, &ci->resources) { entry 246 arch/ia64/pci/pci.c res = entry->res; entry 253 arch/ia64/pci/pci.c resource_list_del(entry); entry 255 arch/ia64/pci/pci.c entry->res); entry 256 arch/ia64/pci/pci.c resource_list_add_tail(entry, entry 260 arch/ia64/pci/pci.c if (resource_is_pcicfg_ioport(entry->res)) entry 261 arch/ia64/pci/pci.c resource_list_destroy_entry(entry); entry 262 arch/ia64/pci/pci.c else if (add_io_space(dev, info, entry)) entry 263 arch/ia64/pci/pci.c resource_list_destroy_entry(entry); entry 274 arch/ia64/pci/pci.c struct resource_entry *entry, *tmp; entry 277 arch/ia64/pci/pci.c resource_list_for_each_entry_safe(entry, tmp, &info->io_resources) { entry 278 arch/ia64/pci/pci.c release_resource(entry->res); entry 279 arch/ia64/pci/pci.c resource_list_destroy_entry(entry); entry 88 arch/m68k/include/asm/sun3mmu.h register unsigned long entry; entry 95 arch/m68k/include/asm/sun3mmu.h entry = c; entry 97 arch/m68k/include/asm/sun3mmu.h return entry; entry 101 arch/m68k/include/asm/sun3mmu.h static inline void sun3_put_segmap(unsigned long addr, unsigned long entry) entry 107 arch/m68k/include/asm/sun3mmu.h SET_CONTROL_BYTE (AC_SEGMAP | (addr & SUN3_CONTROL_MASK), entry); entry 116 arch/m68k/include/asm/sun3mmu.h register unsigned long entry; entry 121 arch/m68k/include/asm/sun3mmu.h GET_CONTROL_WORD (AC_PAGEMAP | (addr & SUN3_CONTROL_MASK), entry); entry 124 arch/m68k/include/asm/sun3mmu.h return entry; entry 128 arch/m68k/include/asm/sun3mmu.h static inline void sun3_put_pte(unsigned long addr, unsigned long entry) entry 134 arch/m68k/include/asm/sun3mmu.h SET_CONTROL_WORD (AC_PAGEMAP | (addr & SUN3_CONTROL_MASK), entry); entry 506 arch/microblaze/include/asm/pgtable.h #define __swp_type(entry) ((entry).val & 0x3f) entry 507 arch/microblaze/include/asm/pgtable.h #define __swp_offset(entry) ((entry).val >> 6) entry 43 arch/mips/boot/ecoff.h int32_t entry; /* entry pt. */ entry 239 arch/mips/boot/elf2ecoff.c a->entry = swab32(a->entry); entry 425 arch/mips/boot/elf2ecoff.c eah.entry = ex.e_entry; entry 196 arch/mips/cavium-octeon/smp.c unsigned long entry = (unsigned long)kernel_entry; entry 199 arch/mips/cavium-octeon/smp.c octeon_processor_relocated_kernel_entry = entry + offset; entry 135 arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h # Jump to the normal Linux SMP entry point entry 2552 arch/mips/include/asm/octeon/cvmx-pip-defs.h uint64_t entry:62; entry 2554 arch/mips/include/asm/octeon/cvmx-pip-defs.h uint64_t entry:62; entry 790 arch/mips/include/asm/sn/sn0/hubio.h u64 entry; entry 211 arch/mips/kernel/machine_kexec.c unsigned long entry; entry 236 arch/mips/kernel/machine_kexec.c for (ptr = &image->head; (entry = *ptr) && !(entry &IND_DONE); entry 237 arch/mips/kernel/machine_kexec.c ptr = (entry & IND_INDIRECTION) ? entry 238 arch/mips/kernel/machine_kexec.c phys_to_virt(entry & PAGE_MASK) : ptr + 1) { entry 26 arch/mips/kernel/perf_event.c static void save_raw_perf_callchain(struct perf_callchain_entry_ctx *entry, entry 35 arch/mips/kernel/perf_event.c perf_callchain_store(entry, addr); entry 36 arch/mips/kernel/perf_event.c if (entry->nr >= entry->max_stack) entry 42 arch/mips/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 55 arch/mips/kernel/perf_event.c save_raw_perf_callchain(entry, sp); entry 59 arch/mips/kernel/perf_event.c perf_callchain_store(entry, pc); entry 60 arch/mips/kernel/perf_event.c if (entry->nr >= entry->max_stack) entry 65 arch/mips/kernel/perf_event.c save_raw_perf_callchain(entry, sp); entry 118 arch/mips/kernel/pm-cps.c cps_nc_entry_fn entry; entry 123 arch/mips/kernel/pm-cps.c entry = per_cpu(nc_asm_enter, core)[state]; entry 124 arch/mips/kernel/pm-cps.c if (!entry) entry 170 arch/mips/kernel/pm-cps.c left = entry(online, nc_core_ready_count); entry 78 arch/mips/kvm/mmu.c unsigned long entry; entry 81 arch/mips/kvm/mmu.c entry = (unsigned long)invalid_pte_table; entry 83 arch/mips/kvm/mmu.c entry = (unsigned long)invalid_pmd_table; entry 90 arch/mips/kvm/mmu.c p[0] = entry; entry 91 arch/mips/kvm/mmu.c p[1] = entry; entry 92 arch/mips/kvm/mmu.c p[2] = entry; entry 93 arch/mips/kvm/mmu.c p[3] = entry; entry 94 arch/mips/kvm/mmu.c p[4] = entry; entry 96 arch/mips/kvm/mmu.c p[-3] = entry; entry 97 arch/mips/kvm/mmu.c p[-2] = entry; entry 98 arch/mips/kvm/mmu.c p[-1] = entry; entry 695 arch/mips/kvm/mmu.c pte_t *ptep, entry, old_pte; entry 765 arch/mips/kvm/mmu.c entry = pfn_pte(pfn, __pgprot(prot_bits)); entry 769 arch/mips/kvm/mmu.c set_pte(ptep, entry); entry 395 arch/mips/kvm/tlb.c int entry; entry 412 arch/mips/kvm/tlb.c for (entry = 0; entry < current_cpu_data.tlbsize; entry++) { entry 413 arch/mips/kvm/tlb.c write_c0_index(entry); entry 423 arch/mips/kvm/tlb.c write_c0_entryhi(UNIQUE_ENTRYHI(entry)); entry 453 arch/mips/kvm/tlb.c int entry; entry 478 arch/mips/kvm/tlb.c for (entry = 0; entry < current_cpu_data.guest.tlbsize; entry++) { entry 480 arch/mips/kvm/tlb.c write_gc0_index(entry); entry 481 arch/mips/kvm/tlb.c write_gc0_entryhi(UNIQUE_GUEST_ENTRYHI(entry)); entry 20 arch/mips/mm/pgtable-64.c unsigned long entry; entry 23 arch/mips/mm/pgtable-64.c entry = (unsigned long)invalid_pud_table; entry 25 arch/mips/mm/pgtable-64.c entry = (unsigned long)invalid_pmd_table; entry 27 arch/mips/mm/pgtable-64.c entry = (unsigned long)invalid_pte_table; entry 34 arch/mips/mm/pgtable-64.c p[0] = entry; entry 35 arch/mips/mm/pgtable-64.c p[1] = entry; entry 36 arch/mips/mm/pgtable-64.c p[2] = entry; entry 37 arch/mips/mm/pgtable-64.c p[3] = entry; entry 38 arch/mips/mm/pgtable-64.c p[4] = entry; entry 40 arch/mips/mm/pgtable-64.c p[-3] = entry; entry 41 arch/mips/mm/pgtable-64.c p[-2] = entry; entry 42 arch/mips/mm/pgtable-64.c p[-1] = entry; entry 43 arch/mips/mm/tlb-r3k.c static void local_flush_tlb_from(int entry) entry 49 arch/mips/mm/tlb-r3k.c while (entry < current_cpu_data.tlbsize) { entry 50 arch/mips/mm/tlb-r3k.c write_c0_index(entry << 8); entry 51 arch/mips/mm/tlb-r3k.c write_c0_entryhi((entry | 0x80000) << 12); entry 52 arch/mips/mm/tlb-r3k.c entry++; /* BARRIER */ entry 59 arch/mips/mm/tlb-r4k.c int entry, ftlbhighset; entry 68 arch/mips/mm/tlb-r4k.c entry = num_wired_entries(); entry 74 arch/mips/mm/tlb-r4k.c if (cpu_has_tlbinv && !entry) { entry 82 arch/mips/mm/tlb-r4k.c for (entry = current_cpu_data.tlbsizevtlb; entry 83 arch/mips/mm/tlb-r4k.c entry < ftlbhighset; entry 84 arch/mips/mm/tlb-r4k.c entry++) { entry 85 arch/mips/mm/tlb-r4k.c write_c0_index(entry); entry 90 arch/mips/mm/tlb-r4k.c while (entry < current_cpu_data.tlbsize) { entry 92 arch/mips/mm/tlb-r4k.c write_c0_entryhi(UNIQUE_ENTRYHI(entry)); entry 93 arch/mips/mm/tlb-r4k.c write_c0_index(entry); entry 96 arch/mips/mm/tlb-r4k.c entry++; entry 2568 arch/mips/mm/tlbex.c unsigned long entry; entry 2582 arch/mips/mm/tlbex.c entry = read_c0_entrylo0(); entry 2585 arch/mips/mm/tlbex.c entry &= ~((1 << MIPS_ENTRYLO_PFN_SHIFT) - 1); entry 2586 arch/mips/mm/tlbex.c entry &= ~(MIPS_ENTRYLO_RI | MIPS_ENTRYLO_XI); entry 2589 arch/mips/mm/tlbex.c pabits = fls_long(entry) + 6; entry 187 arch/mips/pci/msi-octeon.c struct msi_desc *entry; entry 203 arch/mips/pci/msi-octeon.c for_each_pci_msi_entry(entry, dev) { entry 204 arch/mips/pci/msi-octeon.c ret = arch_setup_msi_irq(dev, entry); entry 72 arch/mips/pci/pci-alchemy.c static void mod_wired_entry(int entry, unsigned long entrylo0, entry 82 arch/mips/pci/pci-alchemy.c write_c0_index(entry); entry 34 arch/mips/tools/elf-entry.c uint64_t entry; entry 64 arch/mips/tools/elf-entry.c entry = le32toh(hdr.ehdr32.e_entry); entry 67 arch/mips/tools/elf-entry.c entry = be32toh(hdr.ehdr32.e_entry); entry 74 arch/mips/tools/elf-entry.c entry = (int64_t)(int32_t)entry; entry 80 arch/mips/tools/elf-entry.c entry = le64toh(hdr.ehdr64.e_entry); entry 83 arch/mips/tools/elf-entry.c entry = be64toh(hdr.ehdr64.e_entry); entry 94 arch/mips/tools/elf-entry.c printf("0x%016" PRIx64 "\n", entry); entry 1290 arch/nds32/kernel/perf_event_cpu.c struct perf_callchain_entry_ctx *entry = data; entry 1292 arch/nds32/kernel/perf_event_cpu.c perf_callchain_store(entry, fr->lp); entry 1301 arch/nds32/kernel/perf_event_cpu.c user_backtrace(struct perf_callchain_entry_ctx *entry, unsigned long fp) entry 1320 arch/nds32/kernel/perf_event_cpu.c perf_callchain_store(entry, lp); entry 1325 arch/nds32/kernel/perf_event_cpu.c user_backtrace_opt_size(struct perf_callchain_entry_ctx *entry, entry 1347 arch/nds32/kernel/perf_event_cpu.c perf_callchain_store(entry, lp); entry 1363 arch/nds32/kernel/perf_event_cpu.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, entry 1379 arch/nds32/kernel/perf_event_cpu.c perf_callchain_store(entry, regs->ipc); entry 1384 arch/nds32/kernel/perf_event_cpu.c if (entry->nr < PERF_MAX_STACK_DEPTH && entry 1435 arch/nds32/kernel/perf_event_cpu.c perf_callchain_store(entry, lp); entry 1438 arch/nds32/kernel/perf_event_cpu.c while ((entry->nr < PERF_MAX_STACK_DEPTH) && entry 1443 arch/nds32/kernel/perf_event_cpu.c fp = user_backtrace_opt_size(entry, fp); entry 1451 arch/nds32/kernel/perf_event_cpu.c perf_callchain_store(entry, lp); entry 1452 arch/nds32/kernel/perf_event_cpu.c while ((entry->nr < PERF_MAX_STACK_DEPTH) && entry 1457 arch/nds32/kernel/perf_event_cpu.c fp = user_backtrace(entry, fp); entry 1463 arch/nds32/kernel/perf_event_cpu.c perf_callchain_store(entry, lp); entry 1466 arch/nds32/kernel/perf_event_cpu.c while ((entry->nr < PERF_MAX_STACK_DEPTH) && entry 1470 arch/nds32/kernel/perf_event_cpu.c fp = user_backtrace(entry, fp); entry 1479 arch/nds32/kernel/perf_event_cpu.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 1491 arch/nds32/kernel/perf_event_cpu.c walk_stackframe(&fr, callchain_trace, entry); entry 269 arch/nds32/kernel/traps.c void do_debug_trap(unsigned long entry, unsigned long addr, entry 295 arch/nds32/kernel/traps.c void unhandled_exceptions(unsigned long entry, unsigned long addr, entry 298 arch/nds32/kernel/traps.c pr_emerg("Unhandled Exception: entry: %lx addr:%lx itype:%lx\n", entry, entry 306 arch/nds32/kernel/traps.c extern int do_page_fault(unsigned long entry, unsigned long addr, entry 313 arch/nds32/kernel/traps.c void do_dispatch_tlb_misc(unsigned long entry, unsigned long addr, entry 319 arch/nds32/kernel/traps.c do_page_fault(entry, addr, type, regs); entry 321 arch/nds32/kernel/traps.c unhandled_exceptions(entry, addr, type, regs); entry 337 arch/nds32/kernel/traps.c void do_dispatch_general(unsigned long entry, unsigned long addr, entry 358 arch/nds32/kernel/traps.c do_page_fault(entry, addr, type, regs); entry 376 arch/nds32/kernel/traps.c unhandled_exceptions(entry, addr, type, regs); entry 398 arch/nds32/kernel/traps.c do_debug_trap(entry, addr, itype, regs); entry 401 arch/nds32/kernel/traps.c unhandled_exceptions(entry, addr, type, regs); entry 70 arch/nds32/mm/fault.c void do_page_fault(unsigned long entry, unsigned long addr, entry 99 arch/nds32/mm/fault.c && (entry == ENTRY_PTE_NOT_PRESENT)) entry 106 arch/nds32/mm/fault.c if (entry == ENTRY_GENERAL_EXCPETION entry 168 arch/nds32/mm/fault.c if (entry == ENTRY_PTE_NOT_PRESENT) { entry 174 arch/nds32/mm/fault.c } else if (entry == ENTRY_TLB_MISC) { entry 273 arch/nds32/mm/fault.c tsk->thread.trap_no = entry; entry 290 arch/nds32/mm/fault.c const struct exception_table_entry *entry; entry 292 arch/nds32/mm/fault.c if ((entry = entry 296 arch/nds32/mm/fault.c instruction_pointer(regs) = entry->fixup; entry 342 arch/nds32/mm/fault.c tsk->thread.trap_no = entry; entry 357 arch/openrisc/kernel/traps.c const struct exception_table_entry *entry; entry 378 arch/openrisc/kernel/traps.c if ((entry = search_exception_tables(orig_pc))) { entry 379 arch/openrisc/kernel/traps.c regs->pc = entry->fixup; entry 399 arch/openrisc/kernel/traps.c const struct exception_table_entry *entry; entry 425 arch/openrisc/kernel/traps.c if ((entry = search_exception_tables(orig_pc))) { entry 426 arch/openrisc/kernel/traps.c regs->pc = entry->fixup; entry 228 arch/openrisc/mm/fault.c const struct exception_table_entry *entry; entry 232 arch/openrisc/mm/fault.c if ((entry = search_exception_tables(regs->pc)) != NULL) { entry 234 arch/openrisc/mm/fault.c regs->pc = entry->fixup; entry 33 arch/parisc/boot/compressed/misc.c extern void startup_continue(void *entry, unsigned long cmdline, entry 29 arch/parisc/include/asm/linkage.h .entry ASM_NL\ entry 25 arch/parisc/kernel/alternative.c struct alt_instr *entry; entry 29 arch/parisc/kernel/alternative.c for (entry = start; entry < end; entry++, index++) { entry 34 arch/parisc/kernel/alternative.c from = (u32 *)((ulong)&entry->orig_offset + entry->orig_offset); entry 35 arch/parisc/kernel/alternative.c len = entry->len; entry 36 arch/parisc/kernel/alternative.c cond = entry->cond; entry 37 arch/parisc/kernel/alternative.c replacement = entry->replacement; entry 87 arch/parisc/kernel/alternative.c source = (u32 *)((ulong)&entry->replacement + entry->replacement); entry 21 arch/parisc/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry, entry 24 arch/parisc/kernel/jump_label.c void *addr = (void *)jump_entry_code(entry); entry 28 arch/parisc/kernel/jump_label.c void *target = (void *)jump_entry_target(entry); entry 46 arch/parisc/kernel/jump_label.c void arch_jump_label_transform_static(struct jump_entry *entry, entry 869 arch/parisc/kernel/module.c Elf_Fdesc *entry; entry 872 arch/parisc/kernel/module.c entry = (Elf_Fdesc *)me->init; entry 873 arch/parisc/kernel/module.c printk("FINALIZE, ->init FPTR is %p, GP %lx ADDR %lx\n", entry, entry 874 arch/parisc/kernel/module.c entry->gp, entry->addr); entry 875 arch/parisc/kernel/module.c addr = (u32 *)entry->addr; entry 118 arch/parisc/mm/hugetlbpage.c pte_t *ptep, pte_t entry) entry 127 arch/parisc/mm/hugetlbpage.c set_pte(ptep, entry); entry 131 arch/parisc/mm/hugetlbpage.c pte_val(entry) += PAGE_SIZE; entry 138 arch/parisc/mm/hugetlbpage.c pte_t *ptep, pte_t entry) entry 143 arch/parisc/mm/hugetlbpage.c __set_huge_pte_at(mm, addr, ptep, entry); entry 152 arch/parisc/mm/hugetlbpage.c pte_t entry; entry 155 arch/parisc/mm/hugetlbpage.c entry = *ptep; entry 159 arch/parisc/mm/hugetlbpage.c return entry; entry 15 arch/powerpc/boot/opal.c u64 entry; entry 75 arch/powerpc/boot/opal.c if (getprop(opal_node, "opal-entry-address", &opal.entry, sizeof(u64)) < 0) entry 77 arch/powerpc/boot/opal.c opal.entry = be64_to_cpu(opal.entry); entry 46 arch/powerpc/boot/rs6000.h unsigned char entry[4]; /* entry pt. */ entry 332 arch/powerpc/include/asm/book3s/32/pgtable.h pte_t *ptep, pte_t entry, entry 336 arch/powerpc/include/asm/book3s/32/pgtable.h unsigned long set = pte_val(entry) & entry 375 arch/powerpc/include/asm/book3s/32/pgtable.h #define __swp_type(entry) ((entry).val & 0x1f) entry 376 arch/powerpc/include/asm/book3s/32/pgtable.h #define __swp_offset(entry) ((entry).val >> 5) entry 185 arch/powerpc/include/asm/book3s/64/hash.h static inline void hash__ptep_set_access_flags(pte_t *ptep, pte_t entry) entry 192 arch/powerpc/include/asm/book3s/64/hash.h val = pte_raw(entry) & mask; entry 790 arch/powerpc/include/asm/book3s/64/pgtable.h pte_t *ptep, pte_t entry, entry 795 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__ptep_set_access_flags(vma, ptep, entry, entry 797 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__ptep_set_access_flags(ptep, entry); entry 1239 arch/powerpc/include/asm/book3s/64/pgtable.h pmd_t entry, int dirty); entry 125 arch/powerpc/include/asm/book3s/64/radix.h pte_t entry, unsigned long address, entry 22 arch/powerpc/include/asm/book3s/pgtable.h pte_t *ptep, pte_t entry, int dirty); entry 112 arch/powerpc/include/asm/code-patching.h return ((func_descr_t *)func)->entry; entry 107 arch/powerpc/include/asm/eeh.h list_for_each_entry_safe(edev, tmp, &pe->edevs, entry) entry 146 arch/powerpc/include/asm/eeh.h struct list_head entry; /* Membership in eeh_pe.edevs */ entry 119 arch/powerpc/include/asm/iommu.h #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \ entry 120 arch/powerpc/include/asm/iommu.h ((tbl)->it_ops->useraddrptr((tbl), (entry), false)) entry 121 arch/powerpc/include/asm/iommu.h #define IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry) \ entry 122 arch/powerpc/include/asm/iommu.h ((tbl)->it_ops->useraddrptr((tbl), (entry), true)) entry 211 arch/powerpc/include/asm/iommu.h unsigned long entry, unsigned long *hpa, entry 215 arch/powerpc/include/asm/iommu.h unsigned long entry, unsigned long *hpa, entry 218 arch/powerpc/include/asm/iommu.h unsigned long entry, unsigned long pages); entry 482 arch/powerpc/include/asm/kvm_host.h struct mmio_hpte_cache_entry entry[MMIO_HPTE_CACHE_SIZE]; entry 188 arch/powerpc/include/asm/lppaca.h extern void (*dtl_consumer)(struct dtl_entry *entry, u64 index); entry 316 arch/powerpc/include/asm/nohash/32/pgtable.h pte_t *ptep, pte_t entry, entry 322 arch/powerpc/include/asm/nohash/32/pgtable.h unsigned long set = pte_val(entry) & pte_val(pte_set); entry 323 arch/powerpc/include/asm/nohash/32/pgtable.h unsigned long clr = ~pte_val(entry) & ~pte_val(pte_clr); entry 381 arch/powerpc/include/asm/nohash/32/pgtable.h #define __swp_type(entry) ((entry).val & 0x1f) entry 382 arch/powerpc/include/asm/nohash/32/pgtable.h #define __swp_offset(entry) ((entry).val >> 5) entry 306 arch/powerpc/include/asm/nohash/64/pgtable.h pte_t *ptep, pte_t entry, entry 310 arch/powerpc/include/asm/nohash/64/pgtable.h unsigned long bits = pte_val(entry) & entry 231 arch/powerpc/include/asm/nohash/pgtable.h pte_t *ptep, pte_t entry, int dirty); entry 983 arch/powerpc/include/asm/opal-api.h struct opal_sg_entry entry[]; entry 266 arch/powerpc/include/asm/ppc_asm.h #define _ASM_NOKPROBE_SYMBOL(entry) \ entry 268 arch/powerpc/include/asm/ppc_asm.h PPC_LONG (entry) ; \ entry 271 arch/powerpc/include/asm/ppc_asm.h #define _ASM_NOKPROBE_SYMBOL(entry) entry 56 arch/powerpc/include/asm/rtas.h unsigned long entry; /* physical address pointer */ entry 27 arch/powerpc/include/asm/types.h unsigned long entry; entry 284 arch/powerpc/kernel/asm-offsets.c OFFSET(RTASENTRY, rtas_t, entry); entry 410 arch/powerpc/kernel/eeh_driver.c if (list_is_last(&edev->entry, &edev->pe->edevs)) entry 716 arch/powerpc/kernel/eeh_driver.c edev = list_first_entry(&pe->edevs, struct eeh_dev, entry); entry 388 arch/powerpc/kernel/eeh_pe.c list_add_tail(&edev->entry, &pe->edevs); entry 411 arch/powerpc/kernel/eeh_pe.c list_add_tail(&edev->entry, &pe->edevs); entry 453 arch/powerpc/kernel/eeh_pe.c list_add_tail(&edev->entry, &pe->edevs); entry 484 arch/powerpc/kernel/eeh_pe.c list_del(&edev->entry); entry 609 arch/powerpc/kernel/eeh_pe.c list_for_each_entry(edev, &pe->edevs, entry) { entry 934 arch/powerpc/kernel/eeh_pe.c edev = list_first_entry_or_null(&pe->edevs, struct eeh_dev, entry); entry 299 arch/powerpc/kernel/iommu.c unsigned long entry; entry 303 arch/powerpc/kernel/iommu.c entry = iommu_range_alloc(dev, tbl, npages, NULL, mask, align_order); entry 305 arch/powerpc/kernel/iommu.c if (unlikely(entry == DMA_MAPPING_ERROR)) entry 308 arch/powerpc/kernel/iommu.c entry += tbl->it_offset; /* Offset into real TCE table */ entry 309 arch/powerpc/kernel/iommu.c ret = entry << tbl->it_page_shift; /* Set the return dma address */ entry 312 arch/powerpc/kernel/iommu.c build_fail = tbl->it_ops->set(tbl, entry, npages, entry 339 arch/powerpc/kernel/iommu.c unsigned long entry, free_entry; entry 341 arch/powerpc/kernel/iommu.c entry = dma_addr >> tbl->it_page_shift; entry 342 arch/powerpc/kernel/iommu.c free_entry = entry - tbl->it_offset; entry 345 arch/powerpc/kernel/iommu.c (entry < tbl->it_offset)) { entry 348 arch/powerpc/kernel/iommu.c printk(KERN_INFO "\tentry = 0x%lx\n", entry); entry 365 arch/powerpc/kernel/iommu.c unsigned long entry) entry 371 arch/powerpc/kernel/iommu.c if (entry >= largepool_start) { entry 374 arch/powerpc/kernel/iommu.c unsigned int pool_nr = entry / tbl->poolsize; entry 386 arch/powerpc/kernel/iommu.c unsigned long entry, free_entry; entry 390 arch/powerpc/kernel/iommu.c entry = dma_addr >> tbl->it_page_shift; entry 391 arch/powerpc/kernel/iommu.c free_entry = entry - tbl->it_offset; entry 398 arch/powerpc/kernel/iommu.c tbl->it_ops->clear(tbl, entry, npages); entry 447 arch/powerpc/kernel/iommu.c unsigned long vaddr, npages, entry, slen; entry 462 arch/powerpc/kernel/iommu.c entry = iommu_range_alloc(dev, tbl, npages, &handle, entry 468 arch/powerpc/kernel/iommu.c if (unlikely(entry == DMA_MAPPING_ERROR)) { entry 478 arch/powerpc/kernel/iommu.c entry += tbl->it_offset; entry 479 arch/powerpc/kernel/iommu.c dma_addr = entry << tbl->it_page_shift; entry 483 arch/powerpc/kernel/iommu.c npages, entry, dma_addr); entry 486 arch/powerpc/kernel/iommu.c build_fail = tbl->it_ops->set(tbl, entry, npages, entry 1018 arch/powerpc/kernel/iommu.c unsigned long entry, unsigned long *hpa, entry 1024 arch/powerpc/kernel/iommu.c ret = tbl->it_ops->xchg_no_kill(tbl, entry, hpa, direction, false); entry 1036 arch/powerpc/kernel/iommu.c unsigned long entry, unsigned long pages) entry 1039 arch/powerpc/kernel/iommu.c tbl->it_ops->tce_kill(tbl, entry, pages, false); entry 10 arch/powerpc/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry, entry 13 arch/powerpc/kernel/jump_label.c u32 *addr = (u32 *)(unsigned long)entry->code; entry 16 arch/powerpc/kernel/jump_label.c patch_branch(addr, entry->target, 0); entry 526 arch/powerpc/kernel/kprobes.c const struct exception_table_entry *entry; entry 570 arch/powerpc/kernel/kprobes.c if ((entry = search_exception_tables(regs->nip)) != NULL) { entry 571 arch/powerpc/kernel/kprobes.c regs->nip = extable_fixup(entry); entry 587 arch/powerpc/kernel/kprobes.c unsigned long arch_deref_entry_point(void *entry) entry 590 arch/powerpc/kernel/kprobes.c if (!kernel_text_address((unsigned long)entry)) entry 591 arch/powerpc/kernel/kprobes.c return ppc_global_function_entry(entry); entry 594 arch/powerpc/kernel/kprobes.c return (unsigned long)entry; entry 77 arch/powerpc/kernel/machine_kexec_64.c unsigned long entry; entry 91 arch/powerpc/kernel/machine_kexec_64.c for (entry = ind; !(entry & IND_DONE); entry = *ptr++) { entry 92 arch/powerpc/kernel/machine_kexec_64.c addr = __va(entry & PAGE_MASK); entry 94 arch/powerpc/kernel/machine_kexec_64.c switch (entry & IND_FLAGS) { entry 582 arch/powerpc/kernel/mce_power.c const struct exception_table_entry *entry; entry 584 arch/powerpc/kernel/mce_power.c entry = search_kernel_exception_table(regs->nip); entry 585 arch/powerpc/kernel/mce_power.c if (entry) { entry 587 arch/powerpc/kernel/mce_power.c regs->nip = extable_fixup(entry); entry 161 arch/powerpc/kernel/module_32.c static inline int entry_matches(struct ppc_plt_entry *entry, Elf32_Addr val) entry 163 arch/powerpc/kernel/module_32.c if (entry->jump[0] != (PPC_INST_ADDIS | __PPC_RT(R12) | PPC_HA(val))) entry 165 arch/powerpc/kernel/module_32.c if (entry->jump[1] != (PPC_INST_ADDI | __PPC_RT(R12) | __PPC_RA(R12) | entry 177 arch/powerpc/kernel/module_32.c struct ppc_plt_entry *entry; entry 183 arch/powerpc/kernel/module_32.c entry = (void *)sechdrs[mod->arch.core_plt_section].sh_addr; entry 185 arch/powerpc/kernel/module_32.c entry = (void *)sechdrs[mod->arch.init_plt_section].sh_addr; entry 188 arch/powerpc/kernel/module_32.c while (entry->jump[0]) { entry 189 arch/powerpc/kernel/module_32.c if (entry_matches(entry, val)) return (uint32_t)entry; entry 190 arch/powerpc/kernel/module_32.c entry++; entry 199 arch/powerpc/kernel/module_32.c entry->jump[0] = PPC_INST_ADDIS | __PPC_RT(R12) | PPC_HA(val); entry 200 arch/powerpc/kernel/module_32.c entry->jump[1] = PPC_INST_ADDI | __PPC_RT(R12) | __PPC_RA(R12) | PPC_LO(val); entry 201 arch/powerpc/kernel/module_32.c entry->jump[2] = PPC_INST_MTCTR | __PPC_RS(R12); entry 202 arch/powerpc/kernel/module_32.c entry->jump[3] = PPC_INST_BCTR; entry 204 arch/powerpc/kernel/module_32.c pr_debug("Initialized plt for 0x%x at %p\n", val, entry); entry 205 arch/powerpc/kernel/module_32.c return (uint32_t)entry; entry 400 arch/powerpc/kernel/module_64.c struct ppc64_stub_entry *entry, entry 406 arch/powerpc/kernel/module_64.c memcpy(entry->jump, ppc64_stub_insns, sizeof(ppc64_stub_insns)); entry 409 arch/powerpc/kernel/module_64.c reladdr = (unsigned long)entry - my_r2(sechdrs, me); entry 415 arch/powerpc/kernel/module_64.c pr_debug("Stub %p get data from reladdr %li\n", entry, reladdr); entry 417 arch/powerpc/kernel/module_64.c entry->jump[0] |= PPC_HA(reladdr); entry 418 arch/powerpc/kernel/module_64.c entry->jump[1] |= PPC_LO(reladdr); entry 419 arch/powerpc/kernel/module_64.c entry->funcdata = func_desc(addr); entry 420 arch/powerpc/kernel/module_64.c entry->magic = STUB_MAGIC; entry 768 arch/powerpc/kernel/module_64.c struct ppc64_stub_entry *entry; entry 786 arch/powerpc/kernel/module_64.c num_stubs = sechdrs[me->arch.stubs_section].sh_size / sizeof(*entry); entry 789 arch/powerpc/kernel/module_64.c entry = (void *)sechdrs[me->arch.stubs_section].sh_addr; entry 790 arch/powerpc/kernel/module_64.c for (i = 0; i < num_stubs && stub_func_addr(entry->funcdata); i++, entry++); entry 797 arch/powerpc/kernel/module_64.c memcpy(entry->jump, stub_insns, sizeof(stub_insns)); entry 807 arch/powerpc/kernel/module_64.c entry->jump[1] |= PPC_HA(reladdr); entry 808 arch/powerpc/kernel/module_64.c entry->jump[2] |= PPC_LO(reladdr); entry 811 arch/powerpc/kernel/module_64.c entry->funcdata = func_desc(addr); entry 812 arch/powerpc/kernel/module_64.c entry->magic = STUB_MAGIC; entry 814 arch/powerpc/kernel/module_64.c return (unsigned long)entry; entry 1740 arch/powerpc/kernel/process.c unsigned long entry; entry 1744 arch/powerpc/kernel/process.c entry = start; entry 1766 arch/powerpc/kernel/process.c __get_user(entry, (unsigned long __user *)start); entry 1773 arch/powerpc/kernel/process.c entry += load_addr; entry 1778 arch/powerpc/kernel/process.c regs->nip = entry; entry 138 arch/powerpc/kernel/prom_init.c extern int enter_prom(struct prom_args *args, unsigned long entry); entry 140 arch/powerpc/kernel/prom_init.c static inline int enter_prom(struct prom_args *args, unsigned long entry) entry 142 arch/powerpc/kernel/prom_init.c return ((int (*)(struct prom_args *))entry)(args); entry 1778 arch/powerpc/kernel/prom_init.c u32 base, entry = 0; entry 1807 arch/powerpc/kernel/prom_init.c if (call_prom_ret("call-method", 3, 2, &entry, entry 1810 arch/powerpc/kernel/prom_init.c || entry == 0) { entry 1821 arch/powerpc/kernel/prom_init.c val = cpu_to_be32(entry); entry 1831 arch/powerpc/kernel/prom_init.c prom_debug("rtas entry = 0x%x\n", entry); entry 1845 arch/powerpc/kernel/prom_init.c u32 entry = 0, size = 0, succ = 0; entry 1894 arch/powerpc/kernel/prom_init.c if (call_prom_ret("call-method", 4, 2, &entry, entry 1896 arch/powerpc/kernel/prom_init.c ibmvtpm_inst, size, base) != 0 || entry == 0) { entry 453 arch/powerpc/kernel/rtas.c if (!rtas.entry || token == RTAS_UNKNOWN_SERVICE) entry 1072 arch/powerpc/kernel/rtas.c if (!rtas.entry) entry 1157 arch/powerpc/kernel/rtas.c u32 base, size, entry; entry 1177 arch/powerpc/kernel/rtas.c no_entry = of_property_read_u32(rtas.dev, "linux,rtas-entry", &entry); entry 1178 arch/powerpc/kernel/rtas.c rtas.entry = no_entry ? rtas.base : entry; entry 1214 arch/powerpc/kernel/rtas.c rtas.entry = *entryp; entry 566 arch/powerpc/kernel/rtasd.c struct proc_dir_entry *entry; entry 574 arch/powerpc/kernel/rtasd.c entry = proc_create("powerpc/rtas/error_log", 0400, NULL, entry 576 arch/powerpc/kernel/rtasd.c if (!entry) entry 895 arch/powerpc/kernel/signal_64.c err |= get_user(regs->nip, &funct_desc_ptr->entry); entry 126 arch/powerpc/kernel/trace/ftrace.c unsigned long entry, ptr, tramp; entry 154 arch/powerpc/kernel/trace/ftrace.c entry = ppc_global_function_entry((void *)addr); entry 156 arch/powerpc/kernel/trace/ftrace.c if (ptr != entry) { entry 157 arch/powerpc/kernel/trace/ftrace.c pr_err("addr %lx does not match expected %lx\n", ptr, entry); entry 515 arch/powerpc/kernel/trace/ftrace.c unsigned long entry, ptr, tramp; entry 552 arch/powerpc/kernel/trace/ftrace.c entry = ppc_global_function_entry((void *)addr); entry 554 arch/powerpc/kernel/trace/ftrace.c if (ptr != entry) { entry 555 arch/powerpc/kernel/trace/ftrace.c pr_err("addr %lx does not match expected %lx\n", ptr, entry); entry 618 arch/powerpc/kernel/trace/ftrace.c unsigned long tramp, entry, ptr; entry 621 arch/powerpc/kernel/trace/ftrace.c entry = ppc_global_function_entry((void *)ftrace_caller); entry 624 arch/powerpc/kernel/trace/ftrace.c if (ptr != entry) { entry 626 arch/powerpc/kernel/trace/ftrace.c entry = ppc_global_function_entry((void *)ftrace_regs_caller); entry 627 arch/powerpc/kernel/trace/ftrace.c if (ptr != entry) { entry 705 arch/powerpc/kernel/trace/ftrace.c unsigned long entry, ptr, tramp; entry 728 arch/powerpc/kernel/trace/ftrace.c entry = ppc_global_function_entry((void *)old_addr); entry 732 arch/powerpc/kernel/trace/ftrace.c if (tramp != entry) { entry 742 arch/powerpc/kernel/trace/ftrace.c if (ptr != entry) { entry 743 arch/powerpc/kernel/trace/ftrace.c pr_err("addr %lx does not match expected %lx\n", ptr, entry); entry 771 arch/powerpc/kernel/trace/ftrace.c entry = ppc_global_function_entry((void *)addr); entry 773 arch/powerpc/kernel/trace/ftrace.c if (ptr != entry) { entry 774 arch/powerpc/kernel/trace/ftrace.c pr_err("addr %lx does not match expected %lx\n", ptr, entry); entry 544 arch/powerpc/kernel/traps.c const struct exception_table_entry *entry; entry 548 arch/powerpc/kernel/traps.c && (entry = search_exception_tables(regs->nip)) != NULL) { entry 570 arch/powerpc/kernel/traps.c regs->nip = extable_fixup(entry); entry 227 arch/powerpc/kvm/book3s_64_mmu_radix.c struct prtb_entry entry; entry 234 arch/powerpc/kvm/book3s_64_mmu_radix.c if ((table_index * sizeof(entry)) >= size) entry 238 arch/powerpc/kvm/book3s_64_mmu_radix.c ptbl = (table & PRTB_MASK) + (table_index * sizeof(entry)); entry 239 arch/powerpc/kvm/book3s_64_mmu_radix.c ret = kvm_read_guest(kvm, ptbl, &entry, sizeof(entry)); entry 244 arch/powerpc/kvm/book3s_64_mmu_radix.c root = be64_to_cpu(entry.prtb0); entry 414 arch/powerpc/kvm/book3s_64_vio.c unsigned long entry) entry 419 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_xchg_no_kill(mm, tbl, entry, &hpa, &dir); entry 423 arch/powerpc/kvm/book3s_64_vio.c struct iommu_table *tbl, unsigned long entry) entry 427 arch/powerpc/kvm/book3s_64_vio.c __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); entry 444 arch/powerpc/kvm/book3s_64_vio.c struct iommu_table *tbl, unsigned long entry) entry 450 arch/powerpc/kvm/book3s_64_vio.c if (WARN_ON_ONCE(iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa, entry 457 arch/powerpc/kvm/book3s_64_vio.c ret = kvmppc_tce_iommu_mapped_dec(kvm, tbl, entry); entry 459 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa, &dir); entry 466 arch/powerpc/kvm/book3s_64_vio.c unsigned long entry) entry 470 arch/powerpc/kvm/book3s_64_vio.c unsigned long io_entry = entry * subpages; entry 482 arch/powerpc/kvm/book3s_64_vio.c unsigned long entry, unsigned long ua, entry 487 arch/powerpc/kvm/book3s_64_vio.c __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); entry 505 arch/powerpc/kvm/book3s_64_vio.c ret = iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa, &dir); entry 512 arch/powerpc/kvm/book3s_64_vio.c kvmppc_tce_iommu_mapped_dec(kvm, tbl, entry); entry 521 arch/powerpc/kvm/book3s_64_vio.c unsigned long entry, unsigned long ua, entry 526 arch/powerpc/kvm/book3s_64_vio.c unsigned long io_entry = entry * subpages; entry 546 arch/powerpc/kvm/book3s_64_vio.c unsigned long entry, ua = 0; entry 573 arch/powerpc/kvm/book3s_64_vio.c entry = ioba >> stt->page_shift; entry 578 arch/powerpc/kvm/book3s_64_vio.c stit->tbl, entry); entry 581 arch/powerpc/kvm/book3s_64_vio.c entry, ua, dir); entry 583 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_kill(stit->tbl, entry, 1); entry 586 arch/powerpc/kvm/book3s_64_vio.c kvmppc_clear_tce(vcpu->kvm->mm, stit->tbl, entry); entry 591 arch/powerpc/kvm/book3s_64_vio.c kvmppc_tce_put(stt, entry, tce); entry 606 arch/powerpc/kvm/book3s_64_vio.c unsigned long entry, ua = 0; entry 615 arch/powerpc/kvm/book3s_64_vio.c entry = ioba >> stt->page_shift; entry 673 arch/powerpc/kvm/book3s_64_vio.c stit->tbl, entry + i, ua, entry 678 arch/powerpc/kvm/book3s_64_vio.c entry); entry 683 arch/powerpc/kvm/book3s_64_vio.c kvmppc_tce_put(stt, entry + i, tce); entry 688 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_kill(stit->tbl, entry, npages); entry 718 arch/powerpc/kvm/book3s_64_vio.c unsigned long entry = ioba >> stt->page_shift; entry 722 arch/powerpc/kvm/book3s_64_vio.c stit->tbl, entry + i); entry 731 arch/powerpc/kvm/book3s_64_vio.c kvmppc_clear_tce(vcpu->kvm->mm, stit->tbl, entry); entry 223 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry, unsigned long *hpa, entry 228 arch/powerpc/kvm/book3s_64_vio_hv.c ret = tbl->it_ops->xchg_no_kill(tbl, entry, hpa, direction, true); entry 232 arch/powerpc/kvm/book3s_64_vio_hv.c __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); entry 245 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry, unsigned long pages) entry 248 arch/powerpc/kvm/book3s_64_vio_hv.c tbl->it_ops->tce_kill(tbl, entry, pages, true); entry 252 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry) entry 257 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir); entry 261 arch/powerpc/kvm/book3s_64_vio_hv.c struct iommu_table *tbl, unsigned long entry) entry 265 arch/powerpc/kvm/book3s_64_vio_hv.c __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); entry 283 arch/powerpc/kvm/book3s_64_vio_hv.c struct iommu_table *tbl, unsigned long entry) entry 289 arch/powerpc/kvm/book3s_64_vio_hv.c if (iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir)) entry 299 arch/powerpc/kvm/book3s_64_vio_hv.c ret = kvmppc_rm_tce_iommu_mapped_dec(kvm, tbl, entry); entry 301 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir); entry 308 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry) entry 312 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long io_entry = entry * subpages; entry 324 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry, unsigned long ua, entry 329 arch/powerpc/kvm/book3s_64_vio_hv.c __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); entry 347 arch/powerpc/kvm/book3s_64_vio_hv.c ret = iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir); entry 358 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_tce_iommu_mapped_dec(kvm, tbl, entry); entry 367 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry, unsigned long ua, entry 372 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long io_entry = entry * subpages; entry 392 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry, ua = 0; entry 418 arch/powerpc/kvm/book3s_64_vio_hv.c entry = ioba >> stt->page_shift; entry 423 arch/powerpc/kvm/book3s_64_vio_hv.c stit->tbl, entry); entry 426 arch/powerpc/kvm/book3s_64_vio_hv.c stit->tbl, entry, ua, dir); entry 428 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_kill_rm(stit->tbl, entry, 1); entry 431 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_clear_tce(vcpu->kvm, stit->tbl, entry); entry 436 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_tce_put(stt, entry, tce); entry 483 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long tces, entry, ua = 0; entry 496 arch/powerpc/kvm/book3s_64_vio_hv.c entry = ioba >> stt->page_shift; entry 576 arch/powerpc/kvm/book3s_64_vio_hv.c stit->tbl, entry + i, ua, entry 581 arch/powerpc/kvm/book3s_64_vio_hv.c entry); entry 586 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_tce_put(stt, entry + i, tce); entry 591 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_kill_rm(stit->tbl, entry, npages); entry 625 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry = ioba >> stt->page_shift; entry 629 arch/powerpc/kvm/book3s_64_vio_hv.c stit->tbl, entry + i); entry 638 arch/powerpc/kvm/book3s_64_vio_hv.c kvmppc_rm_clear_tce(vcpu->kvm, stit->tbl, entry); entry 762 arch/powerpc/kvm/book3s_hv_nested.c struct llist_node *entry = ((struct llist_head *) rmapp)->first; entry 774 arch/powerpc/kvm/book3s_hv_nested.c for_each_nest_rmap_safe(cursor, entry, &rmap) { entry 828 arch/powerpc/kvm/book3s_hv_nested.c struct llist_node *entry = ((struct llist_head *) rmapp)->first; entry 838 arch/powerpc/kvm/book3s_hv_nested.c for_each_nest_rmap_safe(cursor, entry, &rmap) entry 866 arch/powerpc/kvm/book3s_hv_nested.c struct llist_node *entry = llist_del_all((struct llist_head *) rmapp); entry 870 arch/powerpc/kvm/book3s_hv_nested.c for_each_nest_rmap_safe(cursor, entry, &rmap) { entry 906 arch/powerpc/kvm/book3s_hv_nested.c struct llist_node *entry; entry 908 arch/powerpc/kvm/book3s_hv_nested.c entry = llist_del_all((struct llist_head *) rmapp); entry 909 arch/powerpc/kvm/book3s_hv_nested.c for_each_nest_rmap_safe(cursor, entry, &rmap) entry 1083 arch/powerpc/kvm/book3s_hv_rm_mmu.c struct mmio_hpte_cache_entry *entry = NULL; entry 1088 arch/powerpc/kvm/book3s_hv_rm_mmu.c entry = &vcpu->arch.mmio_cache.entry[i]; entry 1089 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (entry->mmio_update == mmio_update) { entry 1090 arch/powerpc/kvm/book3s_hv_rm_mmu.c pshift = entry->slb_base_pshift; entry 1091 arch/powerpc/kvm/book3s_hv_rm_mmu.c if ((entry->eaddr >> pshift) == (eaddr >> pshift) && entry 1092 arch/powerpc/kvm/book3s_hv_rm_mmu.c entry->slb_v == slb_v) entry 1093 arch/powerpc/kvm/book3s_hv_rm_mmu.c return entry; entry 1108 arch/powerpc/kvm/book3s_hv_rm_mmu.c return &vcpu->arch.mmio_cache.entry[index]; entry 383 arch/powerpc/kvm/book3s_xive_template.c u32 idx, toggle, entry, irq, hw_num; entry 396 arch/powerpc/kvm/book3s_xive_template.c entry = be32_to_cpup(qpage + idx); entry 399 arch/powerpc/kvm/book3s_xive_template.c if ((entry >> 31) == toggle) entry 401 arch/powerpc/kvm/book3s_xive_template.c irq = entry & 0x7fffffff; entry 419 arch/powerpc/kvm/book3s_xive_template.c qpage[idx] = cpu_to_be32((entry & 0x80000000) | XICS_DUMMY); entry 53 arch/powerpc/kvm/e500.c struct id *entry[NUM_TIDS]; entry 70 arch/powerpc/kvm/e500.c static inline int local_sid_setup_one(struct id *entry) entry 77 arch/powerpc/kvm/e500.c __this_cpu_write(pcpu_sids.entry[sid], entry); entry 78 arch/powerpc/kvm/e500.c entry->val = sid; entry 79 arch/powerpc/kvm/e500.c entry->pentry = this_cpu_ptr(&pcpu_sids.entry[sid]); entry 104 arch/powerpc/kvm/e500.c static inline int local_sid_lookup(struct id *entry) entry 106 arch/powerpc/kvm/e500.c if (entry && entry->val != 0 && entry 107 arch/powerpc/kvm/e500.c __this_cpu_read(pcpu_sids.entry[entry->val]) == entry && entry 108 arch/powerpc/kvm/e500.c entry->pentry == this_cpu_ptr(&pcpu_sids.entry[entry->val])) entry 109 arch/powerpc/kvm/e500.c return entry->val; entry 276 arch/powerpc/kvm/e500.h struct kvmppc_vcpu_e500 *vcpu_e500, int tlbsel, int entry) entry 279 arch/powerpc/kvm/e500.h return &vcpu_e500->gtlb_arch[offset + entry]; entry 481 arch/powerpc/lib/feature-fixups.c static long calc_offset(struct fixup_entry *entry, unsigned int *p) entry 483 arch/powerpc/lib/feature-fixups.c return (unsigned long)p - (unsigned long)entry; entry 63 arch/powerpc/mm/book3s64/iommu_api.c unsigned long entry, chunk; entry 103 arch/powerpc/mm/book3s64/iommu_api.c for (entry = 0; entry < entries; entry += chunk) { entry 104 arch/powerpc/mm/book3s64/iommu_api.c unsigned long n = min(entries - entry, chunk); entry 106 arch/powerpc/mm/book3s64/iommu_api.c ret = get_user_pages(ua + (entry << PAGE_SHIFT), n, entry 108 arch/powerpc/mm/book3s64/iommu_api.c mem->hpages + entry, NULL); entry 344 arch/powerpc/mm/book3s64/iommu_api.c const long entry = (ua - mem->ua) >> PAGE_SHIFT; entry 347 arch/powerpc/mm/book3s64/iommu_api.c if (entry >= mem->entries) entry 358 arch/powerpc/mm/book3s64/iommu_api.c va = &mem->hpas[entry]; entry 368 arch/powerpc/mm/book3s64/iommu_api.c const long entry = (ua - mem->ua) >> PAGE_SHIFT; entry 371 arch/powerpc/mm/book3s64/iommu_api.c if (entry >= mem->entries) entry 382 arch/powerpc/mm/book3s64/iommu_api.c pa = (void *) vmalloc_to_phys(&mem->hpas[entry]); entry 394 arch/powerpc/mm/book3s64/iommu_api.c long entry; entry 405 arch/powerpc/mm/book3s64/iommu_api.c entry = (ua - mem->ua) >> PAGE_SHIFT; entry 406 arch/powerpc/mm/book3s64/iommu_api.c va = &mem->hpas[entry]; entry 36 arch/powerpc/mm/book3s64/pgtable.c pmd_t *pmdp, pmd_t entry, int dirty) entry 43 arch/powerpc/mm/book3s64/pgtable.c changed = !pmd_same(*(pmdp), entry); entry 50 arch/powerpc/mm/book3s64/pgtable.c pmd_pte(entry), address, MMU_PAGE_2M); entry 1033 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t entry, unsigned long address, int psize) entry 1036 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long set = pte_val(entry) & (_PAGE_DIRTY | _PAGE_ACCESSED | entry 1039 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long change = pte_val(entry) ^ pte_val(*ptep); entry 650 arch/powerpc/mm/fault.c const struct exception_table_entry *entry; entry 653 arch/powerpc/mm/fault.c if ((entry = search_exception_tables(regs->nip)) != NULL) { entry 654 arch/powerpc/mm/fault.c regs->nip = extable_fixup(entry); entry 56 arch/powerpc/mm/nohash/44x.c unsigned int entry = tlb_44x_hwater--; entry 70 arch/powerpc/mm/nohash/44x.c "r" (entry), entry 217 arch/powerpc/mm/numa.c const __be32 *entry; entry 219 arch/powerpc/mm/numa.c entry = &associativity[be32_to_cpu(distance_ref_points[i]) - 1]; entry 220 arch/powerpc/mm/numa.c distance_lookup_table[nid][i] = of_read_number(entry, 1); entry 209 arch/powerpc/mm/pgtable.c pte_t *ptep, pte_t entry, int dirty) entry 212 arch/powerpc/mm/pgtable.c entry = set_access_flags_filter(entry, vma, dirty); entry 213 arch/powerpc/mm/pgtable.c changed = !pte_same(*(ptep), entry); entry 216 arch/powerpc/mm/pgtable.c __ptep_set_access_flags(vma, ptep, entry, entry 280 arch/powerpc/mm/ptdump/hashpagetable.c struct mmu_psize_def entry; entry 297 arch/powerpc/mm/ptdump/hashpagetable.c entry = mmu_psize_defs[bps]; entry 299 arch/powerpc/mm/ptdump/hashpagetable.c penc = entry.penc[idx]; entry 73 arch/powerpc/oprofile/cell/spu_profiler.c static void spu_pc_extract(int cpu, int entry) entry 96 arch/powerpc/oprofile/cell/spu_profiler.c samples[spu * TRACE_ARRAY_SIZE + entry] entry 98 arch/powerpc/oprofile/cell/spu_profiler.c samples[(spu + SPUS_PER_TB_ENTRY) * TRACE_ARRAY_SIZE + entry] entry 109 arch/powerpc/oprofile/cell/spu_profiler.c int entry; entry 113 arch/powerpc/oprofile/cell/spu_profiler.c entry = 0; entry 118 arch/powerpc/oprofile/cell/spu_profiler.c spu_pc_extract(cpu, entry); entry 120 arch/powerpc/oprofile/cell/spu_profiler.c entry++; entry 122 arch/powerpc/oprofile/cell/spu_profiler.c if (entry >= TRACE_ARRAY_SIZE) entry 129 arch/powerpc/oprofile/cell/spu_profiler.c return entry; entry 47 arch/powerpc/perf/callchain.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) entry 57 arch/powerpc/perf/callchain.c perf_callchain_store(entry, perf_instruction_pointer(regs)); entry 76 arch/powerpc/perf/callchain.c perf_callchain_store_context(entry, PERF_CONTEXT_KERNEL); entry 98 arch/powerpc/perf/callchain.c perf_callchain_store(entry, next_ip); entry 232 arch/powerpc/perf/callchain.c static void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry, entry 245 arch/powerpc/perf/callchain.c perf_callchain_store(entry, next_ip); entry 247 arch/powerpc/perf/callchain.c while (entry->nr < entry->max_stack) { entry 274 arch/powerpc/perf/callchain.c perf_callchain_store_context(entry, PERF_CONTEXT_USER); entry 275 arch/powerpc/perf/callchain.c perf_callchain_store(entry, next_ip); entry 281 arch/powerpc/perf/callchain.c perf_callchain_store(entry, next_ip); entry 319 arch/powerpc/perf/callchain.c static inline void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry, entry 439 arch/powerpc/perf/callchain.c static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, entry 451 arch/powerpc/perf/callchain.c perf_callchain_store(entry, next_ip); entry 453 arch/powerpc/perf/callchain.c while (entry->nr < entry->max_stack) { entry 473 arch/powerpc/perf/callchain.c perf_callchain_store_context(entry, PERF_CONTEXT_USER); entry 474 arch/powerpc/perf/callchain.c perf_callchain_store(entry, next_ip); entry 480 arch/powerpc/perf/callchain.c perf_callchain_store(entry, next_ip); entry 487 arch/powerpc/perf/callchain.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) entry 490 arch/powerpc/perf/callchain.c perf_callchain_user_64(entry, regs); entry 492 arch/powerpc/perf/callchain.c perf_callchain_user_32(entry, regs); entry 40 arch/powerpc/platforms/4xx/hsta_msi.c struct msi_desc *entry; entry 50 arch/powerpc/platforms/4xx/hsta_msi.c for_each_pci_msi_entry(entry, dev) { entry 78 arch/powerpc/platforms/4xx/hsta_msi.c if (irq_set_msi_desc(hwirq, entry)) { entry 105 arch/powerpc/platforms/4xx/hsta_msi.c struct msi_desc *entry; entry 108 arch/powerpc/platforms/4xx/hsta_msi.c for_each_pci_msi_entry(entry, dev) { entry 109 arch/powerpc/platforms/4xx/hsta_msi.c if (!entry->irq) entry 112 arch/powerpc/platforms/4xx/hsta_msi.c irq = hsta_find_hwirq_offset(entry->irq); entry 116 arch/powerpc/platforms/4xx/hsta_msi.c irq_set_msi_desc(entry->irq, NULL); entry 119 arch/powerpc/platforms/4xx/hsta_msi.c entry->irq, irq); entry 70 arch/powerpc/platforms/4xx/msi.c struct msi_desc *entry; entry 82 arch/powerpc/platforms/4xx/msi.c for_each_pci_msi_entry(entry, dev) { entry 102 arch/powerpc/platforms/4xx/msi.c irq_set_msi_desc(virq, entry); entry 111 arch/powerpc/platforms/4xx/msi.c struct msi_desc *entry; entry 117 arch/powerpc/platforms/4xx/msi.c for_each_pci_msi_entry(entry, dev) { entry 118 arch/powerpc/platforms/4xx/msi.c if (!entry->irq) entry 120 arch/powerpc/platforms/4xx/msi.c hwirq = virq_to_hw(entry->irq); entry 121 arch/powerpc/platforms/4xx/msi.c irq_set_msi_desc(entry->irq, NULL); entry 122 arch/powerpc/platforms/4xx/msi.c irq_dispose_mapping(entry->irq); entry 595 arch/powerpc/platforms/512x/clock-commonclk.c static void mpc512x_clk_setup_mclk(struct mclk_setup_data *entry, size_t idx) entry 602 arch/powerpc/platforms/512x/clock-commonclk.c switch (entry->type) { entry 675 arch/powerpc/platforms/512x/clock-commonclk.c entry->name_mux0, entry 682 arch/powerpc/platforms/512x/clock-commonclk.c entry->name_en0, entry->name_mux0, entry 685 arch/powerpc/platforms/512x/clock-commonclk.c entry->name_div0, entry 686 arch/powerpc/platforms/512x/clock-commonclk.c entry->name_en0, CLK_SET_RATE_GATE, entry 688 arch/powerpc/platforms/512x/clock-commonclk.c if (entry->has_mclk1) { entry 690 arch/powerpc/platforms/512x/clock-commonclk.c entry->name_mclk, entry 691 arch/powerpc/platforms/512x/clock-commonclk.c &entry->parent_names_mux1[0], entry 692 arch/powerpc/platforms/512x/clock-commonclk.c ARRAY_SIZE(entry->parent_names_mux1), entry 696 arch/powerpc/platforms/512x/clock-commonclk.c entry->name_mclk, entry 697 arch/powerpc/platforms/512x/clock-commonclk.c entry->parent_names_mux1[0], entry 202 arch/powerpc/platforms/cell/axon_msi.c struct msi_desc *entry; entry 212 arch/powerpc/platforms/cell/axon_msi.c entry = first_pci_msi_entry(dev); entry 215 arch/powerpc/platforms/cell/axon_msi.c if (entry->msi_attrib.is_64) { entry 256 arch/powerpc/platforms/cell/axon_msi.c struct msi_desc *entry; entry 268 arch/powerpc/platforms/cell/axon_msi.c for_each_pci_msi_entry(entry, dev) { entry 277 arch/powerpc/platforms/cell/axon_msi.c irq_set_msi_desc(virq, entry); entry 287 arch/powerpc/platforms/cell/axon_msi.c struct msi_desc *entry; entry 291 arch/powerpc/platforms/cell/axon_msi.c for_each_pci_msi_entry(entry, dev) { entry 292 arch/powerpc/platforms/cell/axon_msi.c if (!entry->irq) entry 295 arch/powerpc/platforms/cell/axon_msi.c irq_set_msi_desc(entry->irq, NULL); entry 296 arch/powerpc/platforms/cell/axon_msi.c irq_dispose_mapping(entry->irq); entry 38 arch/powerpc/platforms/cell/spu_callbacks.c #define __SYSCALL(nr, entry) entry, entry 1086 arch/powerpc/platforms/cell/spufs/sched.c struct proc_dir_entry *entry; entry 1110 arch/powerpc/platforms/cell/spufs/sched.c entry = proc_create_single("spu_loadavg", 0, NULL, show_spu_loadavg); entry 1111 arch/powerpc/platforms/cell/spufs/sched.c if (!entry) entry 246 arch/powerpc/platforms/embedded6xx/holly.c const struct exception_table_entry *entry; entry 249 arch/powerpc/platforms/embedded6xx/holly.c if ((entry = search_exception_tables(regs->nip)) != NULL) { entry 252 arch/powerpc/platforms/embedded6xx/holly.c regs->nip = extable_fixup(entry); entry 167 arch/powerpc/platforms/embedded6xx/mpc7448_hpc2.c const struct exception_table_entry *entry; entry 170 arch/powerpc/platforms/embedded6xx/mpc7448_hpc2.c if ((entry = search_exception_tables(regs->nip)) != NULL) { entry 173 arch/powerpc/platforms/embedded6xx/mpc7448_hpc2.c regs->nip = extable_fixup(entry); entry 22 arch/powerpc/platforms/pasemi/idle.c void (*entry)(void); entry 26 arch/powerpc/platforms/pasemi/idle.c { .name = "spin", .entry = &idle_spin }, entry 27 arch/powerpc/platforms/pasemi/idle.c { .name = "doze", .entry = &idle_doze }, entry 73 arch/powerpc/platforms/pasemi/idle.c ppc_md.power_save = modes[current_mode].entry; entry 60 arch/powerpc/platforms/pasemi/msi.c struct msi_desc *entry; entry 65 arch/powerpc/platforms/pasemi/msi.c for_each_pci_msi_entry(entry, pdev) { entry 66 arch/powerpc/platforms/pasemi/msi.c if (!entry->irq) entry 69 arch/powerpc/platforms/pasemi/msi.c hwirq = virq_to_hw(entry->irq); entry 70 arch/powerpc/platforms/pasemi/msi.c irq_set_msi_desc(entry->irq, NULL); entry 71 arch/powerpc/platforms/pasemi/msi.c irq_dispose_mapping(entry->irq); entry 81 arch/powerpc/platforms/pasemi/msi.c struct msi_desc *entry; entry 93 arch/powerpc/platforms/pasemi/msi.c for_each_pci_msi_entry(entry, pdev) { entry 121 arch/powerpc/platforms/pasemi/msi.c irq_set_msi_desc(virq, entry); entry 1054 arch/powerpc/platforms/powernv/eeh-powernv.c edev = list_first_entry_or_null(&pe->edevs, struct eeh_dev, entry); entry 357 arch/powerpc/platforms/powernv/opal-lpc.c struct lpc_debugfs_entry *entry; entry 358 arch/powerpc/platforms/powernv/opal-lpc.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 359 arch/powerpc/platforms/powernv/opal-lpc.c if (!entry) entry 361 arch/powerpc/platforms/powernv/opal-lpc.c entry->lpc_type = type; entry 362 arch/powerpc/platforms/powernv/opal-lpc.c debugfs_create_file(fname, 0600, folder, entry, &lpc_fops); entry 43 arch/powerpc/platforms/powernv/opal.c u64 entry; entry 115 arch/powerpc/platforms/powernv/opal.c opal.entry = of_read_number(entryp, entrysz/4); entry 121 arch/powerpc/platforms/powernv/opal.c opal.entry, entryp, entrysz); entry 350 arch/powerpc/platforms/powernv/opal.c if (!opal.entry) entry 369 arch/powerpc/platforms/powernv/opal.c if (!opal.entry) entry 1062 arch/powerpc/platforms/powernv/opal.c sg->entry[i].data = cpu_to_be64(data); entry 1063 arch/powerpc/platforms/powernv/opal.c sg->entry[i].length = cpu_to_be64(length); entry 163 arch/powerpc/platforms/powernv/pci.c struct msi_desc *entry; entry 175 arch/powerpc/platforms/powernv/pci.c for_each_pci_msi_entry(entry, pdev) { entry 176 arch/powerpc/platforms/powernv/pci.c if (!entry->msi_attrib.is_64 && !phb->msi32_support) { entry 195 arch/powerpc/platforms/powernv/pci.c virq, entry->msi_attrib.is_64, &msg); entry 202 arch/powerpc/platforms/powernv/pci.c irq_set_msi_desc(virq, entry); entry 212 arch/powerpc/platforms/powernv/pci.c struct msi_desc *entry; entry 218 arch/powerpc/platforms/powernv/pci.c for_each_pci_msi_entry(entry, pdev) { entry 219 arch/powerpc/platforms/powernv/pci.c if (!entry->irq) entry 221 arch/powerpc/platforms/powernv/pci.c hwirq = virq_to_hw(entry->irq); entry 222 arch/powerpc/platforms/powernv/pci.c irq_set_msi_desc(entry->irq, NULL); entry 223 arch/powerpc/platforms/powernv/pci.c irq_dispose_mapping(entry->irq); entry 114 arch/powerpc/platforms/pseries/msi.c struct msi_desc *entry; entry 116 arch/powerpc/platforms/pseries/msi.c for_each_pci_msi_entry(entry, pdev) { entry 117 arch/powerpc/platforms/pseries/msi.c if (!entry->irq) entry 120 arch/powerpc/platforms/pseries/msi.c irq_set_msi_desc(entry->irq, NULL); entry 121 arch/powerpc/platforms/pseries/msi.c irq_dispose_mapping(entry->irq); entry 202 arch/powerpc/platforms/pseries/msi.c entry); entry 334 arch/powerpc/platforms/pseries/msi.c struct msi_desc *entry; entry 342 arch/powerpc/platforms/pseries/msi.c for_each_pci_msi_entry(entry, pdev) { entry 343 arch/powerpc/platforms/pseries/msi.c if (entry->msi_attrib.entry_nr != expected) { entry 374 arch/powerpc/platforms/pseries/msi.c struct msi_desc *entry; entry 454 arch/powerpc/platforms/pseries/msi.c for_each_pci_msi_entry(entry, pdev) { entry 469 arch/powerpc/platforms/pseries/msi.c irq_set_msi_desc(virq, entry); entry 472 arch/powerpc/platforms/pseries/msi.c __pci_read_msi_msg(entry, &msg); entry 473 arch/powerpc/platforms/pseries/msi.c entry->msg = msg; entry 124 arch/powerpc/sysdev/fsl_msi.c struct msi_desc *entry; entry 128 arch/powerpc/sysdev/fsl_msi.c for_each_pci_msi_entry(entry, pdev) { entry 129 arch/powerpc/sysdev/fsl_msi.c if (!entry->irq) entry 131 arch/powerpc/sysdev/fsl_msi.c hwirq = virq_to_hw(entry->irq); entry 132 arch/powerpc/sysdev/fsl_msi.c msi_data = irq_get_chip_data(entry->irq); entry 133 arch/powerpc/sysdev/fsl_msi.c irq_set_msi_desc(entry->irq, NULL); entry 134 arch/powerpc/sysdev/fsl_msi.c irq_dispose_mapping(entry->irq); entry 185 arch/powerpc/sysdev/fsl_msi.c struct msi_desc *entry; entry 218 arch/powerpc/sysdev/fsl_msi.c for_each_pci_msi_entry(entry, pdev) { entry 255 arch/powerpc/sysdev/fsl_msi.c irq_set_msi_desc(virq, entry); entry 96 arch/powerpc/sysdev/fsl_rio.c const struct exception_table_entry *entry; entry 105 arch/powerpc/sysdev/fsl_rio.c entry = search_exception_tables(regs->nip); entry 106 arch/powerpc/sysdev/fsl_rio.c if (entry) { entry 112 arch/powerpc/sysdev/fsl_rio.c regs->nip = extable_fixup(entry); entry 104 arch/powerpc/sysdev/mpic_u3msi.c struct msi_desc *entry; entry 107 arch/powerpc/sysdev/mpic_u3msi.c for_each_pci_msi_entry(entry, pdev) { entry 108 arch/powerpc/sysdev/mpic_u3msi.c if (!entry->irq) entry 111 arch/powerpc/sysdev/mpic_u3msi.c hwirq = virq_to_hw(entry->irq); entry 112 arch/powerpc/sysdev/mpic_u3msi.c irq_set_msi_desc(entry->irq, NULL); entry 113 arch/powerpc/sysdev/mpic_u3msi.c irq_dispose_mapping(entry->irq); entry 123 arch/powerpc/sysdev/mpic_u3msi.c struct msi_desc *entry; entry 139 arch/powerpc/sysdev/mpic_u3msi.c for_each_pci_msi_entry(entry, pdev) { entry 159 arch/powerpc/sysdev/mpic_u3msi.c irq_set_msi_desc(virq, entry); entry 3105 arch/powerpc/xmon/xmon.c pte_t entry = __pte(pte); entry 3111 arch/powerpc/xmon/xmon.c pte_young(entry) ? "Accessed " : "", entry 3112 arch/powerpc/xmon/xmon.c pte_dirty(entry) ? "Dirty " : "", entry 3113 arch/powerpc/xmon/xmon.c pte_read(entry) ? "Read " : "", entry 3114 arch/powerpc/xmon/xmon.c pte_write(entry) ? "Write " : "", entry 3115 arch/powerpc/xmon/xmon.c pte_exec(entry) ? "Exec " : ""); entry 351 arch/riscv/include/asm/pgtable.h pte_t entry, int dirty) entry 353 arch/riscv/include/asm/pgtable.h if (!pte_same(*ptep, entry)) entry 354 arch/riscv/include/asm/pgtable.h set_pte_at(vma->vm_mm, address, ptep, entry); entry 17 arch/riscv/kernel/perf_callchain.c static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry, entry 39 arch/riscv/kernel/perf_callchain.c perf_callchain_store(entry, ra); entry 60 arch/riscv/kernel/perf_callchain.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry, entry 70 arch/riscv/kernel/perf_callchain.c perf_callchain_store(entry, regs->sepc); entry 72 arch/riscv/kernel/perf_callchain.c fp = user_backtrace(entry, fp, regs->ra); entry 73 arch/riscv/kernel/perf_callchain.c while (fp && !(fp & 0x3) && entry->nr < entry->max_stack) entry 74 arch/riscv/kernel/perf_callchain.c fp = user_backtrace(entry, fp, 0); entry 77 arch/riscv/kernel/perf_callchain.c bool fill_callchain(unsigned long pc, void *entry) entry 79 arch/riscv/kernel/perf_callchain.c return perf_callchain_store(entry, pc); entry 84 arch/riscv/kernel/perf_callchain.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 93 arch/riscv/kernel/perf_callchain.c walk_stackframe(NULL, regs, fill_callchain, entry); entry 15 arch/s390/boot/compressed/decompressor.h void (*entry)(void); entry 19 arch/s390/boot/ipl_report.c #define for_each_rb_entry(entry, rb) \ entry 20 arch/s390/boot/ipl_report.c for (entry = rb->entries; \ entry 21 arch/s390/boot/ipl_report.c (void *) entry + sizeof(*entry) <= (void *) rb + rb->len; \ entry 22 arch/s390/boot/ipl_report.c entry++) entry 40 arch/s390/boot/startup.c void _diag0c_dma(struct hypfs_diag0c_entry *entry); entry 152 arch/s390/boot/startup.c vmlinux.entry += __kaslr_offset; entry 182 arch/s390/boot/startup.c vmlinux.entry(); entry 43 arch/s390/hypfs/hypfs_diag0c.c diag0c_data = kzalloc(struct_size(diag0c_data, entry, cpu_count), entry 50 arch/s390/hypfs/hypfs_diag0c.c diag0c_data->entry[i].cpu = cpu; entry 51 arch/s390/hypfs/hypfs_diag0c.c cpu_vec[cpu] = &diag0c_data->entry[i++]; entry 26 arch/s390/include/asm/debug.h #define DEBUG_DATA(entry) (char *)(entry + 1) /* data is stored behind */ entry 57 arch/s390/include/asm/debug.h debug_entry_t *entry, entry 73 arch/s390/include/asm/debug.h int area, debug_entry_t *entry, char *out_buf); entry 318 arch/s390/include/asm/diag.h void (*diag0c)(struct hypfs_diag0c_entry *entry); entry 95 arch/s390/include/asm/ipl.h struct ipl_rb_component_entry entry; entry 100 arch/s390/include/asm/ipl.h struct ipl_rb_certificate_entry entry; entry 31 arch/s390/include/asm/os_info.h struct os_info_entry entry[2]; entry 101 arch/s390/include/asm/pci.h struct list_head entry; /* list of all zpci_devices, needed for hotplug, etc. */ entry 100 arch/s390/include/asm/pci_dma.h static inline void set_pt_pfaa(unsigned long *entry, void *pfaa) entry 102 arch/s390/include/asm/pci_dma.h *entry &= ZPCI_PTE_FLAG_MASK; entry 103 arch/s390/include/asm/pci_dma.h *entry |= ((unsigned long) pfaa & ZPCI_PTE_ADDR_MASK); entry 106 arch/s390/include/asm/pci_dma.h static inline void set_rt_sto(unsigned long *entry, void *sto) entry 108 arch/s390/include/asm/pci_dma.h *entry &= ZPCI_RTE_FLAG_MASK; entry 109 arch/s390/include/asm/pci_dma.h *entry |= ((unsigned long) sto & ZPCI_RTE_ADDR_MASK); entry 110 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_TABLE_TYPE_RTX; entry 113 arch/s390/include/asm/pci_dma.h static inline void set_st_pto(unsigned long *entry, void *pto) entry 115 arch/s390/include/asm/pci_dma.h *entry &= ZPCI_STE_FLAG_MASK; entry 116 arch/s390/include/asm/pci_dma.h *entry |= ((unsigned long) pto & ZPCI_STE_ADDR_MASK); entry 117 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_TABLE_TYPE_SX; entry 120 arch/s390/include/asm/pci_dma.h static inline void validate_rt_entry(unsigned long *entry) entry 122 arch/s390/include/asm/pci_dma.h *entry &= ~ZPCI_TABLE_VALID_MASK; entry 123 arch/s390/include/asm/pci_dma.h *entry &= ~ZPCI_TABLE_OFFSET_MASK; entry 124 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_TABLE_VALID; entry 125 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_TABLE_LEN_RTX; entry 128 arch/s390/include/asm/pci_dma.h static inline void validate_st_entry(unsigned long *entry) entry 130 arch/s390/include/asm/pci_dma.h *entry &= ~ZPCI_TABLE_VALID_MASK; entry 131 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_TABLE_VALID; entry 134 arch/s390/include/asm/pci_dma.h static inline void invalidate_table_entry(unsigned long *entry) entry 136 arch/s390/include/asm/pci_dma.h *entry &= ~ZPCI_TABLE_VALID_MASK; entry 137 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_TABLE_INVALID; entry 140 arch/s390/include/asm/pci_dma.h static inline void invalidate_pt_entry(unsigned long *entry) entry 142 arch/s390/include/asm/pci_dma.h WARN_ON_ONCE((*entry & ZPCI_PTE_VALID_MASK) == ZPCI_PTE_INVALID); entry 143 arch/s390/include/asm/pci_dma.h *entry &= ~ZPCI_PTE_VALID_MASK; entry 144 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_PTE_INVALID; entry 147 arch/s390/include/asm/pci_dma.h static inline void validate_pt_entry(unsigned long *entry) entry 149 arch/s390/include/asm/pci_dma.h WARN_ON_ONCE((*entry & ZPCI_PTE_VALID_MASK) == ZPCI_PTE_VALID); entry 150 arch/s390/include/asm/pci_dma.h *entry &= ~ZPCI_PTE_VALID_MASK; entry 151 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_PTE_VALID; entry 154 arch/s390/include/asm/pci_dma.h static inline void entry_set_protected(unsigned long *entry) entry 156 arch/s390/include/asm/pci_dma.h *entry &= ~ZPCI_TABLE_PROT_MASK; entry 157 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_TABLE_PROTECTED; entry 160 arch/s390/include/asm/pci_dma.h static inline void entry_clr_protected(unsigned long *entry) entry 162 arch/s390/include/asm/pci_dma.h *entry &= ~ZPCI_TABLE_PROT_MASK; entry 163 arch/s390/include/asm/pci_dma.h *entry |= ZPCI_TABLE_UNPROTECTED; entry 166 arch/s390/include/asm/pci_dma.h static inline int reg_entry_isvalid(unsigned long entry) entry 168 arch/s390/include/asm/pci_dma.h return (entry & ZPCI_TABLE_VALID_MASK) == ZPCI_TABLE_VALID; entry 171 arch/s390/include/asm/pci_dma.h static inline int pt_entry_isvalid(unsigned long entry) entry 173 arch/s390/include/asm/pci_dma.h return (entry & ZPCI_PTE_VALID_MASK) == ZPCI_PTE_VALID; entry 176 arch/s390/include/asm/pci_dma.h static inline int entry_isprotected(unsigned long entry) entry 178 arch/s390/include/asm/pci_dma.h return (entry & ZPCI_TABLE_PROT_MASK) == ZPCI_TABLE_PROTECTED; entry 181 arch/s390/include/asm/pci_dma.h static inline unsigned long *get_rt_sto(unsigned long entry) entry 183 arch/s390/include/asm/pci_dma.h return ((entry & ZPCI_TABLE_TYPE_MASK) == ZPCI_TABLE_TYPE_RTX) entry 184 arch/s390/include/asm/pci_dma.h ? (unsigned long *) (entry & ZPCI_RTE_ADDR_MASK) entry 188 arch/s390/include/asm/pci_dma.h static inline unsigned long *get_st_pto(unsigned long entry) entry 190 arch/s390/include/asm/pci_dma.h return ((entry & ZPCI_TABLE_TYPE_MASK) == ZPCI_TABLE_TYPE_SX) entry 191 arch/s390/include/asm/pci_dma.h ? (unsigned long *) (entry & ZPCI_STE_ADDR_MASK) entry 202 arch/s390/include/asm/pci_dma.h void dma_update_cpu_trans(unsigned long *entry, void *page_addr, int flags); entry 32 arch/s390/include/asm/pgalloc.h static inline void crst_table_init(unsigned long *crst, unsigned long entry) entry 34 arch/s390/include/asm/pgalloc.h memset64((u64 *)crst, entry, _CRST_ENTRIES); entry 1128 arch/s390/include/asm/pgtable.h pte_t entry, int dirty) entry 1130 arch/s390/include/asm/pgtable.h if (pte_same(*ptep, entry)) entry 1132 arch/s390/include/asm/pgtable.h ptep_xchg_direct(vma->vm_mm, addr, ptep, entry); entry 1140 arch/s390/include/asm/pgtable.h pte_t *ptep, pte_t entry); entry 1180 arch/s390/include/asm/pgtable.h pte_t *ptep, pte_t entry) entry 1182 arch/s390/include/asm/pgtable.h if (pte_present(entry)) entry 1183 arch/s390/include/asm/pgtable.h pte_val(entry) &= ~_PAGE_UNUSED; entry 1185 arch/s390/include/asm/pgtable.h ptep_set_pte_at(mm, addr, ptep, entry); entry 1187 arch/s390/include/asm/pgtable.h *ptep = entry; entry 1514 arch/s390/include/asm/pgtable.h pmd_t entry, int dirty) entry 1518 arch/s390/include/asm/pgtable.h entry = pmd_mkyoung(entry); entry 1520 arch/s390/include/asm/pgtable.h entry = pmd_mkdirty(entry); entry 1521 arch/s390/include/asm/pgtable.h if (pmd_val(*pmdp) == pmd_val(entry)) entry 1523 arch/s390/include/asm/pgtable.h pmdp_xchg_direct(vma->vm_mm, addr, pmdp, entry); entry 1546 arch/s390/include/asm/pgtable.h pmd_t *pmdp, pmd_t entry) entry 1549 arch/s390/include/asm/pgtable.h pmd_val(entry) &= ~_SEGMENT_ENTRY_NOEXEC; entry 1550 arch/s390/include/asm/pgtable.h *pmdp = entry; entry 1662 arch/s390/include/asm/pgtable.h static inline unsigned long __swp_type(swp_entry_t entry) entry 1664 arch/s390/include/asm/pgtable.h return (entry.val >> __SWP_TYPE_SHIFT) & __SWP_TYPE_MASK; entry 1667 arch/s390/include/asm/pgtable.h static inline unsigned long __swp_offset(swp_entry_t entry) entry 1669 arch/s390/include/asm/pgtable.h return (entry.val >> __SWP_OFFSET_SHIFT) & __SWP_OFFSET_MASK; entry 429 arch/s390/include/asm/qdio.h void *entry), entry 15 arch/s390/include/asm/vtimer.h struct list_head entry; entry 52 arch/s390/include/uapi/asm/hypfs.h struct hypfs_diag0c_entry entry[]; /* diag0c entry array */ entry 97 arch/s390/kernel/debug.c int area, debug_entry_t *entry, char *out_buf); entry 1391 arch/s390/kernel/debug.c int area, debug_entry_t *entry, char *out_buf) entry 1396 arch/s390/kernel/debug.c memcpy(out_buf, entry, sizeof(debug_entry_t)); entry 1440 arch/s390/kernel/debug.c int area, debug_entry_t *entry, char *out_buf) entry 1448 arch/s390/kernel/debug.c level = entry->id.fields.level; entry 1450 arch/s390/kernel/debug.c sec = (entry->id.stck >> 12) + base - (TOD_UNIX_EPOCH >> 12); entry 1453 arch/s390/kernel/debug.c if (entry->id.fields.exception) entry 1457 arch/s390/kernel/debug.c caller = (unsigned long) entry->caller; entry 1460 arch/s390/kernel/debug.c entry->id.fields.cpuid, (void *)caller); entry 394 arch/s390/kernel/dis.c struct s390_opcode_offset *entry; entry 405 arch/s390/kernel/dis.c entry = &opcode_offset[i]; entry 406 arch/s390/kernel/dis.c if (entry->opcode == code[0]) entry 410 arch/s390/kernel/dis.c opfrag = *(code + entry->byte) & entry->mask; entry 412 arch/s390/kernel/dis.c insn = &opcode[entry->offset]; entry 413 arch/s390/kernel/dis.c for (i = 0; i < entry->count; i++) { entry 1724 arch/s390/kernel/ipl.c comp->entry.addr = kbuf->mem; entry 1725 arch/s390/kernel/ipl.c comp->entry.len = kbuf->memsz; entry 1726 arch/s390/kernel/ipl.c comp->entry.flags = flags; entry 1727 arch/s390/kernel/ipl.c comp->entry.certificate_index = cert; entry 1729 arch/s390/kernel/ipl.c report->size += sizeof(comp->entry); entry 1744 arch/s390/kernel/ipl.c cert->entry.addr = addr; entry 1745 arch/s390/kernel/ipl.c cert->entry.len = len; entry 1748 arch/s390/kernel/ipl.c report->size += sizeof(cert->entry); entry 1749 arch/s390/kernel/ipl.c report->size += cert->entry.len; entry 1804 arch/s390/kernel/ipl.c memcpy(ptr, &comp->entry, sizeof(comp->entry)); entry 1805 arch/s390/kernel/ipl.c ptr += sizeof(comp->entry); entry 1813 arch/s390/kernel/ipl.c memcpy(ptr, &cert->entry, sizeof(cert->entry)); entry 1814 arch/s390/kernel/ipl.c ptr += sizeof(cert->entry); entry 1820 arch/s390/kernel/ipl.c memcpy(ptr, cert->key, cert->entry.len); entry 1821 arch/s390/kernel/ipl.c ptr += cert->entry.len; entry 222 arch/s390/kernel/irq.c struct hlist_node entry; entry 251 arch/s390/kernel/irq.c hlist_add_head_rcu(&p->entry, &ext_int_hash[index]); entry 264 arch/s390/kernel/irq.c hlist_for_each_entry_rcu(p, &ext_int_hash[index], entry) { entry 266 arch/s390/kernel/irq.c hlist_del_rcu(&p->entry); entry 288 arch/s390/kernel/irq.c hlist_for_each_entry_rcu(p, &ext_int_hash[index], entry) { entry 18 arch/s390/kernel/jump_label.c static void jump_label_make_nop(struct jump_entry *entry, struct insn *insn) entry 22 arch/s390/kernel/jump_label.c insn->offset = (jump_entry_target(entry) - jump_entry_code(entry)) >> 1; entry 25 arch/s390/kernel/jump_label.c static void jump_label_make_branch(struct jump_entry *entry, struct insn *insn) entry 29 arch/s390/kernel/jump_label.c insn->offset = (jump_entry_target(entry) - jump_entry_code(entry)) >> 1; entry 32 arch/s390/kernel/jump_label.c static void jump_label_bug(struct jump_entry *entry, struct insn *expected, entry 35 arch/s390/kernel/jump_label.c unsigned char *ipc = (unsigned char *)jump_entry_code(entry); entry 51 arch/s390/kernel/jump_label.c static void __jump_label_transform(struct jump_entry *entry, entry 55 arch/s390/kernel/jump_label.c void *code = (void *)jump_entry_code(entry); entry 59 arch/s390/kernel/jump_label.c jump_label_make_nop(entry, &old); entry 60 arch/s390/kernel/jump_label.c jump_label_make_branch(entry, &new); entry 62 arch/s390/kernel/jump_label.c jump_label_make_branch(entry, &old); entry 63 arch/s390/kernel/jump_label.c jump_label_make_nop(entry, &new); entry 67 arch/s390/kernel/jump_label.c jump_label_bug(entry, &orignop, &new); entry 70 arch/s390/kernel/jump_label.c jump_label_bug(entry, &old, &new); entry 79 arch/s390/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry, entry 82 arch/s390/kernel/jump_label.c __jump_label_transform(entry, type, 0); entry 86 arch/s390/kernel/jump_label.c void arch_jump_label_transform_static(struct jump_entry *entry, entry 89 arch/s390/kernel/jump_label.c __jump_label_transform(entry, type, 1); entry 22 arch/s390/kernel/kexec_elf.c Elf_Addr entry; entry 30 arch/s390/kernel/kexec_elf.c entry = STARTUP_KDUMP_OFFSET; entry 32 arch/s390/kernel/kexec_elf.c entry = ehdr->e_entry; entry 48 arch/s390/kernel/kexec_elf.c if (entry - phdr->p_paddr < phdr->p_memsz) { entry 537 arch/s390/kernel/kprobes.c const struct exception_table_entry *entry; entry 576 arch/s390/kernel/kprobes.c entry = s390_search_extables(regs->psw.addr); entry 577 arch/s390/kernel/kprobes.c if (entry) { entry 578 arch/s390/kernel/kprobes.c regs->psw.addr = extable_fixup(entry); entry 76 arch/s390/kernel/machine_kexec_file.c u64 entry, type; entry 80 arch/s390/kernel/machine_kexec_file.c entry = STARTUP_KDUMP_OFFSET; entry 83 arch/s390/kernel/machine_kexec_file.c entry = STARTUP_NORMAL_OFFSET; entry 87 arch/s390/kernel/machine_kexec_file.c ret = kexec_purgatory_get_set_symbol(image, "kernel_entry", &entry, entry 88 arch/s390/kernel/machine_kexec_file.c sizeof(entry), false); entry 48 arch/s390/kernel/os_info.c os_info.entry[nr].addr = (u64)(unsigned long)ptr; entry 49 arch/s390/kernel/os_info.c os_info.entry[nr].size = size; entry 50 arch/s390/kernel/os_info.c os_info.entry[nr].csum = (__force u32)csum_partial(ptr, size, 0); entry 81 arch/s390/kernel/os_info.c addr = os_info_old->entry[nr].addr; entry 86 arch/s390/kernel/os_info.c size = os_info_old->entry[nr].size; entry 98 arch/s390/kernel/os_info.c if (csum != os_info_old->entry[nr].csum) { entry 102 arch/s390/kernel/os_info.c os_info_old->entry[nr].addr = (u64)(unsigned long)buf_align; entry 108 arch/s390/kernel/os_info.c os_info_old->entry[nr].addr = 0; entry 165 arch/s390/kernel/os_info.c if (!os_info_old->entry[nr].addr) entry 167 arch/s390/kernel/os_info.c *size = (unsigned long) os_info_old->entry[nr].size; entry 168 arch/s390/kernel/os_info.c return (void *)(unsigned long)os_info_old->entry[nr].addr; entry 223 arch/s390/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 231 arch/s390/kernel/perf_event.c if (!addr || perf_callchain_store(entry, addr)) entry 262 arch/s390/kernel/vtime.c list_for_each_entry(tmp, head, entry) { entry 264 arch/s390/kernel/vtime.c list_add_tail(&timer->entry, &tmp->entry); entry 268 arch/s390/kernel/vtime.c list_add_tail(&timer->entry, head); entry 283 arch/s390/kernel/vtime.c list_for_each_entry_safe(timer, tmp, &virt_timer_list, entry) { entry 286 arch/s390/kernel/vtime.c list_move_tail(&timer->entry, &cb_list); entry 292 arch/s390/kernel/vtime.c struct vtimer_list, entry); entry 299 arch/s390/kernel/vtime.c list_for_each_entry_safe(timer, tmp, &cb_list, entry) { entry 300 arch/s390/kernel/vtime.c list_del_init(&timer->entry); entry 316 arch/s390/kernel/vtime.c INIT_LIST_HEAD(&timer->entry); entry 322 arch/s390/kernel/vtime.c return !list_empty(&timer->entry); entry 331 arch/s390/kernel/vtime.c list_add(&timer->entry, &virt_timer_list); entry 384 arch/s390/kernel/vtime.c list_del_init(&timer->entry); entry 422 arch/s390/kernel/vtime.c list_del_init(&timer->entry); entry 337 arch/s390/mm/gmap.c static unsigned long __gmap_segment_gaddr(unsigned long *entry) entry 342 arch/s390/mm/gmap.c offset = (unsigned long) entry / sizeof(unsigned long); entry 345 arch/s390/mm/gmap.c page = virt_to_page((void *)((unsigned long) entry & mask)); entry 358 arch/s390/mm/gmap.c unsigned long *entry; entry 363 arch/s390/mm/gmap.c entry = radix_tree_delete(&gmap->host_to_guest, vmaddr >> PMD_SHIFT); entry 364 arch/s390/mm/gmap.c if (entry) { entry 365 arch/s390/mm/gmap.c flush = (*entry != _SEGMENT_ENTRY_EMPTY); entry 366 arch/s390/mm/gmap.c *entry = _SEGMENT_ENTRY_EMPTY; entry 2358 arch/s390/mm/gmap.c unsigned long *entry, gaddr; entry 2365 arch/s390/mm/gmap.c entry = radix_tree_delete(&gmap->host_to_guest, entry 2367 arch/s390/mm/gmap.c if (entry) { entry 2368 arch/s390/mm/gmap.c pmdp = (pmd_t *)entry; entry 2369 arch/s390/mm/gmap.c gaddr = __gmap_segment_gaddr(entry); entry 2371 arch/s390/mm/gmap.c WARN_ON(*entry & ~(_SEGMENT_ENTRY_HARDWARE_BITS_LARGE | entry 2378 arch/s390/mm/gmap.c *entry = _SEGMENT_ENTRY_EMPTY; entry 2393 arch/s390/mm/gmap.c unsigned long *entry, gaddr; entry 2400 arch/s390/mm/gmap.c entry = radix_tree_delete(&gmap->host_to_guest, entry 2402 arch/s390/mm/gmap.c if (entry) { entry 2403 arch/s390/mm/gmap.c pmdp = (pmd_t *)entry; entry 2404 arch/s390/mm/gmap.c gaddr = __gmap_segment_gaddr(entry); entry 2406 arch/s390/mm/gmap.c WARN_ON(*entry & ~(_SEGMENT_ENTRY_HARDWARE_BITS_LARGE | entry 2415 arch/s390/mm/gmap.c *entry = _SEGMENT_ENTRY_EMPTY; entry 176 arch/s390/mm/pgtable.c static inline void pgste_set_key(pte_t *ptep, pgste_t pgste, pte_t entry, entry 183 arch/s390/mm/pgtable.c if (!mm_uses_skeys(mm) || pte_val(entry) & _PAGE_INVALID) entry 186 arch/s390/mm/pgtable.c address = pte_val(entry) & PAGE_MASK; entry 198 arch/s390/mm/pgtable.c static inline pgste_t pgste_set_pte(pte_t *ptep, pgste_t pgste, pte_t entry) entry 201 arch/s390/mm/pgtable.c if ((pte_val(entry) & _PAGE_PRESENT) && entry 202 arch/s390/mm/pgtable.c (pte_val(entry) & _PAGE_WRITE) && entry 203 arch/s390/mm/pgtable.c !(pte_val(entry) & _PAGE_INVALID)) { entry 209 arch/s390/mm/pgtable.c pte_val(entry) |= _PAGE_DIRTY; entry 210 arch/s390/mm/pgtable.c pte_val(entry) &= ~_PAGE_PROTECT; entry 212 arch/s390/mm/pgtable.c if (!(pte_val(entry) & _PAGE_PROTECT)) entry 217 arch/s390/mm/pgtable.c *ptep = entry; entry 559 arch/s390/mm/pgtable.c pte_t *ptep, pte_t entry) entry 567 arch/s390/mm/pgtable.c pgste_set_key(ptep, pgste, entry, mm); entry 568 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, entry); entry 598 arch/s390/mm/pgtable.c pte_t entry; entry 603 arch/s390/mm/pgtable.c entry = *ptep; entry 605 arch/s390/mm/pgtable.c pte_i = pte_val(entry) & _PAGE_INVALID; entry 606 arch/s390/mm/pgtable.c pte_p = pte_val(entry) & _PAGE_PROTECT; entry 616 arch/s390/mm/pgtable.c pgste = pgste_update_all(entry, pgste, mm); entry 617 arch/s390/mm/pgtable.c pte_val(entry) |= _PAGE_INVALID; entry 621 arch/s390/mm/pgtable.c pte_val(entry) &= ~_PAGE_INVALID; entry 622 arch/s390/mm/pgtable.c pte_val(entry) |= _PAGE_PROTECT; entry 625 arch/s390/mm/pgtable.c pgste = pgste_set_pte(ptep, pgste, entry); entry 671 arch/s390/mm/pgtable.c static void ptep_zap_swap_entry(struct mm_struct *mm, swp_entry_t entry) entry 673 arch/s390/mm/pgtable.c if (!non_swap_entry(entry)) entry 675 arch/s390/mm/pgtable.c else if (is_migration_entry(entry)) { entry 676 arch/s390/mm/pgtable.c struct page *page = migration_entry_to_page(entry); entry 680 arch/s390/mm/pgtable.c free_swap_and_cache(entry); entry 63 arch/s390/pci/pci.c list_for_each_entry(tmp, &zpci_list, entry) { entry 80 arch/s390/pci/pci.c list_for_each_entry_safe(zdev, tmp, &zpci_list, entry) { entry 84 arch/s390/pci/pci.c list_move_tail(&zdev->entry, &remove); entry 88 arch/s390/pci/pci.c list_for_each_entry_safe(zdev, tmp, &remove, entry) entry 465 arch/s390/pci/pci.c unsigned long entry; entry 468 arch/s390/pci/pci.c entry = find_first_zero_bit(zpci_iomap_bitmap, ZPCI_IOMAP_ENTRIES); entry 469 arch/s390/pci/pci.c if (entry == ZPCI_IOMAP_ENTRIES) { entry 473 arch/s390/pci/pci.c set_bit(entry, zpci_iomap_bitmap); entry 475 arch/s390/pci/pci.c return entry; entry 478 arch/s390/pci/pci.c static void zpci_free_iomap(struct zpci_dev *zdev, int entry) entry 481 arch/s390/pci/pci.c memset(&zpci_iomap_start[entry], 0, sizeof(struct zpci_iomap_entry)); entry 482 arch/s390/pci/pci.c clear_bit(entry, zpci_iomap_bitmap); entry 512 arch/s390/pci/pci.c int i, entry; entry 520 arch/s390/pci/pci.c entry = zpci_alloc_iomap(zdev); entry 521 arch/s390/pci/pci.c if (entry < 0) entry 522 arch/s390/pci/pci.c return entry; entry 523 arch/s390/pci/pci.c zdev->bars[i].map_idx = entry; entry 535 arch/s390/pci/pci.c addr = ZPCI_ADDR(entry); entry 540 arch/s390/pci/pci.c zpci_free_iomap(zdev, entry); entry 700 arch/s390/pci/pci.c list_del(&zdev->entry); entry 784 arch/s390/pci/pci.c list_add_tail(&zdev->entry, &zpci_list); entry 360 arch/s390/pci/pci_clp.c static void __clp_add(struct clp_fh_list_entry *entry, void *data) entry 364 arch/s390/pci/pci_clp.c if (!entry->vendor_id) entry 367 arch/s390/pci/pci_clp.c zdev = get_zdev_by_fid(entry->fid); entry 369 arch/s390/pci/pci_clp.c clp_add_pci_device(entry->fid, entry->fh, entry->config_state); entry 372 arch/s390/pci/pci_clp.c static void __clp_update(struct clp_fh_list_entry *entry, void *data) entry 377 arch/s390/pci/pci_clp.c if (!entry->vendor_id) entry 380 arch/s390/pci/pci_clp.c if (fid && *fid != entry->fid) entry 383 arch/s390/pci/pci_clp.c zdev = get_zdev_by_fid(entry->fid); entry 387 arch/s390/pci/pci_clp.c zdev->fh = entry->fh; entry 445 arch/s390/pci/pci_clp.c static void __clp_get_state(struct clp_fh_list_entry *entry, void *data) entry 449 arch/s390/pci/pci_clp.c if (entry->fid != sd->fid) entry 452 arch/s390/pci/pci_clp.c sd->state = entry->config_state; entry 30 arch/s390/pci/pci_dma.c unsigned long *table, *entry; entry 36 arch/s390/pci/pci_dma.c for (entry = table; entry < table + ZPCI_TABLE_ENTRIES; entry++) entry 37 arch/s390/pci/pci_dma.c *entry = ZPCI_TABLE_INVALID; entry 48 arch/s390/pci/pci_dma.c unsigned long *table, *entry; entry 54 arch/s390/pci/pci_dma.c for (entry = table; entry < table + ZPCI_PT_ENTRIES; entry++) entry 55 arch/s390/pci/pci_dma.c *entry = ZPCI_PTE_INVALID; entry 64 arch/s390/pci/pci_dma.c static unsigned long *dma_get_seg_table_origin(unsigned long *entry) entry 68 arch/s390/pci/pci_dma.c if (reg_entry_isvalid(*entry)) entry 69 arch/s390/pci/pci_dma.c sto = get_rt_sto(*entry); entry 75 arch/s390/pci/pci_dma.c set_rt_sto(entry, sto); entry 76 arch/s390/pci/pci_dma.c validate_rt_entry(entry); entry 77 arch/s390/pci/pci_dma.c entry_clr_protected(entry); entry 82 arch/s390/pci/pci_dma.c static unsigned long *dma_get_page_table_origin(unsigned long *entry) entry 86 arch/s390/pci/pci_dma.c if (reg_entry_isvalid(*entry)) entry 87 arch/s390/pci/pci_dma.c pto = get_st_pto(*entry); entry 92 arch/s390/pci/pci_dma.c set_st_pto(entry, pto); entry 93 arch/s390/pci/pci_dma.c validate_st_entry(entry); entry 94 arch/s390/pci/pci_dma.c entry_clr_protected(entry); entry 118 arch/s390/pci/pci_dma.c void dma_update_cpu_trans(unsigned long *entry, void *page_addr, int flags) entry 121 arch/s390/pci/pci_dma.c invalidate_pt_entry(entry); entry 123 arch/s390/pci/pci_dma.c set_pt_pfaa(entry, page_addr); entry 124 arch/s390/pci/pci_dma.c validate_pt_entry(entry); entry 128 arch/s390/pci/pci_dma.c entry_set_protected(entry); entry 130 arch/s390/pci/pci_dma.c entry_clr_protected(entry); entry 139 arch/s390/pci/pci_dma.c unsigned long *entry; entry 152 arch/s390/pci/pci_dma.c entry = dma_walk_cpu_trans(zdev->dma_table, dma_addr); entry 153 arch/s390/pci/pci_dma.c if (!entry) { entry 157 arch/s390/pci/pci_dma.c dma_update_cpu_trans(entry, page_addr, flags); entry 168 arch/s390/pci/pci_dma.c entry = dma_walk_cpu_trans(zdev->dma_table, dma_addr); entry 169 arch/s390/pci/pci_dma.c if (!entry) entry 171 arch/s390/pci/pci_dma.c dma_update_cpu_trans(entry, page_addr, flags); entry 234 arch/s390/pci/pci_dma.c void dma_free_seg_table(unsigned long entry) entry 236 arch/s390/pci/pci_dma.c unsigned long *sto = get_rt_sto(entry); entry 168 arch/s390/pci/pci_insn.c struct zpci_iomap_entry *entry = &zpci_iomap_start[ZPCI_IDX(addr)]; entry 169 arch/s390/pci/pci_insn.c u64 req = ZPCI_CREATE_REQ(entry->fh, entry->bar, len); entry 252 arch/s390/pci/pci_insn.c struct zpci_iomap_entry *entry = &zpci_iomap_start[ZPCI_IDX(addr)]; entry 253 arch/s390/pci/pci_insn.c u64 req = ZPCI_CREATE_REQ(entry->fh, entry->bar, len); entry 332 arch/s390/pci/pci_insn.c struct zpci_iomap_entry *entry = &zpci_iomap_start[ZPCI_IDX(dst)]; entry 333 arch/s390/pci/pci_insn.c u64 req = ZPCI_CREATE_REQ(entry->fh, entry->bar, len); entry 104 arch/s390/pci/pci_irq.c struct msi_desc *entry = irq_get_msi_desc(data->irq); entry 105 arch/s390/pci/pci_irq.c struct msi_msg msg = entry->msg; entry 55 arch/sh/include/asm/tlb_64.h int sh64_put_wired_dtlb_entry(unsigned long long entry); entry 63 arch/sh/include/asm/tlb_64.h #define sh64_put_wired_dtlb_entry(entry) do { } while (0) entry 290 arch/sh/kernel/cpu/sh4/sq.c struct sq_mapping **list, *entry; entry 293 arch/sh/kernel/cpu/sh4/sq.c for (list = &sq_mapping_list; (entry = *list); list = &entry->next) entry 295 arch/sh/kernel/cpu/sh4/sq.c entry->sq_addr, entry->sq_addr + entry->size, entry 296 arch/sh/kernel/cpu/sh4/sq.c entry->addr, entry->name); entry 735 arch/sh/kernel/dwarf.c static int dwarf_parse_cie(void *entry, void *p, unsigned long len, entry 756 arch/sh/kernel/dwarf.c cie->cie_pointer = (unsigned long)entry; entry 865 arch/sh/kernel/dwarf.c static int dwarf_parse_fde(void *entry, u32 entry_type, entry 1028 arch/sh/kernel/dwarf.c void *p, *entry; entry 1036 arch/sh/kernel/dwarf.c entry = eh_frame_start; entry 1038 arch/sh/kernel/dwarf.c while ((char *)entry < eh_frame_end) { entry 1039 arch/sh/kernel/dwarf.c p = entry; entry 1062 arch/sh/kernel/dwarf.c err = dwarf_parse_cie(entry, p, len, end, mod); entry 1068 arch/sh/kernel/dwarf.c err = dwarf_parse_fde(entry, entry_type, p, len, entry 1076 arch/sh/kernel/dwarf.c entry = (char *)entry + len + 4; entry 142 arch/sh/kernel/io_trapped.c pte_t entry; entry 157 arch/sh/kernel/io_trapped.c entry = *pte_k; entry 159 arch/sh/kernel/io_trapped.c return pfn_to_kaddr(pte_pfn(entry)); entry 416 arch/sh/kernel/kprobes.c const struct exception_table_entry *entry; entry 458 arch/sh/kernel/kprobes.c if ((entry = search_exception_tables(regs->pc)) != NULL) { entry 459 arch/sh/kernel/kprobes.c regs->pc = entry->fixup; entry 76 arch/sh/kernel/machine_kexec.c unsigned long entry; entry 85 arch/sh/kernel/machine_kexec.c for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); entry 86 arch/sh/kernel/machine_kexec.c ptr = (entry & IND_INDIRECTION) ? entry 87 arch/sh/kernel/machine_kexec.c phys_to_virt(entry & PAGE_MASK) : ptr + 1) { entry 130 arch/sh/kernel/machine_kexec.c for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); entry 21 arch/sh/kernel/perf_callchain.c struct perf_callchain_entry_ctx *entry = data; entry 24 arch/sh/kernel/perf_callchain.c perf_callchain_store(entry, addr); entry 33 arch/sh/kernel/perf_callchain.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) entry 35 arch/sh/kernel/perf_callchain.c perf_callchain_store(entry, regs->pc); entry 37 arch/sh/kernel/perf_callchain.c unwind_stack(NULL, regs, NULL, &callchain_ops, entry); entry 85 arch/sh/kernel/unwinder.c struct list_head *tmp, *entry = &unwinder_list; entry 95 arch/sh/kernel/unwinder.c entry = tmp; entry 97 arch/sh/kernel/unwinder.c list_add(&ops->list, entry); entry 389 arch/sh/mm/cache-sh5.c pte_t entry; entry 410 arch/sh/mm/cache-sh5.c entry = *pte; entry 411 arch/sh/mm/cache-sh5.c if (pte_none(entry) || !pte_present(entry)) entry 413 arch/sh/mm/cache-sh5.c paddr = pte_val(entry) & PAGE_MASK; entry 48 arch/sh/mm/pmb.c int entry; entry 72 arch/sh/mm/pmb.c static __always_inline unsigned long mk_pmb_entry(unsigned int entry) entry 74 arch/sh/mm/pmb.c return (entry & PMB_E_MASK) << PMB_E_SHIFT; entry 77 arch/sh/mm/pmb.c static __always_inline unsigned long mk_pmb_addr(unsigned int entry) entry 79 arch/sh/mm/pmb.c return mk_pmb_entry(entry) | PMB_ADDR; entry 82 arch/sh/mm/pmb.c static __always_inline unsigned long mk_pmb_data(unsigned int entry) entry 84 arch/sh/mm/pmb.c return mk_pmb_entry(entry) | PMB_DATA; entry 237 arch/sh/mm/pmb.c unsigned long flags, int entry) entry 246 arch/sh/mm/pmb.c if (entry == PMB_NO_ENTRY) { entry 253 arch/sh/mm/pmb.c if (__test_and_set_bit(entry, pmb_map)) { entry 258 arch/sh/mm/pmb.c pos = entry; entry 272 arch/sh/mm/pmb.c pmbe->entry = pos; entry 283 arch/sh/mm/pmb.c __clear_bit(pmbe->entry, pmb_map); entry 285 arch/sh/mm/pmb.c pmbe->entry = PMB_NO_ENTRY; entry 296 arch/sh/mm/pmb.c addr = mk_pmb_addr(pmbe->entry); entry 297 arch/sh/mm/pmb.c data = mk_pmb_data(pmbe->entry); entry 313 arch/sh/mm/pmb.c addr = mk_pmb_addr(pmbe->entry); entry 314 arch/sh/mm/pmb.c data = mk_pmb_data(pmbe->entry); entry 45 arch/sh/mm/tlb-debugfs.c unsigned int nentries, entry; entry 93 arch/sh/mm/tlb-debugfs.c for (entry = 0; entry < nentries; entry++) { entry 100 arch/sh/mm/tlb-debugfs.c val = __raw_readl(addr1 | (entry << MMU_TLB_ENTRY_SHIFT)); entry 105 arch/sh/mm/tlb-debugfs.c val = __raw_readl(addr2 | (entry << MMU_TLB_ENTRY_SHIFT)); entry 109 arch/sh/mm/tlb-debugfs.c val = __raw_readl(data1 | (entry << MMU_TLB_ENTRY_SHIFT)); entry 113 arch/sh/mm/tlb-debugfs.c val = __raw_readl(data2 | (entry << MMU_TLB_ENTRY_SHIFT)); entry 126 arch/sh/mm/tlb-debugfs.c entry, vpn, ppn, asid, entry 128 arch/sh/mm/tlb-debugfs.c (urb <= entry) ? "W" : "-"); entry 59 arch/sh/mm/tlb-sh5.c unsigned long long entry = sh64_next_free_dtlb_entry(); entry 64 arch/sh/mm/tlb-sh5.c return entry; entry 74 arch/sh/mm/tlb-sh5.c int sh64_put_wired_dtlb_entry(unsigned long long entry) entry 76 arch/sh/mm/tlb-sh5.c __flush_tlb_slot(entry); entry 92 arch/sh/mm/tlb-sh5.c if (entry <= DTLB_FIXED) entry 99 arch/sh/mm/tlb-sh5.c if (entry < (cpu_data->dtlb.first - cpu_data->dtlb.step)) entry 104 arch/sh/mm/tlb-sh5.c cpu_data->dtlb.next = entry; entry 152 arch/sh/mm/tlb-sh5.c unsigned long long entry; entry 159 arch/sh/mm/tlb-sh5.c entry = sh64_get_wired_dtlb_entry(); entry 160 arch/sh/mm/tlb-sh5.c dtlb_entries[dtlb_entry++] = entry; entry 165 arch/sh/mm/tlb-sh5.c sh64_setup_tlb_slot(entry, addr, get_asid(), paddr); entry 172 arch/sh/mm/tlb-sh5.c unsigned long long entry; entry 178 arch/sh/mm/tlb-sh5.c entry = dtlb_entries[dtlb_entry--]; entry 180 arch/sh/mm/tlb-sh5.c sh64_teardown_tlb_slot(entry); entry 181 arch/sh/mm/tlb-sh5.c sh64_put_wired_dtlb_entry(entry); entry 29 arch/sh/mm/tlbex_32.c pte_t entry; entry 52 arch/sh/mm/tlbex_32.c entry = *pte; entry 53 arch/sh/mm/tlbex_32.c if (unlikely(pte_none(entry) || pte_not_present(entry))) entry 55 arch/sh/mm/tlbex_32.c if (unlikely(error_code && !pte_write(entry))) entry 59 arch/sh/mm/tlbex_32.c entry = pte_mkdirty(entry); entry 60 arch/sh/mm/tlbex_32.c entry = pte_mkyoung(entry); entry 62 arch/sh/mm/tlbex_32.c set_pte(pte, entry); entry 50 arch/sh/mm/tlbex_64.c pte_t entry; entry 70 arch/sh/mm/tlbex_64.c entry = *pte; entry 71 arch/sh/mm/tlbex_64.c if (pte_none(entry) || !pte_present(entry)) entry 80 arch/sh/mm/tlbex_64.c if ((pte_val(entry) & protection_flags) != protection_flags) entry 51 arch/sparc/include/asm/iommu-common.h unsigned long entry); entry 352 arch/sparc/include/asm/pgtable_32.h static inline unsigned long __swp_type(swp_entry_t entry) entry 354 arch/sparc/include/asm/pgtable_32.h return (entry.val >> SRMMU_SWP_TYPE_SHIFT) & SRMMU_SWP_TYPE_MASK; entry 357 arch/sparc/include/asm/pgtable_32.h static inline unsigned long __swp_offset(swp_entry_t entry) entry 359 arch/sparc/include/asm/pgtable_32.h return (entry.val >> SRMMU_SWP_OFF_SHIFT) & SRMMU_SWP_OFF_MASK; entry 380 arch/sparc/include/asm/pgtable_64.h extern pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma, entry 786 arch/sparc/include/asm/pgtable_64.h static inline pgprot_t pmd_pgprot(pmd_t entry) entry 788 arch/sparc/include/asm/pgtable_64.h unsigned long val = pmd_val(entry); entry 1010 arch/sparc/include/asm/pgtable_64.h #define __swp_type(entry) (((entry).val >> PAGE_SHIFT) & 0xffUL) entry 1011 arch/sparc/include/asm/pgtable_64.h #define __swp_offset(entry) ((entry).val >> (PAGE_SHIFT + 8UL)) entry 169 arch/sparc/include/asm/pgtsrmmu.h register unsigned long entry; entry 172 arch/sparc/include/asm/pgtsrmmu.h "=r" (entry): entry 174 arch/sparc/include/asm/pgtsrmmu.h return entry; entry 119 arch/sparc/include/asm/spitfire.h static inline unsigned long spitfire_get_dtlb_data(int entry) entry 125 arch/sparc/include/asm/spitfire.h : "r" (entry << 3), "i" (ASI_DTLB_DATA_ACCESS)); entry 133 arch/sparc/include/asm/spitfire.h static inline unsigned long spitfire_get_dtlb_tag(int entry) entry 139 arch/sparc/include/asm/spitfire.h : "r" (entry << 3), "i" (ASI_DTLB_TAG_READ)); entry 143 arch/sparc/include/asm/spitfire.h static inline void spitfire_put_dtlb_data(int entry, unsigned long data) entry 148 arch/sparc/include/asm/spitfire.h : "r" (data), "r" (entry << 3), entry 152 arch/sparc/include/asm/spitfire.h static inline unsigned long spitfire_get_itlb_data(int entry) entry 158 arch/sparc/include/asm/spitfire.h : "r" (entry << 3), "i" (ASI_ITLB_DATA_ACCESS)); entry 166 arch/sparc/include/asm/spitfire.h static inline unsigned long spitfire_get_itlb_tag(int entry) entry 172 arch/sparc/include/asm/spitfire.h : "r" (entry << 3), "i" (ASI_ITLB_TAG_READ)); entry 176 arch/sparc/include/asm/spitfire.h static inline void spitfire_put_itlb_data(int entry, unsigned long data) entry 181 arch/sparc/include/asm/spitfire.h : "r" (data), "r" (entry << 3), entry 232 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_ldtlb_data(int entry) entry 239 arch/sparc/include/asm/spitfire.h : "r" ((0 << 16) | (entry << 3)), entry 245 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_litlb_data(int entry) entry 252 arch/sparc/include/asm/spitfire.h : "r" ((0 << 16) | (entry << 3)), entry 258 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_ldtlb_tag(int entry) entry 264 arch/sparc/include/asm/spitfire.h : "r" ((0 << 16) | (entry << 3)), entry 270 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_litlb_tag(int entry) entry 276 arch/sparc/include/asm/spitfire.h : "r" ((0 << 16) | (entry << 3)), entry 282 arch/sparc/include/asm/spitfire.h static inline void cheetah_put_ldtlb_data(int entry, unsigned long data) entry 288 arch/sparc/include/asm/spitfire.h "r" ((0 << 16) | (entry << 3)), entry 292 arch/sparc/include/asm/spitfire.h static inline void cheetah_put_litlb_data(int entry, unsigned long data) entry 298 arch/sparc/include/asm/spitfire.h "r" ((0 << 16) | (entry << 3)), entry 302 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_dtlb_data(int entry, int tlb) entry 309 arch/sparc/include/asm/spitfire.h : "r" ((tlb << 16) | (entry << 3)), "i" (ASI_DTLB_DATA_ACCESS)); entry 314 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_dtlb_tag(int entry, int tlb) entry 320 arch/sparc/include/asm/spitfire.h : "r" ((tlb << 16) | (entry << 3)), "i" (ASI_DTLB_TAG_READ)); entry 324 arch/sparc/include/asm/spitfire.h static inline void cheetah_put_dtlb_data(int entry, unsigned long data, int tlb) entry 330 arch/sparc/include/asm/spitfire.h "r" ((tlb << 16) | (entry << 3)), entry 334 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_itlb_data(int entry) entry 341 arch/sparc/include/asm/spitfire.h : "r" ((2 << 16) | (entry << 3)), entry 347 arch/sparc/include/asm/spitfire.h static inline unsigned long cheetah_get_itlb_tag(int entry) entry 353 arch/sparc/include/asm/spitfire.h : "r" ((2 << 16) | (entry << 3)), "i" (ASI_ITLB_TAG_READ)); entry 357 arch/sparc/include/asm/spitfire.h static inline void cheetah_put_itlb_data(int entry, unsigned long data) entry 362 arch/sparc/include/asm/spitfire.h : "r" (data), "r" ((2 << 16) | (entry << 3)), entry 228 arch/sparc/kernel/iommu-common.c unsigned long entry) entry 235 arch/sparc/kernel/iommu-common.c if (large_pool && entry >= largepool_start) { entry 238 arch/sparc/kernel/iommu-common.c unsigned int pool_nr = entry / tbl->poolsize; entry 251 arch/sparc/kernel/iommu-common.c unsigned long npages, unsigned long entry) entry 257 arch/sparc/kernel/iommu-common.c if (entry == IOMMU_ERROR_CODE) /* use default addr->entry mapping */ entry 258 arch/sparc/kernel/iommu-common.c entry = (dma_addr - iommu->table_map_base) >> shift; entry 259 arch/sparc/kernel/iommu-common.c pool = get_pool(iommu, entry); entry 262 arch/sparc/kernel/iommu-common.c bitmap_clear(iommu->map, entry, npages); entry 57 arch/sparc/kernel/iommu.c int entry; entry 60 arch/sparc/kernel/iommu.c for (entry = 0; entry < 16; entry++) { entry 160 arch/sparc/kernel/iommu.c unsigned long entry; entry 162 arch/sparc/kernel/iommu.c entry = iommu_tbl_range_alloc(dev, &iommu->tbl, npages, NULL, entry 164 arch/sparc/kernel/iommu.c if (unlikely(entry == IOMMU_ERROR_CODE)) entry 167 arch/sparc/kernel/iommu.c return iommu->page_table + entry; entry 479 arch/sparc/kernel/iommu.c unsigned long paddr, npages, entry, out_entry = 0, slen; entry 491 arch/sparc/kernel/iommu.c entry = iommu_tbl_range_alloc(dev, &iommu->tbl, npages, entry 495 arch/sparc/kernel/iommu.c if (unlikely(entry == IOMMU_ERROR_CODE)) { entry 502 arch/sparc/kernel/iommu.c base = iommu->page_table + entry; entry 506 arch/sparc/kernel/iommu.c (entry << IO_PAGE_SHIFT); entry 539 arch/sparc/kernel/iommu.c out_entry = entry; entry 559 arch/sparc/kernel/iommu.c unsigned long vaddr, npages, entry, j; entry 566 arch/sparc/kernel/iommu.c entry = (vaddr - iommu->tbl.table_map_base) entry 568 arch/sparc/kernel/iommu.c base = iommu->page_table + entry; entry 630 arch/sparc/kernel/iommu.c unsigned long npages, entry; entry 638 arch/sparc/kernel/iommu.c entry = ((dma_handle - iommu->tbl.table_map_base) entry 640 arch/sparc/kernel/iommu.c base = iommu->page_table + entry; entry 38 arch/sparc/kernel/iommu_common.h static inline int is_span_boundary(unsigned long entry, entry 48 arch/sparc/kernel/iommu_common.h return iommu_is_span_boundary(entry, nr, shift, boundary_size); entry 12 arch/sparc/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry, entry 15 arch/sparc/kernel/jump_label.c u32 *insn = (u32 *) (unsigned long) entry->code; entry 19 arch/sparc/kernel/jump_label.c s32 off = (s32)entry->target - (s32)entry->code; entry 324 arch/sparc/kernel/kprobes.c const struct exception_table_entry *entry; entry 370 arch/sparc/kernel/kprobes.c entry = search_exception_tables(regs->tpc); entry 371 arch/sparc/kernel/kprobes.c if (entry) { entry 372 arch/sparc/kernel/kprobes.c regs->tpc = entry->fixup; entry 1023 arch/sparc/kernel/ldc.c unsigned long entry, unsigned long npages) entry 1029 arch/sparc/kernel/ldc.c base = iommu->page_table + entry; entry 2020 arch/sparc/kernel/ldc.c long entry; entry 2022 arch/sparc/kernel/ldc.c entry = iommu_tbl_range_alloc(NULL, &iommu->iommu_map_table, entry 2024 arch/sparc/kernel/ldc.c if (unlikely(entry == IOMMU_ERROR_CODE)) entry 2027 arch/sparc/kernel/ldc.c return iommu->page_table + entry; entry 2236 arch/sparc/kernel/ldc.c unsigned long npages, entry; entry 2240 arch/sparc/kernel/ldc.c entry = ldc_cookie_to_index(cookie, iommu); entry 2241 arch/sparc/kernel/ldc.c ldc_demap(iommu, id, cookie, entry, npages); entry 2242 arch/sparc/kernel/ldc.c iommu_tbl_range_free(&iommu->iommu_map_table, cookie, npages, entry); entry 950 arch/sparc/kernel/pci.c struct msi_desc *entry = irq_get_msi_desc(irq); entry 951 arch/sparc/kernel/pci.c struct pci_dev *pdev = msi_desc_to_pci_dev(entry); entry 142 arch/sparc/kernel/pci_impl.h struct msi_desc *entry); entry 124 arch/sparc/kernel/pci_msi.c struct msi_desc *entry) entry 149 arch/sparc/kernel/pci_msi.c (entry->msi_attrib.is_64 ? 1 : 0)); entry 155 arch/sparc/kernel/pci_msi.c if (entry->msi_attrib.is_64) { entry 164 arch/sparc/kernel/pci_msi.c irq_set_msi_desc(*irq_p, entry); entry 57 arch/sparc/kernel/pci_sun4v.c unsigned long entry; /* Index into IOTSB. */ entry 66 arch/sparc/kernel/pci_sun4v.c static inline void iommu_batch_start(struct device *dev, unsigned long prot, unsigned long entry) entry 72 arch/sparc/kernel/pci_sun4v.c p->entry = entry; entry 89 arch/sparc/kernel/pci_sun4v.c unsigned long entry = p->entry; entry 102 arch/sparc/kernel/pci_sun4v.c HV_PCI_TSBID(0, entry), entry 110 arch/sparc/kernel/pci_sun4v.c HV_PCI_TSBID(0, entry), entry 116 arch/sparc/kernel/pci_sun4v.c index_count = HV_PCI_IOTSB_INDEX_COUNT(npages, entry), entry 133 arch/sparc/kernel/pci_sun4v.c entry += num; entry 138 arch/sparc/kernel/pci_sun4v.c p->entry = entry; entry 144 arch/sparc/kernel/pci_sun4v.c static inline void iommu_batch_new_entry(unsigned long entry, u64 mask) entry 148 arch/sparc/kernel/pci_sun4v.c if (p->entry + p->npages == entry) entry 150 arch/sparc/kernel/pci_sun4v.c if (p->entry != ~0UL) entry 152 arch/sparc/kernel/pci_sun4v.c p->entry = entry; entry 190 arch/sparc/kernel/pci_sun4v.c long entry; entry 218 arch/sparc/kernel/pci_sun4v.c entry = iommu_tbl_range_alloc(dev, tbl, npages, NULL, entry 221 arch/sparc/kernel/pci_sun4v.c if (unlikely(entry == IOMMU_ERROR_CODE)) entry 224 arch/sparc/kernel/pci_sun4v.c *dma_addrp = (tbl->table_map_base + (entry << IO_PAGE_SHIFT)); entry 233 arch/sparc/kernel/pci_sun4v.c entry); entry 296 arch/sparc/kernel/pci_sun4v.c unsigned long entry, unsigned long npages) entry 305 arch/sparc/kernel/pci_sun4v.c HV_PCI_TSBID(0, entry), entry 309 arch/sparc/kernel/pci_sun4v.c entry, npages, &num); entry 315 arch/sparc/kernel/pci_sun4v.c entry += num; entry 328 arch/sparc/kernel/pci_sun4v.c unsigned long order, npages, entry; entry 345 arch/sparc/kernel/pci_sun4v.c entry = ((dvma - tbl->table_map_base) >> IO_PAGE_SHIFT); entry 346 arch/sparc/kernel/pci_sun4v.c dma_4v_iommu_demap(dev, devhandle, dvma, iotsb_num, entry, npages); entry 366 arch/sparc/kernel/pci_sun4v.c long entry; entry 384 arch/sparc/kernel/pci_sun4v.c entry = iommu_tbl_range_alloc(dev, tbl, npages, NULL, entry 387 arch/sparc/kernel/pci_sun4v.c if (unlikely(entry == IOMMU_ERROR_CODE)) entry 390 arch/sparc/kernel/pci_sun4v.c bus_addr = (tbl->table_map_base + (entry << IO_PAGE_SHIFT)); entry 402 arch/sparc/kernel/pci_sun4v.c iommu_batch_start(dev, prot, entry); entry 437 arch/sparc/kernel/pci_sun4v.c long entry; entry 462 arch/sparc/kernel/pci_sun4v.c entry = (bus_addr - tbl->table_map_base) >> IO_PAGE_SHIFT; entry 463 arch/sparc/kernel/pci_sun4v.c dma_4v_iommu_demap(dev, devhandle, bus_addr, iotsb_num, entry, npages); entry 523 arch/sparc/kernel/pci_sun4v.c unsigned long paddr, npages, entry, out_entry = 0, slen; entry 534 arch/sparc/kernel/pci_sun4v.c entry = iommu_tbl_range_alloc(dev, tbl, npages, entry 538 arch/sparc/kernel/pci_sun4v.c if (unlikely(entry == IOMMU_ERROR_CODE)) { entry 544 arch/sparc/kernel/pci_sun4v.c iommu_batch_new_entry(entry, mask); entry 547 arch/sparc/kernel/pci_sun4v.c dma_addr = tbl->table_map_base + (entry << IO_PAGE_SHIFT); entry 581 arch/sparc/kernel/pci_sun4v.c out_entry = entry; entry 633 arch/sparc/kernel/pci_sun4v.c unsigned long flags, entry; entry 665 arch/sparc/kernel/pci_sun4v.c entry = ((dma_handle - tbl->table_map_base) >> shift); entry 667 arch/sparc/kernel/pci_sun4v.c entry, npages); entry 1737 arch/sparc/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 1747 arch/sparc/kernel/perf_event.c perf_callchain_store(entry, regs->tpc); entry 1771 arch/sparc/kernel/perf_event.c perf_callchain_store(entry, pc); entry 1779 arch/sparc/kernel/perf_event.c perf_callchain_store(entry, pc); entry 1784 arch/sparc/kernel/perf_event.c } while (entry->nr < entry->max_stack); entry 1797 arch/sparc/kernel/perf_event.c static void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry, entry 1817 arch/sparc/kernel/perf_event.c perf_callchain_store(entry, pc); entry 1818 arch/sparc/kernel/perf_event.c } while (entry->nr < entry->max_stack); entry 1821 arch/sparc/kernel/perf_event.c static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, entry 1849 arch/sparc/kernel/perf_event.c perf_callchain_store(entry, pc); entry 1850 arch/sparc/kernel/perf_event.c } while (entry->nr < entry->max_stack); entry 1854 arch/sparc/kernel/perf_event.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) entry 1859 arch/sparc/kernel/perf_event.c perf_callchain_store(entry, regs->tpc); entry 1869 arch/sparc/kernel/perf_event.c perf_callchain_user_32(entry, regs); entry 1871 arch/sparc/kernel/perf_event.c perf_callchain_user_64(entry, regs); entry 355 arch/sparc/kernel/smp_64.c unsigned long entry = entry 373 arch/sparc/kernel/smp_64.c prom_startcpu_cpuid(cpu, entry, cookie); entry 377 arch/sparc/kernel/smp_64.c prom_startcpu(dp->phandle, entry, cookie); entry 124 arch/sparc/kernel/sun4d_smp.c unsigned long *entry = &sun4d_cpu_startup; entry 140 arch/sparc/kernel/sun4d_smp.c printk(KERN_INFO "Starting CPU %d at %p\n", i, entry); entry 143 arch/sparc/kernel/sun4d_smp.c &smp_penguin_ctable, 0, (char *)entry); entry 81 arch/sparc/kernel/sun4m_smp.c unsigned long *entry = &sun4m_cpu_startup; entry 89 arch/sparc/kernel/sun4m_smp.c entry += ((i - 1) * 3); entry 101 arch/sparc/kernel/sun4m_smp.c printk(KERN_INFO "Starting CPU %d at %p\n", i, entry); entry 103 arch/sparc/kernel/sun4m_smp.c prom_startcpu(cpu_node, &smp_penguin_ctable, 0, (char *)entry); entry 302 arch/sparc/kernel/traps_64.c const struct exception_table_entry *entry; entry 304 arch/sparc/kernel/traps_64.c entry = search_exception_tables(regs->tpc); entry 305 arch/sparc/kernel/traps_64.c if (entry) { entry 310 arch/sparc/kernel/traps_64.c regs->tpc, entry->fixup); entry 312 arch/sparc/kernel/traps_64.c regs->tpc = entry->fixup; entry 351 arch/sparc/kernel/traps_64.c const struct exception_table_entry *entry; entry 353 arch/sparc/kernel/traps_64.c entry = search_exception_tables(regs->tpc); entry 354 arch/sparc/kernel/traps_64.c if (entry) { entry 359 arch/sparc/kernel/traps_64.c regs->tpc, entry->fixup); entry 361 arch/sparc/kernel/traps_64.c regs->tpc = entry->fixup; entry 1726 arch/sparc/kernel/traps_64.c const struct exception_table_entry *entry; entry 1728 arch/sparc/kernel/traps_64.c entry = search_exception_tables(regs->tpc); entry 1729 arch/sparc/kernel/traps_64.c if (entry) { entry 1748 arch/sparc/kernel/traps_64.c regs->tpc = entry->fixup; entry 2053 arch/sparc/kernel/traps_64.c const struct exception_table_entry *entry; entry 2055 arch/sparc/kernel/traps_64.c entry = search_exception_tables(regs->tpc); entry 2056 arch/sparc/kernel/traps_64.c if (entry) { entry 2062 arch/sparc/kernel/traps_64.c ent.err_raddr, entry->fixup); entry 2064 arch/sparc/kernel/traps_64.c regs->tpc = entry->fixup; entry 2677 arch/sparc/kernel/traps_64.c const struct exception_table_entry *entry; entry 2679 arch/sparc/kernel/traps_64.c entry = search_exception_tables(regs->tpc); entry 2680 arch/sparc/kernel/traps_64.c if (entry) { entry 2686 arch/sparc/kernel/traps_64.c regs->tpc, entry->fixup); entry 2688 arch/sparc/kernel/traps_64.c regs->tpc = entry->fixup; entry 263 arch/sparc/kernel/unaligned_64.c const struct exception_table_entry *entry; entry 265 arch/sparc/kernel/unaligned_64.c entry = search_exception_tables(regs->tpc); entry 266 arch/sparc/kernel/unaligned_64.c if (!entry) { entry 287 arch/sparc/kernel/unaligned_64.c regs->tpc = entry->fixup; entry 94 arch/sparc/mm/extable.c const struct exception_table_entry *entry; entry 96 arch/sparc/mm/extable.c entry = search_exception_tables(addr); entry 97 arch/sparc/mm/extable.c if (!entry) entry 101 arch/sparc/mm/extable.c if (!entry->fixup) { entry 102 arch/sparc/mm/extable.c *g2 = (addr - entry->insn) / 4; entry 103 arch/sparc/mm/extable.c return (entry + 1)->fixup; entry 106 arch/sparc/mm/extable.c return entry->fixup; entry 229 arch/sparc/mm/fault_64.c const struct exception_table_entry *entry; entry 231 arch/sparc/mm/fault_64.c entry = search_exception_tables(regs->tpc); entry 232 arch/sparc/mm/fault_64.c if (entry) { entry 233 arch/sparc/mm/fault_64.c regs->tpc = entry->fixup; entry 135 arch/sparc/mm/hugetlbpage.c static pte_t sun4u_hugepage_shift_to_tte(pte_t entry, unsigned int shift) entry 137 arch/sparc/mm/hugetlbpage.c return entry; entry 140 arch/sparc/mm/hugetlbpage.c static pte_t sun4v_hugepage_shift_to_tte(pte_t entry, unsigned int shift) entry 144 arch/sparc/mm/hugetlbpage.c pte_val(entry) = pte_val(entry) & ~_PAGE_SZALL_4V; entry 149 arch/sparc/mm/hugetlbpage.c pte_val(entry) |= _PAGE_PUD_HUGE; entry 153 arch/sparc/mm/hugetlbpage.c pte_val(entry) |= _PAGE_PMD_HUGE; entry 157 arch/sparc/mm/hugetlbpage.c pte_val(entry) |= _PAGE_PMD_HUGE; entry 160 arch/sparc/mm/hugetlbpage.c pte_val(entry) |= _PAGE_PMD_HUGE; entry 169 arch/sparc/mm/hugetlbpage.c pte_val(entry) = pte_val(entry) | hugepage_size; entry 170 arch/sparc/mm/hugetlbpage.c return entry; entry 173 arch/sparc/mm/hugetlbpage.c static pte_t hugepage_shift_to_tte(pte_t entry, unsigned int shift) entry 176 arch/sparc/mm/hugetlbpage.c return sun4v_hugepage_shift_to_tte(entry, shift); entry 178 arch/sparc/mm/hugetlbpage.c return sun4u_hugepage_shift_to_tte(entry, shift); entry 181 arch/sparc/mm/hugetlbpage.c pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma, entry 187 arch/sparc/mm/hugetlbpage.c pte = hugepage_shift_to_tte(entry, shift); entry 201 arch/sparc/mm/hugetlbpage.c static unsigned int sun4v_huge_tte_to_shift(pte_t entry) entry 203 arch/sparc/mm/hugetlbpage.c unsigned long tte_szbits = pte_val(entry) & _PAGE_SZALL_4V; entry 229 arch/sparc/mm/hugetlbpage.c static unsigned int sun4u_huge_tte_to_shift(pte_t entry) entry 231 arch/sparc/mm/hugetlbpage.c unsigned long tte_szbits = pte_val(entry) & _PAGE_SZALL_4U; entry 251 arch/sparc/mm/hugetlbpage.c static unsigned int huge_tte_to_shift(pte_t entry) entry 256 arch/sparc/mm/hugetlbpage.c shift = sun4v_huge_tte_to_shift(entry); entry 258 arch/sparc/mm/hugetlbpage.c shift = sun4u_huge_tte_to_shift(entry); entry 262 arch/sparc/mm/hugetlbpage.c pte_val(entry)); entry 321 arch/sparc/mm/hugetlbpage.c pte_t *ptep, pte_t entry) entry 327 arch/sparc/mm/hugetlbpage.c size = huge_tte_to_size(entry); entry 339 arch/sparc/mm/hugetlbpage.c if (!pte_present(*ptep) && pte_present(entry)) entry 347 arch/sparc/mm/hugetlbpage.c ptep[i] = __pte(pte_val(entry) + (i << shift)); entry 361 arch/sparc/mm/hugetlbpage.c pte_t entry; entry 363 arch/sparc/mm/hugetlbpage.c entry = *ptep; entry 364 arch/sparc/mm/hugetlbpage.c size = huge_tte_to_size(entry); entry 375 arch/sparc/mm/hugetlbpage.c orig_shift = pte_none(entry) ? PAGE_SHIFT : huge_tte_to_shift(entry); entry 377 arch/sparc/mm/hugetlbpage.c if (pte_present(entry)) entry 384 arch/sparc/mm/hugetlbpage.c maybe_tlb_batch_add(mm, addr, ptep, entry, 0, orig_shift); entry 387 arch/sparc/mm/hugetlbpage.c maybe_tlb_batch_add(mm, addr + REAL_HPAGE_SIZE, ptep, entry, 0, entry 390 arch/sparc/mm/hugetlbpage.c return entry; entry 2945 arch/sparc/mm/init_64.c pmd_t entry = *pmd; entry 2947 arch/sparc/mm/init_64.c if (!pmd_large(entry) || !pmd_young(entry)) entry 2950 arch/sparc/mm/init_64.c pte = pmd_val(entry); entry 2986 arch/sparc/mm/init_64.c const struct exception_table_entry *entry; entry 2988 arch/sparc/mm/init_64.c entry = search_exception_tables(regs->tpc); entry 2989 arch/sparc/mm/init_64.c if (entry) { entry 2990 arch/sparc/mm/init_64.c regs->tpc = entry->fixup; entry 419 arch/sparc/mm/srmmu.c static inline void remove_from_ctx_list(struct ctx_list *entry) entry 421 arch/sparc/mm/srmmu.c entry->next->prev = entry->prev; entry 422 arch/sparc/mm/srmmu.c entry->prev->next = entry->next; entry 425 arch/sparc/mm/srmmu.c static inline void add_to_ctx_list(struct ctx_list *head, struct ctx_list *entry) entry 427 arch/sparc/mm/srmmu.c entry->next = head; entry 428 arch/sparc/mm/srmmu.c (entry->prev = head->prev)->next = entry; entry 429 arch/sparc/mm/srmmu.c head->prev = entry; entry 431 arch/sparc/mm/srmmu.c #define add_to_free_ctxlist(entry) add_to_ctx_list(&ctx_free, entry) entry 432 arch/sparc/mm/srmmu.c #define add_to_used_ctxlist(entry) add_to_ctx_list(&ctx_used, entry) entry 247 arch/sparc/mm/tlb.c pmd_t old, entry; entry 249 arch/sparc/mm/tlb.c entry = __pmd(pmd_val(*pmdp) & ~_PAGE_VALID); entry 250 arch/sparc/mm/tlb.c old = pmdp_establish(vma, address, pmdp, entry); entry 258 arch/sparc/mm/tlb.c if ((pmd_val(entry) & _PAGE_PMD_HUGE) && entry 259 arch/sparc/mm/tlb.c !is_huge_zero_page(pmd_page(entry))) entry 441 arch/um/drivers/chan_kern.c const struct chan_type *entry; entry 450 arch/um/drivers/chan_kern.c entry = &chan_table[i]; entry 451 arch/um/drivers/chan_kern.c if (!strncmp(str, entry->key, strlen(entry->key))) { entry 452 arch/um/drivers/chan_kern.c ops = entry->ops; entry 453 arch/um/drivers/chan_kern.c str += strlen(entry->key); entry 348 arch/um/drivers/mconsole_kern.c struct list_head *entry = unplugged_pages.next; entry 351 arch/um/drivers/mconsole_kern.c unplugged = list_entry(entry, entry 567 arch/um/drivers/mconsole_kern.c struct mconsole_output *entry; entry 569 arch/um/drivers/mconsole_kern.c entry = list_entry(ele, struct mconsole_output, list); entry 570 arch/um/drivers/mconsole_kern.c mconsole_reply_len(entry->req, console_buf, n, 0, 1); entry 591 arch/um/drivers/mconsole_kern.c struct mconsole_output entry; entry 594 arch/um/drivers/mconsole_kern.c entry.req = req; entry 596 arch/um/drivers/mconsole_kern.c list_add(&entry.list, &clients); entry 604 arch/um/drivers/mconsole_kern.c list_del(&entry.list); entry 87 arch/um/include/asm/processor-generic.h extern void start_thread(struct pt_regs *regs, unsigned long entry, entry 50 arch/um/include/shared/mem_user.h extern void setup_memory(void *entry); entry 309 arch/x86/boot/compressed/acpi.c u8 *entry; entry 346 arch/x86/boot/compressed/acpi.c entry = (u8 *)(root_table + sizeof(struct acpi_table_header)); entry 350 arch/x86/boot/compressed/acpi.c acpi_table = *(u32 *)entry; entry 352 arch/x86/boot/compressed/acpi.c acpi_table = *(u64 *)entry; entry 360 arch/x86/boot/compressed/acpi.c entry += size; entry 514 arch/x86/boot/compressed/eboot.c struct boot_e820_entry *entry = params->e820_table; entry 587 arch/x86/boot/compressed/eboot.c entry = (struct boot_e820_entry *)e820ext->data; entry 590 arch/x86/boot/compressed/eboot.c entry->addr = d->phys_addr; entry 591 arch/x86/boot/compressed/eboot.c entry->size = d->num_pages << PAGE_SHIFT; entry 592 arch/x86/boot/compressed/eboot.c entry->type = e820_type; entry 593 arch/x86/boot/compressed/eboot.c prev = entry++; entry 576 arch/x86/boot/compressed/kaslr.c static void __process_mem_region(struct mem_vector *entry, entry 585 arch/x86/boot/compressed/kaslr.c if (IS_ENABLED(CONFIG_X86_32) && entry->start >= KERNEL_IMAGE_SIZE) entry 589 arch/x86/boot/compressed/kaslr.c if (entry->start + entry->size < minimum) entry 593 arch/x86/boot/compressed/kaslr.c end = min(entry->size + entry->start, mem_limit); entry 594 arch/x86/boot/compressed/kaslr.c if (entry->start >= end) entry 596 arch/x86/boot/compressed/kaslr.c cur_entry.start = entry->start; entry 597 arch/x86/boot/compressed/kaslr.c cur_entry.size = end - entry->start; entry 680 arch/x86/boot/compressed/kaslr.c struct mem_vector entry; entry 689 arch/x86/boot/compressed/kaslr.c entry.start = clamp(region->start, start, end); entry 691 arch/x86/boot/compressed/kaslr.c entry.size = entry_end - entry.start; entry 693 arch/x86/boot/compressed/kaslr.c __process_mem_region(&entry, minimum, image_size); entry 787 arch/x86/boot/compressed/kaslr.c struct boot_e820_entry *entry; entry 791 arch/x86/boot/compressed/kaslr.c entry = &boot_params->e820_table[i]; entry 793 arch/x86/boot/compressed/kaslr.c if (entry->type != E820_TYPE_RAM) entry 795 arch/x86/boot/compressed/kaslr.c region.start = entry->addr; entry 796 arch/x86/boot/compressed/kaslr.c region.size = entry->size; entry 47 arch/x86/boot/compressed/kaslr_64.c unsigned char *entry; entry 57 arch/x86/boot/compressed/kaslr_64.c entry = pages->pgt_buf + pages->pgt_buf_offset; entry 60 arch/x86/boot/compressed/kaslr_64.c return entry; entry 43 arch/x86/boot/compressed/pgtable_64.c struct boot_e820_entry *entry; entry 77 arch/x86/boot/compressed/pgtable_64.c entry = &boot_params->e820_table[i]; entry 80 arch/x86/boot/compressed/pgtable_64.c if (bios_start <= entry->addr) entry 84 arch/x86/boot/compressed/pgtable_64.c if (entry->type != E820_TYPE_RAM) entry 88 arch/x86/boot/compressed/pgtable_64.c if (bios_start > entry->addr + entry->size) entry 89 arch/x86/boot/compressed/pgtable_64.c new = entry->addr + entry->size; entry 95 arch/x86/boot/compressed/pgtable_64.c if (new - TRAMPOLINE_32BIT_SIZE < entry->addr) entry 28 arch/x86/entry/vsyscall/vsyscall_trace.h #define TRACE_INCLUDE_PATH ../../arch/x86/entry/vsyscall/ entry 2366 arch/x86/events/core.c perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) entry 2376 arch/x86/events/core.c if (perf_callchain_store(entry, regs->ip)) entry 2386 arch/x86/events/core.c if (!addr || perf_callchain_store(entry, addr)) entry 2430 arch/x86/events/core.c perf_callchain_user32(struct pt_regs *regs, struct perf_callchain_entry_ctx *entry) entry 2445 arch/x86/events/core.c while (entry->nr < entry->max_stack) { entry 2460 arch/x86/events/core.c perf_callchain_store(entry, cs_base + frame.return_address); entry 2468 arch/x86/events/core.c perf_callchain_user32(struct pt_regs *regs, struct perf_callchain_entry_ctx *entry) entry 2475 arch/x86/events/core.c perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) entry 2493 arch/x86/events/core.c perf_callchain_store(entry, regs->ip); entry 2498 arch/x86/events/core.c if (perf_callchain_user32(regs, entry)) entry 2502 arch/x86/events/core.c while (entry->nr < entry->max_stack) { entry 2518 arch/x86/events/core.c perf_callchain_store(entry, frame.return_address); entry 1274 arch/x86/events/intel/pt.c list_for_each_entry(filter, filters, entry) { entry 1310 arch/x86/events/intel/pt.c list_for_each_entry(filter, &head->list, entry) { entry 145 arch/x86/include/asm/asm.h # define _ASM_NOKPROBE(entry) \ entry 148 arch/x86/include/asm/asm.h _ASM_PTR (entry); \ entry 126 arch/x86/include/asm/desc.h #define write_ldt_entry(dt, entry, desc) native_write_ldt_entry(dt, entry, desc) entry 127 arch/x86/include/asm/desc.h #define write_gdt_entry(dt, entry, desc, type) native_write_gdt_entry(dt, entry, desc, type) entry 128 arch/x86/include/asm/desc.h #define write_idt_entry(dt, entry, g) native_write_idt_entry(dt, entry, g) entry 141 arch/x86/include/asm/desc.h static inline void native_write_idt_entry(gate_desc *idt, int entry, const gate_desc *gate) entry 143 arch/x86/include/asm/desc.h memcpy(&idt[entry], gate, sizeof(*gate)); entry 146 arch/x86/include/asm/desc.h static inline void native_write_ldt_entry(struct desc_struct *ldt, int entry, const void *desc) entry 148 arch/x86/include/asm/desc.h memcpy(&ldt[entry], desc, 8); entry 152 arch/x86/include/asm/desc.h native_write_gdt_entry(struct desc_struct *gdt, int entry, const void *desc, int type) entry 162 arch/x86/include/asm/desc.h memcpy(&gdt[entry], desc, size); entry 184 arch/x86/include/asm/desc.h static inline void __set_tss_desc(unsigned cpu, unsigned int entry, struct x86_hw_tss *addr) entry 191 arch/x86/include/asm/desc.h write_gdt_entry(d, entry, &tss, DESC_TSS); entry 1114 arch/x86/include/asm/kvm_host.h void (*set_supported_cpuid)(u32 func, struct kvm_cpuid_entry2 *entry); entry 1211 arch/x86/include/asm/kvm_host.h int (*get_msr_feature)(struct kvm_msr_entry *entry); entry 142 arch/x86/include/asm/mce.h struct mce entry[MCE_LOG_LEN]; entry 281 arch/x86/include/asm/paravirt.h static inline void write_ldt_entry(struct desc_struct *dt, int entry, entry 284 arch/x86/include/asm/paravirt.h PVOP_VCALL3(cpu.write_ldt_entry, dt, entry, desc); entry 287 arch/x86/include/asm/paravirt.h static inline void write_gdt_entry(struct desc_struct *dt, int entry, entry 290 arch/x86/include/asm/paravirt.h PVOP_VCALL4(cpu.write_gdt_entry, dt, entry, desc, type); entry 293 arch/x86/include/asm/paravirt.h static inline void write_idt_entry(gate_desc *dt, int entry, const gate_desc *g) entry 295 arch/x86/include/asm/paravirt.h PVOP_VCALL3(cpu.write_idt_entry, dt, entry, g); entry 1113 arch/x86/include/asm/pgtable.h pte_t entry, int dirty); entry 1163 arch/x86/include/asm/pgtable.h pmd_t entry, int dirty); entry 1166 arch/x86/include/asm/pgtable.h pud_t entry, int dirty); entry 99 arch/x86/include/asm/uv/bios.h } entry[1]; /* additional entries follow */ entry 78 arch/x86/kernel/apic/io_apic.c #define for_each_irq_pin(entry, head) \ entry 79 arch/x86/kernel/apic/io_apic.c list_for_each_entry(entry, &head, list) entry 93 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry entry; entry 304 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry entry; entry 314 arch/x86/kernel/apic/io_apic.c return eu.entry; entry 323 arch/x86/kernel/apic/io_apic.c eu.entry = __ioapic_read_entry(apic, pin); entry 326 arch/x86/kernel/apic/io_apic.c return eu.entry; entry 339 arch/x86/kernel/apic/io_apic.c eu.entry = e; entry 361 arch/x86/kernel/apic/io_apic.c union entry_union eu = { .entry.mask = IOAPIC_MASKED }; entry 377 arch/x86/kernel/apic/io_apic.c struct irq_pin_list *entry; entry 380 arch/x86/kernel/apic/io_apic.c for_each_irq_pin(entry, data->irq_2_pin) entry 381 arch/x86/kernel/apic/io_apic.c if (entry->apic == apic && entry->pin == pin) entry 384 arch/x86/kernel/apic/io_apic.c entry = kzalloc_node(sizeof(struct irq_pin_list), GFP_ATOMIC, node); entry 385 arch/x86/kernel/apic/io_apic.c if (!entry) { entry 390 arch/x86/kernel/apic/io_apic.c entry->apic = apic; entry 391 arch/x86/kernel/apic/io_apic.c entry->pin = pin; entry 392 arch/x86/kernel/apic/io_apic.c list_add_tail(&entry->list, &data->irq_2_pin); entry 399 arch/x86/kernel/apic/io_apic.c struct irq_pin_list *tmp, *entry; entry 401 arch/x86/kernel/apic/io_apic.c list_for_each_entry_safe(entry, tmp, &data->irq_2_pin, list) entry 402 arch/x86/kernel/apic/io_apic.c if (entry->apic == apic && entry->pin == pin) { entry 403 arch/x86/kernel/apic/io_apic.c list_del(&entry->list); entry 404 arch/x86/kernel/apic/io_apic.c kfree(entry); entry 423 arch/x86/kernel/apic/io_apic.c struct irq_pin_list *entry; entry 425 arch/x86/kernel/apic/io_apic.c for_each_irq_pin(entry, data->irq_2_pin) { entry 426 arch/x86/kernel/apic/io_apic.c if (entry->apic == oldapic && entry->pin == oldpin) { entry 427 arch/x86/kernel/apic/io_apic.c entry->apic = newapic; entry 428 arch/x86/kernel/apic/io_apic.c entry->pin = newpin; entry 440 arch/x86/kernel/apic/io_apic.c void (*final)(struct irq_pin_list *entry)) entry 443 arch/x86/kernel/apic/io_apic.c struct irq_pin_list *entry; entry 445 arch/x86/kernel/apic/io_apic.c eu.entry = data->entry; entry 448 arch/x86/kernel/apic/io_apic.c data->entry = eu.entry; entry 450 arch/x86/kernel/apic/io_apic.c for_each_irq_pin(entry, data->irq_2_pin) { entry 451 arch/x86/kernel/apic/io_apic.c io_apic_write(entry->apic, 0x10 + 2 * entry->pin, eu.w1); entry 453 arch/x86/kernel/apic/io_apic.c final(entry); entry 457 arch/x86/kernel/apic/io_apic.c static void io_apic_sync(struct irq_pin_list *entry) entry 465 arch/x86/kernel/apic/io_apic.c io_apic = io_apic_base(entry->apic); entry 515 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry entry, entry1; entry 517 arch/x86/kernel/apic/io_apic.c entry = entry1 = __ioapic_read_entry(apic, pin); entry 530 arch/x86/kernel/apic/io_apic.c __ioapic_write_entry(apic, pin, entry); entry 537 arch/x86/kernel/apic/io_apic.c struct irq_pin_list *entry; entry 540 arch/x86/kernel/apic/io_apic.c for_each_irq_pin(entry, data->irq_2_pin) entry 541 arch/x86/kernel/apic/io_apic.c __eoi_ioapic_pin(entry->apic, entry->pin, vector); entry 547 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry entry; entry 550 arch/x86/kernel/apic/io_apic.c entry = ioapic_read_entry(apic, pin); entry 551 arch/x86/kernel/apic/io_apic.c if (entry.delivery_mode == dest_SMI) entry 558 arch/x86/kernel/apic/io_apic.c if (entry.mask == IOAPIC_UNMASKED) { entry 559 arch/x86/kernel/apic/io_apic.c entry.mask = IOAPIC_MASKED; entry 560 arch/x86/kernel/apic/io_apic.c ioapic_write_entry(apic, pin, entry); entry 561 arch/x86/kernel/apic/io_apic.c entry = ioapic_read_entry(apic, pin); entry 564 arch/x86/kernel/apic/io_apic.c if (entry.irr) { entry 572 arch/x86/kernel/apic/io_apic.c if (entry.trigger == IOAPIC_EDGE) { entry 573 arch/x86/kernel/apic/io_apic.c entry.trigger = IOAPIC_LEVEL; entry 574 arch/x86/kernel/apic/io_apic.c ioapic_write_entry(apic, pin, entry); entry 577 arch/x86/kernel/apic/io_apic.c __eoi_ioapic_pin(apic, pin, entry.vector); entry 586 arch/x86/kernel/apic/io_apic.c entry = ioapic_read_entry(apic, pin); entry 587 arch/x86/kernel/apic/io_apic.c if (entry.irr) entry 672 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry entry; entry 674 arch/x86/kernel/apic/io_apic.c entry = ioapics[apic].saved_registers[pin]; entry 675 arch/x86/kernel/apic/io_apic.c if (entry.mask == IOAPIC_UNMASKED) { entry 676 arch/x86/kernel/apic/io_apic.c entry.mask = IOAPIC_MASKED; entry 677 arch/x86/kernel/apic/io_apic.c ioapic_write_entry(apic, pin, entry); entry 951 arch/x86/kernel/apic/io_apic.c data->entry.trigger = data->trigger = info->ioapic_trigger; entry 952 arch/x86/kernel/apic/io_apic.c data->entry.polarity = data->polarity = info->ioapic_polarity; entry 1237 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry entry; entry 1238 arch/x86/kernel/apic/io_apic.c struct IR_IO_APIC_route_entry *ir_entry = (void *)&entry; entry 1242 arch/x86/kernel/apic/io_apic.c entry = ioapic_read_entry(apic, i); entry 1246 arch/x86/kernel/apic/io_apic.c entry.mask == IOAPIC_MASKED ? "disabled" : "enabled ", entry 1247 arch/x86/kernel/apic/io_apic.c entry.trigger == IOAPIC_LEVEL ? "level" : "edge ", entry 1248 arch/x86/kernel/apic/io_apic.c entry.polarity == IOAPIC_POL_LOW ? "low " : "high", entry 1249 arch/x86/kernel/apic/io_apic.c entry.vector, entry.irr, entry.delivery_status); entry 1257 arch/x86/kernel/apic/io_apic.c entry.dest_mode == IOAPIC_DEST_MODE_LOGICAL ? entry 1259 arch/x86/kernel/apic/io_apic.c entry.dest, entry.delivery_mode); entry 1341 arch/x86/kernel/apic/io_apic.c struct irq_pin_list *entry; entry 1355 arch/x86/kernel/apic/io_apic.c for_each_irq_pin(entry, data->irq_2_pin) entry 1356 arch/x86/kernel/apic/io_apic.c pr_cont("-> %d:%d", entry->apic, entry->pin); entry 1379 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry entry = ioapic_read_entry(apic, pin); entry 1384 arch/x86/kernel/apic/io_apic.c if ((entry.mask == 0) && (entry.delivery_mode == dest_ExtINT)) { entry 1425 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry entry; entry 1427 arch/x86/kernel/apic/io_apic.c memset(&entry, 0, sizeof(entry)); entry 1428 arch/x86/kernel/apic/io_apic.c entry.mask = IOAPIC_UNMASKED; entry 1429 arch/x86/kernel/apic/io_apic.c entry.trigger = IOAPIC_EDGE; entry 1430 arch/x86/kernel/apic/io_apic.c entry.polarity = IOAPIC_POL_HIGH; entry 1431 arch/x86/kernel/apic/io_apic.c entry.dest_mode = IOAPIC_DEST_MODE_PHYSICAL; entry 1432 arch/x86/kernel/apic/io_apic.c entry.delivery_mode = dest_ExtINT; entry 1433 arch/x86/kernel/apic/io_apic.c entry.dest = read_apic_id(); entry 1438 arch/x86/kernel/apic/io_apic.c ioapic_write_entry(ioapic_i8259.apic, ioapic_i8259.pin, entry); entry 1707 arch/x86/kernel/apic/io_apic.c struct irq_pin_list *entry; entry 1711 arch/x86/kernel/apic/io_apic.c for_each_irq_pin(entry, data->irq_2_pin) { entry 1715 arch/x86/kernel/apic/io_apic.c pin = entry->pin; entry 1716 arch/x86/kernel/apic/io_apic.c reg = io_apic_read(entry->apic, 0x10 + pin*2); entry 1862 arch/x86/kernel/apic/io_apic.c eoi_ioapic_pin(data->entry.vector, data); entry 1869 arch/x86/kernel/apic/io_apic.c struct irq_pin_list *entry; entry 1877 arch/x86/kernel/apic/io_apic.c mpd->entry.dest = cfg->dest_apicid; entry 1878 arch/x86/kernel/apic/io_apic.c mpd->entry.vector = cfg->vector; entry 1880 arch/x86/kernel/apic/io_apic.c for_each_irq_pin(entry, mpd->irq_2_pin) entry 1881 arch/x86/kernel/apic/io_apic.c __ioapic_write_entry(entry->apic, entry->pin, mpd->entry); entry 2955 arch/x86/kernel/apic/io_apic.c struct IO_APIC_route_entry *entry) entry 2957 arch/x86/kernel/apic/io_apic.c memset(entry, 0, sizeof(*entry)); entry 2958 arch/x86/kernel/apic/io_apic.c entry->delivery_mode = apic->irq_delivery_mode; entry 2959 arch/x86/kernel/apic/io_apic.c entry->dest_mode = apic->irq_dest_mode; entry 2960 arch/x86/kernel/apic/io_apic.c entry->dest = cfg->dest_apicid; entry 2961 arch/x86/kernel/apic/io_apic.c entry->vector = cfg->vector; entry 2962 arch/x86/kernel/apic/io_apic.c entry->trigger = data->trigger; entry 2963 arch/x86/kernel/apic/io_apic.c entry->polarity = data->polarity; entry 2969 arch/x86/kernel/apic/io_apic.c entry->mask = IOAPIC_MASKED; entry 2971 arch/x86/kernel/apic/io_apic.c entry->mask = IOAPIC_UNMASKED; entry 2999 arch/x86/kernel/apic/io_apic.c info->ioapic_entry = &data->entry; entry 1272 arch/x86/kernel/apic/x2apic_uv_x.c for (i = 0; st->entry[i].type != UV_SYSTAB_TYPE_UNUSED; i++) { entry 1273 arch/x86/kernel/apic/x2apic_uv_x.c unsigned long ptr = st->entry[i].offset; entry 1280 arch/x86/kernel/apic/x2apic_uv_x.c switch (st->entry[i].type) { entry 55 arch/x86/kernel/asm-offsets_64.c #define ENTRY(entry) OFFSET(pt_regs_ ## entry, pt_regs, entry) entry 75 arch/x86/kernel/asm-offsets_64.c #define ENTRY(entry) OFFSET(saved_context_ ## entry, saved_context, entry) entry 475 arch/x86/kernel/cpu/mce/core.c static void mce_irq_work_cb(struct irq_work *entry) entry 44 arch/x86/kernel/cpu/mce/dev-mcelog.c unsigned int entry; entry 48 arch/x86/kernel/cpu/mce/dev-mcelog.c entry = mcelog.next; entry 54 arch/x86/kernel/cpu/mce/dev-mcelog.c if (entry >= MCE_LOG_LEN) { entry 59 arch/x86/kernel/cpu/mce/dev-mcelog.c mcelog.next = entry + 1; entry 61 arch/x86/kernel/cpu/mce/dev-mcelog.c memcpy(mcelog.entry + entry, mce, sizeof(struct mce)); entry 62 arch/x86/kernel/cpu/mce/dev-mcelog.c mcelog.entry[entry].finished = 1; entry 224 arch/x86/kernel/cpu/mce/dev-mcelog.c struct mce *m = &mcelog.entry[i]; entry 230 arch/x86/kernel/cpu/mce/dev-mcelog.c memset(mcelog.entry, 0, next * sizeof(struct mce)); entry 41 arch/x86/kernel/cpu/microcode/amd.c struct equiv_cpu_entry *entry; entry 75 arch/x86/kernel/cpu/microcode/amd.c struct equiv_cpu_entry *e = &et->entry[i]; entry 309 arch/x86/kernel/cpu/microcode/amd.c table.entry = (struct equiv_cpu_entry *)(buf + CONTAINER_HDR_SZ); entry 730 arch/x86/kernel/cpu/microcode/amd.c equiv_table.entry = vmalloc(equiv_tbl_len); entry 731 arch/x86/kernel/cpu/microcode/amd.c if (!equiv_table.entry) { entry 736 arch/x86/kernel/cpu/microcode/amd.c memcpy(equiv_table.entry, buf + CONTAINER_HDR_SZ, equiv_tbl_len); entry 745 arch/x86/kernel/cpu/microcode/amd.c vfree(equiv_table.entry); entry 69 arch/x86/kernel/cpu/resctrl/monitor.c struct rmid_entry *entry; entry 71 arch/x86/kernel/cpu/resctrl/monitor.c entry = &rmid_ptrs[rmid]; entry 72 arch/x86/kernel/cpu/resctrl/monitor.c WARN_ON(entry->rmid != rmid); entry 74 arch/x86/kernel/cpu/resctrl/monitor.c return entry; entry 95 arch/x86/kernel/cpu/resctrl/monitor.c static bool rmid_dirty(struct rmid_entry *entry) entry 97 arch/x86/kernel/cpu/resctrl/monitor.c u64 val = __rmid_read(entry->rmid, QOS_L3_OCCUP_EVENT_ID); entry 110 arch/x86/kernel/cpu/resctrl/monitor.c struct rmid_entry *entry; entry 127 arch/x86/kernel/cpu/resctrl/monitor.c entry = __rmid_entry(nrmid); entry 128 arch/x86/kernel/cpu/resctrl/monitor.c if (force_free || !rmid_dirty(entry)) { entry 129 arch/x86/kernel/cpu/resctrl/monitor.c clear_bit(entry->rmid, d->rmid_busy_llc); entry 130 arch/x86/kernel/cpu/resctrl/monitor.c if (!--entry->busy) { entry 132 arch/x86/kernel/cpu/resctrl/monitor.c list_add_tail(&entry->list, &rmid_free_lru); entry 151 arch/x86/kernel/cpu/resctrl/monitor.c struct rmid_entry *entry; entry 158 arch/x86/kernel/cpu/resctrl/monitor.c entry = list_first_entry(&rmid_free_lru, entry 160 arch/x86/kernel/cpu/resctrl/monitor.c list_del(&entry->list); entry 162 arch/x86/kernel/cpu/resctrl/monitor.c return entry->rmid; entry 165 arch/x86/kernel/cpu/resctrl/monitor.c static void add_rmid_to_limbo(struct rmid_entry *entry) entry 174 arch/x86/kernel/cpu/resctrl/monitor.c entry->busy = 0; entry 178 arch/x86/kernel/cpu/resctrl/monitor.c val = __rmid_read(entry->rmid, QOS_L3_OCCUP_EVENT_ID); entry 189 arch/x86/kernel/cpu/resctrl/monitor.c set_bit(entry->rmid, d->rmid_busy_llc); entry 190 arch/x86/kernel/cpu/resctrl/monitor.c entry->busy++; entry 194 arch/x86/kernel/cpu/resctrl/monitor.c if (entry->busy) entry 197 arch/x86/kernel/cpu/resctrl/monitor.c list_add_tail(&entry->list, &rmid_free_lru); entry 202 arch/x86/kernel/cpu/resctrl/monitor.c struct rmid_entry *entry; entry 209 arch/x86/kernel/cpu/resctrl/monitor.c entry = __rmid_entry(rmid); entry 212 arch/x86/kernel/cpu/resctrl/monitor.c add_rmid_to_limbo(entry); entry 214 arch/x86/kernel/cpu/resctrl/monitor.c list_add_tail(&entry->list, &rmid_free_lru); entry 299 arch/x86/kernel/cpu/resctrl/monitor.c struct rdtgroup *rdtgrp, *entry; entry 314 arch/x86/kernel/cpu/resctrl/monitor.c list_for_each_entry(entry, head, mon.crdtgrp_list) { entry 315 arch/x86/kernel/cpu/resctrl/monitor.c if (__mon_event_count(entry->mon.rmid, rr)) entry 361 arch/x86/kernel/cpu/resctrl/monitor.c struct rdtgroup *entry; entry 386 arch/x86/kernel/cpu/resctrl/monitor.c list_for_each_entry(entry, head, mon.crdtgrp_list) { entry 387 arch/x86/kernel/cpu/resctrl/monitor.c cmbm_data = &dom_mbm->mbm_local[entry->mon.rmid]; entry 430 arch/x86/kernel/cpu/resctrl/monitor.c list_for_each_entry(entry, head, mon.crdtgrp_list) { entry 431 arch/x86/kernel/cpu/resctrl/monitor.c cmbm_data = &dom_mbm->mbm_local[entry->mon.rmid]; entry 555 arch/x86/kernel/cpu/resctrl/monitor.c struct rmid_entry *entry = NULL; entry 564 arch/x86/kernel/cpu/resctrl/monitor.c entry = &rmid_ptrs[i]; entry 565 arch/x86/kernel/cpu/resctrl/monitor.c INIT_LIST_HEAD(&entry->list); entry 567 arch/x86/kernel/cpu/resctrl/monitor.c entry->rmid = i; entry 568 arch/x86/kernel/cpu/resctrl/monitor.c list_add_tail(&entry->list, &rmid_free_lru); entry 575 arch/x86/kernel/cpu/resctrl/monitor.c entry = __rmid_entry(0); entry 576 arch/x86/kernel/cpu/resctrl/monitor.c list_del(&entry->list); entry 291 arch/x86/kernel/crash.c static int add_e820_entry(struct boot_params *params, struct e820_entry *entry) entry 299 arch/x86/kernel/crash.c memcpy(¶ms->e820_table[nr_e820_entries], entry, entry 83 arch/x86/kernel/e820.c struct e820_entry *entry = &table->entries[i]; entry 85 arch/x86/kernel/e820.c if (type && entry->type != type) entry 87 arch/x86/kernel/e820.c if (entry->addr >= end || entry->addr + entry->size <= start) entry 118 arch/x86/kernel/e820.c struct e820_entry *entry = &e820_table->entries[i]; entry 120 arch/x86/kernel/e820.c if (type && entry->type != type) entry 124 arch/x86/kernel/e820.c if (entry->addr >= end || entry->addr + entry->size <= start) entry 131 arch/x86/kernel/e820.c if (entry->addr <= start) entry 132 arch/x86/kernel/e820.c start = entry->addr + entry->size; entry 139 arch/x86/kernel/e820.c return entry; entry 158 arch/x86/kernel/e820.c struct e820_entry *entry = __e820__mapped_all(start, end, 0); entry 160 arch/x86/kernel/e820.c return entry ? entry->type : -EINVAL; entry 280 arch/x86/kernel/e820.c struct e820_entry *entry; entry 304 arch/x86/kernel/e820.c return (ap->addr != ap->entry->addr) - (bp->addr != bp->entry->addr); entry 340 arch/x86/kernel/e820.c change_point[chg_idx++]->entry = &entries[i]; entry 342 arch/x86/kernel/e820.c change_point[chg_idx++]->entry = &entries[i]; entry 359 arch/x86/kernel/e820.c if (change_point[chg_idx]->addr == change_point[chg_idx]->entry->addr) { entry 361 arch/x86/kernel/e820.c overlap_list[overlap_entries++] = change_point[chg_idx]->entry; entry 365 arch/x86/kernel/e820.c if (overlap_list[i] == change_point[chg_idx]->entry) entry 409 arch/x86/kernel/e820.c struct boot_e820_entry *entry = entries; entry 412 arch/x86/kernel/e820.c u64 start = entry->addr; entry 413 arch/x86/kernel/e820.c u64 size = entry->size; entry 415 arch/x86/kernel/e820.c u32 type = entry->type; entry 423 arch/x86/kernel/e820.c entry++; entry 467 arch/x86/kernel/e820.c struct e820_entry *entry = &table->entries[i]; entry 471 arch/x86/kernel/e820.c if (entry->type != old_type) entry 474 arch/x86/kernel/e820.c entry_end = entry->addr + entry->size; entry 477 arch/x86/kernel/e820.c if (entry->addr >= start && entry_end <= end) { entry 478 arch/x86/kernel/e820.c entry->type = new_type; entry 479 arch/x86/kernel/e820.c real_updated_size += entry->size; entry 484 arch/x86/kernel/e820.c if (entry->addr < start && entry_end > end) { entry 486 arch/x86/kernel/e820.c __e820__range_add(table, end, entry_end - end, entry->type); entry 487 arch/x86/kernel/e820.c entry->size = start - entry->addr; entry 493 arch/x86/kernel/e820.c final_start = max(start, entry->addr); entry 506 arch/x86/kernel/e820.c entry->size -= final_end - final_start; entry 507 arch/x86/kernel/e820.c if (entry->addr < final_start) entry 510 arch/x86/kernel/e820.c entry->addr = final_end; entry 542 arch/x86/kernel/e820.c struct e820_entry *entry = &e820_table->entries[i]; entry 546 arch/x86/kernel/e820.c if (check_type && entry->type != old_type) entry 549 arch/x86/kernel/e820.c entry_end = entry->addr + entry->size; entry 552 arch/x86/kernel/e820.c if (entry->addr >= start && entry_end <= end) { entry 553 arch/x86/kernel/e820.c real_removed_size += entry->size; entry 554 arch/x86/kernel/e820.c memset(entry, 0, sizeof(*entry)); entry 559 arch/x86/kernel/e820.c if (entry->addr < start && entry_end > end) { entry 560 arch/x86/kernel/e820.c e820__range_add(end, entry_end - end, entry->type); entry 561 arch/x86/kernel/e820.c entry->size = start - entry->addr; entry 567 arch/x86/kernel/e820.c final_start = max(start, entry->addr); entry 578 arch/x86/kernel/e820.c entry->size -= final_end - final_start; entry 579 arch/x86/kernel/e820.c if (entry->addr < final_start) entry 582 arch/x86/kernel/e820.c entry->addr = final_end; entry 744 arch/x86/kernel/e820.c struct e820_entry *entry = &e820_table->entries[i]; entry 746 arch/x86/kernel/e820.c if (pfn < PFN_UP(entry->addr)) entry 747 arch/x86/kernel/e820.c register_nosave_region(pfn, PFN_UP(entry->addr)); entry 749 arch/x86/kernel/e820.c pfn = PFN_DOWN(entry->addr + entry->size); entry 751 arch/x86/kernel/e820.c if (entry->type != E820_TYPE_RAM && entry->type != E820_TYPE_RESERVED_KERN) entry 752 arch/x86/kernel/e820.c register_nosave_region(PFN_UP(entry->addr), pfn); entry 769 arch/x86/kernel/e820.c struct e820_entry *entry = &e820_table->entries[i]; entry 771 arch/x86/kernel/e820.c if (entry->type == E820_TYPE_NVS) entry 772 arch/x86/kernel/e820.c acpi_nvs_register(entry->addr, entry->size); entry 822 arch/x86/kernel/e820.c struct e820_entry *entry = &e820_table->entries[i]; entry 826 arch/x86/kernel/e820.c if (entry->type != type) entry 829 arch/x86/kernel/e820.c start_pfn = entry->addr >> PAGE_SHIFT; entry 830 arch/x86/kernel/e820.c end_pfn = (entry->addr + entry->size) >> PAGE_SHIFT; entry 1029 arch/x86/kernel/e820.c static const char *__init e820_type_to_string(struct e820_entry *entry) entry 1031 arch/x86/kernel/e820.c switch (entry->type) { entry 1044 arch/x86/kernel/e820.c static unsigned long __init e820_type_to_iomem_type(struct e820_entry *entry) entry 1046 arch/x86/kernel/e820.c switch (entry->type) { entry 1059 arch/x86/kernel/e820.c static unsigned long __init e820_type_to_iores_desc(struct e820_entry *entry) entry 1061 arch/x86/kernel/e820.c switch (entry->type) { entry 1119 arch/x86/kernel/e820.c struct e820_entry *entry = e820_table->entries + i; entry 1121 arch/x86/kernel/e820.c end = entry->addr + entry->size - 1; entry 1126 arch/x86/kernel/e820.c res->start = entry->addr; entry 1128 arch/x86/kernel/e820.c res->name = e820_type_to_string(entry); entry 1129 arch/x86/kernel/e820.c res->flags = e820_type_to_iomem_type(entry); entry 1130 arch/x86/kernel/e820.c res->desc = e820_type_to_iores_desc(entry); entry 1137 arch/x86/kernel/e820.c if (do_mark_busy(entry->type, res)) { entry 1146 arch/x86/kernel/e820.c struct e820_entry *entry = e820_table_firmware->entries + i; entry 1148 arch/x86/kernel/e820.c firmware_map_add_early(entry->addr, entry->addr + entry->size, e820_type_to_string(entry)); entry 1190 arch/x86/kernel/e820.c struct e820_entry *entry = &e820_table->entries[i]; entry 1193 arch/x86/kernel/e820.c if (entry->type != E820_TYPE_RAM) entry 1196 arch/x86/kernel/e820.c start = entry->addr + entry->size; entry 1282 arch/x86/kernel/e820.c struct e820_entry *entry = &e820_table->entries[i]; entry 1284 arch/x86/kernel/e820.c end = entry->addr + entry->size; entry 1288 arch/x86/kernel/e820.c if (entry->type != E820_TYPE_RAM && entry->type != E820_TYPE_RESERVED_KERN) entry 1291 arch/x86/kernel/e820.c memblock_add(entry->addr, entry->size); entry 310 arch/x86/kernel/idt.c void *entry; entry 315 arch/x86/kernel/idt.c entry = irq_entries_start + 8 * (i - FIRST_EXTERNAL_VECTOR); entry 316 arch/x86/kernel/idt.c set_intr_gate(i, entry); entry 322 arch/x86/kernel/idt.c entry = spurious_entries_start + 8 * (i - FIRST_SYSTEM_VECTOR); entry 323 arch/x86/kernel/idt.c set_intr_gate(i, entry); entry 38 arch/x86/kernel/jump_label.c static void __jump_label_set_jump_code(struct jump_entry *entry, entry 49 arch/x86/kernel/jump_label.c code->offset = jump_entry_target(entry) - entry 50 arch/x86/kernel/jump_label.c (jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE); entry 60 arch/x86/kernel/jump_label.c if (memcmp((void *)jump_entry_code(entry), expect, JUMP_LABEL_NOP_SIZE)) entry 61 arch/x86/kernel/jump_label.c bug_at((void *)jump_entry_code(entry), line); entry 67 arch/x86/kernel/jump_label.c static void __ref __jump_label_transform(struct jump_entry *entry, entry 73 arch/x86/kernel/jump_label.c __jump_label_set_jump_code(entry, type, &code, init); entry 87 arch/x86/kernel/jump_label.c text_poke_early((void *)jump_entry_code(entry), &code, entry 92 arch/x86/kernel/jump_label.c text_poke_bp((void *)jump_entry_code(entry), &code, JUMP_LABEL_NOP_SIZE, entry 93 arch/x86/kernel/jump_label.c (void *)jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE); entry 96 arch/x86/kernel/jump_label.c void arch_jump_label_transform(struct jump_entry *entry, entry 100 arch/x86/kernel/jump_label.c __jump_label_transform(entry, type, 0); entry 108 arch/x86/kernel/jump_label.c bool arch_jump_label_transform_queue(struct jump_entry *entry, entry 118 arch/x86/kernel/jump_label.c arch_jump_label_transform(entry, type); entry 131 arch/x86/kernel/jump_label.c entry_code = (void *)jump_entry_code(entry); entry 149 arch/x86/kernel/jump_label.c __jump_label_set_jump_code(entry, type, entry 179 arch/x86/kernel/jump_label.c __init_or_module void arch_jump_label_transform_static(struct jump_entry *entry, entry 199 arch/x86/kernel/jump_label.c __jump_label_transform(entry, type, 1); entry 257 arch/x86/kernel/pci-calgary_64.c unsigned long entry; entry 260 arch/x86/kernel/pci-calgary_64.c entry = iommu_range_alloc(dev, tbl, npages); entry 261 arch/x86/kernel/pci-calgary_64.c if (unlikely(entry == DMA_MAPPING_ERROR)) { entry 268 arch/x86/kernel/pci-calgary_64.c ret = (entry << PAGE_SHIFT) | ((unsigned long)vaddr & ~PAGE_MASK); entry 271 arch/x86/kernel/pci-calgary_64.c tce_build(tbl, entry, npages, (unsigned long)vaddr & PAGE_MASK, entry 279 arch/x86/kernel/pci-calgary_64.c unsigned long entry; entry 289 arch/x86/kernel/pci-calgary_64.c entry = dma_addr >> PAGE_SHIFT; entry 291 arch/x86/kernel/pci-calgary_64.c BUG_ON(entry + npages > tbl->it_size); entry 293 arch/x86/kernel/pci-calgary_64.c tce_free(tbl, entry, npages); entry 297 arch/x86/kernel/pci-calgary_64.c bitmap_clear(tbl->it_map, entry, npages); entry 357 arch/x86/kernel/pci-calgary_64.c unsigned long entry; entry 366 arch/x86/kernel/pci-calgary_64.c entry = iommu_range_alloc(dev, tbl, npages); entry 367 arch/x86/kernel/pci-calgary_64.c if (entry == DMA_MAPPING_ERROR) { entry 373 arch/x86/kernel/pci-calgary_64.c s->dma_address = (entry << PAGE_SHIFT) | s->offset; entry 376 arch/x86/kernel/pci-calgary_64.c tce_build(tbl, entry, npages, vaddr & PAGE_MASK, dir); entry 29 arch/x86/kernel/resource.c struct e820_entry *entry; entry 32 arch/x86/kernel/resource.c entry = &e820_table->entries[i]; entry 34 arch/x86/kernel/resource.c resource_clip(avail, entry->addr, entry 35 arch/x86/kernel/resource.c entry->addr + entry->size - 1); entry 166 arch/x86/kvm/cpuid.c struct kvm_cpuid_entry2 *e, *entry; entry 168 arch/x86/kvm/cpuid.c entry = NULL; entry 172 arch/x86/kvm/cpuid.c entry = e; entry 176 arch/x86/kvm/cpuid.c if (entry && (entry->edx & F(NX)) && !is_efer_nx()) { entry 177 arch/x86/kvm/cpuid.c entry->edx &= ~F(NX); entry 289 arch/x86/kvm/cpuid.c static void do_host_cpuid(struct kvm_cpuid_entry2 *entry, u32 function, entry 292 arch/x86/kvm/cpuid.c entry->function = function; entry 293 arch/x86/kvm/cpuid.c entry->index = index; entry 294 arch/x86/kvm/cpuid.c entry->flags = 0; entry 296 arch/x86/kvm/cpuid.c cpuid_count(entry->function, entry->index, entry 297 arch/x86/kvm/cpuid.c &entry->eax, &entry->ebx, &entry->ecx, &entry->edx); entry 301 arch/x86/kvm/cpuid.c entry->flags |= KVM_CPUID_FLAG_STATEFUL_FUNC; entry 315 arch/x86/kvm/cpuid.c entry->flags |= KVM_CPUID_FLAG_SIGNIFCANT_INDEX; entry 320 arch/x86/kvm/cpuid.c static int __do_cpuid_func_emulated(struct kvm_cpuid_entry2 *entry, entry 323 arch/x86/kvm/cpuid.c entry->function = func; entry 324 arch/x86/kvm/cpuid.c entry->index = 0; entry 325 arch/x86/kvm/cpuid.c entry->flags = 0; entry 329 arch/x86/kvm/cpuid.c entry->eax = 7; entry 333 arch/x86/kvm/cpuid.c entry->ecx = F(MOVBE); entry 337 arch/x86/kvm/cpuid.c entry->flags |= KVM_CPUID_FLAG_SIGNIFCANT_INDEX; entry 338 arch/x86/kvm/cpuid.c entry->eax = 0; entry 339 arch/x86/kvm/cpuid.c entry->ecx = F(RDPID); entry 348 arch/x86/kvm/cpuid.c static inline void do_cpuid_7_mask(struct kvm_cpuid_entry2 *entry, int index) entry 384 arch/x86/kvm/cpuid.c entry->eax = min(entry->eax, 1u); entry 385 arch/x86/kvm/cpuid.c entry->ebx &= kvm_cpuid_7_0_ebx_x86_features; entry 386 arch/x86/kvm/cpuid.c cpuid_mask(&entry->ebx, CPUID_7_0_EBX); entry 388 arch/x86/kvm/cpuid.c entry->ebx |= F(TSC_ADJUST); entry 390 arch/x86/kvm/cpuid.c entry->ecx &= kvm_cpuid_7_0_ecx_x86_features; entry 391 arch/x86/kvm/cpuid.c f_la57 = entry->ecx & F(LA57); entry 392 arch/x86/kvm/cpuid.c cpuid_mask(&entry->ecx, CPUID_7_ECX); entry 394 arch/x86/kvm/cpuid.c entry->ecx |= f_la57; entry 395 arch/x86/kvm/cpuid.c entry->ecx |= f_umip; entry 396 arch/x86/kvm/cpuid.c entry->ecx |= f_pku; entry 399 arch/x86/kvm/cpuid.c entry->ecx &= ~F(PKU); entry 401 arch/x86/kvm/cpuid.c entry->edx &= kvm_cpuid_7_0_edx_x86_features; entry 402 arch/x86/kvm/cpuid.c cpuid_mask(&entry->edx, CPUID_7_EDX); entry 404 arch/x86/kvm/cpuid.c entry->edx |= F(SPEC_CTRL); entry 406 arch/x86/kvm/cpuid.c entry->edx |= F(INTEL_STIBP); entry 409 arch/x86/kvm/cpuid.c entry->edx |= F(SPEC_CTRL_SSBD); entry 414 arch/x86/kvm/cpuid.c entry->edx |= F(ARCH_CAPABILITIES); entry 417 arch/x86/kvm/cpuid.c entry->eax &= kvm_cpuid_7_1_eax_x86_features; entry 418 arch/x86/kvm/cpuid.c entry->ebx = 0; entry 419 arch/x86/kvm/cpuid.c entry->ecx = 0; entry 420 arch/x86/kvm/cpuid.c entry->edx = 0; entry 424 arch/x86/kvm/cpuid.c entry->eax = 0; entry 425 arch/x86/kvm/cpuid.c entry->ebx = 0; entry 426 arch/x86/kvm/cpuid.c entry->ecx = 0; entry 427 arch/x86/kvm/cpuid.c entry->edx = 0; entry 432 arch/x86/kvm/cpuid.c static inline int __do_cpuid_func(struct kvm_cpuid_entry2 *entry, u32 function, entry 513 arch/x86/kvm/cpuid.c do_host_cpuid(entry, function, 0); entry 519 arch/x86/kvm/cpuid.c entry->eax = min(entry->eax, 0x1fU); entry 522 arch/x86/kvm/cpuid.c entry->edx &= kvm_cpuid_1_edx_x86_features; entry 523 arch/x86/kvm/cpuid.c cpuid_mask(&entry->edx, CPUID_1_EDX); entry 524 arch/x86/kvm/cpuid.c entry->ecx &= kvm_cpuid_1_ecx_x86_features; entry 525 arch/x86/kvm/cpuid.c cpuid_mask(&entry->ecx, CPUID_1_ECX); entry 528 arch/x86/kvm/cpuid.c entry->ecx |= F(X2APIC); entry 535 arch/x86/kvm/cpuid.c int t, times = entry->eax & 0xff; entry 537 arch/x86/kvm/cpuid.c entry->flags |= KVM_CPUID_FLAG_STATE_READ_NEXT; entry 542 arch/x86/kvm/cpuid.c do_host_cpuid(&entry[t], function, 0); entry 557 arch/x86/kvm/cpuid.c cache_type = entry[i - 1].eax & 0x1f; entry 560 arch/x86/kvm/cpuid.c do_host_cpuid(&entry[i], function, i); entry 566 arch/x86/kvm/cpuid.c entry->eax = 0x4; /* allow ARAT */ entry 567 arch/x86/kvm/cpuid.c entry->ebx = 0; entry 568 arch/x86/kvm/cpuid.c entry->ecx = 0; entry 569 arch/x86/kvm/cpuid.c entry->edx = 0; entry 576 arch/x86/kvm/cpuid.c do_cpuid_7_mask(&entry[i], i); entry 577 arch/x86/kvm/cpuid.c if (i == entry->eax) entry 583 arch/x86/kvm/cpuid.c do_host_cpuid(&entry[i], function, i); entry 613 arch/x86/kvm/cpuid.c entry->eax = eax.full; entry 614 arch/x86/kvm/cpuid.c entry->ebx = cap.events_mask; entry 615 arch/x86/kvm/cpuid.c entry->ecx = 0; entry 616 arch/x86/kvm/cpuid.c entry->edx = edx.full; entry 635 arch/x86/kvm/cpuid.c for (i = 1; entry[i - 1].ecx & 0xff00; ++i) { entry 639 arch/x86/kvm/cpuid.c do_host_cpuid(&entry[i], function, i); entry 648 arch/x86/kvm/cpuid.c entry->eax &= supported; entry 649 arch/x86/kvm/cpuid.c entry->ebx = xstate_required_size(supported, false); entry 650 arch/x86/kvm/cpuid.c entry->ecx = entry->ebx; entry 651 arch/x86/kvm/cpuid.c entry->edx &= supported >> 32; entry 660 arch/x86/kvm/cpuid.c do_host_cpuid(&entry[i], function, idx); entry 662 arch/x86/kvm/cpuid.c entry[i].eax &= kvm_cpuid_D_1_eax_x86_features; entry 663 arch/x86/kvm/cpuid.c cpuid_mask(&entry[i].eax, CPUID_D_1_EAX); entry 664 arch/x86/kvm/cpuid.c entry[i].ebx = 0; entry 665 arch/x86/kvm/cpuid.c if (entry[i].eax & (F(XSAVES)|F(XSAVEC))) entry 666 arch/x86/kvm/cpuid.c entry[i].ebx = entry 670 arch/x86/kvm/cpuid.c if (entry[i].eax == 0 || !(supported & mask)) entry 672 arch/x86/kvm/cpuid.c if (WARN_ON_ONCE(entry[i].ecx & 1)) entry 675 arch/x86/kvm/cpuid.c entry[i].ecx = 0; entry 676 arch/x86/kvm/cpuid.c entry[i].edx = 0; entry 684 arch/x86/kvm/cpuid.c int t, times = entry->eax; entry 692 arch/x86/kvm/cpuid.c do_host_cpuid(&entry[t], function, t); entry 700 arch/x86/kvm/cpuid.c entry->eax = KVM_CPUID_FEATURES; entry 701 arch/x86/kvm/cpuid.c entry->ebx = sigptr[0]; entry 702 arch/x86/kvm/cpuid.c entry->ecx = sigptr[1]; entry 703 arch/x86/kvm/cpuid.c entry->edx = sigptr[2]; entry 707 arch/x86/kvm/cpuid.c entry->eax = (1 << KVM_FEATURE_CLOCKSOURCE) | entry 721 arch/x86/kvm/cpuid.c entry->eax |= (1 << KVM_FEATURE_STEAL_TIME); entry 723 arch/x86/kvm/cpuid.c entry->ebx = 0; entry 724 arch/x86/kvm/cpuid.c entry->ecx = 0; entry 725 arch/x86/kvm/cpuid.c entry->edx = 0; entry 728 arch/x86/kvm/cpuid.c entry->eax = min(entry->eax, 0x8000001f); entry 731 arch/x86/kvm/cpuid.c entry->edx &= kvm_cpuid_8000_0001_edx_x86_features; entry 732 arch/x86/kvm/cpuid.c cpuid_mask(&entry->edx, CPUID_8000_0001_EDX); entry 733 arch/x86/kvm/cpuid.c entry->ecx &= kvm_cpuid_8000_0001_ecx_x86_features; entry 734 arch/x86/kvm/cpuid.c cpuid_mask(&entry->ecx, CPUID_8000_0001_ECX); entry 738 arch/x86/kvm/cpuid.c entry->edx &= (1 << 8); entry 740 arch/x86/kvm/cpuid.c entry->edx &= boot_cpu_data.x86_power; entry 741 arch/x86/kvm/cpuid.c entry->eax = entry->ebx = entry->ecx = 0; entry 744 arch/x86/kvm/cpuid.c unsigned g_phys_as = (entry->eax >> 16) & 0xff; entry 745 arch/x86/kvm/cpuid.c unsigned virt_as = max((entry->eax >> 8) & 0xff, 48U); entry 746 arch/x86/kvm/cpuid.c unsigned phys_as = entry->eax & 0xff; entry 750 arch/x86/kvm/cpuid.c entry->eax = g_phys_as | (virt_as << 8); entry 751 arch/x86/kvm/cpuid.c entry->edx = 0; entry 752 arch/x86/kvm/cpuid.c entry->ebx &= kvm_cpuid_8000_0008_ebx_x86_features; entry 753 arch/x86/kvm/cpuid.c cpuid_mask(&entry->ebx, CPUID_8000_0008_EBX); entry 760 arch/x86/kvm/cpuid.c entry->ebx |= F(AMD_IBPB); entry 762 arch/x86/kvm/cpuid.c entry->ebx |= F(AMD_IBRS); entry 764 arch/x86/kvm/cpuid.c entry->ebx |= F(AMD_STIBP); entry 767 arch/x86/kvm/cpuid.c entry->ebx |= F(AMD_SSBD); entry 769 arch/x86/kvm/cpuid.c entry->ebx |= F(AMD_SSB_NO); entry 776 arch/x86/kvm/cpuid.c entry->ebx |= F(VIRT_SSBD); entry 780 arch/x86/kvm/cpuid.c entry->ecx = entry->edx = 0; entry 788 arch/x86/kvm/cpuid.c entry->eax = min(entry->eax, 0xC0000004); entry 791 arch/x86/kvm/cpuid.c entry->edx &= kvm_cpuid_C000_0001_edx_x86_features; entry 792 arch/x86/kvm/cpuid.c cpuid_mask(&entry->edx, CPUID_C000_0001_EDX); entry 800 arch/x86/kvm/cpuid.c entry->eax = entry->ebx = entry->ecx = entry->edx = 0; entry 804 arch/x86/kvm/cpuid.c kvm_x86_ops->set_supported_cpuid(function, entry); entry 814 arch/x86/kvm/cpuid.c static int do_cpuid_func(struct kvm_cpuid_entry2 *entry, u32 func, entry 821 arch/x86/kvm/cpuid.c return __do_cpuid_func_emulated(entry, func, nent, maxnent); entry 823 arch/x86/kvm/cpuid.c return __do_cpuid_func(entry, func, nent, maxnent); entry 1000 arch/x86/kvm/cpuid.c struct kvm_cpuid_entry2 *entry; entry 1004 arch/x86/kvm/cpuid.c entry = kvm_find_cpuid_entry(vcpu, function, index); entry 1005 arch/x86/kvm/cpuid.c found = entry; entry 1012 arch/x86/kvm/cpuid.c if (!entry && check_limit && !guest_cpuid_is_amd(vcpu) && entry 1017 arch/x86/kvm/cpuid.c entry = kvm_find_cpuid_entry(vcpu, function, index); entry 1020 arch/x86/kvm/cpuid.c if (entry) { entry 1021 arch/x86/kvm/cpuid.c *eax = entry->eax; entry 1022 arch/x86/kvm/cpuid.c *ebx = entry->ebx; entry 1023 arch/x86/kvm/cpuid.c *ecx = entry->ecx; entry 1024 arch/x86/kvm/cpuid.c *edx = entry->edx; entry 1035 arch/x86/kvm/cpuid.c entry = kvm_find_cpuid_entry(vcpu, function, 1); entry 1036 arch/x86/kvm/cpuid.c if (entry) { entry 1038 arch/x86/kvm/cpuid.c *edx = entry->edx; entry 70 arch/x86/kvm/cpuid.h struct kvm_cpuid_entry2 *entry; entry 73 arch/x86/kvm/cpuid.h entry = kvm_find_cpuid_entry(vcpu, cpuid.function, cpuid.index); entry 74 arch/x86/kvm/cpuid.h if (!entry) entry 79 arch/x86/kvm/cpuid.h return &entry->eax; entry 81 arch/x86/kvm/cpuid.h return &entry->ebx; entry 83 arch/x86/kvm/cpuid.h return &entry->ecx; entry 85 arch/x86/kvm/cpuid.h return &entry->edx; entry 176 arch/x86/kvm/ioapic.c union kvm_ioapic_redirect_entry entry; entry 181 arch/x86/kvm/ioapic.c entry = ioapic->redirtbl[irq]; entry 182 arch/x86/kvm/ioapic.c edge = (entry.fields.trig_mode == IOAPIC_EDGE_TRIG); entry 220 arch/x86/kvm/ioapic.c trace_kvm_ioapic_set_irq(entry.bits, irq, ret == 0); entry 334 arch/x86/kvm/ioapic.c union kvm_ioapic_redirect_entry *entry = &ioapic->redirtbl[irq]; entry 338 arch/x86/kvm/ioapic.c if (entry->fields.mask || entry 339 arch/x86/kvm/ioapic.c (entry->fields.trig_mode == IOAPIC_LEVEL_TRIG && entry 340 arch/x86/kvm/ioapic.c entry->fields.remote_irr)) entry 343 arch/x86/kvm/ioapic.c irqe.dest_id = entry->fields.dest_id; entry 344 arch/x86/kvm/ioapic.c irqe.vector = entry->fields.vector; entry 345 arch/x86/kvm/ioapic.c irqe.dest_mode = entry->fields.dest_mode; entry 346 arch/x86/kvm/ioapic.c irqe.trig_mode = entry->fields.trig_mode; entry 347 arch/x86/kvm/ioapic.c irqe.delivery_mode = entry->fields.delivery_mode << 8; entry 370 arch/x86/kvm/ioapic.c entry->fields.remote_irr = 1; entry 401 arch/x86/kvm/irq_comm.c struct kvm_kernel_irq_routing_entry *entry; entry 411 arch/x86/kvm/irq_comm.c hlist_for_each_entry(entry, &table->map[i], link) { entry 414 arch/x86/kvm/irq_comm.c if (entry->type != KVM_IRQ_ROUTING_MSI) entry 417 arch/x86/kvm/irq_comm.c kvm_set_msi_irq(vcpu->kvm, entry, &irq); entry 5438 arch/x86/kvm/mmu.c u64 entry, gentry, *spte; entry 5483 arch/x86/kvm/mmu.c entry = *spte; entry 5489 arch/x86/kvm/mmu.c if (need_remote_flush(entry, *spte)) entry 493 arch/x86/kvm/svm.c u64 *entry = svm->avic_physical_id_cache; entry 495 arch/x86/kvm/svm.c if (!entry) entry 498 arch/x86/kvm/svm.c return (READ_ONCE(*entry) & AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK); entry 1743 arch/x86/kvm/svm.c u64 *entry, new_entry; entry 1760 arch/x86/kvm/svm.c entry = avic_get_physical_id_entry(vcpu, id); entry 1761 arch/x86/kvm/svm.c if (!entry) entry 1767 arch/x86/kvm/svm.c WRITE_ONCE(*entry, new_entry); entry 1769 arch/x86/kvm/svm.c svm->avic_physical_id_cache = entry; entry 2078 arch/x86/kvm/svm.c u64 entry; entry 2093 arch/x86/kvm/svm.c entry = READ_ONCE(*(svm->avic_physical_id_cache)); entry 2094 arch/x86/kvm/svm.c WARN_ON(entry & AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK); entry 2096 arch/x86/kvm/svm.c entry &= ~AVIC_PHYSICAL_ID_ENTRY_HOST_PHYSICAL_ID_MASK; entry 2097 arch/x86/kvm/svm.c entry |= (h_physical_id & AVIC_PHYSICAL_ID_ENTRY_HOST_PHYSICAL_ID_MASK); entry 2099 arch/x86/kvm/svm.c entry &= ~AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK; entry 2101 arch/x86/kvm/svm.c entry |= AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK; entry 2103 arch/x86/kvm/svm.c WRITE_ONCE(*(svm->avic_physical_id_cache), entry); entry 2110 arch/x86/kvm/svm.c u64 entry; entry 2116 arch/x86/kvm/svm.c entry = READ_ONCE(*(svm->avic_physical_id_cache)); entry 2117 arch/x86/kvm/svm.c if (entry & AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK) entry 2120 arch/x86/kvm/svm.c entry &= ~AVIC_PHYSICAL_ID_ENTRY_IS_RUNNING_MASK; entry 2121 arch/x86/kvm/svm.c WRITE_ONCE(*(svm->avic_physical_id_cache), entry); entry 4615 arch/x86/kvm/svm.c u32 *entry, new_entry; entry 4618 arch/x86/kvm/svm.c entry = avic_get_logical_id_entry(vcpu, ldr, flat); entry 4619 arch/x86/kvm/svm.c if (!entry) entry 4622 arch/x86/kvm/svm.c new_entry = READ_ONCE(*entry); entry 4626 arch/x86/kvm/svm.c WRITE_ONCE(*entry, new_entry); entry 4635 arch/x86/kvm/svm.c u32 *entry = avic_get_logical_id_entry(vcpu, svm->ldr_reg, flat); entry 4637 arch/x86/kvm/svm.c if (entry) entry 4638 arch/x86/kvm/svm.c clear_bit(AVIC_LOGICAL_ID_ENTRY_VALID_BIT, (unsigned long *)entry); entry 5958 arch/x86/kvm/svm.c static void svm_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry) entry 5963 arch/x86/kvm/svm.c entry->ecx &= ~bit(X86_FEATURE_X2APIC); entry 5967 arch/x86/kvm/svm.c entry->ecx |= (1 << 2); /* Set SVM bit */ entry 5972 arch/x86/kvm/svm.c entry->ebx |= F(VIRT_SSBD); entry 5975 arch/x86/kvm/svm.c entry->eax = 1; /* SVM revision 1 */ entry 5976 arch/x86/kvm/svm.c entry->ebx = 8; /* Lets support 8 ASIDs in case we add proper entry 5978 arch/x86/kvm/svm.c entry->ecx = 0; /* Reserved */ entry 5979 arch/x86/kvm/svm.c entry->edx = 0; /* Per default do not support any entry 5984 arch/x86/kvm/svm.c entry->edx |= F(NRIPS); entry 5988 arch/x86/kvm/svm.c entry->edx |= F(NPT); entry 5994 arch/x86/kvm/svm.c cpuid(0x8000001f, &entry->eax, &entry->ebx, entry 5995 arch/x86/kvm/svm.c &entry->ecx, &entry->edx); entry 78 arch/x86/kvm/vmx/nested.c struct shadow_vmcs_field entry = shadow_read_only_fields[i]; entry 79 arch/x86/kvm/vmx/nested.c u16 field = entry.encoding; entry 92 arch/x86/kvm/vmx/nested.c entry.offset += sizeof(u32); entry 94 arch/x86/kvm/vmx/nested.c shadow_read_only_fields[j++] = entry; entry 99 arch/x86/kvm/vmx/nested.c struct shadow_vmcs_field entry = shadow_read_write_fields[i]; entry 100 arch/x86/kvm/vmx/nested.c u16 field = entry.encoding; entry 140 arch/x86/kvm/vmx/nested.c entry.offset += sizeof(u32); entry 142 arch/x86/kvm/vmx/nested.c shadow_read_write_fields[j++] = entry; entry 274 arch/x86/kvm/vmx/pmu_intel.c struct kvm_cpuid_entry2 *entry; entry 285 arch/x86/kvm/vmx/pmu_intel.c entry = kvm_find_cpuid_entry(vcpu, 0xa, 0); entry 286 arch/x86/kvm/vmx/pmu_intel.c if (!entry) entry 288 arch/x86/kvm/vmx/pmu_intel.c eax.full = entry->eax; entry 289 arch/x86/kvm/vmx/pmu_intel.c edx.full = entry->edx; entry 300 arch/x86/kvm/vmx/pmu_intel.c pmu->available_event_types = ~entry->ebx & entry 323 arch/x86/kvm/vmx/pmu_intel.c entry = kvm_find_cpuid_entry(vcpu, 7, 0); entry 324 arch/x86/kvm/vmx/pmu_intel.c if (entry && entry 326 arch/x86/kvm/vmx/pmu_intel.c (entry->ebx & (X86_FEATURE_HLE|X86_FEATURE_RTM))) entry 809 arch/x86/kvm/vmx/vmx.c unsigned long entry, unsigned long exit) entry 811 arch/x86/kvm/vmx/vmx.c vm_entry_controls_clearbit(vmx, entry); entry 867 arch/x86/kvm/vmx/vmx.c unsigned long entry, unsigned long exit, entry 874 arch/x86/kvm/vmx/vmx.c vm_entry_controls_setbit(vmx, entry); entry 6221 arch/x86/kvm/vmx/vmx.c unsigned long entry; entry 6235 arch/x86/kvm/vmx/vmx.c entry = gate_offset(desc); entry 6255 arch/x86/kvm/vmx/vmx.c THUNK_TARGET(entry), entry 6941 arch/x86/kvm/vmx/vmx.c struct kvm_cpuid_entry2 *entry; entry 6947 arch/x86/kvm/vmx/vmx.c if (entry && (entry->_reg & (_cpuid_mask))) \ entry 6951 arch/x86/kvm/vmx/vmx.c entry = kvm_find_cpuid_entry(vcpu, 0x1, 0); entry 6967 arch/x86/kvm/vmx/vmx.c entry = kvm_find_cpuid_entry(vcpu, 0x7, 0); entry 7089 arch/x86/kvm/vmx/vmx.c static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry) entry 7092 arch/x86/kvm/vmx/vmx.c entry->ecx |= bit(X86_FEATURE_VMX); entry 961 arch/x86/mm/init.c void update_cache_mode_entry(unsigned entry, enum page_cache_mode cache) entry 964 arch/x86/mm/init.c BUG_ON(!entry && cache != _PAGE_CACHE_MODE_WB); entry 966 arch/x86/mm/init.c __cachemode2pte_tbl[cache] = __cm_idx2pte(entry); entry 967 arch/x86/mm/init.c __pte2cachemode_tbl[entry] = cache; entry 1483 arch/x86/mm/init_64.c pte_t entry; entry 1485 arch/x86/mm/init_64.c entry = pfn_pte(__pa(p) >> PAGE_SHIFT, entry 1487 arch/x86/mm/init_64.c set_pmd(pmd, __pmd(pte_val(entry))); entry 63 arch/x86/mm/kasan_init_64.c pte_t entry; entry 70 arch/x86/mm/kasan_init_64.c entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL); entry 71 arch/x86/mm/kasan_init_64.c set_pte_at(&init_mm, addr, pte, entry); entry 23 arch/x86/mm/mm_internal.h void update_cache_mode_entry(unsigned entry, enum page_cache_mode cache); entry 630 arch/x86/mm/pat.c struct memtype *entry; entry 653 arch/x86/mm/pat.c entry = rbt_memtype_erase(start, end); entry 656 arch/x86/mm/pat.c if (IS_ERR(entry)) { entry 662 arch/x86/mm/pat.c kfree(entry); entry 682 arch/x86/mm/pat.c struct memtype *entry; entry 696 arch/x86/mm/pat.c entry = rbt_memtype_lookup(paddr); entry 697 arch/x86/mm/pat.c if (entry != NULL) entry 698 arch/x86/mm/pat.c rettype = entry->type; entry 481 arch/x86/mm/pgtable.c pte_t entry, int dirty) entry 483 arch/x86/mm/pgtable.c int changed = !pte_same(*ptep, entry); entry 486 arch/x86/mm/pgtable.c set_pte(ptep, entry); entry 494 arch/x86/mm/pgtable.c pmd_t entry, int dirty) entry 496 arch/x86/mm/pgtable.c int changed = !pmd_same(*pmdp, entry); entry 501 arch/x86/mm/pgtable.c set_pmd(pmdp, entry); entry 514 arch/x86/mm/pgtable.c pud_t *pudp, pud_t entry, int dirty) entry 516 arch/x86/mm/pgtable.c int changed = !pud_same(*pudp, entry); entry 521 arch/x86/mm/pgtable.c set_pud(pudp, entry); entry 136 arch/x86/oprofile/op_model_amd.c struct op_entry entry; entry 145 arch/x86/oprofile/op_model_amd.c oprofile_write_reserve(&entry, regs, val, entry 147 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, val); entry 148 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, ctl); entry 150 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, val); entry 151 arch/x86/oprofile/op_model_amd.c oprofile_write_commit(&entry); entry 164 arch/x86/oprofile/op_model_amd.c oprofile_write_reserve(&entry, regs, val, IBS_OP_CODE, entry 166 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, val); entry 168 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, val); entry 170 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, val); entry 172 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, val); entry 174 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, val); entry 176 arch/x86/oprofile/op_model_amd.c oprofile_add_data64(&entry, val); entry 179 arch/x86/oprofile/op_model_amd.c oprofile_add_data(&entry, (unsigned long)val); entry 181 arch/x86/oprofile/op_model_amd.c oprofile_write_commit(&entry); entry 298 arch/x86/pci/acpi.c struct resource_entry *entry, *tmp; entry 303 arch/x86/pci/acpi.c resource_list_for_each_entry_safe(entry, tmp, &ci->resources) entry 304 arch/x86/pci/acpi.c if (resource_is_pcicfg_ioport(entry->res)) entry 305 arch/x86/pci/acpi.c resource_list_destroy_entry(entry); entry 309 arch/x86/pci/acpi.c resource_list_for_each_entry_safe(entry, tmp, &ci->resources) { entry 311 arch/x86/pci/acpi.c "host bridge window %pR (ignored)\n", entry->res); entry 312 arch/x86/pci/acpi.c resource_list_destroy_entry(entry); entry 120 arch/x86/pci/i386.c struct pcibios_fwaddrmap *entry, *next; entry 123 arch/x86/pci/i386.c list_for_each_entry_safe(entry, next, &pcibios_fwaddrmappings, list) { entry 124 arch/x86/pci/i386.c list_del(&entry->list); entry 125 arch/x86/pci/i386.c pci_dev_put(entry->dev); entry 126 arch/x86/pci/i386.c kfree(entry); entry 66 arch/x86/pci/pcbios.c unsigned long entry; /* 32 bit physical address */ entry 96 arch/x86/pci/pcbios.c unsigned long entry; /* %edx */ entry 104 arch/x86/pci/pcbios.c "=d" (entry) entry 112 arch/x86/pci/pcbios.c return address + entry; entry 324 arch/x86/pci/pcbios.c if (check->fields.entry >= 0x100000) { entry 329 arch/x86/pci/pcbios.c unsigned long bios32_entry = check->fields.entry; entry 710 arch/x86/platform/efi/efi.c static inline void *efi_map_next_entry_reverse(void *entry) entry 713 arch/x86/platform/efi/efi.c if (!entry) entry 716 arch/x86/platform/efi/efi.c entry -= efi.memmap.desc_size; entry 717 arch/x86/platform/efi/efi.c if (entry < efi.memmap.map) entry 720 arch/x86/platform/efi/efi.c return entry; entry 734 arch/x86/platform/efi/efi.c static void *efi_map_next_entry(void *entry) entry 754 arch/x86/platform/efi/efi.c return efi_map_next_entry_reverse(entry); entry 758 arch/x86/platform/efi/efi.c if (!entry) entry 761 arch/x86/platform/efi/efi.c entry += efi.memmap.desc_size; entry 762 arch/x86/platform/efi/efi.c if (entry >= efi.memmap.map_end) entry 765 arch/x86/platform/efi/efi.c return entry; entry 77 arch/x86/platform/intel-mid/device_libs/platform_msic.c struct sfi_device_table_entry *entry = info; entry 80 arch/x86/platform/intel-mid/device_libs/platform_msic.c msic_pdata.irq[block] = entry->irq; entry 30 arch/x86/platform/uv/uv_irq.c struct uv_IO_APIC_route_entry *entry; entry 36 arch/x86/platform/uv/uv_irq.c entry = (struct uv_IO_APIC_route_entry *)&mmr_value; entry 37 arch/x86/platform/uv/uv_irq.c entry->vector = cfg->vector; entry 38 arch/x86/platform/uv/uv_irq.c entry->delivery_mode = apic->irq_delivery_mode; entry 39 arch/x86/platform/uv/uv_irq.c entry->dest_mode = apic->irq_dest_mode; entry 40 arch/x86/platform/uv/uv_irq.c entry->polarity = 0; entry 41 arch/x86/platform/uv/uv_irq.c entry->trigger = 0; entry 42 arch/x86/platform/uv/uv_irq.c entry->mask = 0; entry 43 arch/x86/platform/uv/uv_irq.c entry->dest = cfg->dest_apicid; entry 140 arch/x86/platform/uv/uv_irq.c struct uv_IO_APIC_route_entry *entry; entry 143 arch/x86/platform/uv/uv_irq.c entry = (struct uv_IO_APIC_route_entry *)&mmr_value; entry 144 arch/x86/platform/uv/uv_irq.c entry->mask = 1; entry 585 arch/x86/xen/enlighten_pv.c u64 entry = *(u64 *)ptr; entry 587 arch/x86/xen/enlighten_pv.c trace_xen_cpu_write_ldt_entry(dt, entrynum, entry); entry 592 arch/x86/xen/enlighten_pv.c if (HYPERVISOR_update_descriptor(mach_lp.maddr, entry)) entry 646 arch/x86/xen/enlighten_pv.c struct trap_array_entry *entry = trap_array + nr; entry 648 arch/x86/xen/enlighten_pv.c if (*addr == entry->orig) { entry 649 arch/x86/xen/enlighten_pv.c *addr = entry->xen; entry 650 arch/x86/xen/enlighten_pv.c ist_okay = entry->ist_okay; entry 739 arch/x86/xen/enlighten_pv.c gate_desc *entry = (gate_desc *)(desc->address) + in; entry 741 arch/x86/xen/enlighten_pv.c if (cvt_gate_to_trap(in, entry, &traps[out])) entry 779 arch/x86/xen/enlighten_pv.c static void xen_write_gdt_entry(struct desc_struct *dt, int entry, entry 782 arch/x86/xen/enlighten_pv.c trace_xen_cpu_write_gdt_entry(dt, entry, desc, type); entry 793 arch/x86/xen/enlighten_pv.c xmaddr_t maddr = arbitrary_virt_to_machine(&dt[entry]); entry 809 arch/x86/xen/enlighten_pv.c static void __init xen_write_gdt_entry_boot(struct desc_struct *dt, int entry, entry 812 arch/x86/xen/enlighten_pv.c trace_xen_cpu_write_gdt_entry(dt, entry, desc, type); entry 821 arch/x86/xen/enlighten_pv.c xmaddr_t maddr = virt_to_machine(&dt[entry]); entry 824 arch/x86/xen/enlighten_pv.c dt[entry] = *(struct desc_struct *)desc; entry 202 arch/x86/xen/setup.c const struct e820_entry *entry = xen_e820_table.entries; entry 206 arch/x86/xen/setup.c for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { entry 210 arch/x86/xen/setup.c if (entry->type != E820_TYPE_RAM) entry 213 arch/x86/xen/setup.c e_pfn = PFN_DOWN(entry->addr + entry->size); entry 219 arch/x86/xen/setup.c s_pfn = PFN_UP(entry->addr); entry 459 arch/x86/xen/setup.c const struct e820_entry *entry = xen_e820_table.entries; entry 473 arch/x86/xen/setup.c for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { entry 474 arch/x86/xen/setup.c phys_addr_t end = entry->addr + entry->size; entry 475 arch/x86/xen/setup.c if (entry->type == E820_TYPE_RAM || i == xen_e820_table.nr_entries - 1) { entry 479 arch/x86/xen/setup.c if (entry->type == E820_TYPE_RAM) entry 480 arch/x86/xen/setup.c end_pfn = PFN_UP(entry->addr); entry 608 arch/x86/xen/setup.c struct e820_entry *entry = xen_e820_table.entries; entry 611 arch/x86/xen/setup.c for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { entry 612 arch/x86/xen/setup.c if (entry->type == E820_TYPE_UNUSABLE) entry 613 arch/x86/xen/setup.c entry->type = E820_TYPE_RAM; entry 619 arch/x86/xen/setup.c struct e820_entry *entry; entry 627 arch/x86/xen/setup.c entry = xen_e820_table.entries; entry 630 arch/x86/xen/setup.c if (entry->type == E820_TYPE_RAM && entry->addr <= start && entry 631 arch/x86/xen/setup.c (entry->addr + entry->size) >= end) entry 634 arch/x86/xen/setup.c entry++; entry 652 arch/x86/xen/setup.c struct e820_entry *entry = xen_e820_table.entries; entry 654 arch/x86/xen/setup.c for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++, entry++) { entry 655 arch/x86/xen/setup.c if (entry->type != E820_TYPE_RAM || entry->size < size) entry 657 arch/x86/xen/setup.c start = entry->addr; entry 662 arch/x86/xen/setup.c if (start + size > entry->addr + entry->size) entry 201 arch/xtensa/include/asm/asmmacro.h entry sp, (XTENSA_FRAME_SIZE_RESERVE + \ entry 391 arch/xtensa/include/asm/pgtable.h #define __swp_type(entry) (((entry).val >> 6) & 0x1f) entry 392 arch/xtensa/include/asm/pgtable.h #define __swp_offset(entry) ((entry).val >> 11) entry 93 arch/xtensa/include/asm/tlbflush.h static inline void invalidate_itlb_entry_no_isync (unsigned entry) entry 96 arch/xtensa/include/asm/tlbflush.h __asm__ __volatile__ ("iitlb %0\n" : : "a" (entry) ); entry 99 arch/xtensa/include/asm/tlbflush.h static inline void invalidate_dtlb_entry_no_isync (unsigned entry) entry 102 arch/xtensa/include/asm/tlbflush.h __asm__ __volatile__ ("idtlb %0\n" : : "a" (entry) ); entry 130 arch/xtensa/include/asm/tlbflush.h static inline void write_dtlb_entry (pte_t entry, int way) entry 133 arch/xtensa/include/asm/tlbflush.h : : "r" (way), "r" (entry) ); entry 136 arch/xtensa/include/asm/tlbflush.h static inline void write_itlb_entry (pte_t entry, int way) entry 139 arch/xtensa/include/asm/tlbflush.h : : "r" (way), "r" (entry) ); entry 323 arch/xtensa/kernel/perf_event.c struct perf_callchain_entry_ctx *entry = data; entry 325 arch/xtensa/kernel/perf_event.c perf_callchain_store(entry, frame->pc); entry 329 arch/xtensa/kernel/perf_event.c void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 332 arch/xtensa/kernel/perf_event.c xtensa_backtrace_kernel(regs, entry->max_stack, entry 333 arch/xtensa/kernel/perf_event.c callchain_trace, NULL, entry); entry 336 arch/xtensa/kernel/perf_event.c void perf_callchain_user(struct perf_callchain_entry_ctx *entry, entry 339 arch/xtensa/kernel/perf_event.c xtensa_backtrace_user(regs, entry->max_stack, entry 340 arch/xtensa/kernel/perf_event.c callchain_trace, entry); entry 34 arch/xtensa/kernel/syscall.c #define __SYSCALL(nr, entry, nargs)[nr] = (syscall_t)entry, entry 236 arch/xtensa/mm/fault.c const struct exception_table_entry *entry; entry 239 arch/xtensa/mm/fault.c if ((entry = search_exception_tables(regs->pc)) != NULL) { entry 241 arch/xtensa/mm/fault.c current->comm, regs->pc, entry->fixup); entry 243 arch/xtensa/mm/fault.c regs->pc = entry->fixup; entry 376 block/bfq-wf2q.c struct bfq_entity *entry; entry 382 block/bfq-wf2q.c entry = rb_entry(parent, struct bfq_entity, rb_node); entry 384 block/bfq-wf2q.c if (bfq_gt(entry->finish, entity->finish)) entry 1361 block/bfq-wf2q.c struct bfq_entity *entry, *first = NULL; entry 1365 block/bfq-wf2q.c entry = rb_entry(node, struct bfq_entity, rb_node); entry 1367 block/bfq-wf2q.c if (!bfq_gt(entry->start, vtime)) entry 1368 block/bfq-wf2q.c first = entry; entry 1371 block/bfq-wf2q.c entry = rb_entry(node->rb_left, entry 1373 block/bfq-wf2q.c if (!bfq_gt(entry->min_start, vtime)) { entry 68 block/bio.c unsigned int i, entry = -1; entry 76 block/bio.c if (!bslab->slab && entry == -1) entry 77 block/bio.c entry = i; entry 89 block/bio.c if (bio_slab_nr == bio_slab_max && entry == -1) { entry 99 block/bio.c if (entry == -1) entry 100 block/bio.c entry = bio_slab_nr++; entry 102 block/bio.c bslab = &bio_slabs[entry]; entry 104 block/bio.c snprintf(bslab->name, sizeof(bslab->name), "bio-%d", entry); entry 228 block/blk-integrity.c struct integrity_sysfs_entry *entry = entry 231 block/blk-integrity.c return entry->show(bi, page); entry 240 block/blk-integrity.c struct integrity_sysfs_entry *entry = entry 244 block/blk-integrity.c if (entry->store) entry 245 block/blk-integrity.c ret = entry->store(bi, page, count); entry 1135 block/blk-iocost.c list_del_init(&wq_entry->entry); entry 65 block/blk-mq-sysfs.c struct blk_mq_ctx_sysfs_entry *entry; entry 70 block/blk-mq-sysfs.c entry = container_of(attr, struct blk_mq_ctx_sysfs_entry, attr); entry 74 block/blk-mq-sysfs.c if (!entry->show) entry 80 block/blk-mq-sysfs.c res = entry->show(ctx, page); entry 88 block/blk-mq-sysfs.c struct blk_mq_ctx_sysfs_entry *entry; entry 93 block/blk-mq-sysfs.c entry = container_of(attr, struct blk_mq_ctx_sysfs_entry, attr); entry 97 block/blk-mq-sysfs.c if (!entry->store) entry 103 block/blk-mq-sysfs.c res = entry->store(ctx, page, length); entry 111 block/blk-mq-sysfs.c struct blk_mq_hw_ctx_sysfs_entry *entry; entry 116 block/blk-mq-sysfs.c entry = container_of(attr, struct blk_mq_hw_ctx_sysfs_entry, attr); entry 120 block/blk-mq-sysfs.c if (!entry->show) entry 126 block/blk-mq-sysfs.c res = entry->show(hctx, page); entry 135 block/blk-mq-sysfs.c struct blk_mq_hw_ctx_sysfs_entry *entry; entry 140 block/blk-mq-sysfs.c entry = container_of(attr, struct blk_mq_hw_ctx_sysfs_entry, attr); entry 144 block/blk-mq-sysfs.c if (!entry->store) entry 150 block/blk-mq-sysfs.c res = entry->store(hctx, page, length); entry 1094 block/blk-mq.c if (!list_empty(&wait->entry)) { entry 1097 block/blk-mq.c list_del_init(&wait->entry); entry 1136 block/blk-mq.c if (!list_empty_careful(&wait->entry)) entry 1143 block/blk-mq.c if (!list_empty(&wait->entry)) { entry 1169 block/blk-mq.c list_del_init(&wait->entry); entry 1347 block/blk-mq.c (no_tag && list_empty_careful(&hctx->dispatch_wait.entry))) entry 2393 block/blk-mq.c INIT_LIST_HEAD(&hctx->dispatch_wait.entry); entry 228 block/blk-rq-qos.c list_del_init(&curr->entry); entry 256 block/blk-rq-qos.c .entry = LIST_HEAD_INIT(data.wq.entry), entry 796 block/blk-sysfs.c struct queue_sysfs_entry *entry = to_queue(attr); entry 801 block/blk-sysfs.c if (!entry->show) entry 808 block/blk-sysfs.c res = entry->show(q, page); entry 817 block/blk-sysfs.c struct queue_sysfs_entry *entry = to_queue(attr); entry 821 block/blk-sysfs.c if (!entry->store) entry 830 block/blk-sysfs.c res = entry->store(q, page, length); entry 441 block/elevator.c struct elv_fs_entry *entry = to_elv(attr); entry 445 block/elevator.c if (!entry->show) entry 450 block/elevator.c error = e->type ? entry->show(e, page) : -ENOENT; entry 459 block/elevator.c struct elv_fs_entry *entry = to_elv(attr); entry 463 block/elevator.c if (!entry->store) entry 468 block/elevator.c error = e->type ? entry->store(e, page, length) : -ENOENT; entry 497 block/kyber-iosched.c INIT_LIST_HEAD(&khd->domain_wait[i].wait.entry); entry 719 block/kyber-iosched.c if (nr < 0 && list_empty_careful(&wait->wait.entry)) { entry 739 block/kyber-iosched.c if (nr >= 0 && !list_empty_careful(&wait->wait.entry)) { entry 947 block/kyber-iosched.c seq_printf(m, "%d\n", !list_empty_careful(&wait->entry)); \ entry 25 crypto/ablkcipher.c struct list_head entry; entry 44 crypto/ablkcipher.c list_for_each_entry_safe(p, tmp, &walk->buffers, entry) { entry 46 crypto/ablkcipher.c list_del(&p->entry); entry 56 crypto/ablkcipher.c list_add_tail(&p->entry, &walk->buffers); entry 35 crypto/skcipher.c struct list_head entry; entry 185 crypto/skcipher.c list_for_each_entry_safe(p, tmp, &walk->buffers, entry) { entry 204 crypto/skcipher.c list_del(&p->entry); entry 221 crypto/skcipher.c list_add_tail(&p->entry, &walk->buffers); entry 191 drivers/acpi/ac.c struct proc_dir_entry *entry = NULL; entry 203 drivers/acpi/ac.c entry = proc_create_single_data(ACPI_AC_FILE_STATE, S_IRUGO, entry 205 drivers/acpi/ac.c if (!entry) entry 174 drivers/acpi/acpi_video.c struct list_head entry; entry 204 drivers/acpi/acpi_video.c struct list_head entry; entry 1217 drivers/acpi/acpi_video.c list_add_tail(&data->entry, &video->video_device_list); entry 1241 drivers/acpi/acpi_video.c list_for_each_entry(dev, &video->video_device_list, entry) entry 1824 drivers/acpi/acpi_video.c list_for_each_entry(dev, &video->video_device_list, entry) { entry 1860 drivers/acpi/acpi_video.c list_for_each_entry(dev, &video->video_device_list, entry) { entry 1903 drivers/acpi/acpi_video.c list_for_each_entry(dev, &video->video_device_list, entry) entry 1964 drivers/acpi/acpi_video.c list_for_each_entry(dev, &video->video_device_list, entry) entry 1993 drivers/acpi/acpi_video.c list_for_each_entry(dev, &video->video_device_list, entry) entry 2007 drivers/acpi/acpi_video.c list_for_each_entry_safe(dev, next, &video->video_device_list, entry) { entry 2008 drivers/acpi/acpi_video.c list_del(&dev->entry); entry 2078 drivers/acpi/acpi_video.c list_add_tail(&video->entry, &video_bus_head); entry 2111 drivers/acpi/acpi_video.c list_del(&video->entry); entry 2241 drivers/acpi/acpi_video.c list_for_each_entry(video, &video_bus_head, entry) entry 57 drivers/acpi/apei/apei-base.c int __apei_exec_read_register(struct acpi_whea_header *entry, u64 *val) entry 61 drivers/acpi/apei/apei-base.c rc = apei_read(val, &entry->register_region); entry 64 drivers/acpi/apei/apei-base.c *val >>= entry->register_region.bit_offset; entry 65 drivers/acpi/apei/apei-base.c *val &= entry->mask; entry 71 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry) entry 76 drivers/acpi/apei/apei-base.c rc = __apei_exec_read_register(entry, &val); entry 86 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry) entry 90 drivers/acpi/apei/apei-base.c rc = apei_exec_read_register(ctx, entry); entry 93 drivers/acpi/apei/apei-base.c ctx->value = (ctx->value == entry->value); entry 99 drivers/acpi/apei/apei-base.c int __apei_exec_write_register(struct acpi_whea_header *entry, u64 val) entry 103 drivers/acpi/apei/apei-base.c val &= entry->mask; entry 104 drivers/acpi/apei/apei-base.c val <<= entry->register_region.bit_offset; entry 105 drivers/acpi/apei/apei-base.c if (entry->flags & APEI_EXEC_PRESERVE_REGISTER) { entry 107 drivers/acpi/apei/apei-base.c rc = apei_read(&valr, &entry->register_region); entry 110 drivers/acpi/apei/apei-base.c valr &= ~(entry->mask << entry->register_region.bit_offset); entry 113 drivers/acpi/apei/apei-base.c rc = apei_write(val, &entry->register_region); entry 119 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry) entry 121 drivers/acpi/apei/apei-base.c return __apei_exec_write_register(entry, ctx->value); entry 126 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry) entry 130 drivers/acpi/apei/apei-base.c ctx->value = entry->value; entry 131 drivers/acpi/apei/apei-base.c rc = apei_exec_write_register(ctx, entry); entry 138 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry) entry 153 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry; entry 167 drivers/acpi/apei/apei-base.c entry = &ctx->action_table[i]; entry 168 drivers/acpi/apei/apei-base.c if (entry->action != action) entry 171 drivers/acpi/apei/apei-base.c if (entry->instruction >= ctx->instructions || entry 172 drivers/acpi/apei/apei-base.c !ctx->ins_table[entry->instruction].run) { entry 175 drivers/acpi/apei/apei-base.c entry->instruction); entry 178 drivers/acpi/apei/apei-base.c run = ctx->ins_table[entry->instruction].run; entry 179 drivers/acpi/apei/apei-base.c rc = run(ctx, entry); entry 195 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry, entry 205 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry; entry 209 drivers/acpi/apei/apei-base.c entry = ctx->action_table + i; entry 210 drivers/acpi/apei/apei-base.c ins = entry->instruction; entry 219 drivers/acpi/apei/apei-base.c rc = func(ctx, entry, data); entry 228 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry, entry 231 drivers/acpi/apei/apei-base.c u8 ins = entry->instruction; entry 234 drivers/acpi/apei/apei-base.c return apei_map_generic_address(&entry->register_region); entry 261 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry, entry 264 drivers/acpi/apei/apei-base.c u8 ins = entry->instruction; entry 267 drivers/acpi/apei/apei-base.c apei_unmap_generic_address(&entry->register_region); entry 706 drivers/acpi/apei/apei-base.c struct acpi_whea_header *entry, entry 710 drivers/acpi/apei/apei-base.c struct acpi_generic_address *reg = &entry->register_region; entry 711 drivers/acpi/apei/apei-base.c u8 ins = entry->instruction; entry 16 drivers/acpi/apei/apei-internal.h struct acpi_whea_header *entry); entry 83 drivers/acpi/apei/apei-internal.h int __apei_exec_read_register(struct acpi_whea_header *entry, u64 *val); entry 84 drivers/acpi/apei/apei-internal.h int __apei_exec_write_register(struct acpi_whea_header *entry, u64 val); entry 86 drivers/acpi/apei/apei-internal.h struct acpi_whea_header *entry); entry 88 drivers/acpi/apei/apei-internal.h struct acpi_whea_header *entry); entry 90 drivers/acpi/apei/apei-internal.h struct acpi_whea_header *entry); entry 92 drivers/acpi/apei/apei-internal.h struct acpi_whea_header *entry); entry 94 drivers/acpi/apei/apei-internal.h struct acpi_whea_header *entry); entry 208 drivers/acpi/apei/einj.c struct acpi_whea_header *entry; entry 210 drivers/acpi/apei/einj.c entry = EINJ_TAB_ENTRY(einj_tab); entry 212 drivers/acpi/apei/einj.c if (entry->action == ACPI_EINJ_SET_ERROR_TYPE && entry 213 drivers/acpi/apei/einj.c entry->instruction == ACPI_EINJ_WRITE_REGISTER && entry 214 drivers/acpi/apei/einj.c entry->register_region.space_id == entry 216 drivers/acpi/apei/einj.c pa_v4 = get_unaligned(&entry->register_region.address); entry 217 drivers/acpi/apei/einj.c if (entry->action == ACPI_EINJ_SET_ERROR_TYPE_WITH_ADDRESS && entry 218 drivers/acpi/apei/einj.c entry->instruction == ACPI_EINJ_WRITE_REGISTER && entry 219 drivers/acpi/apei/einj.c entry->register_region.space_id == entry 221 drivers/acpi/apei/einj.c pa_v5 = get_unaligned(&entry->register_region.address); entry 222 drivers/acpi/apei/einj.c entry++; entry 270 drivers/acpi/apei/einj.c struct acpi_whea_header *entry; entry 272 drivers/acpi/apei/einj.c entry = (struct acpi_whea_header *) entry 275 drivers/acpi/apei/einj.c if (entry->action == ACPI_EINJ_TRIGGER_ERROR && entry 276 drivers/acpi/apei/einj.c entry->instruction <= ACPI_EINJ_WRITE_REGISTER_VALUE && entry 277 drivers/acpi/apei/einj.c entry->register_region.space_id == entry 279 drivers/acpi/apei/einj.c (entry->register_region.address & param2) == (param1 & param2)) entry 280 drivers/acpi/apei/einj.c return &entry->register_region; entry 281 drivers/acpi/apei/einj.c entry++; entry 113 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 115 drivers/acpi/apei/erst.c return __apei_exec_read_register(entry, &ctx->var1); entry 119 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 121 drivers/acpi/apei/erst.c return __apei_exec_read_register(entry, &ctx->var2); entry 125 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 127 drivers/acpi/apei/erst.c return __apei_exec_write_register(entry, ctx->var1); entry 131 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 138 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 145 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 150 drivers/acpi/apei/erst.c rc = __apei_exec_read_register(entry, &val); entry 154 drivers/acpi/apei/erst.c rc = __apei_exec_write_register(entry, val); entry 159 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 164 drivers/acpi/apei/erst.c rc = __apei_exec_read_register(entry, &val); entry 168 drivers/acpi/apei/erst.c rc = __apei_exec_write_register(entry, val); entry 173 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 190 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 207 drivers/acpi/apei/erst.c rc = __apei_exec_read_register(entry, &val); entry 220 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 225 drivers/acpi/apei/erst.c rc = __apei_exec_read_register(entry, &val); entry 237 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 244 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 246 drivers/acpi/apei/erst.c return __apei_exec_read_register(entry, &ctx->src_base); entry 250 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 252 drivers/acpi/apei/erst.c return __apei_exec_read_register(entry, &ctx->dst_base); entry 256 drivers/acpi/apei/erst.c struct acpi_whea_header *entry) entry 268 drivers/acpi/apei/erst.c rc = __apei_exec_read_register(entry, &offset); entry 277 drivers/acpi/button.c struct proc_dir_entry *entry = NULL; entry 309 drivers/acpi/button.c entry = proc_create_single_data(ACPI_BUTTON_FILE_STATE, S_IRUGO, entry 312 drivers/acpi/button.c if (!entry) { entry 268 drivers/acpi/glue.c struct acpi_device_physical_node *entry; entry 275 drivers/acpi/glue.c list_for_each_entry(entry, &acpi_dev->physical_node_list, node) entry 276 drivers/acpi/glue.c if (entry->dev == dev) { entry 279 drivers/acpi/glue.c list_del(&entry->node); entry 282 drivers/acpi/glue.c acpi_physnode_link_name(physnode_name, entry->node_id); entry 289 drivers/acpi/glue.c kfree(entry); entry 161 drivers/acpi/hmat/hmat.c static u32 hmat_normalize(u16 entry, u64 base, u8 type) entry 168 drivers/acpi/hmat/hmat.c if (entry == 0xffff || !entry) entry 170 drivers/acpi/hmat/hmat.c else if (base > (UINT_MAX / (entry))) entry 177 drivers/acpi/hmat/hmat.c value = entry * base; entry 192 drivers/acpi/numa.c u8 val = slit->entry[d*i + j]; entry 237 drivers/acpi/numa.c slit->entry[slit->locality_count * i + j]); entry 95 drivers/acpi/nvs.c struct nvs_page *entry, *next; entry 103 drivers/acpi/nvs.c entry = kzalloc(sizeof(struct nvs_page), GFP_KERNEL); entry 104 drivers/acpi/nvs.c if (!entry) entry 107 drivers/acpi/nvs.c list_add_tail(&entry->node, &nvs_list); entry 108 drivers/acpi/nvs.c entry->phys_start = start; entry 110 drivers/acpi/nvs.c entry->size = (size < nr_bytes) ? size : nr_bytes; entry 112 drivers/acpi/nvs.c start += entry->size; entry 113 drivers/acpi/nvs.c size -= entry->size; entry 118 drivers/acpi/nvs.c list_for_each_entry_safe(entry, next, &nvs_list, node) { entry 119 drivers/acpi/nvs.c list_del(&entry->node); entry 120 drivers/acpi/nvs.c kfree(entry); entry 130 drivers/acpi/nvs.c struct nvs_page *entry; entry 132 drivers/acpi/nvs.c list_for_each_entry(entry, &nvs_list, node) entry 133 drivers/acpi/nvs.c if (entry->data) { entry 134 drivers/acpi/nvs.c free_page((unsigned long)entry->data); entry 135 drivers/acpi/nvs.c entry->data = NULL; entry 136 drivers/acpi/nvs.c if (entry->kaddr) { entry 137 drivers/acpi/nvs.c if (entry->unmap) { entry 138 drivers/acpi/nvs.c iounmap(entry->kaddr); entry 139 drivers/acpi/nvs.c entry->unmap = false; entry 141 drivers/acpi/nvs.c acpi_os_unmap_iomem(entry->kaddr, entry 142 drivers/acpi/nvs.c entry->size); entry 144 drivers/acpi/nvs.c entry->kaddr = NULL; entry 154 drivers/acpi/nvs.c struct nvs_page *entry; entry 156 drivers/acpi/nvs.c list_for_each_entry(entry, &nvs_list, node) { entry 157 drivers/acpi/nvs.c entry->data = (void *)__get_free_page(GFP_KERNEL); entry 158 drivers/acpi/nvs.c if (!entry->data) { entry 171 drivers/acpi/nvs.c struct nvs_page *entry; entry 175 drivers/acpi/nvs.c list_for_each_entry(entry, &nvs_list, node) entry 176 drivers/acpi/nvs.c if (entry->data) { entry 177 drivers/acpi/nvs.c unsigned long phys = entry->phys_start; entry 178 drivers/acpi/nvs.c unsigned int size = entry->size; entry 180 drivers/acpi/nvs.c entry->kaddr = acpi_os_get_iomem(phys, size); entry 181 drivers/acpi/nvs.c if (!entry->kaddr) { entry 182 drivers/acpi/nvs.c entry->kaddr = acpi_os_ioremap(phys, size); entry 183 drivers/acpi/nvs.c entry->unmap = !!entry->kaddr; entry 185 drivers/acpi/nvs.c if (!entry->kaddr) { entry 189 drivers/acpi/nvs.c memcpy(entry->data, entry->kaddr, entry->size); entry 203 drivers/acpi/nvs.c struct nvs_page *entry; entry 207 drivers/acpi/nvs.c list_for_each_entry(entry, &nvs_list, node) entry 208 drivers/acpi/nvs.c if (entry->data) entry 209 drivers/acpi/nvs.c memcpy(entry->kaddr, entry->data, entry->size); entry 112 drivers/acpi/pci_irq.c static void do_prt_fixups(struct acpi_prt_entry *entry, entry 123 drivers/acpi/pci_irq.c entry->id.segment == quirk->segment && entry 124 drivers/acpi/pci_irq.c entry->id.bus == quirk->bus && entry 125 drivers/acpi/pci_irq.c entry->id.device == quirk->device && entry 126 drivers/acpi/pci_irq.c entry->pin == quirk->pin && entry 132 drivers/acpi/pci_irq.c entry->id.segment, entry->id.bus, entry 133 drivers/acpi/pci_irq.c entry->id.device, pin_name(entry->pin), entry 147 drivers/acpi/pci_irq.c struct acpi_prt_entry *entry; entry 153 drivers/acpi/pci_irq.c entry = kzalloc(sizeof(struct acpi_prt_entry), GFP_KERNEL); entry 154 drivers/acpi/pci_irq.c if (!entry) entry 162 drivers/acpi/pci_irq.c entry->id.segment = segment; entry 163 drivers/acpi/pci_irq.c entry->id.bus = bus; entry 164 drivers/acpi/pci_irq.c entry->id.device = (prt->address >> 16) & 0xFFFF; entry 165 drivers/acpi/pci_irq.c entry->pin = prt->pin + 1; entry 167 drivers/acpi/pci_irq.c do_prt_fixups(entry, prt); entry 169 drivers/acpi/pci_irq.c entry->index = prt->source_index; entry 185 drivers/acpi/pci_irq.c acpi_get_handle(handle, prt->source, &entry->link); entry 197 drivers/acpi/pci_irq.c entry->id.segment, entry->id.bus, entry 198 drivers/acpi/pci_irq.c entry->id.device, pin_name(entry->pin), entry 199 drivers/acpi/pci_irq.c prt->source, entry->index)); entry 201 drivers/acpi/pci_irq.c *entry_ptr = entry; entry 211 drivers/acpi/pci_irq.c struct acpi_pci_routing_table *entry; entry 227 drivers/acpi/pci_irq.c entry = buffer.pointer; entry 228 drivers/acpi/pci_irq.c while (entry && (entry->length > 0)) { entry 230 drivers/acpi/pci_irq.c entry, entry_ptr)) entry 232 drivers/acpi/pci_irq.c entry = (struct acpi_pci_routing_table *) entry 233 drivers/acpi/pci_irq.c ((unsigned long)entry + entry->length); entry 268 drivers/acpi/pci_irq.c struct acpi_prt_entry *entry) entry 285 drivers/acpi/pci_irq.c "IRQ %d\n", entry->index, entry 286 drivers/acpi/pci_irq.c (entry->index % 4) + 16); entry 287 drivers/acpi/pci_irq.c entry->index = (entry->index % 4) + 16; entry 291 drivers/acpi/pci_irq.c "IRQ: unknown mapping\n", entry->index); entry 300 drivers/acpi/pci_irq.c struct acpi_prt_entry *entry = NULL; entry 305 drivers/acpi/pci_irq.c ret = acpi_pci_irq_find_prt_entry(dev, pin, &entry); entry 306 drivers/acpi/pci_irq.c if (!ret && entry) { entry 308 drivers/acpi/pci_irq.c acpi_reroute_boot_interrupt(dev, entry); entry 312 drivers/acpi/pci_irq.c return entry; entry 335 drivers/acpi/pci_irq.c ret = acpi_pci_irq_find_prt_entry(bridge, pin, &entry); entry 336 drivers/acpi/pci_irq.c if (!ret && entry) { entry 341 drivers/acpi/pci_irq.c return entry; entry 397 drivers/acpi/pci_irq.c struct acpi_prt_entry *entry; entry 425 drivers/acpi/pci_irq.c entry = acpi_pci_irq_lookup(dev, pin); entry 426 drivers/acpi/pci_irq.c if (!entry) { entry 436 drivers/acpi/pci_irq.c if (entry) { entry 437 drivers/acpi/pci_irq.c if (entry->link) entry 438 drivers/acpi/pci_irq.c gsi = acpi_pci_link_allocate_irq(entry->link, entry 439 drivers/acpi/pci_irq.c entry->index, entry 443 drivers/acpi/pci_irq.c gsi = entry->index; entry 453 drivers/acpi/pci_irq.c kfree(entry); entry 461 drivers/acpi/pci_irq.c kfree(entry); entry 469 drivers/acpi/pci_irq.c kfree(entry); entry 485 drivers/acpi/pci_irq.c kfree(entry); entry 491 drivers/acpi/pci_irq.c struct acpi_prt_entry *entry; entry 507 drivers/acpi/pci_irq.c entry = acpi_pci_irq_lookup(dev, pin); entry 508 drivers/acpi/pci_irq.c if (!entry) entry 511 drivers/acpi/pci_irq.c if (entry->link) entry 512 drivers/acpi/pci_irq.c gsi = acpi_pci_link_free_irq(entry->link); entry 514 drivers/acpi/pci_irq.c gsi = entry->index; entry 516 drivers/acpi/pci_irq.c kfree(entry); entry 151 drivers/acpi/pci_root.c struct pci_osc_bit_struct *entry; entry 154 drivers/acpi/pci_root.c for (i = 0, entry = table; i < size; i++, entry++) entry 155 drivers/acpi/pci_root.c if (word & entry->bit) entry 157 drivers/acpi/pci_root.c len ? " " : "", entry->desc); entry 681 drivers/acpi/pci_root.c struct resource_entry *tmp, *entry, *entry2; entry 687 drivers/acpi/pci_root.c resource_list_for_each_entry_safe(entry, tmp, &list) { entry 691 drivers/acpi/pci_root.c res1 = entry->res; entry 730 drivers/acpi/pci_root.c resource_list_del(entry); entry 732 drivers/acpi/pci_root.c resource_list_free_entry(entry); entry 734 drivers/acpi/pci_root.c resource_list_add_tail(entry, resources); entry 739 drivers/acpi/pci_root.c struct resource_entry *entry) entry 742 drivers/acpi/pci_root.c struct resource *res = entry->res; entry 744 drivers/acpi/pci_root.c resource_size_t pci_addr = cpu_addr - entry->offset; entry 757 drivers/acpi/pci_root.c entry->offset = port - pci_addr; entry 774 drivers/acpi/pci_root.c struct resource_entry *entry, *tmp; entry 788 drivers/acpi/pci_root.c resource_list_for_each_entry_safe(entry, tmp, list) { entry 789 drivers/acpi/pci_root.c if (entry->res->flags & IORESOURCE_IO) entry 791 drivers/acpi/pci_root.c entry); entry 793 drivers/acpi/pci_root.c if (entry->res->flags & IORESOURCE_DISABLED) entry 794 drivers/acpi/pci_root.c resource_list_destroy_entry(entry); entry 796 drivers/acpi/pci_root.c entry->res->name = info->name; entry 809 drivers/acpi/pci_root.c struct resource_entry *entry, *tmp; entry 812 drivers/acpi/pci_root.c resource_list_for_each_entry_safe(entry, tmp, &info->resources) { entry 813 drivers/acpi/pci_root.c res = entry->res; entry 833 drivers/acpi/pci_root.c resource_list_destroy_entry(entry); entry 841 drivers/acpi/pci_root.c struct resource_entry *entry, *tmp; entry 846 drivers/acpi/pci_root.c resource_list_for_each_entry_safe(entry, tmp, &info->resources) { entry 847 drivers/acpi/pci_root.c res = entry->res; entry 851 drivers/acpi/pci_root.c resource_list_destroy_entry(entry); entry 860 drivers/acpi/pci_root.c struct resource_entry *entry; entry 862 drivers/acpi/pci_root.c resource_list_for_each_entry(entry, &bridge->windows) { entry 863 drivers/acpi/pci_root.c res = entry->res; entry 94 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 99 drivers/acpi/power.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 100 drivers/acpi/power.c if (!entry) entry 103 drivers/acpi/power.c entry->resource = resource; entry 109 drivers/acpi/power.c list_add_tail(&entry->node, &e->node); entry 113 drivers/acpi/power.c list_add_tail(&entry->node, list); entry 119 drivers/acpi/power.c struct acpi_power_resource_entry *entry, *e; entry 121 drivers/acpi/power.c list_for_each_entry_safe(entry, e, list, node) { entry 122 drivers/acpi/power.c list_del(&entry->node); entry 123 drivers/acpi/power.c kfree(entry); entry 211 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 219 drivers/acpi/power.c list_for_each_entry(entry, list, node) { entry 220 drivers/acpi/power.c struct acpi_power_resource *resource = entry->resource; entry 307 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 315 drivers/acpi/power.c list_for_each_entry(entry, resources, node) { entry 316 drivers/acpi/power.c ret = acpi_power_resource_add_dependent(entry->resource, dev); entry 324 drivers/acpi/power.c list_for_each_entry(entry, resources, node) entry 325 drivers/acpi/power.c acpi_power_resource_remove_dependent(entry->resource, dev); entry 342 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 349 drivers/acpi/power.c list_for_each_entry_reverse(entry, resources, node) entry 350 drivers/acpi/power.c acpi_power_resource_remove_dependent(entry->resource, dev); entry 458 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 461 drivers/acpi/power.c list_for_each_entry_reverse(entry, list, node) { entry 462 drivers/acpi/power.c result = acpi_power_off(entry->resource); entry 469 drivers/acpi/power.c list_for_each_entry_continue(entry, list, node) entry 470 drivers/acpi/power.c acpi_power_on(entry->resource); entry 477 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 480 drivers/acpi/power.c list_for_each_entry(entry, list, node) { entry 481 drivers/acpi/power.c result = acpi_power_on(entry->resource); entry 488 drivers/acpi/power.c list_for_each_entry_continue_reverse(entry, list, node) entry 489 drivers/acpi/power.c acpi_power_off(entry->resource); entry 526 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 531 drivers/acpi/power.c list_for_each_entry_reverse(entry, resources, node) { entry 532 drivers/acpi/power.c struct acpi_device *res_dev = &entry->resource->device; entry 545 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 555 drivers/acpi/power.c list_for_each_entry(entry, resources, node) { entry 556 drivers/acpi/power.c struct acpi_device *res_dev = &entry->resource->device; entry 599 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 602 drivers/acpi/power.c list_for_each_entry(entry, list, node) { entry 603 drivers/acpi/power.c struct acpi_power_resource *resource = entry->resource; entry 701 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 712 drivers/acpi/power.c list_for_each_entry(entry, &dev->wakeup.resources, node) { entry 713 drivers/acpi/power.c struct acpi_power_resource *resource = entry->resource; entry 753 drivers/acpi/power.c struct acpi_power_resource_entry *entry; entry 775 drivers/acpi/power.c list_for_each_entry(entry, &dev->wakeup.resources, node) { entry 776 drivers/acpi/power.c struct acpi_power_resource *resource = entry->resource; entry 27 drivers/acpi/pptt.c struct acpi_subtable_header *entry; entry 36 drivers/acpi/pptt.c entry = ACPI_ADD_PTR(struct acpi_subtable_header, table_hdr, pptt_ref); entry 38 drivers/acpi/pptt.c if (entry->length == 0) entry 41 drivers/acpi/pptt.c if (pptt_ref + entry->length > table_hdr->length) entry 44 drivers/acpi/pptt.c return entry; entry 206 drivers/acpi/pptt.c struct acpi_subtable_header *entry; entry 217 drivers/acpi/pptt.c entry = ACPI_ADD_PTR(struct acpi_subtable_header, table_hdr, entry 221 drivers/acpi/pptt.c while ((unsigned long)entry + proc_sz < table_end) { entry 222 drivers/acpi/pptt.c cpu_node = (struct acpi_pptt_processor *)entry; entry 223 drivers/acpi/pptt.c if (entry->type == ACPI_PPTT_TYPE_PROCESSOR && entry 226 drivers/acpi/pptt.c if (entry->length == 0) entry 228 drivers/acpi/pptt.c entry = ACPI_ADD_PTR(struct acpi_subtable_header, entry, entry 229 drivers/acpi/pptt.c entry->length); entry 252 drivers/acpi/pptt.c struct acpi_subtable_header *entry; entry 258 drivers/acpi/pptt.c entry = ACPI_ADD_PTR(struct acpi_subtable_header, table_hdr, entry 263 drivers/acpi/pptt.c while ((unsigned long)entry + proc_sz < table_end) { entry 264 drivers/acpi/pptt.c cpu_node = (struct acpi_pptt_processor *)entry; entry 266 drivers/acpi/pptt.c if (entry->length == 0) { entry 270 drivers/acpi/pptt.c if (entry->type == ACPI_PPTT_TYPE_PROCESSOR && entry 273 drivers/acpi/pptt.c return (struct acpi_pptt_processor *)entry; entry 276 drivers/acpi/pptt.c entry = ACPI_ADD_PTR(struct acpi_subtable_header, entry, entry 277 drivers/acpi/pptt.c entry->length); entry 33 drivers/acpi/proc.c struct acpi_device_physical_node *entry; entry 51 drivers/acpi/proc.c list_for_each_entry(entry, &dev->physical_node_list, entry 53 drivers/acpi/proc.c ldev = get_device(entry->dev); entry 57 drivers/acpi/proc.c if (&entry->node != entry 80 drivers/acpi/proc.c struct acpi_device_physical_node *entry; entry 84 drivers/acpi/proc.c list_for_each_entry(entry, entry 86 drivers/acpi/proc.c if (entry->dev && device_can_wakeup(entry->dev)) { entry 87 drivers/acpi/proc.c bool enable = !device_may_wakeup(entry->dev); entry 88 drivers/acpi/proc.c device_set_wakeup_enable(entry->dev, enable); entry 35 drivers/acpi/processor_core.c static int map_lapic_id(struct acpi_subtable_header *entry, entry 39 drivers/acpi/processor_core.c container_of(entry, struct acpi_madt_local_apic, header); entry 51 drivers/acpi/processor_core.c static int map_x2apic_id(struct acpi_subtable_header *entry, entry 55 drivers/acpi/processor_core.c container_of(entry, struct acpi_madt_local_x2apic, header); entry 68 drivers/acpi/processor_core.c static int map_lsapic_id(struct acpi_subtable_header *entry, entry 72 drivers/acpi/processor_core.c container_of(entry, struct acpi_madt_local_sapic, header); entry 78 drivers/acpi/processor_core.c if ((entry->length < 16) || (lsapic->uid != acpi_id)) entry 90 drivers/acpi/processor_core.c static int map_gicc_mpidr(struct acpi_subtable_header *entry, entry 94 drivers/acpi/processor_core.c container_of(entry, struct acpi_madt_generic_interrupt, header); entry 115 drivers/acpi/processor_core.c unsigned long madt_end, entry; entry 121 drivers/acpi/processor_core.c entry = (unsigned long)madt; entry 122 drivers/acpi/processor_core.c madt_end = entry + madt->header.length; entry 126 drivers/acpi/processor_core.c entry += sizeof(struct acpi_table_madt); entry 127 drivers/acpi/processor_core.c while (entry + sizeof(struct acpi_subtable_header) < madt_end) { entry 129 drivers/acpi/processor_core.c (struct acpi_subtable_header *)entry; entry 143 drivers/acpi/processor_core.c entry += header->length; entry 268 drivers/acpi/processor_core.c static int get_ioapic_id(struct acpi_subtable_header *entry, u32 gsi_base, entry 271 drivers/acpi/processor_core.c struct acpi_madt_io_apic *ioapic = (struct acpi_madt_io_apic *)entry; entry 284 drivers/acpi/processor_core.c unsigned long madt_end, entry; entry 292 drivers/acpi/processor_core.c entry = (unsigned long)madt; entry 293 drivers/acpi/processor_core.c madt_end = entry + madt->header.length; entry 296 drivers/acpi/processor_core.c entry += sizeof(struct acpi_table_madt); entry 297 drivers/acpi/processor_core.c while (entry + sizeof(struct acpi_subtable_header) < madt_end) { entry 298 drivers/acpi/processor_core.c hdr = (struct acpi_subtable_header *)entry; entry 303 drivers/acpi/processor_core.c entry += hdr->length; entry 218 drivers/acpi/tables.c acpi_get_entry_type(struct acpi_subtable_entry *entry) entry 220 drivers/acpi/tables.c switch (entry->type) { entry 222 drivers/acpi/tables.c return entry->hdr->common.type; entry 224 drivers/acpi/tables.c return entry->hdr->hmat.type; entry 230 drivers/acpi/tables.c acpi_get_entry_length(struct acpi_subtable_entry *entry) entry 232 drivers/acpi/tables.c switch (entry->type) { entry 234 drivers/acpi/tables.c return entry->hdr->common.length; entry 236 drivers/acpi/tables.c return entry->hdr->hmat.length; entry 242 drivers/acpi/tables.c acpi_get_subtable_header_length(struct acpi_subtable_entry *entry) entry 244 drivers/acpi/tables.c switch (entry->type) { entry 246 drivers/acpi/tables.c return sizeof(entry->hdr->common); entry 248 drivers/acpi/tables.c return sizeof(entry->hdr->hmat); entry 289 drivers/acpi/tables.c struct acpi_subtable_entry entry; entry 313 drivers/acpi/tables.c entry.type = acpi_get_subtable_type(id); entry 314 drivers/acpi/tables.c entry.hdr = (union acpi_subtable_headers *) entry 316 drivers/acpi/tables.c subtable_len = acpi_get_subtable_header_length(&entry); entry 318 drivers/acpi/tables.c while (((unsigned long)entry.hdr) + subtable_len < table_end) { entry 323 drivers/acpi/tables.c if (acpi_get_entry_type(&entry) != proc[i].id) entry 326 drivers/acpi/tables.c (!errs && proc[i].handler(entry.hdr, table_end))) { entry 341 drivers/acpi/tables.c entry_len = acpi_get_entry_length(&entry); entry 347 drivers/acpi/tables.c entry.hdr = (union acpi_subtable_headers *) entry 348 drivers/acpi/tables.c ((unsigned long)entry.hdr + entry_len); entry 206 drivers/android/binder.c if (cur >= ARRAY_SIZE(log->entry)) entry 208 drivers/android/binder.c e = &log->entry[cur % ARRAY_SIZE(log->entry)]; entry 228 drivers/android/binder.c struct list_head entry; entry 800 drivers/android/binder.c BUG_ON(work->entry.next && !list_empty(&work->entry)); entry 801 drivers/android/binder.c list_add_tail(&work->entry, target_list); entry 862 drivers/android/binder.c list_del_init(&work->entry); entry 886 drivers/android/binder.c w = list_first_entry_or_null(list, struct binder_work, entry); entry 888 drivers/android/binder.c list_del_init(&w->entry); entry 1150 drivers/android/binder.c INIT_LIST_HEAD(&node->work.entry); entry 1220 drivers/android/binder.c if (!node->has_weak_ref && list_empty(&node->work.entry)) { entry 1271 drivers/android/binder.c if (list_empty(&node->work.entry)) { entry 1285 drivers/android/binder.c BUG_ON(!list_empty(&node->work.entry)); entry 2997 drivers/android/binder.c struct binder_work, entry); entry 3920 drivers/android/binder.c INIT_LIST_HEAD(&death->work.entry); entry 3952 drivers/android/binder.c if (list_empty(&death->work.entry)) { entry 3987 drivers/android/binder.c entry) { entry 5391 drivers/android/binder.c BUG_ON(!list_empty(&ref->death->work.entry)); entry 5657 drivers/android/binder.c list_for_each_entry(w, &thread->todo, entry) { entry 5688 drivers/android/binder.c list_for_each_entry(w, &node->async_todo, entry) entry 5762 drivers/android/binder.c list_for_each_entry(w, &proc->todo, entry) entry 5765 drivers/android/binder.c list_for_each_entry(w, &proc->delivered_death, entry) { entry 5923 drivers/android/binder.c list_for_each_entry(w, &proc->todo, entry) { entry 6054 drivers/android/binder.c cur = count < ARRAY_SIZE(log->entry) && !log->full ? entry 6055 drivers/android/binder.c 0 : count % ARRAY_SIZE(log->entry); entry 6056 drivers/android/binder.c if (count > ARRAY_SIZE(log->entry) || log->full) entry 6057 drivers/android/binder.c count = ARRAY_SIZE(log->entry); entry 6059 drivers/android/binder.c unsigned int index = cur++ % ARRAY_SIZE(log->entry); entry 6061 drivers/android/binder.c print_binder_transaction_log_entry(m, &log->entry[index]); entry 52 drivers/android/binder_alloc.c return list_entry(buffer->entry.next, struct binder_buffer, entry); entry 57 drivers/android/binder_alloc.c return list_entry(buffer->entry.prev, struct binder_buffer, entry); entry 63 drivers/android/binder_alloc.c if (list_is_last(&buffer->entry, &alloc->buffers)) entry 474 drivers/android/binder_alloc.c list_add(&new_buffer->entry, &buffer->entry); entry 551 drivers/android/binder_alloc.c BUG_ON(alloc->buffers.next == &buffer->entry); entry 562 drivers/android/binder_alloc.c if (!list_is_last(&buffer->entry, &alloc->buffers)) { entry 590 drivers/android/binder_alloc.c list_del(&buffer->entry); entry 630 drivers/android/binder_alloc.c if (!list_is_last(&buffer->entry, &alloc->buffers)) { entry 638 drivers/android/binder_alloc.c if (alloc->buffers.next != &buffer->entry) { entry 714 drivers/android/binder_alloc.c list_add(&buffer->entry, &alloc->buffers); entry 762 drivers/android/binder_alloc.c struct binder_buffer, entry); entry 765 drivers/android/binder_alloc.c list_del(&buffer->entry); entry 39 drivers/android/binder_alloc.h struct list_head entry; /* free and allocated entries by address */ entry 141 drivers/android/binder_internal.h struct binder_transaction_log_entry entry[32]; entry 3424 drivers/ata/libata-scsi.c u64 entry = sector | entry 3426 drivers/ata/libata-scsi.c buf[i++] = __cpu_to_le64(entry); entry 37 drivers/atm/adummy.c struct list_head entry; entry 177 drivers/atm/adummy.c list_add(&adummy_dev->entry, &adummy_devs); entry 192 drivers/atm/adummy.c list_for_each_entry_safe(adummy_dev, next, &adummy_devs, entry) { entry 559 drivers/atm/fore200e.c struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; entry 574 drivers/atm/fore200e.c fore200e->bus->write(prom_dma, &entry->cp_entry->cmd.prom_block.prom_haddr); entry 576 drivers/atm/fore200e.c *entry->status = STATUS_PENDING; entry 578 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&opcode, (u32 __iomem *)&entry->cp_entry->cmd.prom_block.opcode); entry 580 drivers/atm/fore200e.c ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400); entry 582 drivers/atm/fore200e.c *entry->status = STATUS_FREE; entry 773 drivers/atm/fore200e.c struct host_txq_entry* entry; entry 782 drivers/atm/fore200e.c entry = &txq->host_entry[ txq->tail ]; entry 784 drivers/atm/fore200e.c if ((*entry->status & STATUS_COMPLETE) == 0) { entry 789 drivers/atm/fore200e.c entry, txq->tail, entry->vc_map, entry->skb); entry 792 drivers/atm/fore200e.c kfree(entry->data); entry 795 drivers/atm/fore200e.c dma_unmap_single(fore200e->dev, entry->tpd->tsd[ 0 ].buffer, entry->tpd->tsd[ 0 ].length, entry 798 drivers/atm/fore200e.c vc_map = entry->vc_map; entry 807 drivers/atm/fore200e.c dev_kfree_skb_any(entry->skb); entry 813 drivers/atm/fore200e.c if (vc_map->incarn != entry->incarn) { entry 830 drivers/atm/fore200e.c dev_kfree_skb_any(entry->skb); entry 838 drivers/atm/fore200e.c vcc->pop(vcc, entry->skb); entry 841 drivers/atm/fore200e.c dev_kfree_skb_any(entry->skb); entry 845 drivers/atm/fore200e.c if (*entry->status & STATUS_ERROR) entry 852 drivers/atm/fore200e.c *entry->status = STATUS_FREE; entry 909 drivers/atm/fore200e.c struct host_bsq_entry* entry; entry 925 drivers/atm/fore200e.c entry = &bsq->host_entry[ bsq->head ]; entry 944 drivers/atm/fore200e.c entry->rbd_block->rbd[ i ].buffer_haddr = buffer->data.dma_addr; entry 945 drivers/atm/fore200e.c entry->rbd_block->rbd[ i ].handle = FORE200E_BUF2HDL(buffer); entry 953 drivers/atm/fore200e.c *entry->status = STATUS_PENDING; entry 954 drivers/atm/fore200e.c fore200e->bus->write(entry->rbd_block_dma, &entry->cp_entry->rbd_block_haddr); entry 1090 drivers/atm/fore200e.c struct host_rxq_entry* entry; entry 1096 drivers/atm/fore200e.c entry = &rxq->host_entry[ rxq->head ]; entry 1099 drivers/atm/fore200e.c if ((*entry->status & STATUS_COMPLETE) == 0) entry 1102 drivers/atm/fore200e.c vc_map = FORE200E_VC_MAP(fore200e, entry->rpd->atm_header.vpi, entry->rpd->atm_header.vci); entry 1109 drivers/atm/fore200e.c entry->rpd->atm_header.vpi, entry->rpd->atm_header.vci); entry 1115 drivers/atm/fore200e.c if ((*entry->status & STATUS_ERROR) == 0) { entry 1117 drivers/atm/fore200e.c fore200e_push_rpd(fore200e, vcc, entry->rpd); entry 1122 drivers/atm/fore200e.c entry->rpd->atm_header.vpi, entry->rpd->atm_header.vci); entry 1129 drivers/atm/fore200e.c fore200e_collect_rpd(fore200e, entry->rpd); entry 1132 drivers/atm/fore200e.c fore200e->bus->write(entry->rpd_dma, &entry->cp_entry->rpd_haddr); entry 1133 drivers/atm/fore200e.c *entry->status = STATUS_FREE; entry 1228 drivers/atm/fore200e.c struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; entry 1253 drivers/atm/fore200e.c *entry->status = STATUS_PENDING; entry 1261 drivers/atm/fore200e.c fore200e->bus->write(mtu, &entry->cp_entry->cmd.activate_block.mtu); entry 1262 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&vpvc, (u32 __iomem *)&entry->cp_entry->cmd.activate_block.vpvc); entry 1263 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&activ_opcode, (u32 __iomem *)&entry->cp_entry->cmd.activate_block.opcode); entry 1266 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&vpvc, (u32 __iomem *)&entry->cp_entry->cmd.deactivate_block.vpvc); entry 1267 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&deactiv_opcode, (u32 __iomem *)&entry->cp_entry->cmd.deactivate_block.opcode); entry 1270 drivers/atm/fore200e.c ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400); entry 1272 drivers/atm/fore200e.c *entry->status = STATUS_FREE; entry 1473 drivers/atm/fore200e.c struct host_txq_entry* entry; entry 1561 drivers/atm/fore200e.c entry = &txq->host_entry[ txq->head ]; entry 1563 drivers/atm/fore200e.c if ((*entry->status != STATUS_FREE) || (txq->txing >= QUEUE_SIZE_TX - 2)) { entry 1568 drivers/atm/fore200e.c if (*entry->status != STATUS_FREE) { entry 1597 drivers/atm/fore200e.c entry->incarn = vc_map->incarn; entry 1598 drivers/atm/fore200e.c entry->vc_map = vc_map; entry 1599 drivers/atm/fore200e.c entry->skb = skb; entry 1600 drivers/atm/fore200e.c entry->data = tx_copy ? data : NULL; entry 1602 drivers/atm/fore200e.c tpd = entry->tpd; entry 1657 drivers/atm/fore200e.c tpd_haddr.haddr = entry->tpd_dma >> TPD_HADDR_SHIFT; /* shift the address, as we are in a bitfield */ entry 1659 drivers/atm/fore200e.c *entry->status = STATUS_PENDING; entry 1660 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&tpd_haddr, (u32 __iomem *)&entry->cp_entry->tpd_haddr); entry 1672 drivers/atm/fore200e.c struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; entry 1693 drivers/atm/fore200e.c fore200e->bus->write(stats_dma_addr, &entry->cp_entry->cmd.stats_block.stats_haddr); entry 1695 drivers/atm/fore200e.c *entry->status = STATUS_PENDING; entry 1697 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&opcode, (u32 __iomem *)&entry->cp_entry->cmd.stats_block.opcode); entry 1699 drivers/atm/fore200e.c ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400); entry 1701 drivers/atm/fore200e.c *entry->status = STATUS_FREE; entry 1743 drivers/atm/fore200e.c struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; entry 1757 drivers/atm/fore200e.c fore200e->bus->write(oc3_regs_dma_addr, &entry->cp_entry->cmd.oc3_block.regs_haddr); entry 1759 drivers/atm/fore200e.c *entry->status = STATUS_PENDING; entry 1761 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&opcode, (u32*)&entry->cp_entry->cmd.oc3_block.opcode); entry 1763 drivers/atm/fore200e.c ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400); entry 1765 drivers/atm/fore200e.c *entry->status = STATUS_FREE; entry 1783 drivers/atm/fore200e.c struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; entry 1796 drivers/atm/fore200e.c fore200e->bus->write(0, &entry->cp_entry->cmd.oc3_block.regs_haddr); entry 1798 drivers/atm/fore200e.c *entry->status = STATUS_PENDING; entry 1800 drivers/atm/fore200e.c fore200e->bus->write(*(u32*)&opcode, (u32 __iomem *)&entry->cp_entry->cmd.oc3_block.opcode); entry 1802 drivers/atm/fore200e.c ok = fore200e_poll(fore200e, entry->status, STATUS_COMPLETE, 400); entry 1804 drivers/atm/fore200e.c *entry->status = STATUS_FREE; entry 833 drivers/atm/fore200e.h struct list_head entry; /* next device */ entry 825 drivers/atm/he.c list_add(&heb->entry, &he_dev->rbpl_outstanding); entry 894 drivers/atm/he.c list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry) entry 1573 drivers/atm/he.c list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry) entry 1684 drivers/atm/he.c list_del(&heb->entry); entry 1699 drivers/atm/he.c list_move_tail(&heb->entry, &he_vcc->buffers); entry 1736 drivers/atm/he.c list_for_each_entry(heb, &he_vcc->buffers, entry) entry 1777 drivers/atm/he.c list_for_each_entry_safe(heb, next, &he_vcc->buffers, entry) entry 1822 drivers/atm/he.c list_for_each_entry(__tpd, &he_dev->outstanding_tpds, entry) { entry 1825 drivers/atm/he.c list_del(&__tpd->entry); entry 1913 drivers/atm/he.c list_add(&heb->entry, &he_dev->rbpl_outstanding); entry 2106 drivers/atm/he.c list_add_tail(&tpd->entry, &he_dev->outstanding_tpds); entry 136 drivers/atm/he.h struct list_head entry; entry 224 drivers/atm/he.h struct list_head entry; entry 1223 drivers/atm/horizon.c rx_queue_entry = rd_mem (dev, &dev->rx_q_entry->entry); entry 321 drivers/atm/horizon.h HDW entry; entry 115 drivers/atm/iphase.c IARTN_Q *entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 116 drivers/atm/iphase.c if (!entry) entry 118 drivers/atm/iphase.c entry->data = data; entry 119 drivers/atm/iphase.c entry->next = NULL; entry 121 drivers/atm/iphase.c que->next = que->tail = entry; entry 123 drivers/atm/iphase.c que->tail->next = entry; entry 1545 drivers/base/core.c list_move_tail(&deva->kobj.entry, &devb->kobj.entry); entry 1561 drivers/base/core.c list_move(&deva->kobj.entry, &devb->kobj.entry); entry 1575 drivers/base/core.c list_move_tail(&dev->kobj.entry, &devices_kset->list); entry 1815 drivers/base/core.c list_for_each_entry(k, &dev->class->p->glue_dirs.list, entry) entry 3193 drivers/base/core.c kobj.entry); entry 3206 drivers/base/core.c list_del_init(&dev->kobj.entry); entry 19 drivers/base/devres.c struct list_head entry; entry 109 drivers/base/devres.c INIT_LIST_HEAD(&dr->node.entry); entry 117 drivers/base/devres.c BUG_ON(!list_empty(&node->entry)); entry 118 drivers/base/devres.c list_add_tail(&node->entry, &dev->devres_head); entry 190 drivers/base/devres.c &dev->devres_head, entry) { entry 214 drivers/base/devres.c BUG_ON(!list_empty(&dr->node.entry)); entry 245 drivers/base/devres.c list_for_each_entry_reverse(node, &dev->devres_head, entry) { entry 347 drivers/base/devres.c list_del_init(&dr->node.entry); entry 436 drivers/base/devres.c node = list_entry(cur, struct devres_node, entry); entry 446 drivers/base/devres.c if (&node->entry == first) entry 448 drivers/base/devres.c list_move_tail(&node->entry, todo); entry 467 drivers/base/devres.c node = list_entry(cur, struct devres_node, entry); entry 471 drivers/base/devres.c BUG_ON(!grp || list_empty(&grp->node[0].entry)); entry 474 drivers/base/devres.c if (list_empty(&grp->node[1].entry)) entry 482 drivers/base/devres.c list_move_tail(&grp->node[0].entry, todo); entry 483 drivers/base/devres.c list_del_init(&grp->node[1].entry); entry 505 drivers/base/devres.c list_for_each_entry_safe_reverse(dr, tmp, &todo, node.entry) { entry 557 drivers/base/devres.c INIT_LIST_HEAD(&grp->node[0].entry); entry 558 drivers/base/devres.c INIT_LIST_HEAD(&grp->node[1].entry); entry 577 drivers/base/devres.c list_for_each_entry_reverse(node, &dev->devres_head, entry) { entry 588 drivers/base/devres.c } else if (list_empty(&grp->node[1].entry)) entry 638 drivers/base/devres.c list_del_init(&grp->node[0].entry); entry 639 drivers/base/devres.c list_del_init(&grp->node[1].entry); entry 672 drivers/base/devres.c struct list_head *first = &grp->node[0].entry; entry 675 drivers/base/devres.c if (!list_empty(&grp->node[1].entry)) entry 676 drivers/base/devres.c end = grp->node[1].entry.next; entry 102 drivers/base/power/main.c INIT_LIST_HEAD(&dev->power.entry); entry 138 drivers/base/power/main.c list_add_tail(&dev->power.entry, &dpm_list); entry 156 drivers/base/power/main.c list_del_init(&dev->power.entry); entry 175 drivers/base/power/main.c list_move_tail(&deva->power.entry, &devb->power.entry); entry 189 drivers/base/power/main.c list_move(&deva->power.entry, &devb->power.entry); entry 200 drivers/base/power/main.c list_move_tail(&dev->power.entry, &dpm_list); entry 762 drivers/base/power/main.c list_for_each_entry(dev, &dpm_noirq_list, power.entry) entry 768 drivers/base/power/main.c list_move_tail(&dev->power.entry, &dpm_late_early_list); entry 912 drivers/base/power/main.c list_for_each_entry(dev, &dpm_late_early_list, power.entry) entry 918 drivers/base/power/main.c list_move_tail(&dev->power.entry, &dpm_suspended_list); entry 1073 drivers/base/power/main.c list_for_each_entry(dev, &dpm_suspended_list, power.entry) entry 1094 drivers/base/power/main.c if (!list_empty(&dev->power.entry)) entry 1095 drivers/base/power/main.c list_move_tail(&dev->power.entry, &dpm_prepared_list); entry 1172 drivers/base/power/main.c list_move(&dev->power.entry, &list); entry 1414 drivers/base/power/main.c if (!list_empty(&dev->power.entry)) entry 1415 drivers/base/power/main.c list_move(&dev->power.entry, &dpm_noirq_list); entry 1608 drivers/base/power/main.c if (!list_empty(&dev->power.entry)) entry 1609 drivers/base/power/main.c list_move(&dev->power.entry, &dpm_late_early_list); entry 1886 drivers/base/power/main.c if (!list_empty(&dev->power.entry)) entry 1887 drivers/base/power/main.c list_move(&dev->power.entry, &dpm_suspended_list); entry 2030 drivers/base/power/main.c if (!list_empty(&dev->power.entry)) entry 2031 drivers/base/power/main.c list_move_tail(&dev->power.entry, &dpm_prepared_list); entry 2097 drivers/base/power/main.c list_for_each_entry(dev, &dpm_list, power.entry) entry 102 drivers/base/power/power.h static inline struct device *to_device(struct list_head *entry) entry 104 drivers/base/power/power.h return container_of(entry, struct device, power.entry); entry 197 drivers/base/power/trace.c struct list_head *entry; entry 200 drivers/base/power/trace.c entry = dpm_list.prev; entry 201 drivers/base/power/trace.c while (entry != &dpm_list) { entry 202 drivers/base/power/trace.c struct device * dev = to_device(entry); entry 208 drivers/base/power/trace.c entry = entry->prev; entry 220 drivers/base/power/trace.c struct list_head *entry; entry 227 drivers/base/power/trace.c entry = dpm_list.prev; entry 228 drivers/base/power/trace.c while (size && entry != &dpm_list) { entry 229 drivers/base/power/trace.c struct device *dev = to_device(entry); entry 241 drivers/base/power/trace.c entry = entry->prev; entry 180 drivers/base/power/wakeup.c list_add_rcu(&ws->entry, &wakeup_sources); entry 197 drivers/base/power/wakeup.c list_del_rcu(&ws->entry); entry 356 drivers/base/power/wakeup.c list_for_each_entry_rcu(ws, &wakeup_sources, entry) entry 372 drivers/base/power/wakeup.c list_for_each_entry_rcu(ws, &wakeup_sources, entry) entry 432 drivers/base/power/wakeup.c if (device_is_registered(dev) && !list_empty(&dev->power.entry)) { entry 825 drivers/base/power/wakeup.c list_for_each_entry_rcu(ws, &wakeup_sources, entry) { entry 976 drivers/base/power/wakeup.c list_for_each_entry_rcu(ws, &wakeup_sources, entry) { entry 1055 drivers/base/power/wakeup.c list_for_each_entry_rcu(ws, &wakeup_sources, entry) { entry 1071 drivers/base/power/wakeup.c list_for_each_entry_continue_rcu(ws, &wakeup_sources, entry) { entry 367 drivers/base/regmap/regmap-debugfs.c char *entry; entry 378 drivers/base/regmap/regmap-debugfs.c entry = kmalloc(PAGE_SIZE, GFP_KERNEL); entry 379 drivers/base/regmap/regmap-debugfs.c if (!entry) { entry 397 drivers/base/regmap/regmap-debugfs.c entry_len = snprintf(entry, PAGE_SIZE, "%x-%x\n", entry 402 drivers/base/regmap/regmap-debugfs.c memcpy(buf + buf_pos, entry, entry_len); entry 409 drivers/base/regmap/regmap-debugfs.c kfree(entry); entry 22 drivers/base/swnode.c struct list_head entry; entry 62 drivers/base/swnode.c list_for_each_entry(k, &swnode_kset->list, entry) { entry 537 drivers/base/swnode.c (c && list_is_last(&c->entry, &p->children))) entry 541 drivers/base/swnode.c c = list_next_entry(c, entry); entry 543 drivers/base/swnode.c c = list_first_entry(&p->children, struct swnode, entry); entry 557 drivers/base/swnode.c list_for_each_entry(child, &swnode->children, entry) { entry 647 drivers/base/swnode.c list_for_each_entry(k, &swnode_kset->list, entry) { entry 684 drivers/base/swnode.c list_del(&swnode->entry); entry 730 drivers/base/swnode.c INIT_LIST_HEAD(&swnode->entry); entry 747 drivers/base/swnode.c list_add_tail(&swnode->entry, &parent->children); entry 315 drivers/block/ps3vram.c static void ps3vram_cache_evict(struct ps3_system_bus_device *dev, int entry) entry 320 drivers/block/ps3vram.c if (!(cache->tags[entry].flags & CACHE_PAGE_DIRTY)) entry 323 drivers/block/ps3vram.c dev_dbg(&dev->core, "Flushing %d: 0x%08x\n", entry, entry 324 drivers/block/ps3vram.c cache->tags[entry].address); entry 325 drivers/block/ps3vram.c if (ps3vram_upload(dev, CACHE_OFFSET + entry * cache->page_size, entry 326 drivers/block/ps3vram.c cache->tags[entry].address, DMA_PAGE_SIZE, entry 330 drivers/block/ps3vram.c entry * cache->page_size, cache->tags[entry].address, entry 333 drivers/block/ps3vram.c cache->tags[entry].flags &= ~CACHE_PAGE_DIRTY; entry 336 drivers/block/ps3vram.c static void ps3vram_cache_load(struct ps3_system_bus_device *dev, int entry, entry 342 drivers/block/ps3vram.c dev_dbg(&dev->core, "Fetching %d: 0x%08x\n", entry, address); entry 344 drivers/block/ps3vram.c CACHE_OFFSET + entry * cache->page_size, entry 349 drivers/block/ps3vram.c address, entry * cache->page_size, cache->page_size); entry 352 drivers/block/ps3vram.c cache->tags[entry].address = address; entry 353 drivers/block/ps3vram.c cache->tags[entry].flags |= CACHE_PAGE_PRESENT; entry 450 drivers/block/ps3vram.c unsigned int entry; entry 455 drivers/block/ps3vram.c entry = ps3vram_cache_match(dev, from); entry 456 drivers/block/ps3vram.c cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; entry 491 drivers/block/ps3vram.c unsigned int entry; entry 496 drivers/block/ps3vram.c entry = ps3vram_cache_match(dev, to); entry 497 drivers/block/ps3vram.c cached = CACHE_OFFSET + entry * priv->cache.page_size + offset; entry 507 drivers/block/ps3vram.c priv->cache.tags[entry].flags |= CACHE_PAGE_DIRTY; entry 590 drivers/block/zram/zram_drv.c unsigned long entry, struct bio *parent) entry 598 drivers/block/zram/zram_drv.c bio->bi_iter.bi_sector = entry * (PAGE_SIZE >> 9); entry 775 drivers/block/zram/zram_drv.c unsigned long entry; entry 785 drivers/block/zram/zram_drv.c unsigned long entry = zw->entry; entry 788 drivers/block/zram/zram_drv.c read_from_bdev_async(zram, &zw->bvec, entry, bio); entry 797 drivers/block/zram/zram_drv.c unsigned long entry, struct bio *bio) entry 803 drivers/block/zram/zram_drv.c work.entry = entry; entry 815 drivers/block/zram/zram_drv.c unsigned long entry, struct bio *bio) entry 823 drivers/block/zram/zram_drv.c unsigned long entry, struct bio *parent, bool sync) entry 827 drivers/block/zram/zram_drv.c return read_from_bdev_sync(zram, bvec, entry, parent); entry 829 drivers/block/zram/zram_drv.c return read_from_bdev_async(zram, bvec, entry, parent); entry 834 drivers/block/zram/zram_drv.c unsigned long entry, struct bio *parent, bool sync) entry 1409 drivers/bluetooth/btmrvl_sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; entry 1411 drivers/bluetooth/btmrvl_sdio.c if (entry->mem_ptr) { entry 1412 drivers/bluetooth/btmrvl_sdio.c vfree(entry->mem_ptr); entry 1413 drivers/bluetooth/btmrvl_sdio.c entry->mem_ptr = NULL; entry 1415 drivers/bluetooth/btmrvl_sdio.c entry->mem_size = 0; entry 1438 drivers/bluetooth/btmrvl_sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; entry 1467 drivers/bluetooth/btmrvl_sdio.c BT_INFO("%s_SIZE=0x%x", entry->mem_name, memory_size); entry 1468 drivers/bluetooth/btmrvl_sdio.c entry->mem_ptr = vzalloc(memory_size + 1); entry 1469 drivers/bluetooth/btmrvl_sdio.c entry->mem_size = memory_size; entry 1470 drivers/bluetooth/btmrvl_sdio.c if (!entry->mem_ptr) { entry 1471 drivers/bluetooth/btmrvl_sdio.c BT_ERR("Vzalloc %s failed", entry->mem_name); entry 1476 drivers/bluetooth/btmrvl_sdio.c strlen(entry->mem_name) + entry 1481 drivers/bluetooth/btmrvl_sdio.c dbg_ptr = entry->mem_ptr; entry 1484 drivers/bluetooth/btmrvl_sdio.c doneflag = entry->done_flag; entry 1486 drivers/bluetooth/btmrvl_sdio.c entry->mem_name); entry 1511 drivers/bluetooth/btmrvl_sdio.c entry->mem_name, entry 1512 drivers/bluetooth/btmrvl_sdio.c dbg_ptr - entry->mem_ptr); entry 1538 drivers/bluetooth/btmrvl_sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; entry 1540 drivers/bluetooth/btmrvl_sdio.c if (entry->mem_ptr) { entry 1544 drivers/bluetooth/btmrvl_sdio.c strcpy(fw_dump_ptr, entry->mem_name); entry 1545 drivers/bluetooth/btmrvl_sdio.c fw_dump_ptr += strlen(entry->mem_name); entry 1550 drivers/bluetooth/btmrvl_sdio.c memcpy(fw_dump_ptr, entry->mem_ptr, entry->mem_size); entry 1551 drivers/bluetooth/btmrvl_sdio.c fw_dump_ptr += entry->mem_size; entry 180 drivers/bluetooth/btqca.c le32_to_cpu(tlv_patch->entry)); entry 107 drivers/bluetooth/btqca.h __le32 entry; entry 734 drivers/bluetooth/btrtl.c struct rtl_vendor_config_entry *entry; entry 756 drivers/bluetooth/btrtl.c entry = ((void *)config->entry) + i; entry 758 drivers/bluetooth/btrtl.c switch (le16_to_cpu(entry->offset)) { entry 760 drivers/bluetooth/btrtl.c if (entry->len < sizeof(*device_baudrate)) { entry 765 drivers/bluetooth/btrtl.c *device_baudrate = get_unaligned_le32(entry->data); entry 769 drivers/bluetooth/btrtl.c if (entry->len >= 13) entry 770 drivers/bluetooth/btrtl.c *flow_control = !!(entry->data[12] & BIT(2)); entry 779 drivers/bluetooth/btrtl.c le16_to_cpu(entry->offset), entry->len); entry 783 drivers/bluetooth/btrtl.c i += sizeof(*entry) + entry->len; entry 47 drivers/bluetooth/btrtl.h struct rtl_vendor_config_entry entry[0]; entry 1055 drivers/bluetooth/hci_bcm.c struct resource_entry *entry; entry 1065 drivers/bluetooth/hci_bcm.c resource_list_for_each_entry(entry, &resources) { entry 1066 drivers/bluetooth/hci_bcm.c if (resource_type(entry->res) == IORESOURCE_IRQ) { entry 1067 drivers/bluetooth/hci_bcm.c dev->irq = entry->res->start; entry 558 drivers/bus/moxtet.c struct dentry *root, *entry; entry 565 drivers/bus/moxtet.c entry = debugfs_create_file_unsafe("input", 0444, root, moxtet, entry 567 drivers/bus/moxtet.c if (IS_ERR(entry)) entry 570 drivers/bus/moxtet.c entry = debugfs_create_file_unsafe("output", 0644, root, moxtet, entry 572 drivers/bus/moxtet.c if (IS_ERR(entry)) entry 580 drivers/bus/moxtet.c return PTR_ERR(entry); entry 991 drivers/cdrom/cdrom.c struct cdrom_tocentry entry; entry 1015 drivers/cdrom/cdrom.c entry.cdte_format = CDROM_MSF; entry 1017 drivers/cdrom/cdrom.c entry.cdte_track = i; entry 1018 drivers/cdrom/cdrom.c if (cdi->ops->audio_ioctl(cdi, CDROMREADTOCENTRY, &entry)) { entry 1022 drivers/cdrom/cdrom.c if (entry.cdte_ctrl & CDROM_DATA_TRACK) { entry 1023 drivers/cdrom/cdrom.c if (entry.cdte_format == 0x10) entry 1025 drivers/cdrom/cdrom.c else if (entry.cdte_format == 0x20) entry 1033 drivers/cdrom/cdrom.c i, entry.cdte_format, entry.cdte_ctrl); entry 2669 drivers/cdrom/cdrom.c struct cdrom_tocentry entry; entry 2675 drivers/cdrom/cdrom.c if (copy_from_user(&entry, argp, sizeof(entry))) entry 2678 drivers/cdrom/cdrom.c requested_format = entry.cdte_format; entry 2682 drivers/cdrom/cdrom.c entry.cdte_format = CDROM_MSF; entry 2683 drivers/cdrom/cdrom.c ret = cdi->ops->audio_ioctl(cdi, CDROMREADTOCENTRY, &entry); entry 2686 drivers/cdrom/cdrom.c sanitize_format(&entry.cdte_addr, &entry.cdte_format, requested_format); entry 2688 drivers/cdrom/cdrom.c if (copy_to_user(argp, &entry, sizeof(entry))) entry 91 drivers/cdrom/gdrom.c unsigned int entry[99]; entry 337 drivers/cdrom/gdrom.c data = gd.toc->entry[track - 1]; entry 66 drivers/char/agp/amd-k7-agp.c struct amd_page_map *entry; entry 70 drivers/char/agp/amd-k7-agp.c entry = tables[i]; entry 71 drivers/char/agp/amd-k7-agp.c if (entry != NULL) { entry 72 drivers/char/agp/amd-k7-agp.c if (entry->real != NULL) entry 73 drivers/char/agp/amd-k7-agp.c amd_free_page_map(entry); entry 74 drivers/char/agp/amd-k7-agp.c kfree(entry); entry 84 drivers/char/agp/amd-k7-agp.c struct amd_page_map *entry; entry 94 drivers/char/agp/amd-k7-agp.c entry = kzalloc(sizeof(struct amd_page_map), GFP_KERNEL); entry 95 drivers/char/agp/amd-k7-agp.c tables[i] = entry; entry 96 drivers/char/agp/amd-k7-agp.c if (entry == NULL) { entry 100 drivers/char/agp/amd-k7-agp.c retval = amd_create_page_map(entry); entry 89 drivers/char/agp/ati-agp.c struct ati_page_map *entry; entry 93 drivers/char/agp/ati-agp.c entry = tables[i]; entry 94 drivers/char/agp/ati-agp.c if (entry != NULL) { entry 95 drivers/char/agp/ati-agp.c if (entry->real != NULL) entry 96 drivers/char/agp/ati-agp.c ati_free_page_map(entry); entry 97 drivers/char/agp/ati-agp.c kfree(entry); entry 107 drivers/char/agp/ati-agp.c struct ati_page_map *entry; entry 117 drivers/char/agp/ati-agp.c entry = kzalloc(sizeof(struct ati_page_map), GFP_KERNEL); entry 118 drivers/char/agp/ati-agp.c tables[i] = entry; entry 119 drivers/char/agp/ati-agp.c if (entry == NULL) { entry 123 drivers/char/agp/ati-agp.c retval = ati_create_page_map(entry); entry 54 drivers/char/agp/intel-gtt.c void (*write_entry)(dma_addr_t addr, unsigned int entry, unsigned int flags); entry 319 drivers/char/agp/intel-gtt.c static void i810_write_entry(dma_addr_t addr, unsigned int entry, entry 333 drivers/char/agp/intel-gtt.c writel_relaxed(addr | pte_flags, intel_private.gtt + entry); entry 733 drivers/char/agp/intel-gtt.c static void i830_write_entry(dma_addr_t addr, unsigned int entry, entry 741 drivers/char/agp/intel-gtt.c writel_relaxed(addr | pte_flags, intel_private.gtt + entry); entry 1115 drivers/char/agp/intel-gtt.c unsigned int entry, entry 1126 drivers/char/agp/intel-gtt.c writel_relaxed(addr | pte_flags, intel_private.gtt + entry); entry 77 drivers/char/agp/sworks-agp.c struct serverworks_page_map *entry; entry 81 drivers/char/agp/sworks-agp.c entry = tables[i]; entry 82 drivers/char/agp/sworks-agp.c if (entry != NULL) { entry 83 drivers/char/agp/sworks-agp.c if (entry->real != NULL) { entry 84 drivers/char/agp/sworks-agp.c serverworks_free_page_map(entry); entry 86 drivers/char/agp/sworks-agp.c kfree(entry); entry 95 drivers/char/agp/sworks-agp.c struct serverworks_page_map *entry; entry 105 drivers/char/agp/sworks-agp.c entry = kzalloc(sizeof(struct serverworks_page_map), GFP_KERNEL); entry 106 drivers/char/agp/sworks-agp.c if (entry == NULL) { entry 110 drivers/char/agp/sworks-agp.c tables[i] = entry; entry 111 drivers/char/agp/sworks-agp.c retval = serverworks_create_page_map(entry); entry 204 drivers/char/ipmi/ipmi_devintf.c struct list_head *entry; entry 226 drivers/char/ipmi/ipmi_devintf.c entry = priv->recv_msgs.next; entry 227 drivers/char/ipmi/ipmi_devintf.c msg = list_entry(entry, struct ipmi_recv_msg, link); entry 228 drivers/char/ipmi/ipmi_devintf.c list_del(entry); entry 280 drivers/char/ipmi/ipmi_devintf.c list_add(entry, &priv->recv_msgs); entry 814 drivers/char/ipmi/ipmi_devintf.c struct ipmi_reg_list *entry; entry 816 drivers/char/ipmi/ipmi_devintf.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 817 drivers/char/ipmi/ipmi_devintf.c if (!entry) { entry 821 drivers/char/ipmi/ipmi_devintf.c entry->dev = dev; entry 825 drivers/char/ipmi/ipmi_devintf.c list_add(&entry->link, ®_list); entry 832 drivers/char/ipmi/ipmi_devintf.c struct ipmi_reg_list *entry; entry 835 drivers/char/ipmi/ipmi_devintf.c list_for_each_entry(entry, ®_list, link) { entry 836 drivers/char/ipmi/ipmi_devintf.c if (entry->dev == dev) { entry 837 drivers/char/ipmi/ipmi_devintf.c list_del(&entry->link); entry 838 drivers/char/ipmi/ipmi_devintf.c kfree(entry); entry 893 drivers/char/ipmi/ipmi_devintf.c struct ipmi_reg_list *entry, *entry2; entry 895 drivers/char/ipmi/ipmi_devintf.c list_for_each_entry_safe(entry, entry2, ®_list, link) { entry 896 drivers/char/ipmi/ipmi_devintf.c list_del(&entry->link); entry 897 drivers/char/ipmi/ipmi_devintf.c device_destroy(ipmi_class, entry->dev); entry 898 drivers/char/ipmi/ipmi_devintf.c kfree(entry); entry 3548 drivers/char/ipmi/ipmi_msghandler.c struct list_head *entry; entry 3569 drivers/char/ipmi/ipmi_msghandler.c entry = tmplist.next; entry 3570 drivers/char/ipmi/ipmi_msghandler.c list_del(entry); entry 3571 drivers/char/ipmi/ipmi_msghandler.c msg = list_entry(entry, struct ipmi_smi_msg, link); entry 4504 drivers/char/ipmi/ipmi_msghandler.c struct list_head *entry = NULL; entry 4508 drivers/char/ipmi/ipmi_msghandler.c entry = intf->hp_xmit_msgs.next; entry 4510 drivers/char/ipmi/ipmi_msghandler.c entry = intf->xmit_msgs.next; entry 4512 drivers/char/ipmi/ipmi_msghandler.c if (entry) { entry 4513 drivers/char/ipmi/ipmi_msghandler.c list_del(entry); entry 4514 drivers/char/ipmi/ipmi_msghandler.c newmsg = list_entry(entry, struct ipmi_smi_msg, link); entry 407 drivers/char/xillybus/xillybus_core.c int i, entry, rc; entry 482 drivers/char/xillybus/xillybus_core.c for (entry = 0; entry < entries; entry++, chandesc += 4) { entry 147 drivers/clk/bcm/clk-bcm63xx-gate.c const struct clk_bcm63xx_table_entry *entry, *table; entry 156 drivers/clk/bcm/clk-bcm63xx-gate.c for (entry = table; entry->name; entry++) entry 157 drivers/clk/bcm/clk-bcm63xx-gate.c maxbit = max_t(u8, maxbit, entry->bit); entry 176 drivers/clk/bcm/clk-bcm63xx-gate.c for (entry = table; entry->name; entry++) { entry 179 drivers/clk/bcm/clk-bcm63xx-gate.c clk = clk_hw_register_gate(&pdev->dev, entry->name, NULL, entry 180 drivers/clk/bcm/clk-bcm63xx-gate.c entry->flags, hw->regs, entry->bit, entry 187 drivers/clk/bcm/clk-bcm63xx-gate.c hw->data.hws[entry->bit] = clk; entry 133 drivers/clk/clk-max9485.c const struct max9485_rate *entry; entry 135 drivers/clk/clk-max9485.c for (entry = max9485_rates; entry->out != 0; entry++) entry 136 drivers/clk/clk-max9485.c if (entry->out == rate) entry 139 drivers/clk/clk-max9485.c if (entry->out == 0) entry 144 drivers/clk/clk-max9485.c entry->reg_value); entry 153 drivers/clk/clk-max9485.c const struct max9485_rate *entry; entry 155 drivers/clk/clk-max9485.c for (entry = max9485_rates; entry->out != 0; entry++) entry 156 drivers/clk/clk-max9485.c if (val == entry->reg_value) entry 157 drivers/clk/clk-max9485.c return entry->out; entry 422 drivers/clk/clk.c struct clk_parent_map *entry = &core->parents[index]; entry 425 drivers/clk/clk.c if (entry->hw) { entry 426 drivers/clk/clk.c parent = entry->hw->core; entry 436 drivers/clk/clk.c if (IS_ERR(parent) && PTR_ERR(parent) == -ENOENT && entry->name) entry 437 drivers/clk/clk.c parent = clk_core_lookup(entry->name); entry 442 drivers/clk/clk.c entry->core = parent; entry 65 drivers/clk/tegra/cvb.c const struct cvb_table_freq_entry *entry = &table->entries[i]; entry 67 drivers/clk/tegra/cvb.c if (!entry->freq || (entry->freq > max_freq)) entry 71 drivers/clk/tegra/cvb.c &entry->coefficients); entry 76 drivers/clk/tegra/cvb.c ret = dev_pm_opp_add(dev, entry->freq, dfll_mv * 1000); entry 134 drivers/clk/tegra/cvb.c const struct cvb_table_freq_entry *entry = &table->entries[i]; entry 136 drivers/clk/tegra/cvb.c if (!entry->freq || (entry->freq > max_freq)) entry 139 drivers/clk/tegra/cvb.c dev_pm_opp_remove(dev, entry->freq); entry 232 drivers/clk/ti/clkctrl.c struct omap_clkctrl_clk *entry; entry 241 drivers/clk/ti/clkctrl.c list_for_each_entry(entry, &provider->clocks, node) { entry 242 drivers/clk/ti/clkctrl.c if (entry->reg_offset == clkspec->args[0] && entry 243 drivers/clk/ti/clkctrl.c entry->bit_offset == clkspec->args[1]) { entry 252 drivers/clk/ti/clkctrl.c return entry->clk; entry 460 drivers/cpufreq/acpi-cpufreq.c struct cpufreq_frequency_table *entry; entry 471 drivers/cpufreq/acpi-cpufreq.c entry = &policy->freq_table[index]; entry 472 drivers/cpufreq/acpi-cpufreq.c next_freq = entry->frequency; entry 473 drivers/cpufreq/acpi-cpufreq.c next_perf_state = entry->driver_data; entry 165 drivers/cpufreq/powernow-k7.c static void invalidate_entry(unsigned int entry) entry 167 drivers/cpufreq/powernow-k7.c powernow_table[entry].frequency = CPUFREQ_ENTRY_INVALID; entry 538 drivers/cpufreq/powernow-k8.c unsigned int entry) entry 540 drivers/cpufreq/powernow-k8.c powernow_table[entry].frequency = CPUFREQ_ENTRY_INVALID; entry 1042 drivers/crypto/amcc/crypto4xx_core.c list_add_tail(&alg->entry, &sec_dev->alg_list); entry 1052 drivers/crypto/amcc/crypto4xx_core.c list_for_each_entry_safe(alg, tmp, &sec_dev->alg_list, entry) { entry 1053 drivers/crypto/amcc/crypto4xx_core.c list_del(&alg->entry); entry 144 drivers/crypto/amcc/crypto4xx_core.h struct list_head entry; entry 4503 drivers/crypto/caam/caamalg_qi2.c struct list_head entry; entry 5303 drivers/crypto/caam/caamalg_qi2.c list_add_tail(&t_alg->entry, &hash_list); entry 5322 drivers/crypto/caam/caamalg_qi2.c list_add_tail(&t_alg->entry, &hash_list); entry 5372 drivers/crypto/caam/caamalg_qi2.c list_for_each_entry_safe(t_hash_alg, p, &hash_list, entry) { entry 5374 drivers/crypto/caam/caamalg_qi2.c list_del(&t_hash_alg->entry); entry 1801 drivers/crypto/caam/caamhash.c struct list_head entry; entry 1931 drivers/crypto/caam/caamhash.c list_for_each_entry_safe(t_alg, n, &hash_list, entry) { entry 1933 drivers/crypto/caam/caamhash.c list_del(&t_alg->entry); entry 2044 drivers/crypto/caam/caamhash.c list_add_tail(&t_alg->entry, &hash_list); entry 2064 drivers/crypto/caam/caamhash.c list_add_tail(&t_alg->entry, &hash_list); entry 414 drivers/crypto/cavium/nitrox/nitrox_isr.c ndev->iov.msix.entry = NON_RING_MSIX_BASE; entry 157 drivers/crypto/ccp/ccp-crypto-aes-cmac.c INIT_LIST_HEAD(&rctx->cmd.entry); entry 361 drivers/crypto/ccp/ccp-crypto-aes-cmac.c INIT_LIST_HEAD(&ccp_alg->entry); entry 398 drivers/crypto/ccp/ccp-crypto-aes-cmac.c list_add(&ccp_alg->entry, head); entry 118 drivers/crypto/ccp/ccp-crypto-aes-galois.c INIT_LIST_HEAD(&rctx->cmd.entry); entry 221 drivers/crypto/ccp/ccp-crypto-aes-galois.c INIT_LIST_HEAD(&ccp_aead->entry); entry 242 drivers/crypto/ccp/ccp-crypto-aes-galois.c list_add(&ccp_aead->entry, head); entry 169 drivers/crypto/ccp/ccp-crypto-aes-xts.c INIT_LIST_HEAD(&rctx->cmd.entry); entry 238 drivers/crypto/ccp/ccp-crypto-aes-xts.c INIT_LIST_HEAD(&ccp_alg->entry); entry 270 drivers/crypto/ccp/ccp-crypto-aes-xts.c list_add(&ccp_alg->entry, head); entry 94 drivers/crypto/ccp/ccp-crypto-aes.c INIT_LIST_HEAD(&rctx->cmd.entry); entry 334 drivers/crypto/ccp/ccp-crypto-aes.c INIT_LIST_HEAD(&ccp_alg->entry); entry 355 drivers/crypto/ccp/ccp-crypto-aes.c list_add(&ccp_alg->entry, head); entry 88 drivers/crypto/ccp/ccp-crypto-des3.c INIT_LIST_HEAD(&rctx->cmd.entry); entry 197 drivers/crypto/ccp/ccp-crypto-des3.c INIT_LIST_HEAD(&ccp_alg->entry); entry 218 drivers/crypto/ccp/ccp-crypto-des3.c list_add(&ccp_alg->entry, head); entry 65 drivers/crypto/ccp/ccp-crypto-main.c struct list_head entry; entry 110 drivers/crypto/ccp/ccp-crypto-main.c list_for_each_entry_continue(tmp, &req_queue.cmds, entry) { entry 123 drivers/crypto/ccp/ccp-crypto-main.c if (req_queue.backlog == &crypto_cmd->entry) entry 124 drivers/crypto/ccp/ccp-crypto-main.c req_queue.backlog = crypto_cmd->entry.next; entry 127 drivers/crypto/ccp/ccp-crypto-main.c struct ccp_crypto_cmd, entry); entry 131 drivers/crypto/ccp/ccp-crypto-main.c if (req_queue.backlog == &crypto_cmd->entry) entry 132 drivers/crypto/ccp/ccp-crypto-main.c req_queue.backlog = crypto_cmd->entry.next; entry 137 drivers/crypto/ccp/ccp-crypto-main.c list_del(&crypto_cmd->entry); entry 232 drivers/crypto/ccp/ccp-crypto-main.c list_for_each_entry(tmp, &req_queue.cmds, entry) { entry 249 drivers/crypto/ccp/ccp-crypto-main.c req_queue.backlog = &crypto_cmd->entry; entry 254 drivers/crypto/ccp/ccp-crypto-main.c list_add_tail(&crypto_cmd->entry, &req_queue.cmds); entry 378 drivers/crypto/ccp/ccp-crypto-main.c list_for_each_entry_safe(ahash_alg, ahash_tmp, &hash_algs, entry) { entry 380 drivers/crypto/ccp/ccp-crypto-main.c list_del(&ahash_alg->entry); entry 384 drivers/crypto/ccp/ccp-crypto-main.c list_for_each_entry_safe(ablk_alg, ablk_tmp, &cipher_algs, entry) { entry 386 drivers/crypto/ccp/ccp-crypto-main.c list_del(&ablk_alg->entry); entry 390 drivers/crypto/ccp/ccp-crypto-main.c list_for_each_entry_safe(aead_alg, aead_tmp, &aead_algs, entry) { entry 392 drivers/crypto/ccp/ccp-crypto-main.c list_del(&aead_alg->entry); entry 396 drivers/crypto/ccp/ccp-crypto-main.c list_for_each_entry_safe(akc_alg, akc_tmp, &akcipher_algs, entry) { entry 398 drivers/crypto/ccp/ccp-crypto-main.c list_del(&akc_alg->entry); entry 72 drivers/crypto/ccp/ccp-crypto-rsa.c INIT_LIST_HEAD(&rctx->cmd.entry); entry 256 drivers/crypto/ccp/ccp-crypto-rsa.c INIT_LIST_HEAD(&ccp_alg->entry); entry 271 drivers/crypto/ccp/ccp-crypto-rsa.c list_add(&ccp_alg->entry, head); entry 132 drivers/crypto/ccp/ccp-crypto-sha.c INIT_LIST_HEAD(&rctx->cmd.entry); entry 432 drivers/crypto/ccp/ccp-crypto-sha.c INIT_LIST_HEAD(&ccp_alg->entry); entry 456 drivers/crypto/ccp/ccp-crypto-sha.c list_add(&ccp_alg->entry, head); entry 474 drivers/crypto/ccp/ccp-crypto-sha.c INIT_LIST_HEAD(&ccp_alg->entry); entry 513 drivers/crypto/ccp/ccp-crypto-sha.c list_add(&ccp_alg->entry, head); entry 35 drivers/crypto/ccp/ccp-crypto.h struct list_head entry; entry 43 drivers/crypto/ccp/ccp-crypto.h struct list_head entry; entry 51 drivers/crypto/ccp/ccp-crypto.h struct list_head entry; entry 64 drivers/crypto/ccp/ccp-crypto.h struct list_head entry; entry 556 drivers/crypto/ccp/ccp-dev-v3.c cmd = list_first_entry(&ccp->cmd, struct ccp_cmd, entry); entry 557 drivers/crypto/ccp/ccp-dev-v3.c list_del(&cmd->entry); entry 562 drivers/crypto/ccp/ccp-dev-v3.c cmd = list_first_entry(&ccp->backlog, struct ccp_cmd, entry); entry 563 drivers/crypto/ccp/ccp-dev-v3.c list_del(&cmd->entry); entry 1042 drivers/crypto/ccp/ccp-dev-v5.c cmd = list_first_entry(&ccp->cmd, struct ccp_cmd, entry); entry 1043 drivers/crypto/ccp/ccp-dev-v5.c list_del(&cmd->entry); entry 1048 drivers/crypto/ccp/ccp-dev-v5.c cmd = list_first_entry(&ccp->backlog, struct ccp_cmd, entry); entry 1049 drivers/crypto/ccp/ccp-dev-v5.c list_del(&cmd->entry); entry 140 drivers/crypto/ccp/ccp-dev.c list_add_tail(&ccp->entry, &ccp_units); entry 168 drivers/crypto/ccp/ccp-dev.c if (list_is_last(&ccp_rr->entry, &ccp_units)) entry 170 drivers/crypto/ccp/ccp-dev.c entry); entry 172 drivers/crypto/ccp/ccp-dev.c ccp_rr = list_next_entry(ccp_rr, entry); entry 174 drivers/crypto/ccp/ccp-dev.c list_del(&ccp->entry); entry 215 drivers/crypto/ccp/ccp-dev.c if (list_is_last(&ccp_rr->entry, &ccp_units)) entry 217 drivers/crypto/ccp/ccp-dev.c entry); entry 219 drivers/crypto/ccp/ccp-dev.c ccp_rr = list_next_entry(ccp_rr, entry); entry 259 drivers/crypto/ccp/ccp-dev.c dp = list_first_entry(&ccp_units, struct ccp_device, entry); entry 315 drivers/crypto/ccp/ccp-dev.c list_add_tail(&cmd->entry, &ccp->backlog); entry 322 drivers/crypto/ccp/ccp-dev.c list_add_tail(&cmd->entry, &ccp->cmd); entry 357 drivers/crypto/ccp/ccp-dev.c list_add_tail(&cmd->entry, &ccp->cmd); entry 397 drivers/crypto/ccp/ccp-dev.c cmd = list_first_entry(&ccp->cmd, struct ccp_cmd, entry); entry 398 drivers/crypto/ccp/ccp-dev.c list_del(&cmd->entry); entry 405 drivers/crypto/ccp/ccp-dev.c entry); entry 406 drivers/crypto/ccp/ccp-dev.c list_del(&backlog->entry); entry 227 drivers/crypto/ccp/ccp-dev.h struct list_head entry; entry 233 drivers/crypto/ccp/ccp-dev.h struct list_head entry; entry 340 drivers/crypto/ccp/ccp-dev.h struct list_head entry; entry 67 drivers/crypto/ccp/ccp-dmaengine.c list_for_each_entry_safe(cmd, ctmp, list, entry) { entry 68 drivers/crypto/ccp/ccp-dmaengine.c list_del(&cmd->entry); entry 78 drivers/crypto/ccp/ccp-dmaengine.c list_for_each_entry_safe(desc, dtmp, list, entry) { entry 82 drivers/crypto/ccp/ccp-dmaengine.c list_del(&desc->entry); entry 110 drivers/crypto/ccp/ccp-dmaengine.c list_for_each_entry_safe_reverse(desc, dtmp, list, entry) { entry 119 drivers/crypto/ccp/ccp-dmaengine.c list_del(&desc->entry); entry 144 drivers/crypto/ccp/ccp-dmaengine.c cmd = list_first_entry(&desc->pending, struct ccp_dma_cmd, entry); entry 145 drivers/crypto/ccp/ccp-dmaengine.c list_move(&cmd->entry, &desc->active); entry 165 drivers/crypto/ccp/ccp-dmaengine.c entry); entry 172 drivers/crypto/ccp/ccp-dmaengine.c list_del(&cmd->entry); entry 181 drivers/crypto/ccp/ccp-dmaengine.c list_move(&desc->entry, &chan->complete); entry 185 drivers/crypto/ccp/ccp-dmaengine.c entry); entry 253 drivers/crypto/ccp/ccp-dmaengine.c ? list_first_entry(&chan->pending, struct ccp_dma_desc, entry) entry 310 drivers/crypto/ccp/ccp-dmaengine.c list_del(&desc->entry); entry 311 drivers/crypto/ccp/ccp-dmaengine.c list_add_tail(&desc->entry, &chan->pending); entry 345 drivers/crypto/ccp/ccp-dmaengine.c INIT_LIST_HEAD(&desc->entry); entry 441 drivers/crypto/ccp/ccp-dmaengine.c list_add_tail(&cmd->entry, &desc->pending); entry 466 drivers/crypto/ccp/ccp-dmaengine.c list_add_tail(&desc->entry, &chan->created); entry 561 drivers/crypto/ccp/ccp-dmaengine.c list_for_each_entry(desc, &chan->complete, entry) { entry 600 drivers/crypto/ccp/ccp-dmaengine.c entry); entry 44 drivers/crypto/ccp/psp-dev.h struct list_head entry; entry 48 drivers/crypto/ccp/sp-dev.c list_add_tail(&sp->entry, &sp_units); entry 59 drivers/crypto/ccp/sp-dev.c list_del(&sp->entry); entry 251 drivers/crypto/ccp/sp-dev.c list_for_each_entry(i, &sp_units, entry) { entry 60 drivers/crypto/ccp/sp-dev.h struct list_head entry; entry 43 drivers/crypto/ccp/sp-pci.c sp_pci->msix_entry[v].entry = v; entry 2660 drivers/crypto/ccree/cc_aead.c entry) { entry 2662 drivers/crypto/ccree/cc_aead.c list_del(&t_alg->entry); entry 2718 drivers/crypto/ccree/cc_aead.c list_add_tail(&t_alg->entry, &aead_handle->aead_list); entry 33 drivers/crypto/ccree/cc_buffer_mgr.c union buffer_array_entry entry[MAX_NUM_OF_BUFFERS_IN_MLLI]; entry 213 drivers/crypto/ccree/cc_buffer_mgr.c union buffer_array_entry *entry = &sg_data->entry[i]; entry 218 drivers/crypto/ccree/cc_buffer_mgr.c rc = cc_render_sg_to_mlli(dev, entry->sgl, tot_len, entry 222 drivers/crypto/ccree/cc_buffer_mgr.c rc = cc_render_buff_to_mlli(dev, entry->buffer_dma, entry 260 drivers/crypto/ccree/cc_buffer_mgr.c sgl_data->entry[index].buffer_dma = buffer_dma; entry 281 drivers/crypto/ccree/cc_buffer_mgr.c sgl_data->entry[index].sgl = sgl; entry 1677 drivers/crypto/ccree/cc_cipher.c entry) { entry 1679 drivers/crypto/ccree/cc_cipher.c list_del(&t_alg->entry); entry 1733 drivers/crypto/ccree/cc_cipher.c list_add_tail(&t_alg->entry, entry 168 drivers/crypto/ccree/cc_driver.h struct list_head entry; entry 61 drivers/crypto/ccree/cc_hash.c struct list_head entry; entry 2068 drivers/crypto/ccree/cc_hash.c list_add_tail(&t_alg->entry, entry 2093 drivers/crypto/ccree/cc_hash.c list_add_tail(&t_alg->entry, &hash_handle->hash_list); entry 2112 drivers/crypto/ccree/cc_hash.c &hash_handle->hash_list, entry) { entry 2114 drivers/crypto/ccree/cc_hash.c list_del(&t_hash_alg->entry); entry 89 drivers/crypto/chelsio/chcr_core.c if (list_is_last(&drv_data.last_dev->entry, &drv_data.act_dev)) entry 91 drivers/crypto/chelsio/chcr_core.c struct uld_ctx, entry); entry 94 drivers/crypto/chelsio/chcr_core.c list_next_entry(drv_data.last_dev, entry); entry 108 drivers/crypto/chelsio/chcr_core.c list_move(&u_ctx->entry, &drv_data.act_dev); entry 127 drivers/crypto/chelsio/chcr_core.c list_add_tail(&u_ctx->entry, &drv_data.inact_dev); entry 137 drivers/crypto/chelsio/chcr_core.c if (list_is_last(&drv_data.last_dev->entry, &drv_data.act_dev)) entry 139 drivers/crypto/chelsio/chcr_core.c struct uld_ctx, entry); entry 142 drivers/crypto/chelsio/chcr_core.c list_next_entry(drv_data.last_dev, entry); entry 144 drivers/crypto/chelsio/chcr_core.c list_move(&u_ctx->entry, &drv_data.inact_dev); entry 306 drivers/crypto/chelsio/chcr_core.c list_for_each_entry_safe(u_ctx, tmp, &drv_data.act_dev, entry) { entry 307 drivers/crypto/chelsio/chcr_core.c list_del(&u_ctx->entry); entry 310 drivers/crypto/chelsio/chcr_core.c list_for_each_entry_safe(u_ctx, tmp, &drv_data.inact_dev, entry) { entry 311 drivers/crypto/chelsio/chcr_core.c list_del(&u_ctx->entry); entry 154 drivers/crypto/chelsio/chcr_core.h struct list_head entry; entry 598 drivers/crypto/hifn_795x.c struct list_head entry; entry 2406 drivers/crypto/hifn_795x.c list_add_tail(&alg->entry, &dev->alg_list); entry 2410 drivers/crypto/hifn_795x.c list_del(&alg->entry); entry 2421 drivers/crypto/hifn_795x.c list_for_each_entry_safe(a, n, &dev->alg_list, entry) { entry 2422 drivers/crypto/hifn_795x.c list_del(&a->entry); entry 92 drivers/crypto/n2_core.c struct list_head entry; entry 249 drivers/crypto/n2_core.c struct list_head entry; entry 674 drivers/crypto/n2_core.c struct list_head entry; entry 712 drivers/crypto/n2_core.c struct list_head entry; entry 898 drivers/crypto/n2_core.c INIT_LIST_HEAD(&chunk->entry); entry 927 drivers/crypto/n2_core.c list_add_tail(&chunk->entry, entry 934 drivers/crypto/n2_core.c INIT_LIST_HEAD(&chunk->entry); entry 955 drivers/crypto/n2_core.c list_add_tail(&chunk->entry, &rctx->chunk_list); entry 970 drivers/crypto/n2_core.c list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) { entry 971 drivers/crypto/n2_core.c list_del(&c->entry); entry 997 drivers/crypto/n2_core.c list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) { entry 1001 drivers/crypto/n2_core.c list_del(&c->entry); entry 1055 drivers/crypto/n2_core.c entry) { entry 1061 drivers/crypto/n2_core.c list_del(&c->entry); entry 1068 drivers/crypto/n2_core.c entry) { entry 1090 drivers/crypto/n2_core.c list_del(&c->entry); entry 1351 drivers/crypto/n2_core.c list_for_each_entry_safe(cipher, cipher_tmp, &cipher_algs, entry) { entry 1353 drivers/crypto/n2_core.c list_del(&cipher->entry); entry 1356 drivers/crypto/n2_core.c list_for_each_entry_safe(hmac, hmac_tmp, &hmac_algs, derived.entry) { entry 1358 drivers/crypto/n2_core.c list_del(&hmac->derived.entry); entry 1361 drivers/crypto/n2_core.c list_for_each_entry_safe(alg, alg_tmp, &ahash_algs, entry) { entry 1363 drivers/crypto/n2_core.c list_del(&alg->entry); entry 1398 drivers/crypto/n2_core.c list_add(&p->entry, &cipher_algs); entry 1402 drivers/crypto/n2_core.c list_del(&p->entry); entry 1422 drivers/crypto/n2_core.c INIT_LIST_HEAD(&p->derived.entry); entry 1436 drivers/crypto/n2_core.c list_add(&p->derived.entry, &hmac_algs); entry 1440 drivers/crypto/n2_core.c list_del(&p->derived.entry); entry 1490 drivers/crypto/n2_core.c list_add(&p->entry, &ahash_algs); entry 1494 drivers/crypto/n2_core.c list_del(&p->entry); entry 173 drivers/crypto/nx/nx-842-pseries.c struct nx842_slentry *entry; entry 177 drivers/crypto/nx/nx-842-pseries.c entry = sl->entries; entry 179 drivers/crypto/nx/nx-842-pseries.c entry->ptr = cpu_to_be64(nx842_get_pa((void *)buf)); entry 182 drivers/crypto/nx/nx-842-pseries.c entry->len = cpu_to_be64(entrylen); entry 188 drivers/crypto/nx/nx-842-pseries.c entry++; entry 96 drivers/crypto/picoxcell_crypto.c struct list_head entry; entry 139 drivers/crypto/picoxcell_crypto.c struct list_head entry; entry 1737 drivers/crypto/picoxcell_crypto.c list_add_tail(&engine->algs[i].entry, entry 1754 drivers/crypto/picoxcell_crypto.c list_add_tail(&engine->aeads[i].entry, entry 1788 drivers/crypto/picoxcell_crypto.c list_for_each_entry_safe(aead, an, &engine->registered_aeads, entry) { entry 1789 drivers/crypto/picoxcell_crypto.c list_del(&aead->entry); entry 1793 drivers/crypto/picoxcell_crypto.c list_for_each_entry_safe(alg, next, &engine->registered_algs, entry) { entry 1794 drivers/crypto/picoxcell_crypto.c list_del(&alg->entry); entry 74 drivers/crypto/qat/qat_common/adf_isr.c pci_dev_info->msix_entries.entries[i].entry = i; entry 76 drivers/crypto/qat/qat_common/adf_isr.c pci_dev_info->msix_entries.entries[0].entry = entry 391 drivers/crypto/qce/ablkcipher.c INIT_LIST_HEAD(&tmpl->entry); entry 403 drivers/crypto/qce/ablkcipher.c list_add_tail(&tmpl->entry, &ablkcipher_algs); entry 412 drivers/crypto/qce/ablkcipher.c list_for_each_entry_safe(tmpl, n, &ablkcipher_algs, entry) { entry 414 drivers/crypto/qce/ablkcipher.c list_del(&tmpl->entry); entry 77 drivers/crypto/qce/common.h struct list_head entry; entry 508 drivers/crypto/qce/sha.c INIT_LIST_HEAD(&tmpl->entry); entry 520 drivers/crypto/qce/sha.c list_add_tail(&tmpl->entry, &ahash_algs); entry 529 drivers/crypto/qce/sha.c list_for_each_entry_safe(tmpl, n, &ahash_algs, entry) { entry 531 drivers/crypto/qce/sha.c list_del(&tmpl->entry); entry 3008 drivers/crypto/talitos.c struct list_head entry; entry 3113 drivers/crypto/talitos.c list_for_each_entry_safe(t_alg, n, &priv->alg_list, entry) { entry 3124 drivers/crypto/talitos.c list_del(&t_alg->entry); entry 3487 drivers/crypto/talitos.c list_add_tail(&t_alg->entry, &priv->alg_list); entry 192 drivers/dma/ioat/dca.c u8 entry; entry 198 drivers/dma/ioat/dca.c entry = ioatdca->tag_map[i]; entry 199 drivers/dma/ioat/dca.c if (entry & DCA3_TAG_MAP_BIT_TO_SEL) { entry 200 drivers/dma/ioat/dca.c bit = entry & entry 203 drivers/dma/ioat/dca.c } else if (entry & DCA3_TAG_MAP_BIT_TO_INV) { entry 204 drivers/dma/ioat/dca.c bit = entry & ~DCA3_TAG_MAP_BIT_TO_INV; entry 207 drivers/dma/ioat/dca.c value = (entry & DCA3_TAG_MAP_LITERAL_VAL) ? 1 : 0; entry 426 drivers/dma/ioat/init.c ioat_dma->msix_entries[i].entry = i; entry 44 drivers/dma/ioat/sysfs.c struct ioat_sysfs_entry *entry; entry 47 drivers/dma/ioat/sysfs.c entry = container_of(attr, struct ioat_sysfs_entry, attr); entry 50 drivers/dma/ioat/sysfs.c if (!entry->show) entry 52 drivers/dma/ioat/sysfs.c return entry->show(&ioat_chan->dma_chan, page); entry 59 drivers/dma/ioat/sysfs.c struct ioat_sysfs_entry *entry; entry 62 drivers/dma/ioat/sysfs.c entry = container_of(attr, struct ioat_sysfs_entry, attr); entry 65 drivers/dma/ioat/sysfs.c if (!entry->store) entry 67 drivers/dma/ioat/sysfs.c return entry->store(&ioat_chan->dma_chan, page, count); entry 108 drivers/edac/ghes_edac.c struct memdev_dmi_entry *entry = (struct memdev_dmi_entry *)dh; entry 114 drivers/edac/ghes_edac.c if (entry->size == 0xffff) { entry 118 drivers/edac/ghes_edac.c } else if (entry->size == 0x7fff) { entry 119 drivers/edac/ghes_edac.c dimm->nr_pages = MiB_TO_PAGES(entry->extended_size); entry 121 drivers/edac/ghes_edac.c if (entry->size & BIT(15)) entry 122 drivers/edac/ghes_edac.c dimm->nr_pages = MiB_TO_PAGES((entry->size & 0x7fff) << 10); entry 124 drivers/edac/ghes_edac.c dimm->nr_pages = MiB_TO_PAGES(entry->size); entry 127 drivers/edac/ghes_edac.c switch (entry->memory_type) { entry 129 drivers/edac/ghes_edac.c if (entry->type_detail & BIT(13)) entry 135 drivers/edac/ghes_edac.c if (entry->type_detail & BIT(13)) entry 144 drivers/edac/ghes_edac.c if (entry->type_detail & BIT(12)) entry 146 drivers/edac/ghes_edac.c else if (entry->type_detail & BIT(13)) entry 152 drivers/edac/ghes_edac.c if (entry->type_detail & BIT(12)) entry 154 drivers/edac/ghes_edac.c else if (entry->type_detail & BIT(13)) entry 160 drivers/edac/ghes_edac.c if (entry->type_detail & BIT(6)) entry 162 drivers/edac/ghes_edac.c else if ((entry->type_detail & rdr_mask) == rdr_mask) entry 164 drivers/edac/ghes_edac.c else if (entry->type_detail & BIT(7)) entry 166 drivers/edac/ghes_edac.c else if (entry->type_detail & BIT(9)) entry 176 drivers/edac/ghes_edac.c if (entry->total_width == entry->data_width) entry 194 drivers/edac/ghes_edac.c entry->memory_type, entry->type_detail, entry 195 drivers/edac/ghes_edac.c entry->total_width, entry->data_width); entry 198 drivers/edac/ghes_edac.c dimm->smbios_handle = entry->handle; entry 1148 drivers/edac/sb_edac.c const int entry, entry 1171 drivers/edac/sb_edac.c knl_tad_dram_limit_lo[entry], ®_limit_lo); entry 1173 drivers/edac/sb_edac.c knl_tad_dram_offset_lo[entry], ®_offset_lo); entry 1175 drivers/edac/sb_edac.c knl_tad_dram_hi[entry], ®_hi); entry 1228 drivers/edac/sb_edac.c static u32 knl_get_edc_route(int entry, u32 reg) entry 1230 drivers/edac/sb_edac.c WARN_ON(entry >= KNL_MAX_EDCS); entry 1231 drivers/edac/sb_edac.c return GET_BITFIELD(reg, entry*3, (entry*3)+2); entry 1251 drivers/edac/sb_edac.c static u32 knl_get_mc_route(int entry, u32 reg) entry 1255 drivers/edac/sb_edac.c WARN_ON(entry >= KNL_MAX_CHANNELS); entry 1257 drivers/edac/sb_edac.c mc = GET_BITFIELD(reg, entry*3, (entry*3)+2); entry 1258 drivers/edac/sb_edac.c chan = GET_BITFIELD(reg, (entry*2) + 18, (entry*2) + 18 + 1); entry 723 drivers/edac/thunderx_edac.c lmc->msix_ent.entry = 0; entry 1079 drivers/edac/thunderx_edac.c msix_ent[msix->entry]); entry 1108 drivers/edac/thunderx_edac.c msix_ent[msix->entry]); entry 1172 drivers/edac/thunderx_edac.c msix_ent[msix->entry]); entry 1177 drivers/edac/thunderx_edac.c ctx->link = msix->entry; entry 1191 drivers/edac/thunderx_edac.c msix_ent[msix->entry]); entry 1391 drivers/edac/thunderx_edac.c ocx->msix_ent[i].entry = i; entry 2029 drivers/edac/thunderx_edac.c l2c->msix_ent.entry = 0; entry 875 drivers/extcon/extcon.c list_for_each_entry(sd, &extcon_dev_list, entry) { entry 1256 drivers/extcon/extcon.c list_add(&edev->entry, &extcon_dev_list); entry 1297 drivers/extcon/extcon.c list_del(&edev->entry); entry 1341 drivers/extcon/extcon.c list_for_each_entry(edev, &extcon_dev_list, entry) entry 51 drivers/extcon/extcon.h struct list_head entry; entry 1791 drivers/firewire/ohci.c u32 entry; entry 1794 drivers/firewire/ohci.c entry = ohci->self_id_buffer[i]; entry 1795 drivers/firewire/ohci.c if ((self_id & 0xff000000) == (entry & 0xff000000)) entry 1797 drivers/firewire/ohci.c if ((self_id & 0xff000000) < (entry & 0xff000000)) entry 54 drivers/firmware/dmi-sysfs.c ssize_t (*show)(struct dmi_sysfs_entry *entry, char *buf); entry 69 drivers/firmware/dmi-sysfs.c ssize_t (*show)(struct dmi_sysfs_entry *entry, entry 101 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry = to_entry(kobj); entry 108 drivers/firmware/dmi-sysfs.c return attr->show(entry, buf); entry 119 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry; entry 130 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry = data->entry; entry 133 drivers/firmware/dmi-sysfs.c if (dh->type != entry->dh.type) entry 149 drivers/firmware/dmi-sysfs.c data->ret = data->callback(entry, dh, data->private); entry 159 drivers/firmware/dmi-sysfs.c static ssize_t find_dmi_entry(struct dmi_sysfs_entry *entry, entry 163 drivers/firmware/dmi-sysfs.c .entry = entry, entry 166 drivers/firmware/dmi-sysfs.c .instance_countdown = entry->instance, entry 203 drivers/firmware/dmi-sysfs.c static ssize_t dmi_entry_attr_show_helper(struct dmi_sysfs_entry *entry, entry 212 drivers/firmware/dmi-sysfs.c return attr->show(entry, dh, data->buf); entry 269 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_sel_##_field(struct dmi_sysfs_entry *entry, \ entry 361 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sel_raw_read_io(struct dmi_sysfs_entry *entry, entry 378 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sel_raw_read_phys32(struct dmi_sysfs_entry *entry, entry 399 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sel_raw_read_helper(struct dmi_sysfs_entry *entry, entry 415 drivers/firmware/dmi-sysfs.c return dmi_sel_raw_read_io(entry, &sel, state->buf, entry 418 drivers/firmware/dmi-sysfs.c return dmi_sel_raw_read_phys32(entry, &sel, state->buf, entry 434 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry = to_entry(kobj->parent); entry 441 drivers/firmware/dmi-sysfs.c return find_dmi_entry(entry, dmi_sel_raw_read_helper, &state); entry 449 drivers/firmware/dmi-sysfs.c static int dmi_system_event_log(struct dmi_sysfs_entry *entry) entry 453 drivers/firmware/dmi-sysfs.c entry->child = kzalloc(sizeof(*entry->child), GFP_KERNEL); entry 454 drivers/firmware/dmi-sysfs.c if (!entry->child) entry 456 drivers/firmware/dmi-sysfs.c ret = kobject_init_and_add(entry->child, entry 458 drivers/firmware/dmi-sysfs.c &entry->kobj, entry 463 drivers/firmware/dmi-sysfs.c ret = sysfs_create_bin_file(entry->child, &dmi_sel_raw_attr); entry 470 drivers/firmware/dmi-sysfs.c kobject_del(entry->child); entry 472 drivers/firmware/dmi-sysfs.c kfree(entry->child); entry 480 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_length(struct dmi_sysfs_entry *entry, char *buf) entry 482 drivers/firmware/dmi-sysfs.c return sprintf(buf, "%d\n", entry->dh.length); entry 485 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_handle(struct dmi_sysfs_entry *entry, char *buf) entry 487 drivers/firmware/dmi-sysfs.c return sprintf(buf, "%d\n", entry->dh.handle); entry 490 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_type(struct dmi_sysfs_entry *entry, char *buf) entry 492 drivers/firmware/dmi-sysfs.c return sprintf(buf, "%d\n", entry->dh.type); entry 495 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_instance(struct dmi_sysfs_entry *entry, entry 498 drivers/firmware/dmi-sysfs.c return sprintf(buf, "%d\n", entry->instance); entry 501 drivers/firmware/dmi-sysfs.c static ssize_t dmi_sysfs_entry_position(struct dmi_sysfs_entry *entry, entry 504 drivers/firmware/dmi-sysfs.c return sprintf(buf, "%d\n", entry->position); entry 507 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, length); entry 508 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, handle); entry 509 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, type); entry 510 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, instance); entry 511 drivers/firmware/dmi-sysfs.c static DMI_SYSFS_ATTR(entry, position); entry 522 drivers/firmware/dmi-sysfs.c static ssize_t dmi_entry_raw_read_helper(struct dmi_sysfs_entry *entry, entry 540 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry = to_entry(kobj); entry 547 drivers/firmware/dmi-sysfs.c return find_dmi_entry(entry, dmi_entry_raw_read_helper, &state); entry 557 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry = to_entry(kobj); entry 560 drivers/firmware/dmi-sysfs.c list_del(&entry->list); entry 562 drivers/firmware/dmi-sysfs.c kfree(entry); entry 582 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry; entry 590 drivers/firmware/dmi-sysfs.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 591 drivers/firmware/dmi-sysfs.c if (!entry) { entry 597 drivers/firmware/dmi-sysfs.c memcpy(&entry->dh, dh, sizeof(*dh)); entry 598 drivers/firmware/dmi-sysfs.c entry->instance = instance_counts[dh->type]++; entry 599 drivers/firmware/dmi-sysfs.c entry->position = position_count++; entry 601 drivers/firmware/dmi-sysfs.c entry->kobj.kset = dmi_kset; entry 602 drivers/firmware/dmi-sysfs.c *ret = kobject_init_and_add(&entry->kobj, &dmi_sysfs_entry_ktype, NULL, entry 603 drivers/firmware/dmi-sysfs.c "%d-%d", dh->type, entry->instance); entry 606 drivers/firmware/dmi-sysfs.c kfree(entry); entry 612 drivers/firmware/dmi-sysfs.c list_add_tail(&entry->list, &entry_list); entry 618 drivers/firmware/dmi-sysfs.c *ret = dmi_system_event_log(entry); entry 628 drivers/firmware/dmi-sysfs.c *ret = sysfs_create_bin_file(&entry->kobj, &dmi_entry_raw_attr); entry 634 drivers/firmware/dmi-sysfs.c kobject_put(entry->child); entry 635 drivers/firmware/dmi-sysfs.c kobject_put(&entry->kobj); entry 641 drivers/firmware/dmi-sysfs.c struct dmi_sysfs_entry *entry, *next; entry 644 drivers/firmware/dmi-sysfs.c list_for_each_entry_safe(entry, next, &entry_list, list) { entry 645 drivers/firmware/dmi-sysfs.c kobject_put(entry->child); entry 646 drivers/firmware/dmi-sysfs.c kobject_put(&entry->kobj); entry 50 drivers/firmware/efi/apple-properties.c struct property_entry entry[]) entry 88 drivers/firmware/efi/apple-properties.c entry[i].name = key; entry 89 drivers/firmware/efi/apple-properties.c entry[i].length = val_len - sizeof(val_len); entry 90 drivers/firmware/efi/apple-properties.c entry[i].is_array = !!entry[i].length; entry 91 drivers/firmware/efi/apple-properties.c entry[i].type = DEV_PROP_U8; entry 92 drivers/firmware/efi/apple-properties.c entry[i].pointer.u8_data = ptr + key_len + sizeof(val_len); entry 95 drivers/firmware/efi/apple-properties.c dev_info(dev, "property: %s\n", entry[i].name); entry 97 drivers/firmware/efi/apple-properties.c 16, 1, entry[i].pointer.u8_data, entry 98 drivers/firmware/efi/apple-properties.c entry[i].length, true); entry 121 drivers/firmware/efi/apple-properties.c struct property_entry *entry = NULL; entry 146 drivers/firmware/efi/apple-properties.c entry = kcalloc(dev_header->prop_count + 1, sizeof(*entry), entry 148 drivers/firmware/efi/apple-properties.c if (!entry) { entry 153 drivers/firmware/efi/apple-properties.c unmarshal_key_value_pairs(dev_header, dev, ptr, entry); entry 154 drivers/firmware/efi/apple-properties.c if (!entry[0].name) entry 157 drivers/firmware/efi/apple-properties.c ret = device_add_properties(dev, entry); /* makes deep copy */ entry 161 drivers/firmware/efi/apple-properties.c for (i = 0; entry[i].name; i++) entry 162 drivers/firmware/efi/apple-properties.c kfree(entry[i].name); entry 165 drivers/firmware/efi/apple-properties.c kfree(entry); entry 38 drivers/firmware/efi/efi-pstore.c static int efi_pstore_read_func(struct efivar_entry *entry, entry 49 drivers/firmware/efi/efi-pstore.c if (efi_guidcmp(entry->var.VendorGuid, vendor)) entry 53 drivers/firmware/efi/efi-pstore.c name[i] = entry->var.VariableName[i]; entry 93 drivers/firmware/efi/efi-pstore.c entry->var.DataSize = 1024; entry 94 drivers/firmware/efi/efi-pstore.c __efivar_entry_get(entry, &entry->var.Attributes, entry 95 drivers/firmware/efi/efi-pstore.c &entry->var.DataSize, entry->var.Data); entry 96 drivers/firmware/efi/efi-pstore.c size = entry->var.DataSize; entry 97 drivers/firmware/efi/efi-pstore.c memcpy(record->buf, entry->var.Data, entry 123 drivers/firmware/efi/efi-pstore.c static inline int __efi_pstore_scan_sysfs_exit(struct efivar_entry *entry, entry 126 drivers/firmware/efi/efi-pstore.c if (entry->deleting) { entry 127 drivers/firmware/efi/efi-pstore.c list_del(&entry->list); entry 129 drivers/firmware/efi/efi-pstore.c efivar_unregister(entry); entry 133 drivers/firmware/efi/efi-pstore.c entry->scanning = false; entry 171 drivers/firmware/efi/efi-pstore.c struct efivar_entry *entry, *n; entry 177 drivers/firmware/efi/efi-pstore.c list_for_each_entry_safe(entry, n, head, list) { entry 178 drivers/firmware/efi/efi-pstore.c efi_pstore_scan_sysfs_enter(entry, n, head); entry 180 drivers/firmware/efi/efi-pstore.c size = efi_pstore_read_func(entry, record); entry 181 drivers/firmware/efi/efi-pstore.c ret = efi_pstore_scan_sysfs_exit(entry, n, head, entry 275 drivers/firmware/efi/efi-pstore.c static int efi_pstore_erase_func(struct efivar_entry *entry, void *data) entry 281 drivers/firmware/efi/efi-pstore.c if (efi_guidcmp(entry->var.VendorGuid, vendor)) entry 284 drivers/firmware/efi/efi-pstore.c if (ucs2_strncmp(entry->var.VariableName, efi_name, (size_t)ucs2_len)) entry 287 drivers/firmware/efi/efi-pstore.c if (entry->scanning) { entry 292 drivers/firmware/efi/efi-pstore.c entry->deleting = true; entry 294 drivers/firmware/efi/efi-pstore.c list_del(&entry->list); entry 297 drivers/firmware/efi/efi-pstore.c __efivar_entry_delete(entry); entry 304 drivers/firmware/efi/efi-pstore.c struct efivar_entry *entry = NULL; entry 318 drivers/firmware/efi/efi-pstore.c efi_name, &entry); entry 321 drivers/firmware/efi/efi-pstore.c if (found && !entry->scanning) entry 322 drivers/firmware/efi/efi-pstore.c efivar_unregister(entry); entry 241 drivers/firmware/efi/efi.c struct efivar_entry *entry; entry 250 drivers/firmware/efi/efi.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 251 drivers/firmware/efi/efi.c if (!entry) entry 254 drivers/firmware/efi/efi.c memcpy(entry->var.VariableName, name, name_size); entry 255 drivers/firmware/efi/efi.c memcpy(&entry->var.VendorGuid, &vendor, sizeof(efi_guid_t)); entry 257 drivers/firmware/efi/efi.c efivar_entry_add(entry, list); entry 265 drivers/firmware/efi/efi.c struct efivar_entry *entry, *aux; entry 275 drivers/firmware/efi/efi.c list_for_each_entry_safe(entry, aux, &entries, list) { entry 277 drivers/firmware/efi/efi.c &entry->var.VendorGuid); entry 279 drivers/firmware/efi/efi.c list_del(&entry->list); entry 281 drivers/firmware/efi/efi.c ret = efivar_entry_size(entry, &size); entry 293 drivers/firmware/efi/efi.c ret = efivar_entry_get(entry, NULL, &size, data); entry 311 drivers/firmware/efi/efi.c kfree(entry); entry 613 drivers/firmware/efi/efi.c memblock_reserve(rsv->entry[i].base, entry 614 drivers/firmware/efi/efi.c rsv->entry[i].size); entry 1011 drivers/firmware/efi/efi.c rsv->entry[index].base = addr; entry 1012 drivers/firmware/efi/efi.c rsv->entry[index].size = size; entry 1039 drivers/firmware/efi/efi.c rsv->entry[0].base = addr; entry 1040 drivers/firmware/efi/efi.c rsv->entry[0].size = size; entry 28 drivers/firmware/efi/efibc.c struct efivar_entry *entry; entry 31 drivers/firmware/efi/efibc.c if (size > sizeof(entry->var.Data)) { entry 36 drivers/firmware/efi/efibc.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 37 drivers/firmware/efi/efibc.c if (!entry) { entry 42 drivers/firmware/efi/efibc.c efibc_str_to_str16(name, entry->var.VariableName); entry 43 drivers/firmware/efi/efibc.c efibc_str_to_str16(value, (efi_char16_t *)entry->var.Data); entry 44 drivers/firmware/efi/efibc.c memcpy(&entry->var.VendorGuid, &guid, sizeof(guid)); entry 46 drivers/firmware/efi/efibc.c ret = efivar_entry_set_safe(entry->var.VariableName, entry 47 drivers/firmware/efi/efibc.c entry->var.VendorGuid, entry 51 drivers/firmware/efi/efibc.c false, size, entry->var.Data); entry 57 drivers/firmware/efi/efibc.c kfree(entry); entry 46 drivers/firmware/efi/efivars.c ssize_t (*show) (struct efivar_entry *entry, char *buf); entry 47 drivers/firmware/efi/efivars.c ssize_t (*store)(struct efivar_entry *entry, const char *buf, size_t count); entry 67 drivers/firmware/efi/efivars.c efivar_guid_read(struct efivar_entry *entry, char *buf) entry 69 drivers/firmware/efi/efivars.c struct efi_variable *var = &entry->var; entry 72 drivers/firmware/efi/efivars.c if (!entry || !buf) entry 83 drivers/firmware/efi/efivars.c efivar_attr_read(struct efivar_entry *entry, char *buf) entry 85 drivers/firmware/efi/efivars.c struct efi_variable *var = &entry->var; entry 90 drivers/firmware/efi/efivars.c if (!entry || !buf) entry 93 drivers/firmware/efi/efivars.c ret = efivar_entry_get(entry, &var->Attributes, &size, var->Data); entry 119 drivers/firmware/efi/efivars.c efivar_size_read(struct efivar_entry *entry, char *buf) entry 121 drivers/firmware/efi/efivars.c struct efi_variable *var = &entry->var; entry 126 drivers/firmware/efi/efivars.c if (!entry || !buf) entry 129 drivers/firmware/efi/efivars.c ret = efivar_entry_get(entry, &var->Attributes, &size, var->Data); entry 139 drivers/firmware/efi/efivars.c efivar_data_read(struct efivar_entry *entry, char *buf) entry 141 drivers/firmware/efi/efivars.c struct efi_variable *var = &entry->var; entry 145 drivers/firmware/efi/efivars.c if (!entry || !buf) entry 148 drivers/firmware/efi/efivars.c ret = efivar_entry_get(entry, &var->Attributes, &size, var->Data); entry 201 drivers/firmware/efi/efivars.c efivar_store_raw(struct efivar_entry *entry, const char *buf, size_t count) entry 203 drivers/firmware/efi/efivars.c struct efi_variable *new_var, *var = &entry->var; entry 211 drivers/firmware/efi/efivars.c if (!entry || !buf) entry 231 drivers/firmware/efi/efivars.c copy_out_compat(&entry->var, compat); entry 248 drivers/firmware/efi/efivars.c memcpy(&entry->var, new_var, count); entry 251 drivers/firmware/efi/efivars.c err = efivar_entry_set(entry, attributes, size, data, NULL); entry 261 drivers/firmware/efi/efivars.c efivar_show_raw(struct efivar_entry *entry, char *buf) entry 263 drivers/firmware/efi/efivars.c struct efi_variable *var = &entry->var; entry 269 drivers/firmware/efi/efivars.c if (!entry || !buf) entry 272 drivers/firmware/efi/efivars.c ret = efivar_entry_get(entry, &var->Attributes, &datasize, var->Data); entry 440 drivers/firmware/efi/efivars.c struct efivar_entry *entry; entry 465 drivers/firmware/efi/efivars.c entry = efivar_entry_find(name, vendor, &efivar_sysfs_list, true); entry 466 drivers/firmware/efi/efivars.c if (!entry) entry 468 drivers/firmware/efi/efivars.c else if (__efivar_entry_delete(entry)) entry 476 drivers/firmware/efi/efivars.c if (!entry->scanning) { entry 478 drivers/firmware/efi/efivars.c efivar_unregister(entry); entry 597 drivers/firmware/efi/efivars.c struct efivar_entry *entry = data; entry 602 drivers/firmware/efi/efivars.c memcpy(entry->var.VariableName, name, name_size); entry 603 drivers/firmware/efi/efivars.c memcpy(&(entry->var.VendorGuid), &vendor, sizeof(efi_guid_t)); entry 610 drivers/firmware/efi/efivars.c struct efivar_entry *entry; entry 615 drivers/firmware/efi/efivars.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 616 drivers/firmware/efi/efivars.c if (!entry) entry 619 drivers/firmware/efi/efivars.c err = efivar_init(efivar_update_sysfs_entry, entry, entry 624 drivers/firmware/efi/efivars.c efivar_create_sysfs_entry(entry); entry 627 drivers/firmware/efi/efivars.c kfree(entry); entry 633 drivers/firmware/efi/efivars.c struct efivar_entry *entry; entry 635 drivers/firmware/efi/efivars.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 636 drivers/firmware/efi/efivars.c if (!entry) entry 639 drivers/firmware/efi/efivars.c memcpy(entry->var.VariableName, name, name_size); entry 640 drivers/firmware/efi/efivars.c memcpy(&(entry->var.VendorGuid), &vendor, sizeof(efi_guid_t)); entry 642 drivers/firmware/efi/efivars.c efivar_create_sysfs_entry(entry); entry 647 drivers/firmware/efi/efivars.c static int efivar_sysfs_destroy(struct efivar_entry *entry, void *data) entry 649 drivers/firmware/efi/efivars.c int err = efivar_entry_remove(entry); entry 653 drivers/firmware/efi/efivars.c efivar_unregister(entry); entry 77 drivers/firmware/efi/esrt.c ssize_t (*show)(struct esre_entry *entry, char *buf); entry 78 drivers/firmware/efi/esrt.c ssize_t (*store)(struct esre_entry *entry, entry 95 drivers/firmware/efi/esrt.c struct esre_entry *entry = to_entry(kobj); entry 102 drivers/firmware/efi/esrt.c return attr->show(entry, buf); entry 110 drivers/firmware/efi/esrt.c static ssize_t fw_class_show(struct esre_entry *entry, char *buf) entry 114 drivers/firmware/efi/esrt.c efi_guid_to_str(&entry->esre.esre1->fw_class, str); entry 124 drivers/firmware/efi/esrt.c static ssize_t name##_show(struct esre_entry *entry, char *buf) \ entry 127 drivers/firmware/efi/esrt.c le##size##_to_cpu(entry->esre.esre1->name)); \ entry 151 drivers/firmware/efi/esrt.c struct esre_entry *entry = to_entry(kobj); entry 153 drivers/firmware/efi/esrt.c list_del(&entry->list); entry 154 drivers/firmware/efi/esrt.c kfree(entry); entry 169 drivers/firmware/efi/esrt.c struct esre_entry *entry; entry 171 drivers/firmware/efi/esrt.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 172 drivers/firmware/efi/esrt.c if (!entry) entry 175 drivers/firmware/efi/esrt.c entry->kobj.kset = esrt_kset; entry 180 drivers/firmware/efi/esrt.c entry->esre.esre1 = esre; entry 181 drivers/firmware/efi/esrt.c rc = kobject_init_and_add(&entry->kobj, &esre1_ktype, NULL, entry 184 drivers/firmware/efi/esrt.c kfree(entry); entry 189 drivers/firmware/efi/esrt.c list_add_tail(&entry->list, &entry_list); entry 369 drivers/firmware/efi/esrt.c struct esre_entry *entry, *next; entry 371 drivers/firmware/efi/esrt.c list_for_each_entry_safe(entry, next, &entry_list, list) { entry 372 drivers/firmware/efi/esrt.c kobject_put(&entry->kobj); entry 166 drivers/firmware/efi/memattr.c valid = entry_is_valid((void *)tbl->entry + i * tbl->desc_size, entry 25 drivers/firmware/efi/runtime-map.c ssize_t (*show)(struct efi_runtime_map_entry *entry, char *buf); entry 33 drivers/firmware/efi/runtime-map.c static ssize_t type_show(struct efi_runtime_map_entry *entry, char *buf) entry 35 drivers/firmware/efi/runtime-map.c return snprintf(buf, PAGE_SIZE, "0x%x\n", entry->md.type); entry 38 drivers/firmware/efi/runtime-map.c #define EFI_RUNTIME_FIELD(var) entry->md.var entry 41 drivers/firmware/efi/runtime-map.c static ssize_t name##_show(struct efi_runtime_map_entry *entry, char *buf) \ entry 59 drivers/firmware/efi/runtime-map.c struct efi_runtime_map_entry *entry = to_map_entry(kobj); entry 62 drivers/firmware/efi/runtime-map.c return map_attr->show(entry, buf); entry 89 drivers/firmware/efi/runtime-map.c struct efi_runtime_map_entry *entry; entry 91 drivers/firmware/efi/runtime-map.c entry = to_map_entry(kobj); entry 92 drivers/firmware/efi/runtime-map.c kfree(entry); entry 108 drivers/firmware/efi/runtime-map.c struct efi_runtime_map_entry *entry; entry 116 drivers/firmware/efi/runtime-map.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 117 drivers/firmware/efi/runtime-map.c if (!entry) { entry 123 drivers/firmware/efi/runtime-map.c memcpy(&entry->md, md, sizeof(efi_memory_desc_t)); entry 125 drivers/firmware/efi/runtime-map.c kobject_init(&entry->kobj, &map_ktype); entry 126 drivers/firmware/efi/runtime-map.c entry->kobj.kset = map_kset; entry 127 drivers/firmware/efi/runtime-map.c ret = kobject_add(&entry->kobj, NULL, "%d", nr); entry 129 drivers/firmware/efi/runtime-map.c kobject_put(&entry->kobj); entry 135 drivers/firmware/efi/runtime-map.c return entry; entry 162 drivers/firmware/efi/runtime-map.c struct efi_runtime_map_entry *entry; entry 168 drivers/firmware/efi/runtime-map.c map_entries = kcalloc(efi.memmap.nr_map, sizeof(entry), GFP_KERNEL); entry 176 drivers/firmware/efi/runtime-map.c entry = add_sysfs_runtime_map_entry(efi_kobj, i, md); entry 177 drivers/firmware/efi/runtime-map.c if (IS_ERR(entry)) { entry 178 drivers/firmware/efi/runtime-map.c ret = PTR_ERR(entry); entry 181 drivers/firmware/efi/runtime-map.c *(map_entries + i++) = entry; entry 187 drivers/firmware/efi/runtime-map.c entry = *(map_entries + j); entry 188 drivers/firmware/efi/runtime-map.c kobject_put(&entry->kobj); entry 340 drivers/firmware/efi/vars.c struct efivar_entry *entry, *n; entry 345 drivers/firmware/efi/vars.c list_for_each_entry_safe(entry, n, head, list) { entry 346 drivers/firmware/efi/vars.c strsize2 = ucs2_strsize(entry->var.VariableName, 1024); entry 348 drivers/firmware/efi/vars.c !memcmp(variable_name, &(entry->var.VariableName), entry 350 drivers/firmware/efi/vars.c !efi_guidcmp(entry->var.VendorGuid, entry 526 drivers/firmware/efi/vars.c int efivar_entry_add(struct efivar_entry *entry, struct list_head *head) entry 530 drivers/firmware/efi/vars.c list_add(&entry->list, head); entry 543 drivers/firmware/efi/vars.c int efivar_entry_remove(struct efivar_entry *entry) entry 547 drivers/firmware/efi/vars.c list_del(&entry->list); entry 565 drivers/firmware/efi/vars.c static void efivar_entry_list_del_unlock(struct efivar_entry *entry) entry 567 drivers/firmware/efi/vars.c list_del(&entry->list); entry 586 drivers/firmware/efi/vars.c int __efivar_entry_delete(struct efivar_entry *entry) entry 593 drivers/firmware/efi/vars.c status = __efivars->ops->set_variable(entry->var.VariableName, entry 594 drivers/firmware/efi/vars.c &entry->var.VendorGuid, entry 612 drivers/firmware/efi/vars.c int efivar_entry_delete(struct efivar_entry *entry) entry 625 drivers/firmware/efi/vars.c status = ops->set_variable(entry->var.VariableName, entry 626 drivers/firmware/efi/vars.c &entry->var.VendorGuid, entry 633 drivers/firmware/efi/vars.c efivar_entry_list_del_unlock(entry); entry 659 drivers/firmware/efi/vars.c int efivar_entry_set(struct efivar_entry *entry, u32 attributes, entry 664 drivers/firmware/efi/vars.c efi_char16_t *name = entry->var.VariableName; entry 665 drivers/firmware/efi/vars.c efi_guid_t vendor = entry->var.VendorGuid; entry 817 drivers/firmware/efi/vars.c struct efivar_entry *entry, *n; entry 821 drivers/firmware/efi/vars.c list_for_each_entry_safe(entry, n, head, list) { entry 823 drivers/firmware/efi/vars.c strsize2 = ucs2_strsize(entry->var.VariableName, 1024); entry 825 drivers/firmware/efi/vars.c !memcmp(name, &(entry->var.VariableName), strsize1) && entry 826 drivers/firmware/efi/vars.c !efi_guidcmp(guid, entry->var.VendorGuid)) { entry 836 drivers/firmware/efi/vars.c if (entry->scanning) { entry 841 drivers/firmware/efi/vars.c entry->deleting = true; entry 843 drivers/firmware/efi/vars.c list_del(&entry->list); entry 846 drivers/firmware/efi/vars.c return entry; entry 855 drivers/firmware/efi/vars.c int efivar_entry_size(struct efivar_entry *entry, unsigned long *size) entry 869 drivers/firmware/efi/vars.c status = ops->get_variable(entry->var.VariableName, entry 870 drivers/firmware/efi/vars.c &entry->var.VendorGuid, NULL, size, NULL); entry 891 drivers/firmware/efi/vars.c int __efivar_entry_get(struct efivar_entry *entry, u32 *attributes, entry 899 drivers/firmware/efi/vars.c status = __efivars->ops->get_variable(entry->var.VariableName, entry 900 drivers/firmware/efi/vars.c &entry->var.VendorGuid, entry 914 drivers/firmware/efi/vars.c int efivar_entry_get(struct efivar_entry *entry, u32 *attributes, entry 927 drivers/firmware/efi/vars.c status = __efivars->ops->get_variable(entry->var.VariableName, entry 928 drivers/firmware/efi/vars.c &entry->var.VendorGuid, entry 958 drivers/firmware/efi/vars.c int efivar_entry_set_get_size(struct efivar_entry *entry, u32 attributes, entry 962 drivers/firmware/efi/vars.c efi_char16_t *name = entry->var.VariableName; entry 963 drivers/firmware/efi/vars.c efi_guid_t *vendor = &entry->var.VendorGuid; entry 1018 drivers/firmware/efi/vars.c status = ops->get_variable(entry->var.VariableName, entry 1019 drivers/firmware/efi/vars.c &entry->var.VendorGuid, entry 1023 drivers/firmware/efi/vars.c efivar_entry_list_del_unlock(entry); entry 1089 drivers/firmware/efi/vars.c struct efivar_entry *entry, *n; entry 1093 drivers/firmware/efi/vars.c list_for_each_entry_safe(entry, n, head, list) { entry 1094 drivers/firmware/efi/vars.c err = func(entry, data); entry 1100 drivers/firmware/efi/vars.c *prev = entry; entry 32 drivers/firmware/google/coreboot_table.c return device->entry.tag == driver->tag; entry 92 drivers/firmware/google/coreboot_table.c struct coreboot_table_entry *entry; entry 97 drivers/firmware/google/coreboot_table.c entry = ptr_entry; entry 99 drivers/firmware/google/coreboot_table.c device = kzalloc(sizeof(struct device) + entry->size, GFP_KERNEL); entry 107 drivers/firmware/google/coreboot_table.c memcpy(&device->entry, ptr_entry, entry->size); entry 115 drivers/firmware/google/coreboot_table.c ptr_entry += entry->size; entry 66 drivers/firmware/google/coreboot_table.h struct coreboot_table_entry entry; entry 614 drivers/firmware/google/gsmi.c struct gsmi_log_entry_type_1 entry = { entry 619 drivers/firmware/google/gsmi.c .data_len = sizeof(entry), entry 636 drivers/firmware/google/gsmi.c memcpy(gsmi_dev.data_buf->start, &entry, sizeof(entry)); entry 37 drivers/firmware/google/vpd_decode.c u32 *_consumed, const u8 **entry, u32 *entry_len) entry 49 drivers/firmware/google/vpd_decode.c *entry = input_buf + consumed; entry 253 drivers/firmware/iscsi_ibft.c struct ibft_kobject *entry = data; entry 254 drivers/firmware/iscsi_ibft.c struct ibft_initiator *initiator = entry->initiator; entry 255 drivers/firmware/iscsi_ibft.c void *ibft_loc = entry->header; entry 294 drivers/firmware/iscsi_ibft.c struct ibft_kobject *entry = data; entry 295 drivers/firmware/iscsi_ibft.c struct ibft_nic *nic = entry->nic; entry 296 drivers/firmware/iscsi_ibft.c void *ibft_loc = entry->header; entry 354 drivers/firmware/iscsi_ibft.c struct ibft_kobject *entry = data; entry 355 drivers/firmware/iscsi_ibft.c struct ibft_tgt *tgt = entry->tgt; entry 356 drivers/firmware/iscsi_ibft.c void *ibft_loc = entry->header; entry 418 drivers/firmware/iscsi_ibft.c struct ibft_kobject *entry = data; entry 424 drivers/firmware/iscsi_ibft.c entry->header->header.signature); entry 428 drivers/firmware/iscsi_ibft.c entry->header->header.oem_id); entry 432 drivers/firmware/iscsi_ibft.c entry->header->header.oem_table_id); entry 473 drivers/firmware/iscsi_ibft.c struct ibft_kobject *entry = data; entry 474 drivers/firmware/iscsi_ibft.c struct ibft_nic *nic = entry->nic; entry 527 drivers/firmware/iscsi_ibft.c struct ibft_kobject *entry = data; entry 528 drivers/firmware/iscsi_ibft.c struct ibft_tgt *tgt = entry->tgt; entry 564 drivers/firmware/iscsi_ibft.c struct ibft_kobject *entry = data; entry 565 drivers/firmware/iscsi_ibft.c struct ibft_initiator *init = entry->initiator; entry 43 drivers/firmware/memmap.c static ssize_t start_show(struct firmware_map_entry *entry, char *buf); entry 44 drivers/firmware/memmap.c static ssize_t end_show(struct firmware_map_entry *entry, char *buf); entry 45 drivers/firmware/memmap.c static ssize_t type_show(struct firmware_map_entry *entry, char *buf); entry 56 drivers/firmware/memmap.c ssize_t (*show)(struct firmware_map_entry *entry, char *buf); entry 99 drivers/firmware/memmap.c struct firmware_map_entry *entry = to_memmap_entry(kobj); entry 101 drivers/firmware/memmap.c if (PageReserved(virt_to_page(entry))) { entry 109 drivers/firmware/memmap.c list_add(&entry->list, &map_entries_bootmem); entry 115 drivers/firmware/memmap.c kfree(entry); entry 143 drivers/firmware/memmap.c struct firmware_map_entry *entry) entry 147 drivers/firmware/memmap.c entry->start = start; entry 148 drivers/firmware/memmap.c entry->end = end - 1; entry 149 drivers/firmware/memmap.c entry->type = type; entry 150 drivers/firmware/memmap.c INIT_LIST_HEAD(&entry->list); entry 151 drivers/firmware/memmap.c kobject_init(&entry->kobj, &memmap_ktype); entry 154 drivers/firmware/memmap.c list_add_tail(&entry->list, &map_entries); entry 167 drivers/firmware/memmap.c static inline void firmware_map_remove_entry(struct firmware_map_entry *entry) entry 169 drivers/firmware/memmap.c list_del(&entry->list); entry 175 drivers/firmware/memmap.c static int add_sysfs_fw_map_entry(struct firmware_map_entry *entry) entry 180 drivers/firmware/memmap.c if (entry->kobj.state_in_sysfs) entry 189 drivers/firmware/memmap.c entry->kobj.kset = mmap_kset; entry 190 drivers/firmware/memmap.c if (kobject_add(&entry->kobj, NULL, "%d", map_entries_nr++)) entry 191 drivers/firmware/memmap.c kobject_put(&entry->kobj); entry 199 drivers/firmware/memmap.c static inline void remove_sysfs_fw_map_entry(struct firmware_map_entry *entry) entry 201 drivers/firmware/memmap.c kobject_put(&entry->kobj); entry 221 drivers/firmware/memmap.c struct firmware_map_entry *entry; entry 223 drivers/firmware/memmap.c list_for_each_entry(entry, list, list) entry 224 drivers/firmware/memmap.c if ((entry->start == start) && (entry->end == end) && entry 225 drivers/firmware/memmap.c (!strcmp(entry->type, type))) { entry 226 drivers/firmware/memmap.c return entry; entry 283 drivers/firmware/memmap.c struct firmware_map_entry *entry; entry 285 drivers/firmware/memmap.c entry = firmware_map_find_entry(start, end - 1, type); entry 286 drivers/firmware/memmap.c if (entry) entry 289 drivers/firmware/memmap.c entry = firmware_map_find_entry_bootmem(start, end - 1, type); entry 290 drivers/firmware/memmap.c if (!entry) { entry 291 drivers/firmware/memmap.c entry = kzalloc(sizeof(struct firmware_map_entry), GFP_ATOMIC); entry 292 drivers/firmware/memmap.c if (!entry) entry 297 drivers/firmware/memmap.c list_del(&entry->list); entry 300 drivers/firmware/memmap.c memset(entry, 0, sizeof(*entry)); entry 303 drivers/firmware/memmap.c firmware_map_add_entry(start, end, type, entry); entry 305 drivers/firmware/memmap.c add_sysfs_fw_map_entry(entry); entry 325 drivers/firmware/memmap.c struct firmware_map_entry *entry; entry 327 drivers/firmware/memmap.c entry = memblock_alloc(sizeof(struct firmware_map_entry), entry 329 drivers/firmware/memmap.c if (WARN_ON(!entry)) entry 332 drivers/firmware/memmap.c return firmware_map_add_entry(start, end, type, entry); entry 347 drivers/firmware/memmap.c struct firmware_map_entry *entry; entry 350 drivers/firmware/memmap.c entry = firmware_map_find_entry(start, end - 1, type); entry 351 drivers/firmware/memmap.c if (!entry) { entry 356 drivers/firmware/memmap.c firmware_map_remove_entry(entry); entry 360 drivers/firmware/memmap.c remove_sysfs_fw_map_entry(entry); entry 369 drivers/firmware/memmap.c static ssize_t start_show(struct firmware_map_entry *entry, char *buf) entry 372 drivers/firmware/memmap.c (unsigned long long)entry->start); entry 375 drivers/firmware/memmap.c static ssize_t end_show(struct firmware_map_entry *entry, char *buf) entry 378 drivers/firmware/memmap.c (unsigned long long)entry->end); entry 381 drivers/firmware/memmap.c static ssize_t type_show(struct firmware_map_entry *entry, char *buf) entry 383 drivers/firmware/memmap.c return snprintf(buf, PAGE_SIZE, "%s\n", entry->type); entry 394 drivers/firmware/memmap.c struct firmware_map_entry *entry = to_memmap_entry(kobj); entry 397 drivers/firmware/memmap.c return memmap_attr->show(entry, buf); entry 410 drivers/firmware/memmap.c struct firmware_map_entry *entry; entry 412 drivers/firmware/memmap.c list_for_each_entry(entry, &map_entries, list) entry 413 drivers/firmware/memmap.c add_sysfs_fw_map_entry(entry); entry 29 drivers/firmware/qcom_scm-32.c void *entry; entry 336 drivers/firmware/qcom_scm-32.c int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) entry 358 drivers/firmware/qcom_scm-32.c flags, virt_to_phys(entry)); entry 369 drivers/firmware/qcom_scm-32.c int __qcom_scm_set_warm_boot_addr(struct device *dev, void *entry, entry 385 drivers/firmware/qcom_scm-32.c if (entry == qcom_scm_wb[cpu].entry) entry 394 drivers/firmware/qcom_scm-32.c cmd.addr = cpu_to_le32(virt_to_phys(entry)); entry 400 drivers/firmware/qcom_scm-32.c qcom_scm_wb[cpu].entry = entry; entry 167 drivers/firmware/qcom_scm-64.c int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) entry 181 drivers/firmware/qcom_scm-64.c int __qcom_scm_set_warm_boot_addr(struct device *dev, void *entry, entry 98 drivers/firmware/qcom_scm.c int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) entry 100 drivers/firmware/qcom_scm.c return __qcom_scm_set_cold_boot_addr(entry, cpus); entry 112 drivers/firmware/qcom_scm.c int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus) entry 114 drivers/firmware/qcom_scm.c return __qcom_scm_set_warm_boot_addr(__scm->dev, entry, cpus); entry 18 drivers/firmware/qcom_scm.h extern int __qcom_scm_set_warm_boot_addr(struct device *dev, void *entry, entry 20 drivers/firmware/qcom_scm.h extern int __qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus); entry 356 drivers/firmware/qemu_fw_cfg.c ssize_t (*show)(struct fw_cfg_sysfs_entry *entry, char *buf); entry 371 drivers/firmware/qemu_fw_cfg.c static inline void fw_cfg_sysfs_cache_enlist(struct fw_cfg_sysfs_entry *entry) entry 374 drivers/firmware/qemu_fw_cfg.c list_add_tail(&entry->list, &fw_cfg_entry_cache); entry 378 drivers/firmware/qemu_fw_cfg.c static inline void fw_cfg_sysfs_cache_delist(struct fw_cfg_sysfs_entry *entry) entry 381 drivers/firmware/qemu_fw_cfg.c list_del(&entry->list); entry 387 drivers/firmware/qemu_fw_cfg.c struct fw_cfg_sysfs_entry *entry, *next; entry 389 drivers/firmware/qemu_fw_cfg.c list_for_each_entry_safe(entry, next, &fw_cfg_entry_cache, list) { entry 393 drivers/firmware/qemu_fw_cfg.c kobject_put(&entry->kobj); entry 435 drivers/firmware/qemu_fw_cfg.c struct fw_cfg_sysfs_entry *entry = to_entry(kobj); entry 438 drivers/firmware/qemu_fw_cfg.c return attr->show(entry, buf); entry 448 drivers/firmware/qemu_fw_cfg.c struct fw_cfg_sysfs_entry *entry = to_entry(kobj); entry 450 drivers/firmware/qemu_fw_cfg.c fw_cfg_sysfs_cache_delist(entry); entry 451 drivers/firmware/qemu_fw_cfg.c kfree(entry); entry 466 drivers/firmware/qemu_fw_cfg.c struct fw_cfg_sysfs_entry *entry = to_entry(kobj); entry 468 drivers/firmware/qemu_fw_cfg.c if (pos > entry->size) entry 471 drivers/firmware/qemu_fw_cfg.c if (count > entry->size - pos) entry 472 drivers/firmware/qemu_fw_cfg.c count = entry->size - pos; entry 474 drivers/firmware/qemu_fw_cfg.c return fw_cfg_read_blob(entry->select, buf, pos, count); entry 566 drivers/firmware/qemu_fw_cfg.c list_for_each_entry_safe(k, next, &kset->list, entry) entry 584 drivers/firmware/qemu_fw_cfg.c struct fw_cfg_sysfs_entry *entry; entry 596 drivers/firmware/qemu_fw_cfg.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 597 drivers/firmware/qemu_fw_cfg.c if (!entry) entry 601 drivers/firmware/qemu_fw_cfg.c entry->size = be32_to_cpu(f->size); entry 602 drivers/firmware/qemu_fw_cfg.c entry->select = be16_to_cpu(f->select); entry 603 drivers/firmware/qemu_fw_cfg.c memcpy(entry->name, f->name, FW_CFG_MAX_FILE_PATH); entry 606 drivers/firmware/qemu_fw_cfg.c err = kobject_init_and_add(&entry->kobj, &fw_cfg_sysfs_entry_ktype, entry 607 drivers/firmware/qemu_fw_cfg.c fw_cfg_sel_ko, "%d", entry->select); entry 612 drivers/firmware/qemu_fw_cfg.c err = sysfs_create_bin_file(&entry->kobj, &fw_cfg_sysfs_attr_raw); entry 617 drivers/firmware/qemu_fw_cfg.c fw_cfg_build_symlink(fw_cfg_fname_kset, &entry->kobj, entry->name); entry 620 drivers/firmware/qemu_fw_cfg.c fw_cfg_sysfs_cache_enlist(entry); entry 624 drivers/firmware/qemu_fw_cfg.c kobject_del(&entry->kobj); entry 626 drivers/firmware/qemu_fw_cfg.c kfree(entry); entry 390 drivers/firmware/tegra/bpmp.c struct tegra_bpmp_mrq *entry; entry 392 drivers/firmware/tegra/bpmp.c list_for_each_entry(entry, &bpmp->mrqs, list) entry 393 drivers/firmware/tegra/bpmp.c if (entry->mrq == mrq) entry 394 drivers/firmware/tegra/bpmp.c return entry; entry 440 drivers/firmware/tegra/bpmp.c struct tegra_bpmp_mrq *entry; entry 445 drivers/firmware/tegra/bpmp.c entry = tegra_bpmp_find_mrq(bpmp, mrq); entry 446 drivers/firmware/tegra/bpmp.c if (!entry) { entry 452 drivers/firmware/tegra/bpmp.c entry->handler(mrq, channel, entry->data); entry 460 drivers/firmware/tegra/bpmp.c struct tegra_bpmp_mrq *entry; entry 466 drivers/firmware/tegra/bpmp.c entry = devm_kzalloc(bpmp->dev, sizeof(*entry), GFP_KERNEL); entry 467 drivers/firmware/tegra/bpmp.c if (!entry) entry 472 drivers/firmware/tegra/bpmp.c entry->mrq = mrq; entry 473 drivers/firmware/tegra/bpmp.c entry->handler = handler; entry 474 drivers/firmware/tegra/bpmp.c entry->data = data; entry 475 drivers/firmware/tegra/bpmp.c list_add(&entry->list, &bpmp->mrqs); entry 485 drivers/firmware/tegra/bpmp.c struct tegra_bpmp_mrq *entry; entry 490 drivers/firmware/tegra/bpmp.c entry = tegra_bpmp_find_mrq(bpmp, mrq); entry 491 drivers/firmware/tegra/bpmp.c if (!entry) entry 494 drivers/firmware/tegra/bpmp.c list_del(&entry->list); entry 495 drivers/firmware/tegra/bpmp.c devm_kfree(bpmp->dev, entry); entry 519 drivers/gpio/gpio-thunderx.c txgpio->msix_entries[i].entry = txgpio->base_msi + (2 * i); entry 746 drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c struct acpi_bus_event *entry = (struct acpi_bus_event *)data; entry 748 drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c if (strcmp(entry->device_class, ACPI_AC_CLASS) == 0) { entry 758 drivers/gpu/drm/amd/amdgpu/amdgpu_acpi.c return amdgpu_atif_handler(adev, entry); entry 86 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry; entry 88 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_for_each_entry(entry, &mem->bo_va_list, bo_list) entry 89 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (entry->bo_va->base.vm == avm) entry 433 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry, unsigned long size) entry 436 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->va, entry 437 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->va + size, entry); entry 438 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_vm_bo_rmv(adev, entry->bo_va); entry 439 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_del(&entry->bo_list); entry 440 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c kfree(entry); entry 447 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct ttm_validate_buffer *entry = &mem->validate_list; entry 450 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c INIT_LIST_HEAD(&entry->head); entry 451 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->num_shared = 1; entry 452 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->bo = &bo->tbo; entry 455 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_add_tail(&entry->head, &process_info->userptr_valid_list); entry 457 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_add_tail(&entry->head, &process_info->kfd_bo_list); entry 616 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry; entry 628 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_for_each_entry(entry, &mem->bo_va_list, bo_list) { entry 629 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if ((vm && vm != entry->bo_va->base.vm) || entry 630 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c (entry->is_mapped != map_type entry 650 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_for_each_entry(entry, &mem->bo_va_list, bo_list) { entry 651 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if ((vm && vm != entry->bo_va->base.vm) || entry 652 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c (entry->is_mapped != map_type entry 656 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_vm_get_pd_bo(entry->bo_va->base.vm, &ctx->list, entry 707 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry, entry 710 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo_va *bo_va = entry->bo_va; entry 713 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_vm_bo_unmap(adev, bo_va, entry->va); entry 723 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry, entry 727 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo_va *bo_va = entry->bo_va; entry 740 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry, struct amdgpu_sync *sync, entry 746 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_vm_bo_map(adev, entry->bo_va, entry->va, 0, entry 747 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_bo_size(entry->bo_va->base.bo), entry 748 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->pte_flags); entry 751 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->va, ret); entry 758 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = update_gpuvm_pte(adev, entry, sync); entry 767 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c unmap_bo_from_gpuvm(adev, entry, sync); entry 1231 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry, *tmp; entry 1273 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_for_each_entry_safe(entry, tmp, &mem->bo_va_list, bo_list) entry 1274 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c remove_bo_from_vm((struct amdgpu_device *)entry->kgd_dev, entry 1275 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry, bo_size); entry 1306 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry; entry 1388 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_for_each_entry(entry, &mem->bo_va_list, bo_list) { entry 1389 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (entry->bo_va->base.vm == vm && !entry->is_mapped) { entry 1391 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->va, entry->va + bo_size, entry 1392 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry); entry 1394 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = map_bo_to_gpuvm(adev, entry, ctx.sync, entry 1407 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->is_mapped = true; entry 1443 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct kfd_bo_va_list *entry; entry 1467 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c list_for_each_entry(entry, &mem->bo_va_list, bo_list) { entry 1468 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (entry->bo_va->base.vm == vm && entry->is_mapped) { entry 1470 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->va, entry 1471 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->va + bo_size, entry 1472 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry); entry 1474 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = unmap_bo_from_gpuvm(adev, entry, ctx.sync); entry 1476 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->is_mapped = false; entry 93 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c struct amdgpu_bo_list_entry *entry; entry 114 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c entry = &array[--first_userptr]; entry 116 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c entry = &array[last_entry++]; entry 119 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c entry->priority = min(info[i].bo_priority, entry 121 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c entry->tv.bo = &bo->tbo; entry 232 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ATOM_PPLIB_Clock_Voltage_Dependency_Record *entry; entry 238 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = &atom_table->entries[0]; entry 240 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c amdgpu_table->entries[i].clk = le16_to_cpu(entry->usClockLow) | entry 241 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c (entry->ucClockHigh << 16); entry 242 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c amdgpu_table->entries[i].v = le16_to_cpu(entry->usVoltage); entry 243 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = (ATOM_PPLIB_Clock_Voltage_Dependency_Record *) entry 244 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_Clock_Voltage_Dependency_Record)); entry 397 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ATOM_PPLIB_PhaseSheddingLimits_Record *entry; entry 408 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = &psl->entries[0]; entry 411 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usSclkLow) | (entry->ucSclkHigh << 16); entry 413 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usMclkLow) | (entry->ucMclkHigh << 16); entry 415 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVoltage); entry 416 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = (ATOM_PPLIB_PhaseSheddingLimits_Record *) entry 417 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_PhaseSheddingLimits_Record)); entry 444 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ATOM_PPLIB_CAC_Leakage_Record *entry; entry 451 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = &cac_table->entries[0]; entry 455 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVddc1); entry 457 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVddc2); entry 459 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVddc3); entry 462 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVddc); entry 464 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le32_to_cpu(entry->ulLeakageValue); entry 466 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = (ATOM_PPLIB_CAC_Leakage_Record *) entry 467 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_CAC_Leakage_Record)); entry 495 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record *entry; entry 508 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = &limits->entries[0]; entry 513 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c (entry->ucVCEClockInfoIndex * sizeof(VCEClockInfo))); entry 519 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVoltage); entry 520 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = (ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record *) entry 521 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record)); entry 552 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *entry; entry 563 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = &limits->entries[0]; entry 567 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c (entry->ucUVDClockInfoIndex * sizeof(UVDClockInfo))); entry 573 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVoltage); entry 574 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = (ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *) entry 575 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record)); entry 584 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ATOM_PPLIB_SAMClk_Voltage_Limit_Record *entry; entry 595 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = &limits->entries[0]; entry 598 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usSAMClockLow) | (entry->ucSAMClockHigh << 16); entry 600 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVoltage); entry 601 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = (ATOM_PPLIB_SAMClk_Voltage_Limit_Record *) entry 602 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_SAMClk_Voltage_Limit_Record)); entry 642 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ATOM_PPLIB_ACPClk_Voltage_Limit_Record *entry; entry 653 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = &limits->entries[0]; entry 656 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usACPClockLow) | (entry->ucACPClockHigh << 16); entry 658 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c le16_to_cpu(entry->usVoltage); entry 659 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = (ATOM_PPLIB_ACPClk_Voltage_Limit_Record *) entry 660 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_ACPClk_Voltage_Limit_Record)); entry 63 drivers/gpu/drm/amd/amdgpu/amdgpu_ih.h struct amdgpu_iv_entry *entry); entry 380 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c struct amdgpu_iv_entry entry; entry 386 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c entry.iv_entry = (const uint32_t *)&ih->ring[ring_index]; entry 387 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c amdgpu_ih_decode_iv(adev, &entry); entry 389 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c trace_amdgpu_iv(ih - &adev->irq.ih, &entry); entry 391 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c client_id = entry.client_id; entry 392 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c src_id = entry.src_id; entry 408 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c r = src->funcs->process(adev, src, &entry); entry 420 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c amdgpu_amdkfd_interrupt(adev, entry.iv_entry); entry 78 drivers/gpu/drm/amd/amdgpu/amdgpu_irq.h struct amdgpu_iv_entry *entry); entry 36 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c struct list_head entry; entry 241 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c list_add_tail(&pmu_entry->entry, &amdgpu_pmu_list); entry 273 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c list_for_each_entry_safe(pe, temp, &amdgpu_pmu_list, entry) { entry 275 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c list_del(&pe->entry); entry 1006 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c struct amdgpu_iv_entry entry; entry 1012 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c memcpy(&entry, &data->ring[data->rptr], entry 1023 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c ret = data->cb(obj->adev, &err_data, &entry); entry 1063 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c memcpy(&data->ring[data->wptr], info->entry, entry 369 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h struct amdgpu_iv_entry *entry); entry 446 drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h struct amdgpu_iv_entry *entry; entry 350 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm_pt *entry; entry 370 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c cursor->entry = &vm->root; entry 389 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (!cursor->entry->entries) entry 392 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c BUG_ON(!cursor->entry->base.bo); entry 398 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c cursor->parent = cursor->entry; entry 399 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c cursor->entry = &cursor->entry->entries[idx]; entry 426 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (cursor->entry == &cursor->parent->entries[num_entries - 1]) entry 431 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c ++cursor->entry; entry 450 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c cursor->entry = cursor->parent; entry 511 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm_pt *entry) entry 513 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c return entry && (!start || entry != start->entry); entry 527 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (!cursor->entry) entry 531 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c cursor->entry = NULL; entry 541 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c #define for_each_amdgpu_vm_pt_dfs_safe(adev, vm, start, cursor, entry) \ entry 543 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c (entry) = (cursor).entry, amdgpu_vm_pt_next_dfs((adev), &(cursor));\ entry 544 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_pt_continue_dfs((start), (entry)); \ entry 545 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c (entry) = (cursor).entry, amdgpu_vm_pt_next_dfs((adev), &(cursor))) entry 559 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_list_entry *entry) entry 561 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry->priority = 0; entry 562 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry->tv.bo = &vm->root.base.bo->tbo; entry 564 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry->tv.num_shared = 3; entry 565 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry->user_pages = NULL; entry 566 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_add(&entry->tv.head, validated); entry 855 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm_pt *entry = cursor->entry; entry 860 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (cursor->level < AMDGPU_VM_PTB && !entry->entries) { entry 864 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry->entries = kvmalloc_array(num_entries, entry 865 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c sizeof(*entry->entries), entry 867 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (!entry->entries) entry 871 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (entry->base.bo) entry 884 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_base_init(&entry->base, vm, pt); entry 903 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c static void amdgpu_vm_free_table(struct amdgpu_vm_pt *entry) entry 905 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (entry->base.bo) { entry 906 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry->base.bo->vm_bo = NULL; entry 907 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_del(&entry->base.vm_status); entry 908 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_bo_unref(&entry->base.bo->shadow); entry 909 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_bo_unref(&entry->base.bo); entry 911 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c kvfree(entry->entries); entry 912 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry->entries = NULL; entry 929 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm_pt *entry; entry 933 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c for_each_amdgpu_vm_pt_dfs_safe(adev, vm, start, cursor, entry) entry 934 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_free_table(entry); entry 937 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_free_table(start->entry); entry 1192 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm_pt *entry) entry 1194 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm_pt *parent = amdgpu_vm_pt_parent(entry); entry 1203 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_gmc_get_pde_for_bo(entry->base.bo, level, &pt, &flags); entry 1204 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c pde = (entry - parent->entries) * 8; entry 1220 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm_pt *entry; entry 1222 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c for_each_amdgpu_vm_pt_dfs_safe(adev, vm, NULL, cursor, entry) entry 1223 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (entry->base.bo && !entry->base.moved) entry 1224 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_relocated(&entry->base); entry 1256 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm_pt *entry; entry 1258 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry = list_first_entry(&vm->relocated, struct amdgpu_vm_pt, entry 1260 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_idle(&entry->base); entry 1262 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c r = amdgpu_vm_update_pde(¶ms, vm, entry); entry 1404 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c pt = cursor.entry->base.bo; entry 354 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h struct amdgpu_bo_list_entry *entry); entry 330 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c struct amdgpu_xgmi *entry; entry 369 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c list_for_each_entry(entry, &hive->device_list, head) entry 370 drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c top_info->nodes[count++].node_id = entry->node_id; entry 243 drivers/gpu/drm/amd/amdgpu/cik_ih.c struct amdgpu_iv_entry *entry) entry 254 drivers/gpu/drm/amd/amdgpu/cik_ih.c entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY; entry 255 drivers/gpu/drm/amd/amdgpu/cik_ih.c entry->src_id = dw[0] & 0xff; entry 256 drivers/gpu/drm/amd/amdgpu/cik_ih.c entry->src_data[0] = dw[1] & 0xfffffff; entry 257 drivers/gpu/drm/amd/amdgpu/cik_ih.c entry->ring_id = dw[2] & 0xff; entry 258 drivers/gpu/drm/amd/amdgpu/cik_ih.c entry->vmid = (dw[2] >> 8) & 0xff; entry 259 drivers/gpu/drm/amd/amdgpu/cik_ih.c entry->pasid = (dw[2] >> 16) & 0xffff; entry 1159 drivers/gpu/drm/amd/amdgpu/cik_sdma.c struct amdgpu_iv_entry *entry) entry 1163 drivers/gpu/drm/amd/amdgpu/cik_sdma.c instance_id = (entry->ring_id & 0x3) >> 0; entry 1164 drivers/gpu/drm/amd/amdgpu/cik_sdma.c queue_id = (entry->ring_id & 0xc) >> 2; entry 1200 drivers/gpu/drm/amd/amdgpu/cik_sdma.c struct amdgpu_iv_entry *entry) entry 1205 drivers/gpu/drm/amd/amdgpu/cik_sdma.c instance_id = (entry->ring_id & 0x3) >> 0; entry 222 drivers/gpu/drm/amd/amdgpu/cz_ih.c struct amdgpu_iv_entry *entry) entry 233 drivers/gpu/drm/amd/amdgpu/cz_ih.c entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY; entry 234 drivers/gpu/drm/amd/amdgpu/cz_ih.c entry->src_id = dw[0] & 0xff; entry 235 drivers/gpu/drm/amd/amdgpu/cz_ih.c entry->src_data[0] = dw[1] & 0xfffffff; entry 236 drivers/gpu/drm/amd/amdgpu/cz_ih.c entry->ring_id = dw[2] & 0xff; entry 237 drivers/gpu/drm/amd/amdgpu/cz_ih.c entry->vmid = (dw[2] >> 8) & 0xff; entry 238 drivers/gpu/drm/amd/amdgpu/cz_ih.c entry->pasid = (dw[2] >> 16) & 0xffff; entry 3113 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c struct amdgpu_iv_entry *entry) entry 3120 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c crtc_id = (entry->src_id - 8) >> 1; entry 3211 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c struct amdgpu_iv_entry *entry) entry 3213 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c unsigned crtc = entry->src_id - 1; entry 3217 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c switch (entry->src_data[0]) { entry 3240 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]); entry 3249 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c struct amdgpu_iv_entry *entry) entry 3254 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c if (entry->src_data[0] >= adev->mode_info.num_hpd) { entry 3255 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]); entry 3259 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c hpd = entry->src_data[0]; entry 3239 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c struct amdgpu_iv_entry *entry) entry 3246 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c crtc_id = (entry->src_id - 8) >> 1; entry 3337 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c struct amdgpu_iv_entry *entry) entry 3339 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c unsigned crtc = entry->src_id - 1; entry 3344 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c switch (entry->src_data[0]) { entry 3367 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]); entry 3376 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c struct amdgpu_iv_entry *entry) entry 3381 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c if (entry->src_data[0] >= adev->mode_info.num_hpd) { entry 3382 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]); entry 3386 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c hpd = entry->src_data[0]; entry 2930 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c struct amdgpu_iv_entry *entry) entry 2932 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c unsigned crtc = entry->src_id - 1; entry 2937 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c switch (entry->src_data[0]) { entry 2958 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]); entry 2990 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c struct amdgpu_iv_entry *entry) entry 2997 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c crtc_id = (entry->src_id - 8) >> 1; entry 3043 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c struct amdgpu_iv_entry *entry) entry 3048 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c if (entry->src_data[0] >= adev->mode_info.num_hpd) { entry 3049 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]); entry 3053 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c hpd = entry->src_data[0]; entry 3022 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c struct amdgpu_iv_entry *entry) entry 3024 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c unsigned crtc = entry->src_id - 1; entry 3029 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c switch (entry->src_data[0]) { entry 3050 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]); entry 3082 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c struct amdgpu_iv_entry *entry) entry 3089 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c crtc_id = (entry->src_id - 8) >> 1; entry 3135 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c struct amdgpu_iv_entry *entry) entry 3140 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c if (entry->src_data[0] >= adev->mode_info.num_hpd) { entry 3141 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c DRM_DEBUG("Unhandled interrupt: %d %d\n", entry->src_id, entry->src_data[0]); entry 3145 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c hpd = entry->src_data[0]; entry 4992 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c struct amdgpu_iv_entry *entry) entry 4999 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 5000 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 5001 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c queue_id = (entry->ring_id & 0x70) >> 4; entry 5063 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c struct amdgpu_iv_entry *entry) entry 5069 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 5070 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 5071 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c queue_id = (entry->ring_id & 0x70) >> 4; entry 5098 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c struct amdgpu_iv_entry *entry) entry 5101 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c gfx_v10_0_handle_priv_fault(adev, entry); entry 5107 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c struct amdgpu_iv_entry *entry) entry 5110 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c gfx_v10_0_handle_priv_fault(adev, entry); entry 5161 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c struct amdgpu_iv_entry *entry) entry 5166 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 5167 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 5168 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c queue_id = (entry->ring_id & 0x70) >> 4; entry 3370 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c struct amdgpu_iv_entry *entry) entry 3372 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c switch (entry->ring_id) { entry 3378 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c amdgpu_fence_process(&adev->gfx.compute_ring[entry->ring_id - 1]); entry 3387 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c struct amdgpu_iv_entry *entry) entry 3391 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c switch (entry->ring_id) { entry 3397 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c ring = &adev->gfx.compute_ring[entry->ring_id - 1]; entry 3407 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c struct amdgpu_iv_entry *entry) entry 3410 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c gfx_v6_0_fault(adev, entry); entry 3416 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c struct amdgpu_iv_entry *entry) entry 3419 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c gfx_v6_0_fault(adev, entry); entry 4865 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c struct amdgpu_iv_entry *entry) entry 4872 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 4873 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 4891 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c struct amdgpu_iv_entry *entry) entry 4897 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 4898 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 4916 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c struct amdgpu_iv_entry *entry) entry 4919 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c gfx_v7_0_fault(adev, entry); entry 4925 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c struct amdgpu_iv_entry *entry) entry 4929 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c gfx_v7_0_fault(adev, entry); entry 6729 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c struct amdgpu_iv_entry *entry) entry 6736 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 6737 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 6738 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c queue_id = (entry->ring_id & 0x70) >> 4; entry 6760 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c struct amdgpu_iv_entry *entry) entry 6766 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 6767 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 6768 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c queue_id = (entry->ring_id & 0x70) >> 4; entry 6788 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c struct amdgpu_iv_entry *entry) entry 6791 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c gfx_v8_0_fault(adev, entry); entry 6797 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c struct amdgpu_iv_entry *entry) entry 6800 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c gfx_v8_0_fault(adev, entry); entry 6806 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c struct amdgpu_iv_entry *entry) entry 6893 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c struct amdgpu_iv_entry *entry) entry 6895 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c unsigned ih_data = entry->src_data[0]; entry 4417 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c struct amdgpu_iv_entry *entry); entry 5664 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c struct amdgpu_iv_entry *entry) entry 5671 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 5672 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 5673 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c queue_id = (entry->ring_id & 0x70) >> 4; entry 5695 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c struct amdgpu_iv_entry *entry) entry 5701 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c me_id = (entry->ring_id & 0x0c) >> 2; entry 5702 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c pipe_id = (entry->ring_id & 0x03) >> 0; entry 5703 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c queue_id = (entry->ring_id & 0x70) >> 4; entry 5723 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c struct amdgpu_iv_entry *entry) entry 5726 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c gfx_v9_0_fault(adev, entry); entry 5732 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c struct amdgpu_iv_entry *entry) entry 5735 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c gfx_v9_0_fault(adev, entry); entry 5741 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c struct amdgpu_iv_entry *entry) entry 6161 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c struct amdgpu_iv_entry *entry) entry 6165 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c .entry = entry, entry 129 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c struct amdgpu_iv_entry *entry) entry 131 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c struct amdgpu_vmhub *hub = &adev->vmhub[entry->vmid_src]; entry 135 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c addr = (u64)entry->src_data[0] << 12; entry 136 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c addr |= ((u64)entry->src_data[1] & 0xf) << 44; entry 144 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c if (entry->vmid_src == AMDGPU_GFXHUB_0) entry 155 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c amdgpu_vm_get_task_info(adev, entry->pasid, &task_info); entry 160 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c entry->vmid_src ? "mmhub" : "gfxhub", entry 161 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c entry->src_id, entry->ring_id, entry->vmid, entry 162 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c entry->pasid, task_info.process_name, task_info.tgid, entry 165 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c addr, entry->client_id); entry 1095 drivers/gpu/drm/amd/amdgpu/gmc_v6_0.c struct amdgpu_iv_entry *entry) entry 1111 drivers/gpu/drm/amd/amdgpu/gmc_v6_0.c entry->src_id, entry->src_data[0]); entry 1245 drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c struct amdgpu_iv_entry *entry) entry 1263 drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c entry->src_id, entry->src_data[0]); entry 1269 drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c entry->pasid); entry 1417 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c struct amdgpu_iv_entry *entry) entry 1423 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c entry->src_id, entry->src_data[0]); entry 1444 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c amdgpu_vm_get_task_info(adev, entry->pasid, &task_info); entry 1447 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c entry->src_id, entry->src_data[0], task_info.process_name, entry 1454 drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c entry->pasid); entry 248 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c struct amdgpu_iv_entry *entry) entry 268 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c struct amdgpu_iv_entry *entry) entry 272 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c .entry = entry, entry 331 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c struct amdgpu_iv_entry *entry) entry 334 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c bool retry_fault = !!(entry->src_data[1] & 0x80); entry 339 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c addr = (u64)entry->src_data[0] << 12; entry 340 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c addr |= ((u64)entry->src_data[1] & 0xf) << 44; entry 342 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c if (retry_fault && amdgpu_gmc_filter_faults(adev, addr, entry->pasid, entry 343 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c entry->timestamp)) entry 346 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c if (entry->client_id == SOC15_IH_CLIENTID_VMC) { entry 349 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c } else if (entry->client_id == SOC15_IH_CLIENTID_VMC1) { entry 364 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c if (entry->vmid_src == AMDGPU_GFXHUB_0) entry 375 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c amdgpu_vm_get_task_info(adev, entry->pasid, &task_info); entry 381 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c entry->src_id, entry->ring_id, entry->vmid, entry 382 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c entry->pasid, task_info.process_name, task_info.tgid, entry 385 drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c addr, entry->client_id); entry 222 drivers/gpu/drm/amd/amdgpu/iceland_ih.c struct amdgpu_iv_entry *entry) entry 233 drivers/gpu/drm/amd/amdgpu/iceland_ih.c entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY; entry 234 drivers/gpu/drm/amd/amdgpu/iceland_ih.c entry->src_id = dw[0] & 0xff; entry 235 drivers/gpu/drm/amd/amdgpu/iceland_ih.c entry->src_data[0] = dw[1] & 0xfffffff; entry 236 drivers/gpu/drm/amd/amdgpu/iceland_ih.c entry->ring_id = dw[2] & 0xff; entry 237 drivers/gpu/drm/amd/amdgpu/iceland_ih.c entry->vmid = (dw[2] >> 8) & 0xff; entry 238 drivers/gpu/drm/amd/amdgpu/iceland_ih.c entry->pasid = (dw[2] >> 16) & 0xffff; entry 3188 drivers/gpu/drm/amd/amdgpu/kv_dpm.c struct amdgpu_iv_entry *entry) entry 3192 drivers/gpu/drm/amd/amdgpu/kv_dpm.c if (entry == NULL) entry 3195 drivers/gpu/drm/amd/amdgpu/kv_dpm.c switch (entry->src_id) { entry 292 drivers/gpu/drm/amd/amdgpu/mxgpu_ai.c struct amdgpu_iv_entry *entry) entry 367 drivers/gpu/drm/amd/amdgpu/mxgpu_ai.c struct amdgpu_iv_entry *entry) entry 491 drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c struct amdgpu_iv_entry *entry) entry 543 drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c struct amdgpu_iv_entry *entry) entry 254 drivers/gpu/drm/amd/amdgpu/navi10_ih.c struct amdgpu_iv_entry *entry) entry 269 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->client_id = dw[0] & 0xff; entry 270 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->src_id = (dw[0] >> 8) & 0xff; entry 271 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->ring_id = (dw[0] >> 16) & 0xff; entry 272 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->vmid = (dw[0] >> 24) & 0xf; entry 273 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->vmid_src = (dw[0] >> 31); entry 274 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->timestamp = dw[1] | ((u64)(dw[2] & 0xffff) << 32); entry 275 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->timestamp_src = dw[2] >> 31; entry 276 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->pasid = dw[3] & 0xffff; entry 277 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->pasid_src = dw[3] >> 31; entry 278 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->src_data[0] = dw[4]; entry 279 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->src_data[1] = dw[5]; entry 280 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->src_data[2] = dw[6]; entry 281 drivers/gpu/drm/amd/amdgpu/navi10_ih.c entry->src_data[3] = dw[7]; entry 1052 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c struct amdgpu_iv_entry *entry) entry 1056 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c instance_id = (entry->ring_id & 0x3) >> 0; entry 1057 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c queue_id = (entry->ring_id & 0xc) >> 2; entry 1092 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c struct amdgpu_iv_entry *entry) entry 1097 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c instance_id = (entry->ring_id & 0x3) >> 0; entry 1098 drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c queue_id = (entry->ring_id & 0xc) >> 2; entry 1386 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c struct amdgpu_iv_entry *entry) entry 1390 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c instance_id = (entry->ring_id & 0x3) >> 0; entry 1391 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c queue_id = (entry->ring_id & 0xc) >> 2; entry 1426 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c struct amdgpu_iv_entry *entry) entry 1431 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c instance_id = (entry->ring_id & 0x3) >> 0; entry 1432 drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c queue_id = (entry->ring_id & 0xc) >> 2; entry 1695 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_iv_entry *entry); entry 2002 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_iv_entry *entry) entry 2007 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c instance = sdma_v4_0_irq_id_to_seq(entry->client_id); entry 2008 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c switch (entry->ring_id) { entry 2029 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_iv_entry *entry) entry 2034 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c instance = sdma_v4_0_irq_id_to_seq(entry->client_id); entry 2038 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c switch (entry->src_id) { entry 2058 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_iv_entry *entry) entry 2062 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c .entry = entry, entry 2076 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c struct amdgpu_iv_entry *entry) entry 2082 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c instance = sdma_v4_0_irq_id_to_seq(entry->client_id); entry 2086 drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c switch (entry->ring_id) { entry 1428 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c struct amdgpu_iv_entry *entry) entry 1431 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c switch (entry->client_id) { entry 1433 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c switch (entry->ring_id) { entry 1449 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c switch (entry->ring_id) { entry 1470 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c struct amdgpu_iv_entry *entry) entry 634 drivers/gpu/drm/amd/amdgpu/si_dma.c struct amdgpu_iv_entry *entry) entry 636 drivers/gpu/drm/amd/amdgpu/si_dma.c if (entry->src_id == 224) entry 6057 drivers/gpu/drm/amd/amdgpu/si_dpm.c static void si_convert_mc_registers(const struct si_mc_reg_entry *entry, entry 6065 drivers/gpu/drm/amd/amdgpu/si_dpm.c data->value[i] = cpu_to_be32(entry->mc_data[j]); entry 7554 drivers/gpu/drm/amd/amdgpu/si_dpm.c struct amdgpu_iv_entry *entry) entry 7558 drivers/gpu/drm/amd/amdgpu/si_dpm.c if (entry == NULL) entry 7561 drivers/gpu/drm/amd/amdgpu/si_dpm.c switch (entry->src_id) { entry 126 drivers/gpu/drm/amd/amdgpu/si_ih.c struct amdgpu_iv_entry *entry) entry 136 drivers/gpu/drm/amd/amdgpu/si_ih.c entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY; entry 137 drivers/gpu/drm/amd/amdgpu/si_ih.c entry->src_id = dw[0] & 0xff; entry 138 drivers/gpu/drm/amd/amdgpu/si_ih.c entry->src_data[0] = dw[1] & 0xfffffff; entry 139 drivers/gpu/drm/amd/amdgpu/si_ih.c entry->ring_id = dw[2] & 0xff; entry 140 drivers/gpu/drm/amd/amdgpu/si_ih.c entry->vmid = (dw[2] >> 8) & 0xff; entry 440 drivers/gpu/drm/amd/amdgpu/soc15.c const struct soc15_reg_golden *entry; entry 445 drivers/gpu/drm/amd/amdgpu/soc15.c entry = ®s[i]; entry 446 drivers/gpu/drm/amd/amdgpu/soc15.c reg = adev->reg_offset[entry->hwip][entry->instance][entry->segment] + entry->reg; entry 448 drivers/gpu/drm/amd/amdgpu/soc15.c if (entry->and_mask == 0xffffffff) { entry 449 drivers/gpu/drm/amd/amdgpu/soc15.c tmp = entry->or_mask; entry 452 drivers/gpu/drm/amd/amdgpu/soc15.c tmp &= ~(entry->and_mask); entry 453 drivers/gpu/drm/amd/amdgpu/soc15.c tmp |= (entry->or_mask & entry->and_mask); entry 65 drivers/gpu/drm/amd/amdgpu/soc15.h #define SOC15_REG_ENTRY_OFFSET(entry) (adev->reg_offset[entry.hwip][entry.inst][entry.seg] + entry.reg_offset) entry 224 drivers/gpu/drm/amd/amdgpu/tonga_ih.c struct amdgpu_iv_entry *entry) entry 235 drivers/gpu/drm/amd/amdgpu/tonga_ih.c entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY; entry 236 drivers/gpu/drm/amd/amdgpu/tonga_ih.c entry->src_id = dw[0] & 0xff; entry 237 drivers/gpu/drm/amd/amdgpu/tonga_ih.c entry->src_data[0] = dw[1] & 0xfffffff; entry 238 drivers/gpu/drm/amd/amdgpu/tonga_ih.c entry->ring_id = dw[2] & 0xff; entry 239 drivers/gpu/drm/amd/amdgpu/tonga_ih.c entry->vmid = (dw[2] >> 8) & 0xff; entry 240 drivers/gpu/drm/amd/amdgpu/tonga_ih.c entry->pasid = (dw[2] >> 16) & 0xffff; entry 672 drivers/gpu/drm/amd/amdgpu/uvd_v4_2.c struct amdgpu_iv_entry *entry) entry 595 drivers/gpu/drm/amd/amdgpu/uvd_v5_0.c struct amdgpu_iv_entry *entry) entry 1218 drivers/gpu/drm/amd/amdgpu/uvd_v6_0.c struct amdgpu_iv_entry *entry) entry 1223 drivers/gpu/drm/amd/amdgpu/uvd_v6_0.c switch (entry->src_id) { entry 1243 drivers/gpu/drm/amd/amdgpu/uvd_v6_0.c entry->src_id, entry->src_data[0]); entry 1542 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c struct amdgpu_iv_entry *entry) entry 1546 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c switch (entry->client_id) { entry 1554 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c DRM_ERROR("Unhandled client id: %d\n", entry->client_id); entry 1560 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c switch (entry->src_id) { entry 1573 drivers/gpu/drm/amd/amdgpu/uvd_v7_0.c entry->src_id, entry->src_data[0]); entry 533 drivers/gpu/drm/amd/amdgpu/vce_v2_0.c struct amdgpu_iv_entry *entry) entry 536 drivers/gpu/drm/amd/amdgpu/vce_v2_0.c switch (entry->src_data[0]) { entry 539 drivers/gpu/drm/amd/amdgpu/vce_v2_0.c amdgpu_fence_process(&adev->vce.ring[entry->src_data[0]]); entry 543 drivers/gpu/drm/amd/amdgpu/vce_v2_0.c entry->src_id, entry->src_data[0]); entry 717 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c struct amdgpu_iv_entry *entry) entry 723 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c switch (entry->src_data[0]) { entry 727 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c amdgpu_fence_process(&adev->vce.ring[entry->src_data[0]]); entry 731 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c entry->src_id, entry->src_data[0]); entry 1028 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c struct amdgpu_iv_entry *entry) entry 1032 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c switch (entry->src_data[0]) { entry 1036 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c amdgpu_fence_process(&adev->vce.ring[entry->src_data[0]]); entry 1040 drivers/gpu/drm/amd/amdgpu/vce_v4_0.c entry->src_id, entry->src_data[0]); entry 2113 drivers/gpu/drm/amd/amdgpu/vcn_v1_0.c struct amdgpu_iv_entry *entry) entry 2117 drivers/gpu/drm/amd/amdgpu/vcn_v1_0.c switch (entry->src_id) { entry 2132 drivers/gpu/drm/amd/amdgpu/vcn_v1_0.c entry->src_id, entry->src_data[0]); entry 2069 drivers/gpu/drm/amd/amdgpu/vcn_v2_0.c struct amdgpu_iv_entry *entry) entry 2073 drivers/gpu/drm/amd/amdgpu/vcn_v2_0.c switch (entry->src_id) { entry 2088 drivers/gpu/drm/amd/amdgpu/vcn_v2_0.c entry->src_id, entry->src_data[0]); entry 1328 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c struct amdgpu_iv_entry *entry) entry 1332 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c switch (entry->client_id) { entry 1340 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c DRM_ERROR("Unhandled client id: %d\n", entry->client_id); entry 1346 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c switch (entry->src_id) { entry 1361 drivers/gpu/drm/amd/amdgpu/vcn_v2_5.c entry->src_id, entry->src_data[0]); entry 437 drivers/gpu/drm/amd/amdgpu/vega10_ih.c struct amdgpu_iv_entry *entry) entry 452 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->client_id = dw[0] & 0xff; entry 453 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->src_id = (dw[0] >> 8) & 0xff; entry 454 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->ring_id = (dw[0] >> 16) & 0xff; entry 455 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->vmid = (dw[0] >> 24) & 0xf; entry 456 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->vmid_src = (dw[0] >> 31); entry 457 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->timestamp = dw[1] | ((u64)(dw[2] & 0xffff) << 32); entry 458 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->timestamp_src = dw[2] >> 31; entry 459 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->pasid = dw[3] & 0xffff; entry 460 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->pasid_src = dw[3] >> 31; entry 461 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->src_data[0] = dw[4]; entry 462 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->src_data[1] = dw[5]; entry 463 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->src_data[2] = dw[6]; entry 464 drivers/gpu/drm/amd/amdgpu/vega10_ih.c entry->src_data[3] = dw[7]; entry 539 drivers/gpu/drm/amd/amdgpu/vega10_ih.c struct amdgpu_iv_entry *entry) entry 541 drivers/gpu/drm/amd/amdgpu/vega10_ih.c uint32_t wptr = cpu_to_le32(entry->src_data[0]); entry 543 drivers/gpu/drm/amd/amdgpu/vega10_ih.c switch (entry->ring_id) { entry 312 drivers/gpu/drm/amd/amdkfd/kfd_crat.h uint8_t entry[1]; entry 243 drivers/gpu/drm/amd/amdkfd/kfd_events.c list_for_each_entry(waiter, &ev->wq.head, wait.entry) entry 400 drivers/gpu/drm/amd/amdkfd/kfd_events.c list_for_each_entry(waiter, &ev->wq.head, wait.entry) entry 35 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CLIENT_ID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) & 0xff) entry 36 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_SOURCE_ID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) >> 8 & 0xff) entry 37 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_RING_ID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) >> 16 & 0xff) entry 38 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_VMID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) >> 24 & 0xf) entry 39 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_VMID_TYPE_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[0]) >> 31 & 0x1) entry 40 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_PASID_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[3]) & 0xffff) entry 41 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CONTEXT_ID0_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[4])) entry 42 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CONTEXT_ID1_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[5])) entry 43 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CONTEXT_ID2_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[6])) entry 44 drivers/gpu/drm/amd/amdkfd/soc15_int.h #define SOC15_CONTEXT_ID3_FROM_IH_ENTRY(entry) (le32_to_cpu(entry[7])) entry 114 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c struct list_head *entry; entry 120 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c list_for_each(entry, handler_list) { entry 121 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c handler_data = list_entry(entry, entry 148 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c struct list_head *entry, *tmp; entry 168 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c list_for_each_safe(entry, tmp, hnd_list) { entry 170 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c handler = list_entry(entry, struct amdgpu_dm_irq_handler_data, entry 531 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c struct list_head *entry; entry 537 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c entry, entry 540 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c handler_data = list_entry(entry, entry 563 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c struct amdgpu_iv_entry *entry) entry 569 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c entry->src_id, entry 570 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c entry->src_data[0]); entry 2989 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c uint32_t entry; entry 3004 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c for (entry = 0; entry < 16; entry += 2) { entry 3007 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c link_rate_in_khz = (supported_link_rates[entry+1] * 0x100 + entry 3008 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c supported_link_rates[entry]) * 200; entry 152 drivers/gpu/drm/amd/display/include/logger_interface.h #define DISPLAY_STATS_BEGIN(entry) (void)(entry) entry 156 drivers/gpu/drm/amd/display/include/logger_interface.h #define DISPLAY_STATS_END(entry) (void)(entry) entry 83 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c const struct soc15_baco_cmd_entry *entry, entry 90 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c if ((entry[i].cmd == CMD_WRITE) || entry 91 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c (entry[i].cmd == CMD_READMODIFYWRITE) || entry 92 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c (entry[i].cmd == CMD_WAITFOR)) entry 93 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c reg = adev->reg_offset[entry[i].hwip][entry[i].inst][entry[i].seg] entry 94 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c + entry[i].reg_offset; entry 95 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c if (!baco_cmd_handler(hwmgr, entry[i].cmd, reg, entry[i].mask, entry 96 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.c entry[i].shift, entry[i].val, entry[i].timeout)) entry 48 drivers/gpu/drm/amd/powerplay/hwmgr/common_baco.h const struct soc15_baco_cmd_entry *entry, entry 1374 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c table->entry[i].ucVco_setting = psmu_info->asSclkFcwRangeEntry[i].ucVco_setting; entry 1375 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c table->entry[i].ucPostdiv = psmu_info->asSclkFcwRangeEntry[i].ucPostdiv; entry 1376 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c table->entry[i].usFcw_pcc = entry 1378 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c table->entry[i].usFcw_trans_upper = entry 1380 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.c table->entry[i].usRcw_trans_lower = entry 244 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomctrl.h struct pp_atom_ctrl_sclk_range_table_entry entry[MAX_SCLK_RANGE]; entry 1126 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c const UVDClockInfo *entry = entry 1129 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c uvd_table->entries[i].vclk = ((unsigned long)entry->ucVClkHigh << 16) entry 1130 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c | le16_to_cpu(entry->usVClkLow); entry 1131 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c uvd_table->entries[i].dclk = ((unsigned long)entry->ucDClkHigh << 16) entry 1132 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c | le16_to_cpu(entry->usDClkLow); entry 1158 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c const VCEClockInfo *entry = &array->entries[table->entries[i].ucVCEClockInfoIndex]; entry 1161 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c vce_table->entries[i].evclk = ((unsigned long)entry->ucEVClkHigh << 16) entry 1162 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c | le16_to_cpu(entry->usEVClkLow); entry 1163 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c vce_table->entries[i].ecclk = ((unsigned long)entry->ucECClkHigh << 16) entry 1164 drivers/gpu/drm/amd/powerplay/hwmgr/processpptables.c | le16_to_cpu(entry->usECClkLow); entry 779 drivers/gpu/drm/amd/powerplay/hwmgr/smu10_hwmgr.c unsigned long entry, struct pp_power_state *ps) entry 788 drivers/gpu/drm/amd/powerplay/hwmgr/smu10_hwmgr.c result = pp_tables_get_entry(hwmgr, entry, ps, entry 108 drivers/gpu/drm/amd/powerplay/hwmgr/smu10_hwmgr.h uint32_t entry : 1; entry 1374 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.c unsigned long entry, struct pp_power_state *ps) entry 1383 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.c result = pp_tables_get_entry(hwmgr, entry, ps, entry 131 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.h uint32_t entry : 1; entry 590 drivers/gpu/drm/amd/powerplay/hwmgr/smu_helper.c struct amdgpu_iv_entry *entry) entry 592 drivers/gpu/drm/amd/powerplay/hwmgr/smu_helper.c uint32_t client_id = entry->client_id; entry 593 drivers/gpu/drm/amd/powerplay/hwmgr/smu_helper.c uint32_t src_id = entry->src_id; entry 110 drivers/gpu/drm/amd/powerplay/hwmgr/smu_helper.h struct amdgpu_iv_entry *entry); entry 98 drivers/gpu/drm/amd/powerplay/inc/smu_ucode_xfer_vi.h struct SMU_Entry entry[SMU_MAX_ENTRIES]; entry 1528 drivers/gpu/drm/amd/powerplay/smu_v11_0.c struct amdgpu_iv_entry *entry) entry 1530 drivers/gpu/drm/amd/powerplay/smu_v11_0.c uint32_t client_id = entry->client_id; entry 1531 drivers/gpu/drm/amd/powerplay/smu_v11_0.c uint32_t src_id = entry->src_id; entry 1746 drivers/gpu/drm/amd/powerplay/smumgr/ci_smumgr.c const struct ci_mc_reg_entry *entry, entry 1754 drivers/gpu/drm/amd/powerplay/smumgr/ci_smumgr.c data->value[i] = PP_HOST_TO_SMC_UL(entry->mc_data[j]); entry 1714 drivers/gpu/drm/amd/powerplay/smumgr/iceland_smumgr.c const struct iceland_mc_reg_entry *entry, entry 1722 drivers/gpu/drm/amd/powerplay/smumgr/iceland_smumgr.c data->value[i] = PP_HOST_TO_SMC_UL(entry->mc_data[j]); entry 810 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c table->SclkFcwRangeTable[i].vco_setting = range_table_from_vbios.entry[i].ucVco_setting; entry 811 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c table->SclkFcwRangeTable[i].postdiv = range_table_from_vbios.entry[i].ucPostdiv; entry 812 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c table->SclkFcwRangeTable[i].fcw_pcc = range_table_from_vbios.entry[i].usFcw_pcc; entry 814 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c table->SclkFcwRangeTable[i].fcw_trans_upper = range_table_from_vbios.entry[i].usFcw_trans_upper; entry 815 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c table->SclkFcwRangeTable[i].fcw_trans_lower = range_table_from_vbios.entry[i].usRcw_trans_lower; entry 125 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, entry 127 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, entry 131 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c upper_32_bits(priv->smu_tables.entry[table_id].mc_addr)); entry 134 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c lower_32_bits(priv->smu_tables.entry[table_id].mc_addr)); entry 137 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[table_id].table_id); entry 142 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c memcpy(table, (uint8_t *)priv->smu_tables.entry[table_id].table, entry 143 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[table_id].size); entry 156 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, entry 158 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, entry 161 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c memcpy(priv->smu_tables.entry[table_id].table, table, entry 162 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[table_id].size); entry 166 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c upper_32_bits(priv->smu_tables.entry[table_id].mc_addr)); entry 169 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c lower_32_bits(priv->smu_tables.entry[table_id].mc_addr)); entry 172 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[table_id].table_id); entry 200 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[SMU10_WMTABLE].handle, entry 201 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_WMTABLE].mc_addr, entry 202 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_WMTABLE].table); entry 203 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[SMU10_CLOCKTABLE].handle, entry 204 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_CLOCKTABLE].mc_addr, entry 205 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_CLOCKTABLE].table); entry 248 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_WMTABLE].handle, entry 249 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_WMTABLE].mc_addr, entry 250 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_WMTABLE].table); entry 255 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[SMU10_WMTABLE].version = 0x01; entry 256 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[SMU10_WMTABLE].size = sizeof(Watermarks_t); entry 257 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[SMU10_WMTABLE].table_id = TABLE_WATERMARKS; entry 264 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_CLOCKTABLE].handle, entry 265 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_CLOCKTABLE].mc_addr, entry 266 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_CLOCKTABLE].table); entry 271 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[SMU10_CLOCKTABLE].version = 0x01; entry 272 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[SMU10_CLOCKTABLE].size = sizeof(DpmClocks_t); entry 273 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c priv->smu_tables.entry[SMU10_CLOCKTABLE].table_id = TABLE_DPMCLOCKS; entry 278 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[SMU10_WMTABLE].handle, entry 279 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_WMTABLE].mc_addr, entry 280 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c &priv->smu_tables.entry[SMU10_WMTABLE].table); entry 42 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.h struct smu_table_entry entry[MAX_SMU_TABLE]; entry 307 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c struct SMU_Entry *entry) entry 317 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->version = info.fw_version; entry 318 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->id = (uint16_t)fw_type; entry 319 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->image_addr_high = upper_32_bits(info.mc_addr); entry 320 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->image_addr_low = lower_32_bits(info.mc_addr); entry 321 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->meta_data_addr_high = 0; entry 322 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->meta_data_addr_low = 0; entry 327 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->data_size_byte = info.image_size; entry 328 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->num_register_entries = 0; entry 333 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->flags = 1; entry 335 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c entry->flags = 0; entry 393 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_RLC_G, &toc->entry[toc->num_entries++]), entry 396 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_CP_CE, &toc->entry[toc->num_entries++]), entry 399 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_CP_PFP, &toc->entry[toc->num_entries++]), entry 402 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_CP_ME, &toc->entry[toc->num_entries++]), entry 405 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_CP_MEC, &toc->entry[toc->num_entries++]), entry 408 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_CP_MEC_JT1, &toc->entry[toc->num_entries++]), entry 411 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_CP_MEC_JT2, &toc->entry[toc->num_entries++]), entry 414 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_SDMA0, &toc->entry[toc->num_entries++]), entry 417 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_SDMA1, &toc->entry[toc->num_entries++]), entry 421 drivers/gpu/drm/amd/powerplay/smumgr/smu7_smumgr.c UCODE_ID_MEC_STORAGE, &toc->entry[toc->num_entries++]), entry 586 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c struct smu8_buffer_entry *entry) entry 591 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c entry->data_size = ulsize_byte; entry 592 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c entry->kaddr = (char *) smu8_smu->smu_buffer.kaddr + entry 594 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c entry->mc_addr = smu8_smu->smu_buffer.mc_addr + smu8_smu->smu_buffer_used_bytes; entry 595 drivers/gpu/drm/amd/powerplay/smumgr/smu8_smumgr.c entry->firmware_ID = scratch_type; entry 2092 drivers/gpu/drm/amd/powerplay/smumgr/tonga_smumgr.c const struct tonga_mc_reg_entry *entry, entry 2100 drivers/gpu/drm/amd/powerplay/smumgr/tonga_smumgr.c data->value[i] = PP_HOST_TO_SMC_UL(entry->mc_data[j]); entry 46 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, entry 48 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, entry 52 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c upper_32_bits(priv->smu_tables.entry[table_id].mc_addr)); entry 55 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c lower_32_bits(priv->smu_tables.entry[table_id].mc_addr)); entry 58 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[table_id].table_id); entry 63 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c memcpy(table, priv->smu_tables.entry[table_id].table, entry 64 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[table_id].size); entry 76 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, entry 78 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, entry 81 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c memcpy(priv->smu_tables.entry[table_id].table, table, entry 82 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[table_id].size); entry 86 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c upper_32_bits(priv->smu_tables.entry[table_id].mc_addr)); entry 89 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c lower_32_bits(priv->smu_tables.entry[table_id].mc_addr)); entry 92 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[table_id].table_id); entry 135 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c if (priv->smu_tables.entry[TOOLSTABLE].mc_addr) { entry 138 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c upper_32_bits(priv->smu_tables.entry[TOOLSTABLE].mc_addr)); entry 141 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c lower_32_bits(priv->smu_tables.entry[TOOLSTABLE].mc_addr)); entry 201 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[PPTABLE].handle, entry 202 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[PPTABLE].mc_addr, entry 203 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[PPTABLE].table); entry 207 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[PPTABLE].version = 0x01; entry 208 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[PPTABLE].size = sizeof(PPTable_t); entry 209 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[PPTABLE].table_id = TABLE_PPTABLE; entry 216 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[WMTABLE].handle, entry 217 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[WMTABLE].mc_addr, entry 218 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[WMTABLE].table); entry 223 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[WMTABLE].version = 0x01; entry 224 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[WMTABLE].size = sizeof(Watermarks_t); entry 225 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[WMTABLE].table_id = TABLE_WATERMARKS; entry 232 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSTABLE].handle, entry 233 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSTABLE].mc_addr, entry 234 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSTABLE].table); entry 239 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[AVFSTABLE].version = 0x01; entry 240 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[AVFSTABLE].size = sizeof(AvfsTable_t); entry 241 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[AVFSTABLE].table_id = TABLE_AVFS; entry 249 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[TOOLSTABLE].handle, entry 250 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[TOOLSTABLE].mc_addr, entry 251 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[TOOLSTABLE].table); entry 254 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[TOOLSTABLE].version = 0x01; entry 255 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[TOOLSTABLE].size = tools_size; entry 256 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[TOOLSTABLE].table_id = TABLE_PMSTATUSLOG; entry 264 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSFUSETABLE].handle, entry 265 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSFUSETABLE].mc_addr, entry 266 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSFUSETABLE].table); entry 270 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[AVFSFUSETABLE].version = 0x01; entry 271 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[AVFSFUSETABLE].size = sizeof(AvfsFuseOverride_t); entry 272 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c priv->smu_tables.entry[AVFSFUSETABLE].table_id = TABLE_AVFS_FUSE_OVERRIDE; entry 278 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c if (priv->smu_tables.entry[TOOLSTABLE].table) entry 279 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TOOLSTABLE].handle, entry 280 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[TOOLSTABLE].mc_addr, entry 281 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[TOOLSTABLE].table); entry 283 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[AVFSTABLE].handle, entry 284 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSTABLE].mc_addr, entry 285 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSTABLE].table); entry 287 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[WMTABLE].handle, entry 288 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[WMTABLE].mc_addr, entry 289 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[WMTABLE].table); entry 291 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[PPTABLE].handle, entry 292 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[PPTABLE].mc_addr, entry 293 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[PPTABLE].table); entry 305 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[PPTABLE].handle, entry 306 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[PPTABLE].mc_addr, entry 307 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[PPTABLE].table); entry 308 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[WMTABLE].handle, entry 309 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[WMTABLE].mc_addr, entry 310 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[WMTABLE].table); entry 311 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[AVFSTABLE].handle, entry 312 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSTABLE].mc_addr, entry 313 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSTABLE].table); entry 314 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c if (priv->smu_tables.entry[TOOLSTABLE].table) entry 315 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TOOLSTABLE].handle, entry 316 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[TOOLSTABLE].mc_addr, entry 317 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[TOOLSTABLE].table); entry 318 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[AVFSFUSETABLE].handle, entry 319 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSFUSETABLE].mc_addr, entry 320 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c &priv->smu_tables.entry[AVFSFUSETABLE].table); entry 38 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.h struct smu_table_entry entry[MAX_SMU_TABLE]; entry 49 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, entry 51 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, entry 55 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c upper_32_bits(priv->smu_tables.entry[table_id].mc_addr)) == 0, entry 59 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c lower_32_bits(priv->smu_tables.entry[table_id].mc_addr)) == 0, entry 71 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c memcpy(table, priv->smu_tables.entry[table_id].table, entry 72 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[table_id].size); entry 90 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, entry 92 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, entry 95 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c memcpy(priv->smu_tables.entry[table_id].table, table, entry 96 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[table_id].size); entry 100 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c upper_32_bits(priv->smu_tables.entry[table_id].mc_addr)) == 0, entry 105 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c lower_32_bits(priv->smu_tables.entry[table_id].mc_addr)) == 0, entry 191 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c if (priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr) { entry 194 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c upper_32_bits(priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr))) entry 197 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c lower_32_bits(priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr)); entry 225 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].handle, entry 226 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].mc_addr, entry 227 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].table); entry 231 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_PPTABLE].version = 0x01; entry 232 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_PPTABLE].size = sizeof(PPTable_t); entry 239 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].handle, entry 240 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr, entry 241 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].table); entry 246 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_WATERMARKS].version = 0x01; entry 247 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_WATERMARKS].size = sizeof(Watermarks_t); entry 255 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle, entry 256 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr, entry 257 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].table); entry 261 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_PMSTATUSLOG].version = 0x01; entry 262 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_PMSTATUSLOG].size = tools_size; entry 270 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].handle, entry 271 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].mc_addr, entry 272 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].table); entry 277 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].version = 0x01; entry 278 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].size = sizeof(AvfsFuseOverride_t); entry 285 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].handle, entry 286 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr, entry 287 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].table); entry 291 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_OVERDRIVE].version = 0x01; entry 292 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_OVERDRIVE].size = sizeof(OverDriveTable_t); entry 299 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].handle, entry 300 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr, entry 301 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].table); entry 305 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_SMU_METRICS].version = 0x01; entry 306 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c priv->smu_tables.entry[TABLE_SMU_METRICS].size = sizeof(SmuMetrics_t); entry 311 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_OVERDRIVE].handle, entry 312 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr, entry 313 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].table); entry 315 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].handle, entry 316 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].mc_addr, entry 317 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].table); entry 319 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c if (priv->smu_tables.entry[TABLE_PMSTATUSLOG].table) entry 320 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle, entry 321 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr, entry 322 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].table); entry 324 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_WATERMARKS].handle, entry 325 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr, entry 326 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].table); entry 328 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PPTABLE].handle, entry 329 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].mc_addr, entry 330 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].table); entry 343 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PPTABLE].handle, entry 344 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].mc_addr, entry 345 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].table); entry 346 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_WATERMARKS].handle, entry 347 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr, entry 348 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].table); entry 349 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c if (priv->smu_tables.entry[TABLE_PMSTATUSLOG].table) entry 350 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle, entry 351 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr, entry 352 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].table); entry 353 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].handle, entry 354 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].mc_addr, entry 355 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_AVFS_FUSE_OVERRIDE].table); entry 356 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_OVERDRIVE].handle, entry 357 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr, entry 358 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].table); entry 359 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_SMU_METRICS].handle, entry 360 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr, entry 361 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].table); entry 39 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.h struct smu_table_entry entry[TABLE_COUNT]; entry 171 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, entry 173 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, entry 178 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c upper_32_bits(priv->smu_tables.entry[table_id].mc_addr))) == 0, entry 183 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c lower_32_bits(priv->smu_tables.entry[table_id].mc_addr))) == 0, entry 194 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c memcpy(table, priv->smu_tables.entry[table_id].table, entry 195 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[table_id].size); entry 214 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, entry 216 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, entry 219 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c memcpy(priv->smu_tables.entry[table_id].table, table, entry 220 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[table_id].size); entry 224 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c upper_32_bits(priv->smu_tables.entry[table_id].mc_addr))) == 0, entry 229 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c lower_32_bits(priv->smu_tables.entry[table_id].mc_addr))) == 0, entry 247 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c memcpy(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].table, table, entry 248 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].size); entry 252 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c upper_32_bits(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr))) == 0, entry 257 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c lower_32_bits(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr))) == 0, entry 278 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c upper_32_bits(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr))) == 0, entry 283 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c lower_32_bits(priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr))) == 0, entry 295 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c memcpy(table, priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].table, entry 296 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].size); entry 365 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c if (priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr) { entry 368 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c upper_32_bits(priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr)); entry 372 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c lower_32_bits(priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr)); entry 386 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c upper_32_bits(priv->smu_tables.entry[TABLE_PPTABLE].mc_addr))) == 0, entry 391 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c lower_32_bits(priv->smu_tables.entry[TABLE_PPTABLE].mc_addr))) == 0, entry 423 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].handle, entry 424 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].mc_addr, entry 425 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].table); entry 429 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_PPTABLE].version = 0x01; entry 430 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_PPTABLE].size = sizeof(PPTable_t); entry 437 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].handle, entry 438 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr, entry 439 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].table); entry 443 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_WATERMARKS].version = 0x01; entry 444 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_WATERMARKS].size = sizeof(Watermarks_t); entry 451 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle, entry 452 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr, entry 453 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].table); entry 457 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_PMSTATUSLOG].version = 0x01; entry 458 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_PMSTATUSLOG].size = tools_size; entry 465 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].handle, entry 466 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr, entry 467 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].table); entry 471 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_OVERDRIVE].version = 0x01; entry 472 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_OVERDRIVE].size = sizeof(OverDriveTable_t); entry 479 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].handle, entry 480 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr, entry 481 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].table); entry 485 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_SMU_METRICS].version = 0x01; entry 486 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_SMU_METRICS].size = sizeof(SmuMetrics_t); entry 493 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].handle, entry 494 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr, entry 495 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].table); entry 499 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].version = 0x01; entry 500 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].size = sizeof(DpmActivityMonitorCoeffInt_t); entry 505 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_SMU_METRICS].handle, entry 506 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr, entry 507 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].table); entry 509 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_OVERDRIVE].handle, entry 510 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr, entry 511 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].table); entry 513 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle, entry 514 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr, entry 515 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].table); entry 517 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_WATERMARKS].handle, entry 518 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr, entry 519 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].table); entry 521 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PPTABLE].handle, entry 522 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].mc_addr, entry 523 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].table); entry 536 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PPTABLE].handle, entry 537 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].mc_addr, entry 538 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PPTABLE].table); entry 539 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_WATERMARKS].handle, entry 540 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].mc_addr, entry 541 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_WATERMARKS].table); entry 542 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_PMSTATUSLOG].handle, entry 543 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].mc_addr, entry 544 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_PMSTATUSLOG].table); entry 545 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_OVERDRIVE].handle, entry 546 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].mc_addr, entry 547 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_OVERDRIVE].table); entry 548 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_SMU_METRICS].handle, entry 549 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].mc_addr, entry 550 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_SMU_METRICS].table); entry 551 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c amdgpu_bo_free_kernel(&priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].handle, entry 552 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].mc_addr, entry 553 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c &priv->smu_tables.entry[TABLE_ACTIVITY_MONITOR_COEFF].table); entry 38 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.h struct smu_table_entry entry[TABLE_COUNT]; entry 679 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c range_table_from_vbios.entry[i].ucVco_setting; entry 681 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c range_table_from_vbios.entry[i].ucPostdiv; entry 683 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c range_table_from_vbios.entry[i].usFcw_pcc; entry 686 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c range_table_from_vbios.entry[i].usFcw_trans_upper; entry 688 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c range_table_from_vbios.entry[i].usRcw_trans_lower; entry 63 drivers/gpu/drm/ati_pcigart.c struct drm_sg_mem *entry = dev->sg; entry 69 drivers/gpu/drm/ati_pcigart.c if (!entry) { entry 77 drivers/gpu/drm/ati_pcigart.c pages = (entry->pages <= max_pages) entry 78 drivers/gpu/drm/ati_pcigart.c ? entry->pages : max_pages; entry 81 drivers/gpu/drm/ati_pcigart.c if (!entry->busaddr[i]) entry 83 drivers/gpu/drm/ati_pcigart.c pci_unmap_page(dev->pdev, entry->busaddr[i], entry 103 drivers/gpu/drm/ati_pcigart.c struct drm_sg_mem *entry = dev->sg; entry 111 drivers/gpu/drm/ati_pcigart.c if (!entry) { entry 146 drivers/gpu/drm/ati_pcigart.c pages = (entry->pages <= max_real_pages) entry 147 drivers/gpu/drm/ati_pcigart.c ? entry->pages : max_real_pages; entry 158 drivers/gpu/drm/ati_pcigart.c entry->busaddr[i] = pci_map_page(dev->pdev, entry->pagelist[i], entry 160 drivers/gpu/drm/ati_pcigart.c if (pci_dma_mapping_error(dev->pdev, entry->busaddr[i])) { entry 168 drivers/gpu/drm/ati_pcigart.c page_base = (u32) entry->busaddr[i]; entry 204 drivers/gpu/drm/drm_agpsupport.c struct drm_agp_mem *entry; entry 211 drivers/gpu/drm/drm_agpsupport.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 212 drivers/gpu/drm/drm_agpsupport.c if (!entry) entry 219 drivers/gpu/drm/drm_agpsupport.c kfree(entry); entry 223 drivers/gpu/drm/drm_agpsupport.c entry->handle = (unsigned long)memory->key + 1; entry 224 drivers/gpu/drm/drm_agpsupport.c entry->memory = memory; entry 225 drivers/gpu/drm/drm_agpsupport.c entry->bound = 0; entry 226 drivers/gpu/drm/drm_agpsupport.c entry->pages = pages; entry 227 drivers/gpu/drm/drm_agpsupport.c list_add(&entry->head, &dev->agp->memory); entry 229 drivers/gpu/drm/drm_agpsupport.c request->handle = entry->handle; entry 257 drivers/gpu/drm/drm_agpsupport.c struct drm_agp_mem *entry; entry 259 drivers/gpu/drm/drm_agpsupport.c list_for_each_entry(entry, &dev->agp->memory, head) { entry 260 drivers/gpu/drm/drm_agpsupport.c if (entry->handle == handle) entry 261 drivers/gpu/drm/drm_agpsupport.c return entry; entry 280 drivers/gpu/drm/drm_agpsupport.c struct drm_agp_mem *entry; entry 285 drivers/gpu/drm/drm_agpsupport.c entry = drm_agp_lookup_entry(dev, request->handle); entry 286 drivers/gpu/drm/drm_agpsupport.c if (!entry || !entry->bound) entry 288 drivers/gpu/drm/drm_agpsupport.c ret = drm_unbind_agp(entry->memory); entry 290 drivers/gpu/drm/drm_agpsupport.c entry->bound = 0; entry 319 drivers/gpu/drm/drm_agpsupport.c struct drm_agp_mem *entry; entry 325 drivers/gpu/drm/drm_agpsupport.c entry = drm_agp_lookup_entry(dev, request->handle); entry 326 drivers/gpu/drm/drm_agpsupport.c if (!entry || entry->bound) entry 329 drivers/gpu/drm/drm_agpsupport.c retcode = drm_bind_agp(entry->memory, page); entry 332 drivers/gpu/drm/drm_agpsupport.c entry->bound = dev->agp->base + (page << PAGE_SHIFT); entry 334 drivers/gpu/drm/drm_agpsupport.c dev->agp->base, entry->bound); entry 364 drivers/gpu/drm/drm_agpsupport.c struct drm_agp_mem *entry; entry 368 drivers/gpu/drm/drm_agpsupport.c entry = drm_agp_lookup_entry(dev, request->handle); entry 369 drivers/gpu/drm/drm_agpsupport.c if (!entry) entry 371 drivers/gpu/drm/drm_agpsupport.c if (entry->bound) entry 372 drivers/gpu/drm/drm_agpsupport.c drm_unbind_agp(entry->memory); entry 374 drivers/gpu/drm/drm_agpsupport.c list_del(&entry->head); entry 376 drivers/gpu/drm/drm_agpsupport.c drm_free_agp(entry->memory, entry->pages); entry 377 drivers/gpu/drm/drm_agpsupport.c kfree(entry); entry 447 drivers/gpu/drm/drm_agpsupport.c struct drm_agp_mem *entry, *tempe; entry 454 drivers/gpu/drm/drm_agpsupport.c list_for_each_entry_safe(entry, tempe, &dev->agp->memory, head) { entry 455 drivers/gpu/drm/drm_agpsupport.c if (entry->bound) entry 456 drivers/gpu/drm/drm_agpsupport.c drm_unbind_agp(entry->memory); entry 457 drivers/gpu/drm/drm_agpsupport.c drm_free_agp(entry->memory, entry->pages); entry 458 drivers/gpu/drm/drm_agpsupport.c kfree(entry); entry 55 drivers/gpu/drm/drm_bufs.c struct drm_map_list *entry; entry 56 drivers/gpu/drm/drm_bufs.c list_for_each_entry(entry, &dev->maplist, head) { entry 65 drivers/gpu/drm/drm_bufs.c if (!entry->map || entry 66 drivers/gpu/drm/drm_bufs.c map->type != entry->map->type || entry 67 drivers/gpu/drm/drm_bufs.c entry->master != dev->master) entry 73 drivers/gpu/drm/drm_bufs.c return entry; entry 76 drivers/gpu/drm/drm_bufs.c if ((entry->map->offset & 0xffffffff) == entry 78 drivers/gpu/drm/drm_bufs.c return entry; entry 82 drivers/gpu/drm/drm_bufs.c if (entry->map->offset == map->offset) entry 83 drivers/gpu/drm/drm_bufs.c return entry; entry 270 drivers/gpu/drm/drm_bufs.c struct drm_agp_mem *entry; entry 299 drivers/gpu/drm/drm_bufs.c list_for_each_entry(entry, &dev->agp->memory, head) { entry 300 drivers/gpu/drm/drm_bufs.c if ((map->offset >= entry->bound) && entry 301 drivers/gpu/drm/drm_bufs.c (map->offset + map->size <= entry->bound + entry->pages * PAGE_SIZE)) { entry 673 drivers/gpu/drm/drm_bufs.c struct drm_buf_entry *entry) entry 677 drivers/gpu/drm/drm_bufs.c if (entry->seg_count) { entry 678 drivers/gpu/drm/drm_bufs.c for (i = 0; i < entry->seg_count; i++) { entry 679 drivers/gpu/drm/drm_bufs.c if (entry->seglist[i]) { entry 680 drivers/gpu/drm/drm_bufs.c drm_pci_free(dev, entry->seglist[i]); entry 683 drivers/gpu/drm/drm_bufs.c kfree(entry->seglist); entry 685 drivers/gpu/drm/drm_bufs.c entry->seg_count = 0; entry 688 drivers/gpu/drm/drm_bufs.c if (entry->buf_count) { entry 689 drivers/gpu/drm/drm_bufs.c for (i = 0; i < entry->buf_count; i++) { entry 690 drivers/gpu/drm/drm_bufs.c kfree(entry->buflist[i].dev_private); entry 692 drivers/gpu/drm/drm_bufs.c kfree(entry->buflist); entry 694 drivers/gpu/drm/drm_bufs.c entry->buf_count = 0; entry 714 drivers/gpu/drm/drm_bufs.c struct drm_buf_entry *entry; entry 777 drivers/gpu/drm/drm_bufs.c entry = &dma->bufs[order]; entry 778 drivers/gpu/drm/drm_bufs.c if (entry->buf_count) { entry 790 drivers/gpu/drm/drm_bufs.c entry->buflist = kcalloc(count, sizeof(*entry->buflist), GFP_KERNEL); entry 791 drivers/gpu/drm/drm_bufs.c if (!entry->buflist) { entry 797 drivers/gpu/drm/drm_bufs.c entry->buf_size = size; entry 798 drivers/gpu/drm/drm_bufs.c entry->page_order = page_order; entry 802 drivers/gpu/drm/drm_bufs.c while (entry->buf_count < count) { entry 803 drivers/gpu/drm/drm_bufs.c buf = &entry->buflist[entry->buf_count]; entry 804 drivers/gpu/drm/drm_bufs.c buf->idx = dma->buf_count + entry->buf_count; entry 821 drivers/gpu/drm/drm_bufs.c entry->buf_count = count; entry 822 drivers/gpu/drm/drm_bufs.c drm_cleanup_buf_error(dev, entry); entry 828 drivers/gpu/drm/drm_bufs.c DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address); entry 831 drivers/gpu/drm/drm_bufs.c entry->buf_count++; entry 838 drivers/gpu/drm/drm_bufs.c (dma->buf_count + entry->buf_count) * entry 842 drivers/gpu/drm/drm_bufs.c drm_cleanup_buf_error(dev, entry); entry 849 drivers/gpu/drm/drm_bufs.c for (i = 0; i < entry->buf_count; i++) { entry 850 drivers/gpu/drm/drm_bufs.c dma->buflist[i + dma->buf_count] = &entry->buflist[i]; entry 853 drivers/gpu/drm/drm_bufs.c dma->buf_count += entry->buf_count; entry 854 drivers/gpu/drm/drm_bufs.c dma->seg_count += entry->seg_count; entry 859 drivers/gpu/drm/drm_bufs.c DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count); entry 863 drivers/gpu/drm/drm_bufs.c request->count = entry->buf_count; entry 883 drivers/gpu/drm/drm_bufs.c struct drm_buf_entry *entry; entry 927 drivers/gpu/drm/drm_bufs.c entry = &dma->bufs[order]; entry 928 drivers/gpu/drm/drm_bufs.c if (entry->buf_count) { entry 940 drivers/gpu/drm/drm_bufs.c entry->buflist = kcalloc(count, sizeof(*entry->buflist), GFP_KERNEL); entry 941 drivers/gpu/drm/drm_bufs.c if (!entry->buflist) { entry 947 drivers/gpu/drm/drm_bufs.c entry->seglist = kcalloc(count, sizeof(*entry->seglist), GFP_KERNEL); entry 948 drivers/gpu/drm/drm_bufs.c if (!entry->seglist) { entry 949 drivers/gpu/drm/drm_bufs.c kfree(entry->buflist); entry 962 drivers/gpu/drm/drm_bufs.c kfree(entry->buflist); entry 963 drivers/gpu/drm/drm_bufs.c kfree(entry->seglist); entry 973 drivers/gpu/drm/drm_bufs.c entry->buf_size = size; entry 974 drivers/gpu/drm/drm_bufs.c entry->page_order = page_order; entry 978 drivers/gpu/drm/drm_bufs.c while (entry->buf_count < count) { entry 984 drivers/gpu/drm/drm_bufs.c entry->buf_count = count; entry 985 drivers/gpu/drm/drm_bufs.c entry->seg_count = count; entry 986 drivers/gpu/drm/drm_bufs.c drm_cleanup_buf_error(dev, entry); entry 992 drivers/gpu/drm/drm_bufs.c entry->seglist[entry->seg_count++] = dmah; entry 1001 drivers/gpu/drm/drm_bufs.c offset + size <= total && entry->buf_count < count; entry 1002 drivers/gpu/drm/drm_bufs.c offset += alignment, ++entry->buf_count) { entry 1003 drivers/gpu/drm/drm_bufs.c buf = &entry->buflist[entry->buf_count]; entry 1004 drivers/gpu/drm/drm_bufs.c buf->idx = dma->buf_count + entry->buf_count; entry 1021 drivers/gpu/drm/drm_bufs.c entry->buf_count = count; entry 1022 drivers/gpu/drm/drm_bufs.c entry->seg_count = count; entry 1023 drivers/gpu/drm/drm_bufs.c drm_cleanup_buf_error(dev, entry); entry 1031 drivers/gpu/drm/drm_bufs.c entry->buf_count, buf->address); entry 1037 drivers/gpu/drm/drm_bufs.c (dma->buf_count + entry->buf_count) * entry 1041 drivers/gpu/drm/drm_bufs.c drm_cleanup_buf_error(dev, entry); entry 1049 drivers/gpu/drm/drm_bufs.c for (i = 0; i < entry->buf_count; i++) { entry 1050 drivers/gpu/drm/drm_bufs.c dma->buflist[i + dma->buf_count] = &entry->buflist[i]; entry 1061 drivers/gpu/drm/drm_bufs.c dma->buf_count += entry->buf_count; entry 1062 drivers/gpu/drm/drm_bufs.c dma->seg_count += entry->seg_count; entry 1063 drivers/gpu/drm/drm_bufs.c dma->page_count += entry->seg_count << page_order; entry 1064 drivers/gpu/drm/drm_bufs.c dma->byte_count += PAGE_SIZE * (entry->seg_count << page_order); entry 1068 drivers/gpu/drm/drm_bufs.c request->count = entry->buf_count; entry 1084 drivers/gpu/drm/drm_bufs.c struct drm_buf_entry *entry; entry 1139 drivers/gpu/drm/drm_bufs.c entry = &dma->bufs[order]; entry 1140 drivers/gpu/drm/drm_bufs.c if (entry->buf_count) { entry 1152 drivers/gpu/drm/drm_bufs.c entry->buflist = kcalloc(count, sizeof(*entry->buflist), GFP_KERNEL); entry 1153 drivers/gpu/drm/drm_bufs.c if (!entry->buflist) { entry 1159 drivers/gpu/drm/drm_bufs.c entry->buf_size = size; entry 1160 drivers/gpu/drm/drm_bufs.c entry->page_order = page_order; entry 1164 drivers/gpu/drm/drm_bufs.c while (entry->buf_count < count) { entry 1165 drivers/gpu/drm/drm_bufs.c buf = &entry->buflist[entry->buf_count]; entry 1166 drivers/gpu/drm/drm_bufs.c buf->idx = dma->buf_count + entry->buf_count; entry 1184 drivers/gpu/drm/drm_bufs.c entry->buf_count = count; entry 1185 drivers/gpu/drm/drm_bufs.c drm_cleanup_buf_error(dev, entry); entry 1191 drivers/gpu/drm/drm_bufs.c DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address); entry 1194 drivers/gpu/drm/drm_bufs.c entry->buf_count++; entry 1201 drivers/gpu/drm/drm_bufs.c (dma->buf_count + entry->buf_count) * entry 1205 drivers/gpu/drm/drm_bufs.c drm_cleanup_buf_error(dev, entry); entry 1212 drivers/gpu/drm/drm_bufs.c for (i = 0; i < entry->buf_count; i++) { entry 1213 drivers/gpu/drm/drm_bufs.c dma->buflist[i + dma->buf_count] = &entry->buflist[i]; entry 1216 drivers/gpu/drm/drm_bufs.c dma->buf_count += entry->buf_count; entry 1217 drivers/gpu/drm/drm_bufs.c dma->seg_count += entry->seg_count; entry 1222 drivers/gpu/drm/drm_bufs.c DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count); entry 1226 drivers/gpu/drm/drm_bufs.c request->count = entry->buf_count; entry 1387 drivers/gpu/drm/drm_bufs.c struct drm_buf_entry *entry; entry 1403 drivers/gpu/drm/drm_bufs.c entry = &dma->bufs[order]; entry 1405 drivers/gpu/drm/drm_bufs.c if (request->low_mark < 0 || request->low_mark > entry->buf_count) entry 1407 drivers/gpu/drm/drm_bufs.c if (request->high_mark < 0 || request->high_mark > entry->buf_count) entry 1410 drivers/gpu/drm/drm_bufs.c entry->low_mark = request->low_mark; entry 1411 drivers/gpu/drm/drm_bufs.c entry->high_mark = request->high_mark; entry 1594 drivers/gpu/drm/drm_bufs.c struct drm_map_list *entry; entry 1596 drivers/gpu/drm/drm_bufs.c list_for_each_entry(entry, &dev->maplist, head) { entry 1597 drivers/gpu/drm/drm_bufs.c if (entry->map && entry->map->type == _DRM_SHM && entry 1598 drivers/gpu/drm/drm_bufs.c (entry->map->flags & _DRM_CONTAINS_LOCK)) { entry 1599 drivers/gpu/drm/drm_bufs.c return entry->map; entry 483 drivers/gpu/drm/drm_color_mgmt.c const struct drm_color_lut *entry; entry 489 drivers/gpu/drm/drm_color_mgmt.c entry = lut->data; entry 492 drivers/gpu/drm/drm_color_mgmt.c if (entry[i].red != entry[i].blue || entry 493 drivers/gpu/drm/drm_color_mgmt.c entry[i].red != entry[i].green) { entry 500 drivers/gpu/drm/drm_color_mgmt.c if (entry[i].red < entry[i - 1].red || entry 501 drivers/gpu/drm/drm_color_mgmt.c entry[i].green < entry[i - 1].green || entry 502 drivers/gpu/drm/drm_color_mgmt.c entry[i].blue < entry[i - 1].blue) { entry 282 drivers/gpu/drm/drm_debugfs_crc.c struct drm_crtc_crc_entry *entry; entry 310 drivers/gpu/drm/drm_debugfs_crc.c entry = &crc->entries[crc->tail]; entry 322 drivers/gpu/drm/drm_debugfs_crc.c if (entry->has_frame_counter) entry 323 drivers/gpu/drm/drm_debugfs_crc.c sprintf(buf, "0x%08x", entry->frame); entry 328 drivers/gpu/drm/drm_debugfs_crc.c sprintf(buf + 10 + i * 11, " 0x%08x", entry->crcs[i]); entry 392 drivers/gpu/drm/drm_debugfs_crc.c struct drm_crtc_crc_entry *entry; entry 419 drivers/gpu/drm/drm_debugfs_crc.c entry = &crc->entries[head]; entry 420 drivers/gpu/drm/drm_debugfs_crc.c entry->frame = frame; entry 421 drivers/gpu/drm/drm_debugfs_crc.c entry->has_frame_counter = has_frame; entry 422 drivers/gpu/drm/drm_debugfs_crc.c memcpy(&entry->crcs, crcs, sizeof(*crcs) * crc->values_cnt); entry 1359 drivers/gpu/drm/drm_gem.c struct dma_fence *entry; entry 1371 drivers/gpu/drm/drm_gem.c xa_for_each(fence_array, index, entry) { entry 1372 drivers/gpu/drm/drm_gem.c if (entry->context != fence->context) entry 1375 drivers/gpu/drm/drm_gem.c if (dma_fence_is_later(fence, entry)) { entry 1376 drivers/gpu/drm/drm_gem.c dma_fence_put(entry); entry 65 drivers/gpu/drm/drm_hashtab.c struct drm_hash_item *entry; entry 73 drivers/gpu/drm/drm_hashtab.c hlist_for_each_entry(entry, h_list, head) entry 74 drivers/gpu/drm/drm_hashtab.c DRM_DEBUG("count %d, key: 0x%08lx\n", count++, entry->key); entry 80 drivers/gpu/drm/drm_hashtab.c struct drm_hash_item *entry; entry 86 drivers/gpu/drm/drm_hashtab.c hlist_for_each_entry(entry, h_list, head) { entry 87 drivers/gpu/drm/drm_hashtab.c if (entry->key == key) entry 88 drivers/gpu/drm/drm_hashtab.c return &entry->head; entry 89 drivers/gpu/drm/drm_hashtab.c if (entry->key > key) entry 98 drivers/gpu/drm/drm_hashtab.c struct drm_hash_item *entry; entry 104 drivers/gpu/drm/drm_hashtab.c hlist_for_each_entry_rcu(entry, h_list, head) { entry 105 drivers/gpu/drm/drm_hashtab.c if (entry->key == key) entry 106 drivers/gpu/drm/drm_hashtab.c return &entry->head; entry 107 drivers/gpu/drm/drm_hashtab.c if (entry->key > key) entry 115 drivers/gpu/drm/drm_hashtab.c struct drm_hash_item *entry; entry 124 drivers/gpu/drm/drm_hashtab.c hlist_for_each_entry(entry, h_list, head) { entry 125 drivers/gpu/drm/drm_hashtab.c if (entry->key == key) entry 127 drivers/gpu/drm/drm_hashtab.c if (entry->key > key) entry 129 drivers/gpu/drm/drm_hashtab.c parent = &entry->head; entry 203 drivers/gpu/drm/drm_lease.c void *entry; entry 215 drivers/gpu/drm/drm_lease.c idr_for_each_entry(leases, entry, object) { entry 303 drivers/gpu/drm/drm_lease.c void *entry; entry 316 drivers/gpu/drm/drm_lease.c idr_for_each_entry(&master->leases, entry, object) entry 664 drivers/gpu/drm/drm_lease.c void *entry; entry 687 drivers/gpu/drm/drm_lease.c idr_for_each_entry(object_idr, entry, object) { entry 167 drivers/gpu/drm/drm_lock.c DECLARE_WAITQUEUE(entry, current); entry 188 drivers/gpu/drm/drm_lock.c add_wait_queue(&master->lock.lock_queue, &entry); entry 220 drivers/gpu/drm/drm_lock.c remove_wait_queue(&master->lock.lock_queue, &entry); entry 945 drivers/gpu/drm/drm_mm.c static u64 drm_mm_dump_hole(struct drm_printer *p, const struct drm_mm_node *entry) entry 949 drivers/gpu/drm/drm_mm.c size = entry->hole_size; entry 951 drivers/gpu/drm/drm_mm.c start = drm_mm_hole_node_start(entry); entry 965 drivers/gpu/drm/drm_mm.c const struct drm_mm_node *entry; entry 970 drivers/gpu/drm/drm_mm.c drm_mm_for_each_node(entry, mm) { entry 971 drivers/gpu/drm/drm_mm.c drm_printf(p, "%#018llx-%#018llx: %llu: used\n", entry->start, entry 972 drivers/gpu/drm/drm_mm.c entry->start + entry->size, entry->size); entry 973 drivers/gpu/drm/drm_mm.c total_used += entry->size; entry 974 drivers/gpu/drm/drm_mm.c total_free += drm_mm_dump_hole(p, entry); entry 55 drivers/gpu/drm/drm_scatter.c static void drm_sg_cleanup(struct drm_sg_mem * entry) entry 60 drivers/gpu/drm/drm_scatter.c for (i = 0; i < entry->pages; i++) { entry 61 drivers/gpu/drm/drm_scatter.c page = entry->pagelist[i]; entry 66 drivers/gpu/drm/drm_scatter.c vfree(entry->virtual); entry 68 drivers/gpu/drm/drm_scatter.c kfree(entry->busaddr); entry 69 drivers/gpu/drm/drm_scatter.c kfree(entry->pagelist); entry 70 drivers/gpu/drm/drm_scatter.c kfree(entry); entry 91 drivers/gpu/drm/drm_scatter.c struct drm_sg_mem *entry; entry 105 drivers/gpu/drm/drm_scatter.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 106 drivers/gpu/drm/drm_scatter.c if (!entry) entry 112 drivers/gpu/drm/drm_scatter.c entry->pages = pages; entry 113 drivers/gpu/drm/drm_scatter.c entry->pagelist = kcalloc(pages, sizeof(*entry->pagelist), GFP_KERNEL); entry 114 drivers/gpu/drm/drm_scatter.c if (!entry->pagelist) { entry 115 drivers/gpu/drm/drm_scatter.c kfree(entry); entry 119 drivers/gpu/drm/drm_scatter.c entry->busaddr = kcalloc(pages, sizeof(*entry->busaddr), GFP_KERNEL); entry 120 drivers/gpu/drm/drm_scatter.c if (!entry->busaddr) { entry 121 drivers/gpu/drm/drm_scatter.c kfree(entry->pagelist); entry 122 drivers/gpu/drm/drm_scatter.c kfree(entry); entry 126 drivers/gpu/drm/drm_scatter.c entry->virtual = drm_vmalloc_dma(pages << PAGE_SHIFT); entry 127 drivers/gpu/drm/drm_scatter.c if (!entry->virtual) { entry 128 drivers/gpu/drm/drm_scatter.c kfree(entry->busaddr); entry 129 drivers/gpu/drm/drm_scatter.c kfree(entry->pagelist); entry 130 drivers/gpu/drm/drm_scatter.c kfree(entry); entry 137 drivers/gpu/drm/drm_scatter.c memset(entry->virtual, 0, pages << PAGE_SHIFT); entry 139 drivers/gpu/drm/drm_scatter.c entry->handle = ScatterHandle((unsigned long)entry->virtual); entry 141 drivers/gpu/drm/drm_scatter.c DRM_DEBUG("handle = %08lx\n", entry->handle); entry 142 drivers/gpu/drm/drm_scatter.c DRM_DEBUG("virtual = %p\n", entry->virtual); entry 144 drivers/gpu/drm/drm_scatter.c for (i = (unsigned long)entry->virtual, j = 0; j < pages; entry 146 drivers/gpu/drm/drm_scatter.c entry->pagelist[j] = vmalloc_to_page((void *)i); entry 147 drivers/gpu/drm/drm_scatter.c if (!entry->pagelist[j]) entry 149 drivers/gpu/drm/drm_scatter.c SetPageReserved(entry->pagelist[j]); entry 152 drivers/gpu/drm/drm_scatter.c request->handle = entry->handle; entry 154 drivers/gpu/drm/drm_scatter.c dev->sg = entry; entry 166 drivers/gpu/drm/drm_scatter.c tmp = page_address(entry->pagelist[i]); entry 172 drivers/gpu/drm/drm_scatter.c tmp = (unsigned long *)((u8 *) entry->virtual + entry 184 drivers/gpu/drm/drm_scatter.c tmp = page_address(entry->pagelist[i]); entry 199 drivers/gpu/drm/drm_scatter.c drm_sg_cleanup(entry); entry 207 drivers/gpu/drm/drm_scatter.c struct drm_sg_mem *entry; entry 215 drivers/gpu/drm/drm_scatter.c entry = dev->sg; entry 218 drivers/gpu/drm/drm_scatter.c if (!entry || entry->handle != request->handle) entry 221 drivers/gpu/drm/drm_scatter.c DRM_DEBUG("virtual = %p\n", entry->virtual); entry 223 drivers/gpu/drm/drm_scatter.c drm_sg_cleanup(entry); entry 348 drivers/gpu/drm/drm_vm.c struct drm_sg_mem *entry = dev->sg; entry 354 drivers/gpu/drm/drm_vm.c if (!entry) entry 356 drivers/gpu/drm/drm_vm.c if (!entry->pagelist) entry 362 drivers/gpu/drm/drm_vm.c page = entry->pagelist[page_offset]; entry 267 drivers/gpu/drm/drm_vma_manager.c struct drm_vma_offset_file *new, *entry; entry 274 drivers/gpu/drm/drm_vma_manager.c new = kmalloc(sizeof(*entry), GFP_KERNEL); entry 282 drivers/gpu/drm/drm_vma_manager.c entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb); entry 284 drivers/gpu/drm/drm_vma_manager.c if (tag == entry->vm_tag) { entry 285 drivers/gpu/drm/drm_vma_manager.c entry->vm_count++; entry 287 drivers/gpu/drm/drm_vma_manager.c } else if (tag > entry->vm_tag) { entry 328 drivers/gpu/drm/drm_vma_manager.c struct drm_vma_offset_file *entry; entry 335 drivers/gpu/drm/drm_vma_manager.c entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); entry 336 drivers/gpu/drm/drm_vma_manager.c if (tag == entry->vm_tag) { entry 337 drivers/gpu/drm/drm_vma_manager.c if (!--entry->vm_count) { entry 338 drivers/gpu/drm/drm_vma_manager.c rb_erase(&entry->vm_rb, &node->vm_files); entry 339 drivers/gpu/drm/drm_vma_manager.c kfree(entry); entry 342 drivers/gpu/drm/drm_vma_manager.c } else if (tag > entry->vm_tag) { entry 369 drivers/gpu/drm/drm_vma_manager.c struct drm_vma_offset_file *entry; entry 376 drivers/gpu/drm/drm_vma_manager.c entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); entry 377 drivers/gpu/drm/drm_vma_manager.c if (tag == entry->vm_tag) entry 379 drivers/gpu/drm/drm_vma_manager.c else if (tag > entry->vm_tag) entry 98 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c u32 entry = lower_32_bits(paddr) | MMUv2_PTE_PRESENT; entry 104 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c entry |= (upper_32_bits(paddr) & 0xff) << 4; entry 107 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c entry |= MMUv2_PTE_WRITEABLE; entry 116 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c v2_context->stlb_cpu[mtlb_entry][stlb_entry] = entry; entry 222 drivers/gpu/drm/gma500/intel_bios.c struct bdb_lvds_lfp_data_entry *entry; entry 245 drivers/gpu/drm/gma500/intel_bios.c entry = &lvds_lfp_data->data[lvds_options->panel_type]; entry 246 drivers/gpu/drm/gma500/intel_bios.c dvo_timing = &entry->dvo_timing; entry 549 drivers/gpu/drm/gma500/psb_intel_display.c int entry = 0; entry 555 drivers/gpu/drm/gma500/psb_intel_display.c index_mask |= (1 << entry); entry 556 drivers/gpu/drm/gma500/psb_intel_display.c entry++; entry 178 drivers/gpu/drm/i915/display/intel_bios.c char *entry = (char *)lvds_lfp_data->data + lfp_data_size * index; entry 180 drivers/gpu/drm/i915/display/intel_bios.c return (struct lvds_dvo_timing *)(entry + dvo_timing_offset); entry 307 drivers/gpu/drm/i915/display/intel_bios.c const struct lfp_backlight_data_entry *entry; entry 320 drivers/gpu/drm/i915/display/intel_bios.c entry = &backlight_data->data[panel_type]; entry 322 drivers/gpu/drm/i915/display/intel_bios.c dev_priv->vbt.backlight.present = entry->type == BDB_BACKLIGHT_TYPE_PWM; entry 325 drivers/gpu/drm/i915/display/intel_bios.c entry->type); entry 339 drivers/gpu/drm/i915/display/intel_bios.c dev_priv->vbt.backlight.pwm_freq_hz = entry->pwm_freq_hz; entry 340 drivers/gpu/drm/i915/display/intel_bios.c dev_priv->vbt.backlight.active_low_pwm = entry->active_low_pwm; entry 341 drivers/gpu/drm/i915/display/intel_bios.c dev_priv->vbt.backlight.min_brightness = entry->min_brightness; entry 568 drivers/gpu/drm/i915/display/intel_color.c const struct drm_color_lut *entry = entry 572 drivers/gpu/drm/i915/display/intel_color.c I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_10(entry)); entry 598 drivers/gpu/drm/i915/display/intel_color.c const struct drm_color_lut *entry = entry 601 drivers/gpu/drm/i915/display/intel_color.c I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_10(entry)); entry 819 drivers/gpu/drm/i915/display/intel_color.c const struct drm_color_lut *entry = &lut[i]; entry 822 drivers/gpu/drm/i915/display/intel_color.c ilk_lut_12p4_ldw(entry)); entry 824 drivers/gpu/drm/i915/display/intel_color.c ilk_lut_12p4_udw(entry)); entry 835 drivers/gpu/drm/i915/display/intel_color.c const struct drm_color_lut *entry; entry 852 drivers/gpu/drm/i915/display/intel_color.c entry = &lut[i * 8]; entry 853 drivers/gpu/drm/i915/display/intel_color.c I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_12p4_ldw(entry)); entry 854 drivers/gpu/drm/i915/display/intel_color.c I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_12p4_udw(entry)); entry 870 drivers/gpu/drm/i915/display/intel_color.c entry = &lut[i * 8 * 128]; entry 871 drivers/gpu/drm/i915/display/intel_color.c I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_12p4_ldw(entry)); entry 872 drivers/gpu/drm/i915/display/intel_color.c I915_WRITE(PREC_PAL_DATA(pipe), ilk_lut_12p4_udw(entry)); entry 876 drivers/gpu/drm/i915/display/intel_color.c entry = &lut[256 * 8 * 128]; entry 877 drivers/gpu/drm/i915/display/intel_color.c icl_load_gcmax(crtc_state, entry); entry 14268 drivers/gpu/drm/i915/display/intel_display.c list_del(&wait->wait.entry); entry 15231 drivers/gpu/drm/i915/display/intel_display.c int entry = 0; entry 15235 drivers/gpu/drm/i915/display/intel_display.c index_mask |= (1 << entry); entry 15237 drivers/gpu/drm/i915/display/intel_display.c entry++; entry 351 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c eb_vma_misplaced(const struct drm_i915_gem_exec_object2 *entry, entry 355 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (vma->node.size < entry->pad_to_size) entry 358 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->alignment && !IS_ALIGNED(vma->node.start, entry->alignment)) entry 362 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c vma->node.start != entry->offset) entry 382 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c const struct drm_i915_gem_exec_object2 *entry, entry 391 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c pin_flags = entry->offset & PIN_OFFSET_MASK; entry 411 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c return !eb_vma_misplaced(entry, vma, exec_flags); entry 436 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_i915_gem_exec_object2 *entry, entry 439 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (unlikely(entry->flags & eb->invalid_flags)) entry 442 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (unlikely(entry->alignment && entry 443 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c !is_power_of_2_u64(entry->alignment))) entry 450 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (unlikely(entry->flags & EXEC_OBJECT_PINNED && entry 451 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->offset != gen8_canonical_addr(entry->offset & I915_GTT_PAGE_MASK))) entry 455 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->flags & EXEC_OBJECT_PAD_TO_SIZE) { entry 456 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (unlikely(offset_in_page(entry->pad_to_size))) entry 459 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->pad_to_size = 0; entry 464 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->handle, (int)(entry - eb->exec)); entry 473 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->offset = gen8_noncanonical_addr(entry->offset); entry 476 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->flags &= ~EXEC_OBJECT_NEEDS_FENCE; entry 478 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if ((entry->flags & EXEC_OBJECT_NEEDS_FENCE || entry 481 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->flags |= EXEC_OBJECT_NEEDS_GTT | __EXEC_OBJECT_NEEDS_MAP; entry 484 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (!(entry->flags & EXEC_OBJECT_PINNED)) entry 485 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->flags |= eb->context_flags; entry 495 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_i915_gem_exec_object2 *entry = &eb->exec[i]; entry 501 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c err = eb_validate_vma(eb, entry, vma); entry 507 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c vma->exec_handle = entry->handle; entry 509 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c &eb->buckets[hash_32(entry->handle, entry 513 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->relocation_count) entry 523 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c eb->flags[i] = entry->flags; entry 536 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->relocation_count && entry 546 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (eb_pin_vma(eb, entry, vma)) { entry 547 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->offset != vma->node.start) { entry 548 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->offset = vma->node.start | UPDATE; entry 583 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c struct drm_i915_gem_exec_object2 *entry = exec_entry(eb, vma); entry 603 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c pin_flags |= entry->offset | PIN_OFFSET_FIXED; entry 610 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->pad_to_size, entry->alignment, entry 615 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (entry->offset != vma->node.start) { entry 616 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c entry->offset = vma->node.start | UPDATE; entry 632 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c GEM_BUG_ON(eb_vma_misplaced(entry, vma, exec_flags)); entry 1448 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c const struct drm_i915_gem_exec_object2 *entry = exec_entry(eb, vma); entry 1451 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c urelocs = u64_to_user_ptr(entry->relocs_ptr); entry 1452 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c remain = entry->relocation_count; entry 1533 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c const struct drm_i915_gem_exec_object2 *entry = exec_entry(eb, vma); entry 1535 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c u64_to_ptr(typeof(*relocs), entry->relocs_ptr); entry 1539 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c for (i = 0; i < entry->relocation_count; i++) { entry 1553 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c static int check_relocations(const struct drm_i915_gem_exec_object2 *entry) entry 1559 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c size = entry->relocation_count; entry 1566 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c addr = u64_to_user_ptr(entry->relocs_ptr); entry 1771 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c const struct drm_i915_gem_exec_object2 *entry = entry 1775 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (!entry->relocation_count) entry 1778 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c relocs = u64_to_ptr(typeof(*relocs), entry->relocs_ptr); entry 410 drivers/gpu/drm/i915/gem/i915_gem_pages.c void *entry; entry 426 drivers/gpu/drm/i915/gem/i915_gem_pages.c entry = xa_mk_value(idx); entry 428 drivers/gpu/drm/i915/gem/i915_gem_pages.c ret = radix_tree_insert(&iter->radix, idx + i, entry); entry 527 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, bool ips) entry 529 drivers/gpu/drm/i915/gvt/gtt.c switch (entry->type) { entry 532 drivers/gpu/drm/i915/gvt/gtt.c if (pte_ops->test_pse(entry)) entry 533 drivers/gpu/drm/i915/gvt/gtt.c entry->type = get_pse_type(entry->type); entry 537 drivers/gpu/drm/i915/gvt/gtt.c entry->type = get_pse_type(entry->type); entry 540 drivers/gpu/drm/i915/gvt/gtt.c GEM_BUG_ON(!gtt_type_is_entry(entry->type)); entry 543 drivers/gpu/drm/i915/gvt/gtt.c GEM_BUG_ON(entry->type == GTT_TYPE_INVALID); entry 550 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index, entry 557 drivers/gpu/drm/i915/gvt/gtt.c entry->type = mm->ppgtt_mm.root_entry_type; entry 560 drivers/gpu/drm/i915/gvt/gtt.c entry, index, false, 0, mm->vgpu); entry 561 drivers/gpu/drm/i915/gvt/gtt.c update_entry_type_for_real(pte_ops, entry, false); entry 565 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index) entry 567 drivers/gpu/drm/i915/gvt/gtt.c _ppgtt_get_root_entry(mm, entry, index, true); entry 571 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index) entry 573 drivers/gpu/drm/i915/gvt/gtt.c _ppgtt_get_root_entry(mm, entry, index, false); entry 577 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index, entry 584 drivers/gpu/drm/i915/gvt/gtt.c entry, index, false, 0, mm->vgpu); entry 588 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index) entry 590 drivers/gpu/drm/i915/gvt/gtt.c _ppgtt_set_root_entry(mm, entry, index, true); entry 594 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index) entry 596 drivers/gpu/drm/i915/gvt/gtt.c _ppgtt_set_root_entry(mm, entry, index, false); entry 600 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index) entry 606 drivers/gpu/drm/i915/gvt/gtt.c entry->type = GTT_TYPE_GGTT_PTE; entry 607 drivers/gpu/drm/i915/gvt/gtt.c pte_ops->get_entry(mm->ggtt_mm.virtual_ggtt, entry, index, entry 612 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index) entry 618 drivers/gpu/drm/i915/gvt/gtt.c pte_ops->set_entry(mm->ggtt_mm.virtual_ggtt, entry, index, entry 623 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index) entry 629 drivers/gpu/drm/i915/gvt/gtt.c pte_ops->get_entry(NULL, entry, index, false, 0, mm->vgpu); entry 633 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry, unsigned long index) entry 639 drivers/gpu/drm/i915/gvt/gtt.c pte_ops->set_entry(NULL, entry, index, false, 0, mm->vgpu); entry 975 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry) entry 982 drivers/gpu/drm/i915/gvt/gtt.c pfn = ops->get_pfn(entry); entry 1151 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry) entry 1159 drivers/gpu/drm/i915/gvt/gtt.c pfn = intel_gvt_hypervisor_gfn_to_mfn(vgpu, ops->get_pfn(entry)); entry 1218 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry entry = *se; entry 1229 drivers/gpu/drm/i915/gvt/gtt.c entry.type = GTT_TYPE_PPGTT_PTE_4K_ENTRY; entry 1230 drivers/gpu/drm/i915/gvt/gtt.c ops->set_64k_splited(&entry); entry 1238 drivers/gpu/drm/i915/gvt/gtt.c ops->set_pfn(&entry, dma_addr >> PAGE_SHIFT); entry 1239 drivers/gpu/drm/i915/gvt/gtt.c ppgtt_set_shadow_entry(spt, &entry, index + i); entry 2195 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry *entry) entry 2200 drivers/gpu/drm/i915/gvt/gtt.c pfn = pte_ops->get_pfn(entry); entry 2784 drivers/gpu/drm/i915/gvt/gtt.c struct intel_gvt_gtt_entry entry = {.type = GTT_TYPE_GGTT_PTE}; entry 2789 drivers/gpu/drm/i915/gvt/gtt.c pte_ops->set_pfn(&entry, gvt->gtt.scratch_mfn); entry 2790 drivers/gpu/drm/i915/gvt/gtt.c pte_ops->set_present(&entry); entry 2799 drivers/gpu/drm/i915/gvt/gtt.c ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++); entry 2809 drivers/gpu/drm/i915/gvt/gtt.c ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++); entry 306 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *entry) entry 308 drivers/gpu/drm/i915/gvt/kvmgt.c rb_erase(&entry->gfn_node, &vgpu->vdev.gfn_cache); entry 309 drivers/gpu/drm/i915/gvt/kvmgt.c rb_erase(&entry->dma_addr_node, &vgpu->vdev.dma_addr_cache); entry 310 drivers/gpu/drm/i915/gvt/kvmgt.c kfree(entry); entry 705 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *entry; entry 713 drivers/gpu/drm/i915/gvt/kvmgt.c entry = __gvt_cache_find_gfn(vgpu, iov_pfn); entry 714 drivers/gpu/drm/i915/gvt/kvmgt.c if (!entry) entry 717 drivers/gpu/drm/i915/gvt/kvmgt.c gvt_dma_unmap_page(vgpu, entry->gfn, entry->dma_addr, entry 718 drivers/gpu/drm/i915/gvt/kvmgt.c entry->size); entry 719 drivers/gpu/drm/i915/gvt/kvmgt.c __gvt_cache_remove_entry(vgpu, entry); entry 1889 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *entry; entry 1900 drivers/gpu/drm/i915/gvt/kvmgt.c entry = __gvt_cache_find_gfn(info->vgpu, gfn); entry 1901 drivers/gpu/drm/i915/gvt/kvmgt.c if (!entry) { entry 1909 drivers/gpu/drm/i915/gvt/kvmgt.c } else if (entry->size != size) { entry 1911 drivers/gpu/drm/i915/gvt/kvmgt.c gvt_dma_unmap_page(vgpu, gfn, entry->dma_addr, entry->size); entry 1912 drivers/gpu/drm/i915/gvt/kvmgt.c __gvt_cache_remove_entry(vgpu, entry); entry 1922 drivers/gpu/drm/i915/gvt/kvmgt.c kref_get(&entry->ref); entry 1923 drivers/gpu/drm/i915/gvt/kvmgt.c *dma_addr = entry->dma_addr; entry 1938 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *entry = container_of(ref, typeof(*entry), ref); entry 1940 drivers/gpu/drm/i915/gvt/kvmgt.c gvt_dma_unmap_page(entry->vgpu, entry->gfn, entry->dma_addr, entry 1941 drivers/gpu/drm/i915/gvt/kvmgt.c entry->size); entry 1942 drivers/gpu/drm/i915/gvt/kvmgt.c __gvt_cache_remove_entry(entry->vgpu, entry); entry 1948 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *entry; entry 1956 drivers/gpu/drm/i915/gvt/kvmgt.c entry = __gvt_cache_find_dma_addr(info->vgpu, dma_addr); entry 1957 drivers/gpu/drm/i915/gvt/kvmgt.c if (entry) entry 1958 drivers/gpu/drm/i915/gvt/kvmgt.c kref_put(&entry->ref, __gvt_dma_release); entry 2957 drivers/gpu/drm/i915/i915_debugfs.c struct skl_ddb_entry *entry; entry 2976 drivers/gpu/drm/i915/i915_debugfs.c entry = &crtc_state->wm.skl.plane_ddb_y[plane_id]; entry 2978 drivers/gpu/drm/i915/i915_debugfs.c entry->start, entry->end, entry 2979 drivers/gpu/drm/i915/i915_debugfs.c skl_ddb_entry_size(entry)); entry 2982 drivers/gpu/drm/i915/i915_debugfs.c entry = &crtc_state->wm.skl.plane_ddb_y[PLANE_CURSOR]; entry 2983 drivers/gpu/drm/i915/i915_debugfs.c seq_printf(m, " %-13s%8u%8u%8u\n", "Cursor", entry->start, entry 2984 drivers/gpu/drm/i915/i915_debugfs.c entry->end, skl_ddb_entry_size(entry)); entry 881 drivers/gpu/drm/i915/i915_drv.h static inline u16 skl_ddb_entry_size(const struct skl_ddb_entry *entry) entry 883 drivers/gpu/drm/i915/i915_drv.h return entry->end - entry->start; entry 776 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(atomic_read(px_used(pd)) > 2 * ARRAY_SIZE(pd->entry)); entry 779 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[idx] = to; entry 794 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[idx] = NULL; entry 939 drivers/gpu/drm/i915/i915_gem_gtt.c void **pde = pd->entry; entry 979 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_table *pt = pd->entry[idx]; entry 1057 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_table *pt = pd->entry[idx]; entry 1091 drivers/gpu/drm/i915/i915_gem_gtt.c if (likely(!pd->entry[idx])) entry 1094 drivers/gpu/drm/i915/i915_gem_gtt.c alloc = pt, pt = pd->entry[idx]; entry 1202 drivers/gpu/drm/i915/i915_gem_gtt.c pd = pdp->entry[gen8_pd_index(idx, 2)]; entry 1453 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(count > ARRAY_SIZE(pd->entry)); entry 1455 drivers/gpu/drm/i915/i915_gem_gtt.c pd = __alloc_pd(offsetof(typeof(*pd), entry[count])); entry 1657 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(pd->entry[act_pt] == &vm->scratch[1]); entry 1717 drivers/gpu/drm/i915/i915_gem_gtt.c if (pd->entry[pde] == &vm->scratch[1]) { entry 1718 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[pde] = pt; entry 1726 drivers/gpu/drm/i915/i915_gem_gtt.c pt = pd->entry[pde]; entry 1770 drivers/gpu/drm/i915/i915_gem_gtt.c memset_p(pd->entry, &vm->scratch[1], I915_PDES); entry 1856 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[pde] = scratch; entry 2681 drivers/gpu/drm/i915/i915_gem_gtt.c struct drm_mm_node *entry; entry 2715 drivers/gpu/drm/i915/i915_gem_gtt.c drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) { entry 244 drivers/gpu/drm/i915/i915_gem_gtt.h void *entry[512]; entry 520 drivers/gpu/drm/i915/i915_gem_gtt.h return pd->entry[n]; entry 527 drivers/gpu/drm/i915/i915_gem_gtt.h return pdp->entry[n]; entry 533 drivers/gpu/drm/i915/i915_gem_gtt.h struct i915_page_dma *pt = ppgtt->pd->entry[n]; entry 150 drivers/gpu/drm/i915/i915_sw_fence.c list_for_each_entry_safe(pos, next, &x->head, entry) { entry 154 drivers/gpu/drm/i915/i915_sw_fence.c list_move_tail(&pos->entry, continuation); entry 160 drivers/gpu/drm/i915/i915_sw_fence.c list_for_each_entry_safe(pos, next, &x->head, entry) { entry 240 drivers/gpu/drm/i915/i915_sw_fence.c list_del(&wq->entry); entry 259 drivers/gpu/drm/i915/i915_sw_fence.c list_for_each_entry(wq, &fence->wait.head, entry) { entry 277 drivers/gpu/drm/i915/i915_sw_fence.c list_for_each_entry(wq, &fence->wait.head, entry) { entry 338 drivers/gpu/drm/i915/i915_sw_fence.c INIT_LIST_HEAD(&wq->entry); entry 3974 drivers/gpu/drm/i915/intel_pm.c struct skl_ddb_entry *entry, u32 reg) entry 3977 drivers/gpu/drm/i915/intel_pm.c entry->start = reg & DDB_ENTRY_MASK; entry 3978 drivers/gpu/drm/i915/intel_pm.c entry->end = (reg >> DDB_ENTRY_END_SHIFT) & DDB_ENTRY_MASK; entry 3980 drivers/gpu/drm/i915/intel_pm.c if (entry->end) entry 3981 drivers/gpu/drm/i915/intel_pm.c entry->end += 1; entry 5123 drivers/gpu/drm/i915/intel_pm.c const struct skl_ddb_entry *entry) entry 5125 drivers/gpu/drm/i915/intel_pm.c if (entry->end) entry 5126 drivers/gpu/drm/i915/intel_pm.c I915_WRITE_FW(reg, (entry->end - 1) << 16 | entry->start); entry 821 drivers/gpu/drm/i915/intel_uncore.c static int fw_range_cmp(u32 offset, const struct intel_forcewake_range *entry) entry 823 drivers/gpu/drm/i915/intel_uncore.c if (offset < entry->start) entry 825 drivers/gpu/drm/i915/intel_uncore.c else if (offset > entry->end) entry 853 drivers/gpu/drm/i915/intel_uncore.c const struct intel_forcewake_range *entry; entry 855 drivers/gpu/drm/i915/intel_uncore.c entry = BSEARCH(offset, entry 860 drivers/gpu/drm/i915/intel_uncore.c if (!entry) entry 868 drivers/gpu/drm/i915/intel_uncore.c if (entry->domains == FORCEWAKE_ALL) entry 871 drivers/gpu/drm/i915/intel_uncore.c WARN(entry->domains & ~uncore->fw_domains, entry 873 drivers/gpu/drm/i915/intel_uncore.c entry->domains & ~uncore->fw_domains, offset); entry 875 drivers/gpu/drm/i915/intel_uncore.c return entry->domains; entry 1820 drivers/gpu/drm/i915/intel_uncore.c struct reg_whitelist const *entry; entry 1826 drivers/gpu/drm/i915/intel_uncore.c entry = reg_read_whitelist; entry 1829 drivers/gpu/drm/i915/intel_uncore.c u32 entry_offset = i915_mmio_reg_offset(entry->offset_ldw); entry 1831 drivers/gpu/drm/i915/intel_uncore.c GEM_BUG_ON(!is_power_of_2(entry->size)); entry 1832 drivers/gpu/drm/i915/intel_uncore.c GEM_BUG_ON(entry->size > 8); entry 1833 drivers/gpu/drm/i915/intel_uncore.c GEM_BUG_ON(entry_offset & (entry->size - 1)); entry 1835 drivers/gpu/drm/i915/intel_uncore.c if (INTEL_INFO(i915)->gen_mask & entry->gen_mask && entry 1836 drivers/gpu/drm/i915/intel_uncore.c entry_offset == (reg->offset & -entry->size)) entry 1838 drivers/gpu/drm/i915/intel_uncore.c entry++; entry 1845 drivers/gpu/drm/i915/intel_uncore.c flags = reg->offset & (entry->size - 1); entry 1848 drivers/gpu/drm/i915/intel_uncore.c if (entry->size == 8 && flags == I915_REG_READ_8B_WA) entry 1850 drivers/gpu/drm/i915/intel_uncore.c entry->offset_ldw, entry 1851 drivers/gpu/drm/i915/intel_uncore.c entry->offset_udw); entry 1852 drivers/gpu/drm/i915/intel_uncore.c else if (entry->size == 8 && flags == 0) entry 1854 drivers/gpu/drm/i915/intel_uncore.c entry->offset_ldw); entry 1855 drivers/gpu/drm/i915/intel_uncore.c else if (entry->size == 4 && flags == 0) entry 1856 drivers/gpu/drm/i915/intel_uncore.c reg->val = intel_uncore_read(uncore, entry->offset_ldw); entry 1857 drivers/gpu/drm/i915/intel_uncore.c else if (entry->size == 2 && flags == 0) entry 1859 drivers/gpu/drm/i915/intel_uncore.c entry->offset_ldw); entry 1860 drivers/gpu/drm/i915/intel_uncore.c else if (entry->size == 1 && flags == 0) entry 1862 drivers/gpu/drm/i915/intel_uncore.c entry->offset_ldw); entry 226 drivers/gpu/drm/mga/mga_dma.c drm_mga_freelist_t *entry; entry 235 drivers/gpu/drm/mga/mga_dma.c for (entry = dev_priv->head->next; entry; entry = entry->next) { entry 237 drivers/gpu/drm/mga/mga_dma.c entry, entry->buf->idx, entry->age.head, entry 238 drivers/gpu/drm/mga/mga_dma.c (unsigned long)(entry->age.head - dev_priv->primary->offset)); entry 249 drivers/gpu/drm/mga/mga_dma.c drm_mga_freelist_t *entry; entry 263 drivers/gpu/drm/mga/mga_dma.c entry = kzalloc(sizeof(drm_mga_freelist_t), GFP_KERNEL); entry 264 drivers/gpu/drm/mga/mga_dma.c if (entry == NULL) entry 267 drivers/gpu/drm/mga/mga_dma.c entry->next = dev_priv->head->next; entry 268 drivers/gpu/drm/mga/mga_dma.c entry->prev = dev_priv->head; entry 269 drivers/gpu/drm/mga/mga_dma.c SET_AGE(&entry->age, MGA_BUFFER_FREE, 0); entry 270 drivers/gpu/drm/mga/mga_dma.c entry->buf = buf; entry 273 drivers/gpu/drm/mga/mga_dma.c dev_priv->head->next->prev = entry; entry 274 drivers/gpu/drm/mga/mga_dma.c if (entry->next == NULL) entry 275 drivers/gpu/drm/mga/mga_dma.c dev_priv->tail = entry; entry 277 drivers/gpu/drm/mga/mga_dma.c buf_priv->list_entry = entry; entry 281 drivers/gpu/drm/mga/mga_dma.c dev_priv->head->next = entry; entry 290 drivers/gpu/drm/mga/mga_dma.c drm_mga_freelist_t *entry; entry 294 drivers/gpu/drm/mga/mga_dma.c entry = dev_priv->head; entry 295 drivers/gpu/drm/mga/mga_dma.c while (entry) { entry 296 drivers/gpu/drm/mga/mga_dma.c next = entry->next; entry 297 drivers/gpu/drm/mga/mga_dma.c kfree(entry); entry 298 drivers/gpu/drm/mga/mga_dma.c entry = next; entry 359 drivers/gpu/drm/mga/mga_dma.c drm_mga_freelist_t *head, *entry, *prev; entry 366 drivers/gpu/drm/mga/mga_dma.c entry = buf_priv->list_entry; entry 370 drivers/gpu/drm/mga/mga_dma.c SET_AGE(&entry->age, MGA_BUFFER_FREE, 0); entry 372 drivers/gpu/drm/mga/mga_dma.c prev->next = entry; entry 373 drivers/gpu/drm/mga/mga_dma.c entry->prev = prev; entry 374 drivers/gpu/drm/mga/mga_dma.c entry->next = NULL; entry 377 drivers/gpu/drm/mga/mga_dma.c head->next = entry; entry 378 drivers/gpu/drm/mga/mga_dma.c prev->prev = entry; entry 379 drivers/gpu/drm/mga/mga_dma.c entry->prev = head; entry 380 drivers/gpu/drm/mga/mga_dma.c entry->next = prev; entry 368 drivers/gpu/drm/mga/mga_drv.h drm_mga_freelist_t *entry = (buf_priv)->list_entry; \ entry 370 drivers/gpu/drm/mga/mga_drv.h entry->age.head = (dev_priv->prim.tail + \ entry 372 drivers/gpu/drm/mga/mga_drv.h entry->age.wrap = dev_priv->sarea_priv->last_wrap; \ entry 374 drivers/gpu/drm/mga/mga_drv.h entry->age.head = 0; \ entry 375 drivers/gpu/drm/mga/mga_drv.h entry->age.wrap = 0; \ entry 186 drivers/gpu/drm/msm/adreno/adreno_device.c static inline bool _rev_match(uint8_t entry, uint8_t id) entry 188 drivers/gpu/drm/msm/adreno/adreno_device.c return (entry == ANY_ID) || (entry == id); entry 446 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c struct dentry *entry; entry 448 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c entry = debugfs_create_dir("core_perf", parent); entry 450 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u64("max_core_clk_rate", 0600, entry, entry 452 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u64("core_clk_rate", 0600, entry, entry 454 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u32("enable_bw_release", 0600, entry, entry 456 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u32("threshold_low", 0600, entry, entry 458 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u32("threshold_high", 0600, entry, entry 460 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u32("min_core_ib", 0600, entry, entry 462 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u32("min_llcc_ib", 0600, entry, entry 464 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u32("min_dram_ib", 0600, entry, entry 466 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_file("perf_mode", 0600, entry, entry 468 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u64("fix_core_clk_rate", 0600, entry, entry 470 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u64("fix_core_ib_vote", 0600, entry, entry 472 drivers/gpu/drm/msm/disp/dpu1/dpu_core_perf.c debugfs_create_u64("fix_core_ab_vote", 0600, entry, entry 135 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c struct dentry *entry = debugfs_create_dir("danger", parent); entry 137 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c debugfs_create_file("danger_status", 0600, entry, entry 139 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c debugfs_create_file("safe_status", 0600, entry, entry 226 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c struct dentry *entry; entry 231 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c entry = debugfs_create_dir("debug", minor->debugfs_root); entry 233 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c debugfs_create_x32(DPU_DEBUGFS_HWMASKNAME, 0600, entry, p); entry 235 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c dpu_debugfs_danger_init(dpu_kms, entry); entry 236 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c dpu_debugfs_vbif_init(dpu_kms, entry); entry 237 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c dpu_debugfs_core_irq_init(dpu_kms, entry); entry 239 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c return dpu_core_perf_debugfs_init(dpu_kms, entry); entry 300 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c struct dentry *entry, *debugfs_vbif; entry 303 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c entry = debugfs_create_dir("vbif", debugfs_root); entry 310 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c debugfs_vbif = debugfs_create_dir(vbif_name, entry); entry 357 drivers/gpu/drm/msm/hdmi/hdmi.c #define HDMI_CFG(item, entry) \ entry 358 drivers/gpu/drm/msm/hdmi/hdmi.c .item ## _names = item ##_names_ ## entry, \ entry 359 drivers/gpu/drm/msm/hdmi/hdmi.c .item ## _cnt = ARRAY_SIZE(item ## _names_ ## entry) entry 21 drivers/gpu/drm/msm/msm_submitqueue.c struct msm_gpu_submitqueue *entry; entry 28 drivers/gpu/drm/msm/msm_submitqueue.c list_for_each_entry(entry, &ctx->submitqueues, node) { entry 29 drivers/gpu/drm/msm/msm_submitqueue.c if (entry->id == id) { entry 30 drivers/gpu/drm/msm/msm_submitqueue.c kref_get(&entry->ref); entry 33 drivers/gpu/drm/msm/msm_submitqueue.c return entry; entry 43 drivers/gpu/drm/msm/msm_submitqueue.c struct msm_gpu_submitqueue *entry, *tmp; entry 52 drivers/gpu/drm/msm/msm_submitqueue.c list_for_each_entry_safe(entry, tmp, &ctx->submitqueues, node) entry 53 drivers/gpu/drm/msm/msm_submitqueue.c msm_submitqueue_put(entry); entry 159 drivers/gpu/drm/msm/msm_submitqueue.c struct msm_gpu_submitqueue *entry; entry 173 drivers/gpu/drm/msm/msm_submitqueue.c list_for_each_entry(entry, &ctx->submitqueues, node) { entry 174 drivers/gpu/drm/msm/msm_submitqueue.c if (entry->id == id) { entry 175 drivers/gpu/drm/msm/msm_submitqueue.c list_del(&entry->node); entry 178 drivers/gpu/drm/msm/msm_submitqueue.c msm_submitqueue_put(entry); entry 527 drivers/gpu/drm/nouveau/dispnv04/dac.c nv04_dac_create(struct drm_connector *connector, struct dcb_output *entry) entry 540 drivers/gpu/drm/nouveau/dispnv04/dac.c nv_encoder->dcb = entry; entry 541 drivers/gpu/drm/nouveau/dispnv04/dac.c nv_encoder->or = ffs(entry->or) - 1; entry 555 drivers/gpu/drm/nouveau/dispnv04/dac.c encoder->possible_crtcs = entry->heads; entry 677 drivers/gpu/drm/nouveau/dispnv04/dfp.c nv04_dfp_create(struct drm_connector *connector, struct dcb_output *entry) entry 684 drivers/gpu/drm/nouveau/dispnv04/dfp.c switch (entry->type) { entry 706 drivers/gpu/drm/nouveau/dispnv04/dfp.c nv_encoder->dcb = entry; entry 707 drivers/gpu/drm/nouveau/dispnv04/dfp.c nv_encoder->or = ffs(entry->or) - 1; entry 712 drivers/gpu/drm/nouveau/dispnv04/dfp.c encoder->possible_crtcs = entry->heads; entry 715 drivers/gpu/drm/nouveau/dispnv04/dfp.c if (entry->type == DCB_OUTPUT_TMDS && entry 716 drivers/gpu/drm/nouveau/dispnv04/dfp.c entry->location != DCB_LOC_ON_CHIP) entry 228 drivers/gpu/drm/nouveau/dispnv04/disp.c struct dcb_output *dcbent = &dcb->entry[i]; entry 202 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c nv04_tv_create(struct drm_connector *connector, struct dcb_output *entry) entry 209 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c struct nvkm_i2c_bus *bus = nvkm_i2c_bus_find(i2c, entry->i2c_index); entry 213 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c type = nv04_tv_identify(dev, entry->i2c_index); entry 232 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c encoder->possible_crtcs = entry->heads; entry 234 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c nv_encoder->dcb = entry; entry 235 drivers/gpu/drm/nouveau/dispnv04/tvnv04.c nv_encoder->or = ffs(entry->or) - 1; entry 790 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c nv17_tv_create(struct drm_connector *connector, struct dcb_output *entry) entry 811 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c tv_enc->base.dcb = entry; entry 812 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c tv_enc->base.or = ffs(entry->or) - 1; entry 822 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c encoder->possible_crtcs = entry->heads; entry 2382 drivers/gpu/drm/nouveau/dispnv50/disp.c for (i = 0, dcbe = &dcb->entry[0]; i < dcb->entries; i++, dcbe++) { entry 134 drivers/gpu/drm/nouveau/include/nvif/list.h __list_add(struct list_head *entry, entry 137 drivers/gpu/drm/nouveau/include/nvif/list.h next->prev = entry; entry 138 drivers/gpu/drm/nouveau/include/nvif/list.h entry->next = next; entry 139 drivers/gpu/drm/nouveau/include/nvif/list.h entry->prev = prev; entry 140 drivers/gpu/drm/nouveau/include/nvif/list.h prev->next = entry; entry 159 drivers/gpu/drm/nouveau/include/nvif/list.h list_add(struct list_head *entry, struct list_head *head) entry 161 drivers/gpu/drm/nouveau/include/nvif/list.h __list_add(entry, head, head->next); entry 180 drivers/gpu/drm/nouveau/include/nvif/list.h list_add_tail(struct list_head *entry, struct list_head *head) entry 182 drivers/gpu/drm/nouveau/include/nvif/list.h __list_add(entry, head->prev, head); entry 207 drivers/gpu/drm/nouveau/include/nvif/list.h list_del(struct list_head *entry) entry 209 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(entry->prev, entry->next); entry 213 drivers/gpu/drm/nouveau/include/nvif/list.h list_del_init(struct list_head *entry) entry 215 drivers/gpu/drm/nouveau/include/nvif/list.h __list_del(entry->prev, entry->next); entry 216 drivers/gpu/drm/nouveau/include/nvif/list.h INIT_LIST_HEAD(entry); entry 66 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/dcb.h (struct nvkm_bios *, void *, int index, u16 entry)); entry 974 drivers/gpu/drm/nouveau/nouveau_bios.c u8 entries, *entry; entry 980 drivers/gpu/drm/nouveau/nouveau_bios.c entry = &bios->data[bios->offset + 12]; entry 982 drivers/gpu/drm/nouveau/nouveau_bios.c if (entry[0] == id) { entry 983 drivers/gpu/drm/nouveau/nouveau_bios.c bit->id = entry[0]; entry 984 drivers/gpu/drm/nouveau/nouveau_bios.c bit->version = entry[1]; entry 985 drivers/gpu/drm/nouveau/nouveau_bios.c bit->length = ROM16(entry[2]); entry 986 drivers/gpu/drm/nouveau/nouveau_bios.c bit->offset = ROM16(entry[4]); entry 987 drivers/gpu/drm/nouveau/nouveau_bios.c bit->data = ROMPTR(dev, entry[4]); entry 991 drivers/gpu/drm/nouveau/nouveau_bios.c entry += bios->data[bios->offset + 9]; entry 1373 drivers/gpu/drm/nouveau/nouveau_bios.c struct dcb_output *entry = &dcb->entry[dcb->entries]; entry 1375 drivers/gpu/drm/nouveau/nouveau_bios.c memset(entry, 0, sizeof(struct dcb_output)); entry 1376 drivers/gpu/drm/nouveau/nouveau_bios.c entry->index = dcb->entries++; entry 1378 drivers/gpu/drm/nouveau/nouveau_bios.c return entry; entry 1384 drivers/gpu/drm/nouveau/nouveau_bios.c struct dcb_output *entry = new_dcb_entry(dcb); entry 1386 drivers/gpu/drm/nouveau/nouveau_bios.c entry->type = type; entry 1387 drivers/gpu/drm/nouveau/nouveau_bios.c entry->i2c_index = i2c; entry 1388 drivers/gpu/drm/nouveau/nouveau_bios.c entry->heads = heads; entry 1390 drivers/gpu/drm/nouveau/nouveau_bios.c entry->location = !DCB_LOC_ON_CHIP; /* ie OFF CHIP */ entry 1391 drivers/gpu/drm/nouveau/nouveau_bios.c entry->or = or; entry 1396 drivers/gpu/drm/nouveau/nouveau_bios.c uint32_t conn, uint32_t conf, struct dcb_output *entry) entry 1401 drivers/gpu/drm/nouveau/nouveau_bios.c entry->type = conn & 0xf; entry 1402 drivers/gpu/drm/nouveau/nouveau_bios.c entry->i2c_index = (conn >> 4) & 0xf; entry 1403 drivers/gpu/drm/nouveau/nouveau_bios.c entry->heads = (conn >> 8) & 0xf; entry 1404 drivers/gpu/drm/nouveau/nouveau_bios.c entry->connector = (conn >> 12) & 0xf; entry 1405 drivers/gpu/drm/nouveau/nouveau_bios.c entry->bus = (conn >> 16) & 0xf; entry 1406 drivers/gpu/drm/nouveau/nouveau_bios.c entry->location = (conn >> 20) & 0x3; entry 1407 drivers/gpu/drm/nouveau/nouveau_bios.c entry->or = (conn >> 24) & 0xf; entry 1409 drivers/gpu/drm/nouveau/nouveau_bios.c switch (entry->type) { entry 1415 drivers/gpu/drm/nouveau/nouveau_bios.c entry->crtconf.maxfreq = (dcb->version < 0x30) ? entry 1423 drivers/gpu/drm/nouveau/nouveau_bios.c entry->lvdsconf.use_straps_for_mode = true; entry 1431 drivers/gpu/drm/nouveau/nouveau_bios.c entry->lvdsconf.use_straps_for_mode = true; entry 1437 drivers/gpu/drm/nouveau/nouveau_bios.c entry->lvdsconf.use_power_scripts = true; entry 1441 drivers/gpu/drm/nouveau/nouveau_bios.c entry->lvdsconf.use_acpi_for_edid = true; entry 1443 drivers/gpu/drm/nouveau/nouveau_bios.c entry->lvdsconf.use_power_scripts = true; entry 1444 drivers/gpu/drm/nouveau/nouveau_bios.c entry->lvdsconf.sor.link = (conf & 0x00000030) >> 4; entry 1445 drivers/gpu/drm/nouveau/nouveau_bios.c link = entry->lvdsconf.sor.link; entry 1463 drivers/gpu/drm/nouveau/nouveau_bios.c entry->tvconf.has_component_output = conf & (0x8 << 4); entry 1465 drivers/gpu/drm/nouveau/nouveau_bios.c entry->tvconf.has_component_output = false; entry 1470 drivers/gpu/drm/nouveau/nouveau_bios.c entry->dpconf.sor.link = (conf & 0x00000030) >> 4; entry 1471 drivers/gpu/drm/nouveau/nouveau_bios.c entry->extdev = (conf & 0x0000ff00) >> 8; entry 1474 drivers/gpu/drm/nouveau/nouveau_bios.c entry->dpconf.link_bw = 162000; entry 1477 drivers/gpu/drm/nouveau/nouveau_bios.c entry->dpconf.link_bw = 270000; entry 1480 drivers/gpu/drm/nouveau/nouveau_bios.c entry->dpconf.link_bw = 540000; entry 1484 drivers/gpu/drm/nouveau/nouveau_bios.c entry->dpconf.link_bw = 810000; entry 1490 drivers/gpu/drm/nouveau/nouveau_bios.c entry->dpconf.link_nr = 4; entry 1494 drivers/gpu/drm/nouveau/nouveau_bios.c entry->dpconf.link_nr = 2; entry 1497 drivers/gpu/drm/nouveau/nouveau_bios.c entry->dpconf.link_nr = 1; entry 1500 drivers/gpu/drm/nouveau/nouveau_bios.c link = entry->dpconf.sor.link; entry 1504 drivers/gpu/drm/nouveau/nouveau_bios.c entry->tmdsconf.sor.link = (conf & 0x00000030) >> 4; entry 1505 drivers/gpu/drm/nouveau/nouveau_bios.c entry->extdev = (conf & 0x0000ff00) >> 8; entry 1506 drivers/gpu/drm/nouveau/nouveau_bios.c link = entry->tmdsconf.sor.link; entry 1509 drivers/gpu/drm/nouveau/nouveau_bios.c entry->tmdsconf.slave_addr = (conf & 0x00000700) >> 8; entry 1511 drivers/gpu/drm/nouveau/nouveau_bios.c entry->tmdsconf.slave_addr = (conf & 0x00000070) >> 4; entry 1525 drivers/gpu/drm/nouveau/nouveau_bios.c entry->duallink_possible = entry 1526 drivers/gpu/drm/nouveau/nouveau_bios.c ((1 << (ffs(entry->or) - 1)) * 3 == entry->or); entry 1528 drivers/gpu/drm/nouveau/nouveau_bios.c entry->duallink_possible = (entry->sorconf.link == 3); entry 1533 drivers/gpu/drm/nouveau/nouveau_bios.c entry->i2c_upper_default = true; entry 1535 drivers/gpu/drm/nouveau/nouveau_bios.c entry->hasht = (entry->extdev << 8) | (entry->location << 4) | entry 1536 drivers/gpu/drm/nouveau/nouveau_bios.c entry->type; entry 1537 drivers/gpu/drm/nouveau/nouveau_bios.c entry->hashm = (entry->heads << 8) | (link << 6) | entry->or; entry 1543 drivers/gpu/drm/nouveau/nouveau_bios.c uint32_t conn, uint32_t conf, struct dcb_output *entry) entry 1549 drivers/gpu/drm/nouveau/nouveau_bios.c entry->type = DCB_OUTPUT_ANALOG; entry 1552 drivers/gpu/drm/nouveau/nouveau_bios.c entry->type = DCB_OUTPUT_TV; entry 1557 drivers/gpu/drm/nouveau/nouveau_bios.c entry->type = DCB_OUTPUT_LVDS; entry 1559 drivers/gpu/drm/nouveau/nouveau_bios.c entry->type = DCB_OUTPUT_TMDS; entry 1562 drivers/gpu/drm/nouveau/nouveau_bios.c entry->type = DCB_OUTPUT_LVDS; entry 1569 drivers/gpu/drm/nouveau/nouveau_bios.c entry->i2c_index = (conn & 0x0003c000) >> 14; entry 1570 drivers/gpu/drm/nouveau/nouveau_bios.c entry->heads = ((conn & 0x001c0000) >> 18) + 1; entry 1571 drivers/gpu/drm/nouveau/nouveau_bios.c entry->or = entry->heads; /* same as heads, hopefully safe enough */ entry 1572 drivers/gpu/drm/nouveau/nouveau_bios.c entry->location = (conn & 0x01e00000) >> 21; entry 1573 drivers/gpu/drm/nouveau/nouveau_bios.c entry->bus = (conn & 0x0e000000) >> 25; entry 1574 drivers/gpu/drm/nouveau/nouveau_bios.c entry->duallink_possible = false; entry 1576 drivers/gpu/drm/nouveau/nouveau_bios.c switch (entry->type) { entry 1578 drivers/gpu/drm/nouveau/nouveau_bios.c entry->crtconf.maxfreq = (conf & 0xffff) * 10; entry 1581 drivers/gpu/drm/nouveau/nouveau_bios.c entry->tvconf.has_component_output = false; entry 1585 drivers/gpu/drm/nouveau/nouveau_bios.c entry->lvdsconf.use_straps_for_mode = true; entry 1586 drivers/gpu/drm/nouveau/nouveau_bios.c entry->lvdsconf.use_power_scripts = true; entry 1608 drivers/gpu/drm/nouveau/nouveau_bios.c struct dcb_output *ient = &dcb->entry[i]; entry 1612 drivers/gpu/drm/nouveau/nouveau_bios.c struct dcb_output *jent = &dcb->entry[j]; entry 1632 drivers/gpu/drm/nouveau/nouveau_bios.c if (dcb->entry[i].type == 100) entry 1636 drivers/gpu/drm/nouveau/nouveau_bios.c dcb->entry[newentries] = dcb->entry[i]; entry 1637 drivers/gpu/drm/nouveau/nouveau_bios.c dcb->entry[newentries].index = newentries; entry 1793 drivers/gpu/drm/nouveau/nouveau_bios.c struct dcb_output *entry = new_dcb_entry(dcb); entry 1798 drivers/gpu/drm/nouveau/nouveau_bios.c ret = parse_dcb20_entry(dev, dcb, conn, conf, entry); entry 1800 drivers/gpu/drm/nouveau/nouveau_bios.c ret = parse_dcb15_entry(dev, dcb, conn, conf, entry); entry 1808 drivers/gpu/drm/nouveau/nouveau_bios.c if (entry->type == DCB_OUTPUT_TV && entry 1809 drivers/gpu/drm/nouveau/nouveau_bios.c entry->location == DCB_LOC_ON_CHIP) entry 1810 drivers/gpu/drm/nouveau/nouveau_bios.c entry->i2c_index = 0x0f; entry 1830 drivers/gpu/drm/nouveau/nouveau_bios.c if (dcbt->entry[i].connector) entry 1842 drivers/gpu/drm/nouveau/nouveau_bios.c u8 i2c = dcbt->entry[i].i2c_index; entry 1844 drivers/gpu/drm/nouveau/nouveau_bios.c dcbt->entry[i].connector = idx++; entry 1848 drivers/gpu/drm/nouveau/nouveau_bios.c dcbt->entry[i].connector = map[i2c] - 1; entry 1909 drivers/gpu/drm/nouveau/nouveau_bios.c static int load_nv17_hwsq_ucode_entry(struct drm_device *dev, struct nvbios *bios, uint16_t hwsq_offset, int entry) entry 1926 drivers/gpu/drm/nouveau/nouveau_bios.c if (bios->data[hwsq_offset] <= entry) { entry 1941 drivers/gpu/drm/nouveau/nouveau_bios.c hwsq_entry_offset = hwsq_offset + 2 + entry * bytes_to_write; entry 57 drivers/gpu/drm/nouveau/nouveau_bios.h struct dcb_output entry[DCB_MAX_NUM_ENTRIES]; entry 213 drivers/gpu/drm/nouveau/nouveau_bo.c INIT_LIST_HEAD(&nvbo->entry); entry 23 drivers/gpu/drm/nouveau/nouveau_bo.h struct list_head entry; entry 1289 drivers/gpu/drm/nouveau/nouveau_connector.c u32 entry = ROM16(nv_connector->dcb[0]); entry 1291 drivers/gpu/drm/nouveau/nouveau_connector.c entry |= (u32)ROM16(nv_connector->dcb[2]) << 16; entry 1326 drivers/gpu/drm/nouveau/nouveau_connector.c if (dcbt->entry[i].connector == nv_connector->index) entry 1327 drivers/gpu/drm/nouveau/nouveau_connector.c encoders |= (1 << dcbt->entry[i].type); entry 338 drivers/gpu/drm/nouveau/nouveau_gem.c nvbo = list_entry(op->list.next, struct nouveau_bo, entry); entry 358 drivers/gpu/drm/nouveau/nouveau_gem.c list_del(&nvbo->entry); entry 457 drivers/gpu/drm/nouveau/nouveau_gem.c list_add_tail(&nvbo->entry, &both_list); entry 460 drivers/gpu/drm/nouveau/nouveau_gem.c list_add_tail(&nvbo->entry, &vram_list); entry 463 drivers/gpu/drm/nouveau/nouveau_gem.c list_add_tail(&nvbo->entry, &gart_list); entry 467 drivers/gpu/drm/nouveau/nouveau_gem.c list_add_tail(&nvbo->entry, &both_list); entry 496 drivers/gpu/drm/nouveau/nouveau_gem.c list_for_each_entry(nvbo, list, entry) { entry 32 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c u32 entry = bios->bit_offset + 12; entry 34 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c if (nvbios_rd08(bios, entry + 0) == id) { entry 35 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c bit->id = nvbios_rd08(bios, entry + 0); entry 36 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c bit->version = nvbios_rd08(bios, entry + 1); entry 37 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c bit->length = nvbios_rd16(bios, entry + 2); entry 38 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c bit->offset = nvbios_rd16(bios, entry + 4); entry 42 drivers/gpu/drm/nouveau/nvkm/subdev/bios/bit.c entry += nvbios_rd08(bios, bios->bit_offset + 9); entry 74 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c struct nvbios_extdev_func *entry) entry 76 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c entry->type = nvbios_rd08(bios, offset + 0); entry 77 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c entry->addr = nvbios_rd08(bios, offset + 1); entry 78 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c entry->bus = (nvbios_rd08(bios, offset + 2) >> 4) & 1; entry 86 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c u16 entry; entry 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c if (!(entry = nvbios_extdev_entry(bios, idx, &ver, &len))) entry 91 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c extdev_parse_entry(bios, entry, func); entry 100 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c u16 entry; entry 103 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c while ((entry = nvbios_extdev_entry(bios, i++, &ver, &len))) { entry 104 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c extdev_parse_entry(bios, entry, func); entry 64 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c u32 table, entry; entry 87 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c entry = table + hdr + i * len; entry 91 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c if ((nvbios_rd08(bios, entry + 0x1) & 0xf8) == 0xf8) entry 95 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c rail->extdev_id = nvbios_rd08(bios, entry + 0x2); entry 99 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c rail->mode = nvbios_rd08(bios, entry); entry 100 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c rail->extdev_id = nvbios_rd08(bios, entry + 0x1); entry 122 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c rail->resistors[r].mohm = nvbios_rd08(bios, entry + res_start + r * 2); entry 123 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c rail->resistors[r].enabled = !(nvbios_rd08(bios, entry + res_start + r * 2 + 1) & 0x40); entry 125 drivers/gpu/drm/nouveau/nvkm/subdev/bios/iccsense.c rail->config = nvbios_rd16(bios, entry + res_start + rail->resistor_count * 2); entry 105 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c u8 idx, struct nvbios_power_budget_entry *entry) entry 110 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c || !entry) entry 116 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c entry->min_w = nvbios_rd32(bios, entry_offset + 0x2); entry 117 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c entry->avg_w = nvbios_rd32(bios, entry_offset + 0x6); entry 118 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c entry->max_w = nvbios_rd32(bios, entry_offset + 0xa); entry 120 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c entry->min_w = 0; entry 121 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c entry->max_w = nvbios_rd32(bios, entry_offset + 0x2); entry 122 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c entry->avg_w = entry->max_w; entry 73 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c u32 entry; entry 83 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c while ((entry = nvbios_therm_entry(bios, i++, &ver, &len))) { entry 84 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c s16 value = nvbios_rd16(bios, entry + 1); entry 86 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c switch (nvbios_rd08(bios, entry + 0)) { entry 95 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c offset = ((s8) nvbios_rd08(bios, entry + 2)) / 2; entry 157 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c u32 entry; entry 165 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c while ((entry = nvbios_therm_entry(bios, i++, &ver, &len))) { entry 166 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c s16 value = nvbios_rd16(bios, entry + 1); entry 168 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c switch (nvbios_rd08(bios, entry + 0)) { entry 199 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c fan->linear_min_temp = nvbios_rd08(bios, entry + 1); entry 200 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c fan->linear_max_temp = nvbios_rd08(bios, entry + 2); entry 32 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/gf119.c u16 entry; entry 35 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/gf119.c while ((entry = dcb_gpio_entry(bios, 0, ++ent, &ver, &len))) { entry 36 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/gf119.c u32 data = nvbios_rd32(bios, entry); entry 32 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/nv50.c u16 entry; entry 35 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/nv50.c while ((entry = dcb_gpio_entry(bios, 0, ++ent, &ver, &len))) { entry 37 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/nv50.c u32 data = nvbios_rd32(bios, entry); entry 229 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c struct nvbios_power_budget_entry entry; entry 231 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c budget.cap_entry, &entry); entry 233 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c iccsense->power_w_max = entry.avg_w; entry 234 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c iccsense->power_w_crit = entry.max_w; entry 908 drivers/gpu/drm/omapdrm/dss/dss.c struct dss_debugfs_entry *entry = inode->i_private; entry 910 drivers/gpu/drm/omapdrm/dss/dss.c return single_open(file, entry->show_fn, entry->data); entry 925 drivers/gpu/drm/omapdrm/dss/dss.c struct dss_debugfs_entry *entry; entry 927 drivers/gpu/drm/omapdrm/dss/dss.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 928 drivers/gpu/drm/omapdrm/dss/dss.c if (!entry) entry 931 drivers/gpu/drm/omapdrm/dss/dss.c entry->show_fn = show_fn; entry 932 drivers/gpu/drm/omapdrm/dss/dss.c entry->data = data; entry 933 drivers/gpu/drm/omapdrm/dss/dss.c entry->dentry = debugfs_create_file(name, 0444, dss->debugfs.root, entry 934 drivers/gpu/drm/omapdrm/dss/dss.c entry, &dss_debug_fops); entry 936 drivers/gpu/drm/omapdrm/dss/dss.c return entry; entry 939 drivers/gpu/drm/omapdrm/dss/dss.c void dss_debugfs_remove_file(struct dss_debugfs_entry *entry) entry 941 drivers/gpu/drm/omapdrm/dss/dss.c if (IS_ERR_OR_NULL(entry)) entry 944 drivers/gpu/drm/omapdrm/dss/dss.c debugfs_remove(entry->dentry); entry 945 drivers/gpu/drm/omapdrm/dss/dss.c kfree(entry); entry 287 drivers/gpu/drm/omapdrm/dss/dss.h void dss_debugfs_remove_file(struct dss_debugfs_entry *entry); entry 297 drivers/gpu/drm/omapdrm/dss/dss.h static inline void dss_debugfs_remove_file(struct dss_debugfs_entry *entry) entry 121 drivers/gpu/drm/omapdrm/omap_gem.c struct omap_drm_usergart_entry entry[NUM_USERGART_ENTRIES]; entry 167 drivers/gpu/drm/omapdrm/omap_gem.c enum tiler_fmt fmt, struct omap_drm_usergart_entry *entry) entry 174 drivers/gpu/drm/omapdrm/omap_gem.c (entry->obj_pgoff << PAGE_SHIFT); entry 190 drivers/gpu/drm/omapdrm/omap_gem.c entry->obj = NULL; entry 204 drivers/gpu/drm/omapdrm/omap_gem.c struct omap_drm_usergart_entry *entry = entry 205 drivers/gpu/drm/omapdrm/omap_gem.c &priv->usergart[fmt].entry[i]; entry 207 drivers/gpu/drm/omapdrm/omap_gem.c if (entry->obj == obj) entry 208 drivers/gpu/drm/omapdrm/omap_gem.c omap_gem_evict_entry(obj, fmt, entry); entry 377 drivers/gpu/drm/omapdrm/omap_gem.c struct omap_drm_usergart_entry *entry; entry 416 drivers/gpu/drm/omapdrm/omap_gem.c entry = &priv->usergart[fmt].entry[priv->usergart[fmt].last]; entry 419 drivers/gpu/drm/omapdrm/omap_gem.c if (entry->obj) entry 420 drivers/gpu/drm/omapdrm/omap_gem.c omap_gem_evict_entry(entry->obj, fmt, entry); entry 422 drivers/gpu/drm/omapdrm/omap_gem.c entry->obj = obj; entry 423 drivers/gpu/drm/omapdrm/omap_gem.c entry->obj_pgoff = base_pgoff; entry 431 drivers/gpu/drm/omapdrm/omap_gem.c entry->obj_pgoff += off; entry 450 drivers/gpu/drm/omapdrm/omap_gem.c err = tiler_pin(entry->block, pages, ARRAY_SIZE(pages), 0, true); entry 457 drivers/gpu/drm/omapdrm/omap_gem.c pfn = entry->dma_addr >> PAGE_SHIFT; entry 1349 drivers/gpu/drm/omapdrm/omap_gem.c struct omap_drm_usergart_entry *entry; entry 1352 drivers/gpu/drm/omapdrm/omap_gem.c entry = &usergart[i].entry[j]; entry 1360 drivers/gpu/drm/omapdrm/omap_gem.c entry->dma_addr = tiler_ssptr(block); entry 1361 drivers/gpu/drm/omapdrm/omap_gem.c entry->block = block; entry 1364 drivers/gpu/drm/omapdrm/omap_gem.c &entry->dma_addr, entry 240 drivers/gpu/drm/panel/panel-raydium-rm67191.c const struct cmd_set_entry *entry = &manufacturer_cmd_set[i]; entry 241 drivers/gpu/drm/panel/panel-raydium-rm67191.c u8 buffer[2] = { entry->cmd, entry->param }; entry 163 drivers/gpu/drm/qxl/qxl_release.c struct qxl_bo_list *entry; entry 166 drivers/gpu/drm/qxl/qxl_release.c entry = container_of(release->bos.next, entry 168 drivers/gpu/drm/qxl/qxl_release.c bo = to_qxl_bo(entry->tv.bo); entry 170 drivers/gpu/drm/qxl/qxl_release.c list_del(&entry->tv.head); entry 171 drivers/gpu/drm/qxl/qxl_release.c kfree(entry); entry 211 drivers/gpu/drm/qxl/qxl_release.c struct qxl_bo_list *entry; entry 213 drivers/gpu/drm/qxl/qxl_release.c list_for_each_entry(entry, &release->bos, tv.head) { entry 214 drivers/gpu/drm/qxl/qxl_release.c if (entry->tv.bo == &bo->tbo) entry 218 drivers/gpu/drm/qxl/qxl_release.c entry = kmalloc(sizeof(struct qxl_bo_list), GFP_KERNEL); entry 219 drivers/gpu/drm/qxl/qxl_release.c if (!entry) entry 223 drivers/gpu/drm/qxl/qxl_release.c entry->tv.bo = &bo->tbo; entry 224 drivers/gpu/drm/qxl/qxl_release.c entry->tv.num_shared = 0; entry 225 drivers/gpu/drm/qxl/qxl_release.c list_add_tail(&entry->tv.head, &release->bos); entry 255 drivers/gpu/drm/qxl/qxl_release.c struct qxl_bo_list *entry; entry 267 drivers/gpu/drm/qxl/qxl_release.c list_for_each_entry(entry, &release->bos, tv.head) { entry 268 drivers/gpu/drm/qxl/qxl_release.c struct qxl_bo *bo = to_qxl_bo(entry->tv.bo); entry 434 drivers/gpu/drm/qxl/qxl_release.c struct ttm_validate_buffer *entry; entry 458 drivers/gpu/drm/qxl/qxl_release.c list_for_each_entry(entry, &release->bos, head) { entry 459 drivers/gpu/drm/qxl/qxl_release.c bo = entry->bo; entry 779 drivers/gpu/drm/r128/r128_cce.c drm_r128_freelist_t *entry; entry 792 drivers/gpu/drm/r128/r128_cce.c entry = kmalloc(sizeof(drm_r128_freelist_t), GFP_KERNEL); entry 793 drivers/gpu/drm/r128/r128_cce.c if (!entry) entry 796 drivers/gpu/drm/r128/r128_cce.c entry->age = R128_BUFFER_FREE; entry 797 drivers/gpu/drm/r128/r128_cce.c entry->buf = buf; entry 798 drivers/gpu/drm/r128/r128_cce.c entry->prev = dev_priv->head; entry 799 drivers/gpu/drm/r128/r128_cce.c entry->next = dev_priv->head->next; entry 800 drivers/gpu/drm/r128/r128_cce.c if (!entry->next) entry 801 drivers/gpu/drm/r128/r128_cce.c dev_priv->tail = entry; entry 805 drivers/gpu/drm/r128/r128_cce.c buf_priv->list_entry = entry; entry 807 drivers/gpu/drm/r128/r128_cce.c dev_priv->head->next = entry; entry 810 drivers/gpu/drm/r128/r128_cce.c dev_priv->head->next->prev = entry; entry 4704 drivers/gpu/drm/radeon/ci_dpm.c static void ci_convert_mc_registers(const struct ci_mc_reg_entry *entry, entry 4712 drivers/gpu/drm/radeon/ci_dpm.c data->value[i] = cpu_to_be32(entry->mc_data[j]); entry 814 drivers/gpu/drm/radeon/cypress_dpm.c static void cypress_convert_mc_registers(struct evergreen_mc_reg_entry *entry, entry 822 drivers/gpu/drm/radeon/cypress_dpm.c data->value[i] = cpu_to_be32(entry->mc_data[j]); entry 1032 drivers/gpu/drm/radeon/cypress_dpm.c struct evergreen_mc_reg_entry *entry) entry 1038 drivers/gpu/drm/radeon/cypress_dpm.c entry->mc_data[i] = entry 2943 drivers/gpu/drm/radeon/ni_dpm.c static void ni_convert_mc_registers(struct ni_mc_reg_entry *entry, entry 2951 drivers/gpu/drm/radeon/ni_dpm.c data->value[i] = cpu_to_be32(entry->mc_data[j]); entry 699 drivers/gpu/drm/radeon/r100.c uint64_t entry) entry 702 drivers/gpu/drm/radeon/r100.c gtt[i] = cpu_to_le32(lower_32_bits(entry)); entry 121 drivers/gpu/drm/radeon/r300.c uint64_t entry) entry 128 drivers/gpu/drm/radeon/r300.c writel(entry, ((void __iomem *)ptr) + (i * 4)); entry 826 drivers/gpu/drm/radeon/r600_dpm.c ATOM_PPLIB_Clock_Voltage_Dependency_Record *entry; entry 832 drivers/gpu/drm/radeon/r600_dpm.c entry = &atom_table->entries[0]; entry 834 drivers/gpu/drm/radeon/r600_dpm.c radeon_table->entries[i].clk = le16_to_cpu(entry->usClockLow) | entry 835 drivers/gpu/drm/radeon/r600_dpm.c (entry->ucClockHigh << 16); entry 836 drivers/gpu/drm/radeon/r600_dpm.c radeon_table->entries[i].v = le16_to_cpu(entry->usVoltage); entry 837 drivers/gpu/drm/radeon/r600_dpm.c entry = (ATOM_PPLIB_Clock_Voltage_Dependency_Record *) entry 838 drivers/gpu/drm/radeon/r600_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_Clock_Voltage_Dependency_Record)); entry 990 drivers/gpu/drm/radeon/r600_dpm.c ATOM_PPLIB_PhaseSheddingLimits_Record *entry; entry 1001 drivers/gpu/drm/radeon/r600_dpm.c entry = &psl->entries[0]; entry 1004 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usSclkLow) | (entry->ucSclkHigh << 16); entry 1006 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usMclkLow) | (entry->ucMclkHigh << 16); entry 1008 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVoltage); entry 1009 drivers/gpu/drm/radeon/r600_dpm.c entry = (ATOM_PPLIB_PhaseSheddingLimits_Record *) entry 1010 drivers/gpu/drm/radeon/r600_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_PhaseSheddingLimits_Record)); entry 1037 drivers/gpu/drm/radeon/r600_dpm.c ATOM_PPLIB_CAC_Leakage_Record *entry; entry 1044 drivers/gpu/drm/radeon/r600_dpm.c entry = &cac_table->entries[0]; entry 1048 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVddc1); entry 1050 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVddc2); entry 1052 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVddc3); entry 1055 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVddc); entry 1057 drivers/gpu/drm/radeon/r600_dpm.c le32_to_cpu(entry->ulLeakageValue); entry 1059 drivers/gpu/drm/radeon/r600_dpm.c entry = (ATOM_PPLIB_CAC_Leakage_Record *) entry 1060 drivers/gpu/drm/radeon/r600_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_CAC_Leakage_Record)); entry 1088 drivers/gpu/drm/radeon/r600_dpm.c ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record *entry; entry 1101 drivers/gpu/drm/radeon/r600_dpm.c entry = &limits->entries[0]; entry 1106 drivers/gpu/drm/radeon/r600_dpm.c (entry->ucVCEClockInfoIndex * sizeof(VCEClockInfo))); entry 1112 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVoltage); entry 1113 drivers/gpu/drm/radeon/r600_dpm.c entry = (ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record *) entry 1114 drivers/gpu/drm/radeon/r600_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_VCE_Clock_Voltage_Limit_Record)); entry 1144 drivers/gpu/drm/radeon/r600_dpm.c ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *entry; entry 1155 drivers/gpu/drm/radeon/r600_dpm.c entry = &limits->entries[0]; entry 1159 drivers/gpu/drm/radeon/r600_dpm.c (entry->ucUVDClockInfoIndex * sizeof(UVDClockInfo))); entry 1165 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVoltage); entry 1166 drivers/gpu/drm/radeon/r600_dpm.c entry = (ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *) entry 1167 drivers/gpu/drm/radeon/r600_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record)); entry 1176 drivers/gpu/drm/radeon/r600_dpm.c ATOM_PPLIB_SAMClk_Voltage_Limit_Record *entry; entry 1187 drivers/gpu/drm/radeon/r600_dpm.c entry = &limits->entries[0]; entry 1190 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usSAMClockLow) | (entry->ucSAMClockHigh << 16); entry 1192 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVoltage); entry 1193 drivers/gpu/drm/radeon/r600_dpm.c entry = (ATOM_PPLIB_SAMClk_Voltage_Limit_Record *) entry 1194 drivers/gpu/drm/radeon/r600_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_SAMClk_Voltage_Limit_Record)); entry 1234 drivers/gpu/drm/radeon/r600_dpm.c ATOM_PPLIB_ACPClk_Voltage_Limit_Record *entry; entry 1245 drivers/gpu/drm/radeon/r600_dpm.c entry = &limits->entries[0]; entry 1248 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usACPClockLow) | (entry->ucACPClockHigh << 16); entry 1250 drivers/gpu/drm/radeon/r600_dpm.c le16_to_cpu(entry->usVoltage); entry 1251 drivers/gpu/drm/radeon/r600_dpm.c entry = (ATOM_PPLIB_ACPClk_Voltage_Limit_Record *) entry 1252 drivers/gpu/drm/radeon/r600_dpm.c ((u8 *)entry + sizeof(ATOM_PPLIB_ACPClk_Voltage_Limit_Record)); entry 252 drivers/gpu/drm/radeon/radeon.h uint64_t entry; entry 1865 drivers/gpu/drm/radeon/radeon.h uint64_t entry); entry 685 drivers/gpu/drm/radeon/radeon_acpi.c struct acpi_bus_event *entry = (struct acpi_bus_event *)data; entry 687 drivers/gpu/drm/radeon/radeon_acpi.c if (strcmp(entry->device_class, ACPI_AC_CLASS) == 0) { entry 697 drivers/gpu/drm/radeon/radeon_acpi.c return radeon_atif_handler(rdev, entry); entry 72 drivers/gpu/drm/radeon/radeon_asic.h uint64_t entry); entry 178 drivers/gpu/drm/radeon/radeon_asic.h uint64_t entry); entry 215 drivers/gpu/drm/radeon/radeon_asic.h uint64_t entry); entry 240 drivers/gpu/drm/radeon/radeon_asic.h uint64_t entry); entry 2309 drivers/gpu/drm/radeon/radeon_combios.c uint32_t conn_info, entry, devices; entry 2320 drivers/gpu/drm/radeon/radeon_combios.c entry = conn_info + 2 + i * 2; entry 2322 drivers/gpu/drm/radeon/radeon_combios.c if (!RBIOS16(entry)) entry 2325 drivers/gpu/drm/radeon/radeon_combios.c tmp = RBIOS16(entry); entry 795 drivers/gpu/drm/radeon/radeon_device.c rdev->dummy_page.entry = radeon_gart_get_page_entry(rdev->dummy_page.addr, entry 261 drivers/gpu/drm/radeon/radeon_gart.c rdev->gart.pages_entry[t] = rdev->dummy_page.entry; entry 264 drivers/gpu/drm/radeon/radeon_gart.c rdev->dummy_page.entry); entry 367 drivers/gpu/drm/radeon/radeon_gart.c rdev->gart.pages_entry[i] = rdev->dummy_page.entry; entry 552 drivers/gpu/drm/radeon/radeon_gem.c struct ttm_validate_buffer tv, *entry; entry 573 drivers/gpu/drm/radeon/radeon_gem.c list_for_each_entry(entry, &list, head) { entry 574 drivers/gpu/drm/radeon/radeon_gem.c domain = radeon_mem_type_to_domain(entry->bo->mem.mem_type); entry 222 drivers/gpu/drm/radeon/rs400.c uint32_t entry; entry 224 drivers/gpu/drm/radeon/rs400.c entry = (lower_32_bits(addr) & PAGE_MASK) | entry 227 drivers/gpu/drm/radeon/rs400.c entry |= RS400_PTE_READABLE; entry 229 drivers/gpu/drm/radeon/rs400.c entry |= RS400_PTE_WRITEABLE; entry 231 drivers/gpu/drm/radeon/rs400.c entry |= RS400_PTE_UNSNOOPED; entry 232 drivers/gpu/drm/radeon/rs400.c return entry; entry 236 drivers/gpu/drm/radeon/rs400.c uint64_t entry) entry 239 drivers/gpu/drm/radeon/rs400.c gtt[i] = cpu_to_le32(lower_32_bits(entry)); entry 654 drivers/gpu/drm/radeon/rs600.c uint64_t entry) entry 657 drivers/gpu/drm/radeon/rs600.c writeq(entry, ptr + (i * 8)); entry 598 drivers/gpu/drm/radeon/rv6xx_dpm.c u32 entry, u32 clock) entry 606 drivers/gpu/drm/radeon/rv6xx_dpm.c rv6xx_memory_clock_entry_set_reference_divider(rdev, entry, dividers.ref_div); entry 607 drivers/gpu/drm/radeon/rv6xx_dpm.c rv6xx_memory_clock_entry_set_feedback_divider(rdev, entry, dividers.fb_div); entry 608 drivers/gpu/drm/radeon/rv6xx_dpm.c rv6xx_memory_clock_entry_set_post_divider(rdev, entry, dividers.post_div); entry 611 drivers/gpu/drm/radeon/rv6xx_dpm.c rv6xx_memory_clock_entry_enable_post_divider(rdev, entry, true); entry 613 drivers/gpu/drm/radeon/rv6xx_dpm.c rv6xx_memory_clock_entry_enable_post_divider(rdev, entry, false); entry 702 drivers/gpu/drm/radeon/rv6xx_dpm.c u32 entry, u16 voltage) entry 713 drivers/gpu/drm/radeon/rv6xx_dpm.c r600_voltage_control_program_voltages(rdev, entry, set_pins); entry 5603 drivers/gpu/drm/radeon/si_dpm.c static void si_convert_mc_registers(const struct si_mc_reg_entry *entry, entry 5611 drivers/gpu/drm/radeon/si_dpm.c data->value[i] = cpu_to_be32(entry->mc_data[j]); entry 220 drivers/gpu/drm/savage/savage_bci.c drm_savage_buf_priv_t *entry; entry 234 drivers/gpu/drm/savage/savage_bci.c entry = buf->dev_private; entry 236 drivers/gpu/drm/savage/savage_bci.c SET_AGE(&entry->age, 0, 0); entry 237 drivers/gpu/drm/savage/savage_bci.c entry->buf = buf; entry 239 drivers/gpu/drm/savage/savage_bci.c entry->next = dev_priv->head.next; entry 240 drivers/gpu/drm/savage/savage_bci.c entry->prev = &dev_priv->head; entry 241 drivers/gpu/drm/savage/savage_bci.c dev_priv->head.next->prev = entry; entry 242 drivers/gpu/drm/savage/savage_bci.c dev_priv->head.next = entry; entry 284 drivers/gpu/drm/savage/savage_bci.c drm_savage_buf_priv_t *entry = buf->dev_private, *prev, *next; entry 286 drivers/gpu/drm/savage/savage_bci.c DRM_DEBUG("age=0x%04x wrap=%d\n", entry->age.event, entry->age.wrap); entry 288 drivers/gpu/drm/savage/savage_bci.c if (entry->next != NULL || entry->prev != NULL) { entry 295 drivers/gpu/drm/savage/savage_bci.c prev->next = entry; entry 296 drivers/gpu/drm/savage/savage_bci.c next->prev = entry; entry 297 drivers/gpu/drm/savage/savage_bci.c entry->prev = prev; entry 298 drivers/gpu/drm/savage/savage_bci.c entry->next = next; entry 227 drivers/gpu/drm/sis/sis_mm.c struct drm_map_list *entry; entry 230 drivers/gpu/drm/sis/sis_mm.c list_for_each_entry(entry, &dev->maplist, head) { entry 231 drivers/gpu/drm/sis/sis_mm.c map = entry->map; entry 318 drivers/gpu/drm/sis/sis_mm.c struct sis_memblock *entry, *next; entry 336 drivers/gpu/drm/sis/sis_mm.c list_for_each_entry_safe(entry, next, &file_priv->obj_list, entry 338 drivers/gpu/drm/sis/sis_mm.c list_del(&entry->owner_list); entry 339 drivers/gpu/drm/sis/sis_mm.c if (drm_mm_node_allocated(&entry->mm_node)) entry 340 drivers/gpu/drm/sis/sis_mm.c drm_mm_remove_node(&entry->mm_node); entry 343 drivers/gpu/drm/sis/sis_mm.c sis_free(entry->req.offset); entry 345 drivers/gpu/drm/sis/sis_mm.c kfree(entry); entry 37 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_validate_buffer *entry) entry 39 drivers/gpu/drm/ttm/ttm_execbuf_util.c list_for_each_entry_continue_reverse(entry, list, head) { entry 40 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_buffer_object *bo = entry->bo; entry 48 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_validate_buffer *entry; entry 50 drivers/gpu/drm/ttm/ttm_execbuf_util.c list_for_each_entry(entry, list, head) { entry 51 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_buffer_object *bo = entry->bo; entry 59 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_validate_buffer *entry; entry 65 drivers/gpu/drm/ttm/ttm_execbuf_util.c entry = list_first_entry(list, struct ttm_validate_buffer, head); entry 66 drivers/gpu/drm/ttm/ttm_execbuf_util.c glob = entry->bo->bdev->glob; entry 69 drivers/gpu/drm/ttm/ttm_execbuf_util.c list_for_each_entry(entry, list, head) { entry 70 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_buffer_object *bo = entry->bo; entry 100 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_validate_buffer *entry; entry 106 drivers/gpu/drm/ttm/ttm_execbuf_util.c entry = list_first_entry(list, struct ttm_validate_buffer, head); entry 107 drivers/gpu/drm/ttm/ttm_execbuf_util.c glob = entry->bo->bdev->glob; entry 112 drivers/gpu/drm/ttm/ttm_execbuf_util.c list_for_each_entry(entry, list, head) { entry 113 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_buffer_object *bo = entry->bo; entry 122 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_validate_buffer *safe = entry; entry 123 drivers/gpu/drm/ttm/ttm_execbuf_util.c entry = list_prev_entry(entry, head); entry 130 drivers/gpu/drm/ttm/ttm_execbuf_util.c if (!entry->num_shared) entry 134 drivers/gpu/drm/ttm/ttm_execbuf_util.c entry->num_shared); entry 143 drivers/gpu/drm/ttm/ttm_execbuf_util.c ttm_eu_backoff_reservation_reverse(list, entry); entry 155 drivers/gpu/drm/ttm/ttm_execbuf_util.c if (!ret && entry->num_shared) entry 157 drivers/gpu/drm/ttm/ttm_execbuf_util.c entry->num_shared); entry 172 drivers/gpu/drm/ttm/ttm_execbuf_util.c list_del(&entry->head); entry 173 drivers/gpu/drm/ttm/ttm_execbuf_util.c list_add(&entry->head, list); entry 189 drivers/gpu/drm/ttm/ttm_execbuf_util.c struct ttm_validate_buffer *entry; entry 201 drivers/gpu/drm/ttm/ttm_execbuf_util.c list_for_each_entry(entry, list, head) { entry 202 drivers/gpu/drm/ttm/ttm_execbuf_util.c bo = entry->bo; entry 203 drivers/gpu/drm/ttm/ttm_execbuf_util.c if (entry->num_shared) entry 38 drivers/gpu/drm/udl/udl_drv.h struct list_head entry; entry 150 drivers/gpu/drm/udl/udl_main.c list_add_tail(&unode->entry, &udl->urbs.list); entry 187 drivers/gpu/drm/udl/udl_main.c unode = list_entry(node, struct urb_node, entry); entry 251 drivers/gpu/drm/udl/udl_main.c list_add_tail(&unode->entry, &udl->urbs.list); entry 267 drivers/gpu/drm/udl/udl_main.c struct list_head *entry; entry 283 drivers/gpu/drm/udl/udl_main.c entry = udl->urbs.list.next; entry 284 drivers/gpu/drm/udl/udl_main.c list_del_init(entry); entry 289 drivers/gpu/drm/udl/udl_main.c unode = list_entry(entry, struct urb_node, entry); entry 27 drivers/gpu/drm/vc4/vc4_debugfs.c struct vc4_debugfs_info_entry *entry; entry 32 drivers/gpu/drm/vc4/vc4_debugfs.c list_for_each_entry(entry, &vc4->debugfs_list, link) { entry 33 drivers/gpu/drm/vc4/vc4_debugfs.c int ret = drm_debugfs_create_files(&entry->info, 1, entry 70 drivers/gpu/drm/vc4/vc4_debugfs.c struct vc4_debugfs_info_entry *entry = entry 71 drivers/gpu/drm/vc4/vc4_debugfs.c devm_kzalloc(dev->dev, sizeof(*entry), GFP_KERNEL); entry 73 drivers/gpu/drm/vc4/vc4_debugfs.c if (!entry) entry 76 drivers/gpu/drm/vc4/vc4_debugfs.c entry->info.name = name; entry 77 drivers/gpu/drm/vc4/vc4_debugfs.c entry->info.show = show; entry 78 drivers/gpu/drm/vc4/vc4_debugfs.c entry->info.data = data; entry 80 drivers/gpu/drm/vc4/vc4_debugfs.c list_add(&entry->link, &vc4->debugfs_list); entry 1008 drivers/gpu/drm/vc4/vc4_gem.c list_for_each_entry_safe(cb, cb_temp, &vc4->seqno_cb_list, work.entry) { entry 1010 drivers/gpu/drm/vc4/vc4_gem.c list_del_init(&cb->work.entry); entry 1039 drivers/gpu/drm/vc4/vc4_gem.c list_add_tail(&cb->work.entry, &vc4->seqno_cb_list); entry 164 drivers/gpu/drm/via/via_drv.h DECLARE_WAITQUEUE(entry, current); \ entry 166 drivers/gpu/drm/via/via_drv.h add_wait_queue(&(queue), &entry); \ entry 183 drivers/gpu/drm/via/via_drv.h remove_wait_queue(&(queue), &entry); \ entry 214 drivers/gpu/drm/via/via_mm.c struct via_memblock *entry, *next; entry 231 drivers/gpu/drm/via/via_mm.c list_for_each_entry_safe(entry, next, &file_priv->obj_list, entry 233 drivers/gpu/drm/via/via_mm.c list_del(&entry->owner_list); entry 234 drivers/gpu/drm/via/via_mm.c drm_mm_remove_node(&entry->mm_node); entry 235 drivers/gpu/drm/via/via_mm.c kfree(entry); entry 182 drivers/gpu/drm/virtio/virtgpu_vq.c struct virtio_gpu_vbuffer *entry, *tmp; entry 195 drivers/gpu/drm/virtio/virtgpu_vq.c list_for_each_entry_safe(entry, tmp, &reclaim_list, list) { entry 196 drivers/gpu/drm/virtio/virtgpu_vq.c resp = (struct virtio_gpu_ctrl_hdr *)entry->resp_buf; entry 203 drivers/gpu/drm/virtio/virtgpu_vq.c cmd = (struct virtio_gpu_ctrl_hdr *)entry->buf; entry 220 drivers/gpu/drm/virtio/virtgpu_vq.c if (entry->resp_cb) entry 221 drivers/gpu/drm/virtio/virtgpu_vq.c entry->resp_cb(vgdev, entry); entry 223 drivers/gpu/drm/virtio/virtgpu_vq.c list_del(&entry->list); entry 224 drivers/gpu/drm/virtio/virtgpu_vq.c free_vbuf(vgdev, entry); entry 238 drivers/gpu/drm/virtio/virtgpu_vq.c struct virtio_gpu_vbuffer *entry, *tmp; entry 248 drivers/gpu/drm/virtio/virtgpu_vq.c list_for_each_entry_safe(entry, tmp, &reclaim_list, list) { entry 249 drivers/gpu/drm/virtio/virtgpu_vq.c list_del(&entry->list); entry 250 drivers/gpu/drm/virtio/virtgpu_vq.c free_vbuf(vgdev, entry); entry 353 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry, *next; entry 356 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, &cbs->list, ctx_list) entry 357 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_drop(entry); entry 371 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry; entry 373 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry(entry, &cbs->list, ctx_list) { entry 374 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (!entry->scrubbed) { entry 375 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c (void) vmw_binding_infos[entry->bt].scrub_func entry 376 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c (entry, false); entry 377 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c entry->scrubbed = true; entry 395 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry, *next; entry 398 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, head, res_list) entry 399 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_drop(entry); entry 413 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry; entry 415 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry(entry, head, res_list) { entry 416 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (!entry->scrubbed) { entry 417 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c (void) vmw_binding_infos[entry->bt].scrub_func entry 418 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c (entry, false); entry 419 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c entry->scrubbed = true; entry 423 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry(entry, head, res_list) { entry 425 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_context_binding_state(entry->ctx); entry 447 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry, *next; entry 449 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, &from->list, ctx_list) { entry 450 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_transfer(to, from, entry); entry 451 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_drop(entry); entry 465 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry; entry 468 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry(entry, &cbs->list, ctx_list) { entry 469 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (likely(!entry->scrubbed)) entry 472 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if ((entry->res == NULL || entry->res->id == entry 476 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ret = vmw_binding_infos[entry->bt].scrub_func(entry, true); entry 480 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c entry->scrubbed = false; entry 1238 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_bindinfo *entry, *next; entry 1240 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, &cbs->list, ctx_list) entry 1241 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_drop(entry); entry 345 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *entry; entry 348 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c entry = list_first_entry(&ctx->submitted, entry 352 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c status = vmw_cmdbuf_header_submit(entry); entry 356 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c entry->cb_header->status = SVGA_CB_STATUS_NONE; entry 360 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_del(&entry->list); entry 361 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_add_tail(&entry->list, &ctx->hw_submitted); entry 381 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *entry, *next; entry 385 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_for_each_entry_safe(entry, next, &ctx->hw_submitted, list) { entry 386 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c SVGACBStatus status = entry->cb_header->status; entry 391 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_del(&entry->list); entry 396 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c __vmw_cmdbuf_header_free(entry); entry 400 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c entry->cb_header->status = SVGA_CB_STATUS_NONE; entry 401 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_add_tail(&entry->list, &man->error); entry 405 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c entry->cb_header->status = SVGA_CB_STATUS_NONE; entry 406 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_add_tail(&entry->list, &ctx->preempted); entry 410 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c __vmw_cmdbuf_header_free(entry); entry 414 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c __vmw_cmdbuf_header_free(entry); entry 516 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *entry, *next; entry 529 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_for_each_entry_safe(entry, next, &man->error, list) { entry 530 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c SVGACBHeader *cb_hdr = entry->cb_header; entry 532 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c (entry->cmd + cb_hdr->errorOffset); entry 536 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_del_init(&entry->list); entry 543 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c __vmw_cmdbuf_header_free(entry); entry 558 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c __vmw_cmdbuf_header_free(entry); entry 568 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c entry->cmd += new_start_offset; entry 573 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c list_add_tail(&entry->list, &restart_head[entry->cb_context]); entry 105 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c struct vmw_cmdbuf_res *entry) entry 107 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_del(&entry->head); entry 108 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c WARN_ON(drm_ht_remove_item(&man->resources, &entry->hash)); entry 109 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c vmw_resource_unreference(&entry->res); entry 110 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c kfree(entry); entry 125 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c struct vmw_cmdbuf_res *entry, *next; entry 127 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_for_each_entry_safe(entry, next, list, head) { entry 128 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_del(&entry->head); entry 129 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c if (entry->res->func->commit_notify) entry 130 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c entry->res->func->commit_notify(entry->res, entry 131 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c entry->state); entry 132 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c switch (entry->state) { entry 134 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c entry->state = VMW_CMDBUF_RES_COMMITTED; entry 135 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_add_tail(&entry->head, &entry->man->list); entry 138 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c vmw_resource_unreference(&entry->res); entry 139 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c kfree(entry); entry 162 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c struct vmw_cmdbuf_res *entry, *next; entry 165 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_for_each_entry_safe(entry, next, list, head) { entry 166 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c switch (entry->state) { entry 168 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c vmw_cmdbuf_res_free(entry->man, entry); entry 171 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c ret = drm_ht_insert_item(&entry->man->resources, entry 172 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c &entry->hash); entry 173 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_del(&entry->head); entry 174 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_add_tail(&entry->head, &entry->man->list); entry 175 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c entry->state = VMW_CMDBUF_RES_COMMITTED; entry 248 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c struct vmw_cmdbuf_res *entry; entry 257 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c entry = drm_hash_entry(hash, struct vmw_cmdbuf_res, hash); entry 259 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c switch (entry->state) { entry 261 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c vmw_cmdbuf_res_free(man, entry); entry 265 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c (void) drm_ht_remove_item(&man->resources, &entry->hash); entry 266 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_del(&entry->head); entry 267 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c entry->state = VMW_CMDBUF_RES_DEL; entry 268 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_add_tail(&entry->head, list); entry 269 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c *res_p = entry->res; entry 320 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c struct vmw_cmdbuf_res *entry, *next; entry 322 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c list_for_each_entry_safe(entry, next, &man->list, head) entry 323 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf_res.c vmw_cmdbuf_res_free(man, entry); entry 172 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c struct vmw_ctx_validation_info *entry; entry 174 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c list_for_each_entry(entry, &sw_context->ctx_list, head) { entry 176 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c vmw_binding_state_commit(entry->cur, entry->staged); entry 178 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (entry->staged != sw_context->staged_bindings) entry 179 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c vmw_binding_state_free(entry->staged); entry 458 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c struct vmw_ctx_bindinfo *entry; entry 481 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c list_for_each_entry(entry, binding_list, ctx_list) { entry 482 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (vmw_res_type(entry->res) == vmw_res_view) entry 483 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c ret = vmw_view_res_val_add(sw_context, entry->res); entry 486 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c (sw_context, entry->res, entry 487 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c vmw_binding_dirtying(entry->bt)); entry 3162 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c const struct vmw_cmd_entry *entry; entry 3170 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c entry = &vmw_cmd_entries[cmd_id]; entry 3171 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c *cmd = entry->cmd_name; entry 3209 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c const struct vmw_cmd_entry *entry; entry 3228 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c entry = &vmw_cmd_entries[cmd_id]; entry 3229 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (unlikely(!entry->func)) entry 3232 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (unlikely(!entry->user_allow && !sw_context->kernel)) entry 3235 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (unlikely(entry->gb_disable && gb)) entry 3238 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c if (unlikely(entry->gb_enable && !gb)) entry 3241 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c ret = entry->func(dev_priv, sw_context, header); entry 81 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c struct vmw_legacy_display_unit *entry; entry 91 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c list_for_each_entry(entry, &lds->active, active) { entry 92 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c crtc = &entry->base.crtc; entry 100 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c fb = entry->base.crtc.primary->state->fb; entry 108 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c entry = list_entry(lds->active.next, typeof(*entry), active); entry 109 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c fb = entry->base.crtc.primary->state->fb; entry 120 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c list_for_each_entry(entry, &lds->active, active) { entry 121 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c crtc = &entry->base.crtc; entry 165 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c struct vmw_legacy_display_unit *entry; entry 182 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c list_for_each_entry(entry, &ld->active, active) { entry 183 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c if (entry->base.unit > ldu->base.unit) entry 186 drivers/gpu/drm/vmwgfx/vmwgfx_ldu.c at = &entry->active; entry 550 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c struct vmw_dx_shader *entry, *next; entry 554 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c list_for_each_entry_safe(entry, next, list, cotable_head) { entry 555 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c WARN_ON(vmw_dx_shader_scrub(&entry->res)); entry 557 drivers/gpu/drm/vmwgfx/vmwgfx_shader.c entry->committed = false; entry 436 drivers/gpu/drm/vmwgfx/vmwgfx_so.c struct vmw_view *entry, *next; entry 440 drivers/gpu/drm/vmwgfx/vmwgfx_so.c list_for_each_entry_safe(entry, next, list, cotable_head) entry 441 drivers/gpu/drm/vmwgfx/vmwgfx_so.c WARN_ON(vmw_view_destroy(&entry->res)); entry 456 drivers/gpu/drm/vmwgfx/vmwgfx_so.c struct vmw_view *entry, *next; entry 460 drivers/gpu/drm/vmwgfx/vmwgfx_so.c list_for_each_entry_safe(entry, next, list, srf_head) entry 461 drivers/gpu/drm/vmwgfx/vmwgfx_so.c WARN_ON(vmw_view_destroy(&entry->res)); entry 149 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c struct page *entry, *next; entry 151 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_for_each_entry_safe(entry, next, &ctx->page_list, lru) { entry 152 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_del_init(&entry->lru); entry 153 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c __free_page(entry); entry 188 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c struct vmw_validation_bo_node *entry; entry 190 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_for_each_entry(entry, &ctx->bo_list, base.head) { entry 191 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (entry->base.bo == &vbo->base) { entry 192 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c bo_node = entry; entry 225 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c struct vmw_validation_res_node *entry; entry 227 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_for_each_entry(entry, &ctx->resource_ctx_list, head) { entry 228 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (entry->res == res) { entry 229 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c res_node = entry; entry 234 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_for_each_entry(entry, &ctx->resource_list, head) { entry 235 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (entry->res == res) { entry 236 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c res_node = entry; entry 561 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c struct vmw_validation_bo_node *entry; entry 564 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_for_each_entry(entry, &ctx->bo_list, base.head) { entry 565 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (entry->cpu_blit) { entry 571 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c ret = ttm_bo_validate(entry->base.bo, entry 575 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c (entry->base.bo, intr, entry->as_mob); entry 639 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c struct vmw_validation_bo_node *entry; entry 645 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_for_each_entry(entry, &ctx->bo_list, base.head) entry 646 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c (void) drm_ht_remove_item(ctx->ht, &entry->hash); entry 667 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c struct vmw_validation_bo_node *entry; entry 670 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c list_for_each_entry(entry, &ctx->bo_list, base.head) { entry 671 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c ttm_bo_put(entry->base.bo); entry 672 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c entry->base.bo = NULL; entry 174 drivers/gpu/host1x/dev.c const struct host1x_sid_entry *entry = &info->sid_table[i]; entry 176 drivers/gpu/host1x/dev.c host1x_hypervisor_writel(host, entry->offset, entry->base); entry 177 drivers/gpu/host1x/dev.c host1x_hypervisor_writel(host, entry->limit, entry->base + 4); entry 303 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry = drv_data->device_props; entry 306 drivers/hid/hid-lg4ff.c if (!entry) { entry 311 drivers/hid/hid-lg4ff.c switch (entry->wdata.product_id) { entry 315 drivers/hid/hid-lg4ff.c new_value = lg4ff_adjust_dfp_x_axis(value, entry->wdata.range); entry 330 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry = drv_data->device_props; entry 332 drivers/hid/hid-lg4ff.c if (!entry) entry 336 drivers/hid/hid-lg4ff.c if (entry->wdata.combine) { entry 337 drivers/hid/hid-lg4ff.c switch (entry->wdata.product_id) { entry 409 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 421 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 422 drivers/hid/hid-lg4ff.c if (!entry) { entry 426 drivers/hid/hid-lg4ff.c value = entry->report->field[0]->value; entry 435 drivers/hid/hid-lg4ff.c spin_lock_irqsave(&entry->report_lock, flags); entry 446 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 447 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 459 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 460 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 473 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 483 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 484 drivers/hid/hid-lg4ff.c if (!entry) { entry 488 drivers/hid/hid-lg4ff.c value = entry->report->field[0]->value; entry 491 drivers/hid/hid-lg4ff.c spin_lock_irqsave(&entry->report_lock, flags); entry 501 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 502 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 515 drivers/hid/hid-lg4ff.c switch (entry->wdata.product_id) { entry 532 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 543 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 544 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 551 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 563 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 564 drivers/hid/hid-lg4ff.c if (!entry) { entry 568 drivers/hid/hid-lg4ff.c value = entry->report->field[0]->value; entry 570 drivers/hid/hid-lg4ff.c spin_lock_irqsave(&entry->report_lock, flags); entry 579 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 580 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 586 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 597 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 598 drivers/hid/hid-lg4ff.c if (!entry) { entry 602 drivers/hid/hid-lg4ff.c value = entry->report->field[0]->value; entry 605 drivers/hid/hid-lg4ff.c spin_lock_irqsave(&entry->report_lock, flags); entry 614 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 615 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 621 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 633 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 634 drivers/hid/hid-lg4ff.c if (!entry) { entry 638 drivers/hid/hid-lg4ff.c value = entry->report->field[0]->value; entry 642 drivers/hid/hid-lg4ff.c spin_lock_irqsave(&entry->report_lock, flags); entry 658 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 670 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 671 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 685 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 686 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 765 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 777 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 778 drivers/hid/hid-lg4ff.c if (!entry) { entry 782 drivers/hid/hid-lg4ff.c value = entry->report->field[0]->value; entry 784 drivers/hid/hid-lg4ff.c spin_lock_irqsave(&entry->report_lock, flags); entry 791 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 793 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 801 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 812 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 813 drivers/hid/hid-lg4ff.c if (!entry) { entry 818 drivers/hid/hid-lg4ff.c if (!entry->wdata.real_name) { entry 824 drivers/hid/hid-lg4ff.c if (entry->wdata.alternate_modes & BIT(i)) { entry 828 drivers/hid/hid-lg4ff.c !lg4ff_alternate_modes[i].product_id ? entry->wdata.real_name : lg4ff_alternate_modes[i].name); entry 833 drivers/hid/hid-lg4ff.c if (lg4ff_alternate_modes[i].product_id == entry->wdata.product_id || entry 834 drivers/hid/hid-lg4ff.c (lg4ff_alternate_modes[i].product_id == 0 && entry->wdata.product_id == entry->wdata.real_product_id)) entry 850 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 863 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 864 drivers/hid/hid-lg4ff.c if (!entry) { entry 887 drivers/hid/hid-lg4ff.c if (entry->wdata.alternate_modes & BIT(i)) { entry 890 drivers/hid/hid-lg4ff.c target_product_id = entry->wdata.real_product_id; entry 905 drivers/hid/hid-lg4ff.c if (target_product_id == entry->wdata.product_id) /* Nothing to do */ entry 911 drivers/hid/hid-lg4ff.c entry->wdata.real_name); entry 916 drivers/hid/hid-lg4ff.c if ((entry->wdata.real_product_id == USB_DEVICE_ID_LOGITECH_DFP_WHEEL || entry->wdata.real_product_id == USB_DEVICE_ID_LOGITECH_G25_WHEEL) && entry 917 drivers/hid/hid-lg4ff.c entry->wdata.product_id > target_product_id) { entry 918 drivers/hid/hid-lg4ff.c hid_info(hid, "\"%s\" cannot be switched back into \"%s\" mode\n", entry->wdata.real_name, lg4ff_alternate_modes[i].name); entry 922 drivers/hid/hid-lg4ff.c s = lg4ff_get_mode_switch_command(entry->wdata.real_product_id, target_product_id); entry 937 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 947 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 948 drivers/hid/hid-lg4ff.c if (!entry) { entry 953 drivers/hid/hid-lg4ff.c count = scnprintf(buf, PAGE_SIZE, "%u\n", entry->wdata.combine); entry 961 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 971 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 972 drivers/hid/hid-lg4ff.c if (!entry) { entry 980 drivers/hid/hid-lg4ff.c entry->wdata.combine = combine; entry 990 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 1000 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 1001 drivers/hid/hid-lg4ff.c if (!entry) { entry 1006 drivers/hid/hid-lg4ff.c count = scnprintf(buf, PAGE_SIZE, "%u\n", entry->wdata.range); entry 1016 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 1026 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 1027 drivers/hid/hid-lg4ff.c if (!entry) { entry 1033 drivers/hid/hid-lg4ff.c range = entry->wdata.max_range; entry 1037 drivers/hid/hid-lg4ff.c if (entry->wdata.set_range && range >= entry->wdata.min_range && range <= entry->wdata.max_range) { entry 1038 drivers/hid/hid-lg4ff.c entry->wdata.set_range(hid, range); entry 1039 drivers/hid/hid-lg4ff.c entry->wdata.range = range; entry 1049 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 1059 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 1060 drivers/hid/hid-lg4ff.c if (!entry) { entry 1065 drivers/hid/hid-lg4ff.c if (!entry->wdata.real_tag || !entry->wdata.real_name) { entry 1070 drivers/hid/hid-lg4ff.c count = scnprintf(buf, PAGE_SIZE, "%s: %s\n", entry->wdata.real_tag, entry->wdata.real_name); entry 1085 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 1095 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 1096 drivers/hid/hid-lg4ff.c if (!entry) { entry 1100 drivers/hid/hid-lg4ff.c value = entry->report->field[0]->value; entry 1102 drivers/hid/hid-lg4ff.c spin_lock_irqsave(&entry->report_lock, flags); entry 1110 drivers/hid/hid-lg4ff.c hid_hw_request(hid, entry->report, HID_REQ_SET_REPORT); entry 1111 drivers/hid/hid-lg4ff.c spin_unlock_irqrestore(&entry->report_lock, flags); entry 1120 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 1128 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 1130 drivers/hid/hid-lg4ff.c if (!entry) { entry 1136 drivers/hid/hid-lg4ff.c if (led_cdev != entry->wdata.led[i]) entry 1138 drivers/hid/hid-lg4ff.c state = (entry->wdata.led_state >> i) & 1; entry 1140 drivers/hid/hid-lg4ff.c entry->wdata.led_state &= ~(1 << i); entry 1141 drivers/hid/hid-lg4ff.c lg4ff_set_leds(hid, entry->wdata.led_state); entry 1143 drivers/hid/hid-lg4ff.c entry->wdata.led_state |= 1 << i; entry 1144 drivers/hid/hid-lg4ff.c lg4ff_set_leds(hid, entry->wdata.led_state); entry 1155 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 1163 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 1165 drivers/hid/hid-lg4ff.c if (!entry) { entry 1171 drivers/hid/hid-lg4ff.c if (led_cdev == entry->wdata.led[i]) { entry 1172 drivers/hid/hid-lg4ff.c value = (entry->wdata.led_state >> i) & 1; entry 1263 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 1285 drivers/hid/hid-lg4ff.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1286 drivers/hid/hid-lg4ff.c if (!entry) entry 1288 drivers/hid/hid-lg4ff.c spin_lock_init(&entry->report_lock); entry 1289 drivers/hid/hid-lg4ff.c entry->report = report; entry 1290 drivers/hid/hid-lg4ff.c drv_data->device_props = entry; entry 1350 drivers/hid/hid-lg4ff.c lg4ff_init_wheel_data(&entry->wdata, &lg4ff_devices[i], mmode_wheel, real_product_id); entry 1383 drivers/hid/hid-lg4ff.c entry->wdata.range = entry->wdata.max_range; entry 1384 drivers/hid/hid-lg4ff.c if (entry->wdata.set_range) entry 1385 drivers/hid/hid-lg4ff.c entry->wdata.set_range(hid, entry->wdata.range); entry 1389 drivers/hid/hid-lg4ff.c entry->wdata.led_state = 0; entry 1391 drivers/hid/hid-lg4ff.c entry->wdata.led[j] = NULL; entry 1418 drivers/hid/hid-lg4ff.c entry->wdata.led[j] = led; entry 1426 drivers/hid/hid-lg4ff.c led = entry->wdata.led[j]; entry 1427 drivers/hid/hid-lg4ff.c entry->wdata.led[j] = NULL; entry 1444 drivers/hid/hid-lg4ff.c kfree(entry); entry 1450 drivers/hid/hid-lg4ff.c struct lg4ff_device_entry *entry; entry 1458 drivers/hid/hid-lg4ff.c entry = drv_data->device_props; entry 1459 drivers/hid/hid-lg4ff.c if (!entry) entry 1463 drivers/hid/hid-lg4ff.c if (entry->wdata.alternate_modes) { entry 1478 drivers/hid/hid-lg4ff.c led = entry->wdata.led[j]; entry 1479 drivers/hid/hid-lg4ff.c entry->wdata.led[j] = NULL; entry 1489 drivers/hid/hid-lg4ff.c kfree(entry); entry 2396 drivers/hid/hid-sony.c struct sony_sc *entry; entry 2402 drivers/hid/hid-sony.c list_for_each_entry(entry, &sony_device_list, list_node) { entry 2403 drivers/hid/hid-sony.c ret = memcmp(sc->mac_address, entry->mac_address, entry 2406 drivers/hid/hid-sony.c if (sony_compare_connection_type(sc, entry)) { entry 133 drivers/hsi/clients/cmt_speech.c struct char_queue *entry; entry 142 drivers/hsi/clients/cmt_speech.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 143 drivers/hsi/clients/cmt_speech.c if (!entry) { entry 150 drivers/hsi/clients/cmt_speech.c entry->msg = message; entry 151 drivers/hsi/clients/cmt_speech.c list_add_tail(&entry->list, head); entry 164 drivers/hsi/clients/cmt_speech.c struct char_queue *entry; entry 167 drivers/hsi/clients/cmt_speech.c entry = list_entry(head->next, struct char_queue, list); entry 168 drivers/hsi/clients/cmt_speech.c data = entry->msg; entry 169 drivers/hsi/clients/cmt_speech.c list_del(&entry->list); entry 170 drivers/hsi/clients/cmt_speech.c kfree(entry); entry 1322 drivers/hsi/clients/cmt_speech.c struct char_queue *entry; entry 1327 drivers/hsi/clients/cmt_speech.c entry = list_entry(cursor, struct char_queue, list); entry 1328 drivers/hsi/clients/cmt_speech.c list_del(&entry->list); entry 1329 drivers/hsi/clients/cmt_speech.c kfree(entry); entry 1058 drivers/hv/vmbus_drv.c const struct vmbus_channel_message_table_entry *entry; entry 1075 drivers/hv/vmbus_drv.c entry = &channel_message_table[hdr->msgtype]; entry 1076 drivers/hv/vmbus_drv.c if (entry->handler_type == VMHT_BLOCKING) { entry 1111 drivers/hv/vmbus_drv.c entry->message_handler(hdr); entry 301 drivers/hwmon/applesmc.c static int applesmc_read_entry(const struct applesmc_entry *entry, entry 306 drivers/hwmon/applesmc.c if (entry->len != len) entry 309 drivers/hwmon/applesmc.c ret = read_smc(APPLESMC_READ_CMD, entry->key, buf, len); entry 315 drivers/hwmon/applesmc.c static int applesmc_write_entry(const struct applesmc_entry *entry, entry 320 drivers/hwmon/applesmc.c if (entry->len != len) entry 323 drivers/hwmon/applesmc.c ret = write_smc(APPLESMC_WRITE_CMD, entry->key, buf, len); entry 366 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 370 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_index(middle); entry 371 drivers/hwmon/applesmc.c if (IS_ERR(entry)) { entry 373 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 375 drivers/hwmon/applesmc.c if (strcmp(entry->key, key) < 0) entry 388 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 392 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_index(middle); entry 393 drivers/hwmon/applesmc.c if (IS_ERR(entry)) { entry 395 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 397 drivers/hwmon/applesmc.c if (strcmp(key, entry->key) < 0) entry 426 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 428 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_key(key); entry 429 drivers/hwmon/applesmc.c if (IS_ERR(entry)) entry 430 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 432 drivers/hwmon/applesmc.c return applesmc_read_entry(entry, buffer, len); entry 437 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 439 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_key(key); entry 440 drivers/hwmon/applesmc.c if (IS_ERR(entry)) entry 441 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 443 drivers/hwmon/applesmc.c return applesmc_write_entry(entry, buffer, len); entry 448 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 450 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_key(key); entry 451 drivers/hwmon/applesmc.c if (IS_ERR(entry) && PTR_ERR(entry) != -EINVAL) entry 452 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 454 drivers/hwmon/applesmc.c *value = !IS_ERR(entry); entry 500 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 511 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_index(i); entry 512 drivers/hwmon/applesmc.c if (IS_ERR(entry)) entry 514 drivers/hwmon/applesmc.c if (strcmp(entry->type, TEMP_SENSOR_TYPE)) entry 516 drivers/hwmon/applesmc.c s->index[s->index_count++] = entry->key; entry 734 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 741 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_key(LIGHT_SENSOR_LEFT_KEY); entry 742 drivers/hwmon/applesmc.c if (IS_ERR(entry)) entry 743 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 744 drivers/hwmon/applesmc.c if (entry->len > 10) entry 746 drivers/hwmon/applesmc.c data_length = entry->len; entry 961 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 964 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_index(key_at_index); entry 965 drivers/hwmon/applesmc.c if (IS_ERR(entry)) entry 966 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 967 drivers/hwmon/applesmc.c ret = applesmc_read_entry(entry, sysfsbuf, entry->len); entry 971 drivers/hwmon/applesmc.c return entry->len; entry 977 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 979 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_index(key_at_index); entry 980 drivers/hwmon/applesmc.c if (IS_ERR(entry)) entry 981 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 983 drivers/hwmon/applesmc.c return snprintf(sysfsbuf, PAGE_SIZE, "%d\n", entry->len); entry 989 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 991 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_index(key_at_index); entry 992 drivers/hwmon/applesmc.c if (IS_ERR(entry)) entry 993 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 995 drivers/hwmon/applesmc.c return snprintf(sysfsbuf, PAGE_SIZE, "%s\n", entry->type); entry 1001 drivers/hwmon/applesmc.c const struct applesmc_entry *entry; entry 1003 drivers/hwmon/applesmc.c entry = applesmc_get_entry_by_index(key_at_index); entry 1004 drivers/hwmon/applesmc.c if (IS_ERR(entry)) entry 1005 drivers/hwmon/applesmc.c return PTR_ERR(entry); entry 1007 drivers/hwmon/applesmc.c return snprintf(sysfsbuf, PAGE_SIZE, "%s\n", entry->key); entry 324 drivers/hwmon/k10temp.c const struct tctl_offset *entry = &tctl_offset_table[i]; entry 326 drivers/hwmon/k10temp.c if (boot_cpu_data.x86 == entry->model && entry 327 drivers/hwmon/k10temp.c strstr(boot_cpu_data.x86_model_id, entry->id)) { entry 328 drivers/hwmon/k10temp.c data->temp_offset = entry->offset; entry 2256 drivers/hwmon/pmbus/pmbus_core.c struct pmbus_debugfs_entry *entry = data; entry 2258 drivers/hwmon/pmbus/pmbus_core.c rc = _pmbus_read_byte_data(entry->client, entry->page, entry->reg); entry 2272 drivers/hwmon/pmbus/pmbus_core.c struct pmbus_debugfs_entry *entry = data; entry 2273 drivers/hwmon/pmbus/pmbus_core.c struct pmbus_data *pdata = i2c_get_clientdata(entry->client); entry 2275 drivers/hwmon/pmbus/pmbus_core.c rc = pdata->read_status(entry->client, entry->page); entry 372 drivers/hwmon/pmbus/ucd9000.c struct ucd9000_debugfs_entry *entry = data; entry 373 drivers/hwmon/pmbus/ucd9000.c struct i2c_client *client = entry->client; entry 385 drivers/hwmon/pmbus/ucd9000.c *val = !!(buffer[1] & BIT(entry->index)); entry 101 drivers/hwtracing/coresight/coresight-catu.c #define CATU_ENTRY_ADDR(entry) ((cate_t)(entry) & ~((cate_t)CATU_ENTRY_VALID)) entry 416 drivers/hwtracing/coresight/coresight-etm-perf.c list_for_each_entry(filter, filters, entry) { entry 459 drivers/hwtracing/coresight/coresight-etm-perf.c list_for_each_entry(filter, &head->list, entry) { entry 487 drivers/hwtracing/coresight/coresight-etm-perf.c char entry[sizeof("cpu9999999")]; entry 492 drivers/hwtracing/coresight/coresight-etm-perf.c sprintf(entry, "cpu%d", cpu); entry 498 drivers/hwtracing/coresight/coresight-etm-perf.c ret = sysfs_create_link(&pmu_dev->kobj, &cs_dev->kobj, entry); entry 503 drivers/hwtracing/coresight/coresight-etm-perf.c sysfs_remove_link(&pmu_dev->kobj, entry); entry 96 drivers/hwtracing/coresight/coresight-tmc-etr.c #define ETR_SG_ADDR(entry) \ entry 97 drivers/hwtracing/coresight/coresight-tmc-etr.c (((dma_addr_t)(entry) >> ETR_SG_ADDR_SHIFT) << ETR_SG_PAGE_SHIFT) entry 98 drivers/hwtracing/coresight/coresight-tmc-etr.c #define ETR_SG_ET(entry) ((entry) & ETR_SG_ET_MASK) entry 68 drivers/hwtracing/intel_th/msu.c struct list_head entry; entry 93 drivers/hwtracing/intel_th/msu.c struct list_head entry; entry 176 drivers/hwtracing/intel_th/msu.c struct list_head entry; entry 187 drivers/hwtracing/intel_th/msu.c list_for_each_entry(mbe, &msu_buffer_list, entry) { entry 239 drivers/hwtracing/intel_th/msu.c list_add_tail(&mbe->entry, &msu_buffer_list); entry 254 drivers/hwtracing/intel_th/msu.c list_del(&mbe->entry); entry 302 drivers/hwtracing/intel_th/msu.c return win->entry.next == &win->msc->win_list; entry 315 drivers/hwtracing/intel_th/msu.c entry); entry 317 drivers/hwtracing/intel_th/msu.c return list_next_entry(win, entry); entry 363 drivers/hwtracing/intel_th/msu.c list_for_each_entry(win, &msc->win_list, entry) { entry 398 drivers/hwtracing/intel_th/msu.c return list_first_entry(&msc->win_list, struct msc_window, entry); entry 460 drivers/hwtracing/intel_th/msu.c list_add_tail(&iter->entry, &msc->iter_list); entry 470 drivers/hwtracing/intel_th/msu.c list_del(&iter->entry); entry 655 drivers/hwtracing/intel_th/msu.c list_for_each_entry(win, &msc->win_list, entry) { entry 1082 drivers/hwtracing/intel_th/msu.c entry); entry 1107 drivers/hwtracing/intel_th/msu.c list_add_tail(&win->entry, &msc->win_list); entry 1145 drivers/hwtracing/intel_th/msu.c list_del(&win->entry); entry 1173 drivers/hwtracing/intel_th/msu.c list_for_each_entry(win, &msc->win_list, entry) { entry 1185 drivers/hwtracing/intel_th/msu.c struct msc_window, entry); entry 1187 drivers/hwtracing/intel_th/msu.c next_win = list_next_entry(win, entry); entry 1226 drivers/hwtracing/intel_th/msu.c list_for_each_entry_safe(win, iter, &msc->win_list, entry) entry 1376 drivers/hwtracing/intel_th/msu.c list_for_each_entry(win, &msc->win_list, entry) entry 1687 drivers/hwtracing/intel_th/msu.c first = list_first_entry(&msc->win_list, struct msc_window, entry); entry 1692 drivers/hwtracing/intel_th/msu.c msc->cur_win = list_next_entry(msc->cur_win, entry); entry 1933 drivers/hwtracing/intel_th/msu.c list_for_each_entry(win, &msc->win_list, entry) { entry 373 drivers/hwtracing/stm/core.c struct list_head entry; entry 389 drivers/hwtracing/stm/core.c list_for_each_entry(pe, &stm_pdrv_head, entry) { entry 419 drivers/hwtracing/stm/core.c list_add_tail(&pe->entry, &stm_pdrv_head); entry 439 drivers/hwtracing/stm/core.c list_for_each_entry_safe(pe, iter, &stm_pdrv_head, entry) { entry 441 drivers/hwtracing/stm/core.c list_del(&pe->entry); entry 59 drivers/infiniband/core/agent.c struct ib_agent_port_private *entry; entry 61 drivers/infiniband/core/agent.c list_for_each_entry(entry, &ib_agent_port_list, port_list) { entry 62 drivers/infiniband/core/agent.c if (entry->agent[1]->device == device && entry 63 drivers/infiniband/core/agent.c entry->agent[1]->port_num == port_num) entry 64 drivers/infiniband/core/agent.c return entry; entry 72 drivers/infiniband/core/agent.c struct ib_agent_port_private *entry; entry 76 drivers/infiniband/core/agent.c entry = __ib_get_agent_port(device, port_num); entry 78 drivers/infiniband/core/agent.c return entry; entry 201 drivers/infiniband/core/cache.c static bool is_gid_entry_free(const struct ib_gid_table_entry *entry) entry 203 drivers/infiniband/core/cache.c return !entry; entry 206 drivers/infiniband/core/cache.c static bool is_gid_entry_valid(const struct ib_gid_table_entry *entry) entry 208 drivers/infiniband/core/cache.c return entry && entry->state == GID_TABLE_ENTRY_VALID; entry 213 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry = entry 216 drivers/infiniband/core/cache.c queue_work(ib_wq, &entry->del_work); entry 233 drivers/infiniband/core/cache.c static void free_gid_entry_locked(struct ib_gid_table_entry *entry) entry 235 drivers/infiniband/core/cache.c struct ib_device *device = entry->attr.device; entry 236 drivers/infiniband/core/cache.c u8 port_num = entry->attr.port_num; entry 240 drivers/infiniband/core/cache.c port_num, entry->attr.index, entry->attr.gid.raw); entry 250 drivers/infiniband/core/cache.c if (entry == table->data_vec[entry->attr.index]) entry 251 drivers/infiniband/core/cache.c table->data_vec[entry->attr.index] = NULL; entry 255 drivers/infiniband/core/cache.c if (entry->ndev_storage) entry 256 drivers/infiniband/core/cache.c call_rcu(&entry->ndev_storage->rcu_head, put_gid_ndev); entry 257 drivers/infiniband/core/cache.c kfree(entry); entry 262 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry = entry 265 drivers/infiniband/core/cache.c free_gid_entry_locked(entry); entry 278 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry = entry 280 drivers/infiniband/core/cache.c struct ib_device *device = entry->attr.device; entry 281 drivers/infiniband/core/cache.c u8 port_num = entry->attr.port_num; entry 285 drivers/infiniband/core/cache.c free_gid_entry_locked(entry); entry 292 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry; entry 295 drivers/infiniband/core/cache.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 296 drivers/infiniband/core/cache.c if (!entry) entry 301 drivers/infiniband/core/cache.c entry->ndev_storage = kzalloc(sizeof(*entry->ndev_storage), entry 303 drivers/infiniband/core/cache.c if (!entry->ndev_storage) { entry 304 drivers/infiniband/core/cache.c kfree(entry); entry 308 drivers/infiniband/core/cache.c entry->ndev_storage->ndev = ndev; entry 310 drivers/infiniband/core/cache.c kref_init(&entry->kref); entry 311 drivers/infiniband/core/cache.c memcpy(&entry->attr, attr, sizeof(*attr)); entry 312 drivers/infiniband/core/cache.c INIT_WORK(&entry->del_work, free_gid_work); entry 313 drivers/infiniband/core/cache.c entry->state = GID_TABLE_ENTRY_INVALID; entry 314 drivers/infiniband/core/cache.c return entry; entry 318 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry) entry 320 drivers/infiniband/core/cache.c entry->state = GID_TABLE_ENTRY_VALID; entry 322 drivers/infiniband/core/cache.c dev_dbg(&entry->attr.device->dev, "%s port=%d index=%d gid %pI6\n", entry 323 drivers/infiniband/core/cache.c __func__, entry->attr.port_num, entry->attr.index, entry 324 drivers/infiniband/core/cache.c entry->attr.gid.raw); entry 328 drivers/infiniband/core/cache.c table->data_vec[entry->attr.index] = entry; entry 332 drivers/infiniband/core/cache.c static void get_gid_entry(struct ib_gid_table_entry *entry) entry 334 drivers/infiniband/core/cache.c kref_get(&entry->kref); entry 337 drivers/infiniband/core/cache.c static void put_gid_entry(struct ib_gid_table_entry *entry) entry 339 drivers/infiniband/core/cache.c kref_put(&entry->kref, schedule_free_gid); entry 342 drivers/infiniband/core/cache.c static void put_gid_entry_locked(struct ib_gid_table_entry *entry) entry 344 drivers/infiniband/core/cache.c kref_put(&entry->kref, free_gid_entry); entry 347 drivers/infiniband/core/cache.c static int add_roce_gid(struct ib_gid_table_entry *entry) entry 349 drivers/infiniband/core/cache.c const struct ib_gid_attr *attr = &entry->attr; entry 358 drivers/infiniband/core/cache.c ret = attr->device->ops.add_gid(attr, &entry->context); entry 382 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry; entry 390 drivers/infiniband/core/cache.c entry = table->data_vec[ix]; entry 391 drivers/infiniband/core/cache.c entry->state = GID_TABLE_ENTRY_PENDING_DEL; entry 399 drivers/infiniband/core/cache.c ndev_storage = entry->ndev_storage; entry 401 drivers/infiniband/core/cache.c entry->ndev_storage = NULL; entry 402 drivers/infiniband/core/cache.c rcu_assign_pointer(entry->attr.ndev, NULL); entry 407 drivers/infiniband/core/cache.c ib_dev->ops.del_gid(&entry->attr, &entry->context); entry 409 drivers/infiniband/core/cache.c put_gid_entry_locked(entry); entry 425 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry; entry 443 drivers/infiniband/core/cache.c entry = alloc_gid_entry(attr); entry 444 drivers/infiniband/core/cache.c if (!entry) entry 448 drivers/infiniband/core/cache.c ret = add_roce_gid(entry); entry 453 drivers/infiniband/core/cache.c store_gid_entry(table, entry); entry 457 drivers/infiniband/core/cache.c put_gid_entry(entry); entry 750 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry = table->data_vec[i]; entry 752 drivers/infiniband/core/cache.c if (!is_gid_entry_valid(entry)) entry 755 drivers/infiniband/core/cache.c if (memcmp(gid, &entry->attr.gid, sizeof(*gid))) entry 758 drivers/infiniband/core/cache.c if (filter(gid, &entry->attr, context)) { entry 759 drivers/infiniband/core/cache.c get_gid_entry(entry); entry 760 drivers/infiniband/core/cache.c res = &entry->attr; entry 1238 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry = entry 1241 drivers/infiniband/core/cache.c put_gid_entry(entry); entry 1257 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry = entry 1260 drivers/infiniband/core/cache.c get_gid_entry(entry); entry 1278 drivers/infiniband/core/cache.c struct ib_gid_table_entry *entry = entry 1280 drivers/infiniband/core/cache.c struct ib_device *device = entry->attr.device; entry 1282 drivers/infiniband/core/cache.c u8 port_num = entry->attr.port_num; entry 161 drivers/infiniband/core/device.c void *entry; entry 165 drivers/infiniband/core/device.c entry = xas_find_marked(&xas, ULONG_MAX, filter); entry 166 drivers/infiniband/core/device.c if (xa_is_zero(entry)) entry 168 drivers/infiniband/core/device.c } while (xas_retry(&xas, entry)); entry 171 drivers/infiniband/core/device.c if (entry) { entry 173 drivers/infiniband/core/device.c if (xa_is_zero(entry)) entry 175 drivers/infiniband/core/device.c return entry; entry 179 drivers/infiniband/core/device.c #define xan_for_each_marked(xa, index, entry, filter) \ entry 180 drivers/infiniband/core/device.c for (index = 0, entry = xan_find_marked(xa, &(index), filter); \ entry 181 drivers/infiniband/core/device.c !xa_is_err(entry); \ entry 182 drivers/infiniband/core/device.c (index)++, entry = xan_find_marked(xa, &(index), filter)) entry 60 drivers/infiniband/core/mad.c struct trace_event_raw_ib_mad_send_template *entry) entry 71 drivers/infiniband/core/mad.c entry->sl = attr.sl; entry 73 drivers/infiniband/core/mad.c entry->pkey = pkey; entry 74 drivers/infiniband/core/mad.c entry->rqpn = wr->remote_qpn; entry 75 drivers/infiniband/core/mad.c entry->rqkey = wr->remote_qkey; entry 76 drivers/infiniband/core/mad.c entry->dlid = rdma_ah_get_dlid(&attr); entry 124 drivers/infiniband/core/mad.c struct ib_mad_port_private *entry; entry 126 drivers/infiniband/core/mad.c list_for_each_entry(entry, &ib_mad_port_list, port_list) { entry 127 drivers/infiniband/core/mad.c if (entry->device == device && entry->port_num == port_num) entry 128 drivers/infiniband/core/mad.c return entry; entry 140 drivers/infiniband/core/mad.c struct ib_mad_port_private *entry; entry 144 drivers/infiniband/core/mad.c entry = __ib_get_mad_port(device, port_num); entry 147 drivers/infiniband/core/mad.c return entry; entry 1129 drivers/infiniband/core/nldev.c u32 entry; entry 1142 drivers/infiniband/core/nldev.c .entry = RDMA_NLDEV_ATTR_RES_QP_ENTRY, entry 1149 drivers/infiniband/core/nldev.c .entry = RDMA_NLDEV_ATTR_RES_CM_ID_ENTRY, entry 1157 drivers/infiniband/core/nldev.c .entry = RDMA_NLDEV_ATTR_RES_CQ_ENTRY, entry 1165 drivers/infiniband/core/nldev.c .entry = RDMA_NLDEV_ATTR_RES_MR_ENTRY, entry 1173 drivers/infiniband/core/nldev.c .entry = RDMA_NLDEV_ATTR_RES_PD_ENTRY, entry 1180 drivers/infiniband/core/nldev.c .entry = RDMA_NLDEV_ATTR_STAT_COUNTER_ENTRY, entry 1351 drivers/infiniband/core/nldev.c entry_attr = nla_nest_start_noflag(skb, fe->entry); entry 726 drivers/infiniband/core/rdma_core.c struct ib_uobject *entry; entry 737 drivers/infiniband/core/rdma_core.c xa_for_each(&ufile->idr, id, entry) { entry 738 drivers/infiniband/core/rdma_core.c WARN_ON(entry->object); entry 739 drivers/infiniband/core/rdma_core.c uverbs_uobject_put(entry); entry 354 drivers/infiniband/core/roce_gid_mgmt.c struct sin_list *entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 356 drivers/infiniband/core/roce_gid_mgmt.c if (!entry) entry 359 drivers/infiniband/core/roce_gid_mgmt.c entry->ip.sin_family = AF_INET; entry 360 drivers/infiniband/core/roce_gid_mgmt.c entry->ip.sin_addr.s_addr = ifa->ifa_address; entry 361 drivers/infiniband/core/roce_gid_mgmt.c list_add_tail(&entry->list, &sin_list); entry 397 drivers/infiniband/core/roce_gid_mgmt.c struct sin6_list *entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 399 drivers/infiniband/core/roce_gid_mgmt.c if (!entry) entry 402 drivers/infiniband/core/roce_gid_mgmt.c entry->sin6.sin6_family = AF_INET6; entry 403 drivers/infiniband/core/roce_gid_mgmt.c entry->sin6.sin6_addr = ifp->addr; entry 404 drivers/infiniband/core/roce_gid_mgmt.c list_add_tail(&entry->list, &sin6_list); entry 536 drivers/infiniband/core/roce_gid_mgmt.c struct upper_list *entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 539 drivers/infiniband/core/roce_gid_mgmt.c if (!entry) entry 542 drivers/infiniband/core/roce_gid_mgmt.c list_add_tail(&entry->list, upper_list); entry 544 drivers/infiniband/core/roce_gid_mgmt.c entry->upper = upper; entry 1151 drivers/infiniband/core/sysfs.c list_add_tail(&p->kobj.entry, &coredev->port_list); entry 1312 drivers/infiniband/core/sysfs.c list_for_each_entry_safe(p, t, &coredev->port_list, entry) { entry 1315 drivers/infiniband/core/sysfs.c list_del(&p->entry); entry 1401 drivers/infiniband/core/sysfs.c list_for_each_entry_safe(p, t, &device->coredev.port_list, entry) { entry 481 drivers/infiniband/core/uverbs_cmd.c struct xrcd_table_entry *entry, *scan; entry 485 drivers/infiniband/core/uverbs_cmd.c entry = kmalloc(sizeof *entry, GFP_KERNEL); entry 486 drivers/infiniband/core/uverbs_cmd.c if (!entry) entry 489 drivers/infiniband/core/uverbs_cmd.c entry->xrcd = xrcd; entry 490 drivers/infiniband/core/uverbs_cmd.c entry->inode = inode; entry 501 drivers/infiniband/core/uverbs_cmd.c kfree(entry); entry 506 drivers/infiniband/core/uverbs_cmd.c rb_link_node(&entry->node, parent, p); entry 507 drivers/infiniband/core/uverbs_cmd.c rb_insert_color(&entry->node, &dev->xrcd_tree); entry 515 drivers/infiniband/core/uverbs_cmd.c struct xrcd_table_entry *entry; entry 519 drivers/infiniband/core/uverbs_cmd.c entry = rb_entry(p, struct xrcd_table_entry, node); entry 521 drivers/infiniband/core/uverbs_cmd.c if (inode < entry->inode) entry 523 drivers/infiniband/core/uverbs_cmd.c else if (inode > entry->inode) entry 526 drivers/infiniband/core/uverbs_cmd.c return entry; entry 534 drivers/infiniband/core/uverbs_cmd.c struct xrcd_table_entry *entry; entry 536 drivers/infiniband/core/uverbs_cmd.c entry = xrcd_table_search(dev, inode); entry 537 drivers/infiniband/core/uverbs_cmd.c if (!entry) entry 540 drivers/infiniband/core/uverbs_cmd.c return entry->xrcd; entry 546 drivers/infiniband/core/uverbs_cmd.c struct xrcd_table_entry *entry; entry 548 drivers/infiniband/core/uverbs_cmd.c entry = xrcd_table_search(dev, inode); entry 549 drivers/infiniband/core/uverbs_cmd.c if (entry) { entry 551 drivers/infiniband/core/uverbs_cmd.c rb_erase(&entry->node, &dev->xrcd_tree); entry 552 drivers/infiniband/core/uverbs_cmd.c kfree(entry); entry 372 drivers/infiniband/core/uverbs_ioctl.c void *entry; entry 375 drivers/infiniband/core/uverbs_ioctl.c entry = rcu_dereference_raw(*slot); entry 376 drivers/infiniband/core/uverbs_ioctl.c if (likely(!radix_tree_is_internal_node(entry) && entry)) entry 353 drivers/infiniband/core/uverbs_main.c struct ib_uverbs_event *entry, *tmp; entry 360 drivers/infiniband/core/uverbs_main.c list_for_each_entry_safe(entry, tmp, &file->ev_queue.event_list, list) { entry 361 drivers/infiniband/core/uverbs_main.c if (entry->counter) entry 362 drivers/infiniband/core/uverbs_main.c list_del(&entry->obj_list); entry 363 drivers/infiniband/core/uverbs_main.c kfree(entry); entry 383 drivers/infiniband/core/uverbs_main.c struct ib_uverbs_event *entry, *tmp; entry 386 drivers/infiniband/core/uverbs_main.c list_for_each_entry_safe(entry, tmp, &file->ev_queue.event_list, list) { entry 387 drivers/infiniband/core/uverbs_main.c if (entry->counter) entry 388 drivers/infiniband/core/uverbs_main.c list_del(&entry->obj_list); entry 389 drivers/infiniband/core/uverbs_main.c kfree(entry); entry 421 drivers/infiniband/core/uverbs_main.c struct ib_uverbs_event *entry; entry 433 drivers/infiniband/core/uverbs_main.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 434 drivers/infiniband/core/uverbs_main.c if (!entry) { entry 441 drivers/infiniband/core/uverbs_main.c entry->desc.comp.cq_handle = cq->uobject->user_handle; entry 442 drivers/infiniband/core/uverbs_main.c entry->counter = &uobj->comp_events_reported; entry 444 drivers/infiniband/core/uverbs_main.c list_add_tail(&entry->list, &ev_queue->event_list); entry 445 drivers/infiniband/core/uverbs_main.c list_add_tail(&entry->obj_list, &uobj->comp_list); entry 457 drivers/infiniband/core/uverbs_main.c struct ib_uverbs_event *entry; entry 466 drivers/infiniband/core/uverbs_main.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 467 drivers/infiniband/core/uverbs_main.c if (!entry) { entry 472 drivers/infiniband/core/uverbs_main.c entry->desc.async.element = element; entry 473 drivers/infiniband/core/uverbs_main.c entry->desc.async.event_type = event; entry 474 drivers/infiniband/core/uverbs_main.c entry->desc.async.reserved = 0; entry 475 drivers/infiniband/core/uverbs_main.c entry->counter = counter; entry 477 drivers/infiniband/core/uverbs_main.c list_add_tail(&entry->list, &file->async_file->ev_queue.event_list); entry 479 drivers/infiniband/core/uverbs_main.c list_add_tail(&entry->obj_list, obj_list); entry 55 drivers/infiniband/hw/cxgb3/cxio_hal.c list_for_each_entry(rdev, &rdev_list, entry) entry 65 drivers/infiniband/hw/cxgb3/cxio_hal.c list_for_each_entry(rdev, &rdev_list, entry) entry 191 drivers/infiniband/hw/cxgb3/cxio_hal.c struct cxio_qpid_list *entry; entry 197 drivers/infiniband/hw/cxgb3/cxio_hal.c entry = list_entry(uctx->qpids.next, struct cxio_qpid_list, entry 198 drivers/infiniband/hw/cxgb3/cxio_hal.c entry); entry 199 drivers/infiniband/hw/cxgb3/cxio_hal.c list_del(&entry->entry); entry 200 drivers/infiniband/hw/cxgb3/cxio_hal.c qpid = entry->qpid; entry 201 drivers/infiniband/hw/cxgb3/cxio_hal.c kfree(entry); entry 207 drivers/infiniband/hw/cxgb3/cxio_hal.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 208 drivers/infiniband/hw/cxgb3/cxio_hal.c if (!entry) entry 210 drivers/infiniband/hw/cxgb3/cxio_hal.c entry->qpid = i; entry 211 drivers/infiniband/hw/cxgb3/cxio_hal.c list_add_tail(&entry->entry, &uctx->qpids); entry 223 drivers/infiniband/hw/cxgb3/cxio_hal.c struct cxio_qpid_list *entry; entry 225 drivers/infiniband/hw/cxgb3/cxio_hal.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 226 drivers/infiniband/hw/cxgb3/cxio_hal.c if (!entry) entry 229 drivers/infiniband/hw/cxgb3/cxio_hal.c entry->qpid = qpid; entry 231 drivers/infiniband/hw/cxgb3/cxio_hal.c list_add_tail(&entry->entry, &uctx->qpids); entry 238 drivers/infiniband/hw/cxgb3/cxio_hal.c struct cxio_qpid_list *entry; entry 242 drivers/infiniband/hw/cxgb3/cxio_hal.c entry = list_entry(pos, struct cxio_qpid_list, entry); entry 243 drivers/infiniband/hw/cxgb3/cxio_hal.c list_del_init(&entry->entry); entry 244 drivers/infiniband/hw/cxgb3/cxio_hal.c if (!(entry->qpid & rdev_p->qpmask)) entry 245 drivers/infiniband/hw/cxgb3/cxio_hal.c cxio_hal_put_qpid(rdev_p->rscp, entry->qpid); entry 246 drivers/infiniband/hw/cxgb3/cxio_hal.c kfree(entry); entry 914 drivers/infiniband/hw/cxgb3/cxio_hal.c list_add_tail(&rdev_p->entry, &rdev_list); entry 1008 drivers/infiniband/hw/cxgb3/cxio_hal.c list_del(&rdev_p->entry); entry 1017 drivers/infiniband/hw/cxgb3/cxio_hal.c list_del(&rdev_p->entry); entry 1037 drivers/infiniband/hw/cxgb3/cxio_hal.c list_for_each_entry_safe(rdev, tmp, &rdev_list, entry) entry 90 drivers/infiniband/hw/cxgb3/cxio_hal.h struct list_head entry; entry 113 drivers/infiniband/hw/cxgb3/cxio_hal.h struct list_head entry; entry 53 drivers/infiniband/hw/cxgb3/cxio_resource.c u32 i, j, entry = 0, idx; entry 62 drivers/infiniband/hw/cxgb3/cxio_resource.c kfifo_in(fifo, (unsigned char *) &entry, sizeof(u32)); entry 89 drivers/infiniband/hw/cxgb3/cxio_resource.c if (kfifo_out_locked(fifo, (unsigned char *) &entry, entry 183 drivers/infiniband/hw/cxgb3/cxio_resource.c u32 entry; entry 184 drivers/infiniband/hw/cxgb3/cxio_resource.c if (kfifo_out_locked(fifo, (unsigned char *) &entry, sizeof(u32), lock)) entry 185 drivers/infiniband/hw/cxgb3/cxio_resource.c return entry; entry 191 drivers/infiniband/hw/cxgb3/cxio_resource.c u32 entry) entry 194 drivers/infiniband/hw/cxgb3/cxio_resource.c kfifo_in_locked(fifo, (unsigned char *) &entry, sizeof(u32), lock) entry 160 drivers/infiniband/hw/cxgb3/iwch.c list_add_tail(&rnicp->entry, &dev_list); entry 177 drivers/infiniband/hw/cxgb3/iwch.c list_for_each_entry_safe(dev, tmp, &dev_list, entry) { entry 182 drivers/infiniband/hw/cxgb3/iwch.c list_del(&dev->entry); entry 112 drivers/infiniband/hw/cxgb3/iwch.h struct list_head entry; entry 72 drivers/infiniband/hw/cxgb3/iwch_provider.c list_for_each_entry_safe(mm, tmp, &ucontext->mmaps, entry) entry 200 drivers/infiniband/hw/cxgb3/iwch_provider.h struct list_head entry; entry 215 drivers/infiniband/hw/cxgb3/iwch_provider.h mm = list_entry(pos, struct iwch_mm_entry, entry); entry 217 drivers/infiniband/hw/cxgb3/iwch_provider.h list_del_init(&mm->entry); entry 235 drivers/infiniband/hw/cxgb3/iwch_provider.h list_add_tail(&mm->entry, &ucontext->mmaps); entry 4283 drivers/infiniband/hw/cxgb4/cm.c ep = list_entry(tmp, struct c4iw_ep, entry); entry 4329 drivers/infiniband/hw/cxgb4/cm.c if (!ep->entry.next) { entry 4330 drivers/infiniband/hw/cxgb4/cm.c list_add_tail(&ep->entry, &timeout_list); entry 751 drivers/infiniband/hw/cxgb4/device.c struct c4iw_qid_list *entry; entry 755 drivers/infiniband/hw/cxgb4/device.c entry = list_entry(pos, struct c4iw_qid_list, entry); entry 756 drivers/infiniband/hw/cxgb4/device.c list_del_init(&entry->entry); entry 757 drivers/infiniband/hw/cxgb4/device.c if (!(entry->qid & rdev->qpmask)) { entry 759 drivers/infiniband/hw/cxgb4/device.c entry->qid); entry 764 drivers/infiniband/hw/cxgb4/device.c kfree(entry); entry 768 drivers/infiniband/hw/cxgb4/device.c entry = list_entry(pos, struct c4iw_qid_list, entry); entry 769 drivers/infiniband/hw/cxgb4/device.c list_del_init(&entry->entry); entry 770 drivers/infiniband/hw/cxgb4/device.c kfree(entry); entry 1094 drivers/infiniband/hw/cxgb4/device.c list_add_tail(&ctx->entry, &uld_ctx_list); entry 1558 drivers/infiniband/hw/cxgb4/device.c list_for_each_entry_safe(ctx, tmp, &uld_ctx_list, entry) { entry 104 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct list_head entry; entry 333 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct list_head entry; entry 542 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct list_head entry; entry 557 drivers/infiniband/hw/cxgb4/iw_cxgb4.h mm = list_entry(pos, struct c4iw_mm_entry, entry); entry 559 drivers/infiniband/hw/cxgb4/iw_cxgb4.h list_del_init(&mm->entry); entry 576 drivers/infiniband/hw/cxgb4/iw_cxgb4.h list_add_tail(&mm->entry, &ucontext->mmaps); entry 883 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct list_head entry; entry 950 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void c4iw_put_resource(struct c4iw_id_table *id_table, u32 entry); entry 70 drivers/infiniband/hw/cxgb4/provider.c list_for_each_entry_safe(mm, tmp, &ucontext->mmaps, entry) entry 905 drivers/infiniband/hw/cxgb4/qp.c static void add_to_fc_list(struct list_head *head, struct list_head *entry) entry 907 drivers/infiniband/hw/cxgb4/qp.c if (list_empty(entry)) entry 908 drivers/infiniband/hw/cxgb4/qp.c list_add_tail(entry, head); entry 95 drivers/infiniband/hw/cxgb4/resource.c u32 entry; entry 96 drivers/infiniband/hw/cxgb4/resource.c entry = c4iw_id_alloc(id_table); entry 97 drivers/infiniband/hw/cxgb4/resource.c if (entry == (u32)(-1)) entry 99 drivers/infiniband/hw/cxgb4/resource.c return entry; entry 102 drivers/infiniband/hw/cxgb4/resource.c void c4iw_put_resource(struct c4iw_id_table *id_table, u32 entry) entry 104 drivers/infiniband/hw/cxgb4/resource.c pr_debug("entry 0x%x\n", entry); entry 105 drivers/infiniband/hw/cxgb4/resource.c c4iw_id_free(id_table, entry); entry 110 drivers/infiniband/hw/cxgb4/resource.c struct c4iw_qid_list *entry; entry 116 drivers/infiniband/hw/cxgb4/resource.c entry = list_entry(uctx->cqids.next, struct c4iw_qid_list, entry 117 drivers/infiniband/hw/cxgb4/resource.c entry); entry 118 drivers/infiniband/hw/cxgb4/resource.c list_del(&entry->entry); entry 119 drivers/infiniband/hw/cxgb4/resource.c qid = entry->qid; entry 120 drivers/infiniband/hw/cxgb4/resource.c kfree(entry); entry 129 drivers/infiniband/hw/cxgb4/resource.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 130 drivers/infiniband/hw/cxgb4/resource.c if (!entry) entry 132 drivers/infiniband/hw/cxgb4/resource.c entry->qid = i; entry 133 drivers/infiniband/hw/cxgb4/resource.c list_add_tail(&entry->entry, &uctx->cqids); entry 140 drivers/infiniband/hw/cxgb4/resource.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 141 drivers/infiniband/hw/cxgb4/resource.c if (!entry) entry 143 drivers/infiniband/hw/cxgb4/resource.c entry->qid = qid; entry 144 drivers/infiniband/hw/cxgb4/resource.c list_add_tail(&entry->entry, &uctx->qpids); entry 146 drivers/infiniband/hw/cxgb4/resource.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 147 drivers/infiniband/hw/cxgb4/resource.c if (!entry) entry 149 drivers/infiniband/hw/cxgb4/resource.c entry->qid = i; entry 150 drivers/infiniband/hw/cxgb4/resource.c list_add_tail(&entry->entry, &uctx->qpids); entry 166 drivers/infiniband/hw/cxgb4/resource.c struct c4iw_qid_list *entry; entry 168 drivers/infiniband/hw/cxgb4/resource.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 169 drivers/infiniband/hw/cxgb4/resource.c if (!entry) entry 172 drivers/infiniband/hw/cxgb4/resource.c entry->qid = qid; entry 174 drivers/infiniband/hw/cxgb4/resource.c list_add_tail(&entry->entry, &uctx->cqids); entry 180 drivers/infiniband/hw/cxgb4/resource.c struct c4iw_qid_list *entry; entry 186 drivers/infiniband/hw/cxgb4/resource.c entry = list_entry(uctx->qpids.next, struct c4iw_qid_list, entry 187 drivers/infiniband/hw/cxgb4/resource.c entry); entry 188 drivers/infiniband/hw/cxgb4/resource.c list_del(&entry->entry); entry 189 drivers/infiniband/hw/cxgb4/resource.c qid = entry->qid; entry 190 drivers/infiniband/hw/cxgb4/resource.c kfree(entry); entry 203 drivers/infiniband/hw/cxgb4/resource.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 204 drivers/infiniband/hw/cxgb4/resource.c if (!entry) entry 206 drivers/infiniband/hw/cxgb4/resource.c entry->qid = i; entry 207 drivers/infiniband/hw/cxgb4/resource.c list_add_tail(&entry->entry, &uctx->qpids); entry 214 drivers/infiniband/hw/cxgb4/resource.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 215 drivers/infiniband/hw/cxgb4/resource.c if (!entry) entry 217 drivers/infiniband/hw/cxgb4/resource.c entry->qid = qid; entry 218 drivers/infiniband/hw/cxgb4/resource.c list_add_tail(&entry->entry, &uctx->cqids); entry 220 drivers/infiniband/hw/cxgb4/resource.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 221 drivers/infiniband/hw/cxgb4/resource.c if (!entry) entry 223 drivers/infiniband/hw/cxgb4/resource.c entry->qid = i; entry 224 drivers/infiniband/hw/cxgb4/resource.c list_add_tail(&entry->entry, &uctx->cqids); entry 240 drivers/infiniband/hw/cxgb4/resource.c struct c4iw_qid_list *entry; entry 242 drivers/infiniband/hw/cxgb4/resource.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 243 drivers/infiniband/hw/cxgb4/resource.c if (!entry) entry 246 drivers/infiniband/hw/cxgb4/resource.c entry->qid = qid; entry 248 drivers/infiniband/hw/cxgb4/resource.c list_add_tail(&entry->entry, &uctx->qpids); entry 185 drivers/infiniband/hw/efa/efa_verbs.c struct efa_mmap_entry *entry; entry 188 drivers/infiniband/hw/efa/efa_verbs.c xa_for_each(&ucontext->mmap_xa, mmap_page, entry) { entry 194 drivers/infiniband/hw/efa/efa_verbs.c entry->obj, get_mmap_key(entry), entry->address, entry 195 drivers/infiniband/hw/efa/efa_verbs.c entry->length); entry 196 drivers/infiniband/hw/efa/efa_verbs.c if (entry->mmap_flag == EFA_MMAP_DMA_PAGE) entry 198 drivers/infiniband/hw/efa/efa_verbs.c free_pages_exact(phys_to_virt(entry->address), entry 199 drivers/infiniband/hw/efa/efa_verbs.c entry->length); entry 200 drivers/infiniband/hw/efa/efa_verbs.c kfree(entry); entry 208 drivers/infiniband/hw/efa/efa_verbs.c struct efa_mmap_entry *entry; entry 215 drivers/infiniband/hw/efa/efa_verbs.c entry = xa_load(&ucontext->mmap_xa, mmap_page); entry 216 drivers/infiniband/hw/efa/efa_verbs.c if (!entry || get_mmap_key(entry) != key || entry->length != len) entry 221 drivers/infiniband/hw/efa/efa_verbs.c entry->obj, key, entry->address, entry->length); entry 223 drivers/infiniband/hw/efa/efa_verbs.c return entry; entry 233 drivers/infiniband/hw/efa/efa_verbs.c struct efa_mmap_entry *entry; entry 237 drivers/infiniband/hw/efa/efa_verbs.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 238 drivers/infiniband/hw/efa/efa_verbs.c if (!entry) entry 241 drivers/infiniband/hw/efa/efa_verbs.c entry->obj = obj; entry 242 drivers/infiniband/hw/efa/efa_verbs.c entry->address = address; entry 243 drivers/infiniband/hw/efa/efa_verbs.c entry->length = length; entry 244 drivers/infiniband/hw/efa/efa_verbs.c entry->mmap_flag = mmap_flag; entry 252 drivers/infiniband/hw/efa/efa_verbs.c entry->mmap_page = ucontext->mmap_xa_page; entry 254 drivers/infiniband/hw/efa/efa_verbs.c err = __xa_insert(&ucontext->mmap_xa, entry->mmap_page, entry, entry 264 drivers/infiniband/hw/efa/efa_verbs.c entry->obj, entry->address, entry->length, get_mmap_key(entry)); entry 266 drivers/infiniband/hw/efa/efa_verbs.c return get_mmap_key(entry); entry 270 drivers/infiniband/hw/efa/efa_verbs.c kfree(entry); entry 1595 drivers/infiniband/hw/efa/efa_verbs.c struct efa_mmap_entry *entry; entry 1600 drivers/infiniband/hw/efa/efa_verbs.c entry = mmap_entry_get(dev, ucontext, key, length); entry 1601 drivers/infiniband/hw/efa/efa_verbs.c if (!entry) { entry 1609 drivers/infiniband/hw/efa/efa_verbs.c entry->address, length, entry->mmap_flag); entry 1611 drivers/infiniband/hw/efa/efa_verbs.c pfn = entry->address >> PAGE_SHIFT; entry 1612 drivers/infiniband/hw/efa/efa_verbs.c switch (entry->mmap_flag) { entry 1637 drivers/infiniband/hw/efa/efa_verbs.c entry->address, length, entry->mmap_flag, err); entry 224 drivers/infiniband/hw/hfi1/affinity.c static void node_affinity_destroy(struct hfi1_affinity_node *entry) entry 226 drivers/infiniband/hw/hfi1/affinity.c free_percpu(entry->comp_vect_affinity); entry 227 drivers/infiniband/hw/hfi1/affinity.c kfree(entry); entry 233 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 237 drivers/infiniband/hw/hfi1/affinity.c entry = list_entry(pos, struct hfi1_affinity_node, entry 240 drivers/infiniband/hw/hfi1/affinity.c node_affinity_destroy(entry); entry 248 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 250 drivers/infiniband/hw/hfi1/affinity.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 251 drivers/infiniband/hw/hfi1/affinity.c if (!entry) entry 253 drivers/infiniband/hw/hfi1/affinity.c entry->node = node; entry 254 drivers/infiniband/hw/hfi1/affinity.c entry->comp_vect_affinity = alloc_percpu(u16); entry 255 drivers/infiniband/hw/hfi1/affinity.c INIT_LIST_HEAD(&entry->list); entry 257 drivers/infiniband/hw/hfi1/affinity.c return entry; entry 264 drivers/infiniband/hw/hfi1/affinity.c static void node_affinity_add_tail(struct hfi1_affinity_node *entry) entry 266 drivers/infiniband/hw/hfi1/affinity.c list_add_tail(&entry->list, &node_affinity.list); entry 273 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 276 drivers/infiniband/hw/hfi1/affinity.c entry = list_entry(pos, struct hfi1_affinity_node, list); entry 277 drivers/infiniband/hw/hfi1/affinity.c if (entry->node == node) entry 278 drivers/infiniband/hw/hfi1/affinity.c return entry; entry 362 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry, entry 387 drivers/infiniband/hw/hfi1/affinity.c &entry->def_intr.used); entry 441 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry) entry 469 drivers/infiniband/hw/hfi1/affinity.c cpu = _dev_comp_vect_cpu_get(dd, entry, non_intr_cpus, entry 497 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 500 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); entry 501 drivers/infiniband/hw/hfi1/affinity.c if (!entry) { entry 505 drivers/infiniband/hw/hfi1/affinity.c ret = _dev_comp_vect_mappings_create(dd, entry); entry 534 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry, entry 550 drivers/infiniband/hw/hfi1/affinity.c if (cpumask_weight(&entry->comp_vect_mask) == 1) { entry 556 drivers/infiniband/hw/hfi1/affinity.c cpumask_weight(&entry->comp_vect_mask) / entry 565 drivers/infiniband/hw/hfi1/affinity.c cpumask_weight(&entry->comp_vect_mask) % entry 574 drivers/infiniband/hw/hfi1/affinity.c curr_cpu = per_cpu_affinity_get(&entry->comp_vect_mask, entry 575 drivers/infiniband/hw/hfi1/affinity.c entry->comp_vect_affinity); entry 591 drivers/infiniband/hw/hfi1/affinity.c per_cpu_affinity_put_max(&entry->comp_vect_mask, entry 592 drivers/infiniband/hw/hfi1/affinity.c entry->comp_vect_affinity); entry 601 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry) entry 612 drivers/infiniband/hw/hfi1/affinity.c entry->comp_vect_affinity); entry 635 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 655 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); entry 661 drivers/infiniband/hw/hfi1/affinity.c if (!entry) { entry 662 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_allocate(node); entry 663 drivers/infiniband/hw/hfi1/affinity.c if (!entry) { entry 671 drivers/infiniband/hw/hfi1/affinity.c init_cpu_mask_set(&entry->def_intr); entry 672 drivers/infiniband/hw/hfi1/affinity.c init_cpu_mask_set(&entry->rcv_intr); entry 673 drivers/infiniband/hw/hfi1/affinity.c cpumask_clear(&entry->comp_vect_mask); entry 674 drivers/infiniband/hw/hfi1/affinity.c cpumask_clear(&entry->general_intr_mask); entry 676 drivers/infiniband/hw/hfi1/affinity.c cpumask_and(&entry->def_intr.mask, &node_affinity.real_cpu_mask, entry 680 drivers/infiniband/hw/hfi1/affinity.c possible = cpumask_weight(&entry->def_intr.mask); entry 681 drivers/infiniband/hw/hfi1/affinity.c curr_cpu = cpumask_first(&entry->def_intr.mask); entry 685 drivers/infiniband/hw/hfi1/affinity.c cpumask_set_cpu(curr_cpu, &entry->rcv_intr.mask); entry 686 drivers/infiniband/hw/hfi1/affinity.c cpumask_set_cpu(curr_cpu, &entry->general_intr_mask); entry 693 drivers/infiniband/hw/hfi1/affinity.c cpumask_clear_cpu(curr_cpu, &entry->def_intr.mask); entry 694 drivers/infiniband/hw/hfi1/affinity.c cpumask_set_cpu(curr_cpu, &entry->general_intr_mask); entry 696 drivers/infiniband/hw/hfi1/affinity.c &entry->def_intr.mask); entry 707 drivers/infiniband/hw/hfi1/affinity.c &entry->def_intr.mask); entry 709 drivers/infiniband/hw/hfi1/affinity.c &entry->rcv_intr.mask); entry 711 drivers/infiniband/hw/hfi1/affinity.c &entry->def_intr.mask); entry 721 drivers/infiniband/hw/hfi1/affinity.c if (cpumask_weight(&entry->def_intr.mask) == 0) entry 722 drivers/infiniband/hw/hfi1/affinity.c cpumask_copy(&entry->def_intr.mask, entry 723 drivers/infiniband/hw/hfi1/affinity.c &entry->general_intr_mask); entry 727 drivers/infiniband/hw/hfi1/affinity.c cpumask_and(&entry->comp_vect_mask, entry 729 drivers/infiniband/hw/hfi1/affinity.c cpumask_andnot(&entry->comp_vect_mask, entry 730 drivers/infiniband/hw/hfi1/affinity.c &entry->comp_vect_mask, entry 731 drivers/infiniband/hw/hfi1/affinity.c &entry->rcv_intr.mask); entry 732 drivers/infiniband/hw/hfi1/affinity.c cpumask_andnot(&entry->comp_vect_mask, entry 733 drivers/infiniband/hw/hfi1/affinity.c &entry->comp_vect_mask, entry 734 drivers/infiniband/hw/hfi1/affinity.c &entry->general_intr_mask); entry 741 drivers/infiniband/hw/hfi1/affinity.c if (cpumask_weight(&entry->comp_vect_mask) == 0) entry 742 drivers/infiniband/hw/hfi1/affinity.c cpumask_copy(&entry->comp_vect_mask, entry 743 drivers/infiniband/hw/hfi1/affinity.c &entry->general_intr_mask); entry 746 drivers/infiniband/hw/hfi1/affinity.c ret = _dev_comp_vect_cpu_mask_init(dd, entry, new_entry); entry 751 drivers/infiniband/hw/hfi1/affinity.c node_affinity_add_tail(entry); entry 759 drivers/infiniband/hw/hfi1/affinity.c node_affinity_destroy(entry); entry 766 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 772 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); entry 773 drivers/infiniband/hw/hfi1/affinity.c if (!entry) entry 780 drivers/infiniband/hw/hfi1/affinity.c _dev_comp_vect_cpu_mask_clean_up(dd, entry); entry 795 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 803 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); entry 804 drivers/infiniband/hw/hfi1/affinity.c if (!entry) entry 820 drivers/infiniband/hw/hfi1/affinity.c set = &entry->def_intr; entry 889 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 899 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); entry 905 drivers/infiniband/hw/hfi1/affinity.c set = &entry->def_intr; entry 908 drivers/infiniband/hw/hfi1/affinity.c cpu = cpumask_first(&entry->general_intr_mask); entry 913 drivers/infiniband/hw/hfi1/affinity.c cpu = cpumask_first(&entry->general_intr_mask); entry 915 drivers/infiniband/hw/hfi1/affinity.c set = &entry->rcv_intr; entry 971 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 974 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(dd->node); entry 978 drivers/infiniband/hw/hfi1/affinity.c set = &entry->def_intr; entry 988 drivers/infiniband/hw/hfi1/affinity.c set = &entry->rcv_intr; entry 1040 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node *entry; entry 1114 drivers/infiniband/hw/hfi1/affinity.c entry = node_affinity_lookup(node); entry 1115 drivers/infiniband/hw/hfi1/affinity.c if (entry) { entry 1116 drivers/infiniband/hw/hfi1/affinity.c cpumask_copy(intrs_mask, (entry->def_intr.gen ? entry 1117 drivers/infiniband/hw/hfi1/affinity.c &entry->def_intr.mask : entry 1118 drivers/infiniband/hw/hfi1/affinity.c &entry->def_intr.used)); entry 1119 drivers/infiniband/hw/hfi1/affinity.c cpumask_or(intrs_mask, intrs_mask, (entry->rcv_intr.gen ? entry 1120 drivers/infiniband/hw/hfi1/affinity.c &entry->rcv_intr.mask : entry 1121 drivers/infiniband/hw/hfi1/affinity.c &entry->rcv_intr.used)); entry 1122 drivers/infiniband/hw/hfi1/affinity.c cpumask_or(intrs_mask, intrs_mask, &entry->general_intr_mask); entry 1393 drivers/infiniband/hw/hfi1/chip.c static u64 dev_access_u32_csr(const struct cntr_entry *entry, entry 1397 drivers/infiniband/hw/hfi1/chip.c u64 csr = entry->csr; entry 1399 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_SDMA) { entry 1410 drivers/infiniband/hw/hfi1/chip.c static u64 access_sde_err_cnt(const struct cntr_entry *entry, entry 1420 drivers/infiniband/hw/hfi1/chip.c static u64 access_sde_int_cnt(const struct cntr_entry *entry, entry 1430 drivers/infiniband/hw/hfi1/chip.c static u64 access_sde_idle_int_cnt(const struct cntr_entry *entry, entry 1440 drivers/infiniband/hw/hfi1/chip.c static u64 access_sde_progress_int_cnt(const struct cntr_entry *entry, entry 1451 drivers/infiniband/hw/hfi1/chip.c static u64 dev_access_u64_csr(const struct cntr_entry *entry, void *context, entry 1457 drivers/infiniband/hw/hfi1/chip.c u64 csr = entry->csr; entry 1459 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_VL) { entry 1472 drivers/infiniband/hw/hfi1/chip.c static u64 dc_access_lcb_cntr(const struct cntr_entry *entry, void *context, entry 1476 drivers/infiniband/hw/hfi1/chip.c u32 csr = entry->csr; entry 1496 drivers/infiniband/hw/hfi1/chip.c static u64 port_access_u32_csr(const struct cntr_entry *entry, void *context, entry 1503 drivers/infiniband/hw/hfi1/chip.c return read_write_csr(ppd->dd, entry->csr, mode, data); entry 1506 drivers/infiniband/hw/hfi1/chip.c static u64 port_access_u64_csr(const struct cntr_entry *entry, entry 1511 drivers/infiniband/hw/hfi1/chip.c u64 csr = entry->csr; entry 1513 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_VL) { entry 1546 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_link_dn_cnt(const struct cntr_entry *entry, void *context, entry 1556 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_link_up_cnt(const struct cntr_entry *entry, void *context, entry 1566 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_unknown_frame_cnt(const struct cntr_entry *entry, entry 1577 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_xmit_discards(const struct cntr_entry *entry, entry 1594 drivers/infiniband/hw/hfi1/chip.c static u64 access_xmit_constraint_errs(const struct cntr_entry *entry, entry 1607 drivers/infiniband/hw/hfi1/chip.c static u64 access_rcv_constraint_errs(const struct cntr_entry *entry, entry 1654 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_cpu_intr(const struct cntr_entry *entry, entry 1663 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_cpu_rcv_limit(const struct cntr_entry *entry, entry 1672 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_pio_wait(const struct cntr_entry *entry, entry 1680 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_pio_drain(const struct cntr_entry *entry, entry 1688 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_ctx0_seq_drop(const struct cntr_entry *entry, entry 1696 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_vtx_wait(const struct cntr_entry *entry, entry 1704 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_kmem_wait(const struct cntr_entry *entry, entry 1712 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_send_schedule(const struct cntr_entry *entry, entry 1722 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_pll_lock_fail_err_cnt(const struct cntr_entry *entry, entry 1731 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_mbist_fail_err_cnt(const struct cntr_entry *entry, entry 1740 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_invalid_eep_cmd_err_cnt(const struct cntr_entry *entry, entry 1749 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_efuse_done_parity_err_cnt(const struct cntr_entry *entry, entry 1758 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_efuse_write_err_cnt(const struct cntr_entry *entry, entry 1768 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 1776 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_efuse_csr_parity_err_cnt(const struct cntr_entry *entry, entry 1785 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_fw_auth_failed_err_cnt(const struct cntr_entry *entry, entry 1794 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_key_mismatch_err_cnt(const struct cntr_entry *entry, entry 1803 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_sbus_write_failed_err_cnt(const struct cntr_entry *entry, entry 1813 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 1821 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry, entry 1830 drivers/infiniband/hw/hfi1/chip.c static u64 access_misc_csr_parity_err_cnt(const struct cntr_entry *entry, entry 1844 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 1856 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_msix_csr_parity_err_cnt(const struct cntr_entry *entry, entry 1865 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_int_map_unc_err_cnt(const struct cntr_entry *entry, entry 1874 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_int_map_cor_err_cnt(const struct cntr_entry *entry, entry 1883 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_msix_table_unc_err_cnt(const struct cntr_entry *entry, entry 1892 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_msix_table_cor_err_cnt(const struct cntr_entry *entry, entry 1902 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 1911 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 1919 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_seg_write_bad_addr_err_cnt(const struct cntr_entry *entry, entry 1928 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_seg_read_bad_addr_err_cnt(const struct cntr_entry *entry, entry 1937 drivers/infiniband/hw/hfi1/chip.c static u64 access_la_triggered_cnt(const struct cntr_entry *entry, entry 1945 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_trgt_cpl_timeout_err_cnt(const struct cntr_entry *entry, entry 1954 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_receive_parity_err_cnt(const struct cntr_entry *entry, entry 1964 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 1973 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 1981 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_cpl_dat_q_unc_err_cnt(const struct cntr_entry *entry, entry 1990 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_cpl_hd_q_unc_err_cnt(const struct cntr_entry *entry, entry 1999 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_post_dat_q_unc_err_cnt(const struct cntr_entry *entry, entry 2008 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_post_hd_q_unc_err_cnt(const struct cntr_entry *entry, entry 2017 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_retry_sot_mem_unc_err_cnt(const struct cntr_entry *entry, entry 2026 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_retry_mem_unc_err(const struct cntr_entry *entry, entry 2036 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2044 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_n_post_h_q_parity_err_cnt(const struct cntr_entry *entry, entry 2053 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_cpl_dat_q_cor_err_cnt(const struct cntr_entry *entry, entry 2062 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_cpl_hd_q_cor_err_cnt(const struct cntr_entry *entry, entry 2071 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_post_dat_q_cor_err_cnt(const struct cntr_entry *entry, entry 2080 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_post_hd_q_cor_err_cnt(const struct cntr_entry *entry, entry 2089 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_retry_sot_mem_cor_err_cnt(const struct cntr_entry *entry, entry 2098 drivers/infiniband/hw/hfi1/chip.c static u64 access_pcic_retry_mem_cor_err_cnt(const struct cntr_entry *entry, entry 2108 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2117 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2126 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2135 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2144 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2152 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_csr_cfg_bus_parity_err_cnt(const struct cntr_entry *entry, entry 2162 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2170 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_rspd_data_parity_err_cnt(const struct cntr_entry *entry, entry 2179 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_trgt_access_err_cnt(const struct cntr_entry *entry, entry 2189 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2197 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_csr_write_bad_addr_err_cnt(const struct cntr_entry *entry, entry 2206 drivers/infiniband/hw/hfi1/chip.c static u64 access_cce_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry, entry 2215 drivers/infiniband/hw/hfi1/chip.c static u64 access_ccs_csr_parity_err_cnt(const struct cntr_entry *entry, entry 2228 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_csr_parity_err_cnt(const struct cntr_entry *entry, entry 2237 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_csr_write_bad_addr_err_cnt(const struct cntr_entry *entry, entry 2246 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry, entry 2255 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_csr_unc_err_cnt(const struct cntr_entry *entry, entry 2264 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_dq_fsm_encoding_err_cnt(const struct cntr_entry *entry, entry 2273 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_eq_fsm_encoding_err_cnt(const struct cntr_entry *entry, entry 2282 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_csr_parity_err_cnt(const struct cntr_entry *entry, entry 2291 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_data_cor_err_cnt(const struct cntr_entry *entry, entry 2300 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_data_unc_err_cnt(const struct cntr_entry *entry, entry 2310 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2319 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2327 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_hdr_fifo_rd_cor_err_cnt(const struct cntr_entry *entry, entry 2336 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_hdr_fifo_rd_unc_err_cnt(const struct cntr_entry *entry, entry 2345 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_desc_part2_cor_err_cnt(const struct cntr_entry *entry, entry 2354 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_desc_part2_unc_err_cnt(const struct cntr_entry *entry, entry 2363 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_desc_part1_cor_err_cnt(const struct cntr_entry *entry, entry 2372 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_desc_part1_unc_err_cnt(const struct cntr_entry *entry, entry 2381 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_hq_intr_fsm_err_cnt(const struct cntr_entry *entry, entry 2391 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2400 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2409 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2418 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2427 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2436 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2445 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2454 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2463 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2472 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2481 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2490 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2499 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2507 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_empty_err_cnt(const struct cntr_entry *entry, entry 2516 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_full_err_cnt(const struct cntr_entry *entry, entry 2525 drivers/infiniband/hw/hfi1/chip.c static u64 access_rbuf_bad_lookup_err_cnt(const struct cntr_entry *entry, entry 2534 drivers/infiniband/hw/hfi1/chip.c static u64 access_rbuf_ctx_id_parity_err_cnt(const struct cntr_entry *entry, entry 2543 drivers/infiniband/hw/hfi1/chip.c static u64 access_rbuf_csr_qeopdw_parity_err_cnt(const struct cntr_entry *entry, entry 2553 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2562 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2571 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2580 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2589 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2598 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2607 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2616 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2625 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2633 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_lookup_des_cor_err_cnt(const struct cntr_entry *entry, entry 2642 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_lookup_des_unc_err_cnt(const struct cntr_entry *entry, entry 2652 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2661 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2669 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_free_list_cor_err_cnt(const struct cntr_entry *entry, entry 2678 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rbuf_free_list_unc_err_cnt(const struct cntr_entry *entry, entry 2687 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_fsm_encoding_err_cnt(const struct cntr_entry *entry, entry 2696 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_flag_cor_err_cnt(const struct cntr_entry *entry, entry 2705 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_flag_unc_err_cnt(const struct cntr_entry *entry, entry 2714 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dc_sop_eop_parity_err_cnt(const struct cntr_entry *entry, entry 2723 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_csr_parity_err_cnt(const struct cntr_entry *entry, entry 2733 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2742 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2750 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_data_cor_err_cnt(const struct cntr_entry *entry, entry 2759 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_data_unc_err_cnt(const struct cntr_entry *entry, entry 2768 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_hdr_cor_err_cnt(const struct cntr_entry *entry, entry 2777 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_rcv_hdr_unc_err_cnt(const struct cntr_entry *entry, entry 2786 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dc_intf_parity_err_cnt(const struct cntr_entry *entry, entry 2795 drivers/infiniband/hw/hfi1/chip.c static u64 access_rx_dma_csr_cor_err_cnt(const struct cntr_entry *entry, entry 2809 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2818 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2827 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2836 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2844 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_reserved_31_err_cnt(const struct cntr_entry *entry, entry 2853 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_reserved_30_err_cnt(const struct cntr_entry *entry, entry 2862 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_ppmc_sop_len_err_cnt(const struct cntr_entry *entry, entry 2872 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2880 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_vl_fifo_parity_err_cnt(const struct cntr_entry *entry, entry 2889 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_vlf_sop_parity_err_cnt(const struct cntr_entry *entry, entry 2898 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_vlf_v1_len_parity_err_cnt(const struct cntr_entry *entry, entry 2908 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2917 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2925 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_state_machine_err_cnt(const struct cntr_entry *entry, entry 2934 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_write_data_parity_err_cnt(const struct cntr_entry *entry, entry 2943 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_host_addr_mem_cor_err_cnt(const struct cntr_entry *entry, entry 2952 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_host_addr_mem_unc_err_cnt(const struct cntr_entry *entry, entry 2962 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2970 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_init_sm_in_err_cnt(const struct cntr_entry *entry, entry 2979 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_ppmc_pbl_fifo_err_cnt(const struct cntr_entry *entry, entry 2989 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 2998 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3007 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3016 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3025 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3034 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3043 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3052 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3061 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3069 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_pec_fifo_parity_err_cnt(const struct cntr_entry *entry, entry 3078 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_pcc_fifo_parity_err_cnt(const struct cntr_entry *entry, entry 3087 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_sb_mem_fifo1_err_cnt(const struct cntr_entry *entry, entry 3096 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_sb_mem_fifo0_err_cnt(const struct cntr_entry *entry, entry 3105 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_csr_parity_err_cnt(const struct cntr_entry *entry, entry 3114 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_write_addr_parity_err_cnt(const struct cntr_entry *entry, entry 3123 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_write_bad_ctxt_err_cnt(const struct cntr_entry *entry, entry 3137 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3146 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3154 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_csr_parity_err_cnt(const struct cntr_entry *entry, entry 3163 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_rpy_tag_err_cnt(const struct cntr_entry *entry, entry 3177 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3186 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3194 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_egress_fifo_cor_err_cnt(const struct cntr_entry *entry, entry 3203 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_read_pio_memory_cor_err_cnt(const struct cntr_entry *entry, entry 3213 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3221 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_sb_hdr_cor_err_cnt(const struct cntr_entry *entry, entry 3230 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_credit_overrun_err_cnt(const struct cntr_entry *entry, entry 3239 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo8_cor_err_cnt(const struct cntr_entry *entry, entry 3248 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo7_cor_err_cnt(const struct cntr_entry *entry, entry 3257 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo6_cor_err_cnt(const struct cntr_entry *entry, entry 3266 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo5_cor_err_cnt(const struct cntr_entry *entry, entry 3275 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo4_cor_err_cnt(const struct cntr_entry *entry, entry 3284 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo3_cor_err_cnt(const struct cntr_entry *entry, entry 3293 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo2_cor_err_cnt(const struct cntr_entry *entry, entry 3302 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo1_cor_err_cnt(const struct cntr_entry *entry, entry 3311 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_fifo0_cor_err_cnt(const struct cntr_entry *entry, entry 3320 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_credit_return_vl_err_cnt(const struct cntr_entry *entry, entry 3329 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_hcrc_insertion_err_cnt(const struct cntr_entry *entry, entry 3338 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_egress_fifo_unc_err_cnt(const struct cntr_entry *entry, entry 3347 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_read_pio_memory_unc_err_cnt(const struct cntr_entry *entry, entry 3357 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3365 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_sb_hdr_unc_err_cnt(const struct cntr_entry *entry, entry 3375 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3384 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3393 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3402 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3411 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3420 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3429 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3438 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3447 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3456 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3465 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3474 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3483 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3492 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3501 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3510 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3519 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3528 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3537 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3546 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3555 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3564 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3573 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3582 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3591 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3600 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3608 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_config_parity_err_cnt(const struct cntr_entry *entry, entry 3617 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_sbrd_ctl_csr_parity_err_cnt(const struct cntr_entry *entry, entry 3626 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_launch_csr_parity_err_cnt(const struct cntr_entry *entry, entry 3635 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_illegal_vl_err_cnt(const struct cntr_entry *entry, entry 3645 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3653 drivers/infiniband/hw/hfi1/chip.c static u64 access_egress_reserved_10_err_cnt(const struct cntr_entry *entry, entry 3662 drivers/infiniband/hw/hfi1/chip.c static u64 access_egress_reserved_9_err_cnt(const struct cntr_entry *entry, entry 3672 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3681 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3689 drivers/infiniband/hw/hfi1/chip.c static u64 access_egress_reserved_6_err_cnt(const struct cntr_entry *entry, entry 3699 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3707 drivers/infiniband/hw/hfi1/chip.c static u64 access_tx_linkdown_err_cnt(const struct cntr_entry *entry, entry 3717 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3725 drivers/infiniband/hw/hfi1/chip.c static u64 access_egress_reserved_2_err_cnt(const struct cntr_entry *entry, entry 3735 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3744 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3757 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3765 drivers/infiniband/hw/hfi1/chip.c static u64 access_send_csr_read_bad_addr_err_cnt(const struct cntr_entry *entry, entry 3774 drivers/infiniband/hw/hfi1/chip.c static u64 access_send_csr_parity_cnt(const struct cntr_entry *entry, entry 3788 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3796 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_write_overflow_err_cnt(const struct cntr_entry *entry, entry 3806 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3814 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_disallowed_packet_err_cnt(const struct cntr_entry *entry, entry 3823 drivers/infiniband/hw/hfi1/chip.c static u64 access_pio_inconsistent_sop_err_cnt(const struct cntr_entry *entry, entry 3837 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3846 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3855 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3863 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_assembly_cor_err_cnt(const struct cntr_entry *entry, entry 3872 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_desc_table_cor_err_cnt(const struct cntr_entry *entry, entry 3882 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3891 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3900 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3908 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_assembly_unc_err_cnt(const struct cntr_entry *entry, entry 3917 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_desc_table_unc_err_cnt(const struct cntr_entry *entry, entry 3926 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_timeout_err_cnt(const struct cntr_entry *entry, entry 3935 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_header_length_err_cnt(const struct cntr_entry *entry, entry 3944 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_header_address_err_cnt(const struct cntr_entry *entry, entry 3953 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_header_select_err_cnt(const struct cntr_entry *entry, entry 3962 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_reserved_9_err_cnt(const struct cntr_entry *entry, entry 3972 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 3980 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_length_mismatch_err_cnt(const struct cntr_entry *entry, entry 3989 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_halt_err_cnt(const struct cntr_entry *entry, entry 3997 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_mem_read_err_cnt(const struct cntr_entry *entry, entry 4006 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_first_desc_err_cnt(const struct cntr_entry *entry, entry 4016 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry, entry 4024 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_too_long_err_cnt(const struct cntr_entry *entry, entry 4033 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_gen_mismatch_err_cnt(const struct cntr_entry *entry, entry 4042 drivers/infiniband/hw/hfi1/chip.c static u64 access_sdma_wrong_dw_err_cnt(const struct cntr_entry *entry, entry 4051 drivers/infiniband/hw/hfi1/chip.c static u64 access_dc_rcv_err_cnt(const struct cntr_entry *entry, entry 4058 drivers/infiniband/hw/hfi1/chip.c u64 csr = entry->csr; entry 4074 drivers/infiniband/hw/hfi1/chip.c static u64 access_sw_cpu_##cntr(const struct cntr_entry *entry, \ entry 4088 drivers/infiniband/hw/hfi1/chip.c static u64 access_ibp_##cntr(const struct cntr_entry *entry, \ entry 8285 drivers/infiniband/hw/hfi1/chip.c const struct is_table *entry; entry 8288 drivers/infiniband/hw/hfi1/chip.c for (entry = &is_table[0]; entry->is_name; entry++) { entry 8289 drivers/infiniband/hw/hfi1/chip.c if (source <= entry->end) { entry 8290 drivers/infiniband/hw/hfi1/chip.c trace_hfi1_interrupt(dd, entry, source); entry 8291 drivers/infiniband/hw/hfi1/chip.c entry->is_int(dd, source - entry->start); entry 12076 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry; entry 12088 drivers/infiniband/hw/hfi1/chip.c entry = &dev_cntrs[i]; entry 12089 drivers/infiniband/hw/hfi1/chip.c hfi1_cdbg(CNTR, "reading %s", entry->name); entry 12090 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_DISABLED) { entry 12094 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_VL) { entry 12097 drivers/infiniband/hw/hfi1/chip.c val = entry->rw_cntr(entry, entry 12105 drivers/infiniband/hw/hfi1/chip.c dd->cntrs[entry->offset + j] = entry 12108 drivers/infiniband/hw/hfi1/chip.c } else if (entry->flags & CNTR_SDMA) { entry 12114 drivers/infiniband/hw/hfi1/chip.c entry->rw_cntr(entry, dd, j, entry 12119 drivers/infiniband/hw/hfi1/chip.c dd->cntrs[entry->offset + j] = entry 12123 drivers/infiniband/hw/hfi1/chip.c val = entry->rw_cntr(entry, dd, entry 12126 drivers/infiniband/hw/hfi1/chip.c dd->cntrs[entry->offset] = val; entry 12147 drivers/infiniband/hw/hfi1/chip.c const struct cntr_entry *entry; entry 12154 drivers/infiniband/hw/hfi1/chip.c entry = &port_cntrs[i]; entry 12155 drivers/infiniband/hw/hfi1/chip.c hfi1_cdbg(CNTR, "reading %s", entry->name); entry 12156 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_DISABLED) { entry 12162 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_VL) { entry 12165 drivers/infiniband/hw/hfi1/chip.c val = entry->rw_cntr(entry, ppd, j, entry 12172 drivers/infiniband/hw/hfi1/chip.c ppd->cntrs[entry->offset + j] = val; entry 12175 drivers/infiniband/hw/hfi1/chip.c val = entry->rw_cntr(entry, ppd, entry 12179 drivers/infiniband/hw/hfi1/chip.c ppd->cntrs[entry->offset] = val; entry 12221 drivers/infiniband/hw/hfi1/chip.c static u64 read_dev_port_cntr(struct hfi1_devdata *dd, struct cntr_entry *entry, entry 12227 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_DISABLED) { entry 12228 drivers/infiniband/hw/hfi1/chip.c dd_dev_err(dd, "Counter %s not enabled", entry->name); entry 12232 drivers/infiniband/hw/hfi1/chip.c hfi1_cdbg(CNTR, "cntr: %s vl %d psval 0x%llx", entry->name, vl, *psval); entry 12234 drivers/infiniband/hw/hfi1/chip.c val = entry->rw_cntr(entry, context, vl, CNTR_MODE_R, 0); entry 12237 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_SYNTH) { entry 12243 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_32BIT) { entry 12273 drivers/infiniband/hw/hfi1/chip.c struct cntr_entry *entry, entry 12278 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_DISABLED) { entry 12279 drivers/infiniband/hw/hfi1/chip.c dd_dev_err(dd, "Counter %s not enabled", entry->name); entry 12283 drivers/infiniband/hw/hfi1/chip.c hfi1_cdbg(CNTR, "cntr: %s vl %d psval 0x%llx", entry->name, vl, *psval); entry 12285 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_SYNTH) { entry 12287 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_32BIT) { entry 12288 drivers/infiniband/hw/hfi1/chip.c val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W, entry 12292 drivers/infiniband/hw/hfi1/chip.c val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W, entry 12296 drivers/infiniband/hw/hfi1/chip.c val = entry->rw_cntr(entry, context, vl, CNTR_MODE_W, data); entry 12308 drivers/infiniband/hw/hfi1/chip.c struct cntr_entry *entry; entry 12311 drivers/infiniband/hw/hfi1/chip.c entry = &dev_cntrs[index]; entry 12312 drivers/infiniband/hw/hfi1/chip.c sval = dd->scntrs + entry->offset; entry 12317 drivers/infiniband/hw/hfi1/chip.c return read_dev_port_cntr(dd, entry, sval, dd, vl); entry 12322 drivers/infiniband/hw/hfi1/chip.c struct cntr_entry *entry; entry 12325 drivers/infiniband/hw/hfi1/chip.c entry = &dev_cntrs[index]; entry 12326 drivers/infiniband/hw/hfi1/chip.c sval = dd->scntrs + entry->offset; entry 12331 drivers/infiniband/hw/hfi1/chip.c return write_dev_port_cntr(dd, entry, sval, dd, vl, data); entry 12336 drivers/infiniband/hw/hfi1/chip.c struct cntr_entry *entry; entry 12339 drivers/infiniband/hw/hfi1/chip.c entry = &port_cntrs[index]; entry 12340 drivers/infiniband/hw/hfi1/chip.c sval = ppd->scntrs + entry->offset; entry 12351 drivers/infiniband/hw/hfi1/chip.c return read_dev_port_cntr(ppd->dd, entry, sval, ppd, vl); entry 12356 drivers/infiniband/hw/hfi1/chip.c struct cntr_entry *entry; entry 12359 drivers/infiniband/hw/hfi1/chip.c entry = &port_cntrs[index]; entry 12360 drivers/infiniband/hw/hfi1/chip.c sval = ppd->scntrs + entry->offset; entry 12371 drivers/infiniband/hw/hfi1/chip.c return write_dev_port_cntr(ppd->dd, entry, sval, ppd, vl, data); entry 12382 drivers/infiniband/hw/hfi1/chip.c struct cntr_entry *entry; entry 12392 drivers/infiniband/hw/hfi1/chip.c entry = &dev_cntrs[C_DC_RCV_FLITS]; entry 12393 drivers/infiniband/hw/hfi1/chip.c cur_rx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, CNTR_MODE_R, 0); entry 12395 drivers/infiniband/hw/hfi1/chip.c entry = &dev_cntrs[C_DC_XMIT_FLITS]; entry 12396 drivers/infiniband/hw/hfi1/chip.c cur_tx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, CNTR_MODE_R, 0); entry 12426 drivers/infiniband/hw/hfi1/chip.c entry = &dev_cntrs[i]; entry 12427 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_VL) { entry 12437 drivers/infiniband/hw/hfi1/chip.c entry = &port_cntrs[j]; entry 12438 drivers/infiniband/hw/hfi1/chip.c if (entry->flags & CNTR_VL) { entry 12453 drivers/infiniband/hw/hfi1/chip.c entry = &dev_cntrs[C_DC_XMIT_FLITS]; entry 12454 drivers/infiniband/hw/hfi1/chip.c dd->last_tx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, entry 12457 drivers/infiniband/hw/hfi1/chip.c entry = &dev_cntrs[C_DC_RCV_FLITS]; entry 12458 drivers/infiniband/hw/hfi1/chip.c dd->last_rx = entry->rw_cntr(entry, dd, CNTR_INVALID_VL, entry 298 drivers/infiniband/hw/hfi1/eprom.c struct hfi1_eprom_table_entry *entry; entry 342 drivers/infiniband/hw/hfi1/eprom.c for (entry = NULL, i = 0; i < footer->num_table_entries; i++) { entry 344 drivers/infiniband/hw/hfi1/eprom.c entry = &table[i]; entry 348 drivers/infiniband/hw/hfi1/eprom.c if (!entry) { entry 357 drivers/infiniband/hw/hfi1/eprom.c if (entry->size > (4 * 1024)) { entry 359 drivers/infiniband/hw/hfi1/eprom.c entry->size); entry 365 drivers/infiniband/hw/hfi1/eprom.c if (entry->offset + entry->size < entry->offset) { entry 368 drivers/infiniband/hw/hfi1/eprom.c entry->offset, entry->size); entry 374 drivers/infiniband/hw/hfi1/eprom.c buffer = kmalloc(entry->size, GFP_KERNEL); entry 383 drivers/infiniband/hw/hfi1/eprom.c seg_offset = entry->offset % SEG_SIZE; entry 384 drivers/infiniband/hw/hfi1/eprom.c seg_base = entry->offset - seg_offset; entry 386 drivers/infiniband/hw/hfi1/eprom.c while (ncopied < entry->size) { entry 400 drivers/infiniband/hw/hfi1/eprom.c entry->offset); entry 408 drivers/infiniband/hw/hfi1/eprom.c to_copy = entry->size - ncopied; entry 439 drivers/infiniband/hw/hfi1/eprom.c *size = entry->size; entry 573 drivers/infiniband/hw/hfi1/init.c cce = cc_state->cct.entries[max_ccti].entry; entry 3969 drivers/infiniband/hw/hfi1/mad.c cc_table_attr->ccti_entries[j].entry = entry 3970 drivers/infiniband/hw/hfi1/mad.c cpu_to_be16(entries[i].entry); entry 4021 drivers/infiniband/hw/hfi1/mad.c entries[i].entry = be16_to_cpu(p->ccti_entries[j].entry); entry 315 drivers/infiniband/hw/hfi1/mad.h __be16 entry; /* shift:2, multiplier:14 */ entry 319 drivers/infiniband/hw/hfi1/mad.h u16 entry; /* shift:2, multiplier:14 */ entry 1666 drivers/infiniband/hw/hfi1/tid_rdma.c u64 hfi1_access_sw_tid_wait(const struct cntr_entry *entry, entry 241 drivers/infiniband/hw/hfi1/tid_rdma.h u64 hfi1_access_sw_tid_wait(const struct cntr_entry *entry, entry 714 drivers/infiniband/hw/hfi1/trace_tid.h TP_PROTO(struct rvt_qp *qp, int index, u32 entry), entry 715 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(qp, index, entry) entry 732 drivers/infiniband/hw/hfi1/trace_tid.h TP_PROTO(struct rvt_qp *qp, int index, u32 entry), entry 733 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(qp, index, entry) entry 738 drivers/infiniband/hw/hfi1/trace_tid.h TP_PROTO(struct rvt_qp *qp, int index, u32 entry), entry 739 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(qp, index, entry) entry 3862 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_aeqe *get_aeqe_v1(struct hns_roce_eq *eq, u32 entry) entry 3864 drivers/infiniband/hw/hns/hns_roce_hw_v1.c unsigned long off = (entry & (eq->entries - 1)) * entry 3970 drivers/infiniband/hw/hns/hns_roce_hw_v1.c static struct hns_roce_ceqe *get_ceqe_v1(struct hns_roce_eq *eq, u32 entry) entry 3972 drivers/infiniband/hw/hns/hns_roce_hw_v1.c unsigned long off = (entry & (eq->entries - 1)) * entry 1787 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_link_table_entry *entry; entry 1806 drivers/infiniband/hw/hns/hns_roce_hw_v2.c entry = link_tbl->table.buf; entry 1831 drivers/infiniband/hw/hns/hns_roce_hw_v2.c req_a->head_ba_l = cpu_to_le32(entry[0].blk_ba0); entry 1833 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(entry[0].blk_ba1_nxt_ptr); entry 1839 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(entry[page_num - 1].blk_ba0); entry 1843 drivers/infiniband/hw/hns/hns_roce_hw_v2.c entry[page_num - 1].blk_ba1_nxt_ptr & entry 1848 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (entry[page_num - 2].blk_ba1_nxt_ptr & entry 1864 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_link_table_entry *entry; entry 1906 drivers/infiniband/hw/hns/hns_roce_hw_v2.c entry = link_tbl->table.buf; entry 1915 drivers/infiniband/hw/hns/hns_roce_hw_v2.c entry[i].blk_ba0 = (u32)(t >> 12); entry 1916 drivers/infiniband/hw/hns/hns_roce_hw_v2.c entry[i].blk_ba1_nxt_ptr = (u32)(t >> 44); entry 1919 drivers/infiniband/hw/hns/hns_roce_hw_v2.c entry[i].blk_ba1_nxt_ptr |= entry 4981 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *get_aeqe_v2(struct hns_roce_eq *eq, u32 entry) entry 4987 drivers/infiniband/hw/hns/hns_roce_hw_v2.c off = (entry & (eq->entries - 1)) * HNS_ROCE_AEQ_ENTRY_SIZE; entry 4993 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_aeqe *mhop_get_aeqe(struct hns_roce_eq *eq, u32 entry) entry 5000 drivers/infiniband/hw/hns/hns_roce_hw_v2.c off = (entry & (eq->entries - 1)) * HNS_ROCE_AEQ_ENTRY_SIZE; entry 5111 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *get_ceqe_v2(struct hns_roce_eq *eq, u32 entry) entry 5117 drivers/infiniband/hw/hns/hns_roce_hw_v2.c off = (entry & (eq->entries - 1)) * HNS_ROCE_CEQ_ENTRY_SIZE; entry 5123 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static struct hns_roce_ceqe *mhop_get_ceqe(struct hns_roce_eq *eq, u32 entry) entry 5130 drivers/infiniband/hw/hns/hns_roce_hw_v2.c off = (entry & (eq->entries - 1)) * HNS_ROCE_CEQ_ENTRY_SIZE; entry 306 drivers/infiniband/hw/i40iw/i40iw_ctrl.c static struct i40iw_sc_qp *i40iw_qp_from_entry(struct list_head *entry) entry 308 drivers/infiniband/hw/i40iw/i40iw_ctrl.c if (!entry) entry 311 drivers/infiniband/hw/i40iw/i40iw_ctrl.c return (struct i40iw_sc_qp *)((char *)entry - offsetof(struct i40iw_sc_qp, list)); entry 321 drivers/infiniband/hw/i40iw/i40iw_ctrl.c struct list_head *entry = NULL; entry 328 drivers/infiniband/hw/i40iw/i40iw_ctrl.c entry = head->next; entry 331 drivers/infiniband/hw/i40iw/i40iw_ctrl.c entry = (lastentry != head) ? lastentry->next : NULL; entry 334 drivers/infiniband/hw/i40iw/i40iw_ctrl.c return i40iw_qp_from_entry(entry); entry 3659 drivers/infiniband/hw/i40iw/i40iw_ctrl.c memcpy((char *)sdbuf->va + offset, &info->entry[3], entry 3672 drivers/infiniband/hw/i40iw/i40iw_ctrl.c (LS_64(info->entry[2].cmd, I40IW_CQPSQ_UPESD_SDCMD) | entry 3675 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 56, info->entry[2].data); entry 3679 drivers/infiniband/hw/i40iw/i40iw_ctrl.c (LS_64(info->entry[1].cmd, I40IW_CQPSQ_UPESD_SDCMD) | entry 3682 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 40, info->entry[1].data); entry 3686 drivers/infiniband/hw/i40iw/i40iw_ctrl.c LS_64(info->entry[0].cmd, I40IW_CQPSQ_UPESD_SDCMD)); entry 3688 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 8, info->entry[0].data); entry 115 drivers/infiniband/hw/i40iw/i40iw_hmc.c struct update_sd_entry *entry) entry 117 drivers/infiniband/hw/i40iw/i40iw_hmc.c entry->data = pa | (I40IW_HMC_MAX_BP_COUNT << I40E_PFHMC_SDDATALOW_PMSDBPCOUNT_SHIFT) | entry 121 drivers/infiniband/hw/i40iw/i40iw_hmc.c entry->cmd = (idx | (1 << I40E_PFHMC_SDCMD_PMSDWR_SHIFT) | (1 << 15)); entry 131 drivers/infiniband/hw/i40iw/i40iw_hmc.c struct update_sd_entry *entry) entry 133 drivers/infiniband/hw/i40iw/i40iw_hmc.c entry->data = (I40IW_HMC_MAX_BP_COUNT << entry 137 drivers/infiniband/hw/i40iw/i40iw_hmc.c entry->cmd = (idx | (1 << I40E_PFHMC_SDCMD_PMSDWR_SHIFT) | (1 << 15)); entry 160 drivers/infiniband/hw/i40iw/i40iw_hmc.c i40iw_set_sd_entry(pa, sd_idx, type, sdinfo.entry); entry 162 drivers/infiniband/hw/i40iw/i40iw_hmc.c i40iw_clr_sd_entry(sd_idx, type, sdinfo.entry); entry 200 drivers/infiniband/hw/i40iw/i40iw_hmc.c &sdinfo.entry[sdinfo.cnt]); entry 203 drivers/infiniband/hw/i40iw/i40iw_hmc.c &sdinfo.entry[sdinfo.cnt]); entry 159 drivers/infiniband/hw/i40iw/i40iw_hmc.h struct update_sd_entry entry[I40IW_MAX_SD_ENTRIES]; entry 1457 drivers/infiniband/hw/i40iw/i40iw_main.c iwdev->iw_msixtbl[i].idx = ldev->msix_entries[i].entry; entry 734 drivers/infiniband/hw/i40iw/i40iw_utils.c struct list_head *entry; entry 739 drivers/infiniband/hw/i40iw/i40iw_utils.c entry = (void *)list->next; entry 740 drivers/infiniband/hw/i40iw/i40iw_utils.c list_del(entry); entry 741 drivers/infiniband/hw/i40iw/i40iw_utils.c return (void *)entry; entry 2344 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct ib_wc *entry) entry 2370 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->wc_flags = 0; entry 2371 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->wr_id = cq_poll_info.wr_id; entry 2373 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->status = IB_WC_WR_FLUSH_ERR; entry 2374 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->vendor_err = cq_poll_info.major_err << 16 | cq_poll_info.minor_err; entry 2376 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->status = IB_WC_SUCCESS; entry 2381 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->opcode = IB_WC_RDMA_WRITE; entry 2385 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->opcode = IB_WC_RDMA_READ; entry 2391 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->opcode = IB_WC_SEND; entry 2394 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->opcode = IB_WC_RECV; entry 2397 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->opcode = IB_WC_RECV; entry 2401 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->ex.imm_data = 0; entry 2403 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->qp = (struct ib_qp *)qp->back_qp; entry 2404 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->src_qp = cq_poll_info.qp_id; entry 2412 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry->byte_len = cq_poll_info.bytes_xfered; entry 2413 drivers/infiniband/hw/i40iw/i40iw_verbs.c entry++; entry 584 drivers/infiniband/hw/mlx4/alias_GUID.c int j, k, entry; entry 590 drivers/infiniband/hw/mlx4/alias_GUID.c entry = j * NUM_ALIAS_GUID_IN_REC + k; entry 592 drivers/infiniband/hw/mlx4/alias_GUID.c if (!entry || entry > dev->dev->persist->num_vfs || entry 593 drivers/infiniband/hw/mlx4/alias_GUID.c !mlx4_is_slave_active(dev->dev, entry)) entry 595 drivers/infiniband/hw/mlx4/alias_GUID.c guid = mlx4_get_admin_guid(dev->dev, entry, port); entry 600 drivers/infiniband/hw/mlx4/alias_GUID.c entry, entry 249 drivers/infiniband/hw/mlx4/main.c static void free_gid_entry(struct gid_entry *entry) entry 251 drivers/infiniband/hw/mlx4/main.c memset(&entry->gid, 0, sizeof(entry->gid)); entry 252 drivers/infiniband/hw/mlx4/main.c kfree(entry->ctx); entry 253 drivers/infiniband/hw/mlx4/main.c entry->ctx = NULL; entry 686 drivers/infiniband/hw/mlx4/sysfs.c list_add_tail(&p->kobj.entry, &dev->pkeys.pkey_port_list[slave]); entry 748 drivers/infiniband/hw/mlx4/sysfs.c entry) { entry 749 drivers/infiniband/hw/mlx4/sysfs.c list_del(&p->entry); entry 793 drivers/infiniband/hw/mlx4/sysfs.c entry) { entry 794 drivers/infiniband/hw/mlx4/sysfs.c list_del(&p->entry); entry 2429 drivers/infiniband/hw/mlx5/devx.c void *entry; entry 2434 drivers/infiniband/hw/mlx5/devx.c xa_for_each(&table->event_xa, id, entry) { entry 2435 drivers/infiniband/hw/mlx5/devx.c event = entry; entry 2439 drivers/infiniband/hw/mlx5/devx.c kfree(entry); entry 2504 drivers/infiniband/hw/mlx5/devx.c struct devx_async_data *entry, *tmp; entry 2507 drivers/infiniband/hw/mlx5/devx.c list_for_each_entry_safe(entry, tmp, entry 2509 drivers/infiniband/hw/mlx5/devx.c kvfree(entry); entry 2647 drivers/infiniband/hw/mlx5/devx.c struct devx_async_event_data *entry, *tmp; entry 2668 drivers/infiniband/hw/mlx5/devx.c list_for_each_entry_safe(entry, tmp, entry 2670 drivers/infiniband/hw/mlx5/devx.c kfree(entry); /* read can't come any more */ entry 57 drivers/infiniband/hw/mlx5/mem.c int entry; entry 65 drivers/infiniband/hw/mlx5/mem.c for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { entry 139 drivers/infiniband/hw/mlx5/mem.c int entry; entry 155 drivers/infiniband/hw/mlx5/mem.c for_each_sg(umem->sg_head.sgl, sg, umem->nmap, entry) { entry 1225 drivers/infiniband/hw/mlx5/mlx5_ib.h struct mlx5_ib_mr *mlx5_mr_cache_alloc(struct mlx5_ib_dev *dev, int entry); entry 422 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_mr *mlx5_mr_cache_alloc(struct mlx5_ib_dev *dev, int entry) entry 429 drivers/infiniband/hw/mlx5/mr.c if (entry < 0 || entry >= MAX_MR_CACHE_ENTRIES) { entry 430 drivers/infiniband/hw/mlx5/mr.c mlx5_ib_err(dev, "cache entry %d is out of range\n", entry); entry 434 drivers/infiniband/hw/mlx5/mr.c ent = &cache->ent[entry]; entry 440 drivers/infiniband/hw/mlx5/mr.c err = add_keys(dev, entry, 1); entry 160 drivers/infiniband/hw/mthca/mthca_cq.c int entry) entry 163 drivers/infiniband/hw/mthca/mthca_cq.c return buf->queue.direct.buf + (entry * MTHCA_CQ_ENTRY_SIZE); entry 165 drivers/infiniband/hw/mthca/mthca_cq.c return buf->queue.page_list[entry * MTHCA_CQ_ENTRY_SIZE / PAGE_SIZE].buf entry 166 drivers/infiniband/hw/mthca/mthca_cq.c + (entry * MTHCA_CQ_ENTRY_SIZE) % PAGE_SIZE; entry 169 drivers/infiniband/hw/mthca/mthca_cq.c static inline struct mthca_cqe *get_cqe(struct mthca_cq *cq, int entry) entry 171 drivers/infiniband/hw/mthca/mthca_cq.c return get_cqe_from_buf(&cq->buf, entry); entry 375 drivers/infiniband/hw/mthca/mthca_cq.c struct ib_wc *entry, int *free_cqe) entry 394 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_LOC_LEN_ERR; entry 397 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_LOC_QP_OP_ERR; entry 400 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_LOC_EEC_OP_ERR; entry 403 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_LOC_PROT_ERR; entry 406 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_WR_FLUSH_ERR; entry 409 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_MW_BIND_ERR; entry 412 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_BAD_RESP_ERR; entry 415 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_LOC_ACCESS_ERR; entry 418 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_REM_INV_REQ_ERR; entry 421 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_REM_ACCESS_ERR; entry 424 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_REM_OP_ERR; entry 427 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_RETRY_EXC_ERR; entry 430 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_RNR_RETRY_EXC_ERR; entry 433 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_LOC_RDD_VIOL_ERR; entry 436 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_REM_INV_RD_REQ_ERR; entry 439 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_REM_ABORT_ERR; entry 442 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_INV_EECN_ERR; entry 445 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_INV_EEC_STATE_ERR; entry 448 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_GENERAL_ERR; entry 452 drivers/infiniband/hw/mthca/mthca_cq.c entry->vendor_err = cqe->vendor_err; entry 482 drivers/infiniband/hw/mthca/mthca_cq.c struct ib_wc *entry) entry 531 drivers/infiniband/hw/mthca/mthca_cq.c entry->qp = &(*cur_qp)->ibqp; entry 537 drivers/infiniband/hw/mthca/mthca_cq.c entry->wr_id = (*cur_qp)->wrid[wqe_index + entry 544 drivers/infiniband/hw/mthca/mthca_cq.c entry->wr_id = srq->wrid[wqe_index]; entry 558 drivers/infiniband/hw/mthca/mthca_cq.c entry->wr_id = (*cur_qp)->wrid[wqe_index]; entry 573 drivers/infiniband/hw/mthca/mthca_cq.c entry, &free_cqe); entry 578 drivers/infiniband/hw/mthca/mthca_cq.c entry->wc_flags = 0; entry 581 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_RDMA_WRITE; entry 584 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_RDMA_WRITE; entry 585 drivers/infiniband/hw/mthca/mthca_cq.c entry->wc_flags |= IB_WC_WITH_IMM; entry 588 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_SEND; entry 591 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_SEND; entry 592 drivers/infiniband/hw/mthca/mthca_cq.c entry->wc_flags |= IB_WC_WITH_IMM; entry 595 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_RDMA_READ; entry 596 drivers/infiniband/hw/mthca/mthca_cq.c entry->byte_len = be32_to_cpu(cqe->byte_cnt); entry 599 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_COMP_SWAP; entry 600 drivers/infiniband/hw/mthca/mthca_cq.c entry->byte_len = MTHCA_ATOMIC_BYTE_LEN; entry 603 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_FETCH_ADD; entry 604 drivers/infiniband/hw/mthca/mthca_cq.c entry->byte_len = MTHCA_ATOMIC_BYTE_LEN; entry 607 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = MTHCA_OPCODE_INVALID; entry 611 drivers/infiniband/hw/mthca/mthca_cq.c entry->byte_len = be32_to_cpu(cqe->byte_cnt); entry 615 drivers/infiniband/hw/mthca/mthca_cq.c entry->wc_flags = IB_WC_WITH_IMM; entry 616 drivers/infiniband/hw/mthca/mthca_cq.c entry->ex.imm_data = cqe->imm_etype_pkey_eec; entry 617 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_RECV; entry 621 drivers/infiniband/hw/mthca/mthca_cq.c entry->wc_flags = IB_WC_WITH_IMM; entry 622 drivers/infiniband/hw/mthca/mthca_cq.c entry->ex.imm_data = cqe->imm_etype_pkey_eec; entry 623 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_RECV_RDMA_WITH_IMM; entry 626 drivers/infiniband/hw/mthca/mthca_cq.c entry->wc_flags = 0; entry 627 drivers/infiniband/hw/mthca/mthca_cq.c entry->opcode = IB_WC_RECV; entry 630 drivers/infiniband/hw/mthca/mthca_cq.c entry->slid = be16_to_cpu(cqe->rlid); entry 631 drivers/infiniband/hw/mthca/mthca_cq.c entry->sl = cqe->sl_ipok >> 4; entry 632 drivers/infiniband/hw/mthca/mthca_cq.c entry->src_qp = be32_to_cpu(cqe->rqpn) & 0xffffff; entry 633 drivers/infiniband/hw/mthca/mthca_cq.c entry->dlid_path_bits = cqe->g_mlpath & 0x7f; entry 634 drivers/infiniband/hw/mthca/mthca_cq.c entry->pkey_index = be32_to_cpu(cqe->imm_etype_pkey_eec) >> 16; entry 635 drivers/infiniband/hw/mthca/mthca_cq.c entry->wc_flags |= cqe->g_mlpath & 0x80 ? IB_WC_GRH : 0; entry 638 drivers/infiniband/hw/mthca/mthca_cq.c entry->wc_flags |= (cqe->sl_ipok & 1 && checksum == 0xffff) ? entry 642 drivers/infiniband/hw/mthca/mthca_cq.c entry->status = IB_WC_SUCCESS; entry 655 drivers/infiniband/hw/mthca/mthca_cq.c struct ib_wc *entry) entry 671 drivers/infiniband/hw/mthca/mthca_cq.c &freed, entry + npolled); entry 495 drivers/infiniband/hw/mthca/mthca_dev.h struct ib_wc *entry); entry 228 drivers/infiniband/hw/mthca/mthca_eq.c static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry) entry 230 drivers/infiniband/hw/mthca/mthca_eq.c unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; entry 287 drivers/infiniband/hw/ocrdma/ocrdma.h struct list_head entry; entry 454 drivers/infiniband/hw/ocrdma/ocrdma.h struct list_head entry; entry 209 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c INIT_LIST_HEAD(&mm->entry); entry 212 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c list_add_tail(&mm->entry, &uctx->mm_head); entry 223 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c list_for_each_entry_safe(mm, tmp, &uctx->mm_head, entry) { entry 227 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c list_del(&mm->entry); entry 241 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c list_for_each_entry(mm, &uctx->mm_head, entry) { entry 539 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c list_for_each_entry_safe(mm, tmp, &uctx->mm_head, entry) { entry 540 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c list_del(&mm->entry); entry 309 drivers/infiniband/hw/qedr/qedr.h struct list_head entry; entry 278 drivers/infiniband/hw/qedr/verbs.c INIT_LIST_HEAD(&mm->entry); entry 281 drivers/infiniband/hw/qedr/verbs.c list_add(&mm->entry, &uctx->mm_head); entry 299 drivers/infiniband/hw/qedr/verbs.c list_for_each_entry(mm, &uctx->mm_head, entry) { entry 378 drivers/infiniband/hw/qedr/verbs.c list_for_each_entry_safe(mm, tmp, &uctx->mm_head, entry) { entry 382 drivers/infiniband/hw/qedr/verbs.c list_del(&mm->entry); entry 2205 drivers/infiniband/hw/qib/qib_mad.c p->ccti_entries[i].entry = cpu_to_be16(entries[i].entry); entry 2278 drivers/infiniband/hw/qib/qib_mad.c entries[i].entry = be16_to_cpu(p->ccti_entries[i].entry); entry 262 drivers/infiniband/hw/qib/qib_mad.h __be16 entry; /* shift:2, multiplier:14 */ entry 266 drivers/infiniband/hw/qib/qib_mad.h u16 entry; /* shift:2, multiplier:14 */ entry 67 drivers/infiniband/sw/rdmavt/cq.c bool rvt_cq_enter(struct rvt_cq *cq, struct ib_wc *entry, bool solicited) entry 120 drivers/infiniband/sw/rdmavt/cq.c trace_rvt_cq_enter(cq, entry, head); entry 122 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].wr_id = entry->wr_id; entry 123 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].status = entry->status; entry 124 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].opcode = entry->opcode; entry 125 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].vendor_err = entry->vendor_err; entry 126 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].byte_len = entry->byte_len; entry 127 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].ex.imm_data = entry->ex.imm_data; entry 128 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].qp_num = entry->qp->qp_num; entry 129 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].src_qp = entry->src_qp; entry 130 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].wc_flags = entry->wc_flags; entry 131 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].pkey_index = entry->pkey_index; entry 132 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].slid = ib_lid_cpu16(entry->slid); entry 133 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].sl = entry->sl; entry 134 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].dlid_path_bits = entry->dlid_path_bits; entry 135 drivers/infiniband/sw/rdmavt/cq.c uqueue[head].port_num = entry->port_num; entry 139 drivers/infiniband/sw/rdmavt/cq.c kqueue[head] = *entry; entry 145 drivers/infiniband/sw/rdmavt/cq.c (solicited || entry->status != IB_WC_SUCCESS))) { entry 517 drivers/infiniband/sw/rdmavt/cq.c int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) entry 535 drivers/infiniband/sw/rdmavt/cq.c for (npolled = 0; npolled < num_entries; ++npolled, ++entry) { entry 540 drivers/infiniband/sw/rdmavt/cq.c *entry = wc->kqueue[tail]; entry 59 drivers/infiniband/sw/rdmavt/cq.h int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); entry 245 drivers/infiniband/sw/rdmavt/qp.c int entry; entry 272 drivers/infiniband/sw/rdmavt/qp.c entry = (atomic_inc_return(&wss->clean_entry) - 1) entry 276 drivers/infiniband/sw/rdmavt/qp.c bits = xchg(&wss->entries[entry], 0); entry 290 drivers/infiniband/sw/rdmavt/qp.c u32 entry = page / BITS_PER_LONG; /* assumes this ends up a shift */ entry 293 drivers/infiniband/sw/rdmavt/qp.c if (!test_and_set_bit(nr, &wss->entries[entry])) entry 138 drivers/infiniband/sw/siw/siw_verbs.c void *entry; entry 146 drivers/infiniband/sw/siw/siw_verbs.c xa_for_each(&uctx->xa, index, entry) { entry 169 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct __opa_vnic_mactable_entry *nentry = &node->entry; entry 170 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct opa_veswport_mactable_entry *entry; entry 177 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c entry = &tbl->tbl_entries[node->index - loffset]; entry 178 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c memcpy(entry->mac_addr, nentry->mac_addr, entry 179 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c ARRAY_SIZE(entry->mac_addr)); entry 180 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c memcpy(entry->mac_addr_mask, nentry->mac_addr_mask, entry 181 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c ARRAY_SIZE(entry->mac_addr_mask)); entry 182 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c entry->dlid_sd = cpu_to_be32(nentry->dlid_sd); entry 227 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct opa_veswport_mactable_entry *entry = entry 229 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c u8 *mac_addr = entry->mac_addr; entry 235 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c entry->dlid_sd); entry 248 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c nentry = &node->entry; entry 249 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c memcpy(nentry->mac_addr, entry->mac_addr, entry 251 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c memcpy(nentry->mac_addr_mask, entry->mac_addr_mask, entry 253 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c nentry->dlid_sd = be32_to_cpu(entry->dlid_sd); entry 254 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c key = node->entry.mac_addr[OPA_VNIC_MAC_HASH_IDX]; entry 275 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c memcpy(&new_node->entry, &node->entry, sizeof(node->entry)); entry 276 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c key = new_node->entry.mac_addr[OPA_VNIC_MAC_HASH_IDX]; entry 313 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c struct __opa_vnic_mactable_entry *entry = &node->entry; entry 316 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c if (unlikely(OPA_VNIC_DLID_SD_IS_SRC_MAC(entry->dlid_sd))) entry 319 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c if (!memcmp(node->entry.mac_addr, mac_hdr->h_dest, entry 320 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c ARRAY_SIZE(node->entry.mac_addr))) { entry 322 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c dlid = OPA_VNIC_DLID_SD_GET_DLID(node->entry.dlid_sd); entry 443 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.h struct opa_vnic_iface_mac_entry entry[0]; entry 242 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h struct __opa_vnic_mactable_entry entry; entry 323 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c struct opa_vnic_iface_mac_entry *entry = &macs->entry[count]; entry 329 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c memcpy(entry, ha->addr, sizeof(*entry)); entry 356 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c struct opa_vnic_iface_mac_entry *entry = &macs->entry[count]; entry 369 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c memcpy(entry, ha->addr, sizeof(*entry)); entry 375 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c struct opa_vnic_iface_mac_entry *entry = &macs->entry[count]; entry 381 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c memcpy(entry, ha->addr, sizeof(*entry)); entry 471 drivers/infiniband/ulp/srp/ib_srp.c list_add_tail(&d->entry, &pool->free_list); entry 496 drivers/infiniband/ulp/srp/ib_srp.c d = list_first_entry(&pool->free_list, typeof(*d), entry); entry 497 drivers/infiniband/ulp/srp/ib_srp.c list_del(&d->entry); entry 521 drivers/infiniband/ulp/srp/ib_srp.c list_add(&desc[i]->entry, &pool->free_list); entry 288 drivers/infiniband/ulp/srp/ib_srp.h struct list_head entry; entry 1965 drivers/infiniband/ulp/srpt/ib_srpt.c list_for_each_entry(nexus, &sport->nexus_list, entry) { entry 2005 drivers/infiniband/ulp/srpt/ib_srpt.c list_for_each_entry(nexus, &sport->nexus_list, entry) { entry 2029 drivers/infiniband/ulp/srpt/ib_srpt.c list_for_each_entry(n, &sport->nexus_list, entry) { entry 2037 drivers/infiniband/ulp/srpt/ib_srpt.c list_add_tail_rcu(&tmp_nexus->entry, entry 2939 drivers/infiniband/ulp/srpt/ib_srpt.c list_for_each_entry(nexus, &sport->nexus_list, entry) entry 2967 drivers/infiniband/ulp/srpt/ib_srpt.c list_for_each_entry(nexus, &sport->nexus_list, entry) { entry 2978 drivers/infiniband/ulp/srpt/ib_srpt.c list_for_each_entry_safe(nexus, next_n, &sport->nexus_list, entry) { entry 2979 drivers/infiniband/ulp/srpt/ib_srpt.c list_del(&nexus->entry); entry 343 drivers/infiniband/ulp/srpt/ib_srpt.h struct list_head entry; entry 1294 drivers/input/input.c struct proc_dir_entry *entry; entry 1300 drivers/input/input.c entry = proc_create("devices", 0, proc_bus_input_dir, entry 1302 drivers/input/input.c if (!entry) entry 1305 drivers/input/input.c entry = proc_create("handlers", 0, proc_bus_input_dir, entry 1307 drivers/input/input.c if (!entry) entry 1179 drivers/input/misc/wistron_btns.c struct key_entry *entry) entry 1181 drivers/input/misc/wistron_btns.c switch (entry->type) { entry 1186 drivers/input/misc/wistron_btns.c entry->type = KE_KEY; entry 1187 drivers/input/misc/wistron_btns.c entry->keycode = KEY_WLAN; entry 1193 drivers/input/misc/wistron_btns.c entry->type = KE_KEY; entry 1194 drivers/input/misc/wistron_btns.c entry->keycode = KEY_BLUETOOTH; entry 1199 drivers/input/misc/wistron_btns.c if (entry->code & FE_UNTESTED) entry 101 drivers/input/rmi4/rmi_driver.c struct rmi_function *entry; entry 104 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { entry 105 drivers/input/rmi4/rmi_driver.c retval = reset_one_function(entry); entry 116 drivers/input/rmi4/rmi_driver.c struct rmi_function *entry; entry 119 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { entry 120 drivers/input/rmi4/rmi_driver.c retval = configure_one_function(entry); entry 246 drivers/input/rmi4/rmi_driver.c struct rmi_function *entry; entry 248 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { entry 249 drivers/input/rmi4/rmi_driver.c if (entry->fd.function_number == number) entry 250 drivers/input/rmi4/rmi_driver.c return entry; entry 278 drivers/input/rmi4/rmi_driver.c struct rmi_function *entry; entry 281 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { entry 282 drivers/input/rmi4/rmi_driver.c retval = suspend_one_function(entry); entry 312 drivers/input/rmi4/rmi_driver.c struct rmi_function *entry; entry 315 drivers/input/rmi4/rmi_driver.c list_for_each_entry(entry, &data->function_list, node) { entry 316 drivers/input/rmi4/rmi_driver.c retval = resume_one_function(entry); entry 460 drivers/input/rmi4/rmi_driver.c struct pdt_entry *entry, u16 pdt_address) entry 472 drivers/input/rmi4/rmi_driver.c entry->page_start = pdt_address & RMI4_PAGE_MASK; entry 473 drivers/input/rmi4/rmi_driver.c entry->query_base_addr = buf[0]; entry 474 drivers/input/rmi4/rmi_driver.c entry->command_base_addr = buf[1]; entry 475 drivers/input/rmi4/rmi_driver.c entry->control_base_addr = buf[2]; entry 476 drivers/input/rmi4/rmi_driver.c entry->data_base_addr = buf[3]; entry 477 drivers/input/rmi4/rmi_driver.c entry->interrupt_source_count = buf[4] & RMI_PDT_INT_SOURCE_COUNT_MASK; entry 478 drivers/input/rmi4/rmi_driver.c entry->function_version = (buf[4] & RMI_PDT_FUNCTION_VERSION_MASK) >> 5; entry 479 drivers/input/rmi4/rmi_driver.c entry->function_number = buf[5]; entry 505 drivers/input/rmi4/rmi_driver.c const struct pdt_entry *entry)) entry 544 drivers/input/rmi4/rmi_driver.c void *ctx, const struct pdt_entry *entry)) entry 95 drivers/input/rmi4/rmi_driver.h const struct pdt_entry *entry)); entry 91 drivers/input/rmi4/rmi_smbus.c struct mapping_table_entry *entry = &rmi_smb->mapping_table[i]; entry 93 drivers/input/rmi4/rmi_smbus.c if (le16_to_cpu(entry->rmiaddr) == rmiaddr) { entry 95 drivers/input/rmi4/rmi_smbus.c if (entry->readcount == bytecount) entry 98 drivers/input/rmi4/rmi_smbus.c if (entry->flags & RMI_SMB2_MAP_FLAGS_WE) { entry 172 drivers/input/sparse-keymap.c struct key_entry *map, *entry; entry 185 drivers/input/sparse-keymap.c entry = &map[i]; entry 188 drivers/input/sparse-keymap.c error = setup(dev, entry); entry 193 drivers/input/sparse-keymap.c switch (entry->type) { entry 196 drivers/input/sparse-keymap.c __set_bit(entry->keycode, dev->keybit); entry 202 drivers/input/sparse-keymap.c __set_bit(entry->sw.code, dev->swbit); entry 128 drivers/iommu/amd_iommu.c struct acpihid_map_entry *entry) entry 143 drivers/iommu/amd_iommu.c return strcmp(hid, entry->hid); entry 145 drivers/iommu/amd_iommu.c if (!(*entry->uid)) entry 146 drivers/iommu/amd_iommu.c return strcmp(hid, entry->hid); entry 148 drivers/iommu/amd_iommu.c return (strcmp(hid, entry->hid) || strcmp(uid, entry->uid)); entry 159 drivers/iommu/amd_iommu.c struct acpihid_map_entry **entry) entry 165 drivers/iommu/amd_iommu.c if (entry) entry 166 drivers/iommu/amd_iommu.c *entry = p; entry 313 drivers/iommu/amd_iommu.c struct acpihid_map_entry *p, *entry = NULL; entry 316 drivers/iommu/amd_iommu.c devid = get_acpihid_device_id(dev, &entry); entry 322 drivers/iommu/amd_iommu.c entry->group = p->group; entry 325 drivers/iommu/amd_iommu.c if (!entry->group) entry 326 drivers/iommu/amd_iommu.c entry->group = generic_device_group(dev); entry 328 drivers/iommu/amd_iommu.c iommu_group_ref_get(entry->group); entry 330 drivers/iommu/amd_iommu.c return entry->group; entry 686 drivers/iommu/amd_iommu.c u64 entry[2]; entry 703 drivers/iommu/amd_iommu.c entry[0] = raw[0]; entry 704 drivers/iommu/amd_iommu.c entry[1] = raw[1]; entry 717 drivers/iommu/amd_iommu.c iommu_handle_ppr_entry(iommu, entry); entry 2897 drivers/iommu/amd_iommu.c struct iommu_dev_data *entry; entry 2903 drivers/iommu/amd_iommu.c entry = list_first_entry(&domain->dev_list, entry 2905 drivers/iommu/amd_iommu.c BUG_ON(!entry->domain); entry 2906 drivers/iommu/amd_iommu.c do_detach(entry); entry 3180 drivers/iommu/amd_iommu.c struct unity_map_entry *entry; entry 3187 drivers/iommu/amd_iommu.c list_for_each_entry(entry, &amd_iommu_unity_map, list) { entry 3191 drivers/iommu/amd_iommu.c if (devid < entry->devid_start || devid > entry->devid_end) entry 3195 drivers/iommu/amd_iommu.c length = entry->address_end - entry->address_start; entry 3196 drivers/iommu/amd_iommu.c if (entry->prot & IOMMU_PROT_IR) entry 3198 drivers/iommu/amd_iommu.c if (entry->prot & IOMMU_PROT_IW) entry 3200 drivers/iommu/amd_iommu.c if (entry->prot & IOMMU_UNITY_MAP_FLAG_EXCL_RANGE) entry 3204 drivers/iommu/amd_iommu.c region = iommu_alloc_resv_region(entry->address_start, entry 3231 drivers/iommu/amd_iommu.c struct iommu_resv_region *entry, *next; entry 3233 drivers/iommu/amd_iommu.c list_for_each_entry_safe(entry, next, head, list) entry 3234 drivers/iommu/amd_iommu.c kfree(entry); entry 3879 drivers/iommu/amd_iommu.c struct irte_ga *entry; entry 3891 drivers/iommu/amd_iommu.c entry = (struct irte_ga *)table->table; entry 3892 drivers/iommu/amd_iommu.c entry = &entry[index]; entry 3893 drivers/iommu/amd_iommu.c entry->lo.fields_remap.valid = 0; entry 3894 drivers/iommu/amd_iommu.c entry->hi.val = irte->hi.val; entry 3895 drivers/iommu/amd_iommu.c entry->lo.val = irte->lo.val; entry 3896 drivers/iommu/amd_iommu.c entry->lo.fields_remap.valid = 1; entry 3898 drivers/iommu/amd_iommu.c data->ref = entry; entry 3954 drivers/iommu/amd_iommu.c static void irte_prepare(void *entry, entry 3958 drivers/iommu/amd_iommu.c union irte *irte = (union irte *) entry; entry 3968 drivers/iommu/amd_iommu.c static void irte_ga_prepare(void *entry, entry 3972 drivers/iommu/amd_iommu.c struct irte_ga *irte = (struct irte_ga *) entry; entry 3984 drivers/iommu/amd_iommu.c static void irte_activate(void *entry, u16 devid, u16 index) entry 3986 drivers/iommu/amd_iommu.c union irte *irte = (union irte *) entry; entry 3992 drivers/iommu/amd_iommu.c static void irte_ga_activate(void *entry, u16 devid, u16 index) entry 3994 drivers/iommu/amd_iommu.c struct irte_ga *irte = (struct irte_ga *) entry; entry 4000 drivers/iommu/amd_iommu.c static void irte_deactivate(void *entry, u16 devid, u16 index) entry 4002 drivers/iommu/amd_iommu.c union irte *irte = (union irte *) entry; entry 4008 drivers/iommu/amd_iommu.c static void irte_ga_deactivate(void *entry, u16 devid, u16 index) entry 4010 drivers/iommu/amd_iommu.c struct irte_ga *irte = (struct irte_ga *) entry; entry 4016 drivers/iommu/amd_iommu.c static void irte_set_affinity(void *entry, u16 devid, u16 index, entry 4019 drivers/iommu/amd_iommu.c union irte *irte = (union irte *) entry; entry 4026 drivers/iommu/amd_iommu.c static void irte_ga_set_affinity(void *entry, u16 devid, u16 index, entry 4029 drivers/iommu/amd_iommu.c struct irte_ga *irte = (struct irte_ga *) entry; entry 4171 drivers/iommu/amd_iommu.c struct IO_APIC_route_entry *entry; entry 4179 drivers/iommu/amd_iommu.c iommu->irte_ops->prepare(data->entry, apic->irq_delivery_mode, entry 4186 drivers/iommu/amd_iommu.c entry = info->ioapic_entry; entry 4188 drivers/iommu/amd_iommu.c memset(entry, 0, sizeof(*entry)); entry 4189 drivers/iommu/amd_iommu.c entry->vector = index; entry 4190 drivers/iommu/amd_iommu.c entry->mask = 0; entry 4191 drivers/iommu/amd_iommu.c entry->trigger = info->ioapic_trigger; entry 4192 drivers/iommu/amd_iommu.c entry->polarity = info->ioapic_polarity; entry 4195 drivers/iommu/amd_iommu.c entry->mask = 1; entry 4313 drivers/iommu/amd_iommu.c data->entry = kzalloc(sizeof(union irte), GFP_KERNEL); entry 4315 drivers/iommu/amd_iommu.c data->entry = kzalloc(sizeof(struct irte_ga), entry 4317 drivers/iommu/amd_iommu.c if (!data->entry) { entry 4358 drivers/iommu/amd_iommu.c kfree(data->entry); entry 4381 drivers/iommu/amd_iommu.c iommu->irte_ops->activate(data->entry, irte_info->devid, entry 4395 drivers/iommu/amd_iommu.c iommu->irte_ops->deactivate(data->entry, irte_info->devid, entry 4409 drivers/iommu/amd_iommu.c struct irte_ga *entry = (struct irte_ga *) ir_data->entry; entry 4412 drivers/iommu/amd_iommu.c !entry || entry->lo.fields_vapic.guest_mode) entry 4415 drivers/iommu/amd_iommu.c entry->lo.val = 0; entry 4416 drivers/iommu/amd_iommu.c entry->hi.val = 0; entry 4418 drivers/iommu/amd_iommu.c entry->lo.fields_vapic.guest_mode = 1; entry 4419 drivers/iommu/amd_iommu.c entry->lo.fields_vapic.ga_log_intr = 1; entry 4420 drivers/iommu/amd_iommu.c entry->hi.fields.ga_root_ptr = ir_data->ga_root_ptr; entry 4421 drivers/iommu/amd_iommu.c entry->hi.fields.vector = ir_data->ga_vector; entry 4422 drivers/iommu/amd_iommu.c entry->lo.fields_vapic.ga_tag = ir_data->ga_tag; entry 4425 drivers/iommu/amd_iommu.c ir_data->irq_2_irte.index, entry, ir_data); entry 4432 drivers/iommu/amd_iommu.c struct irte_ga *entry = (struct irte_ga *) ir_data->entry; entry 4436 drivers/iommu/amd_iommu.c !entry || !entry->lo.fields_vapic.guest_mode) entry 4439 drivers/iommu/amd_iommu.c entry->lo.val = 0; entry 4440 drivers/iommu/amd_iommu.c entry->hi.val = 0; entry 4442 drivers/iommu/amd_iommu.c entry->lo.fields_remap.dm = apic->irq_dest_mode; entry 4443 drivers/iommu/amd_iommu.c entry->lo.fields_remap.int_type = apic->irq_delivery_mode; entry 4444 drivers/iommu/amd_iommu.c entry->hi.fields.vector = cfg->vector; entry 4445 drivers/iommu/amd_iommu.c entry->lo.fields_remap.destination = entry 4447 drivers/iommu/amd_iommu.c entry->hi.fields.destination = entry 4451 drivers/iommu/amd_iommu.c ir_data->irq_2_irte.index, entry, ir_data); entry 4522 drivers/iommu/amd_iommu.c iommu->irte_ops->set_affinity(ir_data->entry, irte_info->devid, entry 4596 drivers/iommu/amd_iommu.c struct irte_ga *entry = (struct irte_ga *) ir_data->entry; entry 4600 drivers/iommu/amd_iommu.c !ref || !entry || !entry->lo.fields_vapic.guest_mode) entry 348 drivers/iommu/amd_iommu_init.c u64 entry; entry 353 drivers/iommu/amd_iommu_init.c entry = start | MMIO_EXCL_ENABLE_MASK; entry 355 drivers/iommu/amd_iommu_init.c &entry, sizeof(entry)); entry 357 drivers/iommu/amd_iommu_init.c entry = limit; entry 359 drivers/iommu/amd_iommu_init.c &entry, sizeof(entry)); entry 365 drivers/iommu/amd_iommu_init.c u64 entry; entry 369 drivers/iommu/amd_iommu_init.c entry = iommu_virt_to_phys(amd_iommu_dev_table); entry 370 drivers/iommu/amd_iommu_init.c entry |= (dev_table_size >> 12) - 1; entry 372 drivers/iommu/amd_iommu_init.c &entry, sizeof(entry)); entry 626 drivers/iommu/amd_iommu_init.c u64 entry; entry 630 drivers/iommu/amd_iommu_init.c entry = iommu_virt_to_phys(iommu->cmd_buf); entry 631 drivers/iommu/amd_iommu_init.c entry |= MMIO_CMD_SIZE_512; entry 634 drivers/iommu/amd_iommu_init.c &entry, sizeof(entry)); entry 663 drivers/iommu/amd_iommu_init.c u64 entry; entry 667 drivers/iommu/amd_iommu_init.c entry = iommu_virt_to_phys(iommu->evt_buf) | EVT_LEN_MASK; entry 670 drivers/iommu/amd_iommu_init.c &entry, sizeof(entry)); entry 703 drivers/iommu/amd_iommu_init.c u64 entry; entry 708 drivers/iommu/amd_iommu_init.c entry = iommu_virt_to_phys(iommu->ppr_log) | PPR_LOG_SIZE_512; entry 711 drivers/iommu/amd_iommu_init.c &entry, sizeof(entry)); entry 773 drivers/iommu/amd_iommu_init.c u64 entry; entry 788 drivers/iommu/amd_iommu_init.c entry = iommu_virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512; entry 790 drivers/iommu/amd_iommu_init.c &entry, sizeof(entry)); entry 791 drivers/iommu/amd_iommu_init.c entry = (iommu_virt_to_phys(iommu->ga_log_tail) & entry 794 drivers/iommu/amd_iommu_init.c &entry, sizeof(entry)); entry 864 drivers/iommu/amd_iommu_init.c u64 int_ctl, int_tab_len, entry = 0, last_entry = 0; entry 881 drivers/iommu/amd_iommu_init.c entry = (((u64) hi) << 32) + lo; entry 882 drivers/iommu/amd_iommu_init.c if (last_entry && last_entry != entry) { entry 887 drivers/iommu/amd_iommu_init.c last_entry = entry; entry 889 drivers/iommu/amd_iommu_init.c old_devtb_size = ((entry & ~PAGE_MASK) + 1) << 12; entry 902 drivers/iommu/amd_iommu_init.c old_devtb_phys = __sme_clr(entry) & PAGE_MASK; entry 1008 drivers/iommu/amd_iommu_init.c struct devid_map *entry; entry 1018 drivers/iommu/amd_iommu_init.c list_for_each_entry(entry, list, list) { entry 1019 drivers/iommu/amd_iommu_init.c if (!(entry->id == id && entry->cmd_line)) entry 1025 drivers/iommu/amd_iommu_init.c *devid = entry->devid; entry 1030 drivers/iommu/amd_iommu_init.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1031 drivers/iommu/amd_iommu_init.c if (!entry) entry 1034 drivers/iommu/amd_iommu_init.c entry->id = id; entry 1035 drivers/iommu/amd_iommu_init.c entry->devid = *devid; entry 1036 drivers/iommu/amd_iommu_init.c entry->cmd_line = cmd_line; entry 1038 drivers/iommu/amd_iommu_init.c list_add_tail(&entry->list, list); entry 1046 drivers/iommu/amd_iommu_init.c struct acpihid_map_entry *entry; entry 1049 drivers/iommu/amd_iommu_init.c list_for_each_entry(entry, list, list) { entry 1050 drivers/iommu/amd_iommu_init.c if (strcmp(entry->hid, hid) || entry 1051 drivers/iommu/amd_iommu_init.c (*uid && *entry->uid && strcmp(entry->uid, uid)) || entry 1052 drivers/iommu/amd_iommu_init.c !entry->cmd_line) entry 1057 drivers/iommu/amd_iommu_init.c *devid = entry->devid; entry 1061 drivers/iommu/amd_iommu_init.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1062 drivers/iommu/amd_iommu_init.c if (!entry) entry 1065 drivers/iommu/amd_iommu_init.c memcpy(entry->uid, uid, strlen(uid)); entry 1066 drivers/iommu/amd_iommu_init.c memcpy(entry->hid, hid, strlen(hid)); entry 1067 drivers/iommu/amd_iommu_init.c entry->devid = *devid; entry 1068 drivers/iommu/amd_iommu_init.c entry->cmd_line = cmd_line; entry 1069 drivers/iommu/amd_iommu_init.c entry->root_devid = (entry->devid & (~0x7)); entry 1072 drivers/iommu/amd_iommu_init.c entry->cmd_line ? "cmd" : "ivrs", entry 1073 drivers/iommu/amd_iommu_init.c entry->hid, entry->uid, entry->root_devid); entry 1075 drivers/iommu/amd_iommu_init.c list_add_tail(&entry->list, list); entry 2070 drivers/iommu/amd_iommu_init.c struct unity_map_entry *entry, *next; entry 2072 drivers/iommu/amd_iommu_init.c list_for_each_entry_safe(entry, next, &amd_iommu_unity_map, list) { entry 2073 drivers/iommu/amd_iommu_init.c list_del(&entry->list); entry 2074 drivers/iommu/amd_iommu_init.c kfree(entry); entry 763 drivers/iommu/amd_iommu_types.h struct devid_map *entry; entry 765 drivers/iommu/amd_iommu_types.h list_for_each_entry(entry, &ioapic_map, list) { entry 766 drivers/iommu/amd_iommu_types.h if (entry->id == id) entry 767 drivers/iommu/amd_iommu_types.h return entry->devid; entry 775 drivers/iommu/amd_iommu_types.h struct devid_map *entry; entry 777 drivers/iommu/amd_iommu_types.h list_for_each_entry(entry, &hpet_map, list) { entry 778 drivers/iommu/amd_iommu_types.h if (entry->id == id) entry 779 drivers/iommu/amd_iommu_types.h return entry->devid; entry 876 drivers/iommu/amd_iommu_types.h void *entry; /* Pointer to union irte or struct irte_ga */ entry 2718 drivers/iommu/arm-smmu-v3.c struct iommu_resv_region *entry, *next; entry 2720 drivers/iommu/arm-smmu-v3.c list_for_each_entry_safe(entry, next, head, list) entry 2721 drivers/iommu/arm-smmu-v3.c kfree(entry); entry 1549 drivers/iommu/arm-smmu.c struct iommu_resv_region *entry, *next; entry 1551 drivers/iommu/arm-smmu.c list_for_each_entry_safe(entry, next, head, list) entry 1552 drivers/iommu/arm-smmu.c kfree(entry); entry 851 drivers/iommu/dmar.c dmar_validate_one_drhd(struct acpi_dmar_header *entry, void *arg) entry 857 drivers/iommu/dmar.c drhd = (void *)entry; entry 1079 drivers/iommu/exynos-iommu.c sysmmu_pte_t *entry; entry 1089 drivers/iommu/exynos-iommu.c entry = section_entry(domain->pgtable, iova); entry 1092 drivers/iommu/exynos-iommu.c ret = lv1set_section(domain, entry, iova, paddr, prot, entry 1097 drivers/iommu/exynos-iommu.c pent = alloc_lv2entry(domain, entry, iova, entry 1214 drivers/iommu/exynos-iommu.c sysmmu_pte_t *entry; entry 1220 drivers/iommu/exynos-iommu.c entry = section_entry(domain->pgtable, iova); entry 1222 drivers/iommu/exynos-iommu.c if (lv1ent_section(entry)) { entry 1223 drivers/iommu/exynos-iommu.c phys = section_phys(entry) + section_offs(iova); entry 1224 drivers/iommu/exynos-iommu.c } else if (lv1ent_page(entry)) { entry 1225 drivers/iommu/exynos-iommu.c entry = page_entry(entry, iova); entry 1227 drivers/iommu/exynos-iommu.c if (lv2ent_large(entry)) entry 1228 drivers/iommu/exynos-iommu.c phys = lpage_phys(entry) + lpage_offs(iova); entry 1229 drivers/iommu/exynos-iommu.c else if (lv2ent_small(entry)) entry 1230 drivers/iommu/exynos-iommu.c phys = spage_phys(entry) + spage_offs(iova); entry 1296 drivers/iommu/exynos-iommu.c struct sysmmu_drvdata *data, *entry; entry 1315 drivers/iommu/exynos-iommu.c list_for_each_entry(entry, &owner->controllers, owner_node) entry 1316 drivers/iommu/exynos-iommu.c if (entry == data) entry 43 drivers/iommu/hyperv-iommu.c struct IO_APIC_route_entry *entry; entry 54 drivers/iommu/hyperv-iommu.c entry = data->chip_data; entry 55 drivers/iommu/hyperv-iommu.c entry->dest = cfg->dest_apicid; entry 56 drivers/iommu/hyperv-iommu.c entry->vector = cfg->vector; entry 126 drivers/iommu/hyperv-iommu.c struct IO_APIC_route_entry *entry = irq_data->chip_data; entry 128 drivers/iommu/hyperv-iommu.c entry->dest = cfg->dest_apicid; entry 129 drivers/iommu/hyperv-iommu.c entry->vector = cfg->vector; entry 701 drivers/iommu/intel-iommu.c u64 *entry; entry 703 drivers/iommu/intel-iommu.c entry = &root->lo; entry 707 drivers/iommu/intel-iommu.c entry = &root->hi; entry 711 drivers/iommu/intel-iommu.c if (*entry & 1) entry 712 drivers/iommu/intel-iommu.c context = phys_to_virt(*entry & VTD_PAGE_MASK); entry 724 drivers/iommu/intel-iommu.c *entry = phy_addr | 1; entry 725 drivers/iommu/intel-iommu.c __iommu_flush_cache(iommu, entry, sizeof(*entry)); entry 5726 drivers/iommu/intel-iommu.c struct iommu_resv_region *entry, *next; entry 5728 drivers/iommu/intel-iommu.c list_for_each_entry_safe(entry, next, head, list) entry 5729 drivers/iommu/intel-iommu.c kfree(entry); entry 23 drivers/iommu/intel-pasid.h #define is_pasid_enabled(entry) (((entry)->lo >> 3) & 0x1) entry 24 drivers/iommu/intel-pasid.h #define get_pasid_dir_size(entry) (1 << ((((entry)->lo >> 9) & 0x7) + 7)) entry 239 drivers/iommu/intel_irq_remapping.c struct irte *start, *entry, *end; entry 252 drivers/iommu/intel_irq_remapping.c for (entry = start; entry < end; entry++) { entry 253 drivers/iommu/intel_irq_remapping.c set_64bit(&entry->low, 0); entry 254 drivers/iommu/intel_irq_remapping.c set_64bit(&entry->high, 0); entry 1264 drivers/iommu/intel_irq_remapping.c struct IR_IO_APIC_route_entry *entry; entry 1280 drivers/iommu/intel_irq_remapping.c entry = (struct IR_IO_APIC_route_entry *)info->ioapic_entry; entry 1282 drivers/iommu/intel_irq_remapping.c memset(entry, 0, sizeof(*entry)); entry 1283 drivers/iommu/intel_irq_remapping.c entry->index2 = (index >> 15) & 0x1; entry 1284 drivers/iommu/intel_irq_remapping.c entry->zero = 0; entry 1285 drivers/iommu/intel_irq_remapping.c entry->format = 1; entry 1286 drivers/iommu/intel_irq_remapping.c entry->index = (index & 0x7fff); entry 1291 drivers/iommu/intel_irq_remapping.c entry->vector = info->ioapic_pin; entry 1292 drivers/iommu/intel_irq_remapping.c entry->mask = 0; /* enable IRQ */ entry 1293 drivers/iommu/intel_irq_remapping.c entry->trigger = info->ioapic_trigger; entry 1294 drivers/iommu/intel_irq_remapping.c entry->polarity = info->ioapic_polarity; entry 1296 drivers/iommu/intel_irq_remapping.c entry->mask = 1; /* Mask level triggered irqs. */ entry 1437 drivers/iommu/intel_irq_remapping.c struct irte entry; entry 1439 drivers/iommu/intel_irq_remapping.c memset(&entry, 0, sizeof(entry)); entry 1440 drivers/iommu/intel_irq_remapping.c modify_irte(&data->irq_2_iommu, &entry); entry 348 drivers/iommu/iommu.c struct iommu_resv_region *entry; entry 351 drivers/iommu/iommu.c list_for_each_entry(entry, dev_resv_regions, list) { entry 352 drivers/iommu/iommu.c ret = iommu_insert_resv_region(entry, group_resv_regions); entry 627 drivers/iommu/iommu.c struct iommu_resv_region *entry; entry 643 drivers/iommu/iommu.c list_for_each_entry(entry, &mappings, list) { entry 647 drivers/iommu/iommu.c domain->ops->apply_resv_region(dev, domain, entry); entry 649 drivers/iommu/iommu.c start = ALIGN(entry->start, pg_size); entry 650 drivers/iommu/iommu.c end = ALIGN(entry->start + entry->length, pg_size); entry 652 drivers/iommu/iommu.c if (entry->type != IOMMU_RESV_DIRECT && entry 653 drivers/iommu/iommu.c entry->type != IOMMU_RESV_DIRECT_RELAXABLE) entry 663 drivers/iommu/iommu.c ret = iommu_map(domain, addr, addr, pg_size, entry->prot); entry 811 drivers/iommu/iommu.c struct group_device *entry; entry 814 drivers/iommu/iommu.c list_for_each_entry(entry, &group->devices, list) entry 217 drivers/iommu/s390-iommu.c unsigned long *entry; entry 230 drivers/iommu/s390-iommu.c entry = dma_walk_cpu_trans(s390_domain->dma_table, dma_addr); entry 231 drivers/iommu/s390-iommu.c if (!entry) { entry 235 drivers/iommu/s390-iommu.c dma_update_cpu_trans(entry, page_addr, flags); entry 255 drivers/iommu/s390-iommu.c entry = dma_walk_cpu_trans(s390_domain->dma_table, entry 257 drivers/iommu/s390-iommu.c if (!entry) entry 259 drivers/iommu/s390-iommu.c dma_update_cpu_trans(entry, page_addr, flags); entry 812 drivers/iommu/virtio-iommu.c struct iommu_resv_region *entry, *new_entry, *msi = NULL; entry 817 drivers/iommu/virtio-iommu.c list_for_each_entry(entry, &vdev->resv_regions, list) { entry 818 drivers/iommu/virtio-iommu.c if (entry->type == IOMMU_RESV_MSI) entry 819 drivers/iommu/virtio-iommu.c msi = entry; entry 821 drivers/iommu/virtio-iommu.c new_entry = kmemdup(entry, sizeof(*entry), GFP_KERNEL); entry 845 drivers/iommu/virtio-iommu.c struct iommu_resv_region *entry, *next; entry 847 drivers/iommu/virtio-iommu.c list_for_each_entry_safe(entry, next, head, list) entry 848 drivers/iommu/virtio-iommu.c kfree(entry); entry 197 drivers/irqchip/irq-crossbar.c u32 max = 0, entry, reg_size; entry 242 drivers/irqchip/irq-crossbar.c i, &entry); entry 243 drivers/irqchip/irq-crossbar.c if (entry >= max) { entry 248 drivers/irqchip/irq-crossbar.c cb->irq_map[entry] = IRQ_RESERVED; entry 260 drivers/irqchip/irq-crossbar.c i, &entry); entry 261 drivers/irqchip/irq-crossbar.c if (entry >= max) { entry 266 drivers/irqchip/irq-crossbar.c cb->irq_map[entry] = IRQ_SKIP; entry 62 drivers/irqchip/irq-gic-v2m.c struct list_head entry; entry 184 drivers/irqchip/irq-gic-v2m.c list_for_each_entry(tmp, &v2m_nodes, entry) { entry 269 drivers/irqchip/irq-gic-v2m.c list_for_each_entry_safe(v2m, tmp, &v2m_nodes, entry) { entry 270 drivers/irqchip/irq-gic-v2m.c list_del(&v2m->entry); entry 285 drivers/irqchip/irq-gic-v2m.c v2m = list_first_entry_or_null(&v2m_nodes, struct v2m_data, entry); entry 330 drivers/irqchip/irq-gic-v2m.c INIT_LIST_HEAD(&v2m->entry); entry 397 drivers/irqchip/irq-gic-v2m.c list_add_tail(&v2m->entry, &v2m_nodes); entry 468 drivers/irqchip/irq-gic-v2m.c data = list_first_entry_or_null(&v2m_nodes, struct v2m_data, entry); entry 96 drivers/irqchip/irq-gic-v3-its.c struct list_head entry; entry 146 drivers/irqchip/irq-gic-v3-its.c struct list_head entry; entry 183 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) { entry 1003 drivers/irqchip/irq-gic-v3-its.c its = list_first_entry(&its_nodes, struct its_node, entry); entry 1023 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) { entry 1475 drivers/irqchip/irq-gic-v3-its.c struct list_head entry; entry 1500 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry_safe(range, tmp, &lpi_range_list, entry) { entry 1507 drivers/irqchip/irq-gic-v3-its.c list_del(&range->entry); entry 1524 drivers/irqchip/irq-gic-v3-its.c if (&a->entry == &lpi_range_list || &b->entry == &lpi_range_list) entry 1530 drivers/irqchip/irq-gic-v3-its.c list_del(&a->entry); entry 1544 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry_reverse(old, &lpi_range_list, entry) { entry 1555 drivers/irqchip/irq-gic-v3-its.c list_add(&new->entry, &old->entry); entry 1561 drivers/irqchip/irq-gic-v3-its.c merge_lpi_ranges(new, list_next_entry(new, entry)); entry 2276 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) entry 2289 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(tmp, &its->its_device_list, entry) { entry 2380 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) { entry 2455 drivers/irqchip/irq-gic-v3-its.c INIT_LIST_HEAD(&dev->entry); entry 2458 drivers/irqchip/irq-gic-v3-its.c list_add(&dev->entry, &its->its_device_list); entry 2472 drivers/irqchip/irq-gic-v3-its.c list_del(&its_dev->entry); entry 2871 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) { entry 3146 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) { entry 3172 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) { entry 3364 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) { entry 3385 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry_continue_reverse(its, &its_nodes, entry) { entry 3406 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) { entry 3502 drivers/irqchip/irq-gic-v3-its.c its = list_first_entry(&its_nodes, struct its_node, entry); entry 3608 drivers/irqchip/irq-gic-v3-its.c INIT_LIST_HEAD(&its->entry); entry 3695 drivers/irqchip/irq-gic-v3-its.c list_add(&its->entry, &its_nodes); entry 4007 drivers/irqchip/irq-gic-v3-its.c list_for_each_entry(its, &its_nodes, entry) entry 65 drivers/isdn/mISDN/dsp_pipeline.c struct dsp_element_entry *entry = entry 67 drivers/isdn/mISDN/dsp_pipeline.c list_del(&entry->list); entry 68 drivers/isdn/mISDN/dsp_pipeline.c kfree(entry); entry 73 drivers/isdn/mISDN/dsp_pipeline.c struct dsp_element_entry *entry; entry 79 drivers/isdn/mISDN/dsp_pipeline.c entry = kzalloc(sizeof(struct dsp_element_entry), GFP_ATOMIC); entry 80 drivers/isdn/mISDN/dsp_pipeline.c if (!entry) entry 83 drivers/isdn/mISDN/dsp_pipeline.c entry->elem = elem; entry 85 drivers/isdn/mISDN/dsp_pipeline.c entry->dev.class = elements_class; entry 86 drivers/isdn/mISDN/dsp_pipeline.c entry->dev.release = mISDN_dsp_dev_release; entry 87 drivers/isdn/mISDN/dsp_pipeline.c dev_set_drvdata(&entry->dev, elem); entry 88 drivers/isdn/mISDN/dsp_pipeline.c dev_set_name(&entry->dev, "%s", elem->name); entry 89 drivers/isdn/mISDN/dsp_pipeline.c ret = device_register(&entry->dev); entry 95 drivers/isdn/mISDN/dsp_pipeline.c list_add_tail(&entry->list, &dsp_elements); entry 98 drivers/isdn/mISDN/dsp_pipeline.c ret = device_create_file(&entry->dev, entry 114 drivers/isdn/mISDN/dsp_pipeline.c device_unregister(&entry->dev); entry 117 drivers/isdn/mISDN/dsp_pipeline.c kfree(entry); entry 124 drivers/isdn/mISDN/dsp_pipeline.c struct dsp_element_entry *entry, *n; entry 129 drivers/isdn/mISDN/dsp_pipeline.c list_for_each_entry_safe(entry, n, &dsp_elements, list) entry 130 drivers/isdn/mISDN/dsp_pipeline.c if (entry->elem == elem) { entry 131 drivers/isdn/mISDN/dsp_pipeline.c device_unregister(&entry->dev); entry 159 drivers/isdn/mISDN/dsp_pipeline.c struct dsp_element_entry *entry, *n; entry 165 drivers/isdn/mISDN/dsp_pipeline.c list_for_each_entry_safe(entry, n, &dsp_elements, list) { entry 166 drivers/isdn/mISDN/dsp_pipeline.c list_del(&entry->list); entry 168 drivers/isdn/mISDN/dsp_pipeline.c __func__, entry->elem->name); entry 169 drivers/isdn/mISDN/dsp_pipeline.c kfree(entry); entry 193 drivers/isdn/mISDN/dsp_pipeline.c struct dsp_pipeline_entry *entry, *n; entry 195 drivers/isdn/mISDN/dsp_pipeline.c list_for_each_entry_safe(entry, n, &pipeline->list, list) { entry 196 drivers/isdn/mISDN/dsp_pipeline.c list_del(&entry->list); entry 197 drivers/isdn/mISDN/dsp_pipeline.c if (entry->elem == dsp_hwec) entry 201 drivers/isdn/mISDN/dsp_pipeline.c entry->elem->free(entry->p); entry 202 drivers/isdn/mISDN/dsp_pipeline.c kfree(entry); entry 223 drivers/isdn/mISDN/dsp_pipeline.c struct dsp_element_entry *entry, *n; entry 244 drivers/isdn/mISDN/dsp_pipeline.c list_for_each_entry_safe(entry, n, &dsp_elements, list) entry 245 drivers/isdn/mISDN/dsp_pipeline.c if (!strcmp(entry->elem->name, name)) { entry 246 drivers/isdn/mISDN/dsp_pipeline.c elem = entry->elem; entry 316 drivers/isdn/mISDN/dsp_pipeline.c struct dsp_pipeline_entry *entry; entry 321 drivers/isdn/mISDN/dsp_pipeline.c list_for_each_entry(entry, &pipeline->list, list) entry 322 drivers/isdn/mISDN/dsp_pipeline.c if (entry->elem->process_tx) entry 323 drivers/isdn/mISDN/dsp_pipeline.c entry->elem->process_tx(entry->p, data, len); entry 329 drivers/isdn/mISDN/dsp_pipeline.c struct dsp_pipeline_entry *entry; entry 334 drivers/isdn/mISDN/dsp_pipeline.c list_for_each_entry_reverse(entry, &pipeline->list, list) entry 335 drivers/isdn/mISDN/dsp_pipeline.c if (entry->elem->process_rx) entry 336 drivers/isdn/mISDN/dsp_pipeline.c entry->elem->process_rx(entry->p, data, len, txlen); entry 114 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry; entry 138 drivers/lightnvm/pblk-rb.c entry = &rb->entries[init_entry]; entry 139 drivers/lightnvm/pblk-rb.c entry->data = kaddr; entry 140 drivers/lightnvm/pblk-rb.c entry->cacheline = pblk_cacheline_to_addr(init_entry++); entry 141 drivers/lightnvm/pblk-rb.c entry->w_ctx.flags = PBLK_WRITABLE_ENTRY; entry 145 drivers/lightnvm/pblk-rb.c entry = &rb->entries[init_entry]; entry 146 drivers/lightnvm/pblk-rb.c entry->cacheline = pblk_cacheline_to_addr(init_entry++); entry 147 drivers/lightnvm/pblk-rb.c entry->data = kaddr + (i * rb->seg_size); entry 148 drivers/lightnvm/pblk-rb.c entry->w_ctx.flags = PBLK_WRITABLE_ENTRY; entry 149 drivers/lightnvm/pblk-rb.c bio_list_init(&entry->w_ctx.bios); entry 241 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry; entry 248 drivers/lightnvm/pblk-rb.c entry = &rb->entries[rb->l2p_update]; entry 249 drivers/lightnvm/pblk-rb.c w_ctx = &entry->w_ctx; entry 251 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(entry->w_ctx.flags); entry 260 drivers/lightnvm/pblk-rb.c entry->cacheline); entry 329 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry) entry 331 drivers/lightnvm/pblk-rb.c memcpy(entry->data, data, rb->seg_size); entry 333 drivers/lightnvm/pblk-rb.c entry->w_ctx.lba = w_ctx.lba; entry 334 drivers/lightnvm/pblk-rb.c entry->w_ctx.ppa = w_ctx.ppa; entry 341 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry; entry 344 drivers/lightnvm/pblk-rb.c entry = &rb->entries[ring_pos]; entry 345 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(entry->w_ctx.flags); entry 351 drivers/lightnvm/pblk-rb.c __pblk_rb_write_entry(rb, data, w_ctx, entry); entry 353 drivers/lightnvm/pblk-rb.c pblk_update_map_cache(pblk, w_ctx.lba, entry->cacheline); entry 357 drivers/lightnvm/pblk-rb.c smp_store_release(&entry->w_ctx.flags, flags); entry 365 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry; entry 368 drivers/lightnvm/pblk-rb.c entry = &rb->entries[ring_pos]; entry 369 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(entry->w_ctx.flags); entry 375 drivers/lightnvm/pblk-rb.c __pblk_rb_write_entry(rb, data, w_ctx, entry); entry 377 drivers/lightnvm/pblk-rb.c if (!pblk_update_map_gc(pblk, w_ctx.lba, entry->cacheline, line, paddr)) entry 378 drivers/lightnvm/pblk-rb.c entry->w_ctx.lba = ADDR_EMPTY; entry 383 drivers/lightnvm/pblk-rb.c smp_store_release(&entry->w_ctx.flags, flags); entry 389 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry; entry 405 drivers/lightnvm/pblk-rb.c entry = &rb->entries[flush_point]; entry 411 drivers/lightnvm/pblk-rb.c bio_list_add(&entry->w_ctx.bios, bio); entry 558 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry; entry 577 drivers/lightnvm/pblk-rb.c entry = &rb->entries[pos]; entry 583 drivers/lightnvm/pblk-rb.c flags = READ_ONCE(entry->w_ctx.flags); entry 589 drivers/lightnvm/pblk-rb.c page = virt_to_page(entry->data); entry 595 drivers/lightnvm/pblk-rb.c smp_store_release(&entry->w_ctx.flags, flags); entry 605 drivers/lightnvm/pblk-rb.c smp_store_release(&entry->w_ctx.flags, flags); entry 613 drivers/lightnvm/pblk-rb.c smp_store_release(&entry->w_ctx.flags, flags); entry 648 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry; entry 661 drivers/lightnvm/pblk-rb.c entry = &rb->entries[pos]; entry 662 drivers/lightnvm/pblk-rb.c w_ctx = &entry->w_ctx; entry 677 drivers/lightnvm/pblk-rb.c memcpy(data, entry->data, rb->seg_size); entry 686 drivers/lightnvm/pblk-rb.c unsigned int entry = pblk_rb_ptr_wrap(rb, pos, 0); entry 688 drivers/lightnvm/pblk-rb.c return &rb->entries[entry].w_ctx; entry 765 drivers/lightnvm/pblk-rb.c struct pblk_rb_entry *entry; entry 784 drivers/lightnvm/pblk-rb.c entry = &rb->entries[i]; entry 786 drivers/lightnvm/pblk-rb.c if (!entry->data) { entry 151 drivers/lightnvm/pblk-write.c struct pblk_rb_entry *entry; entry 160 drivers/lightnvm/pblk-write.c entry = &rb->entries[pblk_rb_ptr_wrap(rb, sentry, i)]; entry 161 drivers/lightnvm/pblk-write.c w_ctx = &entry->w_ctx; entry 166 drivers/lightnvm/pblk-write.c if (!pblk_ppa_comp(ppa_l2p, entry->cacheline)) entry 142 drivers/mailbox/tegra-hsp.c struct tegra_hsp_doorbell *entry; entry 144 drivers/mailbox/tegra-hsp.c list_for_each_entry(entry, &hsp->doorbells, list) entry 145 drivers/mailbox/tegra-hsp.c if (entry->master == master) entry 146 drivers/mailbox/tegra-hsp.c return entry; entry 336 drivers/md/dm-bio-prison-v1.c struct dm_deferred_entry *entry; entry 339 drivers/md/dm-bio-prison-v1.c entry = ds->entries + ds->current_entry; entry 340 drivers/md/dm-bio-prison-v1.c entry->count++; entry 343 drivers/md/dm-bio-prison-v1.c return entry; entry 364 drivers/md/dm-bio-prison-v1.c void dm_deferred_entry_dec(struct dm_deferred_entry *entry, struct list_head *head) entry 368 drivers/md/dm-bio-prison-v1.c spin_lock_irqsave(&entry->ds->lock, flags); entry 369 drivers/md/dm-bio-prison-v1.c BUG_ON(!entry->count); entry 370 drivers/md/dm-bio-prison-v1.c --entry->count; entry 371 drivers/md/dm-bio-prison-v1.c __sweep(entry->ds, head); entry 372 drivers/md/dm-bio-prison-v1.c spin_unlock_irqrestore(&entry->ds->lock, flags); entry 133 drivers/md/dm-bio-prison-v1.h void dm_deferred_entry_dec(struct dm_deferred_entry *entry, struct list_head *head); entry 61 drivers/md/dm-cache-policy-smq.c struct entry *begin; entry 62 drivers/md/dm-cache-policy-smq.c struct entry *end; entry 72 drivers/md/dm-cache-policy-smq.c es->begin = vzalloc(array_size(nr_entries, sizeof(struct entry))); entry 85 drivers/md/dm-cache-policy-smq.c static struct entry *__get_entry(struct entry_space *es, unsigned block) entry 87 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 95 drivers/md/dm-cache-policy-smq.c static unsigned to_index(struct entry_space *es, struct entry *e) entry 101 drivers/md/dm-cache-policy-smq.c static struct entry *to_entry(struct entry_space *es, unsigned block) entry 122 drivers/md/dm-cache-policy-smq.c static struct entry *l_head(struct entry_space *es, struct ilist *l) entry 127 drivers/md/dm-cache-policy-smq.c static struct entry *l_tail(struct entry_space *es, struct ilist *l) entry 132 drivers/md/dm-cache-policy-smq.c static struct entry *l_next(struct entry_space *es, struct entry *e) entry 137 drivers/md/dm-cache-policy-smq.c static struct entry *l_prev(struct entry_space *es, struct entry *e) entry 147 drivers/md/dm-cache-policy-smq.c static void l_add_head(struct entry_space *es, struct ilist *l, struct entry *e) entry 149 drivers/md/dm-cache-policy-smq.c struct entry *head = l_head(es, l); entry 163 drivers/md/dm-cache-policy-smq.c static void l_add_tail(struct entry_space *es, struct ilist *l, struct entry *e) entry 165 drivers/md/dm-cache-policy-smq.c struct entry *tail = l_tail(es, l); entry 180 drivers/md/dm-cache-policy-smq.c struct entry *old, struct entry *e) entry 182 drivers/md/dm-cache-policy-smq.c struct entry *prev = l_prev(es, old); entry 197 drivers/md/dm-cache-policy-smq.c static void l_del(struct entry_space *es, struct ilist *l, struct entry *e) entry 199 drivers/md/dm-cache-policy-smq.c struct entry *prev = l_prev(es, e); entry 200 drivers/md/dm-cache-policy-smq.c struct entry *next = l_next(es, e); entry 216 drivers/md/dm-cache-policy-smq.c static struct entry *l_pop_head(struct entry_space *es, struct ilist *l) entry 218 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 229 drivers/md/dm-cache-policy-smq.c static struct entry *l_pop_tail(struct entry_space *es, struct ilist *l) entry 231 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 295 drivers/md/dm-cache-policy-smq.c static void q_push(struct queue *q, struct entry *e) entry 305 drivers/md/dm-cache-policy-smq.c static void q_push_front(struct queue *q, struct entry *e) entry 315 drivers/md/dm-cache-policy-smq.c static void q_push_before(struct queue *q, struct entry *old, struct entry *e) entry 325 drivers/md/dm-cache-policy-smq.c static void q_del(struct queue *q, struct entry *e) entry 335 drivers/md/dm-cache-policy-smq.c static struct entry *q_peek(struct queue *q, unsigned max_level, bool can_cross_sentinel) entry 338 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 357 drivers/md/dm-cache-policy-smq.c static struct entry *q_pop(struct queue *q) entry 359 drivers/md/dm-cache-policy-smq.c struct entry *e = q_peek(q, q->nr_levels, true); entry 372 drivers/md/dm-cache-policy-smq.c static struct entry *__redist_pop_from(struct queue *q, unsigned level) entry 374 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 431 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 470 drivers/md/dm-cache-policy-smq.c static void q_requeue(struct queue *q, struct entry *e, unsigned extra_levels, entry 471 drivers/md/dm-cache-policy-smq.c struct entry *s1, struct entry *s2) entry 473 drivers/md/dm-cache-policy-smq.c struct entry *de; entry 606 drivers/md/dm-cache-policy-smq.c static struct entry *h_head(struct smq_hash_table *ht, unsigned bucket) entry 611 drivers/md/dm-cache-policy-smq.c static struct entry *h_next(struct smq_hash_table *ht, struct entry *e) entry 616 drivers/md/dm-cache-policy-smq.c static void __h_insert(struct smq_hash_table *ht, unsigned bucket, struct entry *e) entry 622 drivers/md/dm-cache-policy-smq.c static void h_insert(struct smq_hash_table *ht, struct entry *e) entry 628 drivers/md/dm-cache-policy-smq.c static struct entry *__h_lookup(struct smq_hash_table *ht, unsigned h, dm_oblock_t oblock, entry 629 drivers/md/dm-cache-policy-smq.c struct entry **prev) entry 631 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 645 drivers/md/dm-cache-policy-smq.c struct entry *e, struct entry *prev) entry 656 drivers/md/dm-cache-policy-smq.c static struct entry *h_lookup(struct smq_hash_table *ht, dm_oblock_t oblock) entry 658 drivers/md/dm-cache-policy-smq.c struct entry *e, *prev; entry 674 drivers/md/dm-cache-policy-smq.c static void h_remove(struct smq_hash_table *ht, struct entry *e) entry 677 drivers/md/dm-cache-policy-smq.c struct entry *prev; entry 712 drivers/md/dm-cache-policy-smq.c static void init_entry(struct entry *e) entry 728 drivers/md/dm-cache-policy-smq.c static struct entry *alloc_entry(struct entry_alloc *ea) entry 730 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 745 drivers/md/dm-cache-policy-smq.c static struct entry *alloc_particular_entry(struct entry_alloc *ea, unsigned i) entry 747 drivers/md/dm-cache-policy-smq.c struct entry *e = __get_entry(ea->es, ea->begin + i); entry 758 drivers/md/dm-cache-policy-smq.c static void free_entry(struct entry_alloc *ea, struct entry *e) entry 773 drivers/md/dm-cache-policy-smq.c static unsigned get_index(struct entry_alloc *ea, struct entry *e) entry 778 drivers/md/dm-cache-policy-smq.c static struct entry *get_entry(struct entry_alloc *ea, unsigned index) entry 862 drivers/md/dm-cache-policy-smq.c static struct entry *get_sentinel(struct entry_alloc *ea, unsigned level, bool which) entry 867 drivers/md/dm-cache-policy-smq.c static struct entry *writeback_sentinel(struct smq_policy *mq, unsigned level) entry 872 drivers/md/dm-cache-policy-smq.c static struct entry *demote_sentinel(struct smq_policy *mq, unsigned level) entry 881 drivers/md/dm-cache-policy-smq.c struct entry *sentinel; entry 894 drivers/md/dm-cache-policy-smq.c struct entry *sentinel; entry 921 drivers/md/dm-cache-policy-smq.c struct entry *sentinel; entry 950 drivers/md/dm-cache-policy-smq.c static void del_queue(struct smq_policy *mq, struct entry *e) entry 955 drivers/md/dm-cache-policy-smq.c static void push_queue(struct smq_policy *mq, struct entry *e) entry 964 drivers/md/dm-cache-policy-smq.c static void push(struct smq_policy *mq, struct entry *e) entry 971 drivers/md/dm-cache-policy-smq.c static void push_queue_front(struct smq_policy *mq, struct entry *e) entry 979 drivers/md/dm-cache-policy-smq.c static void push_front(struct smq_policy *mq, struct entry *e) entry 986 drivers/md/dm-cache-policy-smq.c static dm_cblock_t infer_cblock(struct smq_policy *mq, struct entry *e) entry 991 drivers/md/dm-cache-policy-smq.c static void requeue(struct smq_policy *mq, struct entry *e) entry 1160 drivers/md/dm-cache-policy-smq.c static void mark_pending(struct smq_policy *mq, struct entry *e) entry 1168 drivers/md/dm-cache-policy-smq.c static void clear_pending(struct smq_policy *mq, struct entry *e) entry 1178 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 1201 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 1230 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 1280 drivers/md/dm-cache-policy-smq.c static enum promote_result should_promote(struct smq_policy *mq, struct entry *hs_e, entry 1299 drivers/md/dm-cache-policy-smq.c static struct entry *update_hotspot_queue(struct smq_policy *mq, dm_oblock_t b) entry 1303 drivers/md/dm-cache-policy-smq.c struct entry *e = h_lookup(&mq->hotspot_table, hb); entry 1369 drivers/md/dm-cache-policy-smq.c struct entry *e, *hs_e; entry 1462 drivers/md/dm-cache-policy-smq.c struct entry *e = get_entry(&mq->cache_alloc, entry 1519 drivers/md/dm-cache-policy-smq.c struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); entry 1560 drivers/md/dm-cache-policy-smq.c struct entry *e; entry 1580 drivers/md/dm-cache-policy-smq.c struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); entry 1595 drivers/md/dm-cache-policy-smq.c struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); entry 198 drivers/md/dm-cache-target.c list_for_each_entry_safe(ws, tmp, &work_items, entry) { entry 201 drivers/md/dm-cache-target.c INIT_LIST_HEAD(&ws->entry); /* to avoid a WARN_ON */ entry 246 drivers/md/dm-cache-target.c list_add_tail(&k->ws.entry, &b->work_items); entry 211 drivers/md/dm-log-writes.c static int write_metadata(struct log_writes_c *lc, void *entry, entry 241 drivers/md/dm-log-writes.c memcpy(ptr, entry, entrylen); entry 263 drivers/md/dm-log-writes.c static int write_inline_data(struct log_writes_c *lc, void *entry, entry 335 drivers/md/dm-log-writes.c struct log_write_entry entry; entry 339 drivers/md/dm-log-writes.c entry.sector = cpu_to_le64(block->sector); entry 340 drivers/md/dm-log-writes.c entry.nr_sectors = cpu_to_le64(block->nr_sectors); entry 341 drivers/md/dm-log-writes.c entry.flags = cpu_to_le64(block->flags); entry 342 drivers/md/dm-log-writes.c entry.data_len = cpu_to_le64(block->datalen); entry 345 drivers/md/dm-log-writes.c if (write_metadata(lc, &entry, sizeof(entry), block->data, entry 354 drivers/md/dm-log-writes.c if (write_inline_data(lc, &entry, sizeof(entry), block->data, entry 514 drivers/md/dm-stats.c static void dm_stat_for_entry(struct dm_stat *s, size_t entry, entry 519 drivers/md/dm-stats.c struct dm_stat_shared *shared = &s->stat_shared[entry]; entry 543 drivers/md/dm-stats.c p = &s->stat_percpu[smp_processor_id()][entry]; entry 590 drivers/md/dm-stats.c size_t entry; entry 605 drivers/md/dm-stats.c entry = rel_sector; entry 607 drivers/md/dm-stats.c if (WARN_ON_ONCE(entry >= s->n_entries)) { entry 614 drivers/md/dm-stats.c dm_stat_for_entry(s, entry, bi_rw, fragment_len, entry 617 drivers/md/dm-stats.c entry++; entry 3485 drivers/md/md.c struct rdev_sysfs_entry *entry = container_of(attr, struct rdev_sysfs_entry, attr); entry 3488 drivers/md/md.c if (!entry->show) entry 3492 drivers/md/md.c return entry->show(rdev, page); entry 3499 drivers/md/md.c struct rdev_sysfs_entry *entry = container_of(attr, struct rdev_sysfs_entry, attr); entry 3504 drivers/md/md.c if (!entry->store) entry 3513 drivers/md/md.c rv = entry->store(rdev, page, length); entry 5325 drivers/md/md.c struct md_sysfs_entry *entry = container_of(attr, struct md_sysfs_entry, attr); entry 5329 drivers/md/md.c if (!entry->show) entry 5339 drivers/md/md.c rv = entry->show(mddev, page); entry 5348 drivers/md/md.c struct md_sysfs_entry *entry = container_of(attr, struct md_sysfs_entry, attr); entry 5352 drivers/md/md.c if (!entry->store) entry 5363 drivers/md/md.c rv = entry->store(mddev, page, length); entry 99 drivers/md/persistent-data/dm-array.c unsigned char *entry = (unsigned char *) (ab + 1); entry 101 drivers/md/persistent-data/dm-array.c entry += index * info->value_type.size; entry 103 drivers/md/persistent-data/dm-array.c return entry; entry 775 drivers/md/persistent-data/dm-array.c unsigned entry, max_entries; entry 784 drivers/md/persistent-data/dm-array.c entry = index % max_entries; entry 785 drivers/md/persistent-data/dm-array.c if (entry >= le32_to_cpu(ab->nr_entries)) entry 788 drivers/md/persistent-data/dm-array.c memcpy(value_le, element_at(info, ab, entry), entry 804 drivers/md/persistent-data/dm-array.c unsigned entry; entry 816 drivers/md/persistent-data/dm-array.c entry = index % max_entries; entry 817 drivers/md/persistent-data/dm-array.c if (entry >= le32_to_cpu(ab->nr_entries)) { entry 822 drivers/md/persistent-data/dm-array.c old_value = element_at(info, ab, entry); entry 105 drivers/media/cec/cec-adap.c struct cec_event_entry *entry; entry 116 drivers/media/cec/cec-adap.c entry = &fh->core_events[ev_idx]; entry 118 drivers/media/cec/cec-adap.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 119 drivers/media/cec/cec-adap.c if (entry) { entry 122 drivers/media/cec/cec-adap.c entry->ev.lost_msgs.lost_msgs += entry 126 drivers/media/cec/cec-adap.c entry->ev = *new_ev; entry 127 drivers/media/cec/cec-adap.c entry->ev.ts = ts; entry 131 drivers/media/cec/cec-adap.c list_add_tail(&entry->list, &fh->events[ev_idx]); entry 138 drivers/media/cec/cec-adap.c list_add_tail(&entry->list, &fh->events[ev_idx]); entry 140 drivers/media/cec/cec-adap.c entry = list_first_entry(&fh->events[ev_idx], entry 142 drivers/media/cec/cec-adap.c list_del(&entry->list); entry 143 drivers/media/cec/cec-adap.c kfree(entry); entry 147 drivers/media/cec/cec-adap.c entry = list_first_entry_or_null(&fh->events[ev_idx], entry 149 drivers/media/cec/cec-adap.c if (entry) entry 150 drivers/media/cec/cec-adap.c entry->ev.flags |= CEC_EVENT_FL_DROPPED_EVENTS; entry 237 drivers/media/cec/cec-adap.c struct cec_msg_entry *entry; entry 240 drivers/media/cec/cec-adap.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 241 drivers/media/cec/cec-adap.c if (entry) { entry 242 drivers/media/cec/cec-adap.c entry->msg = *msg; entry 244 drivers/media/cec/cec-adap.c list_add_tail(&entry->list, &fh->msgs); entry 258 drivers/media/cec/cec-adap.c entry = list_first_entry(&fh->msgs, struct cec_msg_entry, list); entry 259 drivers/media/cec/cec-adap.c list_del(&entry->list); entry 260 drivers/media/cec/cec-adap.c kfree(entry); entry 229 drivers/media/cec/cec-api.c struct cec_msg_entry *entry = entry 233 drivers/media/cec/cec-api.c list_del(&entry->list); entry 234 drivers/media/cec/cec-api.c *msg = entry->msg; entry 235 drivers/media/cec/cec-api.c kfree(entry); entry 308 drivers/media/cec/cec-api.c struct cec_event_entry *entry = entry 312 drivers/media/cec/cec-api.c if (entry && entry->ev.ts <= ts) { entry 313 drivers/media/cec/cec-api.c ev = entry; entry 650 drivers/media/cec/cec-api.c struct cec_msg_entry *entry = entry 653 drivers/media/cec/cec-api.c list_del(&entry->list); entry 654 drivers/media/cec/cec-api.c kfree(entry); entry 658 drivers/media/cec/cec-api.c struct cec_event_entry *entry = entry 662 drivers/media/cec/cec-api.c list_del(&entry->list); entry 663 drivers/media/cec/cec-api.c kfree(entry); entry 31 drivers/media/common/siano/smscoreapi.c struct list_head entry; entry 36 drivers/media/common/siano/smscoreapi.c struct list_head entry; entry 42 drivers/media/common/siano/smscoreapi.c struct list_head entry; entry 409 drivers/media/common/siano/smscoreapi.c struct list_head entry; entry 429 drivers/media/common/siano/smscoreapi.c struct smscore_registry_entry_t *entry; entry 436 drivers/media/common/siano/smscoreapi.c entry = (struct smscore_registry_entry_t *) next; entry 437 drivers/media/common/siano/smscoreapi.c if (!strncmp(entry->devpath, devpath, sizeof(entry->devpath))) { entry 439 drivers/media/common/siano/smscoreapi.c return entry; entry 442 drivers/media/common/siano/smscoreapi.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 443 drivers/media/common/siano/smscoreapi.c if (entry) { entry 444 drivers/media/common/siano/smscoreapi.c entry->mode = default_mode; entry 445 drivers/media/common/siano/smscoreapi.c strscpy(entry->devpath, devpath, sizeof(entry->devpath)); entry 446 drivers/media/common/siano/smscoreapi.c list_add(&entry->entry, &g_smscore_registry); entry 450 drivers/media/common/siano/smscoreapi.c return entry; entry 455 drivers/media/common/siano/smscoreapi.c struct smscore_registry_entry_t *entry; entry 457 drivers/media/common/siano/smscoreapi.c entry = smscore_find_registry(devpath); entry 458 drivers/media/common/siano/smscoreapi.c if (entry) entry 459 drivers/media/common/siano/smscoreapi.c return entry->mode; entry 469 drivers/media/common/siano/smscoreapi.c struct smscore_registry_entry_t *entry; entry 471 drivers/media/common/siano/smscoreapi.c entry = smscore_find_registry(devpath); entry 472 drivers/media/common/siano/smscoreapi.c if (entry) entry 473 drivers/media/common/siano/smscoreapi.c return entry->type; entry 482 drivers/media/common/siano/smscoreapi.c struct smscore_registry_entry_t *entry; entry 484 drivers/media/common/siano/smscoreapi.c entry = smscore_find_registry(devpath); entry 485 drivers/media/common/siano/smscoreapi.c if (entry) entry 486 drivers/media/common/siano/smscoreapi.c entry->mode = mode; entry 494 drivers/media/common/siano/smscoreapi.c struct smscore_registry_entry_t *entry; entry 496 drivers/media/common/siano/smscoreapi.c entry = smscore_find_registry(devpath); entry 497 drivers/media/common/siano/smscoreapi.c if (entry) entry 498 drivers/media/common/siano/smscoreapi.c entry->type = type; entry 545 drivers/media/common/siano/smscoreapi.c list_add(¬ifyee->entry, &g_smscore_notifyees); entry 577 drivers/media/common/siano/smscoreapi.c list_del(¬ifyee->entry); entry 605 drivers/media/common/siano/smscoreapi.c list_for_each_entry(elem, &g_smscore_notifyees, entry) { entry 660 drivers/media/common/siano/smscoreapi.c INIT_LIST_HEAD(&dev->entry); entry 736 drivers/media/common/siano/smscoreapi.c list_add(&dev->entry, &g_smscore_devices); entry 1214 drivers/media/common/siano/smscoreapi.c list_del(&cb->entry); entry 1245 drivers/media/common/siano/smscoreapi.c list_del(&coredev->entry); entry 1455 drivers/media/common/siano/smscoreapi.c list_for_each_entry(client, first, entry) { entry 1457 drivers/media/common/siano/smscoreapi.c list_for_each_entry(client_id, firstid, entry) { entry 1640 drivers/media/common/siano/smscoreapi.c list_del(&cb->entry); entry 1667 drivers/media/common/siano/smscoreapi.c list_add_locked(&cb->entry, &coredev->buffers, &coredev->bufferslock); entry 1696 drivers/media/common/siano/smscoreapi.c list_add_locked(&listentry->entry, &client->idlist, entry 1736 drivers/media/common/siano/smscoreapi.c list_add_locked(&newclient->entry, &coredev->clients, entry 1766 drivers/media/common/siano/smscoreapi.c list_del(&identry->entry); entry 1772 drivers/media/common/siano/smscoreapi.c list_del(&client->entry); entry 2142 drivers/media/common/siano/smscoreapi.c list_del(¬ifyee->entry); entry 2149 drivers/media/common/siano/smscoreapi.c struct smscore_registry_entry_t *entry = entry 2153 drivers/media/common/siano/smscoreapi.c list_del(&entry->entry); entry 2154 drivers/media/common/siano/smscoreapi.c kfree(entry); entry 114 drivers/media/common/siano/smscoreapi.h struct list_head entry; entry 153 drivers/media/common/siano/smscoreapi.h struct list_head entry; entry 619 drivers/media/common/siano/smsdvb-main.c list_del(&client->entry); entry 1156 drivers/media/common/siano/smsdvb-main.c list_add(&client->entry, &g_smsdvb_clients); entry 21 drivers/media/common/siano/smsdvb.h struct list_head entry; entry 660 drivers/media/dvb-core/dvb_demux.c struct dvb_demux_feed *entry; entry 662 drivers/media/dvb-core/dvb_demux.c list_for_each_entry(entry, &feed->demux->feed_list, list_head) entry 663 drivers/media/dvb-core/dvb_demux.c if (entry == feed) entry 814 drivers/media/dvb-core/dvbdev.c struct list_head *entry; entry 815 drivers/media/dvb-core/dvbdev.c list_for_each(entry, &dvb_adapter_list) { entry 817 drivers/media/dvb-core/dvbdev.c adap = list_entry(entry, struct dvb_adapter, list_head); entry 618 drivers/media/i2c/m5mols/m5mols_core.c fd->entry[0].length = info->cap.buf_size; entry 619 drivers/media/i2c/m5mols/m5mols_core.c fd->entry[0].pixelcode = info->ffmt[M5MOLS_RESTYPE_CAPTURE].code; entry 622 drivers/media/i2c/m5mols/m5mols_core.c fd->entry[0].flags = V4L2_MBUS_FRAME_DESC_FL_LEN_MAX; entry 637 drivers/media/i2c/m5mols/m5mols_core.c fd->entry[0].flags = V4L2_MBUS_FRAME_DESC_FL_LEN_MAX; entry 639 drivers/media/i2c/m5mols/m5mols_core.c fd->entry[0].length = clamp_t(u32, fd->entry[0].length, entry 643 drivers/media/i2c/m5mols/m5mols_core.c info->cap.buf_size = fd->entry[0].length; entry 1156 drivers/media/i2c/s5c73m3/s5c73m3-core.c fd->entry[i] = state->frame_desc.entry[i]; entry 1172 drivers/media/i2c/s5c73m3/s5c73m3-core.c fd->entry[0].length = 10 * SZ_1M; entry 1173 drivers/media/i2c/s5c73m3/s5c73m3-core.c fd->entry[1].length = max_t(u32, fd->entry[1].length, entry 1179 drivers/media/i2c/s5c73m3/s5c73m3-core.c frame_desc->entry[i] = fd->entry[i]; entry 679 drivers/media/pci/cx18/cx18-ioctl.c e_idx = &idx->entry[idx->entries]; entry 699 drivers/media/pci/cx18/cx18-ioctl.c e_idx = &idx->entry[idx->entries]; entry 137 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct cio2_fbpt_entry entry[CIO2_MAX_LOPS]) entry 151 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[0].first_entry.ctrl = CIO2_FBPT_CTRL_VALID | entry 158 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[CIO2_MAX_LOPS]) entry 162 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[0].first_entry.first_page_offset = 0; entry 163 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[1].second_entry.num_of_pages = entry 165 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[1].second_entry.last_page_available_bytes = CIO2_PAGE_SIZE - 1; entry 168 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[i].lop_page_addr = cio2->dummy_lop_bus_addr >> PAGE_SHIFT; entry 170 drivers/media/pci/intel/ipu3/ipu3-cio2.c cio2_fbpt_entry_enable(cio2, entry); entry 177 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[CIO2_MAX_LOPS]) entry 183 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[0].first_entry.first_page_offset = b->offset; entry 184 drivers/media/pci/intel/ipu3/ipu3-cio2.c remaining = length + entry[0].first_entry.first_page_offset; entry 185 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[1].second_entry.num_of_pages = entry 195 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[1].second_entry.last_page_available_bytes = entry 203 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry->lop_page_addr = b->lop_bus_addr[i] >> PAGE_SHIFT; entry 205 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry++; entry 212 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry->lop_page_addr = cio2->dummy_lop_bus_addr >> PAGE_SHIFT; entry 214 drivers/media/pci/intel/ipu3/ipu3-cio2.c cio2_fbpt_entry_enable(cio2, entry); entry 558 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct cio2_fbpt_entry *const entry = entry 562 drivers/media/pci/intel/ipu3/ipu3-cio2.c if (entry->first_entry.ctrl & CIO2_FBPT_CTRL_VALID) entry 567 drivers/media/pci/intel/ipu3/ipu3-cio2.c unsigned int bytes = entry[1].second_entry.num_of_bytes; entry 584 drivers/media/pci/intel/ipu3/ipu3-cio2.c cio2_fbpt_entry_init_dummy(cio2, entry); entry 904 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct cio2_fbpt_entry *entry; entry 951 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry = &q->fbpt[next * CIO2_MAX_LOPS]; entry 952 drivers/media/pci/intel/ipu3/ipu3-cio2.c cio2_fbpt_entry_init_buf(cio2, b, entry); entry 1257 drivers/media/pci/ivtv/ivtv-ioctl.c struct v4l2_enc_idx_entry *e = idx->entry; entry 277 drivers/media/pci/saa7134/saa7134-core.c list_add_tail(&buf->entry, &q->queue); entry 280 drivers/media/pci/saa7134/saa7134-core.c entry); entry 285 drivers/media/pci/saa7134/saa7134-core.c list_add_tail(&buf->entry, &q->queue); entry 314 drivers/media/pci/saa7134/saa7134-core.c buf = list_entry(q->queue.next, struct saa7134_buf, entry); entry 317 drivers/media/pci/saa7134/saa7134-core.c list_del(&buf->entry); entry 319 drivers/media/pci/saa7134/saa7134-core.c next = list_entry(q->queue.next, struct saa7134_buf, entry); entry 364 drivers/media/pci/saa7134/saa7134-core.c tmp = list_entry(pos, struct saa7134_buf, entry); entry 1394 drivers/media/pci/saa7134/saa7134-core.c entry); entry 137 drivers/media/pci/saa7134/saa7134-ts.c list_for_each_entry_safe(buf, tmp, &dmaq->queue, entry) { entry 138 drivers/media/pci/saa7134/saa7134-ts.c list_del(&buf->entry); entry 989 drivers/media/pci/saa7134/saa7134-video.c list_for_each_entry_safe(buf, tmp, &dmaq->queue, entry) { entry 990 drivers/media/pci/saa7134/saa7134-video.c list_del(&buf->entry); entry 473 drivers/media/pci/saa7134/saa7134.h struct list_head entry; entry 872 drivers/media/platform/exynos4-is/fimc-capture.c fd.entry[i].length = plane_fmt[i].sizeimage; entry 887 drivers/media/platform/exynos4-is/fimc-capture.c plane_fmt[i].sizeimage = fd.entry[i].length; entry 889 drivers/media/platform/exynos4-is/fimc-capture.c if (fd.entry[0].length > FIMC_MAX_JPEG_BUF_SIZE) { entry 891 drivers/media/platform/exynos4-is/fimc-capture.c fd.entry[0].length); entry 516 drivers/media/rc/rc-main.c struct rc_map_table *entry; entry 535 drivers/media/rc/rc-main.c entry = &rc_map->scan[index]; entry 538 drivers/media/rc/rc-main.c ke->keycode = entry->keycode; entry 539 drivers/media/rc/rc-main.c ke->len = sizeof(entry->scancode); entry 540 drivers/media/rc/rc-main.c memcpy(ke->scancode, &entry->scancode, sizeof(entry->scancode)); entry 39 drivers/media/usb/siano/smsusb.c struct list_head entry; entry 1164 drivers/memstick/core/ms_block.c p->entry.disabled_block.start_addr entry 1165 drivers/memstick/core/ms_block.c = be32_to_cpu(p->entry.disabled_block.start_addr); entry 1166 drivers/memstick/core/ms_block.c p->entry.disabled_block.data_size entry 1167 drivers/memstick/core/ms_block.c = be32_to_cpu(p->entry.disabled_block.data_size); entry 1168 drivers/memstick/core/ms_block.c p->entry.cis_idi.start_addr entry 1169 drivers/memstick/core/ms_block.c = be32_to_cpu(p->entry.cis_idi.start_addr); entry 1170 drivers/memstick/core/ms_block.c p->entry.cis_idi.data_size entry 1171 drivers/memstick/core/ms_block.c = be32_to_cpu(p->entry.cis_idi.data_size); entry 1266 drivers/memstick/core/ms_block.c data_size = boot_block->entry.disabled_block.data_size; entry 1268 drivers/memstick/core/ms_block.c boot_block->entry.disabled_block.start_addr; entry 141 drivers/memstick/core/ms_block.h struct ms_system_entry entry; entry 129 drivers/mfd/dln2.c struct dln2_event_cb_entry *i, *entry; entry 133 drivers/mfd/dln2.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 134 drivers/mfd/dln2.c if (!entry) entry 137 drivers/mfd/dln2.c entry->id = id; entry 138 drivers/mfd/dln2.c entry->callback = event_cb; entry 139 drivers/mfd/dln2.c entry->pdev = pdev; entry 151 drivers/mfd/dln2.c list_add_rcu(&entry->list, &dln2->event_cb_list); entry 156 drivers/mfd/dln2.c kfree(entry); entry 155 drivers/mfd/syscon.c struct syscon *entry, *syscon = NULL; entry 159 drivers/mfd/syscon.c list_for_each_entry(entry, &syscon_list, list) entry 160 drivers/mfd/syscon.c if (entry->np == np) { entry 161 drivers/mfd/syscon.c syscon = entry; entry 703 drivers/mfd/timberdale.c msix_entries[i].entry = i; entry 37 drivers/misc/cxl/fault.c unsigned int entry; entry 47 drivers/misc/cxl/fault.c for (entry = 0, sste = primary; entry < 8; entry++, sste++) { entry 98 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry = s->private; entry 99 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = entry->dev_entry; entry 129 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry = s->private; entry 130 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = entry->dev_entry; entry 160 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry = s->private; entry 161 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = entry->dev_entry; entry 195 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry = s->private; entry 196 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = entry->dev_entry; entry 226 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry = s->private; entry 227 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = entry->dev_entry; entry 355 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry = s->private; entry 356 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = entry->dev_entry; entry 464 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry = s->private; entry 465 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = entry->dev_entry; entry 505 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry = s->private; entry 506 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *dev_entry = entry->dev_entry; entry 604 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 605 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 607 drivers/misc/habanalabs/debugfs.c u64 addr = entry->addr; entry 634 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 635 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 636 drivers/misc/habanalabs/debugfs.c u64 addr = entry->addr; entry 663 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 664 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 687 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 688 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 715 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 716 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 724 drivers/misc/habanalabs/debugfs.c rc = hl_debugfs_i2c_read(hdev, entry->i2c_bus, entry->i2c_addr, entry 725 drivers/misc/habanalabs/debugfs.c entry->i2c_reg, &val); entry 729 drivers/misc/habanalabs/debugfs.c entry->i2c_bus, entry->i2c_addr, entry->i2c_reg); entry 743 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 744 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 752 drivers/misc/habanalabs/debugfs.c rc = hl_debugfs_i2c_write(hdev, entry->i2c_bus, entry->i2c_addr, entry 753 drivers/misc/habanalabs/debugfs.c entry->i2c_reg, value); entry 757 drivers/misc/habanalabs/debugfs.c value, entry->i2c_bus, entry->i2c_addr, entry->i2c_reg); entry 767 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 768 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 786 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 787 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 805 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 806 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 832 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = file_inode(f)->i_private; entry 833 drivers/misc/habanalabs/debugfs.c struct hl_device *hdev = entry->hdev; entry 942 drivers/misc/habanalabs/debugfs.c struct hl_debugfs_entry *entry; entry 1031 drivers/misc/habanalabs/debugfs.c for (i = 0, entry = dev_entry->entry_arr ; i < count ; i++, entry++) { entry 1036 drivers/misc/habanalabs/debugfs.c entry, entry 1038 drivers/misc/habanalabs/debugfs.c entry->dent = ent; entry 1039 drivers/misc/habanalabs/debugfs.c entry->info_ent = &hl_debugfs_list[i]; entry 1040 drivers/misc/habanalabs/debugfs.c entry->dev_entry = dev_entry; entry 1046 drivers/misc/habanalabs/debugfs.c struct hl_dbg_device_entry *entry = &hdev->hl_debugfs; entry 1048 drivers/misc/habanalabs/debugfs.c debugfs_remove_recursive(entry->root); entry 1050 drivers/misc/habanalabs/debugfs.c mutex_destroy(&entry->file_mutex); entry 1051 drivers/misc/habanalabs/debugfs.c kfree(entry->entry_arr); entry 40 drivers/misc/hpilo.c static inline int get_entry_id(int entry) entry 42 drivers/misc/hpilo.c return (entry & ENTRY_MASK_DESCRIPTOR) >> ENTRY_BITPOS_DESCRIPTOR; entry 45 drivers/misc/hpilo.c static inline int get_entry_len(int entry) entry 47 drivers/misc/hpilo.c return ((entry & ENTRY_MASK_QWORDS) >> ENTRY_BITPOS_QWORDS) << 3; entry 71 drivers/misc/hpilo.c static int fifo_enqueue(struct ilo_hwinfo *hw, char *fifobar, int entry) entry 81 drivers/misc/hpilo.c (entry & ENTRY_MASK_NOSTATE) | fifo_q->merge; entry 90 drivers/misc/hpilo.c static int fifo_dequeue(struct ilo_hwinfo *hw, char *fifobar, int *entry) entry 100 drivers/misc/hpilo.c if (entry) entry 101 drivers/misc/hpilo.c *entry = c & ENTRY_MASK_NOSTATE; entry 133 drivers/misc/hpilo.c int entry; entry 140 drivers/misc/hpilo.c entry = mk_entry(id, len); entry 141 drivers/misc/hpilo.c return fifo_enqueue(hw, fifobar, entry); entry 148 drivers/misc/hpilo.c int entry = 0, pkt_id = 0; entry 159 drivers/misc/hpilo.c ret = fifo_dequeue(hw, fifobar, &entry); entry 161 drivers/misc/hpilo.c pkt_id = get_entry_id(entry); entry 165 drivers/misc/hpilo.c *len = get_entry_len(entry); entry 580 drivers/misc/ibmasm/ibmasmfs.c struct list_head *entry; entry 583 drivers/misc/ibmasm/ibmasmfs.c list_for_each(entry, &service_processors) { entry 586 drivers/misc/ibmasm/ibmasmfs.c sp = list_entry(entry, struct service_processor, node); entry 36 drivers/misc/mic/host/mic_debugfs.c " ", i, smpt_info->entry[i].dma_addr, entry 37 drivers/misc/mic/host/mic_debugfs.c smpt_info->entry[i].ref_count); entry 63 drivers/misc/mic/host/mic_debugfs.c u16 entry; entry 70 drivers/misc/mic/host/mic_debugfs.c entry = mdev->irq_info.msix_entries[i].entry; entry 73 drivers/misc/mic/host/mic_debugfs.c entry = 0; entry 77 drivers/misc/mic/host/mic_debugfs.c reg = mdev->intr_ops->read_msi_to_src_map(mdev, entry); entry 80 drivers/misc/mic/host/mic_debugfs.c "IRQ:", vector, "Entry:", entry, i, reg); entry 200 drivers/misc/mic/host/mic_intr.c mdev->irq_info.msix_entries[i].entry = i; entry 429 drivers/misc/mic/host/mic_intr.c u16 entry; entry 458 drivers/misc/mic/host/mic_intr.c entry = msix->entry; entry 459 drivers/misc/mic/host/mic_intr.c mdev->irq_info.mic_msi_map[entry] |= BIT(offset); entry 461 drivers/misc/mic/host/mic_intr.c entry, offset, true); entry 462 drivers/misc/mic/host/mic_intr.c cookie = MK_COOKIE(entry, offset); entry 475 drivers/misc/mic/host/mic_intr.c entry = 0; entry 477 drivers/misc/mic/host/mic_intr.c mdev->irq_info.mic_msi_map[entry] |= (1 << offset); entry 479 drivers/misc/mic/host/mic_intr.c entry, offset, true); entry 481 drivers/misc/mic/host/mic_intr.c cookie = MK_COOKIE(entry, intr_cb->cb_id); entry 505 drivers/misc/mic/host/mic_intr.c u32 entry; entry 510 drivers/misc/mic/host/mic_intr.c entry = GET_ENTRY((unsigned long)cookie); entry 513 drivers/misc/mic/host/mic_intr.c if (entry >= mdev->irq_info.num_vectors) { entry 516 drivers/misc/mic/host/mic_intr.c entry, mdev->irq_info.num_vectors); entry 519 drivers/misc/mic/host/mic_intr.c irq = mdev->irq_info.msix_entries[entry].vector; entry 521 drivers/misc/mic/host/mic_intr.c mdev->irq_info.mic_msi_map[entry] &= ~(BIT(offset)); entry 523 drivers/misc/mic/host/mic_intr.c entry, offset, false); entry 534 drivers/misc/mic/host/mic_intr.c mdev->irq_info.mic_msi_map[entry] &= ~(BIT(src_id)); entry 536 drivers/misc/mic/host/mic_intr.c entry, src_id, false); entry 620 drivers/misc/mic/host/mic_intr.c int entry, offset; entry 626 drivers/misc/mic/host/mic_intr.c for (entry = 0; entry < mdev->irq_info.num_vectors; entry++) { entry 628 drivers/misc/mic/host/mic_intr.c if (mdev->irq_info.mic_msi_map[entry] & BIT(offset)) entry 630 drivers/misc/mic/host/mic_intr.c entry, offset, true); entry 74 drivers/misc/mic/host/mic_smpt.c if (!smpt_info->entry[i].ref_count && entry 75 drivers/misc/mic/host/mic_smpt.c (smpt_info->entry[i].dma_addr != addr)) { entry 77 drivers/misc/mic/host/mic_smpt.c smpt_info->entry[i].dma_addr = addr; entry 79 drivers/misc/mic/host/mic_smpt.c smpt_info->entry[i].ref_count += ref[i - spt]; entry 102 drivers/misc/mic/host/mic_smpt.c if (smpt_info->entry[i].dma_addr == addr) { entry 114 drivers/misc/mic/host/mic_smpt.c ae = (smpt_info->entry[i].ref_count == 0) ? ae + 1 : 0; entry 177 drivers/misc/mic/host/mic_smpt.c dma_addr = smpt_info->entry[spt].dma_addr + entry 273 drivers/misc/mic/host/mic_smpt.c smpt_info->entry[i].ref_count -= ref[i - spt]; entry 274 drivers/misc/mic/host/mic_smpt.c if (smpt_info->entry[i].ref_count < 0) entry 354 drivers/misc/mic/host/mic_smpt.c smpt_info->entry = kmalloc_array(smpt_info->info.num_reg, entry 355 drivers/misc/mic/host/mic_smpt.c sizeof(*smpt_info->entry), GFP_KERNEL); entry 356 drivers/misc/mic/host/mic_smpt.c if (!smpt_info->entry) { entry 363 drivers/misc/mic/host/mic_smpt.c smpt_info->entry[i].dma_addr = dma_addr; entry 364 drivers/misc/mic/host/mic_smpt.c smpt_info->entry[i].ref_count = 0; entry 396 drivers/misc/mic/host/mic_smpt.c i, smpt_info->entry[i].dma_addr, entry 397 drivers/misc/mic/host/mic_smpt.c smpt_info->entry[i].ref_count); entry 398 drivers/misc/mic/host/mic_smpt.c if (smpt_info->entry[i].ref_count) entry 402 drivers/misc/mic/host/mic_smpt.c kfree(smpt_info->entry); entry 424 drivers/misc/mic/host/mic_smpt.c dma_addr = mdev->smpt->entry[i].dma_addr; entry 55 drivers/misc/mic/host/mic_smpt.h struct mic_smpt *entry; entry 64 drivers/misc/sgi-gru/grufile.c struct list_head *entry, *next; entry 73 drivers/misc/sgi-gru/grufile.c list_for_each_safe(entry, next, &vdata->vd_head) { entry 75 drivers/misc/sgi-gru/grufile.c list_entry(entry, struct gru_thread_state, ts_next); entry 1092 drivers/misc/sgi-xp/xpc_uv.c int entry; entry 1101 drivers/misc/sgi-xp/xpc_uv.c for (entry = 0; entry < nentries; entry++) { entry 1102 drivers/misc/sgi-xp/xpc_uv.c msg_slot = &ch_uv->send_msg_slots[entry]; entry 1104 drivers/misc/sgi-xp/xpc_uv.c msg_slot->msg_slot_number = entry; entry 1126 drivers/misc/sgi-xp/xpc_uv.c int entry; entry 1135 drivers/misc/sgi-xp/xpc_uv.c for (entry = 0; entry < nentries; entry++) { entry 1137 drivers/misc/sgi-xp/xpc_uv.c entry * ch->entry_size; entry 1139 drivers/misc/sgi-xp/xpc_uv.c msg_slot->hdr.msg_slot_number = entry; entry 1337 drivers/misc/sgi-xp/xpc_uv.c struct xpc_fifo_entry_uv *entry; entry 1340 drivers/misc/sgi-xp/xpc_uv.c entry = xpc_get_fifo_entry_uv(&ch->sn.uv.msg_slot_free_list); entry 1341 drivers/misc/sgi-xp/xpc_uv.c if (entry != NULL) entry 1352 drivers/misc/sgi-xp/xpc_uv.c msg_slot = container_of(entry, struct xpc_send_msg_slot_uv, next); entry 1396 drivers/misc/sgi-xp/xpc_uv.c int entry = msg->hdr.msg_slot_number % ch->local_nentries; entry 1398 drivers/misc/sgi-xp/xpc_uv.c msg_slot = &ch->sn.uv.send_msg_slots[entry]; entry 1628 drivers/misc/sgi-xp/xpc_uv.c int entry; entry 1632 drivers/misc/sgi-xp/xpc_uv.c for (entry = 0; entry < ch->local_nentries; entry++) { entry 1637 drivers/misc/sgi-xp/xpc_uv.c msg_slot = &ch->sn.uv.send_msg_slots[entry]; entry 1649 drivers/misc/sgi-xp/xpc_uv.c struct xpc_fifo_entry_uv *entry; entry 1654 drivers/misc/sgi-xp/xpc_uv.c entry = xpc_get_fifo_entry_uv(&ch->sn.uv.recv_msg_list); entry 1655 drivers/misc/sgi-xp/xpc_uv.c if (entry != NULL) { entry 1656 drivers/misc/sgi-xp/xpc_uv.c msg = container_of(entry, struct xpc_notify_mq_msg_uv, entry 705 drivers/misc/vmw_vmci/vmci_context.c struct vmci_handle_list *entry; entry 724 drivers/misc/vmw_vmci/vmci_context.c list_for_each_entry(entry, &context->notifier_list, node) entry 725 drivers/misc/vmw_vmci/vmci_context.c notifiers[i++] = entry->handle.context; entry 36 drivers/misc/vmw_vmci/vmci_datagram.c struct datagram_entry *entry; entry 59 drivers/misc/vmw_vmci/vmci_datagram.c struct datagram_entry *entry; entry 74 drivers/misc/vmw_vmci/vmci_datagram.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 75 drivers/misc/vmw_vmci/vmci_datagram.c if (!entry) { entry 80 drivers/misc/vmw_vmci/vmci_datagram.c entry->run_delayed = (flags & VMCI_FLAG_DG_DELAYED_CB) ? true : false; entry 81 drivers/misc/vmw_vmci/vmci_datagram.c entry->flags = flags; entry 82 drivers/misc/vmw_vmci/vmci_datagram.c entry->recv_cb = recv_cb; entry 83 drivers/misc/vmw_vmci/vmci_datagram.c entry->client_data = client_data; entry 84 drivers/misc/vmw_vmci/vmci_datagram.c entry->priv_flags = priv_flags; entry 87 drivers/misc/vmw_vmci/vmci_datagram.c result = vmci_resource_add(&entry->resource, entry 93 drivers/misc/vmw_vmci/vmci_datagram.c kfree(entry); entry 97 drivers/misc/vmw_vmci/vmci_datagram.c *out_handle = vmci_resource_handle(&entry->resource); entry 141 drivers/misc/vmw_vmci/vmci_datagram.c dg_info->entry->recv_cb(dg_info->entry->client_data, &dg_info->msg); entry 143 drivers/misc/vmw_vmci/vmci_datagram.c vmci_resource_put(&dg_info->entry->resource); entry 236 drivers/misc/vmw_vmci/vmci_datagram.c dg_info->entry = dst_entry; entry 379 drivers/misc/vmw_vmci/vmci_datagram.c dg_info->entry = dst_entry; entry 461 drivers/misc/vmw_vmci/vmci_datagram.c struct datagram_entry *entry; entry 471 drivers/misc/vmw_vmci/vmci_datagram.c entry = container_of(resource, struct datagram_entry, resource); entry 473 drivers/misc/vmw_vmci/vmci_datagram.c vmci_resource_put(&entry->resource); entry 474 drivers/misc/vmw_vmci/vmci_datagram.c vmci_resource_remove(&entry->resource); entry 475 drivers/misc/vmw_vmci/vmci_datagram.c kfree(entry); entry 91 drivers/misc/vmw_vmci/vmci_doorbell.c struct dbell_entry *entry; entry 99 drivers/misc/vmw_vmci/vmci_doorbell.c entry = container_of(resource, struct dbell_entry, resource); entry 100 drivers/misc/vmw_vmci/vmci_doorbell.c *priv_flags = entry->priv_flags; entry 137 drivers/misc/vmw_vmci/vmci_doorbell.c static void dbell_index_table_add(struct dbell_entry *entry) entry 142 drivers/misc/vmw_vmci/vmci_doorbell.c vmci_resource_get(&entry->resource); entry 186 drivers/misc/vmw_vmci/vmci_doorbell.c entry->idx = new_notify_idx; entry 187 drivers/misc/vmw_vmci/vmci_doorbell.c bucket = VMCI_DOORBELL_HASH(entry->idx); entry 188 drivers/misc/vmw_vmci/vmci_doorbell.c hlist_add_head(&entry->node, &vmci_doorbell_it.entries[bucket]); entry 197 drivers/misc/vmw_vmci/vmci_doorbell.c static void dbell_index_table_remove(struct dbell_entry *entry) entry 201 drivers/misc/vmw_vmci/vmci_doorbell.c hlist_del_init(&entry->node); entry 204 drivers/misc/vmw_vmci/vmci_doorbell.c if (entry->idx == max_notify_idx - 1) { entry 217 drivers/misc/vmw_vmci/vmci_doorbell.c last_notify_idx_released = entry->idx; entry 221 drivers/misc/vmw_vmci/vmci_doorbell.c vmci_resource_put(&entry->resource); entry 282 drivers/misc/vmw_vmci/vmci_doorbell.c struct dbell_entry *entry = container_of(work, entry 285 drivers/misc/vmw_vmci/vmci_doorbell.c entry->notify_cb(entry->client_data); entry 286 drivers/misc/vmw_vmci/vmci_doorbell.c vmci_resource_put(&entry->resource); entry 294 drivers/misc/vmw_vmci/vmci_doorbell.c struct dbell_entry *entry; entry 311 drivers/misc/vmw_vmci/vmci_doorbell.c entry = container_of(resource, struct dbell_entry, resource); entry 312 drivers/misc/vmw_vmci/vmci_doorbell.c if (entry->run_delayed) { entry 313 drivers/misc/vmw_vmci/vmci_doorbell.c if (!schedule_work(&entry->work)) entry 316 drivers/misc/vmw_vmci/vmci_doorbell.c entry->notify_cb(entry->client_data); entry 414 drivers/misc/vmw_vmci/vmci_doorbell.c struct dbell_entry *entry; entry 422 drivers/misc/vmw_vmci/vmci_doorbell.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 423 drivers/misc/vmw_vmci/vmci_doorbell.c if (entry == NULL) { entry 465 drivers/misc/vmw_vmci/vmci_doorbell.c entry->idx = 0; entry 466 drivers/misc/vmw_vmci/vmci_doorbell.c INIT_HLIST_NODE(&entry->node); entry 467 drivers/misc/vmw_vmci/vmci_doorbell.c entry->priv_flags = priv_flags; entry 468 drivers/misc/vmw_vmci/vmci_doorbell.c INIT_WORK(&entry->work, dbell_delayed_dispatch); entry 469 drivers/misc/vmw_vmci/vmci_doorbell.c entry->run_delayed = flags & VMCI_FLAG_DELAYED_CB; entry 470 drivers/misc/vmw_vmci/vmci_doorbell.c entry->notify_cb = notify_cb; entry 471 drivers/misc/vmw_vmci/vmci_doorbell.c entry->client_data = client_data; entry 472 drivers/misc/vmw_vmci/vmci_doorbell.c atomic_set(&entry->active, 0); entry 474 drivers/misc/vmw_vmci/vmci_doorbell.c result = vmci_resource_add(&entry->resource, entry 483 drivers/misc/vmw_vmci/vmci_doorbell.c new_handle = vmci_resource_handle(&entry->resource); entry 485 drivers/misc/vmw_vmci/vmci_doorbell.c dbell_index_table_add(entry); entry 486 drivers/misc/vmw_vmci/vmci_doorbell.c result = dbell_link(new_handle, entry->idx); entry 490 drivers/misc/vmw_vmci/vmci_doorbell.c atomic_set(&entry->active, 1); entry 498 drivers/misc/vmw_vmci/vmci_doorbell.c dbell_index_table_remove(entry); entry 499 drivers/misc/vmw_vmci/vmci_doorbell.c vmci_resource_remove(&entry->resource); entry 501 drivers/misc/vmw_vmci/vmci_doorbell.c kfree(entry); entry 515 drivers/misc/vmw_vmci/vmci_doorbell.c struct dbell_entry *entry; entry 529 drivers/misc/vmw_vmci/vmci_doorbell.c entry = container_of(resource, struct dbell_entry, resource); entry 531 drivers/misc/vmw_vmci/vmci_doorbell.c if (!hlist_unhashed(&entry->node)) { entry 534 drivers/misc/vmw_vmci/vmci_doorbell.c dbell_index_table_remove(entry); entry 561 drivers/misc/vmw_vmci/vmci_doorbell.c vmci_resource_put(&entry->resource); entry 562 drivers/misc/vmw_vmci/vmci_doorbell.c vmci_resource_remove(&entry->resource); entry 564 drivers/misc/vmw_vmci/vmci_doorbell.c kfree(entry); entry 800 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_entry *entry; entry 805 drivers/misc/vmw_vmci/vmci_queue_pair.c list_for_each_entry(entry, &qp_list->head, list_item) { entry 806 drivers/misc/vmw_vmci/vmci_queue_pair.c if (vmci_handle_is_equal(entry->handle, handle)) entry 807 drivers/misc/vmw_vmci/vmci_queue_pair.c return entry; entry 819 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_guest_endpoint *entry; entry 822 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = qp ? container_of( entry 824 drivers/misc/vmw_vmci/vmci_queue_pair.c return entry; entry 833 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_broker_entry *entry; entry 836 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = qp ? container_of( entry 838 drivers/misc/vmw_vmci/vmci_queue_pair.c return entry; entry 879 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_guest_endpoint *entry; entry 890 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 891 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry) { entry 892 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.peer = peer; entry 893 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.flags = flags; entry 894 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.produce_size = produce_size; entry 895 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.consume_size = consume_size; entry 896 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.ref_count = 0; entry 897 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->num_ppns = num_ppns; entry 898 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q = produce_q; entry 899 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q = consume_q; entry 900 drivers/misc/vmw_vmci/vmci_queue_pair.c INIT_LIST_HEAD(&entry->qp.list_item); entry 903 drivers/misc/vmw_vmci/vmci_queue_pair.c result = vmci_resource_add(&entry->resource, entry 906 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.handle = vmci_resource_handle(&entry->resource); entry 908 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_list_find(&qp_guest_endpoints, entry->qp.handle)) { entry 911 drivers/misc/vmw_vmci/vmci_queue_pair.c kfree(entry); entry 912 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = NULL; entry 915 drivers/misc/vmw_vmci/vmci_queue_pair.c return entry; entry 921 drivers/misc/vmw_vmci/vmci_queue_pair.c static void qp_guest_endpoint_destroy(struct qp_guest_endpoint *entry) entry 923 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_free_ppn_set(&entry->ppn_set); entry 924 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_cleanup_queue_mutex(entry->produce_q, entry->consume_q); entry 925 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_free_queue(entry->produce_q, entry->qp.produce_size); entry 926 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_free_queue(entry->consume_q, entry->qp.consume_size); entry 928 drivers/misc/vmw_vmci/vmci_queue_pair.c vmci_resource_remove(&entry->resource); entry 930 drivers/misc/vmw_vmci/vmci_queue_pair.c kfree(entry); entry 937 drivers/misc/vmw_vmci/vmci_queue_pair.c static int qp_alloc_hypercall(const struct qp_guest_endpoint *entry) entry 944 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!entry || entry->num_ppns <= 2) entry 949 drivers/misc/vmw_vmci/vmci_queue_pair.c (size_t) entry->num_ppns * ppn_size; entry 958 drivers/misc/vmw_vmci/vmci_queue_pair.c alloc_msg->handle = entry->qp.handle; entry 959 drivers/misc/vmw_vmci/vmci_queue_pair.c alloc_msg->peer = entry->qp.peer; entry 960 drivers/misc/vmw_vmci/vmci_queue_pair.c alloc_msg->flags = entry->qp.flags; entry 961 drivers/misc/vmw_vmci/vmci_queue_pair.c alloc_msg->produce_size = entry->qp.produce_size; entry 962 drivers/misc/vmw_vmci/vmci_queue_pair.c alloc_msg->consume_size = entry->qp.consume_size; entry 963 drivers/misc/vmw_vmci/vmci_queue_pair.c alloc_msg->num_ppns = entry->num_ppns; entry 966 drivers/misc/vmw_vmci/vmci_queue_pair.c &entry->ppn_set); entry 995 drivers/misc/vmw_vmci/vmci_queue_pair.c static void qp_list_add_entry(struct qp_list *qp_list, struct qp_entry *entry) entry 997 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry) entry 998 drivers/misc/vmw_vmci/vmci_queue_pair.c list_add(&entry->list_item, &qp_list->head); entry 1005 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_entry *entry) entry 1007 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry) entry 1008 drivers/misc/vmw_vmci/vmci_queue_pair.c list_del(&entry->list_item); entry 1018 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_guest_endpoint *entry; entry 1023 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = qp_guest_handle_to_entry(handle); entry 1024 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!entry) { entry 1029 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->qp.flags & VMCI_QPFLAG_LOCAL) { entry 1032 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->qp.ref_count > 1) { entry 1064 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.ref_count--; entry 1065 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->qp.ref_count == 0) entry 1066 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_list_remove_entry(&qp_guest_endpoints, &entry->qp); entry 1069 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry) entry 1070 drivers/misc/vmw_vmci/vmci_queue_pair.c ref_count = entry->qp.ref_count; entry 1075 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_guest_endpoint_destroy(entry); entry 1285 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_broker_entry *entry = NULL; entry 1313 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 1314 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!entry) entry 1333 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.handle = handle; entry 1334 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.peer = peer; entry 1335 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.flags = flags; entry 1336 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.produce_size = guest_produce_size; entry 1337 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.consume_size = guest_consume_size; entry 1338 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.ref_count = 1; entry 1339 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->create_id = context_id; entry 1340 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->attach_id = VMCI_INVALID_ID; entry 1341 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_NEW; entry 1342 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->require_trusted_attach = entry 1344 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->created_by_trusted = entry 1346 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->vmci_page_files = false; entry 1347 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->wakeup_cb = wakeup_cb; entry 1348 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->client_data = client_data; entry 1349 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q = qp_host_alloc_queue(guest_produce_size); entry 1350 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->produce_q == NULL) { entry 1354 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q = qp_host_alloc_queue(guest_consume_size); entry 1355 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->consume_q == NULL) { entry 1360 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_init_queue_mutex(entry->produce_q, entry->consume_q); entry 1362 drivers/misc/vmw_vmci/vmci_queue_pair.c INIT_LIST_HEAD(&entry->qp.list_item); entry 1367 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->local_mem = kcalloc(QPE_NUM_PAGES(entry->qp), entry 1369 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->local_mem == NULL) { entry 1373 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_CREATED_MEM; entry 1374 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q->q_header = entry->local_mem; entry 1375 drivers/misc/vmw_vmci/vmci_queue_pair.c tmp = (u8 *)entry->local_mem + PAGE_SIZE * entry 1376 drivers/misc/vmw_vmci/vmci_queue_pair.c (DIV_ROUND_UP(entry->qp.produce_size, PAGE_SIZE) + 1); entry 1377 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q->q_header = (struct vmci_queue_header *)tmp; entry 1384 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q, entry 1385 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q); entry 1389 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_CREATED_MEM; entry 1398 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_CREATED_NO_MEM; entry 1401 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_list_add_entry(&qp_broker_list, &entry->qp); entry 1403 drivers/misc/vmw_vmci/vmci_queue_pair.c *ent = entry; entry 1406 drivers/misc/vmw_vmci/vmci_queue_pair.c result = vmci_resource_add(&entry->resource, entry 1415 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.handle = vmci_resource_handle(&entry->resource); entry 1417 drivers/misc/vmw_vmci/vmci_queue_pair.c vmci_q_header_init(entry->produce_q->q_header, entry 1418 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.handle); entry 1419 drivers/misc/vmw_vmci/vmci_queue_pair.c vmci_q_header_init(entry->consume_q->q_header, entry 1420 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.handle); entry 1423 drivers/misc/vmw_vmci/vmci_queue_pair.c vmci_ctx_qp_create(context, entry->qp.handle); entry 1428 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry != NULL) { entry 1429 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_host_free_queue(entry->produce_q, guest_produce_size); entry 1430 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_host_free_queue(entry->consume_q, guest_consume_size); entry 1431 drivers/misc/vmw_vmci/vmci_queue_pair.c kfree(entry); entry 1502 drivers/misc/vmw_vmci/vmci_queue_pair.c static int qp_broker_attach(struct qp_broker_entry *entry, entry 1518 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->state != VMCIQPB_CREATED_NO_MEM && entry 1519 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state != VMCIQPB_CREATED_MEM) entry 1523 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!(entry->qp.flags & VMCI_QPFLAG_LOCAL) || entry 1524 drivers/misc/vmw_vmci/vmci_queue_pair.c context_id != entry->create_id) { entry 1527 drivers/misc/vmw_vmci/vmci_queue_pair.c } else if (context_id == entry->create_id || entry 1528 drivers/misc/vmw_vmci/vmci_queue_pair.c context_id == entry->attach_id) { entry 1533 drivers/misc/vmw_vmci/vmci_queue_pair.c VMCI_CONTEXT_IS_VM(entry->create_id)) entry 1541 drivers/misc/vmw_vmci/vmci_queue_pair.c !entry->created_by_trusted) entry 1548 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->require_trusted_attach && entry 1556 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->qp.peer != VMCI_INVALID_ID && entry->qp.peer != context_id) entry 1559 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->create_id == VMCI_HOST_CONTEXT_ID) { entry 1577 drivers/misc/vmw_vmci/vmci_queue_pair.c create_context = vmci_ctx_get(entry->create_id); entry 1585 drivers/misc/vmw_vmci/vmci_queue_pair.c if ((entry->qp.flags & ~VMCI_QP_ASYMM) != (flags & ~VMCI_QP_ASYMM_PEER)) entry 1595 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->qp.produce_size != produce_size || entry 1596 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.consume_size != consume_size) { entry 1599 drivers/misc/vmw_vmci/vmci_queue_pair.c } else if (entry->qp.produce_size != consume_size || entry 1600 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.consume_size != produce_size) { entry 1618 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->state != VMCIQPB_CREATED_NO_MEM) entry 1630 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q, entry 1631 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q); entry 1635 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_ATTACHED_MEM; entry 1637 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_ATTACHED_NO_MEM; entry 1639 drivers/misc/vmw_vmci/vmci_queue_pair.c } else if (entry->state == VMCIQPB_CREATED_NO_MEM) { entry 1650 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_ATTACHED_MEM; entry 1653 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->state == VMCIQPB_ATTACHED_MEM) { entry 1655 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_notify_peer(true, entry->qp.handle, context_id, entry 1656 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->create_id); entry 1659 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->create_id, entry->qp.handle.context, entry 1660 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.handle.resource); entry 1663 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->attach_id = context_id; entry 1664 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.ref_count++; entry 1666 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->wakeup_cb = wakeup_cb; entry 1667 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->client_data = client_data; entry 1675 drivers/misc/vmw_vmci/vmci_queue_pair.c vmci_ctx_qp_create(context, entry->qp.handle); entry 1678 drivers/misc/vmw_vmci/vmci_queue_pair.c *ent = entry; entry 1702 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_broker_entry *entry = NULL; entry 1732 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = qp_broker_handle_to_entry(handle); entry 1734 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!entry) { entry 1743 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_broker_attach(entry, peer, flags, priv_flags, entry 1774 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_broker_entry *entry; entry 1785 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = NULL; entry 1789 drivers/misc/vmw_vmci/vmci_queue_pair.c wakeup_cb, client_data, &entry, &swap); entry 1798 drivers/misc/vmw_vmci/vmci_queue_pair.c *produce_q = entry->consume_q; entry 1799 drivers/misc/vmw_vmci/vmci_queue_pair.c *consume_q = entry->produce_q; entry 1801 drivers/misc/vmw_vmci/vmci_queue_pair.c *produce_q = entry->produce_q; entry 1802 drivers/misc/vmw_vmci/vmci_queue_pair.c *consume_q = entry->consume_q; entry 1805 drivers/misc/vmw_vmci/vmci_queue_pair.c *handle = vmci_resource_handle(&entry->resource); entry 1888 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_entry *entry = entry 1891 drivers/misc/vmw_vmci/vmci_queue_pair.c return entry; entry 1899 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_entry *entry; entry 1904 drivers/misc/vmw_vmci/vmci_queue_pair.c while ((entry = qp_list_get_head(&qp_broker_list))) { entry 1905 drivers/misc/vmw_vmci/vmci_queue_pair.c be = (struct qp_broker_entry *)entry; entry 1907 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_list_remove_entry(&qp_broker_list, entry); entry 1956 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_broker_entry *entry; entry 1981 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = qp_broker_handle_to_entry(handle); entry 1982 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!entry) { entry 1993 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->create_id != context_id && entry 1994 drivers/misc/vmw_vmci/vmci_queue_pair.c (entry->create_id != VMCI_HOST_CONTEXT_ID || entry 1995 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->attach_id != context_id)) { entry 2000 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->state != VMCIQPB_CREATED_NO_MEM && entry 2001 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state != VMCIQPB_ATTACHED_NO_MEM) { entry 2007 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q, entry->consume_q); entry 2011 drivers/misc/vmw_vmci/vmci_queue_pair.c result = qp_host_map_queues(entry->produce_q, entry->consume_q); entry 2013 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_host_unregister_user_memory(entry->produce_q, entry 2014 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q); entry 2018 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->state == VMCIQPB_CREATED_NO_MEM) entry 2019 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_CREATED_MEM; entry 2021 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_ATTACHED_MEM; entry 2023 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->vmci_page_files = true; entry 2025 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->state == VMCIQPB_ATTACHED_MEM) { entry 2027 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_notify_peer(true, handle, context_id, entry->create_id); entry 2030 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->create_id, entry->qp.handle.context, entry 2031 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.handle.resource); entry 2046 drivers/misc/vmw_vmci/vmci_queue_pair.c static void qp_reset_saved_headers(struct qp_broker_entry *entry) entry 2048 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q->saved_header = NULL; entry 2049 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q->saved_header = NULL; entry 2072 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_broker_entry *entry; entry 2092 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = qp_broker_handle_to_entry(handle); entry 2093 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!entry) { entry 2100 drivers/misc/vmw_vmci/vmci_queue_pair.c if (context_id != entry->create_id && context_id != entry->attach_id) { entry 2105 drivers/misc/vmw_vmci/vmci_queue_pair.c if (context_id == entry->create_id) { entry 2106 drivers/misc/vmw_vmci/vmci_queue_pair.c peer_id = entry->attach_id; entry 2107 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->create_id = VMCI_INVALID_ID; entry 2109 drivers/misc/vmw_vmci/vmci_queue_pair.c peer_id = entry->create_id; entry 2110 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->attach_id = VMCI_INVALID_ID; entry 2112 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->qp.ref_count--; entry 2114 drivers/misc/vmw_vmci/vmci_queue_pair.c is_local = entry->qp.flags & VMCI_QPFLAG_LOCAL; entry 2127 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_acquire_queue_mutex(entry->produce_q); entry 2128 drivers/misc/vmw_vmci/vmci_queue_pair.c headers_mapped = entry->produce_q->q_header || entry 2129 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q->q_header; entry 2130 drivers/misc/vmw_vmci/vmci_queue_pair.c if (QPBROKERSTATE_HAS_MEM(entry)) { entry 2133 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q, entry 2134 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q); entry 2140 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_host_unregister_user_memory(entry->produce_q, entry 2141 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q); entry 2146 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_reset_saved_headers(entry); entry 2148 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_release_queue_mutex(entry->produce_q); entry 2150 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!headers_mapped && entry->wakeup_cb) entry 2151 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->wakeup_cb(entry->client_data); entry 2154 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->wakeup_cb) { entry 2155 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->wakeup_cb = NULL; entry 2156 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->client_data = NULL; entry 2160 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->qp.ref_count == 0) { entry 2161 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_list_remove_entry(&qp_broker_list, &entry->qp); entry 2164 drivers/misc/vmw_vmci/vmci_queue_pair.c kfree(entry->local_mem); entry 2166 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_cleanup_queue_mutex(entry->produce_q, entry->consume_q); entry 2167 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_host_free_queue(entry->produce_q, entry->qp.produce_size); entry 2168 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_host_free_queue(entry->consume_q, entry->qp.consume_size); entry 2170 drivers/misc/vmw_vmci/vmci_queue_pair.c vmci_resource_remove(&entry->resource); entry 2172 drivers/misc/vmw_vmci/vmci_queue_pair.c kfree(entry); entry 2178 drivers/misc/vmw_vmci/vmci_queue_pair.c QPBROKERSTATE_HAS_MEM(entry)) { entry 2179 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_SHUTDOWN_MEM; entry 2181 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state = VMCIQPB_SHUTDOWN_NO_MEM; entry 2204 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_broker_entry *entry; entry 2221 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = qp_broker_handle_to_entry(handle); entry 2222 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!entry) { entry 2229 drivers/misc/vmw_vmci/vmci_queue_pair.c if (context_id != entry->create_id && context_id != entry->attach_id) { entry 2240 drivers/misc/vmw_vmci/vmci_queue_pair.c page_store.len = QPE_NUM_PAGES(entry->qp); entry 2242 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_acquire_queue_mutex(entry->produce_q); entry 2243 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_reset_saved_headers(entry); entry 2246 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q, entry 2247 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q); entry 2248 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_release_queue_mutex(entry->produce_q); entry 2252 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state++; entry 2254 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->wakeup_cb) entry 2255 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->wakeup_cb(entry->client_data); entry 2271 drivers/misc/vmw_vmci/vmci_queue_pair.c static int qp_save_headers(struct qp_broker_entry *entry) entry 2275 drivers/misc/vmw_vmci/vmci_queue_pair.c if (entry->produce_q->saved_header != NULL && entry 2276 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q->saved_header != NULL) { entry 2286 drivers/misc/vmw_vmci/vmci_queue_pair.c if (NULL == entry->produce_q->q_header || entry 2287 drivers/misc/vmw_vmci/vmci_queue_pair.c NULL == entry->consume_q->q_header) { entry 2288 drivers/misc/vmw_vmci/vmci_queue_pair.c result = qp_host_map_queues(entry->produce_q, entry->consume_q); entry 2293 drivers/misc/vmw_vmci/vmci_queue_pair.c memcpy(&entry->saved_produce_q, entry->produce_q->q_header, entry 2294 drivers/misc/vmw_vmci/vmci_queue_pair.c sizeof(entry->saved_produce_q)); entry 2295 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->produce_q->saved_header = &entry->saved_produce_q; entry 2296 drivers/misc/vmw_vmci/vmci_queue_pair.c memcpy(&entry->saved_consume_q, entry->consume_q->q_header, entry 2297 drivers/misc/vmw_vmci/vmci_queue_pair.c sizeof(entry->saved_consume_q)); entry 2298 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q->saved_header = &entry->saved_consume_q; entry 2313 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_broker_entry *entry; entry 2330 drivers/misc/vmw_vmci/vmci_queue_pair.c entry = qp_broker_handle_to_entry(handle); entry 2331 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!entry) { entry 2338 drivers/misc/vmw_vmci/vmci_queue_pair.c if (context_id != entry->create_id && context_id != entry->attach_id) { entry 2344 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_acquire_queue_mutex(entry->produce_q); entry 2345 drivers/misc/vmw_vmci/vmci_queue_pair.c result = qp_save_headers(entry); entry 2350 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_host_unmap_queues(gid, entry->produce_q, entry->consume_q); entry 2359 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_host_unregister_user_memory(entry->produce_q, entry 2360 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->consume_q); entry 2365 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->state--; entry 2367 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_release_queue_mutex(entry->produce_q); entry 2385 drivers/misc/vmw_vmci/vmci_queue_pair.c struct qp_entry *entry; entry 2390 drivers/misc/vmw_vmci/vmci_queue_pair.c while ((entry = qp_list_get_head(&qp_guest_endpoints))) { entry 2391 drivers/misc/vmw_vmci/vmci_queue_pair.c ep = (struct qp_guest_endpoint *)entry; entry 2394 drivers/misc/vmw_vmci/vmci_queue_pair.c if (!(entry->flags & VMCI_QPFLAG_LOCAL)) entry 2395 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_detatch_hypercall(entry->handle); entry 2398 drivers/misc/vmw_vmci/vmci_queue_pair.c entry->ref_count = 0; entry 2399 drivers/misc/vmw_vmci/vmci_queue_pair.c qp_list_remove_entry(&qp_guest_endpoints, entry); entry 75 drivers/mtd/nand/bbt.c unsigned int entry) entry 79 drivers/mtd/nand/bbt.c ((entry * bits_per_block) / BITS_PER_LONG); entry 80 drivers/mtd/nand/bbt.c unsigned int offs = (entry * bits_per_block) % BITS_PER_LONG; entry 83 drivers/mtd/nand/bbt.c if (entry >= nanddev_neraseblocks(nand)) entry 107 drivers/mtd/nand/bbt.c int nanddev_bbt_set_block_status(struct nand_device *nand, unsigned int entry, entry 112 drivers/mtd/nand/bbt.c ((entry * bits_per_block) / BITS_PER_LONG); entry 113 drivers/mtd/nand/bbt.c unsigned int offs = (entry * bits_per_block) % BITS_PER_LONG; entry 116 drivers/mtd/nand/bbt.c if (entry >= nanddev_neraseblocks(nand)) entry 25 drivers/mtd/nand/core.c unsigned int entry; entry 28 drivers/mtd/nand/core.c entry = nanddev_bbt_pos_to_entry(nand, pos); entry 29 drivers/mtd/nand/core.c status = nanddev_bbt_get_block_status(nand, entry); entry 37 drivers/mtd/nand/core.c nanddev_bbt_set_block_status(nand, entry, status); entry 64 drivers/mtd/nand/core.c unsigned int entry; entry 78 drivers/mtd/nand/core.c entry = nanddev_bbt_pos_to_entry(nand, pos); entry 79 drivers/mtd/nand/core.c ret = nanddev_bbt_set_block_status(nand, entry, NAND_BBT_BLOCK_WORN); entry 104 drivers/mtd/nand/core.c unsigned int entry; entry 111 drivers/mtd/nand/core.c entry = nanddev_bbt_pos_to_entry(nand, pos); entry 112 drivers/mtd/nand/core.c status = nanddev_bbt_get_block_status(nand, entry); entry 2719 drivers/mtd/nand/raw/marvell_nand.c struct marvell_nand_chip *entry, *temp; entry 2721 drivers/mtd/nand/raw/marvell_nand.c list_for_each_entry_safe(entry, temp, &nfc->chips, node) { entry 2722 drivers/mtd/nand/raw/marvell_nand.c nand_release(&entry->chip); entry 2723 drivers/mtd/nand/raw/marvell_nand.c list_del(&entry->node); entry 78 drivers/mtd/nand/raw/nand_bbt.c uint8_t entry = chip->bbt[block >> BBT_ENTRY_SHIFT]; entry 79 drivers/mtd/nand/raw/nand_bbt.c entry >>= (block & BBT_ENTRY_MASK) * 2; entry 80 drivers/mtd/nand/raw/nand_bbt.c return entry & BBT_ENTRY_MASK; entry 108 drivers/mtd/rfd_ftl.c u16 entry; entry 110 drivers/mtd/rfd_ftl.c entry = le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]); entry 112 drivers/mtd/rfd_ftl.c if (entry == SECTOR_DELETED) entry 115 drivers/mtd/rfd_ftl.c if (entry == SECTOR_FREE) { entry 120 drivers/mtd/rfd_ftl.c if (entry == SECTOR_ZERO) entry 121 drivers/mtd/rfd_ftl.c entry = 0; entry 123 drivers/mtd/rfd_ftl.c if (entry >= part->sector_count) { entry 127 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, block_no, i, entry); entry 131 drivers/mtd/rfd_ftl.c if (part->sector_map[entry] != -1) { entry 134 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, entry); entry 139 drivers/mtd/rfd_ftl.c part->sector_map[entry] = block->offset + entry 354 drivers/mtd/rfd_ftl.c u16 entry = le16_to_cpu(map[HEADER_MAP_OFFSET + i]); entry 358 drivers/mtd/rfd_ftl.c if (entry == SECTOR_FREE || entry == SECTOR_DELETED) entry 361 drivers/mtd/rfd_ftl.c if (entry == SECTOR_ZERO) entry 362 drivers/mtd/rfd_ftl.c entry = 0; entry 365 drivers/mtd/rfd_ftl.c if (entry >= part->sector_count) entry 394 drivers/mtd/rfd_ftl.c entry, sector_data); entry 612 drivers/mtd/rfd_ftl.c u16 entry; entry 647 drivers/mtd/rfd_ftl.c entry = cpu_to_le16(sector == 0 ? SECTOR_ZERO : sector); entry 649 drivers/mtd/rfd_ftl.c part->header_cache[i + HEADER_MAP_OFFSET] = entry; entry 652 drivers/mtd/rfd_ftl.c rc = mtd_write(part->mbd.mtd, addr, sizeof(entry), &retlen, entry 653 drivers/mtd/rfd_ftl.c (u_char *)&entry); entry 655 drivers/mtd/rfd_ftl.c if (!rc && retlen != sizeof(entry)) entry 1656 drivers/mtd/ubi/eba.c struct ubi_eba_entry *entry; entry 1658 drivers/mtd/ubi/eba.c entry = &vol->eba_tbl->entries[aeb->lnum]; entry 1659 drivers/mtd/ubi/eba.c entry->pnum = aeb->pnum; entry 74 drivers/net/bonding/bond_alb.c static inline void tlb_init_table_entry(struct tlb_client_info *entry, int save_load) entry 77 drivers/net/bonding/bond_alb.c entry->load_history = 1 + entry->tx_bytes / entry 79 drivers/net/bonding/bond_alb.c entry->tx_bytes = 0; entry 82 drivers/net/bonding/bond_alb.c entry->tx_slave = NULL; entry 83 drivers/net/bonding/bond_alb.c entry->next = TLB_NULL_INDEX; entry 84 drivers/net/bonding/bond_alb.c entry->prev = TLB_NULL_INDEX; entry 731 drivers/net/bonding/bond_alb.c static void rlb_init_table_entry_dst(struct rlb_client_info *entry) entry 733 drivers/net/bonding/bond_alb.c entry->used_next = RLB_NULL_INDEX; entry 734 drivers/net/bonding/bond_alb.c entry->used_prev = RLB_NULL_INDEX; entry 735 drivers/net/bonding/bond_alb.c entry->assigned = 0; entry 736 drivers/net/bonding/bond_alb.c entry->slave = NULL; entry 737 drivers/net/bonding/bond_alb.c entry->vlan_id = 0; entry 739 drivers/net/bonding/bond_alb.c static void rlb_init_table_entry_src(struct rlb_client_info *entry) entry 741 drivers/net/bonding/bond_alb.c entry->src_first = RLB_NULL_INDEX; entry 742 drivers/net/bonding/bond_alb.c entry->src_prev = RLB_NULL_INDEX; entry 743 drivers/net/bonding/bond_alb.c entry->src_next = RLB_NULL_INDEX; entry 746 drivers/net/bonding/bond_alb.c static void rlb_init_table_entry(struct rlb_client_info *entry) entry 748 drivers/net/bonding/bond_alb.c memset(entry, 0, sizeof(struct rlb_client_info)); entry 749 drivers/net/bonding/bond_alb.c rlb_init_table_entry_dst(entry); entry 750 drivers/net/bonding/bond_alb.c rlb_init_table_entry_src(entry); entry 794 drivers/net/bonding/bond_alb.c struct rlb_client_info *entry = &(bond_info->rx_hashtbl[index]); entry 797 drivers/net/bonding/bond_alb.c rlb_init_table_entry_dst(entry); entry 831 drivers/net/bonding/bond_alb.c struct rlb_client_info *entry = &(bond_info->rx_hashtbl[index]); entry 832 drivers/net/bonding/bond_alb.c u32 next_index = entry->src_next; entry 833 drivers/net/bonding/bond_alb.c if (entry->ip_src == arp->ip_src && entry 834 drivers/net/bonding/bond_alb.c !ether_addr_equal_64bits(arp->mac_src, entry->mac_src)) entry 440 drivers/net/can/mscan/mscan.c struct tx_queue_entry *entry = entry 442 drivers/net/can/mscan/mscan.c u8 mask = entry->mask; entry 450 drivers/net/can/mscan/mscan.c can_get_echo_skb(dev, entry->id); entry 470 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_GAFL_PAGENUM(entry) ((entry) / 16) entry 252 drivers/net/dsa/b53/b53_common.c u32 entry = 0; entry 255 drivers/net/dsa/b53/b53_common.c entry = ((vlan->untag & VA_UNTAG_MASK_25) << entry 258 drivers/net/dsa/b53/b53_common.c entry |= VA_VALID_25_R4 | vid << VA_VID_HIGH_S; entry 260 drivers/net/dsa/b53/b53_common.c entry |= VA_VALID_25; entry 263 drivers/net/dsa/b53/b53_common.c b53_write32(dev, B53_VLAN_PAGE, B53_VLAN_WRITE_25, entry); entry 267 drivers/net/dsa/b53/b53_common.c u16 entry = 0; entry 270 drivers/net/dsa/b53/b53_common.c entry = ((vlan->untag & VA_UNTAG_MASK_65) << entry 273 drivers/net/dsa/b53/b53_common.c b53_write16(dev, B53_VLAN_PAGE, B53_VLAN_WRITE_65, entry); entry 292 drivers/net/dsa/b53/b53_common.c u32 entry = 0; entry 296 drivers/net/dsa/b53/b53_common.c b53_read32(dev, B53_VLAN_PAGE, B53_VLAN_WRITE_25, &entry); entry 299 drivers/net/dsa/b53/b53_common.c vlan->valid = !!(entry & VA_VALID_25_R4); entry 301 drivers/net/dsa/b53/b53_common.c vlan->valid = !!(entry & VA_VALID_25); entry 302 drivers/net/dsa/b53/b53_common.c vlan->members = entry & VA_MEMBER_MASK; entry 303 drivers/net/dsa/b53/b53_common.c vlan->untag = (entry >> VA_UNTAG_S_25) & VA_UNTAG_MASK_25; entry 306 drivers/net/dsa/b53/b53_common.c u16 entry = 0; entry 310 drivers/net/dsa/b53/b53_common.c b53_read16(dev, B53_VLAN_PAGE, B53_VLAN_WRITE_65, &entry); entry 312 drivers/net/dsa/b53/b53_common.c vlan->valid = !!(entry & VA_VALID_65); entry 313 drivers/net/dsa/b53/b53_common.c vlan->members = entry & VA_MEMBER_MASK; entry 314 drivers/net/dsa/b53/b53_common.c vlan->untag = (entry >> VA_UNTAG_S_65) & VA_UNTAG_MASK_65; entry 316 drivers/net/dsa/b53/b53_common.c u32 entry = 0; entry 320 drivers/net/dsa/b53/b53_common.c b53_read32(dev, B53_ARLIO_PAGE, dev->vta_regs[2], &entry); entry 321 drivers/net/dsa/b53/b53_common.c vlan->members = entry & VTE_MEMBERS; entry 322 drivers/net/dsa/b53/b53_common.c vlan->untag = (entry >> VTE_UNTAG_S) & VTE_MEMBERS; entry 1001 drivers/net/dsa/mt7530.c struct mt7530_hw_vlan_entry *entry) entry 1006 drivers/net/dsa/mt7530.c new_members = entry->old_members | BIT(entry->port) | entry 1018 drivers/net/dsa/mt7530.c val = entry->untagged ? MT7530_VLAN_EGRESS_UNTAG : entry 1021 drivers/net/dsa/mt7530.c ETAG_CTRL_P_MASK(entry->port), entry 1022 drivers/net/dsa/mt7530.c ETAG_CTRL_P(entry->port, val)); entry 1037 drivers/net/dsa/mt7530.c struct mt7530_hw_vlan_entry *entry) entry 1042 drivers/net/dsa/mt7530.c new_members = entry->old_members & ~BIT(entry->port); entry 1067 drivers/net/dsa/mt7530.c struct mt7530_hw_vlan_entry *entry, entry 1077 drivers/net/dsa/mt7530.c entry->old_members = (val >> PORT_MEM_SHFT) & PORT_MEM_MASK; entry 1080 drivers/net/dsa/mt7530.c vlan_op(priv, entry); entry 1313 drivers/net/dsa/mv88e6xxx/chip.c struct mv88e6xxx_vtu_entry *entry) entry 1318 drivers/net/dsa/mv88e6xxx/chip.c return chip->info->ops->vtu_getnext(chip, entry); entry 1322 drivers/net/dsa/mv88e6xxx/chip.c struct mv88e6xxx_vtu_entry *entry) entry 1327 drivers/net/dsa/mv88e6xxx/chip.c return chip->info->ops->vtu_loadpurge(chip, entry); entry 1475 drivers/net/dsa/mv88e6xxx/chip.c struct mv88e6xxx_atu_entry entry; entry 1500 drivers/net/dsa/mv88e6xxx/chip.c entry.state = 0; entry 1501 drivers/net/dsa/mv88e6xxx/chip.c ether_addr_copy(entry.mac, addr); entry 1502 drivers/net/dsa/mv88e6xxx/chip.c eth_addr_dec(entry.mac); entry 1504 drivers/net/dsa/mv88e6xxx/chip.c err = mv88e6xxx_g1_atu_getnext(chip, fid, &entry); entry 1509 drivers/net/dsa/mv88e6xxx/chip.c if (!entry.state || !ether_addr_equal(entry.mac, addr)) { entry 1510 drivers/net/dsa/mv88e6xxx/chip.c memset(&entry, 0, sizeof(entry)); entry 1511 drivers/net/dsa/mv88e6xxx/chip.c ether_addr_copy(entry.mac, addr); entry 1516 drivers/net/dsa/mv88e6xxx/chip.c entry.portvec &= ~BIT(port); entry 1517 drivers/net/dsa/mv88e6xxx/chip.c if (!entry.portvec) entry 1518 drivers/net/dsa/mv88e6xxx/chip.c entry.state = 0; entry 1520 drivers/net/dsa/mv88e6xxx/chip.c entry.portvec |= BIT(port); entry 1521 drivers/net/dsa/mv88e6xxx/chip.c entry.state = state; entry 1524 drivers/net/dsa/mv88e6xxx/chip.c return mv88e6xxx_g1_atu_loadpurge(chip, fid, &entry); entry 502 drivers/net/dsa/mv88e6xxx/chip.h struct mv88e6xxx_vtu_entry *entry); entry 504 drivers/net/dsa/mv88e6xxx/chip.h struct mv88e6xxx_vtu_entry *entry); entry 314 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_atu_entry *entry); entry 316 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_atu_entry *entry); entry 324 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_vtu_entry *entry); entry 326 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_vtu_entry *entry); entry 328 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_vtu_entry *entry); entry 330 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_vtu_entry *entry); entry 332 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_vtu_entry *entry); entry 334 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_vtu_entry *entry); entry 336 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_vtu_entry *entry); entry 338 drivers/net/dsa/mv88e6xxx/global1.h struct mv88e6xxx_vtu_entry *entry); entry 128 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry *entry) entry 137 drivers/net/dsa/mv88e6xxx/global1_atu.c entry->state = val & 0xf; entry 138 drivers/net/dsa/mv88e6xxx/global1_atu.c if (entry->state) { entry 139 drivers/net/dsa/mv88e6xxx/global1_atu.c entry->trunk = !!(val & MV88E6XXX_G1_ATU_DATA_TRUNK); entry 140 drivers/net/dsa/mv88e6xxx/global1_atu.c entry->portvec = (val >> 4) & mv88e6xxx_port_mask(chip); entry 147 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry *entry) entry 149 drivers/net/dsa/mv88e6xxx/global1_atu.c u16 data = entry->state & 0xf; entry 151 drivers/net/dsa/mv88e6xxx/global1_atu.c if (entry->state) { entry 152 drivers/net/dsa/mv88e6xxx/global1_atu.c if (entry->trunk) entry 155 drivers/net/dsa/mv88e6xxx/global1_atu.c data |= (entry->portvec & mv88e6xxx_port_mask(chip)) << 4; entry 167 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry *entry) entry 177 drivers/net/dsa/mv88e6xxx/global1_atu.c entry->mac[i * 2] = val >> 8; entry 178 drivers/net/dsa/mv88e6xxx/global1_atu.c entry->mac[i * 2 + 1] = val & 0xff; entry 185 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry *entry) entry 191 drivers/net/dsa/mv88e6xxx/global1_atu.c val = (entry->mac[i * 2] << 8) | entry->mac[i * 2 + 1]; entry 203 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry *entry) entry 212 drivers/net/dsa/mv88e6xxx/global1_atu.c if (!entry->state) { entry 213 drivers/net/dsa/mv88e6xxx/global1_atu.c err = mv88e6xxx_g1_atu_mac_write(chip, entry); entry 222 drivers/net/dsa/mv88e6xxx/global1_atu.c err = mv88e6xxx_g1_atu_data_read(chip, entry); entry 226 drivers/net/dsa/mv88e6xxx/global1_atu.c return mv88e6xxx_g1_atu_mac_read(chip, entry); entry 230 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry *entry) entry 238 drivers/net/dsa/mv88e6xxx/global1_atu.c err = mv88e6xxx_g1_atu_mac_write(chip, entry); entry 242 drivers/net/dsa/mv88e6xxx/global1_atu.c err = mv88e6xxx_g1_atu_data_write(chip, entry); entry 250 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry *entry, entry 260 drivers/net/dsa/mv88e6xxx/global1_atu.c err = mv88e6xxx_g1_atu_data_write(chip, entry); entry 279 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry entry = { entry 283 drivers/net/dsa/mv88e6xxx/global1_atu.c return mv88e6xxx_g1_atu_flushmove(chip, fid, &entry, all); entry 289 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry entry = { 0 }; entry 299 drivers/net/dsa/mv88e6xxx/global1_atu.c entry.state = 0xf, /* Full EntryState means Move */ entry 300 drivers/net/dsa/mv88e6xxx/global1_atu.c entry.portvec = from_port & mask; entry 301 drivers/net/dsa/mv88e6xxx/global1_atu.c entry.portvec |= (to_port & mask) << shift; entry 303 drivers/net/dsa/mv88e6xxx/global1_atu.c return mv88e6xxx_g1_atu_flushmove(chip, fid, &entry, all); entry 318 drivers/net/dsa/mv88e6xxx/global1_atu.c struct mv88e6xxx_atu_entry entry; entry 334 drivers/net/dsa/mv88e6xxx/global1_atu.c err = mv88e6xxx_g1_atu_data_read(chip, &entry); entry 338 drivers/net/dsa/mv88e6xxx/global1_atu.c err = mv88e6xxx_g1_atu_mac_read(chip, &entry); entry 342 drivers/net/dsa/mv88e6xxx/global1_atu.c spid = entry.state; entry 347 drivers/net/dsa/mv88e6xxx/global1_atu.c entry.mac); entry 353 drivers/net/dsa/mv88e6xxx/global1_atu.c entry.mac, entry.portvec, spid); entry 360 drivers/net/dsa/mv88e6xxx/global1_atu.c entry.mac, entry.portvec, spid); entry 367 drivers/net/dsa/mv88e6xxx/global1_atu.c entry.mac, entry.portvec, spid); entry 20 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 29 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->fid = val & MV88E6352_G1_VTU_FID_MASK; entry 35 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 37 drivers/net/dsa/mv88e6xxx/global1_vtu.c u16 val = entry->fid & MV88E6352_G1_VTU_FID_MASK; entry 45 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 54 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->sid = val & MV88E6352_G1_VTU_SID_MASK; entry 60 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 62 drivers/net/dsa/mv88e6xxx/global1_vtu.c u16 val = entry->sid & MV88E6352_G1_VTU_SID_MASK; entry 91 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 100 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->vid = val & 0xfff; entry 103 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->vid |= 0x1000; entry 105 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->valid = !!(val & MV88E6XXX_G1_VTU_VID_VALID); entry 111 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 113 drivers/net/dsa/mv88e6xxx/global1_vtu.c u16 val = entry->vid & 0xfff; entry 115 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->vid & 0x1000) entry 118 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) entry 130 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 150 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->member[i] = (regs[i / 4] >> member_offset) & 0x3; entry 151 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->state[i] = (regs[i / 4] >> state_offset) & 0x3; entry 158 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 168 drivers/net/dsa/mv88e6xxx/global1_vtu.c regs[i / 4] |= (entry->member[i] & 0x3) << member_offset; entry 169 drivers/net/dsa/mv88e6xxx/global1_vtu.c regs[i / 4] |= (entry->state[i] & 0x3) << state_offset; entry 238 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 242 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_sid_write(chip, entry); entry 250 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_sid_read(chip, entry); entry 254 drivers/net/dsa/mv88e6xxx/global1_vtu.c return mv88e6xxx_g1_vtu_vid_read(chip, entry); entry 280 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 295 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (!entry->valid) { entry 296 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_vid_write(chip, entry); entry 305 drivers/net/dsa/mv88e6xxx/global1_vtu.c return mv88e6xxx_g1_vtu_vid_read(chip, entry); entry 309 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 314 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_getnext(chip, entry); entry 318 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) { entry 319 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6185_g1_vtu_data_read(chip, entry); entry 330 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->fid = val & 0x000f; entry 331 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->fid |= (val & 0x0300) >> 4; entry 338 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 343 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_getnext(chip, entry); entry 347 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) { entry 348 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6185_g1_vtu_data_read(chip, entry); entry 359 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->fid = val & 0x000f; entry 360 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry->fid |= (val & 0x0f00) >> 4; entry 367 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 372 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_getnext(chip, entry); entry 376 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) { entry 378 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_stu_get(chip, entry); entry 382 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6185_g1_vtu_data_read(chip, entry); entry 386 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_fid_read(chip, entry); entry 395 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 400 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_getnext(chip, entry); entry 404 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) { entry 405 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6390_g1_vtu_data_read(chip, entry->member); entry 410 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_stu_get(chip, entry); entry 414 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6390_g1_vtu_data_read(chip, entry->state); entry 418 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_fid_read(chip, entry); entry 427 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 436 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_vid_write(chip, entry); entry 440 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) { entry 441 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6185_g1_vtu_data_write(chip, entry); entry 448 drivers/net/dsa/mv88e6xxx/global1_vtu.c op |= entry->fid & 0x000f; entry 449 drivers/net/dsa/mv88e6xxx/global1_vtu.c op |= (entry->fid & 0x0030) << 4; entry 456 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 465 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_vid_write(chip, entry); entry 469 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) { entry 470 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6185_g1_vtu_data_write(chip, entry); entry 477 drivers/net/dsa/mv88e6xxx/global1_vtu.c op |= entry->fid & 0x000f; entry 478 drivers/net/dsa/mv88e6xxx/global1_vtu.c op |= (entry->fid & 0x00f0) << 4; entry 485 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 493 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_vid_write(chip, entry); entry 497 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) { entry 499 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6185_g1_vtu_data_write(chip, entry); entry 503 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_sid_write(chip, entry); entry 513 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_fid_write(chip, entry); entry 523 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry *entry) entry 531 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_vid_write(chip, entry); entry 535 drivers/net/dsa/mv88e6xxx/global1_vtu.c if (entry->valid) { entry 537 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6390_g1_vtu_data_write(chip, entry->state); entry 541 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_sid_write(chip, entry); entry 552 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6390_g1_vtu_data_write(chip, entry->member); entry 556 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_fid_write(chip, entry); entry 579 drivers/net/dsa/mv88e6xxx/global1_vtu.c struct mv88e6xxx_vtu_entry entry; entry 594 drivers/net/dsa/mv88e6xxx/global1_vtu.c err = mv88e6xxx_g1_vtu_vid_read(chip, &entry); entry 602 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry.vid, spid); entry 608 drivers/net/dsa/mv88e6xxx/global1_vtu.c entry.vid, spid); entry 183 drivers/net/dsa/sja1105/sja1105.h int index, void *entry); entry 186 drivers/net/dsa/sja1105/sja1105.h int index, void *entry, bool keep); entry 255 drivers/net/dsa/sja1105/sja1105_dynamic_config.c struct sja1105_l2_lookup_entry *entry = entry_ptr; entry 259 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(cmd, &entry->lockeds, 28, 28, size, op); entry 283 drivers/net/dsa/sja1105/sja1105_dynamic_config.c struct sja1105_l2_lookup_entry *entry = entry_ptr; entry 287 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(cmd, &entry->lockeds, 28, 28, size, op); entry 307 drivers/net/dsa/sja1105/sja1105_dynamic_config.c struct sja1105_mgmt_entry *entry = entry_ptr; entry 316 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->tsreg, 85, 85, size, op); entry 317 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->takets, 84, 84, size, op); entry 318 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->macaddr, 83, 36, size, op); entry 319 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->destports, 35, 31, size, op); entry 320 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->enfport, 30, 30, size, op); entry 340 drivers/net/dsa/sja1105/sja1105_dynamic_config.c struct sja1105_mgmt_entry *entry = entry_ptr; entry 346 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->tsreg, 71, 71, size, op); entry 347 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->takets, 70, 70, size, op); entry 348 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->macaddr, 69, 22, size, op); entry 349 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->destports, 21, 17, size, op); entry 350 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->enfport, 16, 16, size, op); entry 404 drivers/net/dsa/sja1105/sja1105_dynamic_config.c struct sja1105_mac_config_entry *entry = entry_ptr; entry 409 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->speed, 30, 29, size, op); entry 410 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->drpdtag, 23, 23, size, op); entry 411 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->drpuntag, 22, 22, size, op); entry 412 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->retag, 21, 21, size, op); entry 413 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->dyn_learn, 20, 20, size, op); entry 414 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->egress, 19, 19, size, op); entry 415 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->ingress, 18, 18, size, op); entry 416 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->ing_mirr, 17, 17, size, op); entry 417 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->egr_mirr, 16, 16, size, op); entry 418 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->vlanprio, 14, 12, size, op); entry 419 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg1, &entry->vlanid, 11, 0, size, op); entry 420 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg2, &entry->tp_delin, 31, 16, size, op); entry 421 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(reg2, &entry->tp_delout, 15, 0, size, op); entry 454 drivers/net/dsa/sja1105/sja1105_dynamic_config.c struct sja1105_l2_lookup_params_entry *entry = entry_ptr; entry 456 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->poly, 7, 0, entry 476 drivers/net/dsa/sja1105/sja1105_dynamic_config.c struct sja1105_general_params_entry *entry = entry_ptr; entry 479 drivers/net/dsa/sja1105/sja1105_dynamic_config.c sja1105_packing(buf, &entry->mirr_port, 2, 0, size, op); entry 645 drivers/net/dsa/sja1105/sja1105_dynamic_config.c int index, void *entry) entry 686 drivers/net/dsa/sja1105/sja1105_dynamic_config.c ops->entry_packing(packed_buf, entry, PACK); entry 723 drivers/net/dsa/sja1105/sja1105_dynamic_config.c if (entry) entry 724 drivers/net/dsa/sja1105/sja1105_dynamic_config.c ops->entry_packing(packed_buf, entry, UNPACK); entry 730 drivers/net/dsa/sja1105/sja1105_dynamic_config.c int index, void *entry, bool keep) entry 771 drivers/net/dsa/sja1105/sja1105_dynamic_config.c ops->entry_packing(packed_buf, entry, PACK); entry 98 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_avb_params_entry *entry = entry_ptr; entry 100 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->destmeta, 95, 48, size, op); entry 101 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->srcmeta, 47, 0, size, op); entry 109 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_avb_params_entry *entry = entry_ptr; entry 111 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->destmeta, 125, 78, size, op); entry 112 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->srcmeta, 77, 30, size, op); entry 120 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_general_params_entry *entry = entry_ptr; entry 122 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vllupformat, 319, 319, size, op); entry 123 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mirr_ptacu, 318, 318, size, op); entry 124 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->switchid, 317, 315, size, op); entry 125 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->hostprio, 314, 312, size, op); entry 126 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mac_fltres1, 311, 264, size, op); entry 127 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mac_fltres0, 263, 216, size, op); entry 128 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mac_flt1, 215, 168, size, op); entry 129 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mac_flt0, 167, 120, size, op); entry 130 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->incl_srcpt1, 119, 119, size, op); entry 131 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->incl_srcpt0, 118, 118, size, op); entry 132 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->send_meta1, 117, 117, size, op); entry 133 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->send_meta0, 116, 116, size, op); entry 134 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->casc_port, 115, 113, size, op); entry 135 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->host_port, 112, 110, size, op); entry 136 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mirr_port, 109, 107, size, op); entry 137 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlmarker, 106, 75, size, op); entry 138 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlmask, 74, 43, size, op); entry 139 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tpid, 42, 27, size, op); entry 140 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ignore2stf, 26, 26, size, op); entry 141 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tpid2, 25, 10, size, op); entry 153 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_general_params_entry *entry = entry_ptr; entry 155 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vllupformat, 351, 351, size, op); entry 156 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mirr_ptacu, 350, 350, size, op); entry 157 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->switchid, 349, 347, size, op); entry 158 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->hostprio, 346, 344, size, op); entry 159 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mac_fltres1, 343, 296, size, op); entry 160 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mac_fltres0, 295, 248, size, op); entry 161 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mac_flt1, 247, 200, size, op); entry 162 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mac_flt0, 199, 152, size, op); entry 163 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->incl_srcpt1, 151, 151, size, op); entry 164 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->incl_srcpt0, 150, 150, size, op); entry 165 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->send_meta1, 149, 149, size, op); entry 166 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->send_meta0, 148, 148, size, op); entry 167 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->casc_port, 147, 145, size, op); entry 168 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->host_port, 144, 142, size, op); entry 169 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mirr_port, 141, 139, size, op); entry 170 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlmarker, 138, 107, size, op); entry 171 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlmask, 106, 75, size, op); entry 172 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tpid2, 74, 59, size, op); entry 173 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ignore2stf, 58, 58, size, op); entry 174 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tpid, 57, 42, size, op); entry 175 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->queue_ts, 41, 41, size, op); entry 176 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->egrmirrvid, 40, 29, size, op); entry 177 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->egrmirrpcp, 28, 26, size, op); entry 178 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->egrmirrdei, 25, 25, size, op); entry 179 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->replay_port, 24, 22, size, op); entry 188 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_l2_forwarding_params_entry *entry = entry_ptr; entry 191 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->max_dynp, 95, 93, size, op); entry 193 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->part_spc[i], entry 202 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_l2_forwarding_entry *entry = entry_ptr; entry 205 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->bc_domain, 63, 59, size, op); entry 206 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->reach_port, 58, 54, size, op); entry 207 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->fl_domain, 53, 49, size, op); entry 209 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlan_pmap[i], entry 219 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_l2_lookup_params_entry *entry = entry_ptr; entry 221 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->maxage, 31, 17, size, op); entry 222 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->dyn_tbsz, 16, 14, size, op); entry 223 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->poly, 13, 6, size, op); entry 224 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->shared_learn, 5, 5, size, op); entry 225 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->no_enf_hostprt, 4, 4, size, op); entry 226 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->no_mgmt_learn, 3, 3, size, op); entry 235 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_l2_lookup_params_entry *entry = entry_ptr; entry 239 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->maxaddrp[i], entry 241 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->maxage, 57, 43, size, op); entry 242 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->start_dynspc, 42, 33, size, op); entry 243 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->drpnolearn, 32, 28, size, op); entry 244 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->shared_learn, 27, 27, size, op); entry 245 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->no_enf_hostprt, 26, 26, size, op); entry 246 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->no_mgmt_learn, 25, 25, size, op); entry 247 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->use_static, 24, 24, size, op); entry 248 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->owr_dyn, 23, 23, size, op); entry 249 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->learn_once, 22, 22, size, op); entry 257 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_l2_lookup_entry *entry = entry_ptr; entry 259 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlanid, 95, 84, size, op); entry 260 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->macaddr, 83, 36, size, op); entry 261 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->destports, 35, 31, size, op); entry 262 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->enfport, 30, 30, size, op); entry 263 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->index, 29, 20, size, op); entry 271 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_l2_lookup_entry *entry = entry_ptr; entry 273 drivers/net/dsa/sja1105/sja1105_static_config.c if (entry->lockeds) { entry 274 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tsreg, 159, 159, size, op); entry 275 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mirrvlan, 158, 147, size, op); entry 276 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->takets, 146, 146, size, op); entry 277 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mirr, 145, 145, size, op); entry 278 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->retag, 144, 144, size, op); entry 280 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->touched, 159, 159, size, op); entry 281 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->age, 158, 144, size, op); entry 283 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mask_iotag, 143, 143, size, op); entry 284 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mask_vlanid, 142, 131, size, op); entry 285 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->mask_macaddr, 130, 83, size, op); entry 286 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->iotag, 82, 82, size, op); entry 287 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlanid, 81, 70, size, op); entry 288 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->macaddr, 69, 22, size, op); entry 289 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->destports, 21, 17, size, op); entry 290 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->enfport, 16, 16, size, op); entry 291 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->index, 15, 6, size, op); entry 299 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_l2_policing_entry *entry = entry_ptr; entry 301 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->sharindx, 63, 58, size, op); entry 302 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->smax, 57, 42, size, op); entry 303 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->rate, 41, 26, size, op); entry 304 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->maxlen, 25, 15, size, op); entry 305 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->partition, 14, 12, size, op); entry 313 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_mac_config_entry *entry = entry_ptr; entry 317 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->enabled[i], entry 319 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->base[i], entry 321 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->top[i], entry 324 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ifg, 71, 67, size, op); entry 325 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->speed, 66, 65, size, op); entry 326 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tp_delin, 64, 49, size, op); entry 327 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tp_delout, 48, 33, size, op); entry 328 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->maxage, 32, 25, size, op); entry 329 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlanprio, 24, 22, size, op); entry 330 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlanid, 21, 10, size, op); entry 331 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ing_mirr, 9, 9, size, op); entry 332 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->egr_mirr, 8, 8, size, op); entry 333 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->drpnona664, 7, 7, size, op); entry 334 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->drpdtag, 6, 6, size, op); entry 335 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->drpuntag, 5, 5, size, op); entry 336 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->retag, 4, 4, size, op); entry 337 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->dyn_learn, 3, 3, size, op); entry 338 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->egress, 2, 2, size, op); entry 339 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ingress, 1, 1, size, op); entry 347 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_mac_config_entry *entry = entry_ptr; entry 351 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->enabled[i], entry 353 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->base[i], entry 355 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->top[i], entry 358 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ifg, 103, 99, size, op); entry 359 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->speed, 98, 97, size, op); entry 360 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tp_delin, 96, 81, size, op); entry 361 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tp_delout, 80, 65, size, op); entry 362 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->maxage, 64, 57, size, op); entry 363 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlanprio, 56, 54, size, op); entry 364 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlanid, 53, 42, size, op); entry 365 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ing_mirr, 41, 41, size, op); entry 366 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->egr_mirr, 40, 40, size, op); entry 367 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->drpnona664, 39, 39, size, op); entry 368 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->drpdtag, 38, 38, size, op); entry 369 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->drpuntag, 35, 35, size, op); entry 370 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->retag, 34, 34, size, op); entry 371 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->dyn_learn, 33, 33, size, op); entry 372 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->egress, 32, 32, size, op); entry 373 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ingress, 31, 31, size, op); entry 381 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_schedule_entry_points_params_entry *entry = entry_ptr; entry 384 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->clksrc, 31, 30, size, op); entry 385 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->actsubsch, 29, 27, size, op); entry 393 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_schedule_entry_points_entry *entry = entry_ptr; entry 396 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->subschindx, 31, 29, size, op); entry 397 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->delta, 28, 11, size, op); entry 398 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->address, 10, 1, size, op); entry 406 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_schedule_params_entry *entry = entry_ptr; entry 410 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->subscheind[i], entry 419 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_schedule_entry *entry = entry_ptr; entry 421 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->winstindex, 63, 54, size, op); entry 422 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->winend, 53, 53, size, op); entry 423 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->winst, 52, 52, size, op); entry 424 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->destports, 51, 47, size, op); entry 425 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->setvalid, 46, 46, size, op); entry 426 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->txen, 45, 45, size, op); entry 427 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->resmedia_en, 44, 44, size, op); entry 428 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->resmedia, 43, 36, size, op); entry 429 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlindex, 35, 26, size, op); entry 430 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->delta, 25, 8, size, op); entry 438 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_vlan_lookup_entry *entry = entry_ptr; entry 440 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->ving_mirr, 63, 59, size, op); entry 441 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vegr_mirr, 58, 54, size, op); entry 442 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vmemb_port, 53, 49, size, op); entry 443 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlan_bc, 48, 44, size, op); entry 444 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->tag_port, 43, 39, size, op); entry 445 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->vlanid, 38, 27, size, op); entry 453 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_xmii_params_entry *entry = entry_ptr; entry 457 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->xmii_mode[i], entry 459 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->phy_mac[i], entry 469 drivers/net/dsa/sja1105/sja1105_static_config.c struct sja1105_table_header *entry = entry_ptr; entry 471 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->block_id, 31, 24, size, op); entry 472 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->len, 55, 32, size, op); entry 473 drivers/net/dsa/sja1105/sja1105_static_config.c sja1105_packing(buf, &entry->crc, 95, 64, size, op); entry 1013 drivers/net/ethernet/3com/3c515.c int entry = vp->cur_tx % TX_RING_SIZE; entry 1028 drivers/net/ethernet/3com/3c515.c vp->tx_skbuff[entry] = skb; entry 1029 drivers/net/ethernet/3com/3c515.c vp->tx_ring[entry].next = 0; entry 1030 drivers/net/ethernet/3com/3c515.c vp->tx_ring[entry].addr = isa_virt_to_bus(skb->data); entry 1031 drivers/net/ethernet/3com/3c515.c vp->tx_ring[entry].length = skb->len | 0x80000000; entry 1032 drivers/net/ethernet/3com/3c515.c vp->tx_ring[entry].status = skb->len | 0x80000000; entry 1041 drivers/net/ethernet/3com/3c515.c prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]); entry 1043 drivers/net/ethernet/3com/3c515.c outl(isa_virt_to_bus(&vp->tx_ring[entry]), entry 1176 drivers/net/ethernet/3com/3c515.c int entry = dirty_tx % TX_RING_SIZE; entry 1177 drivers/net/ethernet/3com/3c515.c if (inl(ioaddr + DownListPtr) == isa_virt_to_bus(&lp->tx_ring[entry])) entry 1179 drivers/net/ethernet/3com/3c515.c if (lp->tx_skbuff[entry]) { entry 1180 drivers/net/ethernet/3com/3c515.c dev_consume_skb_irq(lp->tx_skbuff[entry]); entry 1181 drivers/net/ethernet/3com/3c515.c lp->tx_skbuff[entry] = NULL; entry 1333 drivers/net/ethernet/3com/3c515.c int entry = vp->cur_rx % RX_RING_SIZE; entry 1340 drivers/net/ethernet/3com/3c515.c while ((rx_status = vp->rx_ring[entry].status) & RxDComplete) { entry 1374 drivers/net/ethernet/3com/3c515.c isa_bus_to_virt(vp->rx_ring[entry].addr), entry 1380 drivers/net/ethernet/3com/3c515.c skb = vp->rx_skbuff[entry]; entry 1381 drivers/net/ethernet/3com/3c515.c vp->rx_skbuff[entry] = NULL; entry 1384 drivers/net/ethernet/3com/3c515.c if (isa_bus_to_virt(vp->rx_ring[entry].addr) != temp) entry 1387 drivers/net/ethernet/3com/3c515.c isa_bus_to_virt(vp->rx_ring[entry].addr), entry 1395 drivers/net/ethernet/3com/3c515.c entry = (++vp->cur_rx) % RX_RING_SIZE; entry 1400 drivers/net/ethernet/3com/3c515.c entry = vp->dirty_rx % RX_RING_SIZE; entry 1401 drivers/net/ethernet/3com/3c515.c if (vp->rx_skbuff[entry] == NULL) { entry 1406 drivers/net/ethernet/3com/3c515.c vp->rx_ring[entry].addr = isa_virt_to_bus(skb->data); entry 1407 drivers/net/ethernet/3com/3c515.c vp->rx_skbuff[entry] = skb; entry 1409 drivers/net/ethernet/3com/3c515.c vp->rx_ring[entry].status = 0; /* Clear complete bit. */ entry 2114 drivers/net/ethernet/3com/3c59x.c int entry = vp->cur_tx % TX_RING_SIZE; entry 2143 drivers/net/ethernet/3com/3c59x.c vp->tx_skbuff[entry] = skb; entry 2145 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].next = 0; entry 2148 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded); entry 2150 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded | AddTCPChksum | AddUDPChksum); entry 2158 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr); entry 2159 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb->len | LAST_FRAG); entry 2168 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr); entry 2169 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb_headlen(skb)); entry 2181 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].frag[i+1].addr), entry 2182 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].frag[i+1].length), entry 2186 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].frag[0].addr), entry 2187 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].frag[0].length), entry 2193 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].frag[i+1].addr = entry 2197 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].frag[i+1].length = cpu_to_le32(skb_frag_size(frag)|LAST_FRAG); entry 2199 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].frag[i+1].length = cpu_to_le32(skb_frag_size(frag)); entry 2206 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].addr = cpu_to_le32(dma_addr); entry 2207 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].length = cpu_to_le32(skb->len | LAST_FRAG); entry 2208 drivers/net/ethernet/3com/3c59x.c vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded); entry 2214 drivers/net/ethernet/3com/3c59x.c prev_entry->next = cpu_to_le32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc)); entry 2216 drivers/net/ethernet/3com/3c59x.c iowrite32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc), ioaddr + DownListPtr); entry 2420 drivers/net/ethernet/3com/3c59x.c int entry = dirty_tx % TX_RING_SIZE; entry 2423 drivers/net/ethernet/3com/3c59x.c vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc)) entry 2426 drivers/net/ethernet/3com/3c59x.c if ((vp->tx_ring[entry].status & DN_COMPLETE) == 0) entry 2430 drivers/net/ethernet/3com/3c59x.c if (vp->tx_skbuff[entry]) { entry 2431 drivers/net/ethernet/3com/3c59x.c struct sk_buff *skb = vp->tx_skbuff[entry]; entry 2435 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].frag[0].addr), entry 2436 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].frag[0].length)&0xFFF, entry 2441 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].frag[i].addr), entry 2442 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].frag[i].length)&0xFFF, entry 2446 drivers/net/ethernet/3com/3c59x.c le32_to_cpu(vp->tx_ring[entry].addr), skb->len, DMA_TO_DEVICE); entry 2451 drivers/net/ethernet/3com/3c59x.c vp->tx_skbuff[entry] = NULL; entry 2592 drivers/net/ethernet/3com/3c59x.c int entry = vp->cur_rx % RX_RING_SIZE; entry 2600 drivers/net/ethernet/3com/3c59x.c while ((rx_status = le32_to_cpu(vp->rx_ring[entry].status)) & RxDComplete){ entry 2618 drivers/net/ethernet/3com/3c59x.c dma_addr_t dma = le32_to_cpu(vp->rx_ring[entry].addr); entry 2631 drivers/net/ethernet/3com/3c59x.c skb_put_data(skb, vp->rx_skbuff[entry]->data, entry 2654 drivers/net/ethernet/3com/3c59x.c skb = vp->rx_skbuff[entry]; entry 2655 drivers/net/ethernet/3com/3c59x.c vp->rx_skbuff[entry] = newskb; entry 2656 drivers/net/ethernet/3com/3c59x.c vp->rx_ring[entry].addr = cpu_to_le32(newdma); entry 2676 drivers/net/ethernet/3com/3c59x.c vp->rx_ring[entry].status = 0; /* Clear complete bit. */ entry 2678 drivers/net/ethernet/3com/3c59x.c entry = (++vp->cur_rx) % RX_RING_SIZE; entry 1190 drivers/net/ethernet/adaptec/starfire.c unsigned int entry; entry 1212 drivers/net/ethernet/adaptec/starfire.c entry = np->cur_tx % TX_RING_SIZE; entry 1218 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].skb = skb; entry 1220 drivers/net/ethernet/adaptec/starfire.c if (entry >= TX_RING_SIZE - skb_num_frags(skb)) { entry 1234 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping = entry 1239 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping = entry 1246 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping)) { entry 1251 drivers/net/ethernet/adaptec/starfire.c np->tx_ring[entry].addr = cpu_to_dma(np->tx_info[entry].mapping); entry 1252 drivers/net/ethernet/adaptec/starfire.c np->tx_ring[entry].status = cpu_to_le32(status); entry 1256 drivers/net/ethernet/adaptec/starfire.c entry, status); entry 1258 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].used_slots = TX_RING_SIZE - entry; entry 1259 drivers/net/ethernet/adaptec/starfire.c np->cur_tx += np->tx_info[entry].used_slots; entry 1260 drivers/net/ethernet/adaptec/starfire.c entry = 0; entry 1262 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].used_slots = 1; entry 1263 drivers/net/ethernet/adaptec/starfire.c np->cur_tx += np->tx_info[entry].used_slots; entry 1264 drivers/net/ethernet/adaptec/starfire.c entry++; entry 1277 drivers/net/ethernet/adaptec/starfire.c writel(entry * (sizeof(starfire_tx_desc) / 8), np->base + TxProducerIdx); entry 1286 drivers/net/ethernet/adaptec/starfire.c entry = prev_tx % TX_RING_SIZE; entry 1287 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].skb = NULL; entry 1290 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping, entry 1293 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping = 0; entry 1294 drivers/net/ethernet/adaptec/starfire.c entry = (entry + np->tx_info[entry].used_slots) % TX_RING_SIZE; entry 1297 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping, entry 1301 drivers/net/ethernet/adaptec/starfire.c entry++; entry 1371 drivers/net/ethernet/adaptec/starfire.c u16 entry = (tx_status & 0x7fff) / sizeof(starfire_tx_desc); entry 1372 drivers/net/ethernet/adaptec/starfire.c struct sk_buff *skb = np->tx_info[entry].skb; entry 1373 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].skb = NULL; entry 1375 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping, entry 1378 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping = 0; entry 1379 drivers/net/ethernet/adaptec/starfire.c np->dirty_tx += np->tx_info[entry].used_slots; entry 1380 drivers/net/ethernet/adaptec/starfire.c entry = (entry + np->tx_info[entry].used_slots) % TX_RING_SIZE; entry 1385 drivers/net/ethernet/adaptec/starfire.c np->tx_info[entry].mapping, entry 1389 drivers/net/ethernet/adaptec/starfire.c entry++; entry 1448 drivers/net/ethernet/adaptec/starfire.c int entry; entry 1470 drivers/net/ethernet/adaptec/starfire.c entry = (desc_status >> 16) & 0x7ff; entry 1480 drivers/net/ethernet/adaptec/starfire.c np->rx_info[entry].mapping, entry 1482 drivers/net/ethernet/adaptec/starfire.c skb_copy_to_linear_data(skb, np->rx_info[entry].skb->data, pkt_len); entry 1484 drivers/net/ethernet/adaptec/starfire.c np->rx_info[entry].mapping, entry 1488 drivers/net/ethernet/adaptec/starfire.c pci_unmap_single(np->pci_dev, np->rx_info[entry].mapping, np->rx_buf_sz, PCI_DMA_FROMDEVICE); entry 1489 drivers/net/ethernet/adaptec/starfire.c skb = np->rx_info[entry].skb; entry 1491 drivers/net/ethernet/adaptec/starfire.c np->rx_info[entry].skb = NULL; entry 1492 drivers/net/ethernet/adaptec/starfire.c np->rx_info[entry].mapping = 0; entry 1595 drivers/net/ethernet/adaptec/starfire.c int entry = -1; entry 1599 drivers/net/ethernet/adaptec/starfire.c entry = np->dirty_rx % RX_RING_SIZE; entry 1600 drivers/net/ethernet/adaptec/starfire.c if (np->rx_info[entry].skb == NULL) { entry 1602 drivers/net/ethernet/adaptec/starfire.c np->rx_info[entry].skb = skb; entry 1605 drivers/net/ethernet/adaptec/starfire.c np->rx_info[entry].mapping = entry 1608 drivers/net/ethernet/adaptec/starfire.c np->rx_info[entry].mapping)) { entry 1610 drivers/net/ethernet/adaptec/starfire.c np->rx_info[entry].skb = NULL; entry 1613 drivers/net/ethernet/adaptec/starfire.c np->rx_ring[entry].rxaddr = entry 1614 drivers/net/ethernet/adaptec/starfire.c cpu_to_dma(np->rx_info[entry].mapping | RxDescValid); entry 1616 drivers/net/ethernet/adaptec/starfire.c if (entry == RX_RING_SIZE - 1) entry 1617 drivers/net/ethernet/adaptec/starfire.c np->rx_ring[entry].rxaddr |= cpu_to_dma(RxDescEndRing); entry 1619 drivers/net/ethernet/adaptec/starfire.c if (entry >= 0) entry 1620 drivers/net/ethernet/adaptec/starfire.c writew(entry, np->base + RxDescQIdx); entry 1545 drivers/net/ethernet/agere/et131x.c u32 entry; entry 1597 drivers/net/ethernet/agere/et131x.c for (entry = 0; entry < fbr->num_entries; entry++) { entry 1598 drivers/net/ethernet/agere/et131x.c fbr_entry->addr_hi = fbr->bus_high[entry]; entry 1599 drivers/net/ethernet/agere/et131x.c fbr_entry->addr_lo = fbr->bus_low[entry]; entry 1600 drivers/net/ethernet/agere/et131x.c fbr_entry->word2 = entry; entry 423 drivers/net/ethernet/altera/altera_sgdma.c #define list_remove_head(list, entry, type, member) \ entry 425 drivers/net/ethernet/altera/altera_sgdma.c entry = NULL; \ entry 427 drivers/net/ethernet/altera/altera_sgdma.c entry = list_entry((list)->next, type, member); \ entry 428 drivers/net/ethernet/altera/altera_sgdma.c list_del_init(&entry->member); \ entry 432 drivers/net/ethernet/altera/altera_sgdma.c #define list_peek_head(list, entry, type, member) \ entry 434 drivers/net/ethernet/altera/altera_sgdma.c entry = NULL; \ entry 436 drivers/net/ethernet/altera/altera_sgdma.c entry = list_entry((list)->next, type, member); \ entry 335 drivers/net/ethernet/altera/altera_tse_main.c unsigned int entry; entry 340 drivers/net/ethernet/altera/altera_tse_main.c entry = priv->rx_prod % rxsize; entry 341 drivers/net/ethernet/altera/altera_tse_main.c if (likely(priv->rx_ring[entry].skb == NULL)) { entry 342 drivers/net/ethernet/altera/altera_tse_main.c ret = tse_init_rx_buffer(priv, &priv->rx_ring[entry], entry 346 drivers/net/ethernet/altera/altera_tse_main.c priv->dmaops->add_rx_desc(priv, &priv->rx_ring[entry]); entry 373 drivers/net/ethernet/altera/altera_tse_main.c unsigned int entry = priv->rx_cons % priv->rx_ring_size; entry 402 drivers/net/ethernet/altera/altera_tse_main.c skb = priv->rx_ring[entry].skb; entry 410 drivers/net/ethernet/altera/altera_tse_main.c priv->rx_ring[entry].skb = NULL; entry 414 drivers/net/ethernet/altera/altera_tse_main.c dma_unmap_single(priv->device, priv->rx_ring[entry].dma_addr, entry 415 drivers/net/ethernet/altera/altera_tse_main.c priv->rx_ring[entry].len, DMA_FROM_DEVICE); entry 434 drivers/net/ethernet/altera/altera_tse_main.c entry = next_entry; entry 448 drivers/net/ethernet/altera/altera_tse_main.c unsigned int entry; entry 458 drivers/net/ethernet/altera/altera_tse_main.c entry = priv->tx_cons % txsize; entry 459 drivers/net/ethernet/altera/altera_tse_main.c tx_buff = &priv->tx_ring[entry]; entry 561 drivers/net/ethernet/altera/altera_tse_main.c unsigned int entry; entry 583 drivers/net/ethernet/altera/altera_tse_main.c entry = priv->tx_prod % txsize; entry 584 drivers/net/ethernet/altera/altera_tse_main.c buffer = &priv->tx_ring[entry]; entry 543 drivers/net/ethernet/amd/7990.c int entry, skblen, len; entry 569 drivers/net/ethernet/amd/7990.c entry = lp->tx_new & lp->tx_ring_mod_mask; entry 570 drivers/net/ethernet/amd/7990.c ib->btx_ring[entry].length = (-len) | 0xf000; entry 571 drivers/net/ethernet/amd/7990.c ib->btx_ring[entry].misc = 0; entry 574 drivers/net/ethernet/amd/7990.c memset((void *)&ib->tx_buf[entry][0], 0, ETH_ZLEN); entry 575 drivers/net/ethernet/amd/7990.c skb_copy_from_linear_data(skb, (void *)&ib->tx_buf[entry][0], skblen); entry 578 drivers/net/ethernet/amd/7990.c ib->btx_ring[entry].tmd1_bits = (LE_T1_POK|LE_T1_OWN); entry 541 drivers/net/ethernet/amd/a2065.c int entry, skblen; entry 559 drivers/net/ethernet/amd/a2065.c entry = lp->tx_new & lp->tx_ring_mod_mask; entry 560 drivers/net/ethernet/amd/a2065.c ib->btx_ring[entry].length = (-skblen) | 0xf000; entry 561 drivers/net/ethernet/amd/a2065.c ib->btx_ring[entry].misc = 0; entry 563 drivers/net/ethernet/amd/a2065.c skb_copy_from_linear_data(skb, (void *)&ib->tx_buf[entry][0], skblen); entry 566 drivers/net/ethernet/amd/a2065.c ib->btx_ring[entry].tmd1_bits = (LE_T1_POK | LE_T1_OWN); entry 164 drivers/net/ethernet/amd/ariadne.c int entry = priv->cur_rx % RX_RING_SIZE; entry 168 drivers/net/ethernet/amd/ariadne.c while (!(lowb(priv->rx_ring[entry]->RMD1) & RF_OWN)) { entry 169 drivers/net/ethernet/amd/ariadne.c int status = lowb(priv->rx_ring[entry]->RMD1); entry 189 drivers/net/ethernet/amd/ariadne.c priv->rx_ring[entry]->RMD1 &= 0xff00 | RF_STP | RF_ENP; entry 192 drivers/net/ethernet/amd/ariadne.c short pkt_len = swapw(priv->rx_ring[entry]->RMD3); entry 198 drivers/net/ethernet/amd/ariadne.c if (lowb(priv->rx_ring[(entry + i) % RX_RING_SIZE]->RMD1) & RF_OWN) entry 203 drivers/net/ethernet/amd/ariadne.c priv->rx_ring[entry]->RMD1 |= RF_OWN; entry 213 drivers/net/ethernet/amd/ariadne.c (const void *)priv->rx_buff[entry], entry 226 drivers/net/ethernet/amd/ariadne.c priv->rx_ring[entry]->RMD1 |= RF_OWN; entry 227 drivers/net/ethernet/amd/ariadne.c entry = (++priv->cur_rx) % RX_RING_SIZE; entry 309 drivers/net/ethernet/amd/ariadne.c int entry = dirty_tx % TX_RING_SIZE; entry 310 drivers/net/ethernet/amd/ariadne.c int status = lowb(priv->tx_ring[entry]->TMD1); entry 315 drivers/net/ethernet/amd/ariadne.c priv->tx_ring[entry]->TMD1 &= 0xff00; entry 319 drivers/net/ethernet/amd/ariadne.c int err_status = priv->tx_ring[entry]->TMD3; entry 548 drivers/net/ethernet/amd/ariadne.c int entry; entry 576 drivers/net/ethernet/amd/ariadne.c entry = priv->cur_tx % TX_RING_SIZE; entry 581 drivers/net/ethernet/amd/ariadne.c priv->tx_ring[entry]->TMD2 = swapw((u_short)-skb->len); entry 582 drivers/net/ethernet/amd/ariadne.c priv->tx_ring[entry]->TMD3 = 0x0000; entry 583 drivers/net/ethernet/amd/ariadne.c memcpyw(priv->tx_buff[entry], (u_short *)skb->data, len); entry 587 drivers/net/ethernet/amd/ariadne.c (void *)priv->tx_buff[entry], entry 591 drivers/net/ethernet/amd/ariadne.c priv->tx_ring[entry]->TMD1 = (priv->tx_ring[entry]->TMD1 & 0xff00) entry 612 drivers/net/ethernet/amd/ariadne.c if (lowb(priv->tx_ring[(entry + 1) % TX_RING_SIZE]->TMD1) != 0) { entry 778 drivers/net/ethernet/amd/atarilance.c int entry, len; entry 815 drivers/net/ethernet/amd/atarilance.c entry = lp->cur_tx & TX_RING_MOD_MASK; entry 816 drivers/net/ethernet/amd/atarilance.c head = &(MEM->tx_head[entry]); entry 838 drivers/net/ethernet/amd/atarilance.c if ((MEM->tx_head[(entry+1) & TX_RING_MOD_MASK].flag & TMD1_OWN) == entry 886 drivers/net/ethernet/amd/atarilance.c int entry = dirty_tx & TX_RING_MOD_MASK; entry 887 drivers/net/ethernet/amd/atarilance.c int status = MEM->tx_head[entry].flag; entry 892 drivers/net/ethernet/amd/atarilance.c MEM->tx_head[entry].flag = 0; entry 896 drivers/net/ethernet/amd/atarilance.c int err_status = MEM->tx_head[entry].misc; entry 965 drivers/net/ethernet/amd/atarilance.c int entry = lp->cur_rx & RX_RING_MOD_MASK; entry 969 drivers/net/ethernet/amd/atarilance.c MEM->rx_head[entry].flag )); entry 972 drivers/net/ethernet/amd/atarilance.c while( (MEM->rx_head[entry].flag & RMD1_OWN) == RMD1_OWN_HOST ) { entry 973 drivers/net/ethernet/amd/atarilance.c struct lance_rx_head *head = &(MEM->rx_head[entry]); entry 1001 drivers/net/ethernet/amd/atarilance.c if (MEM->rx_head[(entry+i) & RX_RING_MOD_MASK].flag & entry 1033 drivers/net/ethernet/amd/atarilance.c entry = (++lp->cur_rx) & RX_RING_MOD_MASK; entry 562 drivers/net/ethernet/amd/declance.c int entry, len; entry 587 drivers/net/ethernet/amd/declance.c entry = lp->rx_new; entry 615 drivers/net/ethernet/amd/declance.c ((lp->rx_buf_ptr_lnc[entry] >> 16) & entry 617 drivers/net/ethernet/amd/declance.c lp->rx_new = (entry + 1) & RX_RING_MOD_MASK; entry 626 drivers/net/ethernet/amd/declance.c lp->rx_buf_ptr_cpu[entry], len); entry 637 drivers/net/ethernet/amd/declance.c ((lp->rx_buf_ptr_lnc[entry] >> 16) & 0xff) | LE_R1_OWN; entry 638 drivers/net/ethernet/amd/declance.c lp->rx_new = (entry + 1) & RX_RING_MOD_MASK; entry 904 drivers/net/ethernet/amd/declance.c int entry, len; entry 918 drivers/net/ethernet/amd/declance.c entry = lp->tx_new; entry 919 drivers/net/ethernet/amd/declance.c *lib_ptr(ib, btx_ring[entry].length, lp->type) = (-len); entry 920 drivers/net/ethernet/amd/declance.c *lib_ptr(ib, btx_ring[entry].misc, lp->type) = 0; entry 922 drivers/net/ethernet/amd/declance.c cp_to_buf(lp->type, lp->tx_buf_ptr_cpu[entry], skb->data, len); entry 925 drivers/net/ethernet/amd/declance.c *lib_ptr(ib, btx_ring[entry].tmd1, lp->type) = entry 926 drivers/net/ethernet/amd/declance.c ((lp->tx_buf_ptr_lnc[entry] >> 16) & 0xff) | entry 928 drivers/net/ethernet/amd/declance.c lp->tx_new = (entry + 1) & TX_RING_MOD_MASK; entry 955 drivers/net/ethernet/amd/lance.c int entry; entry 970 drivers/net/ethernet/amd/lance.c entry = lp->cur_tx & TX_RING_MOD_MASK; entry 980 drivers/net/ethernet/amd/lance.c lp->tx_ring[entry].length = -ETH_ZLEN; entry 983 drivers/net/ethernet/amd/lance.c lp->tx_ring[entry].length = -skb->len; entry 985 drivers/net/ethernet/amd/lance.c lp->tx_ring[entry].length = -skb->len; entry 987 drivers/net/ethernet/amd/lance.c lp->tx_ring[entry].misc = 0x0000; entry 997 drivers/net/ethernet/amd/lance.c skb_copy_from_linear_data(skb, &lp->tx_bounce_buffs[entry], skb->len); entry 998 drivers/net/ethernet/amd/lance.c lp->tx_ring[entry].base = entry 999 drivers/net/ethernet/amd/lance.c ((u32)isa_virt_to_bus((lp->tx_bounce_buffs + entry)) & 0xffffff) | 0x83000000; entry 1002 drivers/net/ethernet/amd/lance.c lp->tx_skbuff[entry] = skb; entry 1003 drivers/net/ethernet/amd/lance.c lp->tx_ring[entry].base = ((u32)isa_virt_to_bus(skb->data) & 0xffffff) | 0x83000000; entry 1051 drivers/net/ethernet/amd/lance.c int entry = dirty_tx & TX_RING_MOD_MASK; entry 1052 drivers/net/ethernet/amd/lance.c int status = lp->tx_ring[entry].base; entry 1057 drivers/net/ethernet/amd/lance.c lp->tx_ring[entry].base = 0; entry 1061 drivers/net/ethernet/amd/lance.c int err_status = lp->tx_ring[entry].misc; entry 1086 drivers/net/ethernet/amd/lance.c if (lp->tx_skbuff[entry]) { entry 1087 drivers/net/ethernet/amd/lance.c dev_consume_skb_irq(lp->tx_skbuff[entry]); entry 1088 drivers/net/ethernet/amd/lance.c lp->tx_skbuff[entry] = NULL; entry 1147 drivers/net/ethernet/amd/lance.c int entry = lp->cur_rx & RX_RING_MOD_MASK; entry 1151 drivers/net/ethernet/amd/lance.c while (lp->rx_ring[entry].base >= 0) { entry 1152 drivers/net/ethernet/amd/lance.c int status = lp->rx_ring[entry].base >> 24; entry 1169 drivers/net/ethernet/amd/lance.c lp->rx_ring[entry].base &= 0x03ffffff; entry 1174 drivers/net/ethernet/amd/lance.c short pkt_len = (lp->rx_ring[entry].msg_length & 0xfff)-4; entry 1189 drivers/net/ethernet/amd/lance.c if (lp->rx_ring[(entry+i) & RX_RING_MOD_MASK].base < 0) entry 1195 drivers/net/ethernet/amd/lance.c lp->rx_ring[entry].base |= 0x80000000; entry 1203 drivers/net/ethernet/amd/lance.c (unsigned char *)isa_bus_to_virt((lp->rx_ring[entry].base & 0x00ffffff)), entry 1213 drivers/net/ethernet/amd/lance.c lp->rx_ring[entry].buf_length = -PKT_BUF_SZ; entry 1214 drivers/net/ethernet/amd/lance.c lp->rx_ring[entry].base |= 0x80000000; entry 1215 drivers/net/ethernet/amd/lance.c entry = (++lp->cur_rx) & RX_RING_MOD_MASK; entry 1181 drivers/net/ethernet/amd/pcnet32.c int entry) entry 1244 drivers/net/ethernet/amd/pcnet32.c skb = lp->rx_skbuff[entry]; entry 1246 drivers/net/ethernet/amd/pcnet32.c lp->rx_dma_addr[entry], entry 1250 drivers/net/ethernet/amd/pcnet32.c lp->rx_skbuff[entry] = newskb; entry 1251 drivers/net/ethernet/amd/pcnet32.c lp->rx_dma_addr[entry] = new_dma_addr; entry 1268 drivers/net/ethernet/amd/pcnet32.c lp->rx_dma_addr[entry], entry 1272 drivers/net/ethernet/amd/pcnet32.c (unsigned char *)(lp->rx_skbuff[entry]->data), entry 1275 drivers/net/ethernet/amd/pcnet32.c lp->rx_dma_addr[entry], entry 1288 drivers/net/ethernet/amd/pcnet32.c int entry = lp->cur_rx & lp->rx_mod_mask; entry 1289 drivers/net/ethernet/amd/pcnet32.c struct pcnet32_rx_head *rxp = &lp->rx_ring[entry]; entry 1294 drivers/net/ethernet/amd/pcnet32.c pcnet32_rx_entry(dev, lp, rxp, entry); entry 1303 drivers/net/ethernet/amd/pcnet32.c entry = (++lp->cur_rx) & lp->rx_mod_mask; entry 1304 drivers/net/ethernet/amd/pcnet32.c rxp = &lp->rx_ring[entry]; entry 1318 drivers/net/ethernet/amd/pcnet32.c int entry = dirty_tx & lp->tx_mod_mask; entry 1319 drivers/net/ethernet/amd/pcnet32.c int status = (short)le16_to_cpu(lp->tx_ring[entry].status); entry 1324 drivers/net/ethernet/amd/pcnet32.c lp->tx_ring[entry].base = 0; entry 1328 drivers/net/ethernet/amd/pcnet32.c int err_status = le32_to_cpu(lp->tx_ring[entry].misc); entry 1365 drivers/net/ethernet/amd/pcnet32.c if (lp->tx_skbuff[entry]) { entry 1367 drivers/net/ethernet/amd/pcnet32.c lp->tx_dma_addr[entry], entry 1368 drivers/net/ethernet/amd/pcnet32.c lp->tx_skbuff[entry]-> entry 1370 drivers/net/ethernet/amd/pcnet32.c dev_kfree_skb_any(lp->tx_skbuff[entry]); entry 1371 drivers/net/ethernet/amd/pcnet32.c lp->tx_skbuff[entry] = NULL; entry 1372 drivers/net/ethernet/amd/pcnet32.c lp->tx_dma_addr[entry] = 0; entry 2504 drivers/net/ethernet/amd/pcnet32.c int entry; entry 2521 drivers/net/ethernet/amd/pcnet32.c entry = lp->cur_tx & lp->tx_mod_mask; entry 2526 drivers/net/ethernet/amd/pcnet32.c lp->tx_ring[entry].length = cpu_to_le16(-skb->len); entry 2528 drivers/net/ethernet/amd/pcnet32.c lp->tx_ring[entry].misc = 0x00000000; entry 2530 drivers/net/ethernet/amd/pcnet32.c lp->tx_dma_addr[entry] = entry 2532 drivers/net/ethernet/amd/pcnet32.c if (pci_dma_mapping_error(lp->pci_dev, lp->tx_dma_addr[entry])) { entry 2537 drivers/net/ethernet/amd/pcnet32.c lp->tx_skbuff[entry] = skb; entry 2538 drivers/net/ethernet/amd/pcnet32.c lp->tx_ring[entry].base = cpu_to_le32(lp->tx_dma_addr[entry]); entry 2540 drivers/net/ethernet/amd/pcnet32.c lp->tx_ring[entry].status = cpu_to_le16(status); entry 2548 drivers/net/ethernet/amd/pcnet32.c if (lp->tx_ring[(entry + 1) & lp->tx_mod_mask].base != 0) { entry 519 drivers/net/ethernet/amd/sun3lance.c int entry, len; entry 613 drivers/net/ethernet/amd/sun3lance.c entry = lp->new_tx; entry 614 drivers/net/ethernet/amd/sun3lance.c head = &(MEM->tx_head[entry]); entry 644 drivers/net/ethernet/amd/sun3lance.c if ((MEM->tx_head[(entry+1) & TX_RING_MOD_MASK].flag & TMD1_OWN) == entry 785 drivers/net/ethernet/amd/sun3lance.c int entry = lp->new_rx; entry 788 drivers/net/ethernet/amd/sun3lance.c while( (MEM->rx_head[entry].flag & RMD1_OWN) == RMD1_OWN_HOST ) { entry 789 drivers/net/ethernet/amd/sun3lance.c struct lance_rx_head *head = &(MEM->rx_head[entry]); entry 841 drivers/net/ethernet/amd/sun3lance.c printk( "%s: RX pkt %d type 0x%04x len %d\n ", dev->name, entry, ((u_short *)data)[6], pkt_len); entry 862 drivers/net/ethernet/amd/sun3lance.c entry = lp->new_rx = (lp->new_rx +1) & RX_RING_MOD_MASK; entry 514 drivers/net/ethernet/amd/sunlance.c int len, entry = lp->rx_new; entry 517 drivers/net/ethernet/amd/sunlance.c for (rd = &ib->brx_ring [entry]; entry 519 drivers/net/ethernet/amd/sunlance.c rd = &ib->brx_ring [entry]) { entry 542 drivers/net/ethernet/amd/sunlance.c lp->rx_new = RX_NEXT(entry); entry 551 drivers/net/ethernet/amd/sunlance.c (unsigned char *)&(ib->rx_buf [entry][0]), entry 561 drivers/net/ethernet/amd/sunlance.c entry = RX_NEXT(entry); entry 564 drivers/net/ethernet/amd/sunlance.c lp->rx_new = entry; entry 683 drivers/net/ethernet/amd/sunlance.c int len, entry; entry 686 drivers/net/ethernet/amd/sunlance.c entry = lp->rx_new; entry 687 drivers/net/ethernet/amd/sunlance.c for (rd = &ib->brx_ring [entry]; entry 689 drivers/net/ethernet/amd/sunlance.c rd = &ib->brx_ring [entry]) { entry 712 drivers/net/ethernet/amd/sunlance.c lp->rx_new = RX_NEXT(entry); entry 720 drivers/net/ethernet/amd/sunlance.c lance_piocopy_to_skb(skb, &(ib->rx_buf[entry][0]), len); entry 729 drivers/net/ethernet/amd/sunlance.c entry = RX_NEXT(entry); entry 732 drivers/net/ethernet/amd/sunlance.c lp->rx_new = entry; entry 882 drivers/net/ethernet/amd/sunlance.c int i, entry; entry 884 drivers/net/ethernet/amd/sunlance.c entry = lp->tx_new & TX_RING_MOD_MASK; entry 887 drivers/net/ethernet/amd/sunlance.c u16 __iomem *packet = (u16 __iomem *) &(ib->tx_buf[entry][0]); entry 895 drivers/net/ethernet/amd/sunlance.c sbus_writew((-ETH_ZLEN) | 0xf000, &ib->btx_ring[entry].length); entry 896 drivers/net/ethernet/amd/sunlance.c sbus_writew(0, &ib->btx_ring[entry].misc); entry 897 drivers/net/ethernet/amd/sunlance.c sbus_writeb(LE_T1_POK|LE_T1_OWN, &ib->btx_ring[entry].tmd1_bits); entry 900 drivers/net/ethernet/amd/sunlance.c u16 *packet = (u16 *) &(ib->tx_buf[entry][0]); entry 907 drivers/net/ethernet/amd/sunlance.c ib->btx_ring[entry].length = (-ETH_ZLEN) | 0xf000; entry 908 drivers/net/ethernet/amd/sunlance.c ib->btx_ring[entry].misc = 0; entry 909 drivers/net/ethernet/amd/sunlance.c ib->btx_ring[entry].tmd1_bits = (LE_T1_POK|LE_T1_OWN); entry 911 drivers/net/ethernet/amd/sunlance.c lp->tx_new = TX_NEXT(entry); entry 1113 drivers/net/ethernet/amd/sunlance.c int entry, skblen, len; entry 1123 drivers/net/ethernet/amd/sunlance.c entry = lp->tx_new & TX_RING_MOD_MASK; entry 1126 drivers/net/ethernet/amd/sunlance.c sbus_writew((-len) | 0xf000, &ib->btx_ring[entry].length); entry 1127 drivers/net/ethernet/amd/sunlance.c sbus_writew(0, &ib->btx_ring[entry].misc); entry 1128 drivers/net/ethernet/amd/sunlance.c lance_piocopy_from_skb(&ib->tx_buf[entry][0], skb->data, skblen); entry 1130 drivers/net/ethernet/amd/sunlance.c lance_piozero(&ib->tx_buf[entry][skblen], len - skblen); entry 1131 drivers/net/ethernet/amd/sunlance.c sbus_writeb(LE_T1_POK | LE_T1_OWN, &ib->btx_ring[entry].tmd1_bits); entry 1134 drivers/net/ethernet/amd/sunlance.c ib->btx_ring [entry].length = (-len) | 0xf000; entry 1135 drivers/net/ethernet/amd/sunlance.c ib->btx_ring [entry].misc = 0; entry 1136 drivers/net/ethernet/amd/sunlance.c skb_copy_from_linear_data(skb, &ib->tx_buf [entry][0], skblen); entry 1138 drivers/net/ethernet/amd/sunlance.c memset((char *) &ib->tx_buf [entry][skblen], 0, len - skblen); entry 1139 drivers/net/ethernet/amd/sunlance.c ib->btx_ring [entry].tmd1_bits = (LE_T1_POK | LE_T1_OWN); entry 1142 drivers/net/ethernet/amd/sunlance.c lp->tx_new = TX_NEXT(entry); entry 54 drivers/net/ethernet/atheros/alx/main.c static void alx_free_txbuf(struct alx_tx_queue *txq, int entry) entry 56 drivers/net/ethernet/atheros/alx/main.c struct alx_buffer *txb = &txq->bufs[entry]; entry 978 drivers/net/ethernet/broadcom/b44.c u32 len, entry, ctrl; entry 1019 drivers/net/ethernet/broadcom/b44.c entry = bp->tx_prod; entry 1020 drivers/net/ethernet/broadcom/b44.c bp->tx_buffers[entry].skb = skb; entry 1021 drivers/net/ethernet/broadcom/b44.c bp->tx_buffers[entry].mapping = mapping; entry 1025 drivers/net/ethernet/broadcom/b44.c if (entry == (B44_TX_RING_SIZE - 1)) entry 1028 drivers/net/ethernet/broadcom/b44.c bp->tx_ring[entry].ctrl = cpu_to_le32(ctrl); entry 1029 drivers/net/ethernet/broadcom/b44.c bp->tx_ring[entry].addr = cpu_to_le32((u32) mapping+bp->dma_offset); entry 1033 drivers/net/ethernet/broadcom/b44.c entry * sizeof(bp->tx_ring[0]), entry 1036 drivers/net/ethernet/broadcom/b44.c entry = NEXT_TX(entry); entry 1038 drivers/net/ethernet/broadcom/b44.c bp->tx_prod = entry; entry 1042 drivers/net/ethernet/broadcom/b44.c bw32(bp, B44_DMATX_PTR, entry * sizeof(struct dma_desc)); entry 1044 drivers/net/ethernet/broadcom/b44.c bw32(bp, B44_DMATX_PTR, entry * sizeof(struct dma_desc)); entry 3683 drivers/net/ethernet/broadcom/bnx2.c const struct bnx2_mips_fw_file_entry *entry) entry 3685 drivers/net/ethernet/broadcom/bnx2.c if (check_fw_section(fw, &entry->text, 4, true) || entry 3686 drivers/net/ethernet/broadcom/bnx2.c check_fw_section(fw, &entry->data, 4, false) || entry 3687 drivers/net/ethernet/broadcom/bnx2.c check_fw_section(fw, &entry->rodata, 4, false)) entry 6288 drivers/net/ethernet/broadcom/bnx2.c msix_ent[i].entry = i; entry 1667 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c bp->msix_table[msix_vec].entry = msix_vec; entry 1669 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c bp->msix_table[0].entry); entry 1675 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c bp->msix_table[msix_vec].entry = msix_vec; entry 1677 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c msix_vec, bp->msix_table[msix_vec].entry); entry 1683 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c bp->msix_table[msix_vec].entry = msix_vec; entry 219 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c struct dcbx_app_priority_entry *entry = entry 223 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c if (GET_FLAGS(entry[index].appBitfield, entry 225 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c GET_FLAGS(entry[index].appBitfield, entry 228 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c } else if (GET_FLAGS(entry[index].appBitfield, entry 230 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c TCP_PORT_ISCSI == entry[index].app_id) { entry 233 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c } else if (GET_FLAGS(entry[index].appBitfield, entry 235 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c ETH_TYPE_FCOE == entry[index].app_id) { entry 244 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry[index].pri_bitmap, entry 1192 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c u8 entry = 0; entry 1206 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry = 1; entry 1209 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry = 0; entry 1215 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c bnx2x_dcbx_add_to_cos_bw(bp, &cos_data->data[entry], entry 1221 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c cos_data->data[entry].strict = entry 1504 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c u8 entry = 0; entry 1545 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry = 0; entry 1553 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry = 1; entry 1554 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c cos_data->data[entry].pri_join_mask |= entry 1557 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c &cos_data->data[entry], entry 1623 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c u8 entry, entry 1635 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c data[entry]; entry 1656 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry++; entry 1672 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c u8 entry, entry 1676 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c if (bnx2x_dcbx_spread_strict_pri(bp, cos_data, entry, entry 1680 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c data[entry]; entry 1702 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c u8 entry = 0; entry 1723 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c data[entry]; entry 1731 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry++; entry 1742 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry += bnx2x_dcbx_cee_fill_strict_pri(bp, cos_data, entry 1743 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c entry, need_num_of_entries, pg->pg_priority); entry 1748 drivers/net/ethernet/broadcom/bnx2x/bnx2x_dcb.c cos_data->num_of_cos = entry; entry 2811 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c struct code_entry *entry, entry 2814 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c size_t size = entry->code_attribute & CODE_IMAGE_LENGTH_MASK; entry 2815 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c u32 type = entry->code_attribute & CODE_IMAGE_TYPE_MASK; entry 2822 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c rc = bnx2x_nvram_crc(bp, entry->nvm_start_addr, size, buff); entry 2833 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c struct code_entry entry; entry 2835 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c rc = bnx2x_nvram_read32(bp, addr, (u32 *)&entry, sizeof(entry)); entry 2839 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c return bnx2x_test_nvram_dir(bp, &entry, buff); entry 2845 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c struct code_entry entry; entry 2850 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c sizeof(entry) * CODE_ENTRY_EXTENDED_DIR_IDX, entry 2851 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c (u32 *)&entry, sizeof(entry)); entry 2855 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c if (!EXTENDED_DIR_EXISTS(entry.code_attribute)) entry 2858 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c rc = bnx2x_nvram_read32(bp, entry.nvm_start_addr, entry 2863 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c dir_offset = entry.nvm_start_addr + 8; entry 672 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int entry) entry 676 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c return XSTORM_ASSERT_LIST_OFFSET(entry); entry 678 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c return TSTORM_ASSERT_LIST_OFFSET(entry); entry 680 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c return CSTORM_ASSERT_LIST_OFFSET(entry); entry 682 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c return USTORM_ASSERT_LIST_OFFSET(entry); entry 9147 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c u8 entry = (BP_VN(bp) + 1)*8; entry 9150 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c EMAC_WR(bp, EMAC_REG_EMAC_MAC_MATCH + entry, val); entry 9154 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c EMAC_WR(bp, EMAC_REG_EMAC_MAC_MATCH + entry + 4, val); entry 10686 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c struct bnx2x_prev_path_list *entry; entry 10691 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c entry = bnx2x_prev_path_get_entry(bp); entry 10692 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c val = !!(entry && (entry->undi & (1 << BP_PORT(bp)))); entry 682 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h bool (*get_entry)(struct bnx2x_credit_pool_obj *o, int *entry); entry 690 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h bool (*put_entry)(struct bnx2x_credit_pool_obj *o, int entry); entry 8048 drivers/net/ethernet/broadcom/bnxt/bnxt.c msix_ent[i].entry = i; entry 7705 drivers/net/ethernet/broadcom/tg3.c static bool tg3_tx_frag_set(struct tg3_napi *tnapi, u32 *entry, u32 *budget, entry 7725 drivers/net/ethernet/broadcom/tg3.c u32 prvidx = *entry; entry 7737 drivers/net/ethernet/broadcom/tg3.c tnapi->tx_buffers[*entry].fragmented = true; entry 7739 drivers/net/ethernet/broadcom/tg3.c tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, entry 7742 drivers/net/ethernet/broadcom/tg3.c prvidx = *entry; entry 7743 drivers/net/ethernet/broadcom/tg3.c *entry = NEXT_TX(*entry); entry 7750 drivers/net/ethernet/broadcom/tg3.c tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, entry 7753 drivers/net/ethernet/broadcom/tg3.c *entry = NEXT_TX(*entry); entry 7760 drivers/net/ethernet/broadcom/tg3.c tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, entry 7762 drivers/net/ethernet/broadcom/tg3.c *entry = NEXT_TX(*entry); entry 7768 drivers/net/ethernet/broadcom/tg3.c static void tg3_tx_skb_unmap(struct tg3_napi *tnapi, u32 entry, int last) entry 7772 drivers/net/ethernet/broadcom/tg3.c struct tg3_tx_ring_info *txb = &tnapi->tx_buffers[entry]; entry 7784 drivers/net/ethernet/broadcom/tg3.c entry = NEXT_TX(entry); entry 7785 drivers/net/ethernet/broadcom/tg3.c txb = &tnapi->tx_buffers[entry]; entry 7791 drivers/net/ethernet/broadcom/tg3.c entry = NEXT_TX(entry); entry 7792 drivers/net/ethernet/broadcom/tg3.c txb = &tnapi->tx_buffers[entry]; entry 7800 drivers/net/ethernet/broadcom/tg3.c entry = NEXT_TX(entry); entry 7801 drivers/net/ethernet/broadcom/tg3.c txb = &tnapi->tx_buffers[entry]; entry 7809 drivers/net/ethernet/broadcom/tg3.c u32 *entry, u32 *budget, entry 7838 drivers/net/ethernet/broadcom/tg3.c u32 save_entry = *entry; entry 7842 drivers/net/ethernet/broadcom/tg3.c tnapi->tx_buffers[*entry].skb = new_skb; entry 7843 drivers/net/ethernet/broadcom/tg3.c dma_unmap_addr_set(&tnapi->tx_buffers[*entry], entry 7846 drivers/net/ethernet/broadcom/tg3.c if (tg3_tx_frag_set(tnapi, entry, budget, new_addr, entry 7918 drivers/net/ethernet/broadcom/tg3.c u32 len, entry, base_flags, mss, vlan = 0; entry 7953 drivers/net/ethernet/broadcom/tg3.c entry = tnapi->tx_prod; entry 8065 drivers/net/ethernet/broadcom/tg3.c tnapi->tx_buffers[entry].skb = skb; entry 8066 drivers/net/ethernet/broadcom/tg3.c dma_unmap_addr_set(&tnapi->tx_buffers[entry], mapping, mapping); entry 8073 drivers/net/ethernet/broadcom/tg3.c if (tg3_tx_frag_set(tnapi, &entry, &budget, mapping, len, base_flags | entry 8096 drivers/net/ethernet/broadcom/tg3.c tnapi->tx_buffers[entry].skb = NULL; entry 8097 drivers/net/ethernet/broadcom/tg3.c dma_unmap_addr_set(&tnapi->tx_buffers[entry], mapping, entry 8103 drivers/net/ethernet/broadcom/tg3.c tg3_tx_frag_set(tnapi, &entry, &budget, mapping, entry 8131 drivers/net/ethernet/broadcom/tg3.c entry = tnapi->tx_prod; entry 8133 drivers/net/ethernet/broadcom/tg3.c if (tigon3_dma_hwbug_workaround(tnapi, &skb, &entry, &budget, entry 8144 drivers/net/ethernet/broadcom/tg3.c tnapi->tx_prod = entry; entry 8160 drivers/net/ethernet/broadcom/tg3.c tw32_tx_mbox(tnapi->prodmbox, entry); entry 11482 drivers/net/ethernet/broadcom/tg3.c msix_ent[i].entry = i; entry 2654 drivers/net/ethernet/brocade/bna/bnad.c bnad->msix_table[i].entry = i; entry 3126 drivers/net/ethernet/brocade/bna/bnad.c int entry; entry 3140 drivers/net/ethernet/brocade/bna/bnad.c entry = 0; entry 3142 drivers/net/ethernet/brocade/bna/bnad.c ether_addr_copy(&mac_list[entry * ETH_ALEN], &ha->addr[0]); entry 3143 drivers/net/ethernet/brocade/bna/bnad.c entry++; entry 3146 drivers/net/ethernet/brocade/bna/bnad.c ret = bna_rx_ucast_listset(bnad->rx_info[0].rx, entry, mac_list); entry 917 drivers/net/ethernet/cadence/macb_main.c unsigned int entry; entry 925 drivers/net/ethernet/cadence/macb_main.c entry = macb_rx_ring_wrap(bp, queue->rx_prepared_head); entry 931 drivers/net/ethernet/cadence/macb_main.c desc = macb_rx_desc(queue, entry); entry 933 drivers/net/ethernet/cadence/macb_main.c if (!queue->rx_skbuff[entry]) { entry 951 drivers/net/ethernet/cadence/macb_main.c queue->rx_skbuff[entry] = skb; entry 953 drivers/net/ethernet/cadence/macb_main.c if (entry == bp->rx_ring_size - 1) entry 1004 drivers/net/ethernet/cadence/macb_main.c unsigned int entry; entry 1014 drivers/net/ethernet/cadence/macb_main.c entry = macb_rx_ring_wrap(bp, queue->rx_tail); entry 1015 drivers/net/ethernet/cadence/macb_main.c desc = macb_rx_desc(queue, entry); entry 1041 drivers/net/ethernet/cadence/macb_main.c skb = queue->rx_skbuff[entry]; entry 1050 drivers/net/ethernet/cadence/macb_main.c queue->rx_skbuff[entry] = NULL; entry 1053 drivers/net/ethernet/cadence/macb_main.c netdev_vdbg(bp->dev, "gem_rx %u (len %u)\n", entry, len); entry 1499 drivers/net/ethernet/cadence/macb_main.c unsigned int len, entry, i, tx_head = queue->tx_head; entry 1525 drivers/net/ethernet/cadence/macb_main.c entry = macb_tx_ring_wrap(bp, tx_head); entry 1526 drivers/net/ethernet/cadence/macb_main.c tx_skb = &queue->tx_skb[entry]; entry 1556 drivers/net/ethernet/cadence/macb_main.c entry = macb_tx_ring_wrap(bp, tx_head); entry 1557 drivers/net/ethernet/cadence/macb_main.c tx_skb = &queue->tx_skb[entry]; entry 1594 drivers/net/ethernet/cadence/macb_main.c entry = macb_tx_ring_wrap(bp, i); entry 1596 drivers/net/ethernet/cadence/macb_main.c desc = macb_tx_desc(queue, entry); entry 1616 drivers/net/ethernet/cadence/macb_main.c entry = macb_tx_ring_wrap(bp, i); entry 1617 drivers/net/ethernet/cadence/macb_main.c tx_skb = &queue->tx_skb[entry]; entry 1618 drivers/net/ethernet/cadence/macb_main.c desc = macb_tx_desc(queue, entry); entry 1625 drivers/net/ethernet/cadence/macb_main.c if (unlikely(entry == (bp->tx_ring_size - 1))) entry 684 drivers/net/ethernet/calxeda/xgmac.c int entry = priv->rx_head; entry 687 drivers/net/ethernet/calxeda/xgmac.c p = priv->dma_rx + entry; entry 689 drivers/net/ethernet/calxeda/xgmac.c if (priv->rx_skbuff[entry] == NULL) { entry 701 drivers/net/ethernet/calxeda/xgmac.c priv->rx_skbuff[entry] = skb; entry 869 drivers/net/ethernet/calxeda/xgmac.c unsigned int entry = priv->tx_tail; entry 870 drivers/net/ethernet/calxeda/xgmac.c struct sk_buff *skb = priv->tx_skbuff[entry]; entry 871 drivers/net/ethernet/calxeda/xgmac.c struct xgmac_dma_desc *p = priv->dma_tx + entry; entry 893 drivers/net/ethernet/calxeda/xgmac.c priv->tx_skbuff[entry] = NULL; entry 894 drivers/net/ethernet/calxeda/xgmac.c priv->tx_tail = dma_ring_incr(entry, DMA_TX_RING_SZ); entry 1077 drivers/net/ethernet/calxeda/xgmac.c unsigned int entry; entry 1091 drivers/net/ethernet/calxeda/xgmac.c entry = priv->tx_head; entry 1092 drivers/net/ethernet/calxeda/xgmac.c desc = priv->dma_tx + entry; entry 1101 drivers/net/ethernet/calxeda/xgmac.c priv->tx_skbuff[entry] = skb; entry 1114 drivers/net/ethernet/calxeda/xgmac.c entry = dma_ring_incr(entry, DMA_TX_RING_SZ); entry 1115 drivers/net/ethernet/calxeda/xgmac.c desc = priv->dma_tx + entry; entry 1116 drivers/net/ethernet/calxeda/xgmac.c priv->tx_skbuff[entry] = skb; entry 1136 drivers/net/ethernet/calxeda/xgmac.c priv->tx_head = dma_ring_incr(entry, DMA_TX_RING_SZ); entry 1150 drivers/net/ethernet/calxeda/xgmac.c entry = priv->tx_head; entry 1152 drivers/net/ethernet/calxeda/xgmac.c entry = dma_ring_incr(entry, DMA_TX_RING_SZ); entry 1153 drivers/net/ethernet/calxeda/xgmac.c desc = priv->dma_tx + entry; entry 1154 drivers/net/ethernet/calxeda/xgmac.c priv->tx_skbuff[entry] = NULL; entry 1168 drivers/net/ethernet/calxeda/xgmac.c unsigned int entry; entry 1180 drivers/net/ethernet/calxeda/xgmac.c entry = priv->rx_tail; entry 1181 drivers/net/ethernet/calxeda/xgmac.c p = priv->dma_rx + entry; entry 1193 drivers/net/ethernet/calxeda/xgmac.c skb = priv->rx_skbuff[entry]; entry 1198 drivers/net/ethernet/calxeda/xgmac.c priv->rx_skbuff[entry] = NULL; entry 1062 drivers/net/ethernet/cavium/liquidio/lio_core.c msix_entries[i].entry = entry 1065 drivers/net/ethernet/cavium/liquidio/lio_core.c msix_entries[oct->num_msix_irqs - 1].entry = entry 1069 drivers/net/ethernet/cavium/liquidio/lio_core.c msix_entries[i].entry = i; entry 3125 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c entries[i].entry = i; entry 19 drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h struct mbox_cmd entry; entry 2835 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c struct mbox_cmd *entry; entry 2855 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c entry = mbox_cmd_log_entry(log, entry_idx); entry 2857 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c if (entry->timestamp == 0) entry 2860 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c memcpy(&mboxlog->entry, entry, sizeof(struct mbox_cmd)); entry 2862 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c flit = entry->cmd[i]; entry 239 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h struct cudbg_qdesc_entry *entry) entry 241 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qtype = type; entry 242 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qid = txq->cntxt_id; entry 243 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->desc_size = sizeof(struct tx_desc); entry 244 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->num_desc = txq->size; entry 245 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->data_size = txq->size * sizeof(struct tx_desc); entry 246 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h memcpy(entry->data, txq->desc, entry->data_size); entry 251 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h struct cudbg_qdesc_entry *entry) entry 253 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qtype = type; entry 254 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qid = rxq->cntxt_id; entry 255 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->desc_size = rxq->iqe_len; entry 256 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->num_desc = rxq->size; entry 257 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->data_size = rxq->size * rxq->iqe_len; entry 258 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h memcpy(entry->data, rxq->desc, entry->data_size); entry 263 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h struct cudbg_qdesc_entry *entry) entry 265 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qtype = type; entry 266 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qid = flq->cntxt_id; entry 267 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->desc_size = sizeof(__be64); entry 268 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->num_desc = flq->size; entry 269 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->data_size = flq->size * sizeof(__be64); entry 270 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h memcpy(entry->data, flq->desc, entry->data_size); entry 1157 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c struct mbox_cmd *entry; entry 1171 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c entry = mbox_cmd_log_entry(log, entry_idx); entry 1174 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (entry->timestamp == 0) entry 1178 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c entry->seqno, entry->timestamp, entry 1179 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c entry->access, entry->execute); entry 1181 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c u64 flit = entry->cmd[i]; entry 1988 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c u16 *entry = v; entry 1991 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c idx * 8, entry[0], entry[1], entry[2], entry[3], entry[4], entry 1992 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c entry[5], entry[6], entry[7]); entry 351 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct hash_mac_addr *entry; entry 354 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c list_for_each_entry(entry, &adap->mac_hlist, list) { entry 355 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c ucast |= is_unicast_ether_addr(entry->addr); entry 356 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c vec |= (1ULL << hash_mac_addr(entry->addr)); entry 406 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct hash_mac_addr *entry, *tmp; entry 411 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c list_for_each_entry_safe(entry, tmp, &adap->mac_hlist, list) { entry 412 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (ether_addr_equal(entry->addr, mac_addr)) { entry 413 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c list_del(&entry->list); entry 414 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(entry); entry 463 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct hash_mac_addr *entry, *new_entry; entry 473 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c list_for_each_entry(entry, &adapter->mac_hlist, list) { entry 474 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (entry->iface_mac) { entry 475 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c ether_addr_copy(entry->addr, addr); entry 5298 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c entries[i].entry = i; entry 6133 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct hash_mac_addr *entry, *tmp; entry 6185 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c list_for_each_entry_safe(entry, tmp, &adapter->mac_hlist, entry 6187 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c list_del(&entry->list); entry 6188 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c kfree(entry); entry 46 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c const struct cxgb4_match_field *entry, entry 71 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c for (j = 0; entry[j].val; j++) { entry 72 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c if (off == entry[j].off) { entry 74 drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32.c err = entry[j].val(fs, val, mask); entry 242 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c struct mbox_cmd *entry; entry 245 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry = mbox_cmd_log_entry(log, log->cursor++); entry 250 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry->cmd[i] = be64_to_cpu(cmd[i]); entry 252 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry->cmd[i++] = 0; entry 253 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry->timestamp = jiffies; entry 254 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry->seqno = log->seqno++; entry 255 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry->access = access; entry 256 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry->execute = execute; entry 289 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c struct mbox_list entry; entry 323 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c list_add_tail(&entry.list, &adap->mlist.list); entry 338 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c list_del(&entry.list); entry 349 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c list) == &entry) entry 371 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c list_del(&entry.list); entry 424 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c list_del(&entry.list); entry 436 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c list_del(&entry.list); entry 247 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct hash_mac_addr *entry; entry 250 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c list_for_each_entry(entry, &adapter->mac_hlist, list) { entry 251 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c ucast |= is_unicast_ether_addr(entry->addr); entry 252 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c vec |= (1ULL << hash_mac_addr(entry->addr)); entry 277 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct hash_mac_addr *new_entry, *entry; entry 287 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c list_for_each_entry(entry, &adapter->mac_hlist, list) { entry 288 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (entry->iface_mac) { entry 289 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c ether_addr_copy(entry->addr, addr); entry 975 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct hash_mac_addr *entry, *tmp; entry 980 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c list_for_each_entry_safe(entry, tmp, &adapter->mac_hlist, list) { entry 981 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (ether_addr_equal(entry->addr, mac_addr)) { entry 982 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c list_del(&entry->list); entry 983 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c kfree(entry); entry 1964 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct mbox_cmd *entry; entry 1978 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c entry = mbox_cmd_log_entry(log, entry_idx); entry 1981 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (entry->timestamp == 0) entry 1985 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c entry->seqno, entry->timestamp, entry 1986 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c entry->access, entry->execute); entry 1988 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c u64 flit = entry->cmd[i]; entry 2873 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c entries[i].entry = i; entry 3332 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct hash_mac_addr *entry, *tmp; entry 3384 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c list_for_each_entry_safe(entry, tmp, &adapter->mac_hlist, entry 3386 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c list_del(&entry->list); entry 3387 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c kfree(entry); entry 91 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c struct mbox_cmd *entry; entry 94 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry = mbox_cmd_log_entry(log, log->cursor++); entry 99 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry->cmd[i] = be64_to_cpu(cmd[i]); entry 101 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry->cmd[i++] = 0; entry 102 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry->timestamp = jiffies; entry 103 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry->seqno = log->seqno++; entry 104 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry->access = access; entry 105 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry->execute = execute; entry 142 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c struct mbox_list entry; entry 166 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c list_add_tail(&entry.list, &adapter->mlist.list); entry 180 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c list_del(&entry.list); entry 191 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c list) == &entry) entry 214 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c list_del(&entry.list); entry 301 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c list_del(&entry.list); entry 311 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c list_del(&entry.list); entry 233 drivers/net/ethernet/cirrus/ep93xx_eth.c int entry; entry 240 drivers/net/ethernet/cirrus/ep93xx_eth.c entry = ep->rx_pointer; entry 241 drivers/net/ethernet/cirrus/ep93xx_eth.c rstat = ep->descs->rstat + entry; entry 255 drivers/net/ethernet/cirrus/ep93xx_eth.c if ((rstat1 & RSTAT1_BUFFER_INDEX) >> 16 != entry) entry 283 drivers/net/ethernet/cirrus/ep93xx_eth.c struct ep93xx_rdesc *rxd = &ep->descs->rdesc[entry]; entry 287 drivers/net/ethernet/cirrus/ep93xx_eth.c skb_copy_to_linear_data(skb, ep->rx_buf[entry], length); entry 303 drivers/net/ethernet/cirrus/ep93xx_eth.c ep->rx_pointer = (entry + 1) & (RX_QUEUE_ENTRIES - 1); entry 335 drivers/net/ethernet/cirrus/ep93xx_eth.c int entry; entry 343 drivers/net/ethernet/cirrus/ep93xx_eth.c entry = ep->tx_pointer; entry 346 drivers/net/ethernet/cirrus/ep93xx_eth.c txd = &ep->descs->tdesc[entry]; entry 348 drivers/net/ethernet/cirrus/ep93xx_eth.c txd->tdesc1 = TDESC1_EOF | (entry << 16) | (skb->len & 0xfff); entry 351 drivers/net/ethernet/cirrus/ep93xx_eth.c skb_copy_and_csum_dev(skb, ep->tx_buf[entry]); entry 376 drivers/net/ethernet/cirrus/ep93xx_eth.c int entry; entry 380 drivers/net/ethernet/cirrus/ep93xx_eth.c entry = ep->tx_clean_pointer; entry 381 drivers/net/ethernet/cirrus/ep93xx_eth.c tstat = ep->descs->tstat + entry; entry 391 drivers/net/ethernet/cirrus/ep93xx_eth.c if ((tstat0 & TSTAT0_BUFFER_INDEX) != entry) entry 395 drivers/net/ethernet/cirrus/ep93xx_eth.c int length = ep->descs->tdesc[entry].tdesc1 & 0xfff; entry 409 drivers/net/ethernet/cirrus/ep93xx_eth.c ep->tx_clean_pointer = (entry + 1) & (TX_QUEUE_ENTRIES - 1); entry 2396 drivers/net/ethernet/cisco/enic/enic_main.c enic->msix_entry[i].entry = i; entry 1205 drivers/net/ethernet/cisco/enic/vnic_dev.c int vnic_dev_classifier(struct vnic_dev *vdev, u8 cmd, u16 *entry, entry 1239 drivers/net/ethernet/cisco/enic/vnic_dev.c action->u.rq_idx = *entry; entry 1242 drivers/net/ethernet/cisco/enic/vnic_dev.c *entry = (u16)a0; entry 1245 drivers/net/ethernet/cisco/enic/vnic_dev.c a0 = *entry; entry 179 drivers/net/ethernet/cisco/enic/vnic_dev.h int vnic_dev_classifier(struct vnic_dev *vdev, u8 cmd, u16 *entry, entry 607 drivers/net/ethernet/dec/tulip/de2104x.c unsigned int entry, tx_free; entry 621 drivers/net/ethernet/dec/tulip/de2104x.c entry = de->tx_head; entry 623 drivers/net/ethernet/dec/tulip/de2104x.c txd = &de->tx_ring[entry]; entry 627 drivers/net/ethernet/dec/tulip/de2104x.c if (entry == (DE_TX_RING_SIZE - 1)) entry 635 drivers/net/ethernet/dec/tulip/de2104x.c de->tx_skb[entry].skb = skb; entry 636 drivers/net/ethernet/dec/tulip/de2104x.c de->tx_skb[entry].mapping = mapping; entry 642 drivers/net/ethernet/dec/tulip/de2104x.c de->tx_head = NEXT_TX(entry); entry 644 drivers/net/ethernet/dec/tulip/de2104x.c entry, skb->len); entry 722 drivers/net/ethernet/dec/tulip/de2104x.c unsigned int entry; entry 751 drivers/net/ethernet/dec/tulip/de2104x.c entry = de->tx_head; entry 754 drivers/net/ethernet/dec/tulip/de2104x.c if (entry != 0) { entry 755 drivers/net/ethernet/dec/tulip/de2104x.c de->tx_skb[entry].skb = DE_DUMMY_SKB; entry 757 drivers/net/ethernet/dec/tulip/de2104x.c dummy_txd = &de->tx_ring[entry]; entry 758 drivers/net/ethernet/dec/tulip/de2104x.c dummy_txd->opts2 = (entry == (DE_TX_RING_SIZE - 1)) ? entry 764 drivers/net/ethernet/dec/tulip/de2104x.c entry = NEXT_TX(entry); entry 767 drivers/net/ethernet/dec/tulip/de2104x.c de->tx_skb[entry].skb = DE_SETUP_SKB; entry 768 drivers/net/ethernet/dec/tulip/de2104x.c de->tx_skb[entry].mapping = mapping = entry 773 drivers/net/ethernet/dec/tulip/de2104x.c txd = &de->tx_ring[entry]; entry 774 drivers/net/ethernet/dec/tulip/de2104x.c if (entry == (DE_TX_RING_SIZE - 1)) entry 789 drivers/net/ethernet/dec/tulip/de2104x.c de->tx_head = NEXT_TX(entry); entry 1605 drivers/net/ethernet/dec/tulip/de4x5.c int entry; entry 1608 drivers/net/ethernet/dec/tulip/de4x5.c for (entry=lp->rx_new; (s32)le32_to_cpu(lp->rx_ring[entry].status)>=0; entry 1609 drivers/net/ethernet/dec/tulip/de4x5.c entry=lp->rx_new) { entry 1610 drivers/net/ethernet/dec/tulip/de4x5.c status = (s32)le32_to_cpu(lp->rx_ring[entry].status); entry 1620 drivers/net/ethernet/dec/tulip/de4x5.c lp->rx_old = entry; entry 1637 drivers/net/ethernet/dec/tulip/de4x5.c short pkt_len = (short)(le32_to_cpu(lp->rx_ring[entry].status) entry 1640 drivers/net/ethernet/dec/tulip/de4x5.c if ((skb = de4x5_alloc_rx_buff(dev, entry, pkt_len)) == NULL) { entry 1659 drivers/net/ethernet/dec/tulip/de4x5.c for (;lp->rx_old!=entry;lp->rx_old=(lp->rx_old + 1)%lp->rxRingSize) { entry 1663 drivers/net/ethernet/dec/tulip/de4x5.c lp->rx_ring[entry].status = cpu_to_le32(R_OWN); entry 1677 drivers/net/ethernet/dec/tulip/de4x5.c de4x5_free_tx_buff(struct de4x5_private *lp, int entry) entry 1679 drivers/net/ethernet/dec/tulip/de4x5.c dma_unmap_single(lp->gendev, le32_to_cpu(lp->tx_ring[entry].buf), entry 1680 drivers/net/ethernet/dec/tulip/de4x5.c le32_to_cpu(lp->tx_ring[entry].des1) & TD_TBS1, entry 1682 drivers/net/ethernet/dec/tulip/de4x5.c if ((u_long) lp->tx_skb[entry] > 1) entry 1683 drivers/net/ethernet/dec/tulip/de4x5.c dev_kfree_skb_irq(lp->tx_skb[entry]); entry 1684 drivers/net/ethernet/dec/tulip/de4x5.c lp->tx_skb[entry] = NULL; entry 1695 drivers/net/ethernet/dec/tulip/de4x5.c int entry; entry 1698 drivers/net/ethernet/dec/tulip/de4x5.c for (entry = lp->tx_old; entry != lp->tx_new; entry = lp->tx_old) { entry 1699 drivers/net/ethernet/dec/tulip/de4x5.c status = (s32)le32_to_cpu(lp->tx_ring[entry].status); entry 1723 drivers/net/ethernet/dec/tulip/de4x5.c if (lp->tx_skb[entry] != NULL) entry 1724 drivers/net/ethernet/dec/tulip/de4x5.c de4x5_free_tx_buff(lp, entry); entry 1899 drivers/net/ethernet/dec/tulip/de4x5.c int entry = (lp->tx_new ? lp->tx_new-1 : lp->txRingSize-1); entry 1906 drivers/net/ethernet/dec/tulip/de4x5.c lp->tx_ring[entry].des1 &= cpu_to_le32(~TD_IC); entry 62 drivers/net/ethernet/dec/tulip/interrupt.c int entry; entry 67 drivers/net/ethernet/dec/tulip/interrupt.c entry = tp->dirty_rx % RX_RING_SIZE; entry 68 drivers/net/ethernet/dec/tulip/interrupt.c if (tp->rx_buffers[entry].skb == NULL) { entry 72 drivers/net/ethernet/dec/tulip/interrupt.c skb = tp->rx_buffers[entry].skb = entry 81 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].skb = NULL; entry 85 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].mapping = mapping; entry 87 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_ring[entry].buffer1 = cpu_to_le32(mapping); entry 90 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_ring[entry].status = cpu_to_le32(DescOwned); entry 116 drivers/net/ethernet/dec/tulip/interrupt.c int entry = tp->cur_rx % RX_RING_SIZE; entry 132 drivers/net/ethernet/dec/tulip/interrupt.c entry, tp->rx_ring[entry].status); entry 144 drivers/net/ethernet/dec/tulip/interrupt.c while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) { entry 145 drivers/net/ethernet/dec/tulip/interrupt.c s32 status = le32_to_cpu(tp->rx_ring[entry].status); entry 153 drivers/net/ethernet/dec/tulip/interrupt.c entry, status); entry 214 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].mapping, entry 217 drivers/net/ethernet/dec/tulip/interrupt.c skb_copy_to_linear_data(skb, tp->rx_buffers[entry].skb->data, entry 222 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].skb->data, entry 226 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].mapping, entry 229 drivers/net/ethernet/dec/tulip/interrupt.c char *temp = skb_put(skb = tp->rx_buffers[entry].skb, entry 233 drivers/net/ethernet/dec/tulip/interrupt.c if (tp->rx_buffers[entry].mapping != entry 234 drivers/net/ethernet/dec/tulip/interrupt.c le32_to_cpu(tp->rx_ring[entry].buffer1)) { entry 237 drivers/net/ethernet/dec/tulip/interrupt.c le32_to_cpu(tp->rx_ring[entry].buffer1), entry 238 drivers/net/ethernet/dec/tulip/interrupt.c (unsigned long long)tp->rx_buffers[entry].mapping, entry 243 drivers/net/ethernet/dec/tulip/interrupt.c pci_unmap_single(tp->pdev, tp->rx_buffers[entry].mapping, entry 246 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].skb = NULL; entry 247 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].mapping = 0; entry 260 drivers/net/ethernet/dec/tulip/interrupt.c entry = (++tp->cur_rx) % RX_RING_SIZE; entry 368 drivers/net/ethernet/dec/tulip/interrupt.c int entry = tp->cur_rx % RX_RING_SIZE; entry 374 drivers/net/ethernet/dec/tulip/interrupt.c entry, tp->rx_ring[entry].status); entry 376 drivers/net/ethernet/dec/tulip/interrupt.c while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) { entry 377 drivers/net/ethernet/dec/tulip/interrupt.c s32 status = le32_to_cpu(tp->rx_ring[entry].status); entry 382 drivers/net/ethernet/dec/tulip/interrupt.c entry, status); entry 440 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].mapping, entry 443 drivers/net/ethernet/dec/tulip/interrupt.c skb_copy_to_linear_data(skb, tp->rx_buffers[entry].skb->data, entry 448 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].skb->data, entry 452 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].mapping, entry 455 drivers/net/ethernet/dec/tulip/interrupt.c char *temp = skb_put(skb = tp->rx_buffers[entry].skb, entry 459 drivers/net/ethernet/dec/tulip/interrupt.c if (tp->rx_buffers[entry].mapping != entry 460 drivers/net/ethernet/dec/tulip/interrupt.c le32_to_cpu(tp->rx_ring[entry].buffer1)) { entry 463 drivers/net/ethernet/dec/tulip/interrupt.c le32_to_cpu(tp->rx_ring[entry].buffer1), entry 464 drivers/net/ethernet/dec/tulip/interrupt.c (long long)tp->rx_buffers[entry].mapping, entry 469 drivers/net/ethernet/dec/tulip/interrupt.c pci_unmap_single(tp->pdev, tp->rx_buffers[entry].mapping, entry 472 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].skb = NULL; entry 473 drivers/net/ethernet/dec/tulip/interrupt.c tp->rx_buffers[entry].mapping = 0; entry 483 drivers/net/ethernet/dec/tulip/interrupt.c entry = (++tp->cur_rx) % RX_RING_SIZE; entry 531 drivers/net/ethernet/dec/tulip/interrupt.c int entry; entry 590 drivers/net/ethernet/dec/tulip/interrupt.c int entry = dirty_tx % TX_RING_SIZE; entry 591 drivers/net/ethernet/dec/tulip/interrupt.c int status = le32_to_cpu(tp->tx_ring[entry].status); entry 597 drivers/net/ethernet/dec/tulip/interrupt.c if (tp->tx_buffers[entry].skb == NULL) { entry 599 drivers/net/ethernet/dec/tulip/interrupt.c if (tp->tx_buffers[entry].mapping) entry 601 drivers/net/ethernet/dec/tulip/interrupt.c tp->tx_buffers[entry].mapping, entry 627 drivers/net/ethernet/dec/tulip/interrupt.c tp->tx_buffers[entry].skb->len; entry 632 drivers/net/ethernet/dec/tulip/interrupt.c pci_unmap_single(tp->pdev, tp->tx_buffers[entry].mapping, entry 633 drivers/net/ethernet/dec/tulip/interrupt.c tp->tx_buffers[entry].skb->len, entry 637 drivers/net/ethernet/dec/tulip/interrupt.c dev_kfree_skb_irq(tp->tx_buffers[entry].skb); entry 638 drivers/net/ethernet/dec/tulip/interrupt.c tp->tx_buffers[entry].skb = NULL; entry 639 drivers/net/ethernet/dec/tulip/interrupt.c tp->tx_buffers[entry].mapping = 0; entry 780 drivers/net/ethernet/dec/tulip/interrupt.c entry = tp->dirty_rx % RX_RING_SIZE; entry 781 drivers/net/ethernet/dec/tulip/interrupt.c if (tp->rx_buffers[entry].skb == NULL) { entry 667 drivers/net/ethernet/dec/tulip/tulip_core.c int entry; entry 675 drivers/net/ethernet/dec/tulip/tulip_core.c entry = tp->cur_tx % TX_RING_SIZE; entry 677 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].skb = skb; entry 680 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].mapping = mapping; entry 681 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].buffer1 = cpu_to_le32(mapping); entry 693 drivers/net/ethernet/dec/tulip/tulip_core.c if (entry == TX_RING_SIZE-1) entry 696 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].length = cpu_to_le32(skb->len | flag); entry 699 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].status = cpu_to_le32(DescOwned); entry 718 drivers/net/ethernet/dec/tulip/tulip_core.c int entry = dirty_tx % TX_RING_SIZE; entry 719 drivers/net/ethernet/dec/tulip/tulip_core.c int status = le32_to_cpu(tp->tx_ring[entry].status); entry 723 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].status = 0; entry 727 drivers/net/ethernet/dec/tulip/tulip_core.c if (tp->tx_buffers[entry].skb == NULL) { entry 729 drivers/net/ethernet/dec/tulip/tulip_core.c if (tp->tx_buffers[entry].mapping) entry 731 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].mapping, entry 737 drivers/net/ethernet/dec/tulip/tulip_core.c pci_unmap_single(tp->pdev, tp->tx_buffers[entry].mapping, entry 738 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].skb->len, entry 742 drivers/net/ethernet/dec/tulip/tulip_core.c dev_kfree_skb_irq(tp->tx_buffers[entry].skb); entry 743 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].skb = NULL; entry 744 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].mapping = 0; entry 1142 drivers/net/ethernet/dec/tulip/tulip_core.c unsigned int entry; entry 1147 drivers/net/ethernet/dec/tulip/tulip_core.c entry = tp->cur_tx++ % TX_RING_SIZE; entry 1149 drivers/net/ethernet/dec/tulip/tulip_core.c if (entry != 0) { entry 1151 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].skb = NULL; entry 1152 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].mapping = 0; entry 1153 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].length = entry 1154 drivers/net/ethernet/dec/tulip/tulip_core.c (entry == TX_RING_SIZE-1) ? cpu_to_le32(DESC_RING_WRAP) : 0; entry 1155 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].buffer1 = 0; entry 1157 drivers/net/ethernet/dec/tulip/tulip_core.c dummy = entry; entry 1158 drivers/net/ethernet/dec/tulip/tulip_core.c entry = tp->cur_tx++ % TX_RING_SIZE; entry 1162 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].skb = NULL; entry 1163 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_buffers[entry].mapping = entry 1168 drivers/net/ethernet/dec/tulip/tulip_core.c if (entry == TX_RING_SIZE-1) entry 1170 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].length = cpu_to_le32(tx_flags); entry 1171 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].buffer1 = entry 1172 drivers/net/ethernet/dec/tulip/tulip_core.c cpu_to_le32(tp->tx_buffers[entry].mapping); entry 1173 drivers/net/ethernet/dec/tulip/tulip_core.c tp->tx_ring[entry].status = cpu_to_le32(DescOwned); entry 999 drivers/net/ethernet/dec/tulip/winbond-840.c unsigned entry; entry 1005 drivers/net/ethernet/dec/tulip/winbond-840.c entry = np->cur_tx % TX_RING_SIZE; entry 1007 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_addr[entry] = pci_map_single(np->pci_dev, entry 1009 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_skbuff[entry] = skb; entry 1011 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_ring[entry].buffer1 = np->tx_addr[entry]; entry 1013 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_ring[entry].length = DescWholePkt | skb->len; entry 1017 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_ring[entry].buffer2 = np->tx_addr[entry]+TX_BUFLIMIT; entry 1018 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_ring[entry].length = DescWholePkt | (len << 11) | TX_BUFLIMIT; entry 1020 drivers/net/ethernet/dec/tulip/winbond-840.c if(entry == TX_RING_SIZE-1) entry 1021 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_ring[entry].length |= DescEndRing; entry 1037 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_ring[entry].status = DescOwned; entry 1053 drivers/net/ethernet/dec/tulip/winbond-840.c np->cur_tx, entry); entry 1062 drivers/net/ethernet/dec/tulip/winbond-840.c int entry = np->dirty_tx % TX_RING_SIZE; entry 1063 drivers/net/ethernet/dec/tulip/winbond-840.c int tx_status = np->tx_ring[entry].status; entry 1084 drivers/net/ethernet/dec/tulip/winbond-840.c entry, tx_status); entry 1086 drivers/net/ethernet/dec/tulip/winbond-840.c np->stats.tx_bytes += np->tx_skbuff[entry]->len; entry 1091 drivers/net/ethernet/dec/tulip/winbond-840.c pci_unmap_single(np->pci_dev,np->tx_addr[entry], entry 1092 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_skbuff[entry]->len, entry 1094 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_q_bytes -= np->tx_skbuff[entry]->len; entry 1095 drivers/net/ethernet/dec/tulip/winbond-840.c dev_kfree_skb_irq(np->tx_skbuff[entry]); entry 1096 drivers/net/ethernet/dec/tulip/winbond-840.c np->tx_skbuff[entry] = NULL; entry 1178 drivers/net/ethernet/dec/tulip/winbond-840.c int entry = np->cur_rx % RX_RING_SIZE; entry 1183 drivers/net/ethernet/dec/tulip/winbond-840.c entry, np->rx_ring[entry].status); entry 1230 drivers/net/ethernet/dec/tulip/winbond-840.c pci_dma_sync_single_for_cpu(np->pci_dev,np->rx_addr[entry], entry 1231 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_skbuff[entry]->len, entry 1233 drivers/net/ethernet/dec/tulip/winbond-840.c skb_copy_to_linear_data(skb, np->rx_skbuff[entry]->data, pkt_len); entry 1235 drivers/net/ethernet/dec/tulip/winbond-840.c pci_dma_sync_single_for_device(np->pci_dev,np->rx_addr[entry], entry 1236 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_skbuff[entry]->len, entry 1239 drivers/net/ethernet/dec/tulip/winbond-840.c pci_unmap_single(np->pci_dev,np->rx_addr[entry], entry 1240 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_skbuff[entry]->len, entry 1242 drivers/net/ethernet/dec/tulip/winbond-840.c skb_put(skb = np->rx_skbuff[entry], pkt_len); entry 1243 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_skbuff[entry] = NULL; entry 1258 drivers/net/ethernet/dec/tulip/winbond-840.c entry = (++np->cur_rx) % RX_RING_SIZE; entry 1259 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_head_desc = &np->rx_ring[entry]; entry 1265 drivers/net/ethernet/dec/tulip/winbond-840.c entry = np->dirty_rx % RX_RING_SIZE; entry 1266 drivers/net/ethernet/dec/tulip/winbond-840.c if (np->rx_skbuff[entry] == NULL) { entry 1268 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_skbuff[entry] = skb; entry 1271 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_addr[entry] = pci_map_single(np->pci_dev, entry 1274 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_ring[entry].buffer1 = np->rx_addr[entry]; entry 1277 drivers/net/ethernet/dec/tulip/winbond-840.c np->rx_ring[entry].status = DescOwned; entry 659 drivers/net/ethernet/dlink/dl2k.c unsigned int entry; entry 670 drivers/net/ethernet/dlink/dl2k.c entry = np->old_rx % RX_RING_SIZE; entry 672 drivers/net/ethernet/dlink/dl2k.c if (np->rx_skbuff[entry] == NULL) { entry 676 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[entry].fraginfo = 0; entry 679 drivers/net/ethernet/dlink/dl2k.c dev->name, entry); entry 682 drivers/net/ethernet/dlink/dl2k.c np->rx_skbuff[entry] = skb; entry 683 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[entry].fraginfo = entry 688 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[entry].fraginfo |= entry 690 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[entry].status = 0; entry 717 drivers/net/ethernet/dlink/dl2k.c unsigned entry; entry 724 drivers/net/ethernet/dlink/dl2k.c entry = np->cur_tx % TX_RING_SIZE; entry 725 drivers/net/ethernet/dlink/dl2k.c np->tx_skbuff[entry] = skb; entry 726 drivers/net/ethernet/dlink/dl2k.c txdesc = &np->tx_ring[entry]; entry 747 drivers/net/ethernet/dlink/dl2k.c if (entry % np->tx_coalesce == 0 || np->speed == 10) entry 748 drivers/net/ethernet/dlink/dl2k.c txdesc->status = cpu_to_le64 (entry | tfc_vlan_tag | entry 753 drivers/net/ethernet/dlink/dl2k.c txdesc->status = cpu_to_le64 (entry | tfc_vlan_tag | entry 772 drivers/net/ethernet/dlink/dl2k.c entry * sizeof (struct netdev_desc)); entry 823 drivers/net/ethernet/dlink/dl2k.c int entry = np->old_tx % TX_RING_SIZE; entry 833 drivers/net/ethernet/dlink/dl2k.c while (entry != np->cur_tx) { entry 836 drivers/net/ethernet/dlink/dl2k.c if (!(np->tx_ring[entry].status & cpu_to_le64(TFDDone))) entry 838 drivers/net/ethernet/dlink/dl2k.c skb = np->tx_skbuff[entry]; entry 840 drivers/net/ethernet/dlink/dl2k.c desc_to_dma(&np->tx_ring[entry]), entry 847 drivers/net/ethernet/dlink/dl2k.c np->tx_skbuff[entry] = NULL; entry 848 drivers/net/ethernet/dlink/dl2k.c entry = (entry + 1) % TX_RING_SIZE; entry 855 drivers/net/ethernet/dlink/dl2k.c np->old_tx = entry; entry 926 drivers/net/ethernet/dlink/dl2k.c int entry = np->cur_rx % RX_RING_SIZE; entry 931 drivers/net/ethernet/dlink/dl2k.c struct netdev_desc *desc = &np->rx_ring[entry]; entry 965 drivers/net/ethernet/dlink/dl2k.c skb_put (skb = np->rx_skbuff[entry], pkt_len); entry 966 drivers/net/ethernet/dlink/dl2k.c np->rx_skbuff[entry] = NULL; entry 973 drivers/net/ethernet/dlink/dl2k.c np->rx_skbuff[entry]->data, entry 991 drivers/net/ethernet/dlink/dl2k.c entry = (entry + 1) % RX_RING_SIZE; entry 994 drivers/net/ethernet/dlink/dl2k.c np->cur_rx = entry; entry 996 drivers/net/ethernet/dlink/dl2k.c entry = np->old_rx; entry 997 drivers/net/ethernet/dlink/dl2k.c while (entry != np->cur_rx) { entry 1000 drivers/net/ethernet/dlink/dl2k.c if (np->rx_skbuff[entry] == NULL) { entry 1003 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[entry].fraginfo = 0; entry 1007 drivers/net/ethernet/dlink/dl2k.c dev->name, entry); entry 1010 drivers/net/ethernet/dlink/dl2k.c np->rx_skbuff[entry] = skb; entry 1011 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[entry].fraginfo = entry 1016 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[entry].fraginfo |= entry 1018 drivers/net/ethernet/dlink/dl2k.c np->rx_ring[entry].status = 0; entry 1019 drivers/net/ethernet/dlink/dl2k.c entry = (entry + 1) % RX_RING_SIZE; entry 1021 drivers/net/ethernet/dlink/dl2k.c np->old_rx = entry; entry 1082 drivers/net/ethernet/dlink/sundance.c int entry = np->cur_task % TX_RING_SIZE; entry 1083 drivers/net/ethernet/dlink/sundance.c txdesc = &np->tx_ring[entry]; entry 1086 drivers/net/ethernet/dlink/sundance.c entry*sizeof(struct netdev_desc)); entry 1103 drivers/net/ethernet/dlink/sundance.c unsigned entry; entry 1106 drivers/net/ethernet/dlink/sundance.c entry = np->cur_tx % TX_RING_SIZE; entry 1107 drivers/net/ethernet/dlink/sundance.c np->tx_skbuff[entry] = skb; entry 1108 drivers/net/ethernet/dlink/sundance.c txdesc = &np->tx_ring[entry]; entry 1111 drivers/net/ethernet/dlink/sundance.c txdesc->status = cpu_to_le32 ((entry << 2) | DisableAlign); entry 1135 drivers/net/ethernet/dlink/sundance.c dev->name, np->cur_tx, entry); entry 1141 drivers/net/ethernet/dlink/sundance.c np->tx_skbuff[entry] = NULL; entry 1271 drivers/net/ethernet/dlink/sundance.c int entry = np->dirty_tx % TX_RING_SIZE; entry 1275 drivers/net/ethernet/dlink/sundance.c np->tx_ring[entry].status) >> 2) & 0xff; entry 1277 drivers/net/ethernet/dlink/sundance.c !(le32_to_cpu(np->tx_ring[entry].status) entry 1283 drivers/net/ethernet/dlink/sundance.c skb = np->tx_skbuff[entry]; entry 1286 drivers/net/ethernet/dlink/sundance.c le32_to_cpu(np->tx_ring[entry].frag[0].addr), entry 1288 drivers/net/ethernet/dlink/sundance.c dev_consume_skb_irq(np->tx_skbuff[entry]); entry 1289 drivers/net/ethernet/dlink/sundance.c np->tx_skbuff[entry] = NULL; entry 1290 drivers/net/ethernet/dlink/sundance.c np->tx_ring[entry].frag[0].addr = 0; entry 1291 drivers/net/ethernet/dlink/sundance.c np->tx_ring[entry].frag[0].length = 0; entry 1297 drivers/net/ethernet/dlink/sundance.c int entry = np->dirty_tx % TX_RING_SIZE; entry 1299 drivers/net/ethernet/dlink/sundance.c if (!(le32_to_cpu(np->tx_ring[entry].status) entry 1302 drivers/net/ethernet/dlink/sundance.c skb = np->tx_skbuff[entry]; entry 1305 drivers/net/ethernet/dlink/sundance.c le32_to_cpu(np->tx_ring[entry].frag[0].addr), entry 1307 drivers/net/ethernet/dlink/sundance.c dev_consume_skb_irq(np->tx_skbuff[entry]); entry 1308 drivers/net/ethernet/dlink/sundance.c np->tx_skbuff[entry] = NULL; entry 1309 drivers/net/ethernet/dlink/sundance.c np->tx_ring[entry].frag[0].addr = 0; entry 1310 drivers/net/ethernet/dlink/sundance.c np->tx_ring[entry].frag[0].length = 0; entry 1334 drivers/net/ethernet/dlink/sundance.c int entry = np->cur_rx % RX_RING_SIZE; entry 1341 drivers/net/ethernet/dlink/sundance.c struct netdev_desc *desc = &(np->rx_ring[entry]); entry 1389 drivers/net/ethernet/dlink/sundance.c skb_copy_to_linear_data(skb, np->rx_skbuff[entry]->data, pkt_len); entry 1398 drivers/net/ethernet/dlink/sundance.c skb_put(skb = np->rx_skbuff[entry], pkt_len); entry 1399 drivers/net/ethernet/dlink/sundance.c np->rx_skbuff[entry] = NULL; entry 1405 drivers/net/ethernet/dlink/sundance.c entry = (entry + 1) % RX_RING_SIZE; entry 1408 drivers/net/ethernet/dlink/sundance.c np->cur_rx = entry; entry 1415 drivers/net/ethernet/dlink/sundance.c np->cur_rx = entry; entry 1428 drivers/net/ethernet/dlink/sundance.c int entry; entry 1435 drivers/net/ethernet/dlink/sundance.c entry = np->dirty_rx % RX_RING_SIZE; entry 1436 drivers/net/ethernet/dlink/sundance.c if (np->rx_skbuff[entry] == NULL) { entry 1438 drivers/net/ethernet/dlink/sundance.c np->rx_skbuff[entry] = skb; entry 1442 drivers/net/ethernet/dlink/sundance.c np->rx_ring[entry].frag[0].addr = cpu_to_le32( entry 1446 drivers/net/ethernet/dlink/sundance.c np->rx_ring[entry].frag[0].addr)) { entry 1448 drivers/net/ethernet/dlink/sundance.c np->rx_skbuff[entry] = NULL; entry 1453 drivers/net/ethernet/dlink/sundance.c np->rx_ring[entry].frag[0].length = entry 1455 drivers/net/ethernet/dlink/sundance.c np->rx_ring[entry].status = 0; entry 641 drivers/net/ethernet/emulex/benet/be.h struct list_head entry; entry 1428 drivers/net/ethernet/emulex/benet/be_main.c u32 *entry; entry 1437 drivers/net/ethernet/emulex/benet/be_main.c entry = txo->q.dma_mem.va; entry 1439 drivers/net/ethernet/emulex/benet/be_main.c if (entry[j] != 0 || entry[j + 1] != 0 || entry 1440 drivers/net/ethernet/emulex/benet/be_main.c entry[j + 2] != 0 || entry[j + 3] != 0) { entry 1442 drivers/net/ethernet/emulex/benet/be_main.c j, entry[j], entry[j + 1], entry 1443 drivers/net/ethernet/emulex/benet/be_main.c entry[j + 2], entry[j + 3]); entry 1447 drivers/net/ethernet/emulex/benet/be_main.c entry = txo->cq.dma_mem.va; entry 1452 drivers/net/ethernet/emulex/benet/be_main.c if (entry[j] != 0 || entry[j + 1] != 0 || entry 1453 drivers/net/ethernet/emulex/benet/be_main.c entry[j + 2] != 0 || entry[j + 3] != 0) { entry 1455 drivers/net/ethernet/emulex/benet/be_main.c j, entry[j], entry[j + 1], entry 1456 drivers/net/ethernet/emulex/benet/be_main.c entry[j + 2], entry[j + 3]); entry 3448 drivers/net/ethernet/emulex/benet/be_main.c adapter->msix_entries[i].entry = i; entry 85 drivers/net/ethernet/emulex/benet/be_roce.c INIT_LIST_HEAD(&adapter->entry); entry 87 drivers/net/ethernet/emulex/benet/be_roce.c list_add_tail(&adapter->entry, &be_adapter_list); entry 110 drivers/net/ethernet/emulex/benet/be_roce.c list_del(&adapter->entry); entry 137 drivers/net/ethernet/emulex/benet/be_roce.c list_for_each_entry(dev, &be_adapter_list, entry) { entry 150 drivers/net/ethernet/emulex/benet/be_roce.c list_for_each_entry(dev, &be_adapter_list, entry) { entry 422 drivers/net/ethernet/ethoc.c unsigned int entry; entry 425 drivers/net/ethernet/ethoc.c entry = priv->num_tx + priv->cur_rx; entry 426 drivers/net/ethernet/ethoc.c ethoc_read_bd(priv, entry, &bd); entry 436 drivers/net/ethernet/ethoc.c ethoc_read_bd(priv, entry, &bd); entry 449 drivers/net/ethernet/ethoc.c void *src = priv->vma[entry]; entry 468 drivers/net/ethernet/ethoc.c ethoc_write_bd(priv, entry, &bd); entry 515 drivers/net/ethernet/ethoc.c unsigned int entry; entry 517 drivers/net/ethernet/ethoc.c entry = priv->dty_tx & (priv->num_tx-1); entry 519 drivers/net/ethernet/ethoc.c ethoc_read_bd(priv, entry, &bd); entry 530 drivers/net/ethernet/ethoc.c ethoc_read_bd(priv, entry, &bd); entry 884 drivers/net/ethernet/ethoc.c unsigned int entry; entry 897 drivers/net/ethernet/ethoc.c entry = priv->cur_tx % priv->num_tx; entry 901 drivers/net/ethernet/ethoc.c ethoc_read_bd(priv, entry, &bd); entry 907 drivers/net/ethernet/ethoc.c dest = priv->vma[entry]; entry 912 drivers/net/ethernet/ethoc.c ethoc_write_bd(priv, entry, &bd); entry 915 drivers/net/ethernet/ethoc.c ethoc_write_bd(priv, entry, &bd); entry 382 drivers/net/ethernet/faraday/ftgmac100.c static int ftgmac100_alloc_rx_buf(struct ftgmac100 *priv, unsigned int entry, entry 410 drivers/net/ethernet/faraday/ftgmac100.c priv->rx_skbs[entry] = skb; entry 419 drivers/net/ethernet/faraday/ftgmac100.c if (entry == (priv->rx_q_entries - 1)) entry 1236 drivers/net/ethernet/freescale/enetc/enetc.c int entry = ENETC_BDR_INT_BASE_IDX + i; entry 1250 drivers/net/ethernet/freescale/enetc/enetc.c enetc_wr(hw, ENETC_SIMSIRRV(i), entry); entry 1255 drivers/net/ethernet/freescale/enetc/enetc.c enetc_wr(hw, ENETC_SIMSITRV(idx), entry); entry 946 drivers/net/ethernet/freescale/fman/fman_memac.c u32 entry; entry 953 drivers/net/ethernet/freescale/fman/fman_memac.c for (entry = 0; entry < HASH_TABLE_SIZE; entry++) entry 954 drivers/net/ethernet/freescale/fman/fman_memac.c iowrite32be(entry | HASH_CTRL_MCAST_EN, entry 957 drivers/net/ethernet/freescale/fman/fman_memac.c for (entry = 0; entry < HASH_TABLE_SIZE; entry++) entry 958 drivers/net/ethernet/freescale/fman/fman_memac.c iowrite32be(entry & ~HASH_CTRL_MCAST_EN, entry 571 drivers/net/ethernet/freescale/fman/fman_tgec.c u32 entry; entry 578 drivers/net/ethernet/freescale/fman/fman_tgec.c for (entry = 0; entry < TGEC_HASH_TABLE_SIZE; entry++) entry 579 drivers/net/ethernet/freescale/fman/fman_tgec.c iowrite32be(entry | TGEC_HASH_MCAST_EN, entry 582 drivers/net/ethernet/freescale/fman/fman_tgec.c for (entry = 0; entry < TGEC_HASH_TABLE_SIZE; entry++) entry 583 drivers/net/ethernet/freescale/fman/fman_tgec.c iowrite32be(entry & ~TGEC_HASH_MCAST_EN, entry 149 drivers/net/ethernet/google/gve/gve_main.c priv->msix_vectors[i].entry = i; entry 191 drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c hwdev->msix_entries[i].entry = i; entry 391 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_msix_attr_cnt_clear(aeq->hwif, aeq->msix_entry.entry); entry 412 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_msix_attr_cnt_clear(ceq->hwif, ceq->msix_entry.entry); entry 436 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ctrl0 = HINIC_AEQ_CTRL_0_SET(msix_entry->entry, INT_IDX) | entry 457 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ctrl0 = HINIC_CEQ_CTRL_0_SET(msix_entry->entry, INTR_IDX) | entry 664 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c struct msix_entry entry) entry 694 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c eq->msix_entry = entry; entry 725 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_msix_attr_set(eq->hwif, eq->msix_entry.entry, entry 733 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c err = request_irq(entry.vector, aeq_interrupt, 0, entry 736 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c err = request_irq(entry.vector, ceq_interrupt, 0, entry 757 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c hinic_set_msix_state(eq->hwif, eq->msix_entry.entry, entry 277 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c struct hinic_wq *wq, struct msix_entry *entry, entry 285 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sq->irq = entry->vector; entry 286 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c sq->msix_entry = entry->entry; entry 380 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c struct hinic_wq *wq, struct msix_entry *entry) entry 390 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c rq->irq = entry->vector; entry 391 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c rq->msix_entry = entry->entry; entry 130 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h struct hinic_wq *wq, struct msix_entry *entry, void *ci_addr, entry 136 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h struct hinic_wq *wq, struct msix_entry *entry); entry 609 drivers/net/ethernet/huawei/hinic/hinic_port.c indir_tbl->entry[i] = indir_table[i]; entry 612 drivers/net/ethernet/huawei/hinic/hinic_port.c temp = (u32 *)&indir_tbl->entry[i - 3]; entry 633 drivers/net/ethernet/huawei/hinic/hinic_port.c memcpy(&indir_tbl->entry[0], &indir_tbl->entry[indir_size], indir_size); entry 294 drivers/net/ethernet/huawei/hinic/hinic_port.h u8 entry[HINIC_RSS_INDIR_SIZE]; entry 930 drivers/net/ethernet/ibm/ehea/ehea_main.c qp_token = EHEA_BMASK_GET(EHEA_EQE_QP_TOKEN, eqe->entry); entry 932 drivers/net/ethernet/ibm/ehea/ehea_main.c eqe->entry, qp_token); entry 1225 drivers/net/ethernet/ibm/ehea/ehea_main.c pr_debug("*eqe=%lx\n", (unsigned long) eqe->entry); entry 1226 drivers/net/ethernet/ibm/ehea/ehea_main.c ehea_parse_eqe(adapter, eqe->entry); entry 190 drivers/net/ethernet/ibm/ehea/ehea_qmr.h u64 entry; entry 2136 drivers/net/ethernet/ibm/ibmvnic.c struct list_head *entry, *tmp_entry; entry 2158 drivers/net/ethernet/ibm/ibmvnic.c list_for_each(entry, &adapter->rwi_list) { entry 2159 drivers/net/ethernet/ibm/ibmvnic.c tmp = list_entry(entry, struct ibmvnic_rwi, list); entry 2179 drivers/net/ethernet/ibm/ibmvnic.c list_for_each_safe(entry, tmp_entry, &adapter->rwi_list) entry 2180 drivers/net/ethernet/ibm/ibmvnic.c list_del(entry); entry 3292 drivers/net/ethernet/ibm/ibmvnic.c union sub_crq *entry = &scrq->msgs[scrq->cur]; entry 3294 drivers/net/ethernet/ibm/ibmvnic.c if (entry->generic.first & IBMVNIC_CRQ_CMD_RSP) entry 3303 drivers/net/ethernet/ibm/ibmvnic.c union sub_crq *entry; entry 3307 drivers/net/ethernet/ibm/ibmvnic.c entry = &scrq->msgs[scrq->cur]; entry 3308 drivers/net/ethernet/ibm/ibmvnic.c if (entry->generic.first & IBMVNIC_CRQ_CMD_RSP) { entry 3312 drivers/net/ethernet/ibm/ibmvnic.c entry = NULL; entry 3316 drivers/net/ethernet/ibm/ibmvnic.c return entry; entry 2063 drivers/net/ethernet/intel/e1000e/netdev.c adapter->msix_entries[i].entry = i; entry 1847 drivers/net/ethernet/intel/fm10k/fm10k_main.c interface->msix_entries[vector].entry = vector; entry 1424 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct msix_entry *entry; entry 1431 drivers/net/ethernet/intel/fm10k/fm10k_pci.c entry = &interface->msix_entries[FM10K_MBX_VECTOR]; entry 1454 drivers/net/ethernet/intel/fm10k/fm10k_pci.c free_irq(entry->vector, interface); entry 1509 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct msix_entry *entry = &interface->msix_entries[FM10K_MBX_VECTOR]; entry 1515 drivers/net/ethernet/intel/fm10k/fm10k_pci.c u32 itr = entry->entry | FM10K_INT_MAP_TIMER0; entry 1523 drivers/net/ethernet/intel/fm10k/fm10k_pci.c err = request_irq(entry->vector, fm10k_msix_mbx_vf, 0, entry 1535 drivers/net/ethernet/intel/fm10k/fm10k_pci.c fm10k_write_reg(hw, FM10K_VFITR(entry->entry), FM10K_ITR_ENABLE); entry 1642 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct msix_entry *entry = &interface->msix_entries[FM10K_MBX_VECTOR]; entry 1648 drivers/net/ethernet/intel/fm10k/fm10k_pci.c u32 mbx_itr = entry->entry | FM10K_INT_MAP_TIMER0; entry 1649 drivers/net/ethernet/intel/fm10k/fm10k_pci.c u32 other_itr = entry->entry | FM10K_INT_MAP_IMMEDIATE; entry 1657 drivers/net/ethernet/intel/fm10k/fm10k_pci.c err = request_irq(entry->vector, fm10k_msix_mbx_pf, 0, entry 1686 drivers/net/ethernet/intel/fm10k/fm10k_pci.c fm10k_write_reg(hw, FM10K_ITR(entry->entry), FM10K_ITR_ENABLE); entry 1723 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct msix_entry *entry; entry 1725 drivers/net/ethernet/intel/fm10k/fm10k_pci.c entry = &interface->msix_entries[NON_Q_VECTORS + vector]; entry 1731 drivers/net/ethernet/intel/fm10k/fm10k_pci.c entry--; entry 1738 drivers/net/ethernet/intel/fm10k/fm10k_pci.c irq_set_affinity_hint(entry->vector, NULL); entry 1743 drivers/net/ethernet/intel/fm10k/fm10k_pci.c free_irq(entry->vector, q_vector); entry 1758 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct msix_entry *entry; entry 1762 drivers/net/ethernet/intel/fm10k/fm10k_pci.c entry = &interface->msix_entries[NON_Q_VECTORS]; entry 1785 drivers/net/ethernet/intel/fm10k/fm10k_pci.c &interface->uc_addr[FM10K_ITR(entry->entry)] : entry 1786 drivers/net/ethernet/intel/fm10k/fm10k_pci.c &interface->uc_addr[FM10K_VFITR(entry->entry)]; entry 1789 drivers/net/ethernet/intel/fm10k/fm10k_pci.c err = request_irq(entry->vector, &fm10k_msix_clean_rings, 0, entry 1799 drivers/net/ethernet/intel/fm10k/fm10k_pci.c irq_set_affinity_hint(entry->vector, &q_vector->affinity_mask); entry 1804 drivers/net/ethernet/intel/fm10k/fm10k_pci.c entry++; entry 1814 drivers/net/ethernet/intel/fm10k/fm10k_pci.c entry--; entry 1822 drivers/net/ethernet/intel/fm10k/fm10k_pci.c irq_set_affinity_hint(entry->vector, NULL); entry 1827 drivers/net/ethernet/intel/fm10k/fm10k_pci.c free_irq(entry->vector, q_vector); entry 392 drivers/net/ethernet/intel/i40e/i40e_ddp.c struct i40e_ddp_old_profile_list *entry; entry 397 drivers/net/ethernet/intel/i40e/i40e_ddp.c entry = list_first_entry(&pf->ddp_old_prof, entry 400 drivers/net/ethernet/intel/i40e/i40e_ddp.c status = i40e_ddp_load(netdev, entry->old_ddp_buf, entry 401 drivers/net/ethernet/intel/i40e/i40e_ddp.c entry->old_ddp_size, false); entry 402 drivers/net/ethernet/intel/i40e/i40e_ddp.c list_del(&entry->list); entry 403 drivers/net/ethernet/intel/i40e/i40e_ddp.c kfree(entry); entry 3576 drivers/net/ethernet/intel/i40e/i40e_ethtool.c struct i40e_flex_pit *entry, *tmp; entry 3580 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_for_each_entry_safe(entry, tmp, &pf->l3_flex_pit_list, list) { entry 3587 drivers/net/ethernet/intel/i40e/i40e_ethtool.c rule->flex_offset == entry->src_offset) { entry 3597 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_del(&entry->list); entry 3598 drivers/net/ethernet/intel/i40e/i40e_ethtool.c kfree(entry); entry 3603 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_for_each_entry_safe(entry, tmp, &pf->l4_flex_pit_list, list) { entry 3613 drivers/net/ethernet/intel/i40e/i40e_ethtool.c rule->flex_offset == entry->src_offset) { entry 3623 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_del(&entry->list); entry 3624 drivers/net/ethernet/intel/i40e/i40e_ethtool.c kfree(entry); entry 3674 drivers/net/ethernet/intel/i40e/i40e_ethtool.c struct i40e_flex_pit *entry; entry 3681 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_for_each_entry(entry, &pf->l4_flex_pit_list, list) entry 3682 drivers/net/ethernet/intel/i40e/i40e_ethtool.c clear_bit(entry->pit_index, &available_index); entry 3684 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_for_each_entry(entry, &pf->l3_flex_pit_list, list) entry 3685 drivers/net/ethernet/intel/i40e/i40e_ethtool.c clear_bit(entry->pit_index, &available_index); entry 3703 drivers/net/ethernet/intel/i40e/i40e_ethtool.c struct i40e_flex_pit *entry; entry 3709 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_for_each_entry(entry, flex_pit_list, list) { entry 3711 drivers/net/ethernet/intel/i40e/i40e_ethtool.c if (entry->src_offset == src_offset) entry 3712 drivers/net/ethernet/intel/i40e/i40e_ethtool.c return entry; entry 3743 drivers/net/ethernet/intel/i40e/i40e_ethtool.c struct i40e_flex_pit *new_pit, *entry; entry 3745 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_pit = kzalloc(sizeof(*entry), GFP_KERNEL); entry 3755 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_for_each_entry(entry, flex_pit_list, list) { entry 3756 drivers/net/ethernet/intel/i40e/i40e_ethtool.c if (new_pit->src_offset < entry->src_offset) { entry 3757 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_add_tail(&new_pit->list, &entry->list); entry 3765 drivers/net/ethernet/intel/i40e/i40e_ethtool.c if (new_pit->src_offset == entry->src_offset) { entry 3771 drivers/net/ethernet/intel/i40e/i40e_ethtool.c if (new_pit->pit_index != entry->pit_index) entry 3811 drivers/net/ethernet/intel/i40e/i40e_ethtool.c struct i40e_flex_pit *entry = NULL; entry 3818 drivers/net/ethernet/intel/i40e/i40e_ethtool.c list_for_each_entry(entry, flex_pit_list, list) { entry 3834 drivers/net/ethernet/intel/i40e/i40e_ethtool.c u16 offset = entry->src_offset + j; entry 3851 drivers/net/ethernet/intel/i40e/i40e_ethtool.c I40E_FLEX_PREP_VAL(entry->pit_index + 50, entry 3853 drivers/net/ethernet/intel/i40e/i40e_ethtool.c entry->src_offset)); entry 3864 drivers/net/ethernet/intel/i40e/i40e_ethtool.c last_offset = list_prev_entry(entry, list)->src_offset + 1; entry 11049 drivers/net/ethernet/intel/i40e/i40e_main.c pf->msix_entries[i].entry = i; entry 1189 drivers/net/ethernet/intel/iavf/iavf_main.c adapter->msix_entries[vector].entry = vector; entry 1508 drivers/net/ethernet/intel/ice/ice_adminq_cmd.h __le16 entry[1]; entry 487 drivers/net/ethernet/intel/ice/ice_common.c (((n) - 1) * sizeof(((struct ice_aqc_fw_logging_data *)0)->entry))) entry 520 drivers/net/ethernet/intel/ice/ice_common.c v = le16_to_cpu(config->entry[i]); entry 626 drivers/net/ethernet/intel/ice/ice_common.c data->entry[chgs++] = cpu_to_le16(val); entry 675 drivers/net/ethernet/intel/ice/ice_common.c v = le16_to_cpu(data->entry[i]); entry 1292 drivers/net/ethernet/intel/ice/ice_flex_pipe.c src = (u8 *)pid->entry; entry 346 drivers/net/ethernet/intel/ice/ice_flex_type.h struct ice_prof_tcam_entry entry[1]; entry 2412 drivers/net/ethernet/intel/ice/ice_main.c pf->msix_entries[i].entry = i; entry 1745 drivers/net/ethernet/intel/ice/ice_switch.c struct ice_aqc_sw_rules_elem *entry = r_iter; entry 1749 drivers/net/ethernet/intel/ice/ice_switch.c status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size, entry 2044 drivers/net/ethernet/intel/ice/ice_switch.c struct ice_fltr_mgmt_list_entry *entry; entry 2047 drivers/net/ethernet/intel/ice/ice_switch.c list_for_each_entry_safe(entry, tmp, rule_head, list_entry) { entry 2048 drivers/net/ethernet/intel/ice/ice_switch.c list_del(&entry->list_entry); entry 2049 drivers/net/ethernet/intel/ice/ice_switch.c devm_kfree(ice_hw_to_dev(hw), entry); entry 1126 drivers/net/ethernet/intel/igb/igb_main.c adapter->msix_entries[i].entry = i; entry 7189 drivers/net/ethernet/intel/igb/igb_main.c static bool igb_mac_entry_can_be_used(const struct igb_mac_addr *entry, entry 7192 drivers/net/ethernet/intel/igb/igb_main.c if (!(entry->state & IGB_MAC_STATE_IN_USE)) entry 7195 drivers/net/ethernet/intel/igb/igb_main.c if ((entry->state & IGB_MAC_STATE_SRC_ADDR) != entry 7199 drivers/net/ethernet/intel/igb/igb_main.c if (!ether_addr_equal(addr, entry->addr)) entry 7354 drivers/net/ethernet/intel/igb/igb_main.c struct vf_mac_filter *entry = NULL; entry 7361 drivers/net/ethernet/intel/igb/igb_main.c entry = list_entry(pos, struct vf_mac_filter, l); entry 7362 drivers/net/ethernet/intel/igb/igb_main.c if (entry->vf == vf) { entry 7363 drivers/net/ethernet/intel/igb/igb_main.c entry->vf = -1; entry 7364 drivers/net/ethernet/intel/igb/igb_main.c entry->free = true; entry 7365 drivers/net/ethernet/intel/igb/igb_main.c igb_del_mac_filter(adapter, entry->vf_mac, vf); entry 7386 drivers/net/ethernet/intel/igb/igb_main.c entry = list_entry(pos, struct vf_mac_filter, l); entry 7387 drivers/net/ethernet/intel/igb/igb_main.c if (entry->free) entry 7391 drivers/net/ethernet/intel/igb/igb_main.c if (entry && entry->free) { entry 7392 drivers/net/ethernet/intel/igb/igb_main.c entry->free = false; entry 7393 drivers/net/ethernet/intel/igb/igb_main.c entry->vf = vf; entry 7394 drivers/net/ethernet/intel/igb/igb_main.c ether_addr_copy(entry->vf_mac, addr); entry 1025 drivers/net/ethernet/intel/igbvf/netdev.c adapter->msix_entries[i].entry = i; entry 2405 drivers/net/ethernet/intel/igc/igc_main.c static bool igc_mac_entry_can_be_used(const struct igc_mac_addr *entry, entry 2408 drivers/net/ethernet/intel/igc/igc_main.c if (!(entry->state & IGC_MAC_STATE_IN_USE)) entry 2411 drivers/net/ethernet/intel/igc/igc_main.c if ((entry->state & IGC_MAC_STATE_SRC_ADDR) != entry 2415 drivers/net/ethernet/intel/igc/igc_main.c if (!ether_addr_equal(addr, entry->addr)) entry 3460 drivers/net/ethernet/intel/igc/igc_main.c adapter->msix_entries[i].entry = i; entry 773 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c adapter->msix_entries[i].entry = i; entry 3233 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct msix_entry *entry = &adapter->msix_entries[vector]; entry 3249 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c err = request_irq(entry->vector, &ixgbe_msix_clean_rings, 0, entry 3259 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c irq_set_affinity_hint(entry->vector, entry 3403 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct msix_entry *entry = &adapter->msix_entries[vector]; entry 3410 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c irq_set_affinity_hint(entry->vector, NULL); entry 3412 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c free_irq(entry->vector, q_vector); entry 10582 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct pci_dev *entry, *pdev = adapter->pdev; entry 10592 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c list_for_each_entry(entry, &adapter->pdev->bus->devices, bus_list) { entry 10594 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (entry->is_virtfn) entry 10603 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if ((entry->vendor != pdev->vendor) || entry 10604 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c (entry->device != pdev->device)) entry 639 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c struct vf_macvlans *entry; entry 645 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry = list_entry(pos, struct vf_macvlans, l); entry 646 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c if (entry->vf == vf) { entry 647 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry->vf = -1; entry 648 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry->free = true; entry 649 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry->is_macvlan = false; entry 651 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry->vf_macvlan, vf); entry 663 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry = NULL; entry 666 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry = list_entry(pos, struct vf_macvlans, l); entry 667 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c if (entry->free) entry 679 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c if (!entry || !entry->free) entry 686 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry->free = false; entry 687 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry->is_macvlan = true; entry 688 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c entry->vf = vf; entry 689 drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c memcpy(entry->vf_macvlan, mac_addr, ETH_ALEN); entry 1540 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct msix_entry *entry = &adapter->msix_entries[vector]; entry 1556 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c err = request_irq(entry->vector, &ixgbevf_msix_clean_rings, 0, entry 2664 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c adapter->msix_entries[vector].entry = vector; entry 1876 drivers/net/ethernet/marvell/mv643xx_eth.c u8 entry; entry 1880 drivers/net/ethernet/marvell/mv643xx_eth.c entry = a[5]; entry 1883 drivers/net/ethernet/marvell/mv643xx_eth.c entry = addr_crc(a); entry 1886 drivers/net/ethernet/marvell/mv643xx_eth.c table[entry >> 2] |= 1 << (8 * (entry & 3)); entry 1049 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c static void mvpp22_port_c2_lookup_disable(struct mvpp2_port *port, int entry) entry 1053 drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c mvpp2_cls_c2_read(port->priv, entry, &c2); entry 60 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_flow_tbl_entry *entry = s->private; entry 62 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c u32 hits = mvpp2_cls_flow_hits(entry->priv, entry->id); entry 73 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_flow_entry *entry = s->private; entry 75 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c u32 hits = mvpp2_cls_lookup_hits(entry->priv, entry->flow); entry 86 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_flow_entry *entry = s->private; entry 90 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c f = mvpp2_cls_flow_get(entry->flow); entry 126 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c const struct mvpp2_dbgfs_flow_entry *entry = s->private; entry 129 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c f = mvpp2_cls_flow_get(entry->flow); entry 142 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_port_flow_entry *entry = s->private; entry 143 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_port *port = entry->port; entry 149 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c f = mvpp2_cls_flow_get(entry->dbg_fe->flow); entry 153 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c flow_index = MVPP2_CLS_FLT_HASH_ENTRY(entry->port->id, f->flow_id); entry 168 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_port_flow_entry *entry = s->private; entry 169 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_port *port = entry->port; entry 174 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c f = mvpp2_cls_flow_get(entry->dbg_fe->flow); entry 178 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c flow_index = MVPP2_CLS_FLT_HASH_ENTRY(entry->port->id, f->flow_id); entry 193 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_c2_entry *entry = s->private; entry 196 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c hits = mvpp2_cls_c2_hit_count(entry->priv, entry->id); entry 207 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_c2_entry *entry = s->private; entry 211 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c mvpp2_cls_c2_read(entry->priv, entry->id, &c2); entry 228 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_c2_entry *entry = s->private; entry 232 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c mvpp2_cls_c2_read(entry->priv, entry->id, &c2); entry 339 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_prs_entry *entry = s->private; entry 340 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2 *priv = entry->priv; entry 342 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c seq_printf(s, "%x\n", priv->prs_shadow[entry->tid].lu); entry 351 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_prs_entry *entry = s->private; entry 355 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c mvpp2_prs_init_from_hw(entry->priv, &pe, entry->tid); entry 369 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_prs_entry *entry = s->private; entry 373 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c mvpp2_prs_init_from_hw(entry->priv, &pe, entry->tid); entry 387 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_prs_entry *entry = s->private; entry 392 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c mvpp2_prs_init_from_hw(entry->priv, &pe, entry->tid); entry 406 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_prs_entry *entry = s->private; entry 409 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c mvpp2_prs_init_from_hw(entry->priv, &pe, entry->tid); entry 420 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_prs_entry *entry = s->private; entry 423 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c val = mvpp2_prs_hits(entry->priv, entry->tid); entry 436 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_prs_entry *entry = s->private; entry 437 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2 *priv = entry->priv; entry 438 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c int tid = entry->tid; entry 449 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_flow_entry *entry) entry 459 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c port_entry->dbg_fe = entry; entry 473 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_flow_entry *entry; entry 482 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry = &priv->dbgfs_entries->flow_entries[flow]; entry 484 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry->flow = flow; entry 485 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry->priv = priv; entry 487 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("dec_hits", 0444, flow_entry_dir, entry, entry 490 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("type", 0444, flow_entry_dir, entry, entry 493 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("id", 0444, flow_entry_dir, entry, entry 499 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c priv->port_list[i], entry); entry 526 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_prs_entry *entry; entry 537 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry = &priv->dbgfs_entries->prs_entries[tid]; entry 539 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry->tid = tid; entry 540 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry->priv = priv; entry 543 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("sram", 0444, prs_entry_dir, entry, entry 546 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("valid", 0644, prs_entry_dir, entry, entry 549 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("lookup_id", 0644, prs_entry_dir, entry, entry 552 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("ai", 0644, prs_entry_dir, entry, entry 555 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("header_data", 0644, prs_entry_dir, entry, entry 558 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("hits", 0444, prs_entry_dir, entry, entry 561 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("pmap", 0444, prs_entry_dir, entry, entry 586 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_c2_entry *entry; entry 599 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry = &priv->dbgfs_entries->c2_entries[id]; entry 601 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry->id = id; entry 602 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry->priv = priv; entry 604 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("hits", 0444, c2_entry_dir, entry, entry 607 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("default_rxq", 0444, c2_entry_dir, entry, entry 610 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("rss_enable", 0444, c2_entry_dir, entry, entry 619 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c struct mvpp2_dbgfs_flow_tbl_entry *entry; entry 632 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry = &priv->dbgfs_entries->flt_entries[id]; entry 634 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry->id = id; entry 635 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c entry->priv = priv; entry 637 drivers/net/ethernet/marvell/mvpp2/mvpp2_debugfs.c debugfs_create_file("hits", 0444, flow_tbl_entry_dir, entry, entry 676 drivers/net/ethernet/marvell/octeontx2/af/mbox.h u16 entry; /* Entry allocated or start index if contiguous. entry 686 drivers/net/ethernet/marvell/octeontx2/af/mbox.h u16 entry; /* Entry index to be freed */ entry 701 drivers/net/ethernet/marvell/octeontx2/af/mbox.h u16 entry; /* MCAM entry to write this match key */ entry 711 drivers/net/ethernet/marvell/octeontx2/af/mbox.h u16 entry; entry 756 drivers/net/ethernet/marvell/octeontx2/af/mbox.h u16 entry; /* Entry and counter to be unmapped */ entry 772 drivers/net/ethernet/marvell/octeontx2/af/mbox.h u16 entry; entry 151 drivers/net/ethernet/marvell/octeontx2/af/rvu.h struct mcam_entry entry; entry 2510 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c free_req.entry = alloc_rsp.entry_list[0]; entry 123 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c static void npc_get_keyword(struct mcam_entry *entry, int idx, entry 138 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 = entry->kw[0]; entry 139 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask = entry->kw_mask[0]; entry 143 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 = entry->kw[1] & CAM_MASK(48); entry 144 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask = entry->kw_mask[1] & CAM_MASK(48); entry 150 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 = (entry->kw[1] >> 48) & CAM_MASK(16); entry 151 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 |= ((entry->kw[2] & CAM_MASK(48)) << 16); entry 152 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask = (entry->kw_mask[1] >> 48) & CAM_MASK(16); entry 153 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask |= ((entry->kw_mask[2] & CAM_MASK(48)) << 16); entry 159 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 = (entry->kw[2] >> 48) & CAM_MASK(16); entry 160 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 |= ((entry->kw[3] & CAM_MASK(32)) << 16); entry 161 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask = (entry->kw_mask[2] >> 48) & CAM_MASK(16); entry 162 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask |= ((entry->kw_mask[3] & CAM_MASK(32)) << 16); entry 168 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 = (entry->kw[3] >> 32) & CAM_MASK(32); entry 169 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 |= ((entry->kw[4] & CAM_MASK(32)) << 32); entry 170 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask = (entry->kw_mask[3] >> 32) & CAM_MASK(32); entry 171 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask |= ((entry->kw_mask[4] & CAM_MASK(32)) << 32); entry 177 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 = (entry->kw[4] >> 32) & CAM_MASK(32); entry 178 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 |= ((entry->kw[5] & CAM_MASK(16)) << 32); entry 179 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask = (entry->kw_mask[4] >> 32) & CAM_MASK(32); entry 180 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask |= ((entry->kw_mask[5] & CAM_MASK(16)) << 32); entry 186 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 = (entry->kw[5] >> 16) & CAM_MASK(48); entry 187 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 |= ((entry->kw[6] & CAM_MASK(16)) << 48); entry 188 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask = (entry->kw_mask[5] >> 16) & CAM_MASK(48); entry 189 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask |= ((entry->kw_mask[6] & CAM_MASK(16)) << 48); entry 193 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c *cam1 = (entry->kw[6] >> 16) & CAM_MASK(48); entry 194 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kw_mask = (entry->kw_mask[6] >> 16) & CAM_MASK(48); entry 204 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c struct mcam_entry *entry, bool enable) entry 230 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_get_keyword(entry, kw, &cam0, &cam1); entry 236 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_get_keyword(entry, kw + 1, &cam0, &cam1); entry 245 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NPC_AF_MCAMEX_BANKX_ACTION(index, actbank), entry->action); entry 249 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry->vtag_action); entry 313 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c struct mcam_entry entry = { {0} }; entry 333 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw[0] = chan; entry 334 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw_mask[0] = 0xFFFULL; entry 337 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw[kwi] = mac; entry 338 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw_mask[kwi] = BIT_ULL(48) - 1; entry 352 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.action = *(u64 *)&action; entry 354 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NIX_INTF_RX, &entry, true); entry 357 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw[0] |= (NPC_LT_LB_STAG | NPC_LT_LB_CTAG) << 20; entry 358 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw_mask[0] |= (NPC_LT_LB_STAG & NPC_LT_LB_CTAG) << 20; entry 360 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.vtag_action = VTAG0_VALID_BIT | entry 365 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c memcpy(&pfvf->entry, &entry, sizeof(entry)); entry 373 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c struct mcam_entry entry = { {0} }; entry 387 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw[0] = chan; entry 388 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw_mask[0] = 0xFFFULL; entry 392 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw[kwi] = BIT_ULL(40); /* LSB bit of 1st byte in DMAC */ entry 393 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw_mask[kwi] = BIT_ULL(40); entry 412 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.action = *(u64 *)&action; entry 414 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NIX_INTF_RX, &entry, true); entry 450 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c struct mcam_entry entry = { {0} }; entry 491 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw[0] = BIT_ULL(13) | chan; entry 492 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.kw_mask[0] = BIT_ULL(13) | 0xFFFULL; entry 508 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry.action = *(u64 *)&action; entry 510 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NIX_INTF_RX, &entry, true); entry 842 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c int kpu, int entry, bool pkind) entry 855 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c reg = NPC_AF_PKINDX_ACTION1(entry); entry 857 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c reg = NPC_AF_KPUX_ENTRYX_ACTION1(kpu, entry); entry 875 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c reg = NPC_AF_PKINDX_ACTION0(entry); entry 877 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c reg = NPC_AF_KPUX_ENTRYX_ACTION0(kpu, entry); entry 884 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c int kpu, int entry) entry 900 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NPC_AF_KPUX_ENTRYX_CAMX(kpu, entry, 0), *(u64 *)&cam0); entry 902 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NPC_AF_KPUX_ENTRYX_CAMX(kpu, entry, 1), *(u64 *)&cam1); entry 913 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c int entry, num_entries, max_entries; entry 925 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (entry = 0; entry < num_entries; entry++) entry 927 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c &profile->cam[entry], kpu, entry); entry 931 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (entry = 0; entry < num_entries; entry++) entry 932 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_config_kpuaction(rvu, blkaddr, &profile->action[entry], entry 933 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c kpu, entry, false); entry 1105 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c int blkaddr, entry, bank, err; entry 1117 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (entry = 0; entry < ((cfg >> 28) & 0xFFFF); entry++) entry 1119 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NPC_AF_MCAMEX_BANKX_CFG(entry, bank), 0); entry 1204 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 pcifunc, int entry) entry 1209 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c if (entry >= mcam->bmap_entries) entry 1212 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c if (pcifunc != mcam->entry2pfvf_map[entry]) entry 1234 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c int blkaddr, u16 entry, u16 cntr) entry 1236 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 index = entry & (mcam->banksize - 1); entry 1237 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 bank = npc_get_bank(mcam, entry); entry 1240 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c mcam->entry2cntr_map[entry] = cntr; entry 1250 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c int blkaddr, u16 entry, u16 cntr) entry 1252 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 index = entry & (mcam->banksize - 1); entry 1253 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 bank = npc_get_bank(mcam, entry); entry 1256 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c mcam->entry2cntr_map[entry] = NPC_MCAM_INVALID_MAP; entry 1269 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 entry, rentry; entry 1271 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry = index; entry 1274 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c __set_bit(entry, mcam->bmap); entry 1285 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 entry, rentry; entry 1287 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry = index; entry 1290 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c __clear_bit(entry, mcam->bmap); entry 1464 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c int entry, next_start; entry 1550 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rsp->entry = mcam->bmap_entries - index - max_contig; entry 1552 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rsp->entry = index; entry 1559 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (entry = 0; entry < req->count; entry++) { entry 1569 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry_list[entry] = index; entry 1610 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (entry = rsp->count - 1; entry >= 0; entry--) { entry 1612 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rsp->entry_list[index++] = entry_list[entry]; entry 1614 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rsp->entry_list[entry] = entry_list[entry]; entry 1619 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (entry = 0; entry < rsp->count; entry++) { entry 1621 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c (rsp->entry + entry) : rsp->entry_list[entry]; entry 1646 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rsp->entry = NPC_MCAM_ENTRY_INVALID; entry 1696 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rc = npc_mcam_verify_entry(mcam, pcifunc, req->entry); entry 1700 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c mcam->entry2pfvf_map[req->entry] = 0; entry 1701 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_mcam_clear_bit(mcam, req->entry); entry 1702 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_enable_mcam_entry(rvu, mcam, blkaddr, req->entry, false); entry 1705 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c cntr = mcam->entry2cntr_map[req->entry]; entry 1708 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c req->entry, cntr); entry 1733 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rc = npc_mcam_verify_entry(mcam, pcifunc, req->entry); entry 1748 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_config_mcam_entry(rvu, mcam, blkaddr, req->entry, req->intf, entry 1753 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c req->entry, req->cntr); entry 1774 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rc = npc_mcam_verify_entry(mcam, pcifunc, req->entry); entry 1779 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_enable_mcam_entry(rvu, mcam, blkaddr, req->entry, true); entry 1797 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rc = npc_mcam_verify_entry(mcam, pcifunc, req->entry); entry 1802 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_enable_mcam_entry(rvu, mcam, blkaddr, req->entry, false); entry 1944 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 index, entry = 0; entry 1963 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c while (entry < mcam->bmap_entries) { entry 1967 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c index = find_next_bit(mcam->bmap, mcam->bmap_entries, entry); entry 1973 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry = index + 1; entry 1986 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 index, entry = 0; entry 2000 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rc = npc_mcam_verify_entry(mcam, req->hdr.pcifunc, req->entry); entry 2004 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c req->entry, req->cntr); entry 2009 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c while (entry < mcam->bmap_entries) { entry 2013 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c index = find_next_bit(mcam->bmap, mcam->bmap_entries, entry); entry 2019 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry = index + 1; entry 2081 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c u16 entry = NPC_MCAM_ENTRY_INVALID; entry 2107 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c entry = entry_rsp.entry; entry 2121 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c mcam->entry2pfvf_map[entry] = 0; entry 2122 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_mcam_clear_bit(mcam, entry); entry 2131 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_config_mcam_entry(rvu, mcam, blkaddr, entry, req->intf, entry 2135 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c npc_map_mcam_entry_and_cntr(rvu, mcam, blkaddr, entry, cntr); entry 2138 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rsp->entry = entry; entry 2206 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c pfvf->entry.action = npc_get_mcam_action(rvu, mcam, blkaddr, index); entry 2209 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NIX_INTF_RX, &pfvf->entry, enable); entry 440 drivers/net/ethernet/marvell/pxa168_eth.c struct addr_table_entry *entry, *start; entry 467 drivers/net/ethernet/marvell/pxa168_eth.c entry = start + hash_function(mac_addr); entry 469 drivers/net/ethernet/marvell/pxa168_eth.c if (!(le32_to_cpu(entry->lo) & HASH_ENTRY_VALID)) { entry 473 drivers/net/ethernet/marvell/pxa168_eth.c if (((le32_to_cpu(entry->lo) & 0xfffffff8) == entry 475 drivers/net/ethernet/marvell/pxa168_eth.c (le32_to_cpu(entry->hi) == new_high)) { entry 479 drivers/net/ethernet/marvell/pxa168_eth.c if (entry == start + 0x7ff) entry 480 drivers/net/ethernet/marvell/pxa168_eth.c entry = start; entry 482 drivers/net/ethernet/marvell/pxa168_eth.c entry++; entry 485 drivers/net/ethernet/marvell/pxa168_eth.c if (((le32_to_cpu(entry->lo) & 0xfffffff8) != (new_low & 0xfffffff8)) && entry 486 drivers/net/ethernet/marvell/pxa168_eth.c (le32_to_cpu(entry->hi) != new_high) && del) entry 504 drivers/net/ethernet/marvell/pxa168_eth.c entry->hi = 0; entry 505 drivers/net/ethernet/marvell/pxa168_eth.c entry->lo = 0; entry 507 drivers/net/ethernet/marvell/pxa168_eth.c entry->hi = cpu_to_le32(new_high); entry 508 drivers/net/ethernet/marvell/pxa168_eth.c entry->lo = cpu_to_le32(new_low); entry 285 drivers/net/ethernet/mellanox/mlx4/alloc.c static void __mlx4_zone_remove_one_entry(struct mlx4_zone_entry *entry) entry 287 drivers/net/ethernet/mellanox/mlx4/alloc.c struct mlx4_zone_allocator *zone_alloc = entry->allocator; entry 289 drivers/net/ethernet/mellanox/mlx4/alloc.c if (!list_empty(&entry->prio_list)) { entry 291 drivers/net/ethernet/mellanox/mlx4/alloc.c if (!list_is_last(&entry->list, &zone_alloc->entries)) { entry 292 drivers/net/ethernet/mellanox/mlx4/alloc.c struct mlx4_zone_entry *next = list_first_entry(&entry->list, entry 296 drivers/net/ethernet/mellanox/mlx4/alloc.c if (next->priority == entry->priority) entry 297 drivers/net/ethernet/mellanox/mlx4/alloc.c list_add_tail(&next->prio_list, &entry->prio_list); entry 300 drivers/net/ethernet/mellanox/mlx4/alloc.c list_del(&entry->prio_list); entry 303 drivers/net/ethernet/mellanox/mlx4/alloc.c list_del(&entry->list); entry 705 drivers/net/ethernet/mellanox/mlx4/en_netdev.c struct mlx4_mac_entry *entry; entry 710 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_for_each_entry_safe(entry, tmp, bucket, hlist) { entry 711 drivers/net/ethernet/mellanox/mlx4/en_netdev.c if (ether_addr_equal_64bits(entry->mac, prev_mac)) { entry 712 drivers/net/ethernet/mellanox/mlx4/en_netdev.c mlx4_en_uc_steer_release(priv, entry->mac, entry 713 drivers/net/ethernet/mellanox/mlx4/en_netdev.c qpn, entry->reg_id); entry 716 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_del_rcu(&entry->hlist); entry 718 drivers/net/ethernet/mellanox/mlx4/en_netdev.c memcpy(entry->mac, new_mac, ETH_ALEN); entry 719 drivers/net/ethernet/mellanox/mlx4/en_netdev.c entry->reg_id = 0; entry 721 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_add_head_rcu(&entry->hlist, entry 726 drivers/net/ethernet/mellanox/mlx4/en_netdev.c &entry->reg_id); entry 1135 drivers/net/ethernet/mellanox/mlx4/en_netdev.c struct mlx4_mac_entry *entry; entry 1152 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_for_each_entry_safe(entry, tmp, bucket, hlist) { entry 1155 drivers/net/ethernet/mellanox/mlx4/en_netdev.c if (ether_addr_equal_64bits(entry->mac, entry 1163 drivers/net/ethernet/mellanox/mlx4/en_netdev.c if (ether_addr_equal_64bits(entry->mac, entry 1168 drivers/net/ethernet/mellanox/mlx4/en_netdev.c mac = mlx4_mac_to_u64(entry->mac); entry 1169 drivers/net/ethernet/mellanox/mlx4/en_netdev.c mlx4_en_uc_steer_release(priv, entry->mac, entry 1171 drivers/net/ethernet/mellanox/mlx4/en_netdev.c entry->reg_id); entry 1174 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_del_rcu(&entry->hlist); entry 1175 drivers/net/ethernet/mellanox/mlx4/en_netdev.c kfree_rcu(entry, rcu); entry 1177 drivers/net/ethernet/mellanox/mlx4/en_netdev.c entry->mac, priv->port); entry 1196 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_for_each_entry(entry, bucket, hlist) { entry 1197 drivers/net/ethernet/mellanox/mlx4/en_netdev.c if (ether_addr_equal_64bits(entry->mac, ha->addr)) { entry 1204 drivers/net/ethernet/mellanox/mlx4/en_netdev.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 1205 drivers/net/ethernet/mellanox/mlx4/en_netdev.c if (!entry) { entry 1212 drivers/net/ethernet/mellanox/mlx4/en_netdev.c memcpy(entry->mac, ha->addr, ETH_ALEN); entry 1217 drivers/net/ethernet/mellanox/mlx4/en_netdev.c kfree(entry); entry 1223 drivers/net/ethernet/mellanox/mlx4/en_netdev.c &entry->reg_id); entry 1228 drivers/net/ethernet/mellanox/mlx4/en_netdev.c kfree(entry); entry 1237 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_add_head_rcu(&entry->hlist, bucket); entry 1302 drivers/net/ethernet/mellanox/mlx4/en_netdev.c struct mlx4_mac_entry *entry; entry 1313 drivers/net/ethernet/mellanox/mlx4/en_netdev.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 1314 drivers/net/ethernet/mellanox/mlx4/en_netdev.c if (!entry) { entry 1319 drivers/net/ethernet/mellanox/mlx4/en_netdev.c memcpy(entry->mac, priv->dev->dev_addr, sizeof(entry->mac)); entry 1320 drivers/net/ethernet/mellanox/mlx4/en_netdev.c memcpy(priv->current_mac, entry->mac, sizeof(priv->current_mac)); entry 1321 drivers/net/ethernet/mellanox/mlx4/en_netdev.c entry->reg_id = reg_id; entry 1322 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_add_head_rcu(&entry->hlist, entry 1323 drivers/net/ethernet/mellanox/mlx4/en_netdev.c &priv->mac_hash[entry->mac[MLX4_EN_MAC_HASH_IDX]]); entry 1343 drivers/net/ethernet/mellanox/mlx4/en_netdev.c struct mlx4_mac_entry *entry; entry 1347 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_for_each_entry_safe(entry, tmp, bucket, hlist) { entry 1348 drivers/net/ethernet/mellanox/mlx4/en_netdev.c mac = mlx4_mac_to_u64(entry->mac); entry 1350 drivers/net/ethernet/mellanox/mlx4/en_netdev.c entry->mac); entry 1351 drivers/net/ethernet/mellanox/mlx4/en_netdev.c mlx4_en_uc_steer_release(priv, entry->mac, entry 1352 drivers/net/ethernet/mellanox/mlx4/en_netdev.c qpn, entry->reg_id); entry 1355 drivers/net/ethernet/mellanox/mlx4/en_netdev.c hlist_del_rcu(&entry->hlist); entry 1356 drivers/net/ethernet/mellanox/mlx4/en_netdev.c kfree_rcu(entry, rcu); entry 51 drivers/net/ethernet/mellanox/mlx4/en_port.c u32 entry; entry 60 drivers/net/ethernet/mellanox/mlx4/en_port.c entry = 0; entry 63 drivers/net/ethernet/mellanox/mlx4/en_port.c entry |= 1 << j; entry 64 drivers/net/ethernet/mellanox/mlx4/en_port.c filter->entry[i] = cpu_to_be32(entry); entry 45 drivers/net/ethernet/mellanox/mlx4/en_port.h __be32 entry[VLAN_FLTR_SIZE]; entry 740 drivers/net/ethernet/mellanox/mlx4/en_rx.c struct mlx4_mac_entry *entry; entry 747 drivers/net/ethernet/mellanox/mlx4/en_rx.c hlist_for_each_entry_rcu(entry, bucket, hlist) { entry 748 drivers/net/ethernet/mellanox/mlx4/en_rx.c if (ether_addr_equal_64bits(entry->mac, entry 106 drivers/net/ethernet/mellanox/mlx4/eq.c static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor, entry 110 drivers/net/ethernet/mellanox/mlx4/eq.c unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; entry 2668 drivers/net/ethernet/mellanox/mlx4/main.c void mlx4_set_admin_guid(struct mlx4_dev *dev, __be64 guid, int entry, int port) entry 2672 drivers/net/ethernet/mellanox/mlx4/main.c priv->mfunc.master.vf_admin[entry].vport[port].guid = guid; entry 2676 drivers/net/ethernet/mellanox/mlx4/main.c __be64 mlx4_get_admin_guid(struct mlx4_dev *dev, int entry, int port) entry 2680 drivers/net/ethernet/mellanox/mlx4/main.c return priv->mfunc.master.vf_admin[entry].vport[port].guid; entry 2684 drivers/net/ethernet/mellanox/mlx4/main.c void mlx4_set_random_admin_guid(struct mlx4_dev *dev, int entry, int port) entry 2690 drivers/net/ethernet/mellanox/mlx4/main.c if (entry == 0) entry 2696 drivers/net/ethernet/mellanox/mlx4/main.c priv->mfunc.master.vf_admin[entry].vport[port].guid = guid; entry 2952 drivers/net/ethernet/mellanox/mlx4/main.c entries[i].entry = i; entry 3127 drivers/net/ethernet/mellanox/mlx4/main.c struct mlx4_steer_index *entry, *tmp_entry; entry 3140 drivers/net/ethernet/mellanox/mlx4/main.c list_for_each_entry_safe(entry, tmp_entry, entry 3143 drivers/net/ethernet/mellanox/mlx4/main.c list_del(&entry->list); entry 3145 drivers/net/ethernet/mellanox/mlx4/main.c &entry->duplicates, entry 3150 drivers/net/ethernet/mellanox/mlx4/main.c kfree(entry); entry 244 drivers/net/ethernet/mellanox/mlx4/mcg.c struct mlx4_steer_index *tmp_entry, *entry = NULL; entry 259 drivers/net/ethernet/mellanox/mlx4/mcg.c entry = tmp_entry; entry 263 drivers/net/ethernet/mellanox/mlx4/mcg.c if (unlikely(!entry)) { entry 271 drivers/net/ethernet/mellanox/mlx4/mcg.c list_for_each_entry(dqp, &entry->duplicates, list) { entry 281 drivers/net/ethernet/mellanox/mlx4/mcg.c list_add_tail(&dqp->list, &entry->duplicates); entry 293 drivers/net/ethernet/mellanox/mlx4/mcg.c struct mlx4_steer_index *tmp_entry, *entry = NULL; entry 309 drivers/net/ethernet/mellanox/mlx4/mcg.c entry = tmp_entry; entry 313 drivers/net/ethernet/mellanox/mlx4/mcg.c if (unlikely(!entry)) { entry 317 drivers/net/ethernet/mellanox/mlx4/mcg.c list_for_each_entry_safe(dqp, tmp_dqp, &entry->duplicates, list) { entry 373 drivers/net/ethernet/mellanox/mlx4/mcg.c struct mlx4_steer_index *entry = NULL, *tmp_entry; entry 389 drivers/net/ethernet/mellanox/mlx4/mcg.c list_for_each_entry_safe(entry, tmp_entry, &s_steer->steer_entries[steer], list) { entry 390 drivers/net/ethernet/mellanox/mlx4/mcg.c if (entry->index == index) { entry 391 drivers/net/ethernet/mellanox/mlx4/mcg.c if (list_empty(&entry->duplicates) || entry 398 drivers/net/ethernet/mellanox/mlx4/mcg.c list_del(&entry->list); entry 400 drivers/net/ethernet/mellanox/mlx4/mcg.c &entry->duplicates, entry 405 drivers/net/ethernet/mellanox/mlx4/mcg.c kfree(entry); entry 424 drivers/net/ethernet/mellanox/mlx4/mcg.c struct mlx4_steer_index *entry; entry 465 drivers/net/ethernet/mellanox/mlx4/mcg.c list_for_each_entry(entry, entry 468 drivers/net/ethernet/mellanox/mlx4/mcg.c err = mlx4_READ_ENTRY(dev, entry->index, mailbox); entry 489 drivers/net/ethernet/mellanox/mlx4/mcg.c &entry->duplicates); entry 506 drivers/net/ethernet/mellanox/mlx4/mcg.c err = mlx4_WRITE_ENTRY(dev, entry->index, entry 555 drivers/net/ethernet/mellanox/mlx4/mcg.c struct mlx4_steer_index *entry, *tmp_entry; entry 600 drivers/net/ethernet/mellanox/mlx4/mcg.c list_for_each_entry_safe(entry, tmp_entry, entry 604 drivers/net/ethernet/mellanox/mlx4/mcg.c list_for_each_entry(dqp, &entry->duplicates, list) { entry 620 drivers/net/ethernet/mellanox/mlx4/mcg.c entry->index, entry 629 drivers/net/ethernet/mellanox/mlx4/mcg.c qpn, entry->index); entry 630 drivers/net/ethernet/mellanox/mlx4/mcg.c list_del(&entry->list); entry 631 drivers/net/ethernet/mellanox/mlx4/mcg.c kfree(entry); entry 644 drivers/net/ethernet/mellanox/mlx4/mcg.c qpn, entry->index); entry 659 drivers/net/ethernet/mellanox/mlx4/mcg.c entry->index, entry 457 drivers/net/ethernet/mellanox/mlx4/mlx4.h __be32 entry[VLAN_FLTR_SIZE]; entry 50 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h static inline struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry) entry 52 drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE); entry 165 drivers/net/ethernet/mellanox/mlx5/core/rl.c struct mlx5_rl_entry *entry; entry 177 drivers/net/ethernet/mellanox/mlx5/core/rl.c entry = find_rl_entry(table, rl); entry 178 drivers/net/ethernet/mellanox/mlx5/core/rl.c if (!entry) { entry 184 drivers/net/ethernet/mellanox/mlx5/core/rl.c if (entry->refcount) { entry 186 drivers/net/ethernet/mellanox/mlx5/core/rl.c entry->refcount++; entry 189 drivers/net/ethernet/mellanox/mlx5/core/rl.c err = mlx5_set_pp_rate_limit_cmd(dev, entry->index, rl); entry 196 drivers/net/ethernet/mellanox/mlx5/core/rl.c entry->rl = *rl; entry 197 drivers/net/ethernet/mellanox/mlx5/core/rl.c entry->refcount = 1; entry 199 drivers/net/ethernet/mellanox/mlx5/core/rl.c *index = entry->index; entry 210 drivers/net/ethernet/mellanox/mlx5/core/rl.c struct mlx5_rl_entry *entry = NULL; entry 218 drivers/net/ethernet/mellanox/mlx5/core/rl.c entry = find_rl_entry(table, rl); entry 219 drivers/net/ethernet/mellanox/mlx5/core/rl.c if (!entry || !entry->refcount) { entry 225 drivers/net/ethernet/mellanox/mlx5/core/rl.c entry->refcount--; entry 226 drivers/net/ethernet/mellanox/mlx5/core/rl.c if (!entry->refcount) { entry 228 drivers/net/ethernet/mellanox/mlx5/core/rl.c mlx5_set_pp_rate_limit_cmd(dev, entry->index, &reset_rl); entry 229 drivers/net/ethernet/mellanox/mlx5/core/rl.c entry->rl = reset_rl; entry 177 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c struct mlxsw_sp1_acl_tcam_entry *entry = entry_priv; entry 180 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c &chunk->cchunk, &entry->centry, entry 190 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c struct mlxsw_sp1_acl_tcam_entry *entry = entry_priv; entry 193 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c &chunk->cchunk, &entry->centry); entry 228 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c struct mlxsw_sp1_acl_tcam_entry *entry = entry_priv; entry 231 drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c offset = mlxsw_sp_acl_ctcam_entry_offset(&entry->centry); entry 207 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c struct mlxsw_sp2_acl_tcam_entry *entry = entry_priv; entry 209 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c entry->act_block = rulei->act_block; entry 211 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c &chunk->achunk, &entry->aentry, entry 221 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c struct mlxsw_sp2_acl_tcam_entry *entry = entry_priv; entry 224 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c &entry->aentry); entry 233 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c struct mlxsw_sp2_acl_tcam_entry *entry = entry_priv; entry 235 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c entry->act_block = rulei->act_block; entry 238 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c &entry->aentry, rulei); entry 246 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c struct mlxsw_sp2_acl_tcam_entry *entry = entry_priv; entry 248 drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c return mlxsw_afa_block_activity_get(entry->act_block, activity); entry 251 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c struct mlxsw_sp_acl_tcam_entry *entry; entry 1107 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c struct mlxsw_sp_acl_tcam_entry *entry; entry 1110 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c entry = kzalloc(sizeof(*entry) + ops->entry_priv_size, GFP_KERNEL); entry 1111 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c if (!entry) entry 1113 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c entry->ventry = ventry; entry 1114 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c entry->chunk = chunk; entry 1117 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c entry->priv, ventry->rulei); entry 1121 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c return entry; entry 1124 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c kfree(entry); entry 1129 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c struct mlxsw_sp_acl_tcam_entry *entry) entry 1133 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c ops->entry_del(mlxsw_sp, entry->chunk->region->priv, entry 1134 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c entry->chunk->priv, entry->priv); entry 1135 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c kfree(entry); entry 1141 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c struct mlxsw_sp_acl_tcam_entry *entry, entry 1147 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c entry->priv, rulei); entry 1152 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c struct mlxsw_sp_acl_tcam_entry *entry, entry 1157 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c return ops->entry_activity_get(mlxsw_sp, entry->chunk->region->priv, entry 1158 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c entry->priv, activity); entry 1180 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c ventry->entry = mlxsw_sp_acl_tcam_entry_create(mlxsw_sp, ventry, entry 1182 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c if (IS_ERR(ventry->entry)) { entry 1184 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c err = PTR_ERR(ventry->entry); entry 1208 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c mlxsw_sp_acl_tcam_entry_destroy(mlxsw_sp, ventry->entry); entry 1222 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c ventry->entry, rulei); entry 1231 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c ventry->entry, activity); entry 1243 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c if (ventry->entry->chunk == chunk) entry 1252 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c mlxsw_sp_acl_tcam_entry_destroy(mlxsw_sp, ventry->entry); entry 1253 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c ventry->entry = new_entry; entry 123 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c static int mlxsw_sp_erif_entry_prepare(struct devlink_dpipe_entry *entry, entry 129 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->match_values = match_value; entry 130 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->match_values_count = 1; entry 132 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->action_values = action_value; entry 133 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->action_values_count = 1; entry 154 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c struct devlink_dpipe_entry *entry, entry 164 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c rif_value = entry->match_values->value; entry 166 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->match_values->mapping_value = mlxsw_sp_rif_dev_ifindex(rif); entry 167 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->match_values->mapping_valid = true; entry 170 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c action_value = entry->action_values->value; entry 173 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->counter_valid = false; entry 174 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->counter = 0; entry 175 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->index = mlxsw_sp_rif_index(rif); entry 184 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->counter = cnt; entry 185 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->counter_valid = true; entry 197 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c struct devlink_dpipe_entry entry = {0}; entry 207 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = mlxsw_sp_erif_entry_prepare(&entry, &match_value, &match, entry 225 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = mlxsw_sp_erif_entry_get(mlxsw_sp, &entry, rif, entry 229 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = devlink_dpipe_entry_ctx_append(dump_ctx, &entry); entry 246 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c devlink_dpipe_entry_clear(&entry); entry 252 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c devlink_dpipe_entry_clear(&entry); entry 400 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host_entry_prepare(struct devlink_dpipe_entry *entry, entry 410 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->match_values = match_values; entry 411 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->match_values_count = MLXSW_SP_DPIPE_TABLE_HOST_MATCH_COUNT; entry 413 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->action_values = action_value; entry 414 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->action_values_count = 1; entry 455 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c __mlxsw_sp_dpipe_table_host_entry_fill(struct devlink_dpipe_entry *entry, entry 464 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_HOST_MATCH_RIF]; entry 472 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_HOST_MATCH_DIP]; entry 476 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c value = entry->action_values; entry 482 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host4_entry_fill(struct devlink_dpipe_entry *entry, entry 491 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c __mlxsw_sp_dpipe_table_host_entry_fill(entry, rif, ha, &dip); entry 495 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host6_entry_fill(struct devlink_dpipe_entry *entry, entry 505 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c __mlxsw_sp_dpipe_table_host_entry_fill(entry, rif, ha, dip); entry 510 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c struct devlink_dpipe_entry *entry, entry 519 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host4_entry_fill(entry, neigh_entry, rif); entry 522 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host6_entry_fill(entry, neigh_entry, rif); entry 530 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c &entry->counter); entry 532 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->counter_valid = true; entry 537 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c struct devlink_dpipe_entry *entry, entry 579 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_host_entry_fill(mlxsw_sp, entry, entry 582 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->index = neigh_count; entry 583 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = devlink_dpipe_entry_ctx_append(dump_ctx, entry); entry 624 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c struct devlink_dpipe_entry entry = {0}; entry 634 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = mlxsw_sp_dpipe_table_host_entry_prepare(&entry, match_values, entry 640 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = mlxsw_sp_dpipe_table_host_entries_get(mlxsw_sp, &entry, entry 644 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c devlink_dpipe_entry_clear(&entry); entry 967 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_adj_entry_prepare(struct devlink_dpipe_entry *entry, entry 977 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->match_values = match_values; entry 978 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->match_values_count = MLXSW_SP_DPIPE_TABLE_ADJ_MATCH_COUNT; entry 980 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->action_values = action_values; entry 981 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->action_values_count = MLXSW_SP_DPIPE_TABLE_ADJ_ACTION_COUNT; entry 1032 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c __mlxsw_sp_dpipe_table_adj_entry_fill(struct devlink_dpipe_entry *entry, entry 1041 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_ADJ_MATCH_INDEX]; entry 1045 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_ADJ_MATCH_SIZE]; entry 1049 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c value = &entry->match_values[MLXSW_SP_DPIPE_TABLE_ADJ_MATCH_HASH_INDEX]; entry 1053 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c value = &entry->action_values[MLXSW_SP_DPIPE_TABLE_ADJ_ACTION_DST_MAC]; entry 1056 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c value = &entry->action_values[MLXSW_SP_DPIPE_TABLE_ADJ_ACTION_ERIF_PORT]; entry 1065 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c struct devlink_dpipe_entry *entry) entry 1075 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c __mlxsw_sp_dpipe_table_adj_entry_fill(entry, adj_index, adj_size, entry 1077 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = mlxsw_sp_nexthop_counter_get(mlxsw_sp, nh, &entry->counter); entry 1079 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->counter_valid = true; entry 1084 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c struct devlink_dpipe_entry *entry, entry 1113 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c mlxsw_sp_dpipe_table_adj_entry_fill(mlxsw_sp, nh, entry); entry 1114 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c entry->index = entry_index; entry 1115 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = devlink_dpipe_entry_ctx_append(dump_ctx, entry); entry 1151 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c struct devlink_dpipe_entry entry = {0}; entry 1165 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = mlxsw_sp_dpipe_table_adj_entry_prepare(&entry, entry 1171 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c err = mlxsw_sp_dpipe_table_adj_entries_get(mlxsw_sp, &entry, entry 1174 drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c devlink_dpipe_entry_clear(&entry); entry 2641 drivers/net/ethernet/micrel/ksz884x.c int entry; entry 2645 drivers/net/ethernet/micrel/ksz884x.c for (entry = 0; entry < VLAN_TABLE_ENTRIES; entry++) { entry 2646 drivers/net/ethernet/micrel/ksz884x.c sw_r_vlan_table(hw, entry, entry 2647 drivers/net/ethernet/micrel/ksz884x.c &sw->vlan_table[entry].vid, entry 2648 drivers/net/ethernet/micrel/ksz884x.c &sw->vlan_table[entry].fid, entry 2649 drivers/net/ethernet/micrel/ksz884x.c &sw->vlan_table[entry].member); entry 2828 drivers/net/ethernet/micrel/ksz884x.c struct ksz_mac_table *entry; entry 2832 drivers/net/ethernet/micrel/ksz884x.c entry = &hw->ksz_switch->mac_table[i]; entry 2834 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr, entry->ports, entry 2835 drivers/net/ethernet/micrel/ksz884x.c entry->override, 0, entry 2836 drivers/net/ethernet/micrel/ksz884x.c entry->use_fid, entry->fid); entry 2848 drivers/net/ethernet/micrel/ksz884x.c struct ksz_mac_table *entry; entry 2850 drivers/net/ethernet/micrel/ksz884x.c entry = &hw->ksz_switch->mac_table[STP_ENTRY]; entry 2851 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr[0] = 0x01; entry 2852 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr[1] = 0x80; entry 2853 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr[2] = 0xC2; entry 2854 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr[3] = 0x00; entry 2855 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr[4] = 0x00; entry 2856 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr[5] = 0x00; entry 2857 drivers/net/ethernet/micrel/ksz884x.c entry->ports = HOST_MASK; entry 2858 drivers/net/ethernet/micrel/ksz884x.c entry->override = 1; entry 2859 drivers/net/ethernet/micrel/ksz884x.c entry->valid = 1; entry 2861 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr, entry->ports, entry 2862 drivers/net/ethernet/micrel/ksz884x.c entry->override, entry->valid, entry 2863 drivers/net/ethernet/micrel/ksz884x.c entry->use_fid, entry->fid); entry 2874 drivers/net/ethernet/micrel/ksz884x.c struct ksz_mac_table *entry; entry 2878 drivers/net/ethernet/micrel/ksz884x.c entry = &hw->ksz_switch->mac_table[i]; entry 2879 drivers/net/ethernet/micrel/ksz884x.c entry->valid = 0; entry 2881 drivers/net/ethernet/micrel/ksz884x.c entry->mac_addr, entry->ports, entry 2882 drivers/net/ethernet/micrel/ksz884x.c entry->override, entry->valid, entry 2883 drivers/net/ethernet/micrel/ksz884x.c entry->use_fid, entry->fid); entry 469 drivers/net/ethernet/microchip/lan743x_main.c msix_entries[index].entry = index; entry 811 drivers/net/ethernet/mscc/ocelot.c static int ocelot_fdb_do_dump(struct ocelot_mact_entry *entry, entry 836 drivers/net/ethernet/mscc/ocelot.c if (nla_put(dump->skb, NDA_LLADDR, ETH_ALEN, entry->mac)) entry 839 drivers/net/ethernet/mscc/ocelot.c if (entry->vid && nla_put_u16(dump->skb, NDA_VLAN, entry->vid)) entry 854 drivers/net/ethernet/mscc/ocelot.c struct ocelot_mact_entry *entry) entry 895 drivers/net/ethernet/mscc/ocelot.c entry->vid = (mach >> 16) & 0xfff; entry 896 drivers/net/ethernet/mscc/ocelot.c ether_addr_copy(entry->mac, mac); entry 914 drivers/net/ethernet/mscc/ocelot.c struct ocelot_mact_entry entry; entry 921 drivers/net/ethernet/mscc/ocelot.c ret = ocelot_mact_read(port, i, j, &entry); entry 930 drivers/net/ethernet/mscc/ocelot.c ret = ocelot_fdb_do_dump(&entry, &dump); entry 2213 drivers/net/ethernet/mscc/ocelot.c struct ocelot_skb *entry; entry 2227 drivers/net/ethernet/mscc/ocelot.c entry = list_entry(pos, struct ocelot_skb, head); entry 2230 drivers/net/ethernet/mscc/ocelot.c dev_kfree_skb_any(entry->skb); entry 2231 drivers/net/ethernet/mscc/ocelot.c kfree(entry); entry 83 drivers/net/ethernet/mscc/ocelot_ace.c u32 entry[VCAP_ENTRY_WIDTH]; /* ENTRY_DAT */ entry 138 drivers/net/ethernet/mscc/ocelot_ace.c ocelot_write_rix(oc, data->entry[i], S2_CACHE_ENTRY_DAT, i); entry 149 drivers/net/ethernet/mscc/ocelot_ace.c data->entry[i] = ocelot_read_rix(oc, S2_CACHE_ENTRY_DAT, i); entry 248 drivers/net/ethernet/mscc/ocelot_ace.c vcap_data_set(data->entry, offset + data->key_offset, width, value); entry 203 drivers/net/ethernet/mscc/ocelot_board.c struct ocelot_skb *entry; entry 224 drivers/net/ethernet/mscc/ocelot_board.c entry = list_entry(pos, struct ocelot_skb, head); entry 225 drivers/net/ethernet/mscc/ocelot_board.c if (entry->id != id) entry 228 drivers/net/ethernet/mscc/ocelot_board.c skb = entry->skb; entry 231 drivers/net/ethernet/mscc/ocelot_board.c kfree(entry); entry 157 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct mcp_slot *entry; entry 943 drivers/net/ethernet/myricom/myri10ge/myri10ge.c bytes = mgp->max_intr_slots * sizeof(*mgp->ss[0].rx_done.entry); entry 1048 drivers/net/ethernet/myricom/myri10ge/myri10ge.c memset(ss->rx_done.entry, 0, bytes); entry 1469 drivers/net/ethernet/myricom/myri10ge/myri10ge.c while (rx_done->entry[idx].length != 0 && work_done < budget) { entry 1470 drivers/net/ethernet/myricom/myri10ge/myri10ge.c length = ntohs(rx_done->entry[idx].length); entry 1471 drivers/net/ethernet/myricom/myri10ge/myri10ge.c rx_done->entry[idx].length = 0; entry 1472 drivers/net/ethernet/myricom/myri10ge/myri10ge.c checksum = csum_unfold(rx_done->entry[idx].checksum); entry 3566 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (ss->rx_done.entry != NULL) { entry 3568 drivers/net/ethernet/myricom/myri10ge/myri10ge.c sizeof(*ss->rx_done.entry); entry 3570 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->rx_done.entry, ss->rx_done.bus); entry 3571 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->rx_done.entry = NULL; entry 3603 drivers/net/ethernet/myricom/myri10ge/myri10ge.c bytes = mgp->max_intr_slots * sizeof(*ss->rx_done.entry); entry 3604 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->rx_done.entry = dma_alloc_coherent(&pdev->dev, bytes, entry 3607 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (ss->rx_done.entry == NULL) entry 3717 drivers/net/ethernet/myricom/myri10ge/myri10ge.c mgp->msix_vectors[i].entry = i; entry 1935 drivers/net/ethernet/natsemi/natsemi.c int entry = np->dirty_rx % RX_RING_SIZE; entry 1936 drivers/net/ethernet/natsemi/natsemi.c if (np->rx_skbuff[entry] == NULL) { entry 1939 drivers/net/ethernet/natsemi/natsemi.c np->rx_skbuff[entry] = skb; entry 1942 drivers/net/ethernet/natsemi/natsemi.c np->rx_dma[entry] = pci_map_single(np->pci_dev, entry 1945 drivers/net/ethernet/natsemi/natsemi.c np->rx_dma[entry])) { entry 1947 drivers/net/ethernet/natsemi/natsemi.c np->rx_skbuff[entry] = NULL; entry 1950 drivers/net/ethernet/natsemi/natsemi.c np->rx_ring[entry].addr = cpu_to_le32(np->rx_dma[entry]); entry 1952 drivers/net/ethernet/natsemi/natsemi.c np->rx_ring[entry].cmd_status = cpu_to_le32(np->rx_buf_sz); entry 2094 drivers/net/ethernet/natsemi/natsemi.c unsigned entry; entry 2101 drivers/net/ethernet/natsemi/natsemi.c entry = np->cur_tx % TX_RING_SIZE; entry 2103 drivers/net/ethernet/natsemi/natsemi.c np->tx_skbuff[entry] = skb; entry 2104 drivers/net/ethernet/natsemi/natsemi.c np->tx_dma[entry] = pci_map_single(np->pci_dev, entry 2106 drivers/net/ethernet/natsemi/natsemi.c if (pci_dma_mapping_error(np->pci_dev, np->tx_dma[entry])) { entry 2107 drivers/net/ethernet/natsemi/natsemi.c np->tx_skbuff[entry] = NULL; entry 2113 drivers/net/ethernet/natsemi/natsemi.c np->tx_ring[entry].addr = cpu_to_le32(np->tx_dma[entry]); entry 2118 drivers/net/ethernet/natsemi/natsemi.c np->tx_ring[entry].cmd_status = cpu_to_le32(DescOwn | skb->len); entry 2138 drivers/net/ethernet/natsemi/natsemi.c dev->name, np->cur_tx, entry); entry 2148 drivers/net/ethernet/natsemi/natsemi.c int entry = np->dirty_tx % TX_RING_SIZE; entry 2149 drivers/net/ethernet/natsemi/natsemi.c if (np->tx_ring[entry].cmd_status & cpu_to_le32(DescOwn)) entry 2155 drivers/net/ethernet/natsemi/natsemi.c le32_to_cpu(np->tx_ring[entry].cmd_status)); entry 2156 drivers/net/ethernet/natsemi/natsemi.c if (np->tx_ring[entry].cmd_status & cpu_to_le32(DescPktOK)) { entry 2158 drivers/net/ethernet/natsemi/natsemi.c dev->stats.tx_bytes += np->tx_skbuff[entry]->len; entry 2161 drivers/net/ethernet/natsemi/natsemi.c le32_to_cpu(np->tx_ring[entry].cmd_status); entry 2172 drivers/net/ethernet/natsemi/natsemi.c pci_unmap_single(np->pci_dev,np->tx_dma[entry], entry 2173 drivers/net/ethernet/natsemi/natsemi.c np->tx_skbuff[entry]->len, entry 2176 drivers/net/ethernet/natsemi/natsemi.c dev_consume_skb_irq(np->tx_skbuff[entry]); entry 2177 drivers/net/ethernet/natsemi/natsemi.c np->tx_skbuff[entry] = NULL; entry 2285 drivers/net/ethernet/natsemi/natsemi.c int entry = np->cur_rx % RX_RING_SIZE; entry 2297 drivers/net/ethernet/natsemi/natsemi.c entry, desc_status); entry 2363 drivers/net/ethernet/natsemi/natsemi.c np->rx_dma[entry], entry 2367 drivers/net/ethernet/natsemi/natsemi.c np->rx_skbuff[entry]->data, pkt_len); entry 2370 drivers/net/ethernet/natsemi/natsemi.c np->rx_dma[entry], entry 2374 drivers/net/ethernet/natsemi/natsemi.c pci_unmap_single(np->pci_dev, np->rx_dma[entry], entry 2377 drivers/net/ethernet/natsemi/natsemi.c skb_put(skb = np->rx_skbuff[entry], pkt_len); entry 2378 drivers/net/ethernet/natsemi/natsemi.c np->rx_skbuff[entry] = NULL; entry 2385 drivers/net/ethernet/natsemi/natsemi.c entry = (++np->cur_rx) % RX_RING_SIZE; entry 2386 drivers/net/ethernet/natsemi/natsemi.c np->rx_head_desc = &np->rx_ring[entry]; entry 242 drivers/net/ethernet/natsemi/sonic.c int entry; entry 267 drivers/net/ethernet/natsemi/sonic.c entry = lp->next_tx; entry 269 drivers/net/ethernet/natsemi/sonic.c sonic_tda_put(dev, entry, SONIC_TD_STATUS, 0); /* clear status */ entry 270 drivers/net/ethernet/natsemi/sonic.c sonic_tda_put(dev, entry, SONIC_TD_FRAG_COUNT, 1); /* single fragment */ entry 271 drivers/net/ethernet/natsemi/sonic.c sonic_tda_put(dev, entry, SONIC_TD_PKTSIZE, length); /* length of packet */ entry 272 drivers/net/ethernet/natsemi/sonic.c sonic_tda_put(dev, entry, SONIC_TD_FRAG_PTR_L, laddr & 0xffff); entry 273 drivers/net/ethernet/natsemi/sonic.c sonic_tda_put(dev, entry, SONIC_TD_FRAG_PTR_H, laddr >> 16); entry 274 drivers/net/ethernet/natsemi/sonic.c sonic_tda_put(dev, entry, SONIC_TD_FRAG_SIZE, length); entry 275 drivers/net/ethernet/natsemi/sonic.c sonic_tda_put(dev, entry, SONIC_TD_LINK, entry 276 drivers/net/ethernet/natsemi/sonic.c sonic_tda_get(dev, entry, SONIC_TD_LINK) | SONIC_EOL); entry 279 drivers/net/ethernet/natsemi/sonic.c lp->tx_len[entry] = length; entry 280 drivers/net/ethernet/natsemi/sonic.c lp->tx_laddr[entry] = laddr; entry 281 drivers/net/ethernet/natsemi/sonic.c lp->tx_skb[entry] = skb; entry 286 drivers/net/ethernet/natsemi/sonic.c lp->eol_tx = entry; entry 288 drivers/net/ethernet/natsemi/sonic.c lp->next_tx = (entry + 1) & SONIC_TDS_MASK; entry 339 drivers/net/ethernet/natsemi/sonic.c int entry = lp->cur_tx; entry 353 drivers/net/ethernet/natsemi/sonic.c while (lp->tx_skb[entry] != NULL) { entry 354 drivers/net/ethernet/natsemi/sonic.c if ((td_status = sonic_tda_get(dev, entry, SONIC_TD_STATUS)) == 0) entry 359 drivers/net/ethernet/natsemi/sonic.c lp->stats.tx_bytes += sonic_tda_get(dev, entry, SONIC_TD_PKTSIZE); entry 374 drivers/net/ethernet/natsemi/sonic.c dev_consume_skb_irq(lp->tx_skb[entry]); entry 375 drivers/net/ethernet/natsemi/sonic.c lp->tx_skb[entry] = NULL; entry 377 drivers/net/ethernet/natsemi/sonic.c dma_unmap_single(lp->device, lp->tx_laddr[entry], lp->tx_len[entry], DMA_TO_DEVICE); entry 378 drivers/net/ethernet/natsemi/sonic.c lp->tx_laddr[entry] = (dma_addr_t)0; entry 381 drivers/net/ethernet/natsemi/sonic.c if (sonic_tda_get(dev, entry, SONIC_TD_LINK) & SONIC_EOL) { entry 382 drivers/net/ethernet/natsemi/sonic.c entry = (entry + 1) & SONIC_TDS_MASK; entry 385 drivers/net/ethernet/natsemi/sonic.c entry = (entry + 1) & SONIC_TDS_MASK; entry 388 drivers/net/ethernet/natsemi/sonic.c if (freed_some || lp->tx_skb[entry] == NULL) entry 390 drivers/net/ethernet/natsemi/sonic.c lp->cur_tx = entry; entry 490 drivers/net/ethernet/natsemi/sonic.c unsigned int entry = sonic_rr_entry(dev, SONIC_READ(SONIC_RWP)); entry 498 drivers/net/ethernet/natsemi/sonic.c buf = (sonic_rra_get(dev, entry, SONIC_RR_BUFADR_H) << 16) | entry 499 drivers/net/ethernet/natsemi/sonic.c sonic_rra_get(dev, entry, SONIC_RR_BUFADR_L); entry 504 drivers/net/ethernet/natsemi/sonic.c entry = (entry + 1) & SONIC_RRS_MASK; entry 505 drivers/net/ethernet/natsemi/sonic.c } while (entry != end); entry 509 drivers/net/ethernet/natsemi/sonic.c sonic_rra_put(dev, entry, SONIC_RR_BUFADR_H, new_addr >> 16); entry 510 drivers/net/ethernet/natsemi/sonic.c sonic_rra_put(dev, entry, SONIC_RR_BUFADR_L, new_addr & 0xffff); entry 512 drivers/net/ethernet/natsemi/sonic.c entry = (entry + 1) & SONIC_RRS_MASK; entry 514 drivers/net/ethernet/natsemi/sonic.c SONIC_WRITE(SONIC_RWP, sonic_rr_addr(dev, entry)); entry 523 drivers/net/ethernet/natsemi/sonic.c int entry = lp->cur_rx; entry 527 drivers/net/ethernet/natsemi/sonic.c while (sonic_rda_get(dev, entry, SONIC_RD_IN_USE) == 0) { entry 528 drivers/net/ethernet/natsemi/sonic.c u16 status = sonic_rda_get(dev, entry, SONIC_RD_STATUS); entry 534 drivers/net/ethernet/natsemi/sonic.c u32 addr = (sonic_rda_get(dev, entry, entry 536 drivers/net/ethernet/natsemi/sonic.c sonic_rda_get(dev, entry, SONIC_RD_PKTPTR_L); entry 537 drivers/net/ethernet/natsemi/sonic.c int i = index_from_addr(lp, addr, entry); entry 552 drivers/net/ethernet/natsemi/sonic.c pkt_len = sonic_rda_get(dev, entry, entry 577 drivers/net/ethernet/natsemi/sonic.c sonic_rda_put(dev, entry, SONIC_RD_STATUS, 0); entry 578 drivers/net/ethernet/natsemi/sonic.c sonic_rda_put(dev, entry, SONIC_RD_IN_USE, 1); entry 580 drivers/net/ethernet/natsemi/sonic.c prev_entry = entry; entry 581 drivers/net/ethernet/natsemi/sonic.c entry = (entry + 1) & SONIC_RDS_MASK; entry 584 drivers/net/ethernet/natsemi/sonic.c lp->cur_rx = entry; entry 378 drivers/net/ethernet/natsemi/sonic.h static inline void sonic_cda_put(struct net_device* dev, int entry, entry 383 drivers/net/ethernet/natsemi/sonic.h (entry * SIZEOF_SONIC_CD) + offset, val); entry 386 drivers/net/ethernet/natsemi/sonic.h static inline __u16 sonic_cda_get(struct net_device* dev, int entry, entry 391 drivers/net/ethernet/natsemi/sonic.h (entry * SIZEOF_SONIC_CD) + offset); entry 406 drivers/net/ethernet/natsemi/sonic.h static inline void sonic_tda_put(struct net_device* dev, int entry, entry 411 drivers/net/ethernet/natsemi/sonic.h (entry * SIZEOF_SONIC_TD) + offset, val); entry 414 drivers/net/ethernet/natsemi/sonic.h static inline __u16 sonic_tda_get(struct net_device* dev, int entry, entry 419 drivers/net/ethernet/natsemi/sonic.h (entry * SIZEOF_SONIC_TD) + offset); entry 422 drivers/net/ethernet/natsemi/sonic.h static inline void sonic_rda_put(struct net_device* dev, int entry, entry 427 drivers/net/ethernet/natsemi/sonic.h (entry * SIZEOF_SONIC_RD) + offset, val); entry 430 drivers/net/ethernet/natsemi/sonic.h static inline __u16 sonic_rda_get(struct net_device* dev, int entry, entry 435 drivers/net/ethernet/natsemi/sonic.h (entry * SIZEOF_SONIC_RD) + offset); entry 438 drivers/net/ethernet/natsemi/sonic.h static inline void sonic_rra_put(struct net_device* dev, int entry, entry 443 drivers/net/ethernet/natsemi/sonic.h (entry * SIZEOF_SONIC_RR) + offset, val); entry 446 drivers/net/ethernet/natsemi/sonic.h static inline __u16 sonic_rra_get(struct net_device* dev, int entry, entry 451 drivers/net/ethernet/natsemi/sonic.h (entry * SIZEOF_SONIC_RR) + offset); entry 454 drivers/net/ethernet/natsemi/sonic.h static inline u16 sonic_rr_addr(struct net_device *dev, int entry) entry 459 drivers/net/ethernet/natsemi/sonic.h entry * SIZEOF_SONIC_RR * SONIC_BUS_SCALE(lp->dma_bitmode); entry 3749 drivers/net/ethernet/neterion/s2io.c nic->entries[0].entry = 0; entry 3750 drivers/net/ethernet/neterion/s2io.c nic->s2io_entries[0].entry = 0; entry 3756 drivers/net/ethernet/neterion/s2io.c nic->entries[i].entry = ((i - 1) * 8) + 1; entry 3757 drivers/net/ethernet/neterion/s2io.c nic->s2io_entries[i].entry = ((i - 1) * 8) + 1; entry 833 drivers/net/ethernet/neterion/s2io.h u16 entry; entry 1137 drivers/net/ethernet/neterion/vxge/vxge-config.c struct __vxge_hw_blockpool_entry *entry = NULL; entry 1159 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = kzalloc(sizeof(struct __vxge_hw_blockpool_entry), entry 1161 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry == NULL) { entry 1166 drivers/net/ethernet/neterion/vxge/vxge-config.c list_add(&entry->item, &blockpool->free_entry_list); entry 1192 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = (struct __vxge_hw_blockpool_entry *) entry 1197 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry == NULL) entry 1198 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = entry 1201 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry != NULL) { entry 1202 drivers/net/ethernet/neterion/vxge/vxge-config.c list_del(&entry->item); entry 1203 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->length = VXGE_HW_BLOCK_SIZE; entry 1204 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->memblock = memblock; entry 1205 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->dma_addr = dma_addr; entry 1206 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->acc_handle = acc_handle; entry 1207 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->dma_handle = dma_handle; entry 1208 drivers/net/ethernet/neterion/vxge/vxge-config.c list_add(&entry->item, entry 2257 drivers/net/ethernet/neterion/vxge/vxge-config.c struct __vxge_hw_blockpool_entry *entry = NULL; entry 2277 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = (struct __vxge_hw_blockpool_entry *) entry 2282 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry == NULL) entry 2283 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = vmalloc(sizeof(struct __vxge_hw_blockpool_entry)); entry 2285 drivers/net/ethernet/neterion/vxge/vxge-config.c list_del(&entry->item); entry 2287 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry) { entry 2288 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->length = length; entry 2289 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->memblock = block_addr; entry 2290 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->dma_addr = dma_addr; entry 2291 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->acc_handle = acc_handle; entry 2292 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->dma_handle = dma_h; entry 2293 drivers/net/ethernet/neterion/vxge/vxge-config.c list_add(&entry->item, &blockpool->free_block_list); entry 2347 drivers/net/ethernet/neterion/vxge/vxge-config.c struct __vxge_hw_blockpool_entry *entry = NULL; entry 2376 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = (struct __vxge_hw_blockpool_entry *) entry 2381 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry != NULL) { entry 2382 drivers/net/ethernet/neterion/vxge/vxge-config.c list_del(&entry->item); entry 2383 drivers/net/ethernet/neterion/vxge/vxge-config.c dma_object->addr = entry->dma_addr; entry 2384 drivers/net/ethernet/neterion/vxge/vxge-config.c dma_object->handle = entry->dma_handle; entry 2385 drivers/net/ethernet/neterion/vxge/vxge-config.c dma_object->acc_handle = entry->acc_handle; entry 2386 drivers/net/ethernet/neterion/vxge/vxge-config.c memblock = entry->memblock; entry 2388 drivers/net/ethernet/neterion/vxge/vxge-config.c list_add(&entry->item, entry 2441 drivers/net/ethernet/neterion/vxge/vxge-config.c struct __vxge_hw_blockpool_entry *entry = NULL; entry 2454 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = (struct __vxge_hw_blockpool_entry *) entry 2459 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry == NULL) entry 2460 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = vmalloc(sizeof( entry 2463 drivers/net/ethernet/neterion/vxge/vxge-config.c list_del(&entry->item); entry 2465 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry != NULL) { entry 2466 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->length = size; entry 2467 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->memblock = memblock; entry 2468 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->dma_addr = dma_object->addr; entry 2469 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->acc_handle = dma_object->acc_handle; entry 2470 drivers/net/ethernet/neterion/vxge/vxge-config.c entry->dma_handle = dma_object->handle; entry 2471 drivers/net/ethernet/neterion/vxge/vxge-config.c list_add(&entry->item, entry 4763 drivers/net/ethernet/neterion/vxge/vxge-config.c struct __vxge_hw_blockpool_entry *entry = NULL; entry 4771 drivers/net/ethernet/neterion/vxge/vxge-config.c entry = (struct __vxge_hw_blockpool_entry *) entry 4776 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry != NULL) { entry 4777 drivers/net/ethernet/neterion/vxge/vxge-config.c list_del(&entry->item); entry 4782 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry != NULL) entry 4785 drivers/net/ethernet/neterion/vxge/vxge-config.c return entry; entry 4944 drivers/net/ethernet/neterion/vxge/vxge-config.c struct __vxge_hw_blockpool_entry *entry) entry 4950 drivers/net/ethernet/neterion/vxge/vxge-config.c if (entry->length == blockpool->block_size) { entry 4951 drivers/net/ethernet/neterion/vxge/vxge-config.c list_add(&entry->item, &blockpool->free_block_list); entry 658 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; entry 659 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, &vpath->mac_addr_list) { entry 660 drivers/net/ethernet/neterion/vxge/vxge-main.c if (((struct vxge_mac_addrs *)entry)->macaddr == del_mac) entry 1068 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; entry 1075 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, &vpath->mac_addr_list) { entry 1076 drivers/net/ethernet/neterion/vxge/vxge-main.c if (((struct vxge_mac_addrs *)entry)->macaddr == del_mac) { entry 1077 drivers/net/ethernet/neterion/vxge/vxge-main.c list_del(entry); entry 1078 drivers/net/ethernet/neterion/vxge/vxge-main.c kfree((struct vxge_mac_addrs *)entry); entry 1131 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; entry 1195 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, list_head) { entry 1196 drivers/net/ethernet/neterion/vxge/vxge-main.c mac_entry = (struct vxge_mac_addrs *)entry; entry 1237 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, list_head) { entry 1238 drivers/net/ethernet/neterion/vxge/vxge-main.c mac_entry = (struct vxge_mac_addrs *)entry; entry 1446 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; entry 1451 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, &vpath->mac_addr_list) { entry 1454 drivers/net/ethernet/neterion/vxge/vxge-main.c ((struct vxge_mac_addrs *)entry)->macaddr; entry 1456 drivers/net/ethernet/neterion/vxge/vxge-main.c ((struct vxge_mac_addrs *)entry)->state = entry 1471 drivers/net/ethernet/neterion/vxge/vxge-main.c ((struct vxge_mac_addrs *)entry)->state entry 2325 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->entries[j].entry = msix_intr_vect; entry 2326 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->vxge_entries[j].entry = msix_intr_vect; entry 2331 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->entries[j].entry = msix_intr_vect + 1; entry 2332 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->vxge_entries[j].entry = msix_intr_vect + 1; entry 2338 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->entries[j].entry = VXGE_ALARM_MSIX_ID; entry 2339 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->vxge_entries[j].entry = VXGE_ALARM_MSIX_ID; entry 2474 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->entries[intr_cnt].entry, entry 2489 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->entries[intr_cnt].entry, entry 2535 drivers/net/ethernet/neterion/vxge/vxge-main.c vdev->entries[intr_cnt].entry, entry 2909 drivers/net/ethernet/neterion/vxge/vxge-main.c struct list_head *entry, *next; entry 2913 drivers/net/ethernet/neterion/vxge/vxge-main.c list_for_each_safe(entry, next, &vpath->mac_addr_list) { entry 2914 drivers/net/ethernet/neterion/vxge/vxge-main.c list_del(entry); entry 2915 drivers/net/ethernet/neterion/vxge/vxge-main.c kfree((struct vxge_mac_addrs *)entry); entry 4311 drivers/net/ethernet/neterion/vxge/vxge-main.c struct vxge_mac_addrs *entry; entry 4685 drivers/net/ethernet/neterion/vxge/vxge-main.c entry = kzalloc(sizeof(struct vxge_mac_addrs), GFP_KERNEL); entry 4686 drivers/net/ethernet/neterion/vxge/vxge-main.c if (NULL == entry) { entry 4693 drivers/net/ethernet/neterion/vxge/vxge-main.c macaddr = (u8 *)&entry->macaddr; entry 4695 drivers/net/ethernet/neterion/vxge/vxge-main.c list_add(&entry->item, &vdev->vpaths[i].mac_addr_list); entry 198 drivers/net/ethernet/neterion/vxge/vxge-main.h u16 entry; entry 145 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c struct nfp_fl_lag_group *entry; entry 150 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c list_for_each_entry(entry, &lag->group_list, list) entry 151 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c if (entry->master_ndev == master) entry 152 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c return entry; entry 267 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c struct nfp_fl_lag_group *entry, *storage; entry 278 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c list_for_each_entry_safe(entry, storage, &lag->group_list, list) { entry 285 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c if (entry->to_remove) { entry 287 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c err = nfp_fl_lag_config_group(lag, entry, NULL, 0, entry 290 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c entry->to_remove = false; entry 291 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c entry->offloaded = false; entry 300 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c if (entry->to_destroy) { entry 302 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c entry->group_id); entry 303 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c list_del(&entry->list); entry 304 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c kfree(entry); entry 309 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c acti_netdevs = kmalloc_array(entry->slave_cnt, entry 319 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c for_each_netdev_in_bond_rcu(entry->master_ndev, iter_netdev) { entry 333 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c if (slaves > entry->slave_cnt) entry 342 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c entry->dirty = true; entry 351 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c if (slaves != entry->slave_cnt || !entry->dirty) { entry 356 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c err = nfp_fl_lag_config_group(lag, entry, acti_netdevs, entry 359 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c entry->offloaded = true; entry 360 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c entry->dirty = false; entry 678 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c struct nfp_fl_lag_group *entry, *storage; entry 686 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c list_for_each_entry_safe(entry, storage, &lag->group_list, list) { entry 687 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c list_del(&entry->list); entry 688 drivers/net/ethernet/netronome/nfp/flower/lag_conf.c kfree(entry); entry 42 drivers/net/ethernet/netronome/nfp/flower/main.c struct net_device *entry; entry 46 drivers/net/ethernet/netronome/nfp/flower/main.c idr_for_each_entry(&priv->internal_ports.port_ids, entry, i) entry 47 drivers/net/ethernet/netronome/nfp/flower/main.c if (entry == netdev) { entry 149 drivers/net/ethernet/netronome/nfp/flower/main.c struct nfp_flower_non_repr_priv *entry; entry 153 drivers/net/ethernet/netronome/nfp/flower/main.c list_for_each_entry(entry, &priv->non_repr_priv, list) entry 154 drivers/net/ethernet/netronome/nfp/flower/main.c if (entry->netdev == netdev) entry 155 drivers/net/ethernet/netronome/nfp/flower/main.c return entry; entry 170 drivers/net/ethernet/netronome/nfp/flower/main.c struct nfp_flower_non_repr_priv *entry; entry 172 drivers/net/ethernet/netronome/nfp/flower/main.c entry = nfp_flower_non_repr_priv_lookup(app, netdev); entry 173 drivers/net/ethernet/netronome/nfp/flower/main.c if (entry) entry 176 drivers/net/ethernet/netronome/nfp/flower/main.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 177 drivers/net/ethernet/netronome/nfp/flower/main.c if (!entry) entry 180 drivers/net/ethernet/netronome/nfp/flower/main.c entry->netdev = netdev; entry 181 drivers/net/ethernet/netronome/nfp/flower/main.c list_add(&entry->list, &priv->non_repr_priv); entry 184 drivers/net/ethernet/netronome/nfp/flower/main.c __nfp_flower_non_repr_priv_get(entry); entry 185 drivers/net/ethernet/netronome/nfp/flower/main.c return entry; entry 201 drivers/net/ethernet/netronome/nfp/flower/main.c struct nfp_flower_non_repr_priv *entry; entry 203 drivers/net/ethernet/netronome/nfp/flower/main.c entry = nfp_flower_non_repr_priv_lookup(app, netdev); entry 204 drivers/net/ethernet/netronome/nfp/flower/main.c if (!entry) entry 207 drivers/net/ethernet/netronome/nfp/flower/main.c __nfp_flower_non_repr_priv_put(entry); entry 230 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c struct nfp_ipv4_route_entry *entry; entry 235 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = list_entry(ptr, struct nfp_ipv4_route_entry, list); entry 236 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry->ipv4_addr == ipv4_addr) { entry 248 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c struct nfp_ipv4_route_entry *entry; entry 253 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = list_entry(ptr, struct nfp_ipv4_route_entry, list); entry 254 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry->ipv4_addr == ipv4_addr) { entry 259 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 260 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (!entry) { entry 266 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->ipv4_addr = ipv4_addr; entry 267 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c list_add_tail(&entry->list, &priv->tun.neigh_off_list); entry 274 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c struct nfp_ipv4_route_entry *entry; entry 279 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = list_entry(ptr, struct nfp_ipv4_route_entry, list); entry 280 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry->ipv4_addr == ipv4_addr) { entry 281 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c list_del(&entry->list); entry 282 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c kfree(entry); entry 427 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c struct nfp_ipv4_addr_entry *entry; entry 441 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = list_entry(ptr, struct nfp_ipv4_addr_entry, list); entry 442 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c payload.ipv4_addr[count++] = entry->ipv4_addr; entry 455 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c struct nfp_ipv4_addr_entry *entry; entry 460 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = list_entry(ptr, struct nfp_ipv4_addr_entry, list); entry 461 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry->ipv4_addr == ipv4) { entry 462 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->ref_count++; entry 468 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 469 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (!entry) { entry 474 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->ipv4_addr = ipv4; entry 475 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->ref_count = 1; entry 476 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c list_add_tail(&entry->list, &priv->tun.ipv4_off_list); entry 485 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c struct nfp_ipv4_addr_entry *entry; entry 490 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = list_entry(ptr, struct nfp_ipv4_addr_entry, list); entry 491 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry->ipv4_addr == ipv4) { entry 492 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->ref_count--; entry 493 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (!entry->ref_count) { entry 494 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c list_del(&entry->list); entry 495 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c kfree(entry); entry 564 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c nfp_tunnel_offloaded_macs_inc_ref_and_link(struct nfp_tun_offloaded_mac *entry, entry 578 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c list_add_tail(&repr_priv->mac_list, &entry->repr_list); entry 580 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->bridge_count++; entry 583 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->ref_count++; entry 592 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c struct nfp_tun_offloaded_mac *entry; entry 595 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = nfp_tunnel_lookup_offloaded_macs(app, netdev->dev_addr); entry 596 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry && nfp_tunnel_is_mac_idx_global(entry->index)) { entry 597 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry->bridge_count || entry 599 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c nfp_tunnel_offloaded_macs_inc_ref_and_link(entry, entry 605 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c nfp_mac_idx = entry->index | NFP_TUN_PRE_TUN_IDX_BIT; entry 610 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry || !port) { entry 628 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (!entry) { entry 629 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 630 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (!entry) { entry 635 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c ether_addr_copy(entry->addr, netdev->dev_addr); entry 636 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c INIT_LIST_HEAD(&entry->repr_list); entry 639 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c &entry->ht_node, entry 650 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (!entry->ref_count) entry 655 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->index = nfp_mac_idx; entry 656 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c nfp_tunnel_offloaded_macs_inc_ref_and_link(entry, netdev, mod); entry 661 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c rhashtable_remove_fast(&priv->tun.offloaded_macs, &entry->ht_node, entry 664 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c kfree(entry); entry 678 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c struct nfp_tun_offloaded_mac *entry; entry 682 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry = nfp_tunnel_lookup_offloaded_macs(app, mac); entry 683 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (!entry) entry 686 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->ref_count--; entry 695 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->bridge_count--; entry 697 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (!entry->bridge_count && entry->ref_count) { entry 700 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c nfp_mac_idx = entry->index & ~NFP_TUN_PRE_TUN_IDX_BIT; entry 708 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->index = nfp_mac_idx; entry 714 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry->ref_count == 1 && list_is_singular(&entry->repr_list)) { entry 718 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c repr_priv = list_first_entry(&entry->repr_list, entry 731 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c ida_idx = nfp_tunnel_get_ida_from_global_mac_idx(entry->index); entry 733 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c entry->index = nfp_mac_idx; entry 737 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (entry->ref_count) entry 741 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c &entry->ht_node, entry 744 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c if (nfp_tunnel_is_mac_idx_global(entry->index)) { entry 745 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c ida_idx = nfp_tunnel_get_ida_from_global_mac_idx(entry->index); entry 749 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c kfree(entry); entry 406 drivers/net/ethernet/netronome/nfp/nfp_net_common.c irq_entries[i].entry = i; entry 537 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct msix_entry *entry; entry 539 drivers/net/ethernet/netronome/nfp/nfp_net_common.c entry = &nn->irq_entries[NFP_NET_IRQ_LSC_IDX]; entry 543 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_net_irq_unmask(nn, entry->entry); entry 622 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct msix_entry *entry; entry 625 drivers/net/ethernet/netronome/nfp/nfp_net_common.c entry = &nn->irq_entries[vector_idx]; entry 628 drivers/net/ethernet/netronome/nfp/nfp_net_common.c err = request_irq(entry->vector, handler, 0, name, nn); entry 631 drivers/net/ethernet/netronome/nfp/nfp_net_common.c entry->vector, err); entry 634 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, ctrl_offset, entry->entry); entry 635 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_net_irq_unmask(nn, entry->entry); entry 2318 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct msix_entry *entry; entry 2320 drivers/net/ethernet/netronome/nfp/nfp_net_common.c entry = &nn->irq_entries[NFP_NET_NON_Q_VECTORS + r]; entry 2324 drivers/net/ethernet/netronome/nfp/nfp_net_common.c r_vec->irq_entry = entry->entry; entry 2325 drivers/net/ethernet/netronome/nfp/nfp_net_common.c r_vec->irq_vector = entry->vector; entry 82 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c struct list_head entry; entry 167 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c list_for_each_entry_safe(cache, ctmp, &cpp->area_cache_list, entry) { entry 168 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c list_del(&cache->entry); entry 814 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c list_add_tail(&cache->entry, &cpp->area_cache_list); entry 849 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c list_for_each_entry(cache, &cpp->area_cache_list, entry) { entry 858 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c struct nfp_cpp_area_cache, entry); entry 908 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c list_del(&cache->entry); entry 909 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c list_add(&cache->entry, &cpp->area_cache_list); entry 204 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c nfp_eth_calc_port_type(struct nfp_cpp *cpp, struct nfp_eth_table_port *entry) entry 206 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c if (entry->interface == NFP_INTERFACE_NONE) { entry 207 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c entry->port_type = PORT_NONE; entry 209 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c } else if (entry->interface == NFP_INTERFACE_RJ45) { entry 210 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c entry->port_type = PORT_TP; entry 214 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c if (entry->media == NFP_MEDIA_FIBRE) entry 215 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c entry->port_type = PORT_FIBRE; entry 217 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c entry->port_type = PORT_DA; entry 71 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c struct nfp_resource_entry entry; entry 88 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c ret = nfp_cpp_read(cpp, cpp_id, addr, &entry, sizeof(entry)); entry 89 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c if (ret != sizeof(entry)) entry 92 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c if (entry.mutex.key != key) entry 99 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c res->cpp_id = NFP_CPP_ID(entry.region.cpp_target, entry 100 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c entry.region.cpp_action, entry 101 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c entry.region.cpp_token); entry 102 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c res->addr = (u64)entry.region.page_offset << 8; entry 103 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c res->size = (u64)entry.region.page_size << 8; entry 3996 drivers/net/ethernet/nvidia/forcedeth.c np->msi_x_entry[i].entry = i; entry 996 drivers/net/ethernet/packetengines/hamachi.c int entry = hmp->dirty_tx % TX_RING_SIZE; entry 999 drivers/net/ethernet/packetengines/hamachi.c if (hmp->tx_ring[entry].status_n_length & cpu_to_le32(DescOwn)) entry 1002 drivers/net/ethernet/packetengines/hamachi.c skb = hmp->tx_skbuff[entry]; entry 1005 drivers/net/ethernet/packetengines/hamachi.c leXX_to_cpu(hmp->tx_ring[entry].addr), entry 1008 drivers/net/ethernet/packetengines/hamachi.c hmp->tx_skbuff[entry] = NULL; entry 1010 drivers/net/ethernet/packetengines/hamachi.c hmp->tx_ring[entry].status_n_length = 0; entry 1011 drivers/net/ethernet/packetengines/hamachi.c if (entry >= TX_RING_SIZE-1) entry 1208 drivers/net/ethernet/packetengines/hamachi.c unsigned entry; entry 1232 drivers/net/ethernet/packetengines/hamachi.c entry = hmp->cur_tx % TX_RING_SIZE; entry 1234 drivers/net/ethernet/packetengines/hamachi.c hmp->tx_skbuff[entry] = skb; entry 1236 drivers/net/ethernet/packetengines/hamachi.c hmp->tx_ring[entry].addr = cpu_to_leXX(pci_map_single(hmp->pci_dev, entry 1248 drivers/net/ethernet/packetengines/hamachi.c if (entry >= TX_RING_SIZE-1) /* Wrap ring */ entry 1249 drivers/net/ethernet/packetengines/hamachi.c hmp->tx_ring[entry].status_n_length = cpu_to_le32(DescOwn | entry 1252 drivers/net/ethernet/packetengines/hamachi.c hmp->tx_ring[entry].status_n_length = cpu_to_le32(DescOwn | entry 1282 drivers/net/ethernet/packetengines/hamachi.c dev->name, hmp->cur_tx, entry); entry 1328 drivers/net/ethernet/packetengines/hamachi.c int entry = hmp->dirty_tx % TX_RING_SIZE; entry 1331 drivers/net/ethernet/packetengines/hamachi.c if (hmp->tx_ring[entry].status_n_length & cpu_to_le32(DescOwn)) entry 1333 drivers/net/ethernet/packetengines/hamachi.c skb = hmp->tx_skbuff[entry]; entry 1337 drivers/net/ethernet/packetengines/hamachi.c leXX_to_cpu(hmp->tx_ring[entry].addr), entry 1341 drivers/net/ethernet/packetengines/hamachi.c hmp->tx_skbuff[entry] = NULL; entry 1343 drivers/net/ethernet/packetengines/hamachi.c hmp->tx_ring[entry].status_n_length = 0; entry 1344 drivers/net/ethernet/packetengines/hamachi.c if (entry >= TX_RING_SIZE-1) entry 1398 drivers/net/ethernet/packetengines/hamachi.c int entry = hmp->cur_rx % RX_RING_SIZE; entry 1403 drivers/net/ethernet/packetengines/hamachi.c entry, hmp->rx_ring[entry].status_n_length); entry 1408 drivers/net/ethernet/packetengines/hamachi.c struct hamachi_desc *desc = &(hmp->rx_ring[entry]); entry 1420 drivers/net/ethernet/packetengines/hamachi.c buf_addr = (u8 *) hmp->rx_skbuff[entry]->data; entry 1487 drivers/net/ethernet/packetengines/hamachi.c leXX_to_cpu(hmp->rx_ring[entry].addr), entry 1493 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_skbuff[entry]->data, pkt_len); entry 1497 drivers/net/ethernet/packetengines/hamachi.c + entry*sizeof(*desc), pkt_len); entry 1500 drivers/net/ethernet/packetengines/hamachi.c leXX_to_cpu(hmp->rx_ring[entry].addr), entry 1505 drivers/net/ethernet/packetengines/hamachi.c leXX_to_cpu(hmp->rx_ring[entry].addr), entry 1507 drivers/net/ethernet/packetengines/hamachi.c skb_put(skb = hmp->rx_skbuff[entry], pkt_len); entry 1508 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_skbuff[entry] = NULL; entry 1573 drivers/net/ethernet/packetengines/hamachi.c entry = (++hmp->cur_rx) % RX_RING_SIZE; entry 1580 drivers/net/ethernet/packetengines/hamachi.c entry = hmp->dirty_rx % RX_RING_SIZE; entry 1581 drivers/net/ethernet/packetengines/hamachi.c desc = &(hmp->rx_ring[entry]); entry 1582 drivers/net/ethernet/packetengines/hamachi.c if (hmp->rx_skbuff[entry] == NULL) { entry 1585 drivers/net/ethernet/packetengines/hamachi.c hmp->rx_skbuff[entry] = skb; entry 1593 drivers/net/ethernet/packetengines/hamachi.c if (entry >= RX_RING_SIZE-1) entry 808 drivers/net/ethernet/packetengines/yellowfin.c unsigned entry; entry 817 drivers/net/ethernet/packetengines/yellowfin.c entry = yp->cur_tx % TX_RING_SIZE; entry 825 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_skbuff[entry] = NULL; entry 831 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_skbuff[entry] = skb; entry 834 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_ring[entry].addr = cpu_to_le32(pci_map_single(yp->pci_dev, entry 836 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_ring[entry].result_status = 0; entry 837 drivers/net/ethernet/packetengines/yellowfin.c if (entry >= TX_RING_SIZE-1) { entry 843 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_ring[entry+1].dbdma_cmd = cpu_to_le32(CMD_STOP); entry 844 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_ring[entry].dbdma_cmd = entry 849 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_ring[entry<<1].request_cnt = len; entry 850 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_ring[entry<<1].addr = cpu_to_le32(pci_map_single(yp->pci_dev, entry 862 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_ring[entry<<1].dbdma_cmd = entry 863 drivers/net/ethernet/packetengines/yellowfin.c cpu_to_le32( ((entry % 6) == 0 ? CMD_TX_PKT|INTR_ALWAYS|BRANCH_IFTRUE : entry 879 drivers/net/ethernet/packetengines/yellowfin.c yp->cur_tx, entry); entry 917 drivers/net/ethernet/packetengines/yellowfin.c int entry = yp->dirty_tx % TX_RING_SIZE; entry 920 drivers/net/ethernet/packetengines/yellowfin.c if (yp->tx_ring[entry].result_status == 0) entry 922 drivers/net/ethernet/packetengines/yellowfin.c skb = yp->tx_skbuff[entry]; entry 926 drivers/net/ethernet/packetengines/yellowfin.c pci_unmap_single(yp->pci_dev, le32_to_cpu(yp->tx_ring[entry].addr), entry 929 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_skbuff[entry] = NULL; entry 944 drivers/net/ethernet/packetengines/yellowfin.c int entry = dirty_tx % TX_RING_SIZE; entry 945 drivers/net/ethernet/packetengines/yellowfin.c u16 tx_errs = yp->tx_status[entry].tx_errs; entry 951 drivers/net/ethernet/packetengines/yellowfin.c entry, entry 952 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_status[entry].tx_cnt, entry 953 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_status[entry].tx_errs, entry 954 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_status[entry].total_tx_cnt, entry 955 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_status[entry].paused); entry 959 drivers/net/ethernet/packetengines/yellowfin.c skb = yp->tx_skbuff[entry]; entry 984 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_ring[entry<<1].addr, skb->len, entry 987 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_skbuff[entry] = 0; entry 989 drivers/net/ethernet/packetengines/yellowfin.c yp->tx_status[entry].tx_errs = 0; entry 1036 drivers/net/ethernet/packetengines/yellowfin.c int entry = yp->cur_rx % RX_RING_SIZE; entry 1041 drivers/net/ethernet/packetengines/yellowfin.c entry, yp->rx_ring[entry].result_status); entry 1043 drivers/net/ethernet/packetengines/yellowfin.c entry, yp->rx_ring[entry].dbdma_cmd, yp->rx_ring[entry].addr, entry 1044 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_ring[entry].result_status); entry 1049 drivers/net/ethernet/packetengines/yellowfin.c struct yellowfin_desc *desc = &yp->rx_ring[entry]; entry 1050 drivers/net/ethernet/packetengines/yellowfin.c struct sk_buff *rx_skb = yp->rx_skbuff[entry]; entry 1100 drivers/net/ethernet/packetengines/yellowfin.c entry * yf_size), entry 1103 drivers/net/ethernet/packetengines/yellowfin.c entry * yf_size), entry 1125 drivers/net/ethernet/packetengines/yellowfin.c le32_to_cpu(yp->rx_ring[entry].addr), entry 1128 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_skbuff[entry] = NULL; entry 1146 drivers/net/ethernet/packetengines/yellowfin.c entry = (++yp->cur_rx) % RX_RING_SIZE; entry 1151 drivers/net/ethernet/packetengines/yellowfin.c entry = yp->dirty_rx % RX_RING_SIZE; entry 1152 drivers/net/ethernet/packetengines/yellowfin.c if (yp->rx_skbuff[entry] == NULL) { entry 1156 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_skbuff[entry] = skb; entry 1158 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_ring[entry].addr = cpu_to_le32(pci_map_single(yp->pci_dev, entry 1161 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_ring[entry].dbdma_cmd = cpu_to_le32(CMD_STOP); entry 1162 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_ring[entry].result_status = 0; /* Clear complete bit. */ entry 1163 drivers/net/ethernet/packetengines/yellowfin.c if (entry != 0) entry 1164 drivers/net/ethernet/packetengines/yellowfin.c yp->rx_ring[entry - 1].dbdma_cmd = entry 2312 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c struct netxen_minidump_entry *entry, int esize) entry 2315 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.driver_flags |= NX_DUMP_SKIP; entry 2318 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c if (esize != entry->hdr.entry_capture_size) { entry 2319 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.entry_capture_size = esize; entry 2320 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.driver_flags |= NX_DUMP_SIZE_ERR; entry 2323 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.entry_type, entry->hdr.entry_capture_mask, entry 2324 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c esize, entry->hdr.entry_capture_size); entry 2339 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c struct netxen_minidump_entry *entry; entry 2348 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry = (struct netxen_minidump_entry *) ((char *) template_buff + entry 2357 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c if (!(entry->hdr.entry_capture_mask & capture_mask)) { entry 2358 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.driver_flags |= NX_DUMP_SKIP; entry 2359 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry = (struct netxen_minidump_entry *) entry 2360 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c ((char *) entry + entry->hdr.entry_size); entry 2363 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c switch (entry->hdr.entry_type) { entry 2365 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.driver_flags |= NX_DUMP_SKIP; entry 2368 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.driver_flags |= NX_DUMP_SKIP; entry 2373 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c template_hdr, (void *)entry); entry 2375 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.driver_flags |= NX_DUMP_SKIP; entry 2380 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (void *) entry, (void *) dbuff); entry 2382 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (adapter, entry, esize); entry 2391 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (void *) entry, (void *) dbuff); entry 2393 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (adapter, entry, esize); entry 2402 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (void *) entry, (void *) dbuff); entry 2404 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (adapter, entry, esize); entry 2415 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (void *) entry, (void *) dbuff); entry 2417 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (adapter, entry, esize); entry 2426 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (void *) entry, (void *) dbuff); entry 2428 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (adapter, entry, esize); entry 2436 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (void *) entry, (void *) dbuff); entry 2438 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (adapter, entry, esize); entry 2446 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (void *) entry, (void *) dbuff); entry 2448 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (adapter, entry, esize); entry 2456 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (void *) entry, (void *) dbuff); entry 2458 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c (adapter, entry, esize); entry 2464 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry->hdr.driver_flags |= NX_DUMP_SKIP; entry 2468 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c entry = (struct netxen_minidump_entry *) entry 2469 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c ((char *) entry + entry->hdr.entry_size); entry 455 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c adapter->msix_entries[i].entry = i; entry 294 drivers/net/ethernet/qlogic/qed/qed_dcbx.c protocol_id = QED_MFW_GET_FIELD(p_tbl[i].entry, entry 296 drivers/net/ethernet/qlogic/qed/qed_dcbx.c priority_map = QED_MFW_GET_FIELD(p_tbl[i].entry, entry 305 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if (qed_dcbx_get_app_protocol_type(p_hwfn, p_tbl[i].entry, entry 486 drivers/net/ethernet/qlogic/qed/qed_dcbx.c struct qed_app_entry *entry; entry 497 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry = &p_params->app_entry[i]; entry 502 drivers/net/ethernet/qlogic/qed/qed_dcbx.c sf_ieee = QED_MFW_GET_FIELD(p_tbl[i].entry, entry 507 drivers/net/ethernet/qlogic/qed/qed_dcbx.c val = QED_MFW_GET_FIELD(p_tbl[i].entry, entry 509 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->sf_ieee = val ? entry 514 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->sf_ieee = QED_DCBX_SF_IEEE_ETHTYPE; entry 517 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->sf_ieee = QED_DCBX_SF_IEEE_TCP_PORT; entry 520 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->sf_ieee = QED_DCBX_SF_IEEE_UDP_PORT; entry 523 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->sf_ieee = QED_DCBX_SF_IEEE_TCP_UDP_PORT; entry 527 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->ethtype = !(QED_MFW_GET_FIELD(p_tbl[i].entry, entry 531 drivers/net/ethernet/qlogic/qed/qed_dcbx.c pri_map = QED_MFW_GET_FIELD(p_tbl[i].entry, DCBX_APP_PRI_MAP); entry 532 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->prio = ffs(pri_map) - 1; entry 533 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->proto_id = QED_MFW_GET_FIELD(p_tbl[i].entry, entry 535 drivers/net/ethernet/qlogic/qed/qed_dcbx.c qed_dcbx_get_app_protocol_type(p_hwfn, p_tbl[i].entry, entry 536 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry->proto_id, entry 537 drivers/net/ethernet/qlogic/qed/qed_dcbx.c &entry->proto_type, ieee); entry 1128 drivers/net/ethernet/qlogic/qed/qed_dcbx.c u32 *entry; entry 1146 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry = &p_app->app_pri_tbl[i].entry; entry 1147 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry = 0; entry 1149 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry &= ~(DCBX_APP_SF_IEEE_MASK | DCBX_APP_SF_MASK); entry 1152 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_IEEE_ETHTYPE << entry 1154 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_ETHTYPE << entry 1158 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_IEEE_TCP_PORT << entry 1160 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_PORT << entry 1164 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_IEEE_UDP_PORT << entry 1166 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_PORT << entry 1170 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_IEEE_TCP_UDP_PORT << entry 1172 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_PORT << entry 1177 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry &= ~DCBX_APP_SF_MASK; entry 1179 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_ETHTYPE << entry 1182 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)DCBX_APP_SF_PORT << entry 1186 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry &= ~DCBX_APP_PROTOCOL_ID_MASK; entry 1187 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)p_params->app_entry[i].proto_id << entry 1189 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry &= ~DCBX_APP_PRI_MAP_MASK; entry 1190 drivers/net/ethernet/qlogic/qed/qed_dcbx.c *entry |= ((u32)(p_params->app_entry[i].prio) << entry 1756 drivers/net/ethernet/qlogic/qed/qed_dcbx.c struct qed_app_entry *entry; entry 1767 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry = &dcbx_info->operational.params.app_entry[i]; entry 1768 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if ((entry->ethtype == ethtype) && (entry->proto_id == idval)) { entry 1769 drivers/net/ethernet/qlogic/qed/qed_dcbx.c prio = entry->prio; entry 1790 drivers/net/ethernet/qlogic/qed/qed_dcbx.c struct qed_app_entry *entry; entry 1802 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry = &dcbx_set.config.params.app_entry[i]; entry 1803 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if ((entry->ethtype == ethtype) && (entry->proto_id == idval)) entry 1806 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if (!entry->proto_id) { entry 2282 drivers/net/ethernet/qlogic/qed/qed_dcbx.c struct qed_app_entry *entry; entry 2307 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry = &dcbx_info->operational.params.app_entry[i]; entry 2308 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if ((entry->sf_ieee == sf_ieee) && entry 2309 drivers/net/ethernet/qlogic/qed/qed_dcbx.c (entry->proto_id == app->protocol)) { entry 2310 drivers/net/ethernet/qlogic/qed/qed_dcbx.c prio = entry->prio; entry 2334 drivers/net/ethernet/qlogic/qed/qed_dcbx.c struct qed_app_entry *entry; entry 2370 drivers/net/ethernet/qlogic/qed/qed_dcbx.c entry = &dcbx_set.config.params.app_entry[i]; entry 2371 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if ((entry->sf_ieee == sf_ieee) && entry 2372 drivers/net/ethernet/qlogic/qed/qed_dcbx.c (entry->proto_id == app->protocol)) entry 2375 drivers/net/ethernet/qlogic/qed/qed_dcbx.c if (!entry->proto_id) { entry 2530 drivers/net/ethernet/qlogic/qed/qed_dev.c p_block = &p_igu_info->entry[igu_sb_id]; entry 1005 drivers/net/ethernet/qlogic/qed/qed_hsi.h struct event_ring_entry entry; entry 11999 drivers/net/ethernet/qlogic/qed/qed_hsi.h u32 entry; entry 1580 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_hwfn->hw_info.p_igu_info->entry[igu_id]; entry 1600 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_hwfn->hw_info.p_igu_info->entry[igu_id]; entry 1651 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_info->entry[sb_info->igu_sb_id]; entry 1703 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_info->entry[sb_info->igu_sb_id]; entry 1936 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_hwfn->hw_info.p_igu_info->entry[igu_sb_id]; entry 1989 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_info->entry[igu_sb_id]; entry 2076 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_info->entry[igu_sb_id]; entry 2149 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_hwfn->hw_info.p_igu_info->entry[igu_sb_id]; entry 2186 drivers/net/ethernet/qlogic/qed/qed_int.c p_block = &p_igu_info->entry[igu_sb_id]; entry 232 drivers/net/ethernet/qlogic/qed/qed_int.h struct qed_igu_block entry[MAX_TOT_SB_PER_PATH]; entry 538 drivers/net/ethernet/qlogic/qed/qed_main.c int_params->msix_table[i].entry = i; entry 949 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_info->entry[igu_id].status |= QED_IGU_STATUS_FREE; entry 163 drivers/net/ethernet/qlogic/qede/qede.h struct list_head entry; entry 125 drivers/net/ethernet/qlogic/qede/qede_rdma.c INIT_LIST_HEAD(&edev->rdma_info.entry); entry 127 drivers/net/ethernet/qlogic/qede/qede_rdma.c list_add_tail(&edev->rdma_info.entry, &qedr_dev_list); entry 152 drivers/net/ethernet/qlogic/qede/qede_rdma.c list_del(&edev->rdma_info.entry); entry 219 drivers/net/ethernet/qlogic/qede/qede_rdma.c list_for_each_entry(edev, &qedr_dev_list, rdma_info.entry) { entry 242 drivers/net/ethernet/qlogic/qede/qede_rdma.c list_for_each_entry(edev, &qedr_dev_list, rdma_info.entry) { entry 1798 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c struct qlc_83xx_entry *entry; entry 1800 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry = (struct qlc_83xx_entry *)((char *)p_hdr + entry 1803 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c for (i = 0; i < p_hdr->count; i++, entry++) { entry 1804 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c qlcnic_83xx_wrt_reg_indirect(p_dev, entry->arg1, entry 1805 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry->arg2); entry 1816 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c struct qlc_83xx_entry *entry; entry 1818 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry = (struct qlc_83xx_entry *)((char *)p_hdr + entry 1821 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c for (i = 0; i < p_hdr->count; i++, entry++) { entry 1822 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c qlcnic_83xx_read_write_crb_reg(p_dev, entry->arg1, entry 1823 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry->arg2); entry 1834 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c struct qlc_83xx_entry *entry; entry 1842 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry = (struct qlc_83xx_entry *)((char *)poll + entry 1847 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c for (i = 0; i < p_hdr->count; i++, entry++) entry 1848 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c qlcnic_83xx_poll_reg(p_dev, entry->arg1, entry 1852 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c for (i = 0; i < p_hdr->count; i++, entry++) { entry 1853 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c arg1 = entry->arg1; entry 1854 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c arg2 = entry->arg2; entry 1878 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c struct qlc_83xx_quad_entry *entry; entry 1883 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry = (struct qlc_83xx_quad_entry *)((char *)poll + entry 1887 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c for (i = 0; i < p_hdr->count; i++, entry++) { entry 1888 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c qlcnic_83xx_wrt_reg_indirect(p_dev, entry->dr_addr, entry 1889 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry->dr_value); entry 1890 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c qlcnic_83xx_wrt_reg_indirect(p_dev, entry->ar_addr, entry 1891 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry->ar_value); entry 1893 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c qlcnic_83xx_poll_reg(p_dev, entry->ar_addr, delay, entry 1903 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c struct qlc_83xx_entry *entry; entry 1909 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry = (struct qlc_83xx_entry *)((char *)rmw_hdr + entry 1912 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c for (i = 0; i < p_hdr->count; i++, entry++) { entry 1913 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c qlcnic_83xx_rmw_crb_reg(p_dev, entry->arg1, entry 1914 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry->arg2, rmw_hdr); entry 1932 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c struct qlc_83xx_quad_entry *entry; entry 1939 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry = (struct qlc_83xx_quad_entry *)((char *)poll + entry 1943 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c for (i = 0; i < p_hdr->count; i++, entry++) { entry 1944 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c qlcnic_83xx_wrt_reg_indirect(p_dev, entry->ar_addr, entry 1945 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry->ar_value); entry 1947 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c if (!qlcnic_83xx_poll_reg(p_dev, entry->ar_addr, delay, entry 1950 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c addr = entry->dr_addr; entry 1999 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c char *entry = p_buff; entry 2007 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c p_hdr = (struct qlc_83xx_entry_hdr *)entry; entry 2045 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c entry += p_hdr->size; entry 715 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c adapter->msix_entries[vector].entry = vector; entry 774 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c adapter->msix_entries[vector].entry = vector; entry 386 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 390 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __crb *crb = &entry->region.crb; entry 404 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 407 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __ctrl *ctr = &entry->region.ctrl; entry 502 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 506 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __mux *mux = &entry->region.mux; entry 520 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 524 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __queue *que = &entry->region.que; entry 543 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 548 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __ocm *ocm = &entry->region.ocm; entry 560 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 564 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __mem *rom = &entry->region.mem; entry 590 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 594 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __cache *l1 = &entry->region.cache; entry 615 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 620 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __cache *l2 = &entry->region.cache; entry 856 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 860 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __mem *mem = &entry->region.mem; entry 870 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.mask); entry 879 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.mask); entry 887 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 889 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.flags |= QLCNIC_DUMP_SKIP; entry 894 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, u32 size) entry 897 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c if (size != entry->hdr.cap_size) { entry 900 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.type, entry->hdr.mask, size, entry 901 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.cap_size); entry 908 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, entry 911 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __pollrdmwr *poll = &entry->region.pollrdmwr; entry 953 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 955 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __pollrd *pollrd = &entry->region.pollrd; entry 988 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 990 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __mux2 *mux2 = &entry->region.mux2; entry 1019 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry, __le32 *buffer) entry 1022 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct __mem *rom = &entry->region.mem; entry 1295 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c struct qlcnic_dump_entry *entry; entry 1353 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry = tmpl_hdr + entry_offset; entry 1354 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c if (!(entry->hdr.mask & fw_dump->cap_mask)) { entry 1355 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.flags |= QLCNIC_DUMP_SKIP; entry 1356 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry_offset += entry->hdr.offset; entry 1363 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c if (entry->hdr.type == fw_dump_ops[ops_index].opcode) entry 1370 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.type); entry 1371 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.flags |= QLCNIC_DUMP_SKIP; entry 1372 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry_offset += entry->hdr.offset; entry 1377 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c dump = fw_dump_ops[ops_index].handler(adapter, entry, buffer); entry 1378 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c if (!qlcnic_valid_dump_entry(dev, entry, dump)) { entry 1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry->hdr.flags |= QLCNIC_DUMP_SKIP; entry 1380 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry_offset += entry->hdr.offset; entry 1384 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c buf_offset += entry->hdr.cap_size; entry 1385 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c entry_offset += entry->hdr.offset; entry 1525 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_async_cmd *entry; entry 1532 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c entry = list_entry(head->next, struct qlcnic_async_cmd, entry 1534 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c list_del(&entry->list); entry 1535 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(entry->cmd); entry 1536 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(entry); entry 1598 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_async_cmd *entry, *tmp; entry 1611 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c list_for_each_entry_safe(entry, tmp, &del_list, list) { entry 1612 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c list_del(&entry->list); entry 1613 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c cmd = entry->cmd; entry 1615 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(entry); entry 1628 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_async_cmd *entry = NULL; entry 1630 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 1631 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (!entry) entry 1634 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c entry->cmd = cmd; entry 1637 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c list_add_tail(&entry->list, &bc->async_cmd_list); entry 1640 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c return entry; entry 1646 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_async_cmd *entry = NULL; entry 1648 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c entry = qlcnic_sriov_alloc_async_cmd(bc, cmd); entry 1649 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c if (!entry) { entry 53 drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c int rc, entry; entry 71 drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c for (entry = 0; entry < RMNET_MAX_LOGICAL_EP; entry++) entry 72 drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c INIT_HLIST_HEAD(&port->muxed_ep[entry]); entry 735 drivers/net/ethernet/realtek/8139cp.c unsigned entry; entry 751 drivers/net/ethernet/realtek/8139cp.c entry = cp->tx_head; entry 752 drivers/net/ethernet/realtek/8139cp.c eor = (entry == (CP_TX_RING_SIZE - 1)) ? RingEnd : 0; entry 779 drivers/net/ethernet/realtek/8139cp.c struct cp_desc *txd = &cp->tx_ring[entry]; entry 797 drivers/net/ethernet/realtek/8139cp.c cp->tx_skb[entry] = skb; entry 798 drivers/net/ethernet/realtek/8139cp.c cp->tx_opts[entry] = opts1; entry 800 drivers/net/ethernet/realtek/8139cp.c entry, skb->len); entry 805 drivers/net/ethernet/realtek/8139cp.c int frag, first_entry = entry; entry 817 drivers/net/ethernet/realtek/8139cp.c cp->tx_skb[entry] = skb; entry 824 drivers/net/ethernet/realtek/8139cp.c entry = NEXT_TX(entry); entry 831 drivers/net/ethernet/realtek/8139cp.c unwind_tx_frag_mapping(cp, skb, first_entry, entry); entry 835 drivers/net/ethernet/realtek/8139cp.c eor = (entry == (CP_TX_RING_SIZE - 1)) ? RingEnd : 0; entry 842 drivers/net/ethernet/realtek/8139cp.c txd = &cp->tx_ring[entry]; entry 850 drivers/net/ethernet/realtek/8139cp.c cp->tx_opts[entry] = ctrl; entry 851 drivers/net/ethernet/realtek/8139cp.c cp->tx_skb[entry] = skb; entry 865 drivers/net/ethernet/realtek/8139cp.c first_entry, entry, skb->len); entry 867 drivers/net/ethernet/realtek/8139cp.c cp->tx_head = NEXT_TX(entry); entry 1719 drivers/net/ethernet/realtek/8139too.c unsigned int entry; entry 1724 drivers/net/ethernet/realtek/8139too.c entry = tp->cur_tx % NUM_TX_DESC; entry 1729 drivers/net/ethernet/realtek/8139too.c memset(tp->tx_buf[entry], 0, ETH_ZLEN); entry 1730 drivers/net/ethernet/realtek/8139too.c skb_copy_and_csum_dev(skb, tp->tx_buf[entry]); entry 1745 drivers/net/ethernet/realtek/8139too.c RTL_W32_F (TxStatus0 + (entry * sizeof (u32)), entry 1755 drivers/net/ethernet/realtek/8139too.c len, entry); entry 1773 drivers/net/ethernet/realtek/8139too.c int entry = dirty_tx % NUM_TX_DESC; entry 1776 drivers/net/ethernet/realtek/8139too.c txstatus = RTL_R32 (TxStatus0 + (entry * sizeof (u32))); entry 5633 drivers/net/ethernet/realtek/r8169_main.c unsigned int entry = (start + i) % NUM_TX_DESC; entry 5634 drivers/net/ethernet/realtek/r8169_main.c struct ring_info *tx_skb = tp->tx_skb + entry; entry 5641 drivers/net/ethernet/realtek/r8169_main.c tp->TxDescArray + entry); entry 5686 drivers/net/ethernet/realtek/r8169_main.c static __le32 rtl8169_get_txd_opts1(u32 opts0, u32 len, unsigned int entry) entry 5690 drivers/net/ethernet/realtek/r8169_main.c if (entry == NUM_TX_DESC - 1) entry 5700 drivers/net/ethernet/realtek/r8169_main.c unsigned int cur_frag, entry; entry 5704 drivers/net/ethernet/realtek/r8169_main.c entry = tp->cur_tx; entry 5711 drivers/net/ethernet/realtek/r8169_main.c entry = (entry + 1) % NUM_TX_DESC; entry 5713 drivers/net/ethernet/realtek/r8169_main.c txd = tp->TxDescArray + entry; entry 5724 drivers/net/ethernet/realtek/r8169_main.c txd->opts1 = rtl8169_get_txd_opts1(opts[0], len, entry); entry 5728 drivers/net/ethernet/realtek/r8169_main.c tp->tx_skb[entry].len = len; entry 5732 drivers/net/ethernet/realtek/r8169_main.c tp->tx_skb[entry].skb = skb; entry 5884 drivers/net/ethernet/realtek/r8169_main.c unsigned int entry = tp->cur_tx % NUM_TX_DESC; entry 5885 drivers/net/ethernet/realtek/r8169_main.c struct TxDesc *txd = tp->TxDescArray + entry; entry 5919 drivers/net/ethernet/realtek/r8169_main.c tp->tx_skb[entry].len = len; entry 5929 drivers/net/ethernet/realtek/r8169_main.c tp->tx_skb[entry].skb = skb; entry 5941 drivers/net/ethernet/realtek/r8169_main.c txd->opts1 = rtl8169_get_txd_opts1(opts[0], len, entry); entry 5977 drivers/net/ethernet/realtek/r8169_main.c rtl8169_unmap_tx_skb(d, tp->tx_skb + entry, txd); entry 6067 drivers/net/ethernet/realtek/r8169_main.c unsigned int entry = dirty_tx % NUM_TX_DESC; entry 6068 drivers/net/ethernet/realtek/r8169_main.c struct ring_info *tx_skb = tp->tx_skb + entry; entry 6071 drivers/net/ethernet/realtek/r8169_main.c status = le32_to_cpu(tp->TxDescArray[entry].opts1); entry 6082 drivers/net/ethernet/realtek/r8169_main.c tp->TxDescArray + entry); entry 6149 drivers/net/ethernet/realtek/r8169_main.c unsigned int entry = cur_rx % NUM_RX_DESC; entry 6150 drivers/net/ethernet/realtek/r8169_main.c const void *rx_buf = page_address(tp->Rx_databuff[entry]); entry 6151 drivers/net/ethernet/realtek/r8169_main.c struct RxDesc *desc = tp->RxDescArray + entry; entry 181 drivers/net/ethernet/renesas/ravb_main.c int entry; entry 187 drivers/net/ethernet/renesas/ravb_main.c entry = priv->dirty_tx[q] % (priv->num_tx_ring[q] * entry 189 drivers/net/ethernet/renesas/ravb_main.c desc = &priv->tx_ring[q][entry]; entry 197 drivers/net/ethernet/renesas/ravb_main.c if (priv->tx_skb[q][entry / num_tx_desc]) { entry 201 drivers/net/ethernet/renesas/ravb_main.c if (entry % num_tx_desc == num_tx_desc - 1) { entry 202 drivers/net/ethernet/renesas/ravb_main.c entry /= num_tx_desc; entry 203 drivers/net/ethernet/renesas/ravb_main.c dev_kfree_skb_any(priv->tx_skb[q][entry]); entry 204 drivers/net/ethernet/renesas/ravb_main.c priv->tx_skb[q][entry] = NULL; entry 536 drivers/net/ethernet/renesas/ravb_main.c int entry = priv->cur_rx[q] % priv->num_rx_ring[q]; entry 550 drivers/net/ethernet/renesas/ravb_main.c desc = &priv->rx_ring[q][entry]; entry 581 drivers/net/ethernet/renesas/ravb_main.c skb = priv->rx_skb[q][entry]; entry 582 drivers/net/ethernet/renesas/ravb_main.c priv->rx_skb[q][entry] = NULL; entry 609 drivers/net/ethernet/renesas/ravb_main.c entry = (++priv->cur_rx[q]) % priv->num_rx_ring[q]; entry 610 drivers/net/ethernet/renesas/ravb_main.c desc = &priv->rx_ring[q][entry]; entry 615 drivers/net/ethernet/renesas/ravb_main.c entry = priv->dirty_rx[q] % priv->num_rx_ring[q]; entry 616 drivers/net/ethernet/renesas/ravb_main.c desc = &priv->rx_ring[q][entry]; entry 619 drivers/net/ethernet/renesas/ravb_main.c if (!priv->rx_skb[q][entry]) { entry 636 drivers/net/ethernet/renesas/ravb_main.c priv->rx_skb[q][entry] = skb; entry 1482 drivers/net/ethernet/renesas/ravb_main.c u32 entry; entry 1498 drivers/net/ethernet/renesas/ravb_main.c entry = priv->cur_tx[q] % (priv->num_tx_ring[q] * num_tx_desc); entry 1499 drivers/net/ethernet/renesas/ravb_main.c priv->tx_skb[q][entry / num_tx_desc] = skb; entry 1503 drivers/net/ethernet/renesas/ravb_main.c entry / num_tx_desc * DPTR_ALIGN; entry 1527 drivers/net/ethernet/renesas/ravb_main.c desc = &priv->tx_ring[q][entry]; entry 1540 drivers/net/ethernet/renesas/ravb_main.c desc = &priv->tx_ring[q][entry]; entry 1599 drivers/net/ethernet/renesas/ravb_main.c priv->tx_skb[q][entry / num_tx_desc] = NULL; entry 1271 drivers/net/ethernet/renesas/sh_eth.c int entry; entry 1275 drivers/net/ethernet/renesas/sh_eth.c entry = mdp->dirty_tx % mdp->num_tx_ring; entry 1276 drivers/net/ethernet/renesas/sh_eth.c txdesc = &mdp->tx_ring[entry]; entry 1284 drivers/net/ethernet/renesas/sh_eth.c entry, le32_to_cpu(txdesc->status)); entry 1286 drivers/net/ethernet/renesas/sh_eth.c if (mdp->tx_skbuff[entry]) { entry 1291 drivers/net/ethernet/renesas/sh_eth.c dev_kfree_skb_irq(mdp->tx_skbuff[entry]); entry 1292 drivers/net/ethernet/renesas/sh_eth.c mdp->tx_skbuff[entry] = NULL; entry 1296 drivers/net/ethernet/renesas/sh_eth.c if (entry >= mdp->num_tx_ring - 1) entry 1624 drivers/net/ethernet/renesas/sh_eth.c int entry = mdp->cur_rx % mdp->num_rx_ring; entry 1636 drivers/net/ethernet/renesas/sh_eth.c rxdesc = &mdp->rx_ring[entry]; entry 1648 drivers/net/ethernet/renesas/sh_eth.c entry, desc_status, pkt_len); entry 1662 drivers/net/ethernet/renesas/sh_eth.c skb = mdp->rx_skbuff[entry]; entry 1684 drivers/net/ethernet/renesas/sh_eth.c mdp->rx_skbuff[entry] = NULL; entry 1700 drivers/net/ethernet/renesas/sh_eth.c entry = (++mdp->cur_rx) % mdp->num_rx_ring; entry 1701 drivers/net/ethernet/renesas/sh_eth.c rxdesc = &mdp->rx_ring[entry]; entry 1706 drivers/net/ethernet/renesas/sh_eth.c entry = mdp->dirty_rx % mdp->num_rx_ring; entry 1707 drivers/net/ethernet/renesas/sh_eth.c rxdesc = &mdp->rx_ring[entry]; entry 1712 drivers/net/ethernet/renesas/sh_eth.c if (mdp->rx_skbuff[entry] == NULL) { entry 1723 drivers/net/ethernet/renesas/sh_eth.c mdp->rx_skbuff[entry] = skb; entry 1729 drivers/net/ethernet/renesas/sh_eth.c if (entry >= mdp->num_rx_ring - 1) entry 2525 drivers/net/ethernet/renesas/sh_eth.c u32 entry; entry 2542 drivers/net/ethernet/renesas/sh_eth.c entry = mdp->cur_tx % mdp->num_tx_ring; entry 2543 drivers/net/ethernet/renesas/sh_eth.c mdp->tx_skbuff[entry] = skb; entry 2544 drivers/net/ethernet/renesas/sh_eth.c txdesc = &mdp->tx_ring[entry]; entry 2558 drivers/net/ethernet/renesas/sh_eth.c if (entry >= mdp->num_tx_ring - 1) entry 2676 drivers/net/ethernet/renesas/sh_eth.c static u32 sh_eth_tsu_get_post_mask(int entry) entry 2678 drivers/net/ethernet/renesas/sh_eth.c return 0x0f << (28 - ((entry % 8) * 4)); entry 2681 drivers/net/ethernet/renesas/sh_eth.c static u32 sh_eth_tsu_get_post_bit(struct sh_eth_private *mdp, int entry) entry 2683 drivers/net/ethernet/renesas/sh_eth.c return (0x08 >> (mdp->port << 1)) << (28 - ((entry % 8) * 4)); entry 2687 drivers/net/ethernet/renesas/sh_eth.c int entry) entry 2690 drivers/net/ethernet/renesas/sh_eth.c int reg = TSU_POST1 + entry / 8; entry 2694 drivers/net/ethernet/renesas/sh_eth.c sh_eth_tsu_write(mdp, tmp | sh_eth_tsu_get_post_bit(mdp, entry), reg); entry 2698 drivers/net/ethernet/renesas/sh_eth.c int entry) entry 2701 drivers/net/ethernet/renesas/sh_eth.c int reg = TSU_POST1 + entry / 8; entry 2704 drivers/net/ethernet/renesas/sh_eth.c post_mask = sh_eth_tsu_get_post_mask(entry); entry 2705 drivers/net/ethernet/renesas/sh_eth.c ref_mask = sh_eth_tsu_get_post_bit(mdp, entry) & ~post_mask; entry 2785 drivers/net/ethernet/renesas/sh_eth.c int entry; entry 2788 drivers/net/ethernet/renesas/sh_eth.c entry = sh_eth_tsu_find_entry(ndev, blank); entry 2789 drivers/net/ethernet/renesas/sh_eth.c return (entry < 0) ? -ENOMEM : entry; entry 2793 drivers/net/ethernet/renesas/sh_eth.c int entry) entry 2801 drivers/net/ethernet/renesas/sh_eth.c ~(1 << (31 - entry)), TSU_TEN); entry 2804 drivers/net/ethernet/renesas/sh_eth.c ret = sh_eth_tsu_write_entry(ndev, reg_offset + entry * 8, blank); entry 2698 drivers/net/ethernet/rocker/rocker_main.c rocker->msix_entries[i].entry = i; entry 93 drivers/net/ethernet/rocker/rocker_ofdpa.c struct hlist_node entry; entry 103 drivers/net/ethernet/rocker/rocker_ofdpa.c struct hlist_node entry; entry 129 drivers/net/ethernet/rocker/rocker_ofdpa.c struct hlist_node entry; entry 141 drivers/net/ethernet/rocker/rocker_ofdpa.c struct hlist_node entry; entry 148 drivers/net/ethernet/rocker/rocker_ofdpa.c struct hlist_node entry; entry 306 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_flow_tbl_entry *entry) entry 309 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ig_port.in_pport)) entry 312 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ig_port.in_pport_mask)) entry 315 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ig_port.goto_tbl)) entry 323 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_flow_tbl_entry *entry) entry 326 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.in_pport)) entry 329 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.vlan_id)) entry 332 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.vlan_id_mask)) entry 335 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.goto_tbl)) entry 337 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->key.vlan.untagged && entry 339 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.new_vlan_id)) entry 347 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_flow_tbl_entry *entry) entry 350 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.in_pport)) entry 353 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.in_pport_mask)) entry 356 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.eth_type)) entry 359 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->key.term_mac.eth_dst)) entry 362 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->key.term_mac.eth_dst_mask)) entry 365 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.vlan_id)) entry 368 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.vlan_id_mask)) entry 371 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.goto_tbl)) entry 373 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->key.term_mac.copy_to_cpu && entry 375 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.copy_to_cpu)) entry 383 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_flow_tbl_entry *entry) entry 386 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.eth_type)) entry 389 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.dst4)) entry 392 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.dst4_mask)) entry 395 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.goto_tbl)) entry 398 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.group_id)) entry 406 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_flow_tbl_entry *entry) entry 408 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->key.bridge.has_eth_dst && entry 410 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->key.bridge.eth_dst)) entry 412 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->key.bridge.has_eth_dst_mask && entry 414 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->key.bridge.eth_dst_mask)) entry 416 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->key.bridge.vlan_id && entry 418 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.vlan_id)) entry 420 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->key.bridge.tunnel_id && entry 422 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.tunnel_id)) entry 425 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.goto_tbl)) entry 428 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.group_id)) entry 430 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->key.bridge.copy_to_cpu && entry 432 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.copy_to_cpu)) entry 440 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_flow_tbl_entry *entry) entry 443 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.in_pport)) entry 446 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.in_pport_mask)) entry 449 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->key.acl.eth_src)) entry 452 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->key.acl.eth_src_mask)) entry 455 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->key.acl.eth_dst)) entry 458 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->key.acl.eth_dst_mask)) entry 461 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.eth_type)) entry 464 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.vlan_id)) entry 467 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.vlan_id_mask)) entry 470 drivers/net/ethernet/rocker/rocker_ofdpa.c switch (ntohs(entry->key.acl.eth_type)) { entry 474 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.ip_proto)) entry 478 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.ip_proto_mask)) entry 481 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.ip_tos & 0x3f)) entry 485 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.ip_tos_mask & 0x3f)) entry 488 drivers/net/ethernet/rocker/rocker_ofdpa.c (entry->key.acl.ip_tos & 0xc0) >> 6)) entry 492 drivers/net/ethernet/rocker/rocker_ofdpa.c (entry->key.acl.ip_tos_mask & 0xc0) >> 6)) entry 497 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->key.acl.group_id != ROCKER_GROUP_NONE && entry 499 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.group_id)) entry 509 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_flow_tbl_entry *entry = priv; entry 513 drivers/net/ethernet/rocker/rocker_ofdpa.c if (rocker_tlv_put_u16(desc_info, ROCKER_TLV_CMD_TYPE, entry->cmd)) entry 519 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.tbl_id)) entry 522 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.priority)) entry 527 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->cookie)) entry 530 drivers/net/ethernet/rocker/rocker_ofdpa.c switch (entry->key.tbl_id) { entry 532 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_flow_tbl_add_ig_port(desc_info, entry); entry 535 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_flow_tbl_add_vlan(desc_info, entry); entry 538 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_flow_tbl_add_term_mac(desc_info, entry); entry 541 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_flow_tbl_add_ucast_routing(desc_info, entry); entry 544 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_flow_tbl_add_bridge(desc_info, entry); entry 547 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_flow_tbl_add_acl(desc_info, entry); entry 566 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_flow_tbl_entry *entry = priv; entry 569 drivers/net/ethernet/rocker/rocker_ofdpa.c if (rocker_tlv_put_u16(desc_info, ROCKER_TLV_CMD_TYPE, entry->cmd)) entry 575 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->cookie)) entry 584 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_group_tbl_entry *entry) entry 587 drivers/net/ethernet/rocker/rocker_ofdpa.c ROCKER_GROUP_PORT_GET(entry->group_id))) entry 590 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l2_interface.pop_vlan)) entry 598 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_group_tbl_entry *entry) entry 601 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l2_rewrite.group_id)) entry 603 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!is_zero_ether_addr(entry->l2_rewrite.eth_src) && entry 605 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->l2_rewrite.eth_src)) entry 607 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!is_zero_ether_addr(entry->l2_rewrite.eth_dst) && entry 609 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->l2_rewrite.eth_dst)) entry 611 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->l2_rewrite.vlan_id && entry 613 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l2_rewrite.vlan_id)) entry 621 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_group_tbl_entry *entry) entry 627 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->group_count)) entry 635 drivers/net/ethernet/rocker/rocker_ofdpa.c for (i = 0; i < entry->group_count; i++) entry 637 drivers/net/ethernet/rocker/rocker_ofdpa.c if (rocker_tlv_put_u32(desc_info, i + 1, entry->group_ids[i])) entry 647 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_group_tbl_entry *entry) entry 649 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!is_zero_ether_addr(entry->l3_unicast.eth_src) && entry 651 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->l3_unicast.eth_src)) entry 653 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!is_zero_ether_addr(entry->l3_unicast.eth_dst) && entry 655 drivers/net/ethernet/rocker/rocker_ofdpa.c ETH_ALEN, entry->l3_unicast.eth_dst)) entry 657 drivers/net/ethernet/rocker/rocker_ofdpa.c if (entry->l3_unicast.vlan_id && entry 659 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l3_unicast.vlan_id)) entry 662 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l3_unicast.ttl_check)) entry 665 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l3_unicast.group_id)) entry 675 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_group_tbl_entry *entry = priv; entry 679 drivers/net/ethernet/rocker/rocker_ofdpa.c if (rocker_tlv_put_u16(desc_info, ROCKER_TLV_CMD_TYPE, entry->cmd)) entry 686 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->group_id)) entry 689 drivers/net/ethernet/rocker/rocker_ofdpa.c switch (ROCKER_GROUP_TYPE_GET(entry->group_id)) { entry 691 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_group_tbl_add_l2_interface(desc_info, entry); entry 694 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_group_tbl_add_l2_rewrite(desc_info, entry); entry 698 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_group_tbl_add_group_ids(desc_info, entry); entry 701 drivers/net/ethernet/rocker/rocker_ofdpa.c err = ofdpa_cmd_group_tbl_add_l3_unicast(desc_info, entry); entry 720 drivers/net/ethernet/rocker/rocker_ofdpa.c const struct ofdpa_group_tbl_entry *entry = priv; entry 723 drivers/net/ethernet/rocker/rocker_ofdpa.c if (rocker_tlv_put_u16(desc_info, ROCKER_TLV_CMD_TYPE, entry->cmd)) entry 729 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->group_id)) entry 748 drivers/net/ethernet/rocker/rocker_ofdpa.c entry, match->key_crc32) { entry 772 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&found->entry); entry 782 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_add(ofdpa->flow_tbl, &found->entry, found->key_crc32); entry 807 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&found->entry); entry 827 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_flow_tbl_entry *entry) entry 830 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_flow_tbl_del(ofdpa_port, flags, entry); entry 832 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_flow_tbl_add(ofdpa_port, flags, entry); entry 839 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_flow_tbl_entry *entry; entry 841 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 842 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 845 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.priority = OFDPA_PRIORITY_IG_PORT; entry 846 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_INGRESS_PORT; entry 847 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ig_port.in_pport = in_pport; entry 848 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ig_port.in_pport_mask = in_pport_mask; entry 849 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ig_port.goto_tbl = goto_tbl; entry 851 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_flow_tbl_do(ofdpa_port, flags, entry); entry 861 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_flow_tbl_entry *entry; entry 863 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 864 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 867 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.priority = OFDPA_PRIORITY_VLAN; entry 868 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_VLAN; entry 869 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.in_pport = in_pport; entry 870 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.vlan_id = vlan_id; entry 871 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.vlan_id_mask = vlan_id_mask; entry 872 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.goto_tbl = goto_tbl; entry 874 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.untagged = untagged; entry 875 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan.new_vlan_id = new_vlan_id; entry 877 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_flow_tbl_do(ofdpa_port, flags, entry); entry 887 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_flow_tbl_entry *entry; entry 889 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 890 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 894 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.priority = OFDPA_PRIORITY_TERM_MAC_MCAST; entry 895 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.goto_tbl = entry 898 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.priority = OFDPA_PRIORITY_TERM_MAC_UCAST; entry 899 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.goto_tbl = entry 903 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_TERMINATION_MAC; entry 904 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.in_pport = in_pport; entry 905 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.in_pport_mask = in_pport_mask; entry 906 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.eth_type = eth_type; entry 907 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->key.term_mac.eth_dst, eth_dst); entry 908 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->key.term_mac.eth_dst_mask, eth_dst_mask); entry 909 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.vlan_id = vlan_id; entry 910 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.vlan_id_mask = vlan_id_mask; entry 911 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.term_mac.copy_to_cpu = copy_to_cpu; entry 913 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_flow_tbl_do(ofdpa_port, flags, entry); entry 923 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_flow_tbl_entry *entry; entry 929 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 930 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 933 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_BRIDGING; entry 936 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.has_eth_dst = 1; entry 937 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->key.bridge.eth_dst, eth_dst); entry 940 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.has_eth_dst_mask = 1; entry 941 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->key.bridge.eth_dst_mask, eth_dst_mask); entry 960 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.priority = priority; entry 961 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.vlan_id = vlan_id; entry 962 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.tunnel_id = tunnel_id; entry 963 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.goto_tbl = goto_tbl; entry 964 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.group_id = group_id; entry 965 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.bridge.copy_to_cpu = copy_to_cpu; entry 967 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_flow_tbl_do(ofdpa_port, flags, entry); entry 977 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_flow_tbl_entry *entry; entry 979 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 980 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 983 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_UNICAST_ROUTING; entry 984 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.priority = priority; entry 985 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.eth_type = eth_type; entry 986 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.dst4 = dst; entry 987 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.dst4_mask = dst_mask; entry 988 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.goto_tbl = goto_tbl; entry 989 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.ucast_routing.group_id = group_id; entry 990 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key_len = offsetof(struct ofdpa_flow_tbl_key, entry 992 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->fi = fi; entry 994 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_flow_tbl_do(ofdpa_port, flags, entry); entry 1007 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_flow_tbl_entry *entry; entry 1009 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1010 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 1021 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.priority = priority; entry 1022 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.tbl_id = ROCKER_OF_DPA_TABLE_ID_ACL_POLICY; entry 1023 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.in_pport = in_pport; entry 1024 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.in_pport_mask = in_pport_mask; entry 1027 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->key.acl.eth_src, eth_src); entry 1029 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->key.acl.eth_src_mask, eth_src_mask); entry 1031 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->key.acl.eth_dst, eth_dst); entry 1033 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->key.acl.eth_dst_mask, eth_dst_mask); entry 1035 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.eth_type = eth_type; entry 1036 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.vlan_id = vlan_id; entry 1037 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.vlan_id_mask = vlan_id_mask; entry 1038 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.ip_proto = ip_proto; entry 1039 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.ip_proto_mask = ip_proto_mask; entry 1040 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.ip_tos = ip_tos; entry 1041 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.ip_tos_mask = ip_tos_mask; entry 1042 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.acl.group_id = group_id; entry 1044 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_flow_tbl_do(ofdpa_port, flags, entry); entry 1054 drivers/net/ethernet/rocker/rocker_ofdpa.c entry, match->group_id) { entry 1062 drivers/net/ethernet/rocker/rocker_ofdpa.c static void ofdpa_group_tbl_entry_free(struct ofdpa_group_tbl_entry *entry) entry 1064 drivers/net/ethernet/rocker/rocker_ofdpa.c switch (ROCKER_GROUP_TYPE_GET(entry->group_id)) { entry 1067 drivers/net/ethernet/rocker/rocker_ofdpa.c kfree(entry->group_ids); entry 1072 drivers/net/ethernet/rocker/rocker_ofdpa.c kfree(entry); entry 1087 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&found->entry); entry 1096 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_add(ofdpa->group_tbl, &found->entry, found->group_id); entry 1119 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&found->entry); entry 1139 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_group_tbl_entry *entry) entry 1142 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_group_tbl_del(ofdpa_port, flags, entry); entry 1144 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_group_tbl_add(ofdpa_port, flags, entry); entry 1151 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_group_tbl_entry *entry; entry 1153 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1154 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 1157 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->group_id = ROCKER_GROUP_L2_INTERFACE(vlan_id, out_pport); entry 1158 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l2_interface.pop_vlan = pop_vlan; entry 1160 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_group_tbl_do(ofdpa_port, flags, entry); entry 1167 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_group_tbl_entry *entry; entry 1169 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1170 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 1173 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->group_id = group_id; entry 1174 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->group_count = group_count; entry 1176 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->group_ids = kcalloc(group_count, sizeof(u32), GFP_KERNEL); entry 1177 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry->group_ids) { entry 1178 drivers/net/ethernet/rocker/rocker_ofdpa.c kfree(entry); entry 1181 drivers/net/ethernet/rocker/rocker_ofdpa.c memcpy(entry->group_ids, group_ids, group_count * sizeof(u32)); entry 1183 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_group_tbl_do(ofdpa_port, flags, entry); entry 1200 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_group_tbl_entry *entry; entry 1202 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1203 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 1206 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->group_id = ROCKER_GROUP_L3_UNICAST(index); entry 1208 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->l3_unicast.eth_src, src_mac); entry 1210 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->l3_unicast.eth_dst, dst_mac); entry 1211 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l3_unicast.vlan_id = vlan_id; entry 1212 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l3_unicast.ttl_check = ttl_check; entry 1213 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->l3_unicast.group_id = ROCKER_GROUP_L2_INTERFACE(vlan_id, pport); entry 1215 drivers/net/ethernet/rocker/rocker_ofdpa.c return ofdpa_group_tbl_do(ofdpa_port, flags, entry); entry 1224 drivers/net/ethernet/rocker/rocker_ofdpa.c entry, be32_to_cpu(ip_addr)) entry 1232 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_neigh_tbl_entry *entry) entry 1234 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->index = ofdpa->neigh_tbl_next_index++; entry 1235 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->ref_count++; entry 1236 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_add(ofdpa->neigh_tbl, &entry->entry, entry 1237 drivers/net/ethernet/rocker/rocker_ofdpa.c be32_to_cpu(entry->ip_addr)); entry 1240 drivers/net/ethernet/rocker/rocker_ofdpa.c static void ofdpa_neigh_del(struct ofdpa_neigh_tbl_entry *entry) entry 1242 drivers/net/ethernet/rocker/rocker_ofdpa.c if (--entry->ref_count == 0) { entry 1243 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&entry->entry); entry 1244 drivers/net/ethernet/rocker/rocker_ofdpa.c kfree(entry); entry 1248 drivers/net/ethernet/rocker/rocker_ofdpa.c static void ofdpa_neigh_update(struct ofdpa_neigh_tbl_entry *entry, entry 1252 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->eth_dst, eth_dst); entry 1253 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->ttl_check = ttl_check; entry 1255 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->ref_count++; entry 1263 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_neigh_tbl_entry *entry; entry 1276 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1277 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 1289 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->ip_addr = ip_addr; entry 1290 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->dev = ofdpa_port->dev; entry 1291 drivers/net/ethernet/rocker/rocker_ofdpa.c ether_addr_copy(entry->eth_dst, eth_dst); entry 1292 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->ttl_check = true; entry 1293 drivers/net/ethernet/rocker/rocker_ofdpa.c ofdpa_neigh_add(ofdpa, entry); entry 1295 drivers/net/ethernet/rocker/rocker_ofdpa.c memcpy(entry, found, sizeof(*entry)); entry 1299 drivers/net/ethernet/rocker/rocker_ofdpa.c memcpy(entry, found, sizeof(*entry)); entry 1316 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->index, entry 1318 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->eth_dst, entry 1320 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->ttl_check, entry 1324 drivers/net/ethernet/rocker/rocker_ofdpa.c err, entry->index); entry 1329 drivers/net/ethernet/rocker/rocker_ofdpa.c group_id = ROCKER_GROUP_L3_UNICAST(entry->index); entry 1338 drivers/net/ethernet/rocker/rocker_ofdpa.c err, &entry->ip_addr, group_id); entry 1343 drivers/net/ethernet/rocker/rocker_ofdpa.c kfree(entry); entry 1380 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_neigh_tbl_entry *entry; entry 1389 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1390 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 1402 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->ip_addr = ip_addr; entry 1403 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->dev = ofdpa_port->dev; entry 1404 drivers/net/ethernet/rocker/rocker_ofdpa.c ofdpa_neigh_add(ofdpa, entry); entry 1405 drivers/net/ethernet/rocker/rocker_ofdpa.c *index = entry->index; entry 1421 drivers/net/ethernet/rocker/rocker_ofdpa.c kfree(entry); entry 1889 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_possible(ofdpa->fdb_tbl, found, entry, match->key_crc32) entry 1925 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&found->entry); entry 1928 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_add(ofdpa->fdb_tbl, &fdb->entry, entry 1963 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, found, entry) { entry 1973 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&found->entry); entry 1986 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_fdb_tbl_entry *entry; entry 1997 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, entry, entry) { entry 1998 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry->learned) entry 2000 drivers/net/ethernet/rocker/rocker_ofdpa.c ofdpa_port = entry->key.ofdpa_port; entry 2001 drivers/net/ethernet/rocker/rocker_ofdpa.c expires = entry->touched + ofdpa_port->ageing_time; entry 2004 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.addr, entry 2005 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->key.vlan_id); entry 2006 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&entry->entry); entry 2217 drivers/net/ethernet/rocker/rocker_ofdpa.c entry, ifindex) { entry 2229 drivers/net/ethernet/rocker/rocker_ofdpa.c struct ofdpa_internal_vlan_tbl_entry *entry; entry 2234 drivers/net/ethernet/rocker/rocker_ofdpa.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 2235 drivers/net/ethernet/rocker/rocker_ofdpa.c if (!entry) entry 2238 drivers/net/ethernet/rocker/rocker_ofdpa.c entry->ifindex = ifindex; entry 2244 drivers/net/ethernet/rocker/rocker_ofdpa.c kfree(entry); entry 2248 drivers/net/ethernet/rocker/rocker_ofdpa.c found = entry; entry 2249 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_add(ofdpa->internal_vlan_tbl, &found->entry, found->ifindex); entry 2334 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&found->entry); entry 2392 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry) entry 2393 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&flow_entry->entry); entry 2397 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->group_tbl, bkt, tmp, group_entry, entry) entry 2398 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&group_entry->entry); entry 2402 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, fdb_entry, entry) entry 2403 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&fdb_entry->entry); entry 2408 drivers/net/ethernet/rocker/rocker_ofdpa.c tmp, internal_vlan_entry, entry) entry 2409 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&internal_vlan_entry->entry); entry 2413 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->neigh_tbl, bkt, tmp, neigh_entry, entry) entry 2414 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_del(&neigh_entry->entry); entry 2787 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry) { entry 743 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c unsigned int entry = tqueue->dirty_tx % tx_rsize; entry 744 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sk_buff *skb = tqueue->tx_skbuff[entry]; entry 747 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c p = tqueue->dma_tx + entry; entry 757 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c if (likely(tqueue->tx_skbuff_dma[entry])) { entry 759 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tqueue->tx_skbuff_dma[entry], entry 762 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tqueue->tx_skbuff_dma[entry] = 0; entry 767 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tqueue->tx_skbuff[entry] = NULL; entry 1268 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c unsigned int entry, frag_num; entry 1306 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c entry = tqueue->cur_tx % tx_rsize; entry 1307 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tx_desc = tqueue->dma_tx + entry; entry 1314 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tqueue->tx_skbuff[entry] = skb; entry 1331 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c entry = (++tqueue->cur_tx) % tx_rsize; entry 1332 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c first_desc = tqueue->dma_tx + entry; entry 1353 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c entry = (++tqueue->cur_tx) % tx_rsize; entry 1354 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tx_desc = tqueue->dma_tx + entry; entry 1358 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tqueue->tx_skbuff_dma[entry] = tx_desc->tdes01; entry 1359 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tqueue->tx_skbuff[entry] = NULL; entry 1398 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c tqueue->dirty_tx % tx_rsize, entry, entry 1437 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c unsigned int entry = priv->rxq[qnum]->dirty_rx % rxsize; entry 1440 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c p = priv->rxq[qnum]->dma_rx + entry; entry 1442 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c if (likely(priv->rxq[qnum]->rx_skbuff[entry] == NULL)) { entry 1450 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c priv->rxq[qnum]->rx_skbuff[entry] = skb; entry 1451 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c priv->rxq[qnum]->rx_skbuff_dma[entry] = entry 1456 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c priv->rxq[qnum]->rx_skbuff_dma[entry]; entry 1479 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c unsigned int entry = priv->rxq[qnum]->cur_rx; entry 1490 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c p = priv->rxq[qnum]->dma_rx + entry; entry 1507 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c entry = next_entry; entry 1513 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c skb = priv->rxq[qnum]->rx_skbuff[entry]; entry 1519 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c priv->rxq[qnum]->rx_skbuff[entry] = NULL; entry 1531 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c entry = next_entry; entry 589 drivers/net/ethernet/seeq/sgiseeq.c int len, entry; entry 604 drivers/net/ethernet/seeq/sgiseeq.c entry = sp->tx_new; entry 605 drivers/net/ethernet/seeq/sgiseeq.c td = &sp->tx_desc[entry]; entry 104 drivers/net/ethernet/sfc/ef10.c } *entry; entry 4028 drivers/net/ethernet/sfc/ef10.c return (struct efx_filter_spec *)(table->entry[filter_idx].spec & entry 4036 drivers/net/ethernet/sfc/ef10.c return table->entry[filter_idx].spec & EFX_EF10_FILTER_FLAGS; entry 4045 drivers/net/ethernet/sfc/ef10.c table->entry[filter_idx].spec = (unsigned long)spec | flags; entry 4400 drivers/net/ethernet/sfc/ef10.c table->entry[ins_index].spec &= entry 4419 drivers/net/ethernet/sfc/ef10.c rc = efx_ef10_filter_push(efx, spec, &table->entry[ins_index].handle, entry 4474 drivers/net/ethernet/sfc/ef10.c table->entry[i].handle); entry 4550 drivers/net/ethernet/sfc/ef10.c table->entry[filter_idx].spec &= ~EFX_EF10_FILTER_FLAG_AUTO_OLD; entry 4569 drivers/net/ethernet/sfc/ef10.c &table->entry[filter_idx].handle, entry 4583 drivers/net/ethernet/sfc/ef10.c table->entry[filter_idx].handle); entry 4698 drivers/net/ethernet/sfc/ef10.c if (table->entry[filter_idx].spec && entry 4996 drivers/net/ethernet/sfc/ef10.c table->entry = vzalloc(array_size(HUNT_FILTER_TBL_ROWS, entry 4997 drivers/net/ethernet/sfc/ef10.c sizeof(*table->entry))); entry 4998 drivers/net/ethernet/sfc/ef10.c if (!table->entry) { entry 5090 drivers/net/ethernet/sfc/ef10.c &table->entry[filter_idx].handle, entry 5158 drivers/net/ethernet/sfc/ef10.c table->entry[filter_idx].handle); entry 5168 drivers/net/ethernet/sfc/ef10.c vfree(table->entry); entry 5181 drivers/net/ethernet/sfc/ef10.c if (!table->entry[filter_idx].spec) entry 5185 drivers/net/ethernet/sfc/ef10.c table->entry[filter_idx].spec |= EFX_EF10_FILTER_FLAG_AUTO_OLD; entry 5505 drivers/net/ethernet/sfc/ef10.c if (READ_ONCE(table->entry[i].spec) & entry 6364 drivers/net/ethernet/sfc/ef10.c efx_dword_t entry; entry 6366 drivers/net/ethernet/sfc/ef10.c EFX_POPULATE_DWORD_2(entry, entry 6373 drivers/net/ethernet/sfc/ef10.c num_entries++) = entry; entry 1463 drivers/net/ethernet/sfc/efx.c xentries[i].entry = i; entry 3554 drivers/net/ethernet/sfc/efx.c const struct pci_device_id *entry) entry 3566 drivers/net/ethernet/sfc/efx.c efx->type = (const struct efx_nic_type *) entry->driver_data; entry 1385 drivers/net/ethernet/sfc/falcon/efx.c xentries[i].entry = i; entry 2915 drivers/net/ethernet/sfc/falcon/efx.c const struct pci_device_id *entry) entry 2927 drivers/net/ethernet/sfc/falcon/efx.c efx->type = (const struct ef4_nic_type *) entry->driver_data; entry 162 drivers/net/ethernet/sfc/mcdi_mon.c efx_dword_t *entry) entry 179 drivers/net/ethernet/sfc/mcdi_mon.c *entry = ((efx_dword_t *)hwmon->dma_buf.addr)[index]; entry 192 drivers/net/ethernet/sfc/mcdi_mon.c efx_dword_t entry; entry 196 drivers/net/ethernet/sfc/mcdi_mon.c rc = efx_mcdi_mon_get_entry(dev, mon_attr->index, &entry); entry 200 drivers/net/ethernet/sfc/mcdi_mon.c state = EFX_DWORD_FIELD(entry, MC_CMD_SENSOR_VALUE_ENTRY_TYPEDEF_STATE); entry 204 drivers/net/ethernet/sfc/mcdi_mon.c value = EFX_DWORD_FIELD(entry, MC_CMD_SENSOR_VALUE_ENTRY_TYPEDEF_VALUE); entry 256 drivers/net/ethernet/sfc/mcdi_mon.c efx_dword_t entry; entry 260 drivers/net/ethernet/sfc/mcdi_mon.c rc = efx_mcdi_mon_get_entry(dev, mon_attr->index, &entry); entry 264 drivers/net/ethernet/sfc/mcdi_mon.c state = EFX_DWORD_FIELD(entry, MC_CMD_SENSOR_VALUE_ENTRY_TYPEDEF_STATE); entry 802 drivers/net/ethernet/sgi/ioc3-eth.c static inline void ioc3_tx_unmap(struct ioc3_private *ip, int entry) entry 807 drivers/net/ethernet/sgi/ioc3-eth.c desc = &ip->txr[entry]; entry 652 drivers/net/ethernet/silan/sc92031.c unsigned entry; entry 657 drivers/net/ethernet/silan/sc92031.c entry = priv->tx_tail % NUM_TX_DESC; entry 658 drivers/net/ethernet/silan/sc92031.c tx_status = ioread32(port_base + TxStatus0 + entry * 4); entry 937 drivers/net/ethernet/silan/sc92031.c unsigned entry; entry 954 drivers/net/ethernet/silan/sc92031.c entry = priv->tx_head++ % NUM_TX_DESC; entry 956 drivers/net/ethernet/silan/sc92031.c skb_copy_and_csum_dev(skb, priv->tx_bufs + entry * TX_BUF_SIZE); entry 960 drivers/net/ethernet/silan/sc92031.c memset(priv->tx_bufs + entry * TX_BUF_SIZE + len, entry 974 drivers/net/ethernet/silan/sc92031.c iowrite32(priv->tx_bufs_dma_addr + entry * TX_BUF_SIZE, entry 975 drivers/net/ethernet/silan/sc92031.c port_base + TxAddr0 + entry * 4); entry 976 drivers/net/ethernet/silan/sc92031.c iowrite32(tx_status, port_base + TxStatus0 + entry * 4); entry 585 drivers/net/ethernet/sis/sis190.c unsigned int entry = cur_rx % NUM_RX_DESC; entry 586 drivers/net/ethernet/sis/sis190.c struct RxDesc *desc = tp->RxDescRing + entry; entry 599 drivers/net/ethernet/sis/sis190.c struct sk_buff *skb = tp->Rx_skbuff[entry]; entry 621 drivers/net/ethernet/sis/sis190.c tp->Rx_skbuff[entry] = NULL; entry 699 drivers/net/ethernet/sis/sis190.c unsigned int entry = dirty_tx % NUM_TX_DESC; entry 700 drivers/net/ethernet/sis/sis190.c struct TxDesc *txd = tp->TxDescRing + entry; entry 707 drivers/net/ethernet/sis/sis190.c skb = tp->Tx_skbuff[entry]; entry 716 drivers/net/ethernet/sis/sis190.c tp->Tx_skbuff[entry] = NULL; entry 1176 drivers/net/ethernet/sis/sis190.c u32 len, entry, dirty_tx; entry 1190 drivers/net/ethernet/sis/sis190.c entry = tp->cur_tx % NUM_TX_DESC; entry 1191 drivers/net/ethernet/sis/sis190.c desc = tp->TxDescRing + entry; entry 1207 drivers/net/ethernet/sis/sis190.c tp->Tx_skbuff[entry] = skb; entry 1213 drivers/net/ethernet/sis/sis190.c if (entry == (NUM_TX_DESC - 1)) entry 1603 drivers/net/ethernet/sis/sis900.c unsigned int entry; entry 1611 drivers/net/ethernet/sis/sis900.c entry = sis_priv->cur_tx % NUM_TX_DESC; entry 1612 drivers/net/ethernet/sis/sis900.c sis_priv->tx_skbuff[entry] = skb; entry 1615 drivers/net/ethernet/sis/sis900.c sis_priv->tx_ring[entry].bufptr = pci_map_single(sis_priv->pci_dev, entry 1618 drivers/net/ethernet/sis/sis900.c sis_priv->tx_ring[entry].bufptr))) { entry 1620 drivers/net/ethernet/sis/sis900.c sis_priv->tx_skbuff[entry] = NULL; entry 1625 drivers/net/ethernet/sis/sis900.c sis_priv->tx_ring[entry].cmdsts = (OWN | INTR | skb->len); entry 1653 drivers/net/ethernet/sis/sis900.c net_dev->name, skb->data, (int)skb->len, entry); entry 1734 drivers/net/ethernet/sis/sis900.c unsigned int entry = sis_priv->cur_rx % NUM_RX_DESC; entry 1735 drivers/net/ethernet/sis/sis900.c u32 rx_status = sis_priv->rx_ring[entry].cmdsts; entry 1776 drivers/net/ethernet/sis/sis900.c sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE; entry 1782 drivers/net/ethernet/sis/sis900.c sis_priv->rx_ring[entry].bufptr, RX_BUF_SIZE, entry 1794 drivers/net/ethernet/sis/sis900.c skb = sis_priv->rx_skbuff[entry]; entry 1802 drivers/net/ethernet/sis/sis900.c if (sis_priv->rx_skbuff[entry] == NULL) { entry 1814 drivers/net/ethernet/sis/sis900.c rx_skb = sis_priv->rx_skbuff[entry]; entry 1826 drivers/net/ethernet/sis/sis900.c sis_priv->rx_skbuff[entry] = skb; entry 1827 drivers/net/ethernet/sis/sis900.c sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE; entry 1828 drivers/net/ethernet/sis/sis900.c sis_priv->rx_ring[entry].bufptr = entry 1832 drivers/net/ethernet/sis/sis900.c sis_priv->rx_ring[entry].bufptr))) { entry 1834 drivers/net/ethernet/sis/sis900.c sis_priv->rx_skbuff[entry] = NULL; entry 1839 drivers/net/ethernet/sis/sis900.c entry = sis_priv->cur_rx % NUM_RX_DESC; entry 1840 drivers/net/ethernet/sis/sis900.c rx_status = sis_priv->rx_ring[entry].cmdsts; entry 1848 drivers/net/ethernet/sis/sis900.c entry = sis_priv->dirty_rx % NUM_RX_DESC; entry 1850 drivers/net/ethernet/sis/sis900.c if (sis_priv->rx_skbuff[entry] == NULL) { entry 1860 drivers/net/ethernet/sis/sis900.c sis_priv->rx_skbuff[entry] = skb; entry 1861 drivers/net/ethernet/sis/sis900.c sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE; entry 1862 drivers/net/ethernet/sis/sis900.c sis_priv->rx_ring[entry].bufptr = entry 1866 drivers/net/ethernet/sis/sis900.c sis_priv->rx_ring[entry].bufptr))) { entry 1868 drivers/net/ethernet/sis/sis900.c sis_priv->rx_skbuff[entry] = NULL; entry 1895 drivers/net/ethernet/sis/sis900.c unsigned int entry; entry 1898 drivers/net/ethernet/sis/sis900.c entry = sis_priv->dirty_tx % NUM_TX_DESC; entry 1899 drivers/net/ethernet/sis/sis900.c tx_status = sis_priv->tx_ring[entry].cmdsts; entry 1930 drivers/net/ethernet/sis/sis900.c skb = sis_priv->tx_skbuff[entry]; entry 1932 drivers/net/ethernet/sis/sis900.c sis_priv->tx_ring[entry].bufptr, skb->len, entry 1935 drivers/net/ethernet/sis/sis900.c sis_priv->tx_skbuff[entry] = NULL; entry 1936 drivers/net/ethernet/sis/sis900.c sis_priv->tx_ring[entry].bufptr = 0; entry 1937 drivers/net/ethernet/sis/sis900.c sis_priv->tx_ring[entry].cmdsts = 0; entry 941 drivers/net/ethernet/smsc/epic100.c int entry, free_count; entry 954 drivers/net/ethernet/smsc/epic100.c entry = ep->cur_tx % TX_RING_SIZE; entry 956 drivers/net/ethernet/smsc/epic100.c ep->tx_skbuff[entry] = skb; entry 957 drivers/net/ethernet/smsc/epic100.c ep->tx_ring[entry].bufaddr = pci_map_single(ep->pci_dev, skb->data, entry 970 drivers/net/ethernet/smsc/epic100.c ep->tx_ring[entry].buflength = ctrl_word | skb->len; entry 971 drivers/net/ethernet/smsc/epic100.c ep->tx_ring[entry].txstatus = entry 985 drivers/net/ethernet/smsc/epic100.c skb->len, entry, ctrl_word, er32(TxSTAT)); entry 1023 drivers/net/ethernet/smsc/epic100.c int entry = dirty_tx % TX_RING_SIZE; entry 1024 drivers/net/ethernet/smsc/epic100.c int txstatus = ep->tx_ring[entry].txstatus; entry 1032 drivers/net/ethernet/smsc/epic100.c dev->stats.tx_bytes += ep->tx_skbuff[entry]->len; entry 1037 drivers/net/ethernet/smsc/epic100.c skb = ep->tx_skbuff[entry]; entry 1038 drivers/net/ethernet/smsc/epic100.c pci_unmap_single(ep->pci_dev, ep->tx_ring[entry].bufaddr, entry 1041 drivers/net/ethernet/smsc/epic100.c ep->tx_skbuff[entry] = NULL; entry 1133 drivers/net/ethernet/smsc/epic100.c int entry = ep->cur_rx % RX_RING_SIZE; entry 1138 drivers/net/ethernet/smsc/epic100.c netdev_dbg(dev, " In epic_rx(), entry %d %8.8x.\n", entry, entry 1139 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[entry].rxstatus); entry 1145 drivers/net/ethernet/smsc/epic100.c while ((ep->rx_ring[entry].rxstatus & DescOwn) == 0) { entry 1146 drivers/net/ethernet/smsc/epic100.c int status = ep->rx_ring[entry].rxstatus; entry 1181 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[entry].bufaddr, entry 1184 drivers/net/ethernet/smsc/epic100.c skb_copy_to_linear_data(skb, ep->rx_skbuff[entry]->data, pkt_len); entry 1187 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[entry].bufaddr, entry 1192 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[entry].bufaddr, entry 1194 drivers/net/ethernet/smsc/epic100.c skb_put(skb = ep->rx_skbuff[entry], pkt_len); entry 1195 drivers/net/ethernet/smsc/epic100.c ep->rx_skbuff[entry] = NULL; entry 1203 drivers/net/ethernet/smsc/epic100.c entry = (++ep->cur_rx) % RX_RING_SIZE; entry 1208 drivers/net/ethernet/smsc/epic100.c entry = ep->dirty_rx % RX_RING_SIZE; entry 1209 drivers/net/ethernet/smsc/epic100.c if (ep->rx_skbuff[entry] == NULL) { entry 1211 drivers/net/ethernet/smsc/epic100.c skb = ep->rx_skbuff[entry] = netdev_alloc_skb(dev, ep->rx_buf_sz + 2); entry 1215 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[entry].bufaddr = pci_map_single(ep->pci_dev, entry 1220 drivers/net/ethernet/smsc/epic100.c ep->rx_ring[entry].rxstatus = DescOwn; entry 631 drivers/net/ethernet/socionext/netsec.c struct netsec_de *entry; entry 639 drivers/net/ethernet/socionext/netsec.c entry = dring->vaddr + DESC_SZ * tail; entry 641 drivers/net/ethernet/socionext/netsec.c while (!(entry->attr & (1U << NETSEC_TX_SHIFT_OWN_FIELD)) && entry 647 drivers/net/ethernet/socionext/netsec.c eop = (entry->attr >> NETSEC_TX_LAST) & 1; entry 675 drivers/net/ethernet/socionext/netsec.c entry->attr = 1U << NETSEC_TX_SHIFT_OWN_FIELD; entry 680 drivers/net/ethernet/socionext/netsec.c entry = dring->vaddr + DESC_SZ * tail; entry 296 drivers/net/ethernet/socionext/sni_ave.c static u32 ave_desc_read(struct net_device *ndev, enum desc_id id, int entry, entry 303 drivers/net/ethernet/socionext/sni_ave.c + entry * priv->desc_size + offset; entry 309 drivers/net/ethernet/socionext/sni_ave.c int entry) entry 311 drivers/net/ethernet/socionext/sni_ave.c return ave_desc_read(ndev, id, entry, AVE_DESC_OFS_CMDSTS); entry 315 drivers/net/ethernet/socionext/sni_ave.c int entry, int offset, u32 val) entry 321 drivers/net/ethernet/socionext/sni_ave.c + entry * priv->desc_size + offset; entry 327 drivers/net/ethernet/socionext/sni_ave.c int entry, u32 val) entry 329 drivers/net/ethernet/socionext/sni_ave.c ave_desc_write(ndev, id, entry, AVE_DESC_OFS_CMDSTS, val); entry 333 drivers/net/ethernet/socionext/sni_ave.c int entry, dma_addr_t paddr) entry 337 drivers/net/ethernet/socionext/sni_ave.c ave_desc_write(ndev, id, entry, AVE_DESC_OFS_ADDRL, entry 341 drivers/net/ethernet/socionext/sni_ave.c entry, AVE_DESC_OFS_ADDRU, entry 578 drivers/net/ethernet/socionext/sni_ave.c static int ave_rxdesc_prepare(struct net_device *ndev, int entry) entry 585 drivers/net/ethernet/socionext/sni_ave.c skb = priv->rx.desc[entry].skbs; entry 597 drivers/net/ethernet/socionext/sni_ave.c ave_desc_write_cmdsts(ndev, AVE_DESCID_RX, entry, entry 609 drivers/net/ethernet/socionext/sni_ave.c ret = ave_dma_map(ndev, &priv->rx.desc[entry], entry 618 drivers/net/ethernet/socionext/sni_ave.c priv->rx.desc[entry].skbs = skb; entry 621 drivers/net/ethernet/socionext/sni_ave.c ave_desc_write_addr(ndev, AVE_DESCID_RX, entry, paddr); entry 624 drivers/net/ethernet/socionext/sni_ave.c ave_desc_write_cmdsts(ndev, AVE_DESCID_RX, entry, entry 981 drivers/net/ethernet/socionext/sni_ave.c static int ave_pfsel_start(struct net_device *ndev, unsigned int entry) entry 986 drivers/net/ethernet/socionext/sni_ave.c if (WARN_ON(entry > AVE_PF_SIZE)) entry 990 drivers/net/ethernet/socionext/sni_ave.c writel(val | BIT(entry), priv->base + AVE_PFEN); entry 995 drivers/net/ethernet/socionext/sni_ave.c static int ave_pfsel_stop(struct net_device *ndev, unsigned int entry) entry 1000 drivers/net/ethernet/socionext/sni_ave.c if (WARN_ON(entry > AVE_PF_SIZE)) entry 1004 drivers/net/ethernet/socionext/sni_ave.c writel(val & ~BIT(entry), priv->base + AVE_PFEN); entry 1010 drivers/net/ethernet/socionext/sni_ave.c unsigned int entry, entry 1016 drivers/net/ethernet/socionext/sni_ave.c if (WARN_ON(entry > AVE_PF_SIZE)) entry 1021 drivers/net/ethernet/socionext/sni_ave.c ave_pfsel_stop(ndev, entry); entry 1025 drivers/net/ethernet/socionext/sni_ave.c AVE_PKTF(entry), AVE_PKTF(entry) + 4); entry 1029 drivers/net/ethernet/socionext/sni_ave.c priv->base + AVE_PFMBYTE(entry)); entry 1030 drivers/net/ethernet/socionext/sni_ave.c writel(AVE_PFMBYTE_MASK1, priv->base + AVE_PFMBYTE(entry) + 4); entry 1033 drivers/net/ethernet/socionext/sni_ave.c writel(AVE_PFMBIT_MASK, priv->base + AVE_PFMBIT(entry)); entry 1036 drivers/net/ethernet/socionext/sni_ave.c writel(0, priv->base + AVE_PFSEL(entry)); entry 1039 drivers/net/ethernet/socionext/sni_ave.c ave_pfsel_start(ndev, entry); entry 1045 drivers/net/ethernet/socionext/sni_ave.c unsigned int entry, u32 rxring) entry 1049 drivers/net/ethernet/socionext/sni_ave.c if (WARN_ON(entry > AVE_PF_SIZE)) entry 1052 drivers/net/ethernet/socionext/sni_ave.c ave_pfsel_stop(ndev, entry); entry 1055 drivers/net/ethernet/socionext/sni_ave.c writel(AVE_PFMBYTE_MASK0, priv->base + AVE_PFMBYTE(entry)); entry 1056 drivers/net/ethernet/socionext/sni_ave.c writel(AVE_PFMBYTE_MASK1, priv->base + AVE_PFMBYTE(entry) + 4); entry 1059 drivers/net/ethernet/socionext/sni_ave.c writel(AVE_PFMBIT_MASK, priv->base + AVE_PFMBIT(entry)); entry 1062 drivers/net/ethernet/socionext/sni_ave.c writel(rxring, priv->base + AVE_PFSEL(entry)); entry 1064 drivers/net/ethernet/socionext/sni_ave.c ave_pfsel_start(ndev, entry); entry 1266 drivers/net/ethernet/socionext/sni_ave.c int entry; entry 1293 drivers/net/ethernet/socionext/sni_ave.c for (entry = 0; entry < priv->tx.ndesc; entry++) { entry 1294 drivers/net/ethernet/socionext/sni_ave.c ave_desc_write_cmdsts(ndev, AVE_DESCID_TX, entry, 0); entry 1295 drivers/net/ethernet/socionext/sni_ave.c ave_desc_write_addr(ndev, AVE_DESCID_TX, entry, 0); entry 1304 drivers/net/ethernet/socionext/sni_ave.c for (entry = 0; entry < priv->rx.ndesc; entry++) { entry 1305 drivers/net/ethernet/socionext/sni_ave.c if (ave_rxdesc_prepare(ndev, entry)) entry 1354 drivers/net/ethernet/socionext/sni_ave.c int entry; entry 1368 drivers/net/ethernet/socionext/sni_ave.c for (entry = 0; entry < priv->tx.ndesc; entry++) { entry 1369 drivers/net/ethernet/socionext/sni_ave.c if (!priv->tx.desc[entry].skbs) entry 1372 drivers/net/ethernet/socionext/sni_ave.c ave_dma_unmap(ndev, &priv->tx.desc[entry], DMA_TO_DEVICE); entry 1373 drivers/net/ethernet/socionext/sni_ave.c dev_kfree_skb_any(priv->tx.desc[entry].skbs); entry 1374 drivers/net/ethernet/socionext/sni_ave.c priv->tx.desc[entry].skbs = NULL; entry 1380 drivers/net/ethernet/socionext/sni_ave.c for (entry = 0; entry < priv->rx.ndesc; entry++) { entry 1381 drivers/net/ethernet/socionext/sni_ave.c if (!priv->rx.desc[entry].skbs) entry 1384 drivers/net/ethernet/socionext/sni_ave.c ave_dma_unmap(ndev, &priv->rx.desc[entry], DMA_FROM_DEVICE); entry 1385 drivers/net/ethernet/socionext/sni_ave.c dev_kfree_skb_any(priv->rx.desc[entry].skbs); entry 1386 drivers/net/ethernet/socionext/sni_ave.c priv->rx.desc[entry].skbs = NULL; entry 22 drivers/net/ethernet/stmicro/stmmac/chain_mode.c unsigned int entry = tx_q->cur_tx; entry 27 drivers/net/ethernet/stmicro/stmmac/chain_mode.c desc = tx_q->dma_tx + entry; entry 41 drivers/net/ethernet/stmicro/stmmac/chain_mode.c tx_q->tx_skbuff_dma[entry].buf = des2; entry 42 drivers/net/ethernet/stmicro/stmmac/chain_mode.c tx_q->tx_skbuff_dma[entry].len = bmax; entry 48 drivers/net/ethernet/stmicro/stmmac/chain_mode.c tx_q->tx_skbuff[entry] = NULL; entry 49 drivers/net/ethernet/stmicro/stmmac/chain_mode.c entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE); entry 50 drivers/net/ethernet/stmicro/stmmac/chain_mode.c desc = tx_q->dma_tx + entry; entry 59 drivers/net/ethernet/stmicro/stmmac/chain_mode.c tx_q->tx_skbuff_dma[entry].buf = des2; entry 60 drivers/net/ethernet/stmicro/stmmac/chain_mode.c tx_q->tx_skbuff_dma[entry].len = bmax; entry 72 drivers/net/ethernet/stmicro/stmmac/chain_mode.c tx_q->tx_skbuff_dma[entry].buf = des2; entry 73 drivers/net/ethernet/stmicro/stmmac/chain_mode.c tx_q->tx_skbuff_dma[entry].len = len; entry 81 drivers/net/ethernet/stmicro/stmmac/chain_mode.c tx_q->cur_tx = entry; entry 83 drivers/net/ethernet/stmicro/stmmac/chain_mode.c return entry; entry 148 drivers/net/ethernet/stmicro/stmmac/chain_mode.c unsigned int entry = tx_q->dirty_tx; entry 150 drivers/net/ethernet/stmicro/stmmac/chain_mode.c if (tx_q->tx_skbuff_dma[entry].last_segment && !priv->extend_desc && entry 331 drivers/net/ethernet/stmicro/stmmac/dwmac5.c struct stmmac_tc_entry *entry, entry 336 drivers/net/ethernet/stmicro/stmmac/dwmac5.c for (i = 0; i < (sizeof(entry->val) / sizeof(u32)); i++) { entry 337 drivers/net/ethernet/stmicro/stmmac/dwmac5.c int real_pos = pos * (sizeof(entry->val) / sizeof(u32)) + i; entry 347 drivers/net/ethernet/stmicro/stmmac/dwmac5.c val = *((u32 *)&entry->val + i); entry 376 drivers/net/ethernet/stmicro/stmmac/dwmac5.c struct stmmac_tc_entry *entry; entry 382 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry = &entries[i]; entry 385 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (!entry->in_use) entry 388 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (entry->in_hw) entry 391 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (entry->is_last) entry 394 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (entry->is_frag) entry 397 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (entry->prio < curr_prio) entry 400 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (entry->prio < min_prio) { entry 401 drivers/net/ethernet/stmicro/stmmac/dwmac5.c min_prio = entry->prio; entry 415 drivers/net/ethernet/stmicro/stmmac/dwmac5.c struct stmmac_tc_entry *entry, *frag; entry 432 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry = &entries[i]; entry 433 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry->in_hw = false; entry 438 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry = dwmac5_rxp_get_next_entry(entries, count, curr_prio); entry 439 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (!entry) entry 442 drivers/net/ethernet/stmicro/stmmac/dwmac5.c curr_prio = entry->prio; entry 443 drivers/net/ethernet/stmicro/stmmac/dwmac5.c frag = entry->frag_ptr; entry 447 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry->val.af = 0; entry 448 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry->val.rf = 0; entry 449 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry->val.nc = 1; entry 450 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry->val.ok_index = nve + 2; entry 453 drivers/net/ethernet/stmicro/stmmac/dwmac5.c ret = dwmac5_rxp_update_single_entry(ioaddr, entry, nve); entry 457 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry->table_pos = nve++; entry 458 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry->in_hw = true; entry 474 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry = &entries[i]; entry 475 drivers/net/ethernet/stmicro/stmmac/dwmac5.c if (!entry->is_last) entry 478 drivers/net/ethernet/stmicro/stmmac/dwmac5.c ret = dwmac5_rxp_update_single_entry(ioaddr, entry, nve); entry 482 drivers/net/ethernet/stmicro/stmmac/dwmac5.c entry->table_pos = nve++; entry 896 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c struct stmmac_tc_entry *entry, entry 901 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c for (i = 0; i < (sizeof(entry->val) / sizeof(u32)); i++) { entry 902 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c int real_pos = pos * (sizeof(entry->val) / sizeof(u32)) + i; entry 912 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c val = *((u32 *)&entry->val + i); entry 941 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c struct stmmac_tc_entry *entry; entry 947 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry = &entries[i]; entry 950 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (!entry->in_use) entry 953 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (entry->in_hw) entry 956 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (entry->is_last) entry 959 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (entry->is_frag) entry 962 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (entry->prio < curr_prio) entry 965 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (entry->prio < min_prio) { entry 966 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c min_prio = entry->prio; entry 981 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c struct stmmac_tc_entry *entry, *frag; entry 998 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry = &entries[i]; entry 999 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry->in_hw = false; entry 1004 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry = dwxgmac3_rxp_get_next_entry(entries, count, curr_prio); entry 1005 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (!entry) entry 1008 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c curr_prio = entry->prio; entry 1009 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c frag = entry->frag_ptr; entry 1013 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry->val.af = 0; entry 1014 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry->val.rf = 0; entry 1015 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry->val.nc = 1; entry 1016 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry->val.ok_index = nve + 2; entry 1019 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, nve); entry 1023 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry->table_pos = nve++; entry 1024 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry->in_hw = true; entry 1040 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry = &entries[i]; entry 1041 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c if (!entry->is_last) entry 1044 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, nve); entry 1048 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c entry->table_pos = nve++; entry 223 drivers/net/ethernet/stmicro/stmmac/hwif.c const struct stmmac_hwif_entry *entry; entry 259 drivers/net/ethernet/stmicro/stmmac/hwif.c entry = &stmmac_hw[i]; entry 261 drivers/net/ethernet/stmicro/stmmac/hwif.c if (needs_gmac ^ entry->gmac) entry 263 drivers/net/ethernet/stmicro/stmmac/hwif.c if (needs_gmac4 ^ entry->gmac4) entry 265 drivers/net/ethernet/stmicro/stmmac/hwif.c if (needs_xgmac ^ entry->xgmac) entry 268 drivers/net/ethernet/stmicro/stmmac/hwif.c if (priv->synopsys_id < entry->min_id) entry 272 drivers/net/ethernet/stmicro/stmmac/hwif.c mac->desc = mac->desc ? : entry->desc; entry 273 drivers/net/ethernet/stmicro/stmmac/hwif.c mac->dma = mac->dma ? : entry->dma; entry 274 drivers/net/ethernet/stmicro/stmmac/hwif.c mac->mac = mac->mac ? : entry->mac; entry 275 drivers/net/ethernet/stmicro/stmmac/hwif.c mac->ptp = mac->ptp ? : entry->hwtimestamp; entry 276 drivers/net/ethernet/stmicro/stmmac/hwif.c mac->mode = mac->mode ? : entry->mode; entry 277 drivers/net/ethernet/stmicro/stmmac/hwif.c mac->tc = mac->tc ? : entry->tc; entry 278 drivers/net/ethernet/stmicro/stmmac/hwif.c mac->mmc = mac->mmc ? : entry->mmc; entry 281 drivers/net/ethernet/stmicro/stmmac/hwif.c priv->ptpaddr = priv->ioaddr + entry->regs.ptp_off; entry 282 drivers/net/ethernet/stmicro/stmmac/hwif.c priv->mmcaddr = priv->ioaddr + entry->regs.mmc_off; entry 286 drivers/net/ethernet/stmicro/stmmac/hwif.c ret = entry->setup(priv); entry 292 drivers/net/ethernet/stmicro/stmmac/hwif.c priv->hwif_quirks = entry->quirks; entry 22 drivers/net/ethernet/stmicro/stmmac/ring_mode.c unsigned int entry = tx_q->cur_tx; entry 27 drivers/net/ethernet/stmicro/stmmac/ring_mode.c desc = (struct dma_desc *)(tx_q->dma_etx + entry); entry 29 drivers/net/ethernet/stmicro/stmmac/ring_mode.c desc = tx_q->dma_tx + entry; entry 46 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].buf = des2; entry 47 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].len = bmax; entry 48 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].is_jumbo = true; entry 53 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff[entry] = NULL; entry 54 drivers/net/ethernet/stmicro/stmmac/ring_mode.c entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE); entry 57 drivers/net/ethernet/stmicro/stmmac/ring_mode.c desc = (struct dma_desc *)(tx_q->dma_etx + entry); entry 59 drivers/net/ethernet/stmicro/stmmac/ring_mode.c desc = tx_q->dma_tx + entry; entry 66 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].buf = des2; entry 67 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].len = len; entry 68 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].is_jumbo = true; entry 80 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].buf = des2; entry 81 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].len = nopaged_len; entry 82 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->tx_skbuff_dma[entry].is_jumbo = true; entry 89 drivers/net/ethernet/stmicro/stmmac/ring_mode.c tx_q->cur_tx = entry; entry 91 drivers/net/ethernet/stmicro/stmmac/ring_mode.c return entry; entry 124 drivers/net/ethernet/stmicro/stmmac/ring_mode.c unsigned int entry = tx_q->dirty_tx; entry 127 drivers/net/ethernet/stmicro/stmmac/ring_mode.c if (unlikely(tx_q->tx_skbuff_dma[entry].is_jumbo || entry 128 drivers/net/ethernet/stmicro/stmmac/ring_mode.c (tx_q->tx_skbuff_dma[entry].last_segment && entry 1874 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c unsigned int entry, count = 0; entry 1880 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry = tx_q->dirty_tx; entry 1881 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c while ((entry != tx_q->cur_tx) && (count < budget)) { entry 1882 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct sk_buff *skb = tx_q->tx_skbuff[entry]; entry 1887 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c p = (struct dma_desc *)(tx_q->dma_etx + entry); entry 1889 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c p = tx_q->dma_tx + entry; entry 1916 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c if (likely(tx_q->tx_skbuff_dma[entry].buf)) { entry 1917 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c if (tx_q->tx_skbuff_dma[entry].map_as_page) entry 1919 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].buf, entry 1920 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].len, entry 1924 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].buf, entry 1925 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].len, entry 1927 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].buf = 0; entry 1928 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].len = 0; entry 1929 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].map_as_page = false; entry 1934 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].last_segment = false; entry 1935 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].is_jumbo = false; entry 1941 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff[entry] = NULL; entry 1946 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE); entry 1948 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->dirty_tx = entry; entry 3140 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c int entry; entry 3168 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry = tx_q->cur_tx; entry 3169 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c first_entry = entry; entry 3175 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c desc = (struct dma_desc *)(tx_q->dma_etx + entry); entry 3177 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c desc = tx_q->dma_tx + entry; entry 3190 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry = stmmac_jumbo_frm(priv, tx_q, skb, csum_insertion); entry 3191 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c if (unlikely(entry < 0) && (entry != -EINVAL)) entry 3200 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE); entry 3201 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c WARN_ON(tx_q->tx_skbuff[entry]); entry 3204 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c desc = (struct dma_desc *)(tx_q->dma_etx + entry); entry 3206 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c desc = tx_q->dma_tx + entry; entry 3213 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].buf = des; entry 3217 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].map_as_page = true; entry 3218 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].len = len; entry 3219 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff_dma[entry].last_segment = last_segment; entry 3227 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->tx_skbuff[entry] = skb; entry 3241 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c desc = &tx_q->dma_etx[entry].basic; entry 3243 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c desc = &tx_q->dma_tx[entry]; entry 3255 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry = STMMAC_GET_ENTRY(entry, DMA_TX_SIZE); entry 3256 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c tx_q->cur_tx = entry; entry 3264 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry, first, nfrags); entry 3388 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c unsigned int entry = rx_q->dirty_rx; entry 3393 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c struct stmmac_rx_buffer *buf = &rx_q->buf_pool[entry]; entry 3398 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c p = (struct dma_desc *)(rx_q->dma_erx + entry); entry 3400 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c p = rx_q->dma_rx + entry; entry 3440 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry = STMMAC_GET_ENTRY(entry, DMA_RX_SIZE); entry 3442 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c rx_q->dirty_rx = entry; entry 3482 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c int entry; entry 3501 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c entry = next_entry; entry 3502 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c buf = &rx_q->buf_pool[entry]; entry 3505 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c p = (struct dma_desc *)(rx_q->dma_erx + entry); entry 3507 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c p = rx_q->dma_rx + entry; entry 3529 drivers/net/ethernet/stmicro/stmmac/stmmac_main.c &priv->xstats, rx_q->dma_erx + entry); entry 14 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c static void tc_fill_all_pass_entry(struct stmmac_tc_entry *entry) entry 16 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c memset(entry, 0, sizeof(*entry)); entry 17 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->in_use = true; entry 18 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->is_last = true; entry 19 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->is_frag = false; entry 20 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->prio = ~0x0; entry 21 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->handle = 0; entry 22 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.match_data = 0x0; entry 23 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.match_en = 0x0; entry 24 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.af = 1; entry 25 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.dma_ch_no = 0x0; entry 32 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_tc_entry *entry, *first = NULL, *dup = NULL; entry 37 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry = &priv->tc_entries[i]; entry 38 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (!entry->in_use && !first && free) entry 39 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c first = entry; entry 40 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if ((entry->handle == loc) && !free && !entry->is_frag) entry 41 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c dup = entry; entry 57 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c static int tc_fill_actions(struct stmmac_tc_entry *entry, entry 61 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_tc_entry *action_entry = entry; entry 94 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_tc_entry *entry, *frag = NULL; entry 124 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry = tc_find_entry(priv, cls, true); entry 125 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (!entry) entry 135 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->frag_ptr = frag; entry 136 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.match_en = (mask << (rem * 8)) & entry 138 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.match_data = (data << (rem * 8)) & entry 140 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.frame_offset = real_off; entry 141 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->prio = prio; entry 151 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->frag_ptr = NULL; entry 152 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.match_en = mask; entry 153 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.match_data = data; entry 154 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->val.frame_offset = real_off; entry 155 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->prio = prio; entry 158 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = tc_fill_actions(entry, frag, cls); entry 167 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->in_use = false; entry 174 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_tc_entry *entry; entry 176 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry = tc_find_entry(priv, cls, false); entry 177 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (!entry) entry 180 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->in_use = false; entry 181 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (entry->frag_ptr) { entry 182 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry = entry->frag_ptr; entry 183 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->is_frag = false; entry 184 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->in_use = false; entry 372 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_flow_entry *entry) entry 383 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->action |= STMMAC_FLOW_ACTION_DROP; entry 396 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_flow_entry *entry) entry 407 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->ip_proto = match.key->ip_proto; entry 413 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_flow_entry *entry) entry 417 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c bool inv = entry->action & STMMAC_FLOW_ACTION_DROP; entry 429 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, true, entry 437 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, true, entry 448 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_flow_entry *entry) entry 452 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c bool inv = entry->action & STMMAC_FLOW_ACTION_DROP; entry 462 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c switch (entry->ip_proto) { entry 477 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, true, entry 485 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, true, entry 491 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->is_l4 = true; entry 502 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_flow_entry *entry = &priv->flow_entries[i]; entry 504 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (entry->cookie == cls->cookie) entry 505 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c return entry; entry 506 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (get_free && (entry->in_use == false)) entry 507 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c return entry; entry 515 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_flow_entry *entry); entry 525 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_flow_entry *entry = tc_find_flow(priv, cls, false); entry 529 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (!entry) { entry 530 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry = tc_find_flow(priv, cls, true); entry 531 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (!entry) entry 535 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = tc_parse_flow_actions(priv, &rule->action, entry); entry 540 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = tc_flow_parsers[i].fn(priv, cls, entry); entry 542 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->in_use = true; entry 547 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (!entry->in_use) entry 550 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->cookie = cls->cookie; entry 557 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c struct stmmac_flow_entry *entry = tc_find_flow(priv, cls, false); entry 560 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (!entry || !entry->in_use) entry 563 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c if (entry->is_l4) { entry 564 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, false, entry 567 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, false, entry 571 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->in_use = false; entry 572 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->cookie = 0; entry 573 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c entry->is_l4 = false; entry 623 drivers/net/ethernet/sun/cassini.c struct list_head *entry; entry 640 drivers/net/ethernet/sun/cassini.c entry = cp->rx_spare_list.next; entry 641 drivers/net/ethernet/sun/cassini.c list_del(entry); entry 656 drivers/net/ethernet/sun/cassini.c return list_entry(entry, cas_page_t, list); entry 1848 drivers/net/ethernet/sun/cassini.c int entry, count; entry 1853 drivers/net/ethernet/sun/cassini.c entry = cp->tx_old[ring]; entry 1855 drivers/net/ethernet/sun/cassini.c count = TX_BUFF_COUNT(ring, entry, limit); entry 1856 drivers/net/ethernet/sun/cassini.c while (entry != limit) { entry 1857 drivers/net/ethernet/sun/cassini.c struct sk_buff *skb = skbs[entry]; entry 1864 drivers/net/ethernet/sun/cassini.c entry = TX_DESC_NEXT(ring, entry); entry 1870 drivers/net/ethernet/sun/cassini.c + cp->tx_tiny_use[ring][entry].nbufs + 1; entry 1875 drivers/net/ethernet/sun/cassini.c "tx[%d] done, slot %d\n", ring, entry); entry 1877 drivers/net/ethernet/sun/cassini.c skbs[entry] = NULL; entry 1878 drivers/net/ethernet/sun/cassini.c cp->tx_tiny_use[ring][entry].nbufs = 0; entry 1881 drivers/net/ethernet/sun/cassini.c struct cas_tx_desc *txd = txds + entry; entry 1888 drivers/net/ethernet/sun/cassini.c entry = TX_DESC_NEXT(ring, entry); entry 1891 drivers/net/ethernet/sun/cassini.c if (cp->tx_tiny_use[ring][entry].used) { entry 1892 drivers/net/ethernet/sun/cassini.c cp->tx_tiny_use[ring][entry].used = 0; entry 1893 drivers/net/ethernet/sun/cassini.c entry = TX_DESC_NEXT(ring, entry); entry 1903 drivers/net/ethernet/sun/cassini.c cp->tx_old[ring] = entry; entry 1942 drivers/net/ethernet/sun/cassini.c int entry, const u64 *words, entry 2181 drivers/net/ethernet/sun/cassini.c int entry; entry 2183 drivers/net/ethernet/sun/cassini.c entry = cp->rx_old[ring]; entry 2186 drivers/net/ethernet/sun/cassini.c cp->init_rxds[ring][entry].buffer = cpu_to_le64(new->dma_addr); entry 2187 drivers/net/ethernet/sun/cassini.c cp->init_rxds[ring][entry].index = entry 2191 drivers/net/ethernet/sun/cassini.c entry = RX_DESC_ENTRY(ring, entry + 1); entry 2192 drivers/net/ethernet/sun/cassini.c cp->rx_old[ring] = entry; entry 2194 drivers/net/ethernet/sun/cassini.c if (entry % 4) entry 2198 drivers/net/ethernet/sun/cassini.c writel(entry, cp->regs + REG_RX_KICK); entry 2201 drivers/net/ethernet/sun/cassini.c writel(entry, cp->regs + REG_PLUS_RX_KICK1); entry 2208 drivers/net/ethernet/sun/cassini.c unsigned int entry, last, count, released; entry 2212 drivers/net/ethernet/sun/cassini.c entry = cp->rx_old[ring]; entry 2215 drivers/net/ethernet/sun/cassini.c "rxd[%d] interrupt, done: %d\n", ring, entry); entry 2218 drivers/net/ethernet/sun/cassini.c count = entry & 0x3; entry 2219 drivers/net/ethernet/sun/cassini.c last = RX_DESC_ENTRY(ring, num ? entry + num - 4: entry - 4); entry 2221 drivers/net/ethernet/sun/cassini.c while (entry != last) { entry 2223 drivers/net/ethernet/sun/cassini.c if (page_count(page[entry]->buffer) > 1) { entry 2233 drivers/net/ethernet/sun/cassini.c cp->rx_old[ring] = entry; entry 2238 drivers/net/ethernet/sun/cassini.c list_add(&page[entry]->list, &cp->rx_inuse_list); entry 2240 drivers/net/ethernet/sun/cassini.c cp->init_rxds[ring][entry].buffer = entry 2242 drivers/net/ethernet/sun/cassini.c page[entry] = new; entry 2247 drivers/net/ethernet/sun/cassini.c cluster = entry; entry 2251 drivers/net/ethernet/sun/cassini.c entry = RX_DESC_ENTRY(ring, entry + 1); entry 2253 drivers/net/ethernet/sun/cassini.c cp->rx_old[ring] = entry; entry 2282 drivers/net/ethernet/sun/cassini.c int entry, drops; entry 2290 drivers/net/ethernet/sun/cassini.c entry = cp->rx_new[ring]; entry 2293 drivers/net/ethernet/sun/cassini.c struct cas_rx_comp *rxc = rxcs + entry; entry 2332 drivers/net/ethernet/sun/cassini.c len = cas_rx_process_pkt(cp, rxc, entry, words, &skb); entry 2379 drivers/net/ethernet/sun/cassini.c entry = RX_COMP_ENTRY(ring, entry + 1 + entry 2386 drivers/net/ethernet/sun/cassini.c cp->rx_new[ring] = entry; entry 2399 drivers/net/ethernet/sun/cassini.c int last, entry; entry 2402 drivers/net/ethernet/sun/cassini.c entry = cp->rx_new[ring]; entry 2405 drivers/net/ethernet/sun/cassini.c ring, readl(cp->regs + REG_RX_COMP_HEAD), entry); entry 2408 drivers/net/ethernet/sun/cassini.c while (last != entry) { entry 2716 drivers/net/ethernet/sun/cassini.c static inline int cas_intme(int ring, int entry) entry 2719 drivers/net/ethernet/sun/cassini.c if (!(entry & ((TX_DESC_RINGN_SIZE(ring) >> 1) - 1))) entry 2725 drivers/net/ethernet/sun/cassini.c static void cas_write_txd(struct cas *cp, int ring, int entry, entry 2728 drivers/net/ethernet/sun/cassini.c struct cas_tx_desc *txd = cp->init_txds[ring] + entry; entry 2731 drivers/net/ethernet/sun/cassini.c if (cas_intme(ring, entry)) entry 2740 drivers/net/ethernet/sun/cassini.c const int entry) entry 2742 drivers/net/ethernet/sun/cassini.c return cp->tx_tiny_bufs[ring] + TX_TINY_BUF_LEN*entry; entry 2746 drivers/net/ethernet/sun/cassini.c const int entry, const int tentry) entry 2749 drivers/net/ethernet/sun/cassini.c cp->tx_tiny_use[ring][entry].used = 1; entry 2750 drivers/net/ethernet/sun/cassini.c return cp->tx_tiny_dvma[ring] + TX_TINY_BUF_LEN*entry; entry 2757 drivers/net/ethernet/sun/cassini.c int entry, nr_frags, frag, tabort, tentry; entry 2784 drivers/net/ethernet/sun/cassini.c entry = cp->tx_new[ring]; entry 2785 drivers/net/ethernet/sun/cassini.c cp->tx_skbs[ring][entry] = skb; entry 2793 drivers/net/ethernet/sun/cassini.c tentry = entry; entry 2797 drivers/net/ethernet/sun/cassini.c cas_write_txd(cp, ring, entry, mapping, len - tabort, entry 2799 drivers/net/ethernet/sun/cassini.c entry = TX_DESC_NEXT(ring, entry); entry 2802 drivers/net/ethernet/sun/cassini.c tx_tiny_buf(cp, ring, entry), tabort); entry 2803 drivers/net/ethernet/sun/cassini.c mapping = tx_tiny_map(cp, ring, entry, tentry); entry 2804 drivers/net/ethernet/sun/cassini.c cas_write_txd(cp, ring, entry, mapping, tabort, ctrl, entry 2807 drivers/net/ethernet/sun/cassini.c cas_write_txd(cp, ring, entry, mapping, len, ctrl | entry 2810 drivers/net/ethernet/sun/cassini.c entry = TX_DESC_NEXT(ring, entry); entry 2824 drivers/net/ethernet/sun/cassini.c cas_write_txd(cp, ring, entry, mapping, len - tabort, entry 2826 drivers/net/ethernet/sun/cassini.c entry = TX_DESC_NEXT(ring, entry); entry 2829 drivers/net/ethernet/sun/cassini.c memcpy(tx_tiny_buf(cp, ring, entry), entry 2833 drivers/net/ethernet/sun/cassini.c mapping = tx_tiny_map(cp, ring, entry, tentry); entry 2837 drivers/net/ethernet/sun/cassini.c cas_write_txd(cp, ring, entry, mapping, len, ctrl, entry 2839 drivers/net/ethernet/sun/cassini.c entry = TX_DESC_NEXT(ring, entry); entry 2842 drivers/net/ethernet/sun/cassini.c cp->tx_new[ring] = entry; entry 2848 drivers/net/ethernet/sun/cassini.c ring, entry, skb->len, TX_BUFFS_AVAIL(cp, ring)); entry 2849 drivers/net/ethernet/sun/cassini.c writel(entry, cp->regs + REG_TX_KICKN(ring)); entry 9030 drivers/net/ethernet/sun/niu.c msi_vec[i].entry = i; entry 957 drivers/net/ethernet/sun/sunbmac.c int len, entry; entry 966 drivers/net/ethernet/sun/sunbmac.c entry = bp->tx_new; entry 967 drivers/net/ethernet/sun/sunbmac.c DTX(("bigmac_start_xmit: len(%d) entry(%d)\n", len, entry)); entry 968 drivers/net/ethernet/sun/sunbmac.c bp->bmac_block->be_txd[entry].tx_flags = TXD_UPDATE; entry 969 drivers/net/ethernet/sun/sunbmac.c bp->tx_skbs[entry] = skb; entry 970 drivers/net/ethernet/sun/sunbmac.c bp->bmac_block->be_txd[entry].tx_addr = mapping; entry 971 drivers/net/ethernet/sun/sunbmac.c bp->bmac_block->be_txd[entry].tx_flags = entry 973 drivers/net/ethernet/sun/sunbmac.c bp->tx_new = NEXT_TX(entry); entry 645 drivers/net/ethernet/sun/sungem.c int entry, limit; entry 647 drivers/net/ethernet/sun/sungem.c entry = gp->tx_old; entry 649 drivers/net/ethernet/sun/sungem.c while (entry != limit) { entry 658 drivers/net/ethernet/sun/sungem.c gp->dev->name, entry); entry 659 drivers/net/ethernet/sun/sungem.c skb = gp->tx_skbs[entry]; entry 661 drivers/net/ethernet/sun/sungem.c int last = entry + skb_shinfo(skb)->nr_frags; entry 662 drivers/net/ethernet/sun/sungem.c int walk = entry; entry 676 drivers/net/ethernet/sun/sungem.c gp->tx_skbs[entry] = NULL; entry 680 drivers/net/ethernet/sun/sungem.c txd = &gp->init_block->txd[entry]; entry 686 drivers/net/ethernet/sun/sungem.c entry = NEXT_TX(entry); entry 692 drivers/net/ethernet/sun/sungem.c gp->tx_old = entry; entry 760 drivers/net/ethernet/sun/sungem.c int entry, drops, work_done = 0; entry 767 drivers/net/ethernet/sun/sungem.c entry = gp->rx_new; entry 771 drivers/net/ethernet/sun/sungem.c struct gem_rxd *rxd = &gp->init_block->rxd[entry]; entry 790 drivers/net/ethernet/sun/sungem.c if (entry == done) { entry 792 drivers/net/ethernet/sun/sungem.c if (entry == done) entry 799 drivers/net/ethernet/sun/sungem.c skb = gp->rx_skbs[entry]; entry 827 drivers/net/ethernet/sun/sungem.c gp->rx_skbs[entry] = new_skb; entry 871 drivers/net/ethernet/sun/sungem.c entry = NEXT_RX(entry); entry 874 drivers/net/ethernet/sun/sungem.c gem_post_rxds(gp, entry); entry 876 drivers/net/ethernet/sun/sungem.c gp->rx_new = entry; entry 991 drivers/net/ethernet/sun/sungem.c static __inline__ int gem_intme(int entry) entry 994 drivers/net/ethernet/sun/sungem.c if (!(entry & ((TX_RING_SIZE>>1)-1))) entry 1004 drivers/net/ethernet/sun/sungem.c int entry; entry 1026 drivers/net/ethernet/sun/sungem.c entry = gp->tx_new; entry 1027 drivers/net/ethernet/sun/sungem.c gp->tx_skbs[entry] = skb; entry 1030 drivers/net/ethernet/sun/sungem.c struct gem_txd *txd = &gp->init_block->txd[entry]; entry 1040 drivers/net/ethernet/sun/sungem.c if (gem_intme(entry)) entry 1045 drivers/net/ethernet/sun/sungem.c entry = NEXT_TX(entry); entry 1051 drivers/net/ethernet/sun/sungem.c int frag, first_entry = entry; entry 1054 drivers/net/ethernet/sun/sungem.c if (gem_intme(entry)) entry 1064 drivers/net/ethernet/sun/sungem.c entry = NEXT_TX(entry); entry 1079 drivers/net/ethernet/sun/sungem.c txd = &gp->init_block->txd[entry]; entry 1084 drivers/net/ethernet/sun/sungem.c if (gem_intme(entry)) entry 1087 drivers/net/ethernet/sun/sungem.c entry = NEXT_TX(entry); entry 1096 drivers/net/ethernet/sun/sungem.c gp->tx_new = entry; entry 1111 drivers/net/ethernet/sun/sungem.c dev->name, entry, skb->len); entry 2268 drivers/net/ethernet/sun/sunhme.c u32 first_len, u32 first_entry, u32 entry) entry 2275 drivers/net/ethernet/sun/sunhme.c while (first_entry != entry) { entry 2290 drivers/net/ethernet/sun/sunhme.c int entry; entry 2313 drivers/net/ethernet/sun/sunhme.c entry = hp->tx_new; entry 2314 drivers/net/ethernet/sun/sunhme.c SXD(("SX<l[%d]e[%d]>", len, entry)); entry 2315 drivers/net/ethernet/sun/sunhme.c hp->tx_skbs[entry] = skb; entry 2325 drivers/net/ethernet/sun/sunhme.c hme_write_txd(hp, &hp->happy_block->happy_meal_txd[entry], entry 2328 drivers/net/ethernet/sun/sunhme.c entry = NEXT_TX(entry); entry 2331 drivers/net/ethernet/sun/sunhme.c int frag, first_entry = entry; entry 2341 drivers/net/ethernet/sun/sunhme.c entry = NEXT_TX(entry); entry 2352 drivers/net/ethernet/sun/sunhme.c first_entry, entry); entry 2358 drivers/net/ethernet/sun/sunhme.c hme_write_txd(hp, &hp->happy_block->happy_meal_txd[entry], entry 2361 drivers/net/ethernet/sun/sunhme.c entry = NEXT_TX(entry); entry 2368 drivers/net/ethernet/sun/sunhme.c hp->tx_new = entry; entry 579 drivers/net/ethernet/sun/sunqe.c int len, entry; entry 586 drivers/net/ethernet/sun/sunqe.c entry = qep->tx_new; entry 588 drivers/net/ethernet/sun/sunqe.c txbuf = &qbufs->tx_buf[entry & (TX_RING_SIZE - 1)][0]; entry 590 drivers/net/ethernet/sun/sunqe.c qebuf_offset(tx_buf, (entry & (TX_RING_SIZE - 1))); entry 593 drivers/net/ethernet/sun/sunqe.c qep->qe_block->qe_txd[entry].tx_flags = TXD_UPDATE; entry 597 drivers/net/ethernet/sun/sunqe.c qep->qe_block->qe_txd[entry].tx_addr = txbuf_dvma; entry 598 drivers/net/ethernet/sun/sunqe.c qep->qe_block->qe_txd[entry].tx_flags = entry 600 drivers/net/ethernet/sun/sunqe.c qep->tx_new = NEXT_TX(entry); entry 474 drivers/net/ethernet/ti/netcp_core.c struct netcp_hook_list *entry; entry 478 drivers/net/ethernet/ti/netcp_core.c entry = devm_kzalloc(netcp_priv->dev, sizeof(*entry), GFP_KERNEL); entry 479 drivers/net/ethernet/ti/netcp_core.c if (!entry) entry 482 drivers/net/ethernet/ti/netcp_core.c entry->hook_rtn = hook_rtn; entry 483 drivers/net/ethernet/ti/netcp_core.c entry->hook_data = hook_data; entry 484 drivers/net/ethernet/ti/netcp_core.c entry->order = order; entry 491 drivers/net/ethernet/ti/netcp_core.c __list_add(&entry->list, next->list.prev, &next->list); entry 523 drivers/net/ethernet/ti/netcp_core.c struct netcp_hook_list *entry; entry 527 drivers/net/ethernet/ti/netcp_core.c entry = devm_kzalloc(netcp_priv->dev, sizeof(*entry), GFP_KERNEL); entry 528 drivers/net/ethernet/ti/netcp_core.c if (!entry) entry 531 drivers/net/ethernet/ti/netcp_core.c entry->hook_rtn = hook_rtn; entry 532 drivers/net/ethernet/ti/netcp_core.c entry->hook_data = hook_data; entry 533 drivers/net/ethernet/ti/netcp_core.c entry->order = order; entry 540 drivers/net/ethernet/ti/netcp_core.c __list_add(&entry->list, next->list.prev, &next->list); entry 1246 drivers/net/ethernet/via/via-rhine.c struct rhine_skb_dma *sd, int entry) entry 1248 drivers/net/ethernet/via/via-rhine.c rp->rx_skbuff_dma[entry] = sd->dma; entry 1249 drivers/net/ethernet/via/via-rhine.c rp->rx_skbuff[entry] = sd->skb; entry 1251 drivers/net/ethernet/via/via-rhine.c rp->rx_ring[entry].addr = cpu_to_le32(sd->dma); entry 1787 drivers/net/ethernet/via/via-rhine.c unsigned entry; entry 1793 drivers/net/ethernet/via/via-rhine.c entry = rp->cur_tx % TX_RING_SIZE; entry 1798 drivers/net/ethernet/via/via-rhine.c rp->tx_skbuff[entry] = skb; entry 1806 drivers/net/ethernet/via/via-rhine.c rp->tx_skbuff[entry] = NULL; entry 1812 drivers/net/ethernet/via/via-rhine.c skb_copy_and_csum_dev(skb, rp->tx_buf[entry]); entry 1814 drivers/net/ethernet/via/via-rhine.c memset(rp->tx_buf[entry] + skb->len, 0, entry 1816 drivers/net/ethernet/via/via-rhine.c rp->tx_skbuff_dma[entry] = 0; entry 1817 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[entry].addr = cpu_to_le32(rp->tx_bufs_dma + entry 1818 drivers/net/ethernet/via/via-rhine.c (rp->tx_buf[entry] - entry 1821 drivers/net/ethernet/via/via-rhine.c rp->tx_skbuff_dma[entry] = entry 1824 drivers/net/ethernet/via/via-rhine.c if (dma_mapping_error(hwdev, rp->tx_skbuff_dma[entry])) { entry 1826 drivers/net/ethernet/via/via-rhine.c rp->tx_skbuff_dma[entry] = 0; entry 1830 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[entry].addr = cpu_to_le32(rp->tx_skbuff_dma[entry]); entry 1833 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[entry].desc_length = entry 1842 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[entry].tx_status = cpu_to_le32((vid_pcp) << 16); entry 1844 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[entry].desc_length |= cpu_to_le32(0x020000); entry 1847 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[entry].tx_status = 0; entry 1852 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[entry].tx_status |= cpu_to_le32(DescOwn); entry 1884 drivers/net/ethernet/via/via-rhine.c rp->cur_tx - 1, entry); entry 1943 drivers/net/ethernet/via/via-rhine.c unsigned int entry = dirty_tx % TX_RING_SIZE; entry 1944 drivers/net/ethernet/via/via-rhine.c u32 txstatus = le32_to_cpu(rp->tx_ring[entry].tx_status); entry 1947 drivers/net/ethernet/via/via-rhine.c entry, txstatus); entry 1950 drivers/net/ethernet/via/via-rhine.c skb = rp->tx_skbuff[entry]; entry 1966 drivers/net/ethernet/via/via-rhine.c rp->tx_ring[entry].tx_status = cpu_to_le32(DescOwn); entry 1984 drivers/net/ethernet/via/via-rhine.c if (rp->tx_skbuff_dma[entry]) { entry 1986 drivers/net/ethernet/via/via-rhine.c rp->tx_skbuff_dma[entry], entry 1993 drivers/net/ethernet/via/via-rhine.c rp->tx_skbuff[entry] = NULL; entry 2045 drivers/net/ethernet/via/via-rhine.c int entry = rp->cur_rx % RX_RING_SIZE; entry 2049 drivers/net/ethernet/via/via-rhine.c entry, le32_to_cpu(rp->rx_ring[entry].rx_status)); entry 2053 drivers/net/ethernet/via/via-rhine.c struct rx_desc *desc = rp->rx_ring + entry; entry 2068 drivers/net/ethernet/via/via-rhine.c entry, data_size, entry 2103 drivers/net/ethernet/via/via-rhine.c rp->rx_skbuff_dma[entry], entry 2108 drivers/net/ethernet/via/via-rhine.c rp->rx_skbuff[entry]->data, entry 2112 drivers/net/ethernet/via/via-rhine.c rp->rx_skbuff_dma[entry], entry 2121 drivers/net/ethernet/via/via-rhine.c skb = rp->rx_skbuff[entry]; entry 2124 drivers/net/ethernet/via/via-rhine.c rp->rx_skbuff_dma[entry], entry 2127 drivers/net/ethernet/via/via-rhine.c rhine_skb_dma_nic_store(rp, &sd, entry); entry 2145 drivers/net/ethernet/via/via-rhine.c entry = (++rp->cur_rx) % RX_RING_SIZE; entry 2158 drivers/net/ethernet/via/via-rhine.c int entry = rp->dirty_tx % TX_RING_SIZE; entry 2170 drivers/net/ethernet/via/via-rhine.c iowrite32(rp->tx_ring_dma + entry * sizeof(struct tx_desc), entry 2176 drivers/net/ethernet/via/via-rhine.c if (rp->tx_ring[entry].desc_length & cpu_to_le32(0x020000)) entry 2480 drivers/net/fddi/defxx.c p_addr = &bp->cmd_req_virt->addr_filter_set.entry[0]; entry 3107 drivers/net/fddi/defxx.c int entry; entry 3109 drivers/net/fddi/defxx.c entry = bp->rcv_xmt_reg.index.rcv_comp; entry 3111 drivers/net/fddi/defxx.c p_buff = (char *) (((struct sk_buff *)bp->p_rcv_buff_va[entry])->data); entry 3113 drivers/net/fddi/defxx.c p_buff = bp->p_rcv_buff_va[entry]; entry 3115 drivers/net/fddi/defxx.c dma_addr = bp->descr_block_virt->rcv_data[entry].long_1; entry 3165 drivers/net/fddi/defxx.c skb = (struct sk_buff *)bp->p_rcv_buff_va[entry]; entry 3171 drivers/net/fddi/defxx.c bp->p_rcv_buff_va[entry] = (char *)newskb; entry 3172 drivers/net/fddi/defxx.c bp->descr_block_virt->rcv_data[entry].long_1 = (u32)new_dma_addr; entry 643 drivers/net/fddi/defxx.h PI_LAN_ADDR entry[PI_CMD_ADDR_FILTER_K_SIZE]; entry 665 drivers/net/fddi/defxx.h PI_LAN_ADDR entry[PI_CMD_ADDR_FILTER_K_SIZE]; entry 137 drivers/net/macvlan.c struct macvlan_source_entry *entry; entry 141 drivers/net/macvlan.c hlist_for_each_entry_rcu(entry, h, hlist) { entry 142 drivers/net/macvlan.c if (ether_addr_equal_64bits(entry->addr, addr) && entry 143 drivers/net/macvlan.c entry->vlan == vlan) entry 144 drivers/net/macvlan.c return entry; entry 153 drivers/net/macvlan.c struct macvlan_source_entry *entry; entry 156 drivers/net/macvlan.c entry = macvlan_hash_lookup_source(vlan, addr); entry 157 drivers/net/macvlan.c if (entry) entry 160 drivers/net/macvlan.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 161 drivers/net/macvlan.c if (!entry) entry 164 drivers/net/macvlan.c ether_addr_copy(entry->addr, addr); entry 165 drivers/net/macvlan.c entry->vlan = vlan; entry 167 drivers/net/macvlan.c hlist_add_head_rcu(&entry->hlist, h); entry 182 drivers/net/macvlan.c static void macvlan_hash_del_source(struct macvlan_source_entry *entry) entry 184 drivers/net/macvlan.c hlist_del_rcu(&entry->hlist); entry 185 drivers/net/macvlan.c kfree_rcu(entry, rcu); entry 386 drivers/net/macvlan.c struct macvlan_source_entry *entry; entry 388 drivers/net/macvlan.c entry = hlist_entry(h, struct macvlan_source_entry, entry 390 drivers/net/macvlan.c if (entry->vlan == vlan) entry 391 drivers/net/macvlan.c macvlan_hash_del_source(entry); entry 427 drivers/net/macvlan.c struct macvlan_source_entry *entry; entry 431 drivers/net/macvlan.c hlist_for_each_entry_rcu(entry, h, hlist) { entry 432 drivers/net/macvlan.c if (ether_addr_equal_64bits(entry->addr, addr)) entry 433 drivers/net/macvlan.c macvlan_forward_source_one(skb, entry->vlan); entry 1323 drivers/net/macvlan.c struct macvlan_source_entry *entry; entry 1338 drivers/net/macvlan.c entry = macvlan_hash_lookup_source(vlan, addr); entry 1339 drivers/net/macvlan.c if (entry) { entry 1340 drivers/net/macvlan.c macvlan_hash_del_source(entry); entry 1586 drivers/net/macvlan.c struct macvlan_source_entry *entry; entry 1588 drivers/net/macvlan.c hlist_for_each_entry_rcu(entry, h, hlist) { entry 1589 drivers/net/macvlan.c if (entry->vlan != vlan) entry 1591 drivers/net/macvlan.c if (nla_put(skb, IFLA_MACVLAN_MACADDR, ETH_ALEN, entry->addr)) entry 47 drivers/net/netdevsim/bpf.c } entry[NSIM_BPF_MAX_KEYS]; entry 338 drivers/net/netdevsim/bpf.c for (i = 0; i < ARRAY_SIZE(nmap->entry); i++) entry 339 drivers/net/netdevsim/bpf.c if (nsim_map_key_match(&offmap->map, &nmap->entry[i], key)) entry 350 drivers/net/netdevsim/bpf.c nmap->entry[idx].key = kmalloc(offmap->map.key_size, GFP_USER); entry 351 drivers/net/netdevsim/bpf.c if (!nmap->entry[idx].key) entry 353 drivers/net/netdevsim/bpf.c nmap->entry[idx].value = kmalloc(offmap->map.value_size, GFP_USER); entry 354 drivers/net/netdevsim/bpf.c if (!nmap->entry[idx].value) { entry 355 drivers/net/netdevsim/bpf.c kfree(nmap->entry[idx].key); entry 356 drivers/net/netdevsim/bpf.c nmap->entry[idx].key = NULL; entry 379 drivers/net/netdevsim/bpf.c for (; idx < ARRAY_SIZE(nmap->entry); idx++) { entry 380 drivers/net/netdevsim/bpf.c if (nmap->entry[idx].key) { entry 381 drivers/net/netdevsim/bpf.c memcpy(next_key, nmap->entry[idx].key, entry 389 drivers/net/netdevsim/bpf.c if (idx == ARRAY_SIZE(nmap->entry)) entry 404 drivers/net/netdevsim/bpf.c memcpy(value, nmap->entry[idx].value, offmap->map.value_size); entry 431 drivers/net/netdevsim/bpf.c for (idx = 0; idx < ARRAY_SIZE(nmap->entry); idx++) entry 432 drivers/net/netdevsim/bpf.c if (!nmap->entry[idx].key) entry 434 drivers/net/netdevsim/bpf.c if (idx == ARRAY_SIZE(nmap->entry)) { entry 444 drivers/net/netdevsim/bpf.c memcpy(nmap->entry[idx].key, key, offmap->map.key_size); entry 445 drivers/net/netdevsim/bpf.c memcpy(nmap->entry[idx].value, value, offmap->map.value_size); entry 464 drivers/net/netdevsim/bpf.c kfree(nmap->entry[idx].key); entry 465 drivers/net/netdevsim/bpf.c kfree(nmap->entry[idx].value); entry 466 drivers/net/netdevsim/bpf.c memset(&nmap->entry[idx], 0, sizeof(nmap->entry[idx])); entry 505 drivers/net/netdevsim/bpf.c for (i = 0; i < ARRAY_SIZE(nmap->entry); i++) { entry 511 drivers/net/netdevsim/bpf.c key = nmap->entry[i].key; entry 523 drivers/net/netdevsim/bpf.c kfree(nmap->entry[i].key); entry 524 drivers/net/netdevsim/bpf.c kfree(nmap->entry[i].value); entry 535 drivers/net/netdevsim/bpf.c for (i = 0; i < ARRAY_SIZE(nmap->entry); i++) { entry 536 drivers/net/netdevsim/bpf.c kfree(nmap->entry[i].key); entry 537 drivers/net/netdevsim/bpf.c kfree(nmap->entry[i].value); entry 45 drivers/net/netdevsim/fib.c struct nsim_fib_entry *entry; entry 49 drivers/net/netdevsim/fib.c entry = &fib_data->ipv4.fib; entry 52 drivers/net/netdevsim/fib.c entry = &fib_data->ipv4.rules; entry 55 drivers/net/netdevsim/fib.c entry = &fib_data->ipv6.fib; entry 58 drivers/net/netdevsim/fib.c entry = &fib_data->ipv6.rules; entry 64 drivers/net/netdevsim/fib.c return max ? entry->max : entry->num; entry 71 drivers/net/netdevsim/fib.c struct nsim_fib_entry *entry; entry 76 drivers/net/netdevsim/fib.c entry = &fib_data->ipv4.fib; entry 79 drivers/net/netdevsim/fib.c entry = &fib_data->ipv4.rules; entry 82 drivers/net/netdevsim/fib.c entry = &fib_data->ipv6.fib; entry 85 drivers/net/netdevsim/fib.c entry = &fib_data->ipv6.rules; entry 94 drivers/net/netdevsim/fib.c if (val < entry->num) { entry 98 drivers/net/netdevsim/fib.c entry->max = val; entry 104 drivers/net/netdevsim/fib.c static int nsim_fib_rule_account(struct nsim_fib_entry *entry, bool add, entry 110 drivers/net/netdevsim/fib.c if (entry->num < entry->max) { entry 111 drivers/net/netdevsim/fib.c entry->num++; entry 117 drivers/net/netdevsim/fib.c entry->num--; entry 141 drivers/net/netdevsim/fib.c static int nsim_fib_account(struct nsim_fib_entry *entry, bool add, entry 147 drivers/net/netdevsim/fib.c if (entry->num < entry->max) { entry 148 drivers/net/netdevsim/fib.c entry->num++; entry 154 drivers/net/netdevsim/fib.c entry->num--; entry 2239 drivers/net/usb/lan78xx.c struct skb_data *entry; entry 2244 drivers/net/usb/lan78xx.c entry = (struct skb_data *)skb->cb; entry 2245 drivers/net/usb/lan78xx.c if (entry->state != unlink_start) entry 2250 drivers/net/usb/lan78xx.c entry->state = unlink_start; entry 2251 drivers/net/usb/lan78xx.c urb = entry->urb; entry 2775 drivers/net/usb/lan78xx.c struct skb_data *entry = (struct skb_data *)skb->cb; entry 2778 drivers/net/usb/lan78xx.c old_state = entry->state; entry 2779 drivers/net/usb/lan78xx.c entry->state = state; entry 2796 drivers/net/usb/lan78xx.c struct skb_data *entry = (struct skb_data *)skb->cb; entry 2797 drivers/net/usb/lan78xx.c struct lan78xx_net *dev = entry->dev; entry 2800 drivers/net/usb/lan78xx.c dev->net->stats.tx_packets += entry->num_of_packet; entry 2801 drivers/net/usb/lan78xx.c dev->net->stats.tx_bytes += entry->length; entry 2822 drivers/net/usb/lan78xx.c "tx err %d\n", entry->urb->status); entry 2835 drivers/net/usb/lan78xx.c struct skb_data *entry = (struct skb_data *)newsk->cb; entry 2838 drivers/net/usb/lan78xx.c entry->state = state; entry 3190 drivers/net/usb/lan78xx.c struct skb_data *entry; entry 3201 drivers/net/usb/lan78xx.c entry = (struct skb_data *)skb->cb; entry 3202 drivers/net/usb/lan78xx.c entry->urb = urb; entry 3203 drivers/net/usb/lan78xx.c entry->dev = dev; entry 3204 drivers/net/usb/lan78xx.c entry->length = 0; entry 3250 drivers/net/usb/lan78xx.c struct skb_data *entry = (struct skb_data *)skb->cb; entry 3251 drivers/net/usb/lan78xx.c struct lan78xx_net *dev = entry->dev; entry 3257 drivers/net/usb/lan78xx.c entry->urb = NULL; entry 3279 drivers/net/usb/lan78xx.c entry->urb = urb; entry 3287 drivers/net/usb/lan78xx.c entry->urb = urb; entry 3321 drivers/net/usb/lan78xx.c struct skb_data *entry; entry 3376 drivers/net/usb/lan78xx.c entry = (struct skb_data *)skb->cb; entry 3377 drivers/net/usb/lan78xx.c entry->urb = urb; entry 3378 drivers/net/usb/lan78xx.c entry->dev = dev; entry 3379 drivers/net/usb/lan78xx.c entry->length = length; entry 3380 drivers/net/usb/lan78xx.c entry->num_of_packet = count; entry 3471 drivers/net/usb/lan78xx.c struct skb_data *entry; entry 3474 drivers/net/usb/lan78xx.c entry = (struct skb_data *)(skb->cb); entry 3475 drivers/net/usb/lan78xx.c switch (entry->state) { entry 3477 drivers/net/usb/lan78xx.c entry->state = rx_cleanup; entry 3481 drivers/net/usb/lan78xx.c usb_free_urb(entry->urb); entry 3485 drivers/net/usb/lan78xx.c usb_free_urb(entry->urb); entry 3489 drivers/net/usb/lan78xx.c netdev_dbg(dev->net, "skb state %d\n", entry->state); entry 411 drivers/net/usb/usbnet.c struct skb_data *entry = (struct skb_data *) newsk->cb; entry 414 drivers/net/usb/usbnet.c entry->state = state; entry 428 drivers/net/usb/usbnet.c struct skb_data *entry = (struct skb_data *) skb->cb; entry 431 drivers/net/usb/usbnet.c old_state = entry->state; entry 432 drivers/net/usb/usbnet.c entry->state = state; entry 471 drivers/net/usb/usbnet.c struct skb_data *entry; entry 493 drivers/net/usb/usbnet.c entry = (struct skb_data *) skb->cb; entry 494 drivers/net/usb/usbnet.c entry->urb = urb; entry 495 drivers/net/usb/usbnet.c entry->dev = dev; entry 496 drivers/net/usb/usbnet.c entry->length = 0; entry 578 drivers/net/usb/usbnet.c struct skb_data *entry = (struct skb_data *) skb->cb; entry 579 drivers/net/usb/usbnet.c struct usbnet *dev = entry->dev; entry 585 drivers/net/usb/usbnet.c entry->urb = NULL; entry 624 drivers/net/usb/usbnet.c entry->urb = urb; entry 712 drivers/net/usb/usbnet.c struct skb_data *entry; entry 717 drivers/net/usb/usbnet.c entry = (struct skb_data *) skb->cb; entry 718 drivers/net/usb/usbnet.c if (entry->state != unlink_start) entry 723 drivers/net/usb/usbnet.c entry->state = unlink_start; entry 724 drivers/net/usb/usbnet.c urb = entry->urb; entry 1244 drivers/net/usb/usbnet.c struct skb_data *entry = (struct skb_data *) skb->cb; entry 1245 drivers/net/usb/usbnet.c struct usbnet *dev = entry->dev; entry 1252 drivers/net/usb/usbnet.c stats64->tx_packets += entry->packets; entry 1253 drivers/net/usb/usbnet.c stats64->tx_bytes += entry->length; entry 1285 drivers/net/usb/usbnet.c "tx err %d\n", entry->urb->status); entry 1352 drivers/net/usb/usbnet.c struct skb_data *entry; entry 1378 drivers/net/usb/usbnet.c entry = (struct skb_data *) skb->cb; entry 1379 drivers/net/usb/usbnet.c entry->urb = urb; entry 1380 drivers/net/usb/usbnet.c entry->dev = dev; entry 1418 drivers/net/usb/usbnet.c entry->length += length; entry 1419 drivers/net/usb/usbnet.c if (WARN_ON_ONCE(entry->length <= 0)) entry 1420 drivers/net/usb/usbnet.c entry->length = length; entry 1521 drivers/net/usb/usbnet.c struct skb_data *entry; entry 1524 drivers/net/usb/usbnet.c entry = (struct skb_data *) skb->cb; entry 1525 drivers/net/usb/usbnet.c switch (entry->state) { entry 1527 drivers/net/usb/usbnet.c entry->state = rx_cleanup; entry 1531 drivers/net/usb/usbnet.c kfree(entry->urb->sg); entry 1534 drivers/net/usb/usbnet.c usb_free_urb (entry->urb); entry 1538 drivers/net/usb/usbnet.c netdev_dbg(dev->net, "bogus skb state %d\n", entry->state); entry 3138 drivers/net/vmxnet3/vmxnet3_drv.c adapter->intr.msix_entries[i].entry = i; entry 426 drivers/net/wan/lapbether.c struct list_head *entry, *tmp; entry 432 drivers/net/wan/lapbether.c list_for_each_safe(entry, tmp, &lapbeth_devices) { entry 433 drivers/net/wan/lapbether.c lapbeth = list_entry(entry, struct lapbethdev, node); entry 1401 drivers/net/wan/lmc/lmc_main.c int entry; entry 1410 drivers/net/wan/lmc/lmc_main.c entry = sc->lmc_next_tx % LMC_TXDESCS; entry 1412 drivers/net/wan/lmc/lmc_main.c sc->lmc_txq[entry] = skb; entry 1413 drivers/net/wan/lmc/lmc_main.c sc->lmc_txring[entry].buffer1 = virt_to_bus (skb->data); entry 1452 drivers/net/wan/lmc/lmc_main.c LMC_EVENT_LOG(LMC_EVENT_TBUSY1, entry, 0); entry 1457 drivers/net/wan/lmc/lmc_main.c if (entry == LMC_TXDESCS - 1) /* last descriptor in ring */ entry 1461 drivers/net/wan/lmc/lmc_main.c flag = sc->lmc_txring[entry].length = (skb->len) | flag | entry 1472 drivers/net/wan/lmc/lmc_main.c LMC_EVENT_LOG(LMC_EVENT_XMT, flag, entry); entry 1473 drivers/net/wan/lmc/lmc_main.c sc->lmc_txring[entry].status = 0x80000000; entry 546 drivers/net/wimax/i2400m/rx.c } entry[I2400M_ROQ_LOG_LENGTH]; entry 601 drivers/net/wimax/i2400m/rx.c e = &roq->log->entry[cnt_idx]; entry 626 drivers/net/wimax/i2400m/rx.c e = &roq->log->entry[cnt_idx]; entry 311 drivers/net/wireless/admtek/adm8211.c unsigned int entry = dirty_tx % priv->tx_ring_size; entry 312 drivers/net/wireless/admtek/adm8211.c u32 status = le32_to_cpu(priv->tx_ring[entry].status); entry 321 drivers/net/wireless/admtek/adm8211.c info = &priv->tx_buffers[entry]; entry 354 drivers/net/wireless/admtek/adm8211.c unsigned int entry = priv->cur_rx % priv->rx_ring_size; entry 361 drivers/net/wireless/admtek/adm8211.c while (!(priv->rx_ring[entry].status & cpu_to_le32(RDES0_STATUS_OWN))) { entry 365 drivers/net/wireless/admtek/adm8211.c status = le32_to_cpu(priv->rx_ring[entry].status); entry 367 drivers/net/wireless/admtek/adm8211.c rssi = le32_to_cpu(priv->rx_ring[entry].length) & entry 387 drivers/net/wireless/admtek/adm8211.c priv->rx_buffers[entry].mapping, entry 390 drivers/net/wireless/admtek/adm8211.c skb_tail_pointer(priv->rx_buffers[entry].skb), entry 394 drivers/net/wireless/admtek/adm8211.c priv->rx_buffers[entry].mapping, entry 400 drivers/net/wireless/admtek/adm8211.c skb = priv->rx_buffers[entry].skb; entry 404 drivers/net/wireless/admtek/adm8211.c priv->rx_buffers[entry].mapping, entry 406 drivers/net/wireless/admtek/adm8211.c priv->rx_buffers[entry].skb = newskb; entry 407 drivers/net/wireless/admtek/adm8211.c priv->rx_buffers[entry].mapping = entry 413 drivers/net/wireless/admtek/adm8211.c priv->rx_buffers[entry].mapping)) { entry 414 drivers/net/wireless/admtek/adm8211.c priv->rx_buffers[entry].skb = NULL; entry 424 drivers/net/wireless/admtek/adm8211.c priv->rx_ring[entry].buffer1 = entry 425 drivers/net/wireless/admtek/adm8211.c cpu_to_le32(priv->rx_buffers[entry].mapping); entry 428 drivers/net/wireless/admtek/adm8211.c priv->rx_ring[entry].status = cpu_to_le32(RDES0_STATUS_OWN | entry 430 drivers/net/wireless/admtek/adm8211.c priv->rx_ring[entry].length = entry 432 drivers/net/wireless/admtek/adm8211.c (entry == priv->rx_ring_size - 1 ? entry 452 drivers/net/wireless/admtek/adm8211.c entry = (++priv->cur_rx) % priv->rx_ring_size; entry 1632 drivers/net/wireless/admtek/adm8211.c unsigned int entry; entry 1650 drivers/net/wireless/admtek/adm8211.c entry = priv->cur_tx % priv->tx_ring_size; entry 1652 drivers/net/wireless/admtek/adm8211.c priv->tx_buffers[entry].skb = skb; entry 1653 drivers/net/wireless/admtek/adm8211.c priv->tx_buffers[entry].mapping = mapping; entry 1654 drivers/net/wireless/admtek/adm8211.c priv->tx_buffers[entry].hdrlen = hdrlen; entry 1655 drivers/net/wireless/admtek/adm8211.c priv->tx_ring[entry].buffer1 = cpu_to_le32(mapping); entry 1657 drivers/net/wireless/admtek/adm8211.c if (entry == priv->tx_ring_size - 1) entry 1659 drivers/net/wireless/admtek/adm8211.c priv->tx_ring[entry].length = cpu_to_le32(flag | skb->len); entry 1663 drivers/net/wireless/admtek/adm8211.c priv->tx_ring[entry].status = cpu_to_le32(flag); entry 205 drivers/net/wireless/ath/ath.h bool ath_hw_keyreset(struct ath_common *common, u16 entry); entry 1823 drivers/net/wireless/ath/ath10k/pci.c const struct service_to_pipe *entry; entry 1830 drivers/net/wireless/ath/ath10k/pci.c entry = &target_service_to_ce_map_wlan[i]; entry 1832 drivers/net/wireless/ath/ath10k/pci.c if (__le32_to_cpu(entry->service_id) != service_id) entry 1835 drivers/net/wireless/ath/ath10k/pci.c switch (__le32_to_cpu(entry->pipedir)) { entry 1840 drivers/net/wireless/ath/ath10k/pci.c *dl_pipe = __le32_to_cpu(entry->pipenum); entry 1845 drivers/net/wireless/ath/ath10k/pci.c *ul_pipe = __le32_to_cpu(entry->pipenum); entry 1851 drivers/net/wireless/ath/ath10k/pci.c *dl_pipe = __le32_to_cpu(entry->pipenum); entry 1852 drivers/net/wireless/ath/ath10k/pci.c *ul_pipe = __le32_to_cpu(entry->pipenum); entry 770 drivers/net/wireless/ath/ath10k/snoc.c const struct service_to_pipe *entry; entry 777 drivers/net/wireless/ath/ath10k/snoc.c entry = &target_service_to_ce_map_wlan[i]; entry 779 drivers/net/wireless/ath/ath10k/snoc.c if (__le32_to_cpu(entry->service_id) != service_id) entry 782 drivers/net/wireless/ath/ath10k/snoc.c switch (__le32_to_cpu(entry->pipedir)) { entry 787 drivers/net/wireless/ath/ath10k/snoc.c *dl_pipe = __le32_to_cpu(entry->pipenum); entry 792 drivers/net/wireless/ath/ath10k/snoc.c *ul_pipe = __le32_to_cpu(entry->pipenum); entry 798 drivers/net/wireless/ath/ath10k/snoc.c *dl_pipe = __le32_to_cpu(entry->pipenum); entry 799 drivers/net/wireless/ath/ath10k/snoc.c *ul_pipe = __le32_to_cpu(entry->pipenum); entry 198 drivers/net/wireless/ath/ath5k/phy.c u16 entry; entry 237 drivers/net/wireless/ath/ath5k/phy.c entry = ((first_bit - 1) / 8) + offset; entry 244 drivers/net/wireless/ath/ath5k/phy.c position = 0, entry++) { entry 253 drivers/net/wireless/ath/ath5k/phy.c rfb[entry] &= ~mask; entry 254 drivers/net/wireless/ath/ath5k/phy.c rfb[entry] |= ((data << position) << (col * 8)) & mask; entry 257 drivers/net/wireless/ath/ath5k/phy.c data |= (((rfb[entry] & mask) >> (col * 8)) >> position) entry 551 drivers/net/wireless/ath/ath6kl/htc_pipe.c struct htc_pipe_txcredit_alloc *entry; entry 561 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry = &target->pipe.txcredit_alloc[0]; entry 570 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 571 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 573 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->service_id = WMI_DATA_VO_SVC; entry 574 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc = (credits - 6); entry 575 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (entry->credit_alloc == 0) entry 576 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc++; entry 578 drivers/net/wireless/ath/ath6kl/htc_pipe.c credits -= (int) entry->credit_alloc; entry 582 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 583 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->service_id = WMI_CONTROL_SVC; entry 584 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc = credit_per_maxmsg; entry 585 drivers/net/wireless/ath/ath6kl/htc_pipe.c credits -= (int) entry->credit_alloc; entry 590 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 591 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 592 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->service_id = WMI_DATA_BE_SVC; entry 593 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc = (u8) credits; entry 596 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 597 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->service_id = WMI_DATA_VI_SVC; entry 598 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc = credits / 4; entry 599 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (entry->credit_alloc == 0) entry 600 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc++; entry 602 drivers/net/wireless/ath/ath6kl/htc_pipe.c credits -= (int) entry->credit_alloc; entry 606 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 607 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->service_id = WMI_DATA_VO_SVC; entry 608 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc = credits / 4; entry 609 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (entry->credit_alloc == 0) entry 610 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc++; entry 612 drivers/net/wireless/ath/ath6kl/htc_pipe.c credits -= (int) entry->credit_alloc; entry 616 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 617 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->service_id = WMI_CONTROL_SVC; entry 618 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc = credit_per_maxmsg; entry 619 drivers/net/wireless/ath/ath6kl/htc_pipe.c credits -= (int) entry->credit_alloc; entry 623 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 624 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->service_id = WMI_DATA_BK_SVC; entry 625 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc = credit_per_maxmsg; entry 626 drivers/net/wireless/ath/ath6kl/htc_pipe.c credits -= (int) entry->credit_alloc; entry 631 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry++; entry 632 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->service_id = WMI_DATA_BE_SVC; entry 633 drivers/net/wireless/ath/ath6kl/htc_pipe.c entry->credit_alloc = (u8) credits; entry 89 drivers/net/wireless/ath/ath6kl/main.c struct ath6kl_mgmt_buff *entry, *tmp; entry 97 drivers/net/wireless/ath/ath6kl/main.c list_for_each_entry_safe(entry, tmp, &sta->mgmt_psq, list) { entry 98 drivers/net/wireless/ath/ath6kl/main.c kfree(entry); entry 293 drivers/net/wireless/ath/ath9k/ar9003_paprd.c u32 *entry = ah->paprd_gain_table_entries; entry 299 drivers/net/wireless/ath/ath9k/ar9003_paprd.c entry[i] = REG_READ(ah, reg); entry 300 drivers/net/wireless/ath/ath9k/ar9003_paprd.c index[i] = (entry[i] >> 24) & 0xff; entry 29 drivers/net/wireless/ath/ath9k/mci.c struct ath_mci_profile_info *entry; entry 34 drivers/net/wireless/ath/ath9k/mci.c list_for_each_entry(entry, &mci->info, list) { entry 35 drivers/net/wireless/ath/ath9k/mci.c if (entry->conn_handle == info->conn_handle) entry 36 drivers/net/wireless/ath/ath9k/mci.c return entry; entry 45 drivers/net/wireless/ath/ath9k/mci.c struct ath_mci_profile_info *entry; entry 56 drivers/net/wireless/ath/ath9k/mci.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 57 drivers/net/wireless/ath/ath9k/mci.c if (!entry) entry 60 drivers/net/wireless/ath/ath9k/mci.c memcpy(entry, info, 10); entry 62 drivers/net/wireless/ath/ath9k/mci.c list_add_tail(&entry->list, &mci->info); entry 75 drivers/net/wireless/ath/ath9k/mci.c struct ath_mci_profile_info *entry) entry 77 drivers/net/wireless/ath/ath9k/mci.c if (!entry) entry 80 drivers/net/wireless/ath/ath9k/mci.c DEC_PROF(mci, entry); entry 81 drivers/net/wireless/ath/ath9k/mci.c list_del(&entry->list); entry 82 drivers/net/wireless/ath/ath9k/mci.c kfree(entry); entry 301 drivers/net/wireless/ath/ath9k/mci.c struct ath_mci_profile_info *entry = NULL; entry 303 drivers/net/wireless/ath/ath9k/mci.c entry = ath_mci_find_profile(mci, info); entry 304 drivers/net/wireless/ath/ath9k/mci.c if (entry) { entry 313 drivers/net/wireless/ath/ath9k/mci.c if (entry->type != info->type) { entry 314 drivers/net/wireless/ath/ath9k/mci.c DEC_PROF(mci, entry); entry 317 drivers/net/wireless/ath/ath9k/mci.c memcpy(entry, info, 10); entry 321 drivers/net/wireless/ath/ath9k/mci.c if (!entry && !ath_mci_add_profile(common, mci, info)) entry 324 drivers/net/wireless/ath/ath9k/mci.c ath_mci_del_profile(common, mci, entry); entry 576 drivers/net/wireless/ath/carl9170/rx.c struct carl9170_bar_list_entry *entry; entry 590 drivers/net/wireless/ath/carl9170/rx.c list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) { entry 591 drivers/net/wireless/ath/carl9170/rx.c struct sk_buff *entry_skb = entry->skb; entry 609 drivers/net/wireless/ath/carl9170/rx.c list_del_rcu(&entry->list); entry 611 drivers/net/wireless/ath/carl9170/rx.c kfree_rcu(entry, head); entry 454 drivers/net/wireless/ath/carl9170/tx.c struct carl9170_bar_list_entry *entry; entry 458 drivers/net/wireless/ath/carl9170/tx.c list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) { entry 459 drivers/net/wireless/ath/carl9170/tx.c if (entry->skb == skb) { entry 461 drivers/net/wireless/ath/carl9170/tx.c list_del_rcu(&entry->list); entry 463 drivers/net/wireless/ath/carl9170/tx.c kfree_rcu(entry, head); entry 1326 drivers/net/wireless/ath/carl9170/tx.c struct carl9170_bar_list_entry *entry; entry 1329 drivers/net/wireless/ath/carl9170/tx.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 1330 drivers/net/wireless/ath/carl9170/tx.c if (!WARN_ON_ONCE(!entry)) { entry 1331 drivers/net/wireless/ath/carl9170/tx.c entry->skb = skb; entry 1333 drivers/net/wireless/ath/carl9170/tx.c list_add_tail_rcu(&entry->list, &ar->bar_list[queue]); entry 42 drivers/net/wireless/ath/key.c bool ath_hw_keyreset(struct ath_common *common, u16 entry) entry 47 drivers/net/wireless/ath/key.c if (entry >= common->keymax) { entry 49 drivers/net/wireless/ath/key.c entry); entry 53 drivers/net/wireless/ath/key.c keyType = REG_READ(ah, AR_KEYTABLE_TYPE(entry)); entry 57 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY0(entry), 0); entry 58 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY1(entry), 0); entry 59 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY2(entry), 0); entry 60 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY3(entry), 0); entry 61 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY4(entry), 0); entry 62 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_TYPE(entry), AR_KEYTABLE_TYPE_CLR); entry 63 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_MAC0(entry), 0); entry 64 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_MAC1(entry), 0); entry 67 drivers/net/wireless/ath/key.c u16 micentry = entry + 64; entry 88 drivers/net/wireless/ath/key.c u16 entry, const u8 *mac) entry 94 drivers/net/wireless/ath/key.c if (entry >= common->keymax) { entry 96 drivers/net/wireless/ath/key.c entry); entry 121 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_MAC0(entry), macLo); entry 122 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_MAC1(entry), macHi | unicast_flag); entry 129 drivers/net/wireless/ath/key.c static bool ath_hw_set_keycache_entry(struct ath_common *common, u16 entry, entry 137 drivers/net/wireless/ath/key.c if (entry >= common->keymax) { entry 139 drivers/net/wireless/ath/key.c entry); entry 157 drivers/net/wireless/ath/key.c if (entry + 64 >= common->keymax) { entry 159 drivers/net/wireless/ath/key.c "entry %u inappropriate for TKIP\n", entry); entry 200 drivers/net/wireless/ath/key.c u16 micentry = entry + 64; entry 208 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY0(entry), ~key0); entry 209 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY1(entry), ~key1); entry 212 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY2(entry), key2); entry 213 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY3(entry), key3); entry 216 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY4(entry), key4); entry 217 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_TYPE(entry), keyType); entry 220 drivers/net/wireless/ath/key.c (void) ath_hw_keysetmac(common, entry, mac); entry 311 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY0(entry), key0); entry 312 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY1(entry), key1); entry 319 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY0(entry), key0); entry 320 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY1(entry), key1); entry 323 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY2(entry), key2); entry 324 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY3(entry), key3); entry 327 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_KEY4(entry), key4); entry 328 drivers/net/wireless/ath/key.c REG_WRITE(ah, AR_KEYTABLE_TYPE(entry), keyType); entry 333 drivers/net/wireless/ath/key.c (void) ath_hw_keysetmac(common, entry, mac); entry 84 drivers/net/wireless/ath/wcn36xx/smd.c struct wcn36xx_hal_cfg *entry; entry 87 drivers/net/wireless/ath/wcn36xx/smd.c if (*len + sizeof(*entry) + sizeof(u32) >= WCN36XX_HAL_BUF_SIZE) { entry 92 drivers/net/wireless/ath/wcn36xx/smd.c entry = (struct wcn36xx_hal_cfg *) (wcn->hal_buf + *len); entry 93 drivers/net/wireless/ath/wcn36xx/smd.c entry->id = id; entry 94 drivers/net/wireless/ath/wcn36xx/smd.c entry->len = sizeof(u32); entry 95 drivers/net/wireless/ath/wcn36xx/smd.c entry->pad_bytes = 0; entry 96 drivers/net/wireless/ath/wcn36xx/smd.c entry->reserve = 0; entry 98 drivers/net/wireless/ath/wcn36xx/smd.c val = (u32 *) (entry + 1); entry 101 drivers/net/wireless/ath/wcn36xx/smd.c *len += sizeof(*entry) + sizeof(u32); entry 2333 drivers/net/wireless/broadcom/b43/phy_g.c static s8 b43_tssi2dbm_entry(s8 entry[], u8 index, entry 2350 drivers/net/wireless/broadcom/b43/phy_g.c entry[index] = clamp_val(b43_tssi2dbm_ad(m1 * f, 8192), -127, 128); entry 4929 drivers/net/wireless/broadcom/b43/phy_n.c u16 scale, entry; entry 4938 drivers/net/wireless/broadcom/b43/phy_n.c entry = ((scale & 0xFF) << 8) | ladder_lo[i].g_env; entry 4939 drivers/net/wireless/broadcom/b43/phy_n.c b43_ntab_write(dev, B43_NTAB16(15, i), entry); entry 4942 drivers/net/wireless/broadcom/b43/phy_n.c entry = ((scale & 0xFF) << 8) | ladder_iq[i].g_env; entry 4943 drivers/net/wireless/broadcom/b43/phy_n.c b43_ntab_write(dev, B43_NTAB16(15, i + 32), entry); entry 12 drivers/net/wireless/broadcom/b43/ppr.c #define ppr_for_each_entry(ppr, i, entry) \ entry 13 drivers/net/wireless/broadcom/b43/ppr.c for (i = 0, entry = &(ppr)->__all_rates[i]; \ entry 15 drivers/net/wireless/broadcom/b43/ppr.c i++, entry++) entry 1942 drivers/net/wireless/broadcom/b43legacy/phy.c s8 b43legacy_tssi2dbm_entry(s8 entry [], u8 index, s16 pab0, s16 pab1, s16 pab2) entry 1963 drivers/net/wireless/broadcom/b43legacy/phy.c entry[index] = clamp_val(b43legacy_tssi2dbm_ad(m1 * f, 8192), entry 56 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c u32 entry; entry 95 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c nvp->entry = nvp->pos; entry 114 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c if (strncmp(&nvp->data[nvp->entry], "RAW1", 4) == 0) entry 118 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c if (strncmp(&nvp->data[nvp->entry], "devpath", 7) == 0) entry 120 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c if (strncmp(&nvp->data[nvp->entry], "pcie/", 5) == 0) entry 122 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c if (strncmp(&nvp->data[nvp->entry], "boardrev", 8) == 0) entry 147 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c skv = (u8 *)&nvp->data[nvp->entry]; entry 748 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 754 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.nodes[0]; entry 756 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->occupied && !memcmp(entry->ea, ea, ETH_ALEN)) entry 757 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c return entry; entry 758 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry++; entry 767 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry = &fws->desc.other; entry 777 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = ifp->fws_desc; entry 781 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = brcmf_fws_macdesc_lookup(fws, da); entry 782 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (IS_ERR(entry)) entry 783 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = ifp->fws_desc; entry 786 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c return entry; entry 790 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry, entry 799 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->mac_handle) { entry 800 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if_entry = &fws->desc.iface[entry->interface_id]; entry 807 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c closed = entry->state == BRCMF_FWS_STATE_CLOSE && entry 808 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c !entry->requested_credit && !entry->requested_packet; entry 811 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c return closed || !(entry->ac_bitmap & BIT(fifo)); entry 815 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry, entry 818 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->occupied && (ifidx == -1 || ifidx == entry->interface_id)) { entry 819 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_psq_flush(fws, &entry->psq, ifidx); entry 820 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->occupied = !!(entry->psq.len); entry 877 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry = brcmf_skbcb(skb)->mac; entry 885 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->name, brcmf_skb_if_flags_get_field(skb, INDEX), entry 888 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->send_tim_signal) entry 910 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->send_tim_signal) { entry 911 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->send_tim_signal = 0; entry 914 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c wlh[2] = entry->mac_handle; entry 915 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c wlh[3] = entry->traffic_pending_bmp; entry 917 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->mac_handle, entry->traffic_pending_bmp); entry 919 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->traffic_lastreported_bmp = entry->traffic_pending_bmp; entry 928 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry, entry 939 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (brcmu_pktq_mlen(&entry->psq, 3 << (fifo * 2)) == 0) entry 940 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->traffic_pending_bmp &= ~NBITVAL(fifo); entry 942 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->traffic_pending_bmp |= NBITVAL(fifo); entry 944 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->send_tim_signal = false; entry 945 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->traffic_lastreported_bmp != entry->traffic_pending_bmp) entry 946 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->send_tim_signal = true; entry 947 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (send_immediately && entry->send_tim_signal && entry 948 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->state == BRCMF_FWS_STATE_CLOSE) { entry 960 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c skcb->mac = entry; entry 1006 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry, *existing; entry 1015 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.nodes[mac_handle & 0x1F]; entry 1017 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->occupied) { entry 1019 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->name, addr); entry 1021 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_cleanup(fws, entry, -1); entry 1022 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_deinit(entry); entry 1031 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (!entry->occupied) { entry 1033 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->mac_handle = mac_handle; entry 1034 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_init(entry, addr, ifidx); entry 1035 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_set_name(fws, entry); entry 1036 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmu_pktq_init(&entry->psq, BRCMF_FWS_PSQ_PREC_COUNT, entry 1039 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_dbg(TRACE, "add %s mac %pM\n", entry->name, addr); entry 1044 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry != existing) { entry 1047 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c memcpy(entry, existing, entry 1049 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->mac_handle = mac_handle; entry 1051 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_set_name(fws, entry); entry 1053 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_dbg(TRACE, "relocate %s mac %pM\n", entry->name, entry 1057 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c WARN_ON(entry->mac_handle != mac_handle); entry 1067 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 1072 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.nodes[mac_handle & 0x1F]; entry 1073 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (!entry->occupied) { entry 1079 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->requested_credit = 0; entry 1080 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->requested_packet = 0; entry 1082 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->state = BRCMF_FWS_STATE_OPEN; entry 1085 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->state = BRCMF_FWS_STATE_CLOSE; entry 1086 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_tim_update(fws, entry, BRCMF_FWS_FIFO_AC_BK, false); entry 1087 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_tim_update(fws, entry, BRCMF_FWS_FIFO_AC_BE, false); entry 1088 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_tim_update(fws, entry, BRCMF_FWS_FIFO_AC_VI, false); entry 1089 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_tim_update(fws, entry, BRCMF_FWS_FIFO_AC_VO, true); entry 1099 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 1110 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.iface[ifidx]; entry 1111 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (!entry->occupied) { entry 1117 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->name); entry 1121 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->state = BRCMF_FWS_STATE_OPEN; entry 1125 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->state = BRCMF_FWS_STATE_CLOSE; entry 1144 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 1146 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.nodes[data[1] & 0x1F]; entry 1147 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (!entry->occupied) { entry 1156 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_get_tlv_name(type), type, entry->name, entry 1160 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->requested_credit = data[0]; entry 1162 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->requested_packet = data[0]; entry 1164 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->ac_bitmap = data[2]; entry 1170 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_use_req_credit(struct brcmf_fws_mac_descriptor *entry, entry 1173 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->requested_credit > 0) { entry 1174 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->requested_credit--; entry 1177 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->state != BRCMF_FWS_STATE_CLOSE) entry 1179 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c } else if (entry->requested_packet > 0) { entry 1180 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->requested_packet--; entry 1183 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->state != BRCMF_FWS_STATE_CLOSE) entry 1193 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry = brcmf_skbcb(skb)->mac; entry 1196 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c (entry->state == BRCMF_FWS_STATE_CLOSE)) entry 1197 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->requested_credit++; entry 1254 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 1262 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = brcmf_skbcb(p)->mac; entry 1263 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry == NULL) { entry 1275 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c pq = &entry->psq; entry 1320 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c } else if (brcmu_pktq_penq(&entry->psq, prec, p) == NULL) { entry 1336 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_tim_update(fws, entry, fifo, true); entry 1337 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_flow_control_check(fws, &entry->psq, entry 1345 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 1358 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &table[(node_pos + i) % num_nodes]; entry 1359 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (!entry->occupied || entry 1360 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_closed(fws, entry, fifo)) entry 1363 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->suppressed) entry 1367 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c p = brcmu_pktq_mdeq(&entry->psq, pmsk << (fifo * 2), &prec_out); entry 1369 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->suppressed) { entry 1370 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->suppr_transit_count) entry 1372 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->suppressed = false; entry 1373 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c p = brcmu_pktq_mdeq(&entry->psq, entry 1380 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_use_req_credit(entry, p); entry 1384 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_flow_control_check(fws, &entry->psq, entry 1392 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_tim_update(fws, entry, fifo, false); entry 1413 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry = brcmf_skbcb(skb)->mac; entry 1420 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (!entry->suppressed) { entry 1421 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->suppressed = true; entry 1422 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->suppr_transit_count = entry->transit_count; entry 1424 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->name, entry->transit_count); entry 1427 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->generation = genbit; entry 1461 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry = NULL; entry 1491 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = skcb->mac; entry 1492 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (WARN_ON(!entry)) { entry 1496 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->transit_count--; entry 1497 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->suppressed && entry->suppr_transit_count) entry 1498 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->suppr_transit_count--; entry 1500 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_dbg(DATA, "%s flags %d htod %X seq %X\n", entry->name, entry 1976 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry = skcb->mac; entry 1980 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_skb_htod_tag_set_field(p, GENERATION, entry->generation); entry 1997 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 2002 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = brcmf_skbcb(skb)->mac; entry 2003 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->occupied) { entry 2008 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c pktout = brcmu_pktq_penq_head(&entry->psq, qidx, skb); entry 2010 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c bphy_err(drvr, "%s queue %d full\n", entry->name, qidx); entry 2014 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c bphy_err(drvr, "%s entry removed\n", entry->name); entry 2058 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 2063 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = skcb->mac; entry 2064 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (IS_ERR(entry)) entry 2065 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c return PTR_ERR(entry); entry 2068 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->transit_count++; entry 2069 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->suppressed) entry 2070 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->suppr_transit_count++; entry 2075 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_dbg(DATA, "%s flags %X htod %X bus_tx %d\n", entry->name, entry 2078 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->transit_count--; entry 2079 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (entry->suppressed) entry 2080 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry->suppr_transit_count--; entry 2159 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry = ifp->fws_desc; entry 2162 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (!entry) entry 2165 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_init(entry, ifp->mac_addr, ifp->ifidx); entry 2171 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry; entry 2176 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c entry = &fws->desc.iface[ifp->ifidx]; entry 2177 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c ifp->fws_desc = entry; entry 2178 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_init(entry, ifp->mac_addr, ifp->ifidx); entry 2179 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_set_name(fws, entry); entry 2180 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmu_pktq_init(&entry->psq, BRCMF_FWS_PSQ_PREC_COUNT, entry 2182 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_dbg(TRACE, "added %s\n", entry->name); entry 2187 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct brcmf_fws_mac_descriptor *entry = ifp->fws_desc; entry 2190 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (!entry) entry 2195 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_dbg(TRACE, "deleting %s\n", entry->name); entry 2198 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c brcmf_fws_macdesc_deinit(entry); entry 176 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c struct brcms_debugfs_entry *entry = inode->i_private; entry 178 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c return single_open(f, entry->read, entry->drvr); entry 195 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c struct brcms_debugfs_entry *entry; entry 197 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c entry = devm_kzalloc(dev, sizeof(*entry), GFP_KERNEL); entry 198 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c if (!entry) entry 201 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c entry->read = read_fn; entry 202 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c entry->drvr = drvr; entry 204 drivers/net/wireless/broadcom/brcm80211/brcmsmac/debug.c debugfs_create_file(fn, 0444, dentry, entry, &brcms_debugfs_def_ops); entry 1596 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c int i, entry; entry 1601 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c for (entry = 0; entry < wl->fw.hdr_num_entries[i]; entry 1602 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c entry++, hdr++) { entry 1628 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c int i, entry; entry 1633 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c for (entry = 0; entry < wl->fw.hdr_num_entries[i]; entry 1634 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c entry++, hdr++) { entry 1670 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c int entry; entry 1696 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c for (entry = 0; entry < wl->fw.hdr_num_entries[i] && entry 1697 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c !rc; entry++, ucode_hdr++) { entry 4518 drivers/net/wireless/cisco/airo.c struct proc_dir_entry *entry; entry 4529 drivers/net/wireless/cisco/airo.c entry = proc_create_data("StatsDelta", 0444 & proc_perm, entry 4531 drivers/net/wireless/cisco/airo.c if (!entry) entry 4533 drivers/net/wireless/cisco/airo.c proc_set_user(entry, proc_kuid, proc_kgid); entry 4536 drivers/net/wireless/cisco/airo.c entry = proc_create_data("Stats", 0444 & proc_perm, entry 4538 drivers/net/wireless/cisco/airo.c if (!entry) entry 4540 drivers/net/wireless/cisco/airo.c proc_set_user(entry, proc_kuid, proc_kgid); entry 4543 drivers/net/wireless/cisco/airo.c entry = proc_create_data("Status", 0444 & proc_perm, entry 4545 drivers/net/wireless/cisco/airo.c if (!entry) entry 4547 drivers/net/wireless/cisco/airo.c proc_set_user(entry, proc_kuid, proc_kgid); entry 4550 drivers/net/wireless/cisco/airo.c entry = proc_create_data("Config", proc_perm, entry 4552 drivers/net/wireless/cisco/airo.c if (!entry) entry 4554 drivers/net/wireless/cisco/airo.c proc_set_user(entry, proc_kuid, proc_kgid); entry 4557 drivers/net/wireless/cisco/airo.c entry = proc_create_data("SSID", proc_perm, entry 4559 drivers/net/wireless/cisco/airo.c if (!entry) entry 4561 drivers/net/wireless/cisco/airo.c proc_set_user(entry, proc_kuid, proc_kgid); entry 4564 drivers/net/wireless/cisco/airo.c entry = proc_create_data("APList", proc_perm, entry 4566 drivers/net/wireless/cisco/airo.c if (!entry) entry 4568 drivers/net/wireless/cisco/airo.c proc_set_user(entry, proc_kuid, proc_kgid); entry 4571 drivers/net/wireless/cisco/airo.c entry = proc_create_data("BSSList", proc_perm, entry 4573 drivers/net/wireless/cisco/airo.c if (!entry) entry 4575 drivers/net/wireless/cisco/airo.c proc_set_user(entry, proc_kuid, proc_kgid); entry 4578 drivers/net/wireless/cisco/airo.c entry = proc_create_data("WepKey", proc_perm, entry 4580 drivers/net/wireless/cisco/airo.c if (!entry) entry 4582 drivers/net/wireless/cisco/airo.c proc_set_user(entry, proc_kuid, proc_kgid); entry 3896 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct ipw_station_entry entry; entry 3916 drivers/net/wireless/intel/ipw2x00/ipw2200.c entry.reserved = 0; entry 3917 drivers/net/wireless/intel/ipw2x00/ipw2200.c entry.support_mode = 0; entry 3918 drivers/net/wireless/intel/ipw2x00/ipw2200.c memcpy(entry.mac_addr, bssid, ETH_ALEN); entry 3920 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_write_direct(priv, IPW_STATION_TABLE_LOWER + i * sizeof(entry), entry 3921 drivers/net/wireless/intel/ipw2x00/ipw2200.c &entry, sizeof(entry)); entry 8138 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct ipw_ibss_seq *entry = NULL; entry 8143 drivers/net/wireless/intel/ipw2x00/ipw2200.c entry = entry 8145 drivers/net/wireless/intel/ipw2x00/ipw2200.c if (ether_addr_equal(entry->mac, mac)) entry 8149 drivers/net/wireless/intel/ipw2x00/ipw2200.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 8150 drivers/net/wireless/intel/ipw2x00/ipw2200.c if (!entry) { entry 8155 drivers/net/wireless/intel/ipw2x00/ipw2200.c memcpy(entry->mac, mac, ETH_ALEN); entry 8156 drivers/net/wireless/intel/ipw2x00/ipw2200.c entry->seq_num = seq; entry 8157 drivers/net/wireless/intel/ipw2x00/ipw2200.c entry->frag_num = frag; entry 8158 drivers/net/wireless/intel/ipw2x00/ipw2200.c entry->packet_time = jiffies; entry 8159 drivers/net/wireless/intel/ipw2x00/ipw2200.c list_add(&entry->list, entry 8163 drivers/net/wireless/intel/ipw2x00/ipw2200.c last_seq = &entry->seq_num; entry 8164 drivers/net/wireless/intel/ipw2x00/ipw2200.c last_frag = &entry->frag_num; entry 8165 drivers/net/wireless/intel/ipw2x00/ipw2200.c last_time = &entry->packet_time; entry 60 drivers/net/wireless/intel/ipw2x00/libipw_rx.c struct libipw_frag_entry *entry; entry 64 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry = &ieee->frag_cache[i]; entry 65 drivers/net/wireless/intel/ipw2x00/libipw_rx.c if (entry->skb != NULL && entry 66 drivers/net/wireless/intel/ipw2x00/libipw_rx.c time_after(jiffies, entry->first_frag_time + 2 * HZ)) { entry 69 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry->seq, entry->last_frag); entry 70 drivers/net/wireless/intel/ipw2x00/libipw_rx.c dev_kfree_skb_any(entry->skb); entry 71 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry->skb = NULL; entry 74 drivers/net/wireless/intel/ipw2x00/libipw_rx.c if (entry->skb != NULL && entry->seq == seq && entry 75 drivers/net/wireless/intel/ipw2x00/libipw_rx.c (entry->last_frag + 1 == frag || frag == -1) && entry 76 drivers/net/wireless/intel/ipw2x00/libipw_rx.c ether_addr_equal(entry->src_addr, src) && entry 77 drivers/net/wireless/intel/ipw2x00/libipw_rx.c ether_addr_equal(entry->dst_addr, dst)) entry 78 drivers/net/wireless/intel/ipw2x00/libipw_rx.c return entry; entry 91 drivers/net/wireless/intel/ipw2x00/libipw_rx.c struct libipw_frag_entry *entry; entry 107 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry = &ieee->frag_cache[ieee->frag_next_idx]; entry 112 drivers/net/wireless/intel/ipw2x00/libipw_rx.c if (entry->skb != NULL) entry 113 drivers/net/wireless/intel/ipw2x00/libipw_rx.c dev_kfree_skb_any(entry->skb); entry 115 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry->first_frag_time = jiffies; entry 116 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry->seq = seq; entry 117 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry->last_frag = frag; entry 118 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry->skb = skb; entry 119 drivers/net/wireless/intel/ipw2x00/libipw_rx.c memcpy(entry->src_addr, hdr->addr2, ETH_ALEN); entry 120 drivers/net/wireless/intel/ipw2x00/libipw_rx.c memcpy(entry->dst_addr, hdr->addr1, ETH_ALEN); entry 124 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry = libipw_frag_cache_find(ieee, seq, frag, hdr->addr2, entry 126 drivers/net/wireless/intel/ipw2x00/libipw_rx.c if (entry != NULL) { entry 127 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry->last_frag = frag; entry 128 drivers/net/wireless/intel/ipw2x00/libipw_rx.c skb = entry->skb; entry 141 drivers/net/wireless/intel/ipw2x00/libipw_rx.c struct libipw_frag_entry *entry; entry 146 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry = libipw_frag_cache_find(ieee, seq, -1, hdr->addr2, entry 149 drivers/net/wireless/intel/ipw2x00/libipw_rx.c if (entry == NULL) { entry 155 drivers/net/wireless/intel/ipw2x00/libipw_rx.c entry->skb = NULL; entry 1652 drivers/net/wireless/intel/iwlwifi/fw/dbg.c struct iwl_fw_ini_dump_entry *entry; entry 1666 drivers/net/wireless/intel/iwlwifi/fw/dbg.c entry = kmalloc(sizeof(*entry) + sizeof(*tlv) + size, GFP_KERNEL); entry 1667 drivers/net/wireless/intel/iwlwifi/fw/dbg.c if (!entry) entry 1670 drivers/net/wireless/intel/iwlwifi/fw/dbg.c entry->size = sizeof(*tlv) + size; entry 1672 drivers/net/wireless/intel/iwlwifi/fw/dbg.c tlv = (void *)entry->data; entry 1708 drivers/net/wireless/intel/iwlwifi/fw/dbg.c list_add_tail(&entry->list, list); entry 1710 drivers/net/wireless/intel/iwlwifi/fw/dbg.c return entry->size; entry 1713 drivers/net/wireless/intel/iwlwifi/fw/dbg.c kfree(entry); entry 1722 drivers/net/wireless/intel/iwlwifi/fw/dbg.c struct iwl_fw_ini_dump_entry *entry; entry 1728 drivers/net/wireless/intel/iwlwifi/fw/dbg.c entry = kmalloc(sizeof(*entry) + size, GFP_KERNEL); entry 1729 drivers/net/wireless/intel/iwlwifi/fw/dbg.c if (!entry) entry 1732 drivers/net/wireless/intel/iwlwifi/fw/dbg.c entry->size = size; entry 1734 drivers/net/wireless/intel/iwlwifi/fw/dbg.c tlv = (void *)entry->data; entry 1786 drivers/net/wireless/intel/iwlwifi/fw/dbg.c list_add(&entry->list, list); entry 1788 drivers/net/wireless/intel/iwlwifi/fw/dbg.c return entry->size; entry 1904 drivers/net/wireless/intel/iwlwifi/fw/dbg.c struct iwl_fw_ini_dump_entry *entry; entry 1916 drivers/net/wireless/intel/iwlwifi/fw/dbg.c entry = kmalloc(sizeof(*entry) + sizeof(*hdr), GFP_KERNEL); entry 1917 drivers/net/wireless/intel/iwlwifi/fw/dbg.c if (!entry) entry 1920 drivers/net/wireless/intel/iwlwifi/fw/dbg.c entry->size = sizeof(*hdr); entry 1924 drivers/net/wireless/intel/iwlwifi/fw/dbg.c kfree(entry); entry 1928 drivers/net/wireless/intel/iwlwifi/fw/dbg.c hdr = (void *)entry->data; entry 1930 drivers/net/wireless/intel/iwlwifi/fw/dbg.c hdr->file_len = cpu_to_le32(size + entry->size); entry 1932 drivers/net/wireless/intel/iwlwifi/fw/dbg.c list_add(&entry->list, list); entry 1986 drivers/net/wireless/intel/iwlwifi/fw/dbg.c struct iwl_fw_ini_dump_entry *entry = entry 1987 drivers/net/wireless/intel/iwlwifi/fw/dbg.c list_entry(list->next, typeof(*entry), list); entry 1989 drivers/net/wireless/intel/iwlwifi/fw/dbg.c list_del(&entry->list); entry 1990 drivers/net/wireless/intel/iwlwifi/fw/dbg.c kfree(entry); entry 2007 drivers/net/wireless/intel/iwlwifi/fw/dbg.c struct iwl_fw_ini_dump_entry *entry; entry 2011 drivers/net/wireless/intel/iwlwifi/fw/dbg.c list_for_each_entry(entry, &dump_list, list) { entry 2013 drivers/net/wireless/intel/iwlwifi/fw/dbg.c entry->data, entry->size, offs); entry 2014 drivers/net/wireless/intel/iwlwifi/fw/dbg.c offs += entry->size; entry 160 drivers/net/wireless/intel/iwlwifi/iwl-devtrace-io.h __field(u32, entry) entry 167 drivers/net/wireless/intel/iwlwifi/iwl-devtrace-io.h __entry->entry = msix_entry->entry; entry 173 drivers/net/wireless/intel/iwlwifi/iwl-devtrace-io.h __entry->entry, __entry->defirq, entry 169 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c struct iwl_phy_db_entry *entry = entry 171 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c if (!entry) entry 174 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c kfree(entry->data); entry 175 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry->data = NULL; entry 176 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry->size = 0; entry 208 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c struct iwl_phy_db_entry *entry; entry 244 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry = iwl_phy_db_get_section(phy_db, type, chg_id); entry 245 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c if (!entry) entry 248 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c kfree(entry->data); entry 249 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry->data = kmemdup(phy_db_notif->data, size, GFP_ATOMIC); entry 250 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c if (!entry->data) { entry 251 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry->size = 0; entry 255 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry->size = size; entry 329 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c struct iwl_phy_db_entry *entry; entry 341 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry = iwl_phy_db_get_section(phy_db, type, ch_group_id); entry 342 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c if (!entry) entry 345 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c *data = entry->data; entry 346 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c *size = entry->size; entry 388 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c struct iwl_phy_db_entry *entry; entry 392 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry = iwl_phy_db_get_section(phy_db, entry 395 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c if (!entry) entry 398 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c if (!entry->size) entry 404 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry->size, entry 405 drivers/net/wireless/intel/iwlwifi/iwl-phy-db.c entry->data); entry 436 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c struct iwl_mvm_loc_entry *entry; entry 438 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c list_for_each_entry(entry, &mvm->ftm_initiator.loc_list, list) { entry 439 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c if (!ether_addr_equal_unaligned(res->addr, entry->addr)) entry 442 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c if (entry->lci_len) { entry 443 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c res->ftm.lci_len = entry->lci_len; entry 444 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c res->ftm.lci = entry->buf; entry 447 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c if (entry->civic_len) { entry 448 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c res->ftm.civicloc_len = entry->civic_len; entry 449 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c res->ftm.civicloc = entry->buf + entry->lci_len; entry 635 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c struct iwl_mvm_loc_entry *entry; entry 665 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c entry = kmalloc(sizeof(*entry) + lci_len + civic_len, GFP_KERNEL); entry 666 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c if (!entry) entry 669 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c memcpy(entry->addr, mgmt->bssid, ETH_ALEN); entry 671 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c entry->lci_len = lci_len; entry 673 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c memcpy(entry->buf, lci, lci_len); entry 675 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c entry->civic_len = civic_len; entry 677 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c memcpy(entry->buf + lci_len, civic, civic_len); entry 679 drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c list_add_tail(&entry->list, &mvm->ftm_initiator.loc_list); entry 824 drivers/net/wireless/intel/iwlwifi/mvm/fw.c union acpi_object *entry; entry 826 drivers/net/wireless/intel/iwlwifi/mvm/fw.c entry = &wifi_pkg->package.elements[idx++]; entry 827 drivers/net/wireless/intel/iwlwifi/mvm/fw.c if ((entry->type != ACPI_TYPE_INTEGER) || entry 828 drivers/net/wireless/intel/iwlwifi/mvm/fw.c (entry->integer.value > U8_MAX)) { entry 833 drivers/net/wireless/intel/iwlwifi/mvm/fw.c mvm->geo_profiles[i].values[j] = entry->integer.value; entry 916 drivers/net/wireless/intel/iwlwifi/mvm/ops.c struct iwl_async_handler_entry *entry, *tmp; entry 919 drivers/net/wireless/intel/iwlwifi/mvm/ops.c list_for_each_entry_safe(entry, tmp, &mvm->async_handlers_list, list) { entry 920 drivers/net/wireless/intel/iwlwifi/mvm/ops.c iwl_free_rxb(&entry->rxb); entry 921 drivers/net/wireless/intel/iwlwifi/mvm/ops.c list_del(&entry->list); entry 922 drivers/net/wireless/intel/iwlwifi/mvm/ops.c kfree(entry); entry 931 drivers/net/wireless/intel/iwlwifi/mvm/ops.c struct iwl_async_handler_entry *entry, *tmp; entry 944 drivers/net/wireless/intel/iwlwifi/mvm/ops.c list_for_each_entry_safe(entry, tmp, &local_list, list) { entry 945 drivers/net/wireless/intel/iwlwifi/mvm/ops.c if (entry->context == RX_HANDLER_ASYNC_LOCKED) entry 947 drivers/net/wireless/intel/iwlwifi/mvm/ops.c entry->fn(mvm, &entry->rxb); entry 948 drivers/net/wireless/intel/iwlwifi/mvm/ops.c iwl_free_rxb(&entry->rxb); entry 949 drivers/net/wireless/intel/iwlwifi/mvm/ops.c list_del(&entry->list); entry 950 drivers/net/wireless/intel/iwlwifi/mvm/ops.c if (entry->context == RX_HANDLER_ASYNC_LOCKED) entry 952 drivers/net/wireless/intel/iwlwifi/mvm/ops.c kfree(entry); entry 1006 drivers/net/wireless/intel/iwlwifi/mvm/ops.c struct iwl_async_handler_entry *entry; entry 1016 drivers/net/wireless/intel/iwlwifi/mvm/ops.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 1018 drivers/net/wireless/intel/iwlwifi/mvm/ops.c if (!entry) entry 1021 drivers/net/wireless/intel/iwlwifi/mvm/ops.c entry->rxb._page = rxb_steal_page(rxb); entry 1022 drivers/net/wireless/intel/iwlwifi/mvm/ops.c entry->rxb._offset = rxb->_offset; entry 1023 drivers/net/wireless/intel/iwlwifi/mvm/ops.c entry->rxb._rx_page_order = rxb->_rx_page_order; entry 1024 drivers/net/wireless/intel/iwlwifi/mvm/ops.c entry->fn = rx_h->fn; entry 1025 drivers/net/wireless/intel/iwlwifi/mvm/ops.c entry->context = rx_h->context; entry 1027 drivers/net/wireless/intel/iwlwifi/mvm/ops.c list_add_tail(&entry->list, &mvm->async_handlers_list); entry 616 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct msix_entry *entry) entry 626 drivers/net/wireless/intel/iwlwifi/pcie/internal.h iwl_write32(trans, CSR_MSIX_AUTOMASK_ST_AD, BIT(entry->entry)); entry 1552 drivers/net/wireless/intel/iwlwifi/pcie/rx.c static struct iwl_trans_pcie *iwl_pcie_get_trans_pcie(struct msix_entry *entry) entry 1554 drivers/net/wireless/intel/iwlwifi/pcie/rx.c u8 queue = entry->entry; entry 1555 drivers/net/wireless/intel/iwlwifi/pcie/rx.c struct msix_entry *entries = entry - queue; entry 1566 drivers/net/wireless/intel/iwlwifi/pcie/rx.c struct msix_entry *entry = dev_id; entry 1567 drivers/net/wireless/intel/iwlwifi/pcie/rx.c struct iwl_trans_pcie *trans_pcie = iwl_pcie_get_trans_pcie(entry); entry 1570 drivers/net/wireless/intel/iwlwifi/pcie/rx.c trace_iwlwifi_dev_irq_msix(trans->dev, entry, false, 0, 0); entry 1572 drivers/net/wireless/intel/iwlwifi/pcie/rx.c if (WARN_ON(entry->entry >= trans->num_rx_queues)) entry 1578 drivers/net/wireless/intel/iwlwifi/pcie/rx.c iwl_pcie_rx_handle(trans, entry->entry); entry 1581 drivers/net/wireless/intel/iwlwifi/pcie/rx.c iwl_pcie_clear_irq(trans, entry); entry 2097 drivers/net/wireless/intel/iwlwifi/pcie/rx.c struct msix_entry *entry = dev_id; entry 2098 drivers/net/wireless/intel/iwlwifi/pcie/rx.c struct iwl_trans_pcie *trans_pcie = iwl_pcie_get_trans_pcie(entry); entry 2115 drivers/net/wireless/intel/iwlwifi/pcie/rx.c trace_iwlwifi_dev_irq_msix(trans->dev, entry, true, inta_fh, inta_hw); entry 2257 drivers/net/wireless/intel/iwlwifi/pcie/rx.c iwl_pcie_clear_irq(trans, entry); entry 1644 drivers/net/wireless/intel/iwlwifi/pcie/trans.c trans_pcie->msix_entries[i].entry = i; entry 224 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c struct prism2_frag_entry *entry; entry 228 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry = &local->frag_cache[i]; entry 229 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c if (entry->skb != NULL && entry 230 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c time_after(jiffies, entry->first_frag_time + 2 * HZ)) { entry 233 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c local->dev->name, entry->seq, entry->last_frag); entry 234 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c dev_kfree_skb(entry->skb); entry 235 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry->skb = NULL; entry 238 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c if (entry->skb != NULL && entry->seq == seq && entry 239 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c (entry->last_frag + 1 == frag || frag == -1) && entry 240 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c memcmp(entry->src_addr, src, ETH_ALEN) == 0 && entry 241 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c memcmp(entry->dst_addr, dst, ETH_ALEN) == 0) entry 242 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c return entry; entry 256 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c struct prism2_frag_entry *entry; entry 272 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry = &local->frag_cache[local->frag_next_idx]; entry 277 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c if (entry->skb != NULL) entry 278 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c dev_kfree_skb(entry->skb); entry 280 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry->first_frag_time = jiffies; entry 281 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry->seq = seq; entry 282 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry->last_frag = frag; entry 283 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry->skb = skb; entry 284 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c memcpy(entry->src_addr, hdr->addr2, ETH_ALEN); entry 285 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c memcpy(entry->dst_addr, hdr->addr1, ETH_ALEN); entry 289 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry = prism2_frag_cache_find(local, seq, frag, hdr->addr2, entry 291 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c if (entry != NULL) { entry 292 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry->last_frag = frag; entry 293 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c skb = entry->skb; entry 307 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c struct prism2_frag_entry *entry; entry 312 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry = prism2_frag_cache_find(local, seq, -1, hdr->addr2, hdr->addr1); entry 314 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c if (entry == NULL) { entry 321 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c entry->skb = NULL; entry 325 drivers/net/wireless/intersil/hostap/hostap_ap.c struct mac_entry *entry; entry 348 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = v; entry 349 drivers/net/wireless/intersil/hostap/hostap_ap.c seq_printf(m, "%pM\n", entry->addr); entry 381 drivers/net/wireless/intersil/hostap/hostap_ap.c struct mac_entry *entry; entry 383 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = kmalloc(sizeof(struct mac_entry), GFP_KERNEL); entry 384 drivers/net/wireless/intersil/hostap/hostap_ap.c if (entry == NULL) entry 387 drivers/net/wireless/intersil/hostap/hostap_ap.c memcpy(entry->addr, mac, ETH_ALEN); entry 390 drivers/net/wireless/intersil/hostap/hostap_ap.c list_add_tail(&entry->list, &mac_restrictions->mac_list); entry 401 drivers/net/wireless/intersil/hostap/hostap_ap.c struct mac_entry *entry; entry 406 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = list_entry(ptr, struct mac_entry, list); entry 408 drivers/net/wireless/intersil/hostap/hostap_ap.c if (ether_addr_equal(entry->addr, mac)) { entry 410 drivers/net/wireless/intersil/hostap/hostap_ap.c kfree(entry); entry 424 drivers/net/wireless/intersil/hostap/hostap_ap.c struct mac_entry *entry; entry 431 drivers/net/wireless/intersil/hostap/hostap_ap.c list_for_each_entry(entry, &mac_restrictions->mac_list, list) { entry 432 drivers/net/wireless/intersil/hostap/hostap_ap.c if (ether_addr_equal(entry->addr, mac)) { entry 449 drivers/net/wireless/intersil/hostap/hostap_ap.c struct mac_entry *entry; entry 458 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = list_entry(ptr, struct mac_entry, list); entry 460 drivers/net/wireless/intersil/hostap/hostap_ap.c kfree(entry); entry 1069 drivers/net/wireless/intersil/hostap/hostap_ap.c struct add_sta_proc_data *entry, *prev; entry 1071 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = ap->add_sta_proc_entries; entry 1074 drivers/net/wireless/intersil/hostap/hostap_ap.c while (entry) { entry 1076 drivers/net/wireless/intersil/hostap/hostap_ap.c sta = ap_get_sta(ap, entry->addr); entry 1090 drivers/net/wireless/intersil/hostap/hostap_ap.c prev = entry; entry 1091 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = entry->next; entry 1120 drivers/net/wireless/intersil/hostap/hostap_ap.c struct add_sta_proc_data *entry; entry 1123 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 1124 drivers/net/wireless/intersil/hostap/hostap_ap.c if (entry) { entry 1125 drivers/net/wireless/intersil/hostap/hostap_ap.c memcpy(entry->addr, sta->addr, ETH_ALEN); entry 1126 drivers/net/wireless/intersil/hostap/hostap_ap.c entry->next = ap->add_sta_proc_entries; entry 1127 drivers/net/wireless/intersil/hostap/hostap_ap.c ap->add_sta_proc_entries = entry; entry 1956 drivers/net/wireless/intersil/hostap/hostap_ap.c struct wds_oper_data *entry, *prev; entry 1959 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = local->ap->wds_oper_entries; entry 1963 drivers/net/wireless/intersil/hostap/hostap_ap.c while (entry) { entry 1967 drivers/net/wireless/intersil/hostap/hostap_ap.c entry->type == WDS_ADD ? "adding" : "removing", entry 1968 drivers/net/wireless/intersil/hostap/hostap_ap.c entry->addr); entry 1969 drivers/net/wireless/intersil/hostap/hostap_ap.c if (entry->type == WDS_ADD) entry 1970 drivers/net/wireless/intersil/hostap/hostap_ap.c prism2_wds_add(local, entry->addr, 0); entry 1971 drivers/net/wireless/intersil/hostap/hostap_ap.c else if (entry->type == WDS_DEL) entry 1972 drivers/net/wireless/intersil/hostap/hostap_ap.c prism2_wds_del(local, entry->addr, 0, 1); entry 1974 drivers/net/wireless/intersil/hostap/hostap_ap.c prev = entry; entry 1975 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = entry->next; entry 3255 drivers/net/wireless/intersil/hostap/hostap_ap.c struct wds_oper_data *entry; entry 3257 drivers/net/wireless/intersil/hostap/hostap_ap.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 3258 drivers/net/wireless/intersil/hostap/hostap_ap.c if (!entry) entry 3260 drivers/net/wireless/intersil/hostap/hostap_ap.c memcpy(entry->addr, addr, ETH_ALEN); entry 3261 drivers/net/wireless/intersil/hostap/hostap_ap.c entry->type = type; entry 3263 drivers/net/wireless/intersil/hostap/hostap_ap.c entry->next = local->ap->wds_oper_entries; entry 3264 drivers/net/wireless/intersil/hostap/hostap_ap.c local->ap->wds_oper_entries = entry; entry 171 drivers/net/wireless/intersil/hostap/hostap_hw.c struct hostap_cmd_queue *entry, entry 175 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->del_req = 1; entry 176 drivers/net/wireless/intersil/hostap/hostap_hw.c if (!list_empty(&entry->list)) { entry 177 drivers/net/wireless/intersil/hostap/hostap_hw.c list_del_init(&entry->list); entry 182 drivers/net/wireless/intersil/hostap/hostap_hw.c if (refcount_dec_and_test(&entry->usecnt) && entry->del_req) entry 183 drivers/net/wireless/intersil/hostap/hostap_hw.c kfree(entry); entry 196 drivers/net/wireless/intersil/hostap/hostap_hw.c struct hostap_cmd_queue *entry, entry 202 drivers/net/wireless/intersil/hostap/hostap_hw.c __hostap_cmd_queue_free(local, entry, del_req); entry 215 drivers/net/wireless/intersil/hostap/hostap_hw.c struct hostap_cmd_queue *entry; entry 219 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = list_entry(ptr, struct hostap_cmd_queue, list); entry 220 drivers/net/wireless/intersil/hostap/hostap_hw.c refcount_inc(&entry->usecnt); entry 223 drivers/net/wireless/intersil/hostap/hostap_hw.c local->dev->name, entry->type, entry->cmd, entry 224 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->param0); entry 225 drivers/net/wireless/intersil/hostap/hostap_hw.c __hostap_cmd_queue_free(local, entry, 1); entry 244 drivers/net/wireless/intersil/hostap/hostap_hw.c struct hostap_cmd_queue *entry) entry 258 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry->issued) { entry 260 drivers/net/wireless/intersil/hostap/hostap_hw.c dev->name, entry); entry 288 drivers/net/wireless/intersil/hostap/hostap_hw.c HFA384X_OUTW(entry->param0, HFA384X_PARAM0_OFF); entry 289 drivers/net/wireless/intersil/hostap/hostap_hw.c HFA384X_OUTW(entry->param1, HFA384X_PARAM1_OFF); entry 290 drivers/net/wireless/intersil/hostap/hostap_hw.c HFA384X_OUTW(entry->cmd, HFA384X_CMD_OFF); entry 291 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->issued = 1; entry 317 drivers/net/wireless/intersil/hostap/hostap_hw.c struct hostap_cmd_queue *entry; entry 338 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 339 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry == NULL) entry 342 drivers/net/wireless/intersil/hostap/hostap_hw.c refcount_set(&entry->usecnt, 1); entry 343 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->type = CMD_SLEEP; entry 344 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->cmd = cmd; entry 345 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->param0 = param0; entry 347 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->param1 = *param1; entry 348 drivers/net/wireless/intersil/hostap/hostap_hw.c init_waitqueue_head(&entry->compl); entry 352 drivers/net/wireless/intersil/hostap/hostap_hw.c add_wait_queue(&entry->compl, &wait); entry 358 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->issuing = 1; entry 359 drivers/net/wireless/intersil/hostap/hostap_hw.c list_add_tail(&entry->list, &local->cmd_queue); entry 371 drivers/net/wireless/intersil/hostap/hostap_hw.c if (hfa384x_cmd_issue(dev, entry)) entry 378 drivers/net/wireless/intersil/hostap/hostap_hw.c if (!err && entry->type != CMD_COMPLETED) { entry 397 drivers/net/wireless/intersil/hostap/hostap_hw.c remove_wait_queue(&entry->compl, &wait); entry 417 drivers/net/wireless/intersil/hostap/hostap_hw.c if (!list_empty(&entry->list)) { entry 419 drivers/net/wireless/intersil/hostap/hostap_hw.c "(entry=%p, type=%d, res=%d)\n", dev->name, entry, entry 420 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->type, res); entry 421 drivers/net/wireless/intersil/hostap/hostap_hw.c list_del_init(&entry->list); entry 433 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry->type != CMD_COMPLETED) { entry 438 drivers/net/wireless/intersil/hostap/hostap_hw.c res, entry, entry->type, entry->cmd, entry->param0, reg, entry 453 drivers/net/wireless/intersil/hostap/hostap_hw.c *resp0 = entry->resp0; entry 455 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry->res) { entry 458 drivers/net/wireless/intersil/hostap/hostap_hw.c dev->name, cmd, entry->res, entry->resp0); entry 462 drivers/net/wireless/intersil/hostap/hostap_hw.c res = entry->res; entry 464 drivers/net/wireless/intersil/hostap/hostap_hw.c hostap_cmd_queue_free(local, entry, 1); entry 493 drivers/net/wireless/intersil/hostap/hostap_hw.c struct hostap_cmd_queue *entry; entry 504 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 505 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry == NULL) entry 508 drivers/net/wireless/intersil/hostap/hostap_hw.c refcount_set(&entry->usecnt, 1); entry 509 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->type = CMD_CALLBACK; entry 510 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->cmd = cmd; entry 511 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->param0 = param0; entry 512 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->callback = callback; entry 513 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->context = context; entry 518 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->issuing = 1; entry 519 drivers/net/wireless/intersil/hostap/hostap_hw.c list_add_tail(&entry->list, &local->cmd_queue); entry 523 drivers/net/wireless/intersil/hostap/hostap_hw.c if (issue && hfa384x_cmd_issue(dev, entry)) entry 528 drivers/net/wireless/intersil/hostap/hostap_hw.c hostap_cmd_queue_free(local, entry, ret); entry 649 drivers/net/wireless/intersil/hostap/hostap_hw.c struct hostap_cmd_queue *entry = NULL; entry 656 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = list_entry(local->cmd_queue.next, entry 658 drivers/net/wireless/intersil/hostap/hostap_hw.c refcount_inc(&entry->usecnt); entry 659 drivers/net/wireless/intersil/hostap/hostap_hw.c list_del_init(&entry->list); entry 662 drivers/net/wireless/intersil/hostap/hostap_hw.c if (!entry->issued) { entry 665 drivers/net/wireless/intersil/hostap/hostap_hw.c __hostap_cmd_queue_free(local, entry, 1); entry 666 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = NULL; entry 671 drivers/net/wireless/intersil/hostap/hostap_hw.c if (!entry) { entry 678 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->resp0 = HFA384X_INW(HFA384X_RESP0_OFF); entry 679 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->res = (HFA384X_INW(HFA384X_STATUS_OFF) & entry 685 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry->type == CMD_SLEEP) { entry 686 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->type = CMD_COMPLETED; entry 687 drivers/net/wireless/intersil/hostap/hostap_hw.c wake_up_interruptible(&entry->compl); entry 688 drivers/net/wireless/intersil/hostap/hostap_hw.c } else if (entry->type == CMD_CALLBACK) { entry 689 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry->callback) entry 690 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->callback(dev, entry->context, entry->resp0, entry 691 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->res); entry 694 drivers/net/wireless/intersil/hostap/hostap_hw.c dev->name, entry->type); entry 696 drivers/net/wireless/intersil/hostap/hostap_hw.c hostap_cmd_queue_free(local, entry, 1); entry 699 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = NULL; entry 702 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = list_entry(local->cmd_queue.next, entry 704 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry->issuing) { entry 707 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = NULL; entry 709 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry) entry 710 drivers/net/wireless/intersil/hostap/hostap_hw.c refcount_inc(&entry->usecnt); entry 714 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry) { entry 717 drivers/net/wireless/intersil/hostap/hostap_hw.c int res = hfa384x_cmd_issue(dev, entry); entry 719 drivers/net/wireless/intersil/hostap/hostap_hw.c __hostap_cmd_queue_free(local, entry, res); entry 2975 drivers/net/wireless/intersil/hostap/hostap_hw.c struct set_tim_data *entry = entry 2977 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry->aid == aid) { entry 2980 drivers/net/wireless/intersil/hostap/hostap_hw.c local->dev->name, aid, entry->set, set); entry 2981 drivers/net/wireless/intersil/hostap/hostap_hw.c entry->set = set; entry 3000 drivers/net/wireless/intersil/hostap/hostap_hw.c struct set_tim_data *entry; entry 3004 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = NULL; entry 3007 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = list_entry(local->set_tim_list.next, entry 3009 drivers/net/wireless/intersil/hostap/hostap_hw.c list_del(&entry->list); entry 3012 drivers/net/wireless/intersil/hostap/hostap_hw.c if (!entry) entry 3016 drivers/net/wireless/intersil/hostap/hostap_hw.c local->dev->name, entry->aid, entry->set); entry 3018 drivers/net/wireless/intersil/hostap/hostap_hw.c val = entry->aid; entry 3019 drivers/net/wireless/intersil/hostap/hostap_hw.c if (entry->set) entry 3024 drivers/net/wireless/intersil/hostap/hostap_hw.c local->dev->name, entry->aid, entry->set); entry 3027 drivers/net/wireless/intersil/hostap/hostap_hw.c kfree(entry); entry 3037 drivers/net/wireless/intersil/hostap/hostap_hw.c struct set_tim_data *entry; entry 3038 drivers/net/wireless/intersil/hostap/hostap_hw.c entry = list_entry(ptr, struct set_tim_data, list); entry 3039 drivers/net/wireless/intersil/hostap/hostap_hw.c list_del(&entry->list); entry 3040 drivers/net/wireless/intersil/hostap/hostap_hw.c kfree(entry); entry 172 drivers/net/wireless/intersil/hostap/hostap_info.c struct hfa384x_hostscan_result *selected, *entry; entry 210 drivers/net/wireless/intersil/hostap/hostap_info.c entry = &local->last_scan_results[i]; entry 211 drivers/net/wireless/intersil/hostap/hostap_info.c if (memcmp(local->preferred_ap, entry->bssid, 6) == 0) entry 215 drivers/net/wireless/intersil/hostap/hostap_info.c selected = entry; entry 646 drivers/net/wireless/intersil/hostap/hostap_ioctl.c struct hfa384x_hostscan_result *entry; entry 658 drivers/net/wireless/intersil/hostap/hostap_ioctl.c entry = &local->last_scan_results[i]; entry 659 drivers/net/wireless/intersil/hostap/hostap_ioctl.c if (ether_addr_equal(local->preferred_ap, entry->bssid)) { entry 660 drivers/net/wireless/intersil/hostap/hostap_ioctl.c req.channel = entry->chid; entry 1958 drivers/net/wireless/intersil/hostap/hostap_ioctl.c int entry, hostscan; entry 1972 drivers/net/wireless/intersil/hostap/hostap_ioctl.c for (entry = 0; entry < local->last_scan_results_count; entry++) { entry 1974 drivers/net/wireless/intersil/hostap/hostap_ioctl.c scan = &local->last_scan_results[entry]; entry 239 drivers/net/wireless/intersil/hostap/hostap_main.c struct hostap_tx_callback_info *entry; entry 241 drivers/net/wireless/intersil/hostap/hostap_main.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 242 drivers/net/wireless/intersil/hostap/hostap_main.c if (entry == NULL) entry 245 drivers/net/wireless/intersil/hostap/hostap_main.c entry->func = func; entry 246 drivers/net/wireless/intersil/hostap/hostap_main.c entry->data = data; entry 249 drivers/net/wireless/intersil/hostap/hostap_main.c entry->idx = local->tx_callback ? local->tx_callback->idx + 1 : 1; entry 250 drivers/net/wireless/intersil/hostap/hostap_main.c entry->next = local->tx_callback; entry 251 drivers/net/wireless/intersil/hostap/hostap_main.c local->tx_callback = entry; entry 254 drivers/net/wireless/intersil/hostap/hostap_main.c return entry->idx; entry 273 drivers/net/wireless/intersil/hostap/hostap_proc.c unsigned long entry; entry 284 drivers/net/wireless/intersil/hostap/hostap_proc.c entry = (unsigned long)v - 2; entry 285 drivers/net/wireless/intersil/hostap/hostap_proc.c scanres = &local->last_scan_results[entry]; entry 236 drivers/net/wireless/intersil/p54/eeprom.c struct p54_channel_entry *entry = NULL; entry 245 drivers/net/wireless/intersil/p54/eeprom.c entry = &list->channels[i]; entry 262 drivers/net/wireless/intersil/p54/eeprom.c entry = &list->channels[i]; entry 263 drivers/net/wireless/intersil/p54/eeprom.c entry->freq = freq; entry 264 drivers/net/wireless/intersil/p54/eeprom.c entry->band = band; entry 265 drivers/net/wireless/intersil/p54/eeprom.c entry->index = ieee80211_frequency_to_channel(freq); entry 266 drivers/net/wireless/intersil/p54/eeprom.c entry->max_power = 0; entry 267 drivers/net/wireless/intersil/p54/eeprom.c entry->data = 0; entry 271 drivers/net/wireless/intersil/p54/eeprom.c if (entry) entry 272 drivers/net/wireless/intersil/p54/eeprom.c entry->data |= data; entry 274 drivers/net/wireless/intersil/p54/eeprom.c return entry; entry 521 drivers/net/wireless/intersil/p54/eeprom.c struct p54_rssi_db_entry *entry; entry 550 drivers/net/wireless/intersil/p54/eeprom.c db_len = sizeof(*entry) * entries; entry 557 drivers/net/wireless/intersil/p54/eeprom.c priv->rssi_db->entry_size = sizeof(*entry); entry 560 drivers/net/wireless/intersil/p54/eeprom.c entry = (void *)((unsigned long)priv->rssi_db->data + priv->rssi_db->offset); entry 565 drivers/net/wireless/intersil/p54/eeprom.c entry[i].freq = le16_to_cpu(cal[i].freq); entry 566 drivers/net/wireless/intersil/p54/eeprom.c entry[i].mul = (s16) le16_to_cpu(cal[i].mul); entry 567 drivers/net/wireless/intersil/p54/eeprom.c entry[i].add = (s16) le16_to_cpu(cal[i].add); entry 583 drivers/net/wireless/intersil/p54/eeprom.c entry[i].freq = freq; entry 584 drivers/net/wireless/intersil/p54/eeprom.c entry[i].mul = (s16) le16_to_cpu(cal[i].mul); entry 585 drivers/net/wireless/intersil/p54/eeprom.c entry[i].add = (s16) le16_to_cpu(cal[i].add); entry 590 drivers/net/wireless/intersil/p54/eeprom.c sort(entry, entries, sizeof(*entry), p54_compare_rssichan, NULL); entry 606 drivers/net/wireless/intersil/p54/eeprom.c struct p54_rssi_db_entry *entry; entry 612 drivers/net/wireless/intersil/p54/eeprom.c entry = (void *)(priv->rssi_db->data + priv->rssi_db->offset); entry 614 drivers/net/wireless/intersil/p54/eeprom.c if (!same_band(freq, entry[i].freq)) entry 623 drivers/net/wireless/intersil/p54/eeprom.c if (abs(freq - entry[i].freq) < entry 624 drivers/net/wireless/intersil/p54/eeprom.c abs(freq - entry[found].freq)) { entry 632 drivers/net/wireless/intersil/p54/eeprom.c return found < 0 ? &p54_rssi_default : &entry[found]; entry 733 drivers/net/wireless/intersil/p54/eeprom.c struct pda_entry *entry; entry 742 drivers/net/wireless/intersil/p54/eeprom.c entry = (void *)wrap->data + le16_to_cpu(wrap->len); entry 745 drivers/net/wireless/intersil/p54/eeprom.c while ((u8 *)entry <= end - sizeof(*entry)) { entry 746 drivers/net/wireless/intersil/p54/eeprom.c entry_len = le16_to_cpu(entry->len); entry 750 drivers/net/wireless/intersil/p54/eeprom.c if ((u8 *)entry + sizeof(*entry) + data_len > end) entry 753 drivers/net/wireless/intersil/p54/eeprom.c switch (le16_to_cpu(entry->code)) { entry 757 drivers/net/wireless/intersil/p54/eeprom.c SET_IEEE80211_PERM_ADDR(dev, entry->data); entry 762 drivers/net/wireless/intersil/p54/eeprom.c err = p54_convert_output_limits(dev, entry->data, entry 769 drivers/net/wireless/intersil/p54/eeprom.c (struct pda_pa_curve_data *)entry->data; entry 794 drivers/net/wireless/intersil/p54/eeprom.c priv->iq_autocal = kmemdup(entry->data, data_len, entry 804 drivers/net/wireless/intersil/p54/eeprom.c p54_parse_default_country(dev, entry->data, data_len); entry 807 drivers/net/wireless/intersil/p54/eeprom.c tmp = entry->data; entry 808 drivers/net/wireless/intersil/p54/eeprom.c while ((u8 *)tmp < entry->data + data_len) { entry 818 drivers/net/wireless/intersil/p54/eeprom.c priv->version = *(u8 *)(entry->data + 1); entry 823 drivers/net/wireless/intersil/p54/eeprom.c err = p54_parse_rssical(dev, entry->data, data_len, entry 824 drivers/net/wireless/intersil/p54/eeprom.c le16_to_cpu(entry->code)); entry 829 drivers/net/wireless/intersil/p54/eeprom.c struct pda_custom_wrapper *pda = (void *) entry->data; entry 850 drivers/net/wireless/intersil/p54/eeprom.c struct pda_custom_wrapper *pda = (void *) entry->data; entry 857 drivers/net/wireless/intersil/p54/eeprom.c struct pda_custom_wrapper *pda = (void *) entry->data; entry 864 drivers/net/wireless/intersil/p54/eeprom.c crc16 = ~crc_ccitt(crc16, (u8 *) entry, sizeof(*entry)); entry 865 drivers/net/wireless/intersil/p54/eeprom.c if (crc16 != le16_to_cpup((__le16 *)entry->data)) { entry 878 drivers/net/wireless/intersil/p54/eeprom.c crc16 = crc_ccitt(crc16, (u8 *)entry, (entry_len + 1) * 2); entry 879 drivers/net/wireless/intersil/p54/eeprom.c entry = (void *)entry + (entry_len + 1) * 2; entry 274 drivers/net/wireless/intersil/p54/fwio.c tim->entry[0] = cpu_to_le16(set ? (aid | 0x8000) : aid); entry 402 drivers/net/wireless/intersil/p54/fwio.c void *entry; entry 470 drivers/net/wireless/intersil/p54/fwio.c entry = (void *)(priv->curve_data->data + priv->curve_data->offset); entry 472 drivers/net/wireless/intersil/p54/fwio.c if (*((__le16 *)entry) != freq) { entry 473 drivers/net/wireless/intersil/p54/fwio.c entry += priv->curve_data->entry_size; entry 479 drivers/net/wireless/intersil/p54/fwio.c entry + sizeof(__le16), entry 486 drivers/net/wireless/intersil/p54/fwio.c entry += sizeof(__le16); entry 489 drivers/net/wireless/intersil/p54/fwio.c memcpy(chan->curve_data, entry, entry 664 drivers/net/wireless/intersil/p54/fwio.c rxkey->entry = slot; entry 437 drivers/net/wireless/intersil/p54/lmac.h u8 entry; entry 503 drivers/net/wireless/intersil/p54/lmac.h __le16 entry[8]; entry 422 drivers/net/wireless/intersil/p54/p54spi.c struct p54s_tx_info *entry; entry 433 drivers/net/wireless/intersil/p54/p54spi.c entry = list_entry(priv->tx_pending.next, entry 436 drivers/net/wireless/intersil/p54/p54spi.c list_del_init(&entry->tx_list); entry 440 drivers/net/wireless/intersil/p54/p54spi.c dinfo = container_of((void *) entry, struct p54s_tx_info, entry 225 drivers/net/wireless/intersil/p54/p54usb.c struct urb *entry = NULL; entry 236 drivers/net/wireless/intersil/p54/p54usb.c entry = usb_alloc_urb(0, GFP_KERNEL); entry 237 drivers/net/wireless/intersil/p54/p54usb.c if (!entry) { entry 242 drivers/net/wireless/intersil/p54/p54usb.c usb_fill_bulk_urb(entry, priv->udev, entry 247 drivers/net/wireless/intersil/p54/p54usb.c info->urb = entry; entry 251 drivers/net/wireless/intersil/p54/p54usb.c usb_anchor_urb(entry, &priv->submitted); entry 252 drivers/net/wireless/intersil/p54/p54usb.c ret = usb_submit_urb(entry, GFP_KERNEL); entry 255 drivers/net/wireless/intersil/p54/p54usb.c usb_unanchor_urb(entry); entry 258 drivers/net/wireless/intersil/p54/p54usb.c usb_free_urb(entry); entry 259 drivers/net/wireless/intersil/p54/p54usb.c entry = NULL; entry 265 drivers/net/wireless/intersil/p54/p54usb.c usb_free_urb(entry); entry 81 drivers/net/wireless/intersil/p54/txrx.c struct sk_buff *entry, *target_skb = NULL; entry 105 drivers/net/wireless/intersil/p54/txrx.c skb_queue_walk(&priv->tx_queue, entry) { entry 107 drivers/net/wireless/intersil/p54/txrx.c info = IEEE80211_SKB_CB(entry); entry 112 drivers/net/wireless/intersil/p54/txrx.c target_skb = entry->prev; entry 251 drivers/net/wireless/intersil/p54/txrx.c struct sk_buff *entry; entry 255 drivers/net/wireless/intersil/p54/txrx.c skb_queue_walk(&priv->tx_queue, entry) { entry 256 drivers/net/wireless/intersil/p54/txrx.c struct p54_hdr *hdr = (struct p54_hdr *) entry->data; entry 259 drivers/net/wireless/intersil/p54/txrx.c __skb_unlink(entry, &priv->tx_queue); entry 261 drivers/net/wireless/intersil/p54/txrx.c p54_tx_qos_accounting_free(priv, entry); entry 262 drivers/net/wireless/intersil/p54/txrx.c return entry; entry 406 drivers/net/wireless/intersil/p54/txrx.c struct sk_buff *entry; entry 410 drivers/net/wireless/intersil/p54/txrx.c entry = p54_find_and_unlink_skb(priv, hdr->req_id); entry 411 drivers/net/wireless/intersil/p54/txrx.c if (unlikely(!entry)) entry 414 drivers/net/wireless/intersil/p54/txrx.c frame_len = entry->len; entry 415 drivers/net/wireless/intersil/p54/txrx.c info = IEEE80211_SKB_CB(entry); entry 416 drivers/net/wireless/intersil/p54/txrx.c entry_hdr = (struct p54_hdr *) entry->data; entry 426 drivers/net/wireless/intersil/p54/txrx.c dev_kfree_skb_any(entry); entry 487 drivers/net/wireless/intersil/p54/txrx.c skb_trim(entry, frame_len); entry 488 drivers/net/wireless/intersil/p54/txrx.c skb_pull(entry, sizeof(*hdr) + pad + sizeof(*entry_data)); entry 489 drivers/net/wireless/intersil/p54/txrx.c ieee80211_tx_status_irqsafe(priv->hw, entry); entry 1796 drivers/net/wireless/intersil/prism54/isl_ioctl.c struct mac_entry *entry; entry 1807 drivers/net/wireless/intersil/prism54/isl_ioctl.c entry = list_entry(ptr, struct mac_entry, _list); entry 1809 drivers/net/wireless/intersil/prism54/isl_ioctl.c kfree(entry); entry 1827 drivers/net/wireless/intersil/prism54/isl_ioctl.c struct mac_entry *entry; entry 1833 drivers/net/wireless/intersil/prism54/isl_ioctl.c entry = kmalloc(sizeof (struct mac_entry), GFP_KERNEL); entry 1834 drivers/net/wireless/intersil/prism54/isl_ioctl.c if (entry == NULL) entry 1837 drivers/net/wireless/intersil/prism54/isl_ioctl.c memcpy(entry->addr, addr->sa_data, ETH_ALEN); entry 1840 drivers/net/wireless/intersil/prism54/isl_ioctl.c kfree(entry); entry 1843 drivers/net/wireless/intersil/prism54/isl_ioctl.c list_add_tail(&entry->_list, &acl->mac_list); entry 1856 drivers/net/wireless/intersil/prism54/isl_ioctl.c struct mac_entry *entry; entry 1864 drivers/net/wireless/intersil/prism54/isl_ioctl.c list_for_each_entry(entry, &acl->mac_list, _list) { entry 1865 drivers/net/wireless/intersil/prism54/isl_ioctl.c if (ether_addr_equal(entry->addr, addr->sa_data)) { entry 1866 drivers/net/wireless/intersil/prism54/isl_ioctl.c list_del(&entry->_list); entry 1868 drivers/net/wireless/intersil/prism54/isl_ioctl.c kfree(entry); entry 1883 drivers/net/wireless/intersil/prism54/isl_ioctl.c struct mac_entry *entry; entry 1891 drivers/net/wireless/intersil/prism54/isl_ioctl.c list_for_each_entry(entry, &acl->mac_list, _list) { entry 1892 drivers/net/wireless/intersil/prism54/isl_ioctl.c memcpy(dst->sa_data, entry->addr, ETH_ALEN); entry 1958 drivers/net/wireless/intersil/prism54/isl_ioctl.c struct mac_entry *entry; entry 1969 drivers/net/wireless/intersil/prism54/isl_ioctl.c list_for_each_entry(entry, &acl->mac_list, _list) { entry 1970 drivers/net/wireless/intersil/prism54/isl_ioctl.c if (memcmp(entry->addr, mac, ETH_ALEN) == 0) { entry 3783 drivers/net/wireless/mac80211_hwsim.c struct mac80211_hwsim_data *entry, *tmp; entry 3787 drivers/net/wireless/mac80211_hwsim.c list_for_each_entry_safe(entry, tmp, &hwsim_radios, list) { entry 3788 drivers/net/wireless/mac80211_hwsim.c if (entry->destroy_on_close && entry->portid == portid) { entry 3789 drivers/net/wireless/mac80211_hwsim.c list_move(&entry->list, &list); entry 3790 drivers/net/wireless/mac80211_hwsim.c rhashtable_remove_fast(&hwsim_radios_rht, &entry->rht, entry 3797 drivers/net/wireless/mac80211_hwsim.c list_for_each_entry_safe(entry, tmp, &list, list) { entry 3798 drivers/net/wireless/mac80211_hwsim.c list_del(&entry->list); entry 3799 drivers/net/wireless/mac80211_hwsim.c mac80211_hwsim_del_radio(entry, wiphy_name(entry->hw->wiphy), entry 1204 drivers/net/wireless/marvell/mwifiex/main.c struct memory_type_mapping *entry = entry 1207 drivers/net/wireless/marvell/mwifiex/main.c if (entry->mem_ptr) { entry 1209 drivers/net/wireless/marvell/mwifiex/main.c strlen(entry->mem_name) + entry 1211 drivers/net/wireless/marvell/mwifiex/main.c (entry->mem_size + 1) + entry 1236 drivers/net/wireless/marvell/mwifiex/main.c struct memory_type_mapping *entry = entry 1239 drivers/net/wireless/marvell/mwifiex/main.c if (entry->mem_ptr) { entry 1243 drivers/net/wireless/marvell/mwifiex/main.c strcpy(fw_dump_ptr, entry->mem_name); entry 1244 drivers/net/wireless/marvell/mwifiex/main.c fw_dump_ptr += strlen(entry->mem_name); entry 1249 drivers/net/wireless/marvell/mwifiex/main.c memcpy(fw_dump_ptr, entry->mem_ptr, entry->mem_size); entry 1250 drivers/net/wireless/marvell/mwifiex/main.c fw_dump_ptr += entry->mem_size; entry 1260 drivers/net/wireless/marvell/mwifiex/main.c struct memory_type_mapping *entry = entry 1263 drivers/net/wireless/marvell/mwifiex/main.c vfree(entry->mem_ptr); entry 1264 drivers/net/wireless/marvell/mwifiex/main.c entry->mem_ptr = NULL; entry 1265 drivers/net/wireless/marvell/mwifiex/main.c entry->mem_size = 0; entry 2673 drivers/net/wireless/marvell/mwifiex/pcie.c struct memory_type_mapping *entry = entry 2676 drivers/net/wireless/marvell/mwifiex/pcie.c if (entry->mem_ptr) { entry 2677 drivers/net/wireless/marvell/mwifiex/pcie.c vfree(entry->mem_ptr); entry 2678 drivers/net/wireless/marvell/mwifiex/pcie.c entry->mem_ptr = NULL; entry 2680 drivers/net/wireless/marvell/mwifiex/pcie.c entry->mem_size = 0; entry 2701 drivers/net/wireless/marvell/mwifiex/pcie.c struct memory_type_mapping *entry = entry 2731 drivers/net/wireless/marvell/mwifiex/pcie.c "%s_SIZE=0x%x\n", entry->mem_name, memory_size); entry 2732 drivers/net/wireless/marvell/mwifiex/pcie.c entry->mem_ptr = vmalloc(memory_size + 1); entry 2733 drivers/net/wireless/marvell/mwifiex/pcie.c entry->mem_size = memory_size; entry 2734 drivers/net/wireless/marvell/mwifiex/pcie.c if (!entry->mem_ptr) { entry 2736 drivers/net/wireless/marvell/mwifiex/pcie.c "Vmalloc %s failed\n", entry->mem_name); entry 2739 drivers/net/wireless/marvell/mwifiex/pcie.c dbg_ptr = entry->mem_ptr; entry 2742 drivers/net/wireless/marvell/mwifiex/pcie.c doneflag = entry->done_flag; entry 2744 drivers/net/wireless/marvell/mwifiex/pcie.c entry->mem_name); entry 2765 drivers/net/wireless/marvell/mwifiex/pcie.c memcpy(tmp_ptr, entry->mem_ptr, memory_size); entry 2766 drivers/net/wireless/marvell/mwifiex/pcie.c vfree(entry->mem_ptr); entry 2767 drivers/net/wireless/marvell/mwifiex/pcie.c entry->mem_ptr = tmp_ptr; entry 2769 drivers/net/wireless/marvell/mwifiex/pcie.c dbg_ptr = entry->mem_ptr + memory_size; entry 2771 drivers/net/wireless/marvell/mwifiex/pcie.c end_ptr = entry->mem_ptr + memory_size; entry 2779 drivers/net/wireless/marvell/mwifiex/pcie.c entry->mem_name, dbg_ptr - entry->mem_ptr); entry 3027 drivers/net/wireless/marvell/mwifiex/pcie.c card->msix_entries[i].entry = i; entry 2313 drivers/net/wireless/marvell/mwifiex/sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; entry 2315 drivers/net/wireless/marvell/mwifiex/sdio.c if (entry->mem_ptr) { entry 2316 drivers/net/wireless/marvell/mwifiex/sdio.c vfree(entry->mem_ptr); entry 2317 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_ptr = NULL; entry 2319 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_size = 0; entry 2341 drivers/net/wireless/marvell/mwifiex/sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; entry 2372 drivers/net/wireless/marvell/mwifiex/sdio.c "%s_SIZE=0x%x\n", entry->mem_name, memory_size); entry 2373 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_ptr = vmalloc(memory_size + 1); entry 2374 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_size = memory_size; entry 2375 drivers/net/wireless/marvell/mwifiex/sdio.c if (!entry->mem_ptr) { entry 2377 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_name); entry 2380 drivers/net/wireless/marvell/mwifiex/sdio.c dbg_ptr = entry->mem_ptr; entry 2383 drivers/net/wireless/marvell/mwifiex/sdio.c doneflag = entry->done_flag; entry 2386 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_name); entry 2413 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_name, dbg_ptr - entry->mem_ptr); entry 2426 drivers/net/wireless/marvell/mwifiex/sdio.c struct memory_type_mapping *entry = &generic_mem_type_map[0]; entry 2436 drivers/net/wireless/marvell/mwifiex/sdio.c if (entry->mem_ptr) { entry 2437 drivers/net/wireless/marvell/mwifiex/sdio.c vfree(entry->mem_ptr); entry 2438 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_ptr = NULL; entry 2440 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_size = 0; entry 2473 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_ptr = vmalloc(0xf0000 + 1); entry 2474 drivers/net/wireless/marvell/mwifiex/sdio.c if (!entry->mem_ptr) { entry 2478 drivers/net/wireless/marvell/mwifiex/sdio.c dbg_ptr = entry->mem_ptr; entry 2479 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_size = 0xf0000; entry 2480 drivers/net/wireless/marvell/mwifiex/sdio.c end_ptr = dbg_ptr + entry->mem_size; entry 2482 drivers/net/wireless/marvell/mwifiex/sdio.c done_flag = entry->done_flag; entry 2484 drivers/net/wireless/marvell/mwifiex/sdio.c "Start %s output, please wait...\n", entry->mem_name); entry 2501 drivers/net/wireless/marvell/mwifiex/sdio.c tmp_ptr = vmalloc(entry->mem_size + 0x4000 + 1); entry 2505 drivers/net/wireless/marvell/mwifiex/sdio.c memcpy(tmp_ptr, entry->mem_ptr, entry 2506 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_size); entry 2507 drivers/net/wireless/marvell/mwifiex/sdio.c vfree(entry->mem_ptr); entry 2508 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_ptr = tmp_ptr; entry 2510 drivers/net/wireless/marvell/mwifiex/sdio.c dbg_ptr = entry->mem_ptr + entry->mem_size; entry 2511 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_size += 0x4000; entry 2512 drivers/net/wireless/marvell/mwifiex/sdio.c end_ptr = entry->mem_ptr + entry->mem_size; entry 2516 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_size = dbg_ptr - entry->mem_ptr; entry 2518 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_name, entry->mem_size); entry 2528 drivers/net/wireless/marvell/mwifiex/sdio.c if (entry->mem_ptr) { entry 2529 drivers/net/wireless/marvell/mwifiex/sdio.c vfree(entry->mem_ptr); entry 2530 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_ptr = NULL; entry 2532 drivers/net/wireless/marvell/mwifiex/sdio.c entry->mem_size = 0; entry 30 drivers/net/wireless/mediatek/mt76/dma.c size = q->ndesc * sizeof(*q->entry); entry 31 drivers/net/wireless/mediatek/mt76/dma.c q->entry = devm_kzalloc(dev->dev, size, GFP_KERNEL); entry 32 drivers/net/wireless/mediatek/mt76/dma.c if (!q->entry) entry 57 drivers/net/wireless/mediatek/mt76/dma.c q->entry[q->head].txwi = DMA_DUMMY_DATA; entry 58 drivers/net/wireless/mediatek/mt76/dma.c q->entry[q->head].skip_buf0 = true; entry 88 drivers/net/wireless/mediatek/mt76/dma.c q->entry[idx].txwi = txwi; entry 89 drivers/net/wireless/mediatek/mt76/dma.c q->entry[idx].skb = skb; entry 98 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_queue_entry *e = &q->entry[idx]; entry 143 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_queue_entry entry; entry 158 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_tx_cleanup_idx(dev, q, q->tail, &entry); entry 159 drivers/net/wireless/mediatek/mt76/dma.c if (entry.schedule) entry 160 drivers/net/wireless/mediatek/mt76/dma.c n_swq_queued[entry.qid]++; entry 165 drivers/net/wireless/mediatek/mt76/dma.c if (entry.skb) entry 166 drivers/net/wireless/mediatek/mt76/dma.c dev->drv->tx_complete_skb(dev, qid, &entry); entry 168 drivers/net/wireless/mediatek/mt76/dma.c if (entry.txwi) { entry 170 drivers/net/wireless/mediatek/mt76/dma.c mt76_put_txwi(dev, entry.txwi); entry 209 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_queue_entry *e = &q->entry[idx]; entry 112 drivers/net/wireless/mediatek/mt76/mt76.h struct mt76_queue_entry *entry; entry 462 drivers/net/wireless/mediatek/mt76/tx.c hwq->entry[idx].qid = sq - dev->q_tx; entry 463 drivers/net/wireless/mediatek/mt76/tx.c hwq->entry[idx].schedule = true; entry 399 drivers/net/wireless/mediatek/mt76/usb.c urb = q->entry[q->head].urb; entry 517 drivers/net/wireless/mediatek/mt76/usb.c if (WARN_ONCE(q->entry[q->tail].urb != urb, "rx urb mismatch")) entry 571 drivers/net/wireless/mediatek/mt76/usb.c err = mt76u_submit_rx_buf(dev, q->entry[i].urb); entry 593 drivers/net/wireless/mediatek/mt76/usb.c q->entry = devm_kcalloc(dev->dev, entry 594 drivers/net/wireless/mediatek/mt76/usb.c MT_NUM_RX_ENTRIES, sizeof(*q->entry), entry 596 drivers/net/wireless/mediatek/mt76/usb.c if (!q->entry) entry 603 drivers/net/wireless/mediatek/mt76/usb.c err = mt76u_rx_urb_alloc(dev, &q->entry[i]); entry 618 drivers/net/wireless/mediatek/mt76/usb.c mt76u_urb_free(q->entry[i].urb); entry 634 drivers/net/wireless/mediatek/mt76/usb.c usb_poison_urb(q->entry[i].urb); entry 646 drivers/net/wireless/mediatek/mt76/usb.c usb_unpoison_urb(q->entry[i].urb); entry 655 drivers/net/wireless/mediatek/mt76/usb.c struct mt76_queue_entry entry; entry 668 drivers/net/wireless/mediatek/mt76/usb.c if (!q->entry[q->head].done) entry 671 drivers/net/wireless/mediatek/mt76/usb.c if (q->entry[q->head].schedule) { entry 672 drivers/net/wireless/mediatek/mt76/usb.c q->entry[q->head].schedule = false; entry 676 drivers/net/wireless/mediatek/mt76/usb.c entry = q->entry[q->head]; entry 677 drivers/net/wireless/mediatek/mt76/usb.c q->entry[q->head].done = false; entry 681 drivers/net/wireless/mediatek/mt76/usb.c dev->drv->tx_complete_skb(dev, i, &entry); entry 787 drivers/net/wireless/mediatek/mt76/usb.c err = mt76u_tx_setup_buffers(dev, tx_info.skb, q->entry[idx].urb); entry 792 drivers/net/wireless/mediatek/mt76/usb.c q->entry[idx].urb, mt76u_complete_tx, entry 793 drivers/net/wireless/mediatek/mt76/usb.c &q->entry[idx]); entry 796 drivers/net/wireless/mediatek/mt76/usb.c q->entry[idx].skb = tx_info.skb; entry 808 drivers/net/wireless/mediatek/mt76/usb.c urb = q->entry[q->first].urb; entry 845 drivers/net/wireless/mediatek/mt76/usb.c q->entry = devm_kcalloc(dev->dev, entry 846 drivers/net/wireless/mediatek/mt76/usb.c MT_NUM_TX_ENTRIES, sizeof(*q->entry), entry 848 drivers/net/wireless/mediatek/mt76/usb.c if (!q->entry) entry 853 drivers/net/wireless/mediatek/mt76/usb.c err = mt76u_urb_alloc(dev, &q->entry[j], entry 870 drivers/net/wireless/mediatek/mt76/usb.c usb_free_urb(q->entry[j].urb); entry 876 drivers/net/wireless/mediatek/mt76/usb.c struct mt76_queue_entry entry; entry 888 drivers/net/wireless/mediatek/mt76/usb.c usb_kill_urb(q->entry[j].urb); entry 902 drivers/net/wireless/mediatek/mt76/usb.c entry = q->entry[q->head]; entry 906 drivers/net/wireless/mediatek/mt76/usb.c dev->drv->tx_complete_skb(dev, i, &entry); entry 714 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static bool rt2400pci_get_entry_state(struct queue_entry *entry) entry 716 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 719 drivers/net/wireless/ralink/rt2x00/rt2400pci.c if (entry->queue->qid == QID_RX) { entry 731 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static void rt2400pci_clear_entry(struct queue_entry *entry) entry 733 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 734 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 737 drivers/net/wireless/ralink/rt2x00/rt2400pci.c if (entry->queue->qid == QID_RX) { entry 739 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_set_field32(&word, RXD_W2_BUFFER_LENGTH, entry->skb->len); entry 1094 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static void rt2400pci_write_tx_desc(struct queue_entry *entry, entry 1097 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1098 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 1165 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static void rt2400pci_write_beacon(struct queue_entry *entry, entry 1168 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1179 drivers/net/wireless/ralink/rt2x00/rt2400pci.c if (rt2x00queue_map_txskb(entry)) { entry 1190 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2400pci_write_tx_desc(entry, txdesc); entry 1195 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry); entry 1207 drivers/net/wireless/ralink/rt2x00/rt2400pci.c static void rt2400pci_fill_rxdone(struct queue_entry *entry, entry 1210 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1211 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 1254 drivers/net/wireless/ralink/rt2x00/rt2400pci.c entry->queue->rt2x00dev->rssi_offset; entry 1270 drivers/net/wireless/ralink/rt2x00/rt2400pci.c struct queue_entry *entry; entry 1275 drivers/net/wireless/ralink/rt2x00/rt2400pci.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE); entry 1276 drivers/net/wireless/ralink/rt2x00/rt2400pci.c entry_priv = entry->priv_data; entry 1300 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00lib_txdone(entry, &txdesc); entry 803 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static bool rt2500pci_get_entry_state(struct queue_entry *entry) entry 805 drivers/net/wireless/ralink/rt2x00/rt2500pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 808 drivers/net/wireless/ralink/rt2x00/rt2500pci.c if (entry->queue->qid == QID_RX) { entry 820 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static void rt2500pci_clear_entry(struct queue_entry *entry) entry 822 drivers/net/wireless/ralink/rt2x00/rt2500pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 823 drivers/net/wireless/ralink/rt2x00/rt2500pci.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 826 drivers/net/wireless/ralink/rt2x00/rt2500pci.c if (entry->queue->qid == QID_RX) { entry 1247 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static void rt2500pci_write_tx_desc(struct queue_entry *entry, entry 1250 drivers/net/wireless/ralink/rt2x00/rt2500pci.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1251 drivers/net/wireless/ralink/rt2x00/rt2500pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 1264 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_set_field32(&word, TXD_W2_AIFS, entry->queue->aifs); entry 1265 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_set_field32(&word, TXD_W2_CWMIN, entry->queue->cw_min); entry 1266 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_set_field32(&word, TXD_W2_CWMAX, entry->queue->cw_max); entry 1317 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static void rt2500pci_write_beacon(struct queue_entry *entry, entry 1320 drivers/net/wireless/ralink/rt2x00/rt2500pci.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1331 drivers/net/wireless/ralink/rt2x00/rt2500pci.c if (rt2x00queue_map_txskb(entry)) { entry 1339 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2500pci_write_tx_desc(entry, txdesc); entry 1344 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry); entry 1356 drivers/net/wireless/ralink/rt2x00/rt2500pci.c static void rt2500pci_fill_rxdone(struct queue_entry *entry, entry 1359 drivers/net/wireless/ralink/rt2x00/rt2500pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 1379 drivers/net/wireless/ralink/rt2x00/rt2500pci.c entry->queue->rt2x00dev->rssi_offset; entry 1398 drivers/net/wireless/ralink/rt2x00/rt2500pci.c struct queue_entry *entry; entry 1403 drivers/net/wireless/ralink/rt2x00/rt2500pci.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE); entry 1404 drivers/net/wireless/ralink/rt2x00/rt2500pci.c entry_priv = entry->priv_data; entry 1428 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00lib_txdone(entry, &txdesc); entry 1056 drivers/net/wireless/ralink/rt2x00/rt2500usb.c static void rt2500usb_write_tx_desc(struct queue_entry *entry, entry 1059 drivers/net/wireless/ralink/rt2x00/rt2500usb.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1060 drivers/net/wireless/ralink/rt2x00/rt2500usb.c __le32 *txd = (__le32 *) entry->skb->data; entry 1086 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00_set_field32(&word, TXD_W1_AIFS, entry->queue->aifs); entry 1087 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min); entry 1088 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max); entry 1118 drivers/net/wireless/ralink/rt2x00/rt2500usb.c static void rt2500usb_write_beacon(struct queue_entry *entry, entry 1121 drivers/net/wireless/ralink/rt2x00/rt2500usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1123 drivers/net/wireless/ralink/rt2x00/rt2500usb.c struct queue_entry_priv_usb_bcn *bcn_priv = entry->priv_data; entry 1124 drivers/net/wireless/ralink/rt2x00/rt2500usb.c int pipe = usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint); entry 1139 drivers/net/wireless/ralink/rt2x00/rt2500usb.c skb_push(entry->skb, TXD_DESC_SIZE); entry 1140 drivers/net/wireless/ralink/rt2x00/rt2500usb.c memset(entry->skb->data, 0, TXD_DESC_SIZE); entry 1145 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2500usb_write_tx_desc(entry, txdesc); entry 1150 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry); entry 1157 drivers/net/wireless/ralink/rt2x00/rt2500usb.c length = rt2x00dev->ops->lib->get_tx_data_len(entry); entry 1160 drivers/net/wireless/ralink/rt2x00/rt2500usb.c entry->skb->data, length, rt2500usb_beacondone, entry 1161 drivers/net/wireless/ralink/rt2x00/rt2500usb.c entry); entry 1171 drivers/net/wireless/ralink/rt2x00/rt2500usb.c entry); entry 1199 drivers/net/wireless/ralink/rt2x00/rt2500usb.c static int rt2500usb_get_tx_data_len(struct queue_entry *entry) entry 1207 drivers/net/wireless/ralink/rt2x00/rt2500usb.c length = roundup(entry->skb->len, 2); entry 1208 drivers/net/wireless/ralink/rt2x00/rt2500usb.c length += (2 * !(length % entry->queue->usb_maxpacket)); entry 1216 drivers/net/wireless/ralink/rt2x00/rt2500usb.c static void rt2500usb_fill_rxdone(struct queue_entry *entry, entry 1219 drivers/net/wireless/ralink/rt2x00/rt2500usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1220 drivers/net/wireless/ralink/rt2x00/rt2500usb.c struct queue_entry_priv_usb *entry_priv = entry->priv_data; entry 1221 drivers/net/wireless/ralink/rt2x00/rt2500usb.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1223 drivers/net/wireless/ralink/rt2x00/rt2500usb.c (__le32 *)(entry->skb->data + entry 1225 drivers/net/wireless/ralink/rt2x00/rt2500usb.c entry->queue->desc_size)); entry 1286 drivers/net/wireless/ralink/rt2x00/rt2500usb.c skb_trim(entry->skb, rxdesc->size); entry 1294 drivers/net/wireless/ralink/rt2x00/rt2500usb.c struct queue_entry *entry = (struct queue_entry *)urb->context; entry 1295 drivers/net/wireless/ralink/rt2x00/rt2500usb.c struct queue_entry_priv_usb_bcn *bcn_priv = entry->priv_data; entry 1297 drivers/net/wireless/ralink/rt2x00/rt2500usb.c if (!test_bit(DEVICE_STATE_ENABLED_RADIO, &entry->queue->rt2x00dev->flags)) entry 1309 drivers/net/wireless/ralink/rt2x00/rt2500usb.c dev_kfree_skb(entry->skb); entry 1310 drivers/net/wireless/ralink/rt2x00/rt2500usb.c entry->skb = NULL; entry 773 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_write_tx_data(struct queue_entry *entry, entry 776 drivers/net/wireless/ralink/rt2x00/rt2800lib.c __le32 *txwi = rt2800_drv_get_txwi(entry); entry 816 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_set_field32(&word, TXWI_W1_PACKETID_QUEUE, entry->queue->qid); entry 817 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_set_field32(&word, TXWI_W1_PACKETID_ENTRY, (entry->entry_idx % 3) + 1); entry 829 drivers/net/wireless/ralink/rt2x00/rt2800lib.c for (i = 2; i < entry->queue->winfo_size / sizeof(__le32); i++) entry 877 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_process_rxwi(struct queue_entry *entry, entry 880 drivers/net/wireless/ralink/rt2x00/rt2800lib.c __le32 *rxwi = (__le32 *) entry->skb->data; entry 914 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rxdesc->rssi = rt2800_agc_to_rssi(entry->queue->rt2x00dev, word); entry 918 drivers/net/wireless/ralink/rt2x00/rt2800lib.c skb_pull(entry->skb, entry->queue->winfo_size); entry 955 drivers/net/wireless/ralink/rt2x00/rt2800lib.c static bool rt2800_txdone_entry_check(struct queue_entry *entry, u32 reg) entry 967 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (test_bit(ENTRY_DATA_IO_FAILED, &entry->flags)) entry 979 drivers/net/wireless/ralink/rt2x00/rt2800lib.c txwi = rt2800_drv_get_txwi(entry); entry 987 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_dbg(entry->queue->rt2x00dev, entry 989 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->queue->qid, entry->entry_idx); entry 996 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_txdone_entry(struct queue_entry *entry, u32 status, __le32 *txwi, entry 999 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1001 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1090 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00lib_txdone_nomatch(entry, &txdesc); entry 1093 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00lib_txdone(entry, &txdesc); entry 1101 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct queue_entry *entry; entry 1120 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE); entry 1122 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (unlikely(test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags) || entry 1123 drivers/net/wireless/ralink/rt2x00/rt2800lib.c !test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags))) { entry 1125 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->entry_idx, qid); entry 1129 drivers/net/wireless/ralink/rt2x00/rt2800lib.c match = rt2800_txdone_entry_check(entry, reg); entry 1130 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2800_txdone_entry(entry, reg, rt2800_drv_get_txwi(entry), match); entry 1136 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct queue_entry *entry) entry 1141 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (!test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags)) entry 1149 drivers/net/wireless/ralink/rt2x00/rt2800lib.c ret = time_after(jiffies, entry->last_action + tout); entry 1151 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_dbg(entry->queue->rt2x00dev, entry 1153 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->entry_idx, entry->queue->qid); entry 1160 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct queue_entry *entry; entry 1163 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE); entry 1164 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (rt2800_entry_txstatus_timeout(rt2x00dev, entry)) entry 1192 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct queue_entry *entry; entry 1203 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE); entry 1205 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags) || entry 1206 drivers/net/wireless/ralink/rt2x00/rt2800lib.c !test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags)) entry 1209 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (test_bit(ENTRY_DATA_IO_FAILED, &entry->flags) || entry 1210 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2800_entry_txstatus_timeout(rt2x00dev, entry)) entry 1211 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00lib_txdone_noinfo(entry, TXDONE_FAILURE); entry 1291 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct queue_entry *entry; entry 1300 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry = &queue->entries[i]; entry 1301 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (!test_bit(ENTRY_BCN_ENABLED, &entry->flags)) entry 1303 drivers/net/wireless/ralink/rt2x00/rt2800lib.c off = rt2800_get_beacon_offset(rt2x00dev, entry->entry_idx); entry 1320 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_write_beacon(struct queue_entry *entry, struct txentry_desc *txdesc) entry 1322 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1323 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1327 drivers/net/wireless/ralink/rt2x00/rt2800lib.c const int txwi_desc_size = entry->queue->winfo_size; entry 1341 drivers/net/wireless/ralink/rt2x00/rt2800lib.c memset(skb_push(entry->skb, txwi_desc_size), 0, txwi_desc_size); entry 1347 drivers/net/wireless/ralink/rt2x00/rt2800lib.c skbdesc->desc = entry->skb->data; entry 1353 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2800_write_tx_data(entry, txdesc); entry 1358 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry); entry 1363 drivers/net/wireless/ralink/rt2x00/rt2800lib.c padding_len = roundup(entry->skb->len, 4) - entry->skb->len; entry 1364 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (padding_len && skb_pad(entry->skb, padding_len)) { entry 1367 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->skb = NULL; entry 1372 drivers/net/wireless/ralink/rt2x00/rt2800lib.c beacon_base = rt2800_hw_beacon_base(rt2x00dev, entry->entry_idx); entry 1374 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2800_register_multiwrite(rt2x00dev, beacon_base, entry->skb->data, entry 1375 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->skb->len + padding_len); entry 1376 drivers/net/wireless/ralink/rt2x00/rt2800lib.c __set_bit(ENTRY_BCN_ENABLED, &entry->flags); entry 1391 drivers/net/wireless/ralink/rt2x00/rt2800lib.c dev_kfree_skb_any(entry->skb); entry 1392 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->skb = NULL; entry 1414 drivers/net/wireless/ralink/rt2x00/rt2800lib.c void rt2800_clear_beacon(struct queue_entry *entry) entry 1416 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1431 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2800_clear_beacon_register(rt2x00dev, entry->entry_idx); entry 1432 drivers/net/wireless/ralink/rt2x00/rt2800lib.c __clear_bit(ENTRY_BCN_ENABLED, &entry->flags); entry 1863 drivers/net/wireless/ralink/rt2x00/rt2800lib.c struct queue_entry *entry; entry 1872 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry = &queue->entries[i]; entry 1873 drivers/net/wireless/ralink/rt2x00/rt2800lib.c clear_bit(ENTRY_BCN_ASSIGNED, &entry->flags); entry 67 drivers/net/wireless/ralink/rt2x00/rt2800lib.h __le32 *(*drv_get_txwi)(struct queue_entry *entry); entry 163 drivers/net/wireless/ralink/rt2x00/rt2800lib.h static inline __le32 *rt2800_drv_get_txwi(struct queue_entry *entry) entry 165 drivers/net/wireless/ralink/rt2x00/rt2800lib.h const struct rt2800_ops *rt2800ops = entry->queue->rt2x00dev->ops->drv; entry 167 drivers/net/wireless/ralink/rt2x00/rt2800lib.h return rt2800ops->drv_get_txwi(entry); entry 189 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_write_tx_data(struct queue_entry *entry, entry 191 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_process_rxwi(struct queue_entry *entry, struct rxdone_entry_desc *txdesc); entry 193 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_txdone_entry(struct queue_entry *entry, u32 status, __le32 *txwi, entry 202 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_write_beacon(struct queue_entry *entry, struct txentry_desc *txdesc); entry 203 drivers/net/wireless/ralink/rt2x00/rt2800lib.h void rt2800_clear_beacon(struct queue_entry *entry); entry 30 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct queue_entry *entry; entry 45 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE); entry 46 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c idx = entry->entry_idx; entry 61 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c __le32 *rt2800mmio_get_txwi(struct queue_entry *entry) entry 63 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c return (__le32 *) entry->skb->data; entry 67 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c void rt2800mmio_write_tx_desc(struct queue_entry *entry, entry 70 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 71 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 74 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c const unsigned int txwi_size = entry->queue->winfo_size; entry 92 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00_set_field32(&word, TXD_W1_SD_LEN1, entry->skb->len); entry 124 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c void rt2800mmio_fill_rxdone(struct queue_entry *entry, entry 127 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 182 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2800_process_rxwi(entry, rxdesc); entry 454 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct queue_entry *entry; entry 462 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry = rt2x00queue_get_entry(queue, Q_INDEX); entry 464 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry->entry_idx); entry 469 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry = rt2x00queue_get_entry(queue, Q_INDEX); entry 471 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry->entry_idx); entry 607 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c bool rt2800mmio_get_entry_state(struct queue_entry *entry) entry 609 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 612 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c if (entry->queue->qid == QID_RX) { entry 624 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c void rt2800mmio_clear_entry(struct queue_entry *entry) entry 626 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 627 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 628 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 631 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c if (entry->queue->qid == QID_RX) { entry 645 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry->entry_idx); entry 652 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c if (entry->queue->length == 1) entry 120 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h __le32 *rt2800mmio_get_txwi(struct queue_entry *entry); entry 121 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h void rt2800mmio_write_tx_desc(struct queue_entry *entry, entry 125 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h void rt2800mmio_fill_rxdone(struct queue_entry *entry, entry 147 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h bool rt2800mmio_get_entry_state(struct queue_entry *entry); entry 148 drivers/net/wireless/ralink/rt2x00/rt2800mmio.h void rt2800mmio_clear_entry(struct queue_entry *entry); entry 155 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static void rt2800usb_tx_dma_done(struct queue_entry *entry) entry 157 drivers/net/wireless/ralink/rt2x00/rt2800usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 384 drivers/net/wireless/ralink/rt2x00/rt2800usb.c struct queue_entry *entry; entry 386 drivers/net/wireless/ralink/rt2x00/rt2800usb.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE); entry 387 drivers/net/wireless/ralink/rt2x00/rt2800usb.c return entry->entry_idx; entry 393 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static __le32 *rt2800usb_get_txwi(struct queue_entry *entry) entry 395 drivers/net/wireless/ralink/rt2x00/rt2800usb.c if (entry->queue->qid == QID_BEACON) entry 396 drivers/net/wireless/ralink/rt2x00/rt2800usb.c return (__le32 *) (entry->skb->data); entry 398 drivers/net/wireless/ralink/rt2x00/rt2800usb.c return (__le32 *) (entry->skb->data + TXINFO_DESC_SIZE); entry 401 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static void rt2800usb_write_tx_desc(struct queue_entry *entry, entry 404 drivers/net/wireless/ralink/rt2x00/rt2800usb.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 405 drivers/net/wireless/ralink/rt2x00/rt2800usb.c __le32 *txi = (__le32 *) entry->skb->data; entry 419 drivers/net/wireless/ralink/rt2x00/rt2800usb.c roundup(entry->skb->len, 4) - TXINFO_DESC_SIZE); entry 434 drivers/net/wireless/ralink/rt2x00/rt2800usb.c skbdesc->desc_len = TXINFO_DESC_SIZE + entry->queue->winfo_size; entry 440 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static int rt2800usb_get_tx_data_len(struct queue_entry *entry) entry 450 drivers/net/wireless/ralink/rt2x00/rt2800usb.c return roundup(entry->skb->len, 4) + 4; entry 481 drivers/net/wireless/ralink/rt2x00/rt2800usb.c static void rt2800usb_fill_rxdone(struct queue_entry *entry, entry 484 drivers/net/wireless/ralink/rt2x00/rt2800usb.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 485 drivers/net/wireless/ralink/rt2x00/rt2800usb.c __le32 *rxi = (__le32 *)entry->skb->data; entry 507 drivers/net/wireless/ralink/rt2x00/rt2800usb.c skb_pull(entry->skb, RXINFO_DESC_SIZE); entry 514 drivers/net/wireless/ralink/rt2x00/rt2800usb.c rx_pkt_len > entry->queue->data_size)) { entry 515 drivers/net/wireless/ralink/rt2x00/rt2800usb.c rt2x00_err(entry->queue->rt2x00dev, entry 520 drivers/net/wireless/ralink/rt2x00/rt2800usb.c rxd = (__le32 *)(entry->skb->data + rx_pkt_len); entry 571 drivers/net/wireless/ralink/rt2x00/rt2800usb.c skb_trim(entry->skb, rx_pkt_len); entry 576 drivers/net/wireless/ralink/rt2x00/rt2800usb.c rt2800_process_rxwi(entry, rxdesc); entry 546 drivers/net/wireless/ralink/rt2x00/rt2x00.h bool (*get_entry_state) (struct queue_entry *entry); entry 547 drivers/net/wireless/ralink/rt2x00/rt2x00.h void (*clear_entry) (struct queue_entry *entry); entry 572 drivers/net/wireless/ralink/rt2x00/rt2x00.h void (*tx_dma_done) (struct queue_entry *entry); entry 577 drivers/net/wireless/ralink/rt2x00/rt2x00.h void (*write_tx_desc) (struct queue_entry *entry, entry 579 drivers/net/wireless/ralink/rt2x00/rt2x00.h void (*write_tx_data) (struct queue_entry *entry, entry 581 drivers/net/wireless/ralink/rt2x00/rt2x00.h void (*write_beacon) (struct queue_entry *entry, entry 583 drivers/net/wireless/ralink/rt2x00/rt2x00.h void (*clear_beacon) (struct queue_entry *entry); entry 584 drivers/net/wireless/ralink/rt2x00/rt2x00.h int (*get_tx_data_len) (struct queue_entry *entry); entry 589 drivers/net/wireless/ralink/rt2x00/rt2x00.h void (*fill_rxdone) (struct queue_entry *entry, entry 1020 drivers/net/wireless/ralink/rt2x00/rt2x00.h struct queue_entry *entry; entry 1285 drivers/net/wireless/ralink/rt2x00/rt2x00.h int rt2x00queue_map_txskb(struct queue_entry *entry); entry 1291 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00queue_unmap_skb(struct queue_entry *entry); entry 1404 drivers/net/wireless/ralink/rt2x00/rt2x00.h enum rt2x00_dump_type type, struct queue_entry *entry); entry 1408 drivers/net/wireless/ralink/rt2x00/rt2x00.h struct queue_entry *entry) entry 1425 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_dmastart(struct queue_entry *entry); entry 1426 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_dmadone(struct queue_entry *entry); entry 1427 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_txdone(struct queue_entry *entry, entry 1429 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_txdone_nomatch(struct queue_entry *entry, entry 1431 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_txdone_noinfo(struct queue_entry *entry, u32 status); entry 1432 drivers/net/wireless/ralink/rt2x00/rt2x00.h void rt2x00lib_rxdone(struct queue_entry *entry, gfp_t gfp); entry 131 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c enum rt2x00_dump_type type, struct queue_entry *entry) entry 134 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c struct sk_buff *skb = entry->skb; entry 171 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c dump_hdr->queue_index = entry->queue->qid; entry 172 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c dump_hdr->entry_index = entry->entry_idx; entry 246 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_dmastart(struct queue_entry *entry) entry 248 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c set_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags); entry 249 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_index_inc(entry, Q_INDEX); entry 253 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_dmadone(struct queue_entry *entry) entry 255 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c set_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags); entry 256 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c clear_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags); entry 257 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_index_inc(entry, Q_INDEX_DMA_DONE); entry 261 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c static inline int rt2x00lib_txdone_bar_status(struct queue_entry *entry) entry 263 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 264 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct ieee80211_bar *bar = (void *) entry->skb->data; entry 287 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c if (bar_entry->entry != entry) entry 385 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct queue_entry *entry) entry 390 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->skb = NULL; entry 391 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->flags = 0; entry 393 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00dev->ops->lib->clear_entry(entry); entry 395 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_index_inc(entry, Q_INDEX_DONE); entry 404 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c spin_lock_bh(&entry->queue->tx_lock); entry 405 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c if (!rt2x00queue_threshold(entry->queue)) entry 406 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_unpause_queue(entry->queue); entry 407 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c spin_unlock_bh(&entry->queue->tx_lock); entry 410 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_txdone_nomatch(struct queue_entry *entry, entry 413 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 414 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 421 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_unmap_skb(entry); entry 432 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_TXDONE, entry); entry 440 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00lib_txdone_bar_status(entry) || entry 455 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c dev_kfree_skb_any(entry->skb); entry 456 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00lib_clear_entry(rt2x00dev, entry); entry 460 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_txdone(struct queue_entry *entry, entry 463 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 464 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(entry->skb); entry 465 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 473 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_unmap_skb(entry); entry 478 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c skb_pull(entry->skb, rt2x00dev->extra_tx_headroom); entry 488 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c header_length = ieee80211_get_hdrlen_from_skb(entry->skb); entry 494 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_remove_l2pad(entry->skb, header_length); entry 503 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00crypto_tx_insert_iv(entry->skb, header_length); entry 509 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_TXDONE, entry); entry 517 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00lib_txdone_bar_status(entry) || entry 537 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c ieee80211_tx_status(rt2x00dev->hw, entry->skb); entry 539 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c ieee80211_tx_status_ni(rt2x00dev->hw, entry->skb); entry 541 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c dev_kfree_skb_any(entry->skb); entry 544 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00lib_clear_entry(rt2x00dev, entry); entry 548 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_txdone_noinfo(struct queue_entry *entry, u32 status) entry 556 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00lib_txdone(entry, &txdesc); entry 601 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct rt2x00_bar_list_entry *entry; entry 611 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c list_for_each_entry_rcu(entry, &rt2x00dev->bar_list, list) { entry 613 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c if (ba->start_seq_num != entry->start_seq_num) entry 620 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c if (!TID_CHECK(ba->control, entry->control)) entry 625 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c if (!ether_addr_equal_64bits(ba->ra, entry->ta)) entry 628 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c if (!ether_addr_equal_64bits(ba->ta, entry->ra)) entry 633 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->block_acked = 1; entry 737 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c void rt2x00lib_rxdone(struct queue_entry *entry, gfp_t gfp) entry 739 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 750 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c if (test_bit(ENTRY_DATA_IO_FAILED, &entry->flags)) entry 757 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c skb = rt2x00queue_alloc_rxskb(entry, gfp); entry 764 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_unmap_skb(entry); entry 770 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00dev->ops->lib->fill_rxdone(entry, &rxdesc); entry 777 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rxdesc.size > entry->queue->data_size)) { entry 779 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rxdesc.size, entry->queue->data_size); entry 780 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c dev_kfree_skb(entry->skb); entry 788 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c header_length = ieee80211_get_hdrlen_from_skb(entry->skb); entry 798 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00crypto_rx_insert_iv(entry->skb, header_length, entry 803 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_remove_l2pad(entry->skb, header_length); entry 806 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c skb_trim(entry->skb, rxdesc.size); entry 820 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00lib_rxdone_check_ps(rt2x00dev, entry->skb, &rxdesc); entry 826 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00lib_rxdone_check_ba(rt2x00dev, entry->skb, &rxdesc); entry 831 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00link_update_stats(rt2x00dev, entry->skb, &rxdesc); entry 833 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_RXDONE, entry); entry 839 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rx_status = IEEE80211_SKB_RXCB(entry->skb); entry 859 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c ieee80211_rx_ni(rt2x00dev->hw, entry->skb); entry 865 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->skb = skb; entry 868 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->flags = 0; entry 869 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00queue_index_inc(entry, Q_INDEX_DONE); entry 872 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00dev->ops->lib->clear_entry(entry); entry 966 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c static void rt2x00lib_channel(struct ieee80211_channel *entry, entry 971 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->band = channel <= 14 ? NL80211_BAND_2GHZ : NL80211_BAND_5GHZ; entry 972 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->center_freq = ieee80211_channel_to_frequency(channel, entry 973 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->band); entry 974 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->hw_value = value; entry 975 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->max_power = tx_power; entry 976 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->max_antenna_gain = 0xff; entry 979 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c static void rt2x00lib_rate(struct ieee80211_rate *entry, entry 982 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->flags = 0; entry 983 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->bitrate = rate->bitrate; entry 984 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->hw_value = index; entry 985 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->hw_value_short = index; entry 988 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c entry->flags |= IEEE80211_RATE_SHORT_PREAMBLE; entry 94 drivers/net/wireless/ralink/rt2x00/rt2x00lib.h struct sk_buff *rt2x00queue_alloc_rxskb(struct queue_entry *entry, gfp_t gfp); entry 100 drivers/net/wireless/ralink/rt2x00/rt2x00lib.h void rt2x00queue_free_skb(struct queue_entry *entry); entry 174 drivers/net/wireless/ralink/rt2x00/rt2x00lib.h void rt2x00queue_index_inc(struct queue_entry *entry, enum queue_index index); entry 189 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c struct queue_entry *entry = NULL; entry 207 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c entry = &queue->entries[i]; entry 208 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c if (!test_and_set_bit(ENTRY_BCN_ASSIGNED, &entry->flags)) entry 226 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c intf->beacon = entry; entry 52 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c struct queue_entry *entry; entry 58 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c entry = rt2x00queue_get_entry(queue, Q_INDEX); entry 59 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c entry_priv = entry->priv_data; entry 61 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c if (rt2x00dev->ops->lib->get_entry_state(entry)) entry 67 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c skbdesc = get_skb_frame_desc(entry->skb); entry 69 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c skbdesc->desc_len = entry->queue->desc_size; entry 75 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c rt2x00lib_dmastart(entry); entry 76 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c rt2x00lib_dmadone(entry); entry 81 drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c rt2x00lib_rxdone(entry, GFP_ATOMIC); entry 23 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct sk_buff *rt2x00queue_alloc_rxskb(struct queue_entry *entry, gfp_t gfp) entry 25 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct data_queue *queue = entry->queue; entry 93 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c int rt2x00queue_map_txskb(struct queue_entry *entry) entry 95 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct device *dev = entry->queue->rt2x00dev->dev; entry 96 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 99 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c dma_map_single(dev, entry->skb->data, entry->skb->len, DMA_TO_DEVICE); entry 105 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00lib_dmadone(entry); entry 110 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c void rt2x00queue_unmap_skb(struct queue_entry *entry) entry 112 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct device *dev = entry->queue->rt2x00dev->dev; entry 113 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 116 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c dma_unmap_single(dev, skbdesc->skb_dma, entry->skb->len, entry 120 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c dma_unmap_single(dev, skbdesc->skb_dma, entry->skb->len, entry 127 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c void rt2x00queue_free_skb(struct queue_entry *entry) entry 129 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c if (!entry->skb) entry 132 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00queue_unmap_skb(entry); entry 133 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c dev_kfree_skb_any(entry->skb); entry 134 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c entry->skb = NULL; entry 487 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c static int rt2x00queue_write_tx_data(struct queue_entry *entry, entry 490 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 498 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00dev->ops->lib->get_entry_state(entry))) { entry 502 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c entry->queue->qid, DRV_PROJECT); entry 509 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c skb_push(entry->skb, rt2x00dev->extra_tx_headroom); entry 510 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c memset(entry->skb->data, 0, rt2x00dev->extra_tx_headroom); entry 516 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00dev->ops->lib->write_tx_data(entry, txdesc); entry 522 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00queue_map_txskb(entry)) entry 528 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c static void rt2x00queue_write_tx_descriptor(struct queue_entry *entry, entry 531 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct data_queue *queue = entry->queue; entry 533 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c queue->rt2x00dev->ops->lib->write_tx_desc(entry, txdesc); entry 539 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00debug_dump_frame(queue->rt2x00dev, DUMP_FRAME_TX, entry); entry 559 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c static void rt2x00queue_bar_check(struct queue_entry *entry) entry 561 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 562 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct ieee80211_bar *bar = (void *) (entry->skb->data + entry 579 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c bar_entry->entry = entry; entry 605 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct queue_entry *entry; entry 672 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c entry = rt2x00queue_get_entry(queue, Q_INDEX); entry 675 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c &entry->flags))) { entry 684 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c entry->skb = skb; entry 691 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c if (unlikely(rt2x00queue_write_tx_data(entry, &txdesc))) { entry 692 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c clear_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags); entry 693 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c entry->skb = NULL; entry 701 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00queue_bar_check(entry); entry 703 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c set_bit(ENTRY_DATA_PENDING, &entry->flags); entry 705 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00queue_index_inc(entry, Q_INDEX); entry 706 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00queue_write_tx_descriptor(entry, &txdesc); entry 790 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c bool (*fn)(struct queue_entry *entry, entry 844 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct queue_entry *entry; entry 855 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c entry = &queue->entries[queue->index[index]]; entry 859 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c return entry; entry 863 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c void rt2x00queue_index_inc(struct queue_entry *entry, enum queue_index index) entry 865 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct data_queue *queue = entry->queue; entry 880 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c entry->last_action = jiffies; entry 584 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h bool (*fn)(struct queue_entry *entry, entry 626 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h static inline int rt2x00queue_dma_timeout(struct queue_entry *entry) entry 628 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h if (!test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags)) entry 630 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h return time_after(jiffies, entry->last_action + msecs_to_jiffies(100)); entry 229 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c static void rt2x00usb_work_txdone_entry(struct queue_entry *entry) entry 239 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (test_bit(ENTRY_DATA_IO_FAILED, &entry->flags)) entry 240 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_txdone_noinfo(entry, TXDONE_FAILURE); entry 242 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_txdone_noinfo(entry, TXDONE_UNKNOWN); entry 250 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry *entry; entry 254 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE); entry 256 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags) || entry 257 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c !test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags)) entry 260 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00usb_work_txdone_entry(entry); entry 267 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry *entry = (struct queue_entry *)urb->context; entry 268 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 270 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (!test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags)) entry 276 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c set_bit(ENTRY_DATA_IO_FAILED, &entry->flags); entry 280 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_dmadone(entry); entry 283 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00dev->ops->lib->tx_dma_done(entry); entry 293 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c static bool rt2x00usb_kick_tx_entry(struct queue_entry *entry, void *data) entry 295 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 297 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry_priv_usb *entry_priv = entry->priv_data; entry 301 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (!test_and_clear_bit(ENTRY_DATA_PENDING, &entry->flags) || entry 302 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c test_bit(ENTRY_DATA_STATUS_PENDING, &entry->flags)) entry 310 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c length = rt2x00dev->ops->lib->get_tx_data_len(entry); entry 312 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c status = skb_padto(entry->skb, length); entry 316 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c set_bit(ENTRY_DATA_IO_FAILED, &entry->flags); entry 317 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_dmadone(entry); entry 323 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint), entry 324 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c entry->skb->data, length, entry 325 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00usb_interrupt_txdone, entry); entry 331 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c set_bit(ENTRY_DATA_IO_FAILED, &entry->flags); entry 332 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_dmadone(entry); entry 345 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry *entry; entry 350 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c entry = rt2x00queue_get_entry(rt2x00dev->rx, Q_INDEX_DONE); entry 352 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags)) entry 358 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c skbdesc = get_skb_frame_desc(entry->skb); entry 360 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c skbdesc->desc_len = entry->queue->desc_size; entry 365 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_rxdone(entry, GFP_KERNEL); entry 371 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry *entry = (struct queue_entry *)urb->context; entry 372 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 374 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (!test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags)) entry 382 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (urb->actual_length < entry->queue->desc_size || urb->status) entry 383 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c set_bit(ENTRY_DATA_IO_FAILED, &entry->flags); entry 388 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_dmadone(entry); entry 396 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c static bool rt2x00usb_kick_rx_entry(struct queue_entry *entry, void *data) entry 398 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 400 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry_priv_usb *entry_priv = entry->priv_data; entry 403 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (test_and_set_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags)) entry 406 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_dmastart(entry); entry 409 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c usb_rcvbulkpipe(usb_dev, entry->queue->usb_endpoint), entry 410 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c entry->skb->data, entry->skb->len, entry 411 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00usb_interrupt_rxdone, entry); entry 417 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c set_bit(ENTRY_DATA_IO_FAILED, &entry->flags); entry 418 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00lib_dmadone(entry); entry 452 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c static bool rt2x00usb_flush_entry(struct queue_entry *entry, void *data) entry 454 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 455 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry_priv_usb *entry_priv = entry->priv_data; entry 456 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry_priv_usb_bcn *bcn_priv = entry->priv_data; entry 458 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (!test_bit(ENTRY_OWNER_DEVICE_DATA, &entry->flags)) entry 466 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if ((entry->queue->qid == QID_BEACON) && entry 535 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct queue_entry *entry; entry 537 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE); entry 538 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c return rt2x00queue_dma_timeout(entry); entry 567 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c void rt2x00usb_clear_entry(struct queue_entry *entry) entry 569 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c entry->flags = 0; entry 571 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (entry->queue->qid == QID_RX) entry 572 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c rt2x00usb_kick_rx_entry(entry, NULL); entry 391 drivers/net/wireless/ralink/rt2x00/rt2x00usb.h void rt2x00usb_clear_entry(struct queue_entry *entry); entry 1287 drivers/net/wireless/ralink/rt2x00/rt61pci.c static bool rt61pci_get_entry_state(struct queue_entry *entry) entry 1289 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 1292 drivers/net/wireless/ralink/rt2x00/rt61pci.c if (entry->queue->qid == QID_RX) { entry 1304 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_clear_entry(struct queue_entry *entry) entry 1306 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 1307 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1310 drivers/net/wireless/ralink/rt2x00/rt61pci.c if (entry->queue->qid == QID_RX) { entry 1775 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_write_tx_desc(struct queue_entry *entry, entry 1778 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1779 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 1787 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid); entry 1788 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_set_field32(&word, TXD_W1_AIFSN, entry->queue->aifs); entry 1789 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min); entry 1790 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max); entry 1812 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_set_field32(&word, TXD_W5_PID_TYPE, entry->queue->qid); entry 1813 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_set_field32(&word, TXD_W5_PID_SUBTYPE, entry->entry_idx); entry 1815 drivers/net/wireless/ralink/rt2x00/rt61pci.c TXPOWER_TO_DEV(entry->queue->rt2x00dev->tx_power)); entry 1819 drivers/net/wireless/ralink/rt2x00/rt61pci.c if (entry->queue->qid != QID_BEACON) { entry 1865 drivers/net/wireless/ralink/rt2x00/rt61pci.c skbdesc->desc_len = (entry->queue->qid == QID_BEACON) ? TXINFO_SIZE : entry 1872 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_write_beacon(struct queue_entry *entry, entry 1875 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1876 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 1893 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt61pci_write_tx_desc(entry, txdesc); entry 1898 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry); entry 1903 drivers/net/wireless/ralink/rt2x00/rt61pci.c padding_len = roundup(entry->skb->len, 4) - entry->skb->len; entry 1904 drivers/net/wireless/ralink/rt2x00/rt61pci.c if (padding_len && skb_pad(entry->skb, padding_len)) { entry 1907 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry->skb = NULL; entry 1912 drivers/net/wireless/ralink/rt2x00/rt61pci.c beacon_base = HW_BEACON_OFFSET(entry->entry_idx); entry 1916 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry->skb->data, entry 1917 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry->skb->len + padding_len); entry 1933 drivers/net/wireless/ralink/rt2x00/rt61pci.c dev_kfree_skb_any(entry->skb); entry 1934 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry->skb = NULL; entry 1937 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_clear_beacon(struct queue_entry *entry) entry 1939 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1955 drivers/net/wireless/ralink/rt2x00/rt61pci.c HW_BEACON_OFFSET(entry->entry_idx), 0); entry 1994 drivers/net/wireless/ralink/rt2x00/rt61pci.c static void rt61pci_fill_rxdone(struct queue_entry *entry, entry 1997 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1998 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct queue_entry_priv_mmio *entry_priv = entry->priv_data; entry 2062 drivers/net/wireless/ralink/rt2x00/rt61pci.c struct queue_entry *entry; entry 2103 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry = &queue->entries[index]; entry 2104 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry_priv = entry->priv_data; entry 2112 drivers/net/wireless/ralink/rt2x00/rt61pci.c while (entry != entry_done) { entry 2146 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00lib_txdone(entry, &txdesc); entry 1444 drivers/net/wireless/ralink/rt2x00/rt73usb.c static void rt73usb_write_tx_desc(struct queue_entry *entry, entry 1447 drivers/net/wireless/ralink/rt2x00/rt73usb.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1448 drivers/net/wireless/ralink/rt2x00/rt73usb.c __le32 *txd = (__le32 *) entry->skb->data; entry 1481 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid); entry 1482 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_set_field32(&word, TXD_W1_AIFSN, entry->queue->aifs); entry 1483 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min); entry 1484 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max); entry 1506 drivers/net/wireless/ralink/rt2x00/rt73usb.c TXPOWER_TO_DEV(entry->queue->rt2x00dev->tx_power)); entry 1521 drivers/net/wireless/ralink/rt2x00/rt73usb.c static void rt73usb_write_beacon(struct queue_entry *entry, entry 1524 drivers/net/wireless/ralink/rt2x00/rt73usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1541 drivers/net/wireless/ralink/rt2x00/rt73usb.c skb_push(entry->skb, TXD_DESC_SIZE); entry 1542 drivers/net/wireless/ralink/rt2x00/rt73usb.c memset(entry->skb->data, 0, TXD_DESC_SIZE); entry 1547 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt73usb_write_tx_desc(entry, txdesc); entry 1552 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00debug_dump_frame(rt2x00dev, DUMP_FRAME_BEACON, entry); entry 1557 drivers/net/wireless/ralink/rt2x00/rt73usb.c padding_len = roundup(entry->skb->len, 4) - entry->skb->len; entry 1558 drivers/net/wireless/ralink/rt2x00/rt73usb.c if (padding_len && skb_pad(entry->skb, padding_len)) { entry 1561 drivers/net/wireless/ralink/rt2x00/rt73usb.c entry->skb = NULL; entry 1566 drivers/net/wireless/ralink/rt2x00/rt73usb.c beacon_base = HW_BEACON_OFFSET(entry->entry_idx); entry 1567 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00usb_register_multiwrite(rt2x00dev, beacon_base, entry->skb->data, entry 1568 drivers/net/wireless/ralink/rt2x00/rt73usb.c entry->skb->len + padding_len); entry 1584 drivers/net/wireless/ralink/rt2x00/rt73usb.c dev_kfree_skb(entry->skb); entry 1585 drivers/net/wireless/ralink/rt2x00/rt73usb.c entry->skb = NULL; entry 1588 drivers/net/wireless/ralink/rt2x00/rt73usb.c static void rt73usb_clear_beacon(struct queue_entry *entry) entry 1590 drivers/net/wireless/ralink/rt2x00/rt73usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1606 drivers/net/wireless/ralink/rt2x00/rt73usb.c beacon_base = HW_BEACON_OFFSET(entry->entry_idx); entry 1615 drivers/net/wireless/ralink/rt2x00/rt73usb.c static int rt73usb_get_tx_data_len(struct queue_entry *entry) entry 1623 drivers/net/wireless/ralink/rt2x00/rt73usb.c length = roundup(entry->skb->len, 4); entry 1624 drivers/net/wireless/ralink/rt2x00/rt73usb.c length += (4 * !(length % entry->queue->usb_maxpacket)); entry 1667 drivers/net/wireless/ralink/rt2x00/rt73usb.c static void rt73usb_fill_rxdone(struct queue_entry *entry, entry 1670 drivers/net/wireless/ralink/rt2x00/rt73usb.c struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; entry 1671 drivers/net/wireless/ralink/rt2x00/rt73usb.c struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb); entry 1672 drivers/net/wireless/ralink/rt2x00/rt73usb.c __le32 *rxd = (__le32 *)entry->skb->data; entry 1742 drivers/net/wireless/ralink/rt2x00/rt73usb.c skb_pull(entry->skb, entry->queue->desc_size); entry 1743 drivers/net/wireless/ralink/rt2x00/rt73usb.c skb_trim(entry->skb, rxdesc->size); entry 221 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c void *entry = priv->rx_ring + priv->rx_idx * priv->rx_ring_sz; entry 227 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8187se_rx_desc *desc = entry; entry 240 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8180_rx_desc *desc = entry; entry 331 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c cmd_desc = entry; entry 348 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8180_tx_desc *entry = &ring->desc[ring->idx]; entry 351 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c u32 flags = le32_to_cpu(entry->flags); entry 358 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c pci_unmap_single(priv->pdev, le32_to_cpu(entry->tx_buf), entry 462 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8180_tx_desc *entry; entry 545 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry = &ring->desc[idx]; entry 548 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->frame_duration = cpu_to_le16(frame_duration); entry 549 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->frame_len_se = cpu_to_le16(skb->len); entry 552 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->flags3 = cpu_to_le16(1<<4); entry 554 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->frame_len = cpu_to_le32(skb->len); entry 556 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->rts_duration = rts_duration; entry 557 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->plcp_len = cpu_to_le16(plcp_len); entry 558 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->tx_buf = cpu_to_le32(mapping); entry 560 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->retry_limit = info->control.rates[0].count - 1; entry 566 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->flags = cpu_to_le32(tx_flags); entry 999 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl818x_rx_cmd_desc *entry; entry 1019 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry = priv->rx_ring + priv->rx_ring_sz*i; entry 1039 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->rx_buf = cpu_to_le32(*mapping); entry 1040 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->flags = cpu_to_le32(RTL818X_RX_DESC_FLAG_OWN | entry 1043 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry->flags |= cpu_to_le32(RTL818X_RX_DESC_FLAG_EOR); entry 1103 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8180_tx_desc *entry = &ring->desc[ring->idx]; entry 1106 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c pci_unmap_single(priv->pdev, le32_to_cpu(entry->tx_buf), entry 417 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c struct urb *entry = NULL; entry 428 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c entry = usb_alloc_urb(0, GFP_KERNEL); entry 429 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c if (!entry) { entry 433 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_fill_bulk_urb(entry, priv->udev, entry 439 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c info->urb = entry; entry 442 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_anchor_urb(entry, &priv->anchored); entry 443 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c ret = usb_submit_urb(entry, GFP_KERNEL); entry 446 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_unanchor_urb(entry); entry 447 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_put_urb(entry); entry 450 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_put_urb(entry); entry 550 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c struct urb *entry; entry 553 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c entry = usb_alloc_urb(0, GFP_KERNEL); entry 554 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c if (!entry) entry 557 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_fill_bulk_urb(entry, priv->udev, usb_rcvbulkpipe(priv->udev, 9), entry 561 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_anchor_urb(entry, &priv->anchored); entry 562 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c ret = usb_submit_urb(entry, GFP_KERNEL); entry 564 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_unanchor_urb(entry); entry 565 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c usb_free_urb(entry); entry 1936 drivers/net/wireless/realtek/rtlwifi/base.c struct rtl_bssid_entry *entry, *next; entry 1938 drivers/net/wireless/realtek/rtlwifi/base.c list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) { entry 1939 drivers/net/wireless/realtek/rtlwifi/base.c list_del(&entry->list); entry 1940 drivers/net/wireless/realtek/rtlwifi/base.c kfree(entry); entry 1967 drivers/net/wireless/realtek/rtlwifi/base.c struct rtl_bssid_entry *entry, *next; entry 1972 drivers/net/wireless/realtek/rtlwifi/base.c list_for_each_entry_safe(entry, next, &rtlpriv->scan_list.list, list) { entry 1974 drivers/net/wireless/realtek/rtlwifi/base.c if (jiffies_to_msecs(jiffies - entry->age) < 180000) entry 1977 drivers/net/wireless/realtek/rtlwifi/base.c list_del(&entry->list); entry 1982 drivers/net/wireless/realtek/rtlwifi/base.c entry->bssid, rtlpriv->scan_list.num); entry 1983 drivers/net/wireless/realtek/rtlwifi/base.c kfree(entry); entry 1998 drivers/net/wireless/realtek/rtlwifi/base.c struct rtl_bssid_entry *entry; entry 2012 drivers/net/wireless/realtek/rtlwifi/base.c list_for_each_entry(entry, &rtlpriv->scan_list.list, list) { entry 2013 drivers/net/wireless/realtek/rtlwifi/base.c if (memcmp(entry->bssid, hdr->addr3, ETH_ALEN) == 0) { entry 2014 drivers/net/wireless/realtek/rtlwifi/base.c list_del_init(&entry->list); entry 2024 drivers/net/wireless/realtek/rtlwifi/base.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 2026 drivers/net/wireless/realtek/rtlwifi/base.c if (!entry) entry 2029 drivers/net/wireless/realtek/rtlwifi/base.c memcpy(entry->bssid, hdr->addr3, ETH_ALEN); entry 2037 drivers/net/wireless/realtek/rtlwifi/base.c entry->age = jiffies; entry 2039 drivers/net/wireless/realtek/rtlwifi/base.c list_add_tail(&entry->list, &rtlpriv->scan_list.list); entry 538 drivers/net/wireless/realtek/rtlwifi/pci.c u8 *entry; entry 541 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->buffer_desc[ring->idx]); entry 543 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->desc[ring->idx]); entry 552 drivers/net/wireless/realtek/rtlwifi/pci.c get_desc(hw, (u8 *)entry, true, entry 629 drivers/net/wireless/realtek/rtlwifi/pci.c struct sk_buff *new_skb, u8 *entry, entry 657 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false, entry 661 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false, entry 664 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false, entry 667 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false, entry 1085 drivers/net/wireless/realtek/rtlwifi/pci.c u8 *entry; entry 1091 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->buffer_desc[ring->idx]); entry 1093 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->desc[ring->idx]); entry 1097 drivers/net/wireless/realtek/rtlwifi/pci.c hw, (u8 *)entry, true, HW_DESC_TXBUFF_ADDR), entry 1280 drivers/net/wireless/realtek/rtlwifi/pci.c struct rtl_rx_buffer_desc *entry = NULL; entry 1297 drivers/net/wireless/realtek/rtlwifi/pci.c entry = &rtlpci->rx_ring[rxring_idx].buffer_desc[i]; entry 1298 drivers/net/wireless/realtek/rtlwifi/pci.c if (!_rtl_pci_init_one_rxdesc(hw, NULL, (u8 *)entry, entry 1303 drivers/net/wireless/realtek/rtlwifi/pci.c struct rtl_rx_desc *entry = NULL; entry 1321 drivers/net/wireless/realtek/rtlwifi/pci.c entry = &rtlpci->rx_ring[rxring_idx].desc[i]; entry 1322 drivers/net/wireless/realtek/rtlwifi/pci.c if (!_rtl_pci_init_one_rxdesc(hw, NULL, (u8 *)entry, entry 1327 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false, entry 1342 drivers/net/wireless/realtek/rtlwifi/pci.c u8 *entry; entry 1346 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->buffer_desc[ring->idx]); entry 1348 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->desc[ring->idx]); entry 1351 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpriv->cfg->ops->get_desc(hw, (u8 *)entry, entry 1474 drivers/net/wireless/realtek/rtlwifi/pci.c struct rtl_rx_desc *entry = NULL; entry 1478 drivers/net/wireless/realtek/rtlwifi/pci.c entry = &rtlpci->rx_ring[rxring_idx].desc[i]; entry 1480 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpriv->cfg->ops->get_desc(hw, (u8 *)entry, entry 1482 drivers/net/wireless/realtek/rtlwifi/pci.c memset((u8 *)entry, 0, entry 1487 drivers/net/wireless/realtek/rtlwifi/pci.c (u8 *)entry, false, entry 1492 drivers/net/wireless/realtek/rtlwifi/pci.c (u8 *)entry, false, entry 1496 drivers/net/wireless/realtek/rtlwifi/pci.c (u8 *)entry, false, entry 1500 drivers/net/wireless/realtek/rtlwifi/pci.c (u8 *)entry, false, entry 1505 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpriv->cfg->ops->set_desc(hw, (u8 *)entry, false, entry 1521 drivers/net/wireless/realtek/rtlwifi/pci.c u8 *entry; entry 1525 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->buffer_desc entry 1528 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->desc[ring->idx]); entry 1533 drivers/net/wireless/realtek/rtlwifi/pci.c entry, entry 75 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c struct rtl_tx_desc *entry = &ring->desc[ring->idx]; entry 81 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c (u8 *)entry, true, HW_DESC_TXBUFF_ADDR), entry 814 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); entry 815 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/trx.c u8 own = (u8)rtl88ee_get_desc(hw, entry, true, HW_DESC_OWN); entry 712 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); entry 713 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/trx.c u8 own = (u8)rtl92ce_get_desc(hw, entry, true, HW_DESC_OWN); entry 831 drivers/net/wireless/realtek/rtlwifi/rtl8192de/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); entry 832 drivers/net/wireless/realtek/rtlwifi/rtl8192de/trx.c u8 own = (u8)rtl92de_get_desc(hw, entry, true, HW_DESC_OWN); entry 219 drivers/net/wireless/realtek/rtlwifi/rtl8192se/sw.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); entry 220 drivers/net/wireless/realtek/rtlwifi/rtl8192se/sw.c u8 own = (u8)rtl92se_get_desc(hw, entry, true, HW_DESC_OWN); entry 673 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); entry 674 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/trx.c u8 own = (u8)rtl8723e_get_desc(hw, entry, true, HW_DESC_OWN); entry 37 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c struct rtl_tx_desc *entry = &ring->desc[ring->idx]; entry 43 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c (u8 *)entry, true, HW_DESC_TXBUFF_ADDR), entry 726 drivers/net/wireless/realtek/rtlwifi/rtl8723be/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); entry 727 drivers/net/wireless/realtek/rtlwifi/rtl8723be/trx.c u8 own = (u8)rtl8723be_get_desc(hw, entry, true, HW_DESC_OWN); entry 33 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c struct rtl_tx_desc *entry = &ring->desc[ring->idx]; entry 39 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c (u8 *)entry, true, HW_DESC_TXBUFF_ADDR), entry 966 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); entry 967 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/trx.c u8 own = (u8)rtl8821ae_get_desc(hw, entry, true, HW_DESC_OWN); entry 448 drivers/net/wireless/st/cw1200/sta.c .entry[0].ie_id = WLAN_EID_VENDOR_SPECIFIC, entry 449 drivers/net/wireless/st/cw1200/sta.c .entry[0].flags = WSM_BEACON_FILTER_IE_HAS_CHANGED | entry 452 drivers/net/wireless/st/cw1200/sta.c .entry[0].oui[0] = 0x50, entry 453 drivers/net/wireless/st/cw1200/sta.c .entry[0].oui[1] = 0x6F, entry 454 drivers/net/wireless/st/cw1200/sta.c .entry[0].oui[2] = 0x9A, entry 455 drivers/net/wireless/st/cw1200/sta.c .entry[1].ie_id = WLAN_EID_HT_OPERATION, entry 456 drivers/net/wireless/st/cw1200/sta.c .entry[1].flags = WSM_BEACON_FILTER_IE_HAS_CHANGED | entry 459 drivers/net/wireless/st/cw1200/sta.c .entry[2].ie_id = WLAN_EID_ERP_INFO, entry 460 drivers/net/wireless/st/cw1200/sta.c .entry[2].flags = WSM_BEACON_FILTER_IE_HAS_CHANGED | entry 1545 drivers/net/wireless/st/cw1200/sta.c struct cw1200_link_entry *entry; entry 1558 drivers/net/wireless/st/cw1200/sta.c entry = &priv->link_id_db[sta_priv->link_id - 1]; entry 1563 drivers/net/wireless/st/cw1200/sta.c entry->status = CW1200_LINK_HARD; entry 1564 drivers/net/wireless/st/cw1200/sta.c while ((skb = skb_dequeue(&entry->rx_queue))) entry 1576 drivers/net/wireless/st/cw1200/sta.c struct cw1200_link_entry *entry; entry 1581 drivers/net/wireless/st/cw1200/sta.c entry = &priv->link_id_db[sta_priv->link_id - 1]; entry 1583 drivers/net/wireless/st/cw1200/sta.c entry->status = CW1200_LINK_RESERVE; entry 1584 drivers/net/wireless/st/cw1200/sta.c entry->timestamp = jiffies; entry 235 drivers/net/wireless/st/cw1200/txrx.c struct tx_policy_cache_entry *entry) entry 237 drivers/net/wireless/st/cw1200/txrx.c ++entry->policy.usage_count; entry 238 drivers/net/wireless/st/cw1200/txrx.c list_move(&entry->link, &cache->used); entry 242 drivers/net/wireless/st/cw1200/txrx.c struct tx_policy_cache_entry *entry) entry 244 drivers/net/wireless/st/cw1200/txrx.c int ret = --entry->policy.usage_count; entry 246 drivers/net/wireless/st/cw1200/txrx.c list_move(&entry->link, &cache->free); entry 254 drivers/net/wireless/st/cw1200/txrx.c struct tx_policy_cache_entry *entry; entry 261 drivers/net/wireless/st/cw1200/txrx.c entry = &cache->cache[idx]; entry 265 drivers/net/wireless/st/cw1200/txrx.c if (WARN_ON(entry->policy.usage_count)) { entry 266 drivers/net/wireless/st/cw1200/txrx.c entry->policy.usage_count = 0; entry 267 drivers/net/wireless/st/cw1200/txrx.c list_move(&entry->link, &cache->free); entry 269 drivers/net/wireless/st/cw1200/txrx.c memset(&entry->policy, 0, sizeof(entry->policy)); entry 316 drivers/net/wireless/st/cw1200/txrx.c struct tx_policy_cache_entry *entry; entry 321 drivers/net/wireless/st/cw1200/txrx.c entry = list_entry(cache->free.prev, entry 323 drivers/net/wireless/st/cw1200/txrx.c entry->policy = wanted; entry 324 drivers/net/wireless/st/cw1200/txrx.c idx = entry - cache->cache; entry 326 drivers/net/wireless/st/cw1200/txrx.c tx_policy_dump(&entry->policy); entry 1012 drivers/net/wireless/st/cw1200/txrx.c struct cw1200_link_entry *entry = NULL; entry 1026 drivers/net/wireless/st/cw1200/txrx.c entry = &priv->link_id_db[link_id - 1]; entry 1027 drivers/net/wireless/st/cw1200/txrx.c if (entry->status == CW1200_LINK_SOFT && entry 1030 drivers/net/wireless/st/cw1200/txrx.c entry->timestamp = jiffies; entry 1209 drivers/net/wireless/st/cw1200/txrx.c if (entry->status == CW1200_LINK_SOFT) entry 1210 drivers/net/wireless/st/cw1200/txrx.c skb_queue_tail(&entry->rx_queue, skb); entry 1346 drivers/net/wireless/st/cw1200/txrx.c struct cw1200_link_entry *entry = &priv->link_id_db[ret - 1]; entry 1348 drivers/net/wireless/st/cw1200/txrx.c entry->status = CW1200_LINK_RESERVE; entry 1349 drivers/net/wireless/st/cw1200/txrx.c memcpy(&entry->mac, mac, ETH_ALEN); entry 1350 drivers/net/wireless/st/cw1200/txrx.c memset(&entry->buffered, 0, CW1200_MAX_TID); entry 1351 drivers/net/wireless/st/cw1200/txrx.c skb_queue_head_init(&entry->rx_queue); entry 1401 drivers/net/wireless/st/cw1200/wsm.h struct wsm_beacon_filter_table_entry entry[10]; entry 214 drivers/net/xen-netback/common.h struct list_head entry; entry 38 drivers/net/xen-netback/hash.c struct xenvif_hash_cache_entry *new, *entry, *oldest; entry 42 drivers/net/xen-netback/hash.c new = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 54 drivers/net/xen-netback/hash.c list_for_each_entry_rcu(entry, &vif->hash.cache.list, link) { entry 56 drivers/net/xen-netback/hash.c if (entry->len == len && entry 57 drivers/net/xen-netback/hash.c memcmp(entry->tag, tag, len) == 0) entry 59 drivers/net/xen-netback/hash.c if (!oldest || entry->seq < oldest->seq) entry 60 drivers/net/xen-netback/hash.c oldest = entry; entry 97 drivers/net/xen-netback/hash.c struct xenvif_hash_cache_entry *entry; entry 105 drivers/net/xen-netback/hash.c list_for_each_entry_rcu(entry, &vif->hash.cache.list, link) { entry 106 drivers/net/xen-netback/hash.c list_del_rcu(&entry->link); entry 108 drivers/net/xen-netback/hash.c kfree_rcu(entry, rcu); entry 117 drivers/net/xen-netback/hash.c struct xenvif_hash_cache_entry *entry; entry 131 drivers/net/xen-netback/hash.c list_for_each_entry_rcu(entry, &vif->hash.cache.list, link) { entry 132 drivers/net/xen-netback/hash.c if (entry->len == len && entry 133 drivers/net/xen-netback/hash.c memcmp(entry->tag, data, len) == 0) { entry 134 drivers/net/xen-netback/hash.c val = entry->val; entry 135 drivers/net/xen-netback/hash.c entry->seq = atomic_inc_return(&vif->hash.cache.seq); entry 735 drivers/net/xen-netback/netback.c list_add_tail_rcu(&mcast->entry, &vif->fe_mcast_addr); entry 745 drivers/net/xen-netback/netback.c list_for_each_entry_rcu(mcast, &vif->fe_mcast_addr, entry) { entry 748 drivers/net/xen-netback/netback.c list_del_rcu(&mcast->entry); entry 760 drivers/net/xen-netback/netback.c list_for_each_entry_rcu(mcast, &vif->fe_mcast_addr, entry) { entry 781 drivers/net/xen-netback/netback.c entry); entry 783 drivers/net/xen-netback/netback.c list_del(&mcast->entry); entry 606 drivers/ntb/hw/amd/ntb_hw_amd.c ndev->msix[i].entry = i; entry 392 drivers/ntb/hw/intel/ntb_hw_gen1.c ndev->msix[i].entry = i; entry 191 drivers/ntb/msi.c struct msi_desc *entry; entry 195 drivers/ntb/msi.c static int ntb_msi_set_desc(struct ntb_dev *ntb, struct msi_desc *entry, entry 200 drivers/ntb/msi.c addr = entry->msg.address_lo + entry 201 drivers/ntb/msi.c ((uint64_t)entry->msg.address_hi << 32); entry 206 drivers/ntb/msi.c entry->irq, addr, ntb->msi->base_addr, entry 212 drivers/ntb/msi.c msi_desc->data = entry->msg.data; entry 217 drivers/ntb/msi.c static void ntb_msi_write_msg(struct msi_desc *entry, void *data) entry 221 drivers/ntb/msi.c WARN_ON(ntb_msi_set_desc(dr->ntb, entry, dr->msi_desc)); entry 231 drivers/ntb/msi.c dr->entry->write_msi_msg = NULL; entry 232 drivers/ntb/msi.c dr->entry->write_msi_msg_data = NULL; entry 235 drivers/ntb/msi.c static int ntbm_msi_setup_callback(struct ntb_dev *ntb, struct msi_desc *entry, entry 246 drivers/ntb/msi.c dr->entry = entry; entry 251 drivers/ntb/msi.c dr->entry->write_msi_msg = ntb_msi_write_msg; entry 252 drivers/ntb/msi.c dr->entry->write_msi_msg_data = dr; entry 284 drivers/ntb/msi.c struct msi_desc *entry; entry 291 drivers/ntb/msi.c for_each_pci_msi_entry(entry, ntb->pdev) { entry 292 drivers/ntb/msi.c desc = irq_to_desc(entry->irq); entry 296 drivers/ntb/msi.c ret = devm_request_threaded_irq(&ntb->dev, entry->irq, handler, entry 301 drivers/ntb/msi.c if (ntb_msi_set_desc(ntb, entry, msi_desc)) { entry 302 drivers/ntb/msi.c devm_free_irq(&ntb->dev, entry->irq, dev_id); entry 306 drivers/ntb/msi.c ret = ntbm_msi_setup_callback(ntb, entry, msi_desc); entry 308 drivers/ntb/msi.c devm_free_irq(&ntb->dev, entry->irq, dev_id); entry 313 drivers/ntb/msi.c return entry->irq; entry 325 drivers/ntb/msi.c return dr->ntb == ntb && dr->entry == data; entry 339 drivers/ntb/msi.c struct msi_desc *entry = irq_get_msi_desc(irq); entry 341 drivers/ntb/msi.c entry->write_msi_msg = NULL; entry 342 drivers/ntb/msi.c entry->write_msi_msg_data = NULL; entry 345 drivers/ntb/msi.c ntbm_msi_callback_match, entry)); entry 109 drivers/ntb/ntb_transport.c struct list_head entry; entry 128 drivers/ntb/ntb_transport.c unsigned int entry; entry 217 drivers/ntb/ntb_transport.c struct list_head entry; entry 223 drivers/ntb/ntb_transport.c struct list_head entry; entry 280 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry); entry 281 drivers/ntb/ntb_transport.c static void ntb_memcpy_tx(struct ntb_queue_entry *entry, void __iomem *offset); entry 282 drivers/ntb/ntb_transport.c static int ntb_async_rx_submit(struct ntb_queue_entry *entry, void *offset); entry 283 drivers/ntb/ntb_transport.c static void ntb_memcpy_rx(struct ntb_queue_entry *entry, void *offset); entry 330 drivers/ntb/ntb_transport.c list_add_tail(&nt->entry, &ntb_transport_list); entry 338 drivers/ntb/ntb_transport.c list_for_each_entry_safe(client_dev, cd, &nt->client_devs, entry) { entry 341 drivers/ntb/ntb_transport.c list_del(&client_dev->entry); entry 345 drivers/ntb/ntb_transport.c list_del(&nt->entry); entry 367 drivers/ntb/ntb_transport.c list_for_each_entry(nt, &ntb_transport_list, entry) entry 368 drivers/ntb/ntb_transport.c list_for_each_entry_safe(client, cd, &nt->client_devs, entry) entry 371 drivers/ntb/ntb_transport.c list_del(&client->entry); entry 393 drivers/ntb/ntb_transport.c list_for_each_entry(nt, &ntb_transport_list, entry) { entry 419 drivers/ntb/ntb_transport.c list_add_tail(&client_dev->entry, &nt->client_devs); entry 529 drivers/ntb/ntb_transport.c qp->remote_rx_info->entry); entry 564 drivers/ntb/ntb_transport.c static void ntb_list_add(spinlock_t *lock, struct list_head *entry, entry 570 drivers/ntb/ntb_transport.c list_add_tail(entry, list); entry 577 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 582 drivers/ntb/ntb_transport.c entry = NULL; entry 585 drivers/ntb/ntb_transport.c entry = list_first_entry(list, struct ntb_queue_entry, entry); entry 586 drivers/ntb/ntb_transport.c list_del(&entry->entry); entry 591 drivers/ntb/ntb_transport.c return entry; entry 598 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 604 drivers/ntb/ntb_transport.c entry = NULL; entry 606 drivers/ntb/ntb_transport.c entry = list_first_entry(list, struct ntb_queue_entry, entry); entry 607 drivers/ntb/ntb_transport.c list_move_tail(&entry->entry, to_list); entry 612 drivers/ntb/ntb_transport.c return entry; entry 621 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 659 drivers/ntb/ntb_transport.c entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node); entry 660 drivers/ntb/ntb_transport.c if (!entry) entry 663 drivers/ntb/ntb_transport.c entry->qp = qp; entry 664 drivers/ntb/ntb_transport.c ntb_list_add(&qp->ntb_rx_q_lock, &entry->entry, entry 669 drivers/ntb/ntb_transport.c qp->remote_rx_info->entry = qp->rx_max_entry - 1; entry 1441 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 1449 drivers/ntb/ntb_transport.c entry = list_first_entry(&qp->rx_post_q, entry 1450 drivers/ntb/ntb_transport.c struct ntb_queue_entry, entry); entry 1451 drivers/ntb/ntb_transport.c if (!(entry->flags & DESC_DONE_FLAG)) entry 1454 drivers/ntb/ntb_transport.c entry->rx_hdr->flags = 0; entry 1455 drivers/ntb/ntb_transport.c iowrite32(entry->rx_index, &qp->rx_info->entry); entry 1457 drivers/ntb/ntb_transport.c cb_data = entry->cb_data; entry 1458 drivers/ntb/ntb_transport.c len = entry->len; entry 1460 drivers/ntb/ntb_transport.c list_move_tail(&entry->entry, &qp->rx_free_q); entry 1476 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry = data; entry 1485 drivers/ntb/ntb_transport.c entry->errors++; entry 1489 drivers/ntb/ntb_transport.c struct ntb_transport_qp *qp = entry->qp; entry 1493 drivers/ntb/ntb_transport.c ntb_memcpy_rx(entry, offset); entry 1504 drivers/ntb/ntb_transport.c entry->flags |= DESC_DONE_FLAG; entry 1506 drivers/ntb/ntb_transport.c ntb_complete_rxc(entry->qp); entry 1509 drivers/ntb/ntb_transport.c static void ntb_memcpy_rx(struct ntb_queue_entry *entry, void *offset) entry 1511 drivers/ntb/ntb_transport.c void *buf = entry->buf; entry 1512 drivers/ntb/ntb_transport.c size_t len = entry->len; entry 1519 drivers/ntb/ntb_transport.c ntb_rx_copy_callback(entry, NULL); entry 1522 drivers/ntb/ntb_transport.c static int ntb_async_rx_submit(struct ntb_queue_entry *entry, void *offset) entry 1525 drivers/ntb/ntb_transport.c struct ntb_transport_qp *qp = entry->qp; entry 1531 drivers/ntb/ntb_transport.c void *buf = entry->buf; entry 1533 drivers/ntb/ntb_transport.c len = entry->len; entry 1567 drivers/ntb/ntb_transport.c txd->callback_param = entry; entry 1590 drivers/ntb/ntb_transport.c static void ntb_async_rx(struct ntb_queue_entry *entry, void *offset) entry 1592 drivers/ntb/ntb_transport.c struct ntb_transport_qp *qp = entry->qp; entry 1599 drivers/ntb/ntb_transport.c if (entry->len < copy_bytes) entry 1602 drivers/ntb/ntb_transport.c res = ntb_async_rx_submit(entry, offset); entry 1606 drivers/ntb/ntb_transport.c if (!entry->retries) entry 1612 drivers/ntb/ntb_transport.c ntb_memcpy_rx(entry, offset); entry 1619 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 1649 drivers/ntb/ntb_transport.c entry = ntb_list_mv(&qp->ntb_rx_q_lock, &qp->rx_pend_q, &qp->rx_post_q); entry 1650 drivers/ntb/ntb_transport.c if (!entry) { entry 1656 drivers/ntb/ntb_transport.c entry->rx_hdr = hdr; entry 1657 drivers/ntb/ntb_transport.c entry->rx_index = qp->rx_index; entry 1659 drivers/ntb/ntb_transport.c if (hdr->len > entry->len) { entry 1662 drivers/ntb/ntb_transport.c hdr->len, entry->len); entry 1665 drivers/ntb/ntb_transport.c entry->len = -EIO; entry 1666 drivers/ntb/ntb_transport.c entry->flags |= DESC_DONE_FLAG; entry 1672 drivers/ntb/ntb_transport.c qp->rx_index, hdr->ver, hdr->len, entry->len); entry 1677 drivers/ntb/ntb_transport.c entry->len = hdr->len; entry 1679 drivers/ntb/ntb_transport.c ntb_async_rx(entry, offset); entry 1730 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry = data; entry 1731 drivers/ntb/ntb_transport.c struct ntb_transport_qp *qp = entry->qp; entry 1732 drivers/ntb/ntb_transport.c struct ntb_payload_header __iomem *hdr = entry->tx_hdr; entry 1741 drivers/ntb/ntb_transport.c entry->errors++; entry 1747 drivers/ntb/ntb_transport.c entry->tx_index; entry 1750 drivers/ntb/ntb_transport.c ntb_memcpy_tx(entry, offset); entry 1761 drivers/ntb/ntb_transport.c iowrite32(entry->flags | DESC_DONE_FLAG, &hdr->flags); entry 1772 drivers/ntb/ntb_transport.c if (entry->len > 0) { entry 1773 drivers/ntb/ntb_transport.c qp->tx_bytes += entry->len; entry 1776 drivers/ntb/ntb_transport.c qp->tx_handler(qp, qp->cb_data, entry->cb_data, entry 1777 drivers/ntb/ntb_transport.c entry->len); entry 1780 drivers/ntb/ntb_transport.c ntb_list_add(&qp->ntb_tx_free_q_lock, &entry->entry, &qp->tx_free_q); entry 1783 drivers/ntb/ntb_transport.c static void ntb_memcpy_tx(struct ntb_queue_entry *entry, void __iomem *offset) entry 1790 drivers/ntb/ntb_transport.c __copy_from_user_inatomic_nocache(offset, entry->buf, entry->len); entry 1792 drivers/ntb/ntb_transport.c memcpy_toio(offset, entry->buf, entry->len); entry 1798 drivers/ntb/ntb_transport.c ntb_tx_copy_callback(entry, NULL); entry 1802 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry) entry 1807 drivers/ntb/ntb_transport.c size_t len = entry->len; entry 1808 drivers/ntb/ntb_transport.c void *buf = entry->buf; entry 1815 drivers/ntb/ntb_transport.c dest = qp->tx_mw_dma_addr + qp->tx_max_frame * entry->tx_index; entry 1840 drivers/ntb/ntb_transport.c txd->callback_param = entry; entry 1861 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry) entry 1868 drivers/ntb/ntb_transport.c entry->tx_index = qp->tx_index; entry 1869 drivers/ntb/ntb_transport.c offset = qp->tx_mw + qp->tx_max_frame * entry->tx_index; entry 1871 drivers/ntb/ntb_transport.c entry->tx_hdr = hdr; entry 1873 drivers/ntb/ntb_transport.c iowrite32(entry->len, &hdr->len); entry 1879 drivers/ntb/ntb_transport.c if (entry->len < copy_bytes) entry 1882 drivers/ntb/ntb_transport.c res = ntb_async_tx_submit(qp, entry); entry 1886 drivers/ntb/ntb_transport.c if (!entry->retries) entry 1892 drivers/ntb/ntb_transport.c ntb_memcpy_tx(entry, offset); entry 1897 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry) entry 1899 drivers/ntb/ntb_transport.c if (qp->tx_index == qp->remote_rx_info->entry) { entry 1904 drivers/ntb/ntb_transport.c if (entry->len > qp->tx_max_frame - sizeof(struct ntb_payload_header)) { entry 1908 drivers/ntb/ntb_transport.c ntb_list_add(&qp->ntb_tx_free_q_lock, &entry->entry, entry 1913 drivers/ntb/ntb_transport.c ntb_async_tx(qp, entry); entry 1926 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 1935 drivers/ntb/ntb_transport.c entry = ntb_list_rm(&qp->ntb_tx_free_q_lock, &qp->tx_free_q); entry 1936 drivers/ntb/ntb_transport.c if (entry) entry 1941 drivers/ntb/ntb_transport.c if (!entry) entry 1944 drivers/ntb/ntb_transport.c entry->cb_data = NULL; entry 1945 drivers/ntb/ntb_transport.c entry->buf = NULL; entry 1946 drivers/ntb/ntb_transport.c entry->len = 0; entry 1947 drivers/ntb/ntb_transport.c entry->flags = LINK_DOWN_FLAG; entry 1949 drivers/ntb/ntb_transport.c rc = ntb_process_tx(qp, entry); entry 1983 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 2054 drivers/ntb/ntb_transport.c entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node); entry 2055 drivers/ntb/ntb_transport.c if (!entry) entry 2058 drivers/ntb/ntb_transport.c entry->qp = qp; entry 2059 drivers/ntb/ntb_transport.c ntb_list_add(&qp->ntb_rx_q_lock, &entry->entry, entry 2065 drivers/ntb/ntb_transport.c entry = kzalloc_node(sizeof(*entry), GFP_KERNEL, node); entry 2066 drivers/ntb/ntb_transport.c if (!entry) entry 2069 drivers/ntb/ntb_transport.c entry->qp = qp; entry 2070 drivers/ntb/ntb_transport.c ntb_list_add(&qp->ntb_tx_free_q_lock, &entry->entry, entry 2082 drivers/ntb/ntb_transport.c while ((entry = ntb_list_rm(&qp->ntb_tx_free_q_lock, &qp->tx_free_q))) entry 2083 drivers/ntb/ntb_transport.c kfree(entry); entry 2086 drivers/ntb/ntb_transport.c while ((entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_free_q))) entry 2087 drivers/ntb/ntb_transport.c kfree(entry); entry 2111 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 2168 drivers/ntb/ntb_transport.c while ((entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_free_q))) entry 2169 drivers/ntb/ntb_transport.c kfree(entry); entry 2171 drivers/ntb/ntb_transport.c while ((entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_pend_q))) { entry 2173 drivers/ntb/ntb_transport.c kfree(entry); entry 2176 drivers/ntb/ntb_transport.c while ((entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_post_q))) { entry 2178 drivers/ntb/ntb_transport.c kfree(entry); entry 2181 drivers/ntb/ntb_transport.c while ((entry = ntb_list_rm(&qp->ntb_tx_free_q_lock, &qp->tx_free_q))) entry 2182 drivers/ntb/ntb_transport.c kfree(entry); entry 2202 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 2208 drivers/ntb/ntb_transport.c entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_pend_q); entry 2209 drivers/ntb/ntb_transport.c if (!entry) entry 2212 drivers/ntb/ntb_transport.c buf = entry->cb_data; entry 2213 drivers/ntb/ntb_transport.c *len = entry->len; entry 2215 drivers/ntb/ntb_transport.c ntb_list_add(&qp->ntb_rx_q_lock, &entry->entry, &qp->rx_free_q); entry 2236 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 2241 drivers/ntb/ntb_transport.c entry = ntb_list_rm(&qp->ntb_rx_q_lock, &qp->rx_free_q); entry 2242 drivers/ntb/ntb_transport.c if (!entry) entry 2245 drivers/ntb/ntb_transport.c entry->cb_data = cb; entry 2246 drivers/ntb/ntb_transport.c entry->buf = data; entry 2247 drivers/ntb/ntb_transport.c entry->len = len; entry 2248 drivers/ntb/ntb_transport.c entry->flags = 0; entry 2249 drivers/ntb/ntb_transport.c entry->retries = 0; entry 2250 drivers/ntb/ntb_transport.c entry->errors = 0; entry 2251 drivers/ntb/ntb_transport.c entry->rx_index = 0; entry 2253 drivers/ntb/ntb_transport.c ntb_list_add(&qp->ntb_rx_q_lock, &entry->entry, &qp->rx_pend_q); entry 2278 drivers/ntb/ntb_transport.c struct ntb_queue_entry *entry; entry 2284 drivers/ntb/ntb_transport.c entry = ntb_list_rm(&qp->ntb_tx_free_q_lock, &qp->tx_free_q); entry 2285 drivers/ntb/ntb_transport.c if (!entry) { entry 2290 drivers/ntb/ntb_transport.c entry->cb_data = cb; entry 2291 drivers/ntb/ntb_transport.c entry->buf = data; entry 2292 drivers/ntb/ntb_transport.c entry->len = len; entry 2293 drivers/ntb/ntb_transport.c entry->flags = 0; entry 2294 drivers/ntb/ntb_transport.c entry->errors = 0; entry 2295 drivers/ntb/ntb_transport.c entry->retries = 0; entry 2296 drivers/ntb/ntb_transport.c entry->tx_index = 0; entry 2298 drivers/ntb/ntb_transport.c rc = ntb_process_tx(qp, entry); entry 2300 drivers/ntb/ntb_transport.c ntb_list_add(&qp->ntb_tx_free_q_lock, &entry->entry, entry 2421 drivers/ntb/ntb_transport.c unsigned int tail = qp->remote_rx_info->entry; entry 441 drivers/nvme/host/core.c list_del_init(&head->entry); entry 2502 drivers/nvme/host/core.c list_del(&subsys->entry); entry 2532 drivers/nvme/host/core.c list_for_each_entry(subsys, &nvme_subsystems, entry) { entry 2674 drivers/nvme/host/core.c list_add_tail(&subsys->entry, &nvme_subsystems); entry 3301 drivers/nvme/host/core.c list_for_each_entry(h, &subsys->nsheads, entry) { entry 3316 drivers/nvme/host/core.c list_for_each_entry(h, &subsys->nsheads, entry) { entry 3367 drivers/nvme/host/core.c list_add_tail(&head->entry, &ctrl->subsys->nsheads); entry 499 drivers/nvme/host/fabrics.c list_add_tail(&ops->entry, &nvmf_transports); entry 518 drivers/nvme/host/fabrics.c list_del(&ops->entry); entry 530 drivers/nvme/host/fabrics.c list_for_each_entry(ops, &nvmf_transports, entry) { entry 142 drivers/nvme/host/fabrics.h struct list_head entry; entry 20 drivers/nvme/host/multipath.c list_for_each_entry(h, &subsys->nsheads, entry) entry 30 drivers/nvme/host/multipath.c list_for_each_entry(h, &subsys->nsheads, entry) entry 40 drivers/nvme/host/multipath.c list_for_each_entry(h, &subsys->nsheads, entry) entry 300 drivers/nvme/host/nvme.h struct list_head entry; entry 339 drivers/nvme/host/nvme.h struct list_head entry; entry 41 drivers/nvme/host/rdma.c struct list_head entry; entry 336 drivers/nvme/host/rdma.c list_del(&ndev->entry); entry 359 drivers/nvme/host/rdma.c list_for_each_entry(ndev, &device_list, entry) { entry 386 drivers/nvme/host/rdma.c list_add(&ndev->entry, &device_list); entry 2089 drivers/nvme/host/rdma.c list_for_each_entry(ndev, &device_list, entry) { entry 38 drivers/nvme/host/tcp.c struct list_head entry; entry 255 drivers/nvme/host/tcp.c list_add_tail(&req->entry, &queue->send_list); entry 268 drivers/nvme/host/tcp.c struct nvme_tcp_request, entry); entry 270 drivers/nvme/host/tcp.c list_del(&req->entry); entry 638 drivers/nvme/target/configfs.c list_for_each_entry(p, &port->subsystems, entry) { entry 649 drivers/nvme/target/configfs.c list_add_tail(&link->entry, &port->subsystems); entry 669 drivers/nvme/target/configfs.c list_for_each_entry(p, &port->subsystems, entry) { entry 677 drivers/nvme/target/configfs.c list_del(&p->entry); entry 724 drivers/nvme/target/configfs.c list_for_each_entry(p, &subsys->hosts, entry) { entry 728 drivers/nvme/target/configfs.c list_add_tail(&link->entry, &subsys->hosts); entry 747 drivers/nvme/target/configfs.c list_for_each_entry(p, &subsys->hosts, entry) { entry 755 drivers/nvme/target/configfs.c list_del(&p->entry); entry 1001 drivers/nvme/target/configfs.c INIT_LIST_HEAD(&port->entry); entry 1207 drivers/nvme/target/configfs.c INIT_LIST_HEAD(&port->entry); entry 159 drivers/nvme/target/core.c struct nvmet_async_event, entry); entry 168 drivers/nvme/target/core.c list_del(&aen->entry); entry 190 drivers/nvme/target/core.c list_add_tail(&aen->entry, &ctrl->async_events); entry 258 drivers/nvme/target/core.c list_for_each_entry(p, &port->subsystems, entry) entry 1157 drivers/nvme/target/core.c list_for_each_entry(p, &subsys->hosts, entry) { entry 1378 drivers/nvme/target/core.c list_for_each_entry(p, &port->subsystems, entry) { entry 75 drivers/nvme/target/discovery.c list_for_each_entry(s, &port->subsystems, entry) { entry 85 drivers/nvme/target/discovery.c if (list_empty(&port->entry)) { entry 86 drivers/nvme/target/discovery.c list_add_tail(&port->entry, &parent->referrals); entry 96 drivers/nvme/target/discovery.c if (!list_empty(&port->entry)) { entry 98 drivers/nvme/target/discovery.c list_del_init(&port->entry); entry 150 drivers/nvme/target/discovery.c list_for_each_entry(p, &req->port->subsystems, entry) { entry 155 drivers/nvme/target/discovery.c list_for_each_entry(r, &req->port->referrals, entry) entry 195 drivers/nvme/target/discovery.c list_for_each_entry(p, &req->port->subsystems, entry) { entry 208 drivers/nvme/target/discovery.c list_for_each_entry(r, &req->port->referrals, entry) { entry 556 drivers/nvme/target/loop.c list_for_each_entry(p, &nvme_loop_ports, entry) { entry 651 drivers/nvme/target/loop.c list_add_tail(&port->entry, &nvme_loop_ports); entry 659 drivers/nvme/target/loop.c list_del_init(&port->entry); entry 129 drivers/nvme/target/nvmet.h struct list_head entry; entry 259 drivers/nvme/target/nvmet.h struct list_head entry; entry 264 drivers/nvme/target/nvmet.h struct list_head entry; entry 340 drivers/nvme/target/nvmet.h struct list_head entry; entry 112 drivers/nvme/target/rdma.c struct list_head entry; entry 905 drivers/nvme/target/rdma.c list_del(&ndev->entry); entry 924 drivers/nvme/target/rdma.c list_for_each_entry(ndev, &device_list, entry) { entry 959 drivers/nvme/target/rdma.c list_add(&ndev->entry, &device_list); entry 1624 drivers/nvme/target/rdma.c list_for_each_entry(ndev, &device_list, entry) { entry 65 drivers/nvme/target/tcp.c struct list_head entry; entry 186 drivers/nvme/target/tcp.c struct nvmet_tcp_cmd, entry); entry 189 drivers/nvme/target/tcp.c list_del_init(&cmd->entry); entry 204 drivers/nvme/target/tcp.c list_add_tail(&cmd->entry, &cmd->queue->free_list); entry 453 drivers/nvme/target/tcp.c list_add(&cmd->entry, &queue->resp_send_list); entry 462 drivers/nvme/target/tcp.c struct nvmet_tcp_cmd, entry); entry 467 drivers/nvme/target/tcp.c struct nvmet_tcp_cmd, entry); entry 472 drivers/nvme/target/tcp.c list_del_init(&queue->snd_cmd->entry); entry 1241 drivers/nvme/target/tcp.c list_add_tail(&c->entry, &queue->free_list); entry 331 drivers/oprofile/buffer_sync.c static void add_data(struct op_entry *entry, struct mm_struct *mm) entry 337 drivers/oprofile/buffer_sync.c if (!op_cpu_buffer_get_data(entry, &code)) entry 339 drivers/oprofile/buffer_sync.c if (!op_cpu_buffer_get_data(entry, &pc)) entry 341 drivers/oprofile/buffer_sync.c if (!op_cpu_buffer_get_size(entry)) entry 364 drivers/oprofile/buffer_sync.c while (op_cpu_buffer_get_data(entry, &val)) entry 505 drivers/oprofile/buffer_sync.c struct op_entry entry; entry 516 drivers/oprofile/buffer_sync.c sample = op_cpu_buffer_read_entry(&entry, cpu); entry 534 drivers/oprofile/buffer_sync.c && op_cpu_buffer_get_data(&entry, &val)) { entry 544 drivers/oprofile/buffer_sync.c if (op_cpu_buffer_get_size(&entry)) entry 545 drivers/oprofile/buffer_sync.c add_data(&entry, mm); entry 144 drivers/oprofile/cpu_buffer.c *op_cpu_buffer_write_reserve(struct op_entry *entry, unsigned long size) entry 146 drivers/oprofile/cpu_buffer.c entry->event = ring_buffer_lock_reserve entry 148 drivers/oprofile/cpu_buffer.c size * sizeof(entry->sample->data[0])); entry 149 drivers/oprofile/cpu_buffer.c if (!entry->event) entry 151 drivers/oprofile/cpu_buffer.c entry->sample = ring_buffer_event_data(entry->event); entry 152 drivers/oprofile/cpu_buffer.c entry->size = size; entry 153 drivers/oprofile/cpu_buffer.c entry->data = entry->sample->data; entry 155 drivers/oprofile/cpu_buffer.c return entry->sample; entry 158 drivers/oprofile/cpu_buffer.c int op_cpu_buffer_write_commit(struct op_entry *entry) entry 160 drivers/oprofile/cpu_buffer.c return ring_buffer_unlock_commit(op_ring_buffer, entry->event); entry 163 drivers/oprofile/cpu_buffer.c struct op_sample *op_cpu_buffer_read_entry(struct op_entry *entry, int cpu) entry 170 drivers/oprofile/cpu_buffer.c entry->event = e; entry 171 drivers/oprofile/cpu_buffer.c entry->sample = ring_buffer_event_data(e); entry 172 drivers/oprofile/cpu_buffer.c entry->size = (ring_buffer_event_length(e) - sizeof(struct op_sample)) entry 173 drivers/oprofile/cpu_buffer.c / sizeof(entry->sample->data[0]); entry 174 drivers/oprofile/cpu_buffer.c entry->data = entry->sample->data; entry 175 drivers/oprofile/cpu_buffer.c return entry->sample; entry 187 drivers/oprofile/cpu_buffer.c struct op_entry entry; entry 221 drivers/oprofile/cpu_buffer.c sample = op_cpu_buffer_write_reserve(&entry, size); entry 229 drivers/oprofile/cpu_buffer.c op_cpu_buffer_add_data(&entry, (unsigned long)task); entry 231 drivers/oprofile/cpu_buffer.c op_cpu_buffer_write_commit(&entry); entry 240 drivers/oprofile/cpu_buffer.c struct op_entry entry; entry 243 drivers/oprofile/cpu_buffer.c sample = op_cpu_buffer_write_reserve(&entry, 0); entry 250 drivers/oprofile/cpu_buffer.c return op_cpu_buffer_write_commit(&entry); entry 357 drivers/oprofile/cpu_buffer.c oprofile_write_reserve(struct op_entry *entry, struct pt_regs * const regs, entry 370 drivers/oprofile/cpu_buffer.c sample = op_cpu_buffer_write_reserve(entry, size + 2); entry 376 drivers/oprofile/cpu_buffer.c op_cpu_buffer_add_data(entry, code); entry 377 drivers/oprofile/cpu_buffer.c op_cpu_buffer_add_data(entry, pc); entry 382 drivers/oprofile/cpu_buffer.c entry->event = NULL; entry 386 drivers/oprofile/cpu_buffer.c int oprofile_add_data(struct op_entry *entry, unsigned long val) entry 388 drivers/oprofile/cpu_buffer.c if (!entry->event) entry 390 drivers/oprofile/cpu_buffer.c return op_cpu_buffer_add_data(entry, val); entry 393 drivers/oprofile/cpu_buffer.c int oprofile_add_data64(struct op_entry *entry, u64 val) entry 395 drivers/oprofile/cpu_buffer.c if (!entry->event) entry 397 drivers/oprofile/cpu_buffer.c if (op_cpu_buffer_get_size(entry) < 2) entry 403 drivers/oprofile/cpu_buffer.c if (!op_cpu_buffer_add_data(entry, (u32)val)) entry 405 drivers/oprofile/cpu_buffer.c return op_cpu_buffer_add_data(entry, (u32)(val >> 32)); entry 408 drivers/oprofile/cpu_buffer.c int oprofile_write_commit(struct op_entry *entry) entry 410 drivers/oprofile/cpu_buffer.c if (!entry->event) entry 412 drivers/oprofile/cpu_buffer.c return op_cpu_buffer_write_commit(entry); entry 78 drivers/oprofile/cpu_buffer.h *op_cpu_buffer_write_reserve(struct op_entry *entry, unsigned long size); entry 79 drivers/oprofile/cpu_buffer.h int op_cpu_buffer_write_commit(struct op_entry *entry); entry 80 drivers/oprofile/cpu_buffer.h struct op_sample *op_cpu_buffer_read_entry(struct op_entry *entry, int cpu); entry 85 drivers/oprofile/cpu_buffer.h int op_cpu_buffer_add_data(struct op_entry *entry, unsigned long val) entry 87 drivers/oprofile/cpu_buffer.h if (!entry->size) entry 89 drivers/oprofile/cpu_buffer.h *entry->data = val; entry 90 drivers/oprofile/cpu_buffer.h entry->size--; entry 91 drivers/oprofile/cpu_buffer.h entry->data++; entry 92 drivers/oprofile/cpu_buffer.h return entry->size; entry 97 drivers/oprofile/cpu_buffer.h int op_cpu_buffer_get_size(struct op_entry *entry) entry 99 drivers/oprofile/cpu_buffer.h return entry->size; entry 104 drivers/oprofile/cpu_buffer.h int op_cpu_buffer_get_data(struct op_entry *entry, unsigned long *val) entry 106 drivers/oprofile/cpu_buffer.h int size = entry->size; entry 109 drivers/oprofile/cpu_buffer.h *val = *entry->data; entry 110 drivers/oprofile/cpu_buffer.h entry->size--; entry 111 drivers/oprofile/cpu_buffer.h entry->data++; entry 98 drivers/parisc/pdc_stable.c ssize_t (*show)(struct pdcspath_entry *entry, char *buf); entry 99 drivers/parisc/pdc_stable.c ssize_t (*store)(struct pdcspath_entry *entry, const char *buf, size_t count); entry 139 drivers/parisc/pdc_stable.c pdcspath_fetch(struct pdcspath_entry *entry) entry 143 drivers/parisc/pdc_stable.c if (!entry) entry 146 drivers/parisc/pdc_stable.c devpath = &entry->devpath; entry 149 drivers/parisc/pdc_stable.c entry, devpath, entry->addr); entry 152 drivers/parisc/pdc_stable.c if (pdc_stable_read(entry->addr, devpath, sizeof(*devpath)) != PDC_OK) entry 158 drivers/parisc/pdc_stable.c entry->dev = hwpath_to_device((struct hardware_path *)devpath); entry 160 drivers/parisc/pdc_stable.c entry->ready = 1; entry 162 drivers/parisc/pdc_stable.c DPRINTK("%s: device: 0x%p\n", __func__, entry->dev); entry 180 drivers/parisc/pdc_stable.c pdcspath_store(struct pdcspath_entry *entry) entry 184 drivers/parisc/pdc_stable.c BUG_ON(!entry); entry 186 drivers/parisc/pdc_stable.c devpath = &entry->devpath; entry 191 drivers/parisc/pdc_stable.c if (!entry->ready) { entry 193 drivers/parisc/pdc_stable.c BUG_ON(!entry->dev); entry 194 drivers/parisc/pdc_stable.c device_to_hwpath(entry->dev, (struct hardware_path *)devpath); entry 199 drivers/parisc/pdc_stable.c entry, devpath, entry->addr); entry 202 drivers/parisc/pdc_stable.c if (pdc_stable_write(entry->addr, devpath, sizeof(*devpath)) != PDC_OK) entry 208 drivers/parisc/pdc_stable.c entry->ready = 2; entry 210 drivers/parisc/pdc_stable.c DPRINTK("%s: device: 0x%p\n", __func__, entry->dev); entry 221 drivers/parisc/pdc_stable.c pdcspath_hwpath_read(struct pdcspath_entry *entry, char *buf) entry 227 drivers/parisc/pdc_stable.c if (!entry || !buf) entry 230 drivers/parisc/pdc_stable.c read_lock(&entry->rw_lock); entry 231 drivers/parisc/pdc_stable.c devpath = &entry->devpath; entry 232 drivers/parisc/pdc_stable.c i = entry->ready; entry 233 drivers/parisc/pdc_stable.c read_unlock(&entry->rw_lock); entry 264 drivers/parisc/pdc_stable.c pdcspath_hwpath_write(struct pdcspath_entry *entry, const char *buf, size_t count) entry 272 drivers/parisc/pdc_stable.c if (!entry || !buf || !count) entry 309 drivers/parisc/pdc_stable.c "hardware path: %s\n", __func__, entry->name, buf); entry 314 drivers/parisc/pdc_stable.c write_lock(&entry->rw_lock); entry 315 drivers/parisc/pdc_stable.c entry->ready = 0; entry 316 drivers/parisc/pdc_stable.c entry->dev = dev; entry 319 drivers/parisc/pdc_stable.c pdcspath_store(entry); entry 322 drivers/parisc/pdc_stable.c sysfs_remove_link(&entry->kobj, "device"); entry 323 drivers/parisc/pdc_stable.c write_unlock(&entry->rw_lock); entry 325 drivers/parisc/pdc_stable.c ret = sysfs_create_link(&entry->kobj, &entry->dev->kobj, "device"); entry 329 drivers/parisc/pdc_stable.c entry->name, buf); entry 342 drivers/parisc/pdc_stable.c pdcspath_layer_read(struct pdcspath_entry *entry, char *buf) entry 348 drivers/parisc/pdc_stable.c if (!entry || !buf) entry 351 drivers/parisc/pdc_stable.c read_lock(&entry->rw_lock); entry 352 drivers/parisc/pdc_stable.c devpath = &entry->devpath; entry 353 drivers/parisc/pdc_stable.c i = entry->ready; entry 354 drivers/parisc/pdc_stable.c read_unlock(&entry->rw_lock); entry 380 drivers/parisc/pdc_stable.c pdcspath_layer_write(struct pdcspath_entry *entry, const char *buf, size_t count) entry 386 drivers/parisc/pdc_stable.c if (!entry || !buf || !count) entry 412 drivers/parisc/pdc_stable.c write_lock(&entry->rw_lock); entry 416 drivers/parisc/pdc_stable.c memcpy(&entry->devpath.layers, &layers, sizeof(layers)); entry 419 drivers/parisc/pdc_stable.c pdcspath_store(entry); entry 420 drivers/parisc/pdc_stable.c write_unlock(&entry->rw_lock); entry 423 drivers/parisc/pdc_stable.c entry->name, buf); entry 437 drivers/parisc/pdc_stable.c struct pdcspath_entry *entry = to_pdcspath_entry(kobj); entry 442 drivers/parisc/pdc_stable.c ret = pdcs_attr->show(entry, buf); entry 458 drivers/parisc/pdc_stable.c struct pdcspath_entry *entry = to_pdcspath_entry(kobj); entry 466 drivers/parisc/pdc_stable.c ret = pdcs_attr->store(entry, buf, count); entry 964 drivers/parisc/pdc_stable.c struct pdcspath_entry *entry; entry 968 drivers/parisc/pdc_stable.c for (i = 0; (entry = pdcspath_entries[i]); i++) entry 969 drivers/parisc/pdc_stable.c rwlock_init(&entry->rw_lock); entry 971 drivers/parisc/pdc_stable.c for (i = 0; (entry = pdcspath_entries[i]); i++) { entry 972 drivers/parisc/pdc_stable.c write_lock(&entry->rw_lock); entry 973 drivers/parisc/pdc_stable.c err = pdcspath_fetch(entry); entry 974 drivers/parisc/pdc_stable.c write_unlock(&entry->rw_lock); entry 979 drivers/parisc/pdc_stable.c entry->kobj.kset = paths_kset; entry 980 drivers/parisc/pdc_stable.c err = kobject_init_and_add(&entry->kobj, &ktype_pdcspath, NULL, entry 981 drivers/parisc/pdc_stable.c "%s", entry->name); entry 986 drivers/parisc/pdc_stable.c write_lock(&entry->rw_lock); entry 987 drivers/parisc/pdc_stable.c entry->ready = 2; entry 988 drivers/parisc/pdc_stable.c write_unlock(&entry->rw_lock); entry 991 drivers/parisc/pdc_stable.c if (entry->dev) { entry 992 drivers/parisc/pdc_stable.c err = sysfs_create_link(&entry->kobj, &entry->dev->kobj, "device"); entry 996 drivers/parisc/pdc_stable.c kobject_uevent(&entry->kobj, KOBJ_ADD); entry 1009 drivers/parisc/pdc_stable.c struct pdcspath_entry *entry; entry 1011 drivers/parisc/pdc_stable.c for (i = 0; (entry = pdcspath_entries[i]); i++) { entry 1012 drivers/parisc/pdc_stable.c read_lock(&entry->rw_lock); entry 1013 drivers/parisc/pdc_stable.c if (entry->ready >= 2) entry 1014 drivers/parisc/pdc_stable.c kobject_put(&entry->kobj); entry 1015 drivers/parisc/pdc_stable.c read_unlock(&entry->rw_lock); entry 22 drivers/pci/bus.c struct resource_entry *entry; entry 24 drivers/pci/bus.c entry = resource_list_create_entry(res, 0); entry 25 drivers/pci/bus.c if (!entry) { entry 30 drivers/pci/bus.c entry->offset = offset; entry 31 drivers/pci/bus.c resource_list_add_tail(entry, resources); entry 199 drivers/pci/controller/pci-xgene.c struct resource_entry *entry; entry 220 drivers/pci/controller/pci-xgene.c entry = list_first_entry(&list, struct resource_entry, node); entry 221 drivers/pci/controller/pci-xgene.c *res = *entry->res; entry 1190 drivers/pci/controller/pcie-iproc.c struct resource_entry *entry, *tmp; entry 1205 drivers/pci/controller/pcie-iproc.c entry = resource_list_create_entry(res, 0); entry 1206 drivers/pci/controller/pcie-iproc.c if (!entry) entry 1209 drivers/pci/controller/pcie-iproc.c entry->offset = res->start - range->cpu_addr; entry 1210 drivers/pci/controller/pcie-iproc.c resource_list_add(entry, head); entry 713 drivers/pci/hotplug/pnv_php.c struct msix_entry entry; entry 724 drivers/pci/hotplug/pnv_php.c entry.entry = (pcie_flag & PCI_EXP_FLAGS_IRQ) >> 9; entry 725 drivers/pci/hotplug/pnv_php.c if (entry.entry >= nr_entries) entry 729 drivers/pci/hotplug/pnv_php.c ret = pci_enable_msix_exact(pdev, &entry, 1); entry 735 drivers/pci/hotplug/pnv_php.c return entry.vector; entry 93 drivers/pci/msi.c struct msi_desc *entry; entry 105 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) { entry 106 drivers/pci/msi.c ret = arch_setup_msi_irq(dev, entry); entry 123 drivers/pci/msi.c struct msi_desc *entry; entry 125 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) entry 126 drivers/pci/msi.c if (entry->irq) entry 127 drivers/pci/msi.c for (i = 0; i < entry->nvec_used; i++) entry 128 drivers/pci/msi.c arch_teardown_msi_irq(entry->irq + i); entry 138 drivers/pci/msi.c struct msi_desc *entry; entry 140 drivers/pci/msi.c entry = NULL; entry 142 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) { entry 143 drivers/pci/msi.c if (irq == entry->irq) entry 147 drivers/pci/msi.c entry = irq_get_msi_desc(irq); entry 150 drivers/pci/msi.c if (entry) entry 151 drivers/pci/msi.c __pci_write_msi_msg(entry, &entry->msg); entry 270 drivers/pci/msi.c struct msi_desc *entry; entry 272 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) entry 273 drivers/pci/msi.c default_restore_msi_irq(dev, entry->irq); entry 276 drivers/pci/msi.c void __pci_read_msi_msg(struct msi_desc *entry, struct msi_msg *msg) entry 278 drivers/pci/msi.c struct pci_dev *dev = msi_desc_to_pci_dev(entry); entry 282 drivers/pci/msi.c if (entry->msi_attrib.is_msix) { entry 283 drivers/pci/msi.c void __iomem *base = pci_msix_desc_addr(entry); entry 299 drivers/pci/msi.c if (entry->msi_attrib.is_64) { entry 311 drivers/pci/msi.c void __pci_write_msi_msg(struct msi_desc *entry, struct msi_msg *msg) entry 313 drivers/pci/msi.c struct pci_dev *dev = msi_desc_to_pci_dev(entry); entry 317 drivers/pci/msi.c } else if (entry->msi_attrib.is_msix) { entry 318 drivers/pci/msi.c void __iomem *base = pci_msix_desc_addr(entry); entry 332 drivers/pci/msi.c msgctl |= entry->msi_attrib.multiple << 4; entry 337 drivers/pci/msi.c if (entry->msi_attrib.is_64) { entry 349 drivers/pci/msi.c entry->msg = *msg; entry 351 drivers/pci/msi.c if (entry->write_msi_msg) entry 352 drivers/pci/msi.c entry->write_msi_msg(entry, entry->write_msi_msg_data); entry 358 drivers/pci/msi.c struct msi_desc *entry = irq_get_msi_desc(irq); entry 360 drivers/pci/msi.c __pci_write_msi_msg(entry, msg); entry 367 drivers/pci/msi.c struct msi_desc *entry, *tmp; entry 372 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) entry 373 drivers/pci/msi.c if (entry->irq) entry 374 drivers/pci/msi.c for (i = 0; i < entry->nvec_used; i++) entry 375 drivers/pci/msi.c BUG_ON(irq_has_action(entry->irq + i)); entry 379 drivers/pci/msi.c list_for_each_entry_safe(entry, tmp, msi_list, list) { entry 380 drivers/pci/msi.c if (entry->msi_attrib.is_msix) { entry 381 drivers/pci/msi.c if (list_is_last(&entry->list, msi_list)) entry 382 drivers/pci/msi.c iounmap(entry->mask_base); entry 385 drivers/pci/msi.c list_del(&entry->list); entry 386 drivers/pci/msi.c free_msi_entry(entry); entry 415 drivers/pci/msi.c struct msi_desc *entry; entry 420 drivers/pci/msi.c entry = irq_get_msi_desc(dev->irq); entry 427 drivers/pci/msi.c msi_mask_irq(entry, msi_mask(entry->msi_attrib.multi_cap), entry 428 drivers/pci/msi.c entry->masked); entry 430 drivers/pci/msi.c control |= (entry->msi_attrib.multiple << 4) | PCI_MSI_FLAGS_ENABLE; entry 436 drivers/pci/msi.c struct msi_desc *entry; entry 448 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) entry 449 drivers/pci/msi.c msix_mask_irq(entry, entry->masked); entry 464 drivers/pci/msi.c struct msi_desc *entry; entry 472 drivers/pci/msi.c entry = irq_get_msi_desc(irq); entry 473 drivers/pci/msi.c if (entry) entry 475 drivers/pci/msi.c entry->msi_attrib.is_msix ? "msix" : "msi"); entry 487 drivers/pci/msi.c struct msi_desc *entry; entry 494 drivers/pci/msi.c for_each_pci_msi_entry(entry, pdev) entry 495 drivers/pci/msi.c num_msi += entry->nvec_used; entry 503 drivers/pci/msi.c for_each_pci_msi_entry(entry, pdev) { entry 504 drivers/pci/msi.c for (i = 0; i < entry->nvec_used; i++) { entry 512 drivers/pci/msi.c entry->irq + i); entry 561 drivers/pci/msi.c struct msi_desc *entry; entry 568 drivers/pci/msi.c entry = alloc_msi_entry(&dev->dev, nvec, masks); entry 569 drivers/pci/msi.c if (!entry) entry 574 drivers/pci/msi.c entry->msi_attrib.is_msix = 0; entry 575 drivers/pci/msi.c entry->msi_attrib.is_64 = !!(control & PCI_MSI_FLAGS_64BIT); entry 576 drivers/pci/msi.c entry->msi_attrib.is_virtual = 0; entry 577 drivers/pci/msi.c entry->msi_attrib.entry_nr = 0; entry 578 drivers/pci/msi.c entry->msi_attrib.maskbit = !!(control & PCI_MSI_FLAGS_MASKBIT); entry 579 drivers/pci/msi.c entry->msi_attrib.default_irq = dev->irq; /* Save IOAPIC IRQ */ entry 580 drivers/pci/msi.c entry->msi_attrib.multi_cap = (control & PCI_MSI_FLAGS_QMASK) >> 1; entry 581 drivers/pci/msi.c entry->msi_attrib.multiple = ilog2(__roundup_pow_of_two(nvec)); entry 584 drivers/pci/msi.c entry->mask_pos = dev->msi_cap + PCI_MSI_MASK_64; entry 586 drivers/pci/msi.c entry->mask_pos = dev->msi_cap + PCI_MSI_MASK_32; entry 589 drivers/pci/msi.c if (entry->msi_attrib.maskbit) entry 590 drivers/pci/msi.c pci_read_config_dword(dev, entry->mask_pos, &entry->masked); entry 594 drivers/pci/msi.c return entry; entry 599 drivers/pci/msi.c struct msi_desc *entry; entry 601 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) { entry 602 drivers/pci/msi.c if (!dev->no_64bit_msi || !entry->msg.address_hi) entry 626 drivers/pci/msi.c struct msi_desc *entry; entry 632 drivers/pci/msi.c entry = msi_setup_entry(dev, nvec, affd); entry 633 drivers/pci/msi.c if (!entry) entry 637 drivers/pci/msi.c mask = msi_mask(entry->msi_attrib.multi_cap); entry 638 drivers/pci/msi.c msi_mask_irq(entry, mask, mask); entry 640 drivers/pci/msi.c list_add_tail(&entry->list, dev_to_msi_list(&dev->dev)); entry 645 drivers/pci/msi.c msi_mask_irq(entry, mask, ~mask); entry 652 drivers/pci/msi.c msi_mask_irq(entry, mask, ~mask); entry 659 drivers/pci/msi.c msi_mask_irq(entry, mask, ~mask); entry 670 drivers/pci/msi.c dev->irq = entry->irq; entry 699 drivers/pci/msi.c struct msi_desc *entry; entry 707 drivers/pci/msi.c entry = alloc_msi_entry(&dev->dev, 1, curmsk); entry 708 drivers/pci/msi.c if (!entry) { entry 718 drivers/pci/msi.c entry->msi_attrib.is_msix = 1; entry 719 drivers/pci/msi.c entry->msi_attrib.is_64 = 1; entry 721 drivers/pci/msi.c entry->msi_attrib.entry_nr = entries[i].entry; entry 723 drivers/pci/msi.c entry->msi_attrib.entry_nr = i; entry 725 drivers/pci/msi.c entry->msi_attrib.is_virtual = entry 726 drivers/pci/msi.c entry->msi_attrib.entry_nr >= vec_count; entry 728 drivers/pci/msi.c entry->msi_attrib.default_irq = dev->irq; entry 729 drivers/pci/msi.c entry->mask_base = base; entry 731 drivers/pci/msi.c list_add_tail(&entry->list, dev_to_msi_list(&dev->dev)); entry 744 drivers/pci/msi.c struct msi_desc *entry; entry 748 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) { entry 750 drivers/pci/msi.c entries[i++].vector = entry->irq; entry 752 drivers/pci/msi.c desc_addr = pci_msix_desc_addr(entry); entry 754 drivers/pci/msi.c entry->masked = readl(desc_addr + entry 757 drivers/pci/msi.c entry->masked = 0; entry 759 drivers/pci/msi.c msix_mask_irq(entry, 1); entry 831 drivers/pci/msi.c struct msi_desc *entry; entry 834 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) { entry 835 drivers/pci/msi.c if (entry->irq != 0) entry 988 drivers/pci/msi.c if (entries[i].entry >= nr_entries) entry 991 drivers/pci/msi.c if (entries[i].entry == entries[j].entry) entry 1007 drivers/pci/msi.c struct msi_desc *entry; entry 1018 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) { entry 1020 drivers/pci/msi.c __pci_msix_desc_mask_irq(entry, 1); entry 1261 drivers/pci/msi.c struct msi_desc *entry; entry 1264 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) { entry 1266 drivers/pci/msi.c return entry->irq; entry 1274 drivers/pci/msi.c struct msi_desc *entry = first_pci_msi_entry(dev); entry 1276 drivers/pci/msi.c if (WARN_ON_ONCE(nr >= entry->nvec_used)) entry 1295 drivers/pci/msi.c struct msi_desc *entry; entry 1298 drivers/pci/msi.c for_each_pci_msi_entry(entry, dev) { entry 1300 drivers/pci/msi.c return &entry->affinity->mask; entry 1306 drivers/pci/msi.c struct msi_desc *entry = first_pci_msi_entry(dev); entry 1308 drivers/pci/msi.c if (WARN_ON_ONCE(!entry || !entry->affinity || entry 1309 drivers/pci/msi.c nr >= entry->nvec_used)) entry 1312 drivers/pci/msi.c return &entry->affinity[nr].mask; entry 299 drivers/pci/p2pdma.c const struct pci_p2pdma_whitelist_entry *entry; entry 309 drivers/pci/p2pdma.c for (entry = pci_p2pdma_whitelist; entry->vendor; entry++) { entry 310 drivers/pci/p2pdma.c if (vendor != entry->vendor || device != entry->device) entry 312 drivers/pci/p2pdma.c if (entry->flags & REQ_SAME_HOST_BRIDGE && !same_host_bridge) entry 34 drivers/pci/pci-acpi.c struct resource_entry *entry; entry 55 drivers/pci/pci-acpi.c entry = list_first_entry(&list, struct resource_entry, node); entry 56 drivers/pci/pci-acpi.c *res = *entry->res; entry 1025 drivers/pci/pcie/aer.c struct aer_recover_entry entry; entry 1028 drivers/pci/pcie/aer.c while (kfifo_get(&aer_recover_ring, &entry)) { entry 1029 drivers/pci/pcie/aer.c pdev = pci_get_domain_bus_and_slot(entry.domain, entry.bus, entry 1030 drivers/pci/pcie/aer.c entry.devfn); entry 1033 drivers/pci/pcie/aer.c entry.domain, entry.bus, entry 1034 drivers/pci/pcie/aer.c PCI_SLOT(entry.devfn), PCI_FUNC(entry.devfn)); entry 1037 drivers/pci/pcie/aer.c cper_print_aer(pdev, entry.severity, entry.regs); entry 1038 drivers/pci/pcie/aer.c if (entry.severity == AER_NONFATAL) entry 1041 drivers/pci/pcie/aer.c else if (entry.severity == AER_FATAL) entry 1059 drivers/pci/pcie/aer.c struct aer_recover_entry entry = { entry 1067 drivers/pci/pcie/aer.c if (kfifo_in_spinlocked(&aer_recover_ring, &entry, 1, entry 262 drivers/pci/xen-pcifront.c struct msi_desc *entry; entry 271 drivers/pci/xen-pcifront.c for_each_pci_msi_entry(entry, dev) { entry 272 drivers/pci/xen-pcifront.c op.msix_entries[i].entry = entry->msi_attrib.entry_nr; entry 1078 drivers/pcmcia/cistpl.c cistpl_cftable_entry_t *entry) entry 1084 drivers/pcmcia/cistpl.c entry->index = *p & 0x3f; entry 1085 drivers/pcmcia/cistpl.c entry->flags = 0; entry 1087 drivers/pcmcia/cistpl.c entry->flags |= CISTPL_CFTABLE_DEFAULT; entry 1092 drivers/pcmcia/cistpl.c entry->flags |= CISTPL_CFTABLE_BVDS; entry 1094 drivers/pcmcia/cistpl.c entry->flags |= CISTPL_CFTABLE_WP; entry 1096 drivers/pcmcia/cistpl.c entry->flags |= CISTPL_CFTABLE_RDYBSY; entry 1098 drivers/pcmcia/cistpl.c entry->flags |= CISTPL_CFTABLE_MWAIT; entry 1099 drivers/pcmcia/cistpl.c entry->interface = *p & 0x0f; entry 1101 drivers/pcmcia/cistpl.c entry->interface = 0; entry 1110 drivers/pcmcia/cistpl.c p = parse_power(p, q, &entry->vcc); entry 1114 drivers/pcmcia/cistpl.c entry->vcc.present = 0; entry 1116 drivers/pcmcia/cistpl.c p = parse_power(p, q, &entry->vpp1); entry 1120 drivers/pcmcia/cistpl.c entry->vpp1.present = 0; entry 1122 drivers/pcmcia/cistpl.c p = parse_power(p, q, &entry->vpp2); entry 1126 drivers/pcmcia/cistpl.c entry->vpp2.present = 0; entry 1130 drivers/pcmcia/cistpl.c p = parse_timing(p, q, &entry->timing); entry 1134 drivers/pcmcia/cistpl.c entry->timing.wait = 0; entry 1135 drivers/pcmcia/cistpl.c entry->timing.ready = 0; entry 1136 drivers/pcmcia/cistpl.c entry->timing.reserved = 0; entry 1141 drivers/pcmcia/cistpl.c p = parse_io(p, q, &entry->io); entry 1145 drivers/pcmcia/cistpl.c entry->io.nwin = 0; entry 1149 drivers/pcmcia/cistpl.c p = parse_irq(p, q, &entry->irq); entry 1153 drivers/pcmcia/cistpl.c entry->irq.IRQInfo1 = 0; entry 1157 drivers/pcmcia/cistpl.c entry->mem.nwin = 0; entry 1160 drivers/pcmcia/cistpl.c entry->mem.nwin = 1; entry 1161 drivers/pcmcia/cistpl.c entry->mem.win[0].len = get_unaligned_le16(p) << 8; entry 1162 drivers/pcmcia/cistpl.c entry->mem.win[0].card_addr = 0; entry 1163 drivers/pcmcia/cistpl.c entry->mem.win[0].host_addr = 0; entry 1169 drivers/pcmcia/cistpl.c entry->mem.nwin = 1; entry 1170 drivers/pcmcia/cistpl.c entry->mem.win[0].len = get_unaligned_le16(p) << 8; entry 1171 drivers/pcmcia/cistpl.c entry->mem.win[0].card_addr = get_unaligned_le16(p + 2) << 8; entry 1172 drivers/pcmcia/cistpl.c entry->mem.win[0].host_addr = 0; entry 1178 drivers/pcmcia/cistpl.c p = parse_mem(p, q, &entry->mem); entry 1188 drivers/pcmcia/cistpl.c entry->flags |= (*p << 8); entry 1195 drivers/pcmcia/cistpl.c entry->subtuples = q-p; entry 66 drivers/perf/thunderx2_pmu.c struct list_head entry; entry 603 drivers/perf/thunderx2_pmu.c list_add(&tx2_pmu->entry, &tx2_pmus); entry 652 drivers/perf/thunderx2_pmu.c INIT_LIST_HEAD(&tx2_pmu->entry); entry 802 drivers/perf/thunderx2_pmu.c list_for_each_entry_safe(tx2_pmu, temp, &tx2_pmus, entry) { entry 808 drivers/perf/thunderx2_pmu.c list_del(&tx2_pmu->entry); entry 538 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pd-entry-vld, 0x07), entry 539 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(sref-entry-vld, 0x08), entry 507 drivers/pinctrl/pinctrl-rza1.c const struct rza1_bidir_entry *entry = &table[port]; entry 511 drivers/pinctrl/pinctrl-rza1.c for (i = 0; i < entry->npins; ++i) { entry 512 drivers/pinctrl/pinctrl-rza1.c bidir_pin = &entry->pins[i]; entry 564 drivers/platform/x86/dell-wmi.c const struct key_entry *entry = &dell_wmi_keymap_type_0010[i]; entry 572 drivers/platform/x86/dell-wmi.c have_scancode(entry->code | (0x0010 << 16), entry 577 drivers/platform/x86/dell-wmi.c keymap[pos] = *entry; entry 9934 drivers/platform/x86/thinkpad_acpi.c struct proc_dir_entry *entry; entry 9986 drivers/platform/x86/thinkpad_acpi.c entry = proc_create_data(ibm->name, mode, proc_dir, entry 9988 drivers/platform/x86/thinkpad_acpi.c if (!entry) { entry 2476 drivers/power/supply/ab8500_fg.c struct ab8500_fg_sysfs_entry *entry; entry 2479 drivers/power/supply/ab8500_fg.c entry = container_of(attr, struct ab8500_fg_sysfs_entry, attr); entry 2482 drivers/power/supply/ab8500_fg.c if (!entry->show) entry 2485 drivers/power/supply/ab8500_fg.c return entry->show(di, buf); entry 2491 drivers/power/supply/ab8500_fg.c struct ab8500_fg_sysfs_entry *entry; entry 2494 drivers/power/supply/ab8500_fg.c entry = container_of(attr, struct ab8500_fg_sysfs_entry, attr); entry 2497 drivers/power/supply/ab8500_fg.c if (!entry->store) entry 2500 drivers/power/supply/ab8500_fg.c return entry->store(di, buf, count); entry 1841 drivers/power/supply/abx500_chargalg.c struct abx500_chargalg_sysfs_entry *entry = container_of(attr, entry 1847 drivers/power/supply/abx500_chargalg.c if (!entry->show) entry 1850 drivers/power/supply/abx500_chargalg.c return entry->show(di, buf); entry 1856 drivers/power/supply/abx500_chargalg.c struct abx500_chargalg_sysfs_entry *entry = container_of(attr, entry 1862 drivers/power/supply/abx500_chargalg.c if (!entry->store) entry 1865 drivers/power/supply/abx500_chargalg.c return entry->store(di, buf, length); entry 745 drivers/power/supply/charger-manager.c list_for_each_entry(cm, &cm_list, entry) { entry 768 drivers/power/supply/charger-manager.c list_for_each_entry(cm, &cm_list, entry) { entry 1071 drivers/power/supply/charger-manager.c list_for_each_entry(cm, &cm_list, entry) { entry 1790 drivers/power/supply/charger-manager.c list_add(&cm->entry, &cm_list); entry 1840 drivers/power/supply/charger-manager.c list_del(&cm->entry); entry 1889 drivers/power/supply/charger-manager.c list_for_each_entry(cm, &cm_list, entry) { entry 2022 drivers/power/supply/charger-manager.c list_for_each_entry(cm, &cm_list, entry) { entry 727 drivers/rapidio/devices/tsi721.c entries[TSI721_VECT_IDB].entry = TSI721_MSIX_SR2PC_IDBQ_RCV(IDB_QUEUE); entry 728 drivers/rapidio/devices/tsi721.c entries[TSI721_VECT_PWRX].entry = TSI721_MSIX_SRIO_MAC_INT; entry 737 drivers/rapidio/devices/tsi721.c entries[TSI721_VECT_IMB0_RCV + i].entry = entry 739 drivers/rapidio/devices/tsi721.c entries[TSI721_VECT_IMB0_INT + i].entry = entry 741 drivers/rapidio/devices/tsi721.c entries[TSI721_VECT_OMB0_DONE + i].entry = entry 743 drivers/rapidio/devices/tsi721.c entries[TSI721_VECT_OMB0_INT + i].entry = entry 754 drivers/rapidio/devices/tsi721.c entries[TSI721_VECT_DMA0_DONE + i].entry = entry 756 drivers/rapidio/devices/tsi721.c entries[TSI721_VECT_DMA0_INT + i].entry = entry 41 drivers/rapidio/switches/idt_gen3.c u32 entry = route_port; entry 45 drivers/rapidio/switches/idt_gen3.c __func__, table, route_destid, entry); entry 51 drivers/rapidio/switches/idt_gen3.c entry = RIO_RT_ENTRY_DROP_PKT; entry 57 drivers/rapidio/switches/idt_gen3.c entry); entry 74 drivers/rapidio/switches/idt_gen3.c entry); entry 1174 drivers/remoteproc/remoteproc_core.c struct rproc_mem_entry *entry, *tmp; entry 1180 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(entry, tmp, &rproc->carveouts, node) { entry 1181 drivers/remoteproc/remoteproc_core.c if (entry->alloc) { entry 1182 drivers/remoteproc/remoteproc_core.c ret = entry->alloc(rproc, entry); entry 1185 drivers/remoteproc/remoteproc_core.c entry->name, ret); entry 1190 drivers/remoteproc/remoteproc_core.c if (entry->rsc_offset != FW_RSC_ADDR_ANY) { entry 1192 drivers/remoteproc/remoteproc_core.c rsc = (void *)rproc->table_ptr + entry->rsc_offset; entry 1213 drivers/remoteproc/remoteproc_core.c if (entry->va) entry 1214 drivers/remoteproc/remoteproc_core.c pa = (u64)rproc_va_to_pa(entry->va); entry 1216 drivers/remoteproc/remoteproc_core.c pa = (u64)entry->dma; entry 1223 drivers/remoteproc/remoteproc_core.c rsc->da = entry->da; entry 1224 drivers/remoteproc/remoteproc_core.c rsc->len = entry->len; entry 1237 drivers/remoteproc/remoteproc_core.c struct rproc_dump_segment *entry, *tmp; entry 1239 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(entry, tmp, &rproc->dump_segments, node) { entry 1240 drivers/remoteproc/remoteproc_core.c list_del(&entry->node); entry 1241 drivers/remoteproc/remoteproc_core.c kfree(entry); entry 1254 drivers/remoteproc/remoteproc_core.c struct rproc_mem_entry *entry, *tmp; entry 1268 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(entry, tmp, &rproc->mappings, node) { entry 1271 drivers/remoteproc/remoteproc_core.c unmapped = iommu_unmap(rproc->domain, entry->da, entry->len); entry 1272 drivers/remoteproc/remoteproc_core.c if (unmapped != entry->len) { entry 1274 drivers/remoteproc/remoteproc_core.c dev_err(dev, "failed to unmap %u/%zu\n", entry->len, entry 1278 drivers/remoteproc/remoteproc_core.c list_del(&entry->node); entry 1279 drivers/remoteproc/remoteproc_core.c kfree(entry); entry 1283 drivers/remoteproc/remoteproc_core.c list_for_each_entry_safe(entry, tmp, &rproc->carveouts, node) { entry 1284 drivers/remoteproc/remoteproc_core.c if (entry->release) entry 1285 drivers/remoteproc/remoteproc_core.c entry->release(rproc, entry); entry 1286 drivers/remoteproc/remoteproc_core.c list_del(&entry->node); entry 1287 drivers/remoteproc/remoteproc_core.c kfree(entry); entry 170 drivers/reset/core.c struct reset_control_lookup *entry; entry 175 drivers/reset/core.c entry = &lookup[i]; entry 177 drivers/reset/core.c if (!entry->dev_id || !entry->provider) { entry 183 drivers/reset/core.c list_add_tail(&entry->list, &reset_lookup_list); entry 1193 drivers/rpmsg/qcom_smd.c struct qcom_smd_alloc_entry *entry; entry 1209 drivers/rpmsg/qcom_smd.c entry = &alloc_tbl[i]; entry 1210 drivers/rpmsg/qcom_smd.c eflags = le32_to_cpu(entry->flags); entry 1214 drivers/rpmsg/qcom_smd.c if (entry->ref_count == 0) entry 1217 drivers/rpmsg/qcom_smd.c if (!entry->name[0]) entry 1226 drivers/rpmsg/qcom_smd.c cid = le32_to_cpu(entry->cid); entry 1230 drivers/rpmsg/qcom_smd.c channel = qcom_smd_create_channel(edge, info_id, fifo_id, entry->name); entry 6031 drivers/s390/block/dasd_eckd.c struct dasd_ckd_path_group_entry *entry; entry 6051 drivers/s390/block/dasd_eckd.c entry = (struct dasd_ckd_path_group_entry *) entry 6052 drivers/s390/block/dasd_eckd.c (info->entry + i * info->entry_size); entry 6053 drivers/s390/block/dasd_eckd.c if (entry->status_flags & DASD_ECKD_PG_GROUPED) entry 6067 drivers/s390/block/dasd_eckd.c struct dasd_ckd_path_group_entry *entry; entry 6087 drivers/s390/block/dasd_eckd.c entry = (struct dasd_ckd_path_group_entry *) entry 6088 drivers/s390/block/dasd_eckd.c (info->entry + i * info->entry_size); entry 6090 drivers/s390/block/dasd_eckd.c seq_printf(m, "pgid %*phN\n", 11, entry->pgid); entry 6092 drivers/s390/block/dasd_eckd.c seq_printf(m, "status_flags %02x\n", entry->status_flags); entry 6094 drivers/s390/block/dasd_eckd.c memcpy(&sysplex, &entry->sysplex_name, sizeof(sysplex) - 1); entry 6098 drivers/s390/block/dasd_eckd.c seq_printf(m, "supported_cylinder %d\n", entry->cylinder); entry 6101 drivers/s390/block/dasd_eckd.c entry->timestamp); entry 512 drivers/s390/block/dasd_eckd.h __u8 entry[16390]; entry 112 drivers/s390/block/dcssblk.c struct segment_info *entry, *temp; entry 115 drivers/s390/block/dcssblk.c list_for_each_entry_safe(entry, temp, &dev_info->seg_list, lh) { entry 116 drivers/s390/block/dcssblk.c list_del(&entry->lh); entry 117 drivers/s390/block/dcssblk.c kfree(entry); entry 133 drivers/s390/block/dcssblk.c struct dcssblk_dev_info *entry; entry 140 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dcssblk_devices, lh) entry 141 drivers/s390/block/dcssblk.c if (minor == entry->gd->first_minor) entry 159 drivers/s390/block/dcssblk.c struct dcssblk_dev_info *entry; entry 161 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dcssblk_devices, lh) { entry 162 drivers/s390/block/dcssblk.c if (!strcmp(name, entry->segment_name)) { entry 163 drivers/s390/block/dcssblk.c return entry; entry 178 drivers/s390/block/dcssblk.c struct segment_info *entry; entry 181 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 182 drivers/s390/block/dcssblk.c if (!strcmp(name, entry->segment_name)) entry 183 drivers/s390/block/dcssblk.c return entry; entry 196 drivers/s390/block/dcssblk.c struct segment_info *entry; entry 199 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 200 drivers/s390/block/dcssblk.c if (highest_addr < entry->end) entry 201 drivers/s390/block/dcssblk.c highest_addr = entry->end; entry 214 drivers/s390/block/dcssblk.c struct segment_info *entry; entry 218 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 220 drivers/s390/block/dcssblk.c lowest_addr = entry->start; entry 223 drivers/s390/block/dcssblk.c if (lowest_addr > entry->start) entry 224 drivers/s390/block/dcssblk.c lowest_addr = entry->start; entry 237 drivers/s390/block/dcssblk.c struct segment_info *sort_list, *entry, temp; entry 248 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 249 drivers/s390/block/dcssblk.c memcpy(&sort_list[i], entry, sizeof(struct segment_info)); entry 348 drivers/s390/block/dcssblk.c struct segment_info *entry, *temp; entry 361 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 362 drivers/s390/block/dcssblk.c rc = segment_modify_shared(entry->segment_name, entry 386 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 387 drivers/s390/block/dcssblk.c rc = segment_modify_shared(entry->segment_name, entry 407 drivers/s390/block/dcssblk.c temp = entry; entry 408 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 409 drivers/s390/block/dcssblk.c if (entry != temp) entry 410 drivers/s390/block/dcssblk.c segment_unload(entry->segment_name); entry 454 drivers/s390/block/dcssblk.c struct segment_info *entry; entry 466 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 467 drivers/s390/block/dcssblk.c if (entry->segment_type == SEG_TYPE_EN || entry 468 drivers/s390/block/dcssblk.c entry->segment_type == SEG_TYPE_SN) entry 471 drivers/s390/block/dcssblk.c entry->segment_name); entry 473 drivers/s390/block/dcssblk.c segment_save(entry->segment_name); entry 512 drivers/s390/block/dcssblk.c struct segment_info *entry; entry 518 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 519 drivers/s390/block/dcssblk.c strcpy(&buf[i], entry->segment_name); entry 520 drivers/s390/block/dcssblk.c i += strlen(entry->segment_name); entry 745 drivers/s390/block/dcssblk.c struct segment_info *entry; entry 794 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) entry 795 drivers/s390/block/dcssblk.c segment_unload(entry->segment_name); entry 830 drivers/s390/block/dcssblk.c struct segment_info *entry; entry 841 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 842 drivers/s390/block/dcssblk.c if (entry->segment_type == SEG_TYPE_EN || entry 843 drivers/s390/block/dcssblk.c entry->segment_type == SEG_TYPE_SN) entry 845 drivers/s390/block/dcssblk.c " be saved\n", entry->segment_name); entry 847 drivers/s390/block/dcssblk.c segment_save(entry->segment_name); entry 1019 drivers/s390/block/dcssblk.c struct segment_info *entry; entry 1024 drivers/s390/block/dcssblk.c list_for_each_entry(entry, &dev_info->seg_list, lh) { entry 1025 drivers/s390/block/dcssblk.c segment_unload(entry->segment_name); entry 1026 drivers/s390/block/dcssblk.c rc = segment_load(entry->segment_name, SEGMENT_SHARED, entry 1030 drivers/s390/block/dcssblk.c segment_warning(rc, entry->segment_name); entry 1033 drivers/s390/block/dcssblk.c if (start != entry->start || end != entry->end) { entry 1036 drivers/s390/block/dcssblk.c entry->segment_name); entry 91 drivers/s390/char/monwriter.c struct mon_buf *entry, *next; entry 93 drivers/s390/char/monwriter.c list_for_each_entry_safe(entry, next, &monpriv->list, list) entry 94 drivers/s390/char/monwriter.c if ((entry->hdr.mon_function == monhdr->mon_function || entry 96 drivers/s390/char/monwriter.c entry->hdr.applid == monhdr->applid && entry 97 drivers/s390/char/monwriter.c entry->hdr.record_num == monhdr->record_num && entry 98 drivers/s390/char/monwriter.c entry->hdr.version == monhdr->version && entry 99 drivers/s390/char/monwriter.c entry->hdr.release == monhdr->release && entry 100 drivers/s390/char/monwriter.c entry->hdr.mod_level == monhdr->mod_level) entry 101 drivers/s390/char/monwriter.c return entry; entry 209 drivers/s390/char/monwriter.c struct mon_buf *entry, *next; entry 211 drivers/s390/char/monwriter.c list_for_each_entry_safe(entry, next, &monpriv->list, list) { entry 212 drivers/s390/char/monwriter.c if (entry->hdr.mon_function != MONWRITE_GEN_EVENT) entry 213 drivers/s390/char/monwriter.c monwrite_diag(&entry->hdr, entry->data, entry 216 drivers/s390/char/monwriter.c list_del(&entry->list); entry 217 drivers/s390/char/monwriter.c kfree(entry->data); entry 218 drivers/s390/char/monwriter.c kfree(entry); entry 413 drivers/s390/cio/blacklist.c struct proc_dir_entry *entry; entry 415 drivers/s390/cio/blacklist.c entry = proc_create("cio_ignore", S_IFREG | S_IRUGO | S_IWUSR, NULL, entry 417 drivers/s390/cio/blacklist.c if (!entry) entry 161 drivers/s390/cio/chp.c struct cmg_entry *entry, reference_buf; entry 171 drivers/s390/cio/chp.c entry = area + (idx * sizeof(struct cmg_entry)); entry 173 drivers/s390/cio/chp.c memcpy(buf, entry, sizeof(*entry)); entry 174 drivers/s390/cio/chp.c memcpy(&reference_buf, entry, sizeof(*entry)); entry 1383 drivers/s390/cio/css.c struct proc_dir_entry *entry; entry 1385 drivers/s390/cio/css.c entry = proc_create("cio_settle", S_IWUSR, NULL, entry 1387 drivers/s390/cio/css.c if (!entry) entry 256 drivers/s390/cio/qdio.h struct list_head entry; entry 34 drivers/s390/cio/qdio_debug.c struct qdio_dbf_entry *entry; entry 38 drivers/s390/cio/qdio_debug.c list_for_each_entry(entry, &qdio_dbf_list, dbf_list) { entry 39 drivers/s390/cio/qdio_debug.c if (strcmp(entry->dbf_name, name) == 0) { entry 40 drivers/s390/cio/qdio_debug.c rc = entry->dbf_info; entry 50 drivers/s390/cio/qdio_debug.c struct qdio_dbf_entry *entry, *tmp; entry 53 drivers/s390/cio/qdio_debug.c list_for_each_entry_safe(entry, tmp, &qdio_dbf_list, dbf_list) { entry 54 drivers/s390/cio/qdio_debug.c list_del(&entry->dbf_list); entry 55 drivers/s390/cio/qdio_debug.c debug_unregister(entry->dbf_info); entry 56 drivers/s390/cio/qdio_debug.c kfree(entry); entry 1784 drivers/s390/cio/qdio_main.c void *entry), entry 154 drivers/s390/cio/qdio_setup.c INIT_LIST_HEAD(&q->entry); entry 183 drivers/s390/cio/qdio_setup.c INIT_LIST_HEAD(&q->entry); entry 80 drivers/s390/cio/qdio_thinint.c list_add_rcu(&irq_ptr->input_qs[0]->entry, &tiq_list); entry 93 drivers/s390/cio/qdio_thinint.c list_del_rcu(&q->entry); entry 96 drivers/s390/cio/qdio_thinint.c INIT_LIST_HEAD(&q->entry); entry 194 drivers/s390/cio/qdio_thinint.c list_for_each_entry_rcu(q, &tiq_list, entry) { entry 298 drivers/s390/cio/trace.h memcpy(&entry->request, chsc, entry 302 drivers/s390/cio/trace.h memcpy(&entry->response, chsc, entry 181 drivers/s390/net/ism.h struct smcd_event entry[15]; entry 392 drivers/s390/net/ism_drv.c struct smcd_event *entry; entry 395 drivers/s390/net/ism_drv.c if (++(ism->ieq_idx) == ARRAY_SIZE(ism->ieq->entry)) entry 398 drivers/s390/net/ism_drv.c entry = &ism->ieq->entry[ism->ieq_idx]; entry 399 drivers/s390/net/ism_drv.c debug_event(ism_debug_info, 2, entry, sizeof(*entry)); entry 400 drivers/s390/net/ism_drv.c smcd_handle_event(ism->smcd, entry); entry 967 drivers/s390/net/qeth_core.h struct qeth_buffer_pool_entry *entry) entry 969 drivers/s390/net/qeth_core.h list_add_tail(&entry->list, &card->qdio.in_buf_pool.entry_list); entry 2571 drivers/s390/net/qeth_core_main.c struct qeth_buffer_pool_entry *entry; entry 2575 drivers/s390/net/qeth_core_main.c list_for_each_entry(entry, entry 2577 drivers/s390/net/qeth_core_main.c qeth_put_buffer_pool_entry(card, entry); entry 2585 drivers/s390/net/qeth_core_main.c struct qeth_buffer_pool_entry *entry; entry 2593 drivers/s390/net/qeth_core_main.c entry = list_entry(plh, struct qeth_buffer_pool_entry, list); entry 2596 drivers/s390/net/qeth_core_main.c if (page_count(virt_to_page(entry->elements[i])) > 1) { entry 2602 drivers/s390/net/qeth_core_main.c list_del_init(&entry->list); entry 2603 drivers/s390/net/qeth_core_main.c return entry; entry 2608 drivers/s390/net/qeth_core_main.c entry = list_entry(card->qdio.in_buf_pool.entry_list.next, entry 2611 drivers/s390/net/qeth_core_main.c if (page_count(virt_to_page(entry->elements[i])) > 1) { entry 2616 drivers/s390/net/qeth_core_main.c free_page((unsigned long)entry->elements[i]); entry 2617 drivers/s390/net/qeth_core_main.c entry->elements[i] = page_address(page); entry 2622 drivers/s390/net/qeth_core_main.c list_del_init(&entry->list); entry 2623 drivers/s390/net/qeth_core_main.c return entry; entry 5522 drivers/s390/net/qeth_core_main.c struct qeth_dbf_entry *entry; entry 5526 drivers/s390/net/qeth_core_main.c list_for_each_entry(entry, &qeth_dbf_list, dbf_list) { entry 5527 drivers/s390/net/qeth_core_main.c if (strcmp(entry->dbf_name, name) == 0) { entry 5528 drivers/s390/net/qeth_core_main.c rc = entry->dbf_info; entry 5566 drivers/s390/net/qeth_core_main.c struct qeth_dbf_entry *entry, *tmp; entry 5569 drivers/s390/net/qeth_core_main.c list_for_each_entry_safe(entry, tmp, &qeth_dbf_list, dbf_list) { entry 5570 drivers/s390/net/qeth_core_main.c list_del(&entry->dbf_list); entry 5571 drivers/s390/net/qeth_core_main.c debug_unregister(entry->dbf_info); entry 5572 drivers/s390/net/qeth_core_main.c kfree(entry); entry 710 drivers/s390/net/qeth_core_mpc.h struct qeth_sbp_port_entry entry[]; entry 718 drivers/s390/net/qeth_core_mpc.h struct qeth_sbp_port_entry entry[]; entry 755 drivers/s390/net/qeth_core_mpc.h struct qeth_ipacmd_addr_change_entry entry[]; entry 1160 drivers/s390/net/qeth_l2_main.c struct qeth_sbp_port_entry *entry = &data->qports.entry[0]; entry 1174 drivers/s390/net/qeth_l2_main.c data->card->options.sbp.role = entry->role; entry 1179 drivers/s390/net/qeth_l2_main.c (entry->role == QETH_SBP_ROLE_NONE) ? "none" : entry 1180 drivers/s390/net/qeth_l2_main.c (entry->role == QETH_SBP_ROLE_PRIMARY) ? "primary" : entry 1181 drivers/s390/net/qeth_l2_main.c (entry->role == QETH_SBP_ROLE_SECONDARY) ? "secondary" : entry 1184 drivers/s390/net/qeth_l2_main.c (entry->state == QETH_SBP_STATE_INACTIVE) ? "inactive" : entry 1185 drivers/s390/net/qeth_l2_main.c (entry->state == QETH_SBP_STATE_STANDBY) ? "standby" : entry 1186 drivers/s390/net/qeth_l2_main.c (entry->state == QETH_SBP_STATE_ACTIVE) ? "active" : entry 1248 drivers/s390/net/qeth_l2_main.c struct qeth_ipacmd_addr_change_entry *entry = entry 1249 drivers/s390/net/qeth_l2_main.c &data->hostevs.entry[i]; entry 1252 drivers/s390/net/qeth_l2_main.c entry->change_code, entry 1253 drivers/s390/net/qeth_l2_main.c &entry->token, &entry->addr_lnid); entry 1490 drivers/s390/net/qeth_l2_main.c *cbctl->data.qports.role = qports->entry[0].role; entry 1492 drivers/s390/net/qeth_l2_main.c *cbctl->data.qports.state = qports->entry[0].state; entry 1616 drivers/s390/net/qeth_l2_main.c enum qdio_brinfo_entry_type type, void *entry) entry 1627 drivers/s390/net/qeth_l2_main.c l2entry = (struct qdio_brinfo_entry_l2 *)entry; entry 110 drivers/s390/net/qeth_l3.h struct list_head entry; entry 123 drivers/s390/net/qeth_l3_main.c list_for_each_entry(ipatoe, &card->ipato.entries, entry) { entry 579 drivers/s390/net/qeth_l3_main.c list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) { entry 580 drivers/s390/net/qeth_l3_main.c list_del(&ipatoe->entry); entry 598 drivers/s390/net/qeth_l3_main.c list_for_each_entry(ipatoe, &card->ipato.entries, entry) { entry 610 drivers/s390/net/qeth_l3_main.c list_add_tail(&new->entry, &card->ipato.entries); entry 630 drivers/s390/net/qeth_l3_main.c list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) { entry 636 drivers/s390/net/qeth_l3_main.c list_del(&ipatoe->entry); entry 1755 drivers/s390/net/qeth_l3_main.c struct qeth_arp_cache_entry *entry, entry 1785 drivers/s390/net/qeth_l3_main.c ether_addr_copy(cmd_entry->macaddr, entry->macaddr); entry 1786 drivers/s390/net/qeth_l3_main.c memcpy(cmd_entry->ipaddr, entry->ipaddr, 4); entry 456 drivers/s390/net/qeth_l3_sys.c list_for_each_entry(ipatoe, &card->ipato.entries, entry) { entry 285 drivers/s390/scsi/zfcp_dbf.c struct list_head *entry; entry 300 drivers/s390/scsi/zfcp_dbf.c list_for_each(entry, &adapter->erp_ready_head) entry 303 drivers/s390/scsi/zfcp_dbf.c list_for_each(entry, &adapter->erp_running_head) entry 298 drivers/sbus/char/bbc_i2c.c int entry; entry 328 drivers/sbus/char/bbc_i2c.c entry = 0; entry 330 drivers/sbus/char/bbc_i2c.c dp && entry < 8; entry 331 drivers/sbus/char/bbc_i2c.c dp = dp->sibling, entry++) { entry 335 drivers/sbus/char/bbc_i2c.c bp->devs[entry].device = child_op; entry 336 drivers/sbus/char/bbc_i2c.c bp->devs[entry].client_claimed = 0; entry 345 drivers/sbus/char/bbc_i2c.c bp->index, bp->i2c_control_regs, entry, bp->own, bp->clock); entry 1529 drivers/scsi/aacraid/aacraid.h struct list_head entry; entry 2689 drivers/scsi/aacraid/aacraid.h int aac_consumer_get(struct aac_dev * dev, struct aac_queue * q, struct aac_entry **entry); entry 176 drivers/scsi/aacraid/commctrl.c struct list_head * entry; entry 206 drivers/scsi/aacraid/commctrl.c entry = dev->fib_list.next; entry 207 drivers/scsi/aacraid/commctrl.c while (entry != &dev->fib_list) { entry 208 drivers/scsi/aacraid/commctrl.c context = list_entry(entry, struct aac_fib_context, next); entry 212 drivers/scsi/aacraid/commctrl.c entry = dev->fib_list.next; entry 214 drivers/scsi/aacraid/commctrl.c entry = entry->next; entry 244 drivers/scsi/aacraid/commctrl.c struct list_head * entry; entry 256 drivers/scsi/aacraid/commctrl.c entry = dev->fib_list.next; entry 259 drivers/scsi/aacraid/commctrl.c while (entry != &dev->fib_list) { entry 260 drivers/scsi/aacraid/commctrl.c fibctx = list_entry(entry, struct aac_fib_context, next); entry 267 drivers/scsi/aacraid/commctrl.c entry = entry->next; entry 292 drivers/scsi/aacraid/commctrl.c entry = fibctx->fib_list.next; entry 293 drivers/scsi/aacraid/commctrl.c list_del(entry); entry 295 drivers/scsi/aacraid/commctrl.c fib = list_entry(entry, struct fib, fiblink); entry 346 drivers/scsi/aacraid/commctrl.c struct list_head * entry; entry 350 drivers/scsi/aacraid/commctrl.c entry = fibctx->fib_list.next; entry 351 drivers/scsi/aacraid/commctrl.c list_del(entry); entry 352 drivers/scsi/aacraid/commctrl.c fib = list_entry(entry, struct fib, fiblink); entry 388 drivers/scsi/aacraid/commctrl.c struct list_head * entry; entry 397 drivers/scsi/aacraid/commctrl.c entry = dev->fib_list.next; entry 400 drivers/scsi/aacraid/commctrl.c while(entry != &dev->fib_list) { entry 401 drivers/scsi/aacraid/commctrl.c fibctx = list_entry(entry, struct aac_fib_context, next); entry 407 drivers/scsi/aacraid/commctrl.c entry = entry->next; entry 357 drivers/scsi/aacraid/commsup.c static int aac_get_entry (struct aac_dev * dev, u32 qid, struct aac_entry **entry, u32 * index, unsigned long *nonotify) entry 398 drivers/scsi/aacraid/commsup.c *entry = q->base + *index; entry 421 drivers/scsi/aacraid/commsup.c struct aac_entry * entry = NULL; entry 426 drivers/scsi/aacraid/commsup.c while (!aac_get_entry(dev, qid, &entry, index, nonotify)) { entry 432 drivers/scsi/aacraid/commsup.c entry->size = cpu_to_le32(le16_to_cpu(hw_fib->header.Size)); entry 435 drivers/scsi/aacraid/commsup.c while (!aac_get_entry(dev, qid, &entry, index, nonotify)) { entry 441 drivers/scsi/aacraid/commsup.c entry->size = cpu_to_le32(le16_to_cpu(hw_fib->header.Size)); entry 442 drivers/scsi/aacraid/commsup.c entry->addr = hw_fib->header.SenderFibAddress; entry 452 drivers/scsi/aacraid/commsup.c entry->addr = cpu_to_le32(fibptr->hw_fib_pa); entry 798 drivers/scsi/aacraid/commsup.c int aac_consumer_get(struct aac_dev * dev, struct aac_queue * q, struct aac_entry **entry) entry 814 drivers/scsi/aacraid/commsup.c *entry = q->base + index; entry 1725 drivers/scsi/aacraid/commsup.c struct list_head * entry; entry 1748 drivers/scsi/aacraid/commsup.c entry = aac->fib_list.next; entry 1756 drivers/scsi/aacraid/commsup.c while (entry != &aac->fib_list) { entry 1760 drivers/scsi/aacraid/commsup.c struct aac_fib_context *fibctx = list_entry(entry, struct aac_fib_context, next); entry 1780 drivers/scsi/aacraid/commsup.c entry = entry->next; entry 1824 drivers/scsi/aacraid/commsup.c entry = entry->next; entry 2017 drivers/scsi/aacraid/commsup.c struct list_head *entry; entry 2029 drivers/scsi/aacraid/commsup.c entry = dev->fib_list.next; entry 2030 drivers/scsi/aacraid/commsup.c while (entry != &dev->fib_list) { entry 2031 drivers/scsi/aacraid/commsup.c entry = entry->next; entry 2077 drivers/scsi/aacraid/commsup.c struct list_head *entry; entry 2087 drivers/scsi/aacraid/commsup.c entry = dev->fib_list.next; entry 2097 drivers/scsi/aacraid/commsup.c while (entry != &dev->fib_list) { entry 2101 drivers/scsi/aacraid/commsup.c fibctx = list_entry(entry, struct aac_fib_context, entry 2120 drivers/scsi/aacraid/commsup.c entry = entry->next; entry 2131 drivers/scsi/aacraid/commsup.c entry = entry->next; entry 2157 drivers/scsi/aacraid/commsup.c entry = entry->next; entry 2179 drivers/scsi/aacraid/commsup.c struct list_head *entry; entry 2187 drivers/scsi/aacraid/commsup.c entry = dev->queues->queue[HostNormCmdQueue].cmdq.next; entry 2188 drivers/scsi/aacraid/commsup.c list_del(entry); entry 2193 drivers/scsi/aacraid/commsup.c fib = list_entry(entry, struct fib, fiblink); entry 42 drivers/scsi/aacraid/dpcsup.c struct aac_entry *entry; entry 55 drivers/scsi/aacraid/dpcsup.c while(aac_consumer_get(dev, q, &entry)) entry 58 drivers/scsi/aacraid/dpcsup.c u32 index = le32_to_cpu(entry->addr); entry 160 drivers/scsi/aacraid/dpcsup.c struct aac_entry *entry; entry 170 drivers/scsi/aacraid/dpcsup.c while(aac_consumer_get(dev, q, &entry)) entry 177 drivers/scsi/aacraid/dpcsup.c index = le32_to_cpu(entry->addr) / sizeof(struct hw_fib); entry 333 drivers/scsi/aacraid/linit.c struct partition *entry = first; entry 339 drivers/scsi/aacraid/linit.c end_head = entry->end_head; entry 340 drivers/scsi/aacraid/linit.c end_sec = entry->end_sector & 0x3f; entry 355 drivers/scsi/aacraid/linit.c entry++; entry 1121 drivers/scsi/aacraid/linit.c list_for_each_entry(aac, &aac_devices, entry) { entry 1606 drivers/scsi/aacraid/linit.c list_for_each_entry(aac, &aac_devices, entry) { entry 1609 drivers/scsi/aacraid/linit.c insert = &aac->entry; entry 1669 drivers/scsi/aacraid/linit.c INIT_LIST_HEAD(&aac->entry); entry 1765 drivers/scsi/aacraid/linit.c list_add(&aac->entry, insert); entry 1948 drivers/scsi/aacraid/linit.c list_del(&aac->entry); entry 79 drivers/scsi/aacraid/src.c struct list_head *entry; entry 101 drivers/scsi/aacraid/src.c entry = dev->sync_fib_list.next; entry 102 drivers/scsi/aacraid/src.c dev->sync_fib = list_entry(entry, entry 105 drivers/scsi/aacraid/src.c list_del(entry); entry 107 drivers/scsi/aic7xxx/aic7770.c struct aic7770_identity *entry; entry 111 drivers/scsi/aic7xxx/aic7770.c entry = &aic7770_ident_table[i]; entry 112 drivers/scsi/aic7xxx/aic7770.c if (entry->full_id == (id & entry->id_mask)) entry 113 drivers/scsi/aic7xxx/aic7770.c return (entry); entry 119 drivers/scsi/aic7xxx/aic7770.c aic7770_config(struct ahc_softc *ahc, struct aic7770_identity *entry, u_int io) entry 127 drivers/scsi/aic7xxx/aic7770.c error = entry->setup(ahc); entry 144 drivers/scsi/aic7xxx/aic7770.c ahc->description = entry->name; entry 4341 drivers/scsi/aic7xxx/aic79xx_core.c const struct ahd_phase_table_entry *entry; entry 4349 drivers/scsi/aic7xxx/aic79xx_core.c for (entry = ahd_phase_table; entry < last_entry; entry++) { entry 4350 drivers/scsi/aic7xxx/aic79xx_core.c if (phase == entry->phase) entry 4353 drivers/scsi/aic7xxx/aic79xx_core.c return (entry); entry 9648 drivers/scsi/aic7xxx/aic79xx_core.c int entry; entry 9650 drivers/scsi/aic7xxx/aic79xx_core.c for (entry = 0; entry < num_entries; entry++) { entry 9651 drivers/scsi/aic7xxx/aic79xx_core.c if (((value & table[entry].mask) entry 9652 drivers/scsi/aic7xxx/aic79xx_core.c != table[entry].value) entry 9653 drivers/scsi/aic7xxx/aic79xx_core.c || ((printed_mask & table[entry].mask) entry 9654 drivers/scsi/aic7xxx/aic79xx_core.c == table[entry].mask)) entry 9659 drivers/scsi/aic7xxx/aic79xx_core.c table[entry].name); entry 9660 drivers/scsi/aic7xxx/aic79xx_core.c printed_mask |= table[entry].mask; entry 9664 drivers/scsi/aic7xxx/aic79xx_core.c if (entry >= num_entries) entry 162 drivers/scsi/aic7xxx/aic79xx_osm_pci.c const struct ahd_pci_identity *entry; entry 168 drivers/scsi/aic7xxx/aic79xx_osm_pci.c entry = ahd_find_pci_device(pci); entry 169 drivers/scsi/aic7xxx/aic79xx_osm_pci.c if (entry == NULL) entry 208 drivers/scsi/aic7xxx/aic79xx_osm_pci.c error = ahd_pci_config(ahd, entry); entry 258 drivers/scsi/aic7xxx/aic79xx_pci.c const struct ahd_pci_identity *entry; entry 277 drivers/scsi/aic7xxx/aic79xx_pci.c entry = &ahd_pci_ident_table[i]; entry 278 drivers/scsi/aic7xxx/aic79xx_pci.c if (entry->full_id == (full_id & entry->id_mask)) { entry 280 drivers/scsi/aic7xxx/aic79xx_pci.c if (entry->name == NULL) entry 282 drivers/scsi/aic7xxx/aic79xx_pci.c return (entry); entry 289 drivers/scsi/aic7xxx/aic79xx_pci.c ahd_pci_config(struct ahd_softc *ahd, const struct ahd_pci_identity *entry) entry 296 drivers/scsi/aic7xxx/aic79xx_pci.c ahd->description = entry->name; entry 305 drivers/scsi/aic7xxx/aic79xx_pci.c error = entry->setup(ahd); entry 2797 drivers/scsi/aic7xxx/aic7xxx_core.c const struct ahc_phase_table_entry *entry; entry 2805 drivers/scsi/aic7xxx/aic7xxx_core.c for (entry = ahc_phase_table; entry < last_entry; entry++) { entry 2806 drivers/scsi/aic7xxx/aic7xxx_core.c if (phase == entry->phase) entry 2809 drivers/scsi/aic7xxx/aic7xxx_core.c return (entry); entry 7106 drivers/scsi/aic7xxx/aic7xxx_core.c int entry; entry 7108 drivers/scsi/aic7xxx/aic7xxx_core.c for (entry = 0; entry < num_entries; entry++) { entry 7109 drivers/scsi/aic7xxx/aic7xxx_core.c if (((value & table[entry].mask) entry 7110 drivers/scsi/aic7xxx/aic7xxx_core.c != table[entry].value) entry 7111 drivers/scsi/aic7xxx/aic7xxx_core.c || ((printed_mask & table[entry].mask) entry 7112 drivers/scsi/aic7xxx/aic7xxx_core.c == table[entry].mask)) entry 7117 drivers/scsi/aic7xxx/aic7xxx_core.c table[entry].name); entry 7118 drivers/scsi/aic7xxx/aic7xxx_core.c printed_mask |= table[entry].mask; entry 7122 drivers/scsi/aic7xxx/aic7xxx_core.c if (entry >= num_entries) entry 209 drivers/scsi/aic7xxx/aic7xxx_osm_pci.c const struct ahc_pci_identity *entry; entry 215 drivers/scsi/aic7xxx/aic7xxx_osm_pci.c entry = ahc_find_pci_device(pci); entry 216 drivers/scsi/aic7xxx/aic7xxx_osm_pci.c if (entry == NULL) entry 254 drivers/scsi/aic7xxx/aic7xxx_osm_pci.c error = ahc_pci_config(ahc, entry); entry 671 drivers/scsi/aic7xxx/aic7xxx_pci.c const struct ahc_pci_identity *entry; entry 693 drivers/scsi/aic7xxx/aic7xxx_pci.c entry = &ahc_pci_ident_table[i]; entry 694 drivers/scsi/aic7xxx/aic7xxx_pci.c if (entry->full_id == (full_id & entry->id_mask)) { entry 696 drivers/scsi/aic7xxx/aic7xxx_pci.c if (entry->name == NULL) entry 698 drivers/scsi/aic7xxx/aic7xxx_pci.c return (entry); entry 705 drivers/scsi/aic7xxx/aic7xxx_pci.c ahc_pci_config(struct ahc_softc *ahc, const struct ahc_pci_identity *entry) entry 717 drivers/scsi/aic7xxx/aic7xxx_pci.c error = entry->setup(ahc); entry 721 drivers/scsi/aic7xxx/aic7xxx_pci.c ahc->description = entry->name; entry 34 drivers/scsi/aic94xx/aic94xx_sds.c struct asd_ocm_dir_ent entry[15]; entry 183 drivers/scsi/aic94xx/aic94xx_sds.c if (dir->entry[i].type == type) entry 188 drivers/scsi/aic94xx/aic94xx_sds.c ent = &dir->entry[i]; entry 1126 drivers/scsi/bfa/bfad.c bfa_msix(&bfad->bfa, vec->msix.entry); entry 1153 drivers/scsi/bfa/bfad.c bfad->msix_tab[bfad->nvec].msix.entry = i; entry 1155 drivers/scsi/bfa/bfad.c msix_entries[bfad->nvec].entry = i; entry 135 drivers/scsi/bfa/bfad_im.h static inline void bfad_im_post_vendor_event(struct bfa_aen_entry_s *entry, entry 148 drivers/scsi/bfa/bfad_im.h entry->aen_tv_sec = ts.tv_sec; entry 149 drivers/scsi/bfa/bfad_im.h entry->aen_tv_usec = ts.tv_nsec / NSEC_PER_USEC; entry 150 drivers/scsi/bfa/bfad_im.h entry->bfad_num = drv->inst_no; entry 151 drivers/scsi/bfa/bfad_im.h entry->seq_num = cnt; entry 152 drivers/scsi/bfa/bfad_im.h entry->aen_category = cat; entry 153 drivers/scsi/bfa/bfad_im.h entry->aen_type = evt; entry 1381 drivers/scsi/cxlflash/main.c u64 entry, entry 1388 drivers/scsi/cxlflash/main.c entry = *hrrq_curr; entry 1390 drivers/scsi/cxlflash/main.c if ((entry & SISL_RESP_HANDLE_T_BIT) != toggle) entry 1393 drivers/scsi/cxlflash/main.c entry &= ~SISL_RESP_HANDLE_T_BIT; entry 1396 drivers/scsi/cxlflash/main.c ioasa = (struct sisl_ioasa *)entry; entry 1399 drivers/scsi/cxlflash/main.c ioarcb = (struct sisl_ioarcb *)entry; entry 94 drivers/scsi/device_handler/scsi_dh_alua.c struct list_head entry; entry 866 drivers/scsi/device_handler/scsi_dh_alua.c list_for_each_entry_safe(qdata, tmp, &qdata_list, entry) { entry 867 drivers/scsi/device_handler/scsi_dh_alua.c list_del(&qdata->entry); entry 901 drivers/scsi/device_handler/scsi_dh_alua.c list_add_tail(&qdata->entry, &pg->rtpg_list); entry 228 drivers/scsi/device_handler/scsi_dh_rdac.c struct list_head entry; entry 300 drivers/scsi/device_handler/scsi_dh_rdac.c list_for_each_entry(qdata, list, entry) { entry 574 drivers/scsi/device_handler/scsi_dh_rdac.c list_for_each_entry_safe(qdata, tmp, &list, entry) { entry 575 drivers/scsi/device_handler/scsi_dh_rdac.c list_del(&qdata->entry); entry 601 drivers/scsi/device_handler/scsi_dh_rdac.c list_add_tail(&qdata->entry, &ctlr->ms_head); entry 1795 drivers/scsi/fcoe/fcoe.c struct dcb_app_type *entry = ptr; entry 1801 drivers/scsi/fcoe/fcoe.c if (entry->app.selector != DCB_APP_IDTYPE_ETHTYPE) entry 1804 drivers/scsi/fcoe/fcoe.c netdev = dev_get_by_index(&init_net, entry->ifindex); entry 1815 drivers/scsi/fcoe/fcoe.c if (entry->dcbx & DCB_CAP_DCBX_VER_CEE) entry 1816 drivers/scsi/fcoe/fcoe.c prio = ffs(entry->app.priority) - 1; entry 1818 drivers/scsi/fcoe/fcoe.c prio = entry->app.priority; entry 1823 drivers/scsi/fcoe/fcoe.c if (entry->app.protocol == ETH_P_FIP || entry 1824 drivers/scsi/fcoe/fcoe.c entry->app.protocol == ETH_P_FCOE) entry 1827 drivers/scsi/fcoe/fcoe.c if (entry->app.protocol == ETH_P_FCOE) entry 652 drivers/scsi/gdth.h gdth_hentry_str entry[MAX_HDRIVES]; /* entries */ entry 491 drivers/scsi/gdth_proc.c phg->offset = GDTOFFSOF(gdth_hget_str, entry[0]); entry 498 drivers/scsi/gdth_proc.c k = phg->entry[j].host_drive; entry 501 drivers/scsi/gdth_proc.c ha->hdr[k].ldr_no = phg->entry[j].log_drive; entry 502 drivers/scsi/gdth_proc.c ha->hdr[k].rw_attribs = phg->entry[j].rw_attribs; entry 503 drivers/scsi/gdth_proc.c ha->hdr[k].start_sec = phg->entry[j].start_sec; entry 215 drivers/scsi/hisi_sas/hisi_sas.h struct list_head entry; entry 251 drivers/scsi/hisi_sas/hisi_sas_main.c list_del_init(&slot->entry); entry 498 drivers/scsi/hisi_sas/hisi_sas_main.c list_add_tail(&slot->entry, &sas_dev->list); entry 1020 drivers/scsi/hisi_sas/hisi_sas_main.c list_for_each_entry_safe(slot, slot2, &sas_dev->list, entry) entry 1970 drivers/scsi/hisi_sas/hisi_sas_main.c list_add_tail(&slot->entry, &sas_dev->list); entry 898 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c struct hisi_sas_sge *entry = &sge_page->sge[i]; entry 900 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c entry->addr = cpu_to_le64(sg_dma_address(sg)); entry 901 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c entry->page_ctrl_0 = entry->page_ctrl_1 = 0; entry 902 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c entry->data_len = cpu_to_le32(sg_dma_len(sg)); entry 903 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c entry->data_off = 0; entry 1678 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c struct hisi_sas_sge *entry = &sge_page->sge[i]; entry 1680 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c entry->addr = cpu_to_le64(sg_dma_address(sg)); entry 1681 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c entry->page_ctrl_0 = entry->page_ctrl_1 = 0; entry 1682 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c entry->data_len = cpu_to_le32(sg_dma_len(sg)); entry 1683 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c entry->data_off = 0; entry 830 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c list_for_each_entry_safe(slot, slot2, &sas_dev->list, entry) { entry 1059 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c struct hisi_sas_sge *entry = &sge_page->sge[i]; entry 1061 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->addr = cpu_to_le64(sg_dma_address(sg)); entry 1062 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->page_ctrl_0 = entry->page_ctrl_1 = 0; entry 1063 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->data_len = cpu_to_le32(sg_dma_len(sg)); entry 1064 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->data_off = 0; entry 1085 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c struct hisi_sas_sge *entry = &sge_dif_page->sge[i]; entry 1087 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->addr = cpu_to_le64(sg_dma_address(sg)); entry 1088 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->page_ctrl_0 = 0; entry 1089 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->page_ctrl_1 = 0; entry 1090 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->data_len = cpu_to_le32(sg_dma_len(sg)); entry 1091 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c entry->data_off = 0; entry 1346 drivers/scsi/hpsa.c int entry, struct hpsa_scsi_dev_t *new_entry) entry 1349 drivers/scsi/hpsa.c BUG_ON(entry < 0 || entry >= HPSA_MAX_DEVICES); entry 1352 drivers/scsi/hpsa.c h->dev[entry]->raid_level = new_entry->raid_level; entry 1357 drivers/scsi/hpsa.c h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; entry 1369 drivers/scsi/hpsa.c h->dev[entry]->raid_map = new_entry->raid_map; entry 1370 drivers/scsi/hpsa.c h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; entry 1373 drivers/scsi/hpsa.c h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; entry 1376 drivers/scsi/hpsa.c h->dev[entry]->hba_ioaccel_enabled = new_entry->hba_ioaccel_enabled; entry 1377 drivers/scsi/hpsa.c h->dev[entry]->offload_config = new_entry->offload_config; entry 1378 drivers/scsi/hpsa.c h->dev[entry]->offload_to_mirror = new_entry->offload_to_mirror; entry 1379 drivers/scsi/hpsa.c h->dev[entry]->queue_depth = new_entry->queue_depth; entry 1386 drivers/scsi/hpsa.c h->dev[entry]->offload_to_be_enabled = new_entry->offload_to_be_enabled; entry 1392 drivers/scsi/hpsa.c h->dev[entry]->offload_enabled = 0; entry 1394 drivers/scsi/hpsa.c hpsa_show_dev_msg(KERN_INFO, h, h->dev[entry], "updated"); entry 1399 drivers/scsi/hpsa.c int entry, struct hpsa_scsi_dev_t *new_entry, entry 1404 drivers/scsi/hpsa.c BUG_ON(entry < 0 || entry >= HPSA_MAX_DEVICES); entry 1405 drivers/scsi/hpsa.c removed[*nremoved] = h->dev[entry]; entry 1413 drivers/scsi/hpsa.c new_entry->target = h->dev[entry]->target; entry 1414 drivers/scsi/hpsa.c new_entry->lun = h->dev[entry]->lun; entry 1417 drivers/scsi/hpsa.c h->dev[entry] = new_entry; entry 1425 drivers/scsi/hpsa.c static void hpsa_scsi_remove_entry(struct ctlr_info *h, int entry, entry 1432 drivers/scsi/hpsa.c BUG_ON(entry < 0 || entry >= HPSA_MAX_DEVICES); entry 1434 drivers/scsi/hpsa.c sd = h->dev[entry]; entry 1435 drivers/scsi/hpsa.c removed[*nremoved] = h->dev[entry]; entry 1438 drivers/scsi/hpsa.c for (i = entry; i < h->ndevices-1; i++) entry 1911 drivers/scsi/hpsa.c int i, entry, device_change, changes = 0; entry 1952 drivers/scsi/hpsa.c device_change = hpsa_scsi_find_entry(csd, sd, nsds, &entry); entry 1959 drivers/scsi/hpsa.c hpsa_scsi_replace_entry(h, i, sd[entry], entry 1964 drivers/scsi/hpsa.c sd[entry] = NULL; entry 1966 drivers/scsi/hpsa.c hpsa_scsi_update_entry(h, i, sd[entry]); entry 1991 drivers/scsi/hpsa.c h->ndevices, &entry); entry 150 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_trace_entry *entry; entry 152 drivers/scsi/ibmvscsi/ibmvfc.c entry = &vhost->trace[vhost->trace_index++]; entry 153 drivers/scsi/ibmvscsi/ibmvfc.c entry->evt = evt; entry 154 drivers/scsi/ibmvscsi/ibmvfc.c entry->time = jiffies; entry 155 drivers/scsi/ibmvscsi/ibmvfc.c entry->fmt = evt->crq.format; entry 156 drivers/scsi/ibmvscsi/ibmvfc.c entry->type = IBMVFC_TRC_START; entry 158 drivers/scsi/ibmvscsi/ibmvfc.c switch (entry->fmt) { entry 160 drivers/scsi/ibmvscsi/ibmvfc.c entry->op_code = vfc_cmd->iu.cdb[0]; entry 161 drivers/scsi/ibmvscsi/ibmvfc.c entry->scsi_id = be64_to_cpu(vfc_cmd->tgt_scsi_id); entry 162 drivers/scsi/ibmvscsi/ibmvfc.c entry->lun = scsilun_to_int(&vfc_cmd->iu.lun); entry 163 drivers/scsi/ibmvscsi/ibmvfc.c entry->tmf_flags = vfc_cmd->iu.tmf_flags; entry 164 drivers/scsi/ibmvscsi/ibmvfc.c entry->u.start.xfer_len = be32_to_cpu(vfc_cmd->iu.xfer_len); entry 167 drivers/scsi/ibmvscsi/ibmvfc.c entry->op_code = be32_to_cpu(mad->opcode); entry 184 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_trace_entry *entry = &vhost->trace[vhost->trace_index++]; entry 186 drivers/scsi/ibmvscsi/ibmvfc.c entry->evt = evt; entry 187 drivers/scsi/ibmvscsi/ibmvfc.c entry->time = jiffies; entry 188 drivers/scsi/ibmvscsi/ibmvfc.c entry->fmt = evt->crq.format; entry 189 drivers/scsi/ibmvscsi/ibmvfc.c entry->type = IBMVFC_TRC_END; entry 191 drivers/scsi/ibmvscsi/ibmvfc.c switch (entry->fmt) { entry 193 drivers/scsi/ibmvscsi/ibmvfc.c entry->op_code = vfc_cmd->iu.cdb[0]; entry 194 drivers/scsi/ibmvscsi/ibmvfc.c entry->scsi_id = be64_to_cpu(vfc_cmd->tgt_scsi_id); entry 195 drivers/scsi/ibmvscsi/ibmvfc.c entry->lun = scsilun_to_int(&vfc_cmd->iu.lun); entry 196 drivers/scsi/ibmvscsi/ibmvfc.c entry->tmf_flags = vfc_cmd->iu.tmf_flags; entry 197 drivers/scsi/ibmvscsi/ibmvfc.c entry->u.end.status = be16_to_cpu(vfc_cmd->status); entry 198 drivers/scsi/ibmvscsi/ibmvfc.c entry->u.end.error = be16_to_cpu(vfc_cmd->error); entry 199 drivers/scsi/ibmvscsi/ibmvfc.c entry->u.end.fcp_rsp_flags = vfc_cmd->rsp.flags; entry 200 drivers/scsi/ibmvscsi/ibmvfc.c entry->u.end.rsp_code = vfc_cmd->rsp.data.info.rsp_code; entry 201 drivers/scsi/ibmvscsi/ibmvfc.c entry->u.end.scsi_status = vfc_cmd->rsp.scsi_status; entry 204 drivers/scsi/ibmvscsi/ibmvfc.c entry->op_code = be32_to_cpu(mad->opcode); entry 205 drivers/scsi/ibmvscsi/ibmvfc.c entry->u.end.status = be16_to_cpu(mad->status); entry 3213 drivers/scsi/ipr.c bytes_to_copy = offsetof(struct ipr_sdt, entry) + entry 3250 drivers/scsi/ipr.c if (sdt->entry[i].flags & IPR_SDT_VALID_ENTRY) { entry 3251 drivers/scsi/ipr.c sdt_word = be32_to_cpu(sdt->entry[i].start_token); entry 3253 drivers/scsi/ipr.c bytes_to_copy = be32_to_cpu(sdt->entry[i].end_token); entry 3256 drivers/scsi/ipr.c end_off = be32_to_cpu(sdt->entry[i].end_token); entry 3265 drivers/scsi/ipr.c sdt->entry[i].flags &= ~IPR_SDT_VALID_ENTRY; entry 4303 drivers/scsi/ipr.c sdt_end = offsetof(struct ipr_ioa_dump, sdt.entry) + entry 4307 drivers/scsi/ipr.c sdt_end = offsetof(struct ipr_ioa_dump, sdt.entry) + entry 8548 drivers/scsi/ipr.c if (rc || !(sdt.entry[0].flags & IPR_SDT_VALID_ENTRY) || entry 8557 drivers/scsi/ipr.c length = be32_to_cpu(sdt.entry[0].end_token); entry 8559 drivers/scsi/ipr.c length = (be32_to_cpu(sdt.entry[0].end_token) - entry 8560 drivers/scsi/ipr.c be32_to_cpu(sdt.entry[0].start_token)) & entry 8569 drivers/scsi/ipr.c be32_to_cpu(sdt.entry[0].start_token), entry 1262 drivers/scsi/ipr.h struct ipr_sdt_entry entry[IPR_FMT3_NUM_SDT_ENTRIES]; entry 1267 drivers/scsi/ipr.h struct ipr_sdt_entry entry[1]; entry 28 drivers/scsi/lpfc/lpfc_scsi.h #define list_remove_head(list, entry, type, member) \ entry 30 drivers/scsi/lpfc/lpfc_scsi.h entry = NULL; \ entry 32 drivers/scsi/lpfc/lpfc_scsi.h entry = list_entry((list)->next, type, member); \ entry 33 drivers/scsi/lpfc/lpfc_scsi.h list_del_init(&entry->member); \ entry 3352 drivers/scsi/lpfc/lpfc_sli.c IOCB_t *entry = NULL; entry 3389 drivers/scsi/lpfc/lpfc_sli.c entry = lpfc_resp_iocb(phba, pring); entry 3395 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli_pcimem_bcopy((uint32_t *) entry, entry 3736 drivers/scsi/lpfc/lpfc_sli.c IOCB_t *entry; entry 3787 drivers/scsi/lpfc/lpfc_sli.c entry = lpfc_resp_iocb(phba, pring); entry 3797 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli_pcimem_bcopy(entry, &rspiocbp->iocb, entry 13021 drivers/scsi/lpfc/lpfc_sli.c lpfc_cq_event_setup(struct lpfc_hba *phba, void *entry, int size) entry 13034 drivers/scsi/lpfc/lpfc_sli.c memcpy(&cq_event->cqe, entry, size); entry 139 drivers/scsi/mvsas/mv_init.c list_for_each_entry(mwq, &mvi->wq_list, entry) entry 814 drivers/scsi/mvsas/mv_sas.c list_add_tail(&slot->entry, &tei.port->list); entry 907 drivers/scsi/mvsas/mv_sas.c list_del_init(&slot->entry); entry 1844 drivers/scsi/mvsas/mv_sas.c list_for_each_entry_safe(slot, slot2, &port->list, entry) { entry 1915 drivers/scsi/mvsas/mv_sas.c list_del(&mwq->entry); entry 1931 drivers/scsi/mvsas/mv_sas.c list_add_tail(&mwq->entry, &mvi->wq_list); entry 307 drivers/scsi/mvsas/mv_sas.h struct list_head entry; entry 414 drivers/scsi/mvsas/mv_sas.h struct list_head entry; entry 150 drivers/scsi/mvumi.c INIT_LIST_HEAD(&res->entry); entry 151 drivers/scsi/mvumi.c list_add_tail(&res->entry, &mhba->res_list); entry 160 drivers/scsi/mvumi.c list_for_each_entry_safe(res, tmp, &mhba->res_list, entry) { entry 174 drivers/scsi/mvumi.c list_del(&res->entry); entry 212 drivers/scsi/mvumi.h struct list_head entry; entry 51 drivers/scsi/myrb.c struct myrb_devstate_name_entry *entry = myrb_devstate_name_list; entry 55 drivers/scsi/myrb.c if (entry[i].state == state) entry 56 drivers/scsi/myrb.c return entry[i].name; entry 75 drivers/scsi/myrb.c struct myrb_raidlevel_name_entry *entry = myrb_raidlevel_name_list; entry 79 drivers/scsi/myrb.c if (entry[i].level == level) entry 80 drivers/scsi/myrb.c return entry[i].name; entry 3499 drivers/scsi/myrb.c const struct pci_device_id *entry) entry 3502 drivers/scsi/myrb.c (struct myrb_privdata *)entry->driver_data; entry 3560 drivers/scsi/myrb.c static int myrb_probe(struct pci_dev *dev, const struct pci_device_id *entry) entry 3565 drivers/scsi/myrb.c cb = myrb_detect(dev, entry); entry 49 drivers/scsi/myrs.c struct myrs_devstate_name_entry *entry = myrs_devstate_name_list; entry 53 drivers/scsi/myrs.c if (entry[i].state == state) entry 54 drivers/scsi/myrs.c return entry[i].name; entry 80 drivers/scsi/myrs.c struct myrs_raid_level_name_entry *entry = myrs_raid_level_name_list; entry 84 drivers/scsi/myrs.c if (entry[i].level == level) entry 85 drivers/scsi/myrs.c return entry[i].name; entry 1938 drivers/scsi/myrs.c const struct pci_device_id *entry) entry 2289 drivers/scsi/myrs.c const struct pci_device_id *entry) entry 2292 drivers/scsi/myrs.c (struct myrs_privdata *)entry->driver_data; entry 2297 drivers/scsi/myrs.c cs = myrs_alloc_host(pdev, entry); entry 3172 drivers/scsi/myrs.c myrs_probe(struct pci_dev *dev, const struct pci_device_id *entry) entry 3177 drivers/scsi/myrs.c cs = myrs_detect(dev, entry); entry 2225 drivers/scsi/nsp32.c int entry; entry 2287 drivers/scsi/nsp32.c entry = nsp32_search_period_entry(data, target, get_period); entry 2289 drivers/scsi/nsp32.c if (entry < 0) { entry 2300 drivers/scsi/nsp32.c nsp32_set_sync_entry(data, target, entry, get_offset); entry 2318 drivers/scsi/nsp32.c entry = nsp32_search_period_entry(data, target, get_period); entry 2320 drivers/scsi/nsp32.c if (get_offset == ASYNC_OFFSET || entry < 0) { entry 2324 drivers/scsi/nsp32.c nsp32_set_sync_entry(data, target, entry, get_offset); entry 2427 drivers/scsi/nsp32.c int entry, entry 2432 drivers/scsi/nsp32.c period = data->synct[entry].period_num; entry 2433 drivers/scsi/nsp32.c ackwidth = data->synct[entry].ackwidth; entry 2434 drivers/scsi/nsp32.c sample_rate = data->synct[entry].sample_rate; entry 2987 drivers/scsi/nsp32.c int entry; entry 3036 drivers/scsi/nsp32.c entry = nsp32_search_period_entry(data, target, ret); entry 3037 drivers/scsi/nsp32.c if (entry < 0) { entry 3039 drivers/scsi/nsp32.c entry = 0; entry 3041 drivers/scsi/nsp32.c target->limit_entry = entry; entry 3073 drivers/scsi/nsp32.c int entry, val; entry 3105 drivers/scsi/nsp32.c entry = nsp32_search_period_entry(data, target, val); entry 3106 drivers/scsi/nsp32.c if (entry < 0 || trans_mode == ULTRA20M_MODE) { entry 3108 drivers/scsi/nsp32.c entry = 0; entry 3110 drivers/scsi/nsp32.c target->limit_entry = entry; entry 298 drivers/scsi/pm8001/pm8001_sas.h struct list_head entry; entry 93 drivers/scsi/qla2xxx/qla_bsg.c pri_entry = &pri_cfg->entry[0]; entry 544 drivers/scsi/qla2xxx/qla_def.h struct list_head entry; entry 2650 drivers/scsi/qla2xxx/qla_def.h struct ct_fdmi_hba_attr entry[FDMI_HBA_ATTR_COUNT]; entry 2679 drivers/scsi/qla2xxx/qla_def.h struct ct_fdmiv2_hba_attr entry[FDMIV2_HBA_ATTR_COUNT]; entry 2746 drivers/scsi/qla2xxx/qla_def.h struct ct_fdmiv2_port_attr entry[FDMIV2_PORT_ATTR_COUNT]; entry 2764 drivers/scsi/qla2xxx/qla_def.h struct ct_fdmi_port_attr entry[FDMI_PORT_ATTR_COUNT]; entry 3249 drivers/scsi/qla2xxx/qla_def.h uint16_t entry; entry 2077 drivers/scsi/qla2xxx/qla_fw.h struct qla_fcp_prio_entry entry[1]; /* fcp priority entries */ entry 1074 drivers/scsi/qla2xxx/qla_gs.c uint8_t *entry; entry 1105 drivers/scsi/qla2xxx/qla_gs.c entry = &sns_cmd->p.gid_data[(i * 4) + 16]; entry 1106 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.domain = entry[1]; entry 1107 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.area = entry[2]; entry 1108 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.al_pa = entry[3]; entry 1111 drivers/scsi/qla2xxx/qla_gs.c if (entry[0] & BIT_7) { entry 1112 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.rsvd_1 = entry[0]; entry 3988 drivers/scsi/qla2xxx/qla_init.c msix->entry); entry 3989 drivers/scsi/qla2xxx/qla_init.c icb->msix = cpu_to_le16(msix->entry); entry 8741 drivers/scsi/qla2xxx/qla_init.c pri_entry = &ha->fcp_prio_cfg->entry[0]; entry 3513 drivers/scsi/qla2xxx/qla_isr.c qentry->entry = i; entry 4432 drivers/scsi/qla2xxx/qla_mbx.c mcp->mb[14] = rsp->msix->entry; entry 1102 drivers/scsi/qla2xxx/qla_sup.c struct qla_npiv_entry *entry; entry 1138 drivers/scsi/qla2xxx/qla_sup.c cnt = (sizeof(hdr) + le16_to_cpu(hdr.entries) * sizeof(*entry)) >> 1; entry 1150 drivers/scsi/qla2xxx/qla_sup.c entry = data + sizeof(struct qla_npiv_header); entry 1152 drivers/scsi/qla2xxx/qla_sup.c for (i = 0; cnt; cnt--, entry++, i++) { entry 1157 drivers/scsi/qla2xxx/qla_sup.c memcpy(&ha->npiv_info[i], entry, sizeof(struct qla_npiv_entry)); entry 1159 drivers/scsi/qla2xxx/qla_sup.c flags = le16_to_cpu(entry->flags); entry 1169 drivers/scsi/qla2xxx/qla_sup.c vid.port_name = wwn_to_u64(entry->port_name); entry 1170 drivers/scsi/qla2xxx/qla_sup.c vid.node_name = wwn_to_u64(entry->node_name); entry 1175 drivers/scsi/qla2xxx/qla_sup.c le16_to_cpu(entry->vf_id), entry 1176 drivers/scsi/qla2xxx/qla_sup.c entry->q_qos, entry->f_qos); entry 3623 drivers/scsi/qla2xxx/qla_sup.c ha->isp_ops->read_optrom(vha, &ha->fcp_prio_cfg->entry[0], entry 377 drivers/scsi/qla2xxx/qla_target.c struct imm_ntfy_from_isp *entry = entry 382 drivers/scsi/qla2xxx/qla_target.c if ((entry->u.isp24.vp_index != 0xFF) && entry 383 drivers/scsi/qla2xxx/qla_target.c (entry->u.isp24.nport_handle != 0xFFFF)) { entry 385 drivers/scsi/qla2xxx/qla_target.c entry->u.isp24.vp_index); entry 391 drivers/scsi/qla2xxx/qla_target.c vha->vp_idx, entry->u.isp24.vp_index); entry 406 drivers/scsi/qla2xxx/qla_target.c struct abts_recv_from_24xx *entry = entry 409 drivers/scsi/qla2xxx/qla_target.c entry->vp_index); entry 416 drivers/scsi/qla2xxx/qla_target.c vha->vp_idx, entry->vp_index); entry 450 drivers/scsi/qla2xxx/qla_target.c struct ctio7_from_24xx *entry = (struct ctio7_from_24xx *)pkt; entry 452 drivers/scsi/qla2xxx/qla_target.c entry->vp_index); entry 457 drivers/scsi/qla2xxx/qla_target.c vha->vp_idx, entry->vp_index); entry 467 drivers/scsi/qla2xxx/qla_target.c struct imm_ntfy_from_isp *entry = entry 470 drivers/scsi/qla2xxx/qla_target.c host = qlt_find_host_by_vp_idx(vha, entry->u.isp24.vp_index); entry 475 drivers/scsi/qla2xxx/qla_target.c vha->vp_idx, entry->u.isp24.vp_index); entry 485 drivers/scsi/qla2xxx/qla_target.c struct nack_to_isp *entry = (struct nack_to_isp *)pkt; entry 487 drivers/scsi/qla2xxx/qla_target.c if (0xFF != entry->u.isp24.vp_index) { entry 489 drivers/scsi/qla2xxx/qla_target.c entry->u.isp24.vp_index); entry 496 drivers/scsi/qla2xxx/qla_target.c entry->u.isp24.vp_index); entry 506 drivers/scsi/qla2xxx/qla_target.c struct abts_recv_from_24xx *entry = entry 509 drivers/scsi/qla2xxx/qla_target.c entry->vp_index); entry 514 drivers/scsi/qla2xxx/qla_target.c "vp_index %d\n", vha->vp_idx, entry->vp_index); entry 523 drivers/scsi/qla2xxx/qla_target.c struct abts_resp_to_24xx *entry = entry 526 drivers/scsi/qla2xxx/qla_target.c entry->vp_index); entry 531 drivers/scsi/qla2xxx/qla_target.c "vp_index %d\n", vha->vp_idx, entry->vp_index); entry 1880 drivers/scsi/qla2xxx/qla_target.c struct abts_recv_from_24xx *entry; entry 1892 drivers/scsi/qla2xxx/qla_target.c entry = &mcmd->orig_iocb.abts; entry 1895 drivers/scsi/qla2xxx/qla_target.c entry = (struct abts_recv_from_24xx *)pkt; entry 1904 drivers/scsi/qla2xxx/qla_target.c ctio->nport_handle = entry->nport_handle; entry 1908 drivers/scsi/qla2xxx/qla_target.c ctio->exchange_addr = entry->exchange_addr_to_abort; entry 1912 drivers/scsi/qla2xxx/qla_target.c ctio->initiator_id = entry->fcp_hdr_le.s_id; entry 1919 drivers/scsi/qla2xxx/qla_target.c ctio->initiator_id = entry->fcp_hdr_le.d_id; entry 1925 drivers/scsi/qla2xxx/qla_target.c ctio->u.status1.ox_id = entry->fcp_hdr_le.ox_id; entry 1944 drivers/scsi/qla2xxx/qla_target.c (struct abts_recv_from_24xx *)entry, FCP_TMF_CMPL, true); entry 5655 drivers/scsi/qla2xxx/qla_target.c struct qla_qpair *qpair, struct abts_resp_from_24xx_fw *entry) entry 5666 drivers/scsi/qla2xxx/qla_target.c if (qpair->retry_term_exchg_addr == entry->exchange_addr_to_abort && entry 5676 drivers/scsi/qla2xxx/qla_target.c vha, 0xffff, (uint8_t *)entry, sizeof(*entry)); entry 5687 drivers/scsi/qla2xxx/qla_target.c qpair->retry_term_exchg_addr = entry->exchange_addr_to_abort; entry 5699 drivers/scsi/qla2xxx/qla_target.c struct abts_resp_from_24xx_fw *entry = entry 5719 drivers/scsi/qla2xxx/qla_target.c entry->compl_status); entry 5721 drivers/scsi/qla2xxx/qla_target.c if (le16_to_cpu(entry->compl_status) != ABTS_RESP_COMPL_SUCCESS) { entry 5722 drivers/scsi/qla2xxx/qla_target.c if ((entry->error_subcode1 == 0x1E) && entry 5723 drivers/scsi/qla2xxx/qla_target.c (entry->error_subcode2 == 0)) { entry 5724 drivers/scsi/qla2xxx/qla_target.c if (qlt_chk_unresolv_exchg(vha, rsp->qpair, entry)) { entry 5733 drivers/scsi/qla2xxx/qla_target.c vha->vp_idx, entry->compl_status, entry 5734 drivers/scsi/qla2xxx/qla_target.c entry->error_subcode1, entry 5735 drivers/scsi/qla2xxx/qla_target.c entry->error_subcode2); entry 5766 drivers/scsi/qla2xxx/qla_target.c struct ctio7_from_24xx *entry = (struct ctio7_from_24xx *)pkt; entry 5768 drivers/scsi/qla2xxx/qla_target.c qlt_do_ctio_completion(vha, rsp, entry->handle, entry 5769 drivers/scsi/qla2xxx/qla_target.c le16_to_cpu(entry->status)|(pkt->entry_status << 16), entry 5770 drivers/scsi/qla2xxx/qla_target.c entry); entry 5826 drivers/scsi/qla2xxx/qla_target.c struct ctio_to_2xxx *entry = (struct ctio_to_2xxx *)pkt; entry 5828 drivers/scsi/qla2xxx/qla_target.c qlt_do_ctio_completion(vha, rsp, entry->handle, entry 5829 drivers/scsi/qla2xxx/qla_target.c le16_to_cpu(entry->status)|(pkt->entry_status << 16), entry 5830 drivers/scsi/qla2xxx/qla_target.c entry); entry 5836 drivers/scsi/qla2xxx/qla_target.c struct ctio_to_2xxx *entry = (struct ctio_to_2xxx *)pkt; entry 5838 drivers/scsi/qla2xxx/qla_target.c qlt_do_ctio_completion(vha, rsp, entry->handle, entry 5839 drivers/scsi/qla2xxx/qla_target.c le16_to_cpu(entry->status)|(pkt->entry_status << 16), entry 5840 drivers/scsi/qla2xxx/qla_target.c entry); entry 5851 drivers/scsi/qla2xxx/qla_target.c struct nack_to_isp *entry = (struct nack_to_isp *)pkt; entry 5855 drivers/scsi/qla2xxx/qla_target.c le16_to_cpu(entry->u.isp2x.seq_id), entry 5856 drivers/scsi/qla2xxx/qla_target.c le16_to_cpu(entry->u.isp2x.status)); entry 5858 drivers/scsi/qla2xxx/qla_target.c if (entry->u.isp2x.status != entry 5863 drivers/scsi/qla2xxx/qla_target.c le16_to_cpu(entry->u.isp2x.status)); entry 6822 drivers/scsi/qla2xxx/qla_target.c icb->msix_atio = cpu_to_le16(msix->entry); entry 6827 drivers/scsi/qla2xxx/qla_target.c msix->entry); entry 280 drivers/scsi/qla4xxx/ql4_def.h struct aen entry[MAX_AEN_ENTRIES]; entry 700 drivers/scsi/qla4xxx/ql4_isr.c ha->aen_log.entry[ha->aen_log.count].mbox_sts[i] = entry 120 drivers/scsi/sg.c struct list_head entry; /* list entry */ entry 875 drivers/scsi/sg.c list_for_each_entry(srp, &sfp->rq_list, entry) { entry 1002 drivers/scsi/sg.c list_for_each_entry(srp, &sfp->rq_list, entry) { entry 1016 drivers/scsi/sg.c list_for_each_entry(srp, &sfp->rq_list, entry) { entry 1188 drivers/scsi/sg.c list_for_each_entry(srp, &sfp->rq_list, entry) { entry 2083 drivers/scsi/sg.c list_for_each_entry(resp, &sfp->rq_list, entry) { entry 2119 drivers/scsi/sg.c list_add_tail(&rp->entry, &sfp->rq_list); entry 2137 drivers/scsi/sg.c if (!list_empty(&srp->entry)) { entry 2138 drivers/scsi/sg.c list_del(&srp->entry); entry 2205 drivers/scsi/sg.c srp = list_first_entry(&sfp->rq_list, Sg_request, entry); entry 2207 drivers/scsi/sg.c list_del(&srp->entry); entry 2539 drivers/scsi/sg.c list_for_each_entry(srp, &fp->rq_list, entry) { entry 26 drivers/sh/intc/virq-debugfs.c struct intc_map_entry *entry = intc_irq_xlate_get(i); entry 27 drivers/sh/intc/virq-debugfs.c struct intc_desc_int *desc = entry->desc; entry 33 drivers/sh/intc/virq-debugfs.c seq_printf(m, "0x%05x ", entry->enum_id); entry 27 drivers/sh/intc/virq.c #define for_each_virq(entry, head) \ entry 28 drivers/sh/intc/virq.c for (entry = head; entry; entry = entry->next) entry 86 drivers/sh/intc/virq.c struct intc_virq_list *entry; entry 90 drivers/sh/intc/virq.c for_each_virq(entry, irq_get_handler_data(irq)) { entry 91 drivers/sh/intc/virq.c if (entry->irq == virq) entry 93 drivers/sh/intc/virq.c last = &entry->next; entry 96 drivers/sh/intc/virq.c entry = kzalloc(sizeof(struct intc_virq_list), GFP_ATOMIC); entry 97 drivers/sh/intc/virq.c if (!entry) entry 100 drivers/sh/intc/virq.c entry->irq = virq; entry 103 drivers/sh/intc/virq.c *last = entry; entry 105 drivers/sh/intc/virq.c irq_set_handler_data(irq, entry); entry 115 drivers/sh/intc/virq.c struct intc_virq_list *entry, *vlist = irq_data_get_irq_handler_data(data); entry 120 drivers/sh/intc/virq.c for_each_virq(entry, vlist) { entry 122 drivers/sh/intc/virq.c struct irq_desc *vdesc = irq_to_desc(entry->irq); entry 165 drivers/sh/intc/virq.c struct intc_subgroup_entry *entry; entry 171 drivers/sh/intc/virq.c entry = kmalloc(sizeof(*entry), GFP_NOWAIT); entry 172 drivers/sh/intc/virq.c if (!entry) entry 175 drivers/sh/intc/virq.c entry->pirq = pirq; entry 176 drivers/sh/intc/virq.c entry->enum_id = subgroup->enum_ids[i]; entry 177 drivers/sh/intc/virq.c entry->handle = intc_subgroup_data(subgroup, d, i); entry 179 drivers/sh/intc/virq.c err = radix_tree_insert(&d->tree, entry->enum_id, entry); entry 183 drivers/sh/intc/virq.c radix_tree_tag_set(&d->tree, entry->enum_id, entry 216 drivers/sh/intc/virq.c struct intc_subgroup_entry *entry; entry 219 drivers/sh/intc/virq.c entry = radix_tree_deref_slot((void **)entries[i]); entry 220 drivers/sh/intc/virq.c if (unlikely(!entry)) entry 222 drivers/sh/intc/virq.c if (radix_tree_deref_retry(entry)) entry 234 drivers/sh/intc/virq.c irq, entry->pirq); entry 236 drivers/sh/intc/virq.c intc_irq_xlate_set(irq, entry->enum_id, d); entry 238 drivers/sh/intc/virq.c irq_set_chip_and_handler_name(irq, irq_get_chip(entry->pirq), entry 240 drivers/sh/intc/virq.c irq_set_chip_data(irq, irq_get_chip_data(entry->pirq)); entry 242 drivers/sh/intc/virq.c irq_set_handler_data(irq, (void *)entry->handle); entry 250 drivers/sh/intc/virq.c add_virq_to_pirq(entry->pirq, irq); entry 251 drivers/sh/intc/virq.c irq_set_chained_handler(entry->pirq, intc_virq_handler); entry 253 drivers/sh/intc/virq.c radix_tree_tag_clear(&d->tree, entry->enum_id, entry 118 drivers/soc/fsl/qe/qe_common.c struct muram_block *entry; entry 129 drivers/soc/fsl/qe/qe_common.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 130 drivers/soc/fsl/qe/qe_common.c if (!entry) entry 132 drivers/soc/fsl/qe/qe_common.c entry->start = start; entry 133 drivers/soc/fsl/qe/qe_common.c entry->size = size; entry 134 drivers/soc/fsl/qe/qe_common.c list_add(&entry->head, &muram_block_list); entry 170 drivers/soc/qcom/smem.c struct smem_ptable_entry entry[]; entry 403 drivers/soc/qcom/smem.c struct smem_global_entry *entry; entry 407 drivers/soc/qcom/smem.c entry = &header->toc[item]; entry 408 drivers/soc/qcom/smem.c if (entry->allocated) entry 415 drivers/soc/qcom/smem.c entry->offset = header->free_offset; entry 416 drivers/soc/qcom/smem.c entry->size = cpu_to_le32(size); entry 424 drivers/soc/qcom/smem.c entry->allocated = cpu_to_le32(1); entry 487 drivers/soc/qcom/smem.c struct smem_global_entry *entry; entry 492 drivers/soc/qcom/smem.c entry = &header->toc[item]; entry 493 drivers/soc/qcom/smem.c if (!entry->allocated) entry 496 drivers/soc/qcom/smem.c aux_base = le32_to_cpu(entry->aux_base) & AUX_BASE_MASK; entry 503 drivers/soc/qcom/smem.c *size = le32_to_cpu(entry->size); entry 504 drivers/soc/qcom/smem.c return region->virt_base + le32_to_cpu(entry->offset); entry 712 drivers/soc/qcom/smem.c info = (struct smem_info *)&ptable->entry[ptable->num_entries]; entry 726 drivers/soc/qcom/smem.c struct smem_ptable_entry *entry, u16 host0, u16 host1) entry 731 drivers/soc/qcom/smem.c header = smem->regions[0].virt_base + le32_to_cpu(entry->offset); entry 752 drivers/soc/qcom/smem.c if (size != le32_to_cpu(entry->size)) { entry 754 drivers/soc/qcom/smem.c size, le32_to_cpu(entry->size)); entry 770 drivers/soc/qcom/smem.c struct smem_ptable_entry *entry; entry 785 drivers/soc/qcom/smem.c entry = &ptable->entry[i]; entry 786 drivers/soc/qcom/smem.c if (!le32_to_cpu(entry->offset)) entry 788 drivers/soc/qcom/smem.c if (!le32_to_cpu(entry->size)) entry 791 drivers/soc/qcom/smem.c if (le16_to_cpu(entry->host0) != SMEM_GLOBAL_HOST) entry 794 drivers/soc/qcom/smem.c if (le16_to_cpu(entry->host1) == SMEM_GLOBAL_HOST) { entry 805 drivers/soc/qcom/smem.c header = qcom_smem_partition_header(smem, entry, entry 811 drivers/soc/qcom/smem.c smem->global_cacheline = le32_to_cpu(entry->cacheline); entry 820 drivers/soc/qcom/smem.c struct smem_ptable_entry *entry; entry 831 drivers/soc/qcom/smem.c entry = &ptable->entry[i]; entry 832 drivers/soc/qcom/smem.c if (!le32_to_cpu(entry->offset)) entry 834 drivers/soc/qcom/smem.c if (!le32_to_cpu(entry->size)) entry 837 drivers/soc/qcom/smem.c host0 = le16_to_cpu(entry->host0); entry 838 drivers/soc/qcom/smem.c host1 = le16_to_cpu(entry->host1); entry 856 drivers/soc/qcom/smem.c header = qcom_smem_partition_header(smem, entry, host0, host1); entry 861 drivers/soc/qcom/smem.c smem->cacheline[remote_host] = le32_to_cpu(entry->cacheline); entry 175 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry; entry 202 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->inbound, node) { entry 204 drivers/soc/qcom/smp2p.c if (!strcmp(buf, entry->name)) { entry 205 drivers/soc/qcom/smp2p.c entry->value = &in->entries[i].value; entry 213 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->inbound, node) { entry 215 drivers/soc/qcom/smp2p.c if (!entry->value) entry 218 drivers/soc/qcom/smp2p.c val = readl(entry->value); entry 220 drivers/soc/qcom/smp2p.c status = val ^ entry->last_value; entry 221 drivers/soc/qcom/smp2p.c entry->last_value = val; entry 227 drivers/soc/qcom/smp2p.c for_each_set_bit(i, entry->irq_enabled, 32) { entry 231 drivers/soc/qcom/smp2p.c if ((val & BIT(i) && test_bit(i, entry->irq_rising)) || entry 232 drivers/soc/qcom/smp2p.c (!(val & BIT(i)) && test_bit(i, entry->irq_falling))) { entry 233 drivers/soc/qcom/smp2p.c irq_pin = irq_find_mapping(entry->domain, i); entry 244 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry = irq_data_get_irq_chip_data(irqd); entry 247 drivers/soc/qcom/smp2p.c clear_bit(irq, entry->irq_enabled); entry 252 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry = irq_data_get_irq_chip_data(irqd); entry 255 drivers/soc/qcom/smp2p.c set_bit(irq, entry->irq_enabled); entry 260 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry = irq_data_get_irq_chip_data(irqd); entry 267 drivers/soc/qcom/smp2p.c set_bit(irq, entry->irq_rising); entry 269 drivers/soc/qcom/smp2p.c clear_bit(irq, entry->irq_rising); entry 272 drivers/soc/qcom/smp2p.c set_bit(irq, entry->irq_falling); entry 274 drivers/soc/qcom/smp2p.c clear_bit(irq, entry->irq_falling); entry 290 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry = d->host_data; entry 293 drivers/soc/qcom/smp2p.c irq_set_chip_data(irq, entry); entry 306 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry, entry 309 drivers/soc/qcom/smp2p.c entry->domain = irq_domain_add_linear(node, 32, &smp2p_irq_ops, entry); entry 310 drivers/soc/qcom/smp2p.c if (!entry->domain) { entry 320 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry = data; entry 324 drivers/soc/qcom/smp2p.c spin_lock(&entry->lock); entry 325 drivers/soc/qcom/smp2p.c val = orig = readl(entry->value); entry 328 drivers/soc/qcom/smp2p.c writel(val, entry->value); entry 329 drivers/soc/qcom/smp2p.c spin_unlock(&entry->lock); entry 332 drivers/soc/qcom/smp2p.c qcom_smp2p_kick(entry->smp2p); entry 342 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry, entry 349 drivers/soc/qcom/smp2p.c strlcpy(buf, entry->name, SMP2P_MAX_ENTRY_NAME); entry 353 drivers/soc/qcom/smp2p.c entry->value = &out->entries[out->valid_entries].value; entry 357 drivers/soc/qcom/smp2p.c entry->state = qcom_smem_state_register(node, &smp2p_state_ops, entry); entry 358 drivers/soc/qcom/smp2p.c if (IS_ERR(entry->state)) { entry 360 drivers/soc/qcom/smp2p.c return PTR_ERR(entry->state); entry 443 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry; entry 501 drivers/soc/qcom/smp2p.c entry = devm_kzalloc(&pdev->dev, sizeof(*entry), GFP_KERNEL); entry 502 drivers/soc/qcom/smp2p.c if (!entry) { entry 507 drivers/soc/qcom/smp2p.c entry->smp2p = smp2p; entry 508 drivers/soc/qcom/smp2p.c spin_lock_init(&entry->lock); entry 510 drivers/soc/qcom/smp2p.c ret = of_property_read_string(node, "qcom,entry-name", &entry->name); entry 515 drivers/soc/qcom/smp2p.c ret = qcom_smp2p_inbound_entry(smp2p, entry, node); entry 519 drivers/soc/qcom/smp2p.c list_add(&entry->node, &smp2p->inbound); entry 521 drivers/soc/qcom/smp2p.c ret = qcom_smp2p_outbound_entry(smp2p, entry, node); entry 525 drivers/soc/qcom/smp2p.c list_add(&entry->node, &smp2p->outbound); entry 545 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->inbound, node) entry 546 drivers/soc/qcom/smp2p.c irq_domain_remove(entry->domain); entry 548 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->outbound, node) entry 549 drivers/soc/qcom/smp2p.c qcom_smem_state_unregister(entry->state); entry 566 drivers/soc/qcom/smp2p.c struct smp2p_entry *entry; entry 568 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->inbound, node) entry 569 drivers/soc/qcom/smp2p.c irq_domain_remove(entry->domain); entry 571 drivers/soc/qcom/smp2p.c list_for_each_entry(entry, &smp2p->outbound, node) entry 572 drivers/soc/qcom/smp2p.c qcom_smem_state_unregister(entry->state); entry 200 drivers/soc/qcom/smsm.c struct smsm_entry *entry = data; entry 206 drivers/soc/qcom/smsm.c val = readl(entry->remote_state); entry 207 drivers/soc/qcom/smsm.c changed = val ^ entry->last_value; entry 208 drivers/soc/qcom/smsm.c entry->last_value = val; entry 210 drivers/soc/qcom/smsm.c for_each_set_bit(i, entry->irq_enabled, 32) { entry 215 drivers/soc/qcom/smsm.c if (test_bit(i, entry->irq_rising)) { entry 216 drivers/soc/qcom/smsm.c irq_pin = irq_find_mapping(entry->domain, i); entry 220 drivers/soc/qcom/smsm.c if (test_bit(i, entry->irq_falling)) { entry 221 drivers/soc/qcom/smsm.c irq_pin = irq_find_mapping(entry->domain, i); entry 239 drivers/soc/qcom/smsm.c struct smsm_entry *entry = irq_data_get_irq_chip_data(irqd); entry 241 drivers/soc/qcom/smsm.c struct qcom_smsm *smsm = entry->smsm; entry 244 drivers/soc/qcom/smsm.c if (entry->subscription) { entry 245 drivers/soc/qcom/smsm.c val = readl(entry->subscription + smsm->local_host); entry 247 drivers/soc/qcom/smsm.c writel(val, entry->subscription + smsm->local_host); entry 250 drivers/soc/qcom/smsm.c clear_bit(irq, entry->irq_enabled); entry 264 drivers/soc/qcom/smsm.c struct smsm_entry *entry = irq_data_get_irq_chip_data(irqd); entry 266 drivers/soc/qcom/smsm.c struct qcom_smsm *smsm = entry->smsm; entry 269 drivers/soc/qcom/smsm.c set_bit(irq, entry->irq_enabled); entry 271 drivers/soc/qcom/smsm.c if (entry->subscription) { entry 272 drivers/soc/qcom/smsm.c val = readl(entry->subscription + smsm->local_host); entry 274 drivers/soc/qcom/smsm.c writel(val, entry->subscription + smsm->local_host); entry 285 drivers/soc/qcom/smsm.c struct smsm_entry *entry = irq_data_get_irq_chip_data(irqd); entry 292 drivers/soc/qcom/smsm.c set_bit(irq, entry->irq_rising); entry 294 drivers/soc/qcom/smsm.c clear_bit(irq, entry->irq_rising); entry 297 drivers/soc/qcom/smsm.c set_bit(irq, entry->irq_falling); entry 299 drivers/soc/qcom/smsm.c clear_bit(irq, entry->irq_falling); entry 321 drivers/soc/qcom/smsm.c struct smsm_entry *entry = d->host_data; entry 324 drivers/soc/qcom/smsm.c irq_set_chip_data(irq, entry); entry 382 drivers/soc/qcom/smsm.c struct smsm_entry *entry, entry 397 drivers/soc/qcom/smsm.c "smsm", (void *)entry); entry 403 drivers/soc/qcom/smsm.c entry->domain = irq_domain_add_linear(node, 32, &smsm_irq_ops, entry); entry 404 drivers/soc/qcom/smsm.c if (!entry->domain) { entry 459 drivers/soc/qcom/smsm.c struct smsm_entry *entry; entry 562 drivers/soc/qcom/smsm.c entry = &smsm->entries[id]; entry 564 drivers/soc/qcom/smsm.c entry->smsm = smsm; entry 565 drivers/soc/qcom/smsm.c entry->remote_state = states + id; entry 568 drivers/soc/qcom/smsm.c entry->subscription = intr_mask + id * smsm->num_hosts; entry 569 drivers/soc/qcom/smsm.c writel(0, entry->subscription + smsm->local_host); entry 571 drivers/soc/qcom/smsm.c ret = smsm_inbound_entry(smsm, entry, node); entry 2701 drivers/spi/spi.c INIT_LIST_HEAD(&sres->entry); entry 2720 drivers/spi/spi.c WARN_ON(!list_empty(&sres->entry)); entry 2734 drivers/spi/spi.c WARN_ON(!list_empty(&sres->entry)); entry 2735 drivers/spi/spi.c list_add_tail(&sres->entry, &message->resources); entry 2748 drivers/spi/spi.c list_for_each_entry_safe_reverse(res, tmp, &message->resources, entry) { entry 2752 drivers/spi/spi.c list_del(&res->entry); entry 860 drivers/staging/android/vsoc.c vsoc_dev.msix_entries[i].entry = i; entry 260 drivers/staging/comedi/drivers/daqboard2000.c static void db2k_write_acq_scan_list_entry(struct comedi_device *dev, u16 entry) entry 262 drivers/staging/comedi/drivers/daqboard2000.c writew(entry & 0x00ff, dev->mmio + DB2K_REG_ACQ_SCAN_LIST_FIFO); entry 263 drivers/staging/comedi/drivers/daqboard2000.c writew((entry >> 8) & 0x00ff, entry 468 drivers/staging/comedi/drivers/me4000.c unsigned int entry; entry 472 drivers/staging/comedi/drivers/me4000.c entry = chan | ME4000_AI_LIST_RANGE(range); entry 491 drivers/staging/comedi/drivers/me4000.c entry |= ME4000_AI_LIST_INPUT_DIFFERENTIAL; entry 494 drivers/staging/comedi/drivers/me4000.c entry |= ME4000_AI_LIST_LAST_ENTRY; entry 501 drivers/staging/comedi/drivers/me4000.c outl(entry, dev->iobase + ME4000_AI_CHANNEL_LIST_REG); entry 637 drivers/staging/comedi/drivers/me4000.c unsigned int entry; entry 639 drivers/staging/comedi/drivers/me4000.c entry = chan | ME4000_AI_LIST_RANGE(range); entry 642 drivers/staging/comedi/drivers/me4000.c entry |= ME4000_AI_LIST_INPUT_DIFFERENTIAL; entry 645 drivers/staging/comedi/drivers/me4000.c entry |= ME4000_AI_LIST_LAST_ENTRY; entry 647 drivers/staging/comedi/drivers/me4000.c outl(entry, dev->iobase + ME4000_AI_CHANNEL_LIST_REG); entry 281 drivers/staging/exfat/exfat.h s32 entry; entry 511 drivers/staging/exfat/exfat.h s32 entry; entry 547 drivers/staging/exfat/exfat.h s32 entry, u32 type, u32 start_clu, u64 size); entry 549 drivers/staging/exfat/exfat.h s32 entry, s32 num_entries, entry 556 drivers/staging/exfat/exfat.h struct chain_t *p_dir, s32 entry, entry 559 drivers/staging/exfat/exfat.h struct chain_t *p_dir, s32 entry, entry 562 drivers/staging/exfat/exfat.h struct chain_t *p_dir, s32 entry, entry 845 drivers/staging/exfat/exfat.h s32 fat_init_dir_entry(struct super_block *sb, struct chain_t *p_dir, s32 entry, entry 848 drivers/staging/exfat/exfat.h s32 entry, u32 type, u32 start_clu, u64 size); entry 850 drivers/staging/exfat/exfat.h s32 entry, s32 num_entries, entry 854 drivers/staging/exfat/exfat.h s32 entry, s32 num_entries, entry 865 drivers/staging/exfat/exfat.h s32 entry, s32 order, s32 num_entries); entry 867 drivers/staging/exfat/exfat.h s32 entry, s32 order, s32 num_entries); entry 869 drivers/staging/exfat/exfat.h s32 find_location(struct super_block *sb, struct chain_t *p_dir, s32 entry, entry 874 drivers/staging/exfat/exfat.h s32 entry, sector_t *sector); entry 876 drivers/staging/exfat/exfat.h struct chain_t *p_dir, s32 entry, entry 895 drivers/staging/exfat/exfat.h s32 entry, struct dentry_t *p_entry); entry 897 drivers/staging/exfat/exfat.h s32 entry, struct dentry_t *p_entry); entry 901 drivers/staging/exfat/exfat.h s32 entry); entry 914 drivers/staging/exfat/exfat.h struct chain_t *p_dir, s32 entry, entry 917 drivers/staging/exfat/exfat.h struct chain_t *p_dir, s32 entry, entry 945 drivers/staging/exfat/exfat.h void remove_file(struct inode *inode, struct chain_t *p_dir, s32 entry); entry 1239 drivers/staging/exfat/exfat_core.c s32 fat_init_dir_entry(struct super_block *sb, struct chain_t *p_dir, s32 entry, entry 1245 drivers/staging/exfat/exfat_core.c dos_ep = (struct dos_dentry_t *)get_entry_in_dir(sb, p_dir, entry, entry 1257 drivers/staging/exfat/exfat_core.c s32 entry, u32 type, u32 start_clu, u64 size) entry 1267 drivers/staging/exfat/exfat_core.c file_ep = (struct file_dentry_t *)get_entry_in_dir(sb, p_dir, entry, entry 1272 drivers/staging/exfat/exfat_core.c strm_ep = (struct strm_dentry_t *)get_entry_in_dir(sb, p_dir, entry + 1, entry 1287 drivers/staging/exfat/exfat_core.c s32 entry, s32 num_entries, entry 1298 drivers/staging/exfat/exfat_core.c dos_ep = (struct dos_dentry_t *)get_entry_in_dir(sb, p_dir, entry, entry 1314 drivers/staging/exfat/exfat_core.c entry - i, entry 1325 drivers/staging/exfat/exfat_core.c entry - i, entry 1338 drivers/staging/exfat/exfat_core.c s32 entry, s32 num_entries, entry 1349 drivers/staging/exfat/exfat_core.c file_ep = (struct file_dentry_t *)get_entry_in_dir(sb, p_dir, entry, entry 1357 drivers/staging/exfat/exfat_core.c strm_ep = (struct strm_dentry_t *)get_entry_in_dir(sb, p_dir, entry + 1, entry 1368 drivers/staging/exfat/exfat_core.c entry + i, entry 1378 drivers/staging/exfat/exfat_core.c update_dir_checksum(sb, p_dir, entry); entry 1487 drivers/staging/exfat/exfat_core.c s32 entry, s32 order, s32 num_entries) entry 1495 drivers/staging/exfat/exfat_core.c ep = get_entry_in_dir(sb, p_dir, entry - i, §or); entry 1505 drivers/staging/exfat/exfat_core.c s32 entry, s32 order, s32 num_entries) entry 1513 drivers/staging/exfat/exfat_core.c ep = get_entry_in_dir(sb, p_dir, entry + i, §or); entry 1523 drivers/staging/exfat/exfat_core.c s32 entry) entry 1531 drivers/staging/exfat/exfat_core.c file_ep = (struct file_dentry_t *)get_entry_in_dir(sb, p_dir, entry, entry 1543 drivers/staging/exfat/exfat_core.c ep = get_entry_in_dir(sb, p_dir, entry + i, NULL); entry 1604 drivers/staging/exfat/exfat_core.c s32 find_location(struct super_block *sb, struct chain_t *p_dir, s32 entry, entry 1612 drivers/staging/exfat/exfat_core.c off = entry << DENTRY_SIZE_BITS; entry 1650 drivers/staging/exfat/exfat_core.c s32 entry, sector_t *sector) entry 1656 drivers/staging/exfat/exfat_core.c if (find_location(sb, p_dir, entry, &sec, &off) != FFS_SUCCESS) entry 1688 drivers/staging/exfat/exfat_core.c struct chain_t *p_dir, s32 entry, entry 1708 drivers/staging/exfat/exfat_core.c byte_offset = entry << DENTRY_SIZE_BITS; entry 1976 drivers/staging/exfat/exfat_core.c if (p_fs->hint_uentry.entry == -1) entry 1983 drivers/staging/exfat/exfat_core.c dentry = p_fs->hint_uentry.entry; entry 1985 drivers/staging/exfat/exfat_core.c p_fs->hint_uentry.entry = -1; entry 2012 drivers/staging/exfat/exfat_core.c if (p_fs->hint_uentry.entry == -1) { entry 2014 drivers/staging/exfat/exfat_core.c p_fs->hint_uentry.entry = dentry; entry 2028 drivers/staging/exfat/exfat_core.c p_fs->hint_uentry.entry = -1; entry 2101 drivers/staging/exfat/exfat_core.c if (p_fs->hint_uentry.entry == -1) { entry 2103 drivers/staging/exfat/exfat_core.c p_fs->hint_uentry.entry = p_dir->size << (p_fs->cluster_size_bits - DENTRY_SIZE_BITS); entry 2118 drivers/staging/exfat/exfat_core.c fid->entry + 1, §or); entry 2126 drivers/staging/exfat/exfat_core.c fid->entry); entry 2278 drivers/staging/exfat/exfat_core.c p_fs->hint_uentry.entry = -1; entry 2295 drivers/staging/exfat/exfat_core.c if (p_fs->hint_uentry.entry == -1) { entry 2304 drivers/staging/exfat/exfat_core.c p_fs->hint_uentry.entry = dentry - (num_empty - 1); entry 2352 drivers/staging/exfat/exfat_core.c p_fs->hint_uentry.entry = -1; entry 2387 drivers/staging/exfat/exfat_core.c s32 entry, struct dentry_t *p_entry) entry 2397 drivers/staging/exfat/exfat_core.c for (entry--; entry >= 0; entry--) { entry 2399 drivers/staging/exfat/exfat_core.c entry, NULL); entry 2417 drivers/staging/exfat/exfat_core.c s32 entry, struct dentry_t *p_entry) entry 2425 drivers/staging/exfat/exfat_core.c for (i = 0, entry++; i < file_ep->num_ext; i++, entry++) { entry 2426 drivers/staging/exfat/exfat_core.c ext_ep = get_entry_in_dir(sb, p_dir, entry, NULL); entry 2618 drivers/staging/exfat/exfat_core.c struct chain_t *p_dir, s32 entry, entry 2625 drivers/staging/exfat/exfat_core.c for (entry--, i = 1; entry >= 0; entry--, i++) { entry 2626 drivers/staging/exfat/exfat_core.c ep = (struct ext_dentry_t *)get_entry_in_dir(sb, p_dir, entry, entry 2645 drivers/staging/exfat/exfat_core.c struct chain_t *p_dir, s32 entry, entry 2653 drivers/staging/exfat/exfat_core.c es = get_entry_set_in_dir(sb, p_dir, entry, ES_ALL_ENTRIES, &ep); entry 3291 drivers/staging/exfat/exfat_core.c fid->entry = dentry; entry 3341 drivers/staging/exfat/exfat_core.c fid->entry = dentry; entry 3355 drivers/staging/exfat/exfat_core.c void remove_file(struct inode *inode, struct chain_t *p_dir, s32 entry) entry 3364 drivers/staging/exfat/exfat_core.c ep = get_entry_in_dir(sb, p_dir, entry, §or); entry 3371 drivers/staging/exfat/exfat_core.c num_entries = fs_func->count_ext_entries(sb, p_dir, entry, ep); entry 3381 drivers/staging/exfat/exfat_core.c fs_func->delete_dir_entry(sb, p_dir, entry, 0, num_entries); entry 3465 drivers/staging/exfat/exfat_core.c fid->entry = newentry; entry 3591 drivers/staging/exfat/exfat_core.c fid->entry = newentry; entry 588 drivers/staging/exfat/exfat_super.c fid->entry = dentry; entry 1006 drivers/staging/exfat/exfat_super.c es = get_entry_set_in_dir(sb, &(fid->dir), fid->entry, entry 1012 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); entry 1137 drivers/staging/exfat/exfat_super.c es = get_entry_set_in_dir(sb, &fid->dir, fid->entry, entry 1145 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); entry 1250 drivers/staging/exfat/exfat_super.c dentry = fid->entry; entry 1281 drivers/staging/exfat/exfat_super.c new_entry = new_fid->entry; entry 1365 drivers/staging/exfat/exfat_super.c dentry = fid->entry; entry 1429 drivers/staging/exfat/exfat_super.c (fid->entry == -1)) { entry 1441 drivers/staging/exfat/exfat_super.c es = get_entry_set_in_dir(sb, &(fid->dir), fid->entry, entry 1448 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); entry 1519 drivers/staging/exfat/exfat_super.c (fid->entry == -1)) { entry 1557 drivers/staging/exfat/exfat_super.c es = get_entry_set_in_dir(sb, &(fid->dir), fid->entry, entry 1565 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); entry 1601 drivers/staging/exfat/exfat_super.c p_fs->fs_func->get_uni_name_from_ext_entry(sb, &(fid->dir), fid->entry, entry 1669 drivers/staging/exfat/exfat_super.c (fid->entry == -1)) { entry 1681 drivers/staging/exfat/exfat_super.c es = get_entry_set_in_dir(sb, &(fid->dir), fid->entry, entry 1690 drivers/staging/exfat/exfat_super.c ep = get_entry_in_dir(sb, &(fid->dir), fid->entry, §or); entry 1835 drivers/staging/exfat/exfat_super.c es = get_entry_set_in_dir(sb, &fid->dir, fid->entry, entry 1849 drivers/staging/exfat/exfat_super.c fid->entry, §or); entry 1961 drivers/staging/exfat/exfat_super.c if (fid->entry == -1) { entry 2149 drivers/staging/exfat/exfat_super.c dentry = fid->entry; entry 2384 drivers/staging/exfat/exfat_super.c i_pos = ((loff_t)fid.dir.dir << 32) | (fid.entry & 0xffffffff); entry 2449 drivers/staging/exfat/exfat_super.c i_pos = ((loff_t)fid.dir.dir << 32) | (fid.entry & 0xffffffff); entry 2606 drivers/staging/exfat/exfat_super.c i_pos = ((loff_t)fid.dir.dir << 32) | (fid.entry & 0xffffffff); entry 2666 drivers/staging/exfat/exfat_super.c i_pos = ((loff_t)fid.dir.dir << 32) | (fid.entry & 0xffffffff); entry 2778 drivers/staging/exfat/exfat_super.c (EXFAT_I(old_inode)->fid.entry & 0xffffffff); entry 3795 drivers/staging/exfat/exfat_super.c EXFAT_I(inode)->fid.entry = -1; entry 236 drivers/staging/fsl-dpaa2/ethsw/ethsw.c struct dpsw_fdb_unicast_cfg entry = {0}; entry 239 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.if_egress = port_priv->idx; entry 240 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.type = DPSW_FDB_ENTRY_STATIC; entry 241 drivers/staging/fsl-dpaa2/ethsw/ethsw.c ether_addr_copy(entry.mac_addr, addr); entry 245 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 0, &entry); entry 255 drivers/staging/fsl-dpaa2/ethsw/ethsw.c struct dpsw_fdb_unicast_cfg entry = {0}; entry 258 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.if_egress = port_priv->idx; entry 259 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.type = DPSW_FDB_ENTRY_STATIC; entry 260 drivers/staging/fsl-dpaa2/ethsw/ethsw.c ether_addr_copy(entry.mac_addr, addr); entry 264 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 0, &entry); entry 275 drivers/staging/fsl-dpaa2/ethsw/ethsw.c struct dpsw_fdb_multicast_cfg entry = {0}; entry 278 drivers/staging/fsl-dpaa2/ethsw/ethsw.c ether_addr_copy(entry.mac_addr, addr); entry 279 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.type = DPSW_FDB_ENTRY_STATIC; entry 280 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.num_ifs = 1; entry 281 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.if_id[0] = port_priv->idx; entry 285 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 0, &entry); entry 296 drivers/staging/fsl-dpaa2/ethsw/ethsw.c struct dpsw_fdb_multicast_cfg entry = {0}; entry 299 drivers/staging/fsl-dpaa2/ethsw/ethsw.c ether_addr_copy(entry.mac_addr, addr); entry 300 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.type = DPSW_FDB_ENTRY_STATIC; entry 301 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.num_ifs = 1; entry 302 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.if_id[0] = port_priv->idx; entry 306 drivers/staging/fsl-dpaa2/ethsw/ethsw.c 0, &entry); entry 559 drivers/staging/fsl-dpaa2/ethsw/ethsw.c static int ethsw_fdb_do_dump(struct fdb_dump_entry *entry, entry 562 drivers/staging/fsl-dpaa2/ethsw/ethsw.c int is_dynamic = entry->type & DPSW_FDB_ENTRY_DINAMIC; entry 585 drivers/staging/fsl-dpaa2/ethsw/ethsw.c if (nla_put(dump->skb, NDA_LLADDR, ETH_ALEN, entry->mac_addr)) entry 599 drivers/staging/fsl-dpaa2/ethsw/ethsw.c static int port_fdb_valid_entry(struct fdb_dump_entry *entry, entry 605 drivers/staging/fsl-dpaa2/ethsw/ethsw.c if (entry->type & DPSW_FDB_ENTRY_TYPE_UNICAST) entry 606 drivers/staging/fsl-dpaa2/ethsw/ethsw.c valid = entry->if_info == port_priv->idx; entry 608 drivers/staging/fsl-dpaa2/ethsw/ethsw.c valid = entry->if_mask[idx / 8] & BIT(idx % 8); entry 164 drivers/staging/gasket/gasket_interrupt.c interrupt = interrupt_data->msix_entries[i].entry; entry 189 drivers/staging/gasket/gasket_interrupt.c interrupt_data->msix_entries[i].entry = i; entry 1146 drivers/staging/greybus/camera.c const struct gb_camera_debugfs_entry *entry = entry 1149 drivers/staging/greybus/camera.c if (!strcmp(file->f_path.dentry->d_iname, entry->name)) { entry 1150 drivers/staging/greybus/camera.c file->private_data = (void *)entry; entry 1185 drivers/staging/greybus/camera.c const struct gb_camera_debugfs_entry *entry = entry 1190 drivers/staging/greybus/camera.c debugfs_create_file(entry->name, entry->mask, entry 30 drivers/staging/greybus/raw.c struct list_head entry; entry 83 drivers/staging/greybus/raw.c list_add_tail(&raw_data->entry, &raw->list); entry 239 drivers/staging/greybus/raw.c list_for_each_entry_safe(raw_data, temp, &raw->list, entry) { entry 240 drivers/staging/greybus/raw.c list_del(&raw_data->entry); entry 296 drivers/staging/greybus/raw.c raw_data = list_first_entry(&raw->list, struct raw_data, entry); entry 307 drivers/staging/greybus/raw.c list_del(&raw_data->entry); entry 30 drivers/staging/media/ipu3/ipu3-css-pool.c imgu_dmamap_free(imgu, &pool->entry[i].param); entry 39 drivers/staging/media/ipu3/ipu3-css-pool.c pool->entry[i].valid = false; entry 41 drivers/staging/media/ipu3/ipu3-css-pool.c pool->entry[i].param.vaddr = NULL; entry 45 drivers/staging/media/ipu3/ipu3-css-pool.c if (!imgu_dmamap_alloc(imgu, &pool->entry[i].param, size)) entry 66 drivers/staging/media/ipu3/ipu3-css-pool.c pool->entry[n].valid = true; entry 75 drivers/staging/media/ipu3/ipu3-css-pool.c pool->entry[pool->last].valid = false; entry 96 drivers/staging/media/ipu3/ipu3-css-pool.c if (!pool->entry[i].valid) entry 99 drivers/staging/media/ipu3/ipu3-css-pool.c return &pool->entry[i].param; entry 40 drivers/staging/media/ipu3/ipu3-css-pool.h } entry[IPU3_CSS_POOL_SIZE]; entry 29 drivers/staging/media/tegra-vde/dmabuf-cache.c static void tegra_vde_release_entry(struct tegra_vde_cache_entry *entry) entry 31 drivers/staging/media/tegra-vde/dmabuf-cache.c struct dma_buf *dmabuf = entry->a->dmabuf; entry 33 drivers/staging/media/tegra-vde/dmabuf-cache.c WARN_ON_ONCE(entry->refcnt); entry 35 drivers/staging/media/tegra-vde/dmabuf-cache.c if (entry->vde->domain) entry 36 drivers/staging/media/tegra-vde/dmabuf-cache.c tegra_vde_iommu_unmap(entry->vde, entry->iova); entry 38 drivers/staging/media/tegra-vde/dmabuf-cache.c dma_buf_unmap_attachment(entry->a, entry->sgt, entry->dma_dir); entry 39 drivers/staging/media/tegra-vde/dmabuf-cache.c dma_buf_detach(dmabuf, entry->a); entry 42 drivers/staging/media/tegra-vde/dmabuf-cache.c list_del(&entry->list); entry 43 drivers/staging/media/tegra-vde/dmabuf-cache.c kfree(entry); entry 48 drivers/staging/media/tegra-vde/dmabuf-cache.c struct tegra_vde_cache_entry *entry; entry 51 drivers/staging/media/tegra-vde/dmabuf-cache.c entry = container_of(work, struct tegra_vde_cache_entry, entry 53 drivers/staging/media/tegra-vde/dmabuf-cache.c vde = entry->vde; entry 56 drivers/staging/media/tegra-vde/dmabuf-cache.c tegra_vde_release_entry(entry); entry 68 drivers/staging/media/tegra-vde/dmabuf-cache.c struct tegra_vde_cache_entry *entry; entry 75 drivers/staging/media/tegra-vde/dmabuf-cache.c list_for_each_entry(entry, &vde->map_list, list) { entry 76 drivers/staging/media/tegra-vde/dmabuf-cache.c if (entry->a->dmabuf != dmabuf) entry 79 drivers/staging/media/tegra-vde/dmabuf-cache.c if (!cancel_delayed_work(&entry->dwork)) entry 82 drivers/staging/media/tegra-vde/dmabuf-cache.c if (entry->dma_dir != dma_dir) entry 83 drivers/staging/media/tegra-vde/dmabuf-cache.c entry->dma_dir = DMA_BIDIRECTIONAL; entry 88 drivers/staging/media/tegra-vde/dmabuf-cache.c *addrp = iova_dma_addr(&vde->iova, entry->iova); entry 90 drivers/staging/media/tegra-vde/dmabuf-cache.c *addrp = sg_dma_address(entry->sgt->sgl); entry 115 drivers/staging/media/tegra-vde/dmabuf-cache.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 116 drivers/staging/media/tegra-vde/dmabuf-cache.c if (!entry) { entry 132 drivers/staging/media/tegra-vde/dmabuf-cache.c INIT_DELAYED_WORK(&entry->dwork, tegra_vde_delayed_unmap); entry 133 drivers/staging/media/tegra-vde/dmabuf-cache.c list_add(&entry->list, &vde->map_list); entry 135 drivers/staging/media/tegra-vde/dmabuf-cache.c entry->dma_dir = dma_dir; entry 136 drivers/staging/media/tegra-vde/dmabuf-cache.c entry->iova = iova; entry 137 drivers/staging/media/tegra-vde/dmabuf-cache.c entry->vde = vde; entry 138 drivers/staging/media/tegra-vde/dmabuf-cache.c entry->sgt = sgt; entry 139 drivers/staging/media/tegra-vde/dmabuf-cache.c entry->a = attachment; entry 141 drivers/staging/media/tegra-vde/dmabuf-cache.c entry->refcnt++; entry 143 drivers/staging/media/tegra-vde/dmabuf-cache.c *ap = entry->a; entry 150 drivers/staging/media/tegra-vde/dmabuf-cache.c kfree(entry); entry 165 drivers/staging/media/tegra-vde/dmabuf-cache.c struct tegra_vde_cache_entry *entry; entry 169 drivers/staging/media/tegra-vde/dmabuf-cache.c list_for_each_entry(entry, &vde->map_list, list) { entry 170 drivers/staging/media/tegra-vde/dmabuf-cache.c if (entry->a != a) entry 173 drivers/staging/media/tegra-vde/dmabuf-cache.c WARN_ON_ONCE(!entry->refcnt); entry 175 drivers/staging/media/tegra-vde/dmabuf-cache.c if (--entry->refcnt == 0) { entry 177 drivers/staging/media/tegra-vde/dmabuf-cache.c tegra_vde_release_entry(entry); entry 179 drivers/staging/media/tegra-vde/dmabuf-cache.c schedule_delayed_work(&entry->dwork, 5 * HZ); entry 189 drivers/staging/media/tegra-vde/dmabuf-cache.c struct tegra_vde_cache_entry *entry, *tmp; entry 193 drivers/staging/media/tegra-vde/dmabuf-cache.c list_for_each_entry_safe(entry, tmp, &vde->map_list, list) { entry 194 drivers/staging/media/tegra-vde/dmabuf-cache.c if (entry->refcnt) entry 197 drivers/staging/media/tegra-vde/dmabuf-cache.c if (!cancel_delayed_work(&entry->dwork)) entry 200 drivers/staging/media/tegra-vde/dmabuf-cache.c tegra_vde_release_entry(entry); entry 208 drivers/staging/media/tegra-vde/dmabuf-cache.c struct tegra_vde_cache_entry *entry, *tmp; entry 213 drivers/staging/media/tegra-vde/dmabuf-cache.c list_for_each_entry_safe(entry, tmp, &vde->map_list, list) { entry 214 drivers/staging/media/tegra-vde/dmabuf-cache.c if (!cancel_delayed_work(&entry->dwork)) entry 217 drivers/staging/media/tegra-vde/dmabuf-cache.c tegra_vde_release_entry(entry); entry 341 drivers/staging/octeon-usb/octeon-hcd.c } entry[MAX_CHANNELS + 1]; entry 1210 drivers/staging/octeon-usb/octeon-hcd.c const u32 *ptr = cvmx_phys_to_ptr(fifo->entry[i].address); entry 1211 drivers/staging/octeon-usb/octeon-hcd.c u64 csr_address = USB_FIFO_ADDRESS(fifo->entry[i].channel, entry 1216 drivers/staging/octeon-usb/octeon-hcd.c if (fifo->entry[i].size <= available) { entry 1217 drivers/staging/octeon-usb/octeon-hcd.c words = fifo->entry[i].size; entry 1225 drivers/staging/octeon-usb/octeon-hcd.c fifo->entry[i].address += words * 4; entry 1226 drivers/staging/octeon-usb/octeon-hcd.c fifo->entry[i].size -= words; entry 1327 drivers/staging/octeon-usb/octeon-hcd.c fifo->entry[fifo->head].channel = channel; entry 1328 drivers/staging/octeon-usb/octeon-hcd.c fifo->entry[fifo->head].address = entry 1331 drivers/staging/octeon-usb/octeon-hcd.c fifo->entry[fifo->head].size = (usbc_hctsiz.s.xfersize + 3) >> 2; entry 1611 drivers/staging/qlge/qlge_dbg.c DUMP_QDEV_ARRAY(qdev, "%d", msi_x_entry, i, entry); entry 3339 drivers/staging/qlge/qlge_main.c qdev->msi_x_entry[i].entry = i; entry 584 drivers/staging/rtl8188eu/core/rtw_cmd.c u8 rtw_clearstakey_cmd(struct adapter *padapter, u8 *psta, u8 entry, u8 enqueue) entry 594 drivers/staging/rtl8188eu/core/rtw_cmd.c clear_cam_entry(padapter, entry); entry 625 drivers/staging/rtl8188eu/core/rtw_cmd.c psetstakey_para->id = entry; entry 411 drivers/staging/rtl8188eu/core/rtw_wlan_util.c void write_cam(struct adapter *padapter, u8 entry, u16 ctrl, u8 *mac, u8 *key) entry 417 drivers/staging/rtl8188eu/core/rtw_wlan_util.c addr = entry << 3; entry 441 drivers/staging/rtl8188eu/core/rtw_wlan_util.c void clear_cam_entry(struct adapter *padapter, u8 entry) entry 446 drivers/staging/rtl8188eu/core/rtw_wlan_util.c write_cam(padapter, entry, 0, null_sta, null_key); entry 1378 drivers/staging/rtl8188eu/core/rtw_xmit.c struct xmit_frame *rtw_dequeue_xframe(struct xmit_priv *pxmitpriv, struct hw_xmit *phwxmit_i, int entry) entry 1400 drivers/staging/rtl8188eu/core/rtw_xmit.c for (i = 0; i < entry; i++) { entry 1538 drivers/staging/rtl8188eu/core/rtw_xmit.c void rtw_init_hwxmits(struct hw_xmit *phwxmit, int entry) entry 1542 drivers/staging/rtl8188eu/core/rtw_xmit.c for (i = 0; i < entry; i++, phwxmit++) entry 244 drivers/staging/rtl8188eu/hal/odm_rtl8188e.c struct sta_info *entry; entry 251 drivers/staging/rtl8188eu/hal/odm_rtl8188e.c entry = dm_odm->pODM_StaInfo[i]; entry 252 drivers/staging/rtl8188eu/hal/odm_rtl8188e.c if (IS_STA_VALID(entry)) { entry 305 drivers/staging/rtl8188eu/include/rtw_cmd.h u8 rtw_clearstakey_cmd(struct adapter *padapter, u8 *psta, u8 entry, entry 477 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h void write_cam(struct adapter *padapter, u8 entry, u16 ctrl, u8 *mac, u8 *key); entry 478 drivers/staging/rtl8188eu/include/rtw_mlme_ext.h void clear_cam_entry(struct adapter *padapter, u8 entry); entry 325 drivers/staging/rtl8188eu/include/rtw_xmit.h struct hw_xmit *phwxmit_i, int entry); entry 336 drivers/staging/rtl8188eu/include/rtw_xmit.h void rtw_init_hwxmits(struct hw_xmit *phwxmit, int entry); entry 1281 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c void rtl92e_fill_tx_cmd_desc(struct net_device *dev, struct tx_desc_cmd *entry, entry 1290 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c memset(entry, 0, 12); entry 1291 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c entry->LINIP = cb_desc->bLastIniPkt; entry 1292 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c entry->FirstSeg = 1; entry 1293 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c entry->LastSeg = 1; entry 1295 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c entry->CmdInit = DESC_PACKET_TYPE_INIT; entry 1297 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c struct tx_desc *entry_tmp = (struct tx_desc *)entry; entry 1307 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c entry->TxBufferSize = skb->len; entry 1308 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c entry->TxBuffAddr = mapping; entry 1309 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.c entry->OWN = 1; entry 32 drivers/staging/rtl8192e/rtl8192e/r8192E_dev.h void rtl92e_fill_tx_cmd_desc(struct net_device *dev, struct tx_desc_cmd *entry, entry 1583 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct tx_desc *entry = &ring->desc[ring->idx]; entry 1586 drivers/staging/rtl8192e/rtl8192e/rtl_core.c pci_unmap_single(priv->pdev, entry->TxBuffAddr, entry 1670 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct tx_desc *entry = &ring->desc[ring->idx]; entry 1674 drivers/staging/rtl8192e/rtl8192e/rtl_core.c if (entry->OWN) entry 1680 drivers/staging/rtl8192e/rtl8192e/rtl_core.c pci_unmap_single(priv->pdev, entry->TxBuffAddr, entry 1693 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct tx_desc_cmd *entry; entry 1702 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry = (struct tx_desc_cmd *)&ring->desc[idx]; entry 1706 drivers/staging/rtl8192e/rtl8192e/rtl_core.c priv->ops->tx_fill_cmd_descriptor(dev, entry, tcb_desc, skb); entry 1782 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct rx_desc *entry = NULL; entry 1801 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry = &priv->rx_ring[rx_queue_idx][i]; entry 1815 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry->BufferAddress = *mapping; entry 1817 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry->Length = priv->rxbuffersize; entry 1818 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry->OWN = 1; entry 1821 drivers/staging/rtl8192e/rtl8192e/rtl_core.c if (entry) entry 1822 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry->EOR = 1; entry 1889 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct rx_desc *entry = NULL; entry 1892 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry = &priv->rx_ring[rx_queue_idx][i]; entry 1893 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry->OWN = 1; entry 1905 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct tx_desc *entry = &ring->desc[ring->idx]; entry 1910 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry->TxBuffAddr, entry 283 drivers/staging/rtl8192e/rtl8192e/rtl_core.h struct tx_desc_cmd *entry, entry 61 drivers/staging/rtl8192e/rtllib_rx.c struct rtllib_frag_entry *entry; entry 65 drivers/staging/rtl8192e/rtllib_rx.c entry = &ieee->frag_cache[tid][i]; entry 66 drivers/staging/rtl8192e/rtllib_rx.c if (entry->skb != NULL && entry 67 drivers/staging/rtl8192e/rtllib_rx.c time_after(jiffies, entry->first_frag_time + 2 * HZ)) { entry 70 drivers/staging/rtl8192e/rtllib_rx.c entry->seq, entry->last_frag); entry 71 drivers/staging/rtl8192e/rtllib_rx.c dev_kfree_skb_any(entry->skb); entry 72 drivers/staging/rtl8192e/rtllib_rx.c entry->skb = NULL; entry 75 drivers/staging/rtl8192e/rtllib_rx.c if (entry->skb != NULL && entry->seq == seq && entry 76 drivers/staging/rtl8192e/rtllib_rx.c (entry->last_frag + 1 == frag || frag == -1) && entry 77 drivers/staging/rtl8192e/rtllib_rx.c memcmp(entry->src_addr, src, ETH_ALEN) == 0 && entry 78 drivers/staging/rtl8192e/rtllib_rx.c memcmp(entry->dst_addr, dst, ETH_ALEN) == 0) entry 79 drivers/staging/rtl8192e/rtllib_rx.c return entry; entry 95 drivers/staging/rtl8192e/rtllib_rx.c struct rtllib_frag_entry *entry; entry 128 drivers/staging/rtl8192e/rtllib_rx.c entry = &ieee->frag_cache[tid][ieee->frag_next_idx[tid]]; entry 133 drivers/staging/rtl8192e/rtllib_rx.c if (entry->skb != NULL) entry 134 drivers/staging/rtl8192e/rtllib_rx.c dev_kfree_skb_any(entry->skb); entry 136 drivers/staging/rtl8192e/rtllib_rx.c entry->first_frag_time = jiffies; entry 137 drivers/staging/rtl8192e/rtllib_rx.c entry->seq = seq; entry 138 drivers/staging/rtl8192e/rtllib_rx.c entry->last_frag = frag; entry 139 drivers/staging/rtl8192e/rtllib_rx.c entry->skb = skb; entry 140 drivers/staging/rtl8192e/rtllib_rx.c ether_addr_copy(entry->src_addr, hdr->addr2); entry 141 drivers/staging/rtl8192e/rtllib_rx.c ether_addr_copy(entry->dst_addr, hdr->addr1); entry 146 drivers/staging/rtl8192e/rtllib_rx.c entry = rtllib_frag_cache_find(ieee, seq, frag, tid, hdr->addr2, entry 148 drivers/staging/rtl8192e/rtllib_rx.c if (entry != NULL) { entry 149 drivers/staging/rtl8192e/rtllib_rx.c entry->last_frag = frag; entry 150 drivers/staging/rtl8192e/rtllib_rx.c skb = entry->skb; entry 165 drivers/staging/rtl8192e/rtllib_rx.c struct rtllib_frag_entry *entry; entry 185 drivers/staging/rtl8192e/rtllib_rx.c entry = rtllib_frag_cache_find(ieee, seq, -1, tid, hdr->addr2, entry 188 drivers/staging/rtl8192e/rtllib_rx.c if (entry == NULL) { entry 195 drivers/staging/rtl8192e/rtllib_rx.c entry->skb = NULL; entry 393 drivers/staging/rtl8192e/rtllib_rx.c struct ieee_ibss_seq *entry = NULL; entry 398 drivers/staging/rtl8192e/rtllib_rx.c entry = list_entry(p, struct ieee_ibss_seq, list); entry 399 drivers/staging/rtl8192e/rtllib_rx.c if (!memcmp(entry->mac, mac, ETH_ALEN)) entry 403 drivers/staging/rtl8192e/rtllib_rx.c entry = kmalloc(sizeof(struct ieee_ibss_seq), entry 405 drivers/staging/rtl8192e/rtllib_rx.c if (!entry) entry 408 drivers/staging/rtl8192e/rtllib_rx.c ether_addr_copy(entry->mac, mac); entry 409 drivers/staging/rtl8192e/rtllib_rx.c entry->seq_num[tid] = seq; entry 410 drivers/staging/rtl8192e/rtllib_rx.c entry->frag_num[tid] = frag; entry 411 drivers/staging/rtl8192e/rtllib_rx.c entry->packet_time[tid] = jiffies; entry 412 drivers/staging/rtl8192e/rtllib_rx.c list_add(&entry->list, &ieee->ibss_mac_hash[index]); entry 415 drivers/staging/rtl8192e/rtllib_rx.c last_seq = &entry->seq_num[tid]; entry 416 drivers/staging/rtl8192e/rtllib_rx.c last_frag = &entry->frag_num[tid]; entry 417 drivers/staging/rtl8192e/rtllib_rx.c last_time = &entry->packet_time[tid]; entry 38 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c struct ieee80211_crypt_data *entry; entry 42 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c entry = list_entry(ptr, struct ieee80211_crypt_data, list); entry 44 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c if (atomic_read(&entry->refcnt) != 0 && !force) entry 49 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c if (entry->ops) entry 50 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c entry->ops->deinit(entry->priv); entry 51 drivers/staging/rtl8192u/ieee80211/ieee80211_crypt.c kfree(entry); entry 65 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c struct ieee80211_frag_entry *entry; entry 69 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry = &ieee->frag_cache[tid][i]; entry 70 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (entry->skb && entry 71 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c time_after(jiffies, entry->first_frag_time + 2 * HZ)) { entry 75 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->seq, entry->last_frag); entry 76 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c dev_kfree_skb_any(entry->skb); entry 77 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->skb = NULL; entry 80 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (entry->skb && entry->seq == seq && entry 81 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c (entry->last_frag + 1 == frag || frag == -1) && entry 82 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c memcmp(entry->src_addr, src, ETH_ALEN) == 0 && entry 83 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c memcmp(entry->dst_addr, dst, ETH_ALEN) == 0) entry 84 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c return entry; entry 100 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c struct ieee80211_frag_entry *entry; entry 131 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry = &ieee->frag_cache[tid][ieee->frag_next_idx[tid]]; entry 136 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (entry->skb) entry 137 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c dev_kfree_skb_any(entry->skb); entry 139 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->first_frag_time = jiffies; entry 140 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->seq = seq; entry 141 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->last_frag = frag; entry 142 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->skb = skb; entry 143 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c memcpy(entry->src_addr, hdr->addr2, ETH_ALEN); entry 144 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c memcpy(entry->dst_addr, hdr->addr1, ETH_ALEN); entry 148 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry = ieee80211_frag_cache_find(ieee, seq, frag, tid, hdr->addr2, entry 150 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (entry) { entry 151 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->last_frag = frag; entry 152 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c skb = entry->skb; entry 167 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c struct ieee80211_frag_entry *entry; entry 186 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry = ieee80211_frag_cache_find(ieee, seq, -1, tid, hdr->addr2, entry 189 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (!entry) { entry 196 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->skb = NULL; entry 445 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c struct ieee_ibss_seq *entry = NULL; entry 450 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry = list_entry(p, struct ieee_ibss_seq, list); entry 451 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (!memcmp(entry->mac, mac, ETH_ALEN)) entry 456 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry = kmalloc(sizeof(struct ieee_ibss_seq), GFP_ATOMIC); entry 457 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c if (!entry) entry 459 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c memcpy(entry->mac, mac, ETH_ALEN); entry 460 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->seq_num[tid] = seq; entry 461 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->frag_num[tid] = frag; entry 462 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c entry->packet_time[tid] = jiffies; entry 463 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c list_add(&entry->list, &ieee->ibss_mac_hash[index]); entry 466 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c last_seq = &entry->seq_num[tid]; entry 467 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c last_frag = &entry->frag_num[tid]; entry 468 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c last_time = &entry->packet_time[tid]; entry 719 drivers/staging/rtl8192u/r8192U_core.c struct urb *entry; entry 728 drivers/staging/rtl8192u/r8192U_core.c entry = usb_alloc_urb(0, GFP_KERNEL); entry 729 drivers/staging/rtl8192u/r8192U_core.c if (!entry) { entry 733 drivers/staging/rtl8192u/r8192U_core.c usb_fill_bulk_urb(entry, priv->udev, entry 738 drivers/staging/rtl8192u/r8192U_core.c info->urb = entry; entry 742 drivers/staging/rtl8192u/r8192U_core.c usb_submit_urb(entry, GFP_KERNEL); entry 750 drivers/staging/rtl8192u/r8192U_core.c entry = usb_alloc_urb(0, GFP_KERNEL); entry 751 drivers/staging/rtl8192u/r8192U_core.c if (!entry) { entry 755 drivers/staging/rtl8192u/r8192U_core.c usb_fill_bulk_urb(entry, priv->udev, entry 760 drivers/staging/rtl8192u/r8192U_core.c info->urb = entry; entry 764 drivers/staging/rtl8192u/r8192U_core.c usb_submit_urb(entry, GFP_KERNEL); entry 170 drivers/staging/rtl8712/rtl8712_xmit.c struct hw_xmit *phwxmit_i, sint entry) entry 195 drivers/staging/rtl8712/rtl8712_xmit.c for (i = 0; i < entry; i++) { entry 28 drivers/staging/rtl8712/rtl871x_xmit.c static void init_hwxmits(struct hw_xmit *phwxmit, sint entry); entry 984 drivers/staging/rtl8712/rtl871x_xmit.c static void init_hwxmits(struct hw_xmit *phwxmit, sint entry) entry 988 drivers/staging/rtl8712/rtl871x_xmit.c for (i = 0; i < entry; i++, phwxmit++) { entry 545 drivers/staging/rtl8723bs/core/rtw_wlan_util.c void read_cam(struct adapter *padapter, u8 entry, u8 *get_key) entry 548 drivers/staging/rtl8723bs/core/rtw_wlan_util.c addr = entry << 3; entry 560 drivers/staging/rtl8723bs/core/rtw_wlan_util.c void _write_cam(struct adapter *padapter, u8 entry, u16 ctrl, u8 *mac, u8 *key) entry 566 drivers/staging/rtl8723bs/core/rtw_wlan_util.c addr = entry << 3; entry 589 drivers/staging/rtl8723bs/core/rtw_wlan_util.c void _clear_cam_entry(struct adapter *padapter, u8 entry) entry 594 drivers/staging/rtl8723bs/core/rtw_wlan_util.c _write_cam(padapter, entry, 0, null_sta, null_key); entry 2218 drivers/staging/rtl8723bs/core/rtw_xmit.c void rtw_init_hwxmits(struct hw_xmit *phwxmit, sint entry) entry 2222 drivers/staging/rtl8723bs/core/rtw_xmit.c for (i = 0; i < entry; i++, phwxmit++) { entry 574 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h void read_cam(struct adapter *padapter , u8 entry, u8 *get_key); entry 577 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h void _write_cam(struct adapter *padapter, u8 entry, u16 ctrl, u8 *mac, u8 *key); entry 578 drivers/staging/rtl8723bs/include/rtw_mlme_ext.h void _clear_cam_entry(struct adapter *padapter, u8 entry); entry 483 drivers/staging/rtl8723bs/include/rtw_xmit.h void rtw_init_hwxmits(struct hw_xmit *phwxmit, sint entry); entry 150 drivers/staging/unisys/visorhba/visorhba_main.c struct scsipending *entry; entry 163 drivers/staging/unisys/visorhba/visorhba_main.c entry = &devdata->pending[insert_location]; entry 164 drivers/staging/unisys/visorhba/visorhba_main.c memset(&entry->cmdrsp, 0, sizeof(entry->cmdrsp)); entry 165 drivers/staging/unisys/visorhba/visorhba_main.c entry->cmdtype = cmdtype; entry 167 drivers/staging/unisys/visorhba/visorhba_main.c entry->sent = new; entry 170 drivers/staging/unisys/visorhba/visorhba_main.c entry->sent = &entry->cmdrsp; entry 38 drivers/staging/uwb/est.c const struct uwb_est_entry *entry; entry 243 drivers/staging/uwb/est.c const struct uwb_est_entry *entry, size_t entries) entry 268 drivers/staging/uwb/est.c uwb_est[itr].entry = entry; entry 292 drivers/staging/uwb/est.c const struct uwb_est_entry *entry, size_t entries) entry 300 drivers/staging/uwb/est.c .entry = entry, entry 345 drivers/staging/uwb/est.c const struct uwb_est_entry *entry; entry 355 drivers/staging/uwb/est.c entry = &est->entry[event_low]; entry 356 drivers/staging/uwb/est.c if (entry->size == 0 && entry->offset == 0) { /* unknown? */ entry 362 drivers/staging/uwb/est.c offset = entry->offset; /* extra fries with that? */ entry 364 drivers/staging/uwb/est.c size = entry->size; entry 371 drivers/staging/uwb/est.c switch (entry->type) { entry 383 drivers/staging/uwb/est.c size = entry->size; entry 385 drivers/staging/uwb/est.c switch (entry->type) { entry 48 drivers/staging/uwb/lc-dev.c int uwb_notifs_deregister(struct uwb_rc *rc, struct uwb_notifs_handler *entry) entry 52 drivers/staging/uwb/lc-dev.c list_del(&entry->list_node); entry 27 drivers/staging/vt6655/key.c u32 entry = 0; entry 43 drivers/staging/vt6655/key.c entry = key->hw_key_idx; entry 52 drivers/staging/vt6655/key.c entry = MAX_KEY_TABLE - 1; entry 53 drivers/staging/vt6655/key.c key->hw_key_idx = entry; entry 85 drivers/staging/vt6655/key.c MACvSetKeyEntry(priv, key_mode, entry, key_inx, entry 43 drivers/staging/vt6656/key.c u32 entry = 0; entry 59 drivers/staging/vt6656/key.c entry = key->hw_key_idx; entry 68 drivers/staging/vt6656/key.c entry = MAX_KEY_TABLE - 1; entry 69 drivers/staging/vt6656/key.c key->hw_key_idx = entry; entry 101 drivers/staging/vt6656/key.c vnt_mac_set_keyentry(priv, key_mode, entry, key_inx, bssid, key->key); entry 960 drivers/target/target_core_user.c struct tcmu_cmd_entry *entry; entry 1023 drivers/target/target_core_user.c entry = (void *) mb + CMDR_OFF + cmd_head; entry 1024 drivers/target/target_core_user.c tcmu_hdr_set_op(&entry->hdr.len_op, TCMU_OP_PAD); entry 1025 drivers/target/target_core_user.c tcmu_hdr_set_len(&entry->hdr.len_op, pad_size); entry 1026 drivers/target/target_core_user.c entry->hdr.cmd_id = 0; /* not used for PAD */ entry 1027 drivers/target/target_core_user.c entry->hdr.kflags = 0; entry 1028 drivers/target/target_core_user.c entry->hdr.uflags = 0; entry 1029 drivers/target/target_core_user.c tcmu_flush_dcache_range(entry, sizeof(*entry)); entry 1038 drivers/target/target_core_user.c entry = (void *) mb + CMDR_OFF + cmd_head; entry 1039 drivers/target/target_core_user.c memset(entry, 0, command_size); entry 1040 drivers/target/target_core_user.c tcmu_hdr_set_op(&entry->hdr.len_op, TCMU_OP_CMD); entry 1044 drivers/target/target_core_user.c iov = &entry->req.iov[0]; entry 1051 drivers/target/target_core_user.c entry->req.iov_cnt = iov_cnt; entry 1061 drivers/target/target_core_user.c entry->req.iov_bidi_cnt = iov_cnt; entry 1071 drivers/target/target_core_user.c entry->hdr.cmd_id = tcmu_cmd->cmd_id; entry 1077 drivers/target/target_core_user.c base_command_size = tcmu_cmd_get_base_cmd_size(entry->req.iov_cnt + entry 1078 drivers/target/target_core_user.c entry->req.iov_bidi_cnt); entry 1081 drivers/target/target_core_user.c tcmu_hdr_set_len(&entry->hdr.len_op, command_size); entry 1086 drivers/target/target_core_user.c entry->req.cdb_off = cdb_off; entry 1087 drivers/target/target_core_user.c tcmu_flush_dcache_range(entry, sizeof(*entry)); entry 1130 drivers/target/target_core_user.c static void tcmu_handle_completion(struct tcmu_cmd *cmd, struct tcmu_cmd_entry *entry) entry 1150 drivers/target/target_core_user.c if (entry->hdr.uflags & TCMU_UFLAG_UNKNOWN_OP) { entry 1153 drivers/target/target_core_user.c entry->rsp.scsi_status = SAM_STAT_CHECK_CONDITION; entry 1159 drivers/target/target_core_user.c (entry->hdr.uflags & TCMU_UFLAG_READ_LEN) && entry->rsp.read_len) { entry 1161 drivers/target/target_core_user.c if (entry->rsp.read_len < read_len) entry 1162 drivers/target/target_core_user.c read_len = entry->rsp.read_len; entry 1165 drivers/target/target_core_user.c if (entry->rsp.scsi_status == SAM_STAT_CHECK_CONDITION) { entry 1166 drivers/target/target_core_user.c transport_copy_sense_to_cmd(se_cmd, entry->rsp.sense_buffer); entry 1188 drivers/target/target_core_user.c entry->rsp.scsi_status, read_len); entry 1190 drivers/target/target_core_user.c target_complete_cmd(cmd->se_cmd, entry->rsp.scsi_status); entry 1233 drivers/target/target_core_user.c struct tcmu_cmd_entry *entry = (void *) mb + CMDR_OFF + udev->cmdr_last_cleaned; entry 1235 drivers/target/target_core_user.c tcmu_flush_dcache_range(entry, sizeof(*entry)); entry 1237 drivers/target/target_core_user.c if (tcmu_hdr_get_op(entry->hdr.len_op) == TCMU_OP_PAD) { entry 1239 drivers/target/target_core_user.c tcmu_hdr_get_len(entry->hdr.len_op), entry 1243 drivers/target/target_core_user.c WARN_ON(tcmu_hdr_get_op(entry->hdr.len_op) != TCMU_OP_CMD); entry 1245 drivers/target/target_core_user.c cmd = idr_remove(&udev->commands, entry->hdr.cmd_id); entry 1248 drivers/target/target_core_user.c entry->hdr.cmd_id); entry 1253 drivers/target/target_core_user.c tcmu_handle_completion(cmd, entry); entry 1256 drivers/target/target_core_user.c tcmu_hdr_get_len(entry->hdr.len_op), entry 308 drivers/thunderbolt/eeprom.c const struct tb_drom_entry_generic *entry = entry 314 drivers/thunderbolt/eeprom.c sw->vendor_name = kstrndup(entry->data, entry 321 drivers/thunderbolt/eeprom.c sw->device_name = kstrndup(entry->data, entry 358 drivers/thunderbolt/eeprom.c struct tb_drom_entry_port *entry = (void *) header; entry 359 drivers/thunderbolt/eeprom.c if (header->len != sizeof(*entry)) { entry 365 drivers/thunderbolt/eeprom.c port->link_nr = entry->link_nr; entry 366 drivers/thunderbolt/eeprom.c if (entry->has_dual_link_port) entry 368 drivers/thunderbolt/eeprom.c &port->sw->ports[entry->dual_link_port_nr]; entry 386 drivers/thunderbolt/eeprom.c struct tb_drom_entry_header *entry = (void *) (sw->drom + pos); entry 387 drivers/thunderbolt/eeprom.c if (pos + 1 == drom_size || pos + entry->len > drom_size entry 388 drivers/thunderbolt/eeprom.c || !entry->len) { entry 393 drivers/thunderbolt/eeprom.c switch (entry->type) { entry 395 drivers/thunderbolt/eeprom.c res = tb_drom_parse_entry_generic(sw, entry); entry 398 drivers/thunderbolt/eeprom.c res = tb_drom_parse_entry_port(sw, entry); entry 404 drivers/thunderbolt/eeprom.c pos += entry->len; entry 52 drivers/thunderbolt/property.c static bool tb_property_entry_valid(const struct tb_property_entry *entry, entry 55 drivers/thunderbolt/property.c switch (entry->type) { entry 59 drivers/thunderbolt/property.c if (entry->length > block_len) entry 61 drivers/thunderbolt/property.c if (entry->value + entry->length > block_len) entry 66 drivers/thunderbolt/property.c if (entry->length != 1) entry 96 drivers/thunderbolt/property.c const struct tb_property_entry *entry) entry 102 drivers/thunderbolt/property.c if (!tb_property_entry_valid(entry, block_len)) entry 105 drivers/thunderbolt/property.c parse_dwdata(key, entry, 2); entry 108 drivers/thunderbolt/property.c property = tb_property_alloc(key, entry->type); entry 112 drivers/thunderbolt/property.c property->length = entry->length; entry 116 drivers/thunderbolt/property.c dir = __tb_property_parse_dir(block, block_len, entry->value, entry 117 drivers/thunderbolt/property.c entry->length, false); entry 132 drivers/thunderbolt/property.c parse_dwdata(property->value.data, block + entry->value, entry 133 drivers/thunderbolt/property.c entry->length); entry 143 drivers/thunderbolt/property.c parse_dwdata(property->value.text, block + entry->value, entry 144 drivers/thunderbolt/property.c entry->length); entry 150 drivers/thunderbolt/property.c property->value.immediate = entry->value; entry 353 drivers/thunderbolt/property.c struct tb_property_entry *entry; entry 417 drivers/thunderbolt/property.c entry = pe->entries; entry 424 drivers/thunderbolt/property.c entry = re->entries; entry 430 drivers/thunderbolt/property.c format_dwdata(entry, property->key, 2); entry 431 drivers/thunderbolt/property.c entry->type = property->type; entry 440 drivers/thunderbolt/property.c entry->length = tb_property_dir_length(child, false, entry 442 drivers/thunderbolt/property.c entry->value = dir_end; entry 449 drivers/thunderbolt/property.c entry->length = property->length; entry 450 drivers/thunderbolt/property.c entry->value = data_offset; entry 451 drivers/thunderbolt/property.c data_offset += entry->length; entry 457 drivers/thunderbolt/property.c entry->length = property->length; entry 458 drivers/thunderbolt/property.c entry->value = data_offset; entry 459 drivers/thunderbolt/property.c data_offset += entry->length; entry 463 drivers/thunderbolt/property.c entry->length = property->length; entry 464 drivers/thunderbolt/property.c entry->value = property->value.immediate; entry 471 drivers/thunderbolt/property.c entry++; entry 53 drivers/tty/hvc/hvc_xen.c struct xencons_info *entry, *n, *ret = NULL; entry 58 drivers/tty/hvc/hvc_xen.c list_for_each_entry_safe(entry, n, &xenconsoles, list) { entry 59 drivers/tty/hvc/hvc_xen.c if (entry->vtermno == vtermno) { entry 60 drivers/tty/hvc/hvc_xen.c ret = entry; entry 1038 drivers/tty/serial/msm_serial.c const struct msm_baud_map *entry, *end, *best; entry 1063 drivers/tty/serial/msm_serial.c entry = table; entry 1064 drivers/tty/serial/msm_serial.c while (entry < end) { entry 1065 drivers/tty/serial/msm_serial.c if (entry->divisor <= divisor) { entry 1066 drivers/tty/serial/msm_serial.c result = target / entry->divisor / 16; entry 1072 drivers/tty/serial/msm_serial.c best = entry; entry 1078 drivers/tty/serial/msm_serial.c } else if (entry->divisor > divisor) { entry 1089 drivers/tty/serial/msm_serial.c entry = table; entry 1093 drivers/tty/serial/msm_serial.c entry++; entry 1105 drivers/tty/serial/msm_serial.c const struct msm_baud_map *entry; entry 1111 drivers/tty/serial/msm_serial.c entry = msm_find_best_baud(port, baud, &rate); entry 1113 drivers/tty/serial/msm_serial.c baud = rate / 16 / entry->divisor; entry 1119 drivers/tty/serial/msm_serial.c msm_write(port, entry->code, UART_CSR); entry 1122 drivers/tty/serial/msm_serial.c rxstale = entry->rxstale; entry 105 drivers/uio/uio.c struct map_sysfs_entry *entry; entry 107 drivers/uio/uio.c entry = container_of(attr, struct map_sysfs_entry, attr); entry 109 drivers/uio/uio.c if (!entry->show) entry 112 drivers/uio/uio.c return entry->show(mem, buf); entry 193 drivers/uio/uio.c struct portio_sysfs_entry *entry; entry 195 drivers/uio/uio.c entry = container_of(attr, struct portio_sysfs_entry, attr); entry 197 drivers/uio/uio.c if (!entry->show) entry 200 drivers/uio/uio.c return entry->show(port, buf); entry 229 drivers/usb/early/xhci-dbc.c struct xdbc_erst_entry *entry; entry 249 drivers/usb/early/xhci-dbc.c entry = (struct xdbc_erst_entry *)xdbc.erst_base; entry 251 drivers/usb/early/xhci-dbc.c entry->seg_addr = cpu_to_le64(xdbc.evt_seg.dma); entry 252 drivers/usb/early/xhci-dbc.c entry->seg_size = cpu_to_le32(XDBC_TRBS_PER_SEGMENT); entry 253 drivers/usb/early/xhci-dbc.c entry->__reserved_0 = 0; entry 1537 drivers/usb/gadget/composite.c list_for_each_entry(ext_prop, &d->ext_prop, entry) { entry 1062 drivers/usb/gadget/configfs.c list_add_tail(&ext_prop->entry, &desc->ext_prop); entry 1077 drivers/usb/gadget/configfs.c list_del(&ext_prop->entry); entry 3033 drivers/usb/gadget/function/f_fs.c list_add_tail(&ext_prop->entry, &t->os_desc->ext_prop); entry 3452 drivers/usb/gadget/function/f_fs.c list_for_each_entry(dev, &ffs_devices, entry) { entry 3468 drivers/usb/gadget/function/f_fs.c dev = list_first_entry(&ffs_devices, struct ffs_dev, entry); entry 3657 drivers/usb/gadget/function/f_fs.c list_add(&dev->entry, &ffs_devices); entry 3703 drivers/usb/gadget/function/f_fs.c list_del(&dev->entry); entry 42 drivers/usb/gadget/function/u_fs.h struct list_head entry; entry 850 drivers/usb/gadget/function/uvc_configfs.c struct list_head entry; entry 929 drivers/usb/gadget/function/uvc_configfs.c list_add_tail(&format_ptr->entry, &src_hdr->formats); entry 961 drivers/usb/gadget/function/uvc_configfs.c list_for_each_entry_safe(format_ptr, tmp, &src_hdr->formats, entry) entry 963 drivers/usb/gadget/function/uvc_configfs.c list_del(&format_ptr->entry); entry 1999 drivers/usb/gadget/function/uvc_configfs.c list_for_each_entry(f, &h->formats, entry) { entry 2098 drivers/usb/gadget/function/uvc_configfs.c list_for_each_entry(f, &h->formats, entry) { entry 221 drivers/usb/gadget/legacy/hid.c struct hidg_func_node *entry; entry 228 drivers/usb/gadget/legacy/hid.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 229 drivers/usb/gadget/legacy/hid.c if (!entry) entry 232 drivers/usb/gadget/legacy/hid.c entry->func = func; entry 233 drivers/usb/gadget/legacy/hid.c list_add_tail(&entry->node, &hidg_func_list); entry 375 drivers/usb/host/ehci-dbg.c struct list_head *entry; entry 416 drivers/usb/host/ehci-dbg.c list_for_each(entry, &qh->qtd_list) { entry 419 drivers/usb/host/ehci-dbg.c td = list_entry(entry, struct ehci_qtd, qtd_list); entry 292 drivers/usb/host/ehci-q.c struct list_head *entry, *tmp; entry 322 drivers/usb/host/ehci-q.c list_for_each_safe (entry, tmp, &qh->qtd_list) { entry 327 drivers/usb/host/ehci-q.c qtd = list_entry (entry, struct ehci_qtd, qtd_list); entry 560 drivers/usb/host/ehci-q.c struct list_head *entry, *temp; entry 562 drivers/usb/host/ehci-q.c list_for_each_safe (entry, temp, qtd_list) { entry 565 drivers/usb/host/ehci-q.c qtd = list_entry (entry, struct ehci_qtd, qtd_list); entry 415 drivers/usb/host/ohci-dbg.c struct list_head *entry; entry 433 drivers/usb/host/ohci-dbg.c list_for_each (entry, &ed->td_list) { entry 436 drivers/usb/host/ohci-dbg.c td = list_entry (entry, struct td, td_list); entry 532 drivers/usb/host/ohci-dbg.c struct list_head *entry; entry 536 drivers/usb/host/ohci-dbg.c list_for_each (entry, &ed->td_list) entry 982 drivers/usb/host/ohci-q.c struct list_head *entry, *tmp; entry 1037 drivers/usb/host/ohci-q.c list_for_each_safe (entry, tmp, &ed->td_list) { entry 1044 drivers/usb/host/ohci-q.c td = list_entry (entry, struct td, td_list); entry 1805 drivers/usb/host/xhci-mem.c struct xhci_erst_entry *entry; entry 1817 drivers/usb/host/xhci-mem.c entry = &erst->entries[val]; entry 1818 drivers/usb/host/xhci-mem.c entry->seg_addr = cpu_to_le64(seg->dma); entry 1819 drivers/usb/host/xhci-mem.c entry->seg_size = cpu_to_le32(TRBS_PER_SEGMENT); entry 1820 drivers/usb/host/xhci-mem.c entry->rsvd = 0; entry 334 drivers/usb/storage/ene_ub6250.c struct ms_bootblock_sysent_rec entry[MS_NUMBER_OF_SYSTEM_ENTRY]; entry 992 drivers/usb/storage/ene_ub6250.c EntryOffset = be32_to_cpu(SysEntry->entry[i].dwStart); entry 996 drivers/usb/storage/ene_ub6250.c EntrySize = be32_to_cpu(SysEntry->entry[i].dwSize); entry 1008 drivers/usb/storage/ene_ub6250.c if (SysEntry->entry[i].bType != MS_SYSENT_TYPE_INVALID_BLOCK) entry 1038 drivers/usb/storage/ene_ub6250.c if (SysEntry->entry[i].bType != MS_SYSENT_TYPE_CIS_IDI) entry 210 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long entry = ioba >> tbl->it_page_shift; entry 214 drivers/vfio/vfio_iommu_spapr_tce.c if ((start <= entry) && (entry < end)) { entry 338 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long entry, unsigned long pages); entry 410 drivers/vfio/vfio_iommu_spapr_tce.c struct iommu_table *tbl, unsigned long entry) entry 415 drivers/vfio/vfio_iommu_spapr_tce.c __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry); entry 424 drivers/vfio/vfio_iommu_spapr_tce.c __func__, be64_to_cpu(*pua), entry, ret); entry 433 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long entry, unsigned long pages) entry 438 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long lastentry = entry + pages, firstentry = entry; entry 440 drivers/vfio/vfio_iommu_spapr_tce.c for ( ; entry < lastentry; ++entry) { entry 451 drivers/vfio/vfio_iommu_spapr_tce.c entry); entry 454 drivers/vfio/vfio_iommu_spapr_tce.c entry |= tbl->it_level_size - 1; entry 463 drivers/vfio/vfio_iommu_spapr_tce.c ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry, &oldhpa, entry 472 drivers/vfio/vfio_iommu_spapr_tce.c tce_iommu_unuse_page_v2(container, tbl, entry); entry 501 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long entry, unsigned long tce, unsigned long pages, entry 523 drivers/vfio/vfio_iommu_spapr_tce.c ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry + i, entry 528 drivers/vfio/vfio_iommu_spapr_tce.c __func__, entry << tbl->it_page_shift, entry 540 drivers/vfio/vfio_iommu_spapr_tce.c tce_iommu_clear(container, tbl, entry, i); entry 542 drivers/vfio/vfio_iommu_spapr_tce.c iommu_tce_kill(tbl, entry, pages); entry 549 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long entry, unsigned long tce, unsigned long pages, entry 558 drivers/vfio/vfio_iommu_spapr_tce.c __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry + i); entry 579 drivers/vfio/vfio_iommu_spapr_tce.c ret = iommu_tce_xchg_no_kill(container->mm, tbl, entry + i, entry 583 drivers/vfio/vfio_iommu_spapr_tce.c tce_iommu_unuse_page_v2(container, tbl, entry + i); entry 585 drivers/vfio/vfio_iommu_spapr_tce.c __func__, entry << tbl->it_page_shift, entry 591 drivers/vfio/vfio_iommu_spapr_tce.c tce_iommu_unuse_page_v2(container, tbl, entry + i); entry 599 drivers/vfio/vfio_iommu_spapr_tce.c tce_iommu_clear(container, tbl, entry, i); entry 601 drivers/vfio/vfio_iommu_spapr_tce.c iommu_tce_kill(tbl, entry, pages); entry 666 drivers/vfio/vfio_iommu_type1.c struct vfio_regions *entry, *next; entry 670 drivers/vfio/vfio_iommu_type1.c list_for_each_entry_safe(entry, next, regions, list) { entry 672 drivers/vfio/vfio_iommu_type1.c entry->iova, entry 673 drivers/vfio/vfio_iommu_type1.c entry->phys >> PAGE_SHIFT, entry 674 drivers/vfio/vfio_iommu_type1.c entry->len >> PAGE_SHIFT, entry 676 drivers/vfio/vfio_iommu_type1.c list_del(&entry->list); entry 677 drivers/vfio/vfio_iommu_type1.c kfree(entry); entry 702 drivers/vfio/vfio_iommu_type1.c struct vfio_regions *entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 704 drivers/vfio/vfio_iommu_type1.c if (entry) { entry 709 drivers/vfio/vfio_iommu_type1.c kfree(entry); entry 711 drivers/vfio/vfio_iommu_type1.c entry->iova = *iova; entry 712 drivers/vfio/vfio_iommu_type1.c entry->phys = phys; entry 713 drivers/vfio/vfio_iommu_type1.c entry->len = unmapped; entry 714 drivers/vfio/vfio_iommu_type1.c list_add_tail(&entry->list, unmapped_list); entry 179 drivers/video/backlight/arcxcnn_bl.c u32 prog_val, num_entry, entry, sources[ARCXCNN_LEDEN_BITS]; entry 236 drivers/video/backlight/arcxcnn_bl.c for (entry = 0; entry < num_entry; entry++) { entry 237 drivers/video/backlight/arcxcnn_bl.c u8 onbit = 1 << sources[entry]; entry 407 drivers/video/backlight/backlight.c list_add(&new_bd->entry, &backlight_dev_list); entry 423 drivers/video/backlight/backlight.c list_for_each_entry(bd, &backlight_dev_list, entry) { entry 447 drivers/video/backlight/backlight.c list_del(&bd->entry); entry 83 drivers/video/fbdev/i810/i810_main.c const struct pci_device_id *entry); entry 1839 drivers/video/fbdev/i810/i810_main.c const struct pci_device_id *entry) entry 1865 drivers/video/fbdev/i810/i810_main.c i810_pci_list[entry->driver_data])) { entry 1880 drivers/video/fbdev/i810/i810_main.c i810_pci_list[entry->driver_data])) { entry 2012 drivers/video/fbdev/i810/i810_main.c const struct pci_device_id *entry) entry 2035 drivers/video/fbdev/i810/i810_main.c if ((err = i810_allocate_pci_resource(par, entry))) { entry 2085 drivers/video/fbdev/i810/i810_main.c i810_pci_list[entry->driver_data], entry 77 drivers/video/fbdev/omap/hwa742.c struct list_head entry; entry 241 drivers/video/fbdev/omap/hwa742.c struct hwa742_request, entry); entry 242 drivers/video/fbdev/omap/hwa742.c list_del(&req->entry); entry 245 drivers/video/fbdev/omap/hwa742.c INIT_LIST_HEAD(&req->entry); entry 257 drivers/video/fbdev/omap/hwa742.c list_move(&req->entry, &hwa742.free_req_list); entry 276 drivers/video/fbdev/omap/hwa742.c struct hwa742_request, entry); entry 412 drivers/video/fbdev/omap/hwa742.c list_add_tail(&req->entry, req_head); \ entry 470 drivers/video/fbdev/omap/hwa742.c last = list_entry(req_list.prev, struct hwa742_request, entry); entry 501 drivers/video/fbdev/omap/hwa742.c last = list_entry(req_list.prev, struct hwa742_request, entry); entry 554 drivers/video/fbdev/omap/hwa742.c list_add(&req->entry, &req_list); entry 1002 drivers/video/fbdev/omap/hwa742.c list_add(&hwa742.req_pool[i].entry, &hwa742.free_req_list); entry 77 drivers/video/fbdev/smscufx.c struct list_head entry; entry 1796 drivers/video/fbdev/smscufx.c list_add_tail(&unode->entry, &dev->urbs.list); entry 1833 drivers/video/fbdev/smscufx.c unode = list_entry(node, struct urb_node, entry); entry 1885 drivers/video/fbdev/smscufx.c list_add_tail(&unode->entry, &dev->urbs.list); entry 1902 drivers/video/fbdev/smscufx.c struct list_head *entry; entry 1919 drivers/video/fbdev/smscufx.c entry = dev->urbs.list.next; entry 1920 drivers/video/fbdev/smscufx.c list_del_init(entry); entry 1925 drivers/video/fbdev/smscufx.c unode = list_entry(entry, struct urb_node, entry); entry 1837 drivers/video/fbdev/udlfb.c list_add_tail(&unode->entry, &dlfb->urbs.list); entry 1862 drivers/video/fbdev/udlfb.c unode = list_entry(node, struct urb_node, entry); entry 1923 drivers/video/fbdev/udlfb.c list_add_tail(&unode->entry, &dlfb->urbs.list); entry 1936 drivers/video/fbdev/udlfb.c struct list_head *entry; entry 1952 drivers/video/fbdev/udlfb.c entry = dlfb->urbs.list.next; entry 1953 drivers/video/fbdev/udlfb.c list_del_init(entry); entry 1958 drivers/video/fbdev/udlfb.c unode = list_entry(entry, struct urb_node, entry); entry 991 drivers/video/fbdev/uvesafb.c struct uvesafb_pal_entry entry; entry 999 drivers/video/fbdev/uvesafb.c entry.red = red >> shift; entry 1000 drivers/video/fbdev/uvesafb.c entry.green = green >> shift; entry 1001 drivers/video/fbdev/uvesafb.c entry.blue = blue >> shift; entry 1002 drivers/video/fbdev/uvesafb.c entry.pad = 0; entry 1004 drivers/video/fbdev/uvesafb.c err = uvesafb_setpalette(&entry, 1, regno, info); entry 1070 drivers/video/fbdev/vermilion/vermilion.c struct vml_info *entry; entry 1089 drivers/video/fbdev/vermilion/vermilion.c entry = list_entry(list, struct vml_info, head); entry 1096 drivers/video/fbdev/vermilion/vermilion.c if (!vmlfb_check_var_locked(&entry->info.var, entry)) { entry 1097 drivers/video/fbdev/vermilion/vermilion.c vmlfb_set_par_locked(entry); entry 1107 drivers/video/fbdev/vermilion/vermilion.c save_activate = entry->info.var.activate; entry 1108 drivers/video/fbdev/vermilion/vermilion.c entry->info.var.bits_per_pixel = 16; entry 1109 drivers/video/fbdev/vermilion/vermilion.c vmlfb_set_pref_pixel_format(&entry->info.var); entry 1110 drivers/video/fbdev/vermilion/vermilion.c if (fb_find_mode(&entry->info.var, entry 1111 drivers/video/fbdev/vermilion/vermilion.c &entry->info, entry 1113 drivers/video/fbdev/vermilion/vermilion.c entry->info.var.activate |= entry 1115 drivers/video/fbdev/vermilion/vermilion.c fb_set_var(&entry->info, &entry->info.var); entry 1120 drivers/video/fbdev/vermilion/vermilion.c entry->info.var.activate = save_activate; entry 1123 drivers/video/fbdev/vermilion/vermilion.c vmlfb_blank_locked(entry); entry 1137 drivers/video/fbdev/vermilion/vermilion.c struct vml_info *entry, *next; entry 1146 drivers/video/fbdev/vermilion/vermilion.c list_for_each_entry_safe(entry, next, &global_has_mode, head) { entry 1148 drivers/video/fbdev/vermilion/vermilion.c vmlfb_disable_pipe(entry); entry 1149 drivers/video/fbdev/vermilion/vermilion.c list_move_tail(&entry->head, &global_no_mode); entry 108 drivers/video/fbdev/vesafb.c struct { u_char blue, green, red, pad; } entry; entry 110 drivers/video/fbdev/vesafb.c entry.red = red >> shift; entry 111 drivers/video/fbdev/vesafb.c entry.green = green >> shift; entry 112 drivers/video/fbdev/vesafb.c entry.blue = blue >> shift; entry 113 drivers/video/fbdev/vesafb.c entry.pad = 0; entry 121 drivers/video/fbdev/vesafb.c "D" (&entry), /* EDI */ entry 146 drivers/video/of_display_timing.c struct device_node *entry; entry 165 drivers/video/of_display_timing.c entry = of_parse_phandle(timings_np, "native-mode", 0); entry 167 drivers/video/of_display_timing.c if (!entry) entry 168 drivers/video/of_display_timing.c entry = of_get_next_child(timings_np, NULL); entry 170 drivers/video/of_display_timing.c if (!entry) { entry 175 drivers/video/of_display_timing.c pr_debug("%pOF: using %pOFn as default timing\n", np, entry); entry 177 drivers/video/of_display_timing.c native_mode = entry; entry 197 drivers/video/of_display_timing.c for_each_child_of_node(timings_np, entry) { entry 208 drivers/video/of_display_timing.c r = of_parse_display_timing(entry, dt); entry 220 drivers/video/of_display_timing.c if (native_mode == entry) entry 1021 drivers/vme/bridges/vme_ca91cx42.c struct ca91cx42_dma_entry *entry, *prev; entry 1031 drivers/vme/bridges/vme_ca91cx42.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 1032 drivers/vme/bridges/vme_ca91cx42.c if (!entry) { entry 1038 drivers/vme/bridges/vme_ca91cx42.c if ((unsigned long)&entry->descriptor & CA91CX42_DCPP_M) { entry 1040 drivers/vme/bridges/vme_ca91cx42.c "required: %p\n", &entry->descriptor); entry 1045 drivers/vme/bridges/vme_ca91cx42.c memset(&entry->descriptor, 0, sizeof(entry->descriptor)); entry 1048 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_L2V; entry 1085 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VCT_BLT; entry 1090 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VDW_D8; entry 1093 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VDW_D16; entry 1096 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VDW_D32; entry 1099 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VDW_D64; entry 1109 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VAS_A16; entry 1112 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VAS_A24; entry 1115 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VAS_A32; entry 1118 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VAS_USER1; entry 1121 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_VAS_USER2; entry 1130 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_SUPER_SUPR; entry 1132 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dctl |= CA91CX42_DCTL_PGM_PGM; entry 1134 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dtbc = count; entry 1135 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dla = pci_attr->address; entry 1136 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dva = vme_attr->address; entry 1137 drivers/vme/bridges/vme_ca91cx42.c entry->descriptor.dcpp = CA91CX42_DCPP_NULL; entry 1140 drivers/vme/bridges/vme_ca91cx42.c list_add_tail(&entry->list, &list->entries); entry 1143 drivers/vme/bridges/vme_ca91cx42.c if (entry->list.prev != &list->entries) { entry 1144 drivers/vme/bridges/vme_ca91cx42.c prev = list_entry(entry->list.prev, struct ca91cx42_dma_entry, entry 1147 drivers/vme/bridges/vme_ca91cx42.c desc_ptr = virt_to_bus(&entry->descriptor); entry 1157 drivers/vme/bridges/vme_ca91cx42.c kfree(entry); entry 1180 drivers/vme/bridges/vme_ca91cx42.c struct ca91cx42_dma_entry *entry; entry 1208 drivers/vme/bridges/vme_ca91cx42.c entry = list_first_entry(&list->entries, struct ca91cx42_dma_entry, entry 1211 drivers/vme/bridges/vme_ca91cx42.c bus_addr = virt_to_bus(&entry->descriptor); entry 1274 drivers/vme/bridges/vme_ca91cx42.c struct ca91cx42_dma_entry *entry; entry 1279 drivers/vme/bridges/vme_ca91cx42.c entry = list_entry(pos, struct ca91cx42_dma_entry, list); entry 1280 drivers/vme/bridges/vme_ca91cx42.c kfree(entry); entry 1627 drivers/vme/bridges/vme_tsi148.c struct tsi148_dma_entry *entry, *prev; entry 1638 drivers/vme/bridges/vme_tsi148.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 1639 drivers/vme/bridges/vme_tsi148.c if (!entry) { entry 1645 drivers/vme/bridges/vme_tsi148.c if ((unsigned long)&entry->descriptor & 0x7) { entry 1648 drivers/vme/bridges/vme_tsi148.c &entry->descriptor); entry 1656 drivers/vme/bridges/vme_tsi148.c memset(&entry->descriptor, 0, sizeof(entry->descriptor)); entry 1663 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dsal = cpu_to_be32(pattern_attr->pattern); entry 1674 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dsat = cpu_to_be32(val); entry 1681 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dsau = cpu_to_be32(address_high); entry 1682 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dsal = cpu_to_be32(address_low); entry 1683 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dsat = cpu_to_be32(TSI148_LCSR_DSAT_TYP_PCI); entry 1690 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dsau = cpu_to_be32(address_high); entry 1691 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dsal = cpu_to_be32(address_low); entry 1692 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dsat = cpu_to_be32(TSI148_LCSR_DSAT_TYP_VME); entry 1695 drivers/vme/bridges/vme_tsi148.c tsi148_bridge->parent, &entry->descriptor.dsat, entry 1708 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dnlau = cpu_to_be32(0); entry 1709 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dnlal = cpu_to_be32(TSI148_LCSR_DNLAL_LLA); entry 1718 drivers/vme/bridges/vme_tsi148.c entry->descriptor.ddau = cpu_to_be32(address_high); entry 1719 drivers/vme/bridges/vme_tsi148.c entry->descriptor.ddal = cpu_to_be32(address_low); entry 1720 drivers/vme/bridges/vme_tsi148.c entry->descriptor.ddat = cpu_to_be32(TSI148_LCSR_DDAT_TYP_PCI); entry 1727 drivers/vme/bridges/vme_tsi148.c entry->descriptor.ddau = cpu_to_be32(address_high); entry 1728 drivers/vme/bridges/vme_tsi148.c entry->descriptor.ddal = cpu_to_be32(address_low); entry 1729 drivers/vme/bridges/vme_tsi148.c entry->descriptor.ddat = cpu_to_be32(TSI148_LCSR_DDAT_TYP_VME); entry 1732 drivers/vme/bridges/vme_tsi148.c tsi148_bridge->parent, &entry->descriptor.ddat, entry 1745 drivers/vme/bridges/vme_tsi148.c entry->descriptor.dcnt = cpu_to_be32((u32)count); entry 1748 drivers/vme/bridges/vme_tsi148.c list_add_tail(&entry->list, &list->entries); entry 1750 drivers/vme/bridges/vme_tsi148.c entry->dma_handle = dma_map_single(tsi148_bridge->parent, entry 1751 drivers/vme/bridges/vme_tsi148.c &entry->descriptor, entry 1752 drivers/vme/bridges/vme_tsi148.c sizeof(entry->descriptor), entry 1754 drivers/vme/bridges/vme_tsi148.c if (dma_mapping_error(tsi148_bridge->parent, entry->dma_handle)) { entry 1761 drivers/vme/bridges/vme_tsi148.c if (entry->list.prev != &list->entries) { entry 1762 drivers/vme/bridges/vme_tsi148.c reg_split((unsigned long long)entry->dma_handle, &address_high, entry 1764 drivers/vme/bridges/vme_tsi148.c prev = list_entry(entry->list.prev, struct tsi148_dma_entry, entry 1777 drivers/vme/bridges/vme_tsi148.c kfree(entry); entry 1811 drivers/vme/bridges/vme_tsi148.c struct tsi148_dma_entry *entry; entry 1841 drivers/vme/bridges/vme_tsi148.c entry = list_first_entry(&list->entries, struct tsi148_dma_entry, entry 1846 drivers/vme/bridges/vme_tsi148.c reg_split(entry->dma_handle, &bus_addr_high, &bus_addr_low); entry 1902 drivers/vme/bridges/vme_tsi148.c struct tsi148_dma_entry *entry; entry 1909 drivers/vme/bridges/vme_tsi148.c entry = list_entry(pos, struct tsi148_dma_entry, list); entry 1911 drivers/vme/bridges/vme_tsi148.c dma_unmap_single(tsi148_bridge->parent, entry->dma_handle, entry 1913 drivers/vme/bridges/vme_tsi148.c kfree(entry); entry 53 drivers/vme/vme.c return list_entry(resource->entry, struct vme_master_resource, entry 57 drivers/vme/vme.c return list_entry(resource->entry, struct vme_slave_resource, entry 61 drivers/vme/vme.c return list_entry(resource->entry, struct vme_dma_resource, entry 65 drivers/vme/vme.c return list_entry(resource->entry, struct vme_lm_resource, entry 338 drivers/vme/vme.c resource->entry = &allocated_image->list; entry 383 drivers/vme/vme.c image = list_entry(resource->entry, struct vme_slave_resource, list); entry 432 drivers/vme/vme.c image = list_entry(resource->entry, struct vme_slave_resource, list); entry 459 drivers/vme/vme.c slave_image = list_entry(resource->entry, struct vme_slave_resource, entry 542 drivers/vme/vme.c resource->entry = &allocated_image->list; entry 587 drivers/vme/vme.c image = list_entry(resource->entry, struct vme_master_resource, list); entry 637 drivers/vme/vme.c image = list_entry(resource->entry, struct vme_master_resource, list); entry 681 drivers/vme/vme.c image = list_entry(resource->entry, struct vme_master_resource, list); entry 730 drivers/vme/vme.c image = list_entry(resource->entry, struct vme_master_resource, list); entry 781 drivers/vme/vme.c image = list_entry(resource->entry, struct vme_master_resource, list); entry 809 drivers/vme/vme.c image = list_entry(resource->entry, struct vme_master_resource, list); entry 839 drivers/vme/vme.c master_image = list_entry(resource->entry, struct vme_master_resource, entry 917 drivers/vme/vme.c resource->entry = &allocated_ctrlr->list; entry 956 drivers/vme/vme.c dma_list->parent = list_entry(resource->entry, entry 1231 drivers/vme/vme.c ctrlr = list_entry(resource->entry, struct vme_dma_resource, list); entry 1517 drivers/vme/vme.c resource->entry = &allocated_lm->list; entry 1552 drivers/vme/vme.c lm = list_entry(resource->entry, struct vme_lm_resource, list); entry 1583 drivers/vme/vme.c lm = list_entry(resource->entry, struct vme_lm_resource, list); entry 1619 drivers/vme/vme.c lm = list_entry(resource->entry, struct vme_lm_resource, list); entry 1656 drivers/vme/vme.c lm = list_entry(resource->entry, struct vme_lm_resource, list); entry 1689 drivers/vme/vme.c lm = list_entry(resource->entry, struct vme_lm_resource, list); entry 1721 drivers/vme/vme.c lm = list_entry(resource->entry, struct vme_lm_resource, list); entry 95 drivers/w1/w1_int.c struct w1_master *dev, *entry; entry 115 drivers/w1/w1_int.c list_for_each_entry(entry, &w1_masters, w1_master_entry) { entry 116 drivers/w1/w1_int.c if (entry->id == id) { entry 25 drivers/watchdog/watchdog_pretimeout.c struct list_head entry; entry 36 drivers/watchdog/watchdog_pretimeout.c struct list_head entry; entry 43 drivers/watchdog/watchdog_pretimeout.c list_for_each_entry(priv, &governor_list, entry) entry 57 drivers/watchdog/watchdog_pretimeout.c list_for_each_entry(priv, &governor_list, entry) entry 132 drivers/watchdog/watchdog_pretimeout.c list_add(&priv->entry, &governor_list); entry 139 drivers/watchdog/watchdog_pretimeout.c list_for_each_entry(p, &pretimeout_list, entry) entry 158 drivers/watchdog/watchdog_pretimeout.c list_for_each_entry_safe(priv, t, &governor_list, entry) { entry 160 drivers/watchdog/watchdog_pretimeout.c list_del(&priv->entry); entry 167 drivers/watchdog/watchdog_pretimeout.c list_for_each_entry(p, &pretimeout_list, entry) entry 188 drivers/watchdog/watchdog_pretimeout.c list_add(&p->entry, &pretimeout_list); entry 206 drivers/watchdog/watchdog_pretimeout.c list_for_each_entry_safe(p, t, &pretimeout_list, entry) { entry 208 drivers/watchdog/watchdog_pretimeout.c list_del(&p->entry); entry 25 drivers/watchdog/wdat_wdt.c struct acpi_wdat_entry entry; entry 67 drivers/watchdog/wdat_wdt.c const struct acpi_generic_address *gas = &instr->entry.register_region; entry 92 drivers/watchdog/wdat_wdt.c const struct acpi_generic_address *gas = &instr->entry.register_region; entry 129 drivers/watchdog/wdat_wdt.c const struct acpi_wdat_entry *entry = &instr->entry; entry 135 drivers/watchdog/wdat_wdt.c gas = &entry->register_region; entry 137 drivers/watchdog/wdat_wdt.c preserve = entry->instruction & ACPI_WDAT_PRESERVE_REGISTER; entry 138 drivers/watchdog/wdat_wdt.c flags = entry->instruction & ~ACPI_WDAT_PRESERVE_REGISTER; entry 139 drivers/watchdog/wdat_wdt.c value = entry->value; entry 140 drivers/watchdog/wdat_wdt.c mask = entry->mask; entry 393 drivers/watchdog/wdat_wdt.c instr->entry = entries[i]; entry 163 drivers/xen/grant-table.c static inline grant_ref_t *__gnttab_entry(grant_ref_t entry) entry 165 drivers/xen/grant-table.c return &gnttab_list[(entry) / RPP][(entry) % RPP]; entry 168 drivers/xen/grant-table.c #define gnttab_entry(entry) (*__gnttab_entry(entry)) entry 376 drivers/xen/grant-table.c struct deferred_entry *entry entry 380 drivers/xen/grant-table.c if (entry == first) entry 382 drivers/xen/grant-table.c list_del(&entry->list); entry 384 drivers/xen/grant-table.c if (_gnttab_end_foreign_access_ref(entry->ref, entry->ro)) { entry 385 drivers/xen/grant-table.c put_free_entry(entry->ref); entry 386 drivers/xen/grant-table.c if (entry->page) { entry 388 drivers/xen/grant-table.c entry->ref, page_to_pfn(entry->page)); entry 389 drivers/xen/grant-table.c put_page(entry->page); entry 391 drivers/xen/grant-table.c pr_info("freeing g.e. %#x\n", entry->ref); entry 392 drivers/xen/grant-table.c kfree(entry); entry 393 drivers/xen/grant-table.c entry = NULL; entry 395 drivers/xen/grant-table.c if (!--entry->warn_delay) entry 396 drivers/xen/grant-table.c pr_info("g.e. %#x still pending\n", entry->ref); entry 398 drivers/xen/grant-table.c first = entry; entry 401 drivers/xen/grant-table.c if (entry) entry 402 drivers/xen/grant-table.c list_add_tail(&entry->list, &deferred_list); entry 416 drivers/xen/grant-table.c struct deferred_entry *entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 419 drivers/xen/grant-table.c if (entry) { entry 422 drivers/xen/grant-table.c entry->ref = ref; entry 423 drivers/xen/grant-table.c entry->ro = readonly; entry 424 drivers/xen/grant-table.c entry->page = page; entry 425 drivers/xen/grant-table.c entry->warn_delay = 60; entry 427 drivers/xen/grant-table.c list_add_tail(&entry->list, &deferred_list); entry 124 drivers/xen/mcelog.c struct xen_mce *m = &xen_mcelog.entry[i]; entry 130 drivers/xen/mcelog.c memset(xen_mcelog.entry, 0, num * sizeof(struct xen_mce)); entry 199 drivers/xen/mcelog.c unsigned entry; entry 201 drivers/xen/mcelog.c entry = xen_mcelog.next; entry 208 drivers/xen/mcelog.c if (entry >= XEN_MCE_LOG_LEN) { entry 214 drivers/xen/mcelog.c memcpy(xen_mcelog.entry + entry, mce, sizeof(struct xen_mce)); entry 276 drivers/xen/privcmd.c mmapcmd.entry); entry 41 drivers/xen/xen-pciback/conf_space.c const struct config_field_entry *entry, entry 45 drivers/xen/xen-pciback/conf_space.c const struct config_field *field = entry->field; entry 53 drivers/xen/xen-pciback/conf_space.c entry->data); entry 58 drivers/xen/xen-pciback/conf_space.c entry->data); entry 62 drivers/xen/xen-pciback/conf_space.c ret = field->u.dw.read(dev, offset, value, entry->data); entry 69 drivers/xen/xen-pciback/conf_space.c const struct config_field_entry *entry, entry 73 drivers/xen/xen-pciback/conf_space.c const struct config_field *field = entry->field; entry 79 drivers/xen/xen-pciback/conf_space.c entry->data); entry 84 drivers/xen/xen-pciback/conf_space.c entry->data); entry 89 drivers/xen/xen-pciback/conf_space.c entry->data); entry 240 drivers/xen/xen-pciback/pciback_ops.c entries[i].entry = op->msix_entries[i].entry; entry 247 drivers/xen/xen-pciback/pciback_ops.c op->msix_entries[i].entry = entries[i].entry; entry 35 drivers/xen/xen-pciback/vpci.c struct pci_dev_entry *entry; entry 45 drivers/xen/xen-pciback/vpci.c list_for_each_entry(entry, entry 48 drivers/xen/xen-pciback/vpci.c if (PCI_FUNC(entry->dev->devfn) == PCI_FUNC(devfn)) { entry 49 drivers/xen/xen-pciback/vpci.c dev = entry->dev; entry 235 drivers/xen/xen-pciback/vpci.c struct pci_dev_entry *entry; entry 242 drivers/xen/xen-pciback/vpci.c list_for_each_entry(entry, entry 245 drivers/xen/xen-pciback/vpci.c dev = entry->dev; entry 312 drivers/xen/xen-scsiback.c struct v2p_entry *entry = container_of(kref, struct v2p_entry, kref); entry 313 drivers/xen/xen-scsiback.c struct scsiback_tpg *tpg = entry->tpg; entry 319 drivers/xen/xen-scsiback.c kfree(entry); entry 630 drivers/xen/xen-scsiback.c struct v2p_entry *entry; entry 635 drivers/xen/xen-scsiback.c list_for_each_entry(entry, head, l) { entry 636 drivers/xen/xen-scsiback.c if ((entry->v.chn == v->chn) && entry 637 drivers/xen/xen-scsiback.c (entry->v.tgt == v->tgt) && entry 638 drivers/xen/xen-scsiback.c (entry->v.lun == v->lun)) { entry 639 drivers/xen/xen-scsiback.c kref_get(&entry->kref); entry 643 drivers/xen/xen-scsiback.c entry = NULL; entry 647 drivers/xen/xen-scsiback.c return entry; entry 878 drivers/xen/xen-scsiback.c struct v2p_entry *entry; entry 880 drivers/xen/xen-scsiback.c list_for_each_entry(entry, head, l) entry 881 drivers/xen/xen-scsiback.c if ((entry->v.chn == v->chn) && entry 882 drivers/xen/xen-scsiback.c (entry->v.tgt == v->tgt) && entry 883 drivers/xen/xen-scsiback.c (entry->v.lun == v->lun)) entry 884 drivers/xen/xen-scsiback.c return entry; entry 983 drivers/xen/xen-scsiback.c static void __scsiback_del_translation_entry(struct v2p_entry *entry) entry 985 drivers/xen/xen-scsiback.c list_del(&entry->l); entry 986 drivers/xen/xen-scsiback.c kref_put(&entry->kref, scsiback_free_translation_entry); entry 995 drivers/xen/xen-scsiback.c struct v2p_entry *entry; entry 1001 drivers/xen/xen-scsiback.c entry = scsiback_chk_translation_entry(info, v); entry 1002 drivers/xen/xen-scsiback.c if (entry) entry 1003 drivers/xen/xen-scsiback.c __scsiback_del_translation_entry(entry); entry 1014 drivers/xen/xen-scsiback.c struct v2p_entry *entry; entry 1020 drivers/xen/xen-scsiback.c entry = scsiback_chk_translation_entry(info, vir); entry 1022 drivers/xen/xen-scsiback.c if (entry) entry 1210 drivers/xen/xen-scsiback.c struct v2p_entry *entry, *tmp; entry 1216 drivers/xen/xen-scsiback.c list_for_each_entry_safe(entry, tmp, head, l) entry 1217 drivers/xen/xen-scsiback.c __scsiback_del_translation_entry(entry); entry 103 drivers/zorro/proc.c struct proc_dir_entry *entry; entry 107 drivers/zorro/proc.c entry = proc_create_data(name, 0, proc_bus_zorro_dir, entry 110 drivers/zorro/proc.c if (!entry) entry 112 drivers/zorro/proc.c proc_set_size(entry, sizeof(struct zorro_dev)); entry 92 fs/afs/callback.c struct afs_server_entry *entry = &slist->servers[index]; entry 94 fs/afs/callback.c struct afs_server *server = entry->server; entry 99 fs/afs/callback.c if (vcbi && likely(vcbi == entry->cb_interest)) entry 103 fs/afs/callback.c cbi = afs_get_cb_interest(entry->cb_interest); entry 128 fs/afs/callback.c if (entry->cb_interest) { entry 134 fs/afs/callback.c entry->cb_interest = cbi; entry 146 fs/afs/callback.c if (!entry->cb_interest) { entry 147 fs/afs/callback.c entry->cb_interest = afs_get_cb_interest(new); entry 151 fs/afs/callback.c cbi = afs_get_cb_interest(entry->cb_interest); entry 288 fs/afs/proc.c const struct afs_vlserver_entry *entry; entry 300 fs/afs/proc.c entry = v; entry 301 fs/afs/proc.c vlserver = entry->server; entry 305 fs/afs/proc.c vlserver->name, entry->priority, entry->weight, entry 306 fs/afs/proc.c dns_record_sources[alist ? alist->source : entry->source], entry 307 fs/afs/proc.c dns_lookup_statuses[alist ? alist->status : entry->status]); entry 20 fs/afs/vlclient.c struct afs_vldb_entry *entry; entry 33 fs/afs/vlclient.c entry = call->ret_vldb; entry 40 fs/afs/vlclient.c entry->name[i] = (u8)ntohl(uvldb->name[i]); entry 41 fs/afs/vlclient.c entry->name[i] = 0; entry 42 fs/afs/vlclient.c entry->name_len = strlen(entry->name); entry 59 fs/afs/vlclient.c int n = entry->nr_servers; entry 66 fs/afs/vlclient.c entry->fs_mask[n] |= AFS_VOL_VTM_RW; entry 68 fs/afs/vlclient.c entry->fs_mask[n] |= AFS_VOL_VTM_BAK; entry 71 fs/afs/vlclient.c entry->fs_mask[n] |= AFS_VOL_VTM_RO; entry 72 fs/afs/vlclient.c if (!entry->fs_mask[n]) entry 76 fs/afs/vlclient.c uuid = (struct afs_uuid *)&entry->fs_server[n]; entry 85 fs/afs/vlclient.c entry->nr_servers++; entry 89 fs/afs/vlclient.c entry->vid[i] = ntohl(uvldb->volumeId[i]); entry 92 fs/afs/vlclient.c __set_bit(AFS_VLDB_HAS_RW, &entry->flags); entry 94 fs/afs/vlclient.c __set_bit(AFS_VLDB_HAS_RO, &entry->flags); entry 96 fs/afs/vlclient.c __set_bit(AFS_VLDB_HAS_BAK, &entry->flags); entry 99 fs/afs/vlclient.c entry->error = -ENOMEDIUM; entry 100 fs/afs/vlclient.c __set_bit(AFS_VLDB_QUERY_ERROR, &entry->flags); entry 103 fs/afs/vlclient.c __set_bit(AFS_VLDB_QUERY_VALID, &entry->flags); entry 132 fs/afs/vlclient.c struct afs_vldb_entry *entry; entry 143 fs/afs/vlclient.c entry = kzalloc(sizeof(struct afs_vldb_entry), GFP_KERNEL); entry 144 fs/afs/vlclient.c if (!entry) entry 150 fs/afs/vlclient.c kfree(entry); entry 155 fs/afs/vlclient.c call->ret_vldb = entry; entry 1669 fs/aio.c if (!list_empty(&req->wait.entry)) { entry 1670 fs/aio.c list_del_init(&req->wait.entry); entry 1690 fs/aio.c list_del_init(&req->wait.entry); entry 1769 fs/aio.c INIT_LIST_HEAD(&req->wait.entry); entry 1776 fs/aio.c if (unlikely(list_empty(&req->wait.entry))) { entry 1783 fs/aio.c list_del_init(&req->wait.entry); entry 83 fs/binfmt_flat.c unsigned long entry; /* Start address for this module */ entry 739 fs/binfmt_flat.c textpos, 0x00ffffff&ntohl(hdr->entry), ntohl(hdr->data_start)); entry 752 fs/binfmt_flat.c libinfo->lib_list[id].entry = (0x00ffffff & ntohl(hdr->entry)) + textpos; entry 992 fs/binfmt_flat.c start_addr = libinfo.lib_list[0].entry; entry 1002 fs/binfmt_flat.c start_addr = libinfo.lib_list[i].entry; entry 283 fs/btrfs/delayed-ref.c struct btrfs_delayed_ref_head *entry; entry 292 fs/btrfs/delayed-ref.c entry = rb_entry(parent_node, struct btrfs_delayed_ref_head, entry 295 fs/btrfs/delayed-ref.c if (bytenr < entry->bytenr) { entry 297 fs/btrfs/delayed-ref.c } else if (bytenr > entry->bytenr) { entry 301 fs/btrfs/delayed-ref.c return entry; entry 316 fs/btrfs/delayed-ref.c struct btrfs_delayed_ref_node *entry; entry 323 fs/btrfs/delayed-ref.c entry = rb_entry(parent_node, struct btrfs_delayed_ref_node, entry 325 fs/btrfs/delayed-ref.c comp = comp_refs(ins, entry, true); entry 332 fs/btrfs/delayed-ref.c return entry; entry 345 fs/btrfs/delayed-ref.c struct btrfs_delayed_ref_head *entry; entry 351 fs/btrfs/delayed-ref.c entry = rb_entry(n, struct btrfs_delayed_ref_head, href_node); entry 353 fs/btrfs/delayed-ref.c return entry; entry 367 fs/btrfs/delayed-ref.c struct btrfs_delayed_ref_head *entry; entry 370 fs/btrfs/delayed-ref.c entry = NULL; entry 372 fs/btrfs/delayed-ref.c entry = rb_entry(n, struct btrfs_delayed_ref_head, href_node); entry 374 fs/btrfs/delayed-ref.c if (bytenr < entry->bytenr) entry 376 fs/btrfs/delayed-ref.c else if (bytenr > entry->bytenr) entry 379 fs/btrfs/delayed-ref.c return entry; entry 381 fs/btrfs/delayed-ref.c if (entry && return_bigger) { entry 382 fs/btrfs/delayed-ref.c if (bytenr > entry->bytenr) { entry 383 fs/btrfs/delayed-ref.c n = rb_next(&entry->href_node); entry 386 fs/btrfs/delayed-ref.c entry = rb_entry(n, struct btrfs_delayed_ref_head, entry 389 fs/btrfs/delayed-ref.c return entry; entry 2082 fs/btrfs/extent-tree.c struct btrfs_delayed_ref_node *entry; entry 2089 fs/btrfs/extent-tree.c entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node); entry 2090 fs/btrfs/extent-tree.c first = entry->bytenr; entry 2094 fs/btrfs/extent-tree.c entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node); entry 2095 fs/btrfs/extent-tree.c last = entry->bytenr; entry 2100 fs/btrfs/extent-tree.c entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node); entry 2101 fs/btrfs/extent-tree.c WARN_ON(!entry->in_tree); entry 2103 fs/btrfs/extent-tree.c middle = entry->bytenr; entry 53 fs/btrfs/extent_io.c void btrfs_leak_debug_del(struct list_head *entry) entry 58 fs/btrfs/extent_io.c list_del(entry); entry 107 fs/btrfs/extent_io.c #define btrfs_leak_debug_del(entry) do {} while (0) entry 335 fs/btrfs/extent_io.c struct tree_entry *entry; entry 346 fs/btrfs/extent_io.c entry = rb_entry(parent, struct tree_entry, rb_node); entry 348 fs/btrfs/extent_io.c if (offset < entry->start) entry 350 fs/btrfs/extent_io.c else if (offset > entry->end) entry 390 fs/btrfs/extent_io.c struct tree_entry *entry; entry 395 fs/btrfs/extent_io.c entry = rb_entry(prev, struct tree_entry, rb_node); entry 396 fs/btrfs/extent_io.c prev_entry = entry; entry 398 fs/btrfs/extent_io.c if (offset < entry->start) entry 400 fs/btrfs/extent_io.c else if (offset > entry->end) entry 98 fs/btrfs/extent_map.c struct extent_map *entry = NULL; entry 105 fs/btrfs/extent_map.c entry = rb_entry(parent, struct extent_map, rb_node); entry 107 fs/btrfs/extent_map.c if (em->start < entry->start) { entry 109 fs/btrfs/extent_map.c } else if (em->start >= extent_map_end(entry)) { entry 118 fs/btrfs/extent_map.c while (parent && em->start >= extent_map_end(entry)) { entry 120 fs/btrfs/extent_map.c entry = rb_entry(parent, struct extent_map, rb_node); entry 123 fs/btrfs/extent_map.c if (end > entry->start && em->start < extent_map_end(entry)) entry 127 fs/btrfs/extent_map.c entry = rb_entry(parent, struct extent_map, rb_node); entry 128 fs/btrfs/extent_map.c while (parent && em->start < entry->start) { entry 130 fs/btrfs/extent_map.c entry = rb_entry(parent, struct extent_map, rb_node); entry 133 fs/btrfs/extent_map.c if (end > entry->start && em->start < extent_map_end(entry)) entry 152 fs/btrfs/extent_map.c struct extent_map *entry; entry 156 fs/btrfs/extent_map.c entry = rb_entry(n, struct extent_map, rb_node); entry 158 fs/btrfs/extent_map.c prev_entry = entry; entry 160 fs/btrfs/extent_map.c if (offset < entry->start) entry 162 fs/btrfs/extent_map.c else if (offset >= extent_map_end(entry)) entry 85 fs/btrfs/file.c struct inode_defrag *entry; entry 93 fs/btrfs/file.c entry = rb_entry(parent, struct inode_defrag, rb_node); entry 95 fs/btrfs/file.c ret = __compare_inode_defrag(defrag, entry); entry 105 fs/btrfs/file.c if (defrag->transid < entry->transid) entry 106 fs/btrfs/file.c entry->transid = defrag->transid; entry 107 fs/btrfs/file.c if (defrag->last_offset > entry->last_offset) entry 108 fs/btrfs/file.c entry->last_offset = defrag->last_offset; entry 213 fs/btrfs/file.c struct inode_defrag *entry = NULL; entry 226 fs/btrfs/file.c entry = rb_entry(parent, struct inode_defrag, rb_node); entry 228 fs/btrfs/file.c ret = __compare_inode_defrag(&tmp, entry); entry 237 fs/btrfs/file.c if (parent && __compare_inode_defrag(&tmp, entry) > 0) { entry 240 fs/btrfs/file.c entry = rb_entry(parent, struct inode_defrag, rb_node); entry 242 fs/btrfs/file.c entry = NULL; entry 245 fs/btrfs/file.c if (entry) entry 248 fs/btrfs/file.c return entry; entry 519 fs/btrfs/free-space-cache.c struct btrfs_free_space_entry *entry; entry 524 fs/btrfs/free-space-cache.c entry = io_ctl->cur; entry 525 fs/btrfs/free-space-cache.c entry->offset = cpu_to_le64(offset); entry 526 fs/btrfs/free-space-cache.c entry->bytes = cpu_to_le64(bytes); entry 527 fs/btrfs/free-space-cache.c entry->type = (bitmap) ? BTRFS_FREE_SPACE_BITMAP : entry 587 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, u8 *type) entry 599 fs/btrfs/free-space-cache.c entry->offset = le64_to_cpu(e->offset); entry 600 fs/btrfs/free-space-cache.c entry->bytes = le64_to_cpu(e->bytes); entry 614 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry) entry 622 fs/btrfs/free-space-cache.c copy_page(entry->bitmap, io_ctl->cur); entry 1103 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, *next; entry 1107 fs/btrfs/free-space-cache.c list_for_each_entry_safe(entry, next, bitmap_list, list) { entry 1108 fs/btrfs/free-space-cache.c ret = io_ctl_add_bitmap(io_ctl, entry->bitmap); entry 1111 fs/btrfs/free-space-cache.c list_del_init(&entry->list); entry 1132 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, *next; entry 1134 fs/btrfs/free-space-cache.c list_for_each_entry_safe(entry, next, bitmap_list, list) entry 1135 fs/btrfs/free-space-cache.c list_del_init(&entry->list); entry 1512 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, *prev = NULL; entry 1517 fs/btrfs/free-space-cache.c entry = NULL; entry 1521 fs/btrfs/free-space-cache.c entry = rb_entry(n, struct btrfs_free_space, offset_index); entry 1522 fs/btrfs/free-space-cache.c prev = entry; entry 1524 fs/btrfs/free-space-cache.c if (offset < entry->offset) entry 1526 fs/btrfs/free-space-cache.c else if (offset > entry->offset) entry 1533 fs/btrfs/free-space-cache.c if (!entry) entry 1535 fs/btrfs/free-space-cache.c if (entry->bitmap) entry 1536 fs/btrfs/free-space-cache.c return entry; entry 1545 fs/btrfs/free-space-cache.c entry = rb_entry(n, struct btrfs_free_space, offset_index); entry 1546 fs/btrfs/free-space-cache.c if (entry->offset != offset) entry 1549 fs/btrfs/free-space-cache.c WARN_ON(!entry->bitmap); entry 1550 fs/btrfs/free-space-cache.c return entry; entry 1551 fs/btrfs/free-space-cache.c } else if (entry) { entry 1552 fs/btrfs/free-space-cache.c if (entry->bitmap) { entry 1557 fs/btrfs/free-space-cache.c n = rb_prev(&entry->offset_index); entry 1563 fs/btrfs/free-space-cache.c entry = prev; entry 1566 fs/btrfs/free-space-cache.c return entry; entry 1573 fs/btrfs/free-space-cache.c entry = prev; entry 1574 fs/btrfs/free-space-cache.c if (entry->offset > offset) { entry 1575 fs/btrfs/free-space-cache.c n = rb_prev(&entry->offset_index); entry 1577 fs/btrfs/free-space-cache.c entry = rb_entry(n, struct btrfs_free_space, entry 1579 fs/btrfs/free-space-cache.c ASSERT(entry->offset <= offset); entry 1582 fs/btrfs/free-space-cache.c return entry; entry 1588 fs/btrfs/free-space-cache.c if (entry->bitmap) { entry 1589 fs/btrfs/free-space-cache.c n = rb_prev(&entry->offset_index); entry 1597 fs/btrfs/free-space-cache.c if (entry->offset + BITS_PER_BITMAP * ctl->unit > offset) entry 1598 fs/btrfs/free-space-cache.c return entry; entry 1599 fs/btrfs/free-space-cache.c } else if (entry->offset + entry->bytes > offset) entry 1600 fs/btrfs/free-space-cache.c return entry; entry 1606 fs/btrfs/free-space-cache.c if (entry->bitmap) { entry 1607 fs/btrfs/free-space-cache.c if (entry->offset + BITS_PER_BITMAP * entry 1611 fs/btrfs/free-space-cache.c if (entry->offset + entry->bytes > offset) entry 1615 fs/btrfs/free-space-cache.c n = rb_next(&entry->offset_index); entry 1618 fs/btrfs/free-space-cache.c entry = rb_entry(n, struct btrfs_free_space, offset_index); entry 1620 fs/btrfs/free-space-cache.c return entry; entry 1799 fs/btrfs/free-space-cache.c static inline u64 get_max_extent_size(struct btrfs_free_space *entry) entry 1801 fs/btrfs/free-space-cache.c if (entry->bitmap) entry 1802 fs/btrfs/free-space-cache.c return entry->max_extent_size; entry 1803 fs/btrfs/free-space-cache.c return entry->bytes; entry 1811 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry; entry 1820 fs/btrfs/free-space-cache.c entry = tree_search_offset(ctl, offset_to_bitmap(ctl, *offset), 0, 1); entry 1821 fs/btrfs/free-space-cache.c if (!entry) entry 1824 fs/btrfs/free-space-cache.c for (node = &entry->offset_index; node; node = rb_next(node)) { entry 1825 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); entry 1826 fs/btrfs/free-space-cache.c if (entry->bytes < *bytes) { entry 1827 fs/btrfs/free-space-cache.c *max_extent_size = max(get_max_extent_size(entry), entry 1836 fs/btrfs/free-space-cache.c tmp = entry->offset - ctl->start + align - 1; entry 1839 fs/btrfs/free-space-cache.c align_off = tmp - entry->offset; entry 1842 fs/btrfs/free-space-cache.c tmp = entry->offset; entry 1845 fs/btrfs/free-space-cache.c if (entry->bytes < *bytes + align_off) { entry 1846 fs/btrfs/free-space-cache.c *max_extent_size = max(get_max_extent_size(entry), entry 1851 fs/btrfs/free-space-cache.c if (entry->bitmap) { entry 1854 fs/btrfs/free-space-cache.c ret = search_bitmap(ctl, entry, &tmp, &size, true); entry 1858 fs/btrfs/free-space-cache.c return entry; entry 1861 fs/btrfs/free-space-cache.c max(get_max_extent_size(entry), entry 1868 fs/btrfs/free-space-cache.c *bytes = entry->bytes - align_off; entry 1869 fs/btrfs/free-space-cache.c return entry; entry 2074 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry; entry 2086 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); entry 2087 fs/btrfs/free-space-cache.c if (!entry->bitmap) { entry 2092 fs/btrfs/free-space-cache.c if (entry->offset == offset_to_bitmap(ctl, offset)) { entry 2093 fs/btrfs/free-space-cache.c bytes_added = add_bytes_to_bitmap(ctl, entry, entry 2545 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry; entry 2560 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); entry 2561 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); entry 2562 fs/btrfs/free-space-cache.c rb_erase(&entry->offset_index, &cluster->root); entry 2563 fs/btrfs/free-space-cache.c RB_CLEAR_NODE(&entry->offset_index); entry 2565 fs/btrfs/free-space-cache.c bitmap = (entry->bitmap != NULL); entry 2567 fs/btrfs/free-space-cache.c try_merge_free_space(ctl, entry, false); entry 2568 fs/btrfs/free-space-cache.c steal_from_bitmap(ctl, entry, false); entry 2571 fs/btrfs/free-space-cache.c entry->offset, &entry->offset_index, bitmap); entry 2634 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry = NULL; entry 2641 fs/btrfs/free-space-cache.c entry = find_free_space(ctl, &offset, &bytes_search, entry 2643 fs/btrfs/free-space-cache.c if (!entry) entry 2647 fs/btrfs/free-space-cache.c if (entry->bitmap) { entry 2648 fs/btrfs/free-space-cache.c bitmap_clear_bits(ctl, entry, offset, bytes); entry 2649 fs/btrfs/free-space-cache.c if (!entry->bytes) entry 2650 fs/btrfs/free-space-cache.c free_bitmap(ctl, entry); entry 2652 fs/btrfs/free-space-cache.c unlink_free_space(ctl, entry); entry 2653 fs/btrfs/free-space-cache.c align_gap_len = offset - entry->offset; entry 2654 fs/btrfs/free-space-cache.c align_gap = entry->offset; entry 2656 fs/btrfs/free-space-cache.c entry->offset = offset + bytes; entry 2657 fs/btrfs/free-space-cache.c WARN_ON(entry->bytes < bytes + align_gap_len); entry 2659 fs/btrfs/free-space-cache.c entry->bytes -= bytes + align_gap_len; entry 2660 fs/btrfs/free-space-cache.c if (!entry->bytes) entry 2661 fs/btrfs/free-space-cache.c kmem_cache_free(btrfs_free_space_cachep, entry); entry 2663 fs/btrfs/free-space-cache.c link_free_space(ctl, entry); entry 2719 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, entry 2732 fs/btrfs/free-space-cache.c err = search_bitmap(ctl, entry, &search_start, &search_bytes, true); entry 2734 fs/btrfs/free-space-cache.c *max_extent_size = max(get_max_extent_size(entry), entry 2740 fs/btrfs/free-space-cache.c __bitmap_clear_bits(ctl, entry, ret, bytes); entry 2755 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry = NULL; entry 2770 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); entry 2772 fs/btrfs/free-space-cache.c if (entry->bytes < bytes) entry 2773 fs/btrfs/free-space-cache.c *max_extent_size = max(get_max_extent_size(entry), entry 2776 fs/btrfs/free-space-cache.c if (entry->bytes < bytes || entry 2777 fs/btrfs/free-space-cache.c (!entry->bitmap && entry->offset < min_start)) { entry 2778 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); entry 2781 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, entry 2786 fs/btrfs/free-space-cache.c if (entry->bitmap) { entry 2788 fs/btrfs/free-space-cache.c cluster, entry, bytes, entry 2792 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); entry 2795 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, entry 2801 fs/btrfs/free-space-cache.c ret = entry->offset; entry 2803 fs/btrfs/free-space-cache.c entry->offset += bytes; entry 2804 fs/btrfs/free-space-cache.c entry->bytes -= bytes; entry 2807 fs/btrfs/free-space-cache.c if (entry->bytes == 0) entry 2808 fs/btrfs/free-space-cache.c rb_erase(&entry->offset_index, &cluster->root); entry 2820 fs/btrfs/free-space-cache.c if (entry->bytes == 0) { entry 2822 fs/btrfs/free-space-cache.c if (entry->bitmap) { entry 2824 fs/btrfs/free-space-cache.c entry->bitmap); entry 2828 fs/btrfs/free-space-cache.c kmem_cache_free(btrfs_free_space_cachep, entry); entry 2837 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, entry 2853 fs/btrfs/free-space-cache.c i = offset_to_bit(entry->offset, ctl->unit, entry 2854 fs/btrfs/free-space-cache.c max_t(u64, offset, entry->offset)); entry 2862 fs/btrfs/free-space-cache.c if (entry->max_extent_size && entry 2863 fs/btrfs/free-space-cache.c entry->max_extent_size < cont1_bytes) entry 2867 fs/btrfs/free-space-cache.c for_each_set_bit_from(i, entry->bitmap, BITS_PER_BITMAP) { entry 2868 fs/btrfs/free-space-cache.c next_zero = find_next_zero_bit(entry->bitmap, entry 2882 fs/btrfs/free-space-cache.c entry->max_extent_size = (u64)max_bits * ctl->unit; entry 2901 fs/btrfs/free-space-cache.c cluster->window_start = start * ctl->unit + entry->offset; entry 2902 fs/btrfs/free-space-cache.c rb_erase(&entry->offset_index, &ctl->free_space_offset); entry 2903 fs/btrfs/free-space-cache.c ret = tree_insert_offset(&cluster->root, entry->offset, entry 2904 fs/btrfs/free-space-cache.c &entry->offset_index, 1); entry 2925 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry = NULL; entry 2932 fs/btrfs/free-space-cache.c entry = tree_search_offset(ctl, offset, 0, 1); entry 2933 fs/btrfs/free-space-cache.c if (!entry) entry 2940 fs/btrfs/free-space-cache.c while (entry->bitmap || entry->bytes < min_bytes) { entry 2941 fs/btrfs/free-space-cache.c if (entry->bitmap && list_empty(&entry->list)) entry 2942 fs/btrfs/free-space-cache.c list_add_tail(&entry->list, bitmaps); entry 2943 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); entry 2946 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); entry 2949 fs/btrfs/free-space-cache.c window_free = entry->bytes; entry 2950 fs/btrfs/free-space-cache.c max_extent = entry->bytes; entry 2951 fs/btrfs/free-space-cache.c first = entry; entry 2952 fs/btrfs/free-space-cache.c last = entry; entry 2954 fs/btrfs/free-space-cache.c for (node = rb_next(&entry->offset_index); node; entry 2955 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index)) { entry 2956 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); entry 2958 fs/btrfs/free-space-cache.c if (entry->bitmap) { entry 2959 fs/btrfs/free-space-cache.c if (list_empty(&entry->list)) entry 2960 fs/btrfs/free-space-cache.c list_add_tail(&entry->list, bitmaps); entry 2964 fs/btrfs/free-space-cache.c if (entry->bytes < min_bytes) entry 2967 fs/btrfs/free-space-cache.c last = entry; entry 2968 fs/btrfs/free-space-cache.c window_free += entry->bytes; entry 2969 fs/btrfs/free-space-cache.c if (entry->bytes > max_extent) entry 2970 fs/btrfs/free-space-cache.c max_extent = entry->bytes; entry 2987 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, offset_index); entry 2988 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); entry 2989 fs/btrfs/free-space-cache.c if (entry->bitmap || entry->bytes < min_bytes) entry 2992 fs/btrfs/free-space-cache.c rb_erase(&entry->offset_index, &ctl->free_space_offset); entry 2993 fs/btrfs/free-space-cache.c ret = tree_insert_offset(&cluster->root, entry->offset, entry 2994 fs/btrfs/free-space-cache.c &entry->offset_index, 0); entry 2995 fs/btrfs/free-space-cache.c total_size += entry->bytes; entry 2997 fs/btrfs/free-space-cache.c } while (node && entry != last); entry 3015 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry = NULL; entry 3027 fs/btrfs/free-space-cache.c entry = list_first_entry(bitmaps, struct btrfs_free_space, list); entry 3029 fs/btrfs/free-space-cache.c if (!entry || entry->offset != bitmap_offset) { entry 3030 fs/btrfs/free-space-cache.c entry = tree_search_offset(ctl, bitmap_offset, 1, 0); entry 3031 fs/btrfs/free-space-cache.c if (entry && list_empty(&entry->list)) entry 3032 fs/btrfs/free-space-cache.c list_add(&entry->list, bitmaps); entry 3035 fs/btrfs/free-space-cache.c list_for_each_entry(entry, bitmaps, list) { entry 3036 fs/btrfs/free-space-cache.c if (entry->bytes < bytes) entry 3038 fs/btrfs/free-space-cache.c ret = btrfs_bitmap_cluster(block_group, entry, cluster, offset, entry 3065 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry, *tmp; entry 3118 fs/btrfs/free-space-cache.c list_for_each_entry_safe(entry, tmp, &bitmaps, list) entry 3119 fs/btrfs/free-space-cache.c list_del_init(&entry->list); entry 3199 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry; entry 3218 fs/btrfs/free-space-cache.c entry = tree_search_offset(ctl, start, 0, 1); entry 3219 fs/btrfs/free-space-cache.c if (!entry) { entry 3226 fs/btrfs/free-space-cache.c while (entry->bitmap) { entry 3227 fs/btrfs/free-space-cache.c node = rb_next(&entry->offset_index); entry 3233 fs/btrfs/free-space-cache.c entry = rb_entry(node, struct btrfs_free_space, entry 3237 fs/btrfs/free-space-cache.c if (entry->offset >= end) { entry 3243 fs/btrfs/free-space-cache.c extent_start = entry->offset; entry 3244 fs/btrfs/free-space-cache.c extent_bytes = entry->bytes; entry 3253 fs/btrfs/free-space-cache.c unlink_free_space(ctl, entry); entry 3254 fs/btrfs/free-space-cache.c kmem_cache_free(btrfs_free_space_cachep, entry); entry 3284 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry; entry 3303 fs/btrfs/free-space-cache.c entry = tree_search_offset(ctl, offset, 1, 0); entry 3304 fs/btrfs/free-space-cache.c if (!entry) { entry 3312 fs/btrfs/free-space-cache.c ret2 = search_bitmap(ctl, entry, &start, &bytes, false); entry 3327 fs/btrfs/free-space-cache.c bitmap_clear_bits(ctl, entry, start, bytes); entry 3328 fs/btrfs/free-space-cache.c if (entry->bytes == 0) entry 3329 fs/btrfs/free-space-cache.c free_bitmap(ctl, entry); entry 3436 fs/btrfs/free-space-cache.c struct btrfs_free_space *entry = NULL; entry 3444 fs/btrfs/free-space-cache.c entry = rb_entry(rb_first(&ctl->free_space_offset), entry 3447 fs/btrfs/free-space-cache.c if (!entry->bitmap) { entry 3448 fs/btrfs/free-space-cache.c ino = entry->offset; entry 3450 fs/btrfs/free-space-cache.c unlink_free_space(ctl, entry); entry 3451 fs/btrfs/free-space-cache.c entry->offset++; entry 3452 fs/btrfs/free-space-cache.c entry->bytes--; entry 3453 fs/btrfs/free-space-cache.c if (!entry->bytes) entry 3454 fs/btrfs/free-space-cache.c kmem_cache_free(btrfs_free_space_cachep, entry); entry 3456 fs/btrfs/free-space-cache.c link_free_space(ctl, entry); entry 3462 fs/btrfs/free-space-cache.c ret = search_bitmap(ctl, entry, &offset, &count, true); entry 3467 fs/btrfs/free-space-cache.c bitmap_clear_bits(ctl, entry, offset, 1); entry 3468 fs/btrfs/free-space-cache.c if (entry->bytes == 0) entry 3469 fs/btrfs/free-space-cache.c free_bitmap(ctl, entry); entry 2518 fs/btrfs/inode.c struct sa_defrag_extent_backref *entry; entry 2523 fs/btrfs/inode.c entry = rb_entry(parent, struct sa_defrag_extent_backref, node); entry 2525 fs/btrfs/inode.c ret = backref_comp(backref, entry); entry 4447 fs/btrfs/inode.c struct btrfs_inode *entry; entry 4460 fs/btrfs/inode.c entry = rb_entry(node, struct btrfs_inode, rb_node); entry 4462 fs/btrfs/inode.c if (objectid < btrfs_ino(entry)) entry 4464 fs/btrfs/inode.c else if (objectid > btrfs_ino(entry)) entry 4471 fs/btrfs/inode.c entry = rb_entry(prev, struct btrfs_inode, rb_node); entry 4472 fs/btrfs/inode.c if (objectid <= btrfs_ino(entry)) { entry 4480 fs/btrfs/inode.c entry = rb_entry(node, struct btrfs_inode, rb_node); entry 4481 fs/btrfs/inode.c objectid = btrfs_ino(entry) + 1; entry 4482 fs/btrfs/inode.c inode = igrab(&entry->vfs_inode); entry 5770 fs/btrfs/inode.c struct btrfs_inode *entry; entry 5783 fs/btrfs/inode.c entry = rb_entry(parent, struct btrfs_inode, rb_node); entry 5785 fs/btrfs/inode.c if (ino < btrfs_ino(entry)) entry 5787 fs/btrfs/inode.c else if (ino > btrfs_ino(entry)) entry 5790 fs/btrfs/inode.c WARN_ON(!(entry->vfs_inode.i_state & entry 6074 fs/btrfs/inode.c struct dir_entry *entry = addr; entry 6075 fs/btrfs/inode.c char *name = (char *)(entry + 1); entry 6077 fs/btrfs/inode.c ctx->pos = get_unaligned(&entry->offset); entry 6078 fs/btrfs/inode.c if (!dir_emit(ctx, name, get_unaligned(&entry->name_len), entry 6079 fs/btrfs/inode.c get_unaligned(&entry->ino), entry 6080 fs/btrfs/inode.c get_unaligned(&entry->type))) entry 6083 fs/btrfs/inode.c get_unaligned(&entry->name_len); entry 6135 fs/btrfs/inode.c struct dir_entry *entry; entry 6172 fs/btrfs/inode.c entry = addr; entry 6173 fs/btrfs/inode.c put_unaligned(name_len, &entry->name_len); entry 6174 fs/btrfs/inode.c name_ptr = (char *)(entry + 1); entry 6178 fs/btrfs/inode.c &entry->type); entry 6180 fs/btrfs/inode.c put_unaligned(location.objectid, &entry->ino); entry 6181 fs/btrfs/inode.c put_unaligned(found_key.offset, &entry->offset); entry 10738 fs/btrfs/inode.c struct btrfs_swapfile_pin *sp, *entry; entry 10753 fs/btrfs/inode.c entry = rb_entry(parent, struct btrfs_swapfile_pin, node); entry 10754 fs/btrfs/inode.c if (sp->ptr < entry->ptr || entry 10755 fs/btrfs/inode.c (sp->ptr == entry->ptr && sp->inode < entry->inode)) { entry 10757 fs/btrfs/inode.c } else if (sp->ptr > entry->ptr || entry 10758 fs/btrfs/inode.c (sp->ptr == entry->ptr && sp->inode > entry->inode)) { entry 21 fs/btrfs/ordered-data.c static u64 entry_end(struct btrfs_ordered_extent *entry) entry 23 fs/btrfs/ordered-data.c if (entry->file_offset + entry->len < entry->file_offset) entry 25 fs/btrfs/ordered-data.c return entry->file_offset + entry->len; entry 36 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry; entry 40 fs/btrfs/ordered-data.c entry = rb_entry(parent, struct btrfs_ordered_extent, rb_node); entry 42 fs/btrfs/ordered-data.c if (file_offset < entry->file_offset) entry 44 fs/btrfs/ordered-data.c else if (file_offset >= entry_end(entry)) entry 73 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry; entry 77 fs/btrfs/ordered-data.c entry = rb_entry(n, struct btrfs_ordered_extent, rb_node); entry 79 fs/btrfs/ordered-data.c prev_entry = entry; entry 81 fs/btrfs/ordered-data.c if (file_offset < entry->file_offset) entry 83 fs/btrfs/ordered-data.c else if (file_offset >= entry_end(entry)) entry 120 fs/btrfs/ordered-data.c static int offset_in_entry(struct btrfs_ordered_extent *entry, u64 file_offset) entry 122 fs/btrfs/ordered-data.c if (file_offset < entry->file_offset || entry 123 fs/btrfs/ordered-data.c entry->file_offset + entry->len <= file_offset) entry 128 fs/btrfs/ordered-data.c static int range_overlaps(struct btrfs_ordered_extent *entry, u64 file_offset, entry 131 fs/btrfs/ordered-data.c if (file_offset + len <= entry->file_offset || entry 132 fs/btrfs/ordered-data.c entry->file_offset + entry->len <= file_offset) entry 147 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry; entry 150 fs/btrfs/ordered-data.c entry = rb_entry(tree->last, struct btrfs_ordered_extent, entry 152 fs/btrfs/ordered-data.c if (offset_in_entry(entry, file_offset)) entry 182 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry; entry 185 fs/btrfs/ordered-data.c entry = kmem_cache_zalloc(btrfs_ordered_extent_cache, GFP_NOFS); entry 186 fs/btrfs/ordered-data.c if (!entry) entry 189 fs/btrfs/ordered-data.c entry->file_offset = file_offset; entry 190 fs/btrfs/ordered-data.c entry->start = start; entry 191 fs/btrfs/ordered-data.c entry->len = len; entry 192 fs/btrfs/ordered-data.c entry->disk_len = disk_len; entry 193 fs/btrfs/ordered-data.c entry->bytes_left = len; entry 194 fs/btrfs/ordered-data.c entry->inode = igrab(inode); entry 195 fs/btrfs/ordered-data.c entry->compress_type = compress_type; entry 196 fs/btrfs/ordered-data.c entry->truncated_len = (u64)-1; entry 198 fs/btrfs/ordered-data.c set_bit(type, &entry->flags); entry 203 fs/btrfs/ordered-data.c set_bit(BTRFS_ORDERED_DIRECT, &entry->flags); entry 207 fs/btrfs/ordered-data.c refcount_set(&entry->refs, 1); entry 208 fs/btrfs/ordered-data.c init_waitqueue_head(&entry->wait); entry 209 fs/btrfs/ordered-data.c INIT_LIST_HEAD(&entry->list); entry 210 fs/btrfs/ordered-data.c INIT_LIST_HEAD(&entry->root_extent_list); entry 211 fs/btrfs/ordered-data.c INIT_LIST_HEAD(&entry->work_list); entry 212 fs/btrfs/ordered-data.c init_completion(&entry->completion); entry 213 fs/btrfs/ordered-data.c INIT_LIST_HEAD(&entry->log_list); entry 214 fs/btrfs/ordered-data.c INIT_LIST_HEAD(&entry->trans_list); entry 216 fs/btrfs/ordered-data.c trace_btrfs_ordered_extent_add(inode, entry); entry 220 fs/btrfs/ordered-data.c &entry->rb_node); entry 226 fs/btrfs/ordered-data.c list_add_tail(&entry->root_extent_list, entry 279 fs/btrfs/ordered-data.c void btrfs_add_ordered_sum(struct btrfs_ordered_extent *entry, entry 284 fs/btrfs/ordered-data.c tree = &BTRFS_I(entry->inode)->ordered_tree; entry 286 fs/btrfs/ordered-data.c list_add_tail(&sum->list, &entry->list); entry 309 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry = NULL; entry 324 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); entry 325 fs/btrfs/ordered-data.c if (!offset_in_entry(entry, *file_offset)) { entry 330 fs/btrfs/ordered-data.c dec_start = max(*file_offset, entry->file_offset); entry 331 fs/btrfs/ordered-data.c dec_end = min(*file_offset + io_size, entry->file_offset + entry 332 fs/btrfs/ordered-data.c entry->len); entry 339 fs/btrfs/ordered-data.c if (to_dec > entry->bytes_left) { entry 342 fs/btrfs/ordered-data.c entry->bytes_left, to_dec); entry 344 fs/btrfs/ordered-data.c entry->bytes_left -= to_dec; entry 346 fs/btrfs/ordered-data.c set_bit(BTRFS_ORDERED_IOERR, &entry->flags); entry 348 fs/btrfs/ordered-data.c if (entry->bytes_left == 0) { entry 349 fs/btrfs/ordered-data.c ret = test_and_set_bit(BTRFS_ORDERED_IO_DONE, &entry->flags); entry 351 fs/btrfs/ordered-data.c cond_wake_up_nomb(&entry->wait); entry 356 fs/btrfs/ordered-data.c if (!ret && cached && entry) { entry 357 fs/btrfs/ordered-data.c *cached = entry; entry 358 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); entry 379 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry = NULL; entry 386 fs/btrfs/ordered-data.c entry = *cached; entry 396 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); entry 398 fs/btrfs/ordered-data.c if (!offset_in_entry(entry, file_offset)) { entry 403 fs/btrfs/ordered-data.c if (io_size > entry->bytes_left) { entry 406 fs/btrfs/ordered-data.c entry->bytes_left, io_size); entry 408 fs/btrfs/ordered-data.c entry->bytes_left -= io_size; entry 410 fs/btrfs/ordered-data.c set_bit(BTRFS_ORDERED_IOERR, &entry->flags); entry 412 fs/btrfs/ordered-data.c if (entry->bytes_left == 0) { entry 413 fs/btrfs/ordered-data.c ret = test_and_set_bit(BTRFS_ORDERED_IO_DONE, &entry->flags); entry 415 fs/btrfs/ordered-data.c cond_wake_up_nomb(&entry->wait); entry 420 fs/btrfs/ordered-data.c if (!ret && cached && entry) { entry 421 fs/btrfs/ordered-data.c *cached = entry; entry 422 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); entry 432 fs/btrfs/ordered-data.c void btrfs_put_ordered_extent(struct btrfs_ordered_extent *entry) entry 437 fs/btrfs/ordered-data.c trace_btrfs_ordered_extent_put(entry->inode, entry); entry 439 fs/btrfs/ordered-data.c if (refcount_dec_and_test(&entry->refs)) { entry 440 fs/btrfs/ordered-data.c ASSERT(list_empty(&entry->log_list)); entry 441 fs/btrfs/ordered-data.c ASSERT(list_empty(&entry->trans_list)); entry 442 fs/btrfs/ordered-data.c ASSERT(list_empty(&entry->root_extent_list)); entry 443 fs/btrfs/ordered-data.c ASSERT(RB_EMPTY_NODE(&entry->rb_node)); entry 444 fs/btrfs/ordered-data.c if (entry->inode) entry 445 fs/btrfs/ordered-data.c btrfs_add_delayed_iput(entry->inode); entry 446 fs/btrfs/ordered-data.c while (!list_empty(&entry->list)) { entry 447 fs/btrfs/ordered-data.c cur = entry->list.next; entry 452 fs/btrfs/ordered-data.c kmem_cache_free(btrfs_ordered_extent_cache, entry); entry 461 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry) entry 474 fs/btrfs/ordered-data.c btrfs_delalloc_release_metadata(btrfs_inode, entry->len, false); entry 476 fs/btrfs/ordered-data.c if (test_bit(BTRFS_ORDERED_DIRECT, &entry->flags)) entry 477 fs/btrfs/ordered-data.c percpu_counter_add_batch(&fs_info->dio_bytes, -entry->len, entry 482 fs/btrfs/ordered-data.c node = &entry->rb_node; entry 487 fs/btrfs/ordered-data.c set_bit(BTRFS_ORDERED_COMPLETE, &entry->flags); entry 491 fs/btrfs/ordered-data.c list_del_init(&entry->root_extent_list); entry 494 fs/btrfs/ordered-data.c trace_btrfs_ordered_extent_remove(inode, entry); entry 503 fs/btrfs/ordered-data.c wake_up(&entry->wait); entry 622 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry, entry 625 fs/btrfs/ordered-data.c u64 start = entry->file_offset; entry 626 fs/btrfs/ordered-data.c u64 end = start + entry->len - 1; entry 628 fs/btrfs/ordered-data.c trace_btrfs_ordered_extent_start(inode, entry); entry 635 fs/btrfs/ordered-data.c if (!test_bit(BTRFS_ORDERED_DIRECT, &entry->flags)) entry 638 fs/btrfs/ordered-data.c wait_event(entry->wait, test_bit(BTRFS_ORDERED_COMPLETE, entry 639 fs/btrfs/ordered-data.c &entry->flags)); entry 717 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry = NULL; entry 725 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); entry 726 fs/btrfs/ordered-data.c if (!offset_in_entry(entry, file_offset)) entry 727 fs/btrfs/ordered-data.c entry = NULL; entry 728 fs/btrfs/ordered-data.c if (entry) entry 729 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); entry 732 fs/btrfs/ordered-data.c return entry; entry 743 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry = NULL; entry 755 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); entry 756 fs/btrfs/ordered-data.c if (range_overlaps(entry, file_offset, len)) entry 759 fs/btrfs/ordered-data.c if (entry->file_offset >= file_offset + len) { entry 760 fs/btrfs/ordered-data.c entry = NULL; entry 763 fs/btrfs/ordered-data.c entry = NULL; entry 769 fs/btrfs/ordered-data.c if (entry) entry 770 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); entry 772 fs/btrfs/ordered-data.c return entry; entry 784 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *entry = NULL; entry 792 fs/btrfs/ordered-data.c entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); entry 793 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); entry 796 fs/btrfs/ordered-data.c return entry; entry 153 fs/btrfs/ordered-data.h void btrfs_put_ordered_extent(struct btrfs_ordered_extent *entry); entry 155 fs/btrfs/ordered-data.h struct btrfs_ordered_extent *entry); entry 170 fs/btrfs/ordered-data.h void btrfs_add_ordered_sum(struct btrfs_ordered_extent *entry, entry 175 fs/btrfs/ordered-data.h struct btrfs_ordered_extent *entry, int wait); entry 1550 fs/btrfs/qgroup.c struct btrfs_qgroup_extent_record *entry; entry 1558 fs/btrfs/qgroup.c entry = rb_entry(parent_node, struct btrfs_qgroup_extent_record, entry 1560 fs/btrfs/qgroup.c if (bytenr < entry->bytenr) { entry 1562 fs/btrfs/qgroup.c } else if (bytenr > entry->bytenr) { entry 1565 fs/btrfs/qgroup.c if (record->data_rsv && !entry->data_rsv) { entry 1566 fs/btrfs/qgroup.c entry->data_rsv = record->data_rsv; entry 1567 fs/btrfs/qgroup.c entry->data_rsv_refroot = entry 3814 fs/btrfs/qgroup.c struct btrfs_qgroup_swapped_block *entry; entry 3817 fs/btrfs/qgroup.c rbtree_postorder_for_each_entry_safe(entry, next, cur_root, entry 3819 fs/btrfs/qgroup.c kfree(entry); entry 3899 fs/btrfs/qgroup.c struct btrfs_qgroup_swapped_block *entry; entry 3902 fs/btrfs/qgroup.c entry = rb_entry(parent, struct btrfs_qgroup_swapped_block, entry 3905 fs/btrfs/qgroup.c if (entry->subvol_bytenr < block->subvol_bytenr) { entry 3907 fs/btrfs/qgroup.c } else if (entry->subvol_bytenr > block->subvol_bytenr) { entry 3910 fs/btrfs/qgroup.c if (entry->subvol_generation != entry 3912 fs/btrfs/qgroup.c entry->reloc_bytenr != block->reloc_bytenr || entry 3913 fs/btrfs/qgroup.c entry->reloc_generation != entry 4031 fs/btrfs/qgroup.c struct btrfs_qgroup_extent_record *entry; entry 4036 fs/btrfs/qgroup.c rbtree_postorder_for_each_entry_safe(entry, next, root, node) { entry 4037 fs/btrfs/qgroup.c ulist_free(entry->old_roots); entry 4038 fs/btrfs/qgroup.c kfree(entry); entry 80 fs/btrfs/ref-verify.c struct block_entry *entry; entry 84 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct block_entry, node); entry 85 fs/btrfs/ref-verify.c if (entry->bytenr > be->bytenr) entry 87 fs/btrfs/ref-verify.c else if (entry->bytenr < be->bytenr) entry 90 fs/btrfs/ref-verify.c return entry; entry 101 fs/btrfs/ref-verify.c struct block_entry *entry = NULL; entry 105 fs/btrfs/ref-verify.c entry = rb_entry(n, struct block_entry, node); entry 106 fs/btrfs/ref-verify.c if (entry->bytenr < bytenr) entry 108 fs/btrfs/ref-verify.c else if (entry->bytenr > bytenr) entry 111 fs/btrfs/ref-verify.c return entry; entry 121 fs/btrfs/ref-verify.c struct root_entry *entry; entry 125 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct root_entry, node); entry 126 fs/btrfs/ref-verify.c if (entry->root_objectid > re->root_objectid) entry 128 fs/btrfs/ref-verify.c else if (entry->root_objectid < re->root_objectid) entry 131 fs/btrfs/ref-verify.c return entry; entry 166 fs/btrfs/ref-verify.c struct ref_entry *entry; entry 171 fs/btrfs/ref-verify.c entry = rb_entry(parent_node, struct ref_entry, node); entry 172 fs/btrfs/ref-verify.c cmp = comp_refs(entry, ref); entry 178 fs/btrfs/ref-verify.c return entry; entry 190 fs/btrfs/ref-verify.c struct root_entry *entry = NULL; entry 194 fs/btrfs/ref-verify.c entry = rb_entry(n, struct root_entry, node); entry 195 fs/btrfs/ref-verify.c if (entry->root_objectid < objectid) entry 197 fs/btrfs/ref-verify.c else if (entry->root_objectid > objectid) entry 200 fs/btrfs/ref-verify.c return entry; entry 915 fs/btrfs/ref-verify.c struct block_entry *be = NULL, *entry; entry 924 fs/btrfs/ref-verify.c entry = rb_entry(n, struct block_entry, node); entry 925 fs/btrfs/ref-verify.c if (entry->bytenr < start) { entry 927 fs/btrfs/ref-verify.c } else if (entry->bytenr > start) { entry 930 fs/btrfs/ref-verify.c be = entry; entry 935 fs/btrfs/ref-verify.c (entry->bytenr < start && be->bytenr > start) || entry 936 fs/btrfs/ref-verify.c (entry->bytenr < start && entry->bytenr > be->bytenr)) entry 937 fs/btrfs/ref-verify.c be = entry; entry 287 fs/btrfs/relocation.c struct tree_entry *entry; entry 291 fs/btrfs/relocation.c entry = rb_entry(parent, struct tree_entry, rb_node); entry 293 fs/btrfs/relocation.c if (bytenr < entry->bytenr) entry 295 fs/btrfs/relocation.c else if (bytenr > entry->bytenr) entry 309 fs/btrfs/relocation.c struct tree_entry *entry; entry 312 fs/btrfs/relocation.c entry = rb_entry(n, struct tree_entry, rb_node); entry 314 fs/btrfs/relocation.c if (bytenr < entry->bytenr) entry 316 fs/btrfs/relocation.c else if (bytenr > entry->bytenr) entry 1555 fs/btrfs/relocation.c struct btrfs_inode *entry; entry 1564 fs/btrfs/relocation.c entry = rb_entry(node, struct btrfs_inode, rb_node); entry 1566 fs/btrfs/relocation.c if (objectid < btrfs_ino(entry)) entry 1568 fs/btrfs/relocation.c else if (objectid > btrfs_ino(entry)) entry 1575 fs/btrfs/relocation.c entry = rb_entry(prev, struct btrfs_inode, rb_node); entry 1576 fs/btrfs/relocation.c if (objectid <= btrfs_ino(entry)) { entry 1584 fs/btrfs/relocation.c entry = rb_entry(node, struct btrfs_inode, rb_node); entry 1585 fs/btrfs/relocation.c inode = igrab(&entry->vfs_inode); entry 1591 fs/btrfs/relocation.c objectid = btrfs_ino(entry) + 1; entry 325 fs/btrfs/scrub.c struct full_stripe_lock *entry; entry 333 fs/btrfs/scrub.c entry = rb_entry(parent, struct full_stripe_lock, node); entry 334 fs/btrfs/scrub.c if (fstripe_logical < entry->logical) { entry 336 fs/btrfs/scrub.c } else if (fstripe_logical > entry->logical) { entry 339 fs/btrfs/scrub.c entry->refs++; entry 340 fs/btrfs/scrub.c return entry; entry 370 fs/btrfs/scrub.c struct full_stripe_lock *entry; entry 376 fs/btrfs/scrub.c entry = rb_entry(node, struct full_stripe_lock, node); entry 377 fs/btrfs/scrub.c if (fstripe_logical < entry->logical) entry 379 fs/btrfs/scrub.c else if (fstripe_logical > entry->logical) entry 382 fs/btrfs/scrub.c return entry; entry 2863 fs/btrfs/send.c struct orphan_dir_info *entry, *odi; entry 2867 fs/btrfs/send.c entry = rb_entry(parent, struct orphan_dir_info, node); entry 2868 fs/btrfs/send.c if (dir_ino < entry->ino) { entry 2870 fs/btrfs/send.c } else if (dir_ino > entry->ino) { entry 2873 fs/btrfs/send.c return entry; entry 2893 fs/btrfs/send.c struct orphan_dir_info *entry; entry 2896 fs/btrfs/send.c entry = rb_entry(n, struct orphan_dir_info, node); entry 2897 fs/btrfs/send.c if (dir_ino < entry->ino) entry 2899 fs/btrfs/send.c else if (dir_ino > entry->ino) entry 2902 fs/btrfs/send.c return entry; entry 3022 fs/btrfs/send.c struct waiting_dir_move *entry = get_waiting_dir_move(sctx, ino); entry 3024 fs/btrfs/send.c return entry != NULL; entry 3031 fs/btrfs/send.c struct waiting_dir_move *entry, *dm; entry 3042 fs/btrfs/send.c entry = rb_entry(parent, struct waiting_dir_move, node); entry 3043 fs/btrfs/send.c if (ino < entry->ino) { entry 3045 fs/btrfs/send.c } else if (ino > entry->ino) { entry 3062 fs/btrfs/send.c struct waiting_dir_move *entry; entry 3065 fs/btrfs/send.c entry = rb_entry(n, struct waiting_dir_move, node); entry 3066 fs/btrfs/send.c if (ino < entry->ino) entry 3068 fs/btrfs/send.c else if (ino > entry->ino) entry 3071 fs/btrfs/send.c return entry; entry 3095 fs/btrfs/send.c struct pending_dir_move *entry = NULL, *pm; entry 3112 fs/btrfs/send.c entry = rb_entry(parent, struct pending_dir_move, node); entry 3113 fs/btrfs/send.c if (parent_ino < entry->parent_ino) { entry 3115 fs/btrfs/send.c } else if (parent_ino > entry->parent_ino) { entry 3139 fs/btrfs/send.c list_add_tail(&pm->list, &entry->list); entry 3157 fs/btrfs/send.c struct pending_dir_move *entry; entry 3160 fs/btrfs/send.c entry = rb_entry(n, struct pending_dir_move, node); entry 3161 fs/btrfs/send.c if (parent_ino < entry->parent_ino) entry 3163 fs/btrfs/send.c else if (parent_ino > entry->parent_ino) entry 3166 fs/btrfs/send.c return entry; entry 48 fs/cachefiles/rdwr.c list_del(&wait->entry); entry 41 fs/coda/dir.c static struct dentry *coda_lookup(struct inode *dir, struct dentry *entry, unsigned int flags) entry 44 fs/coda/dir.c const char *name = entry->d_name.name; entry 45 fs/coda/dir.c size_t length = entry->d_name.len; entry 72 fs/coda/dir.c return d_splice_alias(inode, entry); entry 24 fs/configfs/item.c static inline struct config_item *to_item(struct list_head *entry) entry 26 fs/configfs/item.c return container_of(entry, struct config_item, ci_entry); entry 179 fs/configfs/item.c struct list_head *entry; entry 182 fs/configfs/item.c list_for_each(entry, &group->cg_children) { entry 183 fs/configfs/item.c struct config_item *item = to_item(entry); entry 83 fs/dax.c static unsigned long dax_to_pfn(void *entry) entry 85 fs/dax.c return xa_to_value(entry) >> DAX_SHIFT; entry 93 fs/dax.c static bool dax_is_locked(void *entry) entry 95 fs/dax.c return xa_to_value(entry) & DAX_LOCKED; entry 98 fs/dax.c static unsigned int dax_entry_order(void *entry) entry 100 fs/dax.c if (xa_to_value(entry) & DAX_PMD) entry 105 fs/dax.c static unsigned long dax_is_pmd_entry(void *entry) entry 107 fs/dax.c return xa_to_value(entry) & DAX_PMD; entry 110 fs/dax.c static bool dax_is_pte_entry(void *entry) entry 112 fs/dax.c return !(xa_to_value(entry) & DAX_PMD); entry 115 fs/dax.c static int dax_is_zero_entry(void *entry) entry 117 fs/dax.c return xa_to_value(entry) & DAX_ZERO_PAGE; entry 120 fs/dax.c static int dax_is_empty_entry(void *entry) entry 122 fs/dax.c return xa_to_value(entry) & DAX_EMPTY; entry 129 fs/dax.c static bool dax_is_conflict(void *entry) entry 131 fs/dax.c return entry == XA_RETRY_ENTRY; entry 148 fs/dax.c void *entry, struct exceptional_entry_key *key) entry 158 fs/dax.c if (dax_is_pmd_entry(entry)) entry 185 fs/dax.c static void dax_wake_entry(struct xa_state *xas, void *entry, bool wake_all) entry 190 fs/dax.c wq = dax_entry_waitqueue(xas, entry, &key); entry 214 fs/dax.c void *entry; entry 222 fs/dax.c entry = xas_find_conflict(xas); entry 223 fs/dax.c if (!entry || WARN_ON_ONCE(!xa_is_value(entry))) entry 224 fs/dax.c return entry; entry 225 fs/dax.c if (dax_entry_order(entry) < order) entry 227 fs/dax.c if (!dax_is_locked(entry)) entry 228 fs/dax.c return entry; entry 230 fs/dax.c wq = dax_entry_waitqueue(xas, entry, &ewait.key); entry 246 fs/dax.c static void wait_entry_unlocked(struct xa_state *xas, void *entry) entry 254 fs/dax.c wq = dax_entry_waitqueue(xas, entry, &ewait.key); entry 267 fs/dax.c static void put_unlocked_entry(struct xa_state *xas, void *entry) entry 270 fs/dax.c if (entry && !dax_is_conflict(entry)) entry 271 fs/dax.c dax_wake_entry(xas, entry, false); entry 279 fs/dax.c static void dax_unlock_entry(struct xa_state *xas, void *entry) entry 283 fs/dax.c BUG_ON(dax_is_locked(entry)); entry 286 fs/dax.c old = xas_store(xas, entry); entry 289 fs/dax.c dax_wake_entry(xas, entry, false); entry 295 fs/dax.c static void *dax_lock_entry(struct xa_state *xas, void *entry) entry 297 fs/dax.c unsigned long v = xa_to_value(entry); entry 301 fs/dax.c static unsigned long dax_entry_size(void *entry) entry 303 fs/dax.c if (dax_is_zero_entry(entry)) entry 305 fs/dax.c else if (dax_is_empty_entry(entry)) entry 307 fs/dax.c else if (dax_is_pmd_entry(entry)) entry 313 fs/dax.c static unsigned long dax_end_pfn(void *entry) entry 315 fs/dax.c return dax_to_pfn(entry) + dax_entry_size(entry) / PAGE_SIZE; entry 322 fs/dax.c #define for_each_mapped_pfn(entry, pfn) \ entry 323 fs/dax.c for (pfn = dax_to_pfn(entry); \ entry 324 fs/dax.c pfn < dax_end_pfn(entry); pfn++) entry 331 fs/dax.c static void dax_associate_entry(void *entry, struct address_space *mapping, entry 334 fs/dax.c unsigned long size = dax_entry_size(entry), pfn, index; entry 341 fs/dax.c for_each_mapped_pfn(entry, pfn) { entry 350 fs/dax.c static void dax_disassociate_entry(void *entry, struct address_space *mapping, entry 358 fs/dax.c for_each_mapped_pfn(entry, pfn) { entry 368 fs/dax.c static struct page *dax_busy_page(void *entry) entry 372 fs/dax.c for_each_mapped_pfn(entry, pfn) { entry 392 fs/dax.c void *entry; entry 399 fs/dax.c entry = NULL; entry 410 fs/dax.c entry = (void *)~0UL; entry 421 fs/dax.c entry = xas_load(&xas); entry 422 fs/dax.c if (dax_is_locked(entry)) { entry 424 fs/dax.c wait_entry_unlocked(&xas, entry); entry 428 fs/dax.c dax_lock_entry(&xas, entry); entry 433 fs/dax.c return (dax_entry_t)entry; entry 481 fs/dax.c void *entry; entry 485 fs/dax.c entry = get_unlocked_entry(xas, order); entry 487 fs/dax.c if (entry) { entry 488 fs/dax.c if (dax_is_conflict(entry)) entry 490 fs/dax.c if (!xa_is_value(entry)) { entry 496 fs/dax.c if (dax_is_pmd_entry(entry) && entry 497 fs/dax.c (dax_is_zero_entry(entry) || entry 498 fs/dax.c dax_is_empty_entry(entry))) { entry 509 fs/dax.c dax_lock_entry(xas, entry); entry 516 fs/dax.c if (dax_is_zero_entry(entry)) { entry 525 fs/dax.c dax_disassociate_entry(entry, mapping, false); entry 527 fs/dax.c dax_wake_entry(xas, entry, true); entry 529 fs/dax.c entry = NULL; entry 533 fs/dax.c if (entry) { entry 534 fs/dax.c dax_lock_entry(xas, entry); entry 540 fs/dax.c entry = dax_make_entry(pfn_to_pfn_t(0), flags); entry 541 fs/dax.c dax_lock_entry(xas, entry); entry 555 fs/dax.c return entry; entry 579 fs/dax.c void *entry; entry 607 fs/dax.c xas_for_each(&xas, entry, ULONG_MAX) { entry 608 fs/dax.c if (WARN_ON_ONCE(!xa_is_value(entry))) entry 610 fs/dax.c if (unlikely(dax_is_locked(entry))) entry 611 fs/dax.c entry = get_unlocked_entry(&xas, 0); entry 612 fs/dax.c if (entry) entry 613 fs/dax.c page = dax_busy_page(entry); entry 614 fs/dax.c put_unlocked_entry(&xas, entry); entry 635 fs/dax.c void *entry; entry 638 fs/dax.c entry = get_unlocked_entry(&xas, 0); entry 639 fs/dax.c if (!entry || WARN_ON_ONCE(!xa_is_value(entry))) entry 645 fs/dax.c dax_disassociate_entry(entry, mapping, trunc); entry 650 fs/dax.c put_unlocked_entry(&xas, entry); entry 718 fs/dax.c void *entry, pfn_t pfn, unsigned long flags, bool dirty) entry 725 fs/dax.c if (dax_is_zero_entry(entry) && !(flags & DAX_ZERO_PAGE)) { entry 728 fs/dax.c if (dax_is_pmd_entry(entry)) entry 737 fs/dax.c if (dax_is_zero_entry(entry) || dax_is_empty_entry(entry)) { entry 740 fs/dax.c dax_disassociate_entry(entry, mapping, false); entry 751 fs/dax.c WARN_ON_ONCE(old != xa_mk_value(xa_to_value(entry) | entry 753 fs/dax.c entry = new_entry; entry 762 fs/dax.c return entry; entry 850 fs/dax.c struct address_space *mapping, void *entry) entry 859 fs/dax.c if (WARN_ON(!xa_is_value(entry))) entry 862 fs/dax.c if (unlikely(dax_is_locked(entry))) { entry 863 fs/dax.c void *old_entry = entry; entry 865 fs/dax.c entry = get_unlocked_entry(xas, 0); entry 868 fs/dax.c if (!entry || WARN_ON_ONCE(!xa_is_value(entry))) entry 875 fs/dax.c if (dax_to_pfn(old_entry) != dax_to_pfn(entry)) entry 877 fs/dax.c if (WARN_ON_ONCE(dax_is_empty_entry(entry) || entry 878 fs/dax.c dax_is_zero_entry(entry))) { entry 889 fs/dax.c dax_lock_entry(xas, entry); entry 908 fs/dax.c pfn = dax_to_pfn(entry); entry 909 fs/dax.c count = 1UL << dax_entry_order(entry); entry 922 fs/dax.c xas_store(xas, entry); entry 924 fs/dax.c dax_wake_entry(xas, entry, false); entry 930 fs/dax.c put_unlocked_entry(xas, entry); entry 946 fs/dax.c void *entry; entry 965 fs/dax.c xas_for_each_marked(&xas, entry, end_index, PAGECACHE_TAG_TOWRITE) { entry 966 fs/dax.c ret = dax_writeback_one(&xas, dax_dev, mapping, entry); entry 1031 fs/dax.c struct address_space *mapping, void **entry, entry 1039 fs/dax.c *entry = dax_insert_entry(xas, mapping, vmf, *entry, pfn, entry 1260 fs/dax.c void *entry; entry 1277 fs/dax.c entry = grab_mapping_entry(&xas, mapping, 0); entry 1278 fs/dax.c if (xa_is_internal(entry)) { entry 1279 fs/dax.c ret = xa_to_internal(entry); entry 1352 fs/dax.c entry = dax_insert_entry(&xas, mapping, vmf, entry, pfn, entry 1370 fs/dax.c trace_dax_insert_mapping(inode, vmf, entry); entry 1380 fs/dax.c ret = dax_load_hole(&xas, mapping, &entry, vmf); entry 1407 fs/dax.c dax_unlock_entry(&xas, entry); entry 1415 fs/dax.c struct iomap *iomap, void **entry) entry 1433 fs/dax.c *entry = dax_insert_entry(xas, mapping, vmf, *entry, pfn, entry 1456 fs/dax.c trace_dax_pmd_load_hole(inode, vmf, zero_page, *entry); entry 1462 fs/dax.c trace_dax_pmd_load_hole_fallback(inode, vmf, zero_page, *entry); entry 1480 fs/dax.c void *entry; entry 1529 fs/dax.c entry = grab_mapping_entry(&xas, mapping, PMD_ORDER); entry 1530 fs/dax.c if (xa_is_internal(entry)) { entry 1531 fs/dax.c result = xa_to_internal(entry); entry 1568 fs/dax.c entry = dax_insert_entry(&xas, mapping, vmf, entry, pfn, entry 1585 fs/dax.c trace_dax_pmd_insert_mapping(inode, vmf, PMD_SIZE, pfn, entry); entry 1592 fs/dax.c result = dax_pmd_load_hole(&xas, vmf, &iomap, &entry); entry 1615 fs/dax.c dax_unlock_entry(&xas, entry); entry 1674 fs/dax.c void *entry; entry 1678 fs/dax.c entry = get_unlocked_entry(&xas, order); entry 1680 fs/dax.c if (!entry || dax_is_conflict(entry) || entry 1681 fs/dax.c (order == 0 && !dax_is_pte_entry(entry))) { entry 1682 fs/dax.c put_unlocked_entry(&xas, entry); entry 1689 fs/dax.c dax_lock_entry(&xas, entry); entry 1699 fs/dax.c dax_unlock_entry(&xas, entry); entry 1948 fs/dcache.c void d_instantiate(struct dentry *entry, struct inode * inode) entry 1950 fs/dcache.c BUG_ON(!hlist_unhashed(&entry->d_u.d_alias)); entry 1952 fs/dcache.c security_d_instantiate(entry, inode); entry 1954 fs/dcache.c __d_instantiate(entry, inode); entry 1966 fs/dcache.c void d_instantiate_new(struct dentry *entry, struct inode *inode) entry 1968 fs/dcache.c BUG_ON(!hlist_unhashed(&entry->d_u.d_alias)); entry 1971 fs/dcache.c security_d_instantiate(entry, inode); entry 1973 fs/dcache.c __d_instantiate(entry, inode); entry 2458 fs/dcache.c static void __d_rehash(struct dentry *entry) entry 2460 fs/dcache.c struct hlist_bl_head *b = d_hash(entry->d_name.hash); entry 2463 fs/dcache.c hlist_bl_add_head_rcu(&entry->d_hash, b); entry 2474 fs/dcache.c void d_rehash(struct dentry * entry) entry 2476 fs/dcache.c spin_lock(&entry->d_lock); entry 2477 fs/dcache.c __d_rehash(entry); entry 2478 fs/dcache.c spin_unlock(&entry->d_lock); entry 2674 fs/dcache.c void d_add(struct dentry *entry, struct inode *inode) entry 2677 fs/dcache.c security_d_instantiate(entry, inode); entry 2680 fs/dcache.c __d_add(entry, inode); entry 2695 fs/dcache.c struct dentry *d_exact_alias(struct dentry *entry, struct inode *inode) entry 2698 fs/dcache.c unsigned int hash = entry->d_name.hash; entry 2709 fs/dcache.c if (alias->d_parent != entry->d_parent) entry 2711 fs/dcache.c if (!d_same_name(alias, entry->d_parent, &entry->d_name)) entry 1159 fs/debugfs/file.c struct debugfs_devm_entry *entry = inode->i_private; entry 1161 fs/debugfs/file.c return single_open(f, entry->read, entry->dev); entry 1187 fs/debugfs/file.c struct debugfs_devm_entry *entry; entry 1192 fs/debugfs/file.c entry = devm_kzalloc(dev, sizeof(*entry), GFP_KERNEL); entry 1193 fs/debugfs/file.c if (!entry) entry 1196 fs/debugfs/file.c entry->read = read_fn; entry 1197 fs/debugfs/file.c entry->dev = dev; entry 1199 fs/debugfs/file.c return debugfs_create_file(name, S_IRUGO, parent, entry, entry 427 fs/dlm/debug_fs.c unsigned bucket, entry; entry 431 fs/dlm/debug_fs.c entry = n & ((1LL << 32) - 1); entry 456 fs/dlm/debug_fs.c if (!entry--) { entry 1402 fs/dlm/lowcomms.c struct writequeue_entry *entry; entry 1404 fs/dlm/lowcomms.c entry = kmalloc(sizeof(struct writequeue_entry), allocation); entry 1405 fs/dlm/lowcomms.c if (!entry) entry 1408 fs/dlm/lowcomms.c entry->page = alloc_page(allocation); entry 1409 fs/dlm/lowcomms.c if (!entry->page) { entry 1410 fs/dlm/lowcomms.c kfree(entry); entry 1414 fs/dlm/lowcomms.c entry->offset = 0; entry 1415 fs/dlm/lowcomms.c entry->len = 0; entry 1416 fs/dlm/lowcomms.c entry->end = 0; entry 1417 fs/dlm/lowcomms.c entry->users = 0; entry 1418 fs/dlm/lowcomms.c entry->con = con; entry 1420 fs/dlm/lowcomms.c return entry; entry 110 fs/efivarfs/super.c struct efivar_entry *entry; entry 119 fs/efivarfs/super.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 120 fs/efivarfs/super.c if (!entry) entry 123 fs/efivarfs/super.c memcpy(entry->var.VariableName, name16, name_size); entry 124 fs/efivarfs/super.c memcpy(&(entry->var.VendorGuid), &vendor, sizeof(efi_guid_t)); entry 126 fs/efivarfs/super.c len = ucs2_utf8size(entry->var.VariableName); entry 133 fs/efivarfs/super.c ucs2_as_utf8(name, entry->var.VariableName, len); entry 135 fs/efivarfs/super.c if (efivar_variable_is_removable(entry->var.VendorGuid, name, len)) entry 140 fs/efivarfs/super.c efi_guid_to_str(&entry->var.VendorGuid, name + len + 1); entry 155 fs/efivarfs/super.c efivar_entry_size(entry, &size); entry 156 fs/efivarfs/super.c err = efivar_entry_add(entry, &efivarfs_list); entry 164 fs/efivarfs/super.c inode->i_private = entry; entry 165 fs/efivarfs/super.c i_size_write(inode, size + sizeof(entry->var.Attributes)); entry 176 fs/efivarfs/super.c kfree(entry); entry 180 fs/efivarfs/super.c static int efivarfs_destroy(struct efivar_entry *entry, void *data) entry 182 fs/efivarfs/super.c int err = efivar_entry_remove(entry); entry 186 fs/efivarfs/super.c kfree(entry); entry 155 fs/erofs/xattr.c int (*entry)(struct xattr_iter *_it, struct erofs_xattr_entry *entry); entry 219 fs/erofs/xattr.c struct erofs_xattr_entry entry; entry 233 fs/erofs/xattr.c entry = *(struct erofs_xattr_entry *)(it->kaddr + it->ofs); entry 235 fs/erofs/xattr.c unsigned int entry_sz = erofs_xattr_entry_size(&entry); entry 246 fs/erofs/xattr.c value_sz = le16_to_cpu(entry.e_value_size); entry 249 fs/erofs/xattr.c err = op->entry(it, &entry); entry 251 fs/erofs/xattr.c it->ofs += entry.e_name_len + value_sz; entry 258 fs/erofs/xattr.c while (processed < entry.e_name_len) { entry 269 fs/erofs/xattr.c entry.e_name_len - processed); entry 274 fs/erofs/xattr.c it->ofs += entry.e_name_len - processed + value_sz; entry 325 fs/erofs/xattr.c struct erofs_xattr_entry *entry) entry 329 fs/erofs/xattr.c return (it->index != entry->e_name_index || entry 330 fs/erofs/xattr.c it->name.len != entry->e_name_len) ? -ENOATTR : 0; entry 361 fs/erofs/xattr.c .entry = xattr_entrymatch, entry 532 fs/erofs/xattr.c struct erofs_xattr_entry *entry) entry 540 fs/erofs/xattr.c erofs_xattr_handler(entry->e_name_index); entry 549 fs/erofs/xattr.c it->buffer_ofs += prefix_len + entry->e_name_len + 1; entry 554 fs/erofs/xattr.c + entry->e_name_len + 1 > it->buffer_size) entry 584 fs/erofs/xattr.c .entry = xattr_entrylist, entry 1299 fs/eventpoll.c list_del_init(&wait->entry); entry 1928 fs/eventpoll.c if (!list_empty_careful(&wait.entry)) { entry 43 fs/ext2/acl.c ext2_acl_entry *entry = entry 47 fs/ext2/acl.c acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); entry 48 fs/ext2/acl.c acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); entry 64 fs/ext2/acl.c le32_to_cpu(entry->e_id)); entry 72 fs/ext2/acl.c le32_to_cpu(entry->e_id)); entry 107 fs/ext2/acl.c ext2_acl_entry *entry = (ext2_acl_entry *)e; entry 108 fs/ext2/acl.c entry->e_tag = cpu_to_le16(acl_e->e_tag); entry 109 fs/ext2/acl.c entry->e_perm = cpu_to_le16(acl_e->e_perm); entry 112 fs/ext2/acl.c entry->e_id = cpu_to_le32( entry 117 fs/ext2/acl.c entry->e_id = cpu_to_le32( entry 72 fs/ext2/xattr.c #define IS_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0) entry 149 fs/ext2/xattr.c ext2_xattr_entry_valid(struct ext2_xattr_entry *entry, entry 155 fs/ext2/xattr.c next = EXT2_XATTR_NEXT(entry); entry 159 fs/ext2/xattr.c if (entry->e_value_block != 0) entry 162 fs/ext2/xattr.c size = le32_to_cpu(entry->e_value_size); entry 164 fs/ext2/xattr.c le16_to_cpu(entry->e_value_offs) + size > end_offs) entry 172 fs/ext2/xattr.c struct ext2_xattr_entry *entry) entry 176 fs/ext2/xattr.c cmp = name_index - entry->e_name_index; entry 178 fs/ext2/xattr.c cmp = name_len - entry->e_name_len; entry 180 fs/ext2/xattr.c cmp = memcmp(name, entry->e_name, name_len); entry 200 fs/ext2/xattr.c struct ext2_xattr_entry *entry; entry 237 fs/ext2/xattr.c entry = FIRST_ENTRY(bh); entry 238 fs/ext2/xattr.c while (!IS_LAST_ENTRY(entry)) { entry 239 fs/ext2/xattr.c if (!ext2_xattr_entry_valid(entry, end, entry 244 fs/ext2/xattr.c entry); entry 250 fs/ext2/xattr.c entry = EXT2_XATTR_NEXT(entry); entry 257 fs/ext2/xattr.c size = le32_to_cpu(entry->e_value_size); entry 265 fs/ext2/xattr.c memcpy(buffer, bh->b_data + le16_to_cpu(entry->e_value_offs), entry 292 fs/ext2/xattr.c struct ext2_xattr_entry *entry; entry 323 fs/ext2/xattr.c entry = FIRST_ENTRY(bh); entry 324 fs/ext2/xattr.c while (!IS_LAST_ENTRY(entry)) { entry 325 fs/ext2/xattr.c if (!ext2_xattr_entry_valid(entry, end, entry 328 fs/ext2/xattr.c entry = EXT2_XATTR_NEXT(entry); entry 334 fs/ext2/xattr.c for (entry = FIRST_ENTRY(bh); !IS_LAST_ENTRY(entry); entry 335 fs/ext2/xattr.c entry = EXT2_XATTR_NEXT(entry)) { entry 337 fs/ext2/xattr.c ext2_xattr_handler(entry->e_name_index); entry 342 fs/ext2/xattr.c size_t size = prefix_len + entry->e_name_len + 1; entry 351 fs/ext2/xattr.c memcpy(buffer, entry->e_name, entry->e_name_len); entry 352 fs/ext2/xattr.c buffer += entry->e_name_len; entry 990 fs/ext2/xattr.c struct ext2_xattr_entry *entry) entry 993 fs/ext2/xattr.c char *name = entry->e_name; entry 996 fs/ext2/xattr.c for (n=0; n < entry->e_name_len; n++) { entry 1002 fs/ext2/xattr.c if (entry->e_value_block == 0 && entry->e_value_size != 0) { entry 1004 fs/ext2/xattr.c le16_to_cpu(entry->e_value_offs)); entry 1005 fs/ext2/xattr.c for (n = (le32_to_cpu(entry->e_value_size) + entry 1012 fs/ext2/xattr.c entry->e_hash = cpu_to_le32(hash); entry 1026 fs/ext2/xattr.c struct ext2_xattr_entry *entry) entry 1031 fs/ext2/xattr.c ext2_xattr_hash_entry(header, entry); entry 51 fs/ext2/xattr.h #define EXT2_XATTR_NEXT(entry) \ entry 53 fs/ext2/xattr.h (char *)(entry) + EXT2_XATTR_LEN((entry)->e_name_len)) ) entry 41 fs/ext4/acl.c ext4_acl_entry *entry = entry 45 fs/ext4/acl.c acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); entry 46 fs/ext4/acl.c acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); entry 63 fs/ext4/acl.c le32_to_cpu(entry->e_id)); entry 71 fs/ext4/acl.c le32_to_cpu(entry->e_id)); entry 106 fs/ext4/acl.c ext4_acl_entry *entry = (ext4_acl_entry *)e; entry 107 fs/ext4/acl.c entry->e_tag = cpu_to_le16(acl_e->e_tag); entry 108 fs/ext4/acl.c entry->e_perm = cpu_to_le16(acl_e->e_perm); entry 111 fs/ext4/acl.c entry->e_id = cpu_to_le32( entry 116 fs/ext4/acl.c entry->e_id = cpu_to_le32( entry 55 fs/ext4/block_validity.c struct ext4_system_zone *entry, *n; entry 57 fs/ext4/block_validity.c rbtree_postorder_for_each_entry_safe(entry, n, entry 59 fs/ext4/block_validity.c kmem_cache_free(ext4_system_zone_cachep, entry); entry 71 fs/ext4/block_validity.c struct ext4_system_zone *new_entry = NULL, *entry; entry 77 fs/ext4/block_validity.c entry = rb_entry(parent, struct ext4_system_zone, node); entry 78 fs/ext4/block_validity.c if (start_blk < entry->start_blk) entry 80 fs/ext4/block_validity.c else if (start_blk >= (entry->start_blk + entry->count)) entry 83 fs/ext4/block_validity.c if (start_blk + count > (entry->start_blk + entry 84 fs/ext4/block_validity.c entry->count)) entry 85 fs/ext4/block_validity.c entry->count = (start_blk + count - entry 86 fs/ext4/block_validity.c entry->start_blk); entry 110 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); entry 111 fs/ext4/block_validity.c if (can_merge(entry, new_entry)) { entry 112 fs/ext4/block_validity.c new_entry->start_blk = entry->start_blk; entry 113 fs/ext4/block_validity.c new_entry->count += entry->count; entry 115 fs/ext4/block_validity.c kmem_cache_free(ext4_system_zone_cachep, entry); entry 122 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); entry 123 fs/ext4/block_validity.c if (can_merge(new_entry, entry)) { entry 124 fs/ext4/block_validity.c new_entry->count += entry->count; entry 126 fs/ext4/block_validity.c kmem_cache_free(ext4_system_zone_cachep, entry); entry 135 fs/ext4/block_validity.c struct ext4_system_zone *entry; entry 141 fs/ext4/block_validity.c entry = rb_entry(node, struct ext4_system_zone, node); entry 143 fs/ext4/block_validity.c entry->start_blk, entry->start_blk + entry->count - 1); entry 160 fs/ext4/block_validity.c struct ext4_system_zone *entry; entry 175 fs/ext4/block_validity.c entry = rb_entry(n, struct ext4_system_zone, node); entry 176 fs/ext4/block_validity.c if (start_blk + count - 1 < entry->start_blk) entry 178 fs/ext4/block_validity.c else if (start_blk >= (entry->start_blk + entry->count)) entry 3218 fs/ext4/ext4.h const struct qstr *entry, bool quick); entry 33 fs/ext4/inline.c struct ext4_xattr_entry *entry; entry 54 fs/ext4/inline.c entry = IFIRST(header); entry 57 fs/ext4/inline.c for (; !IS_LAST_ENTRY(entry); entry = EXT4_XATTR_NEXT(entry)) { entry 58 fs/ext4/inline.c if (!entry->e_value_inum && entry->e_value_size) { entry 59 fs/ext4/inline.c size_t offs = le16_to_cpu(entry->e_value_offs); entry 65 fs/ext4/inline.c ((void *)entry - (void *)IFIRST(header)) - sizeof(__u32); entry 68 fs/ext4/inline.c entry = (struct ext4_xattr_entry *) entry 71 fs/ext4/inline.c free += EXT4_XATTR_SIZE(le32_to_cpu(entry->e_value_size)); entry 168 fs/ext4/inline.c struct ext4_xattr_entry *entry; entry 191 fs/ext4/inline.c entry = (struct ext4_xattr_entry *)((void *)raw_inode + entry 194 fs/ext4/inline.c (unsigned int)le32_to_cpu(entry->e_value_size)); entry 197 fs/ext4/inline.c (void *)IFIRST(header) + le16_to_cpu(entry->e_value_offs), len); entry 213 fs/ext4/inline.c struct ext4_xattr_entry *entry; entry 242 fs/ext4/inline.c entry = (struct ext4_xattr_entry *)((void *)raw_inode + entry 245 fs/ext4/inline.c memcpy((void *)IFIRST(header) + le16_to_cpu(entry->e_value_offs) + pos, entry 1058 fs/ext4/inline.c struct ext4_xattr_entry *entry; entry 1064 fs/ext4/inline.c entry = (struct ext4_xattr_entry *)((void *)ext4_raw_inode(iloc) + entry 1067 fs/ext4/inline.c return (void *)IFIRST(header) + le16_to_cpu(entry->e_value_offs); entry 2799 fs/ext4/mballoc.c struct ext4_free_data *entry) entry 2806 fs/ext4/mballoc.c entry->efd_count, entry->efd_group, entry); entry 2808 fs/ext4/mballoc.c err = ext4_mb_load_buddy(sb, entry->efd_group, &e4b); entry 2813 fs/ext4/mballoc.c EXT4_SB(sb)->s_mb_free_pending -= entry->efd_count; entry 2818 fs/ext4/mballoc.c count += entry->efd_count; entry 2820 fs/ext4/mballoc.c ext4_lock_group(sb, entry->efd_group); entry 2822 fs/ext4/mballoc.c rb_erase(&entry->efd_node, &(db->bb_free_root)); entry 2823 fs/ext4/mballoc.c mb_free_blocks(NULL, &e4b, entry->efd_start_cluster, entry->efd_count); entry 2841 fs/ext4/mballoc.c ext4_unlock_group(sb, entry->efd_group); entry 2842 fs/ext4/mballoc.c kmem_cache_free(ext4_free_data_cachep, entry); entry 2855 fs/ext4/mballoc.c struct ext4_free_data *entry, *tmp; entry 2864 fs/ext4/mballoc.c list_for_each_entry(entry, &sbi->s_freed_data_list, efd_list) { entry 2865 fs/ext4/mballoc.c if (entry->efd_tid != commit_tid) entry 2867 fs/ext4/mballoc.c cut_pos = &entry->efd_list; entry 2875 fs/ext4/mballoc.c list_for_each_entry(entry, &freed_data_list, efd_list) { entry 2876 fs/ext4/mballoc.c err = ext4_issue_discard(sb, entry->efd_group, entry 2877 fs/ext4/mballoc.c entry->efd_start_cluster, entry 2878 fs/ext4/mballoc.c entry->efd_count, entry 2883 fs/ext4/mballoc.c " with %d", entry->efd_group, entry 2884 fs/ext4/mballoc.c entry->efd_start_cluster, entry 2885 fs/ext4/mballoc.c entry->efd_count, err); entry 2896 fs/ext4/mballoc.c list_for_each_entry_safe(entry, tmp, &freed_data_list, efd_list) entry 2897 fs/ext4/mballoc.c ext4_free_data_in_buddy(sb, entry); entry 3518 fs/ext4/mballoc.c struct ext4_free_data *entry; entry 3524 fs/ext4/mballoc.c entry = rb_entry(n, struct ext4_free_data, efd_node); entry 3525 fs/ext4/mballoc.c ext4_set_bits(bitmap, entry->efd_start_cluster, entry->efd_count); entry 4622 fs/ext4/mballoc.c struct ext4_free_data *entry, entry 4626 fs/ext4/mballoc.c if ((entry->efd_tid != new_entry->efd_tid) || entry 4627 fs/ext4/mballoc.c (entry->efd_group != new_entry->efd_group)) entry 4629 fs/ext4/mballoc.c if (entry->efd_start_cluster + entry->efd_count == entry 4631 fs/ext4/mballoc.c new_entry->efd_start_cluster = entry->efd_start_cluster; entry 4632 fs/ext4/mballoc.c new_entry->efd_count += entry->efd_count; entry 4634 fs/ext4/mballoc.c entry->efd_start_cluster) { entry 4635 fs/ext4/mballoc.c new_entry->efd_count += entry->efd_count; entry 4639 fs/ext4/mballoc.c list_del(&entry->efd_list); entry 4641 fs/ext4/mballoc.c rb_erase(&entry->efd_node, entry_rb_root); entry 4642 fs/ext4/mballoc.c kmem_cache_free(ext4_free_data_cachep, entry); entry 4652 fs/ext4/mballoc.c struct ext4_free_data *entry; entry 4677 fs/ext4/mballoc.c entry = rb_entry(parent, struct ext4_free_data, efd_node); entry 4678 fs/ext4/mballoc.c if (cluster < entry->efd_start_cluster) entry 4680 fs/ext4/mballoc.c else if (cluster >= (entry->efd_start_cluster + entry->efd_count)) entry 4697 fs/ext4/mballoc.c entry = rb_entry(node, struct ext4_free_data, efd_node); entry 4698 fs/ext4/mballoc.c ext4_try_merge_freed_extent(sbi, entry, new_entry, entry 4704 fs/ext4/mballoc.c entry = rb_entry(node, struct ext4_free_data, efd_node); entry 4705 fs/ext4/mballoc.c ext4_try_merge_freed_extent(sbi, entry, new_entry, entry 261 fs/ext4/namei.c static inline ext4_lblk_t dx_get_block(struct dx_entry *entry); entry 262 fs/ext4/namei.c static void dx_set_block(struct dx_entry *entry, ext4_lblk_t value); entry 263 fs/ext4/namei.c static inline unsigned dx_get_hash(struct dx_entry *entry); entry 264 fs/ext4/namei.c static void dx_set_hash(struct dx_entry *entry, unsigned value); entry 532 fs/ext4/namei.c static inline ext4_lblk_t dx_get_block(struct dx_entry *entry) entry 534 fs/ext4/namei.c return le32_to_cpu(entry->block) & 0x0fffffff; entry 537 fs/ext4/namei.c static inline void dx_set_block(struct dx_entry *entry, ext4_lblk_t value) entry 539 fs/ext4/namei.c entry->block = cpu_to_le32(value); entry 542 fs/ext4/namei.c static inline unsigned dx_get_hash(struct dx_entry *entry) entry 544 fs/ext4/namei.c return le32_to_cpu(entry->hash); entry 547 fs/ext4/namei.c static inline void dx_set_hash(struct dx_entry *entry, unsigned value) entry 549 fs/ext4/namei.c entry->hash = cpu_to_le32(value); entry 1283 fs/ext4/namei.c const struct qstr *entry, bool quick) entry 1290 fs/ext4/namei.c ret = utf8_strncasecmp_folded(um, name, entry); entry 1292 fs/ext4/namei.c ret = utf8_strncasecmp(um, name, entry); entry 1301 fs/ext4/namei.c if (name->len != entry->len) entry 1304 fs/ext4/namei.c return !!memcmp(name->name, entry->name, name->len); entry 1348 fs/ext4/namei.c const struct qstr entry = {.name = de->name, .len = de->name_len}; entry 1365 fs/ext4/namei.c return !ext4_ci_compare(parent, &cf, &entry, true); entry 1367 fs/ext4/namei.c return !ext4_ci_compare(parent, fname->usr_fname, &entry, entry 183 fs/ext4/xattr.c ext4_xattr_check_entries(struct ext4_xattr_entry *entry, void *end, entry 186 fs/ext4/xattr.c struct ext4_xattr_entry *e = entry; entry 199 fs/ext4/xattr.c while (!IS_LAST_ENTRY(entry)) { entry 200 fs/ext4/xattr.c u32 size = le32_to_cpu(entry->e_value_size); entry 205 fs/ext4/xattr.c if (size != 0 && entry->e_value_inum == 0) { entry 206 fs/ext4/xattr.c u16 offs = le16_to_cpu(entry->e_value_offs); entry 223 fs/ext4/xattr.c entry = EXT4_XATTR_NEXT(entry); entry 284 fs/ext4/xattr.c struct ext4_xattr_entry *entry, *next; entry 291 fs/ext4/xattr.c for (entry = *pentry; !IS_LAST_ENTRY(entry); entry = next) { entry 292 fs/ext4/xattr.c next = EXT4_XATTR_NEXT(entry); entry 297 fs/ext4/xattr.c cmp = name_index - entry->e_name_index; entry 299 fs/ext4/xattr.c cmp = name_len - entry->e_name_len; entry 301 fs/ext4/xattr.c cmp = memcmp(name, entry->e_name, name_len); entry 305 fs/ext4/xattr.c *pentry = entry; entry 439 fs/ext4/xattr.c struct ext4_xattr_entry *entry, void *buffer, entry 449 fs/ext4/xattr.c if (entry) { entry 454 fs/ext4/xattr.c e_hash = ext4_xattr_hash_entry(entry->e_name, entry->e_name_len, entry 456 fs/ext4/xattr.c if (e_hash != entry->e_hash) entry 466 fs/ext4/xattr.c ext4_xattr_inode_get(struct inode *inode, struct ext4_xattr_entry *entry, entry 473 fs/ext4/xattr.c err = ext4_xattr_inode_iget(inode, le32_to_cpu(entry->e_value_inum), entry 474 fs/ext4/xattr.c le32_to_cpu(entry->e_hash), &ea_inode); entry 493 fs/ext4/xattr.c err = ext4_xattr_inode_verify_hashes(ea_inode, entry, buffer, entry 516 fs/ext4/xattr.c struct ext4_xattr_entry *entry; entry 538 fs/ext4/xattr.c entry = BFIRST(bh); entry 540 fs/ext4/xattr.c error = xattr_find_entry(inode, &entry, end, name_index, name, 1); entry 543 fs/ext4/xattr.c size = le32_to_cpu(entry->e_value_size); entry 550 fs/ext4/xattr.c if (entry->e_value_inum) { entry 551 fs/ext4/xattr.c error = ext4_xattr_inode_get(inode, entry, buffer, entry 556 fs/ext4/xattr.c u16 offset = le16_to_cpu(entry->e_value_offs); entry 576 fs/ext4/xattr.c struct ext4_xattr_entry *entry; entry 594 fs/ext4/xattr.c entry = IFIRST(header); entry 595 fs/ext4/xattr.c error = xattr_find_entry(inode, &entry, end, name_index, name, 0); entry 598 fs/ext4/xattr.c size = le32_to_cpu(entry->e_value_size); entry 605 fs/ext4/xattr.c if (entry->e_value_inum) { entry 606 fs/ext4/xattr.c error = ext4_xattr_inode_get(inode, entry, buffer, entry 611 fs/ext4/xattr.c u16 offset = le16_to_cpu(entry->e_value_offs); entry 659 fs/ext4/xattr.c ext4_xattr_list_entries(struct dentry *dentry, struct ext4_xattr_entry *entry, entry 664 fs/ext4/xattr.c for (; !IS_LAST_ENTRY(entry); entry = EXT4_XATTR_NEXT(entry)) { entry 666 fs/ext4/xattr.c ext4_xattr_handler(entry->e_name_index); entry 671 fs/ext4/xattr.c size_t size = prefix_len + entry->e_name_len + 1; entry 678 fs/ext4/xattr.c memcpy(buffer, entry->e_name, entry->e_name_len); entry 679 fs/ext4/xattr.c buffer += entry->e_name_len; entry 804 fs/ext4/xattr.c struct ext4_xattr_entry *entry; entry 822 fs/ext4/xattr.c for (entry = IFIRST(header); !IS_LAST_ENTRY(entry); entry 823 fs/ext4/xattr.c entry = EXT4_XATTR_NEXT(entry)) entry 824 fs/ext4/xattr.c if (entry->e_value_inum) entry 840 fs/ext4/xattr.c for (entry = BFIRST(bh); !IS_LAST_ENTRY(entry); entry 841 fs/ext4/xattr.c entry = EXT4_XATTR_NEXT(entry)) entry 842 fs/ext4/xattr.c if (entry->e_value_inum) entry 960 fs/ext4/xattr.c struct ext4_xattr_entry *entry = BFIRST(block_bh); entry 962 fs/ext4/xattr.c for (; !IS_LAST_ENTRY(entry); entry = EXT4_XATTR_NEXT(entry)) entry 963 fs/ext4/xattr.c if (entry->e_value_inum) entry 1100 fs/ext4/xattr.c struct ext4_xattr_entry *entry; entry 1105 fs/ext4/xattr.c for (entry = first; !IS_LAST_ENTRY(entry); entry 1106 fs/ext4/xattr.c entry = EXT4_XATTR_NEXT(entry)) { entry 1107 fs/ext4/xattr.c if (!entry->e_value_inum) entry 1109 fs/ext4/xattr.c ea_ino = le32_to_cpu(entry->e_value_inum); entry 1111 fs/ext4/xattr.c le32_to_cpu(entry->e_hash), entry 1127 fs/ext4/xattr.c failed_entry = entry; entry 1129 fs/ext4/xattr.c for (entry = first; entry != failed_entry; entry 1130 fs/ext4/xattr.c entry = EXT4_XATTR_NEXT(entry)) { entry 1131 fs/ext4/xattr.c if (!entry->e_value_inum) entry 1133 fs/ext4/xattr.c ea_ino = le32_to_cpu(entry->e_value_inum); entry 1135 fs/ext4/xattr.c le32_to_cpu(entry->e_hash), entry 1160 fs/ext4/xattr.c struct ext4_xattr_entry *entry; entry 1169 fs/ext4/xattr.c for (entry = first; !IS_LAST_ENTRY(entry); entry 1170 fs/ext4/xattr.c entry = EXT4_XATTR_NEXT(entry)) { entry 1171 fs/ext4/xattr.c if (!entry->e_value_inum) entry 1173 fs/ext4/xattr.c ea_ino = le32_to_cpu(entry->e_value_inum); entry 1175 fs/ext4/xattr.c le32_to_cpu(entry->e_hash), entry 1205 fs/ext4/xattr.c le32_to_cpu(entry->e_value_size)); entry 1213 fs/ext4/xattr.c entry->e_value_inum = 0; entry 1214 fs/ext4/xattr.c entry->e_value_size = 0; entry 2521 fs/ext4/xattr.c static void ext4_xattr_shift_entries(struct ext4_xattr_entry *entry, entry 2525 fs/ext4/xattr.c struct ext4_xattr_entry *last = entry; entry 2548 fs/ext4/xattr.c struct ext4_xattr_entry *entry) entry 2553 fs/ext4/xattr.c size_t value_size = le32_to_cpu(entry->e_value_size); entry 2557 fs/ext4/xattr.c .name_index = entry->e_name_index, entry 2558 fs/ext4/xattr.c .in_inode = !!entry->e_value_inum, entry 2566 fs/ext4/xattr.c b_entry_name = kmalloc(entry->e_name_len + 1, GFP_NOFS); entry 2578 fs/ext4/xattr.c if (entry->e_value_inum) { entry 2579 fs/ext4/xattr.c error = ext4_xattr_inode_get(inode, entry, buffer, value_size); entry 2583 fs/ext4/xattr.c size_t value_offs = le16_to_cpu(entry->e_value_offs); entry 2587 fs/ext4/xattr.c memcpy(b_entry_name, entry->e_name, entry->e_name_len); entry 2588 fs/ext4/xattr.c b_entry_name[entry->e_name_len] = '\0'; entry 2635 fs/ext4/xattr.c struct ext4_xattr_entry *entry; entry 2643 fs/ext4/xattr.c entry = NULL; entry 2663 fs/ext4/xattr.c entry = last; entry 2669 fs/ext4/xattr.c if (entry == NULL) { entry 2672 fs/ext4/xattr.c entry = small_entry; entry 2675 fs/ext4/xattr.c entry_size = EXT4_XATTR_LEN(entry->e_name_len); entry 2677 fs/ext4/xattr.c if (!entry->e_value_inum) entry 2679 fs/ext4/xattr.c le32_to_cpu(entry->e_value_size)); entry 2681 fs/ext4/xattr.c entry); entry 2861 fs/ext4/xattr.c struct ext4_xattr_entry *entry; entry 2915 fs/ext4/xattr.c for (entry = BFIRST(bh); !IS_LAST_ENTRY(entry); entry 2916 fs/ext4/xattr.c entry = EXT4_XATTR_NEXT(entry)) { entry 2917 fs/ext4/xattr.c if (!entry->e_value_inum) entry 2920 fs/ext4/xattr.c le32_to_cpu(entry->e_value_inum), entry 2921 fs/ext4/xattr.c le32_to_cpu(entry->e_hash), entry 2926 fs/ext4/xattr.c le32_to_cpu(entry->e_value_size)); entry 60 fs/ext4/xattr.h #define EXT4_XATTR_NEXT(entry) \ entry 62 fs/ext4/xattr.h (char *)(entry) + EXT4_XATTR_LEN((entry)->e_name_len))) entry 94 fs/ext4/xattr.h #define IS_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0) entry 50 fs/f2fs/acl.c struct f2fs_acl_entry *entry = (struct f2fs_acl_entry *)(hdr + 1); entry 71 fs/f2fs/acl.c if ((char *)entry > end) entry 74 fs/f2fs/acl.c acl->a_entries[i].e_tag = le16_to_cpu(entry->e_tag); entry 75 fs/f2fs/acl.c acl->a_entries[i].e_perm = le16_to_cpu(entry->e_perm); entry 82 fs/f2fs/acl.c entry = (struct f2fs_acl_entry *)((char *)entry + entry 89 fs/f2fs/acl.c le32_to_cpu(entry->e_id)); entry 90 fs/f2fs/acl.c entry = (struct f2fs_acl_entry *)((char *)entry + entry 96 fs/f2fs/acl.c le32_to_cpu(entry->e_id)); entry 97 fs/f2fs/acl.c entry = (struct f2fs_acl_entry *)((char *)entry + entry 104 fs/f2fs/acl.c if ((char *)entry != end) entry 116 fs/f2fs/acl.c struct f2fs_acl_entry *entry; entry 126 fs/f2fs/acl.c entry = (struct f2fs_acl_entry *)(f2fs_acl + 1); entry 130 fs/f2fs/acl.c entry->e_tag = cpu_to_le16(acl->a_entries[i].e_tag); entry 131 fs/f2fs/acl.c entry->e_perm = cpu_to_le16(acl->a_entries[i].e_perm); entry 135 fs/f2fs/acl.c entry->e_id = cpu_to_le32( entry 138 fs/f2fs/acl.c entry = (struct f2fs_acl_entry *)((char *)entry + entry 142 fs/f2fs/acl.c entry->e_id = cpu_to_le32( entry 145 fs/f2fs/acl.c entry = (struct f2fs_acl_entry *)((char *)entry + entry 152 fs/f2fs/acl.c entry = (struct f2fs_acl_entry *)((char *)entry + entry 115 fs/f2fs/dir.c const struct qstr *entry, bool quick) entry 122 fs/f2fs/dir.c ret = utf8_strncasecmp_folded(um, name, entry); entry 124 fs/f2fs/dir.c ret = utf8_strncasecmp(um, name, entry); entry 133 fs/f2fs/dir.c if (name->len != entry->len) entry 136 fs/f2fs/dir.c return !!memcmp(name->name, entry->name, name->len); entry 177 fs/f2fs/dir.c struct qstr entry; entry 184 fs/f2fs/dir.c entry.name = d->filename[bit_pos]; entry 185 fs/f2fs/dir.c entry.len = de->name_len; entry 191 fs/f2fs/dir.c return !f2fs_ci_compare(parent, &cf, &entry, true); entry 193 fs/f2fs/dir.c return !f2fs_ci_compare(parent, fname->usr_fname, &entry, entry 2207 fs/f2fs/f2fs.h void *entry; entry 2209 fs/f2fs/f2fs.h entry = kmem_cache_alloc(cachep, flags); entry 2210 fs/f2fs/f2fs.h if (!entry) entry 2211 fs/f2fs/f2fs.h entry = kmem_cache_alloc(cachep, flags | __GFP_NOFAIL); entry 2212 fs/f2fs/f2fs.h return entry; entry 2964 fs/f2fs/f2fs.h const struct qstr *entry, entry 509 fs/f2fs/gc.c struct f2fs_summary *entry; entry 519 fs/f2fs/gc.c entry = sum; entry 524 fs/f2fs/gc.c for (off = 0; off < sbi->blocks_per_seg; off++, entry++) { entry 525 fs/f2fs/gc.c nid_t nid = le32_to_cpu(entry->nid); entry 996 fs/f2fs/gc.c struct f2fs_summary *entry; entry 1005 fs/f2fs/gc.c entry = sum; entry 1007 fs/f2fs/gc.c for (off = 0; off < sbi->blocks_per_seg; off++, entry++) { entry 1013 fs/f2fs/gc.c nid_t nid = le32_to_cpu(entry->nid); entry 1034 fs/f2fs/gc.c if (!is_alive(sbi, entry, &dni, start_addr + off, &nofs)) entry 1042 fs/f2fs/gc.c ofs_in_node = le16_to_cpu(entry->ofs_in_node); entry 59 fs/f2fs/recovery.c struct fsync_inode_entry *entry; entry 61 fs/f2fs/recovery.c list_for_each_entry(entry, head, list) entry 62 fs/f2fs/recovery.c if (entry->inode->i_ino == ino) entry 63 fs/f2fs/recovery.c return entry; entry 72 fs/f2fs/recovery.c struct fsync_inode_entry *entry; entry 89 fs/f2fs/recovery.c entry = f2fs_kmem_cache_alloc(fsync_entry_slab, GFP_F2FS_ZERO); entry 90 fs/f2fs/recovery.c entry->inode = inode; entry 91 fs/f2fs/recovery.c list_add_tail(&entry->list, head); entry 93 fs/f2fs/recovery.c return entry; entry 99 fs/f2fs/recovery.c static void del_fsync_inode(struct fsync_inode_entry *entry, int drop) entry 103 fs/f2fs/recovery.c f2fs_inode_synced(entry->inode); entry 105 fs/f2fs/recovery.c iput(entry->inode); entry 106 fs/f2fs/recovery.c list_del(&entry->list); entry 107 fs/f2fs/recovery.c kmem_cache_free(fsync_entry_slab, entry); entry 119 fs/f2fs/recovery.c struct fsync_inode_entry *entry; entry 123 fs/f2fs/recovery.c entry = get_fsync_inode(dir_list, pino); entry 124 fs/f2fs/recovery.c if (!entry) { entry 125 fs/f2fs/recovery.c entry = add_fsync_inode(F2FS_I_SB(inode), dir_list, entry 127 fs/f2fs/recovery.c if (IS_ERR(entry)) { entry 128 fs/f2fs/recovery.c dir = ERR_CAST(entry); entry 129 fs/f2fs/recovery.c err = PTR_ERR(entry); entry 134 fs/f2fs/recovery.c dir = entry->inode; entry 315 fs/f2fs/recovery.c struct fsync_inode_entry *entry; entry 334 fs/f2fs/recovery.c entry = get_fsync_inode(head, ino_of_node(page)); entry 335 fs/f2fs/recovery.c if (!entry) { entry 352 fs/f2fs/recovery.c entry = add_fsync_inode(sbi, head, ino_of_node(page), entry 354 fs/f2fs/recovery.c if (IS_ERR(entry)) { entry 355 fs/f2fs/recovery.c err = PTR_ERR(entry); entry 364 fs/f2fs/recovery.c entry->blkaddr = blkaddr; entry 367 fs/f2fs/recovery.c entry->last_dentry = blkaddr; entry 391 fs/f2fs/recovery.c struct fsync_inode_entry *entry, *tmp; entry 393 fs/f2fs/recovery.c list_for_each_entry_safe(entry, tmp, head, list) entry 394 fs/f2fs/recovery.c del_fsync_inode(entry, drop); entry 660 fs/f2fs/recovery.c struct fsync_inode_entry *entry; entry 678 fs/f2fs/recovery.c entry = get_fsync_inode(inode_list, ino_of_node(page)); entry 679 fs/f2fs/recovery.c if (!entry) entry 687 fs/f2fs/recovery.c err = recover_inode(entry->inode, page); entry 693 fs/f2fs/recovery.c if (entry->last_dentry == blkaddr) { entry 694 fs/f2fs/recovery.c err = recover_dentry(entry->inode, page, dir_list); entry 700 fs/f2fs/recovery.c err = do_recover_data(sbi, entry->inode, page); entry 706 fs/f2fs/recovery.c if (entry->blkaddr == blkaddr) entry 707 fs/f2fs/recovery.c list_move_tail(&entry->list, tmp_inode_list); entry 1901 fs/f2fs/segment.c static void release_discard_addr(struct discard_entry *entry) entry 1903 fs/f2fs/segment.c list_del(&entry->list); entry 1904 fs/f2fs/segment.c kmem_cache_free(discard_entry_slab, entry); entry 1910 fs/f2fs/segment.c struct discard_entry *entry, *this; entry 1913 fs/f2fs/segment.c list_for_each_entry_safe(entry, this, head, list) entry 1914 fs/f2fs/segment.c release_discard_addr(entry); entry 1936 fs/f2fs/segment.c struct discard_entry *entry, *this; entry 1996 fs/f2fs/segment.c list_for_each_entry_safe(entry, this, head, list) { entry 1998 fs/f2fs/segment.c bool is_valid = test_bit_le(0, entry->discard_map); entry 2002 fs/f2fs/segment.c next_pos = find_next_zero_bit_le(entry->discard_map, entry 2010 fs/f2fs/segment.c f2fs_issue_discard(sbi, entry->start_blkaddr + cur_pos, entry 2014 fs/f2fs/segment.c next_pos = find_next_bit_le(entry->discard_map, entry 2024 fs/f2fs/segment.c release_discard_addr(entry); entry 210 fs/f2fs/xattr.c struct f2fs_xattr_entry *entry; entry 212 fs/f2fs/xattr.c list_for_each_xattr(entry, base_addr) { entry 213 fs/f2fs/xattr.c if ((void *)(entry) + sizeof(__u32) > last_base_addr || entry 214 fs/f2fs/xattr.c (void *)XATTR_NEXT_ENTRY(entry) > last_base_addr) entry 217 fs/f2fs/xattr.c if (entry->e_name_index != index) entry 219 fs/f2fs/xattr.c if (entry->e_name_len != len) entry 221 fs/f2fs/xattr.c if (!memcmp(entry->e_name, name, len)) entry 224 fs/f2fs/xattr.c return entry; entry 231 fs/f2fs/xattr.c struct f2fs_xattr_entry *entry; entry 235 fs/f2fs/xattr.c list_for_each_xattr(entry, base_addr) { entry 236 fs/f2fs/xattr.c if ((void *)entry + sizeof(__u32) > max_addr || entry 237 fs/f2fs/xattr.c (void *)XATTR_NEXT_ENTRY(entry) > max_addr) { entry 238 fs/f2fs/xattr.c *last_addr = entry; entry 241 fs/f2fs/xattr.c if (entry->e_name_index != index) entry 243 fs/f2fs/xattr.c if (entry->e_name_len != len) entry 245 fs/f2fs/xattr.c if (!memcmp(entry->e_name, name, len)) entry 250 fs/f2fs/xattr.c if (IS_XATTR_LAST_ENTRY(entry) && entry 251 fs/f2fs/xattr.c (void *)entry + sizeof(__u32) > max_addr) { entry 252 fs/f2fs/xattr.c *last_addr = entry; entry 255 fs/f2fs/xattr.c return entry; entry 497 fs/f2fs/xattr.c struct f2fs_xattr_entry *entry = NULL; entry 512 fs/f2fs/xattr.c &entry, &base_addr, &base_size); entry 517 fs/f2fs/xattr.c size = le16_to_cpu(entry->e_value_size); entry 525 fs/f2fs/xattr.c char *pval = entry->e_name + entry->e_name_len; entry 543 fs/f2fs/xattr.c struct f2fs_xattr_entry *entry; entry 556 fs/f2fs/xattr.c list_for_each_xattr(entry, base_addr) { entry 558 fs/f2fs/xattr.c f2fs_xattr_handler(entry->e_name_index); entry 563 fs/f2fs/xattr.c if ((void *)(entry) + sizeof(__u32) > last_base_addr || entry 564 fs/f2fs/xattr.c (void *)XATTR_NEXT_ENTRY(entry) > last_base_addr) { entry 577 fs/f2fs/xattr.c size = prefix_len + entry->e_name_len + 1; entry 585 fs/f2fs/xattr.c memcpy(buffer, entry->e_name, entry->e_name_len); entry 586 fs/f2fs/xattr.c buffer += entry->e_name_len; entry 597 fs/f2fs/xattr.c static bool f2fs_xattr_value_same(struct f2fs_xattr_entry *entry, entry 600 fs/f2fs/xattr.c void *pval = entry->e_name + entry->e_name_len; entry 602 fs/f2fs/xattr.c return (le16_to_cpu(entry->e_value_size) == size) && entry 62 fs/f2fs/xattr.h #define ENTRY_SIZE(entry) (XATTR_ALIGN(sizeof(struct f2fs_xattr_entry) + \ entry 63 fs/f2fs/xattr.h (entry)->e_name_len + le16_to_cpu((entry)->e_value_size))) entry 65 fs/f2fs/xattr.h #define XATTR_NEXT_ENTRY(entry) ((struct f2fs_xattr_entry *)((char *)(entry) +\ entry 66 fs/f2fs/xattr.h ENTRY_SIZE(entry))) entry 68 fs/f2fs/xattr.h #define IS_XATTR_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0) entry 70 fs/f2fs/xattr.h #define list_for_each_xattr(entry, addr) \ entry 71 fs/f2fs/xattr.h for (entry = XATTR_FIRST_ENTRY(addr);\ entry 72 fs/f2fs/xattr.h !IS_XATTR_LAST_ENTRY(entry);\ entry 73 fs/f2fs/xattr.h entry = XATTR_NEXT_ENTRY(entry)) entry 342 fs/fat/fat.h int entry; entry 356 fs/fat/fat.h fatent->entry = 0; entry 362 fs/fat/fat.h static inline void fatent_set_entry(struct fat_entry *fatent, int entry) entry 364 fs/fat/fat.h fatent->entry = entry; entry 379 fs/fat/fat.h static inline bool fat_valid_entry(struct msdos_sb_info *sbi, int entry) entry 381 fs/fat/fat.h return FAT_START_ENT <= entry && entry < sbi->max_cluster; entry 386 fs/fat/fat.h int entry); entry 22 fs/fat/fatent.c static void fat12_ent_blocknr(struct super_block *sb, int entry, entry 26 fs/fat/fatent.c int bytes = entry + (entry >> 1); entry 27 fs/fat/fatent.c WARN_ON(!fat_valid_entry(sbi, entry)); entry 32 fs/fat/fatent.c static void fat_ent_blocknr(struct super_block *sb, int entry, entry 36 fs/fat/fatent.c int bytes = (entry << sbi->fatent_shift); entry 37 fs/fat/fatent.c WARN_ON(!fat_valid_entry(sbi, entry)); entry 124 fs/fat/fatent.c if (fatent->entry & 1) entry 162 fs/fat/fatent.c if (fatent->entry & 1) { entry 197 fs/fat/fatent.c u8 *nextp = ent12_p[1] + 1 + (fatent->entry & 1); entry 199 fs/fat/fatent.c fatent->entry++; entry 229 fs/fat/fatent.c fatent->entry++; entry 241 fs/fat/fatent.c fatent->entry++; entry 347 fs/fat/fatent.c int fat_ent_read(struct inode *inode, struct fat_entry *fatent, int entry) entry 355 fs/fat/fatent.c if (!fat_valid_entry(sbi, entry)) { entry 357 fs/fat/fatent.c fat_fs_error(sb, "invalid access to FAT (entry 0x%08x)", entry); entry 361 fs/fat/fatent.c fatent_set_entry(fatent, entry); entry 362 fs/fat/fatent.c ops->ent_blocknr(sb, entry, &offset, &blocknr); entry 428 fs/fat/fatent.c if (fatent->entry < sbi->max_cluster) entry 442 fs/fat/fatent.c ops->ent_blocknr(sb, fatent->entry, &offset, &blocknr); entry 488 fs/fat/fatent.c if (fatent.entry >= sbi->max_cluster) entry 489 fs/fat/fatent.c fatent.entry = FAT_START_ENT; entry 490 fs/fat/fatent.c fatent_set_entry(&fatent, fatent.entry); entry 498 fs/fat/fatent.c int entry = fatent.entry; entry 503 fs/fat/fatent.c ops->ent_put(&prev_ent, entry); entry 507 fs/fat/fatent.c sbi->prev_free = entry; entry 511 fs/fat/fatent.c cluster[idx_clus] = entry; entry 583 fs/fat/fatent.c if (cluster != fatent.entry + 1) { entry 584 fs/fat/fatent.c int nr_clus = fatent.entry - first_cl + 1; entry 645 fs/fat/fatent.c ops->ent_blocknr(sb, fatent->entry, &offset, &blocknr); entry 670 fs/fat/fatent.c while (fatent.entry < sbi->max_cluster) { entry 736 fs/fat/fatent.c while (fatent.entry <= ent_end) { entry 752 fs/fat/fatent.c u32 clus = fatent.entry - free; entry 763 fs/fat/fatent.c } while (fat_ent_next(sbi, &fatent) && fatent.entry <= ent_end); entry 779 fs/fat/fatent.c u32 clus = fatent.entry - free; entry 59 fs/fs_pin.c if (likely(list_empty(&wait.entry))) entry 334 fs/fuse/control.c list_for_each_entry(fc, &fuse_conn_list, entry) { entry 368 fs/fuse/control.c list_for_each_entry(fc, &fuse_conn_list, entry) entry 2105 fs/fuse/dev.c list_for_each_entry(fud, &fc->devices, entry) { entry 28 fs/fuse/dir.c static inline void __fuse_dentry_settime(struct dentry *entry, u64 time) entry 30 fs/fuse/dir.c entry->d_fsdata = (void *) time; entry 33 fs/fuse/dir.c static inline u64 fuse_dentry_time(const struct dentry *entry) entry 35 fs/fuse/dir.c return (u64)entry->d_fsdata; entry 49 fs/fuse/dir.c static inline u64 fuse_dentry_time(const struct dentry *entry) entry 51 fs/fuse/dir.c return ((union fuse_dentry *) entry->d_fsdata)->time; entry 102 fs/fuse/dir.c void fuse_change_entry_timeout(struct dentry *entry, struct fuse_entry_out *o) entry 104 fs/fuse/dir.c fuse_dentry_settime(entry, entry 156 fs/fuse/dir.c void fuse_invalidate_entry_cache(struct dentry *entry) entry 158 fs/fuse/dir.c fuse_dentry_settime(entry, 0); entry 165 fs/fuse/dir.c static void fuse_invalidate_entry(struct dentry *entry) entry 167 fs/fuse/dir.c d_invalidate(entry); entry 168 fs/fuse/dir.c fuse_invalidate_entry_cache(entry); entry 195 fs/fuse/dir.c static int fuse_dentry_revalidate(struct dentry *entry, unsigned int flags) entry 203 fs/fuse/dir.c inode = d_inode_rcu(entry); entry 206 fs/fuse/dir.c else if (time_before64(fuse_dentry_time(entry), get_jiffies_64()) || entry 230 fs/fuse/dir.c parent = dget_parent(entry); entry 232 fs/fuse/dir.c &entry->d_name, &outarg); entry 259 fs/fuse/dir.c fuse_change_entry_timeout(entry, &outarg); entry 266 fs/fuse/dir.c parent = dget_parent(entry); entry 379 fs/fuse/dir.c static struct dentry *fuse_lookup(struct inode *dir, struct dentry *entry, entry 390 fs/fuse/dir.c err = fuse_lookup_name(dir->i_sb, get_node_id(dir), &entry->d_name, entry 404 fs/fuse/dir.c newent = d_splice_alias(inode, entry); entry 409 fs/fuse/dir.c entry = newent ? newent : entry; entry 411 fs/fuse/dir.c fuse_change_entry_timeout(entry, &outarg); entry 413 fs/fuse/dir.c fuse_invalidate_entry_cache(entry); entry 431 fs/fuse/dir.c static int fuse_create_open(struct inode *dir, struct dentry *entry, entry 473 fs/fuse/dir.c args.in_args[1].size = entry->d_name.len + 1; entry 474 fs/fuse/dir.c args.in_args[1].value = entry->d_name.name; entry 502 fs/fuse/dir.c d_instantiate(entry, inode); entry 503 fs/fuse/dir.c fuse_change_entry_timeout(entry, &outentry); entry 505 fs/fuse/dir.c err = finish_open(file, entry, generic_file_open); entry 524 fs/fuse/dir.c static int fuse_atomic_open(struct inode *dir, struct dentry *entry, entry 532 fs/fuse/dir.c if (d_in_lookup(entry)) { entry 533 fs/fuse/dir.c res = fuse_lookup(dir, entry, 0); entry 538 fs/fuse/dir.c entry = res; entry 541 fs/fuse/dir.c if (!(flags & O_CREAT) || d_really_is_positive(entry)) entry 550 fs/fuse/dir.c err = fuse_create_open(dir, entry, file, flags, mode); entry 560 fs/fuse/dir.c err = fuse_mknod(dir, entry, mode, 0); entry 571 fs/fuse/dir.c struct inode *dir, struct dentry *entry, entry 608 fs/fuse/dir.c d_drop(entry); entry 609 fs/fuse/dir.c d = d_splice_alias(inode, entry); entry 617 fs/fuse/dir.c fuse_change_entry_timeout(entry, &outarg); entry 627 fs/fuse/dir.c static int fuse_mknod(struct inode *dir, struct dentry *entry, umode_t mode, entry 645 fs/fuse/dir.c args.in_args[1].size = entry->d_name.len + 1; entry 646 fs/fuse/dir.c args.in_args[1].value = entry->d_name.name; entry 647 fs/fuse/dir.c return create_new_entry(fc, &args, dir, entry, mode); entry 650 fs/fuse/dir.c static int fuse_create(struct inode *dir, struct dentry *entry, umode_t mode, entry 653 fs/fuse/dir.c return fuse_mknod(dir, entry, mode, 0); entry 656 fs/fuse/dir.c static int fuse_mkdir(struct inode *dir, struct dentry *entry, umode_t mode) entry 672 fs/fuse/dir.c args.in_args[1].size = entry->d_name.len + 1; entry 673 fs/fuse/dir.c args.in_args[1].value = entry->d_name.name; entry 674 fs/fuse/dir.c return create_new_entry(fc, &args, dir, entry, S_IFDIR); entry 677 fs/fuse/dir.c static int fuse_symlink(struct inode *dir, struct dentry *entry, entry 686 fs/fuse/dir.c args.in_args[0].size = entry->d_name.len + 1; entry 687 fs/fuse/dir.c args.in_args[0].value = entry->d_name.name; entry 690 fs/fuse/dir.c return create_new_entry(fc, &args, dir, entry, S_IFLNK); entry 701 fs/fuse/dir.c static int fuse_unlink(struct inode *dir, struct dentry *entry) entry 710 fs/fuse/dir.c args.in_args[0].size = entry->d_name.len + 1; entry 711 fs/fuse/dir.c args.in_args[0].value = entry->d_name.name; entry 714 fs/fuse/dir.c struct inode *inode = d_inode(entry); entry 730 fs/fuse/dir.c fuse_invalidate_entry_cache(entry); entry 733 fs/fuse/dir.c fuse_invalidate_entry(entry); entry 737 fs/fuse/dir.c static int fuse_rmdir(struct inode *dir, struct dentry *entry) entry 746 fs/fuse/dir.c args.in_args[0].size = entry->d_name.len + 1; entry 747 fs/fuse/dir.c args.in_args[0].value = entry->d_name.name; entry 750 fs/fuse/dir.c clear_nlink(d_inode(entry)); entry 752 fs/fuse/dir.c fuse_invalidate_entry_cache(entry); entry 754 fs/fuse/dir.c fuse_invalidate_entry(entry); entry 844 fs/fuse/dir.c static int fuse_link(struct dentry *entry, struct inode *newdir, entry 849 fs/fuse/dir.c struct inode *inode = d_inode(entry); entry 1011 fs/fuse/dir.c struct dentry *entry; entry 1027 fs/fuse/dir.c entry = d_lookup(dir, name); entry 1029 fs/fuse/dir.c if (!entry) entry 1033 fs/fuse/dir.c fuse_invalidate_entry(entry); entry 1035 fs/fuse/dir.c if (child_nodeid != 0 && d_really_is_positive(entry)) { entry 1036 fs/fuse/dir.c inode_lock(d_inode(entry)); entry 1037 fs/fuse/dir.c if (get_node_id(d_inode(entry)) != child_nodeid) { entry 1041 fs/fuse/dir.c if (d_mountpoint(entry)) { entry 1045 fs/fuse/dir.c if (d_is_dir(entry)) { entry 1046 fs/fuse/dir.c shrink_dcache_parent(entry); entry 1047 fs/fuse/dir.c if (!simple_empty(entry)) { entry 1051 fs/fuse/dir.c d_inode(entry)->i_flags |= S_DEAD; entry 1053 fs/fuse/dir.c dont_mount(entry); entry 1054 fs/fuse/dir.c clear_nlink(d_inode(entry)); entry 1057 fs/fuse/dir.c inode_unlock(d_inode(entry)); entry 1059 fs/fuse/dir.c d_delete(entry); entry 1063 fs/fuse/dir.c dput(entry); entry 1627 fs/fuse/dir.c static int fuse_setattr(struct dentry *entry, struct iattr *attr) entry 1629 fs/fuse/dir.c struct inode *inode = d_inode(entry); entry 1670 fs/fuse/dir.c ret = fuse_do_setattr(entry, attr, file); entry 1680 fs/fuse/dir.c if (d_is_dir(entry) && (attr->ia_valid & ATTR_MODE)) entry 1681 fs/fuse/dir.c fuse_invalidate_entry_cache(entry); entry 466 fs/fuse/fuse_i.h struct list_head entry; entry 729 fs/fuse/fuse_i.h struct list_head entry; entry 936 fs/fuse/fuse_i.h void fuse_invalidate_entry_cache(struct dentry *entry); entry 941 fs/fuse/fuse_i.h void fuse_change_entry_timeout(struct dentry *entry, struct fuse_entry_out *o); entry 1070 fs/fuse/fuse_i.h ssize_t fuse_listxattr(struct dentry *entry, char *list, size_t size); entry 390 fs/fuse/inode.c list_del(&fc->entry); entry 615 fs/fuse/inode.c INIT_LIST_HEAD(&fc->entry); entry 675 fs/fuse/inode.c struct dentry *entry; entry 705 fs/fuse/inode.c entry = d_obtain_alias(inode); entry 706 fs/fuse/inode.c if (!IS_ERR(entry) && get_node_id(inode) != FUSE_ROOT_ID) entry 707 fs/fuse/inode.c fuse_invalidate_entry_cache(entry); entry 709 fs/fuse/inode.c return entry; entry 1085 fs/fuse/inode.c list_add_tail(&fud->entry, &fc->devices); entry 1109 fs/fuse/inode.c list_del(&fud->entry); entry 1206 fs/fuse/inode.c list_add_tail(&fc->entry, &fuse_conn_list); entry 107 fs/fuse/xattr.c ssize_t fuse_listxattr(struct dentry *entry, char *list, size_t size) entry 109 fs/fuse/xattr.c struct inode *inode = d_inode(entry); entry 65 fs/hfs/brec.c int hfs_brec_insert(struct hfs_find_data *fd, void *entry, int entry_len) entry 133 fs/hfs/brec.c hfs_bnode_write(node, entry, data_off + key_len, entry_len); entry 155 fs/hfs/brec.c entry = &cnid; entry 86 fs/hfs/catalog.c union hfs_cat_rec entry; entry 109 fs/hfs/catalog.c entry_size = hfs_cat_build_thread(sb, &entry, S_ISDIR(inode->i_mode) ? entry 118 fs/hfs/catalog.c err = hfs_brec_insert(&fd, &entry, entry_size); entry 123 fs/hfs/catalog.c entry_size = hfs_cat_build_record(&entry, cnid, inode); entry 131 fs/hfs/catalog.c err = hfs_brec_insert(&fd, &entry, entry_size); entry 293 fs/hfs/catalog.c union hfs_cat_rec entry; entry 319 fs/hfs/catalog.c if (src_fd.entrylength > sizeof(entry) || src_fd.entrylength < 0) { entry 324 fs/hfs/catalog.c hfs_bnode_read(src_fd.bnode, &entry, src_fd.entryoffset, entry 336 fs/hfs/catalog.c err = hfs_brec_insert(&dst_fd, &entry, src_fd.entrylength); entry 355 fs/hfs/catalog.c type = entry.type; entry 356 fs/hfs/catalog.c if (type == HFS_CDR_FIL && !(entry.file.Flags & HFS_FIL_THD)) entry 370 fs/hfs/catalog.c entry_size = hfs_cat_build_thread(sb, &entry, type == HFS_CDR_FIL ? HFS_CDR_FTH : HFS_CDR_THD, entry 378 fs/hfs/catalog.c err = hfs_brec_insert(&dst_fd, &entry, entry_size); entry 54 fs/hfs/dir.c union hfs_cat_rec entry; entry 77 fs/hfs/dir.c if (fd.entrylength > sizeof(entry) || fd.entrylength < 0) { entry 82 fs/hfs/dir.c hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, fd.entrylength); entry 83 fs/hfs/dir.c if (entry.type != HFS_CDR_THD) { entry 94 fs/hfs/dir.c be32_to_cpu(entry.thread.ParID), DT_DIR)) entry 111 fs/hfs/dir.c if (fd.entrylength > sizeof(entry) || fd.entrylength < 0) { entry 116 fs/hfs/dir.c hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, fd.entrylength); entry 117 fs/hfs/dir.c type = entry.type; entry 126 fs/hfs/dir.c be32_to_cpu(entry.dir.DirID), DT_DIR)) entry 135 fs/hfs/dir.c be32_to_cpu(entry.file.FlNum), DT_REG)) entry 88 fs/hfsplus/attributes.c void hfsplus_destroy_attr_entry(hfsplus_attr_entry *entry) entry 90 fs/hfsplus/attributes.c if (entry) entry 91 fs/hfsplus/attributes.c kmem_cache_free(hfsplus_attr_tree_cachep, entry); entry 96 fs/hfsplus/attributes.c static int hfsplus_attr_build_record(hfsplus_attr_entry *entry, int record_type, entry 104 fs/hfsplus/attributes.c memset(entry, 0, sizeof(*entry)); entry 111 fs/hfsplus/attributes.c memset(entry, 0, sizeof(*entry)); entry 116 fs/hfsplus/attributes.c memset(entry, 0, sizeof(struct hfsplus_attr_inline_data)); entry 117 fs/hfsplus/attributes.c entry->inline_data.record_type = cpu_to_be32(record_type); entry 122 fs/hfsplus/attributes.c entry->inline_data.length = cpu_to_be16(len); entry 123 fs/hfsplus/attributes.c memcpy(entry->inline_data.raw_bytes, value, len); entry 132 fs/hfsplus/attributes.c memset(entry, 0, sizeof(*entry)); entry 63 fs/hfsplus/brec.c int hfs_brec_insert(struct hfs_find_data *fd, void *entry, int entry_len) entry 133 fs/hfsplus/brec.c hfs_bnode_write(node, entry, data_off + key_len, entry_len); entry 155 fs/hfsplus/brec.c entry = &cnid; entry 105 fs/hfsplus/catalog.c static int hfsplus_cat_build_record(hfsplus_cat_entry *entry, entry 113 fs/hfsplus/catalog.c folder = &entry->folder; entry 132 fs/hfsplus/catalog.c file = &entry->file; entry 177 fs/hfsplus/catalog.c hfsplus_cat_entry *entry, int type, entry 182 fs/hfsplus/catalog.c entry->type = cpu_to_be16(type); entry 183 fs/hfsplus/catalog.c entry->thread.reserved = 0; entry 184 fs/hfsplus/catalog.c entry->thread.parentID = cpu_to_be32(parentid); entry 185 fs/hfsplus/catalog.c err = hfsplus_asc2uni(sb, &entry->thread.nodeName, HFSPLUS_MAX_STRLEN, entry 190 fs/hfsplus/catalog.c return 10 + be16_to_cpu(entry->thread.nodeName.length) * 2; entry 258 fs/hfsplus/catalog.c hfsplus_cat_entry entry; entry 277 fs/hfsplus/catalog.c entry_size = hfsplus_fill_cat_thread(sb, &entry, entry 292 fs/hfsplus/catalog.c err = hfs_brec_insert(&fd, &entry, entry_size); entry 300 fs/hfsplus/catalog.c entry_size = hfsplus_cat_build_record(&entry, cnid, inode); entry 308 fs/hfsplus/catalog.c err = hfs_brec_insert(&fd, &entry, entry_size); entry 440 fs/hfsplus/catalog.c hfsplus_cat_entry entry; entry 469 fs/hfsplus/catalog.c if (src_fd.entrylength > sizeof(entry) || src_fd.entrylength < 0) { entry 474 fs/hfsplus/catalog.c hfs_bnode_read(src_fd.bnode, &entry, src_fd.entryoffset, entry 476 fs/hfsplus/catalog.c type = be16_to_cpu(entry.type); entry 491 fs/hfsplus/catalog.c err = hfs_brec_insert(&dst_fd, &entry, src_fd.entrylength); entry 528 fs/hfsplus/catalog.c entry_size = hfsplus_fill_cat_thread(sb, &entry, type, entry 541 fs/hfsplus/catalog.c err = hfs_brec_insert(&dst_fd, &entry, entry_size); entry 36 fs/hfsplus/dir.c hfsplus_cat_entry entry; entry 52 fs/hfsplus/dir.c err = hfs_brec_read(&fd, &entry, sizeof(entry)); entry 62 fs/hfsplus/dir.c type = be16_to_cpu(entry.type); entry 68 fs/hfsplus/dir.c cnid = be32_to_cpu(entry.folder.id); entry 75 fs/hfsplus/dir.c cnid = be32_to_cpu(entry.file.id); entry 76 fs/hfsplus/dir.c if (entry.file.user_info.fdType == entry 78 fs/hfsplus/dir.c entry.file.user_info.fdCreator == entry 81 fs/hfsplus/dir.c (entry.file.create_date == entry 84 fs/hfsplus/dir.c entry.file.create_date == entry 100 fs/hfsplus/dir.c be32_to_cpu(entry.file.permissions.dev); entry 136 fs/hfsplus/dir.c hfsplus_cat_entry entry; entry 164 fs/hfsplus/dir.c if (fd.entrylength > sizeof(entry) || fd.entrylength < 0) { entry 169 fs/hfsplus/dir.c hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, entry 171 fs/hfsplus/dir.c if (be16_to_cpu(entry.type) != HFSPLUS_FOLDER_THREAD) { entry 182 fs/hfsplus/dir.c be32_to_cpu(entry.thread.parentID), DT_DIR)) entry 198 fs/hfsplus/dir.c if (fd.entrylength > sizeof(entry) || fd.entrylength < 0) { entry 203 fs/hfsplus/dir.c hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, entry 205 fs/hfsplus/dir.c type = be16_to_cpu(entry.type); entry 219 fs/hfsplus/dir.c be32_to_cpu(entry.folder.id)) entry 222 fs/hfsplus/dir.c be32_to_cpu(entry.folder.id), DT_DIR)) entry 234 fs/hfsplus/dir.c mode = be16_to_cpu(entry.file.permissions.mode); entry 249 fs/hfsplus/dir.c be32_to_cpu(entry.file.id), type)) entry 379 fs/hfsplus/hfsplus_fs.h void hfsplus_destroy_attr_entry(hfsplus_attr_entry *entry); entry 428 fs/hfsplus/hfsplus_fs.h int hfs_brec_insert(struct hfs_find_data *fd, void *entry, int entry_len); entry 490 fs/hfsplus/inode.c hfsplus_cat_entry entry; entry 498 fs/hfsplus/inode.c struct hfsplus_cat_folder *folder = &entry.folder; entry 502 fs/hfsplus/inode.c hfs_bnode_read(fd->bnode, &entry, fd->entryoffset, entry 519 fs/hfsplus/inode.c struct hfsplus_cat_file *file = &entry.file; entry 523 fs/hfsplus/inode.c hfs_bnode_read(fd->bnode, &entry, fd->entryoffset, entry 560 fs/hfsplus/inode.c hfsplus_cat_entry entry; entry 577 fs/hfsplus/inode.c struct hfsplus_cat_folder *folder = &entry.folder; entry 581 fs/hfsplus/inode.c hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, entry 593 fs/hfsplus/inode.c hfs_bnode_write(fd.bnode, &entry, fd.entryoffset, entry 596 fs/hfsplus/inode.c struct hfsplus_cat_file *file = &entry.file; entry 597 fs/hfsplus/inode.c hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, entry 600 fs/hfsplus/inode.c hfs_bnode_write(fd.bnode, &entry, fd.entryoffset, entry 603 fs/hfsplus/inode.c struct hfsplus_cat_file *file = &entry.file; entry 607 fs/hfsplus/inode.c hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, entry 620 fs/hfsplus/inode.c hfs_bnode_write(fd.bnode, &entry, fd.entryoffset, entry 379 fs/hfsplus/super.c hfsplus_cat_entry entry; entry 526 fs/hfsplus/super.c if (!hfs_brec_read(&fd, &entry, sizeof(entry))) { entry 528 fs/hfsplus/super.c if (entry.type != cpu_to_be16(HFSPLUS_FOLDER)) { entry 532 fs/hfsplus/super.c inode = hfsplus_iget(sb, be32_to_cpu(entry.folder.id)); entry 263 fs/hfsplus/xattr.c hfsplus_cat_entry entry; entry 296 fs/hfsplus/xattr.c hfs_bnode_read(cat_fd.bnode, &entry, cat_fd.entryoffset, entry 298 fs/hfsplus/xattr.c if (be16_to_cpu(entry.type) == HFSPLUS_FOLDER) { entry 300 fs/hfsplus/xattr.c memcpy(&entry.folder.user_info, value, entry 302 fs/hfsplus/xattr.c hfs_bnode_write(cat_fd.bnode, &entry, entry 311 fs/hfsplus/xattr.c } else if (be16_to_cpu(entry.type) == HFSPLUS_FILE) { entry 313 fs/hfsplus/xattr.c memcpy(&entry.file.user_info, value, entry 315 fs/hfsplus/xattr.c hfs_bnode_write(cat_fd.bnode, &entry, entry 494 fs/hfsplus/xattr.c hfsplus_attr_entry *entry; entry 511 fs/hfsplus/xattr.c entry = hfsplus_alloc_attr_entry(); entry 512 fs/hfsplus/xattr.c if (!entry) { entry 557 fs/hfsplus/xattr.c hfs_bnode_read(fd.bnode, entry, fd.entryoffset, entry 563 fs/hfsplus/xattr.c memcpy(value, entry->inline_data.raw_bytes, record_length); entry 572 fs/hfsplus/xattr.c hfsplus_destroy_attr_entry(entry); entry 1729 fs/io_uring.c if (!list_empty(&poll->wait.entry)) { entry 1730 fs/io_uring.c list_del_init(&poll->wait.entry); entry 1840 fs/io_uring.c list_del_init(&poll->wait.entry); entry 1908 fs/io_uring.c INIT_LIST_HEAD(&poll->wait.entry); entry 1918 fs/io_uring.c if (unlikely(list_empty(&poll->wait.entry))) { entry 1925 fs/io_uring.c list_del_init(&poll->wait.entry); entry 1981 fs/io_uring.c struct list_head *entry; entry 2005 fs/io_uring.c entry = ctx->timeout_list.prev; entry 2018 fs/io_uring.c list_for_each_prev(entry, &ctx->timeout_list) { entry 2019 fs/io_uring.c struct io_kiocb *nxt = list_entry(entry, struct io_kiocb, list); entry 2053 fs/io_uring.c list_add(&req->list, entry); entry 2977 fs/io_uring.c .entry = LIST_HEAD_INIT(iowq.wq.entry), entry 59 fs/jffs2/acl.c struct jffs2_acl_entry *entry; entry 86 fs/jffs2/acl.c entry = value; entry 89 fs/jffs2/acl.c acl->a_entries[i].e_tag = je16_to_cpu(entry->e_tag); entry 90 fs/jffs2/acl.c acl->a_entries[i].e_perm = je16_to_cpu(entry->e_perm); entry 105 fs/jffs2/acl.c je32_to_cpu(entry->e_id)); entry 113 fs/jffs2/acl.c je32_to_cpu(entry->e_id)); entry 131 fs/jffs2/acl.c struct jffs2_acl_entry *entry; entry 144 fs/jffs2/acl.c entry = e; entry 145 fs/jffs2/acl.c entry->e_tag = cpu_to_je16(acl_e->e_tag); entry 146 fs/jffs2/acl.c entry->e_perm = cpu_to_je16(acl_e->e_perm); entry 149 fs/jffs2/acl.c entry->e_id = cpu_to_je32( entry 154 fs/jffs2/acl.c entry->e_id = cpu_to_je32( entry 4521 fs/jfs/jfs_dtree.c struct ldtentry *entry; entry 4553 fs/jfs/jfs_dtree.c entry = (struct ldtentry *) & p->slot[entry_si]; entry 4556 fs/jfs/jfs_dtree.c entry->inumber = cpu_to_le32(new_ino); entry 73 fs/lockd/procfs.c struct proc_dir_entry *entry; entry 75 fs/lockd/procfs.c entry = proc_mkdir("fs/lockd", NULL); entry 76 fs/lockd/procfs.c if (!entry) entry 78 fs/lockd/procfs.c entry = proc_create("nlm_end_grace", S_IRUGO|S_IWUSR, entry, entry 80 fs/lockd/procfs.c if (!entry) { entry 77 fs/mbcache.c struct mb_cache_entry *entry, *dup; entry 88 fs/mbcache.c entry = kmem_cache_alloc(mb_entry_cache, mask); entry 89 fs/mbcache.c if (!entry) entry 92 fs/mbcache.c INIT_LIST_HEAD(&entry->e_list); entry 94 fs/mbcache.c atomic_set(&entry->e_refcnt, 1); entry 95 fs/mbcache.c entry->e_key = key; entry 96 fs/mbcache.c entry->e_value = value; entry 97 fs/mbcache.c entry->e_reusable = reusable; entry 98 fs/mbcache.c entry->e_referenced = 0; entry 104 fs/mbcache.c kmem_cache_free(mb_entry_cache, entry); entry 108 fs/mbcache.c hlist_bl_add_head(&entry->e_hash_list, head); entry 112 fs/mbcache.c list_add_tail(&entry->e_list, &cache->c_list); entry 114 fs/mbcache.c atomic_inc(&entry->e_refcnt); entry 122 fs/mbcache.c void __mb_cache_entry_free(struct mb_cache_entry *entry) entry 124 fs/mbcache.c kmem_cache_free(mb_entry_cache, entry); entry 129 fs/mbcache.c struct mb_cache_entry *entry, entry 132 fs/mbcache.c struct mb_cache_entry *old_entry = entry; entry 138 fs/mbcache.c if (entry && !hlist_bl_unhashed(&entry->e_hash_list)) entry 139 fs/mbcache.c node = entry->e_hash_list.next; entry 143 fs/mbcache.c entry = hlist_bl_entry(node, struct mb_cache_entry, entry 145 fs/mbcache.c if (entry->e_key == key && entry->e_reusable) { entry 146 fs/mbcache.c atomic_inc(&entry->e_refcnt); entry 151 fs/mbcache.c entry = NULL; entry 157 fs/mbcache.c return entry; entry 186 fs/mbcache.c struct mb_cache_entry *entry) entry 188 fs/mbcache.c return __entry_find(cache, entry, entry->e_key); entry 203 fs/mbcache.c struct mb_cache_entry *entry; entry 207 fs/mbcache.c hlist_bl_for_each_entry(entry, node, head, e_hash_list) { entry 208 fs/mbcache.c if (entry->e_key == key && entry->e_value == value) { entry 209 fs/mbcache.c atomic_inc(&entry->e_refcnt); entry 213 fs/mbcache.c entry = NULL; entry 216 fs/mbcache.c return entry; entry 231 fs/mbcache.c struct mb_cache_entry *entry; entry 235 fs/mbcache.c hlist_bl_for_each_entry(entry, node, head, e_hash_list) { entry 236 fs/mbcache.c if (entry->e_key == key && entry->e_value == value) { entry 238 fs/mbcache.c hlist_bl_del_init(&entry->e_hash_list); entry 241 fs/mbcache.c if (!list_empty(&entry->e_list)) { entry 242 fs/mbcache.c list_del_init(&entry->e_list); entry 246 fs/mbcache.c atomic_dec(&entry->e_refcnt); entry 249 fs/mbcache.c mb_cache_entry_put(cache, entry); entry 264 fs/mbcache.c struct mb_cache_entry *entry) entry 266 fs/mbcache.c entry->e_referenced = 1; entry 283 fs/mbcache.c struct mb_cache_entry *entry; entry 289 fs/mbcache.c entry = list_first_entry(&cache->c_list, entry 291 fs/mbcache.c if (entry->e_referenced) { entry 292 fs/mbcache.c entry->e_referenced = 0; entry 293 fs/mbcache.c list_move_tail(&entry->e_list, &cache->c_list); entry 296 fs/mbcache.c list_del_init(&entry->e_list); entry 303 fs/mbcache.c head = mb_cache_entry_head(cache, entry->e_key); entry 305 fs/mbcache.c if (!hlist_bl_unhashed(&entry->e_hash_list)) { entry 306 fs/mbcache.c hlist_bl_del_init(&entry->e_hash_list); entry 307 fs/mbcache.c atomic_dec(&entry->e_refcnt); entry 310 fs/mbcache.c if (mb_cache_entry_put(cache, entry)) entry 394 fs/mbcache.c struct mb_cache_entry *entry, *next; entry 402 fs/mbcache.c list_for_each_entry_safe(entry, next, &cache->c_list, e_list) { entry 403 fs/mbcache.c if (!hlist_bl_unhashed(&entry->e_hash_list)) { entry 404 fs/mbcache.c hlist_bl_del_init(&entry->e_hash_list); entry 405 fs/mbcache.c atomic_dec(&entry->e_refcnt); entry 408 fs/mbcache.c list_del(&entry->e_list); entry 409 fs/mbcache.c WARN_ON(atomic_read(&entry->e_refcnt) != 1); entry 410 fs/mbcache.c mb_cache_entry_put(cache, entry); entry 214 fs/nfs/dir.c int nfs_readdir_add_to_array(struct nfs_entry *entry, struct page *page) entry 227 fs/nfs/dir.c cache_entry->cookie = entry->prev_cookie; entry 228 fs/nfs/dir.c cache_entry->ino = entry->ino; entry 229 fs/nfs/dir.c cache_entry->d_type = entry->d_type; entry 230 fs/nfs/dir.c ret = nfs_readdir_make_qstr(&cache_entry->string, entry->name, entry->len); entry 233 fs/nfs/dir.c array->last_cookie = entry->cookie; entry 235 fs/nfs/dir.c if (entry->eof != 0) entry 346 fs/nfs/dir.c struct nfs_entry *entry, struct file *file, struct inode *inode) entry 356 fs/nfs/dir.c error = NFS_PROTO(inode)->readdir(file_dentry(file), cred, entry->cookie, pages, entry 375 fs/nfs/dir.c struct nfs_entry *entry, struct xdr_stream *xdr) entry 379 fs/nfs/dir.c error = desc->decode(xdr, entry, desc->plus); entry 382 fs/nfs/dir.c entry->fattr->time_start = desc->timestamp; entry 383 fs/nfs/dir.c entry->fattr->gencount = desc->gencount; entry 391 fs/nfs/dir.c int nfs_same_file(struct dentry *dentry, struct nfs_entry *entry) entry 404 fs/nfs/dir.c if (entry->fattr->fileid != nfsi->fileid) entry 406 fs/nfs/dir.c if (entry->fh->size && nfs_compare_fh(entry->fh, &nfsi->fh) != 0) entry 457 fs/nfs/dir.c void nfs_prime_dcache(struct dentry *parent, struct nfs_entry *entry) entry 459 fs/nfs/dir.c struct qstr filename = QSTR_INIT(entry->name, entry->len); entry 467 fs/nfs/dir.c if (!(entry->fattr->valid & NFS_ATTR_FATTR_FILEID)) entry 469 fs/nfs/dir.c if (!(entry->fattr->valid & NFS_ATTR_FATTR_FSID)) entry 497 fs/nfs/dir.c &entry->fattr->fsid)) entry 499 fs/nfs/dir.c if (nfs_same_file(dentry, entry)) { entry 500 fs/nfs/dir.c if (!entry->fh->size) entry 503 fs/nfs/dir.c status = nfs_refresh_inode(d_inode(dentry), entry->fattr); entry 505 fs/nfs/dir.c nfs_setsecurity(d_inode(dentry), entry->fattr, entry->label); entry 514 fs/nfs/dir.c if (!entry->fh->size) { entry 519 fs/nfs/dir.c inode = nfs_fhget(dentry->d_sb, entry->fh, entry->fattr, entry->label); entry 535 fs/nfs/dir.c int nfs_readdir_page_filler(nfs_readdir_descriptor_t *desc, struct nfs_entry *entry, entry 556 fs/nfs/dir.c status = xdr_decode(desc, entry, &stream); entry 566 fs/nfs/dir.c nfs_prime_dcache(file_dentry(desc->file), entry); entry 568 fs/nfs/dir.c status = nfs_readdir_add_to_array(entry, page); entry 571 fs/nfs/dir.c } while (!entry->eof); entry 574 fs/nfs/dir.c if (count == 0 || (status == -EBADCOOKIE && entry->eof != 0)) { entry 619 fs/nfs/dir.c struct nfs_entry entry; entry 627 fs/nfs/dir.c entry.prev_cookie = 0; entry 628 fs/nfs/dir.c entry.cookie = desc->last_cookie; entry 629 fs/nfs/dir.c entry.eof = 0; entry 630 fs/nfs/dir.c entry.fh = nfs_alloc_fhandle(); entry 631 fs/nfs/dir.c entry.fattr = nfs_alloc_fattr(); entry 632 fs/nfs/dir.c entry.server = NFS_SERVER(inode); entry 633 fs/nfs/dir.c if (entry.fh == NULL || entry.fattr == NULL) entry 636 fs/nfs/dir.c entry.label = nfs4_label_alloc(NFS_SERVER(inode), GFP_NOWAIT); entry 637 fs/nfs/dir.c if (IS_ERR(entry.label)) { entry 638 fs/nfs/dir.c status = PTR_ERR(entry.label); entry 649 fs/nfs/dir.c status = nfs_readdir_xdr_filler(pages, desc, &entry, file, inode); entry 654 fs/nfs/dir.c status = nfs_readdir_page_filler(desc, &entry, pages, page, pglen); entry 665 fs/nfs/dir.c nfs4_label_free(entry.label); entry 667 fs/nfs/dir.c nfs_free_fattr(entry.fattr); entry 668 fs/nfs/dir.c nfs_free_fhandle(entry.fh); entry 2171 fs/nfs/dir.c static void nfs_access_free_entry(struct nfs_access_entry *entry) entry 2173 fs/nfs/dir.c put_cred(entry->cred); entry 2174 fs/nfs/dir.c kfree_rcu(entry, rcu_head); entry 2269 fs/nfs/dir.c struct nfs_access_entry *entry; entry 2273 fs/nfs/dir.c entry = rb_entry(n, struct nfs_access_entry, rb_node); entry 2275 fs/nfs/dir.c list_move(&entry->lru, head); entry 2304 fs/nfs/dir.c struct nfs_access_entry *entry = entry 2306 fs/nfs/dir.c int cmp = cred_fscmp(cred, entry->cred); entry 2313 fs/nfs/dir.c return entry; entry 2397 fs/nfs/dir.c struct nfs_access_entry *entry; entry 2403 fs/nfs/dir.c entry = rb_entry(parent, struct nfs_access_entry, rb_node); entry 2404 fs/nfs/dir.c cmp = cred_fscmp(set->cred, entry->cred); entry 2421 fs/nfs/dir.c list_del(&entry->lru); entry 2423 fs/nfs/dir.c nfs_access_free_entry(entry); entry 2490 fs/nfs/dir.c void nfs_access_set_mask(struct nfs_access_entry *entry, u32 access_result) entry 2492 fs/nfs/dir.c entry->mask = access_result; entry 930 fs/nfs/nfs2xdr.c int nfs2_decode_dirent(struct xdr_stream *xdr, struct nfs_entry *entry, entry 945 fs/nfs/nfs2xdr.c entry->eof = 1; entry 952 fs/nfs/nfs2xdr.c entry->ino = be32_to_cpup(p); entry 954 fs/nfs/nfs2xdr.c error = decode_filename_inline(xdr, &entry->name, &entry->len); entry 962 fs/nfs/nfs2xdr.c entry->prev_cookie = entry->cookie; entry 966 fs/nfs/nfs2xdr.c entry->cookie = be32_to_cpup(p); entry 968 fs/nfs/nfs2xdr.c entry->d_type = DT_UNKNOWN; entry 191 fs/nfs/nfs3proc.c static int nfs3_proc_access(struct inode *inode, struct nfs_access_entry *entry) entry 195 fs/nfs/nfs3proc.c .access = entry->mask, entry 202 fs/nfs/nfs3proc.c .rpc_cred = entry->cred, entry 214 fs/nfs/nfs3proc.c nfs_access_set_mask(entry, res.access); entry 1966 fs/nfs/nfs3xdr.c int nfs3_decode_dirent(struct xdr_stream *xdr, struct nfs_entry *entry, entry 1969 fs/nfs/nfs3xdr.c struct user_namespace *userns = rpc_userns(entry->server->client); entry 1970 fs/nfs/nfs3xdr.c struct nfs_entry old = *entry; entry 1984 fs/nfs/nfs3xdr.c entry->eof = 1; entry 1988 fs/nfs/nfs3xdr.c error = decode_fileid3(xdr, &entry->ino); entry 1992 fs/nfs/nfs3xdr.c error = decode_inline_filename3(xdr, &entry->name, &entry->len); entry 2000 fs/nfs/nfs3xdr.c entry->d_type = DT_UNKNOWN; entry 2003 fs/nfs/nfs3xdr.c entry->fattr->valid = 0; entry 2004 fs/nfs/nfs3xdr.c error = decode_post_op_attr(xdr, entry->fattr, userns); entry 2007 fs/nfs/nfs3xdr.c if (entry->fattr->valid & NFS_ATTR_FATTR_V3) entry 2008 fs/nfs/nfs3xdr.c entry->d_type = nfs_umode_to_dtype(entry->fattr->mode); entry 2010 fs/nfs/nfs3xdr.c if (entry->fattr->fileid != entry->ino) { entry 2011 fs/nfs/nfs3xdr.c entry->fattr->mounted_on_fileid = entry->ino; entry 2012 fs/nfs/nfs3xdr.c entry->fattr->valid |= NFS_ATTR_FATTR_MOUNTED_ON_FILEID; entry 2020 fs/nfs/nfs3xdr.c error = decode_nfs_fh3(xdr, entry->fh); entry 2027 fs/nfs/nfs3xdr.c zero_nfs_fh3(entry->fh); entry 2030 fs/nfs/nfs3xdr.c entry->prev_cookie = entry->cookie; entry 2031 fs/nfs/nfs3xdr.c entry->cookie = new_cookie; entry 2037 fs/nfs/nfs3xdr.c *entry = old; entry 4341 fs/nfs/nfs4proc.c static int _nfs4_proc_access(struct inode *inode, struct nfs_access_entry *entry) entry 4346 fs/nfs/nfs4proc.c .access = entry->mask, entry 4355 fs/nfs/nfs4proc.c .rpc_cred = entry->cred, entry 4367 fs/nfs/nfs4proc.c nfs_access_set_mask(entry, res.access); entry 4375 fs/nfs/nfs4proc.c static int nfs4_proc_access(struct inode *inode, struct nfs_access_entry *entry) entry 4382 fs/nfs/nfs4proc.c err = _nfs4_proc_access(inode, entry); entry 7105 fs/nfs/nfs4proc.c list_del_init(&wait->entry); entry 7362 fs/nfs/nfs4xdr.c int nfs4_decode_dirent(struct xdr_stream *xdr, struct nfs_entry *entry, entry 7378 fs/nfs/nfs4xdr.c entry->eof = 1; entry 7386 fs/nfs/nfs4xdr.c entry->len = be32_to_cpup(p); entry 7388 fs/nfs/nfs4xdr.c p = xdr_inline_decode(xdr, entry->len); entry 7391 fs/nfs/nfs4xdr.c entry->name = (const char *) p; entry 7398 fs/nfs/nfs4xdr.c entry->ino = 1; entry 7399 fs/nfs/nfs4xdr.c entry->fattr->valid = 0; entry 7407 fs/nfs/nfs4xdr.c if (decode_getfattr_attrs(xdr, bitmap, entry->fattr, entry->fh, entry 7408 fs/nfs/nfs4xdr.c NULL, entry->label, entry->server) < 0) entry 7410 fs/nfs/nfs4xdr.c if (entry->fattr->valid & NFS_ATTR_FATTR_MOUNTED_ON_FILEID) entry 7411 fs/nfs/nfs4xdr.c entry->ino = entry->fattr->mounted_on_fileid; entry 7412 fs/nfs/nfs4xdr.c else if (entry->fattr->valid & NFS_ATTR_FATTR_FILEID) entry 7413 fs/nfs/nfs4xdr.c entry->ino = entry->fattr->fileid; entry 7415 fs/nfs/nfs4xdr.c entry->d_type = DT_UNKNOWN; entry 7416 fs/nfs/nfs4xdr.c if (entry->fattr->valid & NFS_ATTR_FATTR_TYPE) entry 7417 fs/nfs/nfs4xdr.c entry->d_type = nfs_umode_to_dtype(entry->fattr->mode); entry 7419 fs/nfs/nfs4xdr.c entry->prev_cookie = entry->cookie; entry 7420 fs/nfs/nfs4xdr.c entry->cookie = new_cookie; entry 55 fs/nfs_common/nfsacl.c struct posix_acl_entry *entry = entry 58 fs/nfs_common/nfsacl.c *p++ = htonl(entry->e_tag | nfsacl_desc->typeflag); entry 59 fs/nfs_common/nfsacl.c switch(entry->e_tag) { entry 67 fs/nfs_common/nfsacl.c *p++ = htonl(from_kuid(&init_user_ns, entry->e_uid)); entry 70 fs/nfs_common/nfsacl.c *p++ = htonl(from_kgid(&init_user_ns, entry->e_gid)); entry 76 fs/nfs_common/nfsacl.c *p++ = htonl(entry->e_perm & S_IRWXO); entry 151 fs/nfs_common/nfsacl.c struct posix_acl_entry *entry; entry 163 fs/nfs_common/nfsacl.c entry = &nfsacl_desc->acl->a_entries[nfsacl_desc->count++]; entry 164 fs/nfs_common/nfsacl.c entry->e_tag = ntohl(*p++) & ~NFS_ACL_DEFAULT; entry 166 fs/nfs_common/nfsacl.c entry->e_perm = ntohl(*p++); entry 168 fs/nfs_common/nfsacl.c switch(entry->e_tag) { entry 170 fs/nfs_common/nfsacl.c entry->e_uid = make_kuid(&init_user_ns, id); entry 171 fs/nfs_common/nfsacl.c if (!uid_valid(entry->e_uid)) entry 175 fs/nfs_common/nfsacl.c entry->e_gid = make_kgid(&init_user_ns, id); entry 176 fs/nfs_common/nfsacl.c if (!gid_valid(entry->e_gid)) entry 182 fs/nfs_common/nfsacl.c if (entry->e_perm & ~S_IRWXO) entry 187 fs/nfs_common/nfsacl.c entry->e_perm &= S_IRWXO; entry 283 fs/nfsd/nfs4recover.c struct name_list *entry; entry 287 fs/nfsd/nfs4recover.c entry = kmalloc(sizeof(struct name_list), GFP_KERNEL); entry 288 fs/nfsd/nfs4recover.c if (entry == NULL) entry 290 fs/nfsd/nfs4recover.c memcpy(entry->name, name, HEXDIR_LEN - 1); entry 291 fs/nfsd/nfs4recover.c entry->name[HEXDIR_LEN - 1] = '\0'; entry 292 fs/nfsd/nfs4recover.c list_add(&entry->list, &ctx->names); entry 305 fs/nfsd/nfs4recover.c struct name_list *entry, *tmp; entry 321 fs/nfsd/nfs4recover.c list_for_each_entry_safe(entry, tmp, &ctx.names, list) { entry 324 fs/nfsd/nfs4recover.c dentry = lookup_one_len(entry->name, dir, HEXDIR_LEN-1); entry 332 fs/nfsd/nfs4recover.c list_del(&entry->list); entry 333 fs/nfsd/nfs4recover.c kfree(entry); entry 338 fs/nfsd/nfs4recover.c list_for_each_entry_safe(entry, tmp, &ctx.names, list) { entry 339 fs/nfsd/nfs4recover.c dprintk("NFSD: %s. Left entry %s\n", __func__, entry->name); entry 340 fs/nfsd/nfs4recover.c list_del(&entry->list); entry 341 fs/nfsd/nfs4recover.c kfree(entry); entry 1430 fs/nfsd/nfsctl.c struct proc_dir_entry *entry; entry 1432 fs/nfsd/nfsctl.c entry = proc_mkdir("fs/nfs", NULL); entry 1433 fs/nfsd/nfsctl.c if (!entry) entry 1435 fs/nfsd/nfsctl.c entry = proc_create("exports", 0, entry, entry 1437 fs/nfsd/nfsctl.c if (!entry) { entry 78 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 82 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, entry 84 fs/nilfs2/dat.c entry->de_start = cpu_to_le64(NILFS_CNO_MIN); entry 85 fs/nilfs2/dat.c entry->de_end = cpu_to_le64(NILFS_CNO_MAX); entry 86 fs/nilfs2/dat.c entry->de_blocknr = cpu_to_le64(0); entry 102 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 106 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, entry 108 fs/nilfs2/dat.c entry->de_start = cpu_to_le64(NILFS_CNO_MIN); entry 109 fs/nilfs2/dat.c entry->de_end = cpu_to_le64(NILFS_CNO_MIN); entry 110 fs/nilfs2/dat.c entry->de_blocknr = cpu_to_le64(0); entry 129 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 133 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, entry 135 fs/nilfs2/dat.c entry->de_start = cpu_to_le64(nilfs_mdt_cno(dat)); entry 136 fs/nilfs2/dat.c entry->de_blocknr = cpu_to_le64(blocknr); entry 144 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 156 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, entry 158 fs/nilfs2/dat.c blocknr = le64_to_cpu(entry->de_blocknr); entry 175 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 181 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, entry 183 fs/nilfs2/dat.c end = start = le64_to_cpu(entry->de_start); entry 188 fs/nilfs2/dat.c entry->de_end = cpu_to_le64(end); entry 189 fs/nilfs2/dat.c blocknr = le64_to_cpu(entry->de_blocknr); entry 200 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 206 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr, entry 208 fs/nilfs2/dat.c start = le64_to_cpu(entry->de_start); entry 209 fs/nilfs2/dat.c blocknr = le64_to_cpu(entry->de_blocknr); entry 316 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 341 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry(dat, vblocknr, entry_bh, kaddr); entry 342 fs/nilfs2/dat.c if (unlikely(entry->de_blocknr == cpu_to_le64(0))) { entry 346 fs/nilfs2/dat.c (unsigned long long)le64_to_cpu(entry->de_start), entry 347 fs/nilfs2/dat.c (unsigned long long)le64_to_cpu(entry->de_end)); entry 353 fs/nilfs2/dat.c entry->de_blocknr = cpu_to_le64(blocknr); entry 386 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 405 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry(dat, vblocknr, entry_bh, kaddr); entry 406 fs/nilfs2/dat.c blocknr = le64_to_cpu(entry->de_blocknr); entry 423 fs/nilfs2/dat.c struct nilfs_dat_entry *entry; entry 445 fs/nilfs2/dat.c entry = nilfs_palloc_block_get_entry( entry 447 fs/nilfs2/dat.c vinfo->vi_start = le64_to_cpu(entry->de_start); entry 448 fs/nilfs2/dat.c vinfo->vi_end = le64_to_cpu(entry->de_end); entry 449 fs/nilfs2/dat.c vinfo->vi_blocknr = le64_to_cpu(entry->de_blocknr); entry 2198 fs/nilfs2/segment.c list_for_each_entry_safe(wrq, n, &sci->sc_wait_request.head, wq.entry) { entry 45 fs/ntfs/index.c if (ictx->entry) { entry 202 fs/ntfs/index.c ictx->entry = ie; entry 63 fs/ntfs/index.h INDEX_ENTRY *entry; entry 51 fs/ocfs2/acl.c struct ocfs2_acl_entry *entry = entry 54 fs/ocfs2/acl.c acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); entry 55 fs/ocfs2/acl.c acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); entry 60 fs/ocfs2/acl.c le32_to_cpu(entry->e_id)); entry 65 fs/ocfs2/acl.c le32_to_cpu(entry->e_id)); entry 81 fs/ocfs2/acl.c struct ocfs2_acl_entry *entry = NULL; entry 91 fs/ocfs2/acl.c entry = (struct ocfs2_acl_entry *)ocfs2_acl; entry 92 fs/ocfs2/acl.c for (n = 0; n < acl->a_count; n++, entry++) { entry 93 fs/ocfs2/acl.c entry->e_tag = cpu_to_le16(acl->a_entries[n].e_tag); entry 94 fs/ocfs2/acl.c entry->e_perm = cpu_to_le16(acl->a_entries[n].e_perm); entry 97 fs/ocfs2/acl.c entry->e_id = cpu_to_le32( entry 102 fs/ocfs2/acl.c entry->e_id = cpu_to_le32( entry 107 fs/ocfs2/acl.c entry->e_id = cpu_to_le32(ACL_UNDEFINED_ID); entry 102 fs/ocfs2/filecheck.c struct ocfs2_filecheck_sysfs_entry *entry = container_of(kobj, entry 105 fs/ocfs2/filecheck.c complete(&entry->fs_kobj_unregister); entry 149 fs/ocfs2/filecheck.c ocfs2_filecheck_sysfs_free(struct ocfs2_filecheck_sysfs_entry *entry) entry 153 fs/ocfs2/filecheck.c spin_lock(&entry->fs_fcheck->fc_lock); entry 154 fs/ocfs2/filecheck.c while (!list_empty(&entry->fs_fcheck->fc_head)) { entry 155 fs/ocfs2/filecheck.c p = list_first_entry(&entry->fs_fcheck->fc_head, entry 161 fs/ocfs2/filecheck.c spin_unlock(&entry->fs_fcheck->fc_lock); entry 163 fs/ocfs2/filecheck.c kfree(entry->fs_fcheck); entry 164 fs/ocfs2/filecheck.c entry->fs_fcheck = NULL; entry 171 fs/ocfs2/filecheck.c struct ocfs2_filecheck_sysfs_entry *entry = &osb->osb_fc_ent; entry 183 fs/ocfs2/filecheck.c entry->fs_kobj.kset = osb->osb_dev_kset; entry 184 fs/ocfs2/filecheck.c init_completion(&entry->fs_kobj_unregister); entry 185 fs/ocfs2/filecheck.c ret = kobject_init_and_add(&entry->fs_kobj, &ocfs2_ktype_filecheck, entry 188 fs/ocfs2/filecheck.c kobject_put(&entry->fs_kobj); entry 193 fs/ocfs2/filecheck.c entry->fs_fcheck = fcheck; entry 402 fs/ocfs2/filecheck.c struct ocfs2_filecheck_entry *entry) entry 405 fs/ocfs2/filecheck.c entry->fe_done = 1; entry 434 fs/ocfs2/filecheck.c struct ocfs2_filecheck_entry *entry) entry 439 fs/ocfs2/filecheck.c if (entry->fe_type == OCFS2_FILECHECK_TYPE_CHK) entry 440 fs/ocfs2/filecheck.c entry->fe_status = ocfs2_filecheck_handle(osb, entry 441 fs/ocfs2/filecheck.c entry->fe_ino, OCFS2_FI_FLAG_FILECHECK_CHK); entry 442 fs/ocfs2/filecheck.c else if (entry->fe_type == OCFS2_FILECHECK_TYPE_FIX) entry 443 fs/ocfs2/filecheck.c entry->fe_status = ocfs2_filecheck_handle(osb, entry 444 fs/ocfs2/filecheck.c entry->fe_ino, OCFS2_FI_FLAG_FILECHECK_FIX); entry 446 fs/ocfs2/filecheck.c entry->fe_status = OCFS2_FILECHECK_ERR_UNSUPPORTED; entry 448 fs/ocfs2/filecheck.c ocfs2_filecheck_done_entry(ent, entry); entry 457 fs/ocfs2/filecheck.c struct ocfs2_filecheck_entry *entry; entry 472 fs/ocfs2/filecheck.c entry = kmalloc(sizeof(struct ocfs2_filecheck_entry), GFP_NOFS); entry 473 fs/ocfs2/filecheck.c if (!entry) { entry 481 fs/ocfs2/filecheck.c kfree(entry); entry 489 fs/ocfs2/filecheck.c kfree(entry); entry 500 fs/ocfs2/filecheck.c entry->fe_ino = args.fa_ino; entry 501 fs/ocfs2/filecheck.c entry->fe_type = args.fa_type; entry 502 fs/ocfs2/filecheck.c entry->fe_done = 0; entry 503 fs/ocfs2/filecheck.c entry->fe_status = OCFS2_FILECHECK_ERR_INPROGRESS; entry 504 fs/ocfs2/filecheck.c list_add_tail(&entry->fe_list, &ent->fs_fcheck->fc_head); entry 510 fs/ocfs2/filecheck.c ocfs2_filecheck_handle_entry(ent, entry); entry 939 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry = &header->xh_entries[i]; entry 940 fs/ocfs2/xattr.c type = ocfs2_xattr_get_type(entry); entry 942 fs/ocfs2/xattr.c le16_to_cpu(entry->xe_name_offset); entry 947 fs/ocfs2/xattr.c entry->xe_name_len); entry 1075 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry; entry 1083 fs/ocfs2/xattr.c entry = xs->here; entry 1085 fs/ocfs2/xattr.c cmp = name_index - ocfs2_xattr_get_type(entry); entry 1087 fs/ocfs2/xattr.c cmp = name_len - entry->xe_name_len; entry 1090 fs/ocfs2/xattr.c le16_to_cpu(entry->xe_name_offset)), entry 1094 fs/ocfs2/xattr.c entry += 1; entry 1096 fs/ocfs2/xattr.c xs->here = entry; entry 1630 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry = loc->xl_entry; entry 1634 fs/ocfs2/xattr.c namevalue_offset = le16_to_cpu(entry->xe_name_offset); entry 1635 fs/ocfs2/xattr.c namevalue_size = namevalue_size_xe(entry); entry 1936 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry = loc->xl_entry; entry 1950 fs/ocfs2/xattr.c index = ((char *)entry - (char *)&xh->xh_entries) / entry 2270 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry) entry 2279 fs/ocfs2/xattr.c loc->xl_entry = entry; entry 2289 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry) entry 2300 fs/ocfs2/xattr.c loc->xl_entry = entry; entry 2307 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry) entry 2313 fs/ocfs2/xattr.c loc->xl_entry = entry; entry 2376 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry = &header->xh_entries[i]; entry 2378 fs/ocfs2/xattr.c if (ocfs2_xattr_is_local(entry)) entry 2382 fs/ocfs2/xattr.c le16_to_cpu(entry->xe_name_offset); entry 2384 fs/ocfs2/xattr.c (val + OCFS2_XATTR_SIZE(entry->xe_name_len)); entry 4055 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *entry = &bucket_xh(bucket)->xh_entries[i]; entry 4056 fs/ocfs2/xattr.c type = ocfs2_xattr_get_type(entry); entry 4073 fs/ocfs2/xattr.c entry->xe_name_len); entry 119 fs/omfs/dir.c __be64 *entry; entry 127 fs/omfs/dir.c entry = (__be64 *) &bh->b_data[ofs]; entry 128 fs/omfs/dir.c block = be64_to_cpu(*entry); entry 129 fs/omfs/dir.c *entry = cpu_to_be64(inode->i_ino); entry 164 fs/omfs/dir.c __be64 *entry, next; entry 174 fs/omfs/dir.c entry = (__be64 *) &bh->b_data[ofs]; entry 175 fs/omfs/dir.c block = be64_to_cpu(*entry); entry 195 fs/omfs/dir.c entry = &oi->i_sibling; entry 198 fs/omfs/dir.c *entry = next; entry 35 fs/omfs/file.c struct omfs_extent_entry *entry; entry 71 fs/omfs/file.c entry = &oe->e_entry; entry 76 fs/omfs/file.c start = be64_to_cpu(entry->e_cluster); entry 77 fs/omfs/file.c count = be64_to_cpu(entry->e_blocks); entry 80 fs/omfs/file.c entry++; entry 120 fs/omfs/file.c struct omfs_extent_entry *entry = &oe->e_entry; entry 142 fs/omfs/file.c terminator = entry + extent_count - 1; entry 144 fs/omfs/file.c entry = terminator-1; entry 145 fs/omfs/file.c new_block = be64_to_cpu(entry->e_cluster) + entry 146 fs/omfs/file.c be64_to_cpu(entry->e_blocks); entry 149 fs/omfs/file.c be64_add_cpu(&entry->e_blocks, 1); entry 168 fs/omfs/file.c entry = terminator; entry 170 fs/omfs/file.c memcpy(terminator, entry, sizeof(struct omfs_extent_entry)); entry 172 fs/omfs/file.c entry->e_cluster = cpu_to_be64(new_block); entry 173 fs/omfs/file.c entry->e_blocks = cpu_to_be64((u64) new_count); entry 227 fs/omfs/file.c struct omfs_extent_entry *entry; entry 248 fs/omfs/file.c entry = &oe->e_entry; entry 253 fs/omfs/file.c offset = find_block(inode, entry, block, extent_count, &remain); entry 50 fs/orangefs/orangefs-bufmap.c if (likely(list_empty(&wait.entry))) entry 88 fs/orangefs/orangefs-bufmap.c if (likely(list_empty(&wait.entry))) entry 112 fs/orangefs/orangefs-bufmap.c if (!list_empty(&wait.entry)) entry 113 fs/orangefs/orangefs-bufmap.c list_del(&wait.entry); entry 671 fs/posix_acl.c struct posix_acl_xattr_entry *entry = (void *)(header + 1), *end; entry 689 fs/posix_acl.c for (end = entry + count; entry != end; entry++) { entry 690 fs/posix_acl.c switch(le16_to_cpu(entry->e_tag)) { entry 692 fs/posix_acl.c uid = make_kuid(from, le32_to_cpu(entry->e_id)); entry 693 fs/posix_acl.c entry->e_id = cpu_to_le32(from_kuid(to, uid)); entry 696 fs/posix_acl.c gid = make_kgid(from, le32_to_cpu(entry->e_id)); entry 697 fs/posix_acl.c entry->e_id = cpu_to_le32(from_kgid(to, gid)); entry 729 fs/posix_acl.c const struct posix_acl_xattr_entry *entry = (const void *)(header + 1), *end; entry 752 fs/posix_acl.c for (end = entry + count; entry != end; acl_e++, entry++) { entry 753 fs/posix_acl.c acl_e->e_tag = le16_to_cpu(entry->e_tag); entry 754 fs/posix_acl.c acl_e->e_perm = le16_to_cpu(entry->e_perm); entry 766 fs/posix_acl.c le32_to_cpu(entry->e_id)); entry 773 fs/posix_acl.c le32_to_cpu(entry->e_id)); entry 113 fs/proc/namespaces.c const struct proc_ns_operations **entry, **last; entry 122 fs/proc/namespaces.c entry = ns_entries + (ctx->pos - 2); entry 124 fs/proc/namespaces.c while (entry <= last) { entry 125 fs/proc/namespaces.c const struct proc_ns_operations *ops = *entry; entry 130 fs/proc/namespaces.c entry++; entry 147 fs/proc/namespaces.c const struct proc_ns_operations **entry, **last; entry 155 fs/proc/namespaces.c for (entry = ns_entries; entry < last; entry++) { entry 156 fs/proc/namespaces.c if (strlen((*entry)->name) != len) entry 158 fs/proc/namespaces.c if (!memcmp(dentry->d_name.name, (*entry)->name, len)) entry 161 fs/proc/namespaces.c if (entry == last) entry 164 fs/proc/namespaces.c res = proc_ns_instantiate(dentry, task, *entry); entry 113 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 124 fs/proc/proc_sysctl.c entry = &head->ctl_table[ctl_node - head->node]; entry 125 fs/proc/proc_sysctl.c procname = entry->procname; entry 134 fs/proc/proc_sysctl.c return entry; entry 140 fs/proc/proc_sysctl.c static int insert_entry(struct ctl_table_header *head, struct ctl_table *entry) entry 142 fs/proc/proc_sysctl.c struct rb_node *node = &head->node[entry - head->ctl_table].node; entry 145 fs/proc/proc_sysctl.c const char *name = entry->procname; entry 169 fs/proc/proc_sysctl.c pr_cont("/%s\n", entry->procname); entry 179 fs/proc/proc_sysctl.c static void erase_entry(struct ctl_table_header *head, struct ctl_table *entry) entry 181 fs/proc/proc_sysctl.c struct rb_node *node = &head->node[entry - head->ctl_table].node; entry 202 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 203 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++, node++) entry 210 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 211 fs/proc/proc_sysctl.c for (entry = head->ctl_table; entry->procname; entry++) entry 212 fs/proc/proc_sysctl.c erase_entry(head, entry); entry 217 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 236 fs/proc/proc_sysctl.c for (entry = header->ctl_table; entry->procname; entry++) { entry 237 fs/proc/proc_sysctl.c err = insert_entry(header, entry); entry 372 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 375 fs/proc/proc_sysctl.c entry = find_entry(&head, dir, name, namelen); entry 376 fs/proc/proc_sysctl.c if (entry && use_table(head)) entry 379 fs/proc/proc_sysctl.c entry = NULL; entry 381 fs/proc/proc_sysctl.c return entry; entry 400 fs/proc/proc_sysctl.c struct ctl_table *entry = NULL; entry 408 fs/proc/proc_sysctl.c entry = &head->ctl_table[ctl_node - head->node]; entry 411 fs/proc/proc_sysctl.c *pentry = entry; entry 417 fs/proc/proc_sysctl.c struct ctl_table *entry = *pentry; entry 418 fs/proc/proc_sysctl.c struct ctl_node *ctl_node = &head->node[entry - head->ctl_table]; entry 428 fs/proc/proc_sysctl.c entry = &head->ctl_table[ctl_node - head->node]; entry 431 fs/proc/proc_sysctl.c *pentry = entry; entry 782 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 796 fs/proc/proc_sysctl.c for (first_entry(ctl_dir, &h, &entry); h; next_entry(&h, &entry)) { entry 797 fs/proc/proc_sysctl.c if (!scan(h, entry, &pos, file, ctx)) { entry 954 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 956 fs/proc/proc_sysctl.c entry = find_entry(&head, dir, name, namelen); entry 957 fs/proc/proc_sysctl.c if (!entry) entry 959 fs/proc/proc_sysctl.c if (!S_ISDIR(entry->mode)) entry 1071 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 1085 fs/proc/proc_sysctl.c entry = find_entry(&head, dir, procname, strlen(procname)); entry 1087 fs/proc/proc_sysctl.c if (entry && use_table(head)) { entry 1090 fs/proc/proc_sysctl.c *pentry = entry; entry 1165 fs/proc/proc_sysctl.c struct ctl_table *link_table, *entry, *link; entry 1173 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++) { entry 1175 fs/proc/proc_sysctl.c name_bytes += strlen(entry->procname) + 1; entry 1191 fs/proc/proc_sysctl.c for (link = link_table, entry = table; entry->procname; link++, entry++) { entry 1192 fs/proc/proc_sysctl.c int len = strlen(entry->procname) + 1; entry 1193 fs/proc/proc_sysctl.c memcpy(link_name, entry->procname, len); entry 1209 fs/proc/proc_sysctl.c struct ctl_table *entry, *link; entry 1212 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++) { entry 1213 fs/proc/proc_sysctl.c const char *procname = entry->procname; entry 1217 fs/proc/proc_sysctl.c if (S_ISDIR(link->mode) && S_ISDIR(entry->mode)) entry 1225 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++) { entry 1226 fs/proc/proc_sysctl.c const char *procname = entry->procname; entry 1324 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 1328 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++) entry 1417 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 1423 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++) { entry 1424 fs/proc/proc_sysctl.c if (entry->child) entry 1425 fs/proc/proc_sysctl.c nr_subheaders += count_subheaders(entry->child); entry 1437 fs/proc/proc_sysctl.c struct ctl_table *entry, *files; entry 1442 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++) { entry 1443 fs/proc/proc_sysctl.c if (entry->child) entry 1459 fs/proc/proc_sysctl.c for (new = files, entry = table; entry->procname; entry++) { entry 1460 fs/proc/proc_sysctl.c if (entry->child) entry 1462 fs/proc/proc_sysctl.c *new = *entry; entry 1483 fs/proc/proc_sysctl.c for (entry = table; entry->procname; entry++) { entry 1486 fs/proc/proc_sysctl.c if (!entry->child) entry 1490 fs/proc/proc_sysctl.c child_pos = append_path(path, pos, entry->procname); entry 1495 fs/proc/proc_sysctl.c set, entry->child); entry 1619 fs/proc/proc_sysctl.c struct ctl_table *entry; entry 1628 fs/proc/proc_sysctl.c for (entry = header->ctl_table; entry->procname; entry++) { entry 1631 fs/proc/proc_sysctl.c const char *name = entry->procname; entry 1635 fs/proc/proc_sysctl.c ((S_ISDIR(link->mode) && S_ISDIR(entry->mode)) || entry 1338 fs/proc/task_mmu.c swp_entry_t entry; entry 1341 fs/proc/task_mmu.c entry = pte_to_swp_entry(pte); entry 1343 fs/proc/task_mmu.c frame = swp_type(entry) | entry 1344 fs/proc/task_mmu.c (swp_offset(entry) << MAX_SWAPFILES_SHIFT); entry 1346 fs/proc/task_mmu.c if (is_migration_entry(entry)) entry 1347 fs/proc/task_mmu.c page = migration_entry_to_page(entry); entry 1349 fs/proc/task_mmu.c if (is_device_private_entry(entry)) entry 1350 fs/proc/task_mmu.c page = device_private_entry_to_page(entry); entry 1394 fs/proc/task_mmu.c swp_entry_t entry = pmd_to_swp_entry(pmd); entry 1398 fs/proc/task_mmu.c offset = swp_offset(entry) + entry 1400 fs/proc/task_mmu.c frame = swp_type(entry) | entry 1407 fs/proc/task_mmu.c page = migration_entry_to_page(entry); entry 161 fs/reiserfs/journal.c struct list_head *entry = journal->j_bitmap_nodes.next; entry 166 fs/reiserfs/journal.c if (entry != &journal->j_bitmap_nodes) { entry 167 fs/reiserfs/journal.c bn = list_entry(entry, struct reiserfs_bitmap_node, list); entry 168 fs/reiserfs/journal.c list_del(entry); entry 584 fs/reiserfs/journal.c struct list_head *entry = &journal->j_journal_list; entry 587 fs/reiserfs/journal.c if (!list_empty(entry)) { entry 588 fs/reiserfs/journal.c jl = JOURNAL_LIST_ENTRY(entry->next); entry 891 fs/reiserfs/journal.c struct list_head *entry; entry 900 fs/reiserfs/journal.c entry = jl->j_list.prev; entry 902 fs/reiserfs/journal.c other_jl = JOURNAL_LIST_ENTRY(entry); entry 903 fs/reiserfs/journal.c if (entry == &journal->j_journal_list || entry 908 fs/reiserfs/journal.c entry = other_jl->j_list.prev; entry 916 fs/reiserfs/journal.c entry = &first_jl->j_list; entry 918 fs/reiserfs/journal.c other_jl = JOURNAL_LIST_ENTRY(entry); entry 940 fs/reiserfs/journal.c entry = entry->next; entry 941 fs/reiserfs/journal.c if (entry == &journal->j_journal_list) entry 1301 fs/reiserfs/journal.c struct list_head *entry; entry 1311 fs/reiserfs/journal.c entry = journal->j_journal_list.next; entry 1313 fs/reiserfs/journal.c if (entry == &journal->j_journal_list) entry 1315 fs/reiserfs/journal.c other_jl = JOURNAL_LIST_ENTRY(entry); entry 1727 fs/reiserfs/journal.c struct list_head *entry; entry 1755 fs/reiserfs/journal.c entry = jl->j_list.next; entry 1758 fs/reiserfs/journal.c if (entry == &journal->j_journal_list) { entry 1761 fs/reiserfs/journal.c jl = JOURNAL_LIST_ENTRY(entry); entry 3534 fs/reiserfs/journal.c struct list_head *entry; entry 3539 fs/reiserfs/journal.c entry = journal->j_journal_list.prev; entry 3540 fs/reiserfs/journal.c jl = JOURNAL_LIST_ENTRY(entry); entry 3987 fs/reiserfs/journal.c struct list_head *entry, *safe; entry 4307 fs/reiserfs/journal.c list_for_each_safe(entry, safe, &journal->j_journal_list) { entry 4308 fs/reiserfs/journal.c temp_jl = JOURNAL_LIST_ENTRY(entry); entry 88 fs/reiserfs/xattr_acl.c reiserfs_acl_entry *entry = (reiserfs_acl_entry *) value; entry 91 fs/reiserfs/xattr_acl.c acl->a_entries[n].e_tag = le16_to_cpu(entry->e_tag); entry 92 fs/reiserfs/xattr_acl.c acl->a_entries[n].e_perm = le16_to_cpu(entry->e_perm); entry 108 fs/reiserfs/xattr_acl.c le32_to_cpu(entry->e_id)); entry 116 fs/reiserfs/xattr_acl.c le32_to_cpu(entry->e_id)); entry 152 fs/reiserfs/xattr_acl.c reiserfs_acl_entry *entry = (reiserfs_acl_entry *) e; entry 153 fs/reiserfs/xattr_acl.c entry->e_tag = cpu_to_le16(acl->a_entries[n].e_tag); entry 154 fs/reiserfs/xattr_acl.c entry->e_perm = cpu_to_le16(acl->a_entries[n].e_perm); entry 157 fs/reiserfs/xattr_acl.c entry->e_id = cpu_to_le32( entry 162 fs/reiserfs/xattr_acl.c entry->e_id = cpu_to_le32( entry 99 fs/select.c struct poll_table_entry * entry; entry 104 fs/select.c ((unsigned long)((table)->entry+1) > PAGE_SIZE + (unsigned long)(table)) entry 132 fs/select.c static void free_poll_entry(struct poll_table_entry *entry) entry 134 fs/select.c remove_wait_queue(entry->wait_address, &entry->wait); entry 135 fs/select.c fput(entry->filp); entry 145 fs/select.c struct poll_table_entry * entry; entry 148 fs/select.c entry = p->entry; entry 150 fs/select.c entry--; entry 151 fs/select.c free_poll_entry(entry); entry 152 fs/select.c } while (entry > p->entries); entry 175 fs/select.c new_table->entry = new_table->entries; entry 181 fs/select.c return table->entry++; entry 212 fs/select.c struct poll_table_entry *entry; entry 214 fs/select.c entry = container_of(wait, struct poll_table_entry, wait); entry 215 fs/select.c if (key && !(key_to_poll(key) & entry->key)) entry 225 fs/select.c struct poll_table_entry *entry = poll_get_entry(pwq); entry 226 fs/select.c if (!entry) entry 228 fs/select.c entry->filp = get_file(filp); entry 229 fs/select.c entry->wait_address = wait_address; entry 230 fs/select.c entry->key = p->_key; entry 231 fs/select.c init_waitqueue_func_entry(&entry->wait, pollwake); entry 232 fs/select.c entry->wait.private = pwq; entry 233 fs/select.c add_wait_queue(wait_address, &entry->wait); entry 56 fs/squashfs/cache.c struct squashfs_cache_entry *entry; entry 62 fs/squashfs/cache.c if (cache->entry[i].block == block) { entry 90 fs/squashfs/cache.c if (cache->entry[i].refcount == 0) entry 96 fs/squashfs/cache.c entry = &cache->entry[i]; entry 103 fs/squashfs/cache.c entry->block = block; entry 104 fs/squashfs/cache.c entry->refcount = 1; entry 105 fs/squashfs/cache.c entry->pending = 1; entry 106 fs/squashfs/cache.c entry->num_waiters = 0; entry 107 fs/squashfs/cache.c entry->error = 0; entry 110 fs/squashfs/cache.c entry->length = squashfs_read_data(sb, block, length, entry 111 fs/squashfs/cache.c &entry->next_index, entry->actor); entry 115 fs/squashfs/cache.c if (entry->length < 0) entry 116 fs/squashfs/cache.c entry->error = entry->length; entry 118 fs/squashfs/cache.c entry->pending = 0; entry 125 fs/squashfs/cache.c if (entry->num_waiters) { entry 127 fs/squashfs/cache.c wake_up_all(&entry->wait_queue); entry 140 fs/squashfs/cache.c entry = &cache->entry[i]; entry 141 fs/squashfs/cache.c if (entry->refcount == 0) entry 143 fs/squashfs/cache.c entry->refcount++; entry 149 fs/squashfs/cache.c if (entry->pending) { entry 150 fs/squashfs/cache.c entry->num_waiters++; entry 152 fs/squashfs/cache.c wait_event(entry->wait_queue, !entry->pending); entry 161 fs/squashfs/cache.c cache->name, i, entry->block, entry->refcount, entry->error); entry 163 fs/squashfs/cache.c if (entry->error) entry 166 fs/squashfs/cache.c return entry; entry 173 fs/squashfs/cache.c void squashfs_cache_put(struct squashfs_cache_entry *entry) entry 175 fs/squashfs/cache.c struct squashfs_cache *cache = entry->cache; entry 178 fs/squashfs/cache.c entry->refcount--; entry 179 fs/squashfs/cache.c if (entry->refcount == 0) { entry 205 fs/squashfs/cache.c if (cache->entry[i].data) { entry 207 fs/squashfs/cache.c kfree(cache->entry[i].data[j]); entry 208 fs/squashfs/cache.c kfree(cache->entry[i].data); entry 210 fs/squashfs/cache.c kfree(cache->entry[i].actor); entry 213 fs/squashfs/cache.c kfree(cache->entry); entry 234 fs/squashfs/cache.c cache->entry = kcalloc(entries, sizeof(*(cache->entry)), GFP_KERNEL); entry 235 fs/squashfs/cache.c if (cache->entry == NULL) { entry 253 fs/squashfs/cache.c struct squashfs_cache_entry *entry = &cache->entry[i]; entry 255 fs/squashfs/cache.c init_waitqueue_head(&cache->entry[i].wait_queue); entry 256 fs/squashfs/cache.c entry->cache = cache; entry 257 fs/squashfs/cache.c entry->block = SQUASHFS_INVALID_BLK; entry 258 fs/squashfs/cache.c entry->data = kcalloc(cache->pages, sizeof(void *), GFP_KERNEL); entry 259 fs/squashfs/cache.c if (entry->data == NULL) { entry 265 fs/squashfs/cache.c entry->data[j] = kmalloc(PAGE_SIZE, GFP_KERNEL); entry 266 fs/squashfs/cache.c if (entry->data[j] == NULL) { entry 272 fs/squashfs/cache.c entry->actor = squashfs_page_actor_init(entry->data, entry 274 fs/squashfs/cache.c if (entry->actor == NULL) { entry 293 fs/squashfs/cache.c int squashfs_copy_data(void *buffer, struct squashfs_cache_entry *entry, entry 301 fs/squashfs/cache.c return min(length, entry->length - offset); entry 303 fs/squashfs/cache.c while (offset < entry->length) { entry 304 fs/squashfs/cache.c void *buff = entry->data[offset / PAGE_SIZE] entry 306 fs/squashfs/cache.c int bytes = min_t(int, entry->length - offset, entry 336 fs/squashfs/cache.c struct squashfs_cache_entry *entry; entry 344 fs/squashfs/cache.c entry = squashfs_cache_get(sb, msblk->block_cache, *block, 0); entry 345 fs/squashfs/cache.c if (entry->error) { entry 346 fs/squashfs/cache.c res = entry->error; entry 348 fs/squashfs/cache.c } else if (*offset >= entry->length) { entry 353 fs/squashfs/cache.c bytes = squashfs_copy_data(buffer, entry, *offset, length); entry 359 fs/squashfs/cache.c if (*offset == entry->length) { entry 360 fs/squashfs/cache.c *block = entry->next_index; entry 364 fs/squashfs/cache.c squashfs_cache_put(entry); entry 370 fs/squashfs/cache.c squashfs_cache_put(entry); entry 26 fs/squashfs/squashfs_fs_sb.h struct squashfs_cache_entry *entry; entry 44 fs/squashfs/symlink.c struct squashfs_cache_entry *entry; entry 71 fs/squashfs/symlink.c entry = squashfs_cache_get(sb, msblk->block_cache, block, 0); entry 72 fs/squashfs/symlink.c if (entry->error) { entry 76 fs/squashfs/symlink.c squashfs_cache_put(entry); entry 81 fs/squashfs/symlink.c copied = squashfs_copy_data(pageaddr + bytes, entry, offset, entry 86 fs/squashfs/symlink.c block = entry->next_index; entry 88 fs/squashfs/symlink.c squashfs_cache_put(entry); entry 45 fs/squashfs/xattr.c struct squashfs_xattr_entry entry; entry 50 fs/squashfs/xattr.c err = squashfs_read_metadata(sb, &entry, &start, &offset, entry 51 fs/squashfs/xattr.c sizeof(entry)); entry 55 fs/squashfs/xattr.c name_size = le16_to_cpu(entry.size); entry 56 fs/squashfs/xattr.c handler = squashfs_xattr_handler(le16_to_cpu(entry.type)); entry 123 fs/squashfs/xattr.c struct squashfs_xattr_entry entry; entry 127 fs/squashfs/xattr.c err = squashfs_read_metadata(sb, &entry, &start, &offset, entry 128 fs/squashfs/xattr.c sizeof(entry)); entry 132 fs/squashfs/xattr.c name_size = le16_to_cpu(entry.size); entry 133 fs/squashfs/xattr.c type = le16_to_cpu(entry.type); entry 429 fs/sysfs/group.c struct kernfs_node *entry; entry 445 fs/sysfs/group.c entry = kernfs_find_and_get(target_kobj->sd, target_name); entry 446 fs/sysfs/group.c if (!entry) { entry 451 fs/sysfs/group.c link = kernfs_create_link(kobj->sd, target_name, entry); entry 455 fs/sysfs/group.c kernfs_put(entry); entry 136 fs/udf/partition.c struct sparingEntry *entry = &st->mapEntry[i]; entry 137 fs/udf/partition.c u32 origLoc = le32_to_cpu(entry->origLocation); entry 141 fs/udf/partition.c return le32_to_cpu(entry->mappedLocation) + entry 188 fs/udf/partition.c struct sparingEntry *entry = &st->mapEntry[k]; entry 189 fs/udf/partition.c u32 origLoc = le32_to_cpu(entry->origLocation); entry 200 fs/udf/partition.c entry->origLocation = entry 210 fs/udf/partition.c entry->mappedLocation) + entry 218 fs/udf/partition.c entry->mappedLocation) + entry 229 fs/udf/partition.c struct sparingEntry *entry = &st->mapEntry[l]; entry 230 fs/udf/partition.c u32 origLoc = le32_to_cpu(entry->origLocation); entry 140 fs/userfaultfd.c list_del_init(&wq->entry); entry 564 fs/userfaultfd.c if (!list_empty_careful(&uwq.wq.entry)) { entry 570 fs/userfaultfd.c list_del(&uwq.wq.entry); entry 958 fs/userfaultfd.c wq = list_last_entry(&wqh->head, typeof(*wq), entry); entry 1089 fs/userfaultfd.c list_del(&uwq->wq.entry); entry 1111 fs/userfaultfd.c list_move(&uwq->wq.entry, &fork_event); entry 1158 fs/userfaultfd.c wq.entry); entry 1169 fs/userfaultfd.c list_del(&uwq->wq.entry); entry 1902 fs/userfaultfd.c list_for_each_entry(wq, &ctx->fault_pending_wqh.head, entry) { entry 1906 fs/userfaultfd.c list_for_each_entry(wq, &ctx->fault_wqh.head, entry) { entry 883 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 892 fs/xfs/libxfs/xfs_attr_leaf.c entry = xfs_attr3_leaf_entryp(leaf); entry 895 fs/xfs/libxfs/xfs_attr_leaf.c for (i = 0; i < leafhdr.count; entry++, i++) { entry 896 fs/xfs/libxfs/xfs_attr_leaf.c if (entry->flags & XFS_ATTR_INCOMPLETE) entry 898 fs/xfs/libxfs/xfs_attr_leaf.c if (!(entry->flags & XFS_ATTR_LOCAL)) entry 1001 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 1019 fs/xfs/libxfs/xfs_attr_leaf.c entry = xfs_attr3_leaf_entryp(leaf); entry 1051 fs/xfs/libxfs/xfs_attr_leaf.c for (i = 0; i < ichdr.count; entry++, i++) { entry 1052 fs/xfs/libxfs/xfs_attr_leaf.c if (entry->flags & XFS_ATTR_INCOMPLETE) entry 1054 fs/xfs/libxfs/xfs_attr_leaf.c if (!entry->nameidx) entry 1056 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT(entry->flags & XFS_ATTR_LOCAL); entry 1062 fs/xfs/libxfs/xfs_attr_leaf.c nargs.hashval = be32_to_cpu(entry->hashval); entry 1063 fs/xfs/libxfs/xfs_attr_leaf.c nargs.flags = XFS_ATTR_NSP_ONDISK_TO_ARGS(entry->flags); entry 1348 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 1364 fs/xfs/libxfs/xfs_attr_leaf.c entry = &xfs_attr3_leaf_entryp(leaf)[args->index]; entry 1368 fs/xfs/libxfs/xfs_attr_leaf.c memmove(entry + 1, entry, tmp); entry 1370 fs/xfs/libxfs/xfs_attr_leaf.c XFS_DA_LOGRANGE(leaf, entry, tmp + sizeof(*entry))); entry 1387 fs/xfs/libxfs/xfs_attr_leaf.c entry->nameidx = cpu_to_be16(ichdr->freemap[mapindex].base + entry 1389 fs/xfs/libxfs/xfs_attr_leaf.c entry->hashval = cpu_to_be32(args->hashval); entry 1390 fs/xfs/libxfs/xfs_attr_leaf.c entry->flags = tmp ? XFS_ATTR_LOCAL : 0; entry 1391 fs/xfs/libxfs/xfs_attr_leaf.c entry->flags |= XFS_ATTR_NSP_ARGS_TO_ONDISK(args->flags); entry 1393 fs/xfs/libxfs/xfs_attr_leaf.c entry->flags |= XFS_ATTR_INCOMPLETE; entry 1400 fs/xfs/libxfs/xfs_attr_leaf.c XFS_DA_LOGRANGE(leaf, entry, sizeof(*entry))); entry 1402 fs/xfs/libxfs/xfs_attr_leaf.c (be32_to_cpu(entry->hashval) >= be32_to_cpu((entry-1)->hashval))); entry 1404 fs/xfs/libxfs/xfs_attr_leaf.c (be32_to_cpu(entry->hashval) <= be32_to_cpu((entry+1)->hashval))); entry 1413 fs/xfs/libxfs/xfs_attr_leaf.c if (entry->flags & XFS_ATTR_LOCAL) { entry 1424 fs/xfs/libxfs/xfs_attr_leaf.c entry->flags |= XFS_ATTR_INCOMPLETE; entry 1439 fs/xfs/libxfs/xfs_attr_leaf.c if (be16_to_cpu(entry->nameidx) < ichdr->firstused) entry 1440 fs/xfs/libxfs/xfs_attr_leaf.c ichdr->firstused = be16_to_cpu(entry->nameidx); entry 1769 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 1784 fs/xfs/libxfs/xfs_attr_leaf.c half = (max + 1) * sizeof(*entry); entry 1789 fs/xfs/libxfs/xfs_attr_leaf.c entry = xfs_attr3_leaf_entryp(leaf1); entry 1790 fs/xfs/libxfs/xfs_attr_leaf.c for (count = index = 0; count < max; entry++, index++, count++) { entry 1797 fs/xfs/libxfs/xfs_attr_leaf.c tmp = totallen + sizeof(*entry) + entry 1811 fs/xfs/libxfs/xfs_attr_leaf.c entry = xfs_attr3_leaf_entryp(leaf1); entry 1818 fs/xfs/libxfs/xfs_attr_leaf.c tmp = totallen + sizeof(*entry) + xfs_attr_leaf_entsize(leaf1, entry 1831 fs/xfs/libxfs/xfs_attr_leaf.c totallen -= count * sizeof(*entry); entry 1833 fs/xfs/libxfs/xfs_attr_leaf.c totallen -= sizeof(*entry) + entry 1991 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 2007 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT(ichdr.firstused >= ichdr.count * sizeof(*entry) + entry 2010 fs/xfs/libxfs/xfs_attr_leaf.c entry = &xfs_attr3_leaf_entryp(leaf)[args->index]; entry 2012 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT(be16_to_cpu(entry->nameidx) >= ichdr.firstused); entry 2013 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT(be16_to_cpu(entry->nameidx) < args->geo->blksize); entry 2036 fs/xfs/libxfs/xfs_attr_leaf.c be16_to_cpu(entry->nameidx)) { entry 2039 fs/xfs/libxfs/xfs_attr_leaf.c (be16_to_cpu(entry->nameidx) + entsize)) { entry 2060 fs/xfs/libxfs/xfs_attr_leaf.c ichdr.freemap[after].base = be16_to_cpu(entry->nameidx); entry 2068 fs/xfs/libxfs/xfs_attr_leaf.c ichdr.freemap[smallest].base = be16_to_cpu(entry->nameidx); entry 2076 fs/xfs/libxfs/xfs_attr_leaf.c if (be16_to_cpu(entry->nameidx) == ichdr.firstused) entry 2091 fs/xfs/libxfs/xfs_attr_leaf.c memmove(entry, entry + 1, tmp); entry 2094 fs/xfs/libxfs/xfs_attr_leaf.c XFS_DA_LOGRANGE(leaf, entry, tmp + sizeof(xfs_attr_leaf_entry_t))); entry 2096 fs/xfs/libxfs/xfs_attr_leaf.c entry = &xfs_attr3_leaf_entryp(leaf)[ichdr.count]; entry 2097 fs/xfs/libxfs/xfs_attr_leaf.c memset(entry, 0, sizeof(xfs_attr_leaf_entry_t)); entry 2107 fs/xfs/libxfs/xfs_attr_leaf.c entry = xfs_attr3_leaf_entryp(leaf); entry 2108 fs/xfs/libxfs/xfs_attr_leaf.c for (i = ichdr.count - 1; i >= 0; entry++, i--) { entry 2109 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT(be16_to_cpu(entry->nameidx) >= ichdr.firstused); entry 2110 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT(be16_to_cpu(entry->nameidx) < args->geo->blksize); entry 2112 fs/xfs/libxfs/xfs_attr_leaf.c if (be16_to_cpu(entry->nameidx) < tmp) entry 2113 fs/xfs/libxfs/xfs_attr_leaf.c tmp = be16_to_cpu(entry->nameidx); entry 2148 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 2156 fs/xfs/libxfs/xfs_attr_leaf.c entry = xfs_attr3_leaf_entryp(drop_leaf); entry 2161 fs/xfs/libxfs/xfs_attr_leaf.c drop_blk->hashval = be32_to_cpu(entry[drophdr.count - 1].hashval); entry 2243 fs/xfs/libxfs/xfs_attr_leaf.c entry = xfs_attr3_leaf_entryp(save_leaf); entry 2244 fs/xfs/libxfs/xfs_attr_leaf.c save_blk->hashval = be32_to_cpu(entry[savehdr.count - 1].hashval); entry 2271 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 2292 fs/xfs/libxfs/xfs_attr_leaf.c for (entry = &entries[probe]; span > 4; entry = &entries[probe]) { entry 2294 fs/xfs/libxfs/xfs_attr_leaf.c if (be32_to_cpu(entry->hashval) < hashval) entry 2296 fs/xfs/libxfs/xfs_attr_leaf.c else if (be32_to_cpu(entry->hashval) > hashval) entry 2303 fs/xfs/libxfs/xfs_attr_leaf.c if (!(span <= 4 || be32_to_cpu(entry->hashval) == hashval)) entry 2310 fs/xfs/libxfs/xfs_attr_leaf.c while (probe > 0 && be32_to_cpu(entry->hashval) >= hashval) { entry 2311 fs/xfs/libxfs/xfs_attr_leaf.c entry--; entry 2315 fs/xfs/libxfs/xfs_attr_leaf.c be32_to_cpu(entry->hashval) < hashval) { entry 2316 fs/xfs/libxfs/xfs_attr_leaf.c entry++; entry 2319 fs/xfs/libxfs/xfs_attr_leaf.c if (probe == ichdr.count || be32_to_cpu(entry->hashval) != hashval) { entry 2327 fs/xfs/libxfs/xfs_attr_leaf.c for (; probe < ichdr.count && (be32_to_cpu(entry->hashval) == hashval); entry 2328 fs/xfs/libxfs/xfs_attr_leaf.c entry++, probe++) { entry 2337 fs/xfs/libxfs/xfs_attr_leaf.c (entry->flags & XFS_ATTR_INCOMPLETE)) { entry 2340 fs/xfs/libxfs/xfs_attr_leaf.c if (entry->flags & XFS_ATTR_LOCAL) { entry 2347 fs/xfs/libxfs/xfs_attr_leaf.c if (!xfs_attr_namesp_match(args->flags, entry->flags)) entry 2358 fs/xfs/libxfs/xfs_attr_leaf.c if (!xfs_attr_namesp_match(args->flags, entry->flags)) entry 2388 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 2397 fs/xfs/libxfs/xfs_attr_leaf.c entry = &xfs_attr3_leaf_entryp(leaf)[args->index]; entry 2398 fs/xfs/libxfs/xfs_attr_leaf.c if (entry->flags & XFS_ATTR_LOCAL) { entry 2649 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 2669 fs/xfs/libxfs/xfs_attr_leaf.c entry = &xfs_attr3_leaf_entryp(leaf)[args->index]; entry 2670 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT(entry->flags & XFS_ATTR_INCOMPLETE); entry 2677 fs/xfs/libxfs/xfs_attr_leaf.c if (entry->flags & XFS_ATTR_LOCAL) { entry 2686 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT(be32_to_cpu(entry->hashval) == args->hashval); entry 2691 fs/xfs/libxfs/xfs_attr_leaf.c entry->flags &= ~XFS_ATTR_INCOMPLETE; entry 2693 fs/xfs/libxfs/xfs_attr_leaf.c XFS_DA_LOGRANGE(leaf, entry, sizeof(*entry))); entry 2696 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT((entry->flags & XFS_ATTR_LOCAL) == 0); entry 2718 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_attr_leaf_entry *entry; entry 2741 fs/xfs/libxfs/xfs_attr_leaf.c entry = &xfs_attr3_leaf_entryp(leaf)[args->index]; entry 2743 fs/xfs/libxfs/xfs_attr_leaf.c ASSERT((entry->flags & XFS_ATTR_INCOMPLETE) == 0); entry 2744 fs/xfs/libxfs/xfs_attr_leaf.c entry->flags |= XFS_ATTR_INCOMPLETE; entry 2746 fs/xfs/libxfs/xfs_attr_leaf.c XFS_DA_LOGRANGE(leaf, entry, sizeof(*entry))); entry 2747 fs/xfs/libxfs/xfs_attr_leaf.c if ((entry->flags & XFS_ATTR_LOCAL) == 0) { entry 124 fs/xfs/scrub/dabtree.c struct xfs_da_node_entry *entry; entry 139 fs/xfs/scrub/dabtree.c entry = xchk_da_btree_entry(ds, level - 1, blks[level - 1].index); entry 140 fs/xfs/scrub/dabtree.c parent_hash = be32_to_cpu(entry->hashval); entry 110 fs/xfs/xfs_attr_inactive.c struct xfs_attr_leaf_entry *entry; entry 128 fs/xfs/xfs_attr_inactive.c entry = xfs_attr3_leaf_entryp(leaf); entry 129 fs/xfs/xfs_attr_inactive.c for (i = 0; i < ichdr.count; entry++, i++) { entry 130 fs/xfs/xfs_attr_inactive.c if (be16_to_cpu(entry->nameidx) && entry 131 fs/xfs/xfs_attr_inactive.c ((entry->flags & XFS_ATTR_LOCAL) == 0)) { entry 156 fs/xfs/xfs_attr_inactive.c entry = xfs_attr3_leaf_entryp(leaf); entry 157 fs/xfs/xfs_attr_inactive.c for (i = 0; i < ichdr.count; entry++, i++) { entry 158 fs/xfs/xfs_attr_inactive.c if (be16_to_cpu(entry->nameidx) && entry 159 fs/xfs/xfs_attr_inactive.c ((entry->flags & XFS_ATTR_LOCAL) == 0)) { entry 387 fs/xfs/xfs_attr_list.c struct xfs_attr_leaf_entry *entry; entry 404 fs/xfs/xfs_attr_list.c entry = &entries[0]; entry 405 fs/xfs/xfs_attr_list.c for (i = 0; i < ichdr.count; entry++, i++) { entry 406 fs/xfs/xfs_attr_list.c if (be32_to_cpu(entry->hashval) == cursor->hashval) { entry 412 fs/xfs/xfs_attr_list.c } else if (be32_to_cpu(entry->hashval) > entry 423 fs/xfs/xfs_attr_list.c entry = &entries[0]; entry 431 fs/xfs/xfs_attr_list.c for (; i < ichdr.count; entry++, i++) { entry 435 fs/xfs/xfs_attr_list.c if (be32_to_cpu(entry->hashval) != cursor->hashval) { entry 436 fs/xfs/xfs_attr_list.c cursor->hashval = be32_to_cpu(entry->hashval); entry 440 fs/xfs/xfs_attr_list.c if ((entry->flags & XFS_ATTR_INCOMPLETE) && entry 444 fs/xfs/xfs_attr_list.c if (entry->flags & XFS_ATTR_LOCAL) { entry 460 fs/xfs/xfs_attr_list.c context->put_listent(context, entry->flags, entry 209 include/acpi/acpi_bus.h struct proc_dir_entry *entry; entry 212 include/acpi/acpi_bus.h #define acpi_device_dir(d) ((d)->dir.entry) entry 89 include/acpi/actbl3.h u8 entry[1]; /* Real size = localities^2 */ entry 32 include/asm-generic/pgtable.h pte_t entry, int dirty); entry 39 include/asm-generic/pgtable.h pmd_t entry, int dirty); entry 42 include/asm-generic/pgtable.h pud_t entry, int dirty); entry 46 include/asm-generic/pgtable.h pmd_t entry, int dirty) entry 53 include/asm-generic/pgtable.h pud_t entry, int dirty) entry 559 include/asm-generic/vmlinux.lds.h *(.entry.text) \ entry 359 include/drm/drm_mm.h #define drm_mm_for_each_node(entry, mm) \ entry 360 include/drm/drm_mm.h list_for_each_entry(entry, drm_mm_nodes(mm), node_list) entry 371 include/drm/drm_mm.h #define drm_mm_for_each_node_safe(entry, next, mm) \ entry 372 include/drm/drm_mm.h list_for_each_entry_safe(entry, next, drm_mm_nodes(mm), node_list) entry 35 include/drm/drm_os_linux.h DECLARE_WAITQUEUE(entry, current); \ entry 37 include/drm/drm_os_linux.h add_wait_queue(&(queue), &entry); \ entry 54 include/drm/drm_os_linux.h remove_wait_queue(&(queue), &entry); \ entry 211 include/linux/acpi.h #define BAD_MADT_ENTRY(entry, end) ( \ entry 212 include/linux/acpi.h (!entry) || (unsigned long)entry + sizeof(*entry) > end || \ entry 213 include/linux/acpi.h ((struct acpi_subtable_header *)entry)->length < sizeof(*entry)) entry 142 include/linux/atmdev.h struct list_head entry; /* next address */ entry 114 include/linux/backlight.h struct list_head entry; entry 640 include/linux/ccp.h struct list_head entry; entry 675 include/linux/compat.h asmlinkage long compat_sys_kexec_load(compat_ulong_t entry, entry 856 include/linux/device.h const struct device_attribute *entry); entry 955 include/linux/efi.h efi_memory_desc_t entry[0]; entry 1489 include/linux/efi.h int efivar_entry_add(struct efivar_entry *entry, struct list_head *head); entry 1490 include/linux/efi.h int efivar_entry_remove(struct efivar_entry *entry); entry 1492 include/linux/efi.h int __efivar_entry_delete(struct efivar_entry *entry); entry 1493 include/linux/efi.h int efivar_entry_delete(struct efivar_entry *entry); entry 1495 include/linux/efi.h int efivar_entry_size(struct efivar_entry *entry, unsigned long *size); entry 1496 include/linux/efi.h int __efivar_entry_get(struct efivar_entry *entry, u32 *attributes, entry 1498 include/linux/efi.h int efivar_entry_get(struct efivar_entry *entry, u32 *attributes, entry 1500 include/linux/efi.h int efivar_entry_set(struct efivar_entry *entry, u32 attributes, entry 1502 include/linux/efi.h int efivar_entry_set_get_size(struct efivar_entry *entry, u32 attributes, entry 1787 include/linux/efi.h } entry[0]; entry 1791 include/linux/efi.h (count) * sizeof(((struct linux_efi_memreserve *)0)->entry[0])) entry 1794 include/linux/efi.h / sizeof(((struct linux_efi_memreserve *)0)->entry[0])) entry 23 include/linux/flat.h __be32 entry; /* Offset of first executable instruction entry 208 include/linux/hmm.h uint64_t entry) entry 210 include/linux/hmm.h if (entry == range->values[HMM_PFN_NONE]) entry 212 include/linux/hmm.h if (entry == range->values[HMM_PFN_ERROR]) entry 214 include/linux/hmm.h if (entry == range->values[HMM_PFN_SPECIAL]) entry 216 include/linux/hmm.h if (!(entry & range->flags[HMM_PFN_VALID])) entry 218 include/linux/hmm.h return pfn_to_page(entry >> range->pfn_shift); entry 449 include/linux/hugetlb.h static inline pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma, entry 452 include/linux/hugetlb.h return entry; entry 187 include/linux/idr.h #define idr_for_each_entry(idr, entry, id) \ entry 188 include/linux/idr.h for (id = 0; ((entry) = idr_get_next(idr, &(id))) != NULL; id += 1U) entry 201 include/linux/idr.h #define idr_for_each_entry_ul(idr, entry, tmp, id) \ entry 203 include/linux/idr.h tmp <= id && ((entry) = idr_get_next_ul(idr, &(id))) != NULL; \ entry 214 include/linux/idr.h #define idr_for_each_entry_continue(idr, entry, id) \ entry 215 include/linux/idr.h for ((entry) = idr_get_next((idr), &(id)); \ entry 216 include/linux/idr.h entry; \ entry 217 include/linux/idr.h ++id, (entry) = idr_get_next((idr), &(id))) entry 228 include/linux/idr.h #define idr_for_each_entry_continue_ul(idr, entry, tmp, id) \ entry 230 include/linux/idr.h tmp <= id && ((entry) = idr_get_next_ul(idr, &(id))) != NULL; \ entry 122 include/linux/init.h static inline initcall_t initcall_from_entry(initcall_entry_t *entry) entry 124 include/linux/init.h return offset_to_ptr(entry); entry 129 include/linux/init.h static inline initcall_t initcall_from_entry(initcall_entry_t *entry) entry 131 include/linux/init.h return *entry; entry 762 include/linux/irq.h extern int irq_set_msi_desc(unsigned int irq, struct msi_desc *entry); entry 764 include/linux/irq.h struct msi_desc *entry); entry 128 include/linux/jump_label.h static inline unsigned long jump_entry_code(const struct jump_entry *entry) entry 130 include/linux/jump_label.h return (unsigned long)&entry->code + entry->code; entry 133 include/linux/jump_label.h static inline unsigned long jump_entry_target(const struct jump_entry *entry) entry 135 include/linux/jump_label.h return (unsigned long)&entry->target + entry->target; entry 138 include/linux/jump_label.h static inline struct static_key *jump_entry_key(const struct jump_entry *entry) entry 140 include/linux/jump_label.h long offset = entry->key & ~3L; entry 142 include/linux/jump_label.h return (struct static_key *)((unsigned long)&entry->key + offset); entry 147 include/linux/jump_label.h static inline unsigned long jump_entry_code(const struct jump_entry *entry) entry 149 include/linux/jump_label.h return entry->code; entry 152 include/linux/jump_label.h static inline unsigned long jump_entry_target(const struct jump_entry *entry) entry 154 include/linux/jump_label.h return entry->target; entry 157 include/linux/jump_label.h static inline struct static_key *jump_entry_key(const struct jump_entry *entry) entry 159 include/linux/jump_label.h return (struct static_key *)((unsigned long)entry->key & ~3UL); entry 164 include/linux/jump_label.h static inline bool jump_entry_is_branch(const struct jump_entry *entry) entry 166 include/linux/jump_label.h return (unsigned long)entry->key & 1UL; entry 169 include/linux/jump_label.h static inline bool jump_entry_is_init(const struct jump_entry *entry) entry 171 include/linux/jump_label.h return (unsigned long)entry->key & 2UL; entry 174 include/linux/jump_label.h static inline void jump_entry_set_init(struct jump_entry *entry) entry 176 include/linux/jump_label.h entry->key |= 2; entry 214 include/linux/jump_label.h extern void arch_jump_label_transform(struct jump_entry *entry, entry 216 include/linux/jump_label.h extern void arch_jump_label_transform_static(struct jump_entry *entry, entry 218 include/linux/jump_label.h extern bool arch_jump_label_transform_queue(struct jump_entry *entry, entry 246 include/linux/kexec.h kimage_entry_t *entry; entry 380 include/linux/kexec.h static inline void *boot_phys_to_virt(unsigned long entry) entry 382 include/linux/kexec.h return phys_to_virt(boot_phys_to_phys(entry)); entry 67 include/linux/kobject.h struct list_head entry; entry 238 include/linux/kprobes.h extern int kprobe_add_ksym_blacklist(unsigned long entry); entry 36 include/linux/list.h extern bool __list_del_entry_valid(struct list_head *entry); entry 44 include/linux/list.h static inline bool __list_del_entry_valid(struct list_head *entry) entry 117 include/linux/list.h static inline void __list_del_clearprev(struct list_head *entry) entry 119 include/linux/list.h __list_del(entry->prev, entry->next); entry 120 include/linux/list.h entry->prev = NULL; entry 129 include/linux/list.h static inline void __list_del_entry(struct list_head *entry) entry 131 include/linux/list.h if (!__list_del_entry_valid(entry)) entry 134 include/linux/list.h __list_del(entry->prev, entry->next); entry 137 include/linux/list.h static inline void list_del(struct list_head *entry) entry 139 include/linux/list.h __list_del_entry(entry); entry 140 include/linux/list.h entry->next = LIST_POISON1; entry 141 include/linux/list.h entry->prev = LIST_POISON2; entry 188 include/linux/list.h static inline void list_del_init(struct list_head *entry) entry 190 include/linux/list.h __list_del_entry(entry); entry 191 include/linux/list.h INIT_LIST_HEAD(entry); entry 332 include/linux/list.h struct list_head *head, struct list_head *entry) entry 334 include/linux/list.h struct list_head *new_first = entry->next; entry 337 include/linux/list.h list->prev = entry; entry 338 include/linux/list.h entry->next = list; entry 358 include/linux/list.h struct list_head *head, struct list_head *entry) entry 363 include/linux/list.h (head->next != entry && head != entry)) entry 365 include/linux/list.h if (entry == head) entry 368 include/linux/list.h __list_cut_position(list, head, entry); entry 387 include/linux/list.h struct list_head *entry) entry 389 include/linux/list.h if (head->next == entry) { entry 395 include/linux/list.h list->prev = entry->prev; entry 397 include/linux/list.h head->next = entry; entry 398 include/linux/list.h entry->prev = head; entry 190 include/linux/lockdep.h struct list_head entry; entry 218 include/linux/lockdep.h struct hlist_node entry; entry 32 include/linux/mbcache.h void __mb_cache_entry_free(struct mb_cache_entry *entry); entry 34 include/linux/mbcache.h struct mb_cache_entry *entry) entry 36 include/linux/mbcache.h if (!atomic_dec_and_test(&entry->e_refcnt)) entry 38 include/linux/mbcache.h __mb_cache_entry_free(entry); entry 48 include/linux/mbcache.h struct mb_cache_entry *entry); entry 50 include/linux/mbcache.h struct mb_cache_entry *entry); entry 143 include/linux/migrate.h pmd_t *pmd, pmd_t entry, entry 149 include/linux/migrate.h pmd_t *pmd, pmd_t entry, entry 1446 include/linux/mlx4/device.h void mlx4_set_admin_guid(struct mlx4_dev *dev, __be64 guid, int entry, entry 1448 include/linux/mlx4/device.h __be64 mlx4_get_admin_guid(struct mlx4_dev *dev, int entry, int port); entry 1449 include/linux/mlx4/device.h void mlx4_set_random_admin_guid(struct mlx4_dev *dev, int entry, int port); entry 20 include/linux/msi.h void __get_cached_msi_msg(struct msi_desc *entry, struct msi_msg *msg); entry 97 include/linux/msi.h void (*write_msi_msg)(struct msi_desc *entry, void *data); entry 186 include/linux/msi.h void free_msi_entry(struct msi_desc *entry); entry 187 include/linux/msi.h void __pci_read_msi_msg(struct msi_desc *entry, struct msi_msg *msg); entry 188 include/linux/msi.h void __pci_write_msi_msg(struct msi_desc *entry, struct msi_msg *msg); entry 725 include/linux/mtd/nand.h unsigned int entry); entry 726 include/linux/mtd/nand.h int nanddev_bbt_set_block_status(struct nand_device *nand, unsigned int entry, entry 132 include/linux/netfilter.h nf_hook_entry_hookfn(const struct nf_hook_entry *entry, struct sk_buff *skb, entry 135 include/linux/netfilter.h return entry->hook(entry->priv, skb, state); entry 351 include/linux/netfilter.h int nf_reroute(struct sk_buff *skb, struct nf_queue_entry *entry); entry 20 include/linux/netfilter_arp/arp_tables.h struct arpt_entry entry; entry 25 include/linux/netfilter_arp/arp_tables.h struct arpt_entry entry; entry 37 include/linux/netfilter_arp/arp_tables.h .entry = ARPT_ENTRY_INIT(sizeof(struct arpt_standard)), \ entry 45 include/linux/netfilter_arp/arp_tables.h .entry = ARPT_ENTRY_INIT(sizeof(struct arpt_error)), \ entry 27 include/linux/netfilter_bridge/ebtables.h bool (*checkentry)(const char *table, const void *entry, entry 44 include/linux/netfilter_bridge/ebtables.h bool (*checkentry)(const char *table, const void *entry, entry 62 include/linux/netfilter_bridge/ebtables.h bool (*checkentry)(const char *table, const void *entry, entry 33 include/linux/netfilter_ipv4/ip_tables.h struct ipt_entry entry; entry 38 include/linux/netfilter_ipv4/ip_tables.h struct ipt_entry entry; entry 50 include/linux/netfilter_ipv4/ip_tables.h .entry = IPT_ENTRY_INIT(sizeof(struct ipt_standard)), \ entry 58 include/linux/netfilter_ipv4/ip_tables.h .entry = IPT_ENTRY_INIT(sizeof(struct ipt_error)), \ entry 59 include/linux/netfilter_ipv6.h int (*reroute)(struct sk_buff *skb, const struct nf_queue_entry *entry); entry 189 include/linux/oprofile.h void oprofile_write_reserve(struct op_entry *entry, entry 192 include/linux/oprofile.h int oprofile_add_data(struct op_entry *entry, unsigned long val); entry 193 include/linux/oprofile.h int oprofile_add_data64(struct op_entry *entry, u64 val); entry 194 include/linux/oprofile.h int oprofile_write_commit(struct op_entry *entry); entry 1429 include/linux/pci.h u16 entry; /* Driver uses to specify entry, OS writes */ entry 67 include/linux/perf_event.h struct perf_callchain_entry *entry; entry 255 include/linux/perf_event.h struct list_head entry; entry 488 include/linux/perf_event.h struct list_head entry; entry 1190 include/linux/perf_event.h extern void perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs); entry 1191 include/linux/perf_event.h extern void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs); entry 1205 include/linux/perf_event.h struct perf_callchain_entry *entry = ctx->entry; entry 1206 include/linux/perf_event.h entry->ip[entry->nr++] = ip; entry 1218 include/linux/perf_event.h struct perf_callchain_entry *entry = ctx->entry; entry 1219 include/linux/perf_event.h entry->ip[entry->nr++] = ip; entry 588 include/linux/pm.h struct list_head entry; entry 46 include/linux/pm_wakeup.h struct list_head entry; entry 229 include/linux/power/charger-manager.h struct list_head entry; entry 38 include/linux/qcom_scm.h extern int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus); entry 39 include/linux/qcom_scm.h extern int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus); entry 68 include/linux/qcom_scm.h int qcom_scm_set_cold_boot_addr(void *entry, const cpumask_t *cpus) entry 73 include/linux/qcom_scm.h int qcom_scm_set_warm_boot_addr(void *entry, const cpumask_t *cpus) entry 215 include/linux/radix-tree.h void __rcu **slot, void *entry); entry 217 include/linux/radix-tree.h const struct radix_tree_iter *, void __rcu **slot, void *entry); entry 219 include/linux/radix-tree.h void __rcu **slot, void *entry); entry 146 include/linux/rculist.h static inline void list_del_rcu(struct list_head *entry) entry 148 include/linux/rculist.h __list_del_entry(entry); entry 149 include/linux/rculist.h entry->prev = LIST_POISON2; entry 34 include/linux/resource_ext.h static inline void resource_list_add(struct resource_entry *entry, entry 37 include/linux/resource_ext.h list_add(&entry->node, head); entry 40 include/linux/resource_ext.h static inline void resource_list_add_tail(struct resource_entry *entry, entry 43 include/linux/resource_ext.h list_add_tail(&entry->node, head); entry 46 include/linux/resource_ext.h static inline void resource_list_del(struct resource_entry *entry) entry 48 include/linux/resource_ext.h list_del(&entry->node); entry 51 include/linux/resource_ext.h static inline void resource_list_free_entry(struct resource_entry *entry) entry 53 include/linux/resource_ext.h kfree(entry); entry 57 include/linux/resource_ext.h resource_list_destroy_entry(struct resource_entry *entry) entry 59 include/linux/resource_ext.h resource_list_del(entry); entry 60 include/linux/resource_ext.h resource_list_free_entry(entry); entry 63 include/linux/resource_ext.h #define resource_list_for_each_entry(entry, list) \ entry 64 include/linux/resource_ext.h list_for_each_entry((entry), (list), node) entry 66 include/linux/resource_ext.h #define resource_list_for_each_entry_safe(entry, tmp, list) \ entry 67 include/linux/resource_ext.h list_for_each_entry_safe((entry), (tmp), (list), node) entry 562 include/linux/sbitmap.h .entry = LIST_HEAD_INIT((name).wait.entry), \ entry 691 include/linux/spi/spi.h struct list_head entry; entry 405 include/linux/swap.h #define swap_address_space(entry) \ entry 406 include/linux/swap.h (&swapper_spaces[swp_type(entry)][swp_offset(entry) \ entry 412 include/linux/swap.h extern int __add_to_swap_cache(struct page *page, swp_entry_t entry); entry 413 include/linux/swap.h extern void __delete_from_swap_cache(struct page *, swp_entry_t entry); entry 417 include/linux/swap.h extern struct page *lookup_swap_cache(swp_entry_t entry, entry 426 include/linux/swap.h extern struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t flag, entry 428 include/linux/swap.h extern struct page *swapin_readahead(swp_entry_t entry, gfp_t flag, entry 450 include/linux/swap.h extern void put_swap_page(struct page *page, swp_entry_t entry); entry 465 include/linux/swap.h extern int __swap_count(swp_entry_t entry); entry 466 include/linux/swap.h extern int __swp_swapcount(swp_entry_t entry); entry 467 include/linux/swap.h extern int swp_swapcount(swp_entry_t entry); entry 469 include/linux/swap.h extern struct swap_info_struct *swp_swap_info(swp_entry_t entry); entry 475 include/linux/swap.h extern struct swap_info_struct *get_swap_device(swp_entry_t entry); entry 489 include/linux/swap.h static inline struct swap_info_struct *swp_swap_info(swp_entry_t entry) entry 494 include/linux/swap.h #define swap_address_space(entry) (NULL) entry 538 include/linux/swap.h static inline struct page *swap_cluster_readahead(swp_entry_t entry, entry 567 include/linux/swap.h static inline int add_to_swap_cache(struct page *page, swp_entry_t entry, entry 574 include/linux/swap.h swp_entry_t entry) entry 587 include/linux/swap.h static inline int __swap_count(swp_entry_t entry) entry 592 include/linux/swap.h static inline int __swp_swapcount(swp_entry_t entry) entry 597 include/linux/swap.h static inline int swp_swapcount(swp_entry_t entry) entry 612 include/linux/swap.h swp_entry_t entry; entry 613 include/linux/swap.h entry.val = 0; entry 614 include/linux/swap.h return entry; entry 620 include/linux/swap.h extern int split_swap_cluster(swp_entry_t entry); entry 622 include/linux/swap.h static inline int split_swap_cluster(swp_entry_t entry) entry 659 include/linux/swap.h extern void mem_cgroup_swapout(struct page *page, swp_entry_t entry); entry 660 include/linux/swap.h extern int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry); entry 661 include/linux/swap.h extern void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_pages); entry 665 include/linux/swap.h static inline void mem_cgroup_swapout(struct page *page, swp_entry_t entry) entry 670 include/linux/swap.h swp_entry_t entry) entry 675 include/linux/swap.h static inline void mem_cgroup_uncharge_swap(swp_entry_t entry, entry 27 include/linux/swap_slots.h int free_swap_slot(swp_entry_t entry); entry 41 include/linux/swapops.h static inline unsigned swp_type(swp_entry_t entry) entry 43 include/linux/swapops.h return (entry.val >> SWP_TYPE_SHIFT); entry 50 include/linux/swapops.h static inline pgoff_t swp_offset(swp_entry_t entry) entry 52 include/linux/swapops.h return entry.val & SWP_OFFSET_MASK; entry 79 include/linux/swapops.h static inline pte_t swp_entry_to_pte(swp_entry_t entry) entry 83 include/linux/swapops.h arch_entry = __swp_entry(swp_type(entry), swp_offset(entry)); entry 89 include/linux/swapops.h swp_entry_t entry; entry 91 include/linux/swapops.h entry.val = xa_to_value(arg); entry 92 include/linux/swapops.h return entry; entry 95 include/linux/swapops.h static inline void *swp_to_radix_entry(swp_entry_t entry) entry 97 include/linux/swapops.h return xa_mk_value(entry.val); entry 107 include/linux/swapops.h static inline bool is_device_private_entry(swp_entry_t entry) entry 109 include/linux/swapops.h int type = swp_type(entry); entry 113 include/linux/swapops.h static inline void make_device_private_entry_read(swp_entry_t *entry) entry 115 include/linux/swapops.h *entry = swp_entry(SWP_DEVICE_READ, swp_offset(*entry)); entry 118 include/linux/swapops.h static inline bool is_write_device_private_entry(swp_entry_t entry) entry 120 include/linux/swapops.h return unlikely(swp_type(entry) == SWP_DEVICE_WRITE); entry 123 include/linux/swapops.h static inline unsigned long device_private_entry_to_pfn(swp_entry_t entry) entry 125 include/linux/swapops.h return swp_offset(entry); entry 128 include/linux/swapops.h static inline struct page *device_private_entry_to_page(swp_entry_t entry) entry 130 include/linux/swapops.h return pfn_to_page(swp_offset(entry)); entry 138 include/linux/swapops.h static inline void make_device_private_entry_read(swp_entry_t *entry) entry 142 include/linux/swapops.h static inline bool is_device_private_entry(swp_entry_t entry) entry 147 include/linux/swapops.h static inline bool is_write_device_private_entry(swp_entry_t entry) entry 152 include/linux/swapops.h static inline unsigned long device_private_entry_to_pfn(swp_entry_t entry) entry 157 include/linux/swapops.h static inline struct page *device_private_entry_to_page(swp_entry_t entry) entry 172 include/linux/swapops.h static inline int is_migration_entry(swp_entry_t entry) entry 174 include/linux/swapops.h return unlikely(swp_type(entry) == SWP_MIGRATION_READ || entry 175 include/linux/swapops.h swp_type(entry) == SWP_MIGRATION_WRITE); entry 178 include/linux/swapops.h static inline int is_write_migration_entry(swp_entry_t entry) entry 180 include/linux/swapops.h return unlikely(swp_type(entry) == SWP_MIGRATION_WRITE); entry 183 include/linux/swapops.h static inline unsigned long migration_entry_to_pfn(swp_entry_t entry) entry 185 include/linux/swapops.h return swp_offset(entry); entry 188 include/linux/swapops.h static inline struct page *migration_entry_to_page(swp_entry_t entry) entry 190 include/linux/swapops.h struct page *p = pfn_to_page(swp_offset(entry)); entry 199 include/linux/swapops.h static inline void make_migration_entry_read(swp_entry_t *entry) entry 201 include/linux/swapops.h *entry = swp_entry(SWP_MIGRATION_READ, swp_offset(*entry)); entry 218 include/linux/swapops.h static inline unsigned long migration_entry_to_pfn(swp_entry_t entry) entry 223 include/linux/swapops.h static inline struct page *migration_entry_to_page(swp_entry_t entry) entry 235 include/linux/swapops.h static inline int is_write_migration_entry(swp_entry_t entry) entry 263 include/linux/swapops.h static inline pmd_t swp_entry_to_pmd(swp_entry_t entry) entry 267 include/linux/swapops.h arch_entry = __swp_entry(swp_type(entry), swp_offset(entry)); entry 295 include/linux/swapops.h static inline pmd_t swp_entry_to_pmd(swp_entry_t entry) entry 319 include/linux/swapops.h static inline int is_hwpoison_entry(swp_entry_t entry) entry 321 include/linux/swapops.h return swp_type(entry) == SWP_HWPOISON; entry 353 include/linux/swapops.h static inline int non_swap_entry(swp_entry_t entry) entry 355 include/linux/swapops.h return swp_type(entry) >= MAX_SWAPFILES; entry 358 include/linux/swapops.h static inline int non_swap_entry(swp_entry_t entry) entry 600 include/linux/syscalls.h asmlinkage long sys_kexec_load(unsigned long entry, unsigned long nr_segments, entry 16 include/linux/timer.h struct hlist_node entry; entry 76 include/linux/timer.h .entry = { .next = TIMER_ENTRY_STATIC }, \ entry 167 include/linux/timer.h return timer->entry.pprev != NULL; entry 74 include/linux/usb/composite.h struct list_head entry; entry 246 include/linux/usb/usbnet.h struct skb_data *entry = (struct skb_data *) skb->cb; entry 248 include/linux/usb/usbnet.h entry->packets = packets; entry 249 include/linux/usb/usbnet.h entry->length = bytes_delta; entry 80 include/linux/vme.h struct list_head *entry; entry 31 include/linux/wait.h struct list_head entry; entry 49 include/linux/wait.h .entry = { NULL, NULL } } entry 169 include/linux/wait.h list_add(&wq_entry->entry, &wq_head->head); entry 184 include/linux/wait.h list_add_tail(&wq_entry->entry, &wq_head->head); entry 197 include/linux/wait.h list_del(&wq_entry->entry); entry 1135 include/linux/wait.h .entry = LIST_HEAD_INIT((name).entry), \ entry 1144 include/linux/wait.h INIT_LIST_HEAD(&(wait)->entry); \ entry 44 include/linux/wait_bit.h .entry = \ entry 45 include/linux/wait_bit.h LIST_HEAD_INIT((name).wq_entry.entry), \ entry 104 include/linux/workqueue.h struct list_head entry; entry 186 include/linux/workqueue.h .entry = { &(n).entry, &(n).entry }, \ entry 236 include/linux/workqueue.h INIT_LIST_HEAD(&(_work)->entry); \ entry 244 include/linux/workqueue.h INIT_LIST_HEAD(&(_work)->entry); \ entry 65 include/linux/xarray.h static inline unsigned long xa_to_value(const void *entry) entry 67 include/linux/xarray.h return (unsigned long)entry >> 1; entry 77 include/linux/xarray.h static inline bool xa_is_value(const void *entry) entry 79 include/linux/xarray.h return (unsigned long)entry & 1; entry 110 include/linux/xarray.h static inline void *xa_untag_pointer(void *entry) entry 112 include/linux/xarray.h return (void *)((unsigned long)entry & ~3UL); entry 125 include/linux/xarray.h static inline unsigned int xa_pointer_tag(void *entry) entry 127 include/linux/xarray.h return (unsigned long)entry & 3UL; entry 155 include/linux/xarray.h static inline unsigned long xa_to_internal(const void *entry) entry 157 include/linux/xarray.h return (unsigned long)entry >> 2; entry 167 include/linux/xarray.h static inline bool xa_is_internal(const void *entry) entry 169 include/linux/xarray.h return ((unsigned long)entry & 3) == 2; entry 183 include/linux/xarray.h static inline bool xa_is_zero(const void *entry) entry 185 include/linux/xarray.h return unlikely(entry == XA_ZERO_ENTRY); entry 199 include/linux/xarray.h static inline bool xa_is_err(const void *entry) entry 201 include/linux/xarray.h return unlikely(xa_is_internal(entry) && entry 202 include/linux/xarray.h entry >= xa_mk_internal(-MAX_ERRNO)); entry 217 include/linux/xarray.h static inline int xa_err(void *entry) entry 220 include/linux/xarray.h if (xa_is_err(entry)) entry 221 include/linux/xarray.h return (long)entry >> 2; entry 348 include/linux/xarray.h void *xa_store(struct xarray *, unsigned long index, void *entry, gfp_t); entry 351 include/linux/xarray.h void *entry, gfp_t); entry 442 include/linux/xarray.h #define xa_for_each_start(xa, index, entry, start) \ entry 444 include/linux/xarray.h entry = xa_find(xa, &index, ULONG_MAX, XA_PRESENT); \ entry 445 include/linux/xarray.h entry; \ entry 446 include/linux/xarray.h entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT)) entry 469 include/linux/xarray.h #define xa_for_each(xa, index, entry) \ entry 470 include/linux/xarray.h xa_for_each_start(xa, index, entry, 0) entry 496 include/linux/xarray.h #define xa_for_each_marked(xa, index, entry, filter) \ entry 497 include/linux/xarray.h for (index = 0, entry = xa_find(xa, &index, ULONG_MAX, filter); \ entry 498 include/linux/xarray.h entry; entry = xa_find_after(xa, &index, ULONG_MAX, filter)) entry 520 include/linux/xarray.h void *__xa_store(struct xarray *, unsigned long index, void *entry, gfp_t); entry 522 include/linux/xarray.h void *entry, gfp_t); entry 524 include/linux/xarray.h void *entry, gfp_t); entry 525 include/linux/xarray.h int __must_check __xa_alloc(struct xarray *, u32 *id, void *entry, entry 527 include/linux/xarray.h int __must_check __xa_alloc_cyclic(struct xarray *, u32 *id, void *entry, entry 547 include/linux/xarray.h void *entry, gfp_t gfp) entry 552 include/linux/xarray.h curr = __xa_store(xa, index, entry, gfp); entry 573 include/linux/xarray.h void *entry, gfp_t gfp) entry 578 include/linux/xarray.h curr = __xa_store(xa, index, entry, gfp); entry 599 include/linux/xarray.h void *entry; entry 602 include/linux/xarray.h entry = __xa_erase(xa, index); entry 605 include/linux/xarray.h return entry; entry 623 include/linux/xarray.h void *entry; entry 626 include/linux/xarray.h entry = __xa_erase(xa, index); entry 629 include/linux/xarray.h return entry; entry 648 include/linux/xarray.h void *old, void *entry, gfp_t gfp) entry 653 include/linux/xarray.h curr = __xa_cmpxchg(xa, index, old, entry, gfp); entry 675 include/linux/xarray.h void *old, void *entry, gfp_t gfp) entry 680 include/linux/xarray.h curr = __xa_cmpxchg(xa, index, old, entry, gfp); entry 702 include/linux/xarray.h void *old, void *entry, gfp_t gfp) entry 707 include/linux/xarray.h curr = __xa_cmpxchg(xa, index, old, entry, gfp); entry 731 include/linux/xarray.h unsigned long index, void *entry, gfp_t gfp) entry 736 include/linux/xarray.h err = __xa_insert(xa, index, entry, gfp); entry 760 include/linux/xarray.h unsigned long index, void *entry, gfp_t gfp) entry 765 include/linux/xarray.h err = __xa_insert(xa, index, entry, gfp); entry 789 include/linux/xarray.h unsigned long index, void *entry, gfp_t gfp) entry 794 include/linux/xarray.h err = __xa_insert(xa, index, entry, gfp); entry 818 include/linux/xarray.h void *entry, struct xa_limit limit, gfp_t gfp) entry 823 include/linux/xarray.h err = __xa_alloc(xa, id, entry, limit, gfp); entry 847 include/linux/xarray.h void *entry, struct xa_limit limit, gfp_t gfp) entry 852 include/linux/xarray.h err = __xa_alloc(xa, id, entry, limit, gfp); entry 876 include/linux/xarray.h void *entry, struct xa_limit limit, gfp_t gfp) entry 881 include/linux/xarray.h err = __xa_alloc(xa, id, entry, limit, gfp); entry 908 include/linux/xarray.h static inline int xa_alloc_cyclic(struct xarray *xa, u32 *id, void *entry, entry 914 include/linux/xarray.h err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); entry 941 include/linux/xarray.h static inline int xa_alloc_cyclic_bh(struct xarray *xa, u32 *id, void *entry, entry 947 include/linux/xarray.h err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); entry 974 include/linux/xarray.h static inline int xa_alloc_cyclic_irq(struct xarray *xa, u32 *id, void *entry, entry 980 include/linux/xarray.h err = __xa_alloc_cyclic(xa, id, entry, limit, next, gfp); entry 1181 include/linux/xarray.h static inline struct xa_node *xa_to_node(const void *entry) entry 1183 include/linux/xarray.h return (struct xa_node *)((unsigned long)entry - 2); entry 1187 include/linux/xarray.h static inline bool xa_is_node(const void *entry) entry 1189 include/linux/xarray.h return xa_is_internal(entry) && (unsigned long)entry > 4096; entry 1199 include/linux/xarray.h static inline unsigned long xa_to_sibling(const void *entry) entry 1201 include/linux/xarray.h return xa_to_internal(entry); entry 1210 include/linux/xarray.h static inline bool xa_is_sibling(const void *entry) entry 1212 include/linux/xarray.h return IS_ENABLED(CONFIG_XARRAY_MULTI) && xa_is_internal(entry) && entry 1213 include/linux/xarray.h (entry < xa_mk_sibling(XA_CHUNK_SIZE - 1)); entry 1224 include/linux/xarray.h static inline bool xa_is_retry(const void *entry) entry 1226 include/linux/xarray.h return unlikely(entry == XA_RETRY_ENTRY); entry 1235 include/linux/xarray.h static inline bool xa_is_advanced(const void *entry) entry 1237 include/linux/xarray.h return xa_is_internal(entry) && (entry <= XA_RETRY_ENTRY); entry 1447 include/linux/xarray.h static inline bool xas_retry(struct xa_state *xas, const void *entry) entry 1449 include/linux/xarray.h if (xa_is_zero(entry)) entry 1451 include/linux/xarray.h if (!xa_is_retry(entry)) entry 1458 include/linux/xarray.h void *xas_store(struct xa_state *, void *entry); entry 1558 include/linux/xarray.h void *entry; entry 1569 include/linux/xarray.h entry = xa_entry(xas->xa, node, xas->xa_offset + 1); entry 1570 include/linux/xarray.h if (unlikely(xa_is_internal(entry))) entry 1574 include/linux/xarray.h } while (!entry); entry 1576 include/linux/xarray.h return entry; entry 1616 include/linux/xarray.h void *entry; entry 1628 include/linux/xarray.h entry = xa_entry(xas->xa, node, offset); entry 1629 include/linux/xarray.h if (!entry) entry 1631 include/linux/xarray.h return entry; entry 1655 include/linux/xarray.h #define xas_for_each(xas, entry, max) \ entry 1656 include/linux/xarray.h for (entry = xas_find(xas, max); entry; \ entry 1657 include/linux/xarray.h entry = xas_next_entry(xas, max)) entry 1673 include/linux/xarray.h #define xas_for_each_marked(xas, entry, max, mark) \ entry 1674 include/linux/xarray.h for (entry = xas_find_marked(xas, max, mark); entry; \ entry 1675 include/linux/xarray.h entry = xas_next_marked(xas, max, mark)) entry 1690 include/linux/xarray.h #define xas_for_each_conflict(xas, entry) \ entry 1691 include/linux/xarray.h while ((entry = xas_find_conflict(xas))) entry 349 include/media/v4l2-subdev.h struct v4l2_mbus_frame_desc_entry entry[V4L2_FRAME_DESC_ENTRY_MAX]; entry 24 include/net/atmclip.h struct atmarp_entry *entry; /* ATMARP table entry, NULL if IP addr. entry 529 include/net/bluetooth/mgmt.h } entry[0]; entry 821 include/net/devlink.h struct devlink_dpipe_entry *entry); entry 823 include/net/devlink.h void devlink_dpipe_entry_clear(struct devlink_dpipe_entry *entry); entry 392 include/net/flow_offload.h void flow_indr_add_block_cb(struct flow_indr_block_entry *entry); entry 394 include/net/flow_offload.h void flow_indr_del_block_cb(struct flow_indr_block_entry *entry); entry 323 include/net/ip.h c, stats_list[i].entry, \ entry 335 include/net/ip.h c, stats_list[i].entry); \ entry 28 include/net/netfilter/nf_queue.h int (*outfn)(struct nf_queue_entry *entry, entry 35 include/net/netfilter/nf_queue.h void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict); entry 37 include/net/netfilter/nf_queue.h void nf_queue_entry_get_refs(struct nf_queue_entry *entry); entry 38 include/net/netfilter/nf_queue.h void nf_queue_entry_release_refs(struct nf_queue_entry *entry); entry 31 include/net/snmp.h int entry; entry 36 include/net/snmp.h .entry = _entry, \ entry 41 include/net/snmp.h .entry = 0, \ entry 2514 include/rdma/ib_verbs.h struct rdma_restrack_entry *entry); entry 110 include/rdma/rdmavt_cq.h bool rvt_cq_enter(struct rvt_cq *cq, struct ib_wc *entry, bool solicited); entry 210 include/scsi/fc_encode.h struct fc_fdmi_attr_entry *entry; entry 245 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)hba_attrs->attr; entry 250 include/scsi/fc_encode.h &entry->type); entry 251 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 253 include/scsi/fc_encode.h (__be64 *)&entry->value[0]); entry 256 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 261 include/scsi/fc_encode.h &entry->type); entry 262 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 263 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 268 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 273 include/scsi/fc_encode.h &entry->type); entry 274 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 275 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 280 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 285 include/scsi/fc_encode.h &entry->type); entry 286 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 287 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 292 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 297 include/scsi/fc_encode.h &entry->type); entry 298 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 299 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 304 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 309 include/scsi/fc_encode.h &entry->type); entry 310 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 311 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 316 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 321 include/scsi/fc_encode.h &entry->type); entry 322 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 323 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 328 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 333 include/scsi/fc_encode.h &entry->type); entry 334 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 335 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 340 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 345 include/scsi/fc_encode.h &entry->type); entry 346 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 347 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 352 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 357 include/scsi/fc_encode.h &entry->type); entry 358 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 359 include/scsi/fc_encode.h snprintf((char *)&entry->value, entry 388 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)hba_attrs->attr; entry 394 include/scsi/fc_encode.h &entry->type); entry 395 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 396 include/scsi/fc_encode.h memcpy(&entry->value, fc_host_supported_fc4s(lport->host), entry 400 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 405 include/scsi/fc_encode.h &entry->type); entry 406 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 409 include/scsi/fc_encode.h &entry->value); entry 412 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 417 include/scsi/fc_encode.h &entry->type); entry 418 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 420 include/scsi/fc_encode.h &entry->value); entry 423 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 428 include/scsi/fc_encode.h &entry->type); entry 429 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 431 include/scsi/fc_encode.h &entry->value); entry 434 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 439 include/scsi/fc_encode.h &entry->type); entry 440 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 442 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 448 include/scsi/fc_encode.h entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + entry 453 include/scsi/fc_encode.h &entry->type); entry 454 include/scsi/fc_encode.h put_unaligned_be16(len, &entry->len); entry 456 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 461 include/scsi/fc_encode.h strncpy((char *)&entry->value, entry 30 include/sound/info.h void (*read)(struct snd_info_entry *entry, entry 32 include/sound/info.h void (*write)(struct snd_info_entry *entry, entry 37 include/sound/info.h int (*open)(struct snd_info_entry *entry, entry 39 include/sound/info.h int (*release)(struct snd_info_entry *entry, entry 41 include/sound/info.h ssize_t (*read)(struct snd_info_entry *entry, void *file_private_data, entry 44 include/sound/info.h ssize_t (*write)(struct snd_info_entry *entry, void *file_private_data, entry 47 include/sound/info.h loff_t (*llseek)(struct snd_info_entry *entry, entry 50 include/sound/info.h __poll_t (*poll)(struct snd_info_entry *entry, entry 53 include/sound/info.h int (*ioctl)(struct snd_info_entry *entry, void *file_private_data, entry 55 include/sound/info.h int (*mmap)(struct snd_info_entry *entry, void *file_private_data, entry 72 include/sound/info.h void (*private_free)(struct snd_info_entry *entry); entry 120 include/sound/info.h void snd_info_free_entry(struct snd_info_entry *entry); entry 121 include/sound/info.h int snd_info_store_text(struct snd_info_entry *entry); entry 122 include/sound/info.h int snd_info_restore_text(struct snd_info_entry *entry); entry 129 include/sound/info.h int snd_info_register(struct snd_info_entry *entry); entry 139 include/sound/info.h static inline void snd_info_set_text_ops(struct snd_info_entry *entry, entry 143 include/sound/info.h entry->private_data = private_data; entry 144 include/sound/info.h entry->c.text.read = read; entry 151 include/sound/info.h void (*write)(struct snd_info_entry *entry, entry 169 include/sound/info.h static inline void snd_info_free_entry(struct snd_info_entry *entry) { ; } entry 176 include/sound/info.h static inline int snd_info_register(struct snd_info_entry *entry) { return 0; } entry 180 include/sound/info.h static inline void snd_info_set_text_ops(struct snd_info_entry *entry __attribute__((unused)), entry 187 include/sound/info.h void (*write)(struct snd_info_entry *entry, entry 8 include/trace/bpf_probe.h #define __entry entry entry 14 include/trace/events/erofs.h #define show_dev_nid(entry) show_dev(entry->dev), entry->nid entry 11 include/trace/events/f2fs.h #define show_dev_ino(entry) show_dev(entry->dev), (unsigned long)entry->ino entry 20 include/trace/events/ib_mad.h struct trace_event_raw_ib_mad_send_template *entry); entry 49 include/trace/events/task.h memcpy(entry->oldcomm, task->comm, TASK_COMM_LEN); entry 50 include/trace/events/task.h strlcpy(entry->newcomm, comm, TASK_COMM_LEN); entry 8 include/trace/perf.h #define __entry entry entry 37 include/trace/perf.h struct trace_event_raw_##call *entry; \ entry 54 include/trace/perf.h __entry_size = ALIGN(__data_size + sizeof(*entry) + sizeof(u32),\ entry 58 include/trace/perf.h entry = perf_trace_buf_alloc(__entry_size, &__regs, &rctx); \ entry 59 include/trace/perf.h if (!entry) \ entry 68 include/trace/perf.h perf_trace_run_bpf_submit(entry, __entry_size, rctx, \ entry 375 include/trace/trace_events.h struct trace_entry *entry; \ entry 378 include/trace/trace_events.h entry = iter->ent; \ entry 380 include/trace/trace_events.h if (entry->type != event_##call.event.type) { \ entry 385 include/trace/trace_events.h field = (typeof(field))entry; \ entry 474 include/trace/trace_events.h #define __entry entry entry 495 include/trace/trace_events.h offsetof(typeof(*entry), __data); \ entry 533 include/trace/trace_events.h struct trace_event_raw_##call __maybe_unused *entry; \ entry 654 include/trace/trace_events.h #define __entry entry entry 704 include/trace/trace_events.h struct trace_event_raw_##call *entry; \ entry 712 include/trace/trace_events.h entry = trace_event_buffer_reserve(&fbuffer, trace_file, \ entry 713 include/trace/trace_events.h sizeof(*entry) + __data_size); \ entry 715 include/trace/trace_events.h if (!entry) \ entry 140 include/uapi/linux/coff.h char entry[4]; /* entry pt. */ entry 1590 include/uapi/linux/kvm.h __u16 entry; /* The index of entry in the MSI-X table */ entry 23 include/uapi/linux/mpls.h __be32 entry; entry 180 include/uapi/linux/netfilter/x_tables.h #define xt_ematch_foreach(pos, entry) \ entry 181 include/uapi/linux/netfilter/x_tables.h for ((pos) = (struct xt_entry_match *)entry->elems; \ entry 182 include/uapi/linux/netfilter/x_tables.h (pos) < (struct xt_entry_match *)((char *)(entry) + \ entry 183 include/uapi/linux/netfilter/x_tables.h (entry)->target_offset); \ entry 133 include/uapi/linux/netfilter_ipv6/ip6_tables.h struct ip6t_entry entry; entry 138 include/uapi/linux/netfilter_ipv6/ip6_tables.h struct ip6t_entry entry; entry 150 include/uapi/linux/netfilter_ipv6/ip6_tables.h .entry = IP6T_ENTRY_INIT(sizeof(struct ip6t_standard)), \ entry 158 include/uapi/linux/netfilter_ipv6/ip6_tables.h .entry = IP6T_ENTRY_INIT(sizeof(struct ip6t_error)), \ entry 91 include/uapi/linux/ptrace.h } entry; entry 1951 include/uapi/linux/videodev2.h struct v4l2_enc_idx_entry entry[V4L2_ENC_IDX_ENTRIES]; entry 59 include/uapi/xen/privcmd.h struct privcmd_mmap_entry __user *entry; entry 21 include/video/udlfb.h struct list_head entry; entry 67 include/xen/interface/io/pciif.h uint16_t entry; entry 373 include/xen/interface/xen-mca.h struct xen_mce entry[XEN_MCE_LOG_LEN]; entry 811 init/main.c struct blacklist_entry *entry; entry 818 init/main.c entry = memblock_alloc(sizeof(*entry), entry 820 init/main.c if (!entry) entry 822 init/main.c __func__, sizeof(*entry)); entry 823 init/main.c entry->buf = memblock_alloc(strlen(str_entry) + 1, entry 825 init/main.c if (!entry->buf) entry 828 init/main.c strcpy(entry->buf, str_entry); entry 829 init/main.c list_add(&entry->next, &blacklisted_initcalls); entry 838 init/main.c struct blacklist_entry *entry; entry 854 init/main.c list_for_each_entry(entry, &blacklisted_initcalls, next) { entry 855 init/main.c if (!strcmp(fn_name, entry->buf)) { entry 111 kernel/async.c struct async_entry *entry = entry 119 kernel/async.c (long long)entry->cookie, entry 120 kernel/async.c entry->func, task_pid_nr(current)); entry 123 kernel/async.c entry->func(entry->data, entry->cookie); entry 128 kernel/async.c (long long)entry->cookie, entry 129 kernel/async.c entry->func, entry 135 kernel/async.c list_del_init(&entry->domain_list); entry 136 kernel/async.c list_del_init(&entry->global_list); entry 139 kernel/async.c kfree(entry); entry 168 kernel/async.c struct async_entry *entry; entry 173 kernel/async.c entry = kzalloc(sizeof(struct async_entry), GFP_ATOMIC); entry 179 kernel/async.c if (!entry || atomic_read(&entry_count) > MAX_WORK) { entry 180 kernel/async.c kfree(entry); entry 189 kernel/async.c INIT_LIST_HEAD(&entry->domain_list); entry 190 kernel/async.c INIT_LIST_HEAD(&entry->global_list); entry 191 kernel/async.c INIT_WORK(&entry->work, async_run_entry_fn); entry 192 kernel/async.c entry->func = func; entry 193 kernel/async.c entry->data = data; entry 194 kernel/async.c entry->domain = domain; entry 199 kernel/async.c newcookie = entry->cookie = next_cookie++; entry 201 kernel/async.c list_add_tail(&entry->domain_list, &domain->pending); entry 203 kernel/async.c list_add_tail(&entry->global_list, &async_global_pending); entry 212 kernel/async.c queue_work_node(node, system_unbound_wq, &entry->work); entry 237 kernel/audit.h extern int audit_del_rule(struct audit_entry *entry); entry 148 kernel/audit_fsnotify.c struct audit_entry *entry = container_of(rule, struct audit_entry, rule); entry 151 kernel/audit_fsnotify.c audit_del_rule(entry); entry 547 kernel/audit_tree.c struct audit_entry *entry; entry 550 kernel/audit_tree.c entry = container_of(rule, struct audit_entry, rule); entry 556 kernel/audit_tree.c if (entry->rule.exe) entry 557 kernel/audit_tree.c audit_remove_mark(entry->rule.exe); entry 559 kernel/audit_tree.c list_del_rcu(&entry->list); entry 560 kernel/audit_tree.c list_del(&entry->rule.list); entry 561 kernel/audit_tree.c call_rcu(&entry->rcu, audit_free_rule_rcu); entry 64 kernel/audit_watch.c static void audit_watch_free_mark(struct fsnotify_mark *entry) entry 68 kernel/audit_watch.c parent = container_of(entry, struct audit_parent, mark); entry 91 kernel/audit_watch.c struct fsnotify_mark *entry; entry 93 kernel/audit_watch.c entry = fsnotify_find_mark(&inode->i_fsnotify_marks, audit_watch_group); entry 94 kernel/audit_watch.c if (entry) entry 95 kernel/audit_watch.c parent = container_of(entry, struct audit_parent, mark); entry 106 kernel/auditfilter.c struct audit_entry *entry; entry 109 kernel/auditfilter.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 110 kernel/auditfilter.c if (unlikely(!entry)) entry 115 kernel/auditfilter.c kfree(entry); entry 118 kernel/auditfilter.c entry->rule.fields = fields; entry 120 kernel/auditfilter.c return entry; entry 208 kernel/auditfilter.c static int audit_match_signal(struct audit_entry *entry) entry 210 kernel/auditfilter.c struct audit_field *arch = entry->rule.arch_f; entry 216 kernel/auditfilter.c entry->rule.mask) && entry 218 kernel/auditfilter.c entry->rule.mask)); entry 224 kernel/auditfilter.c entry->rule.mask)); entry 227 kernel/auditfilter.c entry->rule.mask)); entry 238 kernel/auditfilter.c struct audit_entry *entry; entry 268 kernel/auditfilter.c entry = audit_init_entry(rule->field_count); entry 269 kernel/auditfilter.c if (!entry) entry 272 kernel/auditfilter.c entry->rule.flags = rule->flags & AUDIT_FILTER_PREPEND; entry 273 kernel/auditfilter.c entry->rule.listnr = listnr; entry 274 kernel/auditfilter.c entry->rule.action = rule->action; entry 275 kernel/auditfilter.c entry->rule.field_count = rule->field_count; entry 278 kernel/auditfilter.c entry->rule.mask[i] = rule->mask[i]; entry 282 kernel/auditfilter.c __u32 *p = &entry->rule.mask[AUDIT_WORD(bit)]; entry 292 kernel/auditfilter.c entry->rule.mask[j] |= class[j]; entry 296 kernel/auditfilter.c return entry; entry 323 kernel/auditfilter.c static int audit_field_valid(struct audit_entry *entry, struct audit_field *f) entry 327 kernel/auditfilter.c if (entry->rule.listnr != AUDIT_FILTER_EXCLUDE && entry 328 kernel/auditfilter.c entry->rule.listnr != AUDIT_FILTER_USER) entry 332 kernel/auditfilter.c if (entry->rule.listnr != AUDIT_FILTER_FS) entry 337 kernel/auditfilter.c switch (entry->rule.listnr) { entry 445 kernel/auditfilter.c struct audit_entry *entry; entry 452 kernel/auditfilter.c entry = audit_to_entry_common(data); entry 453 kernel/auditfilter.c if (IS_ERR(entry)) entry 458 kernel/auditfilter.c struct audit_field *f = &entry->rule.fields[i]; entry 474 kernel/auditfilter.c entry->rule.pflags |= AUDIT_LOGINUID_LEGACY; entry 477 kernel/auditfilter.c err = audit_field_valid(entry, f); entry 504 kernel/auditfilter.c entry->rule.arch_f = f; entry 521 kernel/auditfilter.c entry->rule.buflen += f_val; entry 540 kernel/auditfilter.c err = audit_to_watch(&entry->rule, str, f_val, f->op); entry 545 kernel/auditfilter.c entry->rule.buflen += f_val; entry 553 kernel/auditfilter.c err = audit_make_tree(&entry->rule, str, f->op); entry 557 kernel/auditfilter.c entry->rule.buflen += f_val; entry 561 kernel/auditfilter.c err = audit_to_inode(&entry->rule, f); entry 566 kernel/auditfilter.c if (entry->rule.filterkey || f_val > AUDIT_MAX_KEY_LEN) entry 573 kernel/auditfilter.c entry->rule.buflen += f_val; entry 574 kernel/auditfilter.c entry->rule.filterkey = str; entry 577 kernel/auditfilter.c if (entry->rule.exe || f_val > PATH_MAX) entry 584 kernel/auditfilter.c audit_mark = audit_alloc_mark(&entry->rule, str, f_val); entry 590 kernel/auditfilter.c entry->rule.buflen += f_val; entry 591 kernel/auditfilter.c entry->rule.exe = audit_mark; entry 599 kernel/auditfilter.c if (entry->rule.inode_f && entry->rule.inode_f->op == Audit_not_equal) entry 600 kernel/auditfilter.c entry->rule.inode_f = NULL; entry 603 kernel/auditfilter.c return entry; entry 606 kernel/auditfilter.c if (entry->rule.tree) entry 607 kernel/auditfilter.c audit_put_tree(entry->rule.tree); /* that's the temporary one */ entry 608 kernel/auditfilter.c if (entry->rule.exe) entry 609 kernel/auditfilter.c audit_remove_mark(entry->rule.exe); /* that's the template one */ entry 610 kernel/auditfilter.c audit_free_rule(entry); entry 814 kernel/auditfilter.c struct audit_entry *entry; entry 819 kernel/auditfilter.c entry = audit_init_entry(fcount); entry 820 kernel/auditfilter.c if (unlikely(!entry)) entry 823 kernel/auditfilter.c new = &entry->rule; entry 876 kernel/auditfilter.c audit_free_rule(entry); entry 886 kernel/auditfilter.c return entry; entry 891 kernel/auditfilter.c static struct audit_entry *audit_find_rule(struct audit_entry *entry, entry 898 kernel/auditfilter.c if (entry->rule.inode_f) { entry 899 kernel/auditfilter.c h = audit_hash_ino(entry->rule.inode_f->val); entry 901 kernel/auditfilter.c } else if (entry->rule.watch) { entry 906 kernel/auditfilter.c if (!audit_compare_rule(&entry->rule, &e->rule)) { entry 913 kernel/auditfilter.c *p = list = &audit_filter_list[entry->rule.listnr]; entry 917 kernel/auditfilter.c if (!audit_compare_rule(&entry->rule, &e->rule)) { entry 930 kernel/auditfilter.c static inline int audit_add_rule(struct audit_entry *entry) entry 933 kernel/auditfilter.c struct audit_watch *watch = entry->rule.watch; entry 934 kernel/auditfilter.c struct audit_tree *tree = entry->rule.tree; entry 941 kernel/auditfilter.c switch(entry->rule.listnr) { entry 950 kernel/auditfilter.c e = audit_find_rule(entry, &list); entry 962 kernel/auditfilter.c err = audit_add_watch(&entry->rule, &list); entry 975 kernel/auditfilter.c err = audit_add_tree_rule(&entry->rule); entry 982 kernel/auditfilter.c entry->rule.prio = ~0ULL; entry 983 kernel/auditfilter.c if (entry->rule.listnr == AUDIT_FILTER_EXIT) { entry 984 kernel/auditfilter.c if (entry->rule.flags & AUDIT_FILTER_PREPEND) entry 985 kernel/auditfilter.c entry->rule.prio = ++prio_high; entry 987 kernel/auditfilter.c entry->rule.prio = --prio_low; entry 990 kernel/auditfilter.c if (entry->rule.flags & AUDIT_FILTER_PREPEND) { entry 991 kernel/auditfilter.c list_add(&entry->rule.list, entry 992 kernel/auditfilter.c &audit_rules_list[entry->rule.listnr]); entry 993 kernel/auditfilter.c list_add_rcu(&entry->list, list); entry 994 kernel/auditfilter.c entry->rule.flags &= ~AUDIT_FILTER_PREPEND; entry 996 kernel/auditfilter.c list_add_tail(&entry->rule.list, entry 997 kernel/auditfilter.c &audit_rules_list[entry->rule.listnr]); entry 998 kernel/auditfilter.c list_add_tail_rcu(&entry->list, list); entry 1004 kernel/auditfilter.c if (!audit_match_signal(entry)) entry 1013 kernel/auditfilter.c int audit_del_rule(struct audit_entry *entry) entry 1016 kernel/auditfilter.c struct audit_tree *tree = entry->rule.tree; entry 1023 kernel/auditfilter.c switch(entry->rule.listnr) { entry 1032 kernel/auditfilter.c e = audit_find_rule(entry, &list); entry 1051 kernel/auditfilter.c if (!audit_match_signal(entry)) entry 1126 kernel/auditfilter.c struct audit_entry *entry; entry 1130 kernel/auditfilter.c entry = audit_data_to_entry(data, datasz); entry 1131 kernel/auditfilter.c if (IS_ERR(entry)) entry 1132 kernel/auditfilter.c return PTR_ERR(entry); entry 1133 kernel/auditfilter.c err = audit_add_rule(entry); entry 1134 kernel/auditfilter.c audit_log_rule_change("add_rule", &entry->rule, !err); entry 1137 kernel/auditfilter.c entry = audit_data_to_entry(data, datasz); entry 1138 kernel/auditfilter.c if (IS_ERR(entry)) entry 1139 kernel/auditfilter.c return PTR_ERR(entry); entry 1140 kernel/auditfilter.c err = audit_del_rule(entry); entry 1141 kernel/auditfilter.c audit_log_rule_change("remove_rule", &entry->rule, !err); entry 1149 kernel/auditfilter.c if (entry->rule.exe) entry 1150 kernel/auditfilter.c audit_remove_mark(entry->rule.exe); entry 1151 kernel/auditfilter.c audit_free_rule(entry); entry 1395 kernel/auditfilter.c struct audit_entry *entry = container_of(r, struct audit_entry, rule); entry 1403 kernel/auditfilter.c if (entry->rule.exe) entry 1404 kernel/auditfilter.c audit_remove_mark(entry->rule.exe); entry 1412 kernel/auditfilter.c list_del_rcu(&entry->list); entry 1417 kernel/auditfilter.c list_replace_rcu(&entry->list, &nentry->list); entry 1420 kernel/auditfilter.c call_rcu(&entry->rcu, audit_free_rule_rcu); entry 39 kernel/bpf/stackmap.c static void do_up_read(struct irq_work *entry) entry 43 kernel/bpf/stackmap.c work = container_of(entry, struct stack_map_irq_work, irq_work); entry 58 kernel/configs.c struct proc_dir_entry *entry; entry 61 kernel/configs.c entry = proc_create("config.gz", S_IFREG | S_IRUGO, NULL, entry 63 kernel/configs.c if (!entry) entry 66 kernel/configs.c proc_set_size(entry, &kernel_config_data_end - &kernel_config_data); entry 163 kernel/dma/debug.c static inline void dump_entry_trace(struct dma_debug_entry *entry) entry 166 kernel/dma/debug.c if (entry) { entry 168 kernel/dma/debug.c stack_trace_print(entry->stack_entries, entry->stack_len, 0); entry 214 kernel/dma/debug.c #define err_printk(dev, entry, format, arg...) do { \ entry 221 kernel/dma/debug.c dump_entry_trace(entry); \ entry 233 kernel/dma/debug.c static int hash_fn(struct dma_debug_entry *entry) entry 239 kernel/dma/debug.c return (entry->dev_addr >> HASH_FN_SHIFT) & HASH_FN_MASK; entry 245 kernel/dma/debug.c static struct hash_bucket *get_hash_bucket(struct dma_debug_entry *entry, entry 249 kernel/dma/debug.c int idx = hash_fn(entry); entry 295 kernel/dma/debug.c struct dma_debug_entry *entry, *ret = NULL; entry 298 kernel/dma/debug.c list_for_each_entry(entry, &bucket->list, list) { entry 299 kernel/dma/debug.c if (!match(ref, entry)) entry 314 kernel/dma/debug.c entry->size == ref->size ? ++match_lvl : 0; entry 315 kernel/dma/debug.c entry->type == ref->type ? ++match_lvl : 0; entry 316 kernel/dma/debug.c entry->direction == ref->direction ? ++match_lvl : 0; entry 317 kernel/dma/debug.c entry->sg_call_ents == ref->sg_call_ents ? ++match_lvl : 0; entry 321 kernel/dma/debug.c return entry; entry 328 kernel/dma/debug.c ret = entry; entry 353 kernel/dma/debug.c struct dma_debug_entry *entry, index = *ref; entry 357 kernel/dma/debug.c entry = __hash_bucket_find(*bucket, ref, containing_match); entry 359 kernel/dma/debug.c if (entry) entry 360 kernel/dma/debug.c return entry; entry 378 kernel/dma/debug.c struct dma_debug_entry *entry) entry 380 kernel/dma/debug.c list_add_tail(&entry->list, &bucket->list); entry 386 kernel/dma/debug.c static void hash_bucket_del(struct dma_debug_entry *entry) entry 388 kernel/dma/debug.c list_del(&entry->list); entry 391 kernel/dma/debug.c static unsigned long long phys_addr(struct dma_debug_entry *entry) entry 393 kernel/dma/debug.c if (entry->type == dma_debug_resource) entry 394 kernel/dma/debug.c return __pfn_to_phys(entry->pfn) + entry->offset; entry 396 kernel/dma/debug.c return page_to_phys(pfn_to_page(entry->pfn)) + entry->offset; entry 408 kernel/dma/debug.c struct dma_debug_entry *entry; entry 413 kernel/dma/debug.c list_for_each_entry(entry, &bucket->list, list) { entry 414 kernel/dma/debug.c if (!dev || dev == entry->dev) { entry 415 kernel/dma/debug.c dev_info(entry->dev, entry 417 kernel/dma/debug.c type2name[entry->type], idx, entry 418 kernel/dma/debug.c phys_addr(entry), entry->pfn, entry 419 kernel/dma/debug.c entry->dev_addr, entry->size, entry 420 kernel/dma/debug.c dir2name[entry->direction], entry 421 kernel/dma/debug.c maperr2str[entry->map_err_type]); entry 459 kernel/dma/debug.c static phys_addr_t to_cacheline_number(struct dma_debug_entry *entry) entry 461 kernel/dma/debug.c return (entry->pfn << CACHELINE_PER_PAGE_SHIFT) + entry 462 kernel/dma/debug.c (entry->offset >> L1_CACHE_SHIFT); entry 515 kernel/dma/debug.c static int active_cacheline_insert(struct dma_debug_entry *entry) entry 517 kernel/dma/debug.c phys_addr_t cln = to_cacheline_number(entry); entry 525 kernel/dma/debug.c if (entry->direction == DMA_TO_DEVICE) entry 529 kernel/dma/debug.c rc = radix_tree_insert(&dma_active_cacheline, cln, entry); entry 537 kernel/dma/debug.c static void active_cacheline_remove(struct dma_debug_entry *entry) entry 539 kernel/dma/debug.c phys_addr_t cln = to_cacheline_number(entry); entry 543 kernel/dma/debug.c if (entry->direction == DMA_TO_DEVICE) entry 567 kernel/dma/debug.c struct dma_debug_entry *entry = NULL; entry 587 kernel/dma/debug.c entry = ents[i]; entry 594 kernel/dma/debug.c if (!entry) entry 597 kernel/dma/debug.c cln = to_cacheline_number(entry); entry 598 kernel/dma/debug.c err_printk(entry->dev, entry, entry 607 kernel/dma/debug.c static void add_dma_entry(struct dma_debug_entry *entry) entry 613 kernel/dma/debug.c bucket = get_hash_bucket(entry, &flags); entry 614 kernel/dma/debug.c hash_bucket_add(bucket, entry); entry 617 kernel/dma/debug.c rc = active_cacheline_insert(entry); entry 630 kernel/dma/debug.c struct dma_debug_entry *entry; entry 633 kernel/dma/debug.c entry = (void *)get_zeroed_page(gfp); entry 634 kernel/dma/debug.c if (!entry) entry 638 kernel/dma/debug.c list_add_tail(&entry[i].list, &free_entries); entry 648 kernel/dma/debug.c struct dma_debug_entry *entry; entry 650 kernel/dma/debug.c entry = list_entry(free_entries.next, struct dma_debug_entry, list); entry 651 kernel/dma/debug.c list_del(&entry->list); entry 652 kernel/dma/debug.c memset(entry, 0, sizeof(*entry)); entry 658 kernel/dma/debug.c return entry; entry 680 kernel/dma/debug.c struct dma_debug_entry *entry; entry 694 kernel/dma/debug.c entry = __dma_entry_alloc(); entry 699 kernel/dma/debug.c entry->stack_len = stack_trace_save(entry->stack_entries, entry 700 kernel/dma/debug.c ARRAY_SIZE(entry->stack_entries), entry 703 kernel/dma/debug.c return entry; entry 706 kernel/dma/debug.c static void dma_entry_free(struct dma_debug_entry *entry) entry 710 kernel/dma/debug.c active_cacheline_remove(entry); entry 717 kernel/dma/debug.c list_add(&entry->list, &free_entries); entry 828 kernel/dma/debug.c struct dma_debug_entry *entry; entry 832 kernel/dma/debug.c list_for_each_entry(entry, &bucket->list, list) { entry 835 kernel/dma/debug.c dev_name(entry->dev), entry 836 kernel/dma/debug.c dev_driver_string(entry->dev), entry 837 kernel/dma/debug.c type2name[entry->type], idx, entry 838 kernel/dma/debug.c phys_addr(entry), entry->pfn, entry 839 kernel/dma/debug.c entry->dev_addr, entry->size, entry 840 kernel/dma/debug.c dir2name[entry->direction], entry 841 kernel/dma/debug.c maperr2str[entry->map_err_type]); entry 866 kernel/dma/debug.c struct dma_debug_entry *entry; entry 872 kernel/dma/debug.c list_for_each_entry(entry, &dma_entry_hash[i].list, list) { entry 873 kernel/dma/debug.c if (entry->dev == dev) { entry 875 kernel/dma/debug.c *out_entry = entry; entry 887 kernel/dma/debug.c struct dma_debug_entry *uninitialized_var(entry); entry 895 kernel/dma/debug.c count = device_dma_allocations(dev, &entry); entry 898 kernel/dma/debug.c err_printk(dev, entry, "device driver has pending " entry 904 kernel/dma/debug.c count, entry->dev_addr, entry->size, entry 905 kernel/dma/debug.c dir2name[entry->direction], type2name[entry->type]); entry 999 kernel/dma/debug.c struct dma_debug_entry *entry; entry 1004 kernel/dma/debug.c entry = bucket_find_exact(bucket, ref); entry 1006 kernel/dma/debug.c if (!entry) { entry 1024 kernel/dma/debug.c if (ref->size != entry->size) { entry 1025 kernel/dma/debug.c err_printk(ref->dev, entry, "device driver frees " entry 1029 kernel/dma/debug.c ref->dev_addr, entry->size, ref->size); entry 1032 kernel/dma/debug.c if (ref->type != entry->type) { entry 1033 kernel/dma/debug.c err_printk(ref->dev, entry, "device driver frees " entry 1038 kernel/dma/debug.c type2name[entry->type], type2name[ref->type]); entry 1039 kernel/dma/debug.c } else if ((entry->type == dma_debug_coherent) && entry 1040 kernel/dma/debug.c (phys_addr(ref) != phys_addr(entry))) { entry 1041 kernel/dma/debug.c err_printk(ref->dev, entry, "device driver frees " entry 1047 kernel/dma/debug.c phys_addr(entry), entry 1052 kernel/dma/debug.c ref->sg_call_ents != entry->sg_call_ents) { entry 1053 kernel/dma/debug.c err_printk(ref->dev, entry, "device driver frees " entry 1056 kernel/dma/debug.c entry->sg_call_ents, ref->sg_call_ents); entry 1063 kernel/dma/debug.c if (ref->direction != entry->direction) { entry 1064 kernel/dma/debug.c err_printk(ref->dev, entry, "device driver frees " entry 1069 kernel/dma/debug.c dir2name[entry->direction], entry 1078 kernel/dma/debug.c if (entry->map_err_type == MAP_ERR_NOT_CHECKED) { entry 1079 kernel/dma/debug.c err_printk(ref->dev, entry, entry 1084 kernel/dma/debug.c type2name[entry->type]); entry 1087 kernel/dma/debug.c hash_bucket_del(entry); entry 1088 kernel/dma/debug.c dma_entry_free(entry); entry 1142 kernel/dma/debug.c struct dma_debug_entry *entry; entry 1148 kernel/dma/debug.c entry = bucket_find_contain(&bucket, ref, &flags); entry 1150 kernel/dma/debug.c if (!entry) { entry 1158 kernel/dma/debug.c if (ref->size > entry->size) { entry 1159 kernel/dma/debug.c err_printk(dev, entry, "device driver syncs" entry 1164 kernel/dma/debug.c entry->dev_addr, entry->size, entry 1168 kernel/dma/debug.c if (entry->direction == DMA_BIDIRECTIONAL) entry 1171 kernel/dma/debug.c if (ref->direction != entry->direction) { entry 1172 kernel/dma/debug.c err_printk(dev, entry, "device driver syncs " entry 1176 kernel/dma/debug.c (unsigned long long)ref->dev_addr, entry->size, entry 1177 kernel/dma/debug.c dir2name[entry->direction], entry 1181 kernel/dma/debug.c if (to_cpu && !(entry->direction == DMA_FROM_DEVICE) && entry 1183 kernel/dma/debug.c err_printk(dev, entry, "device driver syncs " entry 1187 kernel/dma/debug.c (unsigned long long)ref->dev_addr, entry->size, entry 1188 kernel/dma/debug.c dir2name[entry->direction], entry 1191 kernel/dma/debug.c if (!to_cpu && !(entry->direction == DMA_TO_DEVICE) && entry 1193 kernel/dma/debug.c err_printk(dev, entry, "device driver syncs " entry 1197 kernel/dma/debug.c (unsigned long long)ref->dev_addr, entry->size, entry 1198 kernel/dma/debug.c dir2name[entry->direction], entry 1202 kernel/dma/debug.c ref->sg_call_ents != entry->sg_call_ents) { entry 1203 kernel/dma/debug.c err_printk(ref->dev, entry, "device driver syncs " entry 1206 kernel/dma/debug.c entry->sg_call_ents, ref->sg_call_ents); entry 1258 kernel/dma/debug.c struct dma_debug_entry *entry; entry 1266 kernel/dma/debug.c entry = dma_entry_alloc(); entry 1267 kernel/dma/debug.c if (!entry) entry 1270 kernel/dma/debug.c entry->dev = dev; entry 1271 kernel/dma/debug.c entry->type = dma_debug_single; entry 1272 kernel/dma/debug.c entry->pfn = page_to_pfn(page); entry 1273 kernel/dma/debug.c entry->offset = offset, entry 1274 kernel/dma/debug.c entry->dev_addr = dma_addr; entry 1275 kernel/dma/debug.c entry->size = size; entry 1276 kernel/dma/debug.c entry->direction = direction; entry 1277 kernel/dma/debug.c entry->map_err_type = MAP_ERR_NOT_CHECKED; entry 1287 kernel/dma/debug.c add_dma_entry(entry); entry 1294 kernel/dma/debug.c struct dma_debug_entry *entry; entry 1305 kernel/dma/debug.c list_for_each_entry(entry, &bucket->list, list) { entry 1306 kernel/dma/debug.c if (!exact_match(&ref, entry)) entry 1319 kernel/dma/debug.c if (entry->map_err_type == MAP_ERR_NOT_CHECKED) { entry 1320 kernel/dma/debug.c entry->map_err_type = MAP_ERR_CHECKED; entry 1349 kernel/dma/debug.c struct dma_debug_entry *entry; entry 1357 kernel/dma/debug.c entry = dma_entry_alloc(); entry 1358 kernel/dma/debug.c if (!entry) entry 1361 kernel/dma/debug.c entry->type = dma_debug_sg; entry 1362 kernel/dma/debug.c entry->dev = dev; entry 1363 kernel/dma/debug.c entry->pfn = page_to_pfn(sg_page(s)); entry 1364 kernel/dma/debug.c entry->offset = s->offset, entry 1365 kernel/dma/debug.c entry->size = sg_dma_len(s); entry 1366 kernel/dma/debug.c entry->dev_addr = sg_dma_address(s); entry 1367 kernel/dma/debug.c entry->direction = direction; entry 1368 kernel/dma/debug.c entry->sg_call_ents = nents; entry 1369 kernel/dma/debug.c entry->sg_mapped_ents = mapped_ents; entry 1379 kernel/dma/debug.c add_dma_entry(entry); entry 1387 kernel/dma/debug.c struct dma_debug_entry *entry; entry 1393 kernel/dma/debug.c entry = bucket_find_exact(bucket, ref); entry 1396 kernel/dma/debug.c if (entry) entry 1397 kernel/dma/debug.c mapped_ents = entry->sg_mapped_ents; entry 1439 kernel/dma/debug.c struct dma_debug_entry *entry; entry 1451 kernel/dma/debug.c entry = dma_entry_alloc(); entry 1452 kernel/dma/debug.c if (!entry) entry 1455 kernel/dma/debug.c entry->type = dma_debug_coherent; entry 1456 kernel/dma/debug.c entry->dev = dev; entry 1457 kernel/dma/debug.c entry->offset = offset_in_page(virt); entry 1458 kernel/dma/debug.c entry->size = size; entry 1459 kernel/dma/debug.c entry->dev_addr = dma_addr; entry 1460 kernel/dma/debug.c entry->direction = DMA_BIDIRECTIONAL; entry 1463 kernel/dma/debug.c entry->pfn = vmalloc_to_pfn(virt); entry 1465 kernel/dma/debug.c entry->pfn = page_to_pfn(virt_to_page(virt)); entry 1467 kernel/dma/debug.c add_dma_entry(entry); entry 1500 kernel/dma/debug.c struct dma_debug_entry *entry; entry 1505 kernel/dma/debug.c entry = dma_entry_alloc(); entry 1506 kernel/dma/debug.c if (!entry) entry 1509 kernel/dma/debug.c entry->type = dma_debug_resource; entry 1510 kernel/dma/debug.c entry->dev = dev; entry 1511 kernel/dma/debug.c entry->pfn = PHYS_PFN(addr); entry 1512 kernel/dma/debug.c entry->offset = offset_in_page(addr); entry 1513 kernel/dma/debug.c entry->size = size; entry 1514 kernel/dma/debug.c entry->dev_addr = dma_addr; entry 1515 kernel/dma/debug.c entry->direction = direction; entry 1516 kernel/dma/debug.c entry->map_err_type = MAP_ERR_NOT_CHECKED; entry 1518 kernel/dma/debug.c add_dma_entry(entry); entry 38 kernel/events/callchain.c __weak void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, entry 43 kernel/events/callchain.c __weak void perf_callchain_user(struct perf_callchain_entry_ctx *entry, entry 181 kernel/events/callchain.c struct perf_callchain_entry *entry; entry 185 kernel/events/callchain.c entry = get_callchain_entry(&rctx); entry 189 kernel/events/callchain.c if (!entry) entry 192 kernel/events/callchain.c ctx.entry = entry; entry 194 kernel/events/callchain.c ctx.nr = entry->nr = init_nr; entry 231 kernel/events/callchain.c return entry; entry 1710 kernel/events/core.c int entry = sizeof(u64); /* value */ entry 1721 kernel/events/core.c entry += sizeof(u64); entry 1728 kernel/events/core.c size += entry * nr; entry 6009 kernel/events/core.c static void perf_pending_event(struct irq_work *entry) entry 6011 kernel/events/core.c struct perf_event *event = container_of(entry, struct perf_event, pending); entry 6892 kernel/events/core.c list_for_each_entry(filter, &ifh->list, entry) { entry 7621 kernel/events/core.c list_for_each_entry(filter, &ifh->list, entry) { entry 8752 kernel/events/core.c struct trace_entry *entry = record; entry 8764 kernel/events/core.c if (event->attr.config != entry->type) entry 9165 kernel/events/core.c INIT_LIST_HEAD(&filter->entry); entry 9166 kernel/events/core.c list_add_tail(&filter->entry, filters); entry 9175 kernel/events/core.c list_for_each_entry_safe(filter, iter, filters, entry) { entry 9177 kernel/events/core.c list_del(&filter->entry); entry 9258 kernel/events/core.c list_for_each_entry(filter, &ifh->list, entry) { entry 9894 kernel/events/core.c list_for_each_entry(pmu, &pmus, entry) { entry 10153 kernel/events/core.c list_add_rcu(&pmu->entry, &pmus); entry 10178 kernel/events/core.c list_del_rcu(&pmu->entry); entry 10282 kernel/events/core.c list_for_each_entry_rcu(pmu, &pmus, entry) { entry 12165 kernel/events/core.c list_for_each_entry(pmu, &pmus, entry) { entry 12193 kernel/events/core.c list_for_each_entry(pmu, &pmus, entry) { entry 12282 kernel/events/core.c list_for_each_entry(pmu, &pmus, entry) { entry 3673 kernel/futex.c static inline int fetch_robust_entry(struct robust_list __user **entry, entry 3682 kernel/futex.c *entry = (void __user *)(uentry & ~1UL); entry 3697 kernel/futex.c struct robust_list __user *entry, *next_entry, *pending; entry 3710 kernel/futex.c if (fetch_robust_entry(&entry, &head->list.next, &pi)) entry 3725 kernel/futex.c while (entry != &head->list) { entry 3730 kernel/futex.c rc = fetch_robust_entry(&next_entry, &entry->next, &next_pi); entry 3735 kernel/futex.c if (entry != pending) { entry 3736 kernel/futex.c if (handle_futex_death((void __user *)entry + futex_offset, entry 3742 kernel/futex.c entry = next_entry; entry 3967 kernel/futex.c compat_fetch_robust_entry(compat_uptr_t *uentry, struct robust_list __user **entry, entry 3973 kernel/futex.c *entry = compat_ptr((*uentry) & ~1); entry 3979 kernel/futex.c static void __user *futex_uaddr(struct robust_list __user *entry, entry 3982 kernel/futex.c compat_uptr_t base = ptr_to_compat(entry); entry 3997 kernel/futex.c struct robust_list __user *entry, *next_entry, *pending; entry 4011 kernel/futex.c if (compat_fetch_robust_entry(&uentry, &entry, &head->list.next, &pi)) entry 4027 kernel/futex.c while (entry != (struct robust_list __user *) &head->list) { entry 4033 kernel/futex.c (compat_uptr_t __user *)&entry->next, &next_pi); entry 4038 kernel/futex.c if (entry != pending) { entry 4039 kernel/futex.c void __user *uaddr = futex_uaddr(entry, futex_offset); entry 4048 kernel/futex.c entry = next_entry; entry 112 kernel/irq/chip.c struct msi_desc *entry) entry 119 kernel/irq/chip.c desc->irq_common_data.msi_desc = entry; entry 120 kernel/irq/chip.c if (entry && !irq_offset) entry 121 kernel/irq/chip.c entry->irq = irq_base; entry 133 kernel/irq/chip.c int irq_set_msi_desc(unsigned int irq, struct msi_desc *entry) entry 135 kernel/irq/chip.c return irq_set_msi_desc_off(irq, 0, entry); entry 53 kernel/irq/msi.c void free_msi_entry(struct msi_desc *entry) entry 55 kernel/irq/msi.c kfree(entry->affinity); entry 56 kernel/irq/msi.c kfree(entry); entry 59 kernel/irq/msi.c void __get_cached_msi_msg(struct msi_desc *entry, struct msi_msg *msg) entry 61 kernel/irq/msi.c *msg = entry->msg; entry 66 kernel/irq/msi.c struct msi_desc *entry = irq_get_msi_desc(irq); entry 68 kernel/irq/msi.c __get_cached_msi_msg(entry, msg); entry 309 kernel/jump_label.c static int addr_conflict(struct jump_entry *entry, void *start, void *end) entry 311 kernel/jump_label.c if (jump_entry_code(entry) <= (unsigned long)end && entry 312 kernel/jump_label.c jump_entry_code(entry) + JUMP_LABEL_NOP_SIZE > (unsigned long)start) entry 339 kernel/jump_label.c void __weak __init_or_module arch_jump_label_transform_static(struct jump_entry *entry, entry 342 kernel/jump_label.c arch_jump_label_transform(entry, type); entry 391 kernel/jump_label.c static enum jump_label_type jump_label_type(struct jump_entry *entry) entry 393 kernel/jump_label.c struct static_key *key = jump_entry_key(entry); entry 395 kernel/jump_label.c bool branch = jump_entry_is_branch(entry); entry 401 kernel/jump_label.c static bool jump_label_can_update(struct jump_entry *entry, bool init) entry 406 kernel/jump_label.c if (!init && jump_entry_is_init(entry)) entry 409 kernel/jump_label.c if (!kernel_text_address(jump_entry_code(entry))) { entry 410 kernel/jump_label.c WARN_ONCE(!jump_entry_is_init(entry), entry 412 kernel/jump_label.c (void *)jump_entry_code(entry)); entry 421 kernel/jump_label.c struct jump_entry *entry, entry 425 kernel/jump_label.c for (; (entry < stop) && (jump_entry_key(entry) == key); entry++) { entry 426 kernel/jump_label.c if (jump_label_can_update(entry, init)) entry 427 kernel/jump_label.c arch_jump_label_transform(entry, jump_label_type(entry)); entry 432 kernel/jump_label.c struct jump_entry *entry, entry 436 kernel/jump_label.c for (; (entry < stop) && (jump_entry_key(entry) == key); entry++) { entry 438 kernel/jump_label.c if (!jump_label_can_update(entry, init)) entry 441 kernel/jump_label.c if (!arch_jump_label_transform_queue(entry, jump_label_type(entry))) { entry 446 kernel/jump_label.c BUG_ON(!arch_jump_label_transform_queue(entry, jump_label_type(entry))); entry 500 kernel/jump_label.c static enum jump_label_type jump_label_init_type(struct jump_entry *entry) entry 502 kernel/jump_label.c struct static_key *key = jump_entry_key(entry); entry 504 kernel/jump_label.c bool branch = jump_entry_is_branch(entry); entry 790 kernel/jump_label.c struct jump_entry *entry; entry 805 kernel/jump_label.c entry = static_key_entries(key); entry 807 kernel/jump_label.c if (entry) entry 808 kernel/jump_label.c __jump_label_update(key, entry, stop, entry 39 kernel/kexec.c static int kimage_alloc_init(struct kimage **rimage, unsigned long entry, entry 50 kernel/kexec.c if ((entry < phys_to_boot_phys(crashk_res.start)) || entry 51 kernel/kexec.c (entry > phys_to_boot_phys(crashk_res.end))) entry 60 kernel/kexec.c image->start = entry; entry 106 kernel/kexec.c static int do_kexec_load(unsigned long entry, unsigned long nr_segments, entry 135 kernel/kexec.c ret = kimage_alloc_init(&image, entry, nr_segments, segments, flags); entry 232 kernel/kexec.c SYSCALL_DEFINE4(kexec_load, unsigned long, entry, unsigned long, nr_segments, entry 257 kernel/kexec.c result = do_kexec_load(entry, nr_segments, segments, flags); entry 265 kernel/kexec.c COMPAT_SYSCALL_DEFINE4(kexec_load, compat_ulong_t, entry, entry 311 kernel/kexec.c result = do_kexec_load(entry, nr_segments, ksegments, flags); entry 264 kernel/kexec_core.c image->entry = &image->head; entry 534 kernel/kexec_core.c static int kimage_add_entry(struct kimage *image, kimage_entry_t entry) entry 536 kernel/kexec_core.c if (*image->entry != 0) entry 537 kernel/kexec_core.c image->entry++; entry 539 kernel/kexec_core.c if (image->entry == image->last_entry) { entry 548 kernel/kexec_core.c *image->entry = virt_to_boot_phys(ind_page) | IND_INDIRECTION; entry 549 kernel/kexec_core.c image->entry = ind_page; entry 553 kernel/kexec_core.c *image->entry = entry; entry 554 kernel/kexec_core.c image->entry++; entry 555 kernel/kexec_core.c *image->entry = 0; entry 594 kernel/kexec_core.c if (*image->entry != 0) entry 595 kernel/kexec_core.c image->entry++; entry 597 kernel/kexec_core.c *image->entry = IND_DONE; entry 600 kernel/kexec_core.c #define for_each_kimage_entry(image, ptr, entry) \ entry 601 kernel/kexec_core.c for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); \ entry 602 kernel/kexec_core.c ptr = (entry & IND_INDIRECTION) ? \ entry 603 kernel/kexec_core.c boot_phys_to_virt((entry & PAGE_MASK)) : ptr + 1) entry 605 kernel/kexec_core.c static void kimage_free_entry(kimage_entry_t entry) entry 609 kernel/kexec_core.c page = boot_pfn_to_page(entry >> PAGE_SHIFT); entry 615 kernel/kexec_core.c kimage_entry_t *ptr, entry; entry 627 kernel/kexec_core.c for_each_kimage_entry(image, ptr, entry) { entry 628 kernel/kexec_core.c if (entry & IND_INDIRECTION) { entry 635 kernel/kexec_core.c ind = entry; entry 636 kernel/kexec_core.c } else if (entry & IND_SOURCE) entry 637 kernel/kexec_core.c kimage_free_entry(entry); entry 662 kernel/kexec_core.c kimage_entry_t *ptr, entry; entry 665 kernel/kexec_core.c for_each_kimage_entry(image, ptr, entry) { entry 666 kernel/kexec_core.c if (entry & IND_DESTINATION) entry 667 kernel/kexec_core.c destination = entry & PAGE_MASK; entry 668 kernel/kexec_core.c else if (entry & IND_SOURCE) { entry 1848 kernel/kprobes.c unsigned long __weak arch_deref_entry_point(void *entry) entry 1850 kernel/kprobes.c return (unsigned long)entry; entry 2146 kernel/kprobes.c int kprobe_add_ksym_blacklist(unsigned long entry) entry 2151 kernel/kprobes.c if (!kernel_text_address(entry) || entry 2152 kernel/kprobes.c !kallsyms_lookup_size_offset(entry, &size, &offset)) entry 2158 kernel/kprobes.c ent->start_addr = entry; entry 2159 kernel/kprobes.c ent->end_addr = entry + size; entry 2169 kernel/kprobes.c unsigned long entry; entry 2172 kernel/kprobes.c for (entry = start; entry < end; entry += ret) { entry 2173 kernel/kprobes.c ret = kprobe_add_ksym_blacklist(entry); entry 2198 kernel/kprobes.c unsigned long entry; entry 2203 kernel/kprobes.c entry = arch_deref_entry_point((void *)*iter); entry 2204 kernel/kprobes.c ret = kprobe_add_ksym_blacklist(entry); entry 924 kernel/locking/lockdep.c list_for_each_entry(e, h, entry) { entry 1011 kernel/locking/lockdep.c hlist_for_each_entry_rcu(chain, head, entry) { entry 1023 kernel/locking/lockdep.c if (!in_any_class_list(&e->entry)) { entry 1038 kernel/locking/lockdep.c if (in_any_class_list(&e->entry)) { entry 1307 kernel/locking/lockdep.c struct lock_list *entry; entry 1312 kernel/locking/lockdep.c entry = alloc_list_entry(); entry 1313 kernel/locking/lockdep.c if (!entry) entry 1316 kernel/locking/lockdep.c entry->class = this; entry 1317 kernel/locking/lockdep.c entry->links_to = links_to; entry 1318 kernel/locking/lockdep.c entry->distance = distance; entry 1319 kernel/locking/lockdep.c entry->trace = trace; entry 1325 kernel/locking/lockdep.c list_add_tail_rcu(&entry->entry, head); entry 1462 kernel/locking/lockdep.c int (*match)(struct lock_list *entry, void *data), entry 1466 kernel/locking/lockdep.c struct lock_list *entry; entry 1496 kernel/locking/lockdep.c list_for_each_entry_rcu(entry, head, entry) { entry 1497 kernel/locking/lockdep.c if (!lock_accessed(entry)) { entry 1499 kernel/locking/lockdep.c mark_lock_accessed(entry, lock); entry 1500 kernel/locking/lockdep.c if (match(entry, data)) { entry 1501 kernel/locking/lockdep.c *target_entry = entry; entry 1506 kernel/locking/lockdep.c if (__cq_enqueue(cq, entry)) { entry 1522 kernel/locking/lockdep.c int (*match)(struct lock_list *entry, void *data), entry 1532 kernel/locking/lockdep.c int (*match)(struct lock_list *entry, void *data), entry 1616 kernel/locking/lockdep.c print_circular_bug_header(struct lock_list *entry, unsigned int depth, entry 1640 kernel/locking/lockdep.c print_circular_bug_entry(entry, depth); entry 1643 kernel/locking/lockdep.c static inline int class_equal(struct lock_list *entry, void *data) entry 1645 kernel/locking/lockdep.c return entry->class == data; entry 1698 kernel/locking/lockdep.c static int noop_count(struct lock_list *entry, void *data) entry 1852 kernel/locking/lockdep.c static inline int usage_accumulate(struct lock_list *entry, void *mask) entry 1854 kernel/locking/lockdep.c *(unsigned long *)mask |= entry->class->usage_mask; entry 1865 kernel/locking/lockdep.c static inline int usage_match(struct lock_list *entry, void *mask) entry 1867 kernel/locking/lockdep.c return entry->class->usage_mask & *(unsigned long *)mask; entry 1949 kernel/locking/lockdep.c struct lock_list *entry = leaf; entry 1956 kernel/locking/lockdep.c print_lock_class_header(entry->class, depth); entry 1958 kernel/locking/lockdep.c print_lock_trace(entry->trace, 2); entry 1961 kernel/locking/lockdep.c if (depth == 0 && (entry != root)) { entry 1966 kernel/locking/lockdep.c entry = get_lock_parent(entry); entry 1968 kernel/locking/lockdep.c } while (entry && (depth >= 0)); entry 2448 kernel/locking/lockdep.c struct lock_list *entry; entry 2504 kernel/locking/lockdep.c list_for_each_entry(entry, &hlock_class(prev)->locks_after, entry) { entry 2505 kernel/locking/lockdep.c if (entry->class == hlock_class(next)) { entry 2507 kernel/locking/lockdep.c entry->distance = 1; entry 2848 kernel/locking/lockdep.c hlist_add_head_rcu(&chain->entry, hash_head); entry 2864 kernel/locking/lockdep.c hlist_for_each_entry_rcu(chain, hash_head, entry) { entry 3131 kernel/locking/lockdep.c struct lock_list *entry = other; entry 3158 kernel/locking/lockdep.c if (depth == 0 && (entry != root)) { entry 3162 kernel/locking/lockdep.c middle = entry; entry 3163 kernel/locking/lockdep.c entry = get_lock_parent(entry); entry 3165 kernel/locking/lockdep.c } while (entry && entry != root && (depth >= 0)); entry 4812 kernel/locking/lockdep.c hlist_del_rcu(&chain->entry); entry 4828 kernel/locking/lockdep.c hlist_add_head_rcu(&new_chain->entry, chainhashentry(chain_key)); entry 4842 kernel/locking/lockdep.c hlist_for_each_entry_rcu(chain, head, entry) { entry 4853 kernel/locking/lockdep.c struct lock_list *entry; entry 4863 kernel/locking/lockdep.c entry = list_entries + i; entry 4864 kernel/locking/lockdep.c if (entry->class != class && entry->links_to != class) entry 4868 kernel/locking/lockdep.c list_del_rcu(&entry->entry); entry 61 kernel/locking/lockdep_proc.c struct lock_list *entry; entry 85 kernel/locking/lockdep_proc.c list_for_each_entry(entry, &class->locks_after, entry) { entry 86 kernel/locking/lockdep_proc.c if (entry->distance == 1) { entry 87 kernel/locking/lockdep_proc.c seq_printf(m, " -> [%p] ", entry->class->key); entry 88 kernel/locking/lockdep_proc.c print_name(m, entry->class); entry 277 kernel/locking/rtmutex.c struct rt_mutex_waiter *entry; entry 282 kernel/locking/rtmutex.c entry = rb_entry(parent, struct rt_mutex_waiter, tree_entry); entry 283 kernel/locking/rtmutex.c if (rt_mutex_waiter_less(waiter, entry)) { entry 310 kernel/locking/rtmutex.c struct rt_mutex_waiter *entry; entry 315 kernel/locking/rtmutex.c entry = rb_entry(parent, struct rt_mutex_waiter, pi_tree_entry); entry 316 kernel/locking/rtmutex.c if (rt_mutex_waiter_less(waiter, entry)) { entry 48 kernel/power/console.c struct pm_vt_switch *entry, *tmp; entry 59 kernel/power/console.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 60 kernel/power/console.c if (!entry) entry 63 kernel/power/console.c entry->required = required; entry 64 kernel/power/console.c entry->dev = dev; entry 66 kernel/power/console.c list_add(&entry->head, &pm_vt_switch_list); entry 109 kernel/power/console.c struct pm_vt_switch *entry; entry 119 kernel/power/console.c list_for_each_entry(entry, &pm_vt_switch_list, head) { entry 120 kernel/power/console.c if (entry->required) entry 528 kernel/profile.c struct proc_dir_entry *entry; entry 550 kernel/profile.c entry = proc_create("profile", S_IWUSR | S_IRUGO, entry 552 kernel/profile.c if (!entry) entry 554 kernel/profile.c proc_set_size(entry, (1 + prof_len) * sizeof(atomic_t)); entry 912 kernel/ptrace.c unsigned long args[ARRAY_SIZE(info->entry.args)]; entry 916 kernel/ptrace.c info->entry.nr = syscall_get_nr(child, regs); entry 919 kernel/ptrace.c info->entry.args[i] = args[i]; entry 922 kernel/ptrace.c return offsetofend(struct ptrace_syscall_info, entry.args); entry 969 kernel/ptrace.c unsigned long actual_size = offsetof(struct ptrace_syscall_info, entry); entry 38 kernel/rcu/srcutiny.c INIT_LIST_HEAD(&ssp->srcu_work.entry); entry 172 kernel/rcu/srcutiny.c else if (list_empty(&ssp->srcu_work.entry)) entry 173 kernel/rcu/srcutiny.c list_add(&ssp->srcu_work.entry, &srcu_boot_list); entry 211 kernel/rcu/srcutiny.c struct srcu_struct, srcu_work.entry); entry 212 kernel/rcu/srcutiny.c list_del_init(&ssp->srcu_work.entry); entry 687 kernel/rcu/srcutree.c else if (list_empty(&ssp->work.work.entry)) entry 688 kernel/rcu/srcutree.c list_add(&ssp->work.work.entry, &srcu_boot_list); entry 1309 kernel/rcu/srcutree.c work.work.entry); entry 1311 kernel/rcu/srcutree.c list_del_init(&ssp->work.work.entry); entry 1625 kernel/resource.c struct resource_entry *entry; entry 1627 kernel/resource.c entry = kzalloc(sizeof(*entry) + extra_size, GFP_KERNEL); entry 1628 kernel/resource.c if (entry) { entry 1629 kernel/resource.c INIT_LIST_HEAD(&entry->node); entry 1630 kernel/resource.c entry->res = res ? res : &entry->__res; entry 1633 kernel/resource.c return entry; entry 1639 kernel/resource.c struct resource_entry *entry, *tmp; entry 1641 kernel/resource.c list_for_each_entry_safe(entry, tmp, head, node) entry 1642 kernel/resource.c resource_list_destroy_entry(entry); entry 450 kernel/sched/deadline.c struct task_struct *entry; entry 457 kernel/sched/deadline.c entry = rb_entry(parent, struct task_struct, entry 459 kernel/sched/deadline.c if (dl_entity_preempt(&p->dl, &entry->dl)) entry 1359 kernel/sched/deadline.c struct sched_dl_entity *entry; entry 1361 kernel/sched/deadline.c entry = rb_entry(leftmost, struct sched_dl_entity, rb_node); entry 1362 kernel/sched/deadline.c dl_rq->earliest_dl.curr = entry->deadline; entry 1363 kernel/sched/deadline.c cpudl_set(&rq->rd->cpudl, rq->cpu, entry->deadline); entry 1407 kernel/sched/deadline.c struct sched_dl_entity *entry; entry 1414 kernel/sched/deadline.c entry = rb_entry(parent, struct sched_dl_entity, rb_node); entry 1415 kernel/sched/deadline.c if (dl_time_before(dl_se->deadline, entry->deadline)) entry 209 kernel/sched/debug.c struct ctl_table *entry = entry 212 kernel/sched/debug.c return entry; entry 217 kernel/sched/debug.c struct ctl_table *entry; entry 225 kernel/sched/debug.c for (entry = *tablep; entry->mode; entry++) { entry 226 kernel/sched/debug.c if (entry->child) entry 227 kernel/sched/debug.c sd_free_ctl_entry(&entry->child); entry 228 kernel/sched/debug.c if (entry->proc_handler == NULL) entry 229 kernel/sched/debug.c kfree(entry->procname); entry 237 kernel/sched/debug.c set_table_entry(struct ctl_table *entry, entry 241 kernel/sched/debug.c entry->procname = procname; entry 242 kernel/sched/debug.c entry->data = data; entry 243 kernel/sched/debug.c entry->maxlen = maxlen; entry 244 kernel/sched/debug.c entry->mode = mode; entry 245 kernel/sched/debug.c entry->proc_handler = proc_handler; entry 271 kernel/sched/debug.c struct ctl_table *entry, *table; entry 278 kernel/sched/debug.c entry = table = sd_alloc_ctl_entry(domain_num + 1); entry 285 kernel/sched/debug.c entry->procname = kstrdup(buf, GFP_KERNEL); entry 286 kernel/sched/debug.c entry->mode = 0555; entry 287 kernel/sched/debug.c entry->child = sd_alloc_ctl_domain_table(sd); entry 288 kernel/sched/debug.c entry++; entry 569 kernel/sched/fair.c struct sched_entity *entry; entry 577 kernel/sched/fair.c entry = rb_entry(parent, struct sched_entity, run_node); entry 582 kernel/sched/fair.c if (entity_before(se, entry)) { entry 76 kernel/sched/wait.c curr = list_next_entry(bookmark, entry); entry 78 kernel/sched/wait.c list_del(&bookmark->entry); entry 81 kernel/sched/wait.c curr = list_first_entry(&wq_head->head, wait_queue_entry_t, entry); entry 83 kernel/sched/wait.c if (&curr->entry == &wq_head->head) entry 86 kernel/sched/wait.c list_for_each_entry_safe_from(curr, next, &wq_head->head, entry) { entry 100 kernel/sched/wait.c (&next->entry != &wq_head->head)) { entry 102 kernel/sched/wait.c list_add_tail(&bookmark->entry, &next->entry); entry 119 kernel/sched/wait.c INIT_LIST_HEAD(&bookmark.entry); entry 228 kernel/sched/wait.c if (list_empty(&wq_entry->entry)) entry 242 kernel/sched/wait.c if (list_empty(&wq_entry->entry)) entry 254 kernel/sched/wait.c INIT_LIST_HEAD(&wq_entry->entry); entry 277 kernel/sched/wait.c list_del_init(&wq_entry->entry); entry 280 kernel/sched/wait.c if (list_empty(&wq_entry->entry)) { entry 303 kernel/sched/wait.c if (likely(list_empty(&wait->entry))) entry 320 kernel/sched/wait.c if (likely(list_empty(&wait->entry))) entry 362 kernel/sched/wait.c if (!list_empty_careful(&wq_entry->entry)) { entry 364 kernel/sched/wait.c list_del_init(&wq_entry->entry); entry 375 kernel/sched/wait.c list_del_init(&wq_entry->entry); entry 184 kernel/sched/wait_bit.c .entry = LIST_HEAD_INIT(wbq_entry->wq_entry.entry), entry 213 kernel/smp.c struct llist_node *entry; entry 220 kernel/smp.c entry = llist_del_all(head); entry 221 kernel/smp.c entry = llist_reverse_order(entry); entry 233 kernel/smp.c llist_for_each_entry(csd, entry, llist) entry 238 kernel/smp.c llist_for_each_entry_safe(csd, csd_next, entry, llist) { entry 834 kernel/time/clocksource.c struct list_head *entry = &clocksource_list; entry 841 kernel/time/clocksource.c entry = &tmp->list; entry 843 kernel/time/clocksource.c list_add(&cs->list, entry); entry 540 kernel/time/timer.c hlist_add_head(&timer->entry, base->vectors + idx); entry 612 kernel/time/timer.c return (timer->entry.pprev == NULL && entry 613 kernel/time/timer.c timer->entry.next == TIMER_ENTRY_STATIC); entry 782 kernel/time/timer.c timer->entry.pprev = NULL; entry 811 kernel/time/timer.c struct hlist_node *entry = &timer->entry; entry 815 kernel/time/timer.c __hlist_del(entry); entry 817 kernel/time/timer.c entry->pprev = NULL; entry 818 kernel/time/timer.c entry->next = LIST_POISON2; entry 829 kernel/time/timer.c if (hlist_is_singular_node(&timer->entry, base->vectors + idx)) entry 1435 kernel/time/timer.c timer = hlist_entry(head->first, struct timer_list, entry); entry 1951 kernel/time/timer.c timer = hlist_entry(head->first, struct timer_list, entry); entry 629 kernel/trace/bpf_trace.c static void do_bpf_send_signal(struct irq_work *entry) entry 633 kernel/trace/bpf_trace.c work = container_of(entry, struct send_signal_irq_work, irq_work); entry 964 kernel/trace/ftrace.c struct dentry *entry; entry 995 kernel/trace/ftrace.c entry = tracefs_create_file("function_profile_enabled", 0644, entry 997 kernel/trace/ftrace.c if (!entry) entry 1129 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 1135 kernel/trace/ftrace.c hlist_for_each_entry_rcu_notrace(entry, hhd, hlist) { entry 1136 kernel/trace/ftrace.c if (entry->ip == ip) entry 1137 kernel/trace/ftrace.c return entry; entry 1162 kernel/trace/ftrace.c struct ftrace_func_entry *entry) entry 1167 kernel/trace/ftrace.c key = ftrace_hash_key(hash, entry->ip); entry 1169 kernel/trace/ftrace.c hlist_add_head(&entry->hlist, hhd); entry 1175 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 1177 kernel/trace/ftrace.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 1178 kernel/trace/ftrace.c if (!entry) entry 1181 kernel/trace/ftrace.c entry->ip = ip; entry 1182 kernel/trace/ftrace.c __add_hash_entry(hash, entry); entry 1189 kernel/trace/ftrace.c struct ftrace_func_entry *entry) entry 1191 kernel/trace/ftrace.c hlist_del(&entry->hlist); entry 1192 kernel/trace/ftrace.c kfree(entry); entry 1198 kernel/trace/ftrace.c struct ftrace_func_entry *entry) entry 1200 kernel/trace/ftrace.c hlist_del_rcu(&entry->hlist); entry 1208 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 1217 kernel/trace/ftrace.c hlist_for_each_entry_safe(entry, tn, hhd, hlist) entry 1218 kernel/trace/ftrace.c free_hash_entry(hash, entry); entry 1330 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 1349 kernel/trace/ftrace.c hlist_for_each_entry(entry, &hash->buckets[i], hlist) { entry 1350 kernel/trace/ftrace.c ret = add_hash_entry(new_hash, entry->ip); entry 1376 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 1409 kernel/trace/ftrace.c hlist_for_each_entry_safe(entry, tn, hhd, hlist) { entry 1410 kernel/trace/ftrace.c remove_hash_entry(src, entry); entry 1411 kernel/trace/ftrace.c __add_hash_entry(new_hash, entry); entry 3714 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 3717 kernel/trace/ftrace.c entry = ftrace_lookup_ip(hash, rec->ip); entry 3720 kernel/trace/ftrace.c if (!entry) entry 3723 kernel/trace/ftrace.c free_hash_entry(hash, entry); entry 3726 kernel/trace/ftrace.c if (entry) entry 4132 kernel/trace/ftrace.c struct ftrace_func_entry entry; entry 4172 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 4175 kernel/trace/ftrace.c entry = ftrace_lookup_ip(&mapper->hash, ip); entry 4176 kernel/trace/ftrace.c if (!entry) entry 4179 kernel/trace/ftrace.c map = (struct ftrace_func_map *)entry; entry 4194 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 4197 kernel/trace/ftrace.c entry = ftrace_lookup_ip(&mapper->hash, ip); entry 4198 kernel/trace/ftrace.c if (entry) entry 4205 kernel/trace/ftrace.c map->entry.ip = ip; entry 4208 kernel/trace/ftrace.c __add_hash_entry(&mapper->hash, &map->entry); entry 4226 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 4230 kernel/trace/ftrace.c entry = ftrace_lookup_ip(&mapper->hash, ip); entry 4231 kernel/trace/ftrace.c if (!entry) entry 4234 kernel/trace/ftrace.c map = (struct ftrace_func_map *)entry; entry 4237 kernel/trace/ftrace.c remove_hash_entry(&mapper->hash, entry); entry 4238 kernel/trace/ftrace.c kfree(entry); entry 4254 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 4266 kernel/trace/ftrace.c hlist_for_each_entry(entry, hhd, hlist) { entry 4267 kernel/trace/ftrace.c map = (struct ftrace_func_map *)entry; entry 4314 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 4381 kernel/trace/ftrace.c hlist_for_each_entry(entry, &hash->buckets[i], hlist) { entry 4382 kernel/trace/ftrace.c if (ftrace_lookup_ip(old_hash, entry->ip)) entry 4391 kernel/trace/ftrace.c entry->ip, data, entry 4443 kernel/trace/ftrace.c hlist_for_each_entry(entry, &hash->buckets[i], hlist) { entry 4444 kernel/trace/ftrace.c if (ftrace_lookup_ip(old_hash, entry->ip)) entry 4446 kernel/trace/ftrace.c probe_ops->free(probe_ops, tr, entry->ip, probe->data); entry 4457 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 4522 kernel/trace/ftrace.c hlist_for_each_entry_safe(entry, tmp, &hash->buckets[i], hlist) { entry 4525 kernel/trace/ftrace.c kallsyms_lookup(entry->ip, NULL, NULL, entry 4531 kernel/trace/ftrace.c remove_hash_entry(hash, entry); entry 4532 kernel/trace/ftrace.c hlist_add_head(&entry->hlist, &hhd); entry 4560 kernel/trace/ftrace.c hlist_for_each_entry_safe(entry, tmp, &hhd, hlist) { entry 4561 kernel/trace/ftrace.c hlist_del(&entry->hlist); entry 4563 kernel/trace/ftrace.c probe_ops->free(probe_ops, tr, entry->ip, probe->data); entry 4564 kernel/trace/ftrace.c kfree(entry); entry 4731 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 4737 kernel/trace/ftrace.c entry = ftrace_lookup_ip(hash, ip); entry 4738 kernel/trace/ftrace.c if (!entry) entry 4740 kernel/trace/ftrace.c free_hash_entry(hash, entry); entry 5117 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 5129 kernel/trace/ftrace.c struct ftrace_func_entry *entry = fgd->entry; entry 5136 kernel/trace/ftrace.c if (entry) { entry 5137 kernel/trace/ftrace.c hlist_for_each_entry_continue(entry, hlist) { entry 5138 kernel/trace/ftrace.c fgd->entry = entry; entry 5139 kernel/trace/ftrace.c return entry; entry 5147 kernel/trace/ftrace.c hlist_for_each_entry(entry, head, hlist) { entry 5148 kernel/trace/ftrace.c fgd->entry = entry; entry 5150 kernel/trace/ftrace.c return entry; entry 5181 kernel/trace/ftrace.c fgd->entry = NULL; entry 5192 kernel/trace/ftrace.c struct ftrace_func_entry *entry = v; entry 5194 kernel/trace/ftrace.c if (!entry) entry 5197 kernel/trace/ftrace.c if (entry == FTRACE_GRAPH_EMPTY) { entry 5207 kernel/trace/ftrace.c seq_printf(m, "%ps\n", (void *)entry->ip); entry 5407 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 5430 kernel/trace/ftrace.c entry = ftrace_lookup_ip(hash, rec->ip); entry 5435 kernel/trace/ftrace.c if (entry) entry 5440 kernel/trace/ftrace.c if (entry) { entry 5441 kernel/trace/ftrace.c free_hash_entry(hash, entry); entry 5712 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 5721 kernel/trace/ftrace.c entry = __ftrace_lookup_ip(hash, rec->ip); entry 5727 kernel/trace/ftrace.c if (entry) entry 5728 kernel/trace/ftrace.c entry->ip = 0; entry 6074 kernel/trace/ftrace.c struct ftrace_func_entry *entry; entry 6076 kernel/trace/ftrace.c entry = ftrace_lookup_ip(hash, func->ip); entry 6082 kernel/trace/ftrace.c if (entry) entry 6083 kernel/trace/ftrace.c entry->ip = 0; entry 88 kernel/trace/ring_buffer_benchmark.c int *entry; entry 95 kernel/trace/ring_buffer_benchmark.c entry = ring_buffer_event_data(event); entry 96 kernel/trace/ring_buffer_benchmark.c if (*entry != cpu) { entry 111 kernel/trace/ring_buffer_benchmark.c int *entry; entry 145 kernel/trace/ring_buffer_benchmark.c entry = ring_buffer_event_data(event); entry 146 kernel/trace/ring_buffer_benchmark.c if (*entry != cpu) { entry 158 kernel/trace/ring_buffer_benchmark.c entry = ring_buffer_event_data(event); entry 159 kernel/trace/ring_buffer_benchmark.c if (*entry != cpu) { entry 250 kernel/trace/ring_buffer_benchmark.c int *entry; entry 259 kernel/trace/ring_buffer_benchmark.c entry = ring_buffer_event_data(event); entry 260 kernel/trace/ring_buffer_benchmark.c *entry = smp_processor_id(); entry 837 kernel/trace/trace.c struct print_entry *entry; entry 850 kernel/trace/trace.c alloc = sizeof(*entry) + size + 2; /* possible \n added */ entry 859 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 860 kernel/trace/trace.c entry->ip = ip; entry 862 kernel/trace/trace.c memcpy(&entry->buf, str, size); entry 865 kernel/trace/trace.c if (entry->buf[size - 1] != '\n') { entry 866 kernel/trace/trace.c entry->buf[size] = '\n'; entry 867 kernel/trace/trace.c entry->buf[size + 1] = '\0'; entry 869 kernel/trace/trace.c entry->buf[size] = '\0'; entry 887 kernel/trace/trace.c struct bputs_entry *entry; entry 907 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 908 kernel/trace/trace.c entry->ip = ip; entry 909 kernel/trace/trace.c entry->str = str; entry 2340 kernel/trace/trace.c tracing_generic_entry_update(struct trace_entry *entry, unsigned short type, entry 2345 kernel/trace/trace.c entry->preempt_count = pc & 0xff; entry 2346 kernel/trace/trace.c entry->pid = (tsk) ? tsk->pid : 0; entry 2347 kernel/trace/trace.c entry->type = type; entry 2348 kernel/trace/trace.c entry->flags = entry 2490 kernel/trace/trace.c struct ring_buffer_event *entry; entry 2497 kernel/trace/trace.c (entry = this_cpu_read(trace_buffered_event))) { entry 2501 kernel/trace/trace.c trace_event_setup(entry, type, flags, pc); entry 2502 kernel/trace/trace.c entry->array[0] = len; entry 2503 kernel/trace/trace.c return entry; entry 2508 kernel/trace/trace.c entry = __trace_buffer_lock_reserve(*current_rb, entry 2516 kernel/trace/trace.c if (!entry && trace_file->flags & EVENT_FILE_FL_TRIGGER_COND) { entry 2518 kernel/trace/trace.c entry = __trace_buffer_lock_reserve(*current_rb, entry 2521 kernel/trace/trace.c return entry; entry 2548 kernel/trace/trace.c iter->ent = fbuffer->entry; entry 2595 kernel/trace/trace.c fbuffer->event, fbuffer->entry, entry 2641 kernel/trace/trace.c struct trace_entry *entry; entry 2644 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 2646 kernel/trace/trace.c export->write(export, entry, size); entry 2768 kernel/trace/trace.c struct ftrace_entry *entry; entry 2770 kernel/trace/trace.c event = __trace_buffer_lock_reserve(buffer, TRACE_FN, sizeof(*entry), entry 2774 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 2775 kernel/trace/trace.c entry->ip = ip; entry 2776 kernel/trace/trace.c entry->parent_ip = parent_ip; entry 2778 kernel/trace/trace.c if (!call_filter_check_discard(call, entry, buffer, event)) { entry 2812 kernel/trace/trace.c struct stack_entry *entry; entry 2859 kernel/trace/trace.c sizeof(*entry) + size, flags, pc); entry 2862 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 2864 kernel/trace/trace.c memcpy(&entry->caller, fstack->calls, size); entry 2865 kernel/trace/trace.c entry->size = nr_entries; entry 2867 kernel/trace/trace.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 2943 kernel/trace/trace.c struct userstack_entry *entry; entry 2966 kernel/trace/trace.c sizeof(*entry), flags, pc); entry 2969 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 2971 kernel/trace/trace.c entry->tgid = current->tgid; entry 2972 kernel/trace/trace.c memset(&entry->caller, 0, sizeof(entry->caller)); entry 2974 kernel/trace/trace.c stack_trace_save_user(entry->caller, FTRACE_STACK_ENTRIES); entry 2975 kernel/trace/trace.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 3111 kernel/trace/trace.c struct bprint_entry *entry; entry 3137 kernel/trace/trace.c size = sizeof(*entry) + sizeof(u32) * len; entry 3143 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 3144 kernel/trace/trace.c entry->ip = ip; entry 3145 kernel/trace/trace.c entry->fmt = fmt; entry 3147 kernel/trace/trace.c memcpy(entry->buf, tbuffer, sizeof(u32) * len); entry 3148 kernel/trace/trace.c if (!call_filter_check_discard(call, entry, buffer, event)) { entry 3172 kernel/trace/trace.c struct print_entry *entry; entry 3195 kernel/trace/trace.c size = sizeof(*entry) + len + 1; entry 3200 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 3201 kernel/trace/trace.c entry->ip = ip; entry 3203 kernel/trace/trace.c memcpy(&entry->buf, tbuffer, len + 1); entry 3204 kernel/trace/trace.c if (!call_filter_check_discard(call, entry, buffer, event)) { entry 3728 kernel/trace/trace.c struct trace_entry *entry; entry 3731 kernel/trace/trace.c entry = iter->ent; entry 3735 kernel/trace/trace.c event = ftrace_find_event(entry->type); entry 3750 kernel/trace/trace.c trace_seq_printf(s, "Unknown type %d\n", entry->type); entry 3759 kernel/trace/trace.c struct trace_entry *entry; entry 3762 kernel/trace/trace.c entry = iter->ent; entry 3766 kernel/trace/trace.c entry->pid, iter->cpu, iter->ts); entry 3771 kernel/trace/trace.c event = ftrace_find_event(entry->type); entry 3775 kernel/trace/trace.c trace_seq_printf(s, "%d ?\n", entry->type); entry 3785 kernel/trace/trace.c struct trace_entry *entry; entry 3788 kernel/trace/trace.c entry = iter->ent; entry 3791 kernel/trace/trace.c SEQ_PUT_HEX_FIELD(s, entry->pid); entry 3798 kernel/trace/trace.c event = ftrace_find_event(entry->type); entry 3814 kernel/trace/trace.c struct trace_entry *entry; entry 3817 kernel/trace/trace.c entry = iter->ent; entry 3820 kernel/trace/trace.c SEQ_PUT_FIELD(s, entry->pid); entry 3827 kernel/trace/trace.c event = ftrace_find_event(entry->type); entry 6387 kernel/trace/trace.c struct print_entry *entry; entry 6409 kernel/trace/trace.c size = sizeof(*entry) + cnt + 2; /* add '\0' and possible '\n' */ entry 6422 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 6423 kernel/trace/trace.c entry->ip = _THIS_IP_; entry 6425 kernel/trace/trace.c len = __copy_from_user_inatomic(&entry->buf, ubuf, cnt); entry 6427 kernel/trace/trace.c memcpy(&entry->buf, FAULTED_STR, FAULTED_SIZE); entry 6436 kernel/trace/trace.c entry->buf[cnt] = '\0'; entry 6437 kernel/trace/trace.c tt = event_triggers_call(tr->trace_marker_file, entry, event); entry 6440 kernel/trace/trace.c if (entry->buf[cnt - 1] != '\n') { entry 6441 kernel/trace/trace.c entry->buf[cnt] = '\n'; entry 6442 kernel/trace/trace.c entry->buf[cnt + 1] = '\0'; entry 6444 kernel/trace/trace.c entry->buf[cnt] = '\0'; entry 6467 kernel/trace/trace.c struct raw_data_entry *entry; entry 6491 kernel/trace/trace.c size = sizeof(*entry) + cnt; entry 6502 kernel/trace/trace.c entry = ring_buffer_event_data(event); entry 6504 kernel/trace/trace.c len = __copy_from_user_inatomic(&entry->id, ubuf, cnt); entry 6506 kernel/trace/trace.c entry->id = -1; entry 6507 kernel/trace/trace.c memcpy(&entry->buf, FAULTED_STR, FAULTED_SIZE); entry 8062 kernel/trace/trace.c topt->entry = trace_create_file(opt->name, 0644, t_options, topt, entry 8122 kernel/trace/trace.c WARN_ONCE(topts[cnt].entry == NULL, entry 369 kernel/trace/trace.h #define IF_ASSIGN(var, entry, etype, id) \ entry 371 kernel/trace/trace.h var = (typeof(var))(entry); \ entry 372 kernel/trace/trace.h WARN_ON(id != 0 && (entry)->type != id); \ entry 443 kernel/trace/trace.h struct dentry *entry; entry 1365 kernel/trace/trace.h struct dentry *entry; entry 1422 kernel/trace/trace.h void *entry, entry 1428 kernel/trace/trace.h *tt = event_triggers_call(file, entry, event); entry 1432 kernel/trace/trace.h !filter_match_preds(file->filter, entry))) { entry 1457 kernel/trace/trace.h void *entry, unsigned long irq_flags, int pc) entry 1461 kernel/trace/trace.h if (!__event_trigger_test_discard(file, buffer, event, entry, &tt)) entry 1488 kernel/trace/trace.h void *entry, unsigned long irq_flags, int pc, entry 1493 kernel/trace/trace.h if (!__event_trigger_test_discard(file, buffer, event, entry, &tt)) entry 37 kernel/trace/trace_branch.c struct trace_branch *entry; entry 65 kernel/trace/trace_branch.c sizeof(*entry), flags, pc); entry 69 kernel/trace/trace_branch.c entry = ring_buffer_event_data(event); entry 77 kernel/trace/trace_branch.c strncpy(entry->func, f->data.func, TRACE_FUNC_SIZE); entry 78 kernel/trace/trace_branch.c strncpy(entry->file, p, TRACE_FILE_SIZE); entry 79 kernel/trace/trace_branch.c entry->func[TRACE_FUNC_SIZE] = 0; entry 80 kernel/trace/trace_branch.c entry->file[TRACE_FILE_SIZE] = 0; entry 81 kernel/trace/trace_branch.c entry->constant = f->constant; entry 82 kernel/trace/trace_branch.c entry->line = f->data.line; entry 83 kernel/trace/trace_branch.c entry->correct = val == expect; entry 85 kernel/trace/trace_branch.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 210 kernel/trace/trace_dynevent.c struct dentry *entry; entry 216 kernel/trace/trace_dynevent.c entry = tracefs_create_file("dynamic_events", 0644, d_tracer, entry 220 kernel/trace/trace_dynevent.c if (!entry) entry 418 kernel/trace/trace_event_perf.c struct trace_entry *entry = record; entry 423 kernel/trace/trace_event_perf.c tracing_generic_entry_update(entry, type, flags, pc); entry 432 kernel/trace/trace_event_perf.c struct ftrace_entry *entry; entry 459 kernel/trace/trace_event_perf.c entry = perf_trace_buf_alloc(ENTRY_SIZE, NULL, &rctx); entry 460 kernel/trace/trace_event_perf.c if (!entry) entry 463 kernel/trace/trace_event_perf.c entry->ip = ip; entry 464 kernel/trace/trace_event_perf.c entry->parent_ip = parent_ip; entry 465 kernel/trace/trace_event_perf.c perf_trace_buf_submit(entry, ENTRY_SIZE, rctx, TRACE_FN, entry 275 kernel/trace/trace_events.c fbuffer->entry = ring_buffer_event_data(fbuffer->event); entry 276 kernel/trace/trace_events.c return fbuffer->entry; entry 700 kernel/trace/trace_events.c tracefs_remove_recursive(dir->entry); entry 1879 kernel/trace/trace_events.c struct dentry *entry; entry 1887 kernel/trace/trace_events.c return dir->entry; entry 1911 kernel/trace/trace_events.c dir->entry = tracefs_create_dir(name, parent); entry 1912 kernel/trace/trace_events.c if (!dir->entry) { entry 1924 kernel/trace/trace_events.c entry = tracefs_create_file("filter", 0644, dir->entry, dir, entry 1926 kernel/trace/trace_events.c if (!entry) { entry 1932 kernel/trace/trace_events.c trace_create_file("enable", 0644, dir->entry, dir, entry 1937 kernel/trace/trace_events.c return dir->entry; entry 2912 kernel/trace/trace_events.c struct dentry *entry; entry 2914 kernel/trace/trace_events.c entry = tracefs_create_file("set_event", 0644, parent, entry 2916 kernel/trace/trace_events.c if (!entry) { entry 2927 kernel/trace/trace_events.c entry = trace_create_file("enable", 0644, d_events, entry 2929 kernel/trace/trace_events.c if (!entry) { entry 2936 kernel/trace/trace_events.c entry = tracefs_create_file("set_event_pid", 0644, parent, entry 2938 kernel/trace/trace_events.c if (!entry) entry 2942 kernel/trace/trace_events.c entry = trace_create_file("header_page", 0444, d_events, entry 2945 kernel/trace/trace_events.c if (!entry) entry 2948 kernel/trace/trace_events.c entry = trace_create_file("header_event", 0444, d_events, entry 2951 kernel/trace/trace_events.c if (!entry) entry 3143 kernel/trace/trace_events.c struct dentry *entry; entry 3154 kernel/trace/trace_events.c entry = tracefs_create_file("available_events", 0444, d_tracer, entry 3156 kernel/trace/trace_events.c if (!entry) entry 3363 kernel/trace/trace_events.c struct ftrace_entry *entry; entry 3380 kernel/trace/trace_events.c TRACE_FN, sizeof(*entry), entry 3384 kernel/trace/trace_events.c entry = ring_buffer_event_data(event); entry 3385 kernel/trace/trace_events.c entry->ip = ip; entry 3386 kernel/trace/trace_events.c entry->parent_ip = parent_ip; entry 3389 kernel/trace/trace_events.c entry, flags, pc); entry 843 kernel/trace/trace_events_hist.c struct synth_trace_event *entry; entry 849 kernel/trace/trace_events_hist.c entry = (struct synth_trace_event *)iter->ent; entry 869 kernel/trace/trace_events_hist.c (char *)&entry->fields[n_u64], entry 880 kernel/trace/trace_events_hist.c entry->fields[n_u64], entry 886 kernel/trace/trace_events_hist.c entry->fields[n_u64], entry 908 kernel/trace/trace_events_hist.c struct synth_trace_event *entry; entry 929 kernel/trace/trace_events_hist.c entry = trace_event_buffer_reserve(&fbuffer, trace_file, entry 930 kernel/trace/trace_events_hist.c sizeof(*entry) + fields_size); entry 931 kernel/trace/trace_events_hist.c if (!entry) entry 938 kernel/trace/trace_events_hist.c char *str_field = (char *)&entry->fields[n_u64]; entry 948 kernel/trace/trace_events_hist.c *(u8 *)&entry->fields[n_u64] = (u8)val; entry 952 kernel/trace/trace_events_hist.c *(u16 *)&entry->fields[n_u64] = (u16)val; entry 956 kernel/trace/trace_events_hist.c *(u32 *)&entry->fields[n_u64] = (u32)val; entry 960 kernel/trace/trace_events_hist.c entry->fields[n_u64] = val; entry 6535 kernel/trace/trace_events_hist.c struct dentry *entry = NULL; entry 6551 kernel/trace/trace_events_hist.c entry = tracefs_create_file("synthetic_events", 0644, d_tracer, entry 6553 kernel/trace/trace_events_hist.c if (!entry) { entry 105 kernel/trace/trace_functions_graph.c struct ftrace_graph_ent_entry *entry; entry 108 kernel/trace/trace_functions_graph.c sizeof(*entry), flags, pc); entry 111 kernel/trace/trace_functions_graph.c entry = ring_buffer_event_data(event); entry 112 kernel/trace/trace_functions_graph.c entry->graph_ent = *trace; entry 113 kernel/trace/trace_functions_graph.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 225 kernel/trace/trace_functions_graph.c struct ftrace_graph_ret_entry *entry; entry 228 kernel/trace/trace_functions_graph.c sizeof(*entry), flags, pc); entry 231 kernel/trace/trace_functions_graph.c entry = ring_buffer_event_data(event); entry 232 kernel/trace/trace_functions_graph.c entry->ret = *trace; entry 233 kernel/trace/trace_functions_graph.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 376 kernel/trace/trace_functions_graph.c static void print_graph_lat_fmt(struct trace_seq *s, struct trace_entry *entry) entry 379 kernel/trace/trace_functions_graph.c trace_print_lat_fmt(s, entry); entry 627 kernel/trace/trace_functions_graph.c struct ftrace_graph_ent_entry *entry, entry 640 kernel/trace/trace_functions_graph.c call = &entry->graph_ent; entry 678 kernel/trace/trace_functions_graph.c struct ftrace_graph_ent_entry *entry, entry 681 kernel/trace/trace_functions_graph.c struct ftrace_graph_ent *call = &entry->graph_ent; entry 1043 kernel/trace/trace_functions_graph.c struct trace_entry *entry = iter->ent; entry 1069 kernel/trace/trace_functions_graph.c switch (entry->type) { entry 1078 kernel/trace/trace_functions_graph.c trace_assign_type(field, entry); entry 1084 kernel/trace/trace_functions_graph.c trace_assign_type(field, entry); entry 1085 kernel/trace/trace_functions_graph.c return print_graph_return(&field->ret, s, entry, iter, flags); entry 1093 kernel/trace/trace_functions_graph.c return print_graph_comment(s, entry, iter, flags); entry 109 kernel/trace/trace_hwlat.c struct hwlat_entry *entry; entry 116 kernel/trace/trace_hwlat.c event = trace_buffer_lock_reserve(buffer, TRACE_HWLAT, sizeof(*entry), entry 120 kernel/trace/trace_hwlat.c entry = ring_buffer_event_data(event); entry 121 kernel/trace/trace_hwlat.c entry->seqnum = sample->seqnum; entry 122 kernel/trace/trace_hwlat.c entry->duration = sample->duration; entry 123 kernel/trace/trace_hwlat.c entry->outer_duration = sample->outer_duration; entry 124 kernel/trace/trace_hwlat.c entry->timestamp = sample->timestamp; entry 125 kernel/trace/trace_hwlat.c entry->nmi_total_ts = sample->nmi_total_ts; entry 126 kernel/trace/trace_hwlat.c entry->nmi_count = sample->nmi_count; entry 128 kernel/trace/trace_hwlat.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 411 kernel/trace/trace_hwlat.c u64 *entry = filp->private_data; entry 415 kernel/trace/trace_hwlat.c if (!entry) entry 421 kernel/trace/trace_hwlat.c val = *entry; entry 1179 kernel/trace/trace_kprobe.c struct kprobe_trace_entry_head *entry; entry 1195 kernel/trace/trace_kprobe.c size = sizeof(*entry) + tk->tp.size + dsize; entry 1203 kernel/trace/trace_kprobe.c entry = ring_buffer_event_data(event); entry 1204 kernel/trace/trace_kprobe.c entry->ip = (unsigned long)tk->rp.kp.addr; entry 1205 kernel/trace/trace_kprobe.c store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize); entry 1208 kernel/trace/trace_kprobe.c entry, irq_flags, pc, regs); entry 1227 kernel/trace/trace_kprobe.c struct kretprobe_trace_entry_head *entry; entry 1243 kernel/trace/trace_kprobe.c size = sizeof(*entry) + tk->tp.size + dsize; entry 1251 kernel/trace/trace_kprobe.c entry = ring_buffer_event_data(event); entry 1252 kernel/trace/trace_kprobe.c entry->func = (unsigned long)tk->rp.kp.addr; entry 1253 kernel/trace/trace_kprobe.c entry->ret_ip = (unsigned long)ri->ret_addr; entry 1254 kernel/trace/trace_kprobe.c store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize); entry 1257 kernel/trace/trace_kprobe.c entry, irq_flags, pc, regs); entry 1377 kernel/trace/trace_kprobe.c struct kprobe_trace_entry_head *entry; entry 1404 kernel/trace/trace_kprobe.c __size = sizeof(*entry) + tk->tp.size + dsize; entry 1408 kernel/trace/trace_kprobe.c entry = perf_trace_buf_alloc(size, NULL, &rctx); entry 1409 kernel/trace/trace_kprobe.c if (!entry) entry 1412 kernel/trace/trace_kprobe.c entry->ip = (unsigned long)tk->rp.kp.addr; entry 1413 kernel/trace/trace_kprobe.c memset(&entry[1], 0, dsize); entry 1414 kernel/trace/trace_kprobe.c store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize); entry 1415 kernel/trace/trace_kprobe.c perf_trace_buf_submit(entry, size, rctx, call->event.type, 1, regs, entry 1427 kernel/trace/trace_kprobe.c struct kretprobe_trace_entry_head *entry; entry 1440 kernel/trace/trace_kprobe.c __size = sizeof(*entry) + tk->tp.size + dsize; entry 1444 kernel/trace/trace_kprobe.c entry = perf_trace_buf_alloc(size, NULL, &rctx); entry 1445 kernel/trace/trace_kprobe.c if (!entry) entry 1448 kernel/trace/trace_kprobe.c entry->func = (unsigned long)tk->rp.kp.addr; entry 1449 kernel/trace/trace_kprobe.c entry->ret_ip = (unsigned long)ri->ret_addr; entry 1450 kernel/trace/trace_kprobe.c store_trace_args(&entry[1], &tk->tp, regs, sizeof(*entry), dsize); entry 1451 kernel/trace/trace_kprobe.c perf_trace_buf_submit(entry, size, rctx, call->event.type, 1, regs, entry 1695 kernel/trace/trace_kprobe.c struct dentry *entry; entry 1709 kernel/trace/trace_kprobe.c entry = tracefs_create_file("kprobe_events", 0644, d_tracer, entry 1713 kernel/trace/trace_kprobe.c if (!entry) entry 1717 kernel/trace/trace_kprobe.c entry = tracefs_create_file("kprobe_profile", 0444, d_tracer, entry 1720 kernel/trace/trace_kprobe.c if (!entry) entry 169 kernel/trace/trace_mmiotrace.c struct trace_entry *entry = iter->ent; entry 177 kernel/trace/trace_mmiotrace.c trace_assign_type(field, entry); entry 214 kernel/trace/trace_mmiotrace.c struct trace_entry *entry = iter->ent; entry 222 kernel/trace/trace_mmiotrace.c trace_assign_type(field, entry); entry 248 kernel/trace/trace_mmiotrace.c struct trace_entry *entry = iter->ent; entry 249 kernel/trace/trace_mmiotrace.c struct print_entry *print = (struct print_entry *)entry; entry 302 kernel/trace/trace_mmiotrace.c struct trace_mmiotrace_rw *entry; entry 306 kernel/trace/trace_mmiotrace.c sizeof(*entry), 0, pc); entry 311 kernel/trace/trace_mmiotrace.c entry = ring_buffer_event_data(event); entry 312 kernel/trace/trace_mmiotrace.c entry->rw = *rw; entry 314 kernel/trace/trace_mmiotrace.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 332 kernel/trace/trace_mmiotrace.c struct trace_mmiotrace_map *entry; entry 336 kernel/trace/trace_mmiotrace.c sizeof(*entry), 0, pc); entry 341 kernel/trace/trace_mmiotrace.c entry = ring_buffer_event_data(event); entry 342 kernel/trace/trace_mmiotrace.c entry->map = *map; entry 344 kernel/trace/trace_mmiotrace.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 28 kernel/trace/trace_output.c struct trace_entry *entry = iter->ent; entry 31 kernel/trace/trace_output.c trace_assign_type(field, entry); entry 41 kernel/trace/trace_output.c struct trace_entry *entry = iter->ent; entry 44 kernel/trace/trace_output.c trace_assign_type(field, entry); entry 54 kernel/trace/trace_output.c struct trace_entry *entry = iter->ent; entry 57 kernel/trace/trace_output.c trace_assign_type(field, entry); entry 283 kernel/trace/trace_output.c struct trace_entry *entry; entry 286 kernel/trace/trace_output.c entry = iter->ent; entry 288 kernel/trace/trace_output.c if (entry->type != event->event.type) { entry 425 kernel/trace/trace_output.c int trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry) entry 434 kernel/trace/trace_output.c nmi = entry->flags & TRACE_FLAG_NMI; entry 435 kernel/trace/trace_output.c hardirq = entry->flags & TRACE_FLAG_HARDIRQ; entry 436 kernel/trace/trace_output.c softirq = entry->flags & TRACE_FLAG_SOFTIRQ; entry 439 kernel/trace/trace_output.c (entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' : entry 440 kernel/trace/trace_output.c (entry->flags & TRACE_FLAG_IRQS_NOSUPPORT) ? 'X' : entry 443 kernel/trace/trace_output.c switch (entry->flags & (TRACE_FLAG_NEED_RESCHED | entry 470 kernel/trace/trace_output.c if (entry->preempt_count) entry 471 kernel/trace/trace_output.c trace_seq_printf(s, "%x", entry->preempt_count); entry 479 kernel/trace/trace_output.c lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu) entry 483 kernel/trace/trace_output.c trace_find_cmdline(entry->pid, comm); entry 486 kernel/trace/trace_output.c comm, entry->pid, cpu); entry 488 kernel/trace/trace_output.c return trace_print_lat_fmt(s, entry); entry 569 kernel/trace/trace_output.c struct trace_entry *entry = iter->ent; entry 574 kernel/trace/trace_output.c trace_find_cmdline(entry->pid, comm); entry 576 kernel/trace/trace_output.c trace_seq_printf(s, "%16s-%-5d ", comm, entry->pid); entry 579 kernel/trace/trace_output.c unsigned int tgid = trace_find_tgid(entry->pid); entry 590 kernel/trace/trace_output.c trace_print_lat_fmt(s, entry); entry 610 kernel/trace/trace_output.c struct trace_entry *entry = iter->ent, entry 624 kernel/trace/trace_output.c trace_find_cmdline(entry->pid, comm); entry 628 kernel/trace/trace_output.c comm, entry->pid, iter->cpu, entry->flags, entry 629 kernel/trace/trace_output.c entry->preempt_count, iter->idx); entry 631 kernel/trace/trace_output.c lat_print_generic(s, entry, iter->cpu); entry 1140 kernel/trace/trace_output.c struct trace_entry *entry = iter->ent; entry 1144 kernel/trace/trace_output.c trace_assign_type(field, entry); entry 1205 kernel/trace/trace_output.c struct trace_entry *entry = iter->ent; entry 1209 kernel/trace/trace_output.c trace_assign_type(field, entry); entry 1249 kernel/trace/trace_output.c struct trace_entry *entry = iter->ent; entry 1253 kernel/trace/trace_output.c trace_assign_type(field, entry); entry 29 kernel/trace/trace_output.h trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry); entry 383 kernel/trace/trace_sched_wakeup.c struct ctx_switch_entry *entry; entry 386 kernel/trace/trace_sched_wakeup.c sizeof(*entry), flags, pc); entry 389 kernel/trace/trace_sched_wakeup.c entry = ring_buffer_event_data(event); entry 390 kernel/trace/trace_sched_wakeup.c entry->prev_pid = prev->pid; entry 391 kernel/trace/trace_sched_wakeup.c entry->prev_prio = prev->prio; entry 392 kernel/trace/trace_sched_wakeup.c entry->prev_state = task_state_index(prev); entry 393 kernel/trace/trace_sched_wakeup.c entry->next_pid = next->pid; entry 394 kernel/trace/trace_sched_wakeup.c entry->next_prio = next->prio; entry 395 kernel/trace/trace_sched_wakeup.c entry->next_state = task_state_index(next); entry 396 kernel/trace/trace_sched_wakeup.c entry->next_cpu = task_cpu(next); entry 398 kernel/trace/trace_sched_wakeup.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 410 kernel/trace/trace_sched_wakeup.c struct ctx_switch_entry *entry; entry 414 kernel/trace/trace_sched_wakeup.c sizeof(*entry), flags, pc); entry 417 kernel/trace/trace_sched_wakeup.c entry = ring_buffer_event_data(event); entry 418 kernel/trace/trace_sched_wakeup.c entry->prev_pid = curr->pid; entry 419 kernel/trace/trace_sched_wakeup.c entry->prev_prio = curr->prio; entry 420 kernel/trace/trace_sched_wakeup.c entry->prev_state = task_state_index(curr); entry 421 kernel/trace/trace_sched_wakeup.c entry->next_pid = wakee->pid; entry 422 kernel/trace/trace_sched_wakeup.c entry->next_prio = wakee->prio; entry 423 kernel/trace/trace_sched_wakeup.c entry->next_state = task_state_index(wakee); entry 424 kernel/trace/trace_sched_wakeup.c entry->next_cpu = task_cpu(wakee); entry 426 kernel/trace/trace_sched_wakeup.c if (!call_filter_check_discard(call, entry, buffer, event)) entry 10 kernel/trace/trace_selftest.c static inline int trace_valid_entry(struct trace_entry *entry) entry 12 kernel/trace/trace_selftest.c switch (entry->type) { entry 29 kernel/trace/trace_selftest.c struct trace_entry *entry; entry 33 kernel/trace/trace_selftest.c entry = ring_buffer_event_data(event); entry 44 kernel/trace/trace_selftest.c if (!trace_valid_entry(entry)) { entry 46 kernel/trace/trace_selftest.c entry->type); entry 25 kernel/trace/trace_syscalls.c struct syscall_metadata *entry = call->data; entry 27 kernel/trace/trace_syscalls.c return &entry->enter_fields; entry 112 kernel/trace/trace_syscalls.c struct syscall_metadata *entry; entry 114 kernel/trace/trace_syscalls.c entry = syscall_nr_to_meta(syscall); entry 115 kernel/trace/trace_syscalls.c if (!entry) entry 118 kernel/trace/trace_syscalls.c return entry->name; entry 129 kernel/trace/trace_syscalls.c struct syscall_metadata *entry; entry 134 kernel/trace/trace_syscalls.c entry = syscall_nr_to_meta(syscall); entry 136 kernel/trace/trace_syscalls.c if (!entry) entry 139 kernel/trace/trace_syscalls.c if (entry->enter_event->event.type != ent->type) { entry 144 kernel/trace/trace_syscalls.c trace_seq_printf(s, "%s(", entry->name); entry 146 kernel/trace/trace_syscalls.c for (i = 0; i < entry->nb_args; i++) { entry 153 kernel/trace/trace_syscalls.c trace_seq_printf(s, "%s ", entry->types[i]); entry 156 kernel/trace/trace_syscalls.c trace_seq_printf(s, "%s: %lx%s", entry->args[i], entry 158 kernel/trace/trace_syscalls.c i == entry->nb_args - 1 ? "" : ", "); entry 176 kernel/trace/trace_syscalls.c struct syscall_metadata *entry; entry 180 kernel/trace/trace_syscalls.c entry = syscall_nr_to_meta(syscall); entry 182 kernel/trace/trace_syscalls.c if (!entry) { entry 187 kernel/trace/trace_syscalls.c if (entry->exit_event->event.type != ent->type) { entry 192 kernel/trace/trace_syscalls.c trace_seq_printf(s, "%s -> 0x%lx\n", entry->name, entry 208 kernel/trace/trace_syscalls.c __set_enter_print_fmt(struct syscall_metadata *entry, char *buf, int len) entry 217 kernel/trace/trace_syscalls.c for (i = 0; i < entry->nb_args; i++) { entry 219 kernel/trace/trace_syscalls.c entry->args[i], sizeof(unsigned long), entry 220 kernel/trace/trace_syscalls.c i == entry->nb_args - 1 ? "" : ", "); entry 224 kernel/trace/trace_syscalls.c for (i = 0; i < entry->nb_args; i++) { entry 226 kernel/trace/trace_syscalls.c ", ((unsigned long)(REC->%s))", entry->args[i]); entry 239 kernel/trace/trace_syscalls.c struct syscall_metadata *entry = call->data; entry 241 kernel/trace/trace_syscalls.c if (entry->enter_event != call) { entry 247 kernel/trace/trace_syscalls.c len = __set_enter_print_fmt(entry, NULL, 0); entry 254 kernel/trace/trace_syscalls.c __set_enter_print_fmt(entry, print_fmt, len + 1); entry 262 kernel/trace/trace_syscalls.c struct syscall_metadata *entry = call->data; entry 264 kernel/trace/trace_syscalls.c if (entry->enter_event == call) entry 312 kernel/trace/trace_syscalls.c struct syscall_trace_enter *entry; entry 338 kernel/trace/trace_syscalls.c size = sizeof(*entry) + sizeof(unsigned long) * sys_data->nb_args; entry 349 kernel/trace/trace_syscalls.c entry = ring_buffer_event_data(event); entry 350 kernel/trace/trace_syscalls.c entry->nr = syscall_nr; entry 352 kernel/trace/trace_syscalls.c memcpy(entry->args, args, sizeof(unsigned long) * sys_data->nb_args); entry 354 kernel/trace/trace_syscalls.c event_trigger_unlock_commit(trace_file, buffer, event, entry, entry 362 kernel/trace/trace_syscalls.c struct syscall_trace_exit *entry; entry 391 kernel/trace/trace_syscalls.c sys_data->exit_event->event.type, sizeof(*entry), entry 396 kernel/trace/trace_syscalls.c entry = ring_buffer_event_data(event); entry 397 kernel/trace/trace_syscalls.c entry->nr = syscall_nr; entry 398 kernel/trace/trace_syscalls.c entry->ret = syscall_get_return_value(current, regs); entry 400 kernel/trace/trace_syscalls.c event_trigger_unlock_commit(trace_file, buffer, event, entry, entry 34 kernel/trace/trace_uprobe.c #define DATAOF_TRACE_ENTRY(entry, is_return) \ entry 35 kernel/trace/trace_uprobe.c ((void*)(entry) + SIZEOF_TRACE_ENTRY(is_return)) entry 933 kernel/trace/trace_uprobe.c struct uprobe_trace_entry_head *entry; entry 955 kernel/trace/trace_uprobe.c entry = ring_buffer_event_data(event); entry 957 kernel/trace/trace_uprobe.c entry->vaddr[0] = func; entry 958 kernel/trace/trace_uprobe.c entry->vaddr[1] = instruction_pointer(regs); entry 959 kernel/trace/trace_uprobe.c data = DATAOF_TRACE_ENTRY(entry, true); entry 961 kernel/trace/trace_uprobe.c entry->vaddr[0] = instruction_pointer(regs); entry 962 kernel/trace/trace_uprobe.c data = DATAOF_TRACE_ENTRY(entry, false); entry 967 kernel/trace/trace_uprobe.c event_trigger_unlock_commit(trace_file, buffer, event, entry, 0, 0); entry 1003 kernel/trace/trace_uprobe.c struct uprobe_trace_entry_head *entry; entry 1008 kernel/trace/trace_uprobe.c entry = (struct uprobe_trace_entry_head *)iter->ent; entry 1017 kernel/trace/trace_uprobe.c entry->vaddr[1], entry->vaddr[0]); entry 1018 kernel/trace/trace_uprobe.c data = DATAOF_TRACE_ENTRY(entry, true); entry 1022 kernel/trace/trace_uprobe.c entry->vaddr[0]); entry 1023 kernel/trace/trace_uprobe.c data = DATAOF_TRACE_ENTRY(entry, false); entry 1026 kernel/trace/trace_uprobe.c if (print_probe_args(s, tu->tp.args, tu->tp.nr_args, data, entry) < 0) entry 1330 kernel/trace/trace_uprobe.c struct uprobe_trace_entry_head *entry; entry 1351 kernel/trace/trace_uprobe.c entry = perf_trace_buf_alloc(size, NULL, &rctx); entry 1352 kernel/trace/trace_uprobe.c if (!entry) entry 1356 kernel/trace/trace_uprobe.c entry->vaddr[0] = func; entry 1357 kernel/trace/trace_uprobe.c entry->vaddr[1] = instruction_pointer(regs); entry 1358 kernel/trace/trace_uprobe.c data = DATAOF_TRACE_ENTRY(entry, true); entry 1360 kernel/trace/trace_uprobe.c entry->vaddr[0] = instruction_pointer(regs); entry 1361 kernel/trace/trace_uprobe.c data = DATAOF_TRACE_ENTRY(entry, false); entry 1372 kernel/trace/trace_uprobe.c perf_trace_buf_submit(entry, size, rctx, call->event.type, 1, regs, entry 517 kernel/trace/tracing_map.c struct tracing_map_entry *entry; entry 527 kernel/trace/tracing_map.c entry = TRACING_MAP_ENTRY(map->map, idx); entry 528 kernel/trace/tracing_map.c test_key = entry->key; entry 531 kernel/trace/tracing_map.c val = READ_ONCE(entry->val); entry 563 kernel/trace/tracing_map.c if (!cmpxchg(&entry->key, 0, key_hash)) { entry 569 kernel/trace/tracing_map.c entry->key = 0; entry 574 kernel/trace/tracing_map.c entry->val = elt; entry 577 kernel/trace/tracing_map.c return entry->val; entry 905 kernel/trace/tracing_map.c static void destroy_sort_entry(struct tracing_map_sort_entry *entry) entry 907 kernel/trace/tracing_map.c if (!entry) entry 910 kernel/trace/tracing_map.c if (entry->elt_copied) entry 911 kernel/trace/tracing_map.c tracing_map_elt_free(entry->elt); entry 913 kernel/trace/tracing_map.c kfree(entry); entry 1074 kernel/trace/tracing_map.c struct tracing_map_entry *entry; entry 1076 kernel/trace/tracing_map.c entry = TRACING_MAP_ENTRY(map->map, i); entry 1078 kernel/trace/tracing_map.c if (!entry->key || !entry->val) entry 1081 kernel/trace/tracing_map.c entries[n_entries] = create_sort_entry(entry->val->key, entry 1082 kernel/trace/tracing_map.c entry->val); entry 825 kernel/workqueue.c return list_first_entry(&pool->idle_list, struct worker, entry); entry 1068 kernel/workqueue.c list_for_each_entry_safe_from(work, n, NULL, entry) { entry 1069 kernel/workqueue.c list_move_tail(&work->entry, head); entry 1156 kernel/workqueue.c struct work_struct, entry); entry 1292 kernel/workqueue.c list_del_init(&work->entry); entry 1332 kernel/workqueue.c list_add_tail(&work->entry, head); entry 1477 kernel/workqueue.c if (WARN_ON(!list_empty(&work->entry))) entry 1632 kernel/workqueue.c WARN_ON_ONCE(!list_empty(&work->entry)); entry 1773 kernel/workqueue.c WARN_ON_ONCE(!list_empty(&worker->entry) && entry 1783 kernel/workqueue.c list_add(&worker->entry, &pool->idle_list); entry 1816 kernel/workqueue.c list_del_init(&worker->entry); entry 1825 kernel/workqueue.c INIT_LIST_HEAD(&worker->entry); entry 1984 kernel/workqueue.c list_del_init(&worker->entry); entry 2000 kernel/workqueue.c worker = list_entry(pool->idle_list.prev, struct worker, entry); entry 2052 kernel/workqueue.c list_for_each_entry(work, &pool->worklist, entry) entry 2216 kernel/workqueue.c list_del_init(&work->entry); entry 2333 kernel/workqueue.c struct work_struct, entry); entry 2373 kernel/workqueue.c WARN_ON_ONCE(!list_empty(&worker->entry)); entry 2412 kernel/workqueue.c struct work_struct, entry); entry 2515 kernel/workqueue.c list_for_each_entry_safe(work, n, &pool->worklist, entry) { entry 2682 kernel/workqueue.c head = target->entry.next; entry 4687 kernel/workqueue.c list_for_each_entry(work, &worker->scheduled, entry) entry 4694 kernel/workqueue.c list_for_each_entry(work, &pool->worklist, entry) { entry 4704 kernel/workqueue.c list_for_each_entry(work, &pool->worklist, entry) { entry 4718 kernel/workqueue.c list_for_each_entry(work, &pwq->delayed_works, entry) { entry 4788 kernel/workqueue.c list_for_each_entry(worker, &pool->idle_list, entry) { entry 27 kernel/workqueue_internal.h struct list_head entry; /* L: while idle */ entry 63 lib/error-inject.c unsigned long entry, offset = 0, size = 0; entry 67 lib/error-inject.c entry = arch_deref_entry_point((void *)iter->addr); entry 69 lib/error-inject.c if (!kernel_text_address(entry) || entry 70 lib/error-inject.c !kallsyms_lookup_size_offset(entry, &size, &offset)) { entry 72 lib/error-inject.c (void *)entry); entry 79 lib/error-inject.c ent->start_addr = entry; entry 80 lib/error-inject.c ent->end_addr = entry + size; entry 231 lib/idr.c void *entry = NULL; entry 237 lib/idr.c entry = rcu_dereference_raw(*slot); entry 238 lib/idr.c if (!entry) entry 240 lib/idr.c if (!xa_is_internal(entry)) entry 242 lib/idr.c if (slot != &idr->idr_rt.xa_head && !xa_is_retry(entry)) entry 250 lib/idr.c return entry; entry 267 lib/idr.c void *entry = idr_get_next_ul(idr, &id); entry 272 lib/idr.c return entry; entry 294 lib/idr.c void *entry; entry 298 lib/idr.c entry = __radix_tree_lookup(&idr->idr_rt, id, &node, &slot); entry 304 lib/idr.c return entry; entry 556 lib/idr.c static void ida_dump_entry(void *entry, unsigned long index) entry 560 lib/idr.c if (!entry) entry 563 lib/idr.c if (xa_is_node(entry)) { entry 564 lib/idr.c struct xa_node *node = xa_to_node(entry); entry 573 lib/idr.c } else if (xa_is_value(entry)) { entry 575 lib/idr.c pr_cont("value: data %lx [%px]\n", xa_to_value(entry), entry); entry 577 lib/idr.c struct ida_bitmap *bitmap = entry; entry 196 lib/kobject.c list_add_tail(&kobj->entry, &kobj->kset->list); entry 207 lib/kobject.c list_del_init(&kobj->entry); entry 217 lib/kobject.c INIT_LIST_HEAD(&kobj->entry); entry 901 lib/kobject.c list_for_each_entry(k, &kset->list, entry) { entry 38 lib/list_debug.c bool __list_del_entry_valid(struct list_head *entry) entry 42 lib/list_debug.c prev = entry->prev; entry 43 lib/list_debug.c next = entry->next; entry 47 lib/list_debug.c entry, LIST_POISON1) || entry 50 lib/list_debug.c entry, LIST_POISON2) || entry 51 lib/list_debug.c CHECK_DATA_CORRUPTION(prev->next != entry, entry 53 lib/list_debug.c entry, prev->next) || entry 54 lib/list_debug.c CHECK_DATA_CORRUPTION(next->prev != entry, entry 56 lib/list_debug.c entry, next->prev)) entry 55 lib/llist.c struct llist_node *entry, *old_entry, *next; entry 57 lib/llist.c entry = smp_load_acquire(&head->first); entry 59 lib/llist.c if (entry == NULL) entry 61 lib/llist.c old_entry = entry; entry 62 lib/llist.c next = READ_ONCE(entry->next); entry 63 lib/llist.c entry = cmpxchg(&head->first, old_entry, next); entry 64 lib/llist.c if (entry == old_entry) entry 68 lib/llist.c return entry; entry 238 lib/logic_pio.c struct logic_pio_hwaddr *entry = find_io_range(addr); \ entry 240 lib/logic_pio.c if (entry && entry->ops) \ entry 241 lib/logic_pio.c ret = entry->ops->in(entry->hostdata, \ entry 254 lib/logic_pio.c struct logic_pio_hwaddr *entry = find_io_range(addr); \ entry 256 lib/logic_pio.c if (entry && entry->ops) \ entry 257 lib/logic_pio.c entry->ops->out(entry->hostdata, \ entry 270 lib/logic_pio.c struct logic_pio_hwaddr *entry = find_io_range(addr); \ entry 272 lib/logic_pio.c if (entry && entry->ops) \ entry 273 lib/logic_pio.c entry->ops->ins(entry->hostdata, \ entry 287 lib/logic_pio.c struct logic_pio_hwaddr *entry = find_io_range(addr); \ entry 289 lib/logic_pio.c if (entry && entry->ops) \ entry 290 lib/logic_pio.c entry->ops->outs(entry->hostdata, \ entry 75 lib/nlattr.c const struct nlattr *entry; entry 78 lib/nlattr.c nla_for_each_attr(entry, head, len, rem) { entry 81 lib/nlattr.c if (nla_len(entry) == 0) entry 84 lib/nlattr.c if (nla_len(entry) < NLA_HDRLEN) { entry 85 lib/nlattr.c NL_SET_ERR_MSG_ATTR(extack, entry, entry 90 lib/nlattr.c ret = __nla_validate(nla_data(entry), nla_len(entry), entry 98 lib/radix-tree.c void __rcu **entry = rcu_dereference_raw(parent->slots[offset]); entry 100 lib/radix-tree.c *nodep = (void *)entry; entry 420 lib/radix-tree.c void *entry; entry 429 lib/radix-tree.c entry = rcu_dereference_raw(root->xa_head); entry 430 lib/radix-tree.c if (!entry && (!is_idr(root) || root_tag_get(root, IDR_FREE))) entry 454 lib/radix-tree.c if (radix_tree_is_internal_node(entry)) { entry 455 lib/radix-tree.c entry_to_node(entry)->parent = node; entry 456 lib/radix-tree.c } else if (xa_is_value(entry)) { entry 464 lib/radix-tree.c node->slots[0] = (void __rcu *)entry; entry 465 lib/radix-tree.c entry = node_to_entry(node); entry 466 lib/radix-tree.c rcu_assign_pointer(root->xa_head, entry); entry 671 lib/radix-tree.c void *entry = rcu_dereference_raw(child->slots[offset]); entry 672 lib/radix-tree.c if (xa_is_node(entry) && child->shift) { entry 673 lib/radix-tree.c child = entry_to_node(entry); entry 1427 lib/radix-tree.c void *entry; entry 1429 lib/radix-tree.c entry = __radix_tree_lookup(root, index, &node, &slot); entry 1432 lib/radix-tree.c if (!entry && (!is_idr(root) || node_tag_get(root, node, IDR_FREE, entry 1436 lib/radix-tree.c if (item && entry != item) entry 1441 lib/radix-tree.c return entry; entry 222 lib/rhashtable.c struct rhash_head *head, *next, *entry; entry 231 lib/rhashtable.c rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash), entry 234 lib/rhashtable.c next = rht_dereference_bucket(entry->next, old_tbl, old_hash); entry 239 lib/rhashtable.c pprev = &entry->next; entry 245 lib/rhashtable.c new_hash = head_hashfn(ht, new_tbl, entry); entry 251 lib/rhashtable.c RCU_INIT_POINTER(entry->next, head); entry 253 lib/rhashtable.c rht_assign_unlock(new_tbl, &new_tbl->buckets[new_hash], entry); entry 677 lib/sbitmap.c list_del_init(&sbq_wait->wait.entry); entry 72 lib/test_xarray.c unsigned order, void *entry, gfp_t gfp) entry 79 lib/test_xarray.c curr = xas_store(&xas, entry); entry 105 lib/test_xarray.c void *entry; entry 138 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 153 lib/test_xarray.c void *entry = xa_load(xa, j); entry 155 lib/test_xarray.c XA_BUG_ON(xa, xa_to_value(entry) != j); entry 157 lib/test_xarray.c XA_BUG_ON(xa, entry); entry 164 lib/test_xarray.c void *entry = xa_load(xa, j); entry 166 lib/test_xarray.c XA_BUG_ON(xa, xa_to_value(entry) != j); entry 168 lib/test_xarray.c XA_BUG_ON(xa, entry); entry 222 lib/test_xarray.c void *entry; entry 230 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) entry 239 lib/test_xarray.c xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0) entry 259 lib/test_xarray.c void *entry; entry 276 lib/test_xarray.c xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0) entry 282 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 403 lib/test_xarray.c void *entry; entry 444 lib/test_xarray.c xa_for_each(xa, index, entry) { entry 470 lib/test_xarray.c void *entry; entry 495 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 496 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(j)); entry 549 lib/test_xarray.c void *entry; entry 555 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 556 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(index)); entry 561 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 562 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(index)); entry 626 lib/test_xarray.c void *entry = xa_load(xa, (1UL << k) - 1); entry 628 lib/test_xarray.c XA_BUG_ON(xa, entry != NULL); entry 630 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(j)); entry 731 lib/test_xarray.c void *entry; entry 762 lib/test_xarray.c xa_for_each(xa, index, entry) { entry 784 lib/test_xarray.c void *entry; entry 800 lib/test_xarray.c entry = xa_mk_index(i); entry 802 lib/test_xarray.c entry = xa_mk_index(i - 0x3fff); entry 803 lib/test_xarray.c XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, entry, limit, entry 805 lib/test_xarray.c XA_BUG_ON(xa, xa_mk_index(id) != entry); entry 823 lib/test_xarray.c xa_for_each(xa, index, entry) entry 846 lib/test_xarray.c void *entry; entry 851 lib/test_xarray.c xas_for_each_conflict(&xas, entry) { entry 852 lib/test_xarray.c XA_BUG_ON(xa, !xa_is_value(entry)); entry 853 lib/test_xarray.c XA_BUG_ON(xa, entry < xa_mk_index(start)); entry 854 lib/test_xarray.c XA_BUG_ON(xa, entry > xa_mk_index(start + (1UL << order) - 1)); entry 945 lib/test_xarray.c void *entry; entry 955 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 998 lib/test_xarray.c void *entry = xa_find(xa, &index, ULONG_MAX, entry 1005 lib/test_xarray.c XA_BUG_ON(xa, entry != NULL); entry 1008 lib/test_xarray.c entry = xa_find(xa, &index, ULONG_MAX, entry 1015 lib/test_xarray.c XA_BUG_ON(xa, entry != NULL); entry 1029 lib/test_xarray.c void *entry; entry 1032 lib/test_xarray.c xa_for_each(xa, index, entry) { entry 1039 lib/test_xarray.c xa_for_each(xa, index, entry) { entry 1040 lib/test_xarray.c XA_BUG_ON(xa, xa_mk_index(index) != entry); entry 1052 lib/test_xarray.c void *entry; entry 1059 lib/test_xarray.c xas_for_each_marked(&xas, entry, k, XA_MARK_0) entry 1076 lib/test_xarray.c void *entry; entry 1080 lib/test_xarray.c entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT); entry 1081 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(ULONG_MAX)); entry 1083 lib/test_xarray.c entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT); entry 1084 lib/test_xarray.c XA_BUG_ON(xa, entry); entry 1109 lib/test_xarray.c void *entry; entry 1112 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 1113 lib/test_xarray.c if (xas_retry(&xas, entry)) entry 1115 lib/test_xarray.c if (entry == item) entry 1124 lib/test_xarray.c return entry ? xas.xa_index : -1; entry 1162 lib/test_xarray.c void *entry; entry 1174 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 1175 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(1UL << count)); entry 1184 lib/test_xarray.c xas_for_each(&xas, entry, ULONG_MAX) { entry 1185 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(1UL << count)); entry 1248 lib/test_xarray.c void *entry = xas_next(&xas); entry 1253 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(i)); entry 1255 lib/test_xarray.c XA_BUG_ON(xa, entry != NULL); entry 1261 lib/test_xarray.c void *entry = xas_prev(&xas); entry 1267 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(i)); entry 1269 lib/test_xarray.c XA_BUG_ON(xa, entry != NULL); entry 1296 lib/test_xarray.c void *entry = xas_prev(&xas); entry 1298 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(i)); entry 1306 lib/test_xarray.c void *entry = xas_next(&xas); entry 1307 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(i)); entry 1320 lib/test_xarray.c void *entry = xas_prev(&xas); entry 1323 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(i)); entry 1325 lib/test_xarray.c XA_BUG_ON(xa, entry != NULL); entry 1333 lib/test_xarray.c void *entry = xas_next(&xas); entry 1335 lib/test_xarray.c XA_BUG_ON(xa, entry != xa_mk_index(i)); entry 1337 lib/test_xarray.c XA_BUG_ON(xa, entry != NULL); entry 1511 lib/test_xarray.c void *entry; entry 1518 lib/test_xarray.c xa_for_each(xa, index, entry) entry 1519 lib/test_xarray.c XA_BUG_ON(xa, xa_is_err(entry)); entry 181 lib/xarray.c void *entry; entry 188 lib/xarray.c entry = xa_head(xas->xa); entry 189 lib/xarray.c if (!xa_is_node(entry)) { entry 193 lib/xarray.c if ((xas->xa_index >> xa_to_node(entry)->shift) > XA_CHUNK_MASK) entry 198 lib/xarray.c return entry; entry 204 lib/xarray.c void *entry = xa_entry(xas->xa, node, offset); entry 207 lib/xarray.c if (xa_is_sibling(entry)) { entry 208 lib/xarray.c offset = xa_to_sibling(entry); entry 209 lib/xarray.c entry = xa_entry(xas->xa, node, offset); entry 213 lib/xarray.c return entry; entry 233 lib/xarray.c void *entry = xas_start(xas); entry 235 lib/xarray.c while (xa_is_node(entry)) { entry 236 lib/xarray.c struct xa_node *node = xa_to_node(entry); entry 240 lib/xarray.c entry = xas_descend(xas, node); entry 244 lib/xarray.c return entry; entry 426 lib/xarray.c static unsigned long max_index(void *entry) entry 428 lib/xarray.c if (!xa_is_node(entry)) entry 430 lib/xarray.c return (XA_CHUNK_SIZE << xa_to_node(entry)->shift) - 1; entry 439 lib/xarray.c void *entry; entry 444 lib/xarray.c entry = xa_entry_locked(xa, node, 0); entry 445 lib/xarray.c if (!entry) entry 447 lib/xarray.c if (!xa_is_node(entry) && node->shift) entry 449 lib/xarray.c if (xa_is_zero(entry) && xa_zero_busy(xa)) entry 450 lib/xarray.c entry = NULL; entry 453 lib/xarray.c RCU_INIT_POINTER(xa->xa_head, entry); entry 459 lib/xarray.c if (!xa_is_node(entry)) entry 463 lib/xarray.c if (!xa_is_node(entry)) entry 465 lib/xarray.c node = xa_to_node(entry); entry 525 lib/xarray.c void *entry = xa_entry_locked(xas->xa, node, offset); entry 527 lib/xarray.c if (node->shift && xa_is_node(entry)) { entry 528 lib/xarray.c node = xa_to_node(entry); entry 532 lib/xarray.c if (entry) entry 638 lib/xarray.c void *entry; entry 645 lib/xarray.c entry = xa_head_locked(xa); entry 647 lib/xarray.c if (!entry && xa_zero_busy(xa)) entry 648 lib/xarray.c entry = XA_ZERO_ENTRY; entry 649 lib/xarray.c shift = xas_expand(xas, entry); entry 654 lib/xarray.c entry = xa_head_locked(xa); entry 662 lib/xarray.c entry = xa_entry_locked(xa, node, offset); entry 666 lib/xarray.c entry = xa_head_locked(xa); entry 672 lib/xarray.c if (!entry) { entry 679 lib/xarray.c } else if (xa_is_node(entry)) { entry 680 lib/xarray.c node = xa_to_node(entry); entry 684 lib/xarray.c entry = xas_descend(xas, node); entry 688 lib/xarray.c return entry; entry 769 lib/xarray.c void *xas_store(struct xa_state *xas, void *entry) entry 777 lib/xarray.c bool value = xa_is_value(entry); entry 779 lib/xarray.c if (entry) { entry 780 lib/xarray.c bool allow_root = !xa_is_node(entry) && !xa_is_zero(entry); entry 791 lib/xarray.c if ((first == entry) && !xas->xa_sibs) entry 802 lib/xarray.c if (!entry) entry 813 lib/xarray.c rcu_assign_pointer(*slot, entry); entry 818 lib/xarray.c count += !next - !entry; entry 820 lib/xarray.c if (entry) { entry 823 lib/xarray.c if (!xa_is_sibling(entry)) entry 824 lib/xarray.c entry = xa_mk_sibling(xas->xa_offset); entry 831 lib/xarray.c if (!entry && (offset > max)) entry 996 lib/xarray.c void *entry; entry 1016 lib/xarray.c entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset); entry 1017 lib/xarray.c if (!xa_is_node(entry)) entry 1018 lib/xarray.c return entry; entry 1020 lib/xarray.c xas->xa_node = xa_to_node(entry); entry 1035 lib/xarray.c void *entry; entry 1055 lib/xarray.c entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset); entry 1056 lib/xarray.c if (!xa_is_node(entry)) entry 1057 lib/xarray.c return entry; entry 1059 lib/xarray.c xas->xa_node = xa_to_node(entry); entry 1083 lib/xarray.c void *entry; entry 1094 lib/xarray.c entry = xas_load(xas); entry 1095 lib/xarray.c if (entry || xas_not_node(xas->xa_node)) entry 1096 lib/xarray.c return entry; entry 1111 lib/xarray.c entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset); entry 1112 lib/xarray.c if (xa_is_node(entry)) { entry 1113 lib/xarray.c xas->xa_node = xa_to_node(entry); entry 1117 lib/xarray.c if (entry && !xa_is_sibling(entry)) entry 1118 lib/xarray.c return entry; entry 1154 lib/xarray.c void *entry; entry 1166 lib/xarray.c entry = xa_head(xas->xa); entry 1168 lib/xarray.c if (xas->xa_index > max_index(entry)) entry 1170 lib/xarray.c if (!xa_is_node(entry)) { entry 1172 lib/xarray.c return entry; entry 1176 lib/xarray.c xas->xa_node = xa_to_node(entry); entry 1191 lib/xarray.c entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset); entry 1192 lib/xarray.c if (xa_is_sibling(entry)) { entry 1193 lib/xarray.c xas->xa_offset = xa_to_sibling(entry); entry 1210 lib/xarray.c entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset); entry 1211 lib/xarray.c if (!entry && !(xa_track_free(xas->xa) && mark == XA_FREE_MARK)) entry 1213 lib/xarray.c if (!xa_is_node(entry)) entry 1214 lib/xarray.c return entry; entry 1215 lib/xarray.c xas->xa_node = xa_to_node(entry); entry 1301 lib/xarray.c void *entry; entry 1305 lib/xarray.c entry = xas_load(&xas); entry 1306 lib/xarray.c if (xa_is_zero(entry)) entry 1307 lib/xarray.c entry = NULL; entry 1308 lib/xarray.c } while (xas_retry(&xas, entry)); entry 1311 lib/xarray.c return entry; entry 1357 lib/xarray.c void *entry; entry 1360 lib/xarray.c entry = __xa_erase(xa, index); entry 1363 lib/xarray.c return entry; entry 1382 lib/xarray.c void *__xa_store(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp) entry 1387 lib/xarray.c if (WARN_ON_ONCE(xa_is_advanced(entry))) entry 1389 lib/xarray.c if (xa_track_free(xa) && !entry) entry 1390 lib/xarray.c entry = XA_ZERO_ENTRY; entry 1393 lib/xarray.c curr = xas_store(&xas, entry); entry 1419 lib/xarray.c void *xa_store(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp) entry 1424 lib/xarray.c curr = __xa_store(xa, index, entry, gfp); entry 1448 lib/xarray.c void *old, void *entry, gfp_t gfp) entry 1453 lib/xarray.c if (WARN_ON_ONCE(xa_is_advanced(entry))) entry 1459 lib/xarray.c xas_store(&xas, entry); entry 1460 lib/xarray.c if (xa_track_free(xa) && entry && !curr) entry 1485 lib/xarray.c int __xa_insert(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp) entry 1490 lib/xarray.c if (WARN_ON_ONCE(xa_is_advanced(entry))) entry 1492 lib/xarray.c if (!entry) entry 1493 lib/xarray.c entry = XA_ZERO_ENTRY; entry 1498 lib/xarray.c xas_store(&xas, entry); entry 1561 lib/xarray.c unsigned long last, void *entry, gfp_t gfp) entry 1565 lib/xarray.c if (WARN_ON_ONCE(xa_is_internal(entry))) entry 1572 lib/xarray.c if (entry) { entry 1583 lib/xarray.c xas_store(&xas, entry); entry 1614 lib/xarray.c int __xa_alloc(struct xarray *xa, u32 *id, void *entry, entry 1619 lib/xarray.c if (WARN_ON_ONCE(xa_is_advanced(entry))) entry 1624 lib/xarray.c if (!entry) entry 1625 lib/xarray.c entry = XA_ZERO_ENTRY; entry 1634 lib/xarray.c xas_store(&xas, entry); entry 1663 lib/xarray.c int __xa_alloc_cyclic(struct xarray *xa, u32 *id, void *entry, entry 1670 lib/xarray.c ret = __xa_alloc(xa, id, entry, limit, gfp); entry 1678 lib/xarray.c ret = __xa_alloc(xa, id, entry, limit, gfp); entry 1705 lib/xarray.c void *entry = xas_load(&xas); entry 1707 lib/xarray.c if (entry) entry 1723 lib/xarray.c void *entry = xas_load(&xas); entry 1725 lib/xarray.c if (entry) entry 1745 lib/xarray.c void *entry; entry 1748 lib/xarray.c entry = xas_start(&xas); entry 1750 lib/xarray.c if (!xa_is_node(entry)) entry 1752 lib/xarray.c entry = xas_descend(&xas, xa_to_node(entry)); entry 1819 lib/xarray.c void *entry; entry 1824 lib/xarray.c entry = xas_find_marked(&xas, max, filter); entry 1826 lib/xarray.c entry = xas_find(&xas, max); entry 1827 lib/xarray.c } while (xas_retry(&xas, entry)); entry 1830 lib/xarray.c if (entry) entry 1832 lib/xarray.c return entry; entry 1869 lib/xarray.c void *entry; entry 1877 lib/xarray.c entry = xas_find_marked(&xas, max, filter); entry 1879 lib/xarray.c entry = xas_find(&xas, max); entry 1885 lib/xarray.c if (!xas_retry(&xas, entry)) entry 1890 lib/xarray.c if (entry) entry 1892 lib/xarray.c return entry; entry 1899 lib/xarray.c void *entry; entry 1903 lib/xarray.c xas_for_each(xas, entry, max) { entry 1904 lib/xarray.c if (xas_retry(xas, entry)) entry 1906 lib/xarray.c dst[i++] = entry; entry 1918 lib/xarray.c void *entry; entry 1922 lib/xarray.c xas_for_each_marked(xas, entry, max, mark) { entry 1923 lib/xarray.c if (xas_retry(xas, entry)) entry 1925 lib/xarray.c dst[i++] = entry; entry 1990 lib/xarray.c void *entry; entry 1994 lib/xarray.c entry = xa_head_locked(xa); entry 2000 lib/xarray.c if (xa_is_node(entry)) entry 2001 lib/xarray.c xas_free_nodes(&xas, xa_to_node(entry)); entry 2039 lib/xarray.c void xa_dump_entry(const void *entry, unsigned long index, unsigned long shift) entry 2041 lib/xarray.c if (!entry) entry 2046 lib/xarray.c if (xa_is_node(entry)) { entry 2048 lib/xarray.c pr_cont("%px\n", entry); entry 2051 lib/xarray.c struct xa_node *node = xa_to_node(entry); entry 2057 lib/xarray.c } else if (xa_is_value(entry)) entry 2058 lib/xarray.c pr_cont("value %ld (0x%lx) [%px]\n", xa_to_value(entry), entry 2059 lib/xarray.c xa_to_value(entry), entry); entry 2060 lib/xarray.c else if (!xa_is_internal(entry)) entry 2061 lib/xarray.c pr_cont("%px\n", entry); entry 2062 lib/xarray.c else if (xa_is_retry(entry)) entry 2063 lib/xarray.c pr_cont("retry (%ld)\n", xa_to_internal(entry)); entry 2064 lib/xarray.c else if (xa_is_sibling(entry)) entry 2065 lib/xarray.c pr_cont("sibling (slot %ld)\n", xa_to_sibling(entry)); entry 2066 lib/xarray.c else if (xa_is_zero(entry)) entry 2067 lib/xarray.c pr_cont("zero (%ld)\n", xa_to_internal(entry)); entry 2069 lib/xarray.c pr_cont("UNKNOWN ENTRY (%px)\n", entry); entry 2074 lib/xarray.c void *entry = xa->xa_head; entry 2077 lib/xarray.c pr_info("xarray: %px head %px flags %x marks %d %d %d\n", xa, entry, entry 2080 lib/xarray.c if (xa_is_node(entry)) entry 2081 lib/xarray.c shift = xa_to_node(entry)->shift + XA_CHUNK_SHIFT; entry 2082 lib/xarray.c xa_dump_entry(entry, 0, shift); entry 1069 mm/filemap.c INIT_LIST_HEAD(&bookmark.entry); entry 1161 mm/filemap.c if (likely(list_empty(&wait->entry))) { entry 1465 mm/filemap.c void *entry = xas_next(&xas); entry 1466 mm/filemap.c if (!entry || xa_is_value(entry)) entry 1501 mm/filemap.c void *entry = xas_prev(&xas); entry 1502 mm/filemap.c if (!entry || xa_is_value(entry)) entry 248 mm/frontswap.c swp_entry_t entry = { .val = page_private(page), }; entry 249 mm/frontswap.c int type = swp_type(entry); entry 251 mm/frontswap.c pgoff_t offset = swp_offset(entry); entry 297 mm/frontswap.c swp_entry_t entry = { .val = page_private(page), }; entry 298 mm/frontswap.c int type = swp_type(entry); entry 300 mm/frontswap.c pgoff_t offset = swp_offset(entry); entry 147 mm/gup.c pte_t entry = *pte; entry 150 mm/gup.c entry = pte_mkdirty(entry); entry 151 mm/gup.c entry = pte_mkyoung(entry); entry 153 mm/gup.c if (!pte_same(*pte, entry)) { entry 154 mm/gup.c set_pte_at(vma->vm_mm, address, pte, entry); entry 198 mm/gup.c swp_entry_t entry; entry 208 mm/gup.c entry = pte_to_swp_entry(pte); entry 209 mm/gup.c if (!is_migration_entry(entry)) entry 477 mm/hmm.c swp_entry_t entry = pte_to_swp_entry(pte); entry 479 mm/hmm.c if (!non_swap_entry(entry)) { entry 492 mm/hmm.c if (is_device_private_entry(entry)) { entry 495 mm/hmm.c cpu_flags |= is_write_device_private_entry(entry) ? entry 502 mm/hmm.c swp_offset(entry)); entry 507 mm/hmm.c if (is_migration_entry(entry)) { entry 737 mm/hmm.c pte_t entry; entry 741 mm/hmm.c entry = huge_ptep_get(pte); entry 746 mm/hmm.c cpu_flags = pte_to_hmm_pfn_flags(range, entry); entry 755 mm/hmm.c pfn = pte_pfn(entry) + ((start & ~hmask) >> PAGE_SHIFT); entry 610 mm/huge_memory.c pmd_t entry; entry 629 mm/huge_memory.c entry = mk_huge_pmd(page, vma->vm_page_prot); entry 630 mm/huge_memory.c entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); entry 635 mm/huge_memory.c set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); entry 695 mm/huge_memory.c pmd_t entry; entry 698 mm/huge_memory.c entry = mk_pmd(zero_page, vma->vm_page_prot); entry 699 mm/huge_memory.c entry = pmd_mkhuge(entry); entry 702 mm/huge_memory.c set_pmd_at(mm, haddr, pmd, entry); entry 774 mm/huge_memory.c pmd_t entry; entry 784 mm/huge_memory.c entry = pmd_mkyoung(*pmd); entry 785 mm/huge_memory.c entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); entry 786 mm/huge_memory.c if (pmdp_set_access_flags(vma, addr, pmd, entry, 1)) entry 793 mm/huge_memory.c entry = pmd_mkhuge(pfn_t_pmd(pfn, prot)); entry 795 mm/huge_memory.c entry = pmd_mkdevmap(entry); entry 797 mm/huge_memory.c entry = pmd_mkyoung(pmd_mkdirty(entry)); entry 798 mm/huge_memory.c entry = maybe_pmd_mkwrite(entry, vma); entry 807 mm/huge_memory.c set_pmd_at(mm, addr, pmd, entry); entry 862 mm/huge_memory.c pud_t entry; entry 872 mm/huge_memory.c entry = pud_mkyoung(*pud); entry 873 mm/huge_memory.c entry = maybe_pud_mkwrite(pud_mkdirty(entry), vma); entry 874 mm/huge_memory.c if (pudp_set_access_flags(vma, addr, pud, entry, 1)) entry 880 mm/huge_memory.c entry = pud_mkhuge(pfn_t_pud(pfn, prot)); entry 882 mm/huge_memory.c entry = pud_mkdevmap(entry); entry 884 mm/huge_memory.c entry = pud_mkyoung(pud_mkdirty(entry)); entry 885 mm/huge_memory.c entry = maybe_pud_mkwrite(entry, vma); entry 887 mm/huge_memory.c set_pud_at(mm, addr, pud, entry); entry 1005 mm/huge_memory.c swp_entry_t entry = pmd_to_swp_entry(pmd); entry 1008 mm/huge_memory.c if (is_write_migration_entry(entry)) { entry 1009 mm/huge_memory.c make_migration_entry_read(&entry); entry 1010 mm/huge_memory.c pmd = swp_entry_to_pmd(entry); entry 1157 mm/huge_memory.c pud_t entry; entry 1165 mm/huge_memory.c entry = pud_mkyoung(orig_pud); entry 1167 mm/huge_memory.c entry = pud_mkdirty(entry); entry 1169 mm/huge_memory.c if (pudp_set_access_flags(vmf->vma, haddr, vmf->pud, entry, write)) entry 1179 mm/huge_memory.c pmd_t entry; entry 1187 mm/huge_memory.c entry = pmd_mkyoung(orig_pmd); entry 1189 mm/huge_memory.c entry = pmd_mkdirty(entry); entry 1191 mm/huge_memory.c if (pmdp_set_access_flags(vmf->vma, haddr, vmf->pmd, entry, write)) entry 1270 mm/huge_memory.c pte_t entry; entry 1271 mm/huge_memory.c entry = mk_pte(pages[i], vma->vm_page_prot); entry 1272 mm/huge_memory.c entry = maybe_mkwrite(pte_mkdirty(entry), vma); entry 1280 mm/huge_memory.c set_pte_at(vma->vm_mm, haddr, vmf->pte, entry); entry 1352 mm/huge_memory.c pmd_t entry; entry 1353 mm/huge_memory.c entry = pmd_mkyoung(orig_pmd); entry 1354 mm/huge_memory.c entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); entry 1355 mm/huge_memory.c if (pmdp_set_access_flags(vma, haddr, vmf->pmd, entry, 1)) entry 1424 mm/huge_memory.c pmd_t entry; entry 1425 mm/huge_memory.c entry = mk_huge_pmd(new_page, vma->vm_page_prot); entry 1426 mm/huge_memory.c entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); entry 1431 mm/huge_memory.c set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); entry 1813 mm/huge_memory.c swp_entry_t entry; entry 1816 mm/huge_memory.c entry = pmd_to_swp_entry(orig_pmd); entry 1817 mm/huge_memory.c page = pfn_to_page(swp_offset(entry)); entry 1929 mm/huge_memory.c pmd_t entry; entry 1942 mm/huge_memory.c swp_entry_t entry = pmd_to_swp_entry(*pmd); entry 1945 mm/huge_memory.c if (is_write_migration_entry(entry)) { entry 1951 mm/huge_memory.c make_migration_entry_read(&entry); entry 1952 mm/huge_memory.c newpmd = swp_entry_to_pmd(entry); entry 1993 mm/huge_memory.c entry = pmdp_invalidate(vma, addr, pmd); entry 1995 mm/huge_memory.c entry = pmd_modify(entry, newprot); entry 1997 mm/huge_memory.c entry = pmd_mk_savedwrite(entry); entry 1999 mm/huge_memory.c set_pmd_at(mm, addr, pmd, entry); entry 2000 mm/huge_memory.c BUG_ON(vma_is_anonymous(vma) && !preserve_write && pmd_write(entry)); entry 2127 mm/huge_memory.c pte_t *pte, entry; entry 2128 mm/huge_memory.c entry = pfn_pte(my_zero_pfn(haddr), vma->vm_page_prot); entry 2129 mm/huge_memory.c entry = pte_mkspecial(entry); entry 2132 mm/huge_memory.c set_pte_at(mm, haddr, pte, entry); entry 2214 mm/huge_memory.c swp_entry_t entry; entry 2216 mm/huge_memory.c entry = pmd_to_swp_entry(old_pmd); entry 2217 mm/huge_memory.c page = pfn_to_page(swp_offset(entry)); entry 2218 mm/huge_memory.c write = is_write_migration_entry(entry); entry 2240 mm/huge_memory.c pte_t entry, *pte; entry 2249 mm/huge_memory.c entry = swp_entry_to_pte(swp_entry); entry 2251 mm/huge_memory.c entry = pte_swp_mksoft_dirty(entry); entry 2253 mm/huge_memory.c entry = mk_pte(page + i, READ_ONCE(vma->vm_page_prot)); entry 2254 mm/huge_memory.c entry = maybe_mkwrite(entry, vma); entry 2256 mm/huge_memory.c entry = pte_wrprotect(entry); entry 2258 mm/huge_memory.c entry = pte_mkold(entry); entry 2260 mm/huge_memory.c entry = pte_mksoft_dirty(entry); entry 2264 mm/huge_memory.c set_pte_at(mm, addr, pte, entry); entry 2514 mm/huge_memory.c swp_entry_t entry = { .val = page_private(head) }; entry 2516 mm/huge_memory.c offset = swp_offset(entry); entry 2517 mm/huge_memory.c swap_cache = swap_address_space(entry); entry 2803 mm/huge_memory.c swp_entry_t entry = { .val = page_private(head) }; entry 2805 mm/huge_memory.c ret = split_swap_cluster(entry); entry 3027 mm/huge_memory.c swp_entry_t entry; entry 3037 mm/huge_memory.c entry = make_migration_entry(page, pmd_write(pmdval)); entry 3038 mm/huge_memory.c pmdswp = swp_entry_to_pmd(entry); entry 3053 mm/huge_memory.c swp_entry_t entry; entry 3058 mm/huge_memory.c entry = pmd_to_swp_entry(*pvmw->pmd); entry 3063 mm/huge_memory.c if (is_write_migration_entry(entry)) entry 3367 mm/hugetlb.c pte_t entry; entry 3370 mm/hugetlb.c entry = huge_pte_mkwrite(huge_pte_mkdirty(mk_huge_pte(page, entry 3373 mm/hugetlb.c entry = huge_pte_wrprotect(mk_huge_pte(page, entry 3376 mm/hugetlb.c entry = pte_mkyoung(entry); entry 3377 mm/hugetlb.c entry = pte_mkhuge(entry); entry 3378 mm/hugetlb.c entry = arch_make_huge_pte(entry, vma, page, writable); entry 3380 mm/hugetlb.c return entry; entry 3386 mm/hugetlb.c pte_t entry; entry 3388 mm/hugetlb.c entry = huge_pte_mkwrite(huge_pte_mkdirty(huge_ptep_get(ptep))); entry 3389 mm/hugetlb.c if (huge_ptep_set_access_flags(vma, address, ptep, entry, 1)) entry 3422 mm/hugetlb.c pte_t *src_pte, *dst_pte, entry, dst_entry; entry 3467 mm/hugetlb.c entry = huge_ptep_get(src_pte); entry 3469 mm/hugetlb.c if (huge_pte_none(entry) || !huge_pte_none(dst_entry)) { entry 3476 mm/hugetlb.c } else if (unlikely(is_hugetlb_entry_migration(entry) || entry 3477 mm/hugetlb.c is_hugetlb_entry_hwpoisoned(entry))) { entry 3478 mm/hugetlb.c swp_entry_t swp_entry = pte_to_swp_entry(entry); entry 3486 mm/hugetlb.c entry = swp_entry_to_pte(swp_entry); entry 3488 mm/hugetlb.c entry, sz); entry 3490 mm/hugetlb.c set_huge_swap_pte_at(dst, addr, dst_pte, entry, sz); entry 3502 mm/hugetlb.c entry = huge_ptep_get(src_pte); entry 3503 mm/hugetlb.c ptepage = pte_page(entry); entry 3506 mm/hugetlb.c set_huge_pte_at(dst, addr, dst_pte, entry); entry 4122 mm/hugetlb.c pte_t *ptep, entry; entry 4136 mm/hugetlb.c entry = huge_ptep_get(ptep); entry 4137 mm/hugetlb.c if (unlikely(is_hugetlb_entry_migration(entry))) { entry 4140 mm/hugetlb.c } else if (unlikely(is_hugetlb_entry_hwpoisoned(entry))) entry 4160 mm/hugetlb.c entry = huge_ptep_get(ptep); entry 4161 mm/hugetlb.c if (huge_pte_none(entry)) { entry 4175 mm/hugetlb.c if (!pte_present(entry)) entry 4186 mm/hugetlb.c if ((flags & FAULT_FLAG_WRITE) && !huge_pte_write(entry)) { entry 4202 mm/hugetlb.c if (unlikely(!pte_same(entry, huge_ptep_get(ptep)))) entry 4210 mm/hugetlb.c page = pte_page(entry); entry 4220 mm/hugetlb.c if (!huge_pte_write(entry)) { entry 4225 mm/hugetlb.c entry = huge_pte_mkdirty(entry); entry 4227 mm/hugetlb.c entry = pte_mkyoung(entry); entry 4228 mm/hugetlb.c if (huge_ptep_set_access_flags(vma, haddr, ptep, entry, entry 4597 mm/hugetlb.c swp_entry_t entry = pte_to_swp_entry(pte); entry 4599 mm/hugetlb.c if (is_write_migration_entry(entry)) { entry 4602 mm/hugetlb.c make_migration_entry_read(&entry); entry 4603 mm/hugetlb.c newpte = swp_entry_to_pte(entry); entry 1063 mm/ksm.c pte_t entry; entry 1081 mm/ksm.c entry = ptep_clear_flush(vma, pvmw.address, pvmw.pte); entry 1087 mm/ksm.c set_pte_at(mm, pvmw.address, pvmw.pte, entry); entry 1090 mm/ksm.c if (pte_dirty(entry)) entry 1093 mm/ksm.c if (pte_protnone(entry)) entry 1094 mm/ksm.c entry = pte_mkclean(pte_clear_savedwrite(entry)); entry 1096 mm/ksm.c entry = pte_mkclean(pte_wrprotect(entry)); entry 1097 mm/ksm.c set_pte_at_notify(mm, pvmw.address, pvmw.pte, entry); entry 195 mm/madvise.c swp_entry_t entry; entry 205 mm/madvise.c entry = pte_to_swp_entry(pte); entry 206 mm/madvise.c if (unlikely(non_swap_entry(entry))) entry 209 mm/madvise.c page = read_swap_cache_async(entry, GFP_HIGHUSER_MOVABLE, entry 598 mm/madvise.c swp_entry_t entry; entry 600 mm/madvise.c entry = pte_to_swp_entry(ptent); entry 601 mm/madvise.c if (non_swap_entry(entry)) entry 604 mm/madvise.c free_swap_and_cache(entry); entry 1983 mm/memcontrol.c INIT_LIST_HEAD(&owait.wait.entry); entry 3145 mm/memcontrol.c static int mem_cgroup_move_swap_account(swp_entry_t entry, entry 3153 mm/memcontrol.c if (swap_cgroup_cmpxchg(entry, old_id, new_id) == old_id) { entry 3161 mm/memcontrol.c static inline int mem_cgroup_move_swap_account(swp_entry_t entry, entry 5394 mm/memcontrol.c pte_t ptent, swp_entry_t *entry) entry 5424 mm/memcontrol.c entry->val = ent.val; entry 5430 mm/memcontrol.c pte_t ptent, swp_entry_t *entry) entry 5437 mm/memcontrol.c unsigned long addr, pte_t ptent, swp_entry_t *entry) entry 5459 mm/memcontrol.c *entry = swp; entry 6615 mm/memcontrol.c swp_entry_t entry = { .val = page_private(page) }; entry 6621 mm/memcontrol.c mem_cgroup_uncharge_swap(entry, nr_pages); entry 7032 mm/memcontrol.c void mem_cgroup_swapout(struct page *page, swp_entry_t entry) entry 7060 mm/memcontrol.c oldid = swap_cgroup_record(entry, mem_cgroup_id(swap_memcg), entry 7100 mm/memcontrol.c int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry) entry 7116 mm/memcontrol.c if (!entry.val) { entry 7134 mm/memcontrol.c oldid = swap_cgroup_record(entry, mem_cgroup_id(memcg), nr_pages); entry 7146 mm/memcontrol.c void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_pages) entry 7154 mm/memcontrol.c id = swap_cgroup_record(entry, 0, nr_pages); entry 1461 mm/memory-failure.c struct memory_failure_entry entry = { entry 1468 mm/memory-failure.c if (kfifo_put(&mf_cpu->fifo, entry)) entry 1481 mm/memory-failure.c struct memory_failure_entry entry = { 0, }; entry 1488 mm/memory-failure.c gotten = kfifo_get(&mf_cpu->fifo, &entry); entry 1492 mm/memory-failure.c if (entry.flags & MF_SOFT_OFFLINE) entry 1493 mm/memory-failure.c soft_offline_page(pfn_to_page(entry.pfn), entry.flags); entry 1495 mm/memory-failure.c memory_failure(entry.pfn, entry.flags); entry 688 mm/memory.c swp_entry_t entry = pte_to_swp_entry(pte); entry 690 mm/memory.c if (likely(!non_swap_entry(entry))) { entry 691 mm/memory.c if (swap_duplicate(entry) < 0) entry 692 mm/memory.c return entry.val; entry 703 mm/memory.c } else if (is_migration_entry(entry)) { entry 704 mm/memory.c page = migration_entry_to_page(entry); entry 708 mm/memory.c if (is_write_migration_entry(entry) && entry 714 mm/memory.c make_migration_entry_read(&entry); entry 715 mm/memory.c pte = swp_entry_to_pte(entry); entry 720 mm/memory.c } else if (is_device_private_entry(entry)) { entry 721 mm/memory.c page = device_private_entry_to_page(entry); entry 743 mm/memory.c if (is_write_device_private_entry(entry) && entry 745 mm/memory.c make_device_private_entry_read(&entry); entry 746 mm/memory.c pte = swp_entry_to_pte(entry); entry 793 mm/memory.c swp_entry_t entry = (swp_entry_t){0}; entry 823 mm/memory.c entry.val = copy_one_pte(dst_mm, src_mm, dst_pte, src_pte, entry 825 mm/memory.c if (entry.val) entry 837 mm/memory.c if (entry.val) { entry 838 mm/memory.c if (add_swap_count_continuation(entry, GFP_KERNEL) < 0) entry 1015 mm/memory.c swp_entry_t entry; entry 1073 mm/memory.c entry = pte_to_swp_entry(ptent); entry 1074 mm/memory.c if (non_swap_entry(entry) && is_device_private_entry(entry)) { entry 1075 mm/memory.c struct page *page = device_private_entry_to_page(entry); entry 1099 mm/memory.c if (!non_swap_entry(entry)) entry 1101 mm/memory.c else if (is_migration_entry(entry)) { entry 1104 mm/memory.c page = migration_entry_to_page(entry); entry 1107 mm/memory.c if (unlikely(!free_swap_and_cache(entry))) entry 1585 mm/memory.c pte_t *pte, entry; entry 1607 mm/memory.c entry = pte_mkyoung(*pte); entry 1608 mm/memory.c entry = maybe_mkwrite(pte_mkdirty(entry), vma); entry 1609 mm/memory.c if (ptep_set_access_flags(vma, addr, pte, entry, 1)) entry 1617 mm/memory.c entry = pte_mkdevmap(pfn_t_pte(pfn, prot)); entry 1619 mm/memory.c entry = pte_mkspecial(pfn_t_pte(pfn, prot)); entry 1622 mm/memory.c entry = pte_mkyoung(entry); entry 1623 mm/memory.c entry = maybe_mkwrite(pte_mkdirty(entry), vma); entry 1626 mm/memory.c set_pte_at(mm, addr, pte, entry); entry 2288 mm/memory.c pte_t entry; entry 2298 mm/memory.c entry = pte_mkyoung(vmf->orig_pte); entry 2299 mm/memory.c entry = maybe_mkwrite(pte_mkdirty(entry), vma); entry 2300 mm/memory.c if (ptep_set_access_flags(vma, vmf->address, vmf->pte, entry, 1)) entry 2327 mm/memory.c pte_t entry; entry 2373 mm/memory.c entry = mk_pte(new_page, vma->vm_page_prot); entry 2374 mm/memory.c entry = maybe_mkwrite(pte_mkdirty(entry), vma); entry 2390 mm/memory.c set_pte_at_notify(mm, vmf->address, vmf->pte, entry); entry 2756 mm/memory.c swp_entry_t entry; entry 2765 mm/memory.c entry = pte_to_swp_entry(vmf->orig_pte); entry 2766 mm/memory.c if (unlikely(non_swap_entry(entry))) { entry 2767 mm/memory.c if (is_migration_entry(entry)) { entry 2770 mm/memory.c } else if (is_device_private_entry(entry)) { entry 2771 mm/memory.c vmf->page = device_private_entry_to_page(entry); entry 2773 mm/memory.c } else if (is_hwpoison_entry(entry)) { entry 2784 mm/memory.c page = lookup_swap_cache(entry, vma, vmf->address); entry 2788 mm/memory.c struct swap_info_struct *si = swp_swap_info(entry); entry 2791 mm/memory.c __swap_count(entry) == 1) { entry 2798 mm/memory.c set_page_private(page, entry.val); entry 2803 mm/memory.c page = swapin_readahead(entry, GFP_HIGHUSER_MOVABLE, entry 2850 mm/memory.c page_private(page) != entry.val)) && swapcache) entry 2916 mm/memory.c swap_free(entry); entry 2972 mm/memory.c pte_t entry; entry 2998 mm/memory.c entry = pte_mkspecial(pfn_pte(my_zero_pfn(vmf->address), entry 3033 mm/memory.c entry = mk_pte(page, vma->vm_page_prot); entry 3035 mm/memory.c entry = pte_mkwrite(pte_mkdirty(entry)); entry 3059 mm/memory.c set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry); entry 3208 mm/memory.c pmd_t entry; entry 3236 mm/memory.c entry = mk_huge_pmd(page, vma->vm_page_prot); entry 3238 mm/memory.c entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); entry 3248 mm/memory.c set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); entry 3288 mm/memory.c pte_t entry; entry 3312 mm/memory.c entry = mk_pte(page, vma->vm_page_prot); entry 3314 mm/memory.c entry = maybe_mkwrite(pte_mkdirty(entry), vma); entry 3325 mm/memory.c set_pte_at(vma->vm_mm, vmf->address, vmf->pte, entry); entry 3823 mm/memory.c pte_t entry; entry 3876 mm/memory.c entry = vmf->orig_pte; entry 3877 mm/memory.c if (unlikely(!pte_same(*vmf->pte, entry))) entry 3880 mm/memory.c if (!pte_write(entry)) entry 3882 mm/memory.c entry = pte_mkdirty(entry); entry 3884 mm/memory.c entry = pte_mkyoung(entry); entry 3885 mm/memory.c if (ptep_set_access_flags(vmf->vma, vmf->address, vmf->pte, entry, entry 563 mm/mempolicy.c pte_t entry; entry 566 mm/mempolicy.c entry = huge_ptep_get(pte); entry 567 mm/mempolicy.c if (!pte_present(entry)) entry 569 mm/mempolicy.c page = pte_page(entry); entry 215 mm/migrate.c swp_entry_t entry; entry 242 mm/migrate.c entry = pte_to_swp_entry(*pvmw.pte); entry 243 mm/migrate.c if (is_write_migration_entry(entry)) entry 248 mm/migrate.c entry = make_device_private_entry(new, pte_write(pte)); entry 249 mm/migrate.c pte = swp_entry_to_pte(entry); entry 311 mm/migrate.c swp_entry_t entry; entry 319 mm/migrate.c entry = pte_to_swp_entry(pte); entry 320 mm/migrate.c if (!is_migration_entry(entry)) entry 323 mm/migrate.c page = migration_entry_to_page(entry); entry 2026 mm/migrate.c pmd_t *pmd, pmd_t entry, entry 2065 mm/migrate.c if (unlikely(!pmd_same(*pmd, entry) || !page_ref_freeze(page, 2))) { entry 2086 mm/migrate.c entry = mk_huge_pmd(new_page, vma->vm_page_prot); entry 2087 mm/migrate.c entry = maybe_pmd_mkwrite(pmd_mkdirty(entry), vma); entry 2109 mm/migrate.c set_pmd_at(mm, start, pmd, entry); entry 2110 mm/migrate.c update_mmu_cache_pmd(vma, address, &entry); entry 2139 mm/migrate.c if (pmd_same(*pmd, entry)) { entry 2140 mm/migrate.c entry = pmd_modify(entry, vma->vm_page_prot); entry 2141 mm/migrate.c set_pmd_at(mm, start, pmd, entry); entry 2142 mm/migrate.c update_mmu_cache_pmd(vma, address, &entry); entry 2249 mm/migrate.c swp_entry_t entry; entry 2268 mm/migrate.c entry = pte_to_swp_entry(pte); entry 2269 mm/migrate.c if (!is_device_private_entry(entry)) entry 2272 mm/migrate.c page = device_private_entry_to_page(entry); entry 2275 mm/migrate.c if (is_write_device_private_entry(entry)) entry 2319 mm/migrate.c entry = make_migration_entry(page, mpfn & entry 2321 mm/migrate.c swp_pte = swp_entry_to_pte(entry); entry 2714 mm/migrate.c pte_t entry; entry 2773 mm/migrate.c entry = swp_entry_to_pte(swp_entry); entry 2776 mm/migrate.c entry = mk_pte(page, vma->vm_page_prot); entry 2778 mm/migrate.c entry = pte_mkwrite(pte_mkdirty(entry)); entry 2818 mm/migrate.c set_pte_at_notify(mm, addr, ptep, entry); entry 2822 mm/migrate.c set_pte_at(mm, addr, ptep, entry); entry 153 mm/mincore.c swp_entry_t entry = pte_to_swp_entry(pte); entry 155 mm/mincore.c if (non_swap_entry(entry)) { entry 163 mm/mincore.c *vec = mincore_page(swap_address_space(entry), entry 164 mm/mincore.c swp_offset(entry)); entry 126 mm/mprotect.c swp_entry_t entry = pte_to_swp_entry(oldpte); entry 128 mm/mprotect.c if (is_write_migration_entry(entry)) { entry 134 mm/mprotect.c make_migration_entry_read(&entry); entry 135 mm/mprotect.c newpte = swp_entry_to_pte(entry); entry 143 mm/mprotect.c if (is_write_device_private_entry(entry)) { entry 150 mm/mprotect.c make_device_private_entry_read(&entry); entry 151 mm/mprotect.c newpte = swp_entry_to_pte(entry); entry 76 mm/page_io.c swp_entry_t entry; entry 108 mm/page_io.c entry.val = page_private(page); entry 109 mm/page_io.c if (disk->fops->swap_slot_free_notify && __swap_count(entry) == 1) { entry 112 mm/page_io.c offset = swp_offset(entry); entry 40 mm/page_vma_mapped.c swp_entry_t entry; entry 43 mm/page_vma_mapped.c entry = pte_to_swp_entry(*pvmw->pte); entry 44 mm/page_vma_mapped.c if (!is_device_private_entry(entry)) entry 86 mm/page_vma_mapped.c swp_entry_t entry; entry 89 mm/page_vma_mapped.c entry = pte_to_swp_entry(*pvmw->pte); entry 91 mm/page_vma_mapped.c if (!is_migration_entry(entry)) entry 94 mm/page_vma_mapped.c pfn = migration_entry_to_pfn(entry); entry 96 mm/page_vma_mapped.c swp_entry_t entry; entry 99 mm/page_vma_mapped.c entry = pte_to_swp_entry(*pvmw->pte); entry 100 mm/page_vma_mapped.c if (!is_device_private_entry(entry)) entry 103 mm/page_vma_mapped.c pfn = device_private_entry_to_pfn(entry); entry 196 mm/page_vma_mapped.c swp_entry_t entry = pmd_to_swp_entry(*pvmw->pmd); entry 198 mm/page_vma_mapped.c if (migration_entry_to_page(entry) != page) entry 57 mm/pgtable-generic.c pte_t entry, int dirty) entry 59 mm/pgtable-generic.c int changed = !pte_same(*ptep, entry); entry 61 mm/pgtable-generic.c set_pte_at(vma->vm_mm, address, ptep, entry); entry 98 mm/pgtable-generic.c pmd_t entry, int dirty) entry 100 mm/pgtable-generic.c int changed = !pmd_same(*pmdp, entry); entry 103 mm/pgtable-generic.c set_pmd_at(vma->vm_mm, address, pmdp, entry); entry 910 mm/rmap.c pte_t entry; entry 917 mm/rmap.c entry = ptep_clear_flush(vma, address, pte); entry 918 mm/rmap.c entry = pte_wrprotect(entry); entry 919 mm/rmap.c entry = pte_mkclean(entry); entry 920 mm/rmap.c set_pte_at(vma->vm_mm, address, pte, entry); entry 925 mm/rmap.c pmd_t entry; entry 931 mm/rmap.c entry = pmdp_invalidate(vma, address, pmd); entry 932 mm/rmap.c entry = pmd_wrprotect(entry); entry 933 mm/rmap.c entry = pmd_mkclean(entry); entry 934 mm/rmap.c set_pmd_at(vma->vm_mm, address, pmd, entry); entry 1461 mm/rmap.c swp_entry_t entry; entry 1471 mm/rmap.c entry = make_migration_entry(page, 0); entry 1472 mm/rmap.c swp_pte = swp_entry_to_pte(entry); entry 1554 mm/rmap.c swp_entry_t entry; entry 1569 mm/rmap.c entry = make_migration_entry(subpage, entry 1571 mm/rmap.c swp_pte = swp_entry_to_pte(entry); entry 1580 mm/rmap.c swp_entry_t entry = { .val = page_private(subpage) }; entry 1617 mm/rmap.c if (swap_duplicate(entry) < 0) { entry 1637 mm/rmap.c swp_pte = swp_entry_to_pte(entry); entry 625 mm/shmem.c void *entry; entry 627 mm/shmem.c entry = xas_find_conflict(&xas); entry 628 mm/shmem.c if (entry != expected) entry 1128 mm/shmem.c swp_entry_t entry; entry 1142 mm/shmem.c entry = radix_to_swp_entry(page); entry 1143 mm/shmem.c if (swp_type(entry) != type) entry 1146 mm/shmem.c !frontswap_test(swap_info[type], swp_offset(entry))) entry 1558 mm/shmem.c swp_entry_t entry; entry 1563 mm/shmem.c entry.val = page_private(oldpage); entry 1564 mm/shmem.c swap_index = swp_offset(entry); entry 1583 mm/shmem.c set_page_private(newpage, entry.val); entry 1986 mm/shmem.c list_del_init(&wait->entry); entry 189 mm/slab.c void *entry[]; /* entry 583 mm/slab.c memcpy(to->entry + to->avail, from->entry + from->avail -nr, entry 700 mm/slab.c free_block(cachep, ac->entry, ac->avail, node, list); entry 770 mm/slab.c ac->entry[ac->avail++] = objp; entry 902 mm/slab.c free_block(cachep, n->shared->entry, entry 964 mm/slab.c free_block(cachep, nc->entry, nc->avail, node, &list); entry 974 mm/slab.c free_block(cachep, shared->entry, entry 2131 mm/slab.c free_block(cachep, ac->entry, tofree, node, list); entry 2133 mm/slab.c memmove(ac->entry, &(ac->entry[tofree]), sizeof(void *) * ac->avail); entry 2148 mm/slab.c free_block(cachep, ac->entry, ac->avail, node, &list); entry 2899 mm/slab.c ac->entry[ac->avail++] = slab_get_obj(cachep, page); entry 2987 mm/slab.c return ac->entry[--ac->avail]; entry 3047 mm/slab.c objp = ac->entry[--ac->avail]; entry 3068 mm/slab.c kmemleak_erase(&ac->entry[ac->avail]); entry 3389 mm/slab.c memcpy(&(shared_array->entry[shared_array->avail]), entry 3390 mm/slab.c ac->entry, sizeof(void *) * batchcount); entry 3396 mm/slab.c free_block(cachep, ac->entry, batchcount, node, &list); entry 3414 mm/slab.c memmove(ac->entry, &(ac->entry[batchcount]), sizeof(void *)*ac->avail); entry 3468 mm/slab.c ac->entry[ac->avail++] = objp; entry 3834 mm/slab.c free_block(cachep, ac->entry, ac->avail, node, &list); entry 147 mm/sparse-vmemmap.c pte_t entry; entry 151 mm/sparse-vmemmap.c entry = pfn_pte(__pa(p) >> PAGE_SHIFT, PAGE_KERNEL); entry 152 mm/sparse-vmemmap.c set_pte_at(&init_mm, addr, pte, entry); entry 278 mm/swap_slots.c int free_swap_slot(swp_entry_t entry) entry 300 mm/swap_slots.c cache->slots_ret[cache->n_ret++] = entry; entry 304 mm/swap_slots.c swapcache_free_entries(&entry, 1); entry 312 mm/swap_slots.c swp_entry_t entry, *pentry; entry 315 mm/swap_slots.c entry.val = 0; entry 319 mm/swap_slots.c get_swap_pages(1, &entry, HPAGE_PMD_NR); entry 340 mm/swap_slots.c entry = *pentry; entry 349 mm/swap_slots.c if (entry.val) entry 353 mm/swap_slots.c get_swap_pages(1, &entry, 1); entry 355 mm/swap_slots.c if (mem_cgroup_try_charge_swap(page, entry)) { entry 356 mm/swap_slots.c put_swap_page(page, entry); entry 357 mm/swap_slots.c entry.val = 0; entry 359 mm/swap_slots.c return entry; entry 79 mm/swap_state.c swp_entry_t entry = swp_entry(i, 1); entry 82 mm/swap_state.c if (!swp_swap_info(entry)) entry 85 mm/swap_state.c si = get_swap_device(entry); entry 114 mm/swap_state.c int add_to_swap_cache(struct page *page, swp_entry_t entry, gfp_t gfp) entry 116 mm/swap_state.c struct address_space *address_space = swap_address_space(entry); entry 117 mm/swap_state.c pgoff_t idx = swp_offset(entry); entry 135 mm/swap_state.c set_page_private(page + i, entry.val + i); entry 158 mm/swap_state.c void __delete_from_swap_cache(struct page *page, swp_entry_t entry) entry 160 mm/swap_state.c struct address_space *address_space = swap_address_space(entry); entry 162 mm/swap_state.c pgoff_t idx = swp_offset(entry); entry 170 mm/swap_state.c void *entry = xas_store(&xas, NULL); entry 171 mm/swap_state.c VM_BUG_ON_PAGE(entry != page, entry); entry 190 mm/swap_state.c swp_entry_t entry; entry 196 mm/swap_state.c entry = get_swap_page(page); entry 197 mm/swap_state.c if (!entry.val) entry 211 mm/swap_state.c err = add_to_swap_cache(page, entry, entry 234 mm/swap_state.c put_swap_page(page, entry); entry 246 mm/swap_state.c swp_entry_t entry = { .val = page_private(page) }; entry 247 mm/swap_state.c struct address_space *address_space = swap_address_space(entry); entry 250 mm/swap_state.c __delete_from_swap_cache(page, entry); entry 253 mm/swap_state.c put_swap_page(page, entry); entry 310 mm/swap_state.c struct page *lookup_swap_cache(swp_entry_t entry, struct vm_area_struct *vma, entry 316 mm/swap_state.c si = get_swap_device(entry); entry 319 mm/swap_state.c page = find_get_page(swap_address_space(entry), swp_offset(entry)); entry 359 mm/swap_state.c struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, entry 374 mm/swap_state.c si = get_swap_device(entry); entry 377 mm/swap_state.c found_page = find_get_page(swap_address_space(entry), entry 378 mm/swap_state.c swp_offset(entry)); entry 391 mm/swap_state.c if (!__swp_swapcount(entry) && swap_slot_cache_enabled) entry 406 mm/swap_state.c err = swapcache_prepare(entry); entry 421 mm/swap_state.c err = add_to_swap_cache(new_page, entry, gfp_mask & GFP_KERNEL); entry 434 mm/swap_state.c put_swap_page(new_page, entry); entry 448 mm/swap_state.c struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, entry 452 mm/swap_state.c struct page *retpage = __read_swap_cache_async(entry, gfp_mask, entry 539 mm/swap_state.c struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask, entry 543 mm/swap_state.c unsigned long entry_offset = swp_offset(entry); entry 547 mm/swap_state.c struct swap_info_struct *si = swp_swap_info(entry); entry 577 mm/swap_state.c swp_entry(swp_type(entry), offset), entry 594 mm/swap_state.c return read_swap_cache_async(entry, gfp_mask, vma, addr, do_poll); entry 645 mm/swap_state.c swp_entry_t entry; entry 663 mm/swap_state.c entry = pte_to_swp_entry(*pte); entry 664 mm/swap_state.c if ((unlikely(non_swap_entry(entry)))) { entry 729 mm/swap_state.c swp_entry_t entry; entry 746 mm/swap_state.c entry = pte_to_swp_entry(pentry); entry 747 mm/swap_state.c if (unlikely(non_swap_entry(entry))) entry 749 mm/swap_state.c page = __read_swap_cache_async(entry, gfp_mask, vma, entry 781 mm/swap_state.c struct page *swapin_readahead(swp_entry_t entry, gfp_t gfp_mask, entry 785 mm/swap_state.c swap_vma_readahead(entry, gfp_mask, vmf) : entry 786 mm/swap_state.c swap_cluster_readahead(entry, gfp_mask, vmf); entry 130 mm/swapfile.c swp_entry_t entry = swp_entry(si->type, offset); entry 134 mm/swapfile.c page = find_get_page(swap_address_space(entry), offset); entry 980 mm/swapfile.c swp_entry_t entry; entry 983 mm/swapfile.c n_ret = scan_swap_map_slots(si, usage, 1, &entry); entry 986 mm/swapfile.c return swp_offset(entry); entry 1104 mm/swapfile.c static struct swap_info_struct *__swap_info_get(swp_entry_t entry) entry 1109 mm/swapfile.c if (!entry.val) entry 1111 mm/swapfile.c p = swp_swap_info(entry); entry 1116 mm/swapfile.c offset = swp_offset(entry); entry 1122 mm/swapfile.c pr_err("swap_info_get: %s%08lx\n", Bad_offset, entry.val); entry 1125 mm/swapfile.c pr_err("swap_info_get: %s%08lx\n", Unused_file, entry.val); entry 1128 mm/swapfile.c pr_err("swap_info_get: %s%08lx\n", Bad_file, entry.val); entry 1133 mm/swapfile.c static struct swap_info_struct *_swap_info_get(swp_entry_t entry) entry 1137 mm/swapfile.c p = __swap_info_get(entry); entry 1140 mm/swapfile.c if (!p->swap_map[swp_offset(entry)]) entry 1145 mm/swapfile.c pr_err("swap_info_get: %s%08lx\n", Unused_offset, entry.val); entry 1151 mm/swapfile.c static struct swap_info_struct *swap_info_get(swp_entry_t entry) entry 1155 mm/swapfile.c p = _swap_info_get(entry); entry 1161 mm/swapfile.c static struct swap_info_struct *swap_info_get_cont(swp_entry_t entry, entry 1166 mm/swapfile.c p = _swap_info_get(entry); entry 1249 mm/swapfile.c struct swap_info_struct *get_swap_device(swp_entry_t entry) entry 1254 mm/swapfile.c if (!entry.val) entry 1256 mm/swapfile.c si = swp_swap_info(entry); entry 1263 mm/swapfile.c offset = swp_offset(entry); entry 1269 mm/swapfile.c pr_err("%s: %s%08lx\n", __func__, Bad_file, entry.val); entry 1278 mm/swapfile.c swp_entry_t entry, unsigned char usage) entry 1281 mm/swapfile.c unsigned long offset = swp_offset(entry); entry 1287 mm/swapfile.c free_swap_slot(entry); entry 1292 mm/swapfile.c static void swap_entry_free(struct swap_info_struct *p, swp_entry_t entry) entry 1295 mm/swapfile.c unsigned long offset = swp_offset(entry); entry 1305 mm/swapfile.c mem_cgroup_uncharge_swap(entry, 1); entry 1313 mm/swapfile.c void swap_free(swp_entry_t entry) entry 1317 mm/swapfile.c p = _swap_info_get(entry); entry 1319 mm/swapfile.c __swap_entry_free(p, entry, 1); entry 1325 mm/swapfile.c void put_swap_page(struct page *page, swp_entry_t entry) entry 1327 mm/swapfile.c unsigned long offset = swp_offset(entry); entry 1336 mm/swapfile.c si = _swap_info_get(entry); entry 1354 mm/swapfile.c mem_cgroup_uncharge_swap(entry, SWAPFILE_CLUSTER); entry 1360 mm/swapfile.c for (i = 0; i < size; i++, entry.val++) { entry 1363 mm/swapfile.c free_swap_slot(entry); entry 1373 mm/swapfile.c int split_swap_cluster(swp_entry_t entry) entry 1377 mm/swapfile.c unsigned long offset = swp_offset(entry); entry 1379 mm/swapfile.c si = _swap_info_get(entry); entry 1434 mm/swapfile.c swp_entry_t entry; entry 1437 mm/swapfile.c entry.val = page_private(page); entry 1438 mm/swapfile.c p = _swap_info_get(entry); entry 1440 mm/swapfile.c offset = swp_offset(entry); entry 1448 mm/swapfile.c int __swap_count(swp_entry_t entry) entry 1451 mm/swapfile.c pgoff_t offset = swp_offset(entry); entry 1454 mm/swapfile.c si = get_swap_device(entry); entry 1462 mm/swapfile.c static int swap_swapcount(struct swap_info_struct *si, swp_entry_t entry) entry 1465 mm/swapfile.c pgoff_t offset = swp_offset(entry); entry 1479 mm/swapfile.c int __swp_swapcount(swp_entry_t entry) entry 1484 mm/swapfile.c si = get_swap_device(entry); entry 1486 mm/swapfile.c count = swap_swapcount(si, entry); entry 1496 mm/swapfile.c int swp_swapcount(swp_entry_t entry) entry 1505 mm/swapfile.c p = _swap_info_get(entry); entry 1509 mm/swapfile.c offset = swp_offset(entry); entry 1539 mm/swapfile.c swp_entry_t entry) entry 1543 mm/swapfile.c unsigned long roffset = swp_offset(entry); entry 1567 mm/swapfile.c swp_entry_t entry; entry 1574 mm/swapfile.c entry.val = page_private(page); entry 1575 mm/swapfile.c si = _swap_info_get(entry); entry 1577 mm/swapfile.c return swap_page_trans_huge_swapped(si, entry); entry 1607 mm/swapfile.c swp_entry_t entry; entry 1609 mm/swapfile.c entry.val = page_private(page); entry 1610 mm/swapfile.c si = _swap_info_get(entry); entry 1613 mm/swapfile.c offset = swp_offset(entry); entry 1673 mm/swapfile.c swp_entry_t entry; entry 1676 mm/swapfile.c entry.val = page_private(page); entry 1677 mm/swapfile.c p = swap_info_get(entry); entry 1732 mm/swapfile.c int free_swap_and_cache(swp_entry_t entry) entry 1737 mm/swapfile.c if (non_swap_entry(entry)) entry 1740 mm/swapfile.c p = _swap_info_get(entry); entry 1742 mm/swapfile.c count = __swap_entry_free(p, entry, 1); entry 1744 mm/swapfile.c !swap_page_trans_huge_swapped(p, entry)) entry 1745 mm/swapfile.c __try_to_reclaim_swap(p, swp_offset(entry), entry 1854 mm/swapfile.c unsigned long addr, swp_entry_t entry, struct page *page) entry 1874 mm/swapfile.c if (unlikely(!pte_same_as_swp(*pte, swp_entry_to_pte(entry)))) { entry 1893 mm/swapfile.c swap_free(entry); entry 1915 mm/swapfile.c swp_entry_t entry; entry 1930 mm/swapfile.c entry = pte_to_swp_entry(*pte); entry 1931 mm/swapfile.c if (swp_type(entry) != type) entry 1934 mm/swapfile.c offset = swp_offset(entry); entry 1943 mm/swapfile.c page = swapin_readahead(entry, GFP_HIGHUSER_MOVABLE, &vmf); entry 1952 mm/swapfile.c ret = unuse_pte(vma, pmd, addr, entry, page); entry 2132 mm/swapfile.c swp_entry_t entry; entry 2184 mm/swapfile.c entry = swp_entry(type, i); entry 2185 mm/swapfile.c page = find_get_page(swap_address_space(entry), i); entry 2257 mm/swapfile.c static sector_t map_swap_entry(swp_entry_t entry, struct block_device **bdev) entry 2263 mm/swapfile.c sis = swp_swap_info(entry); entry 2266 mm/swapfile.c offset = swp_offset(entry); entry 2276 mm/swapfile.c swp_entry_t entry; entry 2277 mm/swapfile.c entry.val = page_private(page); entry 2278 mm/swapfile.c return map_swap_entry(entry, bdev); entry 3377 mm/swapfile.c static int __swap_duplicate(swp_entry_t entry, unsigned char usage) entry 3386 mm/swapfile.c p = get_swap_device(entry); entry 3390 mm/swapfile.c offset = swp_offset(entry); entry 3445 mm/swapfile.c void swap_shmem_alloc(swp_entry_t entry) entry 3447 mm/swapfile.c __swap_duplicate(entry, SWAP_MAP_SHMEM); entry 3457 mm/swapfile.c int swap_duplicate(swp_entry_t entry) entry 3461 mm/swapfile.c while (!err && __swap_duplicate(entry, 1) == -ENOMEM) entry 3462 mm/swapfile.c err = add_swap_count_continuation(entry, GFP_ATOMIC); entry 3474 mm/swapfile.c int swapcache_prepare(swp_entry_t entry) entry 3476 mm/swapfile.c return __swap_duplicate(entry, SWAP_HAS_CACHE); entry 3479 mm/swapfile.c struct swap_info_struct *swp_swap_info(swp_entry_t entry) entry 3481 mm/swapfile.c return swap_type_to_swap_info(swp_type(entry)); entry 3486 mm/swapfile.c swp_entry_t entry = { .val = page_private(page) }; entry 3487 mm/swapfile.c return swp_swap_info(entry); entry 3521 mm/swapfile.c int add_swap_count_continuation(swp_entry_t entry, gfp_t gfp_mask) entry 3538 mm/swapfile.c si = get_swap_device(entry); entry 3548 mm/swapfile.c offset = swp_offset(entry); entry 35 mm/truncate.c pgoff_t index, void *entry) entry 40 mm/truncate.c if (xas_load(&xas) != entry) entry 47 mm/truncate.c void *entry) entry 50 mm/truncate.c __clear_shadow_entry(mapping, index, entry); entry 112 mm/truncate.c pgoff_t index, void *entry) entry 117 mm/truncate.c clear_shadow_entry(mapping, index, entry); entry 126 mm/truncate.c pgoff_t index, void *entry) entry 133 mm/truncate.c clear_shadow_entry(mapping, index, entry); entry 677 mm/util.c swp_entry_t entry; entry 679 mm/util.c entry.val = page_private(page); entry 680 mm/util.c return swap_address_space(entry); entry 199 mm/workingset.c unsigned long entry = xa_to_value(shadow); entry 203 mm/workingset.c workingset = entry & 1; entry 204 mm/workingset.c entry >>= 1; entry 205 mm/workingset.c nid = entry & ((1UL << NODES_SHIFT) - 1); entry 206 mm/workingset.c entry >>= NODES_SHIFT; entry 207 mm/workingset.c memcgid = entry & ((1UL << MEM_CGROUP_ID_SHIFT) - 1); entry 208 mm/workingset.c entry >>= MEM_CGROUP_ID_SHIFT; entry 212 mm/workingset.c *evictionp = entry << bucket_order; entry 250 mm/zswap.c struct zswap_entry *entry; entry 251 mm/zswap.c entry = kmem_cache_alloc(zswap_entry_cache, gfp); entry 252 mm/zswap.c if (!entry) entry 254 mm/zswap.c entry->refcount = 1; entry 255 mm/zswap.c RB_CLEAR_NODE(&entry->rbnode); entry 256 mm/zswap.c return entry; entry 259 mm/zswap.c static void zswap_entry_cache_free(struct zswap_entry *entry) entry 261 mm/zswap.c kmem_cache_free(zswap_entry_cache, entry); entry 270 mm/zswap.c struct zswap_entry *entry; entry 273 mm/zswap.c entry = rb_entry(node, struct zswap_entry, rbnode); entry 274 mm/zswap.c if (entry->offset > offset) entry 276 mm/zswap.c else if (entry->offset < offset) entry 279 mm/zswap.c return entry; entry 288 mm/zswap.c static int zswap_rb_insert(struct rb_root *root, struct zswap_entry *entry, entry 297 mm/zswap.c if (myentry->offset > entry->offset) entry 299 mm/zswap.c else if (myentry->offset < entry->offset) entry 306 mm/zswap.c rb_link_node(&entry->rbnode, parent, link); entry 307 mm/zswap.c rb_insert_color(&entry->rbnode, root); entry 311 mm/zswap.c static void zswap_rb_erase(struct rb_root *root, struct zswap_entry *entry) entry 313 mm/zswap.c if (!RB_EMPTY_NODE(&entry->rbnode)) { entry 314 mm/zswap.c rb_erase(&entry->rbnode, root); entry 315 mm/zswap.c RB_CLEAR_NODE(&entry->rbnode); entry 323 mm/zswap.c static void zswap_free_entry(struct zswap_entry *entry) entry 325 mm/zswap.c if (!entry->length) entry 328 mm/zswap.c zpool_free(entry->pool->zpool, entry->handle); entry 329 mm/zswap.c zswap_pool_put(entry->pool); entry 331 mm/zswap.c zswap_entry_cache_free(entry); entry 337 mm/zswap.c static void zswap_entry_get(struct zswap_entry *entry) entry 339 mm/zswap.c entry->refcount++; entry 346 mm/zswap.c struct zswap_entry *entry) entry 348 mm/zswap.c int refcount = --entry->refcount; entry 352 mm/zswap.c zswap_rb_erase(&tree->rbroot, entry); entry 353 mm/zswap.c zswap_free_entry(entry); entry 361 mm/zswap.c struct zswap_entry *entry; entry 363 mm/zswap.c entry = zswap_rb_search(root, offset); entry 364 mm/zswap.c if (entry) entry 365 mm/zswap.c zswap_entry_get(entry); entry 367 mm/zswap.c return entry; entry 814 mm/zswap.c static int zswap_get_swap_cache_page(swp_entry_t entry, entry 819 mm/zswap.c *retpage = __read_swap_cache_async(entry, GFP_KERNEL, entry 846 mm/zswap.c struct zswap_entry *entry; entry 864 mm/zswap.c entry = zswap_entry_find_get(&tree->rbroot, offset); entry 865 mm/zswap.c if (!entry) { entry 872 mm/zswap.c BUG_ON(offset != entry->offset); entry 891 mm/zswap.c tfm = *get_cpu_ptr(entry->pool->tfm); entry 892 mm/zswap.c ret = crypto_comp_decompress(tfm, src, entry->length, entry 894 mm/zswap.c put_cpu_ptr(entry->pool->tfm); entry 913 mm/zswap.c zswap_entry_put(tree, entry); entry 922 mm/zswap.c if (entry == zswap_rb_search(&tree->rbroot, offset)) entry 923 mm/zswap.c zswap_entry_put(tree, entry); entry 937 mm/zswap.c zswap_entry_put(tree, entry); entry 991 mm/zswap.c struct zswap_entry *entry, *dupentry; entry 1032 mm/zswap.c entry = zswap_entry_cache_alloc(GFP_KERNEL); entry 1033 mm/zswap.c if (!entry) { entry 1043 mm/zswap.c entry->offset = offset; entry 1044 mm/zswap.c entry->length = 0; entry 1045 mm/zswap.c entry->value = value; entry 1053 mm/zswap.c entry->pool = zswap_pool_current_get(); entry 1054 mm/zswap.c if (!entry->pool) { entry 1061 mm/zswap.c tfm = *get_cpu_ptr(entry->pool->tfm); entry 1065 mm/zswap.c put_cpu_ptr(entry->pool->tfm); entry 1072 mm/zswap.c hlen = zpool_evictable(entry->pool->zpool) ? sizeof(zhdr) : 0; entry 1074 mm/zswap.c if (zpool_malloc_support_movable(entry->pool->zpool)) entry 1076 mm/zswap.c ret = zpool_malloc(entry->pool->zpool, hlen + dlen, gfp, &handle); entry 1085 mm/zswap.c buf = zpool_map_handle(entry->pool->zpool, handle, ZPOOL_MM_RW); entry 1088 mm/zswap.c zpool_unmap_handle(entry->pool->zpool, handle); entry 1092 mm/zswap.c entry->offset = offset; entry 1093 mm/zswap.c entry->handle = handle; entry 1094 mm/zswap.c entry->length = dlen; entry 1100 mm/zswap.c ret = zswap_rb_insert(&tree->rbroot, entry, &dupentry); entry 1118 mm/zswap.c zswap_pool_put(entry->pool); entry 1120 mm/zswap.c zswap_entry_cache_free(entry); entry 1133 mm/zswap.c struct zswap_entry *entry; entry 1141 mm/zswap.c entry = zswap_entry_find_get(&tree->rbroot, offset); entry 1142 mm/zswap.c if (!entry) { entry 1149 mm/zswap.c if (!entry->length) { entry 1151 mm/zswap.c zswap_fill_page(dst, entry->value); entry 1158 mm/zswap.c src = zpool_map_handle(entry->pool->zpool, entry->handle, ZPOOL_MM_RO); entry 1159 mm/zswap.c if (zpool_evictable(entry->pool->zpool)) entry 1162 mm/zswap.c tfm = *get_cpu_ptr(entry->pool->tfm); entry 1163 mm/zswap.c ret = crypto_comp_decompress(tfm, src, entry->length, dst, &dlen); entry 1164 mm/zswap.c put_cpu_ptr(entry->pool->tfm); entry 1166 mm/zswap.c zpool_unmap_handle(entry->pool->zpool, entry->handle); entry 1171 mm/zswap.c zswap_entry_put(tree, entry); entry 1181 mm/zswap.c struct zswap_entry *entry; entry 1185 mm/zswap.c entry = zswap_rb_search(&tree->rbroot, offset); entry 1186 mm/zswap.c if (!entry) { entry 1193 mm/zswap.c zswap_rb_erase(&tree->rbroot, entry); entry 1196 mm/zswap.c zswap_entry_put(tree, entry); entry 1205 mm/zswap.c struct zswap_entry *entry, *n; entry 1212 mm/zswap.c rbtree_postorder_for_each_entry_safe(entry, n, &tree->rbroot, rbnode) entry 1213 mm/zswap.c zswap_free_entry(entry); entry 484 net/appletalk/aarp.c struct aarp_entry *entry; entry 499 net/appletalk/aarp.c entry = aarp_alloc(); entry 501 net/appletalk/aarp.c if (!entry) entry 504 net/appletalk/aarp.c entry->expires_at = -1; entry 505 net/appletalk/aarp.c entry->status = ATIF_PROBE; entry 506 net/appletalk/aarp.c entry->target_addr.s_node = sa->s_node; entry 507 net/appletalk/aarp.c entry->target_addr.s_net = sa->s_net; entry 508 net/appletalk/aarp.c entry->dev = atif->dev; entry 513 net/appletalk/aarp.c entry->next = proxies[hash]; entry 514 net/appletalk/aarp.c proxies[hash] = entry; entry 524 net/appletalk/aarp.c if (entry->status & ATIF_PROBE_FAIL) entry 528 net/appletalk/aarp.c if (entry->status & ATIF_PROBE_FAIL) { entry 529 net/appletalk/aarp.c entry->expires_at = jiffies - 1; /* free the entry */ entry 532 net/appletalk/aarp.c entry->status &= ~ATIF_PROBE; entry 924 net/appletalk/aarp.c struct aarp_entry *entry; entry 928 net/appletalk/aarp.c for (entry = table[ct]; entry; entry = entry->next) { entry 932 net/appletalk/aarp.c return entry; entry 965 net/appletalk/aarp.c struct aarp_entry *entry = v; entry 972 net/appletalk/aarp.c entry = iter_next(iter, NULL); entry 975 net/appletalk/aarp.c else if (entry->next) entry 976 net/appletalk/aarp.c entry = entry->next; entry 981 net/appletalk/aarp.c entry = iter_next(iter, NULL); entry 983 net/appletalk/aarp.c return entry; entry 1004 net/appletalk/aarp.c struct aarp_entry *entry = v; entry 1013 net/appletalk/aarp.c ntohs(entry->target_addr.s_net), entry 1014 net/appletalk/aarp.c (unsigned int) entry->target_addr.s_node, entry 1015 net/appletalk/aarp.c entry->dev ? entry->dev->name : "????"); entry 1016 net/appletalk/aarp.c seq_printf(seq, "%pM", entry->hwaddr); entry 1018 net/appletalk/aarp.c dt2str((long)entry->expires_at - (long)now)); entry 1021 net/appletalk/aarp.c dt2str(now - entry->last_sent), entry 1022 net/appletalk/aarp.c entry->xmit_count); entry 59 net/atm/addr.c list_for_each_entry_safe(this, p, head, entry) { entry 60 net/atm/addr.c list_del(&this->entry); entry 84 net/atm/addr.c list_for_each_entry(this, head, entry) { entry 96 net/atm/addr.c list_add(&this->entry, head); entry 119 net/atm/addr.c list_for_each_entry(this, head, entry) { entry 121 net/atm/addr.c list_del(&this->entry); entry 147 net/atm/addr.c list_for_each_entry(this, head, entry) entry 154 net/atm/addr.c list_for_each_entry(this, head, entry) entry 40 net/atm/atm_sysfs.c list_for_each_entry(aaddr, &adev->local, entry) { entry 76 net/atm/clip.c static void link_vcc(struct clip_vcc *clip_vcc, struct atmarp_entry *entry) entry 78 net/atm/clip.c pr_debug("%p to entry %p (neigh %p)\n", clip_vcc, entry, entry->neigh); entry 79 net/atm/clip.c clip_vcc->entry = entry; entry 81 net/atm/clip.c clip_vcc->next = entry->vccs; entry 82 net/atm/clip.c entry->vccs = clip_vcc; entry 83 net/atm/clip.c entry->neigh->used = jiffies; entry 88 net/atm/clip.c struct atmarp_entry *entry = clip_vcc->entry; entry 91 net/atm/clip.c if (!entry) { entry 95 net/atm/clip.c netif_tx_lock_bh(entry->neigh->dev); /* block clip_start_xmit() */ entry 96 net/atm/clip.c entry->neigh->used = jiffies; entry 97 net/atm/clip.c for (walk = &entry->vccs; *walk; walk = &(*walk)->next) entry 102 net/atm/clip.c clip_vcc->entry = NULL; entry 104 net/atm/clip.c netif_wake_queue(entry->neigh->dev); entry 105 net/atm/clip.c if (entry->vccs) entry 107 net/atm/clip.c entry->expires = jiffies - 1; entry 109 net/atm/clip.c error = neigh_update(entry->neigh, NULL, NUD_NONE, entry 115 net/atm/clip.c pr_err("ATMARP: failed (entry %p, vcc 0x%p)\n", entry, clip_vcc); entry 117 net/atm/clip.c netif_tx_unlock_bh(entry->neigh->dev); entry 123 net/atm/clip.c struct atmarp_entry *entry = neighbour_priv(n); entry 128 net/atm/clip.c for (cv = entry->vccs; cv; cv = cv->next) { entry 133 net/atm/clip.c cv, cv->vcc, entry); entry 138 net/atm/clip.c if (entry->vccs || time_before(jiffies, entry->expires)) entry 204 net/atm/clip.c if (clip_vcc->entry) entry 211 net/atm/clip.c skb->dev = clip_vcc->entry ? clip_vcc->entry->neigh->dev : clip_devs; entry 292 net/atm/clip.c struct atmarp_entry *entry = neighbour_priv(neigh); entry 303 net/atm/clip.c entry->neigh = neigh; entry 304 net/atm/clip.c entry->vccs = NULL; entry 305 net/atm/clip.c entry->expires = jiffies - 1; entry 333 net/atm/clip.c struct atmarp_entry *entry; entry 360 net/atm/clip.c entry = neighbour_priv(n); entry 361 net/atm/clip.c if (!entry->vccs) { entry 362 net/atm/clip.c if (time_after(jiffies, entry->expires)) { entry 364 net/atm/clip.c entry->expires = jiffies + ATMARP_RETRY_DELAY * HZ; entry 367 net/atm/clip.c if (entry->neigh->arp_queue.qlen < ATMARP_MAX_UNRES_PACKETS) entry 368 net/atm/clip.c skb_queue_tail(&entry->neigh->arp_queue, skb); entry 375 net/atm/clip.c pr_debug("neigh %p, vccs %p\n", entry, entry->vccs); entry 376 net/atm/clip.c ATM_SKB(skb)->vcc = vcc = entry->vccs->vcc; entry 378 net/atm/clip.c if (entry->vccs->encap) { entry 386 net/atm/clip.c entry->vccs->last_use = jiffies; entry 388 net/atm/clip.c old = xchg(&entry->vccs->xoff, 1); /* assume XOFF ... */ entry 397 net/atm/clip.c entry->vccs->xoff = 0; entry 403 net/atm/clip.c if (!entry->vccs->xoff) entry 428 net/atm/clip.c clip_vcc->entry = NULL; entry 447 net/atm/clip.c struct atmarp_entry *entry; entry 458 net/atm/clip.c if (!clip_vcc->entry) { entry 473 net/atm/clip.c entry = neighbour_priv(neigh); entry 474 net/atm/clip.c if (entry != clip_vcc->entry) { entry 475 net/atm/clip.c if (!clip_vcc->entry) entry 481 net/atm/clip.c link_vcc(clip_vcc, entry); entry 737 net/atm/clip.c struct atmarp_entry *entry, struct clip_vcc *clip_vcc) entry 750 net/atm/clip.c exp = entry->neigh->used; entry 766 net/atm/clip.c if (time_before(jiffies, entry->expires)) entry 770 net/atm/clip.c refcount_read(&entry->neigh->refcnt)); entry 106 net/atm/lec.c static inline void lec_arp_hold(struct lec_arp_table *entry) entry 108 net/atm/lec.c refcount_inc(&entry->usage); entry 111 net/atm/lec.c static inline void lec_arp_put(struct lec_arp_table *entry) entry 113 net/atm/lec.c if (refcount_dec_and_test(&entry->usage)) entry 114 net/atm/lec.c kfree(entry); entry 211 net/atm/lec.c struct lec_arp_table *entry; entry 282 net/atm/lec.c entry = NULL; entry 283 net/atm/lec.c vcc = lec_arp_resolve(priv, dst, is_rdesc, &entry); entry 285 net/atm/lec.c dev->name, vcc, vcc ? vcc->flags : 0, entry); entry 287 net/atm/lec.c if (entry && (entry->tx_wait.qlen < LEC_UNRES_QUE_LEN)) { entry 290 net/atm/lec.c skb_queue_tail(&entry->tx_wait, skb); entry 304 net/atm/lec.c while (entry && (skb2 = skb_dequeue(&entry->tx_wait))) { entry 328 net/atm/lec.c if (entry) entry 329 net/atm/lec.c lec_arp_put(entry); entry 347 net/atm/lec.c struct lec_arp_table *entry; entry 377 net/atm/lec.c entry = lec_arp_find(priv, mesg->content.normal.mac_addr); entry 378 net/atm/lec.c lec_arp_remove(priv, entry); entry 616 net/atm/lec.c struct lec_arp_table *entry; entry 639 net/atm/lec.c entry = lec_arp_find(priv, src); entry 640 net/atm/lec.c if (entry && entry->vcc != vcc) { entry 641 net/atm/lec.c lec_arp_remove(priv, entry); entry 642 net/atm/lec.c lec_arp_put(entry); entry 800 net/atm/lec.c static void lec_info(struct seq_file *seq, struct lec_arp_table *entry) entry 805 net/atm/lec.c seq_printf(seq, "%2.2x", entry->mac_addr[i] & 0xff); entry 808 net/atm/lec.c seq_printf(seq, "%2.2x", entry->atm_addr[i] & 0xff); entry 809 net/atm/lec.c seq_printf(seq, " %s %4.4x", lec_arp_get_status_string(entry->status), entry 810 net/atm/lec.c entry->flags & 0xffff); entry 811 net/atm/lec.c if (entry->vcc) entry 812 net/atm/lec.c seq_printf(seq, "%3d %3d ", entry->vcc->vpi, entry->vcc->vci); entry 815 net/atm/lec.c if (entry->recv_vcc) { entry 816 net/atm/lec.c seq_printf(seq, " %3d %3d", entry->recv_vcc->vpi, entry 817 net/atm/lec.c entry->recv_vcc->vci); entry 979 net/atm/lec.c struct lec_arp_table *entry = hlist_entry(state->node, entry 984 net/atm/lec.c lec_info(seq, entry); entry 1177 net/atm/lec.c struct lec_arp_table *entry = lec_arp_find(priv, mac_addr); entry 1179 net/atm/lec.c if (entry == NULL) entry 1182 net/atm/lec.c kfree(entry->tlvs); entry 1184 net/atm/lec.c entry->tlvs = kmemdup(tlvs, sizeoftlvs, GFP_KERNEL); entry 1185 net/atm/lec.c if (entry->tlvs == NULL) entry 1187 net/atm/lec.c entry->sizeoftlvs = sizeoftlvs; entry 1255 net/atm/lec.c static void lec_arp_clear_vccs(struct lec_arp_table *entry) entry 1257 net/atm/lec.c if (entry->vcc) { entry 1258 net/atm/lec.c struct atm_vcc *vcc = entry->vcc; entry 1267 net/atm/lec.c vcc->push = entry->old_push; entry 1269 net/atm/lec.c entry->vcc = NULL; entry 1271 net/atm/lec.c if (entry->recv_vcc) { entry 1272 net/atm/lec.c entry->recv_vcc->push = entry->old_recv_push; entry 1273 net/atm/lec.c vcc_release_async(entry->recv_vcc, -EPIPE); entry 1274 net/atm/lec.c entry->recv_vcc = NULL; entry 1283 net/atm/lec.c lec_arp_add(struct lec_priv *priv, struct lec_arp_table *entry) entry 1287 net/atm/lec.c tmp = &priv->lec_arp_tables[HASH(entry->mac_addr[ETH_ALEN - 1])]; entry 1288 net/atm/lec.c hlist_add_head(&entry->next, tmp); entry 1290 net/atm/lec.c pr_debug("Added entry:%pM\n", entry->mac_addr); entry 1299 net/atm/lec.c struct lec_arp_table *entry; entry 1317 net/atm/lec.c hlist_for_each_entry(entry, entry 1320 net/atm/lec.c entry->atm_addr, ATM_ESA_LEN) == 0) { entry 1478 net/atm/lec.c struct lec_arp_table *entry; entry 1489 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 1491 net/atm/lec.c lec_arp_remove(priv, entry); entry 1492 net/atm/lec.c lec_arp_put(entry); entry 1497 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 1499 net/atm/lec.c del_timer_sync(&entry->timer); entry 1500 net/atm/lec.c lec_arp_clear_vccs(entry); entry 1501 net/atm/lec.c hlist_del(&entry->next); entry 1502 net/atm/lec.c lec_arp_put(entry); entry 1506 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 1508 net/atm/lec.c del_timer_sync(&entry->timer); entry 1509 net/atm/lec.c lec_arp_clear_vccs(entry); entry 1510 net/atm/lec.c hlist_del(&entry->next); entry 1511 net/atm/lec.c lec_arp_put(entry); entry 1515 net/atm/lec.c hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) { entry 1517 net/atm/lec.c lec_arp_clear_vccs(entry); entry 1518 net/atm/lec.c hlist_del(&entry->next); entry 1519 net/atm/lec.c lec_arp_put(entry); entry 1533 net/atm/lec.c struct lec_arp_table *entry; entry 1538 net/atm/lec.c hlist_for_each_entry(entry, head, next) { entry 1539 net/atm/lec.c if (ether_addr_equal(mac_addr, entry->mac_addr)) entry 1540 net/atm/lec.c return entry; entry 1568 net/atm/lec.c struct lec_arp_table *entry; entry 1570 net/atm/lec.c entry = from_timer(entry, t, timer); entry 1573 net/atm/lec.c if (entry->status == ESI_ARP_PENDING) { entry 1574 net/atm/lec.c if (entry->no_tries <= entry->priv->max_retry_count) { entry 1575 net/atm/lec.c if (entry->is_rdesc) entry 1576 net/atm/lec.c send_to_lecd(entry->priv, l_rdesc_arp_xmt, entry 1577 net/atm/lec.c entry->mac_addr, NULL, NULL); entry 1579 net/atm/lec.c send_to_lecd(entry->priv, l_arp_xmt, entry 1580 net/atm/lec.c entry->mac_addr, NULL, NULL); entry 1581 net/atm/lec.c entry->no_tries++; entry 1583 net/atm/lec.c mod_timer(&entry->timer, jiffies + (1 * HZ)); entry 1609 net/atm/lec.c static bool __lec_arp_check_expire(struct lec_arp_table *entry, entry 1615 net/atm/lec.c if ((entry->flags) & LEC_REMOTE_FLAG && priv->topology_change) entry 1621 net/atm/lec.c now, entry->last_used, time_to_check); entry 1622 net/atm/lec.c if (time_after(now, entry->last_used + time_to_check) && entry 1623 net/atm/lec.c !(entry->flags & LEC_PERMANENT_FLAG) && entry 1624 net/atm/lec.c !(entry->mac_addr[0] & 0x01)) { /* LANE2: 7.1.20 */ entry 1627 net/atm/lec.c lec_arp_remove(priv, entry); entry 1628 net/atm/lec.c lec_arp_put(entry); entry 1631 net/atm/lec.c if ((entry->status == ESI_VC_PENDING || entry 1632 net/atm/lec.c entry->status == ESI_ARP_PENDING) && entry 1633 net/atm/lec.c time_after_eq(now, entry->timestamp + entry 1635 net/atm/lec.c entry->timestamp = jiffies; entry 1636 net/atm/lec.c entry->packets_flooded = 0; entry 1637 net/atm/lec.c if (entry->status == ESI_VC_PENDING) entry 1639 net/atm/lec.c entry->mac_addr, entry 1640 net/atm/lec.c entry->atm_addr, entry 1643 net/atm/lec.c if (entry->status == ESI_FLUSH_PENDING && entry 1644 net/atm/lec.c time_after_eq(now, entry->timestamp + entry 1646 net/atm/lec.c lec_arp_hold(entry); entry 1675 net/atm/lec.c struct lec_arp_table *entry; entry 1684 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 1686 net/atm/lec.c if (__lec_arp_check_expire(entry, now, priv)) { entry 1688 net/atm/lec.c struct atm_vcc *vcc = entry->vcc; entry 1692 net/atm/lec.c while ((skb = skb_dequeue(&entry->tx_wait))) entry 1694 net/atm/lec.c entry->last_used = jiffies; entry 1695 net/atm/lec.c entry->status = ESI_FORWARD_DIRECT; entry 1696 net/atm/lec.c lec_arp_put(entry); entry 1717 net/atm/lec.c struct lec_arp_table *entry; entry 1734 net/atm/lec.c entry = lec_arp_find(priv, mac_to_find); entry 1736 net/atm/lec.c if (entry) { entry 1737 net/atm/lec.c if (entry->status == ESI_FORWARD_DIRECT) { entry 1739 net/atm/lec.c entry->last_used = jiffies; entry 1740 net/atm/lec.c lec_arp_hold(entry); entry 1741 net/atm/lec.c *ret_entry = entry; entry 1742 net/atm/lec.c found = entry->vcc; entry 1749 net/atm/lec.c if (entry->status == ESI_ARP_PENDING) entry 1750 net/atm/lec.c entry->no_tries = 0; entry 1757 net/atm/lec.c if (entry->status != ESI_FLUSH_PENDING && entry 1758 net/atm/lec.c entry->packets_flooded < entry 1760 net/atm/lec.c entry->packets_flooded++; entry 1770 net/atm/lec.c lec_arp_hold(entry); entry 1771 net/atm/lec.c *ret_entry = entry; entry 1772 net/atm/lec.c pr_debug("entry->status %d entry->vcc %p\n", entry->status, entry 1773 net/atm/lec.c entry->vcc); entry 1777 net/atm/lec.c entry = make_entry(priv, mac_to_find); entry 1779 net/atm/lec.c if (!entry) { entry 1783 net/atm/lec.c lec_arp_add(priv, entry); entry 1785 net/atm/lec.c entry->packets_flooded = 1; entry 1786 net/atm/lec.c entry->status = ESI_ARP_PENDING; entry 1787 net/atm/lec.c entry->no_tries = 1; entry 1788 net/atm/lec.c entry->last_used = entry->timestamp = jiffies; entry 1789 net/atm/lec.c entry->is_rdesc = is_rdesc; entry 1790 net/atm/lec.c if (entry->is_rdesc) entry 1795 net/atm/lec.c entry->timer.expires = jiffies + (1 * HZ); entry 1796 net/atm/lec.c entry->timer.function = lec_arp_expire_arp; entry 1797 net/atm/lec.c add_timer(&entry->timer); entry 1812 net/atm/lec.c struct lec_arp_table *entry; entry 1818 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 1820 net/atm/lec.c if (!memcmp(atm_addr, entry->atm_addr, ATM_ESA_LEN) && entry 1822 net/atm/lec.c !(entry->flags & LEC_PERMANENT_FLAG))) { entry 1823 net/atm/lec.c lec_arp_remove(priv, entry); entry 1824 net/atm/lec.c lec_arp_put(entry); entry 1844 net/atm/lec.c struct lec_arp_table *entry, *tmp; entry 1851 net/atm/lec.c entry = lec_arp_find(priv, mac_addr); entry 1852 net/atm/lec.c if (entry == NULL && targetless_le_arp) entry 1858 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 1860 net/atm/lec.c if (memcmp(entry->atm_addr, atm_addr, ATM_ESA_LEN) == 0) { entry 1861 net/atm/lec.c hlist_del(&entry->next); entry 1862 net/atm/lec.c del_timer(&entry->timer); entry 1868 net/atm/lec.c tmp->vcc = entry->vcc; entry 1869 net/atm/lec.c tmp->old_push = entry->old_push; entry 1871 net/atm/lec.c del_timer(&entry->timer); entry 1872 net/atm/lec.c lec_arp_put(entry); entry 1873 net/atm/lec.c entry = tmp; entry 1875 net/atm/lec.c entry->status = ESI_FORWARD_DIRECT; entry 1876 net/atm/lec.c ether_addr_copy(entry->mac_addr, entry 1878 net/atm/lec.c entry->last_used = jiffies; entry 1879 net/atm/lec.c lec_arp_add(priv, entry); entry 1882 net/atm/lec.c entry->flags |= LEC_REMOTE_FLAG; entry 1884 net/atm/lec.c entry->flags &= ~LEC_REMOTE_FLAG; entry 1892 net/atm/lec.c entry = lec_arp_find(priv, mac_addr); entry 1893 net/atm/lec.c if (!entry) { entry 1894 net/atm/lec.c entry = make_entry(priv, mac_addr); entry 1895 net/atm/lec.c if (!entry) entry 1897 net/atm/lec.c entry->status = ESI_UNKNOWN; entry 1898 net/atm/lec.c lec_arp_add(priv, entry); entry 1901 net/atm/lec.c memcpy(entry->atm_addr, atm_addr, ATM_ESA_LEN); entry 1902 net/atm/lec.c del_timer(&entry->timer); entry 1906 net/atm/lec.c if (entry != tmp && entry 1914 net/atm/lec.c entry->vcc = tmp->vcc; entry 1915 net/atm/lec.c entry->old_push = tmp->old_push; entry 1917 net/atm/lec.c entry->status = tmp->status; entry 1923 net/atm/lec.c entry->flags |= LEC_REMOTE_FLAG; entry 1925 net/atm/lec.c entry->flags &= ~LEC_REMOTE_FLAG; entry 1926 net/atm/lec.c if (entry->status == ESI_ARP_PENDING || entry->status == ESI_UNKNOWN) { entry 1927 net/atm/lec.c entry->status = ESI_VC_PENDING; entry 1928 net/atm/lec.c send_to_lecd(priv, l_svc_setup, entry->mac_addr, atm_addr, NULL); entry 1945 net/atm/lec.c struct lec_arp_table *entry; entry 1953 net/atm/lec.c entry = lec_arp_find(priv, bus_mac); entry 1954 net/atm/lec.c if (!entry) { entry 1958 net/atm/lec.c memcpy(entry->atm_addr, ioc_data->atm_addr, ATM_ESA_LEN); entry 1959 net/atm/lec.c entry->recv_vcc = vcc; entry 1960 net/atm/lec.c entry->old_recv_push = old_push; entry 1962 net/atm/lec.c entry = make_entry(priv, bus_mac); entry 1963 net/atm/lec.c if (entry == NULL) entry 1965 net/atm/lec.c del_timer(&entry->timer); entry 1966 net/atm/lec.c memcpy(entry->atm_addr, ioc_data->atm_addr, ATM_ESA_LEN); entry 1967 net/atm/lec.c entry->recv_vcc = vcc; entry 1968 net/atm/lec.c entry->old_recv_push = old_push; entry 1969 net/atm/lec.c hlist_add_head(&entry->next, &priv->mcast_fwds); entry 1987 net/atm/lec.c entry = make_entry(priv, bus_mac); entry 1988 net/atm/lec.c if (entry == NULL) entry 1990 net/atm/lec.c memcpy(entry->atm_addr, ioc_data->atm_addr, ATM_ESA_LEN); entry 1991 net/atm/lec.c eth_zero_addr(entry->mac_addr); entry 1992 net/atm/lec.c entry->recv_vcc = vcc; entry 1993 net/atm/lec.c entry->old_recv_push = old_push; entry 1994 net/atm/lec.c entry->status = ESI_UNKNOWN; entry 1995 net/atm/lec.c entry->timer.expires = jiffies + priv->vcc_timeout_period; entry 1996 net/atm/lec.c entry->timer.function = lec_arp_expire_vcc; entry 1997 net/atm/lec.c hlist_add_head(&entry->next, &priv->lec_no_forward); entry 1998 net/atm/lec.c add_timer(&entry->timer); entry 2014 net/atm/lec.c hlist_for_each_entry(entry, entry 2017 net/atm/lec.c (ioc_data->atm_addr, entry->atm_addr, entry 2021 net/atm/lec.c entry->vcc ? entry->vcc->vci : 0, entry 2022 net/atm/lec.c entry->recv_vcc ? entry->recv_vcc-> entry 2025 net/atm/lec.c del_timer(&entry->timer); entry 2026 net/atm/lec.c entry->vcc = vcc; entry 2027 net/atm/lec.c entry->old_push = old_push; entry 2028 net/atm/lec.c if (entry->status == ESI_VC_PENDING) { entry 2031 net/atm/lec.c entry->status = entry 2034 net/atm/lec.c entry->timestamp = jiffies; entry 2035 net/atm/lec.c entry->status = entry 2040 net/atm/lec.c entry->atm_addr, entry 2070 net/atm/lec.c entry = make_entry(priv, bus_mac); entry 2071 net/atm/lec.c if (!entry) entry 2073 net/atm/lec.c entry->vcc = vcc; entry 2074 net/atm/lec.c entry->old_push = old_push; entry 2075 net/atm/lec.c memcpy(entry->atm_addr, ioc_data->atm_addr, ATM_ESA_LEN); entry 2076 net/atm/lec.c eth_zero_addr(entry->mac_addr); entry 2077 net/atm/lec.c entry->status = ESI_UNKNOWN; entry 2078 net/atm/lec.c hlist_add_head(&entry->next, &priv->lec_arp_empty_ones); entry 2079 net/atm/lec.c entry->timer.expires = jiffies + priv->vcc_timeout_period; entry 2080 net/atm/lec.c entry->timer.function = lec_arp_expire_vcc; entry 2081 net/atm/lec.c add_timer(&entry->timer); entry 2091 net/atm/lec.c struct lec_arp_table *entry; entry 2098 net/atm/lec.c hlist_for_each_entry(entry, entry 2100 net/atm/lec.c if (entry->flush_tran_id == tran_id && entry 2101 net/atm/lec.c entry->status == ESI_FLUSH_PENDING) { entry 2103 net/atm/lec.c struct atm_vcc *vcc = entry->vcc; entry 2105 net/atm/lec.c lec_arp_hold(entry); entry 2108 net/atm/lec.c while ((skb = skb_dequeue(&entry->tx_wait))) entry 2110 net/atm/lec.c entry->last_used = jiffies; entry 2111 net/atm/lec.c entry->status = ESI_FORWARD_DIRECT; entry 2112 net/atm/lec.c lec_arp_put(entry); entry 2127 net/atm/lec.c struct lec_arp_table *entry; entry 2132 net/atm/lec.c hlist_for_each_entry(entry, entry 2134 net/atm/lec.c if (!memcmp(atm_addr, entry->atm_addr, ATM_ESA_LEN)) { entry 2135 net/atm/lec.c entry->flush_tran_id = tran_id; entry 2137 net/atm/lec.c tran_id, entry); entry 2185 net/atm/lec.c struct lec_arp_table *entry; entry 2194 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 2196 net/atm/lec.c if (vcc == entry->vcc) { entry 2197 net/atm/lec.c lec_arp_remove(priv, entry); entry 2198 net/atm/lec.c lec_arp_put(entry); entry 2205 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 2207 net/atm/lec.c if (entry->vcc == vcc) { entry 2208 net/atm/lec.c lec_arp_clear_vccs(entry); entry 2209 net/atm/lec.c del_timer(&entry->timer); entry 2210 net/atm/lec.c hlist_del(&entry->next); entry 2211 net/atm/lec.c lec_arp_put(entry); entry 2215 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 2217 net/atm/lec.c if (entry->recv_vcc == vcc) { entry 2218 net/atm/lec.c lec_arp_clear_vccs(entry); entry 2219 net/atm/lec.c del_timer(&entry->timer); entry 2220 net/atm/lec.c hlist_del(&entry->next); entry 2221 net/atm/lec.c lec_arp_put(entry); entry 2225 net/atm/lec.c hlist_for_each_entry_safe(entry, next, &priv->mcast_fwds, next) { entry 2226 net/atm/lec.c if (entry->recv_vcc == vcc) { entry 2227 net/atm/lec.c lec_arp_clear_vccs(entry); entry 2229 net/atm/lec.c hlist_del(&entry->next); entry 2230 net/atm/lec.c lec_arp_put(entry); entry 2244 net/atm/lec.c struct lec_arp_table *entry, *tmp; entry 2249 net/atm/lec.c hlist_for_each_entry_safe(entry, next, entry 2251 net/atm/lec.c if (vcc == entry->vcc) { entry 2252 net/atm/lec.c del_timer(&entry->timer); entry 2253 net/atm/lec.c ether_addr_copy(entry->mac_addr, src); entry 2254 net/atm/lec.c entry->status = ESI_FORWARD_DIRECT; entry 2255 net/atm/lec.c entry->last_used = jiffies; entry 2262 net/atm/lec.c hlist_del(&entry->next); entry 2263 net/atm/lec.c lec_arp_add(priv, entry); entry 87 net/atm/mpc.c static void purge_egress_shortcut(struct atm_vcc *vcc, eg_cache_entry *entry); entry 179 net/atm/mpc.c struct atm_mpoa_qos *entry; entry 181 net/atm/mpc.c entry = atm_mpoa_search_qos(dst_ip); entry 182 net/atm/mpc.c if (entry != NULL) { entry 183 net/atm/mpc.c entry->qos = *qos; entry 184 net/atm/mpc.c return entry; entry 187 net/atm/mpc.c entry = kmalloc(sizeof(struct atm_mpoa_qos), GFP_KERNEL); entry 188 net/atm/mpc.c if (entry == NULL) { entry 190 net/atm/mpc.c return entry; entry 193 net/atm/mpc.c entry->ipaddr = dst_ip; entry 194 net/atm/mpc.c entry->qos = *qos; entry 196 net/atm/mpc.c entry->next = qos_head; entry 197 net/atm/mpc.c qos_head = entry; entry 199 net/atm/mpc.c return entry; entry 219 net/atm/mpc.c int atm_mpoa_delete_qos(struct atm_mpoa_qos *entry) entry 223 net/atm/mpc.c if (entry == NULL) entry 225 net/atm/mpc.c if (entry == qos_head) { entry 227 net/atm/mpc.c kfree(entry); entry 233 net/atm/mpc.c if (curr->next == entry) { entry 234 net/atm/mpc.c curr->next = entry->next; entry 235 net/atm/mpc.c kfree(entry); entry 494 net/atm/mpc.c in_cache_entry *entry; entry 514 net/atm/mpc.c entry = mpc->in_ops->get(ipaddr, mpc); entry 515 net/atm/mpc.c if (entry == NULL) { entry 516 net/atm/mpc.c entry = mpc->in_ops->add_entry(ipaddr, mpc); entry 517 net/atm/mpc.c if (entry != NULL) entry 518 net/atm/mpc.c mpc->in_ops->put(entry); entry 522 net/atm/mpc.c if (mpc->in_ops->cache_hit(entry, mpc) != OPEN) { entry 525 net/atm/mpc.c mpc->in_ops->put(entry); entry 535 net/atm/mpc.c mpc->in_ops->put(entry); entry 542 net/atm/mpc.c if (entry->ctrl_info.tag != 0) { entry 544 net/atm/mpc.c mpc->dev->name, entry->ctrl_info.tag); entry 545 net/atm/mpc.c tagged_llc_snap_hdr.tag = entry->ctrl_info.tag; entry 559 net/atm/mpc.c atm_account_tx(entry->shortcut, skb); entry 560 net/atm/mpc.c entry->shortcut->send(entry->shortcut, skb); entry 561 net/atm/mpc.c entry->packets_fwded++; entry 562 net/atm/mpc.c mpc->in_ops->put(entry); entry 1083 net/atm/mpc.c in_cache_entry *entry; entry 1085 net/atm/mpc.c entry = mpc->in_ops->get(dst_ip, mpc); entry 1086 net/atm/mpc.c if (entry == NULL) { entry 1087 net/atm/mpc.c entry = mpc->in_ops->add_entry(dst_ip, mpc); entry 1088 net/atm/mpc.c entry->entry_state = INGRESS_RESOLVING; entry 1090 net/atm/mpc.c msg->content.in_info = entry->ctrl_info; entry 1092 net/atm/mpc.c entry->reply_wait = ktime_get_seconds(); entry 1093 net/atm/mpc.c mpc->in_ops->put(entry); entry 1097 net/atm/mpc.c if (entry->entry_state == INGRESS_INVALID) { entry 1098 net/atm/mpc.c entry->entry_state = INGRESS_RESOLVING; entry 1100 net/atm/mpc.c msg->content.in_info = entry->ctrl_info; entry 1102 net/atm/mpc.c entry->reply_wait = ktime_get_seconds(); entry 1103 net/atm/mpc.c mpc->in_ops->put(entry); entry 1109 net/atm/mpc.c mpc->in_ops->put(entry); entry 1118 net/atm/mpc.c in_cache_entry *entry) entry 1129 net/atm/mpc.c entry->shortcut = eg_entry->shortcut; entry 1131 net/atm/mpc.c entry->shortcut = eg_entry->shortcut; entry 1133 net/atm/mpc.c if (entry->shortcut) { entry 1158 net/atm/mpc.c in_cache_entry *entry = mpc->in_ops->get(dst_ip, mpc); entry 1163 net/atm/mpc.c mpc->dev->name, entry); entry 1164 net/atm/mpc.c if (entry == NULL) { entry 1169 net/atm/mpc.c ddprintk_cont(" entry_state = %d ", entry->entry_state); entry 1171 net/atm/mpc.c if (entry->entry_state == INGRESS_RESOLVED) { entry 1173 net/atm/mpc.c mpc->in_ops->put(entry); entry 1177 net/atm/mpc.c entry->ctrl_info = msg->content.in_info; entry 1178 net/atm/mpc.c entry->time = ktime_get_seconds(); entry 1180 net/atm/mpc.c entry->reply_wait = ktime_get_seconds(); entry 1181 net/atm/mpc.c entry->refresh_time = 0; entry 1182 net/atm/mpc.c ddprintk_cont("entry->shortcut = %p\n", entry->shortcut); entry 1184 net/atm/mpc.c if (entry->entry_state == INGRESS_RESOLVING && entry 1185 net/atm/mpc.c entry->shortcut != NULL) { entry 1186 net/atm/mpc.c entry->entry_state = INGRESS_RESOLVED; entry 1187 net/atm/mpc.c mpc->in_ops->put(entry); entry 1191 net/atm/mpc.c if (entry->shortcut != NULL) { entry 1194 net/atm/mpc.c mpc->in_ops->put(entry); entry 1198 net/atm/mpc.c check_qos_and_open_shortcut(msg, mpc, entry); entry 1199 net/atm/mpc.c entry->entry_state = INGRESS_RESOLVED; entry 1200 net/atm/mpc.c mpc->in_ops->put(entry); entry 1210 net/atm/mpc.c in_cache_entry *entry = mpc->in_ops->get_with_mask(dst_ip, mpc, mask); entry 1212 net/atm/mpc.c if (entry == NULL) { entry 1222 net/atm/mpc.c mpc->in_ops->remove_entry(entry, mpc); entry 1224 net/atm/mpc.c mpc->in_ops->put(entry); entry 1225 net/atm/mpc.c entry = mpc->in_ops->get_with_mask(dst_ip, mpc, mask); entry 1226 net/atm/mpc.c } while (entry != NULL); entry 1232 net/atm/mpc.c eg_cache_entry *entry = mpc->eg_ops->get_by_cache_id(cache_id, mpc); entry 1234 net/atm/mpc.c if (entry == NULL) { entry 1241 net/atm/mpc.c mpc->eg_ops->remove_entry(entry, mpc); entry 1244 net/atm/mpc.c mpc->eg_ops->put(entry); entry 1247 net/atm/mpc.c static void purge_egress_shortcut(struct atm_vcc *vcc, eg_cache_entry *entry) entry 1269 net/atm/mpc.c if (entry != NULL) entry 1270 net/atm/mpc.c purge_msg->content.eg_info = entry->ctrl_info; entry 1286 net/atm/mpc.c eg_cache_entry *entry; entry 1297 net/atm/mpc.c entry = mpc->eg_cache; entry 1298 net/atm/mpc.c while (entry != NULL) { entry 1299 net/atm/mpc.c purge_egress_shortcut(entry->shortcut, entry); entry 1300 net/atm/mpc.c entry = entry->next; entry 1312 net/atm/mpc.c eg_cache_entry *entry = mpc->eg_ops->get_by_cache_id(msg->content.eg_info.cache_id, mpc); entry 1316 net/atm/mpc.c mpc->dev->name, entry, holding_time); entry 1317 net/atm/mpc.c if (entry == NULL && holding_time) { entry 1318 net/atm/mpc.c entry = mpc->eg_ops->add_entry(msg, mpc); entry 1319 net/atm/mpc.c mpc->eg_ops->put(entry); entry 1323 net/atm/mpc.c mpc->eg_ops->update(entry, holding_time); entry 1328 net/atm/mpc.c mpc->eg_ops->remove_entry(entry, mpc); entry 1331 net/atm/mpc.c mpc->eg_ops->put(entry); entry 1392 net/atm/mpc.c eg_cache_entry *entry; entry 1398 net/atm/mpc.c entry = mpc->eg_cache; entry 1399 net/atm/mpc.c while (entry != NULL) { entry 1400 net/atm/mpc.c msg->content.eg_info = entry->ctrl_info; entry 1401 net/atm/mpc.c dprintk("cache_id %u\n", entry->ctrl_info.cache_id); entry 1403 net/atm/mpc.c entry = entry->next; entry 38 net/atm/mpoa_caches.c in_cache_entry *entry; entry 41 net/atm/mpoa_caches.c entry = client->in_cache; entry 42 net/atm/mpoa_caches.c while (entry != NULL) { entry 43 net/atm/mpoa_caches.c if (entry->ctrl_info.in_dst_ip == dst_ip) { entry 44 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 46 net/atm/mpoa_caches.c return entry; entry 48 net/atm/mpoa_caches.c entry = entry->next; entry 59 net/atm/mpoa_caches.c in_cache_entry *entry; entry 62 net/atm/mpoa_caches.c entry = client->in_cache; entry 63 net/atm/mpoa_caches.c while (entry != NULL) { entry 64 net/atm/mpoa_caches.c if ((entry->ctrl_info.in_dst_ip & mask) == (dst_ip & mask)) { entry 65 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 67 net/atm/mpoa_caches.c return entry; entry 69 net/atm/mpoa_caches.c entry = entry->next; entry 80 net/atm/mpoa_caches.c in_cache_entry *entry; entry 83 net/atm/mpoa_caches.c entry = client->in_cache; entry 84 net/atm/mpoa_caches.c while (entry != NULL) { entry 85 net/atm/mpoa_caches.c if (entry->shortcut == vcc) { entry 86 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 88 net/atm/mpoa_caches.c return entry; entry 90 net/atm/mpoa_caches.c entry = entry->next; entry 100 net/atm/mpoa_caches.c in_cache_entry *entry = kzalloc(sizeof(in_cache_entry), GFP_KERNEL); entry 102 net/atm/mpoa_caches.c if (entry == NULL) { entry 109 net/atm/mpoa_caches.c refcount_set(&entry->use, 1); entry 112 net/atm/mpoa_caches.c entry->next = client->in_cache; entry 113 net/atm/mpoa_caches.c entry->prev = NULL; entry 115 net/atm/mpoa_caches.c client->in_cache->prev = entry; entry 116 net/atm/mpoa_caches.c client->in_cache = entry; entry 118 net/atm/mpoa_caches.c memcpy(entry->MPS_ctrl_ATM_addr, client->mps_ctrl_addr, ATM_ESA_LEN); entry 119 net/atm/mpoa_caches.c entry->ctrl_info.in_dst_ip = dst_ip; entry 120 net/atm/mpoa_caches.c entry->time = ktime_get_seconds(); entry 121 net/atm/mpoa_caches.c entry->retry_time = client->parameters.mpc_p4; entry 122 net/atm/mpoa_caches.c entry->count = 1; entry 123 net/atm/mpoa_caches.c entry->entry_state = INGRESS_INVALID; entry 124 net/atm/mpoa_caches.c entry->ctrl_info.holding_time = HOLDING_TIME_DEFAULT; entry 125 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 130 net/atm/mpoa_caches.c return entry; entry 133 net/atm/mpoa_caches.c static int cache_hit(in_cache_entry *entry, struct mpoa_client *mpc) entry 138 net/atm/mpoa_caches.c entry->count++; entry 139 net/atm/mpoa_caches.c if (entry->entry_state == INGRESS_RESOLVED && entry->shortcut != NULL) entry 142 net/atm/mpoa_caches.c if (entry->entry_state == INGRESS_REFRESHING) { entry 143 net/atm/mpoa_caches.c if (entry->count > mpc->parameters.mpc_p1) { entry 145 net/atm/mpoa_caches.c msg.content.in_info = entry->ctrl_info; entry 147 net/atm/mpoa_caches.c qos = atm_mpoa_search_qos(entry->ctrl_info.in_dst_ip); entry 151 net/atm/mpoa_caches.c entry->reply_wait = ktime_get_seconds(); entry 152 net/atm/mpoa_caches.c entry->entry_state = INGRESS_RESOLVING; entry 154 net/atm/mpoa_caches.c if (entry->shortcut != NULL) entry 159 net/atm/mpoa_caches.c if (entry->entry_state == INGRESS_RESOLVING && entry->shortcut != NULL) entry 162 net/atm/mpoa_caches.c if (entry->count > mpc->parameters.mpc_p1 && entry 163 net/atm/mpoa_caches.c entry->entry_state == INGRESS_INVALID) { entry 165 net/atm/mpoa_caches.c mpc->dev->name, &entry->ctrl_info.in_dst_ip); entry 166 net/atm/mpoa_caches.c entry->entry_state = INGRESS_RESOLVING; entry 169 net/atm/mpoa_caches.c msg.content.in_info = entry->ctrl_info; entry 170 net/atm/mpoa_caches.c qos = atm_mpoa_search_qos(entry->ctrl_info.in_dst_ip); entry 174 net/atm/mpoa_caches.c entry->reply_wait = ktime_get_seconds(); entry 180 net/atm/mpoa_caches.c static void in_cache_put(in_cache_entry *entry) entry 182 net/atm/mpoa_caches.c if (refcount_dec_and_test(&entry->use)) { entry 183 net/atm/mpoa_caches.c kzfree(entry); entry 190 net/atm/mpoa_caches.c static void in_cache_remove_entry(in_cache_entry *entry, entry 196 net/atm/mpoa_caches.c vcc = entry->shortcut; entry 198 net/atm/mpoa_caches.c &entry->ctrl_info.in_dst_ip); entry 200 net/atm/mpoa_caches.c if (entry->prev != NULL) entry 201 net/atm/mpoa_caches.c entry->prev->next = entry->next; entry 203 net/atm/mpoa_caches.c client->in_cache = entry->next; entry 204 net/atm/mpoa_caches.c if (entry->next != NULL) entry 205 net/atm/mpoa_caches.c entry->next->prev = entry->prev; entry 206 net/atm/mpoa_caches.c client->in_ops->put(entry); entry 228 net/atm/mpoa_caches.c in_cache_entry *entry, *next_entry; entry 234 net/atm/mpoa_caches.c entry = client->in_cache; entry 235 net/atm/mpoa_caches.c while (entry != NULL) { entry 236 net/atm/mpoa_caches.c entry->count = 0; entry 237 net/atm/mpoa_caches.c next_entry = entry->next; entry 238 net/atm/mpoa_caches.c if ((now - entry->time) > entry->ctrl_info.holding_time) { entry 240 net/atm/mpoa_caches.c &entry->ctrl_info.in_dst_ip); entry 241 net/atm/mpoa_caches.c client->in_ops->remove_entry(entry, client); entry 243 net/atm/mpoa_caches.c entry = next_entry; entry 253 net/atm/mpoa_caches.c in_cache_entry *entry; entry 260 net/atm/mpoa_caches.c entry = client->in_cache; entry 261 net/atm/mpoa_caches.c while (entry != NULL) { entry 262 net/atm/mpoa_caches.c if (entry->entry_state == INGRESS_RESOLVING) { entry 264 net/atm/mpoa_caches.c if ((now - entry->hold_down) entry 266 net/atm/mpoa_caches.c entry = entry->next; /* Entry in hold down */ entry 269 net/atm/mpoa_caches.c if ((now - entry->reply_wait) > entry->retry_time) { entry 270 net/atm/mpoa_caches.c entry->retry_time = MPC_C1 * (entry->retry_time); entry 275 net/atm/mpoa_caches.c if (entry->retry_time > client->parameters.mpc_p5) { entry 276 net/atm/mpoa_caches.c entry->hold_down = ktime_get_seconds(); entry 277 net/atm/mpoa_caches.c entry->retry_time = client->parameters.mpc_p4; entry 278 net/atm/mpoa_caches.c entry = entry->next; entry 282 net/atm/mpoa_caches.c memset(&entry->hold_down, 0, sizeof(time64_t)); entry 285 net/atm/mpoa_caches.c msg.content.in_info = entry->ctrl_info; entry 286 net/atm/mpoa_caches.c qos = atm_mpoa_search_qos(entry->ctrl_info.in_dst_ip); entry 290 net/atm/mpoa_caches.c entry->reply_wait = ktime_get_seconds(); entry 293 net/atm/mpoa_caches.c entry = entry->next; entry 302 net/atm/mpoa_caches.c struct in_cache_entry *entry = client->in_cache; entry 308 net/atm/mpoa_caches.c while (entry != NULL) { entry 309 net/atm/mpoa_caches.c if (entry->entry_state == INGRESS_RESOLVED) { entry 310 net/atm/mpoa_caches.c if (!(entry->refresh_time)) entry 311 net/atm/mpoa_caches.c entry->refresh_time = (2 * (entry->ctrl_info.holding_time))/3; entry 312 net/atm/mpoa_caches.c if ((now - entry->reply_wait) > entry 313 net/atm/mpoa_caches.c entry->refresh_time) { entry 315 net/atm/mpoa_caches.c entry->entry_state = INGRESS_REFRESHING; entry 319 net/atm/mpoa_caches.c entry = entry->next; entry 335 net/atm/mpoa_caches.c eg_cache_entry *entry; entry 338 net/atm/mpoa_caches.c entry = mpc->eg_cache; entry 339 net/atm/mpoa_caches.c while (entry != NULL) { entry 340 net/atm/mpoa_caches.c if (entry->ctrl_info.cache_id == cache_id) { entry 341 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 343 net/atm/mpoa_caches.c return entry; entry 345 net/atm/mpoa_caches.c entry = entry->next; entry 356 net/atm/mpoa_caches.c eg_cache_entry *entry; entry 359 net/atm/mpoa_caches.c entry = mpc->eg_cache; entry 360 net/atm/mpoa_caches.c while (entry != NULL) { entry 361 net/atm/mpoa_caches.c if (entry->ctrl_info.tag == tag) { entry 362 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 364 net/atm/mpoa_caches.c return entry; entry 366 net/atm/mpoa_caches.c entry = entry->next; entry 378 net/atm/mpoa_caches.c eg_cache_entry *entry; entry 381 net/atm/mpoa_caches.c entry = mpc->eg_cache; entry 382 net/atm/mpoa_caches.c while (entry != NULL) { entry 383 net/atm/mpoa_caches.c if (entry->shortcut == vcc) { entry 384 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 386 net/atm/mpoa_caches.c return entry; entry 388 net/atm/mpoa_caches.c entry = entry->next; entry 398 net/atm/mpoa_caches.c eg_cache_entry *entry; entry 401 net/atm/mpoa_caches.c entry = mpc->eg_cache; entry 402 net/atm/mpoa_caches.c while (entry != NULL) { entry 403 net/atm/mpoa_caches.c if (entry->latest_ip_addr == ipaddr) { entry 404 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 406 net/atm/mpoa_caches.c return entry; entry 408 net/atm/mpoa_caches.c entry = entry->next; entry 415 net/atm/mpoa_caches.c static void eg_cache_put(eg_cache_entry *entry) entry 417 net/atm/mpoa_caches.c if (refcount_dec_and_test(&entry->use)) { entry 418 net/atm/mpoa_caches.c kzfree(entry); entry 425 net/atm/mpoa_caches.c static void eg_cache_remove_entry(eg_cache_entry *entry, entry 431 net/atm/mpoa_caches.c vcc = entry->shortcut; entry 433 net/atm/mpoa_caches.c if (entry->prev != NULL) entry 434 net/atm/mpoa_caches.c entry->prev->next = entry->next; entry 436 net/atm/mpoa_caches.c client->eg_cache = entry->next; entry 437 net/atm/mpoa_caches.c if (entry->next != NULL) entry 438 net/atm/mpoa_caches.c entry->next->prev = entry->prev; entry 439 net/atm/mpoa_caches.c client->eg_ops->put(entry); entry 459 net/atm/mpoa_caches.c eg_cache_entry *entry = kzalloc(sizeof(eg_cache_entry), GFP_KERNEL); entry 461 net/atm/mpoa_caches.c if (entry == NULL) { entry 469 net/atm/mpoa_caches.c refcount_set(&entry->use, 1); entry 472 net/atm/mpoa_caches.c entry->next = client->eg_cache; entry 473 net/atm/mpoa_caches.c entry->prev = NULL; entry 475 net/atm/mpoa_caches.c client->eg_cache->prev = entry; entry 476 net/atm/mpoa_caches.c client->eg_cache = entry; entry 478 net/atm/mpoa_caches.c memcpy(entry->MPS_ctrl_ATM_addr, client->mps_ctrl_addr, ATM_ESA_LEN); entry 479 net/atm/mpoa_caches.c entry->ctrl_info = msg->content.eg_info; entry 480 net/atm/mpoa_caches.c entry->time = ktime_get_seconds(); entry 481 net/atm/mpoa_caches.c entry->entry_state = EGRESS_RESOLVED; entry 483 net/atm/mpoa_caches.c ntohl(entry->ctrl_info.cache_id)); entry 484 net/atm/mpoa_caches.c dprintk("mps_ip = %pI4\n", &entry->ctrl_info.mps_ip); entry 485 net/atm/mpoa_caches.c refcount_inc(&entry->use); entry 490 net/atm/mpoa_caches.c return entry; entry 493 net/atm/mpoa_caches.c static void update_eg_cache_entry(eg_cache_entry *entry, uint16_t holding_time) entry 495 net/atm/mpoa_caches.c entry->time = ktime_get_seconds(); entry 496 net/atm/mpoa_caches.c entry->entry_state = EGRESS_RESOLVED; entry 497 net/atm/mpoa_caches.c entry->ctrl_info.holding_time = holding_time; entry 502 net/atm/mpoa_caches.c eg_cache_entry *entry, *next_entry; entry 509 net/atm/mpoa_caches.c entry = client->eg_cache; entry 510 net/atm/mpoa_caches.c while (entry != NULL) { entry 511 net/atm/mpoa_caches.c next_entry = entry->next; entry 512 net/atm/mpoa_caches.c if ((now - entry->time) > entry->ctrl_info.holding_time) { entry 514 net/atm/mpoa_caches.c msg.content.eg_info = entry->ctrl_info; entry 516 net/atm/mpoa_caches.c ntohl(entry->ctrl_info.cache_id)); entry 518 net/atm/mpoa_caches.c client->eg_ops->remove_entry(entry, client); entry 520 net/atm/mpoa_caches.c entry = next_entry; entry 43 net/atm/mpoa_caches.h void (*put)(in_cache_entry *entry); entry 46 net/atm/mpoa_caches.h int (*cache_hit)(in_cache_entry *entry, entry 73 net/atm/mpoa_caches.h void (*put)(eg_cache_entry *entry); entry 74 net/atm/mpoa_caches.h void (*remove_entry)(eg_cache_entry *entry, struct mpoa_client *client); entry 75 net/atm/mpoa_caches.h void (*update)(eg_cache_entry *entry, uint16_t holding_time); entry 163 net/atm/proc.c dev = clip_vcc->entry ? clip_vcc->entry->neigh->dev : NULL; entry 489 net/batman-adv/bridge_loop_avoidance.c struct batadv_bla_backbone_gw *entry; entry 493 net/batman-adv/bridge_loop_avoidance.c entry = batadv_backbone_hash_find(bat_priv, orig, vid); entry 495 net/batman-adv/bridge_loop_avoidance.c if (entry) entry 496 net/batman-adv/bridge_loop_avoidance.c return entry; entry 502 net/batman-adv/bridge_loop_avoidance.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 503 net/batman-adv/bridge_loop_avoidance.c if (!entry) entry 506 net/batman-adv/bridge_loop_avoidance.c entry->vid = vid; entry 507 net/batman-adv/bridge_loop_avoidance.c entry->lasttime = jiffies; entry 508 net/batman-adv/bridge_loop_avoidance.c entry->crc = BATADV_BLA_CRC_INIT; entry 509 net/batman-adv/bridge_loop_avoidance.c entry->bat_priv = bat_priv; entry 510 net/batman-adv/bridge_loop_avoidance.c spin_lock_init(&entry->crc_lock); entry 511 net/batman-adv/bridge_loop_avoidance.c atomic_set(&entry->request_sent, 0); entry 512 net/batman-adv/bridge_loop_avoidance.c atomic_set(&entry->wait_periods, 0); entry 513 net/batman-adv/bridge_loop_avoidance.c ether_addr_copy(entry->orig, orig); entry 514 net/batman-adv/bridge_loop_avoidance.c INIT_WORK(&entry->report_work, batadv_bla_loopdetect_report); entry 515 net/batman-adv/bridge_loop_avoidance.c kref_init(&entry->refcount); entry 517 net/batman-adv/bridge_loop_avoidance.c kref_get(&entry->refcount); entry 520 net/batman-adv/bridge_loop_avoidance.c batadv_choose_backbone_gw, entry, entry 521 net/batman-adv/bridge_loop_avoidance.c &entry->hash_entry); entry 525 net/batman-adv/bridge_loop_avoidance.c kfree(entry); entry 538 net/batman-adv/bridge_loop_avoidance.c batadv_bla_send_announce(bat_priv, entry); entry 541 net/batman-adv/bridge_loop_avoidance.c atomic_inc(&entry->request_sent); entry 542 net/batman-adv/bridge_loop_avoidance.c atomic_set(&entry->wait_periods, BATADV_BLA_WAIT_PERIODS); entry 546 net/batman-adv/bridge_loop_avoidance.c return entry; entry 1600 net/batman-adv/bridge_loop_avoidance.c struct batadv_bcast_duplist_entry *entry; entry 1613 net/batman-adv/bridge_loop_avoidance.c entry = &bat_priv->bla.bcast_duplist[curr]; entry 1618 net/batman-adv/bridge_loop_avoidance.c if (batadv_has_timed_out(entry->entrytime, entry 1622 net/batman-adv/bridge_loop_avoidance.c if (entry->crc != crc) entry 1625 net/batman-adv/bridge_loop_avoidance.c if (batadv_compare_eth(entry->orig, bcast_packet->orig)) entry 1639 net/batman-adv/bridge_loop_avoidance.c entry = &bat_priv->bla.bcast_duplist[curr]; entry 1640 net/batman-adv/bridge_loop_avoidance.c entry->crc = crc; entry 1641 net/batman-adv/bridge_loop_avoidance.c entry->entrytime = jiffies; entry 1642 net/batman-adv/bridge_loop_avoidance.c ether_addr_copy(entry->orig, bcast_packet->orig); entry 41 net/batman-adv/fragmentation.c struct batadv_frag_list_entry *entry; entry 44 net/batman-adv/fragmentation.c hlist_for_each_entry_safe(entry, node, head, list) { entry 45 net/batman-adv/fragmentation.c hlist_del(&entry->list); entry 48 net/batman-adv/fragmentation.c kfree_skb(entry->skb); entry 50 net/batman-adv/fragmentation.c consume_skb(entry->skb); entry 52 net/batman-adv/fragmentation.c kfree(entry); entry 252 net/batman-adv/fragmentation.c struct batadv_frag_list_entry *entry; entry 260 net/batman-adv/fragmentation.c entry = hlist_entry(chain->first, struct batadv_frag_list_entry, list); entry 261 net/batman-adv/fragmentation.c hlist_del(&entry->list); entry 262 net/batman-adv/fragmentation.c skb_out = entry->skb; entry 263 net/batman-adv/fragmentation.c kfree(entry); entry 287 net/batman-adv/fragmentation.c hlist_for_each_entry(entry, chain, list) { entry 288 net/batman-adv/fragmentation.c size = entry->skb->len - hdr_size; entry 289 net/batman-adv/fragmentation.c skb_put_data(skb_out, entry->skb->data + hdr_size, size); entry 469 net/batman-adv/translation-table.c struct batadv_tt_change_node *tt_change_node, *entry, *safe; entry 489 net/batman-adv/translation-table.c list_for_each_entry_safe(entry, safe, &bat_priv->tt.changes_list, entry 491 net/batman-adv/translation-table.c if (!batadv_compare_eth(entry->change.addr, common->addr)) entry 501 net/batman-adv/translation-table.c del_op_entry = entry->change.flags & BATADV_TT_CLIENT_DEL; entry 511 net/batman-adv/translation-table.c entry->change.flags = flags; entry 515 net/batman-adv/translation-table.c list_del(&entry->list); entry 516 net/batman-adv/translation-table.c kmem_cache_free(batadv_tt_change_cache, entry); entry 996 net/batman-adv/translation-table.c struct batadv_tt_change_node *entry, *safe; entry 1026 net/batman-adv/translation-table.c list_for_each_entry_safe(entry, safe, &bat_priv->tt.changes_list, entry 1030 net/batman-adv/translation-table.c &entry->change, entry 1034 net/batman-adv/translation-table.c list_del(&entry->list); entry 1035 net/batman-adv/translation-table.c kmem_cache_free(batadv_tt_change_cache, entry); entry 1505 net/batman-adv/translation-table.c struct batadv_tt_change_node *entry, *safe; entry 1509 net/batman-adv/translation-table.c list_for_each_entry_safe(entry, safe, &bat_priv->tt.changes_list, entry 1511 net/batman-adv/translation-table.c list_del(&entry->list); entry 1512 net/batman-adv/translation-table.c kmem_cache_free(batadv_tt_change_cache, entry); entry 1531 net/batman-adv/translation-table.c batadv_tt_global_orig_entry_find(const struct batadv_tt_global_entry *entry, entry 1538 net/batman-adv/translation-table.c head = &entry->orig_list; entry 1566 net/batman-adv/translation-table.c batadv_tt_global_entry_has_orig(const struct batadv_tt_global_entry *entry, entry 1573 net/batman-adv/translation-table.c orig_entry = batadv_tt_global_orig_entry_find(entry, orig_node); entry 228 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry; entry 233 net/bluetooth/6lowpan.c list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) { entry 234 net/bluetooth/6lowpan.c peer = __peer_lookup_conn(entry, conn); entry 246 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry; entry 251 net/bluetooth/6lowpan.c list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) { entry 252 net/bluetooth/6lowpan.c if (conn->hcon->hdev == entry->hdev) { entry 253 net/bluetooth/6lowpan.c dev = entry; entry 490 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry; entry 495 net/bluetooth/6lowpan.c list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) { entry 499 net/bluetooth/6lowpan.c if (entry->netdev != netdev) entry 502 net/bluetooth/6lowpan.c dev = lowpan_btle_dev(entry->netdev); entry 773 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry = container_of(work, entry 777 net/bluetooth/6lowpan.c lowpan_unregister_netdev(entry->netdev); entry 784 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry; entry 804 net/bluetooth/6lowpan.c list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) { entry 805 net/bluetooth/6lowpan.c dev = lowpan_btle_dev(entry->netdev); entry 829 net/bluetooth/6lowpan.c INIT_WORK(&entry->delete_netdev, delete_netdev); entry 830 net/bluetooth/6lowpan.c schedule_work(&entry->delete_netdev); entry 1017 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry; entry 1030 net/bluetooth/6lowpan.c list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) { entry 1031 net/bluetooth/6lowpan.c list_for_each_entry_rcu(peer, &entry->peers, list) { entry 1177 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry; entry 1182 net/bluetooth/6lowpan.c list_for_each_entry(entry, &bt_6lowpan_devices, list) { entry 1183 net/bluetooth/6lowpan.c list_for_each_entry(peer, &entry->peers, list) entry 1208 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry, *tmp, *new_dev; entry 1220 net/bluetooth/6lowpan.c list_for_each_entry_rcu(entry, &bt_6lowpan_devices, list) { entry 1225 net/bluetooth/6lowpan.c new_dev->netdev = entry->netdev; entry 1233 net/bluetooth/6lowpan.c list_for_each_entry_safe(entry, tmp, &devices, list) { entry 1234 net/bluetooth/6lowpan.c ifdown(entry->netdev); entry 1236 net/bluetooth/6lowpan.c entry->netdev->name, entry->netdev); entry 1237 net/bluetooth/6lowpan.c lowpan_unregister_netdev(entry->netdev); entry 1238 net/bluetooth/6lowpan.c kfree(entry); entry 1246 net/bluetooth/6lowpan.c struct lowpan_btle_dev *entry; entry 1254 net/bluetooth/6lowpan.c list_for_each_entry(entry, &bt_6lowpan_devices, list) { entry 1255 net/bluetooth/6lowpan.c if (entry->netdev == netdev) { entry 1258 net/bluetooth/6lowpan.c list_del(&entry->list); entry 2934 net/bluetooth/hci_core.c struct bdaddr_list *entry; entry 2942 net/bluetooth/hci_core.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 2943 net/bluetooth/hci_core.c if (!entry) entry 2946 net/bluetooth/hci_core.c bacpy(&entry->bdaddr, bdaddr); entry 2947 net/bluetooth/hci_core.c entry->bdaddr_type = type; entry 2949 net/bluetooth/hci_core.c list_add(&entry->list, list); entry 2957 net/bluetooth/hci_core.c struct bdaddr_list_with_irk *entry; entry 2965 net/bluetooth/hci_core.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 2966 net/bluetooth/hci_core.c if (!entry) entry 2969 net/bluetooth/hci_core.c bacpy(&entry->bdaddr, bdaddr); entry 2970 net/bluetooth/hci_core.c entry->bdaddr_type = type; entry 2973 net/bluetooth/hci_core.c memcpy(entry->peer_irk, peer_irk, 16); entry 2976 net/bluetooth/hci_core.c memcpy(entry->local_irk, local_irk, 16); entry 2978 net/bluetooth/hci_core.c list_add(&entry->list, list); entry 2985 net/bluetooth/hci_core.c struct bdaddr_list *entry; entry 2992 net/bluetooth/hci_core.c entry = hci_bdaddr_list_lookup(list, bdaddr, type); entry 2993 net/bluetooth/hci_core.c if (!entry) entry 2996 net/bluetooth/hci_core.c list_del(&entry->list); entry 2997 net/bluetooth/hci_core.c kfree(entry); entry 3005 net/bluetooth/hci_core.c struct bdaddr_list_with_irk *entry; entry 3012 net/bluetooth/hci_core.c entry = hci_bdaddr_list_lookup_with_irk(list, bdaddr, type); entry 3013 net/bluetooth/hci_core.c if (!entry) entry 3016 net/bluetooth/hci_core.c list_del(&entry->list); entry 3017 net/bluetooth/hci_core.c kfree(entry); entry 490 net/bluetooth/mgmt.c rp = kmalloc(struct_size(rp, entry, count), GFP_ATOMIC); entry 511 net/bluetooth/mgmt.c rp->entry[count].type = 0x01; entry 513 net/bluetooth/mgmt.c rp->entry[count].type = 0x00; entry 515 net/bluetooth/mgmt.c rp->entry[count].type = 0x02; entry 520 net/bluetooth/mgmt.c rp->entry[count].bus = d->bus; entry 521 net/bluetooth/mgmt.c rp->entry[count++].index = cpu_to_le16(d->id); entry 539 net/bluetooth/mgmt.c struct_size(rp, entry, count)); entry 219 net/bluetooth/rfcomm/tty.c struct rfcomm_dev *dev, *entry; entry 232 net/bluetooth/rfcomm/tty.c list_for_each_entry(entry, &rfcomm_dev_list, list) { entry 233 net/bluetooth/rfcomm/tty.c if (entry->id != dev->id) entry 237 net/bluetooth/rfcomm/tty.c head = &entry->list; entry 242 net/bluetooth/rfcomm/tty.c list_for_each_entry(entry, &rfcomm_dev_list, list) { entry 243 net/bluetooth/rfcomm/tty.c if (entry->id == dev->id) { entry 248 net/bluetooth/rfcomm/tty.c if (entry->id > dev->id - 1) entry 251 net/bluetooth/rfcomm/tty.c head = &entry->list; entry 67 net/bridge/br_mdb.c static void __mdb_entry_to_br_ip(struct br_mdb_entry *entry, struct br_ip *ip) entry 70 net/bridge/br_mdb.c ip->vid = entry->vid; entry 71 net/bridge/br_mdb.c ip->proto = entry->addr.proto; entry 73 net/bridge/br_mdb.c ip->u.ip4 = entry->addr.u.ip4; entry 76 net/bridge/br_mdb.c ip->u.ip6 = entry->addr.u.ip6; entry 266 net/bridge/br_mdb.c struct br_mdb_entry *entry, u32 pid, entry 288 net/bridge/br_mdb.c if (nla_put(skb, MDBA_MDB_ENTRY_INFO, sizeof(*entry), entry)) entry 344 net/bridge/br_mdb.c struct br_mdb_entry *entry, int type) entry 351 net/bridge/br_mdb.c .vid = entry->vid, entry 354 net/bridge/br_mdb.c if (entry->addr.proto == htons(ETH_P_IP)) entry 355 net/bridge/br_mdb.c ip_eth_mc_map(entry->addr.u.ip4, mdb.addr); entry 358 net/bridge/br_mdb.c ipv6_eth_mc_map(&entry->addr.u.ip6, mdb.addr); entry 373 net/bridge/br_mdb.c struct br_mdb_entry *entry, int type) entry 379 net/bridge/br_mdb.c br_mdb_switchdev_host_port(dev, lower_dev, entry, type); entry 383 net/bridge/br_mdb.c struct br_mdb_entry *entry, int type) entry 391 net/bridge/br_mdb.c .vid = entry->vid, entry 398 net/bridge/br_mdb.c port_dev = __dev_get_by_index(net, entry->ifindex); entry 399 net/bridge/br_mdb.c if (entry->addr.proto == htons(ETH_P_IP)) entry 400 net/bridge/br_mdb.c ip_eth_mc_map(entry->addr.u.ip4, mdb.addr); entry 403 net/bridge/br_mdb.c ipv6_eth_mc_map(&entry->addr.u.ip6, mdb.addr); entry 411 net/bridge/br_mdb.c __mdb_entry_to_br_ip(entry, &complete_info->ip); entry 422 net/bridge/br_mdb.c br_mdb_switchdev_host(dev, entry, type); entry 428 net/bridge/br_mdb.c err = nlmsg_populate_mdb_fill(skb, dev, entry, 0, 0, type, NTF_SELF); entry 443 net/bridge/br_mdb.c struct br_mdb_entry entry; entry 445 net/bridge/br_mdb.c memset(&entry, 0, sizeof(entry)); entry 447 net/bridge/br_mdb.c entry.ifindex = port->dev->ifindex; entry 449 net/bridge/br_mdb.c entry.ifindex = dev->ifindex; entry 450 net/bridge/br_mdb.c entry.addr.proto = group->proto; entry 451 net/bridge/br_mdb.c entry.addr.u.ip4 = group->u.ip4; entry 453 net/bridge/br_mdb.c entry.addr.u.ip6 = group->u.ip6; entry 455 net/bridge/br_mdb.c entry.vid = group->vid; entry 456 net/bridge/br_mdb.c __mdb_entry_fill_flags(&entry, flags); entry 457 net/bridge/br_mdb.c __br_mdb_notify(dev, port, &entry, type); entry 527 net/bridge/br_mdb.c static bool is_valid_mdb_entry(struct br_mdb_entry *entry) entry 529 net/bridge/br_mdb.c if (entry->ifindex == 0) entry 532 net/bridge/br_mdb.c if (entry->addr.proto == htons(ETH_P_IP)) { entry 533 net/bridge/br_mdb.c if (!ipv4_is_multicast(entry->addr.u.ip4)) entry 535 net/bridge/br_mdb.c if (ipv4_is_local_multicast(entry->addr.u.ip4)) entry 538 net/bridge/br_mdb.c } else if (entry->addr.proto == htons(ETH_P_IPV6)) { entry 539 net/bridge/br_mdb.c if (ipv6_addr_is_ll_all_nodes(&entry->addr.u.ip6)) entry 544 net/bridge/br_mdb.c if (entry->state != MDB_PERMANENT && entry->state != MDB_TEMPORARY) entry 546 net/bridge/br_mdb.c if (entry->vid >= VLAN_VID_MASK) entry 556 net/bridge/br_mdb.c struct br_mdb_entry *entry; entry 592 net/bridge/br_mdb.c entry = nla_data(tb[MDBA_SET_ENTRY]); entry 593 net/bridge/br_mdb.c if (!is_valid_mdb_entry(entry)) { entry 598 net/bridge/br_mdb.c *pentry = entry; entry 652 net/bridge/br_mdb.c struct br_mdb_entry *entry) entry 662 net/bridge/br_mdb.c if (entry->ifindex != br->dev->ifindex) { entry 663 net/bridge/br_mdb.c dev = __dev_get_by_index(net, entry->ifindex); entry 672 net/bridge/br_mdb.c __mdb_entry_to_br_ip(entry, &ip); entry 675 net/bridge/br_mdb.c ret = br_mdb_add_group(br, p, &ip, entry->state); entry 687 net/bridge/br_mdb.c struct br_mdb_entry *entry; entry 692 net/bridge/br_mdb.c err = br_mdb_parse(skb, nlh, &dev, &entry); entry 698 net/bridge/br_mdb.c if (entry->ifindex != br->dev->ifindex) { entry 699 net/bridge/br_mdb.c pdev = __dev_get_by_index(net, entry->ifindex); entry 714 net/bridge/br_mdb.c if (br_vlan_enabled(br->dev) && vg && entry->vid == 0) { entry 716 net/bridge/br_mdb.c entry->vid = v->vid; entry 717 net/bridge/br_mdb.c err = __br_mdb_add(net, br, entry); entry 720 net/bridge/br_mdb.c __br_mdb_notify(dev, p, entry, RTM_NEWMDB); entry 723 net/bridge/br_mdb.c err = __br_mdb_add(net, br, entry); entry 725 net/bridge/br_mdb.c __br_mdb_notify(dev, p, entry, RTM_NEWMDB); entry 731 net/bridge/br_mdb.c static int __br_mdb_del(struct net_bridge *br, struct br_mdb_entry *entry) entry 742 net/bridge/br_mdb.c __mdb_entry_to_br_ip(entry, &ip); entry 750 net/bridge/br_mdb.c if (entry->ifindex == mp->br->dev->ifindex && mp->host_joined) { entry 761 net/bridge/br_mdb.c if (!p->port || p->port->dev->ifindex != entry->ifindex) entry 767 net/bridge/br_mdb.c __mdb_entry_fill_flags(entry, p->flags); entry 792 net/bridge/br_mdb.c struct br_mdb_entry *entry; entry 797 net/bridge/br_mdb.c err = br_mdb_parse(skb, nlh, &dev, &entry); entry 803 net/bridge/br_mdb.c if (entry->ifindex != br->dev->ifindex) { entry 804 net/bridge/br_mdb.c pdev = __dev_get_by_index(net, entry->ifindex); entry 819 net/bridge/br_mdb.c if (br_vlan_enabled(br->dev) && vg && entry->vid == 0) { entry 821 net/bridge/br_mdb.c entry->vid = v->vid; entry 822 net/bridge/br_mdb.c err = __br_mdb_del(br, entry); entry 824 net/bridge/br_mdb.c __br_mdb_notify(dev, p, entry, RTM_DELMDB); entry 827 net/bridge/br_mdb.c err = __br_mdb_del(br, entry); entry 829 net/bridge/br_mdb.c __br_mdb_notify(dev, p, entry, RTM_DELMDB); entry 2192 net/bridge/br_multicast.c struct br_ip_list *entry; entry 2210 net/bridge/br_multicast.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 2211 net/bridge/br_multicast.c if (!entry) entry 2214 net/bridge/br_multicast.c entry->addr = group->addr; entry 2215 net/bridge/br_multicast.c list_add(&entry->list, br_ip_list); entry 100 net/bridge/netfilter/ebtables.c ebt_dev_check(const char *entry, const struct net_device *device) entry 105 net/bridge/netfilter/ebtables.c if (*entry == '\0') entry 111 net/bridge/netfilter/ebtables.c while (entry[i] != '\0' && entry[i] != 1 && entry[i] == devname[i]) entry 113 net/bridge/netfilter/ebtables.c return devname[i] != entry[i] && entry[i] != 1; entry 167 net/bridge/netfilter/ebtables.c struct ebt_entry *ebt_next_entry(const struct ebt_entry *entry) entry 169 net/bridge/netfilter/ebtables.c return (void *)entry + entry->next_offset; entry 2042 net/bridge/netfilter/ebtables.c static int size_entry_mwt(const struct ebt_entry *entry, const unsigned char *base, entry 2056 net/bridge/netfilter/ebtables.c if (!entry->bitmask) { entry 2058 net/bridge/netfilter/ebtables.c return ebt_buf_add(state, entry, sizeof(struct ebt_entries)); entry 2060 net/bridge/netfilter/ebtables.c if (*total < sizeof(*entry) || entry->next_offset < sizeof(*entry)) entry 2065 net/bridge/netfilter/ebtables.c ret = ebt_buf_add(state, entry, entry 2071 net/bridge/netfilter/ebtables.c memcpy(&offsets[1], &entry->watchers_offset, entry 2082 net/bridge/netfilter/ebtables.c buf_start = (char *) entry; entry 2133 net/bridge/netfilter/ebtables.c if (next_expected_off != entry->next_offset) entry 2136 net/bridge/netfilter/ebtables.c if (*total < entry->next_offset) entry 2138 net/bridge/netfilter/ebtables.c *total -= entry->next_offset; entry 2024 net/core/devlink.c struct devlink_dpipe_entry *entry) entry 2033 net/core/devlink.c if (nla_put_u64_64bit(skb, DEVLINK_ATTR_DPIPE_ENTRY_INDEX, entry->index, entry 2036 net/core/devlink.c if (entry->counter_valid) entry 2038 net/core/devlink.c entry->counter, DEVLINK_ATTR_PAD)) entry 2046 net/core/devlink.c err = devlink_dpipe_match_values_put(skb, entry->match_values, entry 2047 net/core/devlink.c entry->match_values_count); entry 2059 net/core/devlink.c err = devlink_dpipe_action_values_put(skb, entry->action_values, entry 2060 net/core/devlink.c entry->action_values_count); entry 2125 net/core/devlink.c struct devlink_dpipe_entry *entry) entry 2127 net/core/devlink.c return devlink_dpipe_entry_put(dump_ctx->skb, entry); entry 2139 net/core/devlink.c void devlink_dpipe_entry_clear(struct devlink_dpipe_entry *entry) entry 2145 net/core/devlink.c value = entry->action_values; entry 2146 net/core/devlink.c value_count = entry->action_values_count; entry 2152 net/core/devlink.c value = entry->match_values; entry 2153 net/core/devlink.c value_count = entry->match_values_count; entry 411 net/core/flow_dissector.c u32 entry, label; entry 423 net/core/flow_dissector.c entry = ntohl(hdr[0].entry); entry 424 net/core/flow_dissector.c label = (entry & MPLS_LS_LABEL_MASK) >> MPLS_LS_LABEL_SHIFT; entry 433 net/core/flow_dissector.c key_mpls->mpls_ttl = (entry & MPLS_LS_TTL_MASK) entry 435 net/core/flow_dissector.c key_mpls->mpls_tc = (entry & MPLS_LS_TC_MASK) entry 437 net/core/flow_dissector.c key_mpls->mpls_bos = (entry & MPLS_LS_S_MASK) entry 445 net/core/flow_dissector.c key_keyid->keyid = hdr[1].entry & htonl(MPLS_LS_LABEL_MASK); entry 400 net/core/flow_offload.c struct flow_indr_block_entry *entry; entry 403 net/core/flow_offload.c list_for_each_entry(entry, &block_cb_list, list) { entry 404 net/core/flow_offload.c entry->cb(dev, cb, cb_priv, command); entry 501 net/core/flow_offload.c void flow_indr_add_block_cb(struct flow_indr_block_entry *entry) entry 504 net/core/flow_offload.c list_add_tail(&entry->list, &block_cb_list); entry 509 net/core/flow_offload.c void flow_indr_del_block_cb(struct flow_indr_block_entry *entry) entry 512 net/core/flow_offload.c list_del(&entry->list); entry 261 net/core/pktgen.c struct proc_dir_entry *entry; /* proc file */ entry 1925 net/core/pktgen.c proc_remove(pkt_dev->entry); entry 1927 net/core/pktgen.c pkt_dev->entry = proc_create_data(dev->name, 0600, entry 1931 net/core/pktgen.c if (!pkt_dev->entry) entry 3641 net/core/pktgen.c pkt_dev->entry = proc_create_data(ifname, 0600, t->net->proc_dir, entry 3643 net/core/pktgen.c if (!pkt_dev->entry) { entry 3768 net/core/pktgen.c proc_remove(pkt_dev->entry); entry 1807 net/dcb/dcbnl.c struct dcb_app_type *entry; entry 1809 net/dcb/dcbnl.c entry = kmalloc(sizeof(*entry), GFP_ATOMIC); entry 1810 net/dcb/dcbnl.c if (!entry) entry 1813 net/dcb/dcbnl.c memcpy(&entry->app, app, sizeof(*app)); entry 1814 net/dcb/dcbnl.c entry->ifindex = ifindex; entry 1815 net/dcb/dcbnl.c list_add(&entry->list, &dcb_app_list); entry 51 net/dccp/ccids/lib/packet_history.c struct tfrc_tx_hist_entry *entry = kmem_cache_alloc(tfrc_tx_hist_slab, gfp_any()); entry 53 net/dccp/ccids/lib/packet_history.c if (entry == NULL) entry 55 net/dccp/ccids/lib/packet_history.c entry->seqno = seqno; entry 56 net/dccp/ccids/lib/packet_history.c entry->stamp = ktime_get_real(); entry 57 net/dccp/ccids/lib/packet_history.c entry->next = *headp; entry 58 net/dccp/ccids/lib/packet_history.c *headp = entry; entry 97 net/dccp/ccids/lib/packet_history.c static inline void tfrc_rx_hist_entry_from_skb(struct tfrc_rx_hist_entry *entry, entry 103 net/dccp/ccids/lib/packet_history.c entry->tfrchrx_seqno = DCCP_SKB_CB(skb)->dccpd_seq; entry 104 net/dccp/ccids/lib/packet_history.c entry->tfrchrx_ccval = dh->dccph_ccval; entry 105 net/dccp/ccids/lib/packet_history.c entry->tfrchrx_type = dh->dccph_type; entry 106 net/dccp/ccids/lib/packet_history.c entry->tfrchrx_ndp = ndp; entry 107 net/dccp/ccids/lib/packet_history.c entry->tfrchrx_tstamp = ktime_get_real(); entry 114 net/dccp/ccids/lib/packet_history.c struct tfrc_rx_hist_entry *entry = tfrc_rx_hist_last_rcv(h); entry 116 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(entry, skb, ndp); entry 277 net/dccp/feat.c static void dccp_feat_print_entry(struct dccp_feat_entry const *entry) entry 279 net/dccp/feat.c dccp_debug(" * %s %s = ", entry->is_local ? "local" : "remote", entry 280 net/dccp/feat.c dccp_feat_fname(entry->feat_num)); entry 281 net/dccp/feat.c dccp_feat_printval(entry->feat_num, &entry->val); entry 282 net/dccp/feat.c dccp_pr_debug_cat(", state=%s %s\n", dccp_feat_sname[entry->state], entry 283 net/dccp/feat.c entry->needs_confirm ? "(Confirm pending)" : ""); entry 409 net/dccp/feat.c static void dccp_feat_entry_destructor(struct dccp_feat_entry *entry) entry 411 net/dccp/feat.c if (entry != NULL) { entry 412 net/dccp/feat.c dccp_feat_val_destructor(entry->feat_num, &entry->val); entry 413 net/dccp/feat.c kfree(entry); entry 429 net/dccp/feat.c struct dccp_feat_entry *entry; entry 431 net/dccp/feat.c list_for_each_entry(entry, fn_list, node) { entry 432 net/dccp/feat.c if (entry->feat_num == feat_num && entry->is_local == is_local) entry 433 net/dccp/feat.c return entry; entry 434 net/dccp/feat.c else if (entry->feat_num > feat_num) entry 451 net/dccp/feat.c struct dccp_feat_entry *entry; entry 453 net/dccp/feat.c list_for_each_entry(entry, head, node) entry 454 net/dccp/feat.c if (entry->feat_num == feat && entry->is_local == local) { entry 455 net/dccp/feat.c dccp_feat_val_destructor(entry->feat_num, &entry->val); entry 456 net/dccp/feat.c return entry; entry 457 net/dccp/feat.c } else if (entry->feat_num > feat) { entry 458 net/dccp/feat.c head = &entry->node; entry 462 net/dccp/feat.c entry = kmalloc(sizeof(*entry), gfp_any()); entry 463 net/dccp/feat.c if (entry != NULL) { entry 464 net/dccp/feat.c entry->feat_num = feat; entry 465 net/dccp/feat.c entry->is_local = local; entry 466 net/dccp/feat.c list_add_tail(&entry->node, head); entry 468 net/dccp/feat.c return entry; entry 533 net/dccp/feat.c static inline void dccp_feat_list_pop(struct dccp_feat_entry *entry) entry 535 net/dccp/feat.c list_del(&entry->node); entry 536 net/dccp/feat.c dccp_feat_entry_destructor(entry); entry 541 net/dccp/feat.c struct dccp_feat_entry *entry, *next; entry 543 net/dccp/feat.c list_for_each_entry_safe(entry, next, fn_list, node) entry 544 net/dccp/feat.c dccp_feat_entry_destructor(entry); entry 552 net/dccp/feat.c struct dccp_feat_entry *entry, *new; entry 555 net/dccp/feat.c list_for_each_entry(entry, from, node) { entry 556 net/dccp/feat.c new = dccp_feat_clone_entry(entry); entry 775 net/dccp/feat.c struct dccp_feat_entry *entry; entry 777 net/dccp/feat.c entry = dccp_feat_list_lookup(&dp->dccps_featneg, feat, 1); entry 778 net/dccp/feat.c if (entry != NULL) entry 779 net/dccp/feat.c return entry->val.nn; entry 805 net/dccp/feat.c struct dccp_feat_entry *entry; entry 817 net/dccp/feat.c entry = dccp_feat_list_lookup(fn, feat, 1); entry 818 net/dccp/feat.c if (entry != NULL) { entry 820 net/dccp/feat.c (unsigned long long)entry->val.nn, entry 822 net/dccp/feat.c dccp_feat_list_pop(entry); entry 971 net/dccp/feat.c struct dccp_feat_entry *entry; entry 983 net/dccp/feat.c list_for_each_entry(entry, fn, node) entry 984 net/dccp/feat.c if (entry->feat_num == DCCPF_CCID && entry->val.sp.len == 1) entry 985 net/dccp/feat.c ccids[entry->is_local] = entry->val.sp.vec[0]; entry 1001 net/dccp/feat.c struct dccp_feat_entry *entry; entry 1005 net/dccp/feat.c entry = dccp_feat_list_lookup(fn, DCCPF_CCID, is_local); entry 1007 net/dccp/feat.c if (entry != NULL && !entry->empty_confirm) entry 1008 net/dccp/feat.c ccid = entry->val.sp.vec[0]; entry 1102 net/dccp/feat.c struct dccp_feat_entry *entry; entry 1129 net/dccp/feat.c entry = dccp_feat_list_lookup(fn, feat, local); entry 1130 net/dccp/feat.c if (entry == NULL) { entry 1161 net/dccp/feat.c } else if (entry->state == FEAT_UNSTABLE) { /* 6.6.2 */ entry 1165 net/dccp/feat.c if (dccp_feat_reconcile(&entry->val, val, len, server, true)) { entry 1166 net/dccp/feat.c entry->empty_confirm = false; entry 1169 net/dccp/feat.c } else if (entry->state == FEAT_INITIALISING) { entry 1180 net/dccp/feat.c if (!dccp_feat_reconcile(&entry->val, &defval, 1, server, true)) entry 1182 net/dccp/feat.c entry->empty_confirm = true; entry 1184 net/dccp/feat.c entry->needs_confirm = true; entry 1185 net/dccp/feat.c entry->needs_mandatory = false; entry 1186 net/dccp/feat.c entry->state = FEAT_STABLE; entry 1213 net/dccp/feat.c struct dccp_feat_entry *entry = dccp_feat_list_lookup(fn, feat, local); entry 1217 net/dccp/feat.c if (entry == NULL) { /* nothing queued: ignore or handle error */ entry 1226 net/dccp/feat.c if (entry->state != FEAT_CHANGING) /* 6.6.2 */ entry 1239 net/dccp/feat.c dccp_feat_list_pop(entry); entry 1244 net/dccp/feat.c if (len > sizeof(entry->val.nn)) entry 1247 net/dccp/feat.c if (entry->val.nn == dccp_decode_value_var(val, len)) entry 1271 net/dccp/feat.c if (dccp_feat_reconcile(&entry->val, plist, plen, server, 0) != *val) { entry 1275 net/dccp/feat.c entry->val.sp.vec[0] = *val; entry 1278 net/dccp/feat.c entry->state = FEAT_STABLE; entry 1310 net/dccp/feat.c struct dccp_feat_entry *entry; entry 1347 net/dccp/feat.c entry = dccp_feat_list_lookup(fn, feat, local); entry 1348 net/dccp/feat.c if (entry == NULL || entry->state != FEAT_CHANGING) entry 1358 net/dccp/feat.c if (fval.nn != entry->val.nn) entry 1365 net/dccp/feat.c dccp_feat_list_pop(entry); entry 79 net/dccp/feat.h static inline u8 dccp_feat_genopt(struct dccp_feat_entry *entry) entry 81 net/dccp/feat.h if (entry->needs_confirm) entry 82 net/dccp/feat.h return entry->is_local ? DCCPO_CONFIRM_L : DCCPO_CONFIRM_R; entry 83 net/dccp/feat.h return entry->is_local ? DCCPO_CHANGE_L : DCCPO_CHANGE_R; entry 132 net/ipv4/cipso_ipv4.c static void cipso_v4_cache_entry_free(struct cipso_v4_map_cache_entry *entry) entry 134 net/ipv4/cipso_ipv4.c if (entry->lsm_data) entry 135 net/ipv4/cipso_ipv4.c netlbl_secattr_cache_free(entry->lsm_data); entry 136 net/ipv4/cipso_ipv4.c kfree(entry->key); entry 137 net/ipv4/cipso_ipv4.c kfree(entry); entry 196 net/ipv4/cipso_ipv4.c struct cipso_v4_map_cache_entry *entry, *tmp_entry; entry 201 net/ipv4/cipso_ipv4.c list_for_each_entry_safe(entry, entry 204 net/ipv4/cipso_ipv4.c list_del(&entry->list); entry 205 net/ipv4/cipso_ipv4.c cipso_v4_cache_entry_free(entry); entry 239 net/ipv4/cipso_ipv4.c struct cipso_v4_map_cache_entry *entry; entry 249 net/ipv4/cipso_ipv4.c list_for_each_entry(entry, &cipso_v4_cache[bkt].list, list) { entry 250 net/ipv4/cipso_ipv4.c if (entry->hash == hash && entry 251 net/ipv4/cipso_ipv4.c entry->key_len == key_len && entry 252 net/ipv4/cipso_ipv4.c memcmp(entry->key, key, key_len) == 0) { entry 253 net/ipv4/cipso_ipv4.c entry->activity += 1; entry 254 net/ipv4/cipso_ipv4.c refcount_inc(&entry->lsm_data->refcount); entry 255 net/ipv4/cipso_ipv4.c secattr->cache = entry->lsm_data; entry 265 net/ipv4/cipso_ipv4.c if (entry->activity > prev_entry->activity && entry 266 net/ipv4/cipso_ipv4.c entry->activity - prev_entry->activity > entry 268 net/ipv4/cipso_ipv4.c __list_del(entry->list.prev, entry->list.next); entry 269 net/ipv4/cipso_ipv4.c __list_add(&entry->list, entry 277 net/ipv4/cipso_ipv4.c prev_entry = entry; entry 302 net/ipv4/cipso_ipv4.c struct cipso_v4_map_cache_entry *entry = NULL; entry 311 net/ipv4/cipso_ipv4.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 312 net/ipv4/cipso_ipv4.c if (!entry) entry 314 net/ipv4/cipso_ipv4.c entry->key = kmemdup(cipso_ptr, cipso_ptr_len, GFP_ATOMIC); entry 315 net/ipv4/cipso_ipv4.c if (!entry->key) { entry 319 net/ipv4/cipso_ipv4.c entry->key_len = cipso_ptr_len; entry 320 net/ipv4/cipso_ipv4.c entry->hash = cipso_v4_map_cache_hash(cipso_ptr, cipso_ptr_len); entry 322 net/ipv4/cipso_ipv4.c entry->lsm_data = secattr->cache; entry 324 net/ipv4/cipso_ipv4.c bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETS - 1); entry 327 net/ipv4/cipso_ipv4.c list_add(&entry->list, &cipso_v4_cache[bkt].list); entry 333 net/ipv4/cipso_ipv4.c list_add(&entry->list, &cipso_v4_cache[bkt].list); entry 341 net/ipv4/cipso_ipv4.c if (entry) entry 342 net/ipv4/cipso_ipv4.c cipso_v4_cache_entry_free(entry); entry 490 net/ipv4/cipso_ipv4.c static void cipso_v4_doi_free_rcu(struct rcu_head *entry) entry 494 net/ipv4/cipso_ipv4.c doi_def = container_of(entry, struct cipso_v4_doi, rcu); entry 551 net/ipv4/inet_diag.c const struct inet_diag_entry *entry) entry 567 net/ipv4/inet_diag.c yes = entry->sport == op[1].no; entry 570 net/ipv4/inet_diag.c yes = entry->sport >= op[1].no; entry 573 net/ipv4/inet_diag.c yes = entry->sport <= op[1].no; entry 576 net/ipv4/inet_diag.c yes = entry->dport == op[1].no; entry 579 net/ipv4/inet_diag.c yes = entry->dport >= op[1].no; entry 582 net/ipv4/inet_diag.c yes = entry->dport <= op[1].no; entry 585 net/ipv4/inet_diag.c yes = !(entry->userlocks & SOCK_BINDPORT_LOCK); entry 595 net/ipv4/inet_diag.c entry->sport : entry->dport)) { entry 601 net/ipv4/inet_diag.c addr = entry->saddr; entry 603 net/ipv4/inet_diag.c addr = entry->daddr; entry 606 net/ipv4/inet_diag.c cond->family != entry->family) { entry 607 net/ipv4/inet_diag.c if (entry->family == AF_INET6 && entry 632 net/ipv4/inet_diag.c if (ifindex != entry->ifindex) entry 640 net/ipv4/inet_diag.c if ((entry->mark & cond->mask) != cond->mark) entry 659 net/ipv4/inet_diag.c static void entry_fill_addrs(struct inet_diag_entry *entry, entry 664 net/ipv4/inet_diag.c entry->saddr = sk->sk_v6_rcv_saddr.s6_addr32; entry 665 net/ipv4/inet_diag.c entry->daddr = sk->sk_v6_daddr.s6_addr32; entry 669 net/ipv4/inet_diag.c entry->saddr = &sk->sk_rcv_saddr; entry 670 net/ipv4/inet_diag.c entry->daddr = &sk->sk_daddr; entry 677 net/ipv4/inet_diag.c struct inet_diag_entry entry; entry 682 net/ipv4/inet_diag.c entry.family = sk->sk_family; entry 683 net/ipv4/inet_diag.c entry_fill_addrs(&entry, sk); entry 684 net/ipv4/inet_diag.c entry.sport = inet->inet_num; entry 685 net/ipv4/inet_diag.c entry.dport = ntohs(inet->inet_dport); entry 686 net/ipv4/inet_diag.c entry.ifindex = sk->sk_bound_dev_if; entry 687 net/ipv4/inet_diag.c entry.userlocks = sk_fullsock(sk) ? sk->sk_userlocks : 0; entry 689 net/ipv4/inet_diag.c entry.mark = sk->sk_mark; entry 691 net/ipv4/inet_diag.c entry.mark = inet_rsk(inet_reqsk(sk))->ir_mark; entry 693 net/ipv4/inet_diag.c entry.mark = 0; entry 695 net/ipv4/inet_diag.c return inet_diag_bc_run(bc, &entry); entry 177 net/ipv4/netfilter/arp_tables.c struct arpt_entry *arpt_next_entry(const struct arpt_entry *entry) entry 179 net/ipv4/netfilter/arp_tables.c return (void *)entry + entry->next_offset; entry 218 net/ipv4/netfilter/ip_tables.c struct ipt_entry *ipt_next_entry(const struct ipt_entry *entry) entry 220 net/ipv4/netfilter/ip_tables.c return (void *)entry + entry->next_offset; entry 151 net/ipv4/netfilter/ipt_CLUSTERIP.c clusterip_config_find_get(struct net *net, __be32 clusterip, int entry) entry 165 net/ipv4/netfilter/ipt_CLUSTERIP.c else if (entry) { entry 1102 net/ipv4/nexthop.c struct nexthop_grp *entry = nla_data(grps_attr); entry 1103 net/ipv4/nexthop.c u16 num_nh = nla_len(grps_attr) / sizeof(*entry); entry 1133 net/ipv4/nexthop.c nhe = nexthop_find_by_id(net, entry[i].id); entry 1142 net/ipv4/nexthop.c nhg->nh_entries[i].weight = entry[i].weight + 1; entry 411 net/ipv4/proc.c if (snmp4_tcp_list[i].entry == TCP_MIB_MAXCONN) entry 472 net/ipv4/proc.c snmp4_net_list[i].entry)); entry 482 net/ipv4/proc.c snmp4_ipextstats_list[i].entry, entry 98 net/ipv6/calipso.c static void calipso_cache_entry_free(struct calipso_map_cache_entry *entry) entry 100 net/ipv6/calipso.c if (entry->lsm_data) entry 101 net/ipv6/calipso.c netlbl_secattr_cache_free(entry->lsm_data); entry 102 net/ipv6/calipso.c kfree(entry->key); entry 103 net/ipv6/calipso.c kfree(entry); entry 158 net/ipv6/calipso.c struct calipso_map_cache_entry *entry, *tmp_entry; entry 163 net/ipv6/calipso.c list_for_each_entry_safe(entry, entry 166 net/ipv6/calipso.c list_del(&entry->list); entry 167 net/ipv6/calipso.c calipso_cache_entry_free(entry); entry 201 net/ipv6/calipso.c struct calipso_map_cache_entry *entry; entry 211 net/ipv6/calipso.c list_for_each_entry(entry, &calipso_cache[bkt].list, list) { entry 212 net/ipv6/calipso.c if (entry->hash == hash && entry 213 net/ipv6/calipso.c entry->key_len == key_len && entry 214 net/ipv6/calipso.c memcmp(entry->key, key, key_len) == 0) { entry 215 net/ipv6/calipso.c entry->activity += 1; entry 216 net/ipv6/calipso.c refcount_inc(&entry->lsm_data->refcount); entry 217 net/ipv6/calipso.c secattr->cache = entry->lsm_data; entry 227 net/ipv6/calipso.c if (entry->activity > prev_entry->activity && entry 228 net/ipv6/calipso.c entry->activity - prev_entry->activity > entry 230 net/ipv6/calipso.c __list_del(entry->list.prev, entry->list.next); entry 231 net/ipv6/calipso.c __list_add(&entry->list, entry 239 net/ipv6/calipso.c prev_entry = entry; entry 266 net/ipv6/calipso.c struct calipso_map_cache_entry *entry = NULL; entry 275 net/ipv6/calipso.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 276 net/ipv6/calipso.c if (!entry) entry 278 net/ipv6/calipso.c entry->key = kmemdup(calipso_ptr + 2, calipso_ptr_len, GFP_ATOMIC); entry 279 net/ipv6/calipso.c if (!entry->key) { entry 283 net/ipv6/calipso.c entry->key_len = calipso_ptr_len; entry 284 net/ipv6/calipso.c entry->hash = calipso_map_cache_hash(calipso_ptr, calipso_ptr_len); entry 286 net/ipv6/calipso.c entry->lsm_data = secattr->cache; entry 288 net/ipv6/calipso.c bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1); entry 291 net/ipv6/calipso.c list_add(&entry->list, &calipso_cache[bkt].list); entry 297 net/ipv6/calipso.c list_add(&entry->list, &calipso_cache[bkt].list); entry 305 net/ipv6/calipso.c if (entry) entry 306 net/ipv6/calipso.c calipso_cache_entry_free(entry); entry 415 net/ipv6/calipso.c static void calipso_doi_free_rcu(struct rcu_head *entry) entry 419 net/ipv6/calipso.c doi_def = container_of(entry, struct calipso_doi, rcu); entry 78 net/ipv6/netfilter.c const struct nf_queue_entry *entry) entry 80 net/ipv6/netfilter.c struct ip6_rt_info *rt_info = nf_queue_entry_reroute(entry); entry 82 net/ipv6/netfilter.c if (entry->state.hook == NF_INET_LOCAL_OUT) { entry 87 net/ipv6/netfilter.c return ip6_route_me_harder(entry->state.net, skb); entry 243 net/ipv6/netfilter/ip6_tables.c ip6t_next_entry(const struct ip6t_entry *entry) entry 245 net/ipv6/netfilter/ip6_tables.c return (void *)entry + entry->next_offset; entry 196 net/ipv6/proc.c atomic_long_read(smib + itemlist[i].entry)); entry 2027 net/ipv6/route.c struct rt6_info *entry = rt6_ex->rt6i; entry 2033 net/ipv6/route.c if (dst_metric_raw(&entry->dst, RTAX_MTU) && entry 2034 net/ipv6/route.c rt6_mtu_change_route_allowed(idev, entry, mtu)) entry 2035 net/ipv6/route.c dst_metric_set(&entry->dst, RTAX_MTU, mtu); entry 2060 net/ipv6/route.c struct rt6_info *entry = rt6_ex->rt6i; entry 2062 net/ipv6/route.c if ((entry->rt6i_flags & RTF_CACHE_GATEWAY) == entry 2065 net/ipv6/route.c &entry->rt6i_gateway)) { entry 83 net/lapb/lapb_iface.c struct list_head *entry; entry 86 net/lapb/lapb_iface.c list_for_each(entry, &lapb_list) { entry 87 net/lapb/lapb_iface.c lapb = list_entry(entry, struct lapb_cb, node); entry 2090 net/mac80211/rx.c struct ieee80211_fragment_entry *entry; entry 2092 net/mac80211/rx.c entry = &sdata->fragments[sdata->fragment_next++]; entry 2096 net/mac80211/rx.c if (!skb_queue_empty(&entry->skb_list)) entry 2097 net/mac80211/rx.c __skb_queue_purge(&entry->skb_list); entry 2099 net/mac80211/rx.c __skb_queue_tail(&entry->skb_list, *skb); /* no need for locking */ entry 2101 net/mac80211/rx.c entry->first_frag_time = jiffies; entry 2102 net/mac80211/rx.c entry->seq = seq; entry 2103 net/mac80211/rx.c entry->rx_queue = rx_queue; entry 2104 net/mac80211/rx.c entry->last_frag = frag; entry 2105 net/mac80211/rx.c entry->check_sequential_pn = false; entry 2106 net/mac80211/rx.c entry->extra_len = 0; entry 2108 net/mac80211/rx.c return entry; entry 2116 net/mac80211/rx.c struct ieee80211_fragment_entry *entry; entry 2128 net/mac80211/rx.c entry = &sdata->fragments[idx]; entry 2129 net/mac80211/rx.c if (skb_queue_empty(&entry->skb_list) || entry->seq != seq || entry 2130 net/mac80211/rx.c entry->rx_queue != rx_queue || entry 2131 net/mac80211/rx.c entry->last_frag + 1 != frag) entry 2134 net/mac80211/rx.c f_skb = __skb_peek(&entry->skb_list); entry 2146 net/mac80211/rx.c if (time_after(jiffies, entry->first_frag_time + 2 * HZ)) { entry 2147 net/mac80211/rx.c __skb_queue_purge(&entry->skb_list); entry 2150 net/mac80211/rx.c return entry; entry 2163 net/mac80211/rx.c struct ieee80211_fragment_entry *entry; entry 2198 net/mac80211/rx.c entry = ieee80211_reassemble_add(rx->sdata, frag, seq, entry 2211 net/mac80211/rx.c entry->check_sequential_pn = true; entry 2212 net/mac80211/rx.c memcpy(entry->last_pn, entry 2230 net/mac80211/rx.c entry = ieee80211_reassemble_find(rx->sdata, frag, seq, entry 2232 net/mac80211/rx.c if (!entry) { entry 2242 net/mac80211/rx.c if (entry->check_sequential_pn) { entry 2253 net/mac80211/rx.c memcpy(pn, entry->last_pn, IEEE80211_CCMP_PN_LEN); entry 2263 net/mac80211/rx.c memcpy(entry->last_pn, pn, IEEE80211_CCMP_PN_LEN); entry 2267 net/mac80211/rx.c __skb_queue_tail(&entry->skb_list, rx->skb); entry 2268 net/mac80211/rx.c entry->last_frag = frag; entry 2269 net/mac80211/rx.c entry->extra_len += rx->skb->len; entry 2275 net/mac80211/rx.c rx->skb = __skb_dequeue(&entry->skb_list); entry 2276 net/mac80211/rx.c if (skb_tailroom(rx->skb) < entry->extra_len) { entry 2278 net/mac80211/rx.c if (unlikely(pskb_expand_head(rx->skb, 0, entry->extra_len, entry 2281 net/mac80211/rx.c __skb_queue_purge(&entry->skb_list); entry 2285 net/mac80211/rx.c while ((skb = __skb_dequeue(&entry->skb_list))) { entry 353 net/mac802154/llsec.c struct mac802154_llsec_device *entry; entry 364 net/mac802154/llsec.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 365 net/mac802154/llsec.c if (!entry) entry 368 net/mac802154/llsec.c entry->dev = *dev; entry 369 net/mac802154/llsec.c spin_lock_init(&entry->lock); entry 370 net/mac802154/llsec.c INIT_LIST_HEAD(&entry->dev.keys); entry 373 net/mac802154/llsec.c hash_add_rcu(sec->devices_short, &entry->bucket_s, skey); entry 375 net/mac802154/llsec.c INIT_HLIST_NODE(&entry->bucket_s); entry 377 net/mac802154/llsec.c hash_add_rcu(sec->devices_hw, &entry->bucket_hw, hwkey); entry 378 net/mac802154/llsec.c list_add_tail_rcu(&entry->dev.list, &sec->table.devices); entry 490 net/mac802154/llsec.c struct mac802154_llsec_seclevel *entry; entry 495 net/mac802154/llsec.c entry = kmalloc(sizeof(*entry), GFP_KERNEL); entry 496 net/mac802154/llsec.c if (!entry) entry 499 net/mac802154/llsec.c entry->level = *sl; entry 501 net/mac802154/llsec.c list_add_tail_rcu(&entry->level.list, &sec->table.security_levels); entry 189 net/mpls/internal.h unsigned entry = be32_to_cpu(hdr->label_stack_entry); entry 191 net/mpls/internal.h result.label = (entry & MPLS_LS_LABEL_MASK) >> MPLS_LS_LABEL_SHIFT; entry 192 net/mpls/internal.h result.ttl = (entry & MPLS_LS_TTL_MASK) >> MPLS_LS_TTL_SHIFT; entry 193 net/mpls/internal.h result.tc = (entry & MPLS_LS_TC_MASK) >> MPLS_LS_TC_SHIFT; entry 194 net/mpls/internal.h result.bos = (entry & MPLS_LS_S_MASK) >> MPLS_LS_S_SHIFT; entry 2584 net/netfilter/ipvs/ip_vs_ctl.c struct ip_vs_service_entry entry; entry 2595 net/netfilter/ipvs/ip_vs_ctl.c memset(&entry, 0, sizeof(entry)); entry 2596 net/netfilter/ipvs/ip_vs_ctl.c ip_vs_copy_service(&entry, svc); entry 2598 net/netfilter/ipvs/ip_vs_ctl.c &entry, sizeof(entry))) { entry 2614 net/netfilter/ipvs/ip_vs_ctl.c memset(&entry, 0, sizeof(entry)); entry 2615 net/netfilter/ipvs/ip_vs_ctl.c ip_vs_copy_service(&entry, svc); entry 2617 net/netfilter/ipvs/ip_vs_ctl.c &entry, sizeof(entry))) { entry 2647 net/netfilter/ipvs/ip_vs_ctl.c struct ip_vs_dest_entry entry; entry 2650 net/netfilter/ipvs/ip_vs_ctl.c memset(&entry, 0, sizeof(entry)); entry 2661 net/netfilter/ipvs/ip_vs_ctl.c entry.addr = dest->addr.ip; entry 2662 net/netfilter/ipvs/ip_vs_ctl.c entry.port = dest->port; entry 2663 net/netfilter/ipvs/ip_vs_ctl.c entry.conn_flags = atomic_read(&dest->conn_flags); entry 2664 net/netfilter/ipvs/ip_vs_ctl.c entry.weight = atomic_read(&dest->weight); entry 2665 net/netfilter/ipvs/ip_vs_ctl.c entry.u_threshold = dest->u_threshold; entry 2666 net/netfilter/ipvs/ip_vs_ctl.c entry.l_threshold = dest->l_threshold; entry 2667 net/netfilter/ipvs/ip_vs_ctl.c entry.activeconns = atomic_read(&dest->activeconns); entry 2668 net/netfilter/ipvs/ip_vs_ctl.c entry.inactconns = atomic_read(&dest->inactconns); entry 2669 net/netfilter/ipvs/ip_vs_ctl.c entry.persistconns = atomic_read(&dest->persistconns); entry 2671 net/netfilter/ipvs/ip_vs_ctl.c ip_vs_export_stats_user(&entry.stats, &kstats); entry 2673 net/netfilter/ipvs/ip_vs_ctl.c &entry, sizeof(entry))) { entry 2823 net/netfilter/ipvs/ip_vs_ctl.c struct ip_vs_service_entry *entry; entry 2827 net/netfilter/ipvs/ip_vs_ctl.c entry = (struct ip_vs_service_entry *)arg; entry 2828 net/netfilter/ipvs/ip_vs_ctl.c addr.ip = entry->addr; entry 2830 net/netfilter/ipvs/ip_vs_ctl.c if (entry->fwmark) entry 2831 net/netfilter/ipvs/ip_vs_ctl.c svc = __ip_vs_svc_fwm_find(ipvs, AF_INET, entry->fwmark); entry 2834 net/netfilter/ipvs/ip_vs_ctl.c entry->protocol, &addr, entry 2835 net/netfilter/ipvs/ip_vs_ctl.c entry->port); entry 2838 net/netfilter/ipvs/ip_vs_ctl.c ip_vs_copy_service(entry, svc); entry 2839 net/netfilter/ipvs/ip_vs_ctl.c if (copy_to_user(user, entry, sizeof(*entry)) != 0) entry 58 net/netfilter/nf_dup_netdev.c struct flow_action_entry *entry; entry 66 net/netfilter/nf_dup_netdev.c entry = &flow->rule->action.entries[ctx->num_actions++]; entry 67 net/netfilter/nf_dup_netdev.c entry->id = id; entry 68 net/netfilter/nf_dup_netdev.c entry->dev = dev; entry 63 net/netfilter/nf_flow_table_core.c struct flow_offload_entry *entry; entry 70 net/netfilter/nf_flow_table_core.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 71 net/netfilter/nf_flow_table_core.c if (!entry) entry 74 net/netfilter/nf_flow_table_core.c flow = &entry->flow; entry 82 net/netfilter/nf_flow_table_core.c entry->ct = ct; entry 97 net/netfilter/nf_flow_table_core.c kfree(entry); entry 67 net/netfilter/nf_queue.c void nf_queue_entry_release_refs(struct nf_queue_entry *entry) entry 69 net/netfilter/nf_queue.c struct nf_hook_state *state = &entry->state; entry 79 net/netfilter/nf_queue.c nf_queue_entry_release_br_nf_refs(entry->skb); entry 102 net/netfilter/nf_queue.c void nf_queue_entry_get_refs(struct nf_queue_entry *entry) entry 104 net/netfilter/nf_queue.c struct nf_hook_state *state = &entry->state; entry 113 net/netfilter/nf_queue.c nf_queue_entry_get_br_nf_refs(entry->skb); entry 130 net/netfilter/nf_queue.c struct nf_queue_entry *entry) entry 132 net/netfilter/nf_queue.c struct ip_rt_info *rt_info = nf_queue_entry_reroute(entry); entry 134 net/netfilter/nf_queue.c if (entry->state.hook == NF_INET_LOCAL_OUT) { entry 145 net/netfilter/nf_queue.c struct nf_queue_entry *entry) entry 147 net/netfilter/nf_queue.c struct ip6_rt_info *rt_info = nf_queue_entry_reroute(entry); entry 149 net/netfilter/nf_queue.c if (entry->state.hook == NF_INET_LOCAL_OUT) { entry 162 net/netfilter/nf_queue.c struct nf_queue_entry *entry = NULL; entry 186 net/netfilter/nf_queue.c entry = kmalloc(sizeof(*entry) + route_key_size, GFP_ATOMIC); entry 187 net/netfilter/nf_queue.c if (!entry) { entry 197 net/netfilter/nf_queue.c *entry = (struct nf_queue_entry) { entry 201 net/netfilter/nf_queue.c .size = sizeof(*entry) + route_key_size, entry 204 net/netfilter/nf_queue.c nf_queue_entry_get_refs(entry); entry 206 net/netfilter/nf_queue.c switch (entry->state.pf) { entry 208 net/netfilter/nf_queue.c nf_ip_saveroute(skb, entry); entry 211 net/netfilter/nf_queue.c nf_ip6_saveroute(skb, entry); entry 215 net/netfilter/nf_queue.c status = qh->outfn(entry, queuenum); entry 218 net/netfilter/nf_queue.c nf_queue_entry_release_refs(entry); entry 225 net/netfilter/nf_queue.c kfree(entry); entry 292 net/netfilter/nf_queue.c void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict) entry 296 net/netfilter/nf_queue.c struct sk_buff *skb = entry->skb; entry 302 net/netfilter/nf_queue.c net = entry->state.net; entry 303 net/netfilter/nf_queue.c pf = entry->state.pf; entry 305 net/netfilter/nf_queue.c hooks = nf_hook_entries_head(net, pf, entry->state.hook); entry 307 net/netfilter/nf_queue.c nf_queue_entry_release_refs(entry); entry 309 net/netfilter/nf_queue.c i = entry->hook_index; entry 312 net/netfilter/nf_queue.c kfree(entry); entry 320 net/netfilter/nf_queue.c verdict = nf_hook_entry_hookfn(hook_entry, skb, &entry->state); entry 323 net/netfilter/nf_queue.c if (nf_reroute(skb, entry) < 0) entry 330 net/netfilter/nf_queue.c verdict = nf_iterate(skb, &entry->state, hooks, &i); entry 337 net/netfilter/nf_queue.c entry->state.okfn(entry->state.net, entry->state.sk, skb); entry 341 net/netfilter/nf_queue.c err = nf_queue(skb, &entry->state, i, verdict); entry 351 net/netfilter/nf_queue.c kfree(entry); entry 88 net/netfilter/nf_tables_offload.c struct flow_action_entry *entry; entry 91 net/netfilter/nf_tables_offload.c flow_action_for_each(i, entry, &flow->rule->action) { entry 92 net/netfilter/nf_tables_offload.c switch (entry->id) { entry 95 net/netfilter/nf_tables_offload.c dev_put(entry->dev); entry 191 net/netfilter/nfnetlink_queue.c __enqueue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry) entry 193 net/netfilter/nfnetlink_queue.c list_add_tail(&entry->list, &queue->queue_list); entry 198 net/netfilter/nfnetlink_queue.c __dequeue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry) entry 200 net/netfilter/nfnetlink_queue.c list_del(&entry->list); entry 207 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry = NULL, *i; entry 213 net/netfilter/nfnetlink_queue.c entry = i; entry 218 net/netfilter/nfnetlink_queue.c if (entry) entry 219 net/netfilter/nfnetlink_queue.c __dequeue_entry(queue, entry); entry 223 net/netfilter/nfnetlink_queue.c return entry; entry 226 net/netfilter/nfnetlink_queue.c static void nfqnl_reinject(struct nf_queue_entry *entry, unsigned int verdict) entry 237 net/netfilter/nfnetlink_queue.c err = ct_hook->update(entry->state.net, entry->skb); entry 243 net/netfilter/nfnetlink_queue.c nf_reinject(entry, verdict); entry 249 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry, *next; entry 252 net/netfilter/nfnetlink_queue.c list_for_each_entry_safe(entry, next, &queue->queue_list, list) { entry 253 net/netfilter/nfnetlink_queue.c if (!cmpfn || cmpfn(entry, data)) { entry 254 net/netfilter/nfnetlink_queue.c list_del(&entry->list); entry 256 net/netfilter/nfnetlink_queue.c nfqnl_reinject(entry, NF_DROP); entry 321 net/netfilter/nfnetlink_queue.c static u32 nfqnl_get_bridge_size(struct nf_queue_entry *entry) entry 323 net/netfilter/nfnetlink_queue.c struct sk_buff *entskb = entry->skb; entry 326 net/netfilter/nfnetlink_queue.c if (entry->state.pf != PF_BRIDGE || !skb_mac_header_was_set(entskb)) entry 340 net/netfilter/nfnetlink_queue.c static int nfqnl_put_bridge(struct nf_queue_entry *entry, struct sk_buff *skb) entry 342 net/netfilter/nfnetlink_queue.c struct sk_buff *entskb = entry->skb; entry 344 net/netfilter/nfnetlink_queue.c if (entry->state.pf != PF_BRIDGE || !skb_mac_header_was_set(entskb)) entry 376 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry, entry 387 net/netfilter/nfnetlink_queue.c struct sk_buff *entskb = entry->skb; entry 413 net/netfilter/nfnetlink_queue.c size += nfqnl_get_bridge_size(entry); entry 415 net/netfilter/nfnetlink_queue.c if (entry->state.hook <= NF_INET_FORWARD || entry 416 net/netfilter/nfnetlink_queue.c (entry->state.hook == NF_INET_POST_ROUTING && entskb->sk == NULL)) entry 421 net/netfilter/nfnetlink_queue.c outdev = entry->state.out; entry 481 net/netfilter/nfnetlink_queue.c nfmsg->nfgen_family = entry->state.pf; entry 488 net/netfilter/nfnetlink_queue.c pmsg->hook = entry->state.hook; entry 491 net/netfilter/nfnetlink_queue.c indev = entry->state.in; entry 497 net/netfilter/nfnetlink_queue.c if (entry->state.pf == PF_BRIDGE) { entry 531 net/netfilter/nfnetlink_queue.c if (entry->state.pf == PF_BRIDGE) { entry 578 net/netfilter/nfnetlink_queue.c if (nfqnl_put_bridge(entry, skb) < 0) entry 581 net/netfilter/nfnetlink_queue.c if (entry->state.hook <= NF_INET_FORWARD && entskb->tstamp) { entry 638 net/netfilter/nfnetlink_queue.c static bool nf_ct_drop_unconfirmed(const struct nf_queue_entry *entry) entry 642 net/netfilter/nfnetlink_queue.c const struct nf_conn *ct = (void *)skb_nfct(entry->skb); entry 652 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry) entry 659 net/netfilter/nfnetlink_queue.c nskb = nfqnl_build_packet_message(net, queue, entry, &packet_id_ptr); entry 666 net/netfilter/nfnetlink_queue.c if (nf_ct_drop_unconfirmed(entry)) entry 680 net/netfilter/nfnetlink_queue.c entry->id = ++queue->id_sequence; entry 681 net/netfilter/nfnetlink_queue.c *packet_id_ptr = htonl(entry->id); entry 695 net/netfilter/nfnetlink_queue.c __enqueue_entry(queue, entry); entry 705 net/netfilter/nfnetlink_queue.c nfqnl_reinject(entry, NF_ACCEPT); entry 713 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry = kmemdup(e, e->size, GFP_ATOMIC); entry 714 net/netfilter/nfnetlink_queue.c if (entry) entry 715 net/netfilter/nfnetlink_queue.c nf_queue_entry_get_refs(entry); entry 716 net/netfilter/nfnetlink_queue.c return entry; entry 740 net/netfilter/nfnetlink_queue.c static void free_entry(struct nf_queue_entry *entry) entry 742 net/netfilter/nfnetlink_queue.c nf_queue_entry_release_refs(entry); entry 743 net/netfilter/nfnetlink_queue.c kfree(entry); entry 748 net/netfilter/nfnetlink_queue.c struct sk_buff *skb, struct nf_queue_entry *entry) entry 756 net/netfilter/nfnetlink_queue.c struct sk_buff *gso_skb = entry->skb; entry 757 net/netfilter/nfnetlink_queue.c entry->skb = skb; entry 758 net/netfilter/nfnetlink_queue.c ret = __nfqnl_enqueue_packet(net, queue, entry); entry 760 net/netfilter/nfnetlink_queue.c entry->skb = gso_skb; entry 766 net/netfilter/nfnetlink_queue.c entry_seg = nf_queue_entry_dup(entry); entry 777 net/netfilter/nfnetlink_queue.c nfqnl_enqueue_packet(struct nf_queue_entry *entry, unsigned int queuenum) entry 783 net/netfilter/nfnetlink_queue.c struct net *net = entry->state.net; entry 794 net/netfilter/nfnetlink_queue.c skb = entry->skb; entry 796 net/netfilter/nfnetlink_queue.c switch (entry->state.pf) { entry 806 net/netfilter/nfnetlink_queue.c return __nfqnl_enqueue_packet(net, queue, entry); entry 822 net/netfilter/nfnetlink_queue.c segs, entry); entry 832 net/netfilter/nfnetlink_queue.c free_entry(entry); entry 901 net/netfilter/nfnetlink_queue.c dev_cmp(struct nf_queue_entry *entry, unsigned long ifindex) entry 906 net/netfilter/nfnetlink_queue.c physinif = nf_bridge_get_physinif(entry->skb); entry 907 net/netfilter/nfnetlink_queue.c physoutif = nf_bridge_get_physoutif(entry->skb); entry 912 net/netfilter/nfnetlink_queue.c if (entry->state.in) entry 913 net/netfilter/nfnetlink_queue.c if (entry->state.in->ifindex == ifindex) entry 915 net/netfilter/nfnetlink_queue.c if (entry->state.out) entry 916 net/netfilter/nfnetlink_queue.c if (entry->state.out->ifindex == ifindex) entry 1066 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry, *tmp; entry 1088 net/netfilter/nfnetlink_queue.c list_for_each_entry_safe(entry, tmp, &queue->queue_list, list) { entry 1089 net/netfilter/nfnetlink_queue.c if (nfq_id_after(entry->id, maxid)) entry 1091 net/netfilter/nfnetlink_queue.c __dequeue_entry(queue, entry); entry 1092 net/netfilter/nfnetlink_queue.c list_add_tail(&entry->list, &batch_list); entry 1100 net/netfilter/nfnetlink_queue.c list_for_each_entry_safe(entry, tmp, &batch_list, list) { entry 1102 net/netfilter/nfnetlink_queue.c entry->skb->mark = ntohl(nla_get_be32(nfqa[NFQA_MARK])); entry 1104 net/netfilter/nfnetlink_queue.c nfqnl_reinject(entry, verdict); entry 1112 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry, entry 1117 net/netfilter/nfnetlink_queue.c ct = nfnl_ct->get_ct(entry->skb, ctinfo); entry 1126 net/netfilter/nfnetlink_queue.c NETLINK_CB(entry->skb).portid, entry 1131 net/netfilter/nfnetlink_queue.c static int nfqa_parse_bridge(struct nf_queue_entry *entry, entry 1147 net/netfilter/nfnetlink_queue.c __vlan_hwaccel_put_tag(entry->skb, entry 1153 net/netfilter/nfnetlink_queue.c int mac_header_len = entry->skb->network_header - entry 1154 net/netfilter/nfnetlink_queue.c entry->skb->mac_header; entry 1159 net/netfilter/nfnetlink_queue.c memcpy(skb_mac_header(entry->skb), entry 1178 net/netfilter/nfnetlink_queue.c struct nf_queue_entry *entry; entry 1196 net/netfilter/nfnetlink_queue.c entry = find_dequeue_entry(queue, ntohl(vhdr->id)); entry 1197 net/netfilter/nfnetlink_queue.c if (entry == NULL) entry 1205 net/netfilter/nfnetlink_queue.c ct = nfqnl_ct_parse(nfnl_ct, nlh, nfqa, entry, &ctinfo); entry 1208 net/netfilter/nfnetlink_queue.c if (entry->state.pf == PF_BRIDGE) { entry 1209 net/netfilter/nfnetlink_queue.c err = nfqa_parse_bridge(entry, nfqa); entry 1216 net/netfilter/nfnetlink_queue.c int diff = payload_len - entry->skb->len; entry 1219 net/netfilter/nfnetlink_queue.c payload_len, entry, diff) < 0) entry 1223 net/netfilter/nfnetlink_queue.c nfnl_ct->seq_adjust(entry->skb, ct, ctinfo, diff); entry 1227 net/netfilter/nfnetlink_queue.c entry->skb->mark = ntohl(nla_get_be32(nfqa[NFQA_MARK])); entry 1229 net/netfilter/nfnetlink_queue.c nfqnl_reinject(entry, verdict); entry 138 net/netfilter/nft_compat.c union nft_entry *entry, u16 proto, bool inv) entry 144 net/netfilter/nft_compat.c entry->e4.ip.proto = proto; entry 145 net/netfilter/nft_compat.c entry->e4.ip.invflags = inv ? IPT_INV_PROTO : 0; entry 149 net/netfilter/nft_compat.c entry->e6.ipv6.flags |= IP6T_F_PROTO; entry 151 net/netfilter/nft_compat.c entry->e6.ipv6.proto = proto; entry 152 net/netfilter/nft_compat.c entry->e6.ipv6.invflags = inv ? IP6T_INV_PROTO : 0; entry 155 net/netfilter/nft_compat.c entry->ebt.ethproto = (__force __be16)proto; entry 156 net/netfilter/nft_compat.c entry->ebt.invflags = inv ? EBT_IPROTO : 0; entry 161 net/netfilter/nft_compat.c par->entryinfo = entry; entry 382 net/netfilter/nft_compat.c union nft_entry *entry, u16 proto, bool inv) entry 388 net/netfilter/nft_compat.c entry->e4.ip.proto = proto; entry 389 net/netfilter/nft_compat.c entry->e4.ip.invflags = inv ? IPT_INV_PROTO : 0; entry 393 net/netfilter/nft_compat.c entry->e6.ipv6.flags |= IP6T_F_PROTO; entry 395 net/netfilter/nft_compat.c entry->e6.ipv6.proto = proto; entry 396 net/netfilter/nft_compat.c entry->e6.ipv6.invflags = inv ? IP6T_INV_PROTO : 0; entry 399 net/netfilter/nft_compat.c entry->ebt.ethproto = (__force __be16)proto; entry 400 net/netfilter/nft_compat.c entry->ebt.invflags = inv ? EBT_IPROTO : 0; entry 405 net/netfilter/nft_compat.c par->entryinfo = entry; entry 132 net/netfilter/nft_immediate.c struct flow_action_entry *entry; entry 135 net/netfilter/nft_immediate.c entry = &flow->rule->action.entries[ctx->num_actions++]; entry 140 net/netfilter/nft_immediate.c entry->id = FLOW_ACTION_ACCEPT; entry 143 net/netfilter/nft_immediate.c entry->id = FLOW_ACTION_DROP; entry 182 net/netfilter/utils.c static int nf_ip_reroute(struct sk_buff *skb, const struct nf_queue_entry *entry) entry 185 net/netfilter/utils.c const struct ip_rt_info *rt_info = nf_queue_entry_reroute(entry); entry 187 net/netfilter/utils.c if (entry->state.hook == NF_INET_LOCAL_OUT) { entry 194 net/netfilter/utils.c return ip_route_me_harder(entry->state.net, skb, entry 201 net/netfilter/utils.c int nf_reroute(struct sk_buff *skb, struct nf_queue_entry *entry) entry 206 net/netfilter/utils.c switch (entry->state.pf) { entry 208 net/netfilter/utils.c ret = nf_ip_reroute(skb, entry); entry 213 net/netfilter/utils.c ret = v6ops->reroute(skb, entry); entry 32 net/netfilter/xt_IDLETIMER.c struct list_head entry; entry 50 net/netfilter/xt_IDLETIMER.c struct idletimer_tg *entry; entry 52 net/netfilter/xt_IDLETIMER.c list_for_each_entry(entry, &idletimer_tg_list, entry) { entry 53 net/netfilter/xt_IDLETIMER.c if (!strcmp(label, entry->attr.attr.name)) entry 54 net/netfilter/xt_IDLETIMER.c return entry; entry 143 net/netfilter/xt_IDLETIMER.c list_add(&info->timer->entry, &idletimer_tg_list); entry 236 net/netfilter/xt_IDLETIMER.c list_del(&info->timer->entry); entry 148 net/netlabel/netlabel_addrlist.c int netlbl_af4list_add(struct netlbl_af4list *entry, struct list_head *head) entry 152 net/netlabel/netlabel_addrlist.c iter = netlbl_af4list_search(entry->addr, head); entry 154 net/netlabel/netlabel_addrlist.c iter->addr == entry->addr && iter->mask == entry->mask) entry 163 net/netlabel/netlabel_addrlist.c ntohl(entry->mask) > ntohl(iter->mask)) { entry 164 net/netlabel/netlabel_addrlist.c __list_add_rcu(&entry->list, entry 169 net/netlabel/netlabel_addrlist.c list_add_tail_rcu(&entry->list, head); entry 185 net/netlabel/netlabel_addrlist.c int netlbl_af6list_add(struct netlbl_af6list *entry, struct list_head *head) entry 189 net/netlabel/netlabel_addrlist.c iter = netlbl_af6list_search(&entry->addr, head); entry 191 net/netlabel/netlabel_addrlist.c ipv6_addr_equal(&iter->addr, &entry->addr) && entry 192 net/netlabel/netlabel_addrlist.c ipv6_addr_equal(&iter->mask, &entry->mask)) entry 201 net/netlabel/netlabel_addrlist.c ipv6_addr_cmp(&entry->mask, &iter->mask) > 0) { entry 202 net/netlabel/netlabel_addrlist.c __list_add_rcu(&entry->list, entry 207 net/netlabel/netlabel_addrlist.c list_add_tail_rcu(&entry->list, head); entry 221 net/netlabel/netlabel_addrlist.c void netlbl_af4list_remove_entry(struct netlbl_af4list *entry) entry 223 net/netlabel/netlabel_addrlist.c entry->valid = 0; entry 224 net/netlabel/netlabel_addrlist.c list_del_rcu(&entry->list); entry 242 net/netlabel/netlabel_addrlist.c struct netlbl_af4list *entry; entry 244 net/netlabel/netlabel_addrlist.c entry = netlbl_af4list_search_exact(addr, mask, head); entry 245 net/netlabel/netlabel_addrlist.c if (entry == NULL) entry 247 net/netlabel/netlabel_addrlist.c netlbl_af4list_remove_entry(entry); entry 248 net/netlabel/netlabel_addrlist.c return entry; entry 261 net/netlabel/netlabel_addrlist.c void netlbl_af6list_remove_entry(struct netlbl_af6list *entry) entry 263 net/netlabel/netlabel_addrlist.c entry->valid = 0; entry 264 net/netlabel/netlabel_addrlist.c list_del_rcu(&entry->list); entry 283 net/netlabel/netlabel_addrlist.c struct netlbl_af6list *entry; entry 285 net/netlabel/netlabel_addrlist.c entry = netlbl_af6list_search_exact(addr, mask, head); entry 286 net/netlabel/netlabel_addrlist.c if (entry == NULL) entry 288 net/netlabel/netlabel_addrlist.c netlbl_af6list_remove_entry(entry); entry 289 net/netlabel/netlabel_addrlist.c return entry; entry 98 net/netlabel/netlabel_addrlist.h int netlbl_af4list_add(struct netlbl_af4list *entry, entry 102 net/netlabel/netlabel_addrlist.h void netlbl_af4list_remove_entry(struct netlbl_af4list *entry); entry 165 net/netlabel/netlabel_addrlist.h int netlbl_af6list_add(struct netlbl_af6list *entry, entry 170 net/netlabel/netlabel_addrlist.h void netlbl_af6list_remove_entry(struct netlbl_af6list *entry); entry 258 net/netlabel/netlabel_calipso.c static int netlbl_calipso_remove_cb(struct netlbl_dom_map *entry, void *arg) entry 262 net/netlabel/netlabel_calipso.c if (entry->def.type == NETLBL_NLTYPE_CALIPSO && entry 263 net/netlabel/netlabel_calipso.c entry->def.calipso->doi == cb_arg->doi) entry 264 net/netlabel/netlabel_calipso.c return netlbl_domhsh_remove_entry(entry, cb_arg->audit_info); entry 677 net/netlabel/netlabel_cipso_v4.c static int netlbl_cipsov4_remove_cb(struct netlbl_dom_map *entry, void *arg) entry 681 net/netlabel/netlabel_cipso_v4.c if (entry->def.type == NETLBL_NLTYPE_CIPSOV4 && entry 682 net/netlabel/netlabel_cipso_v4.c entry->def.cipso->doi == cb_arg->doi) entry 683 net/netlabel/netlabel_cipso_v4.c return netlbl_domhsh_remove_entry(entry, cb_arg->audit_info); entry 153 net/netlabel/netlabel_cipso_v4.h void netlbl_cipsov4_doi_free(struct rcu_head *entry); entry 64 net/netlabel/netlabel_domainhash.c static void netlbl_domhsh_free_entry(struct rcu_head *entry) entry 74 net/netlabel/netlabel_domainhash.c ptr = container_of(entry, struct netlbl_dom_map, rcu); entry 173 net/netlabel/netlabel_domainhash.c struct netlbl_dom_map *entry; entry 175 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_search(domain, family); entry 176 net/netlabel/netlabel_domainhash.c if (entry != NULL) entry 177 net/netlabel/netlabel_domainhash.c return entry; entry 179 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_rcu_deref(netlbl_domhsh_def_ipv4); entry 180 net/netlabel/netlabel_domainhash.c if (entry != NULL && entry->valid) entry 181 net/netlabel/netlabel_domainhash.c return entry; entry 184 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_rcu_deref(netlbl_domhsh_def_ipv6); entry 185 net/netlabel/netlabel_domainhash.c if (entry != NULL && entry->valid) entry 186 net/netlabel/netlabel_domainhash.c return entry; entry 206 net/netlabel/netlabel_domainhash.c static void netlbl_domhsh_audit_add(struct netlbl_dom_map *entry, entry 220 net/netlabel/netlabel_domainhash.c entry->domain ? entry->domain : "(default)"); entry 238 net/netlabel/netlabel_domainhash.c type = entry->def.type; entry 239 net/netlabel/netlabel_domainhash.c cipsov4 = entry->def.cipso; entry 240 net/netlabel/netlabel_domainhash.c calipso = entry->def.calipso; entry 272 net/netlabel/netlabel_domainhash.c static int netlbl_domhsh_validate(const struct netlbl_dom_map *entry) entry 281 net/netlabel/netlabel_domainhash.c if (entry == NULL) entry 284 net/netlabel/netlabel_domainhash.c if (entry->family != AF_INET && entry->family != AF_INET6 && entry 285 net/netlabel/netlabel_domainhash.c (entry->family != AF_UNSPEC || entry 286 net/netlabel/netlabel_domainhash.c entry->def.type != NETLBL_NLTYPE_UNLABELED)) entry 289 net/netlabel/netlabel_domainhash.c switch (entry->def.type) { entry 291 net/netlabel/netlabel_domainhash.c if (entry->def.cipso != NULL || entry->def.calipso != NULL || entry 292 net/netlabel/netlabel_domainhash.c entry->def.addrsel != NULL) entry 296 net/netlabel/netlabel_domainhash.c if (entry->family != AF_INET || entry 297 net/netlabel/netlabel_domainhash.c entry->def.cipso == NULL) entry 301 net/netlabel/netlabel_domainhash.c if (entry->family != AF_INET6 || entry 302 net/netlabel/netlabel_domainhash.c entry->def.calipso == NULL) entry 306 net/netlabel/netlabel_domainhash.c netlbl_af4list_foreach(iter4, &entry->def.addrsel->list4) { entry 322 net/netlabel/netlabel_domainhash.c netlbl_af6list_foreach(iter6, &entry->def.addrsel->list6) { entry 403 net/netlabel/netlabel_domainhash.c int netlbl_domhsh_add(struct netlbl_dom_map *entry, entry 415 net/netlabel/netlabel_domainhash.c ret_val = netlbl_domhsh_validate(entry); entry 425 net/netlabel/netlabel_domainhash.c if (entry->domain != NULL) entry 426 net/netlabel/netlabel_domainhash.c entry_old = netlbl_domhsh_search(entry->domain, entry->family); entry 428 net/netlabel/netlabel_domainhash.c entry_old = netlbl_domhsh_search_def(entry->domain, entry 429 net/netlabel/netlabel_domainhash.c entry->family); entry 431 net/netlabel/netlabel_domainhash.c entry->valid = 1; entry 433 net/netlabel/netlabel_domainhash.c if (entry->domain != NULL) { entry 434 net/netlabel/netlabel_domainhash.c u32 bkt = netlbl_domhsh_hash(entry->domain); entry 435 net/netlabel/netlabel_domainhash.c list_add_tail_rcu(&entry->list, entry 438 net/netlabel/netlabel_domainhash.c INIT_LIST_HEAD(&entry->list); entry 439 net/netlabel/netlabel_domainhash.c switch (entry->family) { entry 442 net/netlabel/netlabel_domainhash.c entry); entry 446 net/netlabel/netlabel_domainhash.c entry); entry 449 net/netlabel/netlabel_domainhash.c if (entry->def.type != entry 462 net/netlabel/netlabel_domainhash.c entry->family = AF_INET; entry 464 net/netlabel/netlabel_domainhash.c entry); entry 476 net/netlabel/netlabel_domainhash.c if (entry->def.type == NETLBL_NLTYPE_ADDRSELECT) { entry 478 net/netlabel/netlabel_domainhash.c &entry->def.addrsel->list4) entry 479 net/netlabel/netlabel_domainhash.c netlbl_domhsh_audit_add(entry, iter4, NULL, entry 483 net/netlabel/netlabel_domainhash.c &entry->def.addrsel->list6) entry 484 net/netlabel/netlabel_domainhash.c netlbl_domhsh_audit_add(entry, NULL, iter6, entry 488 net/netlabel/netlabel_domainhash.c netlbl_domhsh_audit_add(entry, NULL, NULL, entry 491 net/netlabel/netlabel_domainhash.c entry->def.type == NETLBL_NLTYPE_ADDRSELECT) { entry 500 net/netlabel/netlabel_domainhash.c netlbl_af4list_foreach_rcu(iter4, &entry->def.addrsel->list4) entry 508 net/netlabel/netlabel_domainhash.c netlbl_af6list_foreach_rcu(iter6, &entry->def.addrsel->list6) entry 518 net/netlabel/netlabel_domainhash.c &entry->def.addrsel->list4) { entry 529 net/netlabel/netlabel_domainhash.c &entry->def.addrsel->list6) { entry 559 net/netlabel/netlabel_domainhash.c int netlbl_domhsh_add_default(struct netlbl_dom_map *entry, entry 562 net/netlabel/netlabel_domainhash.c return netlbl_domhsh_add(entry, audit_info); entry 577 net/netlabel/netlabel_domainhash.c int netlbl_domhsh_remove_entry(struct netlbl_dom_map *entry, entry 583 net/netlabel/netlabel_domainhash.c if (entry == NULL) entry 587 net/netlabel/netlabel_domainhash.c if (entry->valid) { entry 588 net/netlabel/netlabel_domainhash.c entry->valid = 0; entry 589 net/netlabel/netlabel_domainhash.c if (entry == rcu_dereference(netlbl_domhsh_def_ipv4)) entry 591 net/netlabel/netlabel_domainhash.c else if (entry == rcu_dereference(netlbl_domhsh_def_ipv6)) entry 594 net/netlabel/netlabel_domainhash.c list_del_rcu(&entry->list); entry 603 net/netlabel/netlabel_domainhash.c entry->domain ? entry->domain : "(default)", entry 616 net/netlabel/netlabel_domainhash.c switch (entry->def.type) { entry 619 net/netlabel/netlabel_domainhash.c &entry->def.addrsel->list4) { entry 625 net/netlabel/netlabel_domainhash.c &entry->def.addrsel->list6) { entry 632 net/netlabel/netlabel_domainhash.c cipso_v4_doi_putdef(entry->def.cipso); entry 636 net/netlabel/netlabel_domainhash.c calipso_doi_putdef(entry->def.calipso); entry 640 net/netlabel/netlabel_domainhash.c call_rcu(&entry->rcu, netlbl_domhsh_free_entry); entry 670 net/netlabel/netlabel_domainhash.c struct netlbl_domaddr4_map *entry; entry 704 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_addr4_entry(entry_addr); entry 705 net/netlabel/netlabel_domainhash.c cipso_v4_doi_putdef(entry->def.cipso); entry 706 net/netlabel/netlabel_domainhash.c kfree(entry); entry 737 net/netlabel/netlabel_domainhash.c struct netlbl_domaddr6_map *entry; entry 769 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_addr6_entry(entry_addr); entry 770 net/netlabel/netlabel_domainhash.c calipso_doi_putdef(entry->def.calipso); entry 771 net/netlabel/netlabel_domainhash.c kfree(entry); entry 797 net/netlabel/netlabel_domainhash.c struct netlbl_dom_map *entry; entry 803 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_search(domain, AF_INET); entry 805 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_search_def(domain, AF_INET); entry 806 net/netlabel/netlabel_domainhash.c ret_val = netlbl_domhsh_remove_entry(entry, audit_info); entry 814 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_search(domain, AF_INET6); entry 816 net/netlabel/netlabel_domainhash.c entry = netlbl_domhsh_search_def(domain, AF_INET6); entry 817 net/netlabel/netlabel_domainhash.c ret_val2 = netlbl_domhsh_remove_entry(entry, audit_info); entry 940 net/netlabel/netlabel_domainhash.c int (*callback) (struct netlbl_dom_map *entry, void *arg), entry 72 net/netlabel/netlabel_domainhash.h int netlbl_domhsh_add(struct netlbl_dom_map *entry, entry 74 net/netlabel/netlabel_domainhash.h int netlbl_domhsh_add_default(struct netlbl_dom_map *entry, entry 76 net/netlabel/netlabel_domainhash.h int netlbl_domhsh_remove_entry(struct netlbl_dom_map *entry, entry 103 net/netlabel/netlabel_domainhash.h int (*callback) (struct netlbl_dom_map *entry, void *arg), entry 102 net/netlabel/netlabel_kapi.c struct netlbl_dom_map *entry; entry 107 net/netlabel/netlabel_kapi.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 108 net/netlabel/netlabel_kapi.c if (entry == NULL) entry 111 net/netlabel/netlabel_kapi.c entry->domain = kstrdup(domain, GFP_ATOMIC); entry 112 net/netlabel/netlabel_kapi.c if (entry->domain == NULL) entry 115 net/netlabel/netlabel_kapi.c entry->family = family; entry 118 net/netlabel/netlabel_kapi.c entry->def.type = NETLBL_NLTYPE_UNLABELED; entry 169 net/netlabel/netlabel_kapi.c entry->def.addrsel = addrmap; entry 170 net/netlabel/netlabel_kapi.c entry->def.type = NETLBL_NLTYPE_ADDRSELECT; entry 176 net/netlabel/netlabel_kapi.c ret_val = netlbl_domhsh_add(entry, audit_info); entry 183 net/netlabel/netlabel_kapi.c kfree(entry->domain); entry 184 net/netlabel/netlabel_kapi.c kfree(entry); entry 331 net/netlabel/netlabel_kapi.c struct netlbl_dom_map *entry; entry 339 net/netlabel/netlabel_kapi.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 340 net/netlabel/netlabel_kapi.c if (entry == NULL) entry 342 net/netlabel/netlabel_kapi.c entry->family = AF_INET; entry 344 net/netlabel/netlabel_kapi.c entry->domain = kstrdup(domain, GFP_ATOMIC); entry 345 net/netlabel/netlabel_kapi.c if (entry->domain == NULL) entry 350 net/netlabel/netlabel_kapi.c entry->def.cipso = doi_def; entry 351 net/netlabel/netlabel_kapi.c entry->def.type = NETLBL_NLTYPE_CIPSOV4; entry 371 net/netlabel/netlabel_kapi.c entry->def.addrsel = addrmap; entry 372 net/netlabel/netlabel_kapi.c entry->def.type = NETLBL_NLTYPE_ADDRSELECT; entry 378 net/netlabel/netlabel_kapi.c ret_val = netlbl_domhsh_add(entry, audit_info); entry 389 net/netlabel/netlabel_kapi.c kfree(entry->domain); entry 391 net/netlabel/netlabel_kapi.c kfree(entry); entry 457 net/netlabel/netlabel_kapi.c struct netlbl_dom_map *entry; entry 465 net/netlabel/netlabel_kapi.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 466 net/netlabel/netlabel_kapi.c if (entry == NULL) entry 468 net/netlabel/netlabel_kapi.c entry->family = AF_INET6; entry 470 net/netlabel/netlabel_kapi.c entry->domain = kstrdup(domain, GFP_ATOMIC); entry 471 net/netlabel/netlabel_kapi.c if (entry->domain == NULL) entry 476 net/netlabel/netlabel_kapi.c entry->def.calipso = doi_def; entry 477 net/netlabel/netlabel_kapi.c entry->def.type = NETLBL_NLTYPE_CALIPSO; entry 501 net/netlabel/netlabel_kapi.c entry->def.addrsel = addrmap; entry 502 net/netlabel/netlabel_kapi.c entry->def.type = NETLBL_NLTYPE_ADDRSELECT; entry 508 net/netlabel/netlabel_kapi.c ret_val = netlbl_domhsh_add(entry, audit_info); entry 519 net/netlabel/netlabel_kapi.c kfree(entry->domain); entry 521 net/netlabel/netlabel_kapi.c kfree(entry); entry 1111 net/netlabel/netlabel_kapi.c struct netlbl_dommap_def *entry; entry 1117 net/netlabel/netlabel_kapi.c entry = netlbl_domhsh_getentry_af4(secattr->domain, entry 1119 net/netlabel/netlabel_kapi.c if (entry == NULL) { entry 1123 net/netlabel/netlabel_kapi.c switch (entry->type) { entry 1126 net/netlabel/netlabel_kapi.c entry->cipso, secattr); entry 1141 net/netlabel/netlabel_kapi.c entry = netlbl_domhsh_getentry_af6(secattr->domain, entry 1143 net/netlabel/netlabel_kapi.c if (entry == NULL) { entry 1147 net/netlabel/netlabel_kapi.c switch (entry->type) { entry 1150 net/netlabel/netlabel_kapi.c entry->calipso, secattr); entry 1186 net/netlabel/netlabel_kapi.c struct netlbl_dommap_def *entry; entry 1192 net/netlabel/netlabel_kapi.c entry = netlbl_domhsh_getentry_af4(secattr->domain, entry 1194 net/netlabel/netlabel_kapi.c if (entry == NULL) { entry 1198 net/netlabel/netlabel_kapi.c switch (entry->type) { entry 1201 net/netlabel/netlabel_kapi.c entry->cipso, secattr); entry 1213 net/netlabel/netlabel_kapi.c entry = netlbl_domhsh_getentry_af6(secattr->domain, entry 1215 net/netlabel/netlabel_kapi.c if (entry == NULL) { entry 1219 net/netlabel/netlabel_kapi.c switch (entry->type) { entry 1222 net/netlabel/netlabel_kapi.c entry->calipso, secattr); entry 1284 net/netlabel/netlabel_kapi.c struct netlbl_dommap_def *entry; entry 1290 net/netlabel/netlabel_kapi.c entry = netlbl_domhsh_getentry_af4(secattr->domain, entry 1292 net/netlabel/netlabel_kapi.c if (entry == NULL) { entry 1296 net/netlabel/netlabel_kapi.c switch (entry->type) { entry 1298 net/netlabel/netlabel_kapi.c ret_val = cipso_v4_skbuff_setattr(skb, entry->cipso, entry 1313 net/netlabel/netlabel_kapi.c entry = netlbl_domhsh_getentry_af6(secattr->domain, entry 1315 net/netlabel/netlabel_kapi.c if (entry == NULL) { entry 1319 net/netlabel/netlabel_kapi.c switch (entry->type) { entry 1321 net/netlabel/netlabel_kapi.c ret_val = calipso_skbuff_setattr(skb, entry->calipso, entry 86 net/netlabel/netlabel_mgmt.c struct netlbl_dom_map *entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 88 net/netlabel/netlabel_mgmt.c if (!entry) entry 90 net/netlabel/netlabel_mgmt.c entry->def.type = nla_get_u32(info->attrs[NLBL_MGMT_A_PROTOCOL]); entry 93 net/netlabel/netlabel_mgmt.c entry->domain = kmalloc(tmp_size, GFP_KERNEL); entry 94 net/netlabel/netlabel_mgmt.c if (entry->domain == NULL) { entry 98 net/netlabel/netlabel_mgmt.c nla_strlcpy(entry->domain, entry 107 net/netlabel/netlabel_mgmt.c switch (entry->def.type) { entry 110 net/netlabel/netlabel_mgmt.c entry->family = entry 113 net/netlabel/netlabel_mgmt.c entry->family = AF_UNSPEC; entry 123 net/netlabel/netlabel_mgmt.c entry->family = AF_INET; entry 124 net/netlabel/netlabel_mgmt.c entry->def.cipso = cipsov4; entry 135 net/netlabel/netlabel_mgmt.c entry->family = AF_INET6; entry 136 net/netlabel/netlabel_mgmt.c entry->def.calipso = calipso; entry 143 net/netlabel/netlabel_mgmt.c if ((entry->family == AF_INET && info->attrs[NLBL_MGMT_A_IPV6ADDR]) || entry 144 net/netlabel/netlabel_mgmt.c (entry->family == AF_INET6 && info->attrs[NLBL_MGMT_A_IPV4ADDR])) entry 181 net/netlabel/netlabel_mgmt.c map->def.type = entry->def.type; entry 191 net/netlabel/netlabel_mgmt.c entry->family = AF_INET; entry 192 net/netlabel/netlabel_mgmt.c entry->def.type = NETLBL_NLTYPE_ADDRSELECT; entry 193 net/netlabel/netlabel_mgmt.c entry->def.addrsel = addrmap; entry 233 net/netlabel/netlabel_mgmt.c map->def.type = entry->def.type; entry 243 net/netlabel/netlabel_mgmt.c entry->family = AF_INET6; entry 244 net/netlabel/netlabel_mgmt.c entry->def.type = NETLBL_NLTYPE_ADDRSELECT; entry 245 net/netlabel/netlabel_mgmt.c entry->def.addrsel = addrmap; entry 249 net/netlabel/netlabel_mgmt.c ret_val = netlbl_domhsh_add(entry, audit_info); entry 263 net/netlabel/netlabel_mgmt.c kfree(entry->domain); entry 265 net/netlabel/netlabel_mgmt.c kfree(entry); entry 281 net/netlabel/netlabel_mgmt.c struct netlbl_dom_map *entry) entry 291 net/netlabel/netlabel_mgmt.c if (entry->domain != NULL) { entry 293 net/netlabel/netlabel_mgmt.c NLBL_MGMT_A_DOMAIN, entry->domain); entry 298 net/netlabel/netlabel_mgmt.c ret_val = nla_put_u16(skb, NLBL_MGMT_A_FAMILY, entry->family); entry 302 net/netlabel/netlabel_mgmt.c switch (entry->def.type) { entry 308 net/netlabel/netlabel_mgmt.c netlbl_af4list_foreach_rcu(iter4, &entry->def.addrsel->list4) { entry 344 net/netlabel/netlabel_mgmt.c netlbl_af6list_foreach_rcu(iter6, &entry->def.addrsel->list6) { entry 383 net/netlabel/netlabel_mgmt.c entry->def.type); entry 387 net/netlabel/netlabel_mgmt.c entry->def.type); entry 391 net/netlabel/netlabel_mgmt.c entry->def.cipso->doi); entry 395 net/netlabel/netlabel_mgmt.c entry->def.type); entry 399 net/netlabel/netlabel_mgmt.c entry->def.calipso->doi); entry 478 net/netlabel/netlabel_mgmt.c static int netlbl_mgmt_listall_cb(struct netlbl_dom_map *entry, void *arg) entry 490 net/netlabel/netlabel_mgmt.c ret_val = netlbl_mgmt_listentry(cb_arg->skb, entry); entry 600 net/netlabel/netlabel_mgmt.c struct netlbl_dom_map *entry; entry 617 net/netlabel/netlabel_mgmt.c entry = netlbl_domhsh_getentry(NULL, family); entry 618 net/netlabel/netlabel_mgmt.c if (entry == NULL) { entry 622 net/netlabel/netlabel_mgmt.c ret_val = netlbl_mgmt_listentry(ans_skb, entry); entry 147 net/netlabel/netlabel_unlabeled.c static void netlbl_unlhsh_free_iface(struct rcu_head *entry) entry 157 net/netlabel/netlabel_unlabeled.c iface = container_of(entry, struct netlbl_unlhsh_iface, rcu); entry 236 net/netlabel/netlabel_unlabeled.c struct netlbl_unlhsh_addr4 *entry; entry 238 net/netlabel/netlabel_unlabeled.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 239 net/netlabel/netlabel_unlabeled.c if (entry == NULL) entry 242 net/netlabel/netlabel_unlabeled.c entry->list.addr = addr->s_addr & mask->s_addr; entry 243 net/netlabel/netlabel_unlabeled.c entry->list.mask = mask->s_addr; entry 244 net/netlabel/netlabel_unlabeled.c entry->list.valid = 1; entry 245 net/netlabel/netlabel_unlabeled.c entry->secid = secid; entry 248 net/netlabel/netlabel_unlabeled.c ret_val = netlbl_af4list_add(&entry->list, &iface->addr4_list); entry 252 net/netlabel/netlabel_unlabeled.c kfree(entry); entry 276 net/netlabel/netlabel_unlabeled.c struct netlbl_unlhsh_addr6 *entry; entry 278 net/netlabel/netlabel_unlabeled.c entry = kzalloc(sizeof(*entry), GFP_ATOMIC); entry 279 net/netlabel/netlabel_unlabeled.c if (entry == NULL) entry 282 net/netlabel/netlabel_unlabeled.c entry->list.addr = *addr; entry 283 net/netlabel/netlabel_unlabeled.c entry->list.addr.s6_addr32[0] &= mask->s6_addr32[0]; entry 284 net/netlabel/netlabel_unlabeled.c entry->list.addr.s6_addr32[1] &= mask->s6_addr32[1]; entry 285 net/netlabel/netlabel_unlabeled.c entry->list.addr.s6_addr32[2] &= mask->s6_addr32[2]; entry 286 net/netlabel/netlabel_unlabeled.c entry->list.addr.s6_addr32[3] &= mask->s6_addr32[3]; entry 287 net/netlabel/netlabel_unlabeled.c entry->list.mask = *mask; entry 288 net/netlabel/netlabel_unlabeled.c entry->list.valid = 1; entry 289 net/netlabel/netlabel_unlabeled.c entry->secid = secid; entry 292 net/netlabel/netlabel_unlabeled.c ret_val = netlbl_af6list_add(&entry->list, &iface->addr6_list); entry 296 net/netlabel/netlabel_unlabeled.c kfree(entry); entry 472 net/netlabel/netlabel_unlabeled.c struct netlbl_unlhsh_addr4 *entry; entry 483 net/netlabel/netlabel_unlabeled.c entry = netlbl_unlhsh_addr4_entry(list_entry); entry 485 net/netlabel/netlabel_unlabeled.c entry = NULL; entry 496 net/netlabel/netlabel_unlabeled.c if (entry != NULL && entry 497 net/netlabel/netlabel_unlabeled.c security_secid_to_secctx(entry->secid, entry 502 net/netlabel/netlabel_unlabeled.c audit_log_format(audit_buf, " res=%u", entry != NULL ? 1 : 0); entry 506 net/netlabel/netlabel_unlabeled.c if (entry == NULL) entry 509 net/netlabel/netlabel_unlabeled.c kfree_rcu(entry, rcu); entry 534 net/netlabel/netlabel_unlabeled.c struct netlbl_unlhsh_addr6 *entry; entry 544 net/netlabel/netlabel_unlabeled.c entry = netlbl_unlhsh_addr6_entry(list_entry); entry 546 net/netlabel/netlabel_unlabeled.c entry = NULL; entry 557 net/netlabel/netlabel_unlabeled.c if (entry != NULL && entry 558 net/netlabel/netlabel_unlabeled.c security_secid_to_secctx(entry->secid, entry 563 net/netlabel/netlabel_unlabeled.c audit_log_format(audit_buf, " res=%u", entry != NULL ? 1 : 0); entry 567 net/netlabel/netlabel_unlabeled.c if (entry == NULL) entry 570 net/netlabel/netlabel_unlabeled.c kfree_rcu(entry, rcu); entry 1528 net/netlabel/netlabel_unlabeled.c struct netlbl_dom_map *entry; entry 1538 net/netlabel/netlabel_unlabeled.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1539 net/netlabel/netlabel_unlabeled.c if (entry == NULL) entry 1541 net/netlabel/netlabel_unlabeled.c entry->family = AF_UNSPEC; entry 1542 net/netlabel/netlabel_unlabeled.c entry->def.type = NETLBL_NLTYPE_UNLABELED; entry 1543 net/netlabel/netlabel_unlabeled.c ret_val = netlbl_domhsh_add_default(entry, &audit_info); entry 37 net/nfc/hci/llc.c list_for_each_entry_safe(llc_engine, n, &llc_engines, entry) { entry 38 net/nfc/hci/llc.c list_del(&llc_engine->entry); entry 59 net/nfc/hci/llc.c INIT_LIST_HEAD(&llc_engine->entry); entry 60 net/nfc/hci/llc.c list_add_tail(&llc_engine->entry, &llc_engines); entry 69 net/nfc/hci/llc.c list_for_each_entry(llc_engine, &llc_engines, entry) { entry 85 net/nfc/hci/llc.c list_del(&llc_engine->entry); entry 30 net/nfc/hci/llc.h struct list_head entry; entry 3387 net/sched/cls_api.c struct flow_action_entry *entry; entry 3390 net/sched/cls_api.c flow_action_for_each(i, entry, flow_action) entry 3391 net/sched/cls_api.c if (entry->destructor) entry 3392 net/sched/cls_api.c entry->destructor(entry->destructor_priv); entry 3396 net/sched/cls_api.c static void tcf_mirred_get_dev(struct flow_action_entry *entry, entry 3400 net/sched/cls_api.c entry->dev = act->ops->get_dev(act, &entry->destructor); entry 3401 net/sched/cls_api.c if (!entry->dev) entry 3403 net/sched/cls_api.c entry->destructor_priv = entry->dev; entry 3414 net/sched/cls_api.c static int tcf_tunnel_encap_get_tunnel(struct flow_action_entry *entry, entry 3417 net/sched/cls_api.c entry->tunnel = tcf_tunnel_info_copy(act); entry 3418 net/sched/cls_api.c if (!entry->tunnel) entry 3420 net/sched/cls_api.c entry->destructor = tcf_tunnel_encap_put_tunnel; entry 3421 net/sched/cls_api.c entry->destructor_priv = entry->tunnel; entry 3425 net/sched/cls_api.c static void tcf_sample_get_group(struct flow_action_entry *entry, entry 3429 net/sched/cls_api.c entry->sample.psample_group = entry 3430 net/sched/cls_api.c act->ops->get_psample_group(act, &entry->destructor); entry 3431 net/sched/cls_api.c entry->destructor_priv = entry->sample.psample_group; entry 3449 net/sched/cls_api.c struct flow_action_entry *entry; entry 3451 net/sched/cls_api.c entry = &flow_action->entries[j]; entry 3453 net/sched/cls_api.c entry->id = FLOW_ACTION_ACCEPT; entry 3455 net/sched/cls_api.c entry->id = FLOW_ACTION_DROP; entry 3457 net/sched/cls_api.c entry->id = FLOW_ACTION_TRAP; entry 3459 net/sched/cls_api.c entry->id = FLOW_ACTION_GOTO; entry 3460 net/sched/cls_api.c entry->chain_index = tcf_gact_goto_chain_index(act); entry 3462 net/sched/cls_api.c entry->id = FLOW_ACTION_REDIRECT; entry 3463 net/sched/cls_api.c tcf_mirred_get_dev(entry, act); entry 3465 net/sched/cls_api.c entry->id = FLOW_ACTION_MIRRED; entry 3466 net/sched/cls_api.c tcf_mirred_get_dev(entry, act); entry 3468 net/sched/cls_api.c entry->id = FLOW_ACTION_REDIRECT_INGRESS; entry 3469 net/sched/cls_api.c tcf_mirred_get_dev(entry, act); entry 3471 net/sched/cls_api.c entry->id = FLOW_ACTION_MIRRED_INGRESS; entry 3472 net/sched/cls_api.c tcf_mirred_get_dev(entry, act); entry 3476 net/sched/cls_api.c entry->id = FLOW_ACTION_VLAN_PUSH; entry 3477 net/sched/cls_api.c entry->vlan.vid = tcf_vlan_push_vid(act); entry 3478 net/sched/cls_api.c entry->vlan.proto = tcf_vlan_push_proto(act); entry 3479 net/sched/cls_api.c entry->vlan.prio = tcf_vlan_push_prio(act); entry 3482 net/sched/cls_api.c entry->id = FLOW_ACTION_VLAN_POP; entry 3485 net/sched/cls_api.c entry->id = FLOW_ACTION_VLAN_MANGLE; entry 3486 net/sched/cls_api.c entry->vlan.vid = tcf_vlan_push_vid(act); entry 3487 net/sched/cls_api.c entry->vlan.proto = tcf_vlan_push_proto(act); entry 3488 net/sched/cls_api.c entry->vlan.prio = tcf_vlan_push_prio(act); entry 3495 net/sched/cls_api.c entry->id = FLOW_ACTION_TUNNEL_ENCAP; entry 3496 net/sched/cls_api.c err = tcf_tunnel_encap_get_tunnel(entry, act); entry 3500 net/sched/cls_api.c entry->id = FLOW_ACTION_TUNNEL_DECAP; entry 3505 net/sched/cls_api.c entry->id = FLOW_ACTION_MANGLE; entry 3508 net/sched/cls_api.c entry->id = FLOW_ACTION_ADD; entry 3514 net/sched/cls_api.c entry->mangle.htype = tcf_pedit_htype(act, k); entry 3515 net/sched/cls_api.c entry->mangle.mask = tcf_pedit_mask(act, k); entry 3516 net/sched/cls_api.c entry->mangle.val = tcf_pedit_val(act, k); entry 3517 net/sched/cls_api.c entry->mangle.offset = tcf_pedit_offset(act, k); entry 3518 net/sched/cls_api.c entry = &flow_action->entries[++j]; entry 3521 net/sched/cls_api.c entry->id = FLOW_ACTION_CSUM; entry 3522 net/sched/cls_api.c entry->csum_flags = tcf_csum_update_flags(act); entry 3524 net/sched/cls_api.c entry->id = FLOW_ACTION_MARK; entry 3525 net/sched/cls_api.c entry->mark = tcf_skbedit_mark(act); entry 3527 net/sched/cls_api.c entry->id = FLOW_ACTION_SAMPLE; entry 3528 net/sched/cls_api.c entry->sample.trunc_size = tcf_sample_trunc_size(act); entry 3529 net/sched/cls_api.c entry->sample.truncate = tcf_sample_truncate(act); entry 3530 net/sched/cls_api.c entry->sample.rate = tcf_sample_rate(act); entry 3531 net/sched/cls_api.c tcf_sample_get_group(entry, act); entry 3533 net/sched/cls_api.c entry->id = FLOW_ACTION_POLICE; entry 3534 net/sched/cls_api.c entry->police.burst = tcf_police_tcfp_burst(act); entry 3535 net/sched/cls_api.c entry->police.rate_bytes_ps = entry 3538 net/sched/cls_api.c entry->id = FLOW_ACTION_CT; entry 3539 net/sched/cls_api.c entry->ct.action = tcf_ct_action(act); entry 3540 net/sched/cls_api.c entry->ct.zone = tcf_ct_zone(act); entry 3544 net/sched/cls_api.c entry->id = FLOW_ACTION_MPLS_PUSH; entry 3545 net/sched/cls_api.c entry->mpls_push.proto = tcf_mpls_proto(act); entry 3546 net/sched/cls_api.c entry->mpls_push.label = tcf_mpls_label(act); entry 3547 net/sched/cls_api.c entry->mpls_push.tc = tcf_mpls_tc(act); entry 3548 net/sched/cls_api.c entry->mpls_push.bos = tcf_mpls_bos(act); entry 3549 net/sched/cls_api.c entry->mpls_push.ttl = tcf_mpls_ttl(act); entry 3552 net/sched/cls_api.c entry->id = FLOW_ACTION_MPLS_POP; entry 3553 net/sched/cls_api.c entry->mpls_pop.proto = tcf_mpls_proto(act); entry 3556 net/sched/cls_api.c entry->id = FLOW_ACTION_MPLS_MANGLE; entry 3557 net/sched/cls_api.c entry->mpls_mangle.label = tcf_mpls_label(act); entry 3558 net/sched/cls_api.c entry->mpls_mangle.tc = tcf_mpls_tc(act); entry 3559 net/sched/cls_api.c entry->mpls_mangle.bos = tcf_mpls_bos(act); entry 3560 net/sched/cls_api.c entry->mpls_mangle.ttl = tcf_mpls_ttl(act); entry 3566 net/sched/cls_api.c entry->id = FLOW_ACTION_PTYPE; entry 3567 net/sched/cls_api.c entry->ptype = tcf_skbedit_ptype(act); entry 531 net/sched/sch_gred.c static void gred_vq_apply(struct gred_sched *table, const struct nlattr *entry) entry 536 net/sched/sch_gred.c nla_parse_nested_deprecated(tb, TCA_GRED_VQ_MAX, entry, entry 560 net/sched/sch_gred.c const struct nlattr *entry, entry 567 net/sched/sch_gred.c err = nla_parse_nested_deprecated(tb, TCA_GRED_VQ_MAX, entry, entry 114 net/sched/sch_taprio.c struct sched_entry *entry, *n; entry 119 net/sched/sch_taprio.c list_for_each_entry_safe(entry, n, &sched->entries, list) { entry 120 net/sched/sch_taprio.c list_del(&entry->list); entry 121 net/sched/sch_taprio.c kfree(entry); entry 155 net/sched/sch_taprio.c struct sched_entry *entry, entry 162 net/sched/sch_taprio.c intv_end = ktime_add_ns(intv_start, entry->interval); entry 195 net/sched/sch_taprio.c struct sched_entry *entry = NULL, *entry_found = NULL; entry 216 net/sched/sch_taprio.c list_for_each_entry(entry, &sched->entries, list) { entry 218 net/sched/sch_taprio.c curr_intv_end = get_interval_end_time(sched, admin, entry, entry 224 net/sched/sch_taprio.c if (!(entry->gate_mask & BIT(tc)) || entry 225 net/sched/sch_taprio.c packet_transmit_time > entry->interval) entry 228 net/sched/sch_taprio.c txtime = entry->next_txtime; entry 235 net/sched/sch_taprio.c entry_found = entry; entry 244 net/sched/sch_taprio.c entry_found = entry; entry 251 net/sched/sch_taprio.c entry_found = entry; entry 266 net/sched/sch_taprio.c struct sched_entry *entry; entry 272 net/sched/sch_taprio.c entry = find_entry_to_transmit(skb, sch, sched, admin, skb->tstamp, entry 276 net/sched/sch_taprio.c return entry; entry 349 net/sched/sch_taprio.c struct sched_entry *entry; entry 376 net/sched/sch_taprio.c entry = find_entry_to_transmit(skb, sch, sched, admin, entry 380 net/sched/sch_taprio.c if (!entry) { entry 385 net/sched/sch_taprio.c txtime = entry->next_txtime; entry 403 net/sched/sch_taprio.c entry->next_txtime = ktime_add(interval_start, sched->cycle_time); entry 406 net/sched/sch_taprio.c entry->next_txtime = transmit_end_time; entry 445 net/sched/sch_taprio.c struct sched_entry *entry; entry 451 net/sched/sch_taprio.c entry = rcu_dereference(q->current_entry); entry 452 net/sched/sch_taprio.c gate_mask = entry ? entry->gate_mask : TAPRIO_ALL_GATES_OPEN; entry 515 net/sched/sch_taprio.c static void taprio_set_budget(struct taprio_sched *q, struct sched_entry *entry) entry 517 net/sched/sch_taprio.c atomic_set(&entry->budget, entry 518 net/sched/sch_taprio.c div64_u64((u64)entry->interval * 1000, entry 527 net/sched/sch_taprio.c struct sched_entry *entry; entry 532 net/sched/sch_taprio.c entry = rcu_dereference(q->current_entry); entry 538 net/sched/sch_taprio.c gate_mask = entry ? entry->gate_mask : TAPRIO_ALL_GATES_OPEN; entry 580 net/sched/sch_taprio.c ktime_after(guard, entry->close_time)) { entry 587 net/sched/sch_taprio.c atomic_sub_return(len, &entry->budget) < 0) { entry 645 net/sched/sch_taprio.c const struct sched_entry *entry) entry 647 net/sched/sch_taprio.c if (list_is_last(&entry->list, &oper->entries)) entry 650 net/sched/sch_taprio.c if (ktime_compare(entry->close_time, oper->cycle_close_time) == 0) entry 695 net/sched/sch_taprio.c struct sched_entry *entry, *next; entry 700 net/sched/sch_taprio.c entry = rcu_dereference_protected(q->current_entry, entry 716 net/sched/sch_taprio.c if (unlikely(!entry || entry->close_time == oper->base_time)) { entry 723 net/sched/sch_taprio.c if (should_restart_cycle(oper, entry)) { entry 729 net/sched/sch_taprio.c next = list_next_entry(entry, list); entry 732 net/sched/sch_taprio.c close_time = ktime_add_ns(entry->close_time, next->interval); entry 780 net/sched/sch_taprio.c static int fill_sched_entry(struct nlattr **tb, struct sched_entry *entry, entry 786 net/sched/sch_taprio.c entry->command = nla_get_u8( entry 790 net/sched/sch_taprio.c entry->gate_mask = nla_get_u32( entry 802 net/sched/sch_taprio.c entry->interval = interval; entry 807 net/sched/sch_taprio.c static int parse_sched_entry(struct nlattr *n, struct sched_entry *entry, entry 820 net/sched/sch_taprio.c entry->index = index; entry 822 net/sched/sch_taprio.c return fill_sched_entry(tb, entry, extack); entry 837 net/sched/sch_taprio.c struct sched_entry *entry; entry 844 net/sched/sch_taprio.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 845 net/sched/sch_taprio.c if (!entry) { entry 850 net/sched/sch_taprio.c err = parse_sched_entry(n, entry, i, extack); entry 852 net/sched/sch_taprio.c kfree(entry); entry 856 net/sched/sch_taprio.c list_add_tail(&entry->list, &sched->entries); entry 892 net/sched/sch_taprio.c struct sched_entry *entry; entry 895 net/sched/sch_taprio.c list_for_each_entry(entry, &new->entries, list) entry 896 net/sched/sch_taprio.c cycle = ktime_add_ns(cycle, entry->interval); entry 1100 net/sched/sch_taprio.c struct sched_entry *entry; entry 1103 net/sched/sch_taprio.c list_for_each_entry(entry, &sched->entries, list) { entry 1104 net/sched/sch_taprio.c entry->next_txtime = ktime_add_ns(base, interval); entry 1105 net/sched/sch_taprio.c interval += entry->interval; entry 1185 net/sched/sch_taprio.c struct sched_entry *entry; entry 1192 net/sched/sch_taprio.c list_for_each_entry(entry, &sched->entries, list) { entry 1195 net/sched/sch_taprio.c e->command = entry->command; entry 1196 net/sched/sch_taprio.c e->interval = entry->interval; entry 1197 net/sched/sch_taprio.c e->gate_mask = entry->gate_mask; entry 1712 net/sched/sch_taprio.c const struct sched_entry *entry) entry 1720 net/sched/sch_taprio.c if (nla_put_u32(msg, TCA_TAPRIO_SCHED_ENTRY_INDEX, entry->index)) entry 1723 net/sched/sch_taprio.c if (nla_put_u8(msg, TCA_TAPRIO_SCHED_ENTRY_CMD, entry->command)) entry 1727 net/sched/sch_taprio.c entry->gate_mask)) entry 1731 net/sched/sch_taprio.c entry->interval)) entry 1745 net/sched/sch_taprio.c struct sched_entry *entry; entry 1764 net/sched/sch_taprio.c list_for_each_entry(entry, &root->entries, list) { entry 1765 net/sched/sch_taprio.c if (dump_entry(msg, entry) < 0) entry 148 net/sctp/inqueue.c struct list_head *entry; entry 152 net/sctp/inqueue.c entry = sctp_list_dequeue(&queue->in_chunk_list); entry 153 net/sctp/inqueue.c if (!entry) entry 156 net/sctp/inqueue.c chunk = list_entry(entry, struct sctp_chunk, list); entry 3835 net/sctp/sm_statefuns.c struct list_head *entry; entry 3840 net/sctp/sm_statefuns.c entry = asoc->addip_chunk_list.next; entry 3841 net/sctp/sm_statefuns.c asconf = list_entry(entry, struct sctp_chunk, list); entry 3843 net/sctp/sm_statefuns.c list_del_init(entry); entry 62 net/sctp/stream_sched.c struct list_head *entry; entry 71 net/sctp/stream_sched.c entry = q->out_chunk_list.next; entry 72 net/sctp/stream_sched.c ch = list_entry(entry, struct sctp_chunk, list); entry 2418 net/socket.c struct mmsghdr __user *entry; entry 2437 net/socket.c entry = mmsg; entry 2455 net/socket.c (struct user_msghdr __user *)entry, entry 2459 net/socket.c err = put_user(err, &entry->msg_len); entry 2460 net/socket.c ++entry; entry 2640 net/socket.c struct mmsghdr __user *entry; entry 2665 net/socket.c entry = mmsg; entry 2682 net/socket.c (struct user_msghdr __user *)entry, entry 2687 net/socket.c err = put_user(err, &entry->msg_len); entry 2688 net/socket.c ++entry; entry 572 net/sunrpc/auth.c *entry, *new; entry 578 net/sunrpc/auth.c hlist_for_each_entry_rcu(entry, &cache->hashtable[nr], cr_hash) { entry 579 net/sunrpc/auth.c if (!entry->cr_ops->crmatch(acred, entry, flags)) entry 581 net/sunrpc/auth.c cred = get_rpccred(entry); entry 597 net/sunrpc/auth.c hlist_for_each_entry(entry, &cache->hashtable[nr], cr_hash) { entry 598 net/sunrpc/auth.c if (!entry->cr_ops->crmatch(acred, entry, flags)) entry 600 net/sunrpc/auth.c cred = get_rpccred(entry); entry 1320 net/sunrpc/cache.c unsigned int hash, entry; entry 1327 net/sunrpc/cache.c entry = n & ((1LL<<32) - 1); entry 1330 net/sunrpc/cache.c if (!entry--) entry 966 net/sunrpc/xprt.c struct rpc_rqst *entry; entry 968 net/sunrpc/xprt.c entry = xprt_request_rb_find(xprt, xid); entry 969 net/sunrpc/xprt.c if (entry != NULL) { entry 971 net/sunrpc/xprt.c entry->rq_rtt = ktime_sub(ktime_get(), entry->rq_xtime); entry 972 net/sunrpc/xprt.c return entry; entry 91 net/wireless/lib80211.c struct lib80211_crypt_data *entry, *next; entry 95 net/wireless/lib80211.c list_for_each_entry_safe(entry, next, &info->crypt_deinit_list, list) { entry 96 net/wireless/lib80211.c if (atomic_read(&entry->refcnt) != 0 && !force) entry 99 net/wireless/lib80211.c list_del(&entry->list); entry 101 net/wireless/lib80211.c if (entry->ops) { entry 102 net/wireless/lib80211.c entry->ops->deinit(entry->priv); entry 103 net/wireless/lib80211.c module_put(entry->ops->owner); entry 105 net/wireless/lib80211.c kfree(entry); entry 779 net/wireless/util.c ret = (ntohl(mpls->entry) & MPLS_LS_TC_MASK) entry 22 net/x25/x25_forward.c struct list_head *entry; entry 49 net/x25/x25_forward.c list_for_each(entry, &x25_forward_list) { entry 50 net/x25/x25_forward.c x25_frwd = list_entry(entry, struct x25_forward, node); entry 95 net/x25/x25_forward.c struct list_head *entry; entry 102 net/x25/x25_forward.c list_for_each(entry, &x25_forward_list) { entry 103 net/x25/x25_forward.c frwd = list_entry(entry, struct x25_forward, node); entry 135 net/x25/x25_forward.c struct list_head *entry, *tmp; entry 139 net/x25/x25_forward.c list_for_each_safe(entry, tmp, &x25_forward_list) { entry 140 net/x25/x25_forward.c fwd = list_entry(entry, struct x25_forward, node); entry 153 net/x25/x25_forward.c struct list_head *entry, *tmp; entry 157 net/x25/x25_forward.c list_for_each_safe(entry, tmp, &x25_forward_list) { entry 158 net/x25/x25_forward.c fwd = list_entry(entry, struct x25_forward, node); entry 295 net/x25/x25_link.c struct list_head *entry, *tmp; entry 299 net/x25/x25_link.c list_for_each_safe(entry, tmp, &x25_neigh_list) { entry 300 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); entry 317 net/x25/x25_link.c struct list_head *entry; entry 320 net/x25/x25_link.c list_for_each(entry, &x25_neigh_list) { entry 321 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); entry 393 net/x25/x25_link.c struct list_head *entry, *tmp; entry 397 net/x25/x25_link.c list_for_each_safe(entry, tmp, &x25_neigh_list) { entry 400 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); entry 30 net/x25/x25_route.c struct list_head *entry; entry 35 net/x25/x25_route.c list_for_each(entry, &x25_route_list) { entry 36 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); entry 81 net/x25/x25_route.c struct list_head *entry; entry 86 net/x25/x25_route.c list_for_each(entry, &x25_route_list) { entry 87 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); entry 107 net/x25/x25_route.c struct list_head *entry, *tmp; entry 111 net/x25/x25_route.c list_for_each_safe(entry, tmp, &x25_route_list) { entry 112 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); entry 152 net/x25/x25_route.c struct list_head *entry; entry 156 net/x25/x25_route.c list_for_each(entry, &x25_route_list) { entry 157 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); entry 213 net/x25/x25_route.c struct list_head *entry, *tmp; entry 216 net/x25/x25_route.c list_for_each_safe(entry, tmp, &x25_route_list) { entry 217 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); entry 642 net/xfrm/xfrm_algo.c int match(const struct xfrm_algo_desc *entry, const void *data), entry 669 net/xfrm/xfrm_algo.c static int xfrm_alg_id_match(const struct xfrm_algo_desc *entry, entry 672 net/xfrm/xfrm_algo.c return entry->desc.sadb_alg_id == (unsigned long)data; entry 696 net/xfrm/xfrm_algo.c static int xfrm_alg_name_match(const struct xfrm_algo_desc *entry, entry 701 net/xfrm/xfrm_algo.c return name && (!strcmp(name, entry->name) || entry 702 net/xfrm/xfrm_algo.c (entry->compat && !strcmp(name, entry->compat))); entry 731 net/xfrm/xfrm_algo.c static int xfrm_aead_name_match(const struct xfrm_algo_desc *entry, entry 737 net/xfrm/xfrm_algo.c return aead->icvbits == entry->uinfo.aead.icv_truncbits && name && entry 738 net/xfrm/xfrm_algo.c !strcmp(name, entry->name); entry 72 samples/bpf/test_lru_dist.c static inline void __list_del_entry(struct list_head *entry) entry 74 samples/bpf/test_lru_dist.c __list_del(entry->prev, entry->next); entry 87 samples/mic/mpssd/mpssd.h char *readsysfs(char *dir, char *entry); entry 88 samples/mic/mpssd/mpssd.h int setsysfs(char *dir, char *entry, char *value); entry 15 samples/mic/mpssd/sysfs.c readsysfs(char *dir, char *entry) entry 24 samples/mic/mpssd/sysfs.c snprintf(filename, PATH_MAX, "%s/%s", MICSYSFSDIR, entry); entry 27 samples/mic/mpssd/sysfs.c "%s/%s/%s", MICSYSFSDIR, dir, entry); entry 57 samples/mic/mpssd/sysfs.c setsysfs(char *dir, char *entry, char *value) entry 64 samples/mic/mpssd/sysfs.c snprintf(filename, PATH_MAX, "%s/%s", MICSYSFSDIR, entry); entry 67 samples/mic/mpssd/sysfs.c MICSYSFSDIR, dir, entry); entry 69 samples/mic/mpssd/sysfs.c oldvalue = readsysfs(dir, entry); entry 1396 scripts/asn1_compiler.c int entry; entry 1404 scripts/asn1_compiler.c e->entry_index = entry = nr_entries; entry 1417 scripts/asn1_compiler.c render_opcode(out, "_jump_target(%u),\n", entry); entry 1424 scripts/asn1_compiler.c render_opcode(out, "_jump_target(%u),\n", entry); entry 1443 scripts/asn1_compiler.c int entry, skippable = 0, outofline = 0; entry 1574 scripts/asn1_compiler.c entry = nr_entries; entry 1582 scripts/asn1_compiler.c render_opcode(out, "_jump_target(%u),\n", entry); entry 891 scripts/dtc/livetree.c char *entry; entry 900 scripts/dtc/livetree.c xasprintf(&entry, "%s:%s:%u", entry 902 scripts/dtc/livetree.c append_to_property(fn, m->ref, entry, strlen(entry) + 1, TYPE_STRING); entry 904 scripts/dtc/livetree.c free(entry); entry 802 scripts/gcc-plugins/gcc-common.h static inline void cgraph_remove_function_insertion_hook(struct cgraph_node_hook_list *entry) entry 804 scripts/gcc-plugins/gcc-common.h symtab->remove_cgraph_insertion_hook(entry); entry 812 scripts/gcc-plugins/gcc-common.h static inline void cgraph_remove_node_removal_hook(struct cgraph_node_hook_list *entry) entry 814 scripts/gcc-plugins/gcc-common.h symtab->remove_cgraph_removal_hook(entry); entry 822 scripts/gcc-plugins/gcc-common.h static inline void cgraph_remove_node_duplication_hook(struct cgraph_2node_hook_list *entry) entry 824 scripts/gcc-plugins/gcc-common.h symtab->remove_cgraph_duplication_hook(entry); entry 789 scripts/gcc-plugins/randomize_layout_plugin.c const struct whitelist_entry *entry; entry 792 scripts/gcc-plugins/randomize_layout_plugin.c for (entry = whitelist; entry->pathname; entry++) { entry 793 scripts/gcc-plugins/randomize_layout_plugin.c if (!strstr(xloc.file, entry->pathname)) entry 796 scripts/gcc-plugins/randomize_layout_plugin.c if (type_name_eq(stmt, lhs_tree, entry->lhs) && type_name_eq(stmt, rhs_tree, entry->rhs)) entry 126 scripts/kconfig/list.h static inline void list_del(struct list_head *entry) entry 128 scripts/kconfig/list.h __list_del(entry->prev, entry->next); entry 129 scripts/kconfig/list.h entry->next = (struct list_head*)LIST_POISON1; entry 130 scripts/kconfig/list.h entry->prev = (struct list_head*)LIST_POISON2; entry 1801 scripts/kconfig/qconf.cc QString entry; entry 1804 scripts/kconfig/qconf.cc entry = "single"; entry 1808 scripts/kconfig/qconf.cc entry = "split"; entry 1812 scripts/kconfig/qconf.cc entry = "full"; entry 1818 scripts/kconfig/qconf.cc configSettings->setValue("/listMode", entry); entry 104 security/integrity/ima/ima.h struct ima_template_entry *entry; entry 134 security/integrity/ima/ima.h int ima_add_template_entry(struct ima_template_entry *entry, int violation, entry 156 security/integrity/ima/ima.h int ima_restore_measurement_entry(struct ima_template_entry *entry); entry 223 security/integrity/ima/ima.h struct ima_template_entry **entry, entry 225 security/integrity/ima/ima.h int ima_store_template(struct ima_template_entry *entry, int violation, entry 228 security/integrity/ima/ima.h void ima_free_template_entry(struct ima_template_entry *entry); entry 23 security/integrity/ima/ima_api.c void ima_free_template_entry(struct ima_template_entry *entry) entry 27 security/integrity/ima/ima_api.c for (i = 0; i < entry->template_desc->num_fields; i++) entry 28 security/integrity/ima/ima_api.c kfree(entry->template_data[i].data); entry 30 security/integrity/ima/ima_api.c kfree(entry); entry 37 security/integrity/ima/ima_api.c struct ima_template_entry **entry, entry 48 security/integrity/ima/ima_api.c *entry = kzalloc(struct_size(*entry, template_data, entry 50 security/integrity/ima/ima_api.c if (!*entry) entry 53 security/integrity/ima/ima_api.c (*entry)->template_desc = template_desc; entry 60 security/integrity/ima/ima_api.c &((*entry)->template_data[i])); entry 64 security/integrity/ima/ima_api.c len = (*entry)->template_data[i].len; entry 65 security/integrity/ima/ima_api.c (*entry)->template_data_len += sizeof(len); entry 66 security/integrity/ima/ima_api.c (*entry)->template_data_len += len; entry 70 security/integrity/ima/ima_api.c ima_free_template_entry(*entry); entry 71 security/integrity/ima/ima_api.c *entry = NULL; entry 91 security/integrity/ima/ima_api.c int ima_store_template(struct ima_template_entry *entry, entry 97 security/integrity/ima/ima_api.c char *template_name = entry->template_desc->name; entry 105 security/integrity/ima/ima_api.c int num_fields = entry->template_desc->num_fields; entry 109 security/integrity/ima/ima_api.c result = ima_calc_field_array_hash(&entry->template_data[0], entry 110 security/integrity/ima/ima_api.c entry->template_desc, entry 118 security/integrity/ima/ima_api.c memcpy(entry->digest, hash.hdr.digest, hash.hdr.length); entry 120 security/integrity/ima/ima_api.c entry->pcr = pcr; entry 121 security/integrity/ima/ima_api.c result = ima_add_template_entry(entry, violation, op, inode, filename); entry 136 security/integrity/ima/ima_api.c struct ima_template_entry *entry; entry 148 security/integrity/ima/ima_api.c result = ima_alloc_init_template(&event_data, &entry, NULL); entry 153 security/integrity/ima/ima_api.c result = ima_store_template(entry, violation, inode, entry 156 security/integrity/ima/ima_api.c ima_free_template_entry(entry); entry 303 security/integrity/ima/ima_api.c struct ima_template_entry *entry; entry 321 security/integrity/ima/ima_api.c result = ima_alloc_init_template(&event_data, &entry, template_desc); entry 328 security/integrity/ima/ima_api.c result = ima_store_template(entry, violation, inode, filename, pcr); entry 334 security/integrity/ima/ima_api.c ima_free_template_entry(entry); entry 139 security/integrity/ima/ima_fs.c e = qe->entry; entry 227 security/integrity/ima/ima_fs.c e = qe->entry; entry 46 security/integrity/ima/ima_init.c struct ima_template_entry *entry; entry 71 security/integrity/ima/ima_init.c result = ima_alloc_init_template(&event_data, &entry, NULL); entry 77 security/integrity/ima/ima_init.c result = ima_store_template(entry, violation, NULL, entry 81 security/integrity/ima/ima_init.c ima_free_template_entry(entry); entry 639 security/integrity/ima/ima_main.c struct ima_template_entry *entry = NULL; entry 667 security/integrity/ima/ima_main.c ret = ima_alloc_init_template(&event_data, &entry, template_desc); entry 671 security/integrity/ima/ima_main.c ret = ima_store_template(entry, violation, NULL, buf, pcr); entry 674 security/integrity/ima/ima_main.c ima_free_template_entry(entry); entry 252 security/integrity/ima/ima_policy.c static void ima_lsm_free_rule(struct ima_rule_entry *entry) entry 257 security/integrity/ima/ima_policy.c kfree(entry->lsm[i].rule); entry 258 security/integrity/ima/ima_policy.c kfree(entry->lsm[i].args_p); entry 260 security/integrity/ima/ima_policy.c kfree(entry); entry 263 security/integrity/ima/ima_policy.c static struct ima_rule_entry *ima_lsm_copy_rule(struct ima_rule_entry *entry) entry 276 security/integrity/ima/ima_policy.c memcpy(nentry, entry, sizeof(*nentry)); entry 280 security/integrity/ima/ima_policy.c if (!entry->lsm[i].args_p) entry 283 security/integrity/ima/ima_policy.c nentry->lsm[i].type = entry->lsm[i].type; entry 284 security/integrity/ima/ima_policy.c nentry->lsm[i].args_p = kstrdup(entry->lsm[i].args_p, entry 295 security/integrity/ima/ima_policy.c (char *)entry->lsm[i].args_p); entry 304 security/integrity/ima/ima_policy.c static int ima_lsm_update_rule(struct ima_rule_entry *entry) entry 308 security/integrity/ima/ima_policy.c nentry = ima_lsm_copy_rule(entry); entry 312 security/integrity/ima/ima_policy.c list_replace_rcu(&entry->list, &nentry->list); entry 314 security/integrity/ima/ima_policy.c ima_lsm_free_rule(entry); entry 326 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry, *e; entry 329 security/integrity/ima/ima_policy.c list_for_each_entry_safe(entry, e, &ima_policy_rules, list) { entry 332 security/integrity/ima/ima_policy.c if (entry->lsm[i].args_p) { entry 340 security/integrity/ima/ima_policy.c result = ima_lsm_update_rule(entry); entry 496 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry; entry 503 security/integrity/ima/ima_policy.c list_for_each_entry_rcu(entry, ima_rules, list) { entry 505 security/integrity/ima/ima_policy.c if (!(entry->action & actmask)) entry 508 security/integrity/ima/ima_policy.c if (!ima_match_rules(entry, inode, cred, secid, func, mask)) entry 511 security/integrity/ima/ima_policy.c action |= entry->flags & IMA_ACTION_FLAGS; entry 513 security/integrity/ima/ima_policy.c action |= entry->action & IMA_DO_MASK; entry 514 security/integrity/ima/ima_policy.c if (entry->action & IMA_APPRAISE) { entry 515 security/integrity/ima/ima_policy.c action |= get_subaction(entry, func); entry 522 security/integrity/ima/ima_policy.c if (entry->action & IMA_DO_MASK) entry 523 security/integrity/ima/ima_policy.c actmask &= ~(entry->action | entry->action << 1); entry 525 security/integrity/ima/ima_policy.c actmask &= ~(entry->action | entry->action >> 1); entry 527 security/integrity/ima/ima_policy.c if ((pcr) && (entry->flags & IMA_PCR)) entry 528 security/integrity/ima/ima_policy.c *pcr = entry->pcr; entry 530 security/integrity/ima/ima_policy.c if (template_desc && entry->template) entry 531 security/integrity/ima/ima_policy.c *template_desc = entry->template; entry 549 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry; entry 551 security/integrity/ima/ima_policy.c list_for_each_entry(entry, ima_rules, list) { entry 552 security/integrity/ima/ima_policy.c if (entry->action & IMA_DO_MASK) entry 553 security/integrity/ima/ima_policy.c ima_policy_flag |= entry->action; entry 580 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry; entry 586 security/integrity/ima/ima_policy.c entry = kmemdup(&entries[i], sizeof(*entry), entry 588 security/integrity/ima/ima_policy.c if (!entry) entry 591 security/integrity/ima/ima_policy.c list_add_tail(&entry->list, &ima_policy_rules); entry 601 security/integrity/ima/ima_policy.c static int ima_parse_rule(char *rule, struct ima_rule_entry *entry); entry 809 security/integrity/ima/ima_policy.c static int ima_lsm_rule_init(struct ima_rule_entry *entry, entry 814 security/integrity/ima/ima_policy.c if (entry->lsm[lsm_rule].rule) entry 817 security/integrity/ima/ima_policy.c entry->lsm[lsm_rule].args_p = match_strdup(args); entry 818 security/integrity/ima/ima_policy.c if (!entry->lsm[lsm_rule].args_p) entry 821 security/integrity/ima/ima_policy.c entry->lsm[lsm_rule].type = audit_type; entry 822 security/integrity/ima/ima_policy.c result = security_filter_rule_init(entry->lsm[lsm_rule].type, entry 824 security/integrity/ima/ima_policy.c entry->lsm[lsm_rule].args_p, entry 825 security/integrity/ima/ima_policy.c &entry->lsm[lsm_rule].rule); entry 826 security/integrity/ima/ima_policy.c if (!entry->lsm[lsm_rule].rule) { entry 828 security/integrity/ima/ima_policy.c (char *)entry->lsm[lsm_rule].args_p); entry 831 security/integrity/ima/ima_policy.c kfree(entry->lsm[lsm_rule].args_p); entry 891 security/integrity/ima/ima_policy.c static int ima_parse_rule(char *rule, struct ima_rule_entry *entry) entry 903 security/integrity/ima/ima_policy.c entry->uid = INVALID_UID; entry 904 security/integrity/ima/ima_policy.c entry->fowner = INVALID_UID; entry 905 security/integrity/ima/ima_policy.c entry->uid_op = &uid_eq; entry 906 security/integrity/ima/ima_policy.c entry->fowner_op = &uid_eq; entry 907 security/integrity/ima/ima_policy.c entry->action = UNKNOWN; entry 922 security/integrity/ima/ima_policy.c if (entry->action != UNKNOWN) entry 925 security/integrity/ima/ima_policy.c entry->action = MEASURE; entry 930 security/integrity/ima/ima_policy.c if (entry->action != UNKNOWN) entry 933 security/integrity/ima/ima_policy.c entry->action = DONT_MEASURE; entry 938 security/integrity/ima/ima_policy.c if (entry->action != UNKNOWN) entry 941 security/integrity/ima/ima_policy.c entry->action = APPRAISE; entry 946 security/integrity/ima/ima_policy.c if (entry->action != UNKNOWN) entry 949 security/integrity/ima/ima_policy.c entry->action = DONT_APPRAISE; entry 954 security/integrity/ima/ima_policy.c if (entry->action != UNKNOWN) entry 957 security/integrity/ima/ima_policy.c entry->action = AUDIT; entry 962 security/integrity/ima/ima_policy.c if (entry->action != UNKNOWN) entry 965 security/integrity/ima/ima_policy.c entry->action = HASH; entry 970 security/integrity/ima/ima_policy.c if (entry->action != UNKNOWN) entry 973 security/integrity/ima/ima_policy.c entry->action = DONT_HASH; entry 978 security/integrity/ima/ima_policy.c if (entry->func) entry 982 security/integrity/ima/ima_policy.c entry->func = FILE_CHECK; entry 985 security/integrity/ima/ima_policy.c entry->func = FILE_CHECK; entry 987 security/integrity/ima/ima_policy.c entry->func = MODULE_CHECK; entry 989 security/integrity/ima/ima_policy.c entry->func = FIRMWARE_CHECK; entry 992 security/integrity/ima/ima_policy.c entry->func = MMAP_CHECK; entry 994 security/integrity/ima/ima_policy.c entry->func = BPRM_CHECK; entry 996 security/integrity/ima/ima_policy.c entry->func = CREDS_CHECK; entry 999 security/integrity/ima/ima_policy.c entry->func = KEXEC_KERNEL_CHECK; entry 1002 security/integrity/ima/ima_policy.c entry->func = KEXEC_INITRAMFS_CHECK; entry 1004 security/integrity/ima/ima_policy.c entry->func = POLICY_CHECK; entry 1006 security/integrity/ima/ima_policy.c entry->func = KEXEC_CMDLINE; entry 1010 security/integrity/ima/ima_policy.c entry->flags |= IMA_FUNC; entry 1015 security/integrity/ima/ima_policy.c if (entry->mask) entry 1023 security/integrity/ima/ima_policy.c entry->mask = MAY_EXEC; entry 1025 security/integrity/ima/ima_policy.c entry->mask = MAY_WRITE; entry 1027 security/integrity/ima/ima_policy.c entry->mask = MAY_READ; entry 1029 security/integrity/ima/ima_policy.c entry->mask = MAY_APPEND; entry 1033 security/integrity/ima/ima_policy.c entry->flags |= (*args[0].from == '^') entry 1039 security/integrity/ima/ima_policy.c if (entry->fsmagic) { entry 1044 security/integrity/ima/ima_policy.c result = kstrtoul(args[0].from, 16, &entry->fsmagic); entry 1046 security/integrity/ima/ima_policy.c entry->flags |= IMA_FSMAGIC; entry 1051 security/integrity/ima/ima_policy.c entry->fsname = kstrdup(args[0].from, GFP_KERNEL); entry 1052 security/integrity/ima/ima_policy.c if (!entry->fsname) { entry 1057 security/integrity/ima/ima_policy.c entry->flags |= IMA_FSNAME; entry 1062 security/integrity/ima/ima_policy.c if (!uuid_is_null(&entry->fsuuid)) { entry 1067 security/integrity/ima/ima_policy.c result = uuid_parse(args[0].from, &entry->fsuuid); entry 1069 security/integrity/ima/ima_policy.c entry->flags |= IMA_FSUUID; entry 1073 security/integrity/ima/ima_policy.c entry->uid_op = &uid_gt; entry 1078 security/integrity/ima/ima_policy.c entry->uid_op = &uid_lt; entry 1087 security/integrity/ima/ima_policy.c args[0].from, entry->uid_op); entry 1089 security/integrity/ima/ima_policy.c if (uid_valid(entry->uid)) { entry 1096 security/integrity/ima/ima_policy.c entry->uid = make_kuid(current_user_ns(), entry 1098 security/integrity/ima/ima_policy.c if (!uid_valid(entry->uid) || entry 1102 security/integrity/ima/ima_policy.c entry->flags |= uid_token entry 1107 security/integrity/ima/ima_policy.c entry->fowner_op = &uid_gt; entry 1111 security/integrity/ima/ima_policy.c entry->fowner_op = &uid_lt; entry 1115 security/integrity/ima/ima_policy.c entry->fowner_op); entry 1117 security/integrity/ima/ima_policy.c if (uid_valid(entry->fowner)) { entry 1124 security/integrity/ima/ima_policy.c entry->fowner = make_kuid(current_user_ns(), (uid_t)lnum); entry 1125 security/integrity/ima/ima_policy.c if (!uid_valid(entry->fowner) || (((uid_t)lnum) != lnum)) entry 1128 security/integrity/ima/ima_policy.c entry->flags |= IMA_FOWNER; entry 1133 security/integrity/ima/ima_policy.c result = ima_lsm_rule_init(entry, args, entry 1139 security/integrity/ima/ima_policy.c result = ima_lsm_rule_init(entry, args, entry 1145 security/integrity/ima/ima_policy.c result = ima_lsm_rule_init(entry, args, entry 1151 security/integrity/ima/ima_policy.c result = ima_lsm_rule_init(entry, args, entry 1157 security/integrity/ima/ima_policy.c result = ima_lsm_rule_init(entry, args, entry 1163 security/integrity/ima/ima_policy.c result = ima_lsm_rule_init(entry, args, entry 1168 security/integrity/ima/ima_policy.c if (entry->action != APPRAISE) { entry 1175 security/integrity/ima/ima_policy.c entry->flags |= IMA_DIGSIG_REQUIRED; entry 1176 security/integrity/ima/ima_policy.c else if (ima_hook_supports_modsig(entry->func) && entry 1178 security/integrity/ima/ima_policy.c entry->flags |= IMA_DIGSIG_REQUIRED | entry 1184 security/integrity/ima/ima_policy.c entry->flags |= IMA_PERMIT_DIRECTIO; entry 1187 security/integrity/ima/ima_policy.c if (entry->action != MEASURE) { entry 1193 security/integrity/ima/ima_policy.c result = kstrtoint(args[0].from, 10, &entry->pcr); entry 1194 security/integrity/ima/ima_policy.c if (result || INVALID_PCR(entry->pcr)) entry 1197 security/integrity/ima/ima_policy.c entry->flags |= IMA_PCR; entry 1202 security/integrity/ima/ima_policy.c if (entry->action != MEASURE) { entry 1207 security/integrity/ima/ima_policy.c if (!template_desc || entry->template) { entry 1220 security/integrity/ima/ima_policy.c entry->template = template_desc; entry 1228 security/integrity/ima/ima_policy.c if (!result && (entry->action == UNKNOWN)) entry 1230 security/integrity/ima/ima_policy.c else if (entry->action == APPRAISE) entry 1231 security/integrity/ima/ima_policy.c temp_ima_appraise |= ima_appraise_flag(entry->func); entry 1233 security/integrity/ima/ima_policy.c if (!result && entry->flags & IMA_MODSIG_ALLOWED) { entry 1234 security/integrity/ima/ima_policy.c template_desc = entry->template ? entry->template : entry 1255 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry; entry 1266 security/integrity/ima/ima_policy.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 1267 security/integrity/ima/ima_policy.c if (!entry) { entry 1273 security/integrity/ima/ima_policy.c INIT_LIST_HEAD(&entry->list); entry 1275 security/integrity/ima/ima_policy.c result = ima_parse_rule(p, entry); entry 1277 security/integrity/ima/ima_policy.c kfree(entry); entry 1284 security/integrity/ima/ima_policy.c list_add_tail(&entry->list, &ima_temp_rules); entry 1297 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry, *tmp; entry 1301 security/integrity/ima/ima_policy.c list_for_each_entry_safe(entry, tmp, &ima_temp_rules, list) { entry 1303 security/integrity/ima/ima_policy.c kfree(entry->lsm[i].args_p); entry 1305 security/integrity/ima/ima_policy.c list_del(&entry->list); entry 1306 security/integrity/ima/ima_policy.c kfree(entry); entry 1331 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry; entry 1334 security/integrity/ima/ima_policy.c list_for_each_entry_rcu(entry, ima_rules, list) { entry 1337 security/integrity/ima/ima_policy.c return entry; entry 1346 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry = v; entry 1349 security/integrity/ima/ima_policy.c entry = list_entry_rcu(entry->list.next, struct ima_rule_entry, list); entry 1353 security/integrity/ima/ima_policy.c return (&entry->list == ima_rules) ? NULL : entry; entry 1376 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry = v; entry 1383 security/integrity/ima/ima_policy.c if (entry->action & MEASURE) entry 1385 security/integrity/ima/ima_policy.c if (entry->action & DONT_MEASURE) entry 1387 security/integrity/ima/ima_policy.c if (entry->action & APPRAISE) entry 1389 security/integrity/ima/ima_policy.c if (entry->action & DONT_APPRAISE) entry 1391 security/integrity/ima/ima_policy.c if (entry->action & AUDIT) entry 1393 security/integrity/ima/ima_policy.c if (entry->action & HASH) entry 1395 security/integrity/ima/ima_policy.c if (entry->action & DONT_HASH) entry 1400 security/integrity/ima/ima_policy.c if (entry->flags & IMA_FUNC) entry 1401 security/integrity/ima/ima_policy.c policy_func_show(m, entry->func); entry 1403 security/integrity/ima/ima_policy.c if ((entry->flags & IMA_MASK) || (entry->flags & IMA_INMASK)) { entry 1404 security/integrity/ima/ima_policy.c if (entry->flags & IMA_MASK) entry 1406 security/integrity/ima/ima_policy.c if (entry->mask & MAY_EXEC) entry 1408 security/integrity/ima/ima_policy.c if (entry->mask & MAY_WRITE) entry 1410 security/integrity/ima/ima_policy.c if (entry->mask & MAY_READ) entry 1412 security/integrity/ima/ima_policy.c if (entry->mask & MAY_APPEND) entry 1417 security/integrity/ima/ima_policy.c if (entry->flags & IMA_FSMAGIC) { entry 1418 security/integrity/ima/ima_policy.c snprintf(tbuf, sizeof(tbuf), "0x%lx", entry->fsmagic); entry 1423 security/integrity/ima/ima_policy.c if (entry->flags & IMA_FSNAME) { entry 1424 security/integrity/ima/ima_policy.c snprintf(tbuf, sizeof(tbuf), "%s", entry->fsname); entry 1429 security/integrity/ima/ima_policy.c if (entry->flags & IMA_PCR) { entry 1430 security/integrity/ima/ima_policy.c snprintf(tbuf, sizeof(tbuf), "%d", entry->pcr); entry 1435 security/integrity/ima/ima_policy.c if (entry->flags & IMA_FSUUID) { entry 1436 security/integrity/ima/ima_policy.c seq_printf(m, "fsuuid=%pU", &entry->fsuuid); entry 1440 security/integrity/ima/ima_policy.c if (entry->flags & IMA_UID) { entry 1441 security/integrity/ima/ima_policy.c snprintf(tbuf, sizeof(tbuf), "%d", __kuid_val(entry->uid)); entry 1442 security/integrity/ima/ima_policy.c if (entry->uid_op == &uid_gt) entry 1444 security/integrity/ima/ima_policy.c else if (entry->uid_op == &uid_lt) entry 1451 security/integrity/ima/ima_policy.c if (entry->flags & IMA_EUID) { entry 1452 security/integrity/ima/ima_policy.c snprintf(tbuf, sizeof(tbuf), "%d", __kuid_val(entry->uid)); entry 1453 security/integrity/ima/ima_policy.c if (entry->uid_op == &uid_gt) entry 1455 security/integrity/ima/ima_policy.c else if (entry->uid_op == &uid_lt) entry 1462 security/integrity/ima/ima_policy.c if (entry->flags & IMA_FOWNER) { entry 1463 security/integrity/ima/ima_policy.c snprintf(tbuf, sizeof(tbuf), "%d", __kuid_val(entry->fowner)); entry 1464 security/integrity/ima/ima_policy.c if (entry->fowner_op == &uid_gt) entry 1466 security/integrity/ima/ima_policy.c else if (entry->fowner_op == &uid_lt) entry 1474 security/integrity/ima/ima_policy.c if (entry->lsm[i].rule) { entry 1478 security/integrity/ima/ima_policy.c (char *)entry->lsm[i].args_p); entry 1482 security/integrity/ima/ima_policy.c (char *)entry->lsm[i].args_p); entry 1486 security/integrity/ima/ima_policy.c (char *)entry->lsm[i].args_p); entry 1490 security/integrity/ima/ima_policy.c (char *)entry->lsm[i].args_p); entry 1494 security/integrity/ima/ima_policy.c (char *)entry->lsm[i].args_p); entry 1498 security/integrity/ima/ima_policy.c (char *)entry->lsm[i].args_p); entry 1503 security/integrity/ima/ima_policy.c if (entry->template) entry 1504 security/integrity/ima/ima_policy.c seq_printf(m, "template=%s ", entry->template->name); entry 1505 security/integrity/ima/ima_policy.c if (entry->flags & IMA_DIGSIG_REQUIRED) { entry 1506 security/integrity/ima/ima_policy.c if (entry->flags & IMA_MODSIG_ALLOWED) entry 1511 security/integrity/ima/ima_policy.c if (entry->flags & IMA_PERMIT_DIRECTIO) entry 1528 security/integrity/ima/ima_policy.c struct ima_rule_entry *entry; entry 1538 security/integrity/ima/ima_policy.c list_for_each_entry_rcu(entry, ima_rules, list) { entry 1539 security/integrity/ima/ima_policy.c if (entry->action != APPRAISE) entry 1546 security/integrity/ima/ima_policy.c if (entry->func && entry->func != func) entry 1553 security/integrity/ima/ima_policy.c if (entry->flags & IMA_DIGSIG_REQUIRED) entry 60 security/integrity/ima/ima_queue.c rc = memcmp(qe->entry->digest, digest_value, TPM_DIGEST_SIZE); entry 61 security/integrity/ima/ima_queue.c if ((rc == 0) && (qe->entry->pcr == pcr)) { entry 75 security/integrity/ima/ima_queue.c static int get_binary_runtime_size(struct ima_template_entry *entry) entry 80 security/integrity/ima/ima_queue.c size += sizeof(entry->digest); entry 82 security/integrity/ima/ima_queue.c size += strlen(entry->template_desc->name); entry 83 security/integrity/ima/ima_queue.c size += sizeof(entry->template_data_len); entry 84 security/integrity/ima/ima_queue.c size += entry->template_data_len; entry 94 security/integrity/ima/ima_queue.c static int ima_add_digest_entry(struct ima_template_entry *entry, entry 105 security/integrity/ima/ima_queue.c qe->entry = entry; entry 112 security/integrity/ima/ima_queue.c key = ima_hash_key(entry->digest); entry 119 security/integrity/ima/ima_queue.c size = get_binary_runtime_size(entry); entry 164 security/integrity/ima/ima_queue.c int ima_add_template_entry(struct ima_template_entry *entry, int violation, entry 176 security/integrity/ima/ima_queue.c memcpy(digest, entry->digest, sizeof(digest)); entry 177 security/integrity/ima/ima_queue.c if (ima_lookup_digest_entry(digest, entry->pcr)) { entry 184 security/integrity/ima/ima_queue.c result = ima_add_digest_entry(entry, 1); entry 194 security/integrity/ima/ima_queue.c tpmresult = ima_pcr_extend(digest, entry->pcr); entry 208 security/integrity/ima/ima_queue.c int ima_restore_measurement_entry(struct ima_template_entry *entry) entry 213 security/integrity/ima/ima_queue.c result = ima_add_digest_entry(entry, 0); entry 304 security/integrity/ima/ima_template.c struct ima_template_entry **entry) entry 309 security/integrity/ima/ima_template.c *entry = kzalloc(struct_size(*entry, template_data, entry 311 security/integrity/ima/ima_template.c if (!*entry) entry 316 security/integrity/ima/ima_template.c (*entry)->template_data, NULL, NULL, entry 319 security/integrity/ima/ima_template.c kfree(*entry); entry 323 security/integrity/ima/ima_template.c (*entry)->template_desc = template_desc; entry 325 security/integrity/ima/ima_template.c struct ima_field_data *field_data = &(*entry)->template_data[i]; entry 328 security/integrity/ima/ima_template.c (*entry)->template_data[i].data = entry 330 security/integrity/ima/ima_template.c if (!(*entry)->template_data[i].data) { entry 334 security/integrity/ima/ima_template.c memcpy((*entry)->template_data[i].data, data, field_data->len); entry 335 security/integrity/ima/ima_template.c (*entry)->template_data_len += sizeof(field_data->len); entry 336 security/integrity/ima/ima_template.c (*entry)->template_data_len += field_data->len; entry 340 security/integrity/ima/ima_template.c ima_free_template_entry(*entry); entry 341 security/integrity/ima/ima_template.c *entry = NULL; entry 360 security/integrity/ima/ima_template.c struct ima_template_entry *entry; entry 446 security/integrity/ima/ima_template.c &entry); entry 450 security/integrity/ima/ima_template.c memcpy(entry->digest, hdr[HDR_DIGEST].data, entry 452 security/integrity/ima/ima_template.c entry->pcr = !ima_canonical_fmt ? *(hdr[HDR_PCR].data) : entry 454 security/integrity/ima/ima_template.c ret = ima_restore_measurement_entry(entry); entry 42 security/selinux/ss/sidtab.c struct sidtab_isid_entry *entry; entry 48 security/selinux/ss/sidtab.c entry = &s->isids[sid - 1]; entry 50 security/selinux/ss/sidtab.c rc = context_cpy(&entry->context, context); entry 54 security/selinux/ss/sidtab.c entry->set = 1; entry 93 security/selinux/ss/sidtab.c union sidtab_entry_inner *entry; entry 105 security/selinux/ss/sidtab.c entry = &s->roots[level]; entry 110 security/selinux/ss/sidtab.c entry = &entry->ptr_inner->entries[leaf_index >> capacity_shift]; entry 113 security/selinux/ss/sidtab.c if (!entry->ptr_inner) { entry 115 security/selinux/ss/sidtab.c entry->ptr_inner = kzalloc(SIDTAB_NODE_ALLOC_SIZE, entry 117 security/selinux/ss/sidtab.c if (!entry->ptr_inner) entry 121 security/selinux/ss/sidtab.c if (!entry->ptr_leaf) { entry 123 security/selinux/ss/sidtab.c entry->ptr_leaf = kzalloc(SIDTAB_NODE_ALLOC_SIZE, entry 125 security/selinux/ss/sidtab.c if (!entry->ptr_leaf) entry 128 security/selinux/ss/sidtab.c return &entry->ptr_leaf->entries[index % SIDTAB_LEAF_ENTRIES].context; entry 173 security/selinux/ss/sidtab.c static int sidtab_find_context(union sidtab_entry_inner entry, entry 181 security/selinux/ss/sidtab.c struct sidtab_node_inner *node = entry.ptr_inner; entry 193 security/selinux/ss/sidtab.c struct sidtab_node_leaf *node = entry.ptr_leaf; entry 344 security/selinux/ss/sidtab.c struct sidtab_isid_entry *entry = &s->isids[i]; entry 346 security/selinux/ss/sidtab.c if (entry->set && context_cmp(context, &entry->context)) { entry 456 security/selinux/ss/sidtab.c static void sidtab_destroy_tree(union sidtab_entry_inner entry, u32 level) entry 461 security/selinux/ss/sidtab.c struct sidtab_node_inner *node = entry.ptr_inner; entry 470 security/selinux/ss/sidtab.c struct sidtab_node_leaf *node = entry.ptr_leaf; entry 369 security/tomoyo/audit.c struct tomoyo_log *entry; entry 378 security/tomoyo/audit.c entry = kzalloc(sizeof(*entry), GFP_NOFS); entry 379 security/tomoyo/audit.c if (!entry) { entry 383 security/tomoyo/audit.c entry->log = buf; entry 389 security/tomoyo/audit.c entry->size = len + tomoyo_round2(sizeof(*entry)); entry 392 security/tomoyo/audit.c tomoyo_memory_used[TOMOYO_MEMORY_AUDIT] + entry->size >= entry 396 security/tomoyo/audit.c tomoyo_memory_used[TOMOYO_MEMORY_AUDIT] += entry->size; entry 397 security/tomoyo/audit.c list_add_tail(&entry->list, &tomoyo_log); entry 403 security/tomoyo/audit.c kfree(entry); entry 494 security/tomoyo/common.c struct tomoyo_profile *entry; entry 501 security/tomoyo/common.c entry = kzalloc(sizeof(*entry), GFP_NOFS); entry 505 security/tomoyo/common.c if (!ptr && tomoyo_memory_ok(entry)) { entry 506 security/tomoyo/common.c ptr = entry; entry 518 security/tomoyo/common.c entry = NULL; entry 522 security/tomoyo/common.c kfree(entry); entry 2065 security/tomoyo/common.c struct tomoyo_query entry = { }; entry 2097 security/tomoyo/common.c entry.query = tomoyo_init_log(r, len, fmt, args); entry 2099 security/tomoyo/common.c if (!entry.query) entry 2101 security/tomoyo/common.c entry.query_len = strlen(entry.query) + 1; entry 2103 security/tomoyo/common.c tomoyo_add_entry(r->domain, entry.query); entry 2106 security/tomoyo/common.c len = tomoyo_round2(entry.query_len); entry 2107 security/tomoyo/common.c entry.domain = r->domain; entry 2114 security/tomoyo/common.c entry.serial = tomoyo_serial++; entry 2115 security/tomoyo/common.c entry.retry = r->retry; entry 2117 security/tomoyo/common.c list_add_tail(&entry.list, &tomoyo_query_list); entry 2123 security/tomoyo/common.c while (entry.timer < 10) { entry 2126 security/tomoyo/common.c (tomoyo_answer_wait, entry.answer || entry 2129 security/tomoyo/common.c entry.timer++; entry 2132 security/tomoyo/common.c list_del(&entry.list); entry 2135 security/tomoyo/common.c switch (entry.answer) { entry 2149 security/tomoyo/common.c kfree(entry.query); entry 515 security/tomoyo/common.h struct tomoyo_path_info entry; entry 1182 security/tomoyo/common.h container_of(name, typeof(*ptr), entry); entry 407 security/tomoyo/condition.c (struct tomoyo_condition *entry) entry 419 security/tomoyo/condition.c if (!tomoyo_same_condition(ptr, entry) || entry 428 security/tomoyo/condition.c if (tomoyo_memory_ok(entry)) { entry 429 security/tomoyo/condition.c atomic_set(&entry->head.users, 1); entry 430 security/tomoyo/condition.c list_add(&entry->head.list, &tomoyo_condition_list); entry 439 security/tomoyo/condition.c tomoyo_del_condition(&entry->head.list); entry 440 security/tomoyo/condition.c kfree(entry); entry 441 security/tomoyo/condition.c entry = ptr; entry 443 security/tomoyo/condition.c return entry; entry 497 security/tomoyo/condition.c struct tomoyo_condition *entry = NULL; entry 551 security/tomoyo/condition.c if (entry) { entry 553 security/tomoyo/condition.c entry->grant_log != TOMOYO_GRANTLOG_AUTO) entry 556 security/tomoyo/condition.c entry->grant_log = TOMOYO_GRANTLOG_YES; entry 558 security/tomoyo/condition.c entry->grant_log = TOMOYO_GRANTLOG_NO; entry 658 security/tomoyo/condition.c if (entry) { entry 661 security/tomoyo/condition.c return tomoyo_commit_condition(entry); entry 663 security/tomoyo/condition.c e.size = sizeof(*entry) entry 669 security/tomoyo/condition.c entry = kzalloc(e.size, GFP_NOFS); entry 670 security/tomoyo/condition.c if (!entry) entry 672 security/tomoyo/condition.c *entry = e; entry 674 security/tomoyo/condition.c condp = (struct tomoyo_condition_element *) (entry + 1); entry 697 security/tomoyo/condition.c if (entry) { entry 698 security/tomoyo/condition.c tomoyo_del_condition(&entry->head.list); entry 699 security/tomoyo/condition.c kfree(entry); entry 39 security/tomoyo/domain.c struct tomoyo_acl_head *entry; entry 44 security/tomoyo/domain.c list_for_each_entry_rcu(entry, list, list, entry 46 security/tomoyo/domain.c if (entry->is_deleted == TOMOYO_GC_IN_PROGRESS) entry 48 security/tomoyo/domain.c if (!check_duplicate(entry, new_entry)) entry 50 security/tomoyo/domain.c entry->is_deleted = param->is_delete; entry 55 security/tomoyo/domain.c entry = tomoyo_commit_ok(new_entry, size); entry 56 security/tomoyo/domain.c if (entry) { entry 57 security/tomoyo/domain.c list_add_tail_rcu(&entry->list, list); entry 104 security/tomoyo/domain.c struct tomoyo_acl_info *entry; entry 123 security/tomoyo/domain.c list_for_each_entry_rcu(entry, list, list, entry 125 security/tomoyo/domain.c if (entry->is_deleted == TOMOYO_GC_IN_PROGRESS) entry 127 security/tomoyo/domain.c if (!tomoyo_same_acl_head(entry, new_entry) || entry 128 security/tomoyo/domain.c !check_duplicate(entry, new_entry)) entry 131 security/tomoyo/domain.c entry->is_deleted = merge_duplicate(entry, new_entry, entry 134 security/tomoyo/domain.c entry->is_deleted = is_delete; entry 139 security/tomoyo/domain.c entry = tomoyo_commit_ok(new_entry, size); entry 140 security/tomoyo/domain.c if (entry) { entry 141 security/tomoyo/domain.c list_add_tail_rcu(&entry->list, list); entry 465 security/tomoyo/domain.c struct tomoyo_policy_namespace *entry; entry 476 security/tomoyo/domain.c entry = kzalloc(sizeof(*entry) + len + 1, GFP_NOFS); entry 477 security/tomoyo/domain.c if (!entry) entry 482 security/tomoyo/domain.c if (!ptr && tomoyo_memory_ok(entry)) { entry 483 security/tomoyo/domain.c char *name = (char *) (entry + 1); entry 485 security/tomoyo/domain.c ptr = entry; entry 488 security/tomoyo/domain.c entry->name = name; entry 489 security/tomoyo/domain.c tomoyo_init_policy_namespace(entry); entry 490 security/tomoyo/domain.c entry = NULL; entry 494 security/tomoyo/domain.c kfree(entry); entry 528 security/tomoyo/domain.c struct tomoyo_domain_info *entry = tomoyo_find_domain(domainname); entry 531 security/tomoyo/domain.c if (entry) { entry 540 security/tomoyo/domain.c !entry->ns->profile_ptr[entry->profile]) entry 543 security/tomoyo/domain.c return entry; entry 576 security/tomoyo/domain.c entry = tomoyo_find_domain(domainname); entry 577 security/tomoyo/domain.c if (!entry) { entry 578 security/tomoyo/domain.c entry = tomoyo_commit_ok(&e, sizeof(e)); entry 579 security/tomoyo/domain.c if (entry) { entry 580 security/tomoyo/domain.c INIT_LIST_HEAD(&entry->acl_info_list); entry 581 security/tomoyo/domain.c list_add_tail_rcu(&entry->list, &tomoyo_domain_list); entry 588 security/tomoyo/domain.c if (entry && transit) { entry 593 security/tomoyo/domain.c tomoyo_init_request_info(&r, entry, entry 597 security/tomoyo/domain.c entry->profile); entry 599 security/tomoyo/domain.c if (test_bit(i, entry->group)) entry 605 security/tomoyo/domain.c return entry; entry 165 security/tomoyo/gc.c struct tomoyo_path_acl *entry entry 166 security/tomoyo/gc.c = container_of(acl, typeof(*entry), head); entry 167 security/tomoyo/gc.c tomoyo_put_name_union(&entry->name); entry 172 security/tomoyo/gc.c struct tomoyo_path2_acl *entry entry 173 security/tomoyo/gc.c = container_of(acl, typeof(*entry), head); entry 174 security/tomoyo/gc.c tomoyo_put_name_union(&entry->name1); entry 175 security/tomoyo/gc.c tomoyo_put_name_union(&entry->name2); entry 180 security/tomoyo/gc.c struct tomoyo_path_number_acl *entry entry 181 security/tomoyo/gc.c = container_of(acl, typeof(*entry), head); entry 182 security/tomoyo/gc.c tomoyo_put_name_union(&entry->name); entry 183 security/tomoyo/gc.c tomoyo_put_number_union(&entry->number); entry 188 security/tomoyo/gc.c struct tomoyo_mkdev_acl *entry entry 189 security/tomoyo/gc.c = container_of(acl, typeof(*entry), head); entry 190 security/tomoyo/gc.c tomoyo_put_name_union(&entry->name); entry 191 security/tomoyo/gc.c tomoyo_put_number_union(&entry->mode); entry 192 security/tomoyo/gc.c tomoyo_put_number_union(&entry->major); entry 193 security/tomoyo/gc.c tomoyo_put_number_union(&entry->minor); entry 198 security/tomoyo/gc.c struct tomoyo_mount_acl *entry entry 199 security/tomoyo/gc.c = container_of(acl, typeof(*entry), head); entry 200 security/tomoyo/gc.c tomoyo_put_name_union(&entry->dev_name); entry 201 security/tomoyo/gc.c tomoyo_put_name_union(&entry->dir_name); entry 202 security/tomoyo/gc.c tomoyo_put_name_union(&entry->fs_type); entry 203 security/tomoyo/gc.c tomoyo_put_number_union(&entry->flags); entry 208 security/tomoyo/gc.c struct tomoyo_env_acl *entry = entry 209 security/tomoyo/gc.c container_of(acl, typeof(*entry), head); entry 211 security/tomoyo/gc.c tomoyo_put_name(entry->env); entry 216 security/tomoyo/gc.c struct tomoyo_inet_acl *entry = entry 217 security/tomoyo/gc.c container_of(acl, typeof(*entry), head); entry 219 security/tomoyo/gc.c tomoyo_put_group(entry->address.group); entry 220 security/tomoyo/gc.c tomoyo_put_number_union(&entry->port); entry 225 security/tomoyo/gc.c struct tomoyo_unix_acl *entry = entry 226 security/tomoyo/gc.c container_of(acl, typeof(*entry), head); entry 228 security/tomoyo/gc.c tomoyo_put_name_union(&entry->name); entry 233 security/tomoyo/gc.c struct tomoyo_task_acl *entry = entry 234 security/tomoyo/gc.c container_of(acl, typeof(*entry), head); entry 236 security/tomoyo/gc.c tomoyo_put_name(entry->domainname); entry 439 security/tomoyo/gc.c head.list)->entry.name)) entry 121 security/tomoyo/memory.c struct tomoyo_group *entry = tomoyo_commit_ok(&e, sizeof(e)); entry 123 security/tomoyo/memory.c if (entry) { entry 124 security/tomoyo/memory.c INIT_LIST_HEAD(&entry->member_list); entry 125 security/tomoyo/memory.c atomic_set(&entry->head.users, 1); entry 126 security/tomoyo/memory.c list_add_tail_rcu(&entry->head.list, list); entry 127 security/tomoyo/memory.c group = entry; entry 167 security/tomoyo/memory.c if (hash != ptr->entry.hash || strcmp(name, ptr->entry.name) || entry 175 security/tomoyo/memory.c ptr->entry.name = ((char *) ptr) + sizeof(*ptr); entry 176 security/tomoyo/memory.c memmove((char *) ptr->entry.name, name, len); entry 178 security/tomoyo/memory.c tomoyo_fill_path_info(&ptr->entry); entry 186 security/tomoyo/memory.c return ptr ? &ptr->entry : NULL; entry 1019 sound/core/compress_offload.c static void snd_compress_proc_info_read(struct snd_info_entry *entry, entry 1022 sound/core/compress_offload.c struct snd_compr *compr = (struct snd_compr *)entry->private_data; entry 1034 sound/core/compress_offload.c struct snd_info_entry *entry; entry 1038 sound/core/compress_offload.c entry = snd_info_create_card_entry(compr->card, name, entry 1040 sound/core/compress_offload.c if (!entry) entry 1042 sound/core/compress_offload.c entry->mode = S_IFDIR | 0555; entry 1043 sound/core/compress_offload.c compr->proc_root = entry; entry 1045 sound/core/compress_offload.c entry = snd_info_create_card_entry(compr->card, "info", entry 1047 sound/core/compress_offload.c if (entry) entry 1048 sound/core/compress_offload.c snd_info_set_text_ops(entry, compr, entry 1050 sound/core/compress_offload.c compr->proc_info_entry = entry; entry 478 sound/core/hwdep.c static void snd_hwdep_proc_read(struct snd_info_entry *entry, entry 494 sound/core/hwdep.c struct snd_info_entry *entry; entry 496 sound/core/hwdep.c if ((entry = snd_info_create_module_entry(THIS_MODULE, "hwdep", NULL)) != NULL) { entry 497 sound/core/hwdep.c entry->c.text.read = snd_hwdep_proc_read; entry 498 sound/core/hwdep.c if (snd_info_register(entry) < 0) { entry 499 sound/core/hwdep.c snd_info_free_entry(entry); entry 500 sound/core/hwdep.c entry = NULL; entry 503 sound/core/hwdep.c snd_hwdep_proc_entry = entry; entry 55 sound/core/info.c struct snd_info_entry *entry; entry 60 sound/core/info.c static void snd_info_disconnect(struct snd_info_entry *entry); entry 74 sound/core/info.c static int alloc_info_private(struct snd_info_entry *entry, entry 79 sound/core/info.c if (!entry || !entry->p) entry 81 sound/core/info.c if (!try_module_get(entry->module)) entry 85 sound/core/info.c module_put(entry->module); entry 88 sound/core/info.c data->entry = entry; entry 108 sound/core/info.c struct snd_info_entry *entry; entry 112 sound/core/info.c entry = data->entry; entry 113 sound/core/info.c mutex_lock(&entry->access); entry 114 sound/core/info.c if (entry->c.ops->llseek) { entry 115 sound/core/info.c offset = entry->c.ops->llseek(entry, entry 121 sound/core/info.c size = entry->size; entry 143 sound/core/info.c mutex_unlock(&entry->access); entry 151 sound/core/info.c struct snd_info_entry *entry = data->entry; entry 158 sound/core/info.c if (pos >= entry->size) entry 160 sound/core/info.c size = entry->size - pos; entry 162 sound/core/info.c size = entry->c.ops->read(entry, data->file_private_data, entry 173 sound/core/info.c struct snd_info_entry *entry = data->entry; entry 181 sound/core/info.c size_t maxsize = entry->size - pos; entry 183 sound/core/info.c size = entry->c.ops->write(entry, data->file_private_data, entry 194 sound/core/info.c struct snd_info_entry *entry = data->entry; entry 197 sound/core/info.c if (entry->c.ops->poll) entry 198 sound/core/info.c return entry->c.ops->poll(entry, entry 201 sound/core/info.c if (entry->c.ops->read) entry 203 sound/core/info.c if (entry->c.ops->write) entry 212 sound/core/info.c struct snd_info_entry *entry = data->entry; entry 214 sound/core/info.c if (!entry->c.ops->ioctl) entry 216 sound/core/info.c return entry->c.ops->ioctl(entry, data->file_private_data, entry 224 sound/core/info.c struct snd_info_entry *entry; entry 229 sound/core/info.c entry = data->entry; entry 230 sound/core/info.c if (!entry->c.ops->mmap) entry 232 sound/core/info.c return entry->c.ops->mmap(entry, data->file_private_data, entry 238 sound/core/info.c struct snd_info_entry *entry = PDE_DATA(inode); entry 243 sound/core/info.c err = alloc_info_private(entry, &data); entry 248 sound/core/info.c if (((mode == O_RDONLY || mode == O_RDWR) && !entry->c.ops->read) || entry 249 sound/core/info.c ((mode == O_WRONLY || mode == O_RDWR) && !entry->c.ops->write)) { entry 254 sound/core/info.c if (entry->c.ops->open) { entry 255 sound/core/info.c err = entry->c.ops->open(entry, mode, &data->file_private_data); entry 266 sound/core/info.c module_put(entry->module); entry 275 sound/core/info.c struct snd_info_entry *entry = data->entry; entry 277 sound/core/info.c if (entry->c.ops->release) entry 278 sound/core/info.c entry->c.ops->release(entry, file->f_flags & O_ACCMODE, entry 280 sound/core/info.c module_put(entry->module); entry 307 sound/core/info.c struct snd_info_entry *entry = data->entry; entry 313 sound/core/info.c if (!entry->c.text.write) entry 322 sound/core/info.c mutex_lock(&entry->access); entry 347 sound/core/info.c mutex_unlock(&entry->access); entry 357 sound/core/info.c struct snd_info_entry *entry = data->entry; entry 359 sound/core/info.c if (!entry->c.text.read) { entry 363 sound/core/info.c entry->c.text.read(entry, data->rbuffer); entry 370 sound/core/info.c struct snd_info_entry *entry = PDE_DATA(inode); entry 375 sound/core/info.c err = alloc_info_private(entry, &data); entry 384 sound/core/info.c if (entry->size) entry 386 sound/core/info.c entry->size); entry 397 sound/core/info.c module_put(entry->module); entry 407 sound/core/info.c struct snd_info_entry *entry = data->entry; entry 409 sound/core/info.c if (data->wbuffer && entry->c.text.write) entry 410 sound/core/info.c entry->c.text.write(entry, data->wbuffer); entry 419 sound/core/info.c module_put(entry->module); entry 437 sound/core/info.c struct snd_info_entry *entry; entry 439 sound/core/info.c entry = snd_info_create_module_entry(mod, name, NULL); entry 440 sound/core/info.c if (!entry) entry 442 sound/core/info.c entry->mode = S_IFDIR | 0555; entry 443 sound/core/info.c if (snd_info_register(entry) < 0) { entry 444 sound/core/info.c snd_info_free_entry(entry); entry 447 sound/core/info.c return entry; entry 492 sound/core/info.c static void snd_card_id_read(struct snd_info_entry *entry, entry 495 sound/core/info.c struct snd_card *card = entry->private_data; entry 507 sound/core/info.c struct snd_info_entry *entry; entry 513 sound/core/info.c entry = create_subdir(card->module, str); entry 514 sound/core/info.c if (!entry) entry 516 sound/core/info.c card->proc_root = entry; entry 685 sound/core/info.c struct snd_info_entry *entry; entry 686 sound/core/info.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 687 sound/core/info.c if (entry == NULL) entry 689 sound/core/info.c entry->name = kstrdup(name, GFP_KERNEL); entry 690 sound/core/info.c if (entry->name == NULL) { entry 691 sound/core/info.c kfree(entry); entry 694 sound/core/info.c entry->mode = S_IFREG | 0444; entry 695 sound/core/info.c entry->content = SNDRV_INFO_CONTENT_TEXT; entry 696 sound/core/info.c mutex_init(&entry->access); entry 697 sound/core/info.c INIT_LIST_HEAD(&entry->children); entry 698 sound/core/info.c INIT_LIST_HEAD(&entry->list); entry 699 sound/core/info.c entry->parent = parent; entry 700 sound/core/info.c entry->module = module; entry 703 sound/core/info.c list_add_tail(&entry->list, &parent->children); entry 706 sound/core/info.c return entry; entry 749 sound/core/info.c static void snd_info_disconnect(struct snd_info_entry *entry) entry 753 sound/core/info.c if (!entry->p) entry 755 sound/core/info.c list_for_each_entry(p, &entry->children, list) entry 757 sound/core/info.c proc_remove(entry->p); entry 758 sound/core/info.c entry->p = NULL; entry 767 sound/core/info.c void snd_info_free_entry(struct snd_info_entry * entry) entry 771 sound/core/info.c if (!entry) entry 773 sound/core/info.c if (entry->p) { entry 775 sound/core/info.c snd_info_disconnect(entry); entry 780 sound/core/info.c list_for_each_entry_safe(p, n, &entry->children, list) entry 783 sound/core/info.c p = entry->parent; entry 786 sound/core/info.c list_del(&entry->list); entry 789 sound/core/info.c kfree(entry->name); entry 790 sound/core/info.c if (entry->private_free) entry 791 sound/core/info.c entry->private_free(entry); entry 792 sound/core/info.c kfree(entry); entry 796 sound/core/info.c static int __snd_info_register(struct snd_info_entry *entry) entry 800 sound/core/info.c if (snd_BUG_ON(!entry)) entry 802 sound/core/info.c root = entry->parent == NULL ? snd_proc_root->p : entry->parent->p; entry 804 sound/core/info.c if (entry->p || !root) entry 806 sound/core/info.c if (S_ISDIR(entry->mode)) { entry 807 sound/core/info.c p = proc_mkdir_mode(entry->name, entry->mode, root); entry 814 sound/core/info.c if (entry->content == SNDRV_INFO_CONTENT_DATA) entry 818 sound/core/info.c p = proc_create_data(entry->name, entry->mode, root, entry 819 sound/core/info.c ops, entry); entry 824 sound/core/info.c proc_set_size(p, entry->size); entry 826 sound/core/info.c entry->p = p; entry 841 sound/core/info.c int snd_info_register(struct snd_info_entry *entry) entry 846 sound/core/info.c if (!entry->p) { entry 847 sound/core/info.c err = __snd_info_register(entry); entry 852 sound/core/info.c list_for_each_entry(p, &entry->children, list) { entry 877 sound/core/info.c void (*write)(struct snd_info_entry *entry, entry 880 sound/core/info.c struct snd_info_entry *entry; entry 882 sound/core/info.c entry = snd_info_create_card_entry(card, name, card->proc_root); entry 883 sound/core/info.c if (!entry) entry 885 sound/core/info.c snd_info_set_text_ops(entry, private_data, read); entry 887 sound/core/info.c entry->mode |= 0200; entry 888 sound/core/info.c entry->c.text.write = write; entry 898 sound/core/info.c static void snd_info_version_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 907 sound/core/info.c struct snd_info_entry *entry; entry 909 sound/core/info.c entry = snd_info_create_module_entry(THIS_MODULE, "version", NULL); entry 910 sound/core/info.c if (entry == NULL) entry 912 sound/core/info.c entry->c.text.read = snd_info_version_read; entry 913 sound/core/info.c return snd_info_register(entry); /* freed in error path */ entry 74 sound/core/info_oss.c static void snd_sndstat_proc_read(struct snd_info_entry *entry, entry 98 sound/core/info_oss.c struct snd_info_entry *entry; entry 101 sound/core/info_oss.c entry = snd_info_create_module_entry(THIS_MODULE, "sndstat", entry 103 sound/core/info_oss.c if (!entry) entry 105 sound/core/info_oss.c entry->c.text.read = snd_sndstat_proc_read; entry 106 sound/core/info_oss.c return snd_info_register(entry); /* freed in error path */ entry 789 sound/core/init.c static void snd_card_info_read(struct snd_info_entry *entry, entry 835 sound/core/init.c static void snd_card_module_info_read(struct snd_info_entry *entry, entry 853 sound/core/init.c struct snd_info_entry *entry; entry 855 sound/core/init.c entry = snd_info_create_module_entry(THIS_MODULE, "cards", NULL); entry 856 sound/core/init.c if (! entry) entry 858 sound/core/init.c entry->c.text.read = snd_card_info_read; entry 859 sound/core/init.c if (snd_info_register(entry) < 0) entry 863 sound/core/init.c entry = snd_info_create_module_entry(THIS_MODULE, "modules", NULL); entry 864 sound/core/init.c if (!entry) entry 866 sound/core/init.c entry->c.text.read = snd_card_module_info_read; entry 867 sound/core/init.c if (snd_info_register(entry) < 0) entry 1142 sound/core/oss/mixer_oss.c static void snd_mixer_oss_proc_read(struct snd_info_entry *entry, entry 1145 sound/core/oss/mixer_oss.c struct snd_mixer_oss *mixer = entry->private_data; entry 1166 sound/core/oss/mixer_oss.c static void snd_mixer_oss_proc_write(struct snd_info_entry *entry, entry 1169 sound/core/oss/mixer_oss.c struct snd_mixer_oss *mixer = entry->private_data; entry 1228 sound/core/oss/mixer_oss.c struct snd_info_entry *entry; entry 1230 sound/core/oss/mixer_oss.c entry = snd_info_create_card_entry(mixer->card, "oss_mixer", entry 1232 sound/core/oss/mixer_oss.c if (! entry) entry 1234 sound/core/oss/mixer_oss.c entry->content = SNDRV_INFO_CONTENT_TEXT; entry 1235 sound/core/oss/mixer_oss.c entry->mode = S_IFREG | 0644; entry 1236 sound/core/oss/mixer_oss.c entry->c.text.read = snd_mixer_oss_proc_read; entry 1237 sound/core/oss/mixer_oss.c entry->c.text.write = snd_mixer_oss_proc_write; entry 1238 sound/core/oss/mixer_oss.c entry->private_data = mixer; entry 1239 sound/core/oss/mixer_oss.c if (snd_info_register(entry) < 0) { entry 1240 sound/core/oss/mixer_oss.c snd_info_free_entry(entry); entry 1241 sound/core/oss/mixer_oss.c entry = NULL; entry 1243 sound/core/oss/mixer_oss.c mixer->proc_entry = entry; entry 2910 sound/core/oss/pcm_oss.c static void snd_pcm_oss_proc_read(struct snd_info_entry *entry, entry 2913 sound/core/oss/pcm_oss.c struct snd_pcm_str *pstr = entry->private_data; entry 2945 sound/core/oss/pcm_oss.c static void snd_pcm_oss_proc_write(struct snd_info_entry *entry, entry 2948 sound/core/oss/pcm_oss.c struct snd_pcm_str *pstr = entry->private_data; entry 3027 sound/core/oss/pcm_oss.c struct snd_info_entry *entry; entry 3031 sound/core/oss/pcm_oss.c if ((entry = snd_info_create_card_entry(pcm->card, "oss", pstr->proc_root)) != NULL) { entry 3032 sound/core/oss/pcm_oss.c entry->content = SNDRV_INFO_CONTENT_TEXT; entry 3033 sound/core/oss/pcm_oss.c entry->mode = S_IFREG | 0644; entry 3034 sound/core/oss/pcm_oss.c entry->c.text.read = snd_pcm_oss_proc_read; entry 3035 sound/core/oss/pcm_oss.c entry->c.text.write = snd_pcm_oss_proc_write; entry 3036 sound/core/oss/pcm_oss.c entry->private_data = pstr; entry 3037 sound/core/oss/pcm_oss.c if (snd_info_register(entry) < 0) { entry 3038 sound/core/oss/pcm_oss.c snd_info_free_entry(entry); entry 3039 sound/core/oss/pcm_oss.c entry = NULL; entry 3042 sound/core/oss/pcm_oss.c pstr->oss.proc_entry = entry; entry 363 sound/core/pcm.c static void snd_pcm_stream_proc_info_read(struct snd_info_entry *entry, entry 366 sound/core/pcm.c snd_pcm_proc_info_read(((struct snd_pcm_str *)entry->private_data)->substream, entry 370 sound/core/pcm.c static void snd_pcm_substream_proc_info_read(struct snd_info_entry *entry, entry 373 sound/core/pcm.c snd_pcm_proc_info_read(entry->private_data, buffer); entry 376 sound/core/pcm.c static void snd_pcm_substream_proc_hw_params_read(struct snd_info_entry *entry, entry 379 sound/core/pcm.c struct snd_pcm_substream *substream = entry->private_data; entry 413 sound/core/pcm.c static void snd_pcm_substream_proc_sw_params_read(struct snd_info_entry *entry, entry 416 sound/core/pcm.c struct snd_pcm_substream *substream = entry->private_data; entry 441 sound/core/pcm.c static void snd_pcm_substream_proc_status_read(struct snd_info_entry *entry, entry 444 sound/core/pcm.c struct snd_pcm_substream *substream = entry->private_data; entry 478 sound/core/pcm.c static void snd_pcm_xrun_injection_write(struct snd_info_entry *entry, entry 481 sound/core/pcm.c struct snd_pcm_substream *substream = entry->private_data; entry 486 sound/core/pcm.c static void snd_pcm_xrun_debug_read(struct snd_info_entry *entry, entry 489 sound/core/pcm.c struct snd_pcm_str *pstr = entry->private_data; entry 493 sound/core/pcm.c static void snd_pcm_xrun_debug_write(struct snd_info_entry *entry, entry 496 sound/core/pcm.c struct snd_pcm_str *pstr = entry->private_data; entry 506 sound/core/pcm.c struct snd_info_entry *entry; entry 511 sound/core/pcm.c entry = snd_info_create_card_entry(pcm->card, name, entry 513 sound/core/pcm.c if (!entry) entry 515 sound/core/pcm.c entry->mode = S_IFDIR | 0555; entry 516 sound/core/pcm.c pstr->proc_root = entry; entry 517 sound/core/pcm.c entry = snd_info_create_card_entry(pcm->card, "info", pstr->proc_root); entry 518 sound/core/pcm.c if (entry) entry 519 sound/core/pcm.c snd_info_set_text_ops(entry, pstr, snd_pcm_stream_proc_info_read); entry 521 sound/core/pcm.c entry = snd_info_create_card_entry(pcm->card, "xrun_debug", entry 523 sound/core/pcm.c if (entry) { entry 524 sound/core/pcm.c snd_info_set_text_ops(entry, pstr, snd_pcm_xrun_debug_read); entry 525 sound/core/pcm.c entry->c.text.write = snd_pcm_xrun_debug_write; entry 526 sound/core/pcm.c entry->mode |= 0200; entry 545 sound/core/pcm.c struct snd_info_entry *entry; entry 547 sound/core/pcm.c entry = snd_info_create_card_entry(substream->pcm->card, name, entry 549 sound/core/pcm.c if (entry) entry 550 sound/core/pcm.c snd_info_set_text_ops(entry, substream, read); entry 551 sound/core/pcm.c return entry; entry 556 sound/core/pcm.c struct snd_info_entry *entry; entry 563 sound/core/pcm.c entry = snd_info_create_card_entry(card, name, entry 565 sound/core/pcm.c if (!entry) entry 567 sound/core/pcm.c entry->mode = S_IFDIR | 0555; entry 568 sound/core/pcm.c substream->proc_root = entry; entry 580 sound/core/pcm.c entry = create_substream_info_entry(substream, "xrun_injection", NULL); entry 581 sound/core/pcm.c if (entry) { entry 582 sound/core/pcm.c entry->c.text.write = snd_pcm_xrun_injection_write; entry 583 sound/core/pcm.c entry->mode = S_IFREG | 0200; entry 1162 sound/core/pcm.c static void snd_pcm_proc_read(struct snd_info_entry *entry, entry 1186 sound/core/pcm.c struct snd_info_entry *entry; entry 1188 sound/core/pcm.c entry = snd_info_create_module_entry(THIS_MODULE, "pcm", NULL); entry 1189 sound/core/pcm.c if (entry) { entry 1190 sound/core/pcm.c snd_info_set_text_ops(entry, NULL, snd_pcm_proc_read); entry 1191 sound/core/pcm.c if (snd_info_register(entry) < 0) { entry 1192 sound/core/pcm.c snd_info_free_entry(entry); entry 1193 sound/core/pcm.c entry = NULL; entry 1196 sound/core/pcm.c snd_pcm_proc_entry = entry; entry 104 sound/core/pcm_memory.c static void snd_pcm_lib_preallocate_proc_read(struct snd_info_entry *entry, entry 107 sound/core/pcm_memory.c struct snd_pcm_substream *substream = entry->private_data; entry 116 sound/core/pcm_memory.c static void snd_pcm_lib_preallocate_max_proc_read(struct snd_info_entry *entry, entry 119 sound/core/pcm_memory.c struct snd_pcm_substream *substream = entry->private_data; entry 128 sound/core/pcm_memory.c static void snd_pcm_lib_preallocate_proc_write(struct snd_info_entry *entry, entry 131 sound/core/pcm_memory.c struct snd_pcm_substream *substream = entry->private_data; entry 172 sound/core/pcm_memory.c struct snd_info_entry *entry; entry 174 sound/core/pcm_memory.c entry = snd_info_create_card_entry(substream->pcm->card, "prealloc", entry 176 sound/core/pcm_memory.c if (entry) { entry 177 sound/core/pcm_memory.c snd_info_set_text_ops(entry, substream, entry 179 sound/core/pcm_memory.c entry->c.text.write = snd_pcm_lib_preallocate_proc_write; entry 180 sound/core/pcm_memory.c entry->mode |= 0200; entry 182 sound/core/pcm_memory.c entry = snd_info_create_card_entry(substream->pcm->card, "prealloc_max", entry 184 sound/core/pcm_memory.c if (entry) entry 185 sound/core/pcm_memory.c snd_info_set_text_ops(entry, substream, entry 1449 sound/core/rawmidi.c static void snd_rawmidi_proc_info_read(struct snd_info_entry *entry, entry 1456 sound/core/rawmidi.c rmidi = entry->private_data; entry 1676 sound/core/rawmidi.c struct snd_info_entry *entry; entry 1733 sound/core/rawmidi.c entry = snd_info_create_card_entry(rmidi->card, name, rmidi->card->proc_root); entry 1734 sound/core/rawmidi.c if (entry) { entry 1735 sound/core/rawmidi.c entry->private_data = rmidi; entry 1736 sound/core/rawmidi.c entry->c.text.read = snd_rawmidi_proc_info_read; entry 1737 sound/core/rawmidi.c if (snd_info_register(entry) < 0) { entry 1738 sound/core/rawmidi.c snd_info_free_entry(entry); entry 1739 sound/core/rawmidi.c entry = NULL; entry 1742 sound/core/rawmidi.c rmidi->proc_entry = entry; entry 259 sound/core/seq/oss/seq_oss.c info_read(struct snd_info_entry *entry, struct snd_info_buffer *buf) entry 273 sound/core/seq/oss/seq_oss.c struct snd_info_entry *entry; entry 275 sound/core/seq/oss/seq_oss.c entry = snd_info_create_module_entry(THIS_MODULE, SNDRV_SEQ_OSS_PROCNAME, snd_seq_root); entry 276 sound/core/seq/oss/seq_oss.c if (entry == NULL) entry 279 sound/core/seq/oss/seq_oss.c entry->content = SNDRV_INFO_CONTENT_TEXT; entry 280 sound/core/seq/oss/seq_oss.c entry->private_data = NULL; entry 281 sound/core/seq/oss/seq_oss.c entry->c.text.read = info_read; entry 282 sound/core/seq/oss/seq_oss.c if (snd_info_register(entry) < 0) { entry 283 sound/core/seq/oss/seq_oss.c snd_info_free_entry(entry); entry 286 sound/core/seq/oss/seq_oss.c info_entry = entry; entry 2445 sound/core/seq/seq_clientmgr.c void snd_seq_info_clients_read(struct snd_info_entry *entry, entry 24 sound/core/seq/seq_info.c struct snd_info_entry *entry; entry 26 sound/core/seq/seq_info.c entry = snd_info_create_module_entry(THIS_MODULE, name, snd_seq_root); entry 27 sound/core/seq/seq_info.c if (entry == NULL) entry 29 sound/core/seq/seq_info.c entry->content = SNDRV_INFO_CONTENT_TEXT; entry 30 sound/core/seq/seq_info.c entry->c.text.read = read; entry 31 sound/core/seq/seq_info.c if (snd_info_register(entry) < 0) { entry 32 sound/core/seq/seq_info.c snd_info_free_entry(entry); entry 35 sound/core/seq/seq_info.c return entry; entry 12 sound/core/seq/seq_info.h void snd_seq_info_clients_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer); entry 13 sound/core/seq/seq_info.h void snd_seq_info_timer_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer); entry 14 sound/core/seq/seq_info.h void snd_seq_info_queues_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer); entry 744 sound/core/seq/seq_queue.c void snd_seq_info_queues_read(struct snd_info_entry *entry, entry 462 sound/core/seq/seq_timer.c void snd_seq_info_timer_read(struct snd_info_entry *entry, entry 74 sound/core/seq_device.c static void snd_seq_device_info(struct snd_info_entry *entry, entry 345 sound/core/sound.c static void snd_minor_info_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 372 sound/core/sound.c struct snd_info_entry *entry; entry 374 sound/core/sound.c entry = snd_info_create_module_entry(THIS_MODULE, "devices", NULL); entry 375 sound/core/sound.c if (!entry) entry 377 sound/core/sound.c entry->c.text.read = snd_minor_info_read; entry 378 sound/core/sound.c return snd_info_register(entry); /* freed in error path */ entry 212 sound/core/sound_oss.c static void snd_minor_info_oss_read(struct snd_info_entry *entry, entry 236 sound/core/sound_oss.c struct snd_info_entry *entry; entry 238 sound/core/sound_oss.c entry = snd_info_create_module_entry(THIS_MODULE, "devices", snd_oss_root); entry 239 sound/core/sound_oss.c if (!entry) entry 241 sound/core/sound_oss.c entry->c.text.read = snd_minor_info_oss_read; entry 242 sound/core/sound_oss.c return snd_info_register(entry); /* freed in error path */ entry 1216 sound/core/timer.c static void snd_timer_proc_read(struct snd_info_entry *entry, entry 1266 sound/core/timer.c struct snd_info_entry *entry; entry 1268 sound/core/timer.c entry = snd_info_create_module_entry(THIS_MODULE, "timers", NULL); entry 1269 sound/core/timer.c if (entry != NULL) { entry 1270 sound/core/timer.c entry->c.text.read = snd_timer_proc_read; entry 1271 sound/core/timer.c if (snd_info_register(entry) < 0) { entry 1272 sound/core/timer.c snd_info_free_entry(entry); entry 1273 sound/core/timer.c entry = NULL; entry 1276 sound/core/timer.c snd_timer_proc_entry = entry; entry 1107 sound/drivers/aloop.c static void print_cable_info(struct snd_info_entry *entry, entry 1110 sound/drivers/aloop.c struct loopback *loopback = entry->private_data; entry 1114 sound/drivers/aloop.c num = entry->name[strlen(entry->name)-1]; entry 973 sound/drivers/dummy.c static void dummy_proc_read(struct snd_info_entry *entry, entry 976 sound/drivers/dummy.c struct snd_dummy *dummy = entry->private_data; entry 995 sound/drivers/dummy.c static void dummy_proc_write(struct snd_info_entry *entry, entry 998 sound/drivers/dummy.c struct snd_dummy *dummy = entry->private_data; entry 12 sound/drivers/opl4/opl4_proc.c static int snd_opl4_mem_proc_open(struct snd_info_entry *entry, entry 15 sound/drivers/opl4/opl4_proc.c struct snd_opl4 *opl4 = entry->private_data; entry 27 sound/drivers/opl4/opl4_proc.c static int snd_opl4_mem_proc_release(struct snd_info_entry *entry, entry 30 sound/drivers/opl4/opl4_proc.c struct snd_opl4 *opl4 = entry->private_data; entry 38 sound/drivers/opl4/opl4_proc.c static ssize_t snd_opl4_mem_proc_read(struct snd_info_entry *entry, entry 43 sound/drivers/opl4/opl4_proc.c struct snd_opl4 *opl4 = entry->private_data; entry 58 sound/drivers/opl4/opl4_proc.c static ssize_t snd_opl4_mem_proc_write(struct snd_info_entry *entry, entry 64 sound/drivers/opl4/opl4_proc.c struct snd_opl4 *opl4 = entry->private_data; entry 88 sound/drivers/opl4/opl4_proc.c struct snd_info_entry *entry; entry 90 sound/drivers/opl4/opl4_proc.c entry = snd_info_create_card_entry(opl4->card, "opl4-mem", opl4->card->proc_root); entry 91 sound/drivers/opl4/opl4_proc.c if (entry) { entry 94 sound/drivers/opl4/opl4_proc.c entry->mode |= 0200; entry 95 sound/drivers/opl4/opl4_proc.c entry->size = 4 * 1024 * 1024; entry 98 sound/drivers/opl4/opl4_proc.c entry->size = 1 * 1024 * 1024; entry 100 sound/drivers/opl4/opl4_proc.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 101 sound/drivers/opl4/opl4_proc.c entry->c.ops = &snd_opl4_mem_proc_ops; entry 102 sound/drivers/opl4/opl4_proc.c entry->module = THIS_MODULE; entry 103 sound/drivers/opl4/opl4_proc.c entry->private_data = opl4; entry 105 sound/drivers/opl4/opl4_proc.c opl4->proc_entry = entry; entry 588 sound/drivers/vx/vx_core.c static void vx_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 590 sound/drivers/vx/vx_core.c struct vx_core *chip = entry->private_data; entry 902 sound/firewire/amdtp-stream.c } *entry, initial_state[] = { entry 927 sound/firewire/amdtp-stream.c entry = &initial_state[s->sfc]; entry 930 sound/firewire/amdtp-stream.c s->ctx_data.rx.data_block_state = entry->data_block; entry 931 sound/firewire/amdtp-stream.c s->ctx_data.rx.syt_offset_state = entry->syt_offset; entry 203 sound/firewire/bebob/bebob.c if (bebob->entry->model_id == MODEL_MAUDIO_FW1814) entry 247 sound/firewire/bebob/bebob.c bebob_probe(struct fw_unit *unit, const struct ieee1394_device_id *entry) entry 252 sound/firewire/bebob/bebob.c if (entry->vendor_id == VEN_FOCUSRITE && entry 253 sound/firewire/bebob/bebob.c entry->model_id == MODEL_FOCUSRITE_SAFFIRE_BOTH) entry 255 sound/firewire/bebob/bebob.c else if (entry->vendor_id == VEN_MAUDIO1 && entry 256 sound/firewire/bebob/bebob.c entry->model_id == MODEL_MAUDIO_AUDIOPHILE_BOTH && entry 260 sound/firewire/bebob/bebob.c spec = (const struct snd_bebob_spec *)entry->driver_data; entry 263 sound/firewire/bebob/bebob.c if (entry->vendor_id == VEN_MAUDIO1 || entry 264 sound/firewire/bebob/bebob.c entry->vendor_id == VEN_MAUDIO2) entry 278 sound/firewire/bebob/bebob.c bebob->entry = entry; entry 287 sound/firewire/bebob/bebob.c if (entry->vendor_id != VEN_MAUDIO1 || entry 288 sound/firewire/bebob/bebob.c (entry->model_id != MODEL_MAUDIO_FW1814 && entry 289 sound/firewire/bebob/bebob.c entry->model_id != MODEL_MAUDIO_PROJECTMIX)) { entry 89 sound/firewire/bebob/bebob.h const struct ieee1394_device_id *entry; entry 35 sound/firewire/bebob/bebob_proc.c proc_read_hw_info(struct snd_info_entry *entry, entry 38 sound/firewire/bebob/bebob_proc.c struct snd_bebob *bebob = entry->private_data; entry 71 sound/firewire/bebob/bebob_proc.c proc_read_meters(struct snd_info_entry *entry, entry 74 sound/firewire/bebob/bebob_proc.c struct snd_bebob *bebob = entry->private_data; entry 104 sound/firewire/bebob/bebob_proc.c proc_read_formation(struct snd_info_entry *entry, entry 107 sound/firewire/bebob/bebob_proc.c struct snd_bebob *bebob = entry->private_data; entry 131 sound/firewire/bebob/bebob_proc.c proc_read_clock(struct snd_info_entry *entry, entry 139 sound/firewire/bebob/bebob_proc.c struct snd_bebob *bebob = entry->private_data; entry 162 sound/firewire/bebob/bebob_proc.c struct snd_info_entry *entry; entry 164 sound/firewire/bebob/bebob_proc.c entry = snd_info_create_card_entry(bebob->card, name, root); entry 165 sound/firewire/bebob/bebob_proc.c if (entry) entry 166 sound/firewire/bebob/bebob_proc.c snd_info_set_text_ops(entry, bebob, op); entry 27 sound/firewire/dice/dice-presonus.c } *entry, entries[] = { entry 44 sound/firewire/dice/dice-presonus.c entry = entries + i; entry 45 sound/firewire/dice/dice-presonus.c if (entry->model_id == model_id) entry 51 sound/firewire/dice/dice-presonus.c memcpy(dice->tx_pcm_chs, entry->spec->tx_pcm_chs, entry 53 sound/firewire/dice/dice-presonus.c memcpy(dice->rx_pcm_chs, entry->spec->rx_pcm_chs, entry 56 sound/firewire/dice/dice-presonus.c if (entry->spec->has_midi) { entry 56 sound/firewire/dice/dice-proc.c static void dice_proc_read(struct snd_info_entry *entry, entry 70 sound/firewire/dice/dice-proc.c struct snd_dice *dice = entry->private_data; entry 245 sound/firewire/dice/dice-proc.c static void dice_proc_read_formation(struct snd_info_entry *entry, entry 253 sound/firewire/dice/dice-proc.c struct snd_dice *dice = entry->private_data; entry 281 sound/firewire/dice/dice-proc.c void (*op)(struct snd_info_entry *entry, entry 284 sound/firewire/dice/dice-proc.c struct snd_info_entry *entry; entry 286 sound/firewire/dice/dice-proc.c entry = snd_info_create_card_entry(dice->card, name, root); entry 287 sound/firewire/dice/dice-proc.c if (entry) entry 288 sound/firewire/dice/dice-proc.c snd_info_set_text_ops(entry, dice, op); entry 63 sound/firewire/dice/dice-tcelectronic.c } *entry, entries[] = { entry 86 sound/firewire/dice/dice-tcelectronic.c entry = entries + i; entry 87 sound/firewire/dice/dice-tcelectronic.c if (entry->model_id == model_id) entry 93 sound/firewire/dice/dice-tcelectronic.c memcpy(dice->tx_pcm_chs, entry->spec->tx_pcm_chs, entry 95 sound/firewire/dice/dice-tcelectronic.c memcpy(dice->rx_pcm_chs, entry->spec->rx_pcm_chs, entry 98 sound/firewire/dice/dice-tcelectronic.c if (entry->spec->has_midi) { entry 196 sound/firewire/dice/dice.c const struct ieee1394_device_id *entry) entry 201 sound/firewire/dice/dice.c if (!entry->driver_data && entry->vendor_id != OUI_SSL) { entry 214 sound/firewire/dice/dice.c if (!entry->driver_data) { entry 218 sound/firewire/dice/dice.c (snd_dice_detect_formats_t)entry->driver_data; entry 25 sound/firewire/digi00x/digi00x-proc.c static void proc_read_clock(struct snd_info_entry *entry, entry 38 sound/firewire/digi00x/digi00x-proc.c struct snd_dg00x *dg00x = entry->private_data; entry 70 sound/firewire/digi00x/digi00x-proc.c struct snd_info_entry *root, *entry; entry 83 sound/firewire/digi00x/digi00x-proc.c entry = snd_info_create_card_entry(dg00x->card, "clock", root); entry 84 sound/firewire/digi00x/digi00x-proc.c if (entry) entry 85 sound/firewire/digi00x/digi00x-proc.c snd_info_set_text_ops(entry, dg00x, proc_read_clock); entry 107 sound/firewire/digi00x/digi00x.c const struct ieee1394_device_id *entry) entry 124 sound/firewire/digi00x/digi00x.c dg00x->is_console = entry->model_id == MODEL_CONSOLE; entry 27 sound/firewire/fireface/ff-proc.c static void proc_dump_status(struct snd_info_entry *entry, entry 30 sound/firewire/fireface/ff-proc.c struct snd_ff *ff = entry->private_data; entry 40 sound/firewire/fireface/ff-proc.c struct snd_info_entry *entry; entry 42 sound/firewire/fireface/ff-proc.c entry = snd_info_create_card_entry(ff->card, name, root); entry 43 sound/firewire/fireface/ff-proc.c if (entry) entry 44 sound/firewire/fireface/ff-proc.c snd_info_set_text_ops(entry, ff, op); entry 90 sound/firewire/fireface/ff.c const struct ieee1394_device_id *entry) entry 104 sound/firewire/fireface/ff.c ff->spec = (const struct snd_ff_spec *)entry->driver_data; entry 282 sound/firewire/fireworks/fireworks.c efw_probe(struct fw_unit *unit, const struct ieee1394_device_id *entry) entry 28 sound/firewire/fireworks/fireworks_proc.c proc_read_hwinfo(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 30 sound/firewire/fireworks/fireworks_proc.c struct snd_efw *efw = entry->private_data; entry 105 sound/firewire/fireworks/fireworks_proc.c proc_read_clock(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 107 sound/firewire/fireworks/fireworks_proc.c struct snd_efw *efw = entry->private_data; entry 127 sound/firewire/fireworks/fireworks_proc.c proc_read_phys_meters(struct snd_info_entry *entry, entry 130 sound/firewire/fireworks/fireworks_proc.c struct snd_efw *efw = entry->private_data; entry 178 sound/firewire/fireworks/fireworks_proc.c proc_read_queues_state(struct snd_info_entry *entry, entry 181 sound/firewire/fireworks/fireworks_proc.c struct snd_efw *efw = entry->private_data; entry 198 sound/firewire/fireworks/fireworks_proc.c struct snd_info_entry *entry; entry 200 sound/firewire/fireworks/fireworks_proc.c entry = snd_info_create_card_entry(efw->card, name, root); entry 201 sound/firewire/fireworks/fireworks_proc.c if (entry) entry 202 sound/firewire/fireworks/fireworks_proc.c snd_info_set_text_ops(entry, efw, op); entry 24 sound/firewire/motu/motu-proc.c static void proc_read_clock(struct snd_info_entry *entry, entry 28 sound/firewire/motu/motu-proc.c struct snd_motu *motu = entry->private_data; entry 42 sound/firewire/motu/motu-proc.c static void proc_read_format(struct snd_info_entry *entry, entry 45 sound/firewire/motu/motu-proc.c struct snd_motu *motu = entry->private_data; entry 86 sound/firewire/motu/motu-proc.c struct snd_info_entry *entry; entry 88 sound/firewire/motu/motu-proc.c entry = snd_info_create_card_entry(motu->card, name, root); entry 89 sound/firewire/motu/motu-proc.c if (entry) entry 90 sound/firewire/motu/motu-proc.c snd_info_set_text_ops(entry, motu, op); entry 120 sound/firewire/motu/motu.c const struct ieee1394_device_id *entry) entry 131 sound/firewire/motu/motu.c motu->spec = (const struct snd_motu_spec *)entry->driver_data; entry 10 sound/firewire/oxfw/oxfw-proc.c static void proc_read_formation(struct snd_info_entry *entry, entry 13 sound/firewire/oxfw/oxfw-proc.c struct snd_oxfw *oxfw = entry->private_data; entry 82 sound/firewire/oxfw/oxfw-proc.c struct snd_info_entry *entry; entry 84 sound/firewire/oxfw/oxfw-proc.c entry = snd_info_create_card_entry(oxfw->card, name, root); entry 85 sound/firewire/oxfw/oxfw-proc.c if (entry) entry 86 sound/firewire/oxfw/oxfw-proc.c snd_info_set_text_ops(entry, oxfw, op); entry 92 sound/firewire/oxfw/oxfw.c if (oxfw->entry->vendor_id == VENDOR_GRIFFIN || entry 93 sound/firewire/oxfw/oxfw.c oxfw->entry->vendor_id == VENDOR_LACIE) { entry 94 sound/firewire/oxfw/oxfw.c info = (const struct compat_info *)oxfw->entry->driver_data; entry 135 sound/firewire/oxfw/oxfw.c if (oxfw->entry->vendor_id == VENDOR_GRIFFIN) entry 137 sound/firewire/oxfw/oxfw.c if (oxfw->entry->vendor_id == VENDOR_LACIE) entry 144 sound/firewire/oxfw/oxfw.c if (oxfw->entry->vendor_id == OUI_STANTON) { entry 156 sound/firewire/oxfw/oxfw.c if (oxfw->entry->vendor_id == VENDOR_TASCAM) { entry 241 sound/firewire/oxfw/oxfw.c const struct ieee1394_device_id *entry) entry 245 sound/firewire/oxfw/oxfw.c if (entry->vendor_id == VENDOR_LOUD && !detect_loud_models(unit)) entry 255 sound/firewire/oxfw/oxfw.c oxfw->entry = entry; entry 281 sound/firewire/oxfw/oxfw.c if (oxfw->entry->vendor_id == OUI_STANTON) entry 64 sound/firewire/oxfw/oxfw.h const struct ieee1394_device_id *entry; entry 158 sound/firewire/tascam/amdtp-tascam.c struct snd_firewire_tascam_change *entry = entry 163 sound/firewire/tascam/amdtp-tascam.c entry->index = index; entry 164 sound/firewire/tascam/amdtp-tascam.c entry->before = before; entry 165 sound/firewire/tascam/amdtp-tascam.c entry->after = after; entry 10 sound/firewire/tascam/tascam-proc.c static void proc_read_firmware(struct snd_info_entry *entry, entry 13 sound/firewire/tascam/tascam-proc.c struct snd_tscm *tscm = entry->private_data; entry 57 sound/firewire/tascam/tascam-proc.c struct snd_info_entry *entry; entry 59 sound/firewire/tascam/tascam-proc.c entry = snd_info_create_card_entry(tscm->card, name, root); entry 60 sound/firewire/tascam/tascam-proc.c if (entry) entry 61 sound/firewire/tascam/tascam-proc.c snd_info_set_text_ops(entry, tscm, op); entry 150 sound/firewire/tascam/tascam.c const struct ieee1394_device_id *entry) entry 466 sound/i2c/other/ak4113.c static void snd_ak4113_proc_regs_read(struct snd_info_entry *entry, entry 469 sound/i2c/other/ak4113.c struct ak4113 *ak4113 = entry->private_data; entry 439 sound/i2c/other/ak4114.c static void snd_ak4114_proc_regs_read(struct snd_info_entry *entry, entry 442 sound/i2c/other/ak4114.c struct ak4114 *ak4114 = entry->private_data; entry 849 sound/i2c/other/ak4xxx-adda.c static void proc_regs_read(struct snd_info_entry *entry, entry 852 sound/i2c/other/ak4xxx-adda.c struct snd_akm4xxx *ak = entry->private_data; entry 102 sound/isa/gus/gus_irq.c static void snd_gus_irq_info_read(struct snd_info_entry *entry, entry 109 sound/isa/gus/gus_irq.c gus = entry->private_data; entry 14 sound/isa/gus/gus_mem.c static void snd_gf1_mem_info_read(struct snd_info_entry *entry, entry 269 sound/isa/gus/gus_mem.c static void snd_gf1_mem_info_read(struct snd_info_entry *entry, entry 278 sound/isa/gus/gus_mem.c gus = entry->private_data; entry 19 sound/isa/gus/gus_mem_proc.c static ssize_t snd_gf1_mem_proc_dump(struct snd_info_entry *entry, entry 24 sound/isa/gus/gus_mem_proc.c struct gus_proc_private *priv = entry->private_data; entry 34 sound/isa/gus/gus_mem_proc.c static void snd_gf1_mem_proc_free(struct snd_info_entry *entry) entry 36 sound/isa/gus/gus_mem_proc.c struct gus_proc_private *priv = entry->private_data; entry 49 sound/isa/gus/gus_mem_proc.c struct snd_info_entry *entry; entry 58 sound/isa/gus/gus_mem_proc.c if (! snd_card_proc_new(gus->card, name, &entry)) { entry 59 sound/isa/gus/gus_mem_proc.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 60 sound/isa/gus/gus_mem_proc.c entry->private_data = priv; entry 61 sound/isa/gus/gus_mem_proc.c entry->private_free = snd_gf1_mem_proc_free; entry 62 sound/isa/gus/gus_mem_proc.c entry->c.ops = &snd_gf1_mem_proc_ops; entry 64 sound/isa/gus/gus_mem_proc.c priv->size = entry->size = gus->gf1.mem_alloc.banks_8[idx].size; entry 76 sound/isa/gus/gus_mem_proc.c if (! snd_card_proc_new(gus->card, name, &entry)) { entry 77 sound/isa/gus/gus_mem_proc.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 78 sound/isa/gus/gus_mem_proc.c entry->private_data = priv; entry 79 sound/isa/gus/gus_mem_proc.c entry->private_free = snd_gf1_mem_proc_free; entry 80 sound/isa/gus/gus_mem_proc.c entry->c.ops = &snd_gf1_mem_proc_ops; entry 82 sound/isa/gus/gus_mem_proc.c priv->size = entry->size = gus->gf1.rom_memory; entry 882 sound/isa/opti9xx/miro.c static void snd_miro_proc_read(struct snd_info_entry * entry, entry 885 sound/isa/opti9xx/miro.c struct snd_miro *miro = (struct snd_miro *) entry->private_data; entry 97 sound/isa/sb/sb16_csp.c static void info_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer); entry 1121 sound/isa/sb/sb16_csp.c static void info_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 1123 sound/isa/sb/sb16_csp.c struct snd_sb_csp *p = entry->private_data; entry 328 sound/pci/ac97/ac97_proc.c static void snd_ac97_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 330 sound/pci/ac97/ac97_proc.c struct snd_ac97 *ac97 = entry->private_data; entry 363 sound/pci/ac97/ac97_proc.c static void snd_ac97_proc_regs_write(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 365 sound/pci/ac97/ac97_proc.c struct snd_ac97 *ac97 = entry->private_data; entry 390 sound/pci/ac97/ac97_proc.c static void snd_ac97_proc_regs_read(struct snd_info_entry *entry, entry 393 sound/pci/ac97/ac97_proc.c struct snd_ac97 *ac97 = entry->private_data; entry 416 sound/pci/ac97/ac97_proc.c struct snd_info_entry *entry; entry 424 sound/pci/ac97/ac97_proc.c entry = snd_info_create_card_entry(ac97->bus->card, name, entry 426 sound/pci/ac97/ac97_proc.c if (entry) entry 427 sound/pci/ac97/ac97_proc.c snd_info_set_text_ops(entry, ac97, snd_ac97_proc_read); entry 428 sound/pci/ac97/ac97_proc.c ac97->proc = entry; entry 430 sound/pci/ac97/ac97_proc.c entry = snd_info_create_card_entry(ac97->bus->card, name, entry 432 sound/pci/ac97/ac97_proc.c if (entry) { entry 433 sound/pci/ac97/ac97_proc.c snd_info_set_text_ops(entry, ac97, snd_ac97_proc_regs_read); entry 435 sound/pci/ac97/ac97_proc.c entry->mode |= 0200; entry 436 sound/pci/ac97/ac97_proc.c entry->c.text.write = snd_ac97_proc_regs_write; entry 439 sound/pci/ac97/ac97_proc.c ac97->proc_regs = entry; entry 452 sound/pci/ac97/ac97_proc.c struct snd_info_entry *entry; entry 456 sound/pci/ac97/ac97_proc.c entry = snd_info_create_card_entry(bus->card, name, entry 458 sound/pci/ac97/ac97_proc.c if (entry) entry 459 sound/pci/ac97/ac97_proc.c entry->mode = S_IFDIR | 0555; entry 460 sound/pci/ac97/ac97_proc.c bus->proc = entry; entry 644 sound/pci/ad1889.c snd_ad1889_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 646 sound/pci/ad1889.c struct snd_ad1889 *chip = entry->private_data; entry 454 sound/pci/ak4531_codec.c static void snd_ak4531_proc_read(struct snd_info_entry *entry, entry 457 sound/pci/ak4531_codec.c struct snd_ak4531 *ak4531 = entry->private_data; entry 2027 sound/pci/ali5451/ali5451.c static void snd_ali_proc_read(struct snd_info_entry *entry, entry 2030 sound/pci/ali5451/ali5451.c struct snd_ali *codec = entry->private_data; entry 2720 sound/pci/asihpi/asihpi.c snd_asihpi_proc_read(struct snd_info_entry *entry, entry 2723 sound/pci/asihpi/asihpi.c struct snd_card_asihpi *asihpi = entry->private_data; entry 1520 sound/pci/atiixp.c static void snd_atiixp_proc_read(struct snd_info_entry *entry, entry 1523 sound/pci/atiixp.c struct atiixp *chip = entry->private_data; entry 1147 sound/pci/atiixp_modem.c static void snd_atiixp_proc_read(struct snd_info_entry *entry, entry 1150 sound/pci/atiixp_modem.c struct atiixp_modem *chip = entry->private_data; entry 258 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_iec958(struct snd_info_entry *entry, entry 261 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 280 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_write32(struct snd_info_entry *entry, entry 283 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 298 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read32(struct snd_info_entry *entry, entry 301 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 314 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read16(struct snd_info_entry *entry, entry 317 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 330 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read8(struct snd_info_entry *entry, entry 333 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 346 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read1(struct snd_info_entry *entry, entry 349 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 364 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_read2(struct snd_info_entry *entry, entry 367 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 382 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_reg_write(struct snd_info_entry *entry, entry 385 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 396 sound/pci/ca0106/ca0106_proc.c static void snd_ca0106_proc_i2c_write(struct snd_info_entry *entry, entry 399 sound/pci/ca0106/ca0106_proc.c struct snd_ca0106 *emu = entry->private_data; entry 2762 sound/pci/cmipci.c static void snd_cmipci_proc_read(struct snd_info_entry *entry, entry 2765 sound/pci/cmipci.c struct cmipci *cm = entry->private_data; entry 1115 sound/pci/cs4281.c static void snd_cs4281_proc_read(struct snd_info_entry *entry, entry 1118 sound/pci/cs4281.c struct cs4281 *chip = entry->private_data; entry 1125 sound/pci/cs4281.c static ssize_t snd_cs4281_BA0_read(struct snd_info_entry *entry, entry 1130 sound/pci/cs4281.c struct cs4281 *chip = entry->private_data; entry 1137 sound/pci/cs4281.c static ssize_t snd_cs4281_BA1_read(struct snd_info_entry *entry, entry 1142 sound/pci/cs4281.c struct cs4281 *chip = entry->private_data; entry 1159 sound/pci/cs4281.c struct snd_info_entry *entry; entry 1162 sound/pci/cs4281.c if (! snd_card_proc_new(chip->card, "cs4281_BA0", &entry)) { entry 1163 sound/pci/cs4281.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 1164 sound/pci/cs4281.c entry->private_data = chip; entry 1165 sound/pci/cs4281.c entry->c.ops = &snd_cs4281_proc_ops_BA0; entry 1166 sound/pci/cs4281.c entry->size = CS4281_BA0_SIZE; entry 1168 sound/pci/cs4281.c if (! snd_card_proc_new(chip->card, "cs4281_BA1", &entry)) { entry 1169 sound/pci/cs4281.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 1170 sound/pci/cs4281.c entry->private_data = chip; entry 1171 sound/pci/cs4281.c entry->c.ops = &snd_cs4281_proc_ops_BA1; entry 1172 sound/pci/cs4281.c entry->size = CS4281_BA1_SIZE; entry 421 sound/pci/cs46xx/cs46xx_lib.c struct dsp_symbol_entry *entry = entry 425 sound/pci/cs46xx/cs46xx_lib.c entry->address = le32_to_cpu(fwdat[fwlen++]); entry 426 sound/pci/cs46xx/cs46xx_lib.c memcpy(entry->symbol_name, &fwdat[fwlen], DSP_MAX_SYMBOL_NAME - 1); entry 428 sound/pci/cs46xx/cs46xx_lib.c entry->symbol_type = le32_to_cpu(fwdat[fwlen++]); entry 441 sound/pci/cs46xx/cs46xx_lib.c struct dsp_segment_desc *entry = &module->segments[i]; entry 444 sound/pci/cs46xx/cs46xx_lib.c entry->segment_type = le32_to_cpu(fwdat[fwlen++]); entry 445 sound/pci/cs46xx/cs46xx_lib.c entry->offset = le32_to_cpu(fwdat[fwlen++]); entry 446 sound/pci/cs46xx/cs46xx_lib.c entry->size = le32_to_cpu(fwdat[fwlen++]); entry 447 sound/pci/cs46xx/cs46xx_lib.c if (fwlen + entry->size > fwsize) entry 449 sound/pci/cs46xx/cs46xx_lib.c entry->data = kmalloc_array(entry->size, 4, GFP_KERNEL); entry 450 sound/pci/cs46xx/cs46xx_lib.c if (!entry->data) entry 452 sound/pci/cs46xx/cs46xx_lib.c memcpy_le32(entry->data, &fwdat[fwlen], entry->size * 4); entry 453 sound/pci/cs46xx/cs46xx_lib.c fwlen += entry->size; entry 2812 sound/pci/cs46xx/cs46xx_lib.c static ssize_t snd_cs46xx_io_read(struct snd_info_entry *entry, entry 2817 sound/pci/cs46xx/cs46xx_lib.c struct snd_cs46xx_region *region = entry->private_data; entry 2830 sound/pci/cs46xx/cs46xx_lib.c struct snd_info_entry *entry; entry 2835 sound/pci/cs46xx/cs46xx_lib.c if (! snd_card_proc_new(card, region->name, &entry)) { entry 2836 sound/pci/cs46xx/cs46xx_lib.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 2837 sound/pci/cs46xx/cs46xx_lib.c entry->private_data = chip; entry 2838 sound/pci/cs46xx/cs46xx_lib.c entry->c.ops = &snd_cs46xx_proc_io_ops; entry 2839 sound/pci/cs46xx/cs46xx_lib.c entry->size = region->size; entry 2840 sound/pci/cs46xx/cs46xx_lib.c entry->mode = S_IFREG | 0400; entry 493 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_symbol_table_read (struct snd_info_entry *entry, entry 496 sound/pci/cs46xx/dsp_spos.c struct snd_cs46xx *chip = entry->private_data; entry 521 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_modules_read (struct snd_info_entry *entry, entry 524 sound/pci/cs46xx/dsp_spos.c struct snd_cs46xx *chip = entry->private_data; entry 544 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_task_tree_read (struct snd_info_entry *entry, entry 547 sound/pci/cs46xx/dsp_spos.c struct snd_cs46xx *chip = entry->private_data; entry 572 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_scb_read (struct snd_info_entry *entry, entry 575 sound/pci/cs46xx/dsp_spos.c struct snd_cs46xx *chip = entry->private_data; entry 605 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_parameter_dump_read (struct snd_info_entry *entry, entry 608 sound/pci/cs46xx/dsp_spos.c struct snd_cs46xx *chip = entry->private_data; entry 633 sound/pci/cs46xx/dsp_spos.c static void cs46xx_dsp_proc_sample_dump_read (struct snd_info_entry *entry, entry 636 sound/pci/cs46xx/dsp_spos.c struct snd_cs46xx *chip = entry->private_data; entry 783 sound/pci/cs46xx/dsp_spos.c struct snd_info_entry *entry; entry 789 sound/pci/cs46xx/dsp_spos.c entry = snd_info_create_card_entry(card, "dsp", card->proc_root); entry 790 sound/pci/cs46xx/dsp_spos.c if (entry) entry 791 sound/pci/cs46xx/dsp_spos.c entry->mode = S_IFDIR | 0555; entry 792 sound/pci/cs46xx/dsp_spos.c ins->proc_dsp_dir = entry; entry 797 sound/pci/cs46xx/dsp_spos.c entry = snd_info_create_card_entry(card, "spos_symbols", entry 799 sound/pci/cs46xx/dsp_spos.c if (entry) entry 800 sound/pci/cs46xx/dsp_spos.c snd_info_set_text_ops(entry, chip, entry 803 sound/pci/cs46xx/dsp_spos.c entry = snd_info_create_card_entry(card, "spos_modules", entry 805 sound/pci/cs46xx/dsp_spos.c if (entry) entry 806 sound/pci/cs46xx/dsp_spos.c snd_info_set_text_ops(entry, chip, entry 809 sound/pci/cs46xx/dsp_spos.c entry = snd_info_create_card_entry(card, "parameter", entry 811 sound/pci/cs46xx/dsp_spos.c if (entry) entry 812 sound/pci/cs46xx/dsp_spos.c snd_info_set_text_ops(entry, chip, entry 815 sound/pci/cs46xx/dsp_spos.c entry = snd_info_create_card_entry(card, "sample", entry 817 sound/pci/cs46xx/dsp_spos.c if (entry) entry 818 sound/pci/cs46xx/dsp_spos.c snd_info_set_text_ops(entry, chip, entry 821 sound/pci/cs46xx/dsp_spos.c entry = snd_info_create_card_entry(card, "task_tree", entry 823 sound/pci/cs46xx/dsp_spos.c if (entry) entry 824 sound/pci/cs46xx/dsp_spos.c snd_info_set_text_ops(entry, chip, entry 827 sound/pci/cs46xx/dsp_spos.c entry = snd_info_create_card_entry(card, "scb_info", entry 829 sound/pci/cs46xx/dsp_spos.c if (entry) entry 830 sound/pci/cs46xx/dsp_spos.c snd_info_set_text_ops(entry, chip, entry 57 sound/pci/cs46xx/dsp_spos_scb_lib.c static void cs46xx_dsp_proc_scb_info_read (struct snd_info_entry *entry, entry 60 sound/pci/cs46xx/dsp_spos_scb_lib.c struct proc_scb_info * scb_info = entry->private_data; entry 236 sound/pci/cs46xx/dsp_spos_scb_lib.c struct snd_info_entry * entry; entry 243 sound/pci/cs46xx/dsp_spos_scb_lib.c entry = snd_info_create_card_entry(ins->snd_card, scb->scb_name, entry 245 sound/pci/cs46xx/dsp_spos_scb_lib.c if (entry) { entry 248 sound/pci/cs46xx/dsp_spos_scb_lib.c snd_info_free_entry(entry); entry 249 sound/pci/cs46xx/dsp_spos_scb_lib.c entry = NULL; entry 255 sound/pci/cs46xx/dsp_spos_scb_lib.c snd_info_set_text_ops(entry, scb_info, entry 259 sound/pci/cs46xx/dsp_spos_scb_lib.c scb->proc_info = entry; entry 158 sound/pci/ctxfi/ctdaio.c struct imapper *entry; entry 162 sound/pci/ctxfi/ctdaio.c entry = kzalloc((sizeof(*entry) * daio->rscl.msr), GFP_KERNEL); entry 163 sound/pci/ctxfi/ctdaio.c if (!entry) entry 170 sound/pci/ctxfi/ctdaio.c for (i = 0; i < daio->rscl.msr; i++, entry++) { entry 171 sound/pci/ctxfi/ctdaio.c entry->slot = input->ops->output_slot(input); entry 172 sound/pci/ctxfi/ctdaio.c entry->user = entry->addr = daio->rscl.ops->index(&daio->rscl); entry 173 sound/pci/ctxfi/ctdaio.c dao->mgr->imap_add(dao->mgr, entry); entry 174 sound/pci/ctxfi/ctdaio.c dao->imappers[i] = entry; entry 187 sound/pci/ctxfi/ctdaio.c struct imapper *entry; entry 191 sound/pci/ctxfi/ctdaio.c entry = kzalloc((sizeof(*entry) * daio->rscr.msr), GFP_KERNEL); entry 192 sound/pci/ctxfi/ctdaio.c if (!entry) entry 199 sound/pci/ctxfi/ctdaio.c for (i = 0; i < daio->rscr.msr; i++, entry++) { entry 200 sound/pci/ctxfi/ctdaio.c entry->slot = input->ops->output_slot(input); entry 201 sound/pci/ctxfi/ctdaio.c entry->user = entry->addr = daio->rscr.ops->index(&daio->rscr); entry 202 sound/pci/ctxfi/ctdaio.c dao->mgr->imap_add(dao->mgr, entry); entry 203 sound/pci/ctxfi/ctdaio.c dao->imappers[daio->rscl.msr + i] = entry; entry 216 sound/pci/ctxfi/ctdaio.c struct imapper *entry; entry 223 sound/pci/ctxfi/ctdaio.c entry = dao->imappers[0]; entry 224 sound/pci/ctxfi/ctdaio.c dao->mgr->imap_delete(dao->mgr, entry); entry 227 sound/pci/ctxfi/ctdaio.c entry = dao->imappers[i]; entry 228 sound/pci/ctxfi/ctdaio.c dao->mgr->imap_delete(dao->mgr, entry); entry 240 sound/pci/ctxfi/ctdaio.c struct imapper *entry; entry 247 sound/pci/ctxfi/ctdaio.c entry = dao->imappers[daio->rscl.msr]; entry 248 sound/pci/ctxfi/ctdaio.c dao->mgr->imap_delete(dao->mgr, entry); entry 251 sound/pci/ctxfi/ctdaio.c entry = dao->imappers[daio->rscl.msr + i]; entry 252 sound/pci/ctxfi/ctdaio.c dao->mgr->imap_delete(dao->mgr, entry); entry 632 sound/pci/ctxfi/ctdaio.c static int daio_map_op(void *data, struct imapper *entry) entry 637 sound/pci/ctxfi/ctdaio.c hw->daio_mgr_set_imaparc(mgr->ctrl_blk, entry->slot); entry 638 sound/pci/ctxfi/ctdaio.c hw->daio_mgr_set_imapnxt(mgr->ctrl_blk, entry->next); entry 639 sound/pci/ctxfi/ctdaio.c hw->daio_mgr_set_imapaddr(mgr->ctrl_blk, entry->addr); entry 645 sound/pci/ctxfi/ctdaio.c static int daio_imap_add(struct daio_mgr *mgr, struct imapper *entry) entry 651 sound/pci/ctxfi/ctdaio.c if (!entry->addr && mgr->init_imap_added) { entry 656 sound/pci/ctxfi/ctdaio.c err = input_mapper_add(&mgr->imappers, entry, daio_map_op, mgr); entry 662 sound/pci/ctxfi/ctdaio.c static int daio_imap_delete(struct daio_mgr *mgr, struct imapper *entry) entry 668 sound/pci/ctxfi/ctdaio.c err = input_mapper_delete(&mgr->imappers, entry, daio_map_op, mgr); entry 691 sound/pci/ctxfi/ctdaio.c struct imapper *entry; entry 705 sound/pci/ctxfi/ctdaio.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 706 sound/pci/ctxfi/ctdaio.c if (!entry) { entry 710 sound/pci/ctxfi/ctdaio.c entry->slot = entry->addr = entry->next = entry->user = 0; entry 711 sound/pci/ctxfi/ctdaio.c list_add(&entry->list, &daio_mgr->imappers); entry 712 sound/pci/ctxfi/ctdaio.c daio_mgr->init_imap = entry; entry 112 sound/pci/ctxfi/ctdaio.h int (*imap_add)(struct daio_mgr *mgr, struct imapper *entry); entry 113 sound/pci/ctxfi/ctdaio.h int (*imap_delete)(struct daio_mgr *mgr, struct imapper *entry); entry 18 sound/pci/ctxfi/ctimap.c int input_mapper_add(struct list_head *mappers, struct imapper *entry, entry 27 sound/pci/ctxfi/ctimap.c entry->next = entry->addr; entry 28 sound/pci/ctxfi/ctimap.c map_op(data, entry); entry 29 sound/pci/ctxfi/ctimap.c list_add(&entry->list, head); entry 35 sound/pci/ctxfi/ctimap.c if (pos_ent->slot > entry->slot) { entry 46 sound/pci/ctxfi/ctimap.c __list_add(&entry->list, pos->prev, pos); entry 50 sound/pci/ctxfi/ctimap.c list_add_tail(&entry->list, head); entry 56 sound/pci/ctxfi/ctimap.c entry->next = pos_ent->addr; entry 57 sound/pci/ctxfi/ctimap.c map_op(data, entry); entry 58 sound/pci/ctxfi/ctimap.c pre_ent->next = entry->addr; entry 64 sound/pci/ctxfi/ctimap.c int input_mapper_delete(struct list_head *mappers, struct imapper *entry, entry 75 sound/pci/ctxfi/ctimap.c pre = (entry->list.prev == head) ? head->prev : entry->list.prev; entry 76 sound/pci/ctxfi/ctimap.c next = (entry->list.next == head) ? head->next : entry->list.next; entry 78 sound/pci/ctxfi/ctimap.c if (pre == &entry->list) { entry 80 sound/pci/ctxfi/ctimap.c entry->next = entry->addr = entry->user = entry->slot = 0; entry 81 sound/pci/ctxfi/ctimap.c map_op(data, entry); entry 82 sound/pci/ctxfi/ctimap.c list_del(&entry->list); entry 91 sound/pci/ctxfi/ctimap.c list_del(&entry->list); entry 98 sound/pci/ctxfi/ctimap.c struct imapper *entry; entry 104 sound/pci/ctxfi/ctimap.c entry = list_entry(pos, struct imapper, list); entry 105 sound/pci/ctxfi/ctimap.c kfree(entry); entry 28 sound/pci/ctxfi/ctimap.h int input_mapper_add(struct list_head *mappers, struct imapper *entry, entry 31 sound/pci/ctxfi/ctimap.h int input_mapper_delete(struct list_head *mappers, struct imapper *entry, entry 619 sound/pci/ctxfi/ctsrc.c struct imapper *entry; entry 628 sound/pci/ctxfi/ctsrc.c entry = &srcimp->imappers[i]; entry 629 sound/pci/ctxfi/ctsrc.c entry->slot = input->ops->output_slot(input); entry 630 sound/pci/ctxfi/ctsrc.c entry->user = src->rsc.ops->index(&src->rsc); entry 631 sound/pci/ctxfi/ctsrc.c entry->addr = srcimp->rsc.ops->index(&srcimp->rsc); entry 632 sound/pci/ctxfi/ctsrc.c srcimp->mgr->imap_add(srcimp->mgr, entry); entry 777 sound/pci/ctxfi/ctsrc.c static int srcimp_map_op(void *data, struct imapper *entry) entry 782 sound/pci/ctxfi/ctsrc.c hw->srcimp_mgr_set_imaparc(mgr->ctrl_blk, entry->slot); entry 783 sound/pci/ctxfi/ctsrc.c hw->srcimp_mgr_set_imapuser(mgr->ctrl_blk, entry->user); entry 784 sound/pci/ctxfi/ctsrc.c hw->srcimp_mgr_set_imapnxt(mgr->ctrl_blk, entry->next); entry 785 sound/pci/ctxfi/ctsrc.c hw->srcimp_mgr_set_imapaddr(mgr->ctrl_blk, entry->addr); entry 791 sound/pci/ctxfi/ctsrc.c static int srcimp_imap_add(struct srcimp_mgr *mgr, struct imapper *entry) entry 797 sound/pci/ctxfi/ctsrc.c if ((0 == entry->addr) && (mgr->init_imap_added)) { entry 802 sound/pci/ctxfi/ctsrc.c err = input_mapper_add(&mgr->imappers, entry, srcimp_map_op, mgr); entry 808 sound/pci/ctxfi/ctsrc.c static int srcimp_imap_delete(struct srcimp_mgr *mgr, struct imapper *entry) entry 814 sound/pci/ctxfi/ctsrc.c err = input_mapper_delete(&mgr->imappers, entry, srcimp_map_op, mgr); entry 829 sound/pci/ctxfi/ctsrc.c struct imapper *entry; entry 843 sound/pci/ctxfi/ctsrc.c entry = kzalloc(sizeof(*entry), GFP_KERNEL); entry 844 sound/pci/ctxfi/ctsrc.c if (!entry) { entry 848 sound/pci/ctxfi/ctsrc.c entry->slot = entry->addr = entry->next = entry->user = 0; entry 849 sound/pci/ctxfi/ctsrc.c list_add(&entry->list, &srcimp_mgr->imappers); entry 850 sound/pci/ctxfi/ctsrc.c srcimp_mgr->init_imap = entry; entry 137 sound/pci/ctxfi/ctsrc.h int (*imap_add)(struct srcimp_mgr *mgr, struct imapper *entry); entry 138 sound/pci/ctxfi/ctsrc.h int (*imap_delete)(struct srcimp_mgr *mgr, struct imapper *entry); entry 32 sound/pci/ctxfi/ctvmem.c struct ct_vm_block *block = NULL, *entry; entry 44 sound/pci/ctxfi/ctvmem.c entry = list_entry(pos, struct ct_vm_block, list); entry 45 sound/pci/ctxfi/ctvmem.c if (entry->size >= size) entry 51 sound/pci/ctxfi/ctvmem.c if (entry->size == size) { entry 53 sound/pci/ctxfi/ctvmem.c list_move(&entry->list, &vm->used); entry 55 sound/pci/ctxfi/ctvmem.c block = entry; entry 63 sound/pci/ctxfi/ctvmem.c block->addr = entry->addr; entry 66 sound/pci/ctxfi/ctvmem.c entry->addr += size; entry 67 sound/pci/ctxfi/ctvmem.c entry->size -= size; entry 77 sound/pci/ctxfi/ctvmem.c struct ct_vm_block *entry, *pre_ent; entry 87 sound/pci/ctxfi/ctvmem.c entry = list_entry(pos, struct ct_vm_block, list); entry 88 sound/pci/ctxfi/ctvmem.c if (entry->addr >= (block->addr + block->size)) entry 93 sound/pci/ctxfi/ctvmem.c entry = block; entry 95 sound/pci/ctxfi/ctvmem.c if ((block->addr + block->size) == entry->addr) { entry 96 sound/pci/ctxfi/ctvmem.c entry->addr = block->addr; entry 97 sound/pci/ctxfi/ctvmem.c entry->size += block->size; entry 101 sound/pci/ctxfi/ctvmem.c entry = block; entry 105 sound/pci/ctxfi/ctvmem.c pos = &entry->list; entry 108 sound/pci/ctxfi/ctvmem.c entry = list_entry(pos, struct ct_vm_block, list); entry 110 sound/pci/ctxfi/ctvmem.c if ((pre_ent->addr + pre_ent->size) > entry->addr) entry 113 sound/pci/ctxfi/ctvmem.c pre_ent->size += entry->size; entry 115 sound/pci/ctxfi/ctvmem.c kfree(entry); entry 219 sound/pci/ctxfi/ctvmem.c struct ct_vm_block *entry; entry 225 sound/pci/ctxfi/ctvmem.c entry = list_entry(pos, struct ct_vm_block, list); entry 226 sound/pci/ctxfi/ctvmem.c kfree(entry); entry 231 sound/pci/ctxfi/ctvmem.c entry = list_entry(pos, struct ct_vm_block, list); entry 232 sound/pci/ctxfi/ctvmem.c kfree(entry); entry 1008 sound/pci/emu10k1/emu10k1x.c static void snd_emu10k1x_proc_reg_read(struct snd_info_entry *entry, entry 1011 sound/pci/emu10k1/emu10k1x.c struct emu10k1x *emu = entry->private_data; entry 1036 sound/pci/emu10k1/emu10k1x.c static void snd_emu10k1x_proc_reg_write(struct snd_info_entry *entry, entry 1039 sound/pci/emu10k1/emu10k1x.c struct emu10k1x *emu = entry->private_data; entry 66 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_read(struct snd_info_entry *entry, entry 172 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 223 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_spdif_read(struct snd_info_entry *entry, entry 226 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 262 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_rates_read(struct snd_info_entry *entry, entry 266 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 276 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_acode_read(struct snd_info_entry *entry, entry 280 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 319 sound/pci/emu10k1/emuproc.c static ssize_t snd_emu10k1_fx8010_read(struct snd_info_entry *entry, entry 324 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 331 sound/pci/emu10k1/emuproc.c if (!strcmp(entry->name, "fx8010_tram_addr")) { entry 334 sound/pci/emu10k1/emuproc.c } else if (!strcmp(entry->name, "fx8010_tram_data")) { entry 336 sound/pci/emu10k1/emuproc.c } else if (!strcmp(entry->name, "fx8010_code")) { entry 362 sound/pci/emu10k1/emuproc.c static void snd_emu10k1_proc_voices_read(struct snd_info_entry *entry, entry 365 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 383 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_emu1010_reg_read(struct snd_info_entry *entry, entry 386 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 397 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_io_reg_read(struct snd_info_entry *entry, entry 400 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 413 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_io_reg_write(struct snd_info_entry *entry, entry 416 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 466 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read(struct snd_info_entry *entry, entry 469 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 490 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_write(struct snd_info_entry *entry, entry 493 sound/pci/emu10k1/emuproc.c struct snd_emu10k1 *emu = entry->private_data; entry 504 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_write00(struct snd_info_entry *entry, entry 507 sound/pci/emu10k1/emuproc.c snd_emu_proc_ptr_reg_write(entry, buffer, 0); entry 510 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_write20(struct snd_info_entry *entry, entry 513 sound/pci/emu10k1/emuproc.c snd_emu_proc_ptr_reg_write(entry, buffer, 0x20); entry 517 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read00a(struct snd_info_entry *entry, entry 520 sound/pci/emu10k1/emuproc.c snd_emu_proc_ptr_reg_read(entry, buffer, 0, 0, 0x40, 64); entry 523 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read00b(struct snd_info_entry *entry, entry 526 sound/pci/emu10k1/emuproc.c snd_emu_proc_ptr_reg_read(entry, buffer, 0, 0x40, 0x40, 64); entry 529 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read20a(struct snd_info_entry *entry, entry 532 sound/pci/emu10k1/emuproc.c snd_emu_proc_ptr_reg_read(entry, buffer, 0x20, 0, 0x40, 4); entry 535 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read20b(struct snd_info_entry *entry, entry 538 sound/pci/emu10k1/emuproc.c snd_emu_proc_ptr_reg_read(entry, buffer, 0x20, 0x40, 0x40, 4); entry 541 sound/pci/emu10k1/emuproc.c static void snd_emu_proc_ptr_reg_read20c(struct snd_info_entry *entry, entry 544 sound/pci/emu10k1/emuproc.c snd_emu_proc_ptr_reg_read(entry, buffer, 0x20, 0x80, 0x20, 4); entry 554 sound/pci/emu10k1/emuproc.c struct snd_info_entry *entry; entry 592 sound/pci/emu10k1/emuproc.c if (! snd_card_proc_new(emu->card, "fx8010_gpr", &entry)) { entry 593 sound/pci/emu10k1/emuproc.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 594 sound/pci/emu10k1/emuproc.c entry->private_data = emu; entry 595 sound/pci/emu10k1/emuproc.c entry->mode = S_IFREG | 0444 /*| S_IWUSR*/; entry 596 sound/pci/emu10k1/emuproc.c entry->size = emu->audigy ? A_TOTAL_SIZE_GPR : TOTAL_SIZE_GPR; entry 597 sound/pci/emu10k1/emuproc.c entry->c.ops = &snd_emu10k1_proc_ops_fx8010; entry 599 sound/pci/emu10k1/emuproc.c if (! snd_card_proc_new(emu->card, "fx8010_tram_data", &entry)) { entry 600 sound/pci/emu10k1/emuproc.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 601 sound/pci/emu10k1/emuproc.c entry->private_data = emu; entry 602 sound/pci/emu10k1/emuproc.c entry->mode = S_IFREG | 0444 /*| S_IWUSR*/; entry 603 sound/pci/emu10k1/emuproc.c entry->size = emu->audigy ? A_TOTAL_SIZE_TANKMEM_DATA : TOTAL_SIZE_TANKMEM_DATA ; entry 604 sound/pci/emu10k1/emuproc.c entry->c.ops = &snd_emu10k1_proc_ops_fx8010; entry 606 sound/pci/emu10k1/emuproc.c if (! snd_card_proc_new(emu->card, "fx8010_tram_addr", &entry)) { entry 607 sound/pci/emu10k1/emuproc.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 608 sound/pci/emu10k1/emuproc.c entry->private_data = emu; entry 609 sound/pci/emu10k1/emuproc.c entry->mode = S_IFREG | 0444 /*| S_IWUSR*/; entry 610 sound/pci/emu10k1/emuproc.c entry->size = emu->audigy ? A_TOTAL_SIZE_TANKMEM_ADDR : TOTAL_SIZE_TANKMEM_ADDR ; entry 611 sound/pci/emu10k1/emuproc.c entry->c.ops = &snd_emu10k1_proc_ops_fx8010; entry 613 sound/pci/emu10k1/emuproc.c if (! snd_card_proc_new(emu->card, "fx8010_code", &entry)) { entry 614 sound/pci/emu10k1/emuproc.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 615 sound/pci/emu10k1/emuproc.c entry->private_data = emu; entry 616 sound/pci/emu10k1/emuproc.c entry->mode = S_IFREG | 0444 /*| S_IWUSR*/; entry 617 sound/pci/emu10k1/emuproc.c entry->size = emu->audigy ? A_TOTAL_SIZE_CODE : TOTAL_SIZE_CODE; entry 618 sound/pci/emu10k1/emuproc.c entry->c.ops = &snd_emu10k1_proc_ops_fx8010; entry 1870 sound/pci/ens1370.c static void snd_ensoniq_proc_read(struct snd_info_entry *entry, entry 1873 sound/pci/ens1370.c struct ensoniq *ensoniq = entry->private_data; entry 738 sound/pci/hda/hda_proc.c static void print_codec_info(struct snd_info_entry *entry, entry 741 sound/pci/hda/hda_proc.c struct hda_codec *codec = entry->private_data; entry 483 sound/pci/hda/patch_hdmi.c static void print_eld_info(struct snd_info_entry *entry, entry 486 sound/pci/hda/patch_hdmi.c struct hdmi_spec_per_pin *per_pin = entry->private_data; entry 493 sound/pci/hda/patch_hdmi.c static void write_eld_info(struct snd_info_entry *entry, entry 496 sound/pci/hda/patch_hdmi.c struct hdmi_spec_per_pin *per_pin = entry->private_data; entry 507 sound/pci/hda/patch_hdmi.c struct snd_info_entry *entry; entry 511 sound/pci/hda/patch_hdmi.c err = snd_card_proc_new(codec->card, name, &entry); entry 515 sound/pci/hda/patch_hdmi.c snd_info_set_text_ops(entry, per_pin, print_eld_info); entry 516 sound/pci/hda/patch_hdmi.c entry->c.text.write = write_eld_info; entry 517 sound/pci/hda/patch_hdmi.c entry->mode |= 0200; entry 518 sound/pci/hda/patch_hdmi.c per_pin->proc_entry = entry; entry 1550 sound/pci/ice1712/ice1712.c static void snd_ice1712_proc_read(struct snd_info_entry *entry, entry 1553 sound/pci/ice1712/ice1712.c struct snd_ice1712 *ice = entry->private_data; entry 1519 sound/pci/ice1712/ice1724.c static void snd_vt1724_proc_read(struct snd_info_entry *entry, entry 1522 sound/pci/ice1712/ice1724.c struct snd_ice1712 *ice = entry->private_data; entry 618 sound/pci/ice1712/pontis.c static void wm_proc_regs_write(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 620 sound/pci/ice1712/pontis.c struct snd_ice1712 *ice = entry->private_data; entry 633 sound/pci/ice1712/pontis.c static void wm_proc_regs_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 635 sound/pci/ice1712/pontis.c struct snd_ice1712 *ice = entry->private_data; entry 652 sound/pci/ice1712/pontis.c static void cs_proc_regs_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 654 sound/pci/ice1712/pontis.c struct snd_ice1712 *ice = entry->private_data; entry 625 sound/pci/ice1712/prodigy192.c static void stac9460_proc_regs_read(struct snd_info_entry *entry, entry 628 sound/pci/ice1712/prodigy192.c struct snd_ice1712 *ice = entry->private_data; entry 861 sound/pci/ice1712/prodigy_hifi.c static void wm_proc_regs_write(struct snd_info_entry *entry, entry 864 sound/pci/ice1712/prodigy_hifi.c struct snd_ice1712 *ice = entry->private_data; entry 877 sound/pci/ice1712/prodigy_hifi.c static void wm_proc_regs_read(struct snd_info_entry *entry, entry 880 sound/pci/ice1712/prodigy_hifi.c struct snd_ice1712 *ice = entry->private_data; entry 474 sound/pci/ice1712/quartet.c static void proc_regs_read(struct snd_info_entry *entry, entry 477 sound/pci/ice1712/quartet.c struct snd_ice1712 *ice = entry->private_data; entry 2815 sound/pci/intel8x0.c static void snd_intel8x0_proc_read(struct snd_info_entry * entry, entry 2818 sound/pci/intel8x0.c struct intel8x0 *chip = entry->private_data; entry 1050 sound/pci/intel8x0m.c static void snd_intel8x0m_proc_read(struct snd_info_entry * entry, entry 1053 sound/pci/intel8x0m.c struct intel8x0m *chip = entry->private_data; entry 2049 sound/pci/korg1212/korg1212.c static void snd_korg1212_proc_read(struct snd_info_entry *entry, entry 2053 sound/pci/korg1212/korg1212.c struct snd_korg1212 *korg1212 = entry->private_data; entry 97 sound/pci/lola/lola_proc.c static void lola_proc_codec_read(struct snd_info_entry *entry, entry 100 sound/pci/lola/lola_proc.c struct lola *chip = entry->private_data; entry 134 sound/pci/lola/lola_proc.c static void lola_proc_codec_rw_write(struct snd_info_entry *entry, entry 137 sound/pci/lola/lola_proc.c struct lola *chip = entry->private_data; entry 149 sound/pci/lola/lola_proc.c static void lola_proc_codec_rw_read(struct snd_info_entry *entry, entry 152 sound/pci/lola/lola_proc.c struct lola *chip = entry->private_data; entry 159 sound/pci/lola/lola_proc.c static void lola_proc_regs_read(struct snd_info_entry *entry, entry 162 sound/pci/lola/lola_proc.c struct lola *chip = entry->private_data; entry 904 sound/pci/lx6464es/lx6464es.c static void lx_proc_levels_read(struct snd_info_entry *entry, entry 910 sound/pci/lx6464es/lx6464es.c struct lx6464es *chip = entry->private_data; entry 1131 sound/pci/mixart/mixart.c static ssize_t snd_mixart_BA0_read(struct snd_info_entry *entry, entry 1136 sound/pci/mixart/mixart.c struct mixart_mgr *mgr = entry->private_data; entry 1147 sound/pci/mixart/mixart.c static ssize_t snd_mixart_BA1_read(struct snd_info_entry *entry, entry 1152 sound/pci/mixart/mixart.c struct mixart_mgr *mgr = entry->private_data; entry 1169 sound/pci/mixart/mixart.c static void snd_mixart_proc_read(struct snd_info_entry *entry, entry 1172 sound/pci/mixart/mixart.c struct snd_mixart *chip = entry->private_data; entry 1207 sound/pci/mixart/mixart.c struct snd_info_entry *entry; entry 1213 sound/pci/mixart/mixart.c if (! snd_card_proc_new(chip->card, "mixart_BA0", &entry)) { entry 1214 sound/pci/mixart/mixart.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 1215 sound/pci/mixart/mixart.c entry->private_data = chip->mgr; entry 1216 sound/pci/mixart/mixart.c entry->c.ops = &snd_mixart_proc_ops_BA0; entry 1217 sound/pci/mixart/mixart.c entry->size = MIXART_BA0_SIZE; entry 1219 sound/pci/mixart/mixart.c if (! snd_card_proc_new(chip->card, "mixart_BA1", &entry)) { entry 1220 sound/pci/mixart/mixart.c entry->content = SNDRV_INFO_CONTENT_DATA; entry 1221 sound/pci/mixart/mixart.c entry->private_data = chip->mgr; entry 1222 sound/pci/mixart/mixart.c entry->c.ops = &snd_mixart_proc_ops_BA1; entry 1223 sound/pci/mixart/mixart.c entry->size = MIXART_BA1_SIZE; entry 187 sound/pci/oxygen/oxygen_lib.c static void oxygen_proc_read(struct snd_info_entry *entry, entry 190 sound/pci/oxygen/oxygen_lib.c struct oxygen *chip = entry->private_data; entry 1234 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_info(struct snd_info_entry *entry, entry 1237 sound/pci/pcxhr/pcxhr.c struct snd_pcxhr *chip = entry->private_data; entry 1305 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_sync(struct snd_info_entry *entry, entry 1308 sound/pci/pcxhr/pcxhr.c struct snd_pcxhr *chip = entry->private_data; entry 1347 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_gpio_read(struct snd_info_entry *entry, entry 1350 sound/pci/pcxhr/pcxhr.c struct snd_pcxhr *chip = entry->private_data; entry 1364 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_gpo_write(struct snd_info_entry *entry, entry 1367 sound/pci/pcxhr/pcxhr.c struct snd_pcxhr *chip = entry->private_data; entry 1392 sound/pci/pcxhr/pcxhr.c static void pcxhr_proc_ltc(struct snd_info_entry *entry, entry 1395 sound/pci/pcxhr/pcxhr.c struct snd_pcxhr *chip = entry->private_data; entry 1890 sound/pci/riptide/riptide.c snd_riptide_proc_read(struct snd_info_entry *entry, entry 1893 sound/pci/riptide/riptide.c struct snd_riptide *chip = entry->private_data; entry 1461 sound/pci/rme32.c snd_rme32_proc_read(struct snd_info_entry * entry, struct snd_info_buffer *buffer) entry 1464 sound/pci/rme32.c struct rme32 *rme32 = (struct rme32 *) entry->private_data; entry 1727 sound/pci/rme96.c snd_rme96_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 1730 sound/pci/rme96.c struct rme96 *rme96 = entry->private_data; entry 3320 sound/pci/rme9652/hdsp.c snd_hdsp_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 3322 sound/pci/rme9652/hdsp.c struct hdsp *hdsp = entry->private_data; entry 4748 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_tco(struct snd_info_entry *entry, entry 4751 sound/pci/rme9652/hdspm.c struct hdspm *hdspm = entry->private_data; entry 4860 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_madi(struct snd_info_entry *entry, entry 4863 sound/pci/rme9652/hdspm.c struct hdspm *hdspm = entry->private_data; entry 5016 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_tco(entry, buffer); entry 5022 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_aes32(struct snd_info_entry * entry, entry 5025 sound/pci/rme9652/hdspm.c struct hdspm *hdspm = entry->private_data; entry 5169 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_tco(entry, buffer); entry 5175 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_raydat(struct snd_info_entry *entry, entry 5178 sound/pci/rme9652/hdspm.c struct hdspm *hdspm = entry->private_data; entry 5233 sound/pci/rme9652/hdspm.c snd_hdspm_proc_read_debug(struct snd_info_entry *entry, entry 5236 sound/pci/rme9652/hdspm.c struct hdspm *hdspm = entry->private_data; entry 5250 sound/pci/rme9652/hdspm.c static void snd_hdspm_proc_ports_in(struct snd_info_entry *entry, entry 5253 sound/pci/rme9652/hdspm.c struct hdspm *hdspm = entry->private_data; entry 5263 sound/pci/rme9652/hdspm.c static void snd_hdspm_proc_ports_out(struct snd_info_entry *entry, entry 5266 sound/pci/rme9652/hdspm.c struct hdspm *hdspm = entry->private_data; entry 1552 sound/pci/rme9652/rme9652.c snd_rme9652_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 1554 sound/pci/rme9652/rme9652.c struct snd_rme9652 *rme9652 = (struct snd_rme9652 *) entry->private_data; entry 1128 sound/pci/sonicvibes.c static void snd_sonicvibes_proc_read(struct snd_info_entry *entry, entry 1131 sound/pci/sonicvibes.c struct sonicvibes *sonic = entry->private_data; entry 3272 sound/pci/trident/trident_main.c static void snd_trident_proc_read(struct snd_info_entry *entry, entry 3275 sound/pci/trident/trident_main.c struct snd_trident *trident = entry->private_data; entry 2119 sound/pci/via82xx.c static void snd_via82xx_proc_read(struct snd_info_entry *entry, entry 2122 sound/pci/via82xx.c struct via82xx *chip = entry->private_data; entry 911 sound/pci/via82xx_modem.c static void snd_via82xx_proc_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 913 sound/pci/via82xx_modem.c struct via82xx_modem *chip = entry->private_data; entry 1961 sound/pci/ymfpci/ymfpci_main.c static void snd_ymfpci_proc_read(struct snd_info_entry *entry, entry 1964 sound/pci/ymfpci/ymfpci_main.c struct snd_ymfpci *chip = entry->private_data; entry 124 sound/pcmcia/pdaudiocf/pdaudiocf_core.c static void pdacf_proc_read(struct snd_info_entry * entry, entry 127 sound/pcmcia/pdaudiocf/pdaudiocf_core.c struct snd_pdacf *chip = entry->private_data; entry 149 sound/soc/codecs/wm_hubs.c struct wm_hubs_dcs_cache **entry) entry 165 sound/soc/codecs/wm_hubs.c *entry = cache; entry 67 sound/soc/intel/baytrail/sst-baytrail-dsp.c template.entry = module->entry_point; entry 131 sound/soc/intel/common/sst-dsp-priv.h u32 entry; /* entry point */ entry 195 sound/soc/intel/common/sst-dsp-priv.h u32 entry; /* module entry point */ entry 495 sound/soc/intel/common/sst-firmware.c sst_module->entry = template->entry; entry 109 sound/soc/intel/haswell/sst-haswell-dsp.c template.entry = entry_point - 4; entry 1160 sound/soc/intel/haswell/sst-haswell-ipc.c map->module_entries[0].entry_point = module->entry; entry 1927 sound/soc/intel/haswell/sst-haswell-ipc.c config.map.module_entries[0].entry_point = module->entry; entry 350 sound/soc/sh/rcar/dma.c const u8 *entry = NULL; entry 358 sound/soc/sh/rcar/dma.c entry = gen2_id_table_ssiu; entry 362 sound/soc/sh/rcar/dma.c entry = gen2_id_table_scu; entry 366 sound/soc/sh/rcar/dma.c entry = gen2_id_table_cmd; entry 371 sound/soc/sh/rcar/dma.c if ((!entry) || (size <= id)) { entry 380 sound/soc/sh/rcar/dma.c return entry[id]; entry 2414 sound/sparc/dbri.c #define CS4215_SINGLE(xname, entry, shift, mask, invert) \ entry 2418 sound/sparc/dbri.c .private_value = (entry) | ((shift) << 8) | ((mask) << 16) | \ entry 2477 sound/sparc/dbri.c static void dbri_regs_read(struct snd_info_entry *entry, entry 2480 sound/sparc/dbri.c struct snd_dbri *dbri = entry->private_data; entry 2489 sound/sparc/dbri.c static void dbri_debug_read(struct snd_info_entry *entry, entry 2492 sound/sparc/dbri.c struct snd_dbri *dbri = entry->private_data; entry 15 sound/synth/emux/emux_proc.c snd_emux_proc_info_read(struct snd_info_entry *entry, entry 21 sound/synth/emux/emux_proc.c emu = entry->private_data; entry 94 sound/synth/emux/emux_proc.c struct snd_info_entry *entry; entry 98 sound/synth/emux/emux_proc.c entry = snd_info_create_card_entry(card, name, card->proc_root); entry 99 sound/synth/emux/emux_proc.c if (entry == NULL) entry 102 sound/synth/emux/emux_proc.c entry->content = SNDRV_INFO_CONTENT_TEXT; entry 103 sound/synth/emux/emux_proc.c entry->private_data = emu; entry 104 sound/synth/emux/emux_proc.c entry->c.text.read = snd_emux_proc_info_read; entry 3239 sound/usb/mixer.c static void snd_usb_mixer_proc_read(struct snd_info_entry *entry, entry 3242 sound/usb/mixer.c struct snd_usb_audio *chip = entry->private_data; entry 398 sound/usb/mixer_quirks.c static void snd_audigy2nx_proc_read(struct snd_info_entry *entry, entry 417 sound/usb/mixer_quirks.c struct usb_mixer_interface *mixer = entry->private_data; entry 33 sound/usb/proc.c static void proc_audio_usbbus_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 35 sound/usb/proc.c struct snd_usb_audio *chip = entry->private_data; entry 40 sound/usb/proc.c static void proc_audio_usbid_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 42 sound/usb/proc.c struct snd_usb_audio *chip = entry->private_data; entry 137 sound/usb/proc.c static void proc_pcm_format_read(struct snd_info_entry *entry, struct snd_info_buffer *buffer) entry 139 sound/usb/proc.c struct snd_usb_stream *stream = entry->private_data; entry 558 tools/hv/hv_kvp_daemon.c struct dirent *entry; entry 569 tools/hv/hv_kvp_daemon.c while ((entry = readdir(dir)) != NULL) { entry 574 tools/hv/hv_kvp_daemon.c KVP_NET_DIR, entry->d_name); entry 591 tools/hv/hv_kvp_daemon.c if_name = strdup(entry->d_name); entry 949 tools/hv/hv_kvp_daemon.c struct dirent *entry; entry 962 tools/hv/hv_kvp_daemon.c while ((entry = readdir(dir)) != NULL) { entry 967 tools/hv/hv_kvp_daemon.c entry->d_name); entry 992 tools/hv/hv_kvp_daemon.c if_name = entry->d_name; entry 66 tools/hv/hv_vss_daemon.c struct dirent *entry; entry 93 tools/hv/hv_vss_daemon.c while ((entry = readdir(dir)) != NULL) { entry 94 tools/hv/hv_vss_daemon.c if (strcmp(entry->d_name, ".") == 0 || entry 95 tools/hv/hv_vss_daemon.c strcmp(entry->d_name, "..") == 0) entry 99 tools/hv/hv_vss_daemon.c entry->d_name); entry 100 tools/include/linux/list.h static inline void __list_del_entry(struct list_head *entry) entry 102 tools/include/linux/list.h __list_del(entry->prev, entry->next); entry 105 tools/include/linux/list.h static inline void list_del(struct list_head *entry) entry 107 tools/include/linux/list.h __list_del(entry->prev, entry->next); entry 108 tools/include/linux/list.h entry->next = LIST_POISON1; entry 109 tools/include/linux/list.h entry->prev = LIST_POISON2; entry 112 tools/include/linux/list.h extern void __list_del_entry(struct list_head *entry); entry 113 tools/include/linux/list.h extern void list_del(struct list_head *entry); entry 143 tools/include/linux/list.h static inline void list_del_init(struct list_head *entry) entry 145 tools/include/linux/list.h __list_del_entry(entry); entry 146 tools/include/linux/list.h INIT_LIST_HEAD(entry); entry 235 tools/include/linux/list.h struct list_head *head, struct list_head *entry) entry 237 tools/include/linux/list.h struct list_head *new_first = entry->next; entry 240 tools/include/linux/list.h list->prev = entry; entry 241 tools/include/linux/list.h entry->next = list; entry 261 tools/include/linux/list.h struct list_head *head, struct list_head *entry) entry 266 tools/include/linux/list.h (head->next != entry && head != entry)) entry 268 tools/include/linux/list.h if (entry == head) entry 271 tools/include/linux/list.h __list_cut_position(list, head, entry); entry 1590 tools/include/uapi/linux/kvm.h __u16 entry; /* The index of entry in the MSI-X table */ entry 9 tools/lib/api/cpu.c char entry[PATH_MAX]; entry 15 tools/lib/api/cpu.c snprintf(entry, sizeof(entry), entry 18 tools/lib/api/cpu.c return sysfs__read_ull(entry, freq); entry 424 tools/lib/api/fs/fs.c int procfs__read_str(const char *entry, char **buf, size_t *sizep) entry 432 tools/lib/api/fs/fs.c snprintf(path, sizeof(path), "%s/%s", procfs, entry); entry 437 tools/lib/api/fs/fs.c static int sysfs__read_ull_base(const char *entry, entry 446 tools/lib/api/fs/fs.c snprintf(path, sizeof(path), "%s/%s", sysfs, entry); entry 451 tools/lib/api/fs/fs.c int sysfs__read_xll(const char *entry, unsigned long long *value) entry 453 tools/lib/api/fs/fs.c return sysfs__read_ull_base(entry, value, 16); entry 456 tools/lib/api/fs/fs.c int sysfs__read_ull(const char *entry, unsigned long long *value) entry 458 tools/lib/api/fs/fs.c return sysfs__read_ull_base(entry, value, 0); entry 461 tools/lib/api/fs/fs.c int sysfs__read_int(const char *entry, int *value) entry 469 tools/lib/api/fs/fs.c snprintf(path, sizeof(path), "%s/%s", sysfs, entry); entry 474 tools/lib/api/fs/fs.c int sysfs__read_str(const char *entry, char **buf, size_t *sizep) entry 482 tools/lib/api/fs/fs.c snprintf(path, sizeof(path), "%s/%s", sysfs, entry); entry 487 tools/lib/api/fs/fs.c int sysfs__read_bool(const char *entry, bool *value) entry 493 tools/lib/api/fs/fs.c ret = sysfs__read_str(entry, &buf, &size); entry 529 tools/lib/api/fs/fs.c int sysfs__write_int(const char *entry, int value) entry 537 tools/lib/api/fs/fs.c if (snprintf(path, sizeof(path), "%s/%s", sysfs, entry) >= PATH_MAX) entry 38 tools/lib/api/fs/fs.h int procfs__read_str(const char *entry, char **buf, size_t *sizep); entry 41 tools/lib/api/fs/fs.h int sysfs__read_int(const char *entry, int *value); entry 42 tools/lib/api/fs/fs.h int sysfs__read_ull(const char *entry, unsigned long long *value); entry 43 tools/lib/api/fs/fs.h int sysfs__read_xll(const char *entry, unsigned long long *value); entry 44 tools/lib/api/fs/fs.h int sysfs__read_str(const char *entry, char **buf, size_t *sizep); entry 45 tools/lib/api/fs/fs.h int sysfs__read_bool(const char *entry, bool *value); entry 47 tools/lib/api/fs/fs.h int sysfs__write_int(const char *entry, int value); entry 19 tools/lib/bpf/hashmap.c struct hashmap_entry *entry) entry 21 tools/lib/bpf/hashmap.c entry->next = *pprev; entry 22 tools/lib/bpf/hashmap.c *pprev = entry; entry 26 tools/lib/bpf/hashmap.c struct hashmap_entry *entry) entry 28 tools/lib/bpf/hashmap.c *pprev = entry->next; entry 29 tools/lib/bpf/hashmap.c entry->next = NULL; entry 121 tools/lib/bpf/hashmap.c struct hashmap_entry **entry) entry 134 tools/lib/bpf/hashmap.c *entry = cur; entry 146 tools/lib/bpf/hashmap.c struct hashmap_entry *entry; entry 157 tools/lib/bpf/hashmap.c hashmap_find_entry(map, key, h, NULL, &entry)) { entry 159 tools/lib/bpf/hashmap.c *old_key = entry->key; entry 161 tools/lib/bpf/hashmap.c *old_value = entry->value; entry 164 tools/lib/bpf/hashmap.c entry->key = key; entry 165 tools/lib/bpf/hashmap.c entry->value = value; entry 182 tools/lib/bpf/hashmap.c entry = malloc(sizeof(struct hashmap_entry)); entry 183 tools/lib/bpf/hashmap.c if (!entry) entry 186 tools/lib/bpf/hashmap.c entry->key = key; entry 187 tools/lib/bpf/hashmap.c entry->value = value; entry 188 tools/lib/bpf/hashmap.c hashmap_add_entry(&map->buckets[h], entry); entry 196 tools/lib/bpf/hashmap.c struct hashmap_entry *entry; entry 200 tools/lib/bpf/hashmap.c if (!hashmap_find_entry(map, key, h, NULL, &entry)) entry 204 tools/lib/bpf/hashmap.c *value = entry->value; entry 211 tools/lib/bpf/hashmap.c struct hashmap_entry **pprev, *entry; entry 215 tools/lib/bpf/hashmap.c if (!hashmap_find_entry(map, key, h, &pprev, &entry)) entry 219 tools/lib/bpf/hashmap.c *old_key = entry->key; entry 221 tools/lib/bpf/hashmap.c *old_value = entry->value; entry 223 tools/lib/bpf/hashmap.c hashmap_del_entry(pprev, entry); entry 224 tools/lib/bpf/hashmap.c free(entry); entry 3121 tools/lib/bpf/libbpf.c struct hashmap_entry *entry; entry 3176 tools/lib/bpf/libbpf.c hashmap__for_each_entry(cand_cache, entry, i) { entry 3177 tools/lib/bpf/libbpf.c bpf_core_free_cands(entry->value); entry 211 tools/objtool/elf.c struct list_head *entry, *tmp; entry 269 tools/objtool/elf.c entry = &sym->sec->symbol_list; entry 276 tools/objtool/elf.c entry = tmp; entry 285 tools/objtool/elf.c entry = tmp; entry 291 tools/objtool/elf.c list_add(&sym->list, entry); entry 71 tools/objtool/special.c static int get_alt_entry(struct elf *elf, struct special_entry *entry, entry 78 tools/objtool/special.c offset = idx * entry->size; entry 80 tools/objtool/special.c alt->group = entry->group; entry 81 tools/objtool/special.c alt->jump_or_nop = entry->jump_or_nop; entry 85 tools/objtool/special.c entry->orig_len); entry 87 tools/objtool/special.c entry->new_len); entry 90 tools/objtool/special.c if (entry->feature) { entry 94 tools/objtool/special.c entry->feature); entry 121 tools/objtool/special.c orig_rela = find_rela_by_dest(sec, offset + entry->orig); entry 123 tools/objtool/special.c WARN_FUNC("can't find orig rela", sec, offset + entry->orig); entry 128 tools/objtool/special.c sec, offset + entry->orig, orig_rela->sym->name); entry 135 tools/objtool/special.c if (!entry->group || alt->new_len) { entry 136 tools/objtool/special.c new_rela = find_rela_by_dest(sec, offset + entry->new); entry 139 tools/objtool/special.c sec, offset + entry->new); entry 161 tools/objtool/special.c struct special_entry *entry; entry 169 tools/objtool/special.c for (entry = entries; entry->sec; entry++) { entry 170 tools/objtool/special.c sec = find_section_by_name(elf, entry->sec); entry 174 tools/objtool/special.c if (sec->len % entry->size != 0) { entry 176 tools/objtool/special.c sec->name, entry->size); entry 180 tools/objtool/special.c nr_entries = sec->len / entry->size; entry 190 tools/objtool/special.c ret = get_alt_entry(elf, entry, sec, idx, alt); entry 89 tools/perf/builtin-annotate.c struct block_range *entry; entry 104 tools/perf/builtin-annotate.c entry = block_range_iter(&iter); entry 105 tools/perf/builtin-annotate.c assert(entry->is_target); entry 106 tools/perf/builtin-annotate.c entry->entry++; entry 109 tools/perf/builtin-annotate.c entry = block_range_iter(&iter); entry 111 tools/perf/builtin-annotate.c entry->coverage++; entry 112 tools/perf/builtin-annotate.c entry->sym = sym; entry 115 tools/perf/builtin-annotate.c notes->max_coverage = max(notes->max_coverage, entry->coverage); entry 122 tools/perf/builtin-annotate.c entry = block_range_iter(&iter); entry 123 tools/perf/builtin-annotate.c assert(entry->is_branch); entry 124 tools/perf/builtin-annotate.c entry->taken++; entry 126 tools/perf/builtin-annotate.c entry->pred++; entry 402 tools/perf/builtin-c2c.c int (*entry)(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp, entry 1280 tools/perf/builtin-c2c.c .entry = dcacheline_entry, entry 1288 tools/perf/builtin-c2c.c .entry = dcacheline_node_entry, entry 1296 tools/perf/builtin-c2c.c .entry = dcacheline_node_count, entry 1306 tools/perf/builtin-c2c.c .entry = offset_entry, entry 1314 tools/perf/builtin-c2c.c .entry = dcacheline_node_entry, entry 1322 tools/perf/builtin-c2c.c .entry = iaddr_entry, entry 1330 tools/perf/builtin-c2c.c .entry = tot_hitm_entry, entry 1338 tools/perf/builtin-c2c.c .entry = lcl_hitm_entry, entry 1346 tools/perf/builtin-c2c.c .entry = rmt_hitm_entry, entry 1354 tools/perf/builtin-c2c.c .entry = rmt_hitm_entry, entry 1362 tools/perf/builtin-c2c.c .entry = lcl_hitm_entry, entry 1370 tools/perf/builtin-c2c.c .entry = store_entry, entry 1378 tools/perf/builtin-c2c.c .entry = st_l1hit_entry, entry 1386 tools/perf/builtin-c2c.c .entry = st_l1miss_entry, entry 1394 tools/perf/builtin-c2c.c .entry = st_l1hit_entry, entry 1402 tools/perf/builtin-c2c.c .entry = st_l1miss_entry, entry 1410 tools/perf/builtin-c2c.c .entry = ld_fbhit_entry, entry 1418 tools/perf/builtin-c2c.c .entry = ld_l1hit_entry, entry 1426 tools/perf/builtin-c2c.c .entry = ld_l2hit_entry, entry 1434 tools/perf/builtin-c2c.c .entry = ld_llchit_entry, entry 1442 tools/perf/builtin-c2c.c .entry = rmt_hit_entry, entry 1450 tools/perf/builtin-c2c.c .entry = ld_llcmiss_entry, entry 1458 tools/perf/builtin-c2c.c .entry = tot_recs_entry, entry 1466 tools/perf/builtin-c2c.c .entry = tot_loads_entry, entry 1479 tools/perf/builtin-c2c.c .entry = percent_hitm_entry, entry 1488 tools/perf/builtin-c2c.c .entry = percent_rmt_hitm_entry, entry 1497 tools/perf/builtin-c2c.c .entry = percent_lcl_hitm_entry, entry 1506 tools/perf/builtin-c2c.c .entry = percent_stores_l1hit_entry, entry 1515 tools/perf/builtin-c2c.c .entry = percent_stores_l1miss_entry, entry 1524 tools/perf/builtin-c2c.c .entry = lcl_dram_entry, entry 1532 tools/perf/builtin-c2c.c .entry = rmt_dram_entry, entry 1540 tools/perf/builtin-c2c.c .entry = pid_entry, entry 1570 tools/perf/builtin-c2c.c .entry = node_entry, entry 1578 tools/perf/builtin-c2c.c .entry = mean_rmt_entry, entry 1586 tools/perf/builtin-c2c.c .entry = mean_lcl_entry, entry 1594 tools/perf/builtin-c2c.c .entry = mean_load_entry, entry 1602 tools/perf/builtin-c2c.c .entry = cpucnt_entry, entry 1615 tools/perf/builtin-c2c.c .entry = cl_idx_entry, entry 1623 tools/perf/builtin-c2c.c .entry = cl_idx_entry, entry 1631 tools/perf/builtin-c2c.c .entry = cl_idx_empty_entry, entry 1772 tools/perf/builtin-c2c.c fmt->entry = dim->se ? c2c_se_entry : dim->entry; entry 1640 tools/perf/builtin-diff.c fmt->entry = hpp__entry_global; entry 426 tools/perf/builtin-ftrace.c struct filter_entry *entry; entry 428 tools/perf/builtin-ftrace.c entry = malloc(sizeof(*entry) + strlen(str) + 1); entry 429 tools/perf/builtin-ftrace.c if (entry == NULL) entry 432 tools/perf/builtin-ftrace.c strcpy(entry->name, str); entry 433 tools/perf/builtin-ftrace.c list_add_tail(&entry->list, head); entry 320 tools/perf/builtin-lock.c struct list_head *entry = lockhashentry(addr); entry 323 tools/perf/builtin-lock.c list_for_each_entry(ret, entry, hash_entry) { entry 342 tools/perf/builtin-lock.c list_add(&new->hash_entry, entry); entry 2853 tools/perf/builtin-trace.c static void trace__init_bpf_map_syscall_args(struct trace *trace, int id, struct bpf_map_syscall_entry *entry) entry 2862 tools/perf/builtin-trace.c entry->string_args_len[arg] = 0; entry 2865 tools/perf/builtin-trace.c entry->string_args_len[arg] = PATH_MAX; entry 2870 tools/perf/builtin-trace.c entry->string_args_len[arg] = 0; entry 3676 tools/perf/builtin-trace.c entry->syscall = source->i; entry 3677 tools/perf/builtin-trace.c entry->stats = stats; entry 3678 tools/perf/builtin-trace.c entry->msecs = stats ? (u64)stats->n * (avg_stats(stats) / NSEC_PER_MSEC) : 0; entry 3761 tools/perf/builtin-trace.c entry->thread = rb_entry(nd, struct thread, rb_node); entry 146 tools/perf/lib/evsel.c int entry = sizeof(u64); /* value */ entry 157 tools/perf/lib/evsel.c entry += sizeof(u64); entry 164 tools/perf/lib/evsel.c size += entry * nr; entry 59 tools/perf/tests/dwarf-unwind.c static int unwind_entry(struct unwind_entry *entry, void *arg) entry 62 tools/perf/tests/dwarf-unwind.c char *symbol = entry->sym ? entry->sym->name : NULL; entry 87 tools/perf/tests/dwarf-unwind.c entry->ip); entry 93 tools/perf/tests/dwarf-unwind.c symbol, entry->ip, funcs[idx]); entry 29 tools/perf/ui/browser.h void (*write)(struct ui_browser *browser, void *entry, int row); entry 31 tools/perf/ui/browser.h bool (*filter)(struct ui_browser *browser, void *entry); entry 46 tools/perf/ui/browsers/annotate.c static bool disasm_line__filter(struct ui_browser *browser, void *entry) entry 49 tools/perf/ui/browsers/annotate.c struct annotation_line *al = list_entry(entry, struct annotation_line, node); entry 96 tools/perf/ui/browsers/annotate.c static void annotate_browser__write(struct ui_browser *browser, void *entry, int row) entry 100 tools/perf/ui/browsers/annotate.c struct annotation_line *al = list_entry(entry, struct annotation_line, node); entry 13 tools/perf/ui/browsers/header.c void *entry, int row) entry 15 tools/perf/ui/browsers/header.c char **arg = entry; entry 1113 tools/perf/ui/browsers/hists.c struct hist_entry *entry, int level, entry 1119 tools/perf/ui/browsers/hists.c u64 total = hists__total_period(entry->hists); entry 1124 tools/perf/ui/browsers/hists.c parent_total = entry->stat_acc->period; entry 1126 tools/perf/ui/browsers/hists.c parent_total = entry->stat.period; entry 1130 tools/perf/ui/browsers/hists.c &entry->sorted_chain, row, entry 1135 tools/perf/ui/browsers/hists.c &entry->sorted_chain, row, entry 1140 tools/perf/ui/browsers/hists.c &entry->sorted_chain, level, row, entry 1146 tools/perf/ui/browsers/hists.c browser->he_selection = entry; entry 1245 tools/perf/ui/browsers/hists.c struct hist_entry *entry, entry 1252 tools/perf/ui/browsers/hists.c bool use_callchain = hist_entry__has_callchains(entry) && symbol_conf.use_callchain; entry 1253 tools/perf/ui/browsers/hists.c off_t row_offset = entry->row_offset; entry 1258 tools/perf/ui/browsers/hists.c browser->he_selection = entry; entry 1259 tools/perf/ui/browsers/hists.c browser->selection = &entry->ms; entry 1263 tools/perf/ui/browsers/hists.c hist_entry__init_have_children(entry); entry 1264 tools/perf/ui/browsers/hists.c folded_sign = hist_entry__folded(entry); entry 1285 tools/perf/ui/browsers/hists.c if (perf_hpp__should_skip(fmt, entry->hists) || entry 1309 tools/perf/ui/browsers/hists.c int ret = fmt->color(fmt, &hpp, entry); entry 1310 tools/perf/ui/browsers/hists.c hist_entry__snprintf_alignment(entry, &hpp, fmt, ret); entry 1317 tools/perf/ui/browsers/hists.c hist_entry__snprintf_alignment(entry, &hpp, fmt, fmt->entry(fmt, &hpp, entry)); entry 1341 tools/perf/ui/browsers/hists.c entry, 1, row, entry 1351 tools/perf/ui/browsers/hists.c struct hist_entry *entry, entry 1359 tools/perf/ui/browsers/hists.c off_t row_offset = entry->row_offset; entry 1368 tools/perf/ui/browsers/hists.c int hierarchy_indent = (entry->hists->nr_hpp_node - 2) * HIERARCHY_INDENT; entry 1371 tools/perf/ui/browsers/hists.c browser->he_selection = entry; entry 1372 tools/perf/ui/browsers/hists.c browser->selection = &entry->ms; entry 1375 tools/perf/ui/browsers/hists.c hist_entry__init_have_children(entry); entry 1376 tools/perf/ui/browsers/hists.c folded_sign = hist_entry__folded(entry); entry 1379 tools/perf/ui/browsers/hists.c if (entry->leaf && row_offset) { entry 1395 tools/perf/ui/browsers/hists.c fmt_node = list_first_entry(&entry->hists->hpp_formats, entry 1405 tools/perf/ui/browsers/hists.c if (perf_hpp__should_skip(fmt, entry->hists) || entry 1427 tools/perf/ui/browsers/hists.c int ret = fmt->color(fmt, &hpp, entry); entry 1428 tools/perf/ui/browsers/hists.c hist_entry__snprintf_alignment(entry, &hpp, fmt, ret); entry 1435 tools/perf/ui/browsers/hists.c int ret = fmt->entry(fmt, &hpp, entry); entry 1436 tools/perf/ui/browsers/hists.c hist_entry__snprintf_alignment(entry, &hpp, fmt, ret); entry 1463 tools/perf/ui/browsers/hists.c perf_hpp_list__for_each_format(entry->hpp_list, fmt) { entry 1479 tools/perf/ui/browsers/hists.c width -= fmt->color(fmt, &hpp, entry); entry 1483 tools/perf/ui/browsers/hists.c width -= fmt->entry(fmt, &hpp, entry); entry 1502 tools/perf/ui/browsers/hists.c if (entry->leaf && folded_sign == '-' && row != browser->b.rows) { entry 1507 tools/perf/ui/browsers/hists.c printed += hist_browser__show_callchain(browser, entry, entry 2024 tools/perf/ui/browsers/hists.c ret = fmt->entry(fmt, &hpp, he); entry 2069 tools/perf/ui/browsers/hists.c ret = fmt->entry(fmt, &hpp, he); entry 2080 tools/perf/ui/browsers/hists.c ret = fmt->entry(fmt, &hpp, he); entry 3221 tools/perf/ui/browsers/hists.c void *entry, int row) entry 3225 tools/perf/ui/browsers/hists.c struct evsel *evsel = list_entry(entry, struct evsel, core.node); entry 3362 tools/perf/ui/browsers/hists.c void *entry) entry 3364 tools/perf/ui/browsers/hists.c struct evsel *evsel = list_entry(entry, struct evsel, core.node); entry 382 tools/perf/ui/gtk/hists.c fmt->entry(fmt, &hpp, h); entry 444 tools/perf/ui/gtk/hists.c fmt->entry(fmt, hpp, he); entry 457 tools/perf/ui/gtk/hists.c ret = fmt->entry(fmt, hpp, he); entry 404 tools/perf/ui/hist.c .entry = hpp__entry_ ## _fn, \ entry 418 tools/perf/ui/hist.c .entry = hpp__entry_ ## _fn, \ entry 431 tools/perf/ui/hist.c .entry = hpp__entry_ ## _fn, \ entry 578 tools/perf/ui/hist.c if (!fmt->entry && !fmt->color) entry 442 tools/perf/ui/stdio/hist.c ret = fmt->entry(fmt, hpp, he); entry 492 tools/perf/ui/stdio/hist.c ret = fmt->entry(fmt, hpp, he); entry 516 tools/perf/ui/stdio/hist.c fmt->entry(fmt, hpp, he); entry 16 tools/perf/ui/tui/util.c void *entry, int row) entry 18 tools/perf/ui/tui/util.c char **arg = entry; entry 1319 tools/perf/util/annotate.c p = 100 *(double)br->entry / branch->coverage; entry 1410 tools/perf/util/auxtrace.c struct auxtrace_cache_entry *entry; entry 1418 tools/perf/util/auxtrace.c hlist_for_each_entry_safe(entry, tmp, &c->hashtable[i], hash) { entry 1419 tools/perf/util/auxtrace.c hlist_del(&entry->hash); entry 1420 tools/perf/util/auxtrace.c auxtrace_cache__free_entry(c, entry); entry 1443 tools/perf/util/auxtrace.c void *entry) entry 1445 tools/perf/util/auxtrace.c free(entry); entry 1449 tools/perf/util/auxtrace.c struct auxtrace_cache_entry *entry) entry 1454 tools/perf/util/auxtrace.c entry->key = key; entry 1455 tools/perf/util/auxtrace.c hlist_add_head(&entry->hash, &c->hashtable[hash_32(key, c->bits)]); entry 1462 tools/perf/util/auxtrace.c struct auxtrace_cache_entry *entry; entry 1469 tools/perf/util/auxtrace.c hlist_for_each_entry(entry, hlist, hash) { entry 1470 tools/perf/util/auxtrace.c if (entry->key == key) entry 1471 tools/perf/util/auxtrace.c return entry; entry 489 tools/perf/util/auxtrace.h void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry); entry 491 tools/perf/util/auxtrace.h struct auxtrace_cache_entry *entry); entry 23 tools/perf/util/block-range.c struct block_range *entry = rb_entry(rb, struct block_range, node); entry 25 tools/perf/util/block-range.c assert(old < entry->start); entry 26 tools/perf/util/block-range.c assert(entry->start <= entry->end); /* single instruction block; jump to a jump */ entry 28 tools/perf/util/block-range.c old = entry->end; entry 37 tools/perf/util/block-range.c struct block_range *entry; entry 41 tools/perf/util/block-range.c entry = rb_entry(parent, struct block_range, node); entry 43 tools/perf/util/block-range.c if (addr < entry->start) entry 45 tools/perf/util/block-range.c else if (addr > entry->end) entry 48 tools/perf/util/block-range.c return entry; entry 85 tools/perf/util/block-range.c struct block_range *next, *entry = NULL; entry 90 tools/perf/util/block-range.c entry = rb_entry(parent, struct block_range, node); entry 92 tools/perf/util/block-range.c if (start < entry->start) entry 94 tools/perf/util/block-range.c else if (start > entry->end) entry 105 tools/perf/util/block-range.c if (!entry) /* tree empty */ entry 112 tools/perf/util/block-range.c if (entry->end < start) { entry 143 tools/perf/util/block-range.c entry = malloc(sizeof(struct block_range)); entry 144 tools/perf/util/block-range.c if (!entry) entry 147 tools/perf/util/block-range.c *entry = (struct block_range){ entry 154 tools/perf/util/block-range.c rb_link_node(&entry->node, parent, p); entry 155 tools/perf/util/block-range.c rb_insert_color(&entry->node, &block_ranges.root); entry 158 tools/perf/util/block-range.c iter.start = entry; entry 159 tools/perf/util/block-range.c iter.end = entry; entry 166 tools/perf/util/block-range.c if (entry->start < start) { /* split: [e->start...][start...] */ entry 172 tools/perf/util/block-range.c .start = entry->start, entry 174 tools/perf/util/block-range.c .is_target = entry->is_target, entry 177 tools/perf/util/block-range.c .coverage = entry->coverage, entry 178 tools/perf/util/block-range.c .entry = entry->entry, entry 181 tools/perf/util/block-range.c entry->start = start; entry 182 tools/perf/util/block-range.c entry->is_target = 1; entry 183 tools/perf/util/block-range.c entry->entry = 0; entry 185 tools/perf/util/block-range.c rb_link_left_of_node(&head->node, &entry->node); entry 189 tools/perf/util/block-range.c } else if (entry->start == start) entry 190 tools/perf/util/block-range.c entry->is_target = 1; entry 192 tools/perf/util/block-range.c iter.start = entry; entry 199 tools/perf/util/block-range.c entry = iter.start; entry 204 tools/perf/util/block-range.c if (end < entry->end) { /* split: [...end][...e->end] */ entry 211 tools/perf/util/block-range.c .end = entry->end, entry 213 tools/perf/util/block-range.c .is_branch = entry->is_branch, entry 215 tools/perf/util/block-range.c .coverage = entry->coverage, entry 216 tools/perf/util/block-range.c .taken = entry->taken, entry 217 tools/perf/util/block-range.c .pred = entry->pred, entry 220 tools/perf/util/block-range.c entry->end = end; entry 221 tools/perf/util/block-range.c entry->is_branch = 1; entry 222 tools/perf/util/block-range.c entry->taken = 0; entry 223 tools/perf/util/block-range.c entry->pred = 0; entry 225 tools/perf/util/block-range.c rb_link_right_of_node(&tail->node, &entry->node); entry 229 tools/perf/util/block-range.c iter.end = entry; entry 236 tools/perf/util/block-range.c if (end == entry->end) { entry 237 tools/perf/util/block-range.c entry->is_branch = 1; entry 238 tools/perf/util/block-range.c iter.end = entry; entry 242 tools/perf/util/block-range.c next = block_range__next(entry); entry 257 tools/perf/util/block-range.c .start = entry->end + 1, entry 263 tools/perf/util/block-range.c rb_link_right_of_node(&tail->node, &entry->node); entry 274 tools/perf/util/block-range.c if (entry->end + 1 != next->start) { entry 280 tools/perf/util/block-range.c .start = entry->end + 1, entry 291 tools/perf/util/block-range.c entry = next; entry 33 tools/perf/util/block-range.h u64 entry; entry 1039 tools/perf/util/dwarf-aux.c Dwarf_Addr entry; entry 1044 tools/perf/util/dwarf-aux.c ret = die_entrypc(sp_die, &entry); entry 1062 tools/perf/util/dwarf-aux.c start -= entry; entry 1063 tools/perf/util/dwarf-aux.c end -= entry; entry 1099 tools/perf/util/dwarf-aux.c Dwarf_Addr entry; entry 1107 tools/perf/util/dwarf-aux.c ret = die_entrypc(sp_die, &entry); entry 1127 tools/perf/util/dwarf-aux.c start -= entry; entry 1128 tools/perf/util/dwarf-aux.c end -= entry; entry 160 tools/perf/util/evlist.c void evlist__add(struct evlist *evlist, struct evsel *entry) entry 162 tools/perf/util/evlist.c entry->evlist = evlist; entry 163 tools/perf/util/evlist.c entry->idx = evlist->core.nr_entries; entry 164 tools/perf/util/evlist.c entry->tracking = !entry->idx; entry 166 tools/perf/util/evlist.c perf_evlist__add(&evlist->core, &entry->core); entry 92 tools/perf/util/evlist.h void evlist__add(struct evlist *evlist, struct evsel *entry); entry 565 tools/perf/util/hist.c struct hist_entry *entry, entry 573 tools/perf/util/hist.c u64 period = entry->stat.period; entry 574 tools/perf/util/hist.c u64 weight = entry->stat.weight; entry 589 tools/perf/util/hist.c cmp = hist_entry__cmp(he, entry); entry 603 tools/perf/util/hist.c mem_info__zput(entry->mem_info); entry 605 tools/perf/util/hist.c block_info__zput(entry->block_info); entry 613 tools/perf/util/hist.c if (he->ms.map != entry->ms.map) { entry 615 tools/perf/util/hist.c he->ms.map = map__get(entry->ms.map); entry 628 tools/perf/util/hist.c he = hist_entry__new(entry, sample_self); entry 684 tools/perf/util/hist.c struct hist_entry entry = { entry 717 tools/perf/util/hist.c }, *he = hists__findnew_entry(hists, &entry, al, sample_self); entry 755 tools/perf/util/hist.c struct hist_entry entry = { entry 758 tools/perf/util/hist.c }, *he = hists__findnew_entry(hists, &entry, al, false); entry 264 tools/perf/util/hist.h int (*entry)(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp, entry 394 tools/perf/util/intel-pt.c struct auxtrace_cache_entry entry; entry 489 tools/perf/util/intel-pt.c err = auxtrace_cache__add(c, offset, &e->entry); entry 14 tools/perf/util/intlist.c const void *entry) entry 16 tools/perf/util/intlist.c int i = (int)((long)entry); entry 42 tools/perf/util/intlist.c static int intlist__node_cmp(struct rb_node *rb_node, const void *entry) entry 44 tools/perf/util/intlist.c int i = (int)((long)entry); entry 37 tools/perf/util/jitdump.c union jr_entry *entry; entry 2456 tools/perf/util/machine.c static int unwind_entry(struct unwind_entry *entry, void *arg) entry 2460 tools/perf/util/machine.c u64 addr = entry->ip; entry 2462 tools/perf/util/machine.c if (symbol_conf.hide_unresolved && entry->sym == NULL) entry 2465 tools/perf/util/machine.c if (append_inlines(cursor, entry->map, entry->sym, entry->ip) == 0) entry 2472 tools/perf/util/machine.c if (entry->map) entry 2473 tools/perf/util/machine.c addr = map__map_ip(entry->map, entry->ip); entry 2475 tools/perf/util/machine.c srcline = callchain_srcline(entry->map, entry->sym, addr); entry 2476 tools/perf/util/machine.c return callchain_cursor_append(cursor, entry->ip, entry 2477 tools/perf/util/machine.c entry->map, entry->sym, entry 17 tools/perf/util/mem2node.c static void phys_entry__insert(struct phys_entry *entry, struct rb_root *root) entry 27 tools/perf/util/mem2node.c if (entry->start < e->start) entry 33 tools/perf/util/mem2node.c rb_link_node(&entry->rb_node, parent, p); entry 34 tools/perf/util/mem2node.c rb_insert_color(&entry->rb_node, root); entry 38 tools/perf/util/mem2node.c phys_entry__init(struct phys_entry *entry, u64 start, u64 bsize, u64 node) entry 40 tools/perf/util/mem2node.c entry->start = start; entry 41 tools/perf/util/mem2node.c entry->end = start + bsize; entry 42 tools/perf/util/mem2node.c entry->node = node; entry 43 tools/perf/util/mem2node.c RB_CLEAR_NODE(&entry->rb_node); entry 120 tools/perf/util/mem2node.c struct phys_entry *entry; entry 125 tools/perf/util/mem2node.c entry = rb_entry(parent, struct phys_entry, rb_node); entry 126 tools/perf/util/mem2node.c if (addr < entry->start) entry 128 tools/perf/util/mem2node.c else if (addr >= entry->end) entry 134 tools/perf/util/mem2node.c entry = NULL; entry 136 tools/perf/util/mem2node.c return entry ? (int) entry->node : -1; entry 50 tools/perf/util/metricgroup.c static int metric_event_cmp(struct rb_node *rb_node, const void *entry) entry 55 tools/perf/util/metricgroup.c const struct metric_event *b = entry; entry 65 tools/perf/util/metricgroup.c const void *entry) entry 71 tools/perf/util/metricgroup.c memcpy(me, entry, sizeof(struct metric_event)); entry 72 tools/perf/util/metricgroup.c me->evsel = ((struct metric_event *)entry)->evsel; entry 227 tools/perf/util/metricgroup.c static int mep_cmp(struct rb_node *rb_node, const void *entry) entry 230 tools/perf/util/metricgroup.c struct mep *b = (struct mep *)entry; entry 236 tools/perf/util/metricgroup.c const void *entry) entry 242 tools/perf/util/metricgroup.c memcpy(me, entry, sizeof(struct mep)); entry 3177 tools/perf/util/probe-event.c struct probe_cache_entry *entry; entry 3187 tools/perf/util/probe-event.c for_each_probe_cache_entry(entry, cache) { entry 3189 tools/perf/util/probe-event.c if (!entry->pev.event || !entry->pev.group) entry 3191 tools/perf/util/probe-event.c if ((!pev->group || strglobmatch(entry->pev.group, pev->group)) && entry 3192 tools/perf/util/probe-event.c strglobmatch(entry->pev.event, pev->event)) { entry 3193 tools/perf/util/probe-event.c ret = probe_cache_entry__get_event(entry, &tmp_tevs); entry 3260 tools/perf/util/probe-event.c struct probe_cache_entry *entry; entry 3276 tools/perf/util/probe-event.c entry = probe_cache__find(cache, pev); entry 3277 tools/perf/util/probe-event.c if (!entry) { entry 3283 tools/perf/util/probe-event.c ret = strlist__nr_entries(entry->tevlist); entry 3298 tools/perf/util/probe-event.c strlist__for_each_entry(node, entry->tevlist) { entry 349 tools/perf/util/probe-file.c static void probe_cache_entry__delete(struct probe_cache_entry *entry) entry 351 tools/perf/util/probe-file.c if (entry) { entry 352 tools/perf/util/probe-file.c BUG_ON(!list_empty(&entry->node)); entry 354 tools/perf/util/probe-file.c strlist__delete(entry->tevlist); entry 355 tools/perf/util/probe-file.c clear_perf_probe_event(&entry->pev); entry 356 tools/perf/util/probe-file.c zfree(&entry->spev); entry 357 tools/perf/util/probe-file.c free(entry); entry 364 tools/perf/util/probe-file.c struct probe_cache_entry *entry = zalloc(sizeof(*entry)); entry 366 tools/perf/util/probe-file.c if (entry) { entry 367 tools/perf/util/probe-file.c INIT_LIST_HEAD(&entry->node); entry 368 tools/perf/util/probe-file.c entry->tevlist = strlist__new(NULL, NULL); entry 369 tools/perf/util/probe-file.c if (!entry->tevlist) entry 370 tools/perf/util/probe-file.c zfree(&entry); entry 372 tools/perf/util/probe-file.c entry->spev = synthesize_perf_probe_command(pev); entry 373 tools/perf/util/probe-file.c if (!entry->spev || entry 374 tools/perf/util/probe-file.c perf_probe_event__copy(&entry->pev, pev) < 0) { entry 375 tools/perf/util/probe-file.c probe_cache_entry__delete(entry); entry 381 tools/perf/util/probe-file.c return entry; entry 384 tools/perf/util/probe-file.c int probe_cache_entry__get_event(struct probe_cache_entry *entry, entry 391 tools/perf/util/probe-file.c ret = strlist__nr_entries(entry->tevlist); entry 400 tools/perf/util/probe-file.c strlist__for_each_entry(node, entry->tevlist) { entry 472 tools/perf/util/probe-file.c struct probe_cache_entry *entry = NULL; entry 494 tools/perf/util/probe-file.c entry = probe_cache_entry__new(NULL); entry 495 tools/perf/util/probe-file.c if (!entry) { entry 500 tools/perf/util/probe-file.c entry->sdt = true; entry 501 tools/perf/util/probe-file.c entry->spev = strdup(buf + 1); entry 502 tools/perf/util/probe-file.c if (entry->spev) entry 504 tools/perf/util/probe-file.c &entry->pev); entry 508 tools/perf/util/probe-file.c probe_cache_entry__delete(entry); entry 511 tools/perf/util/probe-file.c list_add_tail(&entry->node, &pcache->entries); entry 513 tools/perf/util/probe-file.c if (!entry) { entry 517 tools/perf/util/probe-file.c strlist__add(entry->tevlist, buf); entry 538 tools/perf/util/probe-file.c struct probe_cache_entry *entry, *n; entry 540 tools/perf/util/probe-file.c list_for_each_entry_safe(entry, n, &pcache->entries, node) { entry 541 tools/perf/util/probe-file.c list_del_init(&entry->node); entry 542 tools/perf/util/probe-file.c probe_cache_entry__delete(entry); entry 598 tools/perf/util/probe-file.c struct probe_cache_entry *entry = NULL; entry 604 tools/perf/util/probe-file.c for_each_probe_cache_entry(entry, pcache) { entry 606 tools/perf/util/probe-file.c if (entry->pev.event && entry 607 tools/perf/util/probe-file.c streql(entry->pev.event, pev->event) && entry 609 tools/perf/util/probe-file.c streql(entry->pev.group, pev->group))) entry 616 tools/perf/util/probe-file.c (streql(entry->pev.group, pev->group) && entry 617 tools/perf/util/probe-file.c streql(entry->pev.event, pev->event))) || entry 618 tools/perf/util/probe-file.c (!strcmp(entry->spev, cmd))) entry 621 tools/perf/util/probe-file.c entry = NULL; entry 625 tools/perf/util/probe-file.c return entry; entry 632 tools/perf/util/probe-file.c struct probe_cache_entry *entry = NULL; entry 634 tools/perf/util/probe-file.c for_each_probe_cache_entry(entry, pcache) { entry 636 tools/perf/util/probe-file.c if (streql(entry->pev.group, group) && entry 637 tools/perf/util/probe-file.c streql(entry->pev.event, event)) entry 640 tools/perf/util/probe-file.c entry = NULL; entry 643 tools/perf/util/probe-file.c return entry; entry 650 tools/perf/util/probe-file.c struct probe_cache_entry *entry = NULL; entry 660 tools/perf/util/probe-file.c entry = probe_cache__find(pcache, pev); entry 661 tools/perf/util/probe-file.c if (entry) { entry 662 tools/perf/util/probe-file.c list_del_init(&entry->node); entry 663 tools/perf/util/probe-file.c probe_cache_entry__delete(entry); entry 667 tools/perf/util/probe-file.c entry = probe_cache_entry__new(pev); entry 668 tools/perf/util/probe-file.c if (!entry) entry 678 tools/perf/util/probe-file.c strlist__add(entry->tevlist, command); entry 681 tools/perf/util/probe-file.c list_add_tail(&entry->node, &pcache->entries); entry 687 tools/perf/util/probe-file.c probe_cache_entry__delete(entry); entry 819 tools/perf/util/probe-file.c struct probe_cache_entry *entry = NULL; entry 837 tools/perf/util/probe-file.c entry = probe_cache__find_by_name(pcache, sdtgrp, note->name); entry 838 tools/perf/util/probe-file.c if (!entry) { entry 839 tools/perf/util/probe-file.c entry = probe_cache_entry__new(NULL); entry 840 tools/perf/util/probe-file.c if (!entry) { entry 844 tools/perf/util/probe-file.c entry->sdt = true; entry 845 tools/perf/util/probe-file.c ret = asprintf(&entry->spev, "%s:%s=%s", sdtgrp, entry 849 tools/perf/util/probe-file.c entry->pev.event = strdup(note->name); entry 850 tools/perf/util/probe-file.c entry->pev.group = strdup(sdtgrp); entry 851 tools/perf/util/probe-file.c list_add_tail(&entry->node, &pcache->entries); entry 859 tools/perf/util/probe-file.c strlist__add(entry->tevlist, buf); entry 861 tools/perf/util/probe-file.c entry = NULL; entry 863 tools/perf/util/probe-file.c if (entry) { entry 864 tools/perf/util/probe-file.c list_del_init(&entry->node); entry 865 tools/perf/util/probe-file.c probe_cache_entry__delete(entry); entry 872 tools/perf/util/probe-file.c static int probe_cache_entry__write(struct probe_cache_entry *entry, int fd) entry 877 tools/perf/util/probe-file.c const char *prefix = entry->sdt ? "%" : "#"; entry 884 tools/perf/util/probe-file.c pr_debug("Writing cache: %s%s\n", prefix, entry->spev); entry 886 tools/perf/util/probe-file.c iov[1].iov_base = entry->spev; iov[1].iov_len = strlen(entry->spev); entry 892 tools/perf/util/probe-file.c strlist__for_each_entry(snode, entry->tevlist) { entry 914 tools/perf/util/probe-file.c struct probe_cache_entry *entry; entry 926 tools/perf/util/probe-file.c for_each_probe_cache_entry(entry, pcache) { entry 927 tools/perf/util/probe-file.c ret = probe_cache_entry__write(entry, pcache->fd); entry 936 tools/perf/util/probe-file.c static bool probe_cache_entry__compare(struct probe_cache_entry *entry, entry 939 tools/perf/util/probe-file.c char buf[128], *ptr = entry->spev; entry 941 tools/perf/util/probe-file.c if (entry->pev.event) { entry 942 tools/perf/util/probe-file.c snprintf(buf, 128, "%s:%s", entry->pev.group, entry->pev.event); entry 951 tools/perf/util/probe-file.c struct probe_cache_entry *entry, *tmp; entry 953 tools/perf/util/probe-file.c list_for_each_entry_safe(entry, tmp, &pcache->entries, node) { entry 954 tools/perf/util/probe-file.c if (probe_cache_entry__compare(entry, filter)) { entry 955 tools/perf/util/probe-file.c pr_info("Removed cached event: %s\n", entry->spev); entry 956 tools/perf/util/probe-file.c list_del_init(&entry->node); entry 957 tools/perf/util/probe-file.c probe_cache_entry__delete(entry); entry 966 tools/perf/util/probe-file.c struct probe_cache_entry *entry; entry 968 tools/perf/util/probe-file.c for_each_probe_cache_entry(entry, pcache) { entry 969 tools/perf/util/probe-file.c if (probe_cache_entry__compare(entry, filter)) entry 970 tools/perf/util/probe-file.c printf("%s\n", entry->spev); entry 35 tools/perf/util/probe-file.h #define for_each_probe_cache_entry(entry, pcache) \ entry 36 tools/perf/util/probe-file.h list_for_each_entry(entry, &pcache->entries, node) entry 52 tools/perf/util/probe-file.h int probe_cache_entry__get_event(struct probe_cache_entry *entry, entry 61 tools/perf/util/rb_resort.h struct __name##_sorted_entry *entry); \ entry 121 tools/perf/util/rb_resort.h struct __name##_sorted_entry *entry) entry 54 tools/perf/util/rblist.c const void *entry, entry 66 tools/perf/util/rblist.c rc = rblist->node_cmp(parent, entry); entry 78 tools/perf/util/rblist.c new_node = rblist->node_new(rblist, entry); entry 90 tools/perf/util/rblist.c struct rb_node *rblist__find(struct rblist *rblist, const void *entry) entry 92 tools/perf/util/rblist.c return __rblist__findnew(rblist, entry, false); entry 95 tools/perf/util/rblist.c struct rb_node *rblist__findnew(struct rblist *rblist, const void *entry) entry 97 tools/perf/util/rblist.c return __rblist__findnew(rblist, entry, true); entry 26 tools/perf/util/rblist.h int (*node_cmp)(struct rb_node *rbn, const void *entry); entry 36 tools/perf/util/rblist.h struct rb_node *rblist__find(struct rblist *rblist, const void *entry); entry 37 tools/perf/util/rblist.h struct rb_node *rblist__findnew(struct rblist *rblist, const void *entry); entry 1655 tools/perf/util/sort.c struct sort_entry *entry; entry 1659 tools/perf/util/sort.c #define DIM(d, n, func) [d] = { .name = n, .entry = &(func) } entry 1684 tools/perf/util/sort.c #define DIM(d, n, func) [d - __SORT_BRANCH_STACK] = { .name = n, .entry = &(func) } entry 1702 tools/perf/util/sort.c #define DIM(d, n, func) [d - __SORT_MEMORY_MODE] = { .name = n, .entry = &(func) } entry 1889 tools/perf/util/sort.c hse->se = sd->entry; entry 1890 tools/perf/util/sort.c hse->hpp.name = sd->entry->se_header; entry 1893 tools/perf/util/sort.c hse->hpp.entry = __sort__hpp_entry; entry 2242 tools/perf/util/sort.c hde->hpp.entry = __sort__hde_entry; entry 2510 tools/perf/util/sort.c if (sd->entry->se_collapse) entry 2584 tools/perf/util/sort.c if (sd->entry == &sort_parent) { entry 2594 tools/perf/util/sort.c } else if (sd->entry == &sort_sym) { entry 2603 tools/perf/util/sort.c sd->entry->se_collapse = sort__sym_sort; entry 2605 tools/perf/util/sort.c } else if (sd->entry == &sort_dso) { entry 2607 tools/perf/util/sort.c } else if (sd->entry == &sort_socket) { entry 2609 tools/perf/util/sort.c } else if (sd->entry == &sort_thread) { entry 2611 tools/perf/util/sort.c } else if (sd->entry == &sort_comm) { entry 2636 tools/perf/util/sort.c if (sd->entry == &sort_sym_from || sd->entry == &sort_sym_to) entry 2652 tools/perf/util/sort.c if (sd->entry == &sort_mem_dcacheline && cacheline_size() == 0) entry 2655 tools/perf/util/sort.c if (sd->entry == &sort_mem_daddr_sym) entry 37 tools/perf/util/stat-shadow.c static int saved_value_cmp(struct rb_node *rb_node, const void *entry) entry 42 tools/perf/util/stat-shadow.c const struct saved_value *b = entry; entry 78 tools/perf/util/stat-shadow.c const void *entry) entry 84 tools/perf/util/stat-shadow.c memcpy(nd, entry, sizeof(struct saved_value)); entry 15 tools/perf/util/strlist.c struct rb_node *strlist__node_new(struct rblist *rblist, const void *entry) entry 17 tools/perf/util/strlist.c const char *s = entry; entry 55 tools/perf/util/strlist.c static int strlist__node_cmp(struct rb_node *rb_node, const void *entry) entry 57 tools/perf/util/strlist.c const char *str = entry; entry 70 tools/perf/util/strlist.c char entry[1024]; entry 77 tools/perf/util/strlist.c while (fgets(entry, sizeof(entry), fp) != NULL) { entry 78 tools/perf/util/strlist.c const size_t len = strlen(entry); entry 82 tools/perf/util/strlist.c entry[len - 1] = '\0'; entry 84 tools/perf/util/strlist.c err = strlist__add(slist, entry); entry 100 tools/perf/util/strlist.c struct str_node *strlist__find(struct strlist *slist, const char *entry) entry 103 tools/perf/util/strlist.c struct rb_node *rb_node = rblist__find(&slist->rblist, entry); entry 40 tools/perf/util/strlist.h struct str_node *strlist__find(struct strlist *slist, const char *entry); entry 42 tools/perf/util/strlist.h static inline bool strlist__has_entry(struct strlist *slist, const char *entry) entry 44 tools/perf/util/strlist.h return strlist__find(slist, entry) != NULL; entry 902 tools/perf/util/synthetic-events.c struct perf_record_thread_map_entry *entry = &event->thread_map.entries[i]; entry 908 tools/perf/util/synthetic-events.c entry->pid = perf_thread_map__pid(threads, i); entry 909 tools/perf/util/synthetic-events.c strncpy((char *) &entry->comm, comm, sizeof(entry->comm)); entry 47 tools/perf/util/syscalltbl.c const struct syscall *entry = ventry; entry 49 tools/perf/util/syscalltbl.c return strcmp(key, entry->name); entry 190 tools/perf/util/unwind-libdw.c return entry(pc, ui) || !(--ui->max_stack) ? entry 698 tools/perf/util/unwind-libunwind-local.c ret = ips[j] ? entry(ips[j], ui->thread, cb, arg) : 0; entry 20 tools/perf/util/unwind.h typedef int (*unwind_entry_cb_t)(struct unwind_entry *entry, void *arg); entry 226 tools/testing/radix-tree/idr-test.c void *entry; entry 230 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 236 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 242 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 248 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 255 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 257 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 264 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 266 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 275 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 277 tools/testing/radix-tree/idr-test.c idr_for_each_entry(idr, entry, id); entry 311 tools/testing/radix-tree/idr-test.c void *entry = idr_get_next(&find_idr, &id); entry 312 tools/testing/radix-tree/idr-test.c BUG_ON(entry != xa_mk_value(id)); entry 70 tools/testing/radix-tree/iteration_check.c void *entry; entry 77 tools/testing/radix-tree/iteration_check.c xas_for_each_marked(&xas, entry, ULONG_MAX, TAG) { entry 78 tools/testing/radix-tree/iteration_check.c if (xas_retry(&xas, entry)) entry 103 tools/testing/radix-tree/iteration_check.c void *entry; entry 110 tools/testing/radix-tree/iteration_check.c xas_for_each(&xas, entry, ULONG_MAX) { entry 111 tools/testing/radix-tree/iteration_check.c if (xas_retry(&xas, entry)) entry 16 tools/testing/radix-tree/iteration_check_2.c void *entry; entry 23 tools/testing/radix-tree/iteration_check_2.c xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0) entry 21 tools/testing/radix-tree/regression4.c void *entry; entry 28 tools/testing/radix-tree/regression4.c entry = radix_tree_lookup(&mt_tree, 0); entry 30 tools/testing/radix-tree/regression4.c if (entry != &obj0) { entry 31 tools/testing/radix-tree/regression4.c printf("iteration %d bad entry = %p\n", i, entry); entry 258 tools/testing/radix-tree/test.c void *entry; entry 260 tools/testing/radix-tree/test.c xas_for_each(&xas, entry, ULONG_MAX) { entry 261 tools/testing/radix-tree/test.c if (!xa_is_value(entry)) { entry 262 tools/testing/radix-tree/test.c item_free(entry, xas.xa_index); entry 54 tools/testing/selftests/bpf/test_hashmap.c struct hashmap_entry *entry, *tmp; entry 101 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { entry 102 tools/testing/selftests/bpf/test_hashmap.c long k = (long)entry->key; entry 103 tools/testing/selftests/bpf/test_hashmap.c long v = (long)entry->value; entry 144 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry_safe(map, entry, tmp, bkt) { entry 145 tools/testing/selftests/bpf/test_hashmap.c long k = (long)entry->key; entry 146 tools/testing/selftests/bpf/test_hashmap.c long v = (long)entry->value; entry 158 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_key_entry(map, entry, (void *)0) { entry 166 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_key_entry_safe(map, entry, tmp, (void *)0) { entry 170 tools/testing/selftests/bpf/test_hashmap.c k = entry->key; entry 171 tools/testing/selftests/bpf/test_hashmap.c v = entry->value; entry 201 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry_safe(map, entry, tmp, bkt) { entry 205 tools/testing/selftests/bpf/test_hashmap.c k = entry->key; entry 206 tools/testing/selftests/bpf/test_hashmap.c v = entry->value; entry 235 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { entry 237 tools/testing/selftests/bpf/test_hashmap.c (long)entry->key, (long)entry->value); entry 242 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { entry 244 tools/testing/selftests/bpf/test_hashmap.c (long)entry->key, (long)entry->value); entry 260 tools/testing/selftests/bpf/test_hashmap.c struct hashmap_entry *entry; entry 303 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { entry 304 tools/testing/selftests/bpf/test_hashmap.c found_msk |= (long)entry->value; entry 312 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_key_entry(map, entry, k1) { entry 313 tools/testing/selftests/bpf/test_hashmap.c found_msk |= (long)entry->value; entry 321 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_key_entry(map, entry, k2) { entry 322 tools/testing/selftests/bpf/test_hashmap.c found_msk |= (long)entry->value; entry 334 tools/testing/selftests/bpf/test_hashmap.c struct hashmap_entry *entry; entry 357 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { entry 361 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_key_entry(map, entry, k) { entry 60 tools/testing/selftests/kvm/lib/aarch64/processor.c static uint64_t pte_addr(struct kvm_vm *vm, uint64_t entry) entry 63 tools/testing/selftests/kvm/lib/aarch64/processor.c return entry & mask; entry 74 tools/testing/selftests/kvm/lib/s390x/processor.c uint64_t *entry; entry 94 tools/testing/selftests/kvm/lib/s390x/processor.c entry = addr_gpa2hva(vm, vm->pgd); entry 97 tools/testing/selftests/kvm/lib/s390x/processor.c if (entry[idx] & REGION_ENTRY_INVALID) entry 98 tools/testing/selftests/kvm/lib/s390x/processor.c entry[idx] = virt_alloc_region(vm, ri, memslot); entry 99 tools/testing/selftests/kvm/lib/s390x/processor.c entry = addr_gpa2hva(vm, entry[idx] & REGION_ENTRY_ORIGIN); entry 104 tools/testing/selftests/kvm/lib/s390x/processor.c if (!(entry[idx] & PAGE_INVALID)) entry 107 tools/testing/selftests/kvm/lib/s390x/processor.c entry[idx] = gpa; entry 133 tools/testing/selftests/kvm/lib/s390x/processor.c uint64_t *entry; entry 138 tools/testing/selftests/kvm/lib/s390x/processor.c entry = addr_gpa2hva(vm, vm->pgd); entry 141 tools/testing/selftests/kvm/lib/s390x/processor.c TEST_ASSERT(!(entry[idx] & REGION_ENTRY_INVALID), entry 144 tools/testing/selftests/kvm/lib/s390x/processor.c entry = addr_gpa2hva(vm, entry[idx] & REGION_ENTRY_ORIGIN); entry 149 tools/testing/selftests/kvm/lib/s390x/processor.c TEST_ASSERT(!(entry[idx] & PAGE_INVALID), entry 152 tools/testing/selftests/kvm/lib/s390x/processor.c return (entry[idx] & ~0xffful) + (gva & 0xffful); entry 172 tools/testing/selftests/kvm/lib/s390x/processor.c uint64_t addr, *entry; entry 175 tools/testing/selftests/kvm/lib/s390x/processor.c entry = addr_gpa2hva(vm, addr); entry 176 tools/testing/selftests/kvm/lib/s390x/processor.c if (*entry & REGION_ENTRY_INVALID) entry 179 tools/testing/selftests/kvm/lib/s390x/processor.c indent, "", 4 - ((*entry & REGION_ENTRY_TYPE) >> 2), entry 180 tools/testing/selftests/kvm/lib/s390x/processor.c addr, *entry); entry 181 tools/testing/selftests/kvm/lib/s390x/processor.c if (*entry & REGION_ENTRY_TYPE) { entry 183 tools/testing/selftests/kvm/lib/s390x/processor.c *entry & REGION_ENTRY_ORIGIN); entry 186 tools/testing/selftests/kvm/lib/s390x/processor.c *entry & REGION_ENTRY_ORIGIN); entry 752 tools/testing/selftests/kvm/lib/x86_64/processor.c struct kvm_cpuid_entry2 *entry = NULL; entry 759 tools/testing/selftests/kvm/lib/x86_64/processor.c entry = &cpuid->entries[i]; entry 764 tools/testing/selftests/kvm/lib/x86_64/processor.c TEST_ASSERT(entry, "Guest CPUID entry not found: (EAX=%x, ECX=%x).", entry 766 tools/testing/selftests/kvm/lib/x86_64/processor.c return entry; entry 858 tools/testing/selftests/kvm/lib/x86_64/processor.c struct kvm_msr_entry entry; entry 864 tools/testing/selftests/kvm/lib/x86_64/processor.c buffer.entry.index = msr_index; entry 869 tools/testing/selftests/kvm/lib/x86_64/processor.c return buffer.entry.data; entry 892 tools/testing/selftests/kvm/lib/x86_64/processor.c struct kvm_msr_entry entry; entry 899 tools/testing/selftests/kvm/lib/x86_64/processor.c buffer.entry.index = msr_index; entry 900 tools/testing/selftests/kvm/lib/x86_64/processor.c buffer.entry.data = msr_value; entry 1168 tools/testing/selftests/kvm/lib/x86_64/processor.c struct kvm_cpuid_entry2 *entry; entry 1177 tools/testing/selftests/kvm/lib/x86_64/processor.c entry = kvm_get_supported_cpuid_entry(0x80000008); entry 1178 tools/testing/selftests/kvm/lib/x86_64/processor.c *pa_bits = entry->eax & 0xff; entry 1179 tools/testing/selftests/kvm/lib/x86_64/processor.c *va_bits = (entry->eax >> 8) & 0xff; entry 384 tools/testing/selftests/kvm/lib/x86_64/vmx.c struct kvm_cpuid_entry2 *entry = kvm_get_supported_cpuid_entry(1); entry 386 tools/testing/selftests/kvm/lib/x86_64/vmx.c if (!(entry->ecx & CPUID_VMX)) { entry 69 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c struct kvm_cpuid_entry2 *entry; entry 73 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c entry = kvm_get_supported_cpuid_entry(1); entry 74 tools/testing/selftests/kvm/x86_64/cr4_cpuid_sync_test.c if (!(entry->ecx & X86_FEATURE_XSAVE)) { entry 65 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c struct kvm_cpuid_entry2 *entry = &hv_cpuid_entries->entries[i]; entry 67 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c TEST_ASSERT((entry->function >= 0x40000000) && entry 68 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c (entry->function <= 0x4000000A), entry 70 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c entry->function); entry 72 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c TEST_ASSERT(entry->index == 0, entry 75 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c TEST_ASSERT(entry->flags == 0, entry 78 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c TEST_ASSERT(!entry->padding[0] && !entry->padding[1] && entry 79 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c !entry->padding[2], "padding should be zero"); entry 81 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c if (entry->function == 0x40000004) { entry 82 tools/testing/selftests/kvm/x86_64/hyperv_cpuid.c int nononarchcs = !!(entry->eax & (1UL << 18)); entry 57 tools/testing/selftests/kvm/x86_64/vmx_tsc_adjust_test.c struct kvm_msr_entry entry; entry 264 tools/testing/selftests/networking/timestamping/timestamping.c struct iovec entry; entry 273 tools/testing/selftests/networking/timestamping/timestamping.c msg.msg_iov = &entry; entry 275 tools/testing/selftests/networking/timestamping/timestamping.c entry.iov_base = data; entry 276 tools/testing/selftests/networking/timestamping/timestamping.c entry.iov_len = sizeof(data); entry 271 tools/testing/selftests/networking/timestamping/txtimestamp.c struct iovec entry; entry 280 tools/testing/selftests/networking/timestamping/txtimestamp.c memset(&entry, 0, sizeof(entry)); entry 283 tools/testing/selftests/networking/timestamping/txtimestamp.c entry.iov_base = data; entry 284 tools/testing/selftests/networking/timestamping/txtimestamp.c entry.iov_len = cfg_payload_len; entry 285 tools/testing/selftests/networking/timestamping/txtimestamp.c msg.msg_iov = &entry; entry 133 tools/testing/selftests/powerpc/pmu/ebb/ebb.c u64 entry; entry 136 tools/testing/selftests/powerpc/pmu/ebb/ebb.c entry = (u64)ebb_handler; entry 140 tools/testing/selftests/powerpc/pmu/ebb/ebb.c u64 entry; entry 145 tools/testing/selftests/powerpc/pmu/ebb/ebb.c entry = opd->entry; entry 147 tools/testing/selftests/powerpc/pmu/ebb/ebb.c printf("EBB Handler is at %#llx\n", entry); entry 153 tools/testing/selftests/powerpc/pmu/ebb/ebb.c mtspr(SPRN_EBBHR, entry); entry 266 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define _ASM_NOKPROBE_SYMBOL(entry) \ entry 268 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h PPC_LONG (entry) ; \ entry 271 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define _ASM_NOKPROBE_SYMBOL(entry) entry 74 tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c struct extbl_entry *entry = (struct extbl_entry *)__start___ex_table; entry 76 tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c while (entry < (struct extbl_entry *)__stop___ex_table) { entry 79 tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c insn = (unsigned long)&entry->insn + entry->insn; entry 80 tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c fixup = (unsigned long)&entry->fixup + entry->fixup; entry 188 tools/testing/selftests/powerpc/vphn/asm/lppaca.h extern void (*dtl_consumer)(struct dtl_entry *entry, u64 index); entry 118 tools/testing/selftests/ptrace/get_syscall_info.c (void *) &info.entry - (void *) &info; entry 120 tools/testing/selftests/ptrace/get_syscall_info.c (void *) &info.entry.args[6] - (void *) &info; entry 203 tools/testing/selftests/ptrace/get_syscall_info.c ASSERT_EQ(exp_args[0], info.entry.nr) { entry 206 tools/testing/selftests/ptrace/get_syscall_info.c ASSERT_EQ(exp_args[1], info.entry.args[0]) { entry 209 tools/testing/selftests/ptrace/get_syscall_info.c ASSERT_EQ(exp_args[2], info.entry.args[1]) { entry 212 tools/testing/selftests/ptrace/get_syscall_info.c ASSERT_EQ(exp_args[3], info.entry.args[2]) { entry 215 tools/testing/selftests/ptrace/get_syscall_info.c ASSERT_EQ(exp_args[4], info.entry.args[3]) { entry 218 tools/testing/selftests/ptrace/get_syscall_info.c ASSERT_EQ(exp_args[5], info.entry.args[4]) { entry 221 tools/testing/selftests/ptrace/get_syscall_info.c ASSERT_EQ(exp_args[6], info.entry.args[5]) { entry 23 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/workqueues.h struct list_head entry; entry 31 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/workqueues.h struct hlist_node entry; entry 86 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/workqueues.h .entry = { &(n).entry, &(n).entry }, \ entry 1782 tools/testing/selftests/seccomp/seccomp_bpf.c static bool entry; entry 1788 tools/testing/selftests/seccomp/seccomp_bpf.c entry = !entry; entry 1793 tools/testing/selftests/seccomp/seccomp_bpf.c EXPECT_EQ(entry ? PTRACE_EVENTMSG_SYSCALL_ENTRY entry 1796 tools/testing/selftests/seccomp/seccomp_bpf.c if (!entry) entry 163 tools/testing/selftests/vm/compaction_test.c struct map_list *list, *entry; entry 201 tools/testing/selftests/vm/compaction_test.c entry = malloc(sizeof(struct map_list)); entry 202 tools/testing/selftests/vm/compaction_test.c if (!entry) { entry 206 tools/testing/selftests/vm/compaction_test.c entry->map = map; entry 207 tools/testing/selftests/vm/compaction_test.c entry->next = list; entry 208 tools/testing/selftests/vm/compaction_test.c list = entry; entry 219 tools/testing/selftests/vm/compaction_test.c for (entry = list; entry != NULL; entry = entry->next) { entry 220 tools/testing/selftests/vm/compaction_test.c munmap(entry->map, MAP_SIZE); entry 221 tools/testing/selftests/vm/compaction_test.c if (!entry->next) entry 223 tools/testing/selftests/vm/compaction_test.c entry = entry->next; entry 237 tools/usb/testusb.c struct testdev *entry; entry 255 tools/usb/testusb.c entry = calloc(1, sizeof *entry); entry 256 tools/usb/testusb.c if (!entry) entry 259 tools/usb/testusb.c entry->name = strdup(name); entry 260 tools/usb/testusb.c if (!entry->name) { entry 261 tools/usb/testusb.c free(entry); entry 267 tools/usb/testusb.c entry->ifnum = ifnum; entry 272 tools/usb/testusb.c speed(entry->speed), entry->name, entry->ifnum); entry 274 tools/usb/testusb.c entry->next = testdevs; entry 275 tools/usb/testusb.c testdevs = entry; entry 367 tools/usb/testusb.c struct testdev *entry; entry 485 tools/usb/testusb.c for (entry = testdevs; entry; entry = entry->next) { entry 488 tools/usb/testusb.c entry->param = param; entry 489 tools/usb/testusb.c entry->forever = forever; entry 490 tools/usb/testusb.c entry->test = test; entry 493 tools/usb/testusb.c if (strcmp (entry->name, device)) entry 495 tools/usb/testusb.c return handle_testdev (entry) != entry; entry 497 tools/usb/testusb.c status = pthread_create (&entry->thread, 0, handle_testdev, entry); entry 517 tools/usb/testusb.c for (entry = testdevs; entry; entry = entry->next) { entry 520 tools/usb/testusb.c if (pthread_join (entry->thread, &retval)) entry 86 tools/usb/usbip/libsrc/list.h static inline void __list_del_entry(struct list_head *entry) entry 88 tools/usb/usbip/libsrc/list.h __list_del(entry->prev, entry->next); entry 91 tools/usb/usbip/libsrc/list.h static inline void list_del(struct list_head *entry) entry 93 tools/usb/usbip/libsrc/list.h __list_del(entry->prev, entry->next); entry 94 tools/usb/usbip/libsrc/list.h entry->next = LIST_POISON1; entry 95 tools/usb/usbip/libsrc/list.h entry->prev = LIST_POISON2; entry 142 virt/kvm/arm/vgic/vgic-its.c struct list_head entry; entry 544 virt/kvm/arm/vgic/vgic-its.c list_for_each_entry(cte, &dist->lpi_translation_cache, entry) { entry 560 virt/kvm/arm/vgic/vgic-its.c if (!list_is_first(&cte->entry, &dist->lpi_translation_cache)) entry 561 virt/kvm/arm/vgic/vgic-its.c list_move(&cte->entry, &dist->lpi_translation_cache); entry 612 virt/kvm/arm/vgic/vgic-its.c typeof(*cte), entry); entry 630 virt/kvm/arm/vgic/vgic-its.c list_move(&cte->entry, &dist->lpi_translation_cache); entry 644 virt/kvm/arm/vgic/vgic-its.c list_for_each_entry(cte, &dist->lpi_translation_cache, entry) { entry 1846 virt/kvm/arm/vgic/vgic-its.c INIT_LIST_HEAD(&cte->entry); entry 1847 virt/kvm/arm/vgic/vgic-its.c list_add(&cte->entry, &dist->lpi_translation_cache); entry 1859 virt/kvm/arm/vgic/vgic-its.c &dist->lpi_translation_cache, entry) { entry 1860 virt/kvm/arm/vgic/vgic-its.c list_del(&cte->entry); entry 2061 virt/kvm/arm/vgic/vgic-its.c typedef int (*entry_fn_t)(struct vgic_its *its, u32 id, void *entry, entry 2086 virt/kvm/arm/vgic/vgic-its.c char entry[ESZ_MAX]; entry 2089 virt/kvm/arm/vgic/vgic-its.c memset(entry, 0, esz); entry 2095 virt/kvm/arm/vgic/vgic-its.c ret = kvm_read_guest_lock(kvm, gpa, entry, esz); entry 2099 virt/kvm/arm/vgic/vgic-its.c next_offset = fn(its, id, entry, opaque); entry 2293 virt/kvm/arm/vgic/vgic-its.c u64 entry = *(u64 *)ptr; entry 2298 virt/kvm/arm/vgic/vgic-its.c entry = le64_to_cpu(entry); entry 2300 virt/kvm/arm/vgic/vgic-its.c valid = entry >> KVM_ITS_DTE_VALID_SHIFT; entry 2301 virt/kvm/arm/vgic/vgic-its.c num_eventid_bits = (entry & KVM_ITS_DTE_SIZE_MASK) + 1; entry 2302 virt/kvm/arm/vgic/vgic-its.c itt_addr = ((entry & KVM_ITS_DTE_ITTADDR_MASK) entry 2309 virt/kvm/arm/vgic/vgic-its.c offset = (entry & KVM_ITS_DTE_NEXT_MASK) >> KVM_ITS_DTE_NEXT_SHIFT; entry 2391 virt/kvm/arm/vgic/vgic-its.c u64 entry = *(u64 *)addr; entry 2396 virt/kvm/arm/vgic/vgic-its.c entry = le64_to_cpu(entry); entry 2398 virt/kvm/arm/vgic/vgic-its.c if (!(entry & KVM_ITS_L1E_VALID_MASK)) entry 2401 virt/kvm/arm/vgic/vgic-its.c gpa = entry & KVM_ITS_L1E_ADDR_MASK; entry 1782 virt/kvm/kvm_main.c gfn_t entry = 0; entry 1784 virt/kvm/kvm_main.c addr = gfn_to_hva_many(slot, gfn, &entry); entry 1788 virt/kvm/kvm_main.c if (entry < nr_pages)