log 10213 arch/m68k/ifpsp060/src/fplsp.S # if enabled so the operating system can log the event. # log 993 arch/mips/kvm/mips.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log) log 1002 arch/mips/kvm/mips.c r = kvm_get_dirty_log_protect(kvm, log, &flush); log 1006 arch/mips/kvm/mips.c memslot = id_to_memslot(slots, log->slot); log 1016 arch/mips/kvm/mips.c int kvm_vm_ioctl_clear_dirty_log(struct kvm *kvm, struct kvm_clear_dirty_log *log) log 1025 arch/mips/kvm/mips.c r = kvm_clear_dirty_log_protect(kvm, log, &flush); log 1029 arch/mips/kvm/mips.c memslot = id_to_memslot(slots, log->slot); log 279 arch/powerpc/include/asm/kvm_ppc.h int (*get_dirty_log)(struct kvm *kvm, struct kvm_dirty_log *log); log 341 arch/powerpc/include/asm/rtas.h struct pseries_errorlog *get_pseries_errorlog(struct rtas_error_log *log, log 1031 arch/powerpc/kernel/rtas.c struct pseries_errorlog *get_pseries_errorlog(struct rtas_error_log *log, log 1035 arch/powerpc/kernel/rtas.c (struct rtas_ext_event_log_v6 *)log->buffer; log 1038 arch/powerpc/kernel/rtas.c uint32_t ext_log_length = rtas_error_extended_log_length(log); log 1048 arch/powerpc/kernel/rtas.c log_end = log->buffer + ext_log_length; log 287 arch/powerpc/kernel/rtasd.c static void handle_rtas_event(const struct rtas_error_log *log) log 289 arch/powerpc/kernel/rtasd.c if (rtas_error_type(log) != RTAS_TYPE_PRRN || !prrn_is_enabled()) log 295 arch/powerpc/kernel/rtasd.c handle_prrn_event(rtas_error_extended_log_length(log)); log 300 arch/powerpc/kernel/rtasd.c static void handle_rtas_event(const struct rtas_error_log *log) log 832 arch/powerpc/kvm/book3s.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log) log 834 arch/powerpc/kvm/book3s.c return kvm->arch.kvm_ops->get_dirty_log(kvm, log); log 4393 arch/powerpc/kvm/book3s_hv.c struct kvm_dirty_log *log) log 4405 arch/powerpc/kvm/book3s_hv.c if (log->slot >= KVM_USER_MEM_SLOTS) log 4409 arch/powerpc/kvm/book3s_hv.c memslot = id_to_memslot(slots, log->slot); log 4449 arch/powerpc/kvm/book3s_hv.c if (copy_to_user(log->dirty_bitmap, buf, n)) log 1863 arch/powerpc/kvm/book3s_pr.c struct kvm_dirty_log *log) log 1875 arch/powerpc/kvm/book3s_pr.c r = kvm_get_dirty_log(kvm, log, &is_dirty); log 1882 arch/powerpc/kvm/book3s_pr.c memslot = id_to_memslot(slots, log->slot); log 1799 arch/powerpc/kvm/booke.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log) log 2328 arch/powerpc/platforms/cell/spufs/file.c ctx->switch_log = kmalloc(struct_size(ctx->switch_log, log, log 2365 arch/powerpc/platforms/cell/spufs/file.c p = ctx->switch_log->log + ctx->switch_log->tail % SWITCH_LOG_BUFSIZE; log 2490 arch/powerpc/platforms/cell/spufs/file.c p = ctx->switch_log->log + ctx->switch_log->head; log 64 arch/powerpc/platforms/cell/spufs/spufs.h } log[]; log 224 arch/powerpc/platforms/pseries/ras.c static void rtas_parse_epow_errlog(struct rtas_error_log *log) log 231 arch/powerpc/platforms/pseries/ras.c pseries_log = get_pseries_errorlog(log, PSERIES_ELOG_SECT_ID_EPOW); log 613 arch/s390/kvm/kvm-s390.c struct kvm_dirty_log *log) log 627 arch/s390/kvm/kvm-s390.c if (log->slot >= KVM_USER_MEM_SLOTS) log 631 arch/s390/kvm/kvm-s390.c memslot = id_to_memslot(slots, log->slot); log 637 arch/s390/kvm/kvm-s390.c r = kvm_get_dirty_log(kvm, log, &is_dirty); log 88 arch/um/drivers/vhost_user.h u64 desc, used, avail, log; log 604 arch/um/drivers/virtio_uml.c u64 log) log 613 arch/um/drivers/virtio_uml.c .payload.vring_addr.log = log, log 4746 arch/x86/kvm/x86.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log) log 4759 arch/x86/kvm/x86.c r = kvm_get_dirty_log_protect(kvm, log, &flush); log 4773 arch/x86/kvm/x86.c int kvm_vm_ioctl_clear_dirty_log(struct kvm *kvm, struct kvm_clear_dirty_log *log) log 4786 arch/x86/kvm/x86.c r = kvm_clear_dirty_log_protect(kvm, log, &flush); log 105 drivers/acpi/sbs.c static inline int battery_scale(int log) log 108 drivers/acpi/sbs.c while (log--) log 201 drivers/android/binder.c struct binder_transaction_log *log) log 204 drivers/android/binder.c unsigned int cur = atomic_inc_return(&log->cur); log 206 drivers/android/binder.c if (cur >= ARRAY_SIZE(log->entry)) log 207 drivers/android/binder.c log->full = true; log 208 drivers/android/binder.c e = &log->entry[cur % ARRAY_SIZE(log->entry)]; log 6047 drivers/android/binder.c struct binder_transaction_log *log = m->private; log 6048 drivers/android/binder.c unsigned int log_cur = atomic_read(&log->cur); log 6054 drivers/android/binder.c cur = count < ARRAY_SIZE(log->entry) && !log->full ? log 6055 drivers/android/binder.c 0 : count % ARRAY_SIZE(log->entry); log 6056 drivers/android/binder.c if (count > ARRAY_SIZE(log->entry) || log->full) log 6057 drivers/android/binder.c count = ARRAY_SIZE(log->entry); log 6059 drivers/android/binder.c unsigned int index = cur++ % ARRAY_SIZE(log->entry); log 6061 drivers/android/binder.c print_binder_transaction_log_entry(m, &log->entry[index]); log 2042 drivers/ata/libata-core.c unsigned int ata_read_log_page(struct ata_device *dev, u8 log, log 2050 drivers/ata/libata-core.c DPRINTK("read log page - log 0x%x, page 0x%x\n", log, page); log 2071 drivers/ata/libata-core.c tf.lbal = log; log 2090 drivers/ata/libata-core.c static bool ata_log_supported(struct ata_device *dev, u8 log) log 2096 drivers/ata/libata-core.c return get_unaligned_le16(&ap->sector_buf[log * 2]) ? true : false; log 85 drivers/ata/libata.h extern unsigned int ata_read_log_page(struct ata_device *dev, u8 log, log 48 drivers/base/devres.c module_param_named(log, log_devres, int, S_IRUGO | S_IWUSR); log 225 drivers/block/paride/pg.c static int pg_identify(struct pg *dev, int log); log 437 drivers/block/paride/pg.c static int pg_identify(struct pg *dev, int log) log 452 drivers/block/paride/pg.c if (log) { log 613 drivers/block/rsxx/cregs.c log_str = &card->log.buf[card->log.buf_len]; log 614 drivers/block/rsxx/cregs.c cnt = min(cmd->cnt8 - off, LOG_BUF_SIZE8 - card->log.buf_len); log 618 drivers/block/rsxx/cregs.c card->log.buf_len += len; log 625 drivers/block/rsxx/cregs.c (card->log.buf_len == LOG_BUF_SIZE8)) { log 626 drivers/block/rsxx/cregs.c if (card->log.buf_len != 1) /* Don't log blank lines. */ log 627 drivers/block/rsxx/cregs.c hw_log_msg(card, card->log.buf, log 628 drivers/block/rsxx/cregs.c card->log.buf_len); log 629 drivers/block/rsxx/cregs.c card->log.buf_len = 0; log 643 drivers/block/rsxx/cregs.c sizeof(card->log.tmp), card->log.tmp, log 144 drivers/block/rsxx/rsxx_priv.h } log; log 51 drivers/char/tpm/eventlog/acpi.c struct tpm_bios_log *log; log 56 drivers/char/tpm/eventlog/acpi.c log = &chip->log; log 88 drivers/char/tpm/eventlog/acpi.c log->bios_event_log = kmalloc(len, GFP_KERNEL); log 89 drivers/char/tpm/eventlog/acpi.c if (!log->bios_event_log) log 92 drivers/char/tpm/eventlog/acpi.c log->bios_event_log_end = log->bios_event_log + len; log 98 drivers/char/tpm/eventlog/acpi.c memcpy_fromio(log->bios_event_log, virt, len); log 104 drivers/char/tpm/eventlog/acpi.c kfree(log->bios_event_log); log 105 drivers/char/tpm/eventlog/acpi.c log->bios_event_log = NULL; log 78 drivers/char/tpm/eventlog/common.c if (chip->log.bios_event_log != NULL) { log 21 drivers/char/tpm/eventlog/efi.c struct tpm_bios_log *log; log 33 drivers/char/tpm/eventlog/efi.c log = &chip->log; log 52 drivers/char/tpm/eventlog/efi.c log->bios_event_log = kmemdup(log_tbl->log, log_size, GFP_KERNEL); log 53 drivers/char/tpm/eventlog/efi.c if (!log->bios_event_log) { log 58 drivers/char/tpm/eventlog/efi.c log->bios_event_log_end = log->bios_event_log + log_size; log 73 drivers/char/tpm/eventlog/efi.c kfree(log->bios_event_log); log 80 drivers/char/tpm/eventlog/efi.c tmp = krealloc(log->bios_event_log, log 84 drivers/char/tpm/eventlog/efi.c kfree(log->bios_event_log); log 89 drivers/char/tpm/eventlog/efi.c log->bios_event_log = tmp; log 96 drivers/char/tpm/eventlog/efi.c memcpy((void *)log->bios_event_log + log_size, log 99 drivers/char/tpm/eventlog/efi.c log->bios_event_log_end = log->bios_event_log + log 25 drivers/char/tpm/eventlog/of.c struct tpm_bios_log *log; log 29 drivers/char/tpm/eventlog/of.c log = &chip->log; log 67 drivers/char/tpm/eventlog/of.c log->bios_event_log = kmemdup(__va(base), size, GFP_KERNEL); log 68 drivers/char/tpm/eventlog/of.c if (!log->bios_event_log) log 71 drivers/char/tpm/eventlog/of.c log->bios_event_log_end = log->bios_event_log + size; log 74 drivers/char/tpm/eventlog/tpm1.c struct tpm_bios_log *log = &chip->log; log 75 drivers/char/tpm/eventlog/tpm1.c void *addr = log->bios_event_log; log 76 drivers/char/tpm/eventlog/tpm1.c void *limit = log->bios_event_log_end; log 113 drivers/char/tpm/eventlog/tpm1.c struct tpm_bios_log *log = &chip->log; log 114 drivers/char/tpm/eventlog/tpm1.c void *limit = log->bios_event_log_end; log 45 drivers/char/tpm/eventlog/tpm2.c struct tpm_bios_log *log = &chip->log; log 46 drivers/char/tpm/eventlog/tpm2.c void *addr = log->bios_event_log; log 47 drivers/char/tpm/eventlog/tpm2.c void *limit = log->bios_event_log_end; log 92 drivers/char/tpm/eventlog/tpm2.c struct tpm_bios_log *log = &chip->log; log 93 drivers/char/tpm/eventlog/tpm2.c void *limit = log->bios_event_log_end; log 98 drivers/char/tpm/eventlog/tpm2.c event_header = log->bios_event_log; log 132 drivers/char/tpm/eventlog/tpm2.c struct tpm_bios_log *log = &chip->log; log 133 drivers/char/tpm/eventlog/tpm2.c struct tcg_pcr_event *event_header = log->bios_event_log; log 270 drivers/char/tpm/tpm-chip.c kfree(chip->log.bios_event_log); log 129 drivers/edac/i3200_edac.c static unsigned long eccerrlog_syndrome(u64 log) log 131 drivers/edac/i3200_edac.c return (log & I3200_ECCERRLOG_SYNDROME_BITS) >> log 135 drivers/edac/i3200_edac.c static int eccerrlog_row(int channel, u64 log) log 137 drivers/edac/i3200_edac.c u64 rank = ((log & I3200_ECCERRLOG_RANK_BITS) >> log 223 drivers/edac/i3200_edac.c u64 log; log 235 drivers/edac/i3200_edac.c log = info->eccerrlog[channel]; log 236 drivers/edac/i3200_edac.c if (log & I3200_ECCERRLOG_UE) { log 239 drivers/edac/i3200_edac.c eccerrlog_row(channel, log), log 242 drivers/edac/i3200_edac.c } else if (log & I3200_ECCERRLOG_CE) { log 244 drivers/edac/i3200_edac.c 0, 0, eccerrlog_syndrome(log), log 245 drivers/edac/i3200_edac.c eccerrlog_row(channel, log), log 147 drivers/edac/ie31200_edac.c #define IE31200_ECCERRLOG_SYNDROME(log) \ log 148 drivers/edac/ie31200_edac.c ((log & IE31200_ECCERRLOG_SYNDROME_BITS) >> \ log 241 drivers/edac/ie31200_edac.c static int eccerrlog_row(u64 log) log 243 drivers/edac/ie31200_edac.c return ((log & IE31200_ECCERRLOG_RANK_BITS) >> log 300 drivers/edac/ie31200_edac.c u64 log; log 312 drivers/edac/ie31200_edac.c log = info->eccerrlog[channel]; log 313 drivers/edac/ie31200_edac.c if (log & IE31200_ECCERRLOG_UE) { log 316 drivers/edac/ie31200_edac.c eccerrlog_row(log), log 319 drivers/edac/ie31200_edac.c } else if (log & IE31200_ECCERRLOG_CE) { log 322 drivers/edac/ie31200_edac.c IE31200_ECCERRLOG_SYNDROME(log), log 323 drivers/edac/ie31200_edac.c eccerrlog_row(log), log 116 drivers/edac/x38_edac.c static unsigned long eccerrlog_syndrome(u64 log) log 118 drivers/edac/x38_edac.c return (log & X38_ECCERRLOG_SYNDROME_BITS) >> 16; log 121 drivers/edac/x38_edac.c static int eccerrlog_row(int channel, u64 log) log 123 drivers/edac/x38_edac.c return ((log & X38_ECCERRLOG_RANK_BITS) >> 27) | log 207 drivers/edac/x38_edac.c u64 log; log 220 drivers/edac/x38_edac.c log = info->eccerrlog[channel]; log 221 drivers/edac/x38_edac.c if (log & X38_ECCERRLOG_UE) { log 224 drivers/edac/x38_edac.c eccerrlog_row(channel, log), log 227 drivers/edac/x38_edac.c } else if (log & X38_ECCERRLOG_CE) { log 229 drivers/edac/x38_edac.c 0, 0, eccerrlog_syndrome(log), log 230 drivers/edac/x38_edac.c eccerrlog_row(channel, log), log 171 drivers/firmware/efi/libstub/tpm.c memcpy(log_tbl->log, (void *) first_entry_addr, log_size); log 89 drivers/firmware/efi/tpm.c log_tbl->log); log 65 drivers/gpu/drm/i915/gt/uc/intel_guc.c intel_guc_log_init_early(&guc->log); log 113 drivers/gpu/drm/i915/gt/uc/intel_guc.c u32 level = intel_guc_log_get_level(&guc->log); log 154 drivers/gpu/drm/i915/gt/uc/intel_guc.c u32 offset = intel_guc_ggtt_offset(guc, guc->log.vma) >> PAGE_SHIFT; log 262 drivers/gpu/drm/i915/gt/uc/intel_guc.c ret = intel_guc_log_create(&guc->log); log 298 drivers/gpu/drm/i915/gt/uc/intel_guc.c intel_guc_log_destroy(&guc->log); log 324 drivers/gpu/drm/i915/gt/uc/intel_guc.c intel_guc_log_destroy(&guc->log); log 424 drivers/gpu/drm/i915/gt/uc/intel_guc.c intel_guc_log_handle_flush_event(&guc->log); log 28 drivers/gpu/drm/i915/gt/uc/intel_guc.h struct intel_guc_log log; log 13 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_capture_logs(struct intel_guc_log *log); log 58 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static inline struct intel_guc *log_to_guc(struct intel_guc_log *log) log 60 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c return container_of(log, struct intel_guc, log); log 63 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_enable_flush_events(struct intel_guc_log *log) log 65 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c intel_guc_enable_msg(log_to_guc(log), log 70 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_disable_flush_events(struct intel_guc_log *log) log 72 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c intel_guc_disable_msg(log_to_guc(log), log 148 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_move_to_next_buf(struct intel_guc_log *log) log 157 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c relay_reserve(log->relay.channel, log->vma->obj->base.size); log 160 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c relay_flush(log->relay.channel); log 163 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void *guc_get_write_buffer(struct intel_guc_log *log) log 174 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c return relay_reserve(log->relay.channel, 0); log 177 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static bool guc_check_log_buf_overflow(struct intel_guc_log *log, log 181 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c unsigned int prev_full_cnt = log->stats[type].sampled_overflow; log 187 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->stats[type].overflow = full_cnt; log 188 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->stats[type].sampled_overflow += full_cnt - prev_full_cnt; log 192 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->stats[type].sampled_overflow += 16; log 195 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c dev_notice_ratelimited(guc_to_gt(log_to_guc(log))->i915->drm.dev, log 218 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_read_update_log_buffer(struct intel_guc_log *log) log 227 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c mutex_lock(&log->relay.lock); log 229 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c if (WARN_ON(!intel_guc_log_relay_enabled(log))) log 233 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log_buf_state = src_data = log->relay.buf_addr; log 236 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log_buf_snapshot_state = dst_data = guc_get_write_buffer(log); log 244 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->relay.full_count++; log 267 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->stats[type].flush += log_buf_state_local.flush_to_file; log 268 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c new_overflow = guc_check_log_buf_overflow(log, type, full_cnt); log 315 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_move_to_next_buf(log); log 318 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c mutex_unlock(&log->relay.lock); log 323 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c struct intel_guc_log *log = log 326 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_log_capture_logs(log); log 329 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static int guc_log_map(struct intel_guc_log *log) log 333 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c lockdep_assert_held(&log->relay.lock); log 335 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c if (!log->vma) log 343 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c vaddr = i915_gem_object_pin_map(log->vma->obj, I915_MAP_WC); log 347 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->relay.buf_addr = vaddr; log 352 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_unmap(struct intel_guc_log *log) log 354 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c lockdep_assert_held(&log->relay.lock); log 356 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c i915_gem_object_unpin_map(log->vma->obj); log 357 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->relay.buf_addr = NULL; log 360 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_init_early(struct intel_guc_log *log) log 362 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c mutex_init(&log->relay.lock); log 363 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c INIT_WORK(&log->relay.flush_work, capture_logs_work); log 366 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static int guc_log_relay_create(struct intel_guc_log *log) log 368 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c struct intel_guc *guc = log_to_guc(log); log 374 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c lockdep_assert_held(&log->relay.lock); log 375 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c GEM_BUG_ON(!log->vma); log 378 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c subbuf_size = log->vma->size; log 400 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->relay.channel = guc_log_relay_chan; log 405 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_relay_destroy(struct intel_guc_log *log) log 407 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c lockdep_assert_held(&log->relay.lock); log 409 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c relay_close(log->relay.channel); log 410 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->relay.channel = NULL; log 413 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static void guc_log_capture_logs(struct intel_guc_log *log) log 415 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c struct intel_guc *guc = log_to_guc(log); log 419 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_read_update_log_buffer(log); log 429 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c static u32 __get_default_log_level(struct intel_guc_log *log) log 452 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c int intel_guc_log_create(struct intel_guc_log *log) log 454 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c struct intel_guc *guc = log_to_guc(log); log 459 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c GEM_BUG_ON(log->vma); log 489 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->vma = vma; log 491 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->level = __get_default_log_level(log); log 493 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->level, enableddisabled(log->level), log 494 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c yesno(GUC_LOG_LEVEL_IS_VERBOSE(log->level)), log 495 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c GUC_LOG_LEVEL_TO_VERBOSITY(log->level)); log 504 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_destroy(struct intel_guc_log *log) log 506 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c i915_vma_unpin_and_release(&log->vma, 0); log 509 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c int intel_guc_log_set_level(struct intel_guc_log *log, u32 level) log 511 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c struct intel_guc *guc = log_to_guc(log); log 517 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c GEM_BUG_ON(!log->vma); log 528 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c if (log->level == level) log 541 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c log->level = level; log 549 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c bool intel_guc_log_relay_enabled(const struct intel_guc_log *log) log 551 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c return log->relay.buf_addr; log 554 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c int intel_guc_log_relay_open(struct intel_guc_log *log) log 558 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c if (!log->vma) log 561 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c mutex_lock(&log->relay.lock); log 563 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c if (intel_guc_log_relay_enabled(log)) { log 578 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c ret = guc_log_relay_create(log); log 582 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c ret = guc_log_map(log); log 586 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c mutex_unlock(&log->relay.lock); log 588 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_log_enable_flush_events(log); log 595 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c queue_work(system_highpri_wq, &log->relay.flush_work); log 600 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_log_relay_destroy(log); log 602 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c mutex_unlock(&log->relay.lock); log 607 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_relay_flush(struct intel_guc_log *log) log 609 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c struct intel_guc *guc = log_to_guc(log); log 617 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c flush_work(&log->relay.flush_work); log 623 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_log_capture_logs(log); log 626 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_relay_close(struct intel_guc_log *log) log 628 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c struct intel_guc *guc = log_to_guc(log); log 631 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_log_disable_flush_events(log); log 634 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c flush_work(&log->relay.flush_work); log 636 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c mutex_lock(&log->relay.lock); log 637 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c GEM_BUG_ON(!intel_guc_log_relay_enabled(log)); log 638 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_log_unmap(log); log 639 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c guc_log_relay_destroy(log); log 640 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c mutex_unlock(&log->relay.lock); log 643 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c void intel_guc_log_handle_flush_event(struct intel_guc_log *log) log 645 drivers/gpu/drm/i915/gt/uc/intel_guc_log.c queue_work(system_highpri_wq, &log->relay.flush_work); log 63 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_init_early(struct intel_guc_log *log); log 64 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h int intel_guc_log_create(struct intel_guc_log *log); log 65 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_destroy(struct intel_guc_log *log); log 67 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h int intel_guc_log_set_level(struct intel_guc_log *log, u32 level); log 68 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h bool intel_guc_log_relay_enabled(const struct intel_guc_log *log); log 69 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h int intel_guc_log_relay_open(struct intel_guc_log *log); log 70 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_relay_flush(struct intel_guc_log *log); log 71 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_relay_close(struct intel_guc_log *log); log 73 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h void intel_guc_log_handle_flush_event(struct intel_guc_log *log); log 75 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h static inline u32 intel_guc_log_get_level(struct intel_guc_log *log) log 77 drivers/gpu/drm/i915/gt/uc/intel_guc_log.h return log->level; log 114 drivers/gpu/drm/i915/gt/uc/intel_uc.c if (guc->log.vma && !uc->load_err_log) log 115 drivers/gpu/drm/i915/gt/uc/intel_uc.c uc->load_err_log = i915_gem_object_get(guc->log.vma->obj); log 120 drivers/gpu/drm/i915/gt/uc/intel_uc.c struct drm_i915_gem_object *log = fetch_and_zero(&uc->load_err_log); log 122 drivers/gpu/drm/i915/gt/uc/intel_uc.c if (log) log 123 drivers/gpu/drm/i915/gt/uc/intel_uc.c i915_gem_object_put(log); log 1860 drivers/gpu/drm/i915/i915_debugfs.c struct intel_guc_log *log = &dev_priv->gt.uc.guc.log; log 1863 drivers/gpu/drm/i915/i915_debugfs.c if (!intel_guc_log_relay_enabled(log)) { log 1871 drivers/gpu/drm/i915/i915_debugfs.c log->relay.full_count); log 1876 drivers/gpu/drm/i915/i915_debugfs.c log->stats[type].flush, log 1877 drivers/gpu/drm/i915/i915_debugfs.c log->stats[type].sampled_overflow); log 1971 drivers/gpu/drm/i915/i915_debugfs.c u32 *log; log 1979 drivers/gpu/drm/i915/i915_debugfs.c else if (dev_priv->gt.uc.guc.log.vma) log 1980 drivers/gpu/drm/i915/i915_debugfs.c obj = dev_priv->gt.uc.guc.log.vma->obj; log 1985 drivers/gpu/drm/i915/i915_debugfs.c log = i915_gem_object_pin_map(obj, I915_MAP_WC); log 1986 drivers/gpu/drm/i915/i915_debugfs.c if (IS_ERR(log)) { log 1989 drivers/gpu/drm/i915/i915_debugfs.c return PTR_ERR(log); log 1994 drivers/gpu/drm/i915/i915_debugfs.c *(log + i), *(log + i + 1), log 1995 drivers/gpu/drm/i915/i915_debugfs.c *(log + i + 2), *(log + i + 3)); log 2011 drivers/gpu/drm/i915/i915_debugfs.c *val = intel_guc_log_get_level(&dev_priv->gt.uc.guc.log); log 2023 drivers/gpu/drm/i915/i915_debugfs.c return intel_guc_log_set_level(&dev_priv->gt.uc.guc.log, val); log 2034 drivers/gpu/drm/i915/i915_debugfs.c struct intel_guc_log *log = &guc->log; log 2039 drivers/gpu/drm/i915/i915_debugfs.c file->private_data = log; log 2041 drivers/gpu/drm/i915/i915_debugfs.c return intel_guc_log_relay_open(log); log 2050 drivers/gpu/drm/i915/i915_debugfs.c struct intel_guc_log *log = filp->private_data; log 2052 drivers/gpu/drm/i915/i915_debugfs.c intel_guc_log_relay_flush(log); log 2061 drivers/gpu/drm/i915/i915_debugfs.c intel_guc_log_relay_close(&guc->log); log 1488 drivers/gpu/drm/i915/i915_gpu_error.c uc->guc.log.vma, log 36 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/gpio.h u8 log[2]; log 132 drivers/gpu/drm/nouveau/nvkm/engine/pm/gf100.c u32 log = ctr->logic_op; log 142 drivers/gpu/drm/nouveau/nvkm/engine/pm/gf100.c nvkm_wr32(device, dom->addr + 0x044 + (ctr->slot * 0x08), log); log 31 drivers/gpu/drm/nouveau/nvkm/engine/pm/nv40.c u32 log = ctr->logic_op; log 40 drivers/gpu/drm/nouveau/nvkm/engine/pm/nv40.c nvkm_wr32(device, 0x00a420 + dom->addr + (ctr->slot * 0x40), log); log 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c .log[0] = (info & 0x1800) >> 11, log 89 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c .log[1] = (info & 0x6000) >> 13, log 98 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c .log[0] = (info & 0x18000000) >> 27, log 99 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c .log[1] = (info & 0x60000000) >> 29, log 108 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c .log[0] = (info1 & 0x30) >> 4, log 109 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c .log[1] = (info1 & 0xc0) >> 6, log 140 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c .log[0] = !!(conf & 0x02), log 141 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c .log[1] = !(conf & 0x02), log 1598 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->fuc.r_funcMV[0] = (gpio.log[0] ^ 2) << 12; log 1599 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->fuc.r_funcMV[1] = (gpio.log[1] ^ 2) << 12; log 1605 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->fuc.r_func2E[0] = (gpio.log[0] ^ 2) << 12; log 1606 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c ram->fuc.r_func2E[1] = (gpio.log[1] ^ 2) << 12; log 483 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgt215.c if (!(func.log[1] & 1)) log 210 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv50.c if (!(func.log[1] & 1)) log 70 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c .log[0] = 0, log 71 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c .log[1] = 1, log 88 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c int dir = !!(func.log[state] & 0x02); log 89 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c int out = !!(func.log[state] & 0x01); log 106 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c ret = (ret == (func.log[1] & 1)); log 236 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c if (func.line != 16 && func.log[0] & DCB_GPIO_LOG_DIR_IN) { log 50 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fanpwm.c if (card_type <= NV_40 || (fan->func.log[0] & 1)) log 75 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fanpwm.c if (card_type <= NV_40 || (fan->func.log[0] & 1)) log 1390 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h int vmw_host_log(const char *log); log 479 drivers/gpu/drm/vmwgfx/vmwgfx_msg.c int vmw_host_log(const char *log) log 489 drivers/gpu/drm/vmwgfx/vmwgfx_msg.c if (!log) log 492 drivers/gpu/drm/vmwgfx/vmwgfx_msg.c msg = kasprintf(GFP_KERNEL, "log %s", log); log 100 drivers/ide/ide-cd.c int log = 0; log 118 drivers/ide/ide-cd.c log = 1; log 127 drivers/ide/ide-cd.c log = 1; log 138 drivers/ide/ide-cd.c log = 1; log 141 drivers/ide/ide-cd.c return log; log 131 drivers/isdn/hardware/mISDN/avmfritz.c char log[LOG_SIZE + 1]; log 429 drivers/isdn/hardware/mISDN/avmfritz.c snprintf(fc->log, LOG_SIZE, "B%1d-recv %s %d ", log 431 drivers/isdn/hardware/mISDN/avmfritz.c print_hex_dump_bytes(fc->log, DUMP_PREFIX_OFFSET, p, count); log 500 drivers/isdn/hardware/mISDN/avmfritz.c snprintf(fc->log, LOG_SIZE, "B%1d-send %s %d ", log 502 drivers/isdn/hardware/mISDN/avmfritz.c print_hex_dump_bytes(fc->log, DUMP_PREFIX_OFFSET, p, count); log 52 drivers/isdn/hardware/mISDN/ipac.h char log[64]; log 53 drivers/isdn/hardware/mISDN/isar.h u8 log[256]; log 955 drivers/isdn/hardware/mISDN/mISDNipac.c snprintf(hscx->log, 64, "B%1d-recv %s %d ", log 957 drivers/isdn/hardware/mISDN/mISDNipac.c print_hex_dump_bytes(hscx->log, DUMP_PREFIX_OFFSET, p, count); log 972 drivers/isdn/hardware/mISDN/mISDNipac.c p = hscx->log; log 1000 drivers/isdn/hardware/mISDN/mISDNipac.c snprintf(hscx->log, 64, "B%1d-send %s %d ", log 1002 drivers/isdn/hardware/mISDN/mISDNipac.c print_hex_dump_bytes(hscx->log, DUMP_PREFIX_OFFSET, p, count); log 74 drivers/isdn/hardware/mISDN/mISDNisar.c isar->log, 256, 1); log 76 drivers/isdn/hardware/mISDN/mISDNisar.c __func__, l, isar->log); log 103 drivers/isdn/hardware/mISDN/mISDNisar.c 1, isar->log, 256, 1); log 105 drivers/isdn/hardware/mISDN/mISDNisar.c __func__, l, isar->log); log 84 drivers/isdn/hardware/mISDN/netjet.c char log[LOG_SIZE]; log 413 drivers/isdn/hardware/mISDN/netjet.c snprintf(card->log, LOG_SIZE, log 416 drivers/isdn/hardware/mISDN/netjet.c print_hex_dump_bytes(card->log, log 517 drivers/isdn/hardware/mISDN/netjet.c snprintf(card->log, LOG_SIZE, "B%1d-send %s %d ", log 519 drivers/isdn/hardware/mISDN/netjet.c print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, p, count); log 591 drivers/isdn/hardware/mISDN/netjet.c snprintf(card->log, LOG_SIZE, "B%1d-send %s %d ", log 593 drivers/isdn/hardware/mISDN/netjet.c print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, p, count); log 68 drivers/isdn/hardware/mISDN/w6692.c char log[64]; log 266 drivers/isdn/hardware/mISDN/w6692.c snprintf(card->log, 63, "D-recv %s %d ", log 268 drivers/isdn/hardware/mISDN/w6692.c print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, ptr, count); log 301 drivers/isdn/hardware/mISDN/w6692.c snprintf(card->log, 63, "D-send %s %d ", log 303 drivers/isdn/hardware/mISDN/w6692.c print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, ptr, count); log 477 drivers/isdn/hardware/mISDN/w6692.c snprintf(card->log, 63, "B%1d-recv %s %d ", log 479 drivers/isdn/hardware/mISDN/w6692.c print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, ptr, count); log 521 drivers/isdn/hardware/mISDN/w6692.c snprintf(card->log, 63, "B%1d-send %s %d ", log 523 drivers/isdn/hardware/mISDN/w6692.c print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, ptr, count); log 335 drivers/md/dm-crypt.c int log; log 341 drivers/md/dm-crypt.c log = ilog2(bs); log 346 drivers/md/dm-crypt.c if (1 << log != bs) { log 351 drivers/md/dm-crypt.c if (log > 9) { log 356 drivers/md/dm-crypt.c cc->iv_gen_private.benbi.shift = 9 - log; log 190 drivers/md/dm-log-userspace-base.c static int userspace_ctr(struct dm_dirty_log *log, struct dm_target *ti, log 320 drivers/md/dm-log-userspace-base.c log->context = lc; log 326 drivers/md/dm-log-userspace-base.c static void userspace_dtr(struct dm_dirty_log *log) log 328 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 352 drivers/md/dm-log-userspace-base.c static int userspace_presuspend(struct dm_dirty_log *log) log 355 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 363 drivers/md/dm-log-userspace-base.c static int userspace_postsuspend(struct dm_dirty_log *log) log 366 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 380 drivers/md/dm-log-userspace-base.c static int userspace_resume(struct dm_dirty_log *log) log 383 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 392 drivers/md/dm-log-userspace-base.c static uint32_t userspace_get_region_size(struct dm_dirty_log *log) log 394 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 407 drivers/md/dm-log-userspace-base.c static int userspace_is_clean(struct dm_dirty_log *log, region_t region) log 413 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 434 drivers/md/dm-log-userspace-base.c static int userspace_in_sync(struct dm_dirty_log *log, region_t region, log 441 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 562 drivers/md/dm-log-userspace-base.c static int userspace_flush(struct dm_dirty_log *log) log 566 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 648 drivers/md/dm-log-userspace-base.c static void userspace_mark_region(struct dm_dirty_log *log, region_t region) log 651 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 677 drivers/md/dm-log-userspace-base.c static void userspace_clear_region(struct dm_dirty_log *log, region_t region) log 680 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 712 drivers/md/dm-log-userspace-base.c static int userspace_get_resync_work(struct dm_dirty_log *log, region_t *region) log 716 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 739 drivers/md/dm-log-userspace-base.c static void userspace_set_region_sync(struct dm_dirty_log *log, log 742 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 769 drivers/md/dm-log-userspace-base.c static region_t userspace_get_sync_count(struct dm_dirty_log *log) log 774 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 794 drivers/md/dm-log-userspace-base.c static int userspace_status(struct dm_dirty_log *log, status_type_t status_type, log 800 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 809 drivers/md/dm-log-userspace-base.c DMEMIT("%s 1 COM_FAILURE", log->type->name); log 818 drivers/md/dm-log-userspace-base.c DMEMIT("%s %u %s ", log->type->name, lc->usr_argc, lc->uuid); log 832 drivers/md/dm-log-userspace-base.c static int userspace_is_remote_recovering(struct dm_dirty_log *log, log 837 drivers/md/dm-log-userspace-base.c struct log_c *lc = log->context; log 153 drivers/md/dm-log.c struct dm_dirty_log *log; log 155 drivers/md/dm-log.c log = kmalloc(sizeof(*log), GFP_KERNEL); log 156 drivers/md/dm-log.c if (!log) log 161 drivers/md/dm-log.c kfree(log); log 165 drivers/md/dm-log.c log->flush_callback_fn = flush_callback_fn; log 166 drivers/md/dm-log.c log->type = type; log 167 drivers/md/dm-log.c if (type->ctr(log, ti, argc, argv)) { log 168 drivers/md/dm-log.c kfree(log); log 173 drivers/md/dm-log.c return log; log 177 drivers/md/dm-log.c void dm_dirty_log_destroy(struct dm_dirty_log *log) log 179 drivers/md/dm-log.c log->type->dtr(log); log 180 drivers/md/dm-log.c put_type(log->type); log 181 drivers/md/dm-log.c kfree(log); log 316 drivers/md/dm-log.c static int read_header(struct log_c *log) log 320 drivers/md/dm-log.c r = rw_header(log, REQ_OP_READ); log 324 drivers/md/dm-log.c header_from_disk(&log->header, log->disk_header); log 327 drivers/md/dm-log.c if (log->sync != DEFAULTSYNC || log->header.magic != MIRROR_MAGIC) { log 328 drivers/md/dm-log.c log->header.magic = MIRROR_MAGIC; log 329 drivers/md/dm-log.c log->header.version = MIRROR_DISK_VERSION; log 330 drivers/md/dm-log.c log->header.nr_regions = 0; log 334 drivers/md/dm-log.c if (log->header.version == 1) log 335 drivers/md/dm-log.c log->header.version = 2; log 338 drivers/md/dm-log.c if (log->header.version != MIRROR_DISK_VERSION) { log 363 drivers/md/dm-log.c static int create_log_context(struct dm_dirty_log *log, struct dm_target *ti, log 511 drivers/md/dm-log.c log->context = lc; log 516 drivers/md/dm-log.c static int core_ctr(struct dm_dirty_log *log, struct dm_target *ti, log 519 drivers/md/dm-log.c return create_log_context(log, ti, argc, argv, NULL); log 529 drivers/md/dm-log.c static void core_dtr(struct dm_dirty_log *log) log 531 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 542 drivers/md/dm-log.c static int disk_ctr(struct dm_dirty_log *log, struct dm_target *ti, log 557 drivers/md/dm-log.c r = create_log_context(log, ti, argc - 1, argv + 1, dev); log 566 drivers/md/dm-log.c static void disk_dtr(struct dm_dirty_log *log) log 568 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 585 drivers/md/dm-log.c static int disk_resume(struct dm_dirty_log *log) log 589 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 649 drivers/md/dm-log.c static uint32_t core_get_region_size(struct dm_dirty_log *log) log 651 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 655 drivers/md/dm-log.c static int core_resume(struct dm_dirty_log *log) log 657 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 662 drivers/md/dm-log.c static int core_is_clean(struct dm_dirty_log *log, region_t region) log 664 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 668 drivers/md/dm-log.c static int core_in_sync(struct dm_dirty_log *log, region_t region, int block) log 670 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 674 drivers/md/dm-log.c static int core_flush(struct dm_dirty_log *log) log 680 drivers/md/dm-log.c static int disk_flush(struct dm_dirty_log *log) log 683 drivers/md/dm-log.c struct log_c *lc = log->context; log 689 drivers/md/dm-log.c if (lc->touched_cleaned && log->flush_callback_fn && log 690 drivers/md/dm-log.c log->flush_callback_fn(lc->ti)) { log 720 drivers/md/dm-log.c static void core_mark_region(struct dm_dirty_log *log, region_t region) log 722 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 726 drivers/md/dm-log.c static void core_clear_region(struct dm_dirty_log *log, region_t region) log 728 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 733 drivers/md/dm-log.c static int core_get_resync_work(struct dm_dirty_log *log, region_t *region) log 735 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 755 drivers/md/dm-log.c static void core_set_region_sync(struct dm_dirty_log *log, region_t region, log 758 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 770 drivers/md/dm-log.c static region_t core_get_sync_count(struct dm_dirty_log *log) log 772 drivers/md/dm-log.c struct log_c *lc = (struct log_c *) log->context; log 781 drivers/md/dm-log.c static int core_status(struct dm_dirty_log *log, status_type_t status, log 785 drivers/md/dm-log.c struct log_c *lc = log->context; log 789 drivers/md/dm-log.c DMEMIT("1 %s", log->type->name); log 793 drivers/md/dm-log.c DMEMIT("%s %u %u ", log->type->name, log 801 drivers/md/dm-log.c static int disk_status(struct dm_dirty_log *log, status_type_t status, log 805 drivers/md/dm-log.c struct log_c *lc = log->context; log 809 drivers/md/dm-log.c DMEMIT("3 %s %s %c", log->type->name, lc->log_dev->name, log 816 drivers/md/dm-log.c DMEMIT("%s %u %s %u ", log->type->name, log 387 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); log 404 drivers/md/dm-raid1.c (log->type->get_sync_count(log) == ms->nr_regions)) { log 439 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); log 442 drivers/md/dm-raid1.c if (log->type->in_sync(log, region, 0)) log 685 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); log 708 drivers/md/dm-raid1.c if (log->type->is_remote_recovering && log 709 drivers/md/dm-raid1.c log->type->is_remote_recovering(log, region)) { log 1189 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); log 1202 drivers/md/dm-raid1.c r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0); log 1303 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); log 1333 drivers/md/dm-raid1.c if (log->type->presuspend && log->type->presuspend(log)) log 1349 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); log 1351 drivers/md/dm-raid1.c if (log->type->postsuspend && log->type->postsuspend(log)) log 1359 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); log 1362 drivers/md/dm-raid1.c if (log->type->resume && log->type->resume(log)) log 1399 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); log 1412 drivers/md/dm-raid1.c (unsigned long long)log->type->get_sync_count(log), log 1415 drivers/md/dm-raid1.c sz += log->type->status(log, type, result+sz, maxlen-sz); log 1420 drivers/md/dm-raid1.c sz = log->type->status(log, type, result, maxlen); log 62 drivers/md/dm-region-hash.c struct dm_dirty_log *log; log 167 drivers/md/dm-region-hash.c struct dm_dirty_log *log, uint32_t region_size, log 196 drivers/md/dm-region-hash.c rh->log = log; log 251 drivers/md/dm-region-hash.c if (rh->log) log 252 drivers/md/dm-region-hash.c dm_dirty_log_destroy(rh->log); log 262 drivers/md/dm-region-hash.c return rh->log; log 296 drivers/md/dm-region-hash.c nreg->state = rh->log->type->in_sync(rh->log, region, 1) ? log 354 drivers/md/dm-region-hash.c r = rh->log->type->in_sync(rh->log, region, may_block); log 368 drivers/md/dm-region-hash.c rh->log->type->set_region_sync(rh->log, reg->key, success); log 398 drivers/md/dm-region-hash.c struct dm_dirty_log *log = rh->log; log 412 drivers/md/dm-region-hash.c log->type->set_region_sync(log, region, 0); log 484 drivers/md/dm-region-hash.c rh->log->type->clear_region(rh->log, reg->key); log 495 drivers/md/dm-region-hash.c rh->log->type->clear_region(rh->log, reg->key); log 499 drivers/md/dm-region-hash.c rh->log->type->flush(rh->log); log 518 drivers/md/dm-region-hash.c rh->log->type->mark_region(rh->log, reg->key); log 596 drivers/md/dm-region-hash.c r = rh->log->type->get_resync_work(rh->log, ®ion); log 686 drivers/md/dm-region-hash.c return rh->log->type->flush(rh->log); log 212 drivers/md/raid5-cache.c struct r5l_log *log; log 252 drivers/md/raid5-cache.c bool r5c_is_writeback(struct r5l_log *log) log 254 drivers/md/raid5-cache.c return (log != NULL && log 255 drivers/md/raid5-cache.c log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK); log 258 drivers/md/raid5-cache.c static sector_t r5l_ring_add(struct r5l_log *log, sector_t start, sector_t inc) log 261 drivers/md/raid5-cache.c if (start >= log->device_size) log 262 drivers/md/raid5-cache.c start = start - log->device_size; log 266 drivers/md/raid5-cache.c static sector_t r5l_ring_distance(struct r5l_log *log, sector_t start, log 272 drivers/md/raid5-cache.c return end + log->device_size - start; log 275 drivers/md/raid5-cache.c static bool r5l_has_free_space(struct r5l_log *log, sector_t size) log 279 drivers/md/raid5-cache.c used_size = r5l_ring_distance(log, log->last_checkpoint, log 280 drivers/md/raid5-cache.c log->log_start); log 282 drivers/md/raid5-cache.c return log->device_size > used_size + size; log 326 drivers/md/raid5-cache.c void r5l_wake_reclaim(struct r5l_log *log, sector_t space); log 333 drivers/md/raid5-cache.c if (!r5c_is_writeback(conf->log)) log 349 drivers/md/raid5-cache.c r5l_wake_reclaim(conf->log, 0); log 358 drivers/md/raid5-cache.c if (!r5c_is_writeback(conf->log)) log 368 drivers/md/raid5-cache.c r5l_wake_reclaim(conf->log, 0); log 401 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 403 drivers/md/raid5-cache.c if (!r5c_is_writeback(log)) log 407 drivers/md/raid5-cache.c ((conf->max_degraded + 1) * atomic_read(&log->stripe_in_journal_count) + log 418 drivers/md/raid5-cache.c static inline void r5c_update_log_state(struct r5l_log *log) log 420 drivers/md/raid5-cache.c struct r5conf *conf = log->rdev->mddev->private; log 425 drivers/md/raid5-cache.c if (!r5c_is_writeback(log)) log 428 drivers/md/raid5-cache.c free_space = r5l_ring_distance(log, log->log_start, log 429 drivers/md/raid5-cache.c log->last_checkpoint); log 444 drivers/md/raid5-cache.c r5l_wake_reclaim(log, 0); log 454 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 456 drivers/md/raid5-cache.c BUG_ON(!r5c_is_writeback(log)); log 496 drivers/md/raid5-cache.c struct r5l_log *log = sh->raid_conf->log; log 498 drivers/md/raid5-cache.c if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) { log 529 drivers/md/raid5-cache.c static void r5l_log_run_stripes(struct r5l_log *log) log 533 drivers/md/raid5-cache.c lockdep_assert_held(&log->io_list_lock); log 535 drivers/md/raid5-cache.c list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) { log 540 drivers/md/raid5-cache.c list_move_tail(&io->log_sibling, &log->finished_ios); log 545 drivers/md/raid5-cache.c static void r5l_move_to_end_ios(struct r5l_log *log) log 549 drivers/md/raid5-cache.c lockdep_assert_held(&log->io_list_lock); log 551 drivers/md/raid5-cache.c list_for_each_entry_safe(io, next, &log->running_ios, log_sibling) { log 555 drivers/md/raid5-cache.c list_move_tail(&io->log_sibling, &log->io_end_ios); log 564 drivers/md/raid5-cache.c struct r5l_log *log = io->log; log 570 drivers/md/raid5-cache.c md_error(log->rdev->mddev, log->rdev); log 573 drivers/md/raid5-cache.c mempool_free(io->meta_page, &log->meta_pool); log 575 drivers/md/raid5-cache.c spin_lock_irqsave(&log->io_list_lock, flags); log 587 drivers/md/raid5-cache.c if (log->need_cache_flush && !list_empty(&io->stripe_list)) log 588 drivers/md/raid5-cache.c r5l_move_to_end_ios(log); log 590 drivers/md/raid5-cache.c r5l_log_run_stripes(log); log 591 drivers/md/raid5-cache.c if (!list_empty(&log->running_ios)) { log 596 drivers/md/raid5-cache.c io_deferred = list_first_entry(&log->running_ios, log 599 drivers/md/raid5-cache.c schedule_work(&log->deferred_io_work); log 602 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->io_list_lock, flags); log 604 drivers/md/raid5-cache.c if (log->need_cache_flush) log 605 drivers/md/raid5-cache.c md_wakeup_thread(log->rdev->mddev->thread); log 626 drivers/md/raid5-cache.c static void r5l_do_submit_io(struct r5l_log *log, struct r5l_io_unit *io) log 630 drivers/md/raid5-cache.c spin_lock_irqsave(&log->io_list_lock, flags); log 632 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->io_list_lock, flags); log 663 drivers/md/raid5-cache.c struct r5l_log *log = container_of(work, struct r5l_log, log 668 drivers/md/raid5-cache.c spin_lock_irqsave(&log->io_list_lock, flags); log 669 drivers/md/raid5-cache.c if (!list_empty(&log->running_ios)) { log 670 drivers/md/raid5-cache.c io = list_first_entry(&log->running_ios, struct r5l_io_unit, log 677 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->io_list_lock, flags); log 679 drivers/md/raid5-cache.c r5l_do_submit_io(log, io); log 684 drivers/md/raid5-cache.c struct r5l_log *log = container_of(work, struct r5l_log, log 686 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 690 drivers/md/raid5-cache.c if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) log 697 drivers/md/raid5-cache.c conf->log == NULL || log 702 drivers/md/raid5-cache.c log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH; log 708 drivers/md/raid5-cache.c static void r5l_submit_current_io(struct r5l_log *log) log 710 drivers/md/raid5-cache.c struct r5l_io_unit *io = log->current_io; log 721 drivers/md/raid5-cache.c crc = crc32c_le(log->uuid_checksum, block, PAGE_SIZE); log 724 drivers/md/raid5-cache.c log->current_io = NULL; log 725 drivers/md/raid5-cache.c spin_lock_irqsave(&log->io_list_lock, flags); log 727 drivers/md/raid5-cache.c if (io != list_first_entry(&log->running_ios, log 733 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->io_list_lock, flags); log 735 drivers/md/raid5-cache.c r5l_do_submit_io(log, io); log 738 drivers/md/raid5-cache.c static struct bio *r5l_bio_alloc(struct r5l_log *log) log 740 drivers/md/raid5-cache.c struct bio *bio = bio_alloc_bioset(GFP_NOIO, BIO_MAX_PAGES, &log->bs); log 743 drivers/md/raid5-cache.c bio_set_dev(bio, log->rdev->bdev); log 744 drivers/md/raid5-cache.c bio->bi_iter.bi_sector = log->rdev->data_offset + log->log_start; log 749 drivers/md/raid5-cache.c static void r5_reserve_log_entry(struct r5l_log *log, struct r5l_io_unit *io) log 751 drivers/md/raid5-cache.c log->log_start = r5l_ring_add(log, log->log_start, BLOCK_SECTORS); log 753 drivers/md/raid5-cache.c r5c_update_log_state(log); log 761 drivers/md/raid5-cache.c if (log->log_start == 0) log 764 drivers/md/raid5-cache.c io->log_end = log->log_start; log 767 drivers/md/raid5-cache.c static struct r5l_io_unit *r5l_new_meta(struct r5l_log *log) log 772 drivers/md/raid5-cache.c io = mempool_alloc(&log->io_pool, GFP_ATOMIC); log 777 drivers/md/raid5-cache.c io->log = log; log 783 drivers/md/raid5-cache.c io->meta_page = mempool_alloc(&log->meta_pool, GFP_NOIO); log 788 drivers/md/raid5-cache.c block->seq = cpu_to_le64(log->seq); log 789 drivers/md/raid5-cache.c block->position = cpu_to_le64(log->log_start); log 791 drivers/md/raid5-cache.c io->log_start = log->log_start; log 793 drivers/md/raid5-cache.c io->seq = log->seq++; log 795 drivers/md/raid5-cache.c io->current_bio = r5l_bio_alloc(log); log 800 drivers/md/raid5-cache.c r5_reserve_log_entry(log, io); log 802 drivers/md/raid5-cache.c spin_lock_irq(&log->io_list_lock); log 803 drivers/md/raid5-cache.c list_add_tail(&io->log_sibling, &log->running_ios); log 804 drivers/md/raid5-cache.c spin_unlock_irq(&log->io_list_lock); log 809 drivers/md/raid5-cache.c static int r5l_get_meta(struct r5l_log *log, unsigned int payload_size) log 811 drivers/md/raid5-cache.c if (log->current_io && log 812 drivers/md/raid5-cache.c log->current_io->meta_offset + payload_size > PAGE_SIZE) log 813 drivers/md/raid5-cache.c r5l_submit_current_io(log); log 815 drivers/md/raid5-cache.c if (!log->current_io) { log 816 drivers/md/raid5-cache.c log->current_io = r5l_new_meta(log); log 817 drivers/md/raid5-cache.c if (!log->current_io) log 824 drivers/md/raid5-cache.c static void r5l_append_payload_meta(struct r5l_log *log, u16 type, log 829 drivers/md/raid5-cache.c struct r5l_io_unit *io = log->current_io; log 846 drivers/md/raid5-cache.c static void r5l_append_payload_page(struct r5l_log *log, struct page *page) log 848 drivers/md/raid5-cache.c struct r5l_io_unit *io = log->current_io; log 853 drivers/md/raid5-cache.c io->current_bio = r5l_bio_alloc(log); log 861 drivers/md/raid5-cache.c r5_reserve_log_entry(log, io); log 864 drivers/md/raid5-cache.c static void r5l_append_flush_payload(struct r5l_log *log, sector_t sect) log 866 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 880 drivers/md/raid5-cache.c mutex_lock(&log->io_mutex); log 883 drivers/md/raid5-cache.c if (r5l_get_meta(log, meta_size)) { log 884 drivers/md/raid5-cache.c mutex_unlock(&log->io_mutex); log 889 drivers/md/raid5-cache.c io = log->current_io; log 901 drivers/md/raid5-cache.c mutex_unlock(&log->io_mutex); log 904 drivers/md/raid5-cache.c static int r5l_log_stripe(struct r5l_log *log, struct stripe_head *sh, log 918 drivers/md/raid5-cache.c ret = r5l_get_meta(log, meta_size); log 922 drivers/md/raid5-cache.c io = log->current_io; log 934 drivers/md/raid5-cache.c log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK) { log 942 drivers/md/raid5-cache.c r5l_append_payload_meta(log, R5LOG_PAYLOAD_DATA, log 945 drivers/md/raid5-cache.c r5l_append_payload_page(log, sh->dev[i].page); log 949 drivers/md/raid5-cache.c r5l_append_payload_meta(log, R5LOG_PAYLOAD_PARITY, log 952 drivers/md/raid5-cache.c r5l_append_payload_page(log, sh->dev[sh->pd_idx].page); log 953 drivers/md/raid5-cache.c r5l_append_payload_page(log, sh->dev[sh->qd_idx].page); log 955 drivers/md/raid5-cache.c r5l_append_payload_meta(log, R5LOG_PAYLOAD_PARITY, log 958 drivers/md/raid5-cache.c r5l_append_payload_page(log, sh->dev[sh->pd_idx].page); log 966 drivers/md/raid5-cache.c if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) log 972 drivers/md/raid5-cache.c spin_lock_irq(&log->stripe_in_journal_lock); log 974 drivers/md/raid5-cache.c &log->stripe_in_journal_list); log 975 drivers/md/raid5-cache.c spin_unlock_irq(&log->stripe_in_journal_lock); log 976 drivers/md/raid5-cache.c atomic_inc(&log->stripe_in_journal_count); log 982 drivers/md/raid5-cache.c static inline void r5l_add_no_space_stripe(struct r5l_log *log, log 985 drivers/md/raid5-cache.c spin_lock(&log->no_space_stripes_lock); log 986 drivers/md/raid5-cache.c list_add_tail(&sh->log_list, &log->no_space_stripes); log 987 drivers/md/raid5-cache.c spin_unlock(&log->no_space_stripes_lock); log 994 drivers/md/raid5-cache.c int r5l_write_stripe(struct r5l_log *log, struct stripe_head *sh) log 1004 drivers/md/raid5-cache.c if (!log) log 1028 drivers/md/raid5-cache.c sh->dev[i].log_checksum = crc32c_le(log->uuid_checksum, log 1043 drivers/md/raid5-cache.c mutex_lock(&log->io_mutex); log 1047 drivers/md/raid5-cache.c if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) { log 1048 drivers/md/raid5-cache.c if (!r5l_has_free_space(log, reserve)) { log 1049 drivers/md/raid5-cache.c r5l_add_no_space_stripe(log, sh); log 1052 drivers/md/raid5-cache.c ret = r5l_log_stripe(log, sh, data_pages, parity_pages); log 1054 drivers/md/raid5-cache.c spin_lock_irq(&log->io_list_lock); log 1056 drivers/md/raid5-cache.c &log->no_mem_stripes); log 1057 drivers/md/raid5-cache.c spin_unlock_irq(&log->io_list_lock); log 1067 drivers/md/raid5-cache.c r5l_add_no_space_stripe(log, sh); log 1070 drivers/md/raid5-cache.c } else if (!r5l_has_free_space(log, reserve)) { log 1071 drivers/md/raid5-cache.c if (sh->log_start == log->last_checkpoint) log 1074 drivers/md/raid5-cache.c r5l_add_no_space_stripe(log, sh); log 1076 drivers/md/raid5-cache.c ret = r5l_log_stripe(log, sh, data_pages, parity_pages); log 1078 drivers/md/raid5-cache.c spin_lock_irq(&log->io_list_lock); log 1080 drivers/md/raid5-cache.c &log->no_mem_stripes); log 1081 drivers/md/raid5-cache.c spin_unlock_irq(&log->io_list_lock); log 1086 drivers/md/raid5-cache.c mutex_unlock(&log->io_mutex); log 1088 drivers/md/raid5-cache.c r5l_wake_reclaim(log, reserve); log 1092 drivers/md/raid5-cache.c void r5l_write_stripe_run(struct r5l_log *log) log 1094 drivers/md/raid5-cache.c if (!log) log 1096 drivers/md/raid5-cache.c mutex_lock(&log->io_mutex); log 1097 drivers/md/raid5-cache.c r5l_submit_current_io(log); log 1098 drivers/md/raid5-cache.c mutex_unlock(&log->io_mutex); log 1101 drivers/md/raid5-cache.c int r5l_handle_flush_request(struct r5l_log *log, struct bio *bio) log 1103 drivers/md/raid5-cache.c if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) { log 1119 drivers/md/raid5-cache.c mutex_lock(&log->io_mutex); log 1120 drivers/md/raid5-cache.c r5l_get_meta(log, 0); log 1121 drivers/md/raid5-cache.c bio_list_add(&log->current_io->flush_barriers, bio); log 1122 drivers/md/raid5-cache.c log->current_io->has_flush = 1; log 1123 drivers/md/raid5-cache.c log->current_io->has_null_flush = 1; log 1124 drivers/md/raid5-cache.c atomic_inc(&log->current_io->pending_stripe); log 1125 drivers/md/raid5-cache.c r5l_submit_current_io(log); log 1126 drivers/md/raid5-cache.c mutex_unlock(&log->io_mutex); log 1134 drivers/md/raid5-cache.c static void r5l_run_no_space_stripes(struct r5l_log *log) log 1138 drivers/md/raid5-cache.c spin_lock(&log->no_space_stripes_lock); log 1139 drivers/md/raid5-cache.c while (!list_empty(&log->no_space_stripes)) { log 1140 drivers/md/raid5-cache.c sh = list_first_entry(&log->no_space_stripes, log 1146 drivers/md/raid5-cache.c spin_unlock(&log->no_space_stripes_lock); log 1157 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 1161 drivers/md/raid5-cache.c if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) log 1162 drivers/md/raid5-cache.c return log->next_checkpoint; log 1164 drivers/md/raid5-cache.c spin_lock_irqsave(&log->stripe_in_journal_lock, flags); log 1165 drivers/md/raid5-cache.c if (list_empty(&conf->log->stripe_in_journal_list)) { log 1167 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags); log 1168 drivers/md/raid5-cache.c return log->next_checkpoint; log 1170 drivers/md/raid5-cache.c sh = list_first_entry(&conf->log->stripe_in_journal_list, log 1173 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags); log 1177 drivers/md/raid5-cache.c static sector_t r5l_reclaimable_space(struct r5l_log *log) log 1179 drivers/md/raid5-cache.c struct r5conf *conf = log->rdev->mddev->private; log 1181 drivers/md/raid5-cache.c return r5l_ring_distance(log, log->last_checkpoint, log 1185 drivers/md/raid5-cache.c static void r5l_run_no_mem_stripe(struct r5l_log *log) log 1189 drivers/md/raid5-cache.c lockdep_assert_held(&log->io_list_lock); log 1191 drivers/md/raid5-cache.c if (!list_empty(&log->no_mem_stripes)) { log 1192 drivers/md/raid5-cache.c sh = list_first_entry(&log->no_mem_stripes, log 1200 drivers/md/raid5-cache.c static bool r5l_complete_finished_ios(struct r5l_log *log) log 1205 drivers/md/raid5-cache.c lockdep_assert_held(&log->io_list_lock); log 1207 drivers/md/raid5-cache.c list_for_each_entry_safe(io, next, &log->finished_ios, log_sibling) { log 1212 drivers/md/raid5-cache.c log->next_checkpoint = io->log_start; log 1215 drivers/md/raid5-cache.c mempool_free(io, &log->io_pool); log 1216 drivers/md/raid5-cache.c r5l_run_no_mem_stripe(log); log 1226 drivers/md/raid5-cache.c struct r5l_log *log = io->log; log 1227 drivers/md/raid5-cache.c struct r5conf *conf = log->rdev->mddev->private; log 1230 drivers/md/raid5-cache.c spin_lock_irqsave(&log->io_list_lock, flags); log 1233 drivers/md/raid5-cache.c if (!r5l_complete_finished_ios(log)) { log 1234 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->io_list_lock, flags); log 1238 drivers/md/raid5-cache.c if (r5l_reclaimable_space(log) > log->max_free_space || log 1240 drivers/md/raid5-cache.c r5l_wake_reclaim(log, 0); log 1242 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->io_list_lock, flags); log 1243 drivers/md/raid5-cache.c wake_up(&log->iounit_wait); log 1259 drivers/md/raid5-cache.c struct r5l_log *log = container_of(bio, struct r5l_log, log 1265 drivers/md/raid5-cache.c md_error(log->rdev->mddev, log->rdev); log 1267 drivers/md/raid5-cache.c spin_lock_irqsave(&log->io_list_lock, flags); log 1268 drivers/md/raid5-cache.c list_for_each_entry(io, &log->flushing_ios, log_sibling) log 1270 drivers/md/raid5-cache.c list_splice_tail_init(&log->flushing_ios, &log->finished_ios); log 1271 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->io_list_lock, flags); log 1288 drivers/md/raid5-cache.c void r5l_flush_stripe_to_raid(struct r5l_log *log) log 1292 drivers/md/raid5-cache.c if (!log || !log->need_cache_flush) log 1295 drivers/md/raid5-cache.c spin_lock_irq(&log->io_list_lock); log 1297 drivers/md/raid5-cache.c if (!list_empty(&log->flushing_ios)) { log 1298 drivers/md/raid5-cache.c spin_unlock_irq(&log->io_list_lock); log 1301 drivers/md/raid5-cache.c list_splice_tail_init(&log->io_end_ios, &log->flushing_ios); log 1302 drivers/md/raid5-cache.c do_flush = !list_empty(&log->flushing_ios); log 1303 drivers/md/raid5-cache.c spin_unlock_irq(&log->io_list_lock); log 1307 drivers/md/raid5-cache.c bio_reset(&log->flush_bio); log 1308 drivers/md/raid5-cache.c bio_set_dev(&log->flush_bio, log->rdev->bdev); log 1309 drivers/md/raid5-cache.c log->flush_bio.bi_end_io = r5l_log_flush_endio; log 1310 drivers/md/raid5-cache.c log->flush_bio.bi_opf = REQ_OP_WRITE | REQ_PREFLUSH; log 1311 drivers/md/raid5-cache.c submit_bio(&log->flush_bio); log 1314 drivers/md/raid5-cache.c static void r5l_write_super(struct r5l_log *log, sector_t cp); log 1315 drivers/md/raid5-cache.c static void r5l_write_super_and_discard_space(struct r5l_log *log, log 1318 drivers/md/raid5-cache.c struct block_device *bdev = log->rdev->bdev; log 1321 drivers/md/raid5-cache.c r5l_write_super(log, end); log 1326 drivers/md/raid5-cache.c mddev = log->rdev->mddev; log 1346 drivers/md/raid5-cache.c if (log->last_checkpoint < end) { log 1348 drivers/md/raid5-cache.c log->last_checkpoint + log->rdev->data_offset, log 1349 drivers/md/raid5-cache.c end - log->last_checkpoint, GFP_NOIO, 0); log 1352 drivers/md/raid5-cache.c log->last_checkpoint + log->rdev->data_offset, log 1353 drivers/md/raid5-cache.c log->device_size - log->last_checkpoint, log 1355 drivers/md/raid5-cache.c blkdev_issue_discard(bdev, log->rdev->data_offset, end, log 1405 drivers/md/raid5-cache.c if (!conf->log) log 1426 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 1434 drivers/md/raid5-cache.c if (!r5c_is_writeback(log)) log 1470 drivers/md/raid5-cache.c spin_lock_irqsave(&log->stripe_in_journal_lock, flags); log 1472 drivers/md/raid5-cache.c list_for_each_entry(sh, &log->stripe_in_journal_list, r5c) { log 1490 drivers/md/raid5-cache.c spin_unlock_irqrestore(&log->stripe_in_journal_lock, flags); log 1494 drivers/md/raid5-cache.c r5l_run_no_space_stripes(log); log 1499 drivers/md/raid5-cache.c static void r5l_do_reclaim(struct r5l_log *log) log 1501 drivers/md/raid5-cache.c struct r5conf *conf = log->rdev->mddev->private; log 1502 drivers/md/raid5-cache.c sector_t reclaim_target = xchg(&log->reclaim_target, 0); log 1507 drivers/md/raid5-cache.c spin_lock_irq(&log->io_list_lock); log 1508 drivers/md/raid5-cache.c write_super = r5l_reclaimable_space(log) > log->max_free_space || log 1509 drivers/md/raid5-cache.c reclaim_target != 0 || !list_empty(&log->no_space_stripes); log 1516 drivers/md/raid5-cache.c reclaimable = r5l_reclaimable_space(log); log 1518 drivers/md/raid5-cache.c (list_empty(&log->running_ios) && log 1519 drivers/md/raid5-cache.c list_empty(&log->io_end_ios) && log 1520 drivers/md/raid5-cache.c list_empty(&log->flushing_ios) && log 1521 drivers/md/raid5-cache.c list_empty(&log->finished_ios))) log 1524 drivers/md/raid5-cache.c md_wakeup_thread(log->rdev->mddev->thread); log 1525 drivers/md/raid5-cache.c wait_event_lock_irq(log->iounit_wait, log 1526 drivers/md/raid5-cache.c r5l_reclaimable_space(log) > reclaimable, log 1527 drivers/md/raid5-cache.c log->io_list_lock); log 1531 drivers/md/raid5-cache.c spin_unlock_irq(&log->io_list_lock); log 1541 drivers/md/raid5-cache.c r5l_write_super_and_discard_space(log, next_checkpoint); log 1543 drivers/md/raid5-cache.c mutex_lock(&log->io_mutex); log 1544 drivers/md/raid5-cache.c log->last_checkpoint = next_checkpoint; log 1545 drivers/md/raid5-cache.c r5c_update_log_state(log); log 1546 drivers/md/raid5-cache.c mutex_unlock(&log->io_mutex); log 1548 drivers/md/raid5-cache.c r5l_run_no_space_stripes(log); log 1555 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 1557 drivers/md/raid5-cache.c if (!log) log 1560 drivers/md/raid5-cache.c r5l_do_reclaim(log); log 1563 drivers/md/raid5-cache.c void r5l_wake_reclaim(struct r5l_log *log, sector_t space) log 1568 drivers/md/raid5-cache.c if (!log) log 1571 drivers/md/raid5-cache.c target = log->reclaim_target; log 1574 drivers/md/raid5-cache.c } while (cmpxchg(&log->reclaim_target, target, new) != target); log 1575 drivers/md/raid5-cache.c md_wakeup_thread(log->reclaim_thread); log 1578 drivers/md/raid5-cache.c void r5l_quiesce(struct r5l_log *log, int quiesce) log 1584 drivers/md/raid5-cache.c mddev = log->rdev->mddev; log 1586 drivers/md/raid5-cache.c kthread_park(log->reclaim_thread->tsk); log 1587 drivers/md/raid5-cache.c r5l_wake_reclaim(log, MaxSector); log 1588 drivers/md/raid5-cache.c r5l_do_reclaim(log); log 1590 drivers/md/raid5-cache.c kthread_unpark(log->reclaim_thread->tsk); log 1595 drivers/md/raid5-cache.c struct r5l_log *log; log 1599 drivers/md/raid5-cache.c log = rcu_dereference(conf->log); log 1601 drivers/md/raid5-cache.c if (!log) log 1604 drivers/md/raid5-cache.c ret = test_bit(Faulty, &log->rdev->flags); log 1634 drivers/md/raid5-cache.c static int r5l_recovery_allocate_ra_pool(struct r5l_log *log, log 1639 drivers/md/raid5-cache.c ctx->ra_bio = bio_alloc_bioset(GFP_KERNEL, BIO_MAX_PAGES, &log->bs); log 1663 drivers/md/raid5-cache.c static void r5l_recovery_free_ra_pool(struct r5l_log *log, log 1679 drivers/md/raid5-cache.c static int r5l_recovery_fetch_ra_pool(struct r5l_log *log, log 1684 drivers/md/raid5-cache.c bio_set_dev(ctx->ra_bio, log->rdev->bdev); log 1686 drivers/md/raid5-cache.c ctx->ra_bio->bi_iter.bi_sector = log->rdev->data_offset + offset; log 1696 drivers/md/raid5-cache.c offset = r5l_ring_add(log, offset, BLOCK_SECTORS); log 1709 drivers/md/raid5-cache.c static int r5l_recovery_read_page(struct r5l_log *log, log 1718 drivers/md/raid5-cache.c ret = r5l_recovery_fetch_ra_pool(log, ctx, offset); log 1733 drivers/md/raid5-cache.c static int r5l_recovery_read_meta_block(struct r5l_log *log, log 1741 drivers/md/raid5-cache.c ret = r5l_recovery_read_page(log, ctx, page, ctx->pos); log 1755 drivers/md/raid5-cache.c crc = crc32c_le(log->uuid_checksum, mb, PAGE_SIZE); log 1768 drivers/md/raid5-cache.c r5l_recovery_create_empty_meta_block(struct r5l_log *log, log 1783 drivers/md/raid5-cache.c static int r5l_log_write_empty_meta_block(struct r5l_log *log, sector_t pos, log 1792 drivers/md/raid5-cache.c r5l_recovery_create_empty_meta_block(log, page, pos, seq); log 1794 drivers/md/raid5-cache.c mb->checksum = cpu_to_le32(crc32c_le(log->uuid_checksum, log 1796 drivers/md/raid5-cache.c if (!sync_page_io(log->rdev, pos, PAGE_SIZE, page, REQ_OP_WRITE, log 1812 drivers/md/raid5-cache.c static void r5l_recovery_load_data(struct r5l_log *log, log 1818 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 1825 drivers/md/raid5-cache.c r5l_recovery_read_page(log, ctx, sh->dev[dd_idx].page, log_offset); log 1834 drivers/md/raid5-cache.c static void r5l_recovery_load_parity(struct r5l_log *log, log 1840 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 1844 drivers/md/raid5-cache.c r5l_recovery_read_page(log, ctx, sh->dev[sh->pd_idx].page, log_offset); log 1851 drivers/md/raid5-cache.c log, ctx, sh->dev[sh->qd_idx].page, log 1852 drivers/md/raid5-cache.c r5l_ring_add(log, log_offset, BLOCK_SECTORS)); log 1985 drivers/md/raid5-cache.c r5l_recovery_verify_data_checksum(struct r5l_log *log, log 1993 drivers/md/raid5-cache.c r5l_recovery_read_page(log, ctx, page, log_offset); log 1995 drivers/md/raid5-cache.c checksum = crc32c_le(log->uuid_checksum, addr, PAGE_SIZE); log 2005 drivers/md/raid5-cache.c r5l_recovery_verify_data_checksum_for_mb(struct r5l_log *log, log 2008 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 2012 drivers/md/raid5-cache.c sector_t log_offset = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS); log 2027 drivers/md/raid5-cache.c log, ctx, page, log_offset, log 2032 drivers/md/raid5-cache.c log, ctx, page, log_offset, log 2037 drivers/md/raid5-cache.c log, ctx, page, log 2038 drivers/md/raid5-cache.c r5l_ring_add(log, log_offset, log 2052 drivers/md/raid5-cache.c log_offset = r5l_ring_add(log, log_offset, log 2078 drivers/md/raid5-cache.c r5c_recovery_analyze_meta_block(struct r5l_log *log, log 2082 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 2098 drivers/md/raid5-cache.c ret = r5l_recovery_verify_data_checksum_for_mb(log, ctx); log 2106 drivers/md/raid5-cache.c log_offset = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS); log 2189 drivers/md/raid5-cache.c r5l_recovery_load_data(log, sh, ctx, payload, log 2192 drivers/md/raid5-cache.c r5l_recovery_load_parity(log, sh, ctx, payload, log 2197 drivers/md/raid5-cache.c log_offset = r5l_ring_add(log, log_offset, log 2212 drivers/md/raid5-cache.c static void r5c_recovery_load_one_stripe(struct r5l_log *log, log 2243 drivers/md/raid5-cache.c static int r5c_recovery_flush_log(struct r5l_log *log, log 2251 drivers/md/raid5-cache.c if (r5l_recovery_read_meta_block(log, ctx)) log 2254 drivers/md/raid5-cache.c ret = r5c_recovery_analyze_meta_block(log, ctx, log 2263 drivers/md/raid5-cache.c ctx->pos = r5l_ring_add(log, ctx->pos, ctx->meta_total_blocks); log 2277 drivers/md/raid5-cache.c r5c_recovery_load_one_stripe(log, sh); log 2354 drivers/md/raid5-cache.c r5c_recovery_rewrite_data_only_stripes(struct r5l_log *log, log 2358 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 2378 drivers/md/raid5-cache.c r5l_recovery_create_empty_meta_block(log, page, log 2382 drivers/md/raid5-cache.c write_pos = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS); log 2398 drivers/md/raid5-cache.c crc32c_le(log->uuid_checksum, addr, log 2401 drivers/md/raid5-cache.c sync_page_io(log->rdev, write_pos, PAGE_SIZE, log 2403 drivers/md/raid5-cache.c write_pos = r5l_ring_add(log, write_pos, log 2411 drivers/md/raid5-cache.c mb->checksum = cpu_to_le32(crc32c_le(log->uuid_checksum, log 2413 drivers/md/raid5-cache.c sync_page_io(log->rdev, ctx->pos, PAGE_SIZE, page, log 2416 drivers/md/raid5-cache.c list_add_tail(&sh->r5c, &log->stripe_in_journal_list); log 2417 drivers/md/raid5-cache.c atomic_inc(&log->stripe_in_journal_count); log 2422 drivers/md/raid5-cache.c log->next_checkpoint = next_checkpoint; log 2427 drivers/md/raid5-cache.c static void r5c_recovery_flush_data_only_stripes(struct r5l_log *log, log 2430 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 2437 drivers/md/raid5-cache.c log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_BACK; log 2450 drivers/md/raid5-cache.c log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH; log 2453 drivers/md/raid5-cache.c static int r5l_recovery_log(struct r5l_log *log) log 2455 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 2464 drivers/md/raid5-cache.c ctx->pos = log->last_checkpoint; log 2465 drivers/md/raid5-cache.c ctx->seq = log->last_cp_seq; log 2474 drivers/md/raid5-cache.c if (r5l_recovery_allocate_ra_pool(log, ctx) != 0) { log 2479 drivers/md/raid5-cache.c ret = r5c_recovery_flush_log(log, ctx); log 2496 drivers/md/raid5-cache.c log->next_checkpoint = ctx->pos; log 2497 drivers/md/raid5-cache.c r5l_log_write_empty_meta_block(log, ctx->pos, ctx->seq++); log 2498 drivers/md/raid5-cache.c ctx->pos = r5l_ring_add(log, ctx->pos, BLOCK_SECTORS); log 2499 drivers/md/raid5-cache.c } else if (r5c_recovery_rewrite_data_only_stripes(log, ctx)) { log 2506 drivers/md/raid5-cache.c log->log_start = ctx->pos; log 2507 drivers/md/raid5-cache.c log->seq = ctx->seq; log 2508 drivers/md/raid5-cache.c log->last_checkpoint = pos; log 2509 drivers/md/raid5-cache.c r5l_write_super(log, pos); log 2511 drivers/md/raid5-cache.c r5c_recovery_flush_data_only_stripes(log, ctx); log 2514 drivers/md/raid5-cache.c r5l_recovery_free_ra_pool(log, ctx); log 2522 drivers/md/raid5-cache.c static void r5l_write_super(struct r5l_log *log, sector_t cp) log 2524 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 2526 drivers/md/raid5-cache.c log->rdev->journal_tail = cp; log 2540 drivers/md/raid5-cache.c if (!conf || !conf->log) { log 2545 drivers/md/raid5-cache.c switch (conf->log->r5c_journal_mode) { log 2580 drivers/md/raid5-cache.c if (!conf || !conf->log) log 2588 drivers/md/raid5-cache.c conf->log->r5c_journal_mode = mode; log 2639 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 2648 drivers/md/raid5-cache.c BUG_ON(!r5c_is_writeback(log)); log 2698 drivers/md/raid5-cache.c spin_lock(&log->tree_lock); log 2699 drivers/md/raid5-cache.c pslot = radix_tree_lookup_slot(&log->big_stripe_tree, log 2703 drivers/md/raid5-cache.c pslot, &log->tree_lock) >> log 2706 drivers/md/raid5-cache.c &log->big_stripe_tree, pslot, log 2714 drivers/md/raid5-cache.c &log->big_stripe_tree, tree_index, log 2717 drivers/md/raid5-cache.c spin_unlock(&log->tree_lock); log 2722 drivers/md/raid5-cache.c spin_unlock(&log->tree_lock); log 2806 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 2813 drivers/md/raid5-cache.c if (!log || !test_bit(R5_InJournal, &sh->dev[sh->pd_idx].flags)) log 2819 drivers/md/raid5-cache.c if (log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_THROUGH) log 2841 drivers/md/raid5-cache.c spin_lock_irq(&log->stripe_in_journal_lock); log 2843 drivers/md/raid5-cache.c spin_unlock_irq(&log->stripe_in_journal_lock); log 2846 drivers/md/raid5-cache.c atomic_dec(&log->stripe_in_journal_count); log 2847 drivers/md/raid5-cache.c r5c_update_log_state(log); log 2853 drivers/md/raid5-cache.c spin_lock(&log->tree_lock); log 2854 drivers/md/raid5-cache.c pslot = radix_tree_lookup_slot(&log->big_stripe_tree, log 2858 drivers/md/raid5-cache.c pslot, &log->tree_lock) >> log 2861 drivers/md/raid5-cache.c radix_tree_delete(&log->big_stripe_tree, tree_index); log 2864 drivers/md/raid5-cache.c &log->big_stripe_tree, pslot, log 2866 drivers/md/raid5-cache.c spin_unlock(&log->tree_lock); log 2881 drivers/md/raid5-cache.c r5l_append_flush_payload(log, sh->sector); log 2887 drivers/md/raid5-cache.c int r5c_cache_data(struct r5l_log *log, struct stripe_head *sh) log 2895 drivers/md/raid5-cache.c BUG_ON(!log); log 2903 drivers/md/raid5-cache.c sh->dev[i].log_checksum = crc32c_le(log->uuid_checksum, log 2917 drivers/md/raid5-cache.c mutex_lock(&log->io_mutex); log 2923 drivers/md/raid5-cache.c r5l_add_no_space_stripe(log, sh); log 2924 drivers/md/raid5-cache.c else if (!r5l_has_free_space(log, reserve)) { log 2925 drivers/md/raid5-cache.c if (sh->log_start == log->last_checkpoint) log 2928 drivers/md/raid5-cache.c r5l_add_no_space_stripe(log, sh); log 2930 drivers/md/raid5-cache.c ret = r5l_log_stripe(log, sh, pages, 0); log 2932 drivers/md/raid5-cache.c spin_lock_irq(&log->io_list_lock); log 2933 drivers/md/raid5-cache.c list_add_tail(&sh->log_list, &log->no_mem_stripes); log 2934 drivers/md/raid5-cache.c spin_unlock_irq(&log->io_list_lock); log 2938 drivers/md/raid5-cache.c mutex_unlock(&log->io_mutex); log 2945 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 2949 drivers/md/raid5-cache.c if (!log) log 2954 drivers/md/raid5-cache.c slot = radix_tree_lookup(&log->big_stripe_tree, tree_index); log 2958 drivers/md/raid5-cache.c static int r5l_load_log(struct r5l_log *log) log 2960 drivers/md/raid5-cache.c struct md_rdev *rdev = log->rdev; log 2963 drivers/md/raid5-cache.c sector_t cp = log->rdev->journal_tail; log 2988 drivers/md/raid5-cache.c expected_crc = crc32c_le(log->uuid_checksum, mb, PAGE_SIZE); log 2999 drivers/md/raid5-cache.c log->last_cp_seq = prandom_u32(); log 3001 drivers/md/raid5-cache.c r5l_log_write_empty_meta_block(log, cp, log->last_cp_seq); log 3007 drivers/md/raid5-cache.c r5l_write_super(log, cp); log 3009 drivers/md/raid5-cache.c log->last_cp_seq = le64_to_cpu(mb->seq); log 3011 drivers/md/raid5-cache.c log->device_size = round_down(rdev->sectors, BLOCK_SECTORS); log 3012 drivers/md/raid5-cache.c log->max_free_space = log->device_size >> RECLAIM_MAX_FREE_SPACE_SHIFT; log 3013 drivers/md/raid5-cache.c if (log->max_free_space > RECLAIM_MAX_FREE_SPACE) log 3014 drivers/md/raid5-cache.c log->max_free_space = RECLAIM_MAX_FREE_SPACE; log 3015 drivers/md/raid5-cache.c log->last_checkpoint = cp; log 3020 drivers/md/raid5-cache.c log->log_start = r5l_ring_add(log, cp, BLOCK_SECTORS); log 3021 drivers/md/raid5-cache.c log->seq = log->last_cp_seq + 1; log 3022 drivers/md/raid5-cache.c log->next_checkpoint = cp; log 3024 drivers/md/raid5-cache.c ret = r5l_recovery_log(log); log 3026 drivers/md/raid5-cache.c r5c_update_log_state(log); log 3033 drivers/md/raid5-cache.c int r5l_start(struct r5l_log *log) log 3037 drivers/md/raid5-cache.c if (!log) log 3040 drivers/md/raid5-cache.c ret = r5l_load_log(log); log 3042 drivers/md/raid5-cache.c struct mddev *mddev = log->rdev->mddev; log 3053 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 3055 drivers/md/raid5-cache.c if (!log) log 3060 drivers/md/raid5-cache.c conf->log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK) log 3061 drivers/md/raid5-cache.c schedule_work(&log->disable_writeback_work); log 3067 drivers/md/raid5-cache.c struct r5l_log *log; log 3092 drivers/md/raid5-cache.c log = kzalloc(sizeof(*log), GFP_KERNEL); log 3093 drivers/md/raid5-cache.c if (!log) log 3095 drivers/md/raid5-cache.c log->rdev = rdev; log 3097 drivers/md/raid5-cache.c log->need_cache_flush = test_bit(QUEUE_FLAG_WC, &q->queue_flags) != 0; log 3099 drivers/md/raid5-cache.c log->uuid_checksum = crc32c_le(~0, rdev->mddev->uuid, log 3102 drivers/md/raid5-cache.c mutex_init(&log->io_mutex); log 3104 drivers/md/raid5-cache.c spin_lock_init(&log->io_list_lock); log 3105 drivers/md/raid5-cache.c INIT_LIST_HEAD(&log->running_ios); log 3106 drivers/md/raid5-cache.c INIT_LIST_HEAD(&log->io_end_ios); log 3107 drivers/md/raid5-cache.c INIT_LIST_HEAD(&log->flushing_ios); log 3108 drivers/md/raid5-cache.c INIT_LIST_HEAD(&log->finished_ios); log 3109 drivers/md/raid5-cache.c bio_init(&log->flush_bio, NULL, 0); log 3111 drivers/md/raid5-cache.c log->io_kc = KMEM_CACHE(r5l_io_unit, 0); log 3112 drivers/md/raid5-cache.c if (!log->io_kc) log 3115 drivers/md/raid5-cache.c ret = mempool_init_slab_pool(&log->io_pool, R5L_POOL_SIZE, log->io_kc); log 3119 drivers/md/raid5-cache.c ret = bioset_init(&log->bs, R5L_POOL_SIZE, 0, BIOSET_NEED_BVECS); log 3123 drivers/md/raid5-cache.c ret = mempool_init_page_pool(&log->meta_pool, R5L_POOL_SIZE, 0); log 3127 drivers/md/raid5-cache.c spin_lock_init(&log->tree_lock); log 3128 drivers/md/raid5-cache.c INIT_RADIX_TREE(&log->big_stripe_tree, GFP_NOWAIT | __GFP_NOWARN); log 3130 drivers/md/raid5-cache.c log->reclaim_thread = md_register_thread(r5l_reclaim_thread, log 3131 drivers/md/raid5-cache.c log->rdev->mddev, "reclaim"); log 3132 drivers/md/raid5-cache.c if (!log->reclaim_thread) log 3134 drivers/md/raid5-cache.c log->reclaim_thread->timeout = R5C_RECLAIM_WAKEUP_INTERVAL; log 3136 drivers/md/raid5-cache.c init_waitqueue_head(&log->iounit_wait); log 3138 drivers/md/raid5-cache.c INIT_LIST_HEAD(&log->no_mem_stripes); log 3140 drivers/md/raid5-cache.c INIT_LIST_HEAD(&log->no_space_stripes); log 3141 drivers/md/raid5-cache.c spin_lock_init(&log->no_space_stripes_lock); log 3143 drivers/md/raid5-cache.c INIT_WORK(&log->deferred_io_work, r5l_submit_io_async); log 3144 drivers/md/raid5-cache.c INIT_WORK(&log->disable_writeback_work, r5c_disable_writeback_async); log 3146 drivers/md/raid5-cache.c log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH; log 3147 drivers/md/raid5-cache.c INIT_LIST_HEAD(&log->stripe_in_journal_list); log 3148 drivers/md/raid5-cache.c spin_lock_init(&log->stripe_in_journal_lock); log 3149 drivers/md/raid5-cache.c atomic_set(&log->stripe_in_journal_count, 0); log 3151 drivers/md/raid5-cache.c rcu_assign_pointer(conf->log, log); log 3157 drivers/md/raid5-cache.c mempool_exit(&log->meta_pool); log 3159 drivers/md/raid5-cache.c bioset_exit(&log->bs); log 3161 drivers/md/raid5-cache.c mempool_exit(&log->io_pool); log 3163 drivers/md/raid5-cache.c kmem_cache_destroy(log->io_kc); log 3165 drivers/md/raid5-cache.c kfree(log); log 3171 drivers/md/raid5-cache.c struct r5l_log *log = conf->log; log 3173 drivers/md/raid5-cache.c conf->log = NULL; log 3178 drivers/md/raid5-cache.c flush_work(&log->disable_writeback_work); log 3179 drivers/md/raid5-cache.c md_unregister_thread(&log->reclaim_thread); log 3180 drivers/md/raid5-cache.c mempool_exit(&log->meta_pool); log 3181 drivers/md/raid5-cache.c bioset_exit(&log->bs); log 3182 drivers/md/raid5-cache.c mempool_exit(&log->io_pool); log 3183 drivers/md/raid5-cache.c kmem_cache_destroy(log->io_kc); log 3184 drivers/md/raid5-cache.c kfree(log); log 7 drivers/md/raid5-log.h extern int r5l_write_stripe(struct r5l_log *log, struct stripe_head *head_sh); log 8 drivers/md/raid5-log.h extern void r5l_write_stripe_run(struct r5l_log *log); log 9 drivers/md/raid5-log.h extern void r5l_flush_stripe_to_raid(struct r5l_log *log); log 11 drivers/md/raid5-log.h extern int r5l_handle_flush_request(struct r5l_log *log, struct bio *bio); log 12 drivers/md/raid5-log.h extern void r5l_quiesce(struct r5l_log *log, int quiesce); log 14 drivers/md/raid5-log.h extern bool r5c_is_writeback(struct r5l_log *log); log 23 drivers/md/raid5-log.h extern void r5l_wake_reclaim(struct r5l_log *log, sector_t space); log 26 drivers/md/raid5-log.h extern int r5c_cache_data(struct r5l_log *log, struct stripe_head *sh); log 35 drivers/md/raid5-log.h extern int r5l_start(struct r5l_log *log); log 47 drivers/md/raid5-log.h extern int ppl_handle_flush_request(struct r5l_log *log, struct bio *bio); log 64 drivers/md/raid5-log.h if (conf->log) { log 69 drivers/md/raid5-log.h return r5l_write_stripe(conf->log, sh); log 72 drivers/md/raid5-log.h return r5c_cache_data(conf->log, sh); log 85 drivers/md/raid5-log.h if (conf->log) log 93 drivers/md/raid5-log.h if (conf->log) log 94 drivers/md/raid5-log.h r5l_write_stripe_run(conf->log); log 101 drivers/md/raid5-log.h if (conf->log) log 102 drivers/md/raid5-log.h r5l_flush_stripe_to_raid(conf->log); log 111 drivers/md/raid5-log.h if (conf->log) log 112 drivers/md/raid5-log.h ret = r5l_handle_flush_request(conf->log, bio); log 114 drivers/md/raid5-log.h ret = ppl_handle_flush_request(conf->log, bio); log 121 drivers/md/raid5-log.h if (conf->log) log 122 drivers/md/raid5-log.h r5l_quiesce(conf->log, quiesce); log 129 drivers/md/raid5-log.h if (conf->log) log 136 drivers/md/raid5-ppl.c struct ppl_log *log; log 232 drivers/md/raid5-ppl.c static struct ppl_io_unit *ppl_new_iounit(struct ppl_log *log, log 235 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 248 drivers/md/raid5-ppl.c io->log = log; log 266 drivers/md/raid5-ppl.c static int ppl_log_stripe(struct ppl_log *log, struct stripe_head *sh) log 268 drivers/md/raid5-ppl.c struct ppl_io_unit *io = log->current_io; log 279 drivers/md/raid5-ppl.c if (io && (io->pp_size == log->entry_space || log 288 drivers/md/raid5-ppl.c io = ppl_new_iounit(log, sh); log 291 drivers/md/raid5-ppl.c spin_lock_irq(&log->io_list_lock); log 292 drivers/md/raid5-ppl.c list_add_tail(&io->log_sibling, &log->io_list); log 293 drivers/md/raid5-ppl.c spin_unlock_irq(&log->io_list_lock); log 295 drivers/md/raid5-ppl.c log->current_io = io; log 364 drivers/md/raid5-ppl.c struct ppl_log *log; log 373 drivers/md/raid5-ppl.c log = &ppl_conf->child_logs[sh->pd_idx]; log 375 drivers/md/raid5-ppl.c mutex_lock(&log->io_mutex); log 377 drivers/md/raid5-ppl.c if (!log->rdev || test_bit(Faulty, &log->rdev->flags)) { log 378 drivers/md/raid5-ppl.c mutex_unlock(&log->io_mutex); log 386 drivers/md/raid5-ppl.c if (ppl_log_stripe(log, sh)) { log 392 drivers/md/raid5-ppl.c mutex_unlock(&log->io_mutex); log 400 drivers/md/raid5-ppl.c struct ppl_log *log = io->log; log 401 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 407 drivers/md/raid5-ppl.c md_error(ppl_conf->mddev, log->rdev); log 431 drivers/md/raid5-ppl.c struct ppl_log *log = io->log; log 432 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 440 drivers/md/raid5-ppl.c if (!log->rdev || test_bit(Faulty, &log->rdev->flags)) { log 461 drivers/md/raid5-ppl.c if (log->use_multippl && log 462 drivers/md/raid5-ppl.c log->rdev->ppl.sector + log->rdev->ppl.size - log->next_io_sector < log 464 drivers/md/raid5-ppl.c log->next_io_sector = log->rdev->ppl.sector; log 469 drivers/md/raid5-ppl.c bio_set_dev(bio, log->rdev->bdev); log 470 drivers/md/raid5-ppl.c bio->bi_iter.bi_sector = log->next_io_sector; log 475 drivers/md/raid5-ppl.c (unsigned long long)log->next_io_sector); log 477 drivers/md/raid5-ppl.c if (log->use_multippl) log 478 drivers/md/raid5-ppl.c log->next_io_sector += (PPL_HEADER_SIZE + io->pp_size) >> 9; log 480 drivers/md/raid5-ppl.c WARN_ON(log->disk_flush_bitmap != 0); log 488 drivers/md/raid5-ppl.c set_bit(i, &log->disk_flush_bitmap); log 515 drivers/md/raid5-ppl.c static void ppl_submit_current_io(struct ppl_log *log) log 519 drivers/md/raid5-ppl.c spin_lock_irq(&log->io_list_lock); log 521 drivers/md/raid5-ppl.c io = list_first_entry_or_null(&log->io_list, struct ppl_io_unit, log 526 drivers/md/raid5-ppl.c spin_unlock_irq(&log->io_list_lock); log 531 drivers/md/raid5-ppl.c if (io == log->current_io) log 532 drivers/md/raid5-ppl.c log->current_io = NULL; log 541 drivers/md/raid5-ppl.c struct ppl_log *log; log 545 drivers/md/raid5-ppl.c log = &ppl_conf->child_logs[i]; log 547 drivers/md/raid5-ppl.c mutex_lock(&log->io_mutex); log 548 drivers/md/raid5-ppl.c ppl_submit_current_io(log); log 549 drivers/md/raid5-ppl.c mutex_unlock(&log->io_mutex); log 555 drivers/md/raid5-ppl.c struct ppl_log *log = io->log; log 556 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 564 drivers/md/raid5-ppl.c spin_lock(&log->io_list_lock); log 566 drivers/md/raid5-ppl.c spin_unlock(&log->io_list_lock); log 590 drivers/md/raid5-ppl.c struct ppl_log *log = io->log; log 591 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 617 drivers/md/raid5-ppl.c struct ppl_log *log = io->log; log 618 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 626 drivers/md/raid5-ppl.c for_each_set_bit(i, &log->disk_flush_bitmap, raid_disks) { log 654 drivers/md/raid5-ppl.c log->disk_flush_bitmap = 0; log 663 drivers/md/raid5-ppl.c struct ppl_log *log) log 667 drivers/md/raid5-ppl.c io = list_first_entry_or_null(&log->io_list, struct ppl_io_unit, log 680 drivers/md/raid5-ppl.c struct ppl_log *log = &ppl_conf->child_logs[i]; log 682 drivers/md/raid5-ppl.c spin_lock_irq(&log->io_list_lock); log 684 drivers/md/raid5-ppl.c ppl_no_io_unit_submitted(conf, log), log 685 drivers/md/raid5-ppl.c log->io_list_lock); log 686 drivers/md/raid5-ppl.c spin_unlock_irq(&log->io_list_lock); log 691 drivers/md/raid5-ppl.c int ppl_handle_flush_request(struct r5l_log *log, struct bio *bio) log 709 drivers/md/raid5-ppl.c if (io->log->disk_flush_bitmap) log 795 drivers/md/raid5-ppl.c static int ppl_recover_entry(struct ppl_log *log, struct ppl_header_entry *e, log 798 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 928 drivers/md/raid5-ppl.c if (!sync_page_io(log->rdev, log 929 drivers/md/raid5-ppl.c ppl_sector - log->rdev->data_offset + i, log 934 drivers/md/raid5-ppl.c md_error(mddev, log->rdev); log 948 drivers/md/raid5-ppl.c BUG_ON(parity_rdev->bdev->bd_dev != log->rdev->bdev->bd_dev); log 970 drivers/md/raid5-ppl.c static int ppl_recover(struct ppl_log *log, struct ppl_header *pplhdr, log 973 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 974 drivers/md/raid5-ppl.c struct md_rdev *rdev = log->rdev; log 1030 drivers/md/raid5-ppl.c ret = ppl_recover_entry(log, e, ppl_sector); log 1046 drivers/md/raid5-ppl.c static int ppl_write_empty_header(struct ppl_log *log) log 1050 drivers/md/raid5-ppl.c struct md_rdev *rdev = log->rdev; log 1063 drivers/md/raid5-ppl.c log->rdev->ppl.size, GFP_NOIO, 0); log 1065 drivers/md/raid5-ppl.c pplhdr->signature = cpu_to_le32(log->ppl_conf->signature); log 1079 drivers/md/raid5-ppl.c static int ppl_load_distributed(struct ppl_log *log) log 1081 drivers/md/raid5-ppl.c struct ppl_conf *ppl_conf = log->ppl_conf; log 1082 drivers/md/raid5-ppl.c struct md_rdev *rdev = log->rdev; log 1180 drivers/md/raid5-ppl.c ret = ppl_recover(log, pplhdr, pplhdr_offset); log 1184 drivers/md/raid5-ppl.c ret = ppl_write_empty_header(log); log 1203 drivers/md/raid5-ppl.c struct ppl_log *log = &ppl_conf->child_logs[i]; log 1206 drivers/md/raid5-ppl.c if (!log->rdev) log 1209 drivers/md/raid5-ppl.c ret = ppl_load_distributed(log); log 1309 drivers/md/raid5-ppl.c static void ppl_init_child_log(struct ppl_log *log, struct md_rdev *rdev) log 1315 drivers/md/raid5-ppl.c log->use_multippl = true; log 1317 drivers/md/raid5-ppl.c &log->ppl_conf->mddev->flags); log 1318 drivers/md/raid5-ppl.c log->entry_space = PPL_SPACE_SIZE; log 1320 drivers/md/raid5-ppl.c log->use_multippl = false; log 1321 drivers/md/raid5-ppl.c log->entry_space = (log->rdev->ppl.size << 9) - log 1324 drivers/md/raid5-ppl.c log->next_io_sector = rdev->ppl.sector; log 1328 drivers/md/raid5-ppl.c log->wb_cache_on = true; log 1417 drivers/md/raid5-ppl.c struct ppl_log *log = &ppl_conf->child_logs[i]; log 1420 drivers/md/raid5-ppl.c mutex_init(&log->io_mutex); log 1421 drivers/md/raid5-ppl.c spin_lock_init(&log->io_list_lock); log 1422 drivers/md/raid5-ppl.c INIT_LIST_HEAD(&log->io_list); log 1424 drivers/md/raid5-ppl.c log->ppl_conf = ppl_conf; log 1425 drivers/md/raid5-ppl.c log->rdev = rdev; log 1432 drivers/md/raid5-ppl.c ppl_init_child_log(log, rdev); log 1468 drivers/md/raid5-ppl.c struct ppl_log *log; log 1485 drivers/md/raid5-ppl.c log = &ppl_conf->child_logs[rdev->raid_disk]; log 1487 drivers/md/raid5-ppl.c mutex_lock(&log->io_mutex); log 1491 drivers/md/raid5-ppl.c log->rdev = rdev; log 1492 drivers/md/raid5-ppl.c ret = ppl_write_empty_header(log); log 1493 drivers/md/raid5-ppl.c ppl_init_child_log(log, rdev); log 1496 drivers/md/raid5-ppl.c log->rdev = NULL; log 1498 drivers/md/raid5-ppl.c mutex_unlock(&log->io_mutex); log 221 drivers/md/raid5.c if (r5c_is_writeback(conf->log)) log 233 drivers/md/raid5.c (conf->quiesce && r5c_is_writeback(conf->log) && log 271 drivers/md/raid5.c if (!r5c_is_writeback(conf->log)) log 657 drivers/md/raid5.c r5l_wake_reclaim(conf->log, 0); log 1659 drivers/md/raid5.c if (r5c_is_writeback(sh->raid_conf->log)) log 1765 drivers/md/raid5.c r5c_is_writeback(conf->log)); log 1767 drivers/md/raid5.c !r5c_is_writeback(conf->log)) { log 4861 drivers/md/raid5.c if (!r5c_is_writeback(conf->log)) { log 5330 drivers/md/raid5.c bool second_try = !r5c_is_writeback(conf->log) && log 6209 drivers/md/raid5.c r5l_flush_stripe_to_raid(conf->log); log 6265 drivers/md/raid5.c r5l_flush_stripe_to_raid(conf->log); log 6352 drivers/md/raid5.c r5l_flush_stripe_to_raid(conf->log); log 7604 drivers/md/raid5.c if (test_bit(Journal, &rdev->flags) && conf->log) { log 7690 drivers/md/raid5.c if (conf->log) log 7702 drivers/md/raid5.c ret = r5l_start(conf->log); log 8406 drivers/md/raid5.c return r5l_start(conf->log); log 681 drivers/md/raid5.h struct r5l_log *log; log 125 drivers/media/dvb-core/dvb_math.c u64 log; log 132 drivers/media/dvb-core/dvb_math.c log = intlog2(value); log 139 drivers/media/dvb-core/dvb_math.c return (log * 646456993) >> 31; log 1831 drivers/media/v4l2-core/v4l2-ctrls.c .log = std_log, log 2833 drivers/media/v4l2-core/v4l2-ctrls.c ctrl->type_ops->log(ctrl); log 87 drivers/mtd/nand/raw/nandsim.c static uint log = CONFIG_NANDSIM_LOG; log 120 drivers/mtd/nand/raw/nandsim.c module_param(log, uint, 0400); log 145 drivers/mtd/nand/raw/nandsim.c MODULE_PARM_DESC(log, "Perform logging if not zero"); log 173 drivers/mtd/nand/raw/nandsim.c do { if (log) pr_debug(" log: " args); } while(0) log 90 drivers/mtd/tests/nandbiterrs.c static int write_page(int log) log 92 drivers/mtd/tests/nandbiterrs.c if (log) log 99 drivers/mtd/tests/nandbiterrs.c static int rewrite_page(int log) log 104 drivers/mtd/tests/nandbiterrs.c if (log) log 128 drivers/mtd/tests/nandbiterrs.c static int read_page(int log) log 134 drivers/mtd/tests/nandbiterrs.c if (log) log 154 drivers/mtd/tests/nandbiterrs.c static int verify_page(int log) log 158 drivers/mtd/tests/nandbiterrs.c if (log) log 2834 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c struct mbox_cmd_log *log = NULL; log 2842 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c log = padap->mbox_log; log 2851 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c entry_idx = log->cursor + k; log 2852 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c if (entry_idx >= log->size) log 2853 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c entry_idx -= log->size; log 2855 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c entry = mbox_cmd_log_entry(log, entry_idx); log 451 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h static inline struct mbox_cmd *mbox_cmd_log_entry(struct mbox_cmd_log *log, log 454 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h return &((struct mbox_cmd *)&(log)[1])[entry_idx]; log 987 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c struct fw_devlog_e log[0]; /* Firmware Device Log */ log 1009 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c e = &dinfo->log[index]; log 1108 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c dparams->start, dparams->size, (__be32 *)dinfo->log, log 1120 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c struct fw_devlog_e *e = &dinfo->log[index]; log 1156 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c struct mbox_cmd_log *log = adapter->mbox_log; log 1168 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c entry_idx = log->cursor + ((uintptr_t)v - 2); log 1169 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (entry_idx >= log->size) log 1170 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c entry_idx -= log->size; log 1171 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c entry = mbox_cmd_log_entry(log, entry_idx); log 1194 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c struct mbox_cmd_log *log = adapter->mbox_log; log 1196 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c return ((pos <= log->size) ? (void *)(uintptr_t)(pos + 1) : NULL); log 241 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c struct mbox_cmd_log *log = adapter->mbox_log; log 245 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry = mbox_cmd_log_entry(log, log->cursor++); log 246 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (log->cursor == log->size) log 247 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c log->cursor = 0; log 254 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry->seqno = log->seqno++; log 1963 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct mbox_cmd_log *log = adapter->mbox_log; log 1975 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c entry_idx = log->cursor + ((uintptr_t)v - 2); log 1976 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (entry_idx >= log->size) log 1977 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c entry_idx -= log->size; log 1978 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c entry = mbox_cmd_log_entry(log, entry_idx); log 2001 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct mbox_cmd_log *log = adapter->mbox_log; log 2003 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c return ((pos <= log->size) ? (void *)(uintptr_t)(pos + 1) : NULL); log 294 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h static inline struct mbox_cmd *mbox_cmd_log_entry(struct mbox_cmd_log *log, log 297 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h return &((struct mbox_cmd *)&(log)[1])[entry_idx]; log 90 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c struct mbox_cmd_log *log = adapter->mbox_log; log 94 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry = mbox_cmd_log_entry(log, log->cursor++); log 95 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c if (log->cursor == log->size) log 96 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c log->cursor = 0; log 103 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry->seqno = log->seqno++; log 2468 drivers/net/ethernet/pensando/ionic/ionic_if.h struct ionic_log_event log; log 492 drivers/net/wimax/i2400m/rx.c struct i2400m_roq_log *log; log 598 drivers/net/wimax/i2400m/rx.c if (roq->log->in - roq->log->out == I2400M_ROQ_LOG_LENGTH) log 599 drivers/net/wimax/i2400m/rx.c roq->log->out++; log 600 drivers/net/wimax/i2400m/rx.c cnt_idx = roq->log->in++ % I2400M_ROQ_LOG_LENGTH; log 601 drivers/net/wimax/i2400m/rx.c e = &roq->log->entry[cnt_idx]; log 623 drivers/net/wimax/i2400m/rx.c BUG_ON(roq->log->out > roq->log->in); log 624 drivers/net/wimax/i2400m/rx.c for (cnt = roq->log->out; cnt < roq->log->in; cnt++) { log 626 drivers/net/wimax/i2400m/rx.c e = &roq->log->entry[cnt_idx]; log 630 drivers/net/wimax/i2400m/rx.c roq->log->in = roq->log->out = 0; log 934 drivers/net/wimax/i2400m/rx.c kfree(i2400m->rx_roq[0].log); log 1361 drivers/net/wimax/i2400m/rx.c rd = kcalloc(I2400M_RO_CIN + 1, sizeof(*i2400m->rx_roq[0].log), log 1370 drivers/net/wimax/i2400m/rx.c i2400m->rx_roq[itr].log = &rd[itr]; log 352 drivers/net/wireless/broadcom/b43/debugfs.c struct b43_txstatus_log *log = &dev->dfsentry->txstatlog; log 357 drivers/net/wireless/broadcom/b43/debugfs.c if (log->end < 0) { log 365 drivers/net/wireless/broadcom/b43/debugfs.c i = log->end + 1; log 370 drivers/net/wireless/broadcom/b43/debugfs.c stat = &(log->log[i]); log 385 drivers/net/wireless/broadcom/b43/debugfs.c if (i == log->end) log 682 drivers/net/wireless/broadcom/b43/debugfs.c struct b43_txstatus_log *log; log 692 drivers/net/wireless/broadcom/b43/debugfs.c log = &e->txstatlog; log 693 drivers/net/wireless/broadcom/b43/debugfs.c log->log = kcalloc(B43_NR_LOGGED_TXSTATUS, log 695 drivers/net/wireless/broadcom/b43/debugfs.c if (!log->log) { log 700 drivers/net/wireless/broadcom/b43/debugfs.c log->end = -1; log 764 drivers/net/wireless/broadcom/b43/debugfs.c kfree(e->txstatlog.log); log 772 drivers/net/wireless/broadcom/b43/debugfs.c struct b43_txstatus_log *log; log 778 drivers/net/wireless/broadcom/b43/debugfs.c log = &e->txstatlog; log 779 drivers/net/wireless/broadcom/b43/debugfs.c i = log->end + 1; log 782 drivers/net/wireless/broadcom/b43/debugfs.c log->end = i; log 783 drivers/net/wireless/broadcom/b43/debugfs.c cur = &(log->log[i]); log 30 drivers/net/wireless/broadcom/b43/debugfs.h struct b43_txstatus *log; log 127 drivers/net/wireless/broadcom/b43legacy/debugfs.c struct b43legacy_txstatus_log *log = &dev->dfsentry->txstatlog; log 133 drivers/net/wireless/broadcom/b43legacy/debugfs.c spin_lock_irqsave(&log->lock, flags); log 134 drivers/net/wireless/broadcom/b43legacy/debugfs.c if (log->end < 0) { log 142 drivers/net/wireless/broadcom/b43legacy/debugfs.c i = log->end + 1; log 147 drivers/net/wireless/broadcom/b43legacy/debugfs.c stat = &(log->log[i]); log 162 drivers/net/wireless/broadcom/b43legacy/debugfs.c if (i == log->end) log 167 drivers/net/wireless/broadcom/b43legacy/debugfs.c spin_unlock_irqrestore(&log->lock, flags); log 371 drivers/net/wireless/broadcom/b43legacy/debugfs.c struct b43legacy_txstatus_log *log; log 381 drivers/net/wireless/broadcom/b43legacy/debugfs.c log = &e->txstatlog; log 382 drivers/net/wireless/broadcom/b43legacy/debugfs.c log->log = kcalloc(B43legacy_NR_LOGGED_TXSTATUS, log 384 drivers/net/wireless/broadcom/b43legacy/debugfs.c if (!log->log) { log 389 drivers/net/wireless/broadcom/b43legacy/debugfs.c log->end = -1; log 390 drivers/net/wireless/broadcom/b43legacy/debugfs.c spin_lock_init(&log->lock); log 436 drivers/net/wireless/broadcom/b43legacy/debugfs.c kfree(e->txstatlog.log); log 444 drivers/net/wireless/broadcom/b43legacy/debugfs.c struct b43legacy_txstatus_log *log; log 450 drivers/net/wireless/broadcom/b43legacy/debugfs.c log = &e->txstatlog; log 452 drivers/net/wireless/broadcom/b43legacy/debugfs.c spin_lock(&log->lock); log 453 drivers/net/wireless/broadcom/b43legacy/debugfs.c i = log->end + 1; log 456 drivers/net/wireless/broadcom/b43legacy/debugfs.c log->end = i; log 457 drivers/net/wireless/broadcom/b43legacy/debugfs.c cur = &(log->log[i]); log 459 drivers/net/wireless/broadcom/b43legacy/debugfs.c spin_unlock(&log->lock); log 25 drivers/net/wireless/broadcom/b43legacy/debugfs.h struct b43legacy_txstatus *log; log 667 drivers/net/wireless/intel/ipw2x00/ipw2200.c error->log[i].time, log 668 drivers/net/wireless/intel/ipw2x00/ipw2200.c error->log[i].data, error->log[i].event); log 1219 drivers/net/wireless/intel/ipw2x00/ipw2200.c u32 log_len, struct ipw_event *log) log 1226 drivers/net/wireless/intel/ipw2x00/ipw2200.c (u8 *) log, sizeof(*log) * log_len); log 1239 drivers/net/wireless/intel/ipw2x00/ipw2200.c sizeof(*error->log) * log_len, GFP_ATOMIC); log 1251 drivers/net/wireless/intel/ipw2x00/ipw2200.c error->log = (struct ipw_event *)(error->elem + elem_len); log 1253 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_capture_event_log(priv, log_len, error->log); log 1268 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct ipw_event *log; log 1272 drivers/net/wireless/intel/ipw2x00/ipw2200.c log_size = PAGE_SIZE / sizeof(*log) > log_len ? log 1273 drivers/net/wireless/intel/ipw2x00/ipw2200.c sizeof(*log) * log_len : PAGE_SIZE; log 1274 drivers/net/wireless/intel/ipw2x00/ipw2200.c log = kzalloc(log_size, GFP_KERNEL); log 1275 drivers/net/wireless/intel/ipw2x00/ipw2200.c if (!log) { log 1279 drivers/net/wireless/intel/ipw2x00/ipw2200.c log_len = log_size / sizeof(*log); log 1280 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_capture_event_log(priv, log_len, log); log 1286 drivers/net/wireless/intel/ipw2x00/ipw2200.c log[i].time, log[i].event, log[i].data); log 1288 drivers/net/wireless/intel/ipw2x00/ipw2200.c kfree(log); log 1322 drivers/net/wireless/intel/ipw2x00/ipw2200.c priv->error->log[i].time, log 1323 drivers/net/wireless/intel/ipw2x00/ipw2200.c priv->error->log[i].event, log 1324 drivers/net/wireless/intel/ipw2x00/ipw2200.c priv->error->log[i].data); log 1110 drivers/net/wireless/intel/ipw2x00/ipw2200.h struct ipw_event *log; log 1520 drivers/net/wireless/marvell/mwifiex/main.h struct mwifiex_ds_get_stats *log); log 1200 drivers/net/wireless/marvell/mwifiex/sta_ioctl.c struct mwifiex_ds_get_stats *log) log 1203 drivers/net/wireless/marvell/mwifiex/sta_ioctl.c HostCmd_ACT_GEN_GET, 0, log, true); log 207 drivers/nvdimm/btt.c struct log_group *log) log 210 drivers/nvdimm/btt.c arena->logoff + (lane * LOG_GRP_SIZE), log, log 271 drivers/nvdimm/btt.c static u32 log_seq(struct log_group *log, int log_idx) log 273 drivers/nvdimm/btt.c return le32_to_cpu(log->ent[log_idx].seq); log 285 drivers/nvdimm/btt.c static int btt_log_get_old(struct arena_info *a, struct log_group *log) log 296 drivers/nvdimm/btt.c if (log_seq(log, idx0) == 0) { log 297 drivers/nvdimm/btt.c log->ent[idx0].seq = cpu_to_le32(1); log 301 drivers/nvdimm/btt.c if (log_seq(log, idx0) == log_seq(log, idx1)) log 303 drivers/nvdimm/btt.c if (log_seq(log, idx0) + log_seq(log, idx1) > 5) log 306 drivers/nvdimm/btt.c if (log_seq(log, idx0) < log_seq(log, idx1)) { log 307 drivers/nvdimm/btt.c if ((log_seq(log, idx1) - log_seq(log, idx0)) == 1) log 312 drivers/nvdimm/btt.c if ((log_seq(log, idx0) - log_seq(log, idx1)) == 1) log 332 drivers/nvdimm/btt.c struct log_group log; log 334 drivers/nvdimm/btt.c ret = btt_log_group_read(arena, lane, &log); log 338 drivers/nvdimm/btt.c old_ent = btt_log_get_old(arena, &log); log 342 drivers/nvdimm/btt.c old_ent, lane, log.ent[arena->log_index[0]].seq, log 343 drivers/nvdimm/btt.c log.ent[arena->log_index[1]].seq); log 351 drivers/nvdimm/btt.c memcpy(ent, &log.ent[arena->log_index[ret_ent]], LOG_ENT_SIZE); log 628 drivers/nvdimm/btt.c struct log_group log; log 632 drivers/nvdimm/btt.c ret = btt_log_group_read(arena, i, &log); log 638 drivers/nvdimm/btt.c if (ent_is_padding(&log.ent[j])) { log 666 drivers/nvdimm/btt.c if (ent_is_padding(&log.ent[j])) log 678 drivers/nvdimm/btt.c if (!ent_is_padding(&log.ent[j])) log 1317 drivers/nvdimm/btt.c struct log_entry log; log 1388 drivers/nvdimm/btt.c log.lba = cpu_to_le32(premap); log 1389 drivers/nvdimm/btt.c log.old_map = cpu_to_le32(old_postmap); log 1390 drivers/nvdimm/btt.c log.new_map = cpu_to_le32(new_postmap); log 1391 drivers/nvdimm/btt.c log.seq = cpu_to_le32(arena->freelist[lane].seq); log 1393 drivers/nvdimm/btt.c ret = btt_flog_write(arena, lane, sub, &log); log 2700 drivers/nvme/host/core.c void *log, size_t size, u64 offset) log 2714 drivers/nvme/host/core.c return nvme_submit_sync_cmd(ctrl->admin_q, &c, log, size); log 3700 drivers/nvme/host/core.c __le32 *log; log 3703 drivers/nvme/host/core.c log = kzalloc(log_size, GFP_KERNEL); log 3704 drivers/nvme/host/core.c if (!log) log 3713 drivers/nvme/host/core.c error = nvme_get_log(ctrl, NVME_NSID_ALL, NVME_LOG_CHANGED_NS, 0, log, log 3719 drivers/nvme/host/core.c kfree(log); log 3863 drivers/nvme/host/core.c struct nvme_fw_slot_info_log *log; log 3865 drivers/nvme/host/core.c log = kmalloc(sizeof(*log), GFP_KERNEL); log 3866 drivers/nvme/host/core.c if (!log) log 3869 drivers/nvme/host/core.c if (nvme_get_log(ctrl, NVME_NSID_ALL, NVME_LOG_FW_SLOT, 0, log, log 3870 drivers/nvme/host/core.c sizeof(*log), 0)) log 3872 drivers/nvme/host/core.c kfree(log); log 508 drivers/nvme/host/nvme.h void *log, size_t size, u64 offset); log 133 drivers/nvme/target/admin-cmd.c struct nvme_smart_log *log; log 137 drivers/nvme/target/admin-cmd.c if (req->data_len != sizeof(*log)) log 140 drivers/nvme/target/admin-cmd.c log = kzalloc(sizeof(*log), GFP_KERNEL); log 141 drivers/nvme/target/admin-cmd.c if (!log) log 145 drivers/nvme/target/admin-cmd.c status = nvmet_get_smart_log_all(req, log); log 147 drivers/nvme/target/admin-cmd.c status = nvmet_get_smart_log_nsid(req, log); log 153 drivers/nvme/target/admin-cmd.c &log->num_err_log_entries); log 156 drivers/nvme/target/admin-cmd.c status = nvmet_copy_to_sgl(req, 0, log, sizeof(*log)); log 158 drivers/nvme/target/admin-cmd.c kfree(log); log 166 drivers/nvme/target/admin-cmd.c struct nvme_effects_log *log; log 168 drivers/nvme/target/admin-cmd.c log = kzalloc(sizeof(*log), GFP_KERNEL); log 169 drivers/nvme/target/admin-cmd.c if (!log) log 172 drivers/nvme/target/admin-cmd.c log->acs[nvme_admin_get_log_page] = cpu_to_le32(1 << 0); log 173 drivers/nvme/target/admin-cmd.c log->acs[nvme_admin_identify] = cpu_to_le32(1 << 0); log 174 drivers/nvme/target/admin-cmd.c log->acs[nvme_admin_abort_cmd] = cpu_to_le32(1 << 0); log 175 drivers/nvme/target/admin-cmd.c log->acs[nvme_admin_set_features] = cpu_to_le32(1 << 0); log 176 drivers/nvme/target/admin-cmd.c log->acs[nvme_admin_get_features] = cpu_to_le32(1 << 0); log 177 drivers/nvme/target/admin-cmd.c log->acs[nvme_admin_async_event] = cpu_to_le32(1 << 0); log 178 drivers/nvme/target/admin-cmd.c log->acs[nvme_admin_keep_alive] = cpu_to_le32(1 << 0); log 180 drivers/nvme/target/admin-cmd.c log->iocs[nvme_cmd_read] = cpu_to_le32(1 << 0); log 181 drivers/nvme/target/admin-cmd.c log->iocs[nvme_cmd_write] = cpu_to_le32(1 << 0); log 182 drivers/nvme/target/admin-cmd.c log->iocs[nvme_cmd_flush] = cpu_to_le32(1 << 0); log 183 drivers/nvme/target/admin-cmd.c log->iocs[nvme_cmd_dsm] = cpu_to_le32(1 << 0); log 184 drivers/nvme/target/admin-cmd.c log->iocs[nvme_cmd_write_zeroes] = cpu_to_le32(1 << 0); log 186 drivers/nvme/target/admin-cmd.c status = nvmet_copy_to_sgl(req, 0, log, sizeof(*log)); log 188 drivers/nvme/target/admin-cmd.c kfree(log); log 154 drivers/pci/pcie/dpc.c u32 status, mask, sev, syserr, exc, dw0, dw1, dw2, dw3, log, prefix; log 193 drivers/pci/pcie/dpc.c pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_IMPSPEC_LOG, &log); log 194 drivers/pci/pcie/dpc.c pci_err(pdev, "RP PIO ImpSpec Log %#010x\n", log); log 917 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c struct cygnus_mux_log *log; log 928 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c log = &pinctrl->mux_log[i * CYGNUS_NUM_MUX_PER_REG log 930 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c log->mux.offset = i * 4; log 931 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c log->mux.shift = j * 4; log 932 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c log->mux.alt = 0; log 933 drivers/pinctrl/bcm/pinctrl-cygnus-mux.c log->is_configured = false; log 984 drivers/pinctrl/bcm/pinctrl-ns2-mux.c struct ns2_mux_log *log; log 996 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log = &pinctrl->mux_log[0]; log 997 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.base = NS2_PIN_MUX_BASE0; log 998 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.offset = 0; log 999 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.shift = 31; log 1000 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.alt = 0; log 1008 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log = &pinctrl->mux_log[i]; log 1009 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.base = NS2_PIN_MUX_BASE0; log 1010 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.offset = NS2_MUX_PAD_FUNC1_OFFSET; log 1011 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.shift = 32 - (i * 2); log 1012 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.alt = 0; log 1020 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log = &pinctrl->mux_log[(NS2_NUM_IOMUX - NS2_NUM_PWM_MUX) + i]; log 1021 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.base = NS2_PIN_MUX_BASE1; log 1022 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.offset = 0; log 1023 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.shift = i; log 1024 drivers/pinctrl/bcm/pinctrl-ns2-mux.c log->mux.alt = 0; log 538 drivers/pinctrl/bcm/pinctrl-nsp-mux.c struct nsp_mux_log *log; log 549 drivers/pinctrl/bcm/pinctrl-nsp-mux.c log = &pinctrl->mux_log[i]; log 550 drivers/pinctrl/bcm/pinctrl-nsp-mux.c log->mux.base = nsp_pin_groups[i].mux.base; log 551 drivers/pinctrl/bcm/pinctrl-nsp-mux.c log->mux.shift = nsp_pin_groups[i].mux.shift; log 552 drivers/pinctrl/bcm/pinctrl-nsp-mux.c log->mux.alt = 0; log 553 drivers/pinctrl/bcm/pinctrl-nsp-mux.c log->is_configured = false; log 353 drivers/regulator/max77802-regulator.c #define regulator_77802_desc_p_ldo(num, supply, log) { \ log 359 drivers/regulator/max77802-regulator.c .ops = &max77802_ldo_ops_logic##log, \ log 374 drivers/regulator/max77802-regulator.c #define regulator_77802_desc_n_ldo(num, supply, log) { \ log 380 drivers/regulator/max77802-regulator.c .ops = &max77802_ldo_ops_logic##log, \ log 132 drivers/s390/scsi/zfcp_dbf.c goto log; log 148 drivers/s390/scsi/zfcp_dbf.c log: log 424 drivers/s390/scsi/zfcp_fsf.h u8 log[FSF_QTCB_LOG_SIZE]; log 931 drivers/scsi/esas2r/atvda.h struct atto_vda_ae_log log; log 86 drivers/scsi/ibmvscsi/ibmvfc.c int log; log 1474 drivers/scsi/ibmvscsi/ibmvfc.c logerr = cmd_status[index].log; log 140 drivers/scsi/mesh.c struct dbglog log[N_DBG_LOG]; log 179 drivers/scsi/mesh.c struct dbglog log[N_DBG_SLOG]; log 218 drivers/scsi/mesh.c tlp = &tp->log[tp->log_ix]; log 219 drivers/scsi/mesh.c slp = &ms->log[ms->log_ix]; log 251 drivers/scsi/mesh.c lp = &tp->log[i]; log 276 drivers/scsi/mesh.c lp = &ms->log[i]; log 184 drivers/staging/comedi/drivers/usbduxfast.c u8 len, u8 op, u8 out, u8 log) log 192 drivers/staging/comedi/drivers/usbduxfast.c devpriv->duxbuf[1 + 0x18 + index] = log; log 70 drivers/staging/greybus/log.c struct gb_log *log; log 80 drivers/staging/greybus/log.c log = kzalloc(sizeof(*log), GFP_KERNEL); log 81 drivers/staging/greybus/log.c if (!log) log 91 drivers/staging/greybus/log.c log->connection = connection; log 92 drivers/staging/greybus/log.c greybus_set_drvdata(bundle, log); log 103 drivers/staging/greybus/log.c kfree(log); log 109 drivers/staging/greybus/log.c struct gb_log *log = greybus_get_drvdata(bundle); log 110 drivers/staging/greybus/log.c struct gb_connection *connection = log->connection; log 115 drivers/staging/greybus/log.c kfree(log); log 43 drivers/staging/isdn/hysdn/hysdn_proclog.c struct proc_dir_entry *log; /* log entry */ log 330 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log = proc_create_data(pd->log_name, log 352 drivers/staging/isdn/hysdn/hysdn_proclog.c if (pd->log) log 479 drivers/usb/atm/cxacru.c char log[CMD_MAX_CONFIG * 12 + 1]; /* %02x=%08x */ log 491 drivers/usb/atm/cxacru.c snprintf(log + tmp*12, 13, " %02x=%08x", log 494 drivers/usb/atm/cxacru.c atm_info(instance->usbatm, "config%s\n", log); log 504 drivers/usb/host/isp116x.h isp116x_show_regs(isp116x, log, NULL); log 1248 drivers/usb/storage/ene_ub6250.c u16 log; log 1254 drivers/usb/storage/ene_ub6250.c log = info->MS_Lib.Phy2LogMap[phyblk]; log 1256 drivers/usb/storage/ene_ub6250.c if (log < info->MS_Lib.NumberOfLogBlock) log 1257 drivers/usb/storage/ene_ub6250.c info->MS_Lib.Log2PhyMap[log] = MS_LB_NOT_USED; log 1314 drivers/usb/storage/ene_ub6250.c u16 log; log 1320 drivers/usb/storage/ene_ub6250.c log = info->MS_Lib.Phy2LogMap[phyblk]; log 1322 drivers/usb/storage/ene_ub6250.c if (log < info->MS_Lib.NumberOfLogBlock) log 1323 drivers/usb/storage/ene_ub6250.c info->MS_Lib.Log2PhyMap[log] = MS_LB_NOT_USED; log 1030 drivers/vhost/net.c struct vhost_log *log, log 1051 drivers/vhost/net.c &in, log, log_num); log 1066 drivers/vhost/net.c if (unlikely(log)) { log 1068 drivers/vhost/net.c log += *log_num; log 1079 drivers/vhost/net.c if (unlikely(log)) log 1099 drivers/vhost/net.c unsigned uninitialized_var(in), log; log 1138 drivers/vhost/net.c vq->log : NULL; log 1149 drivers/vhost/net.c vhost_len, &in, vq_log, &log, log 1227 drivers/vhost/net.c vhost_log_write(vq, vq_log, log, vhost_len, log 374 drivers/vhost/vhost.c kfree(vq->log); log 375 drivers/vhost/vhost.c vq->log = NULL; log 391 drivers/vhost/vhost.c vq->log = kmalloc_array(dev->iov_limit, sizeof(*vq->log), log 395 drivers/vhost/vhost.c if (!vq->indirect || !vq->log || !vq->heads) log 482 drivers/vhost/vhost.c vq->log = NULL; log 759 drivers/vhost/vhost.c bool log; log 762 drivers/vhost/vhost.c log = log_all || vhost_has_feature(d->vqs[i], VHOST_F_LOG_ALL); log 766 drivers/vhost/vhost.c umem, log); log 1826 drivers/vhost/vhost.c unsigned long log = (unsigned long)addr; log 1829 drivers/vhost/vhost.c int bit = nr + (log % PAGE_SIZE) * 8; log 1832 drivers/vhost/vhost.c r = get_user_pages_fast(log, 1, FOLL_WRITE, &page); log 1855 drivers/vhost/vhost.c u64 log = base + write_page / 8; log 1857 drivers/vhost/vhost.c if ((u64)(unsigned long)log != log) log 1859 drivers/vhost/vhost.c r = set_bit_to_user(bit, (void __user *)(unsigned long)log); log 1933 drivers/vhost/vhost.c int vhost_log_write(struct vhost_virtqueue *vq, struct vhost_log *log, log 1952 drivers/vhost/vhost.c u64 l = min(log[i].len, len); log 1953 drivers/vhost/vhost.c r = log_write(vq->log_base, log[i].addr, l); log 2105 drivers/vhost/vhost.c struct vhost_log *log, unsigned int *log_num, log 2181 drivers/vhost/vhost.c if (unlikely(log && ret)) { log 2182 drivers/vhost/vhost.c log[*log_num].addr = vhost64_to_cpu(vq, desc.addr); log 2183 drivers/vhost/vhost.c log[*log_num].len = vhost32_to_cpu(vq, desc.len); log 2211 drivers/vhost/vhost.c struct vhost_log *log, unsigned int *log_num) log 2269 drivers/vhost/vhost.c if (unlikely(log)) log 2295 drivers/vhost/vhost.c log, log_num, &desc); log 2322 drivers/vhost/vhost.c if (unlikely(log && ret)) { log 2323 drivers/vhost/vhost.c log[*log_num].addr = vhost64_to_cpu(vq, desc.addr); log 2324 drivers/vhost/vhost.c log[*log_num].len = vhost32_to_cpu(vq, desc.len); log 138 drivers/vhost/vhost.h struct vhost_log *log; log 196 drivers/vhost/vhost.h struct vhost_log *log, unsigned int *log_num); log 212 drivers/vhost/vhost.h int vhost_log_write(struct vhost_virtqueue *vq, struct vhost_log *log, log 1404 drivers/virt/vboxguest/vboxguest_core.c static int vbg_ioctl_log(struct vbg_ioctl_log *log) log 1406 drivers/virt/vboxguest/vboxguest_core.c if (log->hdr.size_out != sizeof(log->hdr)) log 1409 drivers/virt/vboxguest/vboxguest_core.c vbg_info("%.*s", (int)(log->hdr.size_in - sizeof(log->hdr)), log 1410 drivers/virt/vboxguest/vboxguest_core.c log->u.in.msg); log 105 fs/btrfs/tree-log.c struct btrfs_root *log, log 289 fs/btrfs/tree-log.c int (*process_func)(struct btrfs_root *log, struct extent_buffer *eb, log 296 fs/btrfs/tree-log.c static int process_one_buffer(struct btrfs_root *log, log 300 fs/btrfs/tree-log.c struct btrfs_fs_info *fs_info = log->fs_info; log 943 fs/btrfs/tree-log.c static noinline int backref_in_log(struct btrfs_root *log, log 962 fs/btrfs/tree-log.c ret = btrfs_search_slot(NULL, log, key, path, 0, 0); log 1418 fs/btrfs/tree-log.c struct btrfs_root *log, log 1505 fs/btrfs/tree-log.c ret = __add_inode_ref(trans, root, path, log, log 2217 fs/btrfs/tree-log.c struct btrfs_root *log, log 2253 fs/btrfs/tree-log.c if (log && dir_key->type == BTRFS_DIR_ITEM_KEY) { log 2254 fs/btrfs/tree-log.c log_di = btrfs_lookup_dir_item(trans, log, log_path, log 2257 fs/btrfs/tree-log.c } else if (log && dir_key->type == BTRFS_DIR_INDEX_KEY) { log 2258 fs/btrfs/tree-log.c log_di = btrfs_lookup_dir_index_item(trans, log, log 2320 fs/btrfs/tree-log.c struct btrfs_root *log, log 2373 fs/btrfs/tree-log.c log_di = btrfs_lookup_xattr(NULL, log, log_path, ino, log 2428 fs/btrfs/tree-log.c struct btrfs_root *log, log 2463 fs/btrfs/tree-log.c ret = find_dir_range(log, path, dirid, key_type, log 2494 fs/btrfs/tree-log.c ret = check_item_in_log(trans, root, log, path, log 2535 fs/btrfs/tree-log.c static int replay_one_buffer(struct btrfs_root *log, struct extent_buffer *eb, log 2584 fs/btrfs/tree-log.c ret = replay_xattr_deletes(wc->trans, root, log, log 2591 fs/btrfs/tree-log.c root, log, path, key.objectid, 0); log 2659 fs/btrfs/tree-log.c ret = add_inode_ref(wc->trans, root, log, path, log 2857 fs/btrfs/tree-log.c struct btrfs_root *log, struct walk_control *wc) log 2859 fs/btrfs/tree-log.c struct btrfs_fs_info *fs_info = log->fs_info; log 2870 fs/btrfs/tree-log.c level = btrfs_header_level(log->node); log 2872 fs/btrfs/tree-log.c path->nodes[level] = log->node; log 2873 fs/btrfs/tree-log.c extent_buffer_get(log->node); log 2877 fs/btrfs/tree-log.c wret = walk_down_log_tree(trans, log, path, &level, wc); log 2885 fs/btrfs/tree-log.c wret = walk_up_log_tree(trans, log, path, &level, wc); log 2896 fs/btrfs/tree-log.c ret = wc->process_func(log, path->nodes[orig_level], wc, log 2917 fs/btrfs/tree-log.c WARN_ON(log->root_key.objectid != log 2936 fs/btrfs/tree-log.c struct btrfs_root *log, log 2939 fs/btrfs/tree-log.c struct btrfs_fs_info *fs_info = log->fs_info; log 2942 fs/btrfs/tree-log.c if (log->log_transid == 1) { log 2945 fs/btrfs/tree-log.c &log->root_key, root_item); log 2948 fs/btrfs/tree-log.c &log->root_key, root_item); log 3044 fs/btrfs/tree-log.c struct btrfs_root *log = root->log_root; log 3101 fs/btrfs/tree-log.c ret = btrfs_write_marked_extents(fs_info, &log->dirty_log_pages, mark); log 3123 fs/btrfs/tree-log.c btrfs_set_root_node(&log->root_item, log->node); log 3124 fs/btrfs/tree-log.c memcpy(&new_root_item, &log->root_item, sizeof(new_root_item)); log 3127 fs/btrfs/tree-log.c log->log_transid = root->log_transid; log 3155 fs/btrfs/tree-log.c ret = update_log_root(trans, log, &new_root_item); log 3174 fs/btrfs/tree-log.c btrfs_wait_tree_log_extents(log, mark); log 3191 fs/btrfs/tree-log.c ret = btrfs_wait_tree_log_extents(log, mark); log 3215 fs/btrfs/tree-log.c btrfs_wait_tree_log_extents(log, mark); log 3231 fs/btrfs/tree-log.c ret = btrfs_wait_tree_log_extents(log, mark); log 3299 fs/btrfs/tree-log.c struct btrfs_root *log) log 3307 fs/btrfs/tree-log.c ret = walk_log_tree(trans, log, &wc); log 3312 fs/btrfs/tree-log.c btrfs_handle_fs_error(log->fs_info, ret, NULL); log 3315 fs/btrfs/tree-log.c clear_extent_bits(&log->dirty_log_pages, 0, (u64)-1, log 3317 fs/btrfs/tree-log.c free_extent_buffer(log->node); log 3318 fs/btrfs/tree-log.c kfree(log); log 3394 fs/btrfs/tree-log.c struct btrfs_root *log; log 3411 fs/btrfs/tree-log.c log = root->log_root; log 3418 fs/btrfs/tree-log.c di = btrfs_lookup_dir_item(trans, log, path, dir_ino, log 3425 fs/btrfs/tree-log.c ret = btrfs_delete_one_dir_name(trans, log, path, di); log 3433 fs/btrfs/tree-log.c di = btrfs_lookup_dir_index_item(trans, log, path, dir_ino, log 3440 fs/btrfs/tree-log.c ret = btrfs_delete_one_dir_name(trans, log, path, di); log 3459 fs/btrfs/tree-log.c ret = btrfs_search_slot(trans, log, &key, path, 0, 1); log 3502 fs/btrfs/tree-log.c struct btrfs_root *log; log 3512 fs/btrfs/tree-log.c log = root->log_root; log 3515 fs/btrfs/tree-log.c ret = btrfs_del_inode_ref(trans, log, name, name_len, btrfs_ino(inode), log 3534 fs/btrfs/tree-log.c struct btrfs_root *log, log 3549 fs/btrfs/tree-log.c ret = btrfs_insert_empty_item(trans, log, path, &key, sizeof(*item)); log 3574 fs/btrfs/tree-log.c struct btrfs_root *log = root->log_root; log 3584 fs/btrfs/tree-log.c log = root->log_root; log 3630 fs/btrfs/tree-log.c ret = overwrite_item(trans, log, dst_path, log 3668 fs/btrfs/tree-log.c ret = overwrite_item(trans, log, dst_path, src, i, log 3726 fs/btrfs/tree-log.c ret = overwrite_item(trans, log, dst_path, log 3746 fs/btrfs/tree-log.c ret = insert_dir_log_key(trans, log, path, key_type, log 3804 fs/btrfs/tree-log.c struct btrfs_root *log, log 3818 fs/btrfs/tree-log.c ret = btrfs_search_slot(trans, log, &key, path, -1, 1); log 3840 fs/btrfs/tree-log.c ret = btrfs_del_items(trans, log, path, start_slot, log 3913 fs/btrfs/tree-log.c struct btrfs_root *log, struct btrfs_path *path, log 3919 fs/btrfs/tree-log.c ret = btrfs_insert_empty_item(trans, log, path, log 3963 fs/btrfs/tree-log.c struct btrfs_root *log = inode->root->log_root; log 3989 fs/btrfs/tree-log.c ret = btrfs_insert_empty_items(trans, log, dst_path, log 4075 fs/btrfs/tree-log.c ret = log_csums(trans, log, sums); log 4148 fs/btrfs/tree-log.c struct btrfs_root *log = root->log_root; log 4158 fs/btrfs/tree-log.c ret = log_extent_csums(trans, inode, log, em); log 4162 fs/btrfs/tree-log.c ret = __btrfs_drop_extents(trans, log, &inode->vfs_inode, path, em->start, log 4173 fs/btrfs/tree-log.c ret = btrfs_insert_empty_item(trans, log, path, &key, log 4467 fs/btrfs/tree-log.c static int logged_inode_size(struct btrfs_root *log, struct btrfs_inode *inode, log 4477 fs/btrfs/tree-log.c ret = btrfs_search_slot(NULL, log, &key, path, 0, 0); log 5036 fs/btrfs/tree-log.c struct btrfs_root *log = root->log_root; log 5114 fs/btrfs/tree-log.c ret = drop_objectid_items(trans, log, path, ino, max_key_type); log 5130 fs/btrfs/tree-log.c err = logged_inode_size(log, inode, path, &logged_isize); log 5138 fs/btrfs/tree-log.c ret = drop_objectid_items(trans, log, path, ino, log 5147 fs/btrfs/tree-log.c log, &inode->vfs_inode, 0, 0); log 5158 fs/btrfs/tree-log.c ret = drop_objectid_items(trans, log, path, ino, log 5323 fs/btrfs/tree-log.c err = log_inode_item(trans, log, dst_path, inode); log 5550 fs/btrfs/tree-log.c struct btrfs_root *log = root->log_root; log 5584 fs/btrfs/tree-log.c ret = btrfs_search_forward(log, &min_key, path, trans->transid); log 5653 fs/btrfs/tree-log.c ret = btrfs_next_leaf(log, path); log 6165 fs/btrfs/tree-log.c struct btrfs_root *log; log 6218 fs/btrfs/tree-log.c log = btrfs_read_fs_root(log_root_tree, &found_key); log 6219 fs/btrfs/tree-log.c if (IS_ERR(log)) { log 6220 fs/btrfs/tree-log.c ret = PTR_ERR(log); log 6247 fs/btrfs/tree-log.c log->node->start, log 6248 fs/btrfs/tree-log.c log->node->len); log 6249 fs/btrfs/tree-log.c free_extent_buffer(log->node); log 6250 fs/btrfs/tree-log.c free_extent_buffer(log->commit_root); log 6251 fs/btrfs/tree-log.c kfree(log); log 6260 fs/btrfs/tree-log.c wc.replay_dest->log_root = log; log 6262 fs/btrfs/tree-log.c ret = walk_log_tree(trans, log, &wc); log 6287 fs/btrfs/tree-log.c free_extent_buffer(log->node); log 6288 fs/btrfs/tree-log.c free_extent_buffer(log->commit_root); log 6289 fs/btrfs/tree-log.c kfree(log); log 364 fs/fs_context.c if (fc->log) log 365 fs/fs_context.c refcount_inc(&fc->log->usage); log 391 fs/fs_context.c struct fc_log *log = fc ? fc->log : NULL; log 418 fs/fs_context.c if (log && within_module_core((unsigned long)p, log->owner)) log 429 fs/fs_context.c if (!log) { log 444 fs/fs_context.c unsigned int logsize = ARRAY_SIZE(log->buffer); log 447 fs/fs_context.c index = log->head & (logsize - 1); log 448 fs/fs_context.c BUILD_BUG_ON(sizeof(log->head) != sizeof(u8) || log 449 fs/fs_context.c sizeof(log->tail) != sizeof(u8)); log 450 fs/fs_context.c if ((u8)(log->head - log->tail) == logsize) { log 452 fs/fs_context.c if (log->need_free & (1 << index)) log 453 fs/fs_context.c kfree(log->buffer[index]); log 454 fs/fs_context.c log->tail++; log 457 fs/fs_context.c log->buffer[index] = q; log 458 fs/fs_context.c log->need_free &= ~(1 << index); log 459 fs/fs_context.c log->need_free |= freeable << index; log 460 fs/fs_context.c log->head++; log 471 fs/fs_context.c struct fc_log *log = fc->log; log 474 fs/fs_context.c if (log) { log 475 fs/fs_context.c if (refcount_dec_and_test(&log->usage)) { log 476 fs/fs_context.c fc->log = NULL; log 478 fs/fs_context.c if (log->need_free & (1 << i)) log 479 fs/fs_context.c kfree(log->buffer[i]); log 480 fs/fs_context.c kfree(log); log 28 fs/fsopen.c struct fc_log *log = fc->log; log 29 fs/fsopen.c unsigned int logsize = ARRAY_SIZE(log->buffer); log 39 fs/fsopen.c if (log->head == log->tail) { log 44 fs/fsopen.c index = log->tail & (logsize - 1); log 45 fs/fsopen.c p = log->buffer[index]; log 46 fs/fsopen.c need_free = log->need_free & (1 << index); log 47 fs/fsopen.c log->buffer[index] = NULL; log 48 fs/fsopen.c log->need_free &= ~(1 << index); log 49 fs/fsopen.c log->tail++; log 100 fs/fsopen.c fc->log = kzalloc(sizeof(*fc->log), GFP_KERNEL); log 101 fs/fsopen.c if (!fc->log) log 103 fs/fsopen.c refcount_set(&fc->log->usage, 1); log 104 fs/fsopen.c fc->log->owner = fc->fs_type->owner; log 32 fs/jfs/file.c jfs_flush_journal(JFS_SBI(inode->i_sb)->log, 1); log 128 fs/jfs/inode.c jfs_flush_journal(JFS_SBI(inode->i_sb)->log, wait); log 421 fs/jfs/jfs_dmap.c struct jfs_log *log; log 436 fs/jfs/jfs_dmap.c log = (struct jfs_log *) JFS_SBI(tblk->sb)->log; log 437 fs/jfs/jfs_dmap.c logdiff(difft, lsn, log); log 528 fs/jfs/jfs_dmap.c LOGSYNC_LOCK(log, flags); log 531 fs/jfs/jfs_dmap.c logdiff(diffp, mp->lsn, log); log 540 fs/jfs/jfs_dmap.c logdiff(difft, tblk->clsn, log); log 541 fs/jfs/jfs_dmap.c logdiff(diffp, mp->clsn, log); log 545 fs/jfs/jfs_dmap.c mp->log = log; log 549 fs/jfs/jfs_dmap.c log->count++; log 554 fs/jfs/jfs_dmap.c LOGSYNC_UNLOCK(log, flags); log 2726 fs/jfs/jfs_imap.c struct jfs_log *log; log 2801 fs/jfs/jfs_imap.c log = JFS_SBI(tblk->sb)->log; log 2802 fs/jfs/jfs_imap.c LOGSYNC_LOCK(log, flags); log 2805 fs/jfs/jfs_imap.c logdiff(difft, lsn, log); log 2806 fs/jfs/jfs_imap.c logdiff(diffp, mp->lsn, log); log 2814 fs/jfs/jfs_imap.c logdiff(difft, tblk->clsn, log); log 2815 fs/jfs/jfs_imap.c logdiff(diffp, mp->clsn, log); log 2819 fs/jfs/jfs_imap.c mp->log = log; log 2822 fs/jfs/jfs_imap.c log->count++; log 2826 fs/jfs/jfs_imap.c LOGSYNC_UNLOCK(log, flags); log 158 fs/jfs/jfs_incore.h struct jfs_log *log; /* log */ log 214 fs/jfs/jfs_incore.h if (JFS_SBI(inode->i_sb)->log) log 80 fs/jfs/jfs_logmgr.c #define LOG_LOCK_INIT(log) mutex_init(&(log)->loglock) log 81 fs/jfs/jfs_logmgr.c #define LOG_LOCK(log) mutex_lock(&((log)->loglock)) log 82 fs/jfs/jfs_logmgr.c #define LOG_UNLOCK(log) mutex_unlock(&((log)->loglock)) log 89 fs/jfs/jfs_logmgr.c #define LOGGC_LOCK_INIT(log) spin_lock_init(&(log)->gclock) log 90 fs/jfs/jfs_logmgr.c #define LOGGC_LOCK(log) spin_lock_irq(&(log)->gclock) log 91 fs/jfs/jfs_logmgr.c #define LOGGC_UNLOCK(log) spin_unlock_irq(&(log)->gclock) log 163 fs/jfs/jfs_logmgr.c static int lmWriteRecord(struct jfs_log * log, struct tblock * tblk, log 166 fs/jfs/jfs_logmgr.c static int lmNextPage(struct jfs_log * log); log 167 fs/jfs/jfs_logmgr.c static int lmLogFileSystem(struct jfs_log * log, struct jfs_sb_info *sbi, log 172 fs/jfs/jfs_logmgr.c static int lbmLogInit(struct jfs_log * log); log 173 fs/jfs/jfs_logmgr.c static void lbmLogShutdown(struct jfs_log * log); log 174 fs/jfs/jfs_logmgr.c static struct lbuf *lbmAllocate(struct jfs_log * log, int); log 177 fs/jfs/jfs_logmgr.c static int lbmRead(struct jfs_log * log, int pn, struct lbuf ** bpp); log 178 fs/jfs/jfs_logmgr.c static void lbmWrite(struct jfs_log * log, struct lbuf * bp, int flag, int cant_block); log 179 fs/jfs/jfs_logmgr.c static void lbmDirectWrite(struct jfs_log * log, struct lbuf * bp, int flag); log 183 fs/jfs/jfs_logmgr.c static void lmGCwrite(struct jfs_log * log, int cant_block); log 184 fs/jfs/jfs_logmgr.c static int lmLogSync(struct jfs_log * log, int hard_sync); log 201 fs/jfs/jfs_logmgr.c static void write_special_inodes(struct jfs_log *log, log 206 fs/jfs/jfs_logmgr.c list_for_each_entry(sbi, &log->sb_list, log_list) { log 225 fs/jfs/jfs_logmgr.c int lmLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 234 fs/jfs/jfs_logmgr.c log, tblk, lrd, tlck); log 236 fs/jfs/jfs_logmgr.c LOG_LOCK(log); log 250 fs/jfs/jfs_logmgr.c lsn = log->lsn; log 252 fs/jfs/jfs_logmgr.c LOGSYNC_LOCK(log, flags); log 258 fs/jfs/jfs_logmgr.c mp->log = log; log 260 fs/jfs/jfs_logmgr.c log->count++; log 263 fs/jfs/jfs_logmgr.c list_add_tail(&mp->synclist, &log->synclist); log 287 fs/jfs/jfs_logmgr.c log->count++; log 297 fs/jfs/jfs_logmgr.c logdiff(diffp, mp->lsn, log); log 298 fs/jfs/jfs_logmgr.c logdiff(difft, tblk->lsn, log); log 308 fs/jfs/jfs_logmgr.c LOGSYNC_UNLOCK(log, flags); log 314 fs/jfs/jfs_logmgr.c lsn = lmWriteRecord(log, tblk, lrd, tlck); log 319 fs/jfs/jfs_logmgr.c logdiff(diffp, lsn, log); log 320 fs/jfs/jfs_logmgr.c if (diffp >= log->nextsync) log 321 fs/jfs/jfs_logmgr.c lsn = lmLogSync(log, 0); log 324 fs/jfs/jfs_logmgr.c log->lsn = lsn; log 326 fs/jfs/jfs_logmgr.c LOG_UNLOCK(log); log 344 fs/jfs/jfs_logmgr.c lmWriteRecord(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 367 fs/jfs/jfs_logmgr.c bp = (struct lbuf *) log->bp; log 369 fs/jfs/jfs_logmgr.c dstoffset = log->eor; log 416 fs/jfs/jfs_logmgr.c lmNextPage(log); log 418 fs/jfs/jfs_logmgr.c bp = log->bp; log 441 fs/jfs/jfs_logmgr.c lmNextPage(log); log 443 fs/jfs/jfs_logmgr.c bp = (struct lbuf *) log->bp; log 495 fs/jfs/jfs_logmgr.c log->eor = dstoffset; log 497 fs/jfs/jfs_logmgr.c lsn = (log->page << L2LOGPSIZE) + dstoffset; log 514 fs/jfs/jfs_logmgr.c LOGGC_LOCK(log); log 518 fs/jfs/jfs_logmgr.c tblk->bp = log->bp; log 519 fs/jfs/jfs_logmgr.c tblk->pn = log->page; log 520 fs/jfs/jfs_logmgr.c tblk->eor = log->eor; log 523 fs/jfs/jfs_logmgr.c list_add_tail(&tblk->cqueue, &log->cqueue); log 525 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 529 fs/jfs/jfs_logmgr.c le16_to_cpu(lrd->type), log->bp, log->page, dstoffset); log 537 fs/jfs/jfs_logmgr.c lmNextPage(log); log 539 fs/jfs/jfs_logmgr.c bp = (struct lbuf *) log->bp; log 560 fs/jfs/jfs_logmgr.c static int lmNextPage(struct jfs_log * log) log 570 fs/jfs/jfs_logmgr.c pn = log->page; log 571 fs/jfs/jfs_logmgr.c bp = log->bp; log 575 fs/jfs/jfs_logmgr.c LOGGC_LOCK(log); log 581 fs/jfs/jfs_logmgr.c if (list_empty(&log->cqueue)) log 584 fs/jfs/jfs_logmgr.c tblk = list_entry(log->cqueue.prev, struct tblock, cqueue); log 600 fs/jfs/jfs_logmgr.c if (log->cflag & logGC_PAGEOUT) { log 608 fs/jfs/jfs_logmgr.c lbmWrite(log, bp, 0, 0); log 613 fs/jfs/jfs_logmgr.c log->cflag |= logGC_PAGEOUT; log 614 fs/jfs/jfs_logmgr.c lmGCwrite(log, 0); log 624 fs/jfs/jfs_logmgr.c lbmWrite(log, bp, lbmWRITE | lbmRELEASE | lbmFREE, 0); log 626 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 634 fs/jfs/jfs_logmgr.c log->page = (pn == log->size - 1) ? 2 : pn + 1; log 635 fs/jfs/jfs_logmgr.c log->eor = LOGPHDRSIZE; /* ? valid page empty/full at logRedo() */ log 638 fs/jfs/jfs_logmgr.c nextbp = lbmAllocate(log, log->page); log 639 fs/jfs/jfs_logmgr.c nextbp->l_eor = log->eor; log 640 fs/jfs/jfs_logmgr.c log->bp = nextbp; log 666 fs/jfs/jfs_logmgr.c int lmGroupCommit(struct jfs_log * log, struct tblock * tblk) log 670 fs/jfs/jfs_logmgr.c LOGGC_LOCK(log); log 677 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 680 fs/jfs/jfs_logmgr.c jfs_info("lmGroup Commit: tblk = 0x%p, gcrtc = %d", tblk, log->gcrtc); log 685 fs/jfs/jfs_logmgr.c if ((!(log->cflag & logGC_PAGEOUT)) && (!list_empty(&log->cqueue)) && log 686 fs/jfs/jfs_logmgr.c (!(tblk->xflag & COMMIT_LAZY) || test_bit(log_FLUSH, &log->flag) log 693 fs/jfs/jfs_logmgr.c log->cflag |= logGC_PAGEOUT; log 695 fs/jfs/jfs_logmgr.c lmGCwrite(log, 0); log 702 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 712 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 718 fs/jfs/jfs_logmgr.c log->gcrtc++; log 722 fs/jfs/jfs_logmgr.c LOGGC_LOCK(log), LOGGC_UNLOCK(log)); log 728 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 745 fs/jfs/jfs_logmgr.c static void lmGCwrite(struct jfs_log * log, int cant_write) log 760 fs/jfs/jfs_logmgr.c gcpn = list_entry(log->cqueue.next, struct tblock, cqueue)->pn; log 762 fs/jfs/jfs_logmgr.c list_for_each_entry(tblk, &log->cqueue, cqueue) { log 785 fs/jfs/jfs_logmgr.c lbmWrite(log, bp, lbmWRITE | lbmRELEASE | lbmGC, log 793 fs/jfs/jfs_logmgr.c lbmWrite(log, bp, lbmWRITE | lbmGC, cant_write); log 813 fs/jfs/jfs_logmgr.c struct jfs_log *log = bp->l_log; log 818 fs/jfs/jfs_logmgr.c spin_lock_irqsave(&log->gclock, flags); log 825 fs/jfs/jfs_logmgr.c list_for_each_entry_safe(tblk, temp, &log->cqueue, cqueue) { log 840 fs/jfs/jfs_logmgr.c if (tblk == log->flush_tblk) { log 842 fs/jfs/jfs_logmgr.c clear_bit(log_FLUSH, &log->flag); log 843 fs/jfs/jfs_logmgr.c log->flush_tblk = NULL; log 859 fs/jfs/jfs_logmgr.c log->gcrtc--; log 878 fs/jfs/jfs_logmgr.c lbmWrite(log, bp, lbmWRITE | lbmRELEASE | lbmFREE, log 891 fs/jfs/jfs_logmgr.c if ((!list_empty(&log->cqueue)) && log 892 fs/jfs/jfs_logmgr.c ((log->gcrtc > 0) || (tblk->bp->l_wqnext != NULL) || log 893 fs/jfs/jfs_logmgr.c test_bit(log_FLUSH, &log->flag) || jfs_tlocks_low)) log 897 fs/jfs/jfs_logmgr.c lmGCwrite(log, 1); log 905 fs/jfs/jfs_logmgr.c log->cflag &= ~logGC_PAGEOUT; log 908 fs/jfs/jfs_logmgr.c spin_unlock_irqrestore(&log->gclock, flags); log 929 fs/jfs/jfs_logmgr.c static int lmLogSync(struct jfs_log * log, int hard_sync) log 943 fs/jfs/jfs_logmgr.c write_special_inodes(log, filemap_fdatawrite); log 945 fs/jfs/jfs_logmgr.c write_special_inodes(log, filemap_flush); log 954 fs/jfs/jfs_logmgr.c if (log->sync == log->syncpt) { log 955 fs/jfs/jfs_logmgr.c LOGSYNC_LOCK(log, flags); log 956 fs/jfs/jfs_logmgr.c if (list_empty(&log->synclist)) log 957 fs/jfs/jfs_logmgr.c log->sync = log->lsn; log 959 fs/jfs/jfs_logmgr.c lp = list_entry(log->synclist.next, log 961 fs/jfs/jfs_logmgr.c log->sync = lp->lsn; log 963 fs/jfs/jfs_logmgr.c LOGSYNC_UNLOCK(log, flags); log 971 fs/jfs/jfs_logmgr.c if (log->sync != log->syncpt) { log 976 fs/jfs/jfs_logmgr.c lrd.log.syncpt.sync = cpu_to_le32(log->sync); log 977 fs/jfs/jfs_logmgr.c lsn = lmWriteRecord(log, NULL, &lrd, NULL); log 979 fs/jfs/jfs_logmgr.c log->syncpt = log->sync; log 981 fs/jfs/jfs_logmgr.c lsn = log->lsn; log 986 fs/jfs/jfs_logmgr.c logsize = log->logsize; log 988 fs/jfs/jfs_logmgr.c logdiff(written, lsn, log); log 1013 fs/jfs/jfs_logmgr.c log->syncpt = log->sync = lsn; log 1014 fs/jfs/jfs_logmgr.c log->nextsync = delta; log 1017 fs/jfs/jfs_logmgr.c log->nextsync = written + more; log 1024 fs/jfs/jfs_logmgr.c if (!test_bit(log_SYNCBARRIER, &log->flag) && log 1025 fs/jfs/jfs_logmgr.c (written > LOGSYNC_BARRIER(logsize)) && log->active) { log 1026 fs/jfs/jfs_logmgr.c set_bit(log_SYNCBARRIER, &log->flag); log 1028 fs/jfs/jfs_logmgr.c log->syncpt); log 1032 fs/jfs/jfs_logmgr.c jfs_flush_journal(log, 0); log 1046 fs/jfs/jfs_logmgr.c void jfs_syncpt(struct jfs_log *log, int hard_sync) log 1047 fs/jfs/jfs_logmgr.c { LOG_LOCK(log); log 1048 fs/jfs/jfs_logmgr.c if (!test_bit(log_QUIESCE, &log->flag)) log 1049 fs/jfs/jfs_logmgr.c lmLogSync(log, hard_sync); log 1050 fs/jfs/jfs_logmgr.c LOG_UNLOCK(log); log 1070 fs/jfs/jfs_logmgr.c struct jfs_log *log; log 1080 fs/jfs/jfs_logmgr.c list_for_each_entry(log, &jfs_external_logs, journal_list) { log 1081 fs/jfs/jfs_logmgr.c if (log->bdev->bd_dev == sbi->logdev) { log 1082 fs/jfs/jfs_logmgr.c if (!uuid_equal(&log->uuid, &sbi->loguuid)) { log 1090 fs/jfs/jfs_logmgr.c if ((rc = lmLogFileSystem(log, sbi, 1))) { log 1098 fs/jfs/jfs_logmgr.c if (!(log = kzalloc(sizeof(struct jfs_log), GFP_KERNEL))) { log 1102 fs/jfs/jfs_logmgr.c INIT_LIST_HEAD(&log->sb_list); log 1103 fs/jfs/jfs_logmgr.c init_waitqueue_head(&log->syncwait); log 1112 fs/jfs/jfs_logmgr.c log); log 1118 fs/jfs/jfs_logmgr.c log->bdev = bdev; log 1119 fs/jfs/jfs_logmgr.c uuid_copy(&log->uuid, &sbi->loguuid); log 1124 fs/jfs/jfs_logmgr.c if ((rc = lmLogInit(log))) log 1127 fs/jfs/jfs_logmgr.c list_add(&log->journal_list, &jfs_external_logs); log 1132 fs/jfs/jfs_logmgr.c if ((rc = lmLogFileSystem(log, sbi, 1))) log 1136 fs/jfs/jfs_logmgr.c LOG_LOCK(log); log 1137 fs/jfs/jfs_logmgr.c list_add(&sbi->log_list, &log->sb_list); log 1138 fs/jfs/jfs_logmgr.c sbi->log = log; log 1139 fs/jfs/jfs_logmgr.c LOG_UNLOCK(log); log 1148 fs/jfs/jfs_logmgr.c list_del(&log->journal_list); log 1149 fs/jfs/jfs_logmgr.c lbmLogShutdown(log); log 1156 fs/jfs/jfs_logmgr.c kfree(log); log 1164 fs/jfs/jfs_logmgr.c struct jfs_log *log; log 1167 fs/jfs/jfs_logmgr.c if (!(log = kzalloc(sizeof(struct jfs_log), GFP_KERNEL))) log 1169 fs/jfs/jfs_logmgr.c INIT_LIST_HEAD(&log->sb_list); log 1170 fs/jfs/jfs_logmgr.c init_waitqueue_head(&log->syncwait); log 1172 fs/jfs/jfs_logmgr.c set_bit(log_INLINELOG, &log->flag); log 1173 fs/jfs/jfs_logmgr.c log->bdev = sb->s_bdev; log 1174 fs/jfs/jfs_logmgr.c log->base = addressPXD(&JFS_SBI(sb)->logpxd); log 1175 fs/jfs/jfs_logmgr.c log->size = lengthPXD(&JFS_SBI(sb)->logpxd) >> log 1177 fs/jfs/jfs_logmgr.c log->l2bsize = sb->s_blocksize_bits; log 1183 fs/jfs/jfs_logmgr.c if ((rc = lmLogInit(log))) { log 1184 fs/jfs/jfs_logmgr.c kfree(log); log 1189 fs/jfs/jfs_logmgr.c list_add(&JFS_SBI(sb)->log_list, &log->sb_list); log 1190 fs/jfs/jfs_logmgr.c JFS_SBI(sb)->log = log; log 1223 fs/jfs/jfs_logmgr.c JFS_SBI(sb)->log = dummy_log; log 1248 fs/jfs/jfs_logmgr.c int lmLogInit(struct jfs_log * log) log 1258 fs/jfs/jfs_logmgr.c jfs_info("lmLogInit: log:0x%p", log); log 1261 fs/jfs/jfs_logmgr.c LOGGC_LOCK_INIT(log); log 1264 fs/jfs/jfs_logmgr.c LOG_LOCK_INIT(log); log 1266 fs/jfs/jfs_logmgr.c LOGSYNC_LOCK_INIT(log); log 1268 fs/jfs/jfs_logmgr.c INIT_LIST_HEAD(&log->synclist); log 1270 fs/jfs/jfs_logmgr.c INIT_LIST_HEAD(&log->cqueue); log 1271 fs/jfs/jfs_logmgr.c log->flush_tblk = NULL; log 1273 fs/jfs/jfs_logmgr.c log->count = 0; log 1278 fs/jfs/jfs_logmgr.c if ((rc = lbmLogInit(log))) log 1281 fs/jfs/jfs_logmgr.c if (!test_bit(log_INLINELOG, &log->flag)) log 1282 fs/jfs/jfs_logmgr.c log->l2bsize = L2LOGPSIZE; log 1285 fs/jfs/jfs_logmgr.c if (log->no_integrity) { log 1291 fs/jfs/jfs_logmgr.c bp = lbmAllocate(log , 0); log 1292 fs/jfs/jfs_logmgr.c log->bp = bp; log 1298 fs/jfs/jfs_logmgr.c if ((rc = lbmRead(log, 1, &bpsuper))) log 1317 fs/jfs/jfs_logmgr.c if (test_bit(log_INLINELOG,&log->flag)) { log 1318 fs/jfs/jfs_logmgr.c if (log->size != le32_to_cpu(logsuper->size)) { log 1323 fs/jfs/jfs_logmgr.c log, (unsigned long long)log->base, log->size); log 1325 fs/jfs/jfs_logmgr.c if (!uuid_equal(&logsuper->uuid, &log->uuid)) { log 1329 fs/jfs/jfs_logmgr.c log->size = le32_to_cpu(logsuper->size); log 1330 fs/jfs/jfs_logmgr.c log->l2bsize = le32_to_cpu(logsuper->l2bsize); log 1332 fs/jfs/jfs_logmgr.c log, (unsigned long long)log->base, log->size); log 1335 fs/jfs/jfs_logmgr.c log->page = le32_to_cpu(logsuper->end) / LOGPSIZE; log 1336 fs/jfs/jfs_logmgr.c log->eor = le32_to_cpu(logsuper->end) - (LOGPSIZE * log->page); log 1342 fs/jfs/jfs_logmgr.c if ((rc = lbmRead(log, log->page, &bp))) log 1348 fs/jfs/jfs_logmgr.c le32_to_cpu(logsuper->end), log->page, log->eor, log 1351 fs/jfs/jfs_logmgr.c log->bp = bp; log 1352 fs/jfs/jfs_logmgr.c bp->l_pn = log->page; log 1353 fs/jfs/jfs_logmgr.c bp->l_eor = log->eor; log 1356 fs/jfs/jfs_logmgr.c if (log->eor >= LOGPSIZE - LOGPTLRSIZE) log 1357 fs/jfs/jfs_logmgr.c lmNextPage(log); log 1373 fs/jfs/jfs_logmgr.c lrd.log.syncpt.sync = 0; log 1374 fs/jfs/jfs_logmgr.c lsn = lmWriteRecord(log, NULL, &lrd, NULL); log 1375 fs/jfs/jfs_logmgr.c bp = log->bp; log 1379 fs/jfs/jfs_logmgr.c lbmWrite(log, bp, lbmWRITE | lbmSYNC, 0); log 1387 fs/jfs/jfs_logmgr.c log->serial = le32_to_cpu(logsuper->serial) + 1; log 1388 fs/jfs/jfs_logmgr.c logsuper->serial = cpu_to_le32(log->serial); log 1389 fs/jfs/jfs_logmgr.c lbmDirectWrite(log, bpsuper, lbmWRITE | lbmRELEASE | lbmSYNC); log 1395 fs/jfs/jfs_logmgr.c log->logsize = (log->size - 2) << L2LOGPSIZE; log 1396 fs/jfs/jfs_logmgr.c log->lsn = lsn; log 1397 fs/jfs/jfs_logmgr.c log->syncpt = lsn; log 1398 fs/jfs/jfs_logmgr.c log->sync = log->syncpt; log 1399 fs/jfs/jfs_logmgr.c log->nextsync = LOGSYNC_DELTA(log->logsize); log 1402 fs/jfs/jfs_logmgr.c log->lsn, log->syncpt, log->sync); log 1407 fs/jfs/jfs_logmgr.c log->clsn = lsn; log 1415 fs/jfs/jfs_logmgr.c log->wqueue = NULL; log 1423 fs/jfs/jfs_logmgr.c lbmLogShutdown(log); log 1445 fs/jfs/jfs_logmgr.c struct jfs_log *log = sbi->log; log 1449 fs/jfs/jfs_logmgr.c jfs_info("lmLogClose: log:0x%p", log); log 1452 fs/jfs/jfs_logmgr.c LOG_LOCK(log); log 1454 fs/jfs/jfs_logmgr.c LOG_UNLOCK(log); log 1455 fs/jfs/jfs_logmgr.c sbi->log = NULL; log 1463 fs/jfs/jfs_logmgr.c if (test_bit(log_INLINELOG, &log->flag)) { log 1467 fs/jfs/jfs_logmgr.c rc = lmLogShutdown(log); log 1468 fs/jfs/jfs_logmgr.c kfree(log); log 1472 fs/jfs/jfs_logmgr.c if (!log->no_integrity) log 1473 fs/jfs/jfs_logmgr.c lmLogFileSystem(log, sbi, 0); log 1475 fs/jfs/jfs_logmgr.c if (!list_empty(&log->sb_list)) log 1485 fs/jfs/jfs_logmgr.c if (log->no_integrity) log 1491 fs/jfs/jfs_logmgr.c list_del(&log->journal_list); log 1492 fs/jfs/jfs_logmgr.c bdev = log->bdev; log 1493 fs/jfs/jfs_logmgr.c rc = lmLogShutdown(log); log 1497 fs/jfs/jfs_logmgr.c kfree(log); log 1516 fs/jfs/jfs_logmgr.c void jfs_flush_journal(struct jfs_log *log, int wait) log 1522 fs/jfs/jfs_logmgr.c if (!log) log 1525 fs/jfs/jfs_logmgr.c jfs_info("jfs_flush_journal: log:0x%p wait=%d", log, wait); log 1527 fs/jfs/jfs_logmgr.c LOGGC_LOCK(log); log 1529 fs/jfs/jfs_logmgr.c if (!list_empty(&log->cqueue)) { log 1534 fs/jfs/jfs_logmgr.c target = list_entry(log->cqueue.prev, struct tblock, cqueue); log 1536 fs/jfs/jfs_logmgr.c if (test_bit(log_FLUSH, &log->flag)) { log 1543 fs/jfs/jfs_logmgr.c if (log->flush_tblk) log 1544 fs/jfs/jfs_logmgr.c log->flush_tblk = target; log 1547 fs/jfs/jfs_logmgr.c log->flush_tblk = target; log 1548 fs/jfs/jfs_logmgr.c set_bit(log_FLUSH, &log->flag); log 1553 fs/jfs/jfs_logmgr.c if (!(log->cflag & logGC_PAGEOUT)) { log 1554 fs/jfs/jfs_logmgr.c log->cflag |= logGC_PAGEOUT; log 1555 fs/jfs/jfs_logmgr.c lmGCwrite(log, 0); log 1559 fs/jfs/jfs_logmgr.c if ((wait > 1) || test_bit(log_SYNCBARRIER, &log->flag)) { log 1561 fs/jfs/jfs_logmgr.c set_bit(log_FLUSH, &log->flag); log 1562 fs/jfs/jfs_logmgr.c log->flush_tblk = NULL; log 1570 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 1572 fs/jfs/jfs_logmgr.c LOGGC_LOCK(log); log 1575 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 1580 fs/jfs/jfs_logmgr.c write_special_inodes(log, filemap_fdatawrite); log 1586 fs/jfs/jfs_logmgr.c if ((!list_empty(&log->cqueue)) || !list_empty(&log->synclist)) { log 1589 fs/jfs/jfs_logmgr.c write_special_inodes(log, filemap_fdatawrite); log 1590 fs/jfs/jfs_logmgr.c if (list_empty(&log->cqueue) && log 1591 fs/jfs/jfs_logmgr.c list_empty(&log->synclist)) log 1595 fs/jfs/jfs_logmgr.c assert(list_empty(&log->cqueue)); log 1598 fs/jfs/jfs_logmgr.c if (!list_empty(&log->synclist)) { log 1602 fs/jfs/jfs_logmgr.c list_for_each_entry(lp, &log->synclist, synclist) { log 1619 fs/jfs/jfs_logmgr.c WARN_ON(!list_empty(&log->synclist)); log 1621 fs/jfs/jfs_logmgr.c clear_bit(log_FLUSH, &log->flag); log 1638 fs/jfs/jfs_logmgr.c int lmLogShutdown(struct jfs_log * log) log 1648 fs/jfs/jfs_logmgr.c jfs_info("lmLogShutdown: log:0x%p", log); log 1650 fs/jfs/jfs_logmgr.c jfs_flush_journal(log, 2); log 1660 fs/jfs/jfs_logmgr.c lrd.log.syncpt.sync = 0; log 1662 fs/jfs/jfs_logmgr.c lsn = lmWriteRecord(log, NULL, &lrd, NULL); log 1663 fs/jfs/jfs_logmgr.c bp = log->bp; log 1666 fs/jfs/jfs_logmgr.c lbmWrite(log, log->bp, lbmWRITE | lbmRELEASE | lbmSYNC, 0); log 1667 fs/jfs/jfs_logmgr.c lbmIOWait(log->bp, lbmFREE); log 1668 fs/jfs/jfs_logmgr.c log->bp = NULL; log 1675 fs/jfs/jfs_logmgr.c if ((rc = lbmRead(log, 1, &bpsuper))) log 1681 fs/jfs/jfs_logmgr.c lbmDirectWrite(log, bpsuper, lbmWRITE | lbmRELEASE | lbmSYNC); log 1685 fs/jfs/jfs_logmgr.c lsn, log->page, log->eor); log 1691 fs/jfs/jfs_logmgr.c lbmLogShutdown(log); log 1714 fs/jfs/jfs_logmgr.c static int lmLogFileSystem(struct jfs_log * log, struct jfs_sb_info *sbi, log 1726 fs/jfs/jfs_logmgr.c if ((rc = lbmRead(log, 1, &bpsuper))) log 1768 fs/jfs/jfs_logmgr.c lbmDirectWrite(log, bpsuper, lbmWRITE | lbmRELEASE | lbmSYNC); log 1794 fs/jfs/jfs_logmgr.c static int lbmLogInit(struct jfs_log * log) log 1799 fs/jfs/jfs_logmgr.c jfs_info("lbmLogInit: log:0x%p", log); log 1802 fs/jfs/jfs_logmgr.c log->bp = NULL; log 1805 fs/jfs/jfs_logmgr.c log->wqueue = NULL; log 1816 fs/jfs/jfs_logmgr.c init_waitqueue_head(&log->free_wait); log 1818 fs/jfs/jfs_logmgr.c log->lbuf_free = NULL; log 1840 fs/jfs/jfs_logmgr.c lbuf->l_log = log; log 1843 fs/jfs/jfs_logmgr.c lbuf->l_freelist = log->lbuf_free; log 1844 fs/jfs/jfs_logmgr.c log->lbuf_free = lbuf; log 1852 fs/jfs/jfs_logmgr.c lbmLogShutdown(log); log 1862 fs/jfs/jfs_logmgr.c static void lbmLogShutdown(struct jfs_log * log) log 1866 fs/jfs/jfs_logmgr.c jfs_info("lbmLogShutdown: log:0x%p", log); log 1868 fs/jfs/jfs_logmgr.c lbuf = log->lbuf_free; log 1883 fs/jfs/jfs_logmgr.c static struct lbuf *lbmAllocate(struct jfs_log * log, int pn) log 1892 fs/jfs/jfs_logmgr.c LCACHE_SLEEP_COND(log->free_wait, (bp = log->lbuf_free), flags); log 1893 fs/jfs/jfs_logmgr.c log->lbuf_free = bp->l_freelist; log 1902 fs/jfs/jfs_logmgr.c bp->l_blkno = log->base + (pn << (L2LOGPSIZE - log->l2bsize)); log 1927 fs/jfs/jfs_logmgr.c struct jfs_log *log = bp->l_log; log 1934 fs/jfs/jfs_logmgr.c bp->l_freelist = log->lbuf_free; log 1935 fs/jfs/jfs_logmgr.c log->lbuf_free = bp; log 1937 fs/jfs/jfs_logmgr.c wake_up(&log->free_wait); log 1969 fs/jfs/jfs_logmgr.c static int lbmRead(struct jfs_log * log, int pn, struct lbuf ** bpp) log 1977 fs/jfs/jfs_logmgr.c *bpp = bp = lbmAllocate(log, pn); log 1984 fs/jfs/jfs_logmgr.c bio->bi_iter.bi_sector = bp->l_blkno << (log->l2bsize - 9); log 1985 fs/jfs/jfs_logmgr.c bio_set_dev(bio, log->bdev); log 1994 fs/jfs/jfs_logmgr.c if (log->no_integrity) { log 2022 fs/jfs/jfs_logmgr.c static void lbmWrite(struct jfs_log * log, struct lbuf * bp, int flag, log 2032 fs/jfs/jfs_logmgr.c log->base + (bp->l_pn << (L2LOGPSIZE - log->l2bsize)); log 2047 fs/jfs/jfs_logmgr.c tail = log->wqueue; log 2053 fs/jfs/jfs_logmgr.c log->wqueue = bp; log 2056 fs/jfs/jfs_logmgr.c log->wqueue = bp; log 2077 fs/jfs/jfs_logmgr.c LOGGC_UNLOCK(log); log 2079 fs/jfs/jfs_logmgr.c LOGGC_LOCK(log); log 2090 fs/jfs/jfs_logmgr.c static void lbmDirectWrite(struct jfs_log * log, struct lbuf * bp, int flag) log 2102 fs/jfs/jfs_logmgr.c log->base + (bp->l_pn << (L2LOGPSIZE - log->l2bsize)); log 2123 fs/jfs/jfs_logmgr.c struct jfs_log *log = bp->l_log; log 2128 fs/jfs/jfs_logmgr.c bio->bi_iter.bi_sector = bp->l_blkno << (log->l2bsize - 9); log 2129 fs/jfs/jfs_logmgr.c bio_set_dev(bio, log->bdev); log 2139 fs/jfs/jfs_logmgr.c if (log->no_integrity) { log 2183 fs/jfs/jfs_logmgr.c struct jfs_log *log; log 2233 fs/jfs/jfs_logmgr.c log = bp->l_log; log 2234 fs/jfs/jfs_logmgr.c log->clsn = (bp->l_pn << L2LOGPSIZE) + bp->l_ceor; log 2242 fs/jfs/jfs_logmgr.c tail = log->wqueue; log 2250 fs/jfs/jfs_logmgr.c log->wqueue = NULL; log 2364 fs/jfs/jfs_logmgr.c int lmLogFormat(struct jfs_log *log, s64 logAddress, int logSize) log 2378 fs/jfs/jfs_logmgr.c sbi = list_entry(log->sb_list.next, struct jfs_sb_info, log_list); log 2381 fs/jfs/jfs_logmgr.c bp = lbmAllocate(log, 1); log 2449 fs/jfs/jfs_logmgr.c lrd_ptr->log.syncpt.sync = 0; log 336 fs/jfs/jfs_logmgr.h } log; log 477 fs/jfs/jfs_logmgr.h #define LOGSYNC_LOCK_INIT(log) spin_lock_init(&(log)->synclock) log 478 fs/jfs/jfs_logmgr.h #define LOGSYNC_LOCK(log, flags) spin_lock_irqsave(&(log)->synclock, flags) log 479 fs/jfs/jfs_logmgr.h #define LOGSYNC_UNLOCK(log, flags) \ log 480 fs/jfs/jfs_logmgr.h spin_unlock_irqrestore(&(log)->synclock, flags) log 483 fs/jfs/jfs_logmgr.h #define logdiff(diff, lsn, log)\ log 485 fs/jfs/jfs_logmgr.h diff = (lsn) - (log)->syncpt;\ log 487 fs/jfs/jfs_logmgr.h diff += (log)->logsize;\ log 492 fs/jfs/jfs_logmgr.h extern int lmLogShutdown(struct jfs_log * log); log 493 fs/jfs/jfs_logmgr.h extern int lmLogInit(struct jfs_log * log); log 494 fs/jfs/jfs_logmgr.h extern int lmLogFormat(struct jfs_log *log, s64 logAddress, int logSize); log 497 fs/jfs/jfs_logmgr.h extern void jfs_flush_journal(struct jfs_log * log, int wait); log 498 fs/jfs/jfs_logmgr.h extern void jfs_syncpt(struct jfs_log *log, int hard_sync); log 182 fs/jfs/jfs_metapage.c mp->log = NULL; log 281 fs/jfs/jfs_metapage.c struct jfs_log *log = mp->log; log 287 fs/jfs/jfs_metapage.c if (!log) log 290 fs/jfs/jfs_metapage.c LOGSYNC_LOCK(log, flags); log 292 fs/jfs/jfs_metapage.c mp->log = NULL; log 295 fs/jfs/jfs_metapage.c log->count--; log 298 fs/jfs/jfs_metapage.c LOGSYNC_UNLOCK(log, flags); log 373 fs/jfs/jfs_metapage.c if (mp->log && !(mp->log->cflag & logGC_PAGEOUT)) log 374 fs/jfs/jfs_metapage.c jfs_flush_journal(mp->log, 0); log 34 fs/jfs/jfs_metapage.h struct jfs_log *log; log 427 fs/jfs/jfs_mount.c j_sb->s_logdev = cpu_to_le32(new_encode_dev(sbi->log->bdev->bd_dev)); log 428 fs/jfs/jfs_mount.c j_sb->s_logserial = cpu_to_le32(sbi->log->serial); log 483 fs/jfs/jfs_mount.c struct jfs_log *log = JFS_SBI(sb)->log; log 491 fs/jfs/jfs_mount.c lmLog(log, NULL, &lrd, NULL); log 151 fs/jfs/jfs_txnmgr.c static int diLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 153 fs/jfs/jfs_txnmgr.c static int dataLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 155 fs/jfs/jfs_txnmgr.c static void dtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 157 fs/jfs/jfs_txnmgr.c static void mapLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 162 fs/jfs/jfs_txnmgr.c static int txLog(struct jfs_log * log, struct tblock * tblk, log 166 fs/jfs/jfs_txnmgr.c static void xtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 352 fs/jfs/jfs_txnmgr.c struct jfs_log *log; log 355 fs/jfs/jfs_txnmgr.c log = JFS_SBI(sb)->log; log 366 fs/jfs/jfs_txnmgr.c if (test_bit(log_SYNCBARRIER, &log->flag) || log 367 fs/jfs/jfs_txnmgr.c test_bit(log_QUIESCE, &log->flag)) { log 369 fs/jfs/jfs_txnmgr.c TXN_SLEEP(&log->syncwait); log 421 fs/jfs/jfs_txnmgr.c ++log->logtid; log 422 fs/jfs/jfs_txnmgr.c tblk->logtid = log->logtid; log 424 fs/jfs/jfs_txnmgr.c ++log->active; log 449 fs/jfs/jfs_txnmgr.c struct jfs_log *log; log 451 fs/jfs/jfs_txnmgr.c log = JFS_SBI(sb)->log; log 460 fs/jfs/jfs_txnmgr.c if (test_bit(log_SYNCBARRIER, &log->flag) || log 461 fs/jfs/jfs_txnmgr.c test_bit(log_QUIESCE, &log->flag)) { log 463 fs/jfs/jfs_txnmgr.c TXN_SLEEP(&log->syncwait); log 490 fs/jfs/jfs_txnmgr.c struct jfs_log *log; log 501 fs/jfs/jfs_txnmgr.c log = JFS_SBI(tblk->sb)->log; log 515 fs/jfs/jfs_txnmgr.c spin_lock_irq(&log->gclock); // LOGGC_LOCK log 517 fs/jfs/jfs_txnmgr.c spin_unlock_irq(&log->gclock); // LOGGC_UNLOCK log 534 fs/jfs/jfs_txnmgr.c if (--log->active == 0) { log 535 fs/jfs/jfs_txnmgr.c clear_bit(log_FLUSH, &log->flag); log 540 fs/jfs/jfs_txnmgr.c if (test_bit(log_SYNCBARRIER, &log->flag)) { log 544 fs/jfs/jfs_txnmgr.c jfs_syncpt(log, 1); log 546 fs/jfs/jfs_txnmgr.c jfs_info("log barrier off: 0x%x", log->lsn); log 549 fs/jfs/jfs_txnmgr.c clear_bit(log_SYNCBARRIER, &log->flag); log 552 fs/jfs/jfs_txnmgr.c TXN_WAKEUP(&log->syncwait); log 901 fs/jfs/jfs_txnmgr.c struct jfs_log *log; log 906 fs/jfs/jfs_txnmgr.c log = JFS_SBI(tblk->sb)->log; log 930 fs/jfs/jfs_txnmgr.c LOGSYNC_LOCK(log, flags); log 932 fs/jfs/jfs_txnmgr.c logdiff(difft, tblk->clsn, log); log 933 fs/jfs/jfs_txnmgr.c logdiff(diffp, mp->clsn, log); log 938 fs/jfs/jfs_txnmgr.c LOGSYNC_UNLOCK(log, flags); log 969 fs/jfs/jfs_txnmgr.c LOGSYNC_LOCK(log, flags); log 970 fs/jfs/jfs_txnmgr.c log->count--; log 972 fs/jfs/jfs_txnmgr.c LOGSYNC_UNLOCK(log, flags); log 1129 fs/jfs/jfs_txnmgr.c struct jfs_log *log; log 1155 fs/jfs/jfs_txnmgr.c log = JFS_SBI(sb)->log; log 1156 fs/jfs/jfs_txnmgr.c cd.log = log; log 1259 fs/jfs/jfs_txnmgr.c if ((rc = txLog(log, tblk, &cd))) log 1298 fs/jfs/jfs_txnmgr.c lmLog(log, tblk, lrd, NULL); log 1300 fs/jfs/jfs_txnmgr.c lmGroupCommit(log, tblk); log 1368 fs/jfs/jfs_txnmgr.c static int txLog(struct jfs_log * log, struct tblock * tblk, struct commit * cd) log 1387 fs/jfs/jfs_txnmgr.c lrd->log.redopage.fileset = cpu_to_le32(JFS_IP(ip)->fileset); log 1388 fs/jfs/jfs_txnmgr.c lrd->log.redopage.inode = cpu_to_le32(ip->i_ino); log 1393 fs/jfs/jfs_txnmgr.c xtLog(log, tblk, lrd, tlck); log 1397 fs/jfs/jfs_txnmgr.c dtLog(log, tblk, lrd, tlck); log 1401 fs/jfs/jfs_txnmgr.c diLog(log, tblk, lrd, tlck, cd); log 1405 fs/jfs/jfs_txnmgr.c mapLog(log, tblk, lrd, tlck); log 1409 fs/jfs/jfs_txnmgr.c dataLog(log, tblk, lrd, tlck); log 1425 fs/jfs/jfs_txnmgr.c static int diLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 1436 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type = cpu_to_le16(LOG_INODE); log 1437 fs/jfs/jfs_txnmgr.c lrd->log.redopage.l2linesize = cpu_to_le16(L2INODESLOTSIZE); log 1439 fs/jfs/jfs_txnmgr.c pxd = &lrd->log.redopage.pxd; log 1450 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 1480 fs/jfs/jfs_txnmgr.c lrd->log.noredoinoext.iagnum = log 1482 fs/jfs/jfs_txnmgr.c lrd->log.noredoinoext.inoext_idx = log 1487 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL)); log 1514 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.type = log 1517 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.type = log 1519 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.nxd = cpu_to_le16(1); log 1520 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.pxd = pxdlock->pxd; log 1522 fs/jfs/jfs_txnmgr.c cpu_to_le32(lmLog(log, tblk, lrd, NULL)); log 1538 fs/jfs/jfs_txnmgr.c static int dataLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 1547 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type = cpu_to_le16(LOG_DATA); log 1548 fs/jfs/jfs_txnmgr.c lrd->log.redopage.l2linesize = cpu_to_le16(L2DATASLOTSIZE); log 1550 fs/jfs/jfs_txnmgr.c pxd = &lrd->log.redopage.pxd; log 1571 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 1584 fs/jfs/jfs_txnmgr.c static void dtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 1594 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type = cpu_to_le16(LOG_DTREE); log 1595 fs/jfs/jfs_txnmgr.c lrd->log.redopage.l2linesize = cpu_to_le16(L2DTSLOTSIZE); log 1597 fs/jfs/jfs_txnmgr.c pxd = &lrd->log.redopage.pxd; log 1600 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type |= cpu_to_le16(LOG_BTROOT); log 1615 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type |= cpu_to_le16(LOG_EXTEND); log 1617 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type |= cpu_to_le16(LOG_NEW); log 1621 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 1650 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 1672 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL)); log 1687 fs/jfs/jfs_txnmgr.c static void xtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 1704 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type = cpu_to_le16(LOG_XTREE); log 1705 fs/jfs/jfs_txnmgr.c lrd->log.redopage.l2linesize = cpu_to_le16(L2XTSLOTSIZE); log 1707 fs/jfs/jfs_txnmgr.c page_pxd = &lrd->log.redopage.pxd; log 1710 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type |= cpu_to_le16(LOG_BTROOT); log 1713 fs/jfs/jfs_txnmgr.c lrd->log.redopage.type |= log 1740 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 1831 fs/jfs/jfs_txnmgr.c cpu_to_le32(lmLog(log, tblk, lrd, NULL)); log 1837 fs/jfs/jfs_txnmgr.c cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 1846 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.type = cpu_to_le16(LOG_FREEXADLIST); log 1849 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.nxd = log 1855 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 1950 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 1964 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.type = cpu_to_le16(LOG_FREEPXD); log 1965 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.nxd = cpu_to_le16(1); log 1966 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.pxd = pxdlock->pxd; log 1969 fs/jfs/jfs_txnmgr.c cpu_to_le32(lmLog(log, tblk, lrd, NULL)); log 1981 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.type = log 1985 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.nxd = log 1992 fs/jfs/jfs_txnmgr.c cpu_to_le32(lmLog(log, tblk, lrd, tlck)); log 2070 fs/jfs/jfs_txnmgr.c static void mapLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd, log 2090 fs/jfs/jfs_txnmgr.c pxd = &lrd->log.redopage.pxd; log 2092 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL)); log 2105 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.type = cpu_to_le16(LOG_FREEPXD); log 2106 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.nxd = cpu_to_le16(1); log 2107 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.pxd = pxdlock->pxd; log 2108 fs/jfs/jfs_txnmgr.c lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL)); log 2132 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.type = log 2135 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.type = log 2137 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.nxd = cpu_to_le16(1); log 2138 fs/jfs/jfs_txnmgr.c lrd->log.updatemap.pxd = pxdlock->pxd; log 2140 fs/jfs/jfs_txnmgr.c cpu_to_le32(lmLog(log, tblk, lrd, NULL)); log 2684 fs/jfs/jfs_txnmgr.c struct jfs_log *log; log 2698 fs/jfs/jfs_txnmgr.c log = (struct jfs_log *) JFS_SBI(tblk->sb)->log; log 2700 fs/jfs/jfs_txnmgr.c spin_lock_irq(&log->gclock); // LOGGC_LOCK log 2705 fs/jfs/jfs_txnmgr.c log->gcrtc--; log 2713 fs/jfs/jfs_txnmgr.c spin_unlock_irq(&log->gclock); // LOGGC_UNLOCK log 2718 fs/jfs/jfs_txnmgr.c spin_unlock_irq(&log->gclock); // LOGGC_UNLOCK log 2825 fs/jfs/jfs_txnmgr.c struct jfs_log *log = mp->log; log 2828 fs/jfs/jfs_txnmgr.c assert(log); log 2846 fs/jfs/jfs_txnmgr.c struct jfs_log *log = JFS_SBI(sb)->log; log 2849 fs/jfs/jfs_txnmgr.c set_bit(log_QUIESCE, &log->flag); log 2890 fs/jfs/jfs_txnmgr.c jfs_flush_journal(log, 0); log 2900 fs/jfs/jfs_txnmgr.c struct jfs_log *log = JFS_SBI(sb)->log; log 2902 fs/jfs/jfs_txnmgr.c clear_bit(log_QUIESCE, &log->flag); log 2903 fs/jfs/jfs_txnmgr.c TXN_WAKEUP(&log->syncwait); log 263 fs/jfs/jfs_txnmgr.h struct jfs_log *log; /* log */ log 44 fs/jfs/jfs_umount.c struct jfs_log *log; log 54 fs/jfs/jfs_umount.c if ((log = sbi->log)) log 58 fs/jfs/jfs_umount.c jfs_flush_journal(log, 2); log 108 fs/jfs/jfs_umount.c if (log) { /* log = NULL if read-only mount */ log 126 fs/jfs/jfs_umount.c struct jfs_log *log = sbi->log; log 128 fs/jfs/jfs_umount.c if (!log) log 136 fs/jfs/jfs_umount.c jfs_flush_journal(log, 2); log 3168 fs/jfs/jfs_xtree.c int freed, log; log 3183 fs/jfs/jfs_xtree.c log = 1; log 3185 fs/jfs/jfs_xtree.c log = 0; log 3245 fs/jfs/jfs_xtree.c if (log) log 3274 fs/jfs/jfs_xtree.c if (log) { log 3352 fs/jfs/jfs_xtree.c if (log) { /* COMMIT_PWMAP */ log 3393 fs/jfs/jfs_xtree.c if (!log) { /* COMMIT_WAMP */ log 3415 fs/jfs/jfs_xtree.c if (log) { /* COMMIT_PWMAP */ log 3439 fs/jfs/jfs_xtree.c if (log) { /* COMMIT_PWMAP */ log 3483 fs/jfs/jfs_xtree.c if (log) { /* COMMIT_PWMAP */ log 3532 fs/jfs/jfs_xtree.c if (log && mp->lid && (tblk->last != mp->lid) && log 3558 fs/jfs/jfs_xtree.c if (log) { /* COMMIT_PWMAP */ log 3597 fs/jfs/jfs_xtree.c if (log) { /* COMMIT_PWMAP */ log 56 fs/jfs/resize.c struct jfs_log *log = sbi->log; log 184 fs/jfs/resize.c if ((rc = lmLogFormat(log, newLogAddress, newLogSize))) log 207 fs/jfs/resize.c lmLogShutdown(log); log 248 fs/jfs/resize.c if ((rc = lmLogFormat(log, newLogAddress, newLogSize))) log 254 fs/jfs/resize.c log->base = newLogAddress; log 255 fs/jfs/resize.c log->size = newLogSize >> (L2LOGPSIZE - sb->s_blocksize_bits); log 256 fs/jfs/resize.c if ((rc = lmLogInit(log))) log 499 fs/jfs/resize.c j_sb->s_logserial = cpu_to_le32(log->serial); log 569 fs/jfs/super.c sbi->log = NULL; log 625 fs/jfs/super.c struct jfs_log *log = sbi->log; log 630 fs/jfs/super.c rc = lmLogShutdown(log); log 655 fs/jfs/super.c struct jfs_log *log = sbi->log; log 664 fs/jfs/super.c rc = lmLogInit(log); log 681 fs/jfs/super.c struct jfs_log *log = JFS_SBI(sb)->log; log 684 fs/jfs/super.c if (log) { log 690 fs/jfs/super.c jfs_flush_journal(log, wait); log 691 fs/jfs/super.c jfs_syncpt(log, 0); log 37 fs/xfs/libxfs/xfs_log_format.h #define XLOG_BTOLSUNIT(log, b) (((b)+(log)->l_mp->m_sb.sb_logsunit-1) / \ log 38 fs/xfs/libxfs/xfs_log_format.h (log)->l_mp->m_sb.sb_logsunit) log 39 fs/xfs/libxfs/xfs_log_format.h #define XLOG_LSUNITTOB(log, su) ((su) * (log)->l_mp->m_sb.sb_logsunit) log 46 fs/xfs/libxfs/xfs_log_format.h #define XLOG_REC_SHIFT(log) \ log 47 fs/xfs/libxfs/xfs_log_format.h BTOBB(1 << (xfs_sb_version_haslogv2(&log->l_mp->m_sb) ? \ log 49 fs/xfs/libxfs/xfs_log_format.h #define XLOG_TOTAL_REC_SHIFT(log) \ log 50 fs/xfs/libxfs/xfs_log_format.h BTOBB(XLOG_MAX_ICLOGS << (xfs_sb_version_haslogv2(&log->l_mp->m_sb) ? \ log 447 fs/xfs/libxfs/xfs_rtbitmap.c int log, /* log2 of extent size */ log 463 fs/xfs/libxfs/xfs_rtbitmap.c so = XFS_SUMOFFS(mp, log, bbno); log 501 fs/xfs/libxfs/xfs_rtbitmap.c if (*sp == 0 && log == mp->m_rsum_cache[bbno]) log 503 fs/xfs/libxfs/xfs_rtbitmap.c if (*sp != 0 && log < mp->m_rsum_cache[bbno]) log 504 fs/xfs/libxfs/xfs_rtbitmap.c mp->m_rsum_cache[bbno] = log; log 517 fs/xfs/libxfs/xfs_rtbitmap.c int log, /* log2 of extent size */ log 523 fs/xfs/libxfs/xfs_rtbitmap.c return xfs_rtmodify_summary_int(mp, tp, log, bbno, log 29 fs/xfs/xfs_log.c struct xlog *log, log 42 fs/xfs/xfs_log.c struct xlog *log, log 46 fs/xfs/xfs_log.c struct xlog *log); log 54 fs/xfs/xfs_log.c struct xlog *log, log 62 fs/xfs/xfs_log.c struct xlog *log, log 66 fs/xfs/xfs_log.c struct xlog *log, log 71 fs/xfs/xfs_log.c struct xlog *log, log 76 fs/xfs/xfs_log.c struct xlog *log, log 80 fs/xfs/xfs_log.c struct xlog *log, log 84 fs/xfs/xfs_log.c struct xlog *log, log 90 fs/xfs/xfs_log.c struct xlog *log, log 94 fs/xfs/xfs_log.c struct xlog *log); log 97 fs/xfs/xfs_log.c struct xlog *log, log 102 fs/xfs/xfs_log.c struct xlog *log, log 114 fs/xfs/xfs_log.c struct xlog *log); log 118 fs/xfs/xfs_log.c struct xlog *log, log 132 fs/xfs/xfs_log.c space += log->l_logsize; log 144 fs/xfs/xfs_log.c struct xlog *log, log 157 fs/xfs/xfs_log.c tmp = log->l_logsize - space; log 194 fs/xfs/xfs_log.c struct xlog *log, log 198 fs/xfs/xfs_log.c if (head == &log->l_write_head) { log 211 fs/xfs/xfs_log.c struct xlog *log, log 242 fs/xfs/xfs_log.c need_bytes = xlog_ticket_reservation(log, head, tic); log 245 fs/xfs/xfs_log.c xlog_grant_push_ail(log, need_bytes); log 250 fs/xfs/xfs_log.c trace_xfs_log_grant_wake_up(log, tic); log 260 fs/xfs/xfs_log.c struct xlog *log, log 269 fs/xfs/xfs_log.c if (XLOG_FORCED_SHUTDOWN(log)) log 271 fs/xfs/xfs_log.c xlog_grant_push_ail(log, need_bytes); log 276 fs/xfs/xfs_log.c XFS_STATS_INC(log->l_mp, xs_sleep_logspace); log 278 fs/xfs/xfs_log.c trace_xfs_log_grant_sleep(log, tic); log 280 fs/xfs/xfs_log.c trace_xfs_log_grant_wake(log, tic); log 283 fs/xfs/xfs_log.c if (XLOG_FORCED_SHUTDOWN(log)) log 285 fs/xfs/xfs_log.c } while (xlog_space_left(log, &head->grant) < need_bytes); log 313 fs/xfs/xfs_log.c struct xlog *log, log 321 fs/xfs/xfs_log.c ASSERT(!(log->l_flags & XLOG_ACTIVE_RECOVERY)); log 329 fs/xfs/xfs_log.c *need_bytes = xlog_ticket_reservation(log, head, tic); log 330 fs/xfs/xfs_log.c free_bytes = xlog_space_left(log, &head->grant); log 333 fs/xfs/xfs_log.c if (!xlog_grant_head_wake(log, head, &free_bytes) || log 335 fs/xfs/xfs_log.c error = xlog_grant_head_wait(log, head, tic, log 341 fs/xfs/xfs_log.c error = xlog_grant_head_wait(log, head, tic, *need_bytes); log 380 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 384 fs/xfs/xfs_log.c if (XLOG_FORCED_SHUTDOWN(log)) log 397 fs/xfs/xfs_log.c xlog_grant_push_ail(log, tic->t_unit_res); log 405 fs/xfs/xfs_log.c trace_xfs_log_regrant(log, tic); log 407 fs/xfs/xfs_log.c error = xlog_grant_head_check(log, &log->l_write_head, tic, log 412 fs/xfs/xfs_log.c xlog_grant_add_space(log, &log->l_write_head.grant, need_bytes); log 413 fs/xfs/xfs_log.c trace_xfs_log_regrant_exit(log, tic); log 414 fs/xfs/xfs_log.c xlog_verify_grant_tail(log); log 445 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 452 fs/xfs/xfs_log.c if (XLOG_FORCED_SHUTDOWN(log)) log 458 fs/xfs/xfs_log.c tic = xlog_ticket_alloc(log, unit_bytes, cnt, client, permanent, 0); log 461 fs/xfs/xfs_log.c xlog_grant_push_ail(log, tic->t_cnt ? tic->t_unit_res * tic->t_cnt log 464 fs/xfs/xfs_log.c trace_xfs_log_reserve(log, tic); log 466 fs/xfs/xfs_log.c error = xlog_grant_head_check(log, &log->l_reserve_head, tic, log 471 fs/xfs/xfs_log.c xlog_grant_add_space(log, &log->l_reserve_head.grant, need_bytes); log 472 fs/xfs/xfs_log.c xlog_grant_add_space(log, &log->l_write_head.grant, need_bytes); log 473 fs/xfs/xfs_log.c trace_xfs_log_reserve_exit(log, tic); log 474 fs/xfs/xfs_log.c xlog_verify_grant_tail(log); log 517 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 520 fs/xfs/xfs_log.c if (XLOG_FORCED_SHUTDOWN(log) || log 526 fs/xfs/xfs_log.c (xlog_commit_record(log, ticket, iclog, &lsn)))) { log 533 fs/xfs/xfs_log.c trace_xfs_log_done_nonperm(log, ticket); log 539 fs/xfs/xfs_log.c xlog_ungrant_log_space(log, ticket); log 541 fs/xfs/xfs_log.c trace_xfs_log_done_perm(log, ticket); log 543 fs/xfs/xfs_log.c xlog_regrant_reserve_log_space(log, ticket); log 829 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 856 fs/xfs/xfs_log.c error = xlog_write(log, &vec, tic, &lsn, NULL, flags); log 865 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 866 fs/xfs/xfs_log.c iclog = log->l_iclog; log 868 fs/xfs/xfs_log.c xlog_state_want_sync(log, iclog); log 869 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 870 fs/xfs/xfs_log.c error = xlog_state_release_iclog(log, iclog); log 872 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 875 fs/xfs/xfs_log.c if (!XLOG_FORCED_SHUTDOWN(log)) { log 876 fs/xfs/xfs_log.c xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); log 882 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 887 fs/xfs/xfs_log.c trace_xfs_log_umount_write(log, tic); log 888 fs/xfs/xfs_log.c xlog_ungrant_log_space(log, tic); log 904 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 916 fs/xfs/xfs_log.c xfs_readonly_buftarg(log->l_targ)) { log 922 fs/xfs/xfs_log.c ASSERT(error || !(XLOG_FORCED_SHUTDOWN(log))); log 925 fs/xfs/xfs_log.c first_iclog = iclog = log->l_iclog; log 934 fs/xfs/xfs_log.c if (! (XLOG_FORCED_SHUTDOWN(log))) { log 950 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 951 fs/xfs/xfs_log.c iclog = log->l_iclog; log 954 fs/xfs/xfs_log.c xlog_state_want_sync(log, iclog); log 955 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 956 fs/xfs/xfs_log.c error = xlog_state_release_iclog(log, iclog); log 958 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 965 fs/xfs/xfs_log.c &log->l_icloglock); log 967 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 1051 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 1054 fs/xfs/xfs_log.c if (XLOG_FORCED_SHUTDOWN(log)) log 1057 fs/xfs/xfs_log.c if (!list_empty_careful(&log->l_write_head.waiters)) { log 1058 fs/xfs/xfs_log.c ASSERT(!(log->l_flags & XLOG_ACTIVE_RECOVERY)); log 1060 fs/xfs/xfs_log.c spin_lock(&log->l_write_head.lock); log 1061 fs/xfs/xfs_log.c free_bytes = xlog_space_left(log, &log->l_write_head.grant); log 1062 fs/xfs/xfs_log.c xlog_grant_head_wake(log, &log->l_write_head, &free_bytes); log 1063 fs/xfs/xfs_log.c spin_unlock(&log->l_write_head.lock); log 1066 fs/xfs/xfs_log.c if (!list_empty_careful(&log->l_reserve_head.waiters)) { log 1067 fs/xfs/xfs_log.c ASSERT(!(log->l_flags & XLOG_ACTIVE_RECOVERY)); log 1069 fs/xfs/xfs_log.c spin_lock(&log->l_reserve_head.lock); log 1070 fs/xfs/xfs_log.c free_bytes = xlog_space_left(log, &log->l_reserve_head.grant); log 1071 fs/xfs/xfs_log.c xlog_grant_head_wake(log, &log->l_reserve_head, &free_bytes); log 1072 fs/xfs/xfs_log.c spin_unlock(&log->l_reserve_head.lock); log 1096 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 1102 fs/xfs/xfs_log.c if (!xlog_cil_empty(log)) log 1105 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 1106 fs/xfs/xfs_log.c switch (log->l_covered_state) { log 1113 fs/xfs/xfs_log.c if (xfs_ail_min_lsn(log->l_ailp)) log 1115 fs/xfs/xfs_log.c if (!xlog_iclogs_empty(log)) log 1119 fs/xfs/xfs_log.c if (log->l_covered_state == XLOG_STATE_COVER_NEED) log 1120 fs/xfs/xfs_log.c log->l_covered_state = XLOG_STATE_COVER_DONE; log 1122 fs/xfs/xfs_log.c log->l_covered_state = XLOG_STATE_COVER_DONE2; log 1128 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 1139 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 1154 fs/xfs/xfs_log.c tail_lsn = atomic64_read(&log->l_last_sync_lsn); log 1155 fs/xfs/xfs_log.c trace_xfs_log_assign_tail_lsn(log, tail_lsn); log 1156 fs/xfs/xfs_log.c atomic64_set(&log->l_tail_lsn, tail_lsn); log 1189 fs/xfs/xfs_log.c struct xlog *log, log 1199 fs/xfs/xfs_log.c xlog_crack_atomic_lsn(&log->l_tail_lsn, &tail_cycle, &tail_bytes); log 1202 fs/xfs/xfs_log.c free_bytes = log->l_logsize - (head_bytes - tail_bytes); log 1214 fs/xfs/xfs_log.c xfs_alert(log->l_mp, "xlog_space_left: head behind tail"); log 1215 fs/xfs/xfs_log.c xfs_alert(log->l_mp, log 1218 fs/xfs/xfs_log.c xfs_alert(log->l_mp, log 1222 fs/xfs/xfs_log.c free_bytes = log->l_logsize; log 1234 fs/xfs/xfs_log.c struct xlog *log = iclog->ic_log; log 1248 fs/xfs/xfs_log.c if (XFS_TEST_ERROR(error, log->l_mp, XFS_ERRTAG_IODONE_IOERR)) { log 1249 fs/xfs/xfs_log.c xfs_alert(log->l_mp, "log I/O error %d", error); log 1250 fs/xfs/xfs_log.c xfs_force_shutdown(log->l_mp, SHUTDOWN_LOG_IO_ERROR); log 1284 fs/xfs/xfs_log.c struct xlog *log) log 1291 fs/xfs/xfs_log.c log->l_iclog_bufs = mp->m_logbufs; log 1292 fs/xfs/xfs_log.c log->l_iclog_size = mp->m_logbsize; log 1297 fs/xfs/xfs_log.c log->l_iclog_heads = log 1299 fs/xfs/xfs_log.c log->l_iclog_hsize = log->l_iclog_heads << BBSHIFT; log 1319 fs/xfs/xfs_log.c struct xlog *log = container_of(to_delayed_work(work), log 1321 fs/xfs/xfs_log.c struct xfs_mount *mp = log->l_mp; log 1359 fs/xfs/xfs_log.c struct xlog *log; log 1367 fs/xfs/xfs_log.c log = kmem_zalloc(sizeof(struct xlog), KM_MAYFAIL); log 1368 fs/xfs/xfs_log.c if (!log) { log 1373 fs/xfs/xfs_log.c log->l_mp = mp; log 1374 fs/xfs/xfs_log.c log->l_targ = log_target; log 1375 fs/xfs/xfs_log.c log->l_logsize = BBTOB(num_bblks); log 1376 fs/xfs/xfs_log.c log->l_logBBstart = blk_offset; log 1377 fs/xfs/xfs_log.c log->l_logBBsize = num_bblks; log 1378 fs/xfs/xfs_log.c log->l_covered_state = XLOG_STATE_COVER_IDLE; log 1379 fs/xfs/xfs_log.c log->l_flags |= XLOG_ACTIVE_RECOVERY; log 1380 fs/xfs/xfs_log.c INIT_DELAYED_WORK(&log->l_work, xfs_log_worker); log 1382 fs/xfs/xfs_log.c log->l_prev_block = -1; log 1384 fs/xfs/xfs_log.c xlog_assign_atomic_lsn(&log->l_tail_lsn, 1, 0); log 1385 fs/xfs/xfs_log.c xlog_assign_atomic_lsn(&log->l_last_sync_lsn, 1, 0); log 1386 fs/xfs/xfs_log.c log->l_curr_cycle = 1; /* 0 is bad since this is initial value */ log 1388 fs/xfs/xfs_log.c xlog_grant_head_init(&log->l_reserve_head); log 1389 fs/xfs/xfs_log.c xlog_grant_head_init(&log->l_write_head); log 1408 fs/xfs/xfs_log.c if (log2_size && log->l_logBBstart > 0 && log 1416 fs/xfs/xfs_log.c log->l_sectBBsize = 1 << log2_size; log 1418 fs/xfs/xfs_log.c xlog_get_iclog_buffer_size(mp, log); log 1420 fs/xfs/xfs_log.c spin_lock_init(&log->l_icloglock); log 1421 fs/xfs/xfs_log.c init_waitqueue_head(&log->l_flush_wait); log 1423 fs/xfs/xfs_log.c iclogp = &log->l_iclog; log 1431 fs/xfs/xfs_log.c ASSERT(log->l_iclog_size >= 4096); log 1432 fs/xfs/xfs_log.c for (i = 0; i < log->l_iclog_bufs; i++) { log 1434 fs/xfs/xfs_log.c size_t bvec_size = howmany(log->l_iclog_size, PAGE_SIZE) * log 1445 fs/xfs/xfs_log.c iclog->ic_data = kmem_alloc_io(log->l_iclog_size, align_mask, log 1450 fs/xfs/xfs_log.c log->l_iclog_bak[i] = &iclog->ic_header; log 1456 fs/xfs/xfs_log.c xfs_sb_version_haslogv2(&log->l_mp->m_sb) ? 2 : 1); log 1457 fs/xfs/xfs_log.c head->h_size = cpu_to_be32(log->l_iclog_size); log 1462 fs/xfs/xfs_log.c iclog->ic_size = log->l_iclog_size - log->l_iclog_hsize; log 1464 fs/xfs/xfs_log.c iclog->ic_log = log; log 1468 fs/xfs/xfs_log.c iclog->ic_datap = (char *)iclog->ic_data + log->l_iclog_hsize; log 1477 fs/xfs/xfs_log.c *iclogp = log->l_iclog; /* complete ring */ log 1478 fs/xfs/xfs_log.c log->l_iclog->ic_prev = prev_iclog; /* re-write 1st prev ptr */ log 1480 fs/xfs/xfs_log.c log->l_ioend_workqueue = alloc_workqueue("xfs-log/%s", log 1483 fs/xfs/xfs_log.c if (!log->l_ioend_workqueue) log 1486 fs/xfs/xfs_log.c error = xlog_cil_init(log); log 1489 fs/xfs/xfs_log.c return log; log 1492 fs/xfs/xfs_log.c destroy_workqueue(log->l_ioend_workqueue); log 1494 fs/xfs/xfs_log.c for (iclog = log->l_iclog; iclog; iclog = prev_iclog) { log 1498 fs/xfs/xfs_log.c if (prev_iclog == log->l_iclog) log 1502 fs/xfs/xfs_log.c kmem_free(log); log 1514 fs/xfs/xfs_log.c struct xlog *log, log 1519 fs/xfs/xfs_log.c struct xfs_mount *mp = log->l_mp; log 1532 fs/xfs/xfs_log.c error = xlog_write(log, &vec, ticket, commitlsnp, iclog, log 1548 fs/xfs/xfs_log.c struct xlog *log, log 1559 fs/xfs/xfs_log.c ASSERT(BTOBB(need_bytes) < log->l_logBBsize); log 1561 fs/xfs/xfs_log.c free_bytes = xlog_space_left(log, &log->l_reserve_head.grant); log 1570 fs/xfs/xfs_log.c free_threshold = max(free_threshold, (log->l_logBBsize >> 2)); log 1575 fs/xfs/xfs_log.c xlog_crack_atomic_lsn(&log->l_tail_lsn, &threshold_cycle, log 1578 fs/xfs/xfs_log.c if (threshold_block >= log->l_logBBsize) { log 1579 fs/xfs/xfs_log.c threshold_block -= log->l_logBBsize; log 1589 fs/xfs/xfs_log.c last_sync_lsn = atomic64_read(&log->l_last_sync_lsn); log 1598 fs/xfs/xfs_log.c if (!XLOG_FORCED_SHUTDOWN(log)) log 1599 fs/xfs/xfs_log.c xfs_ail_push(log->l_ailp, threshold_lsn); log 1607 fs/xfs/xfs_log.c struct xlog *log, log 1627 fs/xfs/xfs_log.c if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) { log 1638 fs/xfs/xfs_log.c for (i = 1; i < log->l_iclog_heads; i++) log 1651 fs/xfs/xfs_log.c struct xlog *log, log 1664 fs/xfs/xfs_log.c if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) { log 1715 fs/xfs/xfs_log.c struct xlog *log, log 1721 fs/xfs/xfs_log.c ASSERT(bno < log->l_logBBsize); log 1748 fs/xfs/xfs_log.c bio_set_dev(&iclog->ic_bio, log->l_targ->bt_bdev); log 1749 fs/xfs/xfs_log.c iclog->ic_bio.bi_iter.bi_sector = log->l_logBBstart + bno; log 1764 fs/xfs/xfs_log.c if (bno + BTOBB(count) > log->l_logBBsize) { log 1767 fs/xfs/xfs_log.c split = bio_split(&iclog->ic_bio, log->l_logBBsize - bno, log 1773 fs/xfs/xfs_log.c iclog->ic_bio.bi_iter.bi_sector = log->l_logBBstart; log 1786 fs/xfs/xfs_log.c struct xlog *log, log 1791 fs/xfs/xfs_log.c unsigned int split_offset = BBTOB(log->l_logBBsize - bno); log 1805 fs/xfs/xfs_log.c struct xlog *log, log 1812 fs/xfs/xfs_log.c use_lsunit = xfs_sb_version_haslogv2(&log->l_mp->m_sb) && log 1813 fs/xfs/xfs_log.c log->l_mp->m_sb.sb_logsunit > 1; log 1816 fs/xfs/xfs_log.c count_init = log->l_iclog_hsize + iclog->ic_offset; log 1821 fs/xfs/xfs_log.c count = XLOG_LSUNITTOB(log, XLOG_BTOLSUNIT(log, count_init)); log 1830 fs/xfs/xfs_log.c ASSERT(*roundoff < log->l_mp->m_sb.sb_logsunit); log 1861 fs/xfs/xfs_log.c struct xlog *log, log 1872 fs/xfs/xfs_log.c count = xlog_calc_iclog_size(log, iclog, &roundoff); log 1875 fs/xfs/xfs_log.c xlog_grant_add_space(log, &log->l_reserve_head.grant, roundoff); log 1876 fs/xfs/xfs_log.c xlog_grant_add_space(log, &log->l_write_head.grant, roundoff); log 1879 fs/xfs/xfs_log.c xlog_pack_data(log, iclog, roundoff); log 1883 fs/xfs/xfs_log.c if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) log 1887 fs/xfs/xfs_log.c XFS_STATS_INC(log->l_mp, xs_log_writes); log 1888 fs/xfs/xfs_log.c XFS_STATS_ADD(log->l_mp, xs_log_blocks, BTOBB(count)); log 1893 fs/xfs/xfs_log.c if (bno + BTOBB(count) > log->l_logBBsize) { log 1894 fs/xfs/xfs_log.c xlog_split_iclog(log, &iclog->ic_header, bno, count); log 1899 fs/xfs/xfs_log.c iclog->ic_header.h_crc = xlog_cksum(log, &iclog->ic_header, log 1909 fs/xfs/xfs_log.c if (XFS_TEST_ERROR(false, log->l_mp, XFS_ERRTAG_LOG_BAD_CRC)) { log 1912 fs/xfs/xfs_log.c xfs_warn(log->l_mp, log 1926 fs/xfs/xfs_log.c if (log->l_targ != log->l_mp->m_ddev_targp || split) { log 1927 fs/xfs/xfs_log.c xfs_blkdev_issue_flush(log->l_mp->m_ddev_targp); log 1931 fs/xfs/xfs_log.c xlog_verify_iclog(log, iclog, count); log 1932 fs/xfs/xfs_log.c xlog_write_iclog(log, iclog, bno, count, need_flush); log 1940 fs/xfs/xfs_log.c struct xlog *log) log 1945 fs/xfs/xfs_log.c xlog_cil_destroy(log); log 1951 fs/xfs/xfs_log.c iclog = log->l_iclog; log 1952 fs/xfs/xfs_log.c for (i = 0; i < log->l_iclog_bufs; i++) { log 1958 fs/xfs/xfs_log.c iclog = log->l_iclog; log 1959 fs/xfs/xfs_log.c for (i = 0; i < log->l_iclog_bufs; i++) { log 1966 fs/xfs/xfs_log.c log->l_mp->m_log = NULL; log 1967 fs/xfs/xfs_log.c destroy_workqueue(log->l_ioend_workqueue); log 1968 fs/xfs/xfs_log.c kmem_free(log); log 1977 fs/xfs/xfs_log.c struct xlog *log, log 1982 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 1987 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 2175 fs/xfs/xfs_log.c struct xlog *log, log 2198 fs/xfs/xfs_log.c xfs_warn(log->l_mp, log 2258 fs/xfs/xfs_log.c struct xlog *log, log 2273 fs/xfs/xfs_log.c xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt); log 2276 fs/xfs/xfs_log.c return xlog_state_release_iclog(log, iclog); log 2284 fs/xfs/xfs_log.c xlog_state_finish_copy(log, iclog, *record_cnt, *data_cnt); log 2288 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 2289 fs/xfs/xfs_log.c xlog_state_want_sync(log, iclog); log 2290 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 2293 fs/xfs/xfs_log.c return xlog_state_release_iclog(log, iclog); log 2343 fs/xfs/xfs_log.c struct xlog *log, log 2382 fs/xfs/xfs_log.c xfs_alert_tag(log->l_mp, XFS_PTAG_LOGRES, log 2384 fs/xfs/xfs_log.c xlog_print_tic_res(log->l_mp, ticket); log 2385 fs/xfs/xfs_log.c xfs_force_shutdown(log->l_mp, SHUTDOWN_LOG_IO_ERROR); log 2395 fs/xfs/xfs_log.c error = xlog_state_get_iclog_space(log, len, &iclog, ticket, log 2437 fs/xfs/xfs_log.c ophdr = xlog_write_setup_ophdr(log, ptr, ticket, flags); log 2450 fs/xfs/xfs_log.c xlog_verify_dest_ptr(log, ptr); log 2470 fs/xfs/xfs_log.c error = xlog_write_copy_finish(log, iclog, flags, log 2511 fs/xfs/xfs_log.c xlog_state_finish_copy(log, iclog, record_cnt, data_cnt); log 2513 fs/xfs/xfs_log.c return xlog_state_release_iclog(log, iclog); log 2546 fs/xfs/xfs_log.c struct xlog *log, log 2557 fs/xfs/xfs_log.c iclog = log->l_iclog; log 2592 fs/xfs/xfs_log.c } while (iclog != log->l_iclog); log 2609 fs/xfs/xfs_log.c switch (log->l_covered_state) { log 2613 fs/xfs/xfs_log.c log->l_covered_state = XLOG_STATE_COVER_NEED; log 2618 fs/xfs/xfs_log.c log->l_covered_state = XLOG_STATE_COVER_NEED2; log 2620 fs/xfs/xfs_log.c log->l_covered_state = XLOG_STATE_COVER_NEED; log 2625 fs/xfs/xfs_log.c log->l_covered_state = XLOG_STATE_COVER_IDLE; log 2627 fs/xfs/xfs_log.c log->l_covered_state = XLOG_STATE_COVER_NEED; log 2638 fs/xfs/xfs_log.c struct xlog *log) log 2640 fs/xfs/xfs_log.c struct xlog_in_core *iclog = log->l_iclog; log 2650 fs/xfs/xfs_log.c } while ((iclog = iclog->ic_next) != log->l_iclog); log 2679 fs/xfs/xfs_log.c struct xlog *log, log 2685 fs/xfs/xfs_log.c ASSERT(XFS_LSN_CMP(atomic64_read(&log->l_last_sync_lsn), log 2691 fs/xfs/xfs_log.c atomic64_set(&log->l_last_sync_lsn, header_lsn); log 2692 fs/xfs/xfs_log.c xlog_grant_push_ail(log, 0); log 2702 fs/xfs/xfs_log.c struct xlog *log, log 2752 fs/xfs/xfs_log.c lowest_lsn = xlog_get_lowest_lsn(log); log 2756 fs/xfs/xfs_log.c xlog_state_set_callback(log, iclog, header_lsn); log 2772 fs/xfs/xfs_log.c struct xlog *log, log 2776 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 2793 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 2811 fs/xfs/xfs_log.c struct xlog *log) log 2813 fs/xfs/xfs_log.c struct xlog_in_core *first_iclog = log->l_iclog; log 2841 fs/xfs/xfs_log.c struct xlog *log, log 2853 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 2863 fs/xfs/xfs_log.c first_iclog = log->l_iclog; log 2864 fs/xfs/xfs_log.c iclog = log->l_iclog; log 2870 fs/xfs/xfs_log.c if (xlog_state_iodone_process_iclog(log, iclog, log 2885 fs/xfs/xfs_log.c xlog_state_do_iclog_callbacks(log, iclog, aborted); log 2887 fs/xfs/xfs_log.c xlog_state_clean_iclog(log, iclog); log 2896 fs/xfs/xfs_log.c xfs_warn(log->l_mp, log 2903 fs/xfs/xfs_log.c xlog_state_callback_check_state(log); log 2905 fs/xfs/xfs_log.c if (log->l_iclog->ic_state & (XLOG_STATE_ACTIVE|XLOG_STATE_IOERROR)) log 2906 fs/xfs/xfs_log.c wake_up_all(&log->l_flush_wait); log 2908 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 2930 fs/xfs/xfs_log.c struct xlog *log = iclog->ic_log; log 2932 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 2953 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 2954 fs/xfs/xfs_log.c xlog_state_do_callback(log, aborted, iclog); /* also cleans log */ log 2978 fs/xfs/xfs_log.c struct xlog *log, log 2991 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 2992 fs/xfs/xfs_log.c if (XLOG_FORCED_SHUTDOWN(log)) { log 2993 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 2997 fs/xfs/xfs_log.c iclog = log->l_iclog; log 2999 fs/xfs/xfs_log.c XFS_STATS_INC(log->l_mp, xs_log_noiclogs); log 3002 fs/xfs/xfs_log.c xlog_wait(&log->l_flush_wait, &log->l_icloglock); log 3017 fs/xfs/xfs_log.c ticket->t_curr_res -= log->l_iclog_hsize; log 3019 fs/xfs/xfs_log.c log->l_iclog_hsize, log 3021 fs/xfs/xfs_log.c head->h_cycle = cpu_to_be32(log->l_curr_cycle); log 3023 fs/xfs/xfs_log.c xlog_assign_lsn(log->l_curr_cycle, log->l_curr_block)); log 3024 fs/xfs/xfs_log.c ASSERT(log->l_curr_block >= 0); log 3037 fs/xfs/xfs_log.c xlog_state_switch_iclogs(log, iclog, iclog->ic_size); log 3048 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3049 fs/xfs/xfs_log.c error = xlog_state_release_iclog(log, iclog); log 3053 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3069 fs/xfs/xfs_log.c xlog_state_switch_iclogs(log, iclog, iclog->ic_size); log 3074 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3089 fs/xfs/xfs_log.c struct xlog *log, log 3092 fs/xfs/xfs_log.c trace_xfs_log_regrant_reserve_enter(log, ticket); log 3097 fs/xfs/xfs_log.c xlog_grant_sub_space(log, &log->l_reserve_head.grant, log 3099 fs/xfs/xfs_log.c xlog_grant_sub_space(log, &log->l_write_head.grant, log 3104 fs/xfs/xfs_log.c trace_xfs_log_regrant_reserve_sub(log, ticket); log 3110 fs/xfs/xfs_log.c xlog_grant_add_space(log, &log->l_reserve_head.grant, log 3113 fs/xfs/xfs_log.c trace_xfs_log_regrant_reserve_exit(log, ticket); log 3136 fs/xfs/xfs_log.c struct xlog *log, log 3144 fs/xfs/xfs_log.c trace_xfs_log_ungrant_enter(log, ticket); log 3145 fs/xfs/xfs_log.c trace_xfs_log_ungrant_sub(log, ticket); log 3157 fs/xfs/xfs_log.c xlog_grant_sub_space(log, &log->l_reserve_head.grant, bytes); log 3158 fs/xfs/xfs_log.c xlog_grant_sub_space(log, &log->l_write_head.grant, bytes); log 3160 fs/xfs/xfs_log.c trace_xfs_log_ungrant_exit(log, ticket); log 3162 fs/xfs/xfs_log.c xfs_log_space_wake(log->l_mp); log 3176 fs/xfs/xfs_log.c struct xlog *log, log 3185 fs/xfs/xfs_log.c if (!atomic_dec_and_lock(&iclog->ic_refcnt, &log->l_icloglock)) log 3189 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3197 fs/xfs/xfs_log.c xfs_lsn_t tail_lsn = xlog_assign_tail_lsn(log->l_mp); log 3201 fs/xfs/xfs_log.c xlog_verify_tail_lsn(log, iclog, tail_lsn); log 3204 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3214 fs/xfs/xfs_log.c xlog_sync(log, iclog); log 3228 fs/xfs/xfs_log.c struct xlog *log, log 3236 fs/xfs/xfs_log.c iclog->ic_header.h_prev_block = cpu_to_be32(log->l_prev_block); log 3237 fs/xfs/xfs_log.c log->l_prev_block = log->l_curr_block; log 3238 fs/xfs/xfs_log.c log->l_prev_cycle = log->l_curr_cycle; log 3241 fs/xfs/xfs_log.c log->l_curr_block += BTOBB(eventual_size)+BTOBB(log->l_iclog_hsize); log 3244 fs/xfs/xfs_log.c if (xfs_sb_version_haslogv2(&log->l_mp->m_sb) && log 3245 fs/xfs/xfs_log.c log->l_mp->m_sb.sb_logsunit > 1) { log 3246 fs/xfs/xfs_log.c uint32_t sunit_bb = BTOBB(log->l_mp->m_sb.sb_logsunit); log 3247 fs/xfs/xfs_log.c log->l_curr_block = roundup(log->l_curr_block, sunit_bb); log 3250 fs/xfs/xfs_log.c if (log->l_curr_block >= log->l_logBBsize) { log 3258 fs/xfs/xfs_log.c log->l_curr_block -= log->l_logBBsize; log 3259 fs/xfs/xfs_log.c ASSERT(log->l_curr_block >= 0); log 3261 fs/xfs/xfs_log.c log->l_curr_cycle++; log 3262 fs/xfs/xfs_log.c if (log->l_curr_cycle == XLOG_HEADER_MAGIC_NUM) log 3263 fs/xfs/xfs_log.c log->l_curr_cycle++; log 3265 fs/xfs/xfs_log.c ASSERT(iclog == log->l_iclog); log 3266 fs/xfs/xfs_log.c log->l_iclog = iclog->ic_next; log 3301 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 3308 fs/xfs/xfs_log.c xlog_cil_force(log); log 3310 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 3311 fs/xfs/xfs_log.c iclog = log->l_iclog; log 3341 fs/xfs/xfs_log.c xlog_state_switch_iclogs(log, iclog, 0); log 3342 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3344 fs/xfs/xfs_log.c if (xlog_state_release_iclog(log, iclog)) log 3347 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 3359 fs/xfs/xfs_log.c xlog_state_switch_iclogs(log, iclog, 0); log 3375 fs/xfs/xfs_log.c xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); log 3381 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3384 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3396 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 3399 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 3400 fs/xfs/xfs_log.c iclog = log->l_iclog; log 3406 fs/xfs/xfs_log.c if (iclog == log->l_iclog) log 3437 fs/xfs/xfs_log.c &log->l_icloglock); log 3441 fs/xfs/xfs_log.c xlog_state_switch_iclogs(log, iclog, 0); log 3442 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3443 fs/xfs/xfs_log.c if (xlog_state_release_iclog(log, iclog)) log 3447 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 3458 fs/xfs/xfs_log.c xlog_wait(&iclog->ic_force_wait, &log->l_icloglock); log 3464 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3467 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3514 fs/xfs/xfs_log.c struct xlog *log, log 3517 fs/xfs/xfs_log.c assert_spin_locked(&log->l_icloglock); log 3520 fs/xfs/xfs_log.c xlog_state_switch_iclogs(log, iclog, 0); log 3565 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 3624 fs/xfs/xfs_log.c iclog_space = log->l_iclog_size - log->l_iclog_hsize; log 3636 fs/xfs/xfs_log.c unit_bytes += log->l_iclog_hsize * num_headers; log 3639 fs/xfs/xfs_log.c unit_bytes += log->l_iclog_hsize; log 3658 fs/xfs/xfs_log.c struct xlog *log, log 3672 fs/xfs/xfs_log.c unit_res = xfs_log_calc_unit_res(log->l_mp, unit_bytes); log 3707 fs/xfs/xfs_log.c struct xlog *log, log 3713 fs/xfs/xfs_log.c for (i = 0; i < log->l_iclog_bufs; i++) { log 3714 fs/xfs/xfs_log.c if (ptr >= log->l_iclog_bak[i] && log 3715 fs/xfs/xfs_log.c ptr <= log->l_iclog_bak[i] + log->l_iclog_size) log 3720 fs/xfs/xfs_log.c xfs_emerg(log->l_mp, "%s: invalid ptr", __func__); log 3736 fs/xfs/xfs_log.c struct xlog *log) log 3741 fs/xfs/xfs_log.c xlog_crack_grant_head(&log->l_write_head.grant, &cycle, &space); log 3742 fs/xfs/xfs_log.c xlog_crack_atomic_lsn(&log->l_tail_lsn, &tail_cycle, &tail_blocks); log 3745 fs/xfs/xfs_log.c !(log->l_flags & XLOG_TAIL_WARN)) { log 3746 fs/xfs/xfs_log.c xfs_alert_tag(log->l_mp, XFS_PTAG_LOGRES, log 3748 fs/xfs/xfs_log.c log->l_flags |= XLOG_TAIL_WARN; log 3752 fs/xfs/xfs_log.c !(log->l_flags & XLOG_TAIL_WARN)) { log 3753 fs/xfs/xfs_log.c xfs_alert_tag(log->l_mp, XFS_PTAG_LOGRES, log 3755 fs/xfs/xfs_log.c log->l_flags |= XLOG_TAIL_WARN; log 3763 fs/xfs/xfs_log.c struct xlog *log, log 3769 fs/xfs/xfs_log.c if (CYCLE_LSN(tail_lsn) == log->l_prev_cycle) { log 3771 fs/xfs/xfs_log.c log->l_logBBsize - (log->l_prev_block - BLOCK_LSN(tail_lsn)); log 3772 fs/xfs/xfs_log.c if (blocks < BTOBB(iclog->ic_offset)+BTOBB(log->l_iclog_hsize)) log 3773 fs/xfs/xfs_log.c xfs_emerg(log->l_mp, "%s: ran out of log space", __func__); log 3775 fs/xfs/xfs_log.c ASSERT(CYCLE_LSN(tail_lsn)+1 == log->l_prev_cycle); log 3777 fs/xfs/xfs_log.c if (BLOCK_LSN(tail_lsn) == log->l_prev_block) log 3778 fs/xfs/xfs_log.c xfs_emerg(log->l_mp, "%s: tail wrapped", __func__); log 3780 fs/xfs/xfs_log.c blocks = BLOCK_LSN(tail_lsn) - log->l_prev_block; log 3782 fs/xfs/xfs_log.c xfs_emerg(log->l_mp, "%s: ran out of log space", __func__); log 3803 fs/xfs/xfs_log.c struct xlog *log, log 3817 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 3818 fs/xfs/xfs_log.c icptr = log->l_iclog; log 3819 fs/xfs/xfs_log.c for (i = 0; i < log->l_iclog_bufs; i++, icptr = icptr->ic_next) log 3822 fs/xfs/xfs_log.c if (icptr != log->l_iclog) log 3823 fs/xfs/xfs_log.c xfs_emerg(log->l_mp, "%s: corrupt iclog ring", __func__); log 3824 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3828 fs/xfs/xfs_log.c xfs_emerg(log->l_mp, "%s: invalid magic num", __func__); log 3834 fs/xfs/xfs_log.c xfs_emerg(log->l_mp, "%s: unexpected magic num", log 3864 fs/xfs/xfs_log.c xfs_warn(log->l_mp, log 3895 fs/xfs/xfs_log.c struct xlog *log) log 3899 fs/xfs/xfs_log.c iclog = log->l_iclog; log 3939 fs/xfs/xfs_log.c struct xlog *log; log 3942 fs/xfs/xfs_log.c log = mp->m_log; log 3948 fs/xfs/xfs_log.c if (!log || log 3949 fs/xfs/xfs_log.c log->l_flags & XLOG_ACTIVE_RECOVERY) { log 3960 fs/xfs/xfs_log.c if (logerror && log->l_iclog->ic_state & XLOG_STATE_IOERROR) { log 3961 fs/xfs/xfs_log.c ASSERT(XLOG_FORCED_SHUTDOWN(log)); log 3979 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 3988 fs/xfs/xfs_log.c log->l_flags |= XLOG_IO_ERROR; log 3989 fs/xfs/xfs_log.c retval = xlog_state_ioerror(log); log 3990 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 3999 fs/xfs/xfs_log.c xlog_grant_head_wake_all(&log->l_reserve_head); log 4000 fs/xfs/xfs_log.c xlog_grant_head_wake_all(&log->l_write_head); log 4008 fs/xfs/xfs_log.c spin_lock(&log->l_cilp->xc_push_lock); log 4009 fs/xfs/xfs_log.c wake_up_all(&log->l_cilp->xc_commit_wait); log 4010 fs/xfs/xfs_log.c spin_unlock(&log->l_cilp->xc_push_lock); log 4011 fs/xfs/xfs_log.c xlog_state_do_callback(log, true, NULL); log 4017 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 4018 fs/xfs/xfs_log.c iclog = log->l_iclog; log 4022 fs/xfs/xfs_log.c } while (iclog != log->l_iclog); log 4023 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 4032 fs/xfs/xfs_log.c struct xlog *log) log 4036 fs/xfs/xfs_log.c iclog = log->l_iclog; log 4044 fs/xfs/xfs_log.c } while (iclog != log->l_iclog); log 4057 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 4079 fs/xfs/xfs_log.c spin_lock(&log->l_icloglock); log 4084 fs/xfs/xfs_log.c log->l_curr_cycle, log->l_curr_block); log 4085 fs/xfs/xfs_log.c spin_unlock(&log->l_icloglock); log 4095 fs/xfs/xfs_log.c struct xlog *log = mp->m_log; log 4097 fs/xfs/xfs_log.c return log->l_flags & XLOG_ACTIVE_RECOVERY; log 36 fs/xfs/xfs_log_cil.c struct xlog *log) log 40 fs/xfs/xfs_log_cil.c tic = xlog_ticket_alloc(log, 0, 1, XFS_TRANSACTION, 0, log 63 fs/xfs/xfs_log_cil.c struct xlog *log) log 65 fs/xfs/xfs_log_cil.c log->l_cilp->xc_ctx->ticket = xlog_cil_ticket_alloc(log); log 66 fs/xfs/xfs_log_cil.c log->l_cilp->xc_ctx->sequence = 1; log 126 fs/xfs/xfs_log_cil.c struct xlog *log, log 226 fs/xfs/xfs_log_cil.c struct xlog *log, log 267 fs/xfs/xfs_log_cil.c lv->lv_item->li_seq = log->l_cilp->xc_ctx->sequence; log 300 fs/xfs/xfs_log_cil.c struct xlog *log, log 375 fs/xfs/xfs_log_cil.c xfs_cil_prepare_item(log, lv, old_lv, diff_len, diff_iovecs); log 388 fs/xfs/xfs_log_cil.c struct xlog *log, log 391 fs/xfs/xfs_log_cil.c struct xfs_cil *cil = log->l_cilp; log 405 fs/xfs/xfs_log_cil.c xlog_cil_insert_format_items(log, tp, &len, &diff_iovecs); log 432 fs/xfs/xfs_log_cil.c iclog_space = log->l_iclog_size - log->l_iclog_hsize; log 437 fs/xfs/xfs_log_cil.c split_res *= log->l_iclog_hsize + sizeof(struct xlog_op_header); log 451 fs/xfs/xfs_log_cil.c xfs_warn(log->l_mp, "Transaction log reservation overrun:"); log 452 fs/xfs/xfs_log_cil.c xfs_warn(log->l_mp, log 455 fs/xfs/xfs_log_cil.c xfs_warn(log->l_mp, " split region headers: %d bytes", log 457 fs/xfs/xfs_log_cil.c xfs_warn(log->l_mp, " ctx ticket: %d bytes", ctx_res); log 484 fs/xfs/xfs_log_cil.c xfs_force_shutdown(log->l_mp, SHUTDOWN_LOG_IO_ERROR); log 644 fs/xfs/xfs_log_cil.c struct xlog *log) log 646 fs/xfs/xfs_log_cil.c struct xfs_cil *cil = log->l_cilp; log 664 fs/xfs/xfs_log_cil.c new_ctx->ticket = xlog_cil_ticket_alloc(log); log 806 fs/xfs/xfs_log_cil.c error = xlog_write(log, &lvhdr, tic, &ctx->start_lsn, NULL, 0); log 822 fs/xfs/xfs_log_cil.c if (XLOG_FORCED_SHUTDOWN(log)) { log 845 fs/xfs/xfs_log_cil.c commit_lsn = xfs_log_done(log->l_mp, tic, &commit_iclog, false); log 870 fs/xfs/xfs_log_cil.c return xfs_log_release_iclog(log->l_mp, commit_iclog); log 903 fs/xfs/xfs_log_cil.c struct xlog *log) log 905 fs/xfs/xfs_log_cil.c struct xfs_cil *cil = log->l_cilp; log 917 fs/xfs/xfs_log_cil.c if (cil->xc_ctx->space_used < XLOG_CIL_SPACE_LIMIT(log)) log 923 fs/xfs/xfs_log_cil.c queue_work(log->l_mp->m_cil_workqueue, &cil->xc_push_work); log 937 fs/xfs/xfs_log_cil.c struct xlog *log, log 940 fs/xfs/xfs_log_cil.c struct xfs_cil *cil = log->l_cilp; log 961 fs/xfs/xfs_log_cil.c queue_work(log->l_mp->m_cil_workqueue, &cil->xc_push_work); log 967 fs/xfs/xfs_log_cil.c struct xlog *log) log 969 fs/xfs/xfs_log_cil.c struct xfs_cil *cil = log->l_cilp; log 999 fs/xfs/xfs_log_cil.c struct xlog *log = mp->m_log; log 1000 fs/xfs/xfs_log_cil.c struct xfs_cil *cil = log->l_cilp; log 1009 fs/xfs/xfs_log_cil.c xlog_cil_alloc_shadow_bufs(log, tp); log 1014 fs/xfs/xfs_log_cil.c xlog_cil_insert_items(log, tp); log 1041 fs/xfs/xfs_log_cil.c xlog_cil_push_background(log); log 1058 fs/xfs/xfs_log_cil.c struct xlog *log, log 1061 fs/xfs/xfs_log_cil.c struct xfs_cil *cil = log->l_cilp; log 1073 fs/xfs/xfs_log_cil.c xlog_cil_push_now(log, sequence); log 1088 fs/xfs/xfs_log_cil.c if (XLOG_FORCED_SHUTDOWN(log)) log 1177 fs/xfs/xfs_log_cil.c struct xlog *log) log 1207 fs/xfs/xfs_log_cil.c cil->xc_log = log; log 1208 fs/xfs/xfs_log_cil.c log->l_cilp = cil; log 1214 fs/xfs/xfs_log_cil.c struct xlog *log) log 1216 fs/xfs/xfs_log_cil.c if (log->l_cilp->xc_ctx) { log 1217 fs/xfs/xfs_log_cil.c if (log->l_cilp->xc_ctx->ticket) log 1218 fs/xfs/xfs_log_cil.c xfs_log_ticket_put(log->l_cilp->xc_ctx->ticket); log 1219 fs/xfs/xfs_log_cil.c kmem_free(log->l_cilp->xc_ctx); log 1222 fs/xfs/xfs_log_cil.c ASSERT(list_empty(&log->l_cilp->xc_cil)); log 1223 fs/xfs/xfs_log_cil.c kmem_free(log->l_cilp); log 332 fs/xfs/xfs_log_priv.h #define XLOG_CIL_SPACE_LIMIT(log) (log->l_logsize >> 3) log 409 fs/xfs/xfs_log_priv.h #define XLOG_BUF_CANCEL_BUCKET(log, blkno) \ log 410 fs/xfs/xfs_log_priv.h ((log)->l_buf_cancel_table + ((uint64_t)blkno % XLOG_BC_TABLE_SIZE)) log 412 fs/xfs/xfs_log_priv.h #define XLOG_FORCED_SHUTDOWN(log) ((log)->l_flags & XLOG_IO_ERROR) log 417 fs/xfs/xfs_log_priv.h struct xlog *log); log 420 fs/xfs/xfs_log_priv.h struct xlog *log); log 424 fs/xfs/xfs_log_priv.h extern __le32 xlog_cksum(struct xlog *log, struct xlog_rec_header *rhead, log 430 fs/xfs/xfs_log_priv.h struct xlog *log, log 450 fs/xfs/xfs_log_priv.h struct xlog *log, log 515 fs/xfs/xfs_log_priv.h int xlog_cil_init(struct xlog *log); log 516 fs/xfs/xfs_log_priv.h void xlog_cil_init_post_recovery(struct xlog *log); log 517 fs/xfs/xfs_log_priv.h void xlog_cil_destroy(struct xlog *log); log 518 fs/xfs/xfs_log_priv.h bool xlog_cil_empty(struct xlog *log); log 525 fs/xfs/xfs_log_priv.h struct xlog *log, log 529 fs/xfs/xfs_log_priv.h xlog_cil_force(struct xlog *log) log 531 fs/xfs/xfs_log_priv.h xlog_cil_force_lsn(log, log->l_cilp->xc_current_sequence); log 564 fs/xfs/xfs_log_priv.h struct xlog *log, log 583 fs/xfs/xfs_log_priv.h cur_cycle = READ_ONCE(log->l_curr_cycle); log 585 fs/xfs/xfs_log_priv.h cur_block = READ_ONCE(log->l_curr_block); log 594 fs/xfs/xfs_log_priv.h spin_lock(&log->l_icloglock); log 595 fs/xfs/xfs_log_priv.h cur_cycle = log->l_curr_cycle; log 596 fs/xfs/xfs_log_priv.h cur_block = log->l_curr_block; log 597 fs/xfs/xfs_log_priv.h spin_unlock(&log->l_icloglock); log 52 fs/xfs/xfs_log_recover.c #define xlog_recover_check_summary(log) log 80 fs/xfs/xfs_log_recover.c struct xlog *log, log 84 fs/xfs/xfs_log_recover.c if (blk_no < 0 || blk_no >= log->l_logBBsize) log 86 fs/xfs/xfs_log_recover.c if (bbcount <= 0 || (blk_no + bbcount) > log->l_logBBsize) log 97 fs/xfs/xfs_log_recover.c struct xlog *log, log 100 fs/xfs/xfs_log_recover.c int align_mask = xfs_buftarg_dma_alignment(log->l_targ); log 106 fs/xfs/xfs_log_recover.c if (!xlog_verify_bno(log, 0, nbblks)) { log 107 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "Invalid block length (0x%x) for buffer", log 109 fs/xfs/xfs_log_recover.c XFS_ERROR_REPORT(__func__, XFS_ERRLEVEL_HIGH, log->l_mp); log 127 fs/xfs/xfs_log_recover.c if (nbblks > 1 && log->l_sectBBsize > 1) log 128 fs/xfs/xfs_log_recover.c nbblks += log->l_sectBBsize; log 129 fs/xfs/xfs_log_recover.c nbblks = round_up(nbblks, log->l_sectBBsize); log 139 fs/xfs/xfs_log_recover.c struct xlog *log, log 142 fs/xfs/xfs_log_recover.c return BBTOB(blk_no & ((xfs_daddr_t)log->l_sectBBsize - 1)); log 147 fs/xfs/xfs_log_recover.c struct xlog *log, log 155 fs/xfs/xfs_log_recover.c if (!xlog_verify_bno(log, blk_no, nbblks)) { log 156 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 159 fs/xfs/xfs_log_recover.c XFS_ERROR_REPORT(__func__, XFS_ERRLEVEL_HIGH, log->l_mp); log 163 fs/xfs/xfs_log_recover.c blk_no = round_down(blk_no, log->l_sectBBsize); log 164 fs/xfs/xfs_log_recover.c nbblks = round_up(nbblks, log->l_sectBBsize); log 167 fs/xfs/xfs_log_recover.c error = xfs_rw_bdev(log->l_targ->bt_bdev, log->l_logBBstart + blk_no, log 169 fs/xfs/xfs_log_recover.c if (error && !XFS_FORCED_SHUTDOWN(log->l_mp)) { log 170 fs/xfs/xfs_log_recover.c xfs_alert(log->l_mp, log 180 fs/xfs/xfs_log_recover.c struct xlog *log, log 185 fs/xfs/xfs_log_recover.c return xlog_do_io(log, blk_no, nbblks, data, REQ_OP_READ); log 190 fs/xfs/xfs_log_recover.c struct xlog *log, log 198 fs/xfs/xfs_log_recover.c error = xlog_do_io(log, blk_no, nbblks, data, REQ_OP_READ); log 200 fs/xfs/xfs_log_recover.c *offset = data + xlog_align(log, blk_no); log 206 fs/xfs/xfs_log_recover.c struct xlog *log, log 211 fs/xfs/xfs_log_recover.c return xlog_do_io(log, blk_no, nbblks, data, REQ_OP_WRITE); log 327 fs/xfs/xfs_log_recover.c struct xlog *log, log 342 fs/xfs/xfs_log_recover.c error = xlog_bread(log, mid_blk, 1, buffer, &offset); log 370 fs/xfs/xfs_log_recover.c struct xlog *log, log 390 fs/xfs/xfs_log_recover.c while (bufblks > log->l_logBBsize) log 392 fs/xfs/xfs_log_recover.c while (!(buffer = xlog_alloc_buffer(log, bufblks))) { log 394 fs/xfs/xfs_log_recover.c if (bufblks < log->l_sectBBsize) log 403 fs/xfs/xfs_log_recover.c error = xlog_bread(log, i, bcount, buffer, &buf); log 439 fs/xfs/xfs_log_recover.c struct xlog *log, log 455 fs/xfs/xfs_log_recover.c buffer = xlog_alloc_buffer(log, num_blks); log 457 fs/xfs/xfs_log_recover.c buffer = xlog_alloc_buffer(log, 1); log 462 fs/xfs/xfs_log_recover.c error = xlog_bread(log, start_blk, num_blks, buffer, &offset); log 471 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 479 fs/xfs/xfs_log_recover.c error = xlog_bread(log, i, 1, buffer, &offset); log 507 fs/xfs/xfs_log_recover.c if ((error = xlog_header_check_mount(log->l_mp, head))) log 517 fs/xfs/xfs_log_recover.c if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) { log 551 fs/xfs/xfs_log_recover.c struct xlog *log, log 560 fs/xfs/xfs_log_recover.c int error, log_bbnum = log->l_logBBsize; log 563 fs/xfs/xfs_log_recover.c error = xlog_find_zeroed(log, &first_blk); log 565 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "empty log check failed"); log 577 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "totally zeroed log"); log 584 fs/xfs/xfs_log_recover.c buffer = xlog_alloc_buffer(log, 1); log 588 fs/xfs/xfs_log_recover.c error = xlog_bread(log, 0, 1, buffer, &offset); log 595 fs/xfs/xfs_log_recover.c error = xlog_bread(log, last_blk, 1, buffer, &offset); log 665 fs/xfs/xfs_log_recover.c error = xlog_find_cycle_start(log, buffer, first_blk, &head_blk, log 678 fs/xfs/xfs_log_recover.c num_scan_bblks = min_t(int, log_bbnum, XLOG_TOTAL_REC_SHIFT(log)); log 685 fs/xfs/xfs_log_recover.c if ((error = xlog_find_verify_cycle(log, log 722 fs/xfs/xfs_log_recover.c if ((error = xlog_find_verify_cycle(log, start_blk, log 738 fs/xfs/xfs_log_recover.c if ((error = xlog_find_verify_cycle(log, log 751 fs/xfs/xfs_log_recover.c num_scan_bblks = XLOG_REC_SHIFT(log); log 756 fs/xfs/xfs_log_recover.c error = xlog_find_verify_log_record(log, start_blk, &head_blk, 0); log 764 fs/xfs/xfs_log_recover.c error = xlog_find_verify_log_record(log, start_blk, &head_blk, 0); log 774 fs/xfs/xfs_log_recover.c error = xlog_find_verify_log_record(log, start_blk, log 802 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "failed to find log head"); log 816 fs/xfs/xfs_log_recover.c struct xlog *log, log 839 fs/xfs/xfs_log_recover.c error = xlog_bread(log, i, 1, buffer, &offset); log 857 fs/xfs/xfs_log_recover.c for (i = log->l_logBBsize - 1; i >= (int) tail_blk; i--) { log 858 fs/xfs/xfs_log_recover.c error = xlog_bread(log, i, 1, buffer, &offset); log 890 fs/xfs/xfs_log_recover.c struct xlog *log, log 911 fs/xfs/xfs_log_recover.c end_blk = head_blk > tail_blk ? head_blk : log->l_logBBsize - 1; log 913 fs/xfs/xfs_log_recover.c error = xlog_bread(log, i, 1, buffer, &offset); log 931 fs/xfs/xfs_log_recover.c error = xlog_bread(log, i, 1, buffer, &offset); log 957 fs/xfs/xfs_log_recover.c struct xlog *log, log 964 fs/xfs/xfs_log_recover.c return tail_blk + (log->l_logBBsize - head_blk); log 988 fs/xfs/xfs_log_recover.c struct xlog *log, log 1001 fs/xfs/xfs_log_recover.c buffer = xlog_alloc_buffer(log, 1); log 1009 fs/xfs/xfs_log_recover.c error = xlog_seek_logrec_hdr(log, head_blk, *tail_blk, 1, buffer, log 1024 fs/xfs/xfs_log_recover.c error = xlog_do_recovery_pass(log, head_blk, *tail_blk, log 1033 fs/xfs/xfs_log_recover.c tail_distance = xlog_tail_distance(log, head_blk, first_bad); log 1038 fs/xfs/xfs_log_recover.c error = xlog_seek_logrec_hdr(log, head_blk, first_bad, 2, log 1045 fs/xfs/xfs_log_recover.c error = xlog_do_recovery_pass(log, head_blk, *tail_blk, log 1050 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 1073 fs/xfs/xfs_log_recover.c struct xlog *log, log 1095 fs/xfs/xfs_log_recover.c tmp_buffer = xlog_alloc_buffer(log, 1); log 1098 fs/xfs/xfs_log_recover.c error = xlog_rseek_logrec_hdr(log, *head_blk, *tail_blk, log 1110 fs/xfs/xfs_log_recover.c error = xlog_do_recovery_pass(log, *head_blk, tmp_rhead_blk, log 1118 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 1130 fs/xfs/xfs_log_recover.c found = xlog_rseek_logrec_hdr(log, first_bad, *tail_blk, 1, log 1156 fs/xfs/xfs_log_recover.c return xlog_verify_tail(log, *head_blk, tail_blk, log 1170 fs/xfs/xfs_log_recover.c struct xlog *log, log 1175 fs/xfs/xfs_log_recover.c div_s64_rem(bno, log->l_logBBsize, &mod); log 1186 fs/xfs/xfs_log_recover.c struct xlog *log, log 1212 fs/xfs/xfs_log_recover.c if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) { log 1228 fs/xfs/xfs_log_recover.c after_umount_blk = xlog_wrap_logbno(log, log 1233 fs/xfs/xfs_log_recover.c umount_data_blk = xlog_wrap_logbno(log, rhead_blk + hblks); log 1234 fs/xfs/xfs_log_recover.c error = xlog_bread(log, umount_data_blk, 1, buffer, &offset); log 1245 fs/xfs/xfs_log_recover.c xlog_assign_atomic_lsn(&log->l_tail_lsn, log 1246 fs/xfs/xfs_log_recover.c log->l_curr_cycle, after_umount_blk); log 1247 fs/xfs/xfs_log_recover.c xlog_assign_atomic_lsn(&log->l_last_sync_lsn, log 1248 fs/xfs/xfs_log_recover.c log->l_curr_cycle, after_umount_blk); log 1260 fs/xfs/xfs_log_recover.c struct xlog *log, log 1276 fs/xfs/xfs_log_recover.c log->l_prev_block = rhead_blk; log 1277 fs/xfs/xfs_log_recover.c log->l_curr_block = (int)head_blk; log 1278 fs/xfs/xfs_log_recover.c log->l_curr_cycle = be32_to_cpu(rhead->h_cycle); log 1280 fs/xfs/xfs_log_recover.c log->l_curr_cycle++; log 1281 fs/xfs/xfs_log_recover.c atomic64_set(&log->l_tail_lsn, be64_to_cpu(rhead->h_tail_lsn)); log 1282 fs/xfs/xfs_log_recover.c atomic64_set(&log->l_last_sync_lsn, be64_to_cpu(rhead->h_lsn)); log 1283 fs/xfs/xfs_log_recover.c xlog_assign_grant_head(&log->l_reserve_head.grant, log->l_curr_cycle, log 1284 fs/xfs/xfs_log_recover.c BBTOB(log->l_curr_block)); log 1285 fs/xfs/xfs_log_recover.c xlog_assign_grant_head(&log->l_write_head.grant, log->l_curr_cycle, log 1286 fs/xfs/xfs_log_recover.c BBTOB(log->l_curr_block)); log 1307 fs/xfs/xfs_log_recover.c struct xlog *log, log 1323 fs/xfs/xfs_log_recover.c if ((error = xlog_find_head(log, head_blk))) log 1327 fs/xfs/xfs_log_recover.c buffer = xlog_alloc_buffer(log, 1); log 1331 fs/xfs/xfs_log_recover.c error = xlog_bread(log, 0, 1, buffer, &offset); log 1347 fs/xfs/xfs_log_recover.c error = xlog_rseek_logrec_hdr(log, *head_blk, *head_blk, 1, buffer, log 1352 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: couldn't find sync record", __func__); log 1360 fs/xfs/xfs_log_recover.c xlog_set_state(log, *head_blk, rhead, rhead_blk, wrapped); log 1361 fs/xfs/xfs_log_recover.c tail_lsn = atomic64_read(&log->l_tail_lsn); log 1367 fs/xfs/xfs_log_recover.c error = xlog_check_unmount_rec(log, head_blk, tail_blk, rhead, log 1385 fs/xfs/xfs_log_recover.c error = xlog_verify_head(log, head_blk, tail_blk, buffer, log 1392 fs/xfs/xfs_log_recover.c xlog_set_state(log, *head_blk, rhead, rhead_blk, log 1394 fs/xfs/xfs_log_recover.c tail_lsn = atomic64_read(&log->l_tail_lsn); log 1395 fs/xfs/xfs_log_recover.c error = xlog_check_unmount_rec(log, head_blk, tail_blk, log 1409 fs/xfs/xfs_log_recover.c log->l_mp->m_flags |= XFS_MOUNT_WAS_CLEAN; log 1430 fs/xfs/xfs_log_recover.c if (!xfs_readonly_buftarg(log->l_targ)) log 1431 fs/xfs/xfs_log_recover.c error = xlog_clear_stale_blocks(log, tail_lsn); log 1437 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "failed to locate log tail"); log 1459 fs/xfs/xfs_log_recover.c struct xlog *log, log 1467 fs/xfs/xfs_log_recover.c int error, log_bbnum = log->l_logBBsize; log 1472 fs/xfs/xfs_log_recover.c buffer = xlog_alloc_buffer(log, 1); log 1475 fs/xfs/xfs_log_recover.c error = xlog_bread(log, 0, 1, buffer, &offset); log 1487 fs/xfs/xfs_log_recover.c error = xlog_bread(log, log_bbnum-1, 1, buffer, &offset); log 1499 fs/xfs/xfs_log_recover.c error = xlog_find_cycle_start(log, buffer, 0, &last_blk, 0); log 1509 fs/xfs/xfs_log_recover.c num_scan_bblks = XLOG_TOTAL_REC_SHIFT(log); log 1522 fs/xfs/xfs_log_recover.c if ((error = xlog_find_verify_cycle(log, start_blk, log 1532 fs/xfs/xfs_log_recover.c error = xlog_find_verify_log_record(log, start_blk, &last_blk, 0); log 1553 fs/xfs/xfs_log_recover.c struct xlog *log, log 1566 fs/xfs/xfs_log_recover.c xfs_sb_version_haslogv2(&log->l_mp->m_sb) ? 2 : 1); log 1570 fs/xfs/xfs_log_recover.c memcpy(&recp->h_fs_uuid, &log->l_mp->m_sb.sb_uuid, sizeof(uuid_t)); log 1575 fs/xfs/xfs_log_recover.c struct xlog *log, log 1585 fs/xfs/xfs_log_recover.c int sectbb = log->l_sectBBsize; log 1598 fs/xfs/xfs_log_recover.c while (bufblks > log->l_logBBsize) log 1600 fs/xfs/xfs_log_recover.c while (!(buffer = xlog_alloc_buffer(log, bufblks))) { log 1612 fs/xfs/xfs_log_recover.c error = xlog_bread_noalign(log, start_block, 1, buffer); log 1631 fs/xfs/xfs_log_recover.c error = xlog_bread_noalign(log, ealign, sectbb, log 1638 fs/xfs/xfs_log_recover.c offset = buffer + xlog_align(log, start_block); log 1640 fs/xfs/xfs_log_recover.c xlog_add_record(log, offset, cycle, i+j, log 1644 fs/xfs/xfs_log_recover.c error = xlog_bwrite(log, start_block, endcount, buffer); log 1674 fs/xfs/xfs_log_recover.c struct xlog *log, log 1685 fs/xfs/xfs_log_recover.c head_cycle = log->l_curr_cycle; log 1686 fs/xfs/xfs_log_recover.c head_block = log->l_curr_block; log 1702 fs/xfs/xfs_log_recover.c if (unlikely(head_block < tail_block || head_block >= log->l_logBBsize)) { log 1704 fs/xfs/xfs_log_recover.c XFS_ERRLEVEL_LOW, log->l_mp); log 1707 fs/xfs/xfs_log_recover.c tail_distance = tail_block + (log->l_logBBsize - head_block); log 1716 fs/xfs/xfs_log_recover.c XFS_ERRLEVEL_LOW, log->l_mp); log 1731 fs/xfs/xfs_log_recover.c max_distance = XLOG_TOTAL_REC_SHIFT(log); log 1741 fs/xfs/xfs_log_recover.c if ((head_block + max_distance) <= log->l_logBBsize) { log 1749 fs/xfs/xfs_log_recover.c error = xlog_write_log_records(log, (head_cycle - 1), log 1762 fs/xfs/xfs_log_recover.c distance = log->l_logBBsize - head_block; log 1763 fs/xfs/xfs_log_recover.c error = xlog_write_log_records(log, (head_cycle - 1), log 1778 fs/xfs/xfs_log_recover.c distance = max_distance - (log->l_logBBsize - head_block); log 1779 fs/xfs/xfs_log_recover.c error = xlog_write_log_records(log, head_cycle, 0, distance, log 1846 fs/xfs/xfs_log_recover.c struct xlog *log, log 1868 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_item_reorder_head(log, log 1890 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_item_reorder_tail(log, log 1895 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 1936 fs/xfs/xfs_log_recover.c struct xlog *log, log 1947 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_buf_not_cancel(log, buf_f); log 1955 fs/xfs/xfs_log_recover.c bucket = XLOG_BUF_CANCEL_BUCKET(log, buf_f->blf_blkno); log 1960 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_buf_cancel_ref_inc(log, buf_f); log 1971 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_buf_cancel_add(log, buf_f); log 1982 fs/xfs/xfs_log_recover.c struct xlog *log, log 1990 fs/xfs/xfs_log_recover.c if (!log->l_buf_cancel_table) { log 1996 fs/xfs/xfs_log_recover.c bucket = XLOG_BUF_CANCEL_BUCKET(log, blkno); log 2022 fs/xfs/xfs_log_recover.c struct xlog *log, log 2029 fs/xfs/xfs_log_recover.c bcp = xlog_peek_buffer_cancelled(log, blkno, len, flags); log 2664 fs/xfs/xfs_log_recover.c struct xlog *log, log 2671 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_buf_dquot_buf(log, buf_f); log 2689 fs/xfs/xfs_log_recover.c if (log->l_quotaoffs_flag & type) log 2721 fs/xfs/xfs_log_recover.c struct xlog *log, log 2727 fs/xfs/xfs_log_recover.c xfs_mount_t *mp = log->l_mp; log 2737 fs/xfs/xfs_log_recover.c if (xlog_check_buffer_cancelled(log, buf_f->blf_blkno, log 2739 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_buf_cancel(log, buf_f); log 2743 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_buf_recover(log, buf_f); log 2780 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_buf_skip(log, buf_f); log 2793 fs/xfs/xfs_log_recover.c dirty = xlog_recover_do_dquot_buffer(mp, log, item, bp, buf_f); log 2817 fs/xfs/xfs_log_recover.c (BBTOB(bp->b_length) != M_IGEO(log->l_mp)->inode_cluster_size)) { log 2913 fs/xfs/xfs_log_recover.c struct xlog *log, log 2919 fs/xfs/xfs_log_recover.c xfs_mount_t *mp = log->l_mp; log 2946 fs/xfs/xfs_log_recover.c if (xlog_check_buffer_cancelled(log, in_f->ilf_blkno, log 2949 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_inode_cancel(log, in_f); log 2952 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_inode_recover(log, in_f); log 3003 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_inode_skip(log, in_f); log 3027 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_inode_skip(log, in_f); log 3167 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: Invalid flag", __func__); log 3181 fs/xfs/xfs_log_recover.c xfs_dinode_calc_crc(log->l_mp, dip); log 3202 fs/xfs/xfs_log_recover.c struct xlog *log, log 3213 fs/xfs/xfs_log_recover.c log->l_quotaoffs_flag |= XFS_DQ_USER; log 3215 fs/xfs/xfs_log_recover.c log->l_quotaoffs_flag |= XFS_DQ_PROJ; log 3217 fs/xfs/xfs_log_recover.c log->l_quotaoffs_flag |= XFS_DQ_GROUP; log 3227 fs/xfs/xfs_log_recover.c struct xlog *log, log 3232 fs/xfs/xfs_log_recover.c xfs_mount_t *mp = log->l_mp; log 3249 fs/xfs/xfs_log_recover.c xfs_alert(log->l_mp, "NULL dquot in %s.", __func__); log 3253 fs/xfs/xfs_log_recover.c xfs_alert(log->l_mp, "dquot too small (%d) in %s.", log 3263 fs/xfs/xfs_log_recover.c if (log->l_quotaoffs_flag & type) log 3340 fs/xfs/xfs_log_recover.c struct xlog *log, log 3345 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 3359 fs/xfs/xfs_log_recover.c spin_lock(&log->l_ailp->ail_lock); log 3366 fs/xfs/xfs_log_recover.c xfs_trans_ail_update(log->l_ailp, &efip->efi_item, lsn); log 3381 fs/xfs/xfs_log_recover.c struct xlog *log, log 3389 fs/xfs/xfs_log_recover.c struct xfs_ail *ailp = log->l_ailp; log 3436 fs/xfs/xfs_log_recover.c struct xlog *log, log 3441 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 3455 fs/xfs/xfs_log_recover.c spin_lock(&log->l_ailp->ail_lock); log 3462 fs/xfs/xfs_log_recover.c xfs_trans_ail_update(log->l_ailp, &ruip->rui_item, lsn); log 3477 fs/xfs/xfs_log_recover.c struct xlog *log, log 3485 fs/xfs/xfs_log_recover.c struct xfs_ail *ailp = log->l_ailp; log 3552 fs/xfs/xfs_log_recover.c struct xlog *log, log 3557 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 3571 fs/xfs/xfs_log_recover.c spin_lock(&log->l_ailp->ail_lock); log 3578 fs/xfs/xfs_log_recover.c xfs_trans_ail_update(log->l_ailp, &cuip->cui_item, lsn); log 3593 fs/xfs/xfs_log_recover.c struct xlog *log, log 3601 fs/xfs/xfs_log_recover.c struct xfs_ail *ailp = log->l_ailp; log 3669 fs/xfs/xfs_log_recover.c struct xlog *log, log 3674 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 3690 fs/xfs/xfs_log_recover.c spin_lock(&log->l_ailp->ail_lock); log 3697 fs/xfs/xfs_log_recover.c xfs_trans_ail_update(log->l_ailp, &buip->bui_item, lsn); log 3712 fs/xfs/xfs_log_recover.c struct xlog *log, log 3720 fs/xfs/xfs_log_recover.c struct xfs_ail *ailp = log->l_ailp; log 3766 fs/xfs/xfs_log_recover.c struct xlog *log, log 3770 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 3785 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad type"); log 3790 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad icl size"); log 3796 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad agno"); log 3801 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad agbno"); log 3806 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad isize"); log 3811 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad count"); log 3816 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad length"); log 3826 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 3833 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 3852 fs/xfs/xfs_log_recover.c if (xlog_check_buffer_cancelled(log, daddr, bb_per_cluster, 0)) log 3871 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_icreate_cancel(log, icl); log 3875 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_icreate_recover(log, icl); log 3882 fs/xfs/xfs_log_recover.c struct xlog *log, log 3886 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 3888 fs/xfs/xfs_log_recover.c if (xlog_peek_buffer_cancelled(log, buf_f->blf_blkno, log 3899 fs/xfs/xfs_log_recover.c struct xlog *log, log 3904 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 3917 fs/xfs/xfs_log_recover.c if (xlog_peek_buffer_cancelled(log, ilfp->ilf_blkno, ilfp->ilf_len, 0)) log 3926 fs/xfs/xfs_log_recover.c struct xlog *log, log 3929 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 3947 fs/xfs/xfs_log_recover.c if (log->l_quotaoffs_flag & type) log 3955 fs/xfs/xfs_log_recover.c if (xlog_peek_buffer_cancelled(log, dq_f->qlf_blkno, len, 0)) log 3964 fs/xfs/xfs_log_recover.c struct xlog *log, log 3969 fs/xfs/xfs_log_recover.c xlog_recover_buffer_ra_pass2(log, item); log 3972 fs/xfs/xfs_log_recover.c xlog_recover_inode_ra_pass2(log, item); log 3975 fs/xfs/xfs_log_recover.c xlog_recover_dquot_ra_pass2(log, item); log 3993 fs/xfs/xfs_log_recover.c struct xlog *log, log 3997 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_item_recover(log, trans, item, XLOG_RECOVER_PASS1); log 4001 fs/xfs/xfs_log_recover.c return xlog_recover_buffer_pass1(log, item); log 4003 fs/xfs/xfs_log_recover.c return xlog_recover_quotaoff_pass1(log, item); log 4018 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: invalid item type (%d)", log 4027 fs/xfs/xfs_log_recover.c struct xlog *log, log 4032 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_item_recover(log, trans, item, XLOG_RECOVER_PASS2); log 4036 fs/xfs/xfs_log_recover.c return xlog_recover_buffer_pass2(log, buffer_list, item, log 4039 fs/xfs/xfs_log_recover.c return xlog_recover_inode_pass2(log, buffer_list, item, log 4042 fs/xfs/xfs_log_recover.c return xlog_recover_efi_pass2(log, item, trans->r_lsn); log 4044 fs/xfs/xfs_log_recover.c return xlog_recover_efd_pass2(log, item); log 4046 fs/xfs/xfs_log_recover.c return xlog_recover_rui_pass2(log, item, trans->r_lsn); log 4048 fs/xfs/xfs_log_recover.c return xlog_recover_rud_pass2(log, item); log 4050 fs/xfs/xfs_log_recover.c return xlog_recover_cui_pass2(log, item, trans->r_lsn); log 4052 fs/xfs/xfs_log_recover.c return xlog_recover_cud_pass2(log, item); log 4054 fs/xfs/xfs_log_recover.c return xlog_recover_bui_pass2(log, item, trans->r_lsn); log 4056 fs/xfs/xfs_log_recover.c return xlog_recover_bud_pass2(log, item); log 4058 fs/xfs/xfs_log_recover.c return xlog_recover_dquot_pass2(log, buffer_list, item, log 4061 fs/xfs/xfs_log_recover.c return xlog_recover_do_icreate_pass2(log, buffer_list, item); log 4066 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: invalid item type (%d)", log 4075 fs/xfs/xfs_log_recover.c struct xlog *log, log 4084 fs/xfs/xfs_log_recover.c error = xlog_recover_commit_pass2(log, trans, log 4101 fs/xfs/xfs_log_recover.c struct xlog *log, log 4117 fs/xfs/xfs_log_recover.c error = xlog_recover_reorder_trans(log, trans, pass); log 4124 fs/xfs/xfs_log_recover.c error = xlog_recover_commit_pass1(log, trans, item); log 4127 fs/xfs/xfs_log_recover.c xlog_recover_ra_pass2(log, item); log 4131 fs/xfs/xfs_log_recover.c error = xlog_recover_items_pass2(log, trans, log 4149 fs/xfs/xfs_log_recover.c error = xlog_recover_items_pass2(log, trans, log 4173 fs/xfs/xfs_log_recover.c struct xlog *log, log 4189 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: bad header length", __func__); log 4210 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_item_add_cont(log, trans, item, 0); log 4229 fs/xfs/xfs_log_recover.c struct xlog *log, log 4243 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: bad header magic number", log 4250 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: bad header length", __func__); log 4283 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 4301 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_item_add(log, trans, item, 0); log 4337 fs/xfs/xfs_log_recover.c struct xlog *log, log 4361 fs/xfs/xfs_log_recover.c error = xlog_recover_add_to_trans(log, trans, dp, len); log 4364 fs/xfs/xfs_log_recover.c error = xlog_recover_add_to_cont_trans(log, trans, dp, len); log 4367 fs/xfs/xfs_log_recover.c error = xlog_recover_commit_trans(log, trans, pass, log 4376 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: Unmount LR", __func__); log 4381 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: bad flag 0x%x", __func__, flags); log 4444 fs/xfs/xfs_log_recover.c struct xlog *log, log 4460 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: bad clientid 0x%x", log 4471 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: bad length 0x%x", __func__, len); log 4505 fs/xfs/xfs_log_recover.c if (log->l_recovery_lsn != trans->r_lsn && log 4510 fs/xfs/xfs_log_recover.c log->l_recovery_lsn = trans->r_lsn; log 4513 fs/xfs/xfs_log_recover.c return xlog_recovery_process_trans(log, trans, dp, len, log 4528 fs/xfs/xfs_log_recover.c struct xlog *log, log 4544 fs/xfs/xfs_log_recover.c if (xlog_header_check_recover(log->l_mp, rhead)) log 4547 fs/xfs/xfs_log_recover.c trace_xfs_log_recover_record(log, rhead, pass); log 4555 fs/xfs/xfs_log_recover.c error = xlog_recover_process_ophdr(log, rhash, rhead, ohead, log 4792 fs/xfs/xfs_log_recover.c struct xlog *log) log 4812 fs/xfs/xfs_log_recover.c error = xfs_trans_alloc_empty(log->l_mp, &parent_tp); log 4816 fs/xfs/xfs_log_recover.c ailp = log->l_ailp; log 4820 fs/xfs/xfs_log_recover.c last_lsn = xlog_assign_lsn(log->l_curr_cycle, log->l_curr_block); log 4850 fs/xfs/xfs_log_recover.c error = xlog_recover_process_efi(log->l_mp, ailp, lip); log 4853 fs/xfs/xfs_log_recover.c error = xlog_recover_process_rui(log->l_mp, ailp, lip); log 4882 fs/xfs/xfs_log_recover.c struct xlog *log) log 4888 fs/xfs/xfs_log_recover.c ailp = log->l_ailp; log 4906 fs/xfs/xfs_log_recover.c xlog_recover_cancel_efi(log->l_mp, ailp, lip); log 4909 fs/xfs/xfs_log_recover.c xlog_recover_cancel_rui(log->l_mp, ailp, lip); log 4912 fs/xfs/xfs_log_recover.c xlog_recover_cancel_cui(log->l_mp, ailp, lip); log 4915 fs/xfs/xfs_log_recover.c xlog_recover_cancel_bui(log->l_mp, ailp, lip); log 5051 fs/xfs/xfs_log_recover.c struct xlog *log) log 5061 fs/xfs/xfs_log_recover.c mp = log->l_mp; log 5105 fs/xfs/xfs_log_recover.c struct xlog *log) log 5115 fs/xfs/xfs_log_recover.c if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) { log 5131 fs/xfs/xfs_log_recover.c struct xlog *log, log 5141 fs/xfs/xfs_log_recover.c crc = xlog_cksum(log, rhead, dp, be32_to_cpu(rhead->h_len)); log 5163 fs/xfs/xfs_log_recover.c if (old_crc || xfs_sb_version_hascrc(&log->l_mp->m_sb)) { log 5164 fs/xfs/xfs_log_recover.c xfs_alert(log->l_mp, log 5175 fs/xfs/xfs_log_recover.c if (xfs_sb_version_hascrc(&log->l_mp->m_sb)) log 5179 fs/xfs/xfs_log_recover.c xlog_unpack_data(rhead, dp, log); log 5181 fs/xfs/xfs_log_recover.c return xlog_recover_process_data(log, rhash, rhead, dp, pass, log 5187 fs/xfs/xfs_log_recover.c struct xlog *log, log 5195 fs/xfs/xfs_log_recover.c XFS_ERRLEVEL_LOW, log->l_mp); log 5201 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, "%s: unrecognised log version (%d).", log 5210 fs/xfs/xfs_log_recover.c XFS_ERRLEVEL_LOW, log->l_mp); log 5213 fs/xfs/xfs_log_recover.c if (unlikely( blkno > log->l_logBBsize || blkno > INT_MAX )) { log 5215 fs/xfs/xfs_log_recover.c XFS_ERRLEVEL_LOW, log->l_mp); log 5231 fs/xfs/xfs_log_recover.c struct xlog *log, log 5260 fs/xfs/xfs_log_recover.c if (xfs_sb_version_haslogv2(&log->l_mp->m_sb)) { log 5266 fs/xfs/xfs_log_recover.c hbp = xlog_alloc_buffer(log, 1); log 5270 fs/xfs/xfs_log_recover.c error = xlog_bread(log, tail_blk, 1, hbp, &offset); log 5275 fs/xfs/xfs_log_recover.c error = xlog_valid_rec_header(log, rhead, tail_blk); log 5293 fs/xfs/xfs_log_recover.c if (h_len <= log->l_mp->m_logbsize && log 5295 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 5297 fs/xfs/xfs_log_recover.c h_size, log->l_mp->m_logbsize); log 5298 fs/xfs/xfs_log_recover.c h_size = log->l_mp->m_logbsize; log 5309 fs/xfs/xfs_log_recover.c hbp = xlog_alloc_buffer(log, hblks); log 5314 fs/xfs/xfs_log_recover.c ASSERT(log->l_sectBBsize == 1); log 5316 fs/xfs/xfs_log_recover.c hbp = xlog_alloc_buffer(log, 1); log 5322 fs/xfs/xfs_log_recover.c dbp = xlog_alloc_buffer(log, BTOBB(h_size)); log 5335 fs/xfs/xfs_log_recover.c while (blk_no < log->l_logBBsize) { log 5342 fs/xfs/xfs_log_recover.c if (blk_no + hblks <= log->l_logBBsize) { log 5344 fs/xfs/xfs_log_recover.c error = xlog_bread(log, blk_no, hblks, hbp, log 5350 fs/xfs/xfs_log_recover.c if (blk_no != log->l_logBBsize) { log 5353 fs/xfs/xfs_log_recover.c split_hblks = log->l_logBBsize - (int)blk_no; log 5355 fs/xfs/xfs_log_recover.c error = xlog_bread(log, blk_no, log 5375 fs/xfs/xfs_log_recover.c error = xlog_bread_noalign(log, 0, log 5382 fs/xfs/xfs_log_recover.c error = xlog_valid_rec_header(log, rhead, log 5397 fs/xfs/xfs_log_recover.c if (blk_no + bblks <= log->l_logBBsize || log 5398 fs/xfs/xfs_log_recover.c blk_no >= log->l_logBBsize) { log 5399 fs/xfs/xfs_log_recover.c rblk_no = xlog_wrap_logbno(log, blk_no); log 5400 fs/xfs/xfs_log_recover.c error = xlog_bread(log, rblk_no, bblks, dbp, log 5409 fs/xfs/xfs_log_recover.c if (blk_no != log->l_logBBsize) { log 5415 fs/xfs/xfs_log_recover.c log->l_logBBsize - (int)blk_no; log 5417 fs/xfs/xfs_log_recover.c error = xlog_bread(log, blk_no, log 5436 fs/xfs/xfs_log_recover.c error = xlog_bread_noalign(log, 0, log 5443 fs/xfs/xfs_log_recover.c error = xlog_recover_process(log, rhash, rhead, offset, log 5452 fs/xfs/xfs_log_recover.c ASSERT(blk_no >= log->l_logBBsize); log 5453 fs/xfs/xfs_log_recover.c blk_no -= log->l_logBBsize; log 5459 fs/xfs/xfs_log_recover.c error = xlog_bread(log, blk_no, hblks, hbp, &offset); log 5464 fs/xfs/xfs_log_recover.c error = xlog_valid_rec_header(log, rhead, blk_no); log 5470 fs/xfs/xfs_log_recover.c error = xlog_bread(log, blk_no+hblks, bblks, dbp, log 5475 fs/xfs/xfs_log_recover.c error = xlog_recover_process(log, rhash, rhead, offset, pass, log 5530 fs/xfs/xfs_log_recover.c struct xlog *log, log 5542 fs/xfs/xfs_log_recover.c log->l_buf_cancel_table = kmem_zalloc(XLOG_BC_TABLE_SIZE * log 5546 fs/xfs/xfs_log_recover.c INIT_LIST_HEAD(&log->l_buf_cancel_table[i]); log 5548 fs/xfs/xfs_log_recover.c error = xlog_do_recovery_pass(log, head_blk, tail_blk, log 5551 fs/xfs/xfs_log_recover.c kmem_free(log->l_buf_cancel_table); log 5552 fs/xfs/xfs_log_recover.c log->l_buf_cancel_table = NULL; log 5559 fs/xfs/xfs_log_recover.c error = xlog_do_recovery_pass(log, head_blk, tail_blk, log 5566 fs/xfs/xfs_log_recover.c ASSERT(list_empty(&log->l_buf_cancel_table[i])); log 5570 fs/xfs/xfs_log_recover.c kmem_free(log->l_buf_cancel_table); log 5571 fs/xfs/xfs_log_recover.c log->l_buf_cancel_table = NULL; log 5581 fs/xfs/xfs_log_recover.c struct xlog *log, log 5585 fs/xfs/xfs_log_recover.c struct xfs_mount *mp = log->l_mp; log 5590 fs/xfs/xfs_log_recover.c trace_xfs_log_recover(log, head_blk, tail_blk); log 5595 fs/xfs/xfs_log_recover.c error = xlog_do_log_recovery(log, head_blk, tail_blk); log 5651 fs/xfs/xfs_log_recover.c xlog_recover_check_summary(log); log 5654 fs/xfs/xfs_log_recover.c log->l_flags &= ~XLOG_ACTIVE_RECOVERY; log 5665 fs/xfs/xfs_log_recover.c struct xlog *log) log 5671 fs/xfs/xfs_log_recover.c error = xlog_find_tail(log, &head_blk, &tail_blk); log 5680 fs/xfs/xfs_log_recover.c if (xfs_sb_version_hascrc(&log->l_mp->m_sb) && log 5681 fs/xfs/xfs_log_recover.c !xfs_log_check_lsn(log->l_mp, log->l_mp->m_sb.sb_lsn)) log 5696 fs/xfs/xfs_log_recover.c if ((error = xfs_dev_is_read_only(log->l_mp, "recovery"))) { log 5707 fs/xfs/xfs_log_recover.c if (XFS_SB_VERSION_NUM(&log->l_mp->m_sb) == XFS_SB_VERSION_5 && log 5708 fs/xfs/xfs_log_recover.c xfs_sb_has_incompat_log_feature(&log->l_mp->m_sb, log 5710 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 5712 fs/xfs/xfs_log_recover.c (log->l_mp->m_sb.sb_features_log_incompat & log 5714 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 5716 fs/xfs/xfs_log_recover.c xfs_warn(log->l_mp, log 5727 fs/xfs/xfs_log_recover.c xfs_notice(log->l_mp, log 5733 fs/xfs/xfs_log_recover.c xfs_notice(log->l_mp, "Starting recovery (logdev: %s)", log 5734 fs/xfs/xfs_log_recover.c log->l_mp->m_logname ? log->l_mp->m_logname log 5737 fs/xfs/xfs_log_recover.c error = xlog_do_recover(log, head_blk, tail_blk); log 5738 fs/xfs/xfs_log_recover.c log->l_flags |= XLOG_RECOVERY_NEEDED; log 5754 fs/xfs/xfs_log_recover.c struct xlog *log) log 5764 fs/xfs/xfs_log_recover.c if (log->l_flags & XLOG_RECOVERY_NEEDED) { log 5766 fs/xfs/xfs_log_recover.c error = xlog_recover_process_intents(log); log 5768 fs/xfs/xfs_log_recover.c xfs_alert(log->l_mp, "Failed to recover intents"); log 5778 fs/xfs/xfs_log_recover.c xfs_log_force(log->l_mp, XFS_LOG_SYNC); log 5780 fs/xfs/xfs_log_recover.c xlog_recover_process_iunlinks(log); log 5782 fs/xfs/xfs_log_recover.c xlog_recover_check_summary(log); log 5784 fs/xfs/xfs_log_recover.c xfs_notice(log->l_mp, "Ending recovery (logdev: %s)", log 5785 fs/xfs/xfs_log_recover.c log->l_mp->m_logname ? log->l_mp->m_logname log 5787 fs/xfs/xfs_log_recover.c log->l_flags &= ~XLOG_RECOVERY_NEEDED; log 5789 fs/xfs/xfs_log_recover.c xfs_info(log->l_mp, "Ending clean mount"); log 5796 fs/xfs/xfs_log_recover.c struct xlog *log) log 5798 fs/xfs/xfs_log_recover.c if (log->l_flags & XLOG_RECOVERY_NEEDED) log 5799 fs/xfs/xfs_log_recover.c xlog_recover_cancel_intents(log); log 5809 fs/xfs/xfs_log_recover.c struct xlog *log) log 5821 fs/xfs/xfs_log_recover.c mp = log->l_mp; log 33 fs/xfs/xfs_rtalloc.c int log, /* log2 of extent size */ log 39 fs/xfs/xfs_rtalloc.c return xfs_rtmodify_summary_int(mp, tp, log, bbno, 0, rbpp, rsb, sum); log 58 fs/xfs/xfs_rtalloc.c int log; /* loop counter, log2 of ext. size */ log 68 fs/xfs/xfs_rtalloc.c for (log = low; log <= high; log++) { log 72 fs/xfs/xfs_rtalloc.c error = xfs_rtget_summary(mp, tp, log, bbno, rbpp, rsb, &sum); log 90 fs/xfs/xfs_rtalloc.c if (mp->m_rsum_cache && log > mp->m_rsum_cache[bbno]) log 91 fs/xfs/xfs_rtalloc.c mp->m_rsum_cache[bbno] = log; log 109 fs/xfs/xfs_rtalloc.c int log; /* summary level number (log length) */ log 114 fs/xfs/xfs_rtalloc.c for (log = omp->m_rsumlevels - 1; log >= 0; log--) { log 118 fs/xfs/xfs_rtalloc.c error = xfs_rtget_summary(omp, tp, log, bbno, &bp, log 124 fs/xfs/xfs_rtalloc.c error = xfs_rtmodify_summary(omp, tp, log, bbno, -sum, log 128 fs/xfs/xfs_rtalloc.c error = xfs_rtmodify_summary(nmp, tp, log, bbno, sum, log 117 fs/xfs/xfs_rtalloc.h int log, xfs_rtblock_t bbno, int delta, log 120 fs/xfs/xfs_rtalloc.h int xfs_rtmodify_summary(struct xfs_mount *mp, struct xfs_trans *tp, int log, log 322 fs/xfs/xfs_sysfs.c struct xlog *log = to_xlog(kobject); log 324 fs/xfs/xfs_sysfs.c spin_lock(&log->l_icloglock); log 325 fs/xfs/xfs_sysfs.c cycle = log->l_curr_cycle; log 326 fs/xfs/xfs_sysfs.c block = log->l_curr_block; log 327 fs/xfs/xfs_sysfs.c spin_unlock(&log->l_icloglock); log 340 fs/xfs/xfs_sysfs.c struct xlog *log = to_xlog(kobject); log 342 fs/xfs/xfs_sysfs.c xlog_crack_atomic_lsn(&log->l_tail_lsn, &cycle, &block); log 355 fs/xfs/xfs_sysfs.c struct xlog *log = to_xlog(kobject); log 357 fs/xfs/xfs_sysfs.c xlog_crack_grant_head(&log->l_reserve_head.grant, &cycle, &bytes); log 369 fs/xfs/xfs_sysfs.c struct xlog *log = to_xlog(kobject); log 371 fs/xfs/xfs_sysfs.c xlog_crack_grant_head(&log->l_write_head.grant, &cycle, &bytes); log 931 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xlog_ticket *tic), log 932 fs/xfs/xfs_trace.h TP_ARGS(log, tic), log 951 fs/xfs/xfs_trace.h __entry->dev = log->l_mp->m_super->s_dev; log 957 fs/xfs/xfs_trace.h __entry->reserveq = list_empty(&log->l_reserve_head.waiters); log 958 fs/xfs/xfs_trace.h __entry->writeq = list_empty(&log->l_write_head.waiters); log 959 fs/xfs/xfs_trace.h xlog_crack_grant_head(&log->l_reserve_head.grant, log 962 fs/xfs/xfs_trace.h xlog_crack_grant_head(&log->l_write_head.grant, log 965 fs/xfs/xfs_trace.h __entry->curr_cycle = log->l_curr_cycle; log 966 fs/xfs/xfs_trace.h __entry->curr_block = log->l_curr_block; log 967 fs/xfs/xfs_trace.h __entry->tail_lsn = atomic64_read(&log->l_tail_lsn); log 996 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xlog_ticket *tic), \ log 997 fs/xfs/xfs_trace.h TP_ARGS(log, tic)) log 1104 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, xfs_lsn_t new_lsn), log 1105 fs/xfs/xfs_trace.h TP_ARGS(log, new_lsn), log 1113 fs/xfs/xfs_trace.h __entry->dev = log->l_mp->m_super->s_dev; log 1115 fs/xfs/xfs_trace.h __entry->old_lsn = atomic64_read(&log->l_tail_lsn); log 1116 fs/xfs/xfs_trace.h __entry->last_sync_lsn = atomic64_read(&log->l_last_sync_lsn); log 1950 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, xfs_daddr_t headblk, xfs_daddr_t tailblk), log 1951 fs/xfs/xfs_trace.h TP_ARGS(log, headblk, tailblk), log 1958 fs/xfs/xfs_trace.h __entry->dev = log->l_mp->m_super->s_dev; log 1968 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xlog_rec_header *rhead, int pass), log 1969 fs/xfs/xfs_trace.h TP_ARGS(log, rhead, pass), log 1978 fs/xfs/xfs_trace.h __entry->dev = log->l_mp->m_super->s_dev; log 1991 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xlog_recover *trans, log 1993 fs/xfs/xfs_trace.h TP_ARGS(log, trans, item, pass), log 2005 fs/xfs/xfs_trace.h __entry->dev = log->l_mp->m_super->s_dev; log 2028 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xlog_recover *trans, \ log 2030 fs/xfs/xfs_trace.h TP_ARGS(log, trans, item, pass)) log 2039 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xfs_buf_log_format *buf_f), log 2040 fs/xfs/xfs_trace.h TP_ARGS(log, buf_f), log 2050 fs/xfs/xfs_trace.h __entry->dev = log->l_mp->m_super->s_dev; log 2069 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xfs_buf_log_format *buf_f), \ log 2070 fs/xfs/xfs_trace.h TP_ARGS(log, buf_f)) log 2083 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xfs_inode_log_format *in_f), log 2084 fs/xfs/xfs_trace.h TP_ARGS(log, in_f), log 2097 fs/xfs/xfs_trace.h __entry->dev = log->l_mp->m_super->s_dev; log 2121 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xfs_inode_log_format *in_f), \ log 2122 fs/xfs/xfs_trace.h TP_ARGS(log, in_f)) log 2129 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xfs_icreate_log *in_f), log 2130 fs/xfs/xfs_trace.h TP_ARGS(log, in_f), log 2141 fs/xfs/xfs_trace.h __entry->dev = log->l_mp->m_super->s_dev; log 2156 fs/xfs/xfs_trace.h TP_PROTO(struct xlog *log, struct xfs_icreate_log *in_f), \ log 2157 fs/xfs/xfs_trace.h TP_ARGS(log, in_f)) log 323 include/linux/bpf_verifier.h static inline bool bpf_verifier_log_full(const struct bpf_verifier_log *log) log 325 include/linux/bpf_verifier.h return log->len_used >= log->len_total - 1; log 334 include/linux/bpf_verifier.h static inline bool bpf_verifier_log_needed(const struct bpf_verifier_log *log) log 336 include/linux/bpf_verifier.h return log->level && log->ubuf && !bpf_verifier_log_full(log); log 369 include/linux/bpf_verifier.h struct bpf_verifier_log log; log 396 include/linux/bpf_verifier.h __printf(2, 0) void bpf_verifier_vlog(struct bpf_verifier_log *log, log 35 include/linux/dm-dirty-log.h int (*ctr)(struct dm_dirty_log *log, struct dm_target *ti, log 37 include/linux/dm-dirty-log.h void (*dtr)(struct dm_dirty_log *log); log 43 include/linux/dm-dirty-log.h int (*presuspend)(struct dm_dirty_log *log); log 44 include/linux/dm-dirty-log.h int (*postsuspend)(struct dm_dirty_log *log); log 45 include/linux/dm-dirty-log.h int (*resume)(struct dm_dirty_log *log); log 51 include/linux/dm-dirty-log.h uint32_t (*get_region_size)(struct dm_dirty_log *log); log 57 include/linux/dm-dirty-log.h int (*is_clean)(struct dm_dirty_log *log, region_t region); log 70 include/linux/dm-dirty-log.h int (*in_sync)(struct dm_dirty_log *log, region_t region, log 77 include/linux/dm-dirty-log.h int (*flush)(struct dm_dirty_log *log); log 85 include/linux/dm-dirty-log.h void (*mark_region)(struct dm_dirty_log *log, region_t region); log 86 include/linux/dm-dirty-log.h void (*clear_region)(struct dm_dirty_log *log, region_t region); log 100 include/linux/dm-dirty-log.h int (*get_resync_work)(struct dm_dirty_log *log, region_t *region); log 107 include/linux/dm-dirty-log.h void (*set_region_sync)(struct dm_dirty_log *log, log 113 include/linux/dm-dirty-log.h region_t (*get_sync_count)(struct dm_dirty_log *log); log 118 include/linux/dm-dirty-log.h int (*status)(struct dm_dirty_log *log, status_type_t status_type, log 129 include/linux/dm-dirty-log.h int (*is_remote_recovering)(struct dm_dirty_log *log, region_t region); log 143 include/linux/dm-dirty-log.h void dm_dirty_log_destroy(struct dm_dirty_log *log); log 41 include/linux/dm-region-hash.h struct dm_dirty_log *log, uint32_t region_size, log 1721 include/linux/efi.h u8 log[]; log 96 include/linux/fs_context.h struct fc_log *log; /* Logging buffer */ log 809 include/linux/kvm_host.h struct kvm_dirty_log *log, int *is_dirty); log 812 include/linux/kvm_host.h struct kvm_dirty_log *log, bool *flush); log 814 include/linux/kvm_host.h struct kvm_clear_dirty_log *log, bool *flush); log 822 include/linux/kvm_host.h struct kvm_dirty_log *log); log 824 include/linux/kvm_host.h struct kvm_clear_dirty_log *log); log 121 include/linux/tpm.h struct tpm_bios_log log; log 110 include/media/v4l2-ctrls.h void (*log)(const struct v4l2_ctrl *ctrl); log 35 include/net/netfilter/nf_log.h } log; log 75 include/net/udp.h unsigned int log; log 253 kernel/bpf/btf.c struct bpf_verifier_log log; log 664 kernel/bpf/btf.c __printf(2, 3) static void __btf_verifier_log(struct bpf_verifier_log *log, log 670 kernel/bpf/btf.c bpf_verifier_vlog(log, fmt, args); log 677 kernel/bpf/btf.c struct bpf_verifier_log *log = &env->log; log 680 kernel/bpf/btf.c if (!bpf_verifier_log_needed(log)) log 684 kernel/bpf/btf.c bpf_verifier_vlog(log, fmt, args); log 693 kernel/bpf/btf.c struct bpf_verifier_log *log = &env->log; log 698 kernel/bpf/btf.c if (!bpf_verifier_log_needed(log)) log 701 kernel/bpf/btf.c __btf_verifier_log(log, "[%u] %s %s%s", log 711 kernel/bpf/btf.c __btf_verifier_log(log, " "); log 713 kernel/bpf/btf.c bpf_verifier_vlog(log, fmt, args); log 717 kernel/bpf/btf.c __btf_verifier_log(log, "\n"); log 731 kernel/bpf/btf.c struct bpf_verifier_log *log = &env->log; log 735 kernel/bpf/btf.c if (!bpf_verifier_log_needed(log)) log 748 kernel/bpf/btf.c __btf_verifier_log(log, log 755 kernel/bpf/btf.c __btf_verifier_log(log, "\t%s type_id=%u bits_offset=%u", log 760 kernel/bpf/btf.c __btf_verifier_log(log, " "); log 762 kernel/bpf/btf.c bpf_verifier_vlog(log, fmt, args); log 766 kernel/bpf/btf.c __btf_verifier_log(log, "\n"); log 775 kernel/bpf/btf.c struct bpf_verifier_log *log = &env->log; log 778 kernel/bpf/btf.c if (!bpf_verifier_log_needed(log)) log 783 kernel/bpf/btf.c __btf_verifier_log(log, "\t type_id=%u offset=%u size=%u", log 786 kernel/bpf/btf.c __btf_verifier_log(log, " "); log 788 kernel/bpf/btf.c bpf_verifier_vlog(log, fmt, args); log 792 kernel/bpf/btf.c __btf_verifier_log(log, "\n"); log 798 kernel/bpf/btf.c struct bpf_verifier_log *log = &env->log; log 802 kernel/bpf/btf.c if (!bpf_verifier_log_needed(log)) log 806 kernel/bpf/btf.c __btf_verifier_log(log, "magic: 0x%x\n", hdr->magic); log 807 kernel/bpf/btf.c __btf_verifier_log(log, "version: %u\n", hdr->version); log 808 kernel/bpf/btf.c __btf_verifier_log(log, "flags: 0x%x\n", hdr->flags); log 809 kernel/bpf/btf.c __btf_verifier_log(log, "hdr_len: %u\n", hdr->hdr_len); log 810 kernel/bpf/btf.c __btf_verifier_log(log, "type_off: %u\n", hdr->type_off); log 811 kernel/bpf/btf.c __btf_verifier_log(log, "type_len: %u\n", hdr->type_len); log 812 kernel/bpf/btf.c __btf_verifier_log(log, "str_off: %u\n", hdr->str_off); log 813 kernel/bpf/btf.c __btf_verifier_log(log, "str_len: %u\n", hdr->str_len); log 814 kernel/bpf/btf.c __btf_verifier_log(log, "btf_total_size: %u\n", btf_data_size); log 3290 kernel/bpf/btf.c struct bpf_verifier_log *log; log 3302 kernel/bpf/btf.c log = &env->log; log 3307 kernel/bpf/btf.c log->level = log_level; log 3308 kernel/bpf/btf.c log->ubuf = log_ubuf; log 3309 kernel/bpf/btf.c log->len_total = log_size; log 3312 kernel/bpf/btf.c if (log->len_total < 128 || log->len_total > UINT_MAX >> 8 || log 3313 kernel/bpf/btf.c !log->level || !log->ubuf) { log 3354 kernel/bpf/btf.c if (log->level && bpf_verifier_log_full(log)) { log 232 kernel/bpf/verifier.c void bpf_verifier_vlog(struct bpf_verifier_log *log, const char *fmt, log 237 kernel/bpf/verifier.c n = vscnprintf(log->kbuf, BPF_VERIFIER_TMP_LOG_SIZE, fmt, args); log 242 kernel/bpf/verifier.c n = min(log->len_total - log->len_used - 1, n); log 243 kernel/bpf/verifier.c log->kbuf[n] = '\0'; log 245 kernel/bpf/verifier.c if (!copy_to_user(log->ubuf + log->len_used, log->kbuf, n + 1)) log 246 kernel/bpf/verifier.c log->len_used += n; log 248 kernel/bpf/verifier.c log->ubuf = NULL; log 260 kernel/bpf/verifier.c if (!bpf_verifier_log_needed(&env->log)) log 264 kernel/bpf/verifier.c bpf_verifier_vlog(&env->log, fmt, args); log 274 kernel/bpf/verifier.c if (!bpf_verifier_log_needed(&env->log)) log 278 kernel/bpf/verifier.c bpf_verifier_vlog(&env->log, fmt, args); log 296 kernel/bpf/verifier.c if (!bpf_verifier_log_needed(&env->log)) log 307 kernel/bpf/verifier.c bpf_verifier_vlog(&env->log, prefix_fmt, args); log 1153 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL2) log 1459 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) { log 1718 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) log 1805 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) { log 2198 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) log 3825 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) { log 3865 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) { log 6016 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) log 7264 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL2) log 7278 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL2) log 7604 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) { log 7622 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL2 || log 7623 kernel/bpf/verifier.c (env->log.level & BPF_LOG_LEVEL && do_print_state)) { log 7624 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL2) log 7635 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_LEVEL) { log 9273 kernel/bpf/verifier.c if (env->log.level & BPF_LOG_STATS) { log 9298 kernel/bpf/verifier.c struct bpf_verifier_log *log; log 9312 kernel/bpf/verifier.c log = &env->log; log 9334 kernel/bpf/verifier.c log->level = attr->log_level; log 9335 kernel/bpf/verifier.c log->ubuf = (char __user *) (unsigned long) attr->log_buf; log 9336 kernel/bpf/verifier.c log->len_total = attr->log_size; log 9340 kernel/bpf/verifier.c if (log->len_total < 128 || log->len_total > UINT_MAX >> 2 || log 9341 kernel/bpf/verifier.c !log->level || !log->ubuf || log->level & ~BPF_LOG_MASK) log 9436 kernel/bpf/verifier.c if (log->level && bpf_verifier_log_full(log)) log 9438 kernel/bpf/verifier.c if (log->level && !log->ubuf) { log 130 kernel/seccomp.c bool log; log 536 kernel/seccomp.c filter->log = true; log 637 kernel/seccomp.c bool log = false; log 643 kernel/seccomp.c log = requested && seccomp_actions_logged & SECCOMP_LOG_TRAP; log 646 kernel/seccomp.c log = requested && seccomp_actions_logged & SECCOMP_LOG_ERRNO; log 649 kernel/seccomp.c log = requested && seccomp_actions_logged & SECCOMP_LOG_TRACE; log 652 kernel/seccomp.c log = requested && seccomp_actions_logged & SECCOMP_LOG_USER_NOTIF; log 655 kernel/seccomp.c log = seccomp_actions_logged & SECCOMP_LOG_LOG; log 658 kernel/seccomp.c log = seccomp_actions_logged & SECCOMP_LOG_KILL_THREAD; log 662 kernel/seccomp.c log = seccomp_actions_logged & SECCOMP_LOG_KILL_PROCESS; log 671 kernel/seccomp.c if (!log) log 909 kernel/seccomp.c seccomp_log(this_syscall, 0, action, match ? match->log : false); log 1562 kernel/seccomp.c if (filter->log) log 1597 kernel/seccomp.c u32 log; log 1623 kernel/seccomp.c if (!(actions_logged & cur->log)) log 1654 kernel/seccomp.c *action_logged = cur->log; log 1215 mm/vmalloc.c unsigned int log; log 1217 mm/vmalloc.c log = fls(num_online_cpus()); log 1219 mm/vmalloc.c return log * (32UL * 1024 * 1024 / PAGE_SIZE); log 87 net/bridge/netfilter/ebt_log.c '0' + loginfo->u.log.level, prefix, log 93 net/bridge/netfilter/ebt_log.c bitmask = loginfo->u.log.logflags; log 186 net/bridge/netfilter/ebt_log.c li.u.log.level = info->loglevel; log 187 net/bridge/netfilter/ebt_log.c li.u.log.logflags = info->bitmask; log 149 net/ipv4/netfilter/ip_tables.c .log = { log 28 net/ipv4/netfilter/nf_log_arp.c .log = { log 27 net/ipv4/netfilter/nf_log_ipv4.c .log = { log 44 net/ipv4/netfilter/nf_log_ipv4.c logflags = info->u.log.logflags; log 280 net/ipv4/netfilter/nf_log_ipv4.c logflags = info->u.log.logflags; log 131 net/ipv4/udp.c struct sock *sk, unsigned int log) log 152 net/ipv4/udp.c __set_bit(udp_sk(sk2)->udp_port_hash >> log, log 255 net/ipv4/udp.c udptable->log); log 265 net/ipv4/udp.c !test_bit(snum >> udptable->log, bitmap) && log 3008 net/ipv4/udp.c &table->log, log 174 net/ipv6/netfilter/ip6_tables.c .log = { log 28 net/ipv6/netfilter/nf_log_ipv6.c .log = { log 50 net/ipv6/netfilter/nf_log_ipv6.c logflags = info->u.log.logflags; log 293 net/ipv6/netfilter/nf_log_ipv6.c logflags = info->u.log.logflags; log 30 net/netfilter/nf_log.c struct nf_logger *log; log 37 net/netfilter/nf_log.c log = nft_log_dereference(loggers[pf][i]); log 38 net/netfilter/nf_log.c if (!strncasecmp(str_logger, log->name, strlen(log->name))) log 39 net/netfilter/nf_log.c return log; log 47 net/netfilter/nf_log.c const struct nf_logger *log; log 53 net/netfilter/nf_log.c log = nft_log_dereference(net->nf.nf_loggers[pf]); log 54 net/netfilter/nf_log.c if (log == NULL) log 66 net/netfilter/nf_log.c const struct nf_logger *log; log 70 net/netfilter/nf_log.c log = nft_log_dereference(net->nf.nf_loggers[i]); log 71 net/netfilter/nf_log.c if (log == logger) log 114 net/netfilter/nf_log.c const struct nf_logger *log; log 119 net/netfilter/nf_log.c log = nft_log_dereference(loggers[i][logger->type]); log 120 net/netfilter/nf_log.c if (log == logger) log 160 net/netfilter/nf_log_common.c '0' + loginfo->u.log.level, prefix, log 101 net/netfilter/nfnetlink_log.c __instance_lookup(struct nfnl_log_net *log, u_int16_t group_num) log 106 net/netfilter/nfnetlink_log.c head = &log->instance_table[instance_hashfn(group_num)]; log 121 net/netfilter/nfnetlink_log.c instance_lookup_get(struct nfnl_log_net *log, u_int16_t group_num) log 126 net/netfilter/nfnetlink_log.c inst = __instance_lookup(log, group_num); log 158 net/netfilter/nfnetlink_log.c struct nfnl_log_net *log = nfnl_log_pernet(net); log 161 net/netfilter/nfnetlink_log.c spin_lock_bh(&log->instances_lock); log 162 net/netfilter/nfnetlink_log.c if (__instance_lookup(log, group_num)) { log 198 net/netfilter/nfnetlink_log.c &log->instance_table[instance_hashfn(group_num)]); log 201 net/netfilter/nfnetlink_log.c spin_unlock_bh(&log->instances_lock); log 206 net/netfilter/nfnetlink_log.c spin_unlock_bh(&log->instances_lock); log 235 net/netfilter/nfnetlink_log.c instance_destroy(struct nfnl_log_net *log, log 238 net/netfilter/nfnetlink_log.c spin_lock_bh(&log->instances_lock); log 240 net/netfilter/nfnetlink_log.c spin_unlock_bh(&log->instances_lock); log 442 net/netfilter/nfnetlink_log.c __build_packet_message(struct nfnl_log_net *log, log 627 net/netfilter/nfnetlink_log.c htonl(atomic_inc_return(&log->global_seq)))) log 689 net/netfilter/nfnetlink_log.c struct nfnl_log_net *log = nfnl_log_pernet(net); log 699 net/netfilter/nfnetlink_log.c inst = instance_lookup_get(log, li->u.ulog.group); log 793 net/netfilter/nfnetlink_log.c __build_packet_message(log, inst, skb, data_len, pf, log 822 net/netfilter/nfnetlink_log.c struct nfnl_log_net *log = nfnl_log_pernet(n->net); log 828 net/netfilter/nfnetlink_log.c spin_lock_bh(&log->instances_lock); log 832 net/netfilter/nfnetlink_log.c struct hlist_head *head = &log->instance_table[i]; log 839 net/netfilter/nfnetlink_log.c spin_unlock_bh(&log->instances_lock); log 881 net/netfilter/nfnetlink_log.c struct nfnl_log_net *log = nfnl_log_pernet(net); log 899 net/netfilter/nfnetlink_log.c inst = instance_lookup_get(log, group_num); log 949 net/netfilter/nfnetlink_log.c instance_destroy(log, inst); log 1018 net/netfilter/nfnetlink_log.c struct nfnl_log_net *log; log 1022 net/netfilter/nfnetlink_log.c log = nfnl_log_pernet(net); log 1025 net/netfilter/nfnetlink_log.c struct hlist_head *head = &log->instance_table[st->bucket]; log 1038 net/netfilter/nfnetlink_log.c struct nfnl_log_net *log; log 1044 net/netfilter/nfnetlink_log.c log = nfnl_log_pernet(net); log 1045 net/netfilter/nfnetlink_log.c head = &log->instance_table[st->bucket]; log 1106 net/netfilter/nfnetlink_log.c struct nfnl_log_net *log = nfnl_log_pernet(net); log 1114 net/netfilter/nfnetlink_log.c INIT_HLIST_HEAD(&log->instance_table[i]); log 1115 net/netfilter/nfnetlink_log.c spin_lock_init(&log->instances_lock); log 1133 net/netfilter/nfnetlink_log.c struct nfnl_log_net *log = nfnl_log_pernet(net); log 1141 net/netfilter/nfnetlink_log.c WARN_ON_ONCE(!hlist_empty(&log->instance_table[i])); log 111 net/netfilter/nft_log.c priv->loginfo.u.log.level == NFT_LOGLEVEL_AUDIT) { log 163 net/netfilter/nft_log.c li->u.log.level = log 166 net/netfilter/nft_log.c li->u.log.level = NFT_LOGLEVEL_WARNING; log 168 net/netfilter/nft_log.c if (li->u.log.level > NFT_LOGLEVEL_AUDIT) { log 174 net/netfilter/nft_log.c li->u.log.logflags = log 176 net/netfilter/nft_log.c if (li->u.log.logflags & ~NF_LOG_MASK) { log 196 net/netfilter/nft_log.c if (li->u.log.level == NFT_LOGLEVEL_AUDIT) log 220 net/netfilter/nft_log.c if (li->u.log.level == NFT_LOGLEVEL_AUDIT) log 236 net/netfilter/nft_log.c if (nla_put_be32(skb, NFTA_LOG_LEVEL, htonl(li->u.log.level))) log 239 net/netfilter/nft_log.c if (li->u.log.logflags) { log 241 net/netfilter/nft_log.c htonl(li->u.log.logflags))) log 36 net/netfilter/xt_LOG.c li.u.log.level = loginfo->level; log 37 net/netfilter/xt_LOG.c li.u.log.logflags = loginfo->logflags; log 1323 net/openvswitch/conntrack.c const struct sw_flow_key *key, bool log) log 1332 net/openvswitch/conntrack.c OVS_NLERR(log, "Unknown helper \"%s\"", name); log 1348 net/openvswitch/conntrack.c OVS_NLERR(log, "Failed to load \"%s\" NAT helper, error: %d", log 1361 net/openvswitch/conntrack.c struct ovs_conntrack_info *info, bool log) log 1386 net/openvswitch/conntrack.c OVS_NLERR(log, "Unknown NAT attribute (type=%d, max=%d)", log 1392 net/openvswitch/conntrack.c OVS_NLERR(log, "NAT attribute type %d has unexpected length (%d != %d)", log 1402 net/openvswitch/conntrack.c OVS_NLERR(log, "Only one type of NAT may be specified"); log 1447 net/openvswitch/conntrack.c OVS_NLERR(log, "Unknown nat attribute (%d)", type); log 1453 net/openvswitch/conntrack.c OVS_NLERR(log, "NAT attribute has %d unknown bytes", rem); log 1459 net/openvswitch/conntrack.c OVS_NLERR(log, log 1466 net/openvswitch/conntrack.c OVS_NLERR(log, log 1507 net/openvswitch/conntrack.c const char **helper, bool log) log 1518 net/openvswitch/conntrack.c OVS_NLERR(log, log 1527 net/openvswitch/conntrack.c OVS_NLERR(log, log 1550 net/openvswitch/conntrack.c OVS_NLERR(log, "ct_mark mask cannot be 0"); log 1562 net/openvswitch/conntrack.c OVS_NLERR(log, "ct_labels mask cannot be 0"); log 1572 net/openvswitch/conntrack.c OVS_NLERR(log, "Invalid conntrack helper"); log 1578 net/openvswitch/conntrack.c int err = parse_nat(a, info, log); log 1593 net/openvswitch/conntrack.c OVS_NLERR(log, "Invalid conntrack timeout"); log 1600 net/openvswitch/conntrack.c OVS_NLERR(log, "Unknown conntrack attr (%d)", log 1608 net/openvswitch/conntrack.c OVS_NLERR(log, log 1615 net/openvswitch/conntrack.c OVS_NLERR(log, log 1621 net/openvswitch/conntrack.c OVS_NLERR(log, "Conntrack attr has %d unknown bytes", rem); log 1650 net/openvswitch/conntrack.c struct sw_flow_actions **sfa, bool log) log 1659 net/openvswitch/conntrack.c OVS_NLERR(log, "ct family unspecified"); log 1669 net/openvswitch/conntrack.c err = parse_ct(attr, &ct_info, &helper, log); log 1676 net/openvswitch/conntrack.c OVS_NLERR(log, "Failed to allocate conntrack template"); log 1692 net/openvswitch/conntrack.c err = ovs_ct_add_helper(&ct_info, helper, key, log); log 1698 net/openvswitch/conntrack.c sizeof(ct_info), log); log 21 net/openvswitch/conntrack.h bool log); log 51 net/openvswitch/conntrack.h struct sw_flow_actions **acts, bool log) log 547 net/openvswitch/datapath.c bool log = !a[OVS_PACKET_ATTR_PROBE]; log 577 net/openvswitch/datapath.c packet, &flow->key, log); log 582 net/openvswitch/datapath.c &flow->key, &acts, log); log 909 net/openvswitch/datapath.c bool log = !a[OVS_FLOW_ATTR_PROBE]; log 914 net/openvswitch/datapath.c OVS_NLERR(log, "Flow key attr not present in new flow."); log 918 net/openvswitch/datapath.c OVS_NLERR(log, "Flow actions attr not present in new flow."); log 934 net/openvswitch/datapath.c a[OVS_FLOW_ATTR_MASK], log); log 940 net/openvswitch/datapath.c &new_flow->key, log); log 952 net/openvswitch/datapath.c &new_flow->key, &acts, log); log 954 net/openvswitch/datapath.c OVS_NLERR(log, "Flow actions may not be safe on all matching packets."); log 1064 net/openvswitch/datapath.c bool log) log 1071 net/openvswitch/datapath.c error = ovs_nla_copy_actions(net, a, &masked_key, &acts, log); log 1073 net/openvswitch/datapath.c OVS_NLERR(log, log 1100 net/openvswitch/datapath.c bool log) log 1108 net/openvswitch/datapath.c a[OVS_FLOW_ATTR_MASK], log); log 1115 net/openvswitch/datapath.c OVS_NLERR(log, log 1122 net/openvswitch/datapath.c &mask, log); log 1149 net/openvswitch/datapath.c bool log = !a[OVS_FLOW_ATTR_PROBE]; log 1152 net/openvswitch/datapath.c ufid_present = ovs_nla_get_ufid(&sfid, a[OVS_FLOW_ATTR_UFID], log); log 1154 net/openvswitch/datapath.c OVS_NLERR(log, log 1160 net/openvswitch/datapath.c &acts, log); log 1250 net/openvswitch/datapath.c bool log = !a[OVS_FLOW_ATTR_PROBE]; log 1253 net/openvswitch/datapath.c ufid_present = ovs_nla_get_ufid(&ufid, a[OVS_FLOW_ATTR_UFID], log); log 1257 net/openvswitch/datapath.c log); log 1259 net/openvswitch/datapath.c OVS_NLERR(log, log 1309 net/openvswitch/datapath.c bool log = !a[OVS_FLOW_ATTR_PROBE]; log 1312 net/openvswitch/datapath.c ufid_present = ovs_nla_get_ufid(&ufid, a[OVS_FLOW_ATTR_UFID], log); log 1316 net/openvswitch/datapath.c NULL, log); log 900 net/openvswitch/flow.c struct sw_flow_key *key, bool log) log 906 net/openvswitch/flow.c err = parse_flow_nlattrs(attr, a, &attrs, log); log 911 net/openvswitch/flow.c err = ovs_nla_get_flow_metadata(net, a, attrs, key, log); log 280 net/openvswitch/flow.h struct sw_flow_key *key, bool log); log 151 net/openvswitch/flow_netlink.c u64 key_attrs, u64 mask_attrs, bool log) log 290 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Missing key (keys=%llx, expected=%llx)", log 298 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Unexpected mask (mask=%llx, allowed=%llx)", log 463 net/openvswitch/flow_netlink.c u64 *attrsp, bool log, bool nz) log 475 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Key type %d is out of range max %d", log 481 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Duplicate key (type %d).", type); log 487 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Key %d has unexpected len %d expected %d", log 498 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Message has %d unknown bytes.", rem); log 508 net/openvswitch/flow_netlink.c bool log) log 510 net/openvswitch/flow_netlink.c return __parse_flow_nlattrs(attr, a, attrsp, log, true); log 514 net/openvswitch/flow_netlink.c u64 *attrsp, bool log) log 516 net/openvswitch/flow_netlink.c return __parse_flow_nlattrs(attr, a, attrsp, log, false); log 521 net/openvswitch/flow_netlink.c bool log) log 526 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Geneve option length err (len %d, max %zu).", log 532 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Geneve opt len %d is not a multiple of 4.", log 556 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Geneve option len %d != mask len %d", log 572 net/openvswitch/flow_netlink.c bool log) log 586 net/openvswitch/flow_netlink.c OVS_NLERR(log, "VXLAN extension %d out of range max %d", log 593 net/openvswitch/flow_netlink.c OVS_NLERR(log, "VXLAN extension %d has unexpected len %d expected %d", log 604 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Unknown VXLAN extension attribute %d", log 610 net/openvswitch/flow_netlink.c OVS_NLERR(log, "VXLAN extension message has %d unknown bytes.", log 628 net/openvswitch/flow_netlink.c bool log) log 636 net/openvswitch/flow_netlink.c OVS_NLERR(log, "ERSPAN option length err (len %d, max %zu).", log 655 net/openvswitch/flow_netlink.c bool log) log 669 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Tunnel attr %d out of range max %d", log 676 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Tunnel attr %d has unexpected len %d expected %d", log 735 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Multiple metadata blocks provided"); log 739 net/openvswitch/flow_netlink.c err = genev_tun_opt_from_nlattr(a, match, is_mask, log); log 748 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Multiple metadata blocks provided"); log 752 net/openvswitch/flow_netlink.c err = vxlan_tun_opt_from_nlattr(a, match, is_mask, log); log 763 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Multiple metadata blocks provided"); log 768 net/openvswitch/flow_netlink.c log); log 780 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Unknown IP tunnel attribute %d", log 794 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IP tunnel attribute has %d unknown bytes.", log 800 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Mixed IPv4 and IPv6 tunnel attributes"); log 806 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IP tunnel dst address not specified"); log 818 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IPv4 tun info is not correct"); log 822 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IPv4 tunnel dst address is zero"); log 827 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IPv6 tunnel dst address is zero"); log 832 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IP tunnel TTL not specified."); log 985 net/openvswitch/flow_netlink.c const struct nlattr **a, bool log) log 998 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Invalid %s frame", (inner) ? "C-VLAN" : "VLAN"); log 1007 net/openvswitch/flow_netlink.c OVS_NLERR(log, "%s TCI does not have VLAN_CFI_MASK bit set.", log 1012 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Truncated %s header has non-zero encap attribute.", log 1023 net/openvswitch/flow_netlink.c const struct nlattr **a, bool log) log 1038 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Encap mask attribute is set for non-%s frame.", log 1050 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Must have an exact match on %s TPID (mask=%x).", log 1055 net/openvswitch/flow_netlink.c OVS_NLERR(log, "%s TCI mask does not have exact match for VLAN_CFI_MASK bit.", log 1066 net/openvswitch/flow_netlink.c bool log) log 1073 net/openvswitch/flow_netlink.c a, log); log 1076 net/openvswitch/flow_netlink.c a, log); log 1091 net/openvswitch/flow_netlink.c err = parse_flow_nlattrs(encap, a, key_attrs, log); log 1093 net/openvswitch/flow_netlink.c err = parse_flow_mask_nlattrs(encap, a, key_attrs, log); log 1100 net/openvswitch/flow_netlink.c bool is_mask, bool log) log 1106 net/openvswitch/flow_netlink.c is_mask, log); log 1113 net/openvswitch/flow_netlink.c is_mask, log); log 1123 net/openvswitch/flow_netlink.c bool is_mask, bool log) log 1132 net/openvswitch/flow_netlink.c OVS_NLERR(log, "EtherType %x is less than min %x", log 1144 net/openvswitch/flow_netlink.c bool is_mask, bool log) log 1174 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Port %d exceeds max allowable %d", log 1193 net/openvswitch/flow_netlink.c is_mask, log) < 0) log 1203 net/openvswitch/flow_netlink.c OVS_NLERR(log, "ct_state flags %08x unsupported", log 1275 net/openvswitch/flow_netlink.c log); log 1376 net/openvswitch/flow_netlink.c bool is_push_nsh, bool log) log 1394 net/openvswitch/flow_netlink.c OVS_NLERR(log, "nsh attr %d is out of range max %d", log 1402 net/openvswitch/flow_netlink.c log, log 1446 net/openvswitch/flow_netlink.c log, log 1455 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Unknown nsh attribute %d", log 1462 net/openvswitch/flow_netlink.c OVS_NLERR(log, "nsh attribute has %d unknown bytes.", rem); log 1497 net/openvswitch/flow_netlink.c bool is_mask, bool log) log 1501 net/openvswitch/flow_netlink.c err = metadata_from_nlattrs(net, match, &attrs, a, is_mask, log); log 1519 net/openvswitch/flow_netlink.c OVS_NLERR(log, "VLAN attribute unexpected."); log 1525 net/openvswitch/flow_netlink.c log); log 1532 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Either Ethernet header or EtherType is required."); log 1541 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IPv4 frag type %d is out of range max %d", log 1565 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IPv6 frag type %d is out of range max %d", log 1571 net/openvswitch/flow_netlink.c OVS_NLERR(log, "IPv6 flow label %x is out of range (max=%x)", log 1603 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Unknown ARP opcode (opcode=%d).", log 1624 net/openvswitch/flow_netlink.c is_mask, false, log) < 0) log 1711 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Unknown key attributes %llx", log 1761 net/openvswitch/flow_netlink.c bool log) log 1769 net/openvswitch/flow_netlink.c err = parse_flow_nlattrs(nla_key, a, &key_attrs, log); log 1773 net/openvswitch/flow_netlink.c err = parse_vlan_from_nlattrs(match, &key_attrs, a, false, log); log 1777 net/openvswitch/flow_netlink.c err = ovs_key_from_nlattrs(net, match, key_attrs, a, false, log); log 1811 net/openvswitch/flow_netlink.c err = parse_flow_mask_nlattrs(nla_mask, a, &mask_attrs, log); log 1819 net/openvswitch/flow_netlink.c err = parse_vlan_from_nlattrs(match, &mask_attrs, a, true, log); log 1824 net/openvswitch/flow_netlink.c log); log 1829 net/openvswitch/flow_netlink.c if (!match_validate(match, key_attrs, mask_attrs, log)) log 1837 net/openvswitch/flow_netlink.c static size_t get_ufid_len(const struct nlattr *attr, bool log) log 1846 net/openvswitch/flow_netlink.c OVS_NLERR(log, "ufid size %u bytes exceeds the range (1, %d)", log 1858 net/openvswitch/flow_netlink.c bool log) log 1860 net/openvswitch/flow_netlink.c sfid->ufid_len = get_ufid_len(attr, log); log 1868 net/openvswitch/flow_netlink.c const struct sw_flow_key *key, bool log) log 1872 net/openvswitch/flow_netlink.c if (ovs_nla_get_ufid(sfid, ufid, log)) log 1912 net/openvswitch/flow_netlink.c u64 attrs, struct sw_flow_key *key, bool log) log 1928 net/openvswitch/flow_netlink.c return metadata_from_nlattrs(net, &match, &attrs, a, false, log); log 2317 net/openvswitch/flow_netlink.c int attr_len, bool log) log 2333 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Flow action size exceeds max %u", log 2356 net/openvswitch/flow_netlink.c int attrtype, void *data, int len, bool log) log 2360 net/openvswitch/flow_netlink.c a = reserve_sfa_size(sfa, nla_attr_size(len), log); log 2375 net/openvswitch/flow_netlink.c int len, bool log) log 2379 net/openvswitch/flow_netlink.c a = __add_action(sfa, attrtype, data, len, log); log 2385 net/openvswitch/flow_netlink.c int attrtype, bool log) log 2390 net/openvswitch/flow_netlink.c err = ovs_nla_add_action(sfa, attrtype, NULL, 0, log); log 2409 net/openvswitch/flow_netlink.c __be16 eth_type, __be16 vlan_tci, bool log); log 2415 net/openvswitch/flow_netlink.c bool log, bool last) log 2442 net/openvswitch/flow_netlink.c start = add_nested_action_start(sfa, OVS_ACTION_ATTR_SAMPLE, log); log 2461 net/openvswitch/flow_netlink.c log); log 2466 net/openvswitch/flow_netlink.c eth_type, vlan_tci, log); log 2481 net/openvswitch/flow_netlink.c bool log, bool last) log 2489 net/openvswitch/flow_netlink.c start = add_nested_action_start(sfa, OVS_ACTION_ATTR_CLONE, log); log 2496 net/openvswitch/flow_netlink.c sizeof(exec), log); log 2501 net/openvswitch/flow_netlink.c eth_type, vlan_tci, log); log 2557 net/openvswitch/flow_netlink.c struct sw_flow_actions **sfa, bool log) log 2570 net/openvswitch/flow_netlink.c opts_type = ip_tun_from_nlattr(nla_data(attr), &match, false, log); log 2591 net/openvswitch/flow_netlink.c start = add_nested_action_start(sfa, OVS_ACTION_ATTR_SET, log); log 2608 net/openvswitch/flow_netlink.c sizeof(*ovs_tun), log); log 2638 net/openvswitch/flow_netlink.c bool is_push_nsh, bool log) log 2646 net/openvswitch/flow_netlink.c is_push_nsh, log); log 2667 net/openvswitch/flow_netlink.c u8 mac_proto, __be16 eth_type, bool masked, bool log) log 2709 net/openvswitch/flow_netlink.c err = validate_and_copy_set_tun(a, sfa, log); log 2795 net/openvswitch/flow_netlink.c if (!validate_nsh(nla_data(a), masked, false, log)) log 2812 net/openvswitch/flow_netlink.c log); log 2816 net/openvswitch/flow_netlink.c at = __add_action(sfa, key_type, NULL, len, log); log 2867 net/openvswitch/flow_netlink.c bool log, bool last) log 2894 net/openvswitch/flow_netlink.c log); log 2905 net/openvswitch/flow_netlink.c sizeof(arg), log); log 2910 net/openvswitch/flow_netlink.c OVS_CHECK_PKT_LEN_ATTR_ACTIONS_IF_LESS_EQUAL, log); log 2915 net/openvswitch/flow_netlink.c eth_type, vlan_tci, log); log 2923 net/openvswitch/flow_netlink.c OVS_CHECK_PKT_LEN_ATTR_ACTIONS_IF_GREATER, log); log 2928 net/openvswitch/flow_netlink.c eth_type, vlan_tci, log); log 2939 net/openvswitch/flow_netlink.c struct sw_flow_actions **sfa, bool log) log 2944 net/openvswitch/flow_netlink.c to = reserve_sfa_size(sfa, from->nla_len, log); log 2955 net/openvswitch/flow_netlink.c __be16 eth_type, __be16 vlan_tci, bool log) log 3091 net/openvswitch/flow_netlink.c false, log); log 3099 net/openvswitch/flow_netlink.c true, log); log 3109 net/openvswitch/flow_netlink.c log, last); log 3117 net/openvswitch/flow_netlink.c err = ovs_ct_copy_action(net, a, key, sfa, log); log 3179 net/openvswitch/flow_netlink.c log, last); log 3191 net/openvswitch/flow_netlink.c vlan_tci, log, log 3200 net/openvswitch/flow_netlink.c OVS_NLERR(log, "Unknown Action type %d", type); log 3204 net/openvswitch/flow_netlink.c err = copy_action(a, sfa, log); log 3219 net/openvswitch/flow_netlink.c struct sw_flow_actions **sfa, bool log) log 3229 net/openvswitch/flow_netlink.c key->eth.vlan.tci, log); log 36 net/openvswitch/flow_netlink.h u64 *attrsp, bool log); log 39 net/openvswitch/flow_netlink.h u64 attrs, struct sw_flow_key *key, bool log); log 47 net/openvswitch/flow_netlink.h bool log); log 52 net/openvswitch/flow_netlink.h bool ovs_nla_get_ufid(struct sw_flow_id *, const struct nlattr *, bool log); log 54 net/openvswitch/flow_netlink.h const struct sw_flow_key *key, bool log); log 59 net/openvswitch/flow_netlink.h struct sw_flow_actions **sfa, bool log); log 61 net/openvswitch/flow_netlink.h void *data, int len, bool log); log 703 net/sched/sch_fq.c static int fq_resize(struct Qdisc *sch, u32 log) log 710 net/sched/sch_fq.c if (q->fq_root && log == q->fq_trees_log) log 714 net/sched/sch_fq.c array = kvmalloc_node(sizeof(struct rb_root) << log, GFP_KERNEL | __GFP_RETRY_MAYFAIL, log 719 net/sched/sch_fq.c for (idx = 0; idx < (1U << log); idx++) log 726 net/sched/sch_fq.c fq_rehash(q, old_fq_root, q->fq_trees_log, array, log); log 729 net/sched/sch_fq.c q->fq_trees_log = log; log 304 security/tomoyo/audit.c char *log; log 383 security/tomoyo/audit.c entry->log = buf; log 456 security/tomoyo/audit.c head->read_buf = ptr->log; log 1085 tools/perf/util/auxtrace.c synth_opts->log = true; log 98 tools/perf/util/auxtrace.h bool log; log 3244 tools/perf/util/intel-pt.c if (pt->synth_opts.log) log 1126 tools/perf/util/s390-cpumsf.c sf->use_logfile = session->itrace_synth_opts->log; log 322 tools/power/acpi/tools/acpidbg/acpidbg.c ACPI_AML_BATCH_DO(fd, read, log, ret); log 324 tools/power/acpi/tools/acpidbg/acpidbg.c ACPI_AML_DO(fd, read, log, ret); log 328 tools/power/acpi/tools/acpidbg/acpidbg.c ACPI_AML_BATCH_DO(STDOUT_FILENO, write, log, ret); log 330 tools/power/acpi/tools/acpidbg/acpidbg.c ACPI_AML_DO(STDOUT_FILENO, write, log, ret); log 864 tools/testing/selftests/bpf/test_verifier.c static bool cmp_str_seq(const char *log, const char *exp) log 882 tools/testing/selftests/bpf/test_verifier.c q = strstr(log, needle); log 888 tools/testing/selftests/bpf/test_verifier.c log = q + len; log 48 tools/testing/selftests/bpf/test_verifier_log.c static int load(char *log, size_t log_len, int log_level) log 57 tools/testing/selftests/bpf/test_verifier_log.c attr.log_buf = ptr_to_u64(log); log 88 tools/testing/selftests/bpf/test_verifier_log.c static void test_log_good(char *log, size_t buf_len, size_t log_len, log 94 tools/testing/selftests/bpf/test_verifier_log.c memset(log, 1, buf_len); log 96 tools/testing/selftests/bpf/test_verifier_log.c ret = load(log, log_len, 1); log 99 tools/testing/selftests/bpf/test_verifier_log.c len = strnlen(log, buf_len); log 110 tools/testing/selftests/bpf/test_verifier_log.c if (strchr(log, 1)) { log 115 tools/testing/selftests/bpf/test_verifier_log.c check_ones(log + len + 1, buf_len - len - 1, log 118 tools/testing/selftests/bpf/test_verifier_log.c if (memcmp(full_log, log, LOG_SIZE)) { log 124 tools/testing/selftests/bpf/test_verifier_log.c static void test_log_bad(char *log, size_t log_len, int log_level) log 128 tools/testing/selftests/bpf/test_verifier_log.c ret = load(log, log_len, log_level); log 130 tools/testing/selftests/bpf/test_verifier_log.c if (log) log 131 tools/testing/selftests/bpf/test_verifier_log.c check_ones(log, LOG_SIZE, log 138 tools/testing/selftests/bpf/test_verifier_log.c char log[LOG_SIZE]; log 142 tools/testing/selftests/bpf/test_verifier_log.c memset(log, 1, LOG_SIZE); log 146 tools/testing/selftests/bpf/test_verifier_log.c test_log_bad(log, LOG_SIZE, 0); log 149 tools/testing/selftests/bpf/test_verifier_log.c test_log_bad(log, 15, 1); log 161 tools/testing/selftests/bpf/test_verifier_log.c test_log_good(log, LOG_SIZE, want_len + 2, want_len, EACCES, full_log); log 168 tools/testing/selftests/bpf/test_verifier_log.c test_log_good(log, LOG_SIZE, want_len + 1 - i, want_len - i, log 76 tools/testing/selftests/kvm/include/kvm_util.h void kvm_vm_get_dirty_log(struct kvm_vm *vm, int slot, void *log); log 77 tools/testing/selftests/kvm/include/kvm_util.h void kvm_vm_clear_dirty_log(struct kvm_vm *vm, int slot, void *log, log 277 tools/testing/selftests/kvm/lib/kvm_util.c void kvm_vm_get_dirty_log(struct kvm_vm *vm, int slot, void *log) log 279 tools/testing/selftests/kvm/lib/kvm_util.c struct kvm_dirty_log args = { .dirty_bitmap = log, .slot = slot }; log 287 tools/testing/selftests/kvm/lib/kvm_util.c void kvm_vm_clear_dirty_log(struct kvm_vm *vm, int slot, void *log, log 290 tools/testing/selftests/kvm/lib/kvm_util.c struct kvm_clear_dirty_log args = { .dirty_bitmap = log, .slot = slot, log 1018 tools/testing/selftests/seccomp/seccomp_bpf.c struct sock_fprog log; log 1073 tools/testing/selftests/seccomp/seccomp_bpf.c FILTER_ALLOC(log); log 1084 tools/testing/selftests/seccomp/seccomp_bpf.c FILTER_FREE(log); log 1102 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1128 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1161 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1184 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1211 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1234 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1254 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1278 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1301 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1321 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1339 tools/testing/selftests/seccomp/seccomp_bpf.c ret = prctl(PR_SET_SECCOMP, SECCOMP_MODE_FILTER, &self->log); log 1245 virt/kvm/arm/arm.c int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log) log 1252 virt/kvm/arm/arm.c r = kvm_get_dirty_log_protect(kvm, log, &flush); log 1261 virt/kvm/arm/arm.c int kvm_vm_ioctl_clear_dirty_log(struct kvm *kvm, struct kvm_clear_dirty_log *log) log 1268 virt/kvm/arm/arm.c r = kvm_clear_dirty_log_protect(kvm, log, &flush); log 1176 virt/kvm/kvm_main.c struct kvm_dirty_log *log, int *is_dirty) log 1184 virt/kvm/kvm_main.c as_id = log->slot >> 16; log 1185 virt/kvm/kvm_main.c id = (u16)log->slot; log 1199 virt/kvm/kvm_main.c if (copy_to_user(log->dirty_bitmap, memslot->dirty_bitmap, n)) log 1232 virt/kvm/kvm_main.c struct kvm_dirty_log *log, bool *flush) log 1241 virt/kvm/kvm_main.c as_id = log->slot >> 16; log 1242 virt/kvm/kvm_main.c id = (u16)log->slot; log 1288 virt/kvm/kvm_main.c if (copy_to_user(log->dirty_bitmap, dirty_bitmap_buffer, n)) log 1302 virt/kvm/kvm_main.c struct kvm_clear_dirty_log *log, bool *flush) log 1312 virt/kvm/kvm_main.c as_id = log->slot >> 16; log 1313 virt/kvm/kvm_main.c id = (u16)log->slot; log 1317 virt/kvm/kvm_main.c if (log->first_page & 63) log 1327 virt/kvm/kvm_main.c n = ALIGN(log->num_pages, BITS_PER_LONG) / 8; log 1329 virt/kvm/kvm_main.c if (log->first_page > memslot->npages || log 1330 virt/kvm/kvm_main.c log->num_pages > memslot->npages - log->first_page || log 1331 virt/kvm/kvm_main.c (log->num_pages < memslot->npages - log->first_page && (log->num_pages & 63))) log 1336 virt/kvm/kvm_main.c if (copy_from_user(dirty_bitmap_buffer, log->dirty_bitmap, n)) log 1340 virt/kvm/kvm_main.c for (offset = log->first_page, i = offset / BITS_PER_LONG, log 1341 virt/kvm/kvm_main.c n = DIV_ROUND_UP(log->num_pages, BITS_PER_LONG); n--; log 3382 virt/kvm/kvm_main.c struct kvm_dirty_log log; log 3385 virt/kvm/kvm_main.c if (copy_from_user(&log, argp, sizeof(log))) log 3387 virt/kvm/kvm_main.c r = kvm_vm_ioctl_get_dirty_log(kvm, &log); log 3392 virt/kvm/kvm_main.c struct kvm_clear_dirty_log log; log 3395 virt/kvm/kvm_main.c if (copy_from_user(&log, argp, sizeof(log))) log 3397 virt/kvm/kvm_main.c r = kvm_vm_ioctl_clear_dirty_log(kvm, &log); log 3558 virt/kvm/kvm_main.c struct kvm_dirty_log log; log 3563 virt/kvm/kvm_main.c log.slot = compat_log.slot; log 3564 virt/kvm/kvm_main.c log.padding1 = compat_log.padding1; log 3565 virt/kvm/kvm_main.c log.padding2 = compat_log.padding2; log 3566 virt/kvm/kvm_main.c log.dirty_bitmap = compat_ptr(compat_log.dirty_bitmap); log 3568 virt/kvm/kvm_main.c r = kvm_vm_ioctl_get_dirty_log(kvm, &log);