oe 708 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t oe:6; oe 714 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t oe:6; oe 739 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t oe:6; oe 745 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t oe:6; oe 3717 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t oe:1; oe 3721 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t oe:1; oe 4132 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t oe:1; oe 4136 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t oe:1; oe 2826 arch/powerpc/xmon/ppc-opc.c #define XO(op, xop, oe, rc) \ oe 2827 arch/powerpc/xmon/ppc-opc.c (OP (op) | ((((unsigned long)(xop)) & 0x1ff) << 1) | ((((unsigned long)(oe)) & 1) << 10) | (((unsigned long)(rc)) & 1)) oe 200 drivers/ata/pata_octeon_cf.c reg_tim.s.oe = t2; oe 36 drivers/gpio/gpio-omap.c u32 oe; oe 109 drivers/gpio/gpio-omap.c bank->context.oe = omap_gpio_rmw(bank->base + bank->regs->direction, oe 969 drivers/gpio/gpio-omap.c bank->context.oe = readl_relaxed(bank->base + bank->regs->direction); oe 1062 drivers/gpio/gpio-omap.c p->context.oe = readl_relaxed(base + regs->direction); oe 1087 drivers/gpio/gpio-omap.c writel_relaxed(bank->context.oe, base + regs->direction); oe 69 drivers/gpio/gpio-tegra.c u32 oe[4]; oe 210 drivers/gpio/gpio-tegra.c u32 cnf, oe; oe 216 drivers/gpio/gpio-tegra.c oe = tegra_gpio_readl(tgi, GPIO_OE(tgi, offset)); oe 218 drivers/gpio/gpio-tegra.c return !(oe & pin_mask); oe 425 drivers/gpio/gpio-tegra.c for (p = 0; p < ARRAY_SIZE(bank->oe); p++) { oe 440 drivers/gpio/gpio-tegra.c tegra_gpio_writel(tgi, bank->oe[p], oe 463 drivers/gpio/gpio-tegra.c for (p = 0; p < ARRAY_SIZE(bank->oe); p++) { oe 470 drivers/gpio/gpio-tegra.c bank->oe[p] = tegra_gpio_readl(tgi, oe 32 drivers/gpu/drm/bridge/thc63lvd1024.c struct gpio_desc *oe; oe 95 drivers/gpu/drm/bridge/thc63lvd1024.c gpiod_set_value(thc63->oe, 1); oe 103 drivers/gpu/drm/bridge/thc63lvd1024.c gpiod_set_value(thc63->oe, 0); oe 173 drivers/gpu/drm/bridge/thc63lvd1024.c thc63->oe = devm_gpiod_get_optional(thc63->dev, "oe", GPIOD_OUT_LOW); oe 174 drivers/gpu/drm/bridge/thc63lvd1024.c if (IS_ERR(thc63->oe)) { oe 176 drivers/gpu/drm/bridge/thc63lvd1024.c PTR_ERR(thc63->oe)); oe 177 drivers/gpu/drm/bridge/thc63lvd1024.c return PTR_ERR(thc63->oe); oe 153 drivers/media/i2c/mt9v111.c struct gpio_desc *oe; oe 375 drivers/media/i2c/mt9v111.c gpiod_set_value(mt9v111->oe, 1); oe 385 drivers/media/i2c/mt9v111.c gpiod_set_value(mt9v111->oe, 0); oe 1128 drivers/media/i2c/mt9v111.c mt9v111->oe = devm_gpiod_get_optional(&client->dev, "enable", oe 1130 drivers/media/i2c/mt9v111.c if (IS_ERR(mt9v111->oe)) { oe 1132 drivers/media/i2c/mt9v111.c PTR_ERR(mt9v111->oe)); oe 1133 drivers/media/i2c/mt9v111.c return PTR_ERR(mt9v111->oe); oe 1256 drivers/media/i2c/mt9v111.c devm_gpiod_put(mt9v111->dev, mt9v111->oe); oe 68 drivers/media/i2c/tvp5150.c u32 oe; oe 859 drivers/media/i2c/tvp5150.c decoder->lock ? decoder->oe : 0); oe 935 drivers/media/i2c/tvp5150.c decoder->oe = TVP5150_MISC_CTL_YCBCR_OE | oe 940 drivers/media/i2c/tvp5150.c decoder->oe = TVP5150_MISC_CTL_YCBCR_OE | oe 1223 drivers/media/i2c/tvp5150.c val = decoder->lock ? decoder->oe : 0; oe 1225 drivers/media/i2c/tvp5150.c val = decoder->oe; oe 1724 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c u32 ie, oe, pu, pd; oe 1743 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c oe = ioread32(bank->base + NPCM7XX_GP_N_OE) & pinmask; oe 1745 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c rc = (ie && !oe); oe 1747 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c rc = (!ie && oe); oe 229 drivers/pinctrl/pinctrl-st.c struct regmap_field *alt, *oe, *pu, *od; oe 244 drivers/pinctrl/pinctrl-st.c const int alt, oe, pu, od, rt; oe 346 drivers/pinctrl/pinctrl-st.c .alt = 0, .oe = 40, .pu = 50, .od = 60, .rt = 100, oe 356 drivers/pinctrl/pinctrl-st.c .oe = -1, /* Not Available */ oe 386 drivers/pinctrl/pinctrl-st.c struct regmap_field *output_enable = pc->oe; oe 579 drivers/pinctrl/pinctrl-st.c if (pc->oe) { oe 580 drivers/pinctrl/pinctrl-st.c regmap_field_read(pc->oe, &oe_value); oe 1147 drivers/pinctrl/pinctrl-st.c pc->oe = st_pc_get_value(dev, regmap, bank/4, data->oe, lsb, msb); oe 39 drivers/staging/comedi/drivers/dt2817.c unsigned int oe = 0; oe 57 drivers/staging/comedi/drivers/dt2817.c oe |= 0x1; oe 59 drivers/staging/comedi/drivers/dt2817.c oe |= 0x2; oe 61 drivers/staging/comedi/drivers/dt2817.c oe |= 0x4; oe 63 drivers/staging/comedi/drivers/dt2817.c oe |= 0x8; oe 65 drivers/staging/comedi/drivers/dt2817.c outb(oe, dev->iobase + DT2817_CR); oe 256 drivers/tty/amiserial.c int oe = 0; oe 327 drivers/tty/amiserial.c oe = 1; oe 331 drivers/tty/amiserial.c if (oe == 1) oe 22 fs/omfs/file.c struct omfs_extent *oe = (struct omfs_extent *) &bh->b_data[offset]; oe 24 fs/omfs/file.c oe->e_next = ~cpu_to_be64(0ULL); oe 25 fs/omfs/file.c oe->e_extent_count = cpu_to_be32(1), oe 26 fs/omfs/file.c oe->e_fill = cpu_to_be32(0x22), oe 27 fs/omfs/file.c oe->e_entry.e_cluster = ~cpu_to_be64(0ULL); oe 28 fs/omfs/file.c oe->e_entry.e_blocks = ~cpu_to_be64(0ULL); oe 34 fs/omfs/file.c struct omfs_extent *oe; oe 56 fs/omfs/file.c oe = (struct omfs_extent *)(&bh->b_data[OMFS_EXTENT_START]); oe 64 fs/omfs/file.c extent_count = be32_to_cpu(oe->e_extent_count); oe 70 fs/omfs/file.c next = be64_to_cpu(oe->e_next); oe 71 fs/omfs/file.c entry = &oe->e_entry; oe 82 fs/omfs/file.c omfs_make_empty_table(bh, (char *) oe - bh->b_data); oe 95 fs/omfs/file.c oe = (struct omfs_extent *) (&bh->b_data[OMFS_EXTENT_CONT]); oe 116 fs/omfs/file.c static int omfs_grow_extent(struct inode *inode, struct omfs_extent *oe, oe 120 fs/omfs/file.c struct omfs_extent_entry *entry = &oe->e_entry; oe 122 fs/omfs/file.c u32 extent_count = be32_to_cpu(oe->e_extent_count); oe 158 fs/omfs/file.c if (be32_to_cpu(oe->e_extent_count) > max_count-1) oe 179 fs/omfs/file.c be32_add_cpu(&oe->e_extent_count, 1); oe 226 fs/omfs/file.c struct omfs_extent *oe; oe 237 fs/omfs/file.c oe = (struct omfs_extent *)(&bh->b_data[OMFS_EXTENT_START]); oe 246 fs/omfs/file.c extent_count = be32_to_cpu(oe->e_extent_count); oe 247 fs/omfs/file.c next = be64_to_cpu(oe->e_next); oe 248 fs/omfs/file.c entry = &oe->e_entry; oe 269 fs/omfs/file.c oe = (struct omfs_extent *) (&bh->b_data[OMFS_EXTENT_CONT]); oe 273 fs/omfs/file.c ret = omfs_grow_extent(inode, oe, &new_block); oe 79 fs/overlayfs/export.c struct ovl_entry *oe = OVL_E(dentry); oe 83 fs/overlayfs/export.c return oe->numlower; oe 94 fs/overlayfs/export.c return oe->lowerstack[0].layer->idx; oe 298 fs/overlayfs/export.c struct ovl_entry *oe; oe 324 fs/overlayfs/export.c oe = ovl_alloc_entry(lower ? 1 : 0); oe 325 fs/overlayfs/export.c if (!oe) oe 329 fs/overlayfs/export.c oe->lowerstack->dentry = dget(lower); oe 330 fs/overlayfs/export.c oe->lowerstack->layer = lowerpath->layer; oe 332 fs/overlayfs/export.c dentry->d_fsdata = oe; oe 348 fs/overlayfs/export.c struct ovl_entry *oe = dentry->d_fsdata; oe 354 fs/overlayfs/export.c for (i = 0; i < oe->numlower; i++) { oe 355 fs/overlayfs/export.c if (oe->lowerstack[i].layer->idx == idx) oe 356 fs/overlayfs/export.c return oe->lowerstack[i].dentry; oe 776 fs/overlayfs/namei.c struct ovl_entry *oe = dentry->d_fsdata; oe 782 fs/overlayfs/namei.c return oe->numlower ? 1 : -1; oe 785 fs/overlayfs/namei.c BUG_ON(idx > oe->numlower); oe 786 fs/overlayfs/namei.c path->dentry = oe->lowerstack[idx - 1].dentry; oe 787 fs/overlayfs/namei.c path->mnt = oe->lowerstack[idx - 1].layer->mnt; oe 789 fs/overlayfs/namei.c return (idx < oe->numlower) ? idx + 1 : -1; oe 816 fs/overlayfs/namei.c struct ovl_entry *oe; oe 1045 fs/overlayfs/namei.c oe = ovl_alloc_entry(ctr); oe 1047 fs/overlayfs/namei.c if (!oe) oe 1050 fs/overlayfs/namei.c memcpy(oe->lowerstack, stack, sizeof(struct ovl_path) * ctr); oe 1051 fs/overlayfs/namei.c dentry->d_fsdata = oe; oe 1097 fs/overlayfs/namei.c kfree(oe); oe 56 fs/overlayfs/super.c static void ovl_entry_stack_free(struct ovl_entry *oe) oe 60 fs/overlayfs/super.c for (i = 0; i < oe->numlower; i++) oe 61 fs/overlayfs/super.c dput(oe->lowerstack[i].dentry); oe 71 fs/overlayfs/super.c struct ovl_entry *oe = dentry->d_fsdata; oe 73 fs/overlayfs/super.c if (oe) { oe 74 fs/overlayfs/super.c ovl_entry_stack_free(oe); oe 75 fs/overlayfs/super.c kfree_rcu(oe, rcu); oe 118 fs/overlayfs/super.c struct ovl_entry *oe = dentry->d_fsdata; oe 122 fs/overlayfs/super.c for (i = 0; i < oe->numlower; i++) { oe 123 fs/overlayfs/super.c struct dentry *d = oe->lowerstack[i].dentry; oe 141 fs/overlayfs/super.c struct ovl_entry *oe = dentry->d_fsdata; oe 145 fs/overlayfs/super.c for (i = 0; i < oe->numlower; i++) { oe 146 fs/overlayfs/super.c struct dentry *d = oe->lowerstack[i].dentry; oe 1200 fs/overlayfs/super.c struct ovl_entry *oe, struct path *upperpath) oe 1210 fs/overlayfs/super.c err = ovl_verify_origin(upperpath->dentry, oe->lowerstack[0].dentry, oe 1421 fs/overlayfs/super.c struct ovl_entry *oe; oe 1471 fs/overlayfs/super.c oe = ovl_alloc_entry(numlower); oe 1472 fs/overlayfs/super.c if (!oe) oe 1476 fs/overlayfs/super.c oe->lowerstack[i].dentry = dget(stack[i].dentry); oe 1477 fs/overlayfs/super.c oe->lowerstack[i].layer = &ofs->lower_layers[i]; oe 1491 fs/overlayfs/super.c return oe; oe 1494 fs/overlayfs/super.c oe = ERR_PTR(err); oe 1573 fs/overlayfs/super.c struct ovl_entry *oe; oe 1632 fs/overlayfs/super.c oe = ovl_get_lowerstack(sb, ofs); oe 1633 fs/overlayfs/super.c err = PTR_ERR(oe); oe 1634 fs/overlayfs/super.c if (IS_ERR(oe)) oe 1642 fs/overlayfs/super.c err = ovl_get_indexdir(sb, ofs, oe, &upperpath); oe 1689 fs/overlayfs/super.c root_dentry->d_fsdata = oe; oe 1710 fs/overlayfs/super.c ovl_entry_stack_free(oe); oe 1711 fs/overlayfs/super.c kfree(oe); oe 97 fs/overlayfs/util.c struct ovl_entry *oe = kzalloc(size, GFP_KERNEL); oe 99 fs/overlayfs/util.c if (oe) oe 100 fs/overlayfs/util.c oe->numlower = numlower; oe 102 fs/overlayfs/util.c return oe; oe 122 fs/overlayfs/util.c struct ovl_entry *oe = dentry->d_fsdata; oe 131 fs/overlayfs/util.c if (oe->numlower) { oe 139 fs/overlayfs/util.c if (oe->numlower > 1) oe 155 fs/overlayfs/util.c struct ovl_entry *oe = dentry->d_fsdata; oe 157 fs/overlayfs/util.c if (oe->numlower) { oe 158 fs/overlayfs/util.c path->mnt = oe->lowerstack[0].layer->mnt; oe 159 fs/overlayfs/util.c path->dentry = oe->lowerstack[0].dentry; oe 167 fs/overlayfs/util.c struct ovl_entry *oe = dentry->d_fsdata; oe 169 fs/overlayfs/util.c if (oe->numlower) { oe 170 fs/overlayfs/util.c path->mnt = oe->lowerstack[oe->numlower - 1].layer->mnt; oe 171 fs/overlayfs/util.c path->dentry = oe->lowerstack[oe->numlower - 1].dentry; oe 196 fs/overlayfs/util.c struct ovl_entry *oe = dentry->d_fsdata; oe 198 fs/overlayfs/util.c return oe->numlower ? oe->lowerstack[0].dentry : NULL; oe 203 fs/overlayfs/util.c struct ovl_entry *oe = dentry->d_fsdata; oe 205 fs/overlayfs/util.c return oe->numlower ? oe->lowerstack[0].layer : NULL; oe 216 fs/overlayfs/util.c struct ovl_entry *oe = dentry->d_fsdata; oe 218 fs/overlayfs/util.c return oe->numlower ? oe->lowerstack[oe->numlower - 1].dentry : NULL; oe 861 fs/overlayfs/util.c struct ovl_entry *oe = dentry->d_fsdata; oe 872 fs/overlayfs/util.c return (oe->numlower > 1); oe 98 lib/lzo/lzo1x_decompress_safe.c unsigned char *oe = op + t; oe 108 lib/lzo/lzo1x_decompress_safe.c op = oe; oe 215 lib/lzo/lzo1x_decompress_safe.c unsigned char *oe = op + t; oe 224 lib/lzo/lzo1x_decompress_safe.c } while (op < oe); oe 225 lib/lzo/lzo1x_decompress_safe.c op = oe; oe 237 lib/lzo/lzo1x_decompress_safe.c } while (op < oe); oe 242 lib/lzo/lzo1x_decompress_safe.c unsigned char *oe = op + t; oe 250 lib/lzo/lzo1x_decompress_safe.c } while (op < oe); oe 79 tools/perf/builtin-inject.c struct ordered_events *oe __maybe_unused) oe 87 tools/perf/builtin-inject.c struct ordered_events *oe __maybe_unused) oe 828 tools/perf/builtin-kvm.c struct ordered_events *oe = &kvm->session->ordered_events; oe 830 tools/perf/builtin-kvm.c oe->next_flush = flush_time; oe 831 tools/perf/builtin-kvm.c err = ordered_events__flush(oe, OE_FLUSH__ROUND); oe 2361 tools/perf/builtin-script.c struct ordered_events *oe __maybe_unused) oe 168 tools/perf/builtin-trace.c } oe; oe 3225 tools/perf/builtin-trace.c u64 first = ordered_events__first_time(&trace->oe.data); oe 3226 tools/perf/builtin-trace.c u64 flush = trace->oe.last - NSEC_PER_SEC; oe 3230 tools/perf/builtin-trace.c return ordered_events__flush_time(&trace->oe.data, flush); oe 3247 tools/perf/builtin-trace.c err = perf_evlist__parse_sample_timestamp(trace->evlist, event, &trace->oe.last); oe 3251 tools/perf/builtin-trace.c err = ordered_events__queue(&trace->oe.data, event, trace->oe.last, 0); oe 3258 tools/perf/builtin-trace.c static int ordered_events__deliver_event(struct ordered_events *oe, oe 3261 tools/perf/builtin-trace.c struct trace *trace = container_of(oe, struct trace, oe.data); oe 3498 tools/perf/builtin-trace.c ordered_events__flush(&trace->oe.data, OE_FLUSH__FINAL); oe 4298 tools/perf/builtin-trace.c ordered_events__init(&trace.oe.data, ordered_events__deliver_event, &trace); oe 4299 tools/perf/builtin-trace.c ordered_events__set_copy_on_queue(&trace.oe.data, true); oe 18 tools/perf/util/ordered-events.c static void queue_event(struct ordered_events *oe, struct ordered_event *new) oe 20 tools/perf/util/ordered-events.c struct ordered_event *last = oe->last; oe 24 tools/perf/util/ordered-events.c ++oe->nr_events; oe 25 tools/perf/util/ordered-events.c oe->last = new; oe 27 tools/perf/util/ordered-events.c pr_oe_time2(timestamp, "queue_event nr_events %u\n", oe->nr_events); oe 30 tools/perf/util/ordered-events.c list_add(&new->list, &oe->events); oe 31 tools/perf/util/ordered-events.c oe->max_timestamp = timestamp; oe 43 tools/perf/util/ordered-events.c if (p == &oe->events) { oe 44 tools/perf/util/ordered-events.c list_add_tail(&new->list, &oe->events); oe 45 tools/perf/util/ordered-events.c oe->max_timestamp = timestamp; oe 54 tools/perf/util/ordered-events.c if (p == &oe->events) { oe 55 tools/perf/util/ordered-events.c list_add(&new->list, &oe->events); oe 64 tools/perf/util/ordered-events.c static union perf_event *__dup_event(struct ordered_events *oe, oe 69 tools/perf/util/ordered-events.c if (oe->cur_alloc_size < oe->max_alloc_size) { oe 72 tools/perf/util/ordered-events.c oe->cur_alloc_size += event->header.size; oe 78 tools/perf/util/ordered-events.c static union perf_event *dup_event(struct ordered_events *oe, oe 81 tools/perf/util/ordered-events.c return oe->copy_on_queue ? __dup_event(oe, event) : event; oe 84 tools/perf/util/ordered-events.c static void __free_dup_event(struct ordered_events *oe, union perf_event *event) oe 87 tools/perf/util/ordered-events.c oe->cur_alloc_size -= event->header.size; oe 92 tools/perf/util/ordered-events.c static void free_dup_event(struct ordered_events *oe, union perf_event *event) oe 94 tools/perf/util/ordered-events.c if (oe->copy_on_queue) oe 95 tools/perf/util/ordered-events.c __free_dup_event(oe, event); oe 99 tools/perf/util/ordered-events.c static struct ordered_event *alloc_event(struct ordered_events *oe, oe 102 tools/perf/util/ordered-events.c struct list_head *cache = &oe->cache; oe 107 tools/perf/util/ordered-events.c new_event = dup_event(oe, event); oe 138 tools/perf/util/ordered-events.c size = sizeof(*oe->buffer) + MAX_SAMPLE_BUFFER * sizeof(*new); oe 143 tools/perf/util/ordered-events.c } else if (oe->buffer) { oe 144 tools/perf/util/ordered-events.c new = &oe->buffer->event[oe->buffer_idx]; oe 145 tools/perf/util/ordered-events.c if (++oe->buffer_idx == MAX_SAMPLE_BUFFER) oe 146 tools/perf/util/ordered-events.c oe->buffer = NULL; oe 147 tools/perf/util/ordered-events.c } else if ((oe->cur_alloc_size + size) < oe->max_alloc_size) { oe 148 tools/perf/util/ordered-events.c oe->buffer = malloc(size); oe 149 tools/perf/util/ordered-events.c if (!oe->buffer) { oe 150 tools/perf/util/ordered-events.c free_dup_event(oe, new_event); oe 155 tools/perf/util/ordered-events.c oe->cur_alloc_size, size, oe->max_alloc_size); oe 157 tools/perf/util/ordered-events.c oe->cur_alloc_size += size; oe 158 tools/perf/util/ordered-events.c list_add(&oe->buffer->list, &oe->to_free); oe 160 tools/perf/util/ordered-events.c oe->buffer_idx = 1; oe 161 tools/perf/util/ordered-events.c new = &oe->buffer->event[0]; oe 163 tools/perf/util/ordered-events.c pr("allocation limit reached %" PRIu64 "B\n", oe->max_alloc_size); oe 172 tools/perf/util/ordered-events.c ordered_events__new_event(struct ordered_events *oe, u64 timestamp, oe 177 tools/perf/util/ordered-events.c new = alloc_event(oe, event); oe 180 tools/perf/util/ordered-events.c queue_event(oe, new); oe 186 tools/perf/util/ordered-events.c void ordered_events__delete(struct ordered_events *oe, struct ordered_event *event) oe 188 tools/perf/util/ordered-events.c list_move(&event->list, &oe->cache); oe 189 tools/perf/util/ordered-events.c oe->nr_events--; oe 190 tools/perf/util/ordered-events.c free_dup_event(oe, event->event); oe 194 tools/perf/util/ordered-events.c int ordered_events__queue(struct ordered_events *oe, union perf_event *event, oe 202 tools/perf/util/ordered-events.c if (timestamp < oe->last_flush) { oe 204 tools/perf/util/ordered-events.c pr_oe_time(oe->last_flush, "last flush, last_flush_type %d\n", oe 205 tools/perf/util/ordered-events.c oe->last_flush_type); oe 207 tools/perf/util/ordered-events.c oe->nr_unordered_events++; oe 210 tools/perf/util/ordered-events.c oevent = ordered_events__new_event(oe, timestamp, event); oe 212 tools/perf/util/ordered-events.c ordered_events__flush(oe, OE_FLUSH__HALF); oe 213 tools/perf/util/ordered-events.c oevent = ordered_events__new_event(oe, timestamp, event); oe 223 tools/perf/util/ordered-events.c static int do_flush(struct ordered_events *oe, bool show_progress) oe 225 tools/perf/util/ordered-events.c struct list_head *head = &oe->events; oe 227 tools/perf/util/ordered-events.c u64 limit = oe->next_flush; oe 228 tools/perf/util/ordered-events.c u64 last_ts = oe->last ? oe->last->timestamp : 0ULL; oe 236 tools/perf/util/ordered-events.c ui_progress__init(&prog, oe->nr_events, "Processing time ordered events..."); oe 244 tools/perf/util/ordered-events.c ret = oe->deliver(oe, iter); oe 248 tools/perf/util/ordered-events.c ordered_events__delete(oe, iter); oe 249 tools/perf/util/ordered-events.c oe->last_flush = iter->timestamp; oe 256 tools/perf/util/ordered-events.c oe->last = NULL; oe 258 tools/perf/util/ordered-events.c oe->last = list_entry(head->prev, struct ordered_event, list); oe 266 tools/perf/util/ordered-events.c static int __ordered_events__flush(struct ordered_events *oe, enum oe_flush how, oe 280 tools/perf/util/ordered-events.c if (oe->nr_events == 0) oe 288 tools/perf/util/ordered-events.c oe->next_flush = ULLONG_MAX; oe 294 tools/perf/util/ordered-events.c struct list_head *head = &oe->events; oe 297 tools/perf/util/ordered-events.c last = oe->last; oe 303 tools/perf/util/ordered-events.c oe->next_flush = first->timestamp; oe 304 tools/perf/util/ordered-events.c oe->next_flush += (last->timestamp - first->timestamp) / 2; oe 309 tools/perf/util/ordered-events.c oe->next_flush = timestamp; oe 319 tools/perf/util/ordered-events.c pr_oe_time(oe->next_flush, "next_flush - ordered_events__flush PRE %s, nr_events %u\n", oe 320 tools/perf/util/ordered-events.c str[how], oe->nr_events); oe 321 tools/perf/util/ordered-events.c pr_oe_time(oe->max_timestamp, "max_timestamp\n"); oe 323 tools/perf/util/ordered-events.c err = do_flush(oe, show_progress); oe 327 tools/perf/util/ordered-events.c oe->next_flush = oe->max_timestamp; oe 329 tools/perf/util/ordered-events.c oe->last_flush_type = how; oe 332 tools/perf/util/ordered-events.c pr_oe_time(oe->next_flush, "next_flush - ordered_events__flush POST %s, nr_events %u\n", oe 333 tools/perf/util/ordered-events.c str[how], oe->nr_events); oe 334 tools/perf/util/ordered-events.c pr_oe_time(oe->last_flush, "last_flush\n"); oe 339 tools/perf/util/ordered-events.c int ordered_events__flush(struct ordered_events *oe, enum oe_flush how) oe 341 tools/perf/util/ordered-events.c return __ordered_events__flush(oe, how, 0); oe 344 tools/perf/util/ordered-events.c int ordered_events__flush_time(struct ordered_events *oe, u64 timestamp) oe 346 tools/perf/util/ordered-events.c return __ordered_events__flush(oe, OE_FLUSH__TIME, timestamp); oe 349 tools/perf/util/ordered-events.c u64 ordered_events__first_time(struct ordered_events *oe) oe 353 tools/perf/util/ordered-events.c if (list_empty(&oe->events)) oe 356 tools/perf/util/ordered-events.c event = list_first_entry(&oe->events, struct ordered_event, list); oe 360 tools/perf/util/ordered-events.c void ordered_events__init(struct ordered_events *oe, ordered_events__deliver_t deliver, oe 363 tools/perf/util/ordered-events.c INIT_LIST_HEAD(&oe->events); oe 364 tools/perf/util/ordered-events.c INIT_LIST_HEAD(&oe->cache); oe 365 tools/perf/util/ordered-events.c INIT_LIST_HEAD(&oe->to_free); oe 366 tools/perf/util/ordered-events.c oe->max_alloc_size = (u64) -1; oe 367 tools/perf/util/ordered-events.c oe->cur_alloc_size = 0; oe 368 tools/perf/util/ordered-events.c oe->deliver = deliver; oe 369 tools/perf/util/ordered-events.c oe->data = data; oe 374 tools/perf/util/ordered-events.c unsigned int max, struct ordered_events *oe) oe 376 tools/perf/util/ordered-events.c if (oe->copy_on_queue) { oe 380 tools/perf/util/ordered-events.c __free_dup_event(oe, buffer->event[i].event); oe 386 tools/perf/util/ordered-events.c void ordered_events__free(struct ordered_events *oe) oe 390 tools/perf/util/ordered-events.c if (list_empty(&oe->to_free)) oe 397 tools/perf/util/ordered-events.c if (oe->buffer) { oe 398 tools/perf/util/ordered-events.c list_del_init(&oe->buffer->list); oe 399 tools/perf/util/ordered-events.c ordered_events_buffer__free(oe->buffer, oe->buffer_idx, oe); oe 403 tools/perf/util/ordered-events.c list_for_each_entry_safe(buffer, tmp, &oe->to_free, list) { oe 405 tools/perf/util/ordered-events.c ordered_events_buffer__free(buffer, MAX_SAMPLE_BUFFER, oe); oe 409 tools/perf/util/ordered-events.c void ordered_events__reinit(struct ordered_events *oe) oe 411 tools/perf/util/ordered-events.c ordered_events__deliver_t old_deliver = oe->deliver; oe 413 tools/perf/util/ordered-events.c ordered_events__free(oe); oe 414 tools/perf/util/ordered-events.c memset(oe, '\0', sizeof(*oe)); oe 415 tools/perf/util/ordered-events.c ordered_events__init(oe, old_deliver, oe->data); oe 27 tools/perf/util/ordered-events.h typedef int (*ordered_events__deliver_t)(struct ordered_events *oe, oe 55 tools/perf/util/ordered-events.h int ordered_events__queue(struct ordered_events *oe, union perf_event *event, oe 57 tools/perf/util/ordered-events.h void ordered_events__delete(struct ordered_events *oe, struct ordered_event *event); oe 58 tools/perf/util/ordered-events.h int ordered_events__flush(struct ordered_events *oe, enum oe_flush how); oe 59 tools/perf/util/ordered-events.h int ordered_events__flush_time(struct ordered_events *oe, u64 timestamp); oe 60 tools/perf/util/ordered-events.h void ordered_events__init(struct ordered_events *oe, ordered_events__deliver_t deliver, oe 62 tools/perf/util/ordered-events.h void ordered_events__free(struct ordered_events *oe); oe 63 tools/perf/util/ordered-events.h void ordered_events__reinit(struct ordered_events *oe); oe 64 tools/perf/util/ordered-events.h u64 ordered_events__first_time(struct ordered_events *oe); oe 67 tools/perf/util/ordered-events.h void ordered_events__set_alloc_size(struct ordered_events *oe, u64 size) oe 69 tools/perf/util/ordered-events.h oe->max_alloc_size = size; oe 73 tools/perf/util/ordered-events.h void ordered_events__set_copy_on_queue(struct ordered_events *oe, bool copy) oe 75 tools/perf/util/ordered-events.h oe->copy_on_queue = copy; oe 177 tools/perf/util/session.c static int ordered_events__deliver_event(struct ordered_events *oe, oe 180 tools/perf/util/session.c struct perf_session *session = container_of(oe, struct perf_session, oe 355 tools/perf/util/session.c struct ordered_events *oe __maybe_unused) oe 363 tools/perf/util/session.c struct ordered_events *oe); oe 988 tools/perf/util/session.c struct ordered_events *oe) oe 992 tools/perf/util/session.c return ordered_events__flush(oe, OE_FLUSH__ROUND); oe 1502 tools/perf/util/session.c struct ordered_events *oe = &session->ordered_events; oe 1536 tools/perf/util/session.c return tool->finished_round(tool, event, oe); oe 1728 tools/perf/util/session.c const struct ordered_events *oe = &session->ordered_events; oe 1739 tools/perf/util/session.c if (oe->nr_unordered_events != 0) oe 1740 tools/perf/util/session.c ui__warning("%u out of order events recorded.\n", oe->nr_unordered_events); oe 1854 tools/perf/util/session.c struct ordered_events *oe = &session->ordered_events; oe 1873 tools/perf/util/session.c ordered_events__set_copy_on_queue(oe, true); oe 1940 tools/perf/util/session.c err = ordered_events__flush(oe, OE_FLUSH__FINAL); oe 2175 tools/perf/util/session.c struct ordered_events *oe = &session->ordered_events; oe 2191 tools/perf/util/session.c err = ordered_events__flush(oe, OE_FLUSH__FINAL); oe 34 tools/perf/util/tool.h struct ordered_events *oe);