rep 277 arch/arm/mach-davinci/board-dm365-evm.c .rep = 1, rep 153 arch/arm/mach-imx/mach-pcm037_eet.c .rep = 0, /* No auto-repeat */ rep 258 arch/arm/mach-omap1/board-h2.c .rep = true, rep 296 arch/arm/mach-omap1/board-h3.c .rep = true, rep 226 arch/arm/mach-omap1/board-htcherald.c .rep = true, rep 268 arch/arm/mach-omap1/board-htcherald.c .rep = true, rep 83 arch/arm/mach-omap1/board-palmte.c .rep = true, rep 80 arch/arm/mach-omap1/board-palmz71.c .rep = true, rep 519 arch/arm/mach-sa1100/assabet.c .rep = 0, rep 405 arch/mips/alchemy/devboards/db1300.c .rep = 1, rep 1602 arch/x86/kvm/hyperv.c bool fast, rep; rep 1633 arch/x86/kvm/hyperv.c rep = !!(rep_cnt || rep_idx); rep 1639 arch/x86/kvm/hyperv.c if (unlikely(rep)) { rep 1646 arch/x86/kvm/hyperv.c if (unlikely(rep)) { rep 1656 arch/x86/kvm/hyperv.c if (unlikely(rep || !vcpu_to_synic(vcpu)->active)) { rep 1676 arch/x86/kvm/hyperv.c if (unlikely(fast || rep)) { rep 1690 arch/x86/kvm/hyperv.c if (unlikely(fast || rep)) { rep 1697 arch/x86/kvm/hyperv.c if (unlikely(rep)) { rep 1704 arch/x86/kvm/hyperv.c if (unlikely(fast || rep)) { rep 42 arch/x86/platform/intel-mid/device_libs/platform_gpio_keys.c .rep = 1, rep 47 arch/xtensa/include/asm/cacheasm.h .rep _reps rep 104 block/blk-zoned.c static bool blkdev_report_zone(struct block_device *bdev, struct blk_zone *rep) rep 108 block/blk-zoned.c if (rep->start < offset) rep 111 block/blk-zoned.c rep->start -= offset; rep 112 block/blk-zoned.c if (rep->start + rep->len > bdev->bd_part->nr_sects) rep 115 block/blk-zoned.c if (rep->type == BLK_ZONE_TYPE_CONVENTIONAL) rep 116 block/blk-zoned.c rep->wp = rep->start + rep->len; rep 118 block/blk-zoned.c rep->wp -= offset; rep 319 block/blk-zoned.c struct blk_zone_report rep; rep 336 block/blk-zoned.c if (copy_from_user(&rep, argp, sizeof(struct blk_zone_report))) rep 339 block/blk-zoned.c if (!rep.nr_zones) rep 342 block/blk-zoned.c rep.nr_zones = min(blkdev_nr_zones(bdev), rep.nr_zones); rep 344 block/blk-zoned.c zones = kvmalloc_array(rep.nr_zones, sizeof(struct blk_zone), rep 349 block/blk-zoned.c ret = blkdev_report_zones(bdev, rep.sector, zones, &rep.nr_zones); rep 353 block/blk-zoned.c if (copy_to_user(argp, &rep, sizeof(struct blk_zone_report))) { rep 358 block/blk-zoned.c if (rep.nr_zones) { rep 360 block/blk-zoned.c sizeof(struct blk_zone) * rep.nr_zones)) rep 379 drivers/gpu/drm/drm_dp_mst_topology.c static void drm_dp_encode_sideband_reply(struct drm_dp_sideband_msg_reply_body *rep, rep 385 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx++] = (rep->reply_type & 0x1) << 7 | (rep->req_type & 0x7f); rep 1432 drivers/gpu/drm/i2c/tda998x_drv.c u8 reg, div, rep, sel_clk; rep 1509 drivers/gpu/drm/i2c/tda998x_drv.c rep = mode->flags & DRM_MODE_FLAG_DBLCLK ? 1 : 0; rep 1511 drivers/gpu/drm/i2c/tda998x_drv.c SEL_CLK_SEL_VRF_CLK(rep ? 2 : 0); rep 1514 drivers/gpu/drm/i2c/tda998x_drv.c tmds_clock = mode->clock * (1 + rep); rep 1554 drivers/gpu/drm/i2c/tda998x_drv.c reg_write(priv, REG_RPT_CNTRL, RPT_CNTRL_REPEAT(rep)); rep 1557 drivers/gpu/drm/i2c/tda998x_drv.c PLL_SERIAL_2_SRL_PR(rep)); rep 199 drivers/gpu/drm/i915/intel_runtime_pm.c unsigned long rep; rep 201 drivers/gpu/drm/i915/intel_runtime_pm.c rep = 1; rep 203 drivers/gpu/drm/i915/intel_runtime_pm.c rep++, i++; rep 205 drivers/gpu/drm/i915/intel_runtime_pm.c drm_printf(p, "Wakeref x%lu taken at:\n%s", rep, buf); rep 1144 drivers/gpu/drm/nouveau/nouveau_connector.c const struct nvif_notify_conn_rep_v0 *rep = notify->data; rep 1148 drivers/gpu/drm/nouveau/nouveau_connector.c bool plugged = (rep->mask != NVIF_NOTIFY_CONN_V0_UNPLUG); rep 1150 drivers/gpu/drm/nouveau/nouveau_connector.c if (rep->mask & NVIF_NOTIFY_CONN_V0_IRQ) { rep 222 drivers/gpu/drm/nouveau/nouveau_gem.c struct drm_nouveau_gem_info *rep) rep 230 drivers/gpu/drm/nouveau/nouveau_gem.c rep->domain = nvbo->valid_domains; rep 232 drivers/gpu/drm/nouveau/nouveau_gem.c rep->domain = NOUVEAU_GEM_DOMAIN_GART; rep 234 drivers/gpu/drm/nouveau/nouveau_gem.c rep->domain = NOUVEAU_GEM_DOMAIN_VRAM; rep 235 drivers/gpu/drm/nouveau/nouveau_gem.c rep->offset = nvbo->bo.offset; rep 241 drivers/gpu/drm/nouveau/nouveau_gem.c rep->offset = vma->addr; rep 244 drivers/gpu/drm/nouveau/nouveau_gem.c rep->size = nvbo->bo.mem.num_pages << PAGE_SHIFT; rep 245 drivers/gpu/drm/nouveau/nouveau_gem.c rep->map_handle = drm_vma_node_offset_addr(&nvbo->bo.base.vma_node); rep 246 drivers/gpu/drm/nouveau/nouveau_gem.c rep->tile_mode = nvbo->mode; rep 247 drivers/gpu/drm/nouveau/nouveau_gem.c rep->tile_flags = nvbo->contig ? 0 : NOUVEAU_GEM_TILE_NONCONTIG; rep 249 drivers/gpu/drm/nouveau/nouveau_gem.c rep->tile_flags |= nvbo->kind << 8; rep 252 drivers/gpu/drm/nouveau/nouveau_gem.c rep->tile_flags |= nvbo->kind << 8 | nvbo->comp << 16; rep 254 drivers/gpu/drm/nouveau/nouveau_gem.c rep->tile_flags |= nvbo->zeta; rep 78 drivers/gpu/drm/nouveau/nouveau_usif.c } *rep = header; rep 83 drivers/gpu/drm/nouveau/nouveau_usif.c if (length == sizeof(rep->v0) && rep->v0.version == 0) { rep 84 drivers/gpu/drm/nouveau/nouveau_usif.c if (WARN_ON(!(ntfy = (void *)(unsigned long)rep->v0.token))) rep 86 drivers/gpu/drm/nouveau/nouveau_usif.c BUG_ON(rep->v0.route != NVDRM_NOTIFY_USIF); rep 98 drivers/gpu/drm/nouveau/nouveau_usif.c switch (rep->v0.version) { rep 100 drivers/gpu/drm/nouveau/nouveau_usif.c struct nvif_notify_rep_v0 *rep = (void *)ntfy->p->e.data; rep 101 drivers/gpu/drm/nouveau/nouveau_usif.c rep->route = ntfy->route; rep 102 drivers/gpu/drm/nouveau/nouveau_usif.c rep->token = ntfy->token; rep 78 drivers/gpu/drm/nouveau/nvkm/core/client.c } rep; rep 86 drivers/gpu/drm/nouveau/nvkm/core/client.c return client->ntfy(¬ify->rep, notify->size, n->data, n->size); rep 157 drivers/gpu/drm/nouveau/nvkm/core/client.c notify->size = sizeof(notify->rep.v0); rep 158 drivers/gpu/drm/nouveau/nvkm/core/client.c notify->rep.v0.version = req->v0.version; rep 159 drivers/gpu/drm/nouveau/nvkm/core/client.c notify->rep.v0.route = req->v0.route; rep 160 drivers/gpu/drm/nouveau/nvkm/core/client.c notify->rep.v0.token = req->v0.token; rep 93 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c struct nvif_notify_head_rep_v0 rep = {}; rep 94 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c nvkm_event_send(&disp->vblank, 1, head, &rep, sizeof(rep)); rep 128 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c } rep; rep 130 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nvkm_event_send(&disp->uevent, 1, chid, &rep, sizeof(rep)); rep 39 drivers/gpu/drm/nouveau/nvkm/engine/disp/conn.c struct nvif_notify_conn_rep_v0 rep; rep 45 drivers/gpu/drm/nouveau/nvkm/engine/disp/conn.c rep.mask = NVIF_NOTIFY_CONN_V0_UNPLUG; rep 47 drivers/gpu/drm/nouveau/nvkm/engine/disp/conn.c rep.mask = NVIF_NOTIFY_CONN_V0_PLUG; rep 48 drivers/gpu/drm/nouveau/nvkm/engine/disp/conn.c rep.version = 0; rep 50 drivers/gpu/drm/nouveau/nvkm/engine/disp/conn.c nvkm_event_send(&disp->hpd, rep.mask, index, &rep, sizeof(rep)); rep 542 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c struct nvif_notify_conn_rep_v0 rep = {}; rep 548 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c rep.mask |= NVIF_NOTIFY_CONN_V0_IRQ; rep 554 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c rep.mask |= NVIF_NOTIFY_CONN_V0_UNPLUG; rep 556 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c rep.mask |= NVIF_NOTIFY_CONN_V0_PLUG; rep 558 drivers/gpu/drm/nouveau/nvkm/engine/disp/dp.c nvkm_event_send(&disp->hpd, rep.mask, conn->index, &rep, sizeof(rep)); rep 213 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c struct nvif_notify_uevent_rep rep = { rep 215 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c nvkm_event_send(&fifo->uevent, 1, 0, &rep, sizeof(rep)); rep 1891 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv50.c int acnt = 0x10, rep, i; rep 2048 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv50.c rep = 0xc; rep 2050 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv50.c rep = 4; rep 2051 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv50.c for (i = 0; i < rep; i++) { rep 156 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c struct nvkm_gpio_ntfy_rep rep = { rep 160 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c nvkm_event_send(&gpio->event, rep.mask, i, &rep, sizeof(rep)); rep 149 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c struct nvkm_i2c_ntfy_rep rep = { rep 152 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c nvkm_event_send(&i2c->event, rep.mask, aux->id, rep 153 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c &rep, sizeof(rep)); rep 823 drivers/gpu/drm/vmwgfx/vmwgfx_bo.c struct drm_vmw_dmabuf_rep *rep = &arg->rep; rep 838 drivers/gpu/drm/vmwgfx/vmwgfx_bo.c rep->handle = handle; rep 839 drivers/gpu/drm/vmwgfx/vmwgfx_bo.c rep->map_handle = drm_vma_node_offset_addr(&vbo->base.base.vma_node); rep 840 drivers/gpu/drm/vmwgfx/vmwgfx_bo.c rep->cur_gmr_id = handle; rep 841 drivers/gpu/drm/vmwgfx/vmwgfx_bo.c rep->cur_gmr_offset = 0; rep 805 drivers/gpu/drm/vmwgfx/vmwgfx_context.c struct drm_vmw_context_arg *rep = &arg->rep; rep 809 drivers/gpu/drm/vmwgfx/vmwgfx_context.c return vmw_context_define(dev, rep, file_priv, false); rep 811 drivers/gpu/drm/vmwgfx/vmwgfx_context.c return vmw_context_define(dev, rep, file_priv, true); rep 91 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_gb_surface_create_rep *rep, rep 96 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_gb_surface_ref_ext_rep *rep, rep 711 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_surface_arg *rep = &arg->rep; rep 881 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->sid = user_srf->prime.base.handle; rep 984 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_surface_create_req *rep = &arg->rep; rep 1001 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->flags = (uint32_t)srf->flags; rep 1002 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->format = srf->format; rep 1003 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c memcpy(rep->mip_levels, srf->mip_levels, sizeof(srf->mip_levels)); rep 1005 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->size_addr; rep 1292 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_gb_surface_create_rep *rep = &arg->rep; rep 1302 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c return vmw_gb_surface_define_internal(dev, &req_ext, rep, file_priv); rep 1319 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_gb_surface_ref_rep *rep = &arg->rep; rep 1328 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq = rep_ext.creq.base; rep 1329 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->crep = rep_ext.crep; rep 1513 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_gb_surface_create_rep *rep = &arg->rep; rep 1515 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c return vmw_gb_surface_define_internal(dev, req, rep, file_priv); rep 1532 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_gb_surface_ref_ext_rep *rep = &arg->rep; rep 1534 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c return vmw_gb_surface_reference_internal(dev, req, rep, file_priv); rep 1549 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_gb_surface_create_rep *rep, rep 1666 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->handle = user_srf->prime.base.handle; rep 1667 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->backup_size = res->backup_size; rep 1669 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->buffer_map_handle = rep 1671 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->buffer_size = res->backup->base.num_pages * PAGE_SIZE; rep 1672 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->buffer_handle = backup_handle; rep 1674 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->buffer_map_handle = 0; rep 1675 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->buffer_size = 0; rep 1676 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->buffer_handle = SVGA3D_INVALID_ID; rep 1698 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c struct drm_vmw_gb_surface_ref_ext_rep *rep, rep 1733 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.svga3d_flags = SVGA3D_FLAGS_LOWER_32(srf->flags); rep 1734 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.format = srf->format; rep 1735 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.mip_levels = srf->mip_levels[0]; rep 1736 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.drm_surface_flags = 0; rep 1737 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.multisample_count = srf->multisample_count; rep 1738 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.autogen_filter = srf->autogen_filter; rep 1739 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.array_size = srf->array_size; rep 1740 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.buffer_handle = backup_handle; rep 1741 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.base.base_size = srf->base_size; rep 1742 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->crep.handle = user_srf->prime.base.handle; rep 1743 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->crep.backup_size = srf->res.backup_size; rep 1744 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->crep.buffer_handle = backup_handle; rep 1745 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->crep.buffer_map_handle = rep 1747 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->crep.buffer_size = srf->res.backup->base.num_pages * PAGE_SIZE; rep 1749 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.version = drm_vmw_gb_surface_v1; rep 1750 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.svga3d_flags_upper_32_bits = rep 1752 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.multisample_pattern = srf->multisample_pattern; rep 1753 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.quality_level = srf->quality_level; rep 1754 drivers/gpu/drm/vmwgfx/vmwgfx_surface.c rep->creq.must_be_zero = 0; rep 1081 drivers/hid/hid-core.c struct hid_report *rep; rep 1113 drivers/hid/hid-core.c list_for_each_entry(rep, &rep_enum->report_list, list) { rep 1114 drivers/hid/hid-core.c for (i = 0; i < rep->maxfield; i++) { rep 1115 drivers/hid/hid-core.c field = rep->field[i]; rep 1149 drivers/hid/hid-core.c struct hid_report *rep; rep 1154 drivers/hid/hid-core.c list_for_each_entry(rep, &rep_enum->report_list, list) { rep 1155 drivers/hid/hid-core.c for (i = 0; i < rep->maxfield; i++) { rep 1157 drivers/hid/hid-core.c if (rep->field[i]->report_count < 1) rep 1160 drivers/hid/hid-core.c for (j = 0; j < rep->field[i]->maxusage; j++) { rep 1161 drivers/hid/hid-core.c usage = &rep->field[i]->usage[j]; rep 1164 drivers/hid/hid-core.c rep->field[i]); rep 1607 drivers/hid/hid-input.c struct hid_report *rep; rep 1611 drivers/hid/hid-input.c list_for_each_entry(rep, &rep_enum->report_list, list) { rep 1613 drivers/hid/hid-input.c rep, true); rep 1616 drivers/hid/hid-input.c ret = __hid_request(hid, rep, HID_REQ_SET_REPORT); rep 1619 drivers/hid/hid-input.c rep, false); rep 1633 drivers/hid/hid-input.c struct hid_report *rep; rep 1638 drivers/hid/hid-input.c list_for_each_entry(rep, &rep_enum->report_list, list) rep 1639 drivers/hid/hid-input.c for (i = 0; i < rep->maxfield; i++) { rep 1641 drivers/hid/hid-input.c if (rep->field[i]->report_count < 1) rep 1644 drivers/hid/hid-input.c for (j = 0; j < rep->field[i]->maxusage; j++) { rep 1645 drivers/hid/hid-input.c usage = &rep->field[i]->usage[j]; rep 1650 drivers/hid/hid-input.c rep->field[i]); rep 1653 drivers/hid/hid-input.c drv->feature_mapping(hid, rep->field[i], usage); rep 1630 drivers/hid/hid-logitech-dj.c struct hid_report *rep; rep 1683 drivers/hid/hid-logitech-dj.c list_for_each_entry(rep, &rep_enum->report_list, list) { rep 1684 drivers/hid/hid-logitech-dj.c if (rep->application == 0xff000001) rep 1696 drivers/hid/hid-logitech-dj.c rep = list_first_entry(&rep_enum->report_list, struct hid_report, list); rep 1698 drivers/hid/hid-logitech-dj.c rep->application, has_hidpp); rep 1705 drivers/hid/hid-logitech-dj.c djrcv_dev->unnumbered_application = rep->application; rep 1485 drivers/hid/hid-multitouch.c struct hid_report *rep; rep 1492 drivers/hid/hid-multitouch.c list_for_each_entry(rep, &rep_enum->report_list, list) { rep 1495 drivers/hid/hid-multitouch.c for (i = 0; i < rep->maxfield; i++) { rep 1497 drivers/hid/hid-multitouch.c if (rep->field[i]->report_count < 1) rep 1500 drivers/hid/hid-multitouch.c for (j = 0; j < rep->field[i]->maxusage; j++) { rep 1501 drivers/hid/hid-multitouch.c usage = &rep->field[i]->usage[j]; rep 1504 drivers/hid/hid-multitouch.c rep->field[i], rep 1515 drivers/hid/hid-multitouch.c hid_hw_request(hdev, rep, HID_REQ_SET_REPORT); rep 104 drivers/hid/intel-ish-hid/ishtp-hid.c static void ishtp_hid_request(struct hid_device *hid, struct hid_report *rep, rep 109 drivers/hid/intel-ish-hid/ishtp-hid.c unsigned int len = ((rep->size - 1) >> 3) + 1 + (rep->id > 0); rep 119 drivers/hid/intel-ish-hid/ishtp-hid.c hid_ishtp_get_report(hid, rep->id, rep->type); rep 130 drivers/hid/intel-ish-hid/ishtp-hid.c hid_output_report(rep, buf + header_size); rep 131 drivers/hid/intel-ish-hid/ishtp-hid.c hid_ishtp_set_feature(hid, buf, len, rep->id); rep 1265 drivers/hid/usbhid/hid-core.c static void usbhid_request(struct hid_device *hid, struct hid_report *rep, int reqtype) rep 1269 drivers/hid/usbhid/hid-core.c usbhid_submit_report(hid, rep, USB_DIR_IN); rep 1272 drivers/hid/usbhid/hid-core.c usbhid_submit_report(hid, rep, USB_DIR_OUT); rep 186 drivers/hwtracing/coresight/coresight-platform.c struct device_node *rep = NULL; rep 198 drivers/hwtracing/coresight/coresight-platform.c rep = of_graph_get_remote_endpoint(ep); rep 199 drivers/hwtracing/coresight/coresight-platform.c if (!rep) rep 201 drivers/hwtracing/coresight/coresight-platform.c rparent = of_coresight_get_port_parent(rep); rep 204 drivers/hwtracing/coresight/coresight-platform.c if (of_graph_parse_endpoint(rep, &rendpoint)) rep 231 drivers/hwtracing/coresight/coresight-platform.c of_node_put(rep); rep 95 drivers/iio/magnetometer/bmc150_magn.c #define BMC150_MAGN_REPXY_TO_REGVAL(rep) (((rep) - 1) / 2) rep 96 drivers/iio/magnetometer/bmc150_magn.c #define BMC150_MAGN_REPZ_TO_REGVAL(rep) ((rep) - 1) rep 187 drivers/infiniband/core/cm.c static CM_COUNTER_ATTR(rep, CM_REP_COUNTER); rep 3676 drivers/infiniband/core/cma.c const struct ib_cm_sidr_rep_event_param *rep = rep 3692 drivers/infiniband/core/cma.c if (rep->status != IB_SIDR_SUCCESS) { rep 3699 drivers/infiniband/core/cma.c ret = cma_set_qkey(id_priv, rep->qkey); rep 3710 drivers/infiniband/core/cma.c rep->sgid_attr); rep 3711 drivers/infiniband/core/cma.c event.param.ud.qp_num = rep->qpn; rep 3712 drivers/infiniband/core/cma.c event.param.ud.qkey = rep->qkey; rep 3949 drivers/infiniband/core/cma.c struct ib_cm_rep_param rep; rep 3960 drivers/infiniband/core/cma.c memset(&rep, 0, sizeof rep); rep 3961 drivers/infiniband/core/cma.c rep.qp_num = id_priv->qp_num; rep 3962 drivers/infiniband/core/cma.c rep.starting_psn = id_priv->seq_num; rep 3963 drivers/infiniband/core/cma.c rep.private_data = conn_param->private_data; rep 3964 drivers/infiniband/core/cma.c rep.private_data_len = conn_param->private_data_len; rep 3965 drivers/infiniband/core/cma.c rep.responder_resources = conn_param->responder_resources; rep 3966 drivers/infiniband/core/cma.c rep.initiator_depth = conn_param->initiator_depth; rep 3967 drivers/infiniband/core/cma.c rep.failover_accepted = 0; rep 3968 drivers/infiniband/core/cma.c rep.flow_control = conn_param->flow_control; rep 3969 drivers/infiniband/core/cma.c rep.rnr_retry_count = min_t(u8, 7, conn_param->rnr_retry_count); rep 3970 drivers/infiniband/core/cma.c rep.srq = id_priv->srq ? 1 : 0; rep 3972 drivers/infiniband/core/cma.c ret = ib_send_cm_rep(id_priv->cm_id.ib, &rep); rep 4006 drivers/infiniband/core/cma.c struct ib_cm_sidr_rep_param rep; rep 4009 drivers/infiniband/core/cma.c memset(&rep, 0, sizeof rep); rep 4010 drivers/infiniband/core/cma.c rep.status = status; rep 4015 drivers/infiniband/core/cma.c rep.qp_num = id_priv->qp_num; rep 4016 drivers/infiniband/core/cma.c rep.qkey = id_priv->qkey; rep 4018 drivers/infiniband/core/cma.c rep.private_data = private_data; rep 4019 drivers/infiniband/core/cma.c rep.private_data_len = private_data_len; rep 4021 drivers/infiniband/core/cma.c return ib_send_cm_sidr_rep(id_priv->cm_id.ib, &rep); rep 1700 drivers/infiniband/hw/cxgb3/iwch_cm.c struct cpl_rdma_ec_status *rep = cplhdr(skb); rep 1704 drivers/infiniband/hw/cxgb3/iwch_cm.c rep->status); rep 1705 drivers/infiniband/hw/cxgb3/iwch_cm.c if (rep->status) { rep 11 drivers/infiniband/hw/mlx5/ib_rep.c mlx5_ib_set_vport_rep(struct mlx5_core_dev *dev, struct mlx5_eswitch_rep *rep) rep 17 drivers/infiniband/hw/mlx5/ib_rep.c vport_index = rep->vport_index; rep 19 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->port[vport_index].rep = rep; rep 20 drivers/infiniband/hw/mlx5/ib_rep.c rep->rep_data[REP_IB].priv = ibdev; rep 23 drivers/infiniband/hw/mlx5/ib_rep.c mlx5_ib_get_rep_netdev(dev->priv.eswitch, rep->vport); rep 30 drivers/infiniband/hw/mlx5/ib_rep.c mlx5_ib_vport_rep_load(struct mlx5_core_dev *dev, struct mlx5_eswitch_rep *rep) rep 37 drivers/infiniband/hw/mlx5/ib_rep.c if (rep->vport == MLX5_VPORT_UPLINK) rep 40 drivers/infiniband/hw/mlx5/ib_rep.c return mlx5_ib_set_vport_rep(dev, rep); rep 54 drivers/infiniband/hw/mlx5/ib_rep.c vport_index = rep->vport_index; rep 55 drivers/infiniband/hw/mlx5/ib_rep.c ibdev->port[vport_index].rep = rep; rep 57 drivers/infiniband/hw/mlx5/ib_rep.c mlx5_ib_get_rep_netdev(dev->priv.eswitch, rep->vport); rep 64 drivers/infiniband/hw/mlx5/ib_rep.c rep->rep_data[REP_IB].priv = ibdev; rep 70 drivers/infiniband/hw/mlx5/ib_rep.c mlx5_ib_vport_rep_unload(struct mlx5_eswitch_rep *rep) rep 72 drivers/infiniband/hw/mlx5/ib_rep.c struct mlx5_ib_dev *dev = mlx5_ib_rep_to_dev(rep); rep 75 drivers/infiniband/hw/mlx5/ib_rep.c port = &dev->port[rep->vport_index]; rep 79 drivers/infiniband/hw/mlx5/ib_rep.c rep->rep_data[REP_IB].priv = NULL; rep 80 drivers/infiniband/hw/mlx5/ib_rep.c port->rep = NULL; rep 82 drivers/infiniband/hw/mlx5/ib_rep.c if (rep->vport == MLX5_VPORT_UPLINK) rep 86 drivers/infiniband/hw/mlx5/ib_rep.c static void *mlx5_ib_vport_get_proto_dev(struct mlx5_eswitch_rep *rep) rep 88 drivers/infiniband/hw/mlx5/ib_rep.c return mlx5_ib_rep_to_dev(rep); rep 144 drivers/infiniband/hw/mlx5/ib_rep.c struct mlx5_eswitch_rep *rep; rep 149 drivers/infiniband/hw/mlx5/ib_rep.c if (!dev->port[port - 1].rep) rep 152 drivers/infiniband/hw/mlx5/ib_rep.c rep = dev->port[port - 1].rep; rep 154 drivers/infiniband/hw/mlx5/ib_rep.c return mlx5_eswitch_add_send_to_vport_rule(esw, rep->vport, rep 73 drivers/infiniband/hw/mlx5/ib_rep.h struct mlx5_ib_dev *mlx5_ib_rep_to_dev(struct mlx5_eswitch_rep *rep) rep 75 drivers/infiniband/hw/mlx5/ib_rep.h return rep->rep_data[REP_IB].priv; rep 56 drivers/infiniband/hw/mlx5/ib_virt.c struct mlx5_hca_vport_context *rep; rep 59 drivers/infiniband/hw/mlx5/ib_virt.c rep = kzalloc(sizeof(*rep), GFP_KERNEL); rep 60 drivers/infiniband/hw/mlx5/ib_virt.c if (!rep) rep 63 drivers/infiniband/hw/mlx5/ib_virt.c err = mlx5_query_hca_vport_context(mdev, 1, 1, vf + 1, rep); rep 70 drivers/infiniband/hw/mlx5/ib_virt.c info->linkstate = mlx_to_net_policy(rep->policy); rep 75 drivers/infiniband/hw/mlx5/ib_virt.c kfree(rep); rep 171 drivers/infiniband/hw/mlx5/main.c if (!port->rep) rep 176 drivers/infiniband/hw/mlx5/main.c port->rep->vport); rep 1300 drivers/infiniband/hw/mlx5/main.c struct mlx5_hca_vport_context *rep; rep 1307 drivers/infiniband/hw/mlx5/main.c rep = kzalloc(sizeof(*rep), GFP_KERNEL); rep 1308 drivers/infiniband/hw/mlx5/main.c if (!rep) { rep 1315 drivers/infiniband/hw/mlx5/main.c err = mlx5_query_hca_vport_context(mdev, 0, port, 0, rep); rep 1319 drivers/infiniband/hw/mlx5/main.c props->lid = rep->lid; rep 1320 drivers/infiniband/hw/mlx5/main.c props->lmc = rep->lmc; rep 1321 drivers/infiniband/hw/mlx5/main.c props->sm_lid = rep->sm_lid; rep 1322 drivers/infiniband/hw/mlx5/main.c props->sm_sl = rep->sm_sl; rep 1323 drivers/infiniband/hw/mlx5/main.c props->state = rep->vport_state; rep 1324 drivers/infiniband/hw/mlx5/main.c props->phys_state = rep->port_physical_state; rep 1325 drivers/infiniband/hw/mlx5/main.c props->port_cap_flags = rep->cap_mask1; rep 1329 drivers/infiniband/hw/mlx5/main.c props->bad_pkey_cntr = rep->pkey_violation_counter; rep 1330 drivers/infiniband/hw/mlx5/main.c props->qkey_viol_cntr = rep->qkey_violation_counter; rep 1331 drivers/infiniband/hw/mlx5/main.c props->subnet_timeout = rep->subnet_timeout; rep 1332 drivers/infiniband/hw/mlx5/main.c props->init_type_reply = rep->init_type_reply; rep 1335 drivers/infiniband/hw/mlx5/main.c props->port_cap_flags2 = rep->cap_mask2; rep 1362 drivers/infiniband/hw/mlx5/main.c kfree(rep); rep 3485 drivers/infiniband/hw/mlx5/main.c struct mlx5_eswitch_rep *rep) rep 3496 drivers/infiniband/hw/mlx5/main.c rep->vport)); rep 3505 drivers/infiniband/hw/mlx5/main.c MLX5_SET(fte_match_set_misc, misc, source_port, rep->vport); rep 3569 drivers/infiniband/hw/mlx5/main.c struct mlx5_eswitch_rep *rep; rep 3571 drivers/infiniband/hw/mlx5/main.c rep = dev->port[flow_attr->port - 1].rep; rep 3572 drivers/infiniband/hw/mlx5/main.c if (!rep) { rep 3577 drivers/infiniband/hw/mlx5/main.c mlx5_ib_set_rule_source_port(dev, spec, rep); rep 5087 drivers/infiniband/hw/mlx5/main.c struct mlx5_hca_vport_context *rep) rep 5096 drivers/infiniband/hw/mlx5/main.c if (rep->grh_required) rep 5126 drivers/infiniband/hw/mlx5/main.c struct mlx5_hca_vport_context rep = {0}; rep 5135 drivers/infiniband/hw/mlx5/main.c &rep); rep 5142 drivers/infiniband/hw/mlx5/main.c immutable->core_cap_flags = get_core_cap_flags(ibdev, &rep); rep 755 drivers/infiniband/hw/mlx5/mlx5_ib.h struct mlx5_eswitch_rep *rep; rep 1313 drivers/infiniband/hw/ocrdma/ocrdma_sli.h struct ocrdma_mbx_rsp rep; rep 720 drivers/infiniband/sw/siw/siw_cm.c struct mpa_rr *rep; rep 733 drivers/infiniband/sw/siw/siw_cm.c rep = &cep->mpa.hdr; rep 735 drivers/infiniband/sw/siw/siw_cm.c if (__mpa_rr_revision(rep->params.bits) > MPA_REVISION_2) { rep 740 drivers/infiniband/sw/siw/siw_cm.c if (memcmp(rep->key, MPA_KEY_REP, 16)) { rep 747 drivers/infiniband/sw/siw/siw_cm.c if (rep->params.bits & MPA_RR_FLAG_REJECT) { rep 753 drivers/infiniband/sw/siw/siw_cm.c if (try_gso && rep->params.bits & MPA_RR_FLAG_GSO_EXP) { rep 757 drivers/infiniband/sw/siw/siw_cm.c if ((rep->params.bits & MPA_RR_FLAG_MARKERS) || rep 758 drivers/infiniband/sw/siw/siw_cm.c (mpa_crc_required && !(rep->params.bits & MPA_RR_FLAG_CRC)) || rep 760 drivers/infiniband/sw/siw/siw_cm.c (rep->params.bits & MPA_RR_FLAG_CRC))) { rep 762 drivers/infiniband/sw/siw/siw_cm.c rep->params.bits & MPA_RR_FLAG_CRC ? 1 : 0, rep 764 drivers/infiniband/sw/siw/siw_cm.c rep->params.bits & MPA_RR_FLAG_MARKERS ? 1 : 0, 0); rep 773 drivers/infiniband/sw/siw/siw_cm.c if (__mpa_rr_revision(rep->params.bits) < MPA_REVISION_2 || rep 774 drivers/infiniband/sw/siw/siw_cm.c !(rep->params.bits & MPA_RR_FLAG_ENHANCED)) { rep 780 drivers/infiniband/sw/siw/siw_cm.c __mpa_rr_revision(rep->params.bits), rep 781 drivers/infiniband/sw/siw/siw_cm.c rep->params.bits & MPA_RR_FLAG_ENHANCED ? rep 859 drivers/infiniband/sw/siw/siw_cm.c if (rep->params.bits & MPA_RR_FLAG_CRC) rep 427 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ib_cm_rep_param rep = {}; rep 432 drivers/infiniband/ulp/ipoib/ipoib_cm.c rep.private_data = &data; rep 433 drivers/infiniband/ulp/ipoib/ipoib_cm.c rep.private_data_len = sizeof(data); rep 434 drivers/infiniband/ulp/ipoib/ipoib_cm.c rep.flow_control = 0; rep 435 drivers/infiniband/ulp/ipoib/ipoib_cm.c rep.rnr_retry_count = req->rnr_retry_count; rep 436 drivers/infiniband/ulp/ipoib/ipoib_cm.c rep.srq = ipoib_cm_has_srq(dev); rep 437 drivers/infiniband/ulp/ipoib/ipoib_cm.c rep.qp_num = qp->qp_num; rep 438 drivers/infiniband/ulp/ipoib/ipoib_cm.c rep.starting_psn = psn; rep 439 drivers/infiniband/ulp/ipoib/ipoib_cm.c return ib_send_cm_rep(cm_id, &rep); rep 1056 drivers/input/evdev.c if (put_user(dev->rep[REP_DELAY], ip)) rep 1058 drivers/input/evdev.c if (put_user(dev->rep[REP_PERIOD], ip + 1)) rep 75 drivers/input/input.c dev->rep[REP_PERIOD] && dev->rep[REP_DELAY] && rep 79 drivers/input/input.c jiffies + msecs_to_jiffies(dev->rep[REP_DELAY])); rep 196 drivers/input/input.c if (dev->rep[REP_PERIOD]) rep 198 drivers/input/input.c msecs_to_jiffies(dev->rep[REP_PERIOD])); rep 347 drivers/input/input.c if (code <= REP_MAX && value >= 0 && dev->rep[code] != value) { rep 348 drivers/input/input.c dev->rep[code] = value; rep 1680 drivers/input/input.c dev->event(dev, EV_REP, REP_PERIOD, dev->rep[REP_PERIOD]); rep 1681 drivers/input/input.c dev->event(dev, EV_REP, REP_DELAY, dev->rep[REP_DELAY]); rep 2127 drivers/input/input.c dev->rep[REP_DELAY] = delay; rep 2128 drivers/input/input.c dev->rep[REP_PERIOD] = period; rep 2202 drivers/input/input.c if (!dev->rep[REP_DELAY] && !dev->rep[REP_PERIOD]) rep 506 drivers/input/keyboard/atkbd.c atkbd->time = jiffies + msecs_to_jiffies(dev->rep[REP_DELAY]) / 2; rep 545 drivers/input/keyboard/atkbd.c while (i < ARRAY_SIZE(period) - 1 && period[i] < dev->rep[REP_PERIOD]) rep 547 drivers/input/keyboard/atkbd.c dev->rep[REP_PERIOD] = period[i]; rep 549 drivers/input/keyboard/atkbd.c while (j < ARRAY_SIZE(delay) - 1 && delay[j] < dev->rep[REP_DELAY]) rep 551 drivers/input/keyboard/atkbd.c dev->rep[REP_DELAY] = delay[j]; rep 1097 drivers/input/keyboard/atkbd.c input_dev->rep[REP_DELAY] = 250; rep 1098 drivers/input/keyboard/atkbd.c input_dev->rep[REP_PERIOD] = 33; rep 226 drivers/input/keyboard/davinci_keyscan.c if (pdata->rep) rep 715 drivers/input/keyboard/gpio_keys.c pdata->rep = device_property_read_bool(dev, "autorepeat"); rep 822 drivers/input/keyboard/gpio_keys.c if (pdata->rep) rep 165 drivers/input/keyboard/gpio_keys_polled.c pdata->rep = device_property_present(dev, "autorepeat"); rep 281 drivers/input/keyboard/gpio_keys_polled.c if (pdata->rep) rep 243 drivers/input/keyboard/omap-keypad.c if (pdata->rep) rep 64 drivers/input/keyboard/spear-keyboard.c bool rep; rep 161 drivers/input/keyboard/spear-keyboard.c kbd->rep = true; rep 218 drivers/input/keyboard/spear-keyboard.c kbd->rep = pdata->rep; rep 247 drivers/input/keyboard/spear-keyboard.c if (kbd->rep) rep 247 drivers/input/keyboard/tca6416-keypad.c if (pdata->rep) rep 361 drivers/input/keyboard/twl4030_keypad.c kp->autorepeat = pdata->rep; rep 398 drivers/input/misc/ati_remote2.c ar2->jiffies = jiffies + msecs_to_jiffies(idev->rep[REP_DELAY]); rep 410 drivers/input/misc/ati_remote2.c ar2->jiffies = jiffies + msecs_to_jiffies(idev->rep[REP_PERIOD]); rep 605 drivers/input/misc/ati_remote2.c idev->rep[REP_DELAY] = 250; rep 606 drivers/input/misc/ati_remote2.c idev->rep[REP_PERIOD] = 33; rep 128 drivers/input/misc/soc_button_array.c gpio_keys_pdata->rep = autorepeat; rep 902 drivers/macintosh/adbhid.c input_dev->rep[REP_DELAY] = 500; /* input layer default: 250 */ rep 903 drivers/macintosh/adbhid.c input_dev->rep[REP_PERIOD] = 66; /* input layer default: 33 */ rep 500 drivers/media/i2c/ir-kbd-i2c.c int rep, i, l, p = 0, s, c = 0; rep 564 drivers/media/i2c/ir-kbd-i2c.c for (rep = c / 3; rep >= 1; rep--) { rep 565 drivers/media/i2c/ir-kbd-i2c.c if (!memcmp(&codes[c - rep * 3], &codes[c - rep * 2], rep) && rep 566 drivers/media/i2c/ir-kbd-i2c.c !cmp_no_trail(&codes[c - rep], &codes[c - rep * 2], rep)) { rep 574 drivers/media/i2c/ir-kbd-i2c.c int leading = c - rep * 3; rep 576 drivers/media/i2c/ir-kbd-i2c.c if (leading >= ARRAY_SIZE(code_block->codes) - 3 - rep) { rep 581 drivers/media/i2c/ir-kbd-i2c.c dev_dbg(&rcdev->dev, "found trailing %d repeat\n", rep); rep 585 drivers/media/i2c/ir-kbd-i2c.c rep); rep 586 drivers/media/i2c/ir-kbd-i2c.c c = leading + 1 + rep; rep 650 drivers/media/platform/sh_veu.c u32 *mant, u32 *frac, u32 *rep) rep 655 drivers/media/platform/sh_veu.c *rep = *mant = *frac = 0; rep 671 drivers/media/platform/sh_veu.c *rep = 1; rep 674 drivers/media/platform/sh_veu.c *rep = 3; rep 677 drivers/media/platform/sh_veu.c *rep = 7; rep 680 drivers/media/platform/sh_veu.c if (*rep) rep 704 drivers/media/platform/sh_veu.c u32 mant, frac, value, rep; rep 706 drivers/media/platform/sh_veu.c sh_veu_calc_scale(veu, size_in, size_out, crop_out, &mant, &frac, &rep); rep 716 drivers/media/platform/sh_veu.c (((rep << 12) | crop_out) << 16); rep 726 drivers/media/platform/sh_veu.c u32 mant, frac, value, rep; rep 728 drivers/media/platform/sh_veu.c sh_veu_calc_scale(veu, size_in, size_out, crop_out, &mant, &frac, &rep); rep 738 drivers/media/platform/sh_veu.c (rep << 12) | crop_out; rep 1658 drivers/media/rc/imon.c if (msec < ictx->idev->rep[REP_DELAY]) { rep 682 drivers/media/rc/rc-main.c if (input->rep[REP_PERIOD]) rep 684 drivers/media/rc/rc-main.c msecs_to_jiffies(input->rep[REP_PERIOD])); rep 791 drivers/media/rc/rc-main.c dev->input_dev->rep[REP_PERIOD] && rep 792 drivers/media/rc/rc-main.c !dev->input_dev->rep[REP_DELAY]) { rep 795 drivers/media/rc/rc-main.c msecs_to_jiffies(dev->input_dev->rep[REP_PERIOD])); rep 1815 drivers/media/rc/rc-main.c dev->input_dev->rep[REP_DELAY] = 0; rep 1817 drivers/media/rc/rc-main.c dev->input_dev->rep[REP_DELAY] = 500; rep 1824 drivers/media/rc/rc-main.c dev->input_dev->rep[REP_PERIOD] = 125; rep 219 drivers/media/usb/dvb-usb/dvb-usb-remote.c input_dev->rep[REP_PERIOD] = d->props.rc.legacy.rc_interval; rep 220 drivers/media/usb/dvb-usb/dvb-usb-remote.c input_dev->rep[REP_DELAY] = d->props.rc.legacy.rc_interval + 150; rep 139 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 144 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c err = mlx5_eswitch_get_vport_stats(esw, rep->vport, &vf_stats); rep 146 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c pr_warn("vport %d error %d reading stats\n", rep->vport, err); rep 242 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 246 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport, rep 409 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep) rep 417 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rpriv = mlx5e_rep_to_rep_priv(rep); rep 426 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep, rep 438 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rpriv = mlx5e_rep_to_rep_priv(rep); rep 448 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport, rep 461 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c mlx5e_sqs2vport_stop(esw, rep); rep 469 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 485 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c err = mlx5e_sqs2vport_start(esw, rep, sqs, num_sqs); rep 498 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 500 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c mlx5e_sqs2vport_stop(esw, rep); rep 1158 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 1168 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport, 1, rep 1181 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 1187 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport, 1, rep 1269 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep; rep 1277 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep = rpriv->rep; rep 1278 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c return (rep->vport == MLX5_VPORT_UPLINK); rep 1410 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 1423 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rep->vport == MLX5_VPORT_UPLINK) rep 1448 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 1451 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rep->vport == MLX5_VPORT_UPLINK) { rep 1480 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rep->vport == MLX5_VPORT_UPLINK) rep 1541 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 1548 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport, rep 1635 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rpriv->rep->vport == MLX5_VPORT_UPLINK) { rep 1673 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rpriv->rep->vport == MLX5_VPORT_UPLINK) { rep 1799 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c return rpriv->rep->vport == MLX5_VPORT_UPLINK || rep 1800 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rpriv->rep->vport == MLX5_VPORT_PF || rep 1801 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c mlx5_eswitch_is_vf_vport(dev->priv.eswitch, rpriv->rep->vport); rep 1814 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 1825 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rep->vport == MLX5_VPORT_UPLINK) { rep 1830 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c dl_port_index = vport_to_devlink_port_index(dev, rep->vport); rep 1831 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c } else if (rep->vport == MLX5_VPORT_PF) { rep 1835 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c dl_port_index = rep->vport; rep 1837 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rpriv->rep->vport)) { rep 1840 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c pfnum, rep->vport - 1); rep 1841 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c dl_port_index = vport_to_devlink_port_index(dev, rep->vport); rep 1856 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c mlx5e_vport_rep_load(struct mlx5_core_dev *dev, struct mlx5_eswitch_rep *rep) rep 1868 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rpriv->rep = rep; rep 1871 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c profile = (rep->vport == MLX5_VPORT_UPLINK) ? rep 1876 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport); rep 1882 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->rep_data[REP_ETH].priv = rpriv; rep 1885 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rep->vport == MLX5_VPORT_UPLINK) { rep 1894 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport); rep 1901 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport); rep 1908 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport); rep 1915 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rep->vport); rep 1933 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rep->vport == MLX5_VPORT_UPLINK) rep 1943 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c mlx5e_vport_rep_unload(struct mlx5_eswitch_rep *rep) rep 1945 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c struct mlx5e_rep_priv *rpriv = mlx5e_rep_to_rep_priv(rep); rep 1957 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c if (rep->vport == MLX5_VPORT_UPLINK) rep 1963 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c static void *mlx5e_vport_rep_get_proto_dev(struct mlx5_eswitch_rep *rep) rep 1967 drivers/net/ethernet/mellanox/mlx5/core/en_rep.c rpriv = mlx5e_rep_to_rep_priv(rep); rep 86 drivers/net/ethernet/mellanox/mlx5/core/en_rep.h struct mlx5_eswitch_rep *rep; rep 97 drivers/net/ethernet/mellanox/mlx5/core/en_rep.h struct mlx5e_rep_priv *mlx5e_rep_to_rep_priv(struct mlx5_eswitch_rep *rep) rep 99 drivers/net/ethernet/mellanox/mlx5/core/en_rep.h return rep->rep_data[REP_ETH].priv; rep 1194 drivers/net/ethernet/mellanox/mlx5/core/en_rx.c struct mlx5_eswitch_rep *rep = rpriv->rep; rep 1224 drivers/net/ethernet/mellanox/mlx5/core/en_rx.c if (rep->vlan && skb_vlan_tag_present(skb)) rep 1197 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c attr->dests[out_index].rep = rpriv->rep; rep 2163 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c struct mlx5_eswitch_rep *rep; rep 2177 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c rep = rpriv->rep; rep 2178 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (rep->vport != MLX5_VPORT_UPLINK && rep 3327 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c attr->dests[attr->out_count].rep = rpriv->rep; rep 3696 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c struct mlx5_eswitch_rep *in_rep = rpriv->rep; rep 3955 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c vport_num = rpriv->rep->vport; rep 388 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h struct mlx5_eswitch_rep *rep; rep 548 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h #define mlx5_esw_for_all_reps(esw, i, rep) \ rep 550 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h (rep) = &(esw)->offloads.vport_reps[i], \ rep 553 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h #define mlx5_esw_for_each_vf_rep(esw, i, rep, nvfs) \ rep 555 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h (rep) = &(esw)->offloads.vport_reps[i], \ rep 558 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h #define mlx5_esw_for_each_vf_rep_reverse(esw, i, rep, nvfs) \ rep 560 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h (rep) = &(esw)->offloads.vport_reps[i], \ rep 570 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h #define mlx5_esw_for_each_host_func_rep(esw, i, rep, nvfs) \ rep 572 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h (rep) = &(esw)->offloads.vport_reps[i], \ rep 575 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h #define mlx5_esw_for_each_host_func_rep_reverse(esw, i, rep, nvfs) \ rep 577 drivers/net/ethernet/mellanox/mlx5/core/eswitch.h (rep) = &(esw)->offloads.vport_reps[i], \ rep 185 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c dest[i].vport.num = attr->dests[j].rep->vport; rep 272 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c dest[i].vport.num = attr->dests[i].rep->vport; rep 354 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 358 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_esw_for_each_host_func_rep(esw, i, rep, esw->esw_funcs.num_vfs) { rep 359 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (atomic_read(&rep->rep_data[REP_ETH].state) != REP_LOADED) rep 362 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c err = __mlx5_eswitch_set_vport_vlan(esw, rep->vport, 0, 0, val); rep 377 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c out_rep = attr->dests[0].rep; rep 398 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c out_rep = attr->dests[0].rep; rep 452 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (attr->dests[0].rep->vport == MLX5_VPORT_UPLINK) { rep 516 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (attr->dests[0].rep->vport == MLX5_VPORT_UPLINK) rep 1403 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 1413 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_esw_for_all_reps(esw, vport_index, rep) { rep 1414 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep->vport = mlx5_eswitch_index_to_vport_num(esw, vport_index); rep 1415 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep->vport_index = vport_index; rep 1418 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c atomic_set(&rep->rep_data[rep_type].state, rep 1426 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep, u8 rep_type) rep 1428 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (atomic_cmpxchg(&rep->rep_data[rep_type].state, rep 1430 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw->offloads.rep_ops[rep_type]->unload(rep); rep 1435 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 1438 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_ECPF); rep 1439 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c __esw_offloads_unload_rep(esw, rep, rep_type); rep 1443 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_PF); rep 1444 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c __esw_offloads_unload_rep(esw, rep, rep_type); rep 1447 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_UPLINK); rep 1448 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c __esw_offloads_unload_rep(esw, rep, rep_type); rep 1454 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 1457 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_esw_for_each_vf_rep_reverse(esw, i, rep, nvports) rep 1458 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c __esw_offloads_unload_rep(esw, rep, rep_type); rep 1486 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep, u8 rep_type) rep 1490 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (atomic_cmpxchg(&rep->rep_data[rep_type].state, rep 1492 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c err = esw->offloads.rep_ops[rep_type]->load(esw->dev, rep); rep 1494 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c atomic_set(&rep->rep_data[rep_type].state, rep 1503 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 1506 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_UPLINK); rep 1507 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c err = __esw_offloads_load_rep(esw, rep, rep_type); rep 1512 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_PF); rep 1513 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c err = __esw_offloads_load_rep(esw, rep, rep_type); rep 1519 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_ECPF); rep 1520 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c err = __esw_offloads_load_rep(esw, rep, rep_type); rep 1529 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_PF); rep 1530 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c __esw_offloads_unload_rep(esw, rep, rep_type); rep 1534 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_UPLINK); rep 1535 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c __esw_offloads_unload_rep(esw, rep, rep_type); rep 1542 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 1545 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_esw_for_each_vf_rep(esw, i, rep, nvports) { rep 1546 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c err = __esw_offloads_load_rep(esw, rep, rep_type); rep 2531 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 2535 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_esw_for_all_reps(esw, i, rep) { rep 2536 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep_data = &rep->rep_data[rep_type]; rep 2544 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 2550 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_esw_for_all_reps(esw, i, rep) rep 2551 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c atomic_set(&rep->rep_data[rep_type].state, REP_UNREGISTERED); rep 2557 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 2559 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, MLX5_VPORT_UPLINK); rep 2560 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c return rep->rep_data[rep_type].priv; rep 2567 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_eswitch_rep *rep; rep 2569 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c rep = mlx5_eswitch_get_rep(esw, vport); rep 2571 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (atomic_read(&rep->rep_data[rep_type].state) == REP_LOADED && rep 2573 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c return esw->offloads.rep_ops[rep_type]->get_proto_dev(rep); rep 664 drivers/net/ethernet/mellanox/mlx5/core/vport.c struct mlx5_hca_vport_context *rep) rep 699 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->field_select = MLX5_GET_PR(hca_vport_context, ctx, field_select); rep 700 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->sm_virt_aware = MLX5_GET_PR(hca_vport_context, ctx, sm_virt_aware); rep 701 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->has_smi = MLX5_GET_PR(hca_vport_context, ctx, has_smi); rep 702 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->has_raw = MLX5_GET_PR(hca_vport_context, ctx, has_raw); rep 703 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->policy = MLX5_GET_PR(hca_vport_context, ctx, vport_state_policy); rep 704 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->phys_state = MLX5_GET_PR(hca_vport_context, ctx, rep 706 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->vport_state = MLX5_GET_PR(hca_vport_context, ctx, vport_state); rep 707 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->port_physical_state = MLX5_GET_PR(hca_vport_context, ctx, rep 709 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->port_guid = MLX5_GET64_PR(hca_vport_context, ctx, port_guid); rep 710 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->node_guid = MLX5_GET64_PR(hca_vport_context, ctx, node_guid); rep 711 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->cap_mask1 = MLX5_GET_PR(hca_vport_context, ctx, cap_mask1); rep 712 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->cap_mask1_perm = MLX5_GET_PR(hca_vport_context, ctx, rep 714 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->cap_mask2 = MLX5_GET_PR(hca_vport_context, ctx, cap_mask2); rep 715 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->cap_mask2_perm = MLX5_GET_PR(hca_vport_context, ctx, rep 717 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->lid = MLX5_GET_PR(hca_vport_context, ctx, lid); rep 718 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->init_type_reply = MLX5_GET_PR(hca_vport_context, ctx, rep 720 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->lmc = MLX5_GET_PR(hca_vport_context, ctx, lmc); rep 721 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->subnet_timeout = MLX5_GET_PR(hca_vport_context, ctx, rep 723 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->sm_lid = MLX5_GET_PR(hca_vport_context, ctx, sm_lid); rep 724 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->sm_sl = MLX5_GET_PR(hca_vport_context, ctx, sm_sl); rep 725 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->qkey_violation_counter = MLX5_GET_PR(hca_vport_context, ctx, rep 727 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->pkey_violation_counter = MLX5_GET_PR(hca_vport_context, ctx, rep 729 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->grh_required = MLX5_GET_PR(hca_vport_context, ctx, grh_required); rep 730 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep->sys_image_guid = MLX5_GET64_PR(hca_vport_context, ctx, rep 742 drivers/net/ethernet/mellanox/mlx5/core/vport.c struct mlx5_hca_vport_context *rep; rep 745 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep = kzalloc(sizeof(*rep), GFP_KERNEL); rep 746 drivers/net/ethernet/mellanox/mlx5/core/vport.c if (!rep) rep 749 drivers/net/ethernet/mellanox/mlx5/core/vport.c err = mlx5_query_hca_vport_context(dev, 0, 1, 0, rep); rep 751 drivers/net/ethernet/mellanox/mlx5/core/vport.c *sys_image_guid = rep->sys_image_guid; rep 753 drivers/net/ethernet/mellanox/mlx5/core/vport.c kfree(rep); rep 761 drivers/net/ethernet/mellanox/mlx5/core/vport.c struct mlx5_hca_vport_context *rep; rep 764 drivers/net/ethernet/mellanox/mlx5/core/vport.c rep = kzalloc(sizeof(*rep), GFP_KERNEL); rep 765 drivers/net/ethernet/mellanox/mlx5/core/vport.c if (!rep) rep 768 drivers/net/ethernet/mellanox/mlx5/core/vport.c err = mlx5_query_hca_vport_context(dev, 0, 1, 0, rep); rep 770 drivers/net/ethernet/mellanox/mlx5/core/vport.c *node_guid = rep->node_guid; rep 772 drivers/net/ethernet/mellanox/mlx5/core/vport.c kfree(rep); rep 1601 drivers/net/wireless/ath/ath5k/eeprom.c struct ath5k_edge_power *rep; rep 1625 drivers/net/wireless/ath/ath5k/eeprom.c rep = ee->ee_ctl_pwr; rep 1641 drivers/net/wireless/ath/ath5k/eeprom.c rep += AR5K_EEPROM_N_EDGES; rep 1647 drivers/net/wireless/ath/ath5k/eeprom.c rep[j].freq = (val >> 8) & fmask; rep 1648 drivers/net/wireless/ath/ath5k/eeprom.c rep[j + 1].freq = val & fmask; rep 1652 drivers/net/wireless/ath/ath5k/eeprom.c rep[j].edge = (val >> 8) & pmask; rep 1653 drivers/net/wireless/ath/ath5k/eeprom.c rep[j].flag = (val >> 14) & 1; rep 1654 drivers/net/wireless/ath/ath5k/eeprom.c rep[j + 1].edge = val & pmask; rep 1655 drivers/net/wireless/ath/ath5k/eeprom.c rep[j + 1].flag = (val >> 6) & 1; rep 1659 drivers/net/wireless/ath/ath5k/eeprom.c rep[0].freq = (val >> 9) & fmask; rep 1660 drivers/net/wireless/ath/ath5k/eeprom.c rep[1].freq = (val >> 2) & fmask; rep 1661 drivers/net/wireless/ath/ath5k/eeprom.c rep[2].freq = (val << 5) & fmask; rep 1664 drivers/net/wireless/ath/ath5k/eeprom.c rep[2].freq |= (val >> 11) & 0x1f; rep 1665 drivers/net/wireless/ath/ath5k/eeprom.c rep[3].freq = (val >> 4) & fmask; rep 1666 drivers/net/wireless/ath/ath5k/eeprom.c rep[4].freq = (val << 3) & fmask; rep 1669 drivers/net/wireless/ath/ath5k/eeprom.c rep[4].freq |= (val >> 13) & 0x7; rep 1670 drivers/net/wireless/ath/ath5k/eeprom.c rep[5].freq = (val >> 6) & fmask; rep 1671 drivers/net/wireless/ath/ath5k/eeprom.c rep[6].freq = (val << 1) & fmask; rep 1674 drivers/net/wireless/ath/ath5k/eeprom.c rep[6].freq |= (val >> 15) & 0x1; rep 1675 drivers/net/wireless/ath/ath5k/eeprom.c rep[7].freq = (val >> 8) & fmask; rep 1677 drivers/net/wireless/ath/ath5k/eeprom.c rep[0].edge = (val >> 2) & pmask; rep 1678 drivers/net/wireless/ath/ath5k/eeprom.c rep[1].edge = (val << 4) & pmask; rep 1681 drivers/net/wireless/ath/ath5k/eeprom.c rep[1].edge |= (val >> 12) & 0xf; rep 1682 drivers/net/wireless/ath/ath5k/eeprom.c rep[2].edge = (val >> 6) & pmask; rep 1683 drivers/net/wireless/ath/ath5k/eeprom.c rep[3].edge = val & pmask; rep 1686 drivers/net/wireless/ath/ath5k/eeprom.c rep[4].edge = (val >> 10) & pmask; rep 1687 drivers/net/wireless/ath/ath5k/eeprom.c rep[5].edge = (val >> 4) & pmask; rep 1688 drivers/net/wireless/ath/ath5k/eeprom.c rep[6].edge = (val << 2) & pmask; rep 1691 drivers/net/wireless/ath/ath5k/eeprom.c rep[6].edge |= (val >> 14) & 0x3; rep 1692 drivers/net/wireless/ath/ath5k/eeprom.c rep[7].edge = (val >> 8) & pmask; rep 1695 drivers/net/wireless/ath/ath5k/eeprom.c rep[j].freq = ath5k_eeprom_bin2freq(ee, rep 1696 drivers/net/wireless/ath/ath5k/eeprom.c rep[j].freq, ctl_mode); rep 1698 drivers/net/wireless/ath/ath5k/eeprom.c rep += AR5K_EEPROM_N_EDGES; rep 2769 drivers/net/wireless/ath/ath5k/phy.c struct ath5k_edge_power *rep = ee->ee_ctl_pwr; rep 2825 drivers/net/wireless/ath/ath5k/phy.c if (target <= rep[rep_idx].freq) rep 2826 drivers/net/wireless/ath/ath5k/phy.c edge_pwr = (s16) rep[rep_idx].edge; rep 407 drivers/net/wireless/ath/ath9k/eeprom_4k.c struct cal_ctl_data_4k *rep; rep 477 drivers/net/wireless/ath/ath9k/eeprom_4k.c rep = &(pEepData->ctlData[i]); rep 481 drivers/net/wireless/ath/ath9k/eeprom_4k.c rep->ctlEdges[ rep 508 drivers/net/wireless/ath/ath9k/eeprom_9287.c struct cal_ctl_data_ar9287 *rep; rep 595 drivers/net/wireless/ath/ath9k/eeprom_9287.c rep = &(pEepData->ctlData[i]); rep 597 drivers/net/wireless/ath/ath9k/eeprom_9287.c rep->ctlEdges[ar5416_get_ntxchains(tx_chainmask) - 1]; rep 929 drivers/net/wireless/ath/ath9k/eeprom_def.c struct cal_ctl_data *rep; rep 1041 drivers/net/wireless/ath/ath9k/eeprom_def.c rep = &(pEepData->ctlData[i]); rep 1044 drivers/net/wireless/ath/ath9k/eeprom_def.c rep->ctlEdges[ar5416_get_ntxchains(tx_chainmask) - 1], rep 120 drivers/platform/x86/pcengines-apuv2.c .rep = 0, rep 134 drivers/platform/x86/surface3_button.c gpio_keys_pdata->rep = autorepeat; rep 1727 drivers/s390/net/qeth_l2_main.c struct qeth_ipacmd_vnicc *rep = &cmd->data.vnicc; rep 1734 drivers/s390/net/qeth_l2_main.c card->options.vnicc.sup_chars = rep->vnicc_cmds.supported; rep 1735 drivers/s390/net/qeth_l2_main.c card->options.vnicc.cur_chars = rep->vnicc_cmds.enabled; rep 1738 drivers/s390/net/qeth_l2_main.c *cbctl->result.sup_cmds = rep->data.query_cmds.sup_cmds; rep 1740 drivers/s390/net/qeth_l2_main.c *cbctl->result.timeout = rep->data.getset_timeout.timeout; rep 1529 drivers/staging/isdn/gigaset/ev-layer.c struct reply_t *rep; rep 1562 drivers/staging/isdn/gigaset/ev-layer.c rep = at_state->replystruct; rep 1597 drivers/staging/isdn/gigaset/ev-layer.c for (;; rep++) { rep 1598 drivers/staging/isdn/gigaset/ev-layer.c rcode = rep->resp_code; rep 1607 drivers/staging/isdn/gigaset/ev-layer.c && ((int) at_state->ConState >= rep->min_ConState) rep 1608 drivers/staging/isdn/gigaset/ev-layer.c && (rep->max_ConState < 0 rep 1609 drivers/staging/isdn/gigaset/ev-layer.c || (int) at_state->ConState <= rep->max_ConState) rep 1610 drivers/staging/isdn/gigaset/ev-layer.c && (rep->parameter < 0 || rep->parameter == ev->parameter)) rep 1614 drivers/staging/isdn/gigaset/ev-layer.c p_command = rep->command; rep 1620 drivers/staging/isdn/gigaset/ev-layer.c do_action(rep->action[curact], cs, bcs, &at_state, &p_command, rep 1628 drivers/staging/isdn/gigaset/ev-layer.c if (rep->new_ConState >= 0) rep 1629 drivers/staging/isdn/gigaset/ev-layer.c at_state->ConState = rep->new_ConState; rep 1648 drivers/staging/isdn/gigaset/ev-layer.c if (!rep->timeout) { rep 1651 drivers/staging/isdn/gigaset/ev-layer.c } else if (rep->timeout > 0) { /* new timeout */ rep 1652 drivers/staging/isdn/gigaset/ev-layer.c at_state->timer_expires = rep->timeout * 10; rep 136 drivers/tty/vt/keyboard.c static char rep; /* flag telling character repeat */ rep 290 drivers/tty/vt/keyboard.c rpt[1].delay = dev->rep[REP_DELAY]; rep 291 drivers/tty/vt/keyboard.c rpt[1].period = dev->rep[REP_PERIOD]; rep 464 drivers/tty/vt/keyboard.c if (rep) rep 472 drivers/tty/vt/keyboard.c if (rep) rep 490 drivers/tty/vt/keyboard.c if (rep || !tty) rep 520 drivers/tty/vt/keyboard.c if (!rep) rep 824 drivers/tty/vt/keyboard.c if (rep) rep 899 drivers/tty/vt/keyboard.c if (up_flag || rep) rep 908 drivers/tty/vt/keyboard.c if (up_flag || rep) rep 1378 drivers/tty/vt/keyboard.c rep = (down == 2); rep 1419 drivers/tty/vt/keyboard.c if (rep && rep 290 fs/cachefiles/namei.c struct dentry *rep, rep 299 fs/cachefiles/namei.c _enter(",'%pd','%pd'", dir, rep); rep 301 fs/cachefiles/namei.c _debug("remove %p from %p", rep, dir); rep 304 fs/cachefiles/namei.c if (!d_is_dir(rep)) { rep 309 fs/cachefiles/namei.c ret = security_path_unlink(&path, rep); rep 313 fs/cachefiles/namei.c trace_cachefiles_unlink(object, rep, why); rep 314 fs/cachefiles/namei.c ret = vfs_unlink(d_inode(dir), rep, NULL); rep 317 fs/cachefiles/namei.c cachefiles_mark_object_buried(cache, rep, why); rep 343 fs/cachefiles/namei.c if (rep->d_parent != dir || IS_DEADDIR(d_inode(rep))) { rep 357 fs/cachefiles/namei.c if (trap == rep) { rep 363 fs/cachefiles/namei.c if (d_mountpoint(rep)) { rep 411 fs/cachefiles/namei.c ret = security_path_rename(&path, rep, &path_to_graveyard, grave, 0); rep 415 fs/cachefiles/namei.c trace_cachefiles_rename(object, rep, grave, why); rep 416 fs/cachefiles/namei.c ret = vfs_rename(d_inode(dir), rep, rep 423 fs/cachefiles/namei.c cachefiles_mark_object_buried(cache, rep, why); rep 54 include/linux/gpio_keys.h unsigned int rep:1; rep 679 include/linux/hid.h #define HID_REPORT_ID(rep) \ rep 680 include/linux/hid.h .report_type = (rep) rep 886 include/linux/hid.h int __hid_request(struct hid_device *hid, struct hid_report *rep, int reqtype); rep 168 include/linux/input.h int rep[REP_CNT]; rep 611 include/linux/mfd/twl.h bool rep; rep 34 include/linux/mlx5/eswitch.h int (*load)(struct mlx5_core_dev *dev, struct mlx5_eswitch_rep *rep); rep 35 include/linux/mlx5/eswitch.h void (*unload)(struct mlx5_eswitch_rep *rep); rep 36 include/linux/mlx5/eswitch.h void *(*get_proto_dev)(struct mlx5_eswitch_rep *rep); rep 97 include/linux/mlx5/vport.h struct mlx5_hca_vport_context *rep); rep 159 include/linux/platform_data/keyboard-spear.h bool rep; rep 19 include/linux/platform_data/keypad-omap.h bool rep; rep 22 include/linux/platform_data/keyscan-davinci.h u8 rep:1; rep 25 include/linux/tca6416_keypad.h unsigned int rep:1; /* enable input subsystem auto repeat */ rep 32 include/linux/vt_kern.h extern int kbd_rate(struct kbd_repeat *rep); rep 23 include/trace/events/rpcrdma.h const struct rpcrdma_rep *rep rep 26 include/trace/events/rpcrdma.h TP_ARGS(rep), rep 29 include/trace/events/rpcrdma.h __field(const void *, rep) rep 37 include/trace/events/rpcrdma.h __entry->rep = rep; rep 38 include/trace/events/rpcrdma.h __entry->r_xprt = rep->rr_rxprt; rep 39 include/trace/events/rpcrdma.h __entry->xid = be32_to_cpu(rep->rr_xid); rep 40 include/trace/events/rpcrdma.h __entry->version = be32_to_cpu(rep->rr_vers); rep 41 include/trace/events/rpcrdma.h __entry->proc = be32_to_cpu(rep->rr_proc); rep 45 include/trace/events/rpcrdma.h __entry->r_xprt, __entry->xid, __entry->rep, rep 53 include/trace/events/rpcrdma.h const struct rpcrdma_rep *rep \ rep 55 include/trace/events/rpcrdma.h TP_ARGS(rep)) rep 277 include/trace/events/rpcrdma.h __field(const void *, rep) rep 285 include/trace/events/rpcrdma.h __entry->rep = rpcr_to_rdmar(rqst)->rl_reply; rep 290 include/trace/events/rpcrdma.h __entry->xid, __entry->rqst, __entry->req, __entry->rep rep 687 include/trace/events/rpcrdma.h const struct rpcrdma_rep *rep rep 690 include/trace/events/rpcrdma.h TP_ARGS(rep), rep 693 include/trace/events/rpcrdma.h __field(const void *, rep) rep 697 include/trace/events/rpcrdma.h __entry->rep = rep; rep 701 include/trace/events/rpcrdma.h __entry->rep rep 804 include/trace/events/rpcrdma.h __field(const void *, rep) rep 811 include/trace/events/rpcrdma.h __entry->rep = container_of(wc->wr_cqe, struct rpcrdma_rep, rep 824 include/trace/events/rpcrdma.h __entry->rep, __entry->byte_len, rep 983 include/trace/events/rpcrdma.h const struct rpcrdma_rep *rep, rep 988 include/trace/events/rpcrdma.h TP_ARGS(task, rep, req, credits), rep 993 include/trace/events/rpcrdma.h __field(const void *, rep) rep 1002 include/trace/events/rpcrdma.h __entry->rep = rep; rep 1004 include/trace/events/rpcrdma.h __entry->xid = be32_to_cpu(rep->rr_xid); rep 1010 include/trace/events/rpcrdma.h __entry->credits, __entry->rep, __entry->req rep 1016 include/trace/events/rpcrdma.h const struct rpcrdma_rep *rep rep 1019 include/trace/events/rpcrdma.h TP_ARGS(rep), rep 1024 include/trace/events/rpcrdma.h __field(const void *, rep) rep 1029 include/trace/events/rpcrdma.h __entry->task_id = rep->rr_rqst->rq_task->tk_pid; rep 1030 include/trace/events/rpcrdma.h __entry->client_id = rep->rr_rqst->rq_task->tk_client->cl_clid; rep 1031 include/trace/events/rpcrdma.h __entry->rep = rep; rep 1032 include/trace/events/rpcrdma.h __entry->xid = be32_to_cpu(rep->rr_xid); rep 1037 include/trace/events/rpcrdma.h __entry->rep rep 1185 include/trace/events/rpcrdma.h __field(const void *, rep) rep 1192 include/trace/events/rpcrdma.h __entry->rep = req->rl_reply; rep 1197 include/trace/events/rpcrdma.h __entry->req, __entry->rep rep 1239 include/trace/events/rpcrdma.h const struct rpcrdma_rep *rep rep 1242 include/trace/events/rpcrdma.h TP_ARGS(rqst, rep), rep 1248 include/trace/events/rpcrdma.h __field(const void *, rep) rep 1255 include/trace/events/rpcrdma.h __entry->rep = rep; rep 1260 include/trace/events/rpcrdma.h __entry->rep rep 243 include/uapi/drm/vmwgfx_drm.h struct drm_vmw_surface_arg rep; rep 271 include/uapi/drm/vmwgfx_drm.h struct drm_vmw_surface_create_req rep; rep 430 include/uapi/drm/vmwgfx_drm.h struct drm_vmw_bo_rep rep; rep 962 include/uapi/drm/vmwgfx_drm.h struct drm_vmw_gb_surface_create_rep rep; rep 1004 include/uapi/drm/vmwgfx_drm.h struct drm_vmw_gb_surface_ref_rep rep; rep 1089 include/uapi/drm/vmwgfx_drm.h struct drm_vmw_context_arg rep; rep 1167 include/uapi/drm/vmwgfx_drm.h struct drm_vmw_gb_surface_create_rep rep; rep 1210 include/uapi/drm/vmwgfx_drm.h struct drm_vmw_gb_surface_ref_ext_rep rep; rep 282 lib/842/842_decompress.c u64 op, rep, tmp, bytes, total; rep 305 lib/842/842_decompress.c ret = next_bits(&p, &rep, REPEAT_BITS); rep 313 lib/842/842_decompress.c rep++; rep 315 lib/842/842_decompress.c if (rep * 8 > p.olen) rep 318 lib/842/842_decompress.c while (rep-- > 0) { rep 710 lib/bch.c const struct gf_poly *a, int *rep) rep 716 lib/bch.c rep[i] = a->c[i] ? mod_s(bch, a_log(bch, a->c[i])+l) : -1; rep 723 lib/bch.c const struct gf_poly *b, int *rep) rep 733 lib/bch.c if (!rep) { rep 734 lib/bch.c rep = bch->cache; rep 735 lib/bch.c gf_poly_logrep(bch, b, rep); rep 743 lib/bch.c m = rep[i]; rep 66 lib/zstd/compress.c U32 rep[ZSTD_REP_NUM]; rep 231 lib/zstd/compress.c cctx->rep[i] = repStartValue[i]; rep 303 lib/zstd/compress.c zc->rep[i] = repStartValue[i]; rep 341 lib/zstd/compress.c cctx->rep[i] = 0; rep 852 lib/zstd/compress.c zc->rep[i] = zc->repToConfirm[i]; rep 1021 lib/zstd/compress.c U32 offset_1 = cctx->rep[0], offset_2 = cctx->rep[1]; rep 1133 lib/zstd/compress.c U32 offset_1 = ctx->rep[0], offset_2 = ctx->rep[1]; rep 1269 lib/zstd/compress.c U32 offset_1 = cctx->rep[0], offset_2 = cctx->rep[1]; rep 1414 lib/zstd/compress.c U32 offset_1 = ctx->rep[0], offset_2 = ctx->rep[1]; rep 1912 lib/zstd/compress.c U32 offset_1 = ctx->rep[0], offset_2 = ctx->rep[1], savedOffset = 0; rep 2080 lib/zstd/compress.c U32 offset_1 = ctx->rep[0], offset_2 = ctx->rep[1]; rep 2690 lib/zstd/compress.c cctx->rep[0] = ZSTD_readLE32(dictPtr + 0); rep 2691 lib/zstd/compress.c cctx->rep[1] = ZSTD_readLE32(dictPtr + 4); rep 2692 lib/zstd/compress.c cctx->rep[2] = ZSTD_readLE32(dictPtr + 8); rep 2708 lib/zstd/compress.c if (cctx->rep[u] == 0) rep 2710 lib/zstd/compress.c if (cctx->rep[u] > dictContentSize) rep 74 lib/zstd/decompress.c U32 rep[ZSTD_REP_NUM]; rep 117 lib/zstd/decompress.c ZSTD_STATIC_ASSERT(sizeof(dctx->entropy.rep) == sizeof(repStartValue)); rep 118 lib/zstd/decompress.c memcpy(dctx->entropy.rep, repStartValue, sizeof(repStartValue)); /* initial repcodes */ rep 1122 lib/zstd/decompress.c seqState.prevOffset[i] = dctx->entropy.rep[i]; rep 1147 lib/zstd/decompress.c dctx->entropy.rep[i] = (U32)(seqState.prevOffset[i]); rep 1385 lib/zstd/decompress.c seqState.prevOffset[i] = dctx->entropy.rep[i]; rep 1429 lib/zstd/decompress.c dctx->entropy.rep[i] = (U32)(seqState.prevOffset[i]); rep 1932 lib/zstd/decompress.c U32 const rep = ZSTD_readLE32(dictPtr); rep 1934 lib/zstd/decompress.c if (rep == 0 || rep >= dictContentSize) rep 1936 lib/zstd/decompress.c entropy->rep[i] = rep; rep 2011 lib/zstd/decompress.c dstDCtx->entropy.rep[0] = ddict->entropy.rep[0]; rep 2012 lib/zstd/decompress.c dstDCtx->entropy.rep[1] = ddict->entropy.rep[1]; rep 2013 lib/zstd/decompress.c dstDCtx->entropy.rep[2] = ddict->entropy.rep[2]; rep 169 lib/zstd/zstd_internal.h U32 rep[ZSTD_REP_NUM]; rep 426 lib/zstd/zstd_opt.h U32 offset, rep[ZSTD_REP_NUM]; rep 435 lib/zstd/zstd_opt.h rep[i] = ctx->rep[i]; rep 450 lib/zstd/zstd_opt.h const S32 repCur = (i == ZSTD_REP_MOVE_OPT) ? (rep[0] - 1) : rep[i]; rep 509 lib/zstd/zstd_opt.h opt[0].rep[i] = rep[i]; rep 540 lib/zstd/zstd_opt.h opt[cur].rep[2] = opt[cur - mlen].rep[1]; rep 541 lib/zstd/zstd_opt.h opt[cur].rep[1] = opt[cur - mlen].rep[0]; rep 542 lib/zstd/zstd_opt.h opt[cur].rep[0] = opt[cur].off - ZSTD_REP_MOVE_OPT; rep 544 lib/zstd/zstd_opt.h opt[cur].rep[2] = (opt[cur].off > 1) ? opt[cur - mlen].rep[1] : opt[cur - mlen].rep[2]; rep 545 lib/zstd/zstd_opt.h opt[cur].rep[1] = (opt[cur].off > 0) ? opt[cur - mlen].rep[0] : opt[cur - mlen].rep[1]; rep 546 lib/zstd/zstd_opt.h opt[cur].rep[0] = rep 547 lib/zstd/zstd_opt.h ((opt[cur].off == ZSTD_REP_MOVE_OPT) && (mlen != 1)) ? (opt[cur - mlen].rep[0] - 1) : (opt[cur - mlen].rep[opt[cur].off]); rep 554 lib/zstd/zstd_opt.h const S32 repCur = (i == ZSTD_REP_MOVE_OPT) ? (opt[cur].rep[0] - 1) : opt[cur].rep[i]; rep 662 lib/zstd/zstd_opt.h rep[2] = rep[1]; rep 663 lib/zstd/zstd_opt.h rep[1] = rep[0]; rep 664 lib/zstd/zstd_opt.h rep[0] = offset - ZSTD_REP_MOVE_OPT; rep 668 lib/zstd/zstd_opt.h best_off = (offset == ZSTD_REP_MOVE_OPT) ? (rep[0] - 1) : (rep[offset]); rep 670 lib/zstd/zstd_opt.h rep[2] = rep[1]; rep 671 lib/zstd/zstd_opt.h rep[1] = rep[0]; rep 672 lib/zstd/zstd_opt.h rep[0] = best_off; rep 688 lib/zstd/zstd_opt.h ctx->repToConfirm[i] = rep[i]; rep 725 lib/zstd/zstd_opt.h U32 offset, rep[ZSTD_REP_NUM]; rep 729 lib/zstd/zstd_opt.h rep[i] = ctx->rep[i]; rep 749 lib/zstd/zstd_opt.h const S32 repCur = (i == ZSTD_REP_MOVE_OPT) ? (rep[0] - 1) : rep[i]; rep 790 lib/zstd/zstd_opt.h opt[0].rep[i] = rep[i]; rep 848 lib/zstd/zstd_opt.h opt[cur].rep[2] = opt[cur - mlen].rep[1]; rep 849 lib/zstd/zstd_opt.h opt[cur].rep[1] = opt[cur - mlen].rep[0]; rep 850 lib/zstd/zstd_opt.h opt[cur].rep[0] = opt[cur].off - ZSTD_REP_MOVE_OPT; rep 852 lib/zstd/zstd_opt.h opt[cur].rep[2] = (opt[cur].off > 1) ? opt[cur - mlen].rep[1] : opt[cur - mlen].rep[2]; rep 853 lib/zstd/zstd_opt.h opt[cur].rep[1] = (opt[cur].off > 0) ? opt[cur - mlen].rep[0] : opt[cur - mlen].rep[1]; rep 854 lib/zstd/zstd_opt.h opt[cur].rep[0] = rep 855 lib/zstd/zstd_opt.h ((opt[cur].off == ZSTD_REP_MOVE_OPT) && (mlen != 1)) ? (opt[cur - mlen].rep[0] - 1) : (opt[cur - mlen].rep[opt[cur].off]); rep 862 lib/zstd/zstd_opt.h const S32 repCur = (i == ZSTD_REP_MOVE_OPT) ? (opt[cur].rep[0] - 1) : opt[cur].rep[i]; rep 976 lib/zstd/zstd_opt.h rep[2] = rep[1]; rep 977 lib/zstd/zstd_opt.h rep[1] = rep[0]; rep 978 lib/zstd/zstd_opt.h rep[0] = offset - ZSTD_REP_MOVE_OPT; rep 982 lib/zstd/zstd_opt.h best_off = (offset == ZSTD_REP_MOVE_OPT) ? (rep[0] - 1) : (rep[offset]); rep 984 lib/zstd/zstd_opt.h rep[2] = rep[1]; rep 985 lib/zstd/zstd_opt.h rep[1] = rep[0]; rep 986 lib/zstd/zstd_opt.h rep[0] = best_off; rep 1003 lib/zstd/zstd_opt.h ctx->repToConfirm[i] = rep[i]; rep 467 net/ceph/crush/mapper.c int rep; rep 484 net/ceph/crush/mapper.c for (rep = stable ? 0 : outpos; rep < numrep && count > 0 ; rep++) { rep 497 net/ceph/crush/mapper.c r = rep + parent_r; rep 658 net/ceph/crush/mapper.c int rep; rep 670 net/ceph/crush/mapper.c for (rep = outpos; rep < endpos; rep++) { rep 671 net/ceph/crush/mapper.c out[rep] = CRUSH_ITEM_UNDEF; rep 673 net/ceph/crush/mapper.c out2[rep] = CRUSH_ITEM_UNDEF; rep 680 net/ceph/crush/mapper.c for (rep = outpos; rep < endpos; rep++) { rep 681 net/ceph/crush/mapper.c dprintk(" %d", out[rep]); rep 685 net/ceph/crush/mapper.c for (rep = outpos; rep < endpos; rep++) { rep 686 net/ceph/crush/mapper.c dprintk(" %d", out2[rep]); rep 691 net/ceph/crush/mapper.c for (rep = outpos; rep < endpos; rep++) { rep 692 net/ceph/crush/mapper.c if (out[rep] != CRUSH_ITEM_UNDEF) rep 707 net/ceph/crush/mapper.c r = rep + parent_r; rep 732 net/ceph/crush/mapper.c out[rep] = CRUSH_ITEM_NONE; rep 734 net/ceph/crush/mapper.c out2[rep] = CRUSH_ITEM_NONE; rep 751 net/ceph/crush/mapper.c out[rep] = CRUSH_ITEM_NONE; rep 753 net/ceph/crush/mapper.c out2[rep] = rep 781 net/ceph/crush/mapper.c out2, rep, rep 785 net/ceph/crush/mapper.c if (out2[rep] == CRUSH_ITEM_NONE) { rep 791 net/ceph/crush/mapper.c out2[rep] = item; rep 801 net/ceph/crush/mapper.c out[rep] = item; rep 807 net/ceph/crush/mapper.c for (rep = outpos; rep < endpos; rep++) { rep 808 net/ceph/crush/mapper.c if (out[rep] == CRUSH_ITEM_UNDEF) { rep 809 net/ceph/crush/mapper.c out[rep] = CRUSH_ITEM_NONE; rep 811 net/ceph/crush/mapper.c if (out2 && out2[rep] == CRUSH_ITEM_UNDEF) { rep 812 net/ceph/crush/mapper.c out2[rep] = CRUSH_ITEM_NONE; rep 822 net/ceph/crush/mapper.c for (rep = outpos; rep < endpos; rep++) { rep 823 net/ceph/crush/mapper.c dprintk(" %d", out[rep]); rep 827 net/ceph/crush/mapper.c for (rep = outpos; rep < endpos; rep++) { rep 828 net/ceph/crush/mapper.c dprintk(" %d", out2[rep]); rep 467 net/ipv4/inet_diag.c struct sk_buff *rep; rep 475 net/ipv4/inet_diag.c rep = nlmsg_new(inet_sk_attr_size(sk, req, net_admin), GFP_KERNEL); rep 476 net/ipv4/inet_diag.c if (!rep) { rep 481 net/ipv4/inet_diag.c err = sk_diag_fill(sk, rep, req, rep 487 net/ipv4/inet_diag.c nlmsg_free(rep); rep 490 net/ipv4/inet_diag.c err = netlink_unicast(net->diag_nlsk, rep, NETLINK_CB(in_skb).portid, rep 95 net/ipv4/raw_diag.c struct sk_buff *rep; rep 103 net/ipv4/raw_diag.c rep = nlmsg_new(nla_total_size(sizeof(struct inet_diag_msg)) + rep 107 net/ipv4/raw_diag.c if (!rep) { rep 112 net/ipv4/raw_diag.c err = inet_sk_diag_fill(sk, NULL, rep, r, rep 120 net/ipv4/raw_diag.c kfree_skb(rep); rep 124 net/ipv4/raw_diag.c err = netlink_unicast(net->diag_nlsk, rep, rep 659 net/ipv4/tcp_ipv4.c } rep; rep 683 net/ipv4/tcp_ipv4.c memset(&rep, 0, sizeof(rep)); rep 684 net/ipv4/tcp_ipv4.c rep.th.dest = th->source; rep 685 net/ipv4/tcp_ipv4.c rep.th.source = th->dest; rep 686 net/ipv4/tcp_ipv4.c rep.th.doff = sizeof(struct tcphdr) / 4; rep 687 net/ipv4/tcp_ipv4.c rep.th.rst = 1; rep 690 net/ipv4/tcp_ipv4.c rep.th.seq = th->ack_seq; rep 692 net/ipv4/tcp_ipv4.c rep.th.ack = 1; rep 693 net/ipv4/tcp_ipv4.c rep.th.ack_seq = htonl(ntohl(th->seq) + th->syn + th->fin + rep 698 net/ipv4/tcp_ipv4.c arg.iov[0].iov_base = (unsigned char *)&rep; rep 699 net/ipv4/tcp_ipv4.c arg.iov[0].iov_len = sizeof(rep.th); rep 738 net/ipv4/tcp_ipv4.c rep.opt[0] = htonl((TCPOPT_NOP << 24) | rep 744 net/ipv4/tcp_ipv4.c rep.th.doff = arg.iov[0].iov_len / 4; rep 746 net/ipv4/tcp_ipv4.c tcp_v4_md5_hash_hdr((__u8 *) &rep.opt[1], rep 748 net/ipv4/tcp_ipv4.c ip_hdr(skb)->daddr, &rep.th); rep 816 net/ipv4/tcp_ipv4.c } rep; rep 822 net/ipv4/tcp_ipv4.c memset(&rep.th, 0, sizeof(struct tcphdr)); rep 825 net/ipv4/tcp_ipv4.c arg.iov[0].iov_base = (unsigned char *)&rep; rep 826 net/ipv4/tcp_ipv4.c arg.iov[0].iov_len = sizeof(rep.th); rep 828 net/ipv4/tcp_ipv4.c rep.opt[0] = htonl((TCPOPT_NOP << 24) | (TCPOPT_NOP << 16) | rep 831 net/ipv4/tcp_ipv4.c rep.opt[1] = htonl(tsval); rep 832 net/ipv4/tcp_ipv4.c rep.opt[2] = htonl(tsecr); rep 837 net/ipv4/tcp_ipv4.c rep.th.dest = th->source; rep 838 net/ipv4/tcp_ipv4.c rep.th.source = th->dest; rep 839 net/ipv4/tcp_ipv4.c rep.th.doff = arg.iov[0].iov_len / 4; rep 840 net/ipv4/tcp_ipv4.c rep.th.seq = htonl(seq); rep 841 net/ipv4/tcp_ipv4.c rep.th.ack_seq = htonl(ack); rep 842 net/ipv4/tcp_ipv4.c rep.th.ack = 1; rep 843 net/ipv4/tcp_ipv4.c rep.th.window = htons(win); rep 849 net/ipv4/tcp_ipv4.c rep.opt[offset++] = htonl((TCPOPT_NOP << 24) | rep 854 net/ipv4/tcp_ipv4.c rep.th.doff = arg.iov[0].iov_len/4; rep 856 net/ipv4/tcp_ipv4.c tcp_v4_md5_hash_hdr((__u8 *) &rep.opt[offset], rep 858 net/ipv4/tcp_ipv4.c ip_hdr(skb)->daddr, &rep.th); rep 36 net/ipv4/udp_diag.c struct sk_buff *rep; rep 67 net/ipv4/udp_diag.c rep = nlmsg_new(nla_total_size(sizeof(struct inet_diag_msg)) + rep 71 net/ipv4/udp_diag.c if (!rep) rep 74 net/ipv4/udp_diag.c err = inet_sk_diag_fill(sk, NULL, rep, req, rep 81 net/ipv4/udp_diag.c kfree_skb(rep); rep 84 net/ipv4/udp_diag.c err = netlink_unicast(net->diag_nlsk, rep, NETLINK_CB(in_skb).portid, rep 1552 net/netfilter/ipset/ip_set_core.c struct nlmsghdr *rep, *nlh = nlmsg_hdr(skb); rep 1565 net/netfilter/ipset/ip_set_core.c rep = __nlmsg_put(skb2, NETLINK_CB(skb).portid, rep 1567 net/netfilter/ipset/ip_set_core.c errmsg = nlmsg_data(rep); rep 2380 net/netlink/af_netlink.c struct nlmsghdr *rep; rep 2419 net/netlink/af_netlink.c rep = __nlmsg_put(skb, NETLINK_CB(in_skb).portid, nlh->nlmsg_seq, rep 2421 net/netlink/af_netlink.c errmsg = nlmsg_data(rep); rep 2446 net/netlink/af_netlink.c nlmsg_end(skb, rep); rep 49 net/netlink/diag.c struct netlink_diag_msg *rep; rep 52 net/netlink/diag.c nlh = nlmsg_put(skb, portid, seq, SOCK_DIAG_BY_FAMILY, sizeof(*rep), rep 57 net/netlink/diag.c rep = nlmsg_data(nlh); rep 58 net/netlink/diag.c rep->ndiag_family = AF_NETLINK; rep 59 net/netlink/diag.c rep->ndiag_type = sk->sk_type; rep 60 net/netlink/diag.c rep->ndiag_protocol = sk->sk_protocol; rep 61 net/netlink/diag.c rep->ndiag_state = sk->sk_state; rep 63 net/netlink/diag.c rep->ndiag_ino = sk_ino; rep 64 net/netlink/diag.c rep->ndiag_portid = nlk->portid; rep 65 net/netlink/diag.c rep->ndiag_dst_portid = nlk->dst_portid; rep 66 net/netlink/diag.c rep->ndiag_dst_group = nlk->dst_group; rep 67 net/netlink/diag.c sock_diag_save_cookie(sk, rep->ndiag_cookie); rep 257 net/sctp/diag.c struct sk_buff *rep; rep 265 net/sctp/diag.c rep = nlmsg_new(inet_assoc_attr_size(assoc), GFP_KERNEL); rep 266 net/sctp/diag.c if (!rep) rep 275 net/sctp/diag.c err = inet_sctp_diag_fill(sk, assoc, rep, req, rep 283 net/sctp/diag.c kfree_skb(rep); rep 287 net/sctp/diag.c err = netlink_unicast(net->diag_nlsk, rep, NETLINK_CB(in_skb).portid, rep 218 net/sunrpc/xprtrdma/backchannel.c struct rpcrdma_rep *rep) rep 228 net/sunrpc/xprtrdma/backchannel.c p = xdr_inline_decode(&rep->rr_stream, 0); rep 229 net/sunrpc/xprtrdma/backchannel.c size = xdr_stream_remaining(&rep->rr_stream); rep 258 net/sunrpc/xprtrdma/backchannel.c req->rl_reply = rep; rep 456 net/sunrpc/xprtrdma/frwr_ops.c void frwr_reminv(struct rpcrdma_rep *rep, struct list_head *mrs) rep 461 net/sunrpc/xprtrdma/frwr_ops.c if (mr->mr_handle == rep->rr_inv_rkey) { rep 610 net/sunrpc/xprtrdma/frwr_ops.c struct rpcrdma_rep *rep = mr->mr_req->rl_reply; rep 618 net/sunrpc/xprtrdma/frwr_ops.c rpcrdma_complete_rqst(rep); rep 555 net/sunrpc/xprtrdma/rpc_rdma.c struct rpcrdma_rep *rep = req->rl_reply; rep 557 net/sunrpc/xprtrdma/rpc_rdma.c rpcrdma_complete_rqst(rep); rep 558 net/sunrpc/xprtrdma/rpc_rdma.c rep->rr_rxprt->rx_stats.reply_waits_for_send++; rep 1020 net/sunrpc/xprtrdma/rpc_rdma.c rpcrdma_is_bcall(struct rpcrdma_xprt *r_xprt, struct rpcrdma_rep *rep) rep 1023 net/sunrpc/xprtrdma/rpc_rdma.c struct xdr_stream *xdr = &rep->rr_stream; rep 1026 net/sunrpc/xprtrdma/rpc_rdma.c if (rep->rr_proc != rdma_msg) rep 1041 net/sunrpc/xprtrdma/rpc_rdma.c if (*p++ != rep->rr_xid) rep 1053 net/sunrpc/xprtrdma/rpc_rdma.c rpcrdma_bc_receive_call(r_xprt, rep); rep 1163 net/sunrpc/xprtrdma/rpc_rdma.c rpcrdma_decode_msg(struct rpcrdma_xprt *r_xprt, struct rpcrdma_rep *rep, rep 1166 net/sunrpc/xprtrdma/rpc_rdma.c struct xdr_stream *xdr = &rep->rr_stream; rep 1193 net/sunrpc/xprtrdma/rpc_rdma.c rpcrdma_decode_nomsg(struct rpcrdma_xprt *r_xprt, struct rpcrdma_rep *rep) rep 1195 net/sunrpc/xprtrdma/rpc_rdma.c struct xdr_stream *xdr = &rep->rr_stream; rep 1218 net/sunrpc/xprtrdma/rpc_rdma.c rpcrdma_decode_error(struct rpcrdma_xprt *r_xprt, struct rpcrdma_rep *rep, rep 1221 net/sunrpc/xprtrdma/rpc_rdma.c struct xdr_stream *xdr = &rep->rr_stream; rep 1236 net/sunrpc/xprtrdma/rpc_rdma.c be32_to_cpu(rep->rr_xid)); rep 1241 net/sunrpc/xprtrdma/rpc_rdma.c be32_to_cpu(rep->rr_xid)); rep 1246 net/sunrpc/xprtrdma/rpc_rdma.c be32_to_cpup(p), be32_to_cpu(rep->rr_xid)); rep 1257 net/sunrpc/xprtrdma/rpc_rdma.c void rpcrdma_complete_rqst(struct rpcrdma_rep *rep) rep 1259 net/sunrpc/xprtrdma/rpc_rdma.c struct rpcrdma_xprt *r_xprt = rep->rr_rxprt; rep 1261 net/sunrpc/xprtrdma/rpc_rdma.c struct rpc_rqst *rqst = rep->rr_rqst; rep 1264 net/sunrpc/xprtrdma/rpc_rdma.c switch (rep->rr_proc) { rep 1266 net/sunrpc/xprtrdma/rpc_rdma.c status = rpcrdma_decode_msg(r_xprt, rep, rqst); rep 1269 net/sunrpc/xprtrdma/rpc_rdma.c status = rpcrdma_decode_nomsg(r_xprt, rep); rep 1272 net/sunrpc/xprtrdma/rpc_rdma.c status = rpcrdma_decode_error(r_xprt, rep, rqst); rep 1292 net/sunrpc/xprtrdma/rpc_rdma.c trace_xprtrdma_reply_hdr(rep); rep 1312 net/sunrpc/xprtrdma/rpc_rdma.c void rpcrdma_reply_handler(struct rpcrdma_rep *rep) rep 1314 net/sunrpc/xprtrdma/rpc_rdma.c struct rpcrdma_xprt *r_xprt = rep->rr_rxprt; rep 1329 net/sunrpc/xprtrdma/rpc_rdma.c xdr_init_decode(&rep->rr_stream, &rep->rr_hdrbuf, rep 1330 net/sunrpc/xprtrdma/rpc_rdma.c rep->rr_hdrbuf.head[0].iov_base, NULL); rep 1331 net/sunrpc/xprtrdma/rpc_rdma.c p = xdr_inline_decode(&rep->rr_stream, 4 * sizeof(*p)); rep 1334 net/sunrpc/xprtrdma/rpc_rdma.c rep->rr_xid = *p++; rep 1335 net/sunrpc/xprtrdma/rpc_rdma.c rep->rr_vers = *p++; rep 1337 net/sunrpc/xprtrdma/rpc_rdma.c rep->rr_proc = *p++; rep 1339 net/sunrpc/xprtrdma/rpc_rdma.c if (rep->rr_vers != rpcrdma_version) rep 1342 net/sunrpc/xprtrdma/rpc_rdma.c if (rpcrdma_is_bcall(r_xprt, rep)) rep 1349 net/sunrpc/xprtrdma/rpc_rdma.c rqst = xprt_lookup_rqst(xprt, rep->rr_xid); rep 1372 net/sunrpc/xprtrdma/rpc_rdma.c req->rl_reply = rep; rep 1373 net/sunrpc/xprtrdma/rpc_rdma.c rep->rr_rqst = rqst; rep 1375 net/sunrpc/xprtrdma/rpc_rdma.c trace_xprtrdma_reply(rqst->rq_task, rep, req, credits); rep 1377 net/sunrpc/xprtrdma/rpc_rdma.c if (rep->rr_wc_flags & IB_WC_WITH_INVALIDATE) rep 1378 net/sunrpc/xprtrdma/rpc_rdma.c frwr_reminv(rep, &req->rl_registered); rep 1387 net/sunrpc/xprtrdma/rpc_rdma.c trace_xprtrdma_reply_vers(rep); rep 1392 net/sunrpc/xprtrdma/rpc_rdma.c trace_xprtrdma_reply_rqst(rep); rep 1396 net/sunrpc/xprtrdma/rpc_rdma.c trace_xprtrdma_reply_short(rep); rep 1399 net/sunrpc/xprtrdma/rpc_rdma.c rpcrdma_recv_buffer_put(rep); rep 153 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_rep *rep = container_of(cqe, struct rpcrdma_rep, rep 155 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_xprt *r_xprt = rep->rr_rxprt; rep 164 net/sunrpc/xprtrdma/verbs.c rpcrdma_set_xdrlen(&rep->rr_hdrbuf, wc->byte_len); rep 165 net/sunrpc/xprtrdma/verbs.c rep->rr_wc_flags = wc->wc_flags; rep 166 net/sunrpc/xprtrdma/verbs.c rep->rr_inv_rkey = wc->ex.invalidate_rkey; rep 168 net/sunrpc/xprtrdma/verbs.c ib_dma_sync_single_for_cpu(rdmab_device(rep->rr_rdmabuf), rep 169 net/sunrpc/xprtrdma/verbs.c rdmab_addr(rep->rr_rdmabuf), rep 172 net/sunrpc/xprtrdma/verbs.c rpcrdma_reply_handler(rep); rep 176 net/sunrpc/xprtrdma/verbs.c rpcrdma_recv_buffer_put(rep); rep 1069 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_rep *rep; rep 1071 net/sunrpc/xprtrdma/verbs.c rep = kzalloc(sizeof(*rep), GFP_KERNEL); rep 1072 net/sunrpc/xprtrdma/verbs.c if (rep == NULL) rep 1075 net/sunrpc/xprtrdma/verbs.c rep->rr_rdmabuf = rpcrdma_regbuf_alloc(r_xprt->rx_ep.rep_inline_recv, rep 1077 net/sunrpc/xprtrdma/verbs.c if (!rep->rr_rdmabuf) rep 1080 net/sunrpc/xprtrdma/verbs.c xdr_buf_init(&rep->rr_hdrbuf, rdmab_data(rep->rr_rdmabuf), rep 1081 net/sunrpc/xprtrdma/verbs.c rdmab_length(rep->rr_rdmabuf)); rep 1082 net/sunrpc/xprtrdma/verbs.c rep->rr_cqe.done = rpcrdma_wc_receive; rep 1083 net/sunrpc/xprtrdma/verbs.c rep->rr_rxprt = r_xprt; rep 1084 net/sunrpc/xprtrdma/verbs.c rep->rr_recv_wr.next = NULL; rep 1085 net/sunrpc/xprtrdma/verbs.c rep->rr_recv_wr.wr_cqe = &rep->rr_cqe; rep 1086 net/sunrpc/xprtrdma/verbs.c rep->rr_recv_wr.sg_list = &rep->rr_rdmabuf->rg_iov; rep 1087 net/sunrpc/xprtrdma/verbs.c rep->rr_recv_wr.num_sge = 1; rep 1088 net/sunrpc/xprtrdma/verbs.c rep->rr_temp = temp; rep 1089 net/sunrpc/xprtrdma/verbs.c list_add(&rep->rr_all, &r_xprt->rx_buf.rb_all_reps); rep 1090 net/sunrpc/xprtrdma/verbs.c return rep; rep 1093 net/sunrpc/xprtrdma/verbs.c kfree(rep); rep 1098 net/sunrpc/xprtrdma/verbs.c static void rpcrdma_rep_destroy(struct rpcrdma_rep *rep) rep 1100 net/sunrpc/xprtrdma/verbs.c list_del(&rep->rr_all); rep 1101 net/sunrpc/xprtrdma/verbs.c rpcrdma_regbuf_free(rep->rr_rdmabuf); rep 1102 net/sunrpc/xprtrdma/verbs.c kfree(rep); rep 1117 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_rep *rep) rep 1119 net/sunrpc/xprtrdma/verbs.c llist_add(&rep->rr_node, &buf->rb_free_reps); rep 1125 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_rep *rep; rep 1127 net/sunrpc/xprtrdma/verbs.c list_for_each_entry(rep, &buf->rb_all_reps, rr_all) rep 1128 net/sunrpc/xprtrdma/verbs.c rpcrdma_regbuf_dma_unmap(rep->rr_rdmabuf); rep 1133 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_rep *rep; rep 1135 net/sunrpc/xprtrdma/verbs.c while ((rep = rpcrdma_rep_get_locked(buf)) != NULL) rep 1136 net/sunrpc/xprtrdma/verbs.c rpcrdma_rep_destroy(rep); rep 1353 net/sunrpc/xprtrdma/verbs.c void rpcrdma_recv_buffer_put(struct rpcrdma_rep *rep) rep 1355 net/sunrpc/xprtrdma/verbs.c rpcrdma_rep_put(&rep->rr_rxprt->rx_buf, rep); rep 1501 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_rep *rep; rep 1517 net/sunrpc/xprtrdma/verbs.c rep = rpcrdma_rep_get_locked(buf); rep 1518 net/sunrpc/xprtrdma/verbs.c if (rep && rep->rr_temp) { rep 1519 net/sunrpc/xprtrdma/verbs.c rpcrdma_rep_destroy(rep); rep 1522 net/sunrpc/xprtrdma/verbs.c if (!rep) rep 1523 net/sunrpc/xprtrdma/verbs.c rep = rpcrdma_rep_create(r_xprt, temp); rep 1524 net/sunrpc/xprtrdma/verbs.c if (!rep) rep 1527 net/sunrpc/xprtrdma/verbs.c rep->rr_recv_wr.next = wr; rep 1528 net/sunrpc/xprtrdma/verbs.c wr = &rep->rr_recv_wr; rep 1535 net/sunrpc/xprtrdma/verbs.c rep = container_of(i, struct rpcrdma_rep, rr_recv_wr); rep 1537 net/sunrpc/xprtrdma/verbs.c if (!rpcrdma_regbuf_dma_map(r_xprt, rep->rr_rdmabuf)) rep 1540 net/sunrpc/xprtrdma/verbs.c trace_xprtrdma_post_recv(rep); rep 1550 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_rep *rep; rep 1552 net/sunrpc/xprtrdma/verbs.c rep = container_of(wr, struct rpcrdma_rep, rr_recv_wr); rep 1554 net/sunrpc/xprtrdma/verbs.c rpcrdma_recv_buffer_put(rep); rep 1563 net/sunrpc/xprtrdma/verbs.c rep = container_of(i, struct rpcrdma_rep, rr_recv_wr); rep 1565 net/sunrpc/xprtrdma/verbs.c rpcrdma_recv_buffer_put(rep); rep 559 net/sunrpc/xprtrdma/xprt_rdma.h void frwr_reminv(struct rpcrdma_rep *rep, struct list_head *mrs); rep 582 net/sunrpc/xprtrdma/xprt_rdma.h void rpcrdma_complete_rqst(struct rpcrdma_rep *rep); rep 583 net/sunrpc/xprtrdma/xprt_rdma.h void rpcrdma_reply_handler(struct rpcrdma_rep *rep); rep 834 net/tipc/bearer.c struct sk_buff *rep; rep 853 net/tipc/bearer.c rep = nlmsg_new(NLMSG_GOODSIZE, GFP_KERNEL); rep 854 net/tipc/bearer.c if (!rep) rep 857 net/tipc/bearer.c msg.skb = rep; rep 873 net/tipc/bearer.c return genlmsg_reply(rep, info); rep 876 net/tipc/bearer.c nlmsg_free(rep); rep 1168 net/tipc/bearer.c struct sk_buff *rep; rep 1184 net/tipc/bearer.c rep = nlmsg_new(NLMSG_GOODSIZE, GFP_KERNEL); rep 1185 net/tipc/bearer.c if (!rep) rep 1188 net/tipc/bearer.c msg.skb = rep; rep 1204 net/tipc/bearer.c return genlmsg_reply(rep, info); rep 1207 net/tipc/bearer.c nlmsg_free(rep); rep 60 net/tipc/netlink_compat.c struct sk_buff *rep; rep 220 net/tipc/netlink_compat.c if (tipc_skb_tailroom(msg->rep) <= 1) { rep 241 net/tipc/netlink_compat.c if ((TIPC_SKB_MAX - msg->rep->len) <= 1) { rep 242 net/tipc/netlink_compat.c char *tail = skb_tail_pointer(msg->rep); rep 265 net/tipc/netlink_compat.c msg->rep = tipc_tlv_alloc(msg->rep_size); rep 266 net/tipc/netlink_compat.c if (!msg->rep) rep 270 net/tipc/netlink_compat.c tipc_tlv_init(msg->rep, msg->rep_type); rep 275 net/tipc/netlink_compat.c kfree_skb(msg->rep); rep 276 net/tipc/netlink_compat.c msg->rep = NULL; rep 283 net/tipc/netlink_compat.c kfree_skb(msg->rep); rep 284 net/tipc/netlink_compat.c msg->rep = NULL; rep 290 net/tipc/netlink_compat.c kfree_skb(msg->rep); rep 291 net/tipc/netlink_compat.c msg->rep = NULL; rep 368 net/tipc/netlink_compat.c msg->rep = tipc_tlv_alloc(0); rep 369 net/tipc/netlink_compat.c if (!msg->rep) rep 389 net/tipc/netlink_compat.c return tipc_add_tlv(msg->rep, TIPC_TLV_BEARER_NAME, rep 475 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " Window:%u packets\n", rep 478 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 486 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 494 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " RX naks:%u defs:%u dups:%u\n", rep 499 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " TX naks:%u acks:%u dups:%u\n", rep 504 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 560 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "\nLink <%s>\n", rep 569 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " ACTIVE"); rep 571 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " STANDBY"); rep 573 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " DEFUNCT"); rep 575 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " MTU:%u Priority:%u", rep 579 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " Tolerance:%u ms Window:%u packets\n", rep 583 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 592 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 601 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 607 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 618 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "-16384:%u%% -32768:%u%% -66000:%u%%\n", rep 626 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 634 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 642 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, rep 671 net/tipc/netlink_compat.c return tipc_add_tlv(msg->rep, TIPC_TLV_LINK_INFO, rep 858 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, header[i]); rep 859 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "\n"); rep 909 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "%-10u ", rep 915 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "%-10u %-10u ", rep 925 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "%-26s ", port_str); rep 930 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "%-10u %s", rep 934 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "\n"); rep 959 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " {%u,%u}", type, lower); rep 961 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " {%u,%u,%u}", type, lower, upper); rep 1025 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "%u:", sock_ref); rep 1039 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " connected to <%u.%u.%u:%u>", rep 1046 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " via {%u,%u}\n", rep 1050 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "\n"); rep 1052 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, " bound to"); rep 1058 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "\n"); rep 1077 net/tipc/netlink_compat.c return tipc_add_tlv(msg->rep, TIPC_TLV_MEDIA_NAME, rep 1100 net/tipc/netlink_compat.c return tipc_add_tlv(msg->rep, TIPC_TLV_NODE_INFO, &node_info, rep 1146 net/tipc/netlink_compat.c return tipc_add_tlv(msg->rep, TIPC_TLV_UNSIGNED, &id, sizeof(id)); rep 1151 net/tipc/netlink_compat.c msg->rep = tipc_tlv_alloc(ULTRA_STRING_MAX_LEN); rep 1152 net/tipc/netlink_compat.c if (!msg->rep) rep 1155 net/tipc/netlink_compat.c tipc_tlv_init(msg->rep, TIPC_TLV_ULTRA_STRING); rep 1156 net/tipc/netlink_compat.c tipc_tlv_sprintf(msg->rep, "TIPC version " TIPC_MOD_VER "\n"); rep 1171 net/tipc/netlink_compat.c msg->rep = tipc_tlv_alloc(0); rep 1172 net/tipc/netlink_compat.c if (!msg->rep) rep 1279 net/tipc/netlink_compat.c msg.rep = tipc_get_err_tlv(TIPC_CFG_NOT_NET_ADMIN); rep 1286 net/tipc/netlink_compat.c msg.rep = tipc_get_err_tlv(TIPC_CFG_NOT_SUPPORTED); rep 1293 net/tipc/netlink_compat.c msg.rep = tipc_get_err_tlv(TIPC_CFG_NOT_SUPPORTED); rep 1295 net/tipc/netlink_compat.c msg.rep = tipc_get_err_tlv(TIPC_CFG_TLV_ERROR); rep 1297 net/tipc/netlink_compat.c if (!msg.rep) rep 1301 net/tipc/netlink_compat.c skb_push(msg.rep, len); rep 1302 net/tipc/netlink_compat.c rep_nlh = nlmsg_hdr(msg.rep); rep 1304 net/tipc/netlink_compat.c rep_nlh->nlmsg_len = msg.rep->len; rep 1305 net/tipc/netlink_compat.c genlmsg_unicast(msg.net, msg.rep, NETLINK_CB(skb).portid); rep 126 net/unix/diag.c struct unix_diag_msg *rep; rep 128 net/unix/diag.c nlh = nlmsg_put(skb, portid, seq, SOCK_DIAG_BY_FAMILY, sizeof(*rep), rep 133 net/unix/diag.c rep = nlmsg_data(nlh); rep 134 net/unix/diag.c rep->udiag_family = AF_UNIX; rep 135 net/unix/diag.c rep->udiag_type = sk->sk_type; rep 136 net/unix/diag.c rep->udiag_state = sk->sk_state; rep 137 net/unix/diag.c rep->pad = 0; rep 138 net/unix/diag.c rep->udiag_ino = sk_ino; rep 139 net/unix/diag.c sock_diag_save_cookie(sk, rep->udiag_cookie); rep 263 net/unix/diag.c struct sk_buff *rep; rep 284 net/unix/diag.c rep = nlmsg_new(sizeof(struct unix_diag_msg) + extra_len, GFP_KERNEL); rep 285 net/unix/diag.c if (!rep) rep 288 net/unix/diag.c err = sk_diag_fill(sk, rep, req, NETLINK_CB(in_skb).portid, rep 291 net/unix/diag.c nlmsg_free(rep); rep 298 net/unix/diag.c err = netlink_unicast(net->diag_nlsk, rep, NETLINK_CB(in_skb).portid, rep 18 net/vmw_vsock/diag.c struct vsock_diag_msg *rep; rep 21 net/vmw_vsock/diag.c nlh = nlmsg_put(skb, portid, seq, SOCK_DIAG_BY_FAMILY, sizeof(*rep), rep 26 net/vmw_vsock/diag.c rep = nlmsg_data(nlh); rep 27 net/vmw_vsock/diag.c rep->vdiag_family = AF_VSOCK; rep 34 net/vmw_vsock/diag.c rep->vdiag_type = sk->sk_type; rep 35 net/vmw_vsock/diag.c rep->vdiag_state = sk->sk_state; rep 36 net/vmw_vsock/diag.c rep->vdiag_shutdown = sk->sk_shutdown; rep 37 net/vmw_vsock/diag.c rep->vdiag_src_cid = vsk->local_addr.svm_cid; rep 38 net/vmw_vsock/diag.c rep->vdiag_src_port = vsk->local_addr.svm_port; rep 39 net/vmw_vsock/diag.c rep->vdiag_dst_cid = vsk->remote_addr.svm_cid; rep 40 net/vmw_vsock/diag.c rep->vdiag_dst_port = vsk->remote_addr.svm_port; rep 41 net/vmw_vsock/diag.c rep->vdiag_ino = sock_i_ino(sk); rep 43 net/vmw_vsock/diag.c sock_diag_save_cookie(sk, rep->vdiag_cookie); rep 1278 sound/usb/midi.c struct snd_usb_midi_endpoint *rep) rep 1287 sound/usb/midi.c rep->in = NULL; rep 1330 sound/usb/midi.c rep->in = ep; rep 1365 sound/usb/midi.c struct snd_usb_midi_endpoint *rep) rep 1373 sound/usb/midi.c rep->out = NULL; rep 1456 sound/usb/midi.c rep->out = ep; rep 107 tools/perf/builtin-report.c struct report *rep = cb; rep 116 tools/perf/builtin-report.c rep->min_percent = pcnt; rep 125 tools/perf/builtin-report.c return perf_config_u64(&rep->queue_size, var, value); rep 140 tools/perf/builtin-report.c struct report *rep = arg; rep 147 tools/perf/builtin-report.c if (!ui__has_annotation() && !rep->symbol_ipc) rep 158 tools/perf/builtin-report.c } else if (rep->mem_mode) { rep 183 tools/perf/builtin-report.c struct report *rep = arg; rep 189 tools/perf/builtin-report.c branch_type_count(&rep->brtype_stat, &bi->flags, rep 192 tools/perf/builtin-report.c if (!ui__has_annotation() && !rep->symbol_ipc) rep 215 tools/perf/builtin-report.c struct report *rep = container_of(session->tool, struct report, tool); rep 231 tools/perf/builtin-report.c setup_forced_leader(rep, session->evlist); rep 241 tools/perf/builtin-report.c struct report *rep = container_of(tool, struct report, tool); rep 251 tools/perf/builtin-report.c if (perf_time__ranges_skip_sample(rep->ptime_range, rep->range_num, rep 256 tools/perf/builtin-report.c if (evswitch__discard(&rep->evswitch, evsel)) rep 268 tools/perf/builtin-report.c if (rep->cpu_list && !test_bit(sample->cpu, rep->cpu_bitmap)) rep 281 tools/perf/builtin-report.c } else if (rep->mem_mode) { rep 292 tools/perf/builtin-report.c if (ui__has_annotation() || rep->symbol_ipc) { rep 294 tools/perf/builtin-report.c rep->nonany_branch_mode); rep 297 tools/perf/builtin-report.c ret = hist_entry_iter__add(&iter, &al, rep->max_stack, rep); rep 311 tools/perf/builtin-report.c struct report *rep = container_of(tool, struct report, tool); rep 313 tools/perf/builtin-report.c if (rep->show_threads) { rep 315 tools/perf/builtin-report.c int err = perf_read_values_add_value(&rep->show_threads_values, rep 329 tools/perf/builtin-report.c static int report__setup_sample_type(struct report *rep) rep 331 tools/perf/builtin-report.c struct perf_session *session = rep->session; rep 399 tools/perf/builtin-report.c rep->nonany_branch_mode = true; rep 416 tools/perf/builtin-report.c static size_t hists__fprintf_nr_sample_events(struct hists *hists, struct report *rep, rep 462 tools/perf/builtin-report.c if (rep->time_str) rep 463 tools/perf/builtin-report.c ret += fprintf(fp, " (time slices: %s)", rep->time_str); rep 470 tools/perf/builtin-report.c if (rep->mem_mode) { rep 483 tools/perf/builtin-report.c struct report *rep, rep 501 tools/perf/builtin-report.c hists__fprintf_nr_sample_events(hists, rep, evname, stdout); rep 502 tools/perf/builtin-report.c hists__fprintf(hists, !quiet, 0, 0, rep->min_percent, stdout, rep 511 tools/perf/builtin-report.c if (rep->show_threads) { rep 512 tools/perf/builtin-report.c bool style = !strcmp(rep->pretty_printing_style, "raw"); rep 513 tools/perf/builtin-report.c perf_read_values_display(stdout, &rep->show_threads_values, rep 515 tools/perf/builtin-report.c perf_read_values_destroy(&rep->show_threads_values); rep 519 tools/perf/builtin-report.c branch_type_stat_display(stdout, &rep->brtype_stat); rep 524 tools/perf/builtin-report.c static void report__warn_kptr_restrict(const struct report *rep) rep 526 tools/perf/builtin-report.c struct map *kernel_map = machine__kernel_map(&rep->session->machines.host); rep 529 tools/perf/builtin-report.c if (perf_evlist__exclude_kernel(rep->session->evlist)) rep 553 tools/perf/builtin-report.c static int report__gtk_browse_hists(struct report *rep, const char *help) rep 565 tools/perf/builtin-report.c return hist_browser(rep->session->evlist, help, NULL, rep->min_percent); rep 568 tools/perf/builtin-report.c static int report__browse_hists(struct report *rep) rep 571 tools/perf/builtin-report.c struct perf_session *session = rep->session; rep 585 tools/perf/builtin-report.c rep->min_percent, rep 587 tools/perf/builtin-report.c true, &rep->annotation_opts); rep 596 tools/perf/builtin-report.c ret = report__gtk_browse_hists(rep, help); rep 599 tools/perf/builtin-report.c ret = perf_evlist__tty_browse_hists(evlist, rep, help); rep 606 tools/perf/builtin-report.c static int report__collapse_hists(struct report *rep) rep 612 tools/perf/builtin-report.c ui_progress__init(&prog, rep->nr_entries, "Merging related events..."); rep 614 tools/perf/builtin-report.c evlist__for_each_entry(rep->session->evlist, pos) { rep 618 tools/perf/builtin-report.c hists->symbol_filter_str = rep->symbol_filter_str; rep 620 tools/perf/builtin-report.c hists->socket_filter = rep->socket_filter; rep 642 tools/perf/builtin-report.c struct report *rep = arg; rep 645 tools/perf/builtin-report.c if (rep->symbol_ipc && sym && !sym->annotate2) { rep 655 tools/perf/builtin-report.c static void report__output_resort(struct report *rep) rep 660 tools/perf/builtin-report.c ui_progress__init(&prog, rep->nr_entries, "Sorting events for output..."); rep 662 tools/perf/builtin-report.c evlist__for_each_entry(rep->session->evlist, pos) { rep 664 tools/perf/builtin-report.c hists__resort_cb, rep); rep 670 tools/perf/builtin-report.c static void stats_setup(struct report *rep) rep 672 tools/perf/builtin-report.c memset(&rep->tool, 0, sizeof(rep->tool)); rep 673 tools/perf/builtin-report.c rep->tool.no_warn = true; rep 676 tools/perf/builtin-report.c static int stats_print(struct report *rep) rep 678 tools/perf/builtin-report.c struct perf_session *session = rep->session; rep 684 tools/perf/builtin-report.c static void tasks_setup(struct report *rep) rep 686 tools/perf/builtin-report.c memset(&rep->tool, 0, sizeof(rep->tool)); rep 687 tools/perf/builtin-report.c rep->tool.ordered_events = true; rep 688 tools/perf/builtin-report.c if (rep->mmaps_mode) { rep 689 tools/perf/builtin-report.c rep->tool.mmap = perf_event__process_mmap; rep 690 tools/perf/builtin-report.c rep->tool.mmap2 = perf_event__process_mmap2; rep 692 tools/perf/builtin-report.c rep->tool.comm = perf_event__process_comm; rep 693 tools/perf/builtin-report.c rep->tool.exit = perf_event__process_exit; rep 694 tools/perf/builtin-report.c rep->tool.fork = perf_event__process_fork; rep 695 tools/perf/builtin-report.c rep->tool.no_warn = true; rep 770 tools/perf/builtin-report.c static int tasks_print(struct report *rep, FILE *fp) rep 772 tools/perf/builtin-report.c struct perf_session *session = rep->session; rep 837 tools/perf/builtin-report.c static int __cmd_report(struct report *rep) rep 840 tools/perf/builtin-report.c struct perf_session *session = rep->session; rep 846 tools/perf/builtin-report.c if (rep->cpu_list) { rep 847 tools/perf/builtin-report.c ret = perf_session__cpu_bitmap(session, rep->cpu_list, rep 848 tools/perf/builtin-report.c rep->cpu_bitmap); rep 853 tools/perf/builtin-report.c session->itrace_synth_opts->cpu_bitmap = rep->cpu_bitmap; rep 856 tools/perf/builtin-report.c if (rep->show_threads) { rep 857 tools/perf/builtin-report.c ret = perf_read_values_init(&rep->show_threads_values); rep 862 tools/perf/builtin-report.c ret = report__setup_sample_type(rep); rep 868 tools/perf/builtin-report.c if (rep->stats_mode) rep 869 tools/perf/builtin-report.c stats_setup(rep); rep 871 tools/perf/builtin-report.c if (rep->tasks_mode) rep 872 tools/perf/builtin-report.c tasks_setup(rep); rep 880 tools/perf/builtin-report.c if (rep->stats_mode) rep 881 tools/perf/builtin-report.c return stats_print(rep); rep 883 tools/perf/builtin-report.c if (rep->tasks_mode) rep 884 tools/perf/builtin-report.c return tasks_print(rep, stdout); rep 886 tools/perf/builtin-report.c report__warn_kptr_restrict(rep); rep 889 tools/perf/builtin-report.c rep->nr_entries += evsel__hists(pos)->nr_entries; rep 905 tools/perf/builtin-report.c ret = report__collapse_hists(rep); rep 918 tools/perf/builtin-report.c rep->nr_entries = 0; rep 920 tools/perf/builtin-report.c rep->nr_entries += evsel__hists(pos)->nr_entries; rep 922 tools/perf/builtin-report.c if (rep->nr_entries == 0) { rep 927 tools/perf/builtin-report.c report__output_resort(rep); rep 929 tools/perf/builtin-report.c return report__browse_hists(rep); rep 1018 tools/perf/builtin-report.c struct report *rep = opt->value; rep 1021 tools/perf/builtin-report.c rep->min_percent = pcnt;