cbs 160 drivers/dma/ppc4xx/adma.c cb->cbc, cb->cbbc, cb->cbs, cbs 713 drivers/dma/ppc4xx/adma.c xor_hw_desc->cbs = 0; cbs 73 drivers/dma/ppc4xx/xor.h u32 cbs; /* status */ cbs 421 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_mst_types.c aconnector->mst_mgr.cbs = &dm_mst_cbs; cbs 1698 drivers/gpu/drm/drm_dp_mst_topology.c port->connector = (*mstb->mgr->cbs->add_connector)(mstb->mgr, cbs 1717 drivers/gpu/drm/drm_dp_mst_topology.c (*mstb->mgr->cbs->register_connector)(port->connector); cbs 3761 drivers/gpu/drm/drm_dp_mst_topology.c mgr->cbs->destroy_connector(mgr, port->connector); cbs 670 drivers/gpu/drm/i915/display/intel_dp_mst.c intel_dp->mst_mgr.cbs = &mst_cbs; cbs 58 drivers/gpu/drm/nouveau/dispnv04/arb.c int found, mclk_extra, mclk_loop, cbs, m1, p1; cbs 70 drivers/gpu/drm/nouveau/dispnv04/arb.c cbs = 128; cbs 94 drivers/gpu/drm/nouveau/dispnv04/arb.c m1 = clwm + cbs - 512; cbs 105 drivers/gpu/drm/nouveau/dispnv04/arb.c fifo->burst = cbs; cbs 1386 drivers/gpu/drm/nouveau/dispnv50/disp.c mstm->mgr.cbs = &nv50_mstm; cbs 657 drivers/gpu/drm/radeon/radeon_dp_mst.c radeon_connector->mst_mgr.cbs = &mst_cbs; cbs 119 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_binding_emit_dirty(struct vmw_ctx_binding_state *cbs); cbs 244 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_cbs_context(const struct vmw_ctx_binding_state *cbs) cbs 246 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (list_empty(&cbs->list)) cbs 249 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c return list_first_entry(&cbs->list, struct vmw_ctx_bindinfo, cbs 262 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_loc(struct vmw_ctx_binding_state *cbs, cbs 268 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c return (struct vmw_ctx_bindinfo *)((u8 *) cbs + offset); cbs 297 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_add(struct vmw_ctx_binding_state *cbs, cbs 302 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_loc(cbs, bi->bt, shader_slot, slot); cbs 310 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_add(&loc->ctx_list, &cbs->list); cbs 321 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_binding_transfer(struct vmw_ctx_binding_state *cbs, cbs 327 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ((unsigned long) cbs + offset); cbs 337 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_add_tail(&loc->ctx_list, &cbs->list); cbs 351 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_state_kill(struct vmw_ctx_binding_state *cbs) cbs 355 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_state_scrub(cbs); cbs 356 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, &cbs->list, ctx_list) cbs 369 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_state_scrub(struct vmw_ctx_binding_state *cbs) cbs 373 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry(entry, &cbs->list, ctx_list) { cbs 381 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c (void) vmw_binding_emit_dirty(cbs); cbs 424 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_binding_state *cbs = cbs 427 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c (void) vmw_binding_emit_dirty(cbs); cbs 463 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c int vmw_binding_rebind_all(struct vmw_ctx_binding_state *cbs) cbs 468 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry(entry, &cbs->list, ctx_list) { cbs 483 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c return vmw_binding_emit_dirty(cbs); cbs 669 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_collect_view_ids(struct vmw_ctx_binding_state *cbs, cbs 677 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_cmd_count = 0; cbs 678 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_first_slot = 0; cbs 684 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_cmd_buffer[cbs->bind_cmd_count++] = cbs 704 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_collect_dirty_view_ids(struct vmw_ctx_binding_state *cbs, cbs 713 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_cmd_count = 0; cbs 716 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_first_slot = i; cbs 720 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_cmd_buffer[cbs->bind_cmd_count++] = cbs 737 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_emit_set_sr(struct vmw_ctx_binding_state *cbs, cbs 741 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c &cbs->per_shader[shader_slot].shader_res[0].bi; cbs 747 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c const struct vmw_resource *ctx = vmw_cbs_context(cbs); cbs 749 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_collect_dirty_view_ids(cbs, loc, cbs 750 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->per_shader[shader_slot].dirty_sr, cbs 752 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (cbs->bind_cmd_count == 0) cbs 755 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c view_id_size = cbs->bind_cmd_count*sizeof(uint32); cbs 764 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cmd->body.startView = cbs->bind_first_slot; cbs 766 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c memcpy(&cmd[1], cbs->bind_cmd_buffer, view_id_size); cbs 769 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c bitmap_clear(cbs->per_shader[shader_slot].dirty_sr, cbs 770 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_first_slot, cbs->bind_cmd_count); cbs 780 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_emit_set_rt(struct vmw_ctx_binding_state *cbs) cbs 782 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c const struct vmw_ctx_bindinfo *loc = &cbs->render_targets[0].bi; cbs 788 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c const struct vmw_resource *ctx = vmw_cbs_context(cbs); cbs 790 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_collect_view_ids(cbs, loc, SVGA3D_MAX_SIMULTANEOUS_RENDER_TARGETS); cbs 791 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c view_id_size = cbs->bind_cmd_count*sizeof(uint32); cbs 800 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (cbs->ds_view.bi.ctx && !cbs->ds_view.bi.scrubbed) cbs 801 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cmd->body.depthStencilViewId = cbs->ds_view.bi.res->id; cbs 805 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c memcpy(&cmd[1], cbs->bind_cmd_buffer, view_id_size); cbs 827 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_collect_so_targets(struct vmw_ctx_binding_state *cbs, cbs 834 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c SVGA3dSoTarget *so_buffer = (SVGA3dSoTarget *) cbs->bind_cmd_buffer; cbs 836 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_cmd_count = 0; cbs 837 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_first_slot = 0; cbs 840 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ++cbs->bind_cmd_count) { cbs 861 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_emit_set_so(struct vmw_ctx_binding_state *cbs) cbs 863 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c const struct vmw_ctx_bindinfo *loc = &cbs->so_targets[0].bi; cbs 869 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c const struct vmw_resource *ctx = vmw_cbs_context(cbs); cbs 871 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_collect_so_targets(cbs, loc, SVGA3D_DX_MAX_SOTARGETS); cbs 872 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (cbs->bind_cmd_count == 0) cbs 875 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c so_target_size = cbs->bind_cmd_count*sizeof(SVGA3dSoTarget); cbs 883 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c memcpy(&cmd[1], cbs->bind_cmd_buffer, so_target_size); cbs 897 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_binding_emit_dirty_ps(struct vmw_ctx_binding_state *cbs) cbs 899 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_dx_shader_bindings *sb = &cbs->per_shader[0]; cbs 907 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ret = vmw_emit_set_sr(cbs, i); cbs 932 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_collect_dirty_vbs(struct vmw_ctx_binding_state *cbs, cbs 940 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c SVGA3dVertexBuffer *vbs = (SVGA3dVertexBuffer *) &cbs->bind_cmd_buffer; cbs 942 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_cmd_count = 0; cbs 945 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_first_slot = i; cbs 958 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_cmd_count++; cbs 973 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_emit_set_vb(struct vmw_ctx_binding_state *cbs) cbs 976 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c &cbs->vertex_buffers[0].bi; cbs 982 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c const struct vmw_resource *ctx = vmw_cbs_context(cbs); cbs 984 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_collect_dirty_vbs(cbs, loc, cbs->dirty_vb, cbs 986 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (cbs->bind_cmd_count == 0) cbs 989 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c set_vb_size = cbs->bind_cmd_count*sizeof(SVGA3dVertexBuffer); cbs 997 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cmd->body.startBuffer = cbs->bind_first_slot; cbs 999 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c memcpy(&cmd[1], cbs->bind_cmd_buffer, set_vb_size); cbs 1002 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c bitmap_clear(cbs->dirty_vb, cbs 1003 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->bind_first_slot, cbs->bind_cmd_count); cbs 1018 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_binding_emit_dirty(struct vmw_ctx_binding_state *cbs) cbs 1023 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c while ((hit = find_next_bit(&cbs->dirty, VMW_BINDING_NUM_BITS, hit)) cbs 1028 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ret = vmw_emit_set_rt(cbs); cbs 1031 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ret = vmw_binding_emit_dirty_ps(cbs); cbs 1034 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ret = vmw_emit_set_so(cbs); cbs 1037 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ret = vmw_emit_set_vb(cbs); cbs 1045 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c __clear_bit(hit, &cbs->dirty); cbs 1063 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_binding_state *cbs = cbs 1066 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c __set_bit(biv->slot, cbs->per_shader[biv->shader_slot].dirty_sr); cbs 1068 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c &cbs->per_shader[biv->shader_slot].dirty); cbs 1069 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c __set_bit(VMW_BINDING_PS_BIT, &cbs->dirty); cbs 1083 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_binding_state *cbs = cbs 1086 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c __set_bit(VMW_BINDING_RT_BIT, &cbs->dirty); cbs 1100 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_binding_state *cbs = cbs 1103 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c __set_bit(VMW_BINDING_SO_BIT, &cbs->dirty); cbs 1119 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_binding_state *cbs = cbs 1122 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c __set_bit(bivb->slot, cbs->dirty_vb); cbs 1123 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c __set_bit(VMW_BINDING_VB_BIT, &cbs->dirty); cbs 1176 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_ctx_binding_state *cbs; cbs 1183 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ret = ttm_mem_global_alloc(vmw_mem_glob(dev_priv), sizeof(*cbs), cbs 1188 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs = vzalloc(sizeof(*cbs)); cbs 1189 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c if (!cbs) { cbs 1190 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ttm_mem_global_free(vmw_mem_glob(dev_priv), sizeof(*cbs)); cbs 1194 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c cbs->dev_priv = dev_priv; cbs 1195 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c INIT_LIST_HEAD(&cbs->list); cbs 1197 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c return cbs; cbs 1206 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_state_free(struct vmw_ctx_binding_state *cbs) cbs 1208 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct vmw_private *dev_priv = cbs->dev_priv; cbs 1210 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vfree(cbs); cbs 1211 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c ttm_mem_global_free(vmw_mem_glob(dev_priv), sizeof(*cbs)); cbs 1223 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct list_head *vmw_binding_state_list(struct vmw_ctx_binding_state *cbs) cbs 1225 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c return &cbs->list; cbs 1236 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs) cbs 1240 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c list_for_each_entry_safe(entry, next, &cbs->list, ctx_list) cbs 192 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_add(struct vmw_ctx_binding_state *cbs, cbs 200 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern int vmw_binding_rebind_all(struct vmw_ctx_binding_state *cbs); cbs 201 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_state_kill(struct vmw_ctx_binding_state *cbs); cbs 202 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_state_scrub(struct vmw_ctx_binding_state *cbs); cbs 205 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_state_free(struct vmw_ctx_binding_state *cbs); cbs 207 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h vmw_binding_state_list(struct vmw_ctx_binding_state *cbs); cbs 208 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs); cbs 37 drivers/gpu/drm/vmwgfx/vmwgfx_context.c struct vmw_ctx_binding_state *cbs; cbs 151 drivers/gpu/drm/vmwgfx/vmwgfx_context.c vmw_binding_state_kill(uctx->cbs); cbs 202 drivers/gpu/drm/vmwgfx/vmwgfx_context.c uctx->cbs = vmw_binding_state_alloc(dev_priv); cbs 203 drivers/gpu/drm/vmwgfx/vmwgfx_context.c if (IS_ERR(uctx->cbs)) { cbs 204 drivers/gpu/drm/vmwgfx/vmwgfx_context.c ret = PTR_ERR(uctx->cbs); cbs 386 drivers/gpu/drm/vmwgfx/vmwgfx_context.c vmw_binding_state_scrub(uctx->cbs); cbs 550 drivers/gpu/drm/vmwgfx/vmwgfx_context.c vmw_binding_state_scrub(uctx->cbs); cbs 684 drivers/gpu/drm/vmwgfx/vmwgfx_context.c if (ctx->cbs) cbs 685 drivers/gpu/drm/vmwgfx/vmwgfx_context.c vmw_binding_state_free(ctx->cbs); cbs 831 drivers/gpu/drm/vmwgfx/vmwgfx_context.c return vmw_binding_state_list(uctx->cbs); cbs 861 drivers/gpu/drm/vmwgfx/vmwgfx_context.c return container_of(ctx, struct vmw_user_context, res)->cbs; cbs 266 drivers/infiniband/hw/qedr/qedr_roce_cm.c struct qed_ll2_cbs cbs; cbs 270 drivers/infiniband/hw/qedr/qedr_roce_cm.c cbs.rx_comp_cb = qedr_ll2_complete_rx_packet; cbs 271 drivers/infiniband/hw/qedr/qedr_roce_cm.c cbs.tx_comp_cb = qedr_ll2_complete_tx_packet; cbs 272 drivers/infiniband/hw/qedr/qedr_roce_cm.c cbs.rx_release_cb = qedr_ll2_release_rx_packet; cbs 273 drivers/infiniband/hw/qedr/qedr_roce_cm.c cbs.tx_release_cb = qedr_ll2_complete_tx_packet; cbs 274 drivers/infiniband/hw/qedr/qedr_roce_cm.c cbs.cookie = dev; cbs 289 drivers/infiniband/hw/qedr/qedr_roce_cm.c data.cbs = &cbs; cbs 498 drivers/iommu/arm-smmu.c struct arm_smmu_cb *cb = &smmu_domain->smmu->cbs[cfg->cbndx]; cbs 549 drivers/iommu/arm-smmu.c struct arm_smmu_cb *cb = &smmu->cbs[idx]; cbs 840 drivers/iommu/arm-smmu.c smmu->cbs[cfg->cbndx].cfg = NULL; cbs 1782 drivers/iommu/arm-smmu.c smmu->cbs = devm_kcalloc(smmu->dev, smmu->num_context_banks, cbs 1783 drivers/iommu/arm-smmu.c sizeof(*smmu->cbs), GFP_KERNEL); cbs 1784 drivers/iommu/arm-smmu.c if (!smmu->cbs) cbs 254 drivers/iommu/arm-smmu.h struct arm_smmu_cb *cbs; cbs 638 drivers/misc/sgi-gru/gru_instructions.h struct gru_control_block_status *cbs = (void *)cb; cbs 640 drivers/misc/sgi-gru/gru_instructions.h return cbs->istatus; cbs 646 drivers/misc/sgi-gru/gru_instructions.h struct gru_control_block_status *cbs = (void *)cb; cbs 648 drivers/misc/sgi-gru/gru_instructions.h return cbs->isubstatus & CBSS_MSG_QUEUE_MASK; cbs 654 drivers/misc/sgi-gru/gru_instructions.h struct gru_control_block_status *cbs = (void *)cb; cbs 656 drivers/misc/sgi-gru/gru_instructions.h return cbs->isubstatus; cbs 667 drivers/misc/sgi-gru/gru_instructions.h struct gru_control_block_status *cbs = (void *)cb; cbs 670 drivers/misc/sgi-gru/gru_instructions.h ret = cbs->istatus; cbs 132 drivers/net/ethernet/amd/xgbe/xgbe-dcb.c ets->cbs = pdata->ets->cbs; cbs 903 drivers/net/ethernet/broadcom/bcmsysport.c cb = &ring->cbs[ring->clean_index]; cbs 1323 drivers/net/ethernet/broadcom/bcmsysport.c cb = &ring->cbs[ring->curr_desc]; cbs 1475 drivers/net/ethernet/broadcom/bcmsysport.c ring->cbs = kcalloc(size, sizeof(struct bcm_sysport_cb), GFP_KERNEL); cbs 1476 drivers/net/ethernet/broadcom/bcmsysport.c if (!ring->cbs) { cbs 1565 drivers/net/ethernet/broadcom/bcmsysport.c if (!ring->cbs) cbs 1573 drivers/net/ethernet/broadcom/bcmsysport.c kfree(ring->cbs); cbs 1574 drivers/net/ethernet/broadcom/bcmsysport.c ring->cbs = NULL; cbs 719 drivers/net/ethernet/broadcom/bcmsysport.h struct bcm_sysport_cb *cbs; /* Transmit control blocks */ cbs 504 drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c ets->cbs = my_ets->cbs; cbs 1240 drivers/net/ethernet/broadcom/genet/bcmgenet.c tx_cb_ptr = ring->cbs; cbs 1257 drivers/net/ethernet/broadcom/genet/bcmgenet.c tx_cb_ptr = ring->cbs; cbs 1947 drivers/net/ethernet/broadcom/genet/bcmgenet.c cb = ring->cbs + i; cbs 2135 drivers/net/ethernet/broadcom/genet/bcmgenet.c ring->cbs = priv->tx_cbs + start_ptr; cbs 2192 drivers/net/ethernet/broadcom/genet/bcmgenet.c ring->cbs = priv->rx_cbs + start_ptr; cbs 563 drivers/net/ethernet/broadcom/genet/bcmgenet.h struct enet_cb *cbs; /* tx ring buffer control block*/ cbs 592 drivers/net/ethernet/broadcom/genet/bcmgenet.h struct enet_cb *cbs; /* Rx ring buffer control block */ cbs 537 drivers/net/ethernet/intel/e100.c struct param_range cbs; cbs 559 drivers/net/ethernet/intel/e100.c struct cb *cbs; cbs 1042 drivers/net/ethernet/intel/e100.c struct param_range cbs = { .min = 64, .max = 256, .count = 128 }; cbs 1050 drivers/net/ethernet/intel/e100.c nic->params.cbs = cbs; cbs 1827 drivers/net/ethernet/intel/e100.c (int)(((void*)cb - (void*)nic->cbs)/sizeof(struct cb)), cbs 1857 drivers/net/ethernet/intel/e100.c if (nic->cbs) { cbs 1858 drivers/net/ethernet/intel/e100.c while (nic->cbs_avail != nic->params.cbs.count) { cbs 1870 drivers/net/ethernet/intel/e100.c dma_pool_free(nic->cbs_pool, nic->cbs, nic->cbs_dma_addr); cbs 1871 drivers/net/ethernet/intel/e100.c nic->cbs = NULL; cbs 1876 drivers/net/ethernet/intel/e100.c nic->cbs; cbs 1882 drivers/net/ethernet/intel/e100.c unsigned int i, count = nic->params.cbs.count; cbs 1888 drivers/net/ethernet/intel/e100.c nic->cbs = dma_pool_zalloc(nic->cbs_pool, GFP_KERNEL, cbs 1890 drivers/net/ethernet/intel/e100.c if (!nic->cbs) cbs 1893 drivers/net/ethernet/intel/e100.c for (cb = nic->cbs, i = 0; i < count; cb++, i++) { cbs 1894 drivers/net/ethernet/intel/e100.c cb->next = (i + 1 < count) ? cb + 1 : nic->cbs; cbs 1895 drivers/net/ethernet/intel/e100.c cb->prev = (i == 0) ? nic->cbs + count - 1 : cb - 1; cbs 1902 drivers/net/ethernet/intel/e100.c nic->cb_to_use = nic->cb_to_send = nic->cb_to_clean = nic->cbs; cbs 2554 drivers/net/ethernet/intel/e100.c struct param_range *cbs = &nic->params.cbs; cbs 2557 drivers/net/ethernet/intel/e100.c ring->tx_max_pending = cbs->max; cbs 2559 drivers/net/ethernet/intel/e100.c ring->tx_pending = cbs->count; cbs 2567 drivers/net/ethernet/intel/e100.c struct param_range *cbs = &nic->params.cbs; cbs 2576 drivers/net/ethernet/intel/e100.c cbs->count = max(ring->tx_pending, cbs->min); cbs 2577 drivers/net/ethernet/intel/e100.c cbs->count = min(cbs->count, cbs->max); cbs 2579 drivers/net/ethernet/intel/e100.c rfds->count, cbs->count); cbs 2942 drivers/net/ethernet/intel/e100.c nic->params.cbs.max * sizeof(struct cb), cbs 17 drivers/net/ethernet/intel/fm10k/fm10k_dcbnl.c ets->cbs = 0; cbs 55 drivers/net/ethernet/intel/i40e/i40e_dcb.c etscfg->cbs = (u8)((buf[offset] & I40E_IEEE_ETS_CBS_MASK) >> cbs 292 drivers/net/ethernet/intel/i40e/i40e_dcb.c etscfg->cbs = 0; cbs 44 drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c ets->cbs = dcbxcfg->etscfg.cbs; cbs 1003 drivers/net/ethernet/intel/i40e/i40e_debugfs.c cfg->etscfg.willing, cfg->etscfg.cbs, cbs 1033 drivers/net/ethernet/intel/i40e/i40e_debugfs.c r_cfg->etscfg.cbs, r_cfg->etscfg.maxtcs); cbs 523 drivers/net/ethernet/intel/i40e/i40e_type.h u8 cbs; cbs 258 drivers/net/ethernet/intel/ice/ice_dcb.c etscfg->cbs = ((buf[0] & ICE_IEEE_ETS_CBS_M) >> ICE_IEEE_ETS_CBS_S); cbs 422 drivers/net/ethernet/intel/ice/ice_dcb.c etscfg->cbs = 0; cbs 296 drivers/net/ethernet/intel/ice/ice_type.h u8 cbs; cbs 499 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c ets->cbs = my_ets->cbs; cbs 297 drivers/net/ethernet/mellanox/mlx4/en_dcb_nl.c ets->cbs = my_ets->cbs; cbs 3351 drivers/net/ethernet/mellanox/mlxsw/reg.h MLXSW_ITEM32(reg, qpcr, cbs, 0x08, 24, 6); cbs 3398 drivers/net/ethernet/mellanox/mlxsw/reg.h bool bytes, u32 cir, u16 cbs) cbs 3409 drivers/net/ethernet/mellanox/mlxsw/reg.h mlxsw_reg_qpcr_cbs_set(payload, cbs); cbs 37 drivers/net/ethernet/mscc/ocelot_police.c u32 cbs; /* CBS in bytes/frames (ignored in SLB mode) */ cbs 47 drivers/net/ethernet/mscc/ocelot_police.c u32 cir = 0, cbs = 0, pir = 0, pbs = 0; cbs 69 drivers/net/ethernet/mscc/ocelot_police.c cbs = conf->cbs; cbs 70 drivers/net/ethernet/mscc/ocelot_police.c if (cir == 0 && cbs == 0) { cbs 76 drivers/net/ethernet/mscc/ocelot_police.c cbs = DIV_ROUND_UP(cbs, 4096); cbs 77 drivers/net/ethernet/mscc/ocelot_police.c cbs = (cbs ? cbs : 1); /* No zero burst size */ cbs 140 drivers/net/ethernet/mscc/ocelot_police.c if (cbs > cbs_max) { cbs 164 drivers/net/ethernet/mscc/ocelot_police.c ANA_POL_CIR_CFG_CIR_BURST(cbs), cbs 47 drivers/net/ethernet/netronome/nfp/flower/qos_conf.c __be32 cbs; cbs 122 drivers/net/ethernet/netronome/nfp/flower/qos_conf.c config->cbs = cpu_to_be32(burst); cbs 2183 drivers/net/ethernet/qlogic/qed/qed_dcbx.c ets->cbs = params->ets_cbs; cbs 2225 drivers/net/ethernet/qlogic/qed/qed_dcbx.c dcbx_set.config.params.ets_cbs = ets->cbs; cbs 2625 drivers/net/ethernet/qlogic/qed/qed_iwarp.c struct qed_ll2_cbs cbs; cbs 2645 drivers/net/ethernet/qlogic/qed/qed_iwarp.c cbs.rx_comp_cb = qed_iwarp_ll2_comp_syn_pkt; cbs 2646 drivers/net/ethernet/qlogic/qed/qed_iwarp.c cbs.rx_release_cb = qed_iwarp_ll2_rel_rx_pkt; cbs 2647 drivers/net/ethernet/qlogic/qed/qed_iwarp.c cbs.tx_comp_cb = qed_iwarp_ll2_comp_tx_pkt; cbs 2648 drivers/net/ethernet/qlogic/qed/qed_iwarp.c cbs.tx_release_cb = qed_iwarp_ll2_rel_tx_pkt; cbs 2649 drivers/net/ethernet/qlogic/qed/qed_iwarp.c cbs.slowpath_cb = NULL; cbs 2650 drivers/net/ethernet/qlogic/qed/qed_iwarp.c cbs.cookie = p_hwfn; cbs 2661 drivers/net/ethernet/qlogic/qed/qed_iwarp.c data.cbs = &cbs; cbs 2708 drivers/net/ethernet/qlogic/qed/qed_iwarp.c cbs.rx_comp_cb = qed_iwarp_ll2_comp_mpa_pkt; cbs 2709 drivers/net/ethernet/qlogic/qed/qed_iwarp.c cbs.slowpath_cb = qed_iwarp_ll2_slowpath; cbs 2725 drivers/net/ethernet/qlogic/qed/qed_iwarp.c data.cbs = &cbs; cbs 81 drivers/net/ethernet/qlogic/qed/qed_ll2.c const struct qed_ll2_cb_ops *cbs; cbs 106 drivers/net/ethernet/qlogic/qed/qed_ll2.c if (cdev->ll2->cbs && cdev->ll2->cbs->tx_cb) cbs 107 drivers/net/ethernet/qlogic/qed/qed_ll2.c cdev->ll2->cbs->tx_cb(cdev->ll2->cb_cookie, skb, cbs 222 drivers/net/ethernet/qlogic/qed/qed_ll2.c if (cdev->ll2->cbs && cdev->ll2->cbs->rx_cb) { cbs 226 drivers/net/ethernet/qlogic/qed/qed_ll2.c cdev->ll2->cbs->rx_cb(cdev->ll2->cb_cookie, skb, cbs 334 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_conn->cbs.tx_release_cb(p_ll2_conn->cbs.cookie, cbs 395 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_conn->cbs.tx_comp_cb(p_ll2_conn->cbs.cookie, cbs 457 drivers/net/ethernet/qlogic/qed/qed_ll2.c if (!p_ll2_conn->cbs.slowpath_cb) { cbs 465 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_conn->cbs.slowpath_cb(p_ll2_conn->cbs.cookie, cbs 513 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_conn->cbs.rx_comp_cb(p_ll2_conn->cbs.cookie, &data); cbs 600 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_conn->cbs.rx_release_cb(p_ll2_conn->cbs.cookie, cbs 1303 drivers/net/ethernet/qlogic/qed/qed_ll2.c qed_ll2_set_cbs(struct qed_ll2_info *p_ll2_info, const struct qed_ll2_cbs *cbs) cbs 1305 drivers/net/ethernet/qlogic/qed/qed_ll2.c if (!cbs || (!cbs->rx_comp_cb || cbs 1306 drivers/net/ethernet/qlogic/qed/qed_ll2.c !cbs->rx_release_cb || cbs 1307 drivers/net/ethernet/qlogic/qed/qed_ll2.c !cbs->tx_comp_cb || !cbs->tx_release_cb || !cbs->cookie)) cbs 1310 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_info->cbs.rx_comp_cb = cbs->rx_comp_cb; cbs 1311 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_info->cbs.rx_release_cb = cbs->rx_release_cb; cbs 1312 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_info->cbs.tx_comp_cb = cbs->tx_comp_cb; cbs 1313 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_info->cbs.tx_release_cb = cbs->tx_release_cb; cbs 1314 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_info->cbs.slowpath_cb = cbs->slowpath_cb; cbs 1315 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_info->cbs.cookie = cbs->cookie; cbs 1392 drivers/net/ethernet/qlogic/qed/qed_ll2.c rc = qed_ll2_set_cbs(p_ll2_info, data->cbs); cbs 2215 drivers/net/ethernet/qlogic/qed/qed_ll2.c cdev->ll2->cbs = ops; cbs 2241 drivers/net/ethernet/qlogic/qed/qed_ll2.c data->cbs = &ll2_cbs; cbs 2426 drivers/net/ethernet/qlogic/qed/qed_ll2.c WARN_ON(!cdev->ll2->cbs); cbs 128 drivers/net/ethernet/qlogic/qed/qed_ll2.h struct qed_ll2_cbs cbs; cbs 246 drivers/video/fbdev/nvidia/nv_hw.c int found, mclk_extra, mclk_loop, cbs, m1, p1; cbs 263 drivers/video/fbdev/nvidia/nv_hw.c cbs = 128; cbs 304 drivers/video/fbdev/nvidia/nv_hw.c cbs * 1000 * 1000 / 16 / nvclk_freq; cbs 307 drivers/video/fbdev/nvidia/nv_hw.c cbs * 1000 * 1000 / (8 * width) / cbs 341 drivers/video/fbdev/nvidia/nv_hw.c m1 = clwm + cbs - 512; cbs 420 drivers/video/fbdev/nvidia/nv_hw.c int found, mclk_extra, mclk_loop, cbs, m1; cbs 441 drivers/video/fbdev/nvidia/nv_hw.c cbs = 512; cbs 559 drivers/video/fbdev/nvidia/nv_hw.c cbs = 512; cbs 572 drivers/video/fbdev/nvidia/nv_hw.c m1 = clwm + cbs - 1024; /* Amount of overfill */ cbs 583 drivers/video/fbdev/nvidia/nv_hw.c if (cbs <= 32) { cbs 588 drivers/video/fbdev/nvidia/nv_hw.c cbs = cbs / 2; cbs 605 drivers/video/fbdev/nvidia/nv_hw.c if (clwm < (1024 - cbs + 8)) cbs 606 drivers/video/fbdev/nvidia/nv_hw.c clwm = 1024 - cbs + 8; cbs 611 drivers/video/fbdev/nvidia/nv_hw.c fifo->graphics_burst_size = cbs; cbs 661 drivers/video/fbdev/riva/riva_hw.c int found, mclk_extra, mclk_loop, cbs, m1, p1; cbs 681 drivers/video/fbdev/riva/riva_hw.c cbs = 128; cbs 722 drivers/video/fbdev/riva/riva_hw.c video_fill_us = cbs*1000*1000 / 16 / nvclk_freq ; cbs 724 drivers/video/fbdev/riva/riva_hw.c video_fill_us = cbs*1000*1000 / (8 * width) / mclk_freq; cbs 755 drivers/video/fbdev/riva/riva_hw.c m1 = clwm + cbs - 512; cbs 848 drivers/video/fbdev/riva/riva_hw.c int found, mclk_extra, mclk_loop, cbs, m1; cbs 873 drivers/video/fbdev/riva/riva_hw.c cbs = 512; cbs 1003 drivers/video/fbdev/riva/riva_hw.c cbs = 512; cbs 1004 drivers/video/fbdev/riva/riva_hw.c us_extra = (cbs * 1000 * 1000)/ (8*width)/mclk_freq ; cbs 1019 drivers/video/fbdev/riva/riva_hw.c m1 = clwm + cbs - 1024; /* Amount of overfill */ cbs 1031 drivers/video/fbdev/riva/riva_hw.c if(cbs <= 32) { cbs 1034 drivers/video/fbdev/riva/riva_hw.c cbs = cbs/2; /* reduce the burst size */ cbs 1051 drivers/video/fbdev/riva/riva_hw.c if(clwm < (1024-cbs+8)) clwm = 1024-cbs+8; cbs 1054 drivers/video/fbdev/riva/riva_hw.c fifo->graphics_lwm = data; fifo->graphics_burst_size = cbs; cbs 172 drivers/virtio/virtio_input.c vq_callback_t *cbs[] = { virtinput_recv_events, cbs 177 drivers/virtio/virtio_input.c err = virtio_find_vqs(vi->vdev, 2, vqs, cbs, names, NULL); cbs 3236 fs/nfsd/nfs4state.c static __be32 nfsd4_check_cb_sec(struct nfsd4_cb_sec *cbs) cbs 3238 fs/nfsd/nfs4state.c switch (cbs->flavor) { cbs 525 fs/nfsd/nfs4xdr.c static __be32 nfsd4_decode_cb_sec(struct nfsd4_compoundargs *argp, struct nfsd4_cb_sec *cbs) cbs 538 fs/nfsd/nfs4xdr.c cbs->flavor = (u32)(-1); cbs 541 fs/nfsd/nfs4xdr.c cbs->flavor = 0; cbs 548 fs/nfsd/nfs4xdr.c if (cbs->flavor == (u32)(-1)) cbs 549 fs/nfsd/nfs4xdr.c cbs->flavor = RPC_AUTH_NULL; cbs 570 fs/nfsd/nfs4xdr.c if (cbs->flavor == (u32)(-1)) { cbs 574 fs/nfsd/nfs4xdr.c cbs->uid = kuid; cbs 575 fs/nfsd/nfs4xdr.c cbs->gid = kgid; cbs 576 fs/nfsd/nfs4xdr.c cbs->flavor = RPC_AUTH_UNIX; cbs 462 include/drm/drm_dp_mst_helper.h const struct drm_dp_mst_topology_cbs *cbs; cbs 186 include/linux/qed/qed_ll2_if.h const struct qed_ll2_cbs *cbs; cbs 61 include/uapi/linux/dcbnl.h __u8 cbs; cbs 16 kernel/bpf/disasm.c static const char *__func_get_name(const struct bpf_insn_cbs *cbs, cbs 27 kernel/bpf/disasm.c if (cbs && cbs->cb_call) cbs 28 kernel/bpf/disasm.c return cbs->cb_call(cbs->private_data, insn); cbs 36 kernel/bpf/disasm.c static const char *__func_imm_name(const struct bpf_insn_cbs *cbs, cbs 40 kernel/bpf/disasm.c if (cbs && cbs->cb_imm) cbs 41 kernel/bpf/disasm.c return cbs->cb_imm(cbs->private_data, insn, full_imm); cbs 117 kernel/bpf/disasm.c void print_bpf_insn(const struct bpf_insn_cbs *cbs, cbs 121 kernel/bpf/disasm.c const bpf_insn_print_t verbose = cbs->cb_print; cbs 127 kernel/bpf/disasm.c verbose(cbs->private_data, "BUG_alu64_%02x\n", insn->code); cbs 129 kernel/bpf/disasm.c print_bpf_end_insn(verbose, cbs->private_data, insn); cbs 131 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) %c%d = -%c%d\n", cbs 136 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) %c%d %s %c%d\n", cbs 143 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) %c%d %s %d\n", cbs 151 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) *(%s *)(r%d %+d) = r%d\n", cbs 157 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) lock *(%s *)(r%d %+d) += r%d\n", cbs 163 kernel/bpf/disasm.c verbose(cbs->private_data, "BUG_%02x\n", insn->code); cbs 166 kernel/bpf/disasm.c verbose(cbs->private_data, "BUG_st_%02x\n", insn->code); cbs 169 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) *(%s *)(r%d %+d) = %d\n", cbs 176 kernel/bpf/disasm.c verbose(cbs->private_data, "BUG_ldx_%02x\n", insn->code); cbs 179 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) r%d = *(%s *)(r%d %+d)\n", cbs 185 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) r0 = *(%s *)skb[%d]\n", cbs 190 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) r0 = *(%s *)skb[r%d + %d]\n", cbs 207 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) r%d = %s\n", cbs 209 kernel/bpf/disasm.c __func_imm_name(cbs, insn, imm, cbs 212 kernel/bpf/disasm.c verbose(cbs->private_data, "BUG_ld_%02x\n", insn->code); cbs 222 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) call pc%s\n", cbs 224 kernel/bpf/disasm.c __func_get_name(cbs, insn, cbs 228 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) call %s#%d\n", insn->code, cbs 229 kernel/bpf/disasm.c __func_get_name(cbs, insn, cbs 234 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) goto pc%+d\n", cbs 237 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) exit\n", insn->code); cbs 239 kernel/bpf/disasm.c verbose(cbs->private_data, cbs 247 kernel/bpf/disasm.c verbose(cbs->private_data, cbs 255 kernel/bpf/disasm.c verbose(cbs->private_data, "(%02x) %s\n", cbs 37 kernel/bpf/disasm.h void print_bpf_insn(const struct bpf_insn_cbs *cbs, cbs 1445 kernel/bpf/verifier.c const struct bpf_insn_cbs cbs = { cbs 1462 kernel/bpf/verifier.c print_bpf_insn(&cbs, insn, env->allow_ptr_leaks); cbs 7636 kernel/bpf/verifier.c const struct bpf_insn_cbs cbs = { cbs 7643 kernel/bpf/verifier.c print_bpf_insn(&cbs, insn, env->allow_ptr_leaks); cbs 6038 kernel/events/core.c int perf_register_guest_info_callbacks(struct perf_guest_info_callbacks *cbs) cbs 6040 kernel/events/core.c perf_guest_cbs = cbs; cbs 6045 kernel/events/core.c int perf_unregister_guest_info_callbacks(struct perf_guest_info_callbacks *cbs) cbs 515 kernel/rcu/srcutree.c bool cbs; cbs 546 kernel/rcu/srcutree.c cbs = false; cbs 549 kernel/rcu/srcutree.c cbs = snp->srcu_have_cbs[idx] == gpseq; cbs 557 kernel/rcu/srcutree.c if (cbs) cbs 662 kernel/rcu/tree_stall.h unsigned long cbs; cbs 681 kernel/rcu/tree_stall.h cbs = rcu_get_n_cbs_cpu(cpu); cbs 682 kernel/rcu/tree_stall.h if (!cbs) cbs 686 kernel/rcu/tree_stall.h pr_cont(" %d: %lu", cpu, cbs); cbs 687 kernel/rcu/tree_stall.h if (cbs <= max_cbs) cbs 689 kernel/rcu/tree_stall.h max_cbs = cbs; cbs 252 net/sched/sch_cbs.c struct tc_cbs_qopt_offload cbs = { }; cbs 266 net/sched/sch_cbs.c cbs.queue = q->queue; cbs 267 net/sched/sch_cbs.c cbs.enable = 0; cbs 269 net/sched/sch_cbs.c err = ops->ndo_setup_tc(dev, TC_SETUP_QDISC_CBS, &cbs); cbs 272 net/sched/sch_cbs.c cbs.queue); cbs 280 net/sched/sch_cbs.c struct tc_cbs_qopt_offload cbs = { }; cbs 288 net/sched/sch_cbs.c cbs.queue = q->queue; cbs 290 net/sched/sch_cbs.c cbs.enable = 1; cbs 291 net/sched/sch_cbs.c cbs.hicredit = opt->hicredit; cbs 292 net/sched/sch_cbs.c cbs.locredit = opt->locredit; cbs 293 net/sched/sch_cbs.c cbs.idleslope = opt->idleslope; cbs 294 net/sched/sch_cbs.c cbs.sendslope = opt->sendslope; cbs 296 net/sched/sch_cbs.c err = ops->ndo_setup_tc(dev, TC_SETUP_QDISC_CBS, &cbs); cbs 209 tools/bpf/bpftool/xlated_dumper.c const struct bpf_insn_cbs cbs = { cbs 258 tools/bpf/bpftool/xlated_dumper.c print_bpf_insn(&cbs, insn + i, true); cbs 294 tools/bpf/bpftool/xlated_dumper.c const struct bpf_insn_cbs cbs = { cbs 340 tools/bpf/bpftool/xlated_dumper.c print_bpf_insn(&cbs, insn + i, true); cbs 357 tools/bpf/bpftool/xlated_dumper.c const struct bpf_insn_cbs cbs = { cbs 369 tools/bpf/bpftool/xlated_dumper.c print_bpf_insn(&cbs, cur, true);