cbs               160 drivers/dma/ppc4xx/adma.c 			cb->cbc, cb->cbbc, cb->cbs,
cbs               713 drivers/dma/ppc4xx/adma.c 	xor_hw_desc->cbs = 0;
cbs                73 drivers/dma/ppc4xx/xor.h 	u32	cbs;		/* status */
cbs               421 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_mst_types.c 	aconnector->mst_mgr.cbs = &dm_mst_cbs;
cbs              1698 drivers/gpu/drm/drm_dp_mst_topology.c 		port->connector = (*mstb->mgr->cbs->add_connector)(mstb->mgr,
cbs              1717 drivers/gpu/drm/drm_dp_mst_topology.c 		(*mstb->mgr->cbs->register_connector)(port->connector);
cbs              3761 drivers/gpu/drm/drm_dp_mst_topology.c 		mgr->cbs->destroy_connector(mgr, port->connector);
cbs               670 drivers/gpu/drm/i915/display/intel_dp_mst.c 	intel_dp->mst_mgr.cbs = &mst_cbs;
cbs                58 drivers/gpu/drm/nouveau/dispnv04/arb.c 	int found, mclk_extra, mclk_loop, cbs, m1, p1;
cbs                70 drivers/gpu/drm/nouveau/dispnv04/arb.c 	cbs = 128;
cbs                94 drivers/gpu/drm/nouveau/dispnv04/arb.c 		m1 = clwm + cbs - 512;
cbs               105 drivers/gpu/drm/nouveau/dispnv04/arb.c 		fifo->burst = cbs;
cbs              1386 drivers/gpu/drm/nouveau/dispnv50/disp.c 	mstm->mgr.cbs = &nv50_mstm;
cbs               657 drivers/gpu/drm/radeon/radeon_dp_mst.c 	radeon_connector->mst_mgr.cbs = &mst_cbs;
cbs               119 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_binding_emit_dirty(struct vmw_ctx_binding_state *cbs);
cbs               244 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_cbs_context(const struct vmw_ctx_binding_state *cbs)
cbs               246 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	if (list_empty(&cbs->list))
cbs               249 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	return list_first_entry(&cbs->list, struct vmw_ctx_bindinfo,
cbs               262 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c vmw_binding_loc(struct vmw_ctx_binding_state *cbs,
cbs               268 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	return (struct vmw_ctx_bindinfo *)((u8 *) cbs + offset);
cbs               297 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_add(struct vmw_ctx_binding_state *cbs,
cbs               302 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		vmw_binding_loc(cbs, bi->bt, shader_slot, slot);
cbs               310 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_add(&loc->ctx_list, &cbs->list);
cbs               321 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_binding_transfer(struct vmw_ctx_binding_state *cbs,
cbs               327 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		((unsigned long) cbs + offset);
cbs               337 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		list_add_tail(&loc->ctx_list, &cbs->list);
cbs               351 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_state_kill(struct vmw_ctx_binding_state *cbs)
cbs               355 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	vmw_binding_state_scrub(cbs);
cbs               356 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, &cbs->list, ctx_list)
cbs               369 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_state_scrub(struct vmw_ctx_binding_state *cbs)
cbs               373 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry(entry, &cbs->list, ctx_list) {
cbs               381 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	(void) vmw_binding_emit_dirty(cbs);
cbs               424 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		struct vmw_ctx_binding_state *cbs =
cbs               427 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		(void) vmw_binding_emit_dirty(cbs);
cbs               463 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c int vmw_binding_rebind_all(struct vmw_ctx_binding_state *cbs)
cbs               468 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry(entry, &cbs->list, ctx_list) {
cbs               483 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	return vmw_binding_emit_dirty(cbs);
cbs               669 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_collect_view_ids(struct vmw_ctx_binding_state *cbs,
cbs               677 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->bind_cmd_count = 0;
cbs               678 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->bind_first_slot = 0;
cbs               684 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		cbs->bind_cmd_buffer[cbs->bind_cmd_count++] =
cbs               704 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_collect_dirty_view_ids(struct vmw_ctx_binding_state *cbs,
cbs               713 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->bind_cmd_count = 0;
cbs               716 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->bind_first_slot = i;
cbs               720 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		cbs->bind_cmd_buffer[cbs->bind_cmd_count++] =
cbs               737 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_emit_set_sr(struct vmw_ctx_binding_state *cbs,
cbs               741 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		&cbs->per_shader[shader_slot].shader_res[0].bi;
cbs               747 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	const struct vmw_resource *ctx = vmw_cbs_context(cbs);
cbs               749 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	vmw_collect_dirty_view_ids(cbs, loc,
cbs               750 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 				   cbs->per_shader[shader_slot].dirty_sr,
cbs               752 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	if (cbs->bind_cmd_count == 0)
cbs               755 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	view_id_size = cbs->bind_cmd_count*sizeof(uint32);
cbs               764 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cmd->body.startView = cbs->bind_first_slot;
cbs               766 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	memcpy(&cmd[1], cbs->bind_cmd_buffer, view_id_size);
cbs               769 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	bitmap_clear(cbs->per_shader[shader_slot].dirty_sr,
cbs               770 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		     cbs->bind_first_slot, cbs->bind_cmd_count);
cbs               780 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_emit_set_rt(struct vmw_ctx_binding_state *cbs)
cbs               782 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	const struct vmw_ctx_bindinfo *loc = &cbs->render_targets[0].bi;
cbs               788 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	const struct vmw_resource *ctx = vmw_cbs_context(cbs);
cbs               790 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	vmw_collect_view_ids(cbs, loc, SVGA3D_MAX_SIMULTANEOUS_RENDER_TARGETS);
cbs               791 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	view_id_size = cbs->bind_cmd_count*sizeof(uint32);
cbs               800 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	if (cbs->ds_view.bi.ctx && !cbs->ds_view.bi.scrubbed)
cbs               801 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		cmd->body.depthStencilViewId = cbs->ds_view.bi.res->id;
cbs               805 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	memcpy(&cmd[1], cbs->bind_cmd_buffer, view_id_size);
cbs               827 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_collect_so_targets(struct vmw_ctx_binding_state *cbs,
cbs               834 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	SVGA3dSoTarget *so_buffer = (SVGA3dSoTarget *) cbs->bind_cmd_buffer;
cbs               836 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->bind_cmd_count = 0;
cbs               837 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->bind_first_slot = 0;
cbs               840 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		    ++cbs->bind_cmd_count) {
cbs               861 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_emit_set_so(struct vmw_ctx_binding_state *cbs)
cbs               863 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	const struct vmw_ctx_bindinfo *loc = &cbs->so_targets[0].bi;
cbs               869 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	const struct vmw_resource *ctx = vmw_cbs_context(cbs);
cbs               871 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	vmw_collect_so_targets(cbs, loc, SVGA3D_DX_MAX_SOTARGETS);
cbs               872 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	if (cbs->bind_cmd_count == 0)
cbs               875 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	so_target_size = cbs->bind_cmd_count*sizeof(SVGA3dSoTarget);
cbs               883 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	memcpy(&cmd[1], cbs->bind_cmd_buffer, so_target_size);
cbs               897 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_binding_emit_dirty_ps(struct vmw_ctx_binding_state *cbs)
cbs               899 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_dx_shader_bindings *sb = &cbs->per_shader[0];
cbs               907 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		ret = vmw_emit_set_sr(cbs, i);
cbs               932 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static void vmw_collect_dirty_vbs(struct vmw_ctx_binding_state *cbs,
cbs               940 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	SVGA3dVertexBuffer *vbs = (SVGA3dVertexBuffer *) &cbs->bind_cmd_buffer;
cbs               942 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->bind_cmd_count = 0;
cbs               945 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->bind_first_slot = i;
cbs               958 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		cbs->bind_cmd_count++;
cbs               973 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_emit_set_vb(struct vmw_ctx_binding_state *cbs)
cbs               976 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		&cbs->vertex_buffers[0].bi;
cbs               982 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	const struct vmw_resource *ctx = vmw_cbs_context(cbs);
cbs               984 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	vmw_collect_dirty_vbs(cbs, loc, cbs->dirty_vb,
cbs               986 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	if (cbs->bind_cmd_count == 0)
cbs               989 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	set_vb_size = cbs->bind_cmd_count*sizeof(SVGA3dVertexBuffer);
cbs               997 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cmd->body.startBuffer = cbs->bind_first_slot;
cbs               999 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	memcpy(&cmd[1], cbs->bind_cmd_buffer, set_vb_size);
cbs              1002 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	bitmap_clear(cbs->dirty_vb,
cbs              1003 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		     cbs->bind_first_slot, cbs->bind_cmd_count);
cbs              1018 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c static int vmw_binding_emit_dirty(struct vmw_ctx_binding_state *cbs)
cbs              1023 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	while ((hit = find_next_bit(&cbs->dirty, VMW_BINDING_NUM_BITS, hit))
cbs              1028 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			ret = vmw_emit_set_rt(cbs);
cbs              1031 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			ret = vmw_binding_emit_dirty_ps(cbs);
cbs              1034 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			ret = vmw_emit_set_so(cbs);
cbs              1037 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 			ret = vmw_emit_set_vb(cbs);
cbs              1045 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		__clear_bit(hit, &cbs->dirty);
cbs              1063 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_binding_state *cbs =
cbs              1066 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	__set_bit(biv->slot, cbs->per_shader[biv->shader_slot].dirty_sr);
cbs              1068 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		  &cbs->per_shader[biv->shader_slot].dirty);
cbs              1069 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	__set_bit(VMW_BINDING_PS_BIT, &cbs->dirty);
cbs              1083 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_binding_state *cbs =
cbs              1086 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	__set_bit(VMW_BINDING_RT_BIT, &cbs->dirty);
cbs              1100 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_binding_state *cbs =
cbs              1103 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	__set_bit(VMW_BINDING_SO_BIT, &cbs->dirty);
cbs              1119 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_binding_state *cbs =
cbs              1122 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	__set_bit(bivb->slot, cbs->dirty_vb);
cbs              1123 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	__set_bit(VMW_BINDING_VB_BIT, &cbs->dirty);
cbs              1176 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_ctx_binding_state *cbs;
cbs              1183 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	ret = ttm_mem_global_alloc(vmw_mem_glob(dev_priv), sizeof(*cbs),
cbs              1188 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs = vzalloc(sizeof(*cbs));
cbs              1189 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	if (!cbs) {
cbs              1190 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 		ttm_mem_global_free(vmw_mem_glob(dev_priv), sizeof(*cbs));
cbs              1194 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	cbs->dev_priv = dev_priv;
cbs              1195 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	INIT_LIST_HEAD(&cbs->list);
cbs              1197 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	return cbs;
cbs              1206 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_state_free(struct vmw_ctx_binding_state *cbs)
cbs              1208 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	struct vmw_private *dev_priv = cbs->dev_priv;
cbs              1210 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	vfree(cbs);
cbs              1211 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	ttm_mem_global_free(vmw_mem_glob(dev_priv), sizeof(*cbs));
cbs              1223 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c struct list_head *vmw_binding_state_list(struct vmw_ctx_binding_state *cbs)
cbs              1225 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	return &cbs->list;
cbs              1236 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs)
cbs              1240 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c 	list_for_each_entry_safe(entry, next, &cbs->list, ctx_list)
cbs               192 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_add(struct vmw_ctx_binding_state *cbs,
cbs               200 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern int vmw_binding_rebind_all(struct vmw_ctx_binding_state *cbs);
cbs               201 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_state_kill(struct vmw_ctx_binding_state *cbs);
cbs               202 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_state_scrub(struct vmw_ctx_binding_state *cbs);
cbs               205 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_state_free(struct vmw_ctx_binding_state *cbs);
cbs               207 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h vmw_binding_state_list(struct vmw_ctx_binding_state *cbs);
cbs               208 drivers/gpu/drm/vmwgfx/vmwgfx_binding.h extern void vmw_binding_state_reset(struct vmw_ctx_binding_state *cbs);
cbs                37 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 	struct vmw_ctx_binding_state *cbs;
cbs               151 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 		vmw_binding_state_kill(uctx->cbs);
cbs               202 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 	uctx->cbs = vmw_binding_state_alloc(dev_priv);
cbs               203 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 	if (IS_ERR(uctx->cbs)) {
cbs               204 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 		ret = PTR_ERR(uctx->cbs);
cbs               386 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 	vmw_binding_state_scrub(uctx->cbs);
cbs               550 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 	vmw_binding_state_scrub(uctx->cbs);
cbs               684 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 	if (ctx->cbs)
cbs               685 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 		vmw_binding_state_free(ctx->cbs);
cbs               831 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 	return vmw_binding_state_list(uctx->cbs);
cbs               861 drivers/gpu/drm/vmwgfx/vmwgfx_context.c 	return container_of(ctx, struct vmw_user_context, res)->cbs;
cbs               266 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	struct qed_ll2_cbs cbs;
cbs               270 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	cbs.rx_comp_cb = qedr_ll2_complete_rx_packet;
cbs               271 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	cbs.tx_comp_cb = qedr_ll2_complete_tx_packet;
cbs               272 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	cbs.rx_release_cb = qedr_ll2_release_rx_packet;
cbs               273 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	cbs.tx_release_cb = qedr_ll2_complete_tx_packet;
cbs               274 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	cbs.cookie = dev;
cbs               289 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	data.cbs = &cbs;
cbs               498 drivers/iommu/arm-smmu.c 	struct arm_smmu_cb *cb = &smmu_domain->smmu->cbs[cfg->cbndx];
cbs               549 drivers/iommu/arm-smmu.c 	struct arm_smmu_cb *cb = &smmu->cbs[idx];
cbs               840 drivers/iommu/arm-smmu.c 	smmu->cbs[cfg->cbndx].cfg = NULL;
cbs              1782 drivers/iommu/arm-smmu.c 	smmu->cbs = devm_kcalloc(smmu->dev, smmu->num_context_banks,
cbs              1783 drivers/iommu/arm-smmu.c 				 sizeof(*smmu->cbs), GFP_KERNEL);
cbs              1784 drivers/iommu/arm-smmu.c 	if (!smmu->cbs)
cbs               254 drivers/iommu/arm-smmu.h 	struct arm_smmu_cb		*cbs;
cbs               638 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_control_block_status *cbs = (void *)cb;
cbs               640 drivers/misc/sgi-gru/gru_instructions.h 	return cbs->istatus;
cbs               646 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_control_block_status *cbs = (void *)cb;
cbs               648 drivers/misc/sgi-gru/gru_instructions.h 	return cbs->isubstatus & CBSS_MSG_QUEUE_MASK;
cbs               654 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_control_block_status *cbs = (void *)cb;
cbs               656 drivers/misc/sgi-gru/gru_instructions.h 	return cbs->isubstatus;
cbs               667 drivers/misc/sgi-gru/gru_instructions.h 	struct gru_control_block_status *cbs = (void *)cb;
cbs               670 drivers/misc/sgi-gru/gru_instructions.h 	ret = cbs->istatus;
cbs               132 drivers/net/ethernet/amd/xgbe/xgbe-dcb.c 		ets->cbs = pdata->ets->cbs;
cbs               903 drivers/net/ethernet/broadcom/bcmsysport.c 		cb = &ring->cbs[ring->clean_index];
cbs              1323 drivers/net/ethernet/broadcom/bcmsysport.c 	cb = &ring->cbs[ring->curr_desc];
cbs              1475 drivers/net/ethernet/broadcom/bcmsysport.c 	ring->cbs = kcalloc(size, sizeof(struct bcm_sysport_cb), GFP_KERNEL);
cbs              1476 drivers/net/ethernet/broadcom/bcmsysport.c 	if (!ring->cbs) {
cbs              1565 drivers/net/ethernet/broadcom/bcmsysport.c 	if (!ring->cbs)
cbs              1573 drivers/net/ethernet/broadcom/bcmsysport.c 	kfree(ring->cbs);
cbs              1574 drivers/net/ethernet/broadcom/bcmsysport.c 	ring->cbs = NULL;
cbs               719 drivers/net/ethernet/broadcom/bcmsysport.h 	struct bcm_sysport_cb *cbs;	/* Transmit control blocks */
cbs               504 drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c 	ets->cbs = my_ets->cbs;
cbs              1240 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	tx_cb_ptr = ring->cbs;
cbs              1257 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	tx_cb_ptr = ring->cbs;
cbs              1947 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		cb = ring->cbs + i;
cbs              2135 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	ring->cbs = priv->tx_cbs + start_ptr;
cbs              2192 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	ring->cbs = priv->rx_cbs + start_ptr;
cbs               563 drivers/net/ethernet/broadcom/genet/bcmgenet.h 	struct enet_cb	*cbs;		/* tx ring buffer control block*/
cbs               592 drivers/net/ethernet/broadcom/genet/bcmgenet.h 	struct enet_cb	*cbs;		/* Rx ring buffer control block */
cbs               537 drivers/net/ethernet/intel/e100.c 	struct param_range cbs;
cbs               559 drivers/net/ethernet/intel/e100.c 	struct cb *cbs;
cbs              1042 drivers/net/ethernet/intel/e100.c 	struct param_range cbs  = { .min = 64, .max = 256, .count = 128 };
cbs              1050 drivers/net/ethernet/intel/e100.c 	nic->params.cbs = cbs;
cbs              1827 drivers/net/ethernet/intel/e100.c 			     (int)(((void*)cb - (void*)nic->cbs)/sizeof(struct cb)),
cbs              1857 drivers/net/ethernet/intel/e100.c 	if (nic->cbs) {
cbs              1858 drivers/net/ethernet/intel/e100.c 		while (nic->cbs_avail != nic->params.cbs.count) {
cbs              1870 drivers/net/ethernet/intel/e100.c 		dma_pool_free(nic->cbs_pool, nic->cbs, nic->cbs_dma_addr);
cbs              1871 drivers/net/ethernet/intel/e100.c 		nic->cbs = NULL;
cbs              1876 drivers/net/ethernet/intel/e100.c 		nic->cbs;
cbs              1882 drivers/net/ethernet/intel/e100.c 	unsigned int i, count = nic->params.cbs.count;
cbs              1888 drivers/net/ethernet/intel/e100.c 	nic->cbs = dma_pool_zalloc(nic->cbs_pool, GFP_KERNEL,
cbs              1890 drivers/net/ethernet/intel/e100.c 	if (!nic->cbs)
cbs              1893 drivers/net/ethernet/intel/e100.c 	for (cb = nic->cbs, i = 0; i < count; cb++, i++) {
cbs              1894 drivers/net/ethernet/intel/e100.c 		cb->next = (i + 1 < count) ? cb + 1 : nic->cbs;
cbs              1895 drivers/net/ethernet/intel/e100.c 		cb->prev = (i == 0) ? nic->cbs + count - 1 : cb - 1;
cbs              1902 drivers/net/ethernet/intel/e100.c 	nic->cb_to_use = nic->cb_to_send = nic->cb_to_clean = nic->cbs;
cbs              2554 drivers/net/ethernet/intel/e100.c 	struct param_range *cbs = &nic->params.cbs;
cbs              2557 drivers/net/ethernet/intel/e100.c 	ring->tx_max_pending = cbs->max;
cbs              2559 drivers/net/ethernet/intel/e100.c 	ring->tx_pending = cbs->count;
cbs              2567 drivers/net/ethernet/intel/e100.c 	struct param_range *cbs = &nic->params.cbs;
cbs              2576 drivers/net/ethernet/intel/e100.c 	cbs->count = max(ring->tx_pending, cbs->min);
cbs              2577 drivers/net/ethernet/intel/e100.c 	cbs->count = min(cbs->count, cbs->max);
cbs              2579 drivers/net/ethernet/intel/e100.c 		   rfds->count, cbs->count);
cbs              2942 drivers/net/ethernet/intel/e100.c 			   nic->params.cbs.max * sizeof(struct cb),
cbs                17 drivers/net/ethernet/intel/fm10k/fm10k_dcbnl.c 	ets->cbs = 0;
cbs                55 drivers/net/ethernet/intel/i40e/i40e_dcb.c 	etscfg->cbs = (u8)((buf[offset] & I40E_IEEE_ETS_CBS_MASK) >>
cbs               292 drivers/net/ethernet/intel/i40e/i40e_dcb.c 	etscfg->cbs = 0;
cbs                44 drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c 	ets->cbs = dcbxcfg->etscfg.cbs;
cbs              1003 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 				 cfg->etscfg.willing, cfg->etscfg.cbs,
cbs              1033 drivers/net/ethernet/intel/i40e/i40e_debugfs.c 				 r_cfg->etscfg.cbs, r_cfg->etscfg.maxtcs);
cbs               523 drivers/net/ethernet/intel/i40e/i40e_type.h 	u8 cbs;
cbs               258 drivers/net/ethernet/intel/ice/ice_dcb.c 	etscfg->cbs = ((buf[0] & ICE_IEEE_ETS_CBS_M) >> ICE_IEEE_ETS_CBS_S);
cbs               422 drivers/net/ethernet/intel/ice/ice_dcb.c 	etscfg->cbs = 0;
cbs               296 drivers/net/ethernet/intel/ice/ice_type.h 	u8 cbs;
cbs               499 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 	ets->cbs = my_ets->cbs;
cbs               297 drivers/net/ethernet/mellanox/mlx4/en_dcb_nl.c 	ets->cbs = my_ets->cbs;
cbs              3351 drivers/net/ethernet/mellanox/mlxsw/reg.h MLXSW_ITEM32(reg, qpcr, cbs, 0x08, 24, 6);
cbs              3398 drivers/net/ethernet/mellanox/mlxsw/reg.h 				       bool bytes, u32 cir, u16 cbs)
cbs              3409 drivers/net/ethernet/mellanox/mlxsw/reg.h 	mlxsw_reg_qpcr_cbs_set(payload, cbs);
cbs                37 drivers/net/ethernet/mscc/ocelot_police.c 	u32  cbs; /* CBS in bytes/frames (ignored in SLB mode) */
cbs                47 drivers/net/ethernet/mscc/ocelot_police.c 	u32 cir = 0, cbs = 0, pir = 0, pbs = 0;
cbs                69 drivers/net/ethernet/mscc/ocelot_police.c 			cbs = conf->cbs;
cbs                70 drivers/net/ethernet/mscc/ocelot_police.c 			if (cir == 0 && cbs == 0) {
cbs                76 drivers/net/ethernet/mscc/ocelot_police.c 				cbs = DIV_ROUND_UP(cbs, 4096);
cbs                77 drivers/net/ethernet/mscc/ocelot_police.c 				cbs = (cbs ? cbs : 1); /* No zero burst size */
cbs               140 drivers/net/ethernet/mscc/ocelot_police.c 	if (cbs > cbs_max) {
cbs               164 drivers/net/ethernet/mscc/ocelot_police.c 			 ANA_POL_CIR_CFG_CIR_BURST(cbs),
cbs                47 drivers/net/ethernet/netronome/nfp/flower/qos_conf.c 	__be32 cbs;
cbs               122 drivers/net/ethernet/netronome/nfp/flower/qos_conf.c 	config->cbs = cpu_to_be32(burst);
cbs              2183 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 	ets->cbs = params->ets_cbs;
cbs              2225 drivers/net/ethernet/qlogic/qed/qed_dcbx.c 	dcbx_set.config.params.ets_cbs = ets->cbs;
cbs              2625 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	struct qed_ll2_cbs cbs;
cbs              2645 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	cbs.rx_comp_cb = qed_iwarp_ll2_comp_syn_pkt;
cbs              2646 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	cbs.rx_release_cb = qed_iwarp_ll2_rel_rx_pkt;
cbs              2647 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	cbs.tx_comp_cb = qed_iwarp_ll2_comp_tx_pkt;
cbs              2648 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	cbs.tx_release_cb = qed_iwarp_ll2_rel_tx_pkt;
cbs              2649 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	cbs.slowpath_cb = NULL;
cbs              2650 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	cbs.cookie = p_hwfn;
cbs              2661 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	data.cbs = &cbs;
cbs              2708 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	cbs.rx_comp_cb = qed_iwarp_ll2_comp_mpa_pkt;
cbs              2709 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	cbs.slowpath_cb = qed_iwarp_ll2_slowpath;
cbs              2725 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	data.cbs = &cbs;
cbs                81 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	const struct qed_ll2_cb_ops *cbs;
cbs               106 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	if (cdev->ll2->cbs && cdev->ll2->cbs->tx_cb)
cbs               107 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		cdev->ll2->cbs->tx_cb(cdev->ll2->cb_cookie, skb,
cbs               222 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	if (cdev->ll2->cbs && cdev->ll2->cbs->rx_cb) {
cbs               226 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		cdev->ll2->cbs->rx_cb(cdev->ll2->cb_cookie, skb,
cbs               334 drivers/net/ethernet/qlogic/qed/qed_ll2.c 			p_ll2_conn->cbs.tx_release_cb(p_ll2_conn->cbs.cookie,
cbs               395 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		p_ll2_conn->cbs.tx_comp_cb(p_ll2_conn->cbs.cookie,
cbs               457 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	if (!p_ll2_conn->cbs.slowpath_cb) {
cbs               465 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_conn->cbs.slowpath_cb(p_ll2_conn->cbs.cookie,
cbs               513 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_conn->cbs.rx_comp_cb(p_ll2_conn->cbs.cookie, &data);
cbs               600 drivers/net/ethernet/qlogic/qed/qed_ll2.c 			p_ll2_conn->cbs.rx_release_cb(p_ll2_conn->cbs.cookie,
cbs              1303 drivers/net/ethernet/qlogic/qed/qed_ll2.c qed_ll2_set_cbs(struct qed_ll2_info *p_ll2_info, const struct qed_ll2_cbs *cbs)
cbs              1305 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	if (!cbs || (!cbs->rx_comp_cb ||
cbs              1306 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		     !cbs->rx_release_cb ||
cbs              1307 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		     !cbs->tx_comp_cb || !cbs->tx_release_cb || !cbs->cookie))
cbs              1310 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_info->cbs.rx_comp_cb = cbs->rx_comp_cb;
cbs              1311 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_info->cbs.rx_release_cb = cbs->rx_release_cb;
cbs              1312 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_info->cbs.tx_comp_cb = cbs->tx_comp_cb;
cbs              1313 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_info->cbs.tx_release_cb = cbs->tx_release_cb;
cbs              1314 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_info->cbs.slowpath_cb = cbs->slowpath_cb;
cbs              1315 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_info->cbs.cookie = cbs->cookie;
cbs              1392 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	rc = qed_ll2_set_cbs(p_ll2_info, data->cbs);
cbs              2215 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	cdev->ll2->cbs = ops;
cbs              2241 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	data->cbs = &ll2_cbs;
cbs              2426 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	WARN_ON(!cdev->ll2->cbs);
cbs               128 drivers/net/ethernet/qlogic/qed/qed_ll2.h 	struct qed_ll2_cbs cbs;
cbs               246 drivers/video/fbdev/nvidia/nv_hw.c 	int found, mclk_extra, mclk_loop, cbs, m1, p1;
cbs               263 drivers/video/fbdev/nvidia/nv_hw.c 	cbs = 128;
cbs               304 drivers/video/fbdev/nvidia/nv_hw.c 				    cbs * 1000 * 1000 / 16 / nvclk_freq;
cbs               307 drivers/video/fbdev/nvidia/nv_hw.c 				    cbs * 1000 * 1000 / (8 * width) /
cbs               341 drivers/video/fbdev/nvidia/nv_hw.c 		m1 = clwm + cbs - 512;
cbs               420 drivers/video/fbdev/nvidia/nv_hw.c 	int found, mclk_extra, mclk_loop, cbs, m1;
cbs               441 drivers/video/fbdev/nvidia/nv_hw.c 	cbs = 512;
cbs               559 drivers/video/fbdev/nvidia/nv_hw.c 					cbs = 512;
cbs               572 drivers/video/fbdev/nvidia/nv_hw.c 		m1 = clwm + cbs - 1024;	/* Amount of overfill */
cbs               583 drivers/video/fbdev/nvidia/nv_hw.c 				if (cbs <= 32) {
cbs               588 drivers/video/fbdev/nvidia/nv_hw.c 					cbs = cbs / 2;
cbs               605 drivers/video/fbdev/nvidia/nv_hw.c 		if (clwm < (1024 - cbs + 8))
cbs               606 drivers/video/fbdev/nvidia/nv_hw.c 			clwm = 1024 - cbs + 8;
cbs               611 drivers/video/fbdev/nvidia/nv_hw.c 		fifo->graphics_burst_size = cbs;
cbs               661 drivers/video/fbdev/riva/riva_hw.c     int found, mclk_extra, mclk_loop, cbs, m1, p1;
cbs               681 drivers/video/fbdev/riva/riva_hw.c     cbs = 128;
cbs               722 drivers/video/fbdev/riva/riva_hw.c                 video_fill_us = cbs*1000*1000 / 16 / nvclk_freq ;
cbs               724 drivers/video/fbdev/riva/riva_hw.c                 video_fill_us = cbs*1000*1000 / (8 * width) / mclk_freq;
cbs               755 drivers/video/fbdev/riva/riva_hw.c         m1 = clwm + cbs - 512;
cbs               848 drivers/video/fbdev/riva/riva_hw.c     int found, mclk_extra, mclk_loop, cbs, m1;
cbs               873 drivers/video/fbdev/riva/riva_hw.c     cbs = 512;
cbs              1003 drivers/video/fbdev/riva/riva_hw.c                   cbs = 512;
cbs              1004 drivers/video/fbdev/riva/riva_hw.c                   us_extra = (cbs * 1000 * 1000)/ (8*width)/mclk_freq ;
cbs              1019 drivers/video/fbdev/riva/riva_hw.c       m1 = clwm + cbs -  1024; /* Amount of overfill */
cbs              1031 drivers/video/fbdev/riva/riva_hw.c             if(cbs <= 32) {
cbs              1034 drivers/video/fbdev/riva/riva_hw.c               cbs = cbs/2;  /* reduce the burst size */
cbs              1051 drivers/video/fbdev/riva/riva_hw.c       if(clwm < (1024-cbs+8)) clwm = 1024-cbs+8;
cbs              1054 drivers/video/fbdev/riva/riva_hw.c       fifo->graphics_lwm = data;   fifo->graphics_burst_size = cbs;
cbs               172 drivers/virtio/virtio_input.c 	vq_callback_t *cbs[] = { virtinput_recv_events,
cbs               177 drivers/virtio/virtio_input.c 	err = virtio_find_vqs(vi->vdev, 2, vqs, cbs, names, NULL);
cbs              3236 fs/nfsd/nfs4state.c static __be32 nfsd4_check_cb_sec(struct nfsd4_cb_sec *cbs)
cbs              3238 fs/nfsd/nfs4state.c 	switch (cbs->flavor) {
cbs               525 fs/nfsd/nfs4xdr.c static __be32 nfsd4_decode_cb_sec(struct nfsd4_compoundargs *argp, struct nfsd4_cb_sec *cbs)
cbs               538 fs/nfsd/nfs4xdr.c 		cbs->flavor = (u32)(-1);
cbs               541 fs/nfsd/nfs4xdr.c 		cbs->flavor = 0;
cbs               548 fs/nfsd/nfs4xdr.c 			if (cbs->flavor == (u32)(-1))
cbs               549 fs/nfsd/nfs4xdr.c 				cbs->flavor = RPC_AUTH_NULL;
cbs               570 fs/nfsd/nfs4xdr.c 			if (cbs->flavor == (u32)(-1)) {
cbs               574 fs/nfsd/nfs4xdr.c 					cbs->uid = kuid;
cbs               575 fs/nfsd/nfs4xdr.c 					cbs->gid = kgid;
cbs               576 fs/nfsd/nfs4xdr.c 					cbs->flavor = RPC_AUTH_UNIX;
cbs               462 include/drm/drm_dp_mst_helper.h 	const struct drm_dp_mst_topology_cbs *cbs;
cbs               186 include/linux/qed/qed_ll2_if.h 	const struct qed_ll2_cbs *cbs;
cbs                61 include/uapi/linux/dcbnl.h 	__u8	cbs;
cbs                16 kernel/bpf/disasm.c static const char *__func_get_name(const struct bpf_insn_cbs *cbs,
cbs                27 kernel/bpf/disasm.c 	if (cbs && cbs->cb_call)
cbs                28 kernel/bpf/disasm.c 		return cbs->cb_call(cbs->private_data, insn);
cbs                36 kernel/bpf/disasm.c static const char *__func_imm_name(const struct bpf_insn_cbs *cbs,
cbs                40 kernel/bpf/disasm.c 	if (cbs && cbs->cb_imm)
cbs                41 kernel/bpf/disasm.c 		return cbs->cb_imm(cbs->private_data, insn, full_imm);
cbs               117 kernel/bpf/disasm.c void print_bpf_insn(const struct bpf_insn_cbs *cbs,
cbs               121 kernel/bpf/disasm.c 	const bpf_insn_print_t verbose = cbs->cb_print;
cbs               127 kernel/bpf/disasm.c 				verbose(cbs->private_data, "BUG_alu64_%02x\n", insn->code);
cbs               129 kernel/bpf/disasm.c 				print_bpf_end_insn(verbose, cbs->private_data, insn);
cbs               131 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) %c%d = -%c%d\n",
cbs               136 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) %c%d %s %c%d\n",
cbs               143 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) %c%d %s %d\n",
cbs               151 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) *(%s *)(r%d %+d) = r%d\n",
cbs               157 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) lock *(%s *)(r%d %+d) += r%d\n",
cbs               163 kernel/bpf/disasm.c 			verbose(cbs->private_data, "BUG_%02x\n", insn->code);
cbs               166 kernel/bpf/disasm.c 			verbose(cbs->private_data, "BUG_st_%02x\n", insn->code);
cbs               169 kernel/bpf/disasm.c 		verbose(cbs->private_data, "(%02x) *(%s *)(r%d %+d) = %d\n",
cbs               176 kernel/bpf/disasm.c 			verbose(cbs->private_data, "BUG_ldx_%02x\n", insn->code);
cbs               179 kernel/bpf/disasm.c 		verbose(cbs->private_data, "(%02x) r%d = *(%s *)(r%d %+d)\n",
cbs               185 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) r0 = *(%s *)skb[%d]\n",
cbs               190 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) r0 = *(%s *)skb[r%d + %d]\n",
cbs               207 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) r%d = %s\n",
cbs               209 kernel/bpf/disasm.c 				__func_imm_name(cbs, insn, imm,
cbs               212 kernel/bpf/disasm.c 			verbose(cbs->private_data, "BUG_ld_%02x\n", insn->code);
cbs               222 kernel/bpf/disasm.c 				verbose(cbs->private_data, "(%02x) call pc%s\n",
cbs               224 kernel/bpf/disasm.c 					__func_get_name(cbs, insn,
cbs               228 kernel/bpf/disasm.c 				verbose(cbs->private_data, "(%02x) call %s#%d\n", insn->code,
cbs               229 kernel/bpf/disasm.c 					__func_get_name(cbs, insn,
cbs               234 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) goto pc%+d\n",
cbs               237 kernel/bpf/disasm.c 			verbose(cbs->private_data, "(%02x) exit\n", insn->code);
cbs               239 kernel/bpf/disasm.c 			verbose(cbs->private_data,
cbs               247 kernel/bpf/disasm.c 			verbose(cbs->private_data,
cbs               255 kernel/bpf/disasm.c 		verbose(cbs->private_data, "(%02x) %s\n",
cbs                37 kernel/bpf/disasm.h void print_bpf_insn(const struct bpf_insn_cbs *cbs,
cbs              1445 kernel/bpf/verifier.c 	const struct bpf_insn_cbs cbs = {
cbs              1462 kernel/bpf/verifier.c 		print_bpf_insn(&cbs, insn, env->allow_ptr_leaks);
cbs              7636 kernel/bpf/verifier.c 			const struct bpf_insn_cbs cbs = {
cbs              7643 kernel/bpf/verifier.c 			print_bpf_insn(&cbs, insn, env->allow_ptr_leaks);
cbs              6038 kernel/events/core.c int perf_register_guest_info_callbacks(struct perf_guest_info_callbacks *cbs)
cbs              6040 kernel/events/core.c 	perf_guest_cbs = cbs;
cbs              6045 kernel/events/core.c int perf_unregister_guest_info_callbacks(struct perf_guest_info_callbacks *cbs)
cbs               515 kernel/rcu/srcutree.c 	bool cbs;
cbs               546 kernel/rcu/srcutree.c 		cbs = false;
cbs               549 kernel/rcu/srcutree.c 			cbs = snp->srcu_have_cbs[idx] == gpseq;
cbs               557 kernel/rcu/srcutree.c 		if (cbs)
cbs               662 kernel/rcu/tree_stall.h 	unsigned long cbs;
cbs               681 kernel/rcu/tree_stall.h 		cbs = rcu_get_n_cbs_cpu(cpu);
cbs               682 kernel/rcu/tree_stall.h 		if (!cbs)
cbs               686 kernel/rcu/tree_stall.h 		pr_cont(" %d: %lu", cpu, cbs);
cbs               687 kernel/rcu/tree_stall.h 		if (cbs <= max_cbs)
cbs               689 kernel/rcu/tree_stall.h 		max_cbs = cbs;
cbs               252 net/sched/sch_cbs.c 	struct tc_cbs_qopt_offload cbs = { };
cbs               266 net/sched/sch_cbs.c 	cbs.queue = q->queue;
cbs               267 net/sched/sch_cbs.c 	cbs.enable = 0;
cbs               269 net/sched/sch_cbs.c 	err = ops->ndo_setup_tc(dev, TC_SETUP_QDISC_CBS, &cbs);
cbs               272 net/sched/sch_cbs.c 			cbs.queue);
cbs               280 net/sched/sch_cbs.c 	struct tc_cbs_qopt_offload cbs = { };
cbs               288 net/sched/sch_cbs.c 	cbs.queue = q->queue;
cbs               290 net/sched/sch_cbs.c 	cbs.enable = 1;
cbs               291 net/sched/sch_cbs.c 	cbs.hicredit = opt->hicredit;
cbs               292 net/sched/sch_cbs.c 	cbs.locredit = opt->locredit;
cbs               293 net/sched/sch_cbs.c 	cbs.idleslope = opt->idleslope;
cbs               294 net/sched/sch_cbs.c 	cbs.sendslope = opt->sendslope;
cbs               296 net/sched/sch_cbs.c 	err = ops->ndo_setup_tc(dev, TC_SETUP_QDISC_CBS, &cbs);
cbs               209 tools/bpf/bpftool/xlated_dumper.c 	const struct bpf_insn_cbs cbs = {
cbs               258 tools/bpf/bpftool/xlated_dumper.c 		print_bpf_insn(&cbs, insn + i, true);
cbs               294 tools/bpf/bpftool/xlated_dumper.c 	const struct bpf_insn_cbs cbs = {
cbs               340 tools/bpf/bpftool/xlated_dumper.c 		print_bpf_insn(&cbs, insn + i, true);
cbs               357 tools/bpf/bpftool/xlated_dumper.c 	const struct bpf_insn_cbs cbs = {
cbs               369 tools/bpf/bpftool/xlated_dumper.c 		print_bpf_insn(&cbs, cur, true);