Searched refs:buf_id (Results 1 - 12 of 12) sorted by relevance

/linux-4.1.27/drivers/gpu/drm/exynos/
H A Dexynos_drm_gsc.c662 static int gsc_src_set_buf_seq(struct gsc_context *ctx, u32 buf_id, gsc_src_set_buf_seq() argument
668 u32 mask = 0x00000001 << buf_id; gsc_src_set_buf_seq()
670 DRM_DEBUG_KMS("buf_id[%d]buf_type[%d]\n", buf_id, buf_type); gsc_src_set_buf_seq()
689 cfg |= masked << buf_id; gsc_src_set_buf_seq()
698 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, gsc_src_set_addr()
713 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]buf_type[%d]\n", gsc_src_set_addr()
714 property->prop_id, buf_id, buf_type); gsc_src_set_addr()
716 if (buf_id > GSC_MAX_SRC) { gsc_src_set_addr()
717 dev_info(ippdrv->dev, "inavlid buf_id %d.\n", buf_id); gsc_src_set_addr()
725 GSC_IN_BASE_ADDR_Y(buf_id)); gsc_src_set_addr()
727 GSC_IN_BASE_ADDR_CB(buf_id)); gsc_src_set_addr()
729 GSC_IN_BASE_ADDR_CR(buf_id)); gsc_src_set_addr()
732 gsc_write(0x0, GSC_IN_BASE_ADDR_Y(buf_id)); gsc_src_set_addr()
733 gsc_write(0x0, GSC_IN_BASE_ADDR_CB(buf_id)); gsc_src_set_addr()
734 gsc_write(0x0, GSC_IN_BASE_ADDR_CR(buf_id)); gsc_src_set_addr()
741 return gsc_src_set_buf_seq(ctx, buf_id, buf_type); gsc_src_set_addr()
1094 static int gsc_dst_set_buf_seq(struct gsc_context *ctx, u32 buf_id, gsc_dst_set_buf_seq() argument
1100 u32 mask = 0x00000001 << buf_id; gsc_dst_set_buf_seq()
1103 DRM_DEBUG_KMS("buf_id[%d]buf_type[%d]\n", buf_id, buf_type); gsc_dst_set_buf_seq()
1125 cfg |= masked << buf_id; gsc_dst_set_buf_seq()
1146 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, gsc_dst_set_addr()
1161 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]buf_type[%d]\n", gsc_dst_set_addr()
1162 property->prop_id, buf_id, buf_type); gsc_dst_set_addr()
1164 if (buf_id > GSC_MAX_DST) { gsc_dst_set_addr()
1165 dev_info(ippdrv->dev, "inavlid buf_id %d.\n", buf_id); gsc_dst_set_addr()
1173 GSC_OUT_BASE_ADDR_Y(buf_id)); gsc_dst_set_addr()
1175 GSC_OUT_BASE_ADDR_CB(buf_id)); gsc_dst_set_addr()
1177 GSC_OUT_BASE_ADDR_CR(buf_id)); gsc_dst_set_addr()
1180 gsc_write(0x0, GSC_OUT_BASE_ADDR_Y(buf_id)); gsc_dst_set_addr()
1181 gsc_write(0x0, GSC_OUT_BASE_ADDR_CB(buf_id)); gsc_dst_set_addr()
1182 gsc_write(0x0, GSC_OUT_BASE_ADDR_CR(buf_id)); gsc_dst_set_addr()
1189 return gsc_dst_set_buf_seq(ctx, buf_id, buf_type); gsc_dst_set_addr()
1217 u32 buf_id = GSC_MAX_SRC; gsc_get_src_buf_index() local
1227 buf_id = i; gsc_get_src_buf_index()
1232 if (buf_id == GSC_MAX_SRC) { gsc_get_src_buf_index()
1237 ret = gsc_src_set_buf_seq(ctx, buf_id, IPP_BUF_DEQUEUE); gsc_get_src_buf_index()
1243 DRM_DEBUG_KMS("cfg[0x%x]curr_index[%d]buf_id[%d]\n", cfg, gsc_get_src_buf_index()
1244 curr_index, buf_id); gsc_get_src_buf_index()
1246 return buf_id; gsc_get_src_buf_index()
1252 u32 buf_id = GSC_MAX_DST; gsc_get_dst_buf_index() local
1262 buf_id = i; gsc_get_dst_buf_index()
1267 if (buf_id == GSC_MAX_DST) { gsc_get_dst_buf_index()
1272 ret = gsc_dst_set_buf_seq(ctx, buf_id, IPP_BUF_DEQUEUE); gsc_get_dst_buf_index()
1278 DRM_DEBUG_KMS("cfg[0x%x]curr_index[%d]buf_id[%d]\n", cfg, gsc_get_dst_buf_index()
1279 curr_index, buf_id); gsc_get_dst_buf_index()
1281 return buf_id; gsc_get_dst_buf_index()
1292 int buf_id[EXYNOS_DRM_OPS_MAX]; gsc_irq_handler() local
1307 buf_id[EXYNOS_DRM_OPS_SRC] = gsc_get_src_buf_index(ctx); gsc_irq_handler()
1308 if (buf_id[EXYNOS_DRM_OPS_SRC] < 0) gsc_irq_handler()
1311 buf_id[EXYNOS_DRM_OPS_DST] = gsc_get_dst_buf_index(ctx); gsc_irq_handler()
1312 if (buf_id[EXYNOS_DRM_OPS_DST] < 0) gsc_irq_handler()
1316 buf_id[EXYNOS_DRM_OPS_SRC], buf_id[EXYNOS_DRM_OPS_DST]); gsc_irq_handler()
1319 event_work->buf_id[EXYNOS_DRM_OPS_SRC] = gsc_irq_handler()
1320 buf_id[EXYNOS_DRM_OPS_SRC]; gsc_irq_handler()
1321 event_work->buf_id[EXYNOS_DRM_OPS_DST] = gsc_irq_handler()
1322 buf_id[EXYNOS_DRM_OPS_DST]; gsc_irq_handler()
697 gsc_src_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) gsc_src_set_addr() argument
1145 gsc_dst_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) gsc_dst_set_addr() argument
H A Dexynos_drm_fimc.c368 int frame_cnt, buf_id; fimc_get_buf_id() local
385 buf_id = frame_cnt - 1; fimc_get_buf_id()
386 DRM_DEBUG_KMS("buf_id[%d]\n", buf_id); fimc_get_buf_id()
388 return buf_id; fimc_get_buf_id()
684 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, fimc_src_set_addr()
700 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]buf_type[%d]\n", fimc_src_set_addr()
701 property->prop_id, buf_id, buf_type); fimc_src_set_addr()
703 if (buf_id > FIMC_MAX_SRC) { fimc_src_set_addr()
704 dev_info(ippdrv->dev, "inavlid buf_id %d.\n", buf_id); fimc_src_set_addr()
1112 static void fimc_dst_set_buf_seq(struct fimc_context *ctx, u32 buf_id, fimc_dst_set_buf_seq() argument
1119 DRM_DEBUG_KMS("buf_id[%d]buf_type[%d]\n", buf_id, buf_type); fimc_dst_set_buf_seq()
1126 cfg |= (1 << buf_id); fimc_dst_set_buf_seq()
1128 cfg &= ~(1 << buf_id); fimc_dst_set_buf_seq()
1143 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, fimc_dst_set_addr()
1159 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]buf_type[%d]\n", fimc_dst_set_addr()
1160 property->prop_id, buf_id, buf_type); fimc_dst_set_addr()
1162 if (buf_id > FIMC_MAX_DST) { fimc_dst_set_addr()
1163 dev_info(ippdrv->dev, "inavlid buf_id %d.\n", buf_id); fimc_dst_set_addr()
1173 EXYNOS_CIOYSA(buf_id)); fimc_dst_set_addr()
1177 EXYNOS_CIOCBSA(buf_id)); fimc_dst_set_addr()
1179 EXYNOS_CIOCRSA(buf_id)); fimc_dst_set_addr()
1182 EXYNOS_CIOCBSA(buf_id)); fimc_dst_set_addr()
1184 EXYNOS_CIOCRSA(buf_id)); fimc_dst_set_addr()
1188 fimc_write(ctx, 0x0, EXYNOS_CIOYSA(buf_id)); fimc_dst_set_addr()
1189 fimc_write(ctx, 0x0, EXYNOS_CIOCBSA(buf_id)); fimc_dst_set_addr()
1190 fimc_write(ctx, 0x0, EXYNOS_CIOCRSA(buf_id)); fimc_dst_set_addr()
1197 fimc_dst_set_buf_seq(ctx, buf_id, buf_type); fimc_dst_set_addr()
1233 int buf_id; fimc_irq_handler() local
1244 buf_id = fimc_get_buf_id(ctx); fimc_irq_handler()
1245 if (buf_id < 0) fimc_irq_handler()
1248 DRM_DEBUG_KMS("buf_id[%d]\n", buf_id); fimc_irq_handler()
1250 fimc_dst_set_buf_seq(ctx, buf_id, IPP_BUF_DEQUEUE); fimc_irq_handler()
1253 event_work->buf_id[EXYNOS_DRM_OPS_DST] = buf_id; fimc_irq_handler()
683 fimc_src_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) fimc_src_set_addr() argument
1142 fimc_dst_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) fimc_dst_set_addr() argument
H A Dexynos_drm_ipp.c68 * @buf_id: id of buffer.
76 u32 buf_id; member in struct:drm_exynos_ipp_mem_node
565 m_node->buf_id = qbuf->buf_id; ipp_get_mem_node()
569 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]\n", qbuf->prop_id, m_node->buf_id); ipp_get_mem_node()
636 DRM_DEBUG_KMS("ops_id[%d]buf_id[%d]\n", qbuf->ops_id, qbuf->buf_id); ipp_get_event()
651 e->event.buf_id[EXYNOS_DRM_OPS_DST] = qbuf->buf_id; ipp_get_event()
684 if (qbuf && (qbuf->buf_id == ipp_put_event()
685 e->event.buf_id[EXYNOS_DRM_OPS_DST])) { ipp_put_event()
754 DRM_DEBUG_KMS("buf_id[%d]\n", qbuf->buf_id); ipp_find_mem_node()
764 if (m_node->buf_id == qbuf->buf_id) list_for_each_entry()
797 m_node->buf_id, IPP_BUF_ENQUEUE); ipp_set_mem_node()
887 if (m_node->buf_id == qbuf->buf_id && ipp_clean_queue_buf()
915 DRM_DEBUG_KMS("prop_id[%d]ops_id[%s]buf_id[%d]buf_type[%d]\n", exynos_drm_ipp_queue_buf()
917 qbuf->buf_id, qbuf->buf_type); exynos_drm_ipp_queue_buf()
1409 struct drm_exynos_ipp_cmd_node *c_node, int *buf_id) ipp_send_event()
1423 DRM_DEBUG_KMS("%s buf_id[%d]\n", i ? "dst" : "src", buf_id[i]); ipp_send_event()
1459 tbuf_id[i] = m_node->buf_id; for_each_ipp_ops()
1460 DRM_DEBUG_KMS("%s buf_id[%d]\n", for_each_ipp_ops()
1472 qbuf.buf_id = buf_id[EXYNOS_DRM_OPS_DST];
1482 tbuf_id[EXYNOS_DRM_OPS_DST] = m_node->buf_id;
1495 tbuf_id[EXYNOS_DRM_OPS_SRC] = m_node->buf_id;
1508 if (tbuf_id[EXYNOS_DRM_OPS_DST] != buf_id[EXYNOS_DRM_OPS_DST])
1509 DRM_ERROR("failed to match buf_id[%d %d]prop_id[%d]\n",
1510 tbuf_id[1], buf_id[1], property->prop_id);
1529 e->event.buf_id[i] = tbuf_id[i];
1537 DRM_DEBUG_KMS("done cmd[%d]prop_id[%d]buf_id[%d]\n",
1562 DRM_DEBUG_KMS("buf_id[%d]\n", event_work->buf_id[EXYNOS_DRM_OPS_DST]); ipp_sched_event()
1588 ret = ipp_send_event(ippdrv, c_node, event_work->buf_id); ipp_sched_event()
1408 ipp_send_event(struct exynos_drm_ippdrv *ippdrv, struct drm_exynos_ipp_cmd_node *c_node, int *buf_id) ipp_send_event() argument
H A Dexynos_drm_ipp.h110 * @buf_id: id of src, dst buffer.
115 u32 buf_id[EXYNOS_DRM_OPS_MAX]; member in struct:drm_exynos_ipp_event_work
134 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id,
H A Dexynos_drm_rotator.c157 event_work->buf_id[EXYNOS_DRM_OPS_DST] = rotator_irq_handler()
271 u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) rotator_src_set_addr()
278 /* Set current buf_id */ rotator_src_set_addr()
279 rot->cur_buf_id[EXYNOS_DRM_OPS_SRC] = buf_id; rotator_src_set_addr()
406 u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) rotator_dst_set_addr()
413 /* Set current buf_id */ rotator_dst_set_addr()
414 rot->cur_buf_id[EXYNOS_DRM_OPS_DST] = buf_id; rotator_dst_set_addr()
269 rotator_src_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) rotator_src_set_addr() argument
404 rotator_dst_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) rotator_dst_set_addr() argument
/linux-4.1.27/drivers/net/wireless/cw1200/
H A Dhwio.c25 #define SDIO_ADDR17BIT(buf_id, mpf, rfu, reg_id_ofs) \
26 ((((buf_id) & 0x1F) << 7) \
34 void *buf, size_t buf_len, int buf_id) __cw1200_reg_read()
47 sdio_reg_addr_17bit = SDIO_ADDR17BIT(buf_id, 0, 0, addr_sdio); __cw1200_reg_read()
55 const void *buf, size_t buf_len, int buf_id) __cw1200_reg_write()
62 sdio_reg_addr_17bit = SDIO_ADDR17BIT(buf_id, 0, 0, addr_sdio); __cw1200_reg_write()
33 __cw1200_reg_read(struct cw1200_common *priv, u16 addr, void *buf, size_t buf_len, int buf_id) __cw1200_reg_read() argument
54 __cw1200_reg_write(struct cw1200_common *priv, u16 addr, const void *buf, size_t buf_len, int buf_id) __cw1200_reg_write() argument
/linux-4.1.27/drivers/crypto/caam/
H A Dcaamrng.c217 static inline int rng_create_job_desc(struct caam_rng_ctx *ctx, int buf_id) rng_create_job_desc() argument
220 struct buf_data *bd = &ctx->bufs[buf_id]; rng_create_job_desc()
255 static int caam_init_buf(struct caam_rng_ctx *ctx, int buf_id) caam_init_buf() argument
257 struct buf_data *bd = &ctx->bufs[buf_id]; caam_init_buf()
260 err = rng_create_job_desc(ctx, buf_id); caam_init_buf()
265 submit_job(ctx, buf_id == ctx->current_buf); caam_init_buf()
/linux-4.1.27/drivers/net/can/mscan/
H A Dmscan.c202 int i, rtr, buf_id; mscan_start_xmit() local
211 buf_id = ffs(i) - 1; mscan_start_xmit()
219 * if buf_id < 3, then current frame will be send out of order, mscan_start_xmit()
224 if (buf_id < priv->prev_buf_id) { mscan_start_xmit()
234 priv->prev_buf_id = buf_id; mscan_start_xmit()
276 out_8(&regs->cantflg, 1 << buf_id); mscan_start_xmit()
281 list_add_tail(&priv->tx_queue[buf_id].list, &priv->tx_head); mscan_start_xmit()
283 can_put_echo_skb(skb, dev, buf_id); mscan_start_xmit()
286 priv->tx_active |= 1 << buf_id; mscan_start_xmit()
/linux-4.1.27/include/uapi/drm/
H A Dexynos_drm.h253 * @buf_id: id of buffer.
261 __u32 buf_id; member in struct:drm_exynos_ipp_queue_buf
347 __u32 buf_id[EXYNOS_DRM_OPS_MAX]; member in struct:drm_exynos_ipp_event
/linux-4.1.27/drivers/edac/
H A Di5400_edac.c519 int buf_id; i5400_proccess_non_recoverable_info() local
546 buf_id = nrec_buf_id(info); i5400_proccess_non_recoverable_info()
553 buf_id, rdwr_str(rdwr), ras, cas); i5400_proccess_non_recoverable_info()
561 bank, buf_id, ras, cas, allErrors, error_name[errnum]); i5400_proccess_non_recoverable_info()
/linux-4.1.27/drivers/media/platform/exynos4-is/
H A Dfimc-capture.c429 int buf_id = (vid_cap->reqbufs_count == 1) ? -1 : buffer_queue() local
432 fimc_hw_set_output_addr(fimc, &buf->paddr, buf_id); buffer_queue()
/linux-4.1.27/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_execbuf.c1420 * @buf_id: Pointer to the user-space backup buffer handle in the command
1433 uint32_t *buf_id, vmw_cmd_switch_backup()
1445 ret = vmw_translate_mob_ptr(dev_priv, sw_context, buf_id, &dma_buf); vmw_cmd_switch_backup()
1427 vmw_cmd_switch_backup(struct vmw_private *dev_priv, struct vmw_sw_context *sw_context, enum vmw_res_type res_type, const struct vmw_user_resource_conv *converter, uint32_t *res_id, uint32_t *buf_id, unsigned long backup_offset) vmw_cmd_switch_backup() argument

Completed in 624 milliseconds