/linux-4.1.27/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_gsc.c | 662 static int gsc_src_set_buf_seq(struct gsc_context *ctx, u32 buf_id, gsc_src_set_buf_seq() argument 668 u32 mask = 0x00000001 << buf_id; gsc_src_set_buf_seq() 670 DRM_DEBUG_KMS("buf_id[%d]buf_type[%d]\n", buf_id, buf_type); gsc_src_set_buf_seq() 689 cfg |= masked << buf_id; gsc_src_set_buf_seq() 698 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, gsc_src_set_addr() 713 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]buf_type[%d]\n", gsc_src_set_addr() 714 property->prop_id, buf_id, buf_type); gsc_src_set_addr() 716 if (buf_id > GSC_MAX_SRC) { gsc_src_set_addr() 717 dev_info(ippdrv->dev, "inavlid buf_id %d.\n", buf_id); gsc_src_set_addr() 725 GSC_IN_BASE_ADDR_Y(buf_id)); gsc_src_set_addr() 727 GSC_IN_BASE_ADDR_CB(buf_id)); gsc_src_set_addr() 729 GSC_IN_BASE_ADDR_CR(buf_id)); gsc_src_set_addr() 732 gsc_write(0x0, GSC_IN_BASE_ADDR_Y(buf_id)); gsc_src_set_addr() 733 gsc_write(0x0, GSC_IN_BASE_ADDR_CB(buf_id)); gsc_src_set_addr() 734 gsc_write(0x0, GSC_IN_BASE_ADDR_CR(buf_id)); gsc_src_set_addr() 741 return gsc_src_set_buf_seq(ctx, buf_id, buf_type); gsc_src_set_addr() 1094 static int gsc_dst_set_buf_seq(struct gsc_context *ctx, u32 buf_id, gsc_dst_set_buf_seq() argument 1100 u32 mask = 0x00000001 << buf_id; gsc_dst_set_buf_seq() 1103 DRM_DEBUG_KMS("buf_id[%d]buf_type[%d]\n", buf_id, buf_type); gsc_dst_set_buf_seq() 1125 cfg |= masked << buf_id; gsc_dst_set_buf_seq() 1146 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, gsc_dst_set_addr() 1161 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]buf_type[%d]\n", gsc_dst_set_addr() 1162 property->prop_id, buf_id, buf_type); gsc_dst_set_addr() 1164 if (buf_id > GSC_MAX_DST) { gsc_dst_set_addr() 1165 dev_info(ippdrv->dev, "inavlid buf_id %d.\n", buf_id); gsc_dst_set_addr() 1173 GSC_OUT_BASE_ADDR_Y(buf_id)); gsc_dst_set_addr() 1175 GSC_OUT_BASE_ADDR_CB(buf_id)); gsc_dst_set_addr() 1177 GSC_OUT_BASE_ADDR_CR(buf_id)); gsc_dst_set_addr() 1180 gsc_write(0x0, GSC_OUT_BASE_ADDR_Y(buf_id)); gsc_dst_set_addr() 1181 gsc_write(0x0, GSC_OUT_BASE_ADDR_CB(buf_id)); gsc_dst_set_addr() 1182 gsc_write(0x0, GSC_OUT_BASE_ADDR_CR(buf_id)); gsc_dst_set_addr() 1189 return gsc_dst_set_buf_seq(ctx, buf_id, buf_type); gsc_dst_set_addr() 1217 u32 buf_id = GSC_MAX_SRC; gsc_get_src_buf_index() local 1227 buf_id = i; gsc_get_src_buf_index() 1232 if (buf_id == GSC_MAX_SRC) { gsc_get_src_buf_index() 1237 ret = gsc_src_set_buf_seq(ctx, buf_id, IPP_BUF_DEQUEUE); gsc_get_src_buf_index() 1243 DRM_DEBUG_KMS("cfg[0x%x]curr_index[%d]buf_id[%d]\n", cfg, gsc_get_src_buf_index() 1244 curr_index, buf_id); gsc_get_src_buf_index() 1246 return buf_id; gsc_get_src_buf_index() 1252 u32 buf_id = GSC_MAX_DST; gsc_get_dst_buf_index() local 1262 buf_id = i; gsc_get_dst_buf_index() 1267 if (buf_id == GSC_MAX_DST) { gsc_get_dst_buf_index() 1272 ret = gsc_dst_set_buf_seq(ctx, buf_id, IPP_BUF_DEQUEUE); gsc_get_dst_buf_index() 1278 DRM_DEBUG_KMS("cfg[0x%x]curr_index[%d]buf_id[%d]\n", cfg, gsc_get_dst_buf_index() 1279 curr_index, buf_id); gsc_get_dst_buf_index() 1281 return buf_id; gsc_get_dst_buf_index() 1292 int buf_id[EXYNOS_DRM_OPS_MAX]; gsc_irq_handler() local 1307 buf_id[EXYNOS_DRM_OPS_SRC] = gsc_get_src_buf_index(ctx); gsc_irq_handler() 1308 if (buf_id[EXYNOS_DRM_OPS_SRC] < 0) gsc_irq_handler() 1311 buf_id[EXYNOS_DRM_OPS_DST] = gsc_get_dst_buf_index(ctx); gsc_irq_handler() 1312 if (buf_id[EXYNOS_DRM_OPS_DST] < 0) gsc_irq_handler() 1316 buf_id[EXYNOS_DRM_OPS_SRC], buf_id[EXYNOS_DRM_OPS_DST]); gsc_irq_handler() 1319 event_work->buf_id[EXYNOS_DRM_OPS_SRC] = gsc_irq_handler() 1320 buf_id[EXYNOS_DRM_OPS_SRC]; gsc_irq_handler() 1321 event_work->buf_id[EXYNOS_DRM_OPS_DST] = gsc_irq_handler() 1322 buf_id[EXYNOS_DRM_OPS_DST]; gsc_irq_handler() 697 gsc_src_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) gsc_src_set_addr() argument 1145 gsc_dst_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) gsc_dst_set_addr() argument
|
H A D | exynos_drm_fimc.c | 368 int frame_cnt, buf_id; fimc_get_buf_id() local 385 buf_id = frame_cnt - 1; fimc_get_buf_id() 386 DRM_DEBUG_KMS("buf_id[%d]\n", buf_id); fimc_get_buf_id() 388 return buf_id; fimc_get_buf_id() 684 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, fimc_src_set_addr() 700 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]buf_type[%d]\n", fimc_src_set_addr() 701 property->prop_id, buf_id, buf_type); fimc_src_set_addr() 703 if (buf_id > FIMC_MAX_SRC) { fimc_src_set_addr() 704 dev_info(ippdrv->dev, "inavlid buf_id %d.\n", buf_id); fimc_src_set_addr() 1112 static void fimc_dst_set_buf_seq(struct fimc_context *ctx, u32 buf_id, fimc_dst_set_buf_seq() argument 1119 DRM_DEBUG_KMS("buf_id[%d]buf_type[%d]\n", buf_id, buf_type); fimc_dst_set_buf_seq() 1126 cfg |= (1 << buf_id); fimc_dst_set_buf_seq() 1128 cfg &= ~(1 << buf_id); fimc_dst_set_buf_seq() 1143 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, fimc_dst_set_addr() 1159 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]buf_type[%d]\n", fimc_dst_set_addr() 1160 property->prop_id, buf_id, buf_type); fimc_dst_set_addr() 1162 if (buf_id > FIMC_MAX_DST) { fimc_dst_set_addr() 1163 dev_info(ippdrv->dev, "inavlid buf_id %d.\n", buf_id); fimc_dst_set_addr() 1173 EXYNOS_CIOYSA(buf_id)); fimc_dst_set_addr() 1177 EXYNOS_CIOCBSA(buf_id)); fimc_dst_set_addr() 1179 EXYNOS_CIOCRSA(buf_id)); fimc_dst_set_addr() 1182 EXYNOS_CIOCBSA(buf_id)); fimc_dst_set_addr() 1184 EXYNOS_CIOCRSA(buf_id)); fimc_dst_set_addr() 1188 fimc_write(ctx, 0x0, EXYNOS_CIOYSA(buf_id)); fimc_dst_set_addr() 1189 fimc_write(ctx, 0x0, EXYNOS_CIOCBSA(buf_id)); fimc_dst_set_addr() 1190 fimc_write(ctx, 0x0, EXYNOS_CIOCRSA(buf_id)); fimc_dst_set_addr() 1197 fimc_dst_set_buf_seq(ctx, buf_id, buf_type); fimc_dst_set_addr() 1233 int buf_id; fimc_irq_handler() local 1244 buf_id = fimc_get_buf_id(ctx); fimc_irq_handler() 1245 if (buf_id < 0) fimc_irq_handler() 1248 DRM_DEBUG_KMS("buf_id[%d]\n", buf_id); fimc_irq_handler() 1250 fimc_dst_set_buf_seq(ctx, buf_id, IPP_BUF_DEQUEUE); fimc_irq_handler() 1253 event_work->buf_id[EXYNOS_DRM_OPS_DST] = buf_id; fimc_irq_handler() 683 fimc_src_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) fimc_src_set_addr() argument 1142 fimc_dst_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) fimc_dst_set_addr() argument
|
H A D | exynos_drm_ipp.c | 68 * @buf_id: id of buffer. 76 u32 buf_id; member in struct:drm_exynos_ipp_mem_node 565 m_node->buf_id = qbuf->buf_id; ipp_get_mem_node() 569 DRM_DEBUG_KMS("prop_id[%d]buf_id[%d]\n", qbuf->prop_id, m_node->buf_id); ipp_get_mem_node() 636 DRM_DEBUG_KMS("ops_id[%d]buf_id[%d]\n", qbuf->ops_id, qbuf->buf_id); ipp_get_event() 651 e->event.buf_id[EXYNOS_DRM_OPS_DST] = qbuf->buf_id; ipp_get_event() 684 if (qbuf && (qbuf->buf_id == ipp_put_event() 685 e->event.buf_id[EXYNOS_DRM_OPS_DST])) { ipp_put_event() 754 DRM_DEBUG_KMS("buf_id[%d]\n", qbuf->buf_id); ipp_find_mem_node() 764 if (m_node->buf_id == qbuf->buf_id) list_for_each_entry() 797 m_node->buf_id, IPP_BUF_ENQUEUE); ipp_set_mem_node() 887 if (m_node->buf_id == qbuf->buf_id && ipp_clean_queue_buf() 915 DRM_DEBUG_KMS("prop_id[%d]ops_id[%s]buf_id[%d]buf_type[%d]\n", exynos_drm_ipp_queue_buf() 917 qbuf->buf_id, qbuf->buf_type); exynos_drm_ipp_queue_buf() 1409 struct drm_exynos_ipp_cmd_node *c_node, int *buf_id) ipp_send_event() 1423 DRM_DEBUG_KMS("%s buf_id[%d]\n", i ? "dst" : "src", buf_id[i]); ipp_send_event() 1459 tbuf_id[i] = m_node->buf_id; for_each_ipp_ops() 1460 DRM_DEBUG_KMS("%s buf_id[%d]\n", for_each_ipp_ops() 1472 qbuf.buf_id = buf_id[EXYNOS_DRM_OPS_DST]; 1482 tbuf_id[EXYNOS_DRM_OPS_DST] = m_node->buf_id; 1495 tbuf_id[EXYNOS_DRM_OPS_SRC] = m_node->buf_id; 1508 if (tbuf_id[EXYNOS_DRM_OPS_DST] != buf_id[EXYNOS_DRM_OPS_DST]) 1509 DRM_ERROR("failed to match buf_id[%d %d]prop_id[%d]\n", 1510 tbuf_id[1], buf_id[1], property->prop_id); 1529 e->event.buf_id[i] = tbuf_id[i]; 1537 DRM_DEBUG_KMS("done cmd[%d]prop_id[%d]buf_id[%d]\n", 1562 DRM_DEBUG_KMS("buf_id[%d]\n", event_work->buf_id[EXYNOS_DRM_OPS_DST]); ipp_sched_event() 1588 ret = ipp_send_event(ippdrv, c_node, event_work->buf_id); ipp_sched_event() 1408 ipp_send_event(struct exynos_drm_ippdrv *ippdrv, struct drm_exynos_ipp_cmd_node *c_node, int *buf_id) ipp_send_event() argument
|
H A D | exynos_drm_ipp.h | 110 * @buf_id: id of src, dst buffer. 115 u32 buf_id[EXYNOS_DRM_OPS_MAX]; member in struct:drm_exynos_ipp_event_work 134 struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id,
|
H A D | exynos_drm_rotator.c | 157 event_work->buf_id[EXYNOS_DRM_OPS_DST] = rotator_irq_handler() 271 u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) rotator_src_set_addr() 278 /* Set current buf_id */ rotator_src_set_addr() 279 rot->cur_buf_id[EXYNOS_DRM_OPS_SRC] = buf_id; rotator_src_set_addr() 406 u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) rotator_dst_set_addr() 413 /* Set current buf_id */ rotator_dst_set_addr() 414 rot->cur_buf_id[EXYNOS_DRM_OPS_DST] = buf_id; rotator_dst_set_addr() 269 rotator_src_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) rotator_src_set_addr() argument 404 rotator_dst_set_addr(struct device *dev, struct drm_exynos_ipp_buf_info *buf_info, u32 buf_id, enum drm_exynos_ipp_buf_type buf_type) rotator_dst_set_addr() argument
|
/linux-4.1.27/drivers/net/wireless/cw1200/ |
H A D | hwio.c | 25 #define SDIO_ADDR17BIT(buf_id, mpf, rfu, reg_id_ofs) \ 26 ((((buf_id) & 0x1F) << 7) \ 34 void *buf, size_t buf_len, int buf_id) __cw1200_reg_read() 47 sdio_reg_addr_17bit = SDIO_ADDR17BIT(buf_id, 0, 0, addr_sdio); __cw1200_reg_read() 55 const void *buf, size_t buf_len, int buf_id) __cw1200_reg_write() 62 sdio_reg_addr_17bit = SDIO_ADDR17BIT(buf_id, 0, 0, addr_sdio); __cw1200_reg_write() 33 __cw1200_reg_read(struct cw1200_common *priv, u16 addr, void *buf, size_t buf_len, int buf_id) __cw1200_reg_read() argument 54 __cw1200_reg_write(struct cw1200_common *priv, u16 addr, const void *buf, size_t buf_len, int buf_id) __cw1200_reg_write() argument
|
/linux-4.1.27/drivers/crypto/caam/ |
H A D | caamrng.c | 217 static inline int rng_create_job_desc(struct caam_rng_ctx *ctx, int buf_id) rng_create_job_desc() argument 220 struct buf_data *bd = &ctx->bufs[buf_id]; rng_create_job_desc() 255 static int caam_init_buf(struct caam_rng_ctx *ctx, int buf_id) caam_init_buf() argument 257 struct buf_data *bd = &ctx->bufs[buf_id]; caam_init_buf() 260 err = rng_create_job_desc(ctx, buf_id); caam_init_buf() 265 submit_job(ctx, buf_id == ctx->current_buf); caam_init_buf()
|
/linux-4.1.27/drivers/net/can/mscan/ |
H A D | mscan.c | 202 int i, rtr, buf_id; mscan_start_xmit() local 211 buf_id = ffs(i) - 1; mscan_start_xmit() 219 * if buf_id < 3, then current frame will be send out of order, mscan_start_xmit() 224 if (buf_id < priv->prev_buf_id) { mscan_start_xmit() 234 priv->prev_buf_id = buf_id; mscan_start_xmit() 276 out_8(®s->cantflg, 1 << buf_id); mscan_start_xmit() 281 list_add_tail(&priv->tx_queue[buf_id].list, &priv->tx_head); mscan_start_xmit() 283 can_put_echo_skb(skb, dev, buf_id); mscan_start_xmit() 286 priv->tx_active |= 1 << buf_id; mscan_start_xmit()
|
/linux-4.1.27/include/uapi/drm/ |
H A D | exynos_drm.h | 253 * @buf_id: id of buffer. 261 __u32 buf_id; member in struct:drm_exynos_ipp_queue_buf 347 __u32 buf_id[EXYNOS_DRM_OPS_MAX]; member in struct:drm_exynos_ipp_event
|
/linux-4.1.27/drivers/edac/ |
H A D | i5400_edac.c | 519 int buf_id; i5400_proccess_non_recoverable_info() local 546 buf_id = nrec_buf_id(info); i5400_proccess_non_recoverable_info() 553 buf_id, rdwr_str(rdwr), ras, cas); i5400_proccess_non_recoverable_info() 561 bank, buf_id, ras, cas, allErrors, error_name[errnum]); i5400_proccess_non_recoverable_info()
|
/linux-4.1.27/drivers/media/platform/exynos4-is/ |
H A D | fimc-capture.c | 429 int buf_id = (vid_cap->reqbufs_count == 1) ? -1 : buffer_queue() local 432 fimc_hw_set_output_addr(fimc, &buf->paddr, buf_id); buffer_queue()
|
/linux-4.1.27/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_execbuf.c | 1420 * @buf_id: Pointer to the user-space backup buffer handle in the command 1433 uint32_t *buf_id, vmw_cmd_switch_backup() 1445 ret = vmw_translate_mob_ptr(dev_priv, sw_context, buf_id, &dma_buf); vmw_cmd_switch_backup() 1427 vmw_cmd_switch_backup(struct vmw_private *dev_priv, struct vmw_sw_context *sw_context, enum vmw_res_type res_type, const struct vmw_user_resource_conv *converter, uint32_t *res_id, uint32_t *buf_id, unsigned long backup_offset) vmw_cmd_switch_backup() argument
|