wpc               703 drivers/net/ieee802154/ca8210.c 	struct work_priv_container *wpc = container_of(
wpc               708 drivers/net/ieee802154/ca8210.c 	struct ca8210_priv *priv = wpc->priv;
wpc               711 drivers/net/ieee802154/ca8210.c 	kfree(wpc);
wpc                70 drivers/staging/most/dim2/hal.c 	u16 wpc;
wpc               390 drivers/staging/most/dim2/hal.c 	g.atx_dbr.wpc = g.atx_dbr.rpc;
wpc               396 drivers/staging/most/dim2/hal.c 	g.atx_dbr.sz_queue[norm_pc(g.atx_dbr.wpc)] = buf_sz;
wpc               397 drivers/staging/most/dim2/hal.c 	g.atx_dbr.wpc++;
wpc               415 drivers/staging/most/dim2/hal.c 	if ((u16)(dbr->wpc - dbr->rpc) >= CDT0_RPC_MASK)
wpc               424 fs/xfs/xfs_aops.c 	struct xfs_writepage_ctx	*wpc,
wpc               428 fs/xfs/xfs_aops.c 	if (offset_fsb < wpc->imap.br_startoff ||
wpc               429 fs/xfs/xfs_aops.c 	    offset_fsb >= wpc->imap.br_startoff + wpc->imap.br_blockcount)
wpc               436 fs/xfs/xfs_aops.c 	if (wpc->fork == XFS_COW_FORK)
wpc               446 fs/xfs/xfs_aops.c 	if (wpc->data_seq != READ_ONCE(ip->i_df.if_seq))
wpc               449 fs/xfs/xfs_aops.c 	    wpc->cow_seq != READ_ONCE(ip->i_cowfp->if_seq))
wpc               464 fs/xfs/xfs_aops.c 	struct xfs_writepage_ctx *wpc,
wpc               477 fs/xfs/xfs_aops.c 		error = xfs_bmapi_convert_delalloc(ip, wpc->fork, offset_fsb,
wpc               478 fs/xfs/xfs_aops.c 				&wpc->imap, wpc->fork == XFS_COW_FORK ?
wpc               479 fs/xfs/xfs_aops.c 					&wpc->cow_seq : &wpc->data_seq);
wpc               482 fs/xfs/xfs_aops.c 	} while (wpc->imap.br_startoff + wpc->imap.br_blockcount <= offset_fsb);
wpc               489 fs/xfs/xfs_aops.c 	struct xfs_writepage_ctx *wpc,
wpc               522 fs/xfs/xfs_aops.c 	if (xfs_imap_valid(wpc, ip, offset_fsb))
wpc               544 fs/xfs/xfs_aops.c 		wpc->cow_seq = READ_ONCE(ip->i_cowfp->if_seq);
wpc               547 fs/xfs/xfs_aops.c 		wpc->fork = XFS_COW_FORK;
wpc               555 fs/xfs/xfs_aops.c 	if (xfs_imap_valid(wpc, ip, offset_fsb)) {
wpc               567 fs/xfs/xfs_aops.c 	wpc->data_seq = READ_ONCE(ip->i_df.if_seq);
wpc               570 fs/xfs/xfs_aops.c 	wpc->fork = XFS_DATA_FORK;
wpc               595 fs/xfs/xfs_aops.c 	wpc->imap = imap;
wpc               596 fs/xfs/xfs_aops.c 	trace_xfs_map_blocks_found(ip, offset, count, wpc->fork, &imap);
wpc               599 fs/xfs/xfs_aops.c 	error = xfs_convert_blocks(wpc, ip, offset_fsb);
wpc               608 fs/xfs/xfs_aops.c 		if (error == -EAGAIN && wpc->fork == XFS_COW_FORK && !retries++)
wpc               619 fs/xfs/xfs_aops.c 	if (wpc->fork != XFS_COW_FORK && cow_fsb != NULLFILEOFF &&
wpc               620 fs/xfs/xfs_aops.c 	    cow_fsb < wpc->imap.br_startoff + wpc->imap.br_blockcount)
wpc               621 fs/xfs/xfs_aops.c 		wpc->imap.br_blockcount = cow_fsb - wpc->imap.br_startoff;
wpc               623 fs/xfs/xfs_aops.c 	ASSERT(wpc->imap.br_startoff <= offset_fsb);
wpc               624 fs/xfs/xfs_aops.c 	ASSERT(wpc->imap.br_startoff + wpc->imap.br_blockcount > offset_fsb);
wpc               625 fs/xfs/xfs_aops.c 	trace_xfs_map_blocks_alloc(ip, offset, count, wpc->fork, &imap);
wpc               760 fs/xfs/xfs_aops.c 	struct xfs_writepage_ctx *wpc,
wpc               772 fs/xfs/xfs_aops.c 	sector = xfs_fsb_to_db(ip, wpc->imap.br_startblock) +
wpc               773 fs/xfs/xfs_aops.c 		((offset - XFS_FSB_TO_B(mp, wpc->imap.br_startoff)) >> 9);
wpc               775 fs/xfs/xfs_aops.c 	if (!wpc->ioend ||
wpc               776 fs/xfs/xfs_aops.c 	    wpc->fork != wpc->ioend->io_fork ||
wpc               777 fs/xfs/xfs_aops.c 	    wpc->imap.br_state != wpc->ioend->io_state ||
wpc               778 fs/xfs/xfs_aops.c 	    sector != bio_end_sector(wpc->ioend->io_bio) ||
wpc               779 fs/xfs/xfs_aops.c 	    offset != wpc->ioend->io_offset + wpc->ioend->io_size) {
wpc               780 fs/xfs/xfs_aops.c 		if (wpc->ioend)
wpc               781 fs/xfs/xfs_aops.c 			list_add(&wpc->ioend->io_list, iolist);
wpc               782 fs/xfs/xfs_aops.c 		wpc->ioend = xfs_alloc_ioend(inode, wpc->fork,
wpc               783 fs/xfs/xfs_aops.c 				wpc->imap.br_state, offset, bdev, sector, wbc);
wpc               786 fs/xfs/xfs_aops.c 	merged = __bio_try_merge_page(wpc->ioend->io_bio, page, len, poff,
wpc               793 fs/xfs/xfs_aops.c 		if (bio_full(wpc->ioend->io_bio, len))
wpc               794 fs/xfs/xfs_aops.c 			wpc->ioend->io_bio = xfs_chain_bio(wpc->ioend->io_bio);
wpc               795 fs/xfs/xfs_aops.c 		bio_add_page(wpc->ioend->io_bio, page, len, poff);
wpc               798 fs/xfs/xfs_aops.c 	wpc->ioend->io_size += len;
wpc               867 fs/xfs/xfs_aops.c 	struct xfs_writepage_ctx *wpc,
wpc               894 fs/xfs/xfs_aops.c 		error = xfs_map_blocks(wpc, inode, file_offset);
wpc               897 fs/xfs/xfs_aops.c 		if (wpc->imap.br_startblock == HOLESTARTBLOCK)
wpc               899 fs/xfs/xfs_aops.c 		xfs_add_to_ioend(inode, file_offset, page, iop, wpc, wbc,
wpc               904 fs/xfs/xfs_aops.c 	ASSERT(wpc->ioend || list_empty(&submit_list));
wpc               984 fs/xfs/xfs_aops.c 	struct xfs_writepage_ctx *wpc = data;
wpc              1079 fs/xfs/xfs_aops.c 	return xfs_writepage_map(wpc, wbc, inode, page, end_offset);
wpc              1092 fs/xfs/xfs_aops.c 	struct xfs_writepage_ctx wpc = { };
wpc              1095 fs/xfs/xfs_aops.c 	ret = xfs_do_writepage(page, wbc, &wpc);
wpc              1096 fs/xfs/xfs_aops.c 	if (wpc.ioend)
wpc              1097 fs/xfs/xfs_aops.c 		ret = xfs_submit_ioend(wbc, wpc.ioend, ret);
wpc              1106 fs/xfs/xfs_aops.c 	struct xfs_writepage_ctx wpc = { };
wpc              1110 fs/xfs/xfs_aops.c 	ret = write_cache_pages(mapping, wbc, xfs_do_writepage, &wpc);
wpc              1111 fs/xfs/xfs_aops.c 	if (wpc.ioend)
wpc              1112 fs/xfs/xfs_aops.c 		ret = xfs_submit_ioend(wbc, wpc.ioend, ret);