Home
last modified time | relevance | path

Searched refs:chunk (Results 1 – 200 of 211) sorted by relevance

12

/linux-4.4.14/net/sctp/
Dinqueue.c59 struct sctp_chunk *chunk, *tmp; in sctp_inq_free() local
62 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) { in sctp_inq_free()
63 list_del_init(&chunk->list); in sctp_inq_free()
64 sctp_chunk_free(chunk); in sctp_inq_free()
79 void sctp_inq_push(struct sctp_inq *q, struct sctp_chunk *chunk) in sctp_inq_push() argument
82 if (chunk->rcvr->dead) { in sctp_inq_push()
83 sctp_chunk_free(chunk); in sctp_inq_push()
92 list_add_tail(&chunk->list, &q->in_chunk_list); in sctp_inq_push()
93 if (chunk->asoc) in sctp_inq_push()
94 chunk->asoc->stats.ipackets++; in sctp_inq_push()
[all …]
Doutput.c61 struct sctp_chunk *chunk);
63 struct sctp_chunk *chunk);
65 struct sctp_chunk *chunk);
67 struct sctp_chunk *chunk,
87 struct sctp_chunk *chunk = NULL; in sctp_packet_config() local
94 chunk = sctp_get_ecne_prepend(packet->transport->asoc); in sctp_packet_config()
99 if (chunk) in sctp_packet_config()
100 sctp_packet_append_chunk(packet, chunk); in sctp_packet_config()
137 struct sctp_chunk *chunk, *tmp; in sctp_packet_free() local
141 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free()
[all …]
Doutqueue.c218 struct sctp_chunk *chunk, *tmp; in __sctp_outq_teardown() local
224 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
227 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
228 sctp_chunk_free(chunk); in __sctp_outq_teardown()
235 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
237 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
238 sctp_chunk_free(chunk); in __sctp_outq_teardown()
244 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
246 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
247 sctp_chunk_free(chunk); in __sctp_outq_teardown()
[all …]
Dchunk.c75 struct sctp_chunk *chunk; in sctp_datamsg_free() local
80 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free()
81 sctp_chunk_free(chunk); in sctp_datamsg_free()
90 struct sctp_chunk *chunk; in sctp_datamsg_destroy() local
102 chunk = list_entry(pos, struct sctp_chunk, frag_list); in sctp_datamsg_destroy()
105 asoc = chunk->asoc; in sctp_datamsg_destroy()
119 if (chunk->has_tsn) in sctp_datamsg_destroy()
124 ev = sctp_ulpevent_make_send_failed(asoc, chunk, sent, in sctp_datamsg_destroy()
130 sctp_chunk_put(chunk); in sctp_datamsg_destroy()
151 static void sctp_datamsg_assign(struct sctp_datamsg *msg, struct sctp_chunk *chunk) in sctp_datamsg_assign() argument
[all …]
Dsm_statefuns.c65 struct sctp_chunk *chunk,
69 struct sctp_chunk *chunk,
73 const struct sctp_chunk *chunk);
77 const struct sctp_chunk *chunk,
98 static struct sctp_sackhdr *sctp_sm_pull_sack(struct sctp_chunk *chunk);
151 struct sctp_chunk *chunk);
168 sctp_chunk_length_valid(struct sctp_chunk *chunk, in sctp_chunk_length_valid() argument
171 __u16 chunk_length = ntohs(chunk->chunk_hdr->length); in sctp_chunk_length_valid()
174 if (unlikely(chunk->pdiscard)) in sctp_chunk_length_valid()
224 struct sctp_chunk *chunk = arg; in sctp_sf_do_4_C() local
[all …]
Dsm_make_chunk.c79 static void *sctp_addto_param(struct sctp_chunk *chunk, int len,
90 static void sctp_control_set_owner_w(struct sctp_chunk *chunk) in sctp_control_set_owner_w() argument
92 struct sctp_association *asoc = chunk->asoc; in sctp_control_set_owner_w()
93 struct sk_buff *skb = chunk->skb; in sctp_control_set_owner_w()
107 int sctp_chunk_iif(const struct sctp_chunk *chunk) in sctp_chunk_iif() argument
112 af = sctp_get_af_specific(ipver2af(ip_hdr(chunk->skb)->version)); in sctp_chunk_iif()
114 iif = af->skb_iif(chunk->skb); in sctp_chunk_iif()
137 void sctp_init_cause(struct sctp_chunk *chunk, __be16 cause_code, in sctp_init_cause() argument
147 chunk->subh.err_hdr = sctp_addto_chunk(chunk, sizeof(sctp_errhdr_t), &err); in sctp_init_cause()
155 static int sctp_init_cause_fixed(struct sctp_chunk *chunk, __be16 cause_code, in sctp_init_cause_fixed() argument
[all …]
Dsm_sideeffect.c102 struct sctp_chunk *chunk) in sctp_do_ecn_ecne_work() argument
134 repl = sctp_make_cwr(asoc, asoc->last_cwr_tsn, chunk); in sctp_do_ecn_ecne_work()
571 struct sctp_chunk *chunk, in sctp_cmd_assoc_failed() argument
579 if (event_type == SCTP_EVENT_T_CHUNK && subtype.chunk == SCTP_CID_ABORT) in sctp_cmd_assoc_failed()
581 (__u16)error, 0, 0, chunk, in sctp_cmd_assoc_failed()
592 abort = sctp_make_violation_max_retrans(asoc, chunk); in sctp_cmd_assoc_failed()
613 struct sctp_chunk *chunk, in sctp_cmd_process_init() argument
624 if (!sctp_process_init(asoc, chunk, sctp_source(chunk), peer_init, gfp)) in sctp_cmd_process_init()
690 struct sctp_chunk *chunk) in sctp_cmd_transport_on() argument
746 hbinfo = (sctp_sender_hb_info_t *) chunk->skb->data; in sctp_cmd_transport_on()
[all …]
Dendpointola.c413 struct sctp_chunk *chunk; in sctp_endpoint_bh_rcv() local
428 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_endpoint_bh_rcv()
429 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_endpoint_bh_rcv()
434 if (first_time && (subtype.chunk == SCTP_CID_AUTH)) { in sctp_endpoint_bh_rcv()
447 chunk->auth_chunk = skb_clone(chunk->skb, in sctp_endpoint_bh_rcv()
449 chunk->auth = 1; in sctp_endpoint_bh_rcv()
460 if (NULL == chunk->asoc) { in sctp_endpoint_bh_rcv()
462 sctp_source(chunk), in sctp_endpoint_bh_rcv()
464 chunk->asoc = asoc; in sctp_endpoint_bh_rcv()
465 chunk->transport = transport; in sctp_endpoint_bh_rcv()
[all …]
Dulpevent.c129 __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) in sctp_ulpevent_make_assoc_change() argument
138 if (chunk) { in sctp_ulpevent_make_assoc_change()
142 skb = skb_copy_expand(chunk->skb, in sctp_ulpevent_make_assoc_change()
158 ntohs(chunk->chunk_hdr->length) - in sctp_ulpevent_make_assoc_change()
371 struct sctp_chunk *chunk, __u16 flags, in sctp_ulpevent_make_remote_error() argument
381 ch = (sctp_errhdr_t *)(chunk->skb->data); in sctp_ulpevent_make_remote_error()
386 skb_pull(chunk->skb, sizeof(sctp_errhdr_t)); in sctp_ulpevent_make_remote_error()
391 skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp); in sctp_ulpevent_make_remote_error()
394 skb_pull(chunk->skb, elen); in sctp_ulpevent_make_remote_error()
427 const struct sctp_association *asoc, struct sctp_chunk *chunk, in sctp_ulpevent_make_send_failed() argument
[all …]
Dinput.c100 struct sctp_chunk *chunk; member
114 struct sctp_chunk *chunk; in sctp_rcv() local
224 chunk = sctp_chunkify(skb, asoc, sk); in sctp_rcv()
225 if (!chunk) in sctp_rcv()
227 SCTP_INPUT_CB(skb)->chunk = chunk; in sctp_rcv()
230 chunk->rcvr = rcvr; in sctp_rcv()
233 chunk->sctp_hdr = sh; in sctp_rcv()
236 sctp_init_addrs(chunk, &src, &dest); in sctp_rcv()
239 chunk->transport = transport; in sctp_rcv()
263 sctp_chunk_free(chunk); in sctp_rcv()
[all …]
Ddebug.c65 if (cid.chunk <= SCTP_CID_BASE_MAX) in sctp_cname()
66 return sctp_cid_tbl[cid.chunk]; in sctp_cname()
68 switch (cid.chunk) { in sctp_cname()
Dauth.c391 struct sctp_chunk *chunk; in sctp_auth_asoc_init_active_key() local
417 list_for_each_entry(chunk, &asoc->outqueue.out_chunk_list, list) { in sctp_auth_asoc_init_active_key()
418 if (sctp_auth_send_cid(chunk->chunk_hdr->type, asoc)) in sctp_auth_asoc_init_active_key()
419 chunk->auth = 1; in sctp_auth_asoc_init_active_key()
638 static int __sctp_auth_cid(sctp_cid_t chunk, struct sctp_chunks_param *param) in __sctp_auth_cid() argument
664 if (param->chunks[i] == chunk) in __sctp_auth_cid()
674 int sctp_auth_send_cid(sctp_cid_t chunk, const struct sctp_association *asoc) in sctp_auth_send_cid() argument
682 return __sctp_auth_cid(chunk, asoc->peer.peer_chunks); in sctp_auth_send_cid()
686 int sctp_auth_recv_cid(sctp_cid_t chunk, const struct sctp_association *asoc) in sctp_auth_recv_cid() argument
694 return __sctp_auth_cid(chunk, in sctp_auth_recv_cid()
Dassociola.c919 struct sctp_chunk *chunk; in sctp_assoc_lookup_tsn() local
941 list_for_each_entry(chunk, &active->transmitted, in sctp_assoc_lookup_tsn()
944 if (key == chunk->subh.data_hdr->tsn) { in sctp_assoc_lookup_tsn()
956 list_for_each_entry(chunk, &transport->transmitted, in sctp_assoc_lookup_tsn()
958 if (key == chunk->subh.data_hdr->tsn) { in sctp_assoc_lookup_tsn()
1001 struct sctp_chunk *chunk; in sctp_assoc_bh_rcv() local
1012 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_assoc_bh_rcv()
1014 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_assoc_bh_rcv()
1023 if (sctp_auth_recv_cid(subtype.chunk, asoc) && !chunk->auth) in sctp_assoc_bh_rcv()
1029 if (sctp_chunk_is_data(chunk)) in sctp_assoc_bh_rcv()
[all …]
Dobjcnt.c52 SCTP_DBG_OBJCNT(chunk);
66 SCTP_DBG_OBJCNT_ENTRY(chunk),
Dulpqueue.c95 int sctp_ulpq_tail_data(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, in sctp_ulpq_tail_data() argument
103 event = sctp_ulpevent_make_rcvmsg(chunk->asoc, chunk, gfp); in sctp_ulpq_tail_data()
1080 void sctp_ulpq_renege(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, in sctp_ulpq_renege() argument
1088 if (chunk) { in sctp_ulpq_renege()
1089 needed = ntohs(chunk->chunk_hdr->length); in sctp_ulpq_renege()
1103 if (chunk && (freed >= needed)) { in sctp_ulpq_renege()
1105 retval = sctp_ulpq_tail_data(ulpq, chunk, gfp); in sctp_ulpq_renege()
Dsocket.c99 struct sctp_chunk *chunk);
148 static inline void sctp_set_owner_w(struct sctp_chunk *chunk) in sctp_set_owner_w() argument
150 struct sctp_association *asoc = chunk->asoc; in sctp_set_owner_w()
156 skb_set_owner_w(chunk->skb, sk); in sctp_set_owner_w()
158 chunk->skb->destructor = sctp_wfree; in sctp_set_owner_w()
160 skb_shinfo(chunk->skb)->destructor_arg = chunk; in sctp_set_owner_w()
162 asoc->sndbuf_used += SCTP_DATA_SNDSIZE(chunk) + in sctp_set_owner_w()
167 sk->sk_wmem_queued += chunk->skb->truesize; in sctp_set_owner_w()
168 sk_mem_charge(sk, chunk->skb->truesize); in sctp_set_owner_w()
411 struct sctp_chunk *chunk) in sctp_send_asconf() argument
[all …]
Dprobe.c137 struct sctp_chunk *chunk = arg; in jsctp_sf_eat_sack() local
138 struct sk_buff *skb = chunk->skb; in jsctp_sf_eat_sack()
DMakefile10 transport.o chunk.o sm_make_chunk.o ulpevent.o \
Dsm_statetable.c86 return sctp_chunk_event_lookup(net, event_subtype.chunk, state); in sctp_sm_lookup_event()
/linux-4.4.14/mm/
Dpercpu-vm.c13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk, in pcpu_chunk_page() argument
17 WARN_ON(chunk->immutable); in pcpu_chunk_page()
19 return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx)); in pcpu_chunk_page()
55 static void pcpu_free_pages(struct pcpu_chunk *chunk, in pcpu_free_pages() argument
82 static int pcpu_alloc_pages(struct pcpu_chunk *chunk, in pcpu_alloc_pages() argument
125 static void pcpu_pre_unmap_flush(struct pcpu_chunk *chunk, in pcpu_pre_unmap_flush() argument
129 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start), in pcpu_pre_unmap_flush()
130 pcpu_chunk_addr(chunk, pcpu_high_unit_cpu, page_end)); in pcpu_pre_unmap_flush()
151 static void pcpu_unmap_pages(struct pcpu_chunk *chunk, in pcpu_unmap_pages() argument
161 page = pcpu_chunk_page(chunk, cpu, i); in pcpu_unmap_pages()
[all …]
Dpercpu.c218 static int pcpu_chunk_slot(const struct pcpu_chunk *chunk) in pcpu_chunk_slot() argument
220 if (chunk->free_size < sizeof(int) || chunk->contig_hint < sizeof(int)) in pcpu_chunk_slot()
223 return pcpu_size_to_slot(chunk->free_size); in pcpu_chunk_slot()
243 static unsigned long pcpu_chunk_addr(struct pcpu_chunk *chunk, in pcpu_chunk_addr() argument
246 return (unsigned long)chunk->base_addr + pcpu_unit_offsets[cpu] + in pcpu_chunk_addr()
250 static void __maybe_unused pcpu_next_unpop(struct pcpu_chunk *chunk, in pcpu_next_unpop() argument
253 *rs = find_next_zero_bit(chunk->populated, end, *rs); in pcpu_next_unpop()
254 *re = find_next_bit(chunk->populated, end, *rs + 1); in pcpu_next_unpop()
257 static void __maybe_unused pcpu_next_pop(struct pcpu_chunk *chunk, in pcpu_next_pop() argument
260 *rs = find_next_bit(chunk->populated, end, *rs); in pcpu_next_pop()
[all …]
Dpercpu-km.c36 static int pcpu_populate_chunk(struct pcpu_chunk *chunk, in pcpu_populate_chunk() argument
42 static void pcpu_depopulate_chunk(struct pcpu_chunk *chunk, in pcpu_depopulate_chunk() argument
51 struct pcpu_chunk *chunk; in pcpu_create_chunk() local
55 chunk = pcpu_alloc_chunk(); in pcpu_create_chunk()
56 if (!chunk) in pcpu_create_chunk()
61 pcpu_free_chunk(chunk); in pcpu_create_chunk()
66 pcpu_set_page_chunk(nth_page(pages, i), chunk); in pcpu_create_chunk()
68 chunk->data = pages; in pcpu_create_chunk()
69 chunk->base_addr = page_address(pages) - pcpu_group_offsets[0]; in pcpu_create_chunk()
72 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk()
[all …]
/linux-4.4.14/drivers/s390/cio/
Ditcw.c181 void *chunk; in itcw_init() local
193 chunk = fit_chunk(&start, end, sizeof(struct itcw), 1, 0); in itcw_init()
194 if (IS_ERR(chunk)) in itcw_init()
195 return chunk; in itcw_init()
196 itcw = chunk; in itcw_init()
209 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init()
210 if (IS_ERR(chunk)) in itcw_init()
211 return chunk; in itcw_init()
212 itcw->tcw = chunk; in itcw_init()
217 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init()
[all …]
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/
Dicm.c54 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_pages() argument
58 if (chunk->nsg > 0) in mlx4_free_icm_pages()
59 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, in mlx4_free_icm_pages()
62 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages()
63 __free_pages(sg_page(&chunk->mem[i]), in mlx4_free_icm_pages()
64 get_order(chunk->mem[i].length)); in mlx4_free_icm_pages()
67 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_coherent() argument
71 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent()
73 chunk->mem[i].length, in mlx4_free_icm_coherent()
74 lowmem_page_address(sg_page(&chunk->mem[i])), in mlx4_free_icm_coherent()
[all …]
Dicm.h64 struct mlx4_icm_chunk *chunk; member
91 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first()
99 return !iter->chunk; in mlx4_icm_last()
104 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next()
105 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
106 iter->chunk = NULL; in mlx4_icm_next()
110 iter->chunk = list_entry(iter->chunk->list.next, in mlx4_icm_next()
118 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_addr()
123 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_size()
Dmr.c724 int chunk; in __mlx4_write_mtt() local
733 chunk = min_t(int, max_mtts_first_page, npages); in __mlx4_write_mtt()
736 err = mlx4_write_mtt_chunk(dev, mtt, start_index, chunk, page_list); in __mlx4_write_mtt()
739 npages -= chunk; in __mlx4_write_mtt()
740 start_index += chunk; in __mlx4_write_mtt()
741 page_list += chunk; in __mlx4_write_mtt()
743 chunk = min_t(int, mtts_per_page, npages); in __mlx4_write_mtt()
753 int chunk; in mlx4_write_mtt() local
767 chunk = min_t(int, MLX4_MAILBOX_SIZE / sizeof(u64) - 2, in mlx4_write_mtt()
771 for (i = 0; i < chunk; ++i) in mlx4_write_mtt()
[all …]
/linux-4.4.14/lib/
Dgenalloc.c39 static inline size_t chunk_size(const struct gen_pool_chunk *chunk) in chunk_size() argument
41 return chunk->end_addr - chunk->start_addr + 1; in chunk_size()
185 struct gen_pool_chunk *chunk; in gen_pool_add_virt() local
190 chunk = kzalloc_node(nbytes, GFP_KERNEL, nid); in gen_pool_add_virt()
191 if (unlikely(chunk == NULL)) in gen_pool_add_virt()
194 chunk->phys_addr = phys; in gen_pool_add_virt()
195 chunk->start_addr = virt; in gen_pool_add_virt()
196 chunk->end_addr = virt + size - 1; in gen_pool_add_virt()
197 atomic_set(&chunk->avail, size); in gen_pool_add_virt()
200 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_virt()
[all …]
Dbitmap.c363 u32 chunk; in __bitmap_parse() local
370 chunk = 0; in __bitmap_parse()
406 if (chunk & ~((1UL << (CHUNKSZ - 4)) - 1)) in __bitmap_parse()
409 chunk = (chunk << 4) | hex_to_bin(c); in __bitmap_parse()
414 if (nchunks == 0 && chunk == 0) in __bitmap_parse()
418 *maskp |= chunk; in __bitmap_parse()
420 nbits += (nchunks == 1) ? nbits_to_hold_value(chunk) : CHUNKSZ; in __bitmap_parse()
DKconfig517 Provides a heler to split scatterlists into chunks, each chunk being a
/linux-4.4.14/kernel/
Daudit_tree.c109 static void free_chunk(struct audit_chunk *chunk) in free_chunk() argument
113 for (i = 0; i < chunk->count; i++) { in free_chunk()
114 if (chunk->owners[i].owner) in free_chunk()
115 put_tree(chunk->owners[i].owner); in free_chunk()
117 kfree(chunk); in free_chunk()
120 void audit_put_chunk(struct audit_chunk *chunk) in audit_put_chunk() argument
122 if (atomic_long_dec_and_test(&chunk->refs)) in audit_put_chunk()
123 free_chunk(chunk); in audit_put_chunk()
128 struct audit_chunk *chunk = container_of(rcu, struct audit_chunk, head); in __put_chunk() local
129 audit_put_chunk(chunk); in __put_chunk()
[all …]
Dauditsc.c210 static int put_tree_ref(struct audit_context *ctx, struct audit_chunk *chunk) in put_tree_ref() argument
215 p->c[--left] = chunk; in put_tree_ref()
223 p->c[30] = chunk; in put_tree_ref()
1588 struct audit_chunk *chunk; in handle_one() local
1596 chunk = audit_tree_lookup(inode); in handle_one()
1598 if (!chunk) in handle_one()
1600 if (likely(put_tree_ref(context, chunk))) in handle_one()
1605 audit_put_chunk(chunk); in handle_one()
1609 put_tree_ref(context, chunk); in handle_one()
1634 struct audit_chunk *chunk; in handle_path() local
[all …]
/linux-4.4.14/drivers/infiniband/hw/mthca/
Dmthca_memfree.c64 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_pages() argument
68 if (chunk->nsg > 0) in mthca_free_icm_pages()
69 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, in mthca_free_icm_pages()
72 for (i = 0; i < chunk->npages; ++i) in mthca_free_icm_pages()
73 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages()
74 get_order(chunk->mem[i].length)); in mthca_free_icm_pages()
77 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_coherent() argument
81 for (i = 0; i < chunk->npages; ++i) { in mthca_free_icm_coherent()
82 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, in mthca_free_icm_coherent()
83 lowmem_page_address(sg_page(&chunk->mem[i])), in mthca_free_icm_coherent()
[all …]
Dmthca_memfree.h76 struct mthca_icm_chunk *chunk; member
103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first()
111 return !iter->chunk; in mthca_icm_last()
116 if (++iter->page_idx >= iter->chunk->nsg) { in mthca_icm_next()
117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next()
118 iter->chunk = NULL; in mthca_icm_next()
122 iter->chunk = list_entry(iter->chunk->list.next, in mthca_icm_next()
130 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mthca_icm_addr()
135 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mthca_icm_size()
Dmthca_mr.c362 int chunk; in mthca_write_mtt() local
369 chunk = min(size, list_len); in mthca_write_mtt()
372 buffer_list, chunk); in mthca_write_mtt()
375 buffer_list, chunk); in mthca_write_mtt()
377 list_len -= chunk; in mthca_write_mtt()
378 start_index += chunk; in mthca_write_mtt()
379 buffer_list += chunk; in mthca_write_mtt()
/linux-4.4.14/drivers/gpu/drm/qxl/
Dqxl_image.c38 struct qxl_drm_chunk *chunk; in qxl_allocate_chunk() local
41 chunk = kmalloc(sizeof(struct qxl_drm_chunk), GFP_KERNEL); in qxl_allocate_chunk()
42 if (!chunk) in qxl_allocate_chunk()
45 ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo); in qxl_allocate_chunk()
47 kfree(chunk); in qxl_allocate_chunk()
51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk()
88 struct qxl_drm_chunk *chunk, *tmp; in qxl_image_free_objects() local
90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects()
91 qxl_bo_unref(&chunk->bo); in qxl_image_free_objects()
92 kfree(chunk); in qxl_image_free_objects()
[all …]
Dqxl_dev.h421 struct qxl_data_chunk chunk; member
482 struct qxl_data_chunk chunk; member
761 struct qxl_data_chunk chunk; member
779 struct qxl_data_chunk chunk; member
Dqxl_draw.c52 dev_clips->chunk.next_chunk = 0; in drawable_set_clipping()
53 dev_clips->chunk.prev_chunk = 0; in drawable_set_clipping()
54 dev_clips->chunk.data_size = sizeof(struct qxl_rect) * num_clips; in drawable_set_clipping()
55 return (struct qxl_rect *)dev_clips->chunk.data; in drawable_set_clipping()
Dqxl_display.c368 cursor->chunk.next_chunk = 0; in qxl_crtc_cursor_set2()
369 cursor->chunk.prev_chunk = 0; in qxl_crtc_cursor_set2()
370 cursor->chunk.data_size = size; in qxl_crtc_cursor_set2()
372 memcpy(cursor->chunk.data, user_ptr, size); in qxl_crtc_cursor_set2()
/linux-4.4.14/arch/mips/dec/prom/
Dconsole.c22 unsigned int chunk = sizeof(buf) - 1; in prom_console_write() local
25 if (chunk > c) in prom_console_write()
26 chunk = c; in prom_console_write()
27 memcpy(buf, s, chunk); in prom_console_write()
28 buf[chunk] = '\0'; in prom_console_write()
30 s += chunk; in prom_console_write()
31 c -= chunk; in prom_console_write()
/linux-4.4.14/sound/core/
Dsgbuf.c71 unsigned int i, pages, chunk, maxpages; in snd_malloc_sgbuf_pages() local
96 chunk = pages; in snd_malloc_sgbuf_pages()
98 if (chunk > maxpages) in snd_malloc_sgbuf_pages()
99 chunk = maxpages; in snd_malloc_sgbuf_pages()
100 chunk <<= PAGE_SHIFT; in snd_malloc_sgbuf_pages()
102 chunk, &tmpb) < 0) { in snd_malloc_sgbuf_pages()
110 chunk = tmpb.bytes >> PAGE_SHIFT; in snd_malloc_sgbuf_pages()
111 for (i = 0; i < chunk; i++) { in snd_malloc_sgbuf_pages()
115 table->addr |= chunk; /* mark head */ in snd_malloc_sgbuf_pages()
121 sgbuf->pages += chunk; in snd_malloc_sgbuf_pages()
[all …]
/linux-4.4.14/drivers/media/usb/usbtv/
Dusbtv.h49 #define USBTV_MAGIC_OK(chunk) ((be32_to_cpu(chunk[0]) & 0xff000000) \ argument
51 #define USBTV_FRAME_ID(chunk) ((be32_to_cpu(chunk[0]) & 0x00ff0000) >> 16) argument
52 #define USBTV_ODD(chunk) ((be32_to_cpu(chunk[0]) & 0x0000f000) >> 15) argument
53 #define USBTV_CHUNK_NO(chunk) (be32_to_cpu(chunk[0]) & 0x00000fff) argument
Dusbtv-video.c275 static void usbtv_image_chunk(struct usbtv *usbtv, __be32 *chunk) in usbtv_image_chunk() argument
283 if (!USBTV_MAGIC_OK(chunk)) in usbtv_image_chunk()
285 frame_id = USBTV_FRAME_ID(chunk); in usbtv_image_chunk()
286 odd = USBTV_ODD(chunk); in usbtv_image_chunk()
287 chunk_no = USBTV_CHUNK_NO(chunk); in usbtv_image_chunk()
312 usbtv_chunk_to_vbuf(frame, &chunk[1], chunk_no, odd); in usbtv_image_chunk()
/linux-4.4.14/drivers/infiniband/hw/usnic/
Dusnic_vnic.c96 struct usnic_vnic_res_chunk *chunk; in usnic_vnic_dump() local
119 chunk = &vnic->chunks[i]; in usnic_vnic_dump()
120 for (j = 0; j < chunk->cnt; j++) { in usnic_vnic_dump()
121 res = chunk->res[j]; in usnic_vnic_dump()
277 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk) in usnic_vnic_put_resources() argument
282 struct usnic_vnic *vnic = chunk->vnic; in usnic_vnic_put_resources()
285 while ((i = --chunk->cnt) >= 0) { in usnic_vnic_put_resources()
286 res = chunk->res[i]; in usnic_vnic_put_resources()
287 chunk->res[i] = NULL; in usnic_vnic_put_resources()
293 kfree(chunk->res); in usnic_vnic_put_resources()
[all …]
Dusnic_uiom.c81 struct usnic_uiom_chunk *chunk, *tmp; in usnic_uiom_put_pages() local
87 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages()
88 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_put_pages()
96 kfree(chunk); in usnic_uiom_put_pages()
105 struct usnic_uiom_chunk *chunk; in usnic_uiom_get_pages() local
159 chunk = kmalloc(sizeof(*chunk) + in usnic_uiom_get_pages()
163 if (!chunk) { in usnic_uiom_get_pages()
168 chunk->nents = min_t(int, ret, USNIC_UIOM_PAGE_CHUNK); in usnic_uiom_get_pages()
169 sg_init_table(chunk->page_list, chunk->nents); in usnic_uiom_get_pages()
170 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_get_pages()
[all …]
Dusnic_ib_verbs.c64 struct usnic_vnic_res_chunk *chunk; in usnic_ib_fill_create_qp_resp() local
89 chunk = usnic_ib_qp_grp_get_chunk(qp_grp, USNIC_VNIC_RES_TYPE_RQ); in usnic_ib_fill_create_qp_resp()
90 if (IS_ERR_OR_NULL(chunk)) { in usnic_ib_fill_create_qp_resp()
94 PTR_ERR(chunk)); in usnic_ib_fill_create_qp_resp()
95 return chunk ? PTR_ERR(chunk) : -ENOMEM; in usnic_ib_fill_create_qp_resp()
98 WARN_ON(chunk->type != USNIC_VNIC_RES_TYPE_RQ); in usnic_ib_fill_create_qp_resp()
99 resp.rq_cnt = chunk->cnt; in usnic_ib_fill_create_qp_resp()
100 for (i = 0; i < chunk->cnt; i++) in usnic_ib_fill_create_qp_resp()
101 resp.rq_idx[i] = chunk->res[i]->vnic_idx; in usnic_ib_fill_create_qp_resp()
103 chunk = usnic_ib_qp_grp_get_chunk(qp_grp, USNIC_VNIC_RES_TYPE_WQ); in usnic_ib_fill_create_qp_resp()
[all …]
Dusnic_vnic.h110 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk);
/linux-4.4.14/drivers/s390/block/
Ddasd_int.h574 struct dasd_mchunk *chunk; in dasd_init_chunklist() local
577 chunk = (struct dasd_mchunk *) mem; in dasd_init_chunklist()
578 chunk->size = size - sizeof(struct dasd_mchunk); in dasd_init_chunklist()
579 list_add(&chunk->list, chunk_list); in dasd_init_chunklist()
585 struct dasd_mchunk *chunk, *tmp; in dasd_alloc_chunk() local
588 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk()
589 if (chunk->size < size) in dasd_alloc_chunk()
591 if (chunk->size > size + sizeof(struct dasd_mchunk)) { in dasd_alloc_chunk()
592 char *endaddr = (char *) (chunk + 1) + chunk->size; in dasd_alloc_chunk()
595 chunk->size -= size + sizeof(struct dasd_mchunk); in dasd_alloc_chunk()
[all …]
/linux-4.4.14/fs/ocfs2/
Dquota_local.c296 int chunk, in ocfs2_add_recovery_chunk() argument
304 rc->rc_chunk = chunk; in ocfs2_add_recovery_chunk()
470 int bit, chunk; in ocfs2_recover_local_quota_file() local
477 chunk = rchunk->rc_chunk; in ocfs2_recover_local_quota_file()
480 ol_quota_chunk_block(sb, chunk), in ocfs2_recover_local_quota_file()
490 ol_dqblk_block(sb, chunk, bit), in ocfs2_recover_local_quota_file()
497 ol_dqblk_block_off(sb, chunk, bit)); in ocfs2_recover_local_quota_file()
816 struct ocfs2_quota_chunk *chunk; in ocfs2_local_free_info() local
824 list_for_each_entry(chunk, &oinfo->dqi_chunk, qc_chunk) { in ocfs2_local_free_info()
826 (chunk->qc_headerbh->b_data); in ocfs2_local_free_info()
[all …]
Dioctl.c448 unsigned int offset = 0, cluster, chunk; in ocfs2_info_freefrag_scan_chain() local
487 for (chunk = 0; chunk < chunks_in_group; chunk++) { in ocfs2_info_freefrag_scan_chain()
/linux-4.4.14/drivers/media/usb/go7007/
Dgo7007-fw.c378 int size = 0, i, off = 0, chunk; in gen_mjpeghdr_to_package() local
388 chunk = mjpeg_frame_header(go, buf + size, 1); in gen_mjpeghdr_to_package()
389 memmove(buf + size, buf + size + 80, chunk - 80); in gen_mjpeghdr_to_package()
390 size += chunk - 80; in gen_mjpeghdr_to_package()
392 for (i = 0; i < size; i += chunk * 2) { in gen_mjpeghdr_to_package()
400 chunk = 28; in gen_mjpeghdr_to_package()
401 if (mem + chunk > 0x4000) in gen_mjpeghdr_to_package()
402 chunk = 0x4000 - mem; in gen_mjpeghdr_to_package()
403 if (i + 2 * chunk > size) in gen_mjpeghdr_to_package()
404 chunk = (size - i) / 2; in gen_mjpeghdr_to_package()
[all …]
/linux-4.4.14/include/net/sctp/
Dsm.h174 __u8 sctp_get_chunk_type(struct sctp_chunk *chunk);
209 const struct sctp_chunk *chunk);
239 const struct sctp_chunk *chunk,
252 struct sctp_chunk *chunk, bool addr_param_needed,
299 static inline __u16 sctp_data_size(struct sctp_chunk *chunk) in sctp_data_size() argument
303 size = ntohs(chunk->chunk_hdr->length); in sctp_data_size()
392 sctp_vtag_verify(const struct sctp_chunk *chunk, in sctp_vtag_verify() argument
401 if (ntohl(chunk->sctp_hdr->vtag) == asoc->c.my_vtag) in sctp_vtag_verify()
411 sctp_vtag_verify_either(const struct sctp_chunk *chunk, in sctp_vtag_verify_either() argument
435 if ((!sctp_test_T_bit(chunk) && in sctp_vtag_verify_either()
[all …]
Dsctp.h432 static inline bool sctp_chunk_pending(const struct sctp_chunk *chunk) in sctp_chunk_pending() argument
434 return !list_empty(&chunk->list); in sctp_chunk_pending()
442 #define sctp_walk_params(pos, chunk, member)\ argument
443 _sctp_walk_params((pos), (chunk), ntohs((chunk)->chunk_hdr.length), member)
445 #define _sctp_walk_params(pos, chunk, end, member)\ argument
446 for (pos.v = chunk->member;\
447 pos.v <= (void *)chunk + end - ntohs(pos.p->length) &&\
461 #define sctp_walk_fwdtsn(pos, chunk)\ argument
462 _sctp_walk_fwdtsn((pos), (chunk), ntohs((chunk)->chunk_hdr->length) - sizeof(struct sctp_fwdtsn_chu…
464 #define _sctp_walk_fwdtsn(pos, chunk, end)\ argument
[all …]
Dulpevent.h85 struct sctp_chunk *chunk,
98 struct sctp_chunk *chunk,
103 struct sctp_chunk *chunk,
121 struct sctp_chunk *chunk,
Dauth.h100 int sctp_auth_send_cid(sctp_cid_t chunk, const struct sctp_association *asoc);
101 int sctp_auth_recv_cid(sctp_cid_t chunk, const struct sctp_association *asoc);
Dconstants.h128 sctp_cid_t chunk; member
139 SCTP_SUBTYPE_CONSTRUCTOR(CHUNK, sctp_cid_t, chunk)
Dcommand.h131 struct sctp_chunk *chunk; member
172 SCTP_ARG_CONSTRUCTOR(CHUNK, struct sctp_chunk *, chunk) in SCTP_ARG_CONSTRUCTOR()
Dstructs.h654 int sctp_user_addto_chunk(struct sctp_chunk *chunk, int len,
663 const union sctp_addr *sctp_source(const struct sctp_chunk *chunk);
1058 int sctp_outq_tail(struct sctp_outq *, struct sctp_chunk *chunk);
1281 struct sctp_chunk *chunk, struct sctp_chunk **err_chunk);
1282 int sctp_process_init(struct sctp_association *, struct sctp_chunk *chunk,
/linux-4.4.14/sound/soc/codecs/
Dsigmadsp.c54 struct sigma_fw_chunk chunk; member
60 struct sigma_fw_chunk chunk; member
68 struct sigma_fw_chunk chunk; member
200 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_control() argument
212 ctrl_chunk = (const struct sigma_fw_chunk_control *)chunk; in sigma_fw_load_control()
238 ctrl->samplerates = le32_to_cpu(chunk->samplerates); in sigma_fw_load_control()
251 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_data() argument
259 data_chunk = (struct sigma_fw_chunk_data *)chunk; in sigma_fw_load_data()
269 data->samplerates = le32_to_cpu(chunk->samplerates); in sigma_fw_load_data()
277 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_samplerates() argument
[all …]
/linux-4.4.14/arch/um/drivers/
Dmconsole_kern.h27 #define CONFIG_CHUNK(str, size, current, chunk, end) \ argument
29 current += strlen(chunk); \
33 strcpy(str, chunk); \
34 str += strlen(chunk); \
/linux-4.4.14/drivers/dma/sh/
Drcar-dmac.c328 struct rcar_dmac_xfer_chunk *chunk; in rcar_dmac_chan_start_xfer() local
352 chunk = list_first_entry(&desc->chunks, in rcar_dmac_chan_start_xfer()
355 chunk->dst_addr & 0xffffffff); in rcar_dmac_chan_start_xfer()
385 struct rcar_dmac_xfer_chunk *chunk = desc->running; in rcar_dmac_chan_start_xfer() local
389 chan->index, chunk, chunk->size, &chunk->src_addr, in rcar_dmac_chan_start_xfer()
390 &chunk->dst_addr); in rcar_dmac_chan_start_xfer()
394 chunk->src_addr >> 32); in rcar_dmac_chan_start_xfer()
396 chunk->dst_addr >> 32); in rcar_dmac_chan_start_xfer()
399 chunk->src_addr & 0xffffffff); in rcar_dmac_chan_start_xfer()
401 chunk->dst_addr & 0xffffffff); in rcar_dmac_chan_start_xfer()
[all …]
Dshdma-base.c75 struct shdma_desc *chunk, *c, *desc = in shdma_tx_submit() local
89 list_for_each_entry_safe(chunk, c, desc->node.prev, node) { in shdma_tx_submit()
94 if (chunk != desc && (chunk->mark == DESC_IDLE || in shdma_tx_submit()
95 chunk->async_tx.cookie > 0 || in shdma_tx_submit()
96 chunk->async_tx.cookie == -EBUSY || in shdma_tx_submit()
97 &chunk->node == &schan->ld_free)) in shdma_tx_submit()
99 chunk->mark = DESC_SUBMITTED; in shdma_tx_submit()
100 if (chunk->chunks == 1) { in shdma_tx_submit()
101 chunk->async_tx.callback = callback; in shdma_tx_submit()
102 chunk->async_tx.callback_param = tx->callback_param; in shdma_tx_submit()
[all …]
/linux-4.4.14/drivers/md/
Ddm-snap.c164 chunk_t chunk) in chunk_to_sector() argument
166 return chunk << store->chunk_shift; in chunk_to_sector()
222 chunk_t chunk; member
237 static void track_chunk(struct dm_snapshot *s, struct bio *bio, chunk_t chunk) in track_chunk() argument
241 c->chunk = chunk; in track_chunk()
245 &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)]); in track_chunk()
259 static int __chunk_is_tracked(struct dm_snapshot *s, chunk_t chunk) in __chunk_is_tracked() argument
267 &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)], node) { in __chunk_is_tracked()
268 if (c->chunk == chunk) { in __chunk_is_tracked()
283 static void __check_for_conflicting_io(struct dm_snapshot *s, chunk_t chunk) in __check_for_conflicting_io() argument
[all …]
Ddm-stripe.c216 sector_t chunk = dm_target_offset(sc->ti, sector); in stripe_map_sector() local
220 chunk_offset = sector_div(chunk, sc->chunk_size); in stripe_map_sector()
222 chunk_offset = chunk & (sc->chunk_size - 1); in stripe_map_sector()
223 chunk >>= sc->chunk_size_shift; in stripe_map_sector()
227 *stripe = sector_div(chunk, sc->stripes); in stripe_map_sector()
229 *stripe = chunk & (sc->stripes - 1); in stripe_map_sector()
230 chunk >>= sc->stripes_shift; in stripe_map_sector()
234 chunk *= sc->chunk_size; in stripe_map_sector()
236 chunk <<= sc->chunk_size_shift; in stripe_map_sector()
238 *result = chunk + chunk_offset; in stripe_map_sector()
Ddm-exception-store.h142 static inline chunk_t dm_chunk_number(chunk_t chunk) in dm_chunk_number() argument
144 return chunk & (chunk_t)((1ULL << DM_CHUNK_NUMBER_BITS) - 1ULL); in dm_chunk_number()
169 static inline chunk_t dm_chunk_number(chunk_t chunk) in dm_chunk_number() argument
171 return chunk; in dm_chunk_number()
Dbitmap.c724 unsigned long chunk) in file_page_index() argument
727 chunk += sizeof(bitmap_super_t) << 3; in file_page_index()
728 return chunk >> PAGE_BIT_SHIFT; in file_page_index()
733 unsigned long chunk) in file_page_offset() argument
736 chunk += sizeof(bitmap_super_t) << 3; in file_page_offset()
737 return chunk & (PAGE_BITS - 1); in file_page_offset()
745 unsigned long chunk) in filemap_get_page() argument
747 if (file_page_index(store, chunk) >= store->file_pages) in filemap_get_page()
749 return store->filemap[file_page_index(store, chunk)]; in filemap_get_page()
908 unsigned long chunk = block >> bitmap->counts.chunkshift; in bitmap_file_set_bit() local
[all …]
Ddm-snap-persistent.c229 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, int rw, in chunk_io() argument
234 .sector = ps->store->chunk_size * chunk, in chunk_io()
287 chunk_t chunk; in area_io() local
289 chunk = area_location(ps, ps->current_area); in area_io()
291 r = chunk_io(ps, ps->area, chunk, rw, 0); in area_io()
519 chunk_t chunk; in read_exceptions() local
534 chunk = area_location(ps, ps->current_area); in read_exceptions()
536 area = dm_bufio_read(client, chunk, &bp); in read_exceptions()
550 dm_bufio_forget(client, chunk); in read_exceptions()
Draid0.c308 sector_t chunk; in map_sector() local
319 chunk = *sector_offset; in map_sector()
321 sector_div(chunk, zone->nb_dev << chunksect_bits); in map_sector()
324 chunk = *sector_offset; in map_sector()
325 sector_div(chunk, chunk_sects * zone->nb_dev); in map_sector()
332 *sector_offset = (chunk * chunk_sects) + sect_in_chunk; in map_sector()
Draid10.c556 sector_t chunk; in __raid10_find_phys() local
569 chunk = r10bio->sector >> geo->chunk_shift; in __raid10_find_phys()
572 chunk *= geo->near_copies; in __raid10_find_phys()
573 stripe = chunk; in __raid10_find_phys()
632 sector_t offset, chunk, vchunk; in raid10_find_virt() local
655 chunk = sector >> geo->chunk_shift; in raid10_find_virt()
656 fc = sector_div(chunk, geo->far_copies); in raid10_find_virt()
668 chunk = sector >> geo->chunk_shift; in raid10_find_virt()
670 vchunk = chunk * geo->raid_disks + dev; in raid10_find_virt()
3387 int layout, chunk, disks; in setup_geo() local
[all …]
DKconfig68 up evenly, one chunk here and one chunk there. This will increase
Dmd.c4133 unsigned long chunk, end_chunk; in bitmap_store() local
4143 chunk = end_chunk = simple_strtoul(buf, &end, 0); in bitmap_store()
4151 bitmap_dirty_bits(mddev->bitmap, chunk, end_chunk); in bitmap_store()
4591 int chunk; in max_sync_store() local
4606 chunk = mddev->chunk_sectors; in max_sync_store()
4607 if (chunk) { in max_sync_store()
4611 if (sector_div(temp, chunk)) in max_sync_store()
/linux-4.4.14/Documentation/device-mapper/
Dstriped.txt10 Parameters: <num devs> <chunk size> [<dev path> <offset>]+
12 <chunk size>: Size of each chunk of data. Must be at least as
19 be a multiple of the chunk size multiplied by the number of underlying devices.
28 # will be called "stripe_dev" and have a chunk-size of 128k.
Dcache.txt141 Policy plug-ins can store a chunk of data per cache block. It's up to
142 the policy how big this chunk is, but it should be kept small. Like the
Dsnapshot.txt65 deferred until the merging snapshot's corresponding chunk(s) have been
/linux-4.4.14/arch/arm64/kernel/
Dsys_compat.c37 unsigned long chunk = min(PAGE_SIZE, end - start); in __do_compat_cache_op() local
42 ret = __flush_cache_user_range(start, start + chunk); in __do_compat_cache_op()
47 start += chunk; in __do_compat_cache_op()
/linux-4.4.14/net/bluetooth/
Daf_bluetooth.c306 int chunk; in bt_sock_stream_recvmsg() local
332 chunk = min_t(unsigned int, skb->len, size); in bt_sock_stream_recvmsg()
333 if (skb_copy_datagram_msg(skb, 0, msg, chunk)) { in bt_sock_stream_recvmsg()
339 copied += chunk; in bt_sock_stream_recvmsg()
340 size -= chunk; in bt_sock_stream_recvmsg()
347 if (chunk <= skb_len) { in bt_sock_stream_recvmsg()
348 __skb_pull(skb, chunk); in bt_sock_stream_recvmsg()
353 chunk -= skb_len; in bt_sock_stream_recvmsg()
356 if (chunk <= frag->len) { in bt_sock_stream_recvmsg()
358 skb->len -= chunk; in bt_sock_stream_recvmsg()
[all …]
/linux-4.4.14/drivers/atm/
Dfore200e.c178 fore200e_chunk_alloc(struct fore200e* fore200e, struct chunk* chunk, int size, int alignment, int d… in fore200e_chunk_alloc() argument
185 chunk->alloc_size = size + alignment; in fore200e_chunk_alloc()
186 chunk->align_size = size; in fore200e_chunk_alloc()
187 chunk->direction = direction; in fore200e_chunk_alloc()
189 chunk->alloc_addr = kzalloc(chunk->alloc_size, GFP_KERNEL | GFP_DMA); in fore200e_chunk_alloc()
190 if (chunk->alloc_addr == NULL) in fore200e_chunk_alloc()
194 offset = FORE200E_ALIGN(chunk->alloc_addr, alignment); in fore200e_chunk_alloc()
196 chunk->align_addr = chunk->alloc_addr + offset; in fore200e_chunk_alloc()
198chunk->dma_addr = fore200e->bus->dma_map(fore200e, chunk->align_addr, chunk->align_size, direction… in fore200e_chunk_alloc()
207 fore200e_chunk_free(struct fore200e* fore200e, struct chunk* chunk) in fore200e_chunk_free() argument
[all …]
Dfore200e.h563 typedef struct chunk { struct
581 struct chunk data; /* data buffer */
603 struct chunk status; /* array of completion status */
613 … struct chunk tpd; /* array of tpds */
614 … struct chunk status; /* arry of completion status */
624 …struct chunk rpd; /* array of rpds */
625 …struct chunk status; /* array of completion status */
634 …struct chunk rbd_block; /* array of rbds …
635 …struct chunk status; /* array of completion status …
811 int (*dma_chunk_alloc)(struct fore200e*, struct chunk*, int, int, int);
[all …]
/linux-4.4.14/drivers/infiniband/hw/cxgb4/
Dresource.c419 unsigned start, chunk, top; in c4iw_ocqp_pool_create() local
426 chunk = rdev->lldi.vr->ocq.size; in c4iw_ocqp_pool_create()
427 top = start + chunk; in c4iw_ocqp_pool_create()
430 chunk = min(top - start + 1, chunk); in c4iw_ocqp_pool_create()
431 if (gen_pool_add(rdev->ocqp_pool, start, chunk, -1)) { in c4iw_ocqp_pool_create()
433 __func__, start, chunk); in c4iw_ocqp_pool_create()
434 if (chunk <= 1024 << MIN_OCQP_SHIFT) { in c4iw_ocqp_pool_create()
440 chunk >>= 1; in c4iw_ocqp_pool_create()
443 __func__, start, chunk); in c4iw_ocqp_pool_create()
444 start += chunk; in c4iw_ocqp_pool_create()
/linux-4.4.14/fs/reiserfs/
Dxattr.c527 size_t chunk; in reiserfs_xattr_set_handle() local
532 chunk = PAGE_CACHE_SIZE; in reiserfs_xattr_set_handle()
534 chunk = buffer_size - buffer_pos; in reiserfs_xattr_set_handle()
549 if (chunk + skip > PAGE_CACHE_SIZE) in reiserfs_xattr_set_handle()
550 chunk = PAGE_CACHE_SIZE - skip; in reiserfs_xattr_set_handle()
557 err = __reiserfs_write_begin(page, page_offset, chunk + skip); in reiserfs_xattr_set_handle()
560 memcpy(data + skip, buffer + buffer_pos, chunk); in reiserfs_xattr_set_handle()
562 page_offset + chunk + in reiserfs_xattr_set_handle()
568 buffer_pos += chunk; in reiserfs_xattr_set_handle()
569 file_pos += chunk; in reiserfs_xattr_set_handle()
[all …]
Djournal.c676 static void write_chunk(struct buffer_chunk *chunk) in write_chunk() argument
679 for (i = 0; i < chunk->nr; i++) { in write_chunk()
680 submit_logged_buffer(chunk->bh[i]); in write_chunk()
682 chunk->nr = 0; in write_chunk()
685 static void write_ordered_chunk(struct buffer_chunk *chunk) in write_ordered_chunk() argument
688 for (i = 0; i < chunk->nr; i++) { in write_ordered_chunk()
689 submit_ordered_buffer(chunk->bh[i]); in write_ordered_chunk()
691 chunk->nr = 0; in write_ordered_chunk()
694 static int add_to_chunk(struct buffer_chunk *chunk, struct buffer_head *bh, in add_to_chunk() argument
698 BUG_ON(chunk->nr >= CHUNK_SIZE); in add_to_chunk()
[all …]
/linux-4.4.14/fs/btrfs/
Dprint-tree.c23 static void print_chunk(struct extent_buffer *eb, struct btrfs_chunk *chunk) in print_chunk() argument
25 int num_stripes = btrfs_chunk_num_stripes(eb, chunk); in print_chunk()
29 btrfs_chunk_length(eb, chunk), btrfs_chunk_owner(eb, chunk), in print_chunk()
30 btrfs_chunk_type(eb, chunk), num_stripes); in print_chunk()
33 btrfs_stripe_devid_nr(eb, chunk, i), in print_chunk()
34 btrfs_stripe_offset_nr(eb, chunk, i)); in print_chunk()
Dvolumes.c2685 struct btrfs_chunk *chunk; in btrfs_del_sys_chunk() local
2707 chunk = (struct btrfs_chunk *)(ptr + len); in btrfs_del_sys_chunk()
2708 num_stripes = btrfs_stack_chunk_num_stripes(chunk); in btrfs_del_sys_chunk()
2877 struct btrfs_chunk *chunk; in btrfs_relocate_sys_chunks() local
2915 chunk = btrfs_item_ptr(leaf, path->slots[0], in btrfs_relocate_sys_chunks()
2917 chunk_type = btrfs_chunk_type(leaf, chunk); in btrfs_relocate_sys_chunks()
3185 struct btrfs_chunk *chunk, in chunk_devid_filter() argument
3189 int num_stripes = btrfs_chunk_num_stripes(leaf, chunk); in chunk_devid_filter()
3193 stripe = btrfs_stripe_nr(chunk, i); in chunk_devid_filter()
3203 struct btrfs_chunk *chunk, in chunk_drange_filter() argument
[all …]
/linux-4.4.14/sound/usb/line6/
Dmidi.c47 unsigned char chunk[LINE6_FALLBACK_MAXPACKETSIZE]; in line6_midi_transmit() local
52 done = snd_rawmidi_transmit_peek(substream, chunk, req); in line6_midi_transmit()
57 line6_midibuf_write(mb, chunk, done); in line6_midi_transmit()
62 done = line6_midibuf_read(mb, chunk, in line6_midi_transmit()
68 send_midi_async(line6, chunk, done); in line6_midi_transmit()
/linux-4.4.14/drivers/net/wireless/ti/wlcore/
Dboot.c179 u8 *p, *chunk; in wl1271_boot_upload_firmware_chunk() local
194 chunk = kmalloc(CHUNK_SIZE, GFP_KERNEL); in wl1271_boot_upload_firmware_chunk()
195 if (!chunk) { in wl1271_boot_upload_firmware_chunk()
226 memcpy(chunk, p, CHUNK_SIZE); in wl1271_boot_upload_firmware_chunk()
229 ret = wlcore_write(wl, addr, chunk, CHUNK_SIZE, false); in wl1271_boot_upload_firmware_chunk()
239 memcpy(chunk, p, fw_data_len % CHUNK_SIZE); in wl1271_boot_upload_firmware_chunk()
242 ret = wlcore_write(wl, addr, chunk, fw_data_len % CHUNK_SIZE, false); in wl1271_boot_upload_firmware_chunk()
245 kfree(chunk); in wl1271_boot_upload_firmware_chunk()
/linux-4.4.14/Documentation/mmc/
Dmmc-async-req.txt61 request in two chunks, prepare the first chunk and start the request,
62 and finally prepare the second chunk and start the transfer.
72 * The first chunk of the request should take the same time
75 * the transfer is delayed, guesstimate max 4k as first chunk size.
84 * of the first chunk. If the MMC runs out of the first data chunk
Dmmc-dev-attrs.txt59 hence "preferred_erase_size" provides a good chunk
/linux-4.4.14/drivers/crypto/
Dn2_core.c674 struct n2_crypto_chunk chunk; member
877 struct n2_crypto_chunk *chunk; in n2_compute_chunks() local
890 chunk = &rctx->chunk; in n2_compute_chunks()
891 INIT_LIST_HEAD(&chunk->entry); in n2_compute_chunks()
893 chunk->iv_paddr = 0UL; in n2_compute_chunks()
894 chunk->arr_len = 0; in n2_compute_chunks()
895 chunk->dest_paddr = 0UL; in n2_compute_chunks()
913 if (chunk->arr_len != 0) { in n2_compute_chunks()
917 chunk->arr_len == N2_CHUNK_ARR_LEN || in n2_compute_chunks()
919 chunk->dest_final = dest_prev; in n2_compute_chunks()
[all …]
/linux-4.4.14/sound/pci/
Des1968.c1374 struct esm_memory *chunk = kmalloc(sizeof(*chunk), GFP_KERNEL); in snd_es1968_new_memory() local
1375 if (chunk == NULL) { in snd_es1968_new_memory()
1379 chunk->buf = buf->buf; in snd_es1968_new_memory()
1380 chunk->buf.bytes -= size; in snd_es1968_new_memory()
1381 chunk->buf.area += size; in snd_es1968_new_memory()
1382 chunk->buf.addr += size; in snd_es1968_new_memory()
1383 chunk->empty = 1; in snd_es1968_new_memory()
1385 list_add(&chunk->list, &buf->list); in snd_es1968_new_memory()
1395 struct esm_memory *chunk; in snd_es1968_free_memory() local
1400 chunk = list_entry(buf->list.prev, struct esm_memory, list); in snd_es1968_free_memory()
[all …]
/linux-4.4.14/arch/x86/xen/
Dp2m.c331 unsigned int i, chunk; in xen_rebuild_p2m_list() local
354 for (pfn = 0; pfn < xen_max_p2m_pfn; pfn += chunk) { in xen_rebuild_p2m_list()
365 chunk = (pfn & (P2M_PER_PAGE * P2M_MID_PER_PAGE - 1)) ? in xen_rebuild_p2m_list()
371 for (i = 1; i < chunk; i++) in xen_rebuild_p2m_list()
374 if (i < chunk) in xen_rebuild_p2m_list()
376 chunk = P2M_PER_PAGE; in xen_rebuild_p2m_list()
378 if (type == P2M_TYPE_PFN || i < chunk) { in xen_rebuild_p2m_list()
392 if (chunk == P2M_PER_PAGE) { in xen_rebuild_p2m_list()
Dsetup.c340 unsigned int i, chunk; in xen_do_set_identity_and_remap_chunk() local
351 chunk = (left < REMAP_SIZE) ? left : REMAP_SIZE; in xen_do_set_identity_and_remap_chunk()
360 xen_remap_buf.size = chunk; in xen_do_set_identity_and_remap_chunk()
361 for (i = 0; i < chunk; i++) in xen_do_set_identity_and_remap_chunk()
368 set_phys_range_identity(ident_pfn_iter, ident_pfn_iter + chunk); in xen_do_set_identity_and_remap_chunk()
370 left -= chunk; in xen_do_set_identity_and_remap_chunk()
/linux-4.4.14/drivers/net/wireless/libertas/
Dif_sdio.c338 u16 size, type, chunk; in if_sdio_card_to_host() local
362 chunk = sdio_align_size(card->func, size); in if_sdio_card_to_host()
364 ret = sdio_readsb(card->func, card->buffer, card->ioport, chunk); in if_sdio_card_to_host()
368 chunk = card->buffer[0] | (card->buffer[1] << 8); in if_sdio_card_to_host()
372 (int)type, (int)chunk); in if_sdio_card_to_host()
374 if (chunk > size) { in if_sdio_card_to_host()
376 (int)chunk, (int)size); in if_sdio_card_to_host()
381 if (chunk < size) { in if_sdio_card_to_host()
383 (int)chunk, (int)size); in if_sdio_card_to_host()
388 ret = if_sdio_handle_cmd(card, card->buffer + 4, chunk - 4); in if_sdio_card_to_host()
[all …]
/linux-4.4.14/sound/hda/
Dhdac_stream.c323 int chunk; in setup_bdle() local
333 chunk = snd_sgbuf_get_chunk_size(dmab, ofs, size); in setup_bdle()
338 if (chunk > remain) in setup_bdle()
339 chunk = remain; in setup_bdle()
341 bdl[2] = cpu_to_le32(chunk); in setup_bdle()
345 size -= chunk; in setup_bdle()
349 ofs += chunk; in setup_bdle()
/linux-4.4.14/arch/s390/crypto/
Dprng.c506 int chunk, n, tmp, ret = 0; in prng_tdes_read() local
534 chunk = min_t(int, nbytes, prng_chunk_size); in prng_tdes_read()
537 n = (chunk + 7) & -8; in prng_tdes_read()
568 if (copy_to_user(ubuf, prng_data->buf, chunk)) in prng_tdes_read()
571 nbytes -= chunk; in prng_tdes_read()
572 ret += chunk; in prng_tdes_read()
573 ubuf += chunk; in prng_tdes_read()
/linux-4.4.14/drivers/zorro/
Dzorro.c102 u32 chunk = start>>Z2RAM_CHUNKSHIFT; in mark_region() local
104 set_bit(chunk, zorro_unused_z2ram); in mark_region()
106 clear_bit(chunk, zorro_unused_z2ram); in mark_region()
/linux-4.4.14/net/rxrpc/
Dar-output.c547 size_t size, chunk, max, space; in rxrpc_send_data() local
566 chunk = max; in rxrpc_send_data()
567 if (chunk > msg_data_left(msg) && !more) in rxrpc_send_data()
568 chunk = msg_data_left(msg); in rxrpc_send_data()
570 space = chunk + call->conn->size_align; in rxrpc_send_data()
575 _debug("SIZE: %zu/%zu/%zu", chunk, space, size); in rxrpc_send_data()
594 sp->remain = chunk; in rxrpc_send_data()
/linux-4.4.14/Documentation/mtd/nand/
Dpxa3xx-nand.txt15 For instance, if we choose a 2048 data chunk and set "BCH" ECC (see below)
30 OOB, one per chunk read.
51 In the BCH mode the ECC code will be calculated for each transferred chunk
55 So, repeating the above scheme, a 2048B data chunk will be followed by 32B
/linux-4.4.14/init/
Ddo_mounts_md.c37 int chunk; member
101 md_setup_args[ent].chunk = 1 << (factor+12); in md_setup()
211 ainfo.chunk_size = md_setup_args[ent].chunk; in md_setup_drive()
/linux-4.4.14/drivers/usb/early/
Dehci-dbgp.c916 int chunk, ret; in early_dbgp_write() local
941 for (chunk = 0; chunk < DBGP_MAX_PACKET && n > 0; in early_dbgp_write()
942 str++, chunk++, n--) { in early_dbgp_write()
945 buf[chunk] = '\r'; in early_dbgp_write()
952 buf[chunk] = *str; in early_dbgp_write()
954 if (chunk > 0) { in early_dbgp_write()
956 dbgp_endpoint_out, buf, chunk); in early_dbgp_write()
/linux-4.4.14/security/keys/
Dkeyring.c233 unsigned long chunk = 0; in keyring_get_key_chunk() local
235 int desc_len = index_key->desc_len, n = sizeof(chunk); in keyring_get_key_chunk()
250 offset += sizeof(chunk) - 1; in keyring_get_key_chunk()
251 offset += (level - 3) * sizeof(chunk); in keyring_get_key_chunk()
259 chunk <<= 8; in keyring_get_key_chunk()
260 chunk |= ((u8*)index_key->description)[--offset]; in keyring_get_key_chunk()
264 chunk <<= 8; in keyring_get_key_chunk()
265 chunk |= (u8)((unsigned long)index_key->type >> in keyring_get_key_chunk()
268 return chunk; in keyring_get_key_chunk()
/linux-4.4.14/drivers/input/touchscreen/
Dwdt87xx_i2c.c686 static int wdt87xx_write_firmware(struct i2c_client *client, const void *chunk) in wdt87xx_write_firmware() argument
688 u32 start_addr = get_unaligned_le32(chunk + FW_CHUNK_TGT_START_OFFSET); in wdt87xx_write_firmware()
689 u32 size = get_unaligned_le32(chunk + FW_CHUNK_PAYLOAD_LEN_OFFSET); in wdt87xx_write_firmware()
690 const void *data = chunk + FW_CHUNK_PAYLOAD_OFFSET; in wdt87xx_write_firmware()
788 const void *chunk; in wdt87xx_load_chunk() local
791 chunk = wdt87xx_get_fw_chunk(fw, ck_id); in wdt87xx_load_chunk()
792 if (!chunk) { in wdt87xx_load_chunk()
798 error = wdt87xx_validate_fw_chunk(chunk, ck_id); in wdt87xx_load_chunk()
805 error = wdt87xx_write_firmware(client, chunk); in wdt87xx_load_chunk()
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/
Damdgpu_cs.c625 struct amdgpu_cs_chunk *chunk; in amdgpu_cs_ib_fill() local
630 chunk = &parser->chunks[i]; in amdgpu_cs_ib_fill()
632 chunk_ib = (struct drm_amdgpu_cs_chunk_ib *)chunk->kdata; in amdgpu_cs_ib_fill()
634 if (chunk->chunk_id != AMDGPU_CHUNK_ID_IB) in amdgpu_cs_ib_fill()
747 struct amdgpu_cs_chunk *chunk; in amdgpu_cs_dependencies() local
750 chunk = &p->chunks[i]; in amdgpu_cs_dependencies()
752 if (chunk->chunk_id != AMDGPU_CHUNK_ID_DEPENDENCIES) in amdgpu_cs_dependencies()
755 deps = (struct drm_amdgpu_cs_chunk_dep *)chunk->kdata; in amdgpu_cs_dependencies()
756 num_deps = chunk->length_dw * 4 / in amdgpu_cs_dependencies()
/linux-4.4.14/drivers/usb/
Dusb-skeleton.c289 size_t chunk = min(available, count); in skel_read() local
309 chunk)) in skel_read()
312 rv = chunk; in skel_read()
314 dev->bulk_in_copied += chunk; in skel_read()
321 skel_do_read_io(dev, count - chunk); in skel_read()
/linux-4.4.14/drivers/dma/
Dat_xdmac.c853 struct data_chunk *chunk) in at_xdmac_interleaved_queue_desc() argument
873 dwidth = at_xdmac_align_width(chan, src | dst | chunk->size); in at_xdmac_interleaved_queue_desc()
874 if (chunk->size >= (AT_XDMAC_MBR_UBC_UBLEN_MAX << dwidth)) { in at_xdmac_interleaved_queue_desc()
877 __func__, chunk->size, in at_xdmac_interleaved_queue_desc()
910 ublen = chunk->size >> dwidth; in at_xdmac_interleaved_queue_desc()
914 desc->lld.mbr_sus = dmaengine_get_src_icg(xt, chunk); in at_xdmac_interleaved_queue_desc()
915 desc->lld.mbr_dus = dmaengine_get_dst_icg(xt, chunk); in at_xdmac_interleaved_queue_desc()
944 struct data_chunk *chunk; in at_xdmac_prep_interleaved() local
982 chunk = xt->sgl + i; in at_xdmac_prep_interleaved()
984 dst_icg = dmaengine_get_dst_icg(xt, chunk); in at_xdmac_prep_interleaved()
[all …]
Dat_hdmac.c743 struct data_chunk *chunk = xt->sgl + i; in atc_prep_dma_interleaved() local
745 if ((chunk->size != xt->sgl->size) || in atc_prep_dma_interleaved()
746 (dmaengine_get_dst_icg(xt, chunk) != dmaengine_get_dst_icg(xt, first)) || in atc_prep_dma_interleaved()
747 (dmaengine_get_src_icg(xt, chunk) != dmaengine_get_src_icg(xt, first))) { in atc_prep_dma_interleaved()
754 len += chunk->size; in atc_prep_dma_interleaved()
/linux-4.4.14/fs/nfs/blocklayout/
Ddev.c158 u64 chunk; in bl_map_stripe() local
162 chunk = div_u64(offset, dev->chunk_size); in bl_map_stripe()
163 div_u64_rem(chunk, dev->nr_children, &chunk_idx); in bl_map_stripe()
173 offset = chunk * dev->chunk_size; in bl_map_stripe()
/linux-4.4.14/drivers/net/wireless/ath/ar5523/
Dar5523.c528 struct ar5523_chunk *chunk; in ar5523_data_rx_cb() local
549 chunk = (struct ar5523_chunk *) data->skb->data; in ar5523_data_rx_cb()
551 if (((chunk->flags & UATH_CFLAGS_FINAL) == 0) || in ar5523_data_rx_cb()
552 chunk->seqnum != 0) { in ar5523_data_rx_cb()
554 chunk->seqnum, chunk->flags, in ar5523_data_rx_cb()
555 be16_to_cpu(chunk->length)); in ar5523_data_rx_cb()
581 skb_reserve(data->skb, sizeof(*chunk)); in ar5523_data_rx_cb()
793 struct ar5523_chunk *chunk; in ar5523_tx_work_locked() local
836 chunk = (struct ar5523_chunk *)skb_push(skb, sizeof(*chunk)); in ar5523_tx_work_locked()
838 chunk->seqnum = 0; in ar5523_tx_work_locked()
[all …]
/linux-4.4.14/net/core/
Ddatagram.c710 int chunk = skb->len - hlen; in skb_copy_and_csum_datagram_msg() local
712 if (!chunk) in skb_copy_and_csum_datagram_msg()
715 if (msg_data_left(msg) < chunk) { in skb_copy_and_csum_datagram_msg()
718 if (skb_copy_datagram_msg(skb, hlen, msg, chunk)) in skb_copy_and_csum_datagram_msg()
723 chunk, &csum)) in skb_copy_and_csum_datagram_msg()
Dskbuff.c4464 unsigned long chunk; in alloc_skb_with_frags() local
4510 chunk = min_t(unsigned long, data_len, in alloc_skb_with_frags()
4512 skb_fill_page_desc(skb, i, page, 0, chunk); in alloc_skb_with_frags()
4513 data_len -= chunk; in alloc_skb_with_frags()
/linux-4.4.14/Documentation/networking/
Dnetconsole.txt162 For example, assuming a lot smaller chunk size, a message "the first
163 chunk, the 2nd chunk." may be split as follows.
165 6,416,1758426,-,ncfrag=0/31;the first chunk,
166 6,416,1758426,-,ncfrag=16/31; the 2nd chunk.
/linux-4.4.14/include/linux/
Ddmaengine.h971 struct data_chunk *chunk) in dmaengine_get_dst_icg() argument
974 chunk->icg, chunk->dst_icg); in dmaengine_get_dst_icg()
978 struct data_chunk *chunk) in dmaengine_get_src_icg() argument
981 chunk->icg, chunk->src_icg); in dmaengine_get_src_icg()
/linux-4.4.14/drivers/nfc/pn544/
Di2c.c677 struct pn544_i2c_fw_secure_frame *chunk; in pn544_hci_i2c_fw_secure_write_frame_cmd() local
684 chunk = (struct pn544_i2c_fw_secure_frame *) buf; in pn544_hci_i2c_fw_secure_write_frame_cmd()
686 chunk->cmd = PN544_FW_CMD_SECURE_CHUNK_WRITE; in pn544_hci_i2c_fw_secure_write_frame_cmd()
688 put_unaligned_be16(datalen, &chunk->be_datalen); in pn544_hci_i2c_fw_secure_write_frame_cmd()
690 memcpy(chunk->data, data, datalen); in pn544_hci_i2c_fw_secure_write_frame_cmd()
692 chunklen = sizeof(chunk->cmd) + sizeof(chunk->be_datalen) + datalen; in pn544_hci_i2c_fw_secure_write_frame_cmd()
/linux-4.4.14/drivers/net/ethernet/sfc/
Dmcdi.c1952 size_t chunk; in efx_mcdi_mtd_read() local
1956 chunk = min_t(size_t, end - offset, EFX_MCDI_NVRAM_LEN_MAX); in efx_mcdi_mtd_read()
1958 buffer, chunk); in efx_mcdi_mtd_read()
1961 offset += chunk; in efx_mcdi_mtd_read()
1962 buffer += chunk; in efx_mcdi_mtd_read()
1975 size_t chunk = part->common.mtd.erasesize; in efx_mcdi_mtd_erase() local
1990 chunk); in efx_mcdi_mtd_erase()
1993 offset += chunk; in efx_mcdi_mtd_erase()
2006 size_t chunk; in efx_mcdi_mtd_write() local
2017 chunk = min_t(size_t, end - offset, EFX_MCDI_NVRAM_LEN_MAX); in efx_mcdi_mtd_write()
[all …]
/linux-4.4.14/net/ipv4/
Dtcp.c1772 int chunk; in tcp_recvmsg() local
1776 chunk = len - tp->ucopy.len; in tcp_recvmsg()
1777 if (chunk != 0) { in tcp_recvmsg()
1778 NET_ADD_STATS_USER(sock_net(sk), LINUX_MIB_TCPDIRECTCOPYFROMBACKLOG, chunk); in tcp_recvmsg()
1779 len -= chunk; in tcp_recvmsg()
1780 copied += chunk; in tcp_recvmsg()
1788 chunk = len - tp->ucopy.len; in tcp_recvmsg()
1789 if (chunk != 0) { in tcp_recvmsg()
1790 NET_ADD_STATS_USER(sock_net(sk), LINUX_MIB_TCPDIRECTCOPYFROMPREQUEUE, chunk); in tcp_recvmsg()
1791 len -= chunk; in tcp_recvmsg()
[all …]
Dtcp_input.c4562 int chunk = min_t(unsigned int, skb->len, in tcp_data_queue() local
4568 if (!skb_copy_datagram_msg(skb, 0, tp->ucopy.msg, chunk)) { in tcp_data_queue()
4569 tp->ucopy.len -= chunk; in tcp_data_queue()
4570 tp->copied_seq += chunk; in tcp_data_queue()
4571 eaten = (chunk == skb->len); in tcp_data_queue()
5087 int chunk = skb->len - hlen; in tcp_copy_to_iovec() local
5092 err = skb_copy_datagram_msg(skb, hlen, tp->ucopy.msg, chunk); in tcp_copy_to_iovec()
5097 tp->ucopy.len -= chunk; in tcp_copy_to_iovec()
5098 tp->copied_seq += chunk; in tcp_copy_to_iovec()
Dtcp_output.c1798 u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); in tcp_tso_should_defer() local
1803 chunk /= win_divisor; in tcp_tso_should_defer()
1804 if (limit >= chunk) in tcp_tso_should_defer()
/linux-4.4.14/sound/pci/lola/
Dlola_pcm.c326 int chunk; in setup_bdle() local
336 chunk = snd_pcm_sgbuf_get_chunk_size(substream, ofs, size); in setup_bdle()
337 bdl[2] = cpu_to_le32(chunk); in setup_bdle()
341 size -= chunk; in setup_bdle()
345 ofs += chunk; in setup_bdle()
/linux-4.4.14/drivers/infiniband/hw/qib/
Dqib_init.c1672 unsigned e, egrcnt, egrperchunk, chunk, egrsize, egroff; in qib_setup_eagerbufs() local
1689 chunk = rcd->rcvegrbuf_chunks; in qib_setup_eagerbufs()
1694 kzalloc_node(chunk * sizeof(rcd->rcvegrbuf[0]), in qib_setup_eagerbufs()
1701 kmalloc_node(chunk * sizeof(rcd->rcvegrbuf_phys[0]), in qib_setup_eagerbufs()
1723 for (e = chunk = 0; chunk < rcd->rcvegrbuf_chunks; chunk++) { in qib_setup_eagerbufs()
1724 dma_addr_t pa = rcd->rcvegrbuf_phys[chunk]; in qib_setup_eagerbufs()
1728 memset(rcd->rcvegrbuf[chunk], 0, size); in qib_setup_eagerbufs()
Dqib_driver.c286 const u32 chunk = etail >> rcd->rcvegrbufs_perchunk_shift; in qib_get_egrbuf() local
289 return rcd->rcvegrbuf[chunk] + (idx << rcd->dd->rcvegrbufsize_shift); in qib_get_egrbuf()
/linux-4.4.14/net/unix/
Daf_unix.c2304 int chunk; in unix_stream_read_generic() local
2385 chunk = min_t(unsigned int, unix_skb_len(skb) - skip, size); in unix_stream_read_generic()
2387 chunk = state->recv_actor(skb, skip, chunk, state); in unix_stream_read_generic()
2391 if (chunk < 0) { in unix_stream_read_generic()
2396 copied += chunk; in unix_stream_read_generic()
2397 size -= chunk; in unix_stream_read_generic()
2413 UNIXCB(skb).consumed += chunk; in unix_stream_read_generic()
2415 sk_peek_offset_bwd(sk, chunk); in unix_stream_read_generic()
2434 sk_peek_offset_fwd(sk, chunk); in unix_stream_read_generic()
2461 int skip, int chunk, in unix_stream_read_actor() argument
[all …]
/linux-4.4.14/drivers/char/
Dmem.c639 size_t chunk = iov_iter_count(iter), n; in read_iter_zero() local
641 if (chunk > PAGE_SIZE) in read_iter_zero()
642 chunk = PAGE_SIZE; /* Just for latency reasons */ in read_iter_zero()
643 n = iov_iter_zero(chunk, iter); in read_iter_zero()
Drandom.c1342 int chunk = min(nbytes, (int)sizeof(unsigned long)); in get_random_bytes_arch() local
1347 memcpy(p, &v, chunk); in get_random_bytes_arch()
1348 p += chunk; in get_random_bytes_arch()
1349 nbytes -= chunk; in get_random_bytes_arch()
/linux-4.4.14/net/caif/
Dcaif_socket.c373 int chunk; in caif_stream_recvmsg() local
424 chunk = min_t(unsigned int, skb->len, size); in caif_stream_recvmsg()
425 if (memcpy_to_msg(msg, skb->data, chunk)) { in caif_stream_recvmsg()
431 copied += chunk; in caif_stream_recvmsg()
432 size -= chunk; in caif_stream_recvmsg()
436 skb_pull(skb, chunk); in caif_stream_recvmsg()
/linux-4.4.14/Documentation/fb/
Dsh7760fb.txt11 * Framebuffer memory must be a large chunk allocated at the top
14 get a large enough contiguous chunk of memory.
/linux-4.4.14/drivers/irqchip/
Dirq-gic-v3-its.c663 static int its_chunk_to_lpi(int chunk) in its_chunk_to_lpi() argument
665 return (chunk << IRQS_PER_CHUNK_SHIFT) + 8192; in its_chunk_to_lpi()
735 int chunk = its_lpi_to_chunk(lpi); in its_lpi_free() local
736 BUG_ON(chunk > lpi_chunks); in its_lpi_free()
737 if (test_bit(chunk, lpi_bitmap)) { in its_lpi_free()
738 clear_bit(chunk, lpi_bitmap); in its_lpi_free()
740 pr_err("Bad LPI chunk %d\n", chunk); in its_lpi_free()
/linux-4.4.14/Documentation/
Ddell_rbu.txt20 image methods. In case of monolithic the driver allocates a contiguous chunk
57 The user creates packets header, gets the chunk of the BIOS image and
58 places it next to the packetheader; now, the packetheader + BIOS image chunk
Dassoc_array.txt139 (1) Get a chunk of index key from caller data:
143 This should return a chunk of caller-supplied index key starting at the
149 (2) Get a chunk of an object's index key.
360 The index key is read in chunks of machine word. Each chunk is subdivided into
Ddma-buf-sharing.txt278 PAGE_SIZE size. Before accessing a chunk it needs to be mapped, which returns
279 a pointer in kernel virtual address space. Afterwards the chunk needs to be
280 unmapped again. There is no limit on how often a given chunk can be mapped
282 before mapping the same chunk again.
DDMA-attributes.txt100 be mapped as contiguous chunk into device dma address space. By
Dramoops.txt32 power of two) and each oops/panic writes a "record_size" chunk of
Dmd.txt12 md=<md device no.>,<raid level>,<chunk size factor>,<fault level>,dev0,dev1,...,devn
30 chunk size factor = (raid-0 and raid-1 only)
31 Set the chunk size as 4k << n.
309 The size, in bytes, of the chunk which will be represented by a
Dxillybus.txt87 their attributes), there isn't one specific chunk of logic being the Xillybus
/linux-4.4.14/drivers/staging/media/lirc/
Dlirc_serial.c385 unsigned char chunk, shifted; in send_pulse_irdeo() local
390 chunk = 3; in send_pulse_irdeo()
392 chunk = 1; in send_pulse_irdeo()
394 shifted = chunk << (i * 3); in send_pulse_irdeo()
/linux-4.4.14/include/drm/
Ddrm_dp_mst_helper.h206 u8 chunk[48]; member
361 u8 chunk[48]; member
/linux-4.4.14/drivers/tty/
Dtty_io.c1111 unsigned int chunk; in do_tty_write() local
1133 chunk = 2048; in do_tty_write()
1135 chunk = 65536; in do_tty_write()
1136 if (count < chunk) in do_tty_write()
1137 chunk = count; in do_tty_write()
1140 if (tty->write_cnt < chunk) { in do_tty_write()
1143 if (chunk < 1024) in do_tty_write()
1144 chunk = 1024; in do_tty_write()
1146 buf_chunk = kmalloc(chunk, GFP_KERNEL); in do_tty_write()
1152 tty->write_cnt = chunk; in do_tty_write()
[all …]
/linux-4.4.14/arch/xtensa/lib/
Dmemcopy.S151 add a8, a8, a3 # a8 = end of last 16B source chunk
227 add a10, a10, a3 # a10 = end of last 16B source chunk
425 sub a8, a3, a8 # a8 = start of first 16B source chunk
502 sub a10, a3, a10 # a10 = start of first 16B source chunk
Dusercopy.S161 add a8, a8, a3 # a8 = end of last 16B source chunk
226 add a12, a12, a3 # a12 = end of last 16B source chunk
Dmemset.S73 add a6, a6, a5 # a6 = end of last 16B chunk
Dstrnlen_user.S77 add a10, a10, a4 # a10 = end of last 4B chunk
/linux-4.4.14/fs/gfs2/
Dbmap.c1013 u64 chunk; in gfs2_journaled_truncate() local
1017 chunk = oldsize - newsize; in gfs2_journaled_truncate()
1018 if (chunk > max_chunk) in gfs2_journaled_truncate()
1019 chunk = max_chunk; in gfs2_journaled_truncate()
1020 truncate_pagecache(inode, oldsize - chunk); in gfs2_journaled_truncate()
1021 oldsize -= chunk; in gfs2_journaled_truncate()
/linux-4.4.14/arch/arm/kernel/
Dtraps.c540 unsigned long chunk = min(PAGE_SIZE, end - start); in __do_cache_op() local
545 ret = flush_cache_user_range(start, start + chunk); in __do_cache_op()
550 start += chunk; in __do_cache_op()
/linux-4.4.14/drivers/memstick/core/
Dmspro_block.c697 static int mspro_block_issue_req(struct memstick_dev *card, int chunk) in mspro_block_issue_req() argument
704 while (chunk) { in mspro_block_issue_req()
712 chunk = __blk_end_request_cur(msb->block_req, -ENOMEM); in mspro_block_issue_req()
739 chunk = 1; in mspro_block_issue_req()
746 int chunk, cnt; in mspro_block_complete_req() local
779 chunk = __blk_end_request(msb->block_req, error, t_len); in mspro_block_complete_req()
781 error = mspro_block_issue_req(card, chunk); in mspro_block_complete_req()
/linux-4.4.14/drivers/gpu/drm/radeon/
Dradeon_cs.c78 struct radeon_cs_chunk *chunk; in radeon_cs_parser_relocs() local
87 chunk = p->chunk_relocs; in radeon_cs_parser_relocs()
90 p->nrelocs = chunk->length_dw / 4; in radeon_cs_parser_relocs()
103 r = (struct drm_radeon_cs_reloc *)&chunk->kdata[i*4]; in radeon_cs_parser_relocs()
/linux-4.4.14/drivers/thermal/ti-soc-thermal/
Dti-bandgap.c1236 void __iomem *chunk; in ti_bandgap_build() local
1241 chunk = devm_ioremap_resource(&pdev->dev, res); in ti_bandgap_build()
1243 bgp->base = chunk; in ti_bandgap_build()
1244 if (IS_ERR(chunk)) in ti_bandgap_build()
1245 return ERR_CAST(chunk); in ti_bandgap_build()
/linux-4.4.14/drivers/staging/lustre/lustre/osc/
Dosc_cache.c633 pgoff_t chunk; in osc_extent_find() local
653 chunk = index >> ppc_bits; in osc_extent_find()
688 if (chunk > ext_chk_end + 1) in osc_extent_find()
701 if (chunk + 1 < ext_chk_start) { in osc_extent_find()
754 if (chunk + 1 == ext_chk_start) { in osc_extent_find()
764 } else if (chunk == ext_chk_end + 1) { in osc_extent_find()
1145 pgoff_t chunk = index >> ppc_bits; in osc_extent_expand() local
1155 if (chunk > end_chunk + 1) { in osc_extent_expand()
1160 if (end_chunk >= chunk) { in osc_extent_expand()
1165 LASSERT(end_chunk + 1 == chunk); in osc_extent_expand()
[all …]
/linux-4.4.14/arch/arm/lib/
Dmemzero.S42 cmp r1, #16 @ 1 we can skip this chunk if we
/linux-4.4.14/drivers/mmc/host/
Dsdhci.c305 size_t blksize, len, chunk; in sdhci_read_block_pio() local
312 chunk = 0; in sdhci_read_block_pio()
327 if (chunk == 0) { in sdhci_read_block_pio()
329 chunk = 4; in sdhci_read_block_pio()
336 chunk--; in sdhci_read_block_pio()
349 size_t blksize, len, chunk; in sdhci_write_block_pio() local
356 chunk = 0; in sdhci_write_block_pio()
372 scratch |= (u32)*buf << (chunk * 8); in sdhci_write_block_pio()
375 chunk++; in sdhci_write_block_pio()
378 if ((chunk == 4) || ((len == 0) && (blksize == 0))) { in sdhci_write_block_pio()
[all …]
/linux-4.4.14/drivers/macintosh/
Dsmu.c899 unsigned int chunk; in smu_read_datablock() local
907 chunk = 0xe; in smu_read_datablock()
910 unsigned int clen = min(len, chunk); in smu_read_datablock()
915 cmd.reply_len = chunk; in smu_read_datablock()
/linux-4.4.14/arch/powerpc/kernel/
Dvio.c393 size_t avail = 0, level, chunk, need; in vio_cmo_balance() local
441 chunk = min_t(size_t, avail, VIO_CMO_BALANCE_CHUNK); in vio_cmo_balance()
442 chunk = min(chunk, (viodev->cmo.desired - in vio_cmo_balance()
444 viodev->cmo.entitled += chunk; in vio_cmo_balance()
Dprom_init.c2016 unsigned long room, chunk; in make_room() local
2026 chunk = alloc_up(room, 0); in make_room()
2027 if (chunk == 0) in make_room()
2030 *mem_end = chunk + room; in make_room()
/linux-4.4.14/drivers/net/wimax/i2400m/
Dfw.c646 static int i2400m_download_chunk(struct i2400m *i2400m, const void *chunk, in i2400m_download_chunk() argument
660 "direct %u do_csum %u)\n", i2400m, chunk, __chunk_len, in i2400m_download_chunk()
663 memcpy(buf->cmd_payload, chunk, __chunk_len); in i2400m_download_chunk()
676 "direct %u do_csum %u) = %d\n", i2400m, chunk, __chunk_len, in i2400m_download_chunk()
/linux-4.4.14/drivers/crypto/qat/qat_common/
Dqat_uclo.c501 char *chunk; in qat_uclo_map_chunk() local
509 chunk = buf + file_chunk->offset; in qat_uclo_map_chunk()
511 chunk, file_chunk->size)) in qat_uclo_map_chunk()
516 obj_hdr->file_buff = chunk; in qat_uclo_map_chunk()
/linux-4.4.14/drivers/staging/rdma/ipath/
Dipath_file_ops.c897 unsigned e, egrcnt, egrperchunk, chunk, egrsize, egroff; in ipath_create_user_egr() local
917 chunk = pd->port_rcvegrbuf_chunks; in ipath_create_user_egr()
920 pd->port_rcvegrbuf = kmalloc(chunk * sizeof(pd->port_rcvegrbuf[0]), in ipath_create_user_egr()
927 kmalloc(chunk * sizeof(pd->port_rcvegrbuf_phys[0]), in ipath_create_user_egr()
947 for (e = chunk = 0; chunk < pd->port_rcvegrbuf_chunks; chunk++) { in ipath_create_user_egr()
948 dma_addr_t pa = pd->port_rcvegrbuf_phys[chunk]; in ipath_create_user_egr()
/linux-4.4.14/drivers/media/dvb-frontends/
Ddrxk_hard.h224 u8 chunk[256]; member
Ddrxk_hard.c449 int chunk = blk_size > state->m_chunk_size ? in write_block() local
451 u8 *adr_buf = &state->chunk[0]; in write_block()
461 if (chunk == state->m_chunk_size) in write_block()
462 chunk -= 2; in write_block()
469 memcpy(&state->chunk[adr_length], p_block, chunk); in write_block()
474 for (i = 0; i < chunk; i++) in write_block()
479 &state->chunk[0], chunk + adr_length); in write_block()
485 p_block += chunk; in write_block()
486 address += (chunk >> 1); in write_block()
487 blk_size -= chunk; in write_block()
/linux-4.4.14/drivers/input/misc/
Dims-pcu.c515 u8 command, int chunk, int len) in ims_pcu_send_cmd_chunk() argument
527 command, chunk, error); in ims_pcu_send_cmd_chunk()
538 int chunk = 0; in ims_pcu_send_command() local
566 ++chunk, count); in ims_pcu_send_command()
584 error = ims_pcu_send_cmd_chunk(pcu, command, ++chunk, count); in ims_pcu_send_command()
597 return ims_pcu_send_cmd_chunk(pcu, command, ++chunk, count); in ims_pcu_send_command()
/linux-4.4.14/net/irda/
Daf_irda.c1430 int chunk; in irda_recvmsg_stream() local
1470 chunk = min_t(unsigned int, skb->len, size); in irda_recvmsg_stream()
1471 if (memcpy_to_msg(msg, skb->data, chunk)) { in irda_recvmsg_stream()
1477 copied += chunk; in irda_recvmsg_stream()
1478 size -= chunk; in irda_recvmsg_stream()
1482 skb_pull(skb, chunk); in irda_recvmsg_stream()
/linux-4.4.14/net/sunrpc/xprtrdma/
Dsvc_rdma_sendto.c196 int chunk = be32_to_cpu(wr_ary->wc_nchunks); in svc_rdma_get_reply_array() local
199 &wr_ary->wc_array[chunk].wc_target.rs_length; in svc_rdma_get_reply_array()
/linux-4.4.14/net/decnet/
Daf_decnet.c1760 unsigned int chunk = skb->len; in dn_recvmsg() local
1763 if ((chunk + copied) > size) in dn_recvmsg()
1764 chunk = size - copied; in dn_recvmsg()
1766 if (memcpy_to_msg(msg, skb->data, chunk)) { in dn_recvmsg()
1770 copied += chunk; in dn_recvmsg()
1773 skb_pull(skb, chunk); in dn_recvmsg()
/linux-4.4.14/drivers/gpu/drm/
Ddrm_dp_mst_topology.c350 memcpy(&msg->chunk[0], replybuf + hdrlen, msg->curchunk_idx); in drm_dp_sideband_msg_build()
352 memcpy(&msg->chunk[msg->curchunk_idx], replybuf, replybuflen); in drm_dp_sideband_msg_build()
358 crc4 = drm_dp_msg_data_crc4(msg->chunk, msg->curchunk_len - 1); in drm_dp_sideband_msg_build()
360 memcpy(&msg->msg[msg->curlen], msg->chunk, msg->curchunk_len - 1); in drm_dp_sideband_msg_build()
1431 u8 chunk[48]; in process_single_tx_qlock() local
1463 drm_dp_encode_sideband_msg_hdr(&hdr, chunk, &idx); in process_single_tx_qlock()
1464 memcpy(&chunk[idx], &txmsg->msg[txmsg->cur_offset], tosend); in process_single_tx_qlock()
1466 drm_dp_crc_sideband_chunk_req(&chunk[idx], tosend); in process_single_tx_qlock()
1469 ret = drm_dp_send_sideband_msg(mgr, up, chunk, idx); in process_single_tx_qlock()
/linux-4.4.14/drivers/net/ethernet/ibm/emac/
Dcore.c1421 int chunk = min(len, MAL_MAX_TX_SIZE); in emac_xmit_split() local
1422 len -= chunk; in emac_xmit_split()
1433 dev->tx_desc[slot].data_len = (u16) chunk; in emac_xmit_split()
1440 pd += chunk; in emac_xmit_split()
1450 int len = skb->len, chunk; in emac_start_xmit_sg() local
1474 chunk = min(len, MAL_MAX_TX_SIZE); in emac_start_xmit_sg()
1477 dev->tx_desc[slot].data_len = (u16) chunk; in emac_start_xmit_sg()
1478 len -= chunk; in emac_start_xmit_sg()
1480 slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags, in emac_start_xmit_sg()
/linux-4.4.14/security/apparmor/
Dpolicy_unpack.c124 static size_t unpack_u16_chunk(struct aa_ext *e, char **chunk) in unpack_u16_chunk() argument
134 *chunk = e->pos; in unpack_u16_chunk()
/linux-4.4.14/crypto/
Ddrbg.c1433 unsigned int chunk = 0; in drbg_generate_long() local
1435 chunk = slice ? drbg_max_request_bytes(drbg) : (buflen - len); in drbg_generate_long()
1437 err = drbg_generate(drbg, buf + len, chunk, addtl); in drbg_generate_long()
1441 len += chunk; in drbg_generate_long()
/linux-4.4.14/Documentation/dmaengine/
Dprovider.txt69 but all of them will require, for every chunk, at least the source and
188 scatter-gather transfer, with a single chunk to transfer, it's a
204 deal with a single chunk to copy or a collection of them, here,
212 - A cyclic transfer is a transfer where the chunk collection will
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/socklnd/
Dsocklnd_cb.c2581 int chunk = ksocknal_data.ksnd_peer_hash_size; in ksocknal_reaper() local
2591 chunk = (chunk * n * p) / in ksocknal_reaper()
2593 if (chunk == 0) in ksocknal_reaper()
2594 chunk = 1; in ksocknal_reaper()
2596 for (i = 0; i < chunk; i++) { in ksocknal_reaper()
/linux-4.4.14/Documentation/DocBook/
Dmtdnand.xml.db14 API-nand-check-erased-ecc-chunk
Dalsa-driver-api.xml.db103 API-snd-pcm-sgbuf-get-chunk-size
Dkernel-api.xml.db658 API-blk-queue-chunk-sectors
/linux-4.4.14/Documentation/arm/Samsung-S3C24XX/
DSuspend.txt128 Defines the size of memory each CRC chunk covers. A smaller value
/linux-4.4.14/Documentation/ko_KR/
DHOWTO491 리눅스 커널 커뮤니티는 한꺼번에 굉장히 큰 코드의 묶음(chunk)을 쉽게
/linux-4.4.14/Documentation/filesystems/cifs/
DTODO14 - T10 copy offload (copy chunk is only mechanism supported)
/linux-4.4.14/Documentation/ABI/testing/
Dsysfs-driver-wacom72 byte chunk encodes the image data for two consecutive lines on
Dsysfs-block129 chunk size. A properly aligned multiple of
/linux-4.4.14/Documentation/filesystems/
Dntfs.txt318 For a striped volume, i.e. raid level 0, you will need to know the chunk size
394 Windows by default uses a stripe chunk size of 64k, so you probably want the
395 "chunk-size 64k" option for each raid-disk, too.
405 chunk-size 64k
Dsysv-fs.txt84 to the free blocks pertaining to the next chunk; the first of these
Dbtrfs.txt141 Specify that 1 metadata chunk should be allocated after every <value>
Dext4.txt264 disks * RAID chunk size in file system blocks.
/linux-4.4.14/Documentation/fpga/
Dfpga-mgr.txt163 whole FPGA image or may be a smaller chunk of an FPGA image. In the latter
/linux-4.4.14/drivers/acpi/apei/
Dghes.c202 struct gen_pool_chunk *chunk, in ghes_estatus_pool_free_chunk_page() argument
205 free_page(chunk->start_addr); in ghes_estatus_pool_free_chunk_page()
/linux-4.4.14/drivers/mtd/nand/
Dnand_base.c1907 int chunk = chip->ecc.bytes + chip->ecc.prepad + chip->ecc.postpad; in nand_read_oob_syndrome() local
1915 pos = eccsize + i * (eccsize + chunk); in nand_read_oob_syndrome()
1922 toread = min_t(int, length, chunk); in nand_read_oob_syndrome()
1966 int chunk = chip->ecc.bytes + chip->ecc.prepad + chip->ecc.postpad; in nand_write_oob_syndrome() local
1977 pos = steps * (eccsize + chunk); in nand_write_oob_syndrome()
1996 pos = eccsize + i * (eccsize + chunk); in nand_write_oob_syndrome()
2001 len = min_t(int, length, chunk); in nand_write_oob_syndrome()
/linux-4.4.14/arch/sh/lib64/
Dcopy_user_memcpy.S28 ! bytes memory chunk to b copied, the rest of the word can be read
Dmemcpy.S30 ! bytes memory chunk to b copied, the rest of the word can be read
/linux-4.4.14/arch/m68k/
DKconfig.cpu361 bool "Use one physical chunk of memory only" if ADVANCED && !SUN3
366 Ignore all but the first contiguous chunk of physical memory for VM
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c3134 int chunk = kiblnd_data.kib_peer_hash_size; in kiblnd_connd() local
3148 chunk = (chunk * n * p) / in kiblnd_connd()
3150 if (chunk == 0) in kiblnd_connd()
3151 chunk = 1; in kiblnd_connd()
3153 for (i = 0; i < chunk; i++) { in kiblnd_connd()
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb3/
Dcxgb3_main.c2455 unsigned int chunk = in cxgb_extension_ioctl() local
2459 t3_mc7_bd_read(mem, t.addr / 8, chunk / 8, in cxgb_extension_ioctl()
2463 if (copy_to_user(useraddr, buf, chunk)) in cxgb_extension_ioctl()
2465 useraddr += chunk; in cxgb_extension_ioctl()
2466 t.addr += chunk; in cxgb_extension_ioctl()
2467 t.len -= chunk; in cxgb_extension_ioctl()
Dsge.c326 unsigned int chunk) in reclaim_completed_tx() argument
330 reclaim = min(chunk, reclaim); in reclaim_completed_tx()
/linux-4.4.14/fs/cramfs/
DREADME58 applied to the i'th blksize-sized chunk of the input data.
/linux-4.4.14/arch/s390/kernel/
Dhead.S96 ssch 0(%r3) # load chunk of 1600 bytes
/linux-4.4.14/Documentation/early-userspace/
DREADME17 - initramfs, a chunk of code that unpacks the compressed cpio image
/linux-4.4.14/drivers/net/ethernet/marvell/
Dskge.c2530 u32 chunk, ram_addr; in skge_up() local
2595 chunk = (hw->ram_size - hw->ram_offset) / (hw->ports * 2); in skge_up()
2596 ram_addr = hw->ram_offset + 2 * chunk * port; in skge_up()
2598 skge_ramset(hw, rxqaddr[port], ram_addr, chunk); in skge_up()
2602 skge_ramset(hw, txqaddr[port], ram_addr+chunk, chunk); in skge_up()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/ce/fuc/
Dcom.fuc208 // get a chunk of stack space, aligned to 256 byte boundary
419 // zero out a chunk of the stack to store the swizzle into
/linux-4.4.14/arch/cris/arch-v10/
DREADME.mm228 happens first is that a virtual address chunk is allocated to the request using
/linux-4.4.14/drivers/scsi/
Dips.h733 IPS_CHUNK chunk[IPS_MAX_CHUNKS]; member
/linux-4.4.14/Documentation/powerpc/
Dfirmware-assisted-dump.txt53 NOTE: The term 'boot memory' means size of the low memory chunk
/linux-4.4.14/fs/ext4/
Dmballoc.c671 ext4_grpblk_t chunk; in ext4_mb_mark_free_simple() local
687 chunk = 1 << min; in ext4_mb_mark_free_simple()
695 len -= chunk; in ext4_mb_mark_free_simple()
696 first += chunk; in ext4_mb_mark_free_simple()
/linux-4.4.14/Documentation/scsi/
Dst.txt274 (PAGE_SIZE << ST_FIRST_ORDER) bytes if the system can give a chunk of
279 buffer as one chunk.
/linux-4.4.14/Documentation/usb/
Dusbmon.txt263 The size may be rounded down to the next chunk (or page). If the requested
DWUSB-Design-overview.txt372 we issue another URB to read into the destination buffer the chunk of
/linux-4.4.14/scripts/
Dcheckpatch.pl4685 for my $chunk (@chunks) {
4686 my ($cond, $block) = @{$chunk};
/linux-4.4.14/drivers/net/wireless/ath/ath10k/
Dwmi.c5217 struct host_memory_chunk *chunk; in ath10k_wmi_put_host_mem_chunks() local
5223 chunk = &chunks->items[i]; in ath10k_wmi_put_host_mem_chunks()
5224 chunk->ptr = __cpu_to_le32(ar->wmi.mem_chunks[i].paddr); in ath10k_wmi_put_host_mem_chunks()
5225 chunk->size = __cpu_to_le32(ar->wmi.mem_chunks[i].len); in ath10k_wmi_put_host_mem_chunks()
5226 chunk->req_id = __cpu_to_le32(ar->wmi.mem_chunks[i].req_id); in ath10k_wmi_put_host_mem_chunks()
/linux-4.4.14/arch/s390/
DKconfig449 hex "NUMA emulation memory chunk size"

12