Home
last modified time | relevance | path

Searched refs:chunk (Results 1 – 196 of 196) sorted by relevance

/linux-4.1.27/net/sctp/
Dinqueue.c59 struct sctp_chunk *chunk, *tmp; in sctp_inq_free() local
62 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) { in sctp_inq_free()
63 list_del_init(&chunk->list); in sctp_inq_free()
64 sctp_chunk_free(chunk); in sctp_inq_free()
79 void sctp_inq_push(struct sctp_inq *q, struct sctp_chunk *chunk) in sctp_inq_push() argument
82 if (chunk->rcvr->dead) { in sctp_inq_push()
83 sctp_chunk_free(chunk); in sctp_inq_push()
92 list_add_tail(&chunk->list, &q->in_chunk_list); in sctp_inq_push()
93 if (chunk->asoc) in sctp_inq_push()
94 chunk->asoc->stats.ipackets++; in sctp_inq_push()
[all …]
Doutput.c61 struct sctp_chunk *chunk);
63 struct sctp_chunk *chunk);
65 struct sctp_chunk *chunk);
67 struct sctp_chunk *chunk,
87 struct sctp_chunk *chunk = NULL; in sctp_packet_config() local
94 chunk = sctp_get_ecne_prepend(packet->transport->asoc); in sctp_packet_config()
99 if (chunk) in sctp_packet_config()
100 sctp_packet_append_chunk(packet, chunk); in sctp_packet_config()
137 struct sctp_chunk *chunk, *tmp; in sctp_packet_free() local
141 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free()
[all …]
Doutqueue.c218 struct sctp_chunk *chunk, *tmp; in __sctp_outq_teardown() local
224 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
227 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
228 sctp_chunk_free(chunk); in __sctp_outq_teardown()
235 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
237 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
238 sctp_chunk_free(chunk); in __sctp_outq_teardown()
244 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown()
246 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown()
247 sctp_chunk_free(chunk); in __sctp_outq_teardown()
[all …]
Dchunk.c75 struct sctp_chunk *chunk; in sctp_datamsg_free() local
80 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free()
81 sctp_chunk_free(chunk); in sctp_datamsg_free()
90 struct sctp_chunk *chunk; in sctp_datamsg_destroy() local
102 chunk = list_entry(pos, struct sctp_chunk, frag_list); in sctp_datamsg_destroy()
105 asoc = chunk->asoc; in sctp_datamsg_destroy()
119 if (chunk->has_tsn) in sctp_datamsg_destroy()
124 ev = sctp_ulpevent_make_send_failed(asoc, chunk, sent, in sctp_datamsg_destroy()
130 sctp_chunk_put(chunk); in sctp_datamsg_destroy()
151 static void sctp_datamsg_assign(struct sctp_datamsg *msg, struct sctp_chunk *chunk) in sctp_datamsg_assign() argument
[all …]
Dsm_statefuns.c65 struct sctp_chunk *chunk,
69 struct sctp_chunk *chunk,
73 const struct sctp_chunk *chunk);
77 const struct sctp_chunk *chunk,
98 static struct sctp_sackhdr *sctp_sm_pull_sack(struct sctp_chunk *chunk);
151 struct sctp_chunk *chunk);
168 sctp_chunk_length_valid(struct sctp_chunk *chunk, in sctp_chunk_length_valid() argument
171 __u16 chunk_length = ntohs(chunk->chunk_hdr->length); in sctp_chunk_length_valid()
174 if (unlikely(chunk->pdiscard)) in sctp_chunk_length_valid()
224 struct sctp_chunk *chunk = arg; in sctp_sf_do_4_C() local
[all …]
Dsm_make_chunk.c79 static void *sctp_addto_param(struct sctp_chunk *chunk, int len,
90 static void sctp_control_set_owner_w(struct sctp_chunk *chunk) in sctp_control_set_owner_w() argument
92 struct sctp_association *asoc = chunk->asoc; in sctp_control_set_owner_w()
93 struct sk_buff *skb = chunk->skb; in sctp_control_set_owner_w()
107 int sctp_chunk_iif(const struct sctp_chunk *chunk) in sctp_chunk_iif() argument
112 af = sctp_get_af_specific(ipver2af(ip_hdr(chunk->skb)->version)); in sctp_chunk_iif()
114 iif = af->skb_iif(chunk->skb); in sctp_chunk_iif()
137 void sctp_init_cause(struct sctp_chunk *chunk, __be16 cause_code, in sctp_init_cause() argument
147 chunk->subh.err_hdr = sctp_addto_chunk(chunk, sizeof(sctp_errhdr_t), &err); in sctp_init_cause()
155 static int sctp_init_cause_fixed(struct sctp_chunk *chunk, __be16 cause_code, in sctp_init_cause_fixed() argument
[all …]
Dsm_sideeffect.c102 struct sctp_chunk *chunk) in sctp_do_ecn_ecne_work() argument
134 repl = sctp_make_cwr(asoc, asoc->last_cwr_tsn, chunk); in sctp_do_ecn_ecne_work()
571 struct sctp_chunk *chunk, in sctp_cmd_assoc_failed() argument
579 if (event_type == SCTP_EVENT_T_CHUNK && subtype.chunk == SCTP_CID_ABORT) in sctp_cmd_assoc_failed()
581 (__u16)error, 0, 0, chunk, in sctp_cmd_assoc_failed()
592 abort = sctp_make_violation_max_retrans(asoc, chunk); in sctp_cmd_assoc_failed()
613 struct sctp_chunk *chunk, in sctp_cmd_process_init() argument
624 if (!sctp_process_init(asoc, chunk, sctp_source(chunk), peer_init, gfp)) in sctp_cmd_process_init()
690 struct sctp_chunk *chunk) in sctp_cmd_transport_on() argument
746 hbinfo = (sctp_sender_hb_info_t *) chunk->skb->data; in sctp_cmd_transport_on()
[all …]
Dendpointola.c413 struct sctp_chunk *chunk; in sctp_endpoint_bh_rcv() local
428 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_endpoint_bh_rcv()
429 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_endpoint_bh_rcv()
434 if (first_time && (subtype.chunk == SCTP_CID_AUTH)) { in sctp_endpoint_bh_rcv()
447 chunk->auth_chunk = skb_clone(chunk->skb, in sctp_endpoint_bh_rcv()
449 chunk->auth = 1; in sctp_endpoint_bh_rcv()
460 if (NULL == chunk->asoc) { in sctp_endpoint_bh_rcv()
462 sctp_source(chunk), in sctp_endpoint_bh_rcv()
464 chunk->asoc = asoc; in sctp_endpoint_bh_rcv()
465 chunk->transport = transport; in sctp_endpoint_bh_rcv()
[all …]
Dulpevent.c129 __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) in sctp_ulpevent_make_assoc_change() argument
138 if (chunk) { in sctp_ulpevent_make_assoc_change()
142 skb = skb_copy_expand(chunk->skb, in sctp_ulpevent_make_assoc_change()
158 ntohs(chunk->chunk_hdr->length) - in sctp_ulpevent_make_assoc_change()
371 struct sctp_chunk *chunk, __u16 flags, in sctp_ulpevent_make_remote_error() argument
381 ch = (sctp_errhdr_t *)(chunk->skb->data); in sctp_ulpevent_make_remote_error()
386 skb_pull(chunk->skb, sizeof(sctp_errhdr_t)); in sctp_ulpevent_make_remote_error()
391 skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp); in sctp_ulpevent_make_remote_error()
394 skb_pull(chunk->skb, elen); in sctp_ulpevent_make_remote_error()
427 const struct sctp_association *asoc, struct sctp_chunk *chunk, in sctp_ulpevent_make_send_failed() argument
[all …]
Dinput.c100 struct sctp_chunk *chunk; member
114 struct sctp_chunk *chunk; in sctp_rcv() local
224 chunk = sctp_chunkify(skb, asoc, sk); in sctp_rcv()
225 if (!chunk) in sctp_rcv()
227 SCTP_INPUT_CB(skb)->chunk = chunk; in sctp_rcv()
230 chunk->rcvr = rcvr; in sctp_rcv()
233 chunk->sctp_hdr = sh; in sctp_rcv()
236 sctp_init_addrs(chunk, &src, &dest); in sctp_rcv()
239 chunk->transport = transport; in sctp_rcv()
263 sctp_chunk_free(chunk); in sctp_rcv()
[all …]
Ddebug.c65 if (cid.chunk <= SCTP_CID_BASE_MAX) in sctp_cname()
66 return sctp_cid_tbl[cid.chunk]; in sctp_cname()
68 switch (cid.chunk) { in sctp_cname()
Dauth.c391 struct sctp_chunk *chunk; in sctp_auth_asoc_init_active_key() local
417 list_for_each_entry(chunk, &asoc->outqueue.out_chunk_list, list) { in sctp_auth_asoc_init_active_key()
418 if (sctp_auth_send_cid(chunk->chunk_hdr->type, asoc)) in sctp_auth_asoc_init_active_key()
419 chunk->auth = 1; in sctp_auth_asoc_init_active_key()
638 static int __sctp_auth_cid(sctp_cid_t chunk, struct sctp_chunks_param *param) in __sctp_auth_cid() argument
664 if (param->chunks[i] == chunk) in __sctp_auth_cid()
674 int sctp_auth_send_cid(sctp_cid_t chunk, const struct sctp_association *asoc) in sctp_auth_send_cid() argument
682 return __sctp_auth_cid(chunk, asoc->peer.peer_chunks); in sctp_auth_send_cid()
686 int sctp_auth_recv_cid(sctp_cid_t chunk, const struct sctp_association *asoc) in sctp_auth_recv_cid() argument
694 return __sctp_auth_cid(chunk, in sctp_auth_recv_cid()
Dassociola.c919 struct sctp_chunk *chunk; in sctp_assoc_lookup_tsn() local
941 list_for_each_entry(chunk, &active->transmitted, in sctp_assoc_lookup_tsn()
944 if (key == chunk->subh.data_hdr->tsn) { in sctp_assoc_lookup_tsn()
956 list_for_each_entry(chunk, &transport->transmitted, in sctp_assoc_lookup_tsn()
958 if (key == chunk->subh.data_hdr->tsn) { in sctp_assoc_lookup_tsn()
1001 struct sctp_chunk *chunk; in sctp_assoc_bh_rcv() local
1012 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_assoc_bh_rcv()
1014 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_assoc_bh_rcv()
1023 if (sctp_auth_recv_cid(subtype.chunk, asoc) && !chunk->auth) in sctp_assoc_bh_rcv()
1029 if (sctp_chunk_is_data(chunk)) in sctp_assoc_bh_rcv()
[all …]
Dobjcnt.c52 SCTP_DBG_OBJCNT(chunk);
66 SCTP_DBG_OBJCNT_ENTRY(chunk),
Dulpqueue.c95 int sctp_ulpq_tail_data(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, in sctp_ulpq_tail_data() argument
103 event = sctp_ulpevent_make_rcvmsg(chunk->asoc, chunk, gfp); in sctp_ulpq_tail_data()
1080 void sctp_ulpq_renege(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, in sctp_ulpq_renege() argument
1088 if (chunk) { in sctp_ulpq_renege()
1089 needed = ntohs(chunk->chunk_hdr->length); in sctp_ulpq_renege()
1103 if (chunk && (freed >= needed)) { in sctp_ulpq_renege()
1105 retval = sctp_ulpq_tail_data(ulpq, chunk, gfp); in sctp_ulpq_renege()
Dsocket.c99 struct sctp_chunk *chunk);
148 static inline void sctp_set_owner_w(struct sctp_chunk *chunk) in sctp_set_owner_w() argument
150 struct sctp_association *asoc = chunk->asoc; in sctp_set_owner_w()
156 skb_set_owner_w(chunk->skb, sk); in sctp_set_owner_w()
158 chunk->skb->destructor = sctp_wfree; in sctp_set_owner_w()
160 skb_shinfo(chunk->skb)->destructor_arg = chunk; in sctp_set_owner_w()
162 asoc->sndbuf_used += SCTP_DATA_SNDSIZE(chunk) + in sctp_set_owner_w()
167 sk->sk_wmem_queued += chunk->skb->truesize; in sctp_set_owner_w()
168 sk_mem_charge(sk, chunk->skb->truesize); in sctp_set_owner_w()
411 struct sctp_chunk *chunk) in sctp_send_asconf() argument
[all …]
Dprobe.c137 struct sctp_chunk *chunk = arg; in jsctp_sf_eat_sack() local
138 struct sk_buff *skb = chunk->skb; in jsctp_sf_eat_sack()
DMakefile10 transport.o chunk.o sm_make_chunk.o ulpevent.o \
Dsm_statetable.c86 return sctp_chunk_event_lookup(net, event_subtype.chunk, state); in sctp_sm_lookup_event()
/linux-4.1.27/mm/
Dpercpu-vm.c13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk, in pcpu_chunk_page() argument
17 WARN_ON(chunk->immutable); in pcpu_chunk_page()
19 return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx)); in pcpu_chunk_page()
55 static void pcpu_free_pages(struct pcpu_chunk *chunk, in pcpu_free_pages() argument
82 static int pcpu_alloc_pages(struct pcpu_chunk *chunk, in pcpu_alloc_pages() argument
125 static void pcpu_pre_unmap_flush(struct pcpu_chunk *chunk, in pcpu_pre_unmap_flush() argument
129 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start), in pcpu_pre_unmap_flush()
130 pcpu_chunk_addr(chunk, pcpu_high_unit_cpu, page_end)); in pcpu_pre_unmap_flush()
151 static void pcpu_unmap_pages(struct pcpu_chunk *chunk, in pcpu_unmap_pages() argument
161 page = pcpu_chunk_page(chunk, cpu, i); in pcpu_unmap_pages()
[all …]
Dpercpu.c218 static int pcpu_chunk_slot(const struct pcpu_chunk *chunk) in pcpu_chunk_slot() argument
220 if (chunk->free_size < sizeof(int) || chunk->contig_hint < sizeof(int)) in pcpu_chunk_slot()
223 return pcpu_size_to_slot(chunk->free_size); in pcpu_chunk_slot()
243 static unsigned long pcpu_chunk_addr(struct pcpu_chunk *chunk, in pcpu_chunk_addr() argument
246 return (unsigned long)chunk->base_addr + pcpu_unit_offsets[cpu] + in pcpu_chunk_addr()
250 static void __maybe_unused pcpu_next_unpop(struct pcpu_chunk *chunk, in pcpu_next_unpop() argument
253 *rs = find_next_zero_bit(chunk->populated, end, *rs); in pcpu_next_unpop()
254 *re = find_next_bit(chunk->populated, end, *rs + 1); in pcpu_next_unpop()
257 static void __maybe_unused pcpu_next_pop(struct pcpu_chunk *chunk, in pcpu_next_pop() argument
260 *rs = find_next_bit(chunk->populated, end, *rs); in pcpu_next_pop()
[all …]
Dpercpu-km.c36 static int pcpu_populate_chunk(struct pcpu_chunk *chunk, in pcpu_populate_chunk() argument
42 static void pcpu_depopulate_chunk(struct pcpu_chunk *chunk, in pcpu_depopulate_chunk() argument
51 struct pcpu_chunk *chunk; in pcpu_create_chunk() local
55 chunk = pcpu_alloc_chunk(); in pcpu_create_chunk()
56 if (!chunk) in pcpu_create_chunk()
61 pcpu_free_chunk(chunk); in pcpu_create_chunk()
66 pcpu_set_page_chunk(nth_page(pages, i), chunk); in pcpu_create_chunk()
68 chunk->data = pages; in pcpu_create_chunk()
69 chunk->base_addr = page_address(pages) - pcpu_group_offsets[0]; in pcpu_create_chunk()
72 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk()
[all …]
/linux-4.1.27/drivers/s390/cio/
Ditcw.c181 void *chunk; in itcw_init() local
193 chunk = fit_chunk(&start, end, sizeof(struct itcw), 1, 0); in itcw_init()
194 if (IS_ERR(chunk)) in itcw_init()
195 return chunk; in itcw_init()
196 itcw = chunk; in itcw_init()
209 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init()
210 if (IS_ERR(chunk)) in itcw_init()
211 return chunk; in itcw_init()
212 itcw->tcw = chunk; in itcw_init()
217 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init()
[all …]
/linux-4.1.27/lib/
Dgenalloc.c39 static inline size_t chunk_size(const struct gen_pool_chunk *chunk) in chunk_size() argument
41 return chunk->end_addr - chunk->start_addr + 1; in chunk_size()
184 struct gen_pool_chunk *chunk; in gen_pool_add_virt() local
189 chunk = kzalloc_node(nbytes, GFP_KERNEL, nid); in gen_pool_add_virt()
190 if (unlikely(chunk == NULL)) in gen_pool_add_virt()
193 chunk->phys_addr = phys; in gen_pool_add_virt()
194 chunk->start_addr = virt; in gen_pool_add_virt()
195 chunk->end_addr = virt + size - 1; in gen_pool_add_virt()
196 atomic_set(&chunk->avail, size); in gen_pool_add_virt()
199 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_virt()
[all …]
Dbitmap.c363 u32 chunk; in __bitmap_parse() local
370 chunk = ndigits = 0; in __bitmap_parse()
405 if (chunk & ~((1UL << (CHUNKSZ - 4)) - 1)) in __bitmap_parse()
408 chunk = (chunk << 4) | hex_to_bin(c); in __bitmap_parse()
413 if (nchunks == 0 && chunk == 0) in __bitmap_parse()
417 *maskp |= chunk; in __bitmap_parse()
419 nbits += (nchunks == 1) ? nbits_to_hold_value(chunk) : CHUNKSZ; in __bitmap_parse()
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Dicm.c54 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_pages() argument
58 if (chunk->nsg > 0) in mlx4_free_icm_pages()
59 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, in mlx4_free_icm_pages()
62 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages()
63 __free_pages(sg_page(&chunk->mem[i]), in mlx4_free_icm_pages()
64 get_order(chunk->mem[i].length)); in mlx4_free_icm_pages()
67 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_coherent() argument
71 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent()
73 chunk->mem[i].length, in mlx4_free_icm_coherent()
74 lowmem_page_address(sg_page(&chunk->mem[i])), in mlx4_free_icm_coherent()
[all …]
Dicm.h64 struct mlx4_icm_chunk *chunk; member
91 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first()
99 return !iter->chunk; in mlx4_icm_last()
104 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next()
105 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
106 iter->chunk = NULL; in mlx4_icm_next()
110 iter->chunk = list_entry(iter->chunk->list.next, in mlx4_icm_next()
118 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_addr()
123 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_size()
Dmr.c724 int chunk; in __mlx4_write_mtt() local
733 chunk = min_t(int, max_mtts_first_page, npages); in __mlx4_write_mtt()
736 err = mlx4_write_mtt_chunk(dev, mtt, start_index, chunk, page_list); in __mlx4_write_mtt()
739 npages -= chunk; in __mlx4_write_mtt()
740 start_index += chunk; in __mlx4_write_mtt()
741 page_list += chunk; in __mlx4_write_mtt()
743 chunk = min_t(int, mtts_per_page, npages); in __mlx4_write_mtt()
753 int chunk; in mlx4_write_mtt() local
767 chunk = min_t(int, MLX4_MAILBOX_SIZE / sizeof(u64) - 2, in mlx4_write_mtt()
771 for (i = 0; i < chunk; ++i) in mlx4_write_mtt()
[all …]
/linux-4.1.27/kernel/
Daudit_tree.c109 static void free_chunk(struct audit_chunk *chunk) in free_chunk() argument
113 for (i = 0; i < chunk->count; i++) { in free_chunk()
114 if (chunk->owners[i].owner) in free_chunk()
115 put_tree(chunk->owners[i].owner); in free_chunk()
117 kfree(chunk); in free_chunk()
120 void audit_put_chunk(struct audit_chunk *chunk) in audit_put_chunk() argument
122 if (atomic_long_dec_and_test(&chunk->refs)) in audit_put_chunk()
123 free_chunk(chunk); in audit_put_chunk()
128 struct audit_chunk *chunk = container_of(rcu, struct audit_chunk, head); in __put_chunk() local
129 audit_put_chunk(chunk); in __put_chunk()
[all …]
Dauditsc.c210 static int put_tree_ref(struct audit_context *ctx, struct audit_chunk *chunk) in put_tree_ref() argument
215 p->c[--left] = chunk; in put_tree_ref()
223 p->c[30] = chunk; in put_tree_ref()
1588 struct audit_chunk *chunk; in handle_one() local
1596 chunk = audit_tree_lookup(inode); in handle_one()
1598 if (!chunk) in handle_one()
1600 if (likely(put_tree_ref(context, chunk))) in handle_one()
1605 audit_put_chunk(chunk); in handle_one()
1609 put_tree_ref(context, chunk); in handle_one()
1634 struct audit_chunk *chunk; in handle_path() local
[all …]
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_memfree.c64 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_pages() argument
68 if (chunk->nsg > 0) in mthca_free_icm_pages()
69 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, in mthca_free_icm_pages()
72 for (i = 0; i < chunk->npages; ++i) in mthca_free_icm_pages()
73 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages()
74 get_order(chunk->mem[i].length)); in mthca_free_icm_pages()
77 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_coherent() argument
81 for (i = 0; i < chunk->npages; ++i) { in mthca_free_icm_coherent()
82 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, in mthca_free_icm_coherent()
83 lowmem_page_address(sg_page(&chunk->mem[i])), in mthca_free_icm_coherent()
[all …]
Dmthca_memfree.h76 struct mthca_icm_chunk *chunk; member
103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first()
111 return !iter->chunk; in mthca_icm_last()
116 if (++iter->page_idx >= iter->chunk->nsg) { in mthca_icm_next()
117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next()
118 iter->chunk = NULL; in mthca_icm_next()
122 iter->chunk = list_entry(iter->chunk->list.next, in mthca_icm_next()
130 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mthca_icm_addr()
135 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mthca_icm_size()
Dmthca_mr.c362 int chunk; in mthca_write_mtt() local
369 chunk = min(size, list_len); in mthca_write_mtt()
372 buffer_list, chunk); in mthca_write_mtt()
375 buffer_list, chunk); in mthca_write_mtt()
377 list_len -= chunk; in mthca_write_mtt()
378 start_index += chunk; in mthca_write_mtt()
379 buffer_list += chunk; in mthca_write_mtt()
/linux-4.1.27/drivers/gpu/drm/qxl/
Dqxl_image.c38 struct qxl_drm_chunk *chunk; in qxl_allocate_chunk() local
41 chunk = kmalloc(sizeof(struct qxl_drm_chunk), GFP_KERNEL); in qxl_allocate_chunk()
42 if (!chunk) in qxl_allocate_chunk()
45 ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo); in qxl_allocate_chunk()
47 kfree(chunk); in qxl_allocate_chunk()
51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk()
88 struct qxl_drm_chunk *chunk, *tmp; in qxl_image_free_objects() local
90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects()
91 qxl_bo_unref(&chunk->bo); in qxl_image_free_objects()
92 kfree(chunk); in qxl_image_free_objects()
[all …]
Dqxl_dev.h421 struct qxl_data_chunk chunk; member
482 struct qxl_data_chunk chunk; member
761 struct qxl_data_chunk chunk; member
779 struct qxl_data_chunk chunk; member
Dqxl_draw.c52 dev_clips->chunk.next_chunk = 0; in drawable_set_clipping()
53 dev_clips->chunk.prev_chunk = 0; in drawable_set_clipping()
54 dev_clips->chunk.data_size = sizeof(struct qxl_rect) * num_clips; in drawable_set_clipping()
55 return (struct qxl_rect *)dev_clips->chunk.data; in drawable_set_clipping()
Dqxl_display.c360 cursor->chunk.next_chunk = 0; in qxl_crtc_cursor_set2()
361 cursor->chunk.prev_chunk = 0; in qxl_crtc_cursor_set2()
362 cursor->chunk.data_size = size; in qxl_crtc_cursor_set2()
364 memcpy(cursor->chunk.data, user_ptr, size); in qxl_crtc_cursor_set2()
/linux-4.1.27/arch/mips/dec/prom/
Dconsole.c22 unsigned int chunk = sizeof(buf) - 1; in prom_console_write() local
25 if (chunk > c) in prom_console_write()
26 chunk = c; in prom_console_write()
27 memcpy(buf, s, chunk); in prom_console_write()
28 buf[chunk] = '\0'; in prom_console_write()
30 s += chunk; in prom_console_write()
31 c -= chunk; in prom_console_write()
/linux-4.1.27/sound/core/
Dsgbuf.c71 unsigned int i, pages, chunk, maxpages; in snd_malloc_sgbuf_pages() local
96 chunk = pages; in snd_malloc_sgbuf_pages()
98 if (chunk > maxpages) in snd_malloc_sgbuf_pages()
99 chunk = maxpages; in snd_malloc_sgbuf_pages()
100 chunk <<= PAGE_SHIFT; in snd_malloc_sgbuf_pages()
102 chunk, &tmpb) < 0) { in snd_malloc_sgbuf_pages()
110 chunk = tmpb.bytes >> PAGE_SHIFT; in snd_malloc_sgbuf_pages()
111 for (i = 0; i < chunk; i++) { in snd_malloc_sgbuf_pages()
115 table->addr |= chunk; /* mark head */ in snd_malloc_sgbuf_pages()
121 sgbuf->pages += chunk; in snd_malloc_sgbuf_pages()
[all …]
/linux-4.1.27/drivers/media/usb/usbtv/
Dusbtv.h48 #define USBTV_MAGIC_OK(chunk) ((be32_to_cpu(chunk[0]) & 0xff000000) \ argument
50 #define USBTV_FRAME_ID(chunk) ((be32_to_cpu(chunk[0]) & 0x00ff0000) >> 16) argument
51 #define USBTV_ODD(chunk) ((be32_to_cpu(chunk[0]) & 0x0000f000) >> 15) argument
52 #define USBTV_CHUNK_NO(chunk) (be32_to_cpu(chunk[0]) & 0x00000fff) argument
Dusbtv-video.c275 static void usbtv_image_chunk(struct usbtv *usbtv, __be32 *chunk) in usbtv_image_chunk() argument
283 if (!USBTV_MAGIC_OK(chunk)) in usbtv_image_chunk()
285 frame_id = USBTV_FRAME_ID(chunk); in usbtv_image_chunk()
286 odd = USBTV_ODD(chunk); in usbtv_image_chunk()
287 chunk_no = USBTV_CHUNK_NO(chunk); in usbtv_image_chunk()
312 usbtv_chunk_to_vbuf(frame, &chunk[1], chunk_no, odd); in usbtv_image_chunk()
/linux-4.1.27/drivers/infiniband/hw/usnic/
Dusnic_vnic.c81 struct usnic_vnic_res_chunk *chunk; in usnic_vnic_dump() local
104 chunk = &vnic->chunks[i]; in usnic_vnic_dump()
105 for (j = 0; j < chunk->cnt; j++) { in usnic_vnic_dump()
106 res = chunk->res[j]; in usnic_vnic_dump()
262 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk) in usnic_vnic_put_resources() argument
267 struct usnic_vnic *vnic = chunk->vnic; in usnic_vnic_put_resources()
270 while ((i = --chunk->cnt) >= 0) { in usnic_vnic_put_resources()
271 res = chunk->res[i]; in usnic_vnic_put_resources()
272 chunk->res[i] = NULL; in usnic_vnic_put_resources()
278 kfree(chunk->res); in usnic_vnic_put_resources()
[all …]
Dusnic_uiom.c81 struct usnic_uiom_chunk *chunk, *tmp; in usnic_uiom_put_pages() local
87 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages()
88 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_put_pages()
96 kfree(chunk); in usnic_uiom_put_pages()
105 struct usnic_uiom_chunk *chunk; in usnic_uiom_get_pages() local
159 chunk = kmalloc(sizeof(*chunk) + in usnic_uiom_get_pages()
163 if (!chunk) { in usnic_uiom_get_pages()
168 chunk->nents = min_t(int, ret, USNIC_UIOM_PAGE_CHUNK); in usnic_uiom_get_pages()
169 sg_init_table(chunk->page_list, chunk->nents); in usnic_uiom_get_pages()
170 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_get_pages()
[all …]
Dusnic_ib_verbs.c49 struct usnic_vnic_res_chunk *chunk; in usnic_ib_fill_create_qp_resp() local
74 chunk = usnic_ib_qp_grp_get_chunk(qp_grp, USNIC_VNIC_RES_TYPE_RQ); in usnic_ib_fill_create_qp_resp()
75 if (IS_ERR_OR_NULL(chunk)) { in usnic_ib_fill_create_qp_resp()
79 PTR_ERR(chunk)); in usnic_ib_fill_create_qp_resp()
80 return chunk ? PTR_ERR(chunk) : -ENOMEM; in usnic_ib_fill_create_qp_resp()
83 WARN_ON(chunk->type != USNIC_VNIC_RES_TYPE_RQ); in usnic_ib_fill_create_qp_resp()
84 resp.rq_cnt = chunk->cnt; in usnic_ib_fill_create_qp_resp()
85 for (i = 0; i < chunk->cnt; i++) in usnic_ib_fill_create_qp_resp()
86 resp.rq_idx[i] = chunk->res[i]->vnic_idx; in usnic_ib_fill_create_qp_resp()
88 chunk = usnic_ib_qp_grp_get_chunk(qp_grp, USNIC_VNIC_RES_TYPE_WQ); in usnic_ib_fill_create_qp_resp()
[all …]
Dusnic_vnic.h95 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk);
/linux-4.1.27/drivers/s390/block/
Ddasd_int.h573 struct dasd_mchunk *chunk; in dasd_init_chunklist() local
576 chunk = (struct dasd_mchunk *) mem; in dasd_init_chunklist()
577 chunk->size = size - sizeof(struct dasd_mchunk); in dasd_init_chunklist()
578 list_add(&chunk->list, chunk_list); in dasd_init_chunklist()
584 struct dasd_mchunk *chunk, *tmp; in dasd_alloc_chunk() local
587 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk()
588 if (chunk->size < size) in dasd_alloc_chunk()
590 if (chunk->size > size + sizeof(struct dasd_mchunk)) { in dasd_alloc_chunk()
591 char *endaddr = (char *) (chunk + 1) + chunk->size; in dasd_alloc_chunk()
594 chunk->size -= size + sizeof(struct dasd_mchunk); in dasd_alloc_chunk()
[all …]
/linux-4.1.27/fs/ocfs2/
Dquota_local.c297 int chunk, in ocfs2_add_recovery_chunk() argument
305 rc->rc_chunk = chunk; in ocfs2_add_recovery_chunk()
471 int bit, chunk; in ocfs2_recover_local_quota_file() local
478 chunk = rchunk->rc_chunk; in ocfs2_recover_local_quota_file()
481 ol_quota_chunk_block(sb, chunk), in ocfs2_recover_local_quota_file()
491 ol_dqblk_block(sb, chunk, bit), in ocfs2_recover_local_quota_file()
498 ol_dqblk_block_off(sb, chunk, bit)); in ocfs2_recover_local_quota_file()
817 struct ocfs2_quota_chunk *chunk; in ocfs2_local_free_info() local
825 list_for_each_entry(chunk, &oinfo->dqi_chunk, qc_chunk) { in ocfs2_local_free_info()
827 (chunk->qc_headerbh->b_data); in ocfs2_local_free_info()
[all …]
Dioctl.c448 unsigned int offset = 0, cluster, chunk; in ocfs2_info_freefrag_scan_chain() local
487 for (chunk = 0; chunk < chunks_in_group; chunk++) { in ocfs2_info_freefrag_scan_chain()
/linux-4.1.27/drivers/media/usb/go7007/
Dgo7007-fw.c378 int size = 0, i, off = 0, chunk; in gen_mjpeghdr_to_package() local
388 chunk = mjpeg_frame_header(go, buf + size, 1); in gen_mjpeghdr_to_package()
389 memmove(buf + size, buf + size + 80, chunk - 80); in gen_mjpeghdr_to_package()
390 size += chunk - 80; in gen_mjpeghdr_to_package()
392 for (i = 0; i < size; i += chunk * 2) { in gen_mjpeghdr_to_package()
400 chunk = 28; in gen_mjpeghdr_to_package()
401 if (mem + chunk > 0x4000) in gen_mjpeghdr_to_package()
402 chunk = 0x4000 - mem; in gen_mjpeghdr_to_package()
403 if (i + 2 * chunk > size) in gen_mjpeghdr_to_package()
404 chunk = (size - i) / 2; in gen_mjpeghdr_to_package()
[all …]
/linux-4.1.27/include/net/sctp/
Dsm.h174 __u8 sctp_get_chunk_type(struct sctp_chunk *chunk);
209 const struct sctp_chunk *chunk);
239 const struct sctp_chunk *chunk,
252 struct sctp_chunk *chunk, bool addr_param_needed,
299 static inline __u16 sctp_data_size(struct sctp_chunk *chunk) in sctp_data_size() argument
303 size = ntohs(chunk->chunk_hdr->length); in sctp_data_size()
392 sctp_vtag_verify(const struct sctp_chunk *chunk, in sctp_vtag_verify() argument
401 if (ntohl(chunk->sctp_hdr->vtag) == asoc->c.my_vtag) in sctp_vtag_verify()
411 sctp_vtag_verify_either(const struct sctp_chunk *chunk, in sctp_vtag_verify_either() argument
435 if ((!sctp_test_T_bit(chunk) && in sctp_vtag_verify_either()
[all …]
Dsctp.h432 static inline bool sctp_chunk_pending(const struct sctp_chunk *chunk) in sctp_chunk_pending() argument
434 return !list_empty(&chunk->list); in sctp_chunk_pending()
442 #define sctp_walk_params(pos, chunk, member)\ argument
443 _sctp_walk_params((pos), (chunk), ntohs((chunk)->chunk_hdr.length), member)
445 #define _sctp_walk_params(pos, chunk, end, member)\ argument
446 for (pos.v = chunk->member;\
447 pos.v <= (void *)chunk + end - ntohs(pos.p->length) &&\
461 #define sctp_walk_fwdtsn(pos, chunk)\ argument
462 _sctp_walk_fwdtsn((pos), (chunk), ntohs((chunk)->chunk_hdr->length) - sizeof(struct sctp_fwdtsn_chu…
464 #define _sctp_walk_fwdtsn(pos, chunk, end)\ argument
[all …]
Dulpevent.h85 struct sctp_chunk *chunk,
98 struct sctp_chunk *chunk,
103 struct sctp_chunk *chunk,
121 struct sctp_chunk *chunk,
Dauth.h100 int sctp_auth_send_cid(sctp_cid_t chunk, const struct sctp_association *asoc);
101 int sctp_auth_recv_cid(sctp_cid_t chunk, const struct sctp_association *asoc);
Dconstants.h128 sctp_cid_t chunk; member
139 SCTP_SUBTYPE_CONSTRUCTOR(CHUNK, sctp_cid_t, chunk)
Dcommand.h131 struct sctp_chunk *chunk; member
172 SCTP_ARG_CONSTRUCTOR(CHUNK, struct sctp_chunk *, chunk) in SCTP_ARG_CONSTRUCTOR()
Dstructs.h654 int sctp_user_addto_chunk(struct sctp_chunk *chunk, int len,
663 const union sctp_addr *sctp_source(const struct sctp_chunk *chunk);
1058 int sctp_outq_tail(struct sctp_outq *, struct sctp_chunk *chunk);
1281 struct sctp_chunk *chunk, struct sctp_chunk **err_chunk);
1282 int sctp_process_init(struct sctp_association *, struct sctp_chunk *chunk,
/linux-4.1.27/sound/soc/codecs/
Dsigmadsp.c54 struct sigma_fw_chunk chunk; member
60 struct sigma_fw_chunk chunk; member
68 struct sigma_fw_chunk chunk; member
200 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_control() argument
212 ctrl_chunk = (const struct sigma_fw_chunk_control *)chunk; in sigma_fw_load_control()
238 ctrl->samplerates = le32_to_cpu(chunk->samplerates); in sigma_fw_load_control()
251 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_data() argument
259 data_chunk = (struct sigma_fw_chunk_data *)chunk; in sigma_fw_load_data()
269 data->samplerates = le32_to_cpu(chunk->samplerates); in sigma_fw_load_data()
277 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_samplerates() argument
[all …]
/linux-4.1.27/arch/um/drivers/
Dmconsole_kern.h27 #define CONFIG_CHUNK(str, size, current, chunk, end) \ argument
29 current += strlen(chunk); \
33 strcpy(str, chunk); \
34 str += strlen(chunk); \
/linux-4.1.27/drivers/dma/sh/
Drcar-dmac.c328 struct rcar_dmac_xfer_chunk *chunk; in rcar_dmac_chan_start_xfer() local
352 chunk = list_first_entry(&desc->chunks, in rcar_dmac_chan_start_xfer()
355 chunk->dst_addr & 0xffffffff); in rcar_dmac_chan_start_xfer()
385 struct rcar_dmac_xfer_chunk *chunk = desc->running; in rcar_dmac_chan_start_xfer() local
389 chan->index, chunk, chunk->size, &chunk->src_addr, in rcar_dmac_chan_start_xfer()
390 &chunk->dst_addr); in rcar_dmac_chan_start_xfer()
394 chunk->src_addr >> 32); in rcar_dmac_chan_start_xfer()
396 chunk->dst_addr >> 32); in rcar_dmac_chan_start_xfer()
399 chunk->src_addr & 0xffffffff); in rcar_dmac_chan_start_xfer()
401 chunk->dst_addr & 0xffffffff); in rcar_dmac_chan_start_xfer()
[all …]
Dshdma-base.c75 struct shdma_desc *chunk, *c, *desc = in shdma_tx_submit() local
89 list_for_each_entry_safe(chunk, c, desc->node.prev, node) { in shdma_tx_submit()
94 if (chunk != desc && (chunk->mark == DESC_IDLE || in shdma_tx_submit()
95 chunk->async_tx.cookie > 0 || in shdma_tx_submit()
96 chunk->async_tx.cookie == -EBUSY || in shdma_tx_submit()
97 &chunk->node == &schan->ld_free)) in shdma_tx_submit()
99 chunk->mark = DESC_SUBMITTED; in shdma_tx_submit()
100 if (chunk->chunks == 1) { in shdma_tx_submit()
101 chunk->async_tx.callback = callback; in shdma_tx_submit()
102 chunk->async_tx.callback_param = tx->callback_param; in shdma_tx_submit()
[all …]
/linux-4.1.27/drivers/md/
Ddm-snap.c157 chunk_t chunk) in chunk_to_sector() argument
159 return chunk << store->chunk_shift; in chunk_to_sector()
215 chunk_t chunk; member
230 static void track_chunk(struct dm_snapshot *s, struct bio *bio, chunk_t chunk) in track_chunk() argument
234 c->chunk = chunk; in track_chunk()
238 &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)]); in track_chunk()
252 static int __chunk_is_tracked(struct dm_snapshot *s, chunk_t chunk) in __chunk_is_tracked() argument
260 &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)], node) { in __chunk_is_tracked()
261 if (c->chunk == chunk) { in __chunk_is_tracked()
276 static void __check_for_conflicting_io(struct dm_snapshot *s, chunk_t chunk) in __check_for_conflicting_io() argument
[all …]
Ddm-stripe.c214 sector_t chunk = dm_target_offset(sc->ti, sector); in stripe_map_sector() local
218 chunk_offset = sector_div(chunk, sc->chunk_size); in stripe_map_sector()
220 chunk_offset = chunk & (sc->chunk_size - 1); in stripe_map_sector()
221 chunk >>= sc->chunk_size_shift; in stripe_map_sector()
225 *stripe = sector_div(chunk, sc->stripes); in stripe_map_sector()
227 *stripe = chunk & (sc->stripes - 1); in stripe_map_sector()
228 chunk >>= sc->stripes_shift; in stripe_map_sector()
232 chunk *= sc->chunk_size; in stripe_map_sector()
234 chunk <<= sc->chunk_size_shift; in stripe_map_sector()
236 *result = chunk + chunk_offset; in stripe_map_sector()
Ddm-exception-store.h141 static inline chunk_t dm_chunk_number(chunk_t chunk) in dm_chunk_number() argument
143 return chunk & (chunk_t)((1ULL << DM_CHUNK_NUMBER_BITS) - 1ULL); in dm_chunk_number()
168 static inline chunk_t dm_chunk_number(chunk_t chunk) in dm_chunk_number() argument
170 return chunk; in dm_chunk_number()
Dbitmap.c725 unsigned long chunk) in file_page_index() argument
728 chunk += sizeof(bitmap_super_t) << 3; in file_page_index()
729 return chunk >> PAGE_BIT_SHIFT; in file_page_index()
734 unsigned long chunk) in file_page_offset() argument
737 chunk += sizeof(bitmap_super_t) << 3; in file_page_offset()
738 return chunk & (PAGE_BITS - 1); in file_page_offset()
746 unsigned long chunk) in filemap_get_page() argument
748 if (file_page_index(store, chunk) >= store->file_pages) in filemap_get_page()
750 return store->filemap[file_page_index(store, chunk)]; in filemap_get_page()
909 unsigned long chunk = block >> bitmap->counts.chunkshift; in bitmap_file_set_bit() local
[all …]
Ddm-snap-persistent.c228 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, int rw, in chunk_io() argument
233 .sector = ps->store->chunk_size * chunk, in chunk_io()
286 chunk_t chunk; in area_io() local
288 chunk = area_location(ps, ps->current_area); in area_io()
290 r = chunk_io(ps, ps->area, chunk, rw, 0); in area_io()
518 chunk_t chunk; in read_exceptions() local
533 chunk = area_location(ps, ps->current_area); in read_exceptions()
535 area = dm_bufio_read(client, chunk, &bp); in read_exceptions()
549 dm_bufio_forget(client, chunk); in read_exceptions()
Draid0.c311 sector_t chunk; in map_sector() local
322 chunk = *sector_offset; in map_sector()
324 sector_div(chunk, zone->nb_dev << chunksect_bits); in map_sector()
327 chunk = *sector_offset; in map_sector()
328 sector_div(chunk, chunk_sects * zone->nb_dev); in map_sector()
335 *sector_offset = (chunk * chunk_sects) + sect_in_chunk; in map_sector()
Draid10.c556 sector_t chunk; in __raid10_find_phys() local
569 chunk = r10bio->sector >> geo->chunk_shift; in __raid10_find_phys()
572 chunk *= geo->near_copies; in __raid10_find_phys()
573 stripe = chunk; in __raid10_find_phys()
632 sector_t offset, chunk, vchunk; in raid10_find_virt() local
655 chunk = sector >> geo->chunk_shift; in raid10_find_virt()
656 fc = sector_div(chunk, geo->far_copies); in raid10_find_virt()
668 chunk = sector >> geo->chunk_shift; in raid10_find_virt()
670 vchunk = chunk * geo->raid_disks + dev; in raid10_find_virt()
3470 int layout, chunk, disks; in setup_geo() local
[all …]
DKconfig68 up evenly, one chunk here and one chunk there. This will increase
Dmd.c4029 unsigned long chunk, end_chunk; in bitmap_store() local
4039 chunk = end_chunk = simple_strtoul(buf, &end, 0); in bitmap_store()
4047 bitmap_dirty_bits(mddev->bitmap, chunk, end_chunk); in bitmap_store()
4480 int chunk; in max_sync_store() local
4495 chunk = mddev->chunk_sectors; in max_sync_store()
4496 if (chunk) { in max_sync_store()
4500 if (sector_div(temp, chunk)) in max_sync_store()
/linux-4.1.27/Documentation/device-mapper/
Dstriped.txt10 Parameters: <num devs> <chunk size> [<dev path> <offset>]+
12 <chunk size>: Size of each chunk of data. Must be at least as
19 be a multiple of the chunk size multiplied by the number of underlying devices.
28 # will be called "stripe_dev" and have a chunk-size of 128k.
Dcache.txt141 Policy plug-ins can store a chunk of data per cache block. It's up to
142 the policy how big this chunk is, but it should be kept small. Like the
Dsnapshot.txt61 deferred until the merging snapshot's corresponding chunk(s) have been
/linux-4.1.27/arch/arm64/kernel/
Dsys_compat.c37 unsigned long chunk = min(PAGE_SIZE, end - start); in __do_compat_cache_op() local
42 ret = __flush_cache_user_range(start, start + chunk); in __do_compat_cache_op()
47 start += chunk; in __do_compat_cache_op()
/linux-4.1.27/net/bluetooth/
Daf_bluetooth.c306 int chunk; in bt_sock_stream_recvmsg() local
332 chunk = min_t(unsigned int, skb->len, size); in bt_sock_stream_recvmsg()
333 if (skb_copy_datagram_msg(skb, 0, msg, chunk)) { in bt_sock_stream_recvmsg()
339 copied += chunk; in bt_sock_stream_recvmsg()
340 size -= chunk; in bt_sock_stream_recvmsg()
347 if (chunk <= skb_len) { in bt_sock_stream_recvmsg()
348 __skb_pull(skb, chunk); in bt_sock_stream_recvmsg()
353 chunk -= skb_len; in bt_sock_stream_recvmsg()
356 if (chunk <= frag->len) { in bt_sock_stream_recvmsg()
358 skb->len -= chunk; in bt_sock_stream_recvmsg()
[all …]
/linux-4.1.27/drivers/atm/
Dfore200e.c178 fore200e_chunk_alloc(struct fore200e* fore200e, struct chunk* chunk, int size, int alignment, int d… in fore200e_chunk_alloc() argument
185 chunk->alloc_size = size + alignment; in fore200e_chunk_alloc()
186 chunk->align_size = size; in fore200e_chunk_alloc()
187 chunk->direction = direction; in fore200e_chunk_alloc()
189 chunk->alloc_addr = kzalloc(chunk->alloc_size, GFP_KERNEL | GFP_DMA); in fore200e_chunk_alloc()
190 if (chunk->alloc_addr == NULL) in fore200e_chunk_alloc()
194 offset = FORE200E_ALIGN(chunk->alloc_addr, alignment); in fore200e_chunk_alloc()
196 chunk->align_addr = chunk->alloc_addr + offset; in fore200e_chunk_alloc()
198chunk->dma_addr = fore200e->bus->dma_map(fore200e, chunk->align_addr, chunk->align_size, direction… in fore200e_chunk_alloc()
207 fore200e_chunk_free(struct fore200e* fore200e, struct chunk* chunk) in fore200e_chunk_free() argument
[all …]
Dfore200e.h563 typedef struct chunk { struct
581 struct chunk data; /* data buffer */
603 struct chunk status; /* array of completion status */
613 … struct chunk tpd; /* array of tpds */
614 … struct chunk status; /* arry of completion status */
624 …struct chunk rpd; /* array of rpds */
625 …struct chunk status; /* array of completion status */
634 …struct chunk rbd_block; /* array of rbds …
635 …struct chunk status; /* array of completion status …
811 int (*dma_chunk_alloc)(struct fore200e*, struct chunk*, int, int, int);
[all …]
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dresource.c419 unsigned start, chunk, top; in c4iw_ocqp_pool_create() local
426 chunk = rdev->lldi.vr->ocq.size; in c4iw_ocqp_pool_create()
427 top = start + chunk; in c4iw_ocqp_pool_create()
430 chunk = min(top - start + 1, chunk); in c4iw_ocqp_pool_create()
431 if (gen_pool_add(rdev->ocqp_pool, start, chunk, -1)) { in c4iw_ocqp_pool_create()
433 __func__, start, chunk); in c4iw_ocqp_pool_create()
434 if (chunk <= 1024 << MIN_OCQP_SHIFT) { in c4iw_ocqp_pool_create()
440 chunk >>= 1; in c4iw_ocqp_pool_create()
443 __func__, start, chunk); in c4iw_ocqp_pool_create()
444 start += chunk; in c4iw_ocqp_pool_create()
/linux-4.1.27/fs/reiserfs/
Dxattr.c527 size_t chunk; in reiserfs_xattr_set_handle() local
532 chunk = PAGE_CACHE_SIZE; in reiserfs_xattr_set_handle()
534 chunk = buffer_size - buffer_pos; in reiserfs_xattr_set_handle()
549 if (chunk + skip > PAGE_CACHE_SIZE) in reiserfs_xattr_set_handle()
550 chunk = PAGE_CACHE_SIZE - skip; in reiserfs_xattr_set_handle()
557 err = __reiserfs_write_begin(page, page_offset, chunk + skip); in reiserfs_xattr_set_handle()
560 memcpy(data + skip, buffer + buffer_pos, chunk); in reiserfs_xattr_set_handle()
562 page_offset + chunk + in reiserfs_xattr_set_handle()
568 buffer_pos += chunk; in reiserfs_xattr_set_handle()
569 file_pos += chunk; in reiserfs_xattr_set_handle()
[all …]
Djournal.c676 static void write_chunk(struct buffer_chunk *chunk) in write_chunk() argument
679 for (i = 0; i < chunk->nr; i++) { in write_chunk()
680 submit_logged_buffer(chunk->bh[i]); in write_chunk()
682 chunk->nr = 0; in write_chunk()
685 static void write_ordered_chunk(struct buffer_chunk *chunk) in write_ordered_chunk() argument
688 for (i = 0; i < chunk->nr; i++) { in write_ordered_chunk()
689 submit_ordered_buffer(chunk->bh[i]); in write_ordered_chunk()
691 chunk->nr = 0; in write_ordered_chunk()
694 static int add_to_chunk(struct buffer_chunk *chunk, struct buffer_head *bh, in add_to_chunk() argument
698 BUG_ON(chunk->nr >= CHUNK_SIZE); in add_to_chunk()
[all …]
/linux-4.1.27/fs/btrfs/
Dprint-tree.c23 static void print_chunk(struct extent_buffer *eb, struct btrfs_chunk *chunk) in print_chunk() argument
25 int num_stripes = btrfs_chunk_num_stripes(eb, chunk); in print_chunk()
29 btrfs_chunk_length(eb, chunk), btrfs_chunk_owner(eb, chunk), in print_chunk()
30 btrfs_chunk_type(eb, chunk), num_stripes); in print_chunk()
33 btrfs_stripe_devid_nr(eb, chunk, i), in print_chunk()
34 btrfs_stripe_offset_nr(eb, chunk, i)); in print_chunk()
Dvolumes.c2538 struct btrfs_chunk *chunk; in btrfs_del_sys_chunk() local
2560 chunk = (struct btrfs_chunk *)(ptr + len); in btrfs_del_sys_chunk()
2561 num_stripes = btrfs_stack_chunk_num_stripes(chunk); in btrfs_del_sys_chunk()
2712 struct btrfs_chunk *chunk; in btrfs_relocate_sys_chunks() local
2745 chunk = btrfs_item_ptr(leaf, path->slots[0], in btrfs_relocate_sys_chunks()
2747 chunk_type = btrfs_chunk_type(leaf, chunk); in btrfs_relocate_sys_chunks()
2979 struct btrfs_chunk *chunk, in chunk_devid_filter() argument
2983 int num_stripes = btrfs_chunk_num_stripes(leaf, chunk); in chunk_devid_filter()
2987 stripe = btrfs_stripe_nr(chunk, i); in chunk_devid_filter()
2997 struct btrfs_chunk *chunk, in chunk_drange_filter() argument
[all …]
/linux-4.1.27/sound/usb/line6/
Dmidi.c47 unsigned char chunk[LINE6_FALLBACK_MAXPACKETSIZE]; in line6_midi_transmit() local
52 done = snd_rawmidi_transmit_peek(substream, chunk, req); in line6_midi_transmit()
57 line6_midibuf_write(mb, chunk, done); in line6_midi_transmit()
62 done = line6_midibuf_read(mb, chunk, in line6_midi_transmit()
68 send_midi_async(line6, chunk, done); in line6_midi_transmit()
/linux-4.1.27/drivers/net/wireless/ti/wlcore/
Dboot.c179 u8 *p, *chunk; in wl1271_boot_upload_firmware_chunk() local
194 chunk = kmalloc(CHUNK_SIZE, GFP_KERNEL); in wl1271_boot_upload_firmware_chunk()
195 if (!chunk) { in wl1271_boot_upload_firmware_chunk()
226 memcpy(chunk, p, CHUNK_SIZE); in wl1271_boot_upload_firmware_chunk()
229 ret = wlcore_write(wl, addr, chunk, CHUNK_SIZE, false); in wl1271_boot_upload_firmware_chunk()
239 memcpy(chunk, p, fw_data_len % CHUNK_SIZE); in wl1271_boot_upload_firmware_chunk()
242 ret = wlcore_write(wl, addr, chunk, fw_data_len % CHUNK_SIZE, false); in wl1271_boot_upload_firmware_chunk()
245 kfree(chunk); in wl1271_boot_upload_firmware_chunk()
/linux-4.1.27/Documentation/mmc/
Dmmc-async-req.txt61 request in two chunks, prepare the first chunk and start the request,
62 and finally prepare the second chunk and start the transfer.
72 * The first chunk of the request should take the same time
75 * the transfer is delayed, guesstimate max 4k as first chunk size.
84 * of the first chunk. If the MMC runs out of the first data chunk
Dmmc-dev-attrs.txt59 hence "preferred_erase_size" provides a good chunk
/linux-4.1.27/drivers/crypto/
Dn2_core.c674 struct n2_crypto_chunk chunk; member
877 struct n2_crypto_chunk *chunk; in n2_compute_chunks() local
890 chunk = &rctx->chunk; in n2_compute_chunks()
891 INIT_LIST_HEAD(&chunk->entry); in n2_compute_chunks()
893 chunk->iv_paddr = 0UL; in n2_compute_chunks()
894 chunk->arr_len = 0; in n2_compute_chunks()
895 chunk->dest_paddr = 0UL; in n2_compute_chunks()
913 if (chunk->arr_len != 0) { in n2_compute_chunks()
917 chunk->arr_len == N2_CHUNK_ARR_LEN || in n2_compute_chunks()
919 chunk->dest_final = dest_prev; in n2_compute_chunks()
[all …]
/linux-4.1.27/sound/pci/
Des1968.c1374 struct esm_memory *chunk = kmalloc(sizeof(*chunk), GFP_KERNEL); in snd_es1968_new_memory() local
1375 if (chunk == NULL) { in snd_es1968_new_memory()
1379 chunk->buf = buf->buf; in snd_es1968_new_memory()
1380 chunk->buf.bytes -= size; in snd_es1968_new_memory()
1381 chunk->buf.area += size; in snd_es1968_new_memory()
1382 chunk->buf.addr += size; in snd_es1968_new_memory()
1383 chunk->empty = 1; in snd_es1968_new_memory()
1385 list_add(&chunk->list, &buf->list); in snd_es1968_new_memory()
1395 struct esm_memory *chunk; in snd_es1968_free_memory() local
1400 chunk = list_entry(buf->list.prev, struct esm_memory, list); in snd_es1968_free_memory()
[all …]
/linux-4.1.27/arch/x86/xen/
Dp2m.c309 unsigned int i, chunk; in xen_rebuild_p2m_list() local
332 for (pfn = 0; pfn < xen_max_p2m_pfn; pfn += chunk) { in xen_rebuild_p2m_list()
343 chunk = (pfn & (P2M_PER_PAGE * P2M_MID_PER_PAGE - 1)) ? in xen_rebuild_p2m_list()
349 for (i = 1; i < chunk; i++) in xen_rebuild_p2m_list()
352 if (i < chunk) in xen_rebuild_p2m_list()
354 chunk = P2M_PER_PAGE; in xen_rebuild_p2m_list()
356 if (type == P2M_TYPE_PFN || i < chunk) { in xen_rebuild_p2m_list()
370 if (chunk == P2M_PER_PAGE) { in xen_rebuild_p2m_list()
Dsetup.c310 unsigned int i, chunk; in xen_do_set_identity_and_remap_chunk() local
321 chunk = (left < REMAP_SIZE) ? left : REMAP_SIZE; in xen_do_set_identity_and_remap_chunk()
330 xen_remap_buf.size = chunk; in xen_do_set_identity_and_remap_chunk()
331 for (i = 0; i < chunk; i++) in xen_do_set_identity_and_remap_chunk()
338 set_phys_range_identity(ident_pfn_iter, ident_pfn_iter + chunk); in xen_do_set_identity_and_remap_chunk()
340 left -= chunk; in xen_do_set_identity_and_remap_chunk()
/linux-4.1.27/drivers/net/wireless/libertas/
Dif_sdio.c338 u16 size, type, chunk; in if_sdio_card_to_host() local
362 chunk = sdio_align_size(card->func, size); in if_sdio_card_to_host()
364 ret = sdio_readsb(card->func, card->buffer, card->ioport, chunk); in if_sdio_card_to_host()
368 chunk = card->buffer[0] | (card->buffer[1] << 8); in if_sdio_card_to_host()
372 (int)type, (int)chunk); in if_sdio_card_to_host()
374 if (chunk > size) { in if_sdio_card_to_host()
376 (int)chunk, (int)size); in if_sdio_card_to_host()
381 if (chunk < size) { in if_sdio_card_to_host()
383 (int)chunk, (int)size); in if_sdio_card_to_host()
388 ret = if_sdio_handle_cmd(card, card->buffer + 4, chunk - 4); in if_sdio_card_to_host()
[all …]
/linux-4.1.27/arch/s390/crypto/
Dprng.c505 int chunk, n, tmp, ret = 0; in prng_tdes_read() local
533 chunk = min_t(int, nbytes, prng_chunk_size); in prng_tdes_read()
536 n = (chunk + 7) & -8; in prng_tdes_read()
567 if (copy_to_user(ubuf, prng_data->buf, chunk)) in prng_tdes_read()
570 nbytes -= chunk; in prng_tdes_read()
571 ret += chunk; in prng_tdes_read()
572 ubuf += chunk; in prng_tdes_read()
/linux-4.1.27/drivers/zorro/
Dzorro.c102 u32 chunk = start>>Z2RAM_CHUNKSHIFT; in mark_region() local
104 set_bit(chunk, zorro_unused_z2ram); in mark_region()
106 clear_bit(chunk, zorro_unused_z2ram); in mark_region()
/linux-4.1.27/net/rxrpc/
Dar-output.c547 size_t size, chunk, max, space; in rxrpc_send_data() local
566 chunk = max; in rxrpc_send_data()
567 if (chunk > msg_data_left(msg) && !more) in rxrpc_send_data()
568 chunk = msg_data_left(msg); in rxrpc_send_data()
570 space = chunk + call->conn->size_align; in rxrpc_send_data()
575 _debug("SIZE: %zu/%zu/%zu", chunk, space, size); in rxrpc_send_data()
594 sp->remain = chunk; in rxrpc_send_data()
/linux-4.1.27/Documentation/mtd/nand/
Dpxa3xx-nand.txt15 For instance, if we choose a 2048 data chunk and set "BCH" ECC (see below)
30 OOB, one per chunk read.
51 In the BCH mode the ECC code will be calculated for each transferred chunk
55 So, repeating the above scheme, a 2048B data chunk will be followed by 32B
/linux-4.1.27/init/
Ddo_mounts_md.c37 int chunk; member
101 md_setup_args[ent].chunk = 1 << (factor+12); in md_setup()
211 ainfo.chunk_size = md_setup_args[ent].chunk; in md_setup_drive()
/linux-4.1.27/drivers/usb/early/
Dehci-dbgp.c916 int chunk, ret; in early_dbgp_write() local
941 for (chunk = 0; chunk < DBGP_MAX_PACKET && n > 0; in early_dbgp_write()
942 str++, chunk++, n--) { in early_dbgp_write()
945 buf[chunk] = '\r'; in early_dbgp_write()
952 buf[chunk] = *str; in early_dbgp_write()
954 if (chunk > 0) { in early_dbgp_write()
956 dbgp_endpoint_out, buf, chunk); in early_dbgp_write()
/linux-4.1.27/security/keys/
Dkeyring.c233 unsigned long chunk = 0; in keyring_get_key_chunk() local
235 int desc_len = index_key->desc_len, n = sizeof(chunk); in keyring_get_key_chunk()
250 offset += sizeof(chunk) - 1; in keyring_get_key_chunk()
251 offset += (level - 3) * sizeof(chunk); in keyring_get_key_chunk()
259 chunk <<= 8; in keyring_get_key_chunk()
260 chunk |= ((u8*)index_key->description)[--offset]; in keyring_get_key_chunk()
264 chunk <<= 8; in keyring_get_key_chunk()
265 chunk |= (u8)((unsigned long)index_key->type >> in keyring_get_key_chunk()
268 return chunk; in keyring_get_key_chunk()
/linux-4.1.27/drivers/usb/
Dusb-skeleton.c289 size_t chunk = min(available, count); in skel_read() local
309 chunk)) in skel_read()
312 rv = chunk; in skel_read()
314 dev->bulk_in_copied += chunk; in skel_read()
321 skel_do_read_io(dev, count - chunk); in skel_read()
/linux-4.1.27/fs/nfs/blocklayout/
Ddev.c153 u64 chunk; in bl_map_stripe() local
157 chunk = div_u64(offset, dev->chunk_size); in bl_map_stripe()
158 div_u64_rem(chunk, dev->nr_children, &chunk_idx); in bl_map_stripe()
168 offset = chunk * dev->chunk_size; in bl_map_stripe()
/linux-4.1.27/drivers/net/ethernet/sfc/
Dmcdi.c1789 size_t chunk; in efx_mcdi_mtd_read() local
1793 chunk = min_t(size_t, end - offset, EFX_MCDI_NVRAM_LEN_MAX); in efx_mcdi_mtd_read()
1795 buffer, chunk); in efx_mcdi_mtd_read()
1798 offset += chunk; in efx_mcdi_mtd_read()
1799 buffer += chunk; in efx_mcdi_mtd_read()
1812 size_t chunk = part->common.mtd.erasesize; in efx_mcdi_mtd_erase() local
1827 chunk); in efx_mcdi_mtd_erase()
1830 offset += chunk; in efx_mcdi_mtd_erase()
1843 size_t chunk; in efx_mcdi_mtd_write() local
1854 chunk = min_t(size_t, end - offset, EFX_MCDI_NVRAM_LEN_MAX); in efx_mcdi_mtd_write()
[all …]
/linux-4.1.27/drivers/net/wireless/ath/ar5523/
Dar5523.c528 struct ar5523_chunk *chunk; in ar5523_data_rx_cb() local
549 chunk = (struct ar5523_chunk *) data->skb->data; in ar5523_data_rx_cb()
551 if (((chunk->flags & UATH_CFLAGS_FINAL) == 0) || in ar5523_data_rx_cb()
552 chunk->seqnum != 0) { in ar5523_data_rx_cb()
554 chunk->seqnum, chunk->flags, in ar5523_data_rx_cb()
555 be16_to_cpu(chunk->length)); in ar5523_data_rx_cb()
581 skb_reserve(data->skb, sizeof(*chunk)); in ar5523_data_rx_cb()
793 struct ar5523_chunk *chunk; in ar5523_tx_work_locked() local
836 chunk = (struct ar5523_chunk *)skb_push(skb, sizeof(*chunk)); in ar5523_tx_work_locked()
838 chunk->seqnum = 0; in ar5523_tx_work_locked()
[all …]
/linux-4.1.27/net/core/
Ddatagram.c710 int chunk = skb->len - hlen; in skb_copy_and_csum_datagram_msg() local
712 if (!chunk) in skb_copy_and_csum_datagram_msg()
715 if (msg_data_left(msg) < chunk) { in skb_copy_and_csum_datagram_msg()
718 if (skb_copy_datagram_msg(skb, hlen, msg, chunk)) in skb_copy_and_csum_datagram_msg()
723 chunk, &csum)) in skb_copy_and_csum_datagram_msg()
Dskbuff.c4377 unsigned long chunk; in alloc_skb_with_frags() local
4423 chunk = min_t(unsigned long, data_len, in alloc_skb_with_frags()
4425 skb_fill_page_desc(skb, i, page, 0, chunk); in alloc_skb_with_frags()
4426 data_len -= chunk; in alloc_skb_with_frags()
/linux-4.1.27/drivers/nfc/pn544/
Di2c.c677 struct pn544_i2c_fw_secure_frame *chunk; in pn544_hci_i2c_fw_secure_write_frame_cmd() local
684 chunk = (struct pn544_i2c_fw_secure_frame *) buf; in pn544_hci_i2c_fw_secure_write_frame_cmd()
686 chunk->cmd = PN544_FW_CMD_SECURE_CHUNK_WRITE; in pn544_hci_i2c_fw_secure_write_frame_cmd()
688 put_unaligned_be16(datalen, &chunk->be_datalen); in pn544_hci_i2c_fw_secure_write_frame_cmd()
690 memcpy(chunk->data, data, datalen); in pn544_hci_i2c_fw_secure_write_frame_cmd()
692 chunklen = sizeof(chunk->cmd) + sizeof(chunk->be_datalen) + datalen; in pn544_hci_i2c_fw_secure_write_frame_cmd()
/linux-4.1.27/sound/pci/lola/
Dlola_pcm.c326 int chunk; in setup_bdle() local
336 chunk = snd_pcm_sgbuf_get_chunk_size(substream, ofs, size); in setup_bdle()
337 bdl[2] = cpu_to_le32(chunk); in setup_bdle()
341 size -= chunk; in setup_bdle()
345 ofs += chunk; in setup_bdle()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_init.c1672 unsigned e, egrcnt, egrperchunk, chunk, egrsize, egroff; in qib_setup_eagerbufs() local
1689 chunk = rcd->rcvegrbuf_chunks; in qib_setup_eagerbufs()
1694 kzalloc_node(chunk * sizeof(rcd->rcvegrbuf[0]), in qib_setup_eagerbufs()
1701 kmalloc_node(chunk * sizeof(rcd->rcvegrbuf_phys[0]), in qib_setup_eagerbufs()
1723 for (e = chunk = 0; chunk < rcd->rcvegrbuf_chunks; chunk++) { in qib_setup_eagerbufs()
1724 dma_addr_t pa = rcd->rcvegrbuf_phys[chunk]; in qib_setup_eagerbufs()
1728 memset(rcd->rcvegrbuf[chunk], 0, size); in qib_setup_eagerbufs()
Dqib_driver.c286 const u32 chunk = etail >> rcd->rcvegrbufs_perchunk_shift; in qib_get_egrbuf() local
289 return rcd->rcvegrbuf[chunk] + (idx << rcd->dd->rcvegrbufsize_shift); in qib_get_egrbuf()
/linux-4.1.27/net/ipv4/
Dtcp.c1741 int chunk; in tcp_recvmsg() local
1745 if ((chunk = len - tp->ucopy.len) != 0) { in tcp_recvmsg()
1746 NET_ADD_STATS_USER(sock_net(sk), LINUX_MIB_TCPDIRECTCOPYFROMBACKLOG, chunk); in tcp_recvmsg()
1747 len -= chunk; in tcp_recvmsg()
1748 copied += chunk; in tcp_recvmsg()
1756 if ((chunk = len - tp->ucopy.len) != 0) { in tcp_recvmsg()
1757 NET_ADD_STATS_USER(sock_net(sk), LINUX_MIB_TCPDIRECTCOPYFROMPREQUEUE, chunk); in tcp_recvmsg()
1758 len -= chunk; in tcp_recvmsg()
1759 copied += chunk; in tcp_recvmsg()
1836 int chunk; in tcp_recvmsg() local
[all …]
Dtcp_input.c4516 int chunk = min_t(unsigned int, skb->len, in tcp_data_queue() local
4522 if (!skb_copy_datagram_msg(skb, 0, tp->ucopy.msg, chunk)) { in tcp_data_queue()
4523 tp->ucopy.len -= chunk; in tcp_data_queue()
4524 tp->copied_seq += chunk; in tcp_data_queue()
4525 eaten = (chunk == skb->len); in tcp_data_queue()
5039 int chunk = skb->len - hlen; in tcp_copy_to_iovec() local
5044 err = skb_copy_datagram_msg(skb, hlen, tp->ucopy.msg, chunk); in tcp_copy_to_iovec()
5049 tp->ucopy.len -= chunk; in tcp_copy_to_iovec()
5050 tp->copied_seq += chunk; in tcp_copy_to_iovec()
Dtcp_output.c1810 u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); in tcp_tso_should_defer() local
1815 chunk /= win_divisor; in tcp_tso_should_defer()
1816 if (limit >= chunk) in tcp_tso_should_defer()
/linux-4.1.27/drivers/char/
Dmem.c639 size_t chunk = iov_iter_count(iter), n; in read_iter_zero() local
641 if (chunk > PAGE_SIZE) in read_iter_zero()
642 chunk = PAGE_SIZE; /* Just for latency reasons */ in read_iter_zero()
643 n = iov_iter_zero(chunk, iter); in read_iter_zero()
Drandom.c1264 int chunk = min(nbytes, (int)sizeof(unsigned long)); in get_random_bytes_arch() local
1269 memcpy(p, &v, chunk); in get_random_bytes_arch()
1270 p += chunk; in get_random_bytes_arch()
1271 nbytes -= chunk; in get_random_bytes_arch()
/linux-4.1.27/net/caif/
Dcaif_socket.c376 int chunk; in caif_stream_recvmsg() local
427 chunk = min_t(unsigned int, skb->len, size); in caif_stream_recvmsg()
428 if (memcpy_to_msg(msg, skb->data, chunk)) { in caif_stream_recvmsg()
434 copied += chunk; in caif_stream_recvmsg()
435 size -= chunk; in caif_stream_recvmsg()
439 skb_pull(skb, chunk); in caif_stream_recvmsg()
/linux-4.1.27/Documentation/fb/
Dsh7760fb.txt11 * Framebuffer memory must be a large chunk allocated at the top
14 get a large enough contiguous chunk of memory.
/linux-4.1.27/Documentation/
Ddell_rbu.txt20 image methods. In case of monolithic the driver allocates a contiguous chunk
57 The user creates packets header, gets the chunk of the BIOS image and
58 places it next to the packetheader; now, the packetheader + BIOS image chunk
Dassoc_array.txt139 (1) Get a chunk of index key from caller data:
143 This should return a chunk of caller-supplied index key starting at the
149 (2) Get a chunk of an object's index key.
360 The index key is read in chunks of machine word. Each chunk is subdivided into
Ddma-buf-sharing.txt278 PAGE_SIZE size. Before accessing a chunk it needs to be mapped, which returns
279 a pointer in kernel virtual address space. Afterwards the chunk needs to be
280 unmapped again. There is no limit on how often a given chunk can be mapped
282 before mapping the same chunk again.
DDMA-attributes.txt100 be mapped as contiguous chunk into device dma address space. By
Dramoops.txt32 power of two) and each oops/panic writes a "record_size" chunk of
Dmd.txt12 md=<md device no.>,<raid level>,<chunk size factor>,<fault level>,dev0,dev1,...,devn
30 chunk size factor = (raid-0 and raid-1 only)
31 Set the chunk size as 4k << n.
309 The size, in bytes, of the chunk which will be represented by a
Dxillybus.txt87 their attributes), there isn't one specific chunk of logic being the Xillybus
DHOWTO198 If you already have a chunk of code that you want to put into the kernel
Dkernel-parameters.txt845 chunk of memory for debug pagealloc if we don't enable
1437 ahash performance varies for different chunk sizes on
2150 used for mtrr cleanup. It is largest continuous chunk
2847 percpu_alloc= Select which percpu first chunk allocator to use.
DDMA-API-HOWTO.txt414 exists (for example) to guarantee that if you allocate a chunk
/linux-4.1.27/drivers/staging/media/lirc/
Dlirc_serial.c399 unsigned char chunk, shifted; in send_pulse_irdeo() local
404 chunk = 3; in send_pulse_irdeo()
406 chunk = 1; in send_pulse_irdeo()
408 shifted = chunk << (i * 3); in send_pulse_irdeo()
/linux-4.1.27/include/drm/
Ddrm_dp_mst_helper.h206 u8 chunk[48]; member
361 u8 chunk[48]; member
/linux-4.1.27/drivers/tty/
Dtty_io.c1114 unsigned int chunk; in do_tty_write() local
1136 chunk = 2048; in do_tty_write()
1138 chunk = 65536; in do_tty_write()
1139 if (count < chunk) in do_tty_write()
1140 chunk = count; in do_tty_write()
1143 if (tty->write_cnt < chunk) { in do_tty_write()
1146 if (chunk < 1024) in do_tty_write()
1147 chunk = 1024; in do_tty_write()
1149 buf_chunk = kmalloc(chunk, GFP_KERNEL); in do_tty_write()
1155 tty->write_cnt = chunk; in do_tty_write()
[all …]
/linux-4.1.27/drivers/irqchip/
Dirq-gic-v3-its.c688 static int its_chunk_to_lpi(int chunk) in its_chunk_to_lpi() argument
690 return (chunk << IRQS_PER_CHUNK_SHIFT) + 8192; in its_chunk_to_lpi()
757 int chunk = its_lpi_to_chunk(lpi); in its_lpi_free() local
758 BUG_ON(chunk > lpi_chunks); in its_lpi_free()
759 if (test_bit(chunk, lpi_bitmap)) { in its_lpi_free()
760 clear_bit(chunk, lpi_bitmap); in its_lpi_free()
762 pr_err("Bad LPI chunk %d\n", chunk); in its_lpi_free()
/linux-4.1.27/arch/xtensa/lib/
Dmemcopy.S151 add a8, a8, a3 # a8 = end of last 16B source chunk
227 add a10, a10, a3 # a10 = end of last 16B source chunk
425 sub a8, a3, a8 # a8 = start of first 16B source chunk
502 sub a10, a3, a10 # a10 = start of first 16B source chunk
Dusercopy.S161 add a8, a8, a3 # a8 = end of last 16B source chunk
226 add a10, a10, a3 # a10 = end of last 16B source chunk
Dmemset.S73 add a6, a6, a5 # a6 = end of last 16B chunk
Dstrnlen_user.S77 add a10, a10, a4 # a10 = end of last 4B chunk
/linux-4.1.27/fs/gfs2/
Dbmap.c1013 u64 chunk; in gfs2_journaled_truncate() local
1017 chunk = oldsize - newsize; in gfs2_journaled_truncate()
1018 if (chunk > max_chunk) in gfs2_journaled_truncate()
1019 chunk = max_chunk; in gfs2_journaled_truncate()
1020 truncate_pagecache(inode, oldsize - chunk); in gfs2_journaled_truncate()
1021 oldsize -= chunk; in gfs2_journaled_truncate()
/linux-4.1.27/drivers/memstick/core/
Dmspro_block.c697 static int mspro_block_issue_req(struct memstick_dev *card, int chunk) in mspro_block_issue_req() argument
704 while (chunk) { in mspro_block_issue_req()
712 chunk = __blk_end_request_cur(msb->block_req, -ENOMEM); in mspro_block_issue_req()
739 chunk = 1; in mspro_block_issue_req()
746 int chunk, cnt; in mspro_block_complete_req() local
779 chunk = __blk_end_request(msb->block_req, error, t_len); in mspro_block_complete_req()
781 error = mspro_block_issue_req(card, chunk); in mspro_block_complete_req()
/linux-4.1.27/drivers/gpu/drm/radeon/
Dradeon_cs.c78 struct radeon_cs_chunk *chunk; in radeon_cs_parser_relocs() local
87 chunk = p->chunk_relocs; in radeon_cs_parser_relocs()
90 p->nrelocs = chunk->length_dw / 4; in radeon_cs_parser_relocs()
103 r = (struct drm_radeon_cs_reloc *)&chunk->kdata[i*4]; in radeon_cs_parser_relocs()
/linux-4.1.27/arch/arm/kernel/
Dtraps.c540 unsigned long chunk = min(PAGE_SIZE, end - start); in __do_cache_op() local
545 ret = flush_cache_user_range(start, start + chunk); in __do_cache_op()
550 start += chunk; in __do_cache_op()
/linux-4.1.27/drivers/thermal/ti-soc-thermal/
Dti-bandgap.c1239 void __iomem *chunk; in ti_bandgap_build() local
1244 chunk = devm_ioremap_resource(&pdev->dev, res); in ti_bandgap_build()
1246 bgp->base = chunk; in ti_bandgap_build()
1247 if (IS_ERR(chunk)) in ti_bandgap_build()
1248 return ERR_CAST(chunk); in ti_bandgap_build()
/linux-4.1.27/drivers/staging/lustre/lustre/osc/
Dosc_cache.c632 pgoff_t chunk; in osc_extent_find() local
652 chunk = index >> ppc_bits; in osc_extent_find()
687 if (chunk > ext_chk_end + 1) in osc_extent_find()
700 if (chunk + 1 < ext_chk_start) { in osc_extent_find()
753 if (chunk + 1 == ext_chk_start) { in osc_extent_find()
763 } else if (chunk == ext_chk_end + 1) { in osc_extent_find()
1144 pgoff_t chunk = index >> ppc_bits; in osc_extent_expand() local
1154 if (chunk > end_chunk + 1) { in osc_extent_expand()
1159 if (end_chunk >= chunk) { in osc_extent_expand()
1164 LASSERT(end_chunk + 1 == chunk); in osc_extent_expand()
[all …]
/linux-4.1.27/sound/pci/hda/
Dhda_controller.c299 int chunk; in setup_bdle() local
309 chunk = snd_sgbuf_get_chunk_size(dmab, ofs, size); in setup_bdle()
313 if (chunk > remain) in setup_bdle()
314 chunk = remain; in setup_bdle()
316 bdl[2] = cpu_to_le32(chunk); in setup_bdle()
320 size -= chunk; in setup_bdle()
324 ofs += chunk; in setup_bdle()
/linux-4.1.27/arch/arm/lib/
Dmemzero.S42 cmp r1, #16 @ 1 we can skip this chunk if we
/linux-4.1.27/drivers/mmc/host/
Dsdhci.c318 size_t blksize, len, chunk; in sdhci_read_block_pio() local
325 chunk = 0; in sdhci_read_block_pio()
341 if (chunk == 0) { in sdhci_read_block_pio()
343 chunk = 4; in sdhci_read_block_pio()
350 chunk--; in sdhci_read_block_pio()
363 size_t blksize, len, chunk; in sdhci_write_block_pio() local
370 chunk = 0; in sdhci_write_block_pio()
387 scratch |= (u32)*buf << (chunk * 8); in sdhci_write_block_pio()
390 chunk++; in sdhci_write_block_pio()
393 if ((chunk == 4) || ((len == 0) && (blksize == 0))) { in sdhci_write_block_pio()
[all …]
/linux-4.1.27/drivers/macintosh/
Dsmu.c899 unsigned int chunk; in smu_read_datablock() local
907 chunk = 0xe; in smu_read_datablock()
910 unsigned int clen = min(len, chunk); in smu_read_datablock()
915 cmd.reply_len = chunk; in smu_read_datablock()
/linux-4.1.27/arch/powerpc/kernel/
Dvio.c393 size_t avail = 0, level, chunk, need; in vio_cmo_balance() local
441 chunk = min_t(size_t, avail, VIO_CMO_BALANCE_CHUNK); in vio_cmo_balance()
442 chunk = min(chunk, (viodev->cmo.desired - in vio_cmo_balance()
444 viodev->cmo.entitled += chunk; in vio_cmo_balance()
Dprom_init.c1988 unsigned long room, chunk; in make_room() local
1998 chunk = alloc_up(room, 0); in make_room()
1999 if (chunk == 0) in make_room()
2002 *mem_end = chunk + room; in make_room()
/linux-4.1.27/net/unix/
Daf_unix.c2104 int chunk; in unix_stream_recvmsg() local
2178 chunk = min_t(unsigned int, unix_skb_len(skb) - skip, size); in unix_stream_recvmsg()
2180 msg, chunk)) { in unix_stream_recvmsg()
2185 copied += chunk; in unix_stream_recvmsg()
2186 size -= chunk; in unix_stream_recvmsg()
2190 UNIXCB(skb).consumed += chunk; in unix_stream_recvmsg()
2192 sk_peek_offset_bwd(sk, chunk); in unix_stream_recvmsg()
2211 sk_peek_offset_fwd(sk, chunk); in unix_stream_recvmsg()
/linux-4.1.27/drivers/net/wimax/i2400m/
Dfw.c646 static int i2400m_download_chunk(struct i2400m *i2400m, const void *chunk, in i2400m_download_chunk() argument
660 "direct %u do_csum %u)\n", i2400m, chunk, __chunk_len, in i2400m_download_chunk()
663 memcpy(buf->cmd_payload, chunk, __chunk_len); in i2400m_download_chunk()
676 "direct %u do_csum %u) = %d\n", i2400m, chunk, __chunk_len, in i2400m_download_chunk()
/linux-4.1.27/drivers/crypto/qat/qat_common/
Dqat_uclo.c522 char *chunk; in qat_uclo_map_chunk() local
530 chunk = buf + file_chunk->offset; in qat_uclo_map_chunk()
532 chunk, file_chunk->size)) in qat_uclo_map_chunk()
537 obj_hdr->file_buff = chunk; in qat_uclo_map_chunk()
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_file_ops.c897 unsigned e, egrcnt, egrperchunk, chunk, egrsize, egroff; in ipath_create_user_egr() local
917 chunk = pd->port_rcvegrbuf_chunks; in ipath_create_user_egr()
920 pd->port_rcvegrbuf = kmalloc(chunk * sizeof(pd->port_rcvegrbuf[0]), in ipath_create_user_egr()
927 kmalloc(chunk * sizeof(pd->port_rcvegrbuf_phys[0]), in ipath_create_user_egr()
947 for (e = chunk = 0; chunk < pd->port_rcvegrbuf_chunks; chunk++) { in ipath_create_user_egr()
948 dma_addr_t pa = pd->port_rcvegrbuf_phys[chunk]; in ipath_create_user_egr()
/linux-4.1.27/drivers/media/dvb-frontends/
Ddrxk_hard.h224 u8 chunk[256]; member
Ddrxk_hard.c449 int chunk = blk_size > state->m_chunk_size ? in write_block() local
451 u8 *adr_buf = &state->chunk[0]; in write_block()
461 if (chunk == state->m_chunk_size) in write_block()
462 chunk -= 2; in write_block()
469 memcpy(&state->chunk[adr_length], p_block, chunk); in write_block()
474 for (i = 0; i < chunk; i++) in write_block()
479 &state->chunk[0], chunk + adr_length); in write_block()
485 p_block += chunk; in write_block()
486 address += (chunk >> 1); in write_block()
487 blk_size -= chunk; in write_block()
/linux-4.1.27/drivers/input/misc/
Dims-pcu.c515 u8 command, int chunk, int len) in ims_pcu_send_cmd_chunk() argument
527 command, chunk, error); in ims_pcu_send_cmd_chunk()
538 int chunk = 0; in ims_pcu_send_command() local
566 ++chunk, count); in ims_pcu_send_command()
584 error = ims_pcu_send_cmd_chunk(pcu, command, ++chunk, count); in ims_pcu_send_command()
597 return ims_pcu_send_cmd_chunk(pcu, command, ++chunk, count); in ims_pcu_send_command()
/linux-4.1.27/net/irda/
Daf_irda.c1430 int chunk; in irda_recvmsg_stream() local
1470 chunk = min_t(unsigned int, skb->len, size); in irda_recvmsg_stream()
1471 if (memcpy_to_msg(msg, skb->data, chunk)) { in irda_recvmsg_stream()
1477 copied += chunk; in irda_recvmsg_stream()
1478 size -= chunk; in irda_recvmsg_stream()
1482 skb_pull(skb, chunk); in irda_recvmsg_stream()
/linux-4.1.27/net/decnet/
Daf_decnet.c1760 unsigned int chunk = skb->len; in dn_recvmsg() local
1763 if ((chunk + copied) > size) in dn_recvmsg()
1764 chunk = size - copied; in dn_recvmsg()
1766 if (memcpy_to_msg(msg, skb->data, chunk)) { in dn_recvmsg()
1770 copied += chunk; in dn_recvmsg()
1773 skb_pull(skb, chunk); in dn_recvmsg()
/linux-4.1.27/drivers/gpu/drm/
Ddrm_dp_mst_topology.c350 memcpy(&msg->chunk[0], replybuf + hdrlen, msg->curchunk_idx); in drm_dp_sideband_msg_build()
352 memcpy(&msg->chunk[msg->curchunk_idx], replybuf, replybuflen); in drm_dp_sideband_msg_build()
358 crc4 = drm_dp_msg_data_crc4(msg->chunk, msg->curchunk_len - 1); in drm_dp_sideband_msg_build()
360 memcpy(&msg->msg[msg->curlen], msg->chunk, msg->curchunk_len - 1); in drm_dp_sideband_msg_build()
1431 u8 chunk[48]; in process_single_tx_qlock() local
1463 drm_dp_encode_sideband_msg_hdr(&hdr, chunk, &idx); in process_single_tx_qlock()
1464 memcpy(&chunk[idx], &txmsg->msg[txmsg->cur_offset], tosend); in process_single_tx_qlock()
1466 drm_dp_crc_sideband_chunk_req(&chunk[idx], tosend); in process_single_tx_qlock()
1469 ret = drm_dp_send_sideband_msg(mgr, up, chunk, idx); in process_single_tx_qlock()
/linux-4.1.27/drivers/net/ethernet/ibm/emac/
Dcore.c1421 int chunk = min(len, MAL_MAX_TX_SIZE); in emac_xmit_split() local
1422 len -= chunk; in emac_xmit_split()
1433 dev->tx_desc[slot].data_len = (u16) chunk; in emac_xmit_split()
1440 pd += chunk; in emac_xmit_split()
1450 int len = skb->len, chunk; in emac_start_xmit_sg() local
1474 chunk = min(len, MAL_MAX_TX_SIZE); in emac_start_xmit_sg()
1477 dev->tx_desc[slot].data_len = (u16) chunk; in emac_start_xmit_sg()
1478 len -= chunk; in emac_start_xmit_sg()
1480 slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags, in emac_start_xmit_sg()
/linux-4.1.27/security/apparmor/
Dpolicy_unpack.c124 static size_t unpack_u16_chunk(struct aa_ext *e, char **chunk) in unpack_u16_chunk() argument
134 *chunk = e->pos; in unpack_u16_chunk()
/linux-4.1.27/Documentation/dmaengine/
Dprovider.txt69 but all of them will require, for every chunk, at least the source and
188 scatter-gather transfer, with a single chunk to transfer, it's a
204 deal with a single chunk to copy or a collection of them, here,
212 - A cyclic transfer is a transfer where the chunk collection will
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/socklnd/
Dsocklnd_cb.c2583 int chunk = ksocknal_data.ksnd_peer_hash_size; in ksocknal_reaper() local
2593 chunk = (chunk * n * p) / in ksocknal_reaper()
2595 if (chunk == 0) in ksocknal_reaper()
2596 chunk = 1; in ksocknal_reaper()
2598 for (i = 0; i < chunk; i++) { in ksocknal_reaper()
/linux-4.1.27/Documentation/ABI/testing/
Dsysfs-driver-wacom72 byte chunk encodes the image data for two consecutive lines on
Dsysfs-block122 chunk size. A properly aligned multiple of
/linux-4.1.27/Documentation/ko_KR/
DHOWTO491 리눅스 커널 커뮤니티는 한꺼번에 굉장히 큰 코드의 묶음(chunk)을 쉽게
/linux-4.1.27/Documentation/arm/Samsung-S3C24XX/
DSuspend.txt128 Defines the size of memory each CRC chunk covers. A smaller value
/linux-4.1.27/Documentation/filesystems/cifs/
DTODO14 - T10 copy offload (copy chunk is only mechanism supported)
/linux-4.1.27/Documentation/filesystems/
Dntfs.txt318 For a striped volume, i.e. raid level 0, you will need to know the chunk size
394 Windows by default uses a stripe chunk size of 64k, so you probably want the
395 "chunk-size 64k" option for each raid-disk, too.
405 chunk-size 64k
Dsysv-fs.txt84 to the free blocks pertaining to the next chunk; the first of these
Dbtrfs.txt141 Specify that 1 metadata chunk should be allocated after every <value>
Dext4.txt264 disks * RAID chunk size in file system blocks.
/linux-4.1.27/crypto/
Ddrbg.c1449 unsigned int chunk = 0; in drbg_generate_long() local
1451 chunk = slice ? drbg_max_request_bytes(drbg) : (buflen - len); in drbg_generate_long()
1452 tmplen = drbg_generate(drbg, buf + len, chunk, addtl); in drbg_generate_long()
/linux-4.1.27/arch/s390/kernel/
Dhead.S108 ssch 0(%r3) # load chunk of 1600 bytes
/linux-4.1.27/drivers/acpi/apei/
Dghes.c202 struct gen_pool_chunk *chunk, in ghes_estatus_pool_free_chunk_page() argument
205 free_page(chunk->start_addr); in ghes_estatus_pool_free_chunk_page()
/linux-4.1.27/arch/m68k/
DKconfig.cpu361 bool "Use one physical chunk of memory only" if ADVANCED && !SUN3
366 Ignore all but the first contiguous chunk of physical memory for VM
/linux-4.1.27/drivers/mtd/nand/
Dnand_base.c1773 int chunk = chip->ecc.bytes + chip->ecc.prepad + chip->ecc.postpad; in nand_read_oob_syndrome() local
1781 pos = eccsize + i * (eccsize + chunk); in nand_read_oob_syndrome()
1788 toread = min_t(int, length, chunk); in nand_read_oob_syndrome()
1832 int chunk = chip->ecc.bytes + chip->ecc.prepad + chip->ecc.postpad; in nand_write_oob_syndrome() local
1843 pos = steps * (eccsize + chunk); in nand_write_oob_syndrome()
1862 pos = eccsize + i * (eccsize + chunk); in nand_write_oob_syndrome()
1867 len = min_t(int, length, chunk); in nand_write_oob_syndrome()
/linux-4.1.27/arch/sh/lib64/
Dmemcpy.S30 ! bytes memory chunk to b copied, the rest of the word can be read
Dcopy_user_memcpy.S28 ! bytes memory chunk to b copied, the rest of the word can be read
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/
Dcxgb3_main.c2455 unsigned int chunk = in cxgb_extension_ioctl() local
2459 t3_mc7_bd_read(mem, t.addr / 8, chunk / 8, in cxgb_extension_ioctl()
2463 if (copy_to_user(useraddr, buf, chunk)) in cxgb_extension_ioctl()
2465 useraddr += chunk; in cxgb_extension_ioctl()
2466 t.addr += chunk; in cxgb_extension_ioctl()
2467 t.len -= chunk; in cxgb_extension_ioctl()
Dsge.c326 unsigned int chunk) in reclaim_completed_tx() argument
330 reclaim = min(chunk, reclaim); in reclaim_completed_tx()
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c3172 int chunk = kiblnd_data.kib_peer_hash_size; in kiblnd_connd() local
3186 chunk = (chunk * n * p) / in kiblnd_connd()
3188 if (chunk == 0) in kiblnd_connd()
3189 chunk = 1; in kiblnd_connd()
3191 for (i = 0; i < chunk; i++) { in kiblnd_connd()
/linux-4.1.27/fs/cramfs/
DREADME58 applied to the i'th blksize-sized chunk of the input data.
/linux-4.1.27/Documentation/early-userspace/
DREADME17 - initramfs, a chunk of code that unpacks the compressed cpio image
/linux-4.1.27/drivers/net/ethernet/marvell/
Dskge.c2530 u32 chunk, ram_addr; in skge_up() local
2595 chunk = (hw->ram_size - hw->ram_offset) / (hw->ports * 2); in skge_up()
2596 ram_addr = hw->ram_offset + 2 * chunk * port; in skge_up()
2598 skge_ramset(hw, rxqaddr[port], ram_addr, chunk); in skge_up()
2602 skge_ramset(hw, txqaddr[port], ram_addr+chunk, chunk); in skge_up()
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/ce/fuc/
Dcom.fuc208 // get a chunk of stack space, aligned to 256 byte boundary
419 // zero out a chunk of the stack to store the swizzle into
/linux-4.1.27/arch/cris/arch-v10/
DREADME.mm228 happens first is that a virtual address chunk is allocated to the request using
/linux-4.1.27/drivers/scsi/
Dips.h733 IPS_CHUNK chunk[IPS_MAX_CHUNKS]; member
/linux-4.1.27/Documentation/powerpc/
Dfirmware-assisted-dump.txt53 NOTE: The term 'boot memory' means size of the low memory chunk
/linux-4.1.27/drivers/net/wireless/ath/ath10k/
Dwmi.c3774 struct host_memory_chunk *chunk; in ath10k_wmi_put_host_mem_chunks() local
3780 chunk = &chunks->items[i]; in ath10k_wmi_put_host_mem_chunks()
3781 chunk->ptr = __cpu_to_le32(ar->wmi.mem_chunks[i].paddr); in ath10k_wmi_put_host_mem_chunks()
3782 chunk->size = __cpu_to_le32(ar->wmi.mem_chunks[i].len); in ath10k_wmi_put_host_mem_chunks()
3783 chunk->req_id = __cpu_to_le32(ar->wmi.mem_chunks[i].req_id); in ath10k_wmi_put_host_mem_chunks()
/linux-4.1.27/Documentation/scsi/
Dst.txt215 (PAGE_SIZE << ST_FIRST_ORDER) bytes if the system can give a chunk of
220 buffer as one chunk.
/linux-4.1.27/fs/ext4/
Dmballoc.c670 ext4_grpblk_t chunk; in ext4_mb_mark_free_simple() local
686 chunk = 1 << min; in ext4_mb_mark_free_simple()
694 len -= chunk; in ext4_mb_mark_free_simple()
695 first += chunk; in ext4_mb_mark_free_simple()
/linux-4.1.27/Documentation/usb/
Dusbmon.txt263 The size may be rounded down to the next chunk (or page). If the requested
DWUSB-Design-overview.txt372 we issue another URB to read into the destination buffer the chunk of
/linux-4.1.27/scripts/
Dcheckpatch.pl4489 for my $chunk (@chunks) {
4490 my ($cond, $block) = @{$chunk};
/linux-4.1.27/drivers/net/wireless/ipw2x00/
Dipw2200.c3211 struct fw_chunk *chunk; in ipw_load_firmware() local
3251 chunk = (struct fw_chunk *)(data + offset); in ipw_load_firmware()
3253 chunk_len = le32_to_cpu(chunk->length); in ipw_load_firmware()
3278 nr, le32_to_cpu(chunk->address), in ipw_load_firmware()
/linux-4.1.27/Documentation/cgroups/
Dcpusets.txt561 4 : search nodes in a chunk of node [on NUMA system]
/linux-4.1.27/net/netfilter/
DKconfig1349 and SCTP chunk types.
/linux-4.1.27/Documentation/block/
Dbiodoc.txt378 which were generated for each such chunk.
/linux-4.1.27/Documentation/networking/
Dip-sysctl.txt1769 a listening sctp socket to a connecting client in the INIT-ACK chunk.
/linux-4.1.27/drivers/video/fbdev/
DKconfig2226 This is a `virtual' frame buffer device. It operates on a chunk of