/linux-4.4.14/net/sctp/ |
D | inqueue.c | 59 struct sctp_chunk *chunk, *tmp; in sctp_inq_free() local 62 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) { in sctp_inq_free() 63 list_del_init(&chunk->list); in sctp_inq_free() 64 sctp_chunk_free(chunk); in sctp_inq_free() 79 void sctp_inq_push(struct sctp_inq *q, struct sctp_chunk *chunk) in sctp_inq_push() argument 82 if (chunk->rcvr->dead) { in sctp_inq_push() 83 sctp_chunk_free(chunk); in sctp_inq_push() 92 list_add_tail(&chunk->list, &q->in_chunk_list); in sctp_inq_push() 93 if (chunk->asoc) in sctp_inq_push() 94 chunk->asoc->stats.ipackets++; in sctp_inq_push() [all …]
|
D | output.c | 61 struct sctp_chunk *chunk); 63 struct sctp_chunk *chunk); 65 struct sctp_chunk *chunk); 67 struct sctp_chunk *chunk, 87 struct sctp_chunk *chunk = NULL; in sctp_packet_config() local 94 chunk = sctp_get_ecne_prepend(packet->transport->asoc); in sctp_packet_config() 99 if (chunk) in sctp_packet_config() 100 sctp_packet_append_chunk(packet, chunk); in sctp_packet_config() 137 struct sctp_chunk *chunk, *tmp; in sctp_packet_free() local 141 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free() [all …]
|
D | outqueue.c | 218 struct sctp_chunk *chunk, *tmp; in __sctp_outq_teardown() local 224 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 227 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 228 sctp_chunk_free(chunk); in __sctp_outq_teardown() 235 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 237 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 238 sctp_chunk_free(chunk); in __sctp_outq_teardown() 244 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 246 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 247 sctp_chunk_free(chunk); in __sctp_outq_teardown() [all …]
|
D | chunk.c | 75 struct sctp_chunk *chunk; in sctp_datamsg_free() local 80 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free() 81 sctp_chunk_free(chunk); in sctp_datamsg_free() 90 struct sctp_chunk *chunk; in sctp_datamsg_destroy() local 102 chunk = list_entry(pos, struct sctp_chunk, frag_list); in sctp_datamsg_destroy() 105 asoc = chunk->asoc; in sctp_datamsg_destroy() 119 if (chunk->has_tsn) in sctp_datamsg_destroy() 124 ev = sctp_ulpevent_make_send_failed(asoc, chunk, sent, in sctp_datamsg_destroy() 130 sctp_chunk_put(chunk); in sctp_datamsg_destroy() 151 static void sctp_datamsg_assign(struct sctp_datamsg *msg, struct sctp_chunk *chunk) in sctp_datamsg_assign() argument [all …]
|
D | sm_statefuns.c | 65 struct sctp_chunk *chunk, 69 struct sctp_chunk *chunk, 73 const struct sctp_chunk *chunk); 77 const struct sctp_chunk *chunk, 98 static struct sctp_sackhdr *sctp_sm_pull_sack(struct sctp_chunk *chunk); 151 struct sctp_chunk *chunk); 168 sctp_chunk_length_valid(struct sctp_chunk *chunk, in sctp_chunk_length_valid() argument 171 __u16 chunk_length = ntohs(chunk->chunk_hdr->length); in sctp_chunk_length_valid() 174 if (unlikely(chunk->pdiscard)) in sctp_chunk_length_valid() 224 struct sctp_chunk *chunk = arg; in sctp_sf_do_4_C() local [all …]
|
D | sm_make_chunk.c | 79 static void *sctp_addto_param(struct sctp_chunk *chunk, int len, 90 static void sctp_control_set_owner_w(struct sctp_chunk *chunk) in sctp_control_set_owner_w() argument 92 struct sctp_association *asoc = chunk->asoc; in sctp_control_set_owner_w() 93 struct sk_buff *skb = chunk->skb; in sctp_control_set_owner_w() 107 int sctp_chunk_iif(const struct sctp_chunk *chunk) in sctp_chunk_iif() argument 112 af = sctp_get_af_specific(ipver2af(ip_hdr(chunk->skb)->version)); in sctp_chunk_iif() 114 iif = af->skb_iif(chunk->skb); in sctp_chunk_iif() 137 void sctp_init_cause(struct sctp_chunk *chunk, __be16 cause_code, in sctp_init_cause() argument 147 chunk->subh.err_hdr = sctp_addto_chunk(chunk, sizeof(sctp_errhdr_t), &err); in sctp_init_cause() 155 static int sctp_init_cause_fixed(struct sctp_chunk *chunk, __be16 cause_code, in sctp_init_cause_fixed() argument [all …]
|
D | sm_sideeffect.c | 102 struct sctp_chunk *chunk) in sctp_do_ecn_ecne_work() argument 134 repl = sctp_make_cwr(asoc, asoc->last_cwr_tsn, chunk); in sctp_do_ecn_ecne_work() 571 struct sctp_chunk *chunk, in sctp_cmd_assoc_failed() argument 579 if (event_type == SCTP_EVENT_T_CHUNK && subtype.chunk == SCTP_CID_ABORT) in sctp_cmd_assoc_failed() 581 (__u16)error, 0, 0, chunk, in sctp_cmd_assoc_failed() 592 abort = sctp_make_violation_max_retrans(asoc, chunk); in sctp_cmd_assoc_failed() 613 struct sctp_chunk *chunk, in sctp_cmd_process_init() argument 624 if (!sctp_process_init(asoc, chunk, sctp_source(chunk), peer_init, gfp)) in sctp_cmd_process_init() 690 struct sctp_chunk *chunk) in sctp_cmd_transport_on() argument 746 hbinfo = (sctp_sender_hb_info_t *) chunk->skb->data; in sctp_cmd_transport_on() [all …]
|
D | endpointola.c | 413 struct sctp_chunk *chunk; in sctp_endpoint_bh_rcv() local 428 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_endpoint_bh_rcv() 429 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_endpoint_bh_rcv() 434 if (first_time && (subtype.chunk == SCTP_CID_AUTH)) { in sctp_endpoint_bh_rcv() 447 chunk->auth_chunk = skb_clone(chunk->skb, in sctp_endpoint_bh_rcv() 449 chunk->auth = 1; in sctp_endpoint_bh_rcv() 460 if (NULL == chunk->asoc) { in sctp_endpoint_bh_rcv() 462 sctp_source(chunk), in sctp_endpoint_bh_rcv() 464 chunk->asoc = asoc; in sctp_endpoint_bh_rcv() 465 chunk->transport = transport; in sctp_endpoint_bh_rcv() [all …]
|
D | ulpevent.c | 129 __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) in sctp_ulpevent_make_assoc_change() argument 138 if (chunk) { in sctp_ulpevent_make_assoc_change() 142 skb = skb_copy_expand(chunk->skb, in sctp_ulpevent_make_assoc_change() 158 ntohs(chunk->chunk_hdr->length) - in sctp_ulpevent_make_assoc_change() 371 struct sctp_chunk *chunk, __u16 flags, in sctp_ulpevent_make_remote_error() argument 381 ch = (sctp_errhdr_t *)(chunk->skb->data); in sctp_ulpevent_make_remote_error() 386 skb_pull(chunk->skb, sizeof(sctp_errhdr_t)); in sctp_ulpevent_make_remote_error() 391 skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp); in sctp_ulpevent_make_remote_error() 394 skb_pull(chunk->skb, elen); in sctp_ulpevent_make_remote_error() 427 const struct sctp_association *asoc, struct sctp_chunk *chunk, in sctp_ulpevent_make_send_failed() argument [all …]
|
D | input.c | 100 struct sctp_chunk *chunk; member 114 struct sctp_chunk *chunk; in sctp_rcv() local 224 chunk = sctp_chunkify(skb, asoc, sk); in sctp_rcv() 225 if (!chunk) in sctp_rcv() 227 SCTP_INPUT_CB(skb)->chunk = chunk; in sctp_rcv() 230 chunk->rcvr = rcvr; in sctp_rcv() 233 chunk->sctp_hdr = sh; in sctp_rcv() 236 sctp_init_addrs(chunk, &src, &dest); in sctp_rcv() 239 chunk->transport = transport; in sctp_rcv() 263 sctp_chunk_free(chunk); in sctp_rcv() [all …]
|
D | debug.c | 65 if (cid.chunk <= SCTP_CID_BASE_MAX) in sctp_cname() 66 return sctp_cid_tbl[cid.chunk]; in sctp_cname() 68 switch (cid.chunk) { in sctp_cname()
|
D | auth.c | 391 struct sctp_chunk *chunk; in sctp_auth_asoc_init_active_key() local 417 list_for_each_entry(chunk, &asoc->outqueue.out_chunk_list, list) { in sctp_auth_asoc_init_active_key() 418 if (sctp_auth_send_cid(chunk->chunk_hdr->type, asoc)) in sctp_auth_asoc_init_active_key() 419 chunk->auth = 1; in sctp_auth_asoc_init_active_key() 638 static int __sctp_auth_cid(sctp_cid_t chunk, struct sctp_chunks_param *param) in __sctp_auth_cid() argument 664 if (param->chunks[i] == chunk) in __sctp_auth_cid() 674 int sctp_auth_send_cid(sctp_cid_t chunk, const struct sctp_association *asoc) in sctp_auth_send_cid() argument 682 return __sctp_auth_cid(chunk, asoc->peer.peer_chunks); in sctp_auth_send_cid() 686 int sctp_auth_recv_cid(sctp_cid_t chunk, const struct sctp_association *asoc) in sctp_auth_recv_cid() argument 694 return __sctp_auth_cid(chunk, in sctp_auth_recv_cid()
|
D | associola.c | 919 struct sctp_chunk *chunk; in sctp_assoc_lookup_tsn() local 941 list_for_each_entry(chunk, &active->transmitted, in sctp_assoc_lookup_tsn() 944 if (key == chunk->subh.data_hdr->tsn) { in sctp_assoc_lookup_tsn() 956 list_for_each_entry(chunk, &transport->transmitted, in sctp_assoc_lookup_tsn() 958 if (key == chunk->subh.data_hdr->tsn) { in sctp_assoc_lookup_tsn() 1001 struct sctp_chunk *chunk; in sctp_assoc_bh_rcv() local 1012 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_assoc_bh_rcv() 1014 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_assoc_bh_rcv() 1023 if (sctp_auth_recv_cid(subtype.chunk, asoc) && !chunk->auth) in sctp_assoc_bh_rcv() 1029 if (sctp_chunk_is_data(chunk)) in sctp_assoc_bh_rcv() [all …]
|
D | objcnt.c | 52 SCTP_DBG_OBJCNT(chunk); 66 SCTP_DBG_OBJCNT_ENTRY(chunk),
|
D | ulpqueue.c | 95 int sctp_ulpq_tail_data(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, in sctp_ulpq_tail_data() argument 103 event = sctp_ulpevent_make_rcvmsg(chunk->asoc, chunk, gfp); in sctp_ulpq_tail_data() 1080 void sctp_ulpq_renege(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, in sctp_ulpq_renege() argument 1088 if (chunk) { in sctp_ulpq_renege() 1089 needed = ntohs(chunk->chunk_hdr->length); in sctp_ulpq_renege() 1103 if (chunk && (freed >= needed)) { in sctp_ulpq_renege() 1105 retval = sctp_ulpq_tail_data(ulpq, chunk, gfp); in sctp_ulpq_renege()
|
D | socket.c | 99 struct sctp_chunk *chunk); 148 static inline void sctp_set_owner_w(struct sctp_chunk *chunk) in sctp_set_owner_w() argument 150 struct sctp_association *asoc = chunk->asoc; in sctp_set_owner_w() 156 skb_set_owner_w(chunk->skb, sk); in sctp_set_owner_w() 158 chunk->skb->destructor = sctp_wfree; in sctp_set_owner_w() 160 skb_shinfo(chunk->skb)->destructor_arg = chunk; in sctp_set_owner_w() 162 asoc->sndbuf_used += SCTP_DATA_SNDSIZE(chunk) + in sctp_set_owner_w() 167 sk->sk_wmem_queued += chunk->skb->truesize; in sctp_set_owner_w() 168 sk_mem_charge(sk, chunk->skb->truesize); in sctp_set_owner_w() 411 struct sctp_chunk *chunk) in sctp_send_asconf() argument [all …]
|
D | probe.c | 137 struct sctp_chunk *chunk = arg; in jsctp_sf_eat_sack() local 138 struct sk_buff *skb = chunk->skb; in jsctp_sf_eat_sack()
|
D | Makefile | 10 transport.o chunk.o sm_make_chunk.o ulpevent.o \
|
D | sm_statetable.c | 86 return sctp_chunk_event_lookup(net, event_subtype.chunk, state); in sctp_sm_lookup_event()
|
/linux-4.4.14/mm/ |
D | percpu-vm.c | 13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk, in pcpu_chunk_page() argument 17 WARN_ON(chunk->immutable); in pcpu_chunk_page() 19 return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx)); in pcpu_chunk_page() 55 static void pcpu_free_pages(struct pcpu_chunk *chunk, in pcpu_free_pages() argument 82 static int pcpu_alloc_pages(struct pcpu_chunk *chunk, in pcpu_alloc_pages() argument 125 static void pcpu_pre_unmap_flush(struct pcpu_chunk *chunk, in pcpu_pre_unmap_flush() argument 129 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start), in pcpu_pre_unmap_flush() 130 pcpu_chunk_addr(chunk, pcpu_high_unit_cpu, page_end)); in pcpu_pre_unmap_flush() 151 static void pcpu_unmap_pages(struct pcpu_chunk *chunk, in pcpu_unmap_pages() argument 161 page = pcpu_chunk_page(chunk, cpu, i); in pcpu_unmap_pages() [all …]
|
D | percpu.c | 218 static int pcpu_chunk_slot(const struct pcpu_chunk *chunk) in pcpu_chunk_slot() argument 220 if (chunk->free_size < sizeof(int) || chunk->contig_hint < sizeof(int)) in pcpu_chunk_slot() 223 return pcpu_size_to_slot(chunk->free_size); in pcpu_chunk_slot() 243 static unsigned long pcpu_chunk_addr(struct pcpu_chunk *chunk, in pcpu_chunk_addr() argument 246 return (unsigned long)chunk->base_addr + pcpu_unit_offsets[cpu] + in pcpu_chunk_addr() 250 static void __maybe_unused pcpu_next_unpop(struct pcpu_chunk *chunk, in pcpu_next_unpop() argument 253 *rs = find_next_zero_bit(chunk->populated, end, *rs); in pcpu_next_unpop() 254 *re = find_next_bit(chunk->populated, end, *rs + 1); in pcpu_next_unpop() 257 static void __maybe_unused pcpu_next_pop(struct pcpu_chunk *chunk, in pcpu_next_pop() argument 260 *rs = find_next_bit(chunk->populated, end, *rs); in pcpu_next_pop() [all …]
|
D | percpu-km.c | 36 static int pcpu_populate_chunk(struct pcpu_chunk *chunk, in pcpu_populate_chunk() argument 42 static void pcpu_depopulate_chunk(struct pcpu_chunk *chunk, in pcpu_depopulate_chunk() argument 51 struct pcpu_chunk *chunk; in pcpu_create_chunk() local 55 chunk = pcpu_alloc_chunk(); in pcpu_create_chunk() 56 if (!chunk) in pcpu_create_chunk() 61 pcpu_free_chunk(chunk); in pcpu_create_chunk() 66 pcpu_set_page_chunk(nth_page(pages, i), chunk); in pcpu_create_chunk() 68 chunk->data = pages; in pcpu_create_chunk() 69 chunk->base_addr = page_address(pages) - pcpu_group_offsets[0]; in pcpu_create_chunk() 72 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk() [all …]
|
/linux-4.4.14/drivers/s390/cio/ |
D | itcw.c | 181 void *chunk; in itcw_init() local 193 chunk = fit_chunk(&start, end, sizeof(struct itcw), 1, 0); in itcw_init() 194 if (IS_ERR(chunk)) in itcw_init() 195 return chunk; in itcw_init() 196 itcw = chunk; in itcw_init() 209 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init() 210 if (IS_ERR(chunk)) in itcw_init() 211 return chunk; in itcw_init() 212 itcw->tcw = chunk; in itcw_init() 217 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init() [all …]
|
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/ |
D | icm.c | 54 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_pages() argument 58 if (chunk->nsg > 0) in mlx4_free_icm_pages() 59 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, in mlx4_free_icm_pages() 62 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages() 63 __free_pages(sg_page(&chunk->mem[i]), in mlx4_free_icm_pages() 64 get_order(chunk->mem[i].length)); in mlx4_free_icm_pages() 67 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_coherent() argument 71 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent() 73 chunk->mem[i].length, in mlx4_free_icm_coherent() 74 lowmem_page_address(sg_page(&chunk->mem[i])), in mlx4_free_icm_coherent() [all …]
|
D | icm.h | 64 struct mlx4_icm_chunk *chunk; member 91 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first() 99 return !iter->chunk; in mlx4_icm_last() 104 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next() 105 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next() 106 iter->chunk = NULL; in mlx4_icm_next() 110 iter->chunk = list_entry(iter->chunk->list.next, in mlx4_icm_next() 118 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_addr() 123 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_size()
|
D | mr.c | 724 int chunk; in __mlx4_write_mtt() local 733 chunk = min_t(int, max_mtts_first_page, npages); in __mlx4_write_mtt() 736 err = mlx4_write_mtt_chunk(dev, mtt, start_index, chunk, page_list); in __mlx4_write_mtt() 739 npages -= chunk; in __mlx4_write_mtt() 740 start_index += chunk; in __mlx4_write_mtt() 741 page_list += chunk; in __mlx4_write_mtt() 743 chunk = min_t(int, mtts_per_page, npages); in __mlx4_write_mtt() 753 int chunk; in mlx4_write_mtt() local 767 chunk = min_t(int, MLX4_MAILBOX_SIZE / sizeof(u64) - 2, in mlx4_write_mtt() 771 for (i = 0; i < chunk; ++i) in mlx4_write_mtt() [all …]
|
/linux-4.4.14/lib/ |
D | genalloc.c | 39 static inline size_t chunk_size(const struct gen_pool_chunk *chunk) in chunk_size() argument 41 return chunk->end_addr - chunk->start_addr + 1; in chunk_size() 185 struct gen_pool_chunk *chunk; in gen_pool_add_virt() local 190 chunk = kzalloc_node(nbytes, GFP_KERNEL, nid); in gen_pool_add_virt() 191 if (unlikely(chunk == NULL)) in gen_pool_add_virt() 194 chunk->phys_addr = phys; in gen_pool_add_virt() 195 chunk->start_addr = virt; in gen_pool_add_virt() 196 chunk->end_addr = virt + size - 1; in gen_pool_add_virt() 197 atomic_set(&chunk->avail, size); in gen_pool_add_virt() 200 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_virt() [all …]
|
D | bitmap.c | 363 u32 chunk; in __bitmap_parse() local 370 chunk = 0; in __bitmap_parse() 406 if (chunk & ~((1UL << (CHUNKSZ - 4)) - 1)) in __bitmap_parse() 409 chunk = (chunk << 4) | hex_to_bin(c); in __bitmap_parse() 414 if (nchunks == 0 && chunk == 0) in __bitmap_parse() 418 *maskp |= chunk; in __bitmap_parse() 420 nbits += (nchunks == 1) ? nbits_to_hold_value(chunk) : CHUNKSZ; in __bitmap_parse()
|
D | Kconfig | 517 Provides a heler to split scatterlists into chunks, each chunk being a
|
/linux-4.4.14/kernel/ |
D | audit_tree.c | 109 static void free_chunk(struct audit_chunk *chunk) in free_chunk() argument 113 for (i = 0; i < chunk->count; i++) { in free_chunk() 114 if (chunk->owners[i].owner) in free_chunk() 115 put_tree(chunk->owners[i].owner); in free_chunk() 117 kfree(chunk); in free_chunk() 120 void audit_put_chunk(struct audit_chunk *chunk) in audit_put_chunk() argument 122 if (atomic_long_dec_and_test(&chunk->refs)) in audit_put_chunk() 123 free_chunk(chunk); in audit_put_chunk() 128 struct audit_chunk *chunk = container_of(rcu, struct audit_chunk, head); in __put_chunk() local 129 audit_put_chunk(chunk); in __put_chunk() [all …]
|
D | auditsc.c | 210 static int put_tree_ref(struct audit_context *ctx, struct audit_chunk *chunk) in put_tree_ref() argument 215 p->c[--left] = chunk; in put_tree_ref() 223 p->c[30] = chunk; in put_tree_ref() 1588 struct audit_chunk *chunk; in handle_one() local 1596 chunk = audit_tree_lookup(inode); in handle_one() 1598 if (!chunk) in handle_one() 1600 if (likely(put_tree_ref(context, chunk))) in handle_one() 1605 audit_put_chunk(chunk); in handle_one() 1609 put_tree_ref(context, chunk); in handle_one() 1634 struct audit_chunk *chunk; in handle_path() local [all …]
|
/linux-4.4.14/drivers/infiniband/hw/mthca/ |
D | mthca_memfree.c | 64 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_pages() argument 68 if (chunk->nsg > 0) in mthca_free_icm_pages() 69 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, in mthca_free_icm_pages() 72 for (i = 0; i < chunk->npages; ++i) in mthca_free_icm_pages() 73 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages() 74 get_order(chunk->mem[i].length)); in mthca_free_icm_pages() 77 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_coherent() argument 81 for (i = 0; i < chunk->npages; ++i) { in mthca_free_icm_coherent() 82 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, in mthca_free_icm_coherent() 83 lowmem_page_address(sg_page(&chunk->mem[i])), in mthca_free_icm_coherent() [all …]
|
D | mthca_memfree.h | 76 struct mthca_icm_chunk *chunk; member 103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first() 111 return !iter->chunk; in mthca_icm_last() 116 if (++iter->page_idx >= iter->chunk->nsg) { in mthca_icm_next() 117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next() 118 iter->chunk = NULL; in mthca_icm_next() 122 iter->chunk = list_entry(iter->chunk->list.next, in mthca_icm_next() 130 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mthca_icm_addr() 135 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mthca_icm_size()
|
D | mthca_mr.c | 362 int chunk; in mthca_write_mtt() local 369 chunk = min(size, list_len); in mthca_write_mtt() 372 buffer_list, chunk); in mthca_write_mtt() 375 buffer_list, chunk); in mthca_write_mtt() 377 list_len -= chunk; in mthca_write_mtt() 378 start_index += chunk; in mthca_write_mtt() 379 buffer_list += chunk; in mthca_write_mtt()
|
/linux-4.4.14/drivers/gpu/drm/qxl/ |
D | qxl_image.c | 38 struct qxl_drm_chunk *chunk; in qxl_allocate_chunk() local 41 chunk = kmalloc(sizeof(struct qxl_drm_chunk), GFP_KERNEL); in qxl_allocate_chunk() 42 if (!chunk) in qxl_allocate_chunk() 45 ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo); in qxl_allocate_chunk() 47 kfree(chunk); in qxl_allocate_chunk() 51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk() 88 struct qxl_drm_chunk *chunk, *tmp; in qxl_image_free_objects() local 90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects() 91 qxl_bo_unref(&chunk->bo); in qxl_image_free_objects() 92 kfree(chunk); in qxl_image_free_objects() [all …]
|
D | qxl_dev.h | 421 struct qxl_data_chunk chunk; member 482 struct qxl_data_chunk chunk; member 761 struct qxl_data_chunk chunk; member 779 struct qxl_data_chunk chunk; member
|
D | qxl_draw.c | 52 dev_clips->chunk.next_chunk = 0; in drawable_set_clipping() 53 dev_clips->chunk.prev_chunk = 0; in drawable_set_clipping() 54 dev_clips->chunk.data_size = sizeof(struct qxl_rect) * num_clips; in drawable_set_clipping() 55 return (struct qxl_rect *)dev_clips->chunk.data; in drawable_set_clipping()
|
D | qxl_display.c | 368 cursor->chunk.next_chunk = 0; in qxl_crtc_cursor_set2() 369 cursor->chunk.prev_chunk = 0; in qxl_crtc_cursor_set2() 370 cursor->chunk.data_size = size; in qxl_crtc_cursor_set2() 372 memcpy(cursor->chunk.data, user_ptr, size); in qxl_crtc_cursor_set2()
|
/linux-4.4.14/arch/mips/dec/prom/ |
D | console.c | 22 unsigned int chunk = sizeof(buf) - 1; in prom_console_write() local 25 if (chunk > c) in prom_console_write() 26 chunk = c; in prom_console_write() 27 memcpy(buf, s, chunk); in prom_console_write() 28 buf[chunk] = '\0'; in prom_console_write() 30 s += chunk; in prom_console_write() 31 c -= chunk; in prom_console_write()
|
/linux-4.4.14/sound/core/ |
D | sgbuf.c | 71 unsigned int i, pages, chunk, maxpages; in snd_malloc_sgbuf_pages() local 96 chunk = pages; in snd_malloc_sgbuf_pages() 98 if (chunk > maxpages) in snd_malloc_sgbuf_pages() 99 chunk = maxpages; in snd_malloc_sgbuf_pages() 100 chunk <<= PAGE_SHIFT; in snd_malloc_sgbuf_pages() 102 chunk, &tmpb) < 0) { in snd_malloc_sgbuf_pages() 110 chunk = tmpb.bytes >> PAGE_SHIFT; in snd_malloc_sgbuf_pages() 111 for (i = 0; i < chunk; i++) { in snd_malloc_sgbuf_pages() 115 table->addr |= chunk; /* mark head */ in snd_malloc_sgbuf_pages() 121 sgbuf->pages += chunk; in snd_malloc_sgbuf_pages() [all …]
|
/linux-4.4.14/drivers/media/usb/usbtv/ |
D | usbtv.h | 49 #define USBTV_MAGIC_OK(chunk) ((be32_to_cpu(chunk[0]) & 0xff000000) \ argument 51 #define USBTV_FRAME_ID(chunk) ((be32_to_cpu(chunk[0]) & 0x00ff0000) >> 16) argument 52 #define USBTV_ODD(chunk) ((be32_to_cpu(chunk[0]) & 0x0000f000) >> 15) argument 53 #define USBTV_CHUNK_NO(chunk) (be32_to_cpu(chunk[0]) & 0x00000fff) argument
|
D | usbtv-video.c | 275 static void usbtv_image_chunk(struct usbtv *usbtv, __be32 *chunk) in usbtv_image_chunk() argument 283 if (!USBTV_MAGIC_OK(chunk)) in usbtv_image_chunk() 285 frame_id = USBTV_FRAME_ID(chunk); in usbtv_image_chunk() 286 odd = USBTV_ODD(chunk); in usbtv_image_chunk() 287 chunk_no = USBTV_CHUNK_NO(chunk); in usbtv_image_chunk() 312 usbtv_chunk_to_vbuf(frame, &chunk[1], chunk_no, odd); in usbtv_image_chunk()
|
/linux-4.4.14/drivers/infiniband/hw/usnic/ |
D | usnic_vnic.c | 96 struct usnic_vnic_res_chunk *chunk; in usnic_vnic_dump() local 119 chunk = &vnic->chunks[i]; in usnic_vnic_dump() 120 for (j = 0; j < chunk->cnt; j++) { in usnic_vnic_dump() 121 res = chunk->res[j]; in usnic_vnic_dump() 277 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk) in usnic_vnic_put_resources() argument 282 struct usnic_vnic *vnic = chunk->vnic; in usnic_vnic_put_resources() 285 while ((i = --chunk->cnt) >= 0) { in usnic_vnic_put_resources() 286 res = chunk->res[i]; in usnic_vnic_put_resources() 287 chunk->res[i] = NULL; in usnic_vnic_put_resources() 293 kfree(chunk->res); in usnic_vnic_put_resources() [all …]
|
D | usnic_uiom.c | 81 struct usnic_uiom_chunk *chunk, *tmp; in usnic_uiom_put_pages() local 87 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages() 88 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_put_pages() 96 kfree(chunk); in usnic_uiom_put_pages() 105 struct usnic_uiom_chunk *chunk; in usnic_uiom_get_pages() local 159 chunk = kmalloc(sizeof(*chunk) + in usnic_uiom_get_pages() 163 if (!chunk) { in usnic_uiom_get_pages() 168 chunk->nents = min_t(int, ret, USNIC_UIOM_PAGE_CHUNK); in usnic_uiom_get_pages() 169 sg_init_table(chunk->page_list, chunk->nents); in usnic_uiom_get_pages() 170 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_get_pages() [all …]
|
D | usnic_ib_verbs.c | 64 struct usnic_vnic_res_chunk *chunk; in usnic_ib_fill_create_qp_resp() local 89 chunk = usnic_ib_qp_grp_get_chunk(qp_grp, USNIC_VNIC_RES_TYPE_RQ); in usnic_ib_fill_create_qp_resp() 90 if (IS_ERR_OR_NULL(chunk)) { in usnic_ib_fill_create_qp_resp() 94 PTR_ERR(chunk)); in usnic_ib_fill_create_qp_resp() 95 return chunk ? PTR_ERR(chunk) : -ENOMEM; in usnic_ib_fill_create_qp_resp() 98 WARN_ON(chunk->type != USNIC_VNIC_RES_TYPE_RQ); in usnic_ib_fill_create_qp_resp() 99 resp.rq_cnt = chunk->cnt; in usnic_ib_fill_create_qp_resp() 100 for (i = 0; i < chunk->cnt; i++) in usnic_ib_fill_create_qp_resp() 101 resp.rq_idx[i] = chunk->res[i]->vnic_idx; in usnic_ib_fill_create_qp_resp() 103 chunk = usnic_ib_qp_grp_get_chunk(qp_grp, USNIC_VNIC_RES_TYPE_WQ); in usnic_ib_fill_create_qp_resp() [all …]
|
D | usnic_vnic.h | 110 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk);
|
/linux-4.4.14/drivers/s390/block/ |
D | dasd_int.h | 574 struct dasd_mchunk *chunk; in dasd_init_chunklist() local 577 chunk = (struct dasd_mchunk *) mem; in dasd_init_chunklist() 578 chunk->size = size - sizeof(struct dasd_mchunk); in dasd_init_chunklist() 579 list_add(&chunk->list, chunk_list); in dasd_init_chunklist() 585 struct dasd_mchunk *chunk, *tmp; in dasd_alloc_chunk() local 588 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk() 589 if (chunk->size < size) in dasd_alloc_chunk() 591 if (chunk->size > size + sizeof(struct dasd_mchunk)) { in dasd_alloc_chunk() 592 char *endaddr = (char *) (chunk + 1) + chunk->size; in dasd_alloc_chunk() 595 chunk->size -= size + sizeof(struct dasd_mchunk); in dasd_alloc_chunk() [all …]
|
/linux-4.4.14/fs/ocfs2/ |
D | quota_local.c | 296 int chunk, in ocfs2_add_recovery_chunk() argument 304 rc->rc_chunk = chunk; in ocfs2_add_recovery_chunk() 470 int bit, chunk; in ocfs2_recover_local_quota_file() local 477 chunk = rchunk->rc_chunk; in ocfs2_recover_local_quota_file() 480 ol_quota_chunk_block(sb, chunk), in ocfs2_recover_local_quota_file() 490 ol_dqblk_block(sb, chunk, bit), in ocfs2_recover_local_quota_file() 497 ol_dqblk_block_off(sb, chunk, bit)); in ocfs2_recover_local_quota_file() 816 struct ocfs2_quota_chunk *chunk; in ocfs2_local_free_info() local 824 list_for_each_entry(chunk, &oinfo->dqi_chunk, qc_chunk) { in ocfs2_local_free_info() 826 (chunk->qc_headerbh->b_data); in ocfs2_local_free_info() [all …]
|
D | ioctl.c | 448 unsigned int offset = 0, cluster, chunk; in ocfs2_info_freefrag_scan_chain() local 487 for (chunk = 0; chunk < chunks_in_group; chunk++) { in ocfs2_info_freefrag_scan_chain()
|
/linux-4.4.14/drivers/media/usb/go7007/ |
D | go7007-fw.c | 378 int size = 0, i, off = 0, chunk; in gen_mjpeghdr_to_package() local 388 chunk = mjpeg_frame_header(go, buf + size, 1); in gen_mjpeghdr_to_package() 389 memmove(buf + size, buf + size + 80, chunk - 80); in gen_mjpeghdr_to_package() 390 size += chunk - 80; in gen_mjpeghdr_to_package() 392 for (i = 0; i < size; i += chunk * 2) { in gen_mjpeghdr_to_package() 400 chunk = 28; in gen_mjpeghdr_to_package() 401 if (mem + chunk > 0x4000) in gen_mjpeghdr_to_package() 402 chunk = 0x4000 - mem; in gen_mjpeghdr_to_package() 403 if (i + 2 * chunk > size) in gen_mjpeghdr_to_package() 404 chunk = (size - i) / 2; in gen_mjpeghdr_to_package() [all …]
|
/linux-4.4.14/include/net/sctp/ |
D | sm.h | 174 __u8 sctp_get_chunk_type(struct sctp_chunk *chunk); 209 const struct sctp_chunk *chunk); 239 const struct sctp_chunk *chunk, 252 struct sctp_chunk *chunk, bool addr_param_needed, 299 static inline __u16 sctp_data_size(struct sctp_chunk *chunk) in sctp_data_size() argument 303 size = ntohs(chunk->chunk_hdr->length); in sctp_data_size() 392 sctp_vtag_verify(const struct sctp_chunk *chunk, in sctp_vtag_verify() argument 401 if (ntohl(chunk->sctp_hdr->vtag) == asoc->c.my_vtag) in sctp_vtag_verify() 411 sctp_vtag_verify_either(const struct sctp_chunk *chunk, in sctp_vtag_verify_either() argument 435 if ((!sctp_test_T_bit(chunk) && in sctp_vtag_verify_either() [all …]
|
D | sctp.h | 432 static inline bool sctp_chunk_pending(const struct sctp_chunk *chunk) in sctp_chunk_pending() argument 434 return !list_empty(&chunk->list); in sctp_chunk_pending() 442 #define sctp_walk_params(pos, chunk, member)\ argument 443 _sctp_walk_params((pos), (chunk), ntohs((chunk)->chunk_hdr.length), member) 445 #define _sctp_walk_params(pos, chunk, end, member)\ argument 446 for (pos.v = chunk->member;\ 447 pos.v <= (void *)chunk + end - ntohs(pos.p->length) &&\ 461 #define sctp_walk_fwdtsn(pos, chunk)\ argument 462 _sctp_walk_fwdtsn((pos), (chunk), ntohs((chunk)->chunk_hdr->length) - sizeof(struct sctp_fwdtsn_chu… 464 #define _sctp_walk_fwdtsn(pos, chunk, end)\ argument [all …]
|
D | ulpevent.h | 85 struct sctp_chunk *chunk, 98 struct sctp_chunk *chunk, 103 struct sctp_chunk *chunk, 121 struct sctp_chunk *chunk,
|
D | auth.h | 100 int sctp_auth_send_cid(sctp_cid_t chunk, const struct sctp_association *asoc); 101 int sctp_auth_recv_cid(sctp_cid_t chunk, const struct sctp_association *asoc);
|
D | constants.h | 128 sctp_cid_t chunk; member 139 SCTP_SUBTYPE_CONSTRUCTOR(CHUNK, sctp_cid_t, chunk)
|
D | command.h | 131 struct sctp_chunk *chunk; member 172 SCTP_ARG_CONSTRUCTOR(CHUNK, struct sctp_chunk *, chunk) in SCTP_ARG_CONSTRUCTOR()
|
D | structs.h | 654 int sctp_user_addto_chunk(struct sctp_chunk *chunk, int len, 663 const union sctp_addr *sctp_source(const struct sctp_chunk *chunk); 1058 int sctp_outq_tail(struct sctp_outq *, struct sctp_chunk *chunk); 1281 struct sctp_chunk *chunk, struct sctp_chunk **err_chunk); 1282 int sctp_process_init(struct sctp_association *, struct sctp_chunk *chunk,
|
/linux-4.4.14/sound/soc/codecs/ |
D | sigmadsp.c | 54 struct sigma_fw_chunk chunk; member 60 struct sigma_fw_chunk chunk; member 68 struct sigma_fw_chunk chunk; member 200 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_control() argument 212 ctrl_chunk = (const struct sigma_fw_chunk_control *)chunk; in sigma_fw_load_control() 238 ctrl->samplerates = le32_to_cpu(chunk->samplerates); in sigma_fw_load_control() 251 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_data() argument 259 data_chunk = (struct sigma_fw_chunk_data *)chunk; in sigma_fw_load_data() 269 data->samplerates = le32_to_cpu(chunk->samplerates); in sigma_fw_load_data() 277 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_samplerates() argument [all …]
|
/linux-4.4.14/arch/um/drivers/ |
D | mconsole_kern.h | 27 #define CONFIG_CHUNK(str, size, current, chunk, end) \ argument 29 current += strlen(chunk); \ 33 strcpy(str, chunk); \ 34 str += strlen(chunk); \
|
/linux-4.4.14/drivers/dma/sh/ |
D | rcar-dmac.c | 328 struct rcar_dmac_xfer_chunk *chunk; in rcar_dmac_chan_start_xfer() local 352 chunk = list_first_entry(&desc->chunks, in rcar_dmac_chan_start_xfer() 355 chunk->dst_addr & 0xffffffff); in rcar_dmac_chan_start_xfer() 385 struct rcar_dmac_xfer_chunk *chunk = desc->running; in rcar_dmac_chan_start_xfer() local 389 chan->index, chunk, chunk->size, &chunk->src_addr, in rcar_dmac_chan_start_xfer() 390 &chunk->dst_addr); in rcar_dmac_chan_start_xfer() 394 chunk->src_addr >> 32); in rcar_dmac_chan_start_xfer() 396 chunk->dst_addr >> 32); in rcar_dmac_chan_start_xfer() 399 chunk->src_addr & 0xffffffff); in rcar_dmac_chan_start_xfer() 401 chunk->dst_addr & 0xffffffff); in rcar_dmac_chan_start_xfer() [all …]
|
D | shdma-base.c | 75 struct shdma_desc *chunk, *c, *desc = in shdma_tx_submit() local 89 list_for_each_entry_safe(chunk, c, desc->node.prev, node) { in shdma_tx_submit() 94 if (chunk != desc && (chunk->mark == DESC_IDLE || in shdma_tx_submit() 95 chunk->async_tx.cookie > 0 || in shdma_tx_submit() 96 chunk->async_tx.cookie == -EBUSY || in shdma_tx_submit() 97 &chunk->node == &schan->ld_free)) in shdma_tx_submit() 99 chunk->mark = DESC_SUBMITTED; in shdma_tx_submit() 100 if (chunk->chunks == 1) { in shdma_tx_submit() 101 chunk->async_tx.callback = callback; in shdma_tx_submit() 102 chunk->async_tx.callback_param = tx->callback_param; in shdma_tx_submit() [all …]
|
/linux-4.4.14/drivers/md/ |
D | dm-snap.c | 164 chunk_t chunk) in chunk_to_sector() argument 166 return chunk << store->chunk_shift; in chunk_to_sector() 222 chunk_t chunk; member 237 static void track_chunk(struct dm_snapshot *s, struct bio *bio, chunk_t chunk) in track_chunk() argument 241 c->chunk = chunk; in track_chunk() 245 &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)]); in track_chunk() 259 static int __chunk_is_tracked(struct dm_snapshot *s, chunk_t chunk) in __chunk_is_tracked() argument 267 &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)], node) { in __chunk_is_tracked() 268 if (c->chunk == chunk) { in __chunk_is_tracked() 283 static void __check_for_conflicting_io(struct dm_snapshot *s, chunk_t chunk) in __check_for_conflicting_io() argument [all …]
|
D | dm-stripe.c | 216 sector_t chunk = dm_target_offset(sc->ti, sector); in stripe_map_sector() local 220 chunk_offset = sector_div(chunk, sc->chunk_size); in stripe_map_sector() 222 chunk_offset = chunk & (sc->chunk_size - 1); in stripe_map_sector() 223 chunk >>= sc->chunk_size_shift; in stripe_map_sector() 227 *stripe = sector_div(chunk, sc->stripes); in stripe_map_sector() 229 *stripe = chunk & (sc->stripes - 1); in stripe_map_sector() 230 chunk >>= sc->stripes_shift; in stripe_map_sector() 234 chunk *= sc->chunk_size; in stripe_map_sector() 236 chunk <<= sc->chunk_size_shift; in stripe_map_sector() 238 *result = chunk + chunk_offset; in stripe_map_sector()
|
D | dm-exception-store.h | 142 static inline chunk_t dm_chunk_number(chunk_t chunk) in dm_chunk_number() argument 144 return chunk & (chunk_t)((1ULL << DM_CHUNK_NUMBER_BITS) - 1ULL); in dm_chunk_number() 169 static inline chunk_t dm_chunk_number(chunk_t chunk) in dm_chunk_number() argument 171 return chunk; in dm_chunk_number()
|
D | bitmap.c | 724 unsigned long chunk) in file_page_index() argument 727 chunk += sizeof(bitmap_super_t) << 3; in file_page_index() 728 return chunk >> PAGE_BIT_SHIFT; in file_page_index() 733 unsigned long chunk) in file_page_offset() argument 736 chunk += sizeof(bitmap_super_t) << 3; in file_page_offset() 737 return chunk & (PAGE_BITS - 1); in file_page_offset() 745 unsigned long chunk) in filemap_get_page() argument 747 if (file_page_index(store, chunk) >= store->file_pages) in filemap_get_page() 749 return store->filemap[file_page_index(store, chunk)]; in filemap_get_page() 908 unsigned long chunk = block >> bitmap->counts.chunkshift; in bitmap_file_set_bit() local [all …]
|
D | dm-snap-persistent.c | 229 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, int rw, in chunk_io() argument 234 .sector = ps->store->chunk_size * chunk, in chunk_io() 287 chunk_t chunk; in area_io() local 289 chunk = area_location(ps, ps->current_area); in area_io() 291 r = chunk_io(ps, ps->area, chunk, rw, 0); in area_io() 519 chunk_t chunk; in read_exceptions() local 534 chunk = area_location(ps, ps->current_area); in read_exceptions() 536 area = dm_bufio_read(client, chunk, &bp); in read_exceptions() 550 dm_bufio_forget(client, chunk); in read_exceptions()
|
D | raid0.c | 308 sector_t chunk; in map_sector() local 319 chunk = *sector_offset; in map_sector() 321 sector_div(chunk, zone->nb_dev << chunksect_bits); in map_sector() 324 chunk = *sector_offset; in map_sector() 325 sector_div(chunk, chunk_sects * zone->nb_dev); in map_sector() 332 *sector_offset = (chunk * chunk_sects) + sect_in_chunk; in map_sector()
|
D | raid10.c | 556 sector_t chunk; in __raid10_find_phys() local 569 chunk = r10bio->sector >> geo->chunk_shift; in __raid10_find_phys() 572 chunk *= geo->near_copies; in __raid10_find_phys() 573 stripe = chunk; in __raid10_find_phys() 632 sector_t offset, chunk, vchunk; in raid10_find_virt() local 655 chunk = sector >> geo->chunk_shift; in raid10_find_virt() 656 fc = sector_div(chunk, geo->far_copies); in raid10_find_virt() 668 chunk = sector >> geo->chunk_shift; in raid10_find_virt() 670 vchunk = chunk * geo->raid_disks + dev; in raid10_find_virt() 3387 int layout, chunk, disks; in setup_geo() local [all …]
|
D | Kconfig | 68 up evenly, one chunk here and one chunk there. This will increase
|
D | md.c | 4133 unsigned long chunk, end_chunk; in bitmap_store() local 4143 chunk = end_chunk = simple_strtoul(buf, &end, 0); in bitmap_store() 4151 bitmap_dirty_bits(mddev->bitmap, chunk, end_chunk); in bitmap_store() 4591 int chunk; in max_sync_store() local 4606 chunk = mddev->chunk_sectors; in max_sync_store() 4607 if (chunk) { in max_sync_store() 4611 if (sector_div(temp, chunk)) in max_sync_store()
|
/linux-4.4.14/Documentation/device-mapper/ |
D | striped.txt | 10 Parameters: <num devs> <chunk size> [<dev path> <offset>]+ 12 <chunk size>: Size of each chunk of data. Must be at least as 19 be a multiple of the chunk size multiplied by the number of underlying devices. 28 # will be called "stripe_dev" and have a chunk-size of 128k.
|
D | cache.txt | 141 Policy plug-ins can store a chunk of data per cache block. It's up to 142 the policy how big this chunk is, but it should be kept small. Like the
|
D | snapshot.txt | 65 deferred until the merging snapshot's corresponding chunk(s) have been
|
/linux-4.4.14/arch/arm64/kernel/ |
D | sys_compat.c | 37 unsigned long chunk = min(PAGE_SIZE, end - start); in __do_compat_cache_op() local 42 ret = __flush_cache_user_range(start, start + chunk); in __do_compat_cache_op() 47 start += chunk; in __do_compat_cache_op()
|
/linux-4.4.14/net/bluetooth/ |
D | af_bluetooth.c | 306 int chunk; in bt_sock_stream_recvmsg() local 332 chunk = min_t(unsigned int, skb->len, size); in bt_sock_stream_recvmsg() 333 if (skb_copy_datagram_msg(skb, 0, msg, chunk)) { in bt_sock_stream_recvmsg() 339 copied += chunk; in bt_sock_stream_recvmsg() 340 size -= chunk; in bt_sock_stream_recvmsg() 347 if (chunk <= skb_len) { in bt_sock_stream_recvmsg() 348 __skb_pull(skb, chunk); in bt_sock_stream_recvmsg() 353 chunk -= skb_len; in bt_sock_stream_recvmsg() 356 if (chunk <= frag->len) { in bt_sock_stream_recvmsg() 358 skb->len -= chunk; in bt_sock_stream_recvmsg() [all …]
|
/linux-4.4.14/drivers/atm/ |
D | fore200e.c | 178 fore200e_chunk_alloc(struct fore200e* fore200e, struct chunk* chunk, int size, int alignment, int d… in fore200e_chunk_alloc() argument 185 chunk->alloc_size = size + alignment; in fore200e_chunk_alloc() 186 chunk->align_size = size; in fore200e_chunk_alloc() 187 chunk->direction = direction; in fore200e_chunk_alloc() 189 chunk->alloc_addr = kzalloc(chunk->alloc_size, GFP_KERNEL | GFP_DMA); in fore200e_chunk_alloc() 190 if (chunk->alloc_addr == NULL) in fore200e_chunk_alloc() 194 offset = FORE200E_ALIGN(chunk->alloc_addr, alignment); in fore200e_chunk_alloc() 196 chunk->align_addr = chunk->alloc_addr + offset; in fore200e_chunk_alloc() 198 …chunk->dma_addr = fore200e->bus->dma_map(fore200e, chunk->align_addr, chunk->align_size, direction… in fore200e_chunk_alloc() 207 fore200e_chunk_free(struct fore200e* fore200e, struct chunk* chunk) in fore200e_chunk_free() argument [all …]
|
D | fore200e.h | 563 typedef struct chunk { struct 581 struct chunk data; /* data buffer */ 603 struct chunk status; /* array of completion status */ 613 … struct chunk tpd; /* array of tpds */ 614 … struct chunk status; /* arry of completion status */ 624 …struct chunk rpd; /* array of rpds */ 625 …struct chunk status; /* array of completion status */ 634 …struct chunk rbd_block; /* array of rbds … 635 …struct chunk status; /* array of completion status … 811 int (*dma_chunk_alloc)(struct fore200e*, struct chunk*, int, int, int); [all …]
|
/linux-4.4.14/drivers/infiniband/hw/cxgb4/ |
D | resource.c | 419 unsigned start, chunk, top; in c4iw_ocqp_pool_create() local 426 chunk = rdev->lldi.vr->ocq.size; in c4iw_ocqp_pool_create() 427 top = start + chunk; in c4iw_ocqp_pool_create() 430 chunk = min(top - start + 1, chunk); in c4iw_ocqp_pool_create() 431 if (gen_pool_add(rdev->ocqp_pool, start, chunk, -1)) { in c4iw_ocqp_pool_create() 433 __func__, start, chunk); in c4iw_ocqp_pool_create() 434 if (chunk <= 1024 << MIN_OCQP_SHIFT) { in c4iw_ocqp_pool_create() 440 chunk >>= 1; in c4iw_ocqp_pool_create() 443 __func__, start, chunk); in c4iw_ocqp_pool_create() 444 start += chunk; in c4iw_ocqp_pool_create()
|
/linux-4.4.14/fs/reiserfs/ |
D | xattr.c | 527 size_t chunk; in reiserfs_xattr_set_handle() local 532 chunk = PAGE_CACHE_SIZE; in reiserfs_xattr_set_handle() 534 chunk = buffer_size - buffer_pos; in reiserfs_xattr_set_handle() 549 if (chunk + skip > PAGE_CACHE_SIZE) in reiserfs_xattr_set_handle() 550 chunk = PAGE_CACHE_SIZE - skip; in reiserfs_xattr_set_handle() 557 err = __reiserfs_write_begin(page, page_offset, chunk + skip); in reiserfs_xattr_set_handle() 560 memcpy(data + skip, buffer + buffer_pos, chunk); in reiserfs_xattr_set_handle() 562 page_offset + chunk + in reiserfs_xattr_set_handle() 568 buffer_pos += chunk; in reiserfs_xattr_set_handle() 569 file_pos += chunk; in reiserfs_xattr_set_handle() [all …]
|
D | journal.c | 676 static void write_chunk(struct buffer_chunk *chunk) in write_chunk() argument 679 for (i = 0; i < chunk->nr; i++) { in write_chunk() 680 submit_logged_buffer(chunk->bh[i]); in write_chunk() 682 chunk->nr = 0; in write_chunk() 685 static void write_ordered_chunk(struct buffer_chunk *chunk) in write_ordered_chunk() argument 688 for (i = 0; i < chunk->nr; i++) { in write_ordered_chunk() 689 submit_ordered_buffer(chunk->bh[i]); in write_ordered_chunk() 691 chunk->nr = 0; in write_ordered_chunk() 694 static int add_to_chunk(struct buffer_chunk *chunk, struct buffer_head *bh, in add_to_chunk() argument 698 BUG_ON(chunk->nr >= CHUNK_SIZE); in add_to_chunk() [all …]
|
/linux-4.4.14/fs/btrfs/ |
D | print-tree.c | 23 static void print_chunk(struct extent_buffer *eb, struct btrfs_chunk *chunk) in print_chunk() argument 25 int num_stripes = btrfs_chunk_num_stripes(eb, chunk); in print_chunk() 29 btrfs_chunk_length(eb, chunk), btrfs_chunk_owner(eb, chunk), in print_chunk() 30 btrfs_chunk_type(eb, chunk), num_stripes); in print_chunk() 33 btrfs_stripe_devid_nr(eb, chunk, i), in print_chunk() 34 btrfs_stripe_offset_nr(eb, chunk, i)); in print_chunk()
|
D | volumes.c | 2685 struct btrfs_chunk *chunk; in btrfs_del_sys_chunk() local 2707 chunk = (struct btrfs_chunk *)(ptr + len); in btrfs_del_sys_chunk() 2708 num_stripes = btrfs_stack_chunk_num_stripes(chunk); in btrfs_del_sys_chunk() 2877 struct btrfs_chunk *chunk; in btrfs_relocate_sys_chunks() local 2915 chunk = btrfs_item_ptr(leaf, path->slots[0], in btrfs_relocate_sys_chunks() 2917 chunk_type = btrfs_chunk_type(leaf, chunk); in btrfs_relocate_sys_chunks() 3185 struct btrfs_chunk *chunk, in chunk_devid_filter() argument 3189 int num_stripes = btrfs_chunk_num_stripes(leaf, chunk); in chunk_devid_filter() 3193 stripe = btrfs_stripe_nr(chunk, i); in chunk_devid_filter() 3203 struct btrfs_chunk *chunk, in chunk_drange_filter() argument [all …]
|
/linux-4.4.14/sound/usb/line6/ |
D | midi.c | 47 unsigned char chunk[LINE6_FALLBACK_MAXPACKETSIZE]; in line6_midi_transmit() local 52 done = snd_rawmidi_transmit_peek(substream, chunk, req); in line6_midi_transmit() 57 line6_midibuf_write(mb, chunk, done); in line6_midi_transmit() 62 done = line6_midibuf_read(mb, chunk, in line6_midi_transmit() 68 send_midi_async(line6, chunk, done); in line6_midi_transmit()
|
/linux-4.4.14/drivers/net/wireless/ti/wlcore/ |
D | boot.c | 179 u8 *p, *chunk; in wl1271_boot_upload_firmware_chunk() local 194 chunk = kmalloc(CHUNK_SIZE, GFP_KERNEL); in wl1271_boot_upload_firmware_chunk() 195 if (!chunk) { in wl1271_boot_upload_firmware_chunk() 226 memcpy(chunk, p, CHUNK_SIZE); in wl1271_boot_upload_firmware_chunk() 229 ret = wlcore_write(wl, addr, chunk, CHUNK_SIZE, false); in wl1271_boot_upload_firmware_chunk() 239 memcpy(chunk, p, fw_data_len % CHUNK_SIZE); in wl1271_boot_upload_firmware_chunk() 242 ret = wlcore_write(wl, addr, chunk, fw_data_len % CHUNK_SIZE, false); in wl1271_boot_upload_firmware_chunk() 245 kfree(chunk); in wl1271_boot_upload_firmware_chunk()
|
/linux-4.4.14/Documentation/mmc/ |
D | mmc-async-req.txt | 61 request in two chunks, prepare the first chunk and start the request, 62 and finally prepare the second chunk and start the transfer. 72 * The first chunk of the request should take the same time 75 * the transfer is delayed, guesstimate max 4k as first chunk size. 84 * of the first chunk. If the MMC runs out of the first data chunk
|
D | mmc-dev-attrs.txt | 59 hence "preferred_erase_size" provides a good chunk
|
/linux-4.4.14/drivers/crypto/ |
D | n2_core.c | 674 struct n2_crypto_chunk chunk; member 877 struct n2_crypto_chunk *chunk; in n2_compute_chunks() local 890 chunk = &rctx->chunk; in n2_compute_chunks() 891 INIT_LIST_HEAD(&chunk->entry); in n2_compute_chunks() 893 chunk->iv_paddr = 0UL; in n2_compute_chunks() 894 chunk->arr_len = 0; in n2_compute_chunks() 895 chunk->dest_paddr = 0UL; in n2_compute_chunks() 913 if (chunk->arr_len != 0) { in n2_compute_chunks() 917 chunk->arr_len == N2_CHUNK_ARR_LEN || in n2_compute_chunks() 919 chunk->dest_final = dest_prev; in n2_compute_chunks() [all …]
|
/linux-4.4.14/sound/pci/ |
D | es1968.c | 1374 struct esm_memory *chunk = kmalloc(sizeof(*chunk), GFP_KERNEL); in snd_es1968_new_memory() local 1375 if (chunk == NULL) { in snd_es1968_new_memory() 1379 chunk->buf = buf->buf; in snd_es1968_new_memory() 1380 chunk->buf.bytes -= size; in snd_es1968_new_memory() 1381 chunk->buf.area += size; in snd_es1968_new_memory() 1382 chunk->buf.addr += size; in snd_es1968_new_memory() 1383 chunk->empty = 1; in snd_es1968_new_memory() 1385 list_add(&chunk->list, &buf->list); in snd_es1968_new_memory() 1395 struct esm_memory *chunk; in snd_es1968_free_memory() local 1400 chunk = list_entry(buf->list.prev, struct esm_memory, list); in snd_es1968_free_memory() [all …]
|
/linux-4.4.14/arch/x86/xen/ |
D | p2m.c | 331 unsigned int i, chunk; in xen_rebuild_p2m_list() local 354 for (pfn = 0; pfn < xen_max_p2m_pfn; pfn += chunk) { in xen_rebuild_p2m_list() 365 chunk = (pfn & (P2M_PER_PAGE * P2M_MID_PER_PAGE - 1)) ? in xen_rebuild_p2m_list() 371 for (i = 1; i < chunk; i++) in xen_rebuild_p2m_list() 374 if (i < chunk) in xen_rebuild_p2m_list() 376 chunk = P2M_PER_PAGE; in xen_rebuild_p2m_list() 378 if (type == P2M_TYPE_PFN || i < chunk) { in xen_rebuild_p2m_list() 392 if (chunk == P2M_PER_PAGE) { in xen_rebuild_p2m_list()
|
D | setup.c | 340 unsigned int i, chunk; in xen_do_set_identity_and_remap_chunk() local 351 chunk = (left < REMAP_SIZE) ? left : REMAP_SIZE; in xen_do_set_identity_and_remap_chunk() 360 xen_remap_buf.size = chunk; in xen_do_set_identity_and_remap_chunk() 361 for (i = 0; i < chunk; i++) in xen_do_set_identity_and_remap_chunk() 368 set_phys_range_identity(ident_pfn_iter, ident_pfn_iter + chunk); in xen_do_set_identity_and_remap_chunk() 370 left -= chunk; in xen_do_set_identity_and_remap_chunk()
|
/linux-4.4.14/drivers/net/wireless/libertas/ |
D | if_sdio.c | 338 u16 size, type, chunk; in if_sdio_card_to_host() local 362 chunk = sdio_align_size(card->func, size); in if_sdio_card_to_host() 364 ret = sdio_readsb(card->func, card->buffer, card->ioport, chunk); in if_sdio_card_to_host() 368 chunk = card->buffer[0] | (card->buffer[1] << 8); in if_sdio_card_to_host() 372 (int)type, (int)chunk); in if_sdio_card_to_host() 374 if (chunk > size) { in if_sdio_card_to_host() 376 (int)chunk, (int)size); in if_sdio_card_to_host() 381 if (chunk < size) { in if_sdio_card_to_host() 383 (int)chunk, (int)size); in if_sdio_card_to_host() 388 ret = if_sdio_handle_cmd(card, card->buffer + 4, chunk - 4); in if_sdio_card_to_host() [all …]
|
/linux-4.4.14/sound/hda/ |
D | hdac_stream.c | 323 int chunk; in setup_bdle() local 333 chunk = snd_sgbuf_get_chunk_size(dmab, ofs, size); in setup_bdle() 338 if (chunk > remain) in setup_bdle() 339 chunk = remain; in setup_bdle() 341 bdl[2] = cpu_to_le32(chunk); in setup_bdle() 345 size -= chunk; in setup_bdle() 349 ofs += chunk; in setup_bdle()
|
/linux-4.4.14/arch/s390/crypto/ |
D | prng.c | 506 int chunk, n, tmp, ret = 0; in prng_tdes_read() local 534 chunk = min_t(int, nbytes, prng_chunk_size); in prng_tdes_read() 537 n = (chunk + 7) & -8; in prng_tdes_read() 568 if (copy_to_user(ubuf, prng_data->buf, chunk)) in prng_tdes_read() 571 nbytes -= chunk; in prng_tdes_read() 572 ret += chunk; in prng_tdes_read() 573 ubuf += chunk; in prng_tdes_read()
|
/linux-4.4.14/drivers/zorro/ |
D | zorro.c | 102 u32 chunk = start>>Z2RAM_CHUNKSHIFT; in mark_region() local 104 set_bit(chunk, zorro_unused_z2ram); in mark_region() 106 clear_bit(chunk, zorro_unused_z2ram); in mark_region()
|
/linux-4.4.14/net/rxrpc/ |
D | ar-output.c | 547 size_t size, chunk, max, space; in rxrpc_send_data() local 566 chunk = max; in rxrpc_send_data() 567 if (chunk > msg_data_left(msg) && !more) in rxrpc_send_data() 568 chunk = msg_data_left(msg); in rxrpc_send_data() 570 space = chunk + call->conn->size_align; in rxrpc_send_data() 575 _debug("SIZE: %zu/%zu/%zu", chunk, space, size); in rxrpc_send_data() 594 sp->remain = chunk; in rxrpc_send_data()
|
/linux-4.4.14/Documentation/mtd/nand/ |
D | pxa3xx-nand.txt | 15 For instance, if we choose a 2048 data chunk and set "BCH" ECC (see below) 30 OOB, one per chunk read. 51 In the BCH mode the ECC code will be calculated for each transferred chunk 55 So, repeating the above scheme, a 2048B data chunk will be followed by 32B
|
/linux-4.4.14/init/ |
D | do_mounts_md.c | 37 int chunk; member 101 md_setup_args[ent].chunk = 1 << (factor+12); in md_setup() 211 ainfo.chunk_size = md_setup_args[ent].chunk; in md_setup_drive()
|
/linux-4.4.14/drivers/usb/early/ |
D | ehci-dbgp.c | 916 int chunk, ret; in early_dbgp_write() local 941 for (chunk = 0; chunk < DBGP_MAX_PACKET && n > 0; in early_dbgp_write() 942 str++, chunk++, n--) { in early_dbgp_write() 945 buf[chunk] = '\r'; in early_dbgp_write() 952 buf[chunk] = *str; in early_dbgp_write() 954 if (chunk > 0) { in early_dbgp_write() 956 dbgp_endpoint_out, buf, chunk); in early_dbgp_write()
|
/linux-4.4.14/security/keys/ |
D | keyring.c | 233 unsigned long chunk = 0; in keyring_get_key_chunk() local 235 int desc_len = index_key->desc_len, n = sizeof(chunk); in keyring_get_key_chunk() 250 offset += sizeof(chunk) - 1; in keyring_get_key_chunk() 251 offset += (level - 3) * sizeof(chunk); in keyring_get_key_chunk() 259 chunk <<= 8; in keyring_get_key_chunk() 260 chunk |= ((u8*)index_key->description)[--offset]; in keyring_get_key_chunk() 264 chunk <<= 8; in keyring_get_key_chunk() 265 chunk |= (u8)((unsigned long)index_key->type >> in keyring_get_key_chunk() 268 return chunk; in keyring_get_key_chunk()
|
/linux-4.4.14/drivers/input/touchscreen/ |
D | wdt87xx_i2c.c | 686 static int wdt87xx_write_firmware(struct i2c_client *client, const void *chunk) in wdt87xx_write_firmware() argument 688 u32 start_addr = get_unaligned_le32(chunk + FW_CHUNK_TGT_START_OFFSET); in wdt87xx_write_firmware() 689 u32 size = get_unaligned_le32(chunk + FW_CHUNK_PAYLOAD_LEN_OFFSET); in wdt87xx_write_firmware() 690 const void *data = chunk + FW_CHUNK_PAYLOAD_OFFSET; in wdt87xx_write_firmware() 788 const void *chunk; in wdt87xx_load_chunk() local 791 chunk = wdt87xx_get_fw_chunk(fw, ck_id); in wdt87xx_load_chunk() 792 if (!chunk) { in wdt87xx_load_chunk() 798 error = wdt87xx_validate_fw_chunk(chunk, ck_id); in wdt87xx_load_chunk() 805 error = wdt87xx_write_firmware(client, chunk); in wdt87xx_load_chunk()
|
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_cs.c | 625 struct amdgpu_cs_chunk *chunk; in amdgpu_cs_ib_fill() local 630 chunk = &parser->chunks[i]; in amdgpu_cs_ib_fill() 632 chunk_ib = (struct drm_amdgpu_cs_chunk_ib *)chunk->kdata; in amdgpu_cs_ib_fill() 634 if (chunk->chunk_id != AMDGPU_CHUNK_ID_IB) in amdgpu_cs_ib_fill() 747 struct amdgpu_cs_chunk *chunk; in amdgpu_cs_dependencies() local 750 chunk = &p->chunks[i]; in amdgpu_cs_dependencies() 752 if (chunk->chunk_id != AMDGPU_CHUNK_ID_DEPENDENCIES) in amdgpu_cs_dependencies() 755 deps = (struct drm_amdgpu_cs_chunk_dep *)chunk->kdata; in amdgpu_cs_dependencies() 756 num_deps = chunk->length_dw * 4 / in amdgpu_cs_dependencies()
|
/linux-4.4.14/drivers/usb/ |
D | usb-skeleton.c | 289 size_t chunk = min(available, count); in skel_read() local 309 chunk)) in skel_read() 312 rv = chunk; in skel_read() 314 dev->bulk_in_copied += chunk; in skel_read() 321 skel_do_read_io(dev, count - chunk); in skel_read()
|
/linux-4.4.14/drivers/dma/ |
D | at_xdmac.c | 853 struct data_chunk *chunk) in at_xdmac_interleaved_queue_desc() argument 873 dwidth = at_xdmac_align_width(chan, src | dst | chunk->size); in at_xdmac_interleaved_queue_desc() 874 if (chunk->size >= (AT_XDMAC_MBR_UBC_UBLEN_MAX << dwidth)) { in at_xdmac_interleaved_queue_desc() 877 __func__, chunk->size, in at_xdmac_interleaved_queue_desc() 910 ublen = chunk->size >> dwidth; in at_xdmac_interleaved_queue_desc() 914 desc->lld.mbr_sus = dmaengine_get_src_icg(xt, chunk); in at_xdmac_interleaved_queue_desc() 915 desc->lld.mbr_dus = dmaengine_get_dst_icg(xt, chunk); in at_xdmac_interleaved_queue_desc() 944 struct data_chunk *chunk; in at_xdmac_prep_interleaved() local 982 chunk = xt->sgl + i; in at_xdmac_prep_interleaved() 984 dst_icg = dmaengine_get_dst_icg(xt, chunk); in at_xdmac_prep_interleaved() [all …]
|
D | at_hdmac.c | 743 struct data_chunk *chunk = xt->sgl + i; in atc_prep_dma_interleaved() local 745 if ((chunk->size != xt->sgl->size) || in atc_prep_dma_interleaved() 746 (dmaengine_get_dst_icg(xt, chunk) != dmaengine_get_dst_icg(xt, first)) || in atc_prep_dma_interleaved() 747 (dmaengine_get_src_icg(xt, chunk) != dmaengine_get_src_icg(xt, first))) { in atc_prep_dma_interleaved() 754 len += chunk->size; in atc_prep_dma_interleaved()
|
/linux-4.4.14/fs/nfs/blocklayout/ |
D | dev.c | 158 u64 chunk; in bl_map_stripe() local 162 chunk = div_u64(offset, dev->chunk_size); in bl_map_stripe() 163 div_u64_rem(chunk, dev->nr_children, &chunk_idx); in bl_map_stripe() 173 offset = chunk * dev->chunk_size; in bl_map_stripe()
|
/linux-4.4.14/drivers/net/wireless/ath/ar5523/ |
D | ar5523.c | 528 struct ar5523_chunk *chunk; in ar5523_data_rx_cb() local 549 chunk = (struct ar5523_chunk *) data->skb->data; in ar5523_data_rx_cb() 551 if (((chunk->flags & UATH_CFLAGS_FINAL) == 0) || in ar5523_data_rx_cb() 552 chunk->seqnum != 0) { in ar5523_data_rx_cb() 554 chunk->seqnum, chunk->flags, in ar5523_data_rx_cb() 555 be16_to_cpu(chunk->length)); in ar5523_data_rx_cb() 581 skb_reserve(data->skb, sizeof(*chunk)); in ar5523_data_rx_cb() 793 struct ar5523_chunk *chunk; in ar5523_tx_work_locked() local 836 chunk = (struct ar5523_chunk *)skb_push(skb, sizeof(*chunk)); in ar5523_tx_work_locked() 838 chunk->seqnum = 0; in ar5523_tx_work_locked() [all …]
|
/linux-4.4.14/net/core/ |
D | datagram.c | 710 int chunk = skb->len - hlen; in skb_copy_and_csum_datagram_msg() local 712 if (!chunk) in skb_copy_and_csum_datagram_msg() 715 if (msg_data_left(msg) < chunk) { in skb_copy_and_csum_datagram_msg() 718 if (skb_copy_datagram_msg(skb, hlen, msg, chunk)) in skb_copy_and_csum_datagram_msg() 723 chunk, &csum)) in skb_copy_and_csum_datagram_msg()
|
D | skbuff.c | 4464 unsigned long chunk; in alloc_skb_with_frags() local 4510 chunk = min_t(unsigned long, data_len, in alloc_skb_with_frags() 4512 skb_fill_page_desc(skb, i, page, 0, chunk); in alloc_skb_with_frags() 4513 data_len -= chunk; in alloc_skb_with_frags()
|
/linux-4.4.14/Documentation/networking/ |
D | netconsole.txt | 162 For example, assuming a lot smaller chunk size, a message "the first 163 chunk, the 2nd chunk." may be split as follows. 165 6,416,1758426,-,ncfrag=0/31;the first chunk, 166 6,416,1758426,-,ncfrag=16/31; the 2nd chunk.
|
/linux-4.4.14/include/linux/ |
D | dmaengine.h | 971 struct data_chunk *chunk) in dmaengine_get_dst_icg() argument 974 chunk->icg, chunk->dst_icg); in dmaengine_get_dst_icg() 978 struct data_chunk *chunk) in dmaengine_get_src_icg() argument 981 chunk->icg, chunk->src_icg); in dmaengine_get_src_icg()
|
/linux-4.4.14/drivers/nfc/pn544/ |
D | i2c.c | 677 struct pn544_i2c_fw_secure_frame *chunk; in pn544_hci_i2c_fw_secure_write_frame_cmd() local 684 chunk = (struct pn544_i2c_fw_secure_frame *) buf; in pn544_hci_i2c_fw_secure_write_frame_cmd() 686 chunk->cmd = PN544_FW_CMD_SECURE_CHUNK_WRITE; in pn544_hci_i2c_fw_secure_write_frame_cmd() 688 put_unaligned_be16(datalen, &chunk->be_datalen); in pn544_hci_i2c_fw_secure_write_frame_cmd() 690 memcpy(chunk->data, data, datalen); in pn544_hci_i2c_fw_secure_write_frame_cmd() 692 chunklen = sizeof(chunk->cmd) + sizeof(chunk->be_datalen) + datalen; in pn544_hci_i2c_fw_secure_write_frame_cmd()
|
/linux-4.4.14/drivers/net/ethernet/sfc/ |
D | mcdi.c | 1952 size_t chunk; in efx_mcdi_mtd_read() local 1956 chunk = min_t(size_t, end - offset, EFX_MCDI_NVRAM_LEN_MAX); in efx_mcdi_mtd_read() 1958 buffer, chunk); in efx_mcdi_mtd_read() 1961 offset += chunk; in efx_mcdi_mtd_read() 1962 buffer += chunk; in efx_mcdi_mtd_read() 1975 size_t chunk = part->common.mtd.erasesize; in efx_mcdi_mtd_erase() local 1990 chunk); in efx_mcdi_mtd_erase() 1993 offset += chunk; in efx_mcdi_mtd_erase() 2006 size_t chunk; in efx_mcdi_mtd_write() local 2017 chunk = min_t(size_t, end - offset, EFX_MCDI_NVRAM_LEN_MAX); in efx_mcdi_mtd_write() [all …]
|
/linux-4.4.14/net/ipv4/ |
D | tcp.c | 1772 int chunk; in tcp_recvmsg() local 1776 chunk = len - tp->ucopy.len; in tcp_recvmsg() 1777 if (chunk != 0) { in tcp_recvmsg() 1778 NET_ADD_STATS_USER(sock_net(sk), LINUX_MIB_TCPDIRECTCOPYFROMBACKLOG, chunk); in tcp_recvmsg() 1779 len -= chunk; in tcp_recvmsg() 1780 copied += chunk; in tcp_recvmsg() 1788 chunk = len - tp->ucopy.len; in tcp_recvmsg() 1789 if (chunk != 0) { in tcp_recvmsg() 1790 NET_ADD_STATS_USER(sock_net(sk), LINUX_MIB_TCPDIRECTCOPYFROMPREQUEUE, chunk); in tcp_recvmsg() 1791 len -= chunk; in tcp_recvmsg() [all …]
|
D | tcp_input.c | 4562 int chunk = min_t(unsigned int, skb->len, in tcp_data_queue() local 4568 if (!skb_copy_datagram_msg(skb, 0, tp->ucopy.msg, chunk)) { in tcp_data_queue() 4569 tp->ucopy.len -= chunk; in tcp_data_queue() 4570 tp->copied_seq += chunk; in tcp_data_queue() 4571 eaten = (chunk == skb->len); in tcp_data_queue() 5087 int chunk = skb->len - hlen; in tcp_copy_to_iovec() local 5092 err = skb_copy_datagram_msg(skb, hlen, tp->ucopy.msg, chunk); in tcp_copy_to_iovec() 5097 tp->ucopy.len -= chunk; in tcp_copy_to_iovec() 5098 tp->copied_seq += chunk; in tcp_copy_to_iovec()
|
D | tcp_output.c | 1798 u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); in tcp_tso_should_defer() local 1803 chunk /= win_divisor; in tcp_tso_should_defer() 1804 if (limit >= chunk) in tcp_tso_should_defer()
|
/linux-4.4.14/sound/pci/lola/ |
D | lola_pcm.c | 326 int chunk; in setup_bdle() local 336 chunk = snd_pcm_sgbuf_get_chunk_size(substream, ofs, size); in setup_bdle() 337 bdl[2] = cpu_to_le32(chunk); in setup_bdle() 341 size -= chunk; in setup_bdle() 345 ofs += chunk; in setup_bdle()
|
/linux-4.4.14/drivers/infiniband/hw/qib/ |
D | qib_init.c | 1672 unsigned e, egrcnt, egrperchunk, chunk, egrsize, egroff; in qib_setup_eagerbufs() local 1689 chunk = rcd->rcvegrbuf_chunks; in qib_setup_eagerbufs() 1694 kzalloc_node(chunk * sizeof(rcd->rcvegrbuf[0]), in qib_setup_eagerbufs() 1701 kmalloc_node(chunk * sizeof(rcd->rcvegrbuf_phys[0]), in qib_setup_eagerbufs() 1723 for (e = chunk = 0; chunk < rcd->rcvegrbuf_chunks; chunk++) { in qib_setup_eagerbufs() 1724 dma_addr_t pa = rcd->rcvegrbuf_phys[chunk]; in qib_setup_eagerbufs() 1728 memset(rcd->rcvegrbuf[chunk], 0, size); in qib_setup_eagerbufs()
|
D | qib_driver.c | 286 const u32 chunk = etail >> rcd->rcvegrbufs_perchunk_shift; in qib_get_egrbuf() local 289 return rcd->rcvegrbuf[chunk] + (idx << rcd->dd->rcvegrbufsize_shift); in qib_get_egrbuf()
|
/linux-4.4.14/net/unix/ |
D | af_unix.c | 2304 int chunk; in unix_stream_read_generic() local 2385 chunk = min_t(unsigned int, unix_skb_len(skb) - skip, size); in unix_stream_read_generic() 2387 chunk = state->recv_actor(skb, skip, chunk, state); in unix_stream_read_generic() 2391 if (chunk < 0) { in unix_stream_read_generic() 2396 copied += chunk; in unix_stream_read_generic() 2397 size -= chunk; in unix_stream_read_generic() 2413 UNIXCB(skb).consumed += chunk; in unix_stream_read_generic() 2415 sk_peek_offset_bwd(sk, chunk); in unix_stream_read_generic() 2434 sk_peek_offset_fwd(sk, chunk); in unix_stream_read_generic() 2461 int skip, int chunk, in unix_stream_read_actor() argument [all …]
|
/linux-4.4.14/drivers/char/ |
D | mem.c | 639 size_t chunk = iov_iter_count(iter), n; in read_iter_zero() local 641 if (chunk > PAGE_SIZE) in read_iter_zero() 642 chunk = PAGE_SIZE; /* Just for latency reasons */ in read_iter_zero() 643 n = iov_iter_zero(chunk, iter); in read_iter_zero()
|
D | random.c | 1342 int chunk = min(nbytes, (int)sizeof(unsigned long)); in get_random_bytes_arch() local 1347 memcpy(p, &v, chunk); in get_random_bytes_arch() 1348 p += chunk; in get_random_bytes_arch() 1349 nbytes -= chunk; in get_random_bytes_arch()
|
/linux-4.4.14/net/caif/ |
D | caif_socket.c | 373 int chunk; in caif_stream_recvmsg() local 424 chunk = min_t(unsigned int, skb->len, size); in caif_stream_recvmsg() 425 if (memcpy_to_msg(msg, skb->data, chunk)) { in caif_stream_recvmsg() 431 copied += chunk; in caif_stream_recvmsg() 432 size -= chunk; in caif_stream_recvmsg() 436 skb_pull(skb, chunk); in caif_stream_recvmsg()
|
/linux-4.4.14/Documentation/fb/ |
D | sh7760fb.txt | 11 * Framebuffer memory must be a large chunk allocated at the top 14 get a large enough contiguous chunk of memory.
|
/linux-4.4.14/drivers/irqchip/ |
D | irq-gic-v3-its.c | 663 static int its_chunk_to_lpi(int chunk) in its_chunk_to_lpi() argument 665 return (chunk << IRQS_PER_CHUNK_SHIFT) + 8192; in its_chunk_to_lpi() 735 int chunk = its_lpi_to_chunk(lpi); in its_lpi_free() local 736 BUG_ON(chunk > lpi_chunks); in its_lpi_free() 737 if (test_bit(chunk, lpi_bitmap)) { in its_lpi_free() 738 clear_bit(chunk, lpi_bitmap); in its_lpi_free() 740 pr_err("Bad LPI chunk %d\n", chunk); in its_lpi_free()
|
/linux-4.4.14/Documentation/ |
D | dell_rbu.txt | 20 image methods. In case of monolithic the driver allocates a contiguous chunk 57 The user creates packets header, gets the chunk of the BIOS image and 58 places it next to the packetheader; now, the packetheader + BIOS image chunk
|
D | assoc_array.txt | 139 (1) Get a chunk of index key from caller data: 143 This should return a chunk of caller-supplied index key starting at the 149 (2) Get a chunk of an object's index key. 360 The index key is read in chunks of machine word. Each chunk is subdivided into
|
D | dma-buf-sharing.txt | 278 PAGE_SIZE size. Before accessing a chunk it needs to be mapped, which returns 279 a pointer in kernel virtual address space. Afterwards the chunk needs to be 280 unmapped again. There is no limit on how often a given chunk can be mapped 282 before mapping the same chunk again.
|
D | DMA-attributes.txt | 100 be mapped as contiguous chunk into device dma address space. By
|
D | ramoops.txt | 32 power of two) and each oops/panic writes a "record_size" chunk of
|
D | md.txt | 12 md=<md device no.>,<raid level>,<chunk size factor>,<fault level>,dev0,dev1,...,devn 30 chunk size factor = (raid-0 and raid-1 only) 31 Set the chunk size as 4k << n. 309 The size, in bytes, of the chunk which will be represented by a
|
D | xillybus.txt | 87 their attributes), there isn't one specific chunk of logic being the Xillybus
|
/linux-4.4.14/drivers/staging/media/lirc/ |
D | lirc_serial.c | 385 unsigned char chunk, shifted; in send_pulse_irdeo() local 390 chunk = 3; in send_pulse_irdeo() 392 chunk = 1; in send_pulse_irdeo() 394 shifted = chunk << (i * 3); in send_pulse_irdeo()
|
/linux-4.4.14/include/drm/ |
D | drm_dp_mst_helper.h | 206 u8 chunk[48]; member 361 u8 chunk[48]; member
|
/linux-4.4.14/drivers/tty/ |
D | tty_io.c | 1111 unsigned int chunk; in do_tty_write() local 1133 chunk = 2048; in do_tty_write() 1135 chunk = 65536; in do_tty_write() 1136 if (count < chunk) in do_tty_write() 1137 chunk = count; in do_tty_write() 1140 if (tty->write_cnt < chunk) { in do_tty_write() 1143 if (chunk < 1024) in do_tty_write() 1144 chunk = 1024; in do_tty_write() 1146 buf_chunk = kmalloc(chunk, GFP_KERNEL); in do_tty_write() 1152 tty->write_cnt = chunk; in do_tty_write() [all …]
|
/linux-4.4.14/arch/xtensa/lib/ |
D | memcopy.S | 151 add a8, a8, a3 # a8 = end of last 16B source chunk 227 add a10, a10, a3 # a10 = end of last 16B source chunk 425 sub a8, a3, a8 # a8 = start of first 16B source chunk 502 sub a10, a3, a10 # a10 = start of first 16B source chunk
|
D | usercopy.S | 161 add a8, a8, a3 # a8 = end of last 16B source chunk 226 add a12, a12, a3 # a12 = end of last 16B source chunk
|
D | memset.S | 73 add a6, a6, a5 # a6 = end of last 16B chunk
|
D | strnlen_user.S | 77 add a10, a10, a4 # a10 = end of last 4B chunk
|
/linux-4.4.14/fs/gfs2/ |
D | bmap.c | 1013 u64 chunk; in gfs2_journaled_truncate() local 1017 chunk = oldsize - newsize; in gfs2_journaled_truncate() 1018 if (chunk > max_chunk) in gfs2_journaled_truncate() 1019 chunk = max_chunk; in gfs2_journaled_truncate() 1020 truncate_pagecache(inode, oldsize - chunk); in gfs2_journaled_truncate() 1021 oldsize -= chunk; in gfs2_journaled_truncate()
|
/linux-4.4.14/arch/arm/kernel/ |
D | traps.c | 540 unsigned long chunk = min(PAGE_SIZE, end - start); in __do_cache_op() local 545 ret = flush_cache_user_range(start, start + chunk); in __do_cache_op() 550 start += chunk; in __do_cache_op()
|
/linux-4.4.14/drivers/memstick/core/ |
D | mspro_block.c | 697 static int mspro_block_issue_req(struct memstick_dev *card, int chunk) in mspro_block_issue_req() argument 704 while (chunk) { in mspro_block_issue_req() 712 chunk = __blk_end_request_cur(msb->block_req, -ENOMEM); in mspro_block_issue_req() 739 chunk = 1; in mspro_block_issue_req() 746 int chunk, cnt; in mspro_block_complete_req() local 779 chunk = __blk_end_request(msb->block_req, error, t_len); in mspro_block_complete_req() 781 error = mspro_block_issue_req(card, chunk); in mspro_block_complete_req()
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
D | radeon_cs.c | 78 struct radeon_cs_chunk *chunk; in radeon_cs_parser_relocs() local 87 chunk = p->chunk_relocs; in radeon_cs_parser_relocs() 90 p->nrelocs = chunk->length_dw / 4; in radeon_cs_parser_relocs() 103 r = (struct drm_radeon_cs_reloc *)&chunk->kdata[i*4]; in radeon_cs_parser_relocs()
|
/linux-4.4.14/drivers/thermal/ti-soc-thermal/ |
D | ti-bandgap.c | 1236 void __iomem *chunk; in ti_bandgap_build() local 1241 chunk = devm_ioremap_resource(&pdev->dev, res); in ti_bandgap_build() 1243 bgp->base = chunk; in ti_bandgap_build() 1244 if (IS_ERR(chunk)) in ti_bandgap_build() 1245 return ERR_CAST(chunk); in ti_bandgap_build()
|
/linux-4.4.14/drivers/staging/lustre/lustre/osc/ |
D | osc_cache.c | 633 pgoff_t chunk; in osc_extent_find() local 653 chunk = index >> ppc_bits; in osc_extent_find() 688 if (chunk > ext_chk_end + 1) in osc_extent_find() 701 if (chunk + 1 < ext_chk_start) { in osc_extent_find() 754 if (chunk + 1 == ext_chk_start) { in osc_extent_find() 764 } else if (chunk == ext_chk_end + 1) { in osc_extent_find() 1145 pgoff_t chunk = index >> ppc_bits; in osc_extent_expand() local 1155 if (chunk > end_chunk + 1) { in osc_extent_expand() 1160 if (end_chunk >= chunk) { in osc_extent_expand() 1165 LASSERT(end_chunk + 1 == chunk); in osc_extent_expand() [all …]
|
/linux-4.4.14/arch/arm/lib/ |
D | memzero.S | 42 cmp r1, #16 @ 1 we can skip this chunk if we
|
/linux-4.4.14/drivers/mmc/host/ |
D | sdhci.c | 305 size_t blksize, len, chunk; in sdhci_read_block_pio() local 312 chunk = 0; in sdhci_read_block_pio() 327 if (chunk == 0) { in sdhci_read_block_pio() 329 chunk = 4; in sdhci_read_block_pio() 336 chunk--; in sdhci_read_block_pio() 349 size_t blksize, len, chunk; in sdhci_write_block_pio() local 356 chunk = 0; in sdhci_write_block_pio() 372 scratch |= (u32)*buf << (chunk * 8); in sdhci_write_block_pio() 375 chunk++; in sdhci_write_block_pio() 378 if ((chunk == 4) || ((len == 0) && (blksize == 0))) { in sdhci_write_block_pio() [all …]
|
/linux-4.4.14/drivers/macintosh/ |
D | smu.c | 899 unsigned int chunk; in smu_read_datablock() local 907 chunk = 0xe; in smu_read_datablock() 910 unsigned int clen = min(len, chunk); in smu_read_datablock() 915 cmd.reply_len = chunk; in smu_read_datablock()
|
/linux-4.4.14/arch/powerpc/kernel/ |
D | vio.c | 393 size_t avail = 0, level, chunk, need; in vio_cmo_balance() local 441 chunk = min_t(size_t, avail, VIO_CMO_BALANCE_CHUNK); in vio_cmo_balance() 442 chunk = min(chunk, (viodev->cmo.desired - in vio_cmo_balance() 444 viodev->cmo.entitled += chunk; in vio_cmo_balance()
|
D | prom_init.c | 2016 unsigned long room, chunk; in make_room() local 2026 chunk = alloc_up(room, 0); in make_room() 2027 if (chunk == 0) in make_room() 2030 *mem_end = chunk + room; in make_room()
|
/linux-4.4.14/drivers/net/wimax/i2400m/ |
D | fw.c | 646 static int i2400m_download_chunk(struct i2400m *i2400m, const void *chunk, in i2400m_download_chunk() argument 660 "direct %u do_csum %u)\n", i2400m, chunk, __chunk_len, in i2400m_download_chunk() 663 memcpy(buf->cmd_payload, chunk, __chunk_len); in i2400m_download_chunk() 676 "direct %u do_csum %u) = %d\n", i2400m, chunk, __chunk_len, in i2400m_download_chunk()
|
/linux-4.4.14/drivers/crypto/qat/qat_common/ |
D | qat_uclo.c | 501 char *chunk; in qat_uclo_map_chunk() local 509 chunk = buf + file_chunk->offset; in qat_uclo_map_chunk() 511 chunk, file_chunk->size)) in qat_uclo_map_chunk() 516 obj_hdr->file_buff = chunk; in qat_uclo_map_chunk()
|
/linux-4.4.14/drivers/staging/rdma/ipath/ |
D | ipath_file_ops.c | 897 unsigned e, egrcnt, egrperchunk, chunk, egrsize, egroff; in ipath_create_user_egr() local 917 chunk = pd->port_rcvegrbuf_chunks; in ipath_create_user_egr() 920 pd->port_rcvegrbuf = kmalloc(chunk * sizeof(pd->port_rcvegrbuf[0]), in ipath_create_user_egr() 927 kmalloc(chunk * sizeof(pd->port_rcvegrbuf_phys[0]), in ipath_create_user_egr() 947 for (e = chunk = 0; chunk < pd->port_rcvegrbuf_chunks; chunk++) { in ipath_create_user_egr() 948 dma_addr_t pa = pd->port_rcvegrbuf_phys[chunk]; in ipath_create_user_egr()
|
/linux-4.4.14/drivers/media/dvb-frontends/ |
D | drxk_hard.h | 224 u8 chunk[256]; member
|
D | drxk_hard.c | 449 int chunk = blk_size > state->m_chunk_size ? in write_block() local 451 u8 *adr_buf = &state->chunk[0]; in write_block() 461 if (chunk == state->m_chunk_size) in write_block() 462 chunk -= 2; in write_block() 469 memcpy(&state->chunk[adr_length], p_block, chunk); in write_block() 474 for (i = 0; i < chunk; i++) in write_block() 479 &state->chunk[0], chunk + adr_length); in write_block() 485 p_block += chunk; in write_block() 486 address += (chunk >> 1); in write_block() 487 blk_size -= chunk; in write_block()
|
/linux-4.4.14/drivers/input/misc/ |
D | ims-pcu.c | 515 u8 command, int chunk, int len) in ims_pcu_send_cmd_chunk() argument 527 command, chunk, error); in ims_pcu_send_cmd_chunk() 538 int chunk = 0; in ims_pcu_send_command() local 566 ++chunk, count); in ims_pcu_send_command() 584 error = ims_pcu_send_cmd_chunk(pcu, command, ++chunk, count); in ims_pcu_send_command() 597 return ims_pcu_send_cmd_chunk(pcu, command, ++chunk, count); in ims_pcu_send_command()
|
/linux-4.4.14/net/irda/ |
D | af_irda.c | 1430 int chunk; in irda_recvmsg_stream() local 1470 chunk = min_t(unsigned int, skb->len, size); in irda_recvmsg_stream() 1471 if (memcpy_to_msg(msg, skb->data, chunk)) { in irda_recvmsg_stream() 1477 copied += chunk; in irda_recvmsg_stream() 1478 size -= chunk; in irda_recvmsg_stream() 1482 skb_pull(skb, chunk); in irda_recvmsg_stream()
|
/linux-4.4.14/net/sunrpc/xprtrdma/ |
D | svc_rdma_sendto.c | 196 int chunk = be32_to_cpu(wr_ary->wc_nchunks); in svc_rdma_get_reply_array() local 199 &wr_ary->wc_array[chunk].wc_target.rs_length; in svc_rdma_get_reply_array()
|
/linux-4.4.14/net/decnet/ |
D | af_decnet.c | 1760 unsigned int chunk = skb->len; in dn_recvmsg() local 1763 if ((chunk + copied) > size) in dn_recvmsg() 1764 chunk = size - copied; in dn_recvmsg() 1766 if (memcpy_to_msg(msg, skb->data, chunk)) { in dn_recvmsg() 1770 copied += chunk; in dn_recvmsg() 1773 skb_pull(skb, chunk); in dn_recvmsg()
|
/linux-4.4.14/drivers/gpu/drm/ |
D | drm_dp_mst_topology.c | 350 memcpy(&msg->chunk[0], replybuf + hdrlen, msg->curchunk_idx); in drm_dp_sideband_msg_build() 352 memcpy(&msg->chunk[msg->curchunk_idx], replybuf, replybuflen); in drm_dp_sideband_msg_build() 358 crc4 = drm_dp_msg_data_crc4(msg->chunk, msg->curchunk_len - 1); in drm_dp_sideband_msg_build() 360 memcpy(&msg->msg[msg->curlen], msg->chunk, msg->curchunk_len - 1); in drm_dp_sideband_msg_build() 1431 u8 chunk[48]; in process_single_tx_qlock() local 1463 drm_dp_encode_sideband_msg_hdr(&hdr, chunk, &idx); in process_single_tx_qlock() 1464 memcpy(&chunk[idx], &txmsg->msg[txmsg->cur_offset], tosend); in process_single_tx_qlock() 1466 drm_dp_crc_sideband_chunk_req(&chunk[idx], tosend); in process_single_tx_qlock() 1469 ret = drm_dp_send_sideband_msg(mgr, up, chunk, idx); in process_single_tx_qlock()
|
/linux-4.4.14/drivers/net/ethernet/ibm/emac/ |
D | core.c | 1421 int chunk = min(len, MAL_MAX_TX_SIZE); in emac_xmit_split() local 1422 len -= chunk; in emac_xmit_split() 1433 dev->tx_desc[slot].data_len = (u16) chunk; in emac_xmit_split() 1440 pd += chunk; in emac_xmit_split() 1450 int len = skb->len, chunk; in emac_start_xmit_sg() local 1474 chunk = min(len, MAL_MAX_TX_SIZE); in emac_start_xmit_sg() 1477 dev->tx_desc[slot].data_len = (u16) chunk; in emac_start_xmit_sg() 1478 len -= chunk; in emac_start_xmit_sg() 1480 slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags, in emac_start_xmit_sg()
|
/linux-4.4.14/security/apparmor/ |
D | policy_unpack.c | 124 static size_t unpack_u16_chunk(struct aa_ext *e, char **chunk) in unpack_u16_chunk() argument 134 *chunk = e->pos; in unpack_u16_chunk()
|
/linux-4.4.14/crypto/ |
D | drbg.c | 1433 unsigned int chunk = 0; in drbg_generate_long() local 1435 chunk = slice ? drbg_max_request_bytes(drbg) : (buflen - len); in drbg_generate_long() 1437 err = drbg_generate(drbg, buf + len, chunk, addtl); in drbg_generate_long() 1441 len += chunk; in drbg_generate_long()
|
/linux-4.4.14/Documentation/dmaengine/ |
D | provider.txt | 69 but all of them will require, for every chunk, at least the source and 188 scatter-gather transfer, with a single chunk to transfer, it's a 204 deal with a single chunk to copy or a collection of them, here, 212 - A cyclic transfer is a transfer where the chunk collection will
|
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/socklnd/ |
D | socklnd_cb.c | 2581 int chunk = ksocknal_data.ksnd_peer_hash_size; in ksocknal_reaper() local 2591 chunk = (chunk * n * p) / in ksocknal_reaper() 2593 if (chunk == 0) in ksocknal_reaper() 2594 chunk = 1; in ksocknal_reaper() 2596 for (i = 0; i < chunk; i++) { in ksocknal_reaper()
|
/linux-4.4.14/Documentation/DocBook/ |
D | mtdnand.xml.db | 14 API-nand-check-erased-ecc-chunk
|
D | alsa-driver-api.xml.db | 103 API-snd-pcm-sgbuf-get-chunk-size
|
D | kernel-api.xml.db | 658 API-blk-queue-chunk-sectors
|
/linux-4.4.14/Documentation/arm/Samsung-S3C24XX/ |
D | Suspend.txt | 128 Defines the size of memory each CRC chunk covers. A smaller value
|
/linux-4.4.14/Documentation/ko_KR/ |
D | HOWTO | 491 리눅스 커널 커뮤니티는 한꺼번에 굉장히 큰 코드의 묶음(chunk)을 쉽게
|
/linux-4.4.14/Documentation/filesystems/cifs/ |
D | TODO | 14 - T10 copy offload (copy chunk is only mechanism supported)
|
/linux-4.4.14/Documentation/ABI/testing/ |
D | sysfs-driver-wacom | 72 byte chunk encodes the image data for two consecutive lines on
|
D | sysfs-block | 129 chunk size. A properly aligned multiple of
|
/linux-4.4.14/Documentation/filesystems/ |
D | ntfs.txt | 318 For a striped volume, i.e. raid level 0, you will need to know the chunk size 394 Windows by default uses a stripe chunk size of 64k, so you probably want the 395 "chunk-size 64k" option for each raid-disk, too. 405 chunk-size 64k
|
D | sysv-fs.txt | 84 to the free blocks pertaining to the next chunk; the first of these
|
D | btrfs.txt | 141 Specify that 1 metadata chunk should be allocated after every <value>
|
D | ext4.txt | 264 disks * RAID chunk size in file system blocks.
|
/linux-4.4.14/Documentation/fpga/ |
D | fpga-mgr.txt | 163 whole FPGA image or may be a smaller chunk of an FPGA image. In the latter
|
/linux-4.4.14/drivers/acpi/apei/ |
D | ghes.c | 202 struct gen_pool_chunk *chunk, in ghes_estatus_pool_free_chunk_page() argument 205 free_page(chunk->start_addr); in ghes_estatus_pool_free_chunk_page()
|
/linux-4.4.14/drivers/mtd/nand/ |
D | nand_base.c | 1907 int chunk = chip->ecc.bytes + chip->ecc.prepad + chip->ecc.postpad; in nand_read_oob_syndrome() local 1915 pos = eccsize + i * (eccsize + chunk); in nand_read_oob_syndrome() 1922 toread = min_t(int, length, chunk); in nand_read_oob_syndrome() 1966 int chunk = chip->ecc.bytes + chip->ecc.prepad + chip->ecc.postpad; in nand_write_oob_syndrome() local 1977 pos = steps * (eccsize + chunk); in nand_write_oob_syndrome() 1996 pos = eccsize + i * (eccsize + chunk); in nand_write_oob_syndrome() 2001 len = min_t(int, length, chunk); in nand_write_oob_syndrome()
|
/linux-4.4.14/arch/sh/lib64/ |
D | copy_user_memcpy.S | 28 ! bytes memory chunk to b copied, the rest of the word can be read
|
D | memcpy.S | 30 ! bytes memory chunk to b copied, the rest of the word can be read
|
/linux-4.4.14/arch/m68k/ |
D | Kconfig.cpu | 361 bool "Use one physical chunk of memory only" if ADVANCED && !SUN3 366 Ignore all but the first contiguous chunk of physical memory for VM
|
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
D | o2iblnd_cb.c | 3134 int chunk = kiblnd_data.kib_peer_hash_size; in kiblnd_connd() local 3148 chunk = (chunk * n * p) / in kiblnd_connd() 3150 if (chunk == 0) in kiblnd_connd() 3151 chunk = 1; in kiblnd_connd() 3153 for (i = 0; i < chunk; i++) { in kiblnd_connd()
|
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_main.c | 2455 unsigned int chunk = in cxgb_extension_ioctl() local 2459 t3_mc7_bd_read(mem, t.addr / 8, chunk / 8, in cxgb_extension_ioctl() 2463 if (copy_to_user(useraddr, buf, chunk)) in cxgb_extension_ioctl() 2465 useraddr += chunk; in cxgb_extension_ioctl() 2466 t.addr += chunk; in cxgb_extension_ioctl() 2467 t.len -= chunk; in cxgb_extension_ioctl()
|
D | sge.c | 326 unsigned int chunk) in reclaim_completed_tx() argument 330 reclaim = min(chunk, reclaim); in reclaim_completed_tx()
|
/linux-4.4.14/fs/cramfs/ |
D | README | 58 applied to the i'th blksize-sized chunk of the input data.
|
/linux-4.4.14/arch/s390/kernel/ |
D | head.S | 96 ssch 0(%r3) # load chunk of 1600 bytes
|
/linux-4.4.14/Documentation/early-userspace/ |
D | README | 17 - initramfs, a chunk of code that unpacks the compressed cpio image
|
/linux-4.4.14/drivers/net/ethernet/marvell/ |
D | skge.c | 2530 u32 chunk, ram_addr; in skge_up() local 2595 chunk = (hw->ram_size - hw->ram_offset) / (hw->ports * 2); in skge_up() 2596 ram_addr = hw->ram_offset + 2 * chunk * port; in skge_up() 2598 skge_ramset(hw, rxqaddr[port], ram_addr, chunk); in skge_up() 2602 skge_ramset(hw, txqaddr[port], ram_addr+chunk, chunk); in skge_up()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/ce/fuc/ |
D | com.fuc | 208 // get a chunk of stack space, aligned to 256 byte boundary 419 // zero out a chunk of the stack to store the swizzle into
|
/linux-4.4.14/arch/cris/arch-v10/ |
D | README.mm | 228 happens first is that a virtual address chunk is allocated to the request using
|
/linux-4.4.14/drivers/scsi/ |
D | ips.h | 733 IPS_CHUNK chunk[IPS_MAX_CHUNKS]; member
|
/linux-4.4.14/Documentation/powerpc/ |
D | firmware-assisted-dump.txt | 53 NOTE: The term 'boot memory' means size of the low memory chunk
|
/linux-4.4.14/fs/ext4/ |
D | mballoc.c | 671 ext4_grpblk_t chunk; in ext4_mb_mark_free_simple() local 687 chunk = 1 << min; in ext4_mb_mark_free_simple() 695 len -= chunk; in ext4_mb_mark_free_simple() 696 first += chunk; in ext4_mb_mark_free_simple()
|
/linux-4.4.14/Documentation/scsi/ |
D | st.txt | 274 (PAGE_SIZE << ST_FIRST_ORDER) bytes if the system can give a chunk of 279 buffer as one chunk.
|
/linux-4.4.14/Documentation/usb/ |
D | usbmon.txt | 263 The size may be rounded down to the next chunk (or page). If the requested
|
D | WUSB-Design-overview.txt | 372 we issue another URB to read into the destination buffer the chunk of
|
/linux-4.4.14/scripts/ |
D | checkpatch.pl | 4685 for my $chunk (@chunks) { 4686 my ($cond, $block) = @{$chunk};
|
/linux-4.4.14/drivers/net/wireless/ath/ath10k/ |
D | wmi.c | 5217 struct host_memory_chunk *chunk; in ath10k_wmi_put_host_mem_chunks() local 5223 chunk = &chunks->items[i]; in ath10k_wmi_put_host_mem_chunks() 5224 chunk->ptr = __cpu_to_le32(ar->wmi.mem_chunks[i].paddr); in ath10k_wmi_put_host_mem_chunks() 5225 chunk->size = __cpu_to_le32(ar->wmi.mem_chunks[i].len); in ath10k_wmi_put_host_mem_chunks() 5226 chunk->req_id = __cpu_to_le32(ar->wmi.mem_chunks[i].req_id); in ath10k_wmi_put_host_mem_chunks()
|
/linux-4.4.14/arch/s390/ |
D | Kconfig | 449 hex "NUMA emulation memory chunk size"
|