/linux-4.1.27/net/sctp/ |
D | inqueue.c | 59 struct sctp_chunk *chunk, *tmp; in sctp_inq_free() local 62 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) { in sctp_inq_free() 63 list_del_init(&chunk->list); in sctp_inq_free() 64 sctp_chunk_free(chunk); in sctp_inq_free() 79 void sctp_inq_push(struct sctp_inq *q, struct sctp_chunk *chunk) in sctp_inq_push() argument 82 if (chunk->rcvr->dead) { in sctp_inq_push() 83 sctp_chunk_free(chunk); in sctp_inq_push() 92 list_add_tail(&chunk->list, &q->in_chunk_list); in sctp_inq_push() 93 if (chunk->asoc) in sctp_inq_push() 94 chunk->asoc->stats.ipackets++; in sctp_inq_push() [all …]
|
D | output.c | 61 struct sctp_chunk *chunk); 63 struct sctp_chunk *chunk); 65 struct sctp_chunk *chunk); 67 struct sctp_chunk *chunk, 87 struct sctp_chunk *chunk = NULL; in sctp_packet_config() local 94 chunk = sctp_get_ecne_prepend(packet->transport->asoc); in sctp_packet_config() 99 if (chunk) in sctp_packet_config() 100 sctp_packet_append_chunk(packet, chunk); in sctp_packet_config() 137 struct sctp_chunk *chunk, *tmp; in sctp_packet_free() local 141 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free() [all …]
|
D | outqueue.c | 218 struct sctp_chunk *chunk, *tmp; in __sctp_outq_teardown() local 224 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 227 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 228 sctp_chunk_free(chunk); in __sctp_outq_teardown() 235 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 237 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 238 sctp_chunk_free(chunk); in __sctp_outq_teardown() 244 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 246 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 247 sctp_chunk_free(chunk); in __sctp_outq_teardown() [all …]
|
D | chunk.c | 75 struct sctp_chunk *chunk; in sctp_datamsg_free() local 80 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free() 81 sctp_chunk_free(chunk); in sctp_datamsg_free() 90 struct sctp_chunk *chunk; in sctp_datamsg_destroy() local 102 chunk = list_entry(pos, struct sctp_chunk, frag_list); in sctp_datamsg_destroy() 105 asoc = chunk->asoc; in sctp_datamsg_destroy() 119 if (chunk->has_tsn) in sctp_datamsg_destroy() 124 ev = sctp_ulpevent_make_send_failed(asoc, chunk, sent, in sctp_datamsg_destroy() 130 sctp_chunk_put(chunk); in sctp_datamsg_destroy() 151 static void sctp_datamsg_assign(struct sctp_datamsg *msg, struct sctp_chunk *chunk) in sctp_datamsg_assign() argument [all …]
|
D | sm_statefuns.c | 65 struct sctp_chunk *chunk, 69 struct sctp_chunk *chunk, 73 const struct sctp_chunk *chunk); 77 const struct sctp_chunk *chunk, 98 static struct sctp_sackhdr *sctp_sm_pull_sack(struct sctp_chunk *chunk); 151 struct sctp_chunk *chunk); 168 sctp_chunk_length_valid(struct sctp_chunk *chunk, in sctp_chunk_length_valid() argument 171 __u16 chunk_length = ntohs(chunk->chunk_hdr->length); in sctp_chunk_length_valid() 174 if (unlikely(chunk->pdiscard)) in sctp_chunk_length_valid() 224 struct sctp_chunk *chunk = arg; in sctp_sf_do_4_C() local [all …]
|
D | sm_make_chunk.c | 79 static void *sctp_addto_param(struct sctp_chunk *chunk, int len, 90 static void sctp_control_set_owner_w(struct sctp_chunk *chunk) in sctp_control_set_owner_w() argument 92 struct sctp_association *asoc = chunk->asoc; in sctp_control_set_owner_w() 93 struct sk_buff *skb = chunk->skb; in sctp_control_set_owner_w() 107 int sctp_chunk_iif(const struct sctp_chunk *chunk) in sctp_chunk_iif() argument 112 af = sctp_get_af_specific(ipver2af(ip_hdr(chunk->skb)->version)); in sctp_chunk_iif() 114 iif = af->skb_iif(chunk->skb); in sctp_chunk_iif() 137 void sctp_init_cause(struct sctp_chunk *chunk, __be16 cause_code, in sctp_init_cause() argument 147 chunk->subh.err_hdr = sctp_addto_chunk(chunk, sizeof(sctp_errhdr_t), &err); in sctp_init_cause() 155 static int sctp_init_cause_fixed(struct sctp_chunk *chunk, __be16 cause_code, in sctp_init_cause_fixed() argument [all …]
|
D | sm_sideeffect.c | 102 struct sctp_chunk *chunk) in sctp_do_ecn_ecne_work() argument 134 repl = sctp_make_cwr(asoc, asoc->last_cwr_tsn, chunk); in sctp_do_ecn_ecne_work() 571 struct sctp_chunk *chunk, in sctp_cmd_assoc_failed() argument 579 if (event_type == SCTP_EVENT_T_CHUNK && subtype.chunk == SCTP_CID_ABORT) in sctp_cmd_assoc_failed() 581 (__u16)error, 0, 0, chunk, in sctp_cmd_assoc_failed() 592 abort = sctp_make_violation_max_retrans(asoc, chunk); in sctp_cmd_assoc_failed() 613 struct sctp_chunk *chunk, in sctp_cmd_process_init() argument 624 if (!sctp_process_init(asoc, chunk, sctp_source(chunk), peer_init, gfp)) in sctp_cmd_process_init() 690 struct sctp_chunk *chunk) in sctp_cmd_transport_on() argument 746 hbinfo = (sctp_sender_hb_info_t *) chunk->skb->data; in sctp_cmd_transport_on() [all …]
|
D | endpointola.c | 413 struct sctp_chunk *chunk; in sctp_endpoint_bh_rcv() local 428 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_endpoint_bh_rcv() 429 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_endpoint_bh_rcv() 434 if (first_time && (subtype.chunk == SCTP_CID_AUTH)) { in sctp_endpoint_bh_rcv() 447 chunk->auth_chunk = skb_clone(chunk->skb, in sctp_endpoint_bh_rcv() 449 chunk->auth = 1; in sctp_endpoint_bh_rcv() 460 if (NULL == chunk->asoc) { in sctp_endpoint_bh_rcv() 462 sctp_source(chunk), in sctp_endpoint_bh_rcv() 464 chunk->asoc = asoc; in sctp_endpoint_bh_rcv() 465 chunk->transport = transport; in sctp_endpoint_bh_rcv() [all …]
|
D | ulpevent.c | 129 __u16 inbound, struct sctp_chunk *chunk, gfp_t gfp) in sctp_ulpevent_make_assoc_change() argument 138 if (chunk) { in sctp_ulpevent_make_assoc_change() 142 skb = skb_copy_expand(chunk->skb, in sctp_ulpevent_make_assoc_change() 158 ntohs(chunk->chunk_hdr->length) - in sctp_ulpevent_make_assoc_change() 371 struct sctp_chunk *chunk, __u16 flags, in sctp_ulpevent_make_remote_error() argument 381 ch = (sctp_errhdr_t *)(chunk->skb->data); in sctp_ulpevent_make_remote_error() 386 skb_pull(chunk->skb, sizeof(sctp_errhdr_t)); in sctp_ulpevent_make_remote_error() 391 skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp); in sctp_ulpevent_make_remote_error() 394 skb_pull(chunk->skb, elen); in sctp_ulpevent_make_remote_error() 427 const struct sctp_association *asoc, struct sctp_chunk *chunk, in sctp_ulpevent_make_send_failed() argument [all …]
|
D | input.c | 100 struct sctp_chunk *chunk; member 114 struct sctp_chunk *chunk; in sctp_rcv() local 224 chunk = sctp_chunkify(skb, asoc, sk); in sctp_rcv() 225 if (!chunk) in sctp_rcv() 227 SCTP_INPUT_CB(skb)->chunk = chunk; in sctp_rcv() 230 chunk->rcvr = rcvr; in sctp_rcv() 233 chunk->sctp_hdr = sh; in sctp_rcv() 236 sctp_init_addrs(chunk, &src, &dest); in sctp_rcv() 239 chunk->transport = transport; in sctp_rcv() 263 sctp_chunk_free(chunk); in sctp_rcv() [all …]
|
D | debug.c | 65 if (cid.chunk <= SCTP_CID_BASE_MAX) in sctp_cname() 66 return sctp_cid_tbl[cid.chunk]; in sctp_cname() 68 switch (cid.chunk) { in sctp_cname()
|
D | auth.c | 391 struct sctp_chunk *chunk; in sctp_auth_asoc_init_active_key() local 417 list_for_each_entry(chunk, &asoc->outqueue.out_chunk_list, list) { in sctp_auth_asoc_init_active_key() 418 if (sctp_auth_send_cid(chunk->chunk_hdr->type, asoc)) in sctp_auth_asoc_init_active_key() 419 chunk->auth = 1; in sctp_auth_asoc_init_active_key() 638 static int __sctp_auth_cid(sctp_cid_t chunk, struct sctp_chunks_param *param) in __sctp_auth_cid() argument 664 if (param->chunks[i] == chunk) in __sctp_auth_cid() 674 int sctp_auth_send_cid(sctp_cid_t chunk, const struct sctp_association *asoc) in sctp_auth_send_cid() argument 682 return __sctp_auth_cid(chunk, asoc->peer.peer_chunks); in sctp_auth_send_cid() 686 int sctp_auth_recv_cid(sctp_cid_t chunk, const struct sctp_association *asoc) in sctp_auth_recv_cid() argument 694 return __sctp_auth_cid(chunk, in sctp_auth_recv_cid()
|
D | associola.c | 919 struct sctp_chunk *chunk; in sctp_assoc_lookup_tsn() local 941 list_for_each_entry(chunk, &active->transmitted, in sctp_assoc_lookup_tsn() 944 if (key == chunk->subh.data_hdr->tsn) { in sctp_assoc_lookup_tsn() 956 list_for_each_entry(chunk, &transport->transmitted, in sctp_assoc_lookup_tsn() 958 if (key == chunk->subh.data_hdr->tsn) { in sctp_assoc_lookup_tsn() 1001 struct sctp_chunk *chunk; in sctp_assoc_bh_rcv() local 1012 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_assoc_bh_rcv() 1014 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_assoc_bh_rcv() 1023 if (sctp_auth_recv_cid(subtype.chunk, asoc) && !chunk->auth) in sctp_assoc_bh_rcv() 1029 if (sctp_chunk_is_data(chunk)) in sctp_assoc_bh_rcv() [all …]
|
D | objcnt.c | 52 SCTP_DBG_OBJCNT(chunk); 66 SCTP_DBG_OBJCNT_ENTRY(chunk),
|
D | ulpqueue.c | 95 int sctp_ulpq_tail_data(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, in sctp_ulpq_tail_data() argument 103 event = sctp_ulpevent_make_rcvmsg(chunk->asoc, chunk, gfp); in sctp_ulpq_tail_data() 1080 void sctp_ulpq_renege(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, in sctp_ulpq_renege() argument 1088 if (chunk) { in sctp_ulpq_renege() 1089 needed = ntohs(chunk->chunk_hdr->length); in sctp_ulpq_renege() 1103 if (chunk && (freed >= needed)) { in sctp_ulpq_renege() 1105 retval = sctp_ulpq_tail_data(ulpq, chunk, gfp); in sctp_ulpq_renege()
|
D | socket.c | 99 struct sctp_chunk *chunk); 148 static inline void sctp_set_owner_w(struct sctp_chunk *chunk) in sctp_set_owner_w() argument 150 struct sctp_association *asoc = chunk->asoc; in sctp_set_owner_w() 156 skb_set_owner_w(chunk->skb, sk); in sctp_set_owner_w() 158 chunk->skb->destructor = sctp_wfree; in sctp_set_owner_w() 160 skb_shinfo(chunk->skb)->destructor_arg = chunk; in sctp_set_owner_w() 162 asoc->sndbuf_used += SCTP_DATA_SNDSIZE(chunk) + in sctp_set_owner_w() 167 sk->sk_wmem_queued += chunk->skb->truesize; in sctp_set_owner_w() 168 sk_mem_charge(sk, chunk->skb->truesize); in sctp_set_owner_w() 411 struct sctp_chunk *chunk) in sctp_send_asconf() argument [all …]
|
D | probe.c | 137 struct sctp_chunk *chunk = arg; in jsctp_sf_eat_sack() local 138 struct sk_buff *skb = chunk->skb; in jsctp_sf_eat_sack()
|
D | Makefile | 10 transport.o chunk.o sm_make_chunk.o ulpevent.o \
|
D | sm_statetable.c | 86 return sctp_chunk_event_lookup(net, event_subtype.chunk, state); in sctp_sm_lookup_event()
|
/linux-4.1.27/mm/ |
D | percpu-vm.c | 13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk, in pcpu_chunk_page() argument 17 WARN_ON(chunk->immutable); in pcpu_chunk_page() 19 return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx)); in pcpu_chunk_page() 55 static void pcpu_free_pages(struct pcpu_chunk *chunk, in pcpu_free_pages() argument 82 static int pcpu_alloc_pages(struct pcpu_chunk *chunk, in pcpu_alloc_pages() argument 125 static void pcpu_pre_unmap_flush(struct pcpu_chunk *chunk, in pcpu_pre_unmap_flush() argument 129 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start), in pcpu_pre_unmap_flush() 130 pcpu_chunk_addr(chunk, pcpu_high_unit_cpu, page_end)); in pcpu_pre_unmap_flush() 151 static void pcpu_unmap_pages(struct pcpu_chunk *chunk, in pcpu_unmap_pages() argument 161 page = pcpu_chunk_page(chunk, cpu, i); in pcpu_unmap_pages() [all …]
|
D | percpu.c | 218 static int pcpu_chunk_slot(const struct pcpu_chunk *chunk) in pcpu_chunk_slot() argument 220 if (chunk->free_size < sizeof(int) || chunk->contig_hint < sizeof(int)) in pcpu_chunk_slot() 223 return pcpu_size_to_slot(chunk->free_size); in pcpu_chunk_slot() 243 static unsigned long pcpu_chunk_addr(struct pcpu_chunk *chunk, in pcpu_chunk_addr() argument 246 return (unsigned long)chunk->base_addr + pcpu_unit_offsets[cpu] + in pcpu_chunk_addr() 250 static void __maybe_unused pcpu_next_unpop(struct pcpu_chunk *chunk, in pcpu_next_unpop() argument 253 *rs = find_next_zero_bit(chunk->populated, end, *rs); in pcpu_next_unpop() 254 *re = find_next_bit(chunk->populated, end, *rs + 1); in pcpu_next_unpop() 257 static void __maybe_unused pcpu_next_pop(struct pcpu_chunk *chunk, in pcpu_next_pop() argument 260 *rs = find_next_bit(chunk->populated, end, *rs); in pcpu_next_pop() [all …]
|
D | percpu-km.c | 36 static int pcpu_populate_chunk(struct pcpu_chunk *chunk, in pcpu_populate_chunk() argument 42 static void pcpu_depopulate_chunk(struct pcpu_chunk *chunk, in pcpu_depopulate_chunk() argument 51 struct pcpu_chunk *chunk; in pcpu_create_chunk() local 55 chunk = pcpu_alloc_chunk(); in pcpu_create_chunk() 56 if (!chunk) in pcpu_create_chunk() 61 pcpu_free_chunk(chunk); in pcpu_create_chunk() 66 pcpu_set_page_chunk(nth_page(pages, i), chunk); in pcpu_create_chunk() 68 chunk->data = pages; in pcpu_create_chunk() 69 chunk->base_addr = page_address(pages) - pcpu_group_offsets[0]; in pcpu_create_chunk() 72 pcpu_chunk_populated(chunk, 0, nr_pages); in pcpu_create_chunk() [all …]
|
/linux-4.1.27/drivers/s390/cio/ |
D | itcw.c | 181 void *chunk; in itcw_init() local 193 chunk = fit_chunk(&start, end, sizeof(struct itcw), 1, 0); in itcw_init() 194 if (IS_ERR(chunk)) in itcw_init() 195 return chunk; in itcw_init() 196 itcw = chunk; in itcw_init() 209 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init() 210 if (IS_ERR(chunk)) in itcw_init() 211 return chunk; in itcw_init() 212 itcw->tcw = chunk; in itcw_init() 217 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init() [all …]
|
/linux-4.1.27/lib/ |
D | genalloc.c | 39 static inline size_t chunk_size(const struct gen_pool_chunk *chunk) in chunk_size() argument 41 return chunk->end_addr - chunk->start_addr + 1; in chunk_size() 184 struct gen_pool_chunk *chunk; in gen_pool_add_virt() local 189 chunk = kzalloc_node(nbytes, GFP_KERNEL, nid); in gen_pool_add_virt() 190 if (unlikely(chunk == NULL)) in gen_pool_add_virt() 193 chunk->phys_addr = phys; in gen_pool_add_virt() 194 chunk->start_addr = virt; in gen_pool_add_virt() 195 chunk->end_addr = virt + size - 1; in gen_pool_add_virt() 196 atomic_set(&chunk->avail, size); in gen_pool_add_virt() 199 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_virt() [all …]
|
D | bitmap.c | 363 u32 chunk; in __bitmap_parse() local 370 chunk = ndigits = 0; in __bitmap_parse() 405 if (chunk & ~((1UL << (CHUNKSZ - 4)) - 1)) in __bitmap_parse() 408 chunk = (chunk << 4) | hex_to_bin(c); in __bitmap_parse() 413 if (nchunks == 0 && chunk == 0) in __bitmap_parse() 417 *maskp |= chunk; in __bitmap_parse() 419 nbits += (nchunks == 1) ? nbits_to_hold_value(chunk) : CHUNKSZ; in __bitmap_parse()
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
D | icm.c | 54 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_pages() argument 58 if (chunk->nsg > 0) in mlx4_free_icm_pages() 59 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, in mlx4_free_icm_pages() 62 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages() 63 __free_pages(sg_page(&chunk->mem[i]), in mlx4_free_icm_pages() 64 get_order(chunk->mem[i].length)); in mlx4_free_icm_pages() 67 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_coherent() argument 71 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent() 73 chunk->mem[i].length, in mlx4_free_icm_coherent() 74 lowmem_page_address(sg_page(&chunk->mem[i])), in mlx4_free_icm_coherent() [all …]
|
D | icm.h | 64 struct mlx4_icm_chunk *chunk; member 91 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first() 99 return !iter->chunk; in mlx4_icm_last() 104 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next() 105 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next() 106 iter->chunk = NULL; in mlx4_icm_next() 110 iter->chunk = list_entry(iter->chunk->list.next, in mlx4_icm_next() 118 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_addr() 123 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_size()
|
D | mr.c | 724 int chunk; in __mlx4_write_mtt() local 733 chunk = min_t(int, max_mtts_first_page, npages); in __mlx4_write_mtt() 736 err = mlx4_write_mtt_chunk(dev, mtt, start_index, chunk, page_list); in __mlx4_write_mtt() 739 npages -= chunk; in __mlx4_write_mtt() 740 start_index += chunk; in __mlx4_write_mtt() 741 page_list += chunk; in __mlx4_write_mtt() 743 chunk = min_t(int, mtts_per_page, npages); in __mlx4_write_mtt() 753 int chunk; in mlx4_write_mtt() local 767 chunk = min_t(int, MLX4_MAILBOX_SIZE / sizeof(u64) - 2, in mlx4_write_mtt() 771 for (i = 0; i < chunk; ++i) in mlx4_write_mtt() [all …]
|
/linux-4.1.27/kernel/ |
D | audit_tree.c | 109 static void free_chunk(struct audit_chunk *chunk) in free_chunk() argument 113 for (i = 0; i < chunk->count; i++) { in free_chunk() 114 if (chunk->owners[i].owner) in free_chunk() 115 put_tree(chunk->owners[i].owner); in free_chunk() 117 kfree(chunk); in free_chunk() 120 void audit_put_chunk(struct audit_chunk *chunk) in audit_put_chunk() argument 122 if (atomic_long_dec_and_test(&chunk->refs)) in audit_put_chunk() 123 free_chunk(chunk); in audit_put_chunk() 128 struct audit_chunk *chunk = container_of(rcu, struct audit_chunk, head); in __put_chunk() local 129 audit_put_chunk(chunk); in __put_chunk() [all …]
|
D | auditsc.c | 210 static int put_tree_ref(struct audit_context *ctx, struct audit_chunk *chunk) in put_tree_ref() argument 215 p->c[--left] = chunk; in put_tree_ref() 223 p->c[30] = chunk; in put_tree_ref() 1588 struct audit_chunk *chunk; in handle_one() local 1596 chunk = audit_tree_lookup(inode); in handle_one() 1598 if (!chunk) in handle_one() 1600 if (likely(put_tree_ref(context, chunk))) in handle_one() 1605 audit_put_chunk(chunk); in handle_one() 1609 put_tree_ref(context, chunk); in handle_one() 1634 struct audit_chunk *chunk; in handle_path() local [all …]
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
D | mthca_memfree.c | 64 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_pages() argument 68 if (chunk->nsg > 0) in mthca_free_icm_pages() 69 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, in mthca_free_icm_pages() 72 for (i = 0; i < chunk->npages; ++i) in mthca_free_icm_pages() 73 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages() 74 get_order(chunk->mem[i].length)); in mthca_free_icm_pages() 77 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_coherent() argument 81 for (i = 0; i < chunk->npages; ++i) { in mthca_free_icm_coherent() 82 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, in mthca_free_icm_coherent() 83 lowmem_page_address(sg_page(&chunk->mem[i])), in mthca_free_icm_coherent() [all …]
|
D | mthca_memfree.h | 76 struct mthca_icm_chunk *chunk; member 103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first() 111 return !iter->chunk; in mthca_icm_last() 116 if (++iter->page_idx >= iter->chunk->nsg) { in mthca_icm_next() 117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next() 118 iter->chunk = NULL; in mthca_icm_next() 122 iter->chunk = list_entry(iter->chunk->list.next, in mthca_icm_next() 130 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mthca_icm_addr() 135 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mthca_icm_size()
|
D | mthca_mr.c | 362 int chunk; in mthca_write_mtt() local 369 chunk = min(size, list_len); in mthca_write_mtt() 372 buffer_list, chunk); in mthca_write_mtt() 375 buffer_list, chunk); in mthca_write_mtt() 377 list_len -= chunk; in mthca_write_mtt() 378 start_index += chunk; in mthca_write_mtt() 379 buffer_list += chunk; in mthca_write_mtt()
|
/linux-4.1.27/drivers/gpu/drm/qxl/ |
D | qxl_image.c | 38 struct qxl_drm_chunk *chunk; in qxl_allocate_chunk() local 41 chunk = kmalloc(sizeof(struct qxl_drm_chunk), GFP_KERNEL); in qxl_allocate_chunk() 42 if (!chunk) in qxl_allocate_chunk() 45 ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo); in qxl_allocate_chunk() 47 kfree(chunk); in qxl_allocate_chunk() 51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk() 88 struct qxl_drm_chunk *chunk, *tmp; in qxl_image_free_objects() local 90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects() 91 qxl_bo_unref(&chunk->bo); in qxl_image_free_objects() 92 kfree(chunk); in qxl_image_free_objects() [all …]
|
D | qxl_dev.h | 421 struct qxl_data_chunk chunk; member 482 struct qxl_data_chunk chunk; member 761 struct qxl_data_chunk chunk; member 779 struct qxl_data_chunk chunk; member
|
D | qxl_draw.c | 52 dev_clips->chunk.next_chunk = 0; in drawable_set_clipping() 53 dev_clips->chunk.prev_chunk = 0; in drawable_set_clipping() 54 dev_clips->chunk.data_size = sizeof(struct qxl_rect) * num_clips; in drawable_set_clipping() 55 return (struct qxl_rect *)dev_clips->chunk.data; in drawable_set_clipping()
|
D | qxl_display.c | 360 cursor->chunk.next_chunk = 0; in qxl_crtc_cursor_set2() 361 cursor->chunk.prev_chunk = 0; in qxl_crtc_cursor_set2() 362 cursor->chunk.data_size = size; in qxl_crtc_cursor_set2() 364 memcpy(cursor->chunk.data, user_ptr, size); in qxl_crtc_cursor_set2()
|
/linux-4.1.27/arch/mips/dec/prom/ |
D | console.c | 22 unsigned int chunk = sizeof(buf) - 1; in prom_console_write() local 25 if (chunk > c) in prom_console_write() 26 chunk = c; in prom_console_write() 27 memcpy(buf, s, chunk); in prom_console_write() 28 buf[chunk] = '\0'; in prom_console_write() 30 s += chunk; in prom_console_write() 31 c -= chunk; in prom_console_write()
|
/linux-4.1.27/sound/core/ |
D | sgbuf.c | 71 unsigned int i, pages, chunk, maxpages; in snd_malloc_sgbuf_pages() local 96 chunk = pages; in snd_malloc_sgbuf_pages() 98 if (chunk > maxpages) in snd_malloc_sgbuf_pages() 99 chunk = maxpages; in snd_malloc_sgbuf_pages() 100 chunk <<= PAGE_SHIFT; in snd_malloc_sgbuf_pages() 102 chunk, &tmpb) < 0) { in snd_malloc_sgbuf_pages() 110 chunk = tmpb.bytes >> PAGE_SHIFT; in snd_malloc_sgbuf_pages() 111 for (i = 0; i < chunk; i++) { in snd_malloc_sgbuf_pages() 115 table->addr |= chunk; /* mark head */ in snd_malloc_sgbuf_pages() 121 sgbuf->pages += chunk; in snd_malloc_sgbuf_pages() [all …]
|
/linux-4.1.27/drivers/media/usb/usbtv/ |
D | usbtv.h | 48 #define USBTV_MAGIC_OK(chunk) ((be32_to_cpu(chunk[0]) & 0xff000000) \ argument 50 #define USBTV_FRAME_ID(chunk) ((be32_to_cpu(chunk[0]) & 0x00ff0000) >> 16) argument 51 #define USBTV_ODD(chunk) ((be32_to_cpu(chunk[0]) & 0x0000f000) >> 15) argument 52 #define USBTV_CHUNK_NO(chunk) (be32_to_cpu(chunk[0]) & 0x00000fff) argument
|
D | usbtv-video.c | 275 static void usbtv_image_chunk(struct usbtv *usbtv, __be32 *chunk) in usbtv_image_chunk() argument 283 if (!USBTV_MAGIC_OK(chunk)) in usbtv_image_chunk() 285 frame_id = USBTV_FRAME_ID(chunk); in usbtv_image_chunk() 286 odd = USBTV_ODD(chunk); in usbtv_image_chunk() 287 chunk_no = USBTV_CHUNK_NO(chunk); in usbtv_image_chunk() 312 usbtv_chunk_to_vbuf(frame, &chunk[1], chunk_no, odd); in usbtv_image_chunk()
|
/linux-4.1.27/drivers/infiniband/hw/usnic/ |
D | usnic_vnic.c | 81 struct usnic_vnic_res_chunk *chunk; in usnic_vnic_dump() local 104 chunk = &vnic->chunks[i]; in usnic_vnic_dump() 105 for (j = 0; j < chunk->cnt; j++) { in usnic_vnic_dump() 106 res = chunk->res[j]; in usnic_vnic_dump() 262 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk) in usnic_vnic_put_resources() argument 267 struct usnic_vnic *vnic = chunk->vnic; in usnic_vnic_put_resources() 270 while ((i = --chunk->cnt) >= 0) { in usnic_vnic_put_resources() 271 res = chunk->res[i]; in usnic_vnic_put_resources() 272 chunk->res[i] = NULL; in usnic_vnic_put_resources() 278 kfree(chunk->res); in usnic_vnic_put_resources() [all …]
|
D | usnic_uiom.c | 81 struct usnic_uiom_chunk *chunk, *tmp; in usnic_uiom_put_pages() local 87 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages() 88 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_put_pages() 96 kfree(chunk); in usnic_uiom_put_pages() 105 struct usnic_uiom_chunk *chunk; in usnic_uiom_get_pages() local 159 chunk = kmalloc(sizeof(*chunk) + in usnic_uiom_get_pages() 163 if (!chunk) { in usnic_uiom_get_pages() 168 chunk->nents = min_t(int, ret, USNIC_UIOM_PAGE_CHUNK); in usnic_uiom_get_pages() 169 sg_init_table(chunk->page_list, chunk->nents); in usnic_uiom_get_pages() 170 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_get_pages() [all …]
|
D | usnic_ib_verbs.c | 49 struct usnic_vnic_res_chunk *chunk; in usnic_ib_fill_create_qp_resp() local 74 chunk = usnic_ib_qp_grp_get_chunk(qp_grp, USNIC_VNIC_RES_TYPE_RQ); in usnic_ib_fill_create_qp_resp() 75 if (IS_ERR_OR_NULL(chunk)) { in usnic_ib_fill_create_qp_resp() 79 PTR_ERR(chunk)); in usnic_ib_fill_create_qp_resp() 80 return chunk ? PTR_ERR(chunk) : -ENOMEM; in usnic_ib_fill_create_qp_resp() 83 WARN_ON(chunk->type != USNIC_VNIC_RES_TYPE_RQ); in usnic_ib_fill_create_qp_resp() 84 resp.rq_cnt = chunk->cnt; in usnic_ib_fill_create_qp_resp() 85 for (i = 0; i < chunk->cnt; i++) in usnic_ib_fill_create_qp_resp() 86 resp.rq_idx[i] = chunk->res[i]->vnic_idx; in usnic_ib_fill_create_qp_resp() 88 chunk = usnic_ib_qp_grp_get_chunk(qp_grp, USNIC_VNIC_RES_TYPE_WQ); in usnic_ib_fill_create_qp_resp() [all …]
|
D | usnic_vnic.h | 95 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk);
|
/linux-4.1.27/drivers/s390/block/ |
D | dasd_int.h | 573 struct dasd_mchunk *chunk; in dasd_init_chunklist() local 576 chunk = (struct dasd_mchunk *) mem; in dasd_init_chunklist() 577 chunk->size = size - sizeof(struct dasd_mchunk); in dasd_init_chunklist() 578 list_add(&chunk->list, chunk_list); in dasd_init_chunklist() 584 struct dasd_mchunk *chunk, *tmp; in dasd_alloc_chunk() local 587 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk() 588 if (chunk->size < size) in dasd_alloc_chunk() 590 if (chunk->size > size + sizeof(struct dasd_mchunk)) { in dasd_alloc_chunk() 591 char *endaddr = (char *) (chunk + 1) + chunk->size; in dasd_alloc_chunk() 594 chunk->size -= size + sizeof(struct dasd_mchunk); in dasd_alloc_chunk() [all …]
|
/linux-4.1.27/fs/ocfs2/ |
D | quota_local.c | 297 int chunk, in ocfs2_add_recovery_chunk() argument 305 rc->rc_chunk = chunk; in ocfs2_add_recovery_chunk() 471 int bit, chunk; in ocfs2_recover_local_quota_file() local 478 chunk = rchunk->rc_chunk; in ocfs2_recover_local_quota_file() 481 ol_quota_chunk_block(sb, chunk), in ocfs2_recover_local_quota_file() 491 ol_dqblk_block(sb, chunk, bit), in ocfs2_recover_local_quota_file() 498 ol_dqblk_block_off(sb, chunk, bit)); in ocfs2_recover_local_quota_file() 817 struct ocfs2_quota_chunk *chunk; in ocfs2_local_free_info() local 825 list_for_each_entry(chunk, &oinfo->dqi_chunk, qc_chunk) { in ocfs2_local_free_info() 827 (chunk->qc_headerbh->b_data); in ocfs2_local_free_info() [all …]
|
D | ioctl.c | 448 unsigned int offset = 0, cluster, chunk; in ocfs2_info_freefrag_scan_chain() local 487 for (chunk = 0; chunk < chunks_in_group; chunk++) { in ocfs2_info_freefrag_scan_chain()
|
/linux-4.1.27/drivers/media/usb/go7007/ |
D | go7007-fw.c | 378 int size = 0, i, off = 0, chunk; in gen_mjpeghdr_to_package() local 388 chunk = mjpeg_frame_header(go, buf + size, 1); in gen_mjpeghdr_to_package() 389 memmove(buf + size, buf + size + 80, chunk - 80); in gen_mjpeghdr_to_package() 390 size += chunk - 80; in gen_mjpeghdr_to_package() 392 for (i = 0; i < size; i += chunk * 2) { in gen_mjpeghdr_to_package() 400 chunk = 28; in gen_mjpeghdr_to_package() 401 if (mem + chunk > 0x4000) in gen_mjpeghdr_to_package() 402 chunk = 0x4000 - mem; in gen_mjpeghdr_to_package() 403 if (i + 2 * chunk > size) in gen_mjpeghdr_to_package() 404 chunk = (size - i) / 2; in gen_mjpeghdr_to_package() [all …]
|
/linux-4.1.27/include/net/sctp/ |
D | sm.h | 174 __u8 sctp_get_chunk_type(struct sctp_chunk *chunk); 209 const struct sctp_chunk *chunk); 239 const struct sctp_chunk *chunk, 252 struct sctp_chunk *chunk, bool addr_param_needed, 299 static inline __u16 sctp_data_size(struct sctp_chunk *chunk) in sctp_data_size() argument 303 size = ntohs(chunk->chunk_hdr->length); in sctp_data_size() 392 sctp_vtag_verify(const struct sctp_chunk *chunk, in sctp_vtag_verify() argument 401 if (ntohl(chunk->sctp_hdr->vtag) == asoc->c.my_vtag) in sctp_vtag_verify() 411 sctp_vtag_verify_either(const struct sctp_chunk *chunk, in sctp_vtag_verify_either() argument 435 if ((!sctp_test_T_bit(chunk) && in sctp_vtag_verify_either() [all …]
|
D | sctp.h | 432 static inline bool sctp_chunk_pending(const struct sctp_chunk *chunk) in sctp_chunk_pending() argument 434 return !list_empty(&chunk->list); in sctp_chunk_pending() 442 #define sctp_walk_params(pos, chunk, member)\ argument 443 _sctp_walk_params((pos), (chunk), ntohs((chunk)->chunk_hdr.length), member) 445 #define _sctp_walk_params(pos, chunk, end, member)\ argument 446 for (pos.v = chunk->member;\ 447 pos.v <= (void *)chunk + end - ntohs(pos.p->length) &&\ 461 #define sctp_walk_fwdtsn(pos, chunk)\ argument 462 _sctp_walk_fwdtsn((pos), (chunk), ntohs((chunk)->chunk_hdr->length) - sizeof(struct sctp_fwdtsn_chu… 464 #define _sctp_walk_fwdtsn(pos, chunk, end)\ argument [all …]
|
D | ulpevent.h | 85 struct sctp_chunk *chunk, 98 struct sctp_chunk *chunk, 103 struct sctp_chunk *chunk, 121 struct sctp_chunk *chunk,
|
D | auth.h | 100 int sctp_auth_send_cid(sctp_cid_t chunk, const struct sctp_association *asoc); 101 int sctp_auth_recv_cid(sctp_cid_t chunk, const struct sctp_association *asoc);
|
D | constants.h | 128 sctp_cid_t chunk; member 139 SCTP_SUBTYPE_CONSTRUCTOR(CHUNK, sctp_cid_t, chunk)
|
D | command.h | 131 struct sctp_chunk *chunk; member 172 SCTP_ARG_CONSTRUCTOR(CHUNK, struct sctp_chunk *, chunk) in SCTP_ARG_CONSTRUCTOR()
|
D | structs.h | 654 int sctp_user_addto_chunk(struct sctp_chunk *chunk, int len, 663 const union sctp_addr *sctp_source(const struct sctp_chunk *chunk); 1058 int sctp_outq_tail(struct sctp_outq *, struct sctp_chunk *chunk); 1281 struct sctp_chunk *chunk, struct sctp_chunk **err_chunk); 1282 int sctp_process_init(struct sctp_association *, struct sctp_chunk *chunk,
|
/linux-4.1.27/sound/soc/codecs/ |
D | sigmadsp.c | 54 struct sigma_fw_chunk chunk; member 60 struct sigma_fw_chunk chunk; member 68 struct sigma_fw_chunk chunk; member 200 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_control() argument 212 ctrl_chunk = (const struct sigma_fw_chunk_control *)chunk; in sigma_fw_load_control() 238 ctrl->samplerates = le32_to_cpu(chunk->samplerates); in sigma_fw_load_control() 251 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_data() argument 259 data_chunk = (struct sigma_fw_chunk_data *)chunk; in sigma_fw_load_data() 269 data->samplerates = le32_to_cpu(chunk->samplerates); in sigma_fw_load_data() 277 const struct sigma_fw_chunk *chunk, unsigned int length) in sigma_fw_load_samplerates() argument [all …]
|
/linux-4.1.27/arch/um/drivers/ |
D | mconsole_kern.h | 27 #define CONFIG_CHUNK(str, size, current, chunk, end) \ argument 29 current += strlen(chunk); \ 33 strcpy(str, chunk); \ 34 str += strlen(chunk); \
|
/linux-4.1.27/drivers/dma/sh/ |
D | rcar-dmac.c | 328 struct rcar_dmac_xfer_chunk *chunk; in rcar_dmac_chan_start_xfer() local 352 chunk = list_first_entry(&desc->chunks, in rcar_dmac_chan_start_xfer() 355 chunk->dst_addr & 0xffffffff); in rcar_dmac_chan_start_xfer() 385 struct rcar_dmac_xfer_chunk *chunk = desc->running; in rcar_dmac_chan_start_xfer() local 389 chan->index, chunk, chunk->size, &chunk->src_addr, in rcar_dmac_chan_start_xfer() 390 &chunk->dst_addr); in rcar_dmac_chan_start_xfer() 394 chunk->src_addr >> 32); in rcar_dmac_chan_start_xfer() 396 chunk->dst_addr >> 32); in rcar_dmac_chan_start_xfer() 399 chunk->src_addr & 0xffffffff); in rcar_dmac_chan_start_xfer() 401 chunk->dst_addr & 0xffffffff); in rcar_dmac_chan_start_xfer() [all …]
|
D | shdma-base.c | 75 struct shdma_desc *chunk, *c, *desc = in shdma_tx_submit() local 89 list_for_each_entry_safe(chunk, c, desc->node.prev, node) { in shdma_tx_submit() 94 if (chunk != desc && (chunk->mark == DESC_IDLE || in shdma_tx_submit() 95 chunk->async_tx.cookie > 0 || in shdma_tx_submit() 96 chunk->async_tx.cookie == -EBUSY || in shdma_tx_submit() 97 &chunk->node == &schan->ld_free)) in shdma_tx_submit() 99 chunk->mark = DESC_SUBMITTED; in shdma_tx_submit() 100 if (chunk->chunks == 1) { in shdma_tx_submit() 101 chunk->async_tx.callback = callback; in shdma_tx_submit() 102 chunk->async_tx.callback_param = tx->callback_param; in shdma_tx_submit() [all …]
|
/linux-4.1.27/drivers/md/ |
D | dm-snap.c | 157 chunk_t chunk) in chunk_to_sector() argument 159 return chunk << store->chunk_shift; in chunk_to_sector() 215 chunk_t chunk; member 230 static void track_chunk(struct dm_snapshot *s, struct bio *bio, chunk_t chunk) in track_chunk() argument 234 c->chunk = chunk; in track_chunk() 238 &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)]); in track_chunk() 252 static int __chunk_is_tracked(struct dm_snapshot *s, chunk_t chunk) in __chunk_is_tracked() argument 260 &s->tracked_chunk_hash[DM_TRACKED_CHUNK_HASH(chunk)], node) { in __chunk_is_tracked() 261 if (c->chunk == chunk) { in __chunk_is_tracked() 276 static void __check_for_conflicting_io(struct dm_snapshot *s, chunk_t chunk) in __check_for_conflicting_io() argument [all …]
|
D | dm-stripe.c | 214 sector_t chunk = dm_target_offset(sc->ti, sector); in stripe_map_sector() local 218 chunk_offset = sector_div(chunk, sc->chunk_size); in stripe_map_sector() 220 chunk_offset = chunk & (sc->chunk_size - 1); in stripe_map_sector() 221 chunk >>= sc->chunk_size_shift; in stripe_map_sector() 225 *stripe = sector_div(chunk, sc->stripes); in stripe_map_sector() 227 *stripe = chunk & (sc->stripes - 1); in stripe_map_sector() 228 chunk >>= sc->stripes_shift; in stripe_map_sector() 232 chunk *= sc->chunk_size; in stripe_map_sector() 234 chunk <<= sc->chunk_size_shift; in stripe_map_sector() 236 *result = chunk + chunk_offset; in stripe_map_sector()
|
D | dm-exception-store.h | 141 static inline chunk_t dm_chunk_number(chunk_t chunk) in dm_chunk_number() argument 143 return chunk & (chunk_t)((1ULL << DM_CHUNK_NUMBER_BITS) - 1ULL); in dm_chunk_number() 168 static inline chunk_t dm_chunk_number(chunk_t chunk) in dm_chunk_number() argument 170 return chunk; in dm_chunk_number()
|
D | bitmap.c | 725 unsigned long chunk) in file_page_index() argument 728 chunk += sizeof(bitmap_super_t) << 3; in file_page_index() 729 return chunk >> PAGE_BIT_SHIFT; in file_page_index() 734 unsigned long chunk) in file_page_offset() argument 737 chunk += sizeof(bitmap_super_t) << 3; in file_page_offset() 738 return chunk & (PAGE_BITS - 1); in file_page_offset() 746 unsigned long chunk) in filemap_get_page() argument 748 if (file_page_index(store, chunk) >= store->file_pages) in filemap_get_page() 750 return store->filemap[file_page_index(store, chunk)]; in filemap_get_page() 909 unsigned long chunk = block >> bitmap->counts.chunkshift; in bitmap_file_set_bit() local [all …]
|
D | dm-snap-persistent.c | 228 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, int rw, in chunk_io() argument 233 .sector = ps->store->chunk_size * chunk, in chunk_io() 286 chunk_t chunk; in area_io() local 288 chunk = area_location(ps, ps->current_area); in area_io() 290 r = chunk_io(ps, ps->area, chunk, rw, 0); in area_io() 518 chunk_t chunk; in read_exceptions() local 533 chunk = area_location(ps, ps->current_area); in read_exceptions() 535 area = dm_bufio_read(client, chunk, &bp); in read_exceptions() 549 dm_bufio_forget(client, chunk); in read_exceptions()
|
D | raid0.c | 311 sector_t chunk; in map_sector() local 322 chunk = *sector_offset; in map_sector() 324 sector_div(chunk, zone->nb_dev << chunksect_bits); in map_sector() 327 chunk = *sector_offset; in map_sector() 328 sector_div(chunk, chunk_sects * zone->nb_dev); in map_sector() 335 *sector_offset = (chunk * chunk_sects) + sect_in_chunk; in map_sector()
|
D | raid10.c | 556 sector_t chunk; in __raid10_find_phys() local 569 chunk = r10bio->sector >> geo->chunk_shift; in __raid10_find_phys() 572 chunk *= geo->near_copies; in __raid10_find_phys() 573 stripe = chunk; in __raid10_find_phys() 632 sector_t offset, chunk, vchunk; in raid10_find_virt() local 655 chunk = sector >> geo->chunk_shift; in raid10_find_virt() 656 fc = sector_div(chunk, geo->far_copies); in raid10_find_virt() 668 chunk = sector >> geo->chunk_shift; in raid10_find_virt() 670 vchunk = chunk * geo->raid_disks + dev; in raid10_find_virt() 3470 int layout, chunk, disks; in setup_geo() local [all …]
|
D | Kconfig | 68 up evenly, one chunk here and one chunk there. This will increase
|
D | md.c | 4029 unsigned long chunk, end_chunk; in bitmap_store() local 4039 chunk = end_chunk = simple_strtoul(buf, &end, 0); in bitmap_store() 4047 bitmap_dirty_bits(mddev->bitmap, chunk, end_chunk); in bitmap_store() 4480 int chunk; in max_sync_store() local 4495 chunk = mddev->chunk_sectors; in max_sync_store() 4496 if (chunk) { in max_sync_store() 4500 if (sector_div(temp, chunk)) in max_sync_store()
|
/linux-4.1.27/Documentation/device-mapper/ |
D | striped.txt | 10 Parameters: <num devs> <chunk size> [<dev path> <offset>]+ 12 <chunk size>: Size of each chunk of data. Must be at least as 19 be a multiple of the chunk size multiplied by the number of underlying devices. 28 # will be called "stripe_dev" and have a chunk-size of 128k.
|
D | cache.txt | 141 Policy plug-ins can store a chunk of data per cache block. It's up to 142 the policy how big this chunk is, but it should be kept small. Like the
|
D | snapshot.txt | 61 deferred until the merging snapshot's corresponding chunk(s) have been
|
/linux-4.1.27/arch/arm64/kernel/ |
D | sys_compat.c | 37 unsigned long chunk = min(PAGE_SIZE, end - start); in __do_compat_cache_op() local 42 ret = __flush_cache_user_range(start, start + chunk); in __do_compat_cache_op() 47 start += chunk; in __do_compat_cache_op()
|
/linux-4.1.27/net/bluetooth/ |
D | af_bluetooth.c | 306 int chunk; in bt_sock_stream_recvmsg() local 332 chunk = min_t(unsigned int, skb->len, size); in bt_sock_stream_recvmsg() 333 if (skb_copy_datagram_msg(skb, 0, msg, chunk)) { in bt_sock_stream_recvmsg() 339 copied += chunk; in bt_sock_stream_recvmsg() 340 size -= chunk; in bt_sock_stream_recvmsg() 347 if (chunk <= skb_len) { in bt_sock_stream_recvmsg() 348 __skb_pull(skb, chunk); in bt_sock_stream_recvmsg() 353 chunk -= skb_len; in bt_sock_stream_recvmsg() 356 if (chunk <= frag->len) { in bt_sock_stream_recvmsg() 358 skb->len -= chunk; in bt_sock_stream_recvmsg() [all …]
|
/linux-4.1.27/drivers/atm/ |
D | fore200e.c | 178 fore200e_chunk_alloc(struct fore200e* fore200e, struct chunk* chunk, int size, int alignment, int d… in fore200e_chunk_alloc() argument 185 chunk->alloc_size = size + alignment; in fore200e_chunk_alloc() 186 chunk->align_size = size; in fore200e_chunk_alloc() 187 chunk->direction = direction; in fore200e_chunk_alloc() 189 chunk->alloc_addr = kzalloc(chunk->alloc_size, GFP_KERNEL | GFP_DMA); in fore200e_chunk_alloc() 190 if (chunk->alloc_addr == NULL) in fore200e_chunk_alloc() 194 offset = FORE200E_ALIGN(chunk->alloc_addr, alignment); in fore200e_chunk_alloc() 196 chunk->align_addr = chunk->alloc_addr + offset; in fore200e_chunk_alloc() 198 …chunk->dma_addr = fore200e->bus->dma_map(fore200e, chunk->align_addr, chunk->align_size, direction… in fore200e_chunk_alloc() 207 fore200e_chunk_free(struct fore200e* fore200e, struct chunk* chunk) in fore200e_chunk_free() argument [all …]
|
D | fore200e.h | 563 typedef struct chunk { struct 581 struct chunk data; /* data buffer */ 603 struct chunk status; /* array of completion status */ 613 … struct chunk tpd; /* array of tpds */ 614 … struct chunk status; /* arry of completion status */ 624 …struct chunk rpd; /* array of rpds */ 625 …struct chunk status; /* array of completion status */ 634 …struct chunk rbd_block; /* array of rbds … 635 …struct chunk status; /* array of completion status … 811 int (*dma_chunk_alloc)(struct fore200e*, struct chunk*, int, int, int); [all …]
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
D | resource.c | 419 unsigned start, chunk, top; in c4iw_ocqp_pool_create() local 426 chunk = rdev->lldi.vr->ocq.size; in c4iw_ocqp_pool_create() 427 top = start + chunk; in c4iw_ocqp_pool_create() 430 chunk = min(top - start + 1, chunk); in c4iw_ocqp_pool_create() 431 if (gen_pool_add(rdev->ocqp_pool, start, chunk, -1)) { in c4iw_ocqp_pool_create() 433 __func__, start, chunk); in c4iw_ocqp_pool_create() 434 if (chunk <= 1024 << MIN_OCQP_SHIFT) { in c4iw_ocqp_pool_create() 440 chunk >>= 1; in c4iw_ocqp_pool_create() 443 __func__, start, chunk); in c4iw_ocqp_pool_create() 444 start += chunk; in c4iw_ocqp_pool_create()
|
/linux-4.1.27/fs/reiserfs/ |
D | xattr.c | 527 size_t chunk; in reiserfs_xattr_set_handle() local 532 chunk = PAGE_CACHE_SIZE; in reiserfs_xattr_set_handle() 534 chunk = buffer_size - buffer_pos; in reiserfs_xattr_set_handle() 549 if (chunk + skip > PAGE_CACHE_SIZE) in reiserfs_xattr_set_handle() 550 chunk = PAGE_CACHE_SIZE - skip; in reiserfs_xattr_set_handle() 557 err = __reiserfs_write_begin(page, page_offset, chunk + skip); in reiserfs_xattr_set_handle() 560 memcpy(data + skip, buffer + buffer_pos, chunk); in reiserfs_xattr_set_handle() 562 page_offset + chunk + in reiserfs_xattr_set_handle() 568 buffer_pos += chunk; in reiserfs_xattr_set_handle() 569 file_pos += chunk; in reiserfs_xattr_set_handle() [all …]
|
D | journal.c | 676 static void write_chunk(struct buffer_chunk *chunk) in write_chunk() argument 679 for (i = 0; i < chunk->nr; i++) { in write_chunk() 680 submit_logged_buffer(chunk->bh[i]); in write_chunk() 682 chunk->nr = 0; in write_chunk() 685 static void write_ordered_chunk(struct buffer_chunk *chunk) in write_ordered_chunk() argument 688 for (i = 0; i < chunk->nr; i++) { in write_ordered_chunk() 689 submit_ordered_buffer(chunk->bh[i]); in write_ordered_chunk() 691 chunk->nr = 0; in write_ordered_chunk() 694 static int add_to_chunk(struct buffer_chunk *chunk, struct buffer_head *bh, in add_to_chunk() argument 698 BUG_ON(chunk->nr >= CHUNK_SIZE); in add_to_chunk() [all …]
|
/linux-4.1.27/fs/btrfs/ |
D | print-tree.c | 23 static void print_chunk(struct extent_buffer *eb, struct btrfs_chunk *chunk) in print_chunk() argument 25 int num_stripes = btrfs_chunk_num_stripes(eb, chunk); in print_chunk() 29 btrfs_chunk_length(eb, chunk), btrfs_chunk_owner(eb, chunk), in print_chunk() 30 btrfs_chunk_type(eb, chunk), num_stripes); in print_chunk() 33 btrfs_stripe_devid_nr(eb, chunk, i), in print_chunk() 34 btrfs_stripe_offset_nr(eb, chunk, i)); in print_chunk()
|
D | volumes.c | 2538 struct btrfs_chunk *chunk; in btrfs_del_sys_chunk() local 2560 chunk = (struct btrfs_chunk *)(ptr + len); in btrfs_del_sys_chunk() 2561 num_stripes = btrfs_stack_chunk_num_stripes(chunk); in btrfs_del_sys_chunk() 2712 struct btrfs_chunk *chunk; in btrfs_relocate_sys_chunks() local 2745 chunk = btrfs_item_ptr(leaf, path->slots[0], in btrfs_relocate_sys_chunks() 2747 chunk_type = btrfs_chunk_type(leaf, chunk); in btrfs_relocate_sys_chunks() 2979 struct btrfs_chunk *chunk, in chunk_devid_filter() argument 2983 int num_stripes = btrfs_chunk_num_stripes(leaf, chunk); in chunk_devid_filter() 2987 stripe = btrfs_stripe_nr(chunk, i); in chunk_devid_filter() 2997 struct btrfs_chunk *chunk, in chunk_drange_filter() argument [all …]
|
/linux-4.1.27/sound/usb/line6/ |
D | midi.c | 47 unsigned char chunk[LINE6_FALLBACK_MAXPACKETSIZE]; in line6_midi_transmit() local 52 done = snd_rawmidi_transmit_peek(substream, chunk, req); in line6_midi_transmit() 57 line6_midibuf_write(mb, chunk, done); in line6_midi_transmit() 62 done = line6_midibuf_read(mb, chunk, in line6_midi_transmit() 68 send_midi_async(line6, chunk, done); in line6_midi_transmit()
|
/linux-4.1.27/drivers/net/wireless/ti/wlcore/ |
D | boot.c | 179 u8 *p, *chunk; in wl1271_boot_upload_firmware_chunk() local 194 chunk = kmalloc(CHUNK_SIZE, GFP_KERNEL); in wl1271_boot_upload_firmware_chunk() 195 if (!chunk) { in wl1271_boot_upload_firmware_chunk() 226 memcpy(chunk, p, CHUNK_SIZE); in wl1271_boot_upload_firmware_chunk() 229 ret = wlcore_write(wl, addr, chunk, CHUNK_SIZE, false); in wl1271_boot_upload_firmware_chunk() 239 memcpy(chunk, p, fw_data_len % CHUNK_SIZE); in wl1271_boot_upload_firmware_chunk() 242 ret = wlcore_write(wl, addr, chunk, fw_data_len % CHUNK_SIZE, false); in wl1271_boot_upload_firmware_chunk() 245 kfree(chunk); in wl1271_boot_upload_firmware_chunk()
|
/linux-4.1.27/Documentation/mmc/ |
D | mmc-async-req.txt | 61 request in two chunks, prepare the first chunk and start the request, 62 and finally prepare the second chunk and start the transfer. 72 * The first chunk of the request should take the same time 75 * the transfer is delayed, guesstimate max 4k as first chunk size. 84 * of the first chunk. If the MMC runs out of the first data chunk
|
D | mmc-dev-attrs.txt | 59 hence "preferred_erase_size" provides a good chunk
|
/linux-4.1.27/drivers/crypto/ |
D | n2_core.c | 674 struct n2_crypto_chunk chunk; member 877 struct n2_crypto_chunk *chunk; in n2_compute_chunks() local 890 chunk = &rctx->chunk; in n2_compute_chunks() 891 INIT_LIST_HEAD(&chunk->entry); in n2_compute_chunks() 893 chunk->iv_paddr = 0UL; in n2_compute_chunks() 894 chunk->arr_len = 0; in n2_compute_chunks() 895 chunk->dest_paddr = 0UL; in n2_compute_chunks() 913 if (chunk->arr_len != 0) { in n2_compute_chunks() 917 chunk->arr_len == N2_CHUNK_ARR_LEN || in n2_compute_chunks() 919 chunk->dest_final = dest_prev; in n2_compute_chunks() [all …]
|
/linux-4.1.27/sound/pci/ |
D | es1968.c | 1374 struct esm_memory *chunk = kmalloc(sizeof(*chunk), GFP_KERNEL); in snd_es1968_new_memory() local 1375 if (chunk == NULL) { in snd_es1968_new_memory() 1379 chunk->buf = buf->buf; in snd_es1968_new_memory() 1380 chunk->buf.bytes -= size; in snd_es1968_new_memory() 1381 chunk->buf.area += size; in snd_es1968_new_memory() 1382 chunk->buf.addr += size; in snd_es1968_new_memory() 1383 chunk->empty = 1; in snd_es1968_new_memory() 1385 list_add(&chunk->list, &buf->list); in snd_es1968_new_memory() 1395 struct esm_memory *chunk; in snd_es1968_free_memory() local 1400 chunk = list_entry(buf->list.prev, struct esm_memory, list); in snd_es1968_free_memory() [all …]
|
/linux-4.1.27/arch/x86/xen/ |
D | p2m.c | 309 unsigned int i, chunk; in xen_rebuild_p2m_list() local 332 for (pfn = 0; pfn < xen_max_p2m_pfn; pfn += chunk) { in xen_rebuild_p2m_list() 343 chunk = (pfn & (P2M_PER_PAGE * P2M_MID_PER_PAGE - 1)) ? in xen_rebuild_p2m_list() 349 for (i = 1; i < chunk; i++) in xen_rebuild_p2m_list() 352 if (i < chunk) in xen_rebuild_p2m_list() 354 chunk = P2M_PER_PAGE; in xen_rebuild_p2m_list() 356 if (type == P2M_TYPE_PFN || i < chunk) { in xen_rebuild_p2m_list() 370 if (chunk == P2M_PER_PAGE) { in xen_rebuild_p2m_list()
|
D | setup.c | 310 unsigned int i, chunk; in xen_do_set_identity_and_remap_chunk() local 321 chunk = (left < REMAP_SIZE) ? left : REMAP_SIZE; in xen_do_set_identity_and_remap_chunk() 330 xen_remap_buf.size = chunk; in xen_do_set_identity_and_remap_chunk() 331 for (i = 0; i < chunk; i++) in xen_do_set_identity_and_remap_chunk() 338 set_phys_range_identity(ident_pfn_iter, ident_pfn_iter + chunk); in xen_do_set_identity_and_remap_chunk() 340 left -= chunk; in xen_do_set_identity_and_remap_chunk()
|
/linux-4.1.27/drivers/net/wireless/libertas/ |
D | if_sdio.c | 338 u16 size, type, chunk; in if_sdio_card_to_host() local 362 chunk = sdio_align_size(card->func, size); in if_sdio_card_to_host() 364 ret = sdio_readsb(card->func, card->buffer, card->ioport, chunk); in if_sdio_card_to_host() 368 chunk = card->buffer[0] | (card->buffer[1] << 8); in if_sdio_card_to_host() 372 (int)type, (int)chunk); in if_sdio_card_to_host() 374 if (chunk > size) { in if_sdio_card_to_host() 376 (int)chunk, (int)size); in if_sdio_card_to_host() 381 if (chunk < size) { in if_sdio_card_to_host() 383 (int)chunk, (int)size); in if_sdio_card_to_host() 388 ret = if_sdio_handle_cmd(card, card->buffer + 4, chunk - 4); in if_sdio_card_to_host() [all …]
|
/linux-4.1.27/arch/s390/crypto/ |
D | prng.c | 505 int chunk, n, tmp, ret = 0; in prng_tdes_read() local 533 chunk = min_t(int, nbytes, prng_chunk_size); in prng_tdes_read() 536 n = (chunk + 7) & -8; in prng_tdes_read() 567 if (copy_to_user(ubuf, prng_data->buf, chunk)) in prng_tdes_read() 570 nbytes -= chunk; in prng_tdes_read() 571 ret += chunk; in prng_tdes_read() 572 ubuf += chunk; in prng_tdes_read()
|
/linux-4.1.27/drivers/zorro/ |
D | zorro.c | 102 u32 chunk = start>>Z2RAM_CHUNKSHIFT; in mark_region() local 104 set_bit(chunk, zorro_unused_z2ram); in mark_region() 106 clear_bit(chunk, zorro_unused_z2ram); in mark_region()
|
/linux-4.1.27/net/rxrpc/ |
D | ar-output.c | 547 size_t size, chunk, max, space; in rxrpc_send_data() local 566 chunk = max; in rxrpc_send_data() 567 if (chunk > msg_data_left(msg) && !more) in rxrpc_send_data() 568 chunk = msg_data_left(msg); in rxrpc_send_data() 570 space = chunk + call->conn->size_align; in rxrpc_send_data() 575 _debug("SIZE: %zu/%zu/%zu", chunk, space, size); in rxrpc_send_data() 594 sp->remain = chunk; in rxrpc_send_data()
|
/linux-4.1.27/Documentation/mtd/nand/ |
D | pxa3xx-nand.txt | 15 For instance, if we choose a 2048 data chunk and set "BCH" ECC (see below) 30 OOB, one per chunk read. 51 In the BCH mode the ECC code will be calculated for each transferred chunk 55 So, repeating the above scheme, a 2048B data chunk will be followed by 32B
|
/linux-4.1.27/init/ |
D | do_mounts_md.c | 37 int chunk; member 101 md_setup_args[ent].chunk = 1 << (factor+12); in md_setup() 211 ainfo.chunk_size = md_setup_args[ent].chunk; in md_setup_drive()
|
/linux-4.1.27/drivers/usb/early/ |
D | ehci-dbgp.c | 916 int chunk, ret; in early_dbgp_write() local 941 for (chunk = 0; chunk < DBGP_MAX_PACKET && n > 0; in early_dbgp_write() 942 str++, chunk++, n--) { in early_dbgp_write() 945 buf[chunk] = '\r'; in early_dbgp_write() 952 buf[chunk] = *str; in early_dbgp_write() 954 if (chunk > 0) { in early_dbgp_write() 956 dbgp_endpoint_out, buf, chunk); in early_dbgp_write()
|
/linux-4.1.27/security/keys/ |
D | keyring.c | 233 unsigned long chunk = 0; in keyring_get_key_chunk() local 235 int desc_len = index_key->desc_len, n = sizeof(chunk); in keyring_get_key_chunk() 250 offset += sizeof(chunk) - 1; in keyring_get_key_chunk() 251 offset += (level - 3) * sizeof(chunk); in keyring_get_key_chunk() 259 chunk <<= 8; in keyring_get_key_chunk() 260 chunk |= ((u8*)index_key->description)[--offset]; in keyring_get_key_chunk() 264 chunk <<= 8; in keyring_get_key_chunk() 265 chunk |= (u8)((unsigned long)index_key->type >> in keyring_get_key_chunk() 268 return chunk; in keyring_get_key_chunk()
|
/linux-4.1.27/drivers/usb/ |
D | usb-skeleton.c | 289 size_t chunk = min(available, count); in skel_read() local 309 chunk)) in skel_read() 312 rv = chunk; in skel_read() 314 dev->bulk_in_copied += chunk; in skel_read() 321 skel_do_read_io(dev, count - chunk); in skel_read()
|
/linux-4.1.27/fs/nfs/blocklayout/ |
D | dev.c | 153 u64 chunk; in bl_map_stripe() local 157 chunk = div_u64(offset, dev->chunk_size); in bl_map_stripe() 158 div_u64_rem(chunk, dev->nr_children, &chunk_idx); in bl_map_stripe() 168 offset = chunk * dev->chunk_size; in bl_map_stripe()
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
D | mcdi.c | 1789 size_t chunk; in efx_mcdi_mtd_read() local 1793 chunk = min_t(size_t, end - offset, EFX_MCDI_NVRAM_LEN_MAX); in efx_mcdi_mtd_read() 1795 buffer, chunk); in efx_mcdi_mtd_read() 1798 offset += chunk; in efx_mcdi_mtd_read() 1799 buffer += chunk; in efx_mcdi_mtd_read() 1812 size_t chunk = part->common.mtd.erasesize; in efx_mcdi_mtd_erase() local 1827 chunk); in efx_mcdi_mtd_erase() 1830 offset += chunk; in efx_mcdi_mtd_erase() 1843 size_t chunk; in efx_mcdi_mtd_write() local 1854 chunk = min_t(size_t, end - offset, EFX_MCDI_NVRAM_LEN_MAX); in efx_mcdi_mtd_write() [all …]
|
/linux-4.1.27/drivers/net/wireless/ath/ar5523/ |
D | ar5523.c | 528 struct ar5523_chunk *chunk; in ar5523_data_rx_cb() local 549 chunk = (struct ar5523_chunk *) data->skb->data; in ar5523_data_rx_cb() 551 if (((chunk->flags & UATH_CFLAGS_FINAL) == 0) || in ar5523_data_rx_cb() 552 chunk->seqnum != 0) { in ar5523_data_rx_cb() 554 chunk->seqnum, chunk->flags, in ar5523_data_rx_cb() 555 be16_to_cpu(chunk->length)); in ar5523_data_rx_cb() 581 skb_reserve(data->skb, sizeof(*chunk)); in ar5523_data_rx_cb() 793 struct ar5523_chunk *chunk; in ar5523_tx_work_locked() local 836 chunk = (struct ar5523_chunk *)skb_push(skb, sizeof(*chunk)); in ar5523_tx_work_locked() 838 chunk->seqnum = 0; in ar5523_tx_work_locked() [all …]
|
/linux-4.1.27/net/core/ |
D | datagram.c | 710 int chunk = skb->len - hlen; in skb_copy_and_csum_datagram_msg() local 712 if (!chunk) in skb_copy_and_csum_datagram_msg() 715 if (msg_data_left(msg) < chunk) { in skb_copy_and_csum_datagram_msg() 718 if (skb_copy_datagram_msg(skb, hlen, msg, chunk)) in skb_copy_and_csum_datagram_msg() 723 chunk, &csum)) in skb_copy_and_csum_datagram_msg()
|
D | skbuff.c | 4377 unsigned long chunk; in alloc_skb_with_frags() local 4423 chunk = min_t(unsigned long, data_len, in alloc_skb_with_frags() 4425 skb_fill_page_desc(skb, i, page, 0, chunk); in alloc_skb_with_frags() 4426 data_len -= chunk; in alloc_skb_with_frags()
|
/linux-4.1.27/drivers/nfc/pn544/ |
D | i2c.c | 677 struct pn544_i2c_fw_secure_frame *chunk; in pn544_hci_i2c_fw_secure_write_frame_cmd() local 684 chunk = (struct pn544_i2c_fw_secure_frame *) buf; in pn544_hci_i2c_fw_secure_write_frame_cmd() 686 chunk->cmd = PN544_FW_CMD_SECURE_CHUNK_WRITE; in pn544_hci_i2c_fw_secure_write_frame_cmd() 688 put_unaligned_be16(datalen, &chunk->be_datalen); in pn544_hci_i2c_fw_secure_write_frame_cmd() 690 memcpy(chunk->data, data, datalen); in pn544_hci_i2c_fw_secure_write_frame_cmd() 692 chunklen = sizeof(chunk->cmd) + sizeof(chunk->be_datalen) + datalen; in pn544_hci_i2c_fw_secure_write_frame_cmd()
|
/linux-4.1.27/sound/pci/lola/ |
D | lola_pcm.c | 326 int chunk; in setup_bdle() local 336 chunk = snd_pcm_sgbuf_get_chunk_size(substream, ofs, size); in setup_bdle() 337 bdl[2] = cpu_to_le32(chunk); in setup_bdle() 341 size -= chunk; in setup_bdle() 345 ofs += chunk; in setup_bdle()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
D | qib_init.c | 1672 unsigned e, egrcnt, egrperchunk, chunk, egrsize, egroff; in qib_setup_eagerbufs() local 1689 chunk = rcd->rcvegrbuf_chunks; in qib_setup_eagerbufs() 1694 kzalloc_node(chunk * sizeof(rcd->rcvegrbuf[0]), in qib_setup_eagerbufs() 1701 kmalloc_node(chunk * sizeof(rcd->rcvegrbuf_phys[0]), in qib_setup_eagerbufs() 1723 for (e = chunk = 0; chunk < rcd->rcvegrbuf_chunks; chunk++) { in qib_setup_eagerbufs() 1724 dma_addr_t pa = rcd->rcvegrbuf_phys[chunk]; in qib_setup_eagerbufs() 1728 memset(rcd->rcvegrbuf[chunk], 0, size); in qib_setup_eagerbufs()
|
D | qib_driver.c | 286 const u32 chunk = etail >> rcd->rcvegrbufs_perchunk_shift; in qib_get_egrbuf() local 289 return rcd->rcvegrbuf[chunk] + (idx << rcd->dd->rcvegrbufsize_shift); in qib_get_egrbuf()
|
/linux-4.1.27/net/ipv4/ |
D | tcp.c | 1741 int chunk; in tcp_recvmsg() local 1745 if ((chunk = len - tp->ucopy.len) != 0) { in tcp_recvmsg() 1746 NET_ADD_STATS_USER(sock_net(sk), LINUX_MIB_TCPDIRECTCOPYFROMBACKLOG, chunk); in tcp_recvmsg() 1747 len -= chunk; in tcp_recvmsg() 1748 copied += chunk; in tcp_recvmsg() 1756 if ((chunk = len - tp->ucopy.len) != 0) { in tcp_recvmsg() 1757 NET_ADD_STATS_USER(sock_net(sk), LINUX_MIB_TCPDIRECTCOPYFROMPREQUEUE, chunk); in tcp_recvmsg() 1758 len -= chunk; in tcp_recvmsg() 1759 copied += chunk; in tcp_recvmsg() 1836 int chunk; in tcp_recvmsg() local [all …]
|
D | tcp_input.c | 4516 int chunk = min_t(unsigned int, skb->len, in tcp_data_queue() local 4522 if (!skb_copy_datagram_msg(skb, 0, tp->ucopy.msg, chunk)) { in tcp_data_queue() 4523 tp->ucopy.len -= chunk; in tcp_data_queue() 4524 tp->copied_seq += chunk; in tcp_data_queue() 4525 eaten = (chunk == skb->len); in tcp_data_queue() 5039 int chunk = skb->len - hlen; in tcp_copy_to_iovec() local 5044 err = skb_copy_datagram_msg(skb, hlen, tp->ucopy.msg, chunk); in tcp_copy_to_iovec() 5049 tp->ucopy.len -= chunk; in tcp_copy_to_iovec() 5050 tp->copied_seq += chunk; in tcp_copy_to_iovec()
|
D | tcp_output.c | 1810 u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); in tcp_tso_should_defer() local 1815 chunk /= win_divisor; in tcp_tso_should_defer() 1816 if (limit >= chunk) in tcp_tso_should_defer()
|
/linux-4.1.27/drivers/char/ |
D | mem.c | 639 size_t chunk = iov_iter_count(iter), n; in read_iter_zero() local 641 if (chunk > PAGE_SIZE) in read_iter_zero() 642 chunk = PAGE_SIZE; /* Just for latency reasons */ in read_iter_zero() 643 n = iov_iter_zero(chunk, iter); in read_iter_zero()
|
D | random.c | 1264 int chunk = min(nbytes, (int)sizeof(unsigned long)); in get_random_bytes_arch() local 1269 memcpy(p, &v, chunk); in get_random_bytes_arch() 1270 p += chunk; in get_random_bytes_arch() 1271 nbytes -= chunk; in get_random_bytes_arch()
|
/linux-4.1.27/net/caif/ |
D | caif_socket.c | 376 int chunk; in caif_stream_recvmsg() local 427 chunk = min_t(unsigned int, skb->len, size); in caif_stream_recvmsg() 428 if (memcpy_to_msg(msg, skb->data, chunk)) { in caif_stream_recvmsg() 434 copied += chunk; in caif_stream_recvmsg() 435 size -= chunk; in caif_stream_recvmsg() 439 skb_pull(skb, chunk); in caif_stream_recvmsg()
|
/linux-4.1.27/Documentation/fb/ |
D | sh7760fb.txt | 11 * Framebuffer memory must be a large chunk allocated at the top 14 get a large enough contiguous chunk of memory.
|
/linux-4.1.27/Documentation/ |
D | dell_rbu.txt | 20 image methods. In case of monolithic the driver allocates a contiguous chunk 57 The user creates packets header, gets the chunk of the BIOS image and 58 places it next to the packetheader; now, the packetheader + BIOS image chunk
|
D | assoc_array.txt | 139 (1) Get a chunk of index key from caller data: 143 This should return a chunk of caller-supplied index key starting at the 149 (2) Get a chunk of an object's index key. 360 The index key is read in chunks of machine word. Each chunk is subdivided into
|
D | dma-buf-sharing.txt | 278 PAGE_SIZE size. Before accessing a chunk it needs to be mapped, which returns 279 a pointer in kernel virtual address space. Afterwards the chunk needs to be 280 unmapped again. There is no limit on how often a given chunk can be mapped 282 before mapping the same chunk again.
|
D | DMA-attributes.txt | 100 be mapped as contiguous chunk into device dma address space. By
|
D | ramoops.txt | 32 power of two) and each oops/panic writes a "record_size" chunk of
|
D | md.txt | 12 md=<md device no.>,<raid level>,<chunk size factor>,<fault level>,dev0,dev1,...,devn 30 chunk size factor = (raid-0 and raid-1 only) 31 Set the chunk size as 4k << n. 309 The size, in bytes, of the chunk which will be represented by a
|
D | xillybus.txt | 87 their attributes), there isn't one specific chunk of logic being the Xillybus
|
D | HOWTO | 198 If you already have a chunk of code that you want to put into the kernel
|
D | kernel-parameters.txt | 845 chunk of memory for debug pagealloc if we don't enable 1437 ahash performance varies for different chunk sizes on 2150 used for mtrr cleanup. It is largest continuous chunk 2847 percpu_alloc= Select which percpu first chunk allocator to use.
|
D | DMA-API-HOWTO.txt | 414 exists (for example) to guarantee that if you allocate a chunk
|
/linux-4.1.27/drivers/staging/media/lirc/ |
D | lirc_serial.c | 399 unsigned char chunk, shifted; in send_pulse_irdeo() local 404 chunk = 3; in send_pulse_irdeo() 406 chunk = 1; in send_pulse_irdeo() 408 shifted = chunk << (i * 3); in send_pulse_irdeo()
|
/linux-4.1.27/include/drm/ |
D | drm_dp_mst_helper.h | 206 u8 chunk[48]; member 361 u8 chunk[48]; member
|
/linux-4.1.27/drivers/tty/ |
D | tty_io.c | 1114 unsigned int chunk; in do_tty_write() local 1136 chunk = 2048; in do_tty_write() 1138 chunk = 65536; in do_tty_write() 1139 if (count < chunk) in do_tty_write() 1140 chunk = count; in do_tty_write() 1143 if (tty->write_cnt < chunk) { in do_tty_write() 1146 if (chunk < 1024) in do_tty_write() 1147 chunk = 1024; in do_tty_write() 1149 buf_chunk = kmalloc(chunk, GFP_KERNEL); in do_tty_write() 1155 tty->write_cnt = chunk; in do_tty_write() [all …]
|
/linux-4.1.27/drivers/irqchip/ |
D | irq-gic-v3-its.c | 688 static int its_chunk_to_lpi(int chunk) in its_chunk_to_lpi() argument 690 return (chunk << IRQS_PER_CHUNK_SHIFT) + 8192; in its_chunk_to_lpi() 757 int chunk = its_lpi_to_chunk(lpi); in its_lpi_free() local 758 BUG_ON(chunk > lpi_chunks); in its_lpi_free() 759 if (test_bit(chunk, lpi_bitmap)) { in its_lpi_free() 760 clear_bit(chunk, lpi_bitmap); in its_lpi_free() 762 pr_err("Bad LPI chunk %d\n", chunk); in its_lpi_free()
|
/linux-4.1.27/arch/xtensa/lib/ |
D | memcopy.S | 151 add a8, a8, a3 # a8 = end of last 16B source chunk 227 add a10, a10, a3 # a10 = end of last 16B source chunk 425 sub a8, a3, a8 # a8 = start of first 16B source chunk 502 sub a10, a3, a10 # a10 = start of first 16B source chunk
|
D | usercopy.S | 161 add a8, a8, a3 # a8 = end of last 16B source chunk 226 add a10, a10, a3 # a10 = end of last 16B source chunk
|
D | memset.S | 73 add a6, a6, a5 # a6 = end of last 16B chunk
|
D | strnlen_user.S | 77 add a10, a10, a4 # a10 = end of last 4B chunk
|
/linux-4.1.27/fs/gfs2/ |
D | bmap.c | 1013 u64 chunk; in gfs2_journaled_truncate() local 1017 chunk = oldsize - newsize; in gfs2_journaled_truncate() 1018 if (chunk > max_chunk) in gfs2_journaled_truncate() 1019 chunk = max_chunk; in gfs2_journaled_truncate() 1020 truncate_pagecache(inode, oldsize - chunk); in gfs2_journaled_truncate() 1021 oldsize -= chunk; in gfs2_journaled_truncate()
|
/linux-4.1.27/drivers/memstick/core/ |
D | mspro_block.c | 697 static int mspro_block_issue_req(struct memstick_dev *card, int chunk) in mspro_block_issue_req() argument 704 while (chunk) { in mspro_block_issue_req() 712 chunk = __blk_end_request_cur(msb->block_req, -ENOMEM); in mspro_block_issue_req() 739 chunk = 1; in mspro_block_issue_req() 746 int chunk, cnt; in mspro_block_complete_req() local 779 chunk = __blk_end_request(msb->block_req, error, t_len); in mspro_block_complete_req() 781 error = mspro_block_issue_req(card, chunk); in mspro_block_complete_req()
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
D | radeon_cs.c | 78 struct radeon_cs_chunk *chunk; in radeon_cs_parser_relocs() local 87 chunk = p->chunk_relocs; in radeon_cs_parser_relocs() 90 p->nrelocs = chunk->length_dw / 4; in radeon_cs_parser_relocs() 103 r = (struct drm_radeon_cs_reloc *)&chunk->kdata[i*4]; in radeon_cs_parser_relocs()
|
/linux-4.1.27/arch/arm/kernel/ |
D | traps.c | 540 unsigned long chunk = min(PAGE_SIZE, end - start); in __do_cache_op() local 545 ret = flush_cache_user_range(start, start + chunk); in __do_cache_op() 550 start += chunk; in __do_cache_op()
|
/linux-4.1.27/drivers/thermal/ti-soc-thermal/ |
D | ti-bandgap.c | 1239 void __iomem *chunk; in ti_bandgap_build() local 1244 chunk = devm_ioremap_resource(&pdev->dev, res); in ti_bandgap_build() 1246 bgp->base = chunk; in ti_bandgap_build() 1247 if (IS_ERR(chunk)) in ti_bandgap_build() 1248 return ERR_CAST(chunk); in ti_bandgap_build()
|
/linux-4.1.27/drivers/staging/lustre/lustre/osc/ |
D | osc_cache.c | 632 pgoff_t chunk; in osc_extent_find() local 652 chunk = index >> ppc_bits; in osc_extent_find() 687 if (chunk > ext_chk_end + 1) in osc_extent_find() 700 if (chunk + 1 < ext_chk_start) { in osc_extent_find() 753 if (chunk + 1 == ext_chk_start) { in osc_extent_find() 763 } else if (chunk == ext_chk_end + 1) { in osc_extent_find() 1144 pgoff_t chunk = index >> ppc_bits; in osc_extent_expand() local 1154 if (chunk > end_chunk + 1) { in osc_extent_expand() 1159 if (end_chunk >= chunk) { in osc_extent_expand() 1164 LASSERT(end_chunk + 1 == chunk); in osc_extent_expand() [all …]
|
/linux-4.1.27/sound/pci/hda/ |
D | hda_controller.c | 299 int chunk; in setup_bdle() local 309 chunk = snd_sgbuf_get_chunk_size(dmab, ofs, size); in setup_bdle() 313 if (chunk > remain) in setup_bdle() 314 chunk = remain; in setup_bdle() 316 bdl[2] = cpu_to_le32(chunk); in setup_bdle() 320 size -= chunk; in setup_bdle() 324 ofs += chunk; in setup_bdle()
|
/linux-4.1.27/arch/arm/lib/ |
D | memzero.S | 42 cmp r1, #16 @ 1 we can skip this chunk if we
|
/linux-4.1.27/drivers/mmc/host/ |
D | sdhci.c | 318 size_t blksize, len, chunk; in sdhci_read_block_pio() local 325 chunk = 0; in sdhci_read_block_pio() 341 if (chunk == 0) { in sdhci_read_block_pio() 343 chunk = 4; in sdhci_read_block_pio() 350 chunk--; in sdhci_read_block_pio() 363 size_t blksize, len, chunk; in sdhci_write_block_pio() local 370 chunk = 0; in sdhci_write_block_pio() 387 scratch |= (u32)*buf << (chunk * 8); in sdhci_write_block_pio() 390 chunk++; in sdhci_write_block_pio() 393 if ((chunk == 4) || ((len == 0) && (blksize == 0))) { in sdhci_write_block_pio() [all …]
|
/linux-4.1.27/drivers/macintosh/ |
D | smu.c | 899 unsigned int chunk; in smu_read_datablock() local 907 chunk = 0xe; in smu_read_datablock() 910 unsigned int clen = min(len, chunk); in smu_read_datablock() 915 cmd.reply_len = chunk; in smu_read_datablock()
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | vio.c | 393 size_t avail = 0, level, chunk, need; in vio_cmo_balance() local 441 chunk = min_t(size_t, avail, VIO_CMO_BALANCE_CHUNK); in vio_cmo_balance() 442 chunk = min(chunk, (viodev->cmo.desired - in vio_cmo_balance() 444 viodev->cmo.entitled += chunk; in vio_cmo_balance()
|
D | prom_init.c | 1988 unsigned long room, chunk; in make_room() local 1998 chunk = alloc_up(room, 0); in make_room() 1999 if (chunk == 0) in make_room() 2002 *mem_end = chunk + room; in make_room()
|
/linux-4.1.27/net/unix/ |
D | af_unix.c | 2104 int chunk; in unix_stream_recvmsg() local 2178 chunk = min_t(unsigned int, unix_skb_len(skb) - skip, size); in unix_stream_recvmsg() 2180 msg, chunk)) { in unix_stream_recvmsg() 2185 copied += chunk; in unix_stream_recvmsg() 2186 size -= chunk; in unix_stream_recvmsg() 2190 UNIXCB(skb).consumed += chunk; in unix_stream_recvmsg() 2192 sk_peek_offset_bwd(sk, chunk); in unix_stream_recvmsg() 2211 sk_peek_offset_fwd(sk, chunk); in unix_stream_recvmsg()
|
/linux-4.1.27/drivers/net/wimax/i2400m/ |
D | fw.c | 646 static int i2400m_download_chunk(struct i2400m *i2400m, const void *chunk, in i2400m_download_chunk() argument 660 "direct %u do_csum %u)\n", i2400m, chunk, __chunk_len, in i2400m_download_chunk() 663 memcpy(buf->cmd_payload, chunk, __chunk_len); in i2400m_download_chunk() 676 "direct %u do_csum %u) = %d\n", i2400m, chunk, __chunk_len, in i2400m_download_chunk()
|
/linux-4.1.27/drivers/crypto/qat/qat_common/ |
D | qat_uclo.c | 522 char *chunk; in qat_uclo_map_chunk() local 530 chunk = buf + file_chunk->offset; in qat_uclo_map_chunk() 532 chunk, file_chunk->size)) in qat_uclo_map_chunk() 537 obj_hdr->file_buff = chunk; in qat_uclo_map_chunk()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
D | ipath_file_ops.c | 897 unsigned e, egrcnt, egrperchunk, chunk, egrsize, egroff; in ipath_create_user_egr() local 917 chunk = pd->port_rcvegrbuf_chunks; in ipath_create_user_egr() 920 pd->port_rcvegrbuf = kmalloc(chunk * sizeof(pd->port_rcvegrbuf[0]), in ipath_create_user_egr() 927 kmalloc(chunk * sizeof(pd->port_rcvegrbuf_phys[0]), in ipath_create_user_egr() 947 for (e = chunk = 0; chunk < pd->port_rcvegrbuf_chunks; chunk++) { in ipath_create_user_egr() 948 dma_addr_t pa = pd->port_rcvegrbuf_phys[chunk]; in ipath_create_user_egr()
|
/linux-4.1.27/drivers/media/dvb-frontends/ |
D | drxk_hard.h | 224 u8 chunk[256]; member
|
D | drxk_hard.c | 449 int chunk = blk_size > state->m_chunk_size ? in write_block() local 451 u8 *adr_buf = &state->chunk[0]; in write_block() 461 if (chunk == state->m_chunk_size) in write_block() 462 chunk -= 2; in write_block() 469 memcpy(&state->chunk[adr_length], p_block, chunk); in write_block() 474 for (i = 0; i < chunk; i++) in write_block() 479 &state->chunk[0], chunk + adr_length); in write_block() 485 p_block += chunk; in write_block() 486 address += (chunk >> 1); in write_block() 487 blk_size -= chunk; in write_block()
|
/linux-4.1.27/drivers/input/misc/ |
D | ims-pcu.c | 515 u8 command, int chunk, int len) in ims_pcu_send_cmd_chunk() argument 527 command, chunk, error); in ims_pcu_send_cmd_chunk() 538 int chunk = 0; in ims_pcu_send_command() local 566 ++chunk, count); in ims_pcu_send_command() 584 error = ims_pcu_send_cmd_chunk(pcu, command, ++chunk, count); in ims_pcu_send_command() 597 return ims_pcu_send_cmd_chunk(pcu, command, ++chunk, count); in ims_pcu_send_command()
|
/linux-4.1.27/net/irda/ |
D | af_irda.c | 1430 int chunk; in irda_recvmsg_stream() local 1470 chunk = min_t(unsigned int, skb->len, size); in irda_recvmsg_stream() 1471 if (memcpy_to_msg(msg, skb->data, chunk)) { in irda_recvmsg_stream() 1477 copied += chunk; in irda_recvmsg_stream() 1478 size -= chunk; in irda_recvmsg_stream() 1482 skb_pull(skb, chunk); in irda_recvmsg_stream()
|
/linux-4.1.27/net/decnet/ |
D | af_decnet.c | 1760 unsigned int chunk = skb->len; in dn_recvmsg() local 1763 if ((chunk + copied) > size) in dn_recvmsg() 1764 chunk = size - copied; in dn_recvmsg() 1766 if (memcpy_to_msg(msg, skb->data, chunk)) { in dn_recvmsg() 1770 copied += chunk; in dn_recvmsg() 1773 skb_pull(skb, chunk); in dn_recvmsg()
|
/linux-4.1.27/drivers/gpu/drm/ |
D | drm_dp_mst_topology.c | 350 memcpy(&msg->chunk[0], replybuf + hdrlen, msg->curchunk_idx); in drm_dp_sideband_msg_build() 352 memcpy(&msg->chunk[msg->curchunk_idx], replybuf, replybuflen); in drm_dp_sideband_msg_build() 358 crc4 = drm_dp_msg_data_crc4(msg->chunk, msg->curchunk_len - 1); in drm_dp_sideband_msg_build() 360 memcpy(&msg->msg[msg->curlen], msg->chunk, msg->curchunk_len - 1); in drm_dp_sideband_msg_build() 1431 u8 chunk[48]; in process_single_tx_qlock() local 1463 drm_dp_encode_sideband_msg_hdr(&hdr, chunk, &idx); in process_single_tx_qlock() 1464 memcpy(&chunk[idx], &txmsg->msg[txmsg->cur_offset], tosend); in process_single_tx_qlock() 1466 drm_dp_crc_sideband_chunk_req(&chunk[idx], tosend); in process_single_tx_qlock() 1469 ret = drm_dp_send_sideband_msg(mgr, up, chunk, idx); in process_single_tx_qlock()
|
/linux-4.1.27/drivers/net/ethernet/ibm/emac/ |
D | core.c | 1421 int chunk = min(len, MAL_MAX_TX_SIZE); in emac_xmit_split() local 1422 len -= chunk; in emac_xmit_split() 1433 dev->tx_desc[slot].data_len = (u16) chunk; in emac_xmit_split() 1440 pd += chunk; in emac_xmit_split() 1450 int len = skb->len, chunk; in emac_start_xmit_sg() local 1474 chunk = min(len, MAL_MAX_TX_SIZE); in emac_start_xmit_sg() 1477 dev->tx_desc[slot].data_len = (u16) chunk; in emac_start_xmit_sg() 1478 len -= chunk; in emac_start_xmit_sg() 1480 slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags, in emac_start_xmit_sg()
|
/linux-4.1.27/security/apparmor/ |
D | policy_unpack.c | 124 static size_t unpack_u16_chunk(struct aa_ext *e, char **chunk) in unpack_u16_chunk() argument 134 *chunk = e->pos; in unpack_u16_chunk()
|
/linux-4.1.27/Documentation/dmaengine/ |
D | provider.txt | 69 but all of them will require, for every chunk, at least the source and 188 scatter-gather transfer, with a single chunk to transfer, it's a 204 deal with a single chunk to copy or a collection of them, here, 212 - A cyclic transfer is a transfer where the chunk collection will
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/socklnd/ |
D | socklnd_cb.c | 2583 int chunk = ksocknal_data.ksnd_peer_hash_size; in ksocknal_reaper() local 2593 chunk = (chunk * n * p) / in ksocknal_reaper() 2595 if (chunk == 0) in ksocknal_reaper() 2596 chunk = 1; in ksocknal_reaper() 2598 for (i = 0; i < chunk; i++) { in ksocknal_reaper()
|
/linux-4.1.27/Documentation/ABI/testing/ |
D | sysfs-driver-wacom | 72 byte chunk encodes the image data for two consecutive lines on
|
D | sysfs-block | 122 chunk size. A properly aligned multiple of
|
/linux-4.1.27/Documentation/ko_KR/ |
D | HOWTO | 491 리눅스 커널 커뮤니티는 한꺼번에 굉장히 큰 코드의 묶음(chunk)을 쉽게
|
/linux-4.1.27/Documentation/arm/Samsung-S3C24XX/ |
D | Suspend.txt | 128 Defines the size of memory each CRC chunk covers. A smaller value
|
/linux-4.1.27/Documentation/filesystems/cifs/ |
D | TODO | 14 - T10 copy offload (copy chunk is only mechanism supported)
|
/linux-4.1.27/Documentation/filesystems/ |
D | ntfs.txt | 318 For a striped volume, i.e. raid level 0, you will need to know the chunk size 394 Windows by default uses a stripe chunk size of 64k, so you probably want the 395 "chunk-size 64k" option for each raid-disk, too. 405 chunk-size 64k
|
D | sysv-fs.txt | 84 to the free blocks pertaining to the next chunk; the first of these
|
D | btrfs.txt | 141 Specify that 1 metadata chunk should be allocated after every <value>
|
D | ext4.txt | 264 disks * RAID chunk size in file system blocks.
|
/linux-4.1.27/crypto/ |
D | drbg.c | 1449 unsigned int chunk = 0; in drbg_generate_long() local 1451 chunk = slice ? drbg_max_request_bytes(drbg) : (buflen - len); in drbg_generate_long() 1452 tmplen = drbg_generate(drbg, buf + len, chunk, addtl); in drbg_generate_long()
|
/linux-4.1.27/arch/s390/kernel/ |
D | head.S | 108 ssch 0(%r3) # load chunk of 1600 bytes
|
/linux-4.1.27/drivers/acpi/apei/ |
D | ghes.c | 202 struct gen_pool_chunk *chunk, in ghes_estatus_pool_free_chunk_page() argument 205 free_page(chunk->start_addr); in ghes_estatus_pool_free_chunk_page()
|
/linux-4.1.27/arch/m68k/ |
D | Kconfig.cpu | 361 bool "Use one physical chunk of memory only" if ADVANCED && !SUN3 366 Ignore all but the first contiguous chunk of physical memory for VM
|
/linux-4.1.27/drivers/mtd/nand/ |
D | nand_base.c | 1773 int chunk = chip->ecc.bytes + chip->ecc.prepad + chip->ecc.postpad; in nand_read_oob_syndrome() local 1781 pos = eccsize + i * (eccsize + chunk); in nand_read_oob_syndrome() 1788 toread = min_t(int, length, chunk); in nand_read_oob_syndrome() 1832 int chunk = chip->ecc.bytes + chip->ecc.prepad + chip->ecc.postpad; in nand_write_oob_syndrome() local 1843 pos = steps * (eccsize + chunk); in nand_write_oob_syndrome() 1862 pos = eccsize + i * (eccsize + chunk); in nand_write_oob_syndrome() 1867 len = min_t(int, length, chunk); in nand_write_oob_syndrome()
|
/linux-4.1.27/arch/sh/lib64/ |
D | memcpy.S | 30 ! bytes memory chunk to b copied, the rest of the word can be read
|
D | copy_user_memcpy.S | 28 ! bytes memory chunk to b copied, the rest of the word can be read
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_main.c | 2455 unsigned int chunk = in cxgb_extension_ioctl() local 2459 t3_mc7_bd_read(mem, t.addr / 8, chunk / 8, in cxgb_extension_ioctl() 2463 if (copy_to_user(useraddr, buf, chunk)) in cxgb_extension_ioctl() 2465 useraddr += chunk; in cxgb_extension_ioctl() 2466 t.addr += chunk; in cxgb_extension_ioctl() 2467 t.len -= chunk; in cxgb_extension_ioctl()
|
D | sge.c | 326 unsigned int chunk) in reclaim_completed_tx() argument 330 reclaim = min(chunk, reclaim); in reclaim_completed_tx()
|
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
D | o2iblnd_cb.c | 3172 int chunk = kiblnd_data.kib_peer_hash_size; in kiblnd_connd() local 3186 chunk = (chunk * n * p) / in kiblnd_connd() 3188 if (chunk == 0) in kiblnd_connd() 3189 chunk = 1; in kiblnd_connd() 3191 for (i = 0; i < chunk; i++) { in kiblnd_connd()
|
/linux-4.1.27/fs/cramfs/ |
D | README | 58 applied to the i'th blksize-sized chunk of the input data.
|
/linux-4.1.27/Documentation/early-userspace/ |
D | README | 17 - initramfs, a chunk of code that unpacks the compressed cpio image
|
/linux-4.1.27/drivers/net/ethernet/marvell/ |
D | skge.c | 2530 u32 chunk, ram_addr; in skge_up() local 2595 chunk = (hw->ram_size - hw->ram_offset) / (hw->ports * 2); in skge_up() 2596 ram_addr = hw->ram_offset + 2 * chunk * port; in skge_up() 2598 skge_ramset(hw, rxqaddr[port], ram_addr, chunk); in skge_up() 2602 skge_ramset(hw, txqaddr[port], ram_addr+chunk, chunk); in skge_up()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/ce/fuc/ |
D | com.fuc | 208 // get a chunk of stack space, aligned to 256 byte boundary 419 // zero out a chunk of the stack to store the swizzle into
|
/linux-4.1.27/arch/cris/arch-v10/ |
D | README.mm | 228 happens first is that a virtual address chunk is allocated to the request using
|
/linux-4.1.27/drivers/scsi/ |
D | ips.h | 733 IPS_CHUNK chunk[IPS_MAX_CHUNKS]; member
|
/linux-4.1.27/Documentation/powerpc/ |
D | firmware-assisted-dump.txt | 53 NOTE: The term 'boot memory' means size of the low memory chunk
|
/linux-4.1.27/drivers/net/wireless/ath/ath10k/ |
D | wmi.c | 3774 struct host_memory_chunk *chunk; in ath10k_wmi_put_host_mem_chunks() local 3780 chunk = &chunks->items[i]; in ath10k_wmi_put_host_mem_chunks() 3781 chunk->ptr = __cpu_to_le32(ar->wmi.mem_chunks[i].paddr); in ath10k_wmi_put_host_mem_chunks() 3782 chunk->size = __cpu_to_le32(ar->wmi.mem_chunks[i].len); in ath10k_wmi_put_host_mem_chunks() 3783 chunk->req_id = __cpu_to_le32(ar->wmi.mem_chunks[i].req_id); in ath10k_wmi_put_host_mem_chunks()
|
/linux-4.1.27/Documentation/scsi/ |
D | st.txt | 215 (PAGE_SIZE << ST_FIRST_ORDER) bytes if the system can give a chunk of 220 buffer as one chunk.
|
/linux-4.1.27/fs/ext4/ |
D | mballoc.c | 670 ext4_grpblk_t chunk; in ext4_mb_mark_free_simple() local 686 chunk = 1 << min; in ext4_mb_mark_free_simple() 694 len -= chunk; in ext4_mb_mark_free_simple() 695 first += chunk; in ext4_mb_mark_free_simple()
|
/linux-4.1.27/Documentation/usb/ |
D | usbmon.txt | 263 The size may be rounded down to the next chunk (or page). If the requested
|
D | WUSB-Design-overview.txt | 372 we issue another URB to read into the destination buffer the chunk of
|
/linux-4.1.27/scripts/ |
D | checkpatch.pl | 4489 for my $chunk (@chunks) { 4490 my ($cond, $block) = @{$chunk};
|
/linux-4.1.27/drivers/net/wireless/ipw2x00/ |
D | ipw2200.c | 3211 struct fw_chunk *chunk; in ipw_load_firmware() local 3251 chunk = (struct fw_chunk *)(data + offset); in ipw_load_firmware() 3253 chunk_len = le32_to_cpu(chunk->length); in ipw_load_firmware() 3278 nr, le32_to_cpu(chunk->address), in ipw_load_firmware()
|
/linux-4.1.27/Documentation/cgroups/ |
D | cpusets.txt | 561 4 : search nodes in a chunk of node [on NUMA system]
|
/linux-4.1.27/net/netfilter/ |
D | Kconfig | 1349 and SCTP chunk types.
|
/linux-4.1.27/Documentation/block/ |
D | biodoc.txt | 378 which were generated for each such chunk.
|
/linux-4.1.27/Documentation/networking/ |
D | ip-sysctl.txt | 1769 a listening sctp socket to a connecting client in the INIT-ACK chunk.
|
/linux-4.1.27/drivers/video/fbdev/ |
D | Kconfig | 2226 This is a `virtual' frame buffer device. It operates on a chunk of
|