/linux-4.1.27/Documentation/watchdog/ |
H A D | Makefile | 1 subdir-y := src
|
/linux-4.1.27/drivers/infiniband/core/ |
H A D | uverbs_marshall.c | 37 struct ib_ah_attr *src) ib_copy_ah_attr_to_user() 39 memcpy(dst->grh.dgid, src->grh.dgid.raw, sizeof src->grh.dgid); ib_copy_ah_attr_to_user() 40 dst->grh.flow_label = src->grh.flow_label; ib_copy_ah_attr_to_user() 41 dst->grh.sgid_index = src->grh.sgid_index; ib_copy_ah_attr_to_user() 42 dst->grh.hop_limit = src->grh.hop_limit; ib_copy_ah_attr_to_user() 43 dst->grh.traffic_class = src->grh.traffic_class; ib_copy_ah_attr_to_user() 45 dst->dlid = src->dlid; ib_copy_ah_attr_to_user() 46 dst->sl = src->sl; ib_copy_ah_attr_to_user() 47 dst->src_path_bits = src->src_path_bits; ib_copy_ah_attr_to_user() 48 dst->static_rate = src->static_rate; ib_copy_ah_attr_to_user() 49 dst->is_global = src->ah_flags & IB_AH_GRH ? 1 : 0; ib_copy_ah_attr_to_user() 50 dst->port_num = src->port_num; ib_copy_ah_attr_to_user() 56 struct ib_qp_attr *src) ib_copy_qp_attr_to_user() 58 dst->qp_state = src->qp_state; ib_copy_qp_attr_to_user() 59 dst->cur_qp_state = src->cur_qp_state; ib_copy_qp_attr_to_user() 60 dst->path_mtu = src->path_mtu; ib_copy_qp_attr_to_user() 61 dst->path_mig_state = src->path_mig_state; ib_copy_qp_attr_to_user() 62 dst->qkey = src->qkey; ib_copy_qp_attr_to_user() 63 dst->rq_psn = src->rq_psn; ib_copy_qp_attr_to_user() 64 dst->sq_psn = src->sq_psn; ib_copy_qp_attr_to_user() 65 dst->dest_qp_num = src->dest_qp_num; ib_copy_qp_attr_to_user() 66 dst->qp_access_flags = src->qp_access_flags; ib_copy_qp_attr_to_user() 68 dst->max_send_wr = src->cap.max_send_wr; ib_copy_qp_attr_to_user() 69 dst->max_recv_wr = src->cap.max_recv_wr; ib_copy_qp_attr_to_user() 70 dst->max_send_sge = src->cap.max_send_sge; ib_copy_qp_attr_to_user() 71 dst->max_recv_sge = src->cap.max_recv_sge; ib_copy_qp_attr_to_user() 72 dst->max_inline_data = src->cap.max_inline_data; ib_copy_qp_attr_to_user() 74 ib_copy_ah_attr_to_user(&dst->ah_attr, &src->ah_attr); ib_copy_qp_attr_to_user() 75 ib_copy_ah_attr_to_user(&dst->alt_ah_attr, &src->alt_ah_attr); ib_copy_qp_attr_to_user() 77 dst->pkey_index = src->pkey_index; ib_copy_qp_attr_to_user() 78 dst->alt_pkey_index = src->alt_pkey_index; ib_copy_qp_attr_to_user() 79 dst->en_sqd_async_notify = src->en_sqd_async_notify; ib_copy_qp_attr_to_user() 80 dst->sq_draining = src->sq_draining; ib_copy_qp_attr_to_user() 81 dst->max_rd_atomic = src->max_rd_atomic; ib_copy_qp_attr_to_user() 82 dst->max_dest_rd_atomic = src->max_dest_rd_atomic; ib_copy_qp_attr_to_user() 83 dst->min_rnr_timer = src->min_rnr_timer; ib_copy_qp_attr_to_user() 84 dst->port_num = src->port_num; ib_copy_qp_attr_to_user() 85 dst->timeout = src->timeout; ib_copy_qp_attr_to_user() 86 dst->retry_cnt = src->retry_cnt; ib_copy_qp_attr_to_user() 87 dst->rnr_retry = src->rnr_retry; ib_copy_qp_attr_to_user() 88 dst->alt_port_num = src->alt_port_num; ib_copy_qp_attr_to_user() 89 dst->alt_timeout = src->alt_timeout; ib_copy_qp_attr_to_user() 95 struct ib_sa_path_rec *src) ib_copy_path_rec_to_user() 97 memcpy(dst->dgid, src->dgid.raw, sizeof src->dgid); ib_copy_path_rec_to_user() 98 memcpy(dst->sgid, src->sgid.raw, sizeof src->sgid); ib_copy_path_rec_to_user() 100 dst->dlid = src->dlid; ib_copy_path_rec_to_user() 101 dst->slid = src->slid; ib_copy_path_rec_to_user() 102 dst->raw_traffic = src->raw_traffic; ib_copy_path_rec_to_user() 103 dst->flow_label = src->flow_label; ib_copy_path_rec_to_user() 104 dst->hop_limit = src->hop_limit; ib_copy_path_rec_to_user() 105 dst->traffic_class = src->traffic_class; ib_copy_path_rec_to_user() 106 dst->reversible = src->reversible; ib_copy_path_rec_to_user() 107 dst->numb_path = src->numb_path; ib_copy_path_rec_to_user() 108 dst->pkey = src->pkey; ib_copy_path_rec_to_user() 109 dst->sl = src->sl; ib_copy_path_rec_to_user() 110 dst->mtu_selector = src->mtu_selector; ib_copy_path_rec_to_user() 111 dst->mtu = src->mtu; ib_copy_path_rec_to_user() 112 dst->rate_selector = src->rate_selector; ib_copy_path_rec_to_user() 113 dst->rate = src->rate; ib_copy_path_rec_to_user() 114 dst->packet_life_time = src->packet_life_time; ib_copy_path_rec_to_user() 115 dst->preference = src->preference; ib_copy_path_rec_to_user() 116 dst->packet_life_time_selector = src->packet_life_time_selector; ib_copy_path_rec_to_user() 121 struct ib_user_path_rec *src) ib_copy_path_rec_from_user() 123 memcpy(dst->dgid.raw, src->dgid, sizeof dst->dgid); ib_copy_path_rec_from_user() 124 memcpy(dst->sgid.raw, src->sgid, sizeof dst->sgid); ib_copy_path_rec_from_user() 126 dst->dlid = src->dlid; ib_copy_path_rec_from_user() 127 dst->slid = src->slid; ib_copy_path_rec_from_user() 128 dst->raw_traffic = src->raw_traffic; ib_copy_path_rec_from_user() 129 dst->flow_label = src->flow_label; ib_copy_path_rec_from_user() 130 dst->hop_limit = src->hop_limit; ib_copy_path_rec_from_user() 131 dst->traffic_class = src->traffic_class; ib_copy_path_rec_from_user() 132 dst->reversible = src->reversible; ib_copy_path_rec_from_user() 133 dst->numb_path = src->numb_path; ib_copy_path_rec_from_user() 134 dst->pkey = src->pkey; ib_copy_path_rec_from_user() 135 dst->sl = src->sl; ib_copy_path_rec_from_user() 136 dst->mtu_selector = src->mtu_selector; ib_copy_path_rec_from_user() 137 dst->mtu = src->mtu; ib_copy_path_rec_from_user() 138 dst->rate_selector = src->rate_selector; ib_copy_path_rec_from_user() 139 dst->rate = src->rate; ib_copy_path_rec_from_user() 140 dst->packet_life_time = src->packet_life_time; ib_copy_path_rec_from_user() 141 dst->preference = src->preference; ib_copy_path_rec_from_user() 142 dst->packet_life_time_selector = src->packet_life_time_selector; ib_copy_path_rec_from_user() 36 ib_copy_ah_attr_to_user(struct ib_uverbs_ah_attr *dst, struct ib_ah_attr *src) ib_copy_ah_attr_to_user() argument 55 ib_copy_qp_attr_to_user(struct ib_uverbs_qp_attr *dst, struct ib_qp_attr *src) ib_copy_qp_attr_to_user() argument 94 ib_copy_path_rec_to_user(struct ib_user_path_rec *dst, struct ib_sa_path_rec *src) ib_copy_path_rec_to_user() argument 120 ib_copy_path_rec_from_user(struct ib_sa_path_rec *dst, struct ib_user_path_rec *src) ib_copy_path_rec_from_user() argument
|
/linux-4.1.27/arch/nios2/lib/ |
H A D | memmove.c | 16 unsigned long dst, src; memmove() local 23 src = (unsigned long) s; memmove() 25 if ((count < 8) || ((dst ^ src) & 3)) memmove() 29 *(char *)dst++ = *(char *)src++; memmove() 33 *(short *)dst = *(short *)src; memmove() 34 src += 2; memmove() 39 *(long *)dst = *(long *)src; memmove() 40 src += 4; memmove() 46 *(char *)dst++ = *(char *)src++; memmove() 49 src = (unsigned long) s + count; memmove() 51 if ((count < 8) || ((dst ^ src) & 3)) memmove() 55 src--; memmove() 58 *(char *)dst = *(char *)src; memmove() 61 src -= 2; memmove() 64 *(short *)dst = *(short *)src; memmove() 67 src -= 4; memmove() 70 *(long *)dst = *(long *)src; memmove() 74 src--; memmove() 76 *(char *)dst = *(char *)src; memmove()
|
/linux-4.1.27/arch/alpha/lib/ |
H A D | srm_printk.c | 14 char *src, *dst; srm_printk() local 23 for (src = buf; *src; ++src) { srm_printk() 24 if (*src == '\n') { srm_printk() 31 for (dst = src + num_lf; src >= buf; ) { srm_printk() 32 if (*src == '\n') { srm_printk() 35 *dst-- = *src--; srm_printk()
|
H A D | csum_partial_copy.c | 102 csum_partial_cfu_aligned(const unsigned long __user *src, unsigned long *dst, csum_partial_cfu_aligned() argument 111 err |= __get_user(word, src); csum_partial_cfu_aligned() 113 src++; csum_partial_cfu_aligned() 124 err |= __get_user(word, src); csum_partial_cfu_aligned() 142 csum_partial_cfu_dest_aligned(const unsigned long __user *src, csum_partial_cfu_dest_aligned() argument 150 unsigned long lastsrc = 7+len+(unsigned long)src; csum_partial_cfu_dest_aligned() 153 err |= __get_user_u(first,src); csum_partial_cfu_dest_aligned() 158 err |= __get_user_u(second, src+1); csum_partial_cfu_dest_aligned() 161 src++; csum_partial_cfu_dest_aligned() 196 csum_partial_cfu_src_aligned(const unsigned long __user *src, csum_partial_cfu_src_aligned() argument 210 err |= __get_user(word, src); csum_partial_cfu_src_aligned() 215 src++; csum_partial_cfu_src_aligned() 224 err |= __get_user(word, src); csum_partial_cfu_src_aligned() 254 csum_partial_cfu_unaligned(const unsigned long __user * src, csum_partial_cfu_unaligned() argument 266 err |= __get_user_u(first, src); csum_partial_cfu_unaligned() 267 lastsrc = 7+len+(unsigned long)src; csum_partial_cfu_unaligned() 273 err |= __get_user_u(second, src+1); csum_partial_cfu_unaligned() 278 src++; csum_partial_cfu_unaligned() 333 csum_partial_copy_from_user(const void __user *src, void *dst, int len, csum_partial_copy_from_user() argument 337 unsigned long soff = 7 & (unsigned long) src; csum_partial_copy_from_user() 341 if (!access_ok(VERIFY_READ, src, len)) { csum_partial_copy_from_user() 349 (const unsigned long __user *) src, csum_partial_copy_from_user() 354 (const unsigned long __user *) src, csum_partial_copy_from_user() 362 (const unsigned long __user *) src, csum_partial_copy_from_user() 368 (const unsigned long __user *) src, csum_partial_copy_from_user() 379 csum_partial_copy_nocheck(const void *src, void *dst, int len, __wsum sum) csum_partial_copy_nocheck() argument 384 checksum = csum_partial_copy_from_user((__force const void __user *)src, csum_partial_copy_nocheck()
|
/linux-4.1.27/sound/pci/ctxfi/ |
H A D | ctsrc.c | 28 static int src_default_config_memrd(struct src *src); 29 static int src_default_config_memwr(struct src *src); 30 static int src_default_config_arcrw(struct src *src); 32 static int (*src_default_config[3])(struct src *) = { 38 static int src_set_state(struct src *src, unsigned int state) src_set_state() argument 42 hw = src->rsc.hw; src_set_state() 43 hw->src_set_state(src->rsc.ctrl_blk, state); src_set_state() 48 static int src_set_bm(struct src *src, unsigned int bm) src_set_bm() argument 52 hw = src->rsc.hw; src_set_bm() 53 hw->src_set_bm(src->rsc.ctrl_blk, bm); src_set_bm() 58 static int src_set_sf(struct src *src, unsigned int sf) src_set_sf() argument 62 hw = src->rsc.hw; src_set_sf() 63 hw->src_set_sf(src->rsc.ctrl_blk, sf); src_set_sf() 68 static int src_set_pm(struct src *src, unsigned int pm) src_set_pm() argument 72 hw = src->rsc.hw; src_set_pm() 73 hw->src_set_pm(src->rsc.ctrl_blk, pm); src_set_pm() 78 static int src_set_rom(struct src *src, unsigned int rom) src_set_rom() argument 82 hw = src->rsc.hw; src_set_rom() 83 hw->src_set_rom(src->rsc.ctrl_blk, rom); src_set_rom() 88 static int src_set_vo(struct src *src, unsigned int vo) src_set_vo() argument 92 hw = src->rsc.hw; src_set_vo() 93 hw->src_set_vo(src->rsc.ctrl_blk, vo); src_set_vo() 98 static int src_set_st(struct src *src, unsigned int st) src_set_st() argument 102 hw = src->rsc.hw; src_set_st() 103 hw->src_set_st(src->rsc.ctrl_blk, st); src_set_st() 108 static int src_set_bp(struct src *src, unsigned int bp) src_set_bp() argument 112 hw = src->rsc.hw; src_set_bp() 113 hw->src_set_bp(src->rsc.ctrl_blk, bp); src_set_bp() 118 static int src_set_cisz(struct src *src, unsigned int cisz) src_set_cisz() argument 122 hw = src->rsc.hw; src_set_cisz() 123 hw->src_set_cisz(src->rsc.ctrl_blk, cisz); src_set_cisz() 128 static int src_set_ca(struct src *src, unsigned int ca) src_set_ca() argument 132 hw = src->rsc.hw; src_set_ca() 133 hw->src_set_ca(src->rsc.ctrl_blk, ca); src_set_ca() 138 static int src_set_sa(struct src *src, unsigned int sa) src_set_sa() argument 142 hw = src->rsc.hw; src_set_sa() 143 hw->src_set_sa(src->rsc.ctrl_blk, sa); src_set_sa() 148 static int src_set_la(struct src *src, unsigned int la) src_set_la() argument 152 hw = src->rsc.hw; src_set_la() 153 hw->src_set_la(src->rsc.ctrl_blk, la); src_set_la() 158 static int src_set_pitch(struct src *src, unsigned int pitch) src_set_pitch() argument 162 hw = src->rsc.hw; src_set_pitch() 163 hw->src_set_pitch(src->rsc.ctrl_blk, pitch); src_set_pitch() 168 static int src_set_clear_zbufs(struct src *src) src_set_clear_zbufs() argument 172 hw = src->rsc.hw; src_set_clear_zbufs() 173 hw->src_set_clear_zbufs(src->rsc.ctrl_blk, 1); src_set_clear_zbufs() 178 static int src_commit_write(struct src *src) src_commit_write() argument 184 hw = src->rsc.hw; src_commit_write() 185 src->rsc.ops->master(&src->rsc); src_commit_write() 186 if (src->rsc.msr > 1) { src_commit_write() 188 dirty = hw->src_get_dirty(src->rsc.ctrl_blk) & conj_mask; src_commit_write() 190 hw->src_commit_write(hw, src->rsc.ops->index(&src->rsc), src_commit_write() 191 src->rsc.ctrl_blk); src_commit_write() 194 if (MEMWR == src->mode) src_commit_write() 197 for (i = 1; i < src->rsc.msr; i++) { src_commit_write() 198 src->rsc.ops->next_conj(&src->rsc); src_commit_write() 199 hw->src_set_dirty(src->rsc.ctrl_blk, dirty); src_commit_write() 200 hw->src_commit_write(hw, src->rsc.ops->index(&src->rsc), src_commit_write() 201 src->rsc.ctrl_blk); src_commit_write() 203 src->rsc.ops->master(&src->rsc); src_commit_write() 208 static int src_get_ca(struct src *src) src_get_ca() argument 212 hw = src->rsc.hw; src_get_ca() 213 return hw->src_get_ca(hw, src->rsc.ops->index(&src->rsc), src_get_ca() 214 src->rsc.ctrl_blk); src_get_ca() 217 static int src_init(struct src *src) src_init() argument 219 src_default_config[src->mode](src); src_init() 224 static struct src *src_next_interleave(struct src *src) src_next_interleave() argument 226 return src->intlv; src_next_interleave() 229 static int src_default_config_memrd(struct src *src) src_default_config_memrd() argument 231 struct hw *hw = src->rsc.hw; src_default_config_memrd() 234 hw->src_set_state(src->rsc.ctrl_blk, SRC_STATE_OFF); src_default_config_memrd() 235 hw->src_set_bm(src->rsc.ctrl_blk, 1); src_default_config_memrd() 236 for (rsr = 0, msr = src->rsc.msr; msr > 1; msr >>= 1) src_default_config_memrd() 239 hw->src_set_rsr(src->rsc.ctrl_blk, rsr); src_default_config_memrd() 240 hw->src_set_sf(src->rsc.ctrl_blk, SRC_SF_S16); src_default_config_memrd() 241 hw->src_set_wr(src->rsc.ctrl_blk, 0); src_default_config_memrd() 242 hw->src_set_pm(src->rsc.ctrl_blk, 0); src_default_config_memrd() 243 hw->src_set_rom(src->rsc.ctrl_blk, 0); src_default_config_memrd() 244 hw->src_set_vo(src->rsc.ctrl_blk, 0); src_default_config_memrd() 245 hw->src_set_st(src->rsc.ctrl_blk, 0); src_default_config_memrd() 246 hw->src_set_ilsz(src->rsc.ctrl_blk, src->multi - 1); src_default_config_memrd() 247 hw->src_set_cisz(src->rsc.ctrl_blk, 0x80); src_default_config_memrd() 248 hw->src_set_sa(src->rsc.ctrl_blk, 0x0); src_default_config_memrd() 249 hw->src_set_la(src->rsc.ctrl_blk, 0x1000); src_default_config_memrd() 250 hw->src_set_ca(src->rsc.ctrl_blk, 0x80); src_default_config_memrd() 251 hw->src_set_pitch(src->rsc.ctrl_blk, 0x1000000); src_default_config_memrd() 252 hw->src_set_clear_zbufs(src->rsc.ctrl_blk, 1); src_default_config_memrd() 254 src->rsc.ops->master(&src->rsc); src_default_config_memrd() 255 hw->src_commit_write(hw, src->rsc.ops->index(&src->rsc), src_default_config_memrd() 256 src->rsc.ctrl_blk); src_default_config_memrd() 258 for (msr = 1; msr < src->rsc.msr; msr++) { src_default_config_memrd() 259 src->rsc.ops->next_conj(&src->rsc); src_default_config_memrd() 260 hw->src_set_pitch(src->rsc.ctrl_blk, 0x1000000); src_default_config_memrd() 261 hw->src_commit_write(hw, src->rsc.ops->index(&src->rsc), src_default_config_memrd() 262 src->rsc.ctrl_blk); src_default_config_memrd() 264 src->rsc.ops->master(&src->rsc); src_default_config_memrd() 269 static int src_default_config_memwr(struct src *src) src_default_config_memwr() argument 271 struct hw *hw = src->rsc.hw; src_default_config_memwr() 273 hw->src_set_state(src->rsc.ctrl_blk, SRC_STATE_OFF); src_default_config_memwr() 274 hw->src_set_bm(src->rsc.ctrl_blk, 1); src_default_config_memwr() 275 hw->src_set_rsr(src->rsc.ctrl_blk, 0); src_default_config_memwr() 276 hw->src_set_sf(src->rsc.ctrl_blk, SRC_SF_S16); src_default_config_memwr() 277 hw->src_set_wr(src->rsc.ctrl_blk, 1); src_default_config_memwr() 278 hw->src_set_pm(src->rsc.ctrl_blk, 0); src_default_config_memwr() 279 hw->src_set_rom(src->rsc.ctrl_blk, 0); src_default_config_memwr() 280 hw->src_set_vo(src->rsc.ctrl_blk, 0); src_default_config_memwr() 281 hw->src_set_st(src->rsc.ctrl_blk, 0); src_default_config_memwr() 282 hw->src_set_ilsz(src->rsc.ctrl_blk, 0); src_default_config_memwr() 283 hw->src_set_cisz(src->rsc.ctrl_blk, 0x80); src_default_config_memwr() 284 hw->src_set_sa(src->rsc.ctrl_blk, 0x0); src_default_config_memwr() 285 hw->src_set_la(src->rsc.ctrl_blk, 0x1000); src_default_config_memwr() 286 hw->src_set_ca(src->rsc.ctrl_blk, 0x80); src_default_config_memwr() 287 hw->src_set_pitch(src->rsc.ctrl_blk, 0x1000000); src_default_config_memwr() 288 hw->src_set_clear_zbufs(src->rsc.ctrl_blk, 1); src_default_config_memwr() 290 src->rsc.ops->master(&src->rsc); src_default_config_memwr() 291 hw->src_commit_write(hw, src->rsc.ops->index(&src->rsc), src_default_config_memwr() 292 src->rsc.ctrl_blk); src_default_config_memwr() 297 static int src_default_config_arcrw(struct src *src) src_default_config_arcrw() argument 299 struct hw *hw = src->rsc.hw; src_default_config_arcrw() 303 hw->src_set_state(src->rsc.ctrl_blk, SRC_STATE_OFF); src_default_config_arcrw() 304 hw->src_set_bm(src->rsc.ctrl_blk, 0); src_default_config_arcrw() 305 for (rsr = 0, msr = src->rsc.msr; msr > 1; msr >>= 1) src_default_config_arcrw() 308 hw->src_set_rsr(src->rsc.ctrl_blk, rsr); src_default_config_arcrw() 309 hw->src_set_sf(src->rsc.ctrl_blk, SRC_SF_F32); src_default_config_arcrw() 310 hw->src_set_wr(src->rsc.ctrl_blk, 0); src_default_config_arcrw() 311 hw->src_set_pm(src->rsc.ctrl_blk, 0); src_default_config_arcrw() 312 hw->src_set_rom(src->rsc.ctrl_blk, 0); src_default_config_arcrw() 313 hw->src_set_vo(src->rsc.ctrl_blk, 0); src_default_config_arcrw() 314 hw->src_set_st(src->rsc.ctrl_blk, 0); src_default_config_arcrw() 315 hw->src_set_ilsz(src->rsc.ctrl_blk, 0); src_default_config_arcrw() 316 hw->src_set_cisz(src->rsc.ctrl_blk, 0x80); src_default_config_arcrw() 317 hw->src_set_sa(src->rsc.ctrl_blk, 0x0); src_default_config_arcrw() 318 /*hw->src_set_sa(src->rsc.ctrl_blk, 0x100);*/ src_default_config_arcrw() 319 hw->src_set_la(src->rsc.ctrl_blk, 0x1000); src_default_config_arcrw() 320 /*hw->src_set_la(src->rsc.ctrl_blk, 0x03ffffe0);*/ src_default_config_arcrw() 321 hw->src_set_ca(src->rsc.ctrl_blk, 0x80); src_default_config_arcrw() 322 hw->src_set_pitch(src->rsc.ctrl_blk, 0x1000000); src_default_config_arcrw() 323 hw->src_set_clear_zbufs(src->rsc.ctrl_blk, 1); src_default_config_arcrw() 325 dirty = hw->src_get_dirty(src->rsc.ctrl_blk); src_default_config_arcrw() 326 src->rsc.ops->master(&src->rsc); src_default_config_arcrw() 327 for (msr = 0; msr < src->rsc.msr; msr++) { src_default_config_arcrw() 328 hw->src_set_dirty(src->rsc.ctrl_blk, dirty); src_default_config_arcrw() 329 hw->src_commit_write(hw, src->rsc.ops->index(&src->rsc), src_default_config_arcrw() 330 src->rsc.ctrl_blk); src_default_config_arcrw() 331 src->rsc.ops->next_conj(&src->rsc); src_default_config_arcrw() 333 src->rsc.ops->master(&src->rsc); src_default_config_arcrw() 360 src_rsc_init(struct src *src, u32 idx, src_rsc_init() argument 365 struct src *p; src_rsc_init() 368 for (i = 0, p = src; i < n; i++, p++) { src_rsc_init() 373 /* Initialize src specific rsc operations */ src_rsc_init() 396 static int src_rsc_uninit(struct src *src, struct src_mgr *mgr) src_rsc_uninit() argument 399 struct src *p; src_rsc_uninit() 401 n = (MEMRD == src->mode) ? src->multi : 1; src_rsc_uninit() 402 for (i = 0, p = src; i < n; i++, p++) { src_rsc_uninit() 416 get_src_rsc(struct src_mgr *mgr, const struct src_desc *desc, struct src **rsrc) get_src_rsc() 420 struct src *src; get_src_rsc() local 425 /* Check whether there are sufficient src resources to meet request. */ get_src_rsc() 439 /* Allocate mem for master src resource */ get_src_rsc() 441 src = kcalloc(desc->multi, sizeof(*src), GFP_KERNEL); get_src_rsc() 443 src = kzalloc(sizeof(*src), GFP_KERNEL); get_src_rsc() 445 if (!src) { get_src_rsc() 450 err = src_rsc_init(src, idx, desc, mgr); get_src_rsc() 454 *rsrc = src; get_src_rsc() 459 kfree(src); get_src_rsc() 471 static int put_src_rsc(struct src_mgr *mgr, struct src *src) put_src_rsc() argument 476 src->rsc.ops->master(&src->rsc); put_src_rsc() 477 if (MEMRD == src->mode) put_src_rsc() 478 mgr_put_resource(&mgr->mgr, src->multi, put_src_rsc() 479 src->rsc.ops->index(&src->rsc)); put_src_rsc() 481 mgr_put_resource(&mgr->mgr, 1, src->rsc.ops->index(&src->rsc)); put_src_rsc() 484 src_rsc_uninit(src, mgr); put_src_rsc() 485 kfree(src); put_src_rsc() 490 static int src_enable_s(struct src_mgr *mgr, struct src *src) src_enable_s() argument 495 src->rsc.ops->master(&src->rsc); src_enable_s() 496 for (i = 0; i < src->rsc.msr; i++) { src_enable_s() 498 src->rsc.ops->index(&src->rsc)); src_enable_s() 499 src->rsc.ops->next_conj(&src->rsc); src_enable_s() 501 src->rsc.ops->master(&src->rsc); src_enable_s() 506 static int src_enable(struct src_mgr *mgr, struct src *src) src_enable() argument 511 src->rsc.ops->master(&src->rsc); src_enable() 512 for (i = 0; i < src->rsc.msr; i++) { src_enable() 514 src->rsc.ops->index(&src->rsc)); src_enable() 515 src->rsc.ops->next_conj(&src->rsc); src_enable() 517 src->rsc.ops->master(&src->rsc); src_enable() 522 static int src_disable(struct src_mgr *mgr, struct src *src) src_disable() argument 527 src->rsc.ops->master(&src->rsc); src_disable() 528 for (i = 0; i < src->rsc.msr; i++) { src_disable() 530 src->rsc.ops->index(&src->rsc)); src_disable() 531 src->rsc.ops->next_conj(&src->rsc); src_disable() 533 src->rsc.ops->master(&src->rsc); src_disable() 621 static int srcimp_map(struct srcimp *srcimp, struct src *src, struct rsc *input) srcimp_map() argument 627 src->rsc.ops->master(&src->rsc); srcimp_map() 634 entry->user = src->rsc.ops->index(&src->rsc); srcimp_map() 880 /* free src input mapper list */ srcimp_mgr_destroy()
|
H A D | ctsrc.h | 38 /* Define the descriptor of a src resource */ 48 struct src { struct 50 struct src *intlv; /* Pointer to next interleaved SRC in a series */ 58 int (*set_state)(struct src *src, unsigned int state); 59 int (*set_bm)(struct src *src, unsigned int bm); 60 int (*set_sf)(struct src *src, unsigned int sf); 61 int (*set_pm)(struct src *src, unsigned int pm); 62 int (*set_rom)(struct src *src, unsigned int rom); 63 int (*set_vo)(struct src *src, unsigned int vo); 64 int (*set_st)(struct src *src, unsigned int st); 65 int (*set_bp)(struct src *src, unsigned int bp); 66 int (*set_cisz)(struct src *src, unsigned int cisz); 67 int (*set_ca)(struct src *src, unsigned int ca); 68 int (*set_sa)(struct src *src, unsigned int sa); 69 int (*set_la)(struct src *src, unsigned int la); 70 int (*set_pitch)(struct src *src, unsigned int pitch); 71 int (*set_clr_zbufs)(struct src *src); 72 int (*commit_write)(struct src *src); 73 int (*get_ca)(struct src *src); 74 int (*init)(struct src *src); 75 struct src* (*next_interleave)(struct src *src); 78 /* Define src resource request description info */ 86 /* Define src manager object */ 92 /* request src resource */ 94 const struct src_desc *desc, struct src **rsrc); 95 /* return src resource */ 96 int (*put_src)(struct src_mgr *mgr, struct src *src); 97 int (*src_enable_s)(struct src_mgr *mgr, struct src *src); 98 int (*src_enable)(struct src_mgr *mgr, struct src *src); 99 int (*src_disable)(struct src_mgr *mgr, struct src *src); 117 int (*map)(struct srcimp *srcimp, struct src *user, struct rsc *input);
|
H A D | ctatc.c | 249 struct src *src; atc_pcm_playback_prepare() local 263 err = src_mgr->get_src(src_mgr, &desc, (struct src **)&apcm->src); atc_pcm_playback_prepare() 269 src = apcm->src; atc_pcm_playback_prepare() 270 src->ops->set_pitch(src, pitch); atc_pcm_playback_prepare() 271 src->ops->set_rom(src, select_rom(pitch)); atc_pcm_playback_prepare() 272 src->ops->set_sf(src, convert_format(apcm->substream->runtime->format, atc_pcm_playback_prepare() 274 src->ops->set_pm(src, (src->ops->next_interleave(src) != NULL)); atc_pcm_playback_prepare() 299 src = apcm->src; atc_pcm_playback_prepare() 303 amixer->ops->setup(amixer, &src->rsc, atc_pcm_playback_prepare() 306 src = src->ops->next_interleave(src); atc_pcm_playback_prepare() 307 if (!src) atc_pcm_playback_prepare() 308 src = apcm->src; atc_pcm_playback_prepare() 364 if (apcm->src) { atc_pcm_release_resources() 365 src_mgr->put_src(src_mgr, apcm->src); atc_pcm_release_resources() 366 apcm->src = NULL; atc_pcm_release_resources() 381 struct src *src = apcm->src; atc_pcm_playback_start() local 387 max_cisz = src->multi * src->rsc.msr; atc_pcm_playback_start() 390 src->ops->set_sa(src, apcm->vm_block->addr); atc_pcm_playback_start() 391 src->ops->set_la(src, apcm->vm_block->addr + apcm->vm_block->size); atc_pcm_playback_start() 392 src->ops->set_ca(src, apcm->vm_block->addr + max_cisz); atc_pcm_playback_start() 393 src->ops->set_cisz(src, max_cisz); atc_pcm_playback_start() 395 src->ops->set_bm(src, 1); atc_pcm_playback_start() 396 src->ops->set_state(src, SRC_STATE_INIT); atc_pcm_playback_start() 397 src->ops->commit_write(src); atc_pcm_playback_start() 405 struct src *src; atc_pcm_stop() local 410 src = apcm->src; atc_pcm_stop() 411 src->ops->set_bm(src, 0); atc_pcm_stop() 412 src->ops->set_state(src, SRC_STATE_OFF); atc_pcm_stop() 413 src->ops->commit_write(src); atc_pcm_stop() 417 src = apcm->srccs[i]; atc_pcm_stop() 418 src->ops->set_bm(src, 0); atc_pcm_stop() 419 src->ops->set_state(src, SRC_STATE_OFF); atc_pcm_stop() 420 src->ops->commit_write(src); atc_pcm_stop() 432 struct src *src = apcm->src; atc_pcm_playback_position() local 436 if (!src) atc_pcm_playback_position() 438 position = src->ops->get_ca(src); atc_pcm_playback_position() 448 max_cisz = src->multi * src->rsc.msr; atc_pcm_playback_position() 510 struct src *src; atc_pcm_capture_get_resources() local 569 (struct src **)&apcm->srccs[i]); atc_pcm_capture_get_resources() 573 src = apcm->srccs[i]; atc_pcm_capture_get_resources() 575 src->ops->set_pitch(src, pitch); atc_pcm_capture_get_resources() 576 src->ops->set_rom(src, select_rom(pitch)); atc_pcm_capture_get_resources() 577 src->ops->set_vo(src, src_node_conf[i/multi].vo); atc_pcm_capture_get_resources() 628 err = src_mgr->get_src(src_mgr, &src_dsc, (struct src **)&apcm->src); atc_pcm_capture_get_resources() 632 src = apcm->src; atc_pcm_capture_get_resources() 633 src->ops->set_pitch(src, pitch); atc_pcm_capture_get_resources() 649 struct src *src; atc_pcm_capture_prepare() local 684 src = apcm->srccs[i]; atc_pcm_capture_prepare() 687 srcimp->ops->map(srcimp, src, out_ports[i%multi]); atc_pcm_capture_prepare() 688 amixer->ops->setup(amixer, &src->rsc, INIT_VOL, NULL); atc_pcm_capture_prepare() 709 srcimp->ops->map(srcimp, apcm->src, atc_pcm_capture_prepare() 716 srcimp->ops->map(srcimp, apcm->src, out_ports[i]); atc_pcm_capture_prepare() 727 struct src *src; atc_pcm_capture_start() local 738 src = apcm->srccs[i]; atc_pcm_capture_start() 739 src->ops->set_pm(src, ((i%multi) != (multi-1))); atc_pcm_capture_start() 740 src_mgr->src_disable(src_mgr, src); atc_pcm_capture_start() 744 src = apcm->src; atc_pcm_capture_start() 745 src->ops->set_sf(src, convert_format(apcm->substream->runtime->format, atc_pcm_capture_start() 747 src->ops->set_sa(src, apcm->vm_block->addr); atc_pcm_capture_start() 748 src->ops->set_la(src, apcm->vm_block->addr + apcm->vm_block->size); atc_pcm_capture_start() 749 src->ops->set_ca(src, apcm->vm_block->addr); atc_pcm_capture_start() 750 src_mgr->src_disable(src_mgr, src); atc_pcm_capture_start() 757 src = apcm->srccs[i]; atc_pcm_capture_start() 758 src->ops->set_state(src, SRC_STATE_RUN); atc_pcm_capture_start() 759 src->ops->commit_write(src); atc_pcm_capture_start() 760 src_mgr->src_enable_s(src_mgr, src); atc_pcm_capture_start() 762 src = apcm->src; atc_pcm_capture_start() 763 src->ops->set_bm(src, 1); atc_pcm_capture_start() 764 src->ops->set_state(src, SRC_STATE_RUN); atc_pcm_capture_start() 765 src->ops->commit_write(src); atc_pcm_capture_start() 766 src_mgr->src_enable_s(src_mgr, src); atc_pcm_capture_start() 778 struct src *src = apcm->src; atc_pcm_capture_position() local 780 if (!src) atc_pcm_capture_position() 782 return src->ops->get_ca(src) - apcm->vm_block->addr; atc_pcm_capture_position() 792 struct src *src; spdif_passthru_playback_get_resources() local 807 err = src_mgr->get_src(src_mgr, &desc, (struct src **)&apcm->src); spdif_passthru_playback_get_resources() 812 src = apcm->src; spdif_passthru_playback_get_resources() 813 src->ops->set_pitch(src, pitch); spdif_passthru_playback_get_resources() 814 src->ops->set_rom(src, select_rom(pitch)); spdif_passthru_playback_get_resources() 815 src->ops->set_sf(src, convert_format(apcm->substream->runtime->format, spdif_passthru_playback_get_resources() 817 src->ops->set_pm(src, (src->ops->next_interleave(src) != NULL)); spdif_passthru_playback_get_resources() 818 src->ops->set_bp(src, 1); spdif_passthru_playback_get_resources() 899 struct src *src; spdif_passthru_playback_prepare() local 919 src = apcm->src; spdif_passthru_playback_prepare() 922 amixer->ops->setup(amixer, &src->rsc, INIT_VOL, NULL); spdif_passthru_playback_prepare() 923 src = src->ops->next_interleave(src); spdif_passthru_playback_prepare() 924 if (!src) spdif_passthru_playback_prepare() 925 src = apcm->src; spdif_passthru_playback_prepare() 945 struct src *src; atc_select_line_in() local 955 src = atc->srcs[2]; atc_select_line_in() 956 mixer->set_input_left(mixer, MIX_LINE_IN, &src->rsc); atc_select_line_in() 957 src = atc->srcs[3]; atc_select_line_in() 958 mixer->set_input_right(mixer, MIX_LINE_IN, &src->rsc); atc_select_line_in() 967 struct src *src; atc_select_mic_in() local 977 src = atc->srcs[2]; atc_select_mic_in() 978 mixer->set_input_left(mixer, MIX_MIC_IN, &src->rsc); atc_select_mic_in() 979 src = atc->srcs[3]; atc_select_mic_in() 980 mixer->set_input_right(mixer, MIX_MIC_IN, &src->rsc); atc_select_mic_in() 1419 (struct src **)&atc->srcs[i]); atc_get_resources() 1453 struct src **srcs, struct srcimp **srcimps) atc_connect_dai() 1456 struct src *src; atc_connect_dai() local 1463 src = srcs[i]; atc_connect_dai() 1465 srcimp->ops->map(srcimp, src, rscs[i]); atc_connect_dai() 1466 src_mgr->src_disable(src_mgr, src); atc_connect_dai() 1471 src = srcs[0]; atc_connect_dai() 1472 src->ops->set_pm(src, 1); atc_connect_dai() 1474 src = srcs[i]; atc_connect_dai() 1475 src->ops->set_state(src, SRC_STATE_RUN); atc_connect_dai() 1476 src->ops->commit_write(src); atc_connect_dai() 1477 src_mgr->src_enable_s(src_mgr, src); atc_connect_dai() 1494 struct src *src; atc_connect_resources() local 1511 (struct src **)&atc->srcs[2], atc_connect_resources() 1513 src = atc->srcs[2]; atc_connect_resources() 1514 mixer->set_input_left(mixer, MIX_LINE_IN, &src->rsc); atc_connect_resources() 1515 src = atc->srcs[3]; atc_connect_resources() 1516 mixer->set_input_right(mixer, MIX_LINE_IN, &src->rsc); atc_connect_resources() 1522 (struct src **)&atc->srcs[4], atc_connect_resources() 1524 src = atc->srcs[4]; atc_connect_resources() 1525 mixer->set_input_left(mixer, MIX_MIC_IN, &src->rsc); atc_connect_resources() 1526 src = atc->srcs[5]; atc_connect_resources() 1527 mixer->set_input_right(mixer, MIX_MIC_IN, &src->rsc); atc_connect_resources() 1532 (struct src **)&atc->srcs[0], atc_connect_resources() 1535 src = atc->srcs[0]; atc_connect_resources() 1536 mixer->set_input_left(mixer, MIX_SPDIF_IN, &src->rsc); atc_connect_resources() 1537 src = atc->srcs[1]; atc_connect_resources() 1538 mixer->set_input_right(mixer, MIX_SPDIF_IN, &src->rsc); atc_connect_resources()
|
/linux-4.1.27/tools/perf/util/include/asm/ |
H A D | uaccess.h | 4 #define __get_user(src, dest) \ 6 (src) = *dest; \
|
/linux-4.1.27/arch/m68k/lib/ |
H A D | memmove.c | 10 void *memmove(void *dest, const void *src, size_t n) memmove() argument 18 if (dest < src) { memmove() 21 const char *csrc = src; memmove() 24 src = csrc; memmove() 29 const short *ssrc = src; memmove() 32 src = ssrc; memmove() 38 const long *lsrc = src; memmove() 44 src = lsrc; memmove() 48 const short *ssrc = src; memmove() 51 src = ssrc; memmove() 55 const char *csrc = src; memmove() 60 src = (const char *)src + n; memmove() 63 const char *csrc = src; memmove() 66 src = csrc; memmove() 71 const short *ssrc = src; memmove() 74 src = ssrc; memmove() 80 const long *lsrc = src; memmove() 86 src = lsrc; memmove() 90 const short *ssrc = src; memmove() 93 src = ssrc; memmove() 97 const char *csrc = src; memmove()
|
/linux-4.1.27/arch/m68k/math-emu/ |
H A D | fp_trig.c | 22 fp_fsin(struct fp_ext *dest, struct fp_ext *src) fp_fsin() argument 26 fp_monadic_check(dest, src); fp_fsin() 32 fp_fcos(struct fp_ext *dest, struct fp_ext *src) fp_fcos() argument 36 fp_monadic_check(dest, src); fp_fcos() 42 fp_ftan(struct fp_ext *dest, struct fp_ext *src) fp_ftan() argument 46 fp_monadic_check(dest, src); fp_ftan() 52 fp_fasin(struct fp_ext *dest, struct fp_ext *src) fp_fasin() argument 56 fp_monadic_check(dest, src); fp_fasin() 62 fp_facos(struct fp_ext *dest, struct fp_ext *src) fp_facos() argument 66 fp_monadic_check(dest, src); fp_facos() 72 fp_fatan(struct fp_ext *dest, struct fp_ext *src) fp_fatan() argument 76 fp_monadic_check(dest, src); fp_fatan() 82 fp_fsinh(struct fp_ext *dest, struct fp_ext *src) fp_fsinh() argument 86 fp_monadic_check(dest, src); fp_fsinh() 92 fp_fcosh(struct fp_ext *dest, struct fp_ext *src) fp_fcosh() argument 96 fp_monadic_check(dest, src); fp_fcosh() 102 fp_ftanh(struct fp_ext *dest, struct fp_ext *src) fp_ftanh() argument 106 fp_monadic_check(dest, src); fp_ftanh() 112 fp_fatanh(struct fp_ext *dest, struct fp_ext *src) fp_fatanh() argument 116 fp_monadic_check(dest, src); fp_fatanh() 122 fp_fsincos0(struct fp_ext *dest, struct fp_ext *src) fp_fsincos0() argument 130 fp_fsincos1(struct fp_ext *dest, struct fp_ext *src) fp_fsincos1() argument 138 fp_fsincos2(struct fp_ext *dest, struct fp_ext *src) fp_fsincos2() argument 146 fp_fsincos3(struct fp_ext *dest, struct fp_ext *src) fp_fsincos3() argument 154 fp_fsincos4(struct fp_ext *dest, struct fp_ext *src) fp_fsincos4() argument 162 fp_fsincos5(struct fp_ext *dest, struct fp_ext *src) fp_fsincos5() argument 170 fp_fsincos6(struct fp_ext *dest, struct fp_ext *src) fp_fsincos6() argument 178 fp_fsincos7(struct fp_ext *dest, struct fp_ext *src) fp_fsincos7() argument
|
H A D | fp_log.c | 25 extern struct fp_ext *fp_fadd(struct fp_ext *dest, const struct fp_ext *src); 26 extern struct fp_ext *fp_fdiv(struct fp_ext *dest, const struct fp_ext *src); 29 fp_fsqrt(struct fp_ext *dest, struct fp_ext *src) fp_fsqrt() argument 36 fp_monadic_check(dest, src); fp_fsqrt() 102 fp_fetoxm1(struct fp_ext *dest, struct fp_ext *src) fp_fetoxm1() argument 106 fp_monadic_check(dest, src); fp_fetoxm1() 112 fp_fetox(struct fp_ext *dest, struct fp_ext *src) fp_fetox() argument 116 fp_monadic_check(dest, src); fp_fetox() 122 fp_ftwotox(struct fp_ext *dest, struct fp_ext *src) fp_ftwotox() argument 126 fp_monadic_check(dest, src); fp_ftwotox() 132 fp_ftentox(struct fp_ext *dest, struct fp_ext *src) fp_ftentox() argument 136 fp_monadic_check(dest, src); fp_ftentox() 142 fp_flogn(struct fp_ext *dest, struct fp_ext *src) fp_flogn() argument 146 fp_monadic_check(dest, src); fp_flogn() 152 fp_flognp1(struct fp_ext *dest, struct fp_ext *src) fp_flognp1() argument 156 fp_monadic_check(dest, src); fp_flognp1() 162 fp_flog10(struct fp_ext *dest, struct fp_ext *src) fp_flog10() argument 166 fp_monadic_check(dest, src); fp_flog10() 172 fp_flog2(struct fp_ext *dest, struct fp_ext *src) fp_flog2() argument 176 fp_monadic_check(dest, src); fp_flog2() 182 fp_fgetexp(struct fp_ext *dest, struct fp_ext *src) fp_fgetexp() argument 186 fp_monadic_check(dest, src); fp_fgetexp() 203 fp_fgetman(struct fp_ext *dest, struct fp_ext *src) fp_fgetman() argument 207 fp_monadic_check(dest, src); fp_fgetman()
|
H A D | fp_arith.h | 22 fp_fabs(struct fp_ext *dest, struct fp_ext *src); 24 fp_fneg(struct fp_ext *dest, struct fp_ext *src); 28 fp_fadd(struct fp_ext *dest, struct fp_ext *src); 30 fp_fsub(struct fp_ext *dest, struct fp_ext *src); 32 fp_fcmp(struct fp_ext *dest, struct fp_ext *src); 34 fp_ftst(struct fp_ext *dest, struct fp_ext *src); 36 fp_fmul(struct fp_ext *dest, struct fp_ext *src); 38 fp_fdiv(struct fp_ext *dest, struct fp_ext *src); 42 fp_fmod(struct fp_ext *dest, struct fp_ext *src); 44 fp_frem(struct fp_ext *dest, struct fp_ext *src); 46 fp_fint(struct fp_ext *dest, struct fp_ext *src); 48 fp_fintrz(struct fp_ext *dest, struct fp_ext *src); 50 fp_fscale(struct fp_ext *dest, struct fp_ext *src);
|
H A D | fp_arith.c | 34 fp_fabs(struct fp_ext *dest, struct fp_ext *src) fp_fabs() argument 38 fp_monadic_check(dest, src); fp_fabs() 46 fp_fneg(struct fp_ext *dest, struct fp_ext *src) fp_fneg() argument 50 fp_monadic_check(dest, src); fp_fneg() 63 fp_fadd(struct fp_ext *dest, struct fp_ext *src) fp_fadd() argument 69 fp_dyadic_check(dest, src); fp_fadd() 73 if (IS_INF(src) && (src->sign != dest->sign)) fp_fadd() 77 if (IS_INF(src)) { fp_fadd() 78 fp_copy_ext(dest, src); fp_fadd() 83 if (IS_ZERO(src)) { fp_fadd() 84 if (src->sign != dest->sign) { fp_fadd() 91 fp_copy_ext(dest, src); fp_fadd() 95 dest->lowmant = src->lowmant = 0; fp_fadd() 97 if ((diff = dest->exp - src->exp) > 0) fp_fadd() 98 fp_denormalize(src, diff); fp_fadd() 102 if (dest->sign == src->sign) { fp_fadd() 103 if (fp_addmant(dest, src)) fp_fadd() 107 if (dest->mant.m64 < src->mant.m64) { fp_fadd() 108 fp_submant(dest, src, dest); fp_fadd() 111 fp_submant(dest, dest, src); fp_fadd() 123 fp_fsub(struct fp_ext *dest, struct fp_ext *src) fp_fsub() argument 127 src->sign = !src->sign; fp_fsub() 128 return fp_fadd(dest, src); fp_fsub() 133 fp_fcmp(struct fp_ext *dest, struct fp_ext *src) fp_fcmp() argument 138 src->sign = !src->sign; fp_fcmp() 139 return fp_fadd(&FPDATA->temp[1], src); fp_fcmp() 143 fp_ftst(struct fp_ext *dest, struct fp_ext *src) fp_ftst() argument 149 return src; fp_ftst() 153 fp_fmul(struct fp_ext *dest, struct fp_ext *src) fp_fmul() argument 160 fp_dyadic_check(dest, src); fp_fmul() 163 dest->sign = src->sign ^ dest->sign; fp_fmul() 167 if (IS_ZERO(src)) fp_fmul() 171 if (IS_INF(src)) { fp_fmul() 175 fp_copy_ext(dest, src); fp_fmul() 182 if (IS_ZERO(dest) || IS_ZERO(src)) { fp_fmul() 190 exp = dest->exp + src->exp - 0x3ffe; fp_fmul() 197 if ((long)src->mant.m32[0] >= 0) fp_fmul() 198 exp -= fp_overnormalize(src); fp_fmul() 201 fp_multiplymant(&temp, dest, src); fp_fmul() 228 of src / dest, the result is actually dest / src. */ 231 fp_fdiv(struct fp_ext *dest, struct fp_ext *src) fp_fdiv() argument 238 fp_dyadic_check(dest, src); fp_fdiv() 241 dest->sign = src->sign ^ dest->sign; fp_fdiv() 246 if (IS_INF(src)) fp_fdiv() 251 if (IS_INF(src)) { fp_fdiv() 263 if (IS_ZERO(src)) fp_fdiv() 268 if (IS_ZERO(src)) { fp_fdiv() 277 exp = dest->exp - src->exp + 0x3fff; fp_fdiv() 284 if ((long)src->mant.m32[0] >= 0) fp_fdiv() 285 exp -= fp_overnormalize(src); fp_fdiv() 288 fp_dividemant(&temp, dest, src); fp_fdiv() 312 fp_fsglmul(struct fp_ext *dest, struct fp_ext *src) fp_fsglmul() argument 318 fp_dyadic_check(dest, src); fp_fsglmul() 321 dest->sign = src->sign ^ dest->sign; fp_fsglmul() 325 if (IS_ZERO(src)) fp_fsglmul() 329 if (IS_INF(src)) { fp_fsglmul() 333 fp_copy_ext(dest, src); fp_fsglmul() 340 if (IS_ZERO(dest) || IS_ZERO(src)) { fp_fsglmul() 348 exp = dest->exp + src->exp - 0x3ffe; fp_fsglmul() 353 src->mant.m32[0] & 0xffffff00); fp_fsglmul() 369 fp_fsgldiv(struct fp_ext *dest, struct fp_ext *src) fp_fsgldiv() argument 376 fp_dyadic_check(dest, src); fp_fsgldiv() 379 dest->sign = src->sign ^ dest->sign; fp_fsgldiv() 384 if (IS_INF(src)) fp_fsgldiv() 389 if (IS_INF(src)) { fp_fsgldiv() 401 if (IS_ZERO(src)) fp_fsgldiv() 406 if (IS_ZERO(src)) { fp_fsgldiv() 415 exp = dest->exp - src->exp + 0x3fff; fp_fsgldiv() 418 src->mant.m32[0] &= 0xffffff00; fp_fsgldiv() 421 if (dest->mant.m32[0] >= src->mant.m32[0]) { fp_fsgldiv() 422 fp_sub64(dest->mant, src->mant); fp_fsgldiv() 423 fp_div64(quot, rem, dest->mant.m32[0], 0, src->mant.m32[0]); fp_fsgldiv() 427 fp_div64(quot, rem, dest->mant.m32[0], 0, src->mant.m32[0]); fp_fsgldiv() 579 modrem_kernel(struct fp_ext *dest, struct fp_ext *src, int mode) modrem_kernel() argument 583 fp_dyadic_check(dest, src); modrem_kernel() 586 if (IS_INF(dest) || IS_ZERO(src)) { modrem_kernel() 590 if (IS_ZERO(dest) || IS_INF(src)) modrem_kernel() 595 fp_fdiv(&tmp, src); /* NOTE: src might be modified */ modrem_kernel() 597 fp_fmul(&tmp, src); modrem_kernel() 610 fmod(src,dest) = (dest - (src * floor(dest / src))) */ 613 fp_fmod(struct fp_ext *dest, struct fp_ext *src) fp_fmod() argument 616 return modrem_kernel(dest, src, FPCR_ROUND_RZ); fp_fmod() 621 frem(src,dest) = (dest - (src * round(dest / src))) 625 fp_frem(struct fp_ext *dest, struct fp_ext *src) fp_frem() argument 628 return modrem_kernel(dest, src, FPCR_ROUND_RN); fp_frem() 632 fp_fint(struct fp_ext *dest, struct fp_ext *src) fp_fint() argument 636 fp_copy_ext(dest, src); fp_fint() 644 fp_fintrz(struct fp_ext *dest, struct fp_ext *src) fp_fintrz() argument 648 fp_copy_ext(dest, src); fp_fintrz() 656 fp_fscale(struct fp_ext *dest, struct fp_ext *src) fp_fscale() argument 662 fp_dyadic_check(dest, src); fp_fscale() 665 if (IS_INF(src)) { fp_fscale() 673 if (IS_ZERO(src) || IS_ZERO(dest)) fp_fscale() 677 if (src->exp >= 0x400c) { fp_fscale() 682 /* src must be rounded with round to zero. */ fp_fscale() 685 scale = fp_conv_ext2long(src); fp_fscale()
|
H A D | multi_arith.h | 82 static inline int fp_addmant(struct fp_ext *dest, struct fp_ext *src) fp_addmant() argument 88 : "g,d" (src->lowmant), "0,0" (dest->lowmant)); fp_addmant() 90 : "d" (src->mant.m32[1]), "0" (dest->mant.m32[1])); fp_addmant() 92 : "d" (src->mant.m32[0]), "0" (dest->mant.m32[0])); fp_addmant() 140 #define fp_addx96(dest, src) ({ \ 149 #define fp_sub64(dest, src) ({ \ 151 : "dm,d" (src.m32[1]), "0,0" (dest.m32[1])); \ 153 : "d" (src.m32[0]), "0" (dest.m32[0])); \ 181 static inline void fp_dividemant(union fp_mant128 *dest, struct fp_ext *src, fp_dividemant() argument 192 if (src->mant.m64 >= div->mant.m64) { fp_dividemant() 193 fp_sub64(src->mant, div->mant); fp_dividemant() 215 if (src->mant.m32[0] == div->mant.m32[0]) { fp_dividemant() 216 fp_div64(first, rem, 0, src->mant.m32[1], div->mant.m32[0]); fp_dividemant() 221 fp_div64(first, rem, src->mant.m32[0], src->mant.m32[1], div->mant.m32[0]); fp_dividemant() 233 src->mant.m32[0] = tmp.m32[1]; fp_dividemant() 234 src->mant.m32[1] = tmp.m32[2]; fp_dividemant() 237 src->mant.m32[0] = tmp.m32[1]; fp_dividemant() 238 src->mant.m32[1] = tmp.m32[2]; fp_dividemant() 244 static inline void fp_putmant128(struct fp_ext *dest, union fp_mant128 *src, fp_putmant128() argument 251 dest->mant.m64 = src->m64[0]; fp_putmant128() 252 dest->lowmant = src->m32[2] >> 24; fp_putmant128() 253 if (src->m32[3] || (src->m32[2] << 8)) fp_putmant128() 258 : "=d" (tmp) : "0" (src->m32[2])); fp_putmant128() 260 : "=d" (dest->mant.m32[1]) : "0" (src->m32[1])); fp_putmant128() 262 : "=d" (dest->mant.m32[0]) : "0" (src->m32[0])); fp_putmant128() 264 if (src->m32[3] || (tmp << 8)) fp_putmant128() 270 : "d" (src->m32[0]), "0" (src->m32[1])); fp_putmant128() 272 : "=d" (dest->mant.m32[1]) : "0" (src->m32[2])); fp_putmant128() 274 : "=d" (tmp) : "0" (src->m32[3])); fp_putmant128() 276 if (src->m32[3] << 7) fp_putmant128() 280 dest->mant.m32[0] = src->m32[1]; fp_putmant128() 281 dest->mant.m32[1] = src->m32[2]; fp_putmant128() 282 dest->lowmant = src->m32[3] >> 24; fp_putmant128() 283 if (src->m32[3] << 8) fp_putmant128()
|
/linux-4.1.27/net/ceph/ |
H A D | armor.c | 4 int ceph_armor(char *dst, const char *src, const char *end); 5 int ceph_unarmor(char *dst, const char *src, const char *end); 36 int ceph_armor(char *dst, const char *src, const char *end) ceph_armor() argument 41 while (src < end) { ceph_armor() 44 a = *src++; ceph_armor() 46 if (src < end) { ceph_armor() 47 b = *src++; ceph_armor() 49 if (src < end) { ceph_armor() 50 c = *src++; ceph_armor() 74 int ceph_unarmor(char *dst, const char *src, const char *end) ceph_unarmor() argument 78 while (src < end) { ceph_unarmor() 81 if (src[0] == '\n') { ceph_unarmor() 82 src++; ceph_unarmor() 85 if (src + 4 > end) ceph_unarmor() 87 a = decode_bits(src[0]); ceph_unarmor() 88 b = decode_bits(src[1]); ceph_unarmor() 89 c = decode_bits(src[2]); ceph_unarmor() 90 d = decode_bits(src[3]); ceph_unarmor() 95 if (src[2] == '=') ceph_unarmor() 98 if (src[3] == '=') ceph_unarmor() 102 src += 4; ceph_unarmor()
|
H A D | crypto.h | 24 const struct ceph_crypto_key *src); 32 const void *src, size_t src_len); 35 const void *src, size_t src_len); 39 const void *src, size_t src_len); 48 int ceph_armor(char *dst, const char *src, const char *end); 49 int ceph_unarmor(char *dst, const char *src, const char *end);
|
/linux-4.1.27/arch/cris/arch-v10/lib/ |
H A D | string.c | 54 register unsigned const char *src __asm__ ("r11") = psrc; memcpy() 57 /* When src is aligned but not dst, this makes a few extra needless memcpy() 68 *dst = *src; memcpy() 69 src++; memcpy() 76 *(short *) dst = *(short *) src; memcpy() 77 src += 2; memcpy() 103 ;; r11 - src \n\ memcpy() 128 : "=r" (dst), "=r" (src), "=r" (n) memcpy() 131 : "0" (dst), "1" (src), "2" (n)); memcpy() 136 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 137 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 138 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 139 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 150 *dst = *src; memcpy() 154 *(short *) dst = *(short *) src; memcpy() 158 *(short *) dst = *(short *) src; dst += 2; src += 2; memcpy() 159 *dst = *src; memcpy() 163 *(long *) dst = *(long *) src; memcpy() 167 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 168 *dst = *src; memcpy() 172 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 173 *(short *) dst = *(short *) src; memcpy() 177 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 178 *(short *) dst = *(short *) src; dst += 2; src += 2; memcpy() 179 *dst = *src; memcpy() 183 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 184 *(long *) dst = *(long *) src; memcpy() 188 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 189 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 190 *dst = *src; memcpy() 194 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 195 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 196 *(short *) dst = *(short *) src; memcpy() 200 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 201 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 202 *(short *) dst = *(short *) src; dst += 2; src += 2; memcpy() 203 *dst = *src; memcpy() 207 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 208 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 209 *(long *) dst = *(long *) src; memcpy() 213 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 214 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 215 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 216 *dst = *src; memcpy() 220 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 221 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 222 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 223 *(short *) dst = *(short *) src; memcpy() 227 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 228 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 229 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 230 *(short *) dst = *(short *) src; dst += 2; src += 2; memcpy() 231 *dst = *src; memcpy()
|
/linux-4.1.27/arch/cris/arch-v32/lib/ |
H A D | string.c | 54 register unsigned const char *src __asm__ ("r11") = psrc; memcpy() 57 /* When src is aligned but not dst, this makes a few extra needless memcpy() 68 *dst = *src; memcpy() 69 src++; memcpy() 76 *(short *) dst = *(short *) src; memcpy() 77 src += 2; memcpy() 103 ;; r11 - src \n\ memcpy() 128 : "=r" (dst), "=r" (src), "=r" (n) memcpy() 131 : "0" (dst), "1" (src), "2" (n)); memcpy() 136 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 137 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 138 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 139 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 150 *dst = *src; memcpy() 154 *(short *) dst = *(short *) src; memcpy() 158 *(short *) dst = *(short *) src; dst += 2; src += 2; memcpy() 159 *dst = *src; memcpy() 163 *(long *) dst = *(long *) src; memcpy() 167 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 168 *dst = *src; memcpy() 172 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 173 *(short *) dst = *(short *) src; memcpy() 177 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 178 *(short *) dst = *(short *) src; dst += 2; src += 2; memcpy() 179 *dst = *src; memcpy() 183 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 184 *(long *) dst = *(long *) src; memcpy() 188 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 189 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 190 *dst = *src; memcpy() 194 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 195 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 196 *(short *) dst = *(short *) src; memcpy() 200 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 201 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 202 *(short *) dst = *(short *) src; dst += 2; src += 2; memcpy() 203 *dst = *src; memcpy() 207 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 208 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 209 *(long *) dst = *(long *) src; memcpy() 213 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 214 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 215 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 216 *dst = *src; memcpy() 220 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 221 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 222 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 223 *(short *) dst = *(short *) src; memcpy() 227 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 228 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 229 *(long *) dst = *(long *) src; dst += 4; src += 4; memcpy() 230 *(short *) dst = *(short *) src; dst += 2; src += 2; memcpy() 231 *dst = *src; memcpy()
|
/linux-4.1.27/include/linux/ |
H A D | fs_stack.h | 11 extern void fsstack_copy_attr_all(struct inode *dest, const struct inode *src); 12 extern void fsstack_copy_inode_size(struct inode *dst, struct inode *src); 16 const struct inode *src) fsstack_copy_attr_atime() 18 dest->i_atime = src->i_atime; fsstack_copy_attr_atime() 22 const struct inode *src) fsstack_copy_attr_times() 24 dest->i_atime = src->i_atime; fsstack_copy_attr_times() 25 dest->i_mtime = src->i_mtime; fsstack_copy_attr_times() 26 dest->i_ctime = src->i_ctime; fsstack_copy_attr_times() 15 fsstack_copy_attr_atime(struct inode *dest, const struct inode *src) fsstack_copy_attr_atime() argument 21 fsstack_copy_attr_times(struct inode *dest, const struct inode *src) fsstack_copy_attr_times() argument
|
H A D | task_io_accounting_ops.h | 49 struct task_io_accounting *src) task_blk_io_accounting_add() 51 dst->read_bytes += src->read_bytes; task_blk_io_accounting_add() 52 dst->write_bytes += src->write_bytes; task_blk_io_accounting_add() 53 dst->cancelled_write_bytes += src->cancelled_write_bytes; task_blk_io_accounting_add() 85 struct task_io_accounting *src) task_blk_io_accounting_add() 93 struct task_io_accounting *src) task_chr_io_accounting_add() 95 dst->rchar += src->rchar; task_chr_io_accounting_add() 96 dst->wchar += src->wchar; task_chr_io_accounting_add() 97 dst->syscr += src->syscr; task_chr_io_accounting_add() 98 dst->syscw += src->syscw; task_chr_io_accounting_add() 102 struct task_io_accounting *src) task_chr_io_accounting_add() 108 struct task_io_accounting *src) task_io_accounting_add() 110 task_chr_io_accounting_add(dst, src); task_io_accounting_add() 111 task_blk_io_accounting_add(dst, src); task_io_accounting_add() 48 task_blk_io_accounting_add(struct task_io_accounting *dst, struct task_io_accounting *src) task_blk_io_accounting_add() argument 84 task_blk_io_accounting_add(struct task_io_accounting *dst, struct task_io_accounting *src) task_blk_io_accounting_add() argument 92 task_chr_io_accounting_add(struct task_io_accounting *dst, struct task_io_accounting *src) task_chr_io_accounting_add() argument 101 task_chr_io_accounting_add(struct task_io_accounting *dst, struct task_io_accounting *src) task_chr_io_accounting_add() argument 107 task_io_accounting_add(struct task_io_accounting *dst, struct task_io_accounting *src) task_io_accounting_add() argument
|
H A D | string_helpers.h | 23 int string_unescape(char *src, char *dst, size_t size, unsigned int flags); 30 static inline int string_unescape_any(char *src, char *dst, size_t size) string_unescape_any() argument 32 return string_unescape(src, dst, size, UNESCAPE_ANY); string_unescape_any() 50 int string_escape_mem(const char *src, size_t isz, char *dst, size_t osz, 53 static inline int string_escape_mem_any_np(const char *src, size_t isz, string_escape_mem_any_np() argument 56 return string_escape_mem(src, isz, dst, osz, ESCAPE_ANY_NP, esc); string_escape_mem_any_np() 59 static inline int string_escape_str(const char *src, char *dst, size_t sz, string_escape_str() argument 62 return string_escape_mem(src, strlen(src), dst, sz, flags, esc); string_escape_str() 65 static inline int string_escape_str_any_np(const char *src, char *dst, string_escape_str_any_np() argument 68 return string_escape_str(src, dst, sz, ESCAPE_ANY_NP, esc); string_escape_str_any_np()
|
H A D | ucs2_string.h | 14 unsigned long ucs2_utf8size(const ucs2_char_t *src); 15 unsigned long ucs2_as_utf8(u8 *dest, const ucs2_char_t *src,
|
H A D | bitmap.h | 33 * bitmap_copy(dst, src, nbits) *dst = *src 38 * bitmap_complement(dst, src, nbits) *dst = ~(*src) 42 * bitmap_empty(src, nbits) Are all bits zero in *src? 43 * bitmap_full(src, nbits) Are all bits set in *src? 44 * bitmap_weight(src, nbits) Hamming Weight: number set bits 49 * bitmap_shift_right(dst, src, n, nbits) *dst = *src >> n 50 * bitmap_shift_left(dst, src, n, nbits) *dst = *src << n 51 * bitmap_remap(dst, src, old, new, nbits) *dst = map(old, new)(src) 94 extern void __bitmap_complement(unsigned long *dst, const unsigned long *src, 96 extern void __bitmap_shift_right(unsigned long *dst, const unsigned long *src, 98 extern void __bitmap_shift_left(unsigned long *dst, const unsigned long *src, 155 extern void bitmap_remap(unsigned long *dst, const unsigned long *src, 167 extern void bitmap_copy_le(unsigned long *dst, const unsigned long *src, unsigned int nbits); 201 static inline void bitmap_copy(unsigned long *dst, const unsigned long *src, bitmap_copy() argument 205 *dst = *src; bitmap_copy() 208 memcpy(dst, src, len); bitmap_copy() 246 static inline void bitmap_complement(unsigned long *dst, const unsigned long *src, bitmap_complement() argument 250 *dst = ~(*src); bitmap_complement() 252 __bitmap_complement(dst, src, nbits); bitmap_complement() 282 static inline int bitmap_empty(const unsigned long *src, unsigned nbits) bitmap_empty() argument 285 return ! (*src & BITMAP_LAST_WORD_MASK(nbits)); bitmap_empty() 287 return find_first_bit(src, nbits) == nbits; bitmap_empty() 290 static inline int bitmap_full(const unsigned long *src, unsigned int nbits) bitmap_full() argument 293 return ! (~(*src) & BITMAP_LAST_WORD_MASK(nbits)); bitmap_full() 295 return find_first_zero_bit(src, nbits) == nbits; bitmap_full() 298 static inline int bitmap_weight(const unsigned long *src, unsigned int nbits) bitmap_weight() argument 301 return hweight_long(*src & BITMAP_LAST_WORD_MASK(nbits)); bitmap_weight() 302 return __bitmap_weight(src, nbits); bitmap_weight() 305 static inline void bitmap_shift_right(unsigned long *dst, const unsigned long *src, bitmap_shift_right() argument 309 *dst = (*src & BITMAP_LAST_WORD_MASK(nbits)) >> shift; bitmap_shift_right() 311 __bitmap_shift_right(dst, src, shift, nbits); bitmap_shift_right() 314 static inline void bitmap_shift_left(unsigned long *dst, const unsigned long *src, bitmap_shift_left() argument 318 *dst = (*src << shift) & BITMAP_LAST_WORD_MASK(nbits); bitmap_shift_left() 320 __bitmap_shift_left(dst, src, shift, nbits); bitmap_shift_left()
|
/linux-4.1.27/lib/raid6/ |
H A D | Makefile | 14 cmd_unroll = $(AWK) -f$(srctree)/$(src)/unroll.awk -vN=$(UNROLL) \ 38 $(obj)/int1.c: $(src)/int.uc $(src)/unroll.awk FORCE 43 $(obj)/int2.c: $(src)/int.uc $(src)/unroll.awk FORCE 48 $(obj)/int4.c: $(src)/int.uc $(src)/unroll.awk FORCE 53 $(obj)/int8.c: $(src)/int.uc $(src)/unroll.awk FORCE 58 $(obj)/int16.c: $(src)/int.uc $(src)/unroll.awk FORCE 63 $(obj)/int32.c: $(src)/int.uc $(src)/unroll.awk FORCE 69 $(obj)/altivec1.c: $(src)/altivec.uc $(src)/unroll.awk FORCE 75 $(obj)/altivec2.c: $(src)/altivec.uc $(src)/unroll.awk FORCE 81 $(obj)/altivec4.c: $(src)/altivec.uc $(src)/unroll.awk FORCE 87 $(obj)/altivec8.c: $(src)/altivec.uc $(src)/unroll.awk FORCE 93 $(obj)/neon1.c: $(src)/neon.uc $(src)/unroll.awk FORCE 99 $(obj)/neon2.c: $(src)/neon.uc $(src)/unroll.awk FORCE 105 $(obj)/neon4.c: $(src)/neon.uc $(src)/unroll.awk FORCE 111 $(obj)/neon8.c: $(src)/neon.uc $(src)/unroll.awk FORCE 116 $(obj)/tilegx8.c: $(src)/tilegx.uc $(src)/unroll.awk FORCE
|
/linux-4.1.27/tools/perf/util/ |
H A D | alias.c | 28 int src, dst, count = 0, size = 16; local 35 for (src = dst = 0; cmdline[src];) { 36 char c = cmdline[src]; 39 while (cmdline[++src] 40 && isspace(cmdline[src])) 49 src++; 52 src++; 55 src++; 56 c = cmdline[src]; 63 src++;
|
H A D | quote.c | 20 static void sq_quote_buf(struct strbuf *dst, const char *src) sq_quote_buf() argument 24 if (dst->buf == src) sq_quote_buf() 28 while (*src) { sq_quote_buf() 29 size_t len = strcspn(src, "'!"); sq_quote_buf() 30 strbuf_add(dst, src, len); sq_quote_buf() 31 src += len; sq_quote_buf() 32 while (need_bs_quote(*src)) { sq_quote_buf() 34 strbuf_addch(dst, *src++); sq_quote_buf()
|
/linux-4.1.27/drivers/media/platform/coda/ |
H A D | Makefile | 1 ccflags-y += -I$(src)
|
/linux-4.1.27/drivers/crypto/qat/qat_dh895xcc/ |
H A D | Makefile | 1 ccflags-y := -I$(src)/../qat_common
|
/linux-4.1.27/drivers/crypto/vmx/ |
H A D | Makefile | 13 $(src)/aesp8-ppc.S: $(src)/aesp8-ppc.pl 16 $(src)/ghashp8-ppc.S: $(src)/ghashp8-ppc.pl
|
/linux-4.1.27/arch/x86/include/asm/crypto/ |
H A D | serpent-sse2.h | 12 const u8 *src, bool xor); 14 const u8 *src); 17 const u8 *src) serpent_enc_blk_xway() 19 __serpent_enc_blk_4way(ctx, dst, src, false); serpent_enc_blk_xway() 23 const u8 *src) serpent_enc_blk_xway_xor() 25 __serpent_enc_blk_4way(ctx, dst, src, true); serpent_enc_blk_xway_xor() 29 const u8 *src) serpent_dec_blk_xway() 31 serpent_dec_blk_4way(ctx, dst, src); serpent_dec_blk_xway() 39 const u8 *src, bool xor); 41 const u8 *src); 44 const u8 *src) serpent_enc_blk_xway() 46 __serpent_enc_blk_8way(ctx, dst, src, false); serpent_enc_blk_xway() 50 const u8 *src) serpent_enc_blk_xway_xor() 52 __serpent_enc_blk_8way(ctx, dst, src, true); serpent_enc_blk_xway_xor() 56 const u8 *src) serpent_dec_blk_xway() 58 serpent_dec_blk_8way(ctx, dst, src); serpent_dec_blk_xway() 16 serpent_enc_blk_xway(struct serpent_ctx *ctx, u8 *dst, const u8 *src) serpent_enc_blk_xway() argument 22 serpent_enc_blk_xway_xor(struct serpent_ctx *ctx, u8 *dst, const u8 *src) serpent_enc_blk_xway_xor() argument 28 serpent_dec_blk_xway(struct serpent_ctx *ctx, u8 *dst, const u8 *src) serpent_dec_blk_xway() argument 43 serpent_enc_blk_xway(struct serpent_ctx *ctx, u8 *dst, const u8 *src) serpent_enc_blk_xway() argument 49 serpent_enc_blk_xway_xor(struct serpent_ctx *ctx, u8 *dst, const u8 *src) serpent_enc_blk_xway_xor() argument 55 serpent_dec_blk_xway(struct serpent_ctx *ctx, u8 *dst, const u8 *src) serpent_dec_blk_xway() argument
|
H A D | camellia.h | 41 const u8 *src, bool xor); 43 const u8 *src); 47 const u8 *src, bool xor); 49 const u8 *src); 53 const u8 *src); 55 const u8 *src); 58 const u8 *src); 60 const u8 *src, le128 *iv); 63 const u8 *src, le128 *iv); 65 const u8 *src, le128 *iv); 68 const u8 *src) camellia_enc_blk() 70 __camellia_enc_blk(ctx, dst, src, false); camellia_enc_blk() 74 const u8 *src) camellia_enc_blk_xor() 76 __camellia_enc_blk(ctx, dst, src, true); camellia_enc_blk_xor() 80 const u8 *src) camellia_enc_blk_2way() 82 __camellia_enc_blk_2way(ctx, dst, src, false); camellia_enc_blk_2way() 86 const u8 *src) camellia_enc_blk_xor_2way() 88 __camellia_enc_blk_2way(ctx, dst, src, true); camellia_enc_blk_xor_2way() 92 extern void camellia_decrypt_cbc_2way(void *ctx, u128 *dst, const u128 *src); 93 extern void camellia_crypt_ctr(void *ctx, u128 *dst, const u128 *src, 95 extern void camellia_crypt_ctr_2way(void *ctx, u128 *dst, const u128 *src, 98 extern void camellia_xts_enc(void *ctx, u128 *dst, const u128 *src, le128 *iv); 99 extern void camellia_xts_dec(void *ctx, u128 *dst, const u128 *src, le128 *iv); 67 camellia_enc_blk(struct camellia_ctx *ctx, u8 *dst, const u8 *src) camellia_enc_blk() argument 73 camellia_enc_blk_xor(struct camellia_ctx *ctx, u8 *dst, const u8 *src) camellia_enc_blk_xor() argument 79 camellia_enc_blk_2way(struct camellia_ctx *ctx, u8 *dst, const u8 *src) camellia_enc_blk_2way() argument 85 camellia_enc_blk_xor_2way(struct camellia_ctx *ctx, u8 *dst, const u8 *src) camellia_enc_blk_xor_2way() argument
|
H A D | aes.h | 8 const u8 *src); 10 const u8 *src);
|
H A D | twofish.h | 21 const u8 *src); 23 const u8 *src); 27 const u8 *src, bool xor); 29 const u8 *src); 32 extern void twofish_dec_blk_cbc_3way(void *ctx, u128 *dst, const u128 *src); 33 extern void twofish_enc_blk_ctr(void *ctx, u128 *dst, const u128 *src, 35 extern void twofish_enc_blk_ctr_3way(void *ctx, u128 *dst, const u128 *src,
|
H A D | glue_helper.h | 13 typedef void (*common_glue_func_t)(void *ctx, u8 *dst, const u8 *src); 14 typedef void (*common_glue_cbc_func_t)(void *ctx, u128 *dst, const u128 *src); 15 typedef void (*common_glue_ctr_func_t)(void *ctx, u128 *dst, const u128 *src, 17 typedef void (*common_glue_xts_func_t)(void *ctx, u128 *dst, const u128 *src, 78 static inline void le128_to_be128(be128 *dst, const le128 *src) le128_to_be128() argument 80 dst->a = cpu_to_be64(le64_to_cpu(src->a)); le128_to_be128() 81 dst->b = cpu_to_be64(le64_to_cpu(src->b)); le128_to_be128() 84 static inline void be128_to_le128(le128 *dst, const be128 *src) be128_to_le128() argument 86 dst->a = cpu_to_le64(be64_to_cpu(src->a)); be128_to_le128() 87 dst->b = cpu_to_le64(be64_to_cpu(src->b)); be128_to_le128() 103 static inline void le128_gf128mul_x_ble(le128 *dst, const le128 *src) le128_gf128mul_x_ble() argument 105 u64 a = le64_to_cpu(src->a); le128_gf128mul_x_ble() 106 u64 b = le64_to_cpu(src->b); le128_gf128mul_x_ble() 116 struct scatterlist *src, unsigned int nbytes); 121 struct scatterlist *src, 127 struct scatterlist *src, 133 struct scatterlist *src, unsigned int nbytes); 138 struct scatterlist *src, unsigned int nbytes, 142 extern void glue_xts_crypt_128bit_one(void *ctx, u128 *dst, const u128 *src,
|
H A D | serpent-avx.h | 20 const u8 *src); 22 const u8 *src); 25 const u8 *src); 27 const u8 *src, le128 *iv); 30 const u8 *src, le128 *iv); 32 const u8 *src, le128 *iv); 34 extern void __serpent_crypt_ctr(void *ctx, u128 *dst, const u128 *src, 37 extern void serpent_xts_enc(void *ctx, u128 *dst, const u128 *src, le128 *iv); 38 extern void serpent_xts_dec(void *ctx, u128 *dst, const u128 *src, le128 *iv);
|
/linux-4.1.27/arch/um/kernel/ |
H A D | maccess.c | 13 long probe_kernel_read(void *dst, const void *src, size_t size) probe_kernel_read() argument 15 void *psrc = (void *)rounddown((unsigned long)src, PAGE_SIZE); probe_kernel_read() 17 if ((unsigned long)src < PAGE_SIZE || size <= 0) probe_kernel_read() 20 if (os_mincore(psrc, size + src - psrc) <= 0) probe_kernel_read() 23 return __probe_kernel_read(dst, src, size); probe_kernel_read()
|
/linux-4.1.27/arch/score/lib/ |
H A D | checksum.S | 34 #define CSUM_BIGCHUNK(src, offset, sum) \ 35 lw r8, [src, offset + 0x00]; \ 36 lw r9, [src, offset + 0x04]; \ 37 lw r10, [src, offset + 0x08]; \ 38 lw r11, [src, offset + 0x0c]; \ 43 lw r8, [src, offset + 0x10]; \ 44 lw r9, [src, offset + 0x14]; \ 45 lw r10, [src, offset + 0x18]; \ 46 lw r11, [src, offset + 0x1c]; \ 52 #define src r4 define 57 /* unknown src alignment and < 8 bytes to go */ 63 andri.c r25,r4 , 0x1 /*Is src 2 bytes aligned?*/ 69 lbu r9, [src] 72 addi src, 0x1 80 /* Still a full word (4byte) to go,and the src is word aligned.*/ 81 andri.c r8, src, 0x3 /*src is 4bytes aligned, so use LW!!*/ 83 lhu r9, [src] 84 addi src, 2 86 lhu r9, [src] 87 addi src, 2 92 lw r9, [src] 93 addi src, 4 99 lhu r9, [src] 100 addi src, 0x2 /* src+=2 */ 106 lbu r9, [src] 143 andri.c r25, src, 0x1 /* odd buffer? */ 147 lbu r8, [src] 151 addi src, 0x1 154 andri.c r8, src, 0x2 /* 4bytes(dword)_aligned? */ 156 lhu r8, [src] 159 addi src, 0x2 166 andri.c r26, src, 0x4 168 lw r8, [src] 171 addi src, 0x4 174 andri.c r26, src, 0x8 176 lw r8, [src, 0x0] 177 lw r9, [src, 0x4] 181 addi src, 0x8 184 andri.c r26, src, 0x10 186 lw r10, [src, 0x08] 187 lw r11, [src, 0x0c] 188 lw r8, [src, 0x00] 189 lw r9, [src, 0x04] 195 addi src, 0x10 203 CSUM_BIGCHUNK(src, 0x00, sum) 204 CSUM_BIGCHUNK(src, 0x20, sum) 205 CSUM_BIGCHUNK(src, 0x40, sum) 206 CSUM_BIGCHUNK(src, 0x60, sum) 208 addi src, 0x80 216 CSUM_BIGCHUNK(src, 0x00, sum) 217 CSUM_BIGCHUNK(src, 0x20, sum) 218 addi src, 0x40 226 CSUM_BIGCHUNK(src, 0x00, sum) 228 addri src, src, 0x20 237 lw r8, [src] 240 addi src, 0x4
|
H A D | checksum_copy.c | 30 unsigned int csum_partial_copy(const char *src, char *dst, csum_partial_copy() argument 33 sum = csum_partial(src, len, sum); csum_partial_copy() 34 memcpy(dst, src, len); csum_partial_copy() 39 unsigned int csum_partial_copy_from_user(const char *src, char *dst, csum_partial_copy_from_user() argument 45 missing = copy_from_user(dst, src, len); csum_partial_copy_from_user()
|
/linux-4.1.27/drivers/media/platform/s5p-tv/ |
H A D | mixer_vp_layer.c | 107 &layer->fmt->plane[0], layer->geo.src.full_width, mxr_vp_buffer_set() 108 layer->geo.src.full_height); mxr_vp_buffer_set() 114 luma_addr[1] = luma_addr[0] + layer->geo.src.full_width; mxr_vp_buffer_set() 146 struct mxr_crop *src = &geo->src; mxr_vp_fix_geometry() local 159 dst->width = clamp(dst->width, 8U, 16 * src->full_width); mxr_vp_fix_geometry() 160 dst->height = clamp(dst->height, 1U, 16 * src->full_height); mxr_vp_fix_geometry() 171 x_center = src->x_offset + src->width / 2; mxr_vp_fix_geometry() 172 y_center = src->y_offset + src->height / 2; mxr_vp_fix_geometry() 175 src->width = clamp(src->width, round_up(dst->width / 16, 4), mxr_vp_fix_geometry() 177 src->height = clamp(src->height, round_up(dst->height / 16, 4), mxr_vp_fix_geometry() 181 src->width = clamp(src->width, 32U, 2047U); mxr_vp_fix_geometry() 182 src->height = clamp(src->height, 4U, 2047U); mxr_vp_fix_geometry() 185 src->x_offset = do_center(x_center, src->width, mxr_vp_fix_geometry() 186 src->full_width, flags); mxr_vp_fix_geometry() 187 src->y_offset = do_center(y_center, src->height, mxr_vp_fix_geometry() 188 src->full_height, flags); mxr_vp_fix_geometry() 191 geo->x_ratio = (src->width << 16) / dst->width; mxr_vp_fix_geometry() 192 geo->y_ratio = (src->height << 16) / dst->height; mxr_vp_fix_geometry() 196 src->full_width = clamp(src->full_width, mxr_vp_fix_geometry() 197 ALIGN(src->width + src->x_offset, 8), 8192U); mxr_vp_fix_geometry() 198 src->full_height = clamp(src->full_height, mxr_vp_fix_geometry() 199 src->height + src->y_offset, 8192U); mxr_vp_fix_geometry()
|
H A D | mixer_grp_layer.c | 152 struct mxr_crop *src = &geo->src; mxr_graph_fix_geometry() local 179 dst->width = min(dst->width, 2 * src->full_width); mxr_graph_fix_geometry() 180 dst->height = min(dst->height, 2 * src->full_height); mxr_graph_fix_geometry() 192 x_center = src->x_offset + src->width / 2; mxr_graph_fix_geometry() 193 y_center = src->y_offset + src->height / 2; mxr_graph_fix_geometry() 195 if (src->full_width < dst->width) mxr_graph_fix_geometry() 196 src->width = dst->width / 2; mxr_graph_fix_geometry() 198 src->width = closest(src->width, dst->width / 2, mxr_graph_fix_geometry() 201 if (src->width == dst->width) mxr_graph_fix_geometry() 206 if (src->full_height < dst->height) mxr_graph_fix_geometry() 207 src->height = dst->height / 2; mxr_graph_fix_geometry() 209 src->height = closest(src->height, dst->height / 2, mxr_graph_fix_geometry() 212 if (src->height == dst->height) mxr_graph_fix_geometry() 218 src->x_offset = do_center(x_center, src->width, mxr_graph_fix_geometry() 219 src->full_width, flags); mxr_graph_fix_geometry() 220 src->y_offset = do_center(y_center, src->height, mxr_graph_fix_geometry() 221 src->full_height, flags); mxr_graph_fix_geometry() 225 src->full_width = clamp_val(src->full_width, mxr_graph_fix_geometry() 226 src->width + src->x_offset, 32767); mxr_graph_fix_geometry() 227 src->full_height = clamp_val(src->full_height, mxr_graph_fix_geometry() 228 src->height + src->y_offset, 2047); mxr_graph_fix_geometry()
|
/linux-4.1.27/arch/x86/crypto/ |
H A D | glue_helper-asm-avx.S | 18 #define load_8way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ 19 vmovdqu (0*16)(src), x0; \ 20 vmovdqu (1*16)(src), x1; \ 21 vmovdqu (2*16)(src), x2; \ 22 vmovdqu (3*16)(src), x3; \ 23 vmovdqu (4*16)(src), x4; \ 24 vmovdqu (5*16)(src), x5; \ 25 vmovdqu (6*16)(src), x6; \ 26 vmovdqu (7*16)(src), x7; 38 #define store_cbc_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ 39 vpxor (0*16)(src), x1, x1; \ 40 vpxor (1*16)(src), x2, x2; \ 41 vpxor (2*16)(src), x3, x3; \ 42 vpxor (3*16)(src), x4, x4; \ 43 vpxor (4*16)(src), x5, x5; \ 44 vpxor (5*16)(src), x6, x6; \ 45 vpxor (6*16)(src), x7, x7; \ 82 #define store_ctr_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ 83 vpxor (0*16)(src), x0, x0; \ 84 vpxor (1*16)(src), x1, x1; \ 85 vpxor (2*16)(src), x2, x2; \ 86 vpxor (3*16)(src), x3, x3; \ 87 vpxor (4*16)(src), x4, x4; \ 88 vpxor (5*16)(src), x5, x5; \ 89 vpxor (6*16)(src), x6, x6; \ 90 vpxor (7*16)(src), x7, x7; \ 100 #define load_xts_8way(iv, src, dst, x0, x1, x2, x3, x4, x5, x6, x7, tiv, t0, \ 106 vpxor (0*16)(src), tiv, x0; \ 111 vpxor (1*16)(src), tiv, x1; \ 115 vpxor (2*16)(src), tiv, x2; \ 119 vpxor (3*16)(src), tiv, x3; \ 123 vpxor (4*16)(src), tiv, x4; \ 127 vpxor (5*16)(src), tiv, x5; \ 131 vpxor (6*16)(src), tiv, x6; \ 135 vpxor (7*16)(src), tiv, x7; \
|
H A D | blowfish_glue.c | 37 asmlinkage void __blowfish_enc_blk(struct bf_ctx *ctx, u8 *dst, const u8 *src, 39 asmlinkage void blowfish_dec_blk(struct bf_ctx *ctx, u8 *dst, const u8 *src); 43 const u8 *src, bool xor); 45 const u8 *src); 47 static inline void blowfish_enc_blk(struct bf_ctx *ctx, u8 *dst, const u8 *src) blowfish_enc_blk() argument 49 __blowfish_enc_blk(ctx, dst, src, false); blowfish_enc_blk() 53 const u8 *src) blowfish_enc_blk_xor() 55 __blowfish_enc_blk(ctx, dst, src, true); blowfish_enc_blk_xor() 59 const u8 *src) blowfish_enc_blk_4way() 61 __blowfish_enc_blk_4way(ctx, dst, src, false); blowfish_enc_blk_4way() 65 const u8 *src) blowfish_enc_blk_xor_4way() 67 __blowfish_enc_blk_4way(ctx, dst, src, true); blowfish_enc_blk_xor_4way() 70 static void blowfish_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) blowfish_encrypt() argument 72 blowfish_enc_blk(crypto_tfm_ctx(tfm), dst, src); blowfish_encrypt() local 75 static void blowfish_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) blowfish_decrypt() argument 77 blowfish_dec_blk(crypto_tfm_ctx(tfm), dst, src); blowfish_decrypt() local 92 u8 *wsrc = walk->src.virt.addr; ecb_crypt() 126 struct scatterlist *src, unsigned int nbytes) ecb_encrypt() 130 blkcipher_walk_init(&walk, dst, src, nbytes); ecb_encrypt() 135 struct scatterlist *src, unsigned int nbytes) ecb_decrypt() 139 blkcipher_walk_init(&walk, dst, src, nbytes); ecb_decrypt() 149 u64 *src = (u64 *)walk->src.virt.addr; __cbc_encrypt() local 154 *dst = *src ^ *iv; __cbc_encrypt() 158 src += 1; __cbc_encrypt() 168 struct scatterlist *src, unsigned int nbytes) cbc_encrypt() 173 blkcipher_walk_init(&walk, dst, src, nbytes); cbc_encrypt() 190 u64 *src = (u64 *)walk->src.virt.addr; __cbc_decrypt() local 196 src += nbytes / bsize - 1; __cbc_decrypt() 199 last_iv = *src; __cbc_decrypt() 205 src -= 4 - 1; __cbc_decrypt() 208 ivs[0] = src[0]; __cbc_decrypt() 209 ivs[1] = src[1]; __cbc_decrypt() 210 ivs[2] = src[2]; __cbc_decrypt() 212 blowfish_dec_blk_4way(ctx, (u8 *)dst, (u8 *)src); __cbc_decrypt() 222 *dst ^= *(src - 1); __cbc_decrypt() 223 src -= 1; __cbc_decrypt() 230 blowfish_dec_blk(ctx, (u8 *)dst, (u8 *)src); __cbc_decrypt() 236 *dst ^= *(src - 1); __cbc_decrypt() 237 src -= 1; __cbc_decrypt() 249 struct scatterlist *src, unsigned int nbytes) cbc_decrypt() 254 blkcipher_walk_init(&walk, dst, src, nbytes); cbc_decrypt() 269 u8 *src = walk->src.virt.addr; ctr_crypt_final() local 274 crypto_xor(keystream, src, nbytes); ctr_crypt_final() 286 u64 *src = (u64 *)walk->src.virt.addr; __ctr_crypt() local 294 if (dst != src) { __ctr_crypt() 295 dst[0] = src[0]; __ctr_crypt() 296 dst[1] = src[1]; __ctr_crypt() 297 dst[2] = src[2]; __ctr_crypt() 298 dst[3] = src[3]; __ctr_crypt() 310 src += 4; __ctr_crypt() 320 if (dst != src) __ctr_crypt() 321 *dst = *src; __ctr_crypt() 327 src += 1; __ctr_crypt() 337 struct scatterlist *src, unsigned int nbytes) ctr_crypt() 342 blkcipher_walk_init(&walk, dst, src, nbytes); ctr_crypt() 52 blowfish_enc_blk_xor(struct bf_ctx *ctx, u8 *dst, const u8 *src) blowfish_enc_blk_xor() argument 58 blowfish_enc_blk_4way(struct bf_ctx *ctx, u8 *dst, const u8 *src) blowfish_enc_blk_4way() argument 64 blowfish_enc_blk_xor_4way(struct bf_ctx *ctx, u8 *dst, const u8 *src) blowfish_enc_blk_xor_4way() argument 125 ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_encrypt() argument 134 ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_decrypt() argument 167 cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_encrypt() argument 248 cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_decrypt() argument 336 ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ctr_crypt() argument
|
H A D | des3_ede_glue.c | 38 const u8 *src); 42 const u8 *src); 45 const u8 *src) des3_ede_enc_blk() 49 des3_ede_x86_64_crypt_blk(enc_ctx, dst, src); des3_ede_enc_blk() 53 const u8 *src) des3_ede_dec_blk() 57 des3_ede_x86_64_crypt_blk(dec_ctx, dst, src); des3_ede_dec_blk() 61 const u8 *src) des3_ede_enc_blk_3way() 65 des3_ede_x86_64_crypt_blk_3way(enc_ctx, dst, src); des3_ede_enc_blk_3way() 69 const u8 *src) des3_ede_dec_blk_3way() 73 des3_ede_x86_64_crypt_blk_3way(dec_ctx, dst, src); des3_ede_dec_blk_3way() 76 static void des3_ede_x86_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) des3_ede_x86_encrypt() argument 78 des3_ede_enc_blk(crypto_tfm_ctx(tfm), dst, src); des3_ede_x86_encrypt() local 81 static void des3_ede_x86_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) des3_ede_x86_decrypt() argument 83 des3_ede_dec_blk(crypto_tfm_ctx(tfm), dst, src); des3_ede_x86_decrypt() local 96 u8 *wsrc = walk->src.virt.addr; ecb_crypt() 131 struct scatterlist *src, unsigned int nbytes) ecb_encrypt() 136 blkcipher_walk_init(&walk, dst, src, nbytes); ecb_encrypt() 141 struct scatterlist *src, unsigned int nbytes) ecb_decrypt() 146 blkcipher_walk_init(&walk, dst, src, nbytes); ecb_decrypt() 156 u64 *src = (u64 *)walk->src.virt.addr; __cbc_encrypt() local 161 *dst = *src ^ *iv; __cbc_encrypt() 165 src += 1; __cbc_encrypt() 175 struct scatterlist *src, unsigned int nbytes) cbc_encrypt() 180 blkcipher_walk_init(&walk, dst, src, nbytes); cbc_encrypt() 197 u64 *src = (u64 *)walk->src.virt.addr; __cbc_decrypt() local 203 src += nbytes / bsize - 1; __cbc_decrypt() 206 last_iv = *src; __cbc_decrypt() 212 src -= 3 - 1; __cbc_decrypt() 215 ivs[0] = src[0]; __cbc_decrypt() 216 ivs[1] = src[1]; __cbc_decrypt() 218 des3_ede_dec_blk_3way(ctx, (u8 *)dst, (u8 *)src); __cbc_decrypt() 227 *dst ^= *(src - 1); __cbc_decrypt() 228 src -= 1; __cbc_decrypt() 235 des3_ede_dec_blk(ctx, (u8 *)dst, (u8 *)src); __cbc_decrypt() 241 *dst ^= *(src - 1); __cbc_decrypt() 242 src -= 1; __cbc_decrypt() 254 struct scatterlist *src, unsigned int nbytes) cbc_decrypt() 259 blkcipher_walk_init(&walk, dst, src, nbytes); cbc_decrypt() 275 u8 *src = walk->src.virt.addr; ctr_crypt_final() local 280 crypto_xor(keystream, src, nbytes); ctr_crypt_final() 292 __be64 *src = (__be64 *)walk->src.virt.addr; __ctr_crypt() local 308 dst[0] = src[0] ^ ctrblocks[0]; __ctr_crypt() 309 dst[1] = src[1] ^ ctrblocks[1]; __ctr_crypt() 310 dst[2] = src[2] ^ ctrblocks[2]; __ctr_crypt() 312 src += 3; __ctr_crypt() 326 dst[0] = src[0] ^ ctrblocks[0]; __ctr_crypt() 328 src += 1; __ctr_crypt() 338 struct scatterlist *src, unsigned int nbytes) ctr_crypt() 343 blkcipher_walk_init(&walk, dst, src, nbytes); ctr_crypt() 44 des3_ede_enc_blk(struct des3_ede_x86_ctx *ctx, u8 *dst, const u8 *src) des3_ede_enc_blk() argument 52 des3_ede_dec_blk(struct des3_ede_x86_ctx *ctx, u8 *dst, const u8 *src) des3_ede_dec_blk() argument 60 des3_ede_enc_blk_3way(struct des3_ede_x86_ctx *ctx, u8 *dst, const u8 *src) des3_ede_enc_blk_3way() argument 68 des3_ede_dec_blk_3way(struct des3_ede_x86_ctx *ctx, u8 *dst, const u8 *src) des3_ede_dec_blk_3way() argument 130 ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_encrypt() argument 140 ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_decrypt() argument 174 cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_encrypt() argument 253 cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_decrypt() argument 337 ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ctr_crypt() argument
|
H A D | glue_helper-asm-avx2.S | 13 #define load_16way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ 14 vmovdqu (0*32)(src), x0; \ 15 vmovdqu (1*32)(src), x1; \ 16 vmovdqu (2*32)(src), x2; \ 17 vmovdqu (3*32)(src), x3; \ 18 vmovdqu (4*32)(src), x4; \ 19 vmovdqu (5*32)(src), x5; \ 20 vmovdqu (6*32)(src), x6; \ 21 vmovdqu (7*32)(src), x7; 33 #define store_cbc_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7, t0) \ 35 vinserti128 $1, (src), t0, t0; \ 37 vpxor (0*32+16)(src), x1, x1; \ 38 vpxor (1*32+16)(src), x2, x2; \ 39 vpxor (2*32+16)(src), x3, x3; \ 40 vpxor (3*32+16)(src), x4, x4; \ 41 vpxor (4*32+16)(src), x5, x5; \ 42 vpxor (5*32+16)(src), x6, x6; \ 43 vpxor (6*32+16)(src), x7, x7; \ 93 #define store_ctr_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ 94 vpxor (0*32)(src), x0, x0; \ 95 vpxor (1*32)(src), x1, x1; \ 96 vpxor (2*32)(src), x2, x2; \ 97 vpxor (3*32)(src), x3, x3; \ 98 vpxor (4*32)(src), x4, x4; \ 99 vpxor (5*32)(src), x5, x5; \ 100 vpxor (6*32)(src), x6, x6; \ 101 vpxor (7*32)(src), x7, x7; \ 123 #define load_xts_16way(iv, src, dst, x0, x1, x2, x3, x4, x5, x6, x7, tiv, \ 135 vpxor (0*32)(src), tiv, x0; \ 140 vpxor (1*32)(src), tiv, x1; \ 144 vpxor (2*32)(src), tiv, x2; \ 148 vpxor (3*32)(src), tiv, x3; \ 152 vpxor (4*32)(src), tiv, x4; \ 156 vpxor (5*32)(src), tiv, x5; \ 160 vpxor (6*32)(src), tiv, x6; \ 164 vpxor (7*32)(src), tiv, x7; \
|
H A D | aes_glue.c | 13 void crypto_aes_encrypt_x86(struct crypto_aes_ctx *ctx, u8 *dst, const u8 *src) crypto_aes_encrypt_x86() argument 15 aes_enc_blk(ctx, dst, src); crypto_aes_encrypt_x86() 19 void crypto_aes_decrypt_x86(struct crypto_aes_ctx *ctx, u8 *dst, const u8 *src) crypto_aes_decrypt_x86() argument 21 aes_dec_blk(ctx, dst, src); crypto_aes_decrypt_x86() 25 static void aes_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) aes_encrypt() argument 27 aes_enc_blk(crypto_tfm_ctx(tfm), dst, src); aes_encrypt() local 30 static void aes_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) aes_decrypt() argument 32 aes_dec_blk(crypto_tfm_ctx(tfm), dst, src); aes_decrypt() local
|
H A D | cast5_avx_glue.c | 41 const u8 *src); 43 const u8 *src); 45 const u8 *src); 46 asmlinkage void cast5_ctr_16way(struct cast5_ctx *ctx, u8 *dst, const u8 *src, 67 void (*fn)(struct cast5_ctx *ctx, u8 *dst, const u8 *src); ecb_crypt() 76 u8 *wsrc = walk->src.virt.addr; ecb_crypt() 115 struct scatterlist *src, unsigned int nbytes) ecb_encrypt() 119 blkcipher_walk_init(&walk, dst, src, nbytes); ecb_encrypt() 124 struct scatterlist *src, unsigned int nbytes) ecb_decrypt() 128 blkcipher_walk_init(&walk, dst, src, nbytes); ecb_decrypt() 138 u64 *src = (u64 *)walk->src.virt.addr; __cbc_encrypt() local 143 *dst = *src ^ *iv; __cbc_encrypt() 147 src += 1; __cbc_encrypt() 157 struct scatterlist *src, unsigned int nbytes) cbc_encrypt() 162 blkcipher_walk_init(&walk, dst, src, nbytes); cbc_encrypt() 179 u64 *src = (u64 *)walk->src.virt.addr; __cbc_decrypt() local 184 src += nbytes / bsize - 1; __cbc_decrypt() 187 last_iv = *src; __cbc_decrypt() 193 src -= CAST5_PARALLEL_BLOCKS - 1; __cbc_decrypt() 196 cast5_cbc_dec_16way(ctx, (u8 *)dst, (u8 *)src); __cbc_decrypt() 202 *dst ^= *(src - 1); __cbc_decrypt() 203 src -= 1; __cbc_decrypt() 210 __cast5_decrypt(ctx, (u8 *)dst, (u8 *)src); __cbc_decrypt() 216 *dst ^= *(src - 1); __cbc_decrypt() 217 src -= 1; __cbc_decrypt() 229 struct scatterlist *src, unsigned int nbytes) cbc_decrypt() 235 blkcipher_walk_init(&walk, dst, src, nbytes); cbc_decrypt() 255 u8 *src = walk->src.virt.addr; ctr_crypt_final() local 260 crypto_xor(keystream, src, nbytes); ctr_crypt_final() 272 u64 *src = (u64 *)walk->src.virt.addr; __ctr_crypt() local 278 cast5_ctr_16way(ctx, (u8 *)dst, (u8 *)src, __ctr_crypt() 281 src += CAST5_PARALLEL_BLOCKS; __ctr_crypt() 294 if (dst != src) __ctr_crypt() 295 *dst = *src; __ctr_crypt() 303 src += 1; __ctr_crypt() 313 struct scatterlist *src, unsigned int nbytes) ctr_crypt() 319 blkcipher_walk_init(&walk, dst, src, nbytes); ctr_crypt() 114 ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_encrypt() argument 123 ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_decrypt() argument 156 cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_encrypt() argument 228 cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_decrypt() argument 312 ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ctr_crypt() argument
|
H A D | glue_helper.c | 48 u8 *wsrc = walk->src.virt.addr; __glue_ecb_crypt_128bit() 83 struct scatterlist *src, unsigned int nbytes) glue_ecb_crypt_128bit() 87 blkcipher_walk_init(&walk, dst, src, nbytes); glue_ecb_crypt_128bit() 99 u128 *src = (u128 *)walk->src.virt.addr; __glue_cbc_encrypt_128bit() local 104 u128_xor(dst, src, iv); __glue_cbc_encrypt_128bit() 108 src += 1; __glue_cbc_encrypt_128bit() 120 struct scatterlist *src, unsigned int nbytes) glue_cbc_encrypt_128bit() 125 blkcipher_walk_init(&walk, dst, src, nbytes); glue_cbc_encrypt_128bit() 145 u128 *src = (u128 *)walk->src.virt.addr; __glue_cbc_decrypt_128bit() local 152 src += nbytes / bsize - 1; __glue_cbc_decrypt_128bit() 155 last_iv = *src; __glue_cbc_decrypt_128bit() 165 src -= num_blocks - 1; __glue_cbc_decrypt_128bit() 168 gctx->funcs[i].fn_u.cbc(ctx, dst, src); __glue_cbc_decrypt_128bit() 174 u128_xor(dst, dst, src - 1); __glue_cbc_decrypt_128bit() 175 src -= 1; __glue_cbc_decrypt_128bit() 194 struct scatterlist *src, unsigned int nbytes) glue_cbc_decrypt_128bit() 201 blkcipher_walk_init(&walk, dst, src, nbytes); glue_cbc_decrypt_128bit() 221 u8 *src = (u8 *)walk->src.virt.addr; glue_ctr_crypt_final_128bit() local 229 memcpy(&tmp, src, nbytes); glue_ctr_crypt_final_128bit() 243 u128 *src = (u128 *)walk->src.virt.addr; __glue_ctr_crypt_128bit() local 258 gctx->funcs[i].fn_u.ctr(ctx, dst, src, &ctrblk); __glue_ctr_crypt_128bit() 260 src += num_blocks; __glue_ctr_crypt_128bit() 277 struct scatterlist *src, unsigned int nbytes) glue_ctr_crypt_128bit() 284 blkcipher_walk_init(&walk, dst, src, nbytes); glue_ctr_crypt_128bit() 313 u128 *src = (u128 *)walk->src.virt.addr; __glue_xts_crypt_128bit() local 325 gctx->funcs[i].fn_u.xts(ctx, dst, src, __glue_xts_crypt_128bit() 328 src += num_blocks; __glue_xts_crypt_128bit() 345 struct scatterlist *src, unsigned int nbytes, glue_xts_crypt_128bit() 346 void (*tweak_fn)(void *ctx, u8 *dst, const u8 *src), glue_xts_crypt_128bit() 354 blkcipher_walk_init(&walk, dst, src, nbytes); glue_xts_crypt_128bit() 382 void glue_xts_crypt_128bit_one(void *ctx, u128 *dst, const u128 *src, le128 *iv, glue_xts_crypt_128bit_one() argument 391 u128_xor(dst, src, (u128 *)&ivblk); glue_xts_crypt_128bit_one() 81 glue_ecb_crypt_128bit(const struct common_glue_ctx *gctx, struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) glue_ecb_crypt_128bit() argument 117 glue_cbc_encrypt_128bit(const common_glue_func_t fn, struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) glue_cbc_encrypt_128bit() argument 191 glue_cbc_decrypt_128bit(const struct common_glue_ctx *gctx, struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) glue_cbc_decrypt_128bit() argument 275 glue_ctr_crypt_128bit(const struct common_glue_ctx *gctx, struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) glue_ctr_crypt_128bit() argument 343 glue_xts_crypt_128bit(const struct common_glue_ctx *gctx, struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes, void (*tweak_fn)(void *ctx, u8 *dst, const u8 *src), void *tweak_ctx, void *crypt_ctx) glue_xts_crypt_128bit() argument
|
H A D | twofish_glue_3way.c | 40 const u8 *src) twofish_enc_blk_3way() 42 __twofish_enc_blk_3way(ctx, dst, src, false); twofish_enc_blk_3way() 46 const u8 *src) twofish_enc_blk_xor_3way() 48 __twofish_enc_blk_3way(ctx, dst, src, true); twofish_enc_blk_xor_3way() 51 void twofish_dec_blk_cbc_3way(void *ctx, u128 *dst, const u128 *src) twofish_dec_blk_cbc_3way() argument 55 ivs[0] = src[0]; twofish_dec_blk_cbc_3way() 56 ivs[1] = src[1]; twofish_dec_blk_cbc_3way() 58 twofish_dec_blk_3way(ctx, (u8 *)dst, (u8 *)src); twofish_dec_blk_cbc_3way() 65 void twofish_enc_blk_ctr(void *ctx, u128 *dst, const u128 *src, le128 *iv) twofish_enc_blk_ctr() argument 69 if (dst != src) twofish_enc_blk_ctr() 70 *dst = *src; twofish_enc_blk_ctr() 80 void twofish_enc_blk_ctr_3way(void *ctx, u128 *dst, const u128 *src, twofish_enc_blk_ctr_3way() argument 85 if (dst != src) { twofish_enc_blk_ctr_3way() 86 dst[0] = src[0]; twofish_enc_blk_ctr_3way() 87 dst[1] = src[1]; twofish_enc_blk_ctr_3way() 88 dst[2] = src[2]; twofish_enc_blk_ctr_3way() 155 struct scatterlist *src, unsigned int nbytes) ecb_encrypt() 157 return glue_ecb_crypt_128bit(&twofish_enc, desc, dst, src, nbytes); ecb_encrypt() 161 struct scatterlist *src, unsigned int nbytes) ecb_decrypt() 163 return glue_ecb_crypt_128bit(&twofish_dec, desc, dst, src, nbytes); ecb_decrypt() 167 struct scatterlist *src, unsigned int nbytes) cbc_encrypt() 170 dst, src, nbytes); cbc_encrypt() 174 struct scatterlist *src, unsigned int nbytes) cbc_decrypt() 176 return glue_cbc_decrypt_128bit(&twofish_dec_cbc, desc, dst, src, cbc_decrypt() 181 struct scatterlist *src, unsigned int nbytes) ctr_crypt() 183 return glue_ctr_crypt_128bit(&twofish_ctr, desc, dst, src, nbytes); ctr_crypt() 232 struct scatterlist *src, unsigned int nbytes) lrw_encrypt() 245 return lrw_crypt(desc, dst, src, nbytes, &req); lrw_encrypt() 249 struct scatterlist *src, unsigned int nbytes) lrw_decrypt() 262 return lrw_crypt(desc, dst, src, nbytes, &req); lrw_decrypt() 300 struct scatterlist *src, unsigned int nbytes) xts_encrypt() 314 return xts_crypt(desc, dst, src, nbytes, &req); xts_encrypt() 318 struct scatterlist *src, unsigned int nbytes) xts_decrypt() 332 return xts_crypt(desc, dst, src, nbytes, &req); xts_decrypt() 39 twofish_enc_blk_3way(struct twofish_ctx *ctx, u8 *dst, const u8 *src) twofish_enc_blk_3way() argument 45 twofish_enc_blk_xor_3way(struct twofish_ctx *ctx, u8 *dst, const u8 *src) twofish_enc_blk_xor_3way() argument 154 ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_encrypt() argument 160 ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_decrypt() argument 166 cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_encrypt() argument 173 cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_decrypt() argument 180 ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ctr_crypt() argument 231 lrw_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) lrw_encrypt() argument 248 lrw_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) lrw_decrypt() argument 299 xts_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) xts_encrypt() argument 317 xts_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) xts_decrypt() argument
|
H A D | camellia_aesni_avx_glue.c | 31 const u8 *src); 35 const u8 *src); 39 const u8 *src); 43 const u8 *src, le128 *iv); 47 const u8 *src, le128 *iv); 51 const u8 *src, le128 *iv); 54 void camellia_xts_enc(void *ctx, u128 *dst, const u128 *src, le128 *iv) camellia_xts_enc() argument 56 glue_xts_crypt_128bit_one(ctx, dst, src, iv, camellia_xts_enc() 61 void camellia_xts_dec(void *ctx, u128 *dst, const u128 *src, le128 *iv) camellia_xts_dec() argument 63 glue_xts_crypt_128bit_one(ctx, dst, src, iv, camellia_xts_dec() 159 struct scatterlist *src, unsigned int nbytes) ecb_encrypt() 161 return glue_ecb_crypt_128bit(&camellia_enc, desc, dst, src, nbytes); ecb_encrypt() 165 struct scatterlist *src, unsigned int nbytes) ecb_decrypt() 167 return glue_ecb_crypt_128bit(&camellia_dec, desc, dst, src, nbytes); ecb_decrypt() 171 struct scatterlist *src, unsigned int nbytes) cbc_encrypt() 174 dst, src, nbytes); cbc_encrypt() 178 struct scatterlist *src, unsigned int nbytes) cbc_decrypt() 180 return glue_cbc_decrypt_128bit(&camellia_dec_cbc, desc, dst, src, cbc_decrypt() 185 struct scatterlist *src, unsigned int nbytes) ctr_crypt() 187 return glue_ctr_crypt_128bit(&camellia_ctr, desc, dst, src, nbytes); ctr_crypt() 263 struct scatterlist *src, unsigned int nbytes) lrw_encrypt() 282 ret = lrw_crypt(desc, dst, src, nbytes, &req); lrw_encrypt() 289 struct scatterlist *src, unsigned int nbytes) lrw_decrypt() 308 ret = lrw_crypt(desc, dst, src, nbytes, &req); lrw_decrypt() 315 struct scatterlist *src, unsigned int nbytes) xts_encrypt() 319 return glue_xts_crypt_128bit(&camellia_enc_xts, desc, dst, src, nbytes, xts_encrypt() 325 struct scatterlist *src, unsigned int nbytes) xts_decrypt() 329 return glue_xts_crypt_128bit(&camellia_dec_xts, desc, dst, src, nbytes, xts_decrypt() 158 ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_encrypt() argument 164 ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_decrypt() argument 170 cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_encrypt() argument 177 cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_decrypt() argument 184 ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ctr_crypt() argument 262 lrw_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) lrw_encrypt() argument 288 lrw_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) lrw_decrypt() argument 314 xts_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) xts_encrypt() argument 324 xts_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) xts_decrypt() argument
|
H A D | serpent_avx2_glue.c | 32 const u8 *src); 34 const u8 *src); 35 asmlinkage void serpent_cbc_dec_16way(void *ctx, u128 *dst, const u128 *src); 37 asmlinkage void serpent_ctr_16way(void *ctx, u128 *dst, const u128 *src, 40 const u8 *src, le128 *iv); 42 const u8 *src, le128 *iv); 141 struct scatterlist *src, unsigned int nbytes) ecb_encrypt() 143 return glue_ecb_crypt_128bit(&serpent_enc, desc, dst, src, nbytes); ecb_encrypt() 147 struct scatterlist *src, unsigned int nbytes) ecb_decrypt() 149 return glue_ecb_crypt_128bit(&serpent_dec, desc, dst, src, nbytes); ecb_decrypt() 153 struct scatterlist *src, unsigned int nbytes) cbc_encrypt() 156 dst, src, nbytes); cbc_encrypt() 160 struct scatterlist *src, unsigned int nbytes) cbc_decrypt() 162 return glue_cbc_decrypt_128bit(&serpent_dec_cbc, desc, dst, src, cbc_decrypt() 167 struct scatterlist *src, unsigned int nbytes) ctr_crypt() 169 return glue_ctr_crypt_128bit(&serpent_ctr, desc, dst, src, nbytes); ctr_crypt() 237 struct scatterlist *src, unsigned int nbytes) lrw_encrypt() 256 ret = lrw_crypt(desc, dst, src, nbytes, &req); lrw_encrypt() 263 struct scatterlist *src, unsigned int nbytes) lrw_decrypt() 282 ret = lrw_crypt(desc, dst, src, nbytes, &req); lrw_decrypt() 289 struct scatterlist *src, unsigned int nbytes) xts_encrypt() 293 return glue_xts_crypt_128bit(&serpent_enc_xts, desc, dst, src, nbytes, xts_encrypt() 299 struct scatterlist *src, unsigned int nbytes) xts_decrypt() 303 return glue_xts_crypt_128bit(&serpent_dec_xts, desc, dst, src, nbytes, xts_decrypt() 140 ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_encrypt() argument 146 ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_decrypt() argument 152 cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_encrypt() argument 159 cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_decrypt() argument 166 ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ctr_crypt() argument 236 lrw_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) lrw_encrypt() argument 262 lrw_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) lrw_decrypt() argument 288 xts_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) xts_encrypt() argument 298 xts_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) xts_decrypt() argument
|
/linux-4.1.27/arch/arm64/lib/ |
H A D | memmove.S | 10 * files/head:/src/aarch64/ 30 * Move a buffer from src to test (alignment handled by the hardware). 31 * If dest <= src, call memcpy, otherwise copy in reverse order. 35 * x1 - src 41 src .req x1 label 61 cmp dstin, src 63 add tmp1, src, count 68 add src, src, count 72 ands tmp2, src, #15 /* Bytes to reach alignment. */ 76 * process the aligned offset length to make the src aligned firstly. 81 ldrb tmp1w, [src, #-1]! 85 ldrh tmp1w, [src, #-2]! 89 ldr tmp1w, [src, #-4]! 93 ldr tmp1, [src, #-8]! 114 ldp A_l, A_h, [src, #-16]! 117 ldp A_l, A_h, [src, #-16]! 120 ldp A_l, A_h, [src, #-16]! 125 ldr tmp1, [src, #-8]! 129 ldr tmp1w, [src, #-4]! 133 ldrh tmp1w, [src, #-2]! 137 ldrb tmp1w, [src, #-1] 150 ldp A_l, A_h, [src, #-16] 152 ldp B_l, B_h, [src, #-32] 153 ldp C_l, C_h, [src, #-48] 156 ldp D_l, D_h, [src, #-64]! 170 ldp A_l, A_h, [src, #-16] 171 ldp B_l, B_h, [src, #-32] 172 ldp C_l, C_h, [src, #-48] 173 ldp D_l, D_h, [src, #-64]! 180 ldp A_l, A_h, [src, #-16] 182 ldp B_l, B_h, [src, #-32] 184 ldp C_l, C_h, [src, #-48] 186 ldp D_l, D_h, [src, #-64]!
|
H A D | memcpy.S | 10 * files/head:/src/aarch64/ 30 * Copy a buffer from src to dest (alignment handled by the hardware) 34 * x1 - src 40 src .req x1 label 65 neg tmp2, src 70 * Copy the leading memory data from src to dst in an increasing 72 * memory data is eliminated when the distance between src and 76 ldrb tmp1w, [src], #1 80 ldrh tmp1w, [src], #2 84 ldr tmp1w, [src], #4 88 ldr tmp1, [src],#8 108 ldp A_l, A_h, [src], #16 111 ldp A_l, A_h, [src], #16 114 ldp A_l, A_h, [src], #16 120 * bytes from (src-16) to (dst-16) and to backward the src to aligned 123 * precondition that src address is at least 16 bytes bigger than dst 129 ldr tmp1, [src], #8 133 ldr tmp1w, [src], #4 137 ldrh tmp1w, [src], #2 141 ldrb tmp1w, [src] 154 ldp A_l, A_h, [src],#16 156 ldp B_l, B_h, [src],#16 157 ldp C_l, C_h, [src],#16 160 ldp D_l, D_h, [src],#16 174 ldp A_l, A_h, [src],#16 175 ldp B_l, B_h, [src],#16 176 ldp C_l, C_h, [src],#16 177 ldp D_l, D_h, [src],#16 184 ldp A_l, A_h, [src],#16 186 ldp B_l, B_h, [src],#16 188 ldp C_l, C_h, [src],#16 190 ldp D_l, D_h, [src],#16
|
/linux-4.1.27/drivers/video/fbdev/core/ |
H A D | syscopyarea.c | 29 const unsigned long *src, unsigned src_idx, int bits, unsigned n) bitcpy() 44 *dst = comp(*src, *dst, first); bitcpy() 49 *dst = comp(*src, *dst, first); bitcpy() 51 src++; bitcpy() 58 *dst++ = *src++; bitcpy() 59 *dst++ = *src++; bitcpy() 60 *dst++ = *src++; bitcpy() 61 *dst++ = *src++; bitcpy() 62 *dst++ = *src++; bitcpy() 63 *dst++ = *src++; bitcpy() 64 *dst++ = *src++; bitcpy() 65 *dst++ = *src++; bitcpy() 69 *dst++ = *src++; bitcpy() 73 *dst = comp(*src, *dst, last); bitcpy() 89 *dst = comp(*src << left, *dst, first); bitcpy() 92 *dst = comp(*src >> right, *dst, first); bitcpy() 95 d0 = *src++; bitcpy() 96 d1 = *src; bitcpy() 108 d0 = *src++; bitcpy() 117 d1 = *src++; bitcpy() 129 d1 = *src++; bitcpy() 132 d1 = *src++; bitcpy() 135 d1 = *src++; bitcpy() 138 d1 = *src++; bitcpy() 144 d1 = *src++; bitcpy() 156 d1 = *src; bitcpy() 171 const unsigned long *src, unsigned src_idx, unsigned bits, bitcpy_rev() 178 src += (src_idx + n - 1) / bits; bitcpy_rev() 193 *dst = comp(*src, *dst, last); bitcpy_rev() 199 *dst = comp(*src, *dst, first); bitcpy_rev() 201 src--; bitcpy_rev() 208 *dst-- = *src--; bitcpy_rev() 209 *dst-- = *src--; bitcpy_rev() 210 *dst-- = *src--; bitcpy_rev() 211 *dst-- = *src--; bitcpy_rev() 212 *dst-- = *src--; bitcpy_rev() 213 *dst-- = *src--; bitcpy_rev() 214 *dst-- = *src--; bitcpy_rev() 215 *dst-- = *src--; bitcpy_rev() 219 *dst-- = *src--; bitcpy_rev() 222 *dst = comp(*src, *dst, last); bitcpy_rev() 236 *dst = comp(*src >> right, *dst, last); bitcpy_rev() 239 *dst = comp(*src << left, *dst, last); bitcpy_rev() 242 *dst = comp(*src << left | *(src-1) >> right, bitcpy_rev() 256 d0 = *src--; bitcpy_rev() 264 d1 = *src--; bitcpy_rev() 279 d1 = *src--; bitcpy_rev() 282 d1 = *src--; bitcpy_rev() 285 d1 = *src--; bitcpy_rev() 288 d1 = *src--; bitcpy_rev() 294 d1 = *src--; bitcpy_rev() 306 d1 = *src; bitcpy_rev() 28 bitcpy(struct fb_info *p, unsigned long *dst, unsigned dst_idx, const unsigned long *src, unsigned src_idx, int bits, unsigned n) bitcpy() argument 170 bitcpy_rev(struct fb_info *p, unsigned long *dst, unsigned dst_idx, const unsigned long *src, unsigned src_idx, unsigned bits, unsigned n) bitcpy_rev() argument
|
H A D | cfbcopyarea.c | 47 const unsigned long __iomem *src, unsigned src_idx, int bits, bitcpy() 59 (char *)src + ((src_idx & (bits - 1))) / 8, n / 8); bitcpy() 73 FB_WRITEL( comp( FB_READL(src), FB_READL(dst), first), dst); bitcpy() 79 FB_WRITEL( comp( FB_READL(src), FB_READL(dst), first), dst); bitcpy() 81 src++; bitcpy() 88 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 89 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 90 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 91 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 92 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 93 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 94 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 95 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 99 FB_WRITEL(FB_READL(src++), dst++); bitcpy() 103 FB_WRITEL( comp( FB_READL(src), FB_READL(dst), last), dst); bitcpy() 117 d0 = FB_READL(src); bitcpy() 127 d1 = FB_READL(src + 1); bitcpy() 140 d0 = FB_READL(src++); bitcpy() 150 d1 = FB_READL(src++); bitcpy() 165 d1 = FB_READL(src++); bitcpy() 168 d1 = FB_READL(src++); bitcpy() 171 d1 = FB_READL(src++); bitcpy() 174 d1 = FB_READL(src++); bitcpy() 180 d1 = FB_READL(src++); bitcpy() 195 d1 = FB_READL(src); bitcpy() 213 const unsigned long __iomem *src, unsigned src_idx, int bits, bitcpy_rev() 225 (char *)src + ((src_idx & (bits - 1))) / 8, n / 8); bitcpy_rev() 230 src += (src_idx + n - 1) / bits; bitcpy_rev() 246 FB_WRITEL( comp( FB_READL(src), FB_READL(dst), last), dst); bitcpy_rev() 252 FB_WRITEL( comp( FB_READL(src), FB_READL(dst), first), dst); bitcpy_rev() 254 src--; bitcpy_rev() 261 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 262 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 263 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 264 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 265 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 266 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 267 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 268 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 272 FB_WRITEL(FB_READL(src--), dst--); bitcpy_rev() 276 FB_WRITEL( comp( FB_READL(src), FB_READL(dst), last), dst); bitcpy_rev() 290 d0 = FB_READL(src); bitcpy_rev() 299 d1 = FB_READL(src - 1); bitcpy_rev() 313 d0 = FB_READL(src--); bitcpy_rev() 322 d1 = FB_READL(src--); bitcpy_rev() 339 d1 = FB_READL(src--); bitcpy_rev() 342 d1 = FB_READL(src--); bitcpy_rev() 345 d1 = FB_READL(src--); bitcpy_rev() 348 d1 = FB_READL(src--); bitcpy_rev() 354 d1 = FB_READL(src--); bitcpy_rev() 369 d1 = FB_READL(src); bitcpy_rev() 46 bitcpy(struct fb_info *p, unsigned long __iomem *dst, unsigned dst_idx, const unsigned long __iomem *src, unsigned src_idx, int bits, unsigned n, u32 bswapmask) bitcpy() argument 212 bitcpy_rev(struct fb_info *p, unsigned long __iomem *dst, unsigned dst_idx, const unsigned long __iomem *src, unsigned src_idx, int bits, unsigned n, u32 bswapmask) bitcpy_rev() argument
|
/linux-4.1.27/arch/blackfin/mm/ |
H A D | maccess.c | 19 long probe_kernel_read(void *dst, const void *src, size_t size) probe_kernel_read() argument 21 unsigned long lsrc = (unsigned long)src; probe_kernel_read() 30 u16 mmr = bfin_read16(src); probe_kernel_read() 34 u32 mmr = bfin_read32(src); probe_kernel_read() 42 return __probe_kernel_read(dst, src, size); probe_kernel_read() 45 if (dma_memcpy(dst, src, size)) probe_kernel_read() 49 if (isram_memcpy(dst, src, size)) probe_kernel_read() 58 long probe_kernel_write(void *dst, const void *src, size_t size) probe_kernel_write() argument 70 memcpy(&mmr, src, sizeof(mmr)); probe_kernel_write() 75 memcpy(&mmr, src, sizeof(mmr)); probe_kernel_write() 83 return __probe_kernel_write(dst, src, size); probe_kernel_write() 86 if (dma_memcpy(dst, src, size)) probe_kernel_write() 90 if (isram_memcpy(dst, src, size)) probe_kernel_write()
|
/linux-4.1.27/arch/xtensa/boot/lib/ |
H A D | zmem.c | 34 void gunzip (void *dst, int dstlen, unsigned char *src, int *lenp) gunzip() argument 41 flags = src[3]; gunzip() 42 if (src[2] != DEFLATED || (flags & RESERVED) != 0) { gunzip() 47 i = 12 + src[10] + (src[11] << 8); gunzip() 49 while (src[i++] != 0) gunzip() 52 while (src[i++] != 0) gunzip() 67 s.next_in = src + i; gunzip()
|
/linux-4.1.27/arch/sparc/lib/ |
H A D | csum_copy_to_user.S | 19 #define STORE(type,src,addr) type##a src, [addr] %asi
|
H A D | PeeCeeI.c | 12 void outsb(unsigned long __addr, const void *src, unsigned long count) outsb() argument 15 const u8 *p = src; outsb() 22 void outsw(unsigned long __addr, const void *src, unsigned long count) outsw() argument 27 __raw_writew(*(u16 *)src, addr); outsw() 28 src += sizeof(u16); outsw() 33 void outsl(unsigned long __addr, const void *src, unsigned long count) outsl() argument 41 switch (((unsigned long)src) & 0x3) { outsl() 43 /* src is naturally aligned */ outsl() 45 __raw_writel(*(u32 *)src, addr); outsl() 46 src += sizeof(u32); outsl() 52 l = (*(u16 *)src) << 16; outsl() 53 l |= *(u16 *)(src + sizeof(u16)); outsl() 55 src += sizeof(u32); outsl() 60 l = (*(u8 *)src) << 24; outsl() 61 l |= (*(u16 *)(src + sizeof(u8))) << 8; outsl() 62 src += sizeof(u8) + sizeof(u16); outsl() 64 l2 = *(u32 *)src; outsl() 68 src += sizeof(u32); outsl() 73 l = (*(u8 *)src) << 24; outsl() 74 src += sizeof(u8); outsl() 76 l2 = *(u32 *)src; outsl() 80 src += sizeof(u32); outsl()
|
H A D | U1copy_to_user.S | 15 #define STORE(type,src,addr) type##a src, [addr] ASI_AIUS 16 #define STORE_BLK(src,addr) stda src, [addr] ASI_BLK_AIUS
|
H A D | U3copy_to_user.S | 15 #define STORE(type,src,addr) type##a src, [addr] ASI_AIUS 16 #define STORE_BLK(src,addr) stda src, [addr] ASI_BLK_AIUS
|
H A D | NG2copy_to_user.S | 27 #define STORE(type,src,addr) type##a src, [addr] ASI_AIUS 29 #define STORE_BLK(src,addr) stda src, [addr] ASI_BLK_AIUS_4V
|
H A D | memcpy.S | 17 #define MOVE_BIGCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ 18 ldd [%src + (offset) + 0x00], %t0; \ 19 ldd [%src + (offset) + 0x08], %t2; \ 20 ldd [%src + (offset) + 0x10], %t4; \ 21 ldd [%src + (offset) + 0x18], %t6; \ 31 #define MOVE_BIGALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ 32 ldd [%src + (offset) + 0x00], %t0; \ 33 ldd [%src + (offset) + 0x08], %t2; \ 34 ldd [%src + (offset) + 0x10], %t4; \ 35 ldd [%src + (offset) + 0x18], %t6; \ 41 #define MOVE_LASTCHUNK(src, dst, offset, t0, t1, t2, t3) \ 42 ldd [%src - (offset) - 0x10], %t0; \ 43 ldd [%src - (offset) - 0x08], %t2; \ 49 #define MOVE_LASTALIGNCHUNK(src, dst, offset, t0, t1, t2, t3) \ 50 ldd [%src - (offset) - 0x10], %t0; \ 51 ldd [%src - (offset) - 0x08], %t2; \ 55 #define MOVE_SHORTCHUNK(src, dst, offset, t0, t1) \ 56 ldub [%src - (offset) - 0x02], %t0; \ 57 ldub [%src - (offset) - 0x01], %t1; \ 62 #define RMOVE_BIGCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ 63 ldd [%src - (offset) - 0x20], %t0; \ 64 ldd [%src - (offset) - 0x18], %t2; \ 65 ldd [%src - (offset) - 0x10], %t4; \ 66 ldd [%src - (offset) - 0x08], %t6; \ 76 #define RMOVE_BIGALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ 77 ldd [%src - (offset) - 0x20], %t0; \ 78 ldd [%src - (offset) - 0x18], %t2; \ 79 ldd [%src - (offset) - 0x10], %t4; \ 80 ldd [%src - (offset) - 0x08], %t6; \ 86 #define RMOVE_LASTCHUNK(src, dst, offset, t0, t1, t2, t3) \ 87 ldd [%src + (offset) + 0x00], %t0; \ 88 ldd [%src + (offset) + 0x08], %t2; \ 94 #define RMOVE_SHORTCHUNK(src, dst, offset, t0, t1) \ 95 ldub [%src + (offset) + 0x00], %t0; \ 96 ldub [%src + (offset) + 0x01], %t1; \ 100 #define SMOVE_CHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \ 101 ldd [%src + (offset) + 0x00], %t0; \ 102 ldd [%src + (offset) + 0x08], %t2; \ 118 #define SMOVE_ALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \ 119 ldd [%src + (offset) + 0x00], %t0; \ 120 ldd [%src + (offset) + 0x08], %t2; \ 176 where %src (=%o1) & 3 is != 0. 177 We need to align it to 4. So, for (%src & 3) 204 FUNC(memcpy) /* %o0=dst %o1=src %o2=len */
|
H A D | GENcopy_to_user.S | 19 #define STORE(type,src,addr) type##a src, [addr] ASI_AIUS
|
H A D | NGcopy_to_user.S | 19 #define STORE(type,src,addr) type##a src, [addr] ASI_AIUS
|
H A D | NG4copy_to_user.S | 23 #define STORE(type,src,addr) type##a src, [addr] %asi
|
/linux-4.1.27/arch/ia64/lib/ |
H A D | Makefile | 28 $(obj)/__divdi3.o: $(src)/idiv64.S FORCE 31 $(obj)/__udivdi3.o: $(src)/idiv64.S FORCE 34 $(obj)/__moddi3.o: $(src)/idiv64.S FORCE 37 $(obj)/__umoddi3.o: $(src)/idiv64.S FORCE 40 $(obj)/__divsi3.o: $(src)/idiv32.S FORCE 43 $(obj)/__udivsi3.o: $(src)/idiv32.S FORCE 46 $(obj)/__modsi3.o: $(src)/idiv32.S FORCE 49 $(obj)/__umodsi3.o: $(src)/idiv32.S FORCE
|
H A D | memcpy.S | 23 # define src r3 41 * First, check if everything (src, dst, len) is a multiple of eight. If 81 mov src=in1 // copy because of rotation 96 (p[0]) ld8 val[0]=[src],8 140 (p[0]) ld1 val[0]=[src],1 183 and t0=-8,src // t0 = src & ~7 184 and t2=7,src // t2 = src & 7 187 adds src2=7,src // src2 = (src + 7) 190 and src2=-8,src2 // src2 = (src + 7) & ~7 191 shl t2=t2,3 // t2 = 8*(src & 7) 194 ld8 t1=[src2] // t1 = 1st source word if src is 8-byte aligned, 2nd otherwise 195 sub t3=64,t2 // t3 = 64-8*(src & 7) 198 add src_end=src,in2 221 1:{ add src=cnt,src // make src point to remainder of source buffer 225 and src2=-8,src // align source pointer 229 and t0=7,src // t0 = src & 7 237 cmp.ne p6,p0=t0,r0 // is src aligned, too? 238 shl t0=t0,LOG_LOOP_SIZE // t0 = 8*(src & 7) 259 // less than 8) and t0 contains the last few bytes of the src buffer:
|
/linux-4.1.27/arch/x86/boot/compressed/ |
H A D | string.c | 4 void *memcpy(void *dest, const void *src, size_t n) memcpy() argument 12 : "0" (n >> 2), "g" (n & 3), "1" (dest), "2" (src) memcpy() 18 void *memcpy(void *dest, const void *src, size_t n) memcpy() argument 26 : "0" (n >> 3), "g" (n & 7), "1" (dest), "2" (src) memcpy()
|
/linux-4.1.27/net/wireless/ |
H A D | Makefile | 18 CFLAGS_trace.o := -I$(src) 22 $(obj)/regdb.c: $(src)/db.txt $(src)/genregdb.awk 23 @$(AWK) -f $(srctree)/$(src)/genregdb.awk < $< > $@
|
/linux-4.1.27/drivers/scsi/aic7xxx/ |
H A D | Makefile | 57 $(obj)/aic7xxx_seq.h: $(src)/aic7xxx.seq $(src)/aic7xxx.reg $(obj)/aicasm/aicasm 58 $(obj)/aicasm/aicasm -I$(src) -r $(obj)/aic7xxx_reg.h \ 60 $(src)/aic7xxx.seq 64 $(obj)/aic7xxx_reg_print.c: $(src)/aic7xxx_reg_print.c_shipped 74 $(obj)/aic79xx_seq.h: $(src)/aic79xx.seq $(src)/aic79xx.reg $(obj)/aicasm/aicasm 75 $(obj)/aicasm/aicasm -I$(src) -r $(obj)/aic79xx_reg.h \ 77 $(src)/aic79xx.seq 81 $(obj)/aic79xx_reg_print.c: $(src)/aic79xx_reg_print.c_shipped 84 $(obj)/aicasm/aicasm: $(src)/aicasm/*.[chyl] 85 $(MAKE) -C $(src)/aicasm
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | ftrace.h | 22 #define safe_load(load, src, dst, error) \ 39 : [tmp_src] "r" (src) \ 44 #define safe_store(store, src, dst, error) \ 61 : [tmp_dst] "r" (dst), [tmp_src] "r" (src)\ 66 #define safe_load_code(dst, src, error) \ 67 safe_load(STR(lw), src, dst, error) 68 #define safe_store_code(src, dst, error) \ 69 safe_store(STR(sw), src, dst, error) 71 #define safe_load_stack(dst, src, error) \ 72 safe_load(STR(PTR_L), src, dst, error) 74 #define safe_store_stack(src, dst, error) \ 75 safe_store(STR(PTR_S), src, dst, error)
|
/linux-4.1.27/arch/powerpc/boot/ |
H A D | string.h | 5 extern char *strcpy(char *dest, const char *src); 6 extern char *strncpy(char *dest, const char *src, size_t n); 7 extern char *strcat(char *dest, const char *src); 15 extern void *memmove(void *dest, const void *src, unsigned long n); 16 extern void *memcpy(void *dest, const void *src, unsigned long n);
|
/linux-4.1.27/arch/frv/kernel/ |
H A D | uaccess.c | 20 long strncpy_from_user(char *dst, const char __user *src, long count) strncpy_from_user() argument 31 if ((unsigned long) src < memory_start) strncpy_from_user() 35 if ((unsigned long) src >= get_addr_limit()) strncpy_from_user() 38 max = get_addr_limit() - (unsigned long) src; strncpy_from_user() 45 for (; count > 0; count--, p++, src++) { strncpy_from_user() 46 __get_user_asm(err, ch, src, "ub", "=r"); strncpy_from_user() 72 long strnlen_user(const char __user *src, long count) strnlen_user() argument 81 if ((unsigned long) src < memory_start) strnlen_user() 85 if ((unsigned long) src >= get_addr_limit()) strnlen_user() 88 for (p = src; count > 0; count--, p++) { strnlen_user() 96 return p - src + 1; /* return length including NUL */ strnlen_user()
|
/linux-4.1.27/lib/lzo/ |
H A D | lzodefs.h | 15 #define COPY4(dst, src) \ 16 put_unaligned(get_unaligned((const u32 *)(src)), (u32 *)(dst)) 18 #define COPY8(dst, src) \ 19 put_unaligned(get_unaligned((const u64 *)(src)), (u64 *)(dst)) 21 #define COPY8(dst, src) \ 22 COPY4(dst, src); COPY4((dst) + 4, (src) + 4)
|
/linux-4.1.27/arch/hexagon/mm/ |
H A D | copy_user_template.S | 34 r3 = or(dst,src) 35 r4 = xor(dst,src) 41 src_dst_sav = combine(src,dst) 54 d_dbuf = memd(src++#8) 84 w_dbuf = memw(src++#4) 110 w_dbuf = memuh(src++#2) 126 w_dbuf = memub(src++#1) 150 if (p0.new) w_dbuf = memub(src) 151 p0 = tstbit(src,#0) 158 if (p0) src = add(src,#1) 162 if (p0.new) w_dbuf = memuh(src) 163 p0 = tstbit(src,#1) 170 if (p0) src = add(src,#2) 174 if (p0.new) w_dbuf = memw(src) 175 p0 = tstbit(src,#2) 182 if (p0) src = add(src,#4)
|
H A D | copy_from_user.S | 23 * The easy way to do this is only speed up src/dest similar alignment. 43 #define src r1 define 61 /* Normal copy loops. Do epilog. Use src-src_sav to compute distance */ 67 r2 += sub(src_sav,src) 77 r2 += sub(src_sav,src) 87 r2 += sub(src_sav,src) 97 r2 += sub(src_sav,src)
|
/linux-4.1.27/arch/x86/include/asm/ |
H A D | uaccess_64.h | 52 int __copy_from_user_nocheck(void *dst, const void __user *src, unsigned size) __copy_from_user_nocheck() argument 57 return copy_user_generic(dst, (__force void *)src, size); __copy_from_user_nocheck() 59 case 1:__get_user_asm(*(u8 *)dst, (u8 __user *)src, __copy_from_user_nocheck() 62 case 2:__get_user_asm(*(u16 *)dst, (u16 __user *)src, __copy_from_user_nocheck() 65 case 4:__get_user_asm(*(u32 *)dst, (u32 __user *)src, __copy_from_user_nocheck() 68 case 8:__get_user_asm(*(u64 *)dst, (u64 __user *)src, __copy_from_user_nocheck() 72 __get_user_asm(*(u64 *)dst, (u64 __user *)src, __copy_from_user_nocheck() 77 (u16 __user *)(8 + (char __user *)src), __copy_from_user_nocheck() 81 __get_user_asm(*(u64 *)dst, (u64 __user *)src, __copy_from_user_nocheck() 86 (u64 __user *)(8 + (char __user *)src), __copy_from_user_nocheck() 90 return copy_user_generic(dst, (__force void *)src, size); __copy_from_user_nocheck() 95 int __copy_from_user(void *dst, const void __user *src, unsigned size) __copy_from_user() argument 98 return __copy_from_user_nocheck(dst, src, size); __copy_from_user() 102 int __copy_to_user_nocheck(void __user *dst, const void *src, unsigned size) __copy_to_user_nocheck() argument 107 return copy_user_generic((__force void *)dst, src, size); __copy_to_user_nocheck() 109 case 1:__put_user_asm(*(u8 *)src, (u8 __user *)dst, __copy_to_user_nocheck() 112 case 2:__put_user_asm(*(u16 *)src, (u16 __user *)dst, __copy_to_user_nocheck() 115 case 4:__put_user_asm(*(u32 *)src, (u32 __user *)dst, __copy_to_user_nocheck() 118 case 8:__put_user_asm(*(u64 *)src, (u64 __user *)dst, __copy_to_user_nocheck() 122 __put_user_asm(*(u64 *)src, (u64 __user *)dst, __copy_to_user_nocheck() 127 __put_user_asm(4[(u16 *)src], 4 + (u16 __user *)dst, __copy_to_user_nocheck() 131 __put_user_asm(*(u64 *)src, (u64 __user *)dst, __copy_to_user_nocheck() 136 __put_user_asm(1[(u64 *)src], 1 + (u64 __user *)dst, __copy_to_user_nocheck() 140 return copy_user_generic((__force void *)dst, src, size); __copy_to_user_nocheck() 145 int __copy_to_user(void __user *dst, const void *src, unsigned size) __copy_to_user() argument 148 return __copy_to_user_nocheck(dst, src, size); __copy_to_user() 152 int __copy_in_user(void __user *dst, const void __user *src, unsigned size) __copy_in_user() argument 159 (__force void *)src, size); __copy_in_user() 163 __get_user_asm(tmp, (u8 __user *)src, __copy_in_user() 172 __get_user_asm(tmp, (u16 __user *)src, __copy_in_user() 182 __get_user_asm(tmp, (u32 __user *)src, __copy_in_user() 191 __get_user_asm(tmp, (u64 __user *)src, __copy_in_user() 200 (__force void *)src, size); __copy_in_user() 205 __copy_from_user_inatomic(void *dst, const void __user *src, unsigned size) __copy_from_user_inatomic() argument 207 return __copy_from_user_nocheck(dst, src, size); __copy_from_user_inatomic() 211 __copy_to_user_inatomic(void __user *dst, const void *src, unsigned size) __copy_to_user_inatomic() argument 213 return __copy_to_user_nocheck(dst, src, size); __copy_to_user_inatomic() 216 extern long __copy_user_nocache(void *dst, const void __user *src, 220 __copy_from_user_nocache(void *dst, const void __user *src, unsigned size) __copy_from_user_nocache() argument 223 return __copy_user_nocache(dst, src, size, 1); __copy_from_user_nocache() 227 __copy_from_user_inatomic_nocache(void *dst, const void __user *src, __copy_from_user_inatomic_nocache() argument 230 return __copy_user_nocache(dst, src, size, 0); __copy_from_user_inatomic_nocache()
|
H A D | string_64.h | 36 #define memcpy(dst, src, len) \ 41 __ret = __memcpy((dst), (src), __len); \ 43 __ret = __builtin_memcpy((dst), (src), __len); \ 52 #define memcpy(dst, src, len) __inline_memcpy((dst), (src), (len)) 60 void *memmove(void *dest, const void *src, size_t count); 61 void *__memmove(void *dest, const void *src, size_t count); 65 char *strcpy(char *dest, const char *src); 66 char *strcat(char *dest, const char *src); 77 #define memcpy(dst, src, len) __memcpy(dst, src, len) 78 #define memmove(dst, src, len) __memmove(dst, src, len)
|
H A D | pvclock.h | 8 cycle_t pvclock_clocksource_read(struct pvclock_vcpu_time_info *src); 9 u8 pvclock_read_flags(struct pvclock_vcpu_time_info *src); 11 unsigned long pvclock_tsc_khz(struct pvclock_vcpu_time_info *src); 63 u64 pvclock_get_nsec_offset(const struct pvclock_vcpu_time_info *src) pvclock_get_nsec_offset() argument 65 u64 delta = __native_read_tsc() - src->tsc_timestamp; pvclock_get_nsec_offset() 66 return pvclock_scale_delta(delta, src->tsc_to_system_mul, pvclock_get_nsec_offset() 67 src->tsc_shift); pvclock_get_nsec_offset() 71 unsigned __pvclock_read_cycles(const struct pvclock_vcpu_time_info *src, __pvclock_read_cycles() argument 78 version = src->version; __pvclock_read_cycles() 86 offset = pvclock_get_nsec_offset(src); __pvclock_read_cycles() 87 ret = src->system_time + offset; __pvclock_read_cycles() 88 ret_flags = src->flags; __pvclock_read_cycles()
|
/linux-4.1.27/mm/ |
H A D | maccess.c | 11 * @src: address to read from 14 * Safely read from address @src to the buffer at @dst. If a kernel fault 18 long __weak probe_kernel_read(void *dst, const void *src, size_t size) 21 long __probe_kernel_read(void *dst, const void *src, size_t size) __probe_kernel_read() argument 29 (__force const void __user *)src, size); __probe_kernel_read() 40 * @src: pointer to the data that shall be written 43 * Safely write to address @dst from the buffer at @src. If a kernel fault 46 long __weak probe_kernel_write(void *dst, const void *src, size_t size) 49 long __probe_kernel_write(void *dst, const void *src, size_t size) __probe_kernel_write() argument 56 ret = __copy_to_user_inatomic((__force void __user *)dst, src, size); __probe_kernel_write()
|
/linux-4.1.27/drivers/phy/ |
H A D | phy-xgene.c | 97 #define SATA_MEM_RESET_RD(src) (((src) & 0x00000020) >> 5) 106 #define REGSPEC_CFG_I_TX_WORDMODE0_SET(dst, src) \ 107 (((dst) & ~0x00070000) | (((u32) (src) << 16) & 0x00070000)) 108 #define REGSPEC_CFG_I_RX_WORDMODE0_SET(dst, src) \ 109 (((dst) & ~0x00e00000) | (((u32) (src) << 21) & 0x00e00000)) 111 #define REGSPEC_CFG_I_CUSTOMER_PIN_MODE0_SET(dst, src) \ 112 (((dst) & ~0x00007fff) | (((u32) (src)) & 0x00007fff)) 114 #define CFG_I_SPD_SEL_CDR_OVR1_SET(dst, src) \ 115 (((dst) & ~0x0000000f) | (((u32) (src)) & 0x0000000f)) 121 #define CFG_IND_ADDR_SET(dst, src) \ 122 (((dst) & ~0x003ffff0) | (((u32) (src) << 4) & 0x003ffff0)) 126 #define I_RESET_B_SET(dst, src) \ 127 (((dst) & ~0x00000001) | (((u32) (src)) & 0x00000001)) 128 #define I_PLL_FBDIV_SET(dst, src) \ 129 (((dst) & ~0x001ff000) | (((u32) (src) << 12) & 0x001ff000)) 130 #define I_CUSTOMEROV_SET(dst, src) \ 131 (((dst) & ~0x00000f80) | (((u32) (src) << 7) & 0x00000f80)) 132 #define O_PLL_LOCK_RD(src) (((src) & 0x40000000) >> 30) 133 #define O_PLL_READY_RD(src) (((src) & 0x80000000) >> 31) 138 #define CMU_REG0_PLL_REF_SEL_SET(dst, src) \ 139 (((dst) & ~0x00002000) | (((u32) (src) << 13) & 0x00002000)) 141 #define CMU_REG0_CAL_COUNT_RESOL_SET(dst, src) \ 142 (((dst) & ~0x000000e0) | (((u32) (src) << 5) & 0x000000e0)) 144 #define CMU_REG1_PLL_CP_SET(dst, src) \ 145 (((dst) & ~0x00003c00) | (((u32) (src) << 10) & 0x00003c00)) 146 #define CMU_REG1_PLL_MANUALCAL_SET(dst, src) \ 147 (((dst) & ~0x00000008) | (((u32) (src) << 3) & 0x00000008)) 148 #define CMU_REG1_PLL_CP_SEL_SET(dst, src) \ 149 (((dst) & ~0x000003e0) | (((u32) (src) << 5) & 0x000003e0)) 151 #define CMU_REG1_REFCLK_CMOS_SEL_SET(dst, src) \ 152 (((dst) & ~0x00000001) | (((u32) (src) << 0) & 0x00000001)) 154 #define CMU_REG2_PLL_REFDIV_SET(dst, src) \ 155 (((dst) & ~0x0000c000) | (((u32) (src) << 14) & 0x0000c000)) 156 #define CMU_REG2_PLL_LFRES_SET(dst, src) \ 157 (((dst) & ~0x0000001e) | (((u32) (src) << 1) & 0x0000001e)) 158 #define CMU_REG2_PLL_FBDIV_SET(dst, src) \ 159 (((dst) & ~0x00003fe0) | (((u32) (src) << 5) & 0x00003fe0)) 161 #define CMU_REG3_VCOVARSEL_SET(dst, src) \ 162 (((dst) & ~0x0000000f) | (((u32) (src) << 0) & 0x0000000f)) 163 #define CMU_REG3_VCO_MOMSEL_INIT_SET(dst, src) \ 164 (((dst) & ~0x000003f0) | (((u32) (src) << 4) & 0x000003f0)) 165 #define CMU_REG3_VCO_MANMOMSEL_SET(dst, src) \ 166 (((dst) & ~0x0000fc00) | (((u32) (src) << 10) & 0x0000fc00)) 169 #define CMU_REG5_PLL_LFSMCAP_SET(dst, src) \ 170 (((dst) & ~0x0000c000) | (((u32) (src) << 14) & 0x0000c000)) 171 #define CMU_REG5_PLL_LOCK_RESOLUTION_SET(dst, src) \ 172 (((dst) & ~0x0000000e) | (((u32) (src) << 1) & 0x0000000e)) 173 #define CMU_REG5_PLL_LFCAP_SET(dst, src) \ 174 (((dst) & ~0x00003000) | (((u32) (src) << 12) & 0x00003000)) 177 #define CMU_REG6_PLL_VREGTRIM_SET(dst, src) \ 178 (((dst) & ~0x00000600) | (((u32) (src) << 9) & 0x00000600)) 179 #define CMU_REG6_MAN_PVT_CAL_SET(dst, src) \ 180 (((dst) & ~0x00000004) | (((u32) (src) << 2) & 0x00000004)) 182 #define CMU_REG7_PLL_CALIB_DONE_RD(src) ((0x00004000 & (u32) (src)) >> 14) 183 #define CMU_REG7_VCO_CAL_FAIL_RD(src) ((0x00000c00 & (u32) (src)) >> 10) 194 #define CMU_REG9_TX_WORD_MODE_CH1_SET(dst, src) \ 195 (((dst) & ~0x00000380) | (((u32) (src) << 7) & 0x00000380)) 196 #define CMU_REG9_TX_WORD_MODE_CH0_SET(dst, src) \ 197 (((dst) & ~0x00000070) | (((u32) (src) << 4) & 0x00000070)) 198 #define CMU_REG9_PLL_POST_DIVBY2_SET(dst, src) \ 199 (((dst) & ~0x00000008) | (((u32) (src) << 3) & 0x00000008)) 200 #define CMU_REG9_VBG_BYPASSB_SET(dst, src) \ 201 (((dst) & ~0x00000004) | (((u32) (src) << 2) & 0x00000004)) 202 #define CMU_REG9_IGEN_BYPASS_SET(dst, src) \ 203 (((dst) & ~0x00000002) | (((u32) (src) << 1) & 0x00000002)) 205 #define CMU_REG10_VREG_REFSEL_SET(dst, src) \ 206 (((dst) & ~0x00000001) | (((u32) (src) << 0) & 0x00000001)) 209 #define CMU_REG12_STATE_DELAY9_SET(dst, src) \ 210 (((dst) & ~0x000000f0) | (((u32) (src) << 4) & 0x000000f0)) 217 #define CMU_REG16_VCOCAL_WAIT_BTW_CODE_SET(dst, src) \ 218 (((dst) & ~0x0000001c) | (((u32) (src) << 2) & 0x0000001c)) 219 #define CMU_REG16_CALIBRATION_DONE_OVERRIDE_SET(dst, src) \ 220 (((dst) & ~0x00000040) | (((u32) (src) << 6) & 0x00000040)) 221 #define CMU_REG16_BYPASS_PLL_LOCK_SET(dst, src) \ 222 (((dst) & ~0x00000020) | (((u32) (src) << 5) & 0x00000020)) 224 #define CMU_REG17_PVT_CODE_R2A_SET(dst, src) \ 225 (((dst) & ~0x00007f00) | (((u32) (src) << 8) & 0x00007f00)) 226 #define CMU_REG17_RESERVED_7_SET(dst, src) \ 227 (((dst) & ~0x000000e0) | (((u32) (src) << 5) & 0x000000e0)) 238 #define CMU_REG26_FORCE_PLL_LOCK_SET(dst, src) \ 239 (((dst) & ~0x00000001) | (((u32) (src) << 0) & 0x00000001)) 244 #define CMU_REG30_LOCK_COUNT_SET(dst, src) \ 245 (((dst) & ~0x00000006) | (((u32) (src) << 1) & 0x00000006)) 246 #define CMU_REG30_PCIE_MODE_SET(dst, src) \ 247 (((dst) & ~0x00000008) | (((u32) (src) << 3) & 0x00000008)) 251 #define CMU_REG32_PVT_CAL_WAIT_SEL_SET(dst, src) \ 252 (((dst) & ~0x00000006) | (((u32) (src) << 1) & 0x00000006)) 253 #define CMU_REG32_IREF_ADJ_SET(dst, src) \ 254 (((dst) & ~0x00000180) | (((u32) (src) << 7) & 0x00000180)) 257 #define CMU_REG34_VCO_CAL_VTH_LO_MAX_SET(dst, src) \ 258 (((dst) & ~0x0000000f) | (((u32) (src) << 0) & 0x0000000f)) 259 #define CMU_REG34_VCO_CAL_VTH_HI_MAX_SET(dst, src) \ 260 (((dst) & ~0x00000f00) | (((u32) (src) << 8) & 0x00000f00)) 261 #define CMU_REG34_VCO_CAL_VTH_LO_MIN_SET(dst, src) \ 262 (((dst) & ~0x000000f0) | (((u32) (src) << 4) & 0x000000f0)) 263 #define CMU_REG34_VCO_CAL_VTH_HI_MIN_SET(dst, src) \ 264 (((dst) & ~0x0000f000) | (((u32) (src) << 12) & 0x0000f000)) 266 #define CMU_REG35_PLL_SSC_MOD_SET(dst, src) \ 267 (((dst) & ~0x0000fe00) | (((u32) (src) << 9) & 0x0000fe00)) 269 #define CMU_REG36_PLL_SSC_EN_SET(dst, src) \ 270 (((dst) & ~0x00000010) | (((u32) (src) << 4) & 0x00000010)) 271 #define CMU_REG36_PLL_SSC_VSTEP_SET(dst, src) \ 272 (((dst) & ~0x0000ffc0) | (((u32) (src) << 6) & 0x0000ffc0)) 273 #define CMU_REG36_PLL_SSC_DSMSEL_SET(dst, src) \ 274 (((dst) & ~0x00000020) | (((u32) (src) << 5) & 0x00000020)) 281 #define RXTX_REG0_CTLE_EQ_HR_SET(dst, src) \ 282 (((dst) & ~0x0000f800) | (((u32) (src) << 11) & 0x0000f800)) 283 #define RXTX_REG0_CTLE_EQ_QR_SET(dst, src) \ 284 (((dst) & ~0x000007c0) | (((u32) (src) << 6) & 0x000007c0)) 285 #define RXTX_REG0_CTLE_EQ_FR_SET(dst, src) \ 286 (((dst) & ~0x0000003e) | (((u32) (src) << 1) & 0x0000003e)) 288 #define RXTX_REG1_RXACVCM_SET(dst, src) \ 289 (((dst) & ~0x0000f000) | (((u32) (src) << 12) & 0x0000f000)) 290 #define RXTX_REG1_CTLE_EQ_SET(dst, src) \ 291 (((dst) & ~0x00000f80) | (((u32) (src) << 7) & 0x00000f80)) 292 #define RXTX_REG1_RXVREG1_SET(dst, src) \ 293 (((dst) & ~0x00000060) | (((u32) (src) << 5) & 0x00000060)) 294 #define RXTX_REG1_RXIREF_ADJ_SET(dst, src) \ 295 (((dst) & ~0x00000006) | (((u32) (src) << 1) & 0x00000006)) 297 #define RXTX_REG2_VTT_ENA_SET(dst, src) \ 298 (((dst) & ~0x00000100) | (((u32) (src) << 8) & 0x00000100)) 299 #define RXTX_REG2_TX_FIFO_ENA_SET(dst, src) \ 300 (((dst) & ~0x00000020) | (((u32) (src) << 5) & 0x00000020)) 301 #define RXTX_REG2_VTT_SEL_SET(dst, src) \ 302 (((dst) & ~0x000000c0) | (((u32) (src) << 6) & 0x000000c0)) 305 #define RXTX_REG4_TX_DATA_RATE_SET(dst, src) \ 306 (((dst) & ~0x0000c000) | (((u32) (src) << 14) & 0x0000c000)) 307 #define RXTX_REG4_TX_WORD_MODE_SET(dst, src) \ 308 (((dst) & ~0x00003800) | (((u32) (src) << 11) & 0x00003800)) 310 #define RXTX_REG5_TX_CN1_SET(dst, src) \ 311 (((dst) & ~0x0000f800) | (((u32) (src) << 11) & 0x0000f800)) 312 #define RXTX_REG5_TX_CP1_SET(dst, src) \ 313 (((dst) & ~0x000007e0) | (((u32) (src) << 5) & 0x000007e0)) 314 #define RXTX_REG5_TX_CN2_SET(dst, src) \ 315 (((dst) & ~0x0000001f) | (((u32) (src) << 0) & 0x0000001f)) 317 #define RXTX_REG6_TXAMP_CNTL_SET(dst, src) \ 318 (((dst) & ~0x00000780) | (((u32) (src) << 7) & 0x00000780)) 319 #define RXTX_REG6_TXAMP_ENA_SET(dst, src) \ 320 (((dst) & ~0x00000040) | (((u32) (src) << 6) & 0x00000040)) 321 #define RXTX_REG6_RX_BIST_ERRCNT_RD_SET(dst, src) \ 322 (((dst) & ~0x00000001) | (((u32) (src) << 0) & 0x00000001)) 323 #define RXTX_REG6_TX_IDLE_SET(dst, src) \ 324 (((dst) & ~0x00000008) | (((u32) (src) << 3) & 0x00000008)) 325 #define RXTX_REG6_RX_BIST_RESYNC_SET(dst, src) \ 326 (((dst) & ~0x00000002) | (((u32) (src) << 1) & 0x00000002)) 330 #define RXTX_REG7_BIST_ENA_RX_SET(dst, src) \ 331 (((dst) & ~0x00000040) | (((u32) (src) << 6) & 0x00000040)) 332 #define RXTX_REG7_RX_WORD_MODE_SET(dst, src) \ 333 (((dst) & ~0x00003800) | (((u32) (src) << 11) & 0x00003800)) 335 #define RXTX_REG8_CDR_LOOP_ENA_SET(dst, src) \ 336 (((dst) & ~0x00004000) | (((u32) (src) << 14) & 0x00004000)) 337 #define RXTX_REG8_CDR_BYPASS_RXLOS_SET(dst, src) \ 338 (((dst) & ~0x00000800) | (((u32) (src) << 11) & 0x00000800)) 339 #define RXTX_REG8_SSC_ENABLE_SET(dst, src) \ 340 (((dst) & ~0x00000200) | (((u32) (src) << 9) & 0x00000200)) 341 #define RXTX_REG8_SD_VREF_SET(dst, src) \ 342 (((dst) & ~0x000000f0) | (((u32) (src) << 4) & 0x000000f0)) 343 #define RXTX_REG8_SD_DISABLE_SET(dst, src) \ 344 (((dst) & ~0x00000100) | (((u32) (src) << 8) & 0x00000100)) 346 #define RXTX_REG7_RESETB_RXD_SET(dst, src) \ 347 (((dst) & ~0x00000100) | (((u32) (src) << 8) & 0x00000100)) 348 #define RXTX_REG7_RESETB_RXA_SET(dst, src) \ 349 (((dst) & ~0x00000080) | (((u32) (src) << 7) & 0x00000080)) 351 #define RXTX_REG7_LOOP_BACK_ENA_CTLE_SET(dst, src) \ 352 (((dst) & ~0x00004000) | (((u32) (src) << 14) & 0x00004000)) 354 #define RXTX_REG11_PHASE_ADJUST_LIMIT_SET(dst, src) \ 355 (((dst) & ~0x0000f800) | (((u32) (src) << 11) & 0x0000f800)) 357 #define RXTX_REG12_LATCH_OFF_ENA_SET(dst, src) \ 358 (((dst) & ~0x00002000) | (((u32) (src) << 13) & 0x00002000)) 359 #define RXTX_REG12_SUMOS_ENABLE_SET(dst, src) \ 360 (((dst) & ~0x00000004) | (((u32) (src) << 2) & 0x00000004)) 362 #define RXTX_REG12_RX_DET_TERM_ENABLE_SET(dst, src) \ 363 (((dst) & ~0x00000002) | (((u32) (src) << 1) & 0x00000002)) 366 #define RXTX_REG14_CLTE_LATCAL_MAN_PROG_SET(dst, src) \ 367 (((dst) & ~0x0000003f) | (((u32) (src) << 0) & 0x0000003f)) 368 #define RXTX_REG14_CTLE_LATCAL_MAN_ENA_SET(dst, src) \ 369 (((dst) & ~0x00000040) | (((u32) (src) << 6) & 0x00000040)) 371 #define RXTX_REG26_PERIOD_ERROR_LATCH_SET(dst, src) \ 372 (((dst) & ~0x00003800) | (((u32) (src) << 11) & 0x00003800)) 373 #define RXTX_REG26_BLWC_ENA_SET(dst, src) \ 374 (((dst) & ~0x00000008) | (((u32) (src) << 3) & 0x00000008)) 376 #define RXTX_REG21_DO_LATCH_CALOUT_RD(src) ((0x0000fc00 & (u32) (src)) >> 10) 377 #define RXTX_REG21_XO_LATCH_CALOUT_RD(src) ((0x000003f0 & (u32) (src)) >> 4) 378 #define RXTX_REG21_LATCH_CAL_FAIL_ODD_RD(src) ((0x0000000f & (u32)(src))) 380 #define RXTX_REG22_SO_LATCH_CALOUT_RD(src) ((0x000003f0 & (u32) (src)) >> 4) 381 #define RXTX_REG22_EO_LATCH_CALOUT_RD(src) ((0x0000fc00 & (u32) (src)) >> 10) 382 #define RXTX_REG22_LATCH_CAL_FAIL_EVEN_RD(src) ((0x0000000f & (u32)(src))) 384 #define RXTX_REG23_DE_LATCH_CALOUT_RD(src) ((0x0000fc00 & (u32) (src)) >> 10) 385 #define RXTX_REG23_XE_LATCH_CALOUT_RD(src) ((0x000003f0 & (u32) (src)) >> 4) 387 #define RXTX_REG24_EE_LATCH_CALOUT_RD(src) ((0x0000fc00 & (u32) (src)) >> 10) 388 #define RXTX_REG24_SE_LATCH_CALOUT_RD(src) ((0x000003f0 & (u32) (src)) >> 4) 393 #define RXTX_REG38_CUSTOMER_PINMODE_INV_SET(dst, src) \ 394 (((dst) & 0x0000fffe) | (((u32) (src) << 1) & 0x0000fffe)) 413 #define RXTX_REG61_ISCAN_INBERT_SET(dst, src) \ 414 (((dst) & ~0x00000010) | (((u32) (src) << 4) & 0x00000010)) 415 #define RXTX_REG61_LOADFREQ_SHIFT_SET(dst, src) \ 416 (((dst) & ~0x00000008) | (((u32) (src) << 3) & 0x00000008)) 417 #define RXTX_REG61_EYE_COUNT_WIDTH_SEL_SET(dst, src) \ 418 (((dst) & ~0x000000c0) | (((u32) (src) << 6) & 0x000000c0)) 419 #define RXTX_REG61_SPD_SEL_CDR_SET(dst, src) \ 420 (((dst) & ~0x00003c00) | (((u32) (src) << 10) & 0x00003c00)) 422 #define RXTX_REG62_PERIOD_H1_QLATCH_SET(dst, src) \ 423 (((dst) & ~0x00003800) | (((u32) (src) << 11) & 0x00003800)) 425 #define RXTX_REG89_MU_TH7_SET(dst, src) \ 426 (((dst) & ~0x0000f800) | (((u32) (src) << 11) & 0x0000f800)) 427 #define RXTX_REG89_MU_TH8_SET(dst, src) \ 428 (((dst) & ~0x000007c0) | (((u32) (src) << 6) & 0x000007c0)) 429 #define RXTX_REG89_MU_TH9_SET(dst, src) \ 430 (((dst) & ~0x0000003e) | (((u32) (src) << 1) & 0x0000003e)) 432 #define RXTX_REG96_MU_FREQ1_SET(dst, src) \ 433 (((dst) & ~0x0000f800) | (((u32) (src) << 11) & 0x0000f800)) 434 #define RXTX_REG96_MU_FREQ2_SET(dst, src) \ 435 (((dst) & ~0x000007c0) | (((u32) (src) << 6) & 0x000007c0)) 436 #define RXTX_REG96_MU_FREQ3_SET(dst, src) \ 437 (((dst) & ~0x0000003e) | (((u32) (src) << 1) & 0x0000003e)) 439 #define RXTX_REG99_MU_PHASE1_SET(dst, src) \ 440 (((dst) & ~0x0000f800) | (((u32) (src) << 11) & 0x0000f800)) 441 #define RXTX_REG99_MU_PHASE2_SET(dst, src) \ 442 (((dst) & ~0x000007c0) | (((u32) (src) << 6) & 0x000007c0)) 443 #define RXTX_REG99_MU_PHASE3_SET(dst, src) \ 444 (((dst) & ~0x0000003e) | (((u32) (src) << 1) & 0x0000003e)) 446 #define RXTX_REG102_FREQLOOP_LIMIT_SET(dst, src) \ 447 (((dst) & ~0x00000060) | (((u32) (src) << 5) & 0x00000060)) 450 #define RXTX_REG121_SUMOS_CAL_CODE_RD(src) ((0x0000003e & (u32)(src)) >> 0x1) 452 #define RXTX_REG125_PQ_REG_SET(dst, src) \ 453 (((dst) & ~0x0000fe00) | (((u32) (src) << 9) & 0x0000fe00)) 454 #define RXTX_REG125_SIGN_PQ_SET(dst, src) \ 455 (((dst) & ~0x00000100) | (((u32) (src) << 8) & 0x00000100)) 456 #define RXTX_REG125_SIGN_PQ_2C_SET(dst, src) \ 457 (((dst) & ~0x00000080) | (((u32) (src) << 7) & 0x00000080)) 458 #define RXTX_REG125_PHZ_MANUALCODE_SET(dst, src) \ 459 (((dst) & ~0x0000007c) | (((u32) (src) << 2) & 0x0000007c)) 460 #define RXTX_REG125_PHZ_MANUAL_SET(dst, src) \ 461 (((dst) & ~0x00000002) | (((u32) (src) << 1) & 0x00000002)) 465 #define RXTX_REG127_FORCE_SUM_CAL_START_SET(dst, src) \ 466 (((dst) & ~0x00000002) | (((u32) (src) << 1) & 0x00000002)) 467 #define RXTX_REG127_FORCE_LAT_CAL_START_SET(dst, src) \ 468 (((dst) & ~0x00000004) | (((u32) (src) << 2) & 0x00000004)) 469 #define RXTX_REG127_LATCH_MAN_CAL_ENA_SET(dst, src) \ 470 (((dst) & ~0x00000008) | (((u32) (src) << 3) & 0x00000008)) 471 #define RXTX_REG127_DO_LATCH_MANCAL_SET(dst, src) \ 472 (((dst) & ~0x0000fc00) | (((u32) (src) << 10) & 0x0000fc00)) 473 #define RXTX_REG127_XO_LATCH_MANCAL_SET(dst, src) \ 474 (((dst) & ~0x000003f0) | (((u32) (src) << 4) & 0x000003f0)) 476 #define RXTX_REG128_LATCH_CAL_WAIT_SEL_SET(dst, src) \ 477 (((dst) & ~0x0000000c) | (((u32) (src) << 2) & 0x0000000c)) 478 #define RXTX_REG128_EO_LATCH_MANCAL_SET(dst, src) \ 479 (((dst) & ~0x0000fc00) | (((u32) (src) << 10) & 0x0000fc00)) 480 #define RXTX_REG128_SO_LATCH_MANCAL_SET(dst, src) \ 481 (((dst) & ~0x000003f0) | (((u32) (src) << 4) & 0x000003f0)) 483 #define RXTX_REG129_DE_LATCH_MANCAL_SET(dst, src) \ 484 (((dst) & ~0x0000fc00) | (((u32) (src) << 10) & 0x0000fc00)) 485 #define RXTX_REG129_XE_LATCH_MANCAL_SET(dst, src) \ 486 (((dst) & ~0x000003f0) | (((u32) (src) << 4) & 0x000003f0)) 488 #define RXTX_REG130_EE_LATCH_MANCAL_SET(dst, src) \ 489 (((dst) & ~0x0000fc00) | (((u32) (src) << 10) & 0x0000fc00)) 490 #define RXTX_REG130_SE_LATCH_MANCAL_SET(dst, src) \ 491 (((dst) & ~0x000003f0) | (((u32) (src) << 4) & 0x000003f0)) 493 #define RXTX_REG145_TX_IDLE_SATA_SET(dst, src) \ 494 (((dst) & ~0x00000001) | (((u32) (src) << 0) & 0x00000001)) 495 #define RXTX_REG145_RXES_ENA_SET(dst, src) \ 496 (((dst) & ~0x00000002) | (((u32) (src) << 1) & 0x00000002)) 497 #define RXTX_REG145_RXDFE_CONFIG_SET(dst, src) \ 498 (((dst) & ~0x0000c000) | (((u32) (src) << 14) & 0x0000c000)) 499 #define RXTX_REG145_RXVWES_LATENA_SET(dst, src) \ 500 (((dst) & ~0x00000004) | (((u32) (src) << 2) & 0x00000004))
|
/linux-4.1.27/include/net/netfilter/ |
H A D | nf_conntrack_tuple.h | 37 struct nf_conntrack_man src; member in struct:nf_conntrack_tuple 78 } src; member in struct:nf_conntrack_tuple_mask 86 &t->src.u3.ip, ntohs(t->src.u.all), nf_ct_dump_tuple_ip() 96 t->src.u3.all, ntohs(t->src.u.all), nf_ct_dump_tuple_ipv6() 103 switch (t->src.l3num) { nf_ct_dump_tuple() 126 return (nf_inet_addr_cmp(&t1->src.u3, &t2->src.u3) && __nf_ct_tuple_src_equal() 127 t1->src.u.all == t2->src.u.all && __nf_ct_tuple_src_equal() 128 t1->src.l3num == t2->src.l3num); __nf_ct_tuple_src_equal() 150 return (nf_inet_addr_cmp(&m1->src.u3, &m2->src.u3) && nf_ct_tuple_mask_equal() 151 m1->src.u.all == m2->src.u.all); nf_ct_tuple_mask_equal() 162 if ((t1->src.u3.all[count] ^ t2->src.u3.all[count]) & nf_ct_tuple_src_mask_cmp() 163 mask->src.u3.all[count]) nf_ct_tuple_src_mask_cmp() 167 if ((t1->src.u.all ^ t2->src.u.all) & mask->src.u.all) nf_ct_tuple_src_mask_cmp() 170 if (t1->src.l3num != t2->src.l3num || nf_ct_tuple_src_mask_cmp()
|
/linux-4.1.27/arch/parisc/math-emu/ |
H A D | sfsqrt.c | 57 register unsigned int src, result; sgl_fsqrt() local 62 src = *srcptr; sgl_fsqrt() 66 if ((src_exponent = Sgl_exponent(src)) == SGL_INFINITY_EXPONENT) { sgl_fsqrt() 70 if (Sgl_isone_signaling(src)) { sgl_fsqrt() 75 Sgl_set_quiet(src); sgl_fsqrt() 81 if (Sgl_iszero_sign(src) || Sgl_isnotzero_mantissa(src)) { sgl_fsqrt() 82 *dstptr = src; sgl_fsqrt() 90 if (Sgl_iszero_exponentmantissa(src)) { sgl_fsqrt() 91 *dstptr = src; sgl_fsqrt() 98 if (Sgl_isone_sign(src)) { sgl_fsqrt() 103 Sgl_makequietnan(src); sgl_fsqrt() 104 *dstptr = src; sgl_fsqrt() 112 even_exponent = Sgl_hidden(src); sgl_fsqrt() 113 Sgl_clear_signexponent_set_hidden(src); sgl_fsqrt() 117 Sgl_clear_signexponent(src); sgl_fsqrt() 119 Sgl_normalize(src,src_exponent); sgl_fsqrt() 125 Sgl_leftshiftby1(src); sgl_fsqrt() 135 while (newbit && Sgl_isnotzero(src)) { sgl_fsqrt() 137 if(sum <= Sgl_all(src)) { sgl_fsqrt() 140 Sgl_subtract(src,sum,src); sgl_fsqrt() 143 Sgl_leftshiftby1(src); sgl_fsqrt() 151 if (Sgl_isnotzero(src)) { sgl_fsqrt() 152 if (!even_exponent && Sgl_islessthan(result,src)) sgl_fsqrt()
|
H A D | frnd.c | 53 register unsigned int src, result; sgl_frnd() local 57 src = *srcptr; sgl_frnd() 61 if ((src_exponent = Sgl_exponent(src)) == SGL_INFINITY_EXPONENT) { sgl_frnd() 65 if (Sgl_isone_signaling(src)) { sgl_frnd() 70 Sgl_set_quiet(src); sgl_frnd() 75 *dstptr = src; sgl_frnd() 82 *dstptr = src; sgl_frnd() 89 Sgl_clear_exponent_set_hidden(src); sgl_frnd() 90 result = src; sgl_frnd() 93 if (Sgl_isinexact_to_fix(src,src_exponent)) { sgl_frnd() 98 if (Sgl_iszero_sign(src)) Sgl_increment(result); sgl_frnd() 101 if (Sgl_isone_sign(src)) Sgl_increment(result); sgl_frnd() 104 if (Sgl_isone_roundbit(src,src_exponent)) sgl_frnd() 105 if (Sgl_isone_stickybit(src,src_exponent) sgl_frnd() 116 result = src; /* set sign */ sgl_frnd() 119 if (Sgl_isnotzero_exponentmantissa(src)) { sgl_frnd() 124 if (Sgl_iszero_sign(src)) sgl_frnd() 128 if (Sgl_isone_sign(src)) sgl_frnd() 133 if (Sgl_isnotzero_mantissa(src)) sgl_frnd()
|
H A D | fcnvfx.c | 64 register unsigned int src, temp; sgl_to_sgl_fcnvfx() local 68 src = *srcptr; sgl_to_sgl_fcnvfx() 69 src_exponent = Sgl_exponent(src) - SGL_BIAS; sgl_to_sgl_fcnvfx() 77 Sgl_isnotzero_mantissa(src) || Sgl_iszero_sign(src)) { sgl_to_sgl_fcnvfx() 78 if (Sgl_iszero_sign(src)) result = 0x7fffffff; sgl_to_sgl_fcnvfx() 93 temp = src; sgl_to_sgl_fcnvfx() 96 if (Sgl_isone_sign(src)) result = -Sgl_all(temp); sgl_to_sgl_fcnvfx() 100 if (Sgl_isinexact_to_fix(src,src_exponent)) { sgl_to_sgl_fcnvfx() 105 if (Sgl_iszero_sign(src)) result++; sgl_to_sgl_fcnvfx() 108 if (Sgl_isone_sign(src)) result--; sgl_to_sgl_fcnvfx() 111 if (Sgl_isone_roundbit(src,src_exponent)) { sgl_to_sgl_fcnvfx() 112 if (Sgl_isone_stickybit(src,src_exponent) sgl_to_sgl_fcnvfx() 114 if (Sgl_iszero_sign(src)) result++; sgl_to_sgl_fcnvfx() 124 if (Sgl_isnotzero_exponentmantissa(src)) { sgl_to_sgl_fcnvfx() 129 if (Sgl_iszero_sign(src)) result++; sgl_to_sgl_fcnvfx() 132 if (Sgl_isone_sign(src)) result--; sgl_to_sgl_fcnvfx() 136 if (Sgl_isnotzero_mantissa(src)) sgl_to_sgl_fcnvfx() 137 if (Sgl_iszero_sign(src)) result++; sgl_to_sgl_fcnvfx() 162 register unsigned int src, temp, resultp2; sgl_to_dbl_fcnvfx() local 165 src = *srcptr; sgl_to_dbl_fcnvfx() 166 src_exponent = Sgl_exponent(src) - SGL_BIAS; sgl_to_dbl_fcnvfx() 174 Sgl_isnotzero_mantissa(src) || Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfx() 175 if (Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfx() 198 temp = src; sgl_to_dbl_fcnvfx() 201 if (Sgl_isone_sign(src)) { sgl_to_dbl_fcnvfx() 206 if (Sgl_isinexact_to_fix(src,src_exponent)) { sgl_to_dbl_fcnvfx() 211 if (Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfx() 216 if (Sgl_isone_sign(src)) { sgl_to_dbl_fcnvfx() 221 if (Sgl_isone_roundbit(src,src_exponent)) sgl_to_dbl_fcnvfx() 222 if (Sgl_isone_stickybit(src,src_exponent) || sgl_to_dbl_fcnvfx() 224 if (Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfx() 237 if (Sgl_isnotzero_exponentmantissa(src)) { sgl_to_dbl_fcnvfx() 242 if (Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfx() 247 if (Sgl_isone_sign(src)) { sgl_to_dbl_fcnvfx() 253 if (Sgl_isnotzero_mantissa(src)) sgl_to_dbl_fcnvfx() 254 if (Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfx()
|
H A D | fcnvxf.c | 64 register int src, dst_exponent; sgl_to_sgl_fcnvxf() local 67 src = *srcptr; sgl_to_sgl_fcnvxf() 71 if (src < 0) { sgl_to_sgl_fcnvxf() 73 Int_negate(src); sgl_to_sgl_fcnvxf() 78 if (src == 0) { sgl_to_sgl_fcnvxf() 93 Find_ms_one_bit(src,dst_exponent); sgl_to_sgl_fcnvxf() 95 if (dst_exponent >= 0) src <<= dst_exponent; sgl_to_sgl_fcnvxf() 96 else src = 1 << 30; sgl_to_sgl_fcnvxf() 97 Sgl_set_mantissa(result, src >> (SGL_EXP_LENGTH-1)); sgl_to_sgl_fcnvxf() 101 if (Int_isinexact_to_sgl(src)) { sgl_to_sgl_fcnvxf() 112 Sgl_roundnearest_from_int(src,result); sgl_to_sgl_fcnvxf() 135 register int src, dst_exponent; sgl_to_dbl_fcnvxf() local 138 src = *srcptr; sgl_to_dbl_fcnvxf() 142 if (src < 0) { sgl_to_dbl_fcnvxf() 144 Int_negate(src); sgl_to_dbl_fcnvxf() 149 if (src == 0) { sgl_to_dbl_fcnvxf() 164 Find_ms_one_bit(src,dst_exponent); sgl_to_dbl_fcnvxf() 166 if (dst_exponent >= 0) src <<= dst_exponent; sgl_to_dbl_fcnvxf() 167 else src = 1 << 30; sgl_to_dbl_fcnvxf() 168 Dbl_set_mantissap1(resultp1, src >> DBL_EXP_LENGTH - 1); sgl_to_dbl_fcnvxf() 169 Dbl_set_mantissap2(resultp2, src << (33-DBL_EXP_LENGTH)); sgl_to_dbl_fcnvxf() 248 * If dst_exponent = -1, src = - 2**63 so we won't need to dbl_to_sgl_fcnvxf() 352 * If dst_exponent = -1, src = - 2**63 so we won't need to dbl_to_dbl_fcnvxf()
|
H A D | fcnvfu.c | 65 register unsigned int src, result; sgl_to_sgl_fcnvfu() local 69 src = *srcptr; sgl_to_sgl_fcnvfu() 70 src_exponent = Sgl_exponent(src) - SGL_BIAS; sgl_to_sgl_fcnvfu() 76 if (Sgl_isone_sign(src)) { sgl_to_sgl_fcnvfu() 96 if (Sgl_isone_sign(src)) { sgl_to_sgl_fcnvfu() 105 Sgl_clear_signexponent_set_hidden(src); sgl_to_sgl_fcnvfu() 106 Suint_from_sgl_mantissa(src,src_exponent,result); sgl_to_sgl_fcnvfu() 109 if (Sgl_isinexact_to_unsigned(src,src_exponent)) { sgl_to_sgl_fcnvfu() 119 if (Sgl_isone_roundbit(src,src_exponent) && sgl_to_sgl_fcnvfu() 120 (Sgl_isone_stickybit(src,src_exponent) || sgl_to_sgl_fcnvfu() 131 if (Sgl_isnotzero_exponentmantissa(src)) { sgl_to_sgl_fcnvfu() 136 if (Sgl_iszero_sign(src)) { sgl_to_sgl_fcnvfu() 141 if (Sgl_isone_sign(src)) { sgl_to_sgl_fcnvfu() 152 Sgl_isnotzero_mantissa(src)) { sgl_to_sgl_fcnvfu() 153 if (Sgl_isone_sign(src)) { sgl_to_sgl_fcnvfu() 187 register unsigned int src, resultp1, resultp2; sgl_to_dbl_fcnvfu() local 190 src = *srcptr; sgl_to_dbl_fcnvfu() 191 src_exponent = Sgl_exponent(src) - SGL_BIAS; sgl_to_dbl_fcnvfu() 197 if (Sgl_isone_sign(src)) { sgl_to_dbl_fcnvfu() 217 if (Sgl_isone_sign(src)) { sgl_to_dbl_fcnvfu() 226 Sgl_clear_signexponent_set_hidden(src); sgl_to_dbl_fcnvfu() 227 Duint_from_sgl_mantissa(src,src_exponent,resultp1,resultp2); sgl_to_dbl_fcnvfu() 230 if (Sgl_isinexact_to_unsigned(src,src_exponent)) { sgl_to_dbl_fcnvfu() 240 if (Sgl_isone_roundbit(src,src_exponent) && sgl_to_dbl_fcnvfu() 241 (Sgl_isone_stickybit(src,src_exponent) || sgl_to_dbl_fcnvfu() 252 if (Sgl_isnotzero_exponentmantissa(src)) { sgl_to_dbl_fcnvfu() 257 if (Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfu() 262 if (Sgl_isone_sign(src)) { sgl_to_dbl_fcnvfu() 273 Sgl_isnotzero_mantissa(src)) { sgl_to_dbl_fcnvfu() 274 if (Sgl_isone_sign(src)) { sgl_to_dbl_fcnvfu()
|
H A D | fcnvfxt.c | 65 register unsigned int src, temp; sgl_to_sgl_fcnvfxt() local 68 src = *srcptr; sgl_to_sgl_fcnvfxt() 69 src_exponent = Sgl_exponent(src) - SGL_BIAS; sgl_to_sgl_fcnvfxt() 77 Sgl_isnotzero_mantissa(src) || Sgl_iszero_sign(src)) { sgl_to_sgl_fcnvfxt() 78 if (Sgl_iszero_sign(src)) result = 0x7fffffff; sgl_to_sgl_fcnvfxt() 93 temp = src; sgl_to_sgl_fcnvfxt() 96 if (Sgl_isone_sign(src)) result = -Sgl_all(temp); sgl_to_sgl_fcnvfxt() 101 if (Sgl_isinexact_to_fix(src,src_exponent)) { sgl_to_sgl_fcnvfxt() 110 if (Sgl_isnotzero_exponentmantissa(src)) { sgl_to_sgl_fcnvfxt() 130 register unsigned int src, temp, resultp2; sgl_to_dbl_fcnvfxt() local 132 src = *srcptr; sgl_to_dbl_fcnvfxt() 133 src_exponent = Sgl_exponent(src) - SGL_BIAS; sgl_to_dbl_fcnvfxt() 141 Sgl_isnotzero_mantissa(src) || Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfxt() 142 if (Sgl_iszero_sign(src)) { sgl_to_dbl_fcnvfxt() 165 temp = src; sgl_to_dbl_fcnvfxt() 168 if (Sgl_isone_sign(src)) { sgl_to_dbl_fcnvfxt() 174 if (Sgl_isinexact_to_fix(src,src_exponent)) { sgl_to_dbl_fcnvfxt() 184 if (Sgl_isnotzero_exponentmantissa(src)) { sgl_to_dbl_fcnvfxt()
|
/linux-4.1.27/security/selinux/ss/ |
H A D | context.h | 40 static inline int mls_context_cpy(struct context *dst, struct context *src) mls_context_cpy() argument 44 dst->range.level[0].sens = src->range.level[0].sens; mls_context_cpy() 45 rc = ebitmap_cpy(&dst->range.level[0].cat, &src->range.level[0].cat); mls_context_cpy() 49 dst->range.level[1].sens = src->range.level[1].sens; mls_context_cpy() 50 rc = ebitmap_cpy(&dst->range.level[1].cat, &src->range.level[1].cat); mls_context_cpy() 58 * Sets both levels in the MLS range of 'dst' to the low level of 'src'. 60 static inline int mls_context_cpy_low(struct context *dst, struct context *src) mls_context_cpy_low() argument 64 dst->range.level[0].sens = src->range.level[0].sens; mls_context_cpy_low() 65 rc = ebitmap_cpy(&dst->range.level[0].cat, &src->range.level[0].cat); mls_context_cpy_low() 69 dst->range.level[1].sens = src->range.level[0].sens; mls_context_cpy_low() 70 rc = ebitmap_cpy(&dst->range.level[1].cat, &src->range.level[0].cat); mls_context_cpy_low() 78 * Sets both levels in the MLS range of 'dst' to the high level of 'src'. 80 static inline int mls_context_cpy_high(struct context *dst, struct context *src) mls_context_cpy_high() argument 84 dst->range.level[0].sens = src->range.level[1].sens; mls_context_cpy_high() 85 rc = ebitmap_cpy(&dst->range.level[0].cat, &src->range.level[1].cat); mls_context_cpy_high() 89 dst->range.level[1].sens = src->range.level[1].sens; mls_context_cpy_high() 90 rc = ebitmap_cpy(&dst->range.level[1].cat, &src->range.level[1].cat); mls_context_cpy_high() 117 static inline int context_cpy(struct context *dst, struct context *src) context_cpy() argument 121 dst->user = src->user; context_cpy() 122 dst->role = src->role; context_cpy() 123 dst->type = src->type; context_cpy() 124 if (src->str) { context_cpy() 125 dst->str = kstrdup(src->str, GFP_ATOMIC); context_cpy() 128 dst->len = src->len; context_cpy() 133 rc = mls_context_cpy(dst, src); context_cpy()
|
/linux-4.1.27/drivers/staging/lustre/lustre/obdclass/ |
H A D | obdo.c | 59 void obdo_from_inode(struct obdo *dst, struct inode *src, u32 valid) obdo_from_inode() argument 65 valid, LTIME_S(src->i_mtime), obdo_from_inode() 66 LTIME_S(src->i_ctime)); obdo_from_inode() 69 dst->o_atime = LTIME_S(src->i_atime); obdo_from_inode() 73 dst->o_mtime = LTIME_S(src->i_mtime); obdo_from_inode() 77 dst->o_ctime = LTIME_S(src->i_ctime); obdo_from_inode() 81 dst->o_size = i_size_read(src); obdo_from_inode() 85 dst->o_blocks = src->i_blocks; obdo_from_inode() 89 dst->o_blksize = 1 << src->i_blkbits; obdo_from_inode() 94 (src->i_mode & S_IFMT); obdo_from_inode() 99 (src->i_mode & S_IALLUGO); obdo_from_inode() 103 dst->o_uid = from_kuid(&init_user_ns, src->i_uid); obdo_from_inode() 107 dst->o_gid = from_kgid(&init_user_ns, src->i_gid); obdo_from_inode() 111 dst->o_flags = src->i_flags; obdo_from_inode() 118 void obdo_cpy_md(struct obdo *dst, struct obdo *src, u32 valid) obdo_cpy_md() argument 120 CDEBUG(D_INODE, "src obdo "DOSTID" valid %#llx, dst obdo "DOSTID"\n", obdo_cpy_md() 121 POSTID(&src->o_oi), src->o_valid, POSTID(&dst->o_oi)); obdo_cpy_md() 123 dst->o_atime = src->o_atime; obdo_cpy_md() 125 dst->o_mtime = src->o_mtime; obdo_cpy_md() 127 dst->o_ctime = src->o_ctime; obdo_cpy_md() 129 dst->o_size = src->o_size; obdo_cpy_md() 131 dst->o_blocks = src->o_blocks; obdo_cpy_md() 133 dst->o_blksize = src->o_blksize; obdo_cpy_md() 135 dst->o_mode = (dst->o_mode & ~S_IFMT) | (src->o_mode & S_IFMT); obdo_cpy_md() 137 dst->o_mode = (dst->o_mode & S_IFMT) | (src->o_mode & ~S_IFMT); obdo_cpy_md() 139 dst->o_uid = src->o_uid; obdo_cpy_md() 141 dst->o_gid = src->o_gid; obdo_cpy_md() 143 dst->o_flags = src->o_flags; obdo_cpy_md() 145 dst->o_parent_seq = src->o_parent_seq; obdo_cpy_md() 146 dst->o_parent_ver = src->o_parent_ver; obdo_cpy_md() 149 dst->o_parent_oid = src->o_parent_oid; obdo_cpy_md() 151 dst->o_handle = src->o_handle; obdo_cpy_md() 153 dst->o_lcookie = src->o_lcookie; obdo_cpy_md() 160 int obdo_cmp_md(struct obdo *dst, struct obdo *src, u32 compare) obdo_cmp_md() argument 165 res |= dst->o_atime != src->o_atime; obdo_cmp_md() 167 res |= dst->o_mtime != src->o_mtime; obdo_cmp_md() 169 res |= dst->o_ctime != src->o_ctime; obdo_cmp_md() 171 res |= dst->o_size != src->o_size; obdo_cmp_md() 173 res |= dst->o_blocks != src->o_blocks; obdo_cmp_md() 175 res |= dst->o_blksize != src->o_blksize; obdo_cmp_md() 177 res |= ((dst->o_mode ^ src->o_mode) & S_IFMT) != 0; obdo_cmp_md() 179 res |= ((dst->o_mode ^ src->o_mode) & ~S_IFMT) != 0; obdo_cmp_md() 181 res |= dst->o_uid != src->o_uid; obdo_cmp_md() 183 res |= dst->o_gid != src->o_gid; obdo_cmp_md() 185 res |= dst->o_flags != src->o_flags; obdo_cmp_md() 187 res |= dst->o_nlink != src->o_nlink; obdo_cmp_md() 189 res |= dst->o_parent_seq != src->o_parent_seq; obdo_cmp_md() 190 res |= dst->o_parent_ver != src->o_parent_ver; obdo_cmp_md() 193 res |= dst->o_parent_oid != src->o_parent_oid; obdo_cmp_md() 196 res |= memcmp(dst->o_inline, src->o_inline); obdo_cmp_md()
|
/linux-4.1.27/arch/mips/cavium-octeon/ |
H A D | octeon-memcpy.S | 21 #define src a1 define 27 * memcpy copies len bytes from src to dst and sets v0 to dst. 29 * - src and dst don't overlap 30 * - src is readable 34 * __copy_user copies up to len bytes from src to dst and sets a2 (len) to 36 * __copy_user assumes that src and dst don't overlap, and that the call is 39 * - src is readable (no exceptions when reading src) 58 * 2- src_entry <= src < AT, and 59 * 3- (dst - src) == (dst_entry - src_entry), 63 * (2) is met by incrementing src by the number of bytes copied 64 * (3) is met by not doing loads between a pair of increments of dst and src 156 LEAF(memcpy) /* a0=dst a1=src a2=len */ 163 * Note: dst & src may be unaligned, len may be 0 170 pref 0, 0(src) 173 and t0, src, ADDRMASK # Check if src unaligned 185 pref 0, 128(src) # We must not prefetch invalid addresses 188 2: pref 0, 256(src) # We must not prefetch invalid addresses 192 EXC( LOAD t0, UNIT(0)(src), l_exc) 193 EXC( LOAD t1, UNIT(1)(src), l_exc_copy) 194 EXC( LOAD t2, UNIT(2)(src), l_exc_copy) 195 EXC( LOAD t3, UNIT(3)(src), l_exc_copy) 201 EXC( LOAD t0, UNIT(4)(src), l_exc_copy) 202 EXC( LOAD t1, UNIT(5)(src), l_exc_copy) 203 EXC( LOAD t2, UNIT(6)(src), l_exc_copy) 204 EXC( LOAD t3, UNIT(7)(src), l_exc_copy) 208 ADD src, src, 16*NBYTES 211 EXC( LOAD t0, UNIT(-8)(src), l_exc_copy) 212 EXC( LOAD t1, UNIT(-7)(src), l_exc_copy) 213 EXC( LOAD t2, UNIT(-6)(src), l_exc_copy) 214 EXC( LOAD t3, UNIT(-5)(src), l_exc_copy) 219 EXC( LOAD t0, UNIT(-4)(src), l_exc_copy) 220 EXC( LOAD t1, UNIT(-3)(src), l_exc_copy) 221 EXC( LOAD t2, UNIT(-2)(src), l_exc_copy) 222 EXC( LOAD t3, UNIT(-1)(src), l_exc_copy) 240 EXC( LOAD t0, UNIT(0)(src), l_exc) 241 EXC( LOAD t1, UNIT(1)(src), l_exc_copy) 242 EXC( LOAD t2, UNIT(2)(src), l_exc_copy) 243 EXC( LOAD t3, UNIT(3)(src), l_exc_copy) 249 EXC( LOAD t0, UNIT(4)(src), l_exc_copy) 250 EXC( LOAD t1, UNIT(5)(src), l_exc_copy) 251 EXC( LOAD t2, UNIT(6)(src), l_exc_copy) 252 EXC( LOAD t3, UNIT(7)(src), l_exc_copy) 257 ADD src, src, 8*NBYTES 267 EXC( LOAD t0, UNIT(0)(src), l_exc) 268 EXC( LOAD t1, UNIT(1)(src), l_exc_copy) 269 EXC( LOAD t2, UNIT(2)(src), l_exc_copy) 270 EXC( LOAD t3, UNIT(3)(src), l_exc_copy) 276 ADD src, src, 4*NBYTES 290 EXC( LOAD t0, 0(src), l_exc) 294 ADD src, src, NBYTES 300 EXC( LOAD t0, 0(src), l_exc) 304 ADD src, src, NBYTES 310 EXC( LOAD t0, 0(src), l_exc) 312 ADD src, src, NBYTES 327 * are to the same unit (unless src is aligned, but it's not). 329 EXC( LDFIRST t0, FIRST(0)(src), l_exc) 330 EXC( LDFIRST t1, FIRST(1)(src), l_exc_copy) 332 EXC( LDREST t0, REST(0)(src), l_exc_copy) 333 EXC( LDREST t1, REST(1)(src), l_exc_copy) 334 EXC( LDFIRST t2, FIRST(2)(src), l_exc_copy) 335 EXC( LDFIRST t3, FIRST(3)(src), l_exc_copy) 336 EXC( LDREST t2, REST(2)(src), l_exc_copy) 337 EXC( LDREST t3, REST(3)(src), l_exc_copy) 338 ADD src, src, 4*NBYTES 352 EXC( LDFIRST t0, FIRST(0)(src), l_exc) 353 EXC( LDREST t0, REST(0)(src), l_exc_copy) 356 ADD src, src, NBYTES 366 EXC( lb t0, N(src), l_exc); \ 377 EXC( lb t0, NBYTES-2(src), l_exc) 388 * Copy bytes from src until faulting load address (or until a 395 * Assumes src < THREAD_BUADDR($28) 400 EXC( lb t1, 0(src), l_exc) 401 ADD src, src, 1 403 bne src, t0, 1b 411 * Here's where we rely on src and dst being incremented in tandem, 413 * dst += (fault addr - src) to put dst at first byte to clear 416 SUB dst, src 422 SUB src, len, 1 425 bnez src, 1b 426 SUB src, src, 1 464 sltu t0, a1, t0 # dst + len <= src -> memcpy 465 sltu t1, a0, t1 # dst >= src + len -> memcpy 473 LEAF(__rmemcpy) /* a0=dst a1=src a2=len */ 475 beqz t0, r_end_bytes_up # src >= dst 478 ADD a1, a2 # src = src + len
|
H A D | Makefile | 12 CFLAGS_octeon-platform.o = -I$(src)/../../../scripts/dtc/libfdt 13 CFLAGS_setup.o = -I$(src)/../../../scripts/dtc/libfdt
|
/linux-4.1.27/arch/microblaze/lib/ |
H A D | memmove.c | 37 const char *src = v_src; memmove() local 48 src += c; memmove() 53 *--dst = *--src; memmove() 60 const char *src = v_src; memmove() local 81 src += c; memmove() 91 *--dst = *--src; memmove() 94 *--dst = *--src; memmove() 97 *--dst = *--src; memmove() 104 switch ((unsigned long)src & 3) { memmove() 107 i_src = (const void *)src; memmove() 112 src = (const void *)i_src; memmove() 116 i_src = (const void *) (((unsigned)src + 4) & ~3); memmove() 138 src = (const void *)i_src; memmove() 139 src += 1; memmove() 143 i_src = (const void *) (((unsigned)src + 4) & ~3); memmove() 165 src = (const void *)i_src; memmove() 166 src += 2; memmove() 170 i_src = (const void *) (((unsigned)src + 4) & ~3); memmove() 192 src = (const void *)i_src; memmove() 193 src += 3; memmove() 203 *--dst = *--src; memmove() 205 *--dst = *--src; memmove() 207 *--dst = *--src; memmove() 209 *--dst = *--src; memmove()
|
H A D | memcpy.c | 38 const char *src = v_src; memcpy() local 43 *dst++ = *src++; memcpy() 50 const char *src = v_src; memcpy() local 69 *dst++ = *src++; memcpy() 72 *dst++ = *src++; memcpy() 75 *dst++ = *src++; memcpy() 83 switch ((unsigned long)src & 3) { memcpy() 85 i_src = (const void *)src; memcpy() 90 src = (const void *)i_src; memcpy() 94 i_src = (const void *) ((unsigned)src & ~3); memcpy() 115 src = (const void *)i_src; memcpy() 116 src -= 3; memcpy() 120 i_src = (const void *) ((unsigned)src & ~3); memcpy() 141 src = (const void *)i_src; memcpy() 142 src -= 2; memcpy() 146 i_src = (const void *) ((unsigned)src & ~3); memcpy() 167 src = (const void *)i_src; memcpy() 168 src -= 1; memcpy() 178 *dst++ = *src++; memcpy() 180 *dst++ = *src++; memcpy() 182 *dst++ = *src++; memcpy()
|
H A D | memset.c | 37 char *src = v_src; memset() local 44 *src++ = c; memset() 51 char *src = v_src; memset() local 68 switch ((unsigned) src & 3) { memset() 70 *src++ = c; memset() 73 *src++ = c; memset() 76 *src++ = c; memset() 80 i_src = (void *)src; memset() 86 src = (void *)i_src; memset() 91 *src++ = c; memset()
|
/linux-4.1.27/arch/blackfin/lib/ |
H A D | strcpy.S | 9 /* void *strcpy(char *dest, const char *src); 11 * R1 = address (src) 26 P1 = R1 ; /* src*/
|
/linux-4.1.27/drivers/isdn/hardware/eicon/ |
H A D | diva.h | 13 const void *src, int length); 16 const void __user *src, int length); 21 int diva_xdi_write(void *adapter, void *os_handle, const void __user *src, 24 void *diva_xdi_open_adapter(void *os_handle, const void __user *src,
|
H A D | um_idi.h | 17 const void *src, 21 const void *src, 31 const void *src,
|
/linux-4.1.27/drivers/misc/cxl/ |
H A D | Makefile | 6 CFLAGS_trace.o := -I$(src)
|
/linux-4.1.27/drivers/scsi/aacraid/ |
H A D | Makefile | 6 dpcsup.o rx.o sa.o rkt.o nark.o src.o
|
/linux-4.1.27/fs/ |
H A D | stack.c | 10 void fsstack_copy_inode_size(struct inode *dst, struct inode *src) fsstack_copy_inode_size() argument 21 i_size = i_size_read(src); fsstack_copy_inode_size() 35 spin_lock(&src->i_lock); fsstack_copy_inode_size() 36 i_blocks = src->i_blocks; fsstack_copy_inode_size() 38 spin_unlock(&src->i_lock); fsstack_copy_inode_size() 63 void fsstack_copy_attr_all(struct inode *dest, const struct inode *src) fsstack_copy_attr_all() argument 65 dest->i_mode = src->i_mode; fsstack_copy_attr_all() 66 dest->i_uid = src->i_uid; fsstack_copy_attr_all() 67 dest->i_gid = src->i_gid; fsstack_copy_attr_all() 68 dest->i_rdev = src->i_rdev; fsstack_copy_attr_all() 69 dest->i_atime = src->i_atime; fsstack_copy_attr_all() 70 dest->i_mtime = src->i_mtime; fsstack_copy_attr_all() 71 dest->i_ctime = src->i_ctime; fsstack_copy_attr_all() 72 dest->i_blkbits = src->i_blkbits; fsstack_copy_attr_all() 73 dest->i_flags = src->i_flags; fsstack_copy_attr_all() 74 set_nlink(dest, src->i_nlink); fsstack_copy_attr_all()
|
/linux-4.1.27/include/linux/netfilter/ipset/ |
H A D | ip_set_getport.h | 4 extern bool ip_set_get_ip4_port(const struct sk_buff *skb, bool src, 8 extern bool ip_set_get_ip6_port(const struct sk_buff *skb, bool src, 11 static inline bool ip_set_get_ip6_port(const struct sk_buff *skb, bool src, ip_set_get_ip6_port() argument 18 extern bool ip_set_get_ip_port(const struct sk_buff *skb, u8 pf, bool src,
|
/linux-4.1.27/arch/parisc/include/asm/ |
H A D | string.h | 8 void * memcpy(void * dest,const void *src,size_t count);
|
/linux-4.1.27/net/ieee802154/ |
H A D | Makefile | 9 CFLAGS_trace.o := -I$(src)
|
/linux-4.1.27/sound/hda/ |
H A D | Makefile | 5 CFLAGS_trace.o := -I$(src)
|
/linux-4.1.27/arch/mips/lib/ |
H A D | memcpy.S | 37 #define src a1 define 43 * memcpy copies len bytes from src to dst and sets v0 to dst. 45 * - src and dst don't overlap 46 * - src is readable 50 * __copy_user copies up to len bytes from src to dst and sets a2 (len) to 52 * __copy_user assumes that src and dst don't overlap, and that the call is 55 * - src is readable (no exceptions when reading src) 74 * 2- src_entry <= src < AT, and 75 * 3- (dst - src) == (dst_entry - src_entry), 79 * (2) is met by incrementing src by the number of bytes copied 80 * (3) is met by not doing loads between a pair of increments of dst and src 273 * Note: dst & src may be unaligned, len may be 0 286 PREFS( 0, 0(src) ) 290 PREFS( 0, 1*32(src) ) 293 and t0, src, ADDRMASK 294 PREFS( 0, 2*32(src) ) 306 * src and dst are aligned; need to compute rem 312 PREFS( 0, 3*32(src) ) 317 LOAD(t0, UNIT(0)(src), .Ll_exc\@) 318 LOAD(t1, UNIT(1)(src), .Ll_exc_copy\@) 319 LOAD(t2, UNIT(2)(src), .Ll_exc_copy\@) 320 LOAD(t3, UNIT(3)(src), .Ll_exc_copy\@) 322 LOAD(t4, UNIT(4)(src), .Ll_exc_copy\@) 323 LOAD(t7, UNIT(5)(src), .Ll_exc_copy\@) 326 LOAD(t0, UNIT(6)(src), .Ll_exc_copy\@) 327 LOAD(t1, UNIT(7)(src), .Ll_exc_copy\@) 328 ADD src, src, 8*NBYTES 336 PREFS( 0, 8*32(src) ) 352 LOAD( t0, UNIT(0)(src), .Ll_exc\@) 353 LOAD( t1, UNIT(1)(src), .Ll_exc_copy\@) 354 LOAD( t2, UNIT(2)(src), .Ll_exc_copy\@) 355 LOAD( t3, UNIT(3)(src), .Ll_exc_copy\@) 357 ADD src, src, 4*NBYTES 375 LOAD(t0, 0(src), .Ll_exc\@) 376 ADD src, src, NBYTES 386 * src and dst are aligned, need to copy rem bytes (rem < NBYTES) 401 LOAD(t0, 0(src), .Ll_exc\@) 410 * t0 = src & ADDRMASK 415 * Set match = (src and dst have same alignment) 418 LDFIRST(t3, FIRST(0)(src), .Ll_exc\@) 420 LDREST(t3, REST(0)(src), .Ll_exc_copy\@) 429 ADD src, src, t2 433 PREFS( 0, 3*32(src) ) 442 * are to the same unit (unless src is aligned, but it's not). 445 LDFIRST(t0, FIRST(0)(src), .Ll_exc\@) 446 LDFIRST(t1, FIRST(1)(src), .Ll_exc_copy\@) 448 LDREST(t0, REST(0)(src), .Ll_exc_copy\@) 449 LDREST(t1, REST(1)(src), .Ll_exc_copy\@) 450 LDFIRST(t2, FIRST(2)(src), .Ll_exc_copy\@) 451 LDFIRST(t3, FIRST(3)(src), .Ll_exc_copy\@) 452 LDREST(t2, REST(2)(src), .Ll_exc_copy\@) 453 LDREST(t3, REST(3)(src), .Ll_exc_copy\@) 454 PREFS( 0, 9*32(src) ) # 0 is PREF_LOAD (not streamed) 455 ADD src, src, 4*NBYTES 476 LDFIRST(t0, FIRST(0)(src), .Ll_exc\@) 477 LDREST(t0, REST(0)(src), .Ll_exc_copy\@) 478 ADD src, src, NBYTES 494 LOADB(t0, N(src), .Ll_exc\@); \ 507 LOADB(t0, NBYTES-2(src), .Ll_exc\@) 526 ADD src, src, 8 538 * Copy bytes from src until faulting load address (or until a 545 * Assumes src < THREAD_BUADDR($28) 551 LOADB(t1, 0(src), .Ll_exc\@) 552 ADD src, src, 1 556 bne src, t0, 1b 566 * Here's where we rely on src and dst being incremented in tandem, 568 * dst += (fault addr - src) to put dst at first byte to clear 571 SUB dst, src 577 SUB src, len, 1 583 bnez src, 1b 584 SUB src, src, 1 589 bnez src, 1b 590 SUB src, src, v1 627 sltu t0, a1, t0 # dst + len <= src -> memcpy 628 sltu t1, a0, t1 # dst >= src + len -> memcpy 636 LEAF(__rmemcpy) /* a0=dst a1=src a2=len */ 638 beqz t0, .Lr_end_bytes_up # src >= dst 641 ADD a1, a2 # src = src + len 688 LEAF(memcpy) /* a0=dst a1=src a2=len */
|
H A D | csum_partial.S | 74 #define CSUM_BIGCHUNK1(src, offset, sum, _t0, _t1, _t2, _t3) \ 75 LOAD _t0, (offset + UNIT(0))(src); \ 76 LOAD _t1, (offset + UNIT(1))(src); \ 77 LOAD _t2, (offset + UNIT(2))(src); \ 78 LOAD _t3, (offset + UNIT(3))(src); \ 85 #define CSUM_BIGCHUNK(src, offset, sum, _t0, _t1, _t2, _t3) \ 86 CSUM_BIGCHUNK1(src, offset, sum, _t0, _t1, _t2, _t3) 88 #define CSUM_BIGCHUNK(src, offset, sum, _t0, _t1, _t2, _t3) \ 89 CSUM_BIGCHUNK1(src, offset, sum, _t0, _t1, _t2, _t3); \ 90 CSUM_BIGCHUNK1(src, offset + 0x10, sum, _t0, _t1, _t2, _t3) 99 #define src a0 define 113 andi t7, src, 0x1 /* odd buffer? */ 117 andi t8, src, 0x2 119 lbu t0, (src) 125 PTR_ADDU src, src, 0x1 126 andi t8, src, 0x2 132 lhu t0, (src) 136 PTR_ADDU src, src, 0x2 142 andi t8, src, 0x4 144 andi t8, src, 0x8 146 LOAD32 t0, 0x00(src) 149 PTR_ADDU src, src, 0x4 150 andi t8, src, 0x8 154 andi t8, src, 0x10 157 ld t0, 0x00(src) 161 lw t0, 0x00(src) 162 lw t1, 0x04(src) 167 PTR_ADDU src, src, 0x8 168 andi t8, src, 0x10 175 ld t0, 0x00(src) 176 ld t1, 0x08(src) 180 CSUM_BIGCHUNK1(src, 0x00, sum, t0, t1, t3, t4) 183 PTR_ADDU src, src, 0x10 191 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4) 192 CSUM_BIGCHUNK(src, 0x20, sum, t0, t1, t3, t4) 193 CSUM_BIGCHUNK(src, 0x40, sum, t0, t1, t3, t4) 194 CSUM_BIGCHUNK(src, 0x60, sum, t0, t1, t3, t4) 197 PTR_ADDU src, src, 0x80 206 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4) 207 CSUM_BIGCHUNK(src, 0x20, sum, t0, t1, t3, t4) 208 PTR_ADDU src, src, 0x40 215 CSUM_BIGCHUNK(src, 0x00, sum, t0, t1, t3, t4) 217 PTR_ADDU src, src, 0x20 225 LOAD32 t0, (src) 229 PTR_ADDU src, src, 0x4 233 /* unknown src alignment and < 8 bytes to go */ 242 ulw t1, (src) 243 PTR_ADDIU src, 4 254 ulhu t1, (src) 255 PTR_ADDIU src, 2 260 lbu t2, (src) 308 * csum_partial_copy_nocheck(src, dst, len, sum) 309 * __csum_partial_copy_kernel(src, dst, len, sum, errp) 315 #define src a0 define 327 * 2- src_entry <= src < AT, and 328 * 3- (dst - src) == (dst_entry - src_entry), 333 * (2) is met by incrementing src by the number of bytes copied 334 * (3) is met by not doing loads between a pair of increments of dst and src 452 PTR_ADDU AT, src, len /* See (1) above. */ 467 * Note: dst & src may be unaligned, len may be 0 480 and t0, src, ADDRMASK 487 * src and dst are aligned; need to compute rem 496 LOAD(t0, UNIT(0)(src), .Ll_exc\@) 497 LOAD(t1, UNIT(1)(src), .Ll_exc_copy\@) 498 LOAD(t2, UNIT(2)(src), .Ll_exc_copy\@) 499 LOAD(t3, UNIT(3)(src), .Ll_exc_copy\@) 500 LOAD(t4, UNIT(4)(src), .Ll_exc_copy\@) 501 LOAD(t5, UNIT(5)(src), .Ll_exc_copy\@) 502 LOAD(t6, UNIT(6)(src), .Ll_exc_copy\@) 503 LOAD(t7, UNIT(7)(src), .Ll_exc_copy\@) 505 ADD src, src, 8*NBYTES 540 LOAD(t0, UNIT(0)(src), .Ll_exc\@) 541 LOAD(t1, UNIT(1)(src), .Ll_exc_copy\@) 542 LOAD(t2, UNIT(2)(src), .Ll_exc_copy\@) 543 LOAD(t3, UNIT(3)(src), .Ll_exc_copy\@) 545 ADD src, src, 4*NBYTES 565 LOAD(t0, 0(src), .Ll_exc\@) 566 ADD src, src, NBYTES 576 * src and dst are aligned, need to copy rem bytes (rem < NBYTES) 591 LOAD(t0, 0(src), .Ll_exc\@) 603 * t0 = src & ADDRMASK 608 * Set match = (src and dst have same alignment) 611 LDFIRST(t3, FIRST(0)(src), .Ll_exc\@) 613 LDREST(t3, REST(0)(src), .Ll_exc_copy\@) 625 ADD src, src, t2 636 * are to the same unit (unless src is aligned, but it's not). 638 LDFIRST(t0, FIRST(0)(src), .Ll_exc\@) 639 LDFIRST(t1, FIRST(1)(src), .Ll_exc_copy\@) 641 LDREST(t0, REST(0)(src), .Ll_exc_copy\@) 642 LDREST(t1, REST(1)(src), .Ll_exc_copy\@) 643 LDFIRST(t2, FIRST(2)(src), .Ll_exc_copy\@) 644 LDFIRST(t3, FIRST(3)(src), .Ll_exc_copy\@) 645 LDREST(t2, REST(2)(src), .Ll_exc_copy\@) 646 LDREST(t3, REST(3)(src), .Ll_exc_copy\@) 647 ADD src, src, 4*NBYTES 670 LDFIRST(t0, FIRST(0)(src), .Ll_exc\@) 671 LDREST(t0, REST(0)(src), .Ll_exc_copy\@) 672 ADD src, src, NBYTES 697 LOADBU(t0, N(src), .Ll_exc_copy\@); \ 713 LOADBU(t0, NBYTES-2(src), .Ll_exc_copy\@) 757 * Copy bytes from src until faulting load address (or until a 764 * Assumes src < THREAD_BUADDR($28) 770 LOADBU(t1, 0(src), .Ll_exc\@) 771 ADD src, src, 1 778 bne src, t0, 1b 787 * Here's where we rely on src and dst being incremented in tandem, 789 * dst += (fault addr - src) to put dst at first byte to clear 792 SUB dst, src 798 SUB src, len, 1 806 bnez src, 1b 807 SUB src, src, 1 810 bnez src, 1b 811 SUB src, src, v1
|
H A D | iomap.c | 131 static inline void mmio_outsb(void __iomem *addr, const u8 *src, int count) mmio_outsb() argument 134 __mem_writeb(*src, addr); mmio_outsb() 135 src++; mmio_outsb() 139 static inline void mmio_outsw(void __iomem *addr, const u16 *src, int count) mmio_outsw() argument 142 __mem_writew(*src, addr); mmio_outsw() 143 src++; mmio_outsw() 147 static inline void mmio_outsl(void __iomem *addr, const u32 *src, int count) mmio_outsl() argument 150 __mem_writel(*src, addr); mmio_outsl() 151 src++; mmio_outsl() 176 void iowrite8_rep(void __iomem *addr, const void *src, unsigned long count) iowrite8_rep() argument 178 mmio_outsb(addr, src, count); iowrite8_rep() 183 void iowrite16_rep(void __iomem *addr, const void *src, unsigned long count) iowrite16_rep() argument 185 mmio_outsw(addr, src, count); iowrite16_rep() 190 void iowrite32_rep(void __iomem *addr, const void *src, unsigned long count) iowrite32_rep() argument 192 mmio_outsl(addr, src, count); iowrite32_rep()
|
/linux-4.1.27/sound/core/ |
H A D | pcm_compat.c | 27 s32 __user *src) snd_pcm_ioctl_delay_compat() 38 if (put_user(delay, src)) snd_pcm_ioctl_delay_compat() 44 u32 __user *src) snd_pcm_ioctl_rewind_compat() 49 if (get_user(frames, src)) snd_pcm_ioctl_rewind_compat() 55 if (put_user(err, src)) snd_pcm_ioctl_rewind_compat() 61 u32 __user *src) snd_pcm_ioctl_forward_compat() 66 if (get_user(frames, src)) snd_pcm_ioctl_forward_compat() 72 if (put_user(err, src)) snd_pcm_ioctl_forward_compat() 123 struct snd_pcm_sw_params32 __user *src) snd_pcm_ioctl_sw_params_compat() 130 if (get_user(params.tstamp_mode, &src->tstamp_mode) || snd_pcm_ioctl_sw_params_compat() 131 get_user(params.period_step, &src->period_step) || snd_pcm_ioctl_sw_params_compat() 132 get_user(params.sleep_min, &src->sleep_min) || snd_pcm_ioctl_sw_params_compat() 133 get_user(params.avail_min, &src->avail_min) || snd_pcm_ioctl_sw_params_compat() 134 get_user(params.xfer_align, &src->xfer_align) || snd_pcm_ioctl_sw_params_compat() 135 get_user(params.start_threshold, &src->start_threshold) || snd_pcm_ioctl_sw_params_compat() 136 get_user(params.stop_threshold, &src->stop_threshold) || snd_pcm_ioctl_sw_params_compat() 137 get_user(params.silence_threshold, &src->silence_threshold) || snd_pcm_ioctl_sw_params_compat() 138 get_user(params.silence_size, &src->silence_size) || snd_pcm_ioctl_sw_params_compat() 139 get_user(params.tstamp_type, &src->tstamp_type) || snd_pcm_ioctl_sw_params_compat() 140 get_user(params.proto, &src->proto)) snd_pcm_ioctl_sw_params_compat() 152 if (boundary && put_user(boundary, &src->boundary)) snd_pcm_ioctl_sw_params_compat() 165 struct snd_pcm_channel_info32 __user *src) snd_pcm_ioctl_channel_info_compat() 170 if (get_user(info.channel, &src->channel) || snd_pcm_ioctl_channel_info_compat() 171 get_user(info.offset, &src->offset) || snd_pcm_ioctl_channel_info_compat() 172 get_user(info.first, &src->first) || snd_pcm_ioctl_channel_info_compat() 173 get_user(info.step, &src->step)) snd_pcm_ioctl_channel_info_compat() 178 if (put_user(info.channel, &src->channel) || snd_pcm_ioctl_channel_info_compat() 179 put_user(info.offset, &src->offset) || snd_pcm_ioctl_channel_info_compat() 180 put_user(info.first, &src->first) || snd_pcm_ioctl_channel_info_compat() 181 put_user(info.step, &src->step)) snd_pcm_ioctl_channel_info_compat() 189 struct snd_pcm_channel_info __user *src); 214 struct snd_pcm_status32 __user *src, snd_pcm_status_user_compat() 227 (u32 __user *)(&src->audio_tstamp_data))) snd_pcm_status_user_compat() 233 if (clear_user(src, sizeof(*src))) snd_pcm_status_user_compat() 235 if (put_user(status.state, &src->state) || snd_pcm_status_user_compat() 236 compat_put_timespec(&status.trigger_tstamp, &src->trigger_tstamp) || snd_pcm_status_user_compat() 237 compat_put_timespec(&status.tstamp, &src->tstamp) || snd_pcm_status_user_compat() 238 put_user(status.appl_ptr, &src->appl_ptr) || snd_pcm_status_user_compat() 239 put_user(status.hw_ptr, &src->hw_ptr) || snd_pcm_status_user_compat() 240 put_user(status.delay, &src->delay) || snd_pcm_status_user_compat() 241 put_user(status.avail, &src->avail) || snd_pcm_status_user_compat() 242 put_user(status.avail_max, &src->avail_max) || snd_pcm_status_user_compat() 243 put_user(status.overrange, &src->overrange) || snd_pcm_status_user_compat() 244 put_user(status.suspended_state, &src->suspended_state) || snd_pcm_status_user_compat() 245 put_user(status.audio_tstamp_data, &src->audio_tstamp_data) || snd_pcm_status_user_compat() 246 compat_put_timespec(&status.audio_tstamp, &src->audio_tstamp) || snd_pcm_status_user_compat() 247 compat_put_timespec(&status.driver_tstamp, &src->driver_tstamp) || snd_pcm_status_user_compat() 248 put_user(status.audio_tstamp_accuracy, &src->audio_tstamp_accuracy)) snd_pcm_status_user_compat() 275 #define put_timespec(src, dst) copy_to_user(dst, src, sizeof(*dst)) 278 struct snd_pcm_status_x32 __user *src, snd_pcm_status_user_x32() 291 (u32 __user *)(&src->audio_tstamp_data))) snd_pcm_status_user_x32() 297 if (clear_user(src, sizeof(*src))) snd_pcm_status_user_x32() 299 if (put_user(status.state, &src->state) || snd_pcm_status_user_x32() 300 put_timespec(&status.trigger_tstamp, &src->trigger_tstamp) || snd_pcm_status_user_x32() 301 put_timespec(&status.tstamp, &src->tstamp) || snd_pcm_status_user_x32() 302 put_user(status.appl_ptr, &src->appl_ptr) || snd_pcm_status_user_x32() 303 put_user(status.hw_ptr, &src->hw_ptr) || snd_pcm_status_user_x32() 304 put_user(status.delay, &src->delay) || snd_pcm_status_user_x32() 305 put_user(status.avail, &src->avail) || snd_pcm_status_user_x32() 306 put_user(status.avail_max, &src->avail_max) || snd_pcm_status_user_x32() 307 put_user(status.overrange, &src->overrange) || snd_pcm_status_user_x32() 308 put_user(status.suspended_state, &src->suspended_state) || snd_pcm_status_user_x32() 309 put_user(status.audio_tstamp_data, &src->audio_tstamp_data) || snd_pcm_status_user_x32() 310 put_timespec(&status.audio_tstamp, &src->audio_tstamp) || snd_pcm_status_user_x32() 311 put_timespec(&status.driver_tstamp, &src->driver_tstamp) || snd_pcm_status_user_x32() 312 put_user(status.audio_tstamp_accuracy, &src->audio_tstamp_accuracy)) snd_pcm_status_user_x32() 488 struct snd_pcm_sync_ptr32 __user *src) snd_pcm_ioctl_sync_ptr_compat() 502 if (get_user(sflags, &src->flags) || snd_pcm_ioctl_sync_ptr_compat() 503 get_user(scontrol.appl_ptr, &src->c.control.appl_ptr) || snd_pcm_ioctl_sync_ptr_compat() 504 get_user(scontrol.avail_min, &src->c.control.avail_min)) snd_pcm_ioctl_sync_ptr_compat() 532 if (put_user(sstatus.state, &src->s.status.state) || snd_pcm_ioctl_sync_ptr_compat() 533 put_user(sstatus.hw_ptr, &src->s.status.hw_ptr) || snd_pcm_ioctl_sync_ptr_compat() 534 compat_put_timespec(&sstatus.tstamp, &src->s.status.tstamp) || snd_pcm_ioctl_sync_ptr_compat() 535 put_user(sstatus.suspended_state, &src->s.status.suspended_state) || snd_pcm_ioctl_sync_ptr_compat() 537 &src->s.status.audio_tstamp) || snd_pcm_ioctl_sync_ptr_compat() 538 put_user(scontrol.appl_ptr, &src->c.control.appl_ptr) || snd_pcm_ioctl_sync_ptr_compat() 539 put_user(scontrol.avail_min, &src->c.control.avail_min)) snd_pcm_ioctl_sync_ptr_compat() 576 struct snd_pcm_sync_ptr_x32 __user *src) snd_pcm_ioctl_sync_ptr_x32() 590 if (get_user(sflags, &src->flags) || snd_pcm_ioctl_sync_ptr_x32() 591 get_user(scontrol.appl_ptr, &src->c.control.appl_ptr) || snd_pcm_ioctl_sync_ptr_x32() 592 get_user(scontrol.avail_min, &src->c.control.avail_min)) snd_pcm_ioctl_sync_ptr_x32() 620 if (put_user(sstatus.state, &src->s.status.state) || snd_pcm_ioctl_sync_ptr_x32() 621 put_user(sstatus.hw_ptr, &src->s.status.hw_ptr) || snd_pcm_ioctl_sync_ptr_x32() 622 put_timespec(&sstatus.tstamp, &src->s.status.tstamp) || snd_pcm_ioctl_sync_ptr_x32() 623 put_user(sstatus.suspended_state, &src->s.status.suspended_state) || snd_pcm_ioctl_sync_ptr_x32() 624 put_timespec(&sstatus.audio_tstamp, &src->s.status.audio_tstamp) || snd_pcm_ioctl_sync_ptr_x32() 625 put_user(scontrol.appl_ptr, &src->c.control.appl_ptr) || snd_pcm_ioctl_sync_ptr_x32() 626 put_user(scontrol.avail_min, &src->c.control.avail_min)) snd_pcm_ioctl_sync_ptr_x32() 26 snd_pcm_ioctl_delay_compat(struct snd_pcm_substream *substream, s32 __user *src) snd_pcm_ioctl_delay_compat() argument 43 snd_pcm_ioctl_rewind_compat(struct snd_pcm_substream *substream, u32 __user *src) snd_pcm_ioctl_rewind_compat() argument 60 snd_pcm_ioctl_forward_compat(struct snd_pcm_substream *substream, u32 __user *src) snd_pcm_ioctl_forward_compat() argument 122 snd_pcm_ioctl_sw_params_compat(struct snd_pcm_substream *substream, struct snd_pcm_sw_params32 __user *src) snd_pcm_ioctl_sw_params_compat() argument 164 snd_pcm_ioctl_channel_info_compat(struct snd_pcm_substream *substream, struct snd_pcm_channel_info32 __user *src) snd_pcm_ioctl_channel_info_compat() argument 213 snd_pcm_status_user_compat(struct snd_pcm_substream *substream, struct snd_pcm_status32 __user *src, bool ext) snd_pcm_status_user_compat() argument 277 snd_pcm_status_user_x32(struct snd_pcm_substream *substream, struct snd_pcm_status_x32 __user *src, bool ext) snd_pcm_status_user_x32() argument 487 snd_pcm_ioctl_sync_ptr_compat(struct snd_pcm_substream *substream, struct snd_pcm_sync_ptr32 __user *src) snd_pcm_ioctl_sync_ptr_compat() argument 575 snd_pcm_ioctl_sync_ptr_x32(struct snd_pcm_substream *substream, struct snd_pcm_sync_ptr_x32 __user *src) snd_pcm_ioctl_sync_ptr_x32() argument
|
H A D | rawmidi_compat.c | 34 struct snd_rawmidi_params32 __user *src) snd_rawmidi_ioctl_params_compat() 41 if (get_user(params.stream, &src->stream) || snd_rawmidi_ioctl_params_compat() 42 get_user(params.buffer_size, &src->buffer_size) || snd_rawmidi_ioctl_params_compat() 43 get_user(params.avail_min, &src->avail_min) || snd_rawmidi_ioctl_params_compat() 44 get_user(val, &src->no_active_sensing)) snd_rawmidi_ioctl_params_compat() 65 struct snd_rawmidi_status32 __user *src) snd_rawmidi_ioctl_status_compat() 72 if (get_user(status.stream, &src->stream)) snd_rawmidi_ioctl_status_compat() 88 if (put_user(status.tstamp.tv_sec, &src->tstamp.tv_sec) || snd_rawmidi_ioctl_status_compat() 89 put_user(status.tstamp.tv_nsec, &src->tstamp.tv_nsec) || snd_rawmidi_ioctl_status_compat() 90 put_user(status.avail, &src->avail) || snd_rawmidi_ioctl_status_compat() 91 put_user(status.xruns, &src->xruns)) snd_rawmidi_ioctl_status_compat() 108 #define put_timespec(src, dst) copy_to_user(dst, src, sizeof(*dst)) 111 struct snd_rawmidi_status_x32 __user *src) snd_rawmidi_ioctl_status_x32() 118 if (get_user(status.stream, &src->stream)) snd_rawmidi_ioctl_status_x32() 134 if (put_timespec(&status.tstamp, &src->tstamp) || snd_rawmidi_ioctl_status_x32() 135 put_user(status.avail, &src->avail) || snd_rawmidi_ioctl_status_x32() 136 put_user(status.xruns, &src->xruns)) snd_rawmidi_ioctl_status_x32() 33 snd_rawmidi_ioctl_params_compat(struct snd_rawmidi_file *rfile, struct snd_rawmidi_params32 __user *src) snd_rawmidi_ioctl_params_compat() argument 64 snd_rawmidi_ioctl_status_compat(struct snd_rawmidi_file *rfile, struct snd_rawmidi_status32 __user *src) snd_rawmidi_ioctl_status_compat() argument 110 snd_rawmidi_ioctl_status_x32(struct snd_rawmidi_file *rfile, struct snd_rawmidi_status_x32 __user *src) snd_rawmidi_ioctl_status_x32() argument
|
H A D | memory.c | 31 * @src: the source pointer on mmio 38 int copy_to_user_fromio(void __user *dst, const volatile void __iomem *src, size_t count) copy_to_user_fromio() argument 41 return copy_to_user(dst, (const void __force*)src, count) ? -EFAULT : 0; copy_to_user_fromio() 48 memcpy_fromio(buf, (void __iomem *)src, c); copy_to_user_fromio() 53 src += c; copy_to_user_fromio() 64 * @src: the source pointer on user-space 71 int copy_from_user_toio(volatile void __iomem *dst, const void __user *src, size_t count) copy_from_user_toio() argument 74 return copy_from_user((void __force *)dst, src, count) ? -EFAULT : 0; copy_from_user_toio() 81 if (copy_from_user(buf, src, c)) copy_from_user_toio() 86 src += c; copy_from_user_toio()
|
/linux-4.1.27/drivers/staging/lustre/lustre/obdclass/linux/ |
H A D | linux-obdo.c | 148 void obdo_refresh_inode(struct inode *dst, struct obdo *src, u32 valid) obdo_refresh_inode() argument 150 valid &= src->o_valid; obdo_refresh_inode() 155 src->o_valid, LTIME_S(dst->i_mtime), obdo_refresh_inode() 156 LTIME_S(dst->i_ctime), src->o_mtime, src->o_ctime); obdo_refresh_inode() 158 if (valid & OBD_MD_FLATIME && src->o_atime > LTIME_S(dst->i_atime)) obdo_refresh_inode() 159 LTIME_S(dst->i_atime) = src->o_atime; obdo_refresh_inode() 160 if (valid & OBD_MD_FLMTIME && src->o_mtime > LTIME_S(dst->i_mtime)) obdo_refresh_inode() 161 LTIME_S(dst->i_mtime) = src->o_mtime; obdo_refresh_inode() 162 if (valid & OBD_MD_FLCTIME && src->o_ctime > LTIME_S(dst->i_ctime)) obdo_refresh_inode() 163 LTIME_S(dst->i_ctime) = src->o_ctime; obdo_refresh_inode() 165 i_size_write(dst, src->o_size); obdo_refresh_inode() 167 if (valid & OBD_MD_FLBLKSZ && src->o_blksize > (1 << dst->i_blkbits)) obdo_refresh_inode() 168 dst->i_blkbits = ffs(src->o_blksize) - 1; obdo_refresh_inode() 174 if (valid & OBD_MD_FLBLOCKS && src->o_blocks > dst->i_blocks) obdo_refresh_inode() 179 dst->i_blocks = src->o_blocks; obdo_refresh_inode() 183 void obdo_to_inode(struct inode *dst, struct obdo *src, u32 valid) obdo_to_inode() argument 185 valid &= src->o_valid; obdo_to_inode() 189 "object "DOSTID", valid %x\n", POSTID(&src->o_oi), valid); obdo_to_inode() 194 src->o_valid, LTIME_S(dst->i_mtime), obdo_to_inode() 195 LTIME_S(dst->i_ctime), src->o_mtime, src->o_ctime); obdo_to_inode() 198 LTIME_S(dst->i_atime) = src->o_atime; obdo_to_inode() 200 LTIME_S(dst->i_mtime) = src->o_mtime; obdo_to_inode() 201 if (valid & OBD_MD_FLCTIME && src->o_ctime > LTIME_S(dst->i_ctime)) obdo_to_inode() 202 LTIME_S(dst->i_ctime) = src->o_ctime; obdo_to_inode() 204 i_size_write(dst, src->o_size); obdo_to_inode() 206 dst->i_blocks = src->o_blocks; obdo_to_inode() 207 if (dst->i_blocks < src->o_blocks) /* overflow */ obdo_to_inode() 212 dst->i_blkbits = ffs(src->o_blksize)-1; obdo_to_inode() 214 dst->i_mode = (dst->i_mode & S_IFMT) | (src->o_mode & ~S_IFMT); obdo_to_inode() 216 dst->i_uid = make_kuid(&init_user_ns, src->o_uid); obdo_to_inode() 218 dst->i_gid = make_kgid(&init_user_ns, src->o_gid); obdo_to_inode() 220 dst->i_flags = src->o_flags; obdo_to_inode()
|
/linux-4.1.27/arch/sh/kernel/ |
H A D | iomap.c | 107 static inline void mmio_outsb(void __iomem *addr, const u8 *src, int count) mmio_outsb() argument 110 __raw_writeb(*src, addr); mmio_outsb() 111 src++; mmio_outsb() 115 static inline void mmio_outsw(void __iomem *addr, const u16 *src, int count) mmio_outsw() argument 118 __raw_writew(*src, addr); mmio_outsw() 119 src++; mmio_outsw() 123 static inline void mmio_outsl(void __iomem *addr, const u32 *src, int count) mmio_outsl() argument 126 __raw_writel(*src, addr); mmio_outsl() 127 src++; mmio_outsl() 149 void iowrite8_rep(void __iomem *addr, const void *src, unsigned long count) iowrite8_rep() argument 151 mmio_outsb(addr, src, count); iowrite8_rep() 155 void iowrite16_rep(void __iomem *addr, const void *src, unsigned long count) iowrite16_rep() argument 157 mmio_outsw(addr, src, count); iowrite16_rep() 161 void iowrite32_rep(void __iomem *addr, const void *src, unsigned long count) iowrite32_rep() argument 163 mmio_outsl(addr, src, count); iowrite32_rep()
|
H A D | process.c | 21 int arch_dup_task_struct(struct task_struct *dst, struct task_struct *src) arch_dup_task_struct() argument 24 unlazy_fpu(src, task_pt_regs(src)); arch_dup_task_struct() 26 *dst = *src; arch_dup_task_struct() 28 if (src->thread.xstate) { arch_dup_task_struct() 33 memcpy(dst->thread.xstate, src->thread.xstate, xstate_size); arch_dup_task_struct()
|
/linux-4.1.27/fs/quota/ |
H A D | quota.c | 189 static void copy_to_if_dqblk(struct if_dqblk *dst, struct qc_dqblk *src) copy_to_if_dqblk() argument 192 dst->dqb_bhardlimit = stoqb(src->d_spc_hardlimit); copy_to_if_dqblk() 193 dst->dqb_bsoftlimit = stoqb(src->d_spc_softlimit); copy_to_if_dqblk() 194 dst->dqb_curspace = src->d_space; copy_to_if_dqblk() 195 dst->dqb_ihardlimit = src->d_ino_hardlimit; copy_to_if_dqblk() 196 dst->dqb_isoftlimit = src->d_ino_softlimit; copy_to_if_dqblk() 197 dst->dqb_curinodes = src->d_ino_count; copy_to_if_dqblk() 198 dst->dqb_btime = src->d_spc_timer; copy_to_if_dqblk() 199 dst->dqb_itime = src->d_ino_timer; copy_to_if_dqblk() 225 static void copy_from_if_dqblk(struct qc_dqblk *dst, struct if_dqblk *src) copy_from_if_dqblk() argument 227 dst->d_spc_hardlimit = qbtos(src->dqb_bhardlimit); copy_from_if_dqblk() 228 dst->d_spc_softlimit = qbtos(src->dqb_bsoftlimit); copy_from_if_dqblk() 229 dst->d_space = src->dqb_curspace; copy_from_if_dqblk() 230 dst->d_ino_hardlimit = src->dqb_ihardlimit; copy_from_if_dqblk() 231 dst->d_ino_softlimit = src->dqb_isoftlimit; copy_from_if_dqblk() 232 dst->d_ino_count = src->dqb_curinodes; copy_from_if_dqblk() 233 dst->d_spc_timer = src->dqb_btime; copy_from_if_dqblk() 234 dst->d_ino_timer = src->dqb_itime; copy_from_if_dqblk() 237 if (src->dqb_valid & QIF_BLIMITS) copy_from_if_dqblk() 239 if (src->dqb_valid & QIF_SPACE) copy_from_if_dqblk() 241 if (src->dqb_valid & QIF_ILIMITS) copy_from_if_dqblk() 243 if (src->dqb_valid & QIF_INODES) copy_from_if_dqblk() 245 if (src->dqb_valid & QIF_BTIME) copy_from_if_dqblk() 247 if (src->dqb_valid & QIF_ITIME) copy_from_if_dqblk() 470 static void copy_from_xfs_dqblk(struct qc_dqblk *dst, struct fs_disk_quota *src) copy_from_xfs_dqblk() argument 472 dst->d_spc_hardlimit = quota_bbtob(src->d_blk_hardlimit); copy_from_xfs_dqblk() 473 dst->d_spc_softlimit = quota_bbtob(src->d_blk_softlimit); copy_from_xfs_dqblk() 474 dst->d_ino_hardlimit = src->d_ino_hardlimit; copy_from_xfs_dqblk() 475 dst->d_ino_softlimit = src->d_ino_softlimit; copy_from_xfs_dqblk() 476 dst->d_space = quota_bbtob(src->d_bcount); copy_from_xfs_dqblk() 477 dst->d_ino_count = src->d_icount; copy_from_xfs_dqblk() 478 dst->d_ino_timer = src->d_itimer; copy_from_xfs_dqblk() 479 dst->d_spc_timer = src->d_btimer; copy_from_xfs_dqblk() 480 dst->d_ino_warns = src->d_iwarns; copy_from_xfs_dqblk() 481 dst->d_spc_warns = src->d_bwarns; copy_from_xfs_dqblk() 482 dst->d_rt_spc_hardlimit = quota_bbtob(src->d_rtb_hardlimit); copy_from_xfs_dqblk() 483 dst->d_rt_spc_softlimit = quota_bbtob(src->d_rtb_softlimit); copy_from_xfs_dqblk() 484 dst->d_rt_space = quota_bbtob(src->d_rtbcount); copy_from_xfs_dqblk() 485 dst->d_rt_spc_timer = src->d_rtbtimer; copy_from_xfs_dqblk() 486 dst->d_rt_spc_warns = src->d_rtbwarns; copy_from_xfs_dqblk() 488 if (src->d_fieldmask & FS_DQ_ISOFT) copy_from_xfs_dqblk() 490 if (src->d_fieldmask & FS_DQ_IHARD) copy_from_xfs_dqblk() 492 if (src->d_fieldmask & FS_DQ_BSOFT) copy_from_xfs_dqblk() 494 if (src->d_fieldmask & FS_DQ_BHARD) copy_from_xfs_dqblk() 496 if (src->d_fieldmask & FS_DQ_RTBSOFT) copy_from_xfs_dqblk() 498 if (src->d_fieldmask & FS_DQ_RTBHARD) copy_from_xfs_dqblk() 500 if (src->d_fieldmask & FS_DQ_BTIMER) copy_from_xfs_dqblk() 502 if (src->d_fieldmask & FS_DQ_ITIMER) copy_from_xfs_dqblk() 504 if (src->d_fieldmask & FS_DQ_RTBTIMER) copy_from_xfs_dqblk() 506 if (src->d_fieldmask & FS_DQ_BWARNS) copy_from_xfs_dqblk() 508 if (src->d_fieldmask & FS_DQ_IWARNS) copy_from_xfs_dqblk() 510 if (src->d_fieldmask & FS_DQ_RTBWARNS) copy_from_xfs_dqblk() 512 if (src->d_fieldmask & FS_DQ_BCOUNT) copy_from_xfs_dqblk() 514 if (src->d_fieldmask & FS_DQ_ICOUNT) copy_from_xfs_dqblk() 516 if (src->d_fieldmask & FS_DQ_RTBCOUNT) copy_from_xfs_dqblk() 521 struct fs_disk_quota *src) copy_qcinfo_from_xfs_dqblk() 524 dst->i_spc_timelimit = src->d_btimer; copy_qcinfo_from_xfs_dqblk() 525 dst->i_ino_timelimit = src->d_itimer; copy_qcinfo_from_xfs_dqblk() 526 dst->i_rt_spc_timelimit = src->d_rtbtimer; copy_qcinfo_from_xfs_dqblk() 527 dst->i_ino_warnlimit = src->d_iwarns; copy_qcinfo_from_xfs_dqblk() 528 dst->i_spc_warnlimit = src->d_bwarns; copy_qcinfo_from_xfs_dqblk() 529 dst->i_rt_spc_warnlimit = src->d_rtbwarns; copy_qcinfo_from_xfs_dqblk() 530 if (src->d_fieldmask & FS_DQ_BWARNS) copy_qcinfo_from_xfs_dqblk() 532 if (src->d_fieldmask & FS_DQ_IWARNS) copy_qcinfo_from_xfs_dqblk() 534 if (src->d_fieldmask & FS_DQ_RTBWARNS) copy_qcinfo_from_xfs_dqblk() 536 if (src->d_fieldmask & FS_DQ_BTIMER) copy_qcinfo_from_xfs_dqblk() 538 if (src->d_fieldmask & FS_DQ_ITIMER) copy_qcinfo_from_xfs_dqblk() 540 if (src->d_fieldmask & FS_DQ_RTBTIMER) copy_qcinfo_from_xfs_dqblk() 577 static void copy_to_xfs_dqblk(struct fs_disk_quota *dst, struct qc_dqblk *src, copy_to_xfs_dqblk() argument 589 dst->d_blk_hardlimit = quota_btobb(src->d_spc_hardlimit); copy_to_xfs_dqblk() 590 dst->d_blk_softlimit = quota_btobb(src->d_spc_softlimit); copy_to_xfs_dqblk() 591 dst->d_ino_hardlimit = src->d_ino_hardlimit; copy_to_xfs_dqblk() 592 dst->d_ino_softlimit = src->d_ino_softlimit; copy_to_xfs_dqblk() 593 dst->d_bcount = quota_btobb(src->d_space); copy_to_xfs_dqblk() 594 dst->d_icount = src->d_ino_count; copy_to_xfs_dqblk() 595 dst->d_itimer = src->d_ino_timer; copy_to_xfs_dqblk() 596 dst->d_btimer = src->d_spc_timer; copy_to_xfs_dqblk() 597 dst->d_iwarns = src->d_ino_warns; copy_to_xfs_dqblk() 598 dst->d_bwarns = src->d_spc_warns; copy_to_xfs_dqblk() 599 dst->d_rtb_hardlimit = quota_btobb(src->d_rt_spc_hardlimit); copy_to_xfs_dqblk() 600 dst->d_rtb_softlimit = quota_btobb(src->d_rt_spc_softlimit); copy_to_xfs_dqblk() 601 dst->d_rtbcount = quota_btobb(src->d_rt_space); copy_to_xfs_dqblk() 602 dst->d_rtbtimer = src->d_rt_spc_timer; copy_to_xfs_dqblk() 603 dst->d_rtbwarns = src->d_rt_spc_warns; copy_to_xfs_dqblk() 520 copy_qcinfo_from_xfs_dqblk(struct qc_info *dst, struct fs_disk_quota *src) copy_qcinfo_from_xfs_dqblk() argument
|
/linux-4.1.27/drivers/of/ |
H A D | Makefile | 20 CFLAGS_fdt.o = -I$(src)/../../scripts/dtc/libfdt 21 CFLAGS_fdt_address.o = -I$(src)/../../scripts/dtc/libfdt
|
/linux-4.1.27/arch/mips/boot/compressed/ |
H A D | string.c | 9 void *memcpy(void *dest, const void *src, size_t n) memcpy() argument 12 const char *s = src; memcpy()
|
/linux-4.1.27/crypto/ |
H A D | cbc.c | 50 u8 *src = walk->src.virt.addr; crypto_cbc_encrypt_segment() local 55 crypto_xor(iv, src, bsize); crypto_cbc_encrypt_segment() 59 src += bsize; crypto_cbc_encrypt_segment() 74 u8 *src = walk->src.virt.addr; crypto_cbc_encrypt_inplace() local 78 crypto_xor(src, iv, bsize); crypto_cbc_encrypt_inplace() 79 fn(crypto_cipher_tfm(tfm), src, src); crypto_cbc_encrypt_inplace() local 80 iv = src; crypto_cbc_encrypt_inplace() 82 src += bsize; crypto_cbc_encrypt_inplace() 91 struct scatterlist *dst, struct scatterlist *src, crypto_cbc_encrypt() 100 blkcipher_walk_init(&walk, dst, src, nbytes); crypto_cbc_encrypt() 104 if (walk.src.virt.addr == walk.dst.virt.addr) crypto_cbc_encrypt() 122 u8 *src = walk->src.virt.addr; crypto_cbc_decrypt_segment() local 127 fn(crypto_cipher_tfm(tfm), dst, src); crypto_cbc_decrypt_segment() local 129 iv = src; crypto_cbc_decrypt_segment() 131 src += bsize; crypto_cbc_decrypt_segment() 148 u8 *src = walk->src.virt.addr; crypto_cbc_decrypt_inplace() local 152 src += nbytes - (nbytes & (bsize - 1)) - bsize; crypto_cbc_decrypt_inplace() 153 memcpy(last_iv, src, bsize); crypto_cbc_decrypt_inplace() 156 fn(crypto_cipher_tfm(tfm), src, src); crypto_cbc_decrypt_inplace() local 159 crypto_xor(src, src - bsize, bsize); crypto_cbc_decrypt_inplace() 160 src -= bsize; crypto_cbc_decrypt_inplace() 163 crypto_xor(src, walk->iv, bsize); crypto_cbc_decrypt_inplace() 170 struct scatterlist *dst, struct scatterlist *src, crypto_cbc_decrypt() 179 blkcipher_walk_init(&walk, dst, src, nbytes); crypto_cbc_decrypt() 183 if (walk.src.virt.addr == walk.dst.virt.addr) crypto_cbc_decrypt() 90 crypto_cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) crypto_cbc_encrypt() argument 169 crypto_cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) crypto_cbc_decrypt() argument
|
H A D | pcbc.c | 53 u8 *src = walk->src.virt.addr; crypto_pcbc_encrypt_segment() local 58 crypto_xor(iv, src, bsize); crypto_pcbc_encrypt_segment() 61 crypto_xor(iv, src, bsize); crypto_pcbc_encrypt_segment() 63 src += bsize; crypto_pcbc_encrypt_segment() 78 u8 *src = walk->src.virt.addr; crypto_pcbc_encrypt_inplace() local 83 memcpy(tmpbuf, src, bsize); crypto_pcbc_encrypt_inplace() 84 crypto_xor(iv, src, bsize); crypto_pcbc_encrypt_inplace() 85 fn(crypto_cipher_tfm(tfm), src, iv); crypto_pcbc_encrypt_inplace() local 87 crypto_xor(iv, src, bsize); crypto_pcbc_encrypt_inplace() 89 src += bsize; crypto_pcbc_encrypt_inplace() 98 struct scatterlist *dst, struct scatterlist *src, crypto_pcbc_encrypt() 107 blkcipher_walk_init(&walk, dst, src, nbytes); crypto_pcbc_encrypt() 111 if (walk.src.virt.addr == walk.dst.virt.addr) crypto_pcbc_encrypt() 131 u8 *src = walk->src.virt.addr; crypto_pcbc_decrypt_segment() local 136 fn(crypto_cipher_tfm(tfm), dst, src); crypto_pcbc_decrypt_segment() local 138 memcpy(iv, src, bsize); crypto_pcbc_decrypt_segment() 141 src += bsize; crypto_pcbc_decrypt_segment() 158 u8 *src = walk->src.virt.addr; crypto_pcbc_decrypt_inplace() local 163 memcpy(tmpbuf, src, bsize); crypto_pcbc_decrypt_inplace() 164 fn(crypto_cipher_tfm(tfm), src, src); crypto_pcbc_decrypt_inplace() local 165 crypto_xor(src, iv, bsize); crypto_pcbc_decrypt_inplace() 167 crypto_xor(iv, src, bsize); crypto_pcbc_decrypt_inplace() 169 src += bsize; crypto_pcbc_decrypt_inplace() 178 struct scatterlist *dst, struct scatterlist *src, crypto_pcbc_decrypt() 187 blkcipher_walk_init(&walk, dst, src, nbytes); crypto_pcbc_decrypt() 191 if (walk.src.virt.addr == walk.dst.virt.addr) crypto_pcbc_decrypt() 97 crypto_pcbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) crypto_pcbc_encrypt() argument 177 crypto_pcbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) crypto_pcbc_decrypt() argument
|
H A D | compress.c | 21 const u8 *src, unsigned int slen, crypto_compress() 24 return tfm->__crt_alg->cra_compress.coa_compress(tfm, src, slen, dst, crypto_compress() 29 const u8 *src, unsigned int slen, crypto_decompress() 32 return tfm->__crt_alg->cra_compress.coa_decompress(tfm, src, slen, dst, crypto_decompress() 20 crypto_compress(struct crypto_tfm *tfm, const u8 *src, unsigned int slen, u8 *dst, unsigned int *dlen) crypto_compress() argument 28 crypto_decompress(struct crypto_tfm *tfm, const u8 *src, unsigned int slen, u8 *dst, unsigned int *dlen) crypto_decompress() argument
|
H A D | cipher.c | 66 u8 *dst, const u8 *src) cipher_crypt_unaligned() 73 memcpy(tmp, src, size); cipher_crypt_unaligned() 79 u8 *dst, const u8 *src) cipher_encrypt_unaligned() 84 if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { cipher_encrypt_unaligned() 85 cipher_crypt_unaligned(cipher->cia_encrypt, tfm, dst, src); cipher_encrypt_unaligned() 89 cipher->cia_encrypt(tfm, dst, src); cipher_encrypt_unaligned() 93 u8 *dst, const u8 *src) cipher_decrypt_unaligned() 98 if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { cipher_decrypt_unaligned() 99 cipher_crypt_unaligned(cipher->cia_decrypt, tfm, dst, src); cipher_decrypt_unaligned() 103 cipher->cia_decrypt(tfm, dst, src); cipher_decrypt_unaligned() 63 cipher_crypt_unaligned(void (*fn)(struct crypto_tfm *, u8 *, const u8 *), struct crypto_tfm *tfm, u8 *dst, const u8 *src) cipher_crypt_unaligned() argument 78 cipher_encrypt_unaligned(struct crypto_tfm *tfm, u8 *dst, const u8 *src) cipher_encrypt_unaligned() argument 92 cipher_decrypt_unaligned(struct crypto_tfm *tfm, u8 *dst, const u8 *src) cipher_decrypt_unaligned() argument
|
H A D | crypto_null.c | 28 static int null_compress(struct crypto_tfm *tfm, const u8 *src, null_compress() argument 33 memcpy(dst, src, slen); null_compress() 68 static void null_crypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) null_crypt() argument 70 memcpy(dst, src, NULL_BLOCK_SIZE); null_crypt() 75 struct scatterlist *src, unsigned int nbytes) skcipher_null_crypt() 80 blkcipher_walk_init(&walk, dst, src, nbytes); skcipher_null_crypt() 84 if (walk.src.virt.addr != walk.dst.virt.addr) skcipher_null_crypt() 85 memcpy(walk.dst.virt.addr, walk.src.virt.addr, skcipher_null_crypt() 73 skcipher_null_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) skcipher_null_crypt() argument
|
/linux-4.1.27/arch/arm/mach-at91/ |
H A D | Makefile | 4 ccflags-$(CONFIG_ARCH_MULTIPLATFORM) := -I$(srctree)/$(src)/include 5 asflags-$(CONFIG_ARCH_MULTIPLATFORM) := -I$(srctree)/$(src)/include
|
/linux-4.1.27/scripts/genksyms/ |
H A D | Makefile | 8 HOSTCFLAGS_parse.tab.o := -I$(src) 9 HOSTCFLAGS_lex.lex.o := -I$(src)
|
/linux-4.1.27/tools/testing/selftests/powerpc/copyloops/ |
H A D | validate.c | 16 static void do_one(char *src, char *dst, unsigned long src_off, do_one() argument 24 srcp = src + MIN_REDZONE + src_off; do_one() 27 memset(src, POISON, BUFLEN); do_one() 39 printf("src: "); do_one() 64 char *src, *dst, *redzone, *fill; test_copy_loop() local 68 src = memalign(BUFLEN, BUFLEN); test_copy_loop() 73 if (!src || !dst || !redzone || !fill) { test_copy_loop() 87 do_one(src, dst, src_off, dst_off, len, test_copy_loop()
|
/linux-4.1.27/include/crypto/ |
H A D | cast5.h | 20 void __cast5_encrypt(struct cast5_ctx *ctx, u8 *dst, const u8 *src); 21 void __cast5_decrypt(struct cast5_ctx *ctx, u8 *dst, const u8 *src);
|
H A D | cast6.h | 21 void __cast6_encrypt(struct cast6_ctx *ctx, u8 *dst, const u8 *src); 22 void __cast6_decrypt(struct cast6_ctx *ctx, u8 *dst, const u8 *src);
|
H A D | serpent.h | 24 void __serpent_encrypt(struct serpent_ctx *ctx, u8 *dst, const u8 *src); 25 void __serpent_decrypt(struct serpent_ctx *ctx, u8 *dst, const u8 *src);
|
H A D | xts.h | 16 void (*tweak_fn)(void *ctx, u8* dst, const u8* src); 24 struct scatterlist *src, unsigned int nbytes,
|
/linux-4.1.27/fs/xfs/ |
H A D | xfs_icache.h | 81 struct xfs_fs_eofblocks *src, xfs_fs_eofblocks_from_user() 84 if (src->eof_version != XFS_EOFBLOCKS_VERSION) xfs_fs_eofblocks_from_user() 87 if (src->eof_flags & ~XFS_EOF_FLAGS_VALID) xfs_fs_eofblocks_from_user() 90 if (memchr_inv(&src->pad32, 0, sizeof(src->pad32)) || xfs_fs_eofblocks_from_user() 91 memchr_inv(src->pad64, 0, sizeof(src->pad64))) xfs_fs_eofblocks_from_user() 94 dst->eof_flags = src->eof_flags; xfs_fs_eofblocks_from_user() 95 dst->eof_prid = src->eof_prid; xfs_fs_eofblocks_from_user() 96 dst->eof_min_file_size = src->eof_min_file_size; xfs_fs_eofblocks_from_user() 100 if (src->eof_flags & XFS_EOF_FLAGS_UID) { xfs_fs_eofblocks_from_user() 101 dst->eof_uid = make_kuid(current_user_ns(), src->eof_uid); xfs_fs_eofblocks_from_user() 107 if (src->eof_flags & XFS_EOF_FLAGS_GID) { xfs_fs_eofblocks_from_user() 108 dst->eof_gid = make_kgid(current_user_ns(), src->eof_gid); xfs_fs_eofblocks_from_user() 80 xfs_fs_eofblocks_from_user( struct xfs_fs_eofblocks *src, struct xfs_eofblocks *dst) xfs_fs_eofblocks_from_user() argument
|
/linux-4.1.27/arch/mn10300/lib/ |
H A D | checksum.c | 54 __wsum csum_partial_copy(const void *src, void *dst, int len, __wsum sum) csum_partial_copy() argument 56 copy_from_user(dst, src, len); csum_partial_copy() 61 __wsum csum_partial_copy_nocheck(const void *src, void *dst, csum_partial_copy_nocheck() argument 64 sum = csum_partial(src, len, sum); csum_partial_copy_nocheck() 65 memcpy(dst, src, len); csum_partial_copy_nocheck() 70 __wsum csum_partial_copy_from_user(const void *src, void *dst, csum_partial_copy_from_user() argument 76 missing = copy_from_user(dst, src, len); csum_partial_copy_from_user() 86 __wsum csum_and_copy_to_user(const void *src, void *dst, csum_and_copy_to_user() argument 92 missing = copy_to_user(dst, src, len); csum_and_copy_to_user() 98 return csum_partial(src, len, sum); csum_and_copy_to_user()
|
H A D | usercopy.c | 33 #define __do_strncpy_from_user(dst, src, count, res) \ 64 :"i"(-EFAULT), "1"(count), "a"(src), "a"(dst) \ 69 __strncpy_from_user(char *dst, const char *src, long count) __strncpy_from_user() argument 72 __do_strncpy_from_user(dst, src, count, res); __strncpy_from_user() 77 strncpy_from_user(char *dst, const char *src, long count) strncpy_from_user() argument 80 if (access_ok(VERIFY_READ, src, 1)) strncpy_from_user() 81 __do_strncpy_from_user(dst, src, count, res); strncpy_from_user()
|
/linux-4.1.27/arch/s390/include/asm/ |
H A D | pci_io.h | 122 static inline u8 zpci_get_max_write_size(u64 src, u64 dst, int len, int max) zpci_get_max_write_size() argument 126 while (!(src & 0x1) && !(dst & 0x1) && ((size << 1) <= count)) { zpci_get_max_write_size() 128 src = src >> 1; zpci_get_max_write_size() 135 const volatile void __iomem *src, zpci_memcpy_fromio() 138 struct zpci_iomap_entry *entry = &zpci_iomap_start[ZPCI_IDX(src)]; zpci_memcpy_fromio() 139 u64 req, offset = ZPCI_OFFSET(src); zpci_memcpy_fromio() 143 size = zpci_get_max_write_size((u64 __force) src, zpci_memcpy_fromio() 157 const void *src, unsigned long n) zpci_memcpy_toio() 163 if (!src) zpci_memcpy_toio() 168 (u64) src, n, 128); zpci_memcpy_toio() 172 rc = zpci_write_block(req, src, offset); zpci_memcpy_toio() 174 rc = zpci_write_single(req, src, offset, size); zpci_memcpy_toio() 178 src += size; zpci_memcpy_toio() 187 u8 *src = kmalloc(count, GFP_KERNEL); zpci_memset_io() local 190 if (src == NULL) zpci_memset_io() 192 memset(src, val, count); zpci_memset_io() 194 rc = zpci_memcpy_toio(dst, src, count); zpci_memset_io() 195 kfree(src); zpci_memset_io() 134 zpci_memcpy_fromio(void *dst, const volatile void __iomem *src, unsigned long n) zpci_memcpy_fromio() argument 156 zpci_memcpy_toio(volatile void __iomem *dst, const void *src, unsigned long n) zpci_memcpy_toio() argument
|
/linux-4.1.27/arch/x86/lib/ |
H A D | csum-wrappers_64.c | 13 * @src: source address (user space) 20 * src and dst are best aligned to 64bits. 23 csum_partial_copy_from_user(const void __user *src, void *dst, csum_partial_copy_from_user() argument 29 if (!likely(access_ok(VERIFY_READ, src, len))) csum_partial_copy_from_user() 40 if (unlikely((unsigned long)src & 6)) { csum_partial_copy_from_user() 41 while (((unsigned long)src & 6) && len >= 2) { csum_partial_copy_from_user() 44 if (__get_user(val16, (const __u16 __user *)src)) csum_partial_copy_from_user() 50 src += 2; csum_partial_copy_from_user() 56 isum = csum_partial_copy_generic((__force const void *)src, csum_partial_copy_from_user() 74 * @src: source address 81 * src and dst are best aligned to 64bits. 84 csum_partial_copy_to_user(const void *src, void __user *dst, csum_partial_copy_to_user() argument 98 __u16 val16 = *(__u16 *)src; csum_partial_copy_to_user() 105 src += 2; csum_partial_copy_to_user() 113 ret = csum_partial_copy_generic(src, (void __force *)dst, csum_partial_copy_to_user() 122 * @src: source address 130 csum_partial_copy_nocheck(const void *src, void *dst, int len, __wsum sum) csum_partial_copy_nocheck() argument 132 return csum_partial_copy_generic(src, dst, len, sum, NULL, NULL); csum_partial_copy_nocheck()
|
H A D | csum-partial_64.c | 71 asm("addq 0*8(%[src]),%[res]\n\t" do_csum() 72 "adcq 1*8(%[src]),%[res]\n\t" do_csum() 73 "adcq 2*8(%[src]),%[res]\n\t" do_csum() 74 "adcq 3*8(%[src]),%[res]\n\t" do_csum() 75 "adcq 4*8(%[src]),%[res]\n\t" do_csum() 76 "adcq 5*8(%[src]),%[res]\n\t" do_csum() 77 "adcq 6*8(%[src]),%[res]\n\t" do_csum() 78 "adcq 7*8(%[src]),%[res]\n\t" do_csum() 81 : [src] "r" (buff), [zero] "r" (zero), do_csum()
|
/linux-4.1.27/drivers/staging/rtl8192u/ieee80211/ |
H A D | dot11d.h | 37 #define cpMacAddr(des, src) ((des)[0] = (src)[0], \ 38 (des)[1] = (src)[1], (des)[2] = (src)[2], \ 39 (des)[3] = (src)[3], (des)[4] = (src)[4], \ 40 (des)[5] = (src)[5])
|
/linux-4.1.27/arch/tile/lib/ |
H A D | memmove.c | 19 void *memmove(void *dest, const void *src, size_t n) memmove() argument 21 if ((const char *)src >= (char *)dest + n memmove() 22 || (char *)dest >= (const char *)src + n) { memmove() 26 return memcpy(dest, src, n); memmove() 35 if (src < dest) { memmove() 37 in = (const uint8_t *)src + n - 1; memmove() 42 in = (const uint8_t *)src; memmove()
|
/linux-4.1.27/lib/ |
H A D | ucs2_string.c | 54 ucs2_utf8size(const ucs2_char_t *src) ucs2_utf8size() argument 59 for (i = 0; i < ucs2_strlen(src); i++) { ucs2_utf8size() 60 u16 c = src[i]; ucs2_utf8size() 76 * ucs2 string src. 82 ucs2_as_utf8(u8 *dest, const ucs2_char_t *src, unsigned long maxlength) ucs2_as_utf8() argument 86 unsigned long limit = ucs2_strnlen(src, maxlength); ucs2_as_utf8() 89 u16 c = src[i]; ucs2_as_utf8()
|
H A D | strncpy_from_user.c | 11 #define IS_UNALIGNED(src, dst) 0 13 #define IS_UNALIGNED(src, dst) \ 14 (((long) dst | (long) src) & (sizeof(long) - 1)) 23 static inline long do_strncpy_from_user(char *dst, const char __user *src, long count, unsigned long max) do_strncpy_from_user() argument 35 if (IS_UNALIGNED(src, dst)) do_strncpy_from_user() 42 if (unlikely(__get_user(c,(unsigned long __user *)(src+res)))) do_strncpy_from_user() 58 if (unlikely(__get_user(c,src+res))) do_strncpy_from_user() 85 * @src: Source address, in user space. 99 long strncpy_from_user(char *dst, const char __user *src, long count) strncpy_from_user() argument 107 src_addr = (unsigned long)src; strncpy_from_user() 110 return do_strncpy_from_user(dst, src, count, max); strncpy_from_user()
|
H A D | iomap_copy.c | 36 const u32 *src = from; __iowrite32_copy() local 37 const u32 *end = src + count; __iowrite32_copy() 39 while (src < end) __iowrite32_copy() 40 __raw_writel(*src++, dst++); __iowrite32_copy() 60 const u64 *src = from; __iowrite64_copy() local 61 const u64 *end = src + count; __iowrite64_copy() 63 while (src < end) __iowrite64_copy() 64 __raw_writeq(*src++, dst++); __iowrite64_copy()
|
H A D | string_helpers.c | 128 static bool unescape_space(char **src, char **dst) unescape_space() argument 130 char *p = *dst, *q = *src; unescape_space() 152 *src += 1; unescape_space() 156 static bool unescape_octal(char **src, char **dst) unescape_octal() argument 158 char *p = *dst, *q = *src; unescape_octal() 165 while (num < 32 && isodigit(*q) && (q - *src < 3)) { unescape_octal() 171 *src = q; unescape_octal() 175 static bool unescape_hex(char **src, char **dst) unescape_hex() argument 177 char *p = *dst, *q = *src; unescape_hex() 195 *src = q; unescape_hex() 199 static bool unescape_special(char **src, char **dst) unescape_special() argument 201 char *p = *dst, *q = *src; unescape_special() 220 *src += 1; unescape_special() 226 * @src: source buffer (escaped) 262 int string_unescape(char *src, char *dst, size_t size, unsigned int flags) string_unescape() argument 266 while (*src && --size) { string_unescape() 267 if (src[0] == '\\' && src[1] != '\0' && size > 1) { string_unescape() 268 src++; string_unescape() 272 unescape_space(&src, &out)) string_unescape() 276 unescape_octal(&src, &out)) string_unescape() 280 unescape_hex(&src, &out)) string_unescape() 284 unescape_special(&src, &out)) string_unescape() 289 *out++ = *src++; string_unescape() 435 * @src: source buffer (unescaped) 487 int string_escape_mem(const char *src, size_t isz, char *dst, size_t osz, string_escape_mem() argument 495 unsigned char c = *src++; string_escape_mem()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib_pio_copy.c | 49 const u64 *src = from; qib_pio_copy() local 50 const u64 *end = src + (count >> 1); qib_pio_copy() 52 while (src < end) qib_pio_copy() 53 __raw_writeq(*src++, dst++); qib_pio_copy() 55 __raw_writel(*(const u32 *)src, dst); qib_pio_copy() 58 const u32 *src = from; qib_pio_copy() 59 const u32 *end = src + count; qib_pio_copy() 61 while (src < end) qib_pio_copy() 62 __raw_writel(*src++, dst++); qib_pio_copy()
|
/linux-4.1.27/drivers/video/console/ |
H A D | fbcon_rotate.c | 26 const u8 *src; fbcon_rotate_font() local 33 src = ops->fontdata = vc->vc_font.data; fbcon_rotate_font() 35 len = (!ops->p->userfont) ? 256 : FNTCHARCNT(src); fbcon_rotate_font() 67 rotate_ud(src, dst, vc->vc_font.width, fbcon_rotate_font() 70 src += s_cellsize; fbcon_rotate_font() 76 rotate_cw(src, dst, vc->vc_font.width, fbcon_rotate_font() 78 src += s_cellsize; fbcon_rotate_font() 84 rotate_ccw(src, dst, vc->vc_font.width, fbcon_rotate_font() 86 src += s_cellsize; fbcon_rotate_font()
|
H A D | softcursor.c | 29 u8 *src, *dst; soft_cursor() local 48 src = ops->cursor_src + sizeof(struct fb_image); soft_cursor() 61 src[i] = image->data[i] ^ cursor->mask[i]; soft_cursor() 66 src[i] = image->data[i] & cursor->mask[i]; soft_cursor() 70 memcpy(src, image->data, dsize); soft_cursor() 72 fb_pad_aligned_buffer(dst, d_pitch, src, s_pitch, image->height); soft_cursor()
|
/linux-4.1.27/drivers/video/fbdev/ |
H A D | atafb_mfb.c | 28 u8 *src, *dest; atafb_mfb_copyarea() local 32 src = (u8 *)info->screen_base + sy * (width >> 3); atafb_mfb_copyarea() 34 fb_memmove(dest, src, height * (width >> 3)); atafb_mfb_copyarea() 36 src = (u8 *)info->screen_base + sy * next_line + (sx >> 3); atafb_mfb_copyarea() 39 fb_memmove(dest, src, width >> 3); atafb_mfb_copyarea() 40 src += next_line; atafb_mfb_copyarea() 44 src = (u8 *)info->screen_base + (sy + height - 1) * next_line + (sx >> 3); atafb_mfb_copyarea() 47 fb_memmove(dest, src, width >> 3); atafb_mfb_copyarea() 48 src -= next_line; atafb_mfb_copyarea()
|
/linux-4.1.27/arch/m68k/kernel/ |
H A D | relocate_kernel.S | 104 movel %d0,%a3 /* a3 = src = entry & PAGE_MASK */ 107 movel %a3@+,%a2@+ /* *dst++ = *src++ */ 108 movel %a3@+,%a2@+ /* *dst++ = *src++ */ 109 movel %a3@+,%a2@+ /* *dst++ = *src++ */ 110 movel %a3@+,%a2@+ /* *dst++ = *src++ */ 111 movel %a3@+,%a2@+ /* *dst++ = *src++ */ 112 movel %a3@+,%a2@+ /* *dst++ = *src++ */ 113 movel %a3@+,%a2@+ /* *dst++ = *src++ */ 114 movel %a3@+,%a2@+ /* *dst++ = *src++ */
|
/linux-4.1.27/arch/powerpc/crypto/ |
H A D | md5-glue.c | 24 extern void ppc_md5_transform(u32 *state, const u8 *src, u32 blocks); 55 const u8 *src = data; ppc_md5_update() local 60 memcpy((char *)sctx->block + offset, src, len); ppc_md5_update() 65 memcpy((char *)sctx->block + offset, src, avail); ppc_md5_update() 68 src += avail; ppc_md5_update() 72 ppc_md5_transform(sctx->hash, src, len >> 6); ppc_md5_update() 73 src += len & ~0x3f; ppc_md5_update() 77 memcpy((char *)sctx->block, src, len); ppc_md5_update() 85 const u8 *src = (const u8 *)sctx->block; ppc_md5_final() local 86 u8 *p = (u8 *)src + offset; ppc_md5_final() 95 ppc_md5_transform(sctx->hash, src, 1); ppc_md5_final() 102 ppc_md5_transform(sctx->hash, src, 1); ppc_md5_final()
|
/linux-4.1.27/sound/soc/sh/rcar/ |
H A D | src.c | 13 #define SRC_NAME "src" 33 #define rsnd_enable_sync_convert(src) ((src)->sen.val) 35 of_get_child_by_name(rsnd_priv_to_dev(priv)->of_node, "rcar_sound,src") 43 ((pos) = (struct rsnd_src *)(priv)->src + i); \ 58 * src.c is caring... 103 * .playback = { .ssi = &rsnd_ssi[0], .src = &rsnd_src[0] }, 112 * .playback = { .ssi = &rsnd_ssi[0], .src = &rsnd_src[0] }, 238 static u32 rsnd_src_convert_rate(struct rsnd_src *src) rsnd_src_convert_rate() argument 240 struct rsnd_mod *mod = &src->mod; rsnd_src_convert_rate() 248 if (!rsnd_enable_sync_convert(src)) rsnd_src_convert_rate() 249 return src->convert_rate; rsnd_src_convert_rate() 251 convert_rate = src->sync.val; rsnd_src_convert_rate() 254 convert_rate = src->convert_rate; rsnd_src_convert_rate() 267 struct rsnd_src *src; rsnd_src_get_ssi_rate() local 271 src = rsnd_mod_to_src(src_mod); rsnd_src_get_ssi_rate() 277 rate = rsnd_src_convert_rate(src); rsnd_src_get_ssi_rate() 290 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_set_convert_rate() local 291 u32 convert_rate = rsnd_src_convert_rate(src); rsnd_src_set_convert_rate() 322 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_hw_params() local 326 src->convert_rate = src->info->convert_rate; rsnd_src_hw_params() 346 src->convert_rate = params_rate(be_params); rsnd_src_hw_params() 356 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_init() local 360 src->err = 0; rsnd_src_init() 363 src->sync.val = 0; rsnd_src_init() 377 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_quit() local 382 if (src->err) rsnd_src_quit() 384 rsnd_mod_name(mod), rsnd_mod_id(mod), src->err); rsnd_src_quit() 386 src->convert_rate = 0; rsnd_src_quit() 389 src->sync.val = 0; rsnd_src_quit() 455 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_set_convert_timing_gen1() local 457 u32 convert_rate = rsnd_src_convert_rate(src); rsnd_src_set_convert_timing_gen1() 511 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_set_convert_rate_gen1() local 526 if (rsnd_src_convert_rate(src)) rsnd_src_set_convert_rate_gen1() 595 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_irq_ctrol_gen2() local 597 int irq = src->info->irq; rsnd_src_irq_ctrol_gen2() 634 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_error_record_gen2() local 636 src->err++; rsnd_src_error_record_gen2() 683 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_interrupt_gen2() local 690 if (src->err < 1024) rsnd_src_interrupt_gen2() 705 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_set_convert_rate_gen2() local 706 u32 convert_rate = rsnd_src_convert_rate(src); rsnd_src_set_convert_rate_gen2() 733 if (rsnd_enable_sync_convert(src)) { rsnd_src_set_convert_rate_gen2() 764 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_set_convert_timing_gen2() local 765 u32 convert_rate = rsnd_src_convert_rate(src); rsnd_src_set_convert_timing_gen2() 781 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_probe_gen2() local 783 int irq = src->info->irq; rsnd_src_probe_gen2() 802 src->info->dma_id); rsnd_src_probe_gen2() 859 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_reconvert_update() local 860 u32 convert_rate = rsnd_src_convert_rate(src); rsnd_src_reconvert_update() 881 struct rsnd_src *src = rsnd_mod_to_src(mod); rsnd_src_pcm_new() local 915 &src->sen, 1); rsnd_src_pcm_new() 924 &src->sync, 192000); rsnd_src_pcm_new() 947 return &((struct rsnd_src *)(priv->src) + id)->mod; rsnd_src_mod_get() 976 dev_err(dev, "src info allocation error\n"); rsnd_of_parse_src() 1000 struct rsnd_src *src; rsnd_src_probe() local 1025 src = devm_kzalloc(dev, sizeof(*src) * nr, GFP_KERNEL); rsnd_src_probe() 1026 if (!src) { rsnd_src_probe() 1032 priv->src = src; rsnd_src_probe() 1034 for_each_rsnd_src(src, priv, i) { for_each_rsnd_src() 1042 src->info = &info->src_info[i]; for_each_rsnd_src() 1044 ret = rsnd_mod_init(&src->mod, ops, clk, RSND_MOD_SRC, i); for_each_rsnd_src() 1055 struct rsnd_src *src; rsnd_src_remove() local 1058 for_each_rsnd_src(src, priv, i) { for_each_rsnd_src() 1059 rsnd_mod_quit(&src->mod); for_each_rsnd_src()
|
H A D | Makefile | 1 snd-soc-rcar-objs := core.o gen.o dma.o src.o adg.o ssi.o dvc.o
|
/linux-4.1.27/drivers/gpu/drm/ |
H A D | drm_rect.c | 55 * @src: source window rectangle 61 * Clip rectangle @dst by rectangle @clip. Clip rectangle @src by the 68 bool drm_rect_clip_scaled(struct drm_rect *src, struct drm_rect *dst, drm_rect_clip_scaled() argument 76 int64_t tmp = src->x1 + (int64_t) diff * hscale; drm_rect_clip_scaled() 77 src->x1 = clamp_t(int64_t, tmp, INT_MIN, INT_MAX); drm_rect_clip_scaled() 81 int64_t tmp = src->y1 + (int64_t) diff * vscale; drm_rect_clip_scaled() 82 src->y1 = clamp_t(int64_t, tmp, INT_MIN, INT_MAX); drm_rect_clip_scaled() 86 int64_t tmp = src->x2 - (int64_t) diff * hscale; drm_rect_clip_scaled() 87 src->x2 = clamp_t(int64_t, tmp, INT_MIN, INT_MAX); drm_rect_clip_scaled() 91 int64_t tmp = src->y2 - (int64_t) diff * vscale; drm_rect_clip_scaled() 92 src->y2 = clamp_t(int64_t, tmp, INT_MIN, INT_MAX); drm_rect_clip_scaled() 99 static int drm_calc_scale(int src, int dst) drm_calc_scale() argument 103 if (src < 0 || dst < 0) drm_calc_scale() 109 scale = src / dst; drm_calc_scale() 116 * @src: source window rectangle 122 * (@src width) / (@dst width). 127 int drm_rect_calc_hscale(const struct drm_rect *src, drm_rect_calc_hscale() argument 131 int src_w = drm_rect_width(src); drm_rect_calc_hscale() 147 * @src: source window rectangle 153 * (@src height) / (@dst height). 158 int drm_rect_calc_vscale(const struct drm_rect *src, drm_rect_calc_vscale() argument 162 int src_h = drm_rect_height(src); drm_rect_calc_vscale() 178 * @src: source window rectangle 184 * (@src width) / (@dst width). 190 * decrease the height of rectangle @src to compensate. 195 int drm_rect_calc_hscale_relaxed(struct drm_rect *src, drm_rect_calc_hscale_relaxed() argument 199 int src_w = drm_rect_width(src); drm_rect_calc_hscale_relaxed() 217 drm_rect_adjust_size(src, max_src_w - src_w, 0); drm_rect_calc_hscale_relaxed() 228 * @src: source window rectangle 234 * (@src height) / (@dst height). 240 * decrease the height of rectangle @src to compensate. 245 int drm_rect_calc_vscale_relaxed(struct drm_rect *src, drm_rect_calc_vscale_relaxed() argument 249 int src_h = drm_rect_height(src); drm_rect_calc_vscale_relaxed() 267 drm_rect_adjust_size(src, 0, max_src_h - src_h); drm_rect_calc_vscale_relaxed()
|
/linux-4.1.27/arch/s390/lib/ |
H A D | string.c | 66 * @src: Where to copy the string from 70 char *strcpy(char *dest, const char *src) strcpy() argument 77 : "+&a" (dest), "+&a" (src) : "d" (r0) strcpy() 86 * @src: Where to copy the string from 94 size_t strlcpy(char *dest, const char *src, size_t size) strlcpy() argument 96 size_t ret = __strend(src) - src; strlcpy() 101 memcpy(dest, src, len); strlcpy() 110 * @src: Where to copy the string from 116 char *strncpy(char *dest, const char *src, size_t n) strncpy() argument 118 size_t len = __strnend(src, n) - src; strncpy() 120 memcpy(dest, src, len); strncpy() 128 * @src: The string to append to it 132 char *strcat(char *dest, const char *src) strcat() argument 142 : "=&a" (dummy), "+a" (dest), "+a" (src) strcat() 151 * @src: The string to append to it 154 size_t strlcat(char *dest, const char *src, size_t n) strlcat() argument 157 size_t len = __strend(src) - src; strlcat() 166 memcpy(dest, src, len); strlcat() 175 * @src: The string to append to it 183 char *strncat(char *dest, const char *src, size_t n) strncat() argument 185 size_t len = __strnend(src, n) - src; strncat() 189 memcpy(p, src, len); strncat()
|
/linux-4.1.27/fs/hfsplus/ |
H A D | bnode.c | 127 struct hfs_bnode *src_node, int src, int len) hfs_bnode_copy() 133 hfs_dbg(BNODE_MOD, "copybytes: %u,%u,%u\n", dst, src, len); hfs_bnode_copy() 137 src += src_node->page_offset; hfs_bnode_copy() 139 src_page = src_node->page + (src >> PAGE_CACHE_SHIFT); hfs_bnode_copy() 140 src &= ~PAGE_CACHE_MASK; hfs_bnode_copy() 144 if (src == dst) { hfs_bnode_copy() 145 l = min_t(int, len, PAGE_CACHE_SIZE - src); hfs_bnode_copy() 146 memcpy(kmap(*dst_page) + src, kmap(*src_page) + src, l); hfs_bnode_copy() 162 src_ptr = kmap(*src_page) + src; hfs_bnode_copy() 164 if (PAGE_CACHE_SIZE - src < PAGE_CACHE_SIZE - dst) { hfs_bnode_copy() 165 l = PAGE_CACHE_SIZE - src; hfs_bnode_copy() 166 src = 0; hfs_bnode_copy() 170 src += l; hfs_bnode_copy() 186 void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len) hfs_bnode_move() argument 191 hfs_dbg(BNODE_MOD, "movebytes: %u,%u,%u\n", dst, src, len); hfs_bnode_move() 194 src += node->page_offset; hfs_bnode_move() 196 if (dst > src) { hfs_bnode_move() 197 src += len - 1; hfs_bnode_move() 198 src_page = node->page + (src >> PAGE_CACHE_SHIFT); hfs_bnode_move() 199 src = (src & ~PAGE_CACHE_MASK) + 1; hfs_bnode_move() 204 if (src == dst) { hfs_bnode_move() 205 while (src < len) { hfs_bnode_move() 206 memmove(kmap(*dst_page), kmap(*src_page), src); hfs_bnode_move() 210 len -= src; hfs_bnode_move() 211 src = PAGE_CACHE_SIZE; hfs_bnode_move() 215 src -= len; hfs_bnode_move() 216 memmove(kmap(*dst_page) + src, hfs_bnode_move() 217 kmap(*src_page) + src, len); hfs_bnode_move() 225 src_ptr = kmap(*src_page) + src; hfs_bnode_move() 227 if (src < dst) { hfs_bnode_move() 228 l = src; hfs_bnode_move() 229 src = PAGE_CACHE_SIZE; hfs_bnode_move() 233 src -= l; hfs_bnode_move() 248 src_page = node->page + (src >> PAGE_CACHE_SHIFT); hfs_bnode_move() 249 src &= ~PAGE_CACHE_MASK; hfs_bnode_move() 253 if (src == dst) { hfs_bnode_move() 254 l = min_t(int, len, PAGE_CACHE_SIZE - src); hfs_bnode_move() 255 memmove(kmap(*dst_page) + src, hfs_bnode_move() 256 kmap(*src_page) + src, l); hfs_bnode_move() 273 src_ptr = kmap(*src_page) + src; hfs_bnode_move() 275 if (PAGE_CACHE_SIZE - src < hfs_bnode_move() 277 l = PAGE_CACHE_SIZE - src; hfs_bnode_move() 278 src = 0; hfs_bnode_move() 282 src += l; hfs_bnode_move() 126 hfs_bnode_copy(struct hfs_bnode *dst_node, int dst, struct hfs_bnode *src_node, int src, int len) hfs_bnode_copy() argument
|
/linux-4.1.27/arch/arm/crypto/ |
H A D | aesbs-glue.c | 106 struct scatterlist *src, unsigned int nbytes) aesbs_cbc_encrypt() 112 blkcipher_walk_init(&walk, dst, src, nbytes); aesbs_cbc_encrypt() 117 u8 *src = walk.src.virt.addr; aesbs_cbc_encrypt() local 119 if (walk.dst.virt.addr == walk.src.virt.addr) { aesbs_cbc_encrypt() 123 crypto_xor(src, iv, AES_BLOCK_SIZE); aesbs_cbc_encrypt() 124 AES_encrypt(src, src, &ctx->enc); aesbs_cbc_encrypt() 125 iv = src; aesbs_cbc_encrypt() 126 src += AES_BLOCK_SIZE; aesbs_cbc_encrypt() 133 crypto_xor(walk.iv, src, AES_BLOCK_SIZE); aesbs_cbc_encrypt() 136 src += AES_BLOCK_SIZE; aesbs_cbc_encrypt() 147 struct scatterlist *src, unsigned int nbytes) aesbs_cbc_decrypt() 153 blkcipher_walk_init(&walk, dst, src, nbytes); aesbs_cbc_decrypt() 158 bsaes_cbc_encrypt(walk.src.virt.addr, walk.dst.virt.addr, aesbs_cbc_decrypt() 166 u8 *src = walk.src.virt.addr; aesbs_cbc_decrypt() local 171 if (walk.dst.virt.addr == walk.src.virt.addr) aesbs_cbc_decrypt() 172 memcpy(bk[blocks & 1], src, AES_BLOCK_SIZE); aesbs_cbc_decrypt() 174 AES_decrypt(src, dst, &ctx->dec.rk); aesbs_cbc_decrypt() 177 if (walk.dst.virt.addr == walk.src.virt.addr) aesbs_cbc_decrypt() 180 iv = src; aesbs_cbc_decrypt() 183 src += AES_BLOCK_SIZE; aesbs_cbc_decrypt() 204 struct scatterlist *dst, struct scatterlist *src, aesbs_ctr_encrypt() 212 blkcipher_walk_init(&walk, dst, src, nbytes); aesbs_ctr_encrypt() 226 bsaes_ctr32_encrypt_blocks(walk.src.virt.addr, aesbs_ctr_encrypt() 240 u8 *tsrc = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; aesbs_ctr_encrypt() 254 struct scatterlist *src, unsigned int nbytes) aesbs_xts_encrypt() 260 blkcipher_walk_init(&walk, dst, src, nbytes); aesbs_xts_encrypt() 268 bsaes_xts_encrypt(walk.src.virt.addr, walk.dst.virt.addr, aesbs_xts_encrypt() 278 struct scatterlist *src, unsigned int nbytes) aesbs_xts_decrypt() 284 blkcipher_walk_init(&walk, dst, src, nbytes); aesbs_xts_decrypt() 292 bsaes_xts_decrypt(walk.src.virt.addr, walk.dst.virt.addr, aesbs_xts_decrypt() 104 aesbs_cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) aesbs_cbc_encrypt() argument 145 aesbs_cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) aesbs_cbc_decrypt() argument 203 aesbs_ctr_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) aesbs_ctr_encrypt() argument 252 aesbs_xts_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) aesbs_xts_encrypt() argument 276 aesbs_xts_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) aesbs_xts_decrypt() argument
|
/linux-4.1.27/arch/sparc/include/asm/ |
H A D | uaccess.h | 12 long strncpy_from_user(char *dest, const char __user *src, long count);
|
H A D | io_32.h | 26 static inline void _memcpy_fromio(void *dst, const volatile void __iomem *src, _memcpy_fromio() argument 32 char tmp = readb(src); _memcpy_fromio() 34 src++; _memcpy_fromio() 38 static inline void _memcpy_toio(volatile void __iomem *dst, const void *src, _memcpy_toio() argument 41 const char *s = src; _memcpy_toio() 97 const volatile void __iomem *src, sbus_memcpy_fromio() 103 char tmp = sbus_readb(src); sbus_memcpy_fromio() 105 src++; sbus_memcpy_fromio() 110 const void *src, sbus_memcpy_toio() 113 const char *s = src; sbus_memcpy_toio() 96 sbus_memcpy_fromio(void *dst, const volatile void __iomem *src, __kernel_size_t n) sbus_memcpy_fromio() argument 109 sbus_memcpy_toio(volatile void __iomem *dst, const void *src, __kernel_size_t n) sbus_memcpy_toio() argument
|
H A D | ide.h | 65 static inline void __ide_outsw(void __iomem *port, const void *src, u32 count) __ide_outsw() argument 68 unsigned long end = (unsigned long)src + (count << 1); __ide_outsw() 70 const u16 *ps = src; __ide_outsw() 73 if(((unsigned long)src) & 0x2) { __ide_outsw() 91 __flush_dcache_range((unsigned long)src, end); __ide_outsw()
|
/linux-4.1.27/arch/s390/mm/ |
H A D | maccess.c | 19 static notrace long s390_kernel_write_odd(void *dst, const void *src, size_t size) s390_kernel_write_odd() argument 37 : "a" (&tmp), "a" (&tmp[offset]), "a" (src) s390_kernel_write_odd() 45 * @src: source address 58 void notrace s390_kernel_write(void *dst, const void *src, size_t size) s390_kernel_write() argument 63 copied = s390_kernel_write_odd(dst, src, size); s390_kernel_write() 65 src += copied; s390_kernel_write() 70 static int __memcpy_real(void *dest, void *src, size_t count) __memcpy_real() argument 74 register unsigned long _src asm("4") = (unsigned long) src; __memcpy_real() 86 : "m" (*((long *) src)) __memcpy_real() 94 int memcpy_real(void *dest, void *src, size_t count) memcpy_real() argument 103 rc = __memcpy_real(dest, src, count); memcpy_real() 111 void memcpy_absolute(void *dest, void *src, size_t count) memcpy_absolute() argument 122 memcpy(dest, src, count); memcpy_absolute() 126 memcpy(dest, src, count); memcpy_absolute() 135 int copy_to_user_real(void __user *dest, void *src, unsigned long count) copy_to_user_real() argument 146 if (memcpy_real(buf, src + offs, size)) copy_to_user_real()
|
/linux-4.1.27/arch/cris/include/asm/ |
H A D | cacheflush.h | 25 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 26 memcpy(dst, src, len) 27 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 28 memcpy(dst, src, len)
|
/linux-4.1.27/arch/arc/mm/ |
H A D | extable.c | 51 long arc_strncpy_from_user_noinline(char *dst, const char __user *src, arc_strncpy_from_user_noinline() argument 54 return __arc_strncpy_from_user(dst, src, count); arc_strncpy_from_user_noinline() 58 long arc_strnlen_user_noinline(const char __user *src, long n) arc_strnlen_user_noinline() argument 60 return __arc_strnlen_user(src, n); arc_strnlen_user_noinline()
|
/linux-4.1.27/fs/gfs2/ |
H A D | Makefile | 1 ccflags-y := -I$(src)
|
/linux-4.1.27/include/asm-generic/ |
H A D | cacheflush.h | 26 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 28 memcpy(dst, src, len); \ 31 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 32 memcpy(dst, src, len)
|
H A D | checksum.h | 19 * the same as csum_partial, but copies from src while it 22 * here even more important to align src and dst on a 32-bit (or even 25 extern __wsum csum_partial_copy(const void *src, void *dst, int len, __wsum sum); 30 * here even more important to align src and dst on a 32-bit (or even 33 extern __wsum csum_partial_copy_from_user(const void __user *src, void *dst, 37 #define csum_partial_copy_nocheck(src, dst, len, sum) \ 38 csum_partial_copy((src), (dst), (len), (sum))
|
/linux-4.1.27/drivers/gpu/drm/msm/dsi/ |
H A D | sfpb.xml.h | 11 - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 676 bytes, from 2014-12-05 15:34:49) 12 - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) 13 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20908 bytes, from 2014-12-08 16:13:00) 14 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 2357 bytes, from 2014-12-08 16:13:00) 15 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 27208 bytes, from 2015-01-13 23:56:11) 16 - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) 17 - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) 18 - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) 19 - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) 20 - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 26848 bytes, from 2015-01-13 23:55:57) 21 - /home/robclark/src/freedreno/envytools/rnndb/edp/edp.xml ( 8253 bytes, from 2014-12-08 16:13:00)
|
/linux-4.1.27/drivers/gpu/drm/msm/hdmi/ |
H A D | qfprom.xml.h | 11 - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 676 bytes, from 2014-12-05 15:34:49) 12 - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) 13 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20908 bytes, from 2014-12-08 16:13:00) 14 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 2357 bytes, from 2014-12-08 16:13:00) 15 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 27208 bytes, from 2015-01-13 23:56:11) 16 - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) 17 - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) 18 - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) 19 - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) 20 - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 26848 bytes, from 2015-01-13 23:55:57) 21 - /home/robclark/src/freedreno/envytools/rnndb/edp/edp.xml ( 8253 bytes, from 2014-12-08 16:13:00)
|
/linux-4.1.27/drivers/gpu/drm/msm/mdp/ |
H A D | mdp_common.xml.h | 11 - /home/robclark/src/freedreno/envytools/rnndb/msm.xml ( 676 bytes, from 2014-12-05 15:34:49) 12 - /home/robclark/src/freedreno/envytools/rnndb/freedreno_copyright.xml ( 1453 bytes, from 2013-03-31 16:51:27) 13 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp4.xml ( 20908 bytes, from 2014-12-08 16:13:00) 14 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp_common.xml ( 2357 bytes, from 2014-12-08 16:13:00) 15 - /home/robclark/src/freedreno/envytools/rnndb/mdp/mdp5.xml ( 27208 bytes, from 2015-01-13 23:56:11) 16 - /home/robclark/src/freedreno/envytools/rnndb/dsi/dsi.xml ( 11712 bytes, from 2013-08-17 17:13:43) 17 - /home/robclark/src/freedreno/envytools/rnndb/dsi/sfpb.xml ( 344 bytes, from 2013-08-11 19:26:32) 18 - /home/robclark/src/freedreno/envytools/rnndb/dsi/mmss_cc.xml ( 1686 bytes, from 2014-10-31 16:48:57) 19 - /home/robclark/src/freedreno/envytools/rnndb/hdmi/qfprom.xml ( 600 bytes, from 2013-07-05 19:21:12) 20 - /home/robclark/src/freedreno/envytools/rnndb/hdmi/hdmi.xml ( 26848 bytes, from 2015-01-13 23:55:57) 21 - /home/robclark/src/freedreno/envytools/rnndb/edp/edp.xml ( 8253 bytes, from 2014-12-08 16:13:00)
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
H A D | Makefile | 13 $(obj)/rn50_reg_safe.h: $(src)/reg_srcs/rn50 $(obj)/mkregtable 16 $(obj)/r100_reg_safe.h: $(src)/reg_srcs/r100 $(obj)/mkregtable 19 $(obj)/r200_reg_safe.h: $(src)/reg_srcs/r200 $(obj)/mkregtable 22 $(obj)/rv515_reg_safe.h: $(src)/reg_srcs/rv515 $(obj)/mkregtable 25 $(obj)/r300_reg_safe.h: $(src)/reg_srcs/r300 $(obj)/mkregtable 28 $(obj)/r420_reg_safe.h: $(src)/reg_srcs/r420 $(obj)/mkregtable 31 $(obj)/rs600_reg_safe.h: $(src)/reg_srcs/rs600 $(obj)/mkregtable 34 $(obj)/r600_reg_safe.h: $(src)/reg_srcs/r600 $(obj)/mkregtable 37 $(obj)/evergreen_reg_safe.h: $(src)/reg_srcs/evergreen $(obj)/mkregtable 40 $(obj)/cayman_reg_safe.h: $(src)/reg_srcs/cayman $(obj)/mkregtable 118 CFLAGS_radeon_trace_points.o := -I$(src)
|
/linux-4.1.27/fs/hfs/ |
H A D | trans.c | 37 const char *src; hfs_mac2asc() local 41 src = in->name; hfs_mac2asc() 52 size = nls_disk->char2uni(src, srclen, &ch); hfs_mac2asc() 57 src += size; hfs_mac2asc() 60 ch = *src++; hfs_mac2asc() 79 *dst++ = (ch = *src++) == '/' ? ':' : ch; hfs_mac2asc() 101 const char *src; hfs_asc2mac() local 105 src = in->name; hfs_asc2mac() 113 size = nls_io->char2uni(src, srclen, &ch); hfs_asc2mac() 118 src += size; hfs_asc2mac() 143 *dst++ = (ch = *src++) == ':' ? '/' : ch; hfs_asc2mac()
|
/linux-4.1.27/drivers/scsi/lpfc/ |
H A D | lpfc_compat.h | 38 lpfc_memcpy_to_slim(void __iomem *dest, void *src, unsigned int bytes) lpfc_memcpy_to_slim() argument 46 src32 = (uint32_t *) src; lpfc_memcpy_to_slim() 60 lpfc_memcpy_from_slim( void *dest, void __iomem *src, unsigned int bytes) lpfc_memcpy_from_slim() argument 68 src32 = (uint32_t __iomem *) src; lpfc_memcpy_from_slim() 83 lpfc_memcpy_to_slim( void __iomem *dest, void *src, unsigned int bytes) lpfc_memcpy_to_slim() argument 86 __iowrite32_copy(dest, src, bytes / sizeof(uint32_t)); lpfc_memcpy_to_slim() 90 lpfc_memcpy_from_slim( void *dest, void __iomem *src, unsigned int bytes) lpfc_memcpy_from_slim() argument 93 memcpy_fromio( dest, src, bytes); lpfc_memcpy_from_slim()
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/ |
H A D | Makefile | 17 ccflags-y += -D__CHECK_ENDIAN__ -I$(src) 22 CFLAGS_iwl-devtrace.o := -I$(src)
|
/linux-4.1.27/include/uapi/linux/netfilter/ |
H A D | xt_HMARK.h | 27 __u16 src; member in struct:hmark_ports::__anon13470 31 __be16 src; member in struct:hmark_ports::__anon13471
|
/linux-4.1.27/arch/powerpc/lib/ |
H A D | checksum_wrappers_64.c | 26 __wsum csum_and_copy_from_user(const void __user *src, void *dst, csum_and_copy_from_user() argument 40 if (unlikely((len < 0) || !access_ok(VERIFY_READ, src, len))) { csum_and_copy_from_user() 46 csum = csum_partial_copy_generic((void __force *)src, dst, csum_and_copy_from_user() 50 int missing = __copy_from_user(dst, src, len); csum_and_copy_from_user() 67 __wsum csum_and_copy_to_user(const void *src, void __user *dst, int len, csum_and_copy_to_user() argument 87 csum = csum_partial_copy_generic(src, (void __force *)dst, csum_and_copy_to_user() 91 csum = csum_partial(src, len, sum); csum_and_copy_to_user() 93 if (copy_to_user(dst, src, len)) { csum_and_copy_to_user()
|
/linux-4.1.27/arch/hexagon/lib/ |
H A D | io.c | 33 const volatile short int *src = (short int *) addr; __raw_readsw() local 40 *dst++ = *src; __raw_readsw() 52 const short int *src = (short int *)data; __raw_writesw() local 59 *dst = *src++; __raw_writesw() 67 const volatile long *src = (long *) addr; __raw_readsl() local 74 *dst++ = *src; __raw_readsl() 81 const long *src = (long *)data; __raw_writesl() local 88 *dst = *src++; __raw_writesl()
|
/linux-4.1.27/security/tomoyo/ |
H A D | Makefile | 6 $(objtree)/scripts/basic/bin2c <$(firstword $(wildcard $(obj)/policy/$(1).conf $(srctree)/$(src)/policy/$(1).conf.default) /dev/null); \ 12 $(obj)/builtin-policy.h: $(wildcard $(obj)/policy/*.conf $(src)/policy/*.conf.default) FORCE
|
/linux-4.1.27/drivers/net/ethernet/apm/xgene/ |
H A D | xgene_enet_sgmac.h | 24 #define PHY_ADDR(src) (((src)<<8) & GENMASK(12, 8)) 25 #define REG_ADDR(src) ((src) & GENMASK(4, 0)) 26 #define PHY_CONTROL(src) ((src) & GENMASK(15, 0))
|
/linux-4.1.27/arch/arm/mach-w90x900/ |
H A D | clksel.c | 62 void nuc900_clock_source(struct device *dev, unsigned char *src) nuc900_clock_source() argument 67 BUG_ON(!src); nuc900_clock_source() 77 if (strcmp(src, "pll0") == 0) nuc900_clock_source() 79 else if (strcmp(src, "pll1") == 0) nuc900_clock_source() 81 else if (strcmp(src, "ext") == 0) nuc900_clock_source() 83 else if (strcmp(src, "oth") == 0) nuc900_clock_source()
|
/linux-4.1.27/arch/parisc/lib/ |
H A D | io.c | 18 void memcpy_toio(volatile void __iomem *dst, const void *src, int count) memcpy_toio() argument 20 if (((unsigned long)dst & 3) != ((unsigned long)src & 3)) memcpy_toio() 23 writeb(*(char *)src, dst++); memcpy_toio() 24 src++; memcpy_toio() 28 __raw_writel(*(u32 *)src, dst); memcpy_toio() 29 src += 4; memcpy_toio() 35 writeb(*(char *)src, dst++); memcpy_toio() 36 src++; memcpy_toio() 51 ** TODO: only look at src alignment and adjust the stores to dest. 53 void memcpy_fromio(void *dst, const volatile void __iomem *src, int count) memcpy_fromio() argument 55 /* first compare alignment of src/dst */ memcpy_fromio() 56 if ( (((unsigned long)dst ^ (unsigned long)src) & 1) || (count < 2) ) memcpy_fromio() 59 if ( (((unsigned long)dst ^ (unsigned long)src) & 2) || (count < 4) ) memcpy_fromio() 63 if ((unsigned long)src & 1) { memcpy_fromio() 64 *(u8 *)dst = readb(src); memcpy_fromio() 65 src++; memcpy_fromio() 71 if ((unsigned long)src & 2) { memcpy_fromio() 72 *(u16 *)dst = __raw_readw(src); memcpy_fromio() 73 src += 2; memcpy_fromio() 79 *(u32 *)dst = __raw_readl(src); memcpy_fromio() 81 src += 4; memcpy_fromio() 87 *(u16 *)dst = __raw_readw(src); memcpy_fromio() 88 src += 2; memcpy_fromio() 95 *(char *)dst = readb(src); memcpy_fromio() 96 src++; memcpy_fromio() 318 void outsb(unsigned long port, const void * src, unsigned long count) outsb() argument 322 p = (const unsigned char *)src; outsb() 336 void outsw (unsigned long port, const void *src, unsigned long count) outsw() argument 341 p = (const unsigned char *)src; outsw() 408 void outsl (unsigned long port, const void *src, unsigned long count) outsl() argument 413 p = (const unsigned char *)src; outsl()
|
H A D | memcpy.c | 161 /* Copy from a not-aligned src to an aligned dst, using shifts. Handles 4 words 165 unsigned long src, unsigned long len) copy_dstaligned() 173 /* prefetch_src((const void *)src); */ copy_dstaligned() 177 sh_1 = 8 * (src % sizeof(unsigned int)); copy_dstaligned() 180 /* Make src aligned by rounding it down. */ copy_dstaligned() 181 src &= -sizeof(unsigned int); copy_dstaligned() 186 /* a1 = ((unsigned int *) src)[0]; copy_dstaligned() 187 a2 = ((unsigned int *) src)[1]; */ copy_dstaligned() 188 ldw(s_space, 0, src, a1, cda_ldw_exc); copy_dstaligned() 189 ldw(s_space, 4, src, a2, cda_ldw_exc); copy_dstaligned() 190 src -= 1 * sizeof(unsigned int); copy_dstaligned() 195 /* a0 = ((unsigned int *) src)[0]; copy_dstaligned() 196 a1 = ((unsigned int *) src)[1]; */ copy_dstaligned() 197 ldw(s_space, 0, src, a0, cda_ldw_exc); copy_dstaligned() 198 ldw(s_space, 4, src, a1, cda_ldw_exc); copy_dstaligned() 199 src -= 0 * sizeof(unsigned int); copy_dstaligned() 206 /* a3 = ((unsigned int *) src)[0]; copy_dstaligned() 207 a0 = ((unsigned int *) src)[1]; */ copy_dstaligned() 208 ldw(s_space, 0, src, a3, cda_ldw_exc); copy_dstaligned() 209 ldw(s_space, 4, src, a0, cda_ldw_exc); copy_dstaligned() 210 src -=-1 * sizeof(unsigned int); copy_dstaligned() 215 /* a2 = ((unsigned int *) src)[0]; copy_dstaligned() 216 a3 = ((unsigned int *) src)[1]; */ copy_dstaligned() 217 ldw(s_space, 0, src, a2, cda_ldw_exc); copy_dstaligned() 218 ldw(s_space, 4, src, a3, cda_ldw_exc); copy_dstaligned() 219 src -=-2 * sizeof(unsigned int); copy_dstaligned() 229 /* prefetch_src((const void *)(src + 4 * sizeof(unsigned int))); */ copy_dstaligned() 231 /* a0 = ((unsigned int *) src)[0]; */ copy_dstaligned() 232 ldw(s_space, 0, src, a0, cda_ldw_exc); copy_dstaligned() 236 /* a1 = ((unsigned int *) src)[1]; */ copy_dstaligned() 237 ldw(s_space, 4, src, a1, cda_ldw_exc); copy_dstaligned() 241 /* a2 = ((unsigned int *) src)[2]; */ copy_dstaligned() 242 ldw(s_space, 8, src, a2, cda_ldw_exc); copy_dstaligned() 246 /* a3 = ((unsigned int *) src)[3]; */ copy_dstaligned() 247 ldw(s_space, 12, src, a3, cda_ldw_exc); copy_dstaligned() 251 src += 4 * sizeof(unsigned int); copy_dstaligned() 282 register unsigned long src, dst, t1, t2, t3; pa_memcpy_internal() local 288 src = (unsigned long)srcp; pa_memcpy_internal() 299 t1 = (src ^ dst); pa_memcpy_internal() 303 /* src and dst have same alignment. */ pa_memcpy_internal() 306 t2 = src & (sizeof(double) - 1); pa_memcpy_internal() 407 t2 = src & (sizeof(unsigned int) - 1); pa_memcpy_internal() 436 src = (unsigned long)pcs; pa_memcpy_internal() 439 ret = copy_dstaligned(dst, src, len / sizeof(unsigned int)); pa_memcpy_internal() 492 unsigned long copy_to_user(void __user *dst, const void *src, unsigned long len) copy_to_user() argument 496 return pa_memcpy((void __force *)dst, src, len); copy_to_user() 500 unsigned long __copy_from_user(void *dst, const void __user *src, unsigned long len) __copy_from_user() argument 504 return pa_memcpy(dst, (void __force *)src, len); __copy_from_user() 507 unsigned long copy_in_user(void __user *dst, const void __user *src, unsigned long len) copy_in_user() argument 511 return pa_memcpy((void __force *)dst, (void __force *)src, len); copy_in_user() 515 void * memcpy(void * dst,const void *src, size_t count) memcpy() argument 519 pa_memcpy(dst, src, count); memcpy() 528 long probe_kernel_read(void *dst, const void *src, size_t size) probe_kernel_read() argument 530 unsigned long addr = (unsigned long)src; probe_kernel_read() 537 return __probe_kernel_read(dst, src, size); probe_kernel_read() 164 copy_dstaligned(unsigned long dst, unsigned long src, unsigned long len) copy_dstaligned() argument
|
/linux-4.1.27/drivers/crypto/ccp/ |
H A D | ccp-ops.c | 122 struct ccp_mem src; member in struct:ccp_op 273 cr[1] = op->src.u.dma.length - 1; ccp_perform_aes() 274 cr[2] = ccp_addr_lo(&op->src.u.dma); ccp_perform_aes() 277 | ccp_addr_hi(&op->src.u.dma); ccp_perform_aes() 303 cr[1] = op->src.u.dma.length - 1; ccp_perform_xts_aes() 304 cr[2] = ccp_addr_lo(&op->src.u.dma); ccp_perform_xts_aes() 307 | ccp_addr_hi(&op->src.u.dma); ccp_perform_xts_aes() 329 cr[1] = op->src.u.dma.length - 1; ccp_perform_sha() 330 cr[2] = ccp_addr_lo(&op->src.u.dma); ccp_perform_sha() 333 | ccp_addr_hi(&op->src.u.dma); ccp_perform_sha() 357 cr[2] = ccp_addr_lo(&op->src.u.dma); ccp_perform_rsa() 360 | ccp_addr_hi(&op->src.u.dma); ccp_perform_rsa() 377 if (op->src.type == CCP_MEMTYPE_SYSTEM) ccp_perform_passthru() 378 cr[1] = op->src.u.dma.length - 1; ccp_perform_passthru() 382 if (op->src.type == CCP_MEMTYPE_SYSTEM) { ccp_perform_passthru() 383 cr[2] = ccp_addr_lo(&op->src.u.dma); ccp_perform_passthru() 385 | ccp_addr_hi(&op->src.u.dma); ccp_perform_passthru() 390 cr[2] = op->src.u.ksb * CCP_KSB_BYTES; ccp_perform_passthru() 418 cr[1] = op->src.u.dma.length - 1; ccp_perform_ecc() 419 cr[2] = ccp_addr_lo(&op->src.u.dma); ccp_perform_ecc() 421 | ccp_addr_hi(&op->src.u.dma); ccp_perform_ecc() 746 static void ccp_prepare_data(struct ccp_data *src, struct ccp_data *dst, ccp_prepare_data() argument 757 sg_src_len = sg_dma_len(src->sg_wa.sg) - src->sg_wa.sg_used; ccp_prepare_data() 758 sg_src_len = min_t(u64, src->sg_wa.bytes_left, sg_src_len); ccp_prepare_data() 762 sg_dst_len = min_t(u64, src->sg_wa.bytes_left, sg_dst_len); ccp_prepare_data() 781 int cp_len = ccp_fill_queue_buf(src); ccp_prepare_data() 784 op->src.u.dma.address = src->dm_wa.dma.address; ccp_prepare_data() 785 op->src.u.dma.offset = 0; ccp_prepare_data() 786 op->src.u.dma.length = (blocksize_op) ? block_size : cp_len; ccp_prepare_data() 791 op->src.u.dma.address = sg_dma_address(src->sg_wa.sg); ccp_prepare_data() 792 op->src.u.dma.offset = src->sg_wa.sg_used; ccp_prepare_data() 793 op->src.u.dma.length = op_len & ~(block_size - 1); ccp_prepare_data() 795 ccp_update_sg_workarea(&src->sg_wa, op->src.u.dma.length); ccp_prepare_data() 807 op->dst.u.dma.length = op->src.u.dma.length; ccp_prepare_data() 814 op->dst.u.dma.length = op->src.u.dma.length; ccp_prepare_data() 819 static void ccp_process_data(struct ccp_data *src, struct ccp_data *dst, ccp_process_data() argument 847 op.src.type = CCP_MEMTYPE_KSB; ccp_copy_to_from_ksb() 848 op.src.u.ksb = ksb; ccp_copy_to_from_ksb() 853 op.src.type = CCP_MEMTYPE_SYSTEM; ccp_copy_to_from_ksb() 854 op.src.u.dma.address = wa->dma.address; ccp_copy_to_from_ksb() 855 op.src.u.dma.length = wa->length; ccp_copy_to_from_ksb() 884 struct ccp_data src; ccp_run_aes_cmac_cmd() local 900 if (!aes->key || !aes->iv || !aes->src) ccp_run_aes_cmac_cmd() 965 ret = ccp_init_data(&src, cmd_q, aes->src, aes->src_len, ccp_run_aes_cmac_cmd() 970 while (src.sg_wa.bytes_left) { ccp_run_aes_cmac_cmd() 971 ccp_prepare_data(&src, NULL, &op, AES_BLOCK_SIZE, true); ccp_run_aes_cmac_cmd() 972 if (aes->cmac_final && !src.sg_wa.bytes_left) { ccp_run_aes_cmac_cmd() 1000 ccp_process_data(&src, NULL, &op); ccp_run_aes_cmac_cmd() 1018 ccp_free_data(&src, cmd_q); ccp_run_aes_cmac_cmd() 1033 struct ccp_data src, dst; ccp_run_aes_cmd() local 1053 if (!aes->key || !aes->src || !aes->dst) ccp_run_aes_cmd() 1122 * and copy the src workarea to the dst workarea. ccp_run_aes_cmd() 1124 if (sg_virt(aes->src) == sg_virt(aes->dst)) ccp_run_aes_cmd() 1127 ret = ccp_init_data(&src, cmd_q, aes->src, aes->src_len, ccp_run_aes_cmd() 1134 dst = src; ccp_run_aes_cmd() 1143 while (src.sg_wa.bytes_left) { ccp_run_aes_cmd() 1144 ccp_prepare_data(&src, &dst, &op, AES_BLOCK_SIZE, true); ccp_run_aes_cmd() 1145 if (!src.sg_wa.bytes_left) { ccp_run_aes_cmd() 1162 ccp_process_data(&src, &dst, &op); ccp_run_aes_cmd() 1186 ccp_free_data(&src, cmd_q); ccp_run_aes_cmd() 1202 struct ccp_data src, dst; ccp_run_xts_aes_cmd() local 1238 if (!xts->key || !xts->iv || !xts->src || !xts->dst) ccp_run_xts_aes_cmd() 1295 * and copy the src workarea to the dst workarea. ccp_run_xts_aes_cmd() 1297 if (sg_virt(xts->src) == sg_virt(xts->dst)) ccp_run_xts_aes_cmd() 1300 ret = ccp_init_data(&src, cmd_q, xts->src, xts->src_len, ccp_run_xts_aes_cmd() 1307 dst = src; ccp_run_xts_aes_cmd() 1316 while (src.sg_wa.bytes_left) { ccp_run_xts_aes_cmd() 1317 ccp_prepare_data(&src, &dst, &op, unit_size, true); ccp_run_xts_aes_cmd() 1318 if (!src.sg_wa.bytes_left) ccp_run_xts_aes_cmd() 1327 ccp_process_data(&src, &dst, &op); ccp_run_xts_aes_cmd() 1349 ccp_free_data(&src, cmd_q); ccp_run_xts_aes_cmd() 1364 struct ccp_data src; ccp_run_sha_cmd() local 1413 if (!sha->src) ccp_run_sha_cmd() 1465 ret = ccp_init_data(&src, cmd_q, sha->src, sha->src_len, ccp_run_sha_cmd() 1470 while (src.sg_wa.bytes_left) { ccp_run_sha_cmd() 1471 ccp_prepare_data(&src, NULL, &op, CCP_SHA_BLOCKSIZE, false); ccp_run_sha_cmd() 1472 if (sha->final && !src.sg_wa.bytes_left) ccp_run_sha_cmd() 1481 ccp_process_data(&src, NULL, &op); ccp_run_sha_cmd() 1541 hmac_cmd.u.sha.src = &sg; ccp_run_sha_cmd() 1557 ccp_free_data(&src, cmd_q); ccp_run_sha_cmd() 1568 struct ccp_dm_workarea exp, src; ccp_run_rsa_cmd() local 1577 if (!rsa->exp || !rsa->mod || !rsa->src || !rsa->dst) ccp_run_rsa_cmd() 1621 ret = ccp_init_dm_workarea(&src, cmd_q, i_len, DMA_TO_DEVICE); ccp_run_rsa_cmd() 1625 ccp_reverse_set_dm_area(&src, rsa->mod, rsa->mod_len, CCP_KSB_BYTES, ccp_run_rsa_cmd() 1627 src.address += o_len; /* Adjust the address for the copy operation */ ccp_run_rsa_cmd() 1628 ccp_reverse_set_dm_area(&src, rsa->src, rsa->src_len, CCP_KSB_BYTES, ccp_run_rsa_cmd() 1630 src.address -= o_len; /* Reset the address to original value */ ccp_run_rsa_cmd() 1639 op.src.u.dma.address = src.dma.address; ccp_run_rsa_cmd() 1640 op.src.u.dma.offset = 0; ccp_run_rsa_cmd() 1641 op.src.u.dma.length = i_len; ccp_run_rsa_cmd() 1661 ccp_dm_free(&src); ccp_run_rsa_cmd() 1677 struct ccp_data src, dst; ccp_run_passthru_cmd() local 1686 if (!pt->src || !pt->dst) ccp_run_passthru_cmd() 1724 * and copy the src workarea to the dst workarea. ccp_run_passthru_cmd() 1726 if (sg_virt(pt->src) == sg_virt(pt->dst)) ccp_run_passthru_cmd() 1729 ret = ccp_init_data(&src, cmd_q, pt->src, pt->src_len, ccp_run_passthru_cmd() 1736 dst = src; ccp_run_passthru_cmd() 1752 for (i = 1; i <= src.sg_wa.dma_count; i++) { ccp_run_passthru_cmd() 1754 (dst.sg_wa.sg->length < src.sg_wa.sg->length)) { ccp_run_passthru_cmd() 1759 if (i == src.sg_wa.dma_count) { ccp_run_passthru_cmd() 1764 op.src.type = CCP_MEMTYPE_SYSTEM; ccp_run_passthru_cmd() 1765 op.src.u.dma.address = sg_dma_address(src.sg_wa.sg); ccp_run_passthru_cmd() 1766 op.src.u.dma.offset = 0; ccp_run_passthru_cmd() 1767 op.src.u.dma.length = sg_dma_len(src.sg_wa.sg); ccp_run_passthru_cmd() 1772 op.dst.u.dma.length = op.src.u.dma.length; ccp_run_passthru_cmd() 1780 dst.sg_wa.sg_used += src.sg_wa.sg->length; ccp_run_passthru_cmd() 1785 src.sg_wa.sg = sg_next(src.sg_wa.sg); ccp_run_passthru_cmd() 1793 ccp_free_data(&src, cmd_q); ccp_run_passthru_cmd() 1805 struct ccp_dm_workarea src, dst; ccp_run_ecc_mm_cmd() local 1832 ret = ccp_init_dm_workarea(&src, cmd_q, CCP_ECC_SRC_BUF_SIZE, ccp_run_ecc_mm_cmd() 1840 save = src.address; ccp_run_ecc_mm_cmd() 1843 ccp_reverse_set_dm_area(&src, ecc->mod, ecc->mod_len, ccp_run_ecc_mm_cmd() 1845 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_mm_cmd() 1848 ccp_reverse_set_dm_area(&src, ecc->u.mm.operand_1, ccp_run_ecc_mm_cmd() 1851 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_mm_cmd() 1855 ccp_reverse_set_dm_area(&src, ecc->u.mm.operand_2, ccp_run_ecc_mm_cmd() 1858 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_mm_cmd() 1862 src.address = save; ccp_run_ecc_mm_cmd() 1871 op.src.u.dma.address = src.dma.address; ccp_run_ecc_mm_cmd() 1872 op.src.u.dma.offset = 0; ccp_run_ecc_mm_cmd() 1873 op.src.u.dma.length = src.length; ccp_run_ecc_mm_cmd() 1900 ccp_dm_free(&src); ccp_run_ecc_mm_cmd() 1908 struct ccp_dm_workarea src, dst; ccp_run_ecc_pm_cmd() local 1951 ret = ccp_init_dm_workarea(&src, cmd_q, CCP_ECC_SRC_BUF_SIZE, ccp_run_ecc_pm_cmd() 1959 save = src.address; ccp_run_ecc_pm_cmd() 1962 ccp_reverse_set_dm_area(&src, ecc->mod, ecc->mod_len, ccp_run_ecc_pm_cmd() 1964 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 1967 ccp_reverse_set_dm_area(&src, ecc->u.pm.point_1.x, ccp_run_ecc_pm_cmd() 1970 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 1971 ccp_reverse_set_dm_area(&src, ecc->u.pm.point_1.y, ccp_run_ecc_pm_cmd() 1974 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 1977 *src.address = 0x01; ccp_run_ecc_pm_cmd() 1978 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 1982 ccp_reverse_set_dm_area(&src, ecc->u.pm.point_2.x, ccp_run_ecc_pm_cmd() 1985 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 1986 ccp_reverse_set_dm_area(&src, ecc->u.pm.point_2.y, ccp_run_ecc_pm_cmd() 1989 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 1992 *src.address = 0x01; ccp_run_ecc_pm_cmd() 1993 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 1996 ccp_reverse_set_dm_area(&src, ecc->u.pm.domain_a, ccp_run_ecc_pm_cmd() 1999 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 2003 ccp_reverse_set_dm_area(&src, ecc->u.pm.scalar, ccp_run_ecc_pm_cmd() 2006 src.address += CCP_ECC_OPERAND_SIZE; ccp_run_ecc_pm_cmd() 2011 src.address = save; ccp_run_ecc_pm_cmd() 2020 op.src.u.dma.address = src.dma.address; ccp_run_ecc_pm_cmd() 2021 op.src.u.dma.offset = 0; ccp_run_ecc_pm_cmd() 2022 op.src.u.dma.length = src.length; ccp_run_ecc_pm_cmd() 2062 ccp_dm_free(&src); ccp_run_ecc_pm_cmd()
|
/linux-4.1.27/drivers/char/hw_random/ |
H A D | xgene-rng.c | 63 #define MAJOR_HW_REV_RD(src) (((src) & 0x0f000000) >> 24) 64 #define MINOR_HW_REV_RD(src) (((src) & 0x00f00000) >> 20) 65 #define HW_PATCH_LEVEL_RD(src) (((src) & 0x000f0000) >> 16) 66 #define MAX_REFILL_CYCLES_SET(dst, src) \ 67 ((dst & ~0xffff0000) | (((u32)src << 16) & 0xffff0000)) 68 #define MIN_REFILL_CYCLES_SET(dst, src) \ 69 ((dst & ~0x000000ff) | (((u32)src) & 0x000000ff)) 70 #define ALARM_THRESHOLD_SET(dst, src) \ 71 ((dst & ~0x000000ff) | (((u32)src) & 0x000000ff)) 72 #define ENABLE_RNG_SET(dst, src) \ 73 ((dst & ~BIT(10)) | (((u32)src << 10) & BIT(10))) 74 #define REGSPEC_TEST_MODE_SET(dst, src) \ 75 ((dst & ~BIT(8)) | (((u32)src << 8) & BIT(8))) 76 #define MONOBIT_FAIL_MASK_SET(dst, src) \ 77 ((dst & ~BIT(7)) | (((u32)src << 7) & BIT(7))) 78 #define POKER_FAIL_MASK_SET(dst, src) \ 79 ((dst & ~BIT(6)) | (((u32)src << 6) & BIT(6))) 80 #define LONG_RUN_FAIL_MASK_SET(dst, src) \ 81 ((dst & ~BIT(5)) | (((u32)src << 5) & BIT(5))) 82 #define RUN_FAIL_MASK_SET(dst, src) \ 83 ((dst & ~BIT(4)) | (((u32)src << 4) & BIT(4))) 84 #define NOISE_FAIL_MASK_SET(dst, src) \ 85 ((dst & ~BIT(3)) | (((u32)src << 3) & BIT(3))) 86 #define STUCK_OUT_MASK_SET(dst, src) \ 87 ((dst & ~BIT(2)) | (((u32)src << 2) & BIT(2))) 88 #define SHUTDOWN_OFLO_MASK_SET(dst, src) \ 89 ((dst & ~BIT(1)) | (((u32)src << 1) & BIT(1)))
|
/linux-4.1.27/net/netfilter/ |
H A D | xt_HMARK.c | 35 __be32 src; member in struct:hmark_tuple 65 u16 src, dst; hmark_swap_ports() local 68 src = ntohs(hp.b16.src); hmark_swap_ports() 71 if (dst > src) hmark_swap_ports() 72 uports->v32 = (dst << 16) | src; hmark_swap_ports() 74 uports->v32 = (src << 16) | dst; hmark_swap_ports() 93 t->src = hmark_addr_mask(otuple->src.l3num, otuple->src.u3.ip6, hmark_ct_set_htuple() 95 t->dst = hmark_addr_mask(otuple->src.l3num, rtuple->src.u3.ip6, hmark_ct_set_htuple() 103 t->uports.b16.src = otuple->src.u.all; hmark_ct_set_htuple() 104 t->uports.b16.dst = rtuple->src.u.all; hmark_ct_set_htuple() 120 u32 src = ntohl(t->src); hmark_hash() local 123 if (dst < src) hmark_hash() 124 swap(src, dst); hmark_hash() 126 hash = jhash_3words(src, dst, t->uports.v32, info->hashrnd); hmark_hash() 194 t->src = hmark_addr6_mask(ip6->saddr.s6_addr32, info->src_mask.ip6); hmark_pkt_set_htuple_ipv6() 271 t->src = ip->saddr & info->src_mask.ip; hmark_pkt_set_htuple_ipv4()
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | pvclock.c | 34 unsigned long pvclock_tsc_khz(struct pvclock_vcpu_time_info *src) pvclock_tsc_khz() argument 38 do_div(pv_tsc_khz, src->tsc_to_system_mul); pvclock_tsc_khz() 39 if (src->tsc_shift < 0) pvclock_tsc_khz() 40 pv_tsc_khz <<= -src->tsc_shift; pvclock_tsc_khz() 42 pv_tsc_khz >>= src->tsc_shift; pvclock_tsc_khz() 61 u8 pvclock_read_flags(struct pvclock_vcpu_time_info *src) pvclock_read_flags() argument 68 version = __pvclock_read_cycles(src, &ret, &flags); pvclock_read_flags() 69 } while ((src->version & 1) || version != src->version); pvclock_read_flags() 74 cycle_t pvclock_clocksource_read(struct pvclock_vcpu_time_info *src) pvclock_clocksource_read() argument 82 version = __pvclock_read_cycles(src, &ret, &flags); pvclock_clocksource_read() 83 } while ((src->version & 1) || version != src->version); pvclock_clocksource_read() 86 src->flags &= ~PVCLOCK_GUEST_STOPPED; pvclock_clocksource_read()
|
/linux-4.1.27/arch/metag/include/asm/ |
H A D | checksum.h | 19 * the same as csum_partial, but copies from src while it 22 * here even more important to align src and dst on a 32-bit (or even 25 extern __wsum csum_partial_copy(const void *src, void *dst, int len, 31 * here even more important to align src and dst on a 32-bit (or even 34 extern __wsum csum_partial_copy_from_user(const void __user *src, void *dst, 37 #define csum_partial_copy_nocheck(src, dst, len, sum) \ 38 csum_partial_copy((src), (dst), (len), (sum))
|
/linux-4.1.27/drivers/block/xen-blkback/ |
H A D | common.h | 397 struct blkif_x86_32_request *src) blkif_get_x86_32_req() 400 dst->operation = src->operation; blkif_get_x86_32_req() 401 switch (src->operation) { blkif_get_x86_32_req() 406 dst->u.rw.nr_segments = src->u.rw.nr_segments; blkif_get_x86_32_req() 407 dst->u.rw.handle = src->u.rw.handle; blkif_get_x86_32_req() 408 dst->u.rw.id = src->u.rw.id; blkif_get_x86_32_req() 409 dst->u.rw.sector_number = src->u.rw.sector_number; blkif_get_x86_32_req() 414 dst->u.rw.seg[i] = src->u.rw.seg[i]; blkif_get_x86_32_req() 417 dst->u.discard.flag = src->u.discard.flag; blkif_get_x86_32_req() 418 dst->u.discard.id = src->u.discard.id; blkif_get_x86_32_req() 419 dst->u.discard.sector_number = src->u.discard.sector_number; blkif_get_x86_32_req() 420 dst->u.discard.nr_sectors = src->u.discard.nr_sectors; blkif_get_x86_32_req() 423 dst->u.indirect.indirect_op = src->u.indirect.indirect_op; blkif_get_x86_32_req() 424 dst->u.indirect.nr_segments = src->u.indirect.nr_segments; blkif_get_x86_32_req() 425 dst->u.indirect.handle = src->u.indirect.handle; blkif_get_x86_32_req() 426 dst->u.indirect.id = src->u.indirect.id; blkif_get_x86_32_req() 427 dst->u.indirect.sector_number = src->u.indirect.sector_number; blkif_get_x86_32_req() 432 src->u.indirect.indirect_grefs[i]; blkif_get_x86_32_req() 439 dst->u.other.id = src->u.other.id; blkif_get_x86_32_req() 445 struct blkif_x86_64_request *src) blkif_get_x86_64_req() 448 dst->operation = src->operation; blkif_get_x86_64_req() 449 switch (src->operation) { blkif_get_x86_64_req() 454 dst->u.rw.nr_segments = src->u.rw.nr_segments; blkif_get_x86_64_req() 455 dst->u.rw.handle = src->u.rw.handle; blkif_get_x86_64_req() 456 dst->u.rw.id = src->u.rw.id; blkif_get_x86_64_req() 457 dst->u.rw.sector_number = src->u.rw.sector_number; blkif_get_x86_64_req() 462 dst->u.rw.seg[i] = src->u.rw.seg[i]; blkif_get_x86_64_req() 465 dst->u.discard.flag = src->u.discard.flag; blkif_get_x86_64_req() 466 dst->u.discard.id = src->u.discard.id; blkif_get_x86_64_req() 467 dst->u.discard.sector_number = src->u.discard.sector_number; blkif_get_x86_64_req() 468 dst->u.discard.nr_sectors = src->u.discard.nr_sectors; blkif_get_x86_64_req() 471 dst->u.indirect.indirect_op = src->u.indirect.indirect_op; blkif_get_x86_64_req() 472 dst->u.indirect.nr_segments = src->u.indirect.nr_segments; blkif_get_x86_64_req() 473 dst->u.indirect.handle = src->u.indirect.handle; blkif_get_x86_64_req() 474 dst->u.indirect.id = src->u.indirect.id; blkif_get_x86_64_req() 475 dst->u.indirect.sector_number = src->u.indirect.sector_number; blkif_get_x86_64_req() 480 src->u.indirect.indirect_grefs[i]; blkif_get_x86_64_req() 487 dst->u.other.id = src->u.other.id; blkif_get_x86_64_req() 396 blkif_get_x86_32_req(struct blkif_request *dst, struct blkif_x86_32_request *src) blkif_get_x86_32_req() argument 444 blkif_get_x86_64_req(struct blkif_request *dst, struct blkif_x86_64_request *src) blkif_get_x86_64_req() argument
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | crash_dump.c | 90 static int copy_from_realmem(void *dest, void *src, size_t count) copy_from_realmem() argument 97 return memcpy_real(dest, src, count); copy_from_realmem() 100 if (memcpy_real(load_real_addr(dest), src, size)) copy_from_realmem() 104 src += size; copy_from_realmem() 121 unsigned long src, int userbuf) copy_oldmem_page_zfcpdump() 125 if (src < sclp_get_hsa_size()) { copy_oldmem_page_zfcpdump() 126 rc = memcpy_hsa(buf, src, csize, userbuf); copy_oldmem_page_zfcpdump() 130 (void *) src, csize); copy_oldmem_page_zfcpdump() 132 rc = memcpy_real(buf, (void *) src, csize); copy_oldmem_page_zfcpdump() 145 unsigned long src, int userbuf) copy_oldmem_page_kdump() 150 if (src < OLDMEM_SIZE) copy_oldmem_page_kdump() 151 src += OLDMEM_BASE; copy_oldmem_page_kdump() 152 else if (src > OLDMEM_BASE && copy_oldmem_page_kdump() 153 src < OLDMEM_BASE + OLDMEM_SIZE) copy_oldmem_page_kdump() 154 src -= OLDMEM_BASE; copy_oldmem_page_kdump() 157 (void *) src, csize); copy_oldmem_page_kdump() 159 rc = copy_from_realmem(buf, (void *) src, csize); copy_oldmem_page_kdump() 169 unsigned long src; copy_oldmem_page() local 173 src = (pfn << PAGE_SHIFT) + offset; copy_oldmem_page() 175 return copy_oldmem_page_kdump(buf, csize, src, userbuf); copy_oldmem_page() 177 return copy_oldmem_page_zfcpdump(buf, csize, src, userbuf); copy_oldmem_page() 248 int copy_from_oldmem(void *dest, void *src, size_t count) copy_from_oldmem() argument 254 if ((unsigned long) src < OLDMEM_SIZE) { copy_from_oldmem() 255 copied = min(count, OLDMEM_SIZE - (unsigned long) src); copy_from_oldmem() 256 rc = copy_from_realmem(dest, src + OLDMEM_BASE, copied); copy_from_oldmem() 262 if ((unsigned long) src < hsa_end) { copy_from_oldmem() 263 copied = min(count, hsa_end - (unsigned long) src); copy_from_oldmem() 264 rc = memcpy_hsa(dest, (unsigned long) src, copied, 0); copy_from_oldmem() 269 return copy_from_realmem(dest + copied, src + copied, count - copied); copy_from_oldmem() 662 void *src = (void *)(unsigned long)*ppos; elfcorehdr_read() local 664 src = elfcorehdr_newmem ? src : src - OLDMEM_BASE; elfcorehdr_read() 665 memcpy(buf, src, count); elfcorehdr_read() 675 void *src = (void *)(unsigned long)*ppos; elfcorehdr_read_notes() local 679 memcpy(buf, src, count); elfcorehdr_read_notes() 681 rc = copy_from_oldmem(buf, src, count); elfcorehdr_read_notes() 120 copy_oldmem_page_zfcpdump(char *buf, size_t csize, unsigned long src, int userbuf) copy_oldmem_page_zfcpdump() argument 144 copy_oldmem_page_kdump(char *buf, size_t csize, unsigned long src, int userbuf) copy_oldmem_page_kdump() argument
|
/linux-4.1.27/arch/s390/net/ |
H A D | bpf_jit_comp.c | 468 case BPF_ALU | BPF_MOV | BPF_X: /* dst = (u32) src */ bpf_jit_insn() 469 /* llgfr %dst,%src */ bpf_jit_insn() 472 case BPF_ALU64 | BPF_MOV | BPF_X: /* dst = src */ bpf_jit_insn() 473 /* lgr %dst,%src */ bpf_jit_insn() 502 case BPF_ALU | BPF_ADD | BPF_X: /* dst = (u32) dst + (u32) src */ bpf_jit_insn() 503 /* ar %dst,%src */ bpf_jit_insn() 507 case BPF_ALU64 | BPF_ADD | BPF_X: /* dst = dst + src */ bpf_jit_insn() 508 /* agr %dst,%src */ bpf_jit_insn() 527 case BPF_ALU | BPF_SUB | BPF_X: /* dst = (u32) dst - (u32) src */ bpf_jit_insn() 528 /* sr %dst,%src */ bpf_jit_insn() 532 case BPF_ALU64 | BPF_SUB | BPF_X: /* dst = dst - src */ bpf_jit_insn() 533 /* sgr %dst,%src */ bpf_jit_insn() 552 case BPF_ALU | BPF_MUL | BPF_X: /* dst = (u32) dst * (u32) src */ bpf_jit_insn() 553 /* msr %dst,%src */ bpf_jit_insn() 557 case BPF_ALU64 | BPF_MUL | BPF_X: /* dst = dst * src */ bpf_jit_insn() 558 /* msgr %dst,%src */ bpf_jit_insn() 577 case BPF_ALU | BPF_DIV | BPF_X: /* dst = (u32) dst / (u32) src */ bpf_jit_insn() 578 case BPF_ALU | BPF_MOD | BPF_X: /* dst = (u32) dst % (u32) src */ bpf_jit_insn() 583 /* ltr %src,%src (if src == 0 goto fail) */ bpf_jit_insn() 591 /* dlr %w0,%src */ bpf_jit_insn() 597 case BPF_ALU64 | BPF_DIV | BPF_X: /* dst = dst / src */ bpf_jit_insn() 598 case BPF_ALU64 | BPF_MOD | BPF_X: /* dst = dst % src */ bpf_jit_insn() 603 /* ltgr %src,%src (if src == 0 goto fail) */ bpf_jit_insn() 664 case BPF_ALU | BPF_AND | BPF_X: /* dst = (u32) dst & (u32) src */ bpf_jit_insn() 665 /* nr %dst,%src */ bpf_jit_insn() 669 case BPF_ALU64 | BPF_AND | BPF_X: /* dst = dst & src */ bpf_jit_insn() 670 /* ngr %dst,%src */ bpf_jit_insn() 686 case BPF_ALU | BPF_OR | BPF_X: /* dst = (u32) dst | (u32) src */ bpf_jit_insn() 687 /* or %dst,%src */ bpf_jit_insn() 691 case BPF_ALU64 | BPF_OR | BPF_X: /* dst = dst | src */ bpf_jit_insn() 692 /* ogr %dst,%src */ bpf_jit_insn() 708 case BPF_ALU | BPF_XOR | BPF_X: /* dst = (u32) dst ^ (u32) src */ bpf_jit_insn() 709 /* xr %dst,%src */ bpf_jit_insn() 713 case BPF_ALU64 | BPF_XOR | BPF_X: /* dst = dst ^ src */ bpf_jit_insn() 714 /* xgr %dst,%src */ bpf_jit_insn() 732 case BPF_ALU | BPF_LSH | BPF_X: /* dst = (u32) dst << (u32) src */ bpf_jit_insn() 733 /* sll %dst,0(%src) */ bpf_jit_insn() 737 case BPF_ALU64 | BPF_LSH | BPF_X: /* dst = dst << src */ bpf_jit_insn() 738 /* sllg %dst,%dst,0(%src) */ bpf_jit_insn() 757 case BPF_ALU | BPF_RSH | BPF_X: /* dst = (u32) dst >> (u32) src */ bpf_jit_insn() 758 /* srl %dst,0(%src) */ bpf_jit_insn() 762 case BPF_ALU64 | BPF_RSH | BPF_X: /* dst = dst >> src */ bpf_jit_insn() 763 /* srlg %dst,%dst,0(%src) */ bpf_jit_insn() 782 case BPF_ALU64 | BPF_ARSH | BPF_X: /* ((s64) dst) >>= src */ bpf_jit_insn() 783 /* srag %dst,%dst,0(%src) */ bpf_jit_insn() 848 /* stcy %src,off(%dst) */ bpf_jit_insn() 852 case BPF_STX | BPF_MEM | BPF_H: /* (u16 *)(dst + off) = src */ bpf_jit_insn() 853 /* sthy %src,off(%dst) */ bpf_jit_insn() 857 case BPF_STX | BPF_MEM | BPF_W: /* *(u32 *)(dst + off) = src */ bpf_jit_insn() 858 /* sty %src,off(%dst) */ bpf_jit_insn() 862 case BPF_STX | BPF_MEM | BPF_DW: /* (u64 *)(dst + off) = src */ bpf_jit_insn() 863 /* stg %src,off(%dst) */ bpf_jit_insn() 898 case BPF_STX | BPF_XADD | BPF_W: /* *(u32 *)(dst + off) += src */ bpf_jit_insn() 899 /* laal %w0,%src,off(%dst) */ bpf_jit_insn() 904 case BPF_STX | BPF_XADD | BPF_DW: /* *(u64 *)(dst + off) += src */ bpf_jit_insn() 905 /* laalg %w0,%src,off(%dst) */ bpf_jit_insn() 913 case BPF_LDX | BPF_MEM | BPF_B: /* dst = *(u8 *)(ul) (src + off) */ bpf_jit_insn() 914 /* llgc %dst,0(off,%src) */ bpf_jit_insn() 918 case BPF_LDX | BPF_MEM | BPF_H: /* dst = *(u16 *)(ul) (src + off) */ bpf_jit_insn() 919 /* llgh %dst,0(off,%src) */ bpf_jit_insn() 923 case BPF_LDX | BPF_MEM | BPF_W: /* dst = *(u32 *)(ul) (src + off) */ bpf_jit_insn() 924 /* llgf %dst,off(%src) */ bpf_jit_insn() 928 case BPF_LDX | BPF_MEM | BPF_DW: /* dst = *(u64 *)(ul) (src + off) */ bpf_jit_insn() 929 /* lg %dst,0(off,%src) */ bpf_jit_insn() 1010 case BPF_JMP | BPF_JSGT | BPF_X: /* ((s64) dst > (s64) src) */ bpf_jit_insn() 1013 case BPF_JMP | BPF_JSGE | BPF_X: /* ((s64) dst >= (s64) src) */ bpf_jit_insn() 1016 case BPF_JMP | BPF_JGT | BPF_X: /* (dst > src) */ bpf_jit_insn() 1019 case BPF_JMP | BPF_JGE | BPF_X: /* (dst >= src) */ bpf_jit_insn() 1022 case BPF_JMP | BPF_JNE | BPF_X: /* (dst != src) */ bpf_jit_insn() 1025 case BPF_JMP | BPF_JEQ | BPF_X: /* (dst == src) */ bpf_jit_insn() 1028 case BPF_JMP | BPF_JSET | BPF_X: /* (dst & src) */ bpf_jit_insn() 1030 /* ngrk %w1,%dst,%src */ bpf_jit_insn() 1046 /* cgrj %dst,%src,mask,off */ bpf_jit_insn() 1050 /* clgrj %dst,%src,mask,off */ bpf_jit_insn() 1062 case BPF_LD | BPF_IND | BPF_B: /* b0 = *(u8 *) (skb->data+imm+src) */ bpf_jit_insn() 1069 case BPF_LD | BPF_IND | BPF_H: /* b0 = *(u16 *) (skb->data+imm+src) */ bpf_jit_insn() 1076 case BPF_LD | BPF_IND | BPF_W: /* b0 = *(u32 *) (skb->data+imm+src) */ bpf_jit_insn() 1107 /* agfr %b2,%src (%src is s32 here) */ bpf_jit_insn()
|
/linux-4.1.27/arch/powerpc/kvm/ |
H A D | mpic.c | 222 struct irq_source src[MAX_IRQ]; member in struct:openpic 304 irq, IVPR_PRIORITY(opp->src[irq].ivpr), priority); IRQ_check() 306 if (IVPR_PRIORITY(opp->src[irq].ivpr) > priority) { IRQ_check() 308 priority = IVPR_PRIORITY(opp->src[irq].ivpr); IRQ_check() 328 struct irq_source *src; IRQ_local_pipe() local 332 src = &opp->src[n_IRQ]; IRQ_local_pipe() 337 if (src->output != ILR_INTTGT_INT) { IRQ_local_pipe() 339 __func__, src->output, n_IRQ, active, was_active, IRQ_local_pipe() 340 dst->outputs_active[src->output]); IRQ_local_pipe() 348 dst->outputs_active[src->output]++ == 0) { IRQ_local_pipe() 350 __func__, src->output, n_CPU, n_IRQ); IRQ_local_pipe() 351 mpic_irq_raise(opp, dst, src->output); IRQ_local_pipe() 355 --dst->outputs_active[src->output] == 0) { IRQ_local_pipe() 357 __func__, src->output, n_CPU, n_IRQ); IRQ_local_pipe() 358 mpic_irq_lower(opp, dst, src->output); IRQ_local_pipe() 365 priority = IVPR_PRIORITY(src->ivpr); IRQ_local_pipe() 414 struct irq_source *src; openpic_update_irq() local 418 src = &opp->src[n_IRQ]; openpic_update_irq() 419 active = src->pending; openpic_update_irq() 421 if ((src->ivpr & IVPR_MASK_MASK) && !src->nomask) { openpic_update_irq() 427 was_active = !!(src->ivpr & IVPR_ACTIVITY_MASK); openpic_update_irq() 439 src->ivpr |= IVPR_ACTIVITY_MASK; openpic_update_irq() 441 src->ivpr &= ~IVPR_ACTIVITY_MASK; openpic_update_irq() 443 if (src->destmask == 0) { openpic_update_irq() 449 if (src->destmask == (1 << src->last_cpu)) { openpic_update_irq() 451 IRQ_local_pipe(opp, src->last_cpu, n_IRQ, active, was_active); openpic_update_irq() 452 } else if (!(src->ivpr & IVPR_MODE_MASK)) { openpic_update_irq() 455 if (src->destmask & (1 << i)) { openpic_update_irq() 462 for (i = src->last_cpu + 1; i != src->last_cpu; i++) { openpic_update_irq() 466 if (src->destmask & (1 << i)) { openpic_update_irq() 469 src->last_cpu = i; openpic_update_irq() 479 struct irq_source *src; openpic_set_irq() local 486 src = &opp->src[n_IRQ]; openpic_set_irq() 488 n_IRQ, level, src->ivpr); openpic_set_irq() 489 if (src->level) { openpic_set_irq() 491 src->pending = level; openpic_set_irq() 496 src->pending = 1; openpic_set_irq() 500 if (src->output != ILR_INTTGT_INT) { openpic_set_irq() 507 src->pending = 0; openpic_set_irq() 527 opp->src[i].ivpr = opp->ivpr_reset; openpic_reset() 529 switch (opp->src[i].type) { openpic_reset() 531 opp->src[i].level = openpic_reset() 536 opp->src[i].ivpr |= IVPR_POLARITY_MASK; openpic_reset() 564 return opp->src[n_IRQ].idr; read_IRQreg_idr() 570 return opp->src[n_IRQ].output; read_IRQreg_ilr() 577 return opp->src[n_IRQ].ivpr; read_IRQreg_ivpr() 583 struct irq_source *src = &opp->src[n_IRQ]; write_IRQreg_idr() local 595 src->idr = val & mask; write_IRQreg_idr() 596 pr_debug("Set IDR %d to 0x%08x\n", n_IRQ, src->idr); write_IRQreg_idr() 599 if (src->idr & crit_mask) { write_IRQreg_idr() 600 if (src->idr & normal_mask) { write_IRQreg_idr() 605 src->output = ILR_INTTGT_CINT; write_IRQreg_idr() 606 src->nomask = true; write_IRQreg_idr() 607 src->destmask = 0; write_IRQreg_idr() 612 if (src->idr & (1UL << n_ci)) write_IRQreg_idr() 613 src->destmask |= 1UL << i; write_IRQreg_idr() 616 src->output = ILR_INTTGT_INT; write_IRQreg_idr() 617 src->nomask = false; write_IRQreg_idr() 618 src->destmask = src->idr & normal_mask; write_IRQreg_idr() 621 src->destmask = src->idr; write_IRQreg_idr() 629 struct irq_source *src = &opp->src[n_IRQ]; write_IRQreg_ilr() local 631 src->output = val & ILR_INTTGT_MASK; write_IRQreg_ilr() 632 pr_debug("Set ILR %d to 0x%08x, output %d\n", n_IRQ, src->idr, write_IRQreg_ilr() 633 src->output); write_IRQreg_ilr() 651 opp->src[n_IRQ].ivpr = write_IRQreg_ivpr() 652 (opp->src[n_IRQ].ivpr & IVPR_ACTIVITY_MASK) | (val & mask); write_IRQreg_ivpr() 658 switch (opp->src[n_IRQ].type) { write_IRQreg_ivpr() 660 opp->src[n_IRQ].level = write_IRQreg_ivpr() 661 !!(opp->src[n_IRQ].ivpr & IVPR_SENSE_MASK); write_IRQreg_ivpr() 665 opp->src[n_IRQ].ivpr &= ~IVPR_SENSE_MASK; write_IRQreg_ivpr() 669 opp->src[n_IRQ].ivpr &= ~(IVPR_POLARITY_MASK | IVPR_SENSE_MASK); write_IRQreg_ivpr() 675 opp->src[n_IRQ].ivpr); write_IRQreg_ivpr() 1028 struct irq_source *src; openpic_cpu_write_internal() local 1050 opp->src[opp->irq_ipi0 + idx].destmask |= val; openpic_cpu_write_internal() 1097 src = &opp->src[n_IRQ]; openpic_cpu_write_internal() 1100 IVPR_PRIORITY(src->ivpr) > dst->servicing.priority)) { openpic_cpu_write_internal() 1130 struct irq_source *src; openpic_iack() local 1143 src = &opp->src[irq]; openpic_iack() 1144 if (!(src->ivpr & IVPR_ACTIVITY_MASK) || openpic_iack() 1145 !(IVPR_PRIORITY(src->ivpr) > dst->ctpr)) { openpic_iack() 1147 __func__, irq, dst->ctpr, src->ivpr); openpic_iack() 1153 retval = IVPR_VECTOR(opp, src->ivpr); openpic_iack() 1156 if (!src->level) { openpic_iack() 1158 src->ivpr &= ~IVPR_ACTIVITY_MASK; openpic_iack() 1159 src->pending = 0; openpic_iack() 1164 src->destmask &= ~(1 << cpu); openpic_iack() 1165 if (src->destmask && !src->level) { openpic_iack() 1170 src->ivpr |= IVPR_ACTIVITY_MASK; openpic_iack() 1325 opp->src[i].level = false; fsl_common_init() 1329 opp->src[i].type = IRQ_TYPE_FSLINT; fsl_common_init() 1330 opp->src[i].level = true; fsl_common_init() 1335 opp->src[i].type = IRQ_TYPE_FSLSPECIAL; fsl_common_init() 1336 opp->src[i].level = false; fsl_common_init() 1596 attr32 = opp->src[attr->attr].pending; mpic_get_attr()
|
/linux-4.1.27/ipc/ |
H A D | msgutil.c | 86 struct msg_msg *load_msg(const void __user *src, size_t len) load_msg() argument 98 if (copy_from_user(msg + 1, src, alen)) load_msg() 103 src = (char __user *)src + alen; load_msg() 105 if (copy_from_user(seg + 1, src, alen)) load_msg() 120 struct msg_msg *copy_msg(struct msg_msg *src, struct msg_msg *dst) copy_msg() argument 123 size_t len = src->m_ts; copy_msg() 127 if (src->m_ts > dst->m_ts) copy_msg() 131 memcpy(dst + 1, src + 1, alen); copy_msg() 133 for (dst_pseg = dst->next, src_pseg = src->next; copy_msg() 142 dst->m_type = src->m_type; copy_msg() 143 dst->m_ts = src->m_ts; copy_msg() 148 struct msg_msg *copy_msg(struct msg_msg *src, struct msg_msg *dst) copy_msg() argument
|
/linux-4.1.27/arch/s390/boot/compressed/ |
H A D | misc.c | 83 void *memcpy(void *dest, const void *src, size_t n) memcpy() argument 85 const char *s = src; memcpy() 93 void *memmove(void *dest, const void *src, size_t n) memmove() argument 95 const char *s = src; memmove() 132 void *src, *dst; check_ipl_parmblock() local 134 src = (void *)(unsigned long) S390_lowcore.ipl_parmblock_ptr; check_ipl_parmblock() 135 if (src + PAGE_SIZE <= start || src >= start + size) check_ipl_parmblock() 138 memmove(dst, src, PAGE_SIZE); check_ipl_parmblock()
|
/linux-4.1.27/arch/metag/mm/ |
H A D | maccess.c | 18 long probe_kernel_write(void *dst, const void *src, size_t size) probe_kernel_write() argument 22 unsigned long lsrc = (unsigned long)src; probe_kernel_write() 23 const u8 *psrc = (u8 *)src; probe_kernel_write() 43 return __probe_kernel_write(dst, src, size); probe_kernel_write() 45 /* If src is unaligned, copy to the aligned bounce buffer first. */ probe_kernel_write()
|
/linux-4.1.27/drivers/video/logo/ |
H A D | Makefile | 25 $(wildcard $(srctree)/$(src)/*$(1).$(2)))) 48 $(obj)/%_mono.c: $(src)/%_mono.pbm $(pnmtologo) FORCE 51 $(obj)/%_vga16.c: $(src)/%_vga16.ppm $(pnmtologo) FORCE 54 $(obj)/%_clut224.c: $(src)/%_clut224.ppm $(pnmtologo) FORCE 57 $(obj)/%_gray256.c: $(src)/%_gray256.pgm $(pnmtologo) FORCE
|
/linux-4.1.27/include/net/ |
H A D | flow_keys.h | 5 * @src: source ip address in case of IPv4 6 * For IPv6 it contains 32bit hash of src address 10 * port16[0]: src port number 18 /* (src,dst) must be grouped, in the same way than in IP header */ 19 __be32 src; member in struct:flow_keys
|
/linux-4.1.27/arch/m32r/lib/ |
H A D | csum_partial_copy.c | 31 csum_partial_copy_nocheck (const void *src, void *dst, int len, __wsum sum) csum_partial_copy_nocheck() argument 33 sum = csum_partial(src, len, sum); csum_partial_copy_nocheck() 34 memcpy(dst, src, len); csum_partial_copy_nocheck() 45 csum_partial_copy_from_user (const void __user *src, void *dst, csum_partial_copy_from_user() argument 50 missing = copy_from_user(dst, src, len); csum_partial_copy_from_user()
|
/linux-4.1.27/drivers/staging/rtl8192e/rtl8192e/ |
H A D | rtl_crypto.h | 77 void (*cia_encrypt)(void *ctx, u8 *dst, const u8 *src); 78 void (*cia_decrypt)(void *ctx, u8 *dst, const u8 *src); 93 int (*coa_compress)(void *ctx, const u8 *src, unsigned int slen, 95 int (*coa_decompress)(void *ctx, const u8 *src, unsigned int slen, 145 struct scatterlist *src, 149 struct scatterlist *src, 153 struct scatterlist *src, 157 struct scatterlist *src, 159 void (*cit_xor_block)(u8 *dst, const u8 *src); 175 const u8 *src, unsigned int slen, 178 const u8 *src, unsigned int slen, 316 struct scatterlist *src, crypto_cipher_encrypt() 320 return tfm->crt_cipher.cit_encrypt(tfm, dst, src, nbytes); crypto_cipher_encrypt() 325 struct scatterlist *src, crypto_cipher_encrypt_iv() 330 return tfm->crt_cipher.cit_encrypt_iv(tfm, dst, src, nbytes, iv); crypto_cipher_encrypt_iv() 335 struct scatterlist *src, crypto_cipher_decrypt() 339 return tfm->crt_cipher.cit_decrypt(tfm, dst, src, nbytes); crypto_cipher_decrypt() 344 struct scatterlist *src, crypto_cipher_decrypt_iv() 349 return tfm->crt_cipher.cit_decrypt_iv(tfm, dst, src, nbytes, iv); crypto_cipher_decrypt_iv() 353 const u8 *src, unsigned int len) crypto_cipher_set_iv() 356 memcpy(tfm->crt_cipher.cit_iv, src, len); crypto_cipher_set_iv() 367 const u8 *src, unsigned int slen, crypto_comp_compress() 371 return tfm->crt_compress.cot_compress(tfm, src, slen, dst, dlen); crypto_comp_compress() 375 const u8 *src, unsigned int slen, crypto_comp_decompress() 379 return tfm->crt_compress.cot_decompress(tfm, src, slen, dst, dlen); crypto_comp_decompress() 314 crypto_cipher_encrypt(struct crypto_tfm *tfm, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) crypto_cipher_encrypt() argument 323 crypto_cipher_encrypt_iv(struct crypto_tfm *tfm, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes, u8 *iv) crypto_cipher_encrypt_iv() argument 333 crypto_cipher_decrypt(struct crypto_tfm *tfm, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) crypto_cipher_decrypt() argument 342 crypto_cipher_decrypt_iv(struct crypto_tfm *tfm, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes, u8 *iv) crypto_cipher_decrypt_iv() argument 352 crypto_cipher_set_iv(struct crypto_tfm *tfm, const u8 *src, unsigned int len) crypto_cipher_set_iv() argument 366 crypto_comp_compress(struct crypto_tfm *tfm, const u8 *src, unsigned int slen, u8 *dst, unsigned int *dlen) crypto_comp_compress() argument 374 crypto_comp_decompress(struct crypto_tfm *tfm, const u8 *src, unsigned int slen, u8 *dst, unsigned int *dlen) crypto_comp_decompress() argument
|
/linux-4.1.27/arch/sparc/crypto/ |
H A D | des_glue.c | 70 static void des_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) des_encrypt() argument 75 des_sparc64_crypt(K, (const u64 *) src, (u64 *) dst); des_encrypt() 78 static void des_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) des_decrypt() argument 83 des_sparc64_crypt(K, (const u64 *) src, (u64 *) dst); des_decrypt() 94 struct scatterlist *dst, struct scatterlist *src, __ecb_crypt() 101 blkcipher_walk_init(&walk, dst, src, nbytes); __ecb_crypt() 113 des_sparc64_ecb_crypt((const u64 *)walk.src.virt.addr, __ecb_crypt() 125 struct scatterlist *dst, struct scatterlist *src, ecb_encrypt() 128 return __ecb_crypt(desc, dst, src, nbytes, true); ecb_encrypt() 132 struct scatterlist *dst, struct scatterlist *src, ecb_decrypt() 135 return __ecb_crypt(desc, dst, src, nbytes, false); ecb_decrypt() 142 struct scatterlist *dst, struct scatterlist *src, cbc_encrypt() 149 blkcipher_walk_init(&walk, dst, src, nbytes); cbc_encrypt() 158 des_sparc64_cbc_encrypt((const u64 *)walk.src.virt.addr, cbc_encrypt() 173 struct scatterlist *dst, struct scatterlist *src, cbc_decrypt() 180 blkcipher_walk_init(&walk, dst, src, nbytes); cbc_decrypt() 189 des_sparc64_cbc_decrypt((const u64 *)walk.src.virt.addr, cbc_decrypt() 240 static void des3_ede_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) des3_ede_encrypt() argument 245 des3_ede_sparc64_crypt(K, (const u64 *) src, (u64 *) dst); des3_ede_encrypt() 248 static void des3_ede_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) des3_ede_decrypt() argument 253 des3_ede_sparc64_crypt(K, (const u64 *) src, (u64 *) dst); des3_ede_decrypt() 262 struct scatterlist *dst, struct scatterlist *src, __ecb3_crypt() 270 blkcipher_walk_init(&walk, dst, src, nbytes); __ecb3_crypt() 283 const u64 *src64 = (const u64 *)walk.src.virt.addr; __ecb3_crypt() 296 struct scatterlist *dst, struct scatterlist *src, ecb3_encrypt() 299 return __ecb3_crypt(desc, dst, src, nbytes, true); ecb3_encrypt() 303 struct scatterlist *dst, struct scatterlist *src, ecb3_decrypt() 306 return __ecb3_crypt(desc, dst, src, nbytes, false); ecb3_decrypt() 314 struct scatterlist *dst, struct scatterlist *src, cbc3_encrypt() 322 blkcipher_walk_init(&walk, dst, src, nbytes); cbc3_encrypt() 332 const u64 *src64 = (const u64 *)walk.src.virt.addr; cbc3_encrypt() 350 struct scatterlist *dst, struct scatterlist *src, cbc3_decrypt() 358 blkcipher_walk_init(&walk, dst, src, nbytes); cbc3_decrypt() 368 const u64 *src64 = (const u64 *)walk.src.virt.addr; cbc3_decrypt() 93 __ecb_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes, bool encrypt) __ecb_crypt() argument 124 ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_encrypt() argument 131 ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb_decrypt() argument 141 cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_encrypt() argument 172 cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc_decrypt() argument 261 __ecb3_crypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes, bool encrypt) __ecb3_crypt() argument 295 ecb3_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb3_encrypt() argument 302 ecb3_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) ecb3_decrypt() argument 313 cbc3_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc3_encrypt() argument 349 cbc3_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst, struct scatterlist *src, unsigned int nbytes) cbc3_decrypt() argument
|
/linux-4.1.27/arch/x86/um/asm/ |
H A D | checksum_32.h | 42 static __inline__ __wsum csum_and_copy_to_user(const void *src, csum_and_copy_to_user() argument 47 if (copy_to_user(dst, src, len)) { csum_and_copy_to_user() 52 return csum_partial(src, len, sum); csum_and_copy_to_user()
|
/linux-4.1.27/arch/blackfin/include/asm/ |
H A D | string.h | 15 extern char *strcpy(char *dest, const char *src); 18 extern char *strncpy(char *dest, const char *src, size_t n); 35 extern void *memmove(void *dest, const void *src, size_t count);
|
/linux-4.1.27/drivers/video/fbdev/sis/ |
H A D | sis_accel.c | 44 0x88, /* dest &= src; DSa, GXand, 0x1 */ 45 0x44, /* dest = src & ~dest; SDna, GXandReverse, 0x2 */ 46 0xCC, /* dest = src; S, GXcopy, 0x3 */ 47 0x22, /* dest &= ~src; DSna, GXandInverted, 0x4 */ 49 0x66, /* dest = ^src; DSx, GXxor, 0x6 */ 50 0xEE, /* dest |= src; DSo, GXor, 0x7 */ 51 0x11, /* dest = ~src & ~dest; DSon, GXnor, 0x8 */ 52 0x99, /* dest ^= ~src ; DSxn, GXequiv, 0x9 */ 54 0xDD, /* dest = src|~dest ; SDno, GXorReverse, 0xB */ 55 0x33, /* dest = ~src; Sn, GXcopyInverted, 0xC */ 56 0xBB, /* dest |= ~src; DSno, GXorInverted, 0xD */ 57 0x77, /* dest = ~src|~dest; DSan, GXnand, 0xE */ 64 0xA0, /* dest &= src; DPa, GXand, 0x1 */ 65 0x50, /* dest = src & ~dest; PDna, GXandReverse, 0x2 */ 66 0xF0, /* dest = src; P, GXcopy, 0x3 */ 67 0x0A, /* dest &= ~src; DPna, GXandInverted, 0x4 */ 69 0x5A, /* dest = ^src; DPx, GXxor, 0x6 */ 70 0xFA, /* dest |= src; DPo, GXor, 0x7 */ 71 0x05, /* dest = ~src & ~dest; DPon, GXnor, 0x8 */ 72 0xA5, /* dest ^= ~src ; DPxn, GXequiv, 0x9 */ 74 0xF5, /* dest = src|~dest ; PDno, GXorReverse, 0xB */ 75 0x0F, /* dest = ~src; Pn, GXcopyInverted, 0xC */ 76 0xAF, /* dest |= ~src; DPno, GXorInverted, 0xD */ 77 0x5F, /* dest = ~src|~dest; DPan, GXnand, 0xE */
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | iomap.c | 92 void iowrite8_rep(void __iomem *addr, const void *src, unsigned long count) iowrite8_rep() argument 94 writesb(addr, src, count); iowrite8_rep() 96 void iowrite16_rep(void __iomem *addr, const void *src, unsigned long count) iowrite16_rep() argument 98 writesw(addr, src, count); iowrite16_rep() 100 void iowrite32_rep(void __iomem *addr, const void *src, unsigned long count) iowrite32_rep() argument 102 writesl(addr, src, count); iowrite32_rep()
|
/linux-4.1.27/drivers/tty/vt/ |
H A D | Makefile | 17 $(obj)/consolemap_deftbl.c: $(src)/$(FONTMAPFILE) 29 $(obj)/defkeymap.c: $(obj)/%.c: $(src)/%.map
|