Searched refs:ea (Results 1 - 170 of 170) sorted by relevance

/linux-4.1.27/arch/powerpc/math-emu/
H A Dstfd.c6 stfd(void *frS, void *ea) stfd() argument
10 printk("%s: S %p, ea %p: ", __func__, frS, ea); stfd()
16 if (copy_to_user(ea, frS, sizeof(double))) stfd()
H A Dstfiwx.c6 stfiwx(u32 *frS, void *ea) stfiwx() argument
9 printk("%s: %p %p\n", __func__, frS, ea); stfiwx()
12 if (copy_to_user(ea, &frS[1], sizeof(frS[1]))) stfiwx()
H A Dlfd.c9 lfd(void *frD, void *ea) lfd() argument
11 if (copy_from_user(frD, ea, sizeof(double))) lfd()
14 printk("%s: D %p, ea %p: ", __func__, frD, ea); lfd()
H A Dlfs.c11 lfs(void *frD, void *ea) lfs() argument
19 printk("%s: D %p, ea %p\n", __func__, frD, ea); lfs()
22 if (copy_from_user(&f, ea, sizeof(float))) lfs()
H A Dstfs.c11 stfs(void *frS, void *ea) stfs() argument
19 printk("%s: S %p, ea %p\n", __func__, frS, ea); stfs()
37 if (copy_to_user(ea, &f, sizeof(float))) stfs()
/linux-4.1.27/fs/gfs2/
H A Dxattr.h16 #define GFS2_EA_REC_LEN(ea) be32_to_cpu((ea)->ea_rec_len)
17 #define GFS2_EA_DATA_LEN(ea) be32_to_cpu((ea)->ea_data_len)
19 #define GFS2_EA_SIZE(ea) \
20 ALIGN(sizeof(struct gfs2_ea_header) + (ea)->ea_name_len + \
21 ((GFS2_EA_IS_STUFFED(ea)) ? GFS2_EA_DATA_LEN(ea) : \
22 (sizeof(__be64) * (ea)->ea_num_ptrs)), 8)
24 #define GFS2_EA_IS_STUFFED(ea) (!(ea)->ea_num_ptrs)
25 #define GFS2_EA_IS_LAST(ea) ((ea)->ea_flags & GFS2_EAFLAG_LAST)
30 #define GFS2_EA2NAME(ea) ((char *)((struct gfs2_ea_header *)(ea) + 1))
31 #define GFS2_EA2DATA(ea) (GFS2_EA2NAME(ea) + (ea)->ea_name_len)
33 #define GFS2_EA2DATAPTRS(ea) \
34 ((__be64 *)(GFS2_EA2NAME(ea) + ALIGN((ea)->ea_name_len, 8)))
36 #define GFS2_EA2NEXT(ea) \
37 ((struct gfs2_ea_header *)((char *)(ea) + GFS2_EA_REC_LEN(ea)))
H A Dxattr.c76 struct gfs2_ea_header *ea,
82 struct gfs2_ea_header *ea, *prev = NULL; ea_foreach_i() local
88 for (ea = GFS2_EA_BH2FIRST(bh);; prev = ea, ea = GFS2_EA2NEXT(ea)) { ea_foreach_i()
89 if (!GFS2_EA_REC_LEN(ea)) ea_foreach_i()
91 if (!(bh->b_data <= (char *)ea && (char *)GFS2_EA2NEXT(ea) <= ea_foreach_i()
94 if (!GFS2_EATYPE_VALID(ea->ea_type)) ea_foreach_i()
97 error = ea_call(ip, bh, ea, prev, data); ea_foreach_i()
101 if (GFS2_EA_IS_LAST(ea)) { ea_foreach_i()
102 if ((char *)GFS2_EA2NEXT(ea) != ea_foreach_i()
167 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_find_i()
172 if (ea->ea_type == GFS2_EATYPE_UNUSED) ea_find_i()
175 if (ea->ea_type == ef->type) { ea_find_i()
176 if (ea->ea_name_len == ef->namel && ea_find_i()
177 !memcmp(GFS2_EA2NAME(ea), ef->name, ea->ea_name_len)) { ea_find_i()
181 el->el_ea = ea; ea_find_i()
214 * @ea:
226 struct gfs2_ea_header *ea, ea_dealloc_unstuffed()
246 if (GFS2_EA_IS_STUFFED(ea)) ea_dealloc_unstuffed()
249 dataptrs = GFS2_EA2DATAPTRS(ea); ea_dealloc_unstuffed()
250 for (x = 0; x < ea->ea_num_ptrs; x++, dataptrs++) { ea_dealloc_unstuffed()
276 dataptrs = GFS2_EA2DATAPTRS(ea); ea_dealloc_unstuffed()
277 for (x = 0; x < ea->ea_num_ptrs; x++, dataptrs++) { ea_dealloc_unstuffed()
300 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_dealloc_unstuffed()
303 if (GFS2_EA_IS_LAST(ea)) ea_dealloc_unstuffed()
306 ea->ea_type = GFS2_EATYPE_UNUSED; ea_dealloc_unstuffed()
307 ea->ea_num_ptrs = 0; ea_dealloc_unstuffed()
326 struct gfs2_ea_header *ea, ea_remove_unstuffed()
339 error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL); ea_remove_unstuffed()
351 static inline unsigned int gfs2_ea_strlen(struct gfs2_ea_header *ea) gfs2_ea_strlen() argument
353 switch (ea->ea_type) { gfs2_ea_strlen()
355 return 5 + ea->ea_name_len + 1; gfs2_ea_strlen()
357 return 7 + ea->ea_name_len + 1; gfs2_ea_strlen()
359 return 9 + ea->ea_name_len + 1; gfs2_ea_strlen()
366 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_list_i()
371 unsigned int ea_size = gfs2_ea_strlen(ea); ea_list_i()
373 if (ea->ea_type == GFS2_EATYPE_UNUSED) ea_list_i()
384 switch (ea->ea_type) { ea_list_i()
402 memcpy(er->er_data + ei->ei_size + l, GFS2_EA2NAME(ea), ea_list_i()
403 ea->ea_name_len); ea_list_i()
455 * @ea: The extended attribute header structure
462 static int gfs2_iter_unstuffed(struct gfs2_inode *ip, struct gfs2_ea_header *ea, gfs2_iter_unstuffed() argument
467 unsigned int amount = GFS2_EA_DATA_LEN(ea); gfs2_iter_unstuffed()
469 __be64 *dataptrs = GFS2_EA2DATAPTRS(ea); gfs2_iter_unstuffed()
623 struct gfs2_ea_header *ea; ea_alloc_blk() local
637 ea = GFS2_EA_BH2FIRST(*bhp); ea_alloc_blk()
638 ea->ea_rec_len = cpu_to_be32(sdp->sd_jbsize); ea_alloc_blk()
639 ea->ea_type = GFS2_EATYPE_UNUSED; ea_alloc_blk()
640 ea->ea_flags = GFS2_EAFLAG_LAST; ea_alloc_blk()
641 ea->ea_num_ptrs = 0; ea_alloc_blk()
649 * ea_write - writes the request info to an ea, creating new blocks if
652 * @ea: the location of the new ea in a block
660 static int ea_write(struct gfs2_inode *ip, struct gfs2_ea_header *ea, ea_write() argument
666 ea->ea_data_len = cpu_to_be32(er->er_data_len); ea_write()
667 ea->ea_name_len = er->er_name_len; ea_write()
668 ea->ea_type = er->er_type; ea_write()
669 ea->__pad = 0; ea_write()
671 memcpy(GFS2_EA2NAME(ea), er->er_name, er->er_name_len); ea_write()
674 ea->ea_num_ptrs = 0; ea_write()
675 memcpy(GFS2_EA2DATA(ea), er->er_data, er->er_data_len); ea_write()
677 __be64 *dataptr = GFS2_EA2DATAPTRS(ea); ea_write()
683 ea->ea_num_ptrs = DIV_ROUND_UP(er->er_data_len, sdp->sd_jbsize); ea_write()
684 for (x = 0; x < ea->ea_num_ptrs; x++) { ea_write()
815 static struct gfs2_ea_header *ea_split_ea(struct gfs2_ea_header *ea) ea_split_ea() argument
817 u32 ea_size = GFS2_EA_SIZE(ea); ea_split_ea()
818 struct gfs2_ea_header *new = (struct gfs2_ea_header *)((char *)ea + ea_split_ea()
820 u32 new_size = GFS2_EA_REC_LEN(ea) - ea_size; ea_split_ea()
821 int last = ea->ea_flags & GFS2_EAFLAG_LAST; ea_split_ea()
823 ea->ea_rec_len = cpu_to_be32(ea_size); ea_split_ea()
824 ea->ea_flags ^= last; ea_split_ea()
835 struct gfs2_ea_header *ea = el->el_ea; ea_set_remove_stuffed() local
841 if (!prev || !GFS2_EA_IS_STUFFED(ea)) { ea_set_remove_stuffed()
842 ea->ea_type = GFS2_EATYPE_UNUSED; ea_set_remove_stuffed()
844 } else if (GFS2_EA2NEXT(prev) != ea) { ea_set_remove_stuffed()
846 gfs2_assert_withdraw(GFS2_SB(&ip->i_inode), GFS2_EA2NEXT(prev) == ea); ea_set_remove_stuffed()
849 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_set_remove_stuffed()
852 if (GFS2_EA_IS_LAST(ea)) ea_set_remove_stuffed()
867 struct gfs2_ea_header *ea, struct ea_set *es) ea_set_simple_noalloc()
880 ea = ea_split_ea(ea); ea_set_simple_noalloc()
882 ea_write(ip, ea, er); ea_set_simple_noalloc()
903 struct gfs2_ea_header *ea = es->es_ea; ea_set_simple_alloc() local
909 ea = ea_split_ea(ea); ea_set_simple_alloc()
911 error = ea_write(ip, ea, er); ea_set_simple_alloc()
922 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_set_simple()
933 if (ea->ea_type == GFS2_EATYPE_UNUSED) { ea_set_simple()
934 if (GFS2_EA_REC_LEN(ea) < size) ea_set_simple()
936 if (!GFS2_EA_IS_STUFFED(ea)) { ea_set_simple()
937 error = ea_remove_unstuffed(ip, bh, ea, prev, 1); ea_set_simple()
942 } else if (GFS2_EA_REC_LEN(ea) - GFS2_EA_SIZE(ea) >= size) ea_set_simple()
948 error = ea_set_simple_noalloc(ip, bh, ea, es); ea_set_simple()
955 es->es_ea = ea; ea_set_simple()
1088 struct gfs2_ea_header *ea = el->el_ea; ea_remove_stuffed() local
1102 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_remove_stuffed()
1105 if (GFS2_EA_IS_LAST(ea)) ea_remove_stuffed()
1108 ea->ea_type = GFS2_EATYPE_UNUSED; ea_remove_stuffed()
1239 struct gfs2_ea_header *ea, char *data) ea_acl_chmod_unstuffed()
1242 unsigned int amount = GFS2_EA_DATA_LEN(ea); ea_acl_chmod_unstuffed()
1250 ret = gfs2_iter_unstuffed(ip, ea, data, NULL); ea_acl_chmod_unstuffed()
166 ea_find_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_find_i() argument
225 ea_dealloc_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_dealloc_unstuffed() argument
325 ea_remove_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, int leave) ea_remove_unstuffed() argument
365 ea_list_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_list_i() argument
866 ea_set_simple_noalloc(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct ea_set *es) ea_set_simple_noalloc() argument
921 ea_set_simple(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_set_simple() argument
1238 ea_acl_chmod_unstuffed(struct gfs2_inode *ip, struct gfs2_ea_header *ea, char *data) ea_acl_chmod_unstuffed() argument
H A Dinode.c441 struct gfs2_ea_header *ea; gfs2_init_xattr() local
448 ea = GFS2_EA_BH2FIRST(bh); gfs2_init_xattr()
449 ea->ea_rec_len = cpu_to_be32(sdp->sd_jbsize); gfs2_init_xattr()
450 ea->ea_type = GFS2_EATYPE_UNUSED; gfs2_init_xattr()
451 ea->ea_flags = GFS2_EAFLAG_LAST; gfs2_init_xattr()
/linux-4.1.27/fs/hpfs/
H A Dea.c2 * linux/fs/hpfs/ea.c
19 struct extended_attribute *ea = (struct extended_attribute *)ex; hpfs_ea_ext_remove() local
26 if (ea_indirect(ea)) { hpfs_ea_ext_remove()
27 if (ea_valuelen(ea) != 8) { hpfs_ea_ext_remove()
28 hpfs_error(s, "ea_indirect(ea) set while ea->valuelen!=8, %s %08x, pos %08x", hpfs_ea_ext_remove()
32 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 9, ex+4)) hpfs_ea_ext_remove()
34 hpfs_ea_remove(s, ea_sec(ea), ea_in_anode(ea), ea_len(ea)); hpfs_ea_ext_remove()
36 pos += ea->namelen + ea_valuelen(ea) + 5; hpfs_ea_ext_remove()
80 struct extended_attribute *ea; hpfs_read_ea() local
82 for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) hpfs_read_ea()
83 if (!strcmp(ea->name, key)) { hpfs_read_ea()
84 if (ea_indirect(ea)) hpfs_read_ea()
86 if (ea_valuelen(ea) >= size) hpfs_read_ea()
88 memcpy(buf, ea_data(ea), ea_valuelen(ea)); hpfs_read_ea()
89 buf[ea_valuelen(ea)] = 0; hpfs_read_ea()
97 ea = (struct extended_attribute *)ex; hpfs_read_ea()
104 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 1 + (ea_indirect(ea) ? 8 : 0), ex + 4)) hpfs_read_ea()
106 if (!strcmp(ea->name, key)) { hpfs_read_ea()
107 if (ea_indirect(ea)) hpfs_read_ea()
109 if (ea_valuelen(ea) >= size) hpfs_read_ea()
111 if (hpfs_ea_read(s, a, ano, pos + 4 + ea->namelen + 1, ea_valuelen(ea), buf)) hpfs_read_ea()
113 buf[ea_valuelen(ea)] = 0; hpfs_read_ea()
116 pos += ea->namelen + ea_valuelen(ea) + 5; hpfs_read_ea()
120 if (ea_len(ea) >= size) hpfs_read_ea()
122 if (hpfs_ea_read(s, ea_sec(ea), ea_in_anode(ea), 0, ea_len(ea), buf)) hpfs_read_ea()
124 buf[ea_len(ea)] = 0; hpfs_read_ea()
135 struct extended_attribute *ea; hpfs_get_ea() local
137 for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) hpfs_get_ea()
138 if (!strcmp(ea->name, key)) { hpfs_get_ea()
139 if (ea_indirect(ea)) hpfs_get_ea()
140 return get_indirect_ea(s, ea_in_anode(ea), ea_sec(ea), *size = ea_len(ea)); hpfs_get_ea()
141 if (!(ret = kmalloc((*size = ea_valuelen(ea)) + 1, GFP_NOFS))) { hpfs_get_ea()
145 memcpy(ret, ea_data(ea), ea_valuelen(ea)); hpfs_get_ea()
146 ret[ea_valuelen(ea)] = 0; hpfs_get_ea()
155 ea = (struct extended_attribute *)ex; hpfs_get_ea()
162 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 1 + (ea_indirect(ea) ? 8 : 0), ex + 4)) hpfs_get_ea()
164 if (!strcmp(ea->name, key)) { hpfs_get_ea()
165 if (ea_indirect(ea)) hpfs_get_ea()
166 return get_indirect_ea(s, ea_in_anode(ea), ea_sec(ea), *size = ea_len(ea)); hpfs_get_ea()
167 if (!(ret = kmalloc((*size = ea_valuelen(ea)) + 1, GFP_NOFS))) { hpfs_get_ea()
171 if (hpfs_ea_read(s, a, ano, pos + 4 + ea->namelen + 1, ea_valuelen(ea), ret)) { hpfs_get_ea()
175 ret[ea_valuelen(ea)] = 0; hpfs_get_ea()
178 pos += ea->namelen + ea_valuelen(ea) + 5; hpfs_get_ea()
185 * when this ea exists, it MUST have the same size as size of data.
198 struct extended_attribute *ea; hpfs_set_ea() local
200 for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) hpfs_set_ea()
201 if (!strcmp(ea->name, key)) { hpfs_set_ea()
202 if (ea_indirect(ea)) { hpfs_set_ea()
203 if (ea_len(ea) == size) hpfs_set_ea()
204 set_indirect_ea(s, ea_in_anode(ea), ea_sec(ea), data, size); hpfs_set_ea()
205 } else if (ea_valuelen(ea) == size) { hpfs_set_ea()
206 memcpy(ea_data(ea), data, size); hpfs_set_ea()
216 ea = (struct extended_attribute *)ex; hpfs_set_ea()
223 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 1 + (ea_indirect(ea) ? 8 : 0), ex + 4)) hpfs_set_ea()
225 if (!strcmp(ea->name, key)) { hpfs_set_ea()
226 if (ea_indirect(ea)) { hpfs_set_ea()
227 if (ea_len(ea) == size) hpfs_set_ea()
228 set_indirect_ea(s, ea_in_anode(ea), ea_sec(ea), data, size); hpfs_set_ea()
231 if (ea_valuelen(ea) == size) hpfs_set_ea()
232 hpfs_ea_write(s, a, ano, pos + 4 + ea->namelen + 1, size, data); hpfs_set_ea()
236 pos += ea->namelen + ea_valuelen(ea) + 5; hpfs_set_ea()
254 ea = fnode_end_ea(fnode); hpfs_set_ea()
255 *(char *)ea = 0; hpfs_set_ea()
256 ea->namelen = strlen(key); hpfs_set_ea()
257 ea->valuelen_lo = size; hpfs_set_ea()
258 ea->valuelen_hi = size >> 8; hpfs_set_ea()
259 strcpy(ea->name, key); hpfs_set_ea()
260 memcpy(ea_data(ea), data, size); hpfs_set_ea()
265 But what .. HPFS.IFS has also bugs in ea management. */ hpfs_set_ea()
297 /* Aargh... don't know how to create ea anodes :-( */ hpfs_set_ea()
H A DMakefile7 hpfs-objs := alloc.o anode.o buffer.o dentry.o dir.o dnode.o ea.o file.o \
H A Dinode.c49 void *ea; hpfs_read_inode() local
62 if ((ea = hpfs_get_ea(i->i_sb, fnode, "UID", &ea_size))) { hpfs_read_inode()
64 i_uid_write(i, le16_to_cpu(*(__le16*)ea)); hpfs_read_inode()
67 kfree(ea); hpfs_read_inode()
69 if ((ea = hpfs_get_ea(i->i_sb, fnode, "GID", &ea_size))) { hpfs_read_inode()
71 i_gid_write(i, le16_to_cpu(*(__le16*)ea)); hpfs_read_inode()
74 kfree(ea); hpfs_read_inode()
76 if ((ea = hpfs_get_ea(i->i_sb, fnode, "SYMLINK", &ea_size))) { hpfs_read_inode()
77 kfree(ea); hpfs_read_inode()
87 if ((ea = hpfs_get_ea(i->i_sb, fnode, "MODE", &ea_size))) { hpfs_read_inode()
91 mode = le16_to_cpu(*(__le16*)ea); hpfs_read_inode()
94 kfree(ea); hpfs_read_inode()
97 if ((ea = hpfs_get_ea(i->i_sb, fnode, "DEV", &ea_size))) { hpfs_read_inode()
99 rdev = le32_to_cpu(*(__le32*)ea); hpfs_read_inode()
100 kfree(ea); hpfs_read_inode()
152 __le32 ea; hpfs_write_inode_ea() local
154 ea = cpu_to_le32(i_uid_read(i)); hpfs_write_inode_ea()
155 hpfs_set_ea(i, fnode, "UID", (char*)&ea, 2); hpfs_write_inode_ea()
159 ea = cpu_to_le32(i_gid_read(i)); hpfs_write_inode_ea()
160 hpfs_set_ea(i, fnode, "GID", (char *)&ea, 2); hpfs_write_inode_ea()
168 ea = cpu_to_le32(i->i_mode); hpfs_write_inode_ea()
170 hpfs_set_ea(i, fnode, "MODE", (char *)&ea, 2); hpfs_write_inode_ea()
174 ea = cpu_to_le32(new_encode_dev(i->i_rdev)); hpfs_write_inode_ea()
175 hpfs_set_ea(i, fnode, "DEV", (char *)&ea, 4); hpfs_write_inode_ea()
H A Dhpfs_fn.h54 unsigned i_ea_mode : 1; /* file's permission is stored in ea */
55 unsigned i_ea_uid : 1; /* file's uid is stored in ea */
56 unsigned i_ea_gid : 1; /* file's gid is stored in ea */
139 static unsigned ea_valuelen(struct extended_attribute *ea) ea_valuelen() argument
141 return ea->valuelen_lo + 256 * ea->valuelen_hi; ea_valuelen()
144 static inline struct extended_attribute *next_ea(struct extended_attribute *ea) next_ea() argument
146 return (struct extended_attribute *)((char *)ea + 5 + ea->namelen + ea_valuelen(ea)); next_ea()
149 static inline secno ea_sec(struct extended_attribute *ea) ea_sec() argument
151 return le32_to_cpu(get_unaligned((__le32 *)((char *)ea + 9 + ea->namelen))); ea_sec()
154 static inline secno ea_len(struct extended_attribute *ea) ea_len() argument
156 return le32_to_cpu(get_unaligned((__le32 *)((char *)ea + 5 + ea->namelen))); ea_len()
159 static inline char *ea_data(struct extended_attribute *ea) ea_data() argument
161 return (char *)((char *)ea + 5 + ea->namelen); ea_data()
252 /* ea.c */
H A Danode.c349 if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, sec, 1, "ea #1")) return -1; hpfs_ea_read()
372 if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, sec, 1, "ea #2")) return -1; hpfs_ea_write()
484 struct extended_attribute *ea; hpfs_remove_fnode() local
490 for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) hpfs_remove_fnode()
491 if (ea_indirect(ea)) hpfs_remove_fnode()
492 hpfs_ea_remove(s, ea_sec(ea), ea_in_anode(ea), ea_len(ea)); hpfs_remove_fnode()
H A Dmap.c145 struct extended_attribute *ea; hpfs_map_fnode() local
176 ea = fnode_ea(fnode); hpfs_map_fnode()
178 while (ea != ea_end) { hpfs_map_fnode()
179 if (ea > ea_end) { hpfs_map_fnode()
184 ea = next_ea(ea); hpfs_map_fnode()
H A Dhpfs.h399 the data btree or some ea or the
400 main ea bootage pointer ea_secno
447 __le32 ea_size_l; /* length of disk-resident ea's */
448 __le32 ea_secno; /* first sector of disk-resident ea's*/
449 __le16 ea_size_s; /* length of fnode-resident ea's */
464 to first fnode-resident ea */
469 u8 ea[316]; /* zero or more EA's, packed together member in struct:fnode
528 /* bit 7 set -> required ea */
545 static inline bool ea_indirect(struct extended_attribute *ea) ea_indirect() argument
547 return ea->flags & EA_indirect; ea_indirect()
550 static inline bool ea_in_anode(struct extended_attribute *ea) ea_in_anode() argument
552 return ea->flags & EA_anode; ea_in_anode()
/linux-4.1.27/arch/powerpc/mm/
H A Dcopro_fault.c36 int copro_handle_mm_fault(struct mm_struct *mm, unsigned long ea, copro_handle_mm_fault() argument
51 vma = find_vma(mm, ea); copro_handle_mm_fault()
55 if (ea < vma->vm_start) { copro_handle_mm_fault()
58 if (expand_stack(vma, ea)) copro_handle_mm_fault()
78 *flt = handle_mm_fault(mm, vma, ea, is_write ? FAULT_FLAG_WRITE : 0); copro_handle_mm_fault()
101 int copro_calculate_slb(struct mm_struct *mm, u64 ea, struct copro_slb *slb) copro_calculate_slb() argument
106 switch (REGION_ID(ea)) { copro_calculate_slb()
108 pr_devel("%s: 0x%llx -- USER_REGION_ID\n", __func__, ea); copro_calculate_slb()
109 psize = get_slice_psize(mm, ea); copro_calculate_slb()
110 ssize = user_segment_size(ea); copro_calculate_slb()
111 vsid = get_vsid(mm->context.id, ea, ssize); copro_calculate_slb()
114 pr_devel("%s: 0x%llx -- VMALLOC_REGION_ID\n", __func__, ea); copro_calculate_slb()
115 if (ea < VMALLOC_END) copro_calculate_slb()
120 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); copro_calculate_slb()
123 pr_devel("%s: 0x%llx -- KERNEL_REGION_ID\n", __func__, ea); copro_calculate_slb()
126 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); copro_calculate_slb()
129 pr_debug("%s: invalid region access at %016llx\n", __func__, ea); copro_calculate_slb()
138 slb->esid = (ea & (ssize == MMU_SEGSIZE_1T ? ESID_MASK_1T : ESID_MASK)) | SLB_ESID_V; copro_calculate_slb()
H A Dhugetlbpage-book3e.c54 static inline int book3e_tlb_exists(unsigned long ea, unsigned long pid) book3e_tlb_exists() argument
66 : "=&r"(found) : "r"(ea)); book3e_tlb_exists()
72 : "=&r"(found) : "r"(ea)); book3e_tlb_exists()
78 void book3e_hugetlb_preload(struct vm_area_struct *vma, unsigned long ea, book3e_hugetlb_preload() argument
91 if (unlikely(is_kernel_addr(ea))) book3e_hugetlb_preload()
97 psize = get_slice_psize(mm, ea); book3e_hugetlb_preload()
112 if (unlikely(book3e_tlb_exists(ea, mm->context.id))) { book3e_hugetlb_preload()
124 mas2 = ea & ~((1UL << shift) - 1); book3e_hugetlb_preload()
H A Dslb.c29 extern void slb_allocate_realmode(unsigned long ea);
30 extern void slb_allocate_user(unsigned long ea);
32 static void slb_allocate(unsigned long ea) slb_allocate() argument
37 slb_allocate_realmode(ea); slb_allocate()
43 static inline unsigned long mk_esid_data(unsigned long ea, int ssize, mk_esid_data() argument
46 return (ea & slb_esid_mask(ssize)) | SLB_ESID_V | slot; mk_esid_data()
49 static inline unsigned long mk_vsid_data(unsigned long ea, int ssize, mk_vsid_data() argument
52 return (get_kernel_vsid(ea, ssize) << slb_vsid_shift(ssize)) | flags | mk_vsid_data()
56 static inline void slb_shadow_update(unsigned long ea, int ssize, slb_shadow_update() argument
67 cpu_to_be64(mk_vsid_data(ea, ssize, flags)); slb_shadow_update()
69 cpu_to_be64(mk_esid_data(ea, ssize, entry)); slb_shadow_update()
77 static inline void create_shadowed_slbe(unsigned long ea, int ssize, create_shadowed_slbe() argument
86 slb_shadow_update(ea, ssize, flags, entry); create_shadowed_slbe()
89 : "r" (mk_vsid_data(ea, ssize, flags)), create_shadowed_slbe()
90 "r" (mk_esid_data(ea, ssize, entry)) create_shadowed_slbe()
H A Dhash_utils_64.c922 static int subpage_protection(struct mm_struct *mm, unsigned long ea) subpage_protection() argument
928 if (ea >= spt->maxaddr) subpage_protection()
930 if (ea < 0x100000000UL) { subpage_protection()
934 sbpm = spt->protptrs[ea >> SBP_L3_SHIFT]; subpage_protection()
938 sbpp = sbpm[(ea >> SBP_L2_SHIFT) & (SBP_L2_COUNT - 1)]; subpage_protection()
941 spp = sbpp[(ea >> PAGE_SHIFT) & (SBP_L1_COUNT - 1)]; subpage_protection()
944 spp >>= 30 - 2 * ((ea >> 12) & 0xf); subpage_protection()
952 static inline int subpage_protection(struct mm_struct *mm, unsigned long ea) subpage_protection() argument
958 void hash_failure_debug(unsigned long ea, unsigned long access, hash_failure_debug() argument
965 ea, access, current->comm); hash_failure_debug()
970 static void check_paca_psize(unsigned long ea, struct mm_struct *mm, check_paca_psize() argument
974 if (psize != get_paca_psize(ea)) { check_paca_psize()
992 int hash_page_mm(struct mm_struct *mm, unsigned long ea, hash_page_mm() argument
1005 DBG_LOW("hash_page(ea=%016lx, access=%lx, trap=%lx\n", hash_page_mm()
1006 ea, access, trap); hash_page_mm()
1009 switch (REGION_ID(ea)) { hash_page_mm()
1017 psize = get_slice_psize(mm, ea); hash_page_mm()
1018 ssize = user_segment_size(ea); hash_page_mm()
1019 vsid = get_vsid(mm->context.id, ea, ssize); hash_page_mm()
1022 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); hash_page_mm()
1023 if (ea < VMALLOC_END) hash_page_mm()
1065 ea &= ~((1ul << mmu_psize_defs[psize].shift) - 1); hash_page_mm()
1069 ptep = __find_linux_pte_or_hugepte(pgdir, ea, &hugeshift); hash_page_mm()
1090 rc = __hash_page_thp(ea, access, vsid, (pmd_t *)ptep, hash_page_mm()
1094 rc = __hash_page_huge(ea, access, vsid, ptep, trap, hash_page_mm()
1107 check_paca_psize(ea, mm, psize, user_region); hash_page_mm()
1122 demote_segment_4k(mm, ea); hash_page_mm()
1132 demote_segment_4k(mm, ea); hash_page_mm()
1134 } else if (ea < VMALLOC_END) { hash_page_mm()
1149 check_paca_psize(ea, mm, psize, user_region); hash_page_mm()
1154 rc = __hash_page_64K(ea, access, vsid, ptep, trap, hash_page_mm()
1159 int spp = subpage_protection(mm, ea); hash_page_mm()
1163 rc = __hash_page_4K(ea, access, vsid, ptep, trap, hash_page_mm()
1171 hash_failure_debug(ea, access, vsid, trap, ssize, psize, hash_page_mm()
1187 int hash_page(unsigned long ea, unsigned long access, unsigned long trap, hash_page() argument
1193 if (REGION_ID(ea) == VMALLOC_REGION_ID) hash_page()
1199 return hash_page_mm(mm, ea, access, trap, flags); hash_page()
1203 void hash_preload(struct mm_struct *mm, unsigned long ea, hash_preload() argument
1213 BUG_ON(REGION_ID(ea) != USER_REGION_ID); hash_preload()
1217 if (unlikely(get_slice_psize(mm, ea) != mm->context.user_psize)) hash_preload()
1221 DBG_LOW("hash_preload(mm=%p, mm->pgdir=%p, ea=%016lx, access=%lx," hash_preload()
1222 " trap=%lx\n", mm, mm->pgd, ea, access, trap); hash_preload()
1230 ssize = user_segment_size(ea); hash_preload()
1231 vsid = get_vsid(mm->context.id, ea, ssize); hash_preload()
1244 ptep = find_linux_pte_or_hugepte(pgdir, ea, &hugepage_shift); hash_preload()
1267 rc = __hash_page_64K(ea, access, vsid, ptep, trap, hash_preload()
1271 rc = __hash_page_4K(ea, access, vsid, ptep, trap, update_flags, hash_preload()
1272 ssize, subpage_protection(mm, ea)); hash_preload()
1278 hash_failure_debug(ea, access, vsid, trap, ssize, hash_preload()
H A Dhugepage-hash64.c21 int __hash_page_thp(unsigned long ea, unsigned long access, unsigned long vsid, __hash_page_thp() argument
81 * Find the slot index details for this ea, using base page size. __hash_page_thp()
84 index = (ea & ~HPAGE_PMD_MASK) >> shift; __hash_page_thp()
87 vpn = hpt_vpn(ea, vsid, ssize); __hash_page_thp()
96 flush_hash_hugepage(vsid, ea, pmdp, MMU_PAGE_64K, __hash_page_thp()
171 hash_failure_debug(ea, access, vsid, trap, ssize, __hash_page_thp()
H A Dpgtable_64.c91 int map_kernel_page(unsigned long ea, unsigned long pa, int flags) map_kernel_page() argument
99 pgdp = pgd_offset_k(ea); map_kernel_page()
100 pudp = pud_alloc(&init_mm, pgdp, ea); map_kernel_page()
103 pmdp = pmd_alloc(&init_mm, pudp, ea); map_kernel_page()
106 ptep = pte_alloc_kernel(pmdp, ea); map_kernel_page()
109 set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, map_kernel_page()
113 pgdp = pgd_offset_k(ea); map_kernel_page()
121 pudp = pud_offset(pgdp, ea); map_kernel_page()
127 pmdp = pmd_offset(pudp, ea); map_kernel_page()
133 ptep = pte_offset_kernel(pmdp, ea); map_kernel_page()
134 set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, map_kernel_page()
143 if (htab_bolt_mapping(ea, ea + PAGE_SIZE, pa, flags, map_kernel_page()
171 void __iomem * __ioremap_at(phys_addr_t pa, void *ea, unsigned long size, __ioremap_at() argument
189 WARN_ON(((unsigned long)ea) & ~PAGE_MASK); __ioremap_at()
193 if (map_kernel_page((unsigned long)ea+i, pa+i, flags)) __ioremap_at()
196 return (void __iomem *)ea; __ioremap_at()
205 void __iounmap_at(void *ea, unsigned long size) __iounmap_at() argument
207 WARN_ON(((unsigned long)ea) & ~PAGE_MASK); __iounmap_at()
210 unmap_kernel_range((unsigned long)ea, size); __iounmap_at()
H A Dppc_mmu_32.c165 void hash_preload(struct mm_struct *mm, unsigned long ea, hash_preload() argument
172 pmd = pmd_offset(pud_offset(pgd_offset(mm, ea), ea), ea); hash_preload()
174 add_hash_page(mm->context.id, ea, pmd_val(*pmd)); hash_preload()
H A Dhugetlbpage-hash64.c21 int __hash_page_huge(unsigned long ea, unsigned long access, unsigned long vsid, __hash_page_huge() argument
33 vpn = hpt_vpn(ea, vsid, ssize); __hash_page_huge()
116 hash_failure_debug(ea, access, vsid, trap, ssize, __hash_page_huge()
H A Dslb_low.S26 /* void slb_allocate_realmode(unsigned long ea);
36 * (ea & ~REGION_MASK) >= PGTABLE_RANGE
66 * context = (MAX_USER_CONTEXT) + ((ea >> 60) - 0xc) + 1
105 * context = (MAX_USER_CONTEXT) + ((ea >> 60) - 0xc) + 1
184 /* void slb_allocate_user(unsigned long ea);
H A Dmmu_decl.h85 extern void hash_preload(struct mm_struct *mm, unsigned long ea,
113 extern int map_kernel_page(unsigned long ea, unsigned long pa, int flags);
H A Dhugetlbpage.c971 pte_t *__find_linux_pte_or_hugepte(pgd_t *pgdir, unsigned long ea, __find_linux_pte_or_hugepte() argument
984 pgdp = pgdir + pgd_index(ea); __find_linux_pte_or_hugepte()
1006 pudp = pud_offset(&pgd, ea); __find_linux_pte_or_hugepte()
1018 pmdp = pmd_offset(&pud, ea); __find_linux_pte_or_hugepte()
1037 return pte_offset_kernel(&pmd, ea); __find_linux_pte_or_hugepte()
1043 ret_pte = hugepte_offset(*hpdp, ea, pdshift); __find_linux_pte_or_hugepte()
H A Dhash_native_64.c370 static void native_hpte_updateboltedpp(unsigned long newpp, unsigned long ea, native_hpte_updateboltedpp() argument
378 vsid = get_kernel_vsid(ea, ssize); native_hpte_updateboltedpp()
379 vpn = hpt_vpn(ea, vsid, ssize); native_hpte_updateboltedpp()
H A Dhash_low_64.S48 * _hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid,
341 /* _hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid,
419 * clrldi r3,r3,64 - SID_SHIFT --> ea & 0xfffffff
436 * clrldi r3,r3,64 - SID_SHIFT_1T --> ea & 0xffffffffff
/linux-4.1.27/fs/jfs/
H A Dxattr.c75 dxd_t new_ea; /* dxd to replace ea when modifying xattr */
76 struct metapage *mp; /* metapage containing ea list */
77 struct jfs_ea_list *xattr; /* buffer containing ea list */
105 static int is_os2_xattr(struct jfs_ea *ea) is_os2_xattr() argument
107 return !is_known_namespace(ea->name); is_os2_xattr()
110 static inline int name_size(struct jfs_ea *ea) name_size() argument
112 if (is_os2_xattr(ea)) name_size()
113 return ea->namelen + XATTR_OS2_PREFIX_LEN; name_size()
115 return ea->namelen; name_size()
118 static inline int copy_name(char *buffer, struct jfs_ea *ea) copy_name() argument
120 int len = ea->namelen; copy_name()
122 if (is_os2_xattr(ea)) { copy_name()
127 memcpy(buffer, ea->name, ea->namelen); copy_name()
128 buffer[ea->namelen] = 0; copy_name()
148 * ea - dxd_t structure to be filled in with necessary EA information
153 * and sets <ea> fields appropriately. Otherwise, returns failure, EA will
159 int size, dxd_t * ea) ea_write_inline()
174 if (!(ji->mode2 & INLINEEA) && !(ji->ea.flag & DXD_INLINE)) ea_write_inline()
177 DXDsize(ea, size); ea_write_inline()
178 DXDlength(ea, 0); ea_write_inline()
179 DXDaddress(ea, 0); ea_write_inline()
181 ea->flag = DXD_INLINE; ea_write_inline()
184 ea->flag = 0; ea_write_inline()
185 DXDsize(ea, 0); ea_write_inline()
186 DXDlength(ea, 0); ea_write_inline()
187 DXDaddress(ea, 0); ea_write_inline()
190 if (ji->ea.flag & DXD_INLINE) ea_write_inline()
208 * ea - dxd_t structure to be filled in appropriately with where the
217 dxd_t * ea) ea_write()
235 if (!ea_write_inline(ip, ealist, size, ea)) ea_write()
302 ea->flag = DXD_EXTENT; ea_write()
303 DXDsize(ea, le32_to_cpu(ealist->size)); ea_write()
304 DXDlength(ea, nblocks); ea_write()
305 DXDaddress(ea, blkno); ea_write()
308 if (ji->ea.flag & DXD_INLINE) ea_write()
335 int ea_size = sizeDXD(&ji->ea); ea_read_inline()
343 if ((sizeDXD(&ji->ea) > sizeof (ji->i_inline_ea))) ea_read_inline()
380 if (ji->ea.flag & DXD_INLINE) ea_read()
383 nbytes = sizeDXD(&ji->ea); ea_read()
393 nblocks = lengthDXD(&ji->ea) << sbi->l2nbperpage; ea_read()
394 blkno = addressDXD(&ji->ea) << sbi->l2nbperpage; ea_read()
446 int ea_size = sizeDXD(&ji->ea); ea_get()
452 /* When fsck.jfs clears a bad ea, it doesn't clear the size */ ea_get()
453 if (ji->ea.flag == 0) ea_get()
475 } else if (ji->ea.flag & DXD_INLINE) { ea_get()
484 if (!(ji->ea.flag & DXD_EXTENT)) { ea_get()
485 jfs_error(sb, "invalid ea.flag\n"); ea_get()
560 ea_buf->mp = read_metapage(inode, addressDXD(&ji->ea), ea_get()
561 lengthDXD(&ji->ea) << sb->s_blocksize_bits, ea_get()
629 /* ->xattr must point to original ea's metapage */ ea_put()
638 if (ji->ea.flag & DXD_EXTENT) { ea_put()
639 invalidate_dxd_metapages(inode, ji->ea); ea_put()
640 old_blocks = lengthDXD(&ji->ea); ea_put()
644 txEA(tid, inode, &ji->ea, &ea_buf->new_ea); ea_put()
647 if (ji->ea.flag & DXD_INLINE) ea_put()
650 ji->ea = ea_buf->new_ea; ea_put()
652 txEA(tid, inode, &ji->ea, NULL); ea_put()
653 if (ji->ea.flag & DXD_INLINE) ea_put()
655 ji->ea.flag = 0; ea_put()
656 ji->ea.size = 0; ea_put()
701 struct jfs_ea *ea, *old_ea = NULL, *next_ea = NULL; __jfs_setxattr() local
735 for (ea = FIRST_EA(ealist); ea < END_EALIST(ealist); __jfs_setxattr()
736 ea = NEXT_EA(ea)) { __jfs_setxattr()
737 if ((namelen == ea->namelen) && __jfs_setxattr()
738 (memcmp(name, ea->name, namelen) == 0)) { __jfs_setxattr()
744 old_ea = ea; __jfs_setxattr()
745 old_ea_size = EA_SIZE(ea); __jfs_setxattr()
746 next_ea = NEXT_EA(ea); __jfs_setxattr()
748 new_size += EA_SIZE(ea); __jfs_setxattr()
767 * We need to allocate more space for merged ea list. __jfs_setxattr()
779 /* Remove old ea of the same name */ __jfs_setxattr()
791 /* Completely new ea list */ __jfs_setxattr()
807 ea = (struct jfs_ea *) ((char *) ealist + xattr_size); __jfs_setxattr()
808 ea->flag = 0; __jfs_setxattr()
809 ea->namelen = namelen; __jfs_setxattr()
810 ea->valuelen = (cpu_to_le16(value_len)); __jfs_setxattr()
811 memcpy(ea->name, name, namelen); __jfs_setxattr()
812 ea->name[namelen] = 0; __jfs_setxattr()
814 memcpy(&ea->name[namelen + 1], value, value_len); __jfs_setxattr()
815 xattr_size += EA_SIZE(ea); __jfs_setxattr()
829 * If we're left with an empty list, there's no ea __jfs_setxattr()
889 struct jfs_ea *ea; __jfs_getxattr() local
911 for (ea = FIRST_EA(ealist); ea < END_EALIST(ealist); ea = NEXT_EA(ea)) __jfs_getxattr()
912 if ((namelen == ea->namelen) && __jfs_getxattr()
913 memcmp(name, ea->name, namelen) == 0) { __jfs_getxattr()
915 size = le16_to_cpu(ea->valuelen); __jfs_getxattr()
922 value = ((char *) &ea->name) + ea->namelen + 1; __jfs_getxattr()
970 static inline int can_list(struct jfs_ea *ea) can_list() argument
972 return (strncmp(ea->name, XATTR_TRUSTED_PREFIX, can_list()
984 struct jfs_ea *ea; jfs_listxattr() local
1001 for (ea = FIRST_EA(ealist); ea < END_EALIST(ealist); ea = NEXT_EA(ea)) { jfs_listxattr()
1002 if (can_list(ea)) jfs_listxattr()
1003 size += name_size(ea) + 1; jfs_listxattr()
1016 for (ea = FIRST_EA(ealist); ea < END_EALIST(ealist); ea = NEXT_EA(ea)) { jfs_listxattr()
1017 if (can_list(ea)) { jfs_listxattr()
1018 int namelen = copy_name(buffer, ea); jfs_listxattr()
158 ea_write_inline(struct inode *ip, struct jfs_ea_list *ealist, int size, dxd_t * ea) ea_write_inline() argument
216 ea_write(struct inode *ip, struct jfs_ea_list *ealist, int size, dxd_t * ea) ea_write() argument
H A Djfs_xattr.h36 struct jfs_ea ea[0]; /* Variable length list */ member in struct:jfs_ea_list
46 #define EA_SIZE(ea) \
47 (sizeof (struct jfs_ea) + (ea)->namelen + 1 + \
48 le16_to_cpu((ea)->valuelen))
49 #define NEXT_EA(ea) ((struct jfs_ea *) (((char *) (ea)) + (EA_SIZE (ea))))
50 #define FIRST_EA(ealist) ((ealist)->ea)
H A Dnamei.c51 dxd_t *ea = &JFS_IP(inode)->ea; free_ea_wmap() local
53 if (ea->flag & DXD_EXTENT) { free_ea_wmap()
55 invalidate_dxd_metapages(inode, *ea); free_ea_wmap()
56 dbFree(inode, addressDXD(ea), lengthDXD(ea)); free_ea_wmap()
58 ea->flag = 0; free_ea_wmap()
408 if (JFS_IP(ip)->ea.flag & DXD_EXTENT) { jfs_rmdir()
410 txEA(tid, ip, &JFS_IP(ip)->ea, NULL); jfs_rmdir()
412 JFS_IP(ip)->ea.flag = 0; jfs_rmdir()
658 if (JFS_IP(ip)->ea.flag & DXD_EXTENT) commitZeroLink()
660 txEA(tid, ip, &JFS_IP(ip)->ea, NULL); commitZeroLink()
718 if (JFS_IP(ip)->ea.flag & DXD_EXTENT) { jfs_free_zero_link()
719 s64 xaddr = addressDXD(&JFS_IP(ip)->ea); jfs_free_zero_link()
720 int xlen = lengthDXD(&JFS_IP(ip)->ea); jfs_free_zero_link()
725 invalidate_dxd_metapages(ip, JFS_IP(ip)->ea); jfs_free_zero_link()
H A Djfs_inode.c140 memset(&jfs_inode->ea, 0, sizeof(dxd_t)); ialloc()
H A Djfs_incore.h45 dxd_t ea; /* dxd describing ea */ member in struct:jfs_inode_info
H A Djfs_dinode.h66 dxd_t di_ea; /* 16: ea descriptor */
H A Djfs_imap.c3096 jfs_ip->ea = dip->di_ea; copy_from_dinode()
3171 dip->di_ea = jfs_ip->ea; copy_to_dinode()
/linux-4.1.27/arch/powerpc/lib/
H A Dsstep.c38 extern int do_lfs(int rn, unsigned long ea);
39 extern int do_lfd(int rn, unsigned long ea);
40 extern int do_stfs(int rn, unsigned long ea);
41 extern int do_stfd(int rn, unsigned long ea);
42 extern int do_lvx(int rn, unsigned long ea);
43 extern int do_stvx(int rn, unsigned long ea);
44 extern int do_lxvd2x(int rn, unsigned long ea);
45 extern int do_stxvd2x(int rn, unsigned long ea);
84 static long __kprobes address_ok(struct pt_regs *regs, unsigned long ea, int nb) address_ok() argument
88 return __access_ok(ea, nb, USER_DS); address_ok()
97 unsigned long ea; dform_ea() local
100 ea = (signed short) instr; /* sign-extend */ dform_ea()
102 ea += regs->gpr[ra]; dform_ea()
104 return truncate_if_32bit(regs->msr, ea); dform_ea()
114 unsigned long ea; dsform_ea() local
117 ea = (signed short) (instr & ~3); /* sign-extend */ dsform_ea()
119 ea += regs->gpr[ra]; dsform_ea()
121 return truncate_if_32bit(regs->msr, ea); dsform_ea()
132 unsigned long ea; xform_ea() local
136 ea = regs->gpr[rb]; xform_ea()
138 ea += regs->gpr[ra]; xform_ea()
140 return truncate_if_32bit(regs->msr, ea); xform_ea()
172 static int __kprobes read_mem_aligned(unsigned long *dest, unsigned long ea, read_mem_aligned() argument
180 err = __get_user(x, (unsigned char __user *) ea); read_mem_aligned()
183 err = __get_user(x, (unsigned short __user *) ea); read_mem_aligned()
186 err = __get_user(x, (unsigned int __user *) ea); read_mem_aligned()
190 err = __get_user(x, (unsigned long __user *) ea); read_mem_aligned()
199 static int __kprobes read_mem_unaligned(unsigned long *dest, unsigned long ea, read_mem_unaligned() argument
215 c = max_align(ea); read_mem_unaligned()
219 err = read_mem_aligned(&b, ea, c); read_mem_unaligned()
223 ea += c; read_mem_unaligned()
247 * Read memory at address ea for nb bytes, return 0 for success
250 static int __kprobes read_mem(unsigned long *dest, unsigned long ea, int nb, read_mem() argument
253 if (!address_ok(regs, ea, nb)) read_mem()
255 if ((ea & (nb - 1)) == 0) read_mem()
256 return read_mem_aligned(dest, ea, nb); read_mem()
257 return read_mem_unaligned(dest, ea, nb, regs); read_mem()
260 static int __kprobes write_mem_aligned(unsigned long val, unsigned long ea, write_mem_aligned() argument
267 err = __put_user(val, (unsigned char __user *) ea); write_mem_aligned()
270 err = __put_user(val, (unsigned short __user *) ea); write_mem_aligned()
273 err = __put_user(val, (unsigned int __user *) ea); write_mem_aligned()
277 err = __put_user(val, (unsigned long __user *) ea); write_mem_aligned()
284 static int __kprobes write_mem_unaligned(unsigned long val, unsigned long ea, write_mem_unaligned() argument
311 c = max_align(ea); write_mem_unaligned()
315 err = write_mem_aligned(val >> (nb - c) * 8, ea, c); write_mem_unaligned()
318 ea += c; write_mem_unaligned()
324 * Write memory at address ea for nb bytes, return 0 for success
327 static int __kprobes write_mem(unsigned long val, unsigned long ea, int nb, write_mem() argument
330 if (!address_ok(regs, ea, nb)) write_mem()
332 if ((ea & (nb - 1)) == 0) write_mem()
333 return write_mem_aligned(val, ea, nb); write_mem()
334 return write_mem_unaligned(val, ea, nb, regs); write_mem()
343 unsigned long ea, int nb, do_fp_load()
363 if (!address_ok(regs, ea, nb)) do_fp_load()
365 if ((ea & 3) == 0) do_fp_load()
366 return (*func)(rn, ea); do_fp_load()
369 err = read_mem_unaligned(&data.ul[0], ea, nb, regs); do_fp_load()
374 err = read_mem_unaligned(&data.ul[0], ea, 4, regs); do_fp_load()
376 err = read_mem_unaligned(&data.ul[1], ea + 4, 4, regs); do_fp_load()
384 unsigned long ea, int nb, do_fp_store()
404 if (!address_ok(regs, ea, nb)) do_fp_store()
406 if ((ea & 3) == 0) do_fp_store()
407 return (*func)(rn, ea); do_fp_store()
415 err = write_mem_unaligned(data.ul[0], ea, nb, regs); do_fp_store()
421 err = write_mem_unaligned(data.ul[0], ea, 4, regs); do_fp_store()
423 err = write_mem_unaligned(data.ul[1], ea + 4, 4, regs); do_fp_store()
432 unsigned long ea, struct pt_regs *regs) do_vec_load()
434 if (!address_ok(regs, ea & ~0xfUL, 16)) do_vec_load()
436 return (*func)(rn, ea); do_vec_load()
440 unsigned long ea, struct pt_regs *regs) do_vec_store()
442 if (!address_ok(regs, ea & ~0xfUL, 16)) do_vec_store()
444 return (*func)(rn, ea); do_vec_store()
450 unsigned long ea, struct pt_regs *regs) do_vsx_load()
455 if (!address_ok(regs, ea, 16)) do_vsx_load()
457 if ((ea & 3) == 0) do_vsx_load()
458 return (*func)(rn, ea); do_vsx_load()
459 err = read_mem_unaligned(&val[0], ea, 8, regs); do_vsx_load()
461 err = read_mem_unaligned(&val[1], ea + 8, 8, regs); do_vsx_load()
468 unsigned long ea, struct pt_regs *regs) do_vsx_store()
473 if (!address_ok(regs, ea, 16)) do_vsx_store()
475 if ((ea & 3) == 0) do_vsx_store()
476 return (*func)(rn, ea); do_vsx_store()
480 err = write_mem_unaligned(val[0], ea, 8, regs); do_vsx_store()
482 err = write_mem_unaligned(val[1], ea + 8, 8, regs); do_vsx_store()
1300 op->ea = xform_ea(instr, regs); analyse_instr()
1305 op->ea = xform_ea(instr, regs); analyse_instr()
1310 op->ea = xform_ea(instr, regs); analyse_instr()
1316 op->ea = xform_ea(instr, regs); analyse_instr()
1322 op->ea = xform_ea(instr, regs); analyse_instr()
1340 op->ea = xform_ea(instr, regs); analyse_instr()
1448 op->ea = 0; analyse_instr()
1450 op->ea = truncate_if_32bit(regs->msr, analyse_instr()
1504 op->ea = 0; analyse_instr()
1506 op->ea = truncate_if_32bit(regs->msr, analyse_instr()
1543 op->ea = dform_ea(instr, regs); analyse_instr()
1549 op->ea = dform_ea(instr, regs); analyse_instr()
1555 op->ea = dform_ea(instr, regs); analyse_instr()
1561 op->ea = dform_ea(instr, regs); analyse_instr()
1567 op->ea = dform_ea(instr, regs); analyse_instr()
1573 op->ea = dform_ea(instr, regs); analyse_instr()
1579 op->ea = dform_ea(instr, regs); analyse_instr()
1586 op->ea = dform_ea(instr, regs); analyse_instr()
1591 op->ea = dform_ea(instr, regs); analyse_instr()
1600 op->ea = dform_ea(instr, regs); analyse_instr()
1608 op->ea = dform_ea(instr, regs); analyse_instr()
1616 op->ea = dform_ea(instr, regs); analyse_instr()
1624 op->ea = dform_ea(instr, regs); analyse_instr()
1630 op->ea = dsform_ea(instr, regs); analyse_instr()
1645 op->ea = dsform_ea(instr, regs); analyse_instr()
1711 static __kprobes int handle_stack_update(unsigned long ea, struct pt_regs *regs) handle_stack_update() argument
1717 if (ea - STACK_INT_FRAME_SIZE <= current->thread.ksp_limit) { handle_stack_update()
1783 if (!address_ok(regs, op.ea, 8)) emulate_step()
1787 __cacheop_user_asmx(op.ea, err, "dcbst"); emulate_step()
1790 __cacheop_user_asmx(op.ea, err, "dcbf"); emulate_step()
1794 prefetchw((void *) op.ea); emulate_step()
1798 prefetch((void *) op.ea); emulate_step()
1801 __cacheop_user_asmx(op.ea, err, "icbi"); emulate_step()
1811 if (op.ea & (size - 1)) emulate_step()
1814 if (!address_ok(regs, op.ea, size)) emulate_step()
1819 __get_user_asmx(val, op.ea, err, "lwarx"); emulate_step()
1822 __get_user_asmx(val, op.ea, err, "ldarx"); emulate_step()
1834 if (op.ea & (size - 1)) emulate_step()
1837 if (!address_ok(regs, op.ea, size)) emulate_step()
1842 __put_user_asmx(op.val, op.ea, err, "stwcx.", cr); emulate_step()
1845 __put_user_asmx(op.val, op.ea, err, "stdcx.", cr); emulate_step()
1859 err = read_mem(&regs->gpr[op.reg], op.ea, size, regs); emulate_step()
1873 err = do_fp_load(op.reg, do_lfs, op.ea, size, regs); emulate_step()
1875 err = do_fp_load(op.reg, do_lfd, op.ea, size, regs); emulate_step()
1882 err = do_vec_load(op.reg, do_lvx, op.ea & ~0xfUL, regs); emulate_step()
1889 err = do_vsx_load(op.reg, do_lxvd2x, op.ea, regs); emulate_step()
1900 err = read_mem(&regs->gpr[rd], op.ea, nb, regs); emulate_step()
1905 op.ea += 4; emulate_step()
1916 op.ea >= regs->gpr[1] - STACK_INT_FRAME_SIZE) { emulate_step()
1917 err = handle_stack_update(op.ea, regs); emulate_step()
1920 err = write_mem(op.val, op.ea, size, regs); emulate_step()
1928 err = do_fp_store(op.reg, do_stfs, op.ea, size, regs); emulate_step()
1930 err = do_fp_store(op.reg, do_stfd, op.ea, size, regs); emulate_step()
1937 err = do_vec_store(op.reg, do_stvx, op.ea & ~0xfUL, regs); emulate_step()
1944 err = do_vsx_store(op.reg, do_stxvd2x, op.ea, regs); emulate_step()
1958 err = write_mem(val, op.ea, nb, regs); emulate_step()
1961 op.ea += 4; emulate_step()
2010 regs->gpr[op.update_reg] = op.ea; emulate_step()
342 do_fp_load(int rn, int (*func)(int, unsigned long), unsigned long ea, int nb, struct pt_regs *regs) do_fp_load() argument
383 do_fp_store(int rn, int (*func)(int, unsigned long), unsigned long ea, int nb, struct pt_regs *regs) do_fp_store() argument
431 do_vec_load(int rn, int (*func)(int, unsigned long), unsigned long ea, struct pt_regs *regs) do_vec_load() argument
439 do_vec_store(int rn, int (*func)(int, unsigned long), unsigned long ea, struct pt_regs *regs) do_vec_store() argument
449 do_vsx_load(int rn, int (*func)(int, unsigned long), unsigned long ea, struct pt_regs *regs) do_vsx_load() argument
467 do_vsx_store(int rn, int (*func)(int, unsigned long), unsigned long ea, struct pt_regs *regs) do_vsx_store() argument
/linux-4.1.27/arch/nios2/mm/
H A Dextable.c18 fixup = search_exception_tables(regs->ea); fixup_exception()
20 regs->ea = fixup->fixup; fixup_exception()
H A Dfault.c55 regs->ea -= 4; do_page_fault()
87 if (!user_mode(regs) && !search_exception_tables(regs->ea)) do_page_fault()
211 pr_alert("ea = %08lx, ra = %08lx, cause = %ld\n", regs->ea, regs->ra, do_page_fault()
/linux-4.1.27/arch/nios2/kernel/
H A Dtraps.c125 fp->ea -= 4; breakpoint_c()
126 _exception(SIGTRAP, fp, TRAP_BRKPT, fp->ea); breakpoint_c()
136 fp->ea -= 4; handle_unaligned_c()
146 pr_alert(" op-code 0x%08lx\n", *(unsigned long *)(fp->ea)); handle_unaligned_c()
158 fp->ea -= 4; handle_illegal_c()
159 _exception(SIGILL, fp, ILL_ILLOPC, fp->ea); handle_illegal_c()
165 fp->ea -= 4; handle_supervisor_instr()
166 _exception(SIGILL, fp, ILL_PRVOPC, fp->ea); handle_supervisor_instr()
172 fp->ea -= 4; handle_diverror_c()
173 _exception(SIGFPE, fp, FPE_INTDIV, fp->ea); handle_diverror_c()
186 regs->ea -= 4; unhandled_exception()
189 pr_emerg("opcode: 0x%08lx\n", *(unsigned long *)(regs->ea)); unhandled_exception()
194 _send_sig(SIGUSR1, 0, fp->ea); handle_trap_1_c()
199 _send_sig(SIGUSR2, 0, fp->ea); handle_trap_2_c()
204 _send_sig(SIGILL, ILL_ILLTRP, fp->ea); handle_trap_3_c()
H A Dmisaligned.c78 fp->ea -= 4; handle_unaligned_c()
87 isn = *(unsigned long *)(fp->ea); handle_unaligned_c()
171 fp->ea, (unsigned int)addr, handle_unaligned_c()
175 fp->ea, handle_unaligned_c()
179 _exception(SIGSEGV, fp, SEGV_MAPERR, fp->ea); handle_unaligned_c()
190 fp->ea += 4; handle_unaligned_c()
194 fp->ea, handle_unaligned_c()
211 pr_err("user unaligned access @ %#lx; isn=0x%08lx ea=0x%08lx ra=0x%08lx sp=0x%08lx\n", handle_unaligned_c()
213 fp->ea, fp->ra, fp->sp); handle_unaligned_c()
217 _exception(SIGBUS, fp, BUS_ADRALN, fp->ea); handle_unaligned_c()
219 fp->ea += 4; /* else advance */ handle_unaligned_c()
H A Dkgdb.c62 { "ea", GDB_SIZEOF_REG, -1 },
65 { "pc", GDB_SIZEOF_REG, offsetof(struct pt_regs, ea) },
114 gdb_regs[GDB_PC] = p->thread.kregs->ea; sleeping_thread_to_gdb_regs()
119 regs->ea = pc; kgdb_arch_set_pc()
135 regs->ea = addr; kgdb_arch_handle_exception()
150 regs->ea -= 4; kgdb_breakpoint_c()
H A Dptrace.c53 REG_O_ONE(&regs->ea, PTR_EA); genregs_get()
56 REG_O_ONE(&regs->ea, PTR_PC); /* use ea for PC */ genregs_get()
99 REG_IN_ONE(&regs->ea, PTR_EA); genregs_set()
102 REG_IN_ONE(&regs->ea, PTR_PC); /* use ea for PC */ genregs_set()
H A Dsignal.c83 err |= __get_user(regs->ea, &gregs[27]); rt_restore_ucontext()
160 err |= __put_user(regs->ea, &gregs[27]); rt_setup_ucontext()
210 regs->ea = (unsigned long) ksig->ka.sa.sa_handler; setup_rt_frame()
244 continue_addr = regs->ea; do_signal()
261 regs->ea = restart_addr; do_signal()
268 if (unlikely(restart && regs->ea == restart_addr)) { do_signal()
275 regs->ea = continue_addr; do_signal()
285 if (unlikely(restart) && regs->ea == restart_addr) { do_signal()
286 regs->ea = continue_addr; do_signal()
H A Dentry.S51 * ea-4 = address of interrupted insn (ea must be preserved).
54 * If ea <= cmpxchg_stw and ea > cmpxchg_ldw then saved EA is set to
59 bgtu ea, et, 1f
62 bltu ea, et, 1f
64 mov ea, et
164 ldwio r24, -4(ea) /* instruction that caused the exception */
346 addi ea, ea, -4 /* re-issue the interrupted instruction */
347 stw ea, PT_EA(sp)
H A Dprocess.c91 pr_notice("ea: %08lx estatus: %08lx\n", show_regs()
92 regs->ea, regs->estatus); show_regs()
169 pr_emerg("PC: %08lx\n", fp->ea); dump()
188 tp = ((unsigned char *) fp->ea) - 0x20; dump()
247 regs->ea = pc; start_thread()
H A Dhead.S107 subi ea, ea, 4
H A Dasm-offsets.c57 OFFSET(PT_EA, pt_regs, ea); main()
H A Dinsnemu.S53 ldw ea, PT_EA(sp)
74 * the exception return address register (ea) as a destination because
88 * The instruction that causes the exception is found at address -4(ea).
158 stw ea, 116(sp)
166 ldw r2,-4(ea) /* r2 = AAAAA,BBBBB,IIIIIIIIIIIIIIII,PPPPPP */
586 ldw ea, 116(sp)
/linux-4.1.27/drivers/misc/cxl/
H A Dfault.c88 u64 ea) cxl_fault_segment()
93 if (!(rc = copro_calculate_slb(mm, ea, &slb))) { cxl_fault_segment()
116 struct mm_struct *mm, u64 ea) cxl_handle_segment_miss()
120 pr_devel("CXL interrupt: Segment fault pe: %i ea: %#llx\n", ctx->pe, ea); cxl_handle_segment_miss()
121 trace_cxl_ste_miss(ctx, ea); cxl_handle_segment_miss()
123 if ((rc = cxl_fault_segment(ctx, mm, ea))) cxl_handle_segment_miss()
222 static void cxl_prefault_one(struct cxl_context *ctx, u64 ea) cxl_prefault_one() argument
240 rc = cxl_fault_segment(ctx, mm, ea); cxl_prefault_one()
246 static u64 next_segment(u64 ea, u64 vsid) next_segment() argument
249 ea |= (1ULL << 40) - 1; next_segment()
251 ea |= (1ULL << 28) - 1; next_segment()
253 return ea + 1; next_segment()
258 u64 ea, last_esid = 0; cxl_prefault_vma() local
278 for (ea = vma->vm_start; ea < vma->vm_end; cxl_prefault_vma()
279 ea = next_segment(ea, slb.vsid)) { cxl_prefault_vma()
280 rc = copro_calculate_slb(mm, ea, &slb); cxl_prefault_vma()
87 cxl_fault_segment(struct cxl_context *ctx, struct mm_struct *mm, u64 ea) cxl_fault_segment() argument
115 cxl_handle_segment_miss(struct cxl_context *ctx, struct mm_struct *mm, u64 ea) cxl_handle_segment_miss() argument
/linux-4.1.27/arch/powerpc/include/asm/
H A Dcopro.h18 int copro_handle_mm_fault(struct mm_struct *mm, unsigned long ea,
21 int copro_calculate_slb(struct mm_struct *mm, u64 ea, struct copro_slb *slb);
H A Dmmu-hash64.h287 static inline unsigned long hpt_vpn(unsigned long ea, hpt_vpn() argument
294 return (vsid << (s_shift - VPN_SHIFT)) | ((ea >> VPN_SHIFT) & mask); hpt_vpn()
323 extern int __hash_page_4K(unsigned long ea, unsigned long access,
326 extern int __hash_page_64K(unsigned long ea, unsigned long access,
331 extern int hash_page_mm(struct mm_struct *mm, unsigned long ea,
334 extern int hash_page(unsigned long ea, unsigned long access, unsigned long trap,
336 int __hash_page_huge(unsigned long ea, unsigned long access, unsigned long vsid,
340 extern int __hash_page_thp(unsigned long ea, unsigned long access,
344 static inline int __hash_page_thp(unsigned long ea, unsigned long access, __hash_page_thp() argument
353 extern void hash_failure_debug(unsigned long ea, unsigned long access,
571 static inline unsigned long get_vsid(unsigned long context, unsigned long ea, get_vsid() argument
577 if ((ea & ~REGION_MASK) >= PGTABLE_RANGE) get_vsid()
582 | (ea >> SID_SHIFT), 256M); get_vsid()
584 | (ea >> SID_SHIFT_1T), 1T); get_vsid()
596 static inline unsigned long get_kernel_vsid(unsigned long ea, int ssize) get_kernel_vsid() argument
603 context = (MAX_USER_CONTEXT) + ((ea >> 60) - 0xc) + 1; get_kernel_vsid()
604 return get_vsid(context, ea, ssize); get_kernel_vsid()
H A Dsstep.h82 unsigned long ea; member in struct:instruction_op
H A Dkvm_ppc.h665 ulong ea; kvmppc_get_ea_indexed() local
668 ea = kvmppc_get_gpr(vcpu, rb); kvmppc_get_ea_indexed()
670 ea += kvmppc_get_gpr(vcpu, ra); kvmppc_get_ea_indexed()
679 ea = (uint32_t)ea; kvmppc_get_ea_indexed()
681 return ea; kvmppc_get_ea_indexed()
H A Dpci-bridge.h257 unsigned long ea = (unsigned long)address; isa_vaddr_is_ioport() local
258 return ea >= ISA_IO_BASE && ea < ISA_IO_END; isa_vaddr_is_ioport()
H A Dpgtable.h250 pte_t *__find_linux_pte_or_hugepte(pgd_t *pgdir, unsigned long ea,
252 static inline pte_t *find_linux_pte_or_hugepte(pgd_t *pgdir, unsigned long ea, find_linux_pte_or_hugepte() argument
259 return __find_linux_pte_or_hugepte(pgdir, ea, shift); find_linux_pte_or_hugepte()
H A Dhugetlb.h86 void book3e_hugetlb_preload(struct vm_area_struct *vma, unsigned long ea,
H A Dmachdep.h47 unsigned long ea,
57 void (*hpte_removebolted)(unsigned long ea,
H A Dpgtable-ppc64.h77 #define REGION_ID(ea) (((unsigned long)(ea)) >> REGION_SHIFT)
H A Dkvm_book3s.h117 extern void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong ea, ulong ea_mask);
H A Dio.h734 extern void __iomem * __ioremap_at(phys_addr_t pa, void *ea,
736 extern void __iounmap_at(void *ea, unsigned long size);
/linux-4.1.27/sound/pci/echoaudio/
H A Dmona.c66 MODULE_FIRMWARE("ea/loader_dsp.fw");
67 MODULE_FIRMWARE("ea/mona_301_dsp.fw");
68 MODULE_FIRMWARE("ea/mona_361_dsp.fw");
69 MODULE_FIRMWARE("ea/mona_301_1_asic_48.fw");
70 MODULE_FIRMWARE("ea/mona_301_1_asic_96.fw");
71 MODULE_FIRMWARE("ea/mona_361_1_asic_48.fw");
72 MODULE_FIRMWARE("ea/mona_361_1_asic_96.fw");
73 MODULE_FIRMWARE("ea/mona_2_asic.fw");
H A Dgina24.c68 MODULE_FIRMWARE("ea/loader_dsp.fw");
69 MODULE_FIRMWARE("ea/gina24_301_dsp.fw");
70 MODULE_FIRMWARE("ea/gina24_361_dsp.fw");
71 MODULE_FIRMWARE("ea/gina24_301_asic.fw");
72 MODULE_FIRMWARE("ea/gina24_361_asic.fw");
H A Dlayla24.c70 MODULE_FIRMWARE("ea/loader_dsp.fw");
71 MODULE_FIRMWARE("ea/layla24_dsp.fw");
72 MODULE_FIRMWARE("ea/layla24_1_asic.fw");
73 MODULE_FIRMWARE("ea/layla24_2A_asic.fw");
74 MODULE_FIRMWARE("ea/layla24_2S_asic.fw");
H A Decho3g.c70 MODULE_FIRMWARE("ea/loader_dsp.fw");
71 MODULE_FIRMWARE("ea/echo3g_dsp.fw");
72 MODULE_FIRMWARE("ea/3g_asic.fw");
H A Dindigo.c60 MODULE_FIRMWARE("ea/loader_dsp.fw");
61 MODULE_FIRMWARE("ea/indigo_dsp.fw");
H A Dindigodj.c60 MODULE_FIRMWARE("ea/loader_dsp.fw");
61 MODULE_FIRMWARE("ea/indigo_dj_dsp.fw");
H A Dindigodjx.c60 MODULE_FIRMWARE("ea/loader_dsp.fw");
61 MODULE_FIRMWARE("ea/indigo_djx_dsp.fw");
H A Dindigoio.c61 MODULE_FIRMWARE("ea/loader_dsp.fw");
62 MODULE_FIRMWARE("ea/indigo_io_dsp.fw");
H A Dindigoiox.c61 MODULE_FIRMWARE("ea/loader_dsp.fw");
62 MODULE_FIRMWARE("ea/indigo_iox_dsp.fw");
H A Dlayla20.c68 MODULE_FIRMWARE("ea/layla20_dsp.fw");
69 MODULE_FIRMWARE("ea/layla20_asic.fw");
H A Dmia.c69 MODULE_FIRMWARE("ea/loader_dsp.fw");
70 MODULE_FIRMWARE("ea/mia_dsp.fw");
H A Ddarla20.c58 MODULE_FIRMWARE("ea/darla20_dsp.fw");
H A Ddarla24.c62 MODULE_FIRMWARE("ea/darla24_dsp.fw");
H A Dgina20.c62 MODULE_FIRMWARE("ea/gina20_dsp.fw");
H A Dechoaudio_dsp.h658 /* Positions for ea. ch. 0x160 32*4 */
H A Dechoaudio.c61 snprintf(name, sizeof(name), "ea/%s", card_fw[fw_index].data); get_firmware()
/linux-4.1.27/fs/udf/
H A Dmisc.c50 uint8_t *ea = NULL, *ad = NULL; udf_add_extendedattr() local
55 ea = iinfo->i_ext.i_data; udf_add_extendedattr()
59 ad = ea; udf_add_extendedattr()
70 eahd = (struct extendedAttrHeaderDesc *)ea; udf_add_extendedattr()
107 memmove(&ea[offset - aal + size], udf_add_extendedattr()
108 &ea[aal], offset - aal); udf_add_extendedattr()
117 memmove(&ea[offset - ial + size], udf_add_extendedattr()
118 &ea[ial], offset - ial); udf_add_extendedattr()
128 memmove(&ea[offset - aal + size], udf_add_extendedattr()
129 &ea[aal], offset - aal); udf_add_extendedattr()
142 return (struct genericFormat *)&ea[offset]; udf_add_extendedattr()
154 uint8_t *ea = NULL; udf_get_extendedattr() local
158 ea = iinfo->i_ext.i_data; udf_get_extendedattr()
162 eahd = (struct extendedAttrHeaderDesc *)ea; udf_get_extendedattr()
179 gaf = (struct genericFormat *)&ea[offset]; udf_get_extendedattr()
H A Dnamei.c859 uint8_t *ea; udf_symlink() local
910 ea = epos.bh->b_data + udf_ext0_offset(inode); udf_symlink()
912 ea = iinfo->i_ext.i_data + iinfo->i_lenEAttr; udf_symlink()
915 pc = (struct pathComponent *)ea; udf_symlink()
934 pc = (struct pathComponent *)(ea + elen); udf_symlink()
/linux-4.1.27/arch/powerpc/platforms/cell/spufs/
H A Dfault.c37 unsigned long ea, int type) spufs_handle_event()
56 info.si_addr = (void __user *)ea; spufs_handle_event()
112 u64 ea, dsisr, access; spufs_handle_class1() local
126 ea = ctx->csa.class_1_dar; spufs_handle_class1()
134 pr_debug("ctx %p: ea %016llx, dsisr %016llx state %d\n", ctx, ea, spufs_handle_class1()
147 ret = hash_page(ea, access, 0x300, dsisr); spufs_handle_class1()
152 ret = copro_handle_mm_fault(current->mm, ea, dsisr, &flt); spufs_handle_class1()
187 spufs_handle_event(ctx, ea, SPE_EVENT_SPE_DATA_STORAGE); spufs_handle_class1()
36 spufs_handle_event(struct spu_context *ctx, unsigned long ea, int type) spufs_handle_event() argument
H A Dhw_ops.c296 out_be64(&prob->mfc_ea_W, cmd->ea); spu_hw_send_mfc_command()
H A Dspufs.h178 uint64_t ea; /* effective address */ member in struct:mfc_dma_command
H A Dfile.c1661 cmd->ea, cmd->size, cmd->tag, cmd->cmd); spufs_check_valid_dma()
1676 if ((cmd->lsa & 0xf) != (cmd->ea &0xf)) { spufs_check_valid_dma()
1677 pr_debug("invalid DMA alignment, ea %llx lsa %x\n", spufs_check_valid_dma()
1678 cmd->ea, cmd->lsa); spufs_check_valid_dma()
H A Dswitch.c778 static inline int send_mfc_dma(struct spu *spu, unsigned long ea, send_mfc_dma() argument
797 out_be64(&prob->mfc_ea_W, ea); send_mfc_dma()
806 ea += transfer_size; send_mfc_dma()
/linux-4.1.27/arch/mips/sgi-ip32/
H A Dip32-setup.c49 static inline void str2eaddr(unsigned char *ea, unsigned char *str) str2eaddr() argument
60 ea[i] = num; str2eaddr()
/linux-4.1.27/arch/powerpc/kvm/
H A De500_emulate.c123 gva_t ea; kvmppc_core_emulate_op_e500() local
152 ea = kvmppc_get_ea_indexed(vcpu, ra, rb); kvmppc_core_emulate_op_e500()
153 emulated = kvmppc_e500_emul_tlbsx(vcpu, ea); kvmppc_core_emulate_op_e500()
158 ea = kvmppc_get_ea_indexed(vcpu, ra, rb); kvmppc_core_emulate_op_e500()
159 emulated = kvmppc_e500_emul_tlbilx(vcpu, type, ea); kvmppc_core_emulate_op_e500()
164 ea = kvmppc_get_ea_indexed(vcpu, ra, rb); kvmppc_core_emulate_op_e500()
165 emulated = kvmppc_e500_emul_tlbivax(vcpu, ea); kvmppc_core_emulate_op_e500()
H A Dtrace_hv.h275 struct kvm_memory_slot *memslot, unsigned long ea,
278 TP_ARGS(vcpu, hptep, memslot, ea, dsisr),
285 __field(unsigned long, ea)
296 __entry->ea = ea;
302 TP_printk("VCPU %d: hpte=0x%lx:0x%lx guest=0x%lx ea=0x%lx,%x slot=0x%llx,0x%x",
305 __entry->ea, __entry->dsisr,
H A De500_mmu.c248 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea) kvmppc_e500_emul_tlbivax() argument
254 ia = (ea >> 2) & 0x1; kvmppc_e500_emul_tlbivax()
257 tlbsel = (ea >> 3) & 0x1; kvmppc_e500_emul_tlbivax()
265 ea &= 0xfffff000; kvmppc_e500_emul_tlbivax()
266 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel, kvmppc_e500_emul_tlbivax()
296 gva_t ea) tlbilx_one()
301 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel, pid, -1); tlbilx_one()
310 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea) kvmppc_e500_emul_tlbilx() argument
319 tlbilx_one(vcpu_e500, pid, ea); kvmppc_e500_emul_tlbilx()
344 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea) kvmppc_e500_emul_tlbsx() argument
353 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel, pid, as); kvmppc_e500_emul_tlbsx()
295 tlbilx_one(struct kvmppc_vcpu_e500 *vcpu_e500, int pid, gva_t ea) tlbilx_one() argument
H A Dbook3s_64_mmu.c454 static void kvmppc_mmu_book3s_64_slbie(struct kvm_vcpu *vcpu, u64 ea) kvmppc_mmu_book3s_64_slbie() argument
459 dprintk("KVM MMU: slbie(0x%llx)\n", ea); kvmppc_mmu_book3s_64_slbie()
461 slbe = kvmppc_mmu_book3s_64_find_slbe(vcpu, ea); kvmppc_mmu_book3s_64_slbie()
466 dprintk("KVM MMU: slbie(0x%llx, 0x%llx)\n", ea, slbe->esid); kvmppc_mmu_book3s_64_slbie()
473 kvmppc_mmu_flush_segment(vcpu, ea & ~(seg_size - 1), seg_size); kvmppc_mmu_book3s_64_slbie()
580 ulong ea = esid << SID_SHIFT; kvmppc_mmu_book3s_64_esid_to_vsid() local
588 slb = kvmppc_mmu_book3s_64_find_slbe(vcpu, ea); kvmppc_mmu_book3s_64_esid_to_vsid()
H A Dbook3s_32_mmu.c353 static void kvmppc_mmu_book3s_32_tlbie(struct kvm_vcpu *vcpu, ulong ea, bool large) kvmppc_mmu_book3s_32_tlbie() argument
360 kvmppc_mmu_pte_flush(v, ea, 0x0FFFF000); kvmppc_mmu_book3s_32_tlbie()
366 ulong ea = esid << SID_SHIFT; kvmppc_mmu_book3s_32_esid_to_vsid() local
372 sr = find_sr(vcpu, ea); kvmppc_mmu_book3s_32_esid_to_vsid()
H A De500.h132 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea);
133 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea);
134 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea);
H A Dbook3s_64_mmu_host.c116 /* and write the mapping ea -> hpa into the pt */ kvmppc_mmu_map_page()
354 void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong ea, ulong seg_size) kvmppc_mmu_flush_segment() argument
362 (svcpu->slb[i].esid & seg_mask) == ea) { kvmppc_mmu_flush_segment()
H A Dtrace_pr.h88 TP_printk("Map: hvpn=%llx pfn=%llx ea=%lx vp=%llx ra=%lx [%x]",
117 TP_printk("Flush: hva=%llx pfn=%llx ea=%lx vp=%llx ra=%lx [%x]",
H A Dbook3s_64_mmu_hv.c299 unsigned long ea) kvmppc_mmu_get_real_addr()
304 return (r & HPTE_R_RPN & ~ra_mask) | (ea & ra_mask); kvmppc_mmu_get_real_addr()
391 unsigned long gpa, gva_t ea, int is_store) kvmppc_hv_emulate_mmio()
431 vcpu->arch.vaddr_accessed = ea; kvmppc_hv_emulate_mmio()
436 unsigned long ea, unsigned long dsisr) kvmppc_book3s_hv_page_fault()
460 if (ea != vcpu->arch.pgfault_addr) kvmppc_book3s_hv_page_fault()
482 gpa = gpa_base | (ea & (psize - 1)); kvmppc_book3s_hv_page_fault()
486 trace_kvm_page_fault_enter(vcpu, hpte, memslot, ea, dsisr); kvmppc_book3s_hv_page_fault()
490 return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, kvmppc_book3s_hv_page_fault()
298 kvmppc_mmu_get_real_addr(unsigned long v, unsigned long r, unsigned long ea) kvmppc_mmu_get_real_addr() argument
390 kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, unsigned long gpa, gva_t ea, int is_store) kvmppc_hv_emulate_mmio() argument
435 kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu, unsigned long ea, unsigned long dsisr) kvmppc_book3s_hv_page_fault() argument
H A Dbook3s_32_mmu_host.c168 /* and write the mapping ea -> hpa into the pt */ kvmppc_mmu_map_page()
/linux-4.1.27/arch/nios2/include/asm/
H A Dprocessor.h84 #define thread_saved_pc(tsk) ((tsk)->thread.kregs->ea)
92 #define KSTK_EIP(tsk) ((tsk)->thread.kregs->ea)
H A Dentry.h61 stw ea, PT_EA(sp) variable
85 ldw ea, PT_EA(sp) variable
H A Dptrace.h44 unsigned long ea; /* Exception return address (pc) */ member in struct:pt_regs
H A Delf.h72 pr_reg[21] = regs->ea; \
/linux-4.1.27/arch/m68k/fpsp040/
H A Dx_unsupp.S16 | if fmove.p fpm,<ea> then pack source (fpm)
17 | and store in users memory <ea>.
H A Dx_fline.S9 | <ea> field. If so, handle here and return. Otherwise, it
42 | fmovecr with non-zero <ea> handling here
H A Dx_snan.S144 movel EXC_EA(%a6),%a0 |get <ea> from exc frame
173 tstl %a0 |check if <ea> is 0
195 tstl %a0 |check if <ea> is 0
217 tstl %a0 |check if <ea> is 0
H A Dgen_except.S95 | Frame is an unimp frame possible resulting from an fmove <ea>,fp0
325 | by doing an fmove <ea>,fp0. The exception operand, in
H A Dx_operr.S272 | Destination is memory. Get <ea> from integer exception frame
H A Dres_func.S10 | For packed move out (fmove.p fpm,<ea>) the operation is
/linux-4.1.27/arch/m68k/math-emu/
H A Dfp_decode.h91 | .long "f<op> <ea>,fpx","fmove fpx,<ea>"
92 | .long "fmovem <ea>,fpcr","fmovem <ea>,fpx"
93 | .long "fmovem fpcr,<ea>","fmovem fpx,<ea>"
102 | decode destination format for fmove reg,ea
107 | decode source register for fmove reg,ea
H A Dfp_scan.S104 fp_fgen_ea: | source is <ea>
/linux-4.1.27/arch/powerpc/platforms/cell/
H A Dspu_base.c164 static int __spu_trap_data_seg(struct spu *spu, unsigned long ea) __spu_trap_data_seg() argument
169 ret = copro_calculate_slb(spu->mm, ea, &slb); __spu_trap_data_seg()
184 extern int hash_page(unsigned long ea, unsigned long access,
186 static int __spu_trap_data_map(struct spu *spu, unsigned long ea, u64 dsisr) __spu_trap_data_map() argument
190 pr_debug("%s, %llx, %lx\n", __func__, dsisr, ea); __spu_trap_data_map()
197 (REGION_ID(ea) != USER_REGION_ID)) { __spu_trap_data_map()
200 ret = hash_page(ea, _PAGE_PRESENT, 0x300, dsisr); __spu_trap_data_map()
209 spu->class_1_dar = ea; __spu_trap_data_map()
222 unsigned long ea = (unsigned long)addr; __spu_kernel_slb() local
225 if (REGION_ID(ea) == KERNEL_REGION_ID) __spu_kernel_slb()
230 slb->vsid = (get_kernel_vsid(ea, MMU_SEGSIZE_256M) << SLB_VSID_SHIFT) | __spu_kernel_slb()
232 slb->esid = (ea & ESID_MASK) | SLB_ESID_V; __spu_kernel_slb()
242 unsigned long ea = (unsigned long)new_addr; __slb_present() local
246 if (!((slbs[i].esid ^ ea) & ESID_MASK)) __slb_present()
/linux-4.1.27/net/appletalk/
H A Daarp.c722 struct elapaarp *ea = aarp_hdr(skb); aarp_rcv() local
737 if (!skb_pull(skb, sizeof(*ea))) aarp_rcv()
740 function = ntohs(ea->function); aarp_rcv()
744 ea->hw_len != ETH_ALEN || ea->pa_len != AARP_PA_ALEN || aarp_rcv()
745 ea->pa_src_zero || ea->pa_dst_zero) aarp_rcv()
749 hash = ea->pa_src_node % (AARP_HASH_SIZE - 1); aarp_rcv()
752 sa.s_node = ea->pa_src_node; aarp_rcv()
753 sa.s_net = ea->pa_src_net; aarp_rcv()
761 ifa->address.s_node == ea->pa_dst_node && aarp_rcv()
762 ifa->address.s_net == ea->pa_dst_net) { aarp_rcv()
768 da.s_node = ea->pa_dst_node; aarp_rcv()
769 da.s_net = ea->pa_dst_net; aarp_rcv()
794 ether_addr_copy(a->hwaddr, ea->hw_src); aarp_rcv()
816 sa.s_node = ea->pa_dst_node; aarp_rcv()
817 sa.s_net = ea->pa_dst_net; aarp_rcv()
858 sa.s_node = ea->pa_src_node; aarp_rcv()
859 sa.s_net = ea->pa_src_net; aarp_rcv()
863 aarp_send_reply(dev, ma, &sa, ea->hw_src); aarp_rcv()
/linux-4.1.27/arch/mips/alchemy/common/
H A Dprom.c92 static inline void str2eaddr(unsigned char *ea, unsigned char *str) str2eaddr() argument
103 ea[i] = num; str2eaddr()
/linux-4.1.27/arch/powerpc/crypto/
H A Daes-tab-4k.S51 .long R(5f, a2, a2, fd), R(45, af, af, ea)
69 .long R(7f, b2, b2, cd), R(ea, 75, 75, 9f)
131 .long R(f3, f4, f4, 07), R(cf, ea, ea, 25)
178 .long R(45, ea, 0e, 98), R(5d, fe, c0, e1)
266 .long R(ea, 9f, 09, d4), R(29, b0, 7c, d6)
277 .long R(9d, 5e, ea, 04), R(01, 8c, 35, 5d)
287 .long R(53, f7, cd, ea), R(5f, fd, aa, 5b)
/linux-4.1.27/arch/openrisc/kernel/
H A Dtraps.c173 unsigned long ea, unsigned long vector) nommu_dump_state()
178 printk("\n\r[nommu_dump_state] :: ea %lx, vector %lx\n\r", ea, vector); nommu_dump_state()
276 void unhandled_exception(struct pt_regs *regs, int ea, int vector) unhandled_exception() argument
279 ea, vector); unhandled_exception()
172 nommu_dump_state(struct pt_regs *regs, unsigned long ea, unsigned long vector) nommu_dump_state() argument
/linux-4.1.27/drivers/media/tuners/
H A Dqt1010_priv.h72 2c ea ?
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmfmac/
H A Dfwil_types.h178 * @ea: ether address of peer station.
184 u8 ea[ETH_ALEN]; /* Station address */ member in struct:brcmf_tdls_iovar_le
420 u8 ea[ETH_ALEN]; /* per station */ member in struct:brcmf_wsec_key
442 u8 ea[ETH_ALEN]; /* per station */ member in struct:brcmf_wsec_key_le
448 u8 ea[ETH_ALEN]; member in struct:brcmf_scb_val_le
464 u8 ea[ETH_ALEN]; /* Station address */ member in struct:brcmf_sta_info_le
H A Dfwsignal.c377 * @ea: ethernet address.
393 u8 ea[ETH_ALEN]; member in struct:brcmf_fws_mac_descriptor
717 "enter: desc %p ea=%pM, ifidx=%u\n", desc, addr, ifidx); brcmf_fws_macdesc_init()
726 memcpy(&desc->ea[0], addr, ETH_ALEN); brcmf_fws_macdesc_init()
733 "enter: ea=%pM, ifidx=%u\n", desc->ea, desc->interface_id); brcmf_fws_macdesc_deinit()
741 brcmf_fws_macdesc_lookup(struct brcmf_fws_info *fws, u8 *ea) brcmf_fws_macdesc_lookup() argument
746 if (ea == NULL) brcmf_fws_macdesc_lookup()
751 if (entry->occupied && !memcmp(entry->ea, ea, ETH_ALEN)) brcmf_fws_macdesc_lookup()
H A Dcfg80211.c452 memcpy(key_le->ea, key->ea, sizeof(key->ea)); convert_key_from_CPU()
1933 memcpy(&scbval.ea, &profile->bssid, ETH_ALEN); brcmf_cfg80211_disconnect()
2066 /* Instead of bcast for ea address for default wep keys, brcmf_add_keyext()
2069 memcpy((char *)&key.ea, (void *)mac_addr, ETH_ALEN); brcmf_add_keyext()
4272 memcpy(&scbval.ea, params->mac, ETH_ALEN); brcmf_cfg80211_del_station()
4547 memcpy(info.ea, peer, ETH_ALEN); brcmf_cfg80211_tdls_oper()
H A Dp2p.c2110 struct brcmf_if *ifp, u8 ea[ETH_ALEN], brcmf_p2p_request_p2p_if()
2121 memcpy(if_request.addr, ea, ETH_ALEN); brcmf_p2p_request_p2p_if()
2109 brcmf_p2p_request_p2p_if(struct brcmf_p2p_info *p2p, struct brcmf_if *ifp, u8 ea[ETH_ALEN], enum brcmf_fil_p2p_if_types iftype) brcmf_p2p_request_p2p_if() argument
/linux-4.1.27/arch/powerpc/platforms/pseries/
H A Dlpar.c362 unsigned long ea, pSeries_lpar_hpte_updateboltedpp()
368 vsid = get_kernel_vsid(ea, ssize); pSeries_lpar_hpte_updateboltedpp()
369 vpn = hpt_vpn(ea, vsid, ssize); pSeries_lpar_hpte_updateboltedpp()
498 static void pSeries_lpar_hpte_removebolted(unsigned long ea, pSeries_lpar_hpte_removebolted() argument
504 vsid = get_kernel_vsid(ea, ssize); pSeries_lpar_hpte_removebolted()
505 vpn = hpt_vpn(ea, vsid, ssize); pSeries_lpar_hpte_removebolted()
361 pSeries_lpar_hpte_updateboltedpp(unsigned long newpp, unsigned long ea, int psize, int ssize) pSeries_lpar_hpte_updateboltedpp() argument
/linux-4.1.27/arch/m68k/ifpsp060/src/
H A Disp.S390 # access these register files for things like data retrieval and <ea> #
433 # muls.l 0100 1100 00 |<ea>| 0*** 1100 0000 0*** #
434 # mulu.l 0100 1100 00 |<ea>| 0*** 0100 0000 0*** #
436 # divs.l 0100 1100 01 |<ea>| 0*** 1100 0000 0*** #
437 # divu.l 0100 1100 01 |<ea>| 0*** 0100 0000 0*** #
444 # cas.w 0000 1100 11 |<ea>| 0000 000* **00 0*** #
445 # cas.l 0000 1110 11 |<ea>| 0000 000* **00 0*** #
452 # chk2.b 0000 0000 11 |<ea>| **** 1000 0000 0000 #
453 # chk2.w 0000 0010 11 |<ea>| **** 1000 0000 0000 #
454 # chk2.l 0000 0100 11 |<ea>| **** 1000 0000 0000 #
456 # cmp2.b 0000 0000 11 |<ea>| **** 0000 0000 0000 #
457 # cmp2.w 0000 0010 11 |<ea>| **** 0000 0000 0000 #
458 # cmp2.l 0000 0100 11 |<ea>| **** 0000 0000 0000 #
925 jmp (tbl_ea_mode.b,%pc,%d0.w*1) # jmp to correct ea mode
1372 # word, long: <ea> of the data is the current extension word #
1393 mov.w %d0,%a0 # return <ea> in a0
1407 mov.l %d0,%a0 # return <ea> in a0
1426 subq.l &0x2,%a0 # adjust <ea>
1579 add.l %d2,%d0 # <ea> += index
1580 add.l %d4,%d0 # <ea> += od
1591 add.l %d4,%d0 # ea += od
1595 add.l %d2,%d3 # ea = (base + bd) + index
1895 bsr.l _calc_ea # calculate <ea>
1910 mov.l %a0,%a2 # save copy of <ea>
1918 mov.l %a2,%a0 # pass <ea> of long upper bound
2039 # _div64(): routine to emulate div{u,s}.l <ea>,Dr:Dq #
2431 bsr.l _calc_ea # calc <ea>
2437 bsr.l _dmem_read_long # fetch divisor from <ea>
2478 # _mul64(): routine to emulate mul{u,s}.l <ea>,Dh:Dl 32x32->64 #
2662 # must calculate the <ea> and go fetch the 32-bit operand.
2665 bsr.l _calc_ea # calculate <ea>
2918 # _compandset(): routine to emulate cas w/ misaligned <ea> #
3010 bsr.l _calc_ea # a0 = calculated <ea>
3011 mov.l %a0,ADDR(%a6) # save <ea> for possible restart
3017 bsr.l _calc_ea # a0 = calculated <ea>
3018 mov.l %a0,ADDR(%a6) # save <ea> for possible restart
3104 mov.l ADDR(%a6),%a0 # load <ea>
H A Dpfpsp.S300 set EXC_EA, 0xc # stacked <ea>
546 set mia7_flg, 0x04 # flag bit: (a7)+ <ea>
547 set mda7_flg, 0x08 # flag bit: -(a7) <ea>
549 set immed_flg, 0x80 # flag bit: &<data> <ea>
1221 # stack an ea. however, we don't need an a7' for this case anyways.
1580 # is the ea mode pre-decrement of the stack pointer from supervisor mode?
1751 cmpi.b SPCOND_FLG(%a6),&mda7_flg # was the <ea> mode -(sp)?
1904 btst &mia7_bit,SPCOND_FLG(%a6) # was ea mode (a7)+
1913 # shift the stack frame "up". we don't really care about the <ea> field.
2027 # shift stack frame "up". who cares about <ea> field.
2135 btst &mda7_bit,SPCOND_FLG(%a6) # was ea mode -(a7)
2388 # - The system stack contains the "Unimplemented <ea>" stk frame #
2435 # see if the FPU is disabled. Since the "Unimp <ea>" exception is taken #
2982 # call fmovm_calc_ea which will go through the ea calc process and,
2998 # the <ea> field is let as undefined.
3029 mov.l 0x2(%sp),0x8(%sp) # store ea
3190 mov.b 1+EXC_OPWORD(%a6),%d1 # extract <ea> mode,reg
3206 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3208 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3222 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3224 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3238 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3240 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3266 # _calc_ea_fout() - fix An if <ea> is -() or ()+; also get <ea> #
3294 # register must get updated by calling _calc_ea_fout(). If the <ea> #
3360 mov.b 1+EXC_OPWORD(%a6),%d1 # extract <ea> mode,reg
3377 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3379 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3394 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3396 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3411 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3413 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3426 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3434 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3482 # in addition, for pre-decrement, the stacked <ea> is incorrect.
3500 bsr.l _calc_ea_fout # find the correct ea,update An
3502 mov.l %a0,EXC_EA(%a6) # stack correct <ea>
3521 bsr.l _calc_ea_fout # find the correct ea,update An
3523 mov.l %a0,EXC_EA(%a6) # stack correct <ea>
3527 cmpi.b SPCOND_FLG(%a6),&mda7_flg # is <ea> mode -(a7)?
3848 # _dcalc_ea(): calc correct <ea> from <ea> stacked on exception #
3855 # d0 = number of bytes to adjust <ea> by #
3862 # The stacked <ea> for FP unimplemented instructions and opclass #
3867 # stacked <ea> value is 8 bytes too big #
3896 mov.l EXC_EA(%a6),%a0 # return <ea>
3903 lea ([USER_FPIAR,%a6],0x4),%a0 # no; return <ea>
3906 # here, the <ea> is stacked correctly. however, we must update the
3912 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
3915 # the <ea> is stacked correctly for all but extended and packed which
3916 # the <ea>s are 8 bytes too large.
3923 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
3929 sub.l &0x8,%a0 # correct <ea>
3930 mov.l %a0,EXC_EA(%a6) # put correct <ea> on stack
3935 # _calc_ea_fout(): calculate correct stacked <ea> for extended #
3948 # For opclass 3 extended and packed data operations, the <ea> #
3952 # So, for -(an), we must subtract 8 off of the stacked <ea> value #
3953 # and return that value as the correct <ea> and store that value in An. #
3954 # For (an)+, the stacked <ea> is correct but we must adjust An by +12. #
3958 # This calc_ea is currently used to retrieve the correct <ea>
3974 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
3978 # : stacked <ea> is correct
4023 # : stacked <ea> = actual <ea> + 8
4234 # If the instruction is "fmovm.x <ea>,Dn", a _mem_read() is used #
4237 # If the instruction is "fmovm.x DN,<ea>", a _mem_write() is used #
4246 # fmovm.x dn, <ea> #
4247 # fmovm.x <ea>, dn #
4250 # 1111 0010 00 |<ea>| 11@& 1000 0$$$ 0000 #
4280 bsr.l fmovm_calc_ea # calculate <ea>
4285 # but, make sure that we've calculated ea and advanced the opword pointer
4310 cmpi.b SPCOND_FLG(%a6),&mda7_flg # is <ea> mode -(a7)?
4320 mov.l %a0,%a1 # move <ea> to a1
4585 jmp (tbl_fea_mode.b,%pc,%d0.w*1) # jmp to correct ea mode
4996 mov.w %d0,%a0 # return <ea> in a0
5010 mov.l %d0,%a0 # return <ea> in a0
5029 subq.l &0x2,%a0 # adjust <ea>
5198 add.l %d2,%d0 # <ea> += index
5199 add.l %d4,%d0 # <ea> += od
5210 add.l %d4,%d0 # ea += od
5214 add.l %d2,%d3 # ea = (base + bd) + index
7086 # For extended precision, the stacked <ea> must be fixed along #
7091 # word or a data register. The <ea> must be fixed as w/ extended #
7142 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
7188 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
7235 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
7275 bsr.l _calc_ea_fout # fix stacked <ea>
7376 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
7423 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
7473 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
7856 bsr.l _calc_ea_fout # fetch the <ea>
12559 # Typically used for an instruction w/ a post-increment <ea>, #
12623 # Typically used for an instruction w/ a pre-decrement <ea>, #
12893 # _dcalc_ea() - calculate the correct <ea> #
12906 # Get the correct <ea> which is the value on the exception stack #
12907 # frame w/ maybe a correction factor if the <ea> is -(an) or (an)+. #
12916 # the stacked <ea> for packed is correct except for -(An).
12921 bsr.l _dcalc_ea # fetch <ea>; correct An
14742 # need to invert adjustment value if the <ea> was predec
H A Dilsp.S755 # _060LSP__cmp2_Ab_(): Emulate "cmp2.b An,<ea>". #
756 # _060LSP__cmp2_Aw_(): Emulate "cmp2.w An,<ea>". #
757 # _060LSP__cmp2_Al_(): Emulate "cmp2.l An,<ea>". #
758 # _060LSP__cmp2_Db_(): Emulate "cmp2.b Dn,<ea>". #
759 # _060LSP__cmp2_Dw_(): Emulate "cmp2.w Dn,<ea>". #
760 # _060LSP__cmp2_Dl_(): Emulate "cmp2.l Dn,<ea>". #
H A Dfpsp.S301 set EXC_EA, 0xc # stacked <ea>
547 set mia7_flg, 0x04 # flag bit: (a7)+ <ea>
548 set mda7_flg, 0x08 # flag bit: -(a7) <ea>
550 set immed_flg, 0x80 # flag bit: &<data> <ea>
1222 # stack an ea. however, we don't need an a7' for this case anyways.
1581 # is the ea mode pre-decrement of the stack pointer from supervisor mode?
1752 cmpi.b SPCOND_FLG(%a6),&mda7_flg # was the <ea> mode -(sp)?
1905 btst &mia7_bit,SPCOND_FLG(%a6) # was ea mode (a7)+
1914 # shift the stack frame "up". we don't really care about the <ea> field.
2028 # shift stack frame "up". who cares about <ea> field.
2136 btst &mda7_bit,SPCOND_FLG(%a6) # was ea mode -(a7)
2389 # - The system stack contains the "Unimplemented <ea>" stk frame #
2436 # see if the FPU is disabled. Since the "Unimp <ea>" exception is taken #
2983 # call fmovm_calc_ea which will go through the ea calc process and,
2999 # the <ea> field is let as undefined.
3030 mov.l 0x2(%sp),0x8(%sp) # store ea
3191 mov.b 1+EXC_OPWORD(%a6),%d1 # extract <ea> mode,reg
3207 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3209 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3223 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3225 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3239 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3241 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3267 # _calc_ea_fout() - fix An if <ea> is -() or ()+; also get <ea> #
3295 # register must get updated by calling _calc_ea_fout(). If the <ea> #
3361 mov.b 1+EXC_OPWORD(%a6),%d1 # extract <ea> mode,reg
3378 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3380 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3395 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3397 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3412 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3414 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3427 cmpi.b %d1,&0x7 # is <ea> mode a data reg?
3435 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result
3483 # in addition, for pre-decrement, the stacked <ea> is incorrect.
3501 bsr.l _calc_ea_fout # find the correct ea,update An
3503 mov.l %a0,EXC_EA(%a6) # stack correct <ea>
3522 bsr.l _calc_ea_fout # find the correct ea,update An
3524 mov.l %a0,EXC_EA(%a6) # stack correct <ea>
3528 cmpi.b SPCOND_FLG(%a6),&mda7_flg # is <ea> mode -(a7)?
3837 # non-zero <ea> field. These may get flagged as "Line F" but should #
3857 # if the F-Line instruction is an "fmovecr" w/ a non-zero <ea>. if
3878 # it's an fmovecr w/ a non-zero <ea> that has entered through
3894 sub.l &0x8,%sp # make room for "Next PC", <ea>
3999 # * * => <ea> of fp unimp instr.
4110 cmpi.b SPCOND_FLG(%a6),&mia7_flg # was the ea mode (sp)+ ?
4113 cmpi.b SPCOND_FLG(%a6),&mda7_flg # was the ea mode -(sp) ?
4332 # fs<cc>.b <ea> #
17431 # only difference is that the <ea> field should hold the PC
18077 # the stacked <ea> is correct with the exception of:
18078 # -> Dn : <ea> is garbage
18084 cmpi.b %d1,&0x18 # is <ea> (An)+ ?
18086 cmpi.b %d1,&0x20 # is <ea> -(An) ?
18090 mov.l EXC_EA(%a6),%a0 # fetch <ea>
18103 mov.l EXC_EA(%a6),%a0 # fetch <ea>
18121 mov.l EXC_EA(%a6),%a0 # fetch <ea>
18185 # If the instruction is "fmovm.x <ea>,Dn", a _mem_read() is used #
18188 # If the instruction is "fmovm.x DN,<ea>", a _mem_write() is used #
18197 # fmovm.x dn, <ea> #
18198 # fmovm.x <ea>, dn #
18201 # 1111 0010 00 |<ea>| 11@& 1000 0$$$ 0000 #
18231 bsr.l fmovm_calc_ea # calculate <ea>
18236 # but, make sure that we've calculated ea and advanced the opword pointer
18261 cmpi.b SPCOND_FLG(%a6),&mda7_flg # is <ea> mode -(a7)?
18271 mov.l %a0,%a1 # move <ea> to a1
18536 jmp (tbl_fea_mode.b,%pc,%d0.w*1) # jmp to correct ea mode
18947 mov.w %d0,%a0 # return <ea> in a0
18961 mov.l %d0,%a0 # return <ea> in a0
18980 subq.l &0x2,%a0 # adjust <ea>
19149 add.l %d2,%d0 # <ea> += index
19150 add.l %d4,%d0 # <ea> += od
19161 add.l %d4,%d0 # ea += od
19165 add.l %d2,%d3 # ea = (base + bd) + index
19335 # _dcalc_ea(): calc correct <ea> from <ea> stacked on exception #
19342 # d0 = number of bytes to adjust <ea> by #
19349 # The stacked <ea> for FP unimplemented instructions and opclass #
19354 # stacked <ea> value is 8 bytes too big #
19383 mov.l EXC_EA(%a6),%a0 # return <ea>
19390 lea ([USER_FPIAR,%a6],0x4),%a0 # no; return <ea>
19393 # here, the <ea> is stacked correctly. however, we must update the
19399 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
19402 # the <ea> is stacked correctly for all but extended and packed which
19403 # the <ea>s are 8 bytes too large.
19410 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
19416 sub.l &0x8,%a0 # correct <ea>
19417 mov.l %a0,EXC_EA(%a6) # put correct <ea> on stack
19422 # _calc_ea_fout(): calculate correct stacked <ea> for extended #
19435 # For opclass 3 extended and packed data operations, the <ea> #
19439 # So, for -(an), we must subtract 8 off of the stacked <ea> value #
19440 # and return that value as the correct <ea> and store that value in An. #
19441 # For (an)+, the stacked <ea> is correct but we must adjust An by +12. #
19445 # This calc_ea is currently used to retrieve the correct <ea>
19461 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
19465 # : stacked <ea> is correct
19510 # : stacked <ea> = actual <ea> + 8
19567 # _dcalc_ea() - calculate <ea>, fixing An in process #
19618 # beq.w op010 # handle <ea> -> fpn
19619 # bgt.w op011 # handle fpn -> <ea>
19669 # OPCLASS '010: <ea> -> reg #
19695 bfextu EXC_OPWORD(%a6){&10:&3}, %d1 # extract <ea> mode field
19788 # (1) calc ea - must read AFTER you know the src type since #
19789 # if the ea is -() or ()+, need to know # of bytes. #
19817 # (1) calc ea #
19823 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0
19853 # (1) calc ea #
19859 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0
19889 # (1) calc ea #
19895 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0
19925 # (1) calc ea #
19931 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0
20007 # (1) calc ea #
20013 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0
20096 # (1) calc ea #
20102 bsr.l _dcalc_ea # calc <ea>
20128 # (1) calc ea #
20185 # For extended precision, the stacked <ea> must be fixed along #
20190 # word or a data register. The <ea> must be fixed as w/ extended #
20241 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
20287 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
20334 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
20374 bsr.l _calc_ea_fout # fix stacked <ea>
20475 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
20522 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
20572 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct
20955 bsr.l _calc_ea_fout # fetch the <ea>
21343 # Typically used for an instruction w/ a post-increment <ea>, #
21407 # Typically used for an instruction w/ a pre-decrement <ea>, #
22933 # _dcalc_ea() - calculate the correct <ea> #
22946 # Get the correct <ea> which is the value on the exception stack #
22947 # frame w/ maybe a correction factor if the <ea> is -(an) or (an)+. #
22956 # the stacked <ea> for packed is correct except for -(An).
22961 bsr.l _dcalc_ea # fetch <ea>; correct An
24782 # need to invert adjustment value if the <ea> was predec
H A Ditest.S120 ### ea
1285 cas.w %d1,%d2,(%a0) # Dc,Du,<ea>
1312 cas.w %d1,%d2,(%a0) # Dc,Du,<ea>
1340 cas.l %d1,%d2,(%a0) # Dc,Du,<ea>
1367 cas.l %d1,%d2,(%a0) # Dc,Du,<ea>
1395 cas.l %d1,%d2,(%a0) # Dc,Du,<ea>
1422 cas.l %d1,%d2,(%a0) # Dc,Du,<ea>
H A Dftest.S467 string "\tUnimplemented <ea>..."
H A Dfplsp.S281 set EXC_EA, 0xc # stacked <ea>
527 set mia7_flg, 0x04 # flag bit: (a7)+ <ea>
528 set mda7_flg, 0x08 # flag bit: -(a7) <ea>
530 set immed_flg, 0x80 # flag bit: &<data> <ea>
/linux-4.1.27/arch/sh/mm/
H A Dcache-sh4.c323 unsigned long a, ea, p; __flush_cache_one() local
353 ea = base_addr + PAGE_SIZE; __flush_cache_one()
367 } while (a < ea); __flush_cache_one()
/linux-4.1.27/arch/mips/pmcs-msp71xx/
H A Dmsp_prom.c86 int str2eaddr(unsigned char *ea, unsigned char *str) str2eaddr() argument
93 ea[index++] = num; str2eaddr()
103 ea[index++] = num; str2eaddr()
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/
H A Dscb.h69 u8 ea[ETH_ALEN]; /* station address */ member in struct:scb
/linux-4.1.27/arch/unicore32/include/asm/
H A Dassembler.h69 .else; .ifc \cond, ea
/linux-4.1.27/drivers/staging/lustre/lustre/mdc/
H A Dmdc_internal.h67 void *ea, int ealen, void *ea2, int ea2len);
142 void *ea, int ealen, void *ea2, int ea2len,
H A Dmdc_lib.c339 void *ea, int ealen, void *ea2, int ea2len) mdc_setattr_pack()
361 if (ea == NULL) { /* Remove LOV EA */ mdc_setattr_pack()
367 memcpy(lum, ea, ealen); mdc_setattr_pack()
338 mdc_setattr_pack(struct ptlrpc_request *req, struct md_op_data *op_data, void *ea, int ealen, void *ea2, int ea2len) mdc_setattr_pack() argument
H A Dmdc_reint.c104 void *ea, int ealen, void *ea2, int ea2len, mdc_setattr()
151 mdc_setattr_pack(req, op_data, ea, ealen, ea2, ea2len); mdc_setattr()
103 mdc_setattr(struct obd_export *exp, struct md_op_data *op_data, void *ea, int ealen, void *ea2, int ea2len, struct ptlrpc_request **request, struct md_open_data **mod) mdc_setattr() argument
/linux-4.1.27/drivers/staging/lustre/lustre/osc/
H A Dosc_internal.h89 struct obdo *oa, struct lov_stripe_md **ea,
92 struct lov_stripe_md **ea, struct obd_trans_info *oti);
H A Dosc_request.c459 struct lov_stripe_md **ea, struct obd_trans_info *oti) osc_real_create()
467 LASSERT(ea); osc_real_create()
469 lsm = *ea; osc_real_create()
524 *ea = lsm; osc_real_create()
541 if (rc && !*ea) osc_real_create()
721 struct obdo *oa, struct lov_stripe_md **ea, osc_create()
727 LASSERT(ea); osc_create()
732 return osc_real_create(exp, oa, ea, oti); osc_create()
736 return osc_real_create(exp, oa, ea, oti); osc_create()
755 struct obdo *oa, struct lov_stripe_md *ea, osc_destroy()
458 osc_real_create(struct obd_export *exp, struct obdo *oa, struct lov_stripe_md **ea, struct obd_trans_info *oti) osc_real_create() argument
720 osc_create(const struct lu_env *env, struct obd_export *exp, struct obdo *oa, struct lov_stripe_md **ea, struct obd_trans_info *oti) osc_create() argument
754 osc_destroy(const struct lu_env *env, struct obd_export *exp, struct obdo *oa, struct lov_stripe_md *ea, struct obd_trans_info *oti, struct obd_export *md_export, void *capa) osc_destroy() argument
/linux-4.1.27/arch/x86/kvm/
H A Demulate.c655 la = seg_base(ctxt, addr.seg) + addr.ea; __linearize()
681 if (addr.ea <= lim) __linearize()
685 if (addr.ea > lim) __linearize()
690 *max_size = (u64)lim + 1 - addr.ea; __linearize()
725 .ea = dst }; assign_eip()
728 addr.ea = dst & ((1UL << (ctxt->op_bytes << 3)) - 1); assign_eip()
731 ctxt->_eip = addr.ea; assign_eip()
796 .ea = ctxt->eip + cur_size }; __do_insn_fetch_bytes()
897 addr.ea += 2; read_descriptor()
1289 op->addr.mem.ea = modrm_ea; decode_modrm()
1291 ctxt->memop.addr.mem.ea = (u32)ctxt->memop.addr.mem.ea; decode_modrm()
1305 op->addr.mem.ea = insn_fetch(u16, ctxt); decode_abs()
1308 op->addr.mem.ea = insn_fetch(u32, ctxt); decode_abs()
1311 op->addr.mem.ea = insn_fetch(u64, ctxt); decode_abs()
1332 ctxt->dst.addr.mem.ea = address_mask(ctxt, fetch_bit_operand()
1333 ctxt->dst.addr.mem.ea + (sv >> 3)); fetch_bit_operand()
1758 addr.ea = reg_read(ctxt, VCPU_REGS_RSP) & stack_mask(ctxt); push()
1777 addr.ea = reg_read(ctxt, VCPU_REGS_RSP) & stack_mask(ctxt); emulate_pop()
2967 op->addr.mem.ea = register_address(ctxt, reg); string_addr_inc()
4275 op->addr.mem.ea = ctxt->_eip; decode_imm()
4360 op->addr.mem.ea = decode_operand()
4414 op->addr.mem.ea = decode_operand()
4423 op->addr.mem.ea = decode_operand()
4432 op->addr.mem.ea = ctxt->_eip; decode_operand()
4736 ctxt->memopp->addr.mem.ea = address_mask(ctxt, x86_decode_insn()
4737 ctxt->memopp->addr.mem.ea + ctxt->_eip); x86_decode_insn()
4993 ctxt->dst.val = ctxt->src.addr.mem.ea; x86_emulate_insn()
/linux-4.1.27/drivers/base/
H A Dcore.c152 struct dev_ext_attribute *ea = to_ext_attr(attr); device_store_ulong() local
157 *(unsigned long *)(ea->var) = new; device_store_ulong()
167 struct dev_ext_attribute *ea = to_ext_attr(attr); device_show_ulong() local
168 return snprintf(buf, PAGE_SIZE, "%lx\n", *(unsigned long *)(ea->var)); device_show_ulong()
176 struct dev_ext_attribute *ea = to_ext_attr(attr); device_store_int() local
181 *(int *)(ea->var) = new; device_store_int()
191 struct dev_ext_attribute *ea = to_ext_attr(attr); device_show_int() local
193 return snprintf(buf, PAGE_SIZE, "%d\n", *(int *)(ea->var)); device_show_int()
200 struct dev_ext_attribute *ea = to_ext_attr(attr); device_store_bool() local
202 if (strtobool(buf, ea->var) < 0) device_store_bool()
212 struct dev_ext_attribute *ea = to_ext_attr(attr); device_show_bool() local
214 return snprintf(buf, PAGE_SIZE, "%d\n", *(bool *)(ea->var)); device_show_bool()
/linux-4.1.27/arch/powerpc/platforms/ps3/
H A Dhtab.c159 static void ps3_hpte_updateboltedpp(unsigned long newpp, unsigned long ea, ps3_hpte_updateboltedpp() argument
/linux-4.1.27/net/ieee802154/
H A Dtrace.h59 TP_printk(WPAN_PHY_PR_FMT ", virtual intf name: %s, type: %d, ea %llx",
/linux-4.1.27/arch/powerpc/platforms/powernv/
H A Dopal.c487 uint64_t ea = get_mce_fault_addr(evt); opal_recover_mce() local
495 } else if (ea && !is_kernel_addr(ea)) { opal_recover_mce()
/linux-4.1.27/drivers/staging/lustre/lustre/lclient/
H A Dlcommon_misc.c110 /* for each osc event update ea size */ cl_ocd_update()
/linux-4.1.27/arch/s390/include/asm/
H A Detr.h17 unsigned int ea : 1; /* ETR alert control */ member in struct:etr_eacr
/linux-4.1.27/arch/m68k/mm/
H A Dmemory.c178 * (Low order bits of the ea of a CINVP/CPUSHP are "don't care"s).
/linux-4.1.27/drivers/net/ethernet/tile/
H A Dtilepro.c938 netio_ipp_address_t ea = { tile_net_open_aux() local
944 ea.pte = hv_pte_set_lotar(ea.pte, epp_lotar); tile_net_open_aux()
945 ea.pte = hv_pte_set_mode(ea.pte, HV_PTE_MODE_CACHE_TILE_L3); tile_net_open_aux()
947 (HV_VirtAddr)&ea, tile_net_open_aux()
948 sizeof(ea), tile_net_open_aux()
/linux-4.1.27/drivers/mmc/card/
H A Dmmc_test.c1927 unsigned int dev_addr, cnt, rnd_addr, range1, range2, last_ea = 0, ea; mmc_test_rnd_perf() local
1944 ea = mmc_test_rnd_num(range1); mmc_test_rnd_perf()
1945 if (ea == last_ea) mmc_test_rnd_perf()
1946 ea -= 1; mmc_test_rnd_perf()
1947 last_ea = ea; mmc_test_rnd_perf()
1948 dev_addr = rnd_addr + test->card->pref_erase * ea + mmc_test_rnd_perf()
/linux-4.1.27/drivers/staging/lustre/lustre/include/
H A Dobd_class.h791 struct obdo *obdo, struct lov_stripe_md **ea, obd_create()
799 rc = OBP(exp->exp_obd, create)(env, exp, obdo, ea, oti); obd_create()
804 struct obdo *obdo, struct lov_stripe_md *ea, obd_destroy()
813 rc = OBP(exp->exp_obd, destroy)(env, exp, obdo, ea, oti, md_exp, capa); obd_destroy()
1639 void *ea, int ealen, void *ea2, int ea2len, md_setattr()
1647 rc = MDP(exp->exp_obd, setattr)(exp, op_data, ea, ealen, md_setattr()
790 obd_create(const struct lu_env *env, struct obd_export *exp, struct obdo *obdo, struct lov_stripe_md **ea, struct obd_trans_info *oti) obd_create() argument
803 obd_destroy(const struct lu_env *env, struct obd_export *exp, struct obdo *obdo, struct lov_stripe_md *ea, struct obd_trans_info *oti, struct obd_export *md_exp, void *capa) obd_destroy() argument
1638 md_setattr(struct obd_export *exp, struct md_op_data *op_data, void *ea, int ealen, void *ea2, int ea2len, struct ptlrpc_request **request, struct md_open_data **mod) md_setattr() argument
H A Dobd.h1164 struct obdo *oa, struct lov_stripe_md **ea,
1167 struct obdo *oa, struct lov_stripe_md *ea,
/linux-4.1.27/drivers/input/serio/
H A Dhyperv-keyboard.c417 * {f912ad6d-2b17-48ea-bd65-f927a61c7684}
/linux-4.1.27/include/uapi/linux/
H A Dgfs2_ondisk.h377 #define GFS2_EAFLAG_LAST 0x01 /* last ea in block */
/linux-4.1.27/arch/s390/kernel/
H A Dtime.c454 * e0 e1 dp p0 p1 ea es sl
497 .p0 = 0, .p1 = 0, ._pad1 = 0, .ea = 0, etr_reset()
1013 eacr.ea = eacr.p0 || eacr.p1; etr_work_fn()
1014 if (!eacr.ea) { etr_work_fn()
/linux-4.1.27/drivers/staging/lustre/lustre/lov/
H A Dlov_obd.c987 struct lov_stripe_md **ea, struct obd_trans_info *oti) lov_recreate()
1002 lsm = *ea; lov_recreate()
1041 struct obdo *src_oa, struct lov_stripe_md **ea, lov_create()
1047 LASSERT(ea != NULL); lov_create()
1065 rc = lov_recreate(exp, src_oa, ea, oti); lov_create()
986 lov_recreate(struct obd_export *exp, struct obdo *src_oa, struct lov_stripe_md **ea, struct obd_trans_info *oti) lov_recreate() argument
1040 lov_create(const struct lu_env *env, struct obd_export *exp, struct obdo *src_oa, struct lov_stripe_md **ea, struct obd_trans_info *oti) lov_create() argument
/linux-4.1.27/drivers/net/ethernet/apple/
H A Dbmac.c1185 bmac_get_station_address(struct net_device *dev, unsigned char *ea) bmac_get_station_address() argument
1194 ea[2*i] = bitrev8(data & 0x0ff); bmac_get_station_address()
1195 ea[2*i+1] = bitrev8((data >> 8) & 0x0ff); bmac_get_station_address()
/linux-4.1.27/arch/x86/include/asm/
H A Dkvm_emulate.h228 ulong ea; member in struct:operand::__anon3047::segmented_address
/linux-4.1.27/drivers/atm/
H A Dfirestream.h457 struct FS_QENTRY *sa, *ea; member in struct:queue
H A Dfirestream.c1429 txq->ea = p; init_q()
/linux-4.1.27/tools/lib/traceevent/
H A Devent-parse.c5275 struct event_format * const * ea = a; events_id_cmp() local
5278 if ((*ea)->id < (*eb)->id) events_id_cmp()
5281 if ((*ea)->id > (*eb)->id) events_id_cmp()
5289 struct event_format * const * ea = a; events_name_cmp() local
5293 res = strcmp((*ea)->name, (*eb)->name); events_name_cmp()
5297 res = strcmp((*ea)->system, (*eb)->system); events_name_cmp()
5306 struct event_format * const * ea = a; events_system_cmp() local
5310 res = strcmp((*ea)->system, (*eb)->system); events_system_cmp()
5314 res = strcmp((*ea)->name, (*eb)->name); events_system_cmp()
H A Dparse-filter.c106 const struct filter_type *ea = a; filter_cmp() local
109 if (ea->event_id < eb->event_id) filter_cmp()
112 if (ea->event_id > eb->event_id) filter_cmp()
/linux-4.1.27/arch/x86/kernel/cpu/
H A Dperf_event_intel_pt.c86 struct dev_ext_attribute *ea = pt_cap_show() local
88 enum pt_capabilities cap = (long)ea->var; pt_cap_show()
/linux-4.1.27/drivers/bus/
H A Darm-ccn.c222 struct dev_ext_attribute *ea = container_of(attr, arm_ccn_pmu_format_show() local
225 return snprintf(buf, PAGE_SIZE, "%s\n", (char *)ea->var); arm_ccn_pmu_format_show()
/linux-4.1.27/drivers/media/dvb-frontends/
H A Dm88ds3103.c486 /* u8tmp1[5:2] => fe[3:0], u8tmp1[1:0] => ea[7:6] */ m88ds3103_set_frontend()
488 /* u8tmp2[5:0] => ea[5:0] */ m88ds3103_set_frontend()
/linux-4.1.27/drivers/video/fbdev/sis/
H A Dinit.h630 0xe9,0x8b,0xdf,0x28,0x00,0xe7,0x04,0xc3, /* e9,8b is ea,8c on 300 */
645 0xe9,0x8b,0xdf,0x28,0x00,0xe7,0x04,0xe3, /* e9,8b is ea,8c on 300 */
/linux-4.1.27/arch/x86/kernel/
H A Duprobes.c139 * ea - formerly jmp seg:ofs
/linux-4.1.27/fs/ext3/
H A Dxattr.c12 * ea-in-inode support by Alex Tomas <alex@clusterfs.com> aka bzzz
/linux-4.1.27/drivers/staging/lustre/lustre/lmv/
H A Dlmv_obd.c2029 void *ea, int ealen, void *ea2, int ea2len, lmv_setattr()
2050 rc = md_setattr(tgt->ltd_exp, op_data, ea, ealen, ea2, lmv_setattr()
2028 lmv_setattr(struct obd_export *exp, struct md_op_data *op_data, void *ea, int ealen, void *ea2, int ea2len, struct ptlrpc_request **request, struct md_open_data **mod) lmv_setattr() argument
/linux-4.1.27/arch/powerpc/kernel/
H A Dhead_fsl_booke.S795 andc r12, r12, r13 /* mask off ea bits within the page */
/linux-4.1.27/fs/ext4/
H A Dxattr.c12 * ea-in-inode support by Alex Tomas <alex@clusterfs.com> aka bzzz
/linux-4.1.27/fs/cifs/
H A Dcifssmb.c6254 cifs_dbg(FYI, "ea length %d\n", list_len); CIFSSMBQAllEAs()
6271 /* account for ea list len */ CIFSSMBQAllEAs()
6382 now use name_len to calculate length of ea name CIFSSMBSetEA()
/linux-4.1.27/drivers/media/usb/gspca/
H A Dzc3xx.c1856 {0xa0, 0xea, ZC3XX_R196_ANTIFLICKERMID}, /* 01,96,ea,cc */
2579 {0xa0, 0xea, ZC3XX_R01F_HSYNC_2}, /* 00,1f,ea,cc */
2674 {0xa0, 0xea, ZC3XX_R01F_HSYNC_2}, /* 00,1f,ea,cc */
5155 {0xa0, 0xea, ZC3XX_R01F_HSYNC_2}, /* 00,1f,ea,cc */
/linux-4.1.27/drivers/staging/lustre/lustre/obdclass/
H A Dlu_object.c1963 * inode, one for ea. Unfortunately setting this high value results in lu_global_init()
/linux-4.1.27/drivers/tty/
H A Dn_gsm.c1163 /* Must be at least one byte following ea */ gsm_control_rls()
/linux-4.1.27/drivers/staging/lustre/lustre/include/lustre/
H A Dlustre_idl.h3552 /** The link ea holds 1 \a link_ea_entry for each hardlink */
/linux-4.1.27/fs/ntfs/
H A Dlayout.h1089 } __attribute__ ((__packed__)) ea; member in union:__anon11167::__anon11168

Completed in 3900 milliseconds