/linux-4.1.27/arch/powerpc/math-emu/ |
H A D | stfd.c | 6 stfd(void *frS, void *ea) stfd() argument 10 printk("%s: S %p, ea %p: ", __func__, frS, ea); stfd() 16 if (copy_to_user(ea, frS, sizeof(double))) stfd()
|
H A D | stfiwx.c | 6 stfiwx(u32 *frS, void *ea) stfiwx() argument 9 printk("%s: %p %p\n", __func__, frS, ea); stfiwx() 12 if (copy_to_user(ea, &frS[1], sizeof(frS[1]))) stfiwx()
|
H A D | lfd.c | 9 lfd(void *frD, void *ea) lfd() argument 11 if (copy_from_user(frD, ea, sizeof(double))) lfd() 14 printk("%s: D %p, ea %p: ", __func__, frD, ea); lfd()
|
H A D | lfs.c | 11 lfs(void *frD, void *ea) lfs() argument 19 printk("%s: D %p, ea %p\n", __func__, frD, ea); lfs() 22 if (copy_from_user(&f, ea, sizeof(float))) lfs()
|
H A D | stfs.c | 11 stfs(void *frS, void *ea) stfs() argument 19 printk("%s: S %p, ea %p\n", __func__, frS, ea); stfs() 37 if (copy_to_user(ea, &f, sizeof(float))) stfs()
|
/linux-4.1.27/fs/gfs2/ |
H A D | xattr.h | 16 #define GFS2_EA_REC_LEN(ea) be32_to_cpu((ea)->ea_rec_len) 17 #define GFS2_EA_DATA_LEN(ea) be32_to_cpu((ea)->ea_data_len) 19 #define GFS2_EA_SIZE(ea) \ 20 ALIGN(sizeof(struct gfs2_ea_header) + (ea)->ea_name_len + \ 21 ((GFS2_EA_IS_STUFFED(ea)) ? GFS2_EA_DATA_LEN(ea) : \ 22 (sizeof(__be64) * (ea)->ea_num_ptrs)), 8) 24 #define GFS2_EA_IS_STUFFED(ea) (!(ea)->ea_num_ptrs) 25 #define GFS2_EA_IS_LAST(ea) ((ea)->ea_flags & GFS2_EAFLAG_LAST) 30 #define GFS2_EA2NAME(ea) ((char *)((struct gfs2_ea_header *)(ea) + 1)) 31 #define GFS2_EA2DATA(ea) (GFS2_EA2NAME(ea) + (ea)->ea_name_len) 33 #define GFS2_EA2DATAPTRS(ea) \ 34 ((__be64 *)(GFS2_EA2NAME(ea) + ALIGN((ea)->ea_name_len, 8))) 36 #define GFS2_EA2NEXT(ea) \ 37 ((struct gfs2_ea_header *)((char *)(ea) + GFS2_EA_REC_LEN(ea)))
|
H A D | xattr.c | 76 struct gfs2_ea_header *ea, 82 struct gfs2_ea_header *ea, *prev = NULL; ea_foreach_i() local 88 for (ea = GFS2_EA_BH2FIRST(bh);; prev = ea, ea = GFS2_EA2NEXT(ea)) { ea_foreach_i() 89 if (!GFS2_EA_REC_LEN(ea)) ea_foreach_i() 91 if (!(bh->b_data <= (char *)ea && (char *)GFS2_EA2NEXT(ea) <= ea_foreach_i() 94 if (!GFS2_EATYPE_VALID(ea->ea_type)) ea_foreach_i() 97 error = ea_call(ip, bh, ea, prev, data); ea_foreach_i() 101 if (GFS2_EA_IS_LAST(ea)) { ea_foreach_i() 102 if ((char *)GFS2_EA2NEXT(ea) != ea_foreach_i() 167 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_find_i() 172 if (ea->ea_type == GFS2_EATYPE_UNUSED) ea_find_i() 175 if (ea->ea_type == ef->type) { ea_find_i() 176 if (ea->ea_name_len == ef->namel && ea_find_i() 177 !memcmp(GFS2_EA2NAME(ea), ef->name, ea->ea_name_len)) { ea_find_i() 181 el->el_ea = ea; ea_find_i() 214 * @ea: 226 struct gfs2_ea_header *ea, ea_dealloc_unstuffed() 246 if (GFS2_EA_IS_STUFFED(ea)) ea_dealloc_unstuffed() 249 dataptrs = GFS2_EA2DATAPTRS(ea); ea_dealloc_unstuffed() 250 for (x = 0; x < ea->ea_num_ptrs; x++, dataptrs++) { ea_dealloc_unstuffed() 276 dataptrs = GFS2_EA2DATAPTRS(ea); ea_dealloc_unstuffed() 277 for (x = 0; x < ea->ea_num_ptrs; x++, dataptrs++) { ea_dealloc_unstuffed() 300 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_dealloc_unstuffed() 303 if (GFS2_EA_IS_LAST(ea)) ea_dealloc_unstuffed() 306 ea->ea_type = GFS2_EATYPE_UNUSED; ea_dealloc_unstuffed() 307 ea->ea_num_ptrs = 0; ea_dealloc_unstuffed() 326 struct gfs2_ea_header *ea, ea_remove_unstuffed() 339 error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL); ea_remove_unstuffed() 351 static inline unsigned int gfs2_ea_strlen(struct gfs2_ea_header *ea) gfs2_ea_strlen() argument 353 switch (ea->ea_type) { gfs2_ea_strlen() 355 return 5 + ea->ea_name_len + 1; gfs2_ea_strlen() 357 return 7 + ea->ea_name_len + 1; gfs2_ea_strlen() 359 return 9 + ea->ea_name_len + 1; gfs2_ea_strlen() 366 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_list_i() 371 unsigned int ea_size = gfs2_ea_strlen(ea); ea_list_i() 373 if (ea->ea_type == GFS2_EATYPE_UNUSED) ea_list_i() 384 switch (ea->ea_type) { ea_list_i() 402 memcpy(er->er_data + ei->ei_size + l, GFS2_EA2NAME(ea), ea_list_i() 403 ea->ea_name_len); ea_list_i() 455 * @ea: The extended attribute header structure 462 static int gfs2_iter_unstuffed(struct gfs2_inode *ip, struct gfs2_ea_header *ea, gfs2_iter_unstuffed() argument 467 unsigned int amount = GFS2_EA_DATA_LEN(ea); gfs2_iter_unstuffed() 469 __be64 *dataptrs = GFS2_EA2DATAPTRS(ea); gfs2_iter_unstuffed() 623 struct gfs2_ea_header *ea; ea_alloc_blk() local 637 ea = GFS2_EA_BH2FIRST(*bhp); ea_alloc_blk() 638 ea->ea_rec_len = cpu_to_be32(sdp->sd_jbsize); ea_alloc_blk() 639 ea->ea_type = GFS2_EATYPE_UNUSED; ea_alloc_blk() 640 ea->ea_flags = GFS2_EAFLAG_LAST; ea_alloc_blk() 641 ea->ea_num_ptrs = 0; ea_alloc_blk() 649 * ea_write - writes the request info to an ea, creating new blocks if 652 * @ea: the location of the new ea in a block 660 static int ea_write(struct gfs2_inode *ip, struct gfs2_ea_header *ea, ea_write() argument 666 ea->ea_data_len = cpu_to_be32(er->er_data_len); ea_write() 667 ea->ea_name_len = er->er_name_len; ea_write() 668 ea->ea_type = er->er_type; ea_write() 669 ea->__pad = 0; ea_write() 671 memcpy(GFS2_EA2NAME(ea), er->er_name, er->er_name_len); ea_write() 674 ea->ea_num_ptrs = 0; ea_write() 675 memcpy(GFS2_EA2DATA(ea), er->er_data, er->er_data_len); ea_write() 677 __be64 *dataptr = GFS2_EA2DATAPTRS(ea); ea_write() 683 ea->ea_num_ptrs = DIV_ROUND_UP(er->er_data_len, sdp->sd_jbsize); ea_write() 684 for (x = 0; x < ea->ea_num_ptrs; x++) { ea_write() 815 static struct gfs2_ea_header *ea_split_ea(struct gfs2_ea_header *ea) ea_split_ea() argument 817 u32 ea_size = GFS2_EA_SIZE(ea); ea_split_ea() 818 struct gfs2_ea_header *new = (struct gfs2_ea_header *)((char *)ea + ea_split_ea() 820 u32 new_size = GFS2_EA_REC_LEN(ea) - ea_size; ea_split_ea() 821 int last = ea->ea_flags & GFS2_EAFLAG_LAST; ea_split_ea() 823 ea->ea_rec_len = cpu_to_be32(ea_size); ea_split_ea() 824 ea->ea_flags ^= last; ea_split_ea() 835 struct gfs2_ea_header *ea = el->el_ea; ea_set_remove_stuffed() local 841 if (!prev || !GFS2_EA_IS_STUFFED(ea)) { ea_set_remove_stuffed() 842 ea->ea_type = GFS2_EATYPE_UNUSED; ea_set_remove_stuffed() 844 } else if (GFS2_EA2NEXT(prev) != ea) { ea_set_remove_stuffed() 846 gfs2_assert_withdraw(GFS2_SB(&ip->i_inode), GFS2_EA2NEXT(prev) == ea); ea_set_remove_stuffed() 849 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_set_remove_stuffed() 852 if (GFS2_EA_IS_LAST(ea)) ea_set_remove_stuffed() 867 struct gfs2_ea_header *ea, struct ea_set *es) ea_set_simple_noalloc() 880 ea = ea_split_ea(ea); ea_set_simple_noalloc() 882 ea_write(ip, ea, er); ea_set_simple_noalloc() 903 struct gfs2_ea_header *ea = es->es_ea; ea_set_simple_alloc() local 909 ea = ea_split_ea(ea); ea_set_simple_alloc() 911 error = ea_write(ip, ea, er); ea_set_simple_alloc() 922 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_set_simple() 933 if (ea->ea_type == GFS2_EATYPE_UNUSED) { ea_set_simple() 934 if (GFS2_EA_REC_LEN(ea) < size) ea_set_simple() 936 if (!GFS2_EA_IS_STUFFED(ea)) { ea_set_simple() 937 error = ea_remove_unstuffed(ip, bh, ea, prev, 1); ea_set_simple() 942 } else if (GFS2_EA_REC_LEN(ea) - GFS2_EA_SIZE(ea) >= size) ea_set_simple() 948 error = ea_set_simple_noalloc(ip, bh, ea, es); ea_set_simple() 955 es->es_ea = ea; ea_set_simple() 1088 struct gfs2_ea_header *ea = el->el_ea; ea_remove_stuffed() local 1102 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_remove_stuffed() 1105 if (GFS2_EA_IS_LAST(ea)) ea_remove_stuffed() 1108 ea->ea_type = GFS2_EATYPE_UNUSED; ea_remove_stuffed() 1239 struct gfs2_ea_header *ea, char *data) ea_acl_chmod_unstuffed() 1242 unsigned int amount = GFS2_EA_DATA_LEN(ea); ea_acl_chmod_unstuffed() 1250 ret = gfs2_iter_unstuffed(ip, ea, data, NULL); ea_acl_chmod_unstuffed() 166 ea_find_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_find_i() argument 225 ea_dealloc_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_dealloc_unstuffed() argument 325 ea_remove_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, int leave) ea_remove_unstuffed() argument 365 ea_list_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_list_i() argument 866 ea_set_simple_noalloc(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct ea_set *es) ea_set_simple_noalloc() argument 921 ea_set_simple(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_set_simple() argument 1238 ea_acl_chmod_unstuffed(struct gfs2_inode *ip, struct gfs2_ea_header *ea, char *data) ea_acl_chmod_unstuffed() argument
|
H A D | inode.c | 441 struct gfs2_ea_header *ea; gfs2_init_xattr() local 448 ea = GFS2_EA_BH2FIRST(bh); gfs2_init_xattr() 449 ea->ea_rec_len = cpu_to_be32(sdp->sd_jbsize); gfs2_init_xattr() 450 ea->ea_type = GFS2_EATYPE_UNUSED; gfs2_init_xattr() 451 ea->ea_flags = GFS2_EAFLAG_LAST; gfs2_init_xattr()
|
/linux-4.1.27/fs/hpfs/ |
H A D | ea.c | 2 * linux/fs/hpfs/ea.c 19 struct extended_attribute *ea = (struct extended_attribute *)ex; hpfs_ea_ext_remove() local 26 if (ea_indirect(ea)) { hpfs_ea_ext_remove() 27 if (ea_valuelen(ea) != 8) { hpfs_ea_ext_remove() 28 hpfs_error(s, "ea_indirect(ea) set while ea->valuelen!=8, %s %08x, pos %08x", hpfs_ea_ext_remove() 32 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 9, ex+4)) hpfs_ea_ext_remove() 34 hpfs_ea_remove(s, ea_sec(ea), ea_in_anode(ea), ea_len(ea)); hpfs_ea_ext_remove() 36 pos += ea->namelen + ea_valuelen(ea) + 5; hpfs_ea_ext_remove() 80 struct extended_attribute *ea; hpfs_read_ea() local 82 for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) hpfs_read_ea() 83 if (!strcmp(ea->name, key)) { hpfs_read_ea() 84 if (ea_indirect(ea)) hpfs_read_ea() 86 if (ea_valuelen(ea) >= size) hpfs_read_ea() 88 memcpy(buf, ea_data(ea), ea_valuelen(ea)); hpfs_read_ea() 89 buf[ea_valuelen(ea)] = 0; hpfs_read_ea() 97 ea = (struct extended_attribute *)ex; hpfs_read_ea() 104 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 1 + (ea_indirect(ea) ? 8 : 0), ex + 4)) hpfs_read_ea() 106 if (!strcmp(ea->name, key)) { hpfs_read_ea() 107 if (ea_indirect(ea)) hpfs_read_ea() 109 if (ea_valuelen(ea) >= size) hpfs_read_ea() 111 if (hpfs_ea_read(s, a, ano, pos + 4 + ea->namelen + 1, ea_valuelen(ea), buf)) hpfs_read_ea() 113 buf[ea_valuelen(ea)] = 0; hpfs_read_ea() 116 pos += ea->namelen + ea_valuelen(ea) + 5; hpfs_read_ea() 120 if (ea_len(ea) >= size) hpfs_read_ea() 122 if (hpfs_ea_read(s, ea_sec(ea), ea_in_anode(ea), 0, ea_len(ea), buf)) hpfs_read_ea() 124 buf[ea_len(ea)] = 0; hpfs_read_ea() 135 struct extended_attribute *ea; hpfs_get_ea() local 137 for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) hpfs_get_ea() 138 if (!strcmp(ea->name, key)) { hpfs_get_ea() 139 if (ea_indirect(ea)) hpfs_get_ea() 140 return get_indirect_ea(s, ea_in_anode(ea), ea_sec(ea), *size = ea_len(ea)); hpfs_get_ea() 141 if (!(ret = kmalloc((*size = ea_valuelen(ea)) + 1, GFP_NOFS))) { hpfs_get_ea() 145 memcpy(ret, ea_data(ea), ea_valuelen(ea)); hpfs_get_ea() 146 ret[ea_valuelen(ea)] = 0; hpfs_get_ea() 155 ea = (struct extended_attribute *)ex; hpfs_get_ea() 162 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 1 + (ea_indirect(ea) ? 8 : 0), ex + 4)) hpfs_get_ea() 164 if (!strcmp(ea->name, key)) { hpfs_get_ea() 165 if (ea_indirect(ea)) hpfs_get_ea() 166 return get_indirect_ea(s, ea_in_anode(ea), ea_sec(ea), *size = ea_len(ea)); hpfs_get_ea() 167 if (!(ret = kmalloc((*size = ea_valuelen(ea)) + 1, GFP_NOFS))) { hpfs_get_ea() 171 if (hpfs_ea_read(s, a, ano, pos + 4 + ea->namelen + 1, ea_valuelen(ea), ret)) { hpfs_get_ea() 175 ret[ea_valuelen(ea)] = 0; hpfs_get_ea() 178 pos += ea->namelen + ea_valuelen(ea) + 5; hpfs_get_ea() 185 * when this ea exists, it MUST have the same size as size of data. 198 struct extended_attribute *ea; hpfs_set_ea() local 200 for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) hpfs_set_ea() 201 if (!strcmp(ea->name, key)) { hpfs_set_ea() 202 if (ea_indirect(ea)) { hpfs_set_ea() 203 if (ea_len(ea) == size) hpfs_set_ea() 204 set_indirect_ea(s, ea_in_anode(ea), ea_sec(ea), data, size); hpfs_set_ea() 205 } else if (ea_valuelen(ea) == size) { hpfs_set_ea() 206 memcpy(ea_data(ea), data, size); hpfs_set_ea() 216 ea = (struct extended_attribute *)ex; hpfs_set_ea() 223 if (hpfs_ea_read(s, a, ano, pos + 4, ea->namelen + 1 + (ea_indirect(ea) ? 8 : 0), ex + 4)) hpfs_set_ea() 225 if (!strcmp(ea->name, key)) { hpfs_set_ea() 226 if (ea_indirect(ea)) { hpfs_set_ea() 227 if (ea_len(ea) == size) hpfs_set_ea() 228 set_indirect_ea(s, ea_in_anode(ea), ea_sec(ea), data, size); hpfs_set_ea() 231 if (ea_valuelen(ea) == size) hpfs_set_ea() 232 hpfs_ea_write(s, a, ano, pos + 4 + ea->namelen + 1, size, data); hpfs_set_ea() 236 pos += ea->namelen + ea_valuelen(ea) + 5; hpfs_set_ea() 254 ea = fnode_end_ea(fnode); hpfs_set_ea() 255 *(char *)ea = 0; hpfs_set_ea() 256 ea->namelen = strlen(key); hpfs_set_ea() 257 ea->valuelen_lo = size; hpfs_set_ea() 258 ea->valuelen_hi = size >> 8; hpfs_set_ea() 259 strcpy(ea->name, key); hpfs_set_ea() 260 memcpy(ea_data(ea), data, size); hpfs_set_ea() 265 But what .. HPFS.IFS has also bugs in ea management. */ hpfs_set_ea() 297 /* Aargh... don't know how to create ea anodes :-( */ hpfs_set_ea()
|
H A D | Makefile | 7 hpfs-objs := alloc.o anode.o buffer.o dentry.o dir.o dnode.o ea.o file.o \
|
H A D | inode.c | 49 void *ea; hpfs_read_inode() local 62 if ((ea = hpfs_get_ea(i->i_sb, fnode, "UID", &ea_size))) { hpfs_read_inode() 64 i_uid_write(i, le16_to_cpu(*(__le16*)ea)); hpfs_read_inode() 67 kfree(ea); hpfs_read_inode() 69 if ((ea = hpfs_get_ea(i->i_sb, fnode, "GID", &ea_size))) { hpfs_read_inode() 71 i_gid_write(i, le16_to_cpu(*(__le16*)ea)); hpfs_read_inode() 74 kfree(ea); hpfs_read_inode() 76 if ((ea = hpfs_get_ea(i->i_sb, fnode, "SYMLINK", &ea_size))) { hpfs_read_inode() 77 kfree(ea); hpfs_read_inode() 87 if ((ea = hpfs_get_ea(i->i_sb, fnode, "MODE", &ea_size))) { hpfs_read_inode() 91 mode = le16_to_cpu(*(__le16*)ea); hpfs_read_inode() 94 kfree(ea); hpfs_read_inode() 97 if ((ea = hpfs_get_ea(i->i_sb, fnode, "DEV", &ea_size))) { hpfs_read_inode() 99 rdev = le32_to_cpu(*(__le32*)ea); hpfs_read_inode() 100 kfree(ea); hpfs_read_inode() 152 __le32 ea; hpfs_write_inode_ea() local 154 ea = cpu_to_le32(i_uid_read(i)); hpfs_write_inode_ea() 155 hpfs_set_ea(i, fnode, "UID", (char*)&ea, 2); hpfs_write_inode_ea() 159 ea = cpu_to_le32(i_gid_read(i)); hpfs_write_inode_ea() 160 hpfs_set_ea(i, fnode, "GID", (char *)&ea, 2); hpfs_write_inode_ea() 168 ea = cpu_to_le32(i->i_mode); hpfs_write_inode_ea() 170 hpfs_set_ea(i, fnode, "MODE", (char *)&ea, 2); hpfs_write_inode_ea() 174 ea = cpu_to_le32(new_encode_dev(i->i_rdev)); hpfs_write_inode_ea() 175 hpfs_set_ea(i, fnode, "DEV", (char *)&ea, 4); hpfs_write_inode_ea()
|
H A D | hpfs_fn.h | 54 unsigned i_ea_mode : 1; /* file's permission is stored in ea */ 55 unsigned i_ea_uid : 1; /* file's uid is stored in ea */ 56 unsigned i_ea_gid : 1; /* file's gid is stored in ea */ 139 static unsigned ea_valuelen(struct extended_attribute *ea) ea_valuelen() argument 141 return ea->valuelen_lo + 256 * ea->valuelen_hi; ea_valuelen() 144 static inline struct extended_attribute *next_ea(struct extended_attribute *ea) next_ea() argument 146 return (struct extended_attribute *)((char *)ea + 5 + ea->namelen + ea_valuelen(ea)); next_ea() 149 static inline secno ea_sec(struct extended_attribute *ea) ea_sec() argument 151 return le32_to_cpu(get_unaligned((__le32 *)((char *)ea + 9 + ea->namelen))); ea_sec() 154 static inline secno ea_len(struct extended_attribute *ea) ea_len() argument 156 return le32_to_cpu(get_unaligned((__le32 *)((char *)ea + 5 + ea->namelen))); ea_len() 159 static inline char *ea_data(struct extended_attribute *ea) ea_data() argument 161 return (char *)((char *)ea + 5 + ea->namelen); ea_data() 252 /* ea.c */
|
H A D | anode.c | 349 if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, sec, 1, "ea #1")) return -1; hpfs_ea_read() 372 if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, sec, 1, "ea #2")) return -1; hpfs_ea_write() 484 struct extended_attribute *ea; hpfs_remove_fnode() local 490 for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) hpfs_remove_fnode() 491 if (ea_indirect(ea)) hpfs_remove_fnode() 492 hpfs_ea_remove(s, ea_sec(ea), ea_in_anode(ea), ea_len(ea)); hpfs_remove_fnode()
|
H A D | map.c | 145 struct extended_attribute *ea; hpfs_map_fnode() local 176 ea = fnode_ea(fnode); hpfs_map_fnode() 178 while (ea != ea_end) { hpfs_map_fnode() 179 if (ea > ea_end) { hpfs_map_fnode() 184 ea = next_ea(ea); hpfs_map_fnode()
|
H A D | hpfs.h | 399 the data btree or some ea or the 400 main ea bootage pointer ea_secno 447 __le32 ea_size_l; /* length of disk-resident ea's */ 448 __le32 ea_secno; /* first sector of disk-resident ea's*/ 449 __le16 ea_size_s; /* length of fnode-resident ea's */ 464 to first fnode-resident ea */ 469 u8 ea[316]; /* zero or more EA's, packed together member in struct:fnode 528 /* bit 7 set -> required ea */ 545 static inline bool ea_indirect(struct extended_attribute *ea) ea_indirect() argument 547 return ea->flags & EA_indirect; ea_indirect() 550 static inline bool ea_in_anode(struct extended_attribute *ea) ea_in_anode() argument 552 return ea->flags & EA_anode; ea_in_anode()
|
/linux-4.1.27/arch/powerpc/mm/ |
H A D | copro_fault.c | 36 int copro_handle_mm_fault(struct mm_struct *mm, unsigned long ea, copro_handle_mm_fault() argument 51 vma = find_vma(mm, ea); copro_handle_mm_fault() 55 if (ea < vma->vm_start) { copro_handle_mm_fault() 58 if (expand_stack(vma, ea)) copro_handle_mm_fault() 78 *flt = handle_mm_fault(mm, vma, ea, is_write ? FAULT_FLAG_WRITE : 0); copro_handle_mm_fault() 101 int copro_calculate_slb(struct mm_struct *mm, u64 ea, struct copro_slb *slb) copro_calculate_slb() argument 106 switch (REGION_ID(ea)) { copro_calculate_slb() 108 pr_devel("%s: 0x%llx -- USER_REGION_ID\n", __func__, ea); copro_calculate_slb() 109 psize = get_slice_psize(mm, ea); copro_calculate_slb() 110 ssize = user_segment_size(ea); copro_calculate_slb() 111 vsid = get_vsid(mm->context.id, ea, ssize); copro_calculate_slb() 114 pr_devel("%s: 0x%llx -- VMALLOC_REGION_ID\n", __func__, ea); copro_calculate_slb() 115 if (ea < VMALLOC_END) copro_calculate_slb() 120 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); copro_calculate_slb() 123 pr_devel("%s: 0x%llx -- KERNEL_REGION_ID\n", __func__, ea); copro_calculate_slb() 126 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); copro_calculate_slb() 129 pr_debug("%s: invalid region access at %016llx\n", __func__, ea); copro_calculate_slb() 138 slb->esid = (ea & (ssize == MMU_SEGSIZE_1T ? ESID_MASK_1T : ESID_MASK)) | SLB_ESID_V; copro_calculate_slb()
|
H A D | hugetlbpage-book3e.c | 54 static inline int book3e_tlb_exists(unsigned long ea, unsigned long pid) book3e_tlb_exists() argument 66 : "=&r"(found) : "r"(ea)); book3e_tlb_exists() 72 : "=&r"(found) : "r"(ea)); book3e_tlb_exists() 78 void book3e_hugetlb_preload(struct vm_area_struct *vma, unsigned long ea, book3e_hugetlb_preload() argument 91 if (unlikely(is_kernel_addr(ea))) book3e_hugetlb_preload() 97 psize = get_slice_psize(mm, ea); book3e_hugetlb_preload() 112 if (unlikely(book3e_tlb_exists(ea, mm->context.id))) { book3e_hugetlb_preload() 124 mas2 = ea & ~((1UL << shift) - 1); book3e_hugetlb_preload()
|
H A D | slb.c | 29 extern void slb_allocate_realmode(unsigned long ea); 30 extern void slb_allocate_user(unsigned long ea); 32 static void slb_allocate(unsigned long ea) slb_allocate() argument 37 slb_allocate_realmode(ea); slb_allocate() 43 static inline unsigned long mk_esid_data(unsigned long ea, int ssize, mk_esid_data() argument 46 return (ea & slb_esid_mask(ssize)) | SLB_ESID_V | slot; mk_esid_data() 49 static inline unsigned long mk_vsid_data(unsigned long ea, int ssize, mk_vsid_data() argument 52 return (get_kernel_vsid(ea, ssize) << slb_vsid_shift(ssize)) | flags | mk_vsid_data() 56 static inline void slb_shadow_update(unsigned long ea, int ssize, slb_shadow_update() argument 67 cpu_to_be64(mk_vsid_data(ea, ssize, flags)); slb_shadow_update() 69 cpu_to_be64(mk_esid_data(ea, ssize, entry)); slb_shadow_update() 77 static inline void create_shadowed_slbe(unsigned long ea, int ssize, create_shadowed_slbe() argument 86 slb_shadow_update(ea, ssize, flags, entry); create_shadowed_slbe() 89 : "r" (mk_vsid_data(ea, ssize, flags)), create_shadowed_slbe() 90 "r" (mk_esid_data(ea, ssize, entry)) create_shadowed_slbe()
|
H A D | hash_utils_64.c | 922 static int subpage_protection(struct mm_struct *mm, unsigned long ea) subpage_protection() argument 928 if (ea >= spt->maxaddr) subpage_protection() 930 if (ea < 0x100000000UL) { subpage_protection() 934 sbpm = spt->protptrs[ea >> SBP_L3_SHIFT]; subpage_protection() 938 sbpp = sbpm[(ea >> SBP_L2_SHIFT) & (SBP_L2_COUNT - 1)]; subpage_protection() 941 spp = sbpp[(ea >> PAGE_SHIFT) & (SBP_L1_COUNT - 1)]; subpage_protection() 944 spp >>= 30 - 2 * ((ea >> 12) & 0xf); subpage_protection() 952 static inline int subpage_protection(struct mm_struct *mm, unsigned long ea) subpage_protection() argument 958 void hash_failure_debug(unsigned long ea, unsigned long access, hash_failure_debug() argument 965 ea, access, current->comm); hash_failure_debug() 970 static void check_paca_psize(unsigned long ea, struct mm_struct *mm, check_paca_psize() argument 974 if (psize != get_paca_psize(ea)) { check_paca_psize() 992 int hash_page_mm(struct mm_struct *mm, unsigned long ea, hash_page_mm() argument 1005 DBG_LOW("hash_page(ea=%016lx, access=%lx, trap=%lx\n", hash_page_mm() 1006 ea, access, trap); hash_page_mm() 1009 switch (REGION_ID(ea)) { hash_page_mm() 1017 psize = get_slice_psize(mm, ea); hash_page_mm() 1018 ssize = user_segment_size(ea); hash_page_mm() 1019 vsid = get_vsid(mm->context.id, ea, ssize); hash_page_mm() 1022 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); hash_page_mm() 1023 if (ea < VMALLOC_END) hash_page_mm() 1065 ea &= ~((1ul << mmu_psize_defs[psize].shift) - 1); hash_page_mm() 1069 ptep = __find_linux_pte_or_hugepte(pgdir, ea, &hugeshift); hash_page_mm() 1090 rc = __hash_page_thp(ea, access, vsid, (pmd_t *)ptep, hash_page_mm() 1094 rc = __hash_page_huge(ea, access, vsid, ptep, trap, hash_page_mm() 1107 check_paca_psize(ea, mm, psize, user_region); hash_page_mm() 1122 demote_segment_4k(mm, ea); hash_page_mm() 1132 demote_segment_4k(mm, ea); hash_page_mm() 1134 } else if (ea < VMALLOC_END) { hash_page_mm() 1149 check_paca_psize(ea, mm, psize, user_region); hash_page_mm() 1154 rc = __hash_page_64K(ea, access, vsid, ptep, trap, hash_page_mm() 1159 int spp = subpage_protection(mm, ea); hash_page_mm() 1163 rc = __hash_page_4K(ea, access, vsid, ptep, trap, hash_page_mm() 1171 hash_failure_debug(ea, access, vsid, trap, ssize, psize, hash_page_mm() 1187 int hash_page(unsigned long ea, unsigned long access, unsigned long trap, hash_page() argument 1193 if (REGION_ID(ea) == VMALLOC_REGION_ID) hash_page() 1199 return hash_page_mm(mm, ea, access, trap, flags); hash_page() 1203 void hash_preload(struct mm_struct *mm, unsigned long ea, hash_preload() argument 1213 BUG_ON(REGION_ID(ea) != USER_REGION_ID); hash_preload() 1217 if (unlikely(get_slice_psize(mm, ea) != mm->context.user_psize)) hash_preload() 1221 DBG_LOW("hash_preload(mm=%p, mm->pgdir=%p, ea=%016lx, access=%lx," hash_preload() 1222 " trap=%lx\n", mm, mm->pgd, ea, access, trap); hash_preload() 1230 ssize = user_segment_size(ea); hash_preload() 1231 vsid = get_vsid(mm->context.id, ea, ssize); hash_preload() 1244 ptep = find_linux_pte_or_hugepte(pgdir, ea, &hugepage_shift); hash_preload() 1267 rc = __hash_page_64K(ea, access, vsid, ptep, trap, hash_preload() 1271 rc = __hash_page_4K(ea, access, vsid, ptep, trap, update_flags, hash_preload() 1272 ssize, subpage_protection(mm, ea)); hash_preload() 1278 hash_failure_debug(ea, access, vsid, trap, ssize, hash_preload()
|
H A D | hugepage-hash64.c | 21 int __hash_page_thp(unsigned long ea, unsigned long access, unsigned long vsid, __hash_page_thp() argument 81 * Find the slot index details for this ea, using base page size. __hash_page_thp() 84 index = (ea & ~HPAGE_PMD_MASK) >> shift; __hash_page_thp() 87 vpn = hpt_vpn(ea, vsid, ssize); __hash_page_thp() 96 flush_hash_hugepage(vsid, ea, pmdp, MMU_PAGE_64K, __hash_page_thp() 171 hash_failure_debug(ea, access, vsid, trap, ssize, __hash_page_thp()
|
H A D | pgtable_64.c | 91 int map_kernel_page(unsigned long ea, unsigned long pa, int flags) map_kernel_page() argument 99 pgdp = pgd_offset_k(ea); map_kernel_page() 100 pudp = pud_alloc(&init_mm, pgdp, ea); map_kernel_page() 103 pmdp = pmd_alloc(&init_mm, pudp, ea); map_kernel_page() 106 ptep = pte_alloc_kernel(pmdp, ea); map_kernel_page() 109 set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, map_kernel_page() 113 pgdp = pgd_offset_k(ea); map_kernel_page() 121 pudp = pud_offset(pgdp, ea); map_kernel_page() 127 pmdp = pmd_offset(pudp, ea); map_kernel_page() 133 ptep = pte_offset_kernel(pmdp, ea); map_kernel_page() 134 set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, map_kernel_page() 143 if (htab_bolt_mapping(ea, ea + PAGE_SIZE, pa, flags, map_kernel_page() 171 void __iomem * __ioremap_at(phys_addr_t pa, void *ea, unsigned long size, __ioremap_at() argument 189 WARN_ON(((unsigned long)ea) & ~PAGE_MASK); __ioremap_at() 193 if (map_kernel_page((unsigned long)ea+i, pa+i, flags)) __ioremap_at() 196 return (void __iomem *)ea; __ioremap_at() 205 void __iounmap_at(void *ea, unsigned long size) __iounmap_at() argument 207 WARN_ON(((unsigned long)ea) & ~PAGE_MASK); __iounmap_at() 210 unmap_kernel_range((unsigned long)ea, size); __iounmap_at()
|
H A D | ppc_mmu_32.c | 165 void hash_preload(struct mm_struct *mm, unsigned long ea, hash_preload() argument 172 pmd = pmd_offset(pud_offset(pgd_offset(mm, ea), ea), ea); hash_preload() 174 add_hash_page(mm->context.id, ea, pmd_val(*pmd)); hash_preload()
|
H A D | hugetlbpage-hash64.c | 21 int __hash_page_huge(unsigned long ea, unsigned long access, unsigned long vsid, __hash_page_huge() argument 33 vpn = hpt_vpn(ea, vsid, ssize); __hash_page_huge() 116 hash_failure_debug(ea, access, vsid, trap, ssize, __hash_page_huge()
|
H A D | slb_low.S | 26 /* void slb_allocate_realmode(unsigned long ea); 36 * (ea & ~REGION_MASK) >= PGTABLE_RANGE 66 * context = (MAX_USER_CONTEXT) + ((ea >> 60) - 0xc) + 1 105 * context = (MAX_USER_CONTEXT) + ((ea >> 60) - 0xc) + 1 184 /* void slb_allocate_user(unsigned long ea);
|
H A D | mmu_decl.h | 85 extern void hash_preload(struct mm_struct *mm, unsigned long ea, 113 extern int map_kernel_page(unsigned long ea, unsigned long pa, int flags);
|
H A D | hugetlbpage.c | 971 pte_t *__find_linux_pte_or_hugepte(pgd_t *pgdir, unsigned long ea, __find_linux_pte_or_hugepte() argument 984 pgdp = pgdir + pgd_index(ea); __find_linux_pte_or_hugepte() 1006 pudp = pud_offset(&pgd, ea); __find_linux_pte_or_hugepte() 1018 pmdp = pmd_offset(&pud, ea); __find_linux_pte_or_hugepte() 1037 return pte_offset_kernel(&pmd, ea); __find_linux_pte_or_hugepte() 1043 ret_pte = hugepte_offset(*hpdp, ea, pdshift); __find_linux_pte_or_hugepte()
|
H A D | hash_native_64.c | 370 static void native_hpte_updateboltedpp(unsigned long newpp, unsigned long ea, native_hpte_updateboltedpp() argument 378 vsid = get_kernel_vsid(ea, ssize); native_hpte_updateboltedpp() 379 vpn = hpt_vpn(ea, vsid, ssize); native_hpte_updateboltedpp()
|
H A D | hash_low_64.S | 48 * _hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid, 341 /* _hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid, 419 * clrldi r3,r3,64 - SID_SHIFT --> ea & 0xfffffff 436 * clrldi r3,r3,64 - SID_SHIFT_1T --> ea & 0xffffffffff
|
/linux-4.1.27/fs/jfs/ |
H A D | xattr.c | 75 dxd_t new_ea; /* dxd to replace ea when modifying xattr */ 76 struct metapage *mp; /* metapage containing ea list */ 77 struct jfs_ea_list *xattr; /* buffer containing ea list */ 105 static int is_os2_xattr(struct jfs_ea *ea) is_os2_xattr() argument 107 return !is_known_namespace(ea->name); is_os2_xattr() 110 static inline int name_size(struct jfs_ea *ea) name_size() argument 112 if (is_os2_xattr(ea)) name_size() 113 return ea->namelen + XATTR_OS2_PREFIX_LEN; name_size() 115 return ea->namelen; name_size() 118 static inline int copy_name(char *buffer, struct jfs_ea *ea) copy_name() argument 120 int len = ea->namelen; copy_name() 122 if (is_os2_xattr(ea)) { copy_name() 127 memcpy(buffer, ea->name, ea->namelen); copy_name() 128 buffer[ea->namelen] = 0; copy_name() 148 * ea - dxd_t structure to be filled in with necessary EA information 153 * and sets <ea> fields appropriately. Otherwise, returns failure, EA will 159 int size, dxd_t * ea) ea_write_inline() 174 if (!(ji->mode2 & INLINEEA) && !(ji->ea.flag & DXD_INLINE)) ea_write_inline() 177 DXDsize(ea, size); ea_write_inline() 178 DXDlength(ea, 0); ea_write_inline() 179 DXDaddress(ea, 0); ea_write_inline() 181 ea->flag = DXD_INLINE; ea_write_inline() 184 ea->flag = 0; ea_write_inline() 185 DXDsize(ea, 0); ea_write_inline() 186 DXDlength(ea, 0); ea_write_inline() 187 DXDaddress(ea, 0); ea_write_inline() 190 if (ji->ea.flag & DXD_INLINE) ea_write_inline() 208 * ea - dxd_t structure to be filled in appropriately with where the 217 dxd_t * ea) ea_write() 235 if (!ea_write_inline(ip, ealist, size, ea)) ea_write() 302 ea->flag = DXD_EXTENT; ea_write() 303 DXDsize(ea, le32_to_cpu(ealist->size)); ea_write() 304 DXDlength(ea, nblocks); ea_write() 305 DXDaddress(ea, blkno); ea_write() 308 if (ji->ea.flag & DXD_INLINE) ea_write() 335 int ea_size = sizeDXD(&ji->ea); ea_read_inline() 343 if ((sizeDXD(&ji->ea) > sizeof (ji->i_inline_ea))) ea_read_inline() 380 if (ji->ea.flag & DXD_INLINE) ea_read() 383 nbytes = sizeDXD(&ji->ea); ea_read() 393 nblocks = lengthDXD(&ji->ea) << sbi->l2nbperpage; ea_read() 394 blkno = addressDXD(&ji->ea) << sbi->l2nbperpage; ea_read() 446 int ea_size = sizeDXD(&ji->ea); ea_get() 452 /* When fsck.jfs clears a bad ea, it doesn't clear the size */ ea_get() 453 if (ji->ea.flag == 0) ea_get() 475 } else if (ji->ea.flag & DXD_INLINE) { ea_get() 484 if (!(ji->ea.flag & DXD_EXTENT)) { ea_get() 485 jfs_error(sb, "invalid ea.flag\n"); ea_get() 560 ea_buf->mp = read_metapage(inode, addressDXD(&ji->ea), ea_get() 561 lengthDXD(&ji->ea) << sb->s_blocksize_bits, ea_get() 629 /* ->xattr must point to original ea's metapage */ ea_put() 638 if (ji->ea.flag & DXD_EXTENT) { ea_put() 639 invalidate_dxd_metapages(inode, ji->ea); ea_put() 640 old_blocks = lengthDXD(&ji->ea); ea_put() 644 txEA(tid, inode, &ji->ea, &ea_buf->new_ea); ea_put() 647 if (ji->ea.flag & DXD_INLINE) ea_put() 650 ji->ea = ea_buf->new_ea; ea_put() 652 txEA(tid, inode, &ji->ea, NULL); ea_put() 653 if (ji->ea.flag & DXD_INLINE) ea_put() 655 ji->ea.flag = 0; ea_put() 656 ji->ea.size = 0; ea_put() 701 struct jfs_ea *ea, *old_ea = NULL, *next_ea = NULL; __jfs_setxattr() local 735 for (ea = FIRST_EA(ealist); ea < END_EALIST(ealist); __jfs_setxattr() 736 ea = NEXT_EA(ea)) { __jfs_setxattr() 737 if ((namelen == ea->namelen) && __jfs_setxattr() 738 (memcmp(name, ea->name, namelen) == 0)) { __jfs_setxattr() 744 old_ea = ea; __jfs_setxattr() 745 old_ea_size = EA_SIZE(ea); __jfs_setxattr() 746 next_ea = NEXT_EA(ea); __jfs_setxattr() 748 new_size += EA_SIZE(ea); __jfs_setxattr() 767 * We need to allocate more space for merged ea list. __jfs_setxattr() 779 /* Remove old ea of the same name */ __jfs_setxattr() 791 /* Completely new ea list */ __jfs_setxattr() 807 ea = (struct jfs_ea *) ((char *) ealist + xattr_size); __jfs_setxattr() 808 ea->flag = 0; __jfs_setxattr() 809 ea->namelen = namelen; __jfs_setxattr() 810 ea->valuelen = (cpu_to_le16(value_len)); __jfs_setxattr() 811 memcpy(ea->name, name, namelen); __jfs_setxattr() 812 ea->name[namelen] = 0; __jfs_setxattr() 814 memcpy(&ea->name[namelen + 1], value, value_len); __jfs_setxattr() 815 xattr_size += EA_SIZE(ea); __jfs_setxattr() 829 * If we're left with an empty list, there's no ea __jfs_setxattr() 889 struct jfs_ea *ea; __jfs_getxattr() local 911 for (ea = FIRST_EA(ealist); ea < END_EALIST(ealist); ea = NEXT_EA(ea)) __jfs_getxattr() 912 if ((namelen == ea->namelen) && __jfs_getxattr() 913 memcmp(name, ea->name, namelen) == 0) { __jfs_getxattr() 915 size = le16_to_cpu(ea->valuelen); __jfs_getxattr() 922 value = ((char *) &ea->name) + ea->namelen + 1; __jfs_getxattr() 970 static inline int can_list(struct jfs_ea *ea) can_list() argument 972 return (strncmp(ea->name, XATTR_TRUSTED_PREFIX, can_list() 984 struct jfs_ea *ea; jfs_listxattr() local 1001 for (ea = FIRST_EA(ealist); ea < END_EALIST(ealist); ea = NEXT_EA(ea)) { jfs_listxattr() 1002 if (can_list(ea)) jfs_listxattr() 1003 size += name_size(ea) + 1; jfs_listxattr() 1016 for (ea = FIRST_EA(ealist); ea < END_EALIST(ealist); ea = NEXT_EA(ea)) { jfs_listxattr() 1017 if (can_list(ea)) { jfs_listxattr() 1018 int namelen = copy_name(buffer, ea); jfs_listxattr() 158 ea_write_inline(struct inode *ip, struct jfs_ea_list *ealist, int size, dxd_t * ea) ea_write_inline() argument 216 ea_write(struct inode *ip, struct jfs_ea_list *ealist, int size, dxd_t * ea) ea_write() argument
|
H A D | jfs_xattr.h | 36 struct jfs_ea ea[0]; /* Variable length list */ member in struct:jfs_ea_list 46 #define EA_SIZE(ea) \ 47 (sizeof (struct jfs_ea) + (ea)->namelen + 1 + \ 48 le16_to_cpu((ea)->valuelen)) 49 #define NEXT_EA(ea) ((struct jfs_ea *) (((char *) (ea)) + (EA_SIZE (ea)))) 50 #define FIRST_EA(ealist) ((ealist)->ea)
|
H A D | namei.c | 51 dxd_t *ea = &JFS_IP(inode)->ea; free_ea_wmap() local 53 if (ea->flag & DXD_EXTENT) { free_ea_wmap() 55 invalidate_dxd_metapages(inode, *ea); free_ea_wmap() 56 dbFree(inode, addressDXD(ea), lengthDXD(ea)); free_ea_wmap() 58 ea->flag = 0; free_ea_wmap() 408 if (JFS_IP(ip)->ea.flag & DXD_EXTENT) { jfs_rmdir() 410 txEA(tid, ip, &JFS_IP(ip)->ea, NULL); jfs_rmdir() 412 JFS_IP(ip)->ea.flag = 0; jfs_rmdir() 658 if (JFS_IP(ip)->ea.flag & DXD_EXTENT) commitZeroLink() 660 txEA(tid, ip, &JFS_IP(ip)->ea, NULL); commitZeroLink() 718 if (JFS_IP(ip)->ea.flag & DXD_EXTENT) { jfs_free_zero_link() 719 s64 xaddr = addressDXD(&JFS_IP(ip)->ea); jfs_free_zero_link() 720 int xlen = lengthDXD(&JFS_IP(ip)->ea); jfs_free_zero_link() 725 invalidate_dxd_metapages(ip, JFS_IP(ip)->ea); jfs_free_zero_link()
|
H A D | jfs_inode.c | 140 memset(&jfs_inode->ea, 0, sizeof(dxd_t)); ialloc()
|
H A D | jfs_incore.h | 45 dxd_t ea; /* dxd describing ea */ member in struct:jfs_inode_info
|
H A D | jfs_dinode.h | 66 dxd_t di_ea; /* 16: ea descriptor */
|
H A D | jfs_imap.c | 3096 jfs_ip->ea = dip->di_ea; copy_from_dinode() 3171 dip->di_ea = jfs_ip->ea; copy_to_dinode()
|
/linux-4.1.27/arch/powerpc/lib/ |
H A D | sstep.c | 38 extern int do_lfs(int rn, unsigned long ea); 39 extern int do_lfd(int rn, unsigned long ea); 40 extern int do_stfs(int rn, unsigned long ea); 41 extern int do_stfd(int rn, unsigned long ea); 42 extern int do_lvx(int rn, unsigned long ea); 43 extern int do_stvx(int rn, unsigned long ea); 44 extern int do_lxvd2x(int rn, unsigned long ea); 45 extern int do_stxvd2x(int rn, unsigned long ea); 84 static long __kprobes address_ok(struct pt_regs *regs, unsigned long ea, int nb) address_ok() argument 88 return __access_ok(ea, nb, USER_DS); address_ok() 97 unsigned long ea; dform_ea() local 100 ea = (signed short) instr; /* sign-extend */ dform_ea() 102 ea += regs->gpr[ra]; dform_ea() 104 return truncate_if_32bit(regs->msr, ea); dform_ea() 114 unsigned long ea; dsform_ea() local 117 ea = (signed short) (instr & ~3); /* sign-extend */ dsform_ea() 119 ea += regs->gpr[ra]; dsform_ea() 121 return truncate_if_32bit(regs->msr, ea); dsform_ea() 132 unsigned long ea; xform_ea() local 136 ea = regs->gpr[rb]; xform_ea() 138 ea += regs->gpr[ra]; xform_ea() 140 return truncate_if_32bit(regs->msr, ea); xform_ea() 172 static int __kprobes read_mem_aligned(unsigned long *dest, unsigned long ea, read_mem_aligned() argument 180 err = __get_user(x, (unsigned char __user *) ea); read_mem_aligned() 183 err = __get_user(x, (unsigned short __user *) ea); read_mem_aligned() 186 err = __get_user(x, (unsigned int __user *) ea); read_mem_aligned() 190 err = __get_user(x, (unsigned long __user *) ea); read_mem_aligned() 199 static int __kprobes read_mem_unaligned(unsigned long *dest, unsigned long ea, read_mem_unaligned() argument 215 c = max_align(ea); read_mem_unaligned() 219 err = read_mem_aligned(&b, ea, c); read_mem_unaligned() 223 ea += c; read_mem_unaligned() 247 * Read memory at address ea for nb bytes, return 0 for success 250 static int __kprobes read_mem(unsigned long *dest, unsigned long ea, int nb, read_mem() argument 253 if (!address_ok(regs, ea, nb)) read_mem() 255 if ((ea & (nb - 1)) == 0) read_mem() 256 return read_mem_aligned(dest, ea, nb); read_mem() 257 return read_mem_unaligned(dest, ea, nb, regs); read_mem() 260 static int __kprobes write_mem_aligned(unsigned long val, unsigned long ea, write_mem_aligned() argument 267 err = __put_user(val, (unsigned char __user *) ea); write_mem_aligned() 270 err = __put_user(val, (unsigned short __user *) ea); write_mem_aligned() 273 err = __put_user(val, (unsigned int __user *) ea); write_mem_aligned() 277 err = __put_user(val, (unsigned long __user *) ea); write_mem_aligned() 284 static int __kprobes write_mem_unaligned(unsigned long val, unsigned long ea, write_mem_unaligned() argument 311 c = max_align(ea); write_mem_unaligned() 315 err = write_mem_aligned(val >> (nb - c) * 8, ea, c); write_mem_unaligned() 318 ea += c; write_mem_unaligned() 324 * Write memory at address ea for nb bytes, return 0 for success 327 static int __kprobes write_mem(unsigned long val, unsigned long ea, int nb, write_mem() argument 330 if (!address_ok(regs, ea, nb)) write_mem() 332 if ((ea & (nb - 1)) == 0) write_mem() 333 return write_mem_aligned(val, ea, nb); write_mem() 334 return write_mem_unaligned(val, ea, nb, regs); write_mem() 343 unsigned long ea, int nb, do_fp_load() 363 if (!address_ok(regs, ea, nb)) do_fp_load() 365 if ((ea & 3) == 0) do_fp_load() 366 return (*func)(rn, ea); do_fp_load() 369 err = read_mem_unaligned(&data.ul[0], ea, nb, regs); do_fp_load() 374 err = read_mem_unaligned(&data.ul[0], ea, 4, regs); do_fp_load() 376 err = read_mem_unaligned(&data.ul[1], ea + 4, 4, regs); do_fp_load() 384 unsigned long ea, int nb, do_fp_store() 404 if (!address_ok(regs, ea, nb)) do_fp_store() 406 if ((ea & 3) == 0) do_fp_store() 407 return (*func)(rn, ea); do_fp_store() 415 err = write_mem_unaligned(data.ul[0], ea, nb, regs); do_fp_store() 421 err = write_mem_unaligned(data.ul[0], ea, 4, regs); do_fp_store() 423 err = write_mem_unaligned(data.ul[1], ea + 4, 4, regs); do_fp_store() 432 unsigned long ea, struct pt_regs *regs) do_vec_load() 434 if (!address_ok(regs, ea & ~0xfUL, 16)) do_vec_load() 436 return (*func)(rn, ea); do_vec_load() 440 unsigned long ea, struct pt_regs *regs) do_vec_store() 442 if (!address_ok(regs, ea & ~0xfUL, 16)) do_vec_store() 444 return (*func)(rn, ea); do_vec_store() 450 unsigned long ea, struct pt_regs *regs) do_vsx_load() 455 if (!address_ok(regs, ea, 16)) do_vsx_load() 457 if ((ea & 3) == 0) do_vsx_load() 458 return (*func)(rn, ea); do_vsx_load() 459 err = read_mem_unaligned(&val[0], ea, 8, regs); do_vsx_load() 461 err = read_mem_unaligned(&val[1], ea + 8, 8, regs); do_vsx_load() 468 unsigned long ea, struct pt_regs *regs) do_vsx_store() 473 if (!address_ok(regs, ea, 16)) do_vsx_store() 475 if ((ea & 3) == 0) do_vsx_store() 476 return (*func)(rn, ea); do_vsx_store() 480 err = write_mem_unaligned(val[0], ea, 8, regs); do_vsx_store() 482 err = write_mem_unaligned(val[1], ea + 8, 8, regs); do_vsx_store() 1300 op->ea = xform_ea(instr, regs); analyse_instr() 1305 op->ea = xform_ea(instr, regs); analyse_instr() 1310 op->ea = xform_ea(instr, regs); analyse_instr() 1316 op->ea = xform_ea(instr, regs); analyse_instr() 1322 op->ea = xform_ea(instr, regs); analyse_instr() 1340 op->ea = xform_ea(instr, regs); analyse_instr() 1448 op->ea = 0; analyse_instr() 1450 op->ea = truncate_if_32bit(regs->msr, analyse_instr() 1504 op->ea = 0; analyse_instr() 1506 op->ea = truncate_if_32bit(regs->msr, analyse_instr() 1543 op->ea = dform_ea(instr, regs); analyse_instr() 1549 op->ea = dform_ea(instr, regs); analyse_instr() 1555 op->ea = dform_ea(instr, regs); analyse_instr() 1561 op->ea = dform_ea(instr, regs); analyse_instr() 1567 op->ea = dform_ea(instr, regs); analyse_instr() 1573 op->ea = dform_ea(instr, regs); analyse_instr() 1579 op->ea = dform_ea(instr, regs); analyse_instr() 1586 op->ea = dform_ea(instr, regs); analyse_instr() 1591 op->ea = dform_ea(instr, regs); analyse_instr() 1600 op->ea = dform_ea(instr, regs); analyse_instr() 1608 op->ea = dform_ea(instr, regs); analyse_instr() 1616 op->ea = dform_ea(instr, regs); analyse_instr() 1624 op->ea = dform_ea(instr, regs); analyse_instr() 1630 op->ea = dsform_ea(instr, regs); analyse_instr() 1645 op->ea = dsform_ea(instr, regs); analyse_instr() 1711 static __kprobes int handle_stack_update(unsigned long ea, struct pt_regs *regs) handle_stack_update() argument 1717 if (ea - STACK_INT_FRAME_SIZE <= current->thread.ksp_limit) { handle_stack_update() 1783 if (!address_ok(regs, op.ea, 8)) emulate_step() 1787 __cacheop_user_asmx(op.ea, err, "dcbst"); emulate_step() 1790 __cacheop_user_asmx(op.ea, err, "dcbf"); emulate_step() 1794 prefetchw((void *) op.ea); emulate_step() 1798 prefetch((void *) op.ea); emulate_step() 1801 __cacheop_user_asmx(op.ea, err, "icbi"); emulate_step() 1811 if (op.ea & (size - 1)) emulate_step() 1814 if (!address_ok(regs, op.ea, size)) emulate_step() 1819 __get_user_asmx(val, op.ea, err, "lwarx"); emulate_step() 1822 __get_user_asmx(val, op.ea, err, "ldarx"); emulate_step() 1834 if (op.ea & (size - 1)) emulate_step() 1837 if (!address_ok(regs, op.ea, size)) emulate_step() 1842 __put_user_asmx(op.val, op.ea, err, "stwcx.", cr); emulate_step() 1845 __put_user_asmx(op.val, op.ea, err, "stdcx.", cr); emulate_step() 1859 err = read_mem(®s->gpr[op.reg], op.ea, size, regs); emulate_step() 1873 err = do_fp_load(op.reg, do_lfs, op.ea, size, regs); emulate_step() 1875 err = do_fp_load(op.reg, do_lfd, op.ea, size, regs); emulate_step() 1882 err = do_vec_load(op.reg, do_lvx, op.ea & ~0xfUL, regs); emulate_step() 1889 err = do_vsx_load(op.reg, do_lxvd2x, op.ea, regs); emulate_step() 1900 err = read_mem(®s->gpr[rd], op.ea, nb, regs); emulate_step() 1905 op.ea += 4; emulate_step() 1916 op.ea >= regs->gpr[1] - STACK_INT_FRAME_SIZE) { emulate_step() 1917 err = handle_stack_update(op.ea, regs); emulate_step() 1920 err = write_mem(op.val, op.ea, size, regs); emulate_step() 1928 err = do_fp_store(op.reg, do_stfs, op.ea, size, regs); emulate_step() 1930 err = do_fp_store(op.reg, do_stfd, op.ea, size, regs); emulate_step() 1937 err = do_vec_store(op.reg, do_stvx, op.ea & ~0xfUL, regs); emulate_step() 1944 err = do_vsx_store(op.reg, do_stxvd2x, op.ea, regs); emulate_step() 1958 err = write_mem(val, op.ea, nb, regs); emulate_step() 1961 op.ea += 4; emulate_step() 2010 regs->gpr[op.update_reg] = op.ea; emulate_step() 342 do_fp_load(int rn, int (*func)(int, unsigned long), unsigned long ea, int nb, struct pt_regs *regs) do_fp_load() argument 383 do_fp_store(int rn, int (*func)(int, unsigned long), unsigned long ea, int nb, struct pt_regs *regs) do_fp_store() argument 431 do_vec_load(int rn, int (*func)(int, unsigned long), unsigned long ea, struct pt_regs *regs) do_vec_load() argument 439 do_vec_store(int rn, int (*func)(int, unsigned long), unsigned long ea, struct pt_regs *regs) do_vec_store() argument 449 do_vsx_load(int rn, int (*func)(int, unsigned long), unsigned long ea, struct pt_regs *regs) do_vsx_load() argument 467 do_vsx_store(int rn, int (*func)(int, unsigned long), unsigned long ea, struct pt_regs *regs) do_vsx_store() argument
|
/linux-4.1.27/arch/nios2/mm/ |
H A D | extable.c | 18 fixup = search_exception_tables(regs->ea); fixup_exception() 20 regs->ea = fixup->fixup; fixup_exception()
|
H A D | fault.c | 55 regs->ea -= 4; do_page_fault() 87 if (!user_mode(regs) && !search_exception_tables(regs->ea)) do_page_fault() 211 pr_alert("ea = %08lx, ra = %08lx, cause = %ld\n", regs->ea, regs->ra, do_page_fault()
|
/linux-4.1.27/arch/nios2/kernel/ |
H A D | traps.c | 125 fp->ea -= 4; breakpoint_c() 126 _exception(SIGTRAP, fp, TRAP_BRKPT, fp->ea); breakpoint_c() 136 fp->ea -= 4; handle_unaligned_c() 146 pr_alert(" op-code 0x%08lx\n", *(unsigned long *)(fp->ea)); handle_unaligned_c() 158 fp->ea -= 4; handle_illegal_c() 159 _exception(SIGILL, fp, ILL_ILLOPC, fp->ea); handle_illegal_c() 165 fp->ea -= 4; handle_supervisor_instr() 166 _exception(SIGILL, fp, ILL_PRVOPC, fp->ea); handle_supervisor_instr() 172 fp->ea -= 4; handle_diverror_c() 173 _exception(SIGFPE, fp, FPE_INTDIV, fp->ea); handle_diverror_c() 186 regs->ea -= 4; unhandled_exception() 189 pr_emerg("opcode: 0x%08lx\n", *(unsigned long *)(regs->ea)); unhandled_exception() 194 _send_sig(SIGUSR1, 0, fp->ea); handle_trap_1_c() 199 _send_sig(SIGUSR2, 0, fp->ea); handle_trap_2_c() 204 _send_sig(SIGILL, ILL_ILLTRP, fp->ea); handle_trap_3_c()
|
H A D | misaligned.c | 78 fp->ea -= 4; handle_unaligned_c() 87 isn = *(unsigned long *)(fp->ea); handle_unaligned_c() 171 fp->ea, (unsigned int)addr, handle_unaligned_c() 175 fp->ea, handle_unaligned_c() 179 _exception(SIGSEGV, fp, SEGV_MAPERR, fp->ea); handle_unaligned_c() 190 fp->ea += 4; handle_unaligned_c() 194 fp->ea, handle_unaligned_c() 211 pr_err("user unaligned access @ %#lx; isn=0x%08lx ea=0x%08lx ra=0x%08lx sp=0x%08lx\n", handle_unaligned_c() 213 fp->ea, fp->ra, fp->sp); handle_unaligned_c() 217 _exception(SIGBUS, fp, BUS_ADRALN, fp->ea); handle_unaligned_c() 219 fp->ea += 4; /* else advance */ handle_unaligned_c()
|
H A D | kgdb.c | 62 { "ea", GDB_SIZEOF_REG, -1 }, 65 { "pc", GDB_SIZEOF_REG, offsetof(struct pt_regs, ea) }, 114 gdb_regs[GDB_PC] = p->thread.kregs->ea; sleeping_thread_to_gdb_regs() 119 regs->ea = pc; kgdb_arch_set_pc() 135 regs->ea = addr; kgdb_arch_handle_exception() 150 regs->ea -= 4; kgdb_breakpoint_c()
|
H A D | ptrace.c | 53 REG_O_ONE(®s->ea, PTR_EA); genregs_get() 56 REG_O_ONE(®s->ea, PTR_PC); /* use ea for PC */ genregs_get() 99 REG_IN_ONE(®s->ea, PTR_EA); genregs_set() 102 REG_IN_ONE(®s->ea, PTR_PC); /* use ea for PC */ genregs_set()
|
H A D | signal.c | 83 err |= __get_user(regs->ea, &gregs[27]); rt_restore_ucontext() 160 err |= __put_user(regs->ea, &gregs[27]); rt_setup_ucontext() 210 regs->ea = (unsigned long) ksig->ka.sa.sa_handler; setup_rt_frame() 244 continue_addr = regs->ea; do_signal() 261 regs->ea = restart_addr; do_signal() 268 if (unlikely(restart && regs->ea == restart_addr)) { do_signal() 275 regs->ea = continue_addr; do_signal() 285 if (unlikely(restart) && regs->ea == restart_addr) { do_signal() 286 regs->ea = continue_addr; do_signal()
|
H A D | entry.S | 51 * ea-4 = address of interrupted insn (ea must be preserved). 54 * If ea <= cmpxchg_stw and ea > cmpxchg_ldw then saved EA is set to 59 bgtu ea, et, 1f 62 bltu ea, et, 1f 64 mov ea, et 164 ldwio r24, -4(ea) /* instruction that caused the exception */ 346 addi ea, ea, -4 /* re-issue the interrupted instruction */ 347 stw ea, PT_EA(sp)
|
H A D | process.c | 91 pr_notice("ea: %08lx estatus: %08lx\n", show_regs() 92 regs->ea, regs->estatus); show_regs() 169 pr_emerg("PC: %08lx\n", fp->ea); dump() 188 tp = ((unsigned char *) fp->ea) - 0x20; dump() 247 regs->ea = pc; start_thread()
|
H A D | head.S | 107 subi ea, ea, 4
|
H A D | asm-offsets.c | 57 OFFSET(PT_EA, pt_regs, ea); main()
|
H A D | insnemu.S | 53 ldw ea, PT_EA(sp) 74 * the exception return address register (ea) as a destination because 88 * The instruction that causes the exception is found at address -4(ea). 158 stw ea, 116(sp) 166 ldw r2,-4(ea) /* r2 = AAAAA,BBBBB,IIIIIIIIIIIIIIII,PPPPPP */ 586 ldw ea, 116(sp)
|
/linux-4.1.27/drivers/misc/cxl/ |
H A D | fault.c | 88 u64 ea) cxl_fault_segment() 93 if (!(rc = copro_calculate_slb(mm, ea, &slb))) { cxl_fault_segment() 116 struct mm_struct *mm, u64 ea) cxl_handle_segment_miss() 120 pr_devel("CXL interrupt: Segment fault pe: %i ea: %#llx\n", ctx->pe, ea); cxl_handle_segment_miss() 121 trace_cxl_ste_miss(ctx, ea); cxl_handle_segment_miss() 123 if ((rc = cxl_fault_segment(ctx, mm, ea))) cxl_handle_segment_miss() 222 static void cxl_prefault_one(struct cxl_context *ctx, u64 ea) cxl_prefault_one() argument 240 rc = cxl_fault_segment(ctx, mm, ea); cxl_prefault_one() 246 static u64 next_segment(u64 ea, u64 vsid) next_segment() argument 249 ea |= (1ULL << 40) - 1; next_segment() 251 ea |= (1ULL << 28) - 1; next_segment() 253 return ea + 1; next_segment() 258 u64 ea, last_esid = 0; cxl_prefault_vma() local 278 for (ea = vma->vm_start; ea < vma->vm_end; cxl_prefault_vma() 279 ea = next_segment(ea, slb.vsid)) { cxl_prefault_vma() 280 rc = copro_calculate_slb(mm, ea, &slb); cxl_prefault_vma() 87 cxl_fault_segment(struct cxl_context *ctx, struct mm_struct *mm, u64 ea) cxl_fault_segment() argument 115 cxl_handle_segment_miss(struct cxl_context *ctx, struct mm_struct *mm, u64 ea) cxl_handle_segment_miss() argument
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | copro.h | 18 int copro_handle_mm_fault(struct mm_struct *mm, unsigned long ea, 21 int copro_calculate_slb(struct mm_struct *mm, u64 ea, struct copro_slb *slb);
|
H A D | mmu-hash64.h | 287 static inline unsigned long hpt_vpn(unsigned long ea, hpt_vpn() argument 294 return (vsid << (s_shift - VPN_SHIFT)) | ((ea >> VPN_SHIFT) & mask); hpt_vpn() 323 extern int __hash_page_4K(unsigned long ea, unsigned long access, 326 extern int __hash_page_64K(unsigned long ea, unsigned long access, 331 extern int hash_page_mm(struct mm_struct *mm, unsigned long ea, 334 extern int hash_page(unsigned long ea, unsigned long access, unsigned long trap, 336 int __hash_page_huge(unsigned long ea, unsigned long access, unsigned long vsid, 340 extern int __hash_page_thp(unsigned long ea, unsigned long access, 344 static inline int __hash_page_thp(unsigned long ea, unsigned long access, __hash_page_thp() argument 353 extern void hash_failure_debug(unsigned long ea, unsigned long access, 571 static inline unsigned long get_vsid(unsigned long context, unsigned long ea, get_vsid() argument 577 if ((ea & ~REGION_MASK) >= PGTABLE_RANGE) get_vsid() 582 | (ea >> SID_SHIFT), 256M); get_vsid() 584 | (ea >> SID_SHIFT_1T), 1T); get_vsid() 596 static inline unsigned long get_kernel_vsid(unsigned long ea, int ssize) get_kernel_vsid() argument 603 context = (MAX_USER_CONTEXT) + ((ea >> 60) - 0xc) + 1; get_kernel_vsid() 604 return get_vsid(context, ea, ssize); get_kernel_vsid()
|
H A D | sstep.h | 82 unsigned long ea; member in struct:instruction_op
|
H A D | kvm_ppc.h | 665 ulong ea; kvmppc_get_ea_indexed() local 668 ea = kvmppc_get_gpr(vcpu, rb); kvmppc_get_ea_indexed() 670 ea += kvmppc_get_gpr(vcpu, ra); kvmppc_get_ea_indexed() 679 ea = (uint32_t)ea; kvmppc_get_ea_indexed() 681 return ea; kvmppc_get_ea_indexed()
|
H A D | pci-bridge.h | 257 unsigned long ea = (unsigned long)address; isa_vaddr_is_ioport() local 258 return ea >= ISA_IO_BASE && ea < ISA_IO_END; isa_vaddr_is_ioport()
|
H A D | pgtable.h | 250 pte_t *__find_linux_pte_or_hugepte(pgd_t *pgdir, unsigned long ea, 252 static inline pte_t *find_linux_pte_or_hugepte(pgd_t *pgdir, unsigned long ea, find_linux_pte_or_hugepte() argument 259 return __find_linux_pte_or_hugepte(pgdir, ea, shift); find_linux_pte_or_hugepte()
|
H A D | hugetlb.h | 86 void book3e_hugetlb_preload(struct vm_area_struct *vma, unsigned long ea,
|
H A D | machdep.h | 47 unsigned long ea, 57 void (*hpte_removebolted)(unsigned long ea,
|
H A D | pgtable-ppc64.h | 77 #define REGION_ID(ea) (((unsigned long)(ea)) >> REGION_SHIFT)
|
H A D | kvm_book3s.h | 117 extern void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong ea, ulong ea_mask);
|
H A D | io.h | 734 extern void __iomem * __ioremap_at(phys_addr_t pa, void *ea, 736 extern void __iounmap_at(void *ea, unsigned long size);
|
/linux-4.1.27/sound/pci/echoaudio/ |
H A D | mona.c | 66 MODULE_FIRMWARE("ea/loader_dsp.fw"); 67 MODULE_FIRMWARE("ea/mona_301_dsp.fw"); 68 MODULE_FIRMWARE("ea/mona_361_dsp.fw"); 69 MODULE_FIRMWARE("ea/mona_301_1_asic_48.fw"); 70 MODULE_FIRMWARE("ea/mona_301_1_asic_96.fw"); 71 MODULE_FIRMWARE("ea/mona_361_1_asic_48.fw"); 72 MODULE_FIRMWARE("ea/mona_361_1_asic_96.fw"); 73 MODULE_FIRMWARE("ea/mona_2_asic.fw");
|
H A D | gina24.c | 68 MODULE_FIRMWARE("ea/loader_dsp.fw"); 69 MODULE_FIRMWARE("ea/gina24_301_dsp.fw"); 70 MODULE_FIRMWARE("ea/gina24_361_dsp.fw"); 71 MODULE_FIRMWARE("ea/gina24_301_asic.fw"); 72 MODULE_FIRMWARE("ea/gina24_361_asic.fw");
|
H A D | layla24.c | 70 MODULE_FIRMWARE("ea/loader_dsp.fw"); 71 MODULE_FIRMWARE("ea/layla24_dsp.fw"); 72 MODULE_FIRMWARE("ea/layla24_1_asic.fw"); 73 MODULE_FIRMWARE("ea/layla24_2A_asic.fw"); 74 MODULE_FIRMWARE("ea/layla24_2S_asic.fw");
|
H A D | echo3g.c | 70 MODULE_FIRMWARE("ea/loader_dsp.fw"); 71 MODULE_FIRMWARE("ea/echo3g_dsp.fw"); 72 MODULE_FIRMWARE("ea/3g_asic.fw");
|
H A D | indigo.c | 60 MODULE_FIRMWARE("ea/loader_dsp.fw"); 61 MODULE_FIRMWARE("ea/indigo_dsp.fw");
|
H A D | indigodj.c | 60 MODULE_FIRMWARE("ea/loader_dsp.fw"); 61 MODULE_FIRMWARE("ea/indigo_dj_dsp.fw");
|
H A D | indigodjx.c | 60 MODULE_FIRMWARE("ea/loader_dsp.fw"); 61 MODULE_FIRMWARE("ea/indigo_djx_dsp.fw");
|
H A D | indigoio.c | 61 MODULE_FIRMWARE("ea/loader_dsp.fw"); 62 MODULE_FIRMWARE("ea/indigo_io_dsp.fw");
|
H A D | indigoiox.c | 61 MODULE_FIRMWARE("ea/loader_dsp.fw"); 62 MODULE_FIRMWARE("ea/indigo_iox_dsp.fw");
|
H A D | layla20.c | 68 MODULE_FIRMWARE("ea/layla20_dsp.fw"); 69 MODULE_FIRMWARE("ea/layla20_asic.fw");
|
H A D | mia.c | 69 MODULE_FIRMWARE("ea/loader_dsp.fw"); 70 MODULE_FIRMWARE("ea/mia_dsp.fw");
|
H A D | darla20.c | 58 MODULE_FIRMWARE("ea/darla20_dsp.fw");
|
H A D | darla24.c | 62 MODULE_FIRMWARE("ea/darla24_dsp.fw");
|
H A D | gina20.c | 62 MODULE_FIRMWARE("ea/gina20_dsp.fw");
|
H A D | echoaudio_dsp.h | 658 /* Positions for ea. ch. 0x160 32*4 */
|
H A D | echoaudio.c | 61 snprintf(name, sizeof(name), "ea/%s", card_fw[fw_index].data); get_firmware()
|
/linux-4.1.27/fs/udf/ |
H A D | misc.c | 50 uint8_t *ea = NULL, *ad = NULL; udf_add_extendedattr() local 55 ea = iinfo->i_ext.i_data; udf_add_extendedattr() 59 ad = ea; udf_add_extendedattr() 70 eahd = (struct extendedAttrHeaderDesc *)ea; udf_add_extendedattr() 107 memmove(&ea[offset - aal + size], udf_add_extendedattr() 108 &ea[aal], offset - aal); udf_add_extendedattr() 117 memmove(&ea[offset - ial + size], udf_add_extendedattr() 118 &ea[ial], offset - ial); udf_add_extendedattr() 128 memmove(&ea[offset - aal + size], udf_add_extendedattr() 129 &ea[aal], offset - aal); udf_add_extendedattr() 142 return (struct genericFormat *)&ea[offset]; udf_add_extendedattr() 154 uint8_t *ea = NULL; udf_get_extendedattr() local 158 ea = iinfo->i_ext.i_data; udf_get_extendedattr() 162 eahd = (struct extendedAttrHeaderDesc *)ea; udf_get_extendedattr() 179 gaf = (struct genericFormat *)&ea[offset]; udf_get_extendedattr()
|
H A D | namei.c | 859 uint8_t *ea; udf_symlink() local 910 ea = epos.bh->b_data + udf_ext0_offset(inode); udf_symlink() 912 ea = iinfo->i_ext.i_data + iinfo->i_lenEAttr; udf_symlink() 915 pc = (struct pathComponent *)ea; udf_symlink() 934 pc = (struct pathComponent *)(ea + elen); udf_symlink()
|
/linux-4.1.27/arch/powerpc/platforms/cell/spufs/ |
H A D | fault.c | 37 unsigned long ea, int type) spufs_handle_event() 56 info.si_addr = (void __user *)ea; spufs_handle_event() 112 u64 ea, dsisr, access; spufs_handle_class1() local 126 ea = ctx->csa.class_1_dar; spufs_handle_class1() 134 pr_debug("ctx %p: ea %016llx, dsisr %016llx state %d\n", ctx, ea, spufs_handle_class1() 147 ret = hash_page(ea, access, 0x300, dsisr); spufs_handle_class1() 152 ret = copro_handle_mm_fault(current->mm, ea, dsisr, &flt); spufs_handle_class1() 187 spufs_handle_event(ctx, ea, SPE_EVENT_SPE_DATA_STORAGE); spufs_handle_class1() 36 spufs_handle_event(struct spu_context *ctx, unsigned long ea, int type) spufs_handle_event() argument
|
H A D | hw_ops.c | 296 out_be64(&prob->mfc_ea_W, cmd->ea); spu_hw_send_mfc_command()
|
H A D | spufs.h | 178 uint64_t ea; /* effective address */ member in struct:mfc_dma_command
|
H A D | file.c | 1661 cmd->ea, cmd->size, cmd->tag, cmd->cmd); spufs_check_valid_dma() 1676 if ((cmd->lsa & 0xf) != (cmd->ea &0xf)) { spufs_check_valid_dma() 1677 pr_debug("invalid DMA alignment, ea %llx lsa %x\n", spufs_check_valid_dma() 1678 cmd->ea, cmd->lsa); spufs_check_valid_dma()
|
H A D | switch.c | 778 static inline int send_mfc_dma(struct spu *spu, unsigned long ea, send_mfc_dma() argument 797 out_be64(&prob->mfc_ea_W, ea); send_mfc_dma() 806 ea += transfer_size; send_mfc_dma()
|
/linux-4.1.27/arch/mips/sgi-ip32/ |
H A D | ip32-setup.c | 49 static inline void str2eaddr(unsigned char *ea, unsigned char *str) str2eaddr() argument 60 ea[i] = num; str2eaddr()
|
/linux-4.1.27/arch/powerpc/kvm/ |
H A D | e500_emulate.c | 123 gva_t ea; kvmppc_core_emulate_op_e500() local 152 ea = kvmppc_get_ea_indexed(vcpu, ra, rb); kvmppc_core_emulate_op_e500() 153 emulated = kvmppc_e500_emul_tlbsx(vcpu, ea); kvmppc_core_emulate_op_e500() 158 ea = kvmppc_get_ea_indexed(vcpu, ra, rb); kvmppc_core_emulate_op_e500() 159 emulated = kvmppc_e500_emul_tlbilx(vcpu, type, ea); kvmppc_core_emulate_op_e500() 164 ea = kvmppc_get_ea_indexed(vcpu, ra, rb); kvmppc_core_emulate_op_e500() 165 emulated = kvmppc_e500_emul_tlbivax(vcpu, ea); kvmppc_core_emulate_op_e500()
|
H A D | trace_hv.h | 275 struct kvm_memory_slot *memslot, unsigned long ea, 278 TP_ARGS(vcpu, hptep, memslot, ea, dsisr), 285 __field(unsigned long, ea) 296 __entry->ea = ea; 302 TP_printk("VCPU %d: hpte=0x%lx:0x%lx guest=0x%lx ea=0x%lx,%x slot=0x%llx,0x%x", 305 __entry->ea, __entry->dsisr,
|
H A D | e500_mmu.c | 248 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea) kvmppc_e500_emul_tlbivax() argument 254 ia = (ea >> 2) & 0x1; kvmppc_e500_emul_tlbivax() 257 tlbsel = (ea >> 3) & 0x1; kvmppc_e500_emul_tlbivax() 265 ea &= 0xfffff000; kvmppc_e500_emul_tlbivax() 266 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel, kvmppc_e500_emul_tlbivax() 296 gva_t ea) tlbilx_one() 301 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel, pid, -1); tlbilx_one() 310 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea) kvmppc_e500_emul_tlbilx() argument 319 tlbilx_one(vcpu_e500, pid, ea); kvmppc_e500_emul_tlbilx() 344 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea) kvmppc_e500_emul_tlbsx() argument 353 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel, pid, as); kvmppc_e500_emul_tlbsx() 295 tlbilx_one(struct kvmppc_vcpu_e500 *vcpu_e500, int pid, gva_t ea) tlbilx_one() argument
|
H A D | book3s_64_mmu.c | 454 static void kvmppc_mmu_book3s_64_slbie(struct kvm_vcpu *vcpu, u64 ea) kvmppc_mmu_book3s_64_slbie() argument 459 dprintk("KVM MMU: slbie(0x%llx)\n", ea); kvmppc_mmu_book3s_64_slbie() 461 slbe = kvmppc_mmu_book3s_64_find_slbe(vcpu, ea); kvmppc_mmu_book3s_64_slbie() 466 dprintk("KVM MMU: slbie(0x%llx, 0x%llx)\n", ea, slbe->esid); kvmppc_mmu_book3s_64_slbie() 473 kvmppc_mmu_flush_segment(vcpu, ea & ~(seg_size - 1), seg_size); kvmppc_mmu_book3s_64_slbie() 580 ulong ea = esid << SID_SHIFT; kvmppc_mmu_book3s_64_esid_to_vsid() local 588 slb = kvmppc_mmu_book3s_64_find_slbe(vcpu, ea); kvmppc_mmu_book3s_64_esid_to_vsid()
|
H A D | book3s_32_mmu.c | 353 static void kvmppc_mmu_book3s_32_tlbie(struct kvm_vcpu *vcpu, ulong ea, bool large) kvmppc_mmu_book3s_32_tlbie() argument 360 kvmppc_mmu_pte_flush(v, ea, 0x0FFFF000); kvmppc_mmu_book3s_32_tlbie() 366 ulong ea = esid << SID_SHIFT; kvmppc_mmu_book3s_32_esid_to_vsid() local 372 sr = find_sr(vcpu, ea); kvmppc_mmu_book3s_32_esid_to_vsid()
|
H A D | e500.h | 132 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea); 133 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea); 134 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea);
|
H A D | book3s_64_mmu_host.c | 116 /* and write the mapping ea -> hpa into the pt */ kvmppc_mmu_map_page() 354 void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong ea, ulong seg_size) kvmppc_mmu_flush_segment() argument 362 (svcpu->slb[i].esid & seg_mask) == ea) { kvmppc_mmu_flush_segment()
|
H A D | trace_pr.h | 88 TP_printk("Map: hvpn=%llx pfn=%llx ea=%lx vp=%llx ra=%lx [%x]", 117 TP_printk("Flush: hva=%llx pfn=%llx ea=%lx vp=%llx ra=%lx [%x]",
|
H A D | book3s_64_mmu_hv.c | 299 unsigned long ea) kvmppc_mmu_get_real_addr() 304 return (r & HPTE_R_RPN & ~ra_mask) | (ea & ra_mask); kvmppc_mmu_get_real_addr() 391 unsigned long gpa, gva_t ea, int is_store) kvmppc_hv_emulate_mmio() 431 vcpu->arch.vaddr_accessed = ea; kvmppc_hv_emulate_mmio() 436 unsigned long ea, unsigned long dsisr) kvmppc_book3s_hv_page_fault() 460 if (ea != vcpu->arch.pgfault_addr) kvmppc_book3s_hv_page_fault() 482 gpa = gpa_base | (ea & (psize - 1)); kvmppc_book3s_hv_page_fault() 486 trace_kvm_page_fault_enter(vcpu, hpte, memslot, ea, dsisr); kvmppc_book3s_hv_page_fault() 490 return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea, kvmppc_book3s_hv_page_fault() 298 kvmppc_mmu_get_real_addr(unsigned long v, unsigned long r, unsigned long ea) kvmppc_mmu_get_real_addr() argument 390 kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, unsigned long gpa, gva_t ea, int is_store) kvmppc_hv_emulate_mmio() argument 435 kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu, unsigned long ea, unsigned long dsisr) kvmppc_book3s_hv_page_fault() argument
|
H A D | book3s_32_mmu_host.c | 168 /* and write the mapping ea -> hpa into the pt */ kvmppc_mmu_map_page()
|
/linux-4.1.27/arch/nios2/include/asm/ |
H A D | processor.h | 84 #define thread_saved_pc(tsk) ((tsk)->thread.kregs->ea) 92 #define KSTK_EIP(tsk) ((tsk)->thread.kregs->ea)
|
H A D | entry.h | 61 stw ea, PT_EA(sp) variable 85 ldw ea, PT_EA(sp) variable
|
H A D | ptrace.h | 44 unsigned long ea; /* Exception return address (pc) */ member in struct:pt_regs
|
H A D | elf.h | 72 pr_reg[21] = regs->ea; \
|
/linux-4.1.27/arch/m68k/fpsp040/ |
H A D | x_unsupp.S | 16 | if fmove.p fpm,<ea> then pack source (fpm) 17 | and store in users memory <ea>.
|
H A D | x_fline.S | 9 | <ea> field. If so, handle here and return. Otherwise, it 42 | fmovecr with non-zero <ea> handling here
|
H A D | x_snan.S | 144 movel EXC_EA(%a6),%a0 |get <ea> from exc frame 173 tstl %a0 |check if <ea> is 0 195 tstl %a0 |check if <ea> is 0 217 tstl %a0 |check if <ea> is 0
|
H A D | gen_except.S | 95 | Frame is an unimp frame possible resulting from an fmove <ea>,fp0 325 | by doing an fmove <ea>,fp0. The exception operand, in
|
H A D | x_operr.S | 272 | Destination is memory. Get <ea> from integer exception frame
|
H A D | res_func.S | 10 | For packed move out (fmove.p fpm,<ea>) the operation is
|
/linux-4.1.27/arch/m68k/math-emu/ |
H A D | fp_decode.h | 91 | .long "f<op> <ea>,fpx","fmove fpx,<ea>" 92 | .long "fmovem <ea>,fpcr","fmovem <ea>,fpx" 93 | .long "fmovem fpcr,<ea>","fmovem fpx,<ea>" 102 | decode destination format for fmove reg,ea 107 | decode source register for fmove reg,ea
|
H A D | fp_scan.S | 104 fp_fgen_ea: | source is <ea>
|
/linux-4.1.27/arch/powerpc/platforms/cell/ |
H A D | spu_base.c | 164 static int __spu_trap_data_seg(struct spu *spu, unsigned long ea) __spu_trap_data_seg() argument 169 ret = copro_calculate_slb(spu->mm, ea, &slb); __spu_trap_data_seg() 184 extern int hash_page(unsigned long ea, unsigned long access, 186 static int __spu_trap_data_map(struct spu *spu, unsigned long ea, u64 dsisr) __spu_trap_data_map() argument 190 pr_debug("%s, %llx, %lx\n", __func__, dsisr, ea); __spu_trap_data_map() 197 (REGION_ID(ea) != USER_REGION_ID)) { __spu_trap_data_map() 200 ret = hash_page(ea, _PAGE_PRESENT, 0x300, dsisr); __spu_trap_data_map() 209 spu->class_1_dar = ea; __spu_trap_data_map() 222 unsigned long ea = (unsigned long)addr; __spu_kernel_slb() local 225 if (REGION_ID(ea) == KERNEL_REGION_ID) __spu_kernel_slb() 230 slb->vsid = (get_kernel_vsid(ea, MMU_SEGSIZE_256M) << SLB_VSID_SHIFT) | __spu_kernel_slb() 232 slb->esid = (ea & ESID_MASK) | SLB_ESID_V; __spu_kernel_slb() 242 unsigned long ea = (unsigned long)new_addr; __slb_present() local 246 if (!((slbs[i].esid ^ ea) & ESID_MASK)) __slb_present()
|
/linux-4.1.27/net/appletalk/ |
H A D | aarp.c | 722 struct elapaarp *ea = aarp_hdr(skb); aarp_rcv() local 737 if (!skb_pull(skb, sizeof(*ea))) aarp_rcv() 740 function = ntohs(ea->function); aarp_rcv() 744 ea->hw_len != ETH_ALEN || ea->pa_len != AARP_PA_ALEN || aarp_rcv() 745 ea->pa_src_zero || ea->pa_dst_zero) aarp_rcv() 749 hash = ea->pa_src_node % (AARP_HASH_SIZE - 1); aarp_rcv() 752 sa.s_node = ea->pa_src_node; aarp_rcv() 753 sa.s_net = ea->pa_src_net; aarp_rcv() 761 ifa->address.s_node == ea->pa_dst_node && aarp_rcv() 762 ifa->address.s_net == ea->pa_dst_net) { aarp_rcv() 768 da.s_node = ea->pa_dst_node; aarp_rcv() 769 da.s_net = ea->pa_dst_net; aarp_rcv() 794 ether_addr_copy(a->hwaddr, ea->hw_src); aarp_rcv() 816 sa.s_node = ea->pa_dst_node; aarp_rcv() 817 sa.s_net = ea->pa_dst_net; aarp_rcv() 858 sa.s_node = ea->pa_src_node; aarp_rcv() 859 sa.s_net = ea->pa_src_net; aarp_rcv() 863 aarp_send_reply(dev, ma, &sa, ea->hw_src); aarp_rcv()
|
/linux-4.1.27/arch/mips/alchemy/common/ |
H A D | prom.c | 92 static inline void str2eaddr(unsigned char *ea, unsigned char *str) str2eaddr() argument 103 ea[i] = num; str2eaddr()
|
/linux-4.1.27/arch/powerpc/crypto/ |
H A D | aes-tab-4k.S | 51 .long R(5f, a2, a2, fd), R(45, af, af, ea) 69 .long R(7f, b2, b2, cd), R(ea, 75, 75, 9f) 131 .long R(f3, f4, f4, 07), R(cf, ea, ea, 25) 178 .long R(45, ea, 0e, 98), R(5d, fe, c0, e1) 266 .long R(ea, 9f, 09, d4), R(29, b0, 7c, d6) 277 .long R(9d, 5e, ea, 04), R(01, 8c, 35, 5d) 287 .long R(53, f7, cd, ea), R(5f, fd, aa, 5b)
|
/linux-4.1.27/arch/openrisc/kernel/ |
H A D | traps.c | 173 unsigned long ea, unsigned long vector) nommu_dump_state() 178 printk("\n\r[nommu_dump_state] :: ea %lx, vector %lx\n\r", ea, vector); nommu_dump_state() 276 void unhandled_exception(struct pt_regs *regs, int ea, int vector) unhandled_exception() argument 279 ea, vector); unhandled_exception() 172 nommu_dump_state(struct pt_regs *regs, unsigned long ea, unsigned long vector) nommu_dump_state() argument
|
/linux-4.1.27/drivers/media/tuners/ |
H A D | qt1010_priv.h | 72 2c ea ?
|
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmfmac/ |
H A D | fwil_types.h | 178 * @ea: ether address of peer station. 184 u8 ea[ETH_ALEN]; /* Station address */ member in struct:brcmf_tdls_iovar_le 420 u8 ea[ETH_ALEN]; /* per station */ member in struct:brcmf_wsec_key 442 u8 ea[ETH_ALEN]; /* per station */ member in struct:brcmf_wsec_key_le 448 u8 ea[ETH_ALEN]; member in struct:brcmf_scb_val_le 464 u8 ea[ETH_ALEN]; /* Station address */ member in struct:brcmf_sta_info_le
|
H A D | fwsignal.c | 377 * @ea: ethernet address. 393 u8 ea[ETH_ALEN]; member in struct:brcmf_fws_mac_descriptor 717 "enter: desc %p ea=%pM, ifidx=%u\n", desc, addr, ifidx); brcmf_fws_macdesc_init() 726 memcpy(&desc->ea[0], addr, ETH_ALEN); brcmf_fws_macdesc_init() 733 "enter: ea=%pM, ifidx=%u\n", desc->ea, desc->interface_id); brcmf_fws_macdesc_deinit() 741 brcmf_fws_macdesc_lookup(struct brcmf_fws_info *fws, u8 *ea) brcmf_fws_macdesc_lookup() argument 746 if (ea == NULL) brcmf_fws_macdesc_lookup() 751 if (entry->occupied && !memcmp(entry->ea, ea, ETH_ALEN)) brcmf_fws_macdesc_lookup()
|
H A D | cfg80211.c | 452 memcpy(key_le->ea, key->ea, sizeof(key->ea)); convert_key_from_CPU() 1933 memcpy(&scbval.ea, &profile->bssid, ETH_ALEN); brcmf_cfg80211_disconnect() 2066 /* Instead of bcast for ea address for default wep keys, brcmf_add_keyext() 2069 memcpy((char *)&key.ea, (void *)mac_addr, ETH_ALEN); brcmf_add_keyext() 4272 memcpy(&scbval.ea, params->mac, ETH_ALEN); brcmf_cfg80211_del_station() 4547 memcpy(info.ea, peer, ETH_ALEN); brcmf_cfg80211_tdls_oper()
|
H A D | p2p.c | 2110 struct brcmf_if *ifp, u8 ea[ETH_ALEN], brcmf_p2p_request_p2p_if() 2121 memcpy(if_request.addr, ea, ETH_ALEN); brcmf_p2p_request_p2p_if() 2109 brcmf_p2p_request_p2p_if(struct brcmf_p2p_info *p2p, struct brcmf_if *ifp, u8 ea[ETH_ALEN], enum brcmf_fil_p2p_if_types iftype) brcmf_p2p_request_p2p_if() argument
|
/linux-4.1.27/arch/powerpc/platforms/pseries/ |
H A D | lpar.c | 362 unsigned long ea, pSeries_lpar_hpte_updateboltedpp() 368 vsid = get_kernel_vsid(ea, ssize); pSeries_lpar_hpte_updateboltedpp() 369 vpn = hpt_vpn(ea, vsid, ssize); pSeries_lpar_hpte_updateboltedpp() 498 static void pSeries_lpar_hpte_removebolted(unsigned long ea, pSeries_lpar_hpte_removebolted() argument 504 vsid = get_kernel_vsid(ea, ssize); pSeries_lpar_hpte_removebolted() 505 vpn = hpt_vpn(ea, vsid, ssize); pSeries_lpar_hpte_removebolted() 361 pSeries_lpar_hpte_updateboltedpp(unsigned long newpp, unsigned long ea, int psize, int ssize) pSeries_lpar_hpte_updateboltedpp() argument
|
/linux-4.1.27/arch/m68k/ifpsp060/src/ |
H A D | isp.S | 390 # access these register files for things like data retrieval and <ea> # 433 # muls.l 0100 1100 00 |<ea>| 0*** 1100 0000 0*** # 434 # mulu.l 0100 1100 00 |<ea>| 0*** 0100 0000 0*** # 436 # divs.l 0100 1100 01 |<ea>| 0*** 1100 0000 0*** # 437 # divu.l 0100 1100 01 |<ea>| 0*** 0100 0000 0*** # 444 # cas.w 0000 1100 11 |<ea>| 0000 000* **00 0*** # 445 # cas.l 0000 1110 11 |<ea>| 0000 000* **00 0*** # 452 # chk2.b 0000 0000 11 |<ea>| **** 1000 0000 0000 # 453 # chk2.w 0000 0010 11 |<ea>| **** 1000 0000 0000 # 454 # chk2.l 0000 0100 11 |<ea>| **** 1000 0000 0000 # 456 # cmp2.b 0000 0000 11 |<ea>| **** 0000 0000 0000 # 457 # cmp2.w 0000 0010 11 |<ea>| **** 0000 0000 0000 # 458 # cmp2.l 0000 0100 11 |<ea>| **** 0000 0000 0000 # 925 jmp (tbl_ea_mode.b,%pc,%d0.w*1) # jmp to correct ea mode 1372 # word, long: <ea> of the data is the current extension word # 1393 mov.w %d0,%a0 # return <ea> in a0 1407 mov.l %d0,%a0 # return <ea> in a0 1426 subq.l &0x2,%a0 # adjust <ea> 1579 add.l %d2,%d0 # <ea> += index 1580 add.l %d4,%d0 # <ea> += od 1591 add.l %d4,%d0 # ea += od 1595 add.l %d2,%d3 # ea = (base + bd) + index 1895 bsr.l _calc_ea # calculate <ea> 1910 mov.l %a0,%a2 # save copy of <ea> 1918 mov.l %a2,%a0 # pass <ea> of long upper bound 2039 # _div64(): routine to emulate div{u,s}.l <ea>,Dr:Dq # 2431 bsr.l _calc_ea # calc <ea> 2437 bsr.l _dmem_read_long # fetch divisor from <ea> 2478 # _mul64(): routine to emulate mul{u,s}.l <ea>,Dh:Dl 32x32->64 # 2662 # must calculate the <ea> and go fetch the 32-bit operand. 2665 bsr.l _calc_ea # calculate <ea> 2918 # _compandset(): routine to emulate cas w/ misaligned <ea> # 3010 bsr.l _calc_ea # a0 = calculated <ea> 3011 mov.l %a0,ADDR(%a6) # save <ea> for possible restart 3017 bsr.l _calc_ea # a0 = calculated <ea> 3018 mov.l %a0,ADDR(%a6) # save <ea> for possible restart 3104 mov.l ADDR(%a6),%a0 # load <ea>
|
H A D | pfpsp.S | 300 set EXC_EA, 0xc # stacked <ea> 546 set mia7_flg, 0x04 # flag bit: (a7)+ <ea> 547 set mda7_flg, 0x08 # flag bit: -(a7) <ea> 549 set immed_flg, 0x80 # flag bit: &<data> <ea> 1221 # stack an ea. however, we don't need an a7' for this case anyways. 1580 # is the ea mode pre-decrement of the stack pointer from supervisor mode? 1751 cmpi.b SPCOND_FLG(%a6),&mda7_flg # was the <ea> mode -(sp)? 1904 btst &mia7_bit,SPCOND_FLG(%a6) # was ea mode (a7)+ 1913 # shift the stack frame "up". we don't really care about the <ea> field. 2027 # shift stack frame "up". who cares about <ea> field. 2135 btst &mda7_bit,SPCOND_FLG(%a6) # was ea mode -(a7) 2388 # - The system stack contains the "Unimplemented <ea>" stk frame # 2435 # see if the FPU is disabled. Since the "Unimp <ea>" exception is taken # 2982 # call fmovm_calc_ea which will go through the ea calc process and, 2998 # the <ea> field is let as undefined. 3029 mov.l 0x2(%sp),0x8(%sp) # store ea 3190 mov.b 1+EXC_OPWORD(%a6),%d1 # extract <ea> mode,reg 3206 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3208 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3222 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3224 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3238 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3240 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3266 # _calc_ea_fout() - fix An if <ea> is -() or ()+; also get <ea> # 3294 # register must get updated by calling _calc_ea_fout(). If the <ea> # 3360 mov.b 1+EXC_OPWORD(%a6),%d1 # extract <ea> mode,reg 3377 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3379 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3394 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3396 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3411 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3413 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3426 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3434 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3482 # in addition, for pre-decrement, the stacked <ea> is incorrect. 3500 bsr.l _calc_ea_fout # find the correct ea,update An 3502 mov.l %a0,EXC_EA(%a6) # stack correct <ea> 3521 bsr.l _calc_ea_fout # find the correct ea,update An 3523 mov.l %a0,EXC_EA(%a6) # stack correct <ea> 3527 cmpi.b SPCOND_FLG(%a6),&mda7_flg # is <ea> mode -(a7)? 3848 # _dcalc_ea(): calc correct <ea> from <ea> stacked on exception # 3855 # d0 = number of bytes to adjust <ea> by # 3862 # The stacked <ea> for FP unimplemented instructions and opclass # 3867 # stacked <ea> value is 8 bytes too big # 3896 mov.l EXC_EA(%a6),%a0 # return <ea> 3903 lea ([USER_FPIAR,%a6],0x4),%a0 # no; return <ea> 3906 # here, the <ea> is stacked correctly. however, we must update the 3912 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 3915 # the <ea> is stacked correctly for all but extended and packed which 3916 # the <ea>s are 8 bytes too large. 3923 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 3929 sub.l &0x8,%a0 # correct <ea> 3930 mov.l %a0,EXC_EA(%a6) # put correct <ea> on stack 3935 # _calc_ea_fout(): calculate correct stacked <ea> for extended # 3948 # For opclass 3 extended and packed data operations, the <ea> # 3952 # So, for -(an), we must subtract 8 off of the stacked <ea> value # 3953 # and return that value as the correct <ea> and store that value in An. # 3954 # For (an)+, the stacked <ea> is correct but we must adjust An by +12. # 3958 # This calc_ea is currently used to retrieve the correct <ea> 3974 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 3978 # : stacked <ea> is correct 4023 # : stacked <ea> = actual <ea> + 8 4234 # If the instruction is "fmovm.x <ea>,Dn", a _mem_read() is used # 4237 # If the instruction is "fmovm.x DN,<ea>", a _mem_write() is used # 4246 # fmovm.x dn, <ea> # 4247 # fmovm.x <ea>, dn # 4250 # 1111 0010 00 |<ea>| 11@& 1000 0$$$ 0000 # 4280 bsr.l fmovm_calc_ea # calculate <ea> 4285 # but, make sure that we've calculated ea and advanced the opword pointer 4310 cmpi.b SPCOND_FLG(%a6),&mda7_flg # is <ea> mode -(a7)? 4320 mov.l %a0,%a1 # move <ea> to a1 4585 jmp (tbl_fea_mode.b,%pc,%d0.w*1) # jmp to correct ea mode 4996 mov.w %d0,%a0 # return <ea> in a0 5010 mov.l %d0,%a0 # return <ea> in a0 5029 subq.l &0x2,%a0 # adjust <ea> 5198 add.l %d2,%d0 # <ea> += index 5199 add.l %d4,%d0 # <ea> += od 5210 add.l %d4,%d0 # ea += od 5214 add.l %d2,%d3 # ea = (base + bd) + index 7086 # For extended precision, the stacked <ea> must be fixed along # 7091 # word or a data register. The <ea> must be fixed as w/ extended # 7142 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 7188 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 7235 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 7275 bsr.l _calc_ea_fout # fix stacked <ea> 7376 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 7423 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 7473 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 7856 bsr.l _calc_ea_fout # fetch the <ea> 12559 # Typically used for an instruction w/ a post-increment <ea>, # 12623 # Typically used for an instruction w/ a pre-decrement <ea>, # 12893 # _dcalc_ea() - calculate the correct <ea> # 12906 # Get the correct <ea> which is the value on the exception stack # 12907 # frame w/ maybe a correction factor if the <ea> is -(an) or (an)+. # 12916 # the stacked <ea> for packed is correct except for -(An). 12921 bsr.l _dcalc_ea # fetch <ea>; correct An 14742 # need to invert adjustment value if the <ea> was predec
|
H A D | ilsp.S | 755 # _060LSP__cmp2_Ab_(): Emulate "cmp2.b An,<ea>". # 756 # _060LSP__cmp2_Aw_(): Emulate "cmp2.w An,<ea>". # 757 # _060LSP__cmp2_Al_(): Emulate "cmp2.l An,<ea>". # 758 # _060LSP__cmp2_Db_(): Emulate "cmp2.b Dn,<ea>". # 759 # _060LSP__cmp2_Dw_(): Emulate "cmp2.w Dn,<ea>". # 760 # _060LSP__cmp2_Dl_(): Emulate "cmp2.l Dn,<ea>". #
|
H A D | fpsp.S | 301 set EXC_EA, 0xc # stacked <ea> 547 set mia7_flg, 0x04 # flag bit: (a7)+ <ea> 548 set mda7_flg, 0x08 # flag bit: -(a7) <ea> 550 set immed_flg, 0x80 # flag bit: &<data> <ea> 1222 # stack an ea. however, we don't need an a7' for this case anyways. 1581 # is the ea mode pre-decrement of the stack pointer from supervisor mode? 1752 cmpi.b SPCOND_FLG(%a6),&mda7_flg # was the <ea> mode -(sp)? 1905 btst &mia7_bit,SPCOND_FLG(%a6) # was ea mode (a7)+ 1914 # shift the stack frame "up". we don't really care about the <ea> field. 2028 # shift stack frame "up". who cares about <ea> field. 2136 btst &mda7_bit,SPCOND_FLG(%a6) # was ea mode -(a7) 2389 # - The system stack contains the "Unimplemented <ea>" stk frame # 2436 # see if the FPU is disabled. Since the "Unimp <ea>" exception is taken # 2983 # call fmovm_calc_ea which will go through the ea calc process and, 2999 # the <ea> field is let as undefined. 3030 mov.l 0x2(%sp),0x8(%sp) # store ea 3191 mov.b 1+EXC_OPWORD(%a6),%d1 # extract <ea> mode,reg 3207 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3209 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3223 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3225 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3239 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3241 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3267 # _calc_ea_fout() - fix An if <ea> is -() or ()+; also get <ea> # 3295 # register must get updated by calling _calc_ea_fout(). If the <ea> # 3361 mov.b 1+EXC_OPWORD(%a6),%d1 # extract <ea> mode,reg 3378 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3380 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3395 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3397 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3412 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3414 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3427 cmpi.b %d1,&0x7 # is <ea> mode a data reg? 3435 mov.l EXC_EA(%a6),%a0 # pass: <ea> of default result 3483 # in addition, for pre-decrement, the stacked <ea> is incorrect. 3501 bsr.l _calc_ea_fout # find the correct ea,update An 3503 mov.l %a0,EXC_EA(%a6) # stack correct <ea> 3522 bsr.l _calc_ea_fout # find the correct ea,update An 3524 mov.l %a0,EXC_EA(%a6) # stack correct <ea> 3528 cmpi.b SPCOND_FLG(%a6),&mda7_flg # is <ea> mode -(a7)? 3837 # non-zero <ea> field. These may get flagged as "Line F" but should # 3857 # if the F-Line instruction is an "fmovecr" w/ a non-zero <ea>. if 3878 # it's an fmovecr w/ a non-zero <ea> that has entered through 3894 sub.l &0x8,%sp # make room for "Next PC", <ea> 3999 # * * => <ea> of fp unimp instr. 4110 cmpi.b SPCOND_FLG(%a6),&mia7_flg # was the ea mode (sp)+ ? 4113 cmpi.b SPCOND_FLG(%a6),&mda7_flg # was the ea mode -(sp) ? 4332 # fs<cc>.b <ea> # 17431 # only difference is that the <ea> field should hold the PC 18077 # the stacked <ea> is correct with the exception of: 18078 # -> Dn : <ea> is garbage 18084 cmpi.b %d1,&0x18 # is <ea> (An)+ ? 18086 cmpi.b %d1,&0x20 # is <ea> -(An) ? 18090 mov.l EXC_EA(%a6),%a0 # fetch <ea> 18103 mov.l EXC_EA(%a6),%a0 # fetch <ea> 18121 mov.l EXC_EA(%a6),%a0 # fetch <ea> 18185 # If the instruction is "fmovm.x <ea>,Dn", a _mem_read() is used # 18188 # If the instruction is "fmovm.x DN,<ea>", a _mem_write() is used # 18197 # fmovm.x dn, <ea> # 18198 # fmovm.x <ea>, dn # 18201 # 1111 0010 00 |<ea>| 11@& 1000 0$$$ 0000 # 18231 bsr.l fmovm_calc_ea # calculate <ea> 18236 # but, make sure that we've calculated ea and advanced the opword pointer 18261 cmpi.b SPCOND_FLG(%a6),&mda7_flg # is <ea> mode -(a7)? 18271 mov.l %a0,%a1 # move <ea> to a1 18536 jmp (tbl_fea_mode.b,%pc,%d0.w*1) # jmp to correct ea mode 18947 mov.w %d0,%a0 # return <ea> in a0 18961 mov.l %d0,%a0 # return <ea> in a0 18980 subq.l &0x2,%a0 # adjust <ea> 19149 add.l %d2,%d0 # <ea> += index 19150 add.l %d4,%d0 # <ea> += od 19161 add.l %d4,%d0 # ea += od 19165 add.l %d2,%d3 # ea = (base + bd) + index 19335 # _dcalc_ea(): calc correct <ea> from <ea> stacked on exception # 19342 # d0 = number of bytes to adjust <ea> by # 19349 # The stacked <ea> for FP unimplemented instructions and opclass # 19354 # stacked <ea> value is 8 bytes too big # 19383 mov.l EXC_EA(%a6),%a0 # return <ea> 19390 lea ([USER_FPIAR,%a6],0x4),%a0 # no; return <ea> 19393 # here, the <ea> is stacked correctly. however, we must update the 19399 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 19402 # the <ea> is stacked correctly for all but extended and packed which 19403 # the <ea>s are 8 bytes too large. 19410 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 19416 sub.l &0x8,%a0 # correct <ea> 19417 mov.l %a0,EXC_EA(%a6) # put correct <ea> on stack 19422 # _calc_ea_fout(): calculate correct stacked <ea> for extended # 19435 # For opclass 3 extended and packed data operations, the <ea> # 19439 # So, for -(an), we must subtract 8 off of the stacked <ea> value # 19440 # and return that value as the correct <ea> and store that value in An. # 19441 # For (an)+, the stacked <ea> is correct but we must adjust An by +12. # 19445 # This calc_ea is currently used to retrieve the correct <ea> 19461 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 19465 # : stacked <ea> is correct 19510 # : stacked <ea> = actual <ea> + 8 19567 # _dcalc_ea() - calculate <ea>, fixing An in process # 19618 # beq.w op010 # handle <ea> -> fpn 19619 # bgt.w op011 # handle fpn -> <ea> 19669 # OPCLASS '010: <ea> -> reg # 19695 bfextu EXC_OPWORD(%a6){&10:&3}, %d1 # extract <ea> mode field 19788 # (1) calc ea - must read AFTER you know the src type since # 19789 # if the ea is -() or ()+, need to know # of bytes. # 19817 # (1) calc ea # 19823 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0 19853 # (1) calc ea # 19859 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0 19889 # (1) calc ea # 19895 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0 19925 # (1) calc ea # 19931 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0 20007 # (1) calc ea # 20013 bsr.l _dcalc_ea # calc <ea>; <ea> in %a0 20096 # (1) calc ea # 20102 bsr.l _dcalc_ea # calc <ea> 20128 # (1) calc ea # 20185 # For extended precision, the stacked <ea> must be fixed along # 20190 # word or a data register. The <ea> must be fixed as w/ extended # 20241 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 20287 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 20334 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 20374 bsr.l _calc_ea_fout # fix stacked <ea> 20475 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 20522 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 20572 mov.l EXC_EA(%a6),%a0 # stacked <ea> is correct 20955 bsr.l _calc_ea_fout # fetch the <ea> 21343 # Typically used for an instruction w/ a post-increment <ea>, # 21407 # Typically used for an instruction w/ a pre-decrement <ea>, # 22933 # _dcalc_ea() - calculate the correct <ea> # 22946 # Get the correct <ea> which is the value on the exception stack # 22947 # frame w/ maybe a correction factor if the <ea> is -(an) or (an)+. # 22956 # the stacked <ea> for packed is correct except for -(An). 22961 bsr.l _dcalc_ea # fetch <ea>; correct An 24782 # need to invert adjustment value if the <ea> was predec
|
H A D | itest.S | 120 ### ea 1285 cas.w %d1,%d2,(%a0) # Dc,Du,<ea> 1312 cas.w %d1,%d2,(%a0) # Dc,Du,<ea> 1340 cas.l %d1,%d2,(%a0) # Dc,Du,<ea> 1367 cas.l %d1,%d2,(%a0) # Dc,Du,<ea> 1395 cas.l %d1,%d2,(%a0) # Dc,Du,<ea> 1422 cas.l %d1,%d2,(%a0) # Dc,Du,<ea>
|
H A D | ftest.S | 467 string "\tUnimplemented <ea>..."
|
H A D | fplsp.S | 281 set EXC_EA, 0xc # stacked <ea> 527 set mia7_flg, 0x04 # flag bit: (a7)+ <ea> 528 set mda7_flg, 0x08 # flag bit: -(a7) <ea> 530 set immed_flg, 0x80 # flag bit: &<data> <ea>
|
/linux-4.1.27/arch/sh/mm/ |
H A D | cache-sh4.c | 323 unsigned long a, ea, p; __flush_cache_one() local 353 ea = base_addr + PAGE_SIZE; __flush_cache_one() 367 } while (a < ea); __flush_cache_one()
|
/linux-4.1.27/arch/mips/pmcs-msp71xx/ |
H A D | msp_prom.c | 86 int str2eaddr(unsigned char *ea, unsigned char *str) str2eaddr() argument 93 ea[index++] = num; str2eaddr() 103 ea[index++] = num; str2eaddr()
|
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/ |
H A D | scb.h | 69 u8 ea[ETH_ALEN]; /* station address */ member in struct:scb
|
/linux-4.1.27/arch/unicore32/include/asm/ |
H A D | assembler.h | 69 .else; .ifc \cond, ea
|
/linux-4.1.27/drivers/staging/lustre/lustre/mdc/ |
H A D | mdc_internal.h | 67 void *ea, int ealen, void *ea2, int ea2len); 142 void *ea, int ealen, void *ea2, int ea2len,
|
H A D | mdc_lib.c | 339 void *ea, int ealen, void *ea2, int ea2len) mdc_setattr_pack() 361 if (ea == NULL) { /* Remove LOV EA */ mdc_setattr_pack() 367 memcpy(lum, ea, ealen); mdc_setattr_pack() 338 mdc_setattr_pack(struct ptlrpc_request *req, struct md_op_data *op_data, void *ea, int ealen, void *ea2, int ea2len) mdc_setattr_pack() argument
|
H A D | mdc_reint.c | 104 void *ea, int ealen, void *ea2, int ea2len, mdc_setattr() 151 mdc_setattr_pack(req, op_data, ea, ealen, ea2, ea2len); mdc_setattr() 103 mdc_setattr(struct obd_export *exp, struct md_op_data *op_data, void *ea, int ealen, void *ea2, int ea2len, struct ptlrpc_request **request, struct md_open_data **mod) mdc_setattr() argument
|
/linux-4.1.27/drivers/staging/lustre/lustre/osc/ |
H A D | osc_internal.h | 89 struct obdo *oa, struct lov_stripe_md **ea, 92 struct lov_stripe_md **ea, struct obd_trans_info *oti);
|
H A D | osc_request.c | 459 struct lov_stripe_md **ea, struct obd_trans_info *oti) osc_real_create() 467 LASSERT(ea); osc_real_create() 469 lsm = *ea; osc_real_create() 524 *ea = lsm; osc_real_create() 541 if (rc && !*ea) osc_real_create() 721 struct obdo *oa, struct lov_stripe_md **ea, osc_create() 727 LASSERT(ea); osc_create() 732 return osc_real_create(exp, oa, ea, oti); osc_create() 736 return osc_real_create(exp, oa, ea, oti); osc_create() 755 struct obdo *oa, struct lov_stripe_md *ea, osc_destroy() 458 osc_real_create(struct obd_export *exp, struct obdo *oa, struct lov_stripe_md **ea, struct obd_trans_info *oti) osc_real_create() argument 720 osc_create(const struct lu_env *env, struct obd_export *exp, struct obdo *oa, struct lov_stripe_md **ea, struct obd_trans_info *oti) osc_create() argument 754 osc_destroy(const struct lu_env *env, struct obd_export *exp, struct obdo *oa, struct lov_stripe_md *ea, struct obd_trans_info *oti, struct obd_export *md_export, void *capa) osc_destroy() argument
|
/linux-4.1.27/arch/x86/kvm/ |
H A D | emulate.c | 655 la = seg_base(ctxt, addr.seg) + addr.ea; __linearize() 681 if (addr.ea <= lim) __linearize() 685 if (addr.ea > lim) __linearize() 690 *max_size = (u64)lim + 1 - addr.ea; __linearize() 725 .ea = dst }; assign_eip() 728 addr.ea = dst & ((1UL << (ctxt->op_bytes << 3)) - 1); assign_eip() 731 ctxt->_eip = addr.ea; assign_eip() 796 .ea = ctxt->eip + cur_size }; __do_insn_fetch_bytes() 897 addr.ea += 2; read_descriptor() 1289 op->addr.mem.ea = modrm_ea; decode_modrm() 1291 ctxt->memop.addr.mem.ea = (u32)ctxt->memop.addr.mem.ea; decode_modrm() 1305 op->addr.mem.ea = insn_fetch(u16, ctxt); decode_abs() 1308 op->addr.mem.ea = insn_fetch(u32, ctxt); decode_abs() 1311 op->addr.mem.ea = insn_fetch(u64, ctxt); decode_abs() 1332 ctxt->dst.addr.mem.ea = address_mask(ctxt, fetch_bit_operand() 1333 ctxt->dst.addr.mem.ea + (sv >> 3)); fetch_bit_operand() 1758 addr.ea = reg_read(ctxt, VCPU_REGS_RSP) & stack_mask(ctxt); push() 1777 addr.ea = reg_read(ctxt, VCPU_REGS_RSP) & stack_mask(ctxt); emulate_pop() 2967 op->addr.mem.ea = register_address(ctxt, reg); string_addr_inc() 4275 op->addr.mem.ea = ctxt->_eip; decode_imm() 4360 op->addr.mem.ea = decode_operand() 4414 op->addr.mem.ea = decode_operand() 4423 op->addr.mem.ea = decode_operand() 4432 op->addr.mem.ea = ctxt->_eip; decode_operand() 4736 ctxt->memopp->addr.mem.ea = address_mask(ctxt, x86_decode_insn() 4737 ctxt->memopp->addr.mem.ea + ctxt->_eip); x86_decode_insn() 4993 ctxt->dst.val = ctxt->src.addr.mem.ea; x86_emulate_insn()
|
/linux-4.1.27/drivers/base/ |
H A D | core.c | 152 struct dev_ext_attribute *ea = to_ext_attr(attr); device_store_ulong() local 157 *(unsigned long *)(ea->var) = new; device_store_ulong() 167 struct dev_ext_attribute *ea = to_ext_attr(attr); device_show_ulong() local 168 return snprintf(buf, PAGE_SIZE, "%lx\n", *(unsigned long *)(ea->var)); device_show_ulong() 176 struct dev_ext_attribute *ea = to_ext_attr(attr); device_store_int() local 181 *(int *)(ea->var) = new; device_store_int() 191 struct dev_ext_attribute *ea = to_ext_attr(attr); device_show_int() local 193 return snprintf(buf, PAGE_SIZE, "%d\n", *(int *)(ea->var)); device_show_int() 200 struct dev_ext_attribute *ea = to_ext_attr(attr); device_store_bool() local 202 if (strtobool(buf, ea->var) < 0) device_store_bool() 212 struct dev_ext_attribute *ea = to_ext_attr(attr); device_show_bool() local 214 return snprintf(buf, PAGE_SIZE, "%d\n", *(bool *)(ea->var)); device_show_bool()
|
/linux-4.1.27/arch/powerpc/platforms/ps3/ |
H A D | htab.c | 159 static void ps3_hpte_updateboltedpp(unsigned long newpp, unsigned long ea, ps3_hpte_updateboltedpp() argument
|
/linux-4.1.27/net/ieee802154/ |
H A D | trace.h | 59 TP_printk(WPAN_PHY_PR_FMT ", virtual intf name: %s, type: %d, ea %llx",
|
/linux-4.1.27/arch/powerpc/platforms/powernv/ |
H A D | opal.c | 487 uint64_t ea = get_mce_fault_addr(evt); opal_recover_mce() local 495 } else if (ea && !is_kernel_addr(ea)) { opal_recover_mce()
|
/linux-4.1.27/drivers/staging/lustre/lustre/lclient/ |
H A D | lcommon_misc.c | 110 /* for each osc event update ea size */ cl_ocd_update()
|
/linux-4.1.27/arch/s390/include/asm/ |
H A D | etr.h | 17 unsigned int ea : 1; /* ETR alert control */ member in struct:etr_eacr
|
/linux-4.1.27/arch/m68k/mm/ |
H A D | memory.c | 178 * (Low order bits of the ea of a CINVP/CPUSHP are "don't care"s).
|
/linux-4.1.27/drivers/net/ethernet/tile/ |
H A D | tilepro.c | 938 netio_ipp_address_t ea = { tile_net_open_aux() local 944 ea.pte = hv_pte_set_lotar(ea.pte, epp_lotar); tile_net_open_aux() 945 ea.pte = hv_pte_set_mode(ea.pte, HV_PTE_MODE_CACHE_TILE_L3); tile_net_open_aux() 947 (HV_VirtAddr)&ea, tile_net_open_aux() 948 sizeof(ea), tile_net_open_aux()
|
/linux-4.1.27/drivers/mmc/card/ |
H A D | mmc_test.c | 1927 unsigned int dev_addr, cnt, rnd_addr, range1, range2, last_ea = 0, ea; mmc_test_rnd_perf() local 1944 ea = mmc_test_rnd_num(range1); mmc_test_rnd_perf() 1945 if (ea == last_ea) mmc_test_rnd_perf() 1946 ea -= 1; mmc_test_rnd_perf() 1947 last_ea = ea; mmc_test_rnd_perf() 1948 dev_addr = rnd_addr + test->card->pref_erase * ea + mmc_test_rnd_perf()
|
/linux-4.1.27/drivers/staging/lustre/lustre/include/ |
H A D | obd_class.h | 791 struct obdo *obdo, struct lov_stripe_md **ea, obd_create() 799 rc = OBP(exp->exp_obd, create)(env, exp, obdo, ea, oti); obd_create() 804 struct obdo *obdo, struct lov_stripe_md *ea, obd_destroy() 813 rc = OBP(exp->exp_obd, destroy)(env, exp, obdo, ea, oti, md_exp, capa); obd_destroy() 1639 void *ea, int ealen, void *ea2, int ea2len, md_setattr() 1647 rc = MDP(exp->exp_obd, setattr)(exp, op_data, ea, ealen, md_setattr() 790 obd_create(const struct lu_env *env, struct obd_export *exp, struct obdo *obdo, struct lov_stripe_md **ea, struct obd_trans_info *oti) obd_create() argument 803 obd_destroy(const struct lu_env *env, struct obd_export *exp, struct obdo *obdo, struct lov_stripe_md *ea, struct obd_trans_info *oti, struct obd_export *md_exp, void *capa) obd_destroy() argument 1638 md_setattr(struct obd_export *exp, struct md_op_data *op_data, void *ea, int ealen, void *ea2, int ea2len, struct ptlrpc_request **request, struct md_open_data **mod) md_setattr() argument
|
H A D | obd.h | 1164 struct obdo *oa, struct lov_stripe_md **ea, 1167 struct obdo *oa, struct lov_stripe_md *ea,
|
/linux-4.1.27/drivers/input/serio/ |
H A D | hyperv-keyboard.c | 417 * {f912ad6d-2b17-48ea-bd65-f927a61c7684}
|
/linux-4.1.27/include/uapi/linux/ |
H A D | gfs2_ondisk.h | 377 #define GFS2_EAFLAG_LAST 0x01 /* last ea in block */
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | time.c | 454 * e0 e1 dp p0 p1 ea es sl 497 .p0 = 0, .p1 = 0, ._pad1 = 0, .ea = 0, etr_reset() 1013 eacr.ea = eacr.p0 || eacr.p1; etr_work_fn() 1014 if (!eacr.ea) { etr_work_fn()
|
/linux-4.1.27/drivers/staging/lustre/lustre/lov/ |
H A D | lov_obd.c | 987 struct lov_stripe_md **ea, struct obd_trans_info *oti) lov_recreate() 1002 lsm = *ea; lov_recreate() 1041 struct obdo *src_oa, struct lov_stripe_md **ea, lov_create() 1047 LASSERT(ea != NULL); lov_create() 1065 rc = lov_recreate(exp, src_oa, ea, oti); lov_create() 986 lov_recreate(struct obd_export *exp, struct obdo *src_oa, struct lov_stripe_md **ea, struct obd_trans_info *oti) lov_recreate() argument 1040 lov_create(const struct lu_env *env, struct obd_export *exp, struct obdo *src_oa, struct lov_stripe_md **ea, struct obd_trans_info *oti) lov_create() argument
|
/linux-4.1.27/drivers/net/ethernet/apple/ |
H A D | bmac.c | 1185 bmac_get_station_address(struct net_device *dev, unsigned char *ea) bmac_get_station_address() argument 1194 ea[2*i] = bitrev8(data & 0x0ff); bmac_get_station_address() 1195 ea[2*i+1] = bitrev8((data >> 8) & 0x0ff); bmac_get_station_address()
|
/linux-4.1.27/arch/x86/include/asm/ |
H A D | kvm_emulate.h | 228 ulong ea; member in struct:operand::__anon3047::segmented_address
|
/linux-4.1.27/drivers/atm/ |
H A D | firestream.h | 457 struct FS_QENTRY *sa, *ea; member in struct:queue
|
H A D | firestream.c | 1429 txq->ea = p; init_q()
|
/linux-4.1.27/tools/lib/traceevent/ |
H A D | event-parse.c | 5275 struct event_format * const * ea = a; events_id_cmp() local 5278 if ((*ea)->id < (*eb)->id) events_id_cmp() 5281 if ((*ea)->id > (*eb)->id) events_id_cmp() 5289 struct event_format * const * ea = a; events_name_cmp() local 5293 res = strcmp((*ea)->name, (*eb)->name); events_name_cmp() 5297 res = strcmp((*ea)->system, (*eb)->system); events_name_cmp() 5306 struct event_format * const * ea = a; events_system_cmp() local 5310 res = strcmp((*ea)->system, (*eb)->system); events_system_cmp() 5314 res = strcmp((*ea)->name, (*eb)->name); events_system_cmp()
|
H A D | parse-filter.c | 106 const struct filter_type *ea = a; filter_cmp() local 109 if (ea->event_id < eb->event_id) filter_cmp() 112 if (ea->event_id > eb->event_id) filter_cmp()
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
H A D | perf_event_intel_pt.c | 86 struct dev_ext_attribute *ea = pt_cap_show() local 88 enum pt_capabilities cap = (long)ea->var; pt_cap_show()
|
/linux-4.1.27/drivers/bus/ |
H A D | arm-ccn.c | 222 struct dev_ext_attribute *ea = container_of(attr, arm_ccn_pmu_format_show() local 225 return snprintf(buf, PAGE_SIZE, "%s\n", (char *)ea->var); arm_ccn_pmu_format_show()
|
/linux-4.1.27/drivers/media/dvb-frontends/ |
H A D | m88ds3103.c | 486 /* u8tmp1[5:2] => fe[3:0], u8tmp1[1:0] => ea[7:6] */ m88ds3103_set_frontend() 488 /* u8tmp2[5:0] => ea[5:0] */ m88ds3103_set_frontend()
|
/linux-4.1.27/drivers/video/fbdev/sis/ |
H A D | init.h | 630 0xe9,0x8b,0xdf,0x28,0x00,0xe7,0x04,0xc3, /* e9,8b is ea,8c on 300 */ 645 0xe9,0x8b,0xdf,0x28,0x00,0xe7,0x04,0xe3, /* e9,8b is ea,8c on 300 */
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | uprobes.c | 139 * ea - formerly jmp seg:ofs
|
/linux-4.1.27/fs/ext3/ |
H A D | xattr.c | 12 * ea-in-inode support by Alex Tomas <alex@clusterfs.com> aka bzzz
|
/linux-4.1.27/drivers/staging/lustre/lustre/lmv/ |
H A D | lmv_obd.c | 2029 void *ea, int ealen, void *ea2, int ea2len, lmv_setattr() 2050 rc = md_setattr(tgt->ltd_exp, op_data, ea, ealen, ea2, lmv_setattr() 2028 lmv_setattr(struct obd_export *exp, struct md_op_data *op_data, void *ea, int ealen, void *ea2, int ea2len, struct ptlrpc_request **request, struct md_open_data **mod) lmv_setattr() argument
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | head_fsl_booke.S | 795 andc r12, r12, r13 /* mask off ea bits within the page */
|
/linux-4.1.27/fs/ext4/ |
H A D | xattr.c | 12 * ea-in-inode support by Alex Tomas <alex@clusterfs.com> aka bzzz
|
/linux-4.1.27/fs/cifs/ |
H A D | cifssmb.c | 6254 cifs_dbg(FYI, "ea length %d\n", list_len); CIFSSMBQAllEAs() 6271 /* account for ea list len */ CIFSSMBQAllEAs() 6382 now use name_len to calculate length of ea name CIFSSMBSetEA()
|
/linux-4.1.27/drivers/media/usb/gspca/ |
H A D | zc3xx.c | 1856 {0xa0, 0xea, ZC3XX_R196_ANTIFLICKERMID}, /* 01,96,ea,cc */ 2579 {0xa0, 0xea, ZC3XX_R01F_HSYNC_2}, /* 00,1f,ea,cc */ 2674 {0xa0, 0xea, ZC3XX_R01F_HSYNC_2}, /* 00,1f,ea,cc */ 5155 {0xa0, 0xea, ZC3XX_R01F_HSYNC_2}, /* 00,1f,ea,cc */
|
/linux-4.1.27/drivers/staging/lustre/lustre/obdclass/ |
H A D | lu_object.c | 1963 * inode, one for ea. Unfortunately setting this high value results in lu_global_init()
|
/linux-4.1.27/drivers/tty/ |
H A D | n_gsm.c | 1163 /* Must be at least one byte following ea */ gsm_control_rls()
|
/linux-4.1.27/drivers/staging/lustre/lustre/include/lustre/ |
H A D | lustre_idl.h | 3552 /** The link ea holds 1 \a link_ea_entry for each hardlink */
|
/linux-4.1.27/fs/ntfs/ |
H A D | layout.h | 1089 } __attribute__ ((__packed__)) ea; member in union:__anon11167::__anon11168
|