ia                433 arch/arm/mach-omap2/clockdomain.c int clkdm_register_autodeps(struct clkdm_autodep *ia)
ia                440 arch/arm/mach-omap2/clockdomain.c 	if (!ia)
ia                446 arch/arm/mach-omap2/clockdomain.c 	autodeps = ia;
ia                183 arch/arm/mach-omap2/clockdomain.h int clkdm_register_autodeps(struct clkdm_autodep *ia);
ia               3516 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t ia:5;
ia               3522 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t ia:5;
ia               3541 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t ia:8;
ia               3545 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t ia:8;
ia                248 arch/powerpc/kvm/e500_mmu.c 	unsigned int ia;
ia                251 arch/powerpc/kvm/e500_mmu.c 	ia = (ea >> 2) & 0x1;
ia                256 arch/powerpc/kvm/e500_mmu.c 	if (ia) {
ia                114 arch/s390/include/asm/cpu_mf.h 	unsigned long long ia;	    /* Instruction Address		 */
ia                 60 arch/s390/include/asm/nmi.h 		u64 ia :  1; /* 23 psw instruction address validity */
ia                 50 arch/s390/include/asm/ptrace.h 	unsigned long ia     : 64; /* Instruction Address */
ia                284 arch/s390/kernel/nmi.c 	if (!mci.ia && !umode) {
ia                292 arch/s390/kernel/nmi.c 	if (!mci.ms || !mci.pm || !mci.ia)
ia               1093 arch/s390/kernel/perf_cpum_sf.c 	psw_bits(regs.psw).ia	= basic->ia;
ia               2101 arch/sparc/mm/init_64.c 		unsigned int *ia = (unsigned int *)(unsigned long)*start;
ia               2103 arch/sparc/mm/init_64.c 		ia[0] = (ia[0] & ~0x3fffff) | (high_bits >> 10);
ia               2104 arch/sparc/mm/init_64.c 		__asm__ __volatile__("flush	%0" : : "r" (ia));
ia               2106 arch/sparc/mm/init_64.c 		ia[1] = (ia[1] & ~0x3fffff) | (low_bits >> 10);
ia               2107 arch/sparc/mm/init_64.c 		__asm__ __volatile__("flush	%0" : : "r" (ia + 1));
ia               2109 arch/sparc/mm/init_64.c 		ia[2] = (ia[2] & ~0x1fff) | (high_bits & 0x3ff);
ia               2110 arch/sparc/mm/init_64.c 		__asm__ __volatile__("flush	%0" : : "r" (ia + 2));
ia               2112 arch/sparc/mm/init_64.c 		ia[3] = (ia[3] & ~0x1fff) | (low_bits & 0x3ff);
ia               2113 arch/sparc/mm/init_64.c 		__asm__ __volatile__("flush	%0" : : "r" (ia + 3));
ia                528 drivers/acpi/hmat/hmat.c 	struct memory_initiator *ia;
ia                532 drivers/acpi/hmat/hmat.c 	ia = list_entry(a, struct memory_initiator, node);
ia                535 drivers/acpi/hmat/hmat.c 	set_bit(ia->processor_pxm, p_nodes);
ia                538 drivers/acpi/hmat/hmat.c 	return ia->processor_pxm - ib->processor_pxm;
ia                822 drivers/atm/iphase.c static u32 ia_phy_read32(struct iadev_priv *ia, unsigned int reg)
ia                824 drivers/atm/iphase.c 	return readl(ia->phy + (reg >> 2));
ia                827 drivers/atm/iphase.c static void ia_phy_write32(struct iadev_priv *ia, unsigned int reg, u32 val)
ia                829 drivers/atm/iphase.c 	writel(val, ia->phy + (reg >> 2));
ia                 31 drivers/regulator/qcom_rpm-regulator.c 	struct request_member ia;		/* average current in mA */
ia                 73 drivers/regulator/qcom_rpm-regulator.c 	.ia             = { 1, 0x00001FFE,  1 },
ia                 84 drivers/regulator/qcom_rpm-regulator.c 	.ia             = { 1, 0x00001FFE,  1 },
ia                113 drivers/regulator/qcom_rpm-regulator.c 	.ia             = { 1, 0x000FFC00, 10 },
ia                124 drivers/regulator/qcom_rpm-regulator.c 	.ia             = { 1, 0x000FFC00, 10 },
ia                392 drivers/regulator/qcom_rpm-regulator.c 	const struct request_member *req = &parts->ia;
ia               1634 drivers/scsi/sym53c8xx_2/sym_fw2.h 	SCR_LOAD_REL (ia, 4),
ia               1638 drivers/scsi/sym53c8xx_2/sym_fw2.h 	SCR_LOAD_REL (ia, 4),
ia               1648 drivers/scsi/sym53c8xx_2/sym_fw2.h 	SCR_REG_REG (ia, SCR_ADD, 8),
ia               1666 drivers/scsi/sym53c8xx_2/sym_fw2.h 	SCR_STORE_REL (ia, 4),
ia               1688 drivers/scsi/sym53c8xx_2/sym_fw2.h 	SCR_LOAD_ABS (ia, 4),
ia               1691 drivers/scsi/sym53c8xx_2/sym_fw2.h 	SCR_STORE_REL (ia, 4),
ia               1696 drivers/scsi/sym53c8xx_2/sym_fw2.h 	SCR_STORE_REL (ia, 4),
ia               1718 drivers/scsi/sym53c8xx_2/sym_fw2.h 	SCR_LOAD_ABS (ia, 4),
ia                602 fs/afs/cmservice.c 		} ia;
ia                613 fs/afs/cmservice.c 	reply.ia.uuid[0] = call->net->uuid.time_low;
ia                614 fs/afs/cmservice.c 	reply.ia.uuid[1] = htonl(ntohs(call->net->uuid.time_mid));
ia                615 fs/afs/cmservice.c 	reply.ia.uuid[2] = htonl(ntohs(call->net->uuid.time_hi_and_version));
ia                616 fs/afs/cmservice.c 	reply.ia.uuid[3] = htonl((s8) call->net->uuid.clock_seq_hi_and_reserved);
ia                617 fs/afs/cmservice.c 	reply.ia.uuid[4] = htonl((s8) call->net->uuid.clock_seq_low);
ia                619 fs/afs/cmservice.c 		reply.ia.uuid[loop + 5] = htonl((s8) call->net->uuid.node[loop]);
ia                 44 fs/debugfs/inode.c static int debugfs_setattr(struct dentry *dentry, struct iattr *ia)
ia                 48 fs/debugfs/inode.c 	if (ret && (ia->ia_valid & (ATTR_MODE | ATTR_UID | ATTR_GID)))
ia                 50 fs/debugfs/inode.c 	return simple_setattr(dentry, ia);
ia                725 fs/ecryptfs/inode.c static int truncate_upper(struct dentry *dentry, struct iattr *ia,
ia                735 fs/ecryptfs/inode.c 	if (unlikely((ia->ia_size == i_size))) {
ia                744 fs/ecryptfs/inode.c 	if (ia->ia_size > i_size) {
ia                753 fs/ecryptfs/inode.c 				    (ia->ia_size - 1), 1);
ia                760 fs/ecryptfs/inode.c 				    - (ia->ia_size & ~PAGE_MASK));
ia                763 fs/ecryptfs/inode.c 			truncate_setsize(inode, ia->ia_size);
ia                764 fs/ecryptfs/inode.c 			lower_ia->ia_size = ia->ia_size;
ia                777 fs/ecryptfs/inode.c 					    ia->ia_size, num_zeros);
ia                786 fs/ecryptfs/inode.c 		truncate_setsize(inode, ia->ia_size);
ia                799 fs/ecryptfs/inode.c 		    upper_size_to_lower_size(crypt_stat, ia->ia_size);
ia                845 fs/ecryptfs/inode.c 	struct iattr ia = { .ia_valid = ATTR_SIZE, .ia_size = new_length };
ia                853 fs/ecryptfs/inode.c 	rc = truncate_upper(dentry, &ia, &lower_ia);
ia                882 fs/ecryptfs/inode.c static int ecryptfs_setattr(struct dentry *dentry, struct iattr *ia)
ia                936 fs/ecryptfs/inode.c 	rc = setattr_prepare(dentry, ia);
ia                939 fs/ecryptfs/inode.c 	if (ia->ia_valid & ATTR_SIZE) {
ia                940 fs/ecryptfs/inode.c 		rc = ecryptfs_inode_newsize_ok(inode, ia->ia_size);
ia                945 fs/ecryptfs/inode.c 	memcpy(&lower_ia, ia, sizeof(lower_ia));
ia                946 fs/ecryptfs/inode.c 	if (ia->ia_valid & ATTR_FILE)
ia                947 fs/ecryptfs/inode.c 		lower_ia.ia_file = ecryptfs_file_to_lower(ia->ia_file);
ia                948 fs/ecryptfs/inode.c 	if (ia->ia_valid & ATTR_SIZE) {
ia                949 fs/ecryptfs/inode.c 		rc = truncate_upper(dentry, ia, &lower_ia);
ia                 41 fs/fat/file.c  	struct iattr ia;
ia                 66 fs/fat/file.c  	ia.ia_valid = ATTR_MODE | ATTR_CTIME;
ia                 67 fs/fat/file.c  	ia.ia_ctime = current_time(inode);
ia                 69 fs/fat/file.c  		ia.ia_mode = fat_make_mode(sbi, attr, S_IRWXUGO);
ia                 71 fs/fat/file.c  		ia.ia_mode = fat_make_mode(sbi, attr,
ia                 93 fs/fat/file.c  	err = security_inode_setattr(file->f_path.dentry, &ia);
ia                 98 fs/fat/file.c  	err = fat_setattr(file->f_path.dentry, &ia);
ia                102 fs/fat/file.c  	fsnotify_change(file->f_path.dentry, ia.ia_valid);
ia                319 fs/fuse/cuse.c 	struct cuse_init_args *ia = container_of(args, typeof(*ia), ap.args);
ia                320 fs/fuse/cuse.c 	struct fuse_args_pages *ap = &ia->ap;
ia                322 fs/fuse/cuse.c 	struct cuse_init_out *arg = &ia->out;
ia                407 fs/fuse/cuse.c 	kfree(ia);
ia                428 fs/fuse/cuse.c 	struct cuse_init_args *ia;
ia                438 fs/fuse/cuse.c 	ia = kzalloc(sizeof(*ia), GFP_KERNEL);
ia                439 fs/fuse/cuse.c 	if (!ia)
ia                442 fs/fuse/cuse.c 	ap = &ia->ap;
ia                443 fs/fuse/cuse.c 	ia->in.major = FUSE_KERNEL_VERSION;
ia                444 fs/fuse/cuse.c 	ia->in.minor = FUSE_KERNEL_MINOR_VERSION;
ia                445 fs/fuse/cuse.c 	ia->in.flags |= CUSE_UNRESTRICTED_IOCTL;
ia                448 fs/fuse/cuse.c 	ap->args.in_args[0].size = sizeof(ia->in);
ia                449 fs/fuse/cuse.c 	ap->args.in_args[0].value = &ia->in;
ia                451 fs/fuse/cuse.c 	ap->args.out_args[0].size = sizeof(ia->out);
ia                452 fs/fuse/cuse.c 	ap->args.out_args[0].value = &ia->out;
ia                457 fs/fuse/cuse.c 	ap->pages = &ia->page;
ia                458 fs/fuse/cuse.c 	ap->descs = &ia->desc;
ia                459 fs/fuse/cuse.c 	ia->page = page;
ia                460 fs/fuse/cuse.c 	ia->desc.length = ap->args.out_args[1].size;
ia                465 fs/fuse/cuse.c 		kfree(ia);
ia                359 fs/fuse/file.c 	struct fuse_io_args ia;
ia                375 fs/fuse/file.c 		curr_index = wpa->ia.write.in.offset >> PAGE_SHIFT;
ia                376 fs/fuse/file.c 		if (idx_from < curr_index + wpa->ia.ap.num_pages &&
ia                551 fs/fuse/file.c void fuse_read_args_fill(struct fuse_io_args *ia, struct file *file, loff_t pos,
ia                555 fs/fuse/file.c 	struct fuse_args *args = &ia->ap.args;
ia                557 fs/fuse/file.c 	ia->read.in.fh = ff->fh;
ia                558 fs/fuse/file.c 	ia->read.in.offset = pos;
ia                559 fs/fuse/file.c 	ia->read.in.size = count;
ia                560 fs/fuse/file.c 	ia->read.in.flags = file->f_flags;
ia                564 fs/fuse/file.c 	args->in_args[0].size = sizeof(ia->read.in);
ia                565 fs/fuse/file.c 	args->in_args[0].value = &ia->read.in;
ia                652 fs/fuse/file.c 	struct fuse_io_args *ia;
ia                654 fs/fuse/file.c 	ia = kzalloc(sizeof(*ia), GFP_KERNEL);
ia                655 fs/fuse/file.c 	if (ia) {
ia                656 fs/fuse/file.c 		ia->io = io;
ia                657 fs/fuse/file.c 		ia->ap.pages = fuse_pages_alloc(npages, GFP_KERNEL,
ia                658 fs/fuse/file.c 						&ia->ap.descs);
ia                659 fs/fuse/file.c 		if (!ia->ap.pages) {
ia                660 fs/fuse/file.c 			kfree(ia);
ia                661 fs/fuse/file.c 			ia = NULL;
ia                664 fs/fuse/file.c 	return ia;
ia                667 fs/fuse/file.c static void fuse_io_free(struct fuse_io_args *ia)
ia                669 fs/fuse/file.c 	kfree(ia->ap.pages);
ia                670 fs/fuse/file.c 	kfree(ia);
ia                676 fs/fuse/file.c 	struct fuse_io_args *ia = container_of(args, typeof(*ia), ap.args);
ia                677 fs/fuse/file.c 	struct fuse_io_priv *io = ia->io;
ia                680 fs/fuse/file.c 	fuse_release_user_pages(&ia->ap, io->should_dirty);
ia                685 fs/fuse/file.c 		if (ia->write.out.size > ia->write.in.size) {
ia                687 fs/fuse/file.c 		} else if (ia->write.in.size != ia->write.out.size) {
ia                688 fs/fuse/file.c 			pos = ia->write.in.offset - io->offset +
ia                689 fs/fuse/file.c 				ia->write.out.size;
ia                694 fs/fuse/file.c 		if (ia->read.in.size != outsize)
ia                695 fs/fuse/file.c 			pos = ia->read.in.offset - io->offset + outsize;
ia                699 fs/fuse/file.c 	fuse_io_free(ia);
ia                703 fs/fuse/file.c 				   struct fuse_io_args *ia, size_t num_bytes)
ia                706 fs/fuse/file.c 	struct fuse_io_priv *io = ia->io;
ia                714 fs/fuse/file.c 	ia->ap.args.end = fuse_aio_complete_req;
ia                715 fs/fuse/file.c 	err = fuse_simple_background(fc, &ia->ap.args, GFP_KERNEL);
ia                717 fs/fuse/file.c 		fuse_aio_complete_req(fc, &ia->ap.args, err);
ia                722 fs/fuse/file.c static ssize_t fuse_send_read(struct fuse_io_args *ia, loff_t pos, size_t count,
ia                725 fs/fuse/file.c 	struct file *file = ia->io->iocb->ki_filp;
ia                729 fs/fuse/file.c 	fuse_read_args_fill(ia, file, pos, count, FUSE_READ);
ia                731 fs/fuse/file.c 		ia->read.in.read_flags |= FUSE_READ_LOCKOWNER;
ia                732 fs/fuse/file.c 		ia->read.in.lock_owner = fuse_lock_owner_id(fc, owner);
ia                735 fs/fuse/file.c 	if (ia->io->async)
ia                736 fs/fuse/file.c 		return fuse_async_req_send(fc, ia, count);
ia                738 fs/fuse/file.c 	return fuse_simple_request(fc, &ia->ap.args);
ia                787 fs/fuse/file.c 	struct fuse_io_args ia = {
ia                810 fs/fuse/file.c 	fuse_read_args_fill(&ia, file, pos, desc.length, FUSE_READ);
ia                811 fs/fuse/file.c 	res = fuse_simple_request(fc, &ia.ap.args);
ia                818 fs/fuse/file.c 		fuse_short_read(inode, attr_ver, res, &ia.ap);
ia                845 fs/fuse/file.c 	struct fuse_io_args *ia = container_of(args, typeof(*ia), ap.args);
ia                846 fs/fuse/file.c 	struct fuse_args_pages *ap = &ia->ap;
ia                847 fs/fuse/file.c 	size_t count = ia->read.in.size;
ia                861 fs/fuse/file.c 			fuse_short_read(inode, ia->read.attr_ver, num_read, ap);
ia                876 fs/fuse/file.c 	if (ia->ff)
ia                877 fs/fuse/file.c 		fuse_file_put(ia->ff, false, false);
ia                879 fs/fuse/file.c 	fuse_io_free(ia);
ia                882 fs/fuse/file.c static void fuse_send_readpages(struct fuse_io_args *ia, struct file *file)
ia                886 fs/fuse/file.c 	struct fuse_args_pages *ap = &ia->ap;
ia                903 fs/fuse/file.c 	fuse_read_args_fill(ia, file, pos, count, FUSE_READ);
ia                904 fs/fuse/file.c 	ia->read.attr_ver = fuse_get_attr_version(fc);
ia                906 fs/fuse/file.c 		ia->ff = fuse_file_get(ff);
ia                919 fs/fuse/file.c 	struct fuse_io_args *ia;
ia                929 fs/fuse/file.c 	struct fuse_io_args *ia = data->ia;
ia                930 fs/fuse/file.c 	struct fuse_args_pages *ap = &ia->ap;
ia                942 fs/fuse/file.c 		fuse_send_readpages(ia, data->file);
ia                943 fs/fuse/file.c 		data->ia = ia = fuse_io_alloc(NULL, data->max_pages);
ia                944 fs/fuse/file.c 		if (!ia) {
ia                948 fs/fuse/file.c 		ap = &ia->ap;
ia                953 fs/fuse/file.c 		fuse_io_free(ia);
ia                982 fs/fuse/file.c 	data.ia = fuse_io_alloc(NULL, data.max_pages);
ia                984 fs/fuse/file.c 	if (!data.ia)
ia                989 fs/fuse/file.c 		if (data.ia->ap.num_pages)
ia                990 fs/fuse/file.c 			fuse_send_readpages(data.ia, file);
ia                992 fs/fuse/file.c 			fuse_io_free(data.ia);
ia               1019 fs/fuse/file.c static void fuse_write_args_fill(struct fuse_io_args *ia, struct fuse_file *ff,
ia               1022 fs/fuse/file.c 	struct fuse_args *args = &ia->ap.args;
ia               1024 fs/fuse/file.c 	ia->write.in.fh = ff->fh;
ia               1025 fs/fuse/file.c 	ia->write.in.offset = pos;
ia               1026 fs/fuse/file.c 	ia->write.in.size = count;
ia               1033 fs/fuse/file.c 		args->in_args[0].size = sizeof(ia->write.in);
ia               1034 fs/fuse/file.c 	args->in_args[0].value = &ia->write.in;
ia               1037 fs/fuse/file.c 	args->out_args[0].size = sizeof(ia->write.out);
ia               1038 fs/fuse/file.c 	args->out_args[0].value = &ia->write.out;
ia               1053 fs/fuse/file.c static ssize_t fuse_send_write(struct fuse_io_args *ia, loff_t pos,
ia               1056 fs/fuse/file.c 	struct kiocb *iocb = ia->io->iocb;
ia               1060 fs/fuse/file.c 	struct fuse_write_in *inarg = &ia->write.in;
ia               1063 fs/fuse/file.c 	fuse_write_args_fill(ia, ff, pos, count);
ia               1070 fs/fuse/file.c 	if (ia->io->async)
ia               1071 fs/fuse/file.c 		return fuse_async_req_send(fc, ia, count);
ia               1073 fs/fuse/file.c 	err = fuse_simple_request(fc, &ia->ap.args);
ia               1074 fs/fuse/file.c 	if (!err && ia->write.out.size > count)
ia               1077 fs/fuse/file.c 	return err ?: ia->write.out.size;
ia               1097 fs/fuse/file.c static ssize_t fuse_send_write_pages(struct fuse_io_args *ia,
ia               1101 fs/fuse/file.c 	struct fuse_args_pages *ap = &ia->ap;
ia               1111 fs/fuse/file.c 	fuse_write_args_fill(ia, ff, pos, count);
ia               1112 fs/fuse/file.c 	ia->write.in.flags = fuse_write_flags(iocb);
ia               1115 fs/fuse/file.c 	if (!err && ia->write.out.size > count)
ia               1119 fs/fuse/file.c 	count = ia->write.out.size;
ia               1228 fs/fuse/file.c 		struct fuse_io_args ia = {};
ia               1229 fs/fuse/file.c 		struct fuse_args_pages *ap = &ia.ap;
ia               1243 fs/fuse/file.c 			err = fuse_send_write_pages(&ia, iocb, inode,
ia               1246 fs/fuse/file.c 				size_t num_written = ia.write.out.size;
ia               1437 fs/fuse/file.c 	struct fuse_io_args *ia;
ia               1441 fs/fuse/file.c 	ia = fuse_io_alloc(io, max_pages);
ia               1442 fs/fuse/file.c 	if (!ia)
ia               1445 fs/fuse/file.c 	ia->io = io;
ia               1460 fs/fuse/file.c 		err = fuse_get_user_pages(&ia->ap, iter, &nbytes, write,
ia               1467 fs/fuse/file.c 				ia->write.in.write_flags |= FUSE_WRITE_KILL_PRIV;
ia               1469 fs/fuse/file.c 			nres = fuse_send_write(ia, pos, nbytes, owner);
ia               1471 fs/fuse/file.c 			nres = fuse_send_read(ia, pos, nbytes, owner);
ia               1475 fs/fuse/file.c 			fuse_release_user_pages(&ia->ap, io->should_dirty);
ia               1476 fs/fuse/file.c 			fuse_io_free(ia);
ia               1478 fs/fuse/file.c 		ia = NULL;
ia               1495 fs/fuse/file.c 			ia = fuse_io_alloc(io, max_pages);
ia               1496 fs/fuse/file.c 			if (!ia)
ia               1500 fs/fuse/file.c 	if (ia)
ia               1501 fs/fuse/file.c 		fuse_io_free(ia);
ia               1595 fs/fuse/file.c 	struct fuse_args_pages *ap = &wpa->ia.ap;
ia               1601 fs/fuse/file.c 	if (wpa->ia.ff)
ia               1602 fs/fuse/file.c 		fuse_file_put(wpa->ia.ff, false, false);
ia               1611 fs/fuse/file.c 	struct fuse_args_pages *ap = &wpa->ia.ap;
ia               1634 fs/fuse/file.c 	struct fuse_write_in *inarg = &wpa->ia.write.in;
ia               1635 fs/fuse/file.c 	struct fuse_args *args = &wpa->ia.ap.args;
ia               1636 fs/fuse/file.c 	__u64 data_size = wpa->ia.ap.num_pages * PAGE_SIZE;
ia               1709 fs/fuse/file.c 		container_of(args, typeof(*wpa), ia.ap.args);
ia               1717 fs/fuse/file.c 		struct fuse_write_in *inarg = &wpa->ia.write.in;
ia               1722 fs/fuse/file.c 		next->ia.ff = fuse_file_get(wpa->ia.ff);
ia               1802 fs/fuse/file.c 		ap = &wpa->ia.ap;
ia               1830 fs/fuse/file.c 	ap = &wpa->ia.ap;
ia               1837 fs/fuse/file.c 	wpa->ia.ff = fuse_write_file_get(fc, fi);
ia               1838 fs/fuse/file.c 	if (!wpa->ia.ff)
ia               1841 fs/fuse/file.c 	fuse_write_args_fill(&wpa->ia, wpa->ia.ff, page_offset(page), 0);
ia               1844 fs/fuse/file.c 	wpa->ia.write.in.write_flags |= FUSE_WRITE_CACHE;
ia               1910 fs/fuse/file.c 	struct fuse_args_pages *ap = &data->wpa->ia.ap;
ia               1939 fs/fuse/file.c 	int num_pages = wpa->ia.ap.num_pages;
ia               1942 fs/fuse/file.c 	wpa->ia.ff = fuse_file_get(data->ff);
ia               1965 fs/fuse/file.c 	struct fuse_args_pages *new_ap = &new_wpa->ia.ap;
ia               1983 fs/fuse/file.c 		curr_index = tmp->ia.write.in.offset >> PAGE_SHIFT;
ia               1985 fs/fuse/file.c 			WARN_ON(tmp->ia.ap.num_pages != 1);
ia               1986 fs/fuse/file.c 			swap(tmp->ia.ap.pages[0], new_ap->pages[0]);
ia               2015 fs/fuse/file.c 	struct fuse_args_pages *ap = &wpa->ia.ap;
ia               2078 fs/fuse/file.c 		ap = &wpa->ia.ap;
ia               2079 fs/fuse/file.c 		fuse_write_args_fill(&wpa->ia, data->ff, page_offset(page), 0);
ia               2080 fs/fuse/file.c 		wpa->ia.write.in.write_flags |= FUSE_WRITE_CACHE;
ia               2151 fs/fuse/file.c 		WARN_ON(!data.wpa->ia.ap.num_pages);
ia                841 fs/fuse/fuse_i.h void fuse_read_args_fill(struct fuse_io_args *ia, struct file *file, loff_t pos,
ia                887 fs/fuse/inode.c 	struct fuse_init_args *ia = container_of(args, typeof(*ia), args);
ia                888 fs/fuse/inode.c 	struct fuse_init_out *arg = &ia->out;
ia                967 fs/fuse/inode.c 	kfree(ia);
ia                975 fs/fuse/inode.c 	struct fuse_init_args *ia;
ia                977 fs/fuse/inode.c 	ia = kzalloc(sizeof(*ia), GFP_KERNEL | __GFP_NOFAIL);
ia                979 fs/fuse/inode.c 	ia->in.major = FUSE_KERNEL_VERSION;
ia                980 fs/fuse/inode.c 	ia->in.minor = FUSE_KERNEL_MINOR_VERSION;
ia                981 fs/fuse/inode.c 	ia->in.max_readahead = fc->sb->s_bdi->ra_pages * PAGE_SIZE;
ia                982 fs/fuse/inode.c 	ia->in.flags |=
ia                992 fs/fuse/inode.c 	ia->args.opcode = FUSE_INIT;
ia                993 fs/fuse/inode.c 	ia->args.in_numargs = 1;
ia                994 fs/fuse/inode.c 	ia->args.in_args[0].size = sizeof(ia->in);
ia                995 fs/fuse/inode.c 	ia->args.in_args[0].value = &ia->in;
ia                996 fs/fuse/inode.c 	ia->args.out_numargs = 1;
ia               1000 fs/fuse/inode.c 	ia->args.out_argvar = 1;
ia               1001 fs/fuse/inode.c 	ia->args.out_args[0].size = sizeof(ia->out);
ia               1002 fs/fuse/inode.c 	ia->args.out_args[0].value = &ia->out;
ia               1003 fs/fuse/inode.c 	ia->args.force = true;
ia               1004 fs/fuse/inode.c 	ia->args.nocreds = true;
ia               1005 fs/fuse/inode.c 	ia->args.end = process_init_reply;
ia               1007 fs/fuse/inode.c 	if (fuse_simple_background(fc, &ia->args, GFP_KERNEL) != 0)
ia               1008 fs/fuse/inode.c 		process_init_reply(fc, &ia->args, -ENOTCONN);
ia                324 fs/fuse/readdir.c 	struct fuse_io_args ia = {};
ia                325 fs/fuse/readdir.c 	struct fuse_args_pages *ap = &ia.ap;
ia                341 fs/fuse/readdir.c 		fuse_read_args_fill(&ia, file, ctx->pos, PAGE_SIZE,
ia                344 fs/fuse/readdir.c 		fuse_read_args_fill(&ia, file, ctx->pos, PAGE_SIZE,
ia                 74 fs/ntfs/dir.c  	INDEX_ALLOCATION *ia;
ia                319 fs/ntfs/dir.c  	ia = (INDEX_ALLOCATION*)(kaddr + ((vcn <<
ia                322 fs/ntfs/dir.c  	if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_SIZE) {
ia                328 fs/ntfs/dir.c  	if (unlikely(!ntfs_is_indx_record(ia->magic))) {
ia                334 fs/ntfs/dir.c  	if (sle64_to_cpu(ia->index_block_vcn) != vcn) {
ia                339 fs/ntfs/dir.c  				sle64_to_cpu(ia->index_block_vcn),
ia                343 fs/ntfs/dir.c  	if (le32_to_cpu(ia->index.allocated_size) + 0x18 !=
ia                350 fs/ntfs/dir.c  				le32_to_cpu(ia->index.allocated_size) + 0x18,
ia                354 fs/ntfs/dir.c  	index_end = (u8*)ia + dir_ni->itype.index.block_size;
ia                363 fs/ntfs/dir.c  	index_end = (u8*)&ia->index + le32_to_cpu(ia->index.index_length);
ia                364 fs/ntfs/dir.c  	if (index_end > (u8*)ia + dir_ni->itype.index.block_size) {
ia                371 fs/ntfs/dir.c  	ie = (INDEX_ENTRY*)((u8*)&ia->index +
ia                372 fs/ntfs/dir.c  			le32_to_cpu(ia->index.entries_offset));
ia                380 fs/ntfs/dir.c  		if ((u8*)ie < (u8*)ia || (u8*)ie +
ia                534 fs/ntfs/dir.c  		if ((ia->index.flags & NODE_MASK) == LEAF_NODE) {
ia                627 fs/ntfs/dir.c  	INDEX_ALLOCATION *ia;
ia                797 fs/ntfs/dir.c  	ia = (INDEX_ALLOCATION*)(kaddr + ((vcn <<
ia                800 fs/ntfs/dir.c  	if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_SIZE) {
ia                806 fs/ntfs/dir.c  	if (unlikely(!ntfs_is_indx_record(ia->magic))) {
ia                812 fs/ntfs/dir.c  	if (sle64_to_cpu(ia->index_block_vcn) != vcn) {
ia                817 fs/ntfs/dir.c  				sle64_to_cpu(ia->index_block_vcn),
ia                821 fs/ntfs/dir.c  	if (le32_to_cpu(ia->index.allocated_size) + 0x18 !=
ia                828 fs/ntfs/dir.c  				le32_to_cpu(ia->index.allocated_size) + 0x18,
ia                832 fs/ntfs/dir.c  	index_end = (u8*)ia + dir_ni->itype.index.block_size;
ia                841 fs/ntfs/dir.c  	index_end = (u8*)&ia->index + le32_to_cpu(ia->index.index_length);
ia                842 fs/ntfs/dir.c  	if (index_end > (u8*)ia + dir_ni->itype.index.block_size) {
ia                849 fs/ntfs/dir.c  	ie = (INDEX_ENTRY*)((u8*)&ia->index +
ia                850 fs/ntfs/dir.c  			le32_to_cpu(ia->index.entries_offset));
ia                858 fs/ntfs/dir.c  		if ((u8*)ie < (u8*)ia || (u8*)ie +
ia                944 fs/ntfs/dir.c  		if ((ia->index.flags & NODE_MASK) == LEAF_NODE) {
ia               1096 fs/ntfs/dir.c  	INDEX_ALLOCATION *ia;
ia               1296 fs/ntfs/dir.c  	ia = (INDEX_ALLOCATION*)(kaddr + (ia_pos & ~PAGE_MASK &
ia               1299 fs/ntfs/dir.c  	if (unlikely((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_SIZE)) {
ia               1305 fs/ntfs/dir.c  	if (unlikely(!ntfs_is_indx_record(ia->magic))) {
ia               1312 fs/ntfs/dir.c  	if (unlikely(sle64_to_cpu(ia->index_block_vcn) != (ia_pos &
ia               1319 fs/ntfs/dir.c  				sle64_to_cpu(ia->index_block_vcn),
ia               1324 fs/ntfs/dir.c  	if (unlikely(le32_to_cpu(ia->index.allocated_size) + 0x18 !=
ia               1332 fs/ntfs/dir.c  				le32_to_cpu(ia->index.allocated_size) + 0x18,
ia               1336 fs/ntfs/dir.c  	index_end = (u8*)ia + ndir->itype.index.block_size;
ia               1346 fs/ntfs/dir.c  	index_end = (u8*)&ia->index + le32_to_cpu(ia->index.index_length);
ia               1347 fs/ntfs/dir.c  	if (unlikely(index_end > (u8*)ia + ndir->itype.index.block_size)) {
ia               1355 fs/ntfs/dir.c  	ie = (INDEX_ENTRY*)((u8*)&ia->index +
ia               1356 fs/ntfs/dir.c  			le32_to_cpu(ia->index.entries_offset));
ia               1365 fs/ntfs/dir.c  				(unsigned long long)((u8*)ie - (u8*)ia));
ia               1367 fs/ntfs/dir.c  		if (unlikely((u8*)ie < (u8*)ia || (u8*)ie +
ia               1376 fs/ntfs/dir.c  		if (ia_pos - ia_start > (u8*)ie - (u8*)ia)
ia               1379 fs/ntfs/dir.c  		actor->pos = (u8*)ie - (u8*)ia +
ia               1380 fs/ntfs/dir.c  				(sle64_to_cpu(ia->index_block_vcn) <<
ia                116 fs/ntfs/index.c 	INDEX_ALLOCATION *ia;
ia                199 fs/ntfs/index.c 			ictx->ia = NULL;
ia                276 fs/ntfs/index.c 	ia = (INDEX_ALLOCATION*)(kaddr + ((vcn <<
ia                279 fs/ntfs/index.c 	if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_SIZE) {
ia                285 fs/ntfs/index.c 	if (unlikely(!ntfs_is_indx_record(ia->magic))) {
ia                291 fs/ntfs/index.c 	if (sle64_to_cpu(ia->index_block_vcn) != vcn) {
ia                296 fs/ntfs/index.c 				sle64_to_cpu(ia->index_block_vcn),
ia                300 fs/ntfs/index.c 	if (le32_to_cpu(ia->index.allocated_size) + 0x18 !=
ia                307 fs/ntfs/index.c 				le32_to_cpu(ia->index.allocated_size) + 0x18,
ia                311 fs/ntfs/index.c 	index_end = (u8*)ia + idx_ni->itype.index.block_size;
ia                320 fs/ntfs/index.c 	index_end = (u8*)&ia->index + le32_to_cpu(ia->index.index_length);
ia                321 fs/ntfs/index.c 	if (index_end > (u8*)ia + idx_ni->itype.index.block_size) {
ia                328 fs/ntfs/index.c 	ie = (INDEX_ENTRY*)((u8*)&ia->index +
ia                329 fs/ntfs/index.c 			le32_to_cpu(ia->index.entries_offset));
ia                337 fs/ntfs/index.c 		if ((u8*)ie < (u8*)ia || (u8*)ie +
ia                368 fs/ntfs/index.c 			ictx->ia = ia;
ia                402 fs/ntfs/index.c 	if ((ia->index.flags & NODE_MASK) == LEAF_NODE) {
ia                 70 fs/ntfs/index.h 	INDEX_ALLOCATION *ia;
ia                129 fs/ntfs/index.h 				(u8*)ictx->ia - (u8*)page_address(ictx->page));
ia                161 fs/ramfs/file-nommu.c static int ramfs_nommu_setattr(struct dentry *dentry, struct iattr *ia)
ia                164 fs/ramfs/file-nommu.c 	unsigned int old_ia_valid = ia->ia_valid;
ia                168 fs/ramfs/file-nommu.c 	ret = setattr_prepare(dentry, ia);
ia                173 fs/ramfs/file-nommu.c 	if (ia->ia_valid & ATTR_SIZE) {
ia                176 fs/ramfs/file-nommu.c 		if (ia->ia_size != size) {
ia                177 fs/ramfs/file-nommu.c 			ret = ramfs_nommu_resize(inode, ia->ia_size, size);
ia                178 fs/ramfs/file-nommu.c 			if (ret < 0 || ia->ia_valid == ATTR_SIZE)
ia                184 fs/ramfs/file-nommu.c 			ia->ia_valid |= ATTR_MTIME|ATTR_CTIME;
ia                188 fs/ramfs/file-nommu.c 	setattr_copy(inode, ia);
ia                190 fs/ramfs/file-nommu.c 	ia->ia_valid = old_ia_valid;
ia                357 fs/xfs/xfs_aops.c 	struct xfs_ioend	*ia;
ia                360 fs/xfs/xfs_aops.c 	ia = container_of(a, struct xfs_ioend, io_list);
ia                362 fs/xfs/xfs_aops.c 	if (ia->io_offset < ib->io_offset)
ia                364 fs/xfs/xfs_aops.c 	else if (ia->io_offset > ib->io_offset)
ia                 23 include/linux/quotaops.h static inline bool is_quota_modification(struct inode *inode, struct iattr *ia)
ia                 25 include/linux/quotaops.h 	return (ia->ia_valid & ATTR_SIZE) ||
ia                 26 include/linux/quotaops.h 		(ia->ia_valid & ATTR_UID && !uid_eq(ia->ia_uid, inode->i_uid)) ||
ia                 27 include/linux/quotaops.h 		(ia->ia_valid & ATTR_GID && !gid_eq(ia->ia_gid, inode->i_gid));
ia                412 net/bluetooth/smp.c 		  const bdaddr_t *ia, u8 _rat, const bdaddr_t *ra, u8 res[16])
ia                418 net/bluetooth/smp.c 	SMP_DBG("iat %u ia %6phN rat %u ra %6phN", _iat, ia, _rat, ra);
ia                443 net/bluetooth/smp.c 	memcpy(p2 + 6, ia, 6);
ia               3588 net/bluetooth/smp.c 	const bdaddr_t ia = { { 0xa6, 0xa5, 0xa4, 0xa3, 0xa2, 0xa1 } };
ia               3595 net/bluetooth/smp.c 	err = smp_c1(k, r, preq, pres, _iat, &ia, _rat, &ra, res);
ia                162 net/sunrpc/xprtrdma/frwr_ops.c int frwr_init_mr(struct rpcrdma_ia *ia, struct rpcrdma_mr *mr)
ia                164 net/sunrpc/xprtrdma/frwr_ops.c 	unsigned int depth = ia->ri_max_frwr_depth;
ia                172 net/sunrpc/xprtrdma/frwr_ops.c 	frmr = ib_alloc_mr(ia->ri_pd, ia->ri_mrtype, depth);
ia                217 net/sunrpc/xprtrdma/frwr_ops.c int frwr_open(struct rpcrdma_ia *ia, struct rpcrdma_ep *ep)
ia                219 net/sunrpc/xprtrdma/frwr_ops.c 	struct ib_device_attr *attrs = &ia->ri_id->device->attrs;
ia                222 net/sunrpc/xprtrdma/frwr_ops.c 	ia->ri_mrtype = IB_MR_TYPE_MEM_REG;
ia                224 net/sunrpc/xprtrdma/frwr_ops.c 		ia->ri_mrtype = IB_MR_TYPE_SG_GAPS;
ia                231 net/sunrpc/xprtrdma/frwr_ops.c 		ia->ri_max_frwr_depth = attrs->max_sge_rd;
ia                233 net/sunrpc/xprtrdma/frwr_ops.c 		ia->ri_max_frwr_depth = attrs->max_fast_reg_page_list_len;
ia                234 net/sunrpc/xprtrdma/frwr_ops.c 	if (ia->ri_max_frwr_depth > RPCRDMA_MAX_DATA_SEGS)
ia                235 net/sunrpc/xprtrdma/frwr_ops.c 		ia->ri_max_frwr_depth = RPCRDMA_MAX_DATA_SEGS;
ia                237 net/sunrpc/xprtrdma/frwr_ops.c 		__func__, ia->ri_max_frwr_depth);
ia                253 net/sunrpc/xprtrdma/frwr_ops.c 	if (ia->ri_max_frwr_depth < RPCRDMA_MAX_DATA_SEGS) {
ia                254 net/sunrpc/xprtrdma/frwr_ops.c 		delta = RPCRDMA_MAX_DATA_SEGS - ia->ri_max_frwr_depth;
ia                257 net/sunrpc/xprtrdma/frwr_ops.c 			delta -= ia->ri_max_frwr_depth;
ia                261 net/sunrpc/xprtrdma/frwr_ops.c 	max_qp_wr = ia->ri_id->device->attrs.max_qp_wr;
ia                281 net/sunrpc/xprtrdma/frwr_ops.c 	ia->ri_max_segs =
ia                282 net/sunrpc/xprtrdma/frwr_ops.c 		DIV_ROUND_UP(RPCRDMA_MAX_DATA_SEGS, ia->ri_max_frwr_depth);
ia                284 net/sunrpc/xprtrdma/frwr_ops.c 	ia->ri_max_segs += 2;
ia                285 net/sunrpc/xprtrdma/frwr_ops.c 	if (ia->ri_max_segs > RPCRDMA_MAX_HDR_SEGS)
ia                286 net/sunrpc/xprtrdma/frwr_ops.c 		ia->ri_max_segs = RPCRDMA_MAX_HDR_SEGS;
ia                301 net/sunrpc/xprtrdma/frwr_ops.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                304 net/sunrpc/xprtrdma/frwr_ops.c 		     (ia->ri_max_segs - 2) * ia->ri_max_frwr_depth);
ia                327 net/sunrpc/xprtrdma/frwr_ops.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                333 net/sunrpc/xprtrdma/frwr_ops.c 	if (nsegs > ia->ri_max_frwr_depth)
ia                334 net/sunrpc/xprtrdma/frwr_ops.c 		nsegs = ia->ri_max_frwr_depth;
ia                347 net/sunrpc/xprtrdma/frwr_ops.c 		if (ia->ri_mrtype == IB_MR_TYPE_SG_GAPS)
ia                356 net/sunrpc/xprtrdma/frwr_ops.c 	dma_nents = ib_dma_map_sg(ia->ri_id->device, mr->mr_sg, mr->mr_nents,
ia                423 net/sunrpc/xprtrdma/frwr_ops.c int frwr_send(struct rpcrdma_ia *ia, struct rpcrdma_req *req)
ia                447 net/sunrpc/xprtrdma/frwr_ops.c 	return ib_post_send(ia->ri_id->qp, post_wr, NULL);
ia                410 net/sunrpc/xprtrdma/transport.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                419 net/sunrpc/xprtrdma/transport.c 	if (test_and_clear_bit(RPCRDMA_IAF_REMOVING, &ia->ri_flags)) {
ia                420 net/sunrpc/xprtrdma/transport.c 		rpcrdma_ia_remove(ia);
ia                426 net/sunrpc/xprtrdma/transport.c 	rpcrdma_ep_disconnect(ep, ia);
ia                 94 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                 99 net/sunrpc/xprtrdma/verbs.c 	ib_drain_rq(ia->ri_id->qp);
ia                104 net/sunrpc/xprtrdma/verbs.c 	ib_drain_sq(ia->ri_id->qp);
ia                221 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                231 net/sunrpc/xprtrdma/verbs.c 		ia->ri_async_rc = 0;
ia                232 net/sunrpc/xprtrdma/verbs.c 		complete(&ia->ri_done);
ia                235 net/sunrpc/xprtrdma/verbs.c 		ia->ri_async_rc = -EPROTO;
ia                236 net/sunrpc/xprtrdma/verbs.c 		complete(&ia->ri_done);
ia                239 net/sunrpc/xprtrdma/verbs.c 		ia->ri_async_rc = -ENETUNREACH;
ia                240 net/sunrpc/xprtrdma/verbs.c 		complete(&ia->ri_done);
ia                245 net/sunrpc/xprtrdma/verbs.c 			ia->ri_id->device->name,
ia                248 net/sunrpc/xprtrdma/verbs.c 		init_completion(&ia->ri_remove_done);
ia                249 net/sunrpc/xprtrdma/verbs.c 		set_bit(RPCRDMA_IAF_REMOVING, &ia->ri_flags);
ia                252 net/sunrpc/xprtrdma/verbs.c 		wait_for_completion(&ia->ri_remove_done);
ia                254 net/sunrpc/xprtrdma/verbs.c 		ia->ri_id = NULL;
ia                289 net/sunrpc/xprtrdma/verbs.c 		ia->ri_id->device->name, rdma_event_msg(event->event));
ia                294 net/sunrpc/xprtrdma/verbs.c rpcrdma_create_id(struct rpcrdma_xprt *xprt, struct rpcrdma_ia *ia)
ia                302 net/sunrpc/xprtrdma/verbs.c 	init_completion(&ia->ri_done);
ia                309 net/sunrpc/xprtrdma/verbs.c 	ia->ri_async_rc = -ETIMEDOUT;
ia                315 net/sunrpc/xprtrdma/verbs.c 	rc = wait_for_completion_interruptible_timeout(&ia->ri_done, wtimeout);
ia                321 net/sunrpc/xprtrdma/verbs.c 	rc = ia->ri_async_rc;
ia                325 net/sunrpc/xprtrdma/verbs.c 	ia->ri_async_rc = -ETIMEDOUT;
ia                329 net/sunrpc/xprtrdma/verbs.c 	rc = wait_for_completion_interruptible_timeout(&ia->ri_done, wtimeout);
ia                334 net/sunrpc/xprtrdma/verbs.c 	rc = ia->ri_async_rc;
ia                359 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_ia *ia = &xprt->rx_ia;
ia                362 net/sunrpc/xprtrdma/verbs.c 	ia->ri_id = rpcrdma_create_id(xprt, ia);
ia                363 net/sunrpc/xprtrdma/verbs.c 	if (IS_ERR(ia->ri_id)) {
ia                364 net/sunrpc/xprtrdma/verbs.c 		rc = PTR_ERR(ia->ri_id);
ia                368 net/sunrpc/xprtrdma/verbs.c 	ia->ri_pd = ib_alloc_pd(ia->ri_id->device, 0);
ia                369 net/sunrpc/xprtrdma/verbs.c 	if (IS_ERR(ia->ri_pd)) {
ia                370 net/sunrpc/xprtrdma/verbs.c 		rc = PTR_ERR(ia->ri_pd);
ia                377 net/sunrpc/xprtrdma/verbs.c 		if (frwr_is_supported(ia->ri_id->device))
ia                382 net/sunrpc/xprtrdma/verbs.c 		       ia->ri_id->device->name, xprt_rdma_memreg_strategy);
ia                390 net/sunrpc/xprtrdma/verbs.c 	rpcrdma_ia_close(ia);
ia                402 net/sunrpc/xprtrdma/verbs.c rpcrdma_ia_remove(struct rpcrdma_ia *ia)
ia                404 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_xprt *r_xprt = container_of(ia, struct rpcrdma_xprt,
ia                419 net/sunrpc/xprtrdma/verbs.c 	if (ia->ri_id->qp) {
ia                421 net/sunrpc/xprtrdma/verbs.c 		rdma_destroy_qp(ia->ri_id);
ia                422 net/sunrpc/xprtrdma/verbs.c 		ia->ri_id->qp = NULL;
ia                439 net/sunrpc/xprtrdma/verbs.c 	ib_dealloc_pd(ia->ri_pd);
ia                440 net/sunrpc/xprtrdma/verbs.c 	ia->ri_pd = NULL;
ia                443 net/sunrpc/xprtrdma/verbs.c 	complete(&ia->ri_remove_done);
ia                454 net/sunrpc/xprtrdma/verbs.c rpcrdma_ia_close(struct rpcrdma_ia *ia)
ia                456 net/sunrpc/xprtrdma/verbs.c 	if (ia->ri_id != NULL && !IS_ERR(ia->ri_id)) {
ia                457 net/sunrpc/xprtrdma/verbs.c 		if (ia->ri_id->qp)
ia                458 net/sunrpc/xprtrdma/verbs.c 			rdma_destroy_qp(ia->ri_id);
ia                459 net/sunrpc/xprtrdma/verbs.c 		rdma_destroy_id(ia->ri_id);
ia                461 net/sunrpc/xprtrdma/verbs.c 	ia->ri_id = NULL;
ia                464 net/sunrpc/xprtrdma/verbs.c 	if (ia->ri_pd && !IS_ERR(ia->ri_pd))
ia                465 net/sunrpc/xprtrdma/verbs.c 		ib_dealloc_pd(ia->ri_pd);
ia                466 net/sunrpc/xprtrdma/verbs.c 	ia->ri_pd = NULL;
ia                478 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                488 net/sunrpc/xprtrdma/verbs.c 	max_sge = min_t(unsigned int, ia->ri_id->device->attrs.max_send_sge,
ia                494 net/sunrpc/xprtrdma/verbs.c 	ia->ri_max_send_sges = max_sge;
ia                496 net/sunrpc/xprtrdma/verbs.c 	rc = frwr_open(ia, ep);
ia                523 net/sunrpc/xprtrdma/verbs.c 	sendcq = ib_alloc_cq_any(ia->ri_id->device, NULL,
ia                531 net/sunrpc/xprtrdma/verbs.c 	recvcq = ib_alloc_cq_any(ia->ri_id->device, NULL,
ia                557 net/sunrpc/xprtrdma/verbs.c 		min_t(int, U8_MAX, ia->ri_id->device->attrs.max_qp_rd_atom);
ia                588 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                590 net/sunrpc/xprtrdma/verbs.c 	if (ia->ri_id && ia->ri_id->qp) {
ia                591 net/sunrpc/xprtrdma/verbs.c 		rpcrdma_ep_disconnect(ep, ia);
ia                592 net/sunrpc/xprtrdma/verbs.c 		rdma_destroy_qp(ia->ri_id);
ia                593 net/sunrpc/xprtrdma/verbs.c 		ia->ri_id->qp = NULL;
ia                609 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                628 net/sunrpc/xprtrdma/verbs.c 	err = rdma_create_qp(ia->ri_id, ia->ri_pd, qp_init_attr);
ia                640 net/sunrpc/xprtrdma/verbs.c 	rpcrdma_ia_close(ia);
ia                648 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                654 net/sunrpc/xprtrdma/verbs.c 	rpcrdma_ep_disconnect(&r_xprt->rx_ep, ia);
ia                657 net/sunrpc/xprtrdma/verbs.c 	id = rpcrdma_create_id(r_xprt, ia);
ia                671 net/sunrpc/xprtrdma/verbs.c 	if (ia->ri_id->device != id->device) {
ia                676 net/sunrpc/xprtrdma/verbs.c 	err = rdma_create_qp(id, ia->ri_pd, qp_init_attr);
ia                682 net/sunrpc/xprtrdma/verbs.c 	old = ia->ri_id;
ia                683 net/sunrpc/xprtrdma/verbs.c 	ia->ri_id = id;
ia                696 net/sunrpc/xprtrdma/verbs.c rpcrdma_ep_connect(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia)
ia                698 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_xprt *r_xprt = container_of(ia, struct rpcrdma_xprt,
ia                709 net/sunrpc/xprtrdma/verbs.c 		rc = rdma_create_qp(ia->ri_id, ia->ri_pd, &qp_init_attr);
ia                731 net/sunrpc/xprtrdma/verbs.c 	rc = rdma_connect(ia->ri_id, &ep->rep_remote_cma);
ia                767 net/sunrpc/xprtrdma/verbs.c rpcrdma_ep_disconnect(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia)
ia                774 net/sunrpc/xprtrdma/verbs.c 	rc = rdma_disconnect(ia->ri_id);
ia                812 net/sunrpc/xprtrdma/verbs.c static struct rpcrdma_sendctx *rpcrdma_sendctx_create(struct rpcrdma_ia *ia)
ia                816 net/sunrpc/xprtrdma/verbs.c 	sc = kzalloc(struct_size(sc, sc_sges, ia->ri_max_send_sges),
ia                949 net/sunrpc/xprtrdma/verbs.c 	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
ia                952 net/sunrpc/xprtrdma/verbs.c 	for (count = 0; count < ia->ri_max_segs; count++) {
ia                960 net/sunrpc/xprtrdma/verbs.c 		rc = frwr_init_mr(ia, mr);
ia               1468 net/sunrpc/xprtrdma/verbs.c rpcrdma_ep_post(struct rpcrdma_ia *ia,
ia               1483 net/sunrpc/xprtrdma/verbs.c 	rc = frwr_send(ia, req);
ia                466 net/sunrpc/xprtrdma/xprt_rdma.h void rpcrdma_ia_remove(struct rpcrdma_ia *ia);
ia                550 net/sunrpc/xprtrdma/xprt_rdma.h int frwr_open(struct rpcrdma_ia *ia, struct rpcrdma_ep *ep);
ia                551 net/sunrpc/xprtrdma/xprt_rdma.h int frwr_init_mr(struct rpcrdma_ia *ia, struct rpcrdma_mr *mr);
ia                558 net/sunrpc/xprtrdma/xprt_rdma.h int frwr_send(struct rpcrdma_ia *ia, struct rpcrdma_req *req);
ia                668 tools/perf/util/annotate.c 	const struct ins *ia = a;
ia                671 tools/perf/util/annotate.c 	return strcmp(ia->name, ib->name);
ia                 28 tools/perf/util/s390-cpumsf-kernel.h 	unsigned long long ia;	    /* Instruction Address		 */
ia                261 tools/perf/util/s390-cpumsf.c 	local.ia = be64toh(basicp->ia);
ia                278 tools/perf/util/s390-cpumsf.c 		      basic->AS, basic->prim_asn, basic->ia, basic->CL,
ia                507 tools/perf/util/s390-cpumsf.c 				.ip = basic->ia,