ia 433 arch/arm/mach-omap2/clockdomain.c int clkdm_register_autodeps(struct clkdm_autodep *ia) ia 440 arch/arm/mach-omap2/clockdomain.c if (!ia) ia 446 arch/arm/mach-omap2/clockdomain.c autodeps = ia; ia 183 arch/arm/mach-omap2/clockdomain.h int clkdm_register_autodeps(struct clkdm_autodep *ia); ia 3516 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t ia:5; ia 3522 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t ia:5; ia 3541 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t ia:8; ia 3545 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t ia:8; ia 248 arch/powerpc/kvm/e500_mmu.c unsigned int ia; ia 251 arch/powerpc/kvm/e500_mmu.c ia = (ea >> 2) & 0x1; ia 256 arch/powerpc/kvm/e500_mmu.c if (ia) { ia 114 arch/s390/include/asm/cpu_mf.h unsigned long long ia; /* Instruction Address */ ia 60 arch/s390/include/asm/nmi.h u64 ia : 1; /* 23 psw instruction address validity */ ia 50 arch/s390/include/asm/ptrace.h unsigned long ia : 64; /* Instruction Address */ ia 284 arch/s390/kernel/nmi.c if (!mci.ia && !umode) { ia 292 arch/s390/kernel/nmi.c if (!mci.ms || !mci.pm || !mci.ia) ia 1093 arch/s390/kernel/perf_cpum_sf.c psw_bits(regs.psw).ia = basic->ia; ia 2101 arch/sparc/mm/init_64.c unsigned int *ia = (unsigned int *)(unsigned long)*start; ia 2103 arch/sparc/mm/init_64.c ia[0] = (ia[0] & ~0x3fffff) | (high_bits >> 10); ia 2104 arch/sparc/mm/init_64.c __asm__ __volatile__("flush %0" : : "r" (ia)); ia 2106 arch/sparc/mm/init_64.c ia[1] = (ia[1] & ~0x3fffff) | (low_bits >> 10); ia 2107 arch/sparc/mm/init_64.c __asm__ __volatile__("flush %0" : : "r" (ia + 1)); ia 2109 arch/sparc/mm/init_64.c ia[2] = (ia[2] & ~0x1fff) | (high_bits & 0x3ff); ia 2110 arch/sparc/mm/init_64.c __asm__ __volatile__("flush %0" : : "r" (ia + 2)); ia 2112 arch/sparc/mm/init_64.c ia[3] = (ia[3] & ~0x1fff) | (low_bits & 0x3ff); ia 2113 arch/sparc/mm/init_64.c __asm__ __volatile__("flush %0" : : "r" (ia + 3)); ia 528 drivers/acpi/hmat/hmat.c struct memory_initiator *ia; ia 532 drivers/acpi/hmat/hmat.c ia = list_entry(a, struct memory_initiator, node); ia 535 drivers/acpi/hmat/hmat.c set_bit(ia->processor_pxm, p_nodes); ia 538 drivers/acpi/hmat/hmat.c return ia->processor_pxm - ib->processor_pxm; ia 822 drivers/atm/iphase.c static u32 ia_phy_read32(struct iadev_priv *ia, unsigned int reg) ia 824 drivers/atm/iphase.c return readl(ia->phy + (reg >> 2)); ia 827 drivers/atm/iphase.c static void ia_phy_write32(struct iadev_priv *ia, unsigned int reg, u32 val) ia 829 drivers/atm/iphase.c writel(val, ia->phy + (reg >> 2)); ia 31 drivers/regulator/qcom_rpm-regulator.c struct request_member ia; /* average current in mA */ ia 73 drivers/regulator/qcom_rpm-regulator.c .ia = { 1, 0x00001FFE, 1 }, ia 84 drivers/regulator/qcom_rpm-regulator.c .ia = { 1, 0x00001FFE, 1 }, ia 113 drivers/regulator/qcom_rpm-regulator.c .ia = { 1, 0x000FFC00, 10 }, ia 124 drivers/regulator/qcom_rpm-regulator.c .ia = { 1, 0x000FFC00, 10 }, ia 392 drivers/regulator/qcom_rpm-regulator.c const struct request_member *req = &parts->ia; ia 1634 drivers/scsi/sym53c8xx_2/sym_fw2.h SCR_LOAD_REL (ia, 4), ia 1638 drivers/scsi/sym53c8xx_2/sym_fw2.h SCR_LOAD_REL (ia, 4), ia 1648 drivers/scsi/sym53c8xx_2/sym_fw2.h SCR_REG_REG (ia, SCR_ADD, 8), ia 1666 drivers/scsi/sym53c8xx_2/sym_fw2.h SCR_STORE_REL (ia, 4), ia 1688 drivers/scsi/sym53c8xx_2/sym_fw2.h SCR_LOAD_ABS (ia, 4), ia 1691 drivers/scsi/sym53c8xx_2/sym_fw2.h SCR_STORE_REL (ia, 4), ia 1696 drivers/scsi/sym53c8xx_2/sym_fw2.h SCR_STORE_REL (ia, 4), ia 1718 drivers/scsi/sym53c8xx_2/sym_fw2.h SCR_LOAD_ABS (ia, 4), ia 602 fs/afs/cmservice.c } ia; ia 613 fs/afs/cmservice.c reply.ia.uuid[0] = call->net->uuid.time_low; ia 614 fs/afs/cmservice.c reply.ia.uuid[1] = htonl(ntohs(call->net->uuid.time_mid)); ia 615 fs/afs/cmservice.c reply.ia.uuid[2] = htonl(ntohs(call->net->uuid.time_hi_and_version)); ia 616 fs/afs/cmservice.c reply.ia.uuid[3] = htonl((s8) call->net->uuid.clock_seq_hi_and_reserved); ia 617 fs/afs/cmservice.c reply.ia.uuid[4] = htonl((s8) call->net->uuid.clock_seq_low); ia 619 fs/afs/cmservice.c reply.ia.uuid[loop + 5] = htonl((s8) call->net->uuid.node[loop]); ia 44 fs/debugfs/inode.c static int debugfs_setattr(struct dentry *dentry, struct iattr *ia) ia 48 fs/debugfs/inode.c if (ret && (ia->ia_valid & (ATTR_MODE | ATTR_UID | ATTR_GID))) ia 50 fs/debugfs/inode.c return simple_setattr(dentry, ia); ia 725 fs/ecryptfs/inode.c static int truncate_upper(struct dentry *dentry, struct iattr *ia, ia 735 fs/ecryptfs/inode.c if (unlikely((ia->ia_size == i_size))) { ia 744 fs/ecryptfs/inode.c if (ia->ia_size > i_size) { ia 753 fs/ecryptfs/inode.c (ia->ia_size - 1), 1); ia 760 fs/ecryptfs/inode.c - (ia->ia_size & ~PAGE_MASK)); ia 763 fs/ecryptfs/inode.c truncate_setsize(inode, ia->ia_size); ia 764 fs/ecryptfs/inode.c lower_ia->ia_size = ia->ia_size; ia 777 fs/ecryptfs/inode.c ia->ia_size, num_zeros); ia 786 fs/ecryptfs/inode.c truncate_setsize(inode, ia->ia_size); ia 799 fs/ecryptfs/inode.c upper_size_to_lower_size(crypt_stat, ia->ia_size); ia 845 fs/ecryptfs/inode.c struct iattr ia = { .ia_valid = ATTR_SIZE, .ia_size = new_length }; ia 853 fs/ecryptfs/inode.c rc = truncate_upper(dentry, &ia, &lower_ia); ia 882 fs/ecryptfs/inode.c static int ecryptfs_setattr(struct dentry *dentry, struct iattr *ia) ia 936 fs/ecryptfs/inode.c rc = setattr_prepare(dentry, ia); ia 939 fs/ecryptfs/inode.c if (ia->ia_valid & ATTR_SIZE) { ia 940 fs/ecryptfs/inode.c rc = ecryptfs_inode_newsize_ok(inode, ia->ia_size); ia 945 fs/ecryptfs/inode.c memcpy(&lower_ia, ia, sizeof(lower_ia)); ia 946 fs/ecryptfs/inode.c if (ia->ia_valid & ATTR_FILE) ia 947 fs/ecryptfs/inode.c lower_ia.ia_file = ecryptfs_file_to_lower(ia->ia_file); ia 948 fs/ecryptfs/inode.c if (ia->ia_valid & ATTR_SIZE) { ia 949 fs/ecryptfs/inode.c rc = truncate_upper(dentry, ia, &lower_ia); ia 41 fs/fat/file.c struct iattr ia; ia 66 fs/fat/file.c ia.ia_valid = ATTR_MODE | ATTR_CTIME; ia 67 fs/fat/file.c ia.ia_ctime = current_time(inode); ia 69 fs/fat/file.c ia.ia_mode = fat_make_mode(sbi, attr, S_IRWXUGO); ia 71 fs/fat/file.c ia.ia_mode = fat_make_mode(sbi, attr, ia 93 fs/fat/file.c err = security_inode_setattr(file->f_path.dentry, &ia); ia 98 fs/fat/file.c err = fat_setattr(file->f_path.dentry, &ia); ia 102 fs/fat/file.c fsnotify_change(file->f_path.dentry, ia.ia_valid); ia 319 fs/fuse/cuse.c struct cuse_init_args *ia = container_of(args, typeof(*ia), ap.args); ia 320 fs/fuse/cuse.c struct fuse_args_pages *ap = &ia->ap; ia 322 fs/fuse/cuse.c struct cuse_init_out *arg = &ia->out; ia 407 fs/fuse/cuse.c kfree(ia); ia 428 fs/fuse/cuse.c struct cuse_init_args *ia; ia 438 fs/fuse/cuse.c ia = kzalloc(sizeof(*ia), GFP_KERNEL); ia 439 fs/fuse/cuse.c if (!ia) ia 442 fs/fuse/cuse.c ap = &ia->ap; ia 443 fs/fuse/cuse.c ia->in.major = FUSE_KERNEL_VERSION; ia 444 fs/fuse/cuse.c ia->in.minor = FUSE_KERNEL_MINOR_VERSION; ia 445 fs/fuse/cuse.c ia->in.flags |= CUSE_UNRESTRICTED_IOCTL; ia 448 fs/fuse/cuse.c ap->args.in_args[0].size = sizeof(ia->in); ia 449 fs/fuse/cuse.c ap->args.in_args[0].value = &ia->in; ia 451 fs/fuse/cuse.c ap->args.out_args[0].size = sizeof(ia->out); ia 452 fs/fuse/cuse.c ap->args.out_args[0].value = &ia->out; ia 457 fs/fuse/cuse.c ap->pages = &ia->page; ia 458 fs/fuse/cuse.c ap->descs = &ia->desc; ia 459 fs/fuse/cuse.c ia->page = page; ia 460 fs/fuse/cuse.c ia->desc.length = ap->args.out_args[1].size; ia 465 fs/fuse/cuse.c kfree(ia); ia 359 fs/fuse/file.c struct fuse_io_args ia; ia 375 fs/fuse/file.c curr_index = wpa->ia.write.in.offset >> PAGE_SHIFT; ia 376 fs/fuse/file.c if (idx_from < curr_index + wpa->ia.ap.num_pages && ia 551 fs/fuse/file.c void fuse_read_args_fill(struct fuse_io_args *ia, struct file *file, loff_t pos, ia 555 fs/fuse/file.c struct fuse_args *args = &ia->ap.args; ia 557 fs/fuse/file.c ia->read.in.fh = ff->fh; ia 558 fs/fuse/file.c ia->read.in.offset = pos; ia 559 fs/fuse/file.c ia->read.in.size = count; ia 560 fs/fuse/file.c ia->read.in.flags = file->f_flags; ia 564 fs/fuse/file.c args->in_args[0].size = sizeof(ia->read.in); ia 565 fs/fuse/file.c args->in_args[0].value = &ia->read.in; ia 652 fs/fuse/file.c struct fuse_io_args *ia; ia 654 fs/fuse/file.c ia = kzalloc(sizeof(*ia), GFP_KERNEL); ia 655 fs/fuse/file.c if (ia) { ia 656 fs/fuse/file.c ia->io = io; ia 657 fs/fuse/file.c ia->ap.pages = fuse_pages_alloc(npages, GFP_KERNEL, ia 658 fs/fuse/file.c &ia->ap.descs); ia 659 fs/fuse/file.c if (!ia->ap.pages) { ia 660 fs/fuse/file.c kfree(ia); ia 661 fs/fuse/file.c ia = NULL; ia 664 fs/fuse/file.c return ia; ia 667 fs/fuse/file.c static void fuse_io_free(struct fuse_io_args *ia) ia 669 fs/fuse/file.c kfree(ia->ap.pages); ia 670 fs/fuse/file.c kfree(ia); ia 676 fs/fuse/file.c struct fuse_io_args *ia = container_of(args, typeof(*ia), ap.args); ia 677 fs/fuse/file.c struct fuse_io_priv *io = ia->io; ia 680 fs/fuse/file.c fuse_release_user_pages(&ia->ap, io->should_dirty); ia 685 fs/fuse/file.c if (ia->write.out.size > ia->write.in.size) { ia 687 fs/fuse/file.c } else if (ia->write.in.size != ia->write.out.size) { ia 688 fs/fuse/file.c pos = ia->write.in.offset - io->offset + ia 689 fs/fuse/file.c ia->write.out.size; ia 694 fs/fuse/file.c if (ia->read.in.size != outsize) ia 695 fs/fuse/file.c pos = ia->read.in.offset - io->offset + outsize; ia 699 fs/fuse/file.c fuse_io_free(ia); ia 703 fs/fuse/file.c struct fuse_io_args *ia, size_t num_bytes) ia 706 fs/fuse/file.c struct fuse_io_priv *io = ia->io; ia 714 fs/fuse/file.c ia->ap.args.end = fuse_aio_complete_req; ia 715 fs/fuse/file.c err = fuse_simple_background(fc, &ia->ap.args, GFP_KERNEL); ia 717 fs/fuse/file.c fuse_aio_complete_req(fc, &ia->ap.args, err); ia 722 fs/fuse/file.c static ssize_t fuse_send_read(struct fuse_io_args *ia, loff_t pos, size_t count, ia 725 fs/fuse/file.c struct file *file = ia->io->iocb->ki_filp; ia 729 fs/fuse/file.c fuse_read_args_fill(ia, file, pos, count, FUSE_READ); ia 731 fs/fuse/file.c ia->read.in.read_flags |= FUSE_READ_LOCKOWNER; ia 732 fs/fuse/file.c ia->read.in.lock_owner = fuse_lock_owner_id(fc, owner); ia 735 fs/fuse/file.c if (ia->io->async) ia 736 fs/fuse/file.c return fuse_async_req_send(fc, ia, count); ia 738 fs/fuse/file.c return fuse_simple_request(fc, &ia->ap.args); ia 787 fs/fuse/file.c struct fuse_io_args ia = { ia 810 fs/fuse/file.c fuse_read_args_fill(&ia, file, pos, desc.length, FUSE_READ); ia 811 fs/fuse/file.c res = fuse_simple_request(fc, &ia.ap.args); ia 818 fs/fuse/file.c fuse_short_read(inode, attr_ver, res, &ia.ap); ia 845 fs/fuse/file.c struct fuse_io_args *ia = container_of(args, typeof(*ia), ap.args); ia 846 fs/fuse/file.c struct fuse_args_pages *ap = &ia->ap; ia 847 fs/fuse/file.c size_t count = ia->read.in.size; ia 861 fs/fuse/file.c fuse_short_read(inode, ia->read.attr_ver, num_read, ap); ia 876 fs/fuse/file.c if (ia->ff) ia 877 fs/fuse/file.c fuse_file_put(ia->ff, false, false); ia 879 fs/fuse/file.c fuse_io_free(ia); ia 882 fs/fuse/file.c static void fuse_send_readpages(struct fuse_io_args *ia, struct file *file) ia 886 fs/fuse/file.c struct fuse_args_pages *ap = &ia->ap; ia 903 fs/fuse/file.c fuse_read_args_fill(ia, file, pos, count, FUSE_READ); ia 904 fs/fuse/file.c ia->read.attr_ver = fuse_get_attr_version(fc); ia 906 fs/fuse/file.c ia->ff = fuse_file_get(ff); ia 919 fs/fuse/file.c struct fuse_io_args *ia; ia 929 fs/fuse/file.c struct fuse_io_args *ia = data->ia; ia 930 fs/fuse/file.c struct fuse_args_pages *ap = &ia->ap; ia 942 fs/fuse/file.c fuse_send_readpages(ia, data->file); ia 943 fs/fuse/file.c data->ia = ia = fuse_io_alloc(NULL, data->max_pages); ia 944 fs/fuse/file.c if (!ia) { ia 948 fs/fuse/file.c ap = &ia->ap; ia 953 fs/fuse/file.c fuse_io_free(ia); ia 982 fs/fuse/file.c data.ia = fuse_io_alloc(NULL, data.max_pages); ia 984 fs/fuse/file.c if (!data.ia) ia 989 fs/fuse/file.c if (data.ia->ap.num_pages) ia 990 fs/fuse/file.c fuse_send_readpages(data.ia, file); ia 992 fs/fuse/file.c fuse_io_free(data.ia); ia 1019 fs/fuse/file.c static void fuse_write_args_fill(struct fuse_io_args *ia, struct fuse_file *ff, ia 1022 fs/fuse/file.c struct fuse_args *args = &ia->ap.args; ia 1024 fs/fuse/file.c ia->write.in.fh = ff->fh; ia 1025 fs/fuse/file.c ia->write.in.offset = pos; ia 1026 fs/fuse/file.c ia->write.in.size = count; ia 1033 fs/fuse/file.c args->in_args[0].size = sizeof(ia->write.in); ia 1034 fs/fuse/file.c args->in_args[0].value = &ia->write.in; ia 1037 fs/fuse/file.c args->out_args[0].size = sizeof(ia->write.out); ia 1038 fs/fuse/file.c args->out_args[0].value = &ia->write.out; ia 1053 fs/fuse/file.c static ssize_t fuse_send_write(struct fuse_io_args *ia, loff_t pos, ia 1056 fs/fuse/file.c struct kiocb *iocb = ia->io->iocb; ia 1060 fs/fuse/file.c struct fuse_write_in *inarg = &ia->write.in; ia 1063 fs/fuse/file.c fuse_write_args_fill(ia, ff, pos, count); ia 1070 fs/fuse/file.c if (ia->io->async) ia 1071 fs/fuse/file.c return fuse_async_req_send(fc, ia, count); ia 1073 fs/fuse/file.c err = fuse_simple_request(fc, &ia->ap.args); ia 1074 fs/fuse/file.c if (!err && ia->write.out.size > count) ia 1077 fs/fuse/file.c return err ?: ia->write.out.size; ia 1097 fs/fuse/file.c static ssize_t fuse_send_write_pages(struct fuse_io_args *ia, ia 1101 fs/fuse/file.c struct fuse_args_pages *ap = &ia->ap; ia 1111 fs/fuse/file.c fuse_write_args_fill(ia, ff, pos, count); ia 1112 fs/fuse/file.c ia->write.in.flags = fuse_write_flags(iocb); ia 1115 fs/fuse/file.c if (!err && ia->write.out.size > count) ia 1119 fs/fuse/file.c count = ia->write.out.size; ia 1228 fs/fuse/file.c struct fuse_io_args ia = {}; ia 1229 fs/fuse/file.c struct fuse_args_pages *ap = &ia.ap; ia 1243 fs/fuse/file.c err = fuse_send_write_pages(&ia, iocb, inode, ia 1246 fs/fuse/file.c size_t num_written = ia.write.out.size; ia 1437 fs/fuse/file.c struct fuse_io_args *ia; ia 1441 fs/fuse/file.c ia = fuse_io_alloc(io, max_pages); ia 1442 fs/fuse/file.c if (!ia) ia 1445 fs/fuse/file.c ia->io = io; ia 1460 fs/fuse/file.c err = fuse_get_user_pages(&ia->ap, iter, &nbytes, write, ia 1467 fs/fuse/file.c ia->write.in.write_flags |= FUSE_WRITE_KILL_PRIV; ia 1469 fs/fuse/file.c nres = fuse_send_write(ia, pos, nbytes, owner); ia 1471 fs/fuse/file.c nres = fuse_send_read(ia, pos, nbytes, owner); ia 1475 fs/fuse/file.c fuse_release_user_pages(&ia->ap, io->should_dirty); ia 1476 fs/fuse/file.c fuse_io_free(ia); ia 1478 fs/fuse/file.c ia = NULL; ia 1495 fs/fuse/file.c ia = fuse_io_alloc(io, max_pages); ia 1496 fs/fuse/file.c if (!ia) ia 1500 fs/fuse/file.c if (ia) ia 1501 fs/fuse/file.c fuse_io_free(ia); ia 1595 fs/fuse/file.c struct fuse_args_pages *ap = &wpa->ia.ap; ia 1601 fs/fuse/file.c if (wpa->ia.ff) ia 1602 fs/fuse/file.c fuse_file_put(wpa->ia.ff, false, false); ia 1611 fs/fuse/file.c struct fuse_args_pages *ap = &wpa->ia.ap; ia 1634 fs/fuse/file.c struct fuse_write_in *inarg = &wpa->ia.write.in; ia 1635 fs/fuse/file.c struct fuse_args *args = &wpa->ia.ap.args; ia 1636 fs/fuse/file.c __u64 data_size = wpa->ia.ap.num_pages * PAGE_SIZE; ia 1709 fs/fuse/file.c container_of(args, typeof(*wpa), ia.ap.args); ia 1717 fs/fuse/file.c struct fuse_write_in *inarg = &wpa->ia.write.in; ia 1722 fs/fuse/file.c next->ia.ff = fuse_file_get(wpa->ia.ff); ia 1802 fs/fuse/file.c ap = &wpa->ia.ap; ia 1830 fs/fuse/file.c ap = &wpa->ia.ap; ia 1837 fs/fuse/file.c wpa->ia.ff = fuse_write_file_get(fc, fi); ia 1838 fs/fuse/file.c if (!wpa->ia.ff) ia 1841 fs/fuse/file.c fuse_write_args_fill(&wpa->ia, wpa->ia.ff, page_offset(page), 0); ia 1844 fs/fuse/file.c wpa->ia.write.in.write_flags |= FUSE_WRITE_CACHE; ia 1910 fs/fuse/file.c struct fuse_args_pages *ap = &data->wpa->ia.ap; ia 1939 fs/fuse/file.c int num_pages = wpa->ia.ap.num_pages; ia 1942 fs/fuse/file.c wpa->ia.ff = fuse_file_get(data->ff); ia 1965 fs/fuse/file.c struct fuse_args_pages *new_ap = &new_wpa->ia.ap; ia 1983 fs/fuse/file.c curr_index = tmp->ia.write.in.offset >> PAGE_SHIFT; ia 1985 fs/fuse/file.c WARN_ON(tmp->ia.ap.num_pages != 1); ia 1986 fs/fuse/file.c swap(tmp->ia.ap.pages[0], new_ap->pages[0]); ia 2015 fs/fuse/file.c struct fuse_args_pages *ap = &wpa->ia.ap; ia 2078 fs/fuse/file.c ap = &wpa->ia.ap; ia 2079 fs/fuse/file.c fuse_write_args_fill(&wpa->ia, data->ff, page_offset(page), 0); ia 2080 fs/fuse/file.c wpa->ia.write.in.write_flags |= FUSE_WRITE_CACHE; ia 2151 fs/fuse/file.c WARN_ON(!data.wpa->ia.ap.num_pages); ia 841 fs/fuse/fuse_i.h void fuse_read_args_fill(struct fuse_io_args *ia, struct file *file, loff_t pos, ia 887 fs/fuse/inode.c struct fuse_init_args *ia = container_of(args, typeof(*ia), args); ia 888 fs/fuse/inode.c struct fuse_init_out *arg = &ia->out; ia 967 fs/fuse/inode.c kfree(ia); ia 975 fs/fuse/inode.c struct fuse_init_args *ia; ia 977 fs/fuse/inode.c ia = kzalloc(sizeof(*ia), GFP_KERNEL | __GFP_NOFAIL); ia 979 fs/fuse/inode.c ia->in.major = FUSE_KERNEL_VERSION; ia 980 fs/fuse/inode.c ia->in.minor = FUSE_KERNEL_MINOR_VERSION; ia 981 fs/fuse/inode.c ia->in.max_readahead = fc->sb->s_bdi->ra_pages * PAGE_SIZE; ia 982 fs/fuse/inode.c ia->in.flags |= ia 992 fs/fuse/inode.c ia->args.opcode = FUSE_INIT; ia 993 fs/fuse/inode.c ia->args.in_numargs = 1; ia 994 fs/fuse/inode.c ia->args.in_args[0].size = sizeof(ia->in); ia 995 fs/fuse/inode.c ia->args.in_args[0].value = &ia->in; ia 996 fs/fuse/inode.c ia->args.out_numargs = 1; ia 1000 fs/fuse/inode.c ia->args.out_argvar = 1; ia 1001 fs/fuse/inode.c ia->args.out_args[0].size = sizeof(ia->out); ia 1002 fs/fuse/inode.c ia->args.out_args[0].value = &ia->out; ia 1003 fs/fuse/inode.c ia->args.force = true; ia 1004 fs/fuse/inode.c ia->args.nocreds = true; ia 1005 fs/fuse/inode.c ia->args.end = process_init_reply; ia 1007 fs/fuse/inode.c if (fuse_simple_background(fc, &ia->args, GFP_KERNEL) != 0) ia 1008 fs/fuse/inode.c process_init_reply(fc, &ia->args, -ENOTCONN); ia 324 fs/fuse/readdir.c struct fuse_io_args ia = {}; ia 325 fs/fuse/readdir.c struct fuse_args_pages *ap = &ia.ap; ia 341 fs/fuse/readdir.c fuse_read_args_fill(&ia, file, ctx->pos, PAGE_SIZE, ia 344 fs/fuse/readdir.c fuse_read_args_fill(&ia, file, ctx->pos, PAGE_SIZE, ia 74 fs/ntfs/dir.c INDEX_ALLOCATION *ia; ia 319 fs/ntfs/dir.c ia = (INDEX_ALLOCATION*)(kaddr + ((vcn << ia 322 fs/ntfs/dir.c if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_SIZE) { ia 328 fs/ntfs/dir.c if (unlikely(!ntfs_is_indx_record(ia->magic))) { ia 334 fs/ntfs/dir.c if (sle64_to_cpu(ia->index_block_vcn) != vcn) { ia 339 fs/ntfs/dir.c sle64_to_cpu(ia->index_block_vcn), ia 343 fs/ntfs/dir.c if (le32_to_cpu(ia->index.allocated_size) + 0x18 != ia 350 fs/ntfs/dir.c le32_to_cpu(ia->index.allocated_size) + 0x18, ia 354 fs/ntfs/dir.c index_end = (u8*)ia + dir_ni->itype.index.block_size; ia 363 fs/ntfs/dir.c index_end = (u8*)&ia->index + le32_to_cpu(ia->index.index_length); ia 364 fs/ntfs/dir.c if (index_end > (u8*)ia + dir_ni->itype.index.block_size) { ia 371 fs/ntfs/dir.c ie = (INDEX_ENTRY*)((u8*)&ia->index + ia 372 fs/ntfs/dir.c le32_to_cpu(ia->index.entries_offset)); ia 380 fs/ntfs/dir.c if ((u8*)ie < (u8*)ia || (u8*)ie + ia 534 fs/ntfs/dir.c if ((ia->index.flags & NODE_MASK) == LEAF_NODE) { ia 627 fs/ntfs/dir.c INDEX_ALLOCATION *ia; ia 797 fs/ntfs/dir.c ia = (INDEX_ALLOCATION*)(kaddr + ((vcn << ia 800 fs/ntfs/dir.c if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_SIZE) { ia 806 fs/ntfs/dir.c if (unlikely(!ntfs_is_indx_record(ia->magic))) { ia 812 fs/ntfs/dir.c if (sle64_to_cpu(ia->index_block_vcn) != vcn) { ia 817 fs/ntfs/dir.c sle64_to_cpu(ia->index_block_vcn), ia 821 fs/ntfs/dir.c if (le32_to_cpu(ia->index.allocated_size) + 0x18 != ia 828 fs/ntfs/dir.c le32_to_cpu(ia->index.allocated_size) + 0x18, ia 832 fs/ntfs/dir.c index_end = (u8*)ia + dir_ni->itype.index.block_size; ia 841 fs/ntfs/dir.c index_end = (u8*)&ia->index + le32_to_cpu(ia->index.index_length); ia 842 fs/ntfs/dir.c if (index_end > (u8*)ia + dir_ni->itype.index.block_size) { ia 849 fs/ntfs/dir.c ie = (INDEX_ENTRY*)((u8*)&ia->index + ia 850 fs/ntfs/dir.c le32_to_cpu(ia->index.entries_offset)); ia 858 fs/ntfs/dir.c if ((u8*)ie < (u8*)ia || (u8*)ie + ia 944 fs/ntfs/dir.c if ((ia->index.flags & NODE_MASK) == LEAF_NODE) { ia 1096 fs/ntfs/dir.c INDEX_ALLOCATION *ia; ia 1296 fs/ntfs/dir.c ia = (INDEX_ALLOCATION*)(kaddr + (ia_pos & ~PAGE_MASK & ia 1299 fs/ntfs/dir.c if (unlikely((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_SIZE)) { ia 1305 fs/ntfs/dir.c if (unlikely(!ntfs_is_indx_record(ia->magic))) { ia 1312 fs/ntfs/dir.c if (unlikely(sle64_to_cpu(ia->index_block_vcn) != (ia_pos & ia 1319 fs/ntfs/dir.c sle64_to_cpu(ia->index_block_vcn), ia 1324 fs/ntfs/dir.c if (unlikely(le32_to_cpu(ia->index.allocated_size) + 0x18 != ia 1332 fs/ntfs/dir.c le32_to_cpu(ia->index.allocated_size) + 0x18, ia 1336 fs/ntfs/dir.c index_end = (u8*)ia + ndir->itype.index.block_size; ia 1346 fs/ntfs/dir.c index_end = (u8*)&ia->index + le32_to_cpu(ia->index.index_length); ia 1347 fs/ntfs/dir.c if (unlikely(index_end > (u8*)ia + ndir->itype.index.block_size)) { ia 1355 fs/ntfs/dir.c ie = (INDEX_ENTRY*)((u8*)&ia->index + ia 1356 fs/ntfs/dir.c le32_to_cpu(ia->index.entries_offset)); ia 1365 fs/ntfs/dir.c (unsigned long long)((u8*)ie - (u8*)ia)); ia 1367 fs/ntfs/dir.c if (unlikely((u8*)ie < (u8*)ia || (u8*)ie + ia 1376 fs/ntfs/dir.c if (ia_pos - ia_start > (u8*)ie - (u8*)ia) ia 1379 fs/ntfs/dir.c actor->pos = (u8*)ie - (u8*)ia + ia 1380 fs/ntfs/dir.c (sle64_to_cpu(ia->index_block_vcn) << ia 116 fs/ntfs/index.c INDEX_ALLOCATION *ia; ia 199 fs/ntfs/index.c ictx->ia = NULL; ia 276 fs/ntfs/index.c ia = (INDEX_ALLOCATION*)(kaddr + ((vcn << ia 279 fs/ntfs/index.c if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_SIZE) { ia 285 fs/ntfs/index.c if (unlikely(!ntfs_is_indx_record(ia->magic))) { ia 291 fs/ntfs/index.c if (sle64_to_cpu(ia->index_block_vcn) != vcn) { ia 296 fs/ntfs/index.c sle64_to_cpu(ia->index_block_vcn), ia 300 fs/ntfs/index.c if (le32_to_cpu(ia->index.allocated_size) + 0x18 != ia 307 fs/ntfs/index.c le32_to_cpu(ia->index.allocated_size) + 0x18, ia 311 fs/ntfs/index.c index_end = (u8*)ia + idx_ni->itype.index.block_size; ia 320 fs/ntfs/index.c index_end = (u8*)&ia->index + le32_to_cpu(ia->index.index_length); ia 321 fs/ntfs/index.c if (index_end > (u8*)ia + idx_ni->itype.index.block_size) { ia 328 fs/ntfs/index.c ie = (INDEX_ENTRY*)((u8*)&ia->index + ia 329 fs/ntfs/index.c le32_to_cpu(ia->index.entries_offset)); ia 337 fs/ntfs/index.c if ((u8*)ie < (u8*)ia || (u8*)ie + ia 368 fs/ntfs/index.c ictx->ia = ia; ia 402 fs/ntfs/index.c if ((ia->index.flags & NODE_MASK) == LEAF_NODE) { ia 70 fs/ntfs/index.h INDEX_ALLOCATION *ia; ia 129 fs/ntfs/index.h (u8*)ictx->ia - (u8*)page_address(ictx->page)); ia 161 fs/ramfs/file-nommu.c static int ramfs_nommu_setattr(struct dentry *dentry, struct iattr *ia) ia 164 fs/ramfs/file-nommu.c unsigned int old_ia_valid = ia->ia_valid; ia 168 fs/ramfs/file-nommu.c ret = setattr_prepare(dentry, ia); ia 173 fs/ramfs/file-nommu.c if (ia->ia_valid & ATTR_SIZE) { ia 176 fs/ramfs/file-nommu.c if (ia->ia_size != size) { ia 177 fs/ramfs/file-nommu.c ret = ramfs_nommu_resize(inode, ia->ia_size, size); ia 178 fs/ramfs/file-nommu.c if (ret < 0 || ia->ia_valid == ATTR_SIZE) ia 184 fs/ramfs/file-nommu.c ia->ia_valid |= ATTR_MTIME|ATTR_CTIME; ia 188 fs/ramfs/file-nommu.c setattr_copy(inode, ia); ia 190 fs/ramfs/file-nommu.c ia->ia_valid = old_ia_valid; ia 357 fs/xfs/xfs_aops.c struct xfs_ioend *ia; ia 360 fs/xfs/xfs_aops.c ia = container_of(a, struct xfs_ioend, io_list); ia 362 fs/xfs/xfs_aops.c if (ia->io_offset < ib->io_offset) ia 364 fs/xfs/xfs_aops.c else if (ia->io_offset > ib->io_offset) ia 23 include/linux/quotaops.h static inline bool is_quota_modification(struct inode *inode, struct iattr *ia) ia 25 include/linux/quotaops.h return (ia->ia_valid & ATTR_SIZE) || ia 26 include/linux/quotaops.h (ia->ia_valid & ATTR_UID && !uid_eq(ia->ia_uid, inode->i_uid)) || ia 27 include/linux/quotaops.h (ia->ia_valid & ATTR_GID && !gid_eq(ia->ia_gid, inode->i_gid)); ia 412 net/bluetooth/smp.c const bdaddr_t *ia, u8 _rat, const bdaddr_t *ra, u8 res[16]) ia 418 net/bluetooth/smp.c SMP_DBG("iat %u ia %6phN rat %u ra %6phN", _iat, ia, _rat, ra); ia 443 net/bluetooth/smp.c memcpy(p2 + 6, ia, 6); ia 3588 net/bluetooth/smp.c const bdaddr_t ia = { { 0xa6, 0xa5, 0xa4, 0xa3, 0xa2, 0xa1 } }; ia 3595 net/bluetooth/smp.c err = smp_c1(k, r, preq, pres, _iat, &ia, _rat, &ra, res); ia 162 net/sunrpc/xprtrdma/frwr_ops.c int frwr_init_mr(struct rpcrdma_ia *ia, struct rpcrdma_mr *mr) ia 164 net/sunrpc/xprtrdma/frwr_ops.c unsigned int depth = ia->ri_max_frwr_depth; ia 172 net/sunrpc/xprtrdma/frwr_ops.c frmr = ib_alloc_mr(ia->ri_pd, ia->ri_mrtype, depth); ia 217 net/sunrpc/xprtrdma/frwr_ops.c int frwr_open(struct rpcrdma_ia *ia, struct rpcrdma_ep *ep) ia 219 net/sunrpc/xprtrdma/frwr_ops.c struct ib_device_attr *attrs = &ia->ri_id->device->attrs; ia 222 net/sunrpc/xprtrdma/frwr_ops.c ia->ri_mrtype = IB_MR_TYPE_MEM_REG; ia 224 net/sunrpc/xprtrdma/frwr_ops.c ia->ri_mrtype = IB_MR_TYPE_SG_GAPS; ia 231 net/sunrpc/xprtrdma/frwr_ops.c ia->ri_max_frwr_depth = attrs->max_sge_rd; ia 233 net/sunrpc/xprtrdma/frwr_ops.c ia->ri_max_frwr_depth = attrs->max_fast_reg_page_list_len; ia 234 net/sunrpc/xprtrdma/frwr_ops.c if (ia->ri_max_frwr_depth > RPCRDMA_MAX_DATA_SEGS) ia 235 net/sunrpc/xprtrdma/frwr_ops.c ia->ri_max_frwr_depth = RPCRDMA_MAX_DATA_SEGS; ia 237 net/sunrpc/xprtrdma/frwr_ops.c __func__, ia->ri_max_frwr_depth); ia 253 net/sunrpc/xprtrdma/frwr_ops.c if (ia->ri_max_frwr_depth < RPCRDMA_MAX_DATA_SEGS) { ia 254 net/sunrpc/xprtrdma/frwr_ops.c delta = RPCRDMA_MAX_DATA_SEGS - ia->ri_max_frwr_depth; ia 257 net/sunrpc/xprtrdma/frwr_ops.c delta -= ia->ri_max_frwr_depth; ia 261 net/sunrpc/xprtrdma/frwr_ops.c max_qp_wr = ia->ri_id->device->attrs.max_qp_wr; ia 281 net/sunrpc/xprtrdma/frwr_ops.c ia->ri_max_segs = ia 282 net/sunrpc/xprtrdma/frwr_ops.c DIV_ROUND_UP(RPCRDMA_MAX_DATA_SEGS, ia->ri_max_frwr_depth); ia 284 net/sunrpc/xprtrdma/frwr_ops.c ia->ri_max_segs += 2; ia 285 net/sunrpc/xprtrdma/frwr_ops.c if (ia->ri_max_segs > RPCRDMA_MAX_HDR_SEGS) ia 286 net/sunrpc/xprtrdma/frwr_ops.c ia->ri_max_segs = RPCRDMA_MAX_HDR_SEGS; ia 301 net/sunrpc/xprtrdma/frwr_ops.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 304 net/sunrpc/xprtrdma/frwr_ops.c (ia->ri_max_segs - 2) * ia->ri_max_frwr_depth); ia 327 net/sunrpc/xprtrdma/frwr_ops.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 333 net/sunrpc/xprtrdma/frwr_ops.c if (nsegs > ia->ri_max_frwr_depth) ia 334 net/sunrpc/xprtrdma/frwr_ops.c nsegs = ia->ri_max_frwr_depth; ia 347 net/sunrpc/xprtrdma/frwr_ops.c if (ia->ri_mrtype == IB_MR_TYPE_SG_GAPS) ia 356 net/sunrpc/xprtrdma/frwr_ops.c dma_nents = ib_dma_map_sg(ia->ri_id->device, mr->mr_sg, mr->mr_nents, ia 423 net/sunrpc/xprtrdma/frwr_ops.c int frwr_send(struct rpcrdma_ia *ia, struct rpcrdma_req *req) ia 447 net/sunrpc/xprtrdma/frwr_ops.c return ib_post_send(ia->ri_id->qp, post_wr, NULL); ia 410 net/sunrpc/xprtrdma/transport.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 419 net/sunrpc/xprtrdma/transport.c if (test_and_clear_bit(RPCRDMA_IAF_REMOVING, &ia->ri_flags)) { ia 420 net/sunrpc/xprtrdma/transport.c rpcrdma_ia_remove(ia); ia 426 net/sunrpc/xprtrdma/transport.c rpcrdma_ep_disconnect(ep, ia); ia 94 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 99 net/sunrpc/xprtrdma/verbs.c ib_drain_rq(ia->ri_id->qp); ia 104 net/sunrpc/xprtrdma/verbs.c ib_drain_sq(ia->ri_id->qp); ia 221 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 231 net/sunrpc/xprtrdma/verbs.c ia->ri_async_rc = 0; ia 232 net/sunrpc/xprtrdma/verbs.c complete(&ia->ri_done); ia 235 net/sunrpc/xprtrdma/verbs.c ia->ri_async_rc = -EPROTO; ia 236 net/sunrpc/xprtrdma/verbs.c complete(&ia->ri_done); ia 239 net/sunrpc/xprtrdma/verbs.c ia->ri_async_rc = -ENETUNREACH; ia 240 net/sunrpc/xprtrdma/verbs.c complete(&ia->ri_done); ia 245 net/sunrpc/xprtrdma/verbs.c ia->ri_id->device->name, ia 248 net/sunrpc/xprtrdma/verbs.c init_completion(&ia->ri_remove_done); ia 249 net/sunrpc/xprtrdma/verbs.c set_bit(RPCRDMA_IAF_REMOVING, &ia->ri_flags); ia 252 net/sunrpc/xprtrdma/verbs.c wait_for_completion(&ia->ri_remove_done); ia 254 net/sunrpc/xprtrdma/verbs.c ia->ri_id = NULL; ia 289 net/sunrpc/xprtrdma/verbs.c ia->ri_id->device->name, rdma_event_msg(event->event)); ia 294 net/sunrpc/xprtrdma/verbs.c rpcrdma_create_id(struct rpcrdma_xprt *xprt, struct rpcrdma_ia *ia) ia 302 net/sunrpc/xprtrdma/verbs.c init_completion(&ia->ri_done); ia 309 net/sunrpc/xprtrdma/verbs.c ia->ri_async_rc = -ETIMEDOUT; ia 315 net/sunrpc/xprtrdma/verbs.c rc = wait_for_completion_interruptible_timeout(&ia->ri_done, wtimeout); ia 321 net/sunrpc/xprtrdma/verbs.c rc = ia->ri_async_rc; ia 325 net/sunrpc/xprtrdma/verbs.c ia->ri_async_rc = -ETIMEDOUT; ia 329 net/sunrpc/xprtrdma/verbs.c rc = wait_for_completion_interruptible_timeout(&ia->ri_done, wtimeout); ia 334 net/sunrpc/xprtrdma/verbs.c rc = ia->ri_async_rc; ia 359 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_ia *ia = &xprt->rx_ia; ia 362 net/sunrpc/xprtrdma/verbs.c ia->ri_id = rpcrdma_create_id(xprt, ia); ia 363 net/sunrpc/xprtrdma/verbs.c if (IS_ERR(ia->ri_id)) { ia 364 net/sunrpc/xprtrdma/verbs.c rc = PTR_ERR(ia->ri_id); ia 368 net/sunrpc/xprtrdma/verbs.c ia->ri_pd = ib_alloc_pd(ia->ri_id->device, 0); ia 369 net/sunrpc/xprtrdma/verbs.c if (IS_ERR(ia->ri_pd)) { ia 370 net/sunrpc/xprtrdma/verbs.c rc = PTR_ERR(ia->ri_pd); ia 377 net/sunrpc/xprtrdma/verbs.c if (frwr_is_supported(ia->ri_id->device)) ia 382 net/sunrpc/xprtrdma/verbs.c ia->ri_id->device->name, xprt_rdma_memreg_strategy); ia 390 net/sunrpc/xprtrdma/verbs.c rpcrdma_ia_close(ia); ia 402 net/sunrpc/xprtrdma/verbs.c rpcrdma_ia_remove(struct rpcrdma_ia *ia) ia 404 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_xprt *r_xprt = container_of(ia, struct rpcrdma_xprt, ia 419 net/sunrpc/xprtrdma/verbs.c if (ia->ri_id->qp) { ia 421 net/sunrpc/xprtrdma/verbs.c rdma_destroy_qp(ia->ri_id); ia 422 net/sunrpc/xprtrdma/verbs.c ia->ri_id->qp = NULL; ia 439 net/sunrpc/xprtrdma/verbs.c ib_dealloc_pd(ia->ri_pd); ia 440 net/sunrpc/xprtrdma/verbs.c ia->ri_pd = NULL; ia 443 net/sunrpc/xprtrdma/verbs.c complete(&ia->ri_remove_done); ia 454 net/sunrpc/xprtrdma/verbs.c rpcrdma_ia_close(struct rpcrdma_ia *ia) ia 456 net/sunrpc/xprtrdma/verbs.c if (ia->ri_id != NULL && !IS_ERR(ia->ri_id)) { ia 457 net/sunrpc/xprtrdma/verbs.c if (ia->ri_id->qp) ia 458 net/sunrpc/xprtrdma/verbs.c rdma_destroy_qp(ia->ri_id); ia 459 net/sunrpc/xprtrdma/verbs.c rdma_destroy_id(ia->ri_id); ia 461 net/sunrpc/xprtrdma/verbs.c ia->ri_id = NULL; ia 464 net/sunrpc/xprtrdma/verbs.c if (ia->ri_pd && !IS_ERR(ia->ri_pd)) ia 465 net/sunrpc/xprtrdma/verbs.c ib_dealloc_pd(ia->ri_pd); ia 466 net/sunrpc/xprtrdma/verbs.c ia->ri_pd = NULL; ia 478 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 488 net/sunrpc/xprtrdma/verbs.c max_sge = min_t(unsigned int, ia->ri_id->device->attrs.max_send_sge, ia 494 net/sunrpc/xprtrdma/verbs.c ia->ri_max_send_sges = max_sge; ia 496 net/sunrpc/xprtrdma/verbs.c rc = frwr_open(ia, ep); ia 523 net/sunrpc/xprtrdma/verbs.c sendcq = ib_alloc_cq_any(ia->ri_id->device, NULL, ia 531 net/sunrpc/xprtrdma/verbs.c recvcq = ib_alloc_cq_any(ia->ri_id->device, NULL, ia 557 net/sunrpc/xprtrdma/verbs.c min_t(int, U8_MAX, ia->ri_id->device->attrs.max_qp_rd_atom); ia 588 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 590 net/sunrpc/xprtrdma/verbs.c if (ia->ri_id && ia->ri_id->qp) { ia 591 net/sunrpc/xprtrdma/verbs.c rpcrdma_ep_disconnect(ep, ia); ia 592 net/sunrpc/xprtrdma/verbs.c rdma_destroy_qp(ia->ri_id); ia 593 net/sunrpc/xprtrdma/verbs.c ia->ri_id->qp = NULL; ia 609 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 628 net/sunrpc/xprtrdma/verbs.c err = rdma_create_qp(ia->ri_id, ia->ri_pd, qp_init_attr); ia 640 net/sunrpc/xprtrdma/verbs.c rpcrdma_ia_close(ia); ia 648 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 654 net/sunrpc/xprtrdma/verbs.c rpcrdma_ep_disconnect(&r_xprt->rx_ep, ia); ia 657 net/sunrpc/xprtrdma/verbs.c id = rpcrdma_create_id(r_xprt, ia); ia 671 net/sunrpc/xprtrdma/verbs.c if (ia->ri_id->device != id->device) { ia 676 net/sunrpc/xprtrdma/verbs.c err = rdma_create_qp(id, ia->ri_pd, qp_init_attr); ia 682 net/sunrpc/xprtrdma/verbs.c old = ia->ri_id; ia 683 net/sunrpc/xprtrdma/verbs.c ia->ri_id = id; ia 696 net/sunrpc/xprtrdma/verbs.c rpcrdma_ep_connect(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia) ia 698 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_xprt *r_xprt = container_of(ia, struct rpcrdma_xprt, ia 709 net/sunrpc/xprtrdma/verbs.c rc = rdma_create_qp(ia->ri_id, ia->ri_pd, &qp_init_attr); ia 731 net/sunrpc/xprtrdma/verbs.c rc = rdma_connect(ia->ri_id, &ep->rep_remote_cma); ia 767 net/sunrpc/xprtrdma/verbs.c rpcrdma_ep_disconnect(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia) ia 774 net/sunrpc/xprtrdma/verbs.c rc = rdma_disconnect(ia->ri_id); ia 812 net/sunrpc/xprtrdma/verbs.c static struct rpcrdma_sendctx *rpcrdma_sendctx_create(struct rpcrdma_ia *ia) ia 816 net/sunrpc/xprtrdma/verbs.c sc = kzalloc(struct_size(sc, sc_sges, ia->ri_max_send_sges), ia 949 net/sunrpc/xprtrdma/verbs.c struct rpcrdma_ia *ia = &r_xprt->rx_ia; ia 952 net/sunrpc/xprtrdma/verbs.c for (count = 0; count < ia->ri_max_segs; count++) { ia 960 net/sunrpc/xprtrdma/verbs.c rc = frwr_init_mr(ia, mr); ia 1468 net/sunrpc/xprtrdma/verbs.c rpcrdma_ep_post(struct rpcrdma_ia *ia, ia 1483 net/sunrpc/xprtrdma/verbs.c rc = frwr_send(ia, req); ia 466 net/sunrpc/xprtrdma/xprt_rdma.h void rpcrdma_ia_remove(struct rpcrdma_ia *ia); ia 550 net/sunrpc/xprtrdma/xprt_rdma.h int frwr_open(struct rpcrdma_ia *ia, struct rpcrdma_ep *ep); ia 551 net/sunrpc/xprtrdma/xprt_rdma.h int frwr_init_mr(struct rpcrdma_ia *ia, struct rpcrdma_mr *mr); ia 558 net/sunrpc/xprtrdma/xprt_rdma.h int frwr_send(struct rpcrdma_ia *ia, struct rpcrdma_req *req); ia 668 tools/perf/util/annotate.c const struct ins *ia = a; ia 671 tools/perf/util/annotate.c return strcmp(ia->name, ib->name); ia 28 tools/perf/util/s390-cpumsf-kernel.h unsigned long long ia; /* Instruction Address */ ia 261 tools/perf/util/s390-cpumsf.c local.ia = be64toh(basicp->ia); ia 278 tools/perf/util/s390-cpumsf.c basic->AS, basic->prim_asn, basic->ia, basic->CL, ia 507 tools/perf/util/s390-cpumsf.c .ip = basic->ia,