/linux-4.4.14/drivers/staging/dgnc/ |
H A D | dgnc_mgmt.c | 165 struct ni_info ni; dgnc_mgmt_ioctl() local 170 if (copy_from_user(&ni, uarg, sizeof(ni))) dgnc_mgmt_ioctl() 173 board = ni.board; dgnc_mgmt_ioctl() 174 channel = ni.channel; dgnc_mgmt_ioctl() 189 memset(&ni, 0, sizeof(ni)); dgnc_mgmt_ioctl() 190 ni.board = board; dgnc_mgmt_ioctl() 191 ni.channel = channel; dgnc_mgmt_ioctl() 198 ni.mstat |= TIOCM_DTR; dgnc_mgmt_ioctl() 199 ni.dtr = TIOCM_DTR; dgnc_mgmt_ioctl() 202 ni.mstat |= TIOCM_RTS; dgnc_mgmt_ioctl() 203 ni.rts = TIOCM_RTS; dgnc_mgmt_ioctl() 206 ni.mstat |= TIOCM_CTS; dgnc_mgmt_ioctl() 207 ni.cts = TIOCM_CTS; dgnc_mgmt_ioctl() 210 ni.mstat |= TIOCM_RI; dgnc_mgmt_ioctl() 211 ni.ri = TIOCM_RI; dgnc_mgmt_ioctl() 214 ni.mstat |= TIOCM_CD; dgnc_mgmt_ioctl() 215 ni.dcd = TIOCM_CD; dgnc_mgmt_ioctl() 218 ni.mstat |= TIOCM_DSR; dgnc_mgmt_ioctl() 220 ni.iflag = ch->ch_c_iflag; dgnc_mgmt_ioctl() 221 ni.oflag = ch->ch_c_oflag; dgnc_mgmt_ioctl() 222 ni.cflag = ch->ch_c_cflag; dgnc_mgmt_ioctl() 223 ni.lflag = ch->ch_c_lflag; dgnc_mgmt_ioctl() 227 ni.hflow = 1; dgnc_mgmt_ioctl() 229 ni.hflow = 0; dgnc_mgmt_ioctl() 233 ni.recv_stopped = 1; dgnc_mgmt_ioctl() 235 ni.recv_stopped = 0; dgnc_mgmt_ioctl() 238 ni.xmit_stopped = 1; dgnc_mgmt_ioctl() 240 ni.xmit_stopped = 0; dgnc_mgmt_ioctl() 242 ni.curtx = ch->ch_txcount; dgnc_mgmt_ioctl() 243 ni.currx = ch->ch_rxcount; dgnc_mgmt_ioctl() 245 ni.baud = ch->ch_old_baud; dgnc_mgmt_ioctl() 249 if (copy_to_user(uarg, &ni, sizeof(ni))) dgnc_mgmt_ioctl()
|
/linux-4.4.14/drivers/staging/lustre/lnet/lnet/ |
H A D | Makefile | 3 lnet-y := api-ni.o config.o nidstrings.o \
|
H A D | api-ni.c | 656 lnet_ni_t *ni; lnet_net2ni_locked() local 661 ni = list_entry(tmp, lnet_ni_t, ni_list); lnet_net2ni_locked() 663 if (LNET_NIDNET(ni->ni_nid) == net) { lnet_net2ni_locked() 664 lnet_ni_addref_locked(ni, cpt); lnet_net2ni_locked() 665 return ni; lnet_net2ni_locked() 675 lnet_ni_t *ni; lnet_net2ni() local 678 ni = lnet_net2ni_locked(net, 0); lnet_net2ni() 681 return ni; lnet_net2ni() 707 struct lnet_ni *ni; lnet_cpt_of_nid_locked() local 715 list_for_each_entry(ni, &the_lnet.ln_nis_cpt, ni_cptlist) { lnet_cpt_of_nid_locked() 716 if (LNET_NIDNET(ni->ni_nid) != LNET_NIDNET(nid)) lnet_cpt_of_nid_locked() 719 LASSERT(ni->ni_cpts != NULL); lnet_cpt_of_nid_locked() 720 return ni->ni_cpts[lnet_nid_cpt_hash lnet_cpt_of_nid_locked() 721 (nid, ni->ni_ncpts)]; lnet_cpt_of_nid_locked() 751 struct lnet_ni *ni; lnet_islocalnet() local 756 ni = lnet_net2ni_locked(net, cpt); lnet_islocalnet() 757 if (ni != NULL) lnet_islocalnet() 758 lnet_ni_decref_locked(ni, cpt); lnet_islocalnet() 762 return ni != NULL; lnet_islocalnet() 768 struct lnet_ni *ni; lnet_nid2ni_locked() local 774 ni = list_entry(tmp, lnet_ni_t, ni_list); lnet_nid2ni_locked() 776 if (ni->ni_nid == nid) { lnet_nid2ni_locked() 777 lnet_ni_addref_locked(ni, cpt); lnet_nid2ni_locked() 778 return ni; lnet_nid2ni_locked() 788 struct lnet_ni *ni; lnet_islocalnid() local 792 ni = lnet_nid2ni_locked(nid, cpt); lnet_islocalnid() 793 if (ni != NULL) lnet_islocalnid() 794 lnet_ni_decref_locked(ni, cpt); lnet_islocalnid() 797 return ni != NULL; lnet_islocalnid() 806 struct lnet_ni *ni; lnet_count_acceptor_nis() local 811 ni = list_entry(tmp, lnet_ni_t, ni_list); lnet_count_acceptor_nis() 813 if (ni->ni_lnd->lnd_accept != NULL) lnet_count_acceptor_nis() 823 lnet_ni_tq_credits(lnet_ni_t *ni) lnet_ni_tq_credits() argument 827 LASSERT(ni->ni_ncpts >= 1); lnet_ni_tq_credits() 829 if (ni->ni_ncpts == 1) lnet_ni_tq_credits() 830 return ni->ni_maxtxcredits; lnet_ni_tq_credits() 832 credits = ni->ni_maxtxcredits / ni->ni_ncpts; lnet_ni_tq_credits() 833 credits = max(credits, 8 * ni->ni_peertxcredits); lnet_ni_tq_credits() 834 credits = min(credits, ni->ni_maxtxcredits); lnet_ni_tq_credits() 844 lnet_ni_t *ni; lnet_shutdown_lndnis() local 858 ni = list_entry(the_lnet.ln_nis.next, lnet_shutdown_lndnis() 861 list_move(&ni->ni_list, &the_lnet.ln_nis_zombie); lnet_shutdown_lndnis() 862 lnet_ni_decref_locked(ni, 0); /* drop ln_nis' ref */ lnet_shutdown_lndnis() 864 if (!list_empty(&ni->ni_cptlist)) { lnet_shutdown_lndnis() 865 list_del_init(&ni->ni_cptlist); lnet_shutdown_lndnis() 866 lnet_ni_decref_locked(ni, 0); lnet_shutdown_lndnis() 901 ni = list_entry(the_lnet.ln_nis_zombie.next, lnet_shutdown_lndnis() 903 list_del_init(&ni->ni_list); lnet_shutdown_lndnis() 904 cfs_percpt_for_each(ref, j, ni->ni_refs) { lnet_shutdown_lndnis() 908 list_add(&ni->ni_list, &the_lnet.ln_nis_zombie); lnet_shutdown_lndnis() 912 if (!list_empty(&ni->ni_list)) { lnet_shutdown_lndnis() 917 libcfs_nid2str(ni->ni_nid)); lnet_shutdown_lndnis() 925 ni->ni_lnd->lnd_refcount--; lnet_shutdown_lndnis() 928 islo = ni->ni_lnd->lnd_type == LOLND; lnet_shutdown_lndnis() 931 (ni->ni_lnd->lnd_shutdown)(ni); lnet_shutdown_lndnis() 938 libcfs_nid2str(ni->ni_nid)); lnet_shutdown_lndnis() 940 lnet_ni_free(ni); lnet_shutdown_lndnis() 960 struct lnet_ni *ni; lnet_startup_lndnis() local 979 ni = list_entry(nilist.next, lnet_ni_t, ni_list); lnet_startup_lndnis() 980 lnd_type = LNET_NETTYP(LNET_NIDNET(ni->ni_nid)); lnet_startup_lndnis() 1016 ni->ni_lnd = lnd; lnet_startup_lndnis() 1018 rc = (lnd->lnd_startup)(ni); lnet_startup_lndnis() 1031 LASSERT(ni->ni_peertimeout <= 0 || lnd->lnd_query != NULL); lnet_startup_lndnis() 1033 list_del(&ni->ni_list); lnet_startup_lndnis() 1037 lnet_ni_addref_locked(ni, 0); lnet_startup_lndnis() 1038 list_add_tail(&ni->ni_list, &the_lnet.ln_nis); lnet_startup_lndnis() 1039 if (ni->ni_cpts != NULL) { lnet_startup_lndnis() 1040 list_add_tail(&ni->ni_cptlist, lnet_startup_lndnis() 1042 lnet_ni_addref_locked(ni, 0); lnet_startup_lndnis() 1048 lnet_ni_addref(ni); lnet_startup_lndnis() 1050 the_lnet.ln_loni = ni; lnet_startup_lndnis() 1054 if (ni->ni_peertxcredits == 0 || lnet_startup_lndnis() 1055 ni->ni_maxtxcredits == 0) { lnet_startup_lndnis() 1058 ni->ni_peertxcredits == 0 ? lnet_startup_lndnis() 1063 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) { lnet_startup_lndnis() 1066 tq->tq_credits = lnet_ni_tq_credits(ni); lnet_startup_lndnis() 1070 libcfs_nid2str(ni->ni_nid), ni->ni_peertxcredits, lnet_startup_lndnis() 1071 lnet_ni_tq_credits(ni) * LNET_CPT_NUMBER, lnet_startup_lndnis() 1072 ni->ni_peerrtrcredits, ni->ni_peertimeout); lnet_startup_lndnis() 1090 ni = list_entry(nilist.next, lnet_ni_t, ni_list); lnet_startup_lndnis() 1091 list_del(&ni->ni_list); lnet_startup_lndnis() 1092 lnet_ni_free(ni); lnet_startup_lndnis() 1342 lnet_ni_t *ni; LNetCtl() local 1412 ni = lnet_net2ni(LNET_NIDNET(id.nid)); LNetCtl() 1413 if (ni == NULL) { LNetCtl() 1416 if (ni->ni_lnd->lnd_ctl == NULL) { LNetCtl() 1420 (void)ni->ni_lnd->lnd_ctl(ni, cmd, arg); LNetCtl() 1423 lnet_ni_decref(ni); LNetCtl() 1429 ni = lnet_net2ni(data->ioc_net); LNetCtl() 1430 if (ni == NULL) LNetCtl() 1433 if (ni->ni_lnd->lnd_ctl == NULL) LNetCtl() 1436 rc = ni->ni_lnd->lnd_ctl(ni, cmd, arg); LNetCtl() 1438 lnet_ni_decref(ni); LNetCtl() 1459 struct lnet_ni *ni; LNetGetId() local 1476 ni = list_entry(tmp, lnet_ni_t, ni_list); LNetGetId() 1478 id->nid = ni->ni_nid; LNetGetId() 1507 lnet_ni_t *ni; lnet_create_ping_info() local 1542 ni = lnet_nid2ni_locked(id.nid, 0); lnet_create_ping_info() 1543 LASSERT(ni != NULL); lnet_create_ping_info() 1545 lnet_ni_lock(ni); lnet_create_ping_info() 1546 LASSERT(ni->ni_status == NULL); lnet_create_ping_info() 1547 ni->ni_status = ns; lnet_create_ping_info() 1548 lnet_ni_unlock(ni); lnet_create_ping_info() 1550 lnet_ni_decref_locked(ni, 0); lnet_create_ping_info() 1561 struct lnet_ni *ni; lnet_destroy_ping_info() local 1565 list_for_each_entry(ni, &the_lnet.ln_nis, ni_list) { lnet_destroy_ping_info() 1566 lnet_ni_lock(ni); lnet_destroy_ping_info() 1567 ni->ni_status = NULL; lnet_destroy_ping_info() 1568 lnet_ni_unlock(ni); lnet_destroy_ping_info()
|
H A D | lo.c | 39 lolnd_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg) lolnd_send() argument 44 return lnet_parse(ni, &lntmsg->msg_hdr, ni->ni_nid, lntmsg, 0); lolnd_send() 48 lolnd_recv(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg, lolnd_recv() argument 80 lnet_finalize(ni, lntmsg, 0); lolnd_recv() 83 lnet_finalize(ni, sendmsg, 0); lolnd_recv() 90 lolnd_shutdown(lnet_ni_t *ni) lolnd_shutdown() argument 99 lolnd_startup(lnet_ni_t *ni) lolnd_startup() argument 101 LASSERT(ni->ni_lnd == &the_lolnd); lolnd_startup()
|
H A D | config.c | 84 lnet_ni_t *ni; lnet_net_unique() local 87 ni = list_entry(tmp, lnet_ni_t, ni_list); list_for_each() 89 if (LNET_NIDNET(ni->ni_nid) == net) list_for_each() 97 lnet_ni_free(struct lnet_ni *ni) lnet_ni_free() argument 99 if (ni->ni_refs != NULL) lnet_ni_free() 100 cfs_percpt_free(ni->ni_refs); lnet_ni_free() 102 if (ni->ni_tx_queues != NULL) lnet_ni_free() 103 cfs_percpt_free(ni->ni_tx_queues); lnet_ni_free() 105 if (ni->ni_cpts != NULL) lnet_ni_free() 106 cfs_expr_list_values_free(ni->ni_cpts, ni->ni_ncpts); lnet_ni_free() 108 LIBCFS_FREE(ni, sizeof(*ni)); lnet_ni_free() 115 struct lnet_ni *ni; lnet_ni_alloc() local 125 LIBCFS_ALLOC(ni, sizeof(*ni)); lnet_ni_alloc() 126 if (ni == NULL) { lnet_ni_alloc() 132 spin_lock_init(&ni->ni_lock); lnet_ni_alloc() 133 INIT_LIST_HEAD(&ni->ni_cptlist); lnet_ni_alloc() 134 ni->ni_refs = cfs_percpt_alloc(lnet_cpt_table(), lnet_ni_alloc() 135 sizeof(*ni->ni_refs[0])); lnet_ni_alloc() 136 if (ni->ni_refs == NULL) lnet_ni_alloc() 139 ni->ni_tx_queues = cfs_percpt_alloc(lnet_cpt_table(), lnet_ni_alloc() 140 sizeof(*ni->ni_tx_queues[0])); lnet_ni_alloc() 141 if (ni->ni_tx_queues == NULL) lnet_ni_alloc() 144 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) lnet_ni_alloc() 148 ni->ni_cpts = NULL; lnet_ni_alloc() 149 ni->ni_ncpts = LNET_CPT_NUMBER; lnet_ni_alloc() 151 rc = cfs_expr_list_values(el, LNET_CPT_NUMBER, &ni->ni_cpts); lnet_ni_alloc() 160 LIBCFS_FREE(ni->ni_cpts, rc * sizeof(ni->ni_cpts[0])); lnet_ni_alloc() 161 ni->ni_cpts = NULL; lnet_ni_alloc() 164 ni->ni_ncpts = rc; lnet_ni_alloc() 168 ni->ni_nid = LNET_MKNID(net, 0); lnet_ni_alloc() 169 ni->ni_last_alive = ktime_get_real_seconds(); lnet_ni_alloc() 170 list_add_tail(&ni->ni_list, nilist); lnet_ni_alloc() 171 return ni; lnet_ni_alloc() 173 lnet_ni_free(ni); lnet_ni_alloc() 185 struct lnet_ni *ni; lnet_parse_networks() local 208 ni = lnet_ni_alloc(LNET_MKNET(LOLND, 0), NULL, nilist); lnet_parse_networks() 209 if (ni == NULL) lnet_parse_networks() 285 ni = lnet_ni_alloc(net, el, nilist); lnet_parse_networks() 286 if (ni == NULL) lnet_parse_networks() 322 ni->ni_interfaces[niface++] = iface; lnet_parse_networks() 353 ni = list_entry(nilist->next, lnet_ni_t, ni_list); lnet_parse_networks() 355 list_del(&ni->ni_list); lnet_parse_networks() 356 lnet_ni_free(ni); lnet_parse_networks()
|
H A D | lib-move.c | 564 lnet_ni_recv(lnet_ni_t *ni, void *private, lnet_msg_t *msg, int delayed, lnet_ni_recv() argument 595 rc = (ni->ni_lnd->lnd_recv)(ni, private, msg, delayed, lnet_ni_recv() 598 lnet_finalize(ni, msg, rc); lnet_ni_recv() 642 lnet_ni_send(lnet_ni_t *ni, lnet_msg_t *msg) lnet_ni_send() argument 648 LASSERT(LNET_NETTYP(LNET_NIDNET(ni->ni_nid)) == LOLND || lnet_ni_send() 651 rc = (ni->ni_lnd->lnd_send)(ni, priv, msg); lnet_ni_send() 653 lnet_finalize(ni, msg, rc); lnet_ni_send() 657 lnet_ni_eager_recv(lnet_ni_t *ni, lnet_msg_t *msg) lnet_ni_eager_recv() argument 664 LASSERT(ni->ni_lnd->lnd_eager_recv != NULL); lnet_ni_eager_recv() 667 rc = (ni->ni_lnd->lnd_eager_recv)(ni, msg->msg_private, msg, lnet_ni_eager_recv() 681 lnet_ni_query_locked(lnet_ni_t *ni, lnet_peer_t *lp) lnet_ni_query_locked() argument 686 LASSERT(ni->ni_lnd->lnd_query != NULL); lnet_ni_query_locked() 689 (ni->ni_lnd->lnd_query)(ni, lp->lp_nid, &last_alive); lnet_ni_query_locked() 787 lnet_ni_t *ni = lp->lp_ni; lnet_post_send_locked() local 789 struct lnet_tx_queue *tq = ni->ni_tx_queues[cpt]; lnet_post_send_locked() 806 lnet_finalize(ni, msg, -EHOSTUNREACH); lnet_post_send_locked() 819 lnet_finalize(ni, msg, -ECANCELED); lnet_post_send_locked() 862 lnet_ni_send(ni, msg); lnet_post_send_locked() 971 struct lnet_ni *ni = txpeer->lp_ni; lnet_return_tx_credits_locked() local 972 struct lnet_tx_queue *tq = ni->ni_tx_queues[msg->msg_tx_cpt]; lnet_return_tx_credits_locked() 986 LASSERT(msg2->msg_txpeer->lp_ni == ni); lnet_return_tx_credits_locked() 1120 lnet_find_route_locked(lnet_ni_t *ni, lnet_nid_t target, lnet_nid_t rtr_nid) lnet_find_route_locked() argument 1147 if (ni != NULL && lp->lp_ni != ni) lnet_find_route_locked() 1193 /* NB: ni != NULL == interface pre-determined (ACK/REPLY) */ lnet_send() 1351 lnet_drop_message(lnet_ni_t *ni, int cpt, void *private, unsigned int nob) lnet_drop_message() argument 1358 lnet_ni_recv(ni, private, NULL, 0, 0, 0, nob); lnet_drop_message() 1362 lnet_recv_put(lnet_ni_t *ni, lnet_msg_t *msg) lnet_recv_put() argument 1376 lnet_ni_recv(ni, msg->msg_private, msg, msg->msg_rx_delayed, lnet_recv_put() 1381 lnet_parse_put(lnet_ni_t *ni, lnet_msg_t *msg) lnet_parse_put() argument 1400 msg->msg_rx_ready_delay = ni->ni_lnd->lnd_eager_recv == NULL; lnet_parse_put() 1409 lnet_recv_put(ni, msg); lnet_parse_put() 1416 rc = lnet_ni_eager_recv(ni, msg); lnet_parse_put() 1431 lnet_parse_get(lnet_ni_t *ni, lnet_msg_t *msg, int rdma_get) lnet_parse_get() argument 1473 lnet_ni_recv(ni, msg->msg_private, msg, 0, lnet_parse_get() 1478 lnet_ni_recv(ni, msg->msg_private, NULL, 0, 0, 0, 0); lnet_parse_get() 1481 rc = lnet_send(ni->ni_nid, msg, LNET_NID_ANY); lnet_parse_get() 1485 libcfs_nid2str(ni->ni_nid), lnet_parse_get() 1488 lnet_finalize(ni, msg, rc); lnet_parse_get() 1495 lnet_parse_reply(lnet_ni_t *ni, lnet_msg_t *msg) lnet_parse_reply() argument 1515 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_reply() 1535 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_reply() 1543 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_reply() 1555 lnet_ni_recv(ni, private, msg, 0, 0, mlength, rlength); lnet_parse_reply() 1560 lnet_parse_ack(lnet_ni_t *ni, lnet_msg_t *msg) lnet_parse_ack() argument 1583 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_ack() 1596 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_ack() 1605 lnet_ni_recv(ni, msg->msg_private, msg, 0, 0, 0, msg->msg_len); lnet_parse_ack() 1610 lnet_parse_forward_locked(lnet_ni_t *ni, lnet_msg_t *msg) lnet_parse_forward_locked() argument 1616 if (ni->ni_lnd->lnd_eager_recv == NULL) { lnet_parse_forward_locked() 1620 rc = lnet_ni_eager_recv(ni, msg); lnet_parse_forward_locked() 1710 lnet_parse(lnet_ni_t *ni, lnet_hdr_t *hdr, lnet_nid_t from_nid, lnet_parse() argument 1731 for_me = (ni->ni_nid == dest_nid); lnet_parse() 1768 ni->ni_last_alive != ktime_get_real_seconds()) { lnet_parse() 1769 lnet_ni_lock(ni); lnet_parse() 1772 ni->ni_last_alive = ktime_get_real_seconds(); lnet_parse() 1773 if (ni->ni_status != NULL && lnet_parse() 1774 ni->ni_status->ns_status == LNET_NI_STATUS_DOWN) lnet_parse() 1775 ni->ni_status->ns_status = LNET_NI_STATUS_UP; lnet_parse() 1776 lnet_ni_unlock(ni); lnet_parse() 1784 if (LNET_NIDNET(dest_nid) == LNET_NIDNET(ni->ni_nid)) { lnet_parse() 1893 rc = lnet_parse_forward_locked(ni, msg); lnet_parse() 1899 lnet_ni_recv(ni, msg->msg_private, msg, 0, lnet_parse() 1909 rc = lnet_parse_ack(ni, msg); lnet_parse() 1912 rc = lnet_parse_put(ni, msg); lnet_parse() 1915 rc = lnet_parse_get(ni, msg, rdma_req); lnet_parse() 1918 rc = lnet_parse_reply(ni, msg); lnet_parse() 1933 lnet_finalize(ni, msg, rc); lnet_parse() 1936 lnet_drop_message(ni, cpt, private, payload_length); lnet_parse() 2144 lnet_create_reply_msg(lnet_ni_t *ni, lnet_msg_t *getmsg) lnet_create_reply_msg() argument 2168 libcfs_nid2str(ni->ni_nid), libcfs_id2str(peer_id)); lnet_create_reply_msg() 2174 libcfs_nid2str(ni->ni_nid), libcfs_id2str(peer_id), lnet_create_reply_msg() 2183 libcfs_nid2str(ni->ni_nid), libcfs_id2str(peer_id), getmd); lnet_create_reply_msg() 2221 lnet_set_reply_msg_len(lnet_ni_t *ni, lnet_msg_t *reply, unsigned int len) lnet_set_reply_msg_len() argument 2354 struct lnet_ni *ni; LNetDist() local 2373 ni = list_entry(e, lnet_ni_t, ni_list); LNetDist() 2375 if (ni->ni_nid == dstnid) { LNetDist() 2389 if (LNET_NIDNET(ni->ni_nid) == dstnet) { LNetDist() 2391 *srcnidp = ni->ni_nid; LNetDist()
|
H A D | router.c | 56 lnet_peer_buffer_credits(lnet_ni_t *ni) lnet_peer_buffer_credits() argument 59 if (ni->ni_peerrtrcredits > 0) lnet_peer_buffer_credits() 60 return ni->ni_peerrtrcredits; lnet_peer_buffer_credits() 66 return ni->ni_peertxcredits; lnet_peer_buffer_credits() 129 lnet_ni_notify_locked(lnet_ni_t *ni, lnet_peer_t *lp) lnet_ni_notify_locked() argument 138 if (lp->lp_notifying || ni == NULL) lnet_ni_notify_locked() 150 if (notifylnd && ni->ni_lnd->lnd_notify != NULL) { lnet_ni_notify_locked() 156 (ni->ni_lnd->lnd_notify)(ni, lp->lp_nid, alive); lnet_ni_notify_locked() 240 lnet_ni_t *ni; lnet_shuffle_seed() local 251 ni = list_entry(tmp, lnet_ni_t, ni_list); lnet_shuffle_seed() 252 lnd_type = LNET_NETTYP(LNET_NIDNET(ni->ni_nid)); lnet_shuffle_seed() 255 seed[0] ^= (LNET_NIDADDR(ni->ni_nid) | lnd_type); lnet_shuffle_seed() 299 lnet_ni_t *ni; lnet_add_route() local 381 ni = route->lr_gateway->lp_ni; lnet_add_route() 385 if (ni->ni_lnd->lnd_notify != NULL) lnet_add_route() 386 (ni->ni_lnd->lnd_notify)(ni, gateway, 1); lnet_add_route() 790 lnet_ni_t *ni; lnet_update_ni_status_locked() local 800 list_for_each_entry(ni, &the_lnet.ln_nis, ni_list) { lnet_update_ni_status_locked() 801 if (ni->ni_lnd->lnd_type == LOLND) lnet_update_ni_status_locked() 804 if (now < ni->ni_last_alive + timeout) lnet_update_ni_status_locked() 807 lnet_ni_lock(ni); lnet_update_ni_status_locked() 809 if (now < ni->ni_last_alive + timeout) { lnet_update_ni_status_locked() 810 lnet_ni_unlock(ni); lnet_update_ni_status_locked() 814 LASSERT(ni->ni_status != NULL); lnet_update_ni_status_locked() 816 if (ni->ni_status->ns_status != LNET_NI_STATUS_DOWN) { lnet_update_ni_status_locked() 818 libcfs_nid2str(ni->ni_nid), timeout); lnet_update_ni_status_locked() 821 ni->ni_status->ns_status = LNET_NI_STATUS_DOWN; lnet_update_ni_status_locked() 823 lnet_ni_unlock(ni); lnet_update_ni_status_locked() 1493 lnet_notify(lnet_ni_t *ni, lnet_nid_t nid, int alive, unsigned long when) lnet_notify() argument 1502 (ni == NULL) ? "userspace" : libcfs_nid2str(ni->ni_nid), lnet_notify() 1506 if (ni != NULL && lnet_notify() 1507 LNET_NIDNET(ni->ni_nid) != LNET_NIDNET(nid)) { lnet_notify() 1510 libcfs_nid2str(ni->ni_nid)); lnet_notify() 1517 (ni == NULL) ? "userspace" : libcfs_nid2str(ni->ni_nid), lnet_notify() 1523 if (ni != NULL && !alive && /* LND telling me she's down */ lnet_notify() 1548 if (ni != NULL && !alive && when < lp->lp_last_alive) lnet_notify() 1551 lnet_notify_locked(lp, ni == NULL, alive, when); lnet_notify() 1553 lnet_ni_notify_locked(ni, lp); lnet_notify()
|
H A D | router_proc.c | 673 lnet_ni_t *ni = NULL; proc_lnet_nis() local 684 ni = a_ni; proc_lnet_nis() 692 if (ni != NULL) { proc_lnet_nis() 701 last_alive = now - ni->ni_last_alive; proc_lnet_nis() 704 if (ni->ni_lnd->lnd_type == LOLND) proc_lnet_nis() 707 lnet_ni_lock(ni); proc_lnet_nis() 708 LASSERT(ni->ni_status != NULL); proc_lnet_nis() 709 stat = (ni->ni_status->ns_status == proc_lnet_nis() 711 lnet_ni_unlock(ni); proc_lnet_nis() 715 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) { proc_lnet_nis() 716 for (j = 0; ni->ni_cpts != NULL && proc_lnet_nis() 717 j < ni->ni_ncpts; j++) { proc_lnet_nis() 718 if (i == ni->ni_cpts[j]) proc_lnet_nis() 722 if (j == ni->ni_ncpts) proc_lnet_nis() 730 libcfs_nid2str(ni->ni_nid), stat, proc_lnet_nis() 731 last_alive, *ni->ni_refs[i], proc_lnet_nis() 732 ni->ni_peertxcredits, proc_lnet_nis() 733 ni->ni_peerrtrcredits, proc_lnet_nis()
|
H A D | acceptor.c | 215 lnet_ni_t *ni; lnet_accept() local 297 ni = lnet_net2ni(LNET_NIDNET(cr.acr_nid)); lnet_accept() 298 if (ni == NULL || /* no matching net */ lnet_accept() 299 ni->ni_nid != cr.acr_nid) { /* right NET, wrong NID! */ lnet_accept() 300 if (ni != NULL) lnet_accept() 301 lnet_ni_decref(ni); lnet_accept() 307 if (ni->ni_lnd->lnd_accept == NULL) { lnet_accept() 309 lnet_ni_decref(ni); lnet_accept() 318 rc = ni->ni_lnd->lnd_accept(ni, sock); lnet_accept() 320 lnet_ni_decref(ni); lnet_accept()
|
H A D | lib-msg.c | 435 lnet_finalize(lnet_ni_t *ni, lnet_msg_t *msg, int status) lnet_finalize() argument
|
/linux-4.4.14/fs/ntfs/ |
H A D | mft.h | 32 extern MFT_RECORD *map_mft_record(ntfs_inode *ni); 33 extern void unmap_mft_record(ntfs_inode *ni); 38 static inline void unmap_extent_mft_record(ntfs_inode *ni) unmap_extent_mft_record() argument 40 unmap_mft_record(ni); unmap_extent_mft_record() 48 * @ni: ntfs inode structure of mft record 55 static inline void flush_dcache_mft_record_page(ntfs_inode *ni) flush_dcache_mft_record_page() argument 57 flush_dcache_page(ni->page); flush_dcache_mft_record_page() 60 extern void __mark_mft_record_dirty(ntfs_inode *ni); 64 * @ni: ntfs inode describing the mapped mft record 66 * Set the mapped (extent) mft record of the (base or extent) ntfs inode @ni, 73 static inline void mark_mft_record_dirty(ntfs_inode *ni) mark_mft_record_dirty() argument 75 if (!NInoTestSetDirty(ni)) mark_mft_record_dirty() 76 __mark_mft_record_dirty(ni); mark_mft_record_dirty() 82 extern int write_mft_record_nolock(ntfs_inode *ni, MFT_RECORD *m, int sync); 86 * @ni: ntfs inode describing the mapped (extent) mft record 102 static inline int write_mft_record(ntfs_inode *ni, MFT_RECORD *m, int sync) write_mft_record() argument 104 struct page *page = ni->page; write_mft_record() 109 err = write_mft_record_nolock(ni, m, sync); write_mft_record() 120 extern int ntfs_extent_mft_record_free(ntfs_inode *ni, MFT_RECORD *m);
|
H A D | inode.c | 62 ntfs_inode *ni; ntfs_test_inode() local 66 ni = NTFS_I(vi); ntfs_test_inode() 67 /* If !NInoAttr(ni), @vi is a normal file or directory inode. */ ntfs_test_inode() 68 if (likely(!NInoAttr(ni))) { ntfs_test_inode() 74 if (ni->type != na->type) ntfs_test_inode() 76 if (ni->name_len != na->name_len) ntfs_test_inode() 78 if (na->name_len && memcmp(ni->name, na->name, ntfs_test_inode() 106 ntfs_inode *ni = NTFS_I(vi); ntfs_init_locked_inode() local 110 ni->type = na->type; ntfs_init_locked_inode() 112 NInoSetMstProtected(ni); ntfs_init_locked_inode() 114 ni->name = na->name; ntfs_init_locked_inode() 115 ni->name_len = na->name_len; ntfs_init_locked_inode() 125 NInoSetAttr(ni); ntfs_init_locked_inode() 139 ni->name = kmalloc(i + sizeof(ntfschar), GFP_ATOMIC); ntfs_init_locked_inode() 140 if (!ni->name) ntfs_init_locked_inode() 142 memcpy(ni->name, na->name, i); ntfs_init_locked_inode() 143 ni->name[na->name_len] = 0; ntfs_init_locked_inode() 323 ntfs_inode *ni; ntfs_alloc_big_inode() local 326 ni = kmem_cache_alloc(ntfs_big_inode_cache, GFP_NOFS); ntfs_alloc_big_inode() 327 if (likely(ni != NULL)) { ntfs_alloc_big_inode() 328 ni->state = 0; ntfs_alloc_big_inode() 329 return VFS_I(ni); ntfs_alloc_big_inode() 343 ntfs_inode *ni = NTFS_I(inode); ntfs_destroy_big_inode() local 346 BUG_ON(ni->page); ntfs_destroy_big_inode() 347 if (!atomic_dec_and_test(&ni->count)) ntfs_destroy_big_inode() 354 ntfs_inode *ni; ntfs_alloc_extent_inode() local 357 ni = kmem_cache_alloc(ntfs_inode_cache, GFP_NOFS); ntfs_alloc_extent_inode() 358 if (likely(ni != NULL)) { ntfs_alloc_extent_inode() 359 ni->state = 0; ntfs_alloc_extent_inode() 360 return ni; ntfs_alloc_extent_inode() 366 static void ntfs_destroy_extent_inode(ntfs_inode *ni) ntfs_destroy_extent_inode() argument 369 BUG_ON(ni->page); ntfs_destroy_extent_inode() 370 if (!atomic_dec_and_test(&ni->count)) ntfs_destroy_extent_inode() 372 kmem_cache_free(ntfs_inode_cache, ni); ntfs_destroy_extent_inode() 384 * @ni: freshly allocated ntfs inode which to initialize 388 * NOTE: ni->mft_no, ni->state, ni->type, ni->name, and ni->name_len are left 393 void __ntfs_init_inode(struct super_block *sb, ntfs_inode *ni) __ntfs_init_inode() argument 396 rwlock_init(&ni->size_lock); __ntfs_init_inode() 397 ni->initialized_size = ni->allocated_size = 0; __ntfs_init_inode() 398 ni->seq_no = 0; __ntfs_init_inode() 399 atomic_set(&ni->count, 1); __ntfs_init_inode() 400 ni->vol = NTFS_SB(sb); __ntfs_init_inode() 401 ntfs_init_runlist(&ni->runlist); __ntfs_init_inode() 402 mutex_init(&ni->mrec_lock); __ntfs_init_inode() 403 ni->page = NULL; __ntfs_init_inode() 404 ni->page_ofs = 0; __ntfs_init_inode() 405 ni->attr_list_size = 0; __ntfs_init_inode() 406 ni->attr_list = NULL; __ntfs_init_inode() 407 ntfs_init_runlist(&ni->attr_list_rl); __ntfs_init_inode() 408 lockdep_set_class(&ni->attr_list_rl.lock, __ntfs_init_inode() 410 ni->itype.index.block_size = 0; __ntfs_init_inode() 411 ni->itype.index.vcn_size = 0; __ntfs_init_inode() 412 ni->itype.index.collation_rule = 0; __ntfs_init_inode() 413 ni->itype.index.block_size_bits = 0; __ntfs_init_inode() 414 ni->itype.index.vcn_size_bits = 0; __ntfs_init_inode() 415 mutex_init(&ni->extent_lock); __ntfs_init_inode() 416 ni->nr_extents = 0; __ntfs_init_inode() 417 ni->ext.base_ntfs_ino = NULL; __ntfs_init_inode() 430 ntfs_inode *ni = ntfs_alloc_extent_inode(); ntfs_new_extent_inode() local 433 if (likely(ni != NULL)) { ntfs_new_extent_inode() 434 __ntfs_init_inode(sb, ni); ntfs_new_extent_inode() 435 lockdep_set_class(&ni->mrec_lock, &extent_inode_mrec_lock_key); ntfs_new_extent_inode() 436 ni->mft_no = mft_no; ntfs_new_extent_inode() 437 ni->type = AT_UNUSED; ntfs_new_extent_inode() 438 ni->name = NULL; ntfs_new_extent_inode() 439 ni->name_len = 0; ntfs_new_extent_inode() 441 return ni; ntfs_new_extent_inode() 552 ntfs_inode *ni; ntfs_read_locked_inode() local 580 ni = NTFS_I(vi); ntfs_read_locked_inode() 582 m = map_mft_record(ni); ntfs_read_locked_inode() 587 ctx = ntfs_attr_get_search_ctx(ni, m); ntfs_read_locked_inode() 603 vi->i_generation = ni->seq_no = le16_to_cpu(m->sequence_number); ntfs_read_locked_inode() 703 NInoSetAttrList(ni); ntfs_read_locked_inode() 728 ni->attr_list_size = (u32)ntfs_attr_size(a); ntfs_read_locked_inode() 729 ni->attr_list = ntfs_malloc_nofs(ni->attr_list_size); ntfs_read_locked_inode() 730 if (!ni->attr_list) { ntfs_read_locked_inode() 737 NInoSetAttrListNonResident(ni); ntfs_read_locked_inode() 747 ni->attr_list_rl.rl = ntfs_mapping_pairs_decompress(vol, ntfs_read_locked_inode() 749 if (IS_ERR(ni->attr_list_rl.rl)) { ntfs_read_locked_inode() 750 err = PTR_ERR(ni->attr_list_rl.rl); ntfs_read_locked_inode() 751 ni->attr_list_rl.rl = NULL; ntfs_read_locked_inode() 757 if ((err = load_attribute_list(vol, &ni->attr_list_rl, ntfs_read_locked_inode() 758 ni->attr_list, ni->attr_list_size, ntfs_read_locked_inode() 775 memcpy(ni->attr_list, (u8*)a + le16_to_cpu( ntfs_read_locked_inode() 827 NInoSetCompressed(ni); ntfs_read_locked_inode() 834 NInoSetEncrypted(ni); ntfs_read_locked_inode() 837 NInoSetSparse(ni); ntfs_read_locked_inode() 862 ni->itype.index.collation_rule = ir->collation_rule; ntfs_read_locked_inode() 863 ni->itype.index.block_size = le32_to_cpu(ir->index_block_size); ntfs_read_locked_inode() 864 if (ni->itype.index.block_size & ntfs_read_locked_inode() 865 (ni->itype.index.block_size - 1)) { ntfs_read_locked_inode() 868 ni->itype.index.block_size); ntfs_read_locked_inode() 871 if (ni->itype.index.block_size > PAGE_CACHE_SIZE) { ntfs_read_locked_inode() 875 ni->itype.index.block_size, ntfs_read_locked_inode() 880 if (ni->itype.index.block_size < NTFS_BLOCK_SIZE) { ntfs_read_locked_inode() 884 ni->itype.index.block_size, ntfs_read_locked_inode() 889 ni->itype.index.block_size_bits = ntfs_read_locked_inode() 890 ffs(ni->itype.index.block_size) - 1; ntfs_read_locked_inode() 892 if (vol->cluster_size <= ni->itype.index.block_size) { ntfs_read_locked_inode() 893 ni->itype.index.vcn_size = vol->cluster_size; ntfs_read_locked_inode() 894 ni->itype.index.vcn_size_bits = vol->cluster_size_bits; ntfs_read_locked_inode() 896 ni->itype.index.vcn_size = vol->sector_size; ntfs_read_locked_inode() 897 ni->itype.index.vcn_size_bits = vol->sector_size_bits; ntfs_read_locked_inode() 901 NInoSetMstProtected(ni); ntfs_read_locked_inode() 902 ni->type = AT_INDEX_ALLOCATION; ntfs_read_locked_inode() 903 ni->name = I30; ntfs_read_locked_inode() 904 ni->name_len = 4; ntfs_read_locked_inode() 908 vi->i_size = ni->initialized_size = ntfs_read_locked_inode() 909 ni->allocated_size = 0; ntfs_read_locked_inode() 912 unmap_mft_record(ni); ntfs_read_locked_inode() 917 NInoSetIndexAllocPresent(ni); ntfs_read_locked_inode() 973 ni->initialized_size = sle64_to_cpu( ntfs_read_locked_inode() 975 ni->allocated_size = sle64_to_cpu( ntfs_read_locked_inode() 982 unmap_mft_record(ni); ntfs_read_locked_inode() 1002 ni->itype.index.block_size_bits)) { ntfs_read_locked_inode() 1020 ni->type = AT_DATA; ntfs_read_locked_inode() 1021 ni->name = NULL; ntfs_read_locked_inode() 1022 ni->name_len = 0; ntfs_read_locked_inode() 1027 vi->i_size = ni->initialized_size = ntfs_read_locked_inode() 1028 ni->allocated_size = 0; ntfs_read_locked_inode() 1060 NInoSetCompressed(ni); ntfs_read_locked_inode() 1080 NInoSetSparse(ni); ntfs_read_locked_inode() 1083 if (NInoCompressed(ni)) { ntfs_read_locked_inode() 1088 NInoSetEncrypted(ni); ntfs_read_locked_inode() 1091 NInoSetNonResident(ni); ntfs_read_locked_inode() 1092 if (NInoCompressed(ni) || NInoSparse(ni)) { ntfs_read_locked_inode() 1093 if (NInoCompressed(ni) && a->data.non_resident. ntfs_read_locked_inode() 1106 ni->itype.compressed.block_size = 1U << ntfs_read_locked_inode() 1110 ni->itype.compressed.block_size_bits = ntfs_read_locked_inode() 1111 ffs(ni->itype. ntfs_read_locked_inode() 1114 ni->itype.compressed.block_clusters = ntfs_read_locked_inode() 1119 ni->itype.compressed.block_size = 0; ntfs_read_locked_inode() 1120 ni->itype.compressed.block_size_bits = ntfs_read_locked_inode() 1122 ni->itype.compressed.block_clusters = ntfs_read_locked_inode() 1125 ni->itype.compressed.size = sle64_to_cpu( ntfs_read_locked_inode() 1137 ni->initialized_size = sle64_to_cpu( ntfs_read_locked_inode() 1139 ni->allocated_size = sle64_to_cpu( ntfs_read_locked_inode() 1142 vi->i_size = ni->initialized_size = le32_to_cpu( ntfs_read_locked_inode() 1144 ni->allocated_size = le32_to_cpu(a->length) - ntfs_read_locked_inode() 1147 if (vi->i_size > ni->allocated_size) { ntfs_read_locked_inode() 1157 unmap_mft_record(ni); ntfs_read_locked_inode() 1164 if (NInoMstProtected(ni)) ntfs_read_locked_inode() 1166 else if (NInoCompressed(ni)) ntfs_read_locked_inode() 1180 if (S_ISREG(vi->i_mode) && (NInoCompressed(ni) || NInoSparse(ni))) ntfs_read_locked_inode() 1181 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_read_locked_inode() 1183 vi->i_blocks = ni->allocated_size >> 9; ntfs_read_locked_inode() 1194 unmap_mft_record(ni); ntfs_read_locked_inode() 1229 ntfs_inode *ni, *base_ni; ntfs_read_locked_attr_inode() local 1239 ni = NTFS_I(vi); ntfs_read_locked_attr_inode() 1250 vi->i_generation = ni->seq_no = base_ni->seq_no; ntfs_read_locked_attr_inode() 1266 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_read_locked_attr_inode() 1273 NInoSetCompressed(ni); ntfs_read_locked_attr_inode() 1274 if ((ni->type != AT_DATA) || (ni->type == AT_DATA && ntfs_read_locked_attr_inode() 1275 ni->name_len)) { ntfs_read_locked_attr_inode() 1303 if (NInoMstProtected(ni) && ni->type != AT_INDEX_ROOT) { ntfs_read_locked_attr_inode() 1308 NInoCompressed(ni) ? "compressed" : ntfs_read_locked_attr_inode() 1313 NInoSetSparse(ni); ntfs_read_locked_attr_inode() 1316 if (NInoCompressed(ni)) { ntfs_read_locked_attr_inode() 1325 if (NInoMstProtected(ni) && ni->type != AT_INDEX_ROOT) { ntfs_read_locked_attr_inode() 1333 if (ni->type != AT_DATA) { ntfs_read_locked_attr_inode() 1338 NInoSetEncrypted(ni); ntfs_read_locked_attr_inode() 1348 if (NInoMstProtected(ni)) { ntfs_read_locked_attr_inode() 1355 vi->i_size = ni->initialized_size = le32_to_cpu( ntfs_read_locked_attr_inode() 1357 ni->allocated_size = le32_to_cpu(a->length) - ntfs_read_locked_attr_inode() 1359 if (vi->i_size > ni->allocated_size) { ntfs_read_locked_attr_inode() 1365 NInoSetNonResident(ni); ntfs_read_locked_attr_inode() 1377 if (NInoCompressed(ni) || NInoSparse(ni)) { ntfs_read_locked_attr_inode() 1378 if (NInoCompressed(ni) && a->data.non_resident. ntfs_read_locked_attr_inode() 1389 ni->itype.compressed.block_size = 1U << ntfs_read_locked_attr_inode() 1393 ni->itype.compressed.block_size_bits = ntfs_read_locked_attr_inode() 1394 ffs(ni->itype.compressed. ntfs_read_locked_attr_inode() 1396 ni->itype.compressed.block_clusters = 1U << ntfs_read_locked_attr_inode() 1400 ni->itype.compressed.block_size = 0; ntfs_read_locked_attr_inode() 1401 ni->itype.compressed.block_size_bits = 0; ntfs_read_locked_attr_inode() 1402 ni->itype.compressed.block_clusters = 0; ntfs_read_locked_attr_inode() 1404 ni->itype.compressed.size = sle64_to_cpu( ntfs_read_locked_attr_inode() 1413 ni->initialized_size = sle64_to_cpu( ntfs_read_locked_attr_inode() 1415 ni->allocated_size = sle64_to_cpu( ntfs_read_locked_attr_inode() 1419 if (NInoMstProtected(ni)) ntfs_read_locked_attr_inode() 1421 else if (NInoCompressed(ni)) ntfs_read_locked_attr_inode() 1423 if ((NInoCompressed(ni) || NInoSparse(ni)) && ni->type != AT_INDEX_ROOT) ntfs_read_locked_attr_inode() 1424 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_read_locked_attr_inode() 1426 vi->i_blocks = ni->allocated_size >> 9; ntfs_read_locked_attr_inode() 1432 ni->ext.base_ntfs_ino = base_ni; ntfs_read_locked_attr_inode() 1433 ni->nr_extents = -1; ntfs_read_locked_attr_inode() 1451 "Run chkdsk.", err, vi->i_ino, ni->type, ni->name_len, ntfs_read_locked_attr_inode() 1496 ntfs_inode *ni, *base_ni, *bni; ntfs_read_locked_index_inode() local 1507 ni = NTFS_I(vi); ntfs_read_locked_index_inode() 1517 vi->i_generation = ni->seq_no = base_ni->seq_no; ntfs_read_locked_index_inode() 1532 err = ntfs_attr_lookup(AT_INDEX_ROOT, ni->name, ni->name_len, ntfs_read_locked_index_inode() 1579 ni->itype.index.collation_rule = ir->collation_rule; ntfs_read_locked_index_inode() 1582 ni->itype.index.block_size = le32_to_cpu(ir->index_block_size); ntfs_read_locked_index_inode() 1583 if (!is_power_of_2(ni->itype.index.block_size)) { ntfs_read_locked_index_inode() 1585 "two.", ni->itype.index.block_size); ntfs_read_locked_index_inode() 1588 if (ni->itype.index.block_size > PAGE_CACHE_SIZE) { ntfs_read_locked_index_inode() 1591 ni->itype.index.block_size, PAGE_CACHE_SIZE); ntfs_read_locked_index_inode() 1595 if (ni->itype.index.block_size < NTFS_BLOCK_SIZE) { ntfs_read_locked_index_inode() 1598 ni->itype.index.block_size, NTFS_BLOCK_SIZE); ntfs_read_locked_index_inode() 1602 ni->itype.index.block_size_bits = ffs(ni->itype.index.block_size) - 1; ntfs_read_locked_index_inode() 1604 if (vol->cluster_size <= ni->itype.index.block_size) { ntfs_read_locked_index_inode() 1605 ni->itype.index.vcn_size = vol->cluster_size; ntfs_read_locked_index_inode() 1606 ni->itype.index.vcn_size_bits = vol->cluster_size_bits; ntfs_read_locked_index_inode() 1608 ni->itype.index.vcn_size = vol->sector_size; ntfs_read_locked_index_inode() 1609 ni->itype.index.vcn_size_bits = vol->sector_size_bits; ntfs_read_locked_index_inode() 1614 vi->i_size = ni->initialized_size = ni->allocated_size = 0; ntfs_read_locked_index_inode() 1622 NInoSetIndexAllocPresent(ni); ntfs_read_locked_index_inode() 1625 err = ntfs_attr_lookup(AT_INDEX_ALLOCATION, ni->name, ni->name_len, ntfs_read_locked_index_inode() 1673 ni->initialized_size = sle64_to_cpu( ntfs_read_locked_index_inode() 1675 ni->allocated_size = sle64_to_cpu(a->data.non_resident.allocated_size); ntfs_read_locked_index_inode() 1685 bvi = ntfs_attr_iget(base_vi, AT_BITMAP, ni->name, ni->name_len); ntfs_read_locked_index_inode() 1700 if ((bvi_size << 3) < (vi->i_size >> ni->itype.index.block_size_bits)) { ntfs_read_locked_index_inode() 1710 vi->i_blocks = ni->allocated_size >> 9; ntfs_read_locked_index_inode() 1716 ni->ext.base_ntfs_ino = base_ni; ntfs_read_locked_index_inode() 1717 ni->nr_extents = -1; ntfs_read_locked_index_inode() 1733 ni->name_len); ntfs_read_locked_index_inode() 1782 ntfs_inode *ni; ntfs_read_inode_mount() local 1794 ni = NTFS_I(vi); ntfs_read_inode_mount() 1797 NInoSetNonResident(ni); ntfs_read_inode_mount() 1798 NInoSetMstProtected(ni); ntfs_read_inode_mount() 1799 NInoSetSparseDisabled(ni); ntfs_read_inode_mount() 1800 ni->type = AT_DATA; ntfs_read_inode_mount() 1801 ni->name = NULL; ntfs_read_inode_mount() 1802 ni->name_len = 0; ntfs_read_inode_mount() 1807 ni->itype.index.block_size = vol->mft_record_size; ntfs_read_inode_mount() 1808 ni->itype.index.block_size_bits = vol->mft_record_size_bits; ntfs_read_inode_mount() 1855 vi->i_generation = ni->seq_no = le16_to_cpu(m->sequence_number); ntfs_read_inode_mount() 1860 ctx = ntfs_attr_get_search_ctx(ni, m); ntfs_read_inode_mount() 1881 NInoSetAttrList(ni); ntfs_read_inode_mount() 1905 ni->attr_list_size = (u32)ntfs_attr_size(a); ntfs_read_inode_mount() 1906 ni->attr_list = ntfs_malloc_nofs(ni->attr_list_size); ntfs_read_inode_mount() 1907 if (!ni->attr_list) { ntfs_read_inode_mount() 1913 NInoSetAttrListNonResident(ni); ntfs_read_inode_mount() 1921 ni->attr_list_rl.rl = ntfs_mapping_pairs_decompress(vol, ntfs_read_inode_mount() 1923 if (IS_ERR(ni->attr_list_rl.rl)) { ntfs_read_inode_mount() 1924 err = PTR_ERR(ni->attr_list_rl.rl); ntfs_read_inode_mount() 1925 ni->attr_list_rl.rl = NULL; ntfs_read_inode_mount() 1932 if ((err = load_attribute_list(vol, &ni->attr_list_rl, ntfs_read_inode_mount() 1933 ni->attr_list, ni->attr_list_size, ntfs_read_inode_mount() 1952 memcpy(ni->attr_list, (u8*)a + le16_to_cpu( ntfs_read_inode_mount() 1969 al_entry = (ATTR_LIST_ENTRY*)ni->attr_list; ntfs_read_inode_mount() 1970 al_end = (u8*)al_entry + ni->attr_list_size; ntfs_read_inode_mount() 1973 if ((u8*)al_entry < ni->attr_list || ntfs_read_inode_mount() 2009 ni->seq_no) ntfs_read_inode_mount() 2052 nrl = ntfs_mapping_pairs_decompress(vol, a, ni->runlist.rl); ntfs_read_inode_mount() 2059 ni->runlist.rl = nrl; ntfs_read_inode_mount() 2077 ni->initialized_size = sle64_to_cpu( ntfs_read_inode_mount() 2079 ni->allocated_size = sle64_to_cpu( ntfs_read_inode_mount() 2180 lockdep_set_class(&ni->runlist.lock, &mft_ni_runlist_lock_key); ntfs_read_inode_mount() 2181 lockdep_set_class(&ni->mrec_lock, &mft_ni_mrec_lock_key); ntfs_read_inode_mount() 2197 static void __ntfs_clear_inode(ntfs_inode *ni) __ntfs_clear_inode() argument 2200 down_write(&ni->runlist.lock); __ntfs_clear_inode() 2201 if (ni->runlist.rl) { __ntfs_clear_inode() 2202 ntfs_free(ni->runlist.rl); __ntfs_clear_inode() 2203 ni->runlist.rl = NULL; __ntfs_clear_inode() 2205 up_write(&ni->runlist.lock); __ntfs_clear_inode() 2207 if (ni->attr_list) { __ntfs_clear_inode() 2208 ntfs_free(ni->attr_list); __ntfs_clear_inode() 2209 ni->attr_list = NULL; __ntfs_clear_inode() 2212 down_write(&ni->attr_list_rl.lock); __ntfs_clear_inode() 2213 if (ni->attr_list_rl.rl) { __ntfs_clear_inode() 2214 ntfs_free(ni->attr_list_rl.rl); __ntfs_clear_inode() 2215 ni->attr_list_rl.rl = NULL; __ntfs_clear_inode() 2217 up_write(&ni->attr_list_rl.lock); __ntfs_clear_inode() 2219 if (ni->name_len && ni->name != I30) { __ntfs_clear_inode() 2221 BUG_ON(!ni->name); __ntfs_clear_inode() 2222 kfree(ni->name); __ntfs_clear_inode() 2226 void ntfs_clear_extent_inode(ntfs_inode *ni) ntfs_clear_extent_inode() argument 2228 ntfs_debug("Entering for inode 0x%lx.", ni->mft_no); ntfs_clear_extent_inode() 2230 BUG_ON(NInoAttr(ni)); ntfs_clear_extent_inode() 2231 BUG_ON(ni->nr_extents != -1); ntfs_clear_extent_inode() 2234 if (NInoDirty(ni)) { ntfs_clear_extent_inode() 2235 if (!is_bad_inode(VFS_I(ni->ext.base_ntfs_ino))) ntfs_clear_extent_inode() 2236 ntfs_error(ni->vol->sb, "Clearing dirty extent inode! " ntfs_clear_extent_inode() 2242 __ntfs_clear_inode(ni); ntfs_clear_extent_inode() 2245 ntfs_destroy_extent_inode(ni); ntfs_clear_extent_inode() 2260 ntfs_inode *ni = NTFS_I(vi); ntfs_evict_big_inode() local 2266 if (NInoDirty(ni)) { ntfs_evict_big_inode() 2272 if (!was_bad && (is_bad_inode(vi) || NInoDirty(ni))) { ntfs_evict_big_inode() 2281 if (ni->nr_extents > 0) { ntfs_evict_big_inode() 2284 for (i = 0; i < ni->nr_extents; i++) ntfs_evict_big_inode() 2285 ntfs_clear_extent_inode(ni->ext.extent_ntfs_inos[i]); ntfs_evict_big_inode() 2286 kfree(ni->ext.extent_ntfs_inos); ntfs_evict_big_inode() 2289 __ntfs_clear_inode(ni); ntfs_evict_big_inode() 2291 if (NInoAttr(ni)) { ntfs_evict_big_inode() 2293 if (ni->nr_extents == -1) { ntfs_evict_big_inode() 2294 iput(VFS_I(ni->ext.base_ntfs_ino)); ntfs_evict_big_inode() 2295 ni->nr_extents = 0; ntfs_evict_big_inode() 2296 ni->ext.base_ntfs_ino = NULL; ntfs_evict_big_inode() 2368 ntfs_inode *base_ni, *ni = NTFS_I(vi); ntfs_truncate() local 2369 ntfs_volume *vol = ni->vol; ntfs_truncate() 2378 BUG_ON(NInoAttr(ni)); ntfs_truncate() 2380 BUG_ON(NInoMstProtected(ni)); ntfs_truncate() 2381 BUG_ON(ni->nr_extents < 0); ntfs_truncate() 2387 down_write(&ni->runlist.lock); ntfs_truncate() 2388 if (!NInoAttr(ni)) ntfs_truncate() 2389 base_ni = ni; ntfs_truncate() 2391 base_ni = ni->ext.base_ntfs_ino; ntfs_truncate() 2409 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_truncate() 2432 if (NInoNonResident(ni)) ntfs_truncate() 2438 read_lock_irqsave(&ni->size_lock, flags); ntfs_truncate() 2439 old_alloc_size = ni->allocated_size; ntfs_truncate() 2440 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_truncate() 2466 err = ntfs_attr_size_bounds_check(vol, ni->type, new_size); ntfs_truncate() 2476 le32_to_cpu(ni->type)); ntfs_truncate() 2483 le32_to_cpu(ni->type)); ntfs_truncate() 2491 if (NInoCompressed(ni) || NInoEncrypted(ni)) { ntfs_truncate() 2494 NInoCompressed(ni) ? "compressed" : ntfs_truncate() 2501 BUG_ON(NInoNonResident(ni)); ntfs_truncate() 2508 write_lock_irqsave(&ni->size_lock, flags); ntfs_truncate() 2510 ni->allocated_size = le32_to_cpu(a->length) - ntfs_truncate() 2540 ni->initialized_size = new_size; ntfs_truncate() 2541 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_truncate() 2557 up_write(&ni->runlist.lock); ntfs_truncate() 2562 err = ntfs_attr_make_non_resident(ni, old_size); ntfs_truncate() 2575 (unsigned)le32_to_cpu(ni->type), err); ntfs_truncate() 2601 if (ni->type == AT_ATTRIBUTE_LIST || ntfs_truncate() 2602 ni->type == AT_STANDARD_INFORMATION) { ntfs_truncate() 2620 BUG_ON(!NInoNonResident(ni)); ntfs_truncate() 2636 (unsigned)le32_to_cpu(ni->type)); ntfs_truncate() 2649 write_lock_irqsave(&ni->size_lock, flags); ntfs_truncate() 2650 if (new_size < ni->initialized_size) { ntfs_truncate() 2651 ni->initialized_size = new_size; ntfs_truncate() 2656 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_truncate() 2690 up_write(&ni->runlist.lock); ntfs_truncate() 2691 err = ntfs_attr_extend_allocation(ni, new_size, ntfs_truncate() 2704 nr_freed = ntfs_cluster_free(ni, new_alloc_size >> ntfs_truncate() 2716 err = ntfs_rl_truncate_nolock(vol, &ni->runlist, ntfs_truncate() 2734 mp_size = ntfs_get_size_for_mapping_pairs(vol, ni->runlist.rl, 0, -1); ntfs_truncate() 2740 (unsigned)le32_to_cpu(ni->type), mp_size, es); ntfs_truncate() 2758 mp_size, ni->runlist.rl, 0, -1, NULL); ntfs_truncate() 2763 vi->i_ino, (unsigned)le32_to_cpu(ni->type), ntfs_truncate() 2771 write_lock_irqsave(&ni->size_lock, flags); ntfs_truncate() 2772 ni->allocated_size = new_alloc_size; ntfs_truncate() 2774 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_truncate() 2776 ni->itype.compressed.size -= nr_freed << ntfs_truncate() 2778 BUG_ON(ni->itype.compressed.size < 0); ntfs_truncate() 2780 ni->itype.compressed.size); ntfs_truncate() 2781 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_truncate() 2785 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_truncate() 2807 up_write(&ni->runlist.lock); ntfs_truncate() 2830 NInoClearTruncateFailed(ni); ntfs_truncate() 2840 NInoSetTruncateFailed(ni); ntfs_truncate() 2848 up_write(&ni->runlist.lock); ntfs_truncate() 2856 NInoSetTruncateFailed(ni); ntfs_truncate() 2908 ntfs_inode *ni = NTFS_I(vi); ntfs_setattr() local 2913 if (NInoCompressed(ni) || NInoEncrypted(ni)) { ntfs_setattr() 2917 NInoCompressed(ni) ? ntfs_setattr() 2969 ntfs_inode *ni = NTFS_I(vi); __ntfs_write_inode() local 2976 ntfs_debug("Entering for %sinode 0x%lx.", NInoAttr(ni) ? "attr " : "", __ntfs_write_inode() 2983 if (NInoAttr(ni)) { __ntfs_write_inode() 2984 NInoClearDirty(ni); __ntfs_write_inode() 2989 m = map_mft_record(ni); __ntfs_write_inode() 2995 ctx = ntfs_attr_get_search_ctx(ni, m); __ntfs_write_inode() 3060 if (NInoDirty(ni)) __ntfs_write_inode() 3061 err = write_mft_record(ni, m, sync); __ntfs_write_inode() 3063 mutex_lock(&ni->extent_lock); __ntfs_write_inode() 3064 if (ni->nr_extents > 0) { __ntfs_write_inode() 3065 ntfs_inode **extent_nis = ni->ext.extent_ntfs_inos; __ntfs_write_inode() 3068 ntfs_debug("Writing %i extent inodes.", ni->nr_extents); __ntfs_write_inode() 3069 for (i = 0; i < ni->nr_extents; i++) { __ntfs_write_inode() 3090 mutex_unlock(&ni->extent_lock); __ntfs_write_inode() 3091 unmap_mft_record(ni); __ntfs_write_inode() 3097 unmap_mft_record(ni); __ntfs_write_inode() 3106 NVolSetErrors(ni->vol); __ntfs_write_inode()
|
H A D | aops.c | 64 ntfs_inode *ni; ntfs_end_buffer_async_read() local 69 ni = NTFS_I(vi); ntfs_end_buffer_async_read() 79 read_lock_irqsave(&ni->size_lock, flags); ntfs_end_buffer_async_read() 80 init_size = ni->initialized_size; ntfs_end_buffer_async_read() 82 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_end_buffer_async_read() 106 ntfs_error(ni->vol->sb, "Buffer I/O error, logical block " ntfs_end_buffer_async_read() 131 * attribute is mst protected, i.e. if NInoMstProteced(ni) is true. ntfs_end_buffer_async_read() 136 if (!NInoMstProtected(ni)) { ntfs_end_buffer_async_read() 144 rec_size = ni->itype.index.block_size; ntfs_end_buffer_async_read() 191 ntfs_inode *ni; ntfs_read_block() local 202 ni = NTFS_I(vi); ntfs_read_block() 203 vol = ni->vol; ntfs_read_block() 206 BUG_ON(!ni->runlist.rl && !ni->mft_no && !NInoAttr(ni)); ntfs_read_block() 233 read_lock_irqsave(&ni->size_lock, flags); ntfs_read_block() 234 lblock = (ni->allocated_size + blocksize - 1) >> blocksize_bits; ntfs_read_block() 235 init_size = ni->initialized_size; ntfs_read_block() 237 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_read_block() 268 down_read(&ni->runlist.lock); ntfs_read_block() 269 rl = ni->runlist.rl; ntfs_read_block() 302 up_read(&ni->runlist.lock); ntfs_read_block() 303 err = ntfs_map_runlist(ni, vcn); ntfs_read_block() 308 up_read(&ni->runlist.lock); ntfs_read_block() 327 "(error code %i).", ni->mft_no, ntfs_read_block() 328 ni->type, (unsigned long long)vcn, ntfs_read_block() 348 up_read(&ni->runlist.lock); ntfs_read_block() 402 ntfs_inode *ni, *base_ni; ntfs_readpage() local 429 ni = NTFS_I(vi); ntfs_readpage() 438 if (ni->type != AT_INDEX_ALLOCATION) { ntfs_readpage() 440 if (NInoEncrypted(ni)) { ntfs_readpage() 441 BUG_ON(ni->type != AT_DATA); ntfs_readpage() 446 if (NInoNonResident(ni) && NInoCompressed(ni)) { ntfs_readpage() 447 BUG_ON(ni->type != AT_DATA); ntfs_readpage() 448 BUG_ON(ni->name_len); ntfs_readpage() 453 if (NInoNonResident(ni)) { ntfs_readpage() 469 if (!NInoAttr(ni)) ntfs_readpage() 470 base_ni = ni; ntfs_readpage() 472 base_ni = ni->ext.base_ntfs_ino; ntfs_readpage() 483 if (unlikely(NInoNonResident(ni))) { ntfs_readpage() 492 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_readpage() 497 read_lock_irqsave(&ni->size_lock, flags); ntfs_readpage() 498 if (unlikely(attr_len > ni->initialized_size)) ntfs_readpage() 499 attr_len = ni->initialized_size; ntfs_readpage() 501 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_readpage() 558 ntfs_inode *ni; ntfs_write_block() local 569 ni = NTFS_I(vi); ntfs_write_block() 570 vol = ni->vol; ntfs_write_block() 573 "0x%lx.", ni->mft_no, ni->type, page->index); ntfs_write_block() 575 BUG_ON(!NInoNonResident(ni)); ntfs_write_block() 576 BUG_ON(NInoMstProtected(ni)); ntfs_write_block() 604 read_lock_irqsave(&ni->size_lock, flags); ntfs_write_block() 606 initialized_size = ni->initialized_size; ntfs_write_block() 607 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_write_block() 724 down_read(&ni->runlist.lock); ntfs_write_block() 725 rl = ni->runlist.rl; ntfs_write_block() 780 up_read(&ni->runlist.lock); ntfs_write_block() 781 err = ntfs_map_runlist(ni, vcn); ntfs_write_block() 786 up_read(&ni->runlist.lock); ntfs_write_block() 807 "determined%s (error code %i).", ni->mft_no, ntfs_write_block() 808 ni->type, (unsigned long long)vcn, ntfs_write_block() 816 up_read(&ni->runlist.lock); ntfs_write_block() 924 ntfs_inode *ni = NTFS_I(vi); ntfs_write_mst_block() local 925 ntfs_volume *vol = ni->vol; ntfs_write_mst_block() 927 unsigned int rec_size = ni->itype.index.block_size; ntfs_write_mst_block() 938 "0x%lx.", vi->i_ino, ni->type, page->index); ntfs_write_mst_block() 939 BUG_ON(!NInoNonResident(ni)); ntfs_write_mst_block() 940 BUG_ON(!NInoMstProtected(ni)); ntfs_write_mst_block() 949 (NInoAttr(ni) && ni->type == AT_INDEX_ALLOCATION))); ntfs_write_mst_block() 963 rec_size_bits = ni->itype.index.block_size_bits; ntfs_write_mst_block() 1030 down_read(&ni->runlist.lock); ntfs_write_mst_block() 1031 rl = ni->runlist.rl; ntfs_write_mst_block() 1060 up_read(&ni->runlist.lock); ntfs_write_mst_block() 1061 err2 = ntfs_map_runlist(ni, vcn); ntfs_write_mst_block() 1070 up_read(&ni->runlist.lock); ntfs_write_mst_block() 1085 ni->mft_no, ni->type, ntfs_write_mst_block() 1114 up_read(&ni->runlist.lock); ntfs_write_mst_block() 1175 ni->type, page->index, ofs); ntfs_write_mst_block() 1222 "and run chkdsk.", vi->i_ino, ni->type, ntfs_write_mst_block() 1303 if (ni->itype.index.block_size == PAGE_CACHE_SIZE) ntfs_write_mst_block() 1357 ntfs_inode *base_ni = NULL, *ni = NTFS_I(vi); ntfs_writepage() local 1387 if (ni->type != AT_INDEX_ALLOCATION) { ntfs_writepage() 1389 if (NInoEncrypted(ni)) { ntfs_writepage() 1391 BUG_ON(ni->type != AT_DATA); ntfs_writepage() 1396 if (NInoNonResident(ni) && NInoCompressed(ni)) { ntfs_writepage() 1397 BUG_ON(ni->type != AT_DATA); ntfs_writepage() 1398 BUG_ON(ni->name_len); ntfs_writepage() 1407 if (NInoNonResident(ni) && NInoSparse(ni)) { ntfs_writepage() 1415 if (NInoNonResident(ni)) { ntfs_writepage() 1423 if (NInoMstProtected(ni)) ntfs_writepage() 1447 if (!NInoAttr(ni)) ntfs_writepage() 1448 base_ni = ni; ntfs_writepage() 1450 base_ni = ni->ext.base_ntfs_ino; ntfs_writepage() 1463 if (unlikely(NInoNonResident(ni))) { ntfs_writepage() 1472 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_writepage() 1528 NVolSetErrors(ni->vol); ntfs_writepage() 1569 ntfs_inode *ni = NTFS_I(mapping->host); ntfs_bmap() local 1570 ntfs_volume *vol = ni->vol; ntfs_bmap() 1575 ni->mft_no, (unsigned long long)block); ntfs_bmap() 1576 if (ni->type != AT_DATA || !NInoNonResident(ni) || NInoEncrypted(ni)) { ntfs_bmap() 1579 (ni->type != AT_DATA) ? "non-data" : ntfs_bmap() 1580 (!NInoNonResident(ni) ? "resident" : ntfs_bmap() 1585 BUG_ON(NInoCompressed(ni)); ntfs_bmap() 1586 BUG_ON(NInoMstProtected(ni)); ntfs_bmap() 1590 read_lock_irqsave(&ni->size_lock, flags); ntfs_bmap() 1591 size = ni->initialized_size; ntfs_bmap() 1592 i_size = i_size_read(VFS_I(ni)); ntfs_bmap() 1593 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_bmap() 1602 down_read(&ni->runlist.lock); ntfs_bmap() 1603 lcn = ntfs_attr_vcn_to_lcn_nolock(ni, ofs >> cluster_size_shift, false); ntfs_bmap() 1604 up_read(&ni->runlist.lock); ntfs_bmap() 1625 "Returning 0.", ni->mft_no); ntfs_bmap() 1630 "Returning 0.", ni->mft_no); ntfs_bmap() 1731 ntfs_inode *ni = NTFS_I(mapping->host); mark_ntfs_record_dirty() local 1736 end = ofs + ni->itype.index.block_size; mark_ntfs_record_dirty() 1737 bh_size = VFS_I(ni)->i_sb->s_blocksize; mark_ntfs_record_dirty()
|
H A D | file.c | 75 * @ni: ntfs inode of the attribute to extend 78 * Extend the initialized size of an attribute described by the ntfs inode @ni 106 * Locking: i_mutex on the vfs inode corrseponsind to the ntfs inode @ni must be 109 static int ntfs_attr_extend_initialized(ntfs_inode *ni, const s64 new_init_size) ntfs_attr_extend_initialized() argument 115 struct inode *vi = VFS_I(ni); ntfs_attr_extend_initialized() 126 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized() 127 old_init_size = ni->initialized_size; ntfs_attr_extend_initialized() 129 BUG_ON(new_init_size > ni->allocated_size); ntfs_attr_extend_initialized() 130 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized() 134 vi->i_ino, (unsigned)le32_to_cpu(ni->type), ntfs_attr_extend_initialized() 137 if (!NInoAttr(ni)) ntfs_attr_extend_initialized() 138 base_ni = ni; ntfs_attr_extend_initialized() 140 base_ni = ni->ext.base_ntfs_ino; ntfs_attr_extend_initialized() 142 if (NInoNonResident(ni)) ntfs_attr_extend_initialized() 156 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_initialized() 177 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized() 179 ni->initialized_size = new_init_size; ntfs_attr_extend_initialized() 180 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized() 200 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_initialized() 244 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized() 245 ni->initialized_size = (s64)(index + 1) << PAGE_CACHE_SHIFT; ntfs_attr_extend_initialized() 246 if (ni->initialized_size > new_init_size) ntfs_attr_extend_initialized() 247 ni->initialized_size = new_init_size; ntfs_attr_extend_initialized() 248 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized() 282 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized() 283 BUG_ON(ni->initialized_size != new_init_size); ntfs_attr_extend_initialized() 284 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized() 297 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_initialized() 319 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized() 320 ni->initialized_size = old_init_size; ntfs_attr_extend_initialized() 321 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized() 340 ntfs_inode *base_ni, *ni = NTFS_I(vi); ntfs_prepare_file_for_write() local 341 ntfs_volume *vol = ni->vol; ntfs_prepare_file_for_write() 345 (unsigned)le32_to_cpu(ni->type), ntfs_prepare_file_for_write() 355 BUG_ON(NInoMstProtected(ni)); ntfs_prepare_file_for_write() 356 BUG_ON(ni->type != AT_DATA); ntfs_prepare_file_for_write() 358 if (NInoEncrypted(ni)) { ntfs_prepare_file_for_write() 368 if (NInoCompressed(ni)) { ntfs_prepare_file_for_write() 370 BUG_ON(ni->name_len); ntfs_prepare_file_for_write() 382 base_ni = ni; ntfs_prepare_file_for_write() 383 if (NInoAttr(ni)) ntfs_prepare_file_for_write() 384 base_ni = ni->ext.base_ntfs_ino; ntfs_prepare_file_for_write() 401 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_file_for_write() 402 ll = ni->allocated_size; ntfs_prepare_file_for_write() 403 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_file_for_write() 412 ll = ntfs_attr_extend_allocation(ni, end, -1, pos); ntfs_prepare_file_for_write() 422 le32_to_cpu(ni->type)); ntfs_prepare_file_for_write() 427 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_file_for_write() 428 ll = ni->allocated_size; ntfs_prepare_file_for_write() 429 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_file_for_write() 437 le32_to_cpu(ni->type), ntfs_prepare_file_for_write() 450 le32_to_cpu(ni->type), ntfs_prepare_file_for_write() 459 le32_to_cpu(ni->type)); ntfs_prepare_file_for_write() 471 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_file_for_write() 472 ll = ni->initialized_size; ntfs_prepare_file_for_write() 473 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_file_for_write() 480 err = ntfs_attr_extend_initialized(ni, pos); ntfs_prepare_file_for_write() 486 (unsigned)le32_to_cpu(ni->type), ntfs_prepare_file_for_write() 593 ntfs_inode *ni, *base_ni = NULL; ntfs_prepare_pages_for_non_resident_write() local 617 ni = NTFS_I(vi); ntfs_prepare_pages_for_non_resident_write() 618 vol = ni->vol; ntfs_prepare_pages_for_non_resident_write() 621 vi->i_ino, ni->type, pages[0]->index, nr_pages, ntfs_prepare_pages_for_non_resident_write() 696 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 697 initialized_size = ni->initialized_size; ntfs_prepare_pages_for_non_resident_write() 698 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 770 read_lock_irqsave(&ni->size_lock, ntfs_prepare_pages_for_non_resident_write() 772 initialized_size = ni->initialized_size; ntfs_prepare_pages_for_non_resident_write() 773 read_unlock_irqrestore(&ni->size_lock, ntfs_prepare_pages_for_non_resident_write() 830 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 831 initialized_size = ni->allocated_size; ntfs_prepare_pages_for_non_resident_write() 832 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 845 down_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 847 rl = ni->runlist.rl; ntfs_prepare_pages_for_non_resident_write() 874 up_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 877 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 899 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 900 down_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 904 err = ntfs_map_runlist_nolock(ni, bh_cpos, ntfs_prepare_pages_for_non_resident_write() 929 ni->mft_no, ni->type, ntfs_prepare_pages_for_non_resident_write() 981 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 982 down_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 990 while (--rl2 >= ni->runlist.rl) { ntfs_prepare_pages_for_non_resident_write() 1005 rl = ntfs_runlists_merge(ni->runlist.rl, rl2); ntfs_prepare_pages_for_non_resident_write() 1020 ni->runlist.rl = rl; ntfs_prepare_pages_for_non_resident_write() 1025 if (!NInoAttr(ni)) ntfs_prepare_pages_for_non_resident_write() 1026 base_ni = ni; ntfs_prepare_pages_for_non_resident_write() 1028 base_ni = ni->ext.base_ntfs_ino; ntfs_prepare_pages_for_non_resident_write() 1041 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_prepare_pages_for_non_resident_write() 1125 (unsigned)le32_to_cpu(ni->type), err); ntfs_prepare_pages_for_non_resident_write() 1137 if (likely(NInoSparse(ni) || NInoCompressed(ni))) { ntfs_prepare_pages_for_non_resident_write() 1147 err = ntfs_attr_lookup(ni->type, ni->name, ntfs_prepare_pages_for_non_resident_write() 1148 ni->name_len, CASE_SENSITIVE, ntfs_prepare_pages_for_non_resident_write() 1157 write_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 1158 ni->itype.compressed.size += vol->cluster_size; ntfs_prepare_pages_for_non_resident_write() 1160 cpu_to_sle64(ni->itype.compressed.size); ntfs_prepare_pages_for_non_resident_write() 1161 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 1184 up_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 1196 up_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 1199 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 1203 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 1204 initialized_size = ni->initialized_size; ntfs_prepare_pages_for_non_resident_write() 1205 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 1244 if (ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_prepare_pages_for_non_resident_write() 1250 write_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 1251 ni->itype.compressed.size += vol->cluster_size; ntfs_prepare_pages_for_non_resident_write() 1252 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write() 1277 if (ntfs_rl_punch_nolock(vol, &ni->runlist, bh_cpos, 1)) { ntfs_prepare_pages_for_non_resident_write() 1318 mapping_pairs_offset), ni->runlist.rl, ntfs_prepare_pages_for_non_resident_write() 1337 up_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 1339 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write() 1405 ntfs_inode *ni, *base_ni; ntfs_commit_pages_after_non_resident_write() local 1415 ni = NTFS_I(vi); ntfs_commit_pages_after_non_resident_write() 1451 read_lock_irqsave(&ni->size_lock, flags); ntfs_commit_pages_after_non_resident_write() 1452 initialized_size = ni->initialized_size; ntfs_commit_pages_after_non_resident_write() 1453 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_commit_pages_after_non_resident_write() 1462 if (!NInoAttr(ni)) ntfs_commit_pages_after_non_resident_write() 1463 base_ni = ni; ntfs_commit_pages_after_non_resident_write() 1465 base_ni = ni->ext.base_ntfs_ino; ntfs_commit_pages_after_non_resident_write() 1474 BUG_ON(!NInoNonResident(ni)); ntfs_commit_pages_after_non_resident_write() 1480 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_commit_pages_after_non_resident_write() 1489 write_lock_irqsave(&ni->size_lock, flags); ntfs_commit_pages_after_non_resident_write() 1490 BUG_ON(end > ni->allocated_size); ntfs_commit_pages_after_non_resident_write() 1491 ni->initialized_size = end; ntfs_commit_pages_after_non_resident_write() 1498 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_commit_pages_after_non_resident_write() 1514 NVolSetErrors(ni->vol); ntfs_commit_pages_after_non_resident_write() 1560 ntfs_inode *ni, *base_ni; ntfs_commit_pages_after_write() local 1575 ni = NTFS_I(vi); ntfs_commit_pages_after_write() 1578 vi->i_ino, ni->type, page->index, nr_pages, ntfs_commit_pages_after_write() 1580 if (NInoNonResident(ni)) ntfs_commit_pages_after_write() 1588 if (!NInoAttr(ni)) ntfs_commit_pages_after_write() 1589 base_ni = ni; ntfs_commit_pages_after_write() 1591 base_ni = ni->ext.base_ntfs_ino; ntfs_commit_pages_after_write() 1592 BUG_ON(NInoNonResident(ni)); ntfs_commit_pages_after_write() 1606 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_commit_pages_after_write() 1648 read_lock_irqsave(&ni->size_lock, flags); ntfs_commit_pages_after_write() 1649 initialized_size = ni->initialized_size; ntfs_commit_pages_after_write() 1650 BUG_ON(end > ni->allocated_size); ntfs_commit_pages_after_write() 1651 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_commit_pages_after_write() 1654 write_lock_irqsave(&ni->size_lock, flags); ntfs_commit_pages_after_write() 1655 ni->initialized_size = end; ntfs_commit_pages_after_write() 1657 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_commit_pages_after_write() 1686 NVolSetErrors(ni->vol); ntfs_commit_pages_after_write() 1750 ntfs_inode *ni = NTFS_I(vi); ntfs_perform_write() local 1751 ntfs_volume *vol = ni->vol; ntfs_perform_write() 1762 (unsigned)le32_to_cpu(ni->type), ntfs_perform_write() 1769 if (unlikely(NInoTruncateFailed(ni))) { ntfs_perform_write() 1774 if (err || NInoTruncateFailed(ni)) { ntfs_perform_write() 1781 (unsigned)le32_to_cpu(ni->type), err); ntfs_perform_write() 1790 if (vol->cluster_size > PAGE_CACHE_SIZE && NInoNonResident(ni)) ntfs_perform_write() 1812 down_read(&ni->runlist.lock); ntfs_perform_write() 1813 lcn = ntfs_attr_vcn_to_lcn_nolock(ni, pos >> ntfs_perform_write() 1815 up_read(&ni->runlist.lock); ntfs_perform_write() 1828 le32_to_cpu(ni->type)); ntfs_perform_write() 1868 if (NInoNonResident(ni)) { ntfs_perform_write()
|
H A D | attrib.c | 40 * @ni: ntfs inode for which to map (part of) a runlist 44 * Map the part of a runlist containing the @vcn of the ntfs inode @ni. 46 * If @ctx is specified, it is an active search context of @ni and its base mft 77 * Locking: - The runlist described by @ni must be locked for writing on entry 79 * - If @ctx is NULL, the base mft record of @ni must not be mapped on 84 int ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn, ntfs_attr_search_ctx *ctx) ntfs_map_runlist_nolock() argument 99 if (!NInoAttr(ni)) ntfs_map_runlist_nolock() 100 base_ni = ni; ntfs_map_runlist_nolock() 102 base_ni = ni->ext.base_ntfs_ino; ntfs_map_runlist_nolock() 121 read_lock_irqsave(&ni->size_lock, flags); ntfs_map_runlist_nolock() 122 allocated_size_vcn = ni->allocated_size >> ntfs_map_runlist_nolock() 123 ni->vol->cluster_size_bits; ntfs_map_runlist_nolock() 124 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_map_runlist_nolock() 134 if (vcn >= allocated_size_vcn || (a->type == ni->type && ntfs_map_runlist_nolock() 135 a->name_length == ni->name_len && ntfs_map_runlist_nolock() 137 ni->name, ni->name_len) && ntfs_map_runlist_nolock() 166 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_map_runlist_nolock() 187 rl = ntfs_mapping_pairs_decompress(ni->vol, a, ni->runlist.rl); ntfs_map_runlist_nolock() 191 ni->runlist.rl = rl; ntfs_map_runlist_nolock() 285 * @ni: ntfs inode for which to map (part of) a runlist 288 * Map the part of a runlist containing the @vcn of the ntfs inode @ni. 298 int ntfs_map_runlist(ntfs_inode *ni, VCN vcn) ntfs_map_runlist() argument 302 down_write(&ni->runlist.lock); ntfs_map_runlist() 304 if (likely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) <= ntfs_map_runlist() 306 err = ntfs_map_runlist_nolock(ni, vcn, NULL); ntfs_map_runlist() 307 up_write(&ni->runlist.lock); ntfs_map_runlist() 313 * @ni: ntfs inode of the attribute whose runlist to search 318 * described by the ntfs inode @ni and return the corresponding logical cluster 341 LCN ntfs_attr_vcn_to_lcn_nolock(ntfs_inode *ni, const VCN vcn, ntfs_attr_vcn_to_lcn_nolock() argument 348 BUG_ON(!ni); ntfs_attr_vcn_to_lcn_nolock() 350 ni->mft_no, (unsigned long long)vcn, ntfs_attr_vcn_to_lcn_nolock() 352 BUG_ON(!NInoNonResident(ni)); ntfs_attr_vcn_to_lcn_nolock() 354 if (!ni->runlist.rl) { ntfs_attr_vcn_to_lcn_nolock() 355 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_vcn_to_lcn_nolock() 356 if (!ni->allocated_size) { ntfs_attr_vcn_to_lcn_nolock() 357 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_vcn_to_lcn_nolock() 360 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_vcn_to_lcn_nolock() 364 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); ntfs_attr_vcn_to_lcn_nolock() 376 up_read(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock() 377 down_write(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock() 378 if (unlikely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) != ntfs_attr_vcn_to_lcn_nolock() 380 up_write(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock() 381 down_read(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock() 385 err = ntfs_map_runlist_nolock(ni, vcn, NULL); ntfs_attr_vcn_to_lcn_nolock() 387 up_write(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock() 388 down_read(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock() 402 ntfs_error(ni->vol->sb, "Failed with error code %lli.", ntfs_attr_vcn_to_lcn_nolock() 409 * @ni: ntfs inode describing the runlist to search 414 * inode @ni and return the address of the runlist element containing the @vcn. 419 * If @ctx is specified, it is an active search context of @ni and its base mft 456 * Locking: - The runlist described by @ni must be locked for writing on entry 459 * - If @ctx is NULL, the base mft record of @ni must not be mapped on 464 runlist_element *ntfs_attr_find_vcn_nolock(ntfs_inode *ni, const VCN vcn, ntfs_attr_find_vcn_nolock() argument 472 BUG_ON(!ni); ntfs_attr_find_vcn_nolock() 474 ni->mft_no, (unsigned long long)vcn, ctx ? "" : "out"); ntfs_attr_find_vcn_nolock() 475 BUG_ON(!NInoNonResident(ni)); ntfs_attr_find_vcn_nolock() 477 if (!ni->runlist.rl) { ntfs_attr_find_vcn_nolock() 478 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_find_vcn_nolock() 479 if (!ni->allocated_size) { ntfs_attr_find_vcn_nolock() 480 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_find_vcn_nolock() 483 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_find_vcn_nolock() 486 rl = ni->runlist.rl; ntfs_attr_find_vcn_nolock() 517 err = ntfs_map_runlist_nolock(ni, vcn, ctx); ntfs_attr_find_vcn_nolock() 528 ntfs_error(ni->vol->sb, "Failed with error code %i.", err); ntfs_attr_find_vcn_nolock() 862 ntfs_inode *base_ni, *ni; ntfs_external_attr_find() local 872 ni = ctx->ntfs_ino; ntfs_external_attr_find() 874 ntfs_debug("Entering for inode 0x%lx, type 0x%x.", ni->mft_no, type); ntfs_external_attr_find() 880 if (ni == base_ni) ntfs_external_attr_find() 981 if (MREF_LE(al_entry->mft_reference) == ni->mft_no) { ntfs_external_attr_find() 982 if (MSEQNO_LE(al_entry->mft_reference) != ni->seq_no) { ntfs_external_attr_find() 992 if (ni != base_ni) ntfs_external_attr_find() 993 unmap_extent_mft_record(ni); ntfs_external_attr_find() 997 ni = ctx->ntfs_ino = base_ni; ntfs_external_attr_find() 1003 al_entry->mft_reference), &ni); ntfs_external_attr_find() 1016 ni = NULL; ntfs_external_attr_find() 1019 ctx->ntfs_ino = ni; ntfs_external_attr_find() 1090 if (ni != base_ni) { ntfs_external_attr_find() 1091 if (ni) ntfs_external_attr_find() 1092 unmap_extent_mft_record(ni); ntfs_external_attr_find() 1123 if (ni != base_ni) ntfs_external_attr_find() 1124 unmap_extent_mft_record(ni); ntfs_external_attr_find() 1212 * @ni: ntfs inode with which to initialize the search context 1215 * Initialize the attribute search context @ctx with @ni and @mrec. 1218 ntfs_inode *ni, MFT_RECORD *mrec) ntfs_attr_init_search_ctx() 1226 .ntfs_ino = ni, ntfs_attr_init_search_ctx() 1263 * @ni: ntfs inode with which to initialize the search context 1266 * Allocate a new attribute search context, initialize it with @ni and @mrec, 1269 ntfs_attr_search_ctx *ntfs_attr_get_search_ctx(ntfs_inode *ni, MFT_RECORD *mrec) ntfs_attr_get_search_ctx() argument 1275 ntfs_attr_init_search_ctx(ctx, ni, mrec); ntfs_attr_get_search_ctx() 1503 * @ni: ntfs inode describing the attribute to convert 1506 * Convert the resident ntfs attribute described by the ntfs inode @ni to a 1535 int ntfs_attr_make_non_resident(ntfs_inode *ni, const u32 data_size) ntfs_attr_make_non_resident() argument 1538 struct inode *vi = VFS_I(ni); ntfs_attr_make_non_resident() 1539 ntfs_volume *vol = ni->vol; ntfs_attr_make_non_resident() 1553 err = ntfs_attr_can_be_non_resident(vol, ni->type); ntfs_attr_make_non_resident() 1567 BUG_ON(NInoCompressed(ni)); ntfs_attr_make_non_resident() 1568 BUG_ON(NInoEncrypted(ni)); ntfs_attr_make_non_resident() 1607 down_write(&ni->runlist.lock); ntfs_attr_make_non_resident() 1608 if (!NInoAttr(ni)) ntfs_attr_make_non_resident() 1609 base_ni = ni; ntfs_attr_make_non_resident() 1611 base_ni = ni->ext.base_ntfs_ino; ntfs_attr_make_non_resident() 1624 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_make_non_resident() 1633 BUG_ON(NInoNonResident(ni)); ntfs_attr_make_non_resident() 1638 if (NInoSparse(ni) || NInoCompressed(ni)) ntfs_attr_make_non_resident() 1695 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_attr_make_non_resident() 1697 if (NInoCompressed(ni) || vol->major_ver < 3) ntfs_attr_make_non_resident() 1712 ni->runlist.rl = rl; ntfs_attr_make_non_resident() 1713 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_make_non_resident() 1714 ni->allocated_size = new_size; ntfs_attr_make_non_resident() 1715 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_attr_make_non_resident() 1716 ni->itype.compressed.size = ni->allocated_size; ntfs_attr_make_non_resident() 1718 ni->itype.compressed.block_size = 1U << (a->data. ntfs_attr_make_non_resident() 1721 ni->itype.compressed.block_size_bits = ntfs_attr_make_non_resident() 1722 ffs(ni->itype.compressed.block_size) - ntfs_attr_make_non_resident() 1724 ni->itype.compressed.block_clusters = 1U << ntfs_attr_make_non_resident() 1727 ni->itype.compressed.block_size = 0; ntfs_attr_make_non_resident() 1728 ni->itype.compressed.block_size_bits = 0; ntfs_attr_make_non_resident() 1729 ni->itype.compressed.block_clusters = 0; ntfs_attr_make_non_resident() 1731 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_attr_make_non_resident() 1733 vi->i_blocks = ni->allocated_size >> 9; ntfs_attr_make_non_resident() 1734 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_make_non_resident() 1741 NInoSetNonResident(ni); ntfs_attr_make_non_resident() 1747 up_write(&ni->runlist.lock); ntfs_attr_make_non_resident() 1790 (unsigned)le32_to_cpu(ni->type), ntfs_attr_make_non_resident() 1792 ((ni->type == AT_DATA) && ntfs_attr_make_non_resident() 1793 !ni->name_len) ? "FILE": "ATTRIBUTE"); ntfs_attr_make_non_resident() 1794 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_make_non_resident() 1795 ni->initialized_size = attr_size; ntfs_attr_make_non_resident() 1797 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_make_non_resident() 1813 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_make_non_resident() 1814 ni->allocated_size = arec_size - mp_ofs; ntfs_attr_make_non_resident() 1815 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_make_non_resident() 1824 ni->runlist.rl = NULL; ntfs_attr_make_non_resident() 1825 up_write(&ni->runlist.lock); ntfs_attr_make_non_resident() 1847 * @ni: ntfs inode of the attribute whose allocation to extend 1852 * Extend the allocated space of an attribute described by the ntfs inode @ni 1855 * inode @ni have sparse support enabled). If @data_start is >= 0, then the 1893 * Locking: This function takes the runlist lock of @ni for writing as well as 1904 s64 ntfs_attr_extend_allocation(ntfs_inode *ni, s64 new_alloc_size, ntfs_attr_extend_allocation() argument 1909 struct inode *vi = VFS_I(ni); ntfs_attr_extend_allocation() 1910 ntfs_volume *vol = ni->vol; ntfs_attr_extend_allocation() 1922 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation() 1923 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation() 1924 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation() 1929 (unsigned)le32_to_cpu(ni->type), ntfs_attr_extend_allocation() 1940 if (NInoNonResident(ni)) { ntfs_attr_extend_allocation() 1948 err = ntfs_attr_size_bounds_check(vol, ni->type, new_alloc_size); ntfs_attr_extend_allocation() 1951 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation() 1952 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation() 1953 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation() 1963 le32_to_cpu(ni->type)); ntfs_attr_extend_allocation() 1973 le32_to_cpu(ni->type)); ntfs_attr_extend_allocation() 1983 if (!NInoAttr(ni)) ntfs_attr_extend_allocation() 1984 base_ni = ni; ntfs_attr_extend_allocation() 1986 base_ni = ni->ext.base_ntfs_ino; ntfs_attr_extend_allocation() 1991 down_write(&ni->runlist.lock); ntfs_attr_extend_allocation() 2004 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2005 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation() 2006 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2011 vcn = NInoNonResident(ni) ? allocated_size >> vol->cluster_size_bits : ntfs_attr_extend_allocation() 2030 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_allocation() 2042 BUG_ON(NInoNonResident(ni)); ntfs_attr_extend_allocation() 2055 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2056 ni->allocated_size = le32_to_cpu(a->length) - ntfs_attr_extend_allocation() 2058 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2076 up_write(&ni->runlist.lock); ntfs_attr_extend_allocation() 2081 err = ntfs_attr_make_non_resident(ni, attr_len); ntfs_attr_extend_allocation() 2091 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2092 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation() 2093 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2100 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation() 2106 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2107 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation() 2108 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2132 if (ni->type == AT_ATTRIBUTE_LIST || ntfs_attr_extend_allocation() 2133 ni->type == AT_STANDARD_INFORMATION) { ntfs_attr_extend_allocation() 2151 BUG_ON(!NInoNonResident(ni)); ntfs_attr_extend_allocation() 2164 if ((start >= 0 && start <= allocated_size) || ni->type != AT_DATA || ntfs_attr_extend_allocation() 2165 !NVolSparseEnabled(vol) || NInoSparseDisabled(ni)) ntfs_attr_extend_allocation() 2172 rl = ni->runlist.rl; ntfs_attr_extend_allocation() 2180 (rl->lcn == LCN_ENOENT && rl > ni->runlist.rl && ntfs_attr_extend_allocation() 2184 rl = ntfs_mapping_pairs_decompress(vol, a, ni->runlist.rl); ntfs_attr_extend_allocation() 2194 (unsigned)le32_to_cpu(ni->type), ntfs_attr_extend_allocation() 2200 ni->runlist.rl = rl; ntfs_attr_extend_allocation() 2213 while (rl->lcn < 0 && rl > ni->runlist.rl) ntfs_attr_extend_allocation() 2230 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation() 2235 rl = ntfs_runlists_merge(ni->runlist.rl, rl2); ntfs_attr_extend_allocation() 2243 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation() 2256 ni->runlist.rl = rl; ntfs_attr_extend_allocation() 2276 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation() 2311 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation() 2330 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_allocation() 2337 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2338 ni->allocated_size = new_alloc_size; ntfs_attr_extend_allocation() 2341 * FIXME: This would fail if @ni is a directory, $MFT, or an index, ntfs_attr_extend_allocation() 2350 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_attr_extend_allocation() 2351 ni->itype.compressed.size += new_alloc_size - allocated_size; ntfs_attr_extend_allocation() 2353 cpu_to_sle64(ni->itype.compressed.size); ntfs_attr_extend_allocation() 2354 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_attr_extend_allocation() 2357 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2371 up_write(&ni->runlist.lock); ntfs_attr_extend_allocation() 2381 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation() 2385 if (ntfs_attr_lookup(ni->type, ni->name, ni->name_len, CASE_SENSITIVE, ntfs_attr_extend_allocation() 2391 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2392 ni->allocated_size = new_alloc_size; ntfs_attr_extend_allocation() 2394 * FIXME: This would fail if @ni is a directory... See above. ntfs_attr_extend_allocation() 2398 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_attr_extend_allocation() 2399 ni->itype.compressed.size += new_alloc_size - ntfs_attr_extend_allocation() 2401 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_attr_extend_allocation() 2404 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation() 2407 up_write(&ni->runlist.lock); ntfs_attr_extend_allocation() 2419 if (ntfs_cluster_free(ni, ll, -1, ctx) < 0) { ntfs_attr_extend_allocation() 2433 if (ntfs_rl_truncate_nolock(vol, &ni->runlist, ll) || IS_ERR(m)) { ntfs_attr_extend_allocation() 2467 up_write(&ni->runlist.lock); ntfs_attr_extend_allocation() 2475 * @ni: ntfs inode describing the attribute to fill 2480 * Fill @cnt bytes of the attribute described by the ntfs inode @ni starting at 2493 int ntfs_attr_set(ntfs_inode *ni, const s64 ofs, const s64 cnt, const u8 val) ntfs_attr_set() argument 2495 ntfs_volume *vol = ni->vol; ntfs_attr_set() 2512 BUG_ON(NInoCompressed(ni)); ntfs_attr_set() 2513 BUG_ON(NInoEncrypted(ni)); ntfs_attr_set() 2514 mapping = VFS_I(ni)->i_mapping; ntfs_attr_set() 2522 if (unlikely(end > i_size_read(VFS_I(ni)))) { ntfs_attr_set() 1217 ntfs_attr_init_search_ctx(ntfs_attr_search_ctx *ctx, ntfs_inode *ni, MFT_RECORD *mrec) ntfs_attr_init_search_ctx() argument
|
H A D | attrib.h | 63 extern int ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn, 65 extern int ntfs_map_runlist(ntfs_inode *ni, VCN vcn); 67 extern LCN ntfs_attr_vcn_to_lcn_nolock(ntfs_inode *ni, const VCN vcn, 70 extern runlist_element *ntfs_attr_find_vcn_nolock(ntfs_inode *ni, 89 extern ntfs_attr_search_ctx *ntfs_attr_get_search_ctx(ntfs_inode *ni, 106 extern int ntfs_attr_make_non_resident(ntfs_inode *ni, const u32 data_size); 108 extern s64 ntfs_attr_extend_allocation(ntfs_inode *ni, s64 new_alloc_size, 111 extern int ntfs_attr_set(ntfs_inode *ni, const s64 ofs, const s64 cnt,
|
H A D | inode.h | 183 static inline int NIno##flag(ntfs_inode *ni) \ 185 return test_bit(NI_##flag, &(ni)->state); \ 187 static inline void NInoSet##flag(ntfs_inode *ni) \ 189 set_bit(NI_##flag, &(ni)->state); \ 191 static inline void NInoClear##flag(ntfs_inode *ni) \ 193 clear_bit(NI_##flag, &(ni)->state); \ 200 static inline int NInoTestSet##flag(ntfs_inode *ni) \ 202 return test_and_set_bit(NI_##flag, &(ni)->state); \ 204 static inline int NInoTestClear##flag(ntfs_inode *ni) \ 206 return test_and_clear_bit(NI_##flag, &(ni)->state); \ 245 static inline struct inode *VFS_I(ntfs_inode *ni) VFS_I() argument 247 return &((big_ntfs_inode *)ni)->vfs_inode; VFS_I() 284 extern void __ntfs_init_inode(struct super_block *sb, ntfs_inode *ni); 288 ntfs_inode *ni = NTFS_I(vi); ntfs_init_big_inode() local 291 __ntfs_init_inode(vi->i_sb, ni); ntfs_init_big_inode() 292 ni->mft_no = vi->i_ino; ntfs_init_big_inode() 297 extern void ntfs_clear_extent_inode(ntfs_inode *ni);
|
H A D | mft.c | 39 * @ni: ntfs inode whose mft record page to map 41 * This maps the page in which the mft record of the ntfs inode @ni is situated 47 static inline MFT_RECORD *map_mft_record_page(ntfs_inode *ni) map_mft_record_page() argument 50 ntfs_volume *vol = ni->vol; map_mft_record_page() 56 BUG_ON(ni->page); map_mft_record_page() 63 index = (u64)ni->mft_no << vol->mft_record_size_bits >> map_mft_record_page() 65 ofs = (ni->mft_no << vol->mft_record_size_bits) & ~PAGE_CACHE_MASK; map_mft_record_page() 79 "driver.", ni->mft_no); map_mft_record_page() 89 ni->page = page; map_mft_record_page() 90 ni->page_ofs = ofs; map_mft_record_page() 94 "Run chkdsk.", ni->mft_no); map_mft_record_page() 100 ni->page = NULL; map_mft_record_page() 101 ni->page_ofs = 0; map_mft_record_page() 107 * @ni: ntfs inode whose MFT record to map 155 MFT_RECORD *map_mft_record(ntfs_inode *ni) map_mft_record() argument 159 ntfs_debug("Entering for mft_no 0x%lx.", ni->mft_no); map_mft_record() 162 atomic_inc(&ni->count); map_mft_record() 165 mutex_lock(&ni->mrec_lock); map_mft_record() 167 m = map_mft_record_page(ni); map_mft_record() 171 mutex_unlock(&ni->mrec_lock); map_mft_record() 172 atomic_dec(&ni->count); map_mft_record() 173 ntfs_error(ni->vol->sb, "Failed with error code %lu.", -PTR_ERR(m)); map_mft_record() 179 * @ni: ntfs inode whose mft record page to unmap 181 * This unmaps the page in which the mft record of the ntfs inode @ni is 191 static inline void unmap_mft_record_page(ntfs_inode *ni) unmap_mft_record_page() argument 193 BUG_ON(!ni->page); unmap_mft_record_page() 196 ntfs_unmap_page(ni->page); unmap_mft_record_page() 197 ni->page = NULL; unmap_mft_record_page() 198 ni->page_ofs = 0; unmap_mft_record_page() 204 * @ni: ntfs inode whose MFT record to unmap 213 void unmap_mft_record(ntfs_inode *ni) unmap_mft_record() argument 215 struct page *page = ni->page; unmap_mft_record() 219 ntfs_debug("Entering for mft_no 0x%lx.", ni->mft_no); unmap_mft_record() 221 unmap_mft_record_page(ni); unmap_mft_record() 222 mutex_unlock(&ni->mrec_lock); unmap_mft_record() 223 atomic_dec(&ni->count); unmap_mft_record() 250 ntfs_inode *ni = NULL; map_extent_mft_record() local 272 ni = extent_nis[i]; map_extent_mft_record() 274 atomic_inc(&ni->count); map_extent_mft_record() 278 if (likely(ni != NULL)) { map_extent_mft_record() 282 m = map_mft_record(ni); map_extent_mft_record() 284 atomic_dec(&ni->count); map_extent_mft_record() 289 *ntfs_ino = ni; map_extent_mft_record() 292 unmap_mft_record(ni); map_extent_mft_record() 304 ni = ntfs_new_extent_inode(base_ni->vol->sb, mft_no); map_extent_mft_record() 305 if (unlikely(!ni)) { map_extent_mft_record() 310 ni->vol = base_ni->vol; map_extent_mft_record() 311 ni->seq_no = seq_no; map_extent_mft_record() 312 ni->nr_extents = -1; map_extent_mft_record() 313 ni->ext.base_ntfs_ino = base_ni; map_extent_mft_record() 315 m = map_mft_record(ni); map_extent_mft_record() 319 ntfs_clear_extent_inode(ni); map_extent_mft_record() 351 base_ni->ext.extent_ntfs_inos[base_ni->nr_extents++] = ni; map_extent_mft_record() 355 *ntfs_ino = ni; map_extent_mft_record() 358 unmap_mft_record(ni); map_extent_mft_record() 366 ntfs_clear_extent_inode(ni); map_extent_mft_record() 374 * @ni: ntfs inode describing the mapped mft record 378 * Set the mapped (extent) mft record of the (base or extent) ntfs inode @ni, 395 void __mark_mft_record_dirty(ntfs_inode *ni) __mark_mft_record_dirty() argument 399 ntfs_debug("Entering for inode 0x%lx.", ni->mft_no); __mark_mft_record_dirty() 400 BUG_ON(NInoAttr(ni)); __mark_mft_record_dirty() 401 mark_ntfs_record_dirty(ni->page, ni->page_ofs); __mark_mft_record_dirty() 403 mutex_lock(&ni->extent_lock); __mark_mft_record_dirty() 404 if (likely(ni->nr_extents >= 0)) __mark_mft_record_dirty() 405 base_ni = ni; __mark_mft_record_dirty() 407 base_ni = ni->ext.base_ntfs_ino; __mark_mft_record_dirty() 408 mutex_unlock(&ni->extent_lock); __mark_mft_record_dirty() 640 * @ni: ntfs inode describing the mapped (extent) mft record 645 * ntfs inode @ni to backing store. If the mft record @m has a counterpart in 648 * We only write the mft record if the ntfs inode @ni is dirty and the first 669 int write_mft_record_nolock(ntfs_inode *ni, MFT_RECORD *m, int sync) write_mft_record_nolock() argument 671 ntfs_volume *vol = ni->vol; write_mft_record_nolock() 672 struct page *page = ni->page; write_mft_record_nolock() 682 ntfs_debug("Entering for inode 0x%lx.", ni->mft_no); write_mft_record_nolock() 683 BUG_ON(NInoAttr(ni)); write_mft_record_nolock() 692 if (!NInoTestClearDirty(ni)) write_mft_record_nolock() 699 m_start = ni->page_ofs; write_mft_record_nolock() 729 vcn = ((VCN)ni->mft_no << vol->mft_record_size_bits) + write_mft_record_nolock() 755 ni->mft_no, (long long)lcn); write_mft_record_nolock() 777 flush_dcache_mft_record_page(ni); write_mft_record_nolock() 791 if (!sync && ni->mft_no < vol->mftmirr_size) write_mft_record_nolock() 792 ntfs_sync_mft_mirror(vol, ni->mft_no, m, sync); write_mft_record_nolock() 809 if (sync && ni->mft_no < vol->mftmirr_size) write_mft_record_nolock() 810 ntfs_sync_mft_mirror(vol, ni->mft_no, m, sync); write_mft_record_nolock() 813 flush_dcache_mft_record_page(ni); write_mft_record_nolock() 819 ni->mft_no); write_mft_record_nolock() 839 mark_mft_record_dirty(ni); write_mft_record_nolock() 931 ntfs_inode *ni, *eni, **extent_nis; ntfs_may_write_mft_record() local 971 ni = NTFS_I(vi); ntfs_may_write_mft_record() 973 atomic_inc(&ni->count); ntfs_may_write_mft_record() 975 if (NInoDirty(ni)) { ntfs_may_write_mft_record() 978 atomic_dec(&ni->count); ntfs_may_write_mft_record() 984 if (unlikely(!mutex_trylock(&ni->mrec_lock))) { ntfs_may_write_mft_record() 987 atomic_dec(&ni->count); ntfs_may_write_mft_record() 997 *locked_ni = ni; ntfs_may_write_mft_record() 1043 ni = NTFS_I(vi); ntfs_may_write_mft_record() 1044 mutex_lock(&ni->extent_lock); ntfs_may_write_mft_record() 1045 if (ni->nr_extents <= 0) { ntfs_may_write_mft_record() 1050 mutex_unlock(&ni->extent_lock); ntfs_may_write_mft_record() 1057 extent_nis = ni->ext.extent_ntfs_inos; ntfs_may_write_mft_record() 1058 for (eni = NULL, i = 0; i < ni->nr_extents; ++i) { ntfs_may_write_mft_record() 1073 mutex_unlock(&ni->extent_lock); ntfs_may_write_mft_record() 1084 mutex_unlock(&ni->extent_lock); ntfs_may_write_mft_record() 2250 ntfs_inode *mft_ni, *mftbmp_ni, *ni; ntfs_mft_record_alloc() local 2593 m_tmp = map_extent_mft_record(base_ni, bit, &ni); ntfs_mft_record_alloc() 2656 ni = NTFS_I(vi); ntfs_mft_record_alloc() 2665 NInoSetMstProtected(ni); ntfs_mft_record_alloc() 2666 ni->type = AT_INDEX_ALLOCATION; ntfs_mft_record_alloc() 2667 ni->name = I30; ntfs_mft_record_alloc() 2668 ni->name_len = 4; ntfs_mft_record_alloc() 2670 ni->itype.index.block_size = 4096; ntfs_mft_record_alloc() 2671 ni->itype.index.block_size_bits = ntfs_ffs(4096) - 1; ntfs_mft_record_alloc() 2672 ni->itype.index.collation_rule = COLLATION_FILE_NAME; ntfs_mft_record_alloc() 2673 if (vol->cluster_size <= ni->itype.index.block_size) { ntfs_mft_record_alloc() 2674 ni->itype.index.vcn_size = vol->cluster_size; ntfs_mft_record_alloc() 2675 ni->itype.index.vcn_size_bits = ntfs_mft_record_alloc() 2678 ni->itype.index.vcn_size = vol->sector_size; ntfs_mft_record_alloc() 2679 ni->itype.index.vcn_size_bits = ntfs_mft_record_alloc() 2686 ni->type = AT_DATA; ntfs_mft_record_alloc() 2687 ni->name = NULL; ntfs_mft_record_alloc() 2688 ni->name_len = 0; ntfs_mft_record_alloc() 2704 vi->i_generation = ni->seq_no = le16_to_cpu(m->sequence_number); ntfs_mft_record_alloc() 2709 atomic_inc(&ni->count); ntfs_mft_record_alloc() 2710 mutex_lock(&ni->mrec_lock); ntfs_mft_record_alloc() 2711 ni->page = page; ntfs_mft_record_alloc() 2712 ni->page_ofs = ofs; ntfs_mft_record_alloc() 2739 return ni; ntfs_mft_record_alloc() 2765 * @ni: ntfs inode of the mapped extent mft record to free 2766 * @m: mapped extent mft record of the ntfs inode @ni 2768 * Free the mapped extent mft record @m of the extent ntfs inode @ni. 2770 * Note that this function unmaps the mft record and closes and destroys @ni 2771 * internally and hence you cannot use either @ni nor @m any more after this 2774 * On success return 0 and on error return -errno. @ni and @m are still valid 2784 int ntfs_extent_mft_record_free(ntfs_inode *ni, MFT_RECORD *m) ntfs_extent_mft_record_free() argument 2786 unsigned long mft_no = ni->mft_no; ntfs_extent_mft_record_free() 2787 ntfs_volume *vol = ni->vol; ntfs_extent_mft_record_free() 2794 BUG_ON(NInoAttr(ni)); ntfs_extent_mft_record_free() 2795 BUG_ON(ni->nr_extents != -1); ntfs_extent_mft_record_free() 2797 mutex_lock(&ni->extent_lock); ntfs_extent_mft_record_free() 2798 base_ni = ni->ext.base_ntfs_ino; ntfs_extent_mft_record_free() 2799 mutex_unlock(&ni->extent_lock); ntfs_extent_mft_record_free() 2809 if (atomic_read(&ni->count) > 2) { ntfs_extent_mft_record_free() 2820 if (ni != extent_nis[i]) ntfs_extent_mft_record_free() 2861 NInoSetDirty(ni); ntfs_extent_mft_record_free() 2862 err = write_mft_record(ni, m, 0); ntfs_extent_mft_record_free() 2870 unmap_extent_mft_record(ni); ntfs_extent_mft_record_free() 2871 ntfs_clear_extent_inode(ni); ntfs_extent_mft_record_free() 2912 extent_nis[base_ni->nr_extents++] = ni; ntfs_extent_mft_record_free() 2914 mark_mft_record_dirty(ni); ntfs_extent_mft_record_free()
|
H A D | lcnalloc.h | 48 extern s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, 53 * @ni: ntfs inode whose runlist describes the clusters to free 54 * @start_vcn: vcn in the runlist of @ni at which to start freeing clusters 59 * described by the ntfs inode @ni. 65 * If @ctx is specified, it is an active search context of @ni and its base mft 95 * Locking: - The runlist described by @ni must be locked for writing on entry 102 * - If @ctx is NULL, the base mft record of @ni must not be mapped on 107 static inline s64 ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, ntfs_cluster_free() argument 110 return __ntfs_cluster_free(ni, start_vcn, count, ctx, false); ntfs_cluster_free()
|
H A D | compress.c | 487 ntfs_inode *ni = NTFS_I(mapping->host); ntfs_read_compressed_block() local 488 ntfs_volume *vol = ni->vol; ntfs_read_compressed_block() 496 u32 cb_size = ni->itype.compressed.block_size; ntfs_read_compressed_block() 511 >> ni->itype.compressed.block_size_bits; ntfs_read_compressed_block() 531 BUG_ON(ni->type != AT_DATA); ntfs_read_compressed_block() 532 BUG_ON(ni->name_len); ntfs_read_compressed_block() 559 read_lock_irqsave(&ni->size_lock, flags); ntfs_read_compressed_block() 560 i_size = i_size_read(VFS_I(ni)); ntfs_read_compressed_block() 561 initialized_size = ni->initialized_size; ntfs_read_compressed_block() 562 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_read_compressed_block() 605 cb_clusters = ni->itype.compressed.block_clusters; ntfs_read_compressed_block() 618 down_read(&ni->runlist.lock); ntfs_read_compressed_block() 619 rl = ni->runlist.rl; ntfs_read_compressed_block() 645 up_read(&ni->runlist.lock); ntfs_read_compressed_block() 646 if (!ntfs_map_runlist(ni, vcn)) ntfs_read_compressed_block() 663 up_read(&ni->runlist.lock); ntfs_read_compressed_block() 878 ni->mft_no, -err); ntfs_read_compressed_block() 912 "0x%lx.", ni->mft_no, page->index); ntfs_read_compressed_block() 946 up_read(&ni->runlist.lock); ntfs_read_compressed_block() 952 up_read(&ni->runlist.lock); ntfs_read_compressed_block()
|
H A D | namei.c | 180 ntfs_inode *ni = NTFS_I(dent_inode); ntfs_lookup() local 198 ni = NTFS_I(dent_inode); ntfs_lookup() 199 m = map_mft_record(ni); ntfs_lookup() 206 ctx = ntfs_attr_get_search_ctx(ni, m); ntfs_lookup() 246 unmap_mft_record(ni); ntfs_lookup() 269 unmap_mft_record(ni); ntfs_lookup() 303 ntfs_inode *ni = NTFS_I(vi); ntfs_get_parent() local 313 mrec = map_mft_record(ni); ntfs_get_parent() 317 ctx = ntfs_attr_get_search_ctx(ni, mrec); ntfs_get_parent() 319 unmap_mft_record(ni); ntfs_get_parent() 327 unmap_mft_record(ni); ntfs_get_parent() 346 unmap_mft_record(ni); ntfs_get_parent()
|
H A D | lcnalloc.c | 791 * @ni: ntfs inode whose runlist describes the clusters to free 792 * @start_vcn: vcn in the runlist of @ni at which to start freeing clusters 798 * described by the vfs inode @ni. 804 * If @ctx is specified, it is an active search context of @ni and its base mft 837 * Locking: - The runlist described by @ni must be locked for writing on entry 844 * - If @ctx is NULL, the base mft record of @ni must not be mapped on 849 s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, s64 count, __ntfs_cluster_free() argument 858 BUG_ON(!ni); __ntfs_cluster_free() 860 "0x%llx.%s", ni->mft_no, (unsigned long long)start_vcn, __ntfs_cluster_free() 863 vol = ni->vol; __ntfs_cluster_free() 880 rl = ntfs_attr_find_vcn_nolock(ni, start_vcn, ctx); __ntfs_cluster_free() 934 rl = ntfs_attr_find_vcn_nolock(ni, vcn, ctx); __ntfs_cluster_free() 1002 delta = __ntfs_cluster_free(ni, start_vcn, total_freed, ctx, true); __ntfs_cluster_free()
|
H A D | super.c | 377 ntfs_inode *ni = NTFS_I(vol->vol_ino); ntfs_write_volume_flags() local 387 BUG_ON(!ni); ntfs_write_volume_flags() 388 m = map_mft_record(ni); ntfs_write_volume_flags() 393 ctx = ntfs_attr_get_search_ctx(ni, m); ntfs_write_volume_flags() 408 unmap_mft_record(ni); ntfs_write_volume_flags() 415 unmap_mft_record(ni); ntfs_write_volume_flags() 3050 ntfs_inode *ni = (ntfs_inode *)foo; ntfs_big_inode_init_once() local 3052 inode_init_once(VFS_I(ni)); ntfs_big_inode_init_once()
|
/linux-4.4.14/fs/f2fs/ |
H A D | node.h | 56 struct node_info ni; /* in-memory node information */ member in struct:nat_entry 59 #define nat_get_nid(nat) (nat->ni.nid) 60 #define nat_set_nid(nat, n) (nat->ni.nid = n) 61 #define nat_get_blkaddr(nat) (nat->ni.blk_addr) 62 #define nat_set_blkaddr(nat, b) (nat->ni.blk_addr = b) 63 #define nat_get_ino(nat) (nat->ni.ino) 64 #define nat_set_ino(nat, i) (nat->ni.ino = i) 65 #define nat_get_version(nat) (nat->ni.version) 66 #define nat_set_version(nat, v) (nat->ni.version = v) 85 ne->ni.flag |= mask; set_nat_flag() 87 ne->ni.flag &= ~mask; set_nat_flag() 93 return ne->ni.flag & mask; get_nat_flag() 104 static inline void node_info_from_raw_nat(struct node_info *ni, node_info_from_raw_nat() argument 107 ni->ino = le32_to_cpu(raw_ne->ino); node_info_from_raw_nat() 108 ni->blk_addr = le32_to_cpu(raw_ne->block_addr); node_info_from_raw_nat() 109 ni->version = raw_ne->version; node_info_from_raw_nat() 113 struct node_info *ni) raw_nat_from_node_info() 115 raw_ne->ino = cpu_to_le32(ni->ino); raw_nat_from_node_info() 116 raw_ne->block_addr = cpu_to_le32(ni->blk_addr); raw_nat_from_node_info() 117 raw_ne->version = ni->version; raw_nat_from_node_info() 112 raw_nat_from_node_info(struct f2fs_nat_entry *raw_ne, struct node_info *ni) raw_nat_from_node_info() argument
|
H A D | node.c | 154 nid_t set = NAT_BLOCK_OFFSET(ne->ni.nid); __set_nat_cache_dirty() 179 nid_t set = NAT_BLOCK_OFFSET(ne->ni.nid); __clear_nat_cache_dirty() 268 node_info_from_raw_nat(&e->ni, ne); cache_nat_entry() 273 static void set_node_addr(struct f2fs_sb_info *sbi, struct node_info *ni, set_node_addr() argument 280 e = __lookup_nat_cache(nm_i, ni->nid); set_node_addr() 282 e = grab_nat_entry(nm_i, ni->nid); set_node_addr() 283 copy_node_info(&e->ni, ni); set_node_addr() 284 f2fs_bug_on(sbi, ni->blk_addr == NEW_ADDR); set_node_addr() 291 copy_node_info(&e->ni, ni); set_node_addr() 292 f2fs_bug_on(sbi, ni->blk_addr != NULL_ADDR); set_node_addr() 296 f2fs_bug_on(sbi, nat_get_blkaddr(e) != ni->blk_addr); set_node_addr() 311 if (nm_i->next_scan_nid > ni->nid) set_node_addr() 312 nm_i->next_scan_nid = ni->nid; set_node_addr() 322 if (ni->nid != ni->ino) set_node_addr() 323 e = __lookup_nat_cache(nm_i, ni->ino); set_node_addr() 325 if (fsync_done && ni->nid == ni->ino) set_node_addr() 354 void get_node_info(struct f2fs_sb_info *sbi, nid_t nid, struct node_info *ni) get_node_info() argument 366 ni->nid = nid; get_node_info() 372 ni->ino = nat_get_ino(e); get_node_info() 373 ni->blk_addr = nat_get_blkaddr(e); get_node_info() 374 ni->version = nat_get_version(e); get_node_info() 387 node_info_from_raw_nat(ni, &ne); get_node_info() 397 node_info_from_raw_nat(ni, &ne); get_node_info() 592 struct node_info ni; truncate_node() local 594 get_node_info(sbi, dn->nid, &ni); truncate_node() 596 f2fs_bug_on(sbi, ni.blk_addr != NULL_ADDR); truncate_node() 599 f2fs_bug_on(sbi, ni.blk_addr == NULL_ADDR); truncate_node() 602 invalidate_blocks(sbi, ni.blk_addr); truncate_node() 604 set_node_addr(sbi, &ni, NULL_ADDR, false); truncate_node() 622 trace_f2fs_truncate_node(dn->inode, dn->nid, ni.blk_addr); truncate_node() 1007 struct node_info ni; read_node_page() local 1016 get_node_info(sbi, page->index, &ni); read_node_page() 1018 if (unlikely(ni.blk_addr == NULL_ADDR)) { read_node_page() 1026 fio.blk_addr = ni.blk_addr; read_node_page() 1304 struct node_info ni; f2fs_write_node_page() local 1333 get_node_info(sbi, nid, &ni); f2fs_write_node_page() 1336 if (unlikely(ni.blk_addr == NULL_ADDR)) { f2fs_write_node_page() 1345 fio.blk_addr = ni.blk_addr; f2fs_write_node_page() 1347 set_node_addr(sbi, &ni, fio.blk_addr, is_fsync_dnode(page)); f2fs_write_node_page() 1585 struct node_info ni; alloc_nid() local 1599 get_node_info(sbi, *nid, &ni); alloc_nid() 1600 if (ni.blk_addr != NULL_ADDR) { alloc_nid() 1718 struct node_info ni; recover_xattr_data() local 1725 get_node_info(sbi, prev_xnid, &ni); recover_xattr_data() 1726 f2fs_bug_on(sbi, ni.blk_addr == NULL_ADDR); recover_xattr_data() 1727 invalidate_blocks(sbi, ni.blk_addr); recover_xattr_data() 1729 set_node_addr(sbi, &ni, NULL_ADDR, false); recover_xattr_data() 1737 get_node_info(sbi, new_xnid, &ni); recover_xattr_data() 1738 ni.ino = inode->i_ino; recover_xattr_data() 1739 set_node_addr(sbi, &ni, NEW_ADDR, false); recover_xattr_data() 1744 set_node_addr(sbi, &ni, blkaddr, false); recover_xattr_data() 1849 node_info_from_raw_nat(&ne->ni, &raw_ne); remove_nats_in_journal() 1922 raw_nat_from_node_info(raw_ne, &ne->ni); __flush_nat_entry_set()
|
H A D | recovery.c | 363 struct node_info ni; do_recover_data() local 394 get_node_info(sbi, dn.nid, &ni); do_recover_data() 395 f2fs_bug_on(sbi, ni.ino != ino_of_node(page)); do_recover_data() 441 ni.version, false); do_recover_data() 450 fill_node_footer(dn.node_page, dn.nid, ni.ino, do_recover_data()
|
H A D | gc.c | 419 struct node_info ni; gc_node_segment() local 442 get_node_info(sbi, nid, &ni); gc_node_segment() 443 if (ni.blk_addr != start_addr + off) { gc_node_segment() 548 struct node_info ni; move_encrypted_block() local 573 get_node_info(fio.sbi, dn.nid, &ni); move_encrypted_block() 574 set_summary(&sum, dn.nid, dn.ofs_in_node, ni.version); move_encrypted_block()
|
H A D | inline.c | 581 struct node_info ni; f2fs_inline_data_fiemap() local 601 get_node_info(F2FS_I_SB(inode), inode->i_ino, &ni); f2fs_inline_data_fiemap() local 602 byteaddr = (__u64)ni.blk_addr << inode->i_sb->s_blocksize_bits; f2fs_inline_data_fiemap()
|
H A D | data.c | 466 struct node_info ni; __allocate_data_block() local 481 get_node_info(sbi, dn->nid, &ni); __allocate_data_block() 482 set_summary(&sum, dn->nid, dn->ofs_in_node, ni.version); __allocate_data_block()
|
H A D | file.c | 866 struct node_info ni; __exchange_data_block() local 880 get_node_info(sbi, dn.nid, &ni); __exchange_data_block() 882 ni.version, true); __exchange_data_block()
|
H A D | segment.c | 1332 struct node_info ni; write_data_page() local 1335 get_node_info(sbi, dn->nid, &ni); write_data_page() 1336 set_summary(&sum, dn->nid, dn->ofs_in_node, ni.version); write_data_page()
|
/linux-4.4.14/include/uapi/linux/netfilter_bridge/ |
H A D | ebt_802_3.h | 28 /* ui has one byte ctrl, ni has two */ 51 struct hdr_ni ni; member in union:ebt_802_3_hdr::__anon14322
|
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/socklnd/ |
H A D | socklnd.c | 50 ksocknal_ip2iface(lnet_ni_t *ni, __u32 ip) ksocknal_ip2iface() argument 52 ksock_net_t *net = ni->ni_data; ksocknal_ip2iface() 103 ksocknal_create_peer(ksock_peer_t **peerp, lnet_ni_t *ni, lnet_process_id_t id) ksocknal_create_peer() argument 105 ksock_net_t *net = ni->ni_data; ksocknal_create_peer() 116 peer->ksnp_ni = ni; ksocknal_create_peer() 176 ksocknal_find_peer_locked(lnet_ni_t *ni, lnet_process_id_t id) ksocknal_find_peer_locked() argument 188 if (peer->ksnp_ni != ni) list_for_each() 204 ksocknal_find_peer(lnet_ni_t *ni, lnet_process_id_t id) ksocknal_find_peer() argument 209 peer = ksocknal_find_peer_locked(ni, id); ksocknal_find_peer() 248 ksocknal_get_peer_info(lnet_ni_t *ni, int index, ksocknal_get_peer_info() argument 267 if (peer->ksnp_ni != ni) ksocknal_get_peer_info() 448 ksocknal_add_peer(lnet_ni_t *ni, lnet_process_id_t id, __u32 ipaddr, int port) ksocknal_add_peer() argument 462 rc = ksocknal_create_peer(&peer, ni, id); ksocknal_add_peer() 474 /* always called with a ref on ni, so shutdown can't have started */ ksocknal_add_peer() 475 LASSERT(((ksock_net_t *) ni->ni_data)->ksnn_shutdown == 0); ksocknal_add_peer() 477 peer2 = ksocknal_find_peer_locked(ni, id); ksocknal_add_peer() 565 ksocknal_del_peer(lnet_ni_t *ni, lnet_process_id_t id, __u32 ip) ksocknal_del_peer() argument 590 if (peer->ksnp_ni != ni) ksocknal_del_peer() 618 ksocknal_txlist_done(ni, &zombies, 1); ksocknal_del_peer() 624 ksocknal_get_conn_by_idx(lnet_ni_t *ni, int index) ksocknal_get_conn_by_idx() argument 640 if (peer->ksnp_ni != ni) ksocknal_get_conn_by_idx() 684 ksocknal_local_ipvec(lnet_ni_t *ni, __u32 *ipaddrs) ksocknal_local_ipvec() argument 686 ksock_net_t *net = ni->ni_data; ksocknal_local_ipvec() 851 lnet_ni_t *ni = peer->ksnp_ni; ksocknal_create_routes() local 852 ksock_net_t *net = ni->ni_data; ksocknal_create_routes() 964 ksocknal_accept(lnet_ni_t *ni, struct socket *sock) ksocknal_accept() argument 981 lnet_ni_addref(ni); ksocknal_accept() 982 cr->ksncr_ni = ni; ksocknal_accept() 1008 ksocknal_create_conn(lnet_ni_t *ni, ksock_route_t *route, ksocknal_create_conn() argument 1077 LASSERT(ni == peer->ksnp_ni); ksocknal_create_conn() 1080 hello->kshm_nips = ksocknal_local_ipvec(ni, hello->kshm_ips); ksocknal_create_conn() 1097 rc = ksocknal_send_hello(ni, conn, peerid.nid, hello); ksocknal_create_conn() 1108 rc = ksocknal_recv_hello(ni, conn, hello, &peerid, &incarnation); ksocknal_create_conn() 1122 rc = ksocknal_create_peer(&peer, ni, peerid); ksocknal_create_conn() 1128 /* called with a ref on ni, so shutdown can't have started */ ksocknal_create_conn() 1129 LASSERT(((ksock_net_t *) ni->ni_data)->ksnn_shutdown == 0); ksocknal_create_conn() 1131 peer2 = ksocknal_find_peer_locked(ni, peerid); ksocknal_create_conn() 1148 if (peerid.nid < ni->ni_nid && ksocknal_create_conn() 1302 rc = ksocknal_send_hello(ni, conn, peerid.nid, hello); ksocknal_create_conn() 1370 ksocknal_send_hello(ni, conn, peerid.nid, hello); ksocknal_create_conn() 1378 ksocknal_txlist_done(ni, &zombies, 1); ksocknal_create_conn() 1761 ksocknal_notify(lnet_ni_t *ni, lnet_nid_t gw_nid, int alive) ksocknal_notify() argument 1784 ksocknal_query(lnet_ni_t *ni, lnet_nid_t nid, unsigned long *when) ksocknal_query() argument 1795 peer = ksocknal_find_peer_locked(ni, id); ksocknal_query() 1832 ksocknal_add_peer(ni, id, LNET_NIDADDR(nid), lnet_acceptor_port()); ksocknal_query() 1836 peer = ksocknal_find_peer_locked(ni, id); ksocknal_query() 1876 static int ksocknal_push(lnet_ni_t *ni, lnet_process_id_t id) ksocknal_push() argument 1925 ksocknal_add_interface(lnet_ni_t *ni, __u32 ipaddress, __u32 netmask) ksocknal_add_interface() argument 1927 ksock_net_t *net = ni->ni_data; ksocknal_add_interface() 1943 iface = ksocknal_ip2iface(ni, ipaddress); ksocknal_add_interface() 2028 ksocknal_del_interface(lnet_ni_t *ni, __u32 ipaddress) ksocknal_del_interface() argument 2030 ksock_net_t *net = ni->ni_data; ksocknal_del_interface() 2062 if (peer->ksnp_ni != ni) ksocknal_del_interface() 2076 ksocknal_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg) ksocknal_ctl() argument 2084 ksock_net_t *net = ni->ni_data; ksocknal_ctl() 2106 return ksocknal_add_interface(ni, ksocknal_ctl() 2111 return ksocknal_del_interface(ni, ksocknal_ctl() 2121 rc = ksocknal_get_peer_info(ni, data->ioc_count, ksocknal_ctl() 2140 return ksocknal_add_peer(ni, id, ksocknal_ctl() 2147 return ksocknal_del_peer(ni, id, ksocknal_ctl() 2154 ksock_conn_t *conn = ksocknal_get_conn_by_idx(ni, data->ioc_count); ksocknal_ctl() 2183 if (data->ioc_nid == ni->ni_nid) ksocknal_ctl() 2188 libcfs_nid2str(ni->ni_nid)); ksocknal_ctl() 2194 return ksocknal_push(ni, id); ksocknal_ctl() 2475 ksocknal_debug_peerhash(lnet_ni_t *ni) ksocknal_debug_peerhash() argument 2487 if (peer->ksnp_ni == ni) ksocknal_debug_peerhash() 2529 ksocknal_shutdown(lnet_ni_t *ni) ksocknal_shutdown() argument 2531 ksock_net_t *net = ni->ni_data; ksocknal_shutdown() 2546 ksocknal_del_peer(ni, anyid, 0); ksocknal_shutdown() 2561 ksocknal_debug_peerhash(ni); ksocknal_shutdown() 2753 ksocknal_startup(lnet_ni_t *ni) ksocknal_startup() argument 2759 LASSERT(ni->ni_lnd == &the_ksocklnd); ksocknal_startup() 2773 ni->ni_data = net; ksocknal_startup() 2774 ni->ni_peertimeout = *ksocknal_tunables.ksnd_peertimeout; ksocknal_startup() 2775 ni->ni_maxtxcredits = *ksocknal_tunables.ksnd_credits; ksocknal_startup() 2776 ni->ni_peertxcredits = *ksocknal_tunables.ksnd_peertxcredits; ksocknal_startup() 2777 ni->ni_peerrtrcredits = *ksocknal_tunables.ksnd_peerrtrcredits; ksocknal_startup() 2779 if (ni->ni_interfaces[0] == NULL) { ksocknal_startup() 2789 if (ni->ni_interfaces[i] == NULL) ksocknal_startup() 2792 rc = lnet_ipif_query(ni->ni_interfaces[i], &up, ksocknal_startup() 2798 ni->ni_interfaces[i], rc); ksocknal_startup() 2804 ni->ni_interfaces[i]); ksocknal_startup() 2809 ni->ni_interfaces[i], IFNAMSIZ); ksocknal_startup() 2815 rc = ksocknal_net_start_threads(net, ni->ni_cpts, ni->ni_ncpts); ksocknal_startup() 2819 ni->ni_nid = LNET_MKNID(LNET_NIDNET(ni->ni_nid), ksocknal_startup()
|
H A D | socklnd.h | 570 void ksocknal_tx_done(lnet_ni_t *ni, ksock_tx_t *tx); 614 int ksocknal_startup(lnet_ni_t *ni); 615 void ksocknal_shutdown(lnet_ni_t *ni); 616 int ksocknal_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg); 617 int ksocknal_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg); 618 int ksocknal_recv(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg, 622 int ksocknal_accept(lnet_ni_t *ni, struct socket *sock); 624 int ksocknal_add_peer(lnet_ni_t *ni, lnet_process_id_t id, __u32 ip, int port); 625 ksock_peer_t *ksocknal_find_peer_locked(lnet_ni_t *ni, lnet_process_id_t id); 626 ksock_peer_t *ksocknal_find_peer(lnet_ni_t *ni, lnet_process_id_t id); 628 int ksocknal_create_conn(lnet_ni_t *ni, ksock_route_t *route, 640 int ksocknal_launch_packet(lnet_ni_t *ni, ksock_tx_t *tx, 647 void ksocknal_txlist_done(lnet_ni_t *ni, struct list_head *txlist, int error); 648 void ksocknal_notify(lnet_ni_t *ni, lnet_nid_t gw_nid, int alive); 649 void ksocknal_query(struct lnet_ni *ni, lnet_nid_t nid, unsigned long *when); 659 int ksocknal_send_hello(lnet_ni_t *ni, ksock_conn_t *conn, 661 int ksocknal_recv_hello(lnet_ni_t *ni, ksock_conn_t *conn,
|
H A D | socklnd_cb.c | 389 ksocknal_tx_done (lnet_ni_t *ni, ksock_tx_t *tx) ksocknal_tx_done() argument 394 LASSERT(ni != NULL || tx->tx_conn != NULL); ksocknal_tx_done() 399 if (ni == NULL && tx->tx_conn != NULL) ksocknal_tx_done() 400 ni = tx->tx_conn->ksnc_peer->ksnp_ni; ksocknal_tx_done() 404 lnet_finalize (ni, lnetmsg, rc); ksocknal_tx_done() 408 ksocknal_txlist_done (lnet_ni_t *ni, struct list_head *txlist, int error) ksocknal_txlist_done() argument 428 ksocknal_tx_done(ni, tx); ksocknal_txlist_done() 830 ksocknal_launch_packet (lnet_ni_t *ni, ksock_tx_t *tx, lnet_process_id_t id) ksocknal_launch_packet() argument 844 peer = ksocknal_find_peer_locked(ni, id); ksocknal_launch_packet() 864 peer = ksocknal_find_peer_locked(ni, id); ksocknal_launch_packet() 881 rc = ksocknal_add_peer(ni, id, ksocknal_launch_packet() 921 ksocknal_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg) ksocknal_send() argument 991 rc = ksocknal_launch_packet(ni, tx, target); ksocknal_send() 1311 ksocknal_recv (lnet_ni_t *ni, void *private, lnet_msg_t *msg, int delayed, ksocknal_recv() argument 1647 ksocknal_send_hello (lnet_ni_t *ni, ksock_conn_t *conn, ksocknal_send_hello() argument 1651 ksock_net_t *net = (ksock_net_t *)ni->ni_data; ksocknal_send_hello() 1658 hello->kshm_src_nid = ni->ni_nid; ksocknal_send_hello() 1685 ksocknal_recv_hello (lnet_ni_t *ni, ksock_conn_t *conn, ksocknal_recv_hello() argument 1748 ksocknal_send_hello(ni, conn, ni->ni_nid, hello); ksocknal_recv_hello() 1782 recv_id.nid = LNET_MKNID(LNET_NIDNET(ni->ni_nid), conn->ksnc_ipaddr); ksocknal_recv_hello()
|
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
H A D | o2iblnd_cb.c | 43 static void kiblnd_unmap_tx(lnet_ni_t *ni, kib_tx_t *tx); 46 kiblnd_tx_done(lnet_ni_t *ni, kib_tx_t *tx) kiblnd_tx_done() argument 49 kib_net_t *net = ni->ni_data; kiblnd_tx_done() 60 kiblnd_unmap_tx(ni, tx); kiblnd_tx_done() 68 LASSERT(ni == tx->tx_conn->ibc_peer->ibp_ni); kiblnd_tx_done() 84 lnet_finalize(ni, lntmsg[i], rc); kiblnd_tx_done() 89 kiblnd_txlist_done(lnet_ni_t *ni, struct list_head *txlist, int status) kiblnd_txlist_done() argument 100 kiblnd_tx_done(ni, tx); kiblnd_txlist_done() 105 kiblnd_get_idle_tx(lnet_ni_t *ni, lnet_nid_t target) kiblnd_get_idle_tx() argument 107 kib_net_t *net = (kib_net_t *)ni->ni_data; kiblnd_get_idle_tx() 248 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_handle_completion() local 267 lnet_set_reply_msg_len(ni, tx->tx_lntmsg[1], status); kiblnd_handle_completion() 279 kiblnd_tx_done(ni, tx); kiblnd_handle_completion() 285 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_send_completion() local 286 kib_tx_t *tx = kiblnd_get_idle_tx(ni, conn->ibc_peer->ibp_nid); kiblnd_send_completion() 296 kiblnd_init_tx_msg(ni, tx, type, sizeof(kib_completion_msg_t)); kiblnd_send_completion() 306 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_handle_rx() local 371 rc = lnet_parse(ni, &msg->ibm_u.immediate.ibim_hdr, kiblnd_handle_rx() 379 rc = lnet_parse(ni, &msg->ibm_u.putreq.ibprm_hdr, kiblnd_handle_rx() 441 rc = lnet_parse(ni, &msg->ibm_u.get.ibgm_hdr, kiblnd_handle_rx() 467 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_rx_complete() local 468 kib_net_t *net = ni->ni_data; kiblnd_rx_complete() 496 msg->ibm_dstnid != ni->ni_nid || kiblnd_rx_complete() 601 static void kiblnd_unmap_tx(lnet_ni_t *ni, kib_tx_t *tx) kiblnd_unmap_tx() argument 603 kib_net_t *net = ni->ni_data; kiblnd_unmap_tx() 619 static int kiblnd_map_tx(lnet_ni_t *ni, kib_tx_t *tx, kib_rdma_desc_t *rd, kiblnd_map_tx() argument 623 kib_net_t *net = ni->ni_data; kiblnd_map_tx() 659 kiblnd_setup_rd_iov(lnet_ni_t *ni, kib_tx_t *tx, kib_rdma_desc_t *rd, kiblnd_setup_rd_iov() argument 662 kib_net_t *net = ni->ni_data; kiblnd_setup_rd_iov() 708 return kiblnd_map_tx(ni, tx, rd, sg - tx->tx_frags); kiblnd_setup_rd_iov() 712 kiblnd_setup_rd_kiov(lnet_ni_t *ni, kib_tx_t *tx, kib_rdma_desc_t *rd, kiblnd_setup_rd_kiov() argument 715 kib_net_t *net = ni->ni_data; kiblnd_setup_rd_kiov() 748 return kiblnd_map_tx(ni, tx, rd, sg - tx->tx_frags); kiblnd_setup_rd_kiov() 888 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_check_sends() local 917 tx = kiblnd_get_idle_tx(ni, conn->ibc_peer->ibp_nid); kiblnd_check_sends() 919 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_NOOP, 0); kiblnd_check_sends() 1011 kiblnd_init_tx_msg(lnet_ni_t *ni, kib_tx_t *tx, int type, int body_nob) kiblnd_init_tx_msg() argument 1296 kiblnd_launch_tx(lnet_ni_t *ni, kib_tx_t *tx, lnet_nid_t nid) kiblnd_launch_tx() argument 1359 rc = kiblnd_create_peer(ni, &peer, nid); kiblnd_launch_tx() 1365 kiblnd_tx_done(ni, tx); kiblnd_launch_tx() 1401 /* always called with a ref on ni, which prevents ni being shutdown */ kiblnd_launch_tx() 1402 LASSERT(((kib_net_t *)ni->ni_data)->ibn_shutdown == 0); kiblnd_launch_tx() 1417 kiblnd_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg) kiblnd_send() argument 1466 tx = kiblnd_get_idle_tx(ni, target.nid); kiblnd_send() 1476 rc = kiblnd_setup_rd_iov(ni, tx, rd, kiblnd_send() 1481 rc = kiblnd_setup_rd_kiov(ni, tx, rd, kiblnd_send() 1488 kiblnd_tx_done(ni, tx); kiblnd_send() 1496 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_GET_REQ, nob); kiblnd_send() 1498 tx->tx_lntmsg[1] = lnet_create_reply_msg(ni, lntmsg); kiblnd_send() 1502 kiblnd_tx_done(ni, tx); kiblnd_send() 1508 kiblnd_launch_tx(ni, tx, target.nid); kiblnd_send() 1518 tx = kiblnd_get_idle_tx(ni, target.nid); kiblnd_send() 1527 rc = kiblnd_setup_rd_iov(ni, tx, tx->tx_rd, kiblnd_send() 1531 rc = kiblnd_setup_rd_kiov(ni, tx, tx->tx_rd, kiblnd_send() 1537 kiblnd_tx_done(ni, tx); kiblnd_send() 1544 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_PUT_REQ, sizeof(kib_putreq_msg_t)); kiblnd_send() 1548 kiblnd_launch_tx(ni, tx, target.nid); kiblnd_send() 1557 tx = kiblnd_get_idle_tx(ni, target.nid); kiblnd_send() 1579 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_IMMEDIATE, nob); kiblnd_send() 1582 kiblnd_launch_tx(ni, tx, target.nid); kiblnd_send() 1587 kiblnd_reply(lnet_ni_t *ni, kib_rx_t *rx, lnet_msg_t *lntmsg) kiblnd_reply() argument 1598 tx = kiblnd_get_idle_tx(ni, rx->rx_conn->ibc_peer->ibp_nid); kiblnd_reply() 1608 rc = kiblnd_setup_rd_iov(ni, tx, tx->tx_rd, kiblnd_reply() 1611 rc = kiblnd_setup_rd_kiov(ni, tx, tx->tx_rd, kiblnd_reply() 1632 lnet_finalize(ni, lntmsg, 0); kiblnd_reply() 1643 kiblnd_tx_done(ni, tx); kiblnd_reply() 1645 lnet_finalize(ni, lntmsg, -EIO); kiblnd_reply() 1649 kiblnd_recv(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg, int delayed, kiblnd_recv() argument 1690 lnet_finalize(ni, lntmsg, 0); kiblnd_recv() 1698 lnet_finalize(ni, lntmsg, 0); kiblnd_recv() 1704 tx = kiblnd_get_idle_tx(ni, conn->ibc_peer->ibp_nid); kiblnd_recv() 1716 rc = kiblnd_setup_rd_iov(ni, tx, rd, kiblnd_recv() 1719 rc = kiblnd_setup_rd_kiov(ni, tx, rd, kiblnd_recv() 1724 kiblnd_tx_done(ni, tx); kiblnd_recv() 1735 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_PUT_ACK, nob); kiblnd_recv() 1749 kiblnd_reply(ni, rx, lntmsg); kiblnd_recv() 2099 lnet_ni_t *ni = peer->ibp_ni; kiblnd_connreq_done() local 2105 kiblnd_txlist_done(ni, &txs, -ECONNABORTED); kiblnd_connreq_done() 2148 lnet_ni_t *ni = NULL; kiblnd_passive_connect() local 2209 ni = lnet_net2ni(LNET_NIDNET(reqmsg->ibm_dstnid)); kiblnd_passive_connect() 2211 if (ni != NULL) { kiblnd_passive_connect() 2212 net = (kib_net_t *)ni->ni_data; kiblnd_passive_connect() 2216 if (ni == NULL || /* no matching net */ kiblnd_passive_connect() 2217 ni->ni_nid != reqmsg->ibm_dstnid || /* right NET, wrong NID! */ kiblnd_passive_connect() 2221 ni == NULL ? "NA" : libcfs_nid2str(ni->ni_nid), kiblnd_passive_connect() 2281 rc = kiblnd_create_peer(ni, &peer, nid); kiblnd_passive_connect() 2313 nid < ni->ni_nid) { kiblnd_passive_connect() 2339 /* I have a ref on ni that prevents it being shutdown */ kiblnd_passive_connect() 2374 kiblnd_pack_msg(ni, ackmsg, version, 0, nid, reqmsg->ibm_srcstamp); kiblnd_passive_connect() 2398 lnet_ni_decref(ni); kiblnd_passive_connect() 2402 if (ni != NULL) kiblnd_passive_connect() 2403 lnet_ni_decref(ni); kiblnd_passive_connect() 2619 lnet_ni_t *ni = peer->ibp_ni; kiblnd_check_connreply() local 2620 kib_net_t *net = ni->ni_data; kiblnd_check_connreply() 2678 if (msg->ibm_dstnid == ni->ni_nid && kiblnd_check_connreply()
|
H A D | o2iblnd.c | 187 void kiblnd_pack_msg(lnet_ni_t *ni, kib_msg_t *msg, int version, kiblnd_pack_msg() argument 190 kib_net_t *net = ni->ni_data; kiblnd_pack_msg() 200 msg->ibm_srcnid = ni->ni_nid; kiblnd_pack_msg() 322 int kiblnd_create_peer(lnet_ni_t *ni, kib_peer_t **peerp, lnet_nid_t nid) kiblnd_create_peer() argument 325 kib_net_t *net = ni->ni_data; kiblnd_create_peer() 340 peer->ibp_ni = ni; kiblnd_create_peer() 352 /* always called with a ref on ni, which prevents ni being shutdown */ kiblnd_create_peer() 423 static int kiblnd_get_peer_info(lnet_ni_t *ni, int index, kiblnd_get_peer_info() argument 442 if (peer->ibp_ni != ni) kiblnd_get_peer_info() 481 static int kiblnd_del_peer(lnet_ni_t *ni, lnet_nid_t nid) kiblnd_del_peer() argument 509 if (peer->ibp_ni != ni) kiblnd_del_peer() 529 kiblnd_txlist_done(ni, &zombies, -EIO); kiblnd_del_peer() 534 static kib_conn_t *kiblnd_get_conn_by_idx(lnet_ni_t *ni, int index) kiblnd_get_conn_by_idx() argument 553 if (peer->ibp_ni != ni) kiblnd_get_conn_by_idx() 957 static int kiblnd_close_matching_conns(lnet_ni_t *ni, lnet_nid_t nid) kiblnd_close_matching_conns() argument 985 if (peer->ibp_ni != ni) kiblnd_close_matching_conns() 1004 int kiblnd_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg) kiblnd_ctl() argument 1014 rc = kiblnd_get_peer_info(ni, data->ioc_count, kiblnd_ctl() 1022 rc = kiblnd_del_peer(ni, data->ioc_nid); kiblnd_ctl() 1029 conn = kiblnd_get_conn_by_idx(ni, data->ioc_count); kiblnd_ctl() 1046 rc = kiblnd_close_matching_conns(ni, data->ioc_nid); kiblnd_ctl() 1057 void kiblnd_query(lnet_ni_t *ni, lnet_nid_t nid, unsigned long *when) kiblnd_query() argument 1083 kiblnd_launch_tx(ni, NULL, nid); kiblnd_query() 2494 void kiblnd_shutdown(lnet_ni_t *ni) kiblnd_shutdown() argument 2496 kib_net_t *net = ni->ni_data; kiblnd_shutdown() 2516 kiblnd_del_peer(ni, LNET_NID_ANY); kiblnd_shutdown() 2524 libcfs_nid2str(ni->ni_nid), kiblnd_shutdown() 2551 ni->ni_data = NULL; kiblnd_shutdown() 2747 int kiblnd_startup(lnet_ni_t *ni) kiblnd_startup() argument 2757 LASSERT(ni->ni_lnd == &the_o2iblnd); kiblnd_startup() 2766 ni->ni_data = net; kiblnd_startup() 2774 ni->ni_peertimeout = *kiblnd_tunables.kib_peertimeout; kiblnd_startup() 2775 ni->ni_maxtxcredits = *kiblnd_tunables.kib_credits; kiblnd_startup() 2776 ni->ni_peertxcredits = *kiblnd_tunables.kib_peertxcredits; kiblnd_startup() 2777 ni->ni_peerrtrcredits = *kiblnd_tunables.kib_peerrtrcredits; kiblnd_startup() 2779 if (ni->ni_interfaces[0] != NULL) { kiblnd_startup() 2783 if (ni->ni_interfaces[1] != NULL) { kiblnd_startup() 2788 ifname = ni->ni_interfaces[0]; kiblnd_startup() 2809 ni->ni_nid = LNET_MKNID(LNET_NIDNET(ni->ni_nid), ibdev->ibd_ifip); kiblnd_startup() 2812 ni->ni_cpts, ni->ni_ncpts); kiblnd_startup() 2816 rc = kiblnd_net_init_pools(net, ni->ni_cpts, ni->ni_ncpts); kiblnd_startup() 2836 kiblnd_shutdown(ni); kiblnd_startup()
|
H A D | o2iblnd.h | 922 int kiblnd_startup(lnet_ni_t *ni); 923 void kiblnd_shutdown(lnet_ni_t *ni); 924 int kiblnd_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg); 925 void kiblnd_query(struct lnet_ni *ni, lnet_nid_t nid, unsigned long *when); 943 int kiblnd_create_peer(lnet_ni_t *ni, kib_peer_t **peerp, lnet_nid_t nid); 964 void kiblnd_launch_tx(lnet_ni_t *ni, kib_tx_t *tx, lnet_nid_t nid); 967 void kiblnd_init_tx_msg(lnet_ni_t *ni, kib_tx_t *tx, int type, int body_nob); 968 void kiblnd_txlist_done(lnet_ni_t *ni, struct list_head *txlist, 976 void kiblnd_pack_msg(lnet_ni_t *ni, kib_msg_t *msg, int version, 981 int kiblnd_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg); 982 int kiblnd_recv(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg, int delayed,
|
/linux-4.4.14/drivers/staging/lustre/include/linux/lnet/ |
H A D | lib-lnet.h | 159 #define lnet_ni_lock(ni) spin_lock(&(ni)->ni_lock) 160 #define lnet_ni_unlock(ni) spin_unlock(&(ni)->ni_lock) 375 lnet_ni_addref_locked(lnet_ni_t *ni, int cpt) lnet_ni_addref_locked() argument 378 LASSERT(*ni->ni_refs[cpt] >= 0); lnet_ni_addref_locked() 380 (*ni->ni_refs[cpt])++; lnet_ni_addref_locked() 384 lnet_ni_addref(lnet_ni_t *ni) lnet_ni_addref() argument 387 lnet_ni_addref_locked(ni, 0); lnet_ni_addref() 392 lnet_ni_decref_locked(lnet_ni_t *ni, int cpt) lnet_ni_decref_locked() argument 395 LASSERT(*ni->ni_refs[cpt] > 0); lnet_ni_decref_locked() 397 (*ni->ni_refs[cpt])--; lnet_ni_decref_locked() 401 lnet_ni_decref(lnet_ni_t *ni) lnet_ni_decref() argument 404 lnet_ni_decref_locked(ni, 0); lnet_ni_decref() 408 void lnet_ni_free(lnet_ni_t *ni); 436 int lnet_notify(lnet_ni_t *ni, lnet_nid_t peer, int alive, unsigned long when); 523 int lnet_parse(lnet_ni_t *ni, lnet_hdr_t *hdr, 525 void lnet_recv(lnet_ni_t *ni, void *private, lnet_msg_t *msg, int delayed, 527 lnet_msg_t *lnet_create_reply_msg(lnet_ni_t *ni, lnet_msg_t *get_msg); 528 void lnet_set_reply_msg_len(lnet_ni_t *ni, lnet_msg_t *msg, unsigned int len); 530 void lnet_finalize(lnet_ni_t *ni, lnet_msg_t *msg, int rc); 656 int lnet_peer_buffer_credits(lnet_ni_t *ni);
|
H A D | lib-types.h | 189 int (*lnd_startup)(struct lnet_ni *ni); 190 void (*lnd_shutdown)(struct lnet_ni *ni); 191 int (*lnd_ctl)(struct lnet_ni *ni, unsigned int cmd, void *arg); 209 int (*lnd_send)(struct lnet_ni *ni, void *private, lnet_msg_t *msg); 216 int (*lnd_recv)(struct lnet_ni *ni, void *private, lnet_msg_t *msg, 228 int (*lnd_eager_recv)(struct lnet_ni *ni, void *private, 232 void (*lnd_notify)(struct lnet_ni *ni, lnet_nid_t peer, int alive); 235 void (*lnd_query)(struct lnet_ni *ni, lnet_nid_t peer, 239 int (*lnd_accept)(struct lnet_ni *ni, struct socket *sock); 598 /* uniquely identifies this ni in this epoch */
|
/linux-4.4.14/fs/lockd/ |
H A D | host.c | 109 static struct nlm_host *nlm_alloc_host(struct nlm_lookup_host_info *ni, nlm_alloc_host() argument 119 nsm = nsm_get_handle(ni->net, ni->sap, ni->salen, nlm_alloc_host() 120 ni->hostname, ni->hostname_len); nlm_alloc_host() 135 memcpy(nlm_addr(host), ni->sap, ni->salen); nlm_alloc_host() 136 host->h_addrlen = ni->salen; nlm_alloc_host() 142 host->h_version = ni->version; nlm_alloc_host() 143 host->h_proto = ni->protocol; nlm_alloc_host() 145 host->h_server = ni->server; nlm_alloc_host() 146 host->h_noresvport = ni->noresvport; nlm_alloc_host() 163 host->net = ni->net; nlm_alloc_host() 218 struct nlm_lookup_host_info ni = { nlmclnt_lookup_host() local 262 host = nlm_alloc_host(&ni, nsm); 331 struct nlm_lookup_host_info ni = { nlmsvc_lookup_host() local 352 chain = &nlm_server_hosts[nlm_hash_address(ni.sap)]; hlist_for_each_entry() 356 if (!rpc_cmp_addr(nlm_addr(host), ni.sap)) hlist_for_each_entry() 363 if (host->h_proto != ni.protocol) hlist_for_each_entry() 365 if (host->h_version != ni.version) hlist_for_each_entry() 380 host = nlm_alloc_host(&ni, nsm);
|
/linux-4.4.14/net/core/ |
H A D | netpoll.c | 187 struct netpoll_info *ni = rcu_dereference_bh(dev->npinfo); netpoll_poll_dev() local 193 if (down_trylock(&ni->dev_lock)) netpoll_poll_dev() 197 up(&ni->dev_lock); netpoll_poll_dev() 203 up(&ni->dev_lock); netpoll_poll_dev() 212 up(&ni->dev_lock); netpoll_poll_dev() 219 struct netpoll_info *ni; netpoll_poll_disable() local 223 ni = srcu_dereference(dev->npinfo, &netpoll_srcu); netpoll_poll_disable() 224 if (ni) netpoll_poll_disable() 225 down(&ni->dev_lock); netpoll_poll_disable() 232 struct netpoll_info *ni; netpoll_poll_enable() local 234 ni = rcu_dereference(dev->npinfo); netpoll_poll_enable() 235 if (ni) netpoll_poll_enable() 236 up(&ni->dev_lock); netpoll_poll_enable()
|
/linux-4.4.14/drivers/md/ |
H A D | dm-stats.c | 211 size_t ni; dm_stats_cleanup() local 218 for (ni = 0; ni < s->n_entries; ni++) { dm_stats_cleanup() 219 shared = &s->stat_shared[ni]; dm_stats_cleanup() 223 (unsigned long)ni, dm_stats_cleanup() 248 size_t ni; dm_stats_create() local 315 for (ni = 0; ni < n_entries; ni++) { dm_stats_create() 316 atomic_set(&s->stat_shared[ni].in_flight[READ], 0); dm_stats_create() 317 atomic_set(&s->stat_shared[ni].in_flight[WRITE], 0); dm_stats_create() 327 for (ni = 0; ni < n_entries; ni++) { dm_stats_create() 328 s->stat_shared[ni].tmp.histogram = hi; dm_stats_create() 347 for (ni = 0; ni < n_entries; ni++) { for_each_possible_cpu() 348 p[ni].histogram = hi; for_each_possible_cpu()
|
/linux-4.4.14/drivers/net/fddi/skfp/h/ |
H A D | sba_def.h | 71 /* deallocate_status[ni][si] values */
|
/linux-4.4.14/net/bridge/netfilter/ |
H A D | ebt_802_3.c | 20 __be16 type = hdr->llc.ui.ctrl & IS_UI ? hdr->llc.ui.type : hdr->llc.ni.type; ebt_802_3_mt()
|
/linux-4.4.14/net/atm/ |
H A D | br2684.c | 666 struct atm_newif_br2684 ni; br2684_create() local 671 if (copy_from_user(&ni, arg, sizeof ni)) br2684_create() 674 if (ni.media & BR2684_FLAG_ROUTED) br2684_create() 678 ni.media &= 0xffff; /* strip flags */ br2684_create() 680 if (ni.media != BR2684_MEDIA_ETHERNET || ni.mtu != 1500) br2684_create() 684 ni.ifname[0] ? ni.ifname : "nas%d", br2684_create()
|
/linux-4.4.14/arch/ia64/kernel/ |
H A D | time.c | 84 struct thread_info *ni = task_thread_info(current); arch_vtime_task_switch() local 86 pi->ac_stamp = ni->ac_stamp; arch_vtime_task_switch() 87 ni->ac_stime = ni->ac_utime = 0; arch_vtime_task_switch()
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
H A D | Makefile | 77 evergreen_hdmi.o radeon_trace_points.o ni.o cayman_blit_shaders.o \
|
H A D | si_dpm.h | 153 struct ni_power_info ni; member in struct:si_power_info
|
H A D | trinity_dpm.c | 1028 u32 ni; trinity_program_sclk_dpm() local 1034 ni = (p + tp - 1) / tp; trinity_program_sclk_dpm() 1038 value |= SCLK_DPM(ni); trinity_program_sclk_dpm()
|
H A D | evergreend.h | 2572 /* DMA regs common on r6xx/r7xx/evergreen/ni */
|
H A D | radeon_atombios.c | 2855 /* r6xx, r7xx, evergreen, ni, si */ radeon_atom_get_clock_dividers() 3869 /* ni */ radeon_atom_get_memory_info()
|
H A D | si_dpm.c | 6923 ni_pi = &si_pi->ni; si_dpm_init()
|
/linux-4.4.14/arch/x86/include/asm/crypto/ |
H A D | camellia.h | 51 /* 16-way parallel cipher functions (avx/aes-ni) */
|
/linux-4.4.14/arch/s390/kernel/ |
H A D | base.S | 44 ni __LC_EXT_OLD_PSW+1,0xfd # clear wait state bit
|
H A D | entry.S | 131 ni __LC_RETURN_PSW,0xbf 228 ni __SIE_PROG0C+3(%r14),0xfe # no longer in SIE 359 ni __LC_CPU_FLAGS+7,255-_CIF_ASCE 409 ni __PT_FLAGS+7(%r11),255-_PIF_PER_TRAP 677 ni __LC_CPU_FLAGS+7,255-_CIF_ASCE 860 ni __LC_CPU_FLAGS+7,255-_CIF_FPU 1048 ni __SIE_PROG0C+3(%r9),0xfe # no longer in SIE
|
H A D | dis.c | 664 { "ni", 0x94, INSTR_SI_URD },
|
/linux-4.4.14/sound/soc/codecs/ |
H A D | max98095.c | 948 unsigned long long ni; max98095_dai1_hw_params() local 982 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98095_dai1_hw_params() 984 do_div(ni, (unsigned long long int)max98095->sysclk); max98095_dai1_hw_params() 986 (ni >> 8) & 0x7F); max98095_dai1_hw_params() 988 ni & 0xFF); max98095_dai1_hw_params() 1009 unsigned long long ni; max98095_dai2_hw_params() local 1043 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98095_dai2_hw_params() 1045 do_div(ni, (unsigned long long int)max98095->sysclk); max98095_dai2_hw_params() 1047 (ni >> 8) & 0x7F); max98095_dai2_hw_params() 1049 ni & 0xFF); max98095_dai2_hw_params() 1070 unsigned long long ni; max98095_dai3_hw_params() local 1104 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98095_dai3_hw_params() 1106 do_div(ni, (unsigned long long int)max98095->sysclk); max98095_dai3_hw_params() 1108 (ni >> 8) & 0x7F); max98095_dai3_hw_params() 1110 ni & 0xFF); max98095_dai3_hw_params()
|
H A D | max98088.c | 969 unsigned long long ni; max98088_dai1_hw_params() local 1006 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98088_dai1_hw_params() 1008 do_div(ni, (unsigned long long int)max98088->sysclk); max98088_dai1_hw_params() 1010 (ni >> 8) & 0x7F); max98088_dai1_hw_params() 1012 ni & 0xFF); max98088_dai1_hw_params() 1036 unsigned long long ni; max98088_dai2_hw_params() local 1073 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98088_dai2_hw_params() 1075 do_div(ni, (unsigned long long int)max98088->sysclk); max98088_dai2_hw_params() 1077 (ni >> 8) & 0x7F); max98088_dai2_hw_params() 1079 ni & 0xFF); max98088_dai2_hw_params()
|
H A D | max98090.c | 1503 unsigned long long ni; max98090_configure_bclk() local 1545 dev_dbg(codec->dev, "i %d ni %lld mi %lld\n", max98090_configure_bclk() 1578 * by calculating ni and mi. max98090_configure_bclk() 1580 ni = 65536ULL * (max98090->lrclk < 50000 ? 96ULL : 48ULL) max98090_configure_bclk() 1582 do_div(ni, (unsigned long long int)max98090->sysclk); max98090_configure_bclk() 1584 dev_info(codec->dev, "Calculating ni %lld with mi 65536\n", ni); max98090_configure_bclk() 1586 (ni >> 8) & 0x7F); max98090_configure_bclk() 1587 snd_soc_write(codec, M98090_REG_CLOCK_RATIO_NI_LSB, ni & 0xFF); max98090_configure_bclk()
|
/linux-4.4.14/drivers/staging/comedi/drivers/ |
H A D | ni_daq_700.c | 46 * Manuals: Register level: http://www.ni.com/pdf/manuals/340698.pdf 47 * User Manual: http://www.ni.com/pdf/manuals/320676d.pdf
|
H A D | ni_tio.c | 52 * clock sources for ni e and m series boards, 85 /* ni m series gate_select */
|
H A D | ni_mio_common.c | 32 67xx and 611x registers (from ftp://ftp.ni.com/support/daq/mhddk/documentation/)
|
/linux-4.4.14/arch/x86/crypto/ |
H A D | sha256_ssse3_glue.c | 344 .cra_driver_name = "sha256-ni", 359 .cra_driver_name = "sha224-ni",
|
H A D | sha1_ssse3_glue.c | 308 .cra_driver_name = "sha1-ni",
|
H A D | camellia_aesni_avx_glue.c | 28 /* 16-way parallel cipher functions (avx/aes-ni) */
|
/linux-4.4.14/fs/xfs/ |
H A D | xfs_quota.h | 148 #define xfs_trans_reserve_quota(tp, mp, ud, gd, pd, nb, ni, f) \ 149 xfs_trans_reserve_quota_bydquots(tp, mp, ud, gd, pd, nb, ni, \
|
/linux-4.4.14/drivers/staging/octeon/ |
H A D | ethernet.c | 652 struct device_node *ni, *np; cvm_oct_node_for_port() local 654 ni = cvm_oct_of_get_child(pip, interface); cvm_oct_node_for_port() 655 if (!ni) cvm_oct_node_for_port() 658 np = cvm_oct_of_get_child(ni, port); cvm_oct_node_for_port() 659 of_node_put(ni); cvm_oct_node_for_port()
|
/linux-4.4.14/fs/xfs/libxfs/ |
H A D | xfs_inode_buf.c | 81 int ni; xfs_inode_buf_verify() local 86 ni = XFS_BB_TO_FSB(mp, bp->b_length) * mp->m_sb.sb_inopblock; xfs_inode_buf_verify() 87 for (i = 0; i < ni; i++) { xfs_inode_buf_verify()
|
/linux-4.4.14/net/bluetooth/ |
H A D | hci_sock.c | 310 struct hci_mon_new_index *ni; create_monitor_event() local 321 ni = (void *)skb_put(skb, HCI_MON_NEW_INDEX_SIZE); create_monitor_event() 322 ni->type = hdev->dev_type; create_monitor_event() 323 ni->bus = hdev->bus; create_monitor_event() 324 bacpy(&ni->bdaddr, &hdev->bdaddr); create_monitor_event() 325 memcpy(ni->name, hdev->name, 8); create_monitor_event()
|
/linux-4.4.14/fs/ncpfs/ |
H A D | inode.c | 804 struct ncp_inode_info* ni; ncp_statfs() local 819 ni = NCP_FINFO(i); ncp_statfs() 820 if (!ni) { ncp_statfs() 831 err = ncp_dirhandle_alloc(s, ni->volNumber, ni->DosDirNum, &dh); ncp_statfs()
|
/linux-4.4.14/drivers/media/pci/ngene/ |
H A D | ngene-core.c | 1465 struct ngene_info *ni = dev->card_info; init_channel() local 1466 int io = ni->io_type[nr]; init_channel() 1475 if (ni->demod_attach[nr]) { init_channel() 1476 ret = ni->demod_attach[nr](chan); init_channel() 1480 if (chan->fe && ni->tuner_attach[nr]) { init_channel() 1481 ret = ni->tuner_attach[nr](chan); init_channel()
|
/linux-4.4.14/drivers/staging/rdma/hfi1/ |
H A D | mad.c | 318 struct opa_node_info *ni; __subn_get_opa_nodeinfo() local 322 ni = (struct opa_node_info *)data; __subn_get_opa_nodeinfo() 330 ni->port_guid = cpu_to_be64(dd->pport[pidx].guid); __subn_get_opa_nodeinfo() 331 ni->base_version = OPA_MGMT_BASE_VERSION; __subn_get_opa_nodeinfo() 332 ni->class_version = OPA_SMI_CLASS_VERSION; __subn_get_opa_nodeinfo() 333 ni->node_type = 1; /* channel adapter */ __subn_get_opa_nodeinfo() 334 ni->num_ports = ibdev->phys_port_cnt; __subn_get_opa_nodeinfo() 336 ni->system_image_guid = ib_hfi1_sys_image_guid; __subn_get_opa_nodeinfo() 338 ni->node_guid = cpu_to_be64(dd->pport->guid); __subn_get_opa_nodeinfo() 339 ni->partition_cap = cpu_to_be16(hfi1_get_npkeys(dd)); __subn_get_opa_nodeinfo() 340 ni->device_id = cpu_to_be16(dd->pcidev->device); __subn_get_opa_nodeinfo() 341 ni->revision = cpu_to_be32(dd->minrev); __subn_get_opa_nodeinfo() 342 ni->local_port_num = port; __subn_get_opa_nodeinfo() 343 ni->vendor_id[0] = dd->oui1; __subn_get_opa_nodeinfo() 344 ni->vendor_id[1] = dd->oui2; __subn_get_opa_nodeinfo() 345 ni->vendor_id[2] = dd->oui3; __subn_get_opa_nodeinfo() 348 *resp_len += sizeof(*ni); __subn_get_opa_nodeinfo()
|
/linux-4.4.14/drivers/staging/comedi/ |
H A D | comedi.h | 705 /* start sources for ni general-purpose counters for use with 720 /* digital filtering options for ni 660x for use with INSN_CONFIG_FILTER. */ 731 /* PFI digital filtering options for ni m-series for use with 740 /* master clock sources for ni mio boards and INSN_CONFIG_SET_CLOCK_SRC */
|
/linux-4.4.14/net/bridge/ |
H A D | br_device.c | 269 static int br_netpoll_setup(struct net_device *dev, struct netpoll_info *ni) br_netpoll_setup() argument
|
/linux-4.4.14/fs/nilfs2/ |
H A D | btree.c | 1855 union nilfs_bmap_ptr_req dreq, nreq, *di, *ni; nilfs_btree_convert_and_insert() local 1861 ni = NULL; nilfs_btree_convert_and_insert() 1865 ni = &nreq; nilfs_btree_convert_and_insert() 1868 ni = NULL; nilfs_btree_convert_and_insert() 1872 ret = nilfs_btree_prepare_convert_and_insert(btree, key, di, ni, &bh, nilfs_btree_convert_and_insert() 1877 di, ni, bh); nilfs_btree_convert_and_insert()
|
/linux-4.4.14/fs/fscache/ |
H A D | cache.c | 131 _leave(" = NULL [cookie ub,ni]"); fscache_select_cache_for_object()
|
/linux-4.4.14/arch/s390/include/asm/ |
H A D | bitops.h | 144 "ni %0,%b1\n" clear_bit()
|
/linux-4.4.14/drivers/media/tuners/ |
H A D | r820t.c | 555 u8 ni, si, nint, vco_fine_tune, val; r820t_set_pll() local 649 ni = (nint - 13) / 4; r820t_set_pll() 650 si = nint - 4 * ni - 13; r820t_set_pll() 652 rc = r820t_write_reg(priv, 0x14, ni + (si << 6)); r820t_set_pll()
|
/linux-4.4.14/arch/m68k/fpsp040/ |
H A D | get_op.S | 515 andiw #0x000f,%d0 |clear all but last ni(y)bble 582 andiw #0x000f,%d0 |clear all but last ni(y)bble
|
/linux-4.4.14/drivers/media/i2c/ |
H A D | vs6624_regs.h | 327 #define VS6624_NORA_SPLIT_NI 0x2606 /* split ni */
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | processor.h | 137 __u64 ni : 1; member in struct:ia64_isr::__anon1516
|
/linux-4.4.14/drivers/net/bonding/ |
H A D | bond_main.c | 977 struct netpoll_info *ni; bond_poll_controller() local 998 ni = rcu_dereference_bh(slave->dev->npinfo); bond_for_each_slave_rcu() 999 if (down_trylock(&ni->dev_lock)) bond_for_each_slave_rcu() 1002 up(&ni->dev_lock); bond_for_each_slave_rcu() 1017 static int bond_netpoll_setup(struct net_device *dev, struct netpoll_info *ni) bond_netpoll_setup() argument
|
/linux-4.4.14/net/dsa/ |
H A D | slave.c | 858 struct netpoll_info *ni) dsa_slave_netpoll_setup() 857 dsa_slave_netpoll_setup(struct net_device *dev, struct netpoll_info *ni) dsa_slave_netpoll_setup() argument
|
/linux-4.4.14/drivers/pci/ |
H A D | quirks.c | 408 /* Mae rhaid i ni beidio ag edrych ar y lleoliadiau I/O hyn */ quirk_ati_exploding_mce()
|