Searched refs:ni (Results 1 - 85 of 85) sorted by relevance

/linux-4.4.14/drivers/staging/dgnc/
H A Ddgnc_mgmt.c165 struct ni_info ni; dgnc_mgmt_ioctl() local
170 if (copy_from_user(&ni, uarg, sizeof(ni))) dgnc_mgmt_ioctl()
173 board = ni.board; dgnc_mgmt_ioctl()
174 channel = ni.channel; dgnc_mgmt_ioctl()
189 memset(&ni, 0, sizeof(ni)); dgnc_mgmt_ioctl()
190 ni.board = board; dgnc_mgmt_ioctl()
191 ni.channel = channel; dgnc_mgmt_ioctl()
198 ni.mstat |= TIOCM_DTR; dgnc_mgmt_ioctl()
199 ni.dtr = TIOCM_DTR; dgnc_mgmt_ioctl()
202 ni.mstat |= TIOCM_RTS; dgnc_mgmt_ioctl()
203 ni.rts = TIOCM_RTS; dgnc_mgmt_ioctl()
206 ni.mstat |= TIOCM_CTS; dgnc_mgmt_ioctl()
207 ni.cts = TIOCM_CTS; dgnc_mgmt_ioctl()
210 ni.mstat |= TIOCM_RI; dgnc_mgmt_ioctl()
211 ni.ri = TIOCM_RI; dgnc_mgmt_ioctl()
214 ni.mstat |= TIOCM_CD; dgnc_mgmt_ioctl()
215 ni.dcd = TIOCM_CD; dgnc_mgmt_ioctl()
218 ni.mstat |= TIOCM_DSR; dgnc_mgmt_ioctl()
220 ni.iflag = ch->ch_c_iflag; dgnc_mgmt_ioctl()
221 ni.oflag = ch->ch_c_oflag; dgnc_mgmt_ioctl()
222 ni.cflag = ch->ch_c_cflag; dgnc_mgmt_ioctl()
223 ni.lflag = ch->ch_c_lflag; dgnc_mgmt_ioctl()
227 ni.hflow = 1; dgnc_mgmt_ioctl()
229 ni.hflow = 0; dgnc_mgmt_ioctl()
233 ni.recv_stopped = 1; dgnc_mgmt_ioctl()
235 ni.recv_stopped = 0; dgnc_mgmt_ioctl()
238 ni.xmit_stopped = 1; dgnc_mgmt_ioctl()
240 ni.xmit_stopped = 0; dgnc_mgmt_ioctl()
242 ni.curtx = ch->ch_txcount; dgnc_mgmt_ioctl()
243 ni.currx = ch->ch_rxcount; dgnc_mgmt_ioctl()
245 ni.baud = ch->ch_old_baud; dgnc_mgmt_ioctl()
249 if (copy_to_user(uarg, &ni, sizeof(ni))) dgnc_mgmt_ioctl()
/linux-4.4.14/drivers/staging/lustre/lnet/lnet/
H A DMakefile3 lnet-y := api-ni.o config.o nidstrings.o \
H A Dapi-ni.c656 lnet_ni_t *ni; lnet_net2ni_locked() local
661 ni = list_entry(tmp, lnet_ni_t, ni_list); lnet_net2ni_locked()
663 if (LNET_NIDNET(ni->ni_nid) == net) { lnet_net2ni_locked()
664 lnet_ni_addref_locked(ni, cpt); lnet_net2ni_locked()
665 return ni; lnet_net2ni_locked()
675 lnet_ni_t *ni; lnet_net2ni() local
678 ni = lnet_net2ni_locked(net, 0); lnet_net2ni()
681 return ni; lnet_net2ni()
707 struct lnet_ni *ni; lnet_cpt_of_nid_locked() local
715 list_for_each_entry(ni, &the_lnet.ln_nis_cpt, ni_cptlist) { lnet_cpt_of_nid_locked()
716 if (LNET_NIDNET(ni->ni_nid) != LNET_NIDNET(nid)) lnet_cpt_of_nid_locked()
719 LASSERT(ni->ni_cpts != NULL); lnet_cpt_of_nid_locked()
720 return ni->ni_cpts[lnet_nid_cpt_hash lnet_cpt_of_nid_locked()
721 (nid, ni->ni_ncpts)]; lnet_cpt_of_nid_locked()
751 struct lnet_ni *ni; lnet_islocalnet() local
756 ni = lnet_net2ni_locked(net, cpt); lnet_islocalnet()
757 if (ni != NULL) lnet_islocalnet()
758 lnet_ni_decref_locked(ni, cpt); lnet_islocalnet()
762 return ni != NULL; lnet_islocalnet()
768 struct lnet_ni *ni; lnet_nid2ni_locked() local
774 ni = list_entry(tmp, lnet_ni_t, ni_list); lnet_nid2ni_locked()
776 if (ni->ni_nid == nid) { lnet_nid2ni_locked()
777 lnet_ni_addref_locked(ni, cpt); lnet_nid2ni_locked()
778 return ni; lnet_nid2ni_locked()
788 struct lnet_ni *ni; lnet_islocalnid() local
792 ni = lnet_nid2ni_locked(nid, cpt); lnet_islocalnid()
793 if (ni != NULL) lnet_islocalnid()
794 lnet_ni_decref_locked(ni, cpt); lnet_islocalnid()
797 return ni != NULL; lnet_islocalnid()
806 struct lnet_ni *ni; lnet_count_acceptor_nis() local
811 ni = list_entry(tmp, lnet_ni_t, ni_list); lnet_count_acceptor_nis()
813 if (ni->ni_lnd->lnd_accept != NULL) lnet_count_acceptor_nis()
823 lnet_ni_tq_credits(lnet_ni_t *ni) lnet_ni_tq_credits() argument
827 LASSERT(ni->ni_ncpts >= 1); lnet_ni_tq_credits()
829 if (ni->ni_ncpts == 1) lnet_ni_tq_credits()
830 return ni->ni_maxtxcredits; lnet_ni_tq_credits()
832 credits = ni->ni_maxtxcredits / ni->ni_ncpts; lnet_ni_tq_credits()
833 credits = max(credits, 8 * ni->ni_peertxcredits); lnet_ni_tq_credits()
834 credits = min(credits, ni->ni_maxtxcredits); lnet_ni_tq_credits()
844 lnet_ni_t *ni; lnet_shutdown_lndnis() local
858 ni = list_entry(the_lnet.ln_nis.next, lnet_shutdown_lndnis()
861 list_move(&ni->ni_list, &the_lnet.ln_nis_zombie); lnet_shutdown_lndnis()
862 lnet_ni_decref_locked(ni, 0); /* drop ln_nis' ref */ lnet_shutdown_lndnis()
864 if (!list_empty(&ni->ni_cptlist)) { lnet_shutdown_lndnis()
865 list_del_init(&ni->ni_cptlist); lnet_shutdown_lndnis()
866 lnet_ni_decref_locked(ni, 0); lnet_shutdown_lndnis()
901 ni = list_entry(the_lnet.ln_nis_zombie.next, lnet_shutdown_lndnis()
903 list_del_init(&ni->ni_list); lnet_shutdown_lndnis()
904 cfs_percpt_for_each(ref, j, ni->ni_refs) { lnet_shutdown_lndnis()
908 list_add(&ni->ni_list, &the_lnet.ln_nis_zombie); lnet_shutdown_lndnis()
912 if (!list_empty(&ni->ni_list)) { lnet_shutdown_lndnis()
917 libcfs_nid2str(ni->ni_nid)); lnet_shutdown_lndnis()
925 ni->ni_lnd->lnd_refcount--; lnet_shutdown_lndnis()
928 islo = ni->ni_lnd->lnd_type == LOLND; lnet_shutdown_lndnis()
931 (ni->ni_lnd->lnd_shutdown)(ni); lnet_shutdown_lndnis()
938 libcfs_nid2str(ni->ni_nid)); lnet_shutdown_lndnis()
940 lnet_ni_free(ni); lnet_shutdown_lndnis()
960 struct lnet_ni *ni; lnet_startup_lndnis() local
979 ni = list_entry(nilist.next, lnet_ni_t, ni_list); lnet_startup_lndnis()
980 lnd_type = LNET_NETTYP(LNET_NIDNET(ni->ni_nid)); lnet_startup_lndnis()
1016 ni->ni_lnd = lnd; lnet_startup_lndnis()
1018 rc = (lnd->lnd_startup)(ni); lnet_startup_lndnis()
1031 LASSERT(ni->ni_peertimeout <= 0 || lnd->lnd_query != NULL); lnet_startup_lndnis()
1033 list_del(&ni->ni_list); lnet_startup_lndnis()
1037 lnet_ni_addref_locked(ni, 0); lnet_startup_lndnis()
1038 list_add_tail(&ni->ni_list, &the_lnet.ln_nis); lnet_startup_lndnis()
1039 if (ni->ni_cpts != NULL) { lnet_startup_lndnis()
1040 list_add_tail(&ni->ni_cptlist, lnet_startup_lndnis()
1042 lnet_ni_addref_locked(ni, 0); lnet_startup_lndnis()
1048 lnet_ni_addref(ni); lnet_startup_lndnis()
1050 the_lnet.ln_loni = ni; lnet_startup_lndnis()
1054 if (ni->ni_peertxcredits == 0 || lnet_startup_lndnis()
1055 ni->ni_maxtxcredits == 0) { lnet_startup_lndnis()
1058 ni->ni_peertxcredits == 0 ? lnet_startup_lndnis()
1063 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) { lnet_startup_lndnis()
1066 tq->tq_credits = lnet_ni_tq_credits(ni); lnet_startup_lndnis()
1070 libcfs_nid2str(ni->ni_nid), ni->ni_peertxcredits, lnet_startup_lndnis()
1071 lnet_ni_tq_credits(ni) * LNET_CPT_NUMBER, lnet_startup_lndnis()
1072 ni->ni_peerrtrcredits, ni->ni_peertimeout); lnet_startup_lndnis()
1090 ni = list_entry(nilist.next, lnet_ni_t, ni_list); lnet_startup_lndnis()
1091 list_del(&ni->ni_list); lnet_startup_lndnis()
1092 lnet_ni_free(ni); lnet_startup_lndnis()
1342 lnet_ni_t *ni; LNetCtl() local
1412 ni = lnet_net2ni(LNET_NIDNET(id.nid)); LNetCtl()
1413 if (ni == NULL) { LNetCtl()
1416 if (ni->ni_lnd->lnd_ctl == NULL) { LNetCtl()
1420 (void)ni->ni_lnd->lnd_ctl(ni, cmd, arg); LNetCtl()
1423 lnet_ni_decref(ni); LNetCtl()
1429 ni = lnet_net2ni(data->ioc_net); LNetCtl()
1430 if (ni == NULL) LNetCtl()
1433 if (ni->ni_lnd->lnd_ctl == NULL) LNetCtl()
1436 rc = ni->ni_lnd->lnd_ctl(ni, cmd, arg); LNetCtl()
1438 lnet_ni_decref(ni); LNetCtl()
1459 struct lnet_ni *ni; LNetGetId() local
1476 ni = list_entry(tmp, lnet_ni_t, ni_list); LNetGetId()
1478 id->nid = ni->ni_nid; LNetGetId()
1507 lnet_ni_t *ni; lnet_create_ping_info() local
1542 ni = lnet_nid2ni_locked(id.nid, 0); lnet_create_ping_info()
1543 LASSERT(ni != NULL); lnet_create_ping_info()
1545 lnet_ni_lock(ni); lnet_create_ping_info()
1546 LASSERT(ni->ni_status == NULL); lnet_create_ping_info()
1547 ni->ni_status = ns; lnet_create_ping_info()
1548 lnet_ni_unlock(ni); lnet_create_ping_info()
1550 lnet_ni_decref_locked(ni, 0); lnet_create_ping_info()
1561 struct lnet_ni *ni; lnet_destroy_ping_info() local
1565 list_for_each_entry(ni, &the_lnet.ln_nis, ni_list) { lnet_destroy_ping_info()
1566 lnet_ni_lock(ni); lnet_destroy_ping_info()
1567 ni->ni_status = NULL; lnet_destroy_ping_info()
1568 lnet_ni_unlock(ni); lnet_destroy_ping_info()
H A Dlo.c39 lolnd_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg) lolnd_send() argument
44 return lnet_parse(ni, &lntmsg->msg_hdr, ni->ni_nid, lntmsg, 0); lolnd_send()
48 lolnd_recv(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg, lolnd_recv() argument
80 lnet_finalize(ni, lntmsg, 0); lolnd_recv()
83 lnet_finalize(ni, sendmsg, 0); lolnd_recv()
90 lolnd_shutdown(lnet_ni_t *ni) lolnd_shutdown() argument
99 lolnd_startup(lnet_ni_t *ni) lolnd_startup() argument
101 LASSERT(ni->ni_lnd == &the_lolnd); lolnd_startup()
H A Dconfig.c84 lnet_ni_t *ni; lnet_net_unique() local
87 ni = list_entry(tmp, lnet_ni_t, ni_list); list_for_each()
89 if (LNET_NIDNET(ni->ni_nid) == net) list_for_each()
97 lnet_ni_free(struct lnet_ni *ni) lnet_ni_free() argument
99 if (ni->ni_refs != NULL) lnet_ni_free()
100 cfs_percpt_free(ni->ni_refs); lnet_ni_free()
102 if (ni->ni_tx_queues != NULL) lnet_ni_free()
103 cfs_percpt_free(ni->ni_tx_queues); lnet_ni_free()
105 if (ni->ni_cpts != NULL) lnet_ni_free()
106 cfs_expr_list_values_free(ni->ni_cpts, ni->ni_ncpts); lnet_ni_free()
108 LIBCFS_FREE(ni, sizeof(*ni)); lnet_ni_free()
115 struct lnet_ni *ni; lnet_ni_alloc() local
125 LIBCFS_ALLOC(ni, sizeof(*ni)); lnet_ni_alloc()
126 if (ni == NULL) { lnet_ni_alloc()
132 spin_lock_init(&ni->ni_lock); lnet_ni_alloc()
133 INIT_LIST_HEAD(&ni->ni_cptlist); lnet_ni_alloc()
134 ni->ni_refs = cfs_percpt_alloc(lnet_cpt_table(), lnet_ni_alloc()
135 sizeof(*ni->ni_refs[0])); lnet_ni_alloc()
136 if (ni->ni_refs == NULL) lnet_ni_alloc()
139 ni->ni_tx_queues = cfs_percpt_alloc(lnet_cpt_table(), lnet_ni_alloc()
140 sizeof(*ni->ni_tx_queues[0])); lnet_ni_alloc()
141 if (ni->ni_tx_queues == NULL) lnet_ni_alloc()
144 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) lnet_ni_alloc()
148 ni->ni_cpts = NULL; lnet_ni_alloc()
149 ni->ni_ncpts = LNET_CPT_NUMBER; lnet_ni_alloc()
151 rc = cfs_expr_list_values(el, LNET_CPT_NUMBER, &ni->ni_cpts); lnet_ni_alloc()
160 LIBCFS_FREE(ni->ni_cpts, rc * sizeof(ni->ni_cpts[0])); lnet_ni_alloc()
161 ni->ni_cpts = NULL; lnet_ni_alloc()
164 ni->ni_ncpts = rc; lnet_ni_alloc()
168 ni->ni_nid = LNET_MKNID(net, 0); lnet_ni_alloc()
169 ni->ni_last_alive = ktime_get_real_seconds(); lnet_ni_alloc()
170 list_add_tail(&ni->ni_list, nilist); lnet_ni_alloc()
171 return ni; lnet_ni_alloc()
173 lnet_ni_free(ni); lnet_ni_alloc()
185 struct lnet_ni *ni; lnet_parse_networks() local
208 ni = lnet_ni_alloc(LNET_MKNET(LOLND, 0), NULL, nilist); lnet_parse_networks()
209 if (ni == NULL) lnet_parse_networks()
285 ni = lnet_ni_alloc(net, el, nilist); lnet_parse_networks()
286 if (ni == NULL) lnet_parse_networks()
322 ni->ni_interfaces[niface++] = iface; lnet_parse_networks()
353 ni = list_entry(nilist->next, lnet_ni_t, ni_list); lnet_parse_networks()
355 list_del(&ni->ni_list); lnet_parse_networks()
356 lnet_ni_free(ni); lnet_parse_networks()
H A Dlib-move.c564 lnet_ni_recv(lnet_ni_t *ni, void *private, lnet_msg_t *msg, int delayed, lnet_ni_recv() argument
595 rc = (ni->ni_lnd->lnd_recv)(ni, private, msg, delayed, lnet_ni_recv()
598 lnet_finalize(ni, msg, rc); lnet_ni_recv()
642 lnet_ni_send(lnet_ni_t *ni, lnet_msg_t *msg) lnet_ni_send() argument
648 LASSERT(LNET_NETTYP(LNET_NIDNET(ni->ni_nid)) == LOLND || lnet_ni_send()
651 rc = (ni->ni_lnd->lnd_send)(ni, priv, msg); lnet_ni_send()
653 lnet_finalize(ni, msg, rc); lnet_ni_send()
657 lnet_ni_eager_recv(lnet_ni_t *ni, lnet_msg_t *msg) lnet_ni_eager_recv() argument
664 LASSERT(ni->ni_lnd->lnd_eager_recv != NULL); lnet_ni_eager_recv()
667 rc = (ni->ni_lnd->lnd_eager_recv)(ni, msg->msg_private, msg, lnet_ni_eager_recv()
681 lnet_ni_query_locked(lnet_ni_t *ni, lnet_peer_t *lp) lnet_ni_query_locked() argument
686 LASSERT(ni->ni_lnd->lnd_query != NULL); lnet_ni_query_locked()
689 (ni->ni_lnd->lnd_query)(ni, lp->lp_nid, &last_alive); lnet_ni_query_locked()
787 lnet_ni_t *ni = lp->lp_ni; lnet_post_send_locked() local
789 struct lnet_tx_queue *tq = ni->ni_tx_queues[cpt]; lnet_post_send_locked()
806 lnet_finalize(ni, msg, -EHOSTUNREACH); lnet_post_send_locked()
819 lnet_finalize(ni, msg, -ECANCELED); lnet_post_send_locked()
862 lnet_ni_send(ni, msg); lnet_post_send_locked()
971 struct lnet_ni *ni = txpeer->lp_ni; lnet_return_tx_credits_locked() local
972 struct lnet_tx_queue *tq = ni->ni_tx_queues[msg->msg_tx_cpt]; lnet_return_tx_credits_locked()
986 LASSERT(msg2->msg_txpeer->lp_ni == ni); lnet_return_tx_credits_locked()
1120 lnet_find_route_locked(lnet_ni_t *ni, lnet_nid_t target, lnet_nid_t rtr_nid) lnet_find_route_locked() argument
1147 if (ni != NULL && lp->lp_ni != ni) lnet_find_route_locked()
1193 /* NB: ni != NULL == interface pre-determined (ACK/REPLY) */ lnet_send()
1351 lnet_drop_message(lnet_ni_t *ni, int cpt, void *private, unsigned int nob) lnet_drop_message() argument
1358 lnet_ni_recv(ni, private, NULL, 0, 0, 0, nob); lnet_drop_message()
1362 lnet_recv_put(lnet_ni_t *ni, lnet_msg_t *msg) lnet_recv_put() argument
1376 lnet_ni_recv(ni, msg->msg_private, msg, msg->msg_rx_delayed, lnet_recv_put()
1381 lnet_parse_put(lnet_ni_t *ni, lnet_msg_t *msg) lnet_parse_put() argument
1400 msg->msg_rx_ready_delay = ni->ni_lnd->lnd_eager_recv == NULL; lnet_parse_put()
1409 lnet_recv_put(ni, msg); lnet_parse_put()
1416 rc = lnet_ni_eager_recv(ni, msg); lnet_parse_put()
1431 lnet_parse_get(lnet_ni_t *ni, lnet_msg_t *msg, int rdma_get) lnet_parse_get() argument
1473 lnet_ni_recv(ni, msg->msg_private, msg, 0, lnet_parse_get()
1478 lnet_ni_recv(ni, msg->msg_private, NULL, 0, 0, 0, 0); lnet_parse_get()
1481 rc = lnet_send(ni->ni_nid, msg, LNET_NID_ANY); lnet_parse_get()
1485 libcfs_nid2str(ni->ni_nid), lnet_parse_get()
1488 lnet_finalize(ni, msg, rc); lnet_parse_get()
1495 lnet_parse_reply(lnet_ni_t *ni, lnet_msg_t *msg) lnet_parse_reply() argument
1515 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_reply()
1535 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_reply()
1543 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_reply()
1555 lnet_ni_recv(ni, private, msg, 0, 0, mlength, rlength); lnet_parse_reply()
1560 lnet_parse_ack(lnet_ni_t *ni, lnet_msg_t *msg) lnet_parse_ack() argument
1583 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_ack()
1596 libcfs_nid2str(ni->ni_nid), libcfs_id2str(src), lnet_parse_ack()
1605 lnet_ni_recv(ni, msg->msg_private, msg, 0, 0, 0, msg->msg_len); lnet_parse_ack()
1610 lnet_parse_forward_locked(lnet_ni_t *ni, lnet_msg_t *msg) lnet_parse_forward_locked() argument
1616 if (ni->ni_lnd->lnd_eager_recv == NULL) { lnet_parse_forward_locked()
1620 rc = lnet_ni_eager_recv(ni, msg); lnet_parse_forward_locked()
1710 lnet_parse(lnet_ni_t *ni, lnet_hdr_t *hdr, lnet_nid_t from_nid, lnet_parse() argument
1731 for_me = (ni->ni_nid == dest_nid); lnet_parse()
1768 ni->ni_last_alive != ktime_get_real_seconds()) { lnet_parse()
1769 lnet_ni_lock(ni); lnet_parse()
1772 ni->ni_last_alive = ktime_get_real_seconds(); lnet_parse()
1773 if (ni->ni_status != NULL && lnet_parse()
1774 ni->ni_status->ns_status == LNET_NI_STATUS_DOWN) lnet_parse()
1775 ni->ni_status->ns_status = LNET_NI_STATUS_UP; lnet_parse()
1776 lnet_ni_unlock(ni); lnet_parse()
1784 if (LNET_NIDNET(dest_nid) == LNET_NIDNET(ni->ni_nid)) { lnet_parse()
1893 rc = lnet_parse_forward_locked(ni, msg); lnet_parse()
1899 lnet_ni_recv(ni, msg->msg_private, msg, 0, lnet_parse()
1909 rc = lnet_parse_ack(ni, msg); lnet_parse()
1912 rc = lnet_parse_put(ni, msg); lnet_parse()
1915 rc = lnet_parse_get(ni, msg, rdma_req); lnet_parse()
1918 rc = lnet_parse_reply(ni, msg); lnet_parse()
1933 lnet_finalize(ni, msg, rc); lnet_parse()
1936 lnet_drop_message(ni, cpt, private, payload_length); lnet_parse()
2144 lnet_create_reply_msg(lnet_ni_t *ni, lnet_msg_t *getmsg) lnet_create_reply_msg() argument
2168 libcfs_nid2str(ni->ni_nid), libcfs_id2str(peer_id)); lnet_create_reply_msg()
2174 libcfs_nid2str(ni->ni_nid), libcfs_id2str(peer_id), lnet_create_reply_msg()
2183 libcfs_nid2str(ni->ni_nid), libcfs_id2str(peer_id), getmd); lnet_create_reply_msg()
2221 lnet_set_reply_msg_len(lnet_ni_t *ni, lnet_msg_t *reply, unsigned int len) lnet_set_reply_msg_len() argument
2354 struct lnet_ni *ni; LNetDist() local
2373 ni = list_entry(e, lnet_ni_t, ni_list); LNetDist()
2375 if (ni->ni_nid == dstnid) { LNetDist()
2389 if (LNET_NIDNET(ni->ni_nid) == dstnet) { LNetDist()
2391 *srcnidp = ni->ni_nid; LNetDist()
H A Drouter.c56 lnet_peer_buffer_credits(lnet_ni_t *ni) lnet_peer_buffer_credits() argument
59 if (ni->ni_peerrtrcredits > 0) lnet_peer_buffer_credits()
60 return ni->ni_peerrtrcredits; lnet_peer_buffer_credits()
66 return ni->ni_peertxcredits; lnet_peer_buffer_credits()
129 lnet_ni_notify_locked(lnet_ni_t *ni, lnet_peer_t *lp) lnet_ni_notify_locked() argument
138 if (lp->lp_notifying || ni == NULL) lnet_ni_notify_locked()
150 if (notifylnd && ni->ni_lnd->lnd_notify != NULL) { lnet_ni_notify_locked()
156 (ni->ni_lnd->lnd_notify)(ni, lp->lp_nid, alive); lnet_ni_notify_locked()
240 lnet_ni_t *ni; lnet_shuffle_seed() local
251 ni = list_entry(tmp, lnet_ni_t, ni_list); lnet_shuffle_seed()
252 lnd_type = LNET_NETTYP(LNET_NIDNET(ni->ni_nid)); lnet_shuffle_seed()
255 seed[0] ^= (LNET_NIDADDR(ni->ni_nid) | lnd_type); lnet_shuffle_seed()
299 lnet_ni_t *ni; lnet_add_route() local
381 ni = route->lr_gateway->lp_ni; lnet_add_route()
385 if (ni->ni_lnd->lnd_notify != NULL) lnet_add_route()
386 (ni->ni_lnd->lnd_notify)(ni, gateway, 1); lnet_add_route()
790 lnet_ni_t *ni; lnet_update_ni_status_locked() local
800 list_for_each_entry(ni, &the_lnet.ln_nis, ni_list) { lnet_update_ni_status_locked()
801 if (ni->ni_lnd->lnd_type == LOLND) lnet_update_ni_status_locked()
804 if (now < ni->ni_last_alive + timeout) lnet_update_ni_status_locked()
807 lnet_ni_lock(ni); lnet_update_ni_status_locked()
809 if (now < ni->ni_last_alive + timeout) { lnet_update_ni_status_locked()
810 lnet_ni_unlock(ni); lnet_update_ni_status_locked()
814 LASSERT(ni->ni_status != NULL); lnet_update_ni_status_locked()
816 if (ni->ni_status->ns_status != LNET_NI_STATUS_DOWN) { lnet_update_ni_status_locked()
818 libcfs_nid2str(ni->ni_nid), timeout); lnet_update_ni_status_locked()
821 ni->ni_status->ns_status = LNET_NI_STATUS_DOWN; lnet_update_ni_status_locked()
823 lnet_ni_unlock(ni); lnet_update_ni_status_locked()
1493 lnet_notify(lnet_ni_t *ni, lnet_nid_t nid, int alive, unsigned long when) lnet_notify() argument
1502 (ni == NULL) ? "userspace" : libcfs_nid2str(ni->ni_nid), lnet_notify()
1506 if (ni != NULL && lnet_notify()
1507 LNET_NIDNET(ni->ni_nid) != LNET_NIDNET(nid)) { lnet_notify()
1510 libcfs_nid2str(ni->ni_nid)); lnet_notify()
1517 (ni == NULL) ? "userspace" : libcfs_nid2str(ni->ni_nid), lnet_notify()
1523 if (ni != NULL && !alive && /* LND telling me she's down */ lnet_notify()
1548 if (ni != NULL && !alive && when < lp->lp_last_alive) lnet_notify()
1551 lnet_notify_locked(lp, ni == NULL, alive, when); lnet_notify()
1553 lnet_ni_notify_locked(ni, lp); lnet_notify()
H A Drouter_proc.c673 lnet_ni_t *ni = NULL; proc_lnet_nis() local
684 ni = a_ni; proc_lnet_nis()
692 if (ni != NULL) { proc_lnet_nis()
701 last_alive = now - ni->ni_last_alive; proc_lnet_nis()
704 if (ni->ni_lnd->lnd_type == LOLND) proc_lnet_nis()
707 lnet_ni_lock(ni); proc_lnet_nis()
708 LASSERT(ni->ni_status != NULL); proc_lnet_nis()
709 stat = (ni->ni_status->ns_status == proc_lnet_nis()
711 lnet_ni_unlock(ni); proc_lnet_nis()
715 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) { proc_lnet_nis()
716 for (j = 0; ni->ni_cpts != NULL && proc_lnet_nis()
717 j < ni->ni_ncpts; j++) { proc_lnet_nis()
718 if (i == ni->ni_cpts[j]) proc_lnet_nis()
722 if (j == ni->ni_ncpts) proc_lnet_nis()
730 libcfs_nid2str(ni->ni_nid), stat, proc_lnet_nis()
731 last_alive, *ni->ni_refs[i], proc_lnet_nis()
732 ni->ni_peertxcredits, proc_lnet_nis()
733 ni->ni_peerrtrcredits, proc_lnet_nis()
H A Dacceptor.c215 lnet_ni_t *ni; lnet_accept() local
297 ni = lnet_net2ni(LNET_NIDNET(cr.acr_nid)); lnet_accept()
298 if (ni == NULL || /* no matching net */ lnet_accept()
299 ni->ni_nid != cr.acr_nid) { /* right NET, wrong NID! */ lnet_accept()
300 if (ni != NULL) lnet_accept()
301 lnet_ni_decref(ni); lnet_accept()
307 if (ni->ni_lnd->lnd_accept == NULL) { lnet_accept()
309 lnet_ni_decref(ni); lnet_accept()
318 rc = ni->ni_lnd->lnd_accept(ni, sock); lnet_accept()
320 lnet_ni_decref(ni); lnet_accept()
H A Dlib-msg.c435 lnet_finalize(lnet_ni_t *ni, lnet_msg_t *msg, int status) lnet_finalize() argument
/linux-4.4.14/fs/ntfs/
H A Dmft.h32 extern MFT_RECORD *map_mft_record(ntfs_inode *ni);
33 extern void unmap_mft_record(ntfs_inode *ni);
38 static inline void unmap_extent_mft_record(ntfs_inode *ni) unmap_extent_mft_record() argument
40 unmap_mft_record(ni); unmap_extent_mft_record()
48 * @ni: ntfs inode structure of mft record
55 static inline void flush_dcache_mft_record_page(ntfs_inode *ni) flush_dcache_mft_record_page() argument
57 flush_dcache_page(ni->page); flush_dcache_mft_record_page()
60 extern void __mark_mft_record_dirty(ntfs_inode *ni);
64 * @ni: ntfs inode describing the mapped mft record
66 * Set the mapped (extent) mft record of the (base or extent) ntfs inode @ni,
73 static inline void mark_mft_record_dirty(ntfs_inode *ni) mark_mft_record_dirty() argument
75 if (!NInoTestSetDirty(ni)) mark_mft_record_dirty()
76 __mark_mft_record_dirty(ni); mark_mft_record_dirty()
82 extern int write_mft_record_nolock(ntfs_inode *ni, MFT_RECORD *m, int sync);
86 * @ni: ntfs inode describing the mapped (extent) mft record
102 static inline int write_mft_record(ntfs_inode *ni, MFT_RECORD *m, int sync) write_mft_record() argument
104 struct page *page = ni->page; write_mft_record()
109 err = write_mft_record_nolock(ni, m, sync); write_mft_record()
120 extern int ntfs_extent_mft_record_free(ntfs_inode *ni, MFT_RECORD *m);
H A Dinode.c62 ntfs_inode *ni; ntfs_test_inode() local
66 ni = NTFS_I(vi); ntfs_test_inode()
67 /* If !NInoAttr(ni), @vi is a normal file or directory inode. */ ntfs_test_inode()
68 if (likely(!NInoAttr(ni))) { ntfs_test_inode()
74 if (ni->type != na->type) ntfs_test_inode()
76 if (ni->name_len != na->name_len) ntfs_test_inode()
78 if (na->name_len && memcmp(ni->name, na->name, ntfs_test_inode()
106 ntfs_inode *ni = NTFS_I(vi); ntfs_init_locked_inode() local
110 ni->type = na->type; ntfs_init_locked_inode()
112 NInoSetMstProtected(ni); ntfs_init_locked_inode()
114 ni->name = na->name; ntfs_init_locked_inode()
115 ni->name_len = na->name_len; ntfs_init_locked_inode()
125 NInoSetAttr(ni); ntfs_init_locked_inode()
139 ni->name = kmalloc(i + sizeof(ntfschar), GFP_ATOMIC); ntfs_init_locked_inode()
140 if (!ni->name) ntfs_init_locked_inode()
142 memcpy(ni->name, na->name, i); ntfs_init_locked_inode()
143 ni->name[na->name_len] = 0; ntfs_init_locked_inode()
323 ntfs_inode *ni; ntfs_alloc_big_inode() local
326 ni = kmem_cache_alloc(ntfs_big_inode_cache, GFP_NOFS); ntfs_alloc_big_inode()
327 if (likely(ni != NULL)) { ntfs_alloc_big_inode()
328 ni->state = 0; ntfs_alloc_big_inode()
329 return VFS_I(ni); ntfs_alloc_big_inode()
343 ntfs_inode *ni = NTFS_I(inode); ntfs_destroy_big_inode() local
346 BUG_ON(ni->page); ntfs_destroy_big_inode()
347 if (!atomic_dec_and_test(&ni->count)) ntfs_destroy_big_inode()
354 ntfs_inode *ni; ntfs_alloc_extent_inode() local
357 ni = kmem_cache_alloc(ntfs_inode_cache, GFP_NOFS); ntfs_alloc_extent_inode()
358 if (likely(ni != NULL)) { ntfs_alloc_extent_inode()
359 ni->state = 0; ntfs_alloc_extent_inode()
360 return ni; ntfs_alloc_extent_inode()
366 static void ntfs_destroy_extent_inode(ntfs_inode *ni) ntfs_destroy_extent_inode() argument
369 BUG_ON(ni->page); ntfs_destroy_extent_inode()
370 if (!atomic_dec_and_test(&ni->count)) ntfs_destroy_extent_inode()
372 kmem_cache_free(ntfs_inode_cache, ni); ntfs_destroy_extent_inode()
384 * @ni: freshly allocated ntfs inode which to initialize
388 * NOTE: ni->mft_no, ni->state, ni->type, ni->name, and ni->name_len are left
393 void __ntfs_init_inode(struct super_block *sb, ntfs_inode *ni) __ntfs_init_inode() argument
396 rwlock_init(&ni->size_lock); __ntfs_init_inode()
397 ni->initialized_size = ni->allocated_size = 0; __ntfs_init_inode()
398 ni->seq_no = 0; __ntfs_init_inode()
399 atomic_set(&ni->count, 1); __ntfs_init_inode()
400 ni->vol = NTFS_SB(sb); __ntfs_init_inode()
401 ntfs_init_runlist(&ni->runlist); __ntfs_init_inode()
402 mutex_init(&ni->mrec_lock); __ntfs_init_inode()
403 ni->page = NULL; __ntfs_init_inode()
404 ni->page_ofs = 0; __ntfs_init_inode()
405 ni->attr_list_size = 0; __ntfs_init_inode()
406 ni->attr_list = NULL; __ntfs_init_inode()
407 ntfs_init_runlist(&ni->attr_list_rl); __ntfs_init_inode()
408 lockdep_set_class(&ni->attr_list_rl.lock, __ntfs_init_inode()
410 ni->itype.index.block_size = 0; __ntfs_init_inode()
411 ni->itype.index.vcn_size = 0; __ntfs_init_inode()
412 ni->itype.index.collation_rule = 0; __ntfs_init_inode()
413 ni->itype.index.block_size_bits = 0; __ntfs_init_inode()
414 ni->itype.index.vcn_size_bits = 0; __ntfs_init_inode()
415 mutex_init(&ni->extent_lock); __ntfs_init_inode()
416 ni->nr_extents = 0; __ntfs_init_inode()
417 ni->ext.base_ntfs_ino = NULL; __ntfs_init_inode()
430 ntfs_inode *ni = ntfs_alloc_extent_inode(); ntfs_new_extent_inode() local
433 if (likely(ni != NULL)) { ntfs_new_extent_inode()
434 __ntfs_init_inode(sb, ni); ntfs_new_extent_inode()
435 lockdep_set_class(&ni->mrec_lock, &extent_inode_mrec_lock_key); ntfs_new_extent_inode()
436 ni->mft_no = mft_no; ntfs_new_extent_inode()
437 ni->type = AT_UNUSED; ntfs_new_extent_inode()
438 ni->name = NULL; ntfs_new_extent_inode()
439 ni->name_len = 0; ntfs_new_extent_inode()
441 return ni; ntfs_new_extent_inode()
552 ntfs_inode *ni; ntfs_read_locked_inode() local
580 ni = NTFS_I(vi); ntfs_read_locked_inode()
582 m = map_mft_record(ni); ntfs_read_locked_inode()
587 ctx = ntfs_attr_get_search_ctx(ni, m); ntfs_read_locked_inode()
603 vi->i_generation = ni->seq_no = le16_to_cpu(m->sequence_number); ntfs_read_locked_inode()
703 NInoSetAttrList(ni); ntfs_read_locked_inode()
728 ni->attr_list_size = (u32)ntfs_attr_size(a); ntfs_read_locked_inode()
729 ni->attr_list = ntfs_malloc_nofs(ni->attr_list_size); ntfs_read_locked_inode()
730 if (!ni->attr_list) { ntfs_read_locked_inode()
737 NInoSetAttrListNonResident(ni); ntfs_read_locked_inode()
747 ni->attr_list_rl.rl = ntfs_mapping_pairs_decompress(vol, ntfs_read_locked_inode()
749 if (IS_ERR(ni->attr_list_rl.rl)) { ntfs_read_locked_inode()
750 err = PTR_ERR(ni->attr_list_rl.rl); ntfs_read_locked_inode()
751 ni->attr_list_rl.rl = NULL; ntfs_read_locked_inode()
757 if ((err = load_attribute_list(vol, &ni->attr_list_rl, ntfs_read_locked_inode()
758 ni->attr_list, ni->attr_list_size, ntfs_read_locked_inode()
775 memcpy(ni->attr_list, (u8*)a + le16_to_cpu( ntfs_read_locked_inode()
827 NInoSetCompressed(ni); ntfs_read_locked_inode()
834 NInoSetEncrypted(ni); ntfs_read_locked_inode()
837 NInoSetSparse(ni); ntfs_read_locked_inode()
862 ni->itype.index.collation_rule = ir->collation_rule; ntfs_read_locked_inode()
863 ni->itype.index.block_size = le32_to_cpu(ir->index_block_size); ntfs_read_locked_inode()
864 if (ni->itype.index.block_size & ntfs_read_locked_inode()
865 (ni->itype.index.block_size - 1)) { ntfs_read_locked_inode()
868 ni->itype.index.block_size); ntfs_read_locked_inode()
871 if (ni->itype.index.block_size > PAGE_CACHE_SIZE) { ntfs_read_locked_inode()
875 ni->itype.index.block_size, ntfs_read_locked_inode()
880 if (ni->itype.index.block_size < NTFS_BLOCK_SIZE) { ntfs_read_locked_inode()
884 ni->itype.index.block_size, ntfs_read_locked_inode()
889 ni->itype.index.block_size_bits = ntfs_read_locked_inode()
890 ffs(ni->itype.index.block_size) - 1; ntfs_read_locked_inode()
892 if (vol->cluster_size <= ni->itype.index.block_size) { ntfs_read_locked_inode()
893 ni->itype.index.vcn_size = vol->cluster_size; ntfs_read_locked_inode()
894 ni->itype.index.vcn_size_bits = vol->cluster_size_bits; ntfs_read_locked_inode()
896 ni->itype.index.vcn_size = vol->sector_size; ntfs_read_locked_inode()
897 ni->itype.index.vcn_size_bits = vol->sector_size_bits; ntfs_read_locked_inode()
901 NInoSetMstProtected(ni); ntfs_read_locked_inode()
902 ni->type = AT_INDEX_ALLOCATION; ntfs_read_locked_inode()
903 ni->name = I30; ntfs_read_locked_inode()
904 ni->name_len = 4; ntfs_read_locked_inode()
908 vi->i_size = ni->initialized_size = ntfs_read_locked_inode()
909 ni->allocated_size = 0; ntfs_read_locked_inode()
912 unmap_mft_record(ni); ntfs_read_locked_inode()
917 NInoSetIndexAllocPresent(ni); ntfs_read_locked_inode()
973 ni->initialized_size = sle64_to_cpu( ntfs_read_locked_inode()
975 ni->allocated_size = sle64_to_cpu( ntfs_read_locked_inode()
982 unmap_mft_record(ni); ntfs_read_locked_inode()
1002 ni->itype.index.block_size_bits)) { ntfs_read_locked_inode()
1020 ni->type = AT_DATA; ntfs_read_locked_inode()
1021 ni->name = NULL; ntfs_read_locked_inode()
1022 ni->name_len = 0; ntfs_read_locked_inode()
1027 vi->i_size = ni->initialized_size = ntfs_read_locked_inode()
1028 ni->allocated_size = 0; ntfs_read_locked_inode()
1060 NInoSetCompressed(ni); ntfs_read_locked_inode()
1080 NInoSetSparse(ni); ntfs_read_locked_inode()
1083 if (NInoCompressed(ni)) { ntfs_read_locked_inode()
1088 NInoSetEncrypted(ni); ntfs_read_locked_inode()
1091 NInoSetNonResident(ni); ntfs_read_locked_inode()
1092 if (NInoCompressed(ni) || NInoSparse(ni)) { ntfs_read_locked_inode()
1093 if (NInoCompressed(ni) && a->data.non_resident. ntfs_read_locked_inode()
1106 ni->itype.compressed.block_size = 1U << ntfs_read_locked_inode()
1110 ni->itype.compressed.block_size_bits = ntfs_read_locked_inode()
1111 ffs(ni->itype. ntfs_read_locked_inode()
1114 ni->itype.compressed.block_clusters = ntfs_read_locked_inode()
1119 ni->itype.compressed.block_size = 0; ntfs_read_locked_inode()
1120 ni->itype.compressed.block_size_bits = ntfs_read_locked_inode()
1122 ni->itype.compressed.block_clusters = ntfs_read_locked_inode()
1125 ni->itype.compressed.size = sle64_to_cpu( ntfs_read_locked_inode()
1137 ni->initialized_size = sle64_to_cpu( ntfs_read_locked_inode()
1139 ni->allocated_size = sle64_to_cpu( ntfs_read_locked_inode()
1142 vi->i_size = ni->initialized_size = le32_to_cpu( ntfs_read_locked_inode()
1144 ni->allocated_size = le32_to_cpu(a->length) - ntfs_read_locked_inode()
1147 if (vi->i_size > ni->allocated_size) { ntfs_read_locked_inode()
1157 unmap_mft_record(ni); ntfs_read_locked_inode()
1164 if (NInoMstProtected(ni)) ntfs_read_locked_inode()
1166 else if (NInoCompressed(ni)) ntfs_read_locked_inode()
1180 if (S_ISREG(vi->i_mode) && (NInoCompressed(ni) || NInoSparse(ni))) ntfs_read_locked_inode()
1181 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_read_locked_inode()
1183 vi->i_blocks = ni->allocated_size >> 9; ntfs_read_locked_inode()
1194 unmap_mft_record(ni); ntfs_read_locked_inode()
1229 ntfs_inode *ni, *base_ni; ntfs_read_locked_attr_inode() local
1239 ni = NTFS_I(vi); ntfs_read_locked_attr_inode()
1250 vi->i_generation = ni->seq_no = base_ni->seq_no; ntfs_read_locked_attr_inode()
1266 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_read_locked_attr_inode()
1273 NInoSetCompressed(ni); ntfs_read_locked_attr_inode()
1274 if ((ni->type != AT_DATA) || (ni->type == AT_DATA && ntfs_read_locked_attr_inode()
1275 ni->name_len)) { ntfs_read_locked_attr_inode()
1303 if (NInoMstProtected(ni) && ni->type != AT_INDEX_ROOT) { ntfs_read_locked_attr_inode()
1308 NInoCompressed(ni) ? "compressed" : ntfs_read_locked_attr_inode()
1313 NInoSetSparse(ni); ntfs_read_locked_attr_inode()
1316 if (NInoCompressed(ni)) { ntfs_read_locked_attr_inode()
1325 if (NInoMstProtected(ni) && ni->type != AT_INDEX_ROOT) { ntfs_read_locked_attr_inode()
1333 if (ni->type != AT_DATA) { ntfs_read_locked_attr_inode()
1338 NInoSetEncrypted(ni); ntfs_read_locked_attr_inode()
1348 if (NInoMstProtected(ni)) { ntfs_read_locked_attr_inode()
1355 vi->i_size = ni->initialized_size = le32_to_cpu( ntfs_read_locked_attr_inode()
1357 ni->allocated_size = le32_to_cpu(a->length) - ntfs_read_locked_attr_inode()
1359 if (vi->i_size > ni->allocated_size) { ntfs_read_locked_attr_inode()
1365 NInoSetNonResident(ni); ntfs_read_locked_attr_inode()
1377 if (NInoCompressed(ni) || NInoSparse(ni)) { ntfs_read_locked_attr_inode()
1378 if (NInoCompressed(ni) && a->data.non_resident. ntfs_read_locked_attr_inode()
1389 ni->itype.compressed.block_size = 1U << ntfs_read_locked_attr_inode()
1393 ni->itype.compressed.block_size_bits = ntfs_read_locked_attr_inode()
1394 ffs(ni->itype.compressed. ntfs_read_locked_attr_inode()
1396 ni->itype.compressed.block_clusters = 1U << ntfs_read_locked_attr_inode()
1400 ni->itype.compressed.block_size = 0; ntfs_read_locked_attr_inode()
1401 ni->itype.compressed.block_size_bits = 0; ntfs_read_locked_attr_inode()
1402 ni->itype.compressed.block_clusters = 0; ntfs_read_locked_attr_inode()
1404 ni->itype.compressed.size = sle64_to_cpu( ntfs_read_locked_attr_inode()
1413 ni->initialized_size = sle64_to_cpu( ntfs_read_locked_attr_inode()
1415 ni->allocated_size = sle64_to_cpu( ntfs_read_locked_attr_inode()
1419 if (NInoMstProtected(ni)) ntfs_read_locked_attr_inode()
1421 else if (NInoCompressed(ni)) ntfs_read_locked_attr_inode()
1423 if ((NInoCompressed(ni) || NInoSparse(ni)) && ni->type != AT_INDEX_ROOT) ntfs_read_locked_attr_inode()
1424 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_read_locked_attr_inode()
1426 vi->i_blocks = ni->allocated_size >> 9; ntfs_read_locked_attr_inode()
1432 ni->ext.base_ntfs_ino = base_ni; ntfs_read_locked_attr_inode()
1433 ni->nr_extents = -1; ntfs_read_locked_attr_inode()
1451 "Run chkdsk.", err, vi->i_ino, ni->type, ni->name_len, ntfs_read_locked_attr_inode()
1496 ntfs_inode *ni, *base_ni, *bni; ntfs_read_locked_index_inode() local
1507 ni = NTFS_I(vi); ntfs_read_locked_index_inode()
1517 vi->i_generation = ni->seq_no = base_ni->seq_no; ntfs_read_locked_index_inode()
1532 err = ntfs_attr_lookup(AT_INDEX_ROOT, ni->name, ni->name_len, ntfs_read_locked_index_inode()
1579 ni->itype.index.collation_rule = ir->collation_rule; ntfs_read_locked_index_inode()
1582 ni->itype.index.block_size = le32_to_cpu(ir->index_block_size); ntfs_read_locked_index_inode()
1583 if (!is_power_of_2(ni->itype.index.block_size)) { ntfs_read_locked_index_inode()
1585 "two.", ni->itype.index.block_size); ntfs_read_locked_index_inode()
1588 if (ni->itype.index.block_size > PAGE_CACHE_SIZE) { ntfs_read_locked_index_inode()
1591 ni->itype.index.block_size, PAGE_CACHE_SIZE); ntfs_read_locked_index_inode()
1595 if (ni->itype.index.block_size < NTFS_BLOCK_SIZE) { ntfs_read_locked_index_inode()
1598 ni->itype.index.block_size, NTFS_BLOCK_SIZE); ntfs_read_locked_index_inode()
1602 ni->itype.index.block_size_bits = ffs(ni->itype.index.block_size) - 1; ntfs_read_locked_index_inode()
1604 if (vol->cluster_size <= ni->itype.index.block_size) { ntfs_read_locked_index_inode()
1605 ni->itype.index.vcn_size = vol->cluster_size; ntfs_read_locked_index_inode()
1606 ni->itype.index.vcn_size_bits = vol->cluster_size_bits; ntfs_read_locked_index_inode()
1608 ni->itype.index.vcn_size = vol->sector_size; ntfs_read_locked_index_inode()
1609 ni->itype.index.vcn_size_bits = vol->sector_size_bits; ntfs_read_locked_index_inode()
1614 vi->i_size = ni->initialized_size = ni->allocated_size = 0; ntfs_read_locked_index_inode()
1622 NInoSetIndexAllocPresent(ni); ntfs_read_locked_index_inode()
1625 err = ntfs_attr_lookup(AT_INDEX_ALLOCATION, ni->name, ni->name_len, ntfs_read_locked_index_inode()
1673 ni->initialized_size = sle64_to_cpu( ntfs_read_locked_index_inode()
1675 ni->allocated_size = sle64_to_cpu(a->data.non_resident.allocated_size); ntfs_read_locked_index_inode()
1685 bvi = ntfs_attr_iget(base_vi, AT_BITMAP, ni->name, ni->name_len); ntfs_read_locked_index_inode()
1700 if ((bvi_size << 3) < (vi->i_size >> ni->itype.index.block_size_bits)) { ntfs_read_locked_index_inode()
1710 vi->i_blocks = ni->allocated_size >> 9; ntfs_read_locked_index_inode()
1716 ni->ext.base_ntfs_ino = base_ni; ntfs_read_locked_index_inode()
1717 ni->nr_extents = -1; ntfs_read_locked_index_inode()
1733 ni->name_len); ntfs_read_locked_index_inode()
1782 ntfs_inode *ni; ntfs_read_inode_mount() local
1794 ni = NTFS_I(vi); ntfs_read_inode_mount()
1797 NInoSetNonResident(ni); ntfs_read_inode_mount()
1798 NInoSetMstProtected(ni); ntfs_read_inode_mount()
1799 NInoSetSparseDisabled(ni); ntfs_read_inode_mount()
1800 ni->type = AT_DATA; ntfs_read_inode_mount()
1801 ni->name = NULL; ntfs_read_inode_mount()
1802 ni->name_len = 0; ntfs_read_inode_mount()
1807 ni->itype.index.block_size = vol->mft_record_size; ntfs_read_inode_mount()
1808 ni->itype.index.block_size_bits = vol->mft_record_size_bits; ntfs_read_inode_mount()
1855 vi->i_generation = ni->seq_no = le16_to_cpu(m->sequence_number); ntfs_read_inode_mount()
1860 ctx = ntfs_attr_get_search_ctx(ni, m); ntfs_read_inode_mount()
1881 NInoSetAttrList(ni); ntfs_read_inode_mount()
1905 ni->attr_list_size = (u32)ntfs_attr_size(a); ntfs_read_inode_mount()
1906 ni->attr_list = ntfs_malloc_nofs(ni->attr_list_size); ntfs_read_inode_mount()
1907 if (!ni->attr_list) { ntfs_read_inode_mount()
1913 NInoSetAttrListNonResident(ni); ntfs_read_inode_mount()
1921 ni->attr_list_rl.rl = ntfs_mapping_pairs_decompress(vol, ntfs_read_inode_mount()
1923 if (IS_ERR(ni->attr_list_rl.rl)) { ntfs_read_inode_mount()
1924 err = PTR_ERR(ni->attr_list_rl.rl); ntfs_read_inode_mount()
1925 ni->attr_list_rl.rl = NULL; ntfs_read_inode_mount()
1932 if ((err = load_attribute_list(vol, &ni->attr_list_rl, ntfs_read_inode_mount()
1933 ni->attr_list, ni->attr_list_size, ntfs_read_inode_mount()
1952 memcpy(ni->attr_list, (u8*)a + le16_to_cpu( ntfs_read_inode_mount()
1969 al_entry = (ATTR_LIST_ENTRY*)ni->attr_list; ntfs_read_inode_mount()
1970 al_end = (u8*)al_entry + ni->attr_list_size; ntfs_read_inode_mount()
1973 if ((u8*)al_entry < ni->attr_list || ntfs_read_inode_mount()
2009 ni->seq_no) ntfs_read_inode_mount()
2052 nrl = ntfs_mapping_pairs_decompress(vol, a, ni->runlist.rl); ntfs_read_inode_mount()
2059 ni->runlist.rl = nrl; ntfs_read_inode_mount()
2077 ni->initialized_size = sle64_to_cpu( ntfs_read_inode_mount()
2079 ni->allocated_size = sle64_to_cpu( ntfs_read_inode_mount()
2180 lockdep_set_class(&ni->runlist.lock, &mft_ni_runlist_lock_key); ntfs_read_inode_mount()
2181 lockdep_set_class(&ni->mrec_lock, &mft_ni_mrec_lock_key); ntfs_read_inode_mount()
2197 static void __ntfs_clear_inode(ntfs_inode *ni) __ntfs_clear_inode() argument
2200 down_write(&ni->runlist.lock); __ntfs_clear_inode()
2201 if (ni->runlist.rl) { __ntfs_clear_inode()
2202 ntfs_free(ni->runlist.rl); __ntfs_clear_inode()
2203 ni->runlist.rl = NULL; __ntfs_clear_inode()
2205 up_write(&ni->runlist.lock); __ntfs_clear_inode()
2207 if (ni->attr_list) { __ntfs_clear_inode()
2208 ntfs_free(ni->attr_list); __ntfs_clear_inode()
2209 ni->attr_list = NULL; __ntfs_clear_inode()
2212 down_write(&ni->attr_list_rl.lock); __ntfs_clear_inode()
2213 if (ni->attr_list_rl.rl) { __ntfs_clear_inode()
2214 ntfs_free(ni->attr_list_rl.rl); __ntfs_clear_inode()
2215 ni->attr_list_rl.rl = NULL; __ntfs_clear_inode()
2217 up_write(&ni->attr_list_rl.lock); __ntfs_clear_inode()
2219 if (ni->name_len && ni->name != I30) { __ntfs_clear_inode()
2221 BUG_ON(!ni->name); __ntfs_clear_inode()
2222 kfree(ni->name); __ntfs_clear_inode()
2226 void ntfs_clear_extent_inode(ntfs_inode *ni) ntfs_clear_extent_inode() argument
2228 ntfs_debug("Entering for inode 0x%lx.", ni->mft_no); ntfs_clear_extent_inode()
2230 BUG_ON(NInoAttr(ni)); ntfs_clear_extent_inode()
2231 BUG_ON(ni->nr_extents != -1); ntfs_clear_extent_inode()
2234 if (NInoDirty(ni)) { ntfs_clear_extent_inode()
2235 if (!is_bad_inode(VFS_I(ni->ext.base_ntfs_ino))) ntfs_clear_extent_inode()
2236 ntfs_error(ni->vol->sb, "Clearing dirty extent inode! " ntfs_clear_extent_inode()
2242 __ntfs_clear_inode(ni); ntfs_clear_extent_inode()
2245 ntfs_destroy_extent_inode(ni); ntfs_clear_extent_inode()
2260 ntfs_inode *ni = NTFS_I(vi); ntfs_evict_big_inode() local
2266 if (NInoDirty(ni)) { ntfs_evict_big_inode()
2272 if (!was_bad && (is_bad_inode(vi) || NInoDirty(ni))) { ntfs_evict_big_inode()
2281 if (ni->nr_extents > 0) { ntfs_evict_big_inode()
2284 for (i = 0; i < ni->nr_extents; i++) ntfs_evict_big_inode()
2285 ntfs_clear_extent_inode(ni->ext.extent_ntfs_inos[i]); ntfs_evict_big_inode()
2286 kfree(ni->ext.extent_ntfs_inos); ntfs_evict_big_inode()
2289 __ntfs_clear_inode(ni); ntfs_evict_big_inode()
2291 if (NInoAttr(ni)) { ntfs_evict_big_inode()
2293 if (ni->nr_extents == -1) { ntfs_evict_big_inode()
2294 iput(VFS_I(ni->ext.base_ntfs_ino)); ntfs_evict_big_inode()
2295 ni->nr_extents = 0; ntfs_evict_big_inode()
2296 ni->ext.base_ntfs_ino = NULL; ntfs_evict_big_inode()
2368 ntfs_inode *base_ni, *ni = NTFS_I(vi); ntfs_truncate() local
2369 ntfs_volume *vol = ni->vol; ntfs_truncate()
2378 BUG_ON(NInoAttr(ni)); ntfs_truncate()
2380 BUG_ON(NInoMstProtected(ni)); ntfs_truncate()
2381 BUG_ON(ni->nr_extents < 0); ntfs_truncate()
2387 down_write(&ni->runlist.lock); ntfs_truncate()
2388 if (!NInoAttr(ni)) ntfs_truncate()
2389 base_ni = ni; ntfs_truncate()
2391 base_ni = ni->ext.base_ntfs_ino; ntfs_truncate()
2409 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_truncate()
2432 if (NInoNonResident(ni)) ntfs_truncate()
2438 read_lock_irqsave(&ni->size_lock, flags); ntfs_truncate()
2439 old_alloc_size = ni->allocated_size; ntfs_truncate()
2440 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_truncate()
2466 err = ntfs_attr_size_bounds_check(vol, ni->type, new_size); ntfs_truncate()
2476 le32_to_cpu(ni->type)); ntfs_truncate()
2483 le32_to_cpu(ni->type)); ntfs_truncate()
2491 if (NInoCompressed(ni) || NInoEncrypted(ni)) { ntfs_truncate()
2494 NInoCompressed(ni) ? "compressed" : ntfs_truncate()
2501 BUG_ON(NInoNonResident(ni)); ntfs_truncate()
2508 write_lock_irqsave(&ni->size_lock, flags); ntfs_truncate()
2510 ni->allocated_size = le32_to_cpu(a->length) - ntfs_truncate()
2540 ni->initialized_size = new_size; ntfs_truncate()
2541 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_truncate()
2557 up_write(&ni->runlist.lock); ntfs_truncate()
2562 err = ntfs_attr_make_non_resident(ni, old_size); ntfs_truncate()
2575 (unsigned)le32_to_cpu(ni->type), err); ntfs_truncate()
2601 if (ni->type == AT_ATTRIBUTE_LIST || ntfs_truncate()
2602 ni->type == AT_STANDARD_INFORMATION) { ntfs_truncate()
2620 BUG_ON(!NInoNonResident(ni)); ntfs_truncate()
2636 (unsigned)le32_to_cpu(ni->type)); ntfs_truncate()
2649 write_lock_irqsave(&ni->size_lock, flags); ntfs_truncate()
2650 if (new_size < ni->initialized_size) { ntfs_truncate()
2651 ni->initialized_size = new_size; ntfs_truncate()
2656 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_truncate()
2690 up_write(&ni->runlist.lock); ntfs_truncate()
2691 err = ntfs_attr_extend_allocation(ni, new_size, ntfs_truncate()
2704 nr_freed = ntfs_cluster_free(ni, new_alloc_size >> ntfs_truncate()
2716 err = ntfs_rl_truncate_nolock(vol, &ni->runlist, ntfs_truncate()
2734 mp_size = ntfs_get_size_for_mapping_pairs(vol, ni->runlist.rl, 0, -1); ntfs_truncate()
2740 (unsigned)le32_to_cpu(ni->type), mp_size, es); ntfs_truncate()
2758 mp_size, ni->runlist.rl, 0, -1, NULL); ntfs_truncate()
2763 vi->i_ino, (unsigned)le32_to_cpu(ni->type), ntfs_truncate()
2771 write_lock_irqsave(&ni->size_lock, flags); ntfs_truncate()
2772 ni->allocated_size = new_alloc_size; ntfs_truncate()
2774 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_truncate()
2776 ni->itype.compressed.size -= nr_freed << ntfs_truncate()
2778 BUG_ON(ni->itype.compressed.size < 0); ntfs_truncate()
2780 ni->itype.compressed.size); ntfs_truncate()
2781 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_truncate()
2785 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_truncate()
2807 up_write(&ni->runlist.lock); ntfs_truncate()
2830 NInoClearTruncateFailed(ni); ntfs_truncate()
2840 NInoSetTruncateFailed(ni); ntfs_truncate()
2848 up_write(&ni->runlist.lock); ntfs_truncate()
2856 NInoSetTruncateFailed(ni); ntfs_truncate()
2908 ntfs_inode *ni = NTFS_I(vi); ntfs_setattr() local
2913 if (NInoCompressed(ni) || NInoEncrypted(ni)) { ntfs_setattr()
2917 NInoCompressed(ni) ? ntfs_setattr()
2969 ntfs_inode *ni = NTFS_I(vi); __ntfs_write_inode() local
2976 ntfs_debug("Entering for %sinode 0x%lx.", NInoAttr(ni) ? "attr " : "", __ntfs_write_inode()
2983 if (NInoAttr(ni)) { __ntfs_write_inode()
2984 NInoClearDirty(ni); __ntfs_write_inode()
2989 m = map_mft_record(ni); __ntfs_write_inode()
2995 ctx = ntfs_attr_get_search_ctx(ni, m); __ntfs_write_inode()
3060 if (NInoDirty(ni)) __ntfs_write_inode()
3061 err = write_mft_record(ni, m, sync); __ntfs_write_inode()
3063 mutex_lock(&ni->extent_lock); __ntfs_write_inode()
3064 if (ni->nr_extents > 0) { __ntfs_write_inode()
3065 ntfs_inode **extent_nis = ni->ext.extent_ntfs_inos; __ntfs_write_inode()
3068 ntfs_debug("Writing %i extent inodes.", ni->nr_extents); __ntfs_write_inode()
3069 for (i = 0; i < ni->nr_extents; i++) { __ntfs_write_inode()
3090 mutex_unlock(&ni->extent_lock); __ntfs_write_inode()
3091 unmap_mft_record(ni); __ntfs_write_inode()
3097 unmap_mft_record(ni); __ntfs_write_inode()
3106 NVolSetErrors(ni->vol); __ntfs_write_inode()
H A Daops.c64 ntfs_inode *ni; ntfs_end_buffer_async_read() local
69 ni = NTFS_I(vi); ntfs_end_buffer_async_read()
79 read_lock_irqsave(&ni->size_lock, flags); ntfs_end_buffer_async_read()
80 init_size = ni->initialized_size; ntfs_end_buffer_async_read()
82 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_end_buffer_async_read()
106 ntfs_error(ni->vol->sb, "Buffer I/O error, logical block " ntfs_end_buffer_async_read()
131 * attribute is mst protected, i.e. if NInoMstProteced(ni) is true. ntfs_end_buffer_async_read()
136 if (!NInoMstProtected(ni)) { ntfs_end_buffer_async_read()
144 rec_size = ni->itype.index.block_size; ntfs_end_buffer_async_read()
191 ntfs_inode *ni; ntfs_read_block() local
202 ni = NTFS_I(vi); ntfs_read_block()
203 vol = ni->vol; ntfs_read_block()
206 BUG_ON(!ni->runlist.rl && !ni->mft_no && !NInoAttr(ni)); ntfs_read_block()
233 read_lock_irqsave(&ni->size_lock, flags); ntfs_read_block()
234 lblock = (ni->allocated_size + blocksize - 1) >> blocksize_bits; ntfs_read_block()
235 init_size = ni->initialized_size; ntfs_read_block()
237 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_read_block()
268 down_read(&ni->runlist.lock); ntfs_read_block()
269 rl = ni->runlist.rl; ntfs_read_block()
302 up_read(&ni->runlist.lock); ntfs_read_block()
303 err = ntfs_map_runlist(ni, vcn); ntfs_read_block()
308 up_read(&ni->runlist.lock); ntfs_read_block()
327 "(error code %i).", ni->mft_no, ntfs_read_block()
328 ni->type, (unsigned long long)vcn, ntfs_read_block()
348 up_read(&ni->runlist.lock); ntfs_read_block()
402 ntfs_inode *ni, *base_ni; ntfs_readpage() local
429 ni = NTFS_I(vi); ntfs_readpage()
438 if (ni->type != AT_INDEX_ALLOCATION) { ntfs_readpage()
440 if (NInoEncrypted(ni)) { ntfs_readpage()
441 BUG_ON(ni->type != AT_DATA); ntfs_readpage()
446 if (NInoNonResident(ni) && NInoCompressed(ni)) { ntfs_readpage()
447 BUG_ON(ni->type != AT_DATA); ntfs_readpage()
448 BUG_ON(ni->name_len); ntfs_readpage()
453 if (NInoNonResident(ni)) { ntfs_readpage()
469 if (!NInoAttr(ni)) ntfs_readpage()
470 base_ni = ni; ntfs_readpage()
472 base_ni = ni->ext.base_ntfs_ino; ntfs_readpage()
483 if (unlikely(NInoNonResident(ni))) { ntfs_readpage()
492 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_readpage()
497 read_lock_irqsave(&ni->size_lock, flags); ntfs_readpage()
498 if (unlikely(attr_len > ni->initialized_size)) ntfs_readpage()
499 attr_len = ni->initialized_size; ntfs_readpage()
501 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_readpage()
558 ntfs_inode *ni; ntfs_write_block() local
569 ni = NTFS_I(vi); ntfs_write_block()
570 vol = ni->vol; ntfs_write_block()
573 "0x%lx.", ni->mft_no, ni->type, page->index); ntfs_write_block()
575 BUG_ON(!NInoNonResident(ni)); ntfs_write_block()
576 BUG_ON(NInoMstProtected(ni)); ntfs_write_block()
604 read_lock_irqsave(&ni->size_lock, flags); ntfs_write_block()
606 initialized_size = ni->initialized_size; ntfs_write_block()
607 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_write_block()
724 down_read(&ni->runlist.lock); ntfs_write_block()
725 rl = ni->runlist.rl; ntfs_write_block()
780 up_read(&ni->runlist.lock); ntfs_write_block()
781 err = ntfs_map_runlist(ni, vcn); ntfs_write_block()
786 up_read(&ni->runlist.lock); ntfs_write_block()
807 "determined%s (error code %i).", ni->mft_no, ntfs_write_block()
808 ni->type, (unsigned long long)vcn, ntfs_write_block()
816 up_read(&ni->runlist.lock); ntfs_write_block()
924 ntfs_inode *ni = NTFS_I(vi); ntfs_write_mst_block() local
925 ntfs_volume *vol = ni->vol; ntfs_write_mst_block()
927 unsigned int rec_size = ni->itype.index.block_size; ntfs_write_mst_block()
938 "0x%lx.", vi->i_ino, ni->type, page->index); ntfs_write_mst_block()
939 BUG_ON(!NInoNonResident(ni)); ntfs_write_mst_block()
940 BUG_ON(!NInoMstProtected(ni)); ntfs_write_mst_block()
949 (NInoAttr(ni) && ni->type == AT_INDEX_ALLOCATION))); ntfs_write_mst_block()
963 rec_size_bits = ni->itype.index.block_size_bits; ntfs_write_mst_block()
1030 down_read(&ni->runlist.lock); ntfs_write_mst_block()
1031 rl = ni->runlist.rl; ntfs_write_mst_block()
1060 up_read(&ni->runlist.lock); ntfs_write_mst_block()
1061 err2 = ntfs_map_runlist(ni, vcn); ntfs_write_mst_block()
1070 up_read(&ni->runlist.lock); ntfs_write_mst_block()
1085 ni->mft_no, ni->type, ntfs_write_mst_block()
1114 up_read(&ni->runlist.lock); ntfs_write_mst_block()
1175 ni->type, page->index, ofs); ntfs_write_mst_block()
1222 "and run chkdsk.", vi->i_ino, ni->type, ntfs_write_mst_block()
1303 if (ni->itype.index.block_size == PAGE_CACHE_SIZE) ntfs_write_mst_block()
1357 ntfs_inode *base_ni = NULL, *ni = NTFS_I(vi); ntfs_writepage() local
1387 if (ni->type != AT_INDEX_ALLOCATION) { ntfs_writepage()
1389 if (NInoEncrypted(ni)) { ntfs_writepage()
1391 BUG_ON(ni->type != AT_DATA); ntfs_writepage()
1396 if (NInoNonResident(ni) && NInoCompressed(ni)) { ntfs_writepage()
1397 BUG_ON(ni->type != AT_DATA); ntfs_writepage()
1398 BUG_ON(ni->name_len); ntfs_writepage()
1407 if (NInoNonResident(ni) && NInoSparse(ni)) { ntfs_writepage()
1415 if (NInoNonResident(ni)) { ntfs_writepage()
1423 if (NInoMstProtected(ni)) ntfs_writepage()
1447 if (!NInoAttr(ni)) ntfs_writepage()
1448 base_ni = ni; ntfs_writepage()
1450 base_ni = ni->ext.base_ntfs_ino; ntfs_writepage()
1463 if (unlikely(NInoNonResident(ni))) { ntfs_writepage()
1472 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_writepage()
1528 NVolSetErrors(ni->vol); ntfs_writepage()
1569 ntfs_inode *ni = NTFS_I(mapping->host); ntfs_bmap() local
1570 ntfs_volume *vol = ni->vol; ntfs_bmap()
1575 ni->mft_no, (unsigned long long)block); ntfs_bmap()
1576 if (ni->type != AT_DATA || !NInoNonResident(ni) || NInoEncrypted(ni)) { ntfs_bmap()
1579 (ni->type != AT_DATA) ? "non-data" : ntfs_bmap()
1580 (!NInoNonResident(ni) ? "resident" : ntfs_bmap()
1585 BUG_ON(NInoCompressed(ni)); ntfs_bmap()
1586 BUG_ON(NInoMstProtected(ni)); ntfs_bmap()
1590 read_lock_irqsave(&ni->size_lock, flags); ntfs_bmap()
1591 size = ni->initialized_size; ntfs_bmap()
1592 i_size = i_size_read(VFS_I(ni)); ntfs_bmap()
1593 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_bmap()
1602 down_read(&ni->runlist.lock); ntfs_bmap()
1603 lcn = ntfs_attr_vcn_to_lcn_nolock(ni, ofs >> cluster_size_shift, false); ntfs_bmap()
1604 up_read(&ni->runlist.lock); ntfs_bmap()
1625 "Returning 0.", ni->mft_no); ntfs_bmap()
1630 "Returning 0.", ni->mft_no); ntfs_bmap()
1731 ntfs_inode *ni = NTFS_I(mapping->host); mark_ntfs_record_dirty() local
1736 end = ofs + ni->itype.index.block_size; mark_ntfs_record_dirty()
1737 bh_size = VFS_I(ni)->i_sb->s_blocksize; mark_ntfs_record_dirty()
H A Dfile.c75 * @ni: ntfs inode of the attribute to extend
78 * Extend the initialized size of an attribute described by the ntfs inode @ni
106 * Locking: i_mutex on the vfs inode corrseponsind to the ntfs inode @ni must be
109 static int ntfs_attr_extend_initialized(ntfs_inode *ni, const s64 new_init_size) ntfs_attr_extend_initialized() argument
115 struct inode *vi = VFS_I(ni); ntfs_attr_extend_initialized()
126 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized()
127 old_init_size = ni->initialized_size; ntfs_attr_extend_initialized()
129 BUG_ON(new_init_size > ni->allocated_size); ntfs_attr_extend_initialized()
130 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized()
134 vi->i_ino, (unsigned)le32_to_cpu(ni->type), ntfs_attr_extend_initialized()
137 if (!NInoAttr(ni)) ntfs_attr_extend_initialized()
138 base_ni = ni; ntfs_attr_extend_initialized()
140 base_ni = ni->ext.base_ntfs_ino; ntfs_attr_extend_initialized()
142 if (NInoNonResident(ni)) ntfs_attr_extend_initialized()
156 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_initialized()
177 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized()
179 ni->initialized_size = new_init_size; ntfs_attr_extend_initialized()
180 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized()
200 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_initialized()
244 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized()
245 ni->initialized_size = (s64)(index + 1) << PAGE_CACHE_SHIFT; ntfs_attr_extend_initialized()
246 if (ni->initialized_size > new_init_size) ntfs_attr_extend_initialized()
247 ni->initialized_size = new_init_size; ntfs_attr_extend_initialized()
248 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized()
282 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized()
283 BUG_ON(ni->initialized_size != new_init_size); ntfs_attr_extend_initialized()
284 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized()
297 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_initialized()
319 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_initialized()
320 ni->initialized_size = old_init_size; ntfs_attr_extend_initialized()
321 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_initialized()
340 ntfs_inode *base_ni, *ni = NTFS_I(vi); ntfs_prepare_file_for_write() local
341 ntfs_volume *vol = ni->vol; ntfs_prepare_file_for_write()
345 (unsigned)le32_to_cpu(ni->type), ntfs_prepare_file_for_write()
355 BUG_ON(NInoMstProtected(ni)); ntfs_prepare_file_for_write()
356 BUG_ON(ni->type != AT_DATA); ntfs_prepare_file_for_write()
358 if (NInoEncrypted(ni)) { ntfs_prepare_file_for_write()
368 if (NInoCompressed(ni)) { ntfs_prepare_file_for_write()
370 BUG_ON(ni->name_len); ntfs_prepare_file_for_write()
382 base_ni = ni; ntfs_prepare_file_for_write()
383 if (NInoAttr(ni)) ntfs_prepare_file_for_write()
384 base_ni = ni->ext.base_ntfs_ino; ntfs_prepare_file_for_write()
401 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_file_for_write()
402 ll = ni->allocated_size; ntfs_prepare_file_for_write()
403 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_file_for_write()
412 ll = ntfs_attr_extend_allocation(ni, end, -1, pos); ntfs_prepare_file_for_write()
422 le32_to_cpu(ni->type)); ntfs_prepare_file_for_write()
427 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_file_for_write()
428 ll = ni->allocated_size; ntfs_prepare_file_for_write()
429 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_file_for_write()
437 le32_to_cpu(ni->type), ntfs_prepare_file_for_write()
450 le32_to_cpu(ni->type), ntfs_prepare_file_for_write()
459 le32_to_cpu(ni->type)); ntfs_prepare_file_for_write()
471 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_file_for_write()
472 ll = ni->initialized_size; ntfs_prepare_file_for_write()
473 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_file_for_write()
480 err = ntfs_attr_extend_initialized(ni, pos); ntfs_prepare_file_for_write()
486 (unsigned)le32_to_cpu(ni->type), ntfs_prepare_file_for_write()
593 ntfs_inode *ni, *base_ni = NULL; ntfs_prepare_pages_for_non_resident_write() local
617 ni = NTFS_I(vi); ntfs_prepare_pages_for_non_resident_write()
618 vol = ni->vol; ntfs_prepare_pages_for_non_resident_write()
621 vi->i_ino, ni->type, pages[0]->index, nr_pages, ntfs_prepare_pages_for_non_resident_write()
696 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
697 initialized_size = ni->initialized_size; ntfs_prepare_pages_for_non_resident_write()
698 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
770 read_lock_irqsave(&ni->size_lock, ntfs_prepare_pages_for_non_resident_write()
772 initialized_size = ni->initialized_size; ntfs_prepare_pages_for_non_resident_write()
773 read_unlock_irqrestore(&ni->size_lock, ntfs_prepare_pages_for_non_resident_write()
830 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
831 initialized_size = ni->allocated_size; ntfs_prepare_pages_for_non_resident_write()
832 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
845 down_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
847 rl = ni->runlist.rl; ntfs_prepare_pages_for_non_resident_write()
874 up_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
877 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
899 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
900 down_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
904 err = ntfs_map_runlist_nolock(ni, bh_cpos, ntfs_prepare_pages_for_non_resident_write()
929 ni->mft_no, ni->type, ntfs_prepare_pages_for_non_resident_write()
981 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
982 down_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
990 while (--rl2 >= ni->runlist.rl) { ntfs_prepare_pages_for_non_resident_write()
1005 rl = ntfs_runlists_merge(ni->runlist.rl, rl2); ntfs_prepare_pages_for_non_resident_write()
1020 ni->runlist.rl = rl; ntfs_prepare_pages_for_non_resident_write()
1025 if (!NInoAttr(ni)) ntfs_prepare_pages_for_non_resident_write()
1026 base_ni = ni; ntfs_prepare_pages_for_non_resident_write()
1028 base_ni = ni->ext.base_ntfs_ino; ntfs_prepare_pages_for_non_resident_write()
1041 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_prepare_pages_for_non_resident_write()
1125 (unsigned)le32_to_cpu(ni->type), err); ntfs_prepare_pages_for_non_resident_write()
1137 if (likely(NInoSparse(ni) || NInoCompressed(ni))) { ntfs_prepare_pages_for_non_resident_write()
1147 err = ntfs_attr_lookup(ni->type, ni->name, ntfs_prepare_pages_for_non_resident_write()
1148 ni->name_len, CASE_SENSITIVE, ntfs_prepare_pages_for_non_resident_write()
1157 write_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
1158 ni->itype.compressed.size += vol->cluster_size; ntfs_prepare_pages_for_non_resident_write()
1160 cpu_to_sle64(ni->itype.compressed.size); ntfs_prepare_pages_for_non_resident_write()
1161 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
1184 up_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
1196 up_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
1199 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
1203 read_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
1204 initialized_size = ni->initialized_size; ntfs_prepare_pages_for_non_resident_write()
1205 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
1244 if (ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_prepare_pages_for_non_resident_write()
1250 write_lock_irqsave(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
1251 ni->itype.compressed.size += vol->cluster_size; ntfs_prepare_pages_for_non_resident_write()
1252 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_prepare_pages_for_non_resident_write()
1277 if (ntfs_rl_punch_nolock(vol, &ni->runlist, bh_cpos, 1)) { ntfs_prepare_pages_for_non_resident_write()
1318 mapping_pairs_offset), ni->runlist.rl, ntfs_prepare_pages_for_non_resident_write()
1337 up_write(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
1339 up_read(&ni->runlist.lock); ntfs_prepare_pages_for_non_resident_write()
1405 ntfs_inode *ni, *base_ni; ntfs_commit_pages_after_non_resident_write() local
1415 ni = NTFS_I(vi); ntfs_commit_pages_after_non_resident_write()
1451 read_lock_irqsave(&ni->size_lock, flags); ntfs_commit_pages_after_non_resident_write()
1452 initialized_size = ni->initialized_size; ntfs_commit_pages_after_non_resident_write()
1453 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_commit_pages_after_non_resident_write()
1462 if (!NInoAttr(ni)) ntfs_commit_pages_after_non_resident_write()
1463 base_ni = ni; ntfs_commit_pages_after_non_resident_write()
1465 base_ni = ni->ext.base_ntfs_ino; ntfs_commit_pages_after_non_resident_write()
1474 BUG_ON(!NInoNonResident(ni)); ntfs_commit_pages_after_non_resident_write()
1480 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_commit_pages_after_non_resident_write()
1489 write_lock_irqsave(&ni->size_lock, flags); ntfs_commit_pages_after_non_resident_write()
1490 BUG_ON(end > ni->allocated_size); ntfs_commit_pages_after_non_resident_write()
1491 ni->initialized_size = end; ntfs_commit_pages_after_non_resident_write()
1498 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_commit_pages_after_non_resident_write()
1514 NVolSetErrors(ni->vol); ntfs_commit_pages_after_non_resident_write()
1560 ntfs_inode *ni, *base_ni; ntfs_commit_pages_after_write() local
1575 ni = NTFS_I(vi); ntfs_commit_pages_after_write()
1578 vi->i_ino, ni->type, page->index, nr_pages, ntfs_commit_pages_after_write()
1580 if (NInoNonResident(ni)) ntfs_commit_pages_after_write()
1588 if (!NInoAttr(ni)) ntfs_commit_pages_after_write()
1589 base_ni = ni; ntfs_commit_pages_after_write()
1591 base_ni = ni->ext.base_ntfs_ino; ntfs_commit_pages_after_write()
1592 BUG_ON(NInoNonResident(ni)); ntfs_commit_pages_after_write()
1606 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_commit_pages_after_write()
1648 read_lock_irqsave(&ni->size_lock, flags); ntfs_commit_pages_after_write()
1649 initialized_size = ni->initialized_size; ntfs_commit_pages_after_write()
1650 BUG_ON(end > ni->allocated_size); ntfs_commit_pages_after_write()
1651 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_commit_pages_after_write()
1654 write_lock_irqsave(&ni->size_lock, flags); ntfs_commit_pages_after_write()
1655 ni->initialized_size = end; ntfs_commit_pages_after_write()
1657 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_commit_pages_after_write()
1686 NVolSetErrors(ni->vol); ntfs_commit_pages_after_write()
1750 ntfs_inode *ni = NTFS_I(vi); ntfs_perform_write() local
1751 ntfs_volume *vol = ni->vol; ntfs_perform_write()
1762 (unsigned)le32_to_cpu(ni->type), ntfs_perform_write()
1769 if (unlikely(NInoTruncateFailed(ni))) { ntfs_perform_write()
1774 if (err || NInoTruncateFailed(ni)) { ntfs_perform_write()
1781 (unsigned)le32_to_cpu(ni->type), err); ntfs_perform_write()
1790 if (vol->cluster_size > PAGE_CACHE_SIZE && NInoNonResident(ni)) ntfs_perform_write()
1812 down_read(&ni->runlist.lock); ntfs_perform_write()
1813 lcn = ntfs_attr_vcn_to_lcn_nolock(ni, pos >> ntfs_perform_write()
1815 up_read(&ni->runlist.lock); ntfs_perform_write()
1828 le32_to_cpu(ni->type)); ntfs_perform_write()
1868 if (NInoNonResident(ni)) { ntfs_perform_write()
H A Dattrib.c40 * @ni: ntfs inode for which to map (part of) a runlist
44 * Map the part of a runlist containing the @vcn of the ntfs inode @ni.
46 * If @ctx is specified, it is an active search context of @ni and its base mft
77 * Locking: - The runlist described by @ni must be locked for writing on entry
79 * - If @ctx is NULL, the base mft record of @ni must not be mapped on
84 int ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn, ntfs_attr_search_ctx *ctx) ntfs_map_runlist_nolock() argument
99 if (!NInoAttr(ni)) ntfs_map_runlist_nolock()
100 base_ni = ni; ntfs_map_runlist_nolock()
102 base_ni = ni->ext.base_ntfs_ino; ntfs_map_runlist_nolock()
121 read_lock_irqsave(&ni->size_lock, flags); ntfs_map_runlist_nolock()
122 allocated_size_vcn = ni->allocated_size >> ntfs_map_runlist_nolock()
123 ni->vol->cluster_size_bits; ntfs_map_runlist_nolock()
124 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_map_runlist_nolock()
134 if (vcn >= allocated_size_vcn || (a->type == ni->type && ntfs_map_runlist_nolock()
135 a->name_length == ni->name_len && ntfs_map_runlist_nolock()
137 ni->name, ni->name_len) && ntfs_map_runlist_nolock()
166 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_map_runlist_nolock()
187 rl = ntfs_mapping_pairs_decompress(ni->vol, a, ni->runlist.rl); ntfs_map_runlist_nolock()
191 ni->runlist.rl = rl; ntfs_map_runlist_nolock()
285 * @ni: ntfs inode for which to map (part of) a runlist
288 * Map the part of a runlist containing the @vcn of the ntfs inode @ni.
298 int ntfs_map_runlist(ntfs_inode *ni, VCN vcn) ntfs_map_runlist() argument
302 down_write(&ni->runlist.lock); ntfs_map_runlist()
304 if (likely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) <= ntfs_map_runlist()
306 err = ntfs_map_runlist_nolock(ni, vcn, NULL); ntfs_map_runlist()
307 up_write(&ni->runlist.lock); ntfs_map_runlist()
313 * @ni: ntfs inode of the attribute whose runlist to search
318 * described by the ntfs inode @ni and return the corresponding logical cluster
341 LCN ntfs_attr_vcn_to_lcn_nolock(ntfs_inode *ni, const VCN vcn, ntfs_attr_vcn_to_lcn_nolock() argument
348 BUG_ON(!ni); ntfs_attr_vcn_to_lcn_nolock()
350 ni->mft_no, (unsigned long long)vcn, ntfs_attr_vcn_to_lcn_nolock()
352 BUG_ON(!NInoNonResident(ni)); ntfs_attr_vcn_to_lcn_nolock()
354 if (!ni->runlist.rl) { ntfs_attr_vcn_to_lcn_nolock()
355 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_vcn_to_lcn_nolock()
356 if (!ni->allocated_size) { ntfs_attr_vcn_to_lcn_nolock()
357 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_vcn_to_lcn_nolock()
360 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_vcn_to_lcn_nolock()
364 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); ntfs_attr_vcn_to_lcn_nolock()
376 up_read(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock()
377 down_write(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock()
378 if (unlikely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) != ntfs_attr_vcn_to_lcn_nolock()
380 up_write(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock()
381 down_read(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock()
385 err = ntfs_map_runlist_nolock(ni, vcn, NULL); ntfs_attr_vcn_to_lcn_nolock()
387 up_write(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock()
388 down_read(&ni->runlist.lock); ntfs_attr_vcn_to_lcn_nolock()
402 ntfs_error(ni->vol->sb, "Failed with error code %lli.", ntfs_attr_vcn_to_lcn_nolock()
409 * @ni: ntfs inode describing the runlist to search
414 * inode @ni and return the address of the runlist element containing the @vcn.
419 * If @ctx is specified, it is an active search context of @ni and its base mft
456 * Locking: - The runlist described by @ni must be locked for writing on entry
459 * - If @ctx is NULL, the base mft record of @ni must not be mapped on
464 runlist_element *ntfs_attr_find_vcn_nolock(ntfs_inode *ni, const VCN vcn, ntfs_attr_find_vcn_nolock() argument
472 BUG_ON(!ni); ntfs_attr_find_vcn_nolock()
474 ni->mft_no, (unsigned long long)vcn, ctx ? "" : "out"); ntfs_attr_find_vcn_nolock()
475 BUG_ON(!NInoNonResident(ni)); ntfs_attr_find_vcn_nolock()
477 if (!ni->runlist.rl) { ntfs_attr_find_vcn_nolock()
478 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_find_vcn_nolock()
479 if (!ni->allocated_size) { ntfs_attr_find_vcn_nolock()
480 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_find_vcn_nolock()
483 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_find_vcn_nolock()
486 rl = ni->runlist.rl; ntfs_attr_find_vcn_nolock()
517 err = ntfs_map_runlist_nolock(ni, vcn, ctx); ntfs_attr_find_vcn_nolock()
528 ntfs_error(ni->vol->sb, "Failed with error code %i.", err); ntfs_attr_find_vcn_nolock()
862 ntfs_inode *base_ni, *ni; ntfs_external_attr_find() local
872 ni = ctx->ntfs_ino; ntfs_external_attr_find()
874 ntfs_debug("Entering for inode 0x%lx, type 0x%x.", ni->mft_no, type); ntfs_external_attr_find()
880 if (ni == base_ni) ntfs_external_attr_find()
981 if (MREF_LE(al_entry->mft_reference) == ni->mft_no) { ntfs_external_attr_find()
982 if (MSEQNO_LE(al_entry->mft_reference) != ni->seq_no) { ntfs_external_attr_find()
992 if (ni != base_ni) ntfs_external_attr_find()
993 unmap_extent_mft_record(ni); ntfs_external_attr_find()
997 ni = ctx->ntfs_ino = base_ni; ntfs_external_attr_find()
1003 al_entry->mft_reference), &ni); ntfs_external_attr_find()
1016 ni = NULL; ntfs_external_attr_find()
1019 ctx->ntfs_ino = ni; ntfs_external_attr_find()
1090 if (ni != base_ni) { ntfs_external_attr_find()
1091 if (ni) ntfs_external_attr_find()
1092 unmap_extent_mft_record(ni); ntfs_external_attr_find()
1123 if (ni != base_ni) ntfs_external_attr_find()
1124 unmap_extent_mft_record(ni); ntfs_external_attr_find()
1212 * @ni: ntfs inode with which to initialize the search context
1215 * Initialize the attribute search context @ctx with @ni and @mrec.
1218 ntfs_inode *ni, MFT_RECORD *mrec) ntfs_attr_init_search_ctx()
1226 .ntfs_ino = ni, ntfs_attr_init_search_ctx()
1263 * @ni: ntfs inode with which to initialize the search context
1266 * Allocate a new attribute search context, initialize it with @ni and @mrec,
1269 ntfs_attr_search_ctx *ntfs_attr_get_search_ctx(ntfs_inode *ni, MFT_RECORD *mrec) ntfs_attr_get_search_ctx() argument
1275 ntfs_attr_init_search_ctx(ctx, ni, mrec); ntfs_attr_get_search_ctx()
1503 * @ni: ntfs inode describing the attribute to convert
1506 * Convert the resident ntfs attribute described by the ntfs inode @ni to a
1535 int ntfs_attr_make_non_resident(ntfs_inode *ni, const u32 data_size) ntfs_attr_make_non_resident() argument
1538 struct inode *vi = VFS_I(ni); ntfs_attr_make_non_resident()
1539 ntfs_volume *vol = ni->vol; ntfs_attr_make_non_resident()
1553 err = ntfs_attr_can_be_non_resident(vol, ni->type); ntfs_attr_make_non_resident()
1567 BUG_ON(NInoCompressed(ni)); ntfs_attr_make_non_resident()
1568 BUG_ON(NInoEncrypted(ni)); ntfs_attr_make_non_resident()
1607 down_write(&ni->runlist.lock); ntfs_attr_make_non_resident()
1608 if (!NInoAttr(ni)) ntfs_attr_make_non_resident()
1609 base_ni = ni; ntfs_attr_make_non_resident()
1611 base_ni = ni->ext.base_ntfs_ino; ntfs_attr_make_non_resident()
1624 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_make_non_resident()
1633 BUG_ON(NInoNonResident(ni)); ntfs_attr_make_non_resident()
1638 if (NInoSparse(ni) || NInoCompressed(ni)) ntfs_attr_make_non_resident()
1695 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_attr_make_non_resident()
1697 if (NInoCompressed(ni) || vol->major_ver < 3) ntfs_attr_make_non_resident()
1712 ni->runlist.rl = rl; ntfs_attr_make_non_resident()
1713 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_make_non_resident()
1714 ni->allocated_size = new_size; ntfs_attr_make_non_resident()
1715 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_attr_make_non_resident()
1716 ni->itype.compressed.size = ni->allocated_size; ntfs_attr_make_non_resident()
1718 ni->itype.compressed.block_size = 1U << (a->data. ntfs_attr_make_non_resident()
1721 ni->itype.compressed.block_size_bits = ntfs_attr_make_non_resident()
1722 ffs(ni->itype.compressed.block_size) - ntfs_attr_make_non_resident()
1724 ni->itype.compressed.block_clusters = 1U << ntfs_attr_make_non_resident()
1727 ni->itype.compressed.block_size = 0; ntfs_attr_make_non_resident()
1728 ni->itype.compressed.block_size_bits = 0; ntfs_attr_make_non_resident()
1729 ni->itype.compressed.block_clusters = 0; ntfs_attr_make_non_resident()
1731 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_attr_make_non_resident()
1733 vi->i_blocks = ni->allocated_size >> 9; ntfs_attr_make_non_resident()
1734 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_make_non_resident()
1741 NInoSetNonResident(ni); ntfs_attr_make_non_resident()
1747 up_write(&ni->runlist.lock); ntfs_attr_make_non_resident()
1790 (unsigned)le32_to_cpu(ni->type), ntfs_attr_make_non_resident()
1792 ((ni->type == AT_DATA) && ntfs_attr_make_non_resident()
1793 !ni->name_len) ? "FILE": "ATTRIBUTE"); ntfs_attr_make_non_resident()
1794 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_make_non_resident()
1795 ni->initialized_size = attr_size; ntfs_attr_make_non_resident()
1797 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_make_non_resident()
1813 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_make_non_resident()
1814 ni->allocated_size = arec_size - mp_ofs; ntfs_attr_make_non_resident()
1815 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_make_non_resident()
1824 ni->runlist.rl = NULL; ntfs_attr_make_non_resident()
1825 up_write(&ni->runlist.lock); ntfs_attr_make_non_resident()
1847 * @ni: ntfs inode of the attribute whose allocation to extend
1852 * Extend the allocated space of an attribute described by the ntfs inode @ni
1855 * inode @ni have sparse support enabled). If @data_start is >= 0, then the
1893 * Locking: This function takes the runlist lock of @ni for writing as well as
1904 s64 ntfs_attr_extend_allocation(ntfs_inode *ni, s64 new_alloc_size, ntfs_attr_extend_allocation() argument
1909 struct inode *vi = VFS_I(ni); ntfs_attr_extend_allocation()
1910 ntfs_volume *vol = ni->vol; ntfs_attr_extend_allocation()
1922 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation()
1923 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation()
1924 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation()
1929 (unsigned)le32_to_cpu(ni->type), ntfs_attr_extend_allocation()
1940 if (NInoNonResident(ni)) { ntfs_attr_extend_allocation()
1948 err = ntfs_attr_size_bounds_check(vol, ni->type, new_alloc_size); ntfs_attr_extend_allocation()
1951 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation()
1952 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation()
1953 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation()
1963 le32_to_cpu(ni->type)); ntfs_attr_extend_allocation()
1973 le32_to_cpu(ni->type)); ntfs_attr_extend_allocation()
1983 if (!NInoAttr(ni)) ntfs_attr_extend_allocation()
1984 base_ni = ni; ntfs_attr_extend_allocation()
1986 base_ni = ni->ext.base_ntfs_ino; ntfs_attr_extend_allocation()
1991 down_write(&ni->runlist.lock); ntfs_attr_extend_allocation()
2004 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2005 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation()
2006 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2011 vcn = NInoNonResident(ni) ? allocated_size >> vol->cluster_size_bits : ntfs_attr_extend_allocation()
2030 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_allocation()
2042 BUG_ON(NInoNonResident(ni)); ntfs_attr_extend_allocation()
2055 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2056 ni->allocated_size = le32_to_cpu(a->length) - ntfs_attr_extend_allocation()
2058 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2076 up_write(&ni->runlist.lock); ntfs_attr_extend_allocation()
2081 err = ntfs_attr_make_non_resident(ni, attr_len); ntfs_attr_extend_allocation()
2091 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2092 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation()
2093 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2100 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation()
2106 read_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2107 allocated_size = ni->allocated_size; ntfs_attr_extend_allocation()
2108 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2132 if (ni->type == AT_ATTRIBUTE_LIST || ntfs_attr_extend_allocation()
2133 ni->type == AT_STANDARD_INFORMATION) { ntfs_attr_extend_allocation()
2151 BUG_ON(!NInoNonResident(ni)); ntfs_attr_extend_allocation()
2164 if ((start >= 0 && start <= allocated_size) || ni->type != AT_DATA || ntfs_attr_extend_allocation()
2165 !NVolSparseEnabled(vol) || NInoSparseDisabled(ni)) ntfs_attr_extend_allocation()
2172 rl = ni->runlist.rl; ntfs_attr_extend_allocation()
2180 (rl->lcn == LCN_ENOENT && rl > ni->runlist.rl && ntfs_attr_extend_allocation()
2184 rl = ntfs_mapping_pairs_decompress(vol, a, ni->runlist.rl); ntfs_attr_extend_allocation()
2194 (unsigned)le32_to_cpu(ni->type), ntfs_attr_extend_allocation()
2200 ni->runlist.rl = rl; ntfs_attr_extend_allocation()
2213 while (rl->lcn < 0 && rl > ni->runlist.rl) ntfs_attr_extend_allocation()
2230 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation()
2235 rl = ntfs_runlists_merge(ni->runlist.rl, rl2); ntfs_attr_extend_allocation()
2243 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation()
2256 ni->runlist.rl = rl; ntfs_attr_extend_allocation()
2276 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation()
2311 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation()
2330 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, ntfs_attr_extend_allocation()
2337 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2338 ni->allocated_size = new_alloc_size; ntfs_attr_extend_allocation()
2341 * FIXME: This would fail if @ni is a directory, $MFT, or an index, ntfs_attr_extend_allocation()
2350 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_attr_extend_allocation()
2351 ni->itype.compressed.size += new_alloc_size - allocated_size; ntfs_attr_extend_allocation()
2353 cpu_to_sle64(ni->itype.compressed.size); ntfs_attr_extend_allocation()
2354 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_attr_extend_allocation()
2357 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2371 up_write(&ni->runlist.lock); ntfs_attr_extend_allocation()
2381 (unsigned)le32_to_cpu(ni->type), err); ntfs_attr_extend_allocation()
2385 if (ntfs_attr_lookup(ni->type, ni->name, ni->name_len, CASE_SENSITIVE, ntfs_attr_extend_allocation()
2391 write_lock_irqsave(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2392 ni->allocated_size = new_alloc_size; ntfs_attr_extend_allocation()
2394 * FIXME: This would fail if @ni is a directory... See above. ntfs_attr_extend_allocation()
2398 if (NInoSparse(ni) || NInoCompressed(ni)) { ntfs_attr_extend_allocation()
2399 ni->itype.compressed.size += new_alloc_size - ntfs_attr_extend_allocation()
2401 vi->i_blocks = ni->itype.compressed.size >> 9; ntfs_attr_extend_allocation()
2404 write_unlock_irqrestore(&ni->size_lock, flags); ntfs_attr_extend_allocation()
2407 up_write(&ni->runlist.lock); ntfs_attr_extend_allocation()
2419 if (ntfs_cluster_free(ni, ll, -1, ctx) < 0) { ntfs_attr_extend_allocation()
2433 if (ntfs_rl_truncate_nolock(vol, &ni->runlist, ll) || IS_ERR(m)) { ntfs_attr_extend_allocation()
2467 up_write(&ni->runlist.lock); ntfs_attr_extend_allocation()
2475 * @ni: ntfs inode describing the attribute to fill
2480 * Fill @cnt bytes of the attribute described by the ntfs inode @ni starting at
2493 int ntfs_attr_set(ntfs_inode *ni, const s64 ofs, const s64 cnt, const u8 val) ntfs_attr_set() argument
2495 ntfs_volume *vol = ni->vol; ntfs_attr_set()
2512 BUG_ON(NInoCompressed(ni)); ntfs_attr_set()
2513 BUG_ON(NInoEncrypted(ni)); ntfs_attr_set()
2514 mapping = VFS_I(ni)->i_mapping; ntfs_attr_set()
2522 if (unlikely(end > i_size_read(VFS_I(ni)))) { ntfs_attr_set()
1217 ntfs_attr_init_search_ctx(ntfs_attr_search_ctx *ctx, ntfs_inode *ni, MFT_RECORD *mrec) ntfs_attr_init_search_ctx() argument
H A Dattrib.h63 extern int ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn,
65 extern int ntfs_map_runlist(ntfs_inode *ni, VCN vcn);
67 extern LCN ntfs_attr_vcn_to_lcn_nolock(ntfs_inode *ni, const VCN vcn,
70 extern runlist_element *ntfs_attr_find_vcn_nolock(ntfs_inode *ni,
89 extern ntfs_attr_search_ctx *ntfs_attr_get_search_ctx(ntfs_inode *ni,
106 extern int ntfs_attr_make_non_resident(ntfs_inode *ni, const u32 data_size);
108 extern s64 ntfs_attr_extend_allocation(ntfs_inode *ni, s64 new_alloc_size,
111 extern int ntfs_attr_set(ntfs_inode *ni, const s64 ofs, const s64 cnt,
H A Dinode.h183 static inline int NIno##flag(ntfs_inode *ni) \
185 return test_bit(NI_##flag, &(ni)->state); \
187 static inline void NInoSet##flag(ntfs_inode *ni) \
189 set_bit(NI_##flag, &(ni)->state); \
191 static inline void NInoClear##flag(ntfs_inode *ni) \
193 clear_bit(NI_##flag, &(ni)->state); \
200 static inline int NInoTestSet##flag(ntfs_inode *ni) \
202 return test_and_set_bit(NI_##flag, &(ni)->state); \
204 static inline int NInoTestClear##flag(ntfs_inode *ni) \
206 return test_and_clear_bit(NI_##flag, &(ni)->state); \
245 static inline struct inode *VFS_I(ntfs_inode *ni) VFS_I() argument
247 return &((big_ntfs_inode *)ni)->vfs_inode; VFS_I()
284 extern void __ntfs_init_inode(struct super_block *sb, ntfs_inode *ni);
288 ntfs_inode *ni = NTFS_I(vi); ntfs_init_big_inode() local
291 __ntfs_init_inode(vi->i_sb, ni); ntfs_init_big_inode()
292 ni->mft_no = vi->i_ino; ntfs_init_big_inode()
297 extern void ntfs_clear_extent_inode(ntfs_inode *ni);
H A Dmft.c39 * @ni: ntfs inode whose mft record page to map
41 * This maps the page in which the mft record of the ntfs inode @ni is situated
47 static inline MFT_RECORD *map_mft_record_page(ntfs_inode *ni) map_mft_record_page() argument
50 ntfs_volume *vol = ni->vol; map_mft_record_page()
56 BUG_ON(ni->page); map_mft_record_page()
63 index = (u64)ni->mft_no << vol->mft_record_size_bits >> map_mft_record_page()
65 ofs = (ni->mft_no << vol->mft_record_size_bits) & ~PAGE_CACHE_MASK; map_mft_record_page()
79 "driver.", ni->mft_no); map_mft_record_page()
89 ni->page = page; map_mft_record_page()
90 ni->page_ofs = ofs; map_mft_record_page()
94 "Run chkdsk.", ni->mft_no); map_mft_record_page()
100 ni->page = NULL; map_mft_record_page()
101 ni->page_ofs = 0; map_mft_record_page()
107 * @ni: ntfs inode whose MFT record to map
155 MFT_RECORD *map_mft_record(ntfs_inode *ni) map_mft_record() argument
159 ntfs_debug("Entering for mft_no 0x%lx.", ni->mft_no); map_mft_record()
162 atomic_inc(&ni->count); map_mft_record()
165 mutex_lock(&ni->mrec_lock); map_mft_record()
167 m = map_mft_record_page(ni); map_mft_record()
171 mutex_unlock(&ni->mrec_lock); map_mft_record()
172 atomic_dec(&ni->count); map_mft_record()
173 ntfs_error(ni->vol->sb, "Failed with error code %lu.", -PTR_ERR(m)); map_mft_record()
179 * @ni: ntfs inode whose mft record page to unmap
181 * This unmaps the page in which the mft record of the ntfs inode @ni is
191 static inline void unmap_mft_record_page(ntfs_inode *ni) unmap_mft_record_page() argument
193 BUG_ON(!ni->page); unmap_mft_record_page()
196 ntfs_unmap_page(ni->page); unmap_mft_record_page()
197 ni->page = NULL; unmap_mft_record_page()
198 ni->page_ofs = 0; unmap_mft_record_page()
204 * @ni: ntfs inode whose MFT record to unmap
213 void unmap_mft_record(ntfs_inode *ni) unmap_mft_record() argument
215 struct page *page = ni->page; unmap_mft_record()
219 ntfs_debug("Entering for mft_no 0x%lx.", ni->mft_no); unmap_mft_record()
221 unmap_mft_record_page(ni); unmap_mft_record()
222 mutex_unlock(&ni->mrec_lock); unmap_mft_record()
223 atomic_dec(&ni->count); unmap_mft_record()
250 ntfs_inode *ni = NULL; map_extent_mft_record() local
272 ni = extent_nis[i]; map_extent_mft_record()
274 atomic_inc(&ni->count); map_extent_mft_record()
278 if (likely(ni != NULL)) { map_extent_mft_record()
282 m = map_mft_record(ni); map_extent_mft_record()
284 atomic_dec(&ni->count); map_extent_mft_record()
289 *ntfs_ino = ni; map_extent_mft_record()
292 unmap_mft_record(ni); map_extent_mft_record()
304 ni = ntfs_new_extent_inode(base_ni->vol->sb, mft_no); map_extent_mft_record()
305 if (unlikely(!ni)) { map_extent_mft_record()
310 ni->vol = base_ni->vol; map_extent_mft_record()
311 ni->seq_no = seq_no; map_extent_mft_record()
312 ni->nr_extents = -1; map_extent_mft_record()
313 ni->ext.base_ntfs_ino = base_ni; map_extent_mft_record()
315 m = map_mft_record(ni); map_extent_mft_record()
319 ntfs_clear_extent_inode(ni); map_extent_mft_record()
351 base_ni->ext.extent_ntfs_inos[base_ni->nr_extents++] = ni; map_extent_mft_record()
355 *ntfs_ino = ni; map_extent_mft_record()
358 unmap_mft_record(ni); map_extent_mft_record()
366 ntfs_clear_extent_inode(ni); map_extent_mft_record()
374 * @ni: ntfs inode describing the mapped mft record
378 * Set the mapped (extent) mft record of the (base or extent) ntfs inode @ni,
395 void __mark_mft_record_dirty(ntfs_inode *ni) __mark_mft_record_dirty() argument
399 ntfs_debug("Entering for inode 0x%lx.", ni->mft_no); __mark_mft_record_dirty()
400 BUG_ON(NInoAttr(ni)); __mark_mft_record_dirty()
401 mark_ntfs_record_dirty(ni->page, ni->page_ofs); __mark_mft_record_dirty()
403 mutex_lock(&ni->extent_lock); __mark_mft_record_dirty()
404 if (likely(ni->nr_extents >= 0)) __mark_mft_record_dirty()
405 base_ni = ni; __mark_mft_record_dirty()
407 base_ni = ni->ext.base_ntfs_ino; __mark_mft_record_dirty()
408 mutex_unlock(&ni->extent_lock); __mark_mft_record_dirty()
640 * @ni: ntfs inode describing the mapped (extent) mft record
645 * ntfs inode @ni to backing store. If the mft record @m has a counterpart in
648 * We only write the mft record if the ntfs inode @ni is dirty and the first
669 int write_mft_record_nolock(ntfs_inode *ni, MFT_RECORD *m, int sync) write_mft_record_nolock() argument
671 ntfs_volume *vol = ni->vol; write_mft_record_nolock()
672 struct page *page = ni->page; write_mft_record_nolock()
682 ntfs_debug("Entering for inode 0x%lx.", ni->mft_no); write_mft_record_nolock()
683 BUG_ON(NInoAttr(ni)); write_mft_record_nolock()
692 if (!NInoTestClearDirty(ni)) write_mft_record_nolock()
699 m_start = ni->page_ofs; write_mft_record_nolock()
729 vcn = ((VCN)ni->mft_no << vol->mft_record_size_bits) + write_mft_record_nolock()
755 ni->mft_no, (long long)lcn); write_mft_record_nolock()
777 flush_dcache_mft_record_page(ni); write_mft_record_nolock()
791 if (!sync && ni->mft_no < vol->mftmirr_size) write_mft_record_nolock()
792 ntfs_sync_mft_mirror(vol, ni->mft_no, m, sync); write_mft_record_nolock()
809 if (sync && ni->mft_no < vol->mftmirr_size) write_mft_record_nolock()
810 ntfs_sync_mft_mirror(vol, ni->mft_no, m, sync); write_mft_record_nolock()
813 flush_dcache_mft_record_page(ni); write_mft_record_nolock()
819 ni->mft_no); write_mft_record_nolock()
839 mark_mft_record_dirty(ni); write_mft_record_nolock()
931 ntfs_inode *ni, *eni, **extent_nis; ntfs_may_write_mft_record() local
971 ni = NTFS_I(vi); ntfs_may_write_mft_record()
973 atomic_inc(&ni->count); ntfs_may_write_mft_record()
975 if (NInoDirty(ni)) { ntfs_may_write_mft_record()
978 atomic_dec(&ni->count); ntfs_may_write_mft_record()
984 if (unlikely(!mutex_trylock(&ni->mrec_lock))) { ntfs_may_write_mft_record()
987 atomic_dec(&ni->count); ntfs_may_write_mft_record()
997 *locked_ni = ni; ntfs_may_write_mft_record()
1043 ni = NTFS_I(vi); ntfs_may_write_mft_record()
1044 mutex_lock(&ni->extent_lock); ntfs_may_write_mft_record()
1045 if (ni->nr_extents <= 0) { ntfs_may_write_mft_record()
1050 mutex_unlock(&ni->extent_lock); ntfs_may_write_mft_record()
1057 extent_nis = ni->ext.extent_ntfs_inos; ntfs_may_write_mft_record()
1058 for (eni = NULL, i = 0; i < ni->nr_extents; ++i) { ntfs_may_write_mft_record()
1073 mutex_unlock(&ni->extent_lock); ntfs_may_write_mft_record()
1084 mutex_unlock(&ni->extent_lock); ntfs_may_write_mft_record()
2250 ntfs_inode *mft_ni, *mftbmp_ni, *ni; ntfs_mft_record_alloc() local
2593 m_tmp = map_extent_mft_record(base_ni, bit, &ni); ntfs_mft_record_alloc()
2656 ni = NTFS_I(vi); ntfs_mft_record_alloc()
2665 NInoSetMstProtected(ni); ntfs_mft_record_alloc()
2666 ni->type = AT_INDEX_ALLOCATION; ntfs_mft_record_alloc()
2667 ni->name = I30; ntfs_mft_record_alloc()
2668 ni->name_len = 4; ntfs_mft_record_alloc()
2670 ni->itype.index.block_size = 4096; ntfs_mft_record_alloc()
2671 ni->itype.index.block_size_bits = ntfs_ffs(4096) - 1; ntfs_mft_record_alloc()
2672 ni->itype.index.collation_rule = COLLATION_FILE_NAME; ntfs_mft_record_alloc()
2673 if (vol->cluster_size <= ni->itype.index.block_size) { ntfs_mft_record_alloc()
2674 ni->itype.index.vcn_size = vol->cluster_size; ntfs_mft_record_alloc()
2675 ni->itype.index.vcn_size_bits = ntfs_mft_record_alloc()
2678 ni->itype.index.vcn_size = vol->sector_size; ntfs_mft_record_alloc()
2679 ni->itype.index.vcn_size_bits = ntfs_mft_record_alloc()
2686 ni->type = AT_DATA; ntfs_mft_record_alloc()
2687 ni->name = NULL; ntfs_mft_record_alloc()
2688 ni->name_len = 0; ntfs_mft_record_alloc()
2704 vi->i_generation = ni->seq_no = le16_to_cpu(m->sequence_number); ntfs_mft_record_alloc()
2709 atomic_inc(&ni->count); ntfs_mft_record_alloc()
2710 mutex_lock(&ni->mrec_lock); ntfs_mft_record_alloc()
2711 ni->page = page; ntfs_mft_record_alloc()
2712 ni->page_ofs = ofs; ntfs_mft_record_alloc()
2739 return ni; ntfs_mft_record_alloc()
2765 * @ni: ntfs inode of the mapped extent mft record to free
2766 * @m: mapped extent mft record of the ntfs inode @ni
2768 * Free the mapped extent mft record @m of the extent ntfs inode @ni.
2770 * Note that this function unmaps the mft record and closes and destroys @ni
2771 * internally and hence you cannot use either @ni nor @m any more after this
2774 * On success return 0 and on error return -errno. @ni and @m are still valid
2784 int ntfs_extent_mft_record_free(ntfs_inode *ni, MFT_RECORD *m) ntfs_extent_mft_record_free() argument
2786 unsigned long mft_no = ni->mft_no; ntfs_extent_mft_record_free()
2787 ntfs_volume *vol = ni->vol; ntfs_extent_mft_record_free()
2794 BUG_ON(NInoAttr(ni)); ntfs_extent_mft_record_free()
2795 BUG_ON(ni->nr_extents != -1); ntfs_extent_mft_record_free()
2797 mutex_lock(&ni->extent_lock); ntfs_extent_mft_record_free()
2798 base_ni = ni->ext.base_ntfs_ino; ntfs_extent_mft_record_free()
2799 mutex_unlock(&ni->extent_lock); ntfs_extent_mft_record_free()
2809 if (atomic_read(&ni->count) > 2) { ntfs_extent_mft_record_free()
2820 if (ni != extent_nis[i]) ntfs_extent_mft_record_free()
2861 NInoSetDirty(ni); ntfs_extent_mft_record_free()
2862 err = write_mft_record(ni, m, 0); ntfs_extent_mft_record_free()
2870 unmap_extent_mft_record(ni); ntfs_extent_mft_record_free()
2871 ntfs_clear_extent_inode(ni); ntfs_extent_mft_record_free()
2912 extent_nis[base_ni->nr_extents++] = ni; ntfs_extent_mft_record_free()
2914 mark_mft_record_dirty(ni); ntfs_extent_mft_record_free()
H A Dlcnalloc.h48 extern s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn,
53 * @ni: ntfs inode whose runlist describes the clusters to free
54 * @start_vcn: vcn in the runlist of @ni at which to start freeing clusters
59 * described by the ntfs inode @ni.
65 * If @ctx is specified, it is an active search context of @ni and its base mft
95 * Locking: - The runlist described by @ni must be locked for writing on entry
102 * - If @ctx is NULL, the base mft record of @ni must not be mapped on
107 static inline s64 ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, ntfs_cluster_free() argument
110 return __ntfs_cluster_free(ni, start_vcn, count, ctx, false); ntfs_cluster_free()
H A Dcompress.c487 ntfs_inode *ni = NTFS_I(mapping->host); ntfs_read_compressed_block() local
488 ntfs_volume *vol = ni->vol; ntfs_read_compressed_block()
496 u32 cb_size = ni->itype.compressed.block_size; ntfs_read_compressed_block()
511 >> ni->itype.compressed.block_size_bits; ntfs_read_compressed_block()
531 BUG_ON(ni->type != AT_DATA); ntfs_read_compressed_block()
532 BUG_ON(ni->name_len); ntfs_read_compressed_block()
559 read_lock_irqsave(&ni->size_lock, flags); ntfs_read_compressed_block()
560 i_size = i_size_read(VFS_I(ni)); ntfs_read_compressed_block()
561 initialized_size = ni->initialized_size; ntfs_read_compressed_block()
562 read_unlock_irqrestore(&ni->size_lock, flags); ntfs_read_compressed_block()
605 cb_clusters = ni->itype.compressed.block_clusters; ntfs_read_compressed_block()
618 down_read(&ni->runlist.lock); ntfs_read_compressed_block()
619 rl = ni->runlist.rl; ntfs_read_compressed_block()
645 up_read(&ni->runlist.lock); ntfs_read_compressed_block()
646 if (!ntfs_map_runlist(ni, vcn)) ntfs_read_compressed_block()
663 up_read(&ni->runlist.lock); ntfs_read_compressed_block()
878 ni->mft_no, -err); ntfs_read_compressed_block()
912 "0x%lx.", ni->mft_no, page->index); ntfs_read_compressed_block()
946 up_read(&ni->runlist.lock); ntfs_read_compressed_block()
952 up_read(&ni->runlist.lock); ntfs_read_compressed_block()
H A Dnamei.c180 ntfs_inode *ni = NTFS_I(dent_inode); ntfs_lookup() local
198 ni = NTFS_I(dent_inode); ntfs_lookup()
199 m = map_mft_record(ni); ntfs_lookup()
206 ctx = ntfs_attr_get_search_ctx(ni, m); ntfs_lookup()
246 unmap_mft_record(ni); ntfs_lookup()
269 unmap_mft_record(ni); ntfs_lookup()
303 ntfs_inode *ni = NTFS_I(vi); ntfs_get_parent() local
313 mrec = map_mft_record(ni); ntfs_get_parent()
317 ctx = ntfs_attr_get_search_ctx(ni, mrec); ntfs_get_parent()
319 unmap_mft_record(ni); ntfs_get_parent()
327 unmap_mft_record(ni); ntfs_get_parent()
346 unmap_mft_record(ni); ntfs_get_parent()
H A Dlcnalloc.c791 * @ni: ntfs inode whose runlist describes the clusters to free
792 * @start_vcn: vcn in the runlist of @ni at which to start freeing clusters
798 * described by the vfs inode @ni.
804 * If @ctx is specified, it is an active search context of @ni and its base mft
837 * Locking: - The runlist described by @ni must be locked for writing on entry
844 * - If @ctx is NULL, the base mft record of @ni must not be mapped on
849 s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, s64 count, __ntfs_cluster_free() argument
858 BUG_ON(!ni); __ntfs_cluster_free()
860 "0x%llx.%s", ni->mft_no, (unsigned long long)start_vcn, __ntfs_cluster_free()
863 vol = ni->vol; __ntfs_cluster_free()
880 rl = ntfs_attr_find_vcn_nolock(ni, start_vcn, ctx); __ntfs_cluster_free()
934 rl = ntfs_attr_find_vcn_nolock(ni, vcn, ctx); __ntfs_cluster_free()
1002 delta = __ntfs_cluster_free(ni, start_vcn, total_freed, ctx, true); __ntfs_cluster_free()
H A Dsuper.c377 ntfs_inode *ni = NTFS_I(vol->vol_ino); ntfs_write_volume_flags() local
387 BUG_ON(!ni); ntfs_write_volume_flags()
388 m = map_mft_record(ni); ntfs_write_volume_flags()
393 ctx = ntfs_attr_get_search_ctx(ni, m); ntfs_write_volume_flags()
408 unmap_mft_record(ni); ntfs_write_volume_flags()
415 unmap_mft_record(ni); ntfs_write_volume_flags()
3050 ntfs_inode *ni = (ntfs_inode *)foo; ntfs_big_inode_init_once() local
3052 inode_init_once(VFS_I(ni)); ntfs_big_inode_init_once()
/linux-4.4.14/fs/f2fs/
H A Dnode.h56 struct node_info ni; /* in-memory node information */ member in struct:nat_entry
59 #define nat_get_nid(nat) (nat->ni.nid)
60 #define nat_set_nid(nat, n) (nat->ni.nid = n)
61 #define nat_get_blkaddr(nat) (nat->ni.blk_addr)
62 #define nat_set_blkaddr(nat, b) (nat->ni.blk_addr = b)
63 #define nat_get_ino(nat) (nat->ni.ino)
64 #define nat_set_ino(nat, i) (nat->ni.ino = i)
65 #define nat_get_version(nat) (nat->ni.version)
66 #define nat_set_version(nat, v) (nat->ni.version = v)
85 ne->ni.flag |= mask; set_nat_flag()
87 ne->ni.flag &= ~mask; set_nat_flag()
93 return ne->ni.flag & mask; get_nat_flag()
104 static inline void node_info_from_raw_nat(struct node_info *ni, node_info_from_raw_nat() argument
107 ni->ino = le32_to_cpu(raw_ne->ino); node_info_from_raw_nat()
108 ni->blk_addr = le32_to_cpu(raw_ne->block_addr); node_info_from_raw_nat()
109 ni->version = raw_ne->version; node_info_from_raw_nat()
113 struct node_info *ni) raw_nat_from_node_info()
115 raw_ne->ino = cpu_to_le32(ni->ino); raw_nat_from_node_info()
116 raw_ne->block_addr = cpu_to_le32(ni->blk_addr); raw_nat_from_node_info()
117 raw_ne->version = ni->version; raw_nat_from_node_info()
112 raw_nat_from_node_info(struct f2fs_nat_entry *raw_ne, struct node_info *ni) raw_nat_from_node_info() argument
H A Dnode.c154 nid_t set = NAT_BLOCK_OFFSET(ne->ni.nid); __set_nat_cache_dirty()
179 nid_t set = NAT_BLOCK_OFFSET(ne->ni.nid); __clear_nat_cache_dirty()
268 node_info_from_raw_nat(&e->ni, ne); cache_nat_entry()
273 static void set_node_addr(struct f2fs_sb_info *sbi, struct node_info *ni, set_node_addr() argument
280 e = __lookup_nat_cache(nm_i, ni->nid); set_node_addr()
282 e = grab_nat_entry(nm_i, ni->nid); set_node_addr()
283 copy_node_info(&e->ni, ni); set_node_addr()
284 f2fs_bug_on(sbi, ni->blk_addr == NEW_ADDR); set_node_addr()
291 copy_node_info(&e->ni, ni); set_node_addr()
292 f2fs_bug_on(sbi, ni->blk_addr != NULL_ADDR); set_node_addr()
296 f2fs_bug_on(sbi, nat_get_blkaddr(e) != ni->blk_addr); set_node_addr()
311 if (nm_i->next_scan_nid > ni->nid) set_node_addr()
312 nm_i->next_scan_nid = ni->nid; set_node_addr()
322 if (ni->nid != ni->ino) set_node_addr()
323 e = __lookup_nat_cache(nm_i, ni->ino); set_node_addr()
325 if (fsync_done && ni->nid == ni->ino) set_node_addr()
354 void get_node_info(struct f2fs_sb_info *sbi, nid_t nid, struct node_info *ni) get_node_info() argument
366 ni->nid = nid; get_node_info()
372 ni->ino = nat_get_ino(e); get_node_info()
373 ni->blk_addr = nat_get_blkaddr(e); get_node_info()
374 ni->version = nat_get_version(e); get_node_info()
387 node_info_from_raw_nat(ni, &ne); get_node_info()
397 node_info_from_raw_nat(ni, &ne); get_node_info()
592 struct node_info ni; truncate_node() local
594 get_node_info(sbi, dn->nid, &ni); truncate_node()
596 f2fs_bug_on(sbi, ni.blk_addr != NULL_ADDR); truncate_node()
599 f2fs_bug_on(sbi, ni.blk_addr == NULL_ADDR); truncate_node()
602 invalidate_blocks(sbi, ni.blk_addr); truncate_node()
604 set_node_addr(sbi, &ni, NULL_ADDR, false); truncate_node()
622 trace_f2fs_truncate_node(dn->inode, dn->nid, ni.blk_addr); truncate_node()
1007 struct node_info ni; read_node_page() local
1016 get_node_info(sbi, page->index, &ni); read_node_page()
1018 if (unlikely(ni.blk_addr == NULL_ADDR)) { read_node_page()
1026 fio.blk_addr = ni.blk_addr; read_node_page()
1304 struct node_info ni; f2fs_write_node_page() local
1333 get_node_info(sbi, nid, &ni); f2fs_write_node_page()
1336 if (unlikely(ni.blk_addr == NULL_ADDR)) { f2fs_write_node_page()
1345 fio.blk_addr = ni.blk_addr; f2fs_write_node_page()
1347 set_node_addr(sbi, &ni, fio.blk_addr, is_fsync_dnode(page)); f2fs_write_node_page()
1585 struct node_info ni; alloc_nid() local
1599 get_node_info(sbi, *nid, &ni); alloc_nid()
1600 if (ni.blk_addr != NULL_ADDR) { alloc_nid()
1718 struct node_info ni; recover_xattr_data() local
1725 get_node_info(sbi, prev_xnid, &ni); recover_xattr_data()
1726 f2fs_bug_on(sbi, ni.blk_addr == NULL_ADDR); recover_xattr_data()
1727 invalidate_blocks(sbi, ni.blk_addr); recover_xattr_data()
1729 set_node_addr(sbi, &ni, NULL_ADDR, false); recover_xattr_data()
1737 get_node_info(sbi, new_xnid, &ni); recover_xattr_data()
1738 ni.ino = inode->i_ino; recover_xattr_data()
1739 set_node_addr(sbi, &ni, NEW_ADDR, false); recover_xattr_data()
1744 set_node_addr(sbi, &ni, blkaddr, false); recover_xattr_data()
1849 node_info_from_raw_nat(&ne->ni, &raw_ne); remove_nats_in_journal()
1922 raw_nat_from_node_info(raw_ne, &ne->ni); __flush_nat_entry_set()
H A Drecovery.c363 struct node_info ni; do_recover_data() local
394 get_node_info(sbi, dn.nid, &ni); do_recover_data()
395 f2fs_bug_on(sbi, ni.ino != ino_of_node(page)); do_recover_data()
441 ni.version, false); do_recover_data()
450 fill_node_footer(dn.node_page, dn.nid, ni.ino, do_recover_data()
H A Dgc.c419 struct node_info ni; gc_node_segment() local
442 get_node_info(sbi, nid, &ni); gc_node_segment()
443 if (ni.blk_addr != start_addr + off) { gc_node_segment()
548 struct node_info ni; move_encrypted_block() local
573 get_node_info(fio.sbi, dn.nid, &ni); move_encrypted_block()
574 set_summary(&sum, dn.nid, dn.ofs_in_node, ni.version); move_encrypted_block()
H A Dinline.c581 struct node_info ni; f2fs_inline_data_fiemap() local
601 get_node_info(F2FS_I_SB(inode), inode->i_ino, &ni); f2fs_inline_data_fiemap() local
602 byteaddr = (__u64)ni.blk_addr << inode->i_sb->s_blocksize_bits; f2fs_inline_data_fiemap()
H A Ddata.c466 struct node_info ni; __allocate_data_block() local
481 get_node_info(sbi, dn->nid, &ni); __allocate_data_block()
482 set_summary(&sum, dn->nid, dn->ofs_in_node, ni.version); __allocate_data_block()
H A Dfile.c866 struct node_info ni; __exchange_data_block() local
880 get_node_info(sbi, dn.nid, &ni); __exchange_data_block()
882 ni.version, true); __exchange_data_block()
H A Dsegment.c1332 struct node_info ni; write_data_page() local
1335 get_node_info(sbi, dn->nid, &ni); write_data_page()
1336 set_summary(&sum, dn->nid, dn->ofs_in_node, ni.version); write_data_page()
/linux-4.4.14/include/uapi/linux/netfilter_bridge/
H A Debt_802_3.h28 /* ui has one byte ctrl, ni has two */
51 struct hdr_ni ni; member in union:ebt_802_3_hdr::__anon14322
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/socklnd/
H A Dsocklnd.c50 ksocknal_ip2iface(lnet_ni_t *ni, __u32 ip) ksocknal_ip2iface() argument
52 ksock_net_t *net = ni->ni_data; ksocknal_ip2iface()
103 ksocknal_create_peer(ksock_peer_t **peerp, lnet_ni_t *ni, lnet_process_id_t id) ksocknal_create_peer() argument
105 ksock_net_t *net = ni->ni_data; ksocknal_create_peer()
116 peer->ksnp_ni = ni; ksocknal_create_peer()
176 ksocknal_find_peer_locked(lnet_ni_t *ni, lnet_process_id_t id) ksocknal_find_peer_locked() argument
188 if (peer->ksnp_ni != ni) list_for_each()
204 ksocknal_find_peer(lnet_ni_t *ni, lnet_process_id_t id) ksocknal_find_peer() argument
209 peer = ksocknal_find_peer_locked(ni, id); ksocknal_find_peer()
248 ksocknal_get_peer_info(lnet_ni_t *ni, int index, ksocknal_get_peer_info() argument
267 if (peer->ksnp_ni != ni) ksocknal_get_peer_info()
448 ksocknal_add_peer(lnet_ni_t *ni, lnet_process_id_t id, __u32 ipaddr, int port) ksocknal_add_peer() argument
462 rc = ksocknal_create_peer(&peer, ni, id); ksocknal_add_peer()
474 /* always called with a ref on ni, so shutdown can't have started */ ksocknal_add_peer()
475 LASSERT(((ksock_net_t *) ni->ni_data)->ksnn_shutdown == 0); ksocknal_add_peer()
477 peer2 = ksocknal_find_peer_locked(ni, id); ksocknal_add_peer()
565 ksocknal_del_peer(lnet_ni_t *ni, lnet_process_id_t id, __u32 ip) ksocknal_del_peer() argument
590 if (peer->ksnp_ni != ni) ksocknal_del_peer()
618 ksocknal_txlist_done(ni, &zombies, 1); ksocknal_del_peer()
624 ksocknal_get_conn_by_idx(lnet_ni_t *ni, int index) ksocknal_get_conn_by_idx() argument
640 if (peer->ksnp_ni != ni) ksocknal_get_conn_by_idx()
684 ksocknal_local_ipvec(lnet_ni_t *ni, __u32 *ipaddrs) ksocknal_local_ipvec() argument
686 ksock_net_t *net = ni->ni_data; ksocknal_local_ipvec()
851 lnet_ni_t *ni = peer->ksnp_ni; ksocknal_create_routes() local
852 ksock_net_t *net = ni->ni_data; ksocknal_create_routes()
964 ksocknal_accept(lnet_ni_t *ni, struct socket *sock) ksocknal_accept() argument
981 lnet_ni_addref(ni); ksocknal_accept()
982 cr->ksncr_ni = ni; ksocknal_accept()
1008 ksocknal_create_conn(lnet_ni_t *ni, ksock_route_t *route, ksocknal_create_conn() argument
1077 LASSERT(ni == peer->ksnp_ni); ksocknal_create_conn()
1080 hello->kshm_nips = ksocknal_local_ipvec(ni, hello->kshm_ips); ksocknal_create_conn()
1097 rc = ksocknal_send_hello(ni, conn, peerid.nid, hello); ksocknal_create_conn()
1108 rc = ksocknal_recv_hello(ni, conn, hello, &peerid, &incarnation); ksocknal_create_conn()
1122 rc = ksocknal_create_peer(&peer, ni, peerid); ksocknal_create_conn()
1128 /* called with a ref on ni, so shutdown can't have started */ ksocknal_create_conn()
1129 LASSERT(((ksock_net_t *) ni->ni_data)->ksnn_shutdown == 0); ksocknal_create_conn()
1131 peer2 = ksocknal_find_peer_locked(ni, peerid); ksocknal_create_conn()
1148 if (peerid.nid < ni->ni_nid && ksocknal_create_conn()
1302 rc = ksocknal_send_hello(ni, conn, peerid.nid, hello); ksocknal_create_conn()
1370 ksocknal_send_hello(ni, conn, peerid.nid, hello); ksocknal_create_conn()
1378 ksocknal_txlist_done(ni, &zombies, 1); ksocknal_create_conn()
1761 ksocknal_notify(lnet_ni_t *ni, lnet_nid_t gw_nid, int alive) ksocknal_notify() argument
1784 ksocknal_query(lnet_ni_t *ni, lnet_nid_t nid, unsigned long *when) ksocknal_query() argument
1795 peer = ksocknal_find_peer_locked(ni, id); ksocknal_query()
1832 ksocknal_add_peer(ni, id, LNET_NIDADDR(nid), lnet_acceptor_port()); ksocknal_query()
1836 peer = ksocknal_find_peer_locked(ni, id); ksocknal_query()
1876 static int ksocknal_push(lnet_ni_t *ni, lnet_process_id_t id) ksocknal_push() argument
1925 ksocknal_add_interface(lnet_ni_t *ni, __u32 ipaddress, __u32 netmask) ksocknal_add_interface() argument
1927 ksock_net_t *net = ni->ni_data; ksocknal_add_interface()
1943 iface = ksocknal_ip2iface(ni, ipaddress); ksocknal_add_interface()
2028 ksocknal_del_interface(lnet_ni_t *ni, __u32 ipaddress) ksocknal_del_interface() argument
2030 ksock_net_t *net = ni->ni_data; ksocknal_del_interface()
2062 if (peer->ksnp_ni != ni) ksocknal_del_interface()
2076 ksocknal_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg) ksocknal_ctl() argument
2084 ksock_net_t *net = ni->ni_data; ksocknal_ctl()
2106 return ksocknal_add_interface(ni, ksocknal_ctl()
2111 return ksocknal_del_interface(ni, ksocknal_ctl()
2121 rc = ksocknal_get_peer_info(ni, data->ioc_count, ksocknal_ctl()
2140 return ksocknal_add_peer(ni, id, ksocknal_ctl()
2147 return ksocknal_del_peer(ni, id, ksocknal_ctl()
2154 ksock_conn_t *conn = ksocknal_get_conn_by_idx(ni, data->ioc_count); ksocknal_ctl()
2183 if (data->ioc_nid == ni->ni_nid) ksocknal_ctl()
2188 libcfs_nid2str(ni->ni_nid)); ksocknal_ctl()
2194 return ksocknal_push(ni, id); ksocknal_ctl()
2475 ksocknal_debug_peerhash(lnet_ni_t *ni) ksocknal_debug_peerhash() argument
2487 if (peer->ksnp_ni == ni) ksocknal_debug_peerhash()
2529 ksocknal_shutdown(lnet_ni_t *ni) ksocknal_shutdown() argument
2531 ksock_net_t *net = ni->ni_data; ksocknal_shutdown()
2546 ksocknal_del_peer(ni, anyid, 0); ksocknal_shutdown()
2561 ksocknal_debug_peerhash(ni); ksocknal_shutdown()
2753 ksocknal_startup(lnet_ni_t *ni) ksocknal_startup() argument
2759 LASSERT(ni->ni_lnd == &the_ksocklnd); ksocknal_startup()
2773 ni->ni_data = net; ksocknal_startup()
2774 ni->ni_peertimeout = *ksocknal_tunables.ksnd_peertimeout; ksocknal_startup()
2775 ni->ni_maxtxcredits = *ksocknal_tunables.ksnd_credits; ksocknal_startup()
2776 ni->ni_peertxcredits = *ksocknal_tunables.ksnd_peertxcredits; ksocknal_startup()
2777 ni->ni_peerrtrcredits = *ksocknal_tunables.ksnd_peerrtrcredits; ksocknal_startup()
2779 if (ni->ni_interfaces[0] == NULL) { ksocknal_startup()
2789 if (ni->ni_interfaces[i] == NULL) ksocknal_startup()
2792 rc = lnet_ipif_query(ni->ni_interfaces[i], &up, ksocknal_startup()
2798 ni->ni_interfaces[i], rc); ksocknal_startup()
2804 ni->ni_interfaces[i]); ksocknal_startup()
2809 ni->ni_interfaces[i], IFNAMSIZ); ksocknal_startup()
2815 rc = ksocknal_net_start_threads(net, ni->ni_cpts, ni->ni_ncpts); ksocknal_startup()
2819 ni->ni_nid = LNET_MKNID(LNET_NIDNET(ni->ni_nid), ksocknal_startup()
H A Dsocklnd.h570 void ksocknal_tx_done(lnet_ni_t *ni, ksock_tx_t *tx);
614 int ksocknal_startup(lnet_ni_t *ni);
615 void ksocknal_shutdown(lnet_ni_t *ni);
616 int ksocknal_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg);
617 int ksocknal_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg);
618 int ksocknal_recv(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg,
622 int ksocknal_accept(lnet_ni_t *ni, struct socket *sock);
624 int ksocknal_add_peer(lnet_ni_t *ni, lnet_process_id_t id, __u32 ip, int port);
625 ksock_peer_t *ksocknal_find_peer_locked(lnet_ni_t *ni, lnet_process_id_t id);
626 ksock_peer_t *ksocknal_find_peer(lnet_ni_t *ni, lnet_process_id_t id);
628 int ksocknal_create_conn(lnet_ni_t *ni, ksock_route_t *route,
640 int ksocknal_launch_packet(lnet_ni_t *ni, ksock_tx_t *tx,
647 void ksocknal_txlist_done(lnet_ni_t *ni, struct list_head *txlist, int error);
648 void ksocknal_notify(lnet_ni_t *ni, lnet_nid_t gw_nid, int alive);
649 void ksocknal_query(struct lnet_ni *ni, lnet_nid_t nid, unsigned long *when);
659 int ksocknal_send_hello(lnet_ni_t *ni, ksock_conn_t *conn,
661 int ksocknal_recv_hello(lnet_ni_t *ni, ksock_conn_t *conn,
H A Dsocklnd_cb.c389 ksocknal_tx_done (lnet_ni_t *ni, ksock_tx_t *tx) ksocknal_tx_done() argument
394 LASSERT(ni != NULL || tx->tx_conn != NULL); ksocknal_tx_done()
399 if (ni == NULL && tx->tx_conn != NULL) ksocknal_tx_done()
400 ni = tx->tx_conn->ksnc_peer->ksnp_ni; ksocknal_tx_done()
404 lnet_finalize (ni, lnetmsg, rc); ksocknal_tx_done()
408 ksocknal_txlist_done (lnet_ni_t *ni, struct list_head *txlist, int error) ksocknal_txlist_done() argument
428 ksocknal_tx_done(ni, tx); ksocknal_txlist_done()
830 ksocknal_launch_packet (lnet_ni_t *ni, ksock_tx_t *tx, lnet_process_id_t id) ksocknal_launch_packet() argument
844 peer = ksocknal_find_peer_locked(ni, id); ksocknal_launch_packet()
864 peer = ksocknal_find_peer_locked(ni, id); ksocknal_launch_packet()
881 rc = ksocknal_add_peer(ni, id, ksocknal_launch_packet()
921 ksocknal_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg) ksocknal_send() argument
991 rc = ksocknal_launch_packet(ni, tx, target); ksocknal_send()
1311 ksocknal_recv (lnet_ni_t *ni, void *private, lnet_msg_t *msg, int delayed, ksocknal_recv() argument
1647 ksocknal_send_hello (lnet_ni_t *ni, ksock_conn_t *conn, ksocknal_send_hello() argument
1651 ksock_net_t *net = (ksock_net_t *)ni->ni_data; ksocknal_send_hello()
1658 hello->kshm_src_nid = ni->ni_nid; ksocknal_send_hello()
1685 ksocknal_recv_hello (lnet_ni_t *ni, ksock_conn_t *conn, ksocknal_recv_hello() argument
1748 ksocknal_send_hello(ni, conn, ni->ni_nid, hello); ksocknal_recv_hello()
1782 recv_id.nid = LNET_MKNID(LNET_NIDNET(ni->ni_nid), conn->ksnc_ipaddr); ksocknal_recv_hello()
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/
H A Do2iblnd_cb.c43 static void kiblnd_unmap_tx(lnet_ni_t *ni, kib_tx_t *tx);
46 kiblnd_tx_done(lnet_ni_t *ni, kib_tx_t *tx) kiblnd_tx_done() argument
49 kib_net_t *net = ni->ni_data; kiblnd_tx_done()
60 kiblnd_unmap_tx(ni, tx); kiblnd_tx_done()
68 LASSERT(ni == tx->tx_conn->ibc_peer->ibp_ni); kiblnd_tx_done()
84 lnet_finalize(ni, lntmsg[i], rc); kiblnd_tx_done()
89 kiblnd_txlist_done(lnet_ni_t *ni, struct list_head *txlist, int status) kiblnd_txlist_done() argument
100 kiblnd_tx_done(ni, tx); kiblnd_txlist_done()
105 kiblnd_get_idle_tx(lnet_ni_t *ni, lnet_nid_t target) kiblnd_get_idle_tx() argument
107 kib_net_t *net = (kib_net_t *)ni->ni_data; kiblnd_get_idle_tx()
248 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_handle_completion() local
267 lnet_set_reply_msg_len(ni, tx->tx_lntmsg[1], status); kiblnd_handle_completion()
279 kiblnd_tx_done(ni, tx); kiblnd_handle_completion()
285 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_send_completion() local
286 kib_tx_t *tx = kiblnd_get_idle_tx(ni, conn->ibc_peer->ibp_nid); kiblnd_send_completion()
296 kiblnd_init_tx_msg(ni, tx, type, sizeof(kib_completion_msg_t)); kiblnd_send_completion()
306 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_handle_rx() local
371 rc = lnet_parse(ni, &msg->ibm_u.immediate.ibim_hdr, kiblnd_handle_rx()
379 rc = lnet_parse(ni, &msg->ibm_u.putreq.ibprm_hdr, kiblnd_handle_rx()
441 rc = lnet_parse(ni, &msg->ibm_u.get.ibgm_hdr, kiblnd_handle_rx()
467 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_rx_complete() local
468 kib_net_t *net = ni->ni_data; kiblnd_rx_complete()
496 msg->ibm_dstnid != ni->ni_nid || kiblnd_rx_complete()
601 static void kiblnd_unmap_tx(lnet_ni_t *ni, kib_tx_t *tx) kiblnd_unmap_tx() argument
603 kib_net_t *net = ni->ni_data; kiblnd_unmap_tx()
619 static int kiblnd_map_tx(lnet_ni_t *ni, kib_tx_t *tx, kib_rdma_desc_t *rd, kiblnd_map_tx() argument
623 kib_net_t *net = ni->ni_data; kiblnd_map_tx()
659 kiblnd_setup_rd_iov(lnet_ni_t *ni, kib_tx_t *tx, kib_rdma_desc_t *rd, kiblnd_setup_rd_iov() argument
662 kib_net_t *net = ni->ni_data; kiblnd_setup_rd_iov()
708 return kiblnd_map_tx(ni, tx, rd, sg - tx->tx_frags); kiblnd_setup_rd_iov()
712 kiblnd_setup_rd_kiov(lnet_ni_t *ni, kib_tx_t *tx, kib_rdma_desc_t *rd, kiblnd_setup_rd_kiov() argument
715 kib_net_t *net = ni->ni_data; kiblnd_setup_rd_kiov()
748 return kiblnd_map_tx(ni, tx, rd, sg - tx->tx_frags); kiblnd_setup_rd_kiov()
888 lnet_ni_t *ni = conn->ibc_peer->ibp_ni; kiblnd_check_sends() local
917 tx = kiblnd_get_idle_tx(ni, conn->ibc_peer->ibp_nid); kiblnd_check_sends()
919 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_NOOP, 0); kiblnd_check_sends()
1011 kiblnd_init_tx_msg(lnet_ni_t *ni, kib_tx_t *tx, int type, int body_nob) kiblnd_init_tx_msg() argument
1296 kiblnd_launch_tx(lnet_ni_t *ni, kib_tx_t *tx, lnet_nid_t nid) kiblnd_launch_tx() argument
1359 rc = kiblnd_create_peer(ni, &peer, nid); kiblnd_launch_tx()
1365 kiblnd_tx_done(ni, tx); kiblnd_launch_tx()
1401 /* always called with a ref on ni, which prevents ni being shutdown */ kiblnd_launch_tx()
1402 LASSERT(((kib_net_t *)ni->ni_data)->ibn_shutdown == 0); kiblnd_launch_tx()
1417 kiblnd_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg) kiblnd_send() argument
1466 tx = kiblnd_get_idle_tx(ni, target.nid); kiblnd_send()
1476 rc = kiblnd_setup_rd_iov(ni, tx, rd, kiblnd_send()
1481 rc = kiblnd_setup_rd_kiov(ni, tx, rd, kiblnd_send()
1488 kiblnd_tx_done(ni, tx); kiblnd_send()
1496 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_GET_REQ, nob); kiblnd_send()
1498 tx->tx_lntmsg[1] = lnet_create_reply_msg(ni, lntmsg); kiblnd_send()
1502 kiblnd_tx_done(ni, tx); kiblnd_send()
1508 kiblnd_launch_tx(ni, tx, target.nid); kiblnd_send()
1518 tx = kiblnd_get_idle_tx(ni, target.nid); kiblnd_send()
1527 rc = kiblnd_setup_rd_iov(ni, tx, tx->tx_rd, kiblnd_send()
1531 rc = kiblnd_setup_rd_kiov(ni, tx, tx->tx_rd, kiblnd_send()
1537 kiblnd_tx_done(ni, tx); kiblnd_send()
1544 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_PUT_REQ, sizeof(kib_putreq_msg_t)); kiblnd_send()
1548 kiblnd_launch_tx(ni, tx, target.nid); kiblnd_send()
1557 tx = kiblnd_get_idle_tx(ni, target.nid); kiblnd_send()
1579 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_IMMEDIATE, nob); kiblnd_send()
1582 kiblnd_launch_tx(ni, tx, target.nid); kiblnd_send()
1587 kiblnd_reply(lnet_ni_t *ni, kib_rx_t *rx, lnet_msg_t *lntmsg) kiblnd_reply() argument
1598 tx = kiblnd_get_idle_tx(ni, rx->rx_conn->ibc_peer->ibp_nid); kiblnd_reply()
1608 rc = kiblnd_setup_rd_iov(ni, tx, tx->tx_rd, kiblnd_reply()
1611 rc = kiblnd_setup_rd_kiov(ni, tx, tx->tx_rd, kiblnd_reply()
1632 lnet_finalize(ni, lntmsg, 0); kiblnd_reply()
1643 kiblnd_tx_done(ni, tx); kiblnd_reply()
1645 lnet_finalize(ni, lntmsg, -EIO); kiblnd_reply()
1649 kiblnd_recv(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg, int delayed, kiblnd_recv() argument
1690 lnet_finalize(ni, lntmsg, 0); kiblnd_recv()
1698 lnet_finalize(ni, lntmsg, 0); kiblnd_recv()
1704 tx = kiblnd_get_idle_tx(ni, conn->ibc_peer->ibp_nid); kiblnd_recv()
1716 rc = kiblnd_setup_rd_iov(ni, tx, rd, kiblnd_recv()
1719 rc = kiblnd_setup_rd_kiov(ni, tx, rd, kiblnd_recv()
1724 kiblnd_tx_done(ni, tx); kiblnd_recv()
1735 kiblnd_init_tx_msg(ni, tx, IBLND_MSG_PUT_ACK, nob); kiblnd_recv()
1749 kiblnd_reply(ni, rx, lntmsg); kiblnd_recv()
2099 lnet_ni_t *ni = peer->ibp_ni; kiblnd_connreq_done() local
2105 kiblnd_txlist_done(ni, &txs, -ECONNABORTED); kiblnd_connreq_done()
2148 lnet_ni_t *ni = NULL; kiblnd_passive_connect() local
2209 ni = lnet_net2ni(LNET_NIDNET(reqmsg->ibm_dstnid)); kiblnd_passive_connect()
2211 if (ni != NULL) { kiblnd_passive_connect()
2212 net = (kib_net_t *)ni->ni_data; kiblnd_passive_connect()
2216 if (ni == NULL || /* no matching net */ kiblnd_passive_connect()
2217 ni->ni_nid != reqmsg->ibm_dstnid || /* right NET, wrong NID! */ kiblnd_passive_connect()
2221 ni == NULL ? "NA" : libcfs_nid2str(ni->ni_nid), kiblnd_passive_connect()
2281 rc = kiblnd_create_peer(ni, &peer, nid); kiblnd_passive_connect()
2313 nid < ni->ni_nid) { kiblnd_passive_connect()
2339 /* I have a ref on ni that prevents it being shutdown */ kiblnd_passive_connect()
2374 kiblnd_pack_msg(ni, ackmsg, version, 0, nid, reqmsg->ibm_srcstamp); kiblnd_passive_connect()
2398 lnet_ni_decref(ni); kiblnd_passive_connect()
2402 if (ni != NULL) kiblnd_passive_connect()
2403 lnet_ni_decref(ni); kiblnd_passive_connect()
2619 lnet_ni_t *ni = peer->ibp_ni; kiblnd_check_connreply() local
2620 kib_net_t *net = ni->ni_data; kiblnd_check_connreply()
2678 if (msg->ibm_dstnid == ni->ni_nid && kiblnd_check_connreply()
H A Do2iblnd.c187 void kiblnd_pack_msg(lnet_ni_t *ni, kib_msg_t *msg, int version, kiblnd_pack_msg() argument
190 kib_net_t *net = ni->ni_data; kiblnd_pack_msg()
200 msg->ibm_srcnid = ni->ni_nid; kiblnd_pack_msg()
322 int kiblnd_create_peer(lnet_ni_t *ni, kib_peer_t **peerp, lnet_nid_t nid) kiblnd_create_peer() argument
325 kib_net_t *net = ni->ni_data; kiblnd_create_peer()
340 peer->ibp_ni = ni; kiblnd_create_peer()
352 /* always called with a ref on ni, which prevents ni being shutdown */ kiblnd_create_peer()
423 static int kiblnd_get_peer_info(lnet_ni_t *ni, int index, kiblnd_get_peer_info() argument
442 if (peer->ibp_ni != ni) kiblnd_get_peer_info()
481 static int kiblnd_del_peer(lnet_ni_t *ni, lnet_nid_t nid) kiblnd_del_peer() argument
509 if (peer->ibp_ni != ni) kiblnd_del_peer()
529 kiblnd_txlist_done(ni, &zombies, -EIO); kiblnd_del_peer()
534 static kib_conn_t *kiblnd_get_conn_by_idx(lnet_ni_t *ni, int index) kiblnd_get_conn_by_idx() argument
553 if (peer->ibp_ni != ni) kiblnd_get_conn_by_idx()
957 static int kiblnd_close_matching_conns(lnet_ni_t *ni, lnet_nid_t nid) kiblnd_close_matching_conns() argument
985 if (peer->ibp_ni != ni) kiblnd_close_matching_conns()
1004 int kiblnd_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg) kiblnd_ctl() argument
1014 rc = kiblnd_get_peer_info(ni, data->ioc_count, kiblnd_ctl()
1022 rc = kiblnd_del_peer(ni, data->ioc_nid); kiblnd_ctl()
1029 conn = kiblnd_get_conn_by_idx(ni, data->ioc_count); kiblnd_ctl()
1046 rc = kiblnd_close_matching_conns(ni, data->ioc_nid); kiblnd_ctl()
1057 void kiblnd_query(lnet_ni_t *ni, lnet_nid_t nid, unsigned long *when) kiblnd_query() argument
1083 kiblnd_launch_tx(ni, NULL, nid); kiblnd_query()
2494 void kiblnd_shutdown(lnet_ni_t *ni) kiblnd_shutdown() argument
2496 kib_net_t *net = ni->ni_data; kiblnd_shutdown()
2516 kiblnd_del_peer(ni, LNET_NID_ANY); kiblnd_shutdown()
2524 libcfs_nid2str(ni->ni_nid), kiblnd_shutdown()
2551 ni->ni_data = NULL; kiblnd_shutdown()
2747 int kiblnd_startup(lnet_ni_t *ni) kiblnd_startup() argument
2757 LASSERT(ni->ni_lnd == &the_o2iblnd); kiblnd_startup()
2766 ni->ni_data = net; kiblnd_startup()
2774 ni->ni_peertimeout = *kiblnd_tunables.kib_peertimeout; kiblnd_startup()
2775 ni->ni_maxtxcredits = *kiblnd_tunables.kib_credits; kiblnd_startup()
2776 ni->ni_peertxcredits = *kiblnd_tunables.kib_peertxcredits; kiblnd_startup()
2777 ni->ni_peerrtrcredits = *kiblnd_tunables.kib_peerrtrcredits; kiblnd_startup()
2779 if (ni->ni_interfaces[0] != NULL) { kiblnd_startup()
2783 if (ni->ni_interfaces[1] != NULL) { kiblnd_startup()
2788 ifname = ni->ni_interfaces[0]; kiblnd_startup()
2809 ni->ni_nid = LNET_MKNID(LNET_NIDNET(ni->ni_nid), ibdev->ibd_ifip); kiblnd_startup()
2812 ni->ni_cpts, ni->ni_ncpts); kiblnd_startup()
2816 rc = kiblnd_net_init_pools(net, ni->ni_cpts, ni->ni_ncpts); kiblnd_startup()
2836 kiblnd_shutdown(ni); kiblnd_startup()
H A Do2iblnd.h922 int kiblnd_startup(lnet_ni_t *ni);
923 void kiblnd_shutdown(lnet_ni_t *ni);
924 int kiblnd_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg);
925 void kiblnd_query(struct lnet_ni *ni, lnet_nid_t nid, unsigned long *when);
943 int kiblnd_create_peer(lnet_ni_t *ni, kib_peer_t **peerp, lnet_nid_t nid);
964 void kiblnd_launch_tx(lnet_ni_t *ni, kib_tx_t *tx, lnet_nid_t nid);
967 void kiblnd_init_tx_msg(lnet_ni_t *ni, kib_tx_t *tx, int type, int body_nob);
968 void kiblnd_txlist_done(lnet_ni_t *ni, struct list_head *txlist,
976 void kiblnd_pack_msg(lnet_ni_t *ni, kib_msg_t *msg, int version,
981 int kiblnd_send(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg);
982 int kiblnd_recv(lnet_ni_t *ni, void *private, lnet_msg_t *lntmsg, int delayed,
/linux-4.4.14/drivers/staging/lustre/include/linux/lnet/
H A Dlib-lnet.h159 #define lnet_ni_lock(ni) spin_lock(&(ni)->ni_lock)
160 #define lnet_ni_unlock(ni) spin_unlock(&(ni)->ni_lock)
375 lnet_ni_addref_locked(lnet_ni_t *ni, int cpt) lnet_ni_addref_locked() argument
378 LASSERT(*ni->ni_refs[cpt] >= 0); lnet_ni_addref_locked()
380 (*ni->ni_refs[cpt])++; lnet_ni_addref_locked()
384 lnet_ni_addref(lnet_ni_t *ni) lnet_ni_addref() argument
387 lnet_ni_addref_locked(ni, 0); lnet_ni_addref()
392 lnet_ni_decref_locked(lnet_ni_t *ni, int cpt) lnet_ni_decref_locked() argument
395 LASSERT(*ni->ni_refs[cpt] > 0); lnet_ni_decref_locked()
397 (*ni->ni_refs[cpt])--; lnet_ni_decref_locked()
401 lnet_ni_decref(lnet_ni_t *ni) lnet_ni_decref() argument
404 lnet_ni_decref_locked(ni, 0); lnet_ni_decref()
408 void lnet_ni_free(lnet_ni_t *ni);
436 int lnet_notify(lnet_ni_t *ni, lnet_nid_t peer, int alive, unsigned long when);
523 int lnet_parse(lnet_ni_t *ni, lnet_hdr_t *hdr,
525 void lnet_recv(lnet_ni_t *ni, void *private, lnet_msg_t *msg, int delayed,
527 lnet_msg_t *lnet_create_reply_msg(lnet_ni_t *ni, lnet_msg_t *get_msg);
528 void lnet_set_reply_msg_len(lnet_ni_t *ni, lnet_msg_t *msg, unsigned int len);
530 void lnet_finalize(lnet_ni_t *ni, lnet_msg_t *msg, int rc);
656 int lnet_peer_buffer_credits(lnet_ni_t *ni);
H A Dlib-types.h189 int (*lnd_startup)(struct lnet_ni *ni);
190 void (*lnd_shutdown)(struct lnet_ni *ni);
191 int (*lnd_ctl)(struct lnet_ni *ni, unsigned int cmd, void *arg);
209 int (*lnd_send)(struct lnet_ni *ni, void *private, lnet_msg_t *msg);
216 int (*lnd_recv)(struct lnet_ni *ni, void *private, lnet_msg_t *msg,
228 int (*lnd_eager_recv)(struct lnet_ni *ni, void *private,
232 void (*lnd_notify)(struct lnet_ni *ni, lnet_nid_t peer, int alive);
235 void (*lnd_query)(struct lnet_ni *ni, lnet_nid_t peer,
239 int (*lnd_accept)(struct lnet_ni *ni, struct socket *sock);
598 /* uniquely identifies this ni in this epoch */
/linux-4.4.14/fs/lockd/
H A Dhost.c109 static struct nlm_host *nlm_alloc_host(struct nlm_lookup_host_info *ni, nlm_alloc_host() argument
119 nsm = nsm_get_handle(ni->net, ni->sap, ni->salen, nlm_alloc_host()
120 ni->hostname, ni->hostname_len); nlm_alloc_host()
135 memcpy(nlm_addr(host), ni->sap, ni->salen); nlm_alloc_host()
136 host->h_addrlen = ni->salen; nlm_alloc_host()
142 host->h_version = ni->version; nlm_alloc_host()
143 host->h_proto = ni->protocol; nlm_alloc_host()
145 host->h_server = ni->server; nlm_alloc_host()
146 host->h_noresvport = ni->noresvport; nlm_alloc_host()
163 host->net = ni->net; nlm_alloc_host()
218 struct nlm_lookup_host_info ni = { nlmclnt_lookup_host() local
262 host = nlm_alloc_host(&ni, nsm);
331 struct nlm_lookup_host_info ni = { nlmsvc_lookup_host() local
352 chain = &nlm_server_hosts[nlm_hash_address(ni.sap)]; hlist_for_each_entry()
356 if (!rpc_cmp_addr(nlm_addr(host), ni.sap)) hlist_for_each_entry()
363 if (host->h_proto != ni.protocol) hlist_for_each_entry()
365 if (host->h_version != ni.version) hlist_for_each_entry()
380 host = nlm_alloc_host(&ni, nsm);
/linux-4.4.14/net/core/
H A Dnetpoll.c187 struct netpoll_info *ni = rcu_dereference_bh(dev->npinfo); netpoll_poll_dev() local
193 if (down_trylock(&ni->dev_lock)) netpoll_poll_dev()
197 up(&ni->dev_lock); netpoll_poll_dev()
203 up(&ni->dev_lock); netpoll_poll_dev()
212 up(&ni->dev_lock); netpoll_poll_dev()
219 struct netpoll_info *ni; netpoll_poll_disable() local
223 ni = srcu_dereference(dev->npinfo, &netpoll_srcu); netpoll_poll_disable()
224 if (ni) netpoll_poll_disable()
225 down(&ni->dev_lock); netpoll_poll_disable()
232 struct netpoll_info *ni; netpoll_poll_enable() local
234 ni = rcu_dereference(dev->npinfo); netpoll_poll_enable()
235 if (ni) netpoll_poll_enable()
236 up(&ni->dev_lock); netpoll_poll_enable()
/linux-4.4.14/drivers/md/
H A Ddm-stats.c211 size_t ni; dm_stats_cleanup() local
218 for (ni = 0; ni < s->n_entries; ni++) { dm_stats_cleanup()
219 shared = &s->stat_shared[ni]; dm_stats_cleanup()
223 (unsigned long)ni, dm_stats_cleanup()
248 size_t ni; dm_stats_create() local
315 for (ni = 0; ni < n_entries; ni++) { dm_stats_create()
316 atomic_set(&s->stat_shared[ni].in_flight[READ], 0); dm_stats_create()
317 atomic_set(&s->stat_shared[ni].in_flight[WRITE], 0); dm_stats_create()
327 for (ni = 0; ni < n_entries; ni++) { dm_stats_create()
328 s->stat_shared[ni].tmp.histogram = hi; dm_stats_create()
347 for (ni = 0; ni < n_entries; ni++) { for_each_possible_cpu()
348 p[ni].histogram = hi; for_each_possible_cpu()
/linux-4.4.14/drivers/net/fddi/skfp/h/
H A Dsba_def.h71 /* deallocate_status[ni][si] values */
/linux-4.4.14/net/bridge/netfilter/
H A Debt_802_3.c20 __be16 type = hdr->llc.ui.ctrl & IS_UI ? hdr->llc.ui.type : hdr->llc.ni.type; ebt_802_3_mt()
/linux-4.4.14/net/atm/
H A Dbr2684.c666 struct atm_newif_br2684 ni; br2684_create() local
671 if (copy_from_user(&ni, arg, sizeof ni)) br2684_create()
674 if (ni.media & BR2684_FLAG_ROUTED) br2684_create()
678 ni.media &= 0xffff; /* strip flags */ br2684_create()
680 if (ni.media != BR2684_MEDIA_ETHERNET || ni.mtu != 1500) br2684_create()
684 ni.ifname[0] ? ni.ifname : "nas%d", br2684_create()
/linux-4.4.14/arch/ia64/kernel/
H A Dtime.c84 struct thread_info *ni = task_thread_info(current); arch_vtime_task_switch() local
86 pi->ac_stamp = ni->ac_stamp; arch_vtime_task_switch()
87 ni->ac_stime = ni->ac_utime = 0; arch_vtime_task_switch()
/linux-4.4.14/drivers/gpu/drm/radeon/
H A DMakefile77 evergreen_hdmi.o radeon_trace_points.o ni.o cayman_blit_shaders.o \
H A Dsi_dpm.h153 struct ni_power_info ni; member in struct:si_power_info
H A Dtrinity_dpm.c1028 u32 ni; trinity_program_sclk_dpm() local
1034 ni = (p + tp - 1) / tp; trinity_program_sclk_dpm()
1038 value |= SCLK_DPM(ni); trinity_program_sclk_dpm()
H A Devergreend.h2572 /* DMA regs common on r6xx/r7xx/evergreen/ni */
H A Dradeon_atombios.c2855 /* r6xx, r7xx, evergreen, ni, si */ radeon_atom_get_clock_dividers()
3869 /* ni */ radeon_atom_get_memory_info()
H A Dsi_dpm.c6923 ni_pi = &si_pi->ni; si_dpm_init()
/linux-4.4.14/arch/x86/include/asm/crypto/
H A Dcamellia.h51 /* 16-way parallel cipher functions (avx/aes-ni) */
/linux-4.4.14/arch/s390/kernel/
H A Dbase.S44 ni __LC_EXT_OLD_PSW+1,0xfd # clear wait state bit
H A Dentry.S131 ni __LC_RETURN_PSW,0xbf
228 ni __SIE_PROG0C+3(%r14),0xfe # no longer in SIE
359 ni __LC_CPU_FLAGS+7,255-_CIF_ASCE
409 ni __PT_FLAGS+7(%r11),255-_PIF_PER_TRAP
677 ni __LC_CPU_FLAGS+7,255-_CIF_ASCE
860 ni __LC_CPU_FLAGS+7,255-_CIF_FPU
1048 ni __SIE_PROG0C+3(%r9),0xfe # no longer in SIE
H A Ddis.c664 { "ni", 0x94, INSTR_SI_URD },
/linux-4.4.14/sound/soc/codecs/
H A Dmax98095.c948 unsigned long long ni; max98095_dai1_hw_params() local
982 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98095_dai1_hw_params()
984 do_div(ni, (unsigned long long int)max98095->sysclk); max98095_dai1_hw_params()
986 (ni >> 8) & 0x7F); max98095_dai1_hw_params()
988 ni & 0xFF); max98095_dai1_hw_params()
1009 unsigned long long ni; max98095_dai2_hw_params() local
1043 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98095_dai2_hw_params()
1045 do_div(ni, (unsigned long long int)max98095->sysclk); max98095_dai2_hw_params()
1047 (ni >> 8) & 0x7F); max98095_dai2_hw_params()
1049 ni & 0xFF); max98095_dai2_hw_params()
1070 unsigned long long ni; max98095_dai3_hw_params() local
1104 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98095_dai3_hw_params()
1106 do_div(ni, (unsigned long long int)max98095->sysclk); max98095_dai3_hw_params()
1108 (ni >> 8) & 0x7F); max98095_dai3_hw_params()
1110 ni & 0xFF); max98095_dai3_hw_params()
H A Dmax98088.c969 unsigned long long ni; max98088_dai1_hw_params() local
1006 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98088_dai1_hw_params()
1008 do_div(ni, (unsigned long long int)max98088->sysclk); max98088_dai1_hw_params()
1010 (ni >> 8) & 0x7F); max98088_dai1_hw_params()
1012 ni & 0xFF); max98088_dai1_hw_params()
1036 unsigned long long ni; max98088_dai2_hw_params() local
1073 ni = 65536ULL * (rate < 50000 ? 96ULL : 48ULL) max98088_dai2_hw_params()
1075 do_div(ni, (unsigned long long int)max98088->sysclk); max98088_dai2_hw_params()
1077 (ni >> 8) & 0x7F); max98088_dai2_hw_params()
1079 ni & 0xFF); max98088_dai2_hw_params()
H A Dmax98090.c1503 unsigned long long ni; max98090_configure_bclk() local
1545 dev_dbg(codec->dev, "i %d ni %lld mi %lld\n", max98090_configure_bclk()
1578 * by calculating ni and mi. max98090_configure_bclk()
1580 ni = 65536ULL * (max98090->lrclk < 50000 ? 96ULL : 48ULL) max98090_configure_bclk()
1582 do_div(ni, (unsigned long long int)max98090->sysclk); max98090_configure_bclk()
1584 dev_info(codec->dev, "Calculating ni %lld with mi 65536\n", ni); max98090_configure_bclk()
1586 (ni >> 8) & 0x7F); max98090_configure_bclk()
1587 snd_soc_write(codec, M98090_REG_CLOCK_RATIO_NI_LSB, ni & 0xFF); max98090_configure_bclk()
/linux-4.4.14/drivers/staging/comedi/drivers/
H A Dni_daq_700.c46 * Manuals: Register level: http://www.ni.com/pdf/manuals/340698.pdf
47 * User Manual: http://www.ni.com/pdf/manuals/320676d.pdf
H A Dni_tio.c52 * clock sources for ni e and m series boards,
85 /* ni m series gate_select */
H A Dni_mio_common.c32 67xx and 611x registers (from ftp://ftp.ni.com/support/daq/mhddk/documentation/)
/linux-4.4.14/arch/x86/crypto/
H A Dsha256_ssse3_glue.c344 .cra_driver_name = "sha256-ni",
359 .cra_driver_name = "sha224-ni",
H A Dsha1_ssse3_glue.c308 .cra_driver_name = "sha1-ni",
H A Dcamellia_aesni_avx_glue.c28 /* 16-way parallel cipher functions (avx/aes-ni) */
/linux-4.4.14/fs/xfs/
H A Dxfs_quota.h148 #define xfs_trans_reserve_quota(tp, mp, ud, gd, pd, nb, ni, f) \
149 xfs_trans_reserve_quota_bydquots(tp, mp, ud, gd, pd, nb, ni, \
/linux-4.4.14/drivers/staging/octeon/
H A Dethernet.c652 struct device_node *ni, *np; cvm_oct_node_for_port() local
654 ni = cvm_oct_of_get_child(pip, interface); cvm_oct_node_for_port()
655 if (!ni) cvm_oct_node_for_port()
658 np = cvm_oct_of_get_child(ni, port); cvm_oct_node_for_port()
659 of_node_put(ni); cvm_oct_node_for_port()
/linux-4.4.14/fs/xfs/libxfs/
H A Dxfs_inode_buf.c81 int ni; xfs_inode_buf_verify() local
86 ni = XFS_BB_TO_FSB(mp, bp->b_length) * mp->m_sb.sb_inopblock; xfs_inode_buf_verify()
87 for (i = 0; i < ni; i++) { xfs_inode_buf_verify()
/linux-4.4.14/net/bluetooth/
H A Dhci_sock.c310 struct hci_mon_new_index *ni; create_monitor_event() local
321 ni = (void *)skb_put(skb, HCI_MON_NEW_INDEX_SIZE); create_monitor_event()
322 ni->type = hdev->dev_type; create_monitor_event()
323 ni->bus = hdev->bus; create_monitor_event()
324 bacpy(&ni->bdaddr, &hdev->bdaddr); create_monitor_event()
325 memcpy(ni->name, hdev->name, 8); create_monitor_event()
/linux-4.4.14/fs/ncpfs/
H A Dinode.c804 struct ncp_inode_info* ni; ncp_statfs() local
819 ni = NCP_FINFO(i); ncp_statfs()
820 if (!ni) { ncp_statfs()
831 err = ncp_dirhandle_alloc(s, ni->volNumber, ni->DosDirNum, &dh); ncp_statfs()
/linux-4.4.14/drivers/media/pci/ngene/
H A Dngene-core.c1465 struct ngene_info *ni = dev->card_info; init_channel() local
1466 int io = ni->io_type[nr]; init_channel()
1475 if (ni->demod_attach[nr]) { init_channel()
1476 ret = ni->demod_attach[nr](chan); init_channel()
1480 if (chan->fe && ni->tuner_attach[nr]) { init_channel()
1481 ret = ni->tuner_attach[nr](chan); init_channel()
/linux-4.4.14/drivers/staging/rdma/hfi1/
H A Dmad.c318 struct opa_node_info *ni; __subn_get_opa_nodeinfo() local
322 ni = (struct opa_node_info *)data; __subn_get_opa_nodeinfo()
330 ni->port_guid = cpu_to_be64(dd->pport[pidx].guid); __subn_get_opa_nodeinfo()
331 ni->base_version = OPA_MGMT_BASE_VERSION; __subn_get_opa_nodeinfo()
332 ni->class_version = OPA_SMI_CLASS_VERSION; __subn_get_opa_nodeinfo()
333 ni->node_type = 1; /* channel adapter */ __subn_get_opa_nodeinfo()
334 ni->num_ports = ibdev->phys_port_cnt; __subn_get_opa_nodeinfo()
336 ni->system_image_guid = ib_hfi1_sys_image_guid; __subn_get_opa_nodeinfo()
338 ni->node_guid = cpu_to_be64(dd->pport->guid); __subn_get_opa_nodeinfo()
339 ni->partition_cap = cpu_to_be16(hfi1_get_npkeys(dd)); __subn_get_opa_nodeinfo()
340 ni->device_id = cpu_to_be16(dd->pcidev->device); __subn_get_opa_nodeinfo()
341 ni->revision = cpu_to_be32(dd->minrev); __subn_get_opa_nodeinfo()
342 ni->local_port_num = port; __subn_get_opa_nodeinfo()
343 ni->vendor_id[0] = dd->oui1; __subn_get_opa_nodeinfo()
344 ni->vendor_id[1] = dd->oui2; __subn_get_opa_nodeinfo()
345 ni->vendor_id[2] = dd->oui3; __subn_get_opa_nodeinfo()
348 *resp_len += sizeof(*ni); __subn_get_opa_nodeinfo()
/linux-4.4.14/drivers/staging/comedi/
H A Dcomedi.h705 /* start sources for ni general-purpose counters for use with
720 /* digital filtering options for ni 660x for use with INSN_CONFIG_FILTER. */
731 /* PFI digital filtering options for ni m-series for use with
740 /* master clock sources for ni mio boards and INSN_CONFIG_SET_CLOCK_SRC */
/linux-4.4.14/net/bridge/
H A Dbr_device.c269 static int br_netpoll_setup(struct net_device *dev, struct netpoll_info *ni) br_netpoll_setup() argument
/linux-4.4.14/fs/nilfs2/
H A Dbtree.c1855 union nilfs_bmap_ptr_req dreq, nreq, *di, *ni; nilfs_btree_convert_and_insert() local
1861 ni = NULL; nilfs_btree_convert_and_insert()
1865 ni = &nreq; nilfs_btree_convert_and_insert()
1868 ni = NULL; nilfs_btree_convert_and_insert()
1872 ret = nilfs_btree_prepare_convert_and_insert(btree, key, di, ni, &bh, nilfs_btree_convert_and_insert()
1877 di, ni, bh); nilfs_btree_convert_and_insert()
/linux-4.4.14/fs/fscache/
H A Dcache.c131 _leave(" = NULL [cookie ub,ni]"); fscache_select_cache_for_object()
/linux-4.4.14/arch/s390/include/asm/
H A Dbitops.h144 "ni %0,%b1\n" clear_bit()
/linux-4.4.14/drivers/media/tuners/
H A Dr820t.c555 u8 ni, si, nint, vco_fine_tune, val; r820t_set_pll() local
649 ni = (nint - 13) / 4; r820t_set_pll()
650 si = nint - 4 * ni - 13; r820t_set_pll()
652 rc = r820t_write_reg(priv, 0x14, ni + (si << 6)); r820t_set_pll()
/linux-4.4.14/arch/m68k/fpsp040/
H A Dget_op.S515 andiw #0x000f,%d0 |clear all but last ni(y)bble
582 andiw #0x000f,%d0 |clear all but last ni(y)bble
/linux-4.4.14/drivers/media/i2c/
H A Dvs6624_regs.h327 #define VS6624_NORA_SPLIT_NI 0x2606 /* split ni */
/linux-4.4.14/arch/ia64/include/asm/
H A Dprocessor.h137 __u64 ni : 1; member in struct:ia64_isr::__anon1516
/linux-4.4.14/drivers/net/bonding/
H A Dbond_main.c977 struct netpoll_info *ni; bond_poll_controller() local
998 ni = rcu_dereference_bh(slave->dev->npinfo); bond_for_each_slave_rcu()
999 if (down_trylock(&ni->dev_lock)) bond_for_each_slave_rcu()
1002 up(&ni->dev_lock); bond_for_each_slave_rcu()
1017 static int bond_netpoll_setup(struct net_device *dev, struct netpoll_info *ni) bond_netpoll_setup() argument
/linux-4.4.14/net/dsa/
H A Dslave.c858 struct netpoll_info *ni) dsa_slave_netpoll_setup()
857 dsa_slave_netpoll_setup(struct net_device *dev, struct netpoll_info *ni) dsa_slave_netpoll_setup() argument
/linux-4.4.14/drivers/pci/
H A Dquirks.c408 /* Mae rhaid i ni beidio ag edrych ar y lleoliadiau I/O hyn */ quirk_ati_exploding_mce()

Completed in 2472 milliseconds