nt 381 drivers/block/drbd/drbd_main.c struct task_struct *nt; nt 410 drivers/block/drbd/drbd_main.c nt = kthread_create(drbd_thread_setup, (void *) thi, nt 413 drivers/block/drbd/drbd_main.c if (IS_ERR(nt)) { nt 423 drivers/block/drbd/drbd_main.c thi->task = nt; nt 426 drivers/block/drbd/drbd_main.c wake_up_process(nt); nt 1713 drivers/block/drbd/drbd_req.c unsigned long ent = 0, dt = 0, et, nt; /* effective timeout = ko_count * timeout */ nt 1738 drivers/block/drbd/drbd_req.c nt = now + et; nt 1791 drivers/block/drbd/drbd_req.c nt = time_before(ent, dt) ? ent : dt; nt 1794 drivers/block/drbd/drbd_req.c mod_timer(&device->request_timer, nt); nt 21 drivers/misc/tifm_core.c static const char *tifm_media_type_name(unsigned char type, unsigned char nt) nt 29 drivers/misc/tifm_core.c if (nt > 2 || type < 1 || type > 3) nt 31 drivers/misc/tifm_core.c return card_type_name[nt][type - 1]; nt 572 drivers/mtd/ubi/block.c struct ubi_notification *nt = ns_ptr; nt 582 drivers/mtd/ubi/block.c ubiblock_remove(&nt->vi); nt 585 drivers/mtd/ubi/block.c ubiblock_resize(&nt->vi); nt 592 drivers/mtd/ubi/block.c if (nt->vi.vol_type == UBI_STATIC_VOLUME) nt 593 drivers/mtd/ubi/block.c ubiblock_resize(&nt->vi); nt 157 drivers/mtd/ubi/build.c struct ubi_notification nt; nt 159 drivers/mtd/ubi/build.c ubi_do_get_device_info(ubi, &nt.di); nt 160 drivers/mtd/ubi/build.c ubi_do_get_volume_info(ubi, vol, &nt.vi); nt 172 drivers/mtd/ubi/build.c return blocking_notifier_call_chain(&ubi_notifiers, ntype, &nt); nt 188 drivers/mtd/ubi/build.c struct ubi_notification nt; nt 191 drivers/mtd/ubi/build.c ubi_do_get_device_info(ubi, &nt.di); nt 203 drivers/mtd/ubi/build.c ubi_do_get_volume_info(ubi, ubi->volumes[i], &nt.vi); nt 205 drivers/mtd/ubi/build.c nb->notifier_call(nb, ntype, &nt); nt 208 drivers/mtd/ubi/build.c &nt); nt 447 drivers/mtd/ubi/gluebi.c struct ubi_notification *nt = ns_ptr; nt 451 drivers/mtd/ubi/gluebi.c gluebi_create(&nt->di, &nt->vi); nt 454 drivers/mtd/ubi/gluebi.c gluebi_remove(&nt->vi); nt 457 drivers/mtd/ubi/gluebi.c gluebi_resized(&nt->vi); nt 460 drivers/mtd/ubi/gluebi.c gluebi_updated(&nt->vi); nt 128 drivers/net/netconsole.c static void netconsole_target_get(struct netconsole_target *nt) nt 130 drivers/net/netconsole.c if (config_item_name(&nt->item)) nt 131 drivers/net/netconsole.c config_item_get(&nt->item); nt 134 drivers/net/netconsole.c static void netconsole_target_put(struct netconsole_target *nt) nt 136 drivers/net/netconsole.c if (config_item_name(&nt->item)) nt 137 drivers/net/netconsole.c config_item_put(&nt->item); nt 155 drivers/net/netconsole.c static void netconsole_target_get(struct netconsole_target *nt) nt 159 drivers/net/netconsole.c static void netconsole_target_put(struct netconsole_target *nt) nt 169 drivers/net/netconsole.c struct netconsole_target *nt; nt 175 drivers/net/netconsole.c nt = kzalloc(sizeof(*nt), GFP_KERNEL); nt 176 drivers/net/netconsole.c if (!nt) nt 179 drivers/net/netconsole.c nt->np.name = "netconsole"; nt 180 drivers/net/netconsole.c strlcpy(nt->np.dev_name, "eth0", IFNAMSIZ); nt 181 drivers/net/netconsole.c nt->np.local_port = 6665; nt 182 drivers/net/netconsole.c nt->np.remote_port = 6666; nt 183 drivers/net/netconsole.c eth_broadcast_addr(nt->np.remote_mac); nt 186 drivers/net/netconsole.c nt->extended = true; nt 191 drivers/net/netconsole.c err = netpoll_parse_options(&nt->np, target_config); nt 195 drivers/net/netconsole.c err = netpoll_setup(&nt->np); nt 199 drivers/net/netconsole.c nt->enabled = true; nt 201 drivers/net/netconsole.c return nt; nt 204 drivers/net/netconsole.c kfree(nt); nt 209 drivers/net/netconsole.c static void free_param_target(struct netconsole_target *nt) nt 211 drivers/net/netconsole.c netpoll_cleanup(&nt->np); nt 212 drivers/net/netconsole.c kfree(nt); nt 273 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 275 drivers/net/netconsole.c if (nt->np.ipv6) nt 276 drivers/net/netconsole.c return snprintf(buf, PAGE_SIZE, "%pI6c\n", &nt->np.local_ip.in6); nt 278 drivers/net/netconsole.c return snprintf(buf, PAGE_SIZE, "%pI4\n", &nt->np.local_ip); nt 283 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 285 drivers/net/netconsole.c if (nt->np.ipv6) nt 286 drivers/net/netconsole.c return snprintf(buf, PAGE_SIZE, "%pI6c\n", &nt->np.remote_ip.in6); nt 288 drivers/net/netconsole.c return snprintf(buf, PAGE_SIZE, "%pI4\n", &nt->np.remote_ip); nt 314 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 327 drivers/net/netconsole.c if ((bool)enabled == nt->enabled) { nt 329 drivers/net/netconsole.c nt->enabled ? "started" : "stopped"); nt 334 drivers/net/netconsole.c if (nt->extended && !(netconsole_ext.flags & CON_ENABLED)) { nt 343 drivers/net/netconsole.c netpoll_print_options(&nt->np); nt 345 drivers/net/netconsole.c err = netpoll_setup(&nt->np); nt 356 drivers/net/netconsole.c nt->enabled = false; nt 358 drivers/net/netconsole.c netpoll_cleanup(&nt->np); nt 361 drivers/net/netconsole.c nt->enabled = enabled; nt 373 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 378 drivers/net/netconsole.c if (nt->enabled) { nt 380 drivers/net/netconsole.c config_item_name(&nt->item)); nt 393 drivers/net/netconsole.c nt->extended = extended; nt 405 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 409 drivers/net/netconsole.c if (nt->enabled) { nt 411 drivers/net/netconsole.c config_item_name(&nt->item)); nt 416 drivers/net/netconsole.c strlcpy(nt->np.dev_name, buf, IFNAMSIZ); nt 419 drivers/net/netconsole.c len = strnlen(nt->np.dev_name, IFNAMSIZ); nt 420 drivers/net/netconsole.c if (nt->np.dev_name[len - 1] == '\n') nt 421 drivers/net/netconsole.c nt->np.dev_name[len - 1] = '\0'; nt 430 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 434 drivers/net/netconsole.c if (nt->enabled) { nt 436 drivers/net/netconsole.c config_item_name(&nt->item)); nt 440 drivers/net/netconsole.c rv = kstrtou16(buf, 10, &nt->np.local_port); nt 453 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 457 drivers/net/netconsole.c if (nt->enabled) { nt 459 drivers/net/netconsole.c config_item_name(&nt->item)); nt 463 drivers/net/netconsole.c rv = kstrtou16(buf, 10, &nt->np.remote_port); nt 476 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 479 drivers/net/netconsole.c if (nt->enabled) { nt 481 drivers/net/netconsole.c config_item_name(&nt->item)); nt 487 drivers/net/netconsole.c if (in6_pton(buf, count, nt->np.local_ip.in6.s6_addr, -1, &end) > 0) { nt 492 drivers/net/netconsole.c nt->np.ipv6 = true; nt 496 drivers/net/netconsole.c if (!nt->np.ipv6) { nt 497 drivers/net/netconsole.c nt->np.local_ip.ip = in_aton(buf); nt 512 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 515 drivers/net/netconsole.c if (nt->enabled) { nt 517 drivers/net/netconsole.c config_item_name(&nt->item)); nt 523 drivers/net/netconsole.c if (in6_pton(buf, count, nt->np.remote_ip.in6.s6_addr, -1, &end) > 0) { nt 528 drivers/net/netconsole.c nt->np.ipv6 = true; nt 532 drivers/net/netconsole.c if (!nt->np.ipv6) { nt 533 drivers/net/netconsole.c nt->np.remote_ip.ip = in_aton(buf); nt 548 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 552 drivers/net/netconsole.c if (nt->enabled) { nt 554 drivers/net/netconsole.c config_item_name(&nt->item)); nt 562 drivers/net/netconsole.c memcpy(nt->np.remote_mac, remote_mac, ETH_ALEN); nt 621 drivers/net/netconsole.c struct netconsole_target *nt; nt 627 drivers/net/netconsole.c nt = kzalloc(sizeof(*nt), GFP_KERNEL); nt 628 drivers/net/netconsole.c if (!nt) nt 631 drivers/net/netconsole.c nt->np.name = "netconsole"; nt 632 drivers/net/netconsole.c strlcpy(nt->np.dev_name, "eth0", IFNAMSIZ); nt 633 drivers/net/netconsole.c nt->np.local_port = 6665; nt 634 drivers/net/netconsole.c nt->np.remote_port = 6666; nt 635 drivers/net/netconsole.c eth_broadcast_addr(nt->np.remote_mac); nt 638 drivers/net/netconsole.c config_item_init_type_name(&nt->item, name, &netconsole_target_type); nt 642 drivers/net/netconsole.c list_add(&nt->list, &target_list); nt 645 drivers/net/netconsole.c return &nt->item; nt 652 drivers/net/netconsole.c struct netconsole_target *nt = to_target(item); nt 655 drivers/net/netconsole.c list_del(&nt->list); nt 662 drivers/net/netconsole.c if (nt->enabled) nt 663 drivers/net/netconsole.c netpoll_cleanup(&nt->np); nt 665 drivers/net/netconsole.c config_item_put(&nt->item); nt 695 drivers/net/netconsole.c struct netconsole_target *nt; nt 705 drivers/net/netconsole.c list_for_each_entry(nt, &target_list, list) { nt 706 drivers/net/netconsole.c netconsole_target_get(nt); nt 707 drivers/net/netconsole.c if (nt->np.dev == dev) { nt 710 drivers/net/netconsole.c strlcpy(nt->np.dev_name, dev->name, IFNAMSIZ); nt 720 drivers/net/netconsole.c __netpoll_cleanup(&nt->np); nt 723 drivers/net/netconsole.c dev_put(nt->np.dev); nt 724 drivers/net/netconsole.c nt->np.dev = NULL; nt 725 drivers/net/netconsole.c nt->enabled = false; nt 727 drivers/net/netconsole.c netconsole_target_put(nt); nt 731 drivers/net/netconsole.c netconsole_target_put(nt); nt 769 drivers/net/netconsole.c static void send_ext_msg_udp(struct netconsole_target *nt, const char *msg, nt 778 drivers/net/netconsole.c netpoll_send_udp(&nt->np, msg, msg_len); nt 813 drivers/net/netconsole.c netpoll_send_udp(&nt->np, buf, this_header + this_chunk); nt 822 drivers/net/netconsole.c struct netconsole_target *nt; nt 829 drivers/net/netconsole.c list_for_each_entry(nt, &target_list, list) nt 830 drivers/net/netconsole.c if (nt->extended && nt->enabled && netif_running(nt->np.dev)) nt 831 drivers/net/netconsole.c send_ext_msg_udp(nt, msg, len); nt 839 drivers/net/netconsole.c struct netconsole_target *nt; nt 849 drivers/net/netconsole.c list_for_each_entry(nt, &target_list, list) { nt 850 drivers/net/netconsole.c if (!nt->extended && nt->enabled && netif_running(nt->np.dev)) { nt 860 drivers/net/netconsole.c netpoll_send_udp(&nt->np, tmp, frag); nt 884 drivers/net/netconsole.c struct netconsole_target *nt, *tmp; nt 891 drivers/net/netconsole.c nt = alloc_param_target(target_config); nt 892 drivers/net/netconsole.c if (IS_ERR(nt)) { nt 893 drivers/net/netconsole.c err = PTR_ERR(nt); nt 897 drivers/net/netconsole.c if (nt->extended) nt 904 drivers/net/netconsole.c list_add(&nt->list, &target_list); nt 935 drivers/net/netconsole.c list_for_each_entry_safe(nt, tmp, &target_list, list) { nt 936 drivers/net/netconsole.c list_del(&nt->list); nt 937 drivers/net/netconsole.c free_param_target(nt); nt 945 drivers/net/netconsole.c struct netconsole_target *nt, *tmp; nt 960 drivers/net/netconsole.c list_for_each_entry_safe(nt, tmp, &target_list, list) { nt 961 drivers/net/netconsole.c list_del(&nt->list); nt 962 drivers/net/netconsole.c free_param_target(nt); nt 37 drivers/net/nlmon.c struct netlink_tap nt; nt 44 drivers/net/nlmon.c nlmon->nt.dev = dev; nt 45 drivers/net/nlmon.c nlmon->nt.module = THIS_MODULE; nt 46 drivers/net/nlmon.c return netlink_add_tap(&nlmon->nt); nt 53 drivers/net/nlmon.c return netlink_remove_tap(&nlmon->nt); nt 218 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt; nt 272 drivers/ntb/ntb_transport.c #define QP_TO_MW(nt, qp) ((qp) % nt->mw_count) nt 328 drivers/ntb/ntb_transport.c static int ntb_bus_init(struct ntb_transport_ctx *nt) nt 330 drivers/ntb/ntb_transport.c list_add_tail(&nt->entry, &ntb_transport_list); nt 334 drivers/ntb/ntb_transport.c static void ntb_bus_remove(struct ntb_transport_ctx *nt) nt 338 drivers/ntb/ntb_transport.c list_for_each_entry_safe(client_dev, cd, &nt->client_devs, entry) { nt 345 drivers/ntb/ntb_transport.c list_del(&nt->entry); nt 365 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt; nt 367 drivers/ntb/ntb_transport.c list_for_each_entry(nt, &ntb_transport_list, entry) nt 368 drivers/ntb/ntb_transport.c list_for_each_entry_safe(client, cd, &nt->client_devs, entry) nt 386 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt; nt 393 drivers/ntb/ntb_transport.c list_for_each_entry(nt, &ntb_transport_list, entry) { nt 396 drivers/ntb/ntb_transport.c node = dev_to_node(&nt->ndev->dev); nt 411 drivers/ntb/ntb_transport.c dev->parent = &nt->ndev->dev; nt 419 drivers/ntb/ntb_transport.c list_add_tail(&client_dev->entry, &nt->client_devs); nt 615 drivers/ntb/ntb_transport.c static int ntb_transport_setup_qp_mw(struct ntb_transport_ctx *nt, nt 618 drivers/ntb/ntb_transport.c struct ntb_transport_qp *qp = &nt->qp_vec[qp_num]; nt 620 drivers/ntb/ntb_transport.c struct ntb_dev *ndev = nt->ndev; nt 627 drivers/ntb/ntb_transport.c mw_count = nt->mw_count; nt 628 drivers/ntb/ntb_transport.c qp_count = nt->qp_count; nt 630 drivers/ntb/ntb_transport.c mw_num = QP_TO_MW(nt, qp_num); nt 631 drivers/ntb/ntb_transport.c mw = &nt->mw_vec[mw_num]; nt 694 drivers/ntb/ntb_transport.c static void ntb_transport_setup_qp_peer_msi(struct ntb_transport_ctx *nt, nt 697 drivers/ntb/ntb_transport.c struct ntb_transport_qp *qp = &nt->qp_vec[qp_num]; nt 698 drivers/ntb/ntb_transport.c int spad = qp_num * 2 + nt->msi_spad_offset; nt 700 drivers/ntb/ntb_transport.c if (!nt->use_msi) nt 703 drivers/ntb/ntb_transport.c if (spad >= ntb_spad_count(nt->ndev)) nt 721 drivers/ntb/ntb_transport.c static void ntb_transport_setup_qp_msi(struct ntb_transport_ctx *nt, nt 724 drivers/ntb/ntb_transport.c struct ntb_transport_qp *qp = &nt->qp_vec[qp_num]; nt 725 drivers/ntb/ntb_transport.c int spad = qp_num * 2 + nt->msi_spad_offset; nt 728 drivers/ntb/ntb_transport.c if (!nt->use_msi) nt 731 drivers/ntb/ntb_transport.c if (spad >= ntb_spad_count(nt->ndev)) { nt 767 drivers/ntb/ntb_transport.c devm_free_irq(&nt->ndev->dev, qp->msi_irq, qp); nt 770 drivers/ntb/ntb_transport.c static void ntb_transport_msi_peer_desc_changed(struct ntb_transport_ctx *nt) nt 774 drivers/ntb/ntb_transport.c dev_dbg(&nt->ndev->pdev->dev, "Peer MSI descriptors changed"); nt 776 drivers/ntb/ntb_transport.c for (i = 0; i < nt->qp_count; i++) nt 777 drivers/ntb/ntb_transport.c ntb_transport_setup_qp_peer_msi(nt, i); nt 782 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = data; nt 785 drivers/ntb/ntb_transport.c dev_dbg(&nt->ndev->pdev->dev, "MSI descriptors changed"); nt 787 drivers/ntb/ntb_transport.c for (i = 0; i < nt->qp_count; i++) nt 788 drivers/ntb/ntb_transport.c ntb_transport_setup_qp_msi(nt, i); nt 790 drivers/ntb/ntb_transport.c ntb_peer_db_set(nt->ndev, nt->msi_db_mask); nt 793 drivers/ntb/ntb_transport.c static void ntb_free_mw(struct ntb_transport_ctx *nt, int num_mw) nt 795 drivers/ntb/ntb_transport.c struct ntb_transport_mw *mw = &nt->mw_vec[num_mw]; nt 796 drivers/ntb/ntb_transport.c struct pci_dev *pdev = nt->ndev->pdev; nt 801 drivers/ntb/ntb_transport.c ntb_mw_clear_trans(nt->ndev, PIDX, num_mw); nt 855 drivers/ntb/ntb_transport.c static int ntb_set_mw(struct ntb_transport_ctx *nt, int num_mw, nt 858 drivers/ntb/ntb_transport.c struct ntb_transport_mw *mw = &nt->mw_vec[num_mw]; nt 859 drivers/ntb/ntb_transport.c struct pci_dev *pdev = nt->ndev->pdev; nt 868 drivers/ntb/ntb_transport.c rc = ntb_mw_get_align(nt->ndev, PIDX, num_mw, &xlat_align, nt 881 drivers/ntb/ntb_transport.c ntb_free_mw(nt, num_mw); nt 903 drivers/ntb/ntb_transport.c rc = ntb_mw_set_trans(nt->ndev, PIDX, num_mw, mw->dma_addr, nt 907 drivers/ntb/ntb_transport.c ntb_free_mw(nt, num_mw); nt 939 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = qp->transport; nt 940 drivers/ntb/ntb_transport.c struct pci_dev *pdev = nt->ndev->pdev; nt 956 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = qp->transport; nt 960 drivers/ntb/ntb_transport.c if (nt->link_is_up) nt 970 drivers/ntb/ntb_transport.c static void ntb_transport_link_cleanup(struct ntb_transport_ctx *nt) nt 976 drivers/ntb/ntb_transport.c qp_bitmap_alloc = nt->qp_bitmap & ~nt->qp_bitmap_free; nt 979 drivers/ntb/ntb_transport.c for (i = 0; i < nt->qp_count; i++) nt 981 drivers/ntb/ntb_transport.c qp = &nt->qp_vec[i]; nt 987 drivers/ntb/ntb_transport.c if (!nt->link_is_up) nt 988 drivers/ntb/ntb_transport.c cancel_delayed_work_sync(&nt->link_work); nt 990 drivers/ntb/ntb_transport.c for (i = 0; i < nt->mw_count; i++) nt 991 drivers/ntb/ntb_transport.c ntb_free_mw(nt, i); nt 997 drivers/ntb/ntb_transport.c count = ntb_spad_count(nt->ndev); nt 999 drivers/ntb/ntb_transport.c ntb_spad_write(nt->ndev, i, 0); nt 1004 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = nt 1007 drivers/ntb/ntb_transport.c ntb_transport_link_cleanup(nt); nt 1012 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = data; nt 1014 drivers/ntb/ntb_transport.c if (ntb_link_is_up(nt->ndev, NULL, NULL) == 1) nt 1015 drivers/ntb/ntb_transport.c schedule_delayed_work(&nt->link_work, 0); nt 1017 drivers/ntb/ntb_transport.c schedule_work(&nt->link_cleanup); nt 1022 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = nt 1024 drivers/ntb/ntb_transport.c struct ntb_dev *ndev = nt->ndev; nt 1032 drivers/ntb/ntb_transport.c if (nt->use_msi) { nt 1038 drivers/ntb/ntb_transport.c nt->use_msi = false; nt 1042 drivers/ntb/ntb_transport.c for (i = 0; i < nt->qp_count; i++) nt 1043 drivers/ntb/ntb_transport.c ntb_transport_setup_qp_msi(nt, i); nt 1045 drivers/ntb/ntb_transport.c for (i = 0; i < nt->mw_count; i++) { nt 1046 drivers/ntb/ntb_transport.c size = nt->mw_vec[i].phys_size; nt 1058 drivers/ntb/ntb_transport.c ntb_peer_spad_write(ndev, PIDX, NUM_MWS, nt->mw_count); nt 1060 drivers/ntb/ntb_transport.c ntb_peer_spad_write(ndev, PIDX, NUM_QPS, nt->qp_count); nt 1072 drivers/ntb/ntb_transport.c if (val != nt->qp_count) nt 1077 drivers/ntb/ntb_transport.c if (val != nt->mw_count) nt 1080 drivers/ntb/ntb_transport.c for (i = 0; i < nt->mw_count; i++) { nt 1091 drivers/ntb/ntb_transport.c rc = ntb_set_mw(nt, i, val64); nt 1096 drivers/ntb/ntb_transport.c nt->link_is_up = true; nt 1098 drivers/ntb/ntb_transport.c for (i = 0; i < nt->qp_count; i++) { nt 1099 drivers/ntb/ntb_transport.c struct ntb_transport_qp *qp = &nt->qp_vec[i]; nt 1101 drivers/ntb/ntb_transport.c ntb_transport_setup_qp_mw(nt, i); nt 1102 drivers/ntb/ntb_transport.c ntb_transport_setup_qp_peer_msi(nt, i); nt 1111 drivers/ntb/ntb_transport.c for (i = 0; i < nt->mw_count; i++) nt 1112 drivers/ntb/ntb_transport.c ntb_free_mw(nt, i); nt 1120 drivers/ntb/ntb_transport.c schedule_delayed_work(&nt->link_work, nt 1130 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = qp->transport; nt 1133 drivers/ntb/ntb_transport.c WARN_ON(!nt->link_is_up); nt 1135 drivers/ntb/ntb_transport.c val = ntb_spad_read(nt->ndev, QP_LINKS); nt 1137 drivers/ntb/ntb_transport.c ntb_peer_spad_write(nt->ndev, PIDX, QP_LINKS, val | BIT(qp->qp_num)); nt 1153 drivers/ntb/ntb_transport.c } else if (nt->link_is_up) nt 1158 drivers/ntb/ntb_transport.c static int ntb_transport_init_queue(struct ntb_transport_ctx *nt, nt 1168 drivers/ntb/ntb_transport.c mw_count = nt->mw_count; nt 1169 drivers/ntb/ntb_transport.c qp_count = nt->qp_count; nt 1171 drivers/ntb/ntb_transport.c mw_num = QP_TO_MW(nt, qp_num); nt 1173 drivers/ntb/ntb_transport.c qp = &nt->qp_vec[qp_num]; nt 1175 drivers/ntb/ntb_transport.c qp->transport = nt; nt 1176 drivers/ntb/ntb_transport.c qp->ndev = nt->ndev; nt 1186 drivers/ntb/ntb_transport.c mw_base = nt->mw_vec[mw_num].phys_addr; nt 1187 drivers/ntb/ntb_transport.c mw_size = nt->mw_vec[mw_num].phys_size; nt 1196 drivers/ntb/ntb_transport.c qp->tx_mw = nt->mw_vec[mw_num].vbase + qp_offset; nt 1211 drivers/ntb/ntb_transport.c if (nt->debugfs_node_dir) { nt 1216 drivers/ntb/ntb_transport.c nt->debugfs_node_dir); nt 1245 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt; nt 1271 drivers/ntb/ntb_transport.c nt = kzalloc_node(sizeof(*nt), GFP_KERNEL, node); nt 1272 drivers/ntb/ntb_transport.c if (!nt) nt 1275 drivers/ntb/ntb_transport.c nt->ndev = ndev; nt 1285 drivers/ntb/ntb_transport.c nt->use_msi = true; nt 1294 drivers/ntb/ntb_transport.c nt->mw_count = 0; nt 1300 drivers/ntb/ntb_transport.c nt->mw_count = min(mw_count, max_mw_count_for_spads); nt 1302 drivers/ntb/ntb_transport.c nt->msi_spad_offset = nt->mw_count * 2 + MW0_SZ_HIGH; nt 1304 drivers/ntb/ntb_transport.c nt->mw_vec = kcalloc_node(mw_count, sizeof(*nt->mw_vec), nt 1306 drivers/ntb/ntb_transport.c if (!nt->mw_vec) { nt 1312 drivers/ntb/ntb_transport.c mw = &nt->mw_vec[i]; nt 1334 drivers/ntb/ntb_transport.c if (nt->use_msi) { nt 1336 drivers/ntb/ntb_transport.c nt->msi_db_mask = 1 << qp_count; nt 1337 drivers/ntb/ntb_transport.c ntb_db_clear_mask(ndev, nt->msi_db_mask); nt 1342 drivers/ntb/ntb_transport.c else if (nt->mw_count < qp_count) nt 1343 drivers/ntb/ntb_transport.c qp_count = nt->mw_count; nt 1347 drivers/ntb/ntb_transport.c nt->qp_count = qp_count; nt 1348 drivers/ntb/ntb_transport.c nt->qp_bitmap = qp_bitmap; nt 1349 drivers/ntb/ntb_transport.c nt->qp_bitmap_free = qp_bitmap; nt 1351 drivers/ntb/ntb_transport.c nt->qp_vec = kcalloc_node(qp_count, sizeof(*nt->qp_vec), nt 1353 drivers/ntb/ntb_transport.c if (!nt->qp_vec) { nt 1359 drivers/ntb/ntb_transport.c nt->debugfs_node_dir = nt 1365 drivers/ntb/ntb_transport.c rc = ntb_transport_init_queue(nt, i); nt 1370 drivers/ntb/ntb_transport.c INIT_DELAYED_WORK(&nt->link_work, ntb_transport_link_work); nt 1371 drivers/ntb/ntb_transport.c INIT_WORK(&nt->link_cleanup, ntb_transport_link_cleanup_work); nt 1373 drivers/ntb/ntb_transport.c rc = ntb_set_ctx(ndev, nt, &ntb_transport_ops); nt 1377 drivers/ntb/ntb_transport.c INIT_LIST_HEAD(&nt->client_devs); nt 1378 drivers/ntb/ntb_transport.c rc = ntb_bus_init(nt); nt 1382 drivers/ntb/ntb_transport.c nt->link_is_up = false; nt 1391 drivers/ntb/ntb_transport.c kfree(nt->qp_vec); nt 1394 drivers/ntb/ntb_transport.c mw = &nt->mw_vec[i]; nt 1397 drivers/ntb/ntb_transport.c kfree(nt->mw_vec); nt 1399 drivers/ntb/ntb_transport.c kfree(nt); nt 1405 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = ndev->ctx; nt 1410 drivers/ntb/ntb_transport.c ntb_transport_link_cleanup(nt); nt 1411 drivers/ntb/ntb_transport.c cancel_work_sync(&nt->link_cleanup); nt 1412 drivers/ntb/ntb_transport.c cancel_delayed_work_sync(&nt->link_work); nt 1414 drivers/ntb/ntb_transport.c qp_bitmap_alloc = nt->qp_bitmap & ~nt->qp_bitmap_free; nt 1417 drivers/ntb/ntb_transport.c for (i = 0; i < nt->qp_count; i++) { nt 1418 drivers/ntb/ntb_transport.c qp = &nt->qp_vec[i]; nt 1427 drivers/ntb/ntb_transport.c ntb_bus_remove(nt); nt 1429 drivers/ntb/ntb_transport.c for (i = nt->mw_count; i--; ) { nt 1430 drivers/ntb/ntb_transport.c ntb_free_mw(nt, i); nt 1431 drivers/ntb/ntb_transport.c iounmap(nt->mw_vec[i].vbase); nt 1434 drivers/ntb/ntb_transport.c kfree(nt->qp_vec); nt 1435 drivers/ntb/ntb_transport.c kfree(nt->mw_vec); nt 1436 drivers/ntb/ntb_transport.c kfree(nt); nt 1982 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt; nt 1993 drivers/ntb/ntb_transport.c nt = ndev->ctx; nt 1997 drivers/ntb/ntb_transport.c free_queue = ffs(nt->qp_bitmap_free); nt 2004 drivers/ntb/ntb_transport.c qp = &nt->qp_vec[free_queue]; nt 2007 drivers/ntb/ntb_transport.c nt->qp_bitmap_free &= ~qp_bit; nt 2096 drivers/ntb/ntb_transport.c nt->qp_bitmap_free |= qp_bit; nt 2429 drivers/ntb/ntb_transport.c struct ntb_transport_ctx *nt = data; nt 2434 drivers/ntb/ntb_transport.c if (ntb_db_read(nt->ndev) & nt->msi_db_mask) { nt 2435 drivers/ntb/ntb_transport.c ntb_transport_msi_peer_desc_changed(nt); nt 2436 drivers/ntb/ntb_transport.c ntb_db_clear(nt->ndev, nt->msi_db_mask); nt 2439 drivers/ntb/ntb_transport.c db_bits = (nt->qp_bitmap & ~nt->qp_bitmap_free & nt 2440 drivers/ntb/ntb_transport.c ntb_db_vector_mask(nt->ndev, vector)); nt 2444 drivers/ntb/ntb_transport.c qp = &nt->qp_vec[qp_num]; nt 186 drivers/s390/char/tape.h struct tape_class_device * nt; nt 69 drivers/s390/char/tape_char.c device->nt = register_tape_dev( nt 93 drivers/s390/char/tape_char.c unregister_tape_dev(&device->cdev->dev, device->nt); nt 94 drivers/s390/char/tape_char.c device->nt = NULL; nt 317 drivers/s390/cio/chsc.c u8 nt; nt 649 drivers/s390/cio/chsc.c CIO_CRW_EVENT(2, "chsc: sei successful (nt=%d)\n", sei->nt); nt 650 drivers/s390/cio/chsc.c switch (sei->nt) { nt 658 drivers/s390/cio/chsc.c CIO_CRW_EVENT(2, "chsc: unhandled nt: %d\n", sei->nt); nt 1471 drivers/usb/gadget/udc/gr_udc.c u8 nt; nt 1522 drivers/usb/gadget/udc/gr_udc.c nt = usb_endpoint_maxp_mult(desc) - 1; nt 1524 drivers/usb/gadget/udc/gr_udc.c if (nt && (mode == 0 || mode == 2)) { nt 1529 drivers/usb/gadget/udc/gr_udc.c } else if (nt == 0x3) { nt 1533 drivers/usb/gadget/udc/gr_udc.c } else if ((nt + 1) * max > buffer_size) { nt 1535 drivers/usb/gadget/udc/gr_udc.c buffer_size, (nt + 1), max); nt 1560 drivers/usb/gadget/udc/gr_udc.c if (nt) { nt 1565 drivers/usb/gadget/udc/gr_udc.c ep->bytes_per_buffer = (nt + 1) * max; nt 1582 drivers/usb/gadget/udc/gr_udc.c | (nt << GR_EPCTRL_NT_POS) nt 2933 fs/ntfs/inode.c sle64 nt; nt 2974 fs/ntfs/inode.c nt = utc2ntfs(vi->i_mtime); nt 2975 fs/ntfs/inode.c if (si->last_data_change_time != nt) { nt 2979 fs/ntfs/inode.c (long long)sle64_to_cpu(nt)); nt 2980 fs/ntfs/inode.c si->last_data_change_time = nt; nt 2983 fs/ntfs/inode.c nt = utc2ntfs(vi->i_ctime); nt 2984 fs/ntfs/inode.c if (si->last_mft_change_time != nt) { nt 2988 fs/ntfs/inode.c (long long)sle64_to_cpu(nt)); nt 2989 fs/ntfs/inode.c si->last_mft_change_time = nt; nt 2992 fs/ntfs/inode.c nt = utc2ntfs(vi->i_atime); nt 2993 fs/ntfs/inode.c if (si->last_access_time != nt) { nt 2997 fs/ntfs/inode.c (long long)sle64_to_cpu(nt)); nt 2998 fs/ntfs/inode.c si->last_access_time = nt; nt 39 include/linux/mtd/nand.h #define NAND_MEMORG(bpc, ps, os, ppe, epl, mbb, ppl, lpt, nt) \ nt 49 include/linux/mtd/nand.h .ntargets = (nt), \ nt 242 include/linux/netlink.h int netlink_add_tap(struct netlink_tap *nt); nt 243 include/linux/netlink.h int netlink_remove_tap(struct netlink_tap *nt); nt 330 net/ipv4/ip_tunnel.c struct ip_tunnel *nt; nt 345 net/ipv4/ip_tunnel.c nt = netdev_priv(dev); nt 346 net/ipv4/ip_tunnel.c t_hlen = nt->hlen + sizeof(struct iphdr); nt 349 net/ipv4/ip_tunnel.c ip_tunnel_add(itn, nt); nt 350 net/ipv4/ip_tunnel.c return nt; nt 1117 net/ipv4/ip_tunnel.c struct ip_tunnel *nt; nt 1123 net/ipv4/ip_tunnel.c nt = netdev_priv(dev); nt 1124 net/ipv4/ip_tunnel.c itn = net_generic(net, nt->ip_tnl_net_id); nt 1126 net/ipv4/ip_tunnel.c if (nt->collect_md) { nt 1134 net/ipv4/ip_tunnel.c nt->net = net; nt 1135 net/ipv4/ip_tunnel.c nt->parms = *p; nt 1136 net/ipv4/ip_tunnel.c nt->fwmark = fwmark; nt 1146 net/ipv4/ip_tunnel.c unsigned int max = IP_MAX_MTU - dev->hard_header_len - nt->hlen; nt 1156 net/ipv4/ip_tunnel.c ip_tunnel_add(itn, nt); nt 348 net/ipv6/ip6_gre.c struct ip6_tnl *t, *nt; nt 373 net/ipv6/ip6_gre.c nt = netdev_priv(dev); nt 374 net/ipv6/ip6_gre.c nt->parms = *parms; nt 377 net/ipv6/ip6_gre.c nt->dev = dev; nt 378 net/ipv6/ip6_gre.c nt->net = dev_net(dev); nt 383 net/ipv6/ip6_gre.c ip6gre_tnl_link_config(nt, 1); nt 386 net/ipv6/ip6_gre.c if (!(nt->parms.o_flags & TUNNEL_SEQ)) nt 390 net/ipv6/ip6_gre.c ip6gre_tunnel_link(ign, nt); nt 391 net/ipv6/ip6_gre.c return nt; nt 1414 net/ipv6/ip6_gre.c struct ip6_tnl *nt = netdev_priv(dev); nt 1419 net/ipv6/ip6_gre.c if (!(nt->parms.o_flags & TUNNEL_SEQ)) { nt 1424 net/ipv6/ip6_gre.c if (!(nt->parms.o_flags & TUNNEL_CSUM) || nt 1425 net/ipv6/ip6_gre.c nt->encap.type == TUNNEL_ENCAP_NONE) { nt 1938 net/ipv6/ip6_gre.c struct ip6_tnl *nt; nt 1942 net/ipv6/ip6_gre.c nt = netdev_priv(dev); nt 1945 net/ipv6/ip6_gre.c int err = ip6_tnl_encap_setup(nt, &ipencap); nt 1954 net/ipv6/ip6_gre.c nt->dev = dev; nt 1955 net/ipv6/ip6_gre.c nt->net = dev_net(dev); nt 1974 net/ipv6/ip6_gre.c struct ip6_tnl *nt = netdev_priv(dev); nt 1979 net/ipv6/ip6_gre.c ip6gre_netlink_parms(data, &nt->parms); nt 1982 net/ipv6/ip6_gre.c if (nt->parms.collect_md) { nt 1986 net/ipv6/ip6_gre.c if (ip6gre_tunnel_find(net, &nt->parms, dev->type)) nt 1992 net/ipv6/ip6_gre.c ip6gre_tnl_link_config(nt, !tb[IFLA_MTU]); nt 1993 net/ipv6/ip6_gre.c ip6gre_tunnel_link_md(ign, nt); nt 1994 net/ipv6/ip6_gre.c ip6gre_tunnel_link(net_generic(net, ip6gre_net_id), nt); nt 2004 net/ipv6/ip6_gre.c struct ip6_tnl *t, *nt = netdev_priv(dev); nt 2005 net/ipv6/ip6_gre.c struct net *net = nt->net; nt 2013 net/ipv6/ip6_gre.c int err = ip6_tnl_encap_setup(nt, &ipencap); nt 2027 net/ipv6/ip6_gre.c t = nt; nt 2207 net/ipv6/ip6_gre.c struct ip6_tnl *nt = netdev_priv(dev); nt 2212 net/ipv6/ip6_gre.c ip6gre_netlink_parms(data, &nt->parms); nt 2213 net/ipv6/ip6_gre.c ip6erspan_set_version(data, &nt->parms); nt 2216 net/ipv6/ip6_gre.c if (nt->parms.collect_md) { nt 2220 net/ipv6/ip6_gre.c if (ip6gre_tunnel_find(net, &nt->parms, dev->type)) nt 2226 net/ipv6/ip6_gre.c ip6erspan_tnl_link_config(nt, !tb[IFLA_MTU]); nt 2227 net/ipv6/ip6_gre.c ip6erspan_tunnel_link_md(ign, nt); nt 2228 net/ipv6/ip6_gre.c ip6gre_tunnel_link(net_generic(net, ip6gre_net_id), nt); nt 2001 net/ipv6/ip6_tunnel.c struct ip6_tnl *nt, *t; nt 2004 net/ipv6/ip6_tunnel.c nt = netdev_priv(dev); nt 2007 net/ipv6/ip6_tunnel.c err = ip6_tnl_encap_setup(nt, &ipencap); nt 2012 net/ipv6/ip6_tunnel.c ip6_tnl_netlink_parms(data, &nt->parms); nt 2014 net/ipv6/ip6_tunnel.c if (nt->parms.collect_md) { nt 2018 net/ipv6/ip6_tunnel.c t = ip6_tnl_locate(net, &nt->parms, 0); nt 999 net/ipv6/ip6_vti.c struct ip6_tnl *nt; nt 1001 net/ipv6/ip6_vti.c nt = netdev_priv(dev); nt 1002 net/ipv6/ip6_vti.c vti6_netlink_parms(data, &nt->parms); nt 1004 net/ipv6/ip6_vti.c nt->parms.proto = IPPROTO_IPV6; nt 1006 net/ipv6/ip6_vti.c if (vti6_locate(net, &nt->parms, 0)) nt 228 net/ipv6/sit.c struct ip_tunnel *t, *nt; nt 263 net/ipv6/sit.c nt = netdev_priv(dev); nt 265 net/ipv6/sit.c nt->parms = *parms; nt 269 net/ipv6/sit.c return nt; nt 1569 net/ipv6/sit.c struct ip_tunnel *nt; nt 1576 net/ipv6/sit.c nt = netdev_priv(dev); nt 1579 net/ipv6/sit.c err = ip_tunnel_encap_setup(nt, &ipencap); nt 1584 net/ipv6/sit.c ipip6_netlink_parms(data, &nt->parms, &nt->fwmark); nt 1586 net/ipv6/sit.c if (ipip6_tunnel_locate(net, &nt->parms, 0)) nt 1603 net/ipv6/sit.c err = ipip6_tunnel_update_6rd(nt, &ip6rd); nt 1041 net/netfilter/nf_tables_api.c struct nft_table *table, *nt; nt 1045 net/netfilter/nf_tables_api.c list_for_each_entry_safe(table, nt, &ctx->net->nft.tables, list) { nt 7691 net/netfilter/nf_tables_api.c struct nft_table *table, *nt; nt 7701 net/netfilter/nf_tables_api.c list_for_each_entry_safe(table, nt, &net->nft.tables, list) { nt 179 net/netlink/af_netlink.c int netlink_add_tap(struct netlink_tap *nt) nt 181 net/netlink/af_netlink.c struct net *net = dev_net(nt->dev); nt 184 net/netlink/af_netlink.c if (unlikely(nt->dev->type != ARPHRD_NETLINK)) nt 188 net/netlink/af_netlink.c list_add_rcu(&nt->list, &nn->netlink_tap_all); nt 191 net/netlink/af_netlink.c __module_get(nt->module); nt 197 net/netlink/af_netlink.c static int __netlink_remove_tap(struct netlink_tap *nt) nt 199 net/netlink/af_netlink.c struct net *net = dev_net(nt->dev); nt 207 net/netlink/af_netlink.c if (nt == tmp) { nt 208 net/netlink/af_netlink.c list_del_rcu(&nt->list); nt 214 net/netlink/af_netlink.c pr_warn("__netlink_remove_tap: %p not found\n", nt); nt 219 net/netlink/af_netlink.c module_put(nt->module); nt 224 net/netlink/af_netlink.c int netlink_remove_tap(struct netlink_tap *nt) nt 228 net/netlink/af_netlink.c ret = __netlink_remove_tap(nt); nt 101 net/switchdev/switchdev.c static int switchdev_port_attr_notify(enum switchdev_notifier_type nt, nt 115 net/switchdev/switchdev.c rc = call_switchdev_blocking_notifiers(nt, dev, nt 221 net/switchdev/switchdev.c static int switchdev_port_obj_notify(enum switchdev_notifier_type nt, nt 236 net/switchdev/switchdev.c rc = call_switchdev_blocking_notifiers(nt, dev, &obj_info.info, extack); nt 89 net/tipc/name_distr.c struct name_table *nt = tipc_name_table(net); nt 94 net/tipc/name_distr.c list_add_tail_rcu(&publ->binding_node, &nt->node_scope); nt 97 net/tipc/name_distr.c write_lock_bh(&nt->cluster_scope_lock); nt 98 net/tipc/name_distr.c list_add_tail(&publ->binding_node, &nt->cluster_scope); nt 99 net/tipc/name_distr.c write_unlock_bh(&nt->cluster_scope_lock); nt 116 net/tipc/name_distr.c struct name_table *nt = tipc_name_table(net); nt 120 net/tipc/name_distr.c write_lock_bh(&nt->cluster_scope_lock); nt 122 net/tipc/name_distr.c write_unlock_bh(&nt->cluster_scope_lock); nt 190 net/tipc/name_distr.c struct name_table *nt = tipc_name_table(net); nt 195 net/tipc/name_distr.c read_lock_bh(&nt->cluster_scope_lock); nt 196 net/tipc/name_distr.c named_distribute(net, &head, dnode, &nt->cluster_scope); nt 198 net/tipc/name_distr.c read_unlock_bh(&nt->cluster_scope_lock); nt 337 net/tipc/name_distr.c struct name_table *nt = tipc_name_table(net); nt 344 net/tipc/name_distr.c list_for_each_entry_rcu(publ, &nt->node_scope, binding_node) nt 346 net/tipc/name_distr.c list_for_each_entry_rcu(publ, &nt->cluster_scope, binding_node) nt 324 net/tipc/name_table.c struct name_table *nt = tipc_name_table(net); nt 328 net/tipc/name_table.c service_head = &nt->services[hash(type)]; nt 341 net/tipc/name_table.c struct name_table *nt = tipc_name_table(net); nt 352 net/tipc/name_table.c sc = tipc_service_create(type, &nt->services[hash(type)]); nt 614 net/tipc/name_table.c struct name_table *nt = tipc_name_table(net); nt 621 net/tipc/name_table.c if (nt->local_publ_count >= TIPC_MAX_PUBL) { nt 629 net/tipc/name_table.c nt->local_publ_count++; nt 646 net/tipc/name_table.c struct name_table *nt = tipc_name_table(net); nt 656 net/tipc/name_table.c nt->local_publ_count--; nt 678 net/tipc/name_table.c struct name_table *nt = tipc_name_table(sub->net); nt 688 net/tipc/name_table.c sc = tipc_service_create(type, &nt->services[hash(type)]); nt 735 net/tipc/name_table.c struct name_table *nt; nt 738 net/tipc/name_table.c nt = kzalloc(sizeof(*nt), GFP_KERNEL); nt 739 net/tipc/name_table.c if (!nt) nt 743 net/tipc/name_table.c INIT_HLIST_HEAD(&nt->services[i]); nt 745 net/tipc/name_table.c INIT_LIST_HEAD(&nt->node_scope); nt 746 net/tipc/name_table.c INIT_LIST_HEAD(&nt->cluster_scope); nt 747 net/tipc/name_table.c rwlock_init(&nt->cluster_scope_lock); nt 748 net/tipc/name_table.c tn->nametbl = nt; nt 777 net/tipc/name_table.c struct name_table *nt = tipc_name_table(net); nt 788 net/tipc/name_table.c if (hlist_empty(&nt->services[i])) nt 790 net/tipc/name_table.c service_head = &nt->services[i]; nt 798 net/tipc/name_table.c kfree(nt); nt 869 net/tipc/netlink_compat.c struct nlattr *nt[TIPC_NLA_NAME_TABLE_MAX + 1]; nt 879 net/tipc/netlink_compat.c err = nla_parse_nested_deprecated(nt, TIPC_NLA_NAME_TABLE_MAX, nt 885 net/tipc/netlink_compat.c if (!nt[TIPC_NLA_NAME_TABLE_PUBL]) nt 889 net/tipc/netlink_compat.c nt[TIPC_NLA_NAME_TABLE_PUBL], NULL, nt 183 tools/perf/util/thread_map.c struct perf_thread_map *threads = NULL, *nt; nt 213 tools/perf/util/thread_map.c nt = perf_thread_map__realloc(threads, total_tasks); nt 214 tools/perf/util/thread_map.c if (nt == NULL) nt 217 tools/perf/util/thread_map.c threads = nt; nt 245 tools/perf/util/thread_map.c struct perf_thread_map *threads = NULL, *nt; nt 272 tools/perf/util/thread_map.c nt = perf_thread_map__realloc(threads, ntasks); nt 274 tools/perf/util/thread_map.c if (nt == NULL) nt 277 tools/perf/util/thread_map.c threads = nt;