Lines Matching refs:connection
112 drbd_queue_work(&peer_device->connection->sender_work, &peer_req->w); in drbd_endio_read_sec_final()
166 wake_asender(peer_device->connection); in drbd_endio_write_sec_final()
357 digest_size = crypto_hash_digestsize(peer_device->connection->csums_tfm); in w_e_send_csum()
362 drbd_csum_ee(peer_device->connection->csums_tfm, peer_req, digest); in w_e_send_csum()
451 &first_peer_device(device)->connection->sender_work, in resync_timer_fn()
589 struct drbd_connection *const connection = peer_device ? peer_device->connection : NULL; in make_resync_request() local
624 mutex_lock(&connection->data.mutex); in make_resync_request()
625 if (connection->data.socket) { in make_resync_request()
626 struct sock *sk = connection->data.socket->sk; in make_resync_request()
636 mutex_unlock(&connection->data.mutex); in make_resync_request()
833 struct drbd_connection *connection = first_peer_device(device)->connection; in ping_peer() local
835 clear_bit(GOT_PING_ACK, &connection->flags); in ping_peer()
836 request_ping(connection); in ping_peer()
837 wait_event(connection->ping_wait, in ping_peer()
838 test_bit(GOT_PING_ACK, &connection->flags) || device->state.conn < C_CONNECTED); in ping_peer()
864 drbd_queue_work(&first_peer_device(device)->connection->sender_work, in drbd_resync_finished()
1132 if (peer_device->connection->csums_tfm) { in w_e_end_csum_rs_req()
1133 digest_size = crypto_hash_digestsize(peer_device->connection->csums_tfm); in w_e_end_csum_rs_req()
1138 drbd_csum_ee(peer_device->connection->csums_tfm, peer_req, digest); in w_e_end_csum_rs_req()
1183 digest_size = crypto_hash_digestsize(peer_device->connection->verify_tfm); in w_e_end_ov_req()
1191 drbd_csum_ee(peer_device->connection->verify_tfm, peer_req, digest); in w_e_end_ov_req()
1255 digest_size = crypto_hash_digestsize(peer_device->connection->verify_tfm); in w_e_end_ov_reply()
1258 drbd_csum_ee(peer_device->connection->verify_tfm, peer_req, digest); in w_e_end_ov_reply()
1304 static int drbd_send_barrier(struct drbd_connection *connection) in drbd_send_barrier() argument
1309 sock = &connection->data; in drbd_send_barrier()
1310 p = conn_prepare_command(connection, sock); in drbd_send_barrier()
1313 p->barrier = connection->send.current_epoch_nr; in drbd_send_barrier()
1315 connection->send.current_epoch_writes = 0; in drbd_send_barrier()
1317 return conn_send_command(connection, sock, P_BARRIER, sizeof(*p), NULL, 0); in drbd_send_barrier()
1328 sock = &first_peer_device(device)->connection->data; in w_send_write_hint()
1334 static void re_init_if_first_write(struct drbd_connection *connection, unsigned int epoch) in re_init_if_first_write() argument
1336 if (!connection->send.seen_any_write_yet) { in re_init_if_first_write()
1337 connection->send.seen_any_write_yet = true; in re_init_if_first_write()
1338 connection->send.current_epoch_nr = epoch; in re_init_if_first_write()
1339 connection->send.current_epoch_writes = 0; in re_init_if_first_write()
1343 static void maybe_send_barrier(struct drbd_connection *connection, unsigned int epoch) in maybe_send_barrier() argument
1346 if (!connection->send.seen_any_write_yet) in maybe_send_barrier()
1348 if (connection->send.current_epoch_nr != epoch) { in maybe_send_barrier()
1349 if (connection->send.current_epoch_writes) in maybe_send_barrier()
1350 drbd_send_barrier(connection); in maybe_send_barrier()
1351 connection->send.current_epoch_nr = epoch; in maybe_send_barrier()
1360 struct drbd_connection *const connection = peer_device->connection; in w_send_out_of_sync() local
1373 maybe_send_barrier(connection, req->epoch); in w_send_out_of_sync()
1391 struct drbd_connection *connection = peer_device->connection; in w_send_dblock() local
1400 re_init_if_first_write(connection, req->epoch); in w_send_dblock()
1401 maybe_send_barrier(connection, req->epoch); in w_send_dblock()
1402 connection->send.current_epoch_writes++; in w_send_dblock()
1420 struct drbd_connection *connection = peer_device->connection; in w_send_read_req() local
1431 maybe_send_barrier(connection, req->epoch); in w_send_read_req()
1635 static bool use_checksum_based_resync(struct drbd_connection *connection, struct drbd_device *devic… in use_checksum_based_resync() argument
1639 csums_after_crash_only = rcu_dereference(connection->net_conf)->csums_after_crash_only; in use_checksum_based_resync()
1641 return connection->agreed_pro_version >= 89 && /* supported? */ in use_checksum_based_resync()
1642 connection->csums_tfm && /* configured? */ in use_checksum_based_resync()
1658 struct drbd_connection *connection = peer_device ? peer_device->connection : NULL; in drbd_start_resync() local
1677 conn_request_state(connection, NS(conn, C_DISCONNECTING), CS_HARD); in drbd_start_resync()
1690 conn_request_state(connection, in drbd_start_resync()
1698 if (current == connection->worker.task) { in drbd_start_resync()
1782 device->use_csums = use_checksum_based_resync(connection, device); in drbd_start_resync()
1794 if (side == C_SYNC_SOURCE && connection->agreed_pro_version < 96) in drbd_start_resync()
1797 if (connection->agreed_pro_version < 95 && device->rs_total == 0) { in drbd_start_resync()
1813 nc = rcu_dereference(connection->net_conf); in drbd_start_resync()
1971 static void do_unqueued_work(struct drbd_connection *connection) in do_unqueued_work() argument
1977 idr_for_each_entry(&connection->peer_devices, peer_device, vnr) { in do_unqueued_work()
2000 static void wait_for_work(struct drbd_connection *connection, struct list_head *work_list) in wait_for_work() argument
2006 dequeue_work_batch(&connection->sender_work, work_list); in wait_for_work()
2017 nc = rcu_dereference(connection->net_conf); in wait_for_work()
2021 mutex_lock(&connection->data.mutex); in wait_for_work()
2022 if (connection->data.socket) in wait_for_work()
2023 drbd_tcp_uncork(connection->data.socket); in wait_for_work()
2024 mutex_unlock(&connection->data.mutex); in wait_for_work()
2029 prepare_to_wait(&connection->sender_work.q_wait, &wait, TASK_INTERRUPTIBLE); in wait_for_work()
2030 spin_lock_irq(&connection->resource->req_lock); in wait_for_work()
2031 spin_lock(&connection->sender_work.q_lock); /* FIXME get rid of this one? */ in wait_for_work()
2032 if (!list_empty(&connection->sender_work.q)) in wait_for_work()
2033 list_splice_tail_init(&connection->sender_work.q, work_list); in wait_for_work()
2034 spin_unlock(&connection->sender_work.q_lock); /* FIXME get rid of this one? */ in wait_for_work()
2036 spin_unlock_irq(&connection->resource->req_lock); in wait_for_work()
2048 atomic_read(&connection->current_tle_nr) != in wait_for_work()
2049 connection->send.current_epoch_nr; in wait_for_work()
2050 spin_unlock_irq(&connection->resource->req_lock); in wait_for_work()
2053 maybe_send_barrier(connection, in wait_for_work()
2054 connection->send.current_epoch_nr + 1); in wait_for_work()
2056 if (test_bit(DEVICE_WORK_PENDING, &connection->flags)) in wait_for_work()
2060 if (get_t_state(&connection->worker) != RUNNING) in wait_for_work()
2068 finish_wait(&connection->sender_work.q_wait, &wait); in wait_for_work()
2072 nc = rcu_dereference(connection->net_conf); in wait_for_work()
2075 mutex_lock(&connection->data.mutex); in wait_for_work()
2076 if (connection->data.socket) { in wait_for_work()
2078 drbd_tcp_cork(connection->data.socket); in wait_for_work()
2080 drbd_tcp_uncork(connection->data.socket); in wait_for_work()
2082 mutex_unlock(&connection->data.mutex); in wait_for_work()
2087 struct drbd_connection *connection = thi->connection; in drbd_worker() local
2097 update_worker_timing_details(connection, wait_for_work); in drbd_worker()
2098 wait_for_work(connection, &work_list); in drbd_worker()
2101 if (test_and_clear_bit(DEVICE_WORK_PENDING, &connection->flags)) { in drbd_worker()
2102 update_worker_timing_details(connection, do_unqueued_work); in drbd_worker()
2103 do_unqueued_work(connection); in drbd_worker()
2109 drbd_warn(connection, "Worker got an unexpected signal\n"); in drbd_worker()
2121 update_worker_timing_details(connection, w->cb); in drbd_worker()
2122 if (w->cb(w, connection->cstate < C_WF_REPORT_PARAMS) == 0) in drbd_worker()
2124 if (connection->cstate >= C_WF_REPORT_PARAMS) in drbd_worker()
2125 conn_request_state(connection, NS(conn, C_NETWORK_FAILURE), CS_HARD); in drbd_worker()
2130 if (test_and_clear_bit(DEVICE_WORK_PENDING, &connection->flags)) { in drbd_worker()
2131 update_worker_timing_details(connection, do_unqueued_work); in drbd_worker()
2132 do_unqueued_work(connection); in drbd_worker()
2137 update_worker_timing_details(connection, w->cb); in drbd_worker()
2140 dequeue_work_batch(&connection->sender_work, &work_list); in drbd_worker()
2141 } while (!list_empty(&work_list) || test_bit(DEVICE_WORK_PENDING, &connection->flags)); in drbd_worker()
2144 idr_for_each_entry(&connection->peer_devices, peer_device, vnr) { in drbd_worker()