Lines Matching refs:nbd

67 	struct nbd_device *nbd;  member
142 struct nbd_device *nbd; member
156 #define nbd_name(nbd) ((nbd)->disk->disk_name) argument
166 static int nbd_dev_dbg_init(struct nbd_device *nbd);
167 static void nbd_dev_dbg_close(struct nbd_device *nbd);
168 static void nbd_config_put(struct nbd_device *nbd);
172 static void nbd_disconnect_and_put(struct nbd_device *nbd);
174 static inline struct device *nbd_to_dev(struct nbd_device *nbd) in nbd_to_dev() argument
176 return disk_to_dev(nbd->disk); in nbd_to_dev()
224 struct nbd_device *nbd = (struct nbd_device *)disk->private_data; in pid_show() local
226 return sprintf(buf, "%d\n", nbd->pid); in pid_show()
238 struct nbd_device *nbd = (struct nbd_device *)disk->private_data; in backend_show() local
240 return sprintf(buf, "%s\n", nbd->backend ?: ""); in backend_show()
248 static void nbd_dev_remove(struct nbd_device *nbd) in nbd_dev_remove() argument
250 struct gendisk *disk = nbd->disk; in nbd_dev_remove()
254 blk_mq_free_tag_set(&nbd->tag_set); in nbd_dev_remove()
261 idr_remove(&nbd_index_idr, nbd->index); in nbd_dev_remove()
263 destroy_workqueue(nbd->recv_workq); in nbd_dev_remove()
264 kfree(nbd); in nbd_dev_remove()
272 static void nbd_put(struct nbd_device *nbd) in nbd_put() argument
274 if (!refcount_dec_and_test(&nbd->refs)) in nbd_put()
278 if (test_bit(NBD_DESTROY_ON_DISCONNECT, &nbd->flags)) in nbd_put()
279 queue_work(nbd_del_wq, &nbd->remove_work); in nbd_put()
281 nbd_dev_remove(nbd); in nbd_put()
290 static void nbd_mark_nsock_dead(struct nbd_device *nbd, struct nbd_sock *nsock, in nbd_mark_nsock_dead() argument
293 if (!nsock->dead && notify && !nbd_disconnected(nbd->config)) { in nbd_mark_nsock_dead()
298 args->index = nbd->index; in nbd_mark_nsock_dead()
304 if (atomic_dec_return(&nbd->config->live_connections) == 0) { in nbd_mark_nsock_dead()
306 &nbd->config->runtime_flags)) { in nbd_mark_nsock_dead()
308 &nbd->config->runtime_flags); in nbd_mark_nsock_dead()
309 dev_info(nbd_to_dev(nbd), in nbd_mark_nsock_dead()
319 static int nbd_set_size(struct nbd_device *nbd, loff_t bytesize, in nbd_set_size() argument
328 nbd->config->bytesize = bytesize; in nbd_set_size()
329 nbd->config->blksize_bits = __ffs(blksize); in nbd_set_size()
331 if (!nbd->pid) in nbd_set_size()
334 if (nbd->config->flags & NBD_FLAG_SEND_TRIM) { in nbd_set_size()
335 nbd->disk->queue->limits.discard_granularity = blksize; in nbd_set_size()
336 nbd->disk->queue->limits.discard_alignment = blksize; in nbd_set_size()
337 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_set_size()
339 blk_queue_logical_block_size(nbd->disk->queue, blksize); in nbd_set_size()
340 blk_queue_physical_block_size(nbd->disk->queue, blksize); in nbd_set_size()
343 set_bit(GD_NEED_PART_SCAN, &nbd->disk->state); in nbd_set_size()
344 if (!set_capacity_and_notify(nbd->disk, bytesize >> 9)) in nbd_set_size()
345 kobject_uevent(&nbd_to_dev(nbd)->kobj, KOBJ_CHANGE); in nbd_set_size()
353 dev_dbg(nbd_to_dev(cmd->nbd), "request %p: %s\n", req, in nbd_complete_rq()
362 static void sock_shutdown(struct nbd_device *nbd) in sock_shutdown() argument
364 struct nbd_config *config = nbd->config; in sock_shutdown()
375 nbd_mark_nsock_dead(nbd, nsock, 0); in sock_shutdown()
378 dev_warn(disk_to_dev(nbd->disk), "shutting down sockets\n"); in sock_shutdown()
401 struct nbd_device *nbd = cmd->nbd; in nbd_xmit_timeout() local
412 if (!refcount_inc_not_zero(&nbd->config_refs)) { in nbd_xmit_timeout()
417 config = nbd->config; in nbd_xmit_timeout()
420 (config->num_connections == 1 && nbd->tag_set.timeout)) { in nbd_xmit_timeout()
421 dev_err_ratelimited(nbd_to_dev(nbd), in nbd_xmit_timeout()
443 nbd_mark_nsock_dead(nbd, nsock, 1); in nbd_xmit_timeout()
448 nbd_config_put(nbd); in nbd_xmit_timeout()
453 if (!nbd->tag_set.timeout) { in nbd_xmit_timeout()
460 …dev_info(nbd_to_dev(nbd), "Possible stuck request %p: control (%s@%llu,%uB). Runtime %u seconds\n", in nbd_xmit_timeout()
470 nbd_config_put(nbd); in nbd_xmit_timeout()
475 nbd_config_put(nbd); in nbd_xmit_timeout()
479 dev_err_ratelimited(nbd_to_dev(nbd), "Connection timed out\n"); in nbd_xmit_timeout()
483 sock_shutdown(nbd); in nbd_xmit_timeout()
484 nbd_config_put(nbd); in nbd_xmit_timeout()
494 static int sock_xmit(struct nbd_device *nbd, int index, int send, in sock_xmit() argument
497 struct nbd_config *config = nbd->config; in sock_xmit()
504 dev_err_ratelimited(disk_to_dev(nbd->disk), in sock_xmit()
550 static int nbd_send_cmd(struct nbd_device *nbd, struct nbd_cmd *cmd, int index) in nbd_send_cmd() argument
553 struct nbd_config *config = nbd->config; in nbd_send_cmd()
574 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
610 trace_nbd_send_request(&request, nbd->index, blk_mq_rq_from_pdu(cmd)); in nbd_send_cmd()
612 dev_dbg(nbd_to_dev(nbd), "request %p: sending control (%s@%llu,%uB)\n", in nbd_send_cmd()
615 result = sock_xmit(nbd, index, 1, &from, in nbd_send_cmd()
632 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
650 dev_dbg(nbd_to_dev(nbd), "request %p: sending %d bytes data\n", in nbd_send_cmd()
661 result = sock_xmit(nbd, index, 1, &from, flags, &sent); in nbd_send_cmd()
673 dev_err(disk_to_dev(nbd->disk), in nbd_send_cmd()
696 static int nbd_read_reply(struct nbd_device *nbd, int index, in nbd_read_reply() argument
705 result = sock_xmit(nbd, index, 0, &to, MSG_WAITALL, NULL); in nbd_read_reply()
707 if (!nbd_disconnected(nbd->config)) in nbd_read_reply()
708 dev_err(disk_to_dev(nbd->disk), in nbd_read_reply()
714 dev_err(disk_to_dev(nbd->disk), "Wrong magic (0x%lx)\n", in nbd_read_reply()
723 static struct nbd_cmd *nbd_handle_reply(struct nbd_device *nbd, int index, in nbd_handle_reply() argument
737 if (hwq < nbd->tag_set.nr_hw_queues) in nbd_handle_reply()
738 req = blk_mq_tag_to_rq(nbd->tag_set.tags[hwq], in nbd_handle_reply()
741 dev_err(disk_to_dev(nbd->disk), "Unexpected reply (%d) %p\n", in nbd_handle_reply()
750 dev_err(disk_to_dev(nbd->disk), "Suspicious reply %d (status %u flags %lu)", in nbd_handle_reply()
756 dev_err(disk_to_dev(nbd->disk), "Unexpected reply %d from different sock %d (expected %d)", in nbd_handle_reply()
762 dev_err(disk_to_dev(nbd->disk), "Double reply on req %p, cmd_cookie %u, handle cookie %u\n", in nbd_handle_reply()
768 dev_err(disk_to_dev(nbd->disk), "Command already handled %p\n", in nbd_handle_reply()
774 dev_err(disk_to_dev(nbd->disk), "Raced with timeout on req %p\n", in nbd_handle_reply()
780 dev_err(disk_to_dev(nbd->disk), "Other side returned error (%d)\n", in nbd_handle_reply()
786 dev_dbg(nbd_to_dev(nbd), "request %p: got reply\n", req); in nbd_handle_reply()
794 result = sock_xmit(nbd, index, 0, &to, MSG_WAITALL, NULL); in nbd_handle_reply()
796 dev_err(disk_to_dev(nbd->disk), "Receive data failed (result %d)\n", in nbd_handle_reply()
804 if (nbd_disconnected(nbd->config)) { in nbd_handle_reply()
811 dev_dbg(nbd_to_dev(nbd), "request %p: got %d bytes data\n", in nbd_handle_reply()
826 struct nbd_device *nbd = args->nbd; in recv_work() local
827 struct nbd_config *config = nbd->config; in recv_work()
828 struct request_queue *q = nbd->disk->queue; in recv_work()
836 if (nbd_read_reply(nbd, args->index, &reply)) in recv_work()
846 dev_err(disk_to_dev(nbd->disk), "%s: no io inflight\n", in recv_work()
851 cmd = nbd_handle_reply(nbd, args->index, &reply); in recv_work()
865 nbd_mark_nsock_dead(nbd, nsock, 1); in recv_work()
868 nbd_config_put(nbd); in recv_work()
894 static void nbd_clear_que(struct nbd_device *nbd) in nbd_clear_que() argument
896 blk_mq_quiesce_queue(nbd->disk->queue); in nbd_clear_que()
897 blk_mq_tagset_busy_iter(&nbd->tag_set, nbd_clear_req, NULL); in nbd_clear_que()
898 blk_mq_unquiesce_queue(nbd->disk->queue); in nbd_clear_que()
899 dev_dbg(disk_to_dev(nbd->disk), "queue cleared\n"); in nbd_clear_que()
902 static int find_fallback(struct nbd_device *nbd, int index) in find_fallback() argument
904 struct nbd_config *config = nbd->config; in find_fallback()
913 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
936 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
945 static int wait_for_reconnect(struct nbd_device *nbd) in wait_for_reconnect() argument
947 struct nbd_config *config = nbd->config; in wait_for_reconnect()
960 struct nbd_device *nbd = cmd->nbd; in nbd_handle_cmd() local
965 if (!refcount_inc_not_zero(&nbd->config_refs)) { in nbd_handle_cmd()
966 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
970 config = nbd->config; in nbd_handle_cmd()
973 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
975 nbd_config_put(nbd); in nbd_handle_cmd()
984 index = find_fallback(nbd, index); in nbd_handle_cmd()
987 if (wait_for_reconnect(nbd)) { in nbd_handle_cmd()
997 sock_shutdown(nbd); in nbd_handle_cmd()
998 nbd_config_put(nbd); in nbd_handle_cmd()
1019 ret = nbd_send_cmd(nbd, cmd, index); in nbd_handle_cmd()
1027 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
1029 nbd_mark_nsock_dead(nbd, nsock, 1); in nbd_handle_cmd()
1035 nbd_config_put(nbd); in nbd_handle_cmd()
1072 static struct socket *nbd_get_socket(struct nbd_device *nbd, unsigned long fd, in nbd_get_socket() argument
1083 dev_err(disk_to_dev(nbd->disk), "Unsupported socket: shutdown callout must be supported.\n"); in nbd_get_socket()
1092 static int nbd_add_socket(struct nbd_device *nbd, unsigned long arg, in nbd_add_socket() argument
1095 struct nbd_config *config = nbd->config; in nbd_add_socket()
1101 sock = nbd_get_socket(nbd, arg, &err); in nbd_add_socket()
1109 blk_mq_freeze_queue(nbd->disk->queue); in nbd_add_socket()
1111 if (!netlink && !nbd->task_setup && in nbd_add_socket()
1113 nbd->task_setup = current; in nbd_add_socket()
1116 (nbd->task_setup != current || in nbd_add_socket()
1118 dev_err(disk_to_dev(nbd->disk), in nbd_add_socket()
1149 blk_mq_unfreeze_queue(nbd->disk->queue); in nbd_add_socket()
1154 blk_mq_unfreeze_queue(nbd->disk->queue); in nbd_add_socket()
1159 static int nbd_reconnect_socket(struct nbd_device *nbd, unsigned long arg) in nbd_reconnect_socket() argument
1161 struct nbd_config *config = nbd->config; in nbd_reconnect_socket()
1167 sock = nbd_get_socket(nbd, arg, &err); in nbd_reconnect_socket()
1189 if (nbd->tag_set.timeout) in nbd_reconnect_socket()
1190 sock->sk->sk_sndtimeo = nbd->tag_set.timeout; in nbd_reconnect_socket()
1192 refcount_inc(&nbd->config_refs); in nbd_reconnect_socket()
1199 args->nbd = nbd; in nbd_reconnect_socket()
1209 queue_work(nbd->recv_workq, &args->work); in nbd_reconnect_socket()
1227 static void nbd_parse_flags(struct nbd_device *nbd) in nbd_parse_flags() argument
1229 struct nbd_config *config = nbd->config; in nbd_parse_flags()
1231 set_disk_ro(nbd->disk, true); in nbd_parse_flags()
1233 set_disk_ro(nbd->disk, false); in nbd_parse_flags()
1235 blk_queue_flag_set(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_parse_flags()
1238 blk_queue_write_cache(nbd->disk->queue, true, true); in nbd_parse_flags()
1240 blk_queue_write_cache(nbd->disk->queue, true, false); in nbd_parse_flags()
1243 blk_queue_write_cache(nbd->disk->queue, false, false); in nbd_parse_flags()
1246 static void send_disconnects(struct nbd_device *nbd) in send_disconnects() argument
1248 struct nbd_config *config = nbd->config; in send_disconnects()
1262 ret = sock_xmit(nbd, i, 1, &from, 0, NULL); in send_disconnects()
1264 dev_err(disk_to_dev(nbd->disk), in send_disconnects()
1270 static int nbd_disconnect(struct nbd_device *nbd) in nbd_disconnect() argument
1272 struct nbd_config *config = nbd->config; in nbd_disconnect()
1274 dev_info(disk_to_dev(nbd->disk), "NBD_DISCONNECT\n"); in nbd_disconnect()
1276 set_bit(NBD_DISCONNECT_REQUESTED, &nbd->flags); in nbd_disconnect()
1277 send_disconnects(nbd); in nbd_disconnect()
1281 static void nbd_clear_sock(struct nbd_device *nbd) in nbd_clear_sock() argument
1283 sock_shutdown(nbd); in nbd_clear_sock()
1284 nbd_clear_que(nbd); in nbd_clear_sock()
1285 nbd->task_setup = NULL; in nbd_clear_sock()
1288 static void nbd_config_put(struct nbd_device *nbd) in nbd_config_put() argument
1290 if (refcount_dec_and_mutex_lock(&nbd->config_refs, in nbd_config_put()
1291 &nbd->config_lock)) { in nbd_config_put()
1292 struct nbd_config *config = nbd->config; in nbd_config_put()
1293 nbd_dev_dbg_close(nbd); in nbd_config_put()
1294 invalidate_disk(nbd->disk); in nbd_config_put()
1295 if (nbd->config->bytesize) in nbd_config_put()
1296 kobject_uevent(&nbd_to_dev(nbd)->kobj, KOBJ_CHANGE); in nbd_config_put()
1299 device_remove_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_config_put()
1300 nbd->pid = 0; in nbd_config_put()
1303 device_remove_file(disk_to_dev(nbd->disk), &backend_attr); in nbd_config_put()
1304 kfree(nbd->backend); in nbd_config_put()
1305 nbd->backend = NULL; in nbd_config_put()
1307 nbd_clear_sock(nbd); in nbd_config_put()
1316 kfree(nbd->config); in nbd_config_put()
1317 nbd->config = NULL; in nbd_config_put()
1319 nbd->tag_set.timeout = 0; in nbd_config_put()
1320 nbd->disk->queue->limits.discard_granularity = 0; in nbd_config_put()
1321 nbd->disk->queue->limits.discard_alignment = 0; in nbd_config_put()
1322 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_config_put()
1323 blk_queue_flag_clear(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_config_put()
1325 mutex_unlock(&nbd->config_lock); in nbd_config_put()
1326 nbd_put(nbd); in nbd_config_put()
1331 static int nbd_start_device(struct nbd_device *nbd) in nbd_start_device() argument
1333 struct nbd_config *config = nbd->config; in nbd_start_device()
1337 if (nbd->pid) in nbd_start_device()
1343 dev_err(disk_to_dev(nbd->disk), "server does not support multiple connections per device.\n"); in nbd_start_device()
1347 blk_mq_update_nr_hw_queues(&nbd->tag_set, config->num_connections); in nbd_start_device()
1348 nbd->pid = task_pid_nr(current); in nbd_start_device()
1350 nbd_parse_flags(nbd); in nbd_start_device()
1352 error = device_create_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_start_device()
1354 dev_err(disk_to_dev(nbd->disk), "device_create_file failed for pid!\n"); in nbd_start_device()
1359 nbd_dev_dbg_init(nbd); in nbd_start_device()
1365 sock_shutdown(nbd); in nbd_start_device()
1375 flush_workqueue(nbd->recv_workq); in nbd_start_device()
1379 if (nbd->tag_set.timeout) in nbd_start_device()
1381 nbd->tag_set.timeout; in nbd_start_device()
1383 refcount_inc(&nbd->config_refs); in nbd_start_device()
1385 args->nbd = nbd; in nbd_start_device()
1387 queue_work(nbd->recv_workq, &args->work); in nbd_start_device()
1389 return nbd_set_size(nbd, config->bytesize, nbd_blksize(config)); in nbd_start_device()
1392 static int nbd_start_device_ioctl(struct nbd_device *nbd, struct block_device *bdev) in nbd_start_device_ioctl() argument
1394 struct nbd_config *config = nbd->config; in nbd_start_device_ioctl()
1397 ret = nbd_start_device(nbd); in nbd_start_device_ioctl()
1402 set_bit(GD_NEED_PART_SCAN, &nbd->disk->state); in nbd_start_device_ioctl()
1403 mutex_unlock(&nbd->config_lock); in nbd_start_device_ioctl()
1407 sock_shutdown(nbd); in nbd_start_device_ioctl()
1408 flush_workqueue(nbd->recv_workq); in nbd_start_device_ioctl()
1410 mutex_lock(&nbd->config_lock); in nbd_start_device_ioctl()
1420 static void nbd_clear_sock_ioctl(struct nbd_device *nbd, in nbd_clear_sock_ioctl() argument
1423 sock_shutdown(nbd); in nbd_clear_sock_ioctl()
1427 &nbd->config->runtime_flags)) in nbd_clear_sock_ioctl()
1428 nbd_config_put(nbd); in nbd_clear_sock_ioctl()
1431 static void nbd_set_cmd_timeout(struct nbd_device *nbd, u64 timeout) in nbd_set_cmd_timeout() argument
1433 nbd->tag_set.timeout = timeout * HZ; in nbd_set_cmd_timeout()
1435 blk_queue_rq_timeout(nbd->disk->queue, timeout * HZ); in nbd_set_cmd_timeout()
1437 blk_queue_rq_timeout(nbd->disk->queue, 30 * HZ); in nbd_set_cmd_timeout()
1441 static int __nbd_ioctl(struct block_device *bdev, struct nbd_device *nbd, in __nbd_ioctl() argument
1444 struct nbd_config *config = nbd->config; in __nbd_ioctl()
1449 return nbd_disconnect(nbd); in __nbd_ioctl()
1451 nbd_clear_sock_ioctl(nbd, bdev); in __nbd_ioctl()
1454 return nbd_add_socket(nbd, arg, false); in __nbd_ioctl()
1456 return nbd_set_size(nbd, config->bytesize, arg); in __nbd_ioctl()
1458 return nbd_set_size(nbd, arg, nbd_blksize(config)); in __nbd_ioctl()
1462 return nbd_set_size(nbd, bytesize, nbd_blksize(config)); in __nbd_ioctl()
1464 nbd_set_cmd_timeout(nbd, arg); in __nbd_ioctl()
1471 return nbd_start_device_ioctl(nbd, bdev); in __nbd_ioctl()
1491 struct nbd_device *nbd = bdev->bd_disk->private_data; in nbd_ioctl() local
1492 struct nbd_config *config = nbd->config; in nbd_ioctl()
1504 mutex_lock(&nbd->config_lock); in nbd_ioctl()
1511 error = __nbd_ioctl(bdev, nbd, cmd, arg); in nbd_ioctl()
1513 dev_err(nbd_to_dev(nbd), "Cannot use ioctl interface on a netlink controlled device.\n"); in nbd_ioctl()
1514 mutex_unlock(&nbd->config_lock); in nbd_ioctl()
1536 struct nbd_device *nbd; in nbd_open() local
1540 nbd = bdev->bd_disk->private_data; in nbd_open()
1541 if (!nbd) { in nbd_open()
1545 if (!refcount_inc_not_zero(&nbd->refs)) { in nbd_open()
1549 if (!refcount_inc_not_zero(&nbd->config_refs)) { in nbd_open()
1552 mutex_lock(&nbd->config_lock); in nbd_open()
1553 if (refcount_inc_not_zero(&nbd->config_refs)) { in nbd_open()
1554 mutex_unlock(&nbd->config_lock); in nbd_open()
1557 config = nbd->config = nbd_alloc_config(); in nbd_open()
1560 mutex_unlock(&nbd->config_lock); in nbd_open()
1563 refcount_set(&nbd->config_refs, 1); in nbd_open()
1564 refcount_inc(&nbd->refs); in nbd_open()
1565 mutex_unlock(&nbd->config_lock); in nbd_open()
1568 } else if (nbd_disconnected(nbd->config)) { in nbd_open()
1579 struct nbd_device *nbd = disk->private_data; in nbd_release() local
1581 if (test_bit(NBD_RT_DISCONNECT_ON_CLOSE, &nbd->config->runtime_flags) && in nbd_release()
1583 nbd_disconnect_and_put(nbd); in nbd_release()
1585 nbd_config_put(nbd); in nbd_release()
1586 nbd_put(nbd); in nbd_release()
1602 struct nbd_device *nbd = s->private; in nbd_dbg_tasks_show() local
1604 if (nbd->pid) in nbd_dbg_tasks_show()
1605 seq_printf(s, "recv: %d\n", nbd->pid); in nbd_dbg_tasks_show()
1614 struct nbd_device *nbd = s->private; in nbd_dbg_flags_show() local
1615 u32 flags = nbd->config->flags; in nbd_dbg_flags_show()
1637 static int nbd_dev_dbg_init(struct nbd_device *nbd) in nbd_dev_dbg_init() argument
1640 struct nbd_config *config = nbd->config; in nbd_dev_dbg_init()
1645 dir = debugfs_create_dir(nbd_name(nbd), nbd_dbg_dir); in nbd_dev_dbg_init()
1647 dev_err(nbd_to_dev(nbd), "Failed to create debugfs dir for '%s'\n", in nbd_dev_dbg_init()
1648 nbd_name(nbd)); in nbd_dev_dbg_init()
1653 debugfs_create_file("tasks", 0444, dir, nbd, &nbd_dbg_tasks_fops); in nbd_dev_dbg_init()
1655 debugfs_create_u32("timeout", 0444, dir, &nbd->tag_set.timeout); in nbd_dev_dbg_init()
1657 debugfs_create_file("flags", 0444, dir, nbd, &nbd_dbg_flags_fops); in nbd_dev_dbg_init()
1662 static void nbd_dev_dbg_close(struct nbd_device *nbd) in nbd_dev_dbg_close() argument
1664 debugfs_remove_recursive(nbd->config->dbg_dir); in nbd_dev_dbg_close()
1687 static int nbd_dev_dbg_init(struct nbd_device *nbd) in nbd_dev_dbg_init() argument
1692 static void nbd_dev_dbg_close(struct nbd_device *nbd) in nbd_dev_dbg_close() argument
1711 cmd->nbd = set->driver_data; in nbd_init_request()
1726 struct nbd_device *nbd; in nbd_dev_add() local
1730 nbd = kzalloc(sizeof(struct nbd_device), GFP_KERNEL); in nbd_dev_add()
1731 if (!nbd) in nbd_dev_add()
1734 nbd->tag_set.ops = &nbd_mq_ops; in nbd_dev_add()
1735 nbd->tag_set.nr_hw_queues = 1; in nbd_dev_add()
1736 nbd->tag_set.queue_depth = 128; in nbd_dev_add()
1737 nbd->tag_set.numa_node = NUMA_NO_NODE; in nbd_dev_add()
1738 nbd->tag_set.cmd_size = sizeof(struct nbd_cmd); in nbd_dev_add()
1739 nbd->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | in nbd_dev_add()
1741 nbd->tag_set.driver_data = nbd; in nbd_dev_add()
1742 INIT_WORK(&nbd->remove_work, nbd_dev_remove_work); in nbd_dev_add()
1743 nbd->backend = NULL; in nbd_dev_add()
1745 err = blk_mq_alloc_tag_set(&nbd->tag_set); in nbd_dev_add()
1751 err = idr_alloc(&nbd_index_idr, nbd, index, index + 1, in nbd_dev_add()
1756 err = idr_alloc(&nbd_index_idr, nbd, 0, 0, GFP_KERNEL); in nbd_dev_add()
1760 nbd->index = index; in nbd_dev_add()
1765 disk = blk_mq_alloc_disk(&nbd->tag_set, NULL); in nbd_dev_add()
1770 nbd->disk = disk; in nbd_dev_add()
1772 nbd->recv_workq = alloc_workqueue("nbd%d-recv", in nbd_dev_add()
1774 WQ_UNBOUND, 0, nbd->index); in nbd_dev_add()
1775 if (!nbd->recv_workq) { in nbd_dev_add()
1776 dev_err(disk_to_dev(nbd->disk), "Could not allocate knbd recv work queue.\n"); in nbd_dev_add()
1794 mutex_init(&nbd->config_lock); in nbd_dev_add()
1795 refcount_set(&nbd->config_refs, 0); in nbd_dev_add()
1800 refcount_set(&nbd->refs, 0); in nbd_dev_add()
1801 INIT_LIST_HEAD(&nbd->list); in nbd_dev_add()
1816 disk->private_data = nbd; in nbd_dev_add()
1825 refcount_set(&nbd->refs, refs); in nbd_dev_add()
1827 return nbd; in nbd_dev_add()
1830 destroy_workqueue(nbd->recv_workq); in nbd_dev_add()
1838 blk_mq_free_tag_set(&nbd->tag_set); in nbd_dev_add()
1840 kfree(nbd); in nbd_dev_add()
1847 struct nbd_device *nbd; in nbd_find_get_unused() local
1852 idr_for_each_entry(&nbd_index_idr, nbd, id) { in nbd_find_get_unused()
1853 if (refcount_read(&nbd->config_refs) || in nbd_find_get_unused()
1854 test_bit(NBD_DESTROY_ON_DISCONNECT, &nbd->flags)) in nbd_find_get_unused()
1856 if (refcount_inc_not_zero(&nbd->refs)) in nbd_find_get_unused()
1857 return nbd; in nbd_find_get_unused()
1890 static int nbd_genl_size_set(struct genl_info *info, struct nbd_device *nbd) in nbd_genl_size_set() argument
1892 struct nbd_config *config = nbd->config; in nbd_genl_size_set()
1903 return nbd_set_size(nbd, bytes, bsize); in nbd_genl_size_set()
1909 struct nbd_device *nbd; in nbd_genl_connect() local
1931 nbd = nbd_find_get_unused(); in nbd_genl_connect()
1933 nbd = idr_find(&nbd_index_idr, index); in nbd_genl_connect()
1934 if (nbd) { in nbd_genl_connect()
1935 if ((test_bit(NBD_DESTROY_ON_DISCONNECT, &nbd->flags) && in nbd_genl_connect()
1936 test_bit(NBD_DISCONNECT_REQUESTED, &nbd->flags)) || in nbd_genl_connect()
1937 !refcount_inc_not_zero(&nbd->refs)) { in nbd_genl_connect()
1947 if (!nbd) { in nbd_genl_connect()
1948 nbd = nbd_dev_add(index, 2); in nbd_genl_connect()
1949 if (IS_ERR(nbd)) { in nbd_genl_connect()
1951 return PTR_ERR(nbd); in nbd_genl_connect()
1955 mutex_lock(&nbd->config_lock); in nbd_genl_connect()
1956 if (refcount_read(&nbd->config_refs)) { in nbd_genl_connect()
1957 mutex_unlock(&nbd->config_lock); in nbd_genl_connect()
1958 nbd_put(nbd); in nbd_genl_connect()
1964 if (WARN_ON(nbd->config)) { in nbd_genl_connect()
1965 mutex_unlock(&nbd->config_lock); in nbd_genl_connect()
1966 nbd_put(nbd); in nbd_genl_connect()
1969 config = nbd->config = nbd_alloc_config(); in nbd_genl_connect()
1970 if (!nbd->config) { in nbd_genl_connect()
1971 mutex_unlock(&nbd->config_lock); in nbd_genl_connect()
1972 nbd_put(nbd); in nbd_genl_connect()
1976 refcount_set(&nbd->config_refs, 1); in nbd_genl_connect()
1979 ret = nbd_genl_size_set(info, nbd); in nbd_genl_connect()
1984 nbd_set_cmd_timeout(nbd, in nbd_genl_connect()
2006 &nbd->flags)) in nbd_genl_connect()
2010 &nbd->flags)) in nbd_genl_connect()
2011 refcount_inc(&nbd->refs); in nbd_genl_connect()
2044 ret = nbd_add_socket(nbd, fd, true); in nbd_genl_connect()
2049 ret = nbd_start_device(nbd); in nbd_genl_connect()
2053 nbd->backend = nla_strdup(info->attrs[NBD_ATTR_BACKEND_IDENTIFIER], in nbd_genl_connect()
2055 if (!nbd->backend) { in nbd_genl_connect()
2060 ret = device_create_file(disk_to_dev(nbd->disk), &backend_attr); in nbd_genl_connect()
2062 dev_err(disk_to_dev(nbd->disk), in nbd_genl_connect()
2068 mutex_unlock(&nbd->config_lock); in nbd_genl_connect()
2071 refcount_inc(&nbd->config_refs); in nbd_genl_connect()
2072 nbd_connect_reply(info, nbd->index); in nbd_genl_connect()
2074 nbd_config_put(nbd); in nbd_genl_connect()
2076 nbd_put(nbd); in nbd_genl_connect()
2080 static void nbd_disconnect_and_put(struct nbd_device *nbd) in nbd_disconnect_and_put() argument
2082 mutex_lock(&nbd->config_lock); in nbd_disconnect_and_put()
2083 nbd_disconnect(nbd); in nbd_disconnect_and_put()
2084 sock_shutdown(nbd); in nbd_disconnect_and_put()
2089 flush_workqueue(nbd->recv_workq); in nbd_disconnect_and_put()
2090 nbd_clear_que(nbd); in nbd_disconnect_and_put()
2091 nbd->task_setup = NULL; in nbd_disconnect_and_put()
2092 mutex_unlock(&nbd->config_lock); in nbd_disconnect_and_put()
2095 &nbd->config->runtime_flags)) in nbd_disconnect_and_put()
2096 nbd_config_put(nbd); in nbd_disconnect_and_put()
2101 struct nbd_device *nbd; in nbd_genl_disconnect() local
2113 nbd = idr_find(&nbd_index_idr, index); in nbd_genl_disconnect()
2114 if (!nbd) { in nbd_genl_disconnect()
2120 if (!refcount_inc_not_zero(&nbd->refs)) { in nbd_genl_disconnect()
2127 if (!refcount_inc_not_zero(&nbd->config_refs)) in nbd_genl_disconnect()
2129 nbd_disconnect_and_put(nbd); in nbd_genl_disconnect()
2130 nbd_config_put(nbd); in nbd_genl_disconnect()
2132 nbd_put(nbd); in nbd_genl_disconnect()
2138 struct nbd_device *nbd = NULL; in nbd_genl_reconfigure() local
2153 nbd = idr_find(&nbd_index_idr, index); in nbd_genl_reconfigure()
2154 if (!nbd) { in nbd_genl_reconfigure()
2160 if (nbd->backend) { in nbd_genl_reconfigure()
2163 nbd->backend)) { in nbd_genl_reconfigure()
2165 dev_err(nbd_to_dev(nbd), in nbd_genl_reconfigure()
2167 nbd->backend); in nbd_genl_reconfigure()
2172 dev_err(nbd_to_dev(nbd), "must specify backend\n"); in nbd_genl_reconfigure()
2176 if (!refcount_inc_not_zero(&nbd->refs)) { in nbd_genl_reconfigure()
2184 if (!refcount_inc_not_zero(&nbd->config_refs)) { in nbd_genl_reconfigure()
2185 dev_err(nbd_to_dev(nbd), in nbd_genl_reconfigure()
2187 nbd_put(nbd); in nbd_genl_reconfigure()
2191 mutex_lock(&nbd->config_lock); in nbd_genl_reconfigure()
2192 config = nbd->config; in nbd_genl_reconfigure()
2194 !nbd->pid) { in nbd_genl_reconfigure()
2195 dev_err(nbd_to_dev(nbd), in nbd_genl_reconfigure()
2201 ret = nbd_genl_size_set(info, nbd); in nbd_genl_reconfigure()
2206 nbd_set_cmd_timeout(nbd, in nbd_genl_reconfigure()
2217 &nbd->flags)) in nbd_genl_reconfigure()
2221 &nbd->flags)) in nbd_genl_reconfigure()
2222 refcount_inc(&nbd->refs); in nbd_genl_reconfigure()
2259 ret = nbd_reconnect_socket(nbd, fd); in nbd_genl_reconfigure()
2265 dev_info(nbd_to_dev(nbd), "reconnected socket\n"); in nbd_genl_reconfigure()
2269 mutex_unlock(&nbd->config_lock); in nbd_genl_reconfigure()
2270 nbd_config_put(nbd); in nbd_genl_reconfigure()
2271 nbd_put(nbd); in nbd_genl_reconfigure()
2273 nbd_put(nbd); in nbd_genl_reconfigure()
2317 static int populate_nbd_status(struct nbd_device *nbd, struct sk_buff *reply) in populate_nbd_status() argument
2330 if (refcount_read(&nbd->config_refs)) in populate_nbd_status()
2335 ret = nla_put_u32(reply, NBD_DEVICE_INDEX, nbd->index); in populate_nbd_status()
2348 struct nbd_device *nbd = ptr; in status_cb() local
2349 return populate_nbd_status(nbd, (struct sk_buff *)data); in status_cb()
2388 struct nbd_device *nbd; in nbd_genl_status() local
2389 nbd = idr_find(&nbd_index_idr, index); in nbd_genl_status()
2390 if (nbd) { in nbd_genl_status()
2391 ret = populate_nbd_status(nbd, reply); in nbd_genl_status()
2518 struct nbd_device *nbd = ptr; in nbd_exit_cb() local
2521 if (refcount_read(&nbd->refs)) in nbd_exit_cb()
2522 list_add_tail(&nbd->list, list); in nbd_exit_cb()
2529 struct nbd_device *nbd; in nbd_cleanup() local
2539 nbd = list_first_entry(&del_list, struct nbd_device, list); in nbd_cleanup()
2540 list_del_init(&nbd->list); in nbd_cleanup()
2541 if (refcount_read(&nbd->refs) != 1) in nbd_cleanup()
2543 nbd_put(nbd); in nbd_cleanup()