Lines Matching refs:disk
120 struct gendisk *disk; member
156 #define nbd_name(nbd) ((nbd)->disk->disk_name)
176 return disk_to_dev(nbd->disk); in nbd_to_dev()
223 struct gendisk *disk = dev_to_disk(dev); in pid_show() local
224 struct nbd_device *nbd = (struct nbd_device *)disk->private_data; in pid_show()
237 struct gendisk *disk = dev_to_disk(dev); in backend_show() local
238 struct nbd_device *nbd = (struct nbd_device *)disk->private_data; in backend_show()
250 struct gendisk *disk = nbd->disk; in nbd_dev_remove() local
252 del_gendisk(disk); in nbd_dev_remove()
253 blk_cleanup_disk(disk); in nbd_dev_remove()
335 nbd->disk->queue->limits.discard_granularity = blksize; in nbd_set_size()
336 nbd->disk->queue->limits.discard_alignment = blksize; in nbd_set_size()
337 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_set_size()
339 blk_queue_logical_block_size(nbd->disk->queue, blksize); in nbd_set_size()
340 blk_queue_physical_block_size(nbd->disk->queue, blksize); in nbd_set_size()
343 set_bit(GD_NEED_PART_SCAN, &nbd->disk->state); in nbd_set_size()
344 if (!set_capacity_and_notify(nbd->disk, bytesize >> 9)) in nbd_set_size()
378 dev_warn(disk_to_dev(nbd->disk), "shutting down sockets\n"); in sock_shutdown()
504 dev_err_ratelimited(disk_to_dev(nbd->disk), in sock_xmit()
574 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
632 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
673 dev_err(disk_to_dev(nbd->disk), in nbd_send_cmd()
708 dev_err(disk_to_dev(nbd->disk), in nbd_read_reply()
714 dev_err(disk_to_dev(nbd->disk), "Wrong magic (0x%lx)\n", in nbd_read_reply()
741 dev_err(disk_to_dev(nbd->disk), "Unexpected reply (%d) %p\n", in nbd_handle_reply()
750 dev_err(disk_to_dev(nbd->disk), "Suspicious reply %d (status %u flags %lu)", in nbd_handle_reply()
756 dev_err(disk_to_dev(nbd->disk), "Unexpected reply %d from different sock %d (expected %d)", in nbd_handle_reply()
762 dev_err(disk_to_dev(nbd->disk), "Double reply on req %p, cmd_cookie %u, handle cookie %u\n", in nbd_handle_reply()
768 dev_err(disk_to_dev(nbd->disk), "Command already handled %p\n", in nbd_handle_reply()
774 dev_err(disk_to_dev(nbd->disk), "Raced with timeout on req %p\n", in nbd_handle_reply()
780 dev_err(disk_to_dev(nbd->disk), "Other side returned error (%d)\n", in nbd_handle_reply()
796 dev_err(disk_to_dev(nbd->disk), "Receive data failed (result %d)\n", in nbd_handle_reply()
828 struct request_queue *q = nbd->disk->queue; in recv_work()
846 dev_err(disk_to_dev(nbd->disk), "%s: no io inflight\n", in recv_work()
896 blk_mq_quiesce_queue(nbd->disk->queue); in nbd_clear_que()
898 blk_mq_unquiesce_queue(nbd->disk->queue); in nbd_clear_que()
899 dev_dbg(disk_to_dev(nbd->disk), "queue cleared\n"); in nbd_clear_que()
913 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
936 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
966 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
973 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
1027 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
1083 dev_err(disk_to_dev(nbd->disk), "Unsupported socket: shutdown callout must be supported.\n"); in nbd_get_socket()
1109 blk_mq_freeze_queue(nbd->disk->queue); in nbd_add_socket()
1118 dev_err(disk_to_dev(nbd->disk), in nbd_add_socket()
1149 blk_mq_unfreeze_queue(nbd->disk->queue); in nbd_add_socket()
1154 blk_mq_unfreeze_queue(nbd->disk->queue); in nbd_add_socket()
1231 set_disk_ro(nbd->disk, true); in nbd_parse_flags()
1233 set_disk_ro(nbd->disk, false); in nbd_parse_flags()
1235 blk_queue_flag_set(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_parse_flags()
1238 blk_queue_write_cache(nbd->disk->queue, true, true); in nbd_parse_flags()
1240 blk_queue_write_cache(nbd->disk->queue, true, false); in nbd_parse_flags()
1243 blk_queue_write_cache(nbd->disk->queue, false, false); in nbd_parse_flags()
1264 dev_err(disk_to_dev(nbd->disk), in send_disconnects()
1274 dev_info(disk_to_dev(nbd->disk), "NBD_DISCONNECT\n"); in nbd_disconnect()
1294 invalidate_disk(nbd->disk); in nbd_config_put()
1299 device_remove_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_config_put()
1303 device_remove_file(disk_to_dev(nbd->disk), &backend_attr); in nbd_config_put()
1320 nbd->disk->queue->limits.discard_granularity = 0; in nbd_config_put()
1321 nbd->disk->queue->limits.discard_alignment = 0; in nbd_config_put()
1322 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_config_put()
1323 blk_queue_flag_clear(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_config_put()
1343 dev_err(disk_to_dev(nbd->disk), "server does not support multiple connections per device.\n"); in nbd_start_device()
1352 error = device_create_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_start_device()
1354 dev_err(disk_to_dev(nbd->disk), "device_create_file failed for pid!\n"); in nbd_start_device()
1402 set_bit(GD_NEED_PART_SCAN, &nbd->disk->state); in nbd_start_device_ioctl()
1435 blk_queue_rq_timeout(nbd->disk->queue, timeout * HZ); in nbd_set_cmd_timeout()
1437 blk_queue_rq_timeout(nbd->disk->queue, 30 * HZ); in nbd_set_cmd_timeout()
1577 static void nbd_release(struct gendisk *disk, fmode_t mode) in nbd_release() argument
1579 struct nbd_device *nbd = disk->private_data; in nbd_release()
1582 disk->part0->bd_openers == 0) in nbd_release()
1727 struct gendisk *disk; in nbd_dev_add() local
1765 disk = blk_mq_alloc_disk(&nbd->tag_set, NULL); in nbd_dev_add()
1766 if (IS_ERR(disk)) { in nbd_dev_add()
1767 err = PTR_ERR(disk); in nbd_dev_add()
1770 nbd->disk = disk; in nbd_dev_add()
1776 dev_err(disk_to_dev(nbd->disk), "Could not allocate knbd recv work queue.\n"); in nbd_dev_add()
1784 blk_queue_flag_set(QUEUE_FLAG_NONROT, disk->queue); in nbd_dev_add()
1785 blk_queue_flag_clear(QUEUE_FLAG_ADD_RANDOM, disk->queue); in nbd_dev_add()
1786 disk->queue->limits.discard_granularity = 0; in nbd_dev_add()
1787 disk->queue->limits.discard_alignment = 0; in nbd_dev_add()
1788 blk_queue_max_discard_sectors(disk->queue, 0); in nbd_dev_add()
1789 blk_queue_max_segment_size(disk->queue, UINT_MAX); in nbd_dev_add()
1790 blk_queue_max_segments(disk->queue, USHRT_MAX); in nbd_dev_add()
1791 blk_queue_max_hw_sectors(disk->queue, 65536); in nbd_dev_add()
1792 disk->queue->limits.max_sectors = 256; in nbd_dev_add()
1802 disk->major = NBD_MAJOR; in nbd_dev_add()
1808 disk->first_minor = index << part_shift; in nbd_dev_add()
1809 if (disk->first_minor < index || disk->first_minor > MINORMASK) { in nbd_dev_add()
1814 disk->minors = 1 << part_shift; in nbd_dev_add()
1815 disk->fops = &nbd_fops; in nbd_dev_add()
1816 disk->private_data = nbd; in nbd_dev_add()
1817 sprintf(disk->disk_name, "nbd%d", index); in nbd_dev_add()
1818 err = add_disk(disk); in nbd_dev_add()
1832 blk_cleanup_disk(disk); in nbd_dev_add()
2060 ret = device_create_file(disk_to_dev(nbd->disk), &backend_attr); in nbd_genl_connect()
2062 dev_err(disk_to_dev(nbd->disk), in nbd_genl_connect()