Lines Matching refs:cl
340 cb->cl->tx_cb_queued++; in mei_tx_cb_enqueue()
352 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue()
353 cb->cl->tx_cb_queued--; in mei_tx_cb_dequeue()
366 static void mei_cl_set_read_by_fp(const struct mei_cl *cl, in mei_cl_set_read_by_fp() argument
371 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_set_read_by_fp()
388 static struct mei_cl_cb *mei_io_cb_init(struct mei_cl *cl, in mei_io_cb_init() argument
400 cb->cl = cl; in mei_io_cb_init()
415 const struct mei_cl *cl) in mei_io_list_flush_cl() argument
420 if (cl == cb->cl) { in mei_io_list_flush_cl()
436 const struct mei_cl *cl, in mei_io_tx_list_free_cl() argument
442 if (cl == cb->cl && (!fp || fp == cb->fp)) in mei_io_tx_list_free_cl()
467 static void mei_cl_free_pending(struct mei_cl *cl) in mei_cl_free_pending() argument
471 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_free_pending()
485 struct mei_cl_cb *mei_cl_alloc_cb(struct mei_cl *cl, size_t length, in mei_cl_alloc_cb() argument
491 cb = mei_io_cb_init(cl, fop_type, fp); in mei_cl_alloc_cb()
520 struct mei_cl_cb *mei_cl_enqueue_ctrl_wr_cb(struct mei_cl *cl, size_t length, in mei_cl_enqueue_ctrl_wr_cb() argument
528 length = max_t(size_t, length, mei_cl_mtu(cl)); in mei_cl_enqueue_ctrl_wr_cb()
530 cb = mei_cl_alloc_cb(cl, length, fop_type, fp); in mei_cl_enqueue_ctrl_wr_cb()
534 list_add_tail(&cb->list, &cl->dev->ctrl_wr_list); in mei_cl_enqueue_ctrl_wr_cb()
547 struct mei_cl_cb *mei_cl_read_cb(struct mei_cl *cl, const struct file *fp) in mei_cl_read_cb() argument
552 spin_lock(&cl->rd_completed_lock); in mei_cl_read_cb()
553 list_for_each_entry(cb, &cl->rd_completed, list) in mei_cl_read_cb()
558 spin_unlock(&cl->rd_completed_lock); in mei_cl_read_cb()
570 int mei_cl_flush_queues(struct mei_cl *cl, const struct file *fp) in mei_cl_flush_queues() argument
574 if (WARN_ON(!cl || !cl->dev)) in mei_cl_flush_queues()
577 dev = cl->dev; in mei_cl_flush_queues()
579 cl_dbg(dev, cl, "remove list entry belonging to cl\n"); in mei_cl_flush_queues()
580 mei_io_tx_list_free_cl(&cl->dev->write_list, cl, fp); in mei_cl_flush_queues()
581 mei_io_tx_list_free_cl(&cl->dev->write_waiting_list, cl, fp); in mei_cl_flush_queues()
584 mei_io_list_flush_cl(&cl->dev->ctrl_wr_list, cl); in mei_cl_flush_queues()
585 mei_io_list_flush_cl(&cl->dev->ctrl_rd_list, cl); in mei_cl_flush_queues()
586 mei_cl_free_pending(cl); in mei_cl_flush_queues()
588 spin_lock(&cl->rd_completed_lock); in mei_cl_flush_queues()
589 mei_io_list_free_fp(&cl->rd_completed, fp); in mei_cl_flush_queues()
590 spin_unlock(&cl->rd_completed_lock); in mei_cl_flush_queues()
601 static void mei_cl_init(struct mei_cl *cl, struct mei_device *dev) in mei_cl_init() argument
603 memset(cl, 0, sizeof(*cl)); in mei_cl_init()
604 init_waitqueue_head(&cl->wait); in mei_cl_init()
605 init_waitqueue_head(&cl->rx_wait); in mei_cl_init()
606 init_waitqueue_head(&cl->tx_wait); in mei_cl_init()
607 init_waitqueue_head(&cl->ev_wait); in mei_cl_init()
608 INIT_LIST_HEAD(&cl->vtag_map); in mei_cl_init()
609 spin_lock_init(&cl->rd_completed_lock); in mei_cl_init()
610 INIT_LIST_HEAD(&cl->rd_completed); in mei_cl_init()
611 INIT_LIST_HEAD(&cl->rd_pending); in mei_cl_init()
612 INIT_LIST_HEAD(&cl->link); in mei_cl_init()
613 cl->writing_state = MEI_IDLE; in mei_cl_init()
614 cl->state = MEI_FILE_UNINITIALIZED; in mei_cl_init()
615 cl->dev = dev; in mei_cl_init()
626 struct mei_cl *cl; in mei_cl_allocate() local
628 cl = kmalloc(sizeof(*cl), GFP_KERNEL); in mei_cl_allocate()
629 if (!cl) in mei_cl_allocate()
632 mei_cl_init(cl, dev); in mei_cl_allocate()
634 return cl; in mei_cl_allocate()
646 int mei_cl_link(struct mei_cl *cl) in mei_cl_link() argument
651 if (WARN_ON(!cl || !cl->dev)) in mei_cl_link()
654 dev = cl->dev; in mei_cl_link()
670 cl->host_client_id = id; in mei_cl_link()
671 list_add_tail(&cl->link, &dev->file_list); in mei_cl_link()
675 cl->state = MEI_FILE_INITIALIZING; in mei_cl_link()
677 cl_dbg(dev, cl, "link cl\n"); in mei_cl_link()
688 int mei_cl_unlink(struct mei_cl *cl) in mei_cl_unlink() argument
693 if (!cl) in mei_cl_unlink()
696 if (WARN_ON(!cl->dev)) in mei_cl_unlink()
699 dev = cl->dev; in mei_cl_unlink()
701 cl_dbg(dev, cl, "unlink client"); in mei_cl_unlink()
707 if (cl->host_client_id) in mei_cl_unlink()
708 clear_bit(cl->host_client_id, dev->host_clients_map); in mei_cl_unlink()
710 list_del_init(&cl->link); in mei_cl_unlink()
712 cl->state = MEI_FILE_UNINITIALIZED; in mei_cl_unlink()
713 cl->writing_state = MEI_IDLE; in mei_cl_unlink()
715 WARN_ON(!list_empty(&cl->rd_completed) || in mei_cl_unlink()
716 !list_empty(&cl->rd_pending) || in mei_cl_unlink()
717 !list_empty(&cl->link)); in mei_cl_unlink()
764 static void mei_cl_wake_all(struct mei_cl *cl) in mei_cl_wake_all() argument
766 struct mei_device *dev = cl->dev; in mei_cl_wake_all()
769 if (waitqueue_active(&cl->rx_wait)) { in mei_cl_wake_all()
770 cl_dbg(dev, cl, "Waking up reading client!\n"); in mei_cl_wake_all()
771 wake_up_interruptible(&cl->rx_wait); in mei_cl_wake_all()
774 if (waitqueue_active(&cl->tx_wait)) { in mei_cl_wake_all()
775 cl_dbg(dev, cl, "Waking up writing client!\n"); in mei_cl_wake_all()
776 wake_up_interruptible(&cl->tx_wait); in mei_cl_wake_all()
779 if (waitqueue_active(&cl->ev_wait)) { in mei_cl_wake_all()
780 cl_dbg(dev, cl, "Waking up waiting for event clients!\n"); in mei_cl_wake_all()
781 wake_up_interruptible(&cl->ev_wait); in mei_cl_wake_all()
784 if (waitqueue_active(&cl->wait)) { in mei_cl_wake_all()
785 cl_dbg(dev, cl, "Waking up ctrl write clients!\n"); in mei_cl_wake_all()
786 wake_up(&cl->wait); in mei_cl_wake_all()
796 static void mei_cl_set_disconnected(struct mei_cl *cl) in mei_cl_set_disconnected() argument
798 struct mei_device *dev = cl->dev; in mei_cl_set_disconnected()
800 if (cl->state == MEI_FILE_DISCONNECTED || in mei_cl_set_disconnected()
801 cl->state <= MEI_FILE_INITIALIZING) in mei_cl_set_disconnected()
804 cl->state = MEI_FILE_DISCONNECTED; in mei_cl_set_disconnected()
805 mei_io_tx_list_free_cl(&dev->write_list, cl, NULL); in mei_cl_set_disconnected()
806 mei_io_tx_list_free_cl(&dev->write_waiting_list, cl, NULL); in mei_cl_set_disconnected()
807 mei_io_list_flush_cl(&dev->ctrl_rd_list, cl); in mei_cl_set_disconnected()
808 mei_io_list_flush_cl(&dev->ctrl_wr_list, cl); in mei_cl_set_disconnected()
809 mei_cl_wake_all(cl); in mei_cl_set_disconnected()
810 cl->rx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
811 cl->tx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
812 cl->timer_count = 0; in mei_cl_set_disconnected()
814 if (!cl->me_cl) in mei_cl_set_disconnected()
817 if (!WARN_ON(cl->me_cl->connect_count == 0)) in mei_cl_set_disconnected()
818 cl->me_cl->connect_count--; in mei_cl_set_disconnected()
820 if (cl->me_cl->connect_count == 0) in mei_cl_set_disconnected()
821 cl->me_cl->tx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
823 mei_me_cl_put(cl->me_cl); in mei_cl_set_disconnected()
824 cl->me_cl = NULL; in mei_cl_set_disconnected()
827 static int mei_cl_set_connecting(struct mei_cl *cl, struct mei_me_client *me_cl) in mei_cl_set_connecting() argument
840 cl->me_cl = me_cl; in mei_cl_set_connecting()
841 cl->state = MEI_FILE_CONNECTING; in mei_cl_set_connecting()
842 cl->me_cl->connect_count++; in mei_cl_set_connecting()
855 static int mei_cl_send_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_send_disconnect() argument
860 dev = cl->dev; in mei_cl_send_disconnect()
862 ret = mei_hbm_cl_disconnect_req(dev, cl); in mei_cl_send_disconnect()
863 cl->status = ret; in mei_cl_send_disconnect()
865 cl->state = MEI_FILE_DISCONNECT_REPLY; in mei_cl_send_disconnect()
870 cl->timer_count = MEI_CONNECT_TIMEOUT; in mei_cl_send_disconnect()
886 int mei_cl_irq_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_disconnect() argument
889 struct mei_device *dev = cl->dev; in mei_cl_irq_disconnect()
902 ret = mei_cl_send_disconnect(cl, cb); in mei_cl_irq_disconnect()
917 static int __mei_cl_disconnect(struct mei_cl *cl) in __mei_cl_disconnect() argument
923 dev = cl->dev; in __mei_cl_disconnect()
925 cl->state = MEI_FILE_DISCONNECTING; in __mei_cl_disconnect()
927 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DISCONNECT, NULL); in __mei_cl_disconnect()
934 rets = mei_cl_send_disconnect(cl, cb); in __mei_cl_disconnect()
936 cl_err(dev, cl, "failed to disconnect.\n"); in __mei_cl_disconnect()
942 wait_event_timeout(cl->wait, in __mei_cl_disconnect()
943 cl->state == MEI_FILE_DISCONNECT_REPLY || in __mei_cl_disconnect()
944 cl->state == MEI_FILE_DISCONNECTED, in __mei_cl_disconnect()
948 rets = cl->status; in __mei_cl_disconnect()
949 if (cl->state != MEI_FILE_DISCONNECT_REPLY && in __mei_cl_disconnect()
950 cl->state != MEI_FILE_DISCONNECTED) { in __mei_cl_disconnect()
951 cl_dbg(dev, cl, "timeout on disconnect from FW client.\n"); in __mei_cl_disconnect()
957 mei_cl_set_disconnected(cl); in __mei_cl_disconnect()
959 cl_dbg(dev, cl, "successfully disconnected from FW client.\n"); in __mei_cl_disconnect()
974 int mei_cl_disconnect(struct mei_cl *cl) in mei_cl_disconnect() argument
979 if (WARN_ON(!cl || !cl->dev)) in mei_cl_disconnect()
982 dev = cl->dev; in mei_cl_disconnect()
984 cl_dbg(dev, cl, "disconnecting"); in mei_cl_disconnect()
986 if (!mei_cl_is_connected(cl)) in mei_cl_disconnect()
989 if (mei_cl_is_fixed_address(cl)) { in mei_cl_disconnect()
990 mei_cl_set_disconnected(cl); in mei_cl_disconnect()
996 cl_dbg(dev, cl, "Device is powering down, don't bother with disconnection\n"); in mei_cl_disconnect()
997 mei_cl_set_disconnected(cl); in mei_cl_disconnect()
1004 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_disconnect()
1008 rets = __mei_cl_disconnect(cl); in mei_cl_disconnect()
1010 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_disconnect()
1026 static bool mei_cl_is_other_connecting(struct mei_cl *cl) in mei_cl_is_other_connecting() argument
1031 dev = cl->dev; in mei_cl_is_other_connecting()
1035 mei_cl_me_id(cl) == mei_cl_me_id(cb->cl)) in mei_cl_is_other_connecting()
1050 static int mei_cl_send_connect(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_send_connect() argument
1055 dev = cl->dev; in mei_cl_send_connect()
1057 ret = mei_hbm_cl_connect_req(dev, cl); in mei_cl_send_connect()
1058 cl->status = ret; in mei_cl_send_connect()
1060 cl->state = MEI_FILE_DISCONNECT_REPLY; in mei_cl_send_connect()
1065 cl->timer_count = MEI_CONNECT_TIMEOUT; in mei_cl_send_connect()
1079 int mei_cl_irq_connect(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_connect() argument
1082 struct mei_device *dev = cl->dev; in mei_cl_irq_connect()
1087 if (mei_cl_is_other_connecting(cl)) in mei_cl_irq_connect()
1098 rets = mei_cl_send_connect(cl, cb); in mei_cl_irq_connect()
1116 int mei_cl_connect(struct mei_cl *cl, struct mei_me_client *me_cl, in mei_cl_connect() argument
1123 if (WARN_ON(!cl || !cl->dev || !me_cl)) in mei_cl_connect()
1126 dev = cl->dev; in mei_cl_connect()
1128 rets = mei_cl_set_connecting(cl, me_cl); in mei_cl_connect()
1132 if (mei_cl_is_fixed_address(cl)) { in mei_cl_connect()
1133 cl->state = MEI_FILE_CONNECTED; in mei_cl_connect()
1141 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_connect()
1145 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_CONNECT, fp); in mei_cl_connect()
1152 if (!mei_cl_is_other_connecting(cl) && mei_hbuf_acquire(dev)) { in mei_cl_connect()
1153 rets = mei_cl_send_connect(cl, cb); in mei_cl_connect()
1159 wait_event_timeout(cl->wait, in mei_cl_connect()
1160 (cl->state == MEI_FILE_CONNECTED || in mei_cl_connect()
1161 cl->state == MEI_FILE_DISCONNECTED || in mei_cl_connect()
1162 cl->state == MEI_FILE_DISCONNECT_REQUIRED || in mei_cl_connect()
1163 cl->state == MEI_FILE_DISCONNECT_REPLY), in mei_cl_connect()
1167 if (!mei_cl_is_connected(cl)) { in mei_cl_connect()
1168 if (cl->state == MEI_FILE_DISCONNECT_REQUIRED) { in mei_cl_connect()
1169 mei_io_list_flush_cl(&dev->ctrl_rd_list, cl); in mei_cl_connect()
1170 mei_io_list_flush_cl(&dev->ctrl_wr_list, cl); in mei_cl_connect()
1174 __mei_cl_disconnect(cl); in mei_cl_connect()
1180 if (!cl->status) in mei_cl_connect()
1181 cl->status = -EFAULT; in mei_cl_connect()
1184 rets = cl->status; in mei_cl_connect()
1186 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_connect()
1193 if (!mei_cl_is_connected(cl)) in mei_cl_connect()
1194 mei_cl_set_disconnected(cl); in mei_cl_connect()
1208 struct mei_cl *cl; in mei_cl_alloc_linked() local
1211 cl = mei_cl_allocate(dev); in mei_cl_alloc_linked()
1212 if (!cl) { in mei_cl_alloc_linked()
1217 ret = mei_cl_link(cl); in mei_cl_alloc_linked()
1221 return cl; in mei_cl_alloc_linked()
1223 kfree(cl); in mei_cl_alloc_linked()
1234 static int mei_cl_tx_flow_ctrl_creds(struct mei_cl *cl) in mei_cl_tx_flow_ctrl_creds() argument
1236 if (WARN_ON(!cl || !cl->me_cl)) in mei_cl_tx_flow_ctrl_creds()
1239 if (cl->tx_flow_ctrl_creds > 0) in mei_cl_tx_flow_ctrl_creds()
1242 if (mei_cl_is_fixed_address(cl)) in mei_cl_tx_flow_ctrl_creds()
1245 if (mei_cl_is_single_recv_buf(cl)) { in mei_cl_tx_flow_ctrl_creds()
1246 if (cl->me_cl->tx_flow_ctrl_creds > 0) in mei_cl_tx_flow_ctrl_creds()
1262 static int mei_cl_tx_flow_ctrl_creds_reduce(struct mei_cl *cl) in mei_cl_tx_flow_ctrl_creds_reduce() argument
1264 if (WARN_ON(!cl || !cl->me_cl)) in mei_cl_tx_flow_ctrl_creds_reduce()
1267 if (mei_cl_is_fixed_address(cl)) in mei_cl_tx_flow_ctrl_creds_reduce()
1270 if (mei_cl_is_single_recv_buf(cl)) { in mei_cl_tx_flow_ctrl_creds_reduce()
1271 if (WARN_ON(cl->me_cl->tx_flow_ctrl_creds <= 0)) in mei_cl_tx_flow_ctrl_creds_reduce()
1273 cl->me_cl->tx_flow_ctrl_creds--; in mei_cl_tx_flow_ctrl_creds_reduce()
1275 if (WARN_ON(cl->tx_flow_ctrl_creds <= 0)) in mei_cl_tx_flow_ctrl_creds_reduce()
1277 cl->tx_flow_ctrl_creds--; in mei_cl_tx_flow_ctrl_creds_reduce()
1317 const struct file *mei_cl_fp_by_vtag(const struct mei_cl *cl, u8 vtag) in mei_cl_fp_by_vtag() argument
1321 list_for_each_entry(vtag_l, &cl->vtag_map, list) in mei_cl_fp_by_vtag()
1323 if ((cl->cldev && mei_cldev_enabled(cl->cldev)) || in mei_cl_fp_by_vtag()
1336 static void mei_cl_reset_read_by_vtag(const struct mei_cl *cl, u8 vtag) in mei_cl_reset_read_by_vtag() argument
1340 list_for_each_entry(vtag_l, &cl->vtag_map, list) { in mei_cl_reset_read_by_vtag()
1354 static void mei_cl_read_vtag_add_fc(struct mei_cl *cl) in mei_cl_read_vtag_add_fc() argument
1358 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_read_vtag_add_fc()
1360 if (mei_cl_enqueue_ctrl_wr_cb(cl, in mei_cl_read_vtag_add_fc()
1361 mei_cl_mtu(cl), in mei_cl_read_vtag_add_fc()
1364 cl->rx_flow_ctrl_creds++; in mei_cl_read_vtag_add_fc()
1379 int mei_cl_vt_support_check(const struct mei_cl *cl) in mei_cl_vt_support_check() argument
1381 struct mei_device *dev = cl->dev; in mei_cl_vt_support_check()
1386 if (!cl->me_cl) in mei_cl_vt_support_check()
1389 return cl->me_cl->props.vt_supported ? 0 : -EOPNOTSUPP; in mei_cl_vt_support_check()
1400 void mei_cl_add_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_add_rd_completed() argument
1404 if (!mei_cl_vt_support_check(cl)) { in mei_cl_add_rd_completed()
1405 fp = mei_cl_fp_by_vtag(cl, cb->vtag); in mei_cl_add_rd_completed()
1412 mei_cl_reset_read_by_vtag(cl, cb->vtag); in mei_cl_add_rd_completed()
1413 mei_cl_read_vtag_add_fc(cl); in mei_cl_add_rd_completed()
1416 spin_lock(&cl->rd_completed_lock); in mei_cl_add_rd_completed()
1417 list_add_tail(&cb->list, &cl->rd_completed); in mei_cl_add_rd_completed()
1418 spin_unlock(&cl->rd_completed_lock); in mei_cl_add_rd_completed()
1428 void mei_cl_del_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_del_rd_completed() argument
1430 spin_lock(&cl->rd_completed_lock); in mei_cl_del_rd_completed()
1432 spin_unlock(&cl->rd_completed_lock); in mei_cl_del_rd_completed()
1474 int mei_cl_irq_notify(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_notify() argument
1477 struct mei_device *dev = cl->dev; in mei_cl_irq_notify()
1492 ret = mei_hbm_cl_notify_req(dev, cl, request); in mei_cl_irq_notify()
1494 cl->status = ret; in mei_cl_irq_notify()
1514 int mei_cl_notify_request(struct mei_cl *cl, in mei_cl_notify_request() argument
1522 if (WARN_ON(!cl || !cl->dev)) in mei_cl_notify_request()
1525 dev = cl->dev; in mei_cl_notify_request()
1528 cl_dbg(dev, cl, "notifications not supported\n"); in mei_cl_notify_request()
1532 if (!mei_cl_is_connected(cl)) in mei_cl_notify_request()
1538 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_notify_request()
1543 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, fop_type, fp); in mei_cl_notify_request()
1550 if (mei_hbm_cl_notify_req(dev, cl, request)) { in mei_cl_notify_request()
1558 wait_event_timeout(cl->wait, in mei_cl_notify_request()
1559 cl->notify_en == request || in mei_cl_notify_request()
1560 cl->status || in mei_cl_notify_request()
1561 !mei_cl_is_connected(cl), in mei_cl_notify_request()
1565 if (cl->notify_en != request && !cl->status) in mei_cl_notify_request()
1566 cl->status = -EFAULT; in mei_cl_notify_request()
1568 rets = cl->status; in mei_cl_notify_request()
1571 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_notify_request()
1586 void mei_cl_notify(struct mei_cl *cl) in mei_cl_notify() argument
1590 if (!cl || !cl->dev) in mei_cl_notify()
1593 dev = cl->dev; in mei_cl_notify()
1595 if (!cl->notify_en) in mei_cl_notify()
1598 cl_dbg(dev, cl, "notify event"); in mei_cl_notify()
1599 cl->notify_ev = true; in mei_cl_notify()
1600 if (!mei_cl_bus_notify_event(cl)) in mei_cl_notify()
1601 wake_up_interruptible(&cl->ev_wait); in mei_cl_notify()
1603 if (cl->ev_async) in mei_cl_notify()
1604 kill_fasync(&cl->ev_async, SIGIO, POLL_PRI); in mei_cl_notify()
1619 int mei_cl_notify_get(struct mei_cl *cl, bool block, bool *notify_ev) in mei_cl_notify_get() argument
1626 if (WARN_ON(!cl || !cl->dev)) in mei_cl_notify_get()
1629 dev = cl->dev; in mei_cl_notify_get()
1632 cl_dbg(dev, cl, "notifications not supported\n"); in mei_cl_notify_get()
1636 if (!mei_cl_is_connected(cl)) in mei_cl_notify_get()
1639 if (cl->notify_ev) in mei_cl_notify_get()
1646 rets = wait_event_interruptible(cl->ev_wait, cl->notify_ev); in mei_cl_notify_get()
1653 *notify_ev = cl->notify_ev; in mei_cl_notify_get()
1654 cl->notify_ev = false; in mei_cl_notify_get()
1667 int mei_cl_read_start(struct mei_cl *cl, size_t length, const struct file *fp) in mei_cl_read_start() argument
1673 if (WARN_ON(!cl || !cl->dev)) in mei_cl_read_start()
1676 dev = cl->dev; in mei_cl_read_start()
1678 if (!mei_cl_is_connected(cl)) in mei_cl_read_start()
1681 if (!mei_me_cl_is_active(cl->me_cl)) { in mei_cl_read_start()
1682 cl_err(dev, cl, "no such me client\n"); in mei_cl_read_start()
1686 if (mei_cl_is_fixed_address(cl)) in mei_cl_read_start()
1690 if (cl->rx_flow_ctrl_creds) { in mei_cl_read_start()
1691 mei_cl_set_read_by_fp(cl, fp); in mei_cl_read_start()
1695 cb = mei_cl_enqueue_ctrl_wr_cb(cl, length, MEI_FOP_READ, fp); in mei_cl_read_start()
1699 mei_cl_set_read_by_fp(cl, fp); in mei_cl_read_start()
1704 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_read_start()
1710 rets = mei_hbm_cl_flow_control_req(dev, cl); in mei_cl_read_start()
1714 list_move_tail(&cb->list, &cl->rd_pending); in mei_cl_read_start()
1716 cl->rx_flow_ctrl_creds++; in mei_cl_read_start()
1719 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_read_start()
1776 mei_hdr->host_addr = mei_cl_host_addr(cb->cl); in mei_msg_hdr_init()
1777 mei_hdr->me_addr = mei_cl_me_id(cb->cl); in mei_msg_hdr_init()
1804 int mei_cl_irq_write(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_write() argument
1821 if (WARN_ON(!cl || !cl->dev)) in mei_cl_irq_write()
1824 dev = cl->dev; in mei_cl_irq_write()
1830 rets = first_chunk ? mei_cl_tx_flow_ctrl_creds(cl) : 1; in mei_cl_irq_write()
1835 cl_dbg(dev, cl, "No flow control credentials: not sending.\n"); in mei_cl_irq_write()
1858 cl_dbg(dev, cl, "Extended Header %d vtag = %d\n", in mei_cl_irq_write()
1896 cl->status = 0; in mei_cl_irq_write()
1897 cl->writing_state = MEI_WRITING; in mei_cl_irq_write()
1901 if (mei_cl_tx_flow_ctrl_creds_reduce(cl)) { in mei_cl_irq_write()
1915 cl->status = rets; in mei_cl_irq_write()
1929 ssize_t mei_cl_write(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_write() argument
1945 if (WARN_ON(!cl || !cl->dev)) in mei_cl_write()
1951 dev = cl->dev; in mei_cl_write()
1956 cl_dbg(dev, cl, "buf_len=%zd\n", buf_len); in mei_cl_write()
1964 cl_err(dev, cl, "rpm: get failed %zd\n", rets); in mei_cl_write()
1969 cl->writing_state = MEI_IDLE; in mei_cl_write()
1972 rets = mei_cl_tx_flow_ctrl_creds(cl); in mei_cl_write()
1983 cl_dbg(dev, cl, "Extended Header %d vtag = %d\n", in mei_cl_write()
1989 cl_dbg(dev, cl, "No flow control credentials: not sending.\n"); in mei_cl_write()
1995 cl_dbg(dev, cl, "Cannot acquire the host buffer: not sending.\n"); in mei_cl_write()
2037 rets = mei_cl_tx_flow_ctrl_creds_reduce(cl); in mei_cl_write()
2041 cl->writing_state = MEI_WRITING; in mei_cl_write()
2053 if (blocking && cl->writing_state != MEI_WRITE_COMPLETE) { in mei_cl_write()
2056 rets = wait_event_interruptible(cl->tx_wait, in mei_cl_write()
2057 cl->writing_state == MEI_WRITE_COMPLETE || in mei_cl_write()
2058 (!mei_cl_is_connected(cl))); in mei_cl_write()
2066 if (cl->writing_state != MEI_WRITE_COMPLETE) { in mei_cl_write()
2074 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_write()
2091 void mei_cl_complete(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_complete() argument
2093 struct mei_device *dev = cl->dev; in mei_cl_complete()
2098 cl->writing_state = MEI_WRITE_COMPLETE; in mei_cl_complete()
2099 if (waitqueue_active(&cl->tx_wait)) { in mei_cl_complete()
2100 wake_up_interruptible(&cl->tx_wait); in mei_cl_complete()
2108 mei_cl_add_rd_completed(cl, cb); in mei_cl_complete()
2109 if (!mei_cl_is_fixed_address(cl) && in mei_cl_complete()
2110 !WARN_ON(!cl->rx_flow_ctrl_creds)) in mei_cl_complete()
2111 cl->rx_flow_ctrl_creds--; in mei_cl_complete()
2112 if (!mei_cl_bus_rx_event(cl)) in mei_cl_complete()
2113 wake_up_interruptible(&cl->rx_wait); in mei_cl_complete()
2122 if (waitqueue_active(&cl->wait)) in mei_cl_complete()
2123 wake_up(&cl->wait); in mei_cl_complete()
2128 mei_cl_set_disconnected(cl); in mei_cl_complete()
2143 struct mei_cl *cl; in mei_cl_all_disconnect() local
2145 list_for_each_entry(cl, &dev->file_list, link) in mei_cl_all_disconnect()
2146 mei_cl_set_disconnected(cl); in mei_cl_all_disconnect()
2151 struct mei_cl *cl; in mei_cl_dma_map_find() local
2153 list_for_each_entry(cl, &dev->file_list, link) in mei_cl_dma_map_find()
2154 if (cl->dma.buffer_id == buffer_id) in mei_cl_dma_map_find()
2155 return cl; in mei_cl_dma_map_find()
2168 int mei_cl_irq_dma_map(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_dma_map() argument
2171 struct mei_device *dev = cl->dev; in mei_cl_irq_dma_map()
2184 ret = mei_hbm_cl_dma_map_req(dev, cl); in mei_cl_irq_dma_map()
2186 cl->status = ret; in mei_cl_irq_dma_map()
2204 int mei_cl_irq_dma_unmap(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_dma_unmap() argument
2207 struct mei_device *dev = cl->dev; in mei_cl_irq_dma_unmap()
2220 ret = mei_hbm_cl_dma_unmap_req(dev, cl); in mei_cl_irq_dma_unmap()
2222 cl->status = ret; in mei_cl_irq_dma_unmap()
2231 static int mei_cl_dma_alloc(struct mei_cl *cl, u8 buf_id, size_t size) in mei_cl_dma_alloc() argument
2233 cl->dma.vaddr = dmam_alloc_coherent(cl->dev->dev, size, in mei_cl_dma_alloc()
2234 &cl->dma.daddr, GFP_KERNEL); in mei_cl_dma_alloc()
2235 if (!cl->dma.vaddr) in mei_cl_dma_alloc()
2238 cl->dma.buffer_id = buf_id; in mei_cl_dma_alloc()
2239 cl->dma.size = size; in mei_cl_dma_alloc()
2244 static void mei_cl_dma_free(struct mei_cl *cl) in mei_cl_dma_free() argument
2246 cl->dma.buffer_id = 0; in mei_cl_dma_free()
2247 dmam_free_coherent(cl->dev->dev, in mei_cl_dma_free()
2248 cl->dma.size, cl->dma.vaddr, cl->dma.daddr); in mei_cl_dma_free()
2249 cl->dma.size = 0; in mei_cl_dma_free()
2250 cl->dma.vaddr = NULL; in mei_cl_dma_free()
2251 cl->dma.daddr = 0; in mei_cl_dma_free()
2271 int mei_cl_dma_alloc_and_map(struct mei_cl *cl, const struct file *fp, in mei_cl_dma_alloc_and_map() argument
2278 if (WARN_ON(!cl || !cl->dev)) in mei_cl_dma_alloc_and_map()
2281 dev = cl->dev; in mei_cl_dma_alloc_and_map()
2284 cl_dbg(dev, cl, "client dma is not supported\n"); in mei_cl_dma_alloc_and_map()
2291 if (mei_cl_is_connected(cl)) in mei_cl_dma_alloc_and_map()
2294 if (cl->dma_mapped) in mei_cl_dma_alloc_and_map()
2298 cl_dbg(dev, cl, "client dma with id %d is already allocated\n", in mei_cl_dma_alloc_and_map()
2299 cl->dma.buffer_id); in mei_cl_dma_alloc_and_map()
2306 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_dma_alloc_and_map()
2310 rets = mei_cl_dma_alloc(cl, buffer_id, size); in mei_cl_dma_alloc_and_map()
2316 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DMA_MAP, fp); in mei_cl_dma_alloc_and_map()
2323 if (mei_hbm_cl_dma_map_req(dev, cl)) { in mei_cl_dma_alloc_and_map()
2331 wait_event_timeout(cl->wait, in mei_cl_dma_alloc_and_map()
2332 cl->dma_mapped || cl->status, in mei_cl_dma_alloc_and_map()
2336 if (!cl->dma_mapped && !cl->status) in mei_cl_dma_alloc_and_map()
2337 cl->status = -EFAULT; in mei_cl_dma_alloc_and_map()
2339 rets = cl->status; in mei_cl_dma_alloc_and_map()
2343 mei_cl_dma_free(cl); in mei_cl_dma_alloc_and_map()
2345 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_dma_alloc_and_map()
2363 int mei_cl_dma_unmap(struct mei_cl *cl, const struct file *fp) in mei_cl_dma_unmap() argument
2369 if (WARN_ON(!cl || !cl->dev)) in mei_cl_dma_unmap()
2372 dev = cl->dev; in mei_cl_dma_unmap()
2375 cl_dbg(dev, cl, "client dma is not supported\n"); in mei_cl_dma_unmap()
2380 if (mei_cl_is_connected(cl)) in mei_cl_dma_unmap()
2383 if (!cl->dma_mapped) in mei_cl_dma_unmap()
2389 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_dma_unmap()
2393 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DMA_UNMAP, fp); in mei_cl_dma_unmap()
2400 if (mei_hbm_cl_dma_unmap_req(dev, cl)) { in mei_cl_dma_unmap()
2408 wait_event_timeout(cl->wait, in mei_cl_dma_unmap()
2409 !cl->dma_mapped || cl->status, in mei_cl_dma_unmap()
2413 if (cl->dma_mapped && !cl->status) in mei_cl_dma_unmap()
2414 cl->status = -EFAULT; in mei_cl_dma_unmap()
2416 rets = cl->status; in mei_cl_dma_unmap()
2419 mei_cl_dma_free(cl); in mei_cl_dma_unmap()
2421 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_dma_unmap()