blob: 28859ae3ee6eb99d6d6175f99c567745dd675d1c [file] [log] [blame]
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +01001#include <linux/kernel.h>
2#include <linux/errno.h>
3#include <linux/file.h>
4#include <linux/slab.h>
5#include <linux/net.h>
6#include <linux/io_uring.h>
7
8#include "io_uring.h"
9#include "notif.h"
Pavel Begunkov68ef5572022-07-12 21:52:41 +010010#include "rsrc.h"
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +010011
Pavel Begunkov6fe42202024-04-19 12:08:42 +010012static const struct ubuf_info_ops io_ubuf_ops;
13
Pavel Begunkov5a569462024-04-19 12:08:41 +010014static void io_notif_tw_complete(struct io_kiocb *notif, struct io_tw_state *ts)
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +010015{
Pavel Begunkov14b146b2022-07-27 10:30:41 +010016 struct io_notif_data *nd = io_notif_to_data(notif);
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +010017
Pavel Begunkov6fe42202024-04-19 12:08:42 +010018 do {
19 notif = cmd_to_io_kiocb(nd);
Pavel Begunkov42385b02022-11-04 10:59:46 +000020
Pavel Begunkov6fe42202024-04-19 12:08:42 +010021 lockdep_assert(refcount_read(&nd->uarg.refcnt) == 0);
22
23 if (unlikely(nd->zc_report) && (nd->zc_copied || !nd->zc_used))
24 notif->cqe.res |= IORING_NOTIF_USAGE_ZC_COPIED;
25
26 if (nd->account_pages && notif->ctx->user) {
27 __io_unaccount_mem(notif->ctx->user, nd->account_pages);
28 nd->account_pages = 0;
29 }
30
31 nd = nd->next;
32 io_req_task_complete(notif, ts);
33 } while (nd);
Pavel Begunkov40725d12022-11-04 10:59:45 +000034}
35
Pavel Begunkov5a569462024-04-19 12:08:41 +010036void io_tx_ubuf_complete(struct sk_buff *skb, struct ubuf_info *uarg,
37 bool success)
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +010038{
Pavel Begunkov14b146b2022-07-27 10:30:41 +010039 struct io_notif_data *nd = container_of(uarg, struct io_notif_data, uarg);
40 struct io_kiocb *notif = cmd_to_io_kiocb(nd);
Pavel Begunkov19352a12024-04-30 16:42:31 +010041 unsigned tw_flags;
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +010042
Stefan Metzmachere307e662022-10-27 20:34:45 +020043 if (nd->zc_report) {
44 if (success && !nd->zc_used && skb)
45 WRITE_ONCE(nd->zc_used, true);
46 else if (!success && !nd->zc_copied)
47 WRITE_ONCE(nd->zc_copied, true);
48 }
Pavel Begunkov99863292024-04-08 00:54:55 +010049
Pavel Begunkov7e58d0a2024-04-15 13:50:11 +010050 if (!refcount_dec_and_test(&uarg->refcnt))
51 return;
52
Pavel Begunkov6fe42202024-04-19 12:08:42 +010053 if (nd->head != nd) {
54 io_tx_ubuf_complete(skb, &nd->head->uarg, success);
55 return;
56 }
Pavel Begunkov19352a12024-04-30 16:42:31 +010057
58 tw_flags = nd->next ? 0 : IOU_F_TWQ_LAZY_WAKE;
Pavel Begunkov7e58d0a2024-04-15 13:50:11 +010059 notif->io_task_work.func = io_notif_tw_complete;
Pavel Begunkov19352a12024-04-30 16:42:31 +010060 __io_req_task_work_add(notif, tw_flags);
Pavel Begunkoveb4a2992022-07-12 21:52:39 +010061}
62
Pavel Begunkov6fe42202024-04-19 12:08:42 +010063static int io_link_skb(struct sk_buff *skb, struct ubuf_info *uarg)
64{
65 struct io_notif_data *nd, *prev_nd;
66 struct io_kiocb *prev_notif, *notif;
67 struct ubuf_info *prev_uarg = skb_zcopy(skb);
68
69 nd = container_of(uarg, struct io_notif_data, uarg);
70 notif = cmd_to_io_kiocb(nd);
71
72 if (!prev_uarg) {
73 net_zcopy_get(&nd->uarg);
74 skb_zcopy_init(skb, &nd->uarg);
75 return 0;
76 }
77 /* handle it separately as we can't link a notif to itself */
78 if (unlikely(prev_uarg == &nd->uarg))
79 return 0;
80 /* we can't join two links together, just request a fresh skb */
81 if (unlikely(nd->head != nd || nd->next))
82 return -EEXIST;
83 /* don't mix zc providers */
84 if (unlikely(prev_uarg->ops != &io_ubuf_ops))
85 return -EEXIST;
86
87 prev_nd = container_of(prev_uarg, struct io_notif_data, uarg);
88 prev_notif = cmd_to_io_kiocb(nd);
89
90 /* make sure all noifications can be finished in the same task_work */
91 if (unlikely(notif->ctx != prev_notif->ctx ||
92 notif->task != prev_notif->task))
93 return -EEXIST;
94
95 nd->head = prev_nd->head;
96 nd->next = prev_nd->next;
97 prev_nd->next = nd;
98 net_zcopy_get(&nd->head->uarg);
99 return 0;
100}
101
Pavel Begunkov7ab4f162024-04-19 12:08:39 +0100102static const struct ubuf_info_ops io_ubuf_ops = {
103 .complete = io_tx_ubuf_complete,
Pavel Begunkov6fe42202024-04-19 12:08:42 +0100104 .link_skb = io_link_skb,
Pavel Begunkov7ab4f162024-04-19 12:08:39 +0100105};
106
Pavel Begunkovb48c3122022-09-01 11:54:04 +0100107struct io_kiocb *io_alloc_notif(struct io_ring_ctx *ctx)
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +0100108 __must_hold(&ctx->uring_lock)
109{
Pavel Begunkov14b146b2022-07-27 10:30:41 +0100110 struct io_kiocb *notif;
111 struct io_notif_data *nd;
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +0100112
Pavel Begunkovc8576f32023-01-23 14:37:16 +0000113 if (unlikely(!io_alloc_req(ctx, &notif)))
Pavel Begunkov14b146b2022-07-27 10:30:41 +0100114 return NULL;
Pavel Begunkov14b146b2022-07-27 10:30:41 +0100115 notif->opcode = IORING_OP_NOP;
116 notif->flags = 0;
117 notif->file = NULL;
118 notif->task = current;
119 io_get_task_refs(1);
120 notif->rsrc_node = NULL;
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +0100121
Pavel Begunkov14b146b2022-07-27 10:30:41 +0100122 nd = io_notif_to_data(notif);
Pavel Begunkov99863292024-04-08 00:54:55 +0100123 nd->zc_report = false;
124 nd->account_pages = 0;
Pavel Begunkov6fe42202024-04-19 12:08:42 +0100125 nd->next = NULL;
126 nd->head = nd;
127
Pavel Begunkov519760d2023-04-15 14:20:08 +0100128 nd->uarg.flags = IO_NOTIF_UBUF_FLAGS;
Pavel Begunkov7ab4f162024-04-19 12:08:39 +0100129 nd->uarg.ops = &io_ubuf_ops;
Pavel Begunkov14b146b2022-07-27 10:30:41 +0100130 refcount_set(&nd->uarg.refcnt, 1);
Pavel Begunkoveb42ceb2022-07-12 21:52:38 +0100131 return notif;
132}