hc
2024-01-03 2f7c68cb55ecb7331f2381deb497c27155f32faf
kernel/drivers/net/ethernet/mellanox/mlx5/core/wq.c
....@@ -34,53 +34,19 @@
3434 #include "wq.h"
3535 #include "mlx5_core.h"
3636
37
-u32 mlx5_wq_cyc_get_size(struct mlx5_wq_cyc *wq)
37
+static u32 wq_get_byte_sz(u8 log_sz, u8 log_stride)
3838 {
39
- return (u32)wq->fbc.sz_m1 + 1;
40
-}
41
-
42
-u32 mlx5_cqwq_get_size(struct mlx5_cqwq *wq)
43
-{
44
- return wq->fbc.sz_m1 + 1;
45
-}
46
-
47
-u32 mlx5_wq_ll_get_size(struct mlx5_wq_ll *wq)
48
-{
49
- return (u32)wq->fbc.sz_m1 + 1;
50
-}
51
-
52
-static u32 mlx5_wq_cyc_get_byte_size(struct mlx5_wq_cyc *wq)
53
-{
54
- return mlx5_wq_cyc_get_size(wq) << wq->fbc.log_stride;
55
-}
56
-
57
-static u32 mlx5_wq_qp_get_byte_size(struct mlx5_wq_qp *wq)
58
-{
59
- return mlx5_wq_cyc_get_byte_size(&wq->rq) +
60
- mlx5_wq_cyc_get_byte_size(&wq->sq);
61
-}
62
-
63
-static u32 mlx5_cqwq_get_byte_size(struct mlx5_cqwq *wq)
64
-{
65
- return mlx5_cqwq_get_size(wq) << wq->fbc.log_stride;
66
-}
67
-
68
-static u32 mlx5_wq_ll_get_byte_size(struct mlx5_wq_ll *wq)
69
-{
70
- return mlx5_wq_ll_get_size(wq) << wq->fbc.log_stride;
39
+ return ((u32)1 << log_sz) << log_stride;
7140 }
7241
7342 int mlx5_wq_cyc_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param,
7443 void *wqc, struct mlx5_wq_cyc *wq,
7544 struct mlx5_wq_ctrl *wq_ctrl)
7645 {
46
+ u8 log_wq_stride = MLX5_GET(wq, wqc, log_wq_stride);
47
+ u8 log_wq_sz = MLX5_GET(wq, wqc, log_wq_sz);
7748 struct mlx5_frag_buf_ctrl *fbc = &wq->fbc;
7849 int err;
79
-
80
- mlx5_fill_fbc(MLX5_GET(wq, wqc, log_wq_stride),
81
- MLX5_GET(wq, wqc, log_wq_sz),
82
- fbc);
83
- wq->sz = wq->fbc.sz_m1 + 1;
8450
8551 err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node);
8652 if (err) {
....@@ -88,15 +54,17 @@
8854 return err;
8955 }
9056
91
- err = mlx5_frag_buf_alloc_node(mdev, mlx5_wq_cyc_get_byte_size(wq),
57
+ wq->db = wq_ctrl->db.db;
58
+
59
+ err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride),
9260 &wq_ctrl->buf, param->buf_numa_node);
9361 if (err) {
9462 mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err);
9563 goto err_db_free;
9664 }
9765
98
- fbc->frag_buf = wq_ctrl->buf;
99
- wq->db = wq_ctrl->db.db;
66
+ mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc);
67
+ wq->sz = mlx5_wq_cyc_get_size(wq);
10068
10169 wq_ctrl->mdev = mdev;
10270
....@@ -108,46 +76,44 @@
10876 return err;
10977 }
11078
111
-static void mlx5_qp_set_frag_buf(struct mlx5_frag_buf *buf,
112
- struct mlx5_wq_qp *qp)
79
+void mlx5_wq_cyc_wqe_dump(struct mlx5_wq_cyc *wq, u16 ix, u8 nstrides)
11380 {
114
- struct mlx5_frag_buf_ctrl *sq_fbc;
115
- struct mlx5_frag_buf *rqb, *sqb;
81
+ size_t len;
82
+ void *wqe;
11683
117
- rqb = &qp->rq.fbc.frag_buf;
118
- *rqb = *buf;
119
- rqb->size = mlx5_wq_cyc_get_byte_size(&qp->rq);
120
- rqb->npages = DIV_ROUND_UP(rqb->size, PAGE_SIZE);
84
+ if (!net_ratelimit())
85
+ return;
12186
122
- sq_fbc = &qp->sq.fbc;
123
- sqb = &sq_fbc->frag_buf;
124
- *sqb = *buf;
125
- sqb->size = mlx5_wq_cyc_get_byte_size(&qp->sq);
126
- sqb->npages = DIV_ROUND_UP(sqb->size, PAGE_SIZE);
127
- sqb->frags += rqb->npages; /* first part is for the rq */
128
- if (sq_fbc->strides_offset)
129
- sqb->frags--;
87
+ nstrides = max_t(u8, nstrides, 1);
88
+
89
+ len = nstrides << wq->fbc.log_stride;
90
+ wqe = mlx5_wq_cyc_get_wqe(wq, ix);
91
+
92
+ pr_info("WQE DUMP: WQ size %d WQ cur size %d, WQE index 0x%x, len: %zu\n",
93
+ mlx5_wq_cyc_get_size(wq), wq->cur_sz, ix, len);
94
+ print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, wqe, len, false);
95
+}
96
+
97
+void mlx5_wq_cyc_reset(struct mlx5_wq_cyc *wq)
98
+{
99
+ wq->wqe_ctr = 0;
100
+ wq->cur_sz = 0;
101
+ mlx5_wq_cyc_update_db_record(wq);
130102 }
131103
132104 int mlx5_wq_qp_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param,
133105 void *qpc, struct mlx5_wq_qp *wq,
134106 struct mlx5_wq_ctrl *wq_ctrl)
135107 {
136
- u16 sq_strides_offset;
137
- u32 rq_pg_remainder;
108
+ u8 log_rq_stride = MLX5_GET(qpc, qpc, log_rq_stride) + 4;
109
+ u8 log_rq_sz = MLX5_GET(qpc, qpc, log_rq_size);
110
+ u8 log_sq_stride = ilog2(MLX5_SEND_WQE_BB);
111
+ u8 log_sq_sz = MLX5_GET(qpc, qpc, log_sq_size);
112
+
113
+ u32 rq_byte_size;
138114 int err;
139115
140
- mlx5_fill_fbc(MLX5_GET(qpc, qpc, log_rq_stride) + 4,
141
- MLX5_GET(qpc, qpc, log_rq_size),
142
- &wq->rq.fbc);
143116
144
- rq_pg_remainder = mlx5_wq_cyc_get_byte_size(&wq->rq) % PAGE_SIZE;
145
- sq_strides_offset = rq_pg_remainder / MLX5_SEND_WQE_BB;
146
-
147
- mlx5_fill_fbc_offset(ilog2(MLX5_SEND_WQE_BB),
148
- MLX5_GET(qpc, qpc, log_sq_size),
149
- sq_strides_offset,
150
- &wq->sq.fbc);
151117
152118 err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node);
153119 if (err) {
....@@ -155,14 +121,32 @@
155121 return err;
156122 }
157123
158
- err = mlx5_frag_buf_alloc_node(mdev, mlx5_wq_qp_get_byte_size(wq),
124
+ err = mlx5_frag_buf_alloc_node(mdev,
125
+ wq_get_byte_sz(log_rq_sz, log_rq_stride) +
126
+ wq_get_byte_sz(log_sq_sz, log_sq_stride),
159127 &wq_ctrl->buf, param->buf_numa_node);
160128 if (err) {
161129 mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err);
162130 goto err_db_free;
163131 }
164132
165
- mlx5_qp_set_frag_buf(&wq_ctrl->buf, wq);
133
+ mlx5_init_fbc(wq_ctrl->buf.frags, log_rq_stride, log_rq_sz, &wq->rq.fbc);
134
+
135
+ rq_byte_size = wq_get_byte_sz(log_rq_sz, log_rq_stride);
136
+
137
+ if (rq_byte_size < PAGE_SIZE) {
138
+ /* SQ starts within the same page of the RQ */
139
+ u16 sq_strides_offset = rq_byte_size / MLX5_SEND_WQE_BB;
140
+
141
+ mlx5_init_fbc_offset(wq_ctrl->buf.frags,
142
+ log_sq_stride, log_sq_sz, sq_strides_offset,
143
+ &wq->sq.fbc);
144
+ } else {
145
+ u16 rq_npages = rq_byte_size >> PAGE_SHIFT;
146
+
147
+ mlx5_init_fbc(wq_ctrl->buf.frags + rq_npages,
148
+ log_sq_stride, log_sq_sz, &wq->sq.fbc);
149
+ }
166150
167151 wq->rq.db = &wq_ctrl->db.db[MLX5_RCV_DBR];
168152 wq->sq.db = &wq_ctrl->db.db[MLX5_SND_DBR];
....@@ -181,9 +165,10 @@
181165 void *cqc, struct mlx5_cqwq *wq,
182166 struct mlx5_wq_ctrl *wq_ctrl)
183167 {
168
+ /* CQE_STRIDE_128 and CQE_STRIDE_128_PAD both mean 128B stride */
169
+ u8 log_wq_stride = MLX5_GET(cqc, cqc, cqe_sz) == CQE_STRIDE_64 ? 6 : 7;
170
+ u8 log_wq_sz = MLX5_GET(cqc, cqc, log_cq_size);
184171 int err;
185
-
186
- mlx5_core_init_cq_frag_buf(&wq->fbc, cqc);
187172
188173 err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node);
189174 if (err) {
....@@ -191,7 +176,9 @@
191176 return err;
192177 }
193178
194
- err = mlx5_frag_buf_alloc_node(mdev, mlx5_cqwq_get_byte_size(wq),
179
+ wq->db = wq_ctrl->db.db;
180
+
181
+ err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride),
195182 &wq_ctrl->buf,
196183 param->buf_numa_node);
197184 if (err) {
....@@ -200,8 +187,7 @@
200187 goto err_db_free;
201188 }
202189
203
- wq->fbc.frag_buf = wq_ctrl->buf;
204
- wq->db = wq_ctrl->db.db;
190
+ mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, &wq->fbc);
205191
206192 wq_ctrl->mdev = mdev;
207193
....@@ -213,18 +199,27 @@
213199 return err;
214200 }
215201
202
+static void mlx5_wq_ll_init_list(struct mlx5_wq_ll *wq)
203
+{
204
+ struct mlx5_wqe_srq_next_seg *next_seg;
205
+ int i;
206
+
207
+ for (i = 0; i < wq->fbc.sz_m1; i++) {
208
+ next_seg = mlx5_wq_ll_get_wqe(wq, i);
209
+ next_seg->next_wqe_index = cpu_to_be16(i + 1);
210
+ }
211
+ next_seg = mlx5_wq_ll_get_wqe(wq, i);
212
+ wq->tail_next = &next_seg->next_wqe_index;
213
+}
214
+
216215 int mlx5_wq_ll_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param,
217216 void *wqc, struct mlx5_wq_ll *wq,
218217 struct mlx5_wq_ctrl *wq_ctrl)
219218 {
219
+ u8 log_wq_stride = MLX5_GET(wq, wqc, log_wq_stride);
220
+ u8 log_wq_sz = MLX5_GET(wq, wqc, log_wq_sz);
220221 struct mlx5_frag_buf_ctrl *fbc = &wq->fbc;
221
- struct mlx5_wqe_srq_next_seg *next_seg;
222222 int err;
223
- int i;
224
-
225
- mlx5_fill_fbc(MLX5_GET(wq, wqc, log_wq_stride),
226
- MLX5_GET(wq, wqc, log_wq_sz),
227
- fbc);
228223
229224 err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node);
230225 if (err) {
....@@ -232,23 +227,18 @@
232227 return err;
233228 }
234229
235
- err = mlx5_frag_buf_alloc_node(mdev, mlx5_wq_ll_get_byte_size(wq),
230
+ wq->db = wq_ctrl->db.db;
231
+
232
+ err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride),
236233 &wq_ctrl->buf, param->buf_numa_node);
237234 if (err) {
238235 mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err);
239236 goto err_db_free;
240237 }
241238
242
- wq->fbc.frag_buf = wq_ctrl->buf;
243
- wq->db = wq_ctrl->db.db;
239
+ mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc);
244240
245
- for (i = 0; i < fbc->sz_m1; i++) {
246
- next_seg = mlx5_wq_ll_get_wqe(wq, i);
247
- next_seg->next_wqe_index = cpu_to_be16(i + 1);
248
- }
249
- next_seg = mlx5_wq_ll_get_wqe(wq, i);
250
- wq->tail_next = &next_seg->next_wqe_index;
251
-
241
+ mlx5_wq_ll_init_list(wq);
252242 wq_ctrl->mdev = mdev;
253243
254244 return 0;
....@@ -259,6 +249,15 @@
259249 return err;
260250 }
261251
252
+void mlx5_wq_ll_reset(struct mlx5_wq_ll *wq)
253
+{
254
+ wq->head = 0;
255
+ wq->wqe_ctr = 0;
256
+ wq->cur_sz = 0;
257
+ mlx5_wq_ll_init_list(wq);
258
+ mlx5_wq_ll_update_db_record(wq);
259
+}
260
+
262261 void mlx5_wq_destroy(struct mlx5_wq_ctrl *wq_ctrl)
263262 {
264263 mlx5_frag_buf_free(wq_ctrl->mdev, &wq_ctrl->buf);