hc
2023-12-11 d2ccde1c8e90d38cee87a1b0309ad2827f3fd30d
kernel/tools/bpf/bpftool/map_perf_ring.c
....@@ -1,4 +1,4 @@
1
-// SPDX-License-Identifier: GPL-2.0-only
1
+// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
22 /* Copyright (C) 2018 Netronome Systems, Inc. */
33 /* This program is free software; you can redistribute it and/or
44 * modify it under the terms of version 2 of the GNU General Public
....@@ -6,7 +6,7 @@
66 */
77 #include <errno.h>
88 #include <fcntl.h>
9
-#include <libbpf.h>
9
+#include <bpf/libbpf.h>
1010 #include <poll.h>
1111 #include <signal.h>
1212 #include <stdbool.h>
....@@ -21,14 +21,14 @@
2121 #include <sys/mman.h>
2222 #include <sys/syscall.h>
2323
24
-#include <bpf.h>
24
+#include <bpf/bpf.h>
2525 #include <perf-sys.h>
2626
2727 #include "main.h"
2828
2929 #define MMAP_PAGE_CNT 16
3030
31
-static bool stop;
31
+static volatile bool stop;
3232
3333 struct event_ring_info {
3434 int fd;
....@@ -39,9 +39,15 @@
3939
4040 struct perf_event_sample {
4141 struct perf_event_header header;
42
- u64 time;
42
+ __u64 time;
4343 __u32 size;
4444 unsigned char data[];
45
+};
46
+
47
+struct perf_event_lost {
48
+ struct perf_event_header header;
49
+ __u64 id;
50
+ __u64 lost;
4551 };
4652
4753 static void int_exit(int signo)
....@@ -50,24 +56,32 @@
5056 stop = true;
5157 }
5258
53
-static enum bpf_perf_event_ret print_bpf_output(void *event, void *priv)
59
+struct event_pipe_ctx {
60
+ bool all_cpus;
61
+ int cpu;
62
+ int idx;
63
+};
64
+
65
+static enum bpf_perf_event_ret
66
+print_bpf_output(void *private_data, int cpu, struct perf_event_header *event)
5467 {
55
- struct event_ring_info *ring = priv;
56
- struct perf_event_sample *e = event;
57
- struct {
58
- struct perf_event_header header;
59
- __u64 id;
60
- __u64 lost;
61
- } *lost = event;
68
+ struct perf_event_sample *e = container_of(event,
69
+ struct perf_event_sample,
70
+ header);
71
+ struct perf_event_lost *lost = container_of(event,
72
+ struct perf_event_lost,
73
+ header);
74
+ struct event_pipe_ctx *ctx = private_data;
75
+ int idx = ctx->all_cpus ? cpu : ctx->idx;
6276
6377 if (json_output) {
6478 jsonw_start_object(json_wtr);
6579 jsonw_name(json_wtr, "type");
6680 jsonw_uint(json_wtr, e->header.type);
6781 jsonw_name(json_wtr, "cpu");
68
- jsonw_uint(json_wtr, ring->cpu);
82
+ jsonw_uint(json_wtr, cpu);
6983 jsonw_name(json_wtr, "index");
70
- jsonw_uint(json_wtr, ring->key);
84
+ jsonw_uint(json_wtr, idx);
7185 if (e->header.type == PERF_RECORD_SAMPLE) {
7286 jsonw_name(json_wtr, "timestamp");
7387 jsonw_uint(json_wtr, e->time);
....@@ -87,7 +101,7 @@
87101 if (e->header.type == PERF_RECORD_SAMPLE) {
88102 printf("== @%lld.%09lld CPU: %d index: %d =====\n",
89103 e->time / 1000000000ULL, e->time % 1000000000ULL,
90
- ring->cpu, ring->key);
104
+ cpu, idx);
91105 fprint_hex(stdout, e->data, e->size, " ");
92106 printf("\n");
93107 } else if (e->header.type == PERF_RECORD_LOST) {
....@@ -101,87 +115,25 @@
101115 return LIBBPF_PERF_EVENT_CONT;
102116 }
103117
104
-static void
105
-perf_event_read(struct event_ring_info *ring, void **buf, size_t *buf_len)
118
+int do_event_pipe(int argc, char **argv)
106119 {
107
- enum bpf_perf_event_ret ret;
108
-
109
- ret = bpf_perf_event_read_simple(ring->mem,
110
- MMAP_PAGE_CNT * get_page_size(),
111
- get_page_size(), buf, buf_len,
112
- print_bpf_output, ring);
113
- if (ret != LIBBPF_PERF_EVENT_CONT) {
114
- fprintf(stderr, "perf read loop failed with %d\n", ret);
115
- stop = true;
116
- }
117
-}
118
-
119
-static int perf_mmap_size(void)
120
-{
121
- return get_page_size() * (MMAP_PAGE_CNT + 1);
122
-}
123
-
124
-static void *perf_event_mmap(int fd)
125
-{
126
- int mmap_size = perf_mmap_size();
127
- void *base;
128
-
129
- base = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
130
- if (base == MAP_FAILED) {
131
- p_err("event mmap failed: %s\n", strerror(errno));
132
- return NULL;
133
- }
134
-
135
- return base;
136
-}
137
-
138
-static void perf_event_unmap(void *mem)
139
-{
140
- if (munmap(mem, perf_mmap_size()))
141
- fprintf(stderr, "Can't unmap ring memory!\n");
142
-}
143
-
144
-static int bpf_perf_event_open(int map_fd, int key, int cpu)
145
-{
146
- struct perf_event_attr attr = {
120
+ struct perf_event_attr perf_attr = {
147121 .sample_type = PERF_SAMPLE_RAW | PERF_SAMPLE_TIME,
148122 .type = PERF_TYPE_SOFTWARE,
149123 .config = PERF_COUNT_SW_BPF_OUTPUT,
124
+ .sample_period = 1,
125
+ .wakeup_events = 1,
150126 };
151
- int pmu_fd;
152
-
153
- pmu_fd = sys_perf_event_open(&attr, -1, cpu, -1, 0);
154
- if (pmu_fd < 0) {
155
- p_err("failed to open perf event %d for CPU %d", key, cpu);
156
- return -1;
157
- }
158
-
159
- if (bpf_map_update_elem(map_fd, &key, &pmu_fd, BPF_ANY)) {
160
- p_err("failed to update map for event %d for CPU %d", key, cpu);
161
- goto err_close;
162
- }
163
- if (ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0)) {
164
- p_err("failed to enable event %d for CPU %d", key, cpu);
165
- goto err_close;
166
- }
167
-
168
- return pmu_fd;
169
-
170
-err_close:
171
- close(pmu_fd);
172
- return -1;
173
-}
174
-
175
-int do_event_pipe(int argc, char **argv)
176
-{
177
- int i, nfds, map_fd, index = -1, cpu = -1;
178127 struct bpf_map_info map_info = {};
179
- struct event_ring_info *rings;
180
- size_t tmp_buf_sz = 0;
181
- void *tmp_buf = NULL;
182
- struct pollfd *pfds;
128
+ struct perf_buffer_raw_opts opts = {};
129
+ struct event_pipe_ctx ctx = {
130
+ .all_cpus = true,
131
+ .cpu = -1,
132
+ .idx = -1,
133
+ };
134
+ struct perf_buffer *pb;
183135 __u32 map_info_len;
184
- bool do_all = true;
136
+ int err, map_fd;
185137
186138 map_info_len = sizeof(map_info);
187139 map_fd = map_parse_fd_and_info(&argc, &argv, &map_info, &map_info_len);
....@@ -203,7 +155,7 @@
203155 char *endptr;
204156
205157 NEXT_ARG();
206
- cpu = strtoul(*argv, &endptr, 0);
158
+ ctx.cpu = strtoul(*argv, &endptr, 0);
207159 if (*endptr) {
208160 p_err("can't parse %s as CPU ID", *argv);
209161 goto err_close_map;
....@@ -214,7 +166,7 @@
214166 char *endptr;
215167
216168 NEXT_ARG();
217
- index = strtoul(*argv, &endptr, 0);
169
+ ctx.idx = strtoul(*argv, &endptr, 0);
218170 if (*endptr) {
219171 p_err("can't parse %s as index", *argv);
220172 goto err_close_map;
....@@ -226,45 +178,32 @@
226178 goto err_close_map;
227179 }
228180
229
- do_all = false;
181
+ ctx.all_cpus = false;
230182 }
231183
232
- if (!do_all) {
233
- if (index == -1 || cpu == -1) {
184
+ if (!ctx.all_cpus) {
185
+ if (ctx.idx == -1 || ctx.cpu == -1) {
234186 p_err("cpu and index must be specified together");
235187 goto err_close_map;
236188 }
237
-
238
- nfds = 1;
239189 } else {
240
- nfds = min(get_possible_cpus(), map_info.max_entries);
241
- cpu = 0;
242
- index = 0;
190
+ ctx.cpu = 0;
191
+ ctx.idx = 0;
243192 }
244193
245
- rings = calloc(nfds, sizeof(rings[0]));
246
- if (!rings)
194
+ opts.attr = &perf_attr;
195
+ opts.event_cb = print_bpf_output;
196
+ opts.ctx = &ctx;
197
+ opts.cpu_cnt = ctx.all_cpus ? 0 : 1;
198
+ opts.cpus = &ctx.cpu;
199
+ opts.map_keys = &ctx.idx;
200
+
201
+ pb = perf_buffer__new_raw(map_fd, MMAP_PAGE_CNT, &opts);
202
+ err = libbpf_get_error(pb);
203
+ if (err) {
204
+ p_err("failed to create perf buffer: %s (%d)",
205
+ strerror(err), err);
247206 goto err_close_map;
248
-
249
- pfds = calloc(nfds, sizeof(pfds[0]));
250
- if (!pfds)
251
- goto err_free_rings;
252
-
253
- for (i = 0; i < nfds; i++) {
254
- rings[i].cpu = cpu + i;
255
- rings[i].key = index + i;
256
-
257
- rings[i].fd = bpf_perf_event_open(map_fd, rings[i].key,
258
- rings[i].cpu);
259
- if (rings[i].fd < 0)
260
- goto err_close_fds_prev;
261
-
262
- rings[i].mem = perf_event_mmap(rings[i].fd);
263
- if (!rings[i].mem)
264
- goto err_close_fds_current;
265
-
266
- pfds[i].fd = rings[i].fd;
267
- pfds[i].events = POLLIN;
268207 }
269208
270209 signal(SIGINT, int_exit);
....@@ -275,34 +214,24 @@
275214 jsonw_start_array(json_wtr);
276215
277216 while (!stop) {
278
- poll(pfds, nfds, 200);
279
- for (i = 0; i < nfds; i++)
280
- perf_event_read(&rings[i], &tmp_buf, &tmp_buf_sz);
217
+ err = perf_buffer__poll(pb, 200);
218
+ if (err < 0 && err != -EINTR) {
219
+ p_err("perf buffer polling failed: %s (%d)",
220
+ strerror(err), err);
221
+ goto err_close_pb;
222
+ }
281223 }
282
- free(tmp_buf);
283224
284225 if (json_output)
285226 jsonw_end_array(json_wtr);
286227
287
- for (i = 0; i < nfds; i++) {
288
- perf_event_unmap(rings[i].mem);
289
- close(rings[i].fd);
290
- }
291
- free(pfds);
292
- free(rings);
228
+ perf_buffer__free(pb);
293229 close(map_fd);
294230
295231 return 0;
296232
297
-err_close_fds_prev:
298
- while (i--) {
299
- perf_event_unmap(rings[i].mem);
300
-err_close_fds_current:
301
- close(rings[i].fd);
302
- }
303
- free(pfds);
304
-err_free_rings:
305
- free(rings);
233
+err_close_pb:
234
+ perf_buffer__free(pb);
306235 err_close_map:
307236 close(map_fd);
308237 return -1;