static int do_dump(int argc, char **argv) { struct btf *btf = NULL; __u32 root_type_ids[2]; int root_type_cnt = 0; __u32 btf_id = -1; const char *src; int fd = -1; int err; if (!REQ_ARGS(2)) { usage(); return -1; } src = GET_ARG(); if (is_prefix(src, "map")) { struct bpf_map_info info = {}; __u32 len = sizeof(info); if (!REQ_ARGS(2)) { usage(); return -1; } fd = map_parse_fd_and_info(&argc, &argv, &info, &len); if (fd < 0) return -1; btf_id = info.btf_id; if (argc && is_prefix(*argv, "key")) { root_type_ids[root_type_cnt++] = info.btf_key_type_id; NEXT_ARG(); } else if (argc && is_prefix(*argv, "value")) { root_type_ids[root_type_cnt++] = info.btf_value_type_id; NEXT_ARG(); } else if (argc && is_prefix(*argv, "all")) { NEXT_ARG(); } else if (argc && is_prefix(*argv, "kv")) { root_type_ids[root_type_cnt++] = info.btf_key_type_id; root_type_ids[root_type_cnt++] = info.btf_value_type_id; NEXT_ARG(); } else { root_type_ids[root_type_cnt++] = info.btf_key_type_id; root_type_ids[root_type_cnt++] = info.btf_value_type_id; } } else if (is_prefix(src, "prog")) { struct bpf_prog_info info = {}; __u32 len = sizeof(info); if (!REQ_ARGS(2)) { usage(); return -1; } fd = prog_parse_fd(&argc, &argv); if (fd < 0) return -1; err = bpf_obj_get_info_by_fd(fd, &info, &len); if (err) { p_err("can't get prog info: %s", strerror(errno)); goto done; } btf_id = info.btf_id; } else if (is_prefix(src, "id")) { char *endptr; btf_id = strtoul(*argv, &endptr, 0); if (*endptr) { p_err("can't parse %s as ID", **argv); return -1; } NEXT_ARG(); } else if (is_prefix(src, "file")) { err = btf_load_from_elf(*argv, &btf); if (err) goto done; NEXT_ARG(); } else { err = -1; p_err("unrecognized BTF source specifier: '%s'", src); goto done; } if (!btf) { err = btf__get_from_id(btf_id, &btf); if (err) { p_err("get btf by id (%u): %s", btf_id, strerror(err)); goto done; } if (!btf) { err = ENOENT; p_err("can't find btf with ID (%u)", btf_id); goto done; } } dump_btf_raw(btf, root_type_ids, root_type_cnt); done: close(fd); btf__free(btf); return err; }
static int bpf_perf_event_open(int map_fd, int key, int cpu) { struct perf_event_attr attr = { .sample_type = PERF_SAMPLE_RAW | PERF_SAMPLE_TIME, .type = PERF_TYPE_SOFTWARE, .config = PERF_COUNT_SW_BPF_OUTPUT, }; int pmu_fd; pmu_fd = sys_perf_event_open(&attr, -1, cpu, -1, 0); if (pmu_fd < 0) { p_err("failed to open perf event %d for CPU %d", key, cpu); return -1; } if (bpf_map_update_elem(map_fd, &key, &pmu_fd, BPF_ANY)) { p_err("failed to update map for event %d for CPU %d", key, cpu); goto err_close; } if (ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0)) { p_err("failed to enable event %d for CPU %d", key, cpu); goto err_close; } return pmu_fd; err_close: close(pmu_fd); return -1; } int do_event_pipe(int argc, char **argv) { int i, nfds, map_fd, index = -1, cpu = -1; struct bpf_map_info map_info = {}; struct event_ring_info *rings; size_t tmp_buf_sz = 0; void *tmp_buf = NULL; struct pollfd *pfds; __u32 map_info_len; bool do_all = true; map_info_len = sizeof(map_info); map_fd = map_parse_fd_and_info(&argc, &argv, &map_info, &map_info_len); if (map_fd < 0) return -1; if (map_info.type != BPF_MAP_TYPE_PERF_EVENT_ARRAY) { p_err("map is not a perf event array"); goto err_close_map; } while (argc) { if (argc < 2) BAD_ARG(); if (is_prefix(*argv, "cpu")) { char *endptr; NEXT_ARG(); cpu = strtoul(*argv, &endptr, 0); if (*endptr) { p_err("can't parse %s as CPU ID", **argv); goto err_close_map; } NEXT_ARG(); } else if (is_prefix(*argv, "index")) { char *endptr; NEXT_ARG(); index = strtoul(*argv, &endptr, 0); if (*endptr) { p_err("can't parse %s as index", **argv); goto err_close_map; } NEXT_ARG(); } else { BAD_ARG(); } do_all = false; } if (!do_all) { if (index == -1 || cpu == -1) { p_err("cpu and index must be specified together"); goto err_close_map; } nfds = 1; } else { nfds = min(get_possible_cpus(), map_info.max_entries); cpu = 0; index = 0; } rings = calloc(nfds, sizeof(rings[0])); if (!rings) goto err_close_map; pfds = calloc(nfds, sizeof(pfds[0])); if (!pfds) goto err_free_rings; for (i = 0; i < nfds; i++) { rings[i].cpu = cpu + i; rings[i].key = index + i; rings[i].fd = bpf_perf_event_open(map_fd, rings[i].key, rings[i].cpu); if (rings[i].fd < 0) goto err_close_fds_prev; rings[i].mem = perf_event_mmap(rings[i].fd); if (!rings[i].mem) goto err_close_fds_current; pfds[i].fd = rings[i].fd; pfds[i].events = POLLIN; } signal(SIGINT, int_exit); signal(SIGHUP, int_exit); signal(SIGTERM, int_exit); if (json_output) jsonw_start_array(json_wtr); while (!stop) { poll(pfds, nfds, 200); for (i = 0; i < nfds; i++) perf_event_read(&rings[i], &tmp_buf, &tmp_buf_sz); } free(tmp_buf); if (json_output) jsonw_end_array(json_wtr); for (i = 0; i < nfds; i++) { perf_event_unmap(rings[i].mem); close(rings[i].fd); } free(pfds); free(rings); close(map_fd); return 0; err_close_fds_prev: while (i--) { perf_event_unmap(rings[i].mem); err_close_fds_current: close(rings[i].fd); } free(pfds); err_free_rings: free(rings); err_close_map: close(map_fd); return -1; }