2 * Copyright (c) 2005 Robert N. M. Watson
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29 #include <sys/param.h>
31 #define LIBMEMSTAT /* Cause vm_page.h not to include opt_vmpage.h */
33 #include <vm/vm_page.h>
36 #include <vm/uma_int.h>
46 static struct nlist namelist[] = {
48 { .n_name = "_uma_kegs" },
49 #define X_MP_MAXCPUS 1
50 { .n_name = "_mp_maxcpus" },
52 { .n_name = "_mp_maxid" },
54 { .n_name = "_all_cpus" },
62 fprintf(stderr, "umastat [-M core [-N system]]\n");
67 kread(kvm_t *kvm, void *kvm_pointer, void *address, size_t size,
72 ret = kvm_read(kvm, (unsigned long)kvm_pointer + offset, address,
75 return (MEMSTAT_ERROR_KVM);
76 if ((size_t)ret != size)
77 return (MEMSTAT_ERROR_KVM_SHORTREAD);
82 kread_string(kvm_t *kvm, const void *kvm_pointer, char *buffer, int buflen)
87 for (i = 0; i < buflen; i++) {
88 ret = kvm_read(kvm, (unsigned long)kvm_pointer + i,
89 &(buffer[i]), sizeof(char));
91 return (MEMSTAT_ERROR_KVM);
92 if ((size_t)ret != sizeof(char))
93 return (MEMSTAT_ERROR_KVM_SHORTREAD);
94 if (buffer[i] == '\0')
103 kread_symbol(kvm_t *kvm, int index, void *address, size_t size,
108 ret = kvm_read(kvm, namelist[index].n_value + offset, address, size);
110 return (MEMSTAT_ERROR_KVM);
111 if ((size_t)ret != size)
112 return (MEMSTAT_ERROR_KVM_SHORTREAD);
116 static const struct flaginfo {
120 { UMA_ZFLAG_MULTI, "multi" },
121 { UMA_ZFLAG_DRAINING, "draining" },
122 { UMA_ZFLAG_BUCKET, "bucket" },
123 { UMA_ZFLAG_INTERNAL, "internal" },
124 { UMA_ZFLAG_FULL, "full" },
125 { UMA_ZFLAG_CACHEONLY, "cacheonly" },
126 { UMA_ZONE_PAGEABLE, "pageable" },
127 { UMA_ZONE_ZINIT, "zinit" },
128 { UMA_ZONE_STATIC, "static" },
129 { UMA_ZONE_OFFPAGE, "offpage" },
130 { UMA_ZONE_MALLOC, "malloc" },
131 { UMA_ZONE_NOFREE, "nofree" },
132 { UMA_ZONE_MTXCLASS, "mtxclass" },
133 { UMA_ZONE_VM, "vm" },
134 { UMA_ZONE_HASH, "hash" },
135 { UMA_ZONE_SECONDARY, "secondary" },
136 { UMA_ZONE_MAXBUCKET, "maxbucket" },
137 { UMA_ZONE_CACHESPREAD, "cachespread" },
138 { UMA_ZONE_VTOSLAB, "vtoslab" },
139 { UMA_ZONE_NODUMP, "nodump" },
140 { UMA_ZONE_PCPU, "pcpu" },
142 static const int flaginfo_count = sizeof(flaginfo) / sizeof(struct flaginfo);
145 uma_print_keg_flags(struct uma_keg *ukp, const char *spaces)
149 if (!ukp->uk_flags) {
150 printf("%suk_flags = 0;\n", spaces);
154 printf("%suk_flags = ", spaces);
155 for (i = 0, count = 0; i < flaginfo_count; i++) {
156 if (ukp->uk_flags & flaginfo[i].fi_flag) {
159 printf("%s", flaginfo[i].fi_name);
167 uma_print_keg_align(struct uma_keg *ukp, const char *spaces)
170 switch(ukp->uk_align) {
172 printf("%suk_align = UMA_ALIGN_PTR;\n", spaces);
177 printf("%suk_align = UMA_ALIGN_LONG;\n", spaces);
181 printf("%suk_align = UMA_ALIGN_INT;\n", spaces);
185 case UMA_ALIGN_SHORT:
186 printf("%suk_align = UMA_ALIGN_SHORT;\n", spaces);
190 printf("%suk_align = UMA_ALIGN_CHAR;\n", spaces);
193 case UMA_ALIGN_CACHE:
194 printf("%suk_align = UMA_ALIGN_CACHE;\n", spaces);
198 printf("%suk_align = %d\n", spaces, ukp->uk_align);
202 LIST_HEAD(bucketlist, uma_bucket);
205 uma_print_bucket(struct uma_bucket *ubp, const char *spaces __unused)
208 printf("{ ub_cnt = %d, ub_entries = %d }", ubp->ub_cnt,
213 uma_print_bucketlist(kvm_t *kvm, struct bucketlist *bucketlist,
214 const char *name, const char *spaces)
216 struct uma_bucket *ubp, ub;
217 uint64_t total_entries, total_cnt;
220 printf("%s%s {", spaces, name);
222 total_entries = total_cnt = 0;
224 for (ubp = LIST_FIRST(bucketlist); ubp != NULL; ubp =
225 LIST_NEXT(&ub, ub_link)) {
226 ret = kread(kvm, ubp, &ub, sizeof(ub), 0);
228 errx(-1, "uma_print_bucketlist: %s", kvm_geterr(kvm));
230 printf("\n%s ", spaces);
231 uma_print_bucket(&ub, "");
233 total_entries += ub.ub_entries;
234 total_cnt += ub.ub_cnt;
239 printf("%s}; // total cnt %ju, total entries %ju\n", spaces,
240 total_cnt, total_entries);
244 uma_print_cache(kvm_t *kvm, struct uma_cache *cache, const char *name,
245 int cpu, const char *spaces, int *ub_cnt_add, int *ub_entries_add)
247 struct uma_bucket ub;
250 printf("%s%s[%d] = {\n", spaces, name, cpu);
251 printf("%s uc_frees = %ju;\n", spaces, cache->uc_frees);
252 printf("%s uc_allocs = %ju;\n", spaces, cache->uc_allocs);
254 if (cache->uc_freebucket != NULL) {
255 ret = kread(kvm, cache->uc_freebucket, &ub, sizeof(ub), 0);
257 errx(-1, "uma_print_cache: %s", kvm_geterr(kvm));
258 printf("%s uc_freebucket ", spaces);
259 uma_print_bucket(&ub, spaces);
261 if (ub_cnt_add != NULL)
262 *ub_cnt_add += ub.ub_cnt;
263 if (ub_entries_add != NULL)
264 *ub_entries_add += ub.ub_entries;
266 printf("%s uc_freebucket = NULL;\n", spaces);
267 if (cache->uc_allocbucket != NULL) {
268 ret = kread(kvm, cache->uc_allocbucket, &ub, sizeof(ub), 0);
270 errx(-1, "uma_print_cache: %s", kvm_geterr(kvm));
271 printf("%s uc_allocbucket ", spaces);
272 uma_print_bucket(&ub, spaces);
274 if (ub_cnt_add != NULL)
275 *ub_cnt_add += ub.ub_cnt;
276 if (ub_entries_add != NULL)
277 *ub_entries_add += ub.ub_entries;
279 printf("%s uc_allocbucket = NULL;\n", spaces);
280 printf("%s};\n", spaces);
284 main(int argc, char *argv[])
286 LIST_HEAD(, uma_keg) uma_kegs;
287 char name[MEMTYPE_MAXNAME];
288 struct uma_keg *kzp, kz;
289 struct uma_zone *uzp, *uzp_userspace;
291 int all_cpus, cpu, mp_maxcpus, mp_maxid, ret, ub_cnt, ub_entries;
292 size_t uzp_userspace_len;
295 char errbuf[_POSIX2_LINE_MAX];
297 memf = nlistf = NULL;
298 while ((ch = getopt(argc, argv, "M:N:")) != -1) {
315 if (nlistf != NULL && memf == NULL)
318 kvm = kvm_openfiles(nlistf, memf, NULL, 0, errbuf);
320 errx(-1, "kvm_openfiles: %s", errbuf);
322 if (kvm_nlist(kvm, namelist) != 0)
323 err(-1, "kvm_nlist");
325 if (namelist[X_UMA_KEGS].n_type == 0 ||
326 namelist[X_UMA_KEGS].n_value == 0)
327 errx(-1, "kvm_nlist return");
329 ret = kread_symbol(kvm, X_MP_MAXCPUS, &mp_maxcpus, sizeof(mp_maxcpus),
332 errx(-1, "kread_symbol: %s", kvm_geterr(kvm));
334 printf("mp_maxcpus = %d\n", mp_maxcpus);
336 ret = kread_symbol(kvm, X_MP_MAXID, &mp_maxid, sizeof(mp_maxid), 0);
338 errx(-1, "kread_symbol: %s", kvm_geterr(kvm));
340 printf("mp_maxid = %d\n", mp_maxid);
342 ret = kread_symbol(kvm, X_ALLCPU, &all_cpus, sizeof(all_cpus), 0);
344 errx(-1, "kread_symbol: %s", kvm_geterr(kvm));
346 printf("all_cpus = %x\n", all_cpus);
348 ret = kread_symbol(kvm, X_UMA_KEGS, &uma_kegs, sizeof(uma_kegs), 0);
350 errx(-1, "kread_symbol: %s", kvm_geterr(kvm));
353 * uma_zone_t ends in an array of mp_maxid cache entries. However,
354 * it is statically declared as an array of size 1, so we need to
355 * provide additional space.
357 uzp_userspace_len = sizeof(struct uma_zone) + mp_maxid *
358 sizeof(struct uma_cache);
359 uzp_userspace = malloc(uzp_userspace_len);
360 if (uzp_userspace == NULL)
363 for (kzp = LIST_FIRST(&uma_kegs); kzp != NULL; kzp =
364 LIST_NEXT(&kz, uk_link)) {
365 ret = kread(kvm, kzp, &kz, sizeof(kz), 0);
368 errx(-1, "kread: %s", kvm_geterr(kvm));
372 uma_print_keg_align(&kz, " ");
373 printf(" uk_pages = %d\n", kz.uk_pages);
374 printf(" uk_free = %d\n", kz.uk_free);
375 printf(" uk_reserve = %d\n", kz.uk_reserve);
376 printf(" uk_size = %d\n", kz.uk_size);
377 printf(" uk_rsize = %d\n", kz.uk_rsize);
378 printf(" uk_maxpages = %d\n", kz.uk_maxpages);
380 printf(" uk_pgoff = %d\n", kz.uk_pgoff);
381 printf(" uk_ppera = %d\n", kz.uk_ppera);
382 printf(" uk_ipers = %d\n", kz.uk_ipers);
383 uma_print_keg_flags(&kz, " ");
385 if (LIST_FIRST(&kz.uk_zones) == NULL) {
386 printf("; No zones.\n");
390 for (uzp = LIST_FIRST(&kz.uk_zones); uzp != NULL; uzp =
391 LIST_NEXT(uzp_userspace, uz_link)) {
393 * We actually copy in twice: once with the base
394 * structure, so that we can then decide if we also
395 * need to copy in the caches. This prevents us
396 * from reading past the end of the base UMA zones,
397 * which is unlikely to cause problems but could.
399 ret = kread(kvm, uzp, uzp_userspace,
400 sizeof(struct uma_zone), 0);
403 errx(-1, "kread: %s", kvm_geterr(kvm));
405 if (!(kz.uk_flags & UMA_ZFLAG_INTERNAL)) {
406 ret = kread(kvm, uzp, uzp_userspace,
407 uzp_userspace_len, 0);
410 errx(-1, "kread: %s",
414 ret = kread_string(kvm, uzp_userspace->uz_name, name,
418 errx(-1, "kread_string: %s", kvm_geterr(kvm));
421 printf(" uz_name = \"%s\";\n", name);
422 printf(" uz_allocs = %lu;\n",
423 uzp_userspace->uz_allocs);
424 printf(" uz_frees = %lu;\n",
425 uzp_userspace->uz_frees);
426 printf(" uz_fails = %lu;\n",
427 uzp_userspace->uz_fails);
428 printf(" uz_sleeps = %ju;\n",
429 uzp_userspace->uz_sleeps);
430 printf(" uz_count = %u;\n",
431 uzp_userspace->uz_count);
432 uma_print_bucketlist(kvm, (void *)
433 &uzp_userspace->uz_buckets, "uz_buckets",
436 if (!(kz.uk_flags & UMA_ZFLAG_INTERNAL)) {
437 ub_cnt = ub_entries = 0;
438 for (cpu = 0; cpu <= mp_maxid; cpu++) {
439 /* if (CPU_ABSENT(cpu)) */
440 if ((all_cpus & (1 << cpu)) == 0)
443 &uzp_userspace->uz_cpu[cpu],
444 "uc_cache", cpu, " ", &ub_cnt,
447 printf(" // %d cache total cnt, %d total "
448 "entries\n", ub_cnt, ub_entries);