Correct BSD License clause numbering from 1-2-4 to 1-2-3.
[dragonfly.git] / sys / vm / vm_meter.c
1 /*
2  * (MPSAFE)
3  *
4  * Copyright (c) 1982, 1986, 1989, 1993
5  *      The Regents of the University of California.  All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  * 3. Neither the name of the University nor the names of its contributors
16  *    may be used to endorse or promote products derived from this software
17  *    without specific prior written permission.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
20  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
23  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
24  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
25  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
26  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
27  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
28  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29  * SUCH DAMAGE.
30  *
31  *      @(#)vm_meter.c  8.4 (Berkeley) 1/4/94
32  * $FreeBSD: src/sys/vm/vm_meter.c,v 1.34.2.7 2002/10/10 19:28:22 dillon Exp $
33  * $DragonFly: src/sys/vm/vm_meter.c,v 1.15 2008/04/28 18:04:08 dillon Exp $
34  */
35
36 #include <sys/param.h>
37 #include <sys/proc.h>
38 #include <sys/systm.h>
39 #include <sys/kernel.h>
40 #include <sys/resource.h>
41 #include <sys/vmmeter.h>
42
43 #include <vm/vm.h>
44 #include <vm/vm_page.h>
45 #include <vm/vm_extern.h>
46 #include <vm/vm_param.h>
47 #include <sys/lock.h>
48 #include <vm/pmap.h>
49 #include <vm/vm_map.h>
50 #include <vm/vm_object.h>
51 #include <sys/sysctl.h>
52
53 struct vmstats vmstats;
54
55 static int maxslp = MAXSLP;
56
57 SYSCTL_UINT(_vm, VM_V_FREE_MIN, v_free_min,
58         CTLFLAG_RW, &vmstats.v_free_min, 0,
59         "Minimum number of pages desired free");
60 SYSCTL_UINT(_vm, VM_V_FREE_TARGET, v_free_target,
61         CTLFLAG_RW, &vmstats.v_free_target, 0,
62         "Number of pages desired free");
63 SYSCTL_UINT(_vm, VM_V_FREE_RESERVED, v_free_reserved,
64         CTLFLAG_RW, &vmstats.v_free_reserved, 0,
65         "Number of pages reserved for deadlock");
66 SYSCTL_UINT(_vm, VM_V_INACTIVE_TARGET, v_inactive_target,
67         CTLFLAG_RW, &vmstats.v_inactive_target, 0,
68         "Number of pages desired inactive");
69 SYSCTL_UINT(_vm, VM_V_CACHE_MIN, v_cache_min,
70         CTLFLAG_RW, &vmstats.v_cache_min, 0,
71         "Min number of pages desired on cache queue");
72 SYSCTL_UINT(_vm, VM_V_CACHE_MAX, v_cache_max,
73         CTLFLAG_RW, &vmstats.v_cache_max, 0,
74         "Max number of pages in cached obj");
75 SYSCTL_UINT(_vm, VM_V_PAGEOUT_FREE_MIN, v_pageout_free_min,
76         CTLFLAG_RW, &vmstats.v_pageout_free_min, 0,
77         "Min number pages reserved for kernel");
78 SYSCTL_UINT(_vm, OID_AUTO, v_free_severe,
79         CTLFLAG_RW, &vmstats.v_free_severe, 0, "");
80
81 SYSCTL_STRUCT(_vm, VM_LOADAVG, loadavg, CTLFLAG_RD, 
82     &averunnable, loadavg, "Machine loadaverage history");
83
84 static int do_vmtotal_callback(struct proc *p, void *data);
85
86 /*
87  * No requirements.
88  */
89 static int
90 do_vmtotal(SYSCTL_HANDLER_ARGS)
91 {
92         struct vmtotal total;
93         struct vmtotal *totalp;
94         struct vm_object marker;
95         vm_object_t object;
96         long collisions;
97         int burst;
98
99         bzero(&total, sizeof(total));
100         totalp = &total;
101         bzero(&marker, sizeof(marker));
102         marker.type = OBJT_MARKER;
103         collisions = vmobj_token.t_collisions;
104
105 #if 0
106         /*
107          * Mark all objects as inactive.
108          */
109         lwkt_gettoken(&vmobj_token);
110         for (object = TAILQ_FIRST(&vm_object_list);
111             object != NULL;
112             object = TAILQ_NEXT(object,object_list)) {
113                 if (object->type == OBJT_MARKER)
114                         continue;
115                 vm_object_clear_flag(object, OBJ_ACTIVE);
116         }
117         lwkt_reltoken(&vmobj_token);
118 #endif
119
120         /*
121          * Calculate process statistics.
122          */
123         allproc_scan(do_vmtotal_callback, totalp);
124
125         /*
126          * Calculate object memory usage statistics.
127          */
128         lwkt_gettoken(&vmobj_token);
129         TAILQ_INSERT_HEAD(&vm_object_list, &marker, object_list);
130         burst = 0;
131
132         for (object = TAILQ_FIRST(&vm_object_list);
133             object != NULL;
134             object = TAILQ_NEXT(object, object_list)) {
135                 /*
136                  * devices, like /dev/mem, will badly skew our totals.
137                  * markers aren't real objects.
138                  */
139                 if (object->type == OBJT_MARKER)
140                         continue;
141                 if (object->type == OBJT_DEVICE)
142                         continue;
143                 if (object->size >= 0x7FFFFFFF) {
144                         /*
145                          * Probably unbounded anonymous memory (really
146                          * bounded by related vm_map_entry structures which
147                          * we do not have access to in this loop).
148                          */
149                         totalp->t_vm += object->resident_page_count;
150                 } else {
151                         /*
152                          * It's questionable how useful this is but...
153                          */
154                         totalp->t_vm += object->size;
155                 }
156                 totalp->t_rm += object->resident_page_count;
157                 if (object->flags & OBJ_ACTIVE) {
158                         totalp->t_avm += object->size;
159                         totalp->t_arm += object->resident_page_count;
160                 }
161                 if (object->shadow_count > 1) {
162                         /* shared object */
163                         totalp->t_vmshr += object->size;
164                         totalp->t_rmshr += object->resident_page_count;
165                         if (object->flags & OBJ_ACTIVE) {
166                                 totalp->t_avmshr += object->size;
167                                 totalp->t_armshr += object->resident_page_count;
168                         }
169                 }
170
171                 /*
172                  * Don't waste time unnecessarily
173                  */
174                 if (++burst < 25)
175                         continue;
176                 burst = 0;
177
178                 /*
179                  * Don't hog the vmobj_token if someone else wants it.
180                  */
181                 TAILQ_REMOVE(&vm_object_list, &marker, object_list);
182                 TAILQ_INSERT_AFTER(&vm_object_list, object,
183                                    &marker, object_list);
184                 object = &marker;
185                 if (collisions != vmobj_token.t_collisions) {
186                         tsleep(&vm_object_list, 0, "breath", 1);
187                         collisions = vmobj_token.t_collisions;
188                 } else {
189                         lwkt_yield();
190                 }
191         }
192
193         TAILQ_REMOVE(&vm_object_list, &marker, object_list);
194         lwkt_reltoken(&vmobj_token);
195
196         totalp->t_free = vmstats.v_free_count + vmstats.v_cache_count;
197
198         return (sysctl_handle_opaque(oidp, totalp, sizeof total, req));
199 }
200
201 /*
202  * The caller must hold proc_token.
203  */
204 static int
205 do_vmtotal_callback(struct proc *p, void *data)
206 {
207         struct vmtotal *totalp = data;
208         struct lwp *lp;
209
210         if (p->p_flags & P_SYSTEM)
211                 return(0);
212
213         FOREACH_LWP_IN_PROC(lp, p) {
214                 switch (lp->lwp_stat) {
215                 case LSSTOP:
216                 case LSSLEEP:
217                         if ((p->p_flags & P_SWAPPEDOUT) == 0) {
218                                 if ((lp->lwp_flags & LWP_SINTR) == 0)
219                                         totalp->t_dw++;
220                                 else if (lp->lwp_slptime < maxslp)
221                                         totalp->t_sl++;
222                         } else if (lp->lwp_slptime < maxslp) {
223                                 totalp->t_sw++;
224                         }
225                         if (lp->lwp_slptime >= maxslp)
226                                 return(0);
227                         break;
228
229                 case LSRUN:
230                         if (p->p_flags & P_SWAPPEDOUT)
231                                 totalp->t_sw++;
232                         else
233                                 totalp->t_rq++;
234                         if (p->p_stat == SIDL)
235                                 return(0);
236                         break;
237
238                 default:
239                         return (0);
240                 }
241
242                 /*
243                  * Set while in vm_fault()
244                  */
245                 if (lp->lwp_flags & LWP_PAGING)
246                         totalp->t_pw++;
247         }
248         return(0);
249 }
250
251 /*
252  * No requirements.
253  */
254 static int
255 do_vmstats(SYSCTL_HANDLER_ARGS)
256 {
257         struct vmstats vms = vmstats;
258         return (sysctl_handle_opaque(oidp, &vms, sizeof(vms), req));
259 }
260
261 /*
262  * No requirements.
263  */
264 static int
265 do_vmmeter(SYSCTL_HANDLER_ARGS)
266 {
267         int boffset = offsetof(struct vmmeter, vmmeter_uint_begin);
268         int eoffset = offsetof(struct vmmeter, vmmeter_uint_end);
269         struct vmmeter vmm;
270         int i;
271
272         bzero(&vmm, sizeof(vmm));
273         for (i = 0; i < ncpus; ++i) {
274                 int off;
275                 struct globaldata *gd = globaldata_find(i);
276
277                 for (off = boffset; off <= eoffset; off += sizeof(u_int)) {
278                         *(u_int *)((char *)&vmm + off) +=
279                                 *(u_int *)((char *)&gd->gd_cnt + off);
280                 }
281                 
282         }
283         vmm.v_intr += vmm.v_ipi + vmm.v_timer;
284         return (sysctl_handle_opaque(oidp, &vmm, sizeof(vmm), req));
285 }
286
287 /*
288  * vcnt() -     accumulate statistics from the cnt structure for each cpu
289  *
290  *      The vmmeter structure is now per-cpu as well as global.  Those
291  *      statistics which can be kept on a per-cpu basis (to avoid cache
292  *      stalls between cpus) can be moved to the per-cpu vmmeter.  Remaining
293  *      statistics, such as v_free_reserved, are left in the global
294  *      structure.
295  *
296  * (sysctl_oid *oidp, void *arg1, int arg2, struct sysctl_req *req)
297  *
298  * No requirements.
299  */
300 static int
301 vcnt(SYSCTL_HANDLER_ARGS)
302 {
303         int i;
304         int count = 0;
305         int offset = arg2;
306
307         for (i = 0; i < ncpus; ++i) {
308                 struct globaldata *gd = globaldata_find(i);
309                 count += *(int *)((char *)&gd->gd_cnt + offset);
310         }
311         return(SYSCTL_OUT(req, &count, sizeof(int)));
312 }
313
314 /*
315  * No requirements.
316  */
317 static int
318 vcnt_intr(SYSCTL_HANDLER_ARGS)
319 {
320         int i;
321         int count = 0;
322
323         for (i = 0; i < ncpus; ++i) {
324                 struct globaldata *gd = globaldata_find(i);
325
326                 count += gd->gd_cnt.v_intr + gd->gd_cnt.v_ipi +
327                          gd->gd_cnt.v_timer;
328         }
329         return(SYSCTL_OUT(req, &count, sizeof(int)));
330 }
331
332 #define VMMETEROFF(var) offsetof(struct vmmeter, var)
333
334 SYSCTL_PROC(_vm, OID_AUTO, vmtotal, CTLTYPE_OPAQUE|CTLFLAG_RD,
335     0, sizeof(struct vmtotal), do_vmtotal, "S,vmtotal", 
336     "System virtual memory aggregate");
337 SYSCTL_PROC(_vm, OID_AUTO, vmstats, CTLTYPE_OPAQUE|CTLFLAG_RD,
338     0, sizeof(struct vmstats), do_vmstats, "S,vmstats", 
339     "System virtual memory statistics");
340 SYSCTL_PROC(_vm, OID_AUTO, vmmeter, CTLTYPE_OPAQUE|CTLFLAG_RD,
341     0, sizeof(struct vmmeter), do_vmmeter, "S,vmmeter", 
342     "System statistics");
343 SYSCTL_NODE(_vm, OID_AUTO, stats, CTLFLAG_RW, 0, "VM meter stats");
344 SYSCTL_NODE(_vm_stats, OID_AUTO, sys, CTLFLAG_RW, 0, "VM meter sys stats");
345 SYSCTL_NODE(_vm_stats, OID_AUTO, vm, CTLFLAG_RW, 0, "VM meter vm stats");
346 SYSCTL_NODE(_vm_stats, OID_AUTO, misc, CTLFLAG_RW, 0, "VM meter misc stats");
347
348 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_swtch, CTLTYPE_UINT|CTLFLAG_RD,
349         0, VMMETEROFF(v_swtch), vcnt, "IU", "Context switches");
350 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intrans_coll, CTLTYPE_UINT|CTLFLAG_RD,
351         0, VMMETEROFF(v_intrans_coll), vcnt, "IU", "Intransit map collisions (total)");
352 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intrans_wait, CTLTYPE_UINT|CTLFLAG_RD,
353         0, VMMETEROFF(v_intrans_wait), vcnt, "IU", "Intransit map collisions which blocked");
354 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_ints, CTLTYPE_UINT|CTLFLAG_RD,
355         0, VMMETEROFF(v_forwarded_ints), vcnt, "IU", "Forwarded interrupts due to MP lock");
356 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_hits, CTLTYPE_UINT|CTLFLAG_RD,
357         0, VMMETEROFF(v_forwarded_hits), vcnt, "IU", "Forwarded hits due to MP lock");
358 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_misses, CTLTYPE_UINT|CTLFLAG_RD,
359         0, VMMETEROFF(v_forwarded_misses), vcnt, "IU", "Forwarded misses due to MP lock");
360 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_trap, CTLTYPE_UINT|CTLFLAG_RD,
361         0, VMMETEROFF(v_trap), vcnt, "IU", "Traps");
362 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_syscall, CTLTYPE_UINT|CTLFLAG_RD,
363         0, VMMETEROFF(v_syscall), vcnt, "IU", "Syscalls");
364 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intr, CTLTYPE_UINT|CTLFLAG_RD,
365         0, VMMETEROFF(v_intr), vcnt_intr, "IU", "Hardware interrupts");
366 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_ipi, CTLTYPE_UINT|CTLFLAG_RD,
367         0, VMMETEROFF(v_ipi), vcnt, "IU", "Inter-processor interrupts");
368 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_timer, CTLTYPE_UINT|CTLFLAG_RD,
369         0, VMMETEROFF(v_timer), vcnt, "IU", "LAPIC timer interrupts");
370 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_soft, CTLTYPE_UINT|CTLFLAG_RD,
371         0, VMMETEROFF(v_soft), vcnt, "IU", "Software interrupts");
372 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vm_faults, CTLTYPE_UINT|CTLFLAG_RD,
373         0, VMMETEROFF(v_vm_faults), vcnt, "IU", "VM faults");
374 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_cow_faults, CTLTYPE_UINT|CTLFLAG_RD,
375         0, VMMETEROFF(v_cow_faults), vcnt, "IU", "COW faults");
376 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_cow_optim, CTLTYPE_UINT|CTLFLAG_RD,
377         0, VMMETEROFF(v_cow_optim), vcnt, "IU", "Optimized COW faults");
378 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_zfod, CTLTYPE_UINT|CTLFLAG_RD,
379         0, VMMETEROFF(v_zfod), vcnt, "IU", "Zero fill");
380 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_ozfod, CTLTYPE_UINT|CTLFLAG_RD,
381         0, VMMETEROFF(v_ozfod), vcnt, "IU", "Optimized zero fill");
382 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swapin, CTLTYPE_UINT|CTLFLAG_RD,
383         0, VMMETEROFF(v_swapin), vcnt, "IU", "Swapin operations");
384 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swapout, CTLTYPE_UINT|CTLFLAG_RD,
385         0, VMMETEROFF(v_swapout), vcnt, "IU", "Swapout operations");
386 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swappgsin, CTLTYPE_UINT|CTLFLAG_RD,
387         0, VMMETEROFF(v_swappgsin), vcnt, "IU", "Swapin pages");
388 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swappgsout, CTLTYPE_UINT|CTLFLAG_RD,
389         0, VMMETEROFF(v_swappgsout), vcnt, "IU", "Swapout pages");
390 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodein, CTLTYPE_UINT|CTLFLAG_RD,
391         0, VMMETEROFF(v_vnodein), vcnt, "IU", "Vnodein operations");
392 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodeout, CTLTYPE_UINT|CTLFLAG_RD,
393         0, VMMETEROFF(v_vnodeout), vcnt, "IU", "Vnodeout operations");
394 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodepgsin, CTLTYPE_UINT|CTLFLAG_RD,
395         0, VMMETEROFF(v_vnodepgsin), vcnt, "IU", "Vnodein pages");
396 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodepgsout, CTLTYPE_UINT|CTLFLAG_RD,
397         0, VMMETEROFF(v_vnodepgsout), vcnt, "IU", "Vnodeout pages");
398 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_intrans, CTLTYPE_UINT|CTLFLAG_RD,
399         0, VMMETEROFF(v_intrans), vcnt, "IU", "In transit page blocking");
400 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_reactivated, CTLTYPE_UINT|CTLFLAG_RD,
401         0, VMMETEROFF(v_reactivated), vcnt, "IU", "Reactivated pages");
402 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pdwakeups, CTLTYPE_UINT|CTLFLAG_RD,
403         0, VMMETEROFF(v_pdwakeups), vcnt, "IU", "Pagedaemon wakeups");
404 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_ppwakeups, CTLTYPE_UINT|CTLFLAG_RD,
405         0, VMMETEROFF(v_ppwakeups), vcnt, "IU", "vm_wait wakeups");
406 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pdpages, CTLTYPE_UINT|CTLFLAG_RD,
407         0, VMMETEROFF(v_pdpages), vcnt, "IU", "Pagedaemon page scans");
408 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_dfree, CTLTYPE_UINT|CTLFLAG_RD,
409         0, VMMETEROFF(v_dfree), vcnt, "IU", "Pages freed by daemon");
410 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pfree, CTLTYPE_UINT|CTLFLAG_RD,
411         0, VMMETEROFF(v_pfree), vcnt, "IU", "Pages freed by exiting processes");
412 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_tfree, CTLTYPE_UINT|CTLFLAG_RD,
413         0, VMMETEROFF(v_tfree), vcnt, "IU", "Total pages freed");
414 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_forks, CTLTYPE_UINT|CTLFLAG_RD,
415         0, VMMETEROFF(v_forks), vcnt, "IU", "Number of fork() calls");
416 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vforks, CTLTYPE_UINT|CTLFLAG_RD,
417         0, VMMETEROFF(v_vforks), vcnt, "IU", "Number of vfork() calls");
418 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_rforks, CTLTYPE_UINT|CTLFLAG_RD,
419         0, VMMETEROFF(v_rforks), vcnt, "IU", "Number of rfork() calls");
420 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_kthreads, CTLTYPE_UINT|CTLFLAG_RD,
421         0, VMMETEROFF(v_kthreads), vcnt, "IU", "Number of fork() calls by kernel");
422 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_forkpages, CTLTYPE_UINT|CTLFLAG_RD,
423         0, VMMETEROFF(v_forkpages), vcnt, "IU", "VM pages affected by fork()");
424 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vforkpages, CTLTYPE_UINT|CTLFLAG_RD,
425         0, VMMETEROFF(v_vforkpages), vcnt, "IU", "VM pages affected by vfork()");
426 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_rforkpages, CTLTYPE_UINT|CTLFLAG_RD,
427         0, VMMETEROFF(v_rforkpages), vcnt, "IU", "VM pages affected by rfork()");
428 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_kthreadpages, CTLTYPE_UINT|CTLFLAG_RD,
429         0, VMMETEROFF(v_kthreadpages), vcnt, "IU", "VM pages affected by fork() by kernel");
430
431 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
432         v_page_size, CTLFLAG_RD, &vmstats.v_page_size, 0,
433         "Page size in bytes");
434 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
435         v_page_count, CTLFLAG_RD, &vmstats.v_page_count, 0, 
436         "Total number of pages in system");
437 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
438         v_free_reserved, CTLFLAG_RD, &vmstats.v_free_reserved, 0,
439         "Number of pages reserved for deadlock");
440 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
441         v_free_target, CTLFLAG_RD, &vmstats.v_free_target, 0,
442         "Number of pages desired free");
443 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
444         v_free_min, CTLFLAG_RD, &vmstats.v_free_min, 0,
445         "Minimum number of pages desired free");
446 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
447         v_free_count, CTLFLAG_RD, &vmstats.v_free_count, 0,
448         "Number of pages free");
449 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
450         v_wire_count, CTLFLAG_RD, &vmstats.v_wire_count, 0,
451         "Number of pages wired down");
452 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
453         v_active_count, CTLFLAG_RD, &vmstats.v_active_count, 0,
454         "Number of pages active");
455 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
456         v_inactive_target, CTLFLAG_RD, &vmstats.v_inactive_target, 0,
457         "Number of pages desired inactive");
458 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
459         v_inactive_count, CTLFLAG_RD, &vmstats.v_inactive_count, 0,
460         "Number of pages inactive");
461 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
462         v_cache_count, CTLFLAG_RD, &vmstats.v_cache_count, 0,
463         "Number of pages on buffer cache queue");
464 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
465         v_cache_min, CTLFLAG_RD, &vmstats.v_cache_min, 0,
466         "Min number of pages desired on cache queue");
467 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
468         v_cache_max, CTLFLAG_RD, &vmstats.v_cache_max, 0,
469         "Max number of pages in cached obj");
470 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
471         v_pageout_free_min, CTLFLAG_RD, &vmstats.v_pageout_free_min, 0,
472         "Min number pages reserved for kernel");
473 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
474         v_interrupt_free_min, CTLFLAG_RD, &vmstats.v_interrupt_free_min, 0,
475         "Reserved number of pages for int code");
476 SYSCTL_INT(_vm_stats_misc, OID_AUTO,
477         zero_page_count, CTLFLAG_RD, &vm_page_zero_count, 0,
478         "Number of zeroing pages");
479
480 /*
481  * No requirements.
482  */
483 static int
484 do_vmmeter_pcpu(SYSCTL_HANDLER_ARGS)
485 {
486         int boffset = offsetof(struct vmmeter, vmmeter_uint_begin);
487         int eoffset = offsetof(struct vmmeter, vmmeter_uint_end);
488         struct globaldata *gd = arg1;
489         struct vmmeter vmm;
490         int off;
491
492         bzero(&vmm, sizeof(vmm));
493         for (off = boffset; off <= eoffset; off += sizeof(u_int)) {
494                 *(u_int *)((char *)&vmm + off) +=
495                         *(u_int *)((char *)&gd->gd_cnt + off);
496         }
497         vmm.v_intr += vmm.v_ipi + vmm.v_timer;
498         return (sysctl_handle_opaque(oidp, &vmm, sizeof(vmm), req));
499 }
500
501 /*
502  * Called from the low level boot code only.
503  */
504 static void
505 vmmeter_init(void *dummy __unused)
506 {
507         int i;
508
509         for (i = 0; i < ncpus; ++i) {
510                 struct sysctl_ctx_list *ctx;
511                 struct sysctl_oid *oid;
512                 struct globaldata *gd;
513                 char name[32];
514
515                 ksnprintf(name, sizeof(name), "cpu%d", i);
516
517                 ctx = kmalloc(sizeof(*ctx), M_TEMP, M_WAITOK);
518                 sysctl_ctx_init(ctx);
519                 oid = SYSCTL_ADD_NODE(ctx, SYSCTL_STATIC_CHILDREN(_vm),
520                                       OID_AUTO, name, CTLFLAG_RD, 0, "");
521
522                 gd = globaldata_find(i);
523                 SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(oid), OID_AUTO,
524                                 "vmmeter", CTLTYPE_OPAQUE|CTLFLAG_RD,
525                                 gd, sizeof(struct vmmeter), do_vmmeter_pcpu,
526                                 "S,vmmeter", "System per-cpu statistics");
527         }
528 }
529 SYSINIT(vmmeter, SI_SUB_PSEUDO, SI_ORDER_ANY, vmmeter_init, 0);