1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
|
/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License (the "License").
* You may not use this file except in compliance with the License.
*
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
* or http://www.opensolaris.org/os/licensing.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
* If applicable, add the following below this CDDL HEADER, with the
* fields enclosed by brackets "[]" replaced with your own identifying
* information: Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*/
/*
* Copyright 2006 Sun Microsystems, Inc. All rights reserved.
* Copyright 2011, 2012, Joyent, Inc. All rights reserved.
*/
/*
* This file implements the code which runs a thread inside zoneadmd to cap
* the associated zone's physical memory. A thread to do this is started
* when the zone boots and is halted when the zone shuts down.
*
* Because of the way that the VM system is currently implemented, there is no
* way to go from the bottom up (page to process to zone). Thus, there is no
* obvious way to hook an rctl into the kernel's paging code to enforce a hard
* memory cap. Instead, we implement a soft physical memory cap which looks
* at the zone's overall rss and once it is over the cap, works from the top
* down (zone to process to page), looking at zone processes, to determine
* what to try to pageout to get the zone under its memory cap.
*
* The code uses the vm_getusage syscall to determine the zone's rss and
* checks that against the zone's zone.max-physical-memory rctl. Once the
* zone goes over its cap, then this thread will work through the zone's
* /proc process list, Pgrab-bing each process and stepping through the
* address space segments attempting to use pr_memcntl(...MS_INVALCURPROC...)
* to pageout pages, until the zone is again under its cap.
*
* Although zone memory capping is implemented as a soft cap by this user-level
* thread, the interfaces around memory caps that are exposed to the user are
* the standard ones; an rctl and kstats. This thread uses the rctl value
* to obtain the cap and works with the zone kernel code to update the kstats.
* If the implementation ever moves into the kernel, these exposed interfaces
* do not need to change.
*
* The thread adaptively sleeps, periodically checking the state of the
* zone. As the zone's rss gets closer to the cap, the thread will wake up
* more often to check the zone's status. Once the zone is over the cap,
* the thread will work to pageout until the zone is under the cap, as shown
* by updated vm_usage data.
*
* NOTE: The pagedata page maps (at least on x86) are not useful. Those flags
* are set by hrm_setbits() and on x86 that code path is only executed by
* segvn_pagelock -> hat_setstat -> hrm_setbits
* segvn_softunlock -^
* On SPARC there is an additional code path which may make this data
* useful (sfmmu_ttesync), but since it is not generic, we ignore the page
* maps. If we ever fix this issue, then we could generalize this mcap code to
* do more with the data on active pages.
*
* For debugging, touch the file {zonepath}/mcap_debug.log. This will
* cause the thread to start logging its actions into that file (it may take
* a minute or two if the thread is currently sleeping). Removing that
* file will cause logging to stop.
*/
#include <sys/mman.h>
#include <sys/param.h>
#include <sys/stat.h>
#include <sys/types.h>
#include <assert.h>
#include <errno.h>
#include <fcntl.h>
#include <libproc.h>
#include <limits.h>
#include <procfs.h>
#include <stdio.h>
#include <stdlib.h>
#include <strings.h>
#include <time.h>
#include <unistd.h>
#include <sys/priocntl.h>
#include <dirent.h>
#include <zone.h>
#include <libzonecfg.h>
#include <thread.h>
#include <values.h>
#include <sys/vm_usage.h>
#include <sys/resource.h>
#include <sys/debug.h>
#include <synch.h>
#include <wait.h>
#include <libcontract.h>
#include <libcontract_priv.h>
#include <sys/contract/process.h>
#include "zoneadmd.h"
/* round up to next y = 2^n */
#define ROUNDUP(x, y) (((x) + ((y) - 1)) & ~((y) - 1))
#define CAP_REFRESH ((uint64_t)300 * NANOSEC) /* every 5 minutes */
static char zonename[ZONENAME_MAX];
static char zonepath[MAXPATHLEN];
static char zoneproc[MAXPATHLEN];
static char debug_log[MAXPATHLEN];
static zoneid_t zid;
static mutex_t shutdown_mx;
static cond_t shutdown_cv;
static int shutting_down = 0;
static thread_t mcap_tid;
static FILE *debug_log_fp = NULL;
static uint64_t zone_rss_cap; /* RSS cap(KB) */
static char over_cmd[2 * BUFSIZ]; /* same size as zone_attr_value */
/*
* Structure to hold current state about a process address space that we're
* working on.
*/
typedef struct {
int pr_curr; /* the # of the mapping we're working on */
int pr_nmap; /* number of mappings in address space */
prxmap_t *pr_xmapp; /* process's xmap array */
} proc_map_t;
typedef struct zsd_vmusage64 {
id_t vmu_zoneid;
uint_t vmu_type;
id_t vmu_id;
/*
* An amd64 kernel will align the following uint64_t members, but a
* 32bit i386 process will not without help.
*/
int vmu_align_next_members_on_8_bytes;
uint64_t vmu_rss_all;
uint64_t vmu_rss_private;
uint64_t vmu_rss_shared;
uint64_t vmu_swap_all;
uint64_t vmu_swap_private;
uint64_t vmu_swap_shared;
} zsd_vmusage64_t;
/*
* Output a debug log message.
*/
/*PRINTFLIKE1*/
static void
debug(char *fmt, ...)
{
va_list ap;
if (debug_log_fp == NULL)
return;
va_start(ap, fmt);
(void) vfprintf(debug_log_fp, fmt, ap);
va_end(ap);
(void) fflush(debug_log_fp);
}
/*
* Like sleep(3C) but can be interupted by cond_signal which is posted when
* we're shutting down the mcap thread.
*/
static void
sleep_shutdown(int secs)
{
timestruc_t to;
to.tv_sec = secs;
to.tv_nsec = 0;
(void) mutex_lock(&shutdown_mx);
if (!shutting_down)
(void) cond_reltimedwait(&shutdown_cv, &shutdown_mx, &to);
(void) mutex_unlock(&shutdown_mx);
}
static boolean_t
proc_issystem(pid_t pid)
{
char pc_clname[PC_CLNMSZ];
if (priocntl(P_PID, pid, PC_GETXPARMS, NULL, PC_KY_CLNAME, pc_clname,
PC_KY_NULL) != -1)
return (strcmp(pc_clname, "SYS") == 0);
return (B_TRUE);
}
/*
* Fork a child that enters the zone and runs the "phys-mcap-cmd" command.
*/
static void
run_over_cmd()
{
int ctfd;
int err;
pid_t childpid;
siginfo_t info;
ctid_t ct;
/*
* Before we enter the zone, we need to create a new process contract
* for the child, as required by zone_enter().
*/
if ((ctfd = open64("/system/contract/process/template", O_RDWR)) == -1)
return;
if (ct_tmpl_set_critical(ctfd, 0) != 0 ||
ct_tmpl_set_informative(ctfd, 0) != 0 ||
ct_pr_tmpl_set_fatal(ctfd, CT_PR_EV_HWERR) != 0 ||
ct_pr_tmpl_set_param(ctfd, CT_PR_PGRPONLY) != 0 ||
ct_tmpl_activate(ctfd) != 0) {
(void) close(ctfd);
return;
}
childpid = fork();
switch (childpid) {
case -1:
(void) ct_tmpl_clear(ctfd);
(void) close(ctfd);
break;
case 0: /* Child */
(void) ct_tmpl_clear(ctfd);
(void) close(ctfd);
if (zone_enter(zid) == -1)
_exit(errno);
err = system(over_cmd);
_exit(err);
break;
default: /* Parent */
if (contract_latest(&ct) == -1)
ct = -1;
(void) ct_tmpl_clear(ctfd);
(void) close(ctfd);
err = waitid(P_PID, childpid, &info, WEXITED);
(void) contract_abandon_id(ct);
if (err == -1 || info.si_status != 0)
debug("over_cmd failed");
break;
}
}
static struct ps_prochandle *
control_proc(pid_t pid)
{
int res;
struct ps_prochandle *ph;
/* Take control of the process. */
if ((ph = Pgrab(pid, 0, &res)) == NULL)
return (NULL);
if (Pcreate_agent(ph) != 0) {
(void) Prelease(ph, 0);
return (NULL);
}
/* Verify agent LWP is actually stopped. */
errno = 0;
while (Pstate(ph) == PS_RUN)
(void) Pwait(ph, 0);
if (Pstate(ph) != PS_STOP) {
Pdestroy_agent(ph);
(void) Prelease(ph, 0);
return (NULL);
}
return (ph);
}
/*
* Get the next mapping.
*/
static prxmap_t *
nextmapping(proc_map_t *pmp)
{
if (pmp->pr_xmapp == NULL || pmp->pr_curr >= pmp->pr_nmap)
return (NULL);
return (&pmp->pr_xmapp[pmp->pr_curr++]);
}
/*
* Initialize the proc_map_t to access the first mapping of an address space.
*/
static prxmap_t *
init_map(proc_map_t *pmp, pid_t pid)
{
int fd;
int res;
struct stat st;
char pathbuf[MAXPATHLEN];
bzero(pmp, sizeof (proc_map_t));
pmp->pr_nmap = -1;
(void) snprintf(pathbuf, sizeof (pathbuf), "%s/%d/xmap", zoneproc, pid);
if ((fd = open(pathbuf, O_RDONLY, 0)) < 0)
return (NULL);
redo:
errno = 0;
if (fstat(fd, &st) != 0)
goto done;
if ((pmp->pr_xmapp = malloc(st.st_size)) == NULL) {
debug("cannot malloc() %ld bytes for xmap", st.st_size);
goto done;
}
(void) bzero(pmp->pr_xmapp, st.st_size);
errno = 0;
if ((res = read(fd, pmp->pr_xmapp, st.st_size)) != st.st_size) {
free(pmp->pr_xmapp);
pmp->pr_xmapp = NULL;
if (res > 0 || errno == E2BIG) {
goto redo;
} else {
debug("pid %ld cannot read xmap\n", pid);
goto done;
}
}
pmp->pr_nmap = st.st_size / sizeof (prxmap_t);
done:
(void) close(fd);
return (nextmapping(pmp));
}
/*
* Attempt to page out a region of the given process's address space. May
* return nonzero if not all of the pages may are pageable, for any reason.
*/
static int
pageout_mapping(struct ps_prochandle *Pr, prxmap_t *pmp)
{
int res;
/*
* We particularly want to avoid the pr_memcntl on anonymous mappings
* which show 0 since that will pull them back off of the free list
* and increase the zone's RSS, even though the process itself has
* them freed up.
*/
if (pmp->pr_mflags & MA_ANON && pmp->pr_anon == 0)
return (0);
else if (pmp->pr_mflags & MA_ISM || pmp->pr_mflags & MA_SHM)
return (0);
errno = 0;
res = pr_memcntl(Pr, (caddr_t)pmp->pr_vaddr, pmp->pr_size, MC_SYNC,
(caddr_t)(MS_ASYNC | MS_INVALCURPROC), 0, 0);
/*
* EBUSY indicates none of the pages have backing store allocated, or
* some pages were locked. Don't care about this.
*/
if (res != 0 && errno == EBUSY)
res = 0;
return (res);
}
/*
* Compute the delta of the process RSS since the last call. If the
* psinfo cannot be obtained, no error is returned; its up to the caller to
* detect the process termination via other means.
*/
static int64_t
rss_delta(int64_t *old_rss, int psfd)
{
int64_t d_rss = 0;
psinfo_t psinfo;
if (pread(psfd, &psinfo, sizeof (psinfo_t), 0) == sizeof (psinfo_t)) {
d_rss = (int64_t)psinfo.pr_rssize - *old_rss;
*old_rss = (int64_t)psinfo.pr_rssize;
}
return (d_rss);
}
/*
* Work through a process paging out mappings until the whole address space was
* examined or the excess is < 0. Return our estimate of the updated excess.
*/
static int64_t
pageout_process(pid_t pid, int64_t excess)
{
int psfd;
prxmap_t *pxmap;
proc_map_t cur;
struct ps_prochandle *ph = NULL;
int unpageable_mappings;
int64_t sum_d_rss, sum_att, d_rss;
int64_t old_rss;
psinfo_t psinfo;
int incr_rss_check = 0;
char pathbuf[MAXPATHLEN];
(void) snprintf(pathbuf, sizeof (pathbuf), "%s/%d/psinfo", zoneproc,
pid);
if ((psfd = open(pathbuf, O_RDONLY, 0000)) < 0)
return (excess);
if (pread(psfd, &psinfo, sizeof (psinfo), 0) != sizeof (psinfo))
goto done;
old_rss = (int64_t)psinfo.pr_rssize;
/* If unscannable, skip it. */
if (psinfo.pr_nlwp == 0 || proc_issystem(pid)) {
debug("pid: %ld system process, skipping %s\n",
pid, psinfo.pr_psargs);
goto done;
}
/* If tiny RSS (16KB), skip it. */
if (old_rss <= 16) {
debug("pid: %ld skipping, RSS %lldKB %s\n",
pid, old_rss, psinfo.pr_psargs);
goto done;
}
/* Get segment residency information. */
pxmap = init_map(&cur, pid);
/* Skip process if it has no mappings. */
if (pxmap == NULL) {
debug("%ld: xmap unreadable; ignoring\n", pid);
goto done;
}
debug("pid %ld: nmap %d sz %dKB rss %lldKB %s\n",
pid, cur.pr_nmap, psinfo.pr_size, old_rss, psinfo.pr_psargs);
/* Take control of the process. */
if ((ph = control_proc(pid)) == NULL) {
debug("%ld: cannot control\n", pid);
goto done;
}
/*
* If the process RSS is not enough to erase the excess then no need
* to incrementally check the RSS delta after each pageout attempt.
* Instead check it after we've tried all of the segements.
*/
if (excess - old_rss < 0)
incr_rss_check = 1;
/*
* Within the process's address space, attempt to page out mappings.
*/
sum_att = sum_d_rss = 0;
unpageable_mappings = 0;
while (excess > 0 && pxmap != NULL && !shutting_down) {
/* Try to page out the mapping. */
if (pageout_mapping(ph, pxmap) < 0) {
debug("pid %ld: exited or unpageable\n", pid);
break;
}
/* attempted is the size of the mapping */
sum_att += pxmap->pr_size / 1024;
/*
* This processes RSS is potentially enough to clear the
* excess so check as we go along to see if we can stop
* paging out partway through the process.
*/
if (incr_rss_check) {
d_rss = rss_delta(&old_rss, psfd);
/*
* If this pageout attempt was unsuccessful (the
* resident portion was not affected), then note it was
* unpageable. Mappings are unpageable when none of the
* pages paged out, such as when they are locked, or
* involved in asynchronous I/O.
*/
if (d_rss >= 0) {
unpageable_mappings++;
} else {
excess += d_rss;
sum_d_rss += d_rss;
}
}
pxmap = nextmapping(&cur);
}
if (!incr_rss_check) {
d_rss = rss_delta(&old_rss, psfd);
if (d_rss < 0) {
excess += d_rss;
sum_d_rss += d_rss;
}
}
debug("pid %ld: unp %d att %lluKB drss %lldKB excess %lldKB\n",
pid, unpageable_mappings, (unsigned long long)sum_att,
(unsigned long long)sum_d_rss, (long long)excess);
done:
/* If a process was grabbed, release it, destroying its agent. */
if (ph != NULL) {
Pdestroy_agent(ph);
(void) Prelease(ph, 0);
}
if (cur.pr_xmapp != NULL)
free(cur.pr_xmapp);
(void) close(psfd);
if (shutting_down)
return (0);
return (excess);
}
/*
* Get the zone's RSS data.
*/
static uint64_t
get_mem_info(int age)
{
uint64_t n = 1;
zsd_vmusage64_t buf;
uint64_t zone_rss;
buf.vmu_id = zid;
if (syscall(SYS_rusagesys, _RUSAGESYS_GETVMUSAGE, VMUSAGE_A_ZONE,
age, (uintptr_t)&buf, (uintptr_t)&n) != 0) {
debug("vmusage failed\n");
(void) sleep_shutdown(1);
return (0);
}
if (n > 1) {
/* This should never happen */
debug("vmusage returned more than one result\n");
(void) sleep_shutdown(1);
return (0);
}
if (buf.vmu_id != zid) {
/* This should never happen */
debug("vmusage returned the incorrect zone\n");
(void) sleep_shutdown(1);
return (0);
}
zone_rss = buf.vmu_rss_all / 1024;
return (zone_rss);
}
/*
* Needed to read the zones physical-memory-cap rctl.
*/
static struct ps_prochandle *
grab_zone_proc()
{
DIR *dirp;
struct dirent *dentp;
struct ps_prochandle *ph = NULL;
int tmp;
if ((dirp = opendir(zoneproc)) == NULL)
return (NULL);
while (!shutting_down && (dentp = readdir(dirp))) {
int pid;
if (strcmp(".", dentp->d_name) == 0 ||
strcmp("..", dentp->d_name) == 0)
continue;
pid = atoi(dentp->d_name);
/* attempt to grab process */
if ((ph = Pgrab(pid, 0, &tmp)) != NULL) {
if (Psetflags(ph, PR_RLC) == 0) {
if (Pcreate_agent(ph) == 0) {
(void) closedir(dirp);
return (ph);
}
}
Prelease(ph, 0);
}
}
(void) closedir(dirp);
return (NULL);
}
static uint64_t
get_zone_cap()
{
rctlblk_t *rblk;
uint64_t mcap;
struct ps_prochandle *ph;
if ((rblk = (rctlblk_t *)malloc(rctlblk_size())) == NULL)
return (UINT64_MAX);
if ((ph = grab_zone_proc()) == NULL) {
free(rblk);
return (UINT64_MAX);
}
if (pr_getrctl(ph, "zone.max-physical-memory", NULL, rblk,
RCTL_FIRST)) {
Pdestroy_agent(ph);
Prelease(ph, 0);
free(rblk);
return (UINT64_MAX);
}
Pdestroy_agent(ph);
Prelease(ph, 0);
mcap = rctlblk_get_value(rblk);
free(rblk);
return (mcap);
}
/*
* check_suspend is invoked at the beginning of every pass through the process
* list or after we've paged out enough so that we think the excess is under
* the cap. The purpose is to periodically check the zone's rss and return
* the excess when the zone is over the cap. The rest of the time this
* function will sleep, periodically waking up to check the current rss.
*
* The age parameter is used to tell us how old the cached rss data can be.
* When first starting up, the cached data can be older, but after we
* start paging out, we want current data.
*
* Depending on the percentage of penetration of the zone's rss into the
* cap we sleep for longer or shorter amounts and accept older cached
* vmusage data. This reduces the impact of this work on the system, which
* is important considering that each zone will be monitoring its rss.
*/
static int64_t
check_suspend(int age, boolean_t new_cycle)
{
static hrtime_t last_cap_read = 0;
static uint64_t addon;
static uint64_t lo_thresh; /* Thresholds for how long to sleep */
static uint64_t hi_thresh; /* when under the cap (80% & 90%). */
static uint64_t prev_zone_rss = 0;
/* Wait a second to give the async pageout a chance to catch up. */
(void) sleep_shutdown(1);
while (!shutting_down) {
int64_t new_excess;
int sleep_time;
hrtime_t now;
struct stat st;
uint64_t zone_rss; /* total RSS(KB) */
/*
* Check if the debug log files exists and enable or disable
* debug.
*/
if (debug_log_fp == NULL) {
if (stat(debug_log, &st) == 0)
debug_log_fp = fopen(debug_log, "w");
} else {
if (stat(debug_log, &st) == -1) {
(void) fclose(debug_log_fp);
debug_log_fp = NULL;
}
}
/*
* If the CAP_REFRESH interval has passed, re-get the current
* cap in case it has been dynamically updated.
*/
now = gethrtime();
if (now - last_cap_read > CAP_REFRESH) {
uint64_t mcap;
last_cap_read = now;
mcap = get_zone_cap();
if (mcap != 0 && mcap != UINT64_MAX)
zone_rss_cap = ROUNDUP(mcap, 1024) / 1024;
else
zone_rss_cap = UINT64_MAX;
lo_thresh = (uint64_t)(zone_rss_cap * .8);
hi_thresh = (uint64_t)(zone_rss_cap * .9);
addon = (uint64_t)(zone_rss_cap * 0.05);
debug("current cap %lluKB lo %lluKB hi %lluKB\n",
zone_rss_cap, lo_thresh, hi_thresh);
}
/* No cap, nothing to do. */
if (zone_rss_cap == 0 || zone_rss_cap == UINT64_MAX) {
debug("no cap, sleep 120 seconds\n");
(void) sleep_shutdown(120);
continue;
}
zone_rss = get_mem_info(age);
/* calculate excess */
new_excess = zone_rss - zone_rss_cap;
debug("rss %lluKB, cap %lluKB, excess %lldKB\n",
zone_rss, zone_rss_cap, new_excess);
/*
* If necessary, updates stats.
*/
/*
* If it looks like we did some paging out since last over the
* cap then update the kstat so we can approximate how much was
* paged out.
*/
if (prev_zone_rss > zone_rss_cap && zone_rss < prev_zone_rss) {
uint64_t diff;
/* assume diff is num bytes we paged out */
diff = (prev_zone_rss - zone_rss) * 1024;
(void) zone_setattr(zid, ZONE_ATTR_PMCAP_PAGEOUT,
&diff, 0);
}
prev_zone_rss = zone_rss;
if (new_excess > 0) {
if (new_cycle) {
uint64_t n = 1;
/* Increment "nover" kstat. */
(void) zone_setattr(zid, ZONE_ATTR_PMCAP_NOVER,
&n, 0);
}
/*
* Once we go over the cap, then we want to
* page out a little extra instead of stopping
* right at the cap. To do this we add 5% to
* the excess so that pageout_proces will work
* a little longer before stopping.
*/
return ((int64_t)(new_excess + addon));
}
/*
* At this point we are under the cap.
*
* Scale the amount of time we sleep before rechecking the
* zone's memory usage. Also, scale the accpetable age of
* cached results from vm_getusage. We do this based on the
* penetration into the capped limit.
*/
if (zone_rss <= lo_thresh) {
sleep_time = 120;
age = 15;
} else if (zone_rss <= hi_thresh) {
sleep_time = 60;
age = 10;
} else {
sleep_time = 30;
age = 5;
}
debug("sleep %d seconds\n", sleep_time);
(void) sleep_shutdown(sleep_time);
}
return (0);
}
/*
* Thread that checks zone's memory usage and when over the cap, goes through
* the zone's process list trying to pageout processes to get under the cap.
*/
static void
mcap_zone()
{
DIR *pdir = NULL;
int age = 10; /* initial cached vmusage can be 10 secs. old */
int64_t excess;
debug("thread startup\n");
/*
* When first starting it is likely lots of other zones are starting
* too because the system is booting. Since we just started the zone
* we're not worried about being over the cap right away, so we let
* things settle a bit and tolerate some older data here to minimize
* the load on the system.
*/
(void) sleep_shutdown(15); /* wait 15 secs. so the zone can get going */
/* Wait until zone's /proc is mounted */
while (!shutting_down) {
struct stat st;
if (stat(zoneproc, &st) == 0 &&
strcmp(st.st_fstype, "proc") == 0)
break;
sleep_shutdown(5);
}
/* Open zone's /proc and walk entries. */
while (!shutting_down) {
if ((pdir = opendir(zoneproc)) != NULL)
break;
sleep_shutdown(5);
}
while (!shutting_down) {
struct dirent *dirent;
/* Wait until we've gone over the cap. */
excess = check_suspend(age, B_TRUE);
debug("starting to scan, excess %lldk\n", (long long)excess);
/*
* After the initial startup, we want the age of the cached
* vmusage to be only 1 second old since we are checking
* the current state after we've gone over the cap and have
* paged out some processes.
*/
age = 1;
if (over_cmd[0] != '\0') {
uint64_t zone_rss; /* total RSS(KB) */
debug("run phys_mcap_cmd: %s\n", over_cmd);
run_over_cmd();
zone_rss = get_mem_info(0);
excess = zone_rss - zone_rss_cap;
debug("rss %lluKB, cap %lluKB, excess %lldKB\n",
zone_rss, zone_rss_cap, excess);
if (excess <= 0)
continue;
}
while (!shutting_down && (dirent = readdir(pdir)) != NULL) {
pid_t pid;
if (strcmp(".", dirent->d_name) == 0 ||
strcmp("..", dirent->d_name) == 0)
continue;
pid = atoi(dirent->d_name);
if (pid == 0 || pid == 1)
continue;
excess = pageout_process(pid, excess);
if (excess <= 0) {
debug("apparently under; excess %lld\n",
(long long)excess);
/* Double check the current excess */
excess = check_suspend(1, B_FALSE);
}
}
debug("process pass done; excess %lld\n", (long long)excess);
rewinddir(pdir);
}
if (pdir != NULL)
(void) closedir(pdir);
debug("thread shutdown\n");
}
static void
get_over_cmd()
{
zone_dochandle_t handle;
struct zone_attrtab attr;
over_cmd[0] = '\0';
if ((handle = zonecfg_init_handle()) == NULL)
return;
if (zonecfg_get_handle(zonename, handle) != Z_OK)
goto done;
if (zonecfg_setattrent(handle) != Z_OK)
goto done;
while (zonecfg_getattrent(handle, &attr) == Z_OK) {
if (strcmp("phys-mcap-cmd", attr.zone_attr_name) != 0)
continue; /* no match */
(void) strlcpy(over_cmd, attr.zone_attr_value,
sizeof (over_cmd));
break;
}
(void) zonecfg_endattrent(handle);
done:
zonecfg_fini_handle(handle);
}
void
create_mcap_thread(zlog_t *zlogp, zoneid_t id)
{
int res;
shutting_down = 0;
zid = id;
(void) getzonenamebyid(zid, zonename, sizeof (zonename));
if (zone_get_zonepath(zonename, zonepath, sizeof (zonepath)) != 0)
zerror(zlogp, B_FALSE, "zone %s missing zonepath", zonename);
(void) snprintf(zoneproc, sizeof (zoneproc), "%s/root/proc", zonepath);
(void) snprintf(debug_log, sizeof (debug_log), "%s/mcap_debug.log",
zonepath);
get_over_cmd();
res = thr_create(NULL, NULL, (void *(*)(void *))mcap_zone, NULL, NULL,
&mcap_tid);
if (res != 0) {
zerror(zlogp, B_FALSE, "error %d creating memory cap thread",
res);
mcap_tid = 0;
}
}
void
destroy_mcap_thread()
{
if (mcap_tid != 0) {
shutting_down = 1;
(void) cond_signal(&shutdown_cv);
(void) thr_join(mcap_tid, NULL, NULL);
mcap_tid = 0;
}
}
|