profiling build fixes
[ghc.git] / rts / LdvProfile.c
1 /* -----------------------------------------------------------------------------
2 *
3 * (c) The GHC Team, 2001
4 * Author: Sungwoo Park
5 *
6 * Lag/Drag/Void profiling.
7 *
8 * ---------------------------------------------------------------------------*/
9
10 #ifdef PROFILING
11
12 #include "PosixSource.h"
13 #include "Rts.h"
14
15 #include "Profiling.h"
16 #include "LdvProfile.h"
17 #include "Stats.h"
18 #include "RtsUtils.h"
19 #include "Schedule.h"
20
21 /* --------------------------------------------------------------------------
22 * Fills in the slop when a *dynamic* closure changes its type.
23 * First calls LDV_recordDead() to declare the closure is dead, and then
24 * fills in the slop.
25 *
26 * Invoked when:
27 * 1) blackholing, UPD_BH_UPDATABLE() and UPD_BH_SINGLE_ENTRY (in
28 * includes/StgMacros.h), threadLazyBlackHole() and
29 * threadSqueezeStack() (in GC.c).
30 * 2) updating with indirection closures, updateWithIndirection()
31 * and updateWithPermIndirection() (in Storage.h).
32 *
33 * LDV_recordDead_FILL_SLOP_DYNAMIC() is not called on 'inherently used'
34 * closures such as TSO. It is not called on PAP because PAP is not updatable.
35 * ----------------------------------------------------------------------- */
36 void
37 LDV_recordDead_FILL_SLOP_DYNAMIC( StgClosure *p )
38 {
39 nat size, i;
40
41 #if defined(__GNUC__) && __GNUC__ < 3 && defined(DEBUG)
42 #error Please use gcc 3.0+ to compile this file with DEBUG; gcc < 3.0 miscompiles it
43 #endif
44
45 if (era > 0) {
46 // very like FILL_SLOP(), except that we call LDV_recordDead().
47 size = closure_sizeW(p);
48
49 LDV_recordDead((StgClosure *)(p), size);
50
51 if (size > sizeofW(StgThunkHeader)) {
52 for (i = 0; i < size - sizeofW(StgThunkHeader); i++) {
53 ((StgThunk *)(p))->payload[i] = 0;
54 }
55 }
56 }
57 }
58
59 /* --------------------------------------------------------------------------
60 * This function is called eventually on every object destroyed during
61 * a garbage collection, whether it is a major garbage collection or
62 * not. If c is an 'inherently used' closure, nothing happens. If c
63 * is an ordinary closure, LDV_recordDead() is called on c with its
64 * proper size which excludes the profiling header portion in the
65 * closure. Returns the size of the closure, including the profiling
66 * header portion, so that the caller can find the next closure.
67 * ----------------------------------------------------------------------- */
68 STATIC_INLINE nat
69 processHeapClosureForDead( StgClosure *c )
70 {
71 nat size;
72 const StgInfoTable *info;
73
74 info = get_itbl(c);
75
76 info = c->header.info;
77 if (IS_FORWARDING_PTR(info)) {
78 // The size of the evacuated closure is currently stored in
79 // the LDV field. See SET_EVACUAEE_FOR_LDV() in
80 // includes/StgLdvProf.h.
81 return LDVW(c);
82 }
83 info = INFO_PTR_TO_STRUCT(info);
84
85 ASSERT(((LDVW(c) & LDV_CREATE_MASK) >> LDV_SHIFT) <= era &&
86 ((LDVW(c) & LDV_CREATE_MASK) >> LDV_SHIFT) > 0);
87 ASSERT(((LDVW(c) & LDV_STATE_MASK) == LDV_STATE_CREATE) ||
88 (
89 (LDVW(c) & LDV_LAST_MASK) <= era &&
90 (LDVW(c) & LDV_LAST_MASK) > 0
91 ));
92
93
94 size = closure_sizeW(c);
95
96 switch (info->type) {
97 /*
98 'inherently used' cases: do nothing.
99 */
100 case TSO:
101 case MVAR_CLEAN:
102 case MVAR_DIRTY:
103 case MUT_ARR_PTRS_CLEAN:
104 case MUT_ARR_PTRS_DIRTY:
105 case MUT_ARR_PTRS_FROZEN:
106 case MUT_ARR_PTRS_FROZEN0:
107 case ARR_WORDS:
108 case WEAK:
109 case MUT_VAR_CLEAN:
110 case MUT_VAR_DIRTY:
111 case BCO:
112 case STABLE_NAME:
113 case TVAR_WATCH_QUEUE:
114 case TVAR:
115 case TREC_HEADER:
116 case TREC_CHUNK:
117 case INVARIANT_CHECK_QUEUE:
118 case ATOMIC_INVARIANT:
119 return size;
120
121 /*
122 ordinary cases: call LDV_recordDead().
123 */
124 case THUNK:
125 case THUNK_1_0:
126 case THUNK_0_1:
127 case THUNK_SELECTOR:
128 case THUNK_2_0:
129 case THUNK_1_1:
130 case THUNK_0_2:
131 case AP:
132 case PAP:
133 case AP_STACK:
134 case CONSTR:
135 case CONSTR_1_0:
136 case CONSTR_0_1:
137 case CONSTR_2_0:
138 case CONSTR_1_1:
139 case CONSTR_0_2:
140 case FUN:
141 case FUN_1_0:
142 case FUN_0_1:
143 case FUN_2_0:
144 case FUN_1_1:
145 case FUN_0_2:
146 case BLACKHOLE:
147 case CAF_BLACKHOLE:
148 case IND_PERM:
149 case IND_OLDGEN_PERM:
150 /*
151 'Ingore' cases
152 */
153 // Why can we ignore IND/IND_OLDGEN closures? We assume that
154 // any census is preceded by a major garbage collection, which
155 // IND/IND_OLDGEN closures cannot survive. Therefore, it is no
156 // use considering IND/IND_OLDGEN closures in the meanwhile
157 // because they will perish before the next census at any
158 // rate.
159 case IND:
160 case IND_OLDGEN:
161 // Found a dead closure: record its size
162 LDV_recordDead(c, size);
163 return size;
164
165 /*
166 Error case
167 */
168 // static objects
169 case IND_STATIC:
170 case CONSTR_STATIC:
171 case FUN_STATIC:
172 case THUNK_STATIC:
173 case CONSTR_NOCAF_STATIC:
174 // stack objects
175 case UPDATE_FRAME:
176 case CATCH_FRAME:
177 case STOP_FRAME:
178 case RET_DYN:
179 case RET_BCO:
180 case RET_SMALL:
181 case RET_BIG:
182 // others
183 case INVALID_OBJECT:
184 default:
185 barf("Invalid object in processHeapClosureForDead(): %d", info->type);
186 return 0;
187 }
188 }
189
190 /* --------------------------------------------------------------------------
191 * Calls processHeapClosureForDead() on every *dead* closures in the
192 * heap blocks starting at bd.
193 * ----------------------------------------------------------------------- */
194 static void
195 processHeapForDead( bdescr *bd )
196 {
197 StgPtr p;
198
199 while (bd != NULL) {
200 p = bd->start;
201 while (p < bd->free) {
202 p += processHeapClosureForDead((StgClosure *)p);
203 while (p < bd->free && !*p) // skip slop
204 p++;
205 }
206 ASSERT(p == bd->free);
207 bd = bd->link;
208 }
209 }
210
211 /* --------------------------------------------------------------------------
212 * Calls processHeapClosureForDead() on every *dead* closures in the nursery.
213 * ----------------------------------------------------------------------- */
214 static void
215 processNurseryForDead( void )
216 {
217 StgPtr p, bdLimit;
218 bdescr *bd;
219
220 bd = MainCapability.r.rNursery->blocks;
221 while (bd->start < bd->free) {
222 p = bd->start;
223 bdLimit = bd->start + BLOCK_SIZE_W;
224 while (p < bd->free && p < bdLimit) {
225 p += processHeapClosureForDead((StgClosure *)p);
226 while (p < bd->free && p < bdLimit && !*p) // skip slop
227 p++;
228 }
229 bd = bd->link;
230 if (bd == NULL)
231 break;
232 }
233 }
234
235 /* --------------------------------------------------------------------------
236 * Calls processHeapClosureForDead() on every *dead* closures in the
237 * small object pool.
238 * ----------------------------------------------------------------------- */
239 static void
240 processSmallObjectPoolForDead( void )
241 {
242 bdescr *bd;
243 StgPtr p;
244
245 for (bd = g0s0->blocks; bd != NULL; bd = bd->link) {
246 p = bd->start;
247 while (p < bd->free) {
248 p += processHeapClosureForDead((StgClosure *)p);
249 while (p < bd->free && !*p) // skip slop
250 p++;
251 }
252 ASSERT(p == bd->free);
253 }
254 }
255
256 /* --------------------------------------------------------------------------
257 * Calls processHeapClosureForDead() on every *dead* closures in the closure
258 * chain.
259 * ----------------------------------------------------------------------- */
260 static void
261 processChainForDead( bdescr *bd )
262 {
263 // Any object still in the chain is dead!
264 while (bd != NULL) {
265 if (!(bd->flags & BF_PINNED)) {
266 processHeapClosureForDead((StgClosure *)bd->start);
267 }
268 bd = bd->link;
269 }
270 }
271
272 /* --------------------------------------------------------------------------
273 * Start a census for *dead* closures, and calls
274 * processHeapClosureForDead() on every closure which died in the
275 * current garbage collection. This function is called from a garbage
276 * collector right before tidying up, when all dead closures are still
277 * stored in the heap and easy to identify. Generations 0 through N
278 * have just beed garbage collected.
279 * ----------------------------------------------------------------------- */
280 void
281 LdvCensusForDead( nat N )
282 {
283 nat g, s;
284
285 // ldvTime == 0 means that LDV profiling is currently turned off.
286 if (era == 0)
287 return;
288
289 if (RtsFlags.GcFlags.generations == 1) {
290 //
291 // Todo: support LDV for two-space garbage collection.
292 //
293 barf("Lag/Drag/Void profiling not supported with -G1");
294 } else {
295 for (g = 0; g <= N; g++)
296 for (s = 0; s < generations[g].n_steps; s++) {
297 if (g == 0 && s == 0) {
298 processSmallObjectPoolForDead();
299 processNurseryForDead();
300 processChainForDead(generations[g].steps[s].large_objects);
301 } else{
302 processHeapForDead(generations[g].steps[s].old_blocks);
303 processChainForDead(generations[g].steps[s].large_objects);
304 }
305 }
306 }
307 }
308
309 /* --------------------------------------------------------------------------
310 * Regard any closure in the current heap as dead or moribund and update
311 * LDV statistics accordingly.
312 * Called from shutdownHaskell() in RtsStartup.c.
313 * Also, stops LDV profiling by resetting ldvTime to 0.
314 * ----------------------------------------------------------------------- */
315 void
316 LdvCensusKillAll( void )
317 {
318 LdvCensusForDead(RtsFlags.GcFlags.generations - 1);
319 }
320
321 #endif /* PROFILING */