Line | % of fetches | Source |
---|---|---|
1 | /******************************************************************************/ | |
2 | #ifdef JEMALLOC_H_TYPES | |
3 | ||
4 | typedef struct tcache_bin_info_s tcache_bin_info_t; | |
5 | typedef struct tcache_bin_s tcache_bin_t; | |
6 | typedef struct tcache_s tcache_t; | |
7 | ||
8 | /* | |
9 | * tcache pointers close to NULL are used to encode state information that is | |
10 | * used for two purposes: preventing thread caching on a per thread basis and | |
11 | * cleaning up during thread shutdown. | |
12 | */ | |
13 | #define TCACHE_STATE_DISABLED ((tcache_t *)(uintptr_t)1) | |
14 | #define TCACHE_STATE_REINCARNATED ((tcache_t *)(uintptr_t)2) | |
15 | #define TCACHE_STATE_PURGATORY ((tcache_t *)(uintptr_t)3) | |
16 | #define TCACHE_STATE_MAX TCACHE_STATE_PURGATORY | |
17 | ||
18 | /* | |
19 | * Absolute maximum number of cache slots for each small bin in the thread | |
20 | * cache. This is an additional constraint beyond that imposed as: twice the | |
21 | * number of regions per run for this size class. | |
22 | * | |
23 | * This constant must be an even number. | |
24 | */ | |
25 | #define TCACHE_NSLOTS_SMALL_MAX 200 | |
26 | ||
27 | /* Number of cache slots for large size classes. */ | |
28 | #define TCACHE_NSLOTS_LARGE 20 | |
29 | ||
30 | /* (1U << opt_lg_tcache_max) is used to compute tcache_maxclass. */ | |
31 | #define LG_TCACHE_MAXCLASS_DEFAULT 15 | |
32 | ||
33 | /* | |
34 | * TCACHE_GC_SWEEP is the approximate number of allocation events between | |
35 | * full GC sweeps. Integer rounding may cause the actual number to be | |
36 | * slightly higher, since GC is performed incrementally. | |
37 | */ | |
38 | #define TCACHE_GC_SWEEP 8192 | |
39 | ||
40 | /* Number of tcache allocation/deallocation events between incremental GCs. */ | |
41 | #define TCACHE_GC_INCR \ | |
42 | ((TCACHE_GC_SWEEP / NBINS) + ((TCACHE_GC_SWEEP / NBINS == 0) ? 0 : 1)) | |
43 | ||
44 | #endif /* JEMALLOC_H_TYPES */ | |
45 | /******************************************************************************/ | |
46 | #ifdef JEMALLOC_H_STRUCTS | |
47 | ||
48 | typedef enum { | |
49 | tcache_enabled_false = 0, /* Enable cast to/from bool. */ | |
50 | tcache_enabled_true = 1, | |
51 | tcache_enabled_default = 2 | |
52 | } tcache_enabled_t; | |
53 | ||
54 | /* | |
55 | * Read-only information associated with each element of tcache_t's tbins array | |
56 | * is stored separately, mainly to reduce memory usage. | |
57 | */ | |
58 | struct tcache_bin_info_s { | |
59 | unsigned ncached_max; /* Upper limit on ncached. */ | |
60 | }; | |
61 | ||
62 | struct tcache_bin_s { | |
63 | tcache_bin_stats_t tstats; | |
64 | int low_water; /* Min # cached since last GC. */ | |
65 | unsigned lg_fill_div; /* Fill (ncached_max >> lg_fill_div). */ | |
66 | unsigned ncached; /* # of cached objects. */ | |
67 | void **avail; /* Stack of available objects. */ | |
68 | }; | |
69 | ||
70 | struct tcache_s { | |
71 | ql_elm(tcache_t) link; /* Used for aggregating stats. */ | |
72 | uint64_t prof_accumbytes;/* Cleared after arena_prof_accum() */ | |
73 | arena_t *arena; /* This thread's arena. */ | |
74 | unsigned ev_cnt; /* Event count since incremental GC. */ | |
75 | unsigned next_gc_bin; /* Next bin to GC. */ | |
76 | tcache_bin_t tbins[1]; /* Dynamically sized. */ | |
77 | /* | |
78 | * The pointer stacks associated with tbins follow as a contiguous | |
79 | * array. During tcache initialization, the avail pointer in each | |
80 | * element of tbins is initialized to point to the proper offset within | |
81 | * this array. | |
82 | */ | |
83 | }; | |
84 | ||
85 | #endif /* JEMALLOC_H_STRUCTS */ | |
86 | /******************************************************************************/ | |
87 | #ifdef JEMALLOC_H_EXTERNS | |
88 | ||
89 | extern bool opt_tcache; | |
90 | extern ssize_t opt_lg_tcache_max; | |
91 | ||
92 | extern tcache_bin_info_t *tcache_bin_info; | |
93 | ||
94 | /* | |
95 | * Number of tcache bins. There are NBINS small-object bins, plus 0 or more | |
96 | * large-object bins. | |
97 | */ | |
98 | extern size_t nhbins; | |
99 | ||
100 | /* Maximum cached size class. */ | |
101 | extern size_t tcache_maxclass; | |
102 | ||
103 | size_t tcache_salloc(const void *ptr); | |
104 | void tcache_event_hard(tcache_t *tcache); | |
105 | void *tcache_alloc_small_hard(tcache_t *tcache, tcache_bin_t *tbin, | |
106 | size_t binind); | |
107 | void tcache_bin_flush_small(tcache_bin_t *tbin, size_t binind, unsigned rem, | |
108 | tcache_t *tcache); | |
109 | void tcache_bin_flush_large(tcache_bin_t *tbin, size_t binind, unsigned rem, | |
110 | tcache_t *tcache); | |
111 | void tcache_arena_associate(tcache_t *tcache, arena_t *arena); | |
112 | void tcache_arena_dissociate(tcache_t *tcache); | |
113 | tcache_t *tcache_create(arena_t *arena); | |
114 | void tcache_destroy(tcache_t *tcache); | |
115 | void tcache_thread_cleanup(void *arg); | |
116 | void tcache_stats_merge(tcache_t *tcache, arena_t *arena); | |
117 | bool tcache_boot0(void); | |
118 | bool tcache_boot1(void); | |
119 | ||
120 | #endif /* JEMALLOC_H_EXTERNS */ | |
121 | /******************************************************************************/ | |
122 | #ifdef JEMALLOC_H_INLINES | |
123 | ||
124 | #ifndef JEMALLOC_ENABLE_INLINE | |
125 | malloc_tsd_protos(JEMALLOC_ATTR(unused), tcache, tcache_t *) | |
126 | malloc_tsd_protos(JEMALLOC_ATTR(unused), tcache_enabled, tcache_enabled_t) | |
127 | ||
128 | void tcache_event(tcache_t *tcache); | |
129 | void tcache_flush(void); | |
130 | bool tcache_enabled_get(void); | |
131 | tcache_t *tcache_get(bool create); | |
132 | void tcache_enabled_set(bool enabled); | |
133 | void *tcache_alloc_easy(tcache_bin_t *tbin); | |
134 | void *tcache_alloc_small(tcache_t *tcache, size_t size, bool zero); | |
135 | void *tcache_alloc_large(tcache_t *tcache, size_t size, bool zero); | |
136 | void tcache_dalloc_small(tcache_t *tcache, void *ptr, size_t binind); | |
137 | void tcache_dalloc_large(tcache_t *tcache, void *ptr, size_t size); | |
138 | #endif | |
139 | ||
140 | #if (defined(JEMALLOC_ENABLE_INLINE) || defined(JEMALLOC_TCACHE_C_)) | |
141 | /* Map of thread-specific caches. */ | |
142 | malloc_tsd_externs(tcache, tcache_t *) | |
143 | malloc_tsd_funcs(JEMALLOC_ALWAYS_INLINE, tcache, tcache_t *, NULL, | |
144 | tcache_thread_cleanup) | |
145 | /* Per thread flag that allows thread caches to be disabled. */ | |
146 | malloc_tsd_externs(tcache_enabled, tcache_enabled_t) | |
147 | malloc_tsd_funcs(JEMALLOC_ALWAYS_INLINE, tcache_enabled, tcache_enabled_t, | |
148 | tcache_enabled_default, malloc_tsd_no_cleanup) | |
149 | ||
150 | JEMALLOC_INLINE void | |
151 | tcache_flush(void) | |
152 | { | |
153 | tcache_t *tcache; | |
154 | ||
155 | cassert(config_tcache); | |
156 | ||
157 | tcache = *tcache_tsd_get(); | |
158 | if ((uintptr_t)tcache <= (uintptr_t)TCACHE_STATE_MAX) | |
159 | return; | |
160 | tcache_destroy(tcache); | |
161 | tcache = NULL; | |
162 | tcache_tsd_set(&tcache); | |
163 | } | |
164 | ||
165 | JEMALLOC_INLINE bool | |
166 | tcache_enabled_get(void) | |
167 | { | |
168 | tcache_enabled_t tcache_enabled; | |
169 | ||
170 | cassert(config_tcache); | |
171 | ||
172 | tcache_enabled = *tcache_enabled_tsd_get(); | |
173 | if (tcache_enabled == tcache_enabled_default) { | |
174 | tcache_enabled = (tcache_enabled_t)opt_tcache; | |
175 | tcache_enabled_tsd_set(&tcache_enabled); | |
176 | } | |
177 | ||
178 | return ((bool)tcache_enabled); | |
179 | } | |
180 | ||
181 | JEMALLOC_INLINE void | |
182 | tcache_enabled_set(bool enabled) | |
183 | { | |
184 | tcache_enabled_t tcache_enabled; | |
185 | tcache_t *tcache; | |
186 | ||
187 | cassert(config_tcache); | |
188 | ||
189 | tcache_enabled = (tcache_enabled_t)enabled; | |
190 | tcache_enabled_tsd_set(&tcache_enabled); | |
191 | tcache = *tcache_tsd_get(); | |
192 | if (enabled) { | |
193 | if (tcache == TCACHE_STATE_DISABLED) { | |
194 | tcache = NULL; | |
195 | tcache_tsd_set(&tcache); | |
196 | } | |
197 | } else /* disabled */ { | |
198 | if (tcache > TCACHE_STATE_MAX) { | |
199 | tcache_destroy(tcache); | |
200 | tcache = NULL; | |
201 | } | |
202 | if (tcache == NULL) { | |
203 | tcache = TCACHE_STATE_DISABLED; | |
204 | tcache_tsd_set(&tcache); | |
205 | } | |
206 | } | |
207 | } | |
208 | ||
209 | JEMALLOC_ALWAYS_INLINE tcache_t * | |
210 | tcache_get(bool create) | |
211 | { | |
212 | tcache_t *tcache; | |
213 | ||
214 | if (config_tcache == false) | |
215 | return (NULL); | |
216 | if (config_lazy_lock && isthreaded == false) | |
217 | return (NULL); | |
218 | ||
219 | tcache = *tcache_tsd_get(); | |
220 | if ((uintptr_t)tcache <= (uintptr_t)TCACHE_STATE_MAX) { | |
221 | if (tcache == TCACHE_STATE_DISABLED) | |
222 | return (NULL); | |
223 | if (tcache == NULL) { | |
224 | if (create == false) { | |
225 | /* | |
226 | * Creating a tcache here would cause | |
227 | * allocation as a side effect of free(). | |
228 | * Ordinarily that would be okay since | |
229 | * tcache_create() failure is a soft failure | |
230 | * that doesn't propagate. However, if TLS | |
231 | * data are freed via free() as in glibc, | |
232 | * subtle corruption could result from setting | |
233 | * a TLS variable after its backing memory is | |
234 | * freed. | |
235 | */ | |
236 | return (NULL); | |
237 | } | |
238 | if (tcache_enabled_get() == false) { | |
239 | tcache_enabled_set(false); /* Memoize. */ | |
240 | return (NULL); | |
241 | } | |
242 | return (tcache_create(choose_arena(NULL))); | |
243 | } | |
244 | if (tcache == TCACHE_STATE_PURGATORY) { | |
245 | /* | |
246 | * Make a note that an allocator function was called | |
247 | * after tcache_thread_cleanup() was called. | |
248 | */ | |
249 | tcache = TCACHE_STATE_REINCARNATED; | |
250 | tcache_tsd_set(&tcache); | |
251 | return (NULL); | |
252 | } | |
253 | if (tcache == TCACHE_STATE_REINCARNATED) | |
254 | return (NULL); | |
255 | not_reached(); | |
256 | } | |
257 | ||
258 | return (tcache); | |
259 | } | |
260 | ||
261 | JEMALLOC_ALWAYS_INLINE void | |
262 | tcache_event(tcache_t *tcache) | |
263 | { | |
264 | ||
265 | if (TCACHE_GC_INCR == 0) | |
266 | return; | |
267 | ||
268 | tcache->ev_cnt++; | |
269 | assert(tcache->ev_cnt <= TCACHE_GC_INCR); | |
270 | if (tcache->ev_cnt == TCACHE_GC_INCR) | |
271 | tcache_event_hard(tcache); | |
272 | } | |
273 | ||
274 | JEMALLOC_ALWAYS_INLINE void * | |
275 | tcache_alloc_easy(tcache_bin_t *tbin) | |
276 | { | |
277 | void *ret; | |
278 | ||
279 | if (tbin->ncached == 0) { | |
280 | tbin->low_water = -1; | |
281 | return (NULL); | |
282 | } | |
283 | tbin->ncached--; | |
284 | if ((int)tbin->ncached < tbin->low_water) | |
285 | tbin->low_water = tbin->ncached; | |
286 | ret = tbin->avail[tbin->ncached]; | |
287 | return (ret); | |
288 | } | |
289 | ||
290 | JEMALLOC_ALWAYS_INLINE void * | |
291 | tcache_alloc_small(tcache_t *tcache, size_t size, bool zero) | |
292 | { | |
293 | void *ret; | |
294 | size_t binind; | |
295 | tcache_bin_t *tbin; | |
296 | ||
297 | binind = SMALL_SIZE2BIN(size); | |
298 | assert(binind < NBINS); | |
299 | tbin = &tcache->tbins[binind]; | |
300 | size = arena_bin_info[binind].reg_size; | |
301 | ret = tcache_alloc_easy(tbin); | |
302 | if (ret == NULL) { | |
303 | ret = tcache_alloc_small_hard(tcache, tbin, binind); | |
304 | if (ret == NULL) | |
305 | return (NULL); | |
306 | } | |
307 | assert(tcache_salloc(ret) == arena_bin_info[binind].reg_size); | |
308 | ||
309 | if (zero == false) { | |
310 | if (config_fill) { | |
311 | if (opt_junk) { | |
312 | arena_alloc_junk_small(ret, | |
313 | &arena_bin_info[binind], false); | |
314 | } else if (opt_zero) | |
315 | memset(ret, 0, size); | |
316 | } | |
317 | VALGRIND_MAKE_MEM_UNDEFINED(ret, size); | |
318 | } else { | |
319 | if (config_fill && opt_junk) { | |
320 | arena_alloc_junk_small(ret, &arena_bin_info[binind], | |
321 | true); | |
322 | } | |
323 | VALGRIND_MAKE_MEM_UNDEFINED(ret, size); | |
324 | memset(ret, 0, size); | |
325 | } | |
326 | ||
327 | if (config_stats) | |
328 | tbin->tstats.nrequests++; | |
329 | if (config_prof) | |
330 | tcache->prof_accumbytes += arena_bin_info[binind].reg_size; | |
331 | tcache_event(tcache); | |
332 | return (ret); | |
333 | } | |
334 | ||
335 | JEMALLOC_ALWAYS_INLINE void * | |
336 | tcache_alloc_large(tcache_t *tcache, size_t size, bool zero) | |
337 | { | |
338 | void *ret; | |
339 | size_t binind; | |
340 | tcache_bin_t *tbin; | |
341 | ||
342 | size = PAGE_CEILING(size); | |
343 | assert(size <= tcache_maxclass); | |
344 | binind = NBINS + (size >> LG_PAGE) - 1; | |
345 | assert(binind < nhbins); | |
346 | tbin = &tcache->tbins[binind]; | |
347 | ret = tcache_alloc_easy(tbin); | |
348 | if (ret == NULL) { | |
349 | /* | |
350 | * Only allocate one large object at a time, because it's quite | |
351 | * expensive to create one and not use it. | |
352 | */ | |
353 | ret = arena_malloc_large(tcache->arena, size, zero); | |
354 | if (ret == NULL) | |
355 | return (NULL); | |
356 | } else { | |
357 | if (config_prof && prof_promote && size == PAGE) { | |
358 | arena_chunk_t *chunk = | |
359 | (arena_chunk_t *)CHUNK_ADDR2BASE(ret); | |
360 | size_t pageind = (((uintptr_t)ret - (uintptr_t)chunk) >> | |
361 | LG_PAGE); | |
362 | arena_mapbits_large_binind_set(chunk, pageind, | |
363 | BININD_INVALID); | |
364 | } | |
365 | if (zero == false) { | |
366 | if (config_fill) { | |
367 | if (opt_junk) | |
368 | memset(ret, 0xa5, size); | |
369 | else if (opt_zero) | |
370 | memset(ret, 0, size); | |
371 | } | |
372 | VALGRIND_MAKE_MEM_UNDEFINED(ret, size); | |
373 | } else { | |
374 | VALGRIND_MAKE_MEM_UNDEFINED(ret, size); | |
375 | memset(ret, 0, size); | |
376 | } | |
377 | ||
378 | if (config_stats) | |
379 | tbin->tstats.nrequests++; | |
380 | if (config_prof) | |
381 | tcache->prof_accumbytes += size; | |
382 | } | |
383 | ||
384 | tcache_event(tcache); | |
385 | return (ret); | |
386 | } | |
387 | ||
388 | JEMALLOC_ALWAYS_INLINE void | |
389 | tcache_dalloc_small(tcache_t *tcache, void *ptr, size_t binind) | |
390 | { | |
391 | tcache_bin_t *tbin; | |
392 | tcache_bin_info_t *tbin_info; | |
393 | ||
394 | assert(tcache_salloc(ptr) <= SMALL_MAXCLASS); | |
395 | ||
396 | if (config_fill && opt_junk) | |
397 | arena_dalloc_junk_small(ptr, &arena_bin_info[binind]); | |
398 | ||
399 | tbin = &tcache->tbins[binind]; | |
400 | tbin_info = &tcache_bin_info[binind]; | |
401 | if (tbin->ncached == tbin_info->ncached_max) { | |
402 | tcache_bin_flush_small(tbin, binind, (tbin_info->ncached_max >> | |
403 | 1), tcache); | |
404 | } | |
405 | assert(tbin->ncached < tbin_info->ncached_max); | |
406 | tbin->avail[tbin->ncached] = ptr; | |
407 | tbin->ncached++; | |
408 | ||
409 | tcache_event(tcache); | |
410 | } | |
411 | ||
412 | JEMALLOC_ALWAYS_INLINE void | |
413 | tcache_dalloc_large(tcache_t *tcache, void *ptr, size_t size) | |
414 | { | |
415 | size_t binind; | |
416 | tcache_bin_t *tbin; | |
417 | tcache_bin_info_t *tbin_info; | |
418 | ||
419 | assert((size & PAGE_MASK) == 0); | |
420 | assert(tcache_salloc(ptr) > SMALL_MAXCLASS); | |
421 | assert(tcache_salloc(ptr) <= tcache_maxclass); | |
422 | ||
423 | binind = NBINS + (size >> LG_PAGE) - 1; | |
424 | ||
425 | if (config_fill && opt_junk) | |
426 | memset(ptr, 0x5a, size); | |
427 | ||
428 | tbin = &tcache->tbins[binind]; | |
429 | tbin_info = &tcache_bin_info[binind]; | |
430 | if (tbin->ncached == tbin_info->ncached_max) { | |
431 | tcache_bin_flush_large(tbin, binind, (tbin_info->ncached_max >> | |
432 | 1), tcache); | |
433 | } | |
434 | assert(tbin->ncached < tbin_info->ncached_max); | |
435 | tbin->avail[tbin->ncached] = ptr; | |
436 | tbin->ncached++; | |
437 | ||
438 | tcache_event(tcache); | |
439 | } | |
440 | #endif | |
441 | ||
442 | #endif /* JEMALLOC_H_INLINES */ | |
443 | /******************************************************************************/ | |
444 |
Copyright (c) 2006-2012 Rogue Wave Software, Inc. All Rights Reserved.
Patents pending.