mirror of
https://github.com/autc04/Retro68.git
synced 2024-12-01 11:52:47 +00:00
308 lines
7.5 KiB
C
308 lines
7.5 KiB
C
// Copyright 2009 The Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
// Central free lists.
|
|
//
|
|
// See malloc.h for an overview.
|
|
//
|
|
// The MCentral doesn't actually contain the list of free objects; the MSpan does.
|
|
// Each MCentral is two lists of MSpans: those with free objects (c->nonempty)
|
|
// and those that are completely allocated (c->mempty).
|
|
//
|
|
// TODO(rsc): tcmalloc uses a "transfer cache" to split the list
|
|
// into sections of class_to_transfercount[sizeclass] objects
|
|
// so that it is faster to move those lists between MCaches and MCentrals.
|
|
|
|
#include "runtime.h"
|
|
#include "arch.h"
|
|
#include "malloc.h"
|
|
|
|
static bool MCentral_Grow(MCentral *c);
|
|
static void MCentral_Free(MCentral *c, MLink *v);
|
|
static void MCentral_ReturnToHeap(MCentral *c, MSpan *s);
|
|
|
|
// Initialize a single central free list.
|
|
void
|
|
runtime_MCentral_Init(MCentral *c, int32 sizeclass)
|
|
{
|
|
c->sizeclass = sizeclass;
|
|
runtime_MSpanList_Init(&c->nonempty);
|
|
runtime_MSpanList_Init(&c->mempty);
|
|
}
|
|
|
|
// Allocate a span to use in an MCache.
|
|
MSpan*
|
|
runtime_MCentral_CacheSpan(MCentral *c)
|
|
{
|
|
MSpan *s;
|
|
int32 cap, n;
|
|
uint32 sg;
|
|
|
|
runtime_lock(c);
|
|
sg = runtime_mheap.sweepgen;
|
|
retry:
|
|
for(s = c->nonempty.next; s != &c->nonempty; s = s->next) {
|
|
if(s->sweepgen == sg-2 && runtime_cas(&s->sweepgen, sg-2, sg-1)) {
|
|
runtime_unlock(c);
|
|
runtime_MSpan_Sweep(s);
|
|
runtime_lock(c);
|
|
// the span could have been moved to heap, retry
|
|
goto retry;
|
|
}
|
|
if(s->sweepgen == sg-1) {
|
|
// the span is being swept by background sweeper, skip
|
|
continue;
|
|
}
|
|
// we have a nonempty span that does not require sweeping, allocate from it
|
|
goto havespan;
|
|
}
|
|
|
|
for(s = c->mempty.next; s != &c->mempty; s = s->next) {
|
|
if(s->sweepgen == sg-2 && runtime_cas(&s->sweepgen, sg-2, sg-1)) {
|
|
// we have an empty span that requires sweeping,
|
|
// sweep it and see if we can free some space in it
|
|
runtime_MSpanList_Remove(s);
|
|
// swept spans are at the end of the list
|
|
runtime_MSpanList_InsertBack(&c->mempty, s);
|
|
runtime_unlock(c);
|
|
runtime_MSpan_Sweep(s);
|
|
runtime_lock(c);
|
|
// the span could be moved to nonempty or heap, retry
|
|
goto retry;
|
|
}
|
|
if(s->sweepgen == sg-1) {
|
|
// the span is being swept by background sweeper, skip
|
|
continue;
|
|
}
|
|
// already swept empty span,
|
|
// all subsequent ones must also be either swept or in process of sweeping
|
|
break;
|
|
}
|
|
|
|
// Replenish central list if empty.
|
|
if(!MCentral_Grow(c)) {
|
|
runtime_unlock(c);
|
|
return nil;
|
|
}
|
|
goto retry;
|
|
|
|
havespan:
|
|
cap = (s->npages << PageShift) / s->elemsize;
|
|
n = cap - s->ref;
|
|
if(n == 0)
|
|
runtime_throw("empty span");
|
|
if(s->freelist == nil)
|
|
runtime_throw("freelist empty");
|
|
c->nfree -= n;
|
|
runtime_MSpanList_Remove(s);
|
|
runtime_MSpanList_InsertBack(&c->mempty, s);
|
|
s->incache = true;
|
|
runtime_unlock(c);
|
|
return s;
|
|
}
|
|
|
|
// Return span from an MCache.
|
|
void
|
|
runtime_MCentral_UncacheSpan(MCentral *c, MSpan *s)
|
|
{
|
|
MLink *v;
|
|
int32 cap, n;
|
|
|
|
runtime_lock(c);
|
|
|
|
s->incache = false;
|
|
|
|
// Move any explicitly freed items from the freebuf to the freelist.
|
|
while((v = s->freebuf) != nil) {
|
|
s->freebuf = v->next;
|
|
runtime_markfreed(v);
|
|
v->next = s->freelist;
|
|
s->freelist = v;
|
|
s->ref--;
|
|
}
|
|
|
|
if(s->ref == 0) {
|
|
// Free back to heap. Unlikely, but possible.
|
|
MCentral_ReturnToHeap(c, s); // unlocks c
|
|
return;
|
|
}
|
|
|
|
cap = (s->npages << PageShift) / s->elemsize;
|
|
n = cap - s->ref;
|
|
if(n > 0) {
|
|
c->nfree += n;
|
|
runtime_MSpanList_Remove(s);
|
|
runtime_MSpanList_Insert(&c->nonempty, s);
|
|
}
|
|
runtime_unlock(c);
|
|
}
|
|
|
|
// Free the list of objects back into the central free list c.
|
|
// Called from runtime_free.
|
|
void
|
|
runtime_MCentral_FreeList(MCentral *c, MLink *start)
|
|
{
|
|
MLink *next;
|
|
|
|
runtime_lock(c);
|
|
for(; start != nil; start = next) {
|
|
next = start->next;
|
|
MCentral_Free(c, start);
|
|
}
|
|
runtime_unlock(c);
|
|
}
|
|
|
|
// Helper: free one object back into the central free list.
|
|
// Caller must hold lock on c on entry. Holds lock on exit.
|
|
static void
|
|
MCentral_Free(MCentral *c, MLink *v)
|
|
{
|
|
MSpan *s;
|
|
|
|
// Find span for v.
|
|
s = runtime_MHeap_Lookup(&runtime_mheap, v);
|
|
if(s == nil || s->ref == 0)
|
|
runtime_throw("invalid free");
|
|
if(s->sweepgen != runtime_mheap.sweepgen)
|
|
runtime_throw("free into unswept span");
|
|
|
|
// If the span is currently being used unsynchronized by an MCache,
|
|
// we can't modify the freelist. Add to the freebuf instead. The
|
|
// items will get moved to the freelist when the span is returned
|
|
// by the MCache.
|
|
if(s->incache) {
|
|
v->next = s->freebuf;
|
|
s->freebuf = v;
|
|
return;
|
|
}
|
|
|
|
// Move span to nonempty if necessary.
|
|
if(s->freelist == nil) {
|
|
runtime_MSpanList_Remove(s);
|
|
runtime_MSpanList_Insert(&c->nonempty, s);
|
|
}
|
|
|
|
// Add the object to span's free list.
|
|
runtime_markfreed(v);
|
|
v->next = s->freelist;
|
|
s->freelist = v;
|
|
s->ref--;
|
|
c->nfree++;
|
|
|
|
// If s is completely freed, return it to the heap.
|
|
if(s->ref == 0) {
|
|
MCentral_ReturnToHeap(c, s); // unlocks c
|
|
runtime_lock(c);
|
|
}
|
|
}
|
|
|
|
// Free n objects from a span s back into the central free list c.
|
|
// Called during sweep.
|
|
// Returns true if the span was returned to heap. Sets sweepgen to
|
|
// the latest generation.
|
|
bool
|
|
runtime_MCentral_FreeSpan(MCentral *c, MSpan *s, int32 n, MLink *start, MLink *end)
|
|
{
|
|
if(s->incache)
|
|
runtime_throw("freespan into cached span");
|
|
runtime_lock(c);
|
|
|
|
// Move to nonempty if necessary.
|
|
if(s->freelist == nil) {
|
|
runtime_MSpanList_Remove(s);
|
|
runtime_MSpanList_Insert(&c->nonempty, s);
|
|
}
|
|
|
|
// Add the objects back to s's free list.
|
|
end->next = s->freelist;
|
|
s->freelist = start;
|
|
s->ref -= n;
|
|
c->nfree += n;
|
|
|
|
// delay updating sweepgen until here. This is the signal that
|
|
// the span may be used in an MCache, so it must come after the
|
|
// linked list operations above (actually, just after the
|
|
// lock of c above.)
|
|
runtime_atomicstore(&s->sweepgen, runtime_mheap.sweepgen);
|
|
|
|
if(s->ref != 0) {
|
|
runtime_unlock(c);
|
|
return false;
|
|
}
|
|
|
|
// s is completely freed, return it to the heap.
|
|
MCentral_ReturnToHeap(c, s); // unlocks c
|
|
return true;
|
|
}
|
|
|
|
void
|
|
runtime_MGetSizeClassInfo(int32 sizeclass, uintptr *sizep, int32 *npagesp, int32 *nobj)
|
|
{
|
|
int32 size;
|
|
int32 npages;
|
|
|
|
npages = runtime_class_to_allocnpages[sizeclass];
|
|
size = runtime_class_to_size[sizeclass];
|
|
*npagesp = npages;
|
|
*sizep = size;
|
|
*nobj = (npages << PageShift) / size;
|
|
}
|
|
|
|
// Fetch a new span from the heap and
|
|
// carve into objects for the free list.
|
|
static bool
|
|
MCentral_Grow(MCentral *c)
|
|
{
|
|
int32 i, n, npages;
|
|
uintptr size;
|
|
MLink **tailp, *v;
|
|
byte *p;
|
|
MSpan *s;
|
|
|
|
runtime_unlock(c);
|
|
runtime_MGetSizeClassInfo(c->sizeclass, &size, &npages, &n);
|
|
s = runtime_MHeap_Alloc(&runtime_mheap, npages, c->sizeclass, 0, 1);
|
|
if(s == nil) {
|
|
// TODO(rsc): Log out of memory
|
|
runtime_lock(c);
|
|
return false;
|
|
}
|
|
|
|
// Carve span into sequence of blocks.
|
|
tailp = &s->freelist;
|
|
p = (byte*)(s->start << PageShift);
|
|
s->limit = (uintptr)(p + size*n);
|
|
for(i=0; i<n; i++) {
|
|
v = (MLink*)p;
|
|
*tailp = v;
|
|
tailp = &v->next;
|
|
p += size;
|
|
}
|
|
*tailp = nil;
|
|
runtime_markspan((byte*)(s->start<<PageShift), size, n, size*n < (s->npages<<PageShift));
|
|
|
|
runtime_lock(c);
|
|
c->nfree += n;
|
|
runtime_MSpanList_Insert(&c->nonempty, s);
|
|
return true;
|
|
}
|
|
|
|
// Return s to the heap. s must be unused (s->ref == 0). Unlocks c.
|
|
static void
|
|
MCentral_ReturnToHeap(MCentral *c, MSpan *s)
|
|
{
|
|
int32 size;
|
|
|
|
size = runtime_class_to_size[c->sizeclass];
|
|
runtime_MSpanList_Remove(s);
|
|
s->needzero = 1;
|
|
s->freelist = nil;
|
|
if(s->ref != 0)
|
|
runtime_throw("ref wrong");
|
|
c->nfree -= (s->npages << PageShift) / size;
|
|
runtime_unlock(c);
|
|
runtime_unmarkspan((byte*)(s->start<<PageShift), s->npages<<PageShift);
|
|
runtime_MHeap_Free(&runtime_mheap, s, 0);
|
|
}
|