// Copyright 2016 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. package runtime import ( "internal/cpu" "runtime/internal/atomic" "runtime/internal/sys" "unsafe" ) // A gcSweepBuf is a set of *mspans. // // gcSweepBuf is safe for concurrent push operations *or* concurrent // pop operations, but not both simultaneously. type gcSweepBuf struct { // A gcSweepBuf is a two-level data structure consisting of a // growable spine that points to fixed-sized blocks. The spine // can be accessed without locks, but adding a block or // growing it requires taking the spine lock. // // Because each mspan covers at least 8K of heap and takes at // most 8 bytes in the gcSweepBuf, the growth of the spine is // quite limited. // // The spine and all blocks are allocated off-heap, which // allows this to be used in the memory manager and avoids the // need for write barriers on all of these. We never release // this memory because there could be concurrent lock-free // access and we're likely to reuse it anyway. (In principle, // we could do this during STW.) spineLock mutex spine unsafe.Pointer // *[N]*gcSweepBlock, accessed atomically spineLen uintptr // Spine array length, accessed atomically spineCap uintptr // Spine array cap, accessed under lock // index is the first unused slot in the logical concatenation // of all blocks. It is accessed atomically. index uint32 } const ( gcSweepBlockEntries = 512 // 4KB on 64-bit gcSweepBufInitSpineCap = 256 // Enough for 1GB heap on 64-bit ) type gcSweepBlock struct { spans [gcSweepBlockEntries]*mspan } // push adds span s to buffer b. push is safe to call concurrently // with other push operations, but NOT to call concurrently with pop. func (b *gcSweepBuf) push(s *mspan) { // Obtain our slot. cursor := uintptr(atomic.Xadd(&b.index, +1) - 1) top, bottom := cursor/gcSweepBlockEntries, cursor%gcSweepBlockEntries // Do we need to add a block? spineLen := atomic.Loaduintptr(&b.spineLen) var block *gcSweepBlock retry: if top < spineLen { spine := atomic.Loadp(unsafe.Pointer(&b.spine)) blockp := add(spine, sys.PtrSize*top) block = (*gcSweepBlock)(atomic.Loadp(blockp)) } else { // Add a new block to the spine, potentially growing // the spine. lock(&b.spineLock) // spineLen cannot change until we release the lock, // but may have changed while we were waiting. spineLen = atomic.Loaduintptr(&b.spineLen) if top < spineLen { unlock(&b.spineLock) goto retry } if spineLen == b.spineCap { // Grow the spine. newCap := b.spineCap * 2 if newCap == 0 { newCap = gcSweepBufInitSpineCap } newSpine := persistentalloc(newCap*sys.PtrSize, cpu.CacheLineSize, &memstats.gc_sys) if b.spineCap != 0 { // Blocks are allocated off-heap, so // no write barriers. memmove(newSpine, b.spine, b.spineCap*sys.PtrSize) } // Spine is allocated off-heap, so no write barrier. atomic.StorepNoWB(unsafe.Pointer(&b.spine), newSpine) b.spineCap = newCap // We can't immediately free the old spine // since a concurrent push with a lower index // could still be reading from it. We let it // leak because even a 1TB heap would waste // less than 2MB of memory on old spines. If // this is a problem, we could free old spines // during STW. } // Allocate a new block and add it to the spine. block = (*gcSweepBlock)(persistentalloc(unsafe.Sizeof(gcSweepBlock{}), cpu.CacheLineSize, &memstats.gc_sys)) blockp := add(b.spine, sys.PtrSize*top) // Blocks are allocated off-heap, so no write barrier. atomic.StorepNoWB(blockp, unsafe.Pointer(block)) atomic.Storeuintptr(&b.spineLen, spineLen+1) unlock(&b.spineLock) } // We have a block. Insert the span atomically, since there may be // concurrent readers via the block API. atomic.StorepNoWB(unsafe.Pointer(&block.spans[bottom]), unsafe.Pointer(s)) } // pop removes and returns a span from buffer b, or nil if b is empty. // pop is safe to call concurrently with other pop operations, but NOT // to call concurrently with push. func (b *gcSweepBuf) pop() *mspan { cursor := atomic.Xadd(&b.index, -1) if int32(cursor) < 0 { atomic.Xadd(&b.index, +1) return nil } // There are no concurrent spine or block modifications during // pop, so we can omit the atomics. top, bottom := cursor/gcSweepBlockEntries, cursor%gcSweepBlockEntries blockp := (**gcSweepBlock)(add(b.spine, sys.PtrSize*uintptr(top))) block := *blockp s := block.spans[bottom] // Clear the pointer for block(i). block.spans[bottom] = nil return s } // numBlocks returns the number of blocks in buffer b. numBlocks is // safe to call concurrently with any other operation. Spans that have // been pushed prior to the call to numBlocks are guaranteed to appear // in some block in the range [0, numBlocks()), assuming there are no // intervening pops. Spans that are pushed after the call may also // appear in these blocks. func (b *gcSweepBuf) numBlocks() int { return int(divRoundUp(uintptr(atomic.Load(&b.index)), gcSweepBlockEntries)) } // block returns the spans in the i'th block of buffer b. block is // safe to call concurrently with push. The block may contain nil // pointers that must be ignored, and each entry in the block must be // loaded atomically. func (b *gcSweepBuf) block(i int) []*mspan { // Perform bounds check before loading spine address since // push ensures the allocated length is at least spineLen. if i < 0 || uintptr(i) >= atomic.Loaduintptr(&b.spineLen) { throw("block index out of range") } // Get block i. spine := atomic.Loadp(unsafe.Pointer(&b.spine)) blockp := add(spine, sys.PtrSize*uintptr(i)) block := (*gcSweepBlock)(atomic.Loadp(blockp)) // Slice the block if necessary. cursor := uintptr(atomic.Load(&b.index)) top, bottom := cursor/gcSweepBlockEntries, cursor%gcSweepBlockEntries var spans []*mspan if uintptr(i) < top { spans = block.spans[:] } else { spans = block.spans[:bottom] } return spans }