mirror of
https://github.com/golang/net.git
synced 2026-04-01 02:47:08 +09:00
fixedBuffer was a bad idea for two reasons: 1. It was fixed at a constant 64KB (the current default flow-control window) which wastes memory on the server when clients upload many small request bodies. 2. A follow-up CL will allow configuring the server's connection and stream receive windows. We want to allow individual streams to use varying amounts of the available connection window. This is not possible when each stream uses a fixedBuffer. dataBuffer grows and shrinks based on current usage. The worst-case fragmentation of dataBuffer is 32KB wasted memory per stream, but I expect that worst-case will be rare. In particular, if the declared size of a stream's request body is under 1KB, then the server will not allocate more than 1KB to process that stream's request body. Updates golang/go#16512 Fixes golang/go#18509 Change-Id: Ibcb18007037e82518a65848ef3baf4937955ac9d Reviewed-on: https://go-review.googlesource.com/37400 Run-TryBot: Tom Bergan <tombergan@google.com> TryBot-Result: Gobot Gobot <gobot@golang.org> Reviewed-by: Brad Fitzpatrick <bradfitz@golang.org>
147 lines
3.9 KiB
Go
147 lines
3.9 KiB
Go
// Copyright 2014 The Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
package http2
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"sync"
|
|
)
|
|
|
|
// Buffer chunks are allocated from a pool to reduce pressure on GC.
|
|
// The maximum wasted space per dataBuffer is 2x the largest size class,
|
|
// which happens when the dataBuffer has multiple chunks and there is
|
|
// one unread byte in both the first and last chunks. We use a few size
|
|
// classes to minimize overheads for servers that typically receive very
|
|
// small request bodies.
|
|
//
|
|
// TODO: Benchmark to determine if the pools are necessary. The GC may have
|
|
// improved enough that we can instead allocate chunks like this:
|
|
// make([]byte, max(16<<10, expectedBytesRemaining))
|
|
var (
|
|
dataChunkSizeClasses = []int{
|
|
1 << 10,
|
|
2 << 10,
|
|
4 << 10,
|
|
8 << 10,
|
|
16 << 10,
|
|
}
|
|
dataChunkPools = [...]sync.Pool{
|
|
{New: func() interface{} { return make([]byte, 1<<10) }},
|
|
{New: func() interface{} { return make([]byte, 2<<10) }},
|
|
{New: func() interface{} { return make([]byte, 4<<10) }},
|
|
{New: func() interface{} { return make([]byte, 8<<10) }},
|
|
{New: func() interface{} { return make([]byte, 16<<10) }},
|
|
}
|
|
)
|
|
|
|
func getDataBufferChunk(size int64) []byte {
|
|
i := 0
|
|
for ; i < len(dataChunkSizeClasses)-1; i++ {
|
|
if size <= int64(dataChunkSizeClasses[i]) {
|
|
break
|
|
}
|
|
}
|
|
return dataChunkPools[i].Get().([]byte)
|
|
}
|
|
|
|
func putDataBufferChunk(p []byte) {
|
|
for i, n := range dataChunkSizeClasses {
|
|
if len(p) == n {
|
|
dataChunkPools[i].Put(p)
|
|
return
|
|
}
|
|
}
|
|
panic(fmt.Sprintf("unexpected buffer len=%v", len(p)))
|
|
}
|
|
|
|
// dataBuffer is an io.ReadWriter backed by a list of data chunks.
|
|
// Each dataBuffer is used to read DATA frames on a single stream.
|
|
// The buffer is divided into chunks so the server can limit the
|
|
// total memory used by a single connection without limiting the
|
|
// request body size on any single stream.
|
|
type dataBuffer struct {
|
|
chunks [][]byte
|
|
r int // next byte to read is chunks[0][r]
|
|
w int // next byte to write is chunks[len(chunks)-1][w]
|
|
size int // total buffered bytes
|
|
expected int64 // we expect at least this many bytes in future Write calls (ignored if <= 0)
|
|
}
|
|
|
|
var errReadEmpty = errors.New("read from empty dataBuffer")
|
|
|
|
// Read copies bytes from the buffer into p.
|
|
// It is an error to read when no data is available.
|
|
func (b *dataBuffer) Read(p []byte) (int, error) {
|
|
if b.size == 0 {
|
|
return 0, errReadEmpty
|
|
}
|
|
var ntotal int
|
|
for len(p) > 0 && b.size > 0 {
|
|
readFrom := b.bytesFromFirstChunk()
|
|
n := copy(p, readFrom)
|
|
p = p[n:]
|
|
ntotal += n
|
|
b.r += n
|
|
b.size -= n
|
|
// If the first chunk has been consumed, advance to the next chunk.
|
|
if b.r == len(b.chunks[0]) {
|
|
putDataBufferChunk(b.chunks[0])
|
|
end := len(b.chunks) - 1
|
|
copy(b.chunks[:end], b.chunks[1:])
|
|
b.chunks[end] = nil
|
|
b.chunks = b.chunks[:end]
|
|
b.r = 0
|
|
}
|
|
}
|
|
return ntotal, nil
|
|
}
|
|
|
|
func (b *dataBuffer) bytesFromFirstChunk() []byte {
|
|
if len(b.chunks) == 1 {
|
|
return b.chunks[0][b.r:b.w]
|
|
}
|
|
return b.chunks[0][b.r:]
|
|
}
|
|
|
|
// Len returns the number of bytes of the unread portion of the buffer.
|
|
func (b *dataBuffer) Len() int {
|
|
return b.size
|
|
}
|
|
|
|
// Write appends p to the buffer.
|
|
func (b *dataBuffer) Write(p []byte) (int, error) {
|
|
ntotal := len(p)
|
|
for len(p) > 0 {
|
|
// If the last chunk is empty, allocate a new chunk. Try to allocate
|
|
// enough to fully copy p plus any additional bytes we expect to
|
|
// receive. However, this may allocate less than len(p).
|
|
want := int64(len(p))
|
|
if b.expected > want {
|
|
want = b.expected
|
|
}
|
|
chunk := b.lastChunkOrAlloc(want)
|
|
n := copy(chunk[b.w:], p)
|
|
p = p[n:]
|
|
b.w += n
|
|
b.size += n
|
|
b.expected -= int64(n)
|
|
}
|
|
return ntotal, nil
|
|
}
|
|
|
|
func (b *dataBuffer) lastChunkOrAlloc(want int64) []byte {
|
|
if len(b.chunks) != 0 {
|
|
last := b.chunks[len(b.chunks)-1]
|
|
if b.w < len(last) {
|
|
return last
|
|
}
|
|
}
|
|
chunk := getDataBufferChunk(want)
|
|
b.chunks = append(b.chunks, chunk)
|
|
b.w = 0
|
|
return chunk
|
|
}
|