This commit is contained in:
unknown
2026-02-04 20:27:13 +08:00
commit 3b042d1dad
9410 changed files with 1488147 additions and 0 deletions

View File

@@ -0,0 +1,52 @@
package bytepool
import (
"sync"
)
var Pool1k = NewPool(1 << 10)
var Pool4k = NewPool(4 << 10)
var Pool16k = NewPool(16 << 10)
var Pool32k = NewPool(32 << 10)
type Buf struct {
Bytes []byte
}
// Pool for get byte slice
type Pool struct {
length int
rawPool *sync.Pool
}
// NewPool 创建新对象
func NewPool(length int) *Pool {
if length < 0 {
length = 1024
}
return &Pool{
length: length,
rawPool: &sync.Pool{
New: func() any {
return &Buf{
Bytes: make([]byte, length),
}
},
},
}
}
// Get 获取一个新的byte slice
func (this *Pool) Get() *Buf {
return this.rawPool.Get().(*Buf)
}
// Put 放回一个使用过的byte slice
func (this *Pool) Put(ptr *Buf) {
this.rawPool.Put(ptr)
}
// Length 单个字节slice长度
func (this *Pool) Length() int {
return this.length
}

View File

@@ -0,0 +1,272 @@
package bytepool_test
import (
"bytes"
"github.com/TeaOSLab/EdgeNode/internal/utils/bytepool"
"runtime"
"sync"
"testing"
)
func TestBytePool_Memory(t *testing.T) {
var stat1 = &runtime.MemStats{}
runtime.ReadMemStats(stat1)
var pool = bytepool.NewPool(32 * 1024)
for i := 0; i < 20480; i++ {
pool.Put(&bytepool.Buf{
Bytes: make([]byte, 32*1024),
})
}
//pool.Purge()
//time.Sleep(60 * time.Second)
runtime.GC()
var stat2 = &runtime.MemStats{}
runtime.ReadMemStats(stat2)
t.Log((stat2.HeapInuse-stat1.HeapInuse)/1024/1024, "MB,")
}
func BenchmarkBytePool_Get(b *testing.B) {
runtime.GOMAXPROCS(1)
var pool = bytepool.NewPool(1)
b.ResetTimer()
for i := 0; i < b.N; i++ {
var buf = pool.Get()
_ = buf
pool.Put(buf)
}
}
func BenchmarkBytePool_Get_Parallel(b *testing.B) {
runtime.GOMAXPROCS(1)
var pool = bytepool.NewPool(1024)
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get()
pool.Put(buf)
}
})
}
func BenchmarkBytePool_Get_Sync(b *testing.B) {
runtime.GOMAXPROCS(1)
var pool = &sync.Pool{
New: func() any {
return make([]byte, 1024)
},
}
b.ReportAllocs()
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get()
pool.Put(buf)
}
})
}
func BenchmarkBytePool_Get_Sync2(b *testing.B) {
runtime.GOMAXPROCS(1)
var pool = &sync.Pool{
New: func() any {
return &bytepool.Buf{
Bytes: make([]byte, 1024),
}
},
}
b.ReportAllocs()
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get()
pool.Put(buf)
}
})
}
func BenchmarkBytePool_Copy_Bytes_4(b *testing.B) {
const size = 4 << 10
var data = bytes.Repeat([]byte{'A'}, size)
var pool = &sync.Pool{
New: func() any {
return make([]byte, size)
},
}
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get().([]byte)
copy(buf, data)
pool.Put(buf)
}
})
}
func BenchmarkBytePool_Copy_Wrapper_4(b *testing.B) {
const size = 4 << 10
var data = bytes.Repeat([]byte{'A'}, size)
var pool = &sync.Pool{
New: func() any {
return &bytepool.Buf{
Bytes: make([]byte, size),
}
},
}
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get().(*bytepool.Buf)
copy(buf.Bytes, data)
pool.Put(buf)
}
})
}
func BenchmarkBytePool_Copy_Bytes_16(b *testing.B) {
const size = 16 << 10
var data = bytes.Repeat([]byte{'A'}, size)
var pool = &sync.Pool{
New: func() any {
return make([]byte, size)
},
}
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get().([]byte)
copy(buf, data)
pool.Put(buf)
}
})
}
func BenchmarkBytePool_Copy_Wrapper_16(b *testing.B) {
const size = 16 << 10
var data = bytes.Repeat([]byte{'A'}, size)
var pool = &sync.Pool{
New: func() any {
return &bytepool.Buf{
Bytes: make([]byte, size),
}
},
}
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get().(*bytepool.Buf)
copy(buf.Bytes, data)
pool.Put(buf)
}
})
}
func BenchmarkBytePool_Copy_Wrapper_Buf_16(b *testing.B) {
const size = 16 << 10
var data = bytes.Repeat([]byte{'A'}, size)
var pool = &sync.Pool{
New: func() any {
return &bytepool.Buf{
Bytes: make([]byte, size),
}
},
}
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var bytesPtr = pool.Get().(*bytepool.Buf)
var buf = bytesPtr.Bytes
copy(buf, data)
pool.Put(bytesPtr)
}
})
}
func BenchmarkBytePool_Copy_Wrapper_BytePool_16(b *testing.B) {
const size = 16 << 10
var data = bytes.Repeat([]byte{'A'}, size)
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var bytesPtr = bytepool.Pool16k.Get()
copy(bytesPtr.Bytes, data)
bytepool.Pool16k.Put(bytesPtr)
}
})
}
func BenchmarkBytePool_Copy_Bytes_32(b *testing.B) {
const size = 32 << 10
var data = bytes.Repeat([]byte{'A'}, size)
var pool = &sync.Pool{
New: func() any {
return make([]byte, size)
},
}
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get().([]byte)
copy(buf, data)
pool.Put(buf)
}
})
}
func BenchmarkBytePool_Copy_Wrapper_32(b *testing.B) {
const size = 32 << 10
var data = bytes.Repeat([]byte{'A'}, size)
var pool = &sync.Pool{
New: func() any {
return &bytepool.Buf{
Bytes: make([]byte, size),
}
},
}
b.ResetTimer()
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
var buf = pool.Get().(*bytepool.Buf)
copy(buf.Bytes, data)
pool.Put(buf)
}
})
}