| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160 | package cryptoimport (	"encoding/binary"	"io"	"github.com/xtls/xray-core/common"	"github.com/xtls/xray-core/common/buf")// ChunkSizeDecoder is a utility class to decode size value from bytes.type ChunkSizeDecoder interface {	SizeBytes() int32	Decode([]byte) (uint16, error)}// ChunkSizeEncoder is a utility class to encode size value into bytes.type ChunkSizeEncoder interface {	SizeBytes() int32	Encode(uint16, []byte) []byte}type PaddingLengthGenerator interface {	MaxPaddingLen() uint16	NextPaddingLen() uint16}type PlainChunkSizeParser struct{}func (PlainChunkSizeParser) SizeBytes() int32 {	return 2}func (PlainChunkSizeParser) Encode(size uint16, b []byte) []byte {	binary.BigEndian.PutUint16(b, size)	return b[:2]}func (PlainChunkSizeParser) Decode(b []byte) (uint16, error) {	return binary.BigEndian.Uint16(b), nil}type AEADChunkSizeParser struct {	Auth *AEADAuthenticator}func (p *AEADChunkSizeParser) SizeBytes() int32 {	return 2 + int32(p.Auth.Overhead())}func (p *AEADChunkSizeParser) Encode(size uint16, b []byte) []byte {	binary.BigEndian.PutUint16(b, size-uint16(p.Auth.Overhead()))	b, err := p.Auth.Seal(b[:0], b[:2])	common.Must(err)	return b}func (p *AEADChunkSizeParser) Decode(b []byte) (uint16, error) {	b, err := p.Auth.Open(b[:0], b)	if err != nil {		return 0, err	}	return binary.BigEndian.Uint16(b) + uint16(p.Auth.Overhead()), nil}type ChunkStreamReader struct {	sizeDecoder ChunkSizeDecoder	reader      *buf.BufferedReader	buffer       []byte	leftOverSize int32	maxNumChunk  uint32	numChunk     uint32}func NewChunkStreamReader(sizeDecoder ChunkSizeDecoder, reader io.Reader) *ChunkStreamReader {	return NewChunkStreamReaderWithChunkCount(sizeDecoder, reader, 0)}func NewChunkStreamReaderWithChunkCount(sizeDecoder ChunkSizeDecoder, reader io.Reader, maxNumChunk uint32) *ChunkStreamReader {	r := &ChunkStreamReader{		sizeDecoder: sizeDecoder,		buffer:      make([]byte, sizeDecoder.SizeBytes()),		maxNumChunk: maxNumChunk,	}	if breader, ok := reader.(*buf.BufferedReader); ok {		r.reader = breader	} else {		r.reader = &buf.BufferedReader{Reader: buf.NewReader(reader)}	}	return r}func (r *ChunkStreamReader) readSize() (uint16, error) {	if _, err := io.ReadFull(r.reader, r.buffer); err != nil {		return 0, err	}	return r.sizeDecoder.Decode(r.buffer)}func (r *ChunkStreamReader) ReadMultiBuffer() (buf.MultiBuffer, error) {	size := r.leftOverSize	if size == 0 {		r.numChunk++		if r.maxNumChunk > 0 && r.numChunk > r.maxNumChunk {			return nil, io.EOF		}		nextSize, err := r.readSize()		if err != nil {			return nil, err		}		if nextSize == 0 {			return nil, io.EOF		}		size = int32(nextSize)	}	r.leftOverSize = size	mb, err := r.reader.ReadAtMost(size)	if !mb.IsEmpty() {		r.leftOverSize -= mb.Len()		return mb, nil	}	return nil, err}type ChunkStreamWriter struct {	sizeEncoder ChunkSizeEncoder	writer      buf.Writer}func NewChunkStreamWriter(sizeEncoder ChunkSizeEncoder, writer io.Writer) *ChunkStreamWriter {	return &ChunkStreamWriter{		sizeEncoder: sizeEncoder,		writer:      buf.NewWriter(writer),	}}func (w *ChunkStreamWriter) WriteMultiBuffer(mb buf.MultiBuffer) error {	const sliceSize = 8192	mbLen := mb.Len()	mb2Write := make(buf.MultiBuffer, 0, mbLen/buf.Size+mbLen/sliceSize+2)	for {		mb2, slice := buf.SplitSize(mb, sliceSize)		mb = mb2		b := buf.New()		w.sizeEncoder.Encode(uint16(slice.Len()), b.Extend(w.sizeEncoder.SizeBytes()))		mb2Write = append(mb2Write, b)		mb2Write = append(mb2Write, slice...)		if mb.IsEmpty() {			break		}	}	return w.writer.WriteMultiBuffer(mb2Write)}
 |