mirror of git://gcc.gnu.org/git/gcc.git
				
				
				
			
		
			
				
	
	
		
			256 lines
		
	
	
		
			6.5 KiB
		
	
	
	
		
			Go
		
	
	
	
			
		
		
	
	
			256 lines
		
	
	
		
			6.5 KiB
		
	
	
	
		
			Go
		
	
	
	
| // Copyright 2009 The Go Authors. All rights reserved.
 | |
| // Use of this source code is governed by a BSD-style
 | |
| // license that can be found in the LICENSE file.
 | |
| 
 | |
| // The wire protocol for HTTP's "chunked" Transfer-Encoding.
 | |
| 
 | |
| // Package internal contains HTTP internals shared by net/http and
 | |
| // net/http/httputil.
 | |
| package internal
 | |
| 
 | |
| import (
 | |
| 	"bufio"
 | |
| 	"bytes"
 | |
| 	"errors"
 | |
| 	"fmt"
 | |
| 	"io"
 | |
| )
 | |
| 
 | |
| const maxLineLength = 4096 // assumed <= bufio.defaultBufSize
 | |
| 
 | |
| var ErrLineTooLong = errors.New("header line too long")
 | |
| 
 | |
| // NewChunkedReader returns a new chunkedReader that translates the data read from r
 | |
| // out of HTTP "chunked" format before returning it.
 | |
| // The chunkedReader returns io.EOF when the final 0-length chunk is read.
 | |
| //
 | |
| // NewChunkedReader is not needed by normal applications. The http package
 | |
| // automatically decodes chunking when reading response bodies.
 | |
| func NewChunkedReader(r io.Reader) io.Reader {
 | |
| 	br, ok := r.(*bufio.Reader)
 | |
| 	if !ok {
 | |
| 		br = bufio.NewReader(r)
 | |
| 	}
 | |
| 	return &chunkedReader{r: br}
 | |
| }
 | |
| 
 | |
| type chunkedReader struct {
 | |
| 	r        *bufio.Reader
 | |
| 	n        uint64 // unread bytes in chunk
 | |
| 	err      error
 | |
| 	buf      [2]byte
 | |
| 	checkEnd bool // whether need to check for \r\n chunk footer
 | |
| }
 | |
| 
 | |
| func (cr *chunkedReader) beginChunk() {
 | |
| 	// chunk-size CRLF
 | |
| 	var line []byte
 | |
| 	line, cr.err = readChunkLine(cr.r)
 | |
| 	if cr.err != nil {
 | |
| 		return
 | |
| 	}
 | |
| 	cr.n, cr.err = parseHexUint(line)
 | |
| 	if cr.err != nil {
 | |
| 		return
 | |
| 	}
 | |
| 	if cr.n == 0 {
 | |
| 		cr.err = io.EOF
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (cr *chunkedReader) chunkHeaderAvailable() bool {
 | |
| 	n := cr.r.Buffered()
 | |
| 	if n > 0 {
 | |
| 		peek, _ := cr.r.Peek(n)
 | |
| 		return bytes.IndexByte(peek, '\n') >= 0
 | |
| 	}
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| func (cr *chunkedReader) Read(b []uint8) (n int, err error) {
 | |
| 	for cr.err == nil {
 | |
| 		if cr.checkEnd {
 | |
| 			if n > 0 && cr.r.Buffered() < 2 {
 | |
| 				// We have some data. Return early (per the io.Reader
 | |
| 				// contract) instead of potentially blocking while
 | |
| 				// reading more.
 | |
| 				break
 | |
| 			}
 | |
| 			if _, cr.err = io.ReadFull(cr.r, cr.buf[:2]); cr.err == nil {
 | |
| 				if string(cr.buf[:]) != "\r\n" {
 | |
| 					cr.err = errors.New("malformed chunked encoding")
 | |
| 					break
 | |
| 				}
 | |
| 			}
 | |
| 			cr.checkEnd = false
 | |
| 		}
 | |
| 		if cr.n == 0 {
 | |
| 			if n > 0 && !cr.chunkHeaderAvailable() {
 | |
| 				// We've read enough. Don't potentially block
 | |
| 				// reading a new chunk header.
 | |
| 				break
 | |
| 			}
 | |
| 			cr.beginChunk()
 | |
| 			continue
 | |
| 		}
 | |
| 		if len(b) == 0 {
 | |
| 			break
 | |
| 		}
 | |
| 		rbuf := b
 | |
| 		if uint64(len(rbuf)) > cr.n {
 | |
| 			rbuf = rbuf[:cr.n]
 | |
| 		}
 | |
| 		var n0 int
 | |
| 		n0, cr.err = cr.r.Read(rbuf)
 | |
| 		n += n0
 | |
| 		b = b[n0:]
 | |
| 		cr.n -= uint64(n0)
 | |
| 		// If we're at the end of a chunk, read the next two
 | |
| 		// bytes to verify they are "\r\n".
 | |
| 		if cr.n == 0 && cr.err == nil {
 | |
| 			cr.checkEnd = true
 | |
| 		}
 | |
| 	}
 | |
| 	return n, cr.err
 | |
| }
 | |
| 
 | |
| // Read a line of bytes (up to \n) from b.
 | |
| // Give up if the line exceeds maxLineLength.
 | |
| // The returned bytes are owned by the bufio.Reader
 | |
| // so they are only valid until the next bufio read.
 | |
| func readChunkLine(b *bufio.Reader) ([]byte, error) {
 | |
| 	p, err := b.ReadSlice('\n')
 | |
| 	if err != nil {
 | |
| 		// We always know when EOF is coming.
 | |
| 		// If the caller asked for a line, there should be a line.
 | |
| 		if err == io.EOF {
 | |
| 			err = io.ErrUnexpectedEOF
 | |
| 		} else if err == bufio.ErrBufferFull {
 | |
| 			err = ErrLineTooLong
 | |
| 		}
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	if len(p) >= maxLineLength {
 | |
| 		return nil, ErrLineTooLong
 | |
| 	}
 | |
| 	p = trimTrailingWhitespace(p)
 | |
| 	p, err = removeChunkExtension(p)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	return p, nil
 | |
| }
 | |
| 
 | |
| func trimTrailingWhitespace(b []byte) []byte {
 | |
| 	for len(b) > 0 && isASCIISpace(b[len(b)-1]) {
 | |
| 		b = b[:len(b)-1]
 | |
| 	}
 | |
| 	return b
 | |
| }
 | |
| 
 | |
| func isASCIISpace(b byte) bool {
 | |
| 	return b == ' ' || b == '\t' || b == '\n' || b == '\r'
 | |
| }
 | |
| 
 | |
| // removeChunkExtension removes any chunk-extension from p.
 | |
| // For example,
 | |
| //     "0" => "0"
 | |
| //     "0;token" => "0"
 | |
| //     "0;token=val" => "0"
 | |
| //     `0;token="quoted string"` => "0"
 | |
| func removeChunkExtension(p []byte) ([]byte, error) {
 | |
| 	semi := bytes.IndexByte(p, ';')
 | |
| 	if semi == -1 {
 | |
| 		return p, nil
 | |
| 	}
 | |
| 	// TODO: care about exact syntax of chunk extensions? We're
 | |
| 	// ignoring and stripping them anyway. For now just never
 | |
| 	// return an error.
 | |
| 	return p[:semi], nil
 | |
| }
 | |
| 
 | |
| // NewChunkedWriter returns a new chunkedWriter that translates writes into HTTP
 | |
| // "chunked" format before writing them to w. Closing the returned chunkedWriter
 | |
| // sends the final 0-length chunk that marks the end of the stream but does
 | |
| // not send the final CRLF that appears after trailers; trailers and the last
 | |
| // CRLF must be written separately.
 | |
| //
 | |
| // NewChunkedWriter is not needed by normal applications. The http
 | |
| // package adds chunking automatically if handlers don't set a
 | |
| // Content-Length header. Using newChunkedWriter inside a handler
 | |
| // would result in double chunking or chunking with a Content-Length
 | |
| // length, both of which are wrong.
 | |
| func NewChunkedWriter(w io.Writer) io.WriteCloser {
 | |
| 	return &chunkedWriter{w}
 | |
| }
 | |
| 
 | |
| // Writing to chunkedWriter translates to writing in HTTP chunked Transfer
 | |
| // Encoding wire format to the underlying Wire chunkedWriter.
 | |
| type chunkedWriter struct {
 | |
| 	Wire io.Writer
 | |
| }
 | |
| 
 | |
| // Write the contents of data as one chunk to Wire.
 | |
| // NOTE: Note that the corresponding chunk-writing procedure in Conn.Write has
 | |
| // a bug since it does not check for success of io.WriteString
 | |
| func (cw *chunkedWriter) Write(data []byte) (n int, err error) {
 | |
| 
 | |
| 	// Don't send 0-length data. It looks like EOF for chunked encoding.
 | |
| 	if len(data) == 0 {
 | |
| 		return 0, nil
 | |
| 	}
 | |
| 
 | |
| 	if _, err = fmt.Fprintf(cw.Wire, "%x\r\n", len(data)); err != nil {
 | |
| 		return 0, err
 | |
| 	}
 | |
| 	if n, err = cw.Wire.Write(data); err != nil {
 | |
| 		return
 | |
| 	}
 | |
| 	if n != len(data) {
 | |
| 		err = io.ErrShortWrite
 | |
| 		return
 | |
| 	}
 | |
| 	if _, err = io.WriteString(cw.Wire, "\r\n"); err != nil {
 | |
| 		return
 | |
| 	}
 | |
| 	if bw, ok := cw.Wire.(*FlushAfterChunkWriter); ok {
 | |
| 		err = bw.Flush()
 | |
| 	}
 | |
| 	return
 | |
| }
 | |
| 
 | |
| func (cw *chunkedWriter) Close() error {
 | |
| 	_, err := io.WriteString(cw.Wire, "0\r\n")
 | |
| 	return err
 | |
| }
 | |
| 
 | |
| // FlushAfterChunkWriter signals from the caller of NewChunkedWriter
 | |
| // that each chunk should be followed by a flush. It is used by the
 | |
| // http.Transport code to keep the buffering behavior for headers and
 | |
| // trailers, but flush out chunks aggressively in the middle for
 | |
| // request bodies which may be generated slowly. See Issue 6574.
 | |
| type FlushAfterChunkWriter struct {
 | |
| 	*bufio.Writer
 | |
| }
 | |
| 
 | |
| func parseHexUint(v []byte) (n uint64, err error) {
 | |
| 	for i, b := range v {
 | |
| 		switch {
 | |
| 		case '0' <= b && b <= '9':
 | |
| 			b = b - '0'
 | |
| 		case 'a' <= b && b <= 'f':
 | |
| 			b = b - 'a' + 10
 | |
| 		case 'A' <= b && b <= 'F':
 | |
| 			b = b - 'A' + 10
 | |
| 		default:
 | |
| 			return 0, errors.New("invalid byte in chunk length")
 | |
| 		}
 | |
| 		if i == 16 {
 | |
| 			return 0, errors.New("http chunk length too large")
 | |
| 		}
 | |
| 		n <<= 4
 | |
| 		n |= uint64(b)
 | |
| 	}
 | |
| 	return
 | |
| }
 |