mirror of
https://github.com/golang/go
synced 2024-11-06 04:26:11 -07:00
ddb648fdf6
These new methods provide support for cases where performance is a primary concern. For example, copying files from an existing zip to a new zip without incurring the decompression and compression overhead. Using an optimized, external compression method and writing the output to a zip archive. And compressing file contents in parallel and then sequentially writing the compressed bytes to a zip archive. TestWriterCopy is copied verbatim from https://github.com/rsc/zipmerge Fixes #34974 Change-Id: Iade5bc245ba34cdbb86364bf59f79f38bb9e2eb6 Reviewed-on: https://go-review.googlesource.com/c/go/+/312310 Run-TryBot: Ian Lance Taylor <iant@golang.org> TryBot-Result: Go Bot <gobot@golang.org> Reviewed-by: Ian Lance Taylor <iant@golang.org> Trust: Carlos Amedee <carlos@golang.org>
635 lines
18 KiB
Go
635 lines
18 KiB
Go
// Copyright 2011 The Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
package zip
|
|
|
|
import (
|
|
"bufio"
|
|
"encoding/binary"
|
|
"errors"
|
|
"hash"
|
|
"hash/crc32"
|
|
"io"
|
|
"strings"
|
|
"unicode/utf8"
|
|
)
|
|
|
|
var (
|
|
errLongName = errors.New("zip: FileHeader.Name too long")
|
|
errLongExtra = errors.New("zip: FileHeader.Extra too long")
|
|
)
|
|
|
|
// Writer implements a zip file writer.
|
|
type Writer struct {
|
|
cw *countWriter
|
|
dir []*header
|
|
last *fileWriter
|
|
closed bool
|
|
compressors map[uint16]Compressor
|
|
comment string
|
|
|
|
// testHookCloseSizeOffset if non-nil is called with the size
|
|
// of offset of the central directory at Close.
|
|
testHookCloseSizeOffset func(size, offset uint64)
|
|
}
|
|
|
|
type header struct {
|
|
*FileHeader
|
|
offset uint64
|
|
raw bool
|
|
}
|
|
|
|
// NewWriter returns a new Writer writing a zip file to w.
|
|
func NewWriter(w io.Writer) *Writer {
|
|
return &Writer{cw: &countWriter{w: bufio.NewWriter(w)}}
|
|
}
|
|
|
|
// SetOffset sets the offset of the beginning of the zip data within the
|
|
// underlying writer. It should be used when the zip data is appended to an
|
|
// existing file, such as a binary executable.
|
|
// It must be called before any data is written.
|
|
func (w *Writer) SetOffset(n int64) {
|
|
if w.cw.count != 0 {
|
|
panic("zip: SetOffset called after data was written")
|
|
}
|
|
w.cw.count = n
|
|
}
|
|
|
|
// Flush flushes any buffered data to the underlying writer.
|
|
// Calling Flush is not normally necessary; calling Close is sufficient.
|
|
func (w *Writer) Flush() error {
|
|
return w.cw.w.(*bufio.Writer).Flush()
|
|
}
|
|
|
|
// SetComment sets the end-of-central-directory comment field.
|
|
// It can only be called before Close.
|
|
func (w *Writer) SetComment(comment string) error {
|
|
if len(comment) > uint16max {
|
|
return errors.New("zip: Writer.Comment too long")
|
|
}
|
|
w.comment = comment
|
|
return nil
|
|
}
|
|
|
|
// Close finishes writing the zip file by writing the central directory.
|
|
// It does not close the underlying writer.
|
|
func (w *Writer) Close() error {
|
|
if w.last != nil && !w.last.closed {
|
|
if err := w.last.close(); err != nil {
|
|
return err
|
|
}
|
|
w.last = nil
|
|
}
|
|
if w.closed {
|
|
return errors.New("zip: writer closed twice")
|
|
}
|
|
w.closed = true
|
|
|
|
// write central directory
|
|
start := w.cw.count
|
|
for _, h := range w.dir {
|
|
var buf [directoryHeaderLen]byte
|
|
b := writeBuf(buf[:])
|
|
b.uint32(uint32(directoryHeaderSignature))
|
|
b.uint16(h.CreatorVersion)
|
|
b.uint16(h.ReaderVersion)
|
|
b.uint16(h.Flags)
|
|
b.uint16(h.Method)
|
|
b.uint16(h.ModifiedTime)
|
|
b.uint16(h.ModifiedDate)
|
|
b.uint32(h.CRC32)
|
|
if h.isZip64() || h.offset >= uint32max {
|
|
// the file needs a zip64 header. store maxint in both
|
|
// 32 bit size fields (and offset later) to signal that the
|
|
// zip64 extra header should be used.
|
|
b.uint32(uint32max) // compressed size
|
|
b.uint32(uint32max) // uncompressed size
|
|
|
|
// append a zip64 extra block to Extra
|
|
var buf [28]byte // 2x uint16 + 3x uint64
|
|
eb := writeBuf(buf[:])
|
|
eb.uint16(zip64ExtraID)
|
|
eb.uint16(24) // size = 3x uint64
|
|
eb.uint64(h.UncompressedSize64)
|
|
eb.uint64(h.CompressedSize64)
|
|
eb.uint64(h.offset)
|
|
h.Extra = append(h.Extra, buf[:]...)
|
|
} else {
|
|
b.uint32(h.CompressedSize)
|
|
b.uint32(h.UncompressedSize)
|
|
}
|
|
|
|
b.uint16(uint16(len(h.Name)))
|
|
b.uint16(uint16(len(h.Extra)))
|
|
b.uint16(uint16(len(h.Comment)))
|
|
b = b[4:] // skip disk number start and internal file attr (2x uint16)
|
|
b.uint32(h.ExternalAttrs)
|
|
if h.offset > uint32max {
|
|
b.uint32(uint32max)
|
|
} else {
|
|
b.uint32(uint32(h.offset))
|
|
}
|
|
if _, err := w.cw.Write(buf[:]); err != nil {
|
|
return err
|
|
}
|
|
if _, err := io.WriteString(w.cw, h.Name); err != nil {
|
|
return err
|
|
}
|
|
if _, err := w.cw.Write(h.Extra); err != nil {
|
|
return err
|
|
}
|
|
if _, err := io.WriteString(w.cw, h.Comment); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
end := w.cw.count
|
|
|
|
records := uint64(len(w.dir))
|
|
size := uint64(end - start)
|
|
offset := uint64(start)
|
|
|
|
if f := w.testHookCloseSizeOffset; f != nil {
|
|
f(size, offset)
|
|
}
|
|
|
|
if records >= uint16max || size >= uint32max || offset >= uint32max {
|
|
var buf [directory64EndLen + directory64LocLen]byte
|
|
b := writeBuf(buf[:])
|
|
|
|
// zip64 end of central directory record
|
|
b.uint32(directory64EndSignature)
|
|
b.uint64(directory64EndLen - 12) // length minus signature (uint32) and length fields (uint64)
|
|
b.uint16(zipVersion45) // version made by
|
|
b.uint16(zipVersion45) // version needed to extract
|
|
b.uint32(0) // number of this disk
|
|
b.uint32(0) // number of the disk with the start of the central directory
|
|
b.uint64(records) // total number of entries in the central directory on this disk
|
|
b.uint64(records) // total number of entries in the central directory
|
|
b.uint64(size) // size of the central directory
|
|
b.uint64(offset) // offset of start of central directory with respect to the starting disk number
|
|
|
|
// zip64 end of central directory locator
|
|
b.uint32(directory64LocSignature)
|
|
b.uint32(0) // number of the disk with the start of the zip64 end of central directory
|
|
b.uint64(uint64(end)) // relative offset of the zip64 end of central directory record
|
|
b.uint32(1) // total number of disks
|
|
|
|
if _, err := w.cw.Write(buf[:]); err != nil {
|
|
return err
|
|
}
|
|
|
|
// store max values in the regular end record to signal
|
|
// that the zip64 values should be used instead
|
|
records = uint16max
|
|
size = uint32max
|
|
offset = uint32max
|
|
}
|
|
|
|
// write end record
|
|
var buf [directoryEndLen]byte
|
|
b := writeBuf(buf[:])
|
|
b.uint32(uint32(directoryEndSignature))
|
|
b = b[4:] // skip over disk number and first disk number (2x uint16)
|
|
b.uint16(uint16(records)) // number of entries this disk
|
|
b.uint16(uint16(records)) // number of entries total
|
|
b.uint32(uint32(size)) // size of directory
|
|
b.uint32(uint32(offset)) // start of directory
|
|
b.uint16(uint16(len(w.comment))) // byte size of EOCD comment
|
|
if _, err := w.cw.Write(buf[:]); err != nil {
|
|
return err
|
|
}
|
|
if _, err := io.WriteString(w.cw, w.comment); err != nil {
|
|
return err
|
|
}
|
|
|
|
return w.cw.w.(*bufio.Writer).Flush()
|
|
}
|
|
|
|
// Create adds a file to the zip file using the provided name.
|
|
// It returns a Writer to which the file contents should be written.
|
|
// The file contents will be compressed using the Deflate method.
|
|
// The name must be a relative path: it must not start with a drive
|
|
// letter (e.g. C:) or leading slash, and only forward slashes are
|
|
// allowed. To create a directory instead of a file, add a trailing
|
|
// slash to the name.
|
|
// The file's contents must be written to the io.Writer before the next
|
|
// call to Create, CreateHeader, or Close.
|
|
func (w *Writer) Create(name string) (io.Writer, error) {
|
|
header := &FileHeader{
|
|
Name: name,
|
|
Method: Deflate,
|
|
}
|
|
return w.CreateHeader(header)
|
|
}
|
|
|
|
// detectUTF8 reports whether s is a valid UTF-8 string, and whether the string
|
|
// must be considered UTF-8 encoding (i.e., not compatible with CP-437, ASCII,
|
|
// or any other common encoding).
|
|
func detectUTF8(s string) (valid, require bool) {
|
|
for i := 0; i < len(s); {
|
|
r, size := utf8.DecodeRuneInString(s[i:])
|
|
i += size
|
|
// Officially, ZIP uses CP-437, but many readers use the system's
|
|
// local character encoding. Most encoding are compatible with a large
|
|
// subset of CP-437, which itself is ASCII-like.
|
|
//
|
|
// Forbid 0x7e and 0x5c since EUC-KR and Shift-JIS replace those
|
|
// characters with localized currency and overline characters.
|
|
if r < 0x20 || r > 0x7d || r == 0x5c {
|
|
if !utf8.ValidRune(r) || (r == utf8.RuneError && size == 1) {
|
|
return false, false
|
|
}
|
|
require = true
|
|
}
|
|
}
|
|
return true, require
|
|
}
|
|
|
|
// prepare performs the bookkeeping operations required at the start of
|
|
// CreateHeader and CreateRaw.
|
|
func (w *Writer) prepare(fh *FileHeader) error {
|
|
if w.last != nil && !w.last.closed {
|
|
if err := w.last.close(); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
if len(w.dir) > 0 && w.dir[len(w.dir)-1].FileHeader == fh {
|
|
// See https://golang.org/issue/11144 confusion.
|
|
return errors.New("archive/zip: invalid duplicate FileHeader")
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// CreateHeader adds a file to the zip archive using the provided FileHeader
|
|
// for the file metadata. Writer takes ownership of fh and may mutate
|
|
// its fields. The caller must not modify fh after calling CreateHeader.
|
|
//
|
|
// This returns a Writer to which the file contents should be written.
|
|
// The file's contents must be written to the io.Writer before the next
|
|
// call to Create, CreateHeader, CreateRaw, or Close.
|
|
func (w *Writer) CreateHeader(fh *FileHeader) (io.Writer, error) {
|
|
if err := w.prepare(fh); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// The ZIP format has a sad state of affairs regarding character encoding.
|
|
// Officially, the name and comment fields are supposed to be encoded
|
|
// in CP-437 (which is mostly compatible with ASCII), unless the UTF-8
|
|
// flag bit is set. However, there are several problems:
|
|
//
|
|
// * Many ZIP readers still do not support UTF-8.
|
|
// * If the UTF-8 flag is cleared, several readers simply interpret the
|
|
// name and comment fields as whatever the local system encoding is.
|
|
//
|
|
// In order to avoid breaking readers without UTF-8 support,
|
|
// we avoid setting the UTF-8 flag if the strings are CP-437 compatible.
|
|
// However, if the strings require multibyte UTF-8 encoding and is a
|
|
// valid UTF-8 string, then we set the UTF-8 bit.
|
|
//
|
|
// For the case, where the user explicitly wants to specify the encoding
|
|
// as UTF-8, they will need to set the flag bit themselves.
|
|
utf8Valid1, utf8Require1 := detectUTF8(fh.Name)
|
|
utf8Valid2, utf8Require2 := detectUTF8(fh.Comment)
|
|
switch {
|
|
case fh.NonUTF8:
|
|
fh.Flags &^= 0x800
|
|
case (utf8Require1 || utf8Require2) && (utf8Valid1 && utf8Valid2):
|
|
fh.Flags |= 0x800
|
|
}
|
|
|
|
fh.CreatorVersion = fh.CreatorVersion&0xff00 | zipVersion20 // preserve compatibility byte
|
|
fh.ReaderVersion = zipVersion20
|
|
|
|
// If Modified is set, this takes precedence over MS-DOS timestamp fields.
|
|
if !fh.Modified.IsZero() {
|
|
// Contrary to the FileHeader.SetModTime method, we intentionally
|
|
// do not convert to UTC, because we assume the user intends to encode
|
|
// the date using the specified timezone. A user may want this control
|
|
// because many legacy ZIP readers interpret the timestamp according
|
|
// to the local timezone.
|
|
//
|
|
// The timezone is only non-UTC if a user directly sets the Modified
|
|
// field directly themselves. All other approaches sets UTC.
|
|
fh.ModifiedDate, fh.ModifiedTime = timeToMsDosTime(fh.Modified)
|
|
|
|
// Use "extended timestamp" format since this is what Info-ZIP uses.
|
|
// Nearly every major ZIP implementation uses a different format,
|
|
// but at least most seem to be able to understand the other formats.
|
|
//
|
|
// This format happens to be identical for both local and central header
|
|
// if modification time is the only timestamp being encoded.
|
|
var mbuf [9]byte // 2*SizeOf(uint16) + SizeOf(uint8) + SizeOf(uint32)
|
|
mt := uint32(fh.Modified.Unix())
|
|
eb := writeBuf(mbuf[:])
|
|
eb.uint16(extTimeExtraID)
|
|
eb.uint16(5) // Size: SizeOf(uint8) + SizeOf(uint32)
|
|
eb.uint8(1) // Flags: ModTime
|
|
eb.uint32(mt) // ModTime
|
|
fh.Extra = append(fh.Extra, mbuf[:]...)
|
|
}
|
|
|
|
var (
|
|
ow io.Writer
|
|
fw *fileWriter
|
|
)
|
|
h := &header{
|
|
FileHeader: fh,
|
|
offset: uint64(w.cw.count),
|
|
}
|
|
|
|
if strings.HasSuffix(fh.Name, "/") {
|
|
// Set the compression method to Store to ensure data length is truly zero,
|
|
// which the writeHeader method always encodes for the size fields.
|
|
// This is necessary as most compression formats have non-zero lengths
|
|
// even when compressing an empty string.
|
|
fh.Method = Store
|
|
fh.Flags &^= 0x8 // we will not write a data descriptor
|
|
|
|
// Explicitly clear sizes as they have no meaning for directories.
|
|
fh.CompressedSize = 0
|
|
fh.CompressedSize64 = 0
|
|
fh.UncompressedSize = 0
|
|
fh.UncompressedSize64 = 0
|
|
|
|
ow = dirWriter{}
|
|
} else {
|
|
fh.Flags |= 0x8 // we will write a data descriptor
|
|
|
|
fw = &fileWriter{
|
|
zipw: w.cw,
|
|
compCount: &countWriter{w: w.cw},
|
|
crc32: crc32.NewIEEE(),
|
|
}
|
|
comp := w.compressor(fh.Method)
|
|
if comp == nil {
|
|
return nil, ErrAlgorithm
|
|
}
|
|
var err error
|
|
fw.comp, err = comp(fw.compCount)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
fw.rawCount = &countWriter{w: fw.comp}
|
|
fw.header = h
|
|
ow = fw
|
|
}
|
|
w.dir = append(w.dir, h)
|
|
if err := writeHeader(w.cw, h); err != nil {
|
|
return nil, err
|
|
}
|
|
// If we're creating a directory, fw is nil.
|
|
w.last = fw
|
|
return ow, nil
|
|
}
|
|
|
|
func writeHeader(w io.Writer, h *header) error {
|
|
const maxUint16 = 1<<16 - 1
|
|
if len(h.Name) > maxUint16 {
|
|
return errLongName
|
|
}
|
|
if len(h.Extra) > maxUint16 {
|
|
return errLongExtra
|
|
}
|
|
|
|
var buf [fileHeaderLen]byte
|
|
b := writeBuf(buf[:])
|
|
b.uint32(uint32(fileHeaderSignature))
|
|
b.uint16(h.ReaderVersion)
|
|
b.uint16(h.Flags)
|
|
b.uint16(h.Method)
|
|
b.uint16(h.ModifiedTime)
|
|
b.uint16(h.ModifiedDate)
|
|
// In raw mode (caller does the compression), the values are either
|
|
// written here or in the trailing data descriptor based on the header
|
|
// flags.
|
|
if h.raw && !h.hasDataDescriptor() {
|
|
b.uint32(h.CRC32)
|
|
b.uint32(uint32(min64(h.CompressedSize64, uint32max)))
|
|
b.uint32(uint32(min64(h.UncompressedSize64, uint32max)))
|
|
} else {
|
|
// When this package handle the compression, these values are
|
|
// always written to the trailing data descriptor.
|
|
b.uint32(0) // crc32
|
|
b.uint32(0) // compressed size
|
|
b.uint32(0) // uncompressed size
|
|
}
|
|
b.uint16(uint16(len(h.Name)))
|
|
b.uint16(uint16(len(h.Extra)))
|
|
if _, err := w.Write(buf[:]); err != nil {
|
|
return err
|
|
}
|
|
if _, err := io.WriteString(w, h.Name); err != nil {
|
|
return err
|
|
}
|
|
_, err := w.Write(h.Extra)
|
|
return err
|
|
}
|
|
|
|
func min64(x, y uint64) uint64 {
|
|
if x < y {
|
|
return x
|
|
}
|
|
return y
|
|
}
|
|
|
|
// CreateRaw adds a file to the zip archive using the provided FileHeader and
|
|
// returns a Writer to which the file contents should be written. The file's
|
|
// contents must be written to the io.Writer before the next call to Create,
|
|
// CreateHeader, CreateRaw, or Close.
|
|
//
|
|
// In contrast to CreateHeader, the bytes passed to Writer are not compressed.
|
|
func (w *Writer) CreateRaw(fh *FileHeader) (io.Writer, error) {
|
|
if err := w.prepare(fh); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
fh.CompressedSize = uint32(min64(fh.CompressedSize64, uint32max))
|
|
fh.UncompressedSize = uint32(min64(fh.UncompressedSize64, uint32max))
|
|
|
|
h := &header{
|
|
FileHeader: fh,
|
|
offset: uint64(w.cw.count),
|
|
raw: true,
|
|
}
|
|
w.dir = append(w.dir, h)
|
|
if err := writeHeader(w.cw, h); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if strings.HasSuffix(fh.Name, "/") {
|
|
w.last = nil
|
|
return dirWriter{}, nil
|
|
}
|
|
|
|
fw := &fileWriter{
|
|
header: h,
|
|
zipw: w.cw,
|
|
}
|
|
w.last = fw
|
|
return fw, nil
|
|
}
|
|
|
|
// Copy copies the file f (obtained from a Reader) into w. It copies the raw
|
|
// form directly bypassing decompression, compression, and validation.
|
|
func (w *Writer) Copy(f *File) error {
|
|
r, err := f.OpenRaw()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
fw, err := w.CreateRaw(&f.FileHeader)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
_, err = io.Copy(fw, r)
|
|
return err
|
|
}
|
|
|
|
// RegisterCompressor registers or overrides a custom compressor for a specific
|
|
// method ID. If a compressor for a given method is not found, Writer will
|
|
// default to looking up the compressor at the package level.
|
|
func (w *Writer) RegisterCompressor(method uint16, comp Compressor) {
|
|
if w.compressors == nil {
|
|
w.compressors = make(map[uint16]Compressor)
|
|
}
|
|
w.compressors[method] = comp
|
|
}
|
|
|
|
func (w *Writer) compressor(method uint16) Compressor {
|
|
comp := w.compressors[method]
|
|
if comp == nil {
|
|
comp = compressor(method)
|
|
}
|
|
return comp
|
|
}
|
|
|
|
type dirWriter struct{}
|
|
|
|
func (dirWriter) Write(b []byte) (int, error) {
|
|
if len(b) == 0 {
|
|
return 0, nil
|
|
}
|
|
return 0, errors.New("zip: write to directory")
|
|
}
|
|
|
|
type fileWriter struct {
|
|
*header
|
|
zipw io.Writer
|
|
rawCount *countWriter
|
|
comp io.WriteCloser
|
|
compCount *countWriter
|
|
crc32 hash.Hash32
|
|
closed bool
|
|
}
|
|
|
|
func (w *fileWriter) Write(p []byte) (int, error) {
|
|
if w.closed {
|
|
return 0, errors.New("zip: write to closed file")
|
|
}
|
|
if w.raw {
|
|
return w.zipw.Write(p)
|
|
}
|
|
w.crc32.Write(p)
|
|
return w.rawCount.Write(p)
|
|
}
|
|
|
|
func (w *fileWriter) close() error {
|
|
if w.closed {
|
|
return errors.New("zip: file closed twice")
|
|
}
|
|
w.closed = true
|
|
if w.raw {
|
|
return w.writeDataDescriptor()
|
|
}
|
|
if err := w.comp.Close(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// update FileHeader
|
|
fh := w.header.FileHeader
|
|
fh.CRC32 = w.crc32.Sum32()
|
|
fh.CompressedSize64 = uint64(w.compCount.count)
|
|
fh.UncompressedSize64 = uint64(w.rawCount.count)
|
|
|
|
if fh.isZip64() {
|
|
fh.CompressedSize = uint32max
|
|
fh.UncompressedSize = uint32max
|
|
fh.ReaderVersion = zipVersion45 // requires 4.5 - File uses ZIP64 format extensions
|
|
} else {
|
|
fh.CompressedSize = uint32(fh.CompressedSize64)
|
|
fh.UncompressedSize = uint32(fh.UncompressedSize64)
|
|
}
|
|
|
|
return w.writeDataDescriptor()
|
|
}
|
|
|
|
func (w *fileWriter) writeDataDescriptor() error {
|
|
if !w.hasDataDescriptor() {
|
|
return nil
|
|
}
|
|
// Write data descriptor. This is more complicated than one would
|
|
// think, see e.g. comments in zipfile.c:putextended() and
|
|
// http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=7073588.
|
|
// The approach here is to write 8 byte sizes if needed without
|
|
// adding a zip64 extra in the local header (too late anyway).
|
|
var buf []byte
|
|
if w.isZip64() {
|
|
buf = make([]byte, dataDescriptor64Len)
|
|
} else {
|
|
buf = make([]byte, dataDescriptorLen)
|
|
}
|
|
b := writeBuf(buf)
|
|
b.uint32(dataDescriptorSignature) // de-facto standard, required by OS X
|
|
b.uint32(w.CRC32)
|
|
if w.isZip64() {
|
|
b.uint64(w.CompressedSize64)
|
|
b.uint64(w.UncompressedSize64)
|
|
} else {
|
|
b.uint32(w.CompressedSize)
|
|
b.uint32(w.UncompressedSize)
|
|
}
|
|
_, err := w.zipw.Write(buf)
|
|
return err
|
|
}
|
|
|
|
type countWriter struct {
|
|
w io.Writer
|
|
count int64
|
|
}
|
|
|
|
func (w *countWriter) Write(p []byte) (int, error) {
|
|
n, err := w.w.Write(p)
|
|
w.count += int64(n)
|
|
return n, err
|
|
}
|
|
|
|
type nopCloser struct {
|
|
io.Writer
|
|
}
|
|
|
|
func (w nopCloser) Close() error {
|
|
return nil
|
|
}
|
|
|
|
type writeBuf []byte
|
|
|
|
func (b *writeBuf) uint8(v uint8) {
|
|
(*b)[0] = v
|
|
*b = (*b)[1:]
|
|
}
|
|
|
|
func (b *writeBuf) uint16(v uint16) {
|
|
binary.LittleEndian.PutUint16(*b, v)
|
|
*b = (*b)[2:]
|
|
}
|
|
|
|
func (b *writeBuf) uint32(v uint32) {
|
|
binary.LittleEndian.PutUint32(*b, v)
|
|
*b = (*b)[4:]
|
|
}
|
|
|
|
func (b *writeBuf) uint64(v uint64) {
|
|
binary.LittleEndian.PutUint64(*b, v)
|
|
*b = (*b)[8:]
|
|
}
|