| // Copyright 2015 Google Inc. All rights reserved. |
| // |
| // Licensed under the Apache License, Version 2.0 (the "License"); |
| // you may not use this file except in compliance with the License. |
| // You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, software |
| // distributed under the License is distributed on an "AS IS" BASIS, |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| // See the License for the specific language governing permissions and |
| // limitations under the License. |
| |
| package zip |
| |
| import ( |
| "bytes" |
| "compress/flate" |
| "crypto/sha256" |
| "encoding/binary" |
| "errors" |
| "fmt" |
| "hash" |
| "hash/crc32" |
| "io" |
| "io/ioutil" |
| "os" |
| "path/filepath" |
| "sort" |
| "strings" |
| "sync" |
| "syscall" |
| "time" |
| |
| "android/soong/response" |
| |
| "github.com/google/blueprint/pathtools" |
| |
| "android/soong/jar" |
| "android/soong/third_party/zip" |
| ) |
| |
| // Sha256HeaderID is a custom Header ID for the `extra` field in |
| // the file header to store the SHA checksum. |
| const Sha256HeaderID = 0x4967 |
| |
| // Sha256HeaderSignature is the signature to verify that the extra |
| // data block is used to store the SHA checksum. |
| const Sha256HeaderSignature = 0x9514 |
| |
| // Block size used during parallel compression of a single file. |
| const parallelBlockSize = 1 * 1024 * 1024 // 1MB |
| |
| // Minimum file size to use parallel compression. It requires more |
| // flate.Writer allocations, since we can't change the dictionary |
| // during Reset |
| const minParallelFileSize = parallelBlockSize * 6 |
| |
| // Size of the ZIP compression window (32KB) |
| const windowSize = 32 * 1024 |
| |
| type nopCloser struct { |
| io.Writer |
| } |
| |
| func (nopCloser) Close() error { |
| return nil |
| } |
| |
| type byteReaderCloser struct { |
| *bytes.Reader |
| io.Closer |
| } |
| |
| type pathMapping struct { |
| dest, src string |
| zipMethod uint16 |
| } |
| |
| type FileArg struct { |
| PathPrefixInZip, SourcePrefixToStrip string |
| ExplicitPathInZip string |
| SourceFiles []string |
| JunkPaths bool |
| GlobDir string |
| } |
| |
| type FileArgsBuilder struct { |
| state FileArg |
| err error |
| fs pathtools.FileSystem |
| |
| fileArgs []FileArg |
| } |
| |
| func NewFileArgsBuilder() *FileArgsBuilder { |
| return &FileArgsBuilder{ |
| fs: pathtools.OsFs, |
| } |
| } |
| |
| func (b *FileArgsBuilder) JunkPaths(v bool) *FileArgsBuilder { |
| b.state.JunkPaths = v |
| b.state.SourcePrefixToStrip = "" |
| return b |
| } |
| |
| func (b *FileArgsBuilder) SourcePrefixToStrip(prefixToStrip string) *FileArgsBuilder { |
| b.state.JunkPaths = false |
| b.state.SourcePrefixToStrip = prefixToStrip |
| return b |
| } |
| |
| func (b *FileArgsBuilder) PathPrefixInZip(rootPrefix string) *FileArgsBuilder { |
| b.state.PathPrefixInZip = rootPrefix |
| return b |
| } |
| |
| func (b *FileArgsBuilder) File(name string) *FileArgsBuilder { |
| if b.err != nil { |
| return b |
| } |
| |
| arg := b.state |
| arg.SourceFiles = []string{name} |
| b.fileArgs = append(b.fileArgs, arg) |
| |
| if b.state.ExplicitPathInZip != "" { |
| b.state.ExplicitPathInZip = "" |
| } |
| return b |
| } |
| |
| func (b *FileArgsBuilder) Dir(name string) *FileArgsBuilder { |
| if b.err != nil { |
| return b |
| } |
| |
| arg := b.state |
| arg.GlobDir = name |
| b.fileArgs = append(b.fileArgs, arg) |
| return b |
| } |
| |
| // List reads the file names from the given file and adds them to the source files list. |
| func (b *FileArgsBuilder) List(name string) *FileArgsBuilder { |
| if b.err != nil { |
| return b |
| } |
| |
| f, err := b.fs.Open(name) |
| if err != nil { |
| b.err = err |
| return b |
| } |
| defer f.Close() |
| |
| list, err := ioutil.ReadAll(f) |
| if err != nil { |
| b.err = err |
| return b |
| } |
| |
| arg := b.state |
| arg.SourceFiles = strings.Fields(string(list)) |
| b.fileArgs = append(b.fileArgs, arg) |
| return b |
| } |
| |
| // RspFile reads the file names from given .rsp file and adds them to the source files list. |
| func (b *FileArgsBuilder) RspFile(name string) *FileArgsBuilder { |
| if b.err != nil { |
| return b |
| } |
| |
| f, err := b.fs.Open(name) |
| if err != nil { |
| b.err = err |
| return b |
| } |
| defer f.Close() |
| |
| arg := b.state |
| arg.SourceFiles, err = response.ReadRspFile(f) |
| if err != nil { |
| b.err = err |
| return b |
| } |
| for i := range arg.SourceFiles { |
| arg.SourceFiles[i] = pathtools.MatchEscape(arg.SourceFiles[i]) |
| } |
| b.fileArgs = append(b.fileArgs, arg) |
| return b |
| } |
| |
| // ExplicitPathInZip sets the path in the zip file for the next File call. |
| func (b *FileArgsBuilder) ExplicitPathInZip(s string) *FileArgsBuilder { |
| b.state.ExplicitPathInZip = s |
| return b |
| } |
| |
| func (b *FileArgsBuilder) Error() error { |
| if b == nil { |
| return nil |
| } |
| return b.err |
| } |
| |
| func (b *FileArgsBuilder) FileArgs() []FileArg { |
| if b == nil { |
| return nil |
| } |
| return b.fileArgs |
| } |
| |
| type IncorrectRelativeRootError struct { |
| RelativeRoot string |
| Path string |
| } |
| |
| func (x IncorrectRelativeRootError) Error() string { |
| return fmt.Sprintf("path %q is outside relative root %q", x.Path, x.RelativeRoot) |
| } |
| |
| type ConflictingFileError struct { |
| Dest string |
| Prev string |
| Src string |
| } |
| |
| func (x ConflictingFileError) Error() string { |
| return fmt.Sprintf("destination %q has two files %q and %q", x.Dest, x.Prev, x.Src) |
| } |
| |
| type ZipWriter struct { |
| time time.Time |
| createdFiles map[string]string |
| createdDirs map[string]string |
| directories bool |
| |
| errors chan error |
| writeOps chan chan *zipEntry |
| |
| cpuRateLimiter *CPURateLimiter |
| memoryRateLimiter *MemoryRateLimiter |
| |
| compressorPool sync.Pool |
| compLevel int |
| |
| followSymlinks pathtools.ShouldFollowSymlinks |
| ignoreMissingFiles bool |
| |
| stderr io.Writer |
| fs pathtools.FileSystem |
| |
| sha256Checksum bool |
| } |
| |
| type zipEntry struct { |
| fh *zip.FileHeader |
| |
| // List of delayed io.Reader |
| futureReaders chan chan io.Reader |
| |
| // Only used for passing into the MemoryRateLimiter to ensure we |
| // release as much memory as much as we request |
| allocatedSize int64 |
| } |
| |
| type ZipArgs struct { |
| FileArgs []FileArg |
| OutputFilePath string |
| EmulateJar bool |
| SrcJar bool |
| AddDirectoryEntriesToZip bool |
| CompressionLevel int |
| ManifestSourcePath string |
| NumParallelJobs int |
| NonDeflatedFiles map[string]bool |
| WriteIfChanged bool |
| StoreSymlinks bool |
| IgnoreMissingFiles bool |
| Sha256Checksum bool |
| DoNotWrite bool |
| Quiet bool |
| |
| Stderr io.Writer |
| Filesystem pathtools.FileSystem |
| } |
| |
| func zipTo(args ZipArgs, w io.Writer) error { |
| if args.EmulateJar { |
| args.AddDirectoryEntriesToZip = true |
| } |
| |
| // Have Glob follow symlinks if they are not being stored as symlinks in the zip file. |
| followSymlinks := pathtools.ShouldFollowSymlinks(!args.StoreSymlinks) |
| |
| z := &ZipWriter{ |
| time: jar.DefaultTime, |
| createdDirs: make(map[string]string), |
| createdFiles: make(map[string]string), |
| directories: args.AddDirectoryEntriesToZip, |
| compLevel: args.CompressionLevel, |
| followSymlinks: followSymlinks, |
| ignoreMissingFiles: args.IgnoreMissingFiles, |
| stderr: args.Stderr, |
| fs: args.Filesystem, |
| sha256Checksum: args.Sha256Checksum, |
| } |
| |
| if z.fs == nil { |
| z.fs = pathtools.OsFs |
| } |
| |
| if z.stderr == nil { |
| z.stderr = os.Stderr |
| } |
| |
| pathMappings := []pathMapping{} |
| |
| noCompression := args.CompressionLevel == 0 |
| |
| for _, fa := range args.FileArgs { |
| var srcs []string |
| for _, s := range fa.SourceFiles { |
| s = strings.TrimSpace(s) |
| if s == "" { |
| continue |
| } |
| |
| result, err := z.fs.Glob(s, nil, followSymlinks) |
| if err != nil { |
| return err |
| } |
| if len(result.Matches) == 0 { |
| err := &os.PathError{ |
| Op: "lstat", |
| Path: s, |
| Err: os.ErrNotExist, |
| } |
| if args.IgnoreMissingFiles { |
| if !args.Quiet { |
| fmt.Fprintln(z.stderr, "warning:", err) |
| } |
| } else { |
| return err |
| } |
| } |
| srcs = append(srcs, result.Matches...) |
| } |
| if fa.GlobDir != "" { |
| if exists, isDir, err := z.fs.Exists(fa.GlobDir); err != nil { |
| return err |
| } else if !exists && !args.IgnoreMissingFiles { |
| err := &os.PathError{ |
| Op: "lstat", |
| Path: fa.GlobDir, |
| Err: os.ErrNotExist, |
| } |
| if args.IgnoreMissingFiles { |
| if !args.Quiet { |
| fmt.Fprintln(z.stderr, "warning:", err) |
| } |
| } else { |
| return err |
| } |
| } else if !isDir && !args.IgnoreMissingFiles { |
| err := &os.PathError{ |
| Op: "lstat", |
| Path: fa.GlobDir, |
| Err: syscall.ENOTDIR, |
| } |
| if args.IgnoreMissingFiles { |
| if !args.Quiet { |
| fmt.Fprintln(z.stderr, "warning:", err) |
| } |
| } else { |
| return err |
| } |
| } |
| result, err := z.fs.Glob(filepath.Join(fa.GlobDir, "**/*"), nil, followSymlinks) |
| if err != nil { |
| return err |
| } |
| srcs = append(srcs, result.Matches...) |
| } |
| for _, src := range srcs { |
| err := fillPathPairs(fa, src, &pathMappings, args.NonDeflatedFiles, noCompression) |
| if err != nil { |
| return err |
| } |
| } |
| } |
| |
| return z.write(w, pathMappings, args.ManifestSourcePath, args.EmulateJar, args.SrcJar, args.NumParallelJobs) |
| } |
| |
| // Zip creates an output zip archive from given sources. |
| func Zip(args ZipArgs) error { |
| if args.OutputFilePath == "" { |
| return fmt.Errorf("output file path must be nonempty") |
| } |
| |
| buf := &bytes.Buffer{} |
| var out io.Writer = buf |
| |
| var zipErr error |
| |
| if args.DoNotWrite { |
| out = io.Discard |
| } else if !args.WriteIfChanged { |
| f, err := os.Create(args.OutputFilePath) |
| if err != nil { |
| return err |
| } |
| |
| defer f.Close() |
| defer func() { |
| if zipErr != nil { |
| os.Remove(args.OutputFilePath) |
| } |
| }() |
| |
| out = f |
| } |
| |
| zipErr = zipTo(args, out) |
| if zipErr != nil { |
| return zipErr |
| } |
| |
| if args.WriteIfChanged && !args.DoNotWrite { |
| err := pathtools.WriteFileIfChanged(args.OutputFilePath, buf.Bytes(), 0666) |
| if err != nil { |
| return err |
| } |
| } |
| |
| return nil |
| } |
| |
| func fillPathPairs(fa FileArg, src string, pathMappings *[]pathMapping, |
| nonDeflatedFiles map[string]bool, noCompression bool) error { |
| |
| var dest string |
| |
| if fa.ExplicitPathInZip != "" { |
| dest = fa.ExplicitPathInZip |
| } else if fa.JunkPaths { |
| dest = filepath.Base(src) |
| } else { |
| var err error |
| dest, err = filepath.Rel(fa.SourcePrefixToStrip, src) |
| if err != nil { |
| return err |
| } |
| if strings.HasPrefix(dest, "../") { |
| return IncorrectRelativeRootError{ |
| Path: src, |
| RelativeRoot: fa.SourcePrefixToStrip, |
| } |
| } |
| } |
| dest = filepath.Join(fa.PathPrefixInZip, dest) |
| |
| zipMethod := zip.Deflate |
| if _, found := nonDeflatedFiles[dest]; found || noCompression { |
| zipMethod = zip.Store |
| } |
| *pathMappings = append(*pathMappings, |
| pathMapping{dest: dest, src: src, zipMethod: zipMethod}) |
| |
| return nil |
| } |
| |
| func jarSort(mappings []pathMapping) { |
| sort.SliceStable(mappings, func(i int, j int) bool { |
| return jar.EntryNamesLess(mappings[i].dest, mappings[j].dest) |
| }) |
| } |
| |
| func (z *ZipWriter) write(f io.Writer, pathMappings []pathMapping, manifest string, emulateJar, srcJar bool, |
| parallelJobs int) error { |
| |
| z.errors = make(chan error) |
| defer close(z.errors) |
| |
| // This channel size can be essentially unlimited -- it's used as a fifo |
| // queue decouple the CPU and IO loads. Directories don't require any |
| // compression time, but still cost some IO. Similar with small files that |
| // can be very fast to compress. Some files that are more difficult to |
| // compress won't take a corresponding longer time writing out. |
| // |
| // The optimum size here depends on your CPU and IO characteristics, and |
| // the the layout of your zip file. 1000 was chosen mostly at random as |
| // something that worked reasonably well for a test file. |
| // |
| // The RateLimit object will put the upper bounds on the number of |
| // parallel compressions and outstanding buffers. |
| z.writeOps = make(chan chan *zipEntry, 1000) |
| z.cpuRateLimiter = NewCPURateLimiter(int64(parallelJobs)) |
| z.memoryRateLimiter = NewMemoryRateLimiter(0) |
| defer func() { |
| z.cpuRateLimiter.Stop() |
| z.memoryRateLimiter.Stop() |
| }() |
| |
| if manifest != "" && !emulateJar { |
| return errors.New("must specify --jar when specifying a manifest via -m") |
| } |
| |
| if emulateJar { |
| // manifest may be empty, in which case addManifest will fill in a default |
| pathMappings = append(pathMappings, pathMapping{jar.ManifestFile, manifest, zip.Deflate}) |
| |
| jarSort(pathMappings) |
| } |
| |
| go func() { |
| var err error |
| defer close(z.writeOps) |
| |
| for _, ele := range pathMappings { |
| if emulateJar && ele.dest == jar.ManifestFile { |
| err = z.addManifest(ele.dest, ele.src, ele.zipMethod) |
| } else { |
| err = z.addFile(ele.dest, ele.src, ele.zipMethod, emulateJar, srcJar) |
| } |
| if err != nil { |
| z.errors <- err |
| return |
| } |
| } |
| }() |
| |
| zipw := zip.NewWriter(f) |
| |
| var currentWriteOpChan chan *zipEntry |
| var currentWriter io.WriteCloser |
| var currentReaders chan chan io.Reader |
| var currentReader chan io.Reader |
| var done bool |
| |
| for !done { |
| var writeOpsChan chan chan *zipEntry |
| var writeOpChan chan *zipEntry |
| var readersChan chan chan io.Reader |
| |
| if currentReader != nil { |
| // Only read and process errors |
| } else if currentReaders != nil { |
| readersChan = currentReaders |
| } else if currentWriteOpChan != nil { |
| writeOpChan = currentWriteOpChan |
| } else { |
| writeOpsChan = z.writeOps |
| } |
| |
| select { |
| case writeOp, ok := <-writeOpsChan: |
| if !ok { |
| done = true |
| } |
| |
| currentWriteOpChan = writeOp |
| |
| case op := <-writeOpChan: |
| currentWriteOpChan = nil |
| |
| var err error |
| if op.fh.Method == zip.Deflate { |
| currentWriter, err = zipw.CreateCompressedHeader(op.fh) |
| } else { |
| var zw io.Writer |
| |
| op.fh.CompressedSize64 = op.fh.UncompressedSize64 |
| |
| zw, err = zipw.CreateHeaderAndroid(op.fh) |
| currentWriter = nopCloser{zw} |
| } |
| if err != nil { |
| return err |
| } |
| |
| currentReaders = op.futureReaders |
| if op.futureReaders == nil { |
| currentWriter.Close() |
| currentWriter = nil |
| } |
| z.memoryRateLimiter.Finish(op.allocatedSize) |
| |
| case futureReader, ok := <-readersChan: |
| if !ok { |
| // Done with reading |
| currentWriter.Close() |
| currentWriter = nil |
| currentReaders = nil |
| } |
| |
| currentReader = futureReader |
| |
| case reader := <-currentReader: |
| _, err := io.Copy(currentWriter, reader) |
| if err != nil { |
| return err |
| } |
| |
| currentReader = nil |
| |
| case err := <-z.errors: |
| return err |
| } |
| } |
| |
| // One last chance to catch an error |
| select { |
| case err := <-z.errors: |
| return err |
| default: |
| zipw.Close() |
| return nil |
| } |
| } |
| |
| // imports (possibly with compression) <src> into the zip at sub-path <dest> |
| func (z *ZipWriter) addFile(dest, src string, method uint16, emulateJar, srcJar bool) error { |
| var fileSize int64 |
| var executable bool |
| |
| var s os.FileInfo |
| var err error |
| if z.followSymlinks { |
| s, err = z.fs.Stat(src) |
| } else { |
| s, err = z.fs.Lstat(src) |
| } |
| |
| if err != nil { |
| if os.IsNotExist(err) && z.ignoreMissingFiles { |
| fmt.Fprintln(z.stderr, "warning:", err) |
| return nil |
| } |
| return err |
| } |
| |
| createParentDirs := func(dest, src string) error { |
| if err := z.writeDirectory(filepath.Dir(dest), src, emulateJar); err != nil { |
| return err |
| } |
| |
| if prev, exists := z.createdDirs[dest]; exists { |
| return fmt.Errorf("destination %q is both a directory %q and a file %q", dest, prev, src) |
| } |
| |
| return nil |
| } |
| |
| checkDuplicateFiles := func(dest, src string) (bool, error) { |
| if prev, exists := z.createdFiles[dest]; exists { |
| if prev != src { |
| return true, ConflictingFileError{ |
| Dest: dest, |
| Prev: prev, |
| Src: src, |
| } |
| } |
| return true, nil |
| } |
| |
| z.createdFiles[dest] = src |
| return false, nil |
| } |
| |
| if s.IsDir() { |
| if z.directories { |
| return z.writeDirectory(dest, src, emulateJar) |
| } |
| return nil |
| } else if s.Mode()&os.ModeSymlink != 0 { |
| err = createParentDirs(dest, src) |
| if err != nil { |
| return err |
| } |
| |
| duplicate, err := checkDuplicateFiles(dest, src) |
| if err != nil { |
| return err |
| } |
| if duplicate { |
| return nil |
| } |
| |
| return z.writeSymlink(dest, src) |
| } else if s.Mode().IsRegular() { |
| r, err := z.fs.Open(src) |
| if err != nil { |
| return err |
| } |
| |
| if srcJar && filepath.Ext(src) == ".java" { |
| // rewrite the destination using the package path if it can be determined |
| pkg, err := jar.JavaPackage(r, src) |
| if err != nil { |
| // ignore errors for now, leaving the file at in its original location in the zip |
| } else { |
| dest = filepath.Join(filepath.Join(strings.Split(pkg, ".")...), filepath.Base(src)) |
| } |
| |
| _, err = r.Seek(0, io.SeekStart) |
| if err != nil { |
| return err |
| } |
| } |
| |
| fileSize = s.Size() |
| executable = s.Mode()&0100 != 0 |
| |
| header := &zip.FileHeader{ |
| Name: dest, |
| Method: method, |
| UncompressedSize64: uint64(fileSize), |
| } |
| |
| mode := os.FileMode(0644) |
| if executable { |
| mode = 0755 |
| } |
| header.SetMode(mode) |
| |
| err = createParentDirs(dest, src) |
| if err != nil { |
| return err |
| } |
| |
| duplicate, err := checkDuplicateFiles(dest, src) |
| if err != nil { |
| return err |
| } |
| if duplicate { |
| return nil |
| } |
| |
| return z.writeFileContents(header, r) |
| } else { |
| return fmt.Errorf("%s is not a file, directory, or symlink", src) |
| } |
| } |
| |
| func (z *ZipWriter) addManifest(dest string, src string, _ uint16) error { |
| if prev, exists := z.createdDirs[dest]; exists { |
| return fmt.Errorf("destination %q is both a directory %q and a file %q", dest, prev, src) |
| } |
| if prev, exists := z.createdFiles[dest]; exists { |
| if prev != src { |
| return ConflictingFileError{ |
| Dest: dest, |
| Prev: prev, |
| Src: src, |
| } |
| } |
| return nil |
| } |
| |
| if err := z.writeDirectory(filepath.Dir(dest), src, true); err != nil { |
| return err |
| } |
| |
| var contents []byte |
| if src != "" { |
| f, err := z.fs.Open(src) |
| if err != nil { |
| return err |
| } |
| |
| contents, err = ioutil.ReadAll(f) |
| f.Close() |
| if err != nil { |
| return err |
| } |
| } |
| |
| fh, buf, err := jar.ManifestFileContents(contents) |
| if err != nil { |
| return err |
| } |
| |
| reader := &byteReaderCloser{bytes.NewReader(buf), ioutil.NopCloser(nil)} |
| |
| return z.writeFileContents(fh, reader) |
| } |
| |
| func (z *ZipWriter) writeFileContents(header *zip.FileHeader, r pathtools.ReaderAtSeekerCloser) (err error) { |
| |
| header.SetModTime(z.time) |
| |
| compressChan := make(chan *zipEntry, 1) |
| z.writeOps <- compressChan |
| |
| // Pre-fill a zipEntry, it will be sent in the compressChan once |
| // we're sure about the Method and CRC. |
| ze := &zipEntry{ |
| fh: header, |
| } |
| |
| ze.allocatedSize = int64(header.UncompressedSize64) |
| z.cpuRateLimiter.Request() |
| z.memoryRateLimiter.Request(ze.allocatedSize) |
| |
| fileSize := int64(header.UncompressedSize64) |
| if fileSize == 0 { |
| fileSize = int64(header.UncompressedSize) |
| } |
| |
| if header.Method == zip.Deflate && fileSize >= minParallelFileSize { |
| wg := new(sync.WaitGroup) |
| |
| // Allocate enough buffer to hold all readers. We'll limit |
| // this based on actual buffer sizes in RateLimit. |
| ze.futureReaders = make(chan chan io.Reader, (fileSize/parallelBlockSize)+1) |
| |
| // Calculate the CRC and SHA256 in the background, since reading |
| // the entire file could take a while. |
| // |
| // We could split this up into chunks as well, but it's faster |
| // than the compression. Due to the Go Zip API, we also need to |
| // know the result before we can begin writing the compressed |
| // data out to the zipfile. |
| // |
| // We calculate SHA256 only if `-sha256` is set. |
| wg.Add(1) |
| go z.checksumFileAsync(r, ze, compressChan, wg) |
| |
| for start := int64(0); start < fileSize; start += parallelBlockSize { |
| sr := io.NewSectionReader(r, start, parallelBlockSize) |
| resultChan := make(chan io.Reader, 1) |
| ze.futureReaders <- resultChan |
| |
| z.cpuRateLimiter.Request() |
| |
| last := !(start+parallelBlockSize < fileSize) |
| var dict []byte |
| if start >= windowSize { |
| dict, err = ioutil.ReadAll(io.NewSectionReader(r, start-windowSize, windowSize)) |
| if err != nil { |
| return err |
| } |
| } |
| |
| wg.Add(1) |
| go z.compressPartialFile(sr, dict, last, resultChan, wg) |
| } |
| |
| close(ze.futureReaders) |
| |
| // Close the file handle after all readers are done |
| go func(wg *sync.WaitGroup, closer io.Closer) { |
| wg.Wait() |
| closer.Close() |
| }(wg, r) |
| } else { |
| go func() { |
| z.compressWholeFile(ze, r, compressChan) |
| r.Close() |
| }() |
| } |
| |
| return nil |
| } |
| |
| func (z *ZipWriter) checksumFileAsync(r io.ReadSeeker, ze *zipEntry, resultChan chan *zipEntry, wg *sync.WaitGroup) { |
| defer wg.Done() |
| defer z.cpuRateLimiter.Finish() |
| |
| z.checksumFile(r, ze) |
| |
| resultChan <- ze |
| close(resultChan) |
| } |
| |
| func (z *ZipWriter) checksumFile(r io.ReadSeeker, ze *zipEntry) { |
| crc := crc32.NewIEEE() |
| writers := []io.Writer{crc} |
| |
| var shaHasher hash.Hash |
| if z.sha256Checksum && !ze.fh.Mode().IsDir() { |
| shaHasher = sha256.New() |
| writers = append(writers, shaHasher) |
| } |
| |
| w := io.MultiWriter(writers...) |
| |
| _, err := io.Copy(w, r) |
| if err != nil { |
| z.errors <- err |
| return |
| } |
| |
| ze.fh.CRC32 = crc.Sum32() |
| if shaHasher != nil { |
| z.appendSHAToExtra(ze, shaHasher.Sum(nil)) |
| } |
| } |
| |
| func (z *ZipWriter) appendSHAToExtra(ze *zipEntry, checksum []byte) { |
| // The block of SHA256 checksum consist of: |
| // - Header ID, equals to Sha256HeaderID (2 bytes) |
| // - Data size (2 bytes) |
| // - Data block: |
| // - Signature, equals to Sha256HeaderSignature (2 bytes) |
| // - Data, SHA checksum value |
| var buf []byte |
| buf = binary.LittleEndian.AppendUint16(buf, Sha256HeaderID) |
| buf = binary.LittleEndian.AppendUint16(buf, uint16(len(checksum)+2)) |
| buf = binary.LittleEndian.AppendUint16(buf, Sha256HeaderSignature) |
| buf = append(buf, checksum...) |
| ze.fh.Extra = append(ze.fh.Extra, buf...) |
| } |
| |
| func (z *ZipWriter) compressPartialFile(r io.Reader, dict []byte, last bool, resultChan chan io.Reader, wg *sync.WaitGroup) { |
| defer wg.Done() |
| |
| result, err := z.compressBlock(r, dict, last) |
| if err != nil { |
| z.errors <- err |
| return |
| } |
| |
| z.cpuRateLimiter.Finish() |
| |
| resultChan <- result |
| } |
| |
| func (z *ZipWriter) compressBlock(r io.Reader, dict []byte, last bool) (*bytes.Buffer, error) { |
| buf := new(bytes.Buffer) |
| var fw *flate.Writer |
| var err error |
| if len(dict) > 0 { |
| // There's no way to Reset a Writer with a new dictionary, so |
| // don't use the Pool |
| fw, err = flate.NewWriterDict(buf, z.compLevel, dict) |
| } else { |
| var ok bool |
| if fw, ok = z.compressorPool.Get().(*flate.Writer); ok { |
| fw.Reset(buf) |
| } else { |
| fw, err = flate.NewWriter(buf, z.compLevel) |
| } |
| defer z.compressorPool.Put(fw) |
| } |
| if err != nil { |
| return nil, err |
| } |
| |
| _, err = io.Copy(fw, r) |
| if err != nil { |
| return nil, err |
| } |
| if last { |
| fw.Close() |
| } else { |
| fw.Flush() |
| } |
| |
| return buf, nil |
| } |
| |
| func (z *ZipWriter) compressWholeFile(ze *zipEntry, r io.ReadSeeker, compressChan chan *zipEntry) { |
| z.checksumFile(r, ze) |
| |
| _, err := r.Seek(0, 0) |
| if err != nil { |
| z.errors <- err |
| return |
| } |
| |
| readFile := func(reader io.ReadSeeker) ([]byte, error) { |
| _, err := reader.Seek(0, 0) |
| if err != nil { |
| return nil, err |
| } |
| |
| buf, err := ioutil.ReadAll(reader) |
| if err != nil { |
| return nil, err |
| } |
| |
| return buf, nil |
| } |
| |
| ze.futureReaders = make(chan chan io.Reader, 1) |
| futureReader := make(chan io.Reader, 1) |
| ze.futureReaders <- futureReader |
| close(ze.futureReaders) |
| |
| if ze.fh.Method == zip.Deflate { |
| compressed, err := z.compressBlock(r, nil, true) |
| if err != nil { |
| z.errors <- err |
| return |
| } |
| if uint64(compressed.Len()) < ze.fh.UncompressedSize64 { |
| futureReader <- compressed |
| } else { |
| buf, err := readFile(r) |
| if err != nil { |
| z.errors <- err |
| return |
| } |
| ze.fh.Method = zip.Store |
| futureReader <- bytes.NewReader(buf) |
| } |
| } else { |
| buf, err := readFile(r) |
| if err != nil { |
| z.errors <- err |
| return |
| } |
| ze.fh.Method = zip.Store |
| futureReader <- bytes.NewReader(buf) |
| } |
| |
| z.cpuRateLimiter.Finish() |
| |
| close(futureReader) |
| |
| compressChan <- ze |
| close(compressChan) |
| } |
| |
| // writeDirectory annotates that dir is a directory created for the src file or directory, and adds |
| // the directory entry to the zip file if directories are enabled. |
| func (z *ZipWriter) writeDirectory(dir string, src string, emulateJar bool) error { |
| // clean the input |
| dir = filepath.Clean(dir) |
| |
| // discover any uncreated directories in the path |
| var zipDirs []string |
| for dir != "" && dir != "." { |
| if _, exists := z.createdDirs[dir]; exists { |
| break |
| } |
| |
| if prev, exists := z.createdFiles[dir]; exists { |
| return fmt.Errorf("destination %q is both a directory %q and a file %q", dir, src, prev) |
| } |
| |
| z.createdDirs[dir] = src |
| // parent directories precede their children |
| zipDirs = append([]string{dir}, zipDirs...) |
| |
| dir = filepath.Dir(dir) |
| } |
| |
| if z.directories { |
| // make a directory entry for each uncreated directory |
| for _, cleanDir := range zipDirs { |
| var dirHeader *zip.FileHeader |
| |
| if emulateJar && cleanDir+"/" == jar.MetaDir { |
| dirHeader = jar.MetaDirFileHeader() |
| } else { |
| dirHeader = &zip.FileHeader{ |
| Name: cleanDir + "/", |
| } |
| dirHeader.SetMode(0755 | os.ModeDir) |
| } |
| |
| dirHeader.SetModTime(z.time) |
| |
| ze := make(chan *zipEntry, 1) |
| ze <- &zipEntry{ |
| fh: dirHeader, |
| } |
| close(ze) |
| z.writeOps <- ze |
| } |
| } |
| |
| return nil |
| } |
| |
| func (z *ZipWriter) writeSymlink(rel, file string) error { |
| fileHeader := &zip.FileHeader{ |
| Name: rel, |
| } |
| fileHeader.SetModTime(z.time) |
| fileHeader.SetMode(0777 | os.ModeSymlink) |
| |
| dest, err := z.fs.Readlink(file) |
| if err != nil { |
| return err |
| } |
| |
| fileHeader.UncompressedSize64 = uint64(len(dest)) |
| fileHeader.CRC32 = crc32.ChecksumIEEE([]byte(dest)) |
| |
| ze := make(chan *zipEntry, 1) |
| futureReaders := make(chan chan io.Reader, 1) |
| futureReader := make(chan io.Reader, 1) |
| futureReaders <- futureReader |
| close(futureReaders) |
| futureReader <- bytes.NewBufferString(dest) |
| close(futureReader) |
| |
| ze <- &zipEntry{ |
| fh: fileHeader, |
| futureReaders: futureReaders, |
| } |
| close(ze) |
| z.writeOps <- ze |
| |
| return nil |
| } |