1
0
mirror of https://github.com/restic/restic.git synced 2024-06-30 08:20:55 +02:00
restic/archiver.go

316 lines
6.3 KiB
Go
Raw Normal View History

2014-12-05 21:45:49 +01:00
package restic
2014-09-23 22:39:12 +02:00
import (
"errors"
"fmt"
"io"
2014-09-23 22:39:12 +02:00
"os"
2014-11-16 22:50:20 +01:00
"sync"
2014-09-23 22:39:12 +02:00
2014-11-23 12:14:56 +01:00
"github.com/juju/arrar"
2014-12-05 21:45:49 +01:00
"github.com/restic/restic/backend"
"github.com/restic/restic/chunker"
2014-09-23 22:39:12 +02:00
)
2014-11-16 22:50:20 +01:00
const (
maxConcurrentFiles = 8
maxConcurrentBlobs = 8
2014-11-16 22:50:20 +01:00
)
2014-09-23 22:39:12 +02:00
type Archiver struct {
2014-12-21 18:10:19 +01:00
s Server
ch *ContentHandler
2014-11-16 22:50:20 +01:00
bl *BlobList // blobs used for the current snapshot
2014-09-23 22:39:12 +02:00
2014-11-16 22:50:20 +01:00
fileToken chan struct{}
2014-11-22 22:05:39 +01:00
blobToken chan struct{}
2014-11-16 22:50:20 +01:00
2014-09-23 22:39:12 +02:00
Error func(dir string, fi os.FileInfo, err error) error
Filter func(item string, fi os.FileInfo) bool
p *Progress
2014-09-23 22:39:12 +02:00
}
func NewArchiver(s Server, bl *BlobList, p *Progress) (*Archiver, error) {
2014-09-23 22:39:12 +02:00
var err error
2014-11-16 22:50:20 +01:00
arch := &Archiver{
2014-12-21 17:02:49 +01:00
s: s,
p: p,
2014-11-16 22:50:20 +01:00
fileToken: make(chan struct{}, maxConcurrentFiles),
2014-11-22 22:05:39 +01:00
blobToken: make(chan struct{}, maxConcurrentBlobs),
2014-11-16 22:50:20 +01:00
}
2014-11-22 22:05:39 +01:00
// fill file and blob token
2014-11-16 22:50:20 +01:00
for i := 0; i < maxConcurrentFiles; i++ {
arch.fileToken <- struct{}{}
}
2014-09-23 22:39:12 +02:00
2014-11-22 22:05:39 +01:00
for i := 0; i < maxConcurrentBlobs; i++ {
arch.blobToken <- struct{}{}
}
2014-09-23 22:39:12 +02:00
// abort on all errors
arch.Error = func(string, os.FileInfo, error) error { return err }
// allow all files
arch.Filter = func(string, os.FileInfo) bool { return true }
2014-11-21 21:21:44 +01:00
arch.bl = NewBlobList()
if bl != nil {
arch.bl.Merge(bl)
}
arch.ch = NewContentHandler(s)
2014-09-23 22:39:12 +02:00
// load all blobs from all snapshots
// TODO: only use bloblist from old snapshot if available
2014-11-23 22:26:01 +01:00
err = arch.ch.LoadAllMaps()
2014-09-23 22:39:12 +02:00
if err != nil {
return nil, err
}
return arch, nil
}
2014-11-21 21:21:44 +01:00
func (arch *Archiver) Save(t backend.Type, data []byte) (Blob, error) {
2014-09-23 22:39:12 +02:00
blob, err := arch.ch.Save(t, data)
if err != nil {
2014-11-21 21:21:44 +01:00
return Blob{}, err
2014-09-23 22:39:12 +02:00
}
// store blob in storage map for current snapshot
2014-11-21 21:21:44 +01:00
arch.bl.Insert(blob)
2014-09-23 22:39:12 +02:00
return blob, nil
}
2014-11-21 21:21:44 +01:00
func (arch *Archiver) SaveJSON(t backend.Type, item interface{}) (Blob, error) {
2014-09-23 22:39:12 +02:00
blob, err := arch.ch.SaveJSON(t, item)
if err != nil {
2014-11-21 21:21:44 +01:00
return Blob{}, err
2014-09-23 22:39:12 +02:00
}
// store blob in storage map for current snapshot
2014-11-21 21:21:44 +01:00
arch.bl.Insert(blob)
2014-09-23 22:39:12 +02:00
return blob, nil
}
// SaveFile stores the content of the file on the backend as a Blob by calling
// Save for each chunk.
func (arch *Archiver) SaveFile(node *Node) error {
file, err := os.Open(node.path)
defer file.Close()
2014-09-23 22:39:12 +02:00
if err != nil {
return err
}
// check file again
fi, err := file.Stat()
if err != nil {
return err
}
if fi.ModTime() != node.ModTime {
2015-01-08 20:49:32 +01:00
e2 := arch.Error(node.path, fi, errors.New("file was updated, using new version\n"))
if e2 == nil {
// create new node
n, err := NodeFromFileInfo(node.path, fi)
if err != nil {
return err
}
// copy node
*node = *n
}
}
var blobs Blobs
// if the file is small enough, store it directly
if node.Size < chunker.MinSize {
// acquire token
token := <-arch.blobToken
defer func() {
arch.blobToken <- token
}()
buf := GetChunkBuf("blob single file")
defer FreeChunkBuf("blob single file", buf)
n, err := io.ReadFull(file, buf)
2014-11-30 16:06:37 +01:00
if err != nil && err != io.ErrUnexpectedEOF && err != io.EOF {
return arrar.Annotate(err, "SaveFile() read small file")
}
if err == io.EOF {
// use empty blob list for empty files
blobs = Blobs{}
} else {
blob, err := arch.ch.Save(backend.Data, buf[:n])
if err != nil {
return arrar.Annotate(err, "SaveFile() save chunk")
}
arch.p.Report(Stat{Bytes: blob.Size})
blobs = Blobs{blob}
}
} else {
// else store all chunks
chnker := chunker.New(file)
2014-11-22 22:05:39 +01:00
chans := [](<-chan Blob){}
defer chnker.Free()
chunks := 0
for {
buf := GetChunkBuf("blob chunker")
chunk, err := chnker.Next(buf)
if err == io.EOF {
FreeChunkBuf("blob chunker", buf)
break
}
if err != nil {
FreeChunkBuf("blob chunker", buf)
2014-11-30 16:06:37 +01:00
return arrar.Annotate(err, "SaveFile() chunker.Next()")
}
chunks++
2014-11-22 22:05:39 +01:00
// acquire token, start goroutine to save chunk
token := <-arch.blobToken
resCh := make(chan Blob, 1)
go func(ch chan<- Blob) {
blob, err := arch.ch.Save(backend.Data, chunk.Data)
// TODO handle error
if err != nil {
panic(err)
}
FreeChunkBuf("blob chunker", buf)
arch.p.Report(Stat{Bytes: blob.Size})
2014-11-22 22:05:39 +01:00
arch.blobToken <- token
ch <- blob
}(resCh)
chans = append(chans, resCh)
}
2014-11-22 22:05:39 +01:00
blobs = []Blob{}
for _, ch := range chans {
blobs = append(blobs, <-ch)
}
if len(blobs) != chunks {
return fmt.Errorf("chunker returned %v chunks, but only %v blobs saved", chunks, len(blobs))
}
2014-09-23 22:39:12 +02:00
}
var bytes uint64
2014-09-23 22:39:12 +02:00
node.Content = make([]backend.ID, len(blobs))
for i, blob := range blobs {
node.Content[i] = blob.ID
2014-11-21 21:21:44 +01:00
arch.bl.Insert(blob)
bytes += blob.Size
}
if bytes != node.Size {
return fmt.Errorf("errors saving node %q: saved %d bytes, wanted %d bytes", node.path, bytes, node.Size)
2014-09-23 22:39:12 +02:00
}
2014-11-30 16:06:37 +01:00
return nil
2014-09-23 22:39:12 +02:00
}
2014-11-21 21:21:44 +01:00
func (arch *Archiver) saveTree(t *Tree) (Blob, error) {
2014-11-16 22:50:20 +01:00
var wg sync.WaitGroup
for _, node := range *t {
if node.tree != nil && node.Subtree == nil {
b, err := arch.saveTree(node.tree)
if err != nil {
2014-11-21 21:21:44 +01:00
return Blob{}, err
}
node.Subtree = b.ID
arch.p.Report(Stat{Dirs: 1})
} else if node.Type == "file" && len(node.Content) == 0 {
// get token
token := <-arch.fileToken
2014-11-16 22:50:20 +01:00
// start goroutine
wg.Add(1)
go func(n *Node) {
defer wg.Done()
defer func() {
arch.fileToken <- token
}()
node.err = arch.SaveFile(n)
arch.p.Report(Stat{Files: 1})
2014-11-16 22:50:20 +01:00
}(node)
2014-09-23 22:39:12 +02:00
}
}
2014-11-16 22:50:20 +01:00
wg.Wait()
2014-12-01 00:06:29 +01:00
// check for invalid file nodes
for _, node := range *t {
if node.Type == "file" && node.Content == nil && node.err == nil {
2014-12-01 00:06:29 +01:00
return Blob{}, fmt.Errorf("node %v has empty content", node.Name)
}
if node.err != nil {
err := arch.Error(node.path, nil, node.err)
if err != nil {
return Blob{}, err
}
// save error message in node
node.Error = node.err.Error()
}
2014-12-01 00:06:29 +01:00
}
blob, err := arch.SaveJSON(backend.Tree, t)
if err != nil {
2014-11-21 21:21:44 +01:00
return Blob{}, err
}
return blob, nil
}
2014-11-30 22:34:21 +01:00
func (arch *Archiver) Snapshot(dir string, t *Tree, parentSnapshot backend.ID) (*Snapshot, backend.ID, error) {
arch.p.Start()
defer arch.p.Done()
sn, err := NewSnapshot(dir)
if err != nil {
return nil, nil, err
}
2014-11-30 22:34:21 +01:00
sn.Parent = parentSnapshot
blob, err := arch.saveTree(t)
2014-09-23 22:39:12 +02:00
if err != nil {
return nil, nil, err
}
sn.Tree = blob.ID
2014-09-23 22:39:12 +02:00
2014-11-23 22:26:01 +01:00
// save bloblist
blob, err = arch.SaveJSON(backend.Map, arch.bl)
if err != nil {
return nil, nil, err
}
sn.Map = blob.Storage
2014-09-23 22:39:12 +02:00
// save snapshot
blob, err = arch.SaveJSON(backend.Snapshot, sn)
if err != nil {
return nil, nil, err
}
return sn, blob.Storage, nil
2014-09-23 22:39:12 +02:00
}