aboutsummaryrefslogtreecommitdiffstats
path: root/swarm/storage/localstore/export.go
diff options
context:
space:
mode:
Diffstat (limited to 'swarm/storage/localstore/export.go')
-rw-r--r--swarm/storage/localstore/export.go204
1 files changed, 0 insertions, 204 deletions
diff --git a/swarm/storage/localstore/export.go b/swarm/storage/localstore/export.go
deleted file mode 100644
index 411392b4e..000000000
--- a/swarm/storage/localstore/export.go
+++ /dev/null
@@ -1,204 +0,0 @@
-// Copyright 2019 The go-ethereum Authors
-// This file is part of the go-ethereum library.
-//
-// The go-ethereum library is free software: you can redistribute it and/or modify
-// it under the terms of the GNU Lesser General Public License as published by
-// the Free Software Foundation, either version 3 of the License, or
-// (at your option) any later version.
-//
-// The go-ethereum library is distributed in the hope that it will be useful,
-// but WITHOUT ANY WARRANTY; without even the implied warranty of
-// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-// GNU Lesser General Public License for more details.
-//
-// You should have received a copy of the GNU Lesser General Public License
-// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
-
-package localstore
-
-import (
- "archive/tar"
- "context"
- "encoding/hex"
- "fmt"
- "io"
- "io/ioutil"
- "sync"
-
- "github.com/ethereum/go-ethereum/swarm/chunk"
- "github.com/ethereum/go-ethereum/swarm/log"
- "github.com/ethereum/go-ethereum/swarm/shed"
-)
-
-const (
- // filename in tar archive that holds the information
- // about exported data format version
- exportVersionFilename = ".swarm-export-version"
- // legacy version for previous LDBStore
- legacyExportVersion = "1"
- // current export format version
- currentExportVersion = "2"
-)
-
-// Export writes a tar structured data to the writer of
-// all chunks in the retrieval data index. It returns the
-// number of chunks exported.
-func (db *DB) Export(w io.Writer) (count int64, err error) {
- tw := tar.NewWriter(w)
- defer tw.Close()
-
- if err := tw.WriteHeader(&tar.Header{
- Name: exportVersionFilename,
- Mode: 0644,
- Size: int64(len(currentExportVersion)),
- }); err != nil {
- return 0, err
- }
- if _, err := tw.Write([]byte(currentExportVersion)); err != nil {
- return 0, err
- }
-
- err = db.retrievalDataIndex.Iterate(func(item shed.Item) (stop bool, err error) {
- hdr := &tar.Header{
- Name: hex.EncodeToString(item.Address),
- Mode: 0644,
- Size: int64(len(item.Data)),
- }
- if err := tw.WriteHeader(hdr); err != nil {
- return false, err
- }
- if _, err := tw.Write(item.Data); err != nil {
- return false, err
- }
- count++
- return false, nil
- }, nil)
-
- return count, err
-}
-
-// Import reads a tar structured data from the reader and
-// stores chunks in the database. It returns the number of
-// chunks imported.
-func (db *DB) Import(r io.Reader, legacy bool) (count int64, err error) {
- tr := tar.NewReader(r)
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- errC := make(chan error)
- doneC := make(chan struct{})
- tokenPool := make(chan struct{}, 100)
- var wg sync.WaitGroup
- go func() {
- var (
- firstFile = true
- // if exportVersionFilename file is not present
- // assume legacy version
- version = legacyExportVersion
- )
- for {
- hdr, err := tr.Next()
- if err != nil {
- if err == io.EOF {
- break
- }
- select {
- case errC <- err:
- case <-ctx.Done():
- }
- }
- if firstFile {
- firstFile = false
- if hdr.Name == exportVersionFilename {
- data, err := ioutil.ReadAll(tr)
- if err != nil {
- select {
- case errC <- err:
- case <-ctx.Done():
- }
- }
- version = string(data)
- continue
- }
- }
-
- if len(hdr.Name) != 64 {
- log.Warn("ignoring non-chunk file", "name", hdr.Name)
- continue
- }
-
- keybytes, err := hex.DecodeString(hdr.Name)
- if err != nil {
- log.Warn("ignoring invalid chunk file", "name", hdr.Name, "err", err)
- continue
- }
-
- data, err := ioutil.ReadAll(tr)
- if err != nil {
- select {
- case errC <- err:
- case <-ctx.Done():
- }
- }
- key := chunk.Address(keybytes)
-
- var ch chunk.Chunk
- switch version {
- case legacyExportVersion:
- // LDBStore Export exported chunk data prefixed with the chunk key.
- // That is not necessary, as the key is in the chunk filename,
- // but backward compatibility needs to be preserved.
- ch = chunk.NewChunk(key, data[32:])
- case currentExportVersion:
- ch = chunk.NewChunk(key, data)
- default:
- select {
- case errC <- fmt.Errorf("unsupported export data version %q", version):
- case <-ctx.Done():
- }
- }
- tokenPool <- struct{}{}
- wg.Add(1)
-
- go func() {
- _, err := db.Put(ctx, chunk.ModePutUpload, ch)
- select {
- case errC <- err:
- case <-ctx.Done():
- wg.Done()
- <-tokenPool
- default:
- _, err := db.Put(ctx, chunk.ModePutUpload, ch)
- if err != nil {
- errC <- err
- }
- wg.Done()
- <-tokenPool
- }
- }()
-
- count++
- }
- wg.Wait()
- close(doneC)
- }()
-
- // wait for all chunks to be stored
- for {
- select {
- case err := <-errC:
- if err != nil {
- return count, err
- }
- case <-ctx.Done():
- return count, ctx.Err()
- default:
- select {
- case <-doneC:
- return count, nil
- default:
- }
- }
- }
-}