aboutsummaryrefslogtreecommitdiffstats
path: root/eth
diff options
context:
space:
mode:
authorPéter Szilágyi <peterke@gmail.com>2016-10-31 23:26:44 +0800
committerPéter Szilágyi <peterke@gmail.com>2016-10-31 23:26:44 +0800
commit8b10617bba7be804ab6516a36ae02dd6175ed46d (patch)
treef762f3b52faad7ef2345bf41ebe4d8711943d78b /eth
parent90b16a3e85d82241f4505a7db0f95c2f312642a6 (diff)
downloadgo-tangerine-8b10617bba7be804ab6516a36ae02dd6175ed46d.tar.gz
go-tangerine-8b10617bba7be804ab6516a36ae02dd6175ed46d.tar.zst
go-tangerine-8b10617bba7be804ab6516a36ae02dd6175ed46d.zip
eth/downloader: reduce fast sync block requirements, fix test
Diffstat (limited to 'eth')
-rw-r--r--eth/downloader/downloader.go10
-rw-r--r--eth/downloader/downloader_test.go31
2 files changed, 28 insertions, 13 deletions
diff --git a/eth/downloader/downloader.go b/eth/downloader/downloader.go
index 1cf69ea71..683794470 100644
--- a/eth/downloader/downloader.go
+++ b/eth/downloader/downloader.go
@@ -67,8 +67,8 @@ var (
fsHeaderCheckFrequency = 100 // Verification frequency of the downloaded headers during fast sync
fsHeaderSafetyNet = 2048 // Number of headers to discard in case a chain violation is detected
fsHeaderForceVerify = 24 // Number of headers to verify before and after the pivot to accept it
- fsPivotInterval = 512 // Number of headers out of which to randomize the pivot point
- fsMinFullBlocks = 1024 // Number of blocks to retrieve fully even in fast sync
+ fsPivotInterval = 256 // Number of headers out of which to randomize the pivot point
+ fsMinFullBlocks = 64 // Number of blocks to retrieve fully even in fast sync
fsCriticalTrials = uint32(32) // Number of times to retry in the cricical section before bailing
)
@@ -480,6 +480,11 @@ func (d *Downloader) spawnSync(origin uint64, fetchers ...func() error) error {
d.queue.Close()
d.cancel()
wg.Wait()
+
+ // If sync failed in the critical section, bump the fail counter
+ if err != nil && d.mode == FastSync && d.fsPivotLock != nil {
+ atomic.AddUint32(&d.fsPivotFails, 1)
+ }
return err
}
@@ -1190,7 +1195,6 @@ func (d *Downloader) processHeaders(origin uint64, td *big.Int) error {
}
}
}
- atomic.AddUint32(&d.fsPivotFails, 1)
}
}
}()
diff --git a/eth/downloader/downloader_test.go b/eth/downloader/downloader_test.go
index df5d7febd..2849712ab 100644
--- a/eth/downloader/downloader_test.go
+++ b/eth/downloader/downloader_test.go
@@ -358,7 +358,7 @@ func (dl *downloadTester) insertBlocks(blocks types.Blocks) (int, error) {
return len(blocks), nil
}
-// insertReceipts injects a new batch of blocks into the simulated chain.
+// insertReceipts injects a new batch of receipts into the simulated chain.
func (dl *downloadTester) insertReceipts(blocks types.Blocks, receipts []types.Receipts) (int, error) {
dl.lock.Lock()
defer dl.lock.Unlock()
@@ -1199,7 +1199,7 @@ func testInvalidHeaderRollback(t *testing.T, protocol int, mode SyncMode) {
defer tester.terminate()
// Create a small enough block chain to download
- targetBlocks := 3*fsHeaderSafetyNet + fsMinFullBlocks
+ targetBlocks := 3*fsHeaderSafetyNet + fsPivotInterval + fsMinFullBlocks
hashes, headers, blocks, receipts := tester.makeChain(targetBlocks, 0, tester.genesis, nil, false)
// Attempt to sync with an attacker that feeds junk during the fast sync phase.
@@ -1722,8 +1722,6 @@ func TestFastCriticalRestartsCont63(t *testing.T) { testFastCriticalRestarts(t,
func TestFastCriticalRestartsCont64(t *testing.T) { testFastCriticalRestarts(t, 64, true) }
func testFastCriticalRestarts(t *testing.T, protocol int, progress bool) {
- t.Parallel()
-
tester := newTester()
defer tester.terminate()
@@ -1731,12 +1729,16 @@ func testFastCriticalRestarts(t *testing.T, protocol int, progress bool) {
targetBlocks := fsMinFullBlocks + 2*fsPivotInterval - 15
hashes, headers, blocks, receipts := tester.makeChain(targetBlocks, 0, tester.genesis, nil, false)
- // Create a tester peer with the critical section state roots missing (force failures)
+ // Create a tester peer with a critical section header missing (force failures)
tester.newPeer("peer", protocol, hashes, headers, blocks, receipts)
+ delete(tester.peerHeaders["peer"], hashes[fsMinFullBlocks-1])
+ tester.downloader.dropPeer = func(id string) {} // We reuse the same "faulty" peer throughout the test
+
+ // Remove all possible pivot state roots and slow down replies (test failure resets later)
for i := 0; i < fsPivotInterval; i++ {
tester.peerMissingStates["peer"][headers[hashes[fsMinFullBlocks+i]].Root] = true
}
- tester.downloader.dropPeer = func(id string) {} // We reuse the same "faulty" peer throughout the test
+ tester.downloader.peers.peers["peer"].getNodeData = tester.peerGetNodeDataFn("peer", 500*time.Millisecond) // Enough to reach the critical section
// Synchronise with the peer a few times and make sure they fail until the retry limit
for i := 0; i < int(fsCriticalTrials)-1; i++ {
@@ -1744,12 +1746,18 @@ func testFastCriticalRestarts(t *testing.T, protocol int, progress bool) {
if err := tester.sync("peer", nil, FastSync); err == nil {
t.Fatalf("failing fast sync succeeded: %v", err)
}
- time.Sleep(100 * time.Millisecond) // Make sure no in-flight requests remain
+ time.Sleep(150 * time.Millisecond) // Make sure no in-flight requests remain
// If it's the first failure, pivot should be locked => reenable all others to detect pivot changes
if i == 0 {
+ if tester.downloader.fsPivotLock == nil {
+ time.Sleep(400 * time.Millisecond) // Make sure the first huge timeout expires too
+ t.Fatalf("pivot block not locked in after critical section failure")
+ }
tester.lock.Lock()
+ tester.peerHeaders["peer"][hashes[fsMinFullBlocks-1]] = headers[hashes[fsMinFullBlocks-1]]
tester.peerMissingStates["peer"] = map[common.Hash]bool{tester.downloader.fsPivotLock.Root: true}
+ tester.downloader.peers.peers["peer"].getNodeData = tester.peerGetNodeDataFn("peer", 0)
tester.lock.Unlock()
}
}
@@ -1762,16 +1770,17 @@ func testFastCriticalRestarts(t *testing.T, protocol int, progress bool) {
if err := tester.sync("peer", nil, FastSync); err != nil {
t.Fatalf("failed to synchronise blocks in progressed fast sync: %v", err)
}
- time.Sleep(100 * time.Millisecond) // Make sure no in-flight requests remain
+ time.Sleep(150 * time.Millisecond) // Make sure no in-flight requests remain
if fails := atomic.LoadUint32(&tester.downloader.fsPivotFails); fails != 1 {
t.Fatalf("progressed pivot trial count mismatch: have %v, want %v", fails, 1)
}
+ assertOwnChain(t, tester, targetBlocks+1)
} else {
if err := tester.sync("peer", nil, FastSync); err == nil {
t.Fatalf("succeeded to synchronise blocks in failed fast sync")
}
- time.Sleep(100 * time.Millisecond) // Make sure no in-flight requests remain
+ time.Sleep(150 * time.Millisecond) // Make sure no in-flight requests remain
if fails := atomic.LoadUint32(&tester.downloader.fsPivotFails); fails != fsCriticalTrials {
t.Fatalf("failed pivot trial count mismatch: have %v, want %v", fails, fsCriticalTrials)
@@ -1781,5 +1790,7 @@ func testFastCriticalRestarts(t *testing.T, protocol int, progress bool) {
if err := tester.sync("peer", nil, FastSync); err != nil {
t.Fatalf("failed to synchronise blocks in slow sync: %v", err)
}
- assertOwnChain(t, tester, targetBlocks+1)
+ // Note, we can't assert the chain here because the test asserter assumes sync
+ // completed using a single mode of operation, whereas fast-then-slow can result
+ // in arbitrary intermediate state that's not cleanly verifiable.
}