Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
M
mybee
Project
Project
Details
Activity
Releases
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
vicotor
mybee
Commits
cd0fb0af
Unverified
Commit
cd0fb0af
authored
May 20, 2021
by
acud
Committed by
GitHub
May 20, 2021
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
chore: swipe migrate v0.5.3 to v0.6.0 (#1716)
parent
f76f08bc
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
272 additions
and
3 deletions
+272
-3
migration.go
pkg/localstore/migration.go
+271
-3
node.go
pkg/node/node.go
+1
-0
No files found.
pkg/localstore/migration.go
View file @
cd0fb0af
...
@@ -17,8 +17,14 @@
...
@@ -17,8 +17,14 @@
package
localstore
package
localstore
import
(
import
(
"encoding/binary"
"errors"
"errors"
"fmt"
"fmt"
"time"
"github.com/ethersphere/bee/pkg/shed"
"github.com/ethersphere/bee/pkg/swarm"
"github.com/syndtr/goleveldb/leveldb"
)
)
var
errMissingCurrentSchema
=
errors
.
New
(
"could not find current db schema"
)
var
errMissingCurrentSchema
=
errors
.
New
(
"could not find current db schema"
)
...
@@ -33,9 +39,7 @@ type migration struct {
...
@@ -33,9 +39,7 @@ type migration struct {
// in order to run data migrations in the correct sequence
// in order to run data migrations in the correct sequence
var
schemaMigrations
=
[]
migration
{
var
schemaMigrations
=
[]
migration
{
{
name
:
DbSchemaCode
,
fn
:
func
(
_
*
DB
)
error
{
return
nil
}},
{
name
:
DbSchemaCode
,
fn
:
func
(
_
*
DB
)
error
{
return
nil
}},
{
name
:
DbSchemaYuj
,
fn
:
func
(
_
*
DB
)
error
{
{
name
:
DbSchemaYuj
,
fn
:
migrateYuj
},
return
errors
.
New
(
"db schema incompatible. you must migrate your data manually. see the release notes for more info"
)
}},
}
}
func
(
db
*
DB
)
migrate
(
schemaName
string
)
error
{
func
(
db
*
DB
)
migrate
(
schemaName
string
)
error
{
...
@@ -101,3 +105,267 @@ func getMigrations(currentSchema, targetSchema string, allSchemeMigrations []mig
...
@@ -101,3 +105,267 @@ func getMigrations(currentSchema, targetSchema string, allSchemeMigrations []mig
}
}
return
migrations
,
nil
return
migrations
,
nil
}
}
// migrateYuj removes all existing database content, unless
// pinned content is detected, in which case it aborts the
// operation for the user to resolve.
func
migrateYuj
(
db
*
DB
)
error
{
pinIndex
,
err
:=
db
.
shed
.
NewIndex
(
"Hash->PinCounter"
,
shed
.
IndexFuncs
{
EncodeKey
:
func
(
fields
shed
.
Item
)
(
key
[]
byte
,
err
error
)
{
return
fields
.
Address
,
nil
},
DecodeKey
:
func
(
key
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
Address
=
key
return
e
,
nil
},
EncodeValue
:
func
(
fields
shed
.
Item
)
(
value
[]
byte
,
err
error
)
{
b
:=
make
([]
byte
,
8
)
binary
.
BigEndian
.
PutUint64
(
b
[
:
8
],
fields
.
PinCounter
)
return
b
,
nil
},
DecodeValue
:
func
(
keyItem
shed
.
Item
,
value
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
PinCounter
=
binary
.
BigEndian
.
Uint64
(
value
[
:
8
])
return
e
,
nil
},
})
if
err
!=
nil
{
return
err
}
hasPinned
:=
false
_
=
pinIndex
.
Iterate
(
func
(
item
shed
.
Item
)
(
stop
bool
,
err
error
)
{
hasPinned
=
true
return
true
,
nil
},
nil
)
if
hasPinned
{
return
errors
.
New
(
"failed to update your node due to the existence of pinned content. Please refer to the release notes on how to safely migrate your pinned content."
)
}
// define the old indexes from the previous schema
// and swipe them clean.
retrievalDataIndex
,
err
:=
db
.
shed
.
NewIndex
(
"Address->StoreTimestamp|BinID|Data"
,
shed
.
IndexFuncs
{
EncodeKey
:
func
(
fields
shed
.
Item
)
(
key
[]
byte
,
err
error
)
{
return
fields
.
Address
,
nil
},
DecodeKey
:
func
(
key
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
Address
=
key
return
e
,
nil
},
EncodeValue
:
func
(
fields
shed
.
Item
)
(
value
[]
byte
,
err
error
)
{
b
:=
make
([]
byte
,
16
)
binary
.
BigEndian
.
PutUint64
(
b
[
:
8
],
fields
.
BinID
)
binary
.
BigEndian
.
PutUint64
(
b
[
8
:
16
],
uint64
(
fields
.
StoreTimestamp
))
value
=
append
(
b
,
fields
.
Data
...
)
return
value
,
nil
},
DecodeValue
:
func
(
keyItem
shed
.
Item
,
value
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
StoreTimestamp
=
int64
(
binary
.
BigEndian
.
Uint64
(
value
[
8
:
16
]))
e
.
BinID
=
binary
.
BigEndian
.
Uint64
(
value
[
:
8
])
e
.
Data
=
value
[
16
:
]
return
e
,
nil
},
})
if
err
!=
nil
{
return
err
}
retrievalAccessIndex
,
err
:=
db
.
shed
.
NewIndex
(
"Address->AccessTimestamp"
,
shed
.
IndexFuncs
{
EncodeKey
:
func
(
fields
shed
.
Item
)
(
key
[]
byte
,
err
error
)
{
return
fields
.
Address
,
nil
},
DecodeKey
:
func
(
key
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
Address
=
key
return
e
,
nil
},
EncodeValue
:
func
(
fields
shed
.
Item
)
(
value
[]
byte
,
err
error
)
{
b
:=
make
([]
byte
,
8
)
binary
.
BigEndian
.
PutUint64
(
b
,
uint64
(
fields
.
AccessTimestamp
))
return
b
,
nil
},
DecodeValue
:
func
(
keyItem
shed
.
Item
,
value
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
AccessTimestamp
=
int64
(
binary
.
BigEndian
.
Uint64
(
value
))
return
e
,
nil
},
})
if
err
!=
nil
{
return
err
}
// pull index allows history and live syncing per po bin
pullIndex
,
err
:=
db
.
shed
.
NewIndex
(
"PO|BinID->Hash|Tag"
,
shed
.
IndexFuncs
{
EncodeKey
:
func
(
fields
shed
.
Item
)
(
key
[]
byte
,
err
error
)
{
key
=
make
([]
byte
,
41
)
key
[
0
]
=
db
.
po
(
swarm
.
NewAddress
(
fields
.
Address
))
binary
.
BigEndian
.
PutUint64
(
key
[
1
:
9
],
fields
.
BinID
)
return
key
,
nil
},
DecodeKey
:
func
(
key
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
BinID
=
binary
.
BigEndian
.
Uint64
(
key
[
1
:
9
])
return
e
,
nil
},
EncodeValue
:
func
(
fields
shed
.
Item
)
(
value
[]
byte
,
err
error
)
{
value
=
make
([]
byte
,
36
)
// 32 bytes address, 4 bytes tag
copy
(
value
,
fields
.
Address
)
if
fields
.
Tag
!=
0
{
binary
.
BigEndian
.
PutUint32
(
value
[
32
:
],
fields
.
Tag
)
}
return
value
,
nil
},
DecodeValue
:
func
(
keyItem
shed
.
Item
,
value
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
Address
=
value
[
:
32
]
if
len
(
value
)
>
32
{
e
.
Tag
=
binary
.
BigEndian
.
Uint32
(
value
[
32
:
])
}
return
e
,
nil
},
})
if
err
!=
nil
{
return
err
}
// create a vector for bin IDs
binIDs
,
err
:=
db
.
shed
.
NewUint64Vector
(
"bin-ids"
)
if
err
!=
nil
{
return
err
}
pushIndex
,
err
:=
db
.
shed
.
NewIndex
(
"StoreTimestamp|Hash->Tags"
,
shed
.
IndexFuncs
{
EncodeKey
:
func
(
fields
shed
.
Item
)
(
key
[]
byte
,
err
error
)
{
key
=
make
([]
byte
,
40
)
binary
.
BigEndian
.
PutUint64
(
key
[
:
8
],
uint64
(
fields
.
StoreTimestamp
))
copy
(
key
[
8
:
],
fields
.
Address
)
return
key
,
nil
},
DecodeKey
:
func
(
key
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
Address
=
key
[
8
:
]
e
.
StoreTimestamp
=
int64
(
binary
.
BigEndian
.
Uint64
(
key
[
:
8
]))
return
e
,
nil
},
EncodeValue
:
func
(
fields
shed
.
Item
)
(
value
[]
byte
,
err
error
)
{
tag
:=
make
([]
byte
,
4
)
binary
.
BigEndian
.
PutUint32
(
tag
,
fields
.
Tag
)
return
tag
,
nil
},
DecodeValue
:
func
(
keyItem
shed
.
Item
,
value
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
if
len
(
value
)
==
4
{
// only values with tag should be decoded
e
.
Tag
=
binary
.
BigEndian
.
Uint32
(
value
)
}
return
e
,
nil
},
})
if
err
!=
nil
{
return
err
}
gcIndex
,
err
:=
db
.
shed
.
NewIndex
(
"AccessTimestamp|BinID|Hash->nil"
,
shed
.
IndexFuncs
{
EncodeKey
:
func
(
fields
shed
.
Item
)
(
key
[]
byte
,
err
error
)
{
b
:=
make
([]
byte
,
16
,
16
+
len
(
fields
.
Address
))
binary
.
BigEndian
.
PutUint64
(
b
[
:
8
],
uint64
(
fields
.
AccessTimestamp
))
binary
.
BigEndian
.
PutUint64
(
b
[
8
:
16
],
fields
.
BinID
)
key
=
append
(
b
,
fields
.
Address
...
)
return
key
,
nil
},
DecodeKey
:
func
(
key
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
AccessTimestamp
=
int64
(
binary
.
BigEndian
.
Uint64
(
key
[
:
8
]))
e
.
BinID
=
binary
.
BigEndian
.
Uint64
(
key
[
8
:
16
])
e
.
Address
=
key
[
16
:
]
return
e
,
nil
},
EncodeValue
:
func
(
fields
shed
.
Item
)
(
value
[]
byte
,
err
error
)
{
return
nil
,
nil
},
DecodeValue
:
func
(
keyItem
shed
.
Item
,
value
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
return
e
,
nil
},
})
if
err
!=
nil
{
return
err
}
// Create a index structure for excluding pinned chunks from gcIndex
gcExcludeIndex
,
err
:=
db
.
shed
.
NewIndex
(
"Hash->nil"
,
shed
.
IndexFuncs
{
EncodeKey
:
func
(
fields
shed
.
Item
)
(
key
[]
byte
,
err
error
)
{
return
fields
.
Address
,
nil
},
DecodeKey
:
func
(
key
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
e
.
Address
=
key
return
e
,
nil
},
EncodeValue
:
func
(
fields
shed
.
Item
)
(
value
[]
byte
,
err
error
)
{
return
nil
,
nil
},
DecodeValue
:
func
(
keyItem
shed
.
Item
,
value
[]
byte
)
(
e
shed
.
Item
,
err
error
)
{
return
e
,
nil
},
})
if
err
!=
nil
{
return
err
}
var
lim
=
10000
count
:=
0
truncate
:=
func
(
i
shed
.
Index
)
error
{
batch
:=
new
(
leveldb
.
Batch
)
count
=
0
err
:=
i
.
Iterate
(
func
(
item
shed
.
Item
)
(
stop
bool
,
err
error
)
{
if
err
=
i
.
DeleteInBatch
(
batch
,
item
);
err
!=
nil
{
return
true
,
err
}
count
++
if
count
%
lim
==
0
{
db
.
logger
.
Debugf
(
"truncate writing batch. processed %d"
,
count
)
err
:=
db
.
shed
.
WriteBatch
(
batch
)
if
err
!=
nil
{
return
true
,
err
}
batch
=
new
(
leveldb
.
Batch
)
}
return
false
,
nil
},
nil
)
if
err
!=
nil
{
return
err
}
return
db
.
shed
.
WriteBatch
(
batch
)
}
start
:=
time
.
Now
()
db
.
logger
.
Debug
(
"truncating indexes"
)
for
_
,
v
:=
range
[]
struct
{
name
string
idx
shed
.
Index
}{
{
"pullsync"
,
pullIndex
},
{
"pushsync"
,
pushIndex
},
{
"gc"
,
gcIndex
},
{
"gcExclude"
,
gcExcludeIndex
},
{
"retrievalAccess"
,
retrievalAccessIndex
},
{
"retrievalData"
,
retrievalDataIndex
},
}
{
db
.
logger
.
Debugf
(
"truncating %s index"
,
v
.
name
)
err
:=
truncate
(
v
.
idx
)
if
err
!=
nil
{
return
fmt
.
Errorf
(
"truncate %s index: %w"
,
v
.
name
,
err
)
}
db
.
logger
.
Debugf
(
"truncated %d %s index entries"
,
count
,
v
.
name
)
}
gcSize
,
err
:=
db
.
shed
.
NewUint64Field
(
"gc-size"
)
if
err
!=
nil
{
return
fmt
.
Errorf
(
"gc size index: %w"
,
err
)
}
err
=
gcSize
.
Put
(
0
)
if
err
!=
nil
{
return
fmt
.
Errorf
(
"put gcsize: %w"
,
err
)
}
for
i
:=
0
;
i
<
int
(
swarm
.
MaxBins
);
i
++
{
if
err
:=
binIDs
.
Put
(
uint64
(
i
),
0
);
err
!=
nil
{
return
fmt
.
Errorf
(
"zero binsIDs: %w"
,
err
)
}
}
db
.
logger
.
Debugf
(
"done truncating indexes. took %s"
,
time
.
Since
(
start
))
return
nil
}
pkg/node/node.go
View file @
cd0fb0af
...
@@ -312,6 +312,7 @@ func NewBee(addr string, swarmAddress swarm.Address, publicKey ecdsa.PublicKey,
...
@@ -312,6 +312,7 @@ func NewBee(addr string, swarmAddress swarm.Address, publicKey ecdsa.PublicKey,
var
path
string
var
path
string
if
o
.
DataDir
!=
""
{
if
o
.
DataDir
!=
""
{
logger
.
Infof
(
"using datadir in: '%s'"
,
o
.
DataDir
)
path
=
filepath
.
Join
(
o
.
DataDir
,
"localstore"
)
path
=
filepath
.
Join
(
o
.
DataDir
,
"localstore"
)
}
}
lo
:=
&
localstore
.
Options
{
lo
:=
&
localstore
.
Options
{
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment