Enforce that the information used to create a bolt11 invoice using `get_bolt11_invoice()` is similar to the related instance of PaymentInfo by requiring a PaymentInfo as argument for `get_bolt11_invoice()`. This way the invoice cannot differ from the created PaymentInfo. This allows to use the information in PaymentInfo for validation of incoming htlcs more reliably. To cover all required information for the creation of a b11 invoice the PaymentInfo class has to be extended with a expiry and min_final_cltv_expiry. This requires a db upgrade.
1763 lines
69 KiB
Python
1763 lines
69 KiB
Python
#!/usr/bin/env python
|
|
#
|
|
# Electrum - lightweight Bitcoin client
|
|
# Copyright (C) 2015 Thomas Voegtlin
|
|
#
|
|
# Permission is hereby granted, free of charge, to any person
|
|
# obtaining a copy of this software and associated documentation files
|
|
# (the "Software"), to deal in the Software without restriction,
|
|
# including without limitation the rights to use, copy, modify, merge,
|
|
# publish, distribute, sublicense, and/or sell copies of the Software,
|
|
# and to permit persons to whom the Software is furnished to do so,
|
|
# subject to the following conditions:
|
|
#
|
|
# The above copyright notice and this permission notice shall be
|
|
# included in all copies or substantial portions of the Software.
|
|
#
|
|
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
# MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
# BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
# ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
# CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
# SOFTWARE.
|
|
import os
|
|
import ast
|
|
import datetime
|
|
import json
|
|
import copy
|
|
import threading
|
|
from collections import defaultdict
|
|
from typing import (Dict, Optional, List, Tuple, Set, Iterable, NamedTuple, Sequence, TYPE_CHECKING,
|
|
Union, AbstractSet)
|
|
import binascii
|
|
import time
|
|
from functools import partial
|
|
|
|
import attr
|
|
|
|
from . import util, bitcoin
|
|
from .util import profiler, WalletFileException, multisig_type, TxMinedInfo, bfh, MyEncoder
|
|
from .invoices import Invoice, Request
|
|
from .keystore import bip44_derivation
|
|
from .transaction import Transaction, TxOutpoint, tx_from_any, PartialTransaction, PartialTxOutput, BadHeaderMagic
|
|
from .logging import Logger
|
|
|
|
from .lnutil import HTLCOwner, ChannelType
|
|
from . import json_db
|
|
from .json_db import StoredDict, JsonDB, locked, modifier, StoredObject, stored_in, stored_as
|
|
from .plugin import run_hook, plugin_loaders
|
|
from .version import ELECTRUM_VERSION
|
|
|
|
if TYPE_CHECKING:
|
|
from .storage import WalletStorage
|
|
|
|
|
|
class WalletRequiresUpgrade(WalletFileException):
|
|
pass
|
|
|
|
|
|
class WalletRequiresSplit(WalletFileException):
|
|
def __init__(self, split_data):
|
|
super().__init__()
|
|
self._split_data = split_data
|
|
|
|
|
|
class WalletUnfinished(WalletFileException):
|
|
def __init__(self, wallet_db: 'WalletDB'):
|
|
super().__init__()
|
|
self._wallet_db = wallet_db
|
|
|
|
|
|
# seed_version is now used for the version of the wallet file
|
|
OLD_SEED_VERSION = 4 # electrum versions < 2.0
|
|
NEW_SEED_VERSION = 11 # electrum versions >= 2.0
|
|
FINAL_SEED_VERSION = 61 # electrum >= 2.7 will set this to prevent
|
|
# old versions from overwriting new format
|
|
|
|
|
|
@stored_in('tx_fees', tuple)
|
|
class TxFeesValue(NamedTuple):
|
|
fee: Optional[int] = None
|
|
is_calculated_by_us: bool = False
|
|
num_inputs: Optional[int] = None
|
|
|
|
|
|
@stored_as('db_metadata')
|
|
@attr.s
|
|
class DBMetadata(StoredObject):
|
|
creation_timestamp = attr.ib(default=None, type=int)
|
|
first_electrum_version_used = attr.ib(default=None, type=str)
|
|
|
|
def to_str(self) -> str:
|
|
ts = self.creation_timestamp
|
|
ver = self.first_electrum_version_used
|
|
if ts is None or ver is None:
|
|
return "unknown"
|
|
date_str = datetime.date.fromtimestamp(ts).isoformat()
|
|
return f"using {ver}, on {date_str}"
|
|
|
|
|
|
# note: subclassing WalletFileException for some specific cases
|
|
# allows the crash reporter to distinguish them and open
|
|
# separate tracking issues
|
|
class WalletFileExceptionVersion51(WalletFileException): pass
|
|
|
|
|
|
# register dicts that require value conversions not handled by constructor
|
|
json_db.register_dict('transactions', lambda x: tx_from_any(x, deserialize=False), None)
|
|
json_db.register_dict('data_loss_protect_remote_pcp', lambda x: bytes.fromhex(x), None)
|
|
json_db.register_dict('contacts', tuple, None)
|
|
# register dicts that require key conversion
|
|
for key in [
|
|
'adds', 'locked_in', 'settles', 'fails', 'fee_updates', 'buckets',
|
|
'unacked_updates', 'unfulfilled_htlcs', 'onion_keys']:
|
|
json_db.register_dict_key(key, int)
|
|
for key in ['log']:
|
|
json_db.register_dict_key(key, lambda x: HTLCOwner(int(x)))
|
|
for key in ['locked_in', 'fails', 'settles']:
|
|
json_db.register_parent_key(key, lambda x: HTLCOwner(int(x)))
|
|
|
|
|
|
class WalletDBUpgrader(Logger):
|
|
def __init__(self, data):
|
|
Logger.__init__(self)
|
|
self.data = data
|
|
|
|
def get(self, key, default=None):
|
|
return self.data.get(key, default)
|
|
|
|
def put(self, key, value):
|
|
if value is not None:
|
|
self.data[key] = value
|
|
else:
|
|
self.data.pop(key, None)
|
|
|
|
def requires_split(self):
|
|
d = self.get('accounts', {})
|
|
return len(d) > 1
|
|
|
|
def get_split_accounts(self):
|
|
result = []
|
|
# backward compatibility with old wallets
|
|
d = self.get('accounts', {})
|
|
if len(d) < 2:
|
|
return
|
|
wallet_type = self.get('wallet_type')
|
|
if wallet_type == 'old':
|
|
assert len(d) == 2
|
|
data1 = copy.deepcopy(self.data)
|
|
data1['accounts'] = {'0': d['0']}
|
|
data1['suffix'] = 'deterministic'
|
|
data2 = copy.deepcopy(self.data)
|
|
data2['accounts'] = {'/x': d['/x']}
|
|
data2['seed'] = None
|
|
data2['seed_version'] = None
|
|
data2['master_public_key'] = None
|
|
data2['wallet_type'] = 'imported'
|
|
data2['suffix'] = 'imported'
|
|
result = [data1, data2]
|
|
|
|
# note: do not add new hardware types here, this code is for converting legacy wallets
|
|
elif wallet_type in ['bip44', 'trezor', 'keepkey', 'ledger', 'btchip']:
|
|
mpk = self.get('master_public_keys')
|
|
for k in d.keys():
|
|
i = int(k)
|
|
x = d[k]
|
|
if x.get("pending"):
|
|
continue
|
|
xpub = mpk["x/%d'"%i]
|
|
new_data = copy.deepcopy(self.data)
|
|
# save account, derivation and xpub at index 0
|
|
new_data['accounts'] = {'0': x}
|
|
new_data['master_public_keys'] = {"x/0'": xpub}
|
|
new_data['derivation'] = bip44_derivation(k)
|
|
new_data['suffix'] = k
|
|
result.append(new_data)
|
|
else:
|
|
raise WalletFileException(f'Unsupported wallet type for split: {wallet_type}')
|
|
return result
|
|
|
|
def requires_upgrade(self):
|
|
return self.get_seed_version() < FINAL_SEED_VERSION
|
|
|
|
@profiler
|
|
def upgrade(self):
|
|
self.logger.info('upgrading wallet format')
|
|
self._convert_imported()
|
|
self._convert_wallet_type()
|
|
self._convert_account()
|
|
self._convert_version_13_b()
|
|
self._convert_version_14()
|
|
self._convert_version_15()
|
|
self._convert_version_16()
|
|
self._convert_version_17()
|
|
self._convert_version_18()
|
|
self._convert_version_19()
|
|
self._convert_version_20()
|
|
self._convert_version_21()
|
|
self._convert_version_22()
|
|
self._convert_version_23()
|
|
self._convert_version_24()
|
|
self._convert_version_25()
|
|
self._convert_version_26()
|
|
self._convert_version_27()
|
|
self._convert_version_28()
|
|
self._convert_version_29()
|
|
self._convert_version_30()
|
|
self._convert_version_31()
|
|
self._convert_version_32()
|
|
self._convert_version_33()
|
|
self._convert_version_34()
|
|
self._convert_version_35()
|
|
self._convert_version_36()
|
|
self._convert_version_37()
|
|
self._convert_version_38()
|
|
self._convert_version_39()
|
|
self._convert_version_40()
|
|
self._convert_version_41()
|
|
self._convert_version_42()
|
|
self._convert_version_43()
|
|
self._convert_version_44()
|
|
self._convert_version_45()
|
|
self._convert_version_46()
|
|
self._convert_version_47()
|
|
self._convert_version_48()
|
|
self._convert_version_49()
|
|
self._convert_version_50()
|
|
self._convert_version_51()
|
|
self._convert_version_52()
|
|
self._convert_version_53()
|
|
self._convert_version_54()
|
|
self._convert_version_55()
|
|
self._convert_version_56()
|
|
self._convert_version_57()
|
|
self._convert_version_58()
|
|
self._convert_version_59()
|
|
self._convert_version_60()
|
|
self._convert_version_61()
|
|
self.put('seed_version', FINAL_SEED_VERSION) # just to be sure
|
|
|
|
def _convert_wallet_type(self):
|
|
if not self._is_upgrade_method_needed(0, 13):
|
|
return
|
|
|
|
wallet_type = self.get('wallet_type')
|
|
if wallet_type == 'btchip': wallet_type = 'ledger'
|
|
if self.get('keystore') or self.get('x1/') or wallet_type=='imported':
|
|
return False
|
|
assert not self.requires_split()
|
|
seed_version = self.get_seed_version()
|
|
seed = self.get('seed')
|
|
xpubs = self.get('master_public_keys')
|
|
xprvs = self.get('master_private_keys', {})
|
|
mpk = self.get('master_public_key')
|
|
keypairs = self.get('keypairs')
|
|
key_type = self.get('key_type')
|
|
if seed_version == OLD_SEED_VERSION or wallet_type == 'old':
|
|
d = {
|
|
'type': 'old',
|
|
'seed': seed,
|
|
'mpk': mpk,
|
|
}
|
|
self.put('wallet_type', 'standard')
|
|
self.put('keystore', d)
|
|
|
|
elif key_type == 'imported':
|
|
d = {
|
|
'type': 'imported',
|
|
'keypairs': keypairs,
|
|
}
|
|
self.put('wallet_type', 'standard')
|
|
self.put('keystore', d)
|
|
|
|
elif wallet_type in ['xpub', 'standard']:
|
|
xpub = xpubs["x/"]
|
|
xprv = xprvs.get("x/")
|
|
d = {
|
|
'type': 'bip32',
|
|
'xpub': xpub,
|
|
'xprv': xprv,
|
|
'seed': seed,
|
|
}
|
|
self.put('wallet_type', 'standard')
|
|
self.put('keystore', d)
|
|
|
|
elif wallet_type in ['bip44']:
|
|
xpub = xpubs["x/0'"]
|
|
xprv = xprvs.get("x/0'")
|
|
d = {
|
|
'type': 'bip32',
|
|
'xpub': xpub,
|
|
'xprv': xprv,
|
|
}
|
|
self.put('wallet_type', 'standard')
|
|
self.put('keystore', d)
|
|
|
|
# note: do not add new hardware types here, this code is for converting legacy wallets
|
|
elif wallet_type in ['trezor', 'keepkey', 'ledger']:
|
|
xpub = xpubs["x/0'"]
|
|
derivation = self.get('derivation', bip44_derivation(0))
|
|
d = {
|
|
'type': 'hardware',
|
|
'hw_type': wallet_type,
|
|
'xpub': xpub,
|
|
'derivation': derivation,
|
|
}
|
|
self.put('wallet_type', 'standard')
|
|
self.put('keystore', d)
|
|
|
|
elif (wallet_type == '2fa') or multisig_type(wallet_type):
|
|
for key in xpubs.keys():
|
|
d = {
|
|
'type': 'bip32',
|
|
'xpub': xpubs[key],
|
|
'xprv': xprvs.get(key),
|
|
}
|
|
if key == 'x1/' and seed:
|
|
d['seed'] = seed
|
|
self.put(key, d)
|
|
else:
|
|
raise WalletFileException('Unable to tell wallet type. Is this even a wallet file?')
|
|
# remove junk
|
|
self.put('master_public_key', None)
|
|
self.put('master_public_keys', None)
|
|
self.put('master_private_keys', None)
|
|
self.put('derivation', None)
|
|
self.put('seed', None)
|
|
self.put('keypairs', None)
|
|
self.put('key_type', None)
|
|
|
|
def _convert_version_13_b(self):
|
|
# version 13 is ambiguous, and has an earlier and a later structure
|
|
if not self._is_upgrade_method_needed(0, 13):
|
|
return
|
|
|
|
if self.get('wallet_type') == 'standard':
|
|
if self.get('keystore').get('type') == 'imported':
|
|
pubkeys = self.get('keystore').get('keypairs').keys()
|
|
d = {'change': []}
|
|
receiving_addresses = []
|
|
for pubkey in pubkeys:
|
|
addr = bitcoin.pubkey_to_address('p2pkh', pubkey)
|
|
receiving_addresses.append(addr)
|
|
d['receiving'] = receiving_addresses
|
|
self.put('addresses', d)
|
|
self.put('pubkeys', None)
|
|
|
|
self.put('seed_version', 13)
|
|
|
|
def _convert_version_14(self):
|
|
# convert imported wallets for 3.0
|
|
if not self._is_upgrade_method_needed(13, 13):
|
|
return
|
|
|
|
if self.get('wallet_type') =='imported':
|
|
addresses = self.get('addresses')
|
|
if type(addresses) is list:
|
|
addresses = dict([(x, None) for x in addresses])
|
|
self.put('addresses', addresses)
|
|
elif self.get('wallet_type') == 'standard':
|
|
if self.get('keystore').get('type')=='imported':
|
|
addresses = set(self.get('addresses').get('receiving'))
|
|
pubkeys = self.get('keystore').get('keypairs').keys()
|
|
assert len(addresses) == len(pubkeys)
|
|
d = {}
|
|
for pubkey in pubkeys:
|
|
addr = bitcoin.pubkey_to_address('p2pkh', pubkey)
|
|
assert addr in addresses
|
|
d[addr] = {
|
|
'pubkey': pubkey,
|
|
'redeem_script': None,
|
|
'type': 'p2pkh'
|
|
}
|
|
self.put('addresses', d)
|
|
self.put('pubkeys', None)
|
|
self.put('wallet_type', 'imported')
|
|
self.put('seed_version', 14)
|
|
|
|
def _convert_version_15(self):
|
|
if not self._is_upgrade_method_needed(14, 14):
|
|
return
|
|
if self.get('seed_type') == 'segwit':
|
|
# should not get here; get_seed_version should have caught this
|
|
raise Exception('unsupported derivation (development segwit, v14)')
|
|
self.put('seed_version', 15)
|
|
|
|
def _convert_version_16(self):
|
|
# fixes issue #3193 for Imported_Wallets with addresses
|
|
# also, previous versions allowed importing any garbage as an address
|
|
# which we now try to remove, see pr #3191
|
|
if not self._is_upgrade_method_needed(15, 15):
|
|
return
|
|
|
|
def remove_address(addr):
|
|
def remove_from_dict(dict_name):
|
|
d = self.get(dict_name, None)
|
|
if d is not None:
|
|
d.pop(addr, None)
|
|
self.put(dict_name, d)
|
|
|
|
def remove_from_list(list_name):
|
|
lst = self.get(list_name, None)
|
|
if lst is not None:
|
|
s = set(lst)
|
|
s -= {addr}
|
|
self.put(list_name, list(s))
|
|
|
|
# note: we don't remove 'addr' from self.get('addresses')
|
|
remove_from_dict('addr_history')
|
|
remove_from_dict('labels')
|
|
remove_from_dict('payment_requests')
|
|
remove_from_list('frozen_addresses')
|
|
|
|
if self.get('wallet_type') == 'imported':
|
|
addresses = self.get('addresses')
|
|
assert isinstance(addresses, dict)
|
|
addresses_new = dict()
|
|
for address, details in addresses.items():
|
|
if not bitcoin.is_address(address):
|
|
remove_address(address)
|
|
continue
|
|
if details is None:
|
|
addresses_new[address] = {}
|
|
else:
|
|
addresses_new[address] = details
|
|
self.put('addresses', addresses_new)
|
|
|
|
self.put('seed_version', 16)
|
|
|
|
def _convert_version_17(self):
|
|
# delete pruned_txo; construct spent_outpoints
|
|
if not self._is_upgrade_method_needed(16, 16):
|
|
return
|
|
|
|
self.put('pruned_txo', None)
|
|
|
|
transactions = self.get('transactions', {}) # txid -> raw_tx
|
|
spent_outpoints = defaultdict(dict)
|
|
for txid, raw_tx in transactions.items():
|
|
tx = Transaction(raw_tx)
|
|
for txin in tx.inputs():
|
|
if txin.is_coinbase_input():
|
|
continue
|
|
prevout_hash = txin.prevout.txid.hex()
|
|
prevout_n = txin.prevout.out_idx
|
|
spent_outpoints[prevout_hash][str(prevout_n)] = txid
|
|
self.put('spent_outpoints', spent_outpoints)
|
|
|
|
self.put('seed_version', 17)
|
|
|
|
def _convert_version_18(self):
|
|
# delete verified_tx3 as its structure changed
|
|
if not self._is_upgrade_method_needed(17, 17):
|
|
return
|
|
self.put('verified_tx3', None)
|
|
self.put('seed_version', 18)
|
|
|
|
def _convert_version_19(self):
|
|
# delete tx_fees as its structure changed
|
|
if not self._is_upgrade_method_needed(18, 18):
|
|
return
|
|
self.put('tx_fees', None)
|
|
self.put('seed_version', 19)
|
|
|
|
def _convert_version_20(self):
|
|
# store 'derivation' (prefix) and 'root_fingerprint' in all xpub-based keystores.
|
|
# store explicit None values if we cannot retroactively determine them
|
|
if not self._is_upgrade_method_needed(19, 19):
|
|
return
|
|
|
|
from .bip32 import BIP32Node, convert_bip32_intpath_to_strpath
|
|
# note: This upgrade method reimplements bip32.root_fp_and_der_prefix_from_xkey.
|
|
# This is done deliberately, to avoid introducing that method as a dependency to this upgrade.
|
|
for ks_name in ('keystore', *['x{}/'.format(i) for i in range(1, 16)]):
|
|
ks = self.get(ks_name, None)
|
|
if ks is None: continue
|
|
xpub = ks.get('xpub', None)
|
|
if xpub is None: continue
|
|
bip32node = BIP32Node.from_xkey(xpub)
|
|
# derivation prefix
|
|
derivation_prefix = ks.get('derivation', None)
|
|
if derivation_prefix is None:
|
|
assert bip32node.depth >= 0, bip32node.depth
|
|
if bip32node.depth == 0:
|
|
derivation_prefix = 'm'
|
|
elif bip32node.depth == 1:
|
|
child_number_int = int.from_bytes(bip32node.child_number, 'big')
|
|
derivation_prefix = convert_bip32_intpath_to_strpath([child_number_int])
|
|
ks['derivation'] = derivation_prefix
|
|
# root fingerprint
|
|
root_fingerprint = ks.get('ckcc_xfp', None)
|
|
if root_fingerprint is not None:
|
|
root_fingerprint = root_fingerprint.to_bytes(4, byteorder="little", signed=False).hex().lower()
|
|
if root_fingerprint is None:
|
|
if bip32node.depth == 0:
|
|
root_fingerprint = bip32node.calc_fingerprint_of_this_node().hex().lower()
|
|
elif bip32node.depth == 1:
|
|
root_fingerprint = bip32node.fingerprint.hex()
|
|
ks['root_fingerprint'] = root_fingerprint
|
|
ks.pop('ckcc_xfp', None)
|
|
self.put(ks_name, ks)
|
|
|
|
self.put('seed_version', 20)
|
|
|
|
def _convert_version_21(self):
|
|
if not self._is_upgrade_method_needed(20, 20):
|
|
return
|
|
channels = self.get('channels')
|
|
if channels:
|
|
for channel in channels:
|
|
channel['state'] = 'OPENING'
|
|
self.put('channels', channels)
|
|
self.put('seed_version', 21)
|
|
|
|
def _convert_version_22(self):
|
|
# construct prevouts_by_scripthash
|
|
if not self._is_upgrade_method_needed(21, 21):
|
|
return
|
|
|
|
from .bitcoin import script_to_scripthash
|
|
transactions = self.get('transactions', {}) # txid -> raw_tx
|
|
prevouts_by_scripthash = defaultdict(list)
|
|
for txid, raw_tx in transactions.items():
|
|
tx = Transaction(raw_tx)
|
|
for idx, txout in enumerate(tx.outputs()):
|
|
outpoint = f"{txid}:{idx}"
|
|
scripthash = script_to_scripthash(txout.scriptpubkey)
|
|
prevouts_by_scripthash[scripthash].append((outpoint, txout.value))
|
|
self.put('prevouts_by_scripthash', prevouts_by_scripthash)
|
|
|
|
self.put('seed_version', 22)
|
|
|
|
def _convert_version_23(self):
|
|
if not self._is_upgrade_method_needed(22, 22):
|
|
return
|
|
channels = self.get('channels', [])
|
|
LOCAL = 1
|
|
REMOTE = -1
|
|
for c in channels:
|
|
# move revocation store from remote_config
|
|
r = c['remote_config'].pop('revocation_store')
|
|
c['revocation_store'] = r
|
|
# convert fee updates
|
|
log = c.get('log', {})
|
|
for sub in LOCAL, REMOTE:
|
|
l = log[str(sub)]['fee_updates']
|
|
d = {}
|
|
for i, fu in enumerate(l):
|
|
d[str(i)] = {
|
|
'rate':fu['rate'],
|
|
'ctn_local':fu['ctns'][str(LOCAL)],
|
|
'ctn_remote':fu['ctns'][str(REMOTE)]
|
|
}
|
|
log[str(int(sub))]['fee_updates'] = d
|
|
self.data['channels'] = channels
|
|
|
|
self.data['seed_version'] = 23
|
|
|
|
def _convert_version_24(self):
|
|
if not self._is_upgrade_method_needed(23, 23):
|
|
return
|
|
channels = self.get('channels', [])
|
|
for c in channels:
|
|
# convert revocation store to dict
|
|
r = c['revocation_store']
|
|
d = {}
|
|
for i in range(49):
|
|
v = r['buckets'][i]
|
|
if v is not None:
|
|
d[str(i)] = v
|
|
r['buckets'] = d
|
|
c['revocation_store'] = r
|
|
# convert channels to dict
|
|
self.data['channels'] = {x['channel_id']: x for x in channels}
|
|
# convert txi & txo
|
|
txi = self.get('txi', {})
|
|
for tx_hash, d in list(txi.items()):
|
|
d2 = {}
|
|
for addr, l in d.items():
|
|
d2[addr] = {}
|
|
for ser, v in l:
|
|
d2[addr][ser] = v
|
|
txi[tx_hash] = d2
|
|
self.data['txi'] = txi
|
|
txo = self.get('txo', {})
|
|
for tx_hash, d in list(txo.items()):
|
|
d2 = {}
|
|
for addr, l in d.items():
|
|
d2[addr] = {}
|
|
for n, v, cb in l:
|
|
d2[addr][str(n)] = (v, cb)
|
|
txo[tx_hash] = d2
|
|
self.data['txo'] = txo
|
|
|
|
self.data['seed_version'] = 24
|
|
|
|
def _convert_version_25(self):
|
|
from .crypto import sha256
|
|
if not self._is_upgrade_method_needed(24, 24):
|
|
return
|
|
# add 'type' field to onchain requests
|
|
PR_TYPE_ONCHAIN = 0
|
|
requests = self.data.get('payment_requests', {})
|
|
for k, r in list(requests.items()):
|
|
if r.get('address') == k:
|
|
requests[k] = {
|
|
'address': r['address'],
|
|
'amount': r.get('amount'),
|
|
'exp': r.get('exp'),
|
|
'id': r.get('id'),
|
|
'memo': r.get('memo'),
|
|
'time': r.get('time'),
|
|
'type': PR_TYPE_ONCHAIN,
|
|
}
|
|
# delete bip70 invoices
|
|
# note: this upgrade was changed ~2 years after-the-fact to delete instead of converting
|
|
invoices = self.data.get('invoices', {})
|
|
for k, r in list(invoices.items()):
|
|
data = r.get("hex")
|
|
pr_id = sha256(bytes.fromhex(data))[0:16].hex()
|
|
if pr_id != k:
|
|
continue
|
|
del invoices[k]
|
|
self.data['seed_version'] = 25
|
|
|
|
def _convert_version_26(self):
|
|
if not self._is_upgrade_method_needed(25, 25):
|
|
return
|
|
channels = self.data.get('channels', {})
|
|
channel_timestamps = self.data.pop('lightning_channel_timestamps', {})
|
|
for channel_id, c in channels.items():
|
|
item = channel_timestamps.get(channel_id)
|
|
if item:
|
|
funding_txid, funding_height, funding_timestamp, closing_txid, closing_height, closing_timestamp = item
|
|
if funding_txid:
|
|
c['funding_height'] = funding_txid, funding_height, funding_timestamp
|
|
if closing_txid:
|
|
c['closing_height'] = closing_txid, closing_height, closing_timestamp
|
|
self.data['seed_version'] = 26
|
|
|
|
def _convert_version_27(self):
|
|
if not self._is_upgrade_method_needed(26, 26):
|
|
return
|
|
channels = self.data.get('channels', {})
|
|
for channel_id, c in channels.items():
|
|
c['local_config']['htlc_minimum_msat'] = 1
|
|
self.data['seed_version'] = 27
|
|
|
|
def _convert_version_28(self):
|
|
if not self._is_upgrade_method_needed(27, 27):
|
|
return
|
|
channels = self.data.get('channels', {})
|
|
for channel_id, c in channels.items():
|
|
c['local_config']['channel_seed'] = None
|
|
self.data['seed_version'] = 28
|
|
|
|
def _convert_version_29(self):
|
|
if not self._is_upgrade_method_needed(28, 28):
|
|
return
|
|
PR_TYPE_ONCHAIN = 0
|
|
requests = self.data.get('payment_requests', {})
|
|
invoices = self.data.get('invoices', {})
|
|
for d in [invoices, requests]:
|
|
for key, r in list(d.items()):
|
|
_type = r.get('type', 0)
|
|
item = {
|
|
'type': _type,
|
|
'message': r.get('message') or r.get('memo', ''),
|
|
'amount': r.get('amount'),
|
|
'exp': r.get('exp') or 0,
|
|
'time': r.get('time', 0),
|
|
}
|
|
if _type == PR_TYPE_ONCHAIN:
|
|
address = r.pop('address', None)
|
|
if address:
|
|
outputs = [(0, address, r.get('amount'))]
|
|
else:
|
|
outputs = r.get('outputs')
|
|
item.update({
|
|
'outputs': outputs,
|
|
'id': r.get('id'),
|
|
'bip70': r.get('bip70'),
|
|
'requestor': r.get('requestor'),
|
|
})
|
|
else:
|
|
item.update({
|
|
'rhash': r['rhash'],
|
|
'invoice': r['invoice'],
|
|
})
|
|
d[key] = item
|
|
self.data['seed_version'] = 29
|
|
|
|
def _convert_version_30(self):
|
|
if not self._is_upgrade_method_needed(29, 29):
|
|
return
|
|
PR_TYPE_ONCHAIN = 0
|
|
PR_TYPE_LN = 2
|
|
requests = self.data.get('payment_requests', {})
|
|
invoices = self.data.get('invoices', {})
|
|
for d in [invoices, requests]:
|
|
for key, item in list(d.items()):
|
|
_type = item['type']
|
|
if _type == PR_TYPE_ONCHAIN:
|
|
item['amount_sat'] = item.pop('amount')
|
|
elif _type == PR_TYPE_LN:
|
|
amount_sat = item.pop('amount')
|
|
item['amount_msat'] = 1000 * amount_sat if amount_sat is not None else None
|
|
item.pop('exp')
|
|
item.pop('message')
|
|
item.pop('rhash')
|
|
item.pop('time')
|
|
else:
|
|
raise Exception(f"unknown invoice type: {_type}")
|
|
self.data['seed_version'] = 30
|
|
|
|
def _convert_version_31(self):
|
|
if not self._is_upgrade_method_needed(30, 30):
|
|
return
|
|
PR_TYPE_ONCHAIN = 0
|
|
requests = self.data.get('payment_requests', {})
|
|
invoices = self.data.get('invoices', {})
|
|
for d in [invoices, requests]:
|
|
for key, item in list(d.items()):
|
|
if item['type'] == PR_TYPE_ONCHAIN:
|
|
item['amount_sat'] = item['amount_sat'] or 0
|
|
item['exp'] = item['exp'] or 0
|
|
item['time'] = item['time'] or 0
|
|
self.data['seed_version'] = 31
|
|
|
|
def _convert_version_32(self):
|
|
if not self._is_upgrade_method_needed(31, 31):
|
|
return
|
|
PR_TYPE_ONCHAIN = 0
|
|
invoices_old = self.data.get('invoices', {})
|
|
invoices_new = {k: item for k, item in invoices_old.items()
|
|
if not (item['type'] == PR_TYPE_ONCHAIN and item['outputs'] is None)}
|
|
self.data['invoices'] = invoices_new
|
|
self.data['seed_version'] = 32
|
|
|
|
def _convert_version_33(self):
|
|
if not self._is_upgrade_method_needed(32, 32):
|
|
return
|
|
PR_TYPE_ONCHAIN = 0
|
|
requests = self.data.get('payment_requests', {})
|
|
invoices = self.data.get('invoices', {})
|
|
for d in [invoices, requests]:
|
|
for key, item in list(d.items()):
|
|
if item['type'] == PR_TYPE_ONCHAIN:
|
|
item['height'] = item.get('height') or 0
|
|
self.data['seed_version'] = 33
|
|
|
|
def _convert_version_34(self):
|
|
if not self._is_upgrade_method_needed(33, 33):
|
|
return
|
|
channels = self.data.get('channels', {})
|
|
for key, item in channels.items():
|
|
item['local_config']['upfront_shutdown_script'] = \
|
|
item['local_config'].get('upfront_shutdown_script') or ""
|
|
item['remote_config']['upfront_shutdown_script'] = \
|
|
item['remote_config'].get('upfront_shutdown_script') or ""
|
|
self.data['seed_version'] = 34
|
|
|
|
def _convert_version_35(self):
|
|
# same as 32, but for payment_requests
|
|
if not self._is_upgrade_method_needed(34, 34):
|
|
return
|
|
PR_TYPE_ONCHAIN = 0
|
|
requests_old = self.data.get('payment_requests', {})
|
|
requests_new = {k: item for k, item in requests_old.items()
|
|
if not (item['type'] == PR_TYPE_ONCHAIN and item['outputs'] is None)}
|
|
self.data['payment_requests'] = requests_new
|
|
self.data['seed_version'] = 35
|
|
|
|
def _convert_version_36(self):
|
|
if not self._is_upgrade_method_needed(35, 35):
|
|
return
|
|
old_frozen_coins = self.data.get('frozen_coins', [])
|
|
new_frozen_coins = {coin: True for coin in old_frozen_coins}
|
|
self.data['frozen_coins'] = new_frozen_coins
|
|
self.data['seed_version'] = 36
|
|
|
|
def _convert_version_37(self):
|
|
if not self._is_upgrade_method_needed(36, 36):
|
|
return
|
|
payments = self.data.get('lightning_payments', {})
|
|
for k, v in list(payments.items()):
|
|
amount_sat, direction, status = v
|
|
amount_msat = amount_sat * 1000 if amount_sat is not None else None
|
|
payments[k] = amount_msat, direction, status
|
|
self.data['lightning_payments'] = payments
|
|
self.data['seed_version'] = 37
|
|
|
|
def _convert_version_38(self):
|
|
if not self._is_upgrade_method_needed(37, 37):
|
|
return
|
|
PR_TYPE_ONCHAIN = 0
|
|
PR_TYPE_LN = 2
|
|
from .bitcoin import TOTAL_COIN_SUPPLY_LIMIT_IN_BTC, COIN
|
|
max_sats = TOTAL_COIN_SUPPLY_LIMIT_IN_BTC * COIN
|
|
requests = self.data.get('payment_requests', {})
|
|
invoices = self.data.get('invoices', {})
|
|
for d in [invoices, requests]:
|
|
for key, item in list(d.items()):
|
|
if item['type'] == PR_TYPE_ONCHAIN:
|
|
amount_sat = item['amount_sat']
|
|
if amount_sat == '!':
|
|
continue
|
|
if not (isinstance(amount_sat, int) and 0 <= amount_sat <= max_sats):
|
|
del d[key]
|
|
elif item['type'] == PR_TYPE_LN:
|
|
amount_msat = item['amount_msat']
|
|
if not amount_msat:
|
|
continue
|
|
if not (isinstance(amount_msat, int) and 0 <= amount_msat <= max_sats * 1000):
|
|
del d[key]
|
|
self.data['seed_version'] = 38
|
|
|
|
def _convert_version_39(self):
|
|
# this upgrade prevents initialization of lightning_privkey2 after lightning_xprv has been set
|
|
if not self._is_upgrade_method_needed(38, 38):
|
|
return
|
|
self.data['imported_channel_backups'] = self.data.pop('channel_backups', {})
|
|
self.data['seed_version'] = 39
|
|
|
|
def _convert_version_40(self):
|
|
# put 'seed_type' into keystores
|
|
if not self._is_upgrade_method_needed(39, 39):
|
|
return
|
|
for ks_name in ('keystore', *['x{}/'.format(i) for i in range(1, 16)]):
|
|
ks = self.data.get(ks_name, None)
|
|
if ks is None: continue
|
|
seed = ks.get('seed')
|
|
if not seed: continue
|
|
seed_type = None
|
|
xpub = ks.get('xpub') or None
|
|
if xpub:
|
|
assert isinstance(xpub, str)
|
|
if xpub[0:4] in ('xpub', 'tpub'):
|
|
seed_type = 'standard'
|
|
elif xpub[0:4] in ('zpub', 'Zpub', 'vpub', 'Vpub'):
|
|
seed_type = 'segwit'
|
|
elif ks.get('type') == 'old':
|
|
seed_type = 'old'
|
|
if seed_type is not None:
|
|
ks['seed_type'] = seed_type
|
|
self.data['seed_version'] = 40
|
|
|
|
def _convert_version_41(self):
|
|
# this is a repeat of upgrade 39, to fix wallet backup files (see #7339)
|
|
if not self._is_upgrade_method_needed(40, 40):
|
|
return
|
|
imported_channel_backups = self.data.pop('channel_backups', {})
|
|
imported_channel_backups.update(self.data.get('imported_channel_backups', {}))
|
|
self.data['imported_channel_backups'] = imported_channel_backups
|
|
self.data['seed_version'] = 41
|
|
|
|
def _convert_version_42(self):
|
|
# in OnchainInvoice['outputs'], convert values from None to 0
|
|
if not self._is_upgrade_method_needed(41, 41):
|
|
return
|
|
PR_TYPE_ONCHAIN = 0
|
|
requests = self.data.get('payment_requests', {})
|
|
invoices = self.data.get('invoices', {})
|
|
for d in [invoices, requests]:
|
|
for key, item in list(d.items()):
|
|
if item['type'] == PR_TYPE_ONCHAIN:
|
|
item['outputs'] = [(_type, addr, (val or 0))
|
|
for _type, addr, val in item['outputs']]
|
|
self.data['seed_version'] = 42
|
|
|
|
def _convert_version_43(self):
|
|
if not self._is_upgrade_method_needed(42, 42):
|
|
return
|
|
channels = self.data.pop('channels', {})
|
|
for k, c in channels.items():
|
|
log = c['log']
|
|
c['fail_htlc_reasons'] = log.pop('fail_htlc_reasons', {})
|
|
c['unfulfilled_htlcs'] = log.pop('unfulfilled_htlcs', {})
|
|
log["1"]['unacked_updates'] = log.pop('unacked_local_updates2', {})
|
|
self.data['channels'] = channels
|
|
self.data['seed_version'] = 43
|
|
|
|
def _convert_version_44(self):
|
|
if not self._is_upgrade_method_needed(43, 43):
|
|
return
|
|
channels = self.data.get('channels', {})
|
|
for key, item in channels.items():
|
|
if bool(item.get('static_remotekey_enabled')):
|
|
channel_type = ChannelType.OPTION_STATIC_REMOTEKEY
|
|
else:
|
|
channel_type = ChannelType(0)
|
|
item.pop('static_remotekey_enabled', None)
|
|
item['channel_type'] = channel_type
|
|
self.data['seed_version'] = 44
|
|
|
|
def _convert_version_45(self):
|
|
from .lnaddr import lndecode
|
|
if not self._is_upgrade_method_needed(44, 44):
|
|
return
|
|
swaps = self.data.get('submarine_swaps', {})
|
|
for key, item in swaps.items():
|
|
item['receive_address'] = None
|
|
# note: we set height to zero
|
|
# the new key for all requests is a wallet address, not done here
|
|
for name in ['invoices', 'payment_requests']:
|
|
invoices = self.data.get(name, {})
|
|
for key, item in invoices.items():
|
|
is_lightning = item['type'] == 2
|
|
lightning_invoice = item['invoice'] if is_lightning else None
|
|
outputs = item['outputs'] if not is_lightning else None
|
|
bip70 = item['bip70'] if not is_lightning else None
|
|
if is_lightning:
|
|
lnaddr = lndecode(item['invoice'])
|
|
amount_msat = lnaddr.get_amount_msat()
|
|
timestamp = lnaddr.date
|
|
exp_delay = lnaddr.get_expiry()
|
|
message = lnaddr.get_description()
|
|
height = 0
|
|
else:
|
|
amount_sat = item['amount_sat']
|
|
amount_msat = amount_sat * 1000 if amount_sat not in [None, '!'] else amount_sat
|
|
message = item['message']
|
|
timestamp = item['time']
|
|
exp_delay = item['exp']
|
|
height = item['height']
|
|
|
|
invoices[key] = {
|
|
'amount_msat':amount_msat,
|
|
'message':message,
|
|
'time':timestamp,
|
|
'exp':exp_delay,
|
|
'height':height,
|
|
'outputs':outputs,
|
|
'bip70':bip70,
|
|
'lightning_invoice':lightning_invoice,
|
|
}
|
|
self.data['seed_version'] = 45
|
|
|
|
def _convert_invoices_keys(self, invoices):
|
|
# recalc keys of outgoing on-chain invoices
|
|
from .crypto import sha256d
|
|
def get_id_from_onchain_outputs(raw_outputs, timestamp):
|
|
outputs = [PartialTxOutput.from_legacy_tuple(*output) for output in raw_outputs]
|
|
outputs_str = "\n".join(f"{txout.scriptpubkey.hex()}, {txout.value}" for txout in outputs)
|
|
return sha256d(outputs_str + "%d" % timestamp).hex()[0:10]
|
|
for key, item in list(invoices.items()):
|
|
is_lightning = item['lightning_invoice'] is not None
|
|
if is_lightning:
|
|
continue
|
|
outputs_raw = item['outputs']
|
|
assert outputs_raw, outputs_raw
|
|
timestamp = item['time']
|
|
newkey = get_id_from_onchain_outputs(outputs_raw, timestamp)
|
|
if newkey != key:
|
|
invoices[newkey] = item
|
|
del invoices[key]
|
|
|
|
def _convert_version_46(self):
|
|
if not self._is_upgrade_method_needed(45, 45):
|
|
return
|
|
invoices = self.data.get('invoices', {})
|
|
self._convert_invoices_keys(invoices)
|
|
self.data['seed_version'] = 46
|
|
|
|
def _convert_version_47(self):
|
|
from .lnaddr import lndecode
|
|
if not self._is_upgrade_method_needed(46, 46):
|
|
return
|
|
# recalc keys of requests
|
|
requests = self.data.get('payment_requests', {})
|
|
for key, item in list(requests.items()):
|
|
lnaddr = item.get('lightning_invoice')
|
|
if lnaddr:
|
|
lnaddr = lndecode(lnaddr)
|
|
rhash = lnaddr.paymenthash.hex()
|
|
if key != rhash:
|
|
requests[rhash] = item
|
|
del requests[key]
|
|
self.data['seed_version'] = 47
|
|
|
|
def _convert_version_48(self):
|
|
# fix possible corruption of invoice amounts, see #7774
|
|
if not self._is_upgrade_method_needed(47, 47):
|
|
return
|
|
invoices = self.data.get('invoices', {})
|
|
for key, item in list(invoices.items()):
|
|
if item['amount_msat'] == 1000 * "!":
|
|
item['amount_msat'] = "!"
|
|
self.data['seed_version'] = 48
|
|
|
|
def _convert_version_49(self):
|
|
if not self._is_upgrade_method_needed(48, 48):
|
|
return
|
|
channels = self.data.get('channels', {})
|
|
legacy_chans = [chan_dict for chan_dict in channels.values()
|
|
if chan_dict['channel_type'] == ChannelType.OPTION_LEGACY_CHANNEL]
|
|
if legacy_chans:
|
|
raise WalletFileException(
|
|
f"This wallet contains {len(legacy_chans)} lightning channels of type 'LEGACY'. "
|
|
f"These channels were created using unreleased development versions of Electrum "
|
|
f"before the first lightning-capable release of 4.0, and are not supported anymore. "
|
|
f"Please use Electrum 4.3.0 to open this wallet, close the channels, "
|
|
f"and delete them from the wallet."
|
|
)
|
|
self.data['seed_version'] = 49
|
|
|
|
def _convert_version_50(self):
|
|
if not self._is_upgrade_method_needed(49, 49):
|
|
return
|
|
requests = self.data.get('payment_requests', {})
|
|
self._convert_invoices_keys(requests)
|
|
self.data['seed_version'] = 50
|
|
|
|
def _convert_version_51(self):
|
|
from .lnaddr import lndecode
|
|
if not self._is_upgrade_method_needed(50, 50):
|
|
return
|
|
requests = self.data.get('payment_requests', {})
|
|
for key, item in list(requests.items()):
|
|
lightning_invoice = item.pop('lightning_invoice')
|
|
if lightning_invoice is None:
|
|
payment_hash = None
|
|
else:
|
|
lnaddr = lndecode(lightning_invoice)
|
|
payment_hash = lnaddr.paymenthash.hex()
|
|
item['payment_hash'] = payment_hash
|
|
self.data['seed_version'] = 51
|
|
|
|
def _detect_insane_version_51(self) -> int:
|
|
"""Returns 0 if file okay,
|
|
error code 1: multisig wallet has old_mpk
|
|
error code 2: multisig wallet has mixed Ypub/Zpub
|
|
"""
|
|
assert self.get('seed_version') == 51
|
|
xpub_type = None
|
|
for ks_name in ['x{}/'.format(i) for i in range(1, 16)]: # having any such field <=> multisig wallet
|
|
ks = self.data.get(ks_name, None)
|
|
if ks is None: continue
|
|
ks_type = ks.get('type')
|
|
if ks_type == "old":
|
|
return 1 # error
|
|
assert ks_type in ("bip32", "hardware"), f"unexpected {ks_type=}"
|
|
xpub = ks.get('xpub') or None
|
|
assert xpub is not None
|
|
assert isinstance(xpub, str)
|
|
if xpub_type is None: # first iter
|
|
xpub_type = xpub[0:4]
|
|
if xpub[0:4] != xpub_type:
|
|
return 2 # error
|
|
# looks okay
|
|
return 0
|
|
|
|
def _convert_version_52(self):
|
|
if not self._is_upgrade_method_needed(51, 51):
|
|
return
|
|
if (error_code := self._detect_insane_version_51()) != 0:
|
|
# should not get here; get_seed_version should have caught this
|
|
raise Exception(f'unsupported wallet file: version_51 with error {error_code}')
|
|
self.data['seed_version'] = 52
|
|
|
|
def _convert_version_53(self):
|
|
if not self._is_upgrade_method_needed(52, 52):
|
|
return
|
|
cbs = self.data.get('imported_channel_backups', {})
|
|
for channel_id, cb in list(cbs.items()):
|
|
if 'local_payment_pubkey' not in cb:
|
|
cb['local_payment_pubkey'] = None
|
|
self.data['seed_version'] = 53
|
|
|
|
def _convert_version_54(self):
|
|
# note: similar to convert_version_38
|
|
if not self._is_upgrade_method_needed(53, 53):
|
|
return
|
|
from .bitcoin import TOTAL_COIN_SUPPLY_LIMIT_IN_BTC, COIN
|
|
max_sats = TOTAL_COIN_SUPPLY_LIMIT_IN_BTC * COIN
|
|
requests = self.data.get('payment_requests', {})
|
|
invoices = self.data.get('invoices', {})
|
|
for d in [invoices, requests]:
|
|
for key, item in list(d.items()):
|
|
amount_msat = item['amount_msat']
|
|
if amount_msat == '!':
|
|
continue
|
|
if not (isinstance(amount_msat, int) and 0 <= amount_msat <= max_sats * 1000):
|
|
del d[key]
|
|
self.data['seed_version'] = 54
|
|
|
|
def _convert_version_55(self):
|
|
if not self._is_upgrade_method_needed(54, 54):
|
|
return
|
|
# do not use '/' in dict keys
|
|
for key in list(self.data.keys()):
|
|
if key.endswith('/'):
|
|
self.data[key[:-1]] = self.data.pop(key)
|
|
self.data['seed_version'] = 55
|
|
|
|
def _convert_version_56(self):
|
|
if not self._is_upgrade_method_needed(55, 55):
|
|
return
|
|
channels = self.data.get('channels', {})
|
|
for key, item in channels.items():
|
|
item['constraints']['flags'] = 0
|
|
for c in ['local_config', 'remote_config']:
|
|
item[c]['announcement_node_sig'] = ''
|
|
item[c]['announcement_bitcoin_sig'] = ''
|
|
item['local_config'].pop('was_announced')
|
|
self.data['seed_version'] = 56
|
|
|
|
def _convert_version_57(self):
|
|
if not self._is_upgrade_method_needed(56, 56):
|
|
return
|
|
# The 'seed_type' field could be present both at the top-level and inside keystores.
|
|
# We delete the one that is top-level.
|
|
self.data.pop('seed_type', None)
|
|
self.data['seed_version'] = 57
|
|
|
|
def _convert_version_58(self):
|
|
# re-construct prevouts_by_scripthash
|
|
# new structure: scripthash -> outpoint -> value
|
|
if not self._is_upgrade_method_needed(57, 57):
|
|
return
|
|
from .bitcoin import script_to_scripthash
|
|
transactions = self.get('transactions', {}) # txid -> raw_tx
|
|
prevouts_by_scripthash = {}
|
|
for txid, raw_tx in transactions.items():
|
|
try:
|
|
tx = PartialTransaction.from_raw_psbt(raw_tx)
|
|
except BadHeaderMagic:
|
|
tx = Transaction(raw_tx)
|
|
for idx, txout in enumerate(tx.outputs()):
|
|
outpoint = f"{txid}:{idx}"
|
|
scripthash = script_to_scripthash(txout.scriptpubkey)
|
|
if scripthash not in prevouts_by_scripthash:
|
|
prevouts_by_scripthash[scripthash] = {}
|
|
prevouts_by_scripthash[scripthash][outpoint] = txout.value
|
|
self.put('prevouts_by_scripthash', prevouts_by_scripthash)
|
|
self.data['seed_version'] = 58
|
|
|
|
def _convert_version_59(self):
|
|
if not self._is_upgrade_method_needed(58, 58):
|
|
return
|
|
channels = self.data.get('channels', {})
|
|
for _key, chan in channels.items():
|
|
chan.pop('fail_htlc_reasons', {})
|
|
unfulfilled_htlcs = {}
|
|
for htlc_id, (local_ctn, remote_ctn, onion_packet_hex, forwarding_key) in chan['unfulfilled_htlcs'].items():
|
|
unfulfilled_htlcs[htlc_id] = (onion_packet_hex, forwarding_key or None)
|
|
chan['unfulfilled_htlcs'] = unfulfilled_htlcs
|
|
self.data['channels'] = channels
|
|
self.data['seed_version'] = 59
|
|
|
|
def _convert_version_60(self):
|
|
if not self._is_upgrade_method_needed(59, 59):
|
|
return
|
|
cbs = self.data.get('imported_channel_backups', {})
|
|
for channel_id, cb in list(cbs.items()):
|
|
if 'multisig_funding_privkey' not in cb:
|
|
cb['multisig_funding_privkey'] = None
|
|
self.data['seed_version'] = 60
|
|
|
|
def _convert_version_61(self):
|
|
if not self._is_upgrade_method_needed(60, 60):
|
|
return
|
|
# adding additional fields to PaymentInfo
|
|
lightning_payments = self.data.get('lightning_payments', {})
|
|
expiry_never = 100 * 365 * 24 * 60 * 60
|
|
migration_time = int(time.time())
|
|
for rhash, (amount_msat, direction, is_paid) in list(lightning_payments.items()):
|
|
new = (amount_msat, direction, is_paid, 147, expiry_never, migration_time)
|
|
lightning_payments[rhash] = new
|
|
self.data['seed_version'] = 61
|
|
|
|
def _convert_imported(self):
|
|
if not self._is_upgrade_method_needed(0, 13):
|
|
return
|
|
|
|
# '/x' is the internal ID for imported accounts
|
|
d = self.get('accounts', {}).get('/x', {}).get('imported',{})
|
|
if not d:
|
|
return False
|
|
addresses = []
|
|
keypairs = {}
|
|
for addr, v in d.items():
|
|
pubkey, privkey = v
|
|
if privkey:
|
|
keypairs[pubkey] = privkey
|
|
else:
|
|
addresses.append(addr)
|
|
if addresses and keypairs:
|
|
raise WalletFileException('mixed addresses and privkeys')
|
|
elif addresses:
|
|
self.put('addresses', addresses)
|
|
self.put('accounts', None)
|
|
elif keypairs:
|
|
self.put('wallet_type', 'standard')
|
|
self.put('key_type', 'imported')
|
|
self.put('keypairs', keypairs)
|
|
self.put('accounts', None)
|
|
else:
|
|
raise WalletFileException('no addresses or privkeys')
|
|
|
|
def _convert_account(self):
|
|
if not self._is_upgrade_method_needed(0, 13):
|
|
return
|
|
self.put('accounts', None)
|
|
|
|
def _is_upgrade_method_needed(self, min_version, max_version):
|
|
assert min_version <= max_version
|
|
cur_version = self.get_seed_version()
|
|
if cur_version > max_version:
|
|
return False
|
|
elif cur_version < min_version:
|
|
raise WalletFileException(
|
|
'storage upgrade: unexpected version {} (should be {}-{})'
|
|
.format(cur_version, min_version, max_version))
|
|
else:
|
|
return True
|
|
|
|
def get_seed_version(self):
|
|
seed_version = self.get('seed_version')
|
|
if not seed_version:
|
|
seed_version = OLD_SEED_VERSION if len(self.get('master_public_key','')) == 128 else NEW_SEED_VERSION
|
|
if seed_version > FINAL_SEED_VERSION:
|
|
raise WalletFileException('This version of Electrum ({}) is too old to open this wallet.\n'
|
|
'(highest supported storage version: {}, version of this file: {})'
|
|
.format(ELECTRUM_VERSION, FINAL_SEED_VERSION, seed_version))
|
|
if seed_version == 14 and self.get('seed_type') == 'segwit':
|
|
self._raise_unsupported_version(seed_version)
|
|
if seed_version == 51 and self._detect_insane_version_51():
|
|
self._raise_unsupported_version(seed_version)
|
|
if seed_version >= 12:
|
|
return seed_version
|
|
if seed_version not in [OLD_SEED_VERSION, NEW_SEED_VERSION]:
|
|
self._raise_unsupported_version(seed_version)
|
|
return seed_version
|
|
|
|
def _raise_unsupported_version(self, seed_version):
|
|
msg = f"Your wallet has an unsupported seed version: {seed_version}."
|
|
if seed_version in [5, 7, 8, 9, 10, 14]:
|
|
msg += "\n\nTo open this wallet, try 'git checkout seed_v%d'"%seed_version
|
|
if seed_version == 6:
|
|
# version 1.9.8 created v6 wallets when an incorrect seed was entered in the restore dialog
|
|
msg += '\n\nThis file was created because of a bug in version 1.9.8.'
|
|
if self.get('master_public_keys') is None and self.get('master_private_keys') is None and self.get('imported_keys') is None:
|
|
# pbkdf2 (at that time an additional dependency) was not included with the binaries, and wallet creation aborted.
|
|
msg += "\nIt does not contain any keys, and can safely be removed."
|
|
else:
|
|
# creation was complete if electrum was run from source
|
|
msg += "\nPlease open this file with Electrum 1.9.8, and move your coins to a new wallet."
|
|
if seed_version == 51:
|
|
error_code = self._detect_insane_version_51()
|
|
assert error_code != 0
|
|
msg += f" ({error_code=})"
|
|
if error_code == 1:
|
|
msg += "\nThis is a multisig wallet containing an old_mpk (pre-bip32 master public key)."
|
|
msg += "\nPlease contact us to help recover it by opening an issue on GitHub."
|
|
elif error_code == 2:
|
|
msg += ("\nThis is a multisig wallet containing mixed xpub/Ypub/Zpub."
|
|
"\nThe script type is determined by the type of the first keystore."
|
|
"\nTo recover, you should re-create the wallet with matching type "
|
|
"(converted if needed) master keys."
|
|
"\nOr you can contact us to help recover it by opening an issue on GitHub.")
|
|
else:
|
|
raise Exception(f"unexpected {error_code=}")
|
|
raise WalletFileExceptionVersion51(msg, should_report_crash=True)
|
|
# generic exception
|
|
raise WalletFileException(msg)
|
|
|
|
|
|
def upgrade_wallet_db(data: dict, do_upgrade: bool) -> Tuple[dict, bool]:
|
|
was_upgraded = False
|
|
|
|
if len(data) == 0:
|
|
# create new DB
|
|
data['seed_version'] = FINAL_SEED_VERSION
|
|
# store this for debugging purposes
|
|
v = DBMetadata(
|
|
creation_timestamp=int(time.time()),
|
|
first_electrum_version_used=ELECTRUM_VERSION,
|
|
)
|
|
assert data.get("db_metadata", None) is None
|
|
data["db_metadata"] = v
|
|
was_upgraded = True
|
|
|
|
dbu = WalletDBUpgrader(data)
|
|
if dbu.requires_split():
|
|
raise WalletRequiresSplit(dbu.get_split_accounts())
|
|
if dbu.requires_upgrade() and do_upgrade:
|
|
dbu.upgrade()
|
|
was_upgraded = True
|
|
if dbu.requires_upgrade():
|
|
raise WalletRequiresUpgrade()
|
|
return dbu.data, was_upgraded
|
|
|
|
|
|
class WalletDB(JsonDB):
|
|
|
|
def __init__(
|
|
self,
|
|
s: str,
|
|
*,
|
|
storage: Optional['WalletStorage'] = None,
|
|
upgrade: bool = False,
|
|
):
|
|
JsonDB.__init__(
|
|
self,
|
|
s,
|
|
storage=storage,
|
|
encoder=MyEncoder,
|
|
upgrader=partial(upgrade_wallet_db, do_upgrade=upgrade),
|
|
)
|
|
# create pointers
|
|
self.load_transactions()
|
|
# load plugins that are conditional on wallet type
|
|
self.load_plugins()
|
|
|
|
@locked
|
|
def get_seed_version(self):
|
|
return self.get('seed_version')
|
|
|
|
def get_db_metadata(self) -> Optional[DBMetadata]:
|
|
# field only present for wallet files created with ver 4.4.0 or later
|
|
return self.get("db_metadata")
|
|
|
|
@locked
|
|
def get_txi_addresses(self, tx_hash: str) -> List[str]:
|
|
"""Returns list of is_mine addresses that appear as inputs in tx."""
|
|
assert isinstance(tx_hash, str)
|
|
return list(self.txi.get(tx_hash, {}).keys())
|
|
|
|
@locked
|
|
def get_txo_addresses(self, tx_hash: str) -> List[str]:
|
|
"""Returns list of is_mine addresses that appear as outputs in tx."""
|
|
assert isinstance(tx_hash, str)
|
|
return list(self.txo.get(tx_hash, {}).keys())
|
|
|
|
@locked
|
|
def get_txi_addr(self, tx_hash: str, address: str) -> Iterable[Tuple[str, int]]:
|
|
"""Returns an iterable of (prev_outpoint, value)."""
|
|
assert isinstance(tx_hash, str)
|
|
assert isinstance(address, str)
|
|
d = self.txi.get(tx_hash, {}).get(address, {})
|
|
return list(d.items())
|
|
|
|
@locked
|
|
def get_txo_addr(self, tx_hash: str, address: str) -> Dict[int, Tuple[int, bool]]:
|
|
"""Returns a dict: output_index -> (value, is_coinbase)."""
|
|
assert isinstance(tx_hash, str)
|
|
assert isinstance(address, str)
|
|
d = self.txo.get(tx_hash, {}).get(address, {})
|
|
return {int(n): (v, cb) for (n, (v, cb)) in d.items()}
|
|
|
|
@modifier
|
|
def add_txi_addr(self, tx_hash: str, addr: str, ser: str, v: int) -> None:
|
|
assert isinstance(tx_hash, str)
|
|
assert isinstance(addr, str)
|
|
assert isinstance(ser, str)
|
|
assert isinstance(v, int)
|
|
if tx_hash not in self.txi:
|
|
self.txi[tx_hash] = {}
|
|
d = self.txi[tx_hash]
|
|
if addr not in d:
|
|
d[addr] = {}
|
|
d[addr][ser] = v
|
|
|
|
@modifier
|
|
def add_txo_addr(self, tx_hash: str, addr: str, n: Union[int, str], v: int, is_coinbase: bool) -> None:
|
|
n = str(n)
|
|
assert isinstance(tx_hash, str)
|
|
assert isinstance(addr, str)
|
|
assert isinstance(n, str)
|
|
assert isinstance(v, int)
|
|
assert isinstance(is_coinbase, bool)
|
|
if tx_hash not in self.txo:
|
|
self.txo[tx_hash] = {}
|
|
d = self.txo[tx_hash]
|
|
if addr not in d:
|
|
d[addr] = {}
|
|
d[addr][n] = (v, is_coinbase)
|
|
|
|
@locked
|
|
def list_txi(self) -> Sequence[str]:
|
|
return list(self.txi.keys())
|
|
|
|
@locked
|
|
def list_txo(self) -> Sequence[str]:
|
|
return list(self.txo.keys())
|
|
|
|
@modifier
|
|
def remove_txi(self, tx_hash: str) -> None:
|
|
assert isinstance(tx_hash, str)
|
|
self.txi.pop(tx_hash, None)
|
|
|
|
@modifier
|
|
def remove_txo(self, tx_hash: str) -> None:
|
|
assert isinstance(tx_hash, str)
|
|
self.txo.pop(tx_hash, None)
|
|
|
|
@locked
|
|
def list_spent_outpoints(self) -> Sequence[Tuple[str, str]]:
|
|
return [(h, n)
|
|
for h in self.spent_outpoints.keys()
|
|
for n in self.get_spent_outpoints(h)
|
|
]
|
|
|
|
@locked
|
|
def get_spent_outpoints(self, prevout_hash: str) -> Sequence[str]:
|
|
assert isinstance(prevout_hash, str)
|
|
return list(self.spent_outpoints.get(prevout_hash, {}).keys())
|
|
|
|
@locked
|
|
def get_spent_outpoint(self, prevout_hash: str, prevout_n: Union[int, str]) -> Optional[str]:
|
|
assert isinstance(prevout_hash, str)
|
|
prevout_n = str(prevout_n)
|
|
return self.spent_outpoints.get(prevout_hash, {}).get(prevout_n)
|
|
|
|
@modifier
|
|
def remove_spent_outpoint(self, prevout_hash: str, prevout_n: Union[int, str]) -> None:
|
|
assert isinstance(prevout_hash, str)
|
|
prevout_n = str(prevout_n)
|
|
self.spent_outpoints[prevout_hash].pop(prevout_n, None)
|
|
if not self.spent_outpoints[prevout_hash]:
|
|
self.spent_outpoints.pop(prevout_hash)
|
|
|
|
@modifier
|
|
def set_spent_outpoint(self, prevout_hash: str, prevout_n: Union[int, str], tx_hash: str) -> None:
|
|
assert isinstance(prevout_hash, str)
|
|
assert isinstance(tx_hash, str)
|
|
prevout_n = str(prevout_n)
|
|
if prevout_hash not in self.spent_outpoints:
|
|
self.spent_outpoints[prevout_hash] = {}
|
|
self.spent_outpoints[prevout_hash][prevout_n] = tx_hash
|
|
|
|
@modifier
|
|
def add_prevout_by_scripthash(self, scripthash: str, *, prevout: TxOutpoint, value: int) -> None:
|
|
assert isinstance(scripthash, str)
|
|
assert isinstance(prevout, TxOutpoint)
|
|
assert isinstance(value, int)
|
|
if scripthash not in self._prevouts_by_scripthash:
|
|
self._prevouts_by_scripthash[scripthash] = dict()
|
|
self._prevouts_by_scripthash[scripthash][prevout.to_str()] = value
|
|
|
|
@modifier
|
|
def remove_prevout_by_scripthash(self, scripthash: str, *, prevout: TxOutpoint, value: int) -> None:
|
|
assert isinstance(scripthash, str)
|
|
assert isinstance(prevout, TxOutpoint)
|
|
assert isinstance(value, int)
|
|
self._prevouts_by_scripthash[scripthash].pop(prevout.to_str(), None)
|
|
if not self._prevouts_by_scripthash[scripthash]:
|
|
self._prevouts_by_scripthash.pop(scripthash)
|
|
|
|
@locked
|
|
def get_prevouts_by_scripthash(self, scripthash: str) -> Set[Tuple[TxOutpoint, int]]:
|
|
assert isinstance(scripthash, str)
|
|
prevouts_and_values = self._prevouts_by_scripthash.get(scripthash, {})
|
|
return {(TxOutpoint.from_str(prevout), value) for prevout, value in prevouts_and_values.items()}
|
|
|
|
@modifier
|
|
def add_transaction(self, tx_hash: str, tx: Transaction) -> None:
|
|
assert isinstance(tx_hash, str)
|
|
assert isinstance(tx, Transaction), tx
|
|
# note that tx might be a PartialTransaction
|
|
# serialize and de-serialize tx now. this might e.g. convert a complete PartialTx to a Tx
|
|
tx = tx_from_any(str(tx))
|
|
if not tx_hash:
|
|
raise Exception("trying to add tx to db without txid")
|
|
if tx_hash != tx.txid():
|
|
raise Exception(f"trying to add tx to db with inconsistent txid: {tx_hash} != {tx.txid()}")
|
|
# don't allow overwriting complete tx with partial tx
|
|
tx_we_already_have = self.transactions.get(tx_hash, None)
|
|
if tx_we_already_have is None or isinstance(tx_we_already_have, PartialTransaction):
|
|
self.transactions[tx_hash] = tx
|
|
|
|
@modifier
|
|
def remove_transaction(self, tx_hash: str) -> Optional[Transaction]:
|
|
assert isinstance(tx_hash, str)
|
|
return self.transactions.pop(tx_hash, None)
|
|
|
|
@locked
|
|
def get_transaction(self, tx_hash: Optional[str]) -> Optional[Transaction]:
|
|
if tx_hash is None:
|
|
return None
|
|
assert isinstance(tx_hash, str)
|
|
return self.transactions.get(tx_hash)
|
|
|
|
@locked
|
|
def list_transactions(self) -> Sequence[str]:
|
|
return list(self.transactions.keys())
|
|
|
|
@locked
|
|
def get_history(self) -> Sequence[str]:
|
|
return list(self.history.keys())
|
|
|
|
def is_addr_in_history(self, addr: str) -> bool:
|
|
# does not mean history is non-empty!
|
|
assert isinstance(addr, str)
|
|
return addr in self.history
|
|
|
|
@locked
|
|
def get_addr_history(self, addr: str) -> Sequence[Tuple[str, int]]:
|
|
assert isinstance(addr, str)
|
|
return self.history.get(addr, [])
|
|
|
|
@modifier
|
|
def set_addr_history(self, addr: str, hist) -> None:
|
|
assert isinstance(addr, str)
|
|
self.history[addr] = hist
|
|
|
|
@modifier
|
|
def remove_addr_history(self, addr: str) -> None:
|
|
assert isinstance(addr, str)
|
|
self.history.pop(addr, None)
|
|
|
|
@locked
|
|
def list_verified_tx(self) -> Sequence[str]:
|
|
return list(self.verified_tx.keys())
|
|
|
|
@locked
|
|
def get_verified_tx(self, txid: str) -> Optional[TxMinedInfo]:
|
|
assert isinstance(txid, str)
|
|
if txid not in self.verified_tx:
|
|
return None
|
|
height, timestamp, txpos, header_hash = self.verified_tx[txid]
|
|
return TxMinedInfo(_height=height,
|
|
conf=None,
|
|
timestamp=timestamp,
|
|
txpos=txpos,
|
|
header_hash=header_hash)
|
|
|
|
@modifier
|
|
def add_verified_tx(self, txid: str, info: TxMinedInfo):
|
|
assert isinstance(txid, str)
|
|
assert isinstance(info, TxMinedInfo)
|
|
height = info._height # number of conf is dynamic and might not be set here
|
|
assert height > 0, height
|
|
self.verified_tx[txid] = (height, info.timestamp, info.txpos, info.header_hash)
|
|
|
|
@modifier
|
|
def remove_verified_tx(self, txid: str):
|
|
assert isinstance(txid, str)
|
|
self.verified_tx.pop(txid, None)
|
|
|
|
def is_in_verified_tx(self, txid: str) -> bool:
|
|
assert isinstance(txid, str)
|
|
return txid in self.verified_tx
|
|
|
|
@modifier
|
|
def add_tx_fee_from_server(self, txid: str, fee_sat: Optional[int]) -> None:
|
|
assert isinstance(txid, str)
|
|
# note: when called with (fee_sat is None), rm currently saved value
|
|
if txid not in self.tx_fees:
|
|
self.tx_fees[txid] = TxFeesValue()
|
|
tx_fees_value = self.tx_fees[txid]
|
|
if tx_fees_value.is_calculated_by_us:
|
|
return
|
|
self.tx_fees[txid] = tx_fees_value._replace(fee=fee_sat, is_calculated_by_us=False)
|
|
|
|
@modifier
|
|
def add_tx_fee_we_calculated(self, txid: str, fee_sat: Optional[int]) -> None:
|
|
assert isinstance(txid, str)
|
|
if fee_sat is None:
|
|
return
|
|
assert isinstance(fee_sat, int)
|
|
if txid not in self.tx_fees:
|
|
self.tx_fees[txid] = TxFeesValue()
|
|
self.tx_fees[txid] = self.tx_fees[txid]._replace(fee=fee_sat, is_calculated_by_us=True)
|
|
|
|
@locked
|
|
def get_tx_fee(self, txid: str, *, trust_server: bool = False) -> Optional[int]:
|
|
assert isinstance(txid, str)
|
|
"""Returns tx_fee."""
|
|
tx_fees_value = self.tx_fees.get(txid)
|
|
if tx_fees_value is None:
|
|
return None
|
|
if not trust_server and not tx_fees_value.is_calculated_by_us:
|
|
return None
|
|
return tx_fees_value.fee
|
|
|
|
@modifier
|
|
def add_num_inputs_to_tx(self, txid: str, num_inputs: int) -> None:
|
|
assert isinstance(txid, str)
|
|
assert isinstance(num_inputs, int)
|
|
if txid not in self.tx_fees:
|
|
self.tx_fees[txid] = TxFeesValue()
|
|
self.tx_fees[txid] = self.tx_fees[txid]._replace(num_inputs=num_inputs)
|
|
|
|
@locked
|
|
def get_num_all_inputs_of_tx(self, txid: str) -> Optional[int]:
|
|
assert isinstance(txid, str)
|
|
tx_fees_value = self.tx_fees.get(txid)
|
|
if tx_fees_value is None:
|
|
return None
|
|
return tx_fees_value.num_inputs
|
|
|
|
@locked
|
|
def get_num_ismine_inputs_of_tx(self, txid: str) -> int:
|
|
assert isinstance(txid, str)
|
|
txins = self.txi.get(txid, {})
|
|
return sum([len(tupls) for addr, tupls in txins.items()])
|
|
|
|
@modifier
|
|
def remove_tx_fee(self, txid: str) -> None:
|
|
assert isinstance(txid, str)
|
|
self.tx_fees.pop(txid, None)
|
|
|
|
@locked
|
|
def num_change_addresses(self) -> int:
|
|
return len(self.change_addresses)
|
|
|
|
@locked
|
|
def num_receiving_addresses(self) -> int:
|
|
return len(self.receiving_addresses)
|
|
|
|
@locked
|
|
def get_change_addresses(self, *, slice_start=None, slice_stop=None) -> List[str]:
|
|
# note: slicing makes a shallow copy
|
|
return self.change_addresses[slice_start:slice_stop]
|
|
|
|
@locked
|
|
def get_receiving_addresses(self, *, slice_start=None, slice_stop=None) -> List[str]:
|
|
# note: slicing makes a shallow copy
|
|
return self.receiving_addresses[slice_start:slice_stop]
|
|
|
|
@modifier
|
|
def add_change_address(self, addr: str) -> None:
|
|
assert isinstance(addr, str)
|
|
self._addr_to_addr_index[addr] = (1, len(self.change_addresses))
|
|
self.change_addresses.append(addr)
|
|
|
|
@modifier
|
|
def add_receiving_address(self, addr: str) -> None:
|
|
assert isinstance(addr, str)
|
|
self._addr_to_addr_index[addr] = (0, len(self.receiving_addresses))
|
|
self.receiving_addresses.append(addr)
|
|
|
|
@locked
|
|
def get_address_index(self, address: str) -> Optional[Sequence[int]]:
|
|
assert isinstance(address, str)
|
|
return self._addr_to_addr_index.get(address)
|
|
|
|
@modifier
|
|
def add_imported_address(self, addr: str, d: dict) -> None:
|
|
assert isinstance(addr, str)
|
|
self.imported_addresses[addr] = d
|
|
|
|
@modifier
|
|
def remove_imported_address(self, addr: str) -> None:
|
|
assert isinstance(addr, str)
|
|
self.imported_addresses.pop(addr)
|
|
|
|
@locked
|
|
def has_imported_address(self, addr: str) -> bool:
|
|
assert isinstance(addr, str)
|
|
return addr in self.imported_addresses
|
|
|
|
@locked
|
|
def get_imported_addresses(self) -> Sequence[str]:
|
|
return list(sorted(self.imported_addresses.keys()))
|
|
|
|
@locked
|
|
def get_imported_address(self, addr: str) -> Optional[dict]:
|
|
assert isinstance(addr, str)
|
|
return self.imported_addresses.get(addr)
|
|
|
|
def load_addresses(self, wallet_type):
|
|
""" called from Abstract_Wallet.__init__ """
|
|
if wallet_type == 'imported':
|
|
self.imported_addresses = self.get_dict('addresses') # type: Dict[str, dict]
|
|
else:
|
|
self.get_dict('addresses')
|
|
for name in ['receiving', 'change']:
|
|
if name not in self.data['addresses']:
|
|
self.data['addresses'][name] = []
|
|
self.change_addresses = self.data['addresses']['change']
|
|
self.receiving_addresses = self.data['addresses']['receiving']
|
|
self._addr_to_addr_index = {} # type: Dict[str, Sequence[int]] # key: address, value: (is_change, index)
|
|
for i, addr in enumerate(self.receiving_addresses):
|
|
self._addr_to_addr_index[addr] = (0, i)
|
|
for i, addr in enumerate(self.change_addresses):
|
|
self._addr_to_addr_index[addr] = (1, i)
|
|
|
|
@profiler
|
|
def load_transactions(self):
|
|
# references in self.data
|
|
# TODO make all these private
|
|
# txid -> address -> prev_outpoint -> value
|
|
self.txi = self.get_dict('txi') # type: Dict[str, Dict[str, Dict[str, int]]]
|
|
# txid -> address -> output_index -> (value, is_coinbase)
|
|
self.txo = self.get_dict('txo') # type: Dict[str, Dict[str, Dict[str, Tuple[int, bool]]]]
|
|
self.transactions = self.get_dict('transactions') # type: Dict[str, Transaction]
|
|
self.spent_outpoints = self.get_dict('spent_outpoints') # txid -> output_index -> next_txid
|
|
self.history = self.get_dict('addr_history') # address -> list of (txid, height)
|
|
self.verified_tx = self.get_dict('verified_tx3') # txid -> (height, timestamp, txpos, header_hash)
|
|
self.tx_fees = self.get_dict('tx_fees') # type: Dict[str, TxFeesValue]
|
|
# scripthash -> outpoint -> value
|
|
self._prevouts_by_scripthash = self.get_dict('prevouts_by_scripthash') # type: Dict[str, Dict[str, int]]
|
|
# remove unreferenced tx
|
|
for tx_hash in list(self.transactions.keys()):
|
|
if not self.get_txi_addresses(tx_hash) and not self.get_txo_addresses(tx_hash):
|
|
self.logger.info(f"removing unreferenced tx: {tx_hash}")
|
|
self.transactions.pop(tx_hash)
|
|
# remove unreferenced outpoints
|
|
for prevout_hash in self.spent_outpoints.keys():
|
|
d = self.spent_outpoints[prevout_hash]
|
|
for prevout_n, spending_txid in list(d.items()):
|
|
if spending_txid not in self.transactions:
|
|
self.logger.info("removing unreferenced spent outpoint")
|
|
d.pop(prevout_n)
|
|
|
|
@modifier
|
|
def clear_history(self):
|
|
self.txi.clear()
|
|
self.txo.clear()
|
|
self.spent_outpoints.clear()
|
|
self.transactions.clear()
|
|
self.history.clear()
|
|
self.verified_tx.clear()
|
|
self.tx_fees.clear()
|
|
self._prevouts_by_scripthash.clear()
|
|
|
|
def _should_convert_to_stored_dict(self, key) -> bool:
|
|
if key == 'keystore':
|
|
return False
|
|
multisig_keystore_names = [('x%d' % i) for i in range(1, 16)]
|
|
if key in multisig_keystore_names:
|
|
return False
|
|
return True
|
|
|
|
@classmethod
|
|
def split_accounts(klass, root_path, split_data):
|
|
from .storage import WalletStorage
|
|
file_list = []
|
|
for data in split_data:
|
|
path = root_path + '.' + data['suffix']
|
|
item_storage = WalletStorage(path)
|
|
db = WalletDB(json.dumps(data), storage=item_storage, upgrade=True)
|
|
db.write()
|
|
file_list.append(path)
|
|
return file_list
|
|
|
|
def get_action(self):
|
|
action = run_hook('get_action', self)
|
|
return action
|
|
|
|
def load_plugins(self):
|
|
wallet_type = self.get('wallet_type')
|
|
if wallet_type in plugin_loaders:
|
|
plugin_loaders[wallet_type]()
|
|
|
|
def get_plugin_storage(self) -> dict:
|
|
return self.get_dict('plugin_data')
|
|
|
|
def prune_uninstalled_plugin_data(self, installed_plugins: AbstractSet[str]) -> None:
|
|
"""Remove plugin data for plugins that are not installed anymore."""
|
|
plugin_storage = self.get_plugin_storage()
|
|
for name in list(plugin_storage.keys()):
|
|
if name not in installed_plugins:
|
|
plugin_storage.pop(name)
|
|
self.logger.info(f"deleting plugin data: {name=}")
|
|
|
|
def set_keystore_encryption(self, enable):
|
|
self.put('use_encryption', enable)
|