220 lines
5.9 KiB
Go
220 lines
5.9 KiB
Go
package bchain
|
|
|
|
import (
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/golang/glog"
|
|
)
|
|
|
|
// addrIndex and outpoint are used also in non utxo mempool
|
|
type addrIndex struct {
|
|
addrID string
|
|
n int32
|
|
}
|
|
|
|
type outpoint struct {
|
|
txid string
|
|
vout int32
|
|
}
|
|
|
|
type txidio struct {
|
|
txid string
|
|
io []addrIndex
|
|
}
|
|
|
|
// UTXOMempool is mempool handle.
|
|
type UTXOMempool struct {
|
|
chain BlockChain
|
|
mux sync.Mutex
|
|
txToInputOutput map[string][]addrIndex
|
|
addrIDToTx map[string][]outpoint
|
|
chanTxid chan string
|
|
chanAddrIndex chan txidio
|
|
onNewTxAddr func(txid string, addr string)
|
|
}
|
|
|
|
// NewUTXOMempool creates new mempool handler.
|
|
// For now there is no cleanup of sync routines, the expectation is that the mempool is created only once per process
|
|
func NewUTXOMempool(chain BlockChain, workers int, subworkers int) *UTXOMempool {
|
|
m := &UTXOMempool{
|
|
chain: chain,
|
|
chanTxid: make(chan string, 1),
|
|
chanAddrIndex: make(chan txidio, 1),
|
|
}
|
|
for i := 0; i < workers; i++ {
|
|
go func(i int) {
|
|
chanInput := make(chan outpoint, 1)
|
|
chanResult := make(chan *addrIndex, 1)
|
|
for j := 0; j < subworkers; j++ {
|
|
go func(j int) {
|
|
for input := range chanInput {
|
|
ai := m.getInputAddress(input)
|
|
chanResult <- ai
|
|
}
|
|
}(j)
|
|
}
|
|
for txid := range m.chanTxid {
|
|
io, ok := m.getTxAddrs(txid, chanInput, chanResult)
|
|
if !ok {
|
|
io = []addrIndex{}
|
|
}
|
|
m.chanAddrIndex <- txidio{txid, io}
|
|
}
|
|
}(i)
|
|
}
|
|
glog.Info("mempool: starting with ", workers, "*", subworkers, " sync workers")
|
|
return m
|
|
}
|
|
|
|
// GetTransactions returns slice of mempool transactions for given address
|
|
func (m *UTXOMempool) GetTransactions(address string) ([]string, error) {
|
|
parser := m.chain.GetChainParser()
|
|
addrID, err := parser.GetAddrIDFromAddress(address)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
m.mux.Lock()
|
|
defer m.mux.Unlock()
|
|
outpoints := m.addrIDToTx[string(addrID)]
|
|
txs := make([]string, 0, len(outpoints))
|
|
for _, o := range outpoints {
|
|
txs = append(txs, o.txid)
|
|
}
|
|
return txs, nil
|
|
}
|
|
|
|
func (m *UTXOMempool) updateMappings(newTxToInputOutput map[string][]addrIndex, newAddrIDToTx map[string][]outpoint) {
|
|
m.mux.Lock()
|
|
defer m.mux.Unlock()
|
|
m.txToInputOutput = newTxToInputOutput
|
|
m.addrIDToTx = newAddrIDToTx
|
|
}
|
|
|
|
func (m *UTXOMempool) getInputAddress(input outpoint) *addrIndex {
|
|
// TODO - possibly get from DB unspenttxs - however some output txs can be also in mempool
|
|
itx, err := m.chain.GetTransactionForMempool(input.txid)
|
|
if err != nil {
|
|
glog.Error("cannot get transaction ", input.txid, ": ", err)
|
|
return nil
|
|
}
|
|
if int(input.vout) >= len(itx.Vout) {
|
|
glog.Error("Vout len in transaction ", input.txid, " ", len(itx.Vout), " input.Vout=", input.vout)
|
|
return nil
|
|
}
|
|
addrID, err := m.chain.GetChainParser().GetAddrIDFromVout(&itx.Vout[input.vout])
|
|
if err != nil {
|
|
glog.Error("error in addrID in ", input.txid, " ", input.vout, ": ", err)
|
|
return nil
|
|
}
|
|
return &addrIndex{string(addrID), ^input.vout}
|
|
|
|
}
|
|
|
|
func (m *UTXOMempool) getTxAddrs(txid string, chanInput chan outpoint, chanResult chan *addrIndex) ([]addrIndex, bool) {
|
|
tx, err := m.chain.GetTransactionForMempool(txid)
|
|
if err != nil {
|
|
glog.Error("cannot get transaction ", txid, ": ", err)
|
|
return nil, false
|
|
}
|
|
glog.V(2).Info("mempool: gettxaddrs ", txid, ", ", len(tx.Vin), " inputs")
|
|
io := make([]addrIndex, 0, len(tx.Vout)+len(tx.Vin))
|
|
for _, output := range tx.Vout {
|
|
addrID, err := m.chain.GetChainParser().GetAddrIDFromVout(&output)
|
|
if err != nil {
|
|
glog.Error("error in addrID in ", txid, " ", output.N, ": ", err)
|
|
continue
|
|
}
|
|
if len(addrID) > 0 {
|
|
io = append(io, addrIndex{string(addrID), int32(output.N)})
|
|
}
|
|
if m.onNewTxAddr != nil && len(output.ScriptPubKey.Addresses) == 1 {
|
|
m.onNewTxAddr(tx.Txid, output.ScriptPubKey.Addresses[0])
|
|
}
|
|
}
|
|
dispatched := 0
|
|
for _, input := range tx.Vin {
|
|
if input.Coinbase != "" {
|
|
continue
|
|
}
|
|
o := outpoint{input.Txid, int32(input.Vout)}
|
|
loop:
|
|
for {
|
|
select {
|
|
// store as many processed results as possible
|
|
case ai := <-chanResult:
|
|
if ai != nil {
|
|
io = append(io, *ai)
|
|
}
|
|
dispatched--
|
|
// send input to be processed
|
|
case chanInput <- o:
|
|
dispatched++
|
|
break loop
|
|
}
|
|
}
|
|
}
|
|
for i := 0; i < dispatched; i++ {
|
|
ai := <-chanResult
|
|
if ai != nil {
|
|
io = append(io, *ai)
|
|
}
|
|
}
|
|
return io, true
|
|
}
|
|
|
|
// Resync gets mempool transactions and maps outputs to transactions.
|
|
// Resync is not reentrant, it should be called from a single thread.
|
|
// Read operations (GetTransactions) are safe.
|
|
func (m *UTXOMempool) Resync(onNewTxAddr func(txid string, addr string)) error {
|
|
start := time.Now()
|
|
glog.V(1).Info("mempool: resync")
|
|
m.onNewTxAddr = onNewTxAddr
|
|
txs, err := m.chain.GetMempool()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
glog.V(2).Info("mempool: resync ", len(txs), " txs")
|
|
// allocate slightly larger capacity of the maps
|
|
newTxToInputOutput := make(map[string][]addrIndex, len(m.txToInputOutput)+5)
|
|
newAddrIDToTx := make(map[string][]outpoint, len(m.addrIDToTx)+5)
|
|
dispatched := 0
|
|
onNewData := func(txid string, io []addrIndex) {
|
|
if len(io) > 0 {
|
|
newTxToInputOutput[txid] = io
|
|
for _, si := range io {
|
|
newAddrIDToTx[si.addrID] = append(newAddrIDToTx[si.addrID], outpoint{txid, si.n})
|
|
}
|
|
}
|
|
}
|
|
// get transaction in parallel using goroutines created in NewUTXOMempool
|
|
for _, txid := range txs {
|
|
io, exists := m.txToInputOutput[txid]
|
|
if !exists {
|
|
loop:
|
|
for {
|
|
select {
|
|
// store as many processed transactions as possible
|
|
case tio := <-m.chanAddrIndex:
|
|
onNewData(tio.txid, tio.io)
|
|
dispatched--
|
|
// send transaction to be processed
|
|
case m.chanTxid <- txid:
|
|
dispatched++
|
|
break loop
|
|
}
|
|
}
|
|
} else {
|
|
onNewData(txid, io)
|
|
}
|
|
}
|
|
for i := 0; i < dispatched; i++ {
|
|
tio := <-m.chanAddrIndex
|
|
onNewData(tio.txid, tio.io)
|
|
}
|
|
m.updateMappings(newTxToInputOutput, newAddrIDToTx)
|
|
m.onNewTxAddr = nil
|
|
glog.Info("mempool: resync finished in ", time.Since(start), ", ", len(m.txToInputOutput), " transactions in mempool")
|
|
return nil
|
|
}
|