eth, les, swarm: fix go vet issues sufraced by log15

This commit is contained in:
Péter Szilágyi 2017-02-22 14:56:09 +02:00
parent d4fd06c3dc
commit 61e6bb1247
No known key found for this signature in database
GPG Key ID: E9AE538CEDF8293D
8 changed files with 10 additions and 10 deletions

View File

@ -938,7 +938,7 @@ func (d *Downloader) fetchNodeData() error {
} }
if err != nil { if err != nil {
// If the node data processing failed, the root hash is very wrong, abort // If the node data processing failed, the root hash is very wrong, abort
log.Error(fmt.Sprintf("peer %d: state processing failed: %v", packet.PeerId(), err)) log.Error(fmt.Sprintf("peer %s: state processing failed: %v", packet.PeerId(), err))
d.cancel() d.cancel()
return return
} }

View File

@ -441,7 +441,7 @@ func (pm *ProtocolManager) handleMsg(p *peer) error {
// Block header query, collect the requested headers and reply // Block header query, collect the requested headers and reply
case AnnounceMsg: case AnnounceMsg:
log.Debug(fmt.Sprint("<=== AnnounceMsg from peer %v:", p.id)) log.Debug(fmt.Sprintf("<=== AnnounceMsg from peer %v:", p.id))
var req announceData var req announceData
if err := msg.Decode(&req); err != nil { if err := msg.Decode(&req); err != nil {

View File

@ -84,7 +84,7 @@ func (self *Api) Resolve(hostPort string, nameresolver bool) (storage.Key, error
err = ErrResolve(err) err = ErrResolve(err)
log.Warn(fmt.Sprintf("DNS error : %v", err)) log.Warn(fmt.Sprintf("DNS error : %v", err))
} }
log.Trace(fmt.Sprintf("host lookup: %v -> %v", err)) log.Trace(fmt.Sprintf("host lookup: %v -> %v", hostPort, contentHash))
return contentHash[:], err return contentHash[:], err
} }
func Parse(uri string) (hostPort, path string) { func Parse(uri string) (hostPort, path string) {

View File

@ -211,7 +211,7 @@ only add if less than requesterCount peers forwarded the same request id so far
note this is done irrespective of status (searching or found) note this is done irrespective of status (searching or found)
*/ */
func (self *Depo) addRequester(rs *storage.RequestStatus, req *retrieveRequestMsgData) { func (self *Depo) addRequester(rs *storage.RequestStatus, req *retrieveRequestMsgData) {
log.Trace(fmt.Sprintf("Depo.addRequester: key %v - add peer to req.Id %v", req.Key.Log(), req.from, req.Id)) log.Trace(fmt.Sprintf("Depo.addRequester: key %v - add peer to req.Id %v", req.Key.Log(), req.Id))
list := rs.Requesters[req.Id] list := rs.Requesters[req.Id]
rs.Requesters[req.Id] = append(list, req) rs.Requesters[req.Id] = append(list, req)
} }

View File

@ -120,7 +120,7 @@ func (self *KadDb) add(nrs []*NodeRecord, proximityBin func(Address) int) {
copy(newnodes[:], nodes[:dbcursor]) copy(newnodes[:], nodes[:dbcursor])
newnodes[dbcursor] = node newnodes[dbcursor] = node
copy(newnodes[dbcursor+1:], nodes[dbcursor:]) copy(newnodes[dbcursor+1:], nodes[dbcursor:])
log.Trace(fmt.Sprintf("new nodes: %v (keys: %v)\nnodes: %v", newnodes, nodes)) log.Trace(fmt.Sprintf("new nodes: %v, nodes: %v", newnodes, nodes))
self.Nodes[index] = newnodes self.Nodes[index] = newnodes
n++ n++
} }
@ -294,7 +294,7 @@ func (self *KadDb) save(path string, cb func(*NodeRecord, Node)) error {
} }
err = ioutil.WriteFile(path, data, os.ModePerm) err = ioutil.WriteFile(path, data, os.ModePerm)
if err != nil { if err != nil {
log.Warn(fmt.Sprintf("unable to save kaddb with %v nodes to %v: err", n, path, err)) log.Warn(fmt.Sprintf("unable to save kaddb with %v nodes to %v: %v", n, path, err))
} else { } else {
log.Info(fmt.Sprintf("saved kaddb with %v nodes to %v", n, path)) log.Info(fmt.Sprintf("saved kaddb with %v nodes to %v", n, path))
} }

View File

@ -217,7 +217,7 @@ LOOP:
// if just switched to db mode and not quitting, then launch dbRead // if just switched to db mode and not quitting, then launch dbRead
// in a parallel go routine to send deliveries from db // in a parallel go routine to send deliveries from db
if inDb == 0 && quit != nil { if inDb == 0 && quit != nil {
log.Trace(fmt.Sprintf("syncDb[%v/%v] start dbRead")) log.Trace(fmt.Sprintf("syncDb[%v/%v] start dbRead", self.key.Log(), self.priority))
go self.dbRead(true, counter, deliver) go self.dbRead(true, counter, deliver)
} }
inDb++ inDb++

View File

@ -480,7 +480,7 @@ LOOP:
stateCopy := *state stateCopy := *state
err := self.unsyncedKeys(unsynced, &stateCopy) err := self.unsyncedKeys(unsynced, &stateCopy)
if err != nil { if err != nil {
log.Warn(fmt.Sprintf("syncer[%v]: unable to send unsynced keys: %v", err)) log.Warn(fmt.Sprintf("syncer[%v]: unable to send unsynced keys: %v", self.key.Log(), err))
} }
self.state = state self.state = state
log.Debug(fmt.Sprintf("syncer[%v]: --> %v keys sent: (total: %v (%v), history: %v), sent sync state: %v", self.key.Log(), len(unsynced), keyCounts, keyCount, historyCnt, stateCopy)) log.Debug(fmt.Sprintf("syncer[%v]: --> %v keys sent: (total: %v (%v), history: %v), sent sync state: %v", self.key.Log(), len(unsynced), keyCounts, keyCount, historyCnt, stateCopy))
@ -553,7 +553,7 @@ LOOP:
log.Trace(fmt.Sprintf("syncer[%v]: (priority %v): request %v (synced = %v)", self.key.Log(), priority, req, state.Synced)) log.Trace(fmt.Sprintf("syncer[%v]: (priority %v): request %v (synced = %v)", self.key.Log(), priority, req, state.Synced))
unsynced = append(unsynced, sreq) unsynced = append(unsynced, sreq)
} else { } else {
log.Warn(fmt.Sprintf("syncer[%v]: (priority %v): error creating request for %v: %v)", self.key.Log(), priority, req, state.Synced, err)) log.Warn(fmt.Sprintf("syncer[%v]: (priority %v): error creating request for %v: %v)", self.key.Log(), priority, req, err))
} }
} }

View File

@ -197,7 +197,7 @@ func (self *Swarm) Start(net *p2p.Server) error {
log.Debug(fmt.Sprintf("Swarm http proxy started on port: %v", self.config.Port)) log.Debug(fmt.Sprintf("Swarm http proxy started on port: %v", self.config.Port))
if self.corsString != "" { if self.corsString != "" {
log.Debug(fmt.Sprintf("Swarm http proxy started with corsdomain:", self.corsString)) log.Debug(fmt.Sprintf("Swarm http proxy started with corsdomain: %v", self.corsString))
} }
return nil return nil