Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
G
Geth-Modification
Project
Project
Details
Activity
Releases
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
张蕾
Geth-Modification
Commits
a4e4c76c
Commit
a4e4c76c
authored
Jun 21, 2017
by
Bas van Kervel
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
whisper: use syncmap for dynamic configuration options
parent
7a11e864
Changes
5
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
445 additions
and
43 deletions
+445
-43
LICENSE
vendor/golang.org/x/sync/LICENSE
+27
-0
PATENTS
vendor/golang.org/x/sync/PATENTS
+22
-0
map.go
vendor/golang.org/x/sync/syncmap/map.go
+372
-0
vendor.json
vendor/vendor.json
+6
-0
whisper.go
whisper/whisperv5/whisper.go
+18
-43
No files found.
vendor/golang.org/x/sync/LICENSE
0 → 100644
View file @
a4e4c76c
Copyright (c) 2009 The Go Authors. All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above
copyright notice, this list of conditions and the following disclaimer
in the documentation and/or other materials provided with the
distribution.
* Neither the name of Google Inc. nor the names of its
contributors may be used to endorse or promote products derived from
this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
vendor/golang.org/x/sync/PATENTS
0 → 100644
View file @
a4e4c76c
Additional IP Rights Grant (Patents)
"This implementation" means the copyrightable works distributed by
Google as part of the Go project.
Google hereby grants to You a perpetual, worldwide, non-exclusive,
no-charge, royalty-free, irrevocable (except as stated in this section)
patent license to make, have made, use, offer to sell, sell, import,
transfer and otherwise run, modify and propagate the contents of this
implementation of Go, where such license applies only to those patent
claims, both currently owned or controlled by Google and acquired in
the future, licensable by Google that are necessarily infringed by this
implementation of Go. This grant does not include claims that would be
infringed only as a consequence of further modification of this
implementation. If you or your agent or exclusive licensee institute or
order or agree to the institution of patent litigation against any
entity (including a cross-claim or counterclaim in a lawsuit) alleging
that this implementation of Go or any code incorporated within this
implementation of Go constitutes direct or contributory patent
infringement, or inducement of patent infringement, then any patent
rights granted to you under this License for this implementation of Go
shall terminate as of the date such litigation is filed.
vendor/golang.org/x/sync/syncmap/map.go
0 → 100644
View file @
a4e4c76c
// Copyright 2016 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
// Package syncmap provides a concurrent map implementation.
// It is a prototype for a proposed addition to the sync package
// in the standard library.
// (https://golang.org/issue/18177)
package
syncmap
import
(
"sync"
"sync/atomic"
"unsafe"
)
// Map is a concurrent map with amortized-constant-time loads, stores, and deletes.
// It is safe for multiple goroutines to call a Map's methods concurrently.
//
// The zero Map is valid and empty.
//
// A Map must not be copied after first use.
type
Map
struct
{
mu
sync
.
Mutex
// read contains the portion of the map's contents that are safe for
// concurrent access (with or without mu held).
//
// The read field itself is always safe to load, but must only be stored with
// mu held.
//
// Entries stored in read may be updated concurrently without mu, but updating
// a previously-expunged entry requires that the entry be copied to the dirty
// map and unexpunged with mu held.
read
atomic
.
Value
// readOnly
// dirty contains the portion of the map's contents that require mu to be
// held. To ensure that the dirty map can be promoted to the read map quickly,
// it also includes all of the non-expunged entries in the read map.
//
// Expunged entries are not stored in the dirty map. An expunged entry in the
// clean map must be unexpunged and added to the dirty map before a new value
// can be stored to it.
//
// If the dirty map is nil, the next write to the map will initialize it by
// making a shallow copy of the clean map, omitting stale entries.
dirty
map
[
interface
{}]
*
entry
// misses counts the number of loads since the read map was last updated that
// needed to lock mu to determine whether the key was present.
//
// Once enough misses have occurred to cover the cost of copying the dirty
// map, the dirty map will be promoted to the read map (in the unamended
// state) and the next store to the map will make a new dirty copy.
misses
int
}
// readOnly is an immutable struct stored atomically in the Map.read field.
type
readOnly
struct
{
m
map
[
interface
{}]
*
entry
amended
bool
// true if the dirty map contains some key not in m.
}
// expunged is an arbitrary pointer that marks entries which have been deleted
// from the dirty map.
var
expunged
=
unsafe
.
Pointer
(
new
(
interface
{}))
// An entry is a slot in the map corresponding to a particular key.
type
entry
struct
{
// p points to the interface{} value stored for the entry.
//
// If p == nil, the entry has been deleted and m.dirty == nil.
//
// If p == expunged, the entry has been deleted, m.dirty != nil, and the entry
// is missing from m.dirty.
//
// Otherwise, the entry is valid and recorded in m.read.m[key] and, if m.dirty
// != nil, in m.dirty[key].
//
// An entry can be deleted by atomic replacement with nil: when m.dirty is
// next created, it will atomically replace nil with expunged and leave
// m.dirty[key] unset.
//
// An entry's associated value can be updated by atomic replacement, provided
// p != expunged. If p == expunged, an entry's associated value can be updated
// only after first setting m.dirty[key] = e so that lookups using the dirty
// map find the entry.
p
unsafe
.
Pointer
// *interface{}
}
func
newEntry
(
i
interface
{})
*
entry
{
return
&
entry
{
p
:
unsafe
.
Pointer
(
&
i
)}
}
// Load returns the value stored in the map for a key, or nil if no
// value is present.
// The ok result indicates whether value was found in the map.
func
(
m
*
Map
)
Load
(
key
interface
{})
(
value
interface
{},
ok
bool
)
{
read
,
_
:=
m
.
read
.
Load
()
.
(
readOnly
)
e
,
ok
:=
read
.
m
[
key
]
if
!
ok
&&
read
.
amended
{
m
.
mu
.
Lock
()
// Avoid reporting a spurious miss if m.dirty got promoted while we were
// blocked on m.mu. (If further loads of the same key will not miss, it's
// not worth copying the dirty map for this key.)
read
,
_
=
m
.
read
.
Load
()
.
(
readOnly
)
e
,
ok
=
read
.
m
[
key
]
if
!
ok
&&
read
.
amended
{
e
,
ok
=
m
.
dirty
[
key
]
// Regardless of whether the entry was present, record a miss: this key
// will take the slow path until the dirty map is promoted to the read
// map.
m
.
missLocked
()
}
m
.
mu
.
Unlock
()
}
if
!
ok
{
return
nil
,
false
}
return
e
.
load
()
}
func
(
e
*
entry
)
load
()
(
value
interface
{},
ok
bool
)
{
p
:=
atomic
.
LoadPointer
(
&
e
.
p
)
if
p
==
nil
||
p
==
expunged
{
return
nil
,
false
}
return
*
(
*
interface
{})(
p
),
true
}
// Store sets the value for a key.
func
(
m
*
Map
)
Store
(
key
,
value
interface
{})
{
read
,
_
:=
m
.
read
.
Load
()
.
(
readOnly
)
if
e
,
ok
:=
read
.
m
[
key
];
ok
&&
e
.
tryStore
(
&
value
)
{
return
}
m
.
mu
.
Lock
()
read
,
_
=
m
.
read
.
Load
()
.
(
readOnly
)
if
e
,
ok
:=
read
.
m
[
key
];
ok
{
if
e
.
unexpungeLocked
()
{
// The entry was previously expunged, which implies that there is a
// non-nil dirty map and this entry is not in it.
m
.
dirty
[
key
]
=
e
}
e
.
storeLocked
(
&
value
)
}
else
if
e
,
ok
:=
m
.
dirty
[
key
];
ok
{
e
.
storeLocked
(
&
value
)
}
else
{
if
!
read
.
amended
{
// We're adding the first new key to the dirty map.
// Make sure it is allocated and mark the read-only map as incomplete.
m
.
dirtyLocked
()
m
.
read
.
Store
(
readOnly
{
m
:
read
.
m
,
amended
:
true
})
}
m
.
dirty
[
key
]
=
newEntry
(
value
)
}
m
.
mu
.
Unlock
()
}
// tryStore stores a value if the entry has not been expunged.
//
// If the entry is expunged, tryStore returns false and leaves the entry
// unchanged.
func
(
e
*
entry
)
tryStore
(
i
*
interface
{})
bool
{
p
:=
atomic
.
LoadPointer
(
&
e
.
p
)
if
p
==
expunged
{
return
false
}
for
{
if
atomic
.
CompareAndSwapPointer
(
&
e
.
p
,
p
,
unsafe
.
Pointer
(
i
))
{
return
true
}
p
=
atomic
.
LoadPointer
(
&
e
.
p
)
if
p
==
expunged
{
return
false
}
}
}
// unexpungeLocked ensures that the entry is not marked as expunged.
//
// If the entry was previously expunged, it must be added to the dirty map
// before m.mu is unlocked.
func
(
e
*
entry
)
unexpungeLocked
()
(
wasExpunged
bool
)
{
return
atomic
.
CompareAndSwapPointer
(
&
e
.
p
,
expunged
,
nil
)
}
// storeLocked unconditionally stores a value to the entry.
//
// The entry must be known not to be expunged.
func
(
e
*
entry
)
storeLocked
(
i
*
interface
{})
{
atomic
.
StorePointer
(
&
e
.
p
,
unsafe
.
Pointer
(
i
))
}
// LoadOrStore returns the existing value for the key if present.
// Otherwise, it stores and returns the given value.
// The loaded result is true if the value was loaded, false if stored.
func
(
m
*
Map
)
LoadOrStore
(
key
,
value
interface
{})
(
actual
interface
{},
loaded
bool
)
{
// Avoid locking if it's a clean hit.
read
,
_
:=
m
.
read
.
Load
()
.
(
readOnly
)
if
e
,
ok
:=
read
.
m
[
key
];
ok
{
actual
,
loaded
,
ok
:=
e
.
tryLoadOrStore
(
value
)
if
ok
{
return
actual
,
loaded
}
}
m
.
mu
.
Lock
()
read
,
_
=
m
.
read
.
Load
()
.
(
readOnly
)
if
e
,
ok
:=
read
.
m
[
key
];
ok
{
if
e
.
unexpungeLocked
()
{
m
.
dirty
[
key
]
=
e
}
actual
,
loaded
,
_
=
e
.
tryLoadOrStore
(
value
)
}
else
if
e
,
ok
:=
m
.
dirty
[
key
];
ok
{
actual
,
loaded
,
_
=
e
.
tryLoadOrStore
(
value
)
m
.
missLocked
()
}
else
{
if
!
read
.
amended
{
// We're adding the first new key to the dirty map.
// Make sure it is allocated and mark the read-only map as incomplete.
m
.
dirtyLocked
()
m
.
read
.
Store
(
readOnly
{
m
:
read
.
m
,
amended
:
true
})
}
m
.
dirty
[
key
]
=
newEntry
(
value
)
actual
,
loaded
=
value
,
false
}
m
.
mu
.
Unlock
()
return
actual
,
loaded
}
// tryLoadOrStore atomically loads or stores a value if the entry is not
// expunged.
//
// If the entry is expunged, tryLoadOrStore leaves the entry unchanged and
// returns with ok==false.
func
(
e
*
entry
)
tryLoadOrStore
(
i
interface
{})
(
actual
interface
{},
loaded
,
ok
bool
)
{
p
:=
atomic
.
LoadPointer
(
&
e
.
p
)
if
p
==
expunged
{
return
nil
,
false
,
false
}
if
p
!=
nil
{
return
*
(
*
interface
{})(
p
),
true
,
true
}
// Copy the interface after the first load to make this method more amenable
// to escape analysis: if we hit the "load" path or the entry is expunged, we
// shouldn't bother heap-allocating.
ic
:=
i
for
{
if
atomic
.
CompareAndSwapPointer
(
&
e
.
p
,
nil
,
unsafe
.
Pointer
(
&
ic
))
{
return
i
,
false
,
true
}
p
=
atomic
.
LoadPointer
(
&
e
.
p
)
if
p
==
expunged
{
return
nil
,
false
,
false
}
if
p
!=
nil
{
return
*
(
*
interface
{})(
p
),
true
,
true
}
}
}
// Delete deletes the value for a key.
func
(
m
*
Map
)
Delete
(
key
interface
{})
{
read
,
_
:=
m
.
read
.
Load
()
.
(
readOnly
)
e
,
ok
:=
read
.
m
[
key
]
if
!
ok
&&
read
.
amended
{
m
.
mu
.
Lock
()
read
,
_
=
m
.
read
.
Load
()
.
(
readOnly
)
e
,
ok
=
read
.
m
[
key
]
if
!
ok
&&
read
.
amended
{
delete
(
m
.
dirty
,
key
)
}
m
.
mu
.
Unlock
()
}
if
ok
{
e
.
delete
()
}
}
func
(
e
*
entry
)
delete
()
(
hadValue
bool
)
{
for
{
p
:=
atomic
.
LoadPointer
(
&
e
.
p
)
if
p
==
nil
||
p
==
expunged
{
return
false
}
if
atomic
.
CompareAndSwapPointer
(
&
e
.
p
,
p
,
nil
)
{
return
true
}
}
}
// Range calls f sequentially for each key and value present in the map.
// If f returns false, range stops the iteration.
//
// Range does not necessarily correspond to any consistent snapshot of the Map's
// contents: no key will be visited more than once, but if the value for any key
// is stored or deleted concurrently, Range may reflect any mapping for that key
// from any point during the Range call.
//
// Range may be O(N) with the number of elements in the map even if f returns
// false after a constant number of calls.
func
(
m
*
Map
)
Range
(
f
func
(
key
,
value
interface
{})
bool
)
{
// We need to be able to iterate over all of the keys that were already
// present at the start of the call to Range.
// If read.amended is false, then read.m satisfies that property without
// requiring us to hold m.mu for a long time.
read
,
_
:=
m
.
read
.
Load
()
.
(
readOnly
)
if
read
.
amended
{
// m.dirty contains keys not in read.m. Fortunately, Range is already O(N)
// (assuming the caller does not break out early), so a call to Range
// amortizes an entire copy of the map: we can promote the dirty copy
// immediately!
m
.
mu
.
Lock
()
read
,
_
=
m
.
read
.
Load
()
.
(
readOnly
)
if
read
.
amended
{
read
=
readOnly
{
m
:
m
.
dirty
}
m
.
read
.
Store
(
read
)
m
.
dirty
=
nil
m
.
misses
=
0
}
m
.
mu
.
Unlock
()
}
for
k
,
e
:=
range
read
.
m
{
v
,
ok
:=
e
.
load
()
if
!
ok
{
continue
}
if
!
f
(
k
,
v
)
{
break
}
}
}
func
(
m
*
Map
)
missLocked
()
{
m
.
misses
++
if
m
.
misses
<
len
(
m
.
dirty
)
{
return
}
m
.
read
.
Store
(
readOnly
{
m
:
m
.
dirty
})
m
.
dirty
=
nil
m
.
misses
=
0
}
func
(
m
*
Map
)
dirtyLocked
()
{
if
m
.
dirty
!=
nil
{
return
}
read
,
_
:=
m
.
read
.
Load
()
.
(
readOnly
)
m
.
dirty
=
make
(
map
[
interface
{}]
*
entry
,
len
(
read
.
m
))
for
k
,
e
:=
range
read
.
m
{
if
!
e
.
tryExpungeLocked
()
{
m
.
dirty
[
k
]
=
e
}
}
}
func
(
e
*
entry
)
tryExpungeLocked
()
(
isExpunged
bool
)
{
p
:=
atomic
.
LoadPointer
(
&
e
.
p
)
for
p
==
nil
{
if
atomic
.
CompareAndSwapPointer
(
&
e
.
p
,
nil
,
expunged
)
{
return
true
}
p
=
atomic
.
LoadPointer
(
&
e
.
p
)
}
return
p
==
expunged
}
vendor/vendor.json
View file @
a4e4c76c
...
...
@@ -501,6 +501,12 @@
"revision"
:
"b4690f45fa1cafc47b1c280c2e75116efe40cc13"
,
"revisionTime"
:
"2017-02-15T08:41:58Z"
},
{
"checksumSHA1"
:
"4TEYFKrAUuwBMqExjQBsnf/CgjQ="
,
"path"
:
"golang.org/x/sync/syncmap"
,
"revision"
:
"f52d1811a62927559de87708c8913c1650ce4f26"
,
"revisionTime"
:
"2017-05-17T20:25:26Z"
},
{
"checksumSHA1"
:
"rTPzsn0jeqfgnQR0OsMKR8JRy5Y="
,
"path"
:
"golang.org/x/sys/unix"
,
...
...
whisper/whisperv5/whisper.go
View file @
a4e4c76c
...
...
@@ -26,8 +26,6 @@ import (
"sync"
"time"
"sync/atomic"
"github.com/ethereum/go-ethereum/common"
"github.com/ethereum/go-ethereum/crypto"
"github.com/ethereum/go-ethereum/log"
...
...
@@ -35,6 +33,7 @@ import (
"github.com/ethereum/go-ethereum/rpc"
"github.com/syndtr/goleveldb/leveldb/errors"
"golang.org/x/crypto/pbkdf2"
"golang.org/x/sync/syncmap"
set
"gopkg.in/fatih/set.v0"
)
...
...
@@ -46,38 +45,12 @@ type Statistics struct {
totalMessagesCleared
int
}
type
settingType
byte
type
settingsMap
map
[
settingType
]
interface
{}
const
(
minPowIdx
settingType
=
iota
// Minimal PoW required by the whisper node
maxMsgSizeIdx
settingType
=
iota
// Maximal message length allowed by the whisper node
OverflowIdx
settingType
=
iota
// Indicator of message queue overflow
minPowIdx
=
iota
// Minimal PoW required by the whisper node
maxMsgSizeIdx
=
iota
// Maximal message length allowed by the whisper node
overflowIdx
=
iota
// Indicator of message queue overflow
)
type
settingsVault
struct
{
vaultMu
sync
.
Mutex
vault
atomic
.
Value
}
func
(
s
*
settingsVault
)
get
(
idx
settingType
)
interface
{}
{
m
:=
s
.
vault
.
Load
()
.
(
settingsMap
)
return
m
[
idx
]
}
func
(
s
*
settingsVault
)
store
(
idx
settingType
,
val
interface
{})
{
s
.
vaultMu
.
Lock
()
defer
s
.
vaultMu
.
Unlock
()
m1
:=
s
.
vault
.
Load
()
.
(
settingsMap
)
m2
:=
make
(
settingsMap
)
for
k
,
v
:=
range
m1
{
m2
[
k
]
=
v
}
m2
[
idx
]
=
val
s
.
vault
.
Store
(
m2
)
}
// Whisper represents a dark communication interface through the Ethereum
// network, using its very own P2P communication layer.
type
Whisper
struct
{
...
...
@@ -99,7 +72,7 @@ type Whisper struct {
p2pMsgQueue
chan
*
Envelope
// Message queue for peer-to-peer messages (not to be forwarded any further)
quit
chan
struct
{}
// Channel used for graceful exit
settings
s
ettingsVault
// holds configuration settings that can be dynamically changed
settings
s
yncmap
.
Map
// holds configuration settings that can be dynamically changed
statsMu
sync
.
Mutex
// guard stats
stats
Statistics
// Statistics of whisper node
...
...
@@ -126,10 +99,9 @@ func New(cfg *Config) *Whisper {
whisper
.
filters
=
NewFilters
(
whisper
)
whisper
.
settings
.
vault
.
Store
(
make
(
settingsMap
))
whisper
.
settings
.
store
(
minPowIdx
,
cfg
.
MinimumAcceptedPOW
)
whisper
.
settings
.
store
(
maxMsgSizeIdx
,
cfg
.
MaxMessageSize
)
whisper
.
settings
.
store
(
OverflowIdx
,
false
)
whisper
.
settings
.
Store
(
minPowIdx
,
cfg
.
MinimumAcceptedPOW
)
whisper
.
settings
.
Store
(
maxMsgSizeIdx
,
cfg
.
MaxMessageSize
)
whisper
.
settings
.
Store
(
overflowIdx
,
false
)
// p2p whisper sub protocol handler
whisper
.
protocol
=
p2p
.
Protocol
{
...
...
@@ -150,17 +122,20 @@ func New(cfg *Config) *Whisper {
}
func
(
w
*
Whisper
)
MinPow
()
float64
{
return
w
.
settings
.
get
(
minPowIdx
)
.
(
float64
)
val
,
_
:=
w
.
settings
.
Load
(
minPowIdx
)
return
val
.
(
float64
)
}
// MaxMessageSize returns the maximum accepted message size.
func
(
w
*
Whisper
)
MaxMessageSize
()
uint32
{
return
w
.
settings
.
get
(
maxMsgSizeIdx
)
.
(
uint32
)
val
,
_
:=
w
.
settings
.
Load
(
maxMsgSizeIdx
)
return
val
.
(
uint32
)
}
// Overflow returns an indication if the message queue is full.
func
(
w
*
Whisper
)
Overflow
()
bool
{
return
w
.
settings
.
get
(
OverflowIdx
)
.
(
bool
)
val
,
_
:=
w
.
settings
.
Load
(
overflowIdx
)
return
val
.
(
bool
)
}
// APIs returns the RPC descriptors the Whisper implementation offers
...
...
@@ -196,7 +171,7 @@ func (w *Whisper) SetMaxMessageSize(size uint32) error {
if
size
>
MaxMessageSize
{
return
fmt
.
Errorf
(
"message size too large [%d>%d]"
,
size
,
MaxMessageSize
)
}
w
.
settings
.
s
tore
(
maxMsgSizeIdx
,
uint32
(
size
))
w
.
settings
.
S
tore
(
maxMsgSizeIdx
,
uint32
(
size
))
return
nil
}
...
...
@@ -205,7 +180,7 @@ func (w *Whisper) SetMinimumPoW(val float64) error {
if
val
<=
0.0
{
return
fmt
.
Errorf
(
"invalid PoW: %f"
,
val
)
}
w
.
settings
.
s
tore
(
minPowIdx
,
val
)
w
.
settings
.
S
tore
(
minPowIdx
,
val
)
return
nil
}
...
...
@@ -679,12 +654,12 @@ func (w *Whisper) checkOverflow() {
if
queueSize
==
messageQueueLimit
{
if
!
w
.
Overflow
()
{
w
.
settings
.
store
(
O
verflowIdx
,
true
)
w
.
settings
.
Store
(
o
verflowIdx
,
true
)
log
.
Warn
(
"message queue overflow"
)
}
}
else
if
queueSize
<=
messageQueueLimit
/
2
{
if
w
.
Overflow
()
{
w
.
settings
.
store
(
O
verflowIdx
,
false
)
w
.
settings
.
Store
(
o
verflowIdx
,
false
)
log
.
Warn
(
"message queue overflow fixed (back to normal)"
)
}
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment