Unusually high number of inactive channels after upgrade to LND v0.19.0-beta #9870
Replies: 15 comments 20 replies
-
This has happened to me with two channels since the update to 0.19.0, with one of them the problem has been solved when his node has been restarted and I suppose that with the second peer it will be solved as soon as he restarts his node. my log, message repeats every 3minutes
peer log
|
Beta Was this translation helpful? Give feedback.
-
My node has been on all 0.19.0-RCs and upgraded to release a few days ago without issue. The above mentioned issues only occurred after channel peers also upgraded to LND 0.19 – and those channels then remained as Adding my logs: Similar from my side as above
This looks like the last time the channel connected before the downtime:
|
Beta Was this translation helpful? Give feedback.
-
Interesting. The very surprising issue I saw on my end -- where two of my OWN nodes, in the SAME datacenter, showed channels between them as |
Beta Was this translation helpful? Give feedback.
-
I'm just focusing now on two computers, in the same datacenter, both running Checking just a few minutes ago, I'm finding the once again these 5 (five) channels are However the other node DOES show some interesting logs, which DO mention the partner node:
|
Beta Was this translation helpful? Give feedback.
-
One of the nodes, despite CMGR=debug & PEER=debug, showed no mention of the other node in the logs. I have now restarted both the underlying machine and the LND & other docker containers, and the channels came back online, which is the same behavior I have been seeing for days.... the channels operate well at first, but then after some number of hours the channels go This stuff is tricky to debug, so ..... I think we need some OTHER people to report similar problems (ideally in this thread) to really dig into this. If so far it's only myself and @Filouman reporting issues, it still could be something flaky with our networking or machines. |
Beta Was this translation helpful? Give feedback.
-
For these channels, are you able to find any sort of errors on the A channel won't ever go to active if we don't get the channel reest message for it. A channel may go from active to inactive, if we get some error at the link (channel) level, that may have us send an error/warning, but not disconnect. |
Beta Was this translation helpful? Give feedback.
-
OK, good, I will put HSWC=debug and then see if I can catch any useful logs when/if the channels go offline again. |
Beta Was this translation helpful? Give feedback.
-
@faket0shi are you aware of what node implementation your partner with the problem channel is running? It seems like you are running 0.19, but what about your partner node? |
Beta Was this translation helpful? Give feedback.
-
Seeing something similar on my node, one node sees the channel as active the other one as inactive trying to pin point the bug. |
Beta Was this translation helpful? Give feedback.
-
Here is some history of one of the "problem" channels, as seen from one side:
The channel is clearly going up and down a few times per day... and a similar channel with an 0.18.x node, in the same datacenter, doesn't show any intermittency like this. |
Beta Was this translation helpful? Give feedback.
-
@Roasbeef Please see log file attached, showings logs where the problem peer (which is my own node, in the same datacenter) is mentioned. I do see a lot of |
Beta Was this translation helpful? Give feedback.
-
@faket0shi @MegalithicBTC do you have the setting
active ? |
Beta Was this translation helpful? Give feedback.
-
@ziggie1984 I have not set I've now restarted both nodes with..
So this should capture sufficient logs that I could see useful logs from both sides next time this happens. thanks |
Beta Was this translation helpful? Give feedback.
-
This pretty interesting, my experience very closely matches @Filouman ...
exact same experience here
exact same experience, it was only when I had ANOTHER node ALSO upgraded to 0.19 that I saw the problem...
Exact same experience... the channel went offline again |
Beta Was this translation helpful? Give feedback.
-
We now have |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
In the last few days there have been some feedback from node-runners about v0.19.0-beta, where in this version some channels are showing
is_active: false
where in previous versions, the channel was reliably active.We have seen this with two of our nodes which we have updated to v.0.19-0-beta, but trying to diagnose it has been tricky.
After first restarting with the new version, we found that a channel between two of our own nodes was down, despite both nodes living in the same datacenter. We solved this by restarting one of the nodes, and now that channel is up.
There are however persistent issues we have been unable to solve.... A few channels with "clearnet" nodes, where we can see (for example on amboss.space) that their other channels are up, but ours are
is_active: false
.We've already set
no-disconnect-on-pong-failure: true
, inlnd.conf
, but that didn't fix the issue.We're now looking for a way to try to get logs from our impacted node to try to understand why the channels have not come back online.
So far, this is what we've tried
Then...
After doing this, nothing of interest appears in LND's log -- even with these subsystems set to
debug
, there is no mention of this peer's public key on the logs.So right now we don't have much beyond this anecdotal "it's not working".... But I wanted to open this discussion to see if anyone else could provide more useful evidence about what might or might not be going wrong.... or suggest certain commands we could run to help debug this.
Beta Was this translation helpful? Give feedback.
All reactions