Skip to content

Commit 5c21c68

Browse files
committed
Migrate FilesystemPersister tests to FilesystemStore
1 parent ca93285 commit 5c21c68

File tree

3 files changed

+213
-2
lines changed

3 files changed

+213
-2
lines changed

lightning-storage/Cargo.toml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,10 @@ rustdoc-args = ["--cfg", "docsrs"]
1717
lightning = { version = "0.0.116", path = "../lightning", default-features = false, features = ["std"] }
1818
libc = "0.2"
1919

20+
[dev-dependencies]
21+
bitcoin = { version = "0.29.0", default-features = false }
22+
lightning = { version = "0.0.116", path = "../lightning", features = ["_test_utils"] }
23+
2024
[target.'cfg(windows)'.dependencies]
2125
windows-sys = { version = "0.48.0", default-features = false, features = ["Win32_Storage_FileSystem", "Win32_Foundation"] }
2226

lightning-storage/src/fs_store.rs

Lines changed: 139 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -261,12 +261,150 @@ impl Read for FilesystemReader {
261261
#[cfg(test)]
262262
mod tests {
263263
use super::*;
264-
use crate::test_utils::do_read_write_remove_list_persist;
264+
use crate::test_utils::{do_read_write_remove_list_persist, do_test_store};
265+
266+
use bitcoin::hashes::hex::FromHex;
267+
use bitcoin::Txid;
268+
269+
use lightning::chain::ChannelMonitorUpdateStatus;
270+
use lightning::chain::chainmonitor::Persist;
271+
use lightning::chain::transaction::OutPoint;
272+
use lightning::check_closed_event;
273+
use lightning::events::{ClosureReason, MessageSendEventsProvider};
274+
use lightning::ln::functional_test_utils::*;
275+
use lightning::util::test_utils;
276+
use lightning::util::persist::read_channel_monitors;
277+
use std::fs;
278+
#[cfg(target_os = "windows")]
279+
use {
280+
lightning::get_event_msg,
281+
lightning::ln::msgs::ChannelMessageHandler,
282+
};
283+
284+
impl Drop for FilesystemStore{
285+
fn drop(&mut self) {
286+
// We test for invalid directory names, so it's OK if directory removal
287+
// fails.
288+
match fs::remove_dir_all(&self.data_dir) {
289+
Err(e) => println!("Failed to remove test persister directory: {}", e),
290+
_ => {}
291+
}
292+
}
293+
}
265294

266295
#[test]
267296
fn read_write_remove_list_persist() {
268297
let temp_path = std::env::temp_dir();
269298
let fs_store = FilesystemStore::new(temp_path);
270299
do_read_write_remove_list_persist(&fs_store);
271300
}
301+
302+
#[test]
303+
fn test_if_monitors_is_not_dir() {
304+
let store = FilesystemStore::new("test_monitors_is_not_dir".into());
305+
306+
fs::create_dir_all(&store.get_data_dir()).unwrap();
307+
let mut path = std::path::PathBuf::from(&store.get_data_dir());
308+
path.push("monitors");
309+
fs::File::create(path).unwrap();
310+
311+
let chanmon_cfgs = create_chanmon_cfgs(1);
312+
let mut node_cfgs = create_node_cfgs(1, &chanmon_cfgs);
313+
let chain_mon_0 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[0].chain_source), &chanmon_cfgs[0].tx_broadcaster, &chanmon_cfgs[0].logger, &chanmon_cfgs[0].fee_estimator, &store, node_cfgs[0].keys_manager);
314+
node_cfgs[0].chain_monitor = chain_mon_0;
315+
let node_chanmgrs = create_node_chanmgrs(1, &node_cfgs, &[None]);
316+
let nodes = create_network(1, &node_cfgs, &node_chanmgrs);
317+
318+
// Check that read_channel_monitors() returns error if monitors/ is not a
319+
// directory.
320+
assert!(read_channel_monitors(&store, nodes[0].keys_manager, nodes[0].keys_manager).is_err());
321+
}
322+
323+
#[test]
324+
fn test_filesystem_store() {
325+
// Create the nodes, giving them FilesystemStores for data stores.
326+
let store_0 = FilesystemStore::new("test_filesystem_store_0".into());
327+
let store_1 = FilesystemStore::new("test_filesystem_store_1".into());
328+
do_test_store(&store_0, &store_1)
329+
}
330+
331+
// Test that if the store's path to channel data is read-only, writing a
332+
// monitor to it results in the store returning a PermanentFailure.
333+
// Windows ignores the read-only flag for folders, so this test is Unix-only.
334+
#[cfg(not(target_os = "windows"))]
335+
#[test]
336+
fn test_readonly_dir_perm_failure() {
337+
let store = FilesystemStore::new("test_readonly_dir_perm_failure".into());
338+
fs::create_dir_all(&store.get_data_dir()).unwrap();
339+
340+
// Set up a dummy channel and force close. This will produce a monitor
341+
// that we can then use to test persistence.
342+
let chanmon_cfgs = create_chanmon_cfgs(2);
343+
let node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
344+
let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
345+
let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
346+
let chan = create_announced_chan_between_nodes(&nodes, 0, 1);
347+
nodes[1].node.force_close_broadcasting_latest_txn(&chan.2, &nodes[0].node.get_our_node_id()).unwrap();
348+
check_closed_event!(nodes[1], 1, ClosureReason::HolderForceClosed, [nodes[0].node.get_our_node_id()], 100000);
349+
let mut added_monitors = nodes[1].chain_monitor.added_monitors.lock().unwrap();
350+
let update_map = nodes[1].chain_monitor.latest_monitor_update_id.lock().unwrap();
351+
let update_id = update_map.get(&added_monitors[0].0.to_channel_id()).unwrap();
352+
353+
// Set the store's directory to read-only, which should result in
354+
// returning a permanent failure when we then attempt to persist a
355+
// channel update.
356+
let path = &store.get_data_dir();
357+
let mut perms = fs::metadata(path).unwrap().permissions();
358+
perms.set_readonly(true);
359+
fs::set_permissions(path, perms).unwrap();
360+
361+
let test_txo = OutPoint {
362+
txid: Txid::from_hex("8984484a580b825b9972d7adb15050b3ab624ccd731946b3eeddb92f4e7ef6be").unwrap(),
363+
index: 0
364+
};
365+
match store.persist_new_channel(test_txo, &added_monitors[0].1, update_id.2) {
366+
ChannelMonitorUpdateStatus::PermanentFailure => {},
367+
_ => panic!("unexpected result from persisting new channel")
368+
}
369+
370+
nodes[1].node.get_and_clear_pending_msg_events();
371+
added_monitors.clear();
372+
}
373+
374+
// Test that if a store's directory name is invalid, monitor persistence
375+
// will fail.
376+
#[cfg(target_os = "windows")]
377+
#[test]
378+
fn test_fail_on_open() {
379+
// Set up a dummy channel and force close. This will produce a monitor
380+
// that we can then use to test persistence.
381+
let chanmon_cfgs = create_chanmon_cfgs(2);
382+
let mut node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
383+
let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
384+
let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
385+
let chan = create_announced_chan_between_nodes(&nodes, 0, 1);
386+
nodes[1].node.force_close_broadcasting_latest_txn(&chan.2, &nodes[0].node.get_our_node_id()).unwrap();
387+
check_closed_event!(nodes[1], 1, ClosureReason::HolderForceClosed, [nodes[0].node.get_our_node_id()], 100000);
388+
let mut added_monitors = nodes[1].chain_monitor.added_monitors.lock().unwrap();
389+
let update_map = nodes[1].chain_monitor.latest_monitor_update_id.lock().unwrap();
390+
let update_id = update_map.get(&added_monitors[0].0.to_channel_id()).unwrap();
391+
392+
// Create the store with an invalid directory name and test that the
393+
// channel fails to open because the directories fail to be created. There
394+
// don't seem to be invalid filename characters on Unix that Rust doesn't
395+
// handle, hence why the test is Windows-only.
396+
let store = FilesystemStore::new(":<>/".into());
397+
398+
let test_txo = OutPoint {
399+
txid: Txid::from_hex("8984484a580b825b9972d7adb15050b3ab624ccd731946b3eeddb92f4e7ef6be").unwrap(),
400+
index: 0
401+
};
402+
match store.persist_new_channel(test_txo, &added_monitors[0].1, update_id.2) {
403+
ChannelMonitorUpdateStatus::PermanentFailure => {},
404+
_ => panic!("unexpected result from persisting new channel")
405+
}
406+
407+
nodes[1].node.get_and_clear_pending_msg_events();
408+
added_monitors.clear();
409+
}
272410
}

lightning-storage/src/test_utils.rs

Lines changed: 70 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,9 @@
1-
use lightning::util::persist::KVStore;
1+
use lightning::util::persist::{KVStore, read_channel_monitors};
2+
use lightning::ln::functional_test_utils::*;
3+
use lightning::chain::channelmonitor::CLOSED_CHANNEL_UPDATE_ID;
4+
use lightning::util::test_utils;
5+
use lightning::{check_closed_broadcast, check_closed_event, check_added_monitors};
6+
use lightning::events::ClosureReason;
27

38
pub(crate) fn do_read_write_remove_list_persist<K: KVStore>(kv_store: &K) {
49
use lightning::util::ser::Readable;
@@ -28,3 +33,67 @@ pub(crate) fn do_read_write_remove_list_persist<K: KVStore>(kv_store: &K) {
2833
let listed_keys = kv_store.list(namespace).unwrap();
2934
assert_eq!(listed_keys.len(), 0);
3035
}
36+
37+
// Integration-test the given KVStore implementation. Test relaying a few payments and check that
38+
// the persisted data is updated the appropriate number of times.
39+
pub(crate) fn do_test_store<K: KVStore>(store_0: &K, store_1: &K) {
40+
let chanmon_cfgs = create_chanmon_cfgs(2);
41+
let mut node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
42+
let chain_mon_0 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[0].chain_source), &chanmon_cfgs[0].tx_broadcaster, &chanmon_cfgs[0].logger, &chanmon_cfgs[0].fee_estimator, store_0, node_cfgs[0].keys_manager);
43+
let chain_mon_1 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[1].chain_source), &chanmon_cfgs[1].tx_broadcaster, &chanmon_cfgs[1].logger, &chanmon_cfgs[1].fee_estimator, store_1, node_cfgs[1].keys_manager);
44+
node_cfgs[0].chain_monitor = chain_mon_0;
45+
node_cfgs[1].chain_monitor = chain_mon_1;
46+
let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
47+
let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
48+
49+
// Check that the persisted channel data is empty before any channels are
50+
// open.
51+
let mut persisted_chan_data_0 = read_channel_monitors(store_0, nodes[0].keys_manager, nodes[0].keys_manager).unwrap();
52+
assert_eq!(persisted_chan_data_0.len(), 0);
53+
let mut persisted_chan_data_1 = read_channel_monitors(store_1, nodes[1].keys_manager, nodes[1].keys_manager).unwrap();
54+
assert_eq!(persisted_chan_data_1.len(), 0);
55+
56+
// Helper to make sure the channel is on the expected update ID.
57+
macro_rules! check_persisted_data {
58+
($expected_update_id: expr) => {
59+
persisted_chan_data_0 = read_channel_monitors(store_0, nodes[0].keys_manager, nodes[0].keys_manager).unwrap();
60+
assert_eq!(persisted_chan_data_0.len(), 1);
61+
for (_, mon) in persisted_chan_data_0.iter() {
62+
assert_eq!(mon.get_latest_update_id(), $expected_update_id);
63+
}
64+
persisted_chan_data_1 = read_channel_monitors(store_1, nodes[1].keys_manager, nodes[1].keys_manager).unwrap();
65+
assert_eq!(persisted_chan_data_1.len(), 1);
66+
for (_, mon) in persisted_chan_data_1.iter() {
67+
assert_eq!(mon.get_latest_update_id(), $expected_update_id);
68+
}
69+
}
70+
}
71+
72+
// Create some initial channel and check that a channel was persisted.
73+
let _ = create_announced_chan_between_nodes(&nodes, 0, 1);
74+
check_persisted_data!(0);
75+
76+
// Send a few payments and make sure the monitors are updated to the latest.
77+
send_payment(&nodes[0], &vec!(&nodes[1])[..], 8000000);
78+
check_persisted_data!(5);
79+
send_payment(&nodes[1], &vec!(&nodes[0])[..], 4000000);
80+
check_persisted_data!(10);
81+
82+
// Force close because cooperative close doesn't result in any persisted
83+
// updates.
84+
nodes[0].node.force_close_broadcasting_latest_txn(&nodes[0].node.list_channels()[0].channel_id, &nodes[1].node.get_our_node_id()).unwrap();
85+
check_closed_event!(nodes[0], 1, ClosureReason::HolderForceClosed, [nodes[1].node.get_our_node_id()], 100000);
86+
check_closed_broadcast!(nodes[0], true);
87+
check_added_monitors!(nodes[0], 1);
88+
89+
let node_txn = nodes[0].tx_broadcaster.txn_broadcasted.lock().unwrap();
90+
assert_eq!(node_txn.len(), 1);
91+
92+
connect_block(&nodes[1], &create_dummy_block(nodes[0].best_block_hash(), 42, vec![node_txn[0].clone(), node_txn[0].clone()]));
93+
check_closed_broadcast!(nodes[1], true);
94+
check_closed_event!(nodes[1], 1, ClosureReason::CommitmentTxConfirmed, [nodes[0].node.get_our_node_id()], 100000);
95+
check_added_monitors!(nodes[1], 1);
96+
97+
// Make sure everything is persisted as expected after close.
98+
check_persisted_data!(CLOSED_CHANNEL_UPDATE_ID);
99+
}

0 commit comments

Comments
 (0)