Skip to content

Commit f508981

Browse files
committed
Migrate FilesystemPersister tests to FilesystemStore
1 parent 832f32c commit f508981

File tree

3 files changed

+212
-2
lines changed

3 files changed

+212
-2
lines changed

lightning-persister/Cargo.toml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -25,3 +25,4 @@ criterion = { version = "0.4", optional = true, default-features = false }
2525

2626
[dev-dependencies]
2727
lightning = { version = "0.0.116", path = "../lightning", features = ["_test_utils"] }
28+
bitcoin = { version = "0.29.0", default-features = false }

lightning-persister/src/fs_store.rs

Lines changed: 139 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -367,7 +367,36 @@ impl KVStore for FilesystemStore {
367367
#[cfg(test)]
368368
mod tests {
369369
use super::*;
370-
use crate::test_utils::do_read_write_remove_list_persist;
370+
use crate::test_utils::{do_read_write_remove_list_persist, do_test_store};
371+
372+
use bitcoin::hashes::hex::FromHex;
373+
use bitcoin::Txid;
374+
375+
use lightning::chain::ChannelMonitorUpdateStatus;
376+
use lightning::chain::chainmonitor::Persist;
377+
use lightning::chain::transaction::OutPoint;
378+
use lightning::check_closed_event;
379+
use lightning::events::{ClosureReason, MessageSendEventsProvider};
380+
use lightning::ln::functional_test_utils::*;
381+
use lightning::util::test_utils;
382+
use lightning::util::persist::read_channel_monitors;
383+
use std::fs;
384+
#[cfg(target_os = "windows")]
385+
use {
386+
lightning::get_event_msg,
387+
lightning::ln::msgs::ChannelMessageHandler,
388+
};
389+
390+
impl Drop for FilesystemStore {
391+
fn drop(&mut self) {
392+
// We test for invalid directory names, so it's OK if directory removal
393+
// fails.
394+
match fs::remove_dir_all(&self.data_dir) {
395+
Err(e) => println!("Failed to remove test persister directory: {}", e),
396+
_ => {}
397+
}
398+
}
399+
}
371400

372401
#[test]
373402
fn read_write_remove_list_persist() {
@@ -376,4 +405,113 @@ mod tests {
376405
let fs_store = FilesystemStore::new(temp_path);
377406
do_read_write_remove_list_persist(&fs_store);
378407
}
408+
409+
#[test]
410+
fn test_if_monitors_is_not_dir() {
411+
let store = FilesystemStore::new("test_monitors_is_not_dir".into());
412+
413+
fs::create_dir_all(&store.get_data_dir()).unwrap();
414+
let mut path = std::path::PathBuf::from(&store.get_data_dir());
415+
path.push("monitors");
416+
fs::File::create(path).unwrap();
417+
418+
let chanmon_cfgs = create_chanmon_cfgs(1);
419+
let mut node_cfgs = create_node_cfgs(1, &chanmon_cfgs);
420+
let chain_mon_0 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[0].chain_source), &chanmon_cfgs[0].tx_broadcaster, &chanmon_cfgs[0].logger, &chanmon_cfgs[0].fee_estimator, &store, node_cfgs[0].keys_manager);
421+
node_cfgs[0].chain_monitor = chain_mon_0;
422+
let node_chanmgrs = create_node_chanmgrs(1, &node_cfgs, &[None]);
423+
let nodes = create_network(1, &node_cfgs, &node_chanmgrs);
424+
425+
// Check that read_channel_monitors() returns error if monitors/ is not a
426+
// directory.
427+
assert!(read_channel_monitors(&store, nodes[0].keys_manager, nodes[0].keys_manager).is_err());
428+
}
429+
430+
#[test]
431+
fn test_filesystem_store() {
432+
// Create the nodes, giving them FilesystemStores for data stores.
433+
let store_0 = FilesystemStore::new("test_filesystem_store_0".into());
434+
let store_1 = FilesystemStore::new("test_filesystem_store_1".into());
435+
do_test_store(&store_0, &store_1)
436+
}
437+
438+
// Test that if the store's path to channel data is read-only, writing a
439+
// monitor to it results in the store returning a PermanentFailure.
440+
// Windows ignores the read-only flag for folders, so this test is Unix-only.
441+
#[cfg(not(target_os = "windows"))]
442+
#[test]
443+
fn test_readonly_dir_perm_failure() {
444+
let store = FilesystemStore::new("test_readonly_dir_perm_failure".into());
445+
fs::create_dir_all(&store.get_data_dir()).unwrap();
446+
447+
// Set up a dummy channel and force close. This will produce a monitor
448+
// that we can then use to test persistence.
449+
let chanmon_cfgs = create_chanmon_cfgs(2);
450+
let node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
451+
let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
452+
let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
453+
let chan = create_announced_chan_between_nodes(&nodes, 0, 1);
454+
nodes[1].node.force_close_broadcasting_latest_txn(&chan.2, &nodes[0].node.get_our_node_id()).unwrap();
455+
check_closed_event!(nodes[1], 1, ClosureReason::HolderForceClosed, [nodes[0].node.get_our_node_id()], 100000);
456+
let mut added_monitors = nodes[1].chain_monitor.added_monitors.lock().unwrap();
457+
let update_map = nodes[1].chain_monitor.latest_monitor_update_id.lock().unwrap();
458+
let update_id = update_map.get(&added_monitors[0].0.to_channel_id()).unwrap();
459+
460+
// Set the store's directory to read-only, which should result in
461+
// returning a permanent failure when we then attempt to persist a
462+
// channel update.
463+
let path = &store.get_data_dir();
464+
let mut perms = fs::metadata(path).unwrap().permissions();
465+
perms.set_readonly(true);
466+
fs::set_permissions(path, perms).unwrap();
467+
468+
let test_txo = OutPoint {
469+
txid: Txid::from_hex("8984484a580b825b9972d7adb15050b3ab624ccd731946b3eeddb92f4e7ef6be").unwrap(),
470+
index: 0
471+
};
472+
match store.persist_new_channel(test_txo, &added_monitors[0].1, update_id.2) {
473+
ChannelMonitorUpdateStatus::PermanentFailure => {},
474+
_ => panic!("unexpected result from persisting new channel")
475+
}
476+
477+
nodes[1].node.get_and_clear_pending_msg_events();
478+
added_monitors.clear();
479+
}
480+
481+
// Test that if a store's directory name is invalid, monitor persistence
482+
// will fail.
483+
#[cfg(target_os = "windows")]
484+
#[test]
485+
fn test_fail_on_open() {
486+
// Set up a dummy channel and force close. This will produce a monitor
487+
// that we can then use to test persistence.
488+
let chanmon_cfgs = create_chanmon_cfgs(2);
489+
let node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
490+
let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
491+
let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
492+
let chan = create_announced_chan_between_nodes(&nodes, 0, 1);
493+
nodes[1].node.force_close_broadcasting_latest_txn(&chan.2, &nodes[0].node.get_our_node_id()).unwrap();
494+
check_closed_event!(nodes[1], 1, ClosureReason::HolderForceClosed, [nodes[0].node.get_our_node_id()], 100000);
495+
let mut added_monitors = nodes[1].chain_monitor.added_monitors.lock().unwrap();
496+
let update_map = nodes[1].chain_monitor.latest_monitor_update_id.lock().unwrap();
497+
let update_id = update_map.get(&added_monitors[0].0.to_channel_id()).unwrap();
498+
499+
// Create the store with an invalid directory name and test that the
500+
// channel fails to open because the directories fail to be created. There
501+
// don't seem to be invalid filename characters on Unix that Rust doesn't
502+
// handle, hence why the test is Windows-only.
503+
let store = FilesystemStore::new(":<>/".into());
504+
505+
let test_txo = OutPoint {
506+
txid: Txid::from_hex("8984484a580b825b9972d7adb15050b3ab624ccd731946b3eeddb92f4e7ef6be").unwrap(),
507+
index: 0
508+
};
509+
match store.persist_new_channel(test_txo, &added_monitors[0].1, update_id.2) {
510+
ChannelMonitorUpdateStatus::PermanentFailure => {},
511+
_ => panic!("unexpected result from persisting new channel")
512+
}
513+
514+
nodes[1].node.get_and_clear_pending_msg_events();
515+
added_monitors.clear();
516+
}
379517
}

lightning-persister/src/test_utils.rs

Lines changed: 72 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,11 @@
1-
use lightning::util::persist::{KVStore, KVSTORE_NAMESPACE_KEY_MAX_LEN};
1+
use lightning::util::persist::{KVStore, KVSTORE_NAMESPACE_KEY_MAX_LEN, read_channel_monitors};
2+
use lightning::ln::functional_test_utils::{connect_block, create_announced_chan_between_nodes,
3+
create_chanmon_cfgs, create_dummy_block, create_network, create_node_cfgs, create_node_chanmgrs,
4+
send_payment};
5+
use lightning::chain::channelmonitor::CLOSED_CHANNEL_UPDATE_ID;
6+
use lightning::util::test_utils;
7+
use lightning::{check_closed_broadcast, check_closed_event, check_added_monitors};
8+
use lightning::events::ClosureReason;
29

310
use std::panic::RefUnwindSafe;
411

@@ -48,3 +55,67 @@ pub(crate) fn do_read_write_remove_list_persist<K: KVStore + RefUnwindSafe>(kv_s
4855
let listed_keys = kv_store.list(&max_chars, &max_chars).unwrap();
4956
assert_eq!(listed_keys.len(), 0);
5057
}
58+
59+
// Integration-test the given KVStore implementation. Test relaying a few payments and check that
60+
// the persisted data is updated the appropriate number of times.
61+
pub(crate) fn do_test_store<K: KVStore>(store_0: &K, store_1: &K) {
62+
let chanmon_cfgs = create_chanmon_cfgs(2);
63+
let mut node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
64+
let chain_mon_0 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[0].chain_source), &chanmon_cfgs[0].tx_broadcaster, &chanmon_cfgs[0].logger, &chanmon_cfgs[0].fee_estimator, store_0, node_cfgs[0].keys_manager);
65+
let chain_mon_1 = test_utils::TestChainMonitor::new(Some(&chanmon_cfgs[1].chain_source), &chanmon_cfgs[1].tx_broadcaster, &chanmon_cfgs[1].logger, &chanmon_cfgs[1].fee_estimator, store_1, node_cfgs[1].keys_manager);
66+
node_cfgs[0].chain_monitor = chain_mon_0;
67+
node_cfgs[1].chain_monitor = chain_mon_1;
68+
let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
69+
let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
70+
71+
// Check that the persisted channel data is empty before any channels are
72+
// open.
73+
let mut persisted_chan_data_0 = read_channel_monitors(store_0, nodes[0].keys_manager, nodes[0].keys_manager).unwrap();
74+
assert_eq!(persisted_chan_data_0.len(), 0);
75+
let mut persisted_chan_data_1 = read_channel_monitors(store_1, nodes[1].keys_manager, nodes[1].keys_manager).unwrap();
76+
assert_eq!(persisted_chan_data_1.len(), 0);
77+
78+
// Helper to make sure the channel is on the expected update ID.
79+
macro_rules! check_persisted_data {
80+
($expected_update_id: expr) => {
81+
persisted_chan_data_0 = read_channel_monitors(store_0, nodes[0].keys_manager, nodes[0].keys_manager).unwrap();
82+
assert_eq!(persisted_chan_data_0.len(), 1);
83+
for (_, mon) in persisted_chan_data_0.iter() {
84+
assert_eq!(mon.get_latest_update_id(), $expected_update_id);
85+
}
86+
persisted_chan_data_1 = read_channel_monitors(store_1, nodes[1].keys_manager, nodes[1].keys_manager).unwrap();
87+
assert_eq!(persisted_chan_data_1.len(), 1);
88+
for (_, mon) in persisted_chan_data_1.iter() {
89+
assert_eq!(mon.get_latest_update_id(), $expected_update_id);
90+
}
91+
}
92+
}
93+
94+
// Create some initial channel and check that a channel was persisted.
95+
let _ = create_announced_chan_between_nodes(&nodes, 0, 1);
96+
check_persisted_data!(0);
97+
98+
// Send a few payments and make sure the monitors are updated to the latest.
99+
send_payment(&nodes[0], &vec!(&nodes[1])[..], 8000000);
100+
check_persisted_data!(5);
101+
send_payment(&nodes[1], &vec!(&nodes[0])[..], 4000000);
102+
check_persisted_data!(10);
103+
104+
// Force close because cooperative close doesn't result in any persisted
105+
// updates.
106+
nodes[0].node.force_close_broadcasting_latest_txn(&nodes[0].node.list_channels()[0].channel_id, &nodes[1].node.get_our_node_id()).unwrap();
107+
check_closed_event!(nodes[0], 1, ClosureReason::HolderForceClosed, [nodes[1].node.get_our_node_id()], 100000);
108+
check_closed_broadcast!(nodes[0], true);
109+
check_added_monitors!(nodes[0], 1);
110+
111+
let node_txn = nodes[0].tx_broadcaster.txn_broadcasted.lock().unwrap();
112+
assert_eq!(node_txn.len(), 1);
113+
114+
connect_block(&nodes[1], &create_dummy_block(nodes[0].best_block_hash(), 42, vec![node_txn[0].clone(), node_txn[0].clone()]));
115+
check_closed_broadcast!(nodes[1], true);
116+
check_closed_event!(nodes[1], 1, ClosureReason::CommitmentTxConfirmed, [nodes[0].node.get_our_node_id()], 100000);
117+
check_added_monitors!(nodes[1], 1);
118+
119+
// Make sure everything is persisted as expected after close.
120+
check_persisted_data!(CLOSED_CHANNEL_UPDATE_ID);
121+
}

0 commit comments

Comments
 (0)