|
17 | 17 |
|
18 | 18 | use std::time::Duration; |
19 | 19 |
|
| 20 | +use databend_common_base::base::tokio::time::sleep; |
20 | 21 | use databend_common_base::base::Stoppable; |
21 | | -// use databend_common_meta_client::ClientHandle; |
22 | | -// use databend_common_meta_client::MetaGrpcClient; |
| 22 | +use databend_common_meta_client::ClientHandle; |
| 23 | +use databend_common_meta_client::MetaGrpcClient; |
23 | 24 | use databend_common_meta_kvapi::kvapi::KVApi; |
24 | 25 | use databend_common_meta_kvapi::kvapi::UpsertKVReq; |
25 | 26 | use log::info; |
@@ -116,108 +117,120 @@ async fn test_kv_api_restart_cluster_write_read() -> anyhow::Result<()> { |
116 | 117 | Ok(()) |
117 | 118 | } |
118 | 119 |
|
119 | | -// FIXME: Disable this test until https://github.com/databendlabs/databend/pull/16704/#issuecomment-2442094481 addressed. |
120 | 120 | /// - Start a cluster of 3. |
121 | 121 | /// - Test upsert kv and read on different nodes. |
122 | 122 | /// - Stop and restart the cluster. |
123 | 123 | /// - Test read kv using same grpc client. |
124 | | -// #[test(harness = meta_service_test_harness)] |
125 | | -// #[fastrace::trace] |
126 | | -// async fn test_kv_api_restart_cluster_token_expired() -> anyhow::Result<()> { |
127 | | -// fn make_key(tc: &MetaSrvTestContext, k: impl std::fmt::Display) -> String { |
128 | | -// let x = &tc.config.raft_config; |
129 | | -// format!("t-restart-cluster-{}-{}-{}", x.config_id, x.id, k) |
130 | | -// } |
131 | | -// |
132 | | -// async fn test_write_read_on_every_node( |
133 | | -// tcs: &[MetaSrvTestContext], |
134 | | -// client: &ClientHandle, |
135 | | -// key_suffix: &str, |
136 | | -// ) -> anyhow::Result<()> { |
137 | | -// info!("--- test write on every node: {}", key_suffix); |
138 | | -// |
139 | | -// for (i, tc) in tcs.iter().enumerate() { |
140 | | -// let k = make_key(tc, key_suffix); |
141 | | -// if i == 0 { |
142 | | -// let res = client.upsert_kv(UpsertKVReq::update(&k, &b(&k))).await?; |
143 | | -// info!("--- upsert res: {:?}", res); |
144 | | -// } else { |
145 | | -// let client = tc.grpc_client().await.unwrap(); |
146 | | -// let res = client.upsert_kv(UpsertKVReq::update(&k, &b(&k))).await?; |
147 | | -// info!("--- upsert res: {:?}", res); |
148 | | -// } |
149 | | -// |
150 | | -// let res = client.get_kv(&k).await?; |
151 | | -// let res = res.unwrap(); |
152 | | -// |
153 | | -// assert_eq!(k.into_bytes(), res.data); |
154 | | -// } |
155 | | -// |
156 | | -// Ok(()) |
157 | | -// } |
158 | | -// |
159 | | -// let tcs = start_metasrv_cluster(&[0, 1, 2]).await?; |
160 | | -// let client = MetaGrpcClient::try_create( |
161 | | -// vec![tcs[0].config.grpc_api_address.clone()], |
162 | | -// "root", |
163 | | -// "xxx", |
164 | | -// None, |
165 | | -// Some(Duration::from_secs(10)), |
166 | | -// None, |
167 | | -// )?; |
168 | | -// |
169 | | -// info!("--- test write on a fresh cluster"); |
170 | | -// let key_suffix = "1st"; |
171 | | -// test_write_read_on_every_node(&tcs, &client, key_suffix).await?; |
172 | | -// |
173 | | -// info!("--- shutdown the cluster"); |
174 | | -// let stopped_tcs = { |
175 | | -// let mut stopped_tcs = vec![]; |
176 | | -// for mut tc in tcs { |
177 | | -// assert!(tc.meta_node.is_none()); |
178 | | -// |
179 | | -// let mut srv = tc.grpc_srv.take().unwrap(); |
180 | | -// srv.stop(None).await?; |
181 | | -// |
182 | | -// stopped_tcs.push(tc); |
183 | | -// } |
184 | | -// stopped_tcs |
185 | | -// }; |
186 | | -// |
187 | | -// info!("--- restart the cluster"); |
188 | | -// let tcs = { |
189 | | -// let mut tcs = vec![]; |
190 | | -// for mut tc in stopped_tcs { |
191 | | -// start_metasrv_with_context(&mut tc).await?; |
192 | | -// tcs.push(tc); |
193 | | -// } |
194 | | -// |
195 | | -// for tc in &tcs { |
196 | | -// info!("--- wait until a leader is observed"); |
197 | | -// // Every tcs[i] contains one meta node in this context. |
198 | | -// let g = tc.grpc_srv.as_ref().unwrap(); |
199 | | -// let meta_node = g.get_meta_node(); |
200 | | -// let metrics = meta_node |
201 | | -// .raft |
202 | | -// .wait(timeout()) |
203 | | -// .metrics(|m| m.current_leader.is_some(), "a leader is observed") |
204 | | -// .await?; |
205 | | -// |
206 | | -// info!("got leader, metrics: {:?}", metrics); |
207 | | -// } |
208 | | -// tcs |
209 | | -// }; |
210 | | -// |
211 | | -// info!("--- read use old client"); |
212 | | -// let tc = &tcs[0]; |
213 | | -// let k = make_key(tc, key_suffix); |
214 | | -// let res = client.get_kv(&k).await?; |
215 | | -// let res = res.unwrap(); |
216 | | -// |
217 | | -// assert_eq!(b(k), res.data); |
218 | | -// |
219 | | -// Ok(()) |
220 | | -// } |
| 124 | +#[test(harness = meta_service_test_harness)] |
| 125 | +#[fastrace::trace] |
| 126 | +async fn test_kv_api_restart_cluster_token_expired() -> anyhow::Result<()> { |
| 127 | + fn make_key(tc: &MetaSrvTestContext, k: impl std::fmt::Display) -> String { |
| 128 | + let x = &tc.config.raft_config; |
| 129 | + format!("t-restart-cluster-{}-{}-{}", x.config_id, x.id, k) |
| 130 | + } |
| 131 | + |
| 132 | + async fn test_write_read_on_every_node( |
| 133 | + tcs: &[MetaSrvTestContext], |
| 134 | + client: &ClientHandle, |
| 135 | + key_suffix: &str, |
| 136 | + ) -> anyhow::Result<()> { |
| 137 | + info!("--- test write on every node: {}", key_suffix); |
| 138 | + |
| 139 | + for (i, tc) in tcs.iter().enumerate() { |
| 140 | + let k = make_key(tc, key_suffix); |
| 141 | + if i == 0 { |
| 142 | + let res = client.upsert_kv(UpsertKVReq::update(&k, &b(&k))).await?; |
| 143 | + info!("--- upsert res: {:?}", res); |
| 144 | + } else { |
| 145 | + let client = tc.grpc_client().await.unwrap(); |
| 146 | + let res = client.upsert_kv(UpsertKVReq::update(&k, &b(&k))).await?; |
| 147 | + info!("--- upsert res: {:?}", res); |
| 148 | + } |
| 149 | + |
| 150 | + let res = client.get_kv(&k).await?; |
| 151 | + let res = res.unwrap(); |
| 152 | + |
| 153 | + assert_eq!(k.into_bytes(), res.data); |
| 154 | + } |
| 155 | + |
| 156 | + Ok(()) |
| 157 | + } |
| 158 | + |
| 159 | + let tcs = start_metasrv_cluster(&[0, 1, 2]).await?; |
| 160 | + let client = MetaGrpcClient::try_create( |
| 161 | + vec![tcs[0].config.grpc_api_address.clone()], |
| 162 | + "root", |
| 163 | + "xxx", |
| 164 | + // Without timeout, the client will not be able to reconnect. |
| 165 | + // This is an issue of the http client. |
| 166 | + Some(Duration::from_secs(1)), |
| 167 | + Some(Duration::from_secs(10)), |
| 168 | + None, |
| 169 | + )?; |
| 170 | + |
| 171 | + info!("--- test write on a fresh cluster"); |
| 172 | + let key_suffix = "1st"; |
| 173 | + test_write_read_on_every_node(&tcs, &client, key_suffix).await?; |
| 174 | + |
| 175 | + info!("--- shutdown the cluster"); |
| 176 | + let stopped_tcs = { |
| 177 | + let mut stopped_tcs = vec![]; |
| 178 | + for mut tc in tcs { |
| 179 | + assert!(tc.meta_node.is_none()); |
| 180 | + |
| 181 | + let mut srv = tc.grpc_srv.take().unwrap(); |
| 182 | + srv.stop(None).await?; |
| 183 | + |
| 184 | + stopped_tcs.push(tc); |
| 185 | + } |
| 186 | + stopped_tcs |
| 187 | + }; |
| 188 | + |
| 189 | + info!("--- restart the cluster"); |
| 190 | + let tcs = { |
| 191 | + let mut tcs = vec![]; |
| 192 | + for mut tc in stopped_tcs { |
| 193 | + info!( |
| 194 | + "--- starting metasrv: {:?}", |
| 195 | + tc.config.raft_config.raft_api_addr().await? |
| 196 | + ); |
| 197 | + start_metasrv_with_context(&mut tc).await?; |
| 198 | + |
| 199 | + info!( |
| 200 | + "--- started metasrv: {:?}", |
| 201 | + tc.config.raft_config.raft_api_addr().await? |
| 202 | + ); |
| 203 | + |
| 204 | + // sleep(Duration::from_secs(3)).await; |
| 205 | + tcs.push(tc); |
| 206 | + } |
| 207 | + |
| 208 | + for tc in &tcs { |
| 209 | + info!("--- wait until a leader is observed"); |
| 210 | + // Every tcs[i] contains one meta node in this context. |
| 211 | + let g = tc.grpc_srv.as_ref().unwrap(); |
| 212 | + let meta_node = g.get_meta_node(); |
| 213 | + let metrics = meta_node |
| 214 | + .raft |
| 215 | + .wait(timeout()) |
| 216 | + .metrics(|m| m.current_leader.is_some(), "a leader is observed") |
| 217 | + .await?; |
| 218 | + |
| 219 | + info!("got leader, metrics: {:?}", metrics); |
| 220 | + } |
| 221 | + tcs |
| 222 | + }; |
| 223 | + |
| 224 | + info!("--- read use old client"); |
| 225 | + let tc = &tcs[0]; |
| 226 | + let k = make_key(tc, key_suffix); |
| 227 | + let res = client.get_kv(&k).await?; |
| 228 | + let res = res.unwrap(); |
| 229 | + |
| 230 | + assert_eq!(b(k), res.data); |
| 231 | + |
| 232 | + Ok(()) |
| 233 | +} |
221 | 234 |
|
222 | 235 | // Election timeout is 8~12 sec. |
223 | 236 | // A raft node waits for a interval of election timeout before starting election |
|
0 commit comments