| |
| |
| |
|
|
| |
| |
| |
|
|
| use futures_util::TryStreamExt; |
| use serde::{ser::Serializer, Serialize}; |
| use tauri::{command, ipc::Channel}; |
| use tokio::{ |
| fs::File, |
| io::{AsyncWriteExt, BufWriter}, |
| }; |
| use tokio_util::codec::{BytesCodec, FramedRead}; |
|
|
| use read_progress_stream::ReadProgressStream; |
|
|
| use std::time::Instant; |
| use std::{collections::HashMap, sync::Arc}; |
|
|
| type Result<T> = std::result::Result<T, Error>; |
|
|
| |
| pub struct TransferStats { |
| accumulated_chunk_len: usize, |
| accumulated_time: u128, |
| pub transfer_speed: u64, |
| pub total_transferred: u64, |
| start_time: Instant, |
| granularity: u32, |
| } |
|
|
| impl TransferStats { |
| |
| pub fn start(granularity: u32) -> Self { |
| Self { |
| accumulated_chunk_len: 0, |
| accumulated_time: 0, |
| transfer_speed: 0, |
| total_transferred: 0, |
| start_time: Instant::now(), |
| granularity, |
| } |
| } |
| |
| pub fn record_chunk_transfer(&mut self, chunk_len: usize) { |
| let now = Instant::now(); |
| let it_took = now.duration_since(self.start_time).as_millis(); |
| self.accumulated_chunk_len += chunk_len; |
| self.total_transferred += chunk_len as u64; |
| self.accumulated_time += it_took; |
|
|
| |
| if self.accumulated_time >= self.granularity as u128 { |
| self.transfer_speed = |
| (self.accumulated_chunk_len as u128 / self.accumulated_time * 1024) as u64; |
| self.accumulated_chunk_len = 0; |
| self.accumulated_time = 0; |
| } |
|
|
| |
| self.start_time = now; |
| } |
| } |
|
|
| |
| impl Default for TransferStats { |
| fn default() -> Self { |
| Self::start(500) |
| } |
| } |
|
|
| #[derive(Debug, thiserror::Error)] |
| pub enum Error { |
| #[error(transparent)] |
| Io(#[from] std::io::Error), |
| #[error(transparent)] |
| Request(#[from] reqwest::Error), |
| #[error("{0}")] |
| ContentLength(String), |
| #[error("request failed with status code {0}: {1}")] |
| HttpErrorCode(u16, String), |
| } |
|
|
| impl Serialize for Error { |
| fn serialize<S>(&self, serializer: S) -> std::result::Result<S::Ok, S::Error> |
| where |
| S: Serializer, |
| { |
| serializer.serialize_str(self.to_string().as_ref()) |
| } |
| } |
|
|
| #[derive(Clone, Serialize)] |
| #[serde(rename_all = "camelCase")] |
| pub struct ProgressPayload { |
| progress: u64, |
| total: u64, |
| transfer_speed: u64, |
| } |
|
|
| #[command] |
| pub async fn download_file( |
| url: &str, |
| file_path: &str, |
| headers: HashMap<String, String>, |
| body: Option<String>, |
| single_threaded: Option<bool>, |
| skip_ssl_verification: Option<bool>, |
| on_progress: Channel<ProgressPayload>, |
| ) -> Result<HashMap<String, String>> { |
| use futures::stream::{self, StreamExt}; |
| use std::cmp::min; |
| use tokio::io::AsyncSeekExt; |
|
|
| const PART_SIZE: u64 = 1024 * 1024; |
|
|
| let client = reqwest::ClientBuilder::new() |
| .danger_accept_invalid_certs(skip_ssl_verification.unwrap_or(false)) |
| .danger_accept_invalid_hostnames(skip_ssl_verification.unwrap_or(false)) |
| .build()?; |
| let force_single = single_threaded.unwrap_or(false); |
|
|
| async fn single_threaded_download( |
| client: &reqwest::Client, |
| url: &str, |
| file_path: &str, |
| headers: &HashMap<String, String>, |
| body: &Option<String>, |
| on_progress: Channel<ProgressPayload>, |
| ) -> Result<HashMap<String, String>> { |
| let mut request = if let Some(body) = body { |
| client.post(url).body(body.clone()) |
| } else { |
| client.get(url) |
| }; |
|
|
| for (key, value) in headers { |
| request = request.header(key, value); |
| } |
|
|
| let response = request.send().await?; |
| if !response.status().is_success() { |
| return Err(Error::HttpErrorCode( |
| response.status().as_u16(), |
| response.text().await.unwrap_or_default(), |
| )); |
| } |
|
|
| let mut resp_headers = HashMap::new(); |
| for (key, value) in response.headers().iter() { |
| if let Ok(val_str) = value.to_str() { |
| resp_headers.insert(key.to_string(), val_str.to_string()); |
| } |
| } |
|
|
| let total = response.content_length().unwrap_or(0); |
| let mut file = BufWriter::new(File::create(file_path).await?); |
| let mut stream = response.bytes_stream(); |
|
|
| let mut stats = TransferStats::default(); |
| while let Some(chunk) = stream.try_next().await? { |
| file.write_all(&chunk).await?; |
| stats.record_chunk_transfer(chunk.len()); |
| let _ = on_progress.send(ProgressPayload { |
| progress: stats.total_transferred, |
| total, |
| transfer_speed: stats.transfer_speed, |
| }); |
| } |
| file.flush().await?; |
|
|
| Ok(resp_headers) |
| } |
|
|
| if force_single { |
| return single_threaded_download(&client, url, file_path, &headers, &body, on_progress) |
| .await; |
| } |
|
|
| |
| let mut range_req = client.get(url).header("Range", "bytes=0-0"); |
| for (key, value) in headers.iter() { |
| range_req = range_req.header(key, value); |
| } |
| let range_resp = range_req.send().await?; |
| let accept_ranges = range_resp |
| .headers() |
| .get("accept-ranges") |
| .map(|v| v.to_str().unwrap_or("")) |
| .unwrap_or("") |
| .eq_ignore_ascii_case("bytes"); |
| let total = range_resp |
| .headers() |
| .get("content-range") |
| .and_then(|v| v.to_str().ok()) |
| .and_then(|s| s.split('/').nth(1)) |
| .and_then(|s| s.parse::<u64>().ok()) |
| .unwrap_or(0); |
|
|
| let mut resp_headers = HashMap::new(); |
| for (key, value) in range_resp.headers().iter() { |
| if let Ok(val_str) = value.to_str() { |
| resp_headers.insert(key.to_string(), val_str.to_string()); |
| } |
| } |
|
|
| if !accept_ranges || total == 0 { |
| return single_threaded_download(&client, url, file_path, &headers, &body, on_progress) |
| .await; |
| } |
|
|
| |
| let part_count = total.div_ceil(PART_SIZE); |
| let file = File::create(file_path).await?; |
| file.set_len(total).await?; |
|
|
| let file = Arc::new(tokio::sync::Mutex::new(file)); |
| let progress = Arc::new(tokio::sync::Mutex::new(TransferStats::default())); |
|
|
| stream::iter(0..part_count) |
| .for_each_concurrent(8, |i| { |
| let client = client.clone(); |
| let file = Arc::clone(&file); |
| let progress = Arc::clone(&progress); |
| let headers = headers.clone(); |
| let url = url.to_string(); |
| let on_progress = on_progress.clone(); |
|
|
| async move { |
| let start = i * PART_SIZE; |
| let end = min(start + PART_SIZE - 1, total - 1); |
| let range_header = format!("bytes={start}-{end}"); |
|
|
| let mut req = client.get(&url).header("Range", range_header); |
| for (key, value) in headers { |
| req = req.header(key, value); |
| } |
|
|
| let resp = match req.send().await { |
| Ok(r) => r, |
| Err(_) => return, |
| }; |
|
|
| if !resp.status().is_success() |
| && resp.status() != reqwest::StatusCode::PARTIAL_CONTENT |
| { |
| return; |
| } |
|
|
| let bytes = match resp.bytes().await { |
| Ok(b) => b, |
| Err(_) => return, |
| }; |
|
|
| { |
| let mut f = file.lock().await; |
| f.seek(std::io::SeekFrom::Start(start)).await.unwrap(); |
| f.write_all(&bytes).await.unwrap(); |
| } |
|
|
| { |
| let mut stat = progress.lock().await; |
| stat.record_chunk_transfer(bytes.len()); |
| let _ = on_progress.send(ProgressPayload { |
| progress: stat.total_transferred, |
| total, |
| transfer_speed: stat.transfer_speed, |
| }); |
| } |
| } |
| }) |
| .await; |
|
|
| Ok(resp_headers) |
| } |
|
|
| #[command] |
| pub async fn upload_file( |
| url: &str, |
| file_path: &str, |
| method: &str, |
| headers: HashMap<String, String>, |
| on_progress: Channel<ProgressPayload>, |
| ) -> Result<String> { |
| let file = File::open(file_path).await?; |
| let file_len = file.metadata().await.unwrap().len(); |
|
|
| let client = reqwest::Client::new(); |
| let mut request = match method.to_uppercase().as_str() { |
| "POST" => client.post(url), |
| "PUT" => client.put(url), |
| _ => return Err(Error::ContentLength("Invalid HTTP method".into())), |
| }; |
|
|
| request = request |
| .header(reqwest::header::CONTENT_LENGTH, file_len) |
| .body(file_to_body(on_progress.clone(), file, file_len)); |
|
|
| for (key, value) in headers { |
| request = request.header(&key, value); |
| } |
|
|
| let response = request.send().await?; |
| if response.status().is_success() { |
| response.text().await.map_err(Into::into) |
| } else { |
| Err(Error::HttpErrorCode( |
| response.status().as_u16(), |
| response.text().await.unwrap_or_default(), |
| )) |
| } |
| } |
|
|
| fn file_to_body(channel: Channel<ProgressPayload>, file: File, file_len: u64) -> reqwest::Body { |
| let stream = FramedRead::new(file, BytesCodec::new()).map_ok(|r| r.freeze()); |
|
|
| let mut stats = TransferStats::default(); |
| reqwest::Body::wrap_stream(ReadProgressStream::new( |
| stream, |
| Box::new(move |progress_chunk, _progress_total| { |
| stats.record_chunk_transfer(progress_chunk as usize); |
| let _ = channel.send(ProgressPayload { |
| progress: stats.total_transferred, |
| total: file_len, |
| transfer_speed: stats.transfer_speed, |
| }); |
| }), |
| )) |
| } |
|
|