#![allow(clippy::needless_return)] // tokio macro-generated code doesn't respect this #![feature(trivial_bounds)] #![feature(min_specialization)] #![feature(iter_advance_by)] #![feature(io_error_more)] #![feature(round_char_boundary)] #![feature(arbitrary_self_types)] #![feature(arbitrary_self_types_pointers)] #![allow(clippy::mutable_key_type)] pub mod attach; pub mod embed; pub mod glob; mod globset; pub mod invalidation; mod invalidator_map; pub mod json; mod mutex_map; mod read_glob; mod retry; pub mod rope; pub mod source_context; pub mod util; pub(crate) mod virtual_fs; mod watcher; use std::{ borrow::Cow, cmp::{Ordering, min}, fmt::{self, Debug, Display, Formatter}, fs::FileType, future::Future, io::{self, BufRead, BufReader, ErrorKind, Read}, mem::take, path::{MAIN_SEPARATOR, Path, PathBuf}, sync::{Arc, LazyLock}, time::Duration, }; use anyhow::{Context, Result, anyhow, bail}; use auto_hash_map::{AutoMap, AutoSet}; use bitflags::bitflags; use dunce::simplified; use glob::Glob; use indexmap::IndexSet; use invalidator_map::InvalidatorMap; use jsonc_parser::{ParseOptions, parse_to_serde_value}; use mime::Mime; use rayon::iter::{IntoParallelIterator, ParallelIterator}; pub use read_glob::ReadGlobResult; use read_glob::{read_glob, track_glob}; use rustc_hash::FxHashSet; use serde::{Deserialize, Serialize}; use serde_json::Value; use tokio::sync::{RwLock, RwLockReadGuard}; use tracing::Instrument; use turbo_rcstr::{RcStr, rcstr}; use turbo_tasks::{ ApplyEffectsContext, Completion, InvalidationReason, Invalidator, NonLocalValue, ReadRef, ResolvedVc, TaskInput, ValueToString, Vc, debug::ValueDebugFormat, effect, mark_session_dependent, mark_stateful, trace::TraceRawVcs, }; use turbo_tasks_hash::{DeterministicHash, DeterministicHasher, hash_xxh3_hash64}; use util::{extract_disk_access, join_path, normalize_path, sys_to_unix, unix_to_sys}; pub use virtual_fs::VirtualFileSystem; use watcher::DiskWatcher; use self::{invalidation::Write, json::UnparsableJson, mutex_map::MutexMap}; use crate::{ attach::AttachedFileSystem, invalidator_map::WriteContent, retry::retry_blocking, rope::{Rope, RopeReader}, }; /// A (somewhat arbitrary) filename limit that we should try to keep output file names below. /// /// For the sake of consistency, this is a fixed constant that is likely to be safe across all /// platforms. /// /// Different operating systems have different limits on file name and file path. See /// [`validate_path_length`] for details. Because this only accounts for a single path segment, and /// not the total path length, this cannot not guarantee a full file path is safe. /// /// To ensure file names are kept within this limit, call /// [`FileSystemPath::truncate_file_name_with_hash`]. pub const MAX_SAFE_FILE_NAME_LENGTH: usize = 200; /// Validate the path, returning the valid path, a modified-but-now-valid path, or bailing with an /// error. /// /// The behaviour of the file system changes depending on the OS, and indeed sometimes the FS /// implementation of the OS itself. /// /// - On Windows the limit for normal file paths is 260 characters, a holdover from the DOS days, /// but Rust will opportunistically rewrite paths to 'UNC' paths for supported path operations /// which can be up to 32767 characters long. /// - On macOS, the limit is traditionally 255 characters for the file name and a second limit of /// 1024 for the entire path (verified by running `getconf PATH_MAX /`). /// - On Linux, the limit differs between kernel (and by extension, distro) and filesystem. On most /// common file systems (e.g. ext4, btrfs, and xfs), individual file names can be up to 255 bytes /// with no hard limit on total path length. [Some legacy POSIX APIs are restricted to the /// `PATH_MAX` value of 4096 bytes in `limits.h`, but most applications support longer /// paths][PATH_MAX]. /// /// For more details, refer to . /// /// Realistically, the output path lengths will be the same across all platforms, so we need to set /// a conservative limit and be particular about when we decide to bump it. Here we have opted for /// 255 characters, because it is the shortest of the three options. /// /// [PATH_MAX]: https://eklitzke.org/path-max-is-tricky pub fn validate_path_length(path: &Path) -> Result> { /// Here we check if the path is too long for windows, and if so, attempt to canonicalize it /// to a UNC path. #[cfg(target_family = "windows")] fn validate_path_length_inner(path: &Path) -> Result> { const MAX_PATH_LENGTH_WINDOWS: usize = 260; const UNC_PREFIX: &str = "\\\\?\\"; if path.starts_with(UNC_PREFIX) { return Ok(path.into()); } if path.as_os_str().len() > MAX_PATH_LENGTH_WINDOWS { let new_path = std::fs::canonicalize(path) .map_err(|_| anyhow!("file is too long, and could not be normalized"))?; return Ok(new_path.into()); } Ok(path.into()) } /// Here we are only going to check if the total length exceeds, or the last segment exceeds. /// This heuristic is primarily to avoid long file names, and it makes the operation much /// cheaper. #[cfg(not(target_family = "windows"))] fn validate_path_length_inner(path: &Path) -> Result> { const MAX_FILE_NAME_LENGTH_UNIX: usize = 255; // macOS reports a limit of 1024, but I (@arlyon) have had issues with paths above 1016 // so we subtract a bit to be safe. on most linux distros this is likely a lot larger than // 1024, but macOS is *special* const MAX_PATH_LENGTH: usize = 1024 - 8; // check the last segment (file name) if path .file_name() .map(|n| n.as_encoded_bytes().len()) .unwrap_or(0) > MAX_FILE_NAME_LENGTH_UNIX { anyhow::bail!( "file name is too long (exceeds {} bytes)", MAX_FILE_NAME_LENGTH_UNIX ); } if path.as_os_str().len() > MAX_PATH_LENGTH { anyhow::bail!("path is too long (exceeds {} bytes)", MAX_PATH_LENGTH); } Ok(path.into()) } validate_path_length_inner(path).with_context(|| { format!( "path length for file {} exceeds max length of filesystem", path.to_string_lossy() ) }) } trait ConcurrencyLimitedExt { type Output; async fn concurrency_limited(self, semaphore: &tokio::sync::Semaphore) -> Self::Output; } impl ConcurrencyLimitedExt for F where F: Future, { type Output = R; async fn concurrency_limited(self, semaphore: &tokio::sync::Semaphore) -> Self::Output { let _permit = semaphore.acquire().await; self.await } } fn create_semaphore() -> tokio::sync::Semaphore { tokio::sync::Semaphore::new(256) } #[turbo_tasks::value_trait] pub trait FileSystem: ValueToString { /// Returns the path to the root of the file system. #[turbo_tasks::function] fn root(self: ResolvedVc) -> Vc { FileSystemPath::new_normalized(self, RcStr::default()).cell() } #[turbo_tasks::function] fn read(self: Vc, fs_path: FileSystemPath) -> Vc; #[turbo_tasks::function] fn read_link(self: Vc, fs_path: FileSystemPath) -> Vc; #[turbo_tasks::function] fn raw_read_dir(self: Vc, fs_path: FileSystemPath) -> Vc; #[turbo_tasks::function] fn write(self: Vc, fs_path: FileSystemPath, content: Vc) -> Vc<()>; #[turbo_tasks::function] fn write_link(self: Vc, fs_path: FileSystemPath, target: Vc) -> Vc<()>; #[turbo_tasks::function] fn metadata(self: Vc, fs_path: FileSystemPath) -> Vc; } #[derive(Default)] struct DiskFileSystemApplyContext { /// A cache of already created directories to avoid creating them multiple times. created_directories: FxHashSet, } #[derive(Serialize, Deserialize, TraceRawVcs, ValueDebugFormat, NonLocalValue)] struct DiskFileSystemInner { pub name: RcStr, pub root: RcStr, #[turbo_tasks(debug_ignore, trace_ignore)] #[serde(skip)] mutex_map: MutexMap, #[turbo_tasks(debug_ignore, trace_ignore)] #[serde(skip)] invalidator_map: InvalidatorMap, #[turbo_tasks(debug_ignore, trace_ignore)] #[serde(skip)] dir_invalidator_map: InvalidatorMap, /// Lock that makes invalidation atomic. It will keep a write lock during /// watcher invalidation and a read lock during other operations. #[turbo_tasks(debug_ignore, trace_ignore)] #[serde(skip)] invalidation_lock: RwLock<()>, /// Semaphore to limit the maximum number of concurrent file operations. #[turbo_tasks(debug_ignore, trace_ignore)] #[serde(skip, default = "create_semaphore")] semaphore: tokio::sync::Semaphore, #[turbo_tasks(debug_ignore, trace_ignore)] watcher: DiskWatcher, } impl DiskFileSystemInner { /// Returns the root as Path fn root_path(&self) -> &Path { simplified(Path::new(&*self.root)) } /// registers the path as an invalidator for the current task, /// has to be called within a turbo-tasks function fn register_read_invalidator(&self, path: &Path) -> Result<()> { let invalidator = turbo_tasks::get_invalidator(); self.invalidator_map .insert(path_to_key(path), invalidator, None); #[cfg(not(any(target_os = "macos", target_os = "windows")))] if let Some(dir) = path.parent() { self.watcher.ensure_watching(dir, self.root_path())?; } Ok(()) } /// registers the path as an invalidator for the current task, /// has to be called within a turbo-tasks function. It removes and returns /// the current list of invalidators. fn register_write_invalidator( &self, path: &Path, invalidator: Invalidator, write_content: WriteContent, ) -> Result)>> { let mut invalidator_map = self.invalidator_map.lock().unwrap(); let invalidators = invalidator_map.entry(path_to_key(path)).or_default(); let old_invalidators = invalidators .extract_if(|i, old_write_content| { i == &invalidator || old_write_content .as_ref() .is_none_or(|old| old != &write_content) }) .filter(|(i, _)| i != &invalidator) .collect::>(); invalidators.insert(invalidator, Some(write_content)); drop(invalidator_map); #[cfg(not(any(target_os = "macos", target_os = "windows")))] if let Some(dir) = path.parent() { self.watcher.ensure_watching(dir, self.root_path())?; } Ok(old_invalidators) } /// registers the path as an invalidator for the current task, /// has to be called within a turbo-tasks function fn register_dir_invalidator(&self, path: &Path) -> Result<()> { let invalidator = turbo_tasks::get_invalidator(); self.dir_invalidator_map .insert(path_to_key(path), invalidator, None); #[cfg(not(any(target_os = "macos", target_os = "windows")))] self.watcher.ensure_watching(path, self.root_path())?; Ok(()) } async fn lock_path(&self, full_path: &Path) -> PathLockGuard<'_> { let lock1 = self.invalidation_lock.read().await; let lock2 = self.mutex_map.lock(full_path.to_path_buf()).await; PathLockGuard(lock1, lock2) } fn invalidate(&self) { let _span = tracing::info_span!("invalidate filesystem", name = &*self.root).entered(); let span = tracing::Span::current(); let handle = tokio::runtime::Handle::current(); let invalidator_map = take(&mut *self.invalidator_map.lock().unwrap()); let dir_invalidator_map = take(&mut *self.dir_invalidator_map.lock().unwrap()); let iter = invalidator_map .into_par_iter() .chain(dir_invalidator_map.into_par_iter()) .flat_map(|(_, invalidators)| invalidators.into_par_iter()); iter.for_each(|(i, _)| { let _span = span.clone().entered(); let _guard = handle.enter(); i.invalidate() }); } /// Invalidates every tracked file in the filesystem. /// /// Calls the given fn invalidate_with_reason( &self, reason: impl Fn(String) -> R + Sync, ) { let _span = tracing::info_span!("invalidate filesystem", name = &*self.root).entered(); let span = tracing::Span::current(); let handle = tokio::runtime::Handle::current(); let invalidator_map = take(&mut *self.invalidator_map.lock().unwrap()); let dir_invalidator_map = take(&mut *self.dir_invalidator_map.lock().unwrap()); let iter = invalidator_map .into_par_iter() .chain(dir_invalidator_map.into_par_iter()) .flat_map(|(path, invalidators)| { let _span = span.clone().entered(); let reason_for_path = reason(path); invalidators .into_par_iter() .map(move |i| (reason_for_path.clone(), i)) }); iter.for_each(|(reason, (invalidator, _))| { let _span = span.clone().entered(); let _guard = handle.enter(); invalidator.invalidate_with_reason(reason) }); } fn invalidate_from_write( &self, full_path: &Path, invalidators: Vec<(Invalidator, Option)>, ) { if !invalidators.is_empty() { if let Some(path) = format_absolute_fs_path(full_path, &self.name, self.root_path()) { if invalidators.len() == 1 { let (invalidator, _) = invalidators.into_iter().next().unwrap(); invalidator.invalidate_with_reason(Write { path }); } else { invalidators.into_iter().for_each(|(invalidator, _)| { invalidator.invalidate_with_reason(Write { path: path.clone() }); }); } } else { invalidators.into_iter().for_each(|(invalidator, _)| { invalidator.invalidate(); }); } } } #[tracing::instrument(level = "info", name = "start filesystem watching", skip_all, fields(path = %self.root))] async fn start_watching_internal( self: &Arc, report_invalidation_reason: bool, poll_interval: Option, ) -> Result<()> { let root_path = self.root_path().to_path_buf(); // create the directory for the filesystem on disk, if it doesn't exist retry_blocking(root_path.clone(), move |path| { let _tracing = tracing::info_span!("create root directory", name = display(path.display())) .entered(); std::fs::create_dir_all(path) }) .concurrency_limited(&self.semaphore) .await?; self.watcher .start_watching(self.clone(), report_invalidation_reason, poll_interval)?; Ok(()) } async fn create_directory(self: &Arc, directory: &Path) -> Result<()> { let already_created = ApplyEffectsContext::with_or_insert_with( DiskFileSystemApplyContext::default, |fs_context| fs_context.created_directories.contains(directory), ); if !already_created { let func = |p: &Path| std::fs::create_dir_all(p); retry_blocking(directory.to_path_buf(), func) .concurrency_limited(&self.semaphore) .instrument(tracing::info_span!( "create directory", name = display(directory.display()) )) .await?; ApplyEffectsContext::with(|fs_context: &mut DiskFileSystemApplyContext| { fs_context .created_directories .insert(directory.to_path_buf()) }); } Ok(()) } } #[turbo_tasks::value(cell = "new", eq = "manual")] pub struct DiskFileSystem { inner: Arc, } impl DiskFileSystem { pub fn name(&self) -> &RcStr { &self.inner.name } pub fn root(&self) -> &RcStr { &self.inner.root } pub fn invalidate(&self) { self.inner.invalidate(); } pub fn invalidate_with_reason( &self, reason: impl Fn(String) -> R + Sync, ) { self.inner.invalidate_with_reason(reason); } pub async fn start_watching(&self, poll_interval: Option) -> Result<()> { self.inner .start_watching_internal(false, poll_interval) .await } pub async fn start_watching_with_invalidation_reason( &self, poll_interval: Option, ) -> Result<()> { self.inner .start_watching_internal(true, poll_interval) .await } pub fn stop_watching(&self) { self.inner.watcher.stop_watching(); } pub async fn to_sys_path(&self, fs_path: FileSystemPath) -> Result { // just in case there's a windows unc path prefix we remove it with `dunce` let path = self.inner.root_path(); Ok(if fs_path.path.is_empty() { path.to_path_buf() } else { path.join(&*unix_to_sys(&fs_path.path)) }) } } #[allow(dead_code, reason = "we need to hold onto the locks")] struct PathLockGuard<'a>( #[allow(dead_code)] RwLockReadGuard<'a, ()>, #[allow(dead_code)] mutex_map::MutexMapGuard<'a, PathBuf>, ); fn format_absolute_fs_path(path: &Path, name: &str, root_path: &Path) -> Option { if let Ok(rel_path) = path.strip_prefix(root_path) { let path = if MAIN_SEPARATOR != '/' { let rel_path = rel_path.to_string_lossy().replace(MAIN_SEPARATOR, "/"); format!("[{name}]/{rel_path}") } else { format!("[{name}]/{}", rel_path.display()) }; Some(path) } else { None } } pub fn path_to_key(path: impl AsRef) -> String { path.as_ref().to_string_lossy().to_string() } #[turbo_tasks::value_impl] impl DiskFileSystem { /// Create a new instance of `DiskFileSystem`. /// # Arguments /// /// * `name` - Name of the filesystem. /// * `root` - Path to the given filesystem's root. Should be /// [canonicalized][std::fs::canonicalize]. /// * `ignored_subpaths` - A list of subpaths that should not trigger invalidation. This should /// be a full path, since it is possible that root & project dir is different and requires to /// ignore specific subpaths from each. #[turbo_tasks::function] pub fn new(name: RcStr, root: RcStr, ignored_subpaths: Vec) -> Result> { mark_stateful(); let instance = DiskFileSystem { inner: Arc::new(DiskFileSystemInner { name, root, mutex_map: Default::default(), invalidation_lock: Default::default(), invalidator_map: InvalidatorMap::new(), dir_invalidator_map: InvalidatorMap::new(), semaphore: create_semaphore(), watcher: DiskWatcher::new( ignored_subpaths.into_iter().map(PathBuf::from).collect(), ), }), }; Ok(Self::cell(instance)) } } impl Debug for DiskFileSystem { fn fmt(&self, f: &mut Formatter) -> fmt::Result { write!(f, "name: {}, root: {}", self.inner.name, self.inner.root) } } #[turbo_tasks::value_impl] impl FileSystem for DiskFileSystem { #[turbo_tasks::function(fs)] async fn read(&self, fs_path: FileSystemPath) -> Result> { mark_session_dependent(); let full_path = self.to_sys_path(fs_path).await?; self.inner.register_read_invalidator(&full_path)?; let _lock = self.inner.lock_path(&full_path).await; let content = match retry_blocking(full_path.clone(), |path: &Path| File::from_path(path)) .concurrency_limited(&self.inner.semaphore) .instrument(tracing::info_span!( "read file", name = display(full_path.display()) )) .await { Ok(file) => FileContent::new(file), Err(e) if e.kind() == ErrorKind::NotFound || e.kind() == ErrorKind::InvalidFilename => { FileContent::NotFound } Err(e) => { bail!(anyhow!(e).context(format!("reading file {}", full_path.display()))) } }; Ok(content.cell()) } #[turbo_tasks::function(fs)] async fn raw_read_dir(&self, fs_path: FileSystemPath) -> Result> { mark_session_dependent(); let full_path = self.to_sys_path(fs_path).await?; self.inner.register_dir_invalidator(&full_path)?; // we use the sync std function here as it's a lot faster (600%) in // node-file-trace let read_dir = match retry_blocking(full_path.clone(), |path| { let _span = tracing::info_span!("read directory", name = display(path.display())).entered(); std::fs::read_dir(path) }) .concurrency_limited(&self.inner.semaphore) .await { Ok(dir) => dir, Err(e) if e.kind() == ErrorKind::NotFound || e.kind() == ErrorKind::NotADirectory || e.kind() == ErrorKind::InvalidFilename => { return Ok(RawDirectoryContent::not_found()); } Err(e) => { bail!(anyhow!(e).context(format!("reading dir {}", full_path.display()))) } }; let entries = read_dir .filter_map(|r| { let e = match r { Ok(e) => e, Err(err) => return Some(Err(err.into())), }; // we filter out any non unicode names let file_name = e.file_name().to_str()?.into(); let entry = match e.file_type() { Ok(t) if t.is_file() => RawDirectoryEntry::File, Ok(t) if t.is_dir() => RawDirectoryEntry::Directory, Ok(t) if t.is_symlink() => RawDirectoryEntry::Symlink, Ok(_) => RawDirectoryEntry::Other, Err(err) => return Some(Err(err.into())), }; Some(anyhow::Ok((file_name, entry))) }) .collect::>() .with_context(|| format!("reading directory item in {}", full_path.display()))?; Ok(RawDirectoryContent::new(entries)) } #[turbo_tasks::function(fs)] async fn read_link(&self, fs_path: FileSystemPath) -> Result> { mark_session_dependent(); let full_path = self.to_sys_path(fs_path.clone()).await?; self.inner.register_read_invalidator(&full_path)?; let _lock = self.inner.lock_path(&full_path).await; let link_path = match retry_blocking(full_path.clone(), |path: &Path| std::fs::read_link(path)) .concurrency_limited(&self.inner.semaphore) .instrument(tracing::info_span!( "read symlink", name = display(full_path.display()) )) .await { Ok(res) => res, Err(_) => return Ok(LinkContent::NotFound.cell()), }; let is_link_absolute = link_path.is_absolute(); let mut file = link_path.clone(); if !is_link_absolute { if let Some(normalized_linked_path) = full_path.parent().and_then(|p| { normalize_path(&sys_to_unix(p.join(&file).to_string_lossy().as_ref())) }) { #[cfg(target_family = "windows")] { file = PathBuf::from(normalized_linked_path); } // `normalize_path` stripped the leading `/` of the path // add it back here or the `strip_prefix` will return `Err` #[cfg(not(target_family = "windows"))] { file = PathBuf::from(format!("/{normalized_linked_path}")); } } else { return Ok(LinkContent::Invalid.cell()); } } // strip the root from the path, it serves two purpose // 1. ensure the linked path is under the root // 2. strip the root path if the linked path is absolute // // we use `dunce::simplify` to strip a potential UNC prefix on windows, on any // other OS this gets compiled away let result = simplified(&file).strip_prefix(simplified(Path::new(&self.inner.root))); let relative_to_root_path = match result { Ok(file) => PathBuf::from(sys_to_unix(&file.to_string_lossy()).as_ref()), Err(_) => return Ok(LinkContent::Invalid.cell()), }; let (target, file_type) = if is_link_absolute { let target_string: RcStr = relative_to_root_path.to_string_lossy().into(); ( target_string.clone(), FileSystemPath::new_normalized(fs_path.fs().to_resolved().await?, target_string) .get_type() .await?, ) } else { let link_path_string_cow = link_path.to_string_lossy(); let link_path_unix: RcStr = sys_to_unix(&link_path_string_cow).into(); ( link_path_unix.clone(), fs_path.parent().join(&link_path_unix)?.get_type().await?, ) }; Ok(LinkContent::Link { target, link_type: { let mut link_type = Default::default(); if link_path.is_absolute() { link_type |= LinkType::ABSOLUTE; } if matches!(&*file_type, FileSystemEntryType::Directory) { link_type |= LinkType::DIRECTORY; } link_type }, } .cell()) } #[turbo_tasks::function(fs)] async fn write(&self, fs_path: FileSystemPath, content: Vc) -> Result<()> { mark_session_dependent(); let full_path = self.to_sys_path(fs_path).await?; let content = content.await?; let inner = self.inner.clone(); let invalidator = turbo_tasks::get_invalidator(); effect(async move { let full_path = validate_path_length(&full_path)?; let _lock = inner.lock_path(&full_path).await; // Track the file, so that we will rewrite it if it ever changes. let old_invalidators = inner.register_write_invalidator( &full_path, invalidator, WriteContent::File(content.clone()), )?; // We perform an untracked comparison here, so that this write is not dependent // on a read's Vc (and the memory it holds). Our untracked read can // be freed immediately. Given this is an output file, it's unlikely any Turbo // code will need to read the file from disk into a Vc, so we're // not wasting cycles. let compare = content .streaming_compare(&full_path) .concurrency_limited(&inner.semaphore) .instrument(tracing::info_span!( "read file before write", name = display(full_path.display()) )) .await?; if compare == FileComparison::Equal { if !old_invalidators.is_empty() { let key = path_to_key(&full_path); for (invalidator, write_content) in old_invalidators { inner .invalidator_map .insert(key.clone(), invalidator, write_content); } } return Ok(()); } match &*content { FileContent::Content(..) => { let create_directory = compare == FileComparison::Create; if create_directory && let Some(parent) = full_path.parent() { inner.create_directory(parent).await.with_context(|| { format!( "failed to create directory {} for write to {}", parent.display(), full_path.display() ) })?; } let full_path_to_write = full_path.clone(); let content = content.clone(); retry_blocking(full_path_to_write.into_owned(), move |full_path| { use std::io::Write; let mut f = std::fs::File::create(full_path)?; let FileContent::Content(file) = &*content else { unreachable!() }; std::io::copy(&mut file.read(), &mut f)?; #[cfg(target_family = "unix")] f.set_permissions(file.meta.permissions.into())?; f.flush()?; static WRITE_VERSION: LazyLock = LazyLock::new(|| { std::env::var_os("TURBO_ENGINE_WRITE_VERSION") .is_some_and(|v| v == "1" || v == "true") }); if *WRITE_VERSION { let mut full_path = full_path.to_owned(); let hash = hash_xxh3_hash64(file); let ext = full_path.extension(); let ext = if let Some(ext) = ext { format!("{:016x}.{}", hash, ext.to_string_lossy()) } else { format!("{hash:016x}") }; full_path.set_extension(ext); let mut f = std::fs::File::create(&full_path)?; std::io::copy(&mut file.read(), &mut f)?; #[cfg(target_family = "unix")] f.set_permissions(file.meta.permissions.into())?; f.flush()?; } Ok::<(), io::Error>(()) }) .concurrency_limited(&inner.semaphore) .instrument(tracing::info_span!( "write file", name = display(full_path.display()) )) .await .with_context(|| format!("failed to write to {}", full_path.display()))?; } FileContent::NotFound => { retry_blocking(full_path.clone().into_owned(), |path| { std::fs::remove_file(path) }) .concurrency_limited(&inner.semaphore) .instrument(tracing::info_span!( "remove file", name = display(full_path.display()) )) .await .or_else(|err| { if err.kind() == ErrorKind::NotFound { Ok(()) } else { Err(err) } }) .with_context(|| anyhow!("removing {} failed", full_path.display()))?; } } inner.invalidate_from_write(&full_path, old_invalidators); Ok(()) }); Ok(()) } #[turbo_tasks::function(fs)] async fn write_link(&self, fs_path: FileSystemPath, target: Vc) -> Result<()> { mark_session_dependent(); let full_path = self.to_sys_path(fs_path).await?; let content = target.await?; let inner = self.inner.clone(); let invalidator = turbo_tasks::get_invalidator(); effect(async move { let full_path = validate_path_length(&full_path)?; let _lock = inner.lock_path(&full_path).await; let old_invalidators = inner.register_write_invalidator( &full_path, invalidator, WriteContent::Link(content.clone()), )?; // TODO(sokra) preform a untracked read here, register an invalidator and get // all existing invalidators let old_content = match retry_blocking(full_path.clone().into_owned(), |path| { std::fs::read_link(path) }) .concurrency_limited(&inner.semaphore) .instrument(tracing::info_span!( "read symlink before write", name = display(full_path.display()) )) .await { Ok(res) => Some((res.is_absolute(), res)), Err(_) => None, }; let is_equal = match (&*content, &old_content) { (LinkContent::Link { target, link_type }, Some((old_is_absolute, old_target))) => { Path::new(&**target) == old_target && link_type.contains(LinkType::ABSOLUTE) == *old_is_absolute } (LinkContent::NotFound, None) => true, _ => false, }; if is_equal { if !old_invalidators.is_empty() { let key = path_to_key(&full_path); for (invalidator, write_content) in old_invalidators { inner .invalidator_map .insert(key.clone(), invalidator, write_content); } } return Ok(()); } match &*content { LinkContent::Link { target, link_type } => { let create_directory = old_content.is_none(); if create_directory && let Some(parent) = full_path.parent() { inner.create_directory(parent).await.with_context(|| { format!( "failed to create directory {} for write link to {}", parent.display(), full_path.display() ) })?; } let link_type = *link_type; let target_path = if link_type.contains(LinkType::ABSOLUTE) { Path::new(&inner.root).join(unix_to_sys(target).as_ref()) } else { PathBuf::from(unix_to_sys(target).as_ref()) }; let full_path = full_path.into_owned(); retry_blocking(target_path, move |target_path| { let _span = tracing::info_span!( "write symlink", name = display(target_path.display()) ) .entered(); // we use the sync std method here because `symlink` is fast // if we put it into a task, it will be slower #[cfg(not(target_family = "windows"))] { std::os::unix::fs::symlink(target_path, &full_path) } #[cfg(target_family = "windows")] { if link_type.contains(LinkType::DIRECTORY) { std::os::windows::fs::symlink_dir(target_path, &full_path) } else { std::os::windows::fs::symlink_file(target_path, &full_path) } } }) .await .with_context(|| format!("create symlink to {target}"))?; } LinkContent::Invalid => { anyhow::bail!("invalid symlink target: {}", full_path.display()) } LinkContent::NotFound => { retry_blocking(full_path.clone().into_owned(), |path| { std::fs::remove_file(path) }) .concurrency_limited(&inner.semaphore) .await .or_else(|err| { if err.kind() == ErrorKind::NotFound { Ok(()) } else { Err(err) } }) .with_context(|| anyhow!("removing {} failed", full_path.display()))?; } } Ok(()) }); Ok(()) } #[turbo_tasks::function(fs)] async fn metadata(&self, fs_path: FileSystemPath) -> Result> { mark_session_dependent(); let full_path = self.to_sys_path(fs_path).await?; self.inner.register_read_invalidator(&full_path)?; let _lock = self.inner.lock_path(&full_path).await; let meta = retry_blocking(full_path.clone(), |path| std::fs::metadata(path)) .concurrency_limited(&self.inner.semaphore) .instrument(tracing::info_span!( "read metadata", name = display(full_path.display()) )) .await .with_context(|| format!("reading metadata for {}", full_path.display()))?; Ok(FileMeta::cell(meta.into())) } } #[turbo_tasks::value_impl] impl ValueToString for DiskFileSystem { #[turbo_tasks::function] fn to_string(&self) -> Vc { Vc::cell(self.inner.name.clone()) } } /// Note: this only works for Unix-style paths (with `/` as a separator). pub fn get_relative_path_to(path: &str, other_path: &str) -> String { fn split(s: &str) -> impl Iterator { let empty = s.is_empty(); let mut iterator = s.split('/'); if empty { iterator.next(); } iterator } let mut self_segments = split(path).peekable(); let mut other_segments = split(other_path).peekable(); while self_segments.peek() == other_segments.peek() { self_segments.next(); if other_segments.next().is_none() { return ".".to_string(); } } let mut result = Vec::new(); if self_segments.peek().is_none() { result.push("."); } else { while self_segments.next().is_some() { result.push(".."); } } for segment in other_segments { result.push(segment); } result.join("/") } #[turbo_tasks::value(shared)] #[derive(Debug, Clone, Hash, TaskInput)] pub struct FileSystemPath { pub fs: ResolvedVc>, pub path: RcStr, } impl FileSystemPath { /// Mimics `ValueToString::to_string`. pub fn value_to_string(&self) -> Vc { value_to_string(self.clone()) } } #[turbo_tasks::function] async fn value_to_string(path: FileSystemPath) -> Result> { Ok(Vc::cell( format!("[{}]/{}", path.fs.to_string().await?, path.path).into(), )) } impl FileSystemPath { pub fn is_inside_ref(&self, other: &FileSystemPath) -> bool { if self.fs == other.fs && self.path.starts_with(&*other.path) { if other.path.is_empty() { true } else { self.path.as_bytes().get(other.path.len()) == Some(&b'/') } } else { false } } pub fn is_inside_or_equal_ref(&self, other: &FileSystemPath) -> bool { if self.fs == other.fs && self.path.starts_with(&*other.path) { if other.path.is_empty() { true } else { matches!( self.path.as_bytes().get(other.path.len()), Some(&b'/') | None ) } } else { false } } pub fn is_root(&self) -> bool { self.path.is_empty() } /// Returns the path of `inner` relative to `self`. /// /// Note: this method always strips the leading `/` from the result. pub fn get_path_to<'a>(&self, inner: &'a FileSystemPath) -> Option<&'a str> { if self.fs != inner.fs { return None; } let path = inner.path.strip_prefix(&*self.path)?; if self.path.is_empty() { Some(path) } else if let Some(stripped) = path.strip_prefix('/') { Some(stripped) } else { None } } pub fn get_relative_path_to(&self, other: &FileSystemPath) -> Option { if self.fs != other.fs { return None; } Some(get_relative_path_to(&self.path, &other.path).into()) } /// Returns the final component of the FileSystemPath, or an empty string /// for the root path. pub fn file_name(&self) -> &str { let (_, file_name) = self.split_file_name(); file_name } /// Returns true if this path has the given extension /// /// slightly faster than `self.extension_ref() == Some(extension)` as we can simply match a /// suffix pub fn has_extension(&self, extension: &str) -> bool { debug_assert!(!extension.contains('/') && extension.starts_with('.')); self.path.ends_with(extension) } /// Returns the extension (without a leading `.`) pub fn extension_ref(&self) -> Option<&str> { let (_, extension) = self.split_extension(); extension } /// Splits the path into two components: /// 1. The path without the extension; /// 2. The extension, if any. fn split_extension(&self) -> (&str, Option<&str>) { if let Some((path_before_extension, extension)) = self.path.rsplit_once('.') { if extension.contains('/') || // The file name begins with a `.` and has no other `.`s within. path_before_extension.ends_with('/') || path_before_extension.is_empty() { (self.path.as_str(), None) } else { (path_before_extension, Some(extension)) } } else { (self.path.as_str(), None) } } /// Splits the path into two components: /// 1. The parent directory, if any; /// 2. The file name; fn split_file_name(&self) -> (Option<&str>, &str) { // Since the path is normalized, we know `parent`, if any, must not be empty. if let Some((parent, file_name)) = self.path.rsplit_once('/') { (Some(parent), file_name) } else { (None, self.path.as_str()) } } /// Splits the path into three components: /// 1. The parent directory, if any; /// 2. The file stem; /// 3. The extension, if any. fn split_file_stem_extension(&self) -> (Option<&str>, &str, Option<&str>) { let (path_before_extension, extension) = self.split_extension(); if let Some((parent, file_stem)) = path_before_extension.rsplit_once('/') { (Some(parent), file_stem, extension) } else { (None, path_before_extension, extension) } } } #[turbo_tasks::value(transparent)] pub struct FileSystemPathOption(Option); #[turbo_tasks::value_impl] impl FileSystemPathOption { #[turbo_tasks::function] pub fn none() -> Vc { Vc::cell(None) } } impl FileSystemPath { /// Create a new FileSystemPath from a path within a FileSystem. The /// /-separated path is expected to be already normalized (this is asserted /// in dev mode). fn new_normalized(fs: ResolvedVc>, path: RcStr) -> Self { // On Windows, the path must be converted to a unix path before creating. But on // Unix, backslashes are a valid char in file names, and the path can be // provided by the user, so we allow it. debug_assert!( MAIN_SEPARATOR != '\\' || !path.contains('\\'), "path {path} must not contain a Windows directory '\\', it must be normalized to Unix \ '/'", ); debug_assert!( normalize_path(&path).as_deref() == Some(&*path), "path {path} must be normalized", ); FileSystemPath { fs, path } } /// Adds a subpath to the current path. The /-separate path argument might /// contain ".." or "." segments, but it must not leave the root of the /// filesystem. pub fn join(&self, path: &str) -> Result { if let Some(path) = join_path(&self.path, path) { Ok(Self::new_normalized(self.fs, path.into())) } else { bail!( "FileSystemPath(\"{}\").join(\"{}\") leaves the filesystem root", self.path, path ); } } /// Adds a suffix to the filename. [path] must not contain `/`. pub fn append(&self, path: &str) -> Result { if path.contains('/') { bail!( "FileSystemPath(\"{}\").append(\"{}\") must not append '/'", self.path, path ) } Ok(Self::new_normalized( self.fs, format!("{}{}", self.path, path).into(), )) } /// Adds a suffix to the basename of the filename. [appending] must not /// contain `/`. Extension will stay intact. pub fn append_to_stem(&self, appending: &str) -> Result { if appending.contains('/') { bail!( "FileSystemPath(\"{}\").append_to_stem(\"{}\") must not append '/'", self.path, appending ) } if let (path, Some(ext)) = self.split_extension() { return Ok(Self::new_normalized( self.fs, format!("{path}{appending}.{ext}").into(), )); } Ok(Self::new_normalized( self.fs, format!("{}{}", self.path, appending).into(), )) } /// Similar to [FileSystemPath::join], but returns an Option that will be /// None when the joined path would leave the filesystem root. #[allow(clippy::needless_borrow)] // for windows build pub fn try_join(&self, path: &str) -> Result> { // TODO(PACK-3279): Remove this once we do not produce invalid paths at the first place. #[cfg(target_os = "windows")] let path = path.replace('\\', "/"); if let Some(path) = join_path(&self.path, &path) { Ok(Some(Self::new_normalized(self.fs, path.into()))) } else { Ok(None) } } /// Similar to [FileSystemPath::join], but returns an Option that will be /// None when the joined path would leave the current path. pub fn try_join_inside(&self, path: &str) -> Result> { if let Some(path) = join_path(&self.path, path) && path.starts_with(&*self.path) { return Ok(Some(Self::new_normalized(self.fs, path.into()))); } Ok(None) } pub fn read_glob(&self, glob: Vc) -> Vc { read_glob(self.clone(), glob) } // Tracks all files and directories matching the glob // Follows symlinks as though they were part of the original hierarchy. pub fn track_glob(&self, glob: Vc, include_dot_files: bool) -> Vc { track_glob(self.clone(), glob, include_dot_files) } pub fn root(&self) -> Vc { self.fs().root() } } impl FileSystemPath { pub fn fs(&self) -> Vc> { *self.fs } pub fn extension(&self) -> &str { self.extension_ref().unwrap_or_default() } pub fn is_inside(&self, other: &FileSystemPath) -> bool { self.is_inside_ref(other) } pub fn is_inside_or_equal(&self, other: &FileSystemPath) -> bool { self.is_inside_or_equal_ref(other) } /// Creates a new [`FileSystemPath`] like `self` but with the given /// extension. pub fn with_extension(&self, extension: &str) -> FileSystemPath { let (path_without_extension, _) = self.split_extension(); Self::new_normalized( self.fs, // Like `Path::with_extension` and `PathBuf::set_extension`, if the extension is empty, // we remove the extension altogether. match extension.is_empty() { true => path_without_extension.into(), false => format!("{path_without_extension}.{extension}").into(), }, ) } /// Extracts the stem (non-extension) portion of self.file_name. /// /// The stem is: /// /// * [`None`], if there is no file name; /// * The entire file name if there is no embedded `.`; /// * The entire file name if the file name begins with `.` and has no other `.`s within; /// * Otherwise, the portion of the file name before the final `.` pub fn file_stem(&self) -> Option<&str> { let (_, file_stem, _) = self.split_file_stem_extension(); if file_stem.is_empty() { return None; } Some(file_stem) } } impl Display for FileSystemPath { fn fmt(&self, f: &mut Formatter<'_>) -> fmt::Result { write!(f, "{}", self.path) } } #[turbo_tasks::function] pub async fn rebase( fs_path: FileSystemPath, old_base: FileSystemPath, new_base: FileSystemPath, ) -> Result> { let new_path; if old_base.path.is_empty() { if new_base.path.is_empty() { new_path = fs_path.path.clone(); } else { new_path = [new_base.path.as_str(), "/", &fs_path.path].concat().into(); } } else { let base_path = [&old_base.path, "/"].concat(); if !fs_path.path.starts_with(&base_path) { bail!( "rebasing {} from {} onto {} doesn't work because it's not part of the source path", fs_path.to_string(), old_base.to_string(), new_base.to_string() ); } if new_base.path.is_empty() { new_path = [&fs_path.path[base_path.len()..]].concat().into(); } else { new_path = [new_base.path.as_str(), &fs_path.path[old_base.path.len()..]] .concat() .into(); } } Ok(new_base.fs.root().await?.join(&new_path)?.cell()) } // Not turbo-tasks functions, only delegating impl FileSystemPath { pub fn read(&self) -> Vc { self.fs().read(self.clone()) } pub fn read_link(&self) -> Vc { self.fs().read_link(self.clone()) } pub fn read_json(&self) -> Vc { self.fs().read(self.clone()).parse_json() } pub fn read_json5(&self) -> Vc { self.fs().read(self.clone()).parse_json5() } /// Reads content of a directory. /// /// DETERMINISM: Result is in random order. Either sort result or do not /// depend on the order. pub fn raw_read_dir(&self) -> Vc { self.fs().raw_read_dir(self.clone()) } pub fn write(&self, content: Vc) -> Vc<()> { self.fs().write(self.clone(), content) } pub fn write_link(&self, target: Vc) -> Vc<()> { self.fs().write_link(self.clone(), target) } pub fn metadata(&self) -> Vc { self.fs().metadata(self.clone()) } pub fn realpath(&self) -> Vc { self.realpath_with_links().path() } pub fn rebase( fs_path: FileSystemPath, old_base: FileSystemPath, new_base: FileSystemPath, ) -> Vc { rebase(fs_path, old_base, new_base) } } impl FileSystemPath { /// Reads content of a directory. /// /// DETERMINISM: Result is in random order. Either sort result or do not /// depend on the order. pub fn read_dir(&self) -> Vc { read_dir(self.clone()) } pub fn parent(&self) -> FileSystemPath { let path = &self.path; if path.is_empty() { return self.clone(); } let p = match str::rfind(path, '/') { Some(index) => path[..index].to_string(), None => "".to_string(), }; FileSystemPath::new_normalized(self.fs, p.into()) } // It is important that get_type uses read_dir and not stat/metadata. // - `get_type` is called very very often during resolving and stat would // make it 1 syscall per call, whereas read_dir would make it 1 syscall per // directory. // - `metadata` allows you to use the "wrong" casing on // case-insensitive filesystems, while read_dir gives you the "correct" // casing. We want to enforce "correct" casing to avoid broken builds on // Vercel deployments (case-sensitive). pub fn get_type(&self) -> Vc { get_type(self.clone()) } pub fn realpath_with_links(&self) -> Vc { realpath_with_links(self.clone()) } } #[turbo_tasks::value_impl] impl ValueToString for FileSystemPath { #[turbo_tasks::function] async fn to_string(&self) -> Result> { Ok(Vc::cell( format!("[{}]/{}", self.fs.to_string().await?, self.path).into(), )) } } #[derive(Clone, Debug)] #[turbo_tasks::value(shared)] pub struct RealPathResult { pub path: FileSystemPath, pub symlinks: Vec, } #[turbo_tasks::value_impl] impl RealPathResult { #[turbo_tasks::function] pub fn path(&self) -> Vc { self.path.clone().cell() } } #[derive(Clone, Copy, Debug, DeterministicHash, PartialOrd, Ord)] #[turbo_tasks::value(shared)] pub enum Permissions { Readable, Writable, Executable, } impl Default for Permissions { fn default() -> Self { Self::Writable } } // Only handle the permissions on unix platform for now #[cfg(target_family = "unix")] impl From for std::fs::Permissions { fn from(perm: Permissions) -> Self { use std::os::unix::fs::PermissionsExt; match perm { Permissions::Readable => std::fs::Permissions::from_mode(0o444), Permissions::Writable => std::fs::Permissions::from_mode(0o664), Permissions::Executable => std::fs::Permissions::from_mode(0o755), } } } #[cfg(target_family = "unix")] impl From for Permissions { fn from(perm: std::fs::Permissions) -> Self { use std::os::unix::fs::PermissionsExt; if perm.readonly() { Permissions::Readable } else { // https://github.com/fitzgen/is_executable/blob/master/src/lib.rs#L96 if perm.mode() & 0o111 != 0 { Permissions::Executable } else { Permissions::Writable } } } } #[cfg(not(target_family = "unix"))] impl From for Permissions { fn from(_: std::fs::Permissions) -> Self { Permissions::default() } } #[turbo_tasks::value(shared)] #[derive(Clone, Debug, DeterministicHash, PartialOrd, Ord)] pub enum FileContent { Content(File), NotFound, } impl From for FileContent { fn from(file: File) -> Self { FileContent::Content(file) } } impl From for Vc { fn from(file: File) -> Self { FileContent::Content(file).cell() } } #[derive(Clone, Debug, Eq, PartialEq)] enum FileComparison { Create, Equal, NotEqual, } impl FileContent { /// Performs a comparison of self's data against a disk file's streamed /// read. async fn streaming_compare(&self, path: &Path) -> Result { let old_file = extract_disk_access( retry_blocking(path.to_path_buf(), |path| std::fs::File::open(path)).await, path, )?; let Some(old_file) = old_file else { return Ok(match self { FileContent::NotFound => FileComparison::Equal, _ => FileComparison::Create, }); }; // We know old file exists, does the new file? let FileContent::Content(new_file) = self else { return Ok(FileComparison::NotEqual); }; let old_meta = extract_disk_access( retry_blocking(path.to_path_buf(), { let file_for_metadata = old_file.try_clone()?; move |_| file_for_metadata.metadata() }) .await, path, )?; let Some(old_meta) = old_meta else { // If we failed to get meta, then the old file has been deleted between the // handle open. In which case, we just pretend the file never // existed. return Ok(FileComparison::Create); }; // If the meta is different, we need to rewrite the file to update it. if new_file.meta != old_meta.into() { return Ok(FileComparison::NotEqual); } // So meta matches, and we have a file handle. Let's stream the contents to see // if they match. let mut new_contents = new_file.read(); let mut old_contents = BufReader::new(old_file); Ok(loop { let new_chunk = new_contents.fill_buf()?; let Ok(old_chunk) = old_contents.fill_buf() else { break FileComparison::NotEqual; }; let len = min(new_chunk.len(), old_chunk.len()); if len == 0 { if new_chunk.len() == old_chunk.len() { break FileComparison::Equal; } else { break FileComparison::NotEqual; } } if new_chunk[0..len] != old_chunk[0..len] { break FileComparison::NotEqual; } new_contents.consume(len); old_contents.consume(len); }) } } bitflags! { #[derive(Default, Serialize, Deserialize, TraceRawVcs, NonLocalValue)] pub struct LinkType: u8 { const DIRECTORY = 0b00000001; const ABSOLUTE = 0b00000010; } } #[turbo_tasks::value(shared)] #[derive(Debug)] pub enum LinkContent { // for the relative link, the target is raw value read from the link // for the absolute link, the target is stripped of the root path while reading // We don't use the `FileSystemPath` here for now, because the `FileSystemPath` is always // normalized, which means in `fn write_link` we couldn't restore the raw value of the file // link because there is only **dist** path in `fn write_link`, and we need the raw path if // we want to restore the link value in `fn write_link` Link { target: RcStr, link_type: LinkType }, Invalid, NotFound, } #[turbo_tasks::value(shared)] #[derive(Clone, DeterministicHash, PartialOrd, Ord)] pub struct File { #[turbo_tasks(debug_ignore)] content: Rope, meta: FileMeta, } impl File { /// Reads a [File] from the given path fn from_path(p: &Path) -> io::Result { let mut file = std::fs::File::open(p)?; let metadata = file.metadata()?; let mut output = Vec::with_capacity(metadata.len() as usize); file.read_to_end(&mut output)?; Ok(File { meta: metadata.into(), content: Rope::from(output), }) } /// Creates a [File] from raw bytes. fn from_bytes(content: Vec) -> Self { File { meta: FileMeta::default(), content: Rope::from(content), } } /// Creates a [File] from a rope. fn from_rope(content: Rope) -> Self { File { meta: FileMeta::default(), content, } } /// Returns the content type associated with this file. pub fn content_type(&self) -> Option<&Mime> { self.meta.content_type.as_ref() } /// Sets the content type associated with this file. pub fn with_content_type(mut self, content_type: Mime) -> Self { self.meta.content_type = Some(content_type); self } /// Returns a Read/AsyncRead/Stream/Iterator to access the File's contents. pub fn read(&self) -> RopeReader { self.content.read() } } impl Debug for File { fn fmt(&self, f: &mut Formatter<'_>) -> fmt::Result { f.debug_struct("File") .field("meta", &self.meta) .field("content (hash)", &hash_xxh3_hash64(&self.content)) .finish() } } impl From for File { fn from(s: RcStr) -> Self { s.into_owned().into() } } impl From for File { fn from(s: String) -> Self { File::from_bytes(s.into_bytes()) } } impl From> for File { fn from(s: ReadRef) -> Self { File::from_bytes(s.as_bytes().to_vec()) } } impl From<&str> for File { fn from(s: &str) -> Self { File::from_bytes(s.as_bytes().to_vec()) } } impl From> for File { fn from(bytes: Vec) -> Self { File::from_bytes(bytes) } } impl From<&[u8]> for File { fn from(bytes: &[u8]) -> Self { File::from_bytes(bytes.to_vec()) } } impl From> for File { fn from(rope: ReadRef) -> Self { File::from_rope(ReadRef::into_owned(rope)) } } impl From for File { fn from(rope: Rope) -> Self { File::from_rope(rope) } } impl File { pub fn new(meta: FileMeta, content: Vec) -> Self { Self { meta, content: Rope::from(content), } } /// Returns the associated [FileMeta] of this file. pub fn meta(&self) -> &FileMeta { &self.meta } /// Returns the immutable contents of this file. pub fn content(&self) -> &Rope { &self.content } } mod mime_option_serde { use std::{fmt, str::FromStr}; use mime::Mime; use serde::{Deserializer, Serializer, de}; pub fn serialize(mime: &Option, serializer: S) -> Result where S: Serializer, { if let Some(mime) = mime { serializer.serialize_str(mime.as_ref()) } else { serializer.serialize_str("") } } pub fn deserialize<'de, D>(deserializer: D) -> Result, D::Error> where D: Deserializer<'de>, { struct Visitor; impl de::Visitor<'_> for Visitor { type Value = Option; fn expecting(&self, formatter: &mut fmt::Formatter) -> fmt::Result { formatter.write_str("a valid MIME type or empty string") } fn visit_str(self, value: &str) -> Result, E> where E: de::Error, { if value.is_empty() { Ok(None) } else { Mime::from_str(value) .map(Some) .map_err(|e| E::custom(format!("{e}"))) } } } deserializer.deserialize_str(Visitor) } } #[turbo_tasks::value(shared)] #[derive(Debug, Clone, Default)] pub struct FileMeta { // Size of the file // len: u64, permissions: Permissions, #[serde(with = "mime_option_serde")] #[turbo_tasks(trace_ignore)] content_type: Option, } impl Ord for FileMeta { fn cmp(&self, other: &Self) -> Ordering { self.permissions .cmp(&other.permissions) .then_with(|| self.content_type.as_ref().cmp(&other.content_type.as_ref())) } } impl PartialOrd for FileMeta { fn partial_cmp(&self, other: &Self) -> Option { Some(self.cmp(other)) } } impl From for FileMeta { fn from(meta: std::fs::Metadata) -> Self { let permissions = meta.permissions().into(); Self { permissions, content_type: None, } } } impl DeterministicHash for FileMeta { fn deterministic_hash(&self, state: &mut H) { self.permissions.deterministic_hash(state); if let Some(content_type) = &self.content_type { content_type.to_string().deterministic_hash(state); } } } impl FileContent { pub fn new(file: File) -> Self { FileContent::Content(file) } pub fn is_content(&self) -> bool { matches!(self, FileContent::Content(_)) } pub fn as_content(&self) -> Option<&File> { match self { FileContent::Content(file) => Some(file), FileContent::NotFound => None, } } pub fn parse_json_ref(&self) -> FileJsonContent { match self { FileContent::Content(file) => { let content = file.content.clone().into_bytes(); let de = &mut serde_json::Deserializer::from_slice(&content); match serde_path_to_error::deserialize(de) { Ok(data) => FileJsonContent::Content(data), Err(e) => FileJsonContent::Unparsable(Box::new( UnparsableJson::from_serde_path_to_error(e), )), } } FileContent::NotFound => FileJsonContent::NotFound, } } pub fn parse_json_with_comments_ref(&self) -> FileJsonContent { match self { FileContent::Content(file) => match file.content.to_str() { Ok(string) => match parse_to_serde_value( &string, &ParseOptions { allow_comments: true, allow_trailing_commas: true, allow_loose_object_property_names: false, }, ) { Ok(data) => match data { Some(value) => FileJsonContent::Content(value), None => FileJsonContent::unparsable( "text content doesn't contain any json data", ), }, Err(e) => FileJsonContent::Unparsable(Box::new( UnparsableJson::from_jsonc_error(e, string.as_ref()), )), }, Err(_) => FileJsonContent::unparsable("binary is not valid utf-8 text"), }, FileContent::NotFound => FileJsonContent::NotFound, } } pub fn parse_json5_ref(&self) -> FileJsonContent { match self { FileContent::Content(file) => match file.content.to_str() { Ok(string) => match parse_to_serde_value( &string, &ParseOptions { allow_comments: true, allow_trailing_commas: true, allow_loose_object_property_names: true, }, ) { Ok(data) => match data { Some(value) => FileJsonContent::Content(value), None => FileJsonContent::unparsable( "text content doesn't contain any json data", ), }, Err(e) => FileJsonContent::Unparsable(Box::new( UnparsableJson::from_jsonc_error(e, string.as_ref()), )), }, Err(_) => FileJsonContent::unparsable("binary is not valid utf-8 text"), }, FileContent::NotFound => FileJsonContent::NotFound, } } pub fn lines_ref(&self) -> FileLinesContent { match self { FileContent::Content(file) => match file.content.to_str() { Ok(string) => { let mut bytes_offset = 0; FileLinesContent::Lines( string .split('\n') .map(|l| { let line = FileLine { content: l.to_string(), bytes_offset, }; bytes_offset += (l.len() + 1) as u32; line }) .collect(), ) } Err(_) => FileLinesContent::Unparsable, }, FileContent::NotFound => FileLinesContent::NotFound, } } } #[turbo_tasks::value_impl] impl FileContent { #[turbo_tasks::function] pub fn len(&self) -> Result>> { Ok(Vc::cell(match self { FileContent::Content(file) => Some(file.content.len() as u64), FileContent::NotFound => None, })) } #[turbo_tasks::function] pub fn parse_json(&self) -> Result> { Ok(self.parse_json_ref().into()) } #[turbo_tasks::function] pub async fn parse_json_with_comments(self: Vc) -> Result> { let this = self.await?; Ok(this.parse_json_with_comments_ref().into()) } #[turbo_tasks::function] pub async fn parse_json5(self: Vc) -> Result> { let this = self.await?; Ok(this.parse_json5_ref().into()) } #[turbo_tasks::function] pub async fn lines(self: Vc) -> Result> { let this = self.await?; Ok(this.lines_ref().into()) } #[turbo_tasks::function] pub async fn hash(self: Vc) -> Result> { Ok(Vc::cell(hash_xxh3_hash64(&self.await?))) } } /// A file's content interpreted as a JSON value. #[turbo_tasks::value(shared, serialization = "none")] pub enum FileJsonContent { Content(Value), Unparsable(Box), NotFound, } #[turbo_tasks::value_impl] impl ValueToString for FileJsonContent { /// Returns the JSON file content as a UTF-8 string. /// /// This operation will only succeed if the file contents are a valid JSON /// value. #[turbo_tasks::function] fn to_string(&self) -> Result> { match self { FileJsonContent::Content(json) => Ok(Vc::cell(json.to_string().into())), FileJsonContent::Unparsable(e) => Err(anyhow!("File is not valid JSON: {}", e)), FileJsonContent::NotFound => Err(anyhow!("File not found")), } } } #[turbo_tasks::value_impl] impl FileJsonContent { #[turbo_tasks::function] pub async fn content(self: Vc) -> Result> { match &*self.await? { FileJsonContent::Content(json) => Ok(Vc::cell(json.clone())), FileJsonContent::Unparsable(e) => Err(anyhow!("File is not valid JSON: {}", e)), FileJsonContent::NotFound => Err(anyhow!("File not found")), } } } impl FileJsonContent { pub fn unparsable(message: &'static str) -> Self { FileJsonContent::Unparsable(Box::new(UnparsableJson { message: Cow::Borrowed(message), path: None, start_location: None, end_location: None, })) } pub fn unparsable_with_message(message: Cow<'static, str>) -> Self { FileJsonContent::Unparsable(Box::new(UnparsableJson { message, path: None, start_location: None, end_location: None, })) } } #[derive(Debug, PartialEq, Eq)] pub struct FileLine { pub content: String, pub bytes_offset: u32, } #[turbo_tasks::value(shared, serialization = "none")] pub enum FileLinesContent { Lines(#[turbo_tasks(trace_ignore)] Vec), Unparsable, NotFound, } #[derive(Hash, Clone, Debug, PartialEq, Eq, TraceRawVcs, Serialize, Deserialize, NonLocalValue)] pub enum RawDirectoryEntry { File, Directory, Symlink, Other, Error, } #[derive(Hash, Clone, Debug, PartialEq, Eq, TraceRawVcs, Serialize, Deserialize, NonLocalValue)] pub enum DirectoryEntry { File(FileSystemPath), Directory(FileSystemPath), Symlink(FileSystemPath), Other(FileSystemPath), Error, } impl DirectoryEntry { /// Handles the `DirectoryEntry::Symlink` variant by checking the symlink target /// type and replacing it with `DirectoryEntry::File` or /// `DirectoryEntry::Directory`. pub async fn resolve_symlink(self) -> Result { if let DirectoryEntry::Symlink(symlink) = &self { let real_path = symlink.realpath().owned().await?; match *real_path.get_type().await? { FileSystemEntryType::Directory => Ok(DirectoryEntry::Directory(real_path)), FileSystemEntryType::File => Ok(DirectoryEntry::File(real_path)), _ => Ok(self), } } else { Ok(self) } } pub fn path(self) -> Option { match self { DirectoryEntry::File(path) | DirectoryEntry::Directory(path) | DirectoryEntry::Symlink(path) | DirectoryEntry::Other(path) => Some(path), DirectoryEntry::Error => None, } } } #[turbo_tasks::value] #[derive(Hash, Clone, Copy, Debug)] pub enum FileSystemEntryType { NotFound, File, Directory, Symlink, Other, Error, } impl From for FileSystemEntryType { fn from(file_type: FileType) -> Self { match file_type { t if t.is_dir() => FileSystemEntryType::Directory, t if t.is_file() => FileSystemEntryType::File, t if t.is_symlink() => FileSystemEntryType::Symlink, _ => FileSystemEntryType::Other, } } } impl From for FileSystemEntryType { fn from(entry: DirectoryEntry) -> Self { FileSystemEntryType::from(&entry) } } impl From<&DirectoryEntry> for FileSystemEntryType { fn from(entry: &DirectoryEntry) -> Self { match entry { DirectoryEntry::File(_) => FileSystemEntryType::File, DirectoryEntry::Directory(_) => FileSystemEntryType::Directory, DirectoryEntry::Symlink(_) => FileSystemEntryType::Symlink, DirectoryEntry::Other(_) => FileSystemEntryType::Other, DirectoryEntry::Error => FileSystemEntryType::Error, } } } impl From for FileSystemEntryType { fn from(entry: RawDirectoryEntry) -> Self { FileSystemEntryType::from(&entry) } } impl From<&RawDirectoryEntry> for FileSystemEntryType { fn from(entry: &RawDirectoryEntry) -> Self { match entry { RawDirectoryEntry::File => FileSystemEntryType::File, RawDirectoryEntry::Directory => FileSystemEntryType::Directory, RawDirectoryEntry::Symlink => FileSystemEntryType::Symlink, RawDirectoryEntry::Other => FileSystemEntryType::Other, RawDirectoryEntry::Error => FileSystemEntryType::Error, } } } #[turbo_tasks::value] #[derive(Debug)] pub enum RawDirectoryContent { // The entry keys are the directory relative file names // e.g. for `/bar/foo`, it will be `foo` Entries(AutoMap), NotFound, } impl RawDirectoryContent { pub fn new(entries: AutoMap) -> Vc { Self::cell(RawDirectoryContent::Entries(entries)) } pub fn not_found() -> Vc { Self::cell(RawDirectoryContent::NotFound) } } #[turbo_tasks::value] #[derive(Debug)] pub enum DirectoryContent { Entries(AutoMap), NotFound, } impl DirectoryContent { pub fn new(entries: AutoMap) -> Vc { Self::cell(DirectoryContent::Entries(entries)) } pub fn not_found() -> Vc { Self::cell(DirectoryContent::NotFound) } } #[turbo_tasks::value(shared)] pub struct NullFileSystem; #[turbo_tasks::value_impl] impl FileSystem for NullFileSystem { #[turbo_tasks::function] fn read(&self, _fs_path: FileSystemPath) -> Vc { FileContent::NotFound.cell() } #[turbo_tasks::function] fn read_link(&self, _fs_path: FileSystemPath) -> Vc { LinkContent::NotFound.into() } #[turbo_tasks::function] fn raw_read_dir(&self, _fs_path: FileSystemPath) -> Vc { RawDirectoryContent::not_found() } #[turbo_tasks::function] fn write(&self, _fs_path: FileSystemPath, _content: Vc) -> Vc<()> { Vc::default() } #[turbo_tasks::function] fn write_link(&self, _fs_path: FileSystemPath, _target: Vc) -> Vc<()> { Vc::default() } #[turbo_tasks::function] fn metadata(&self, _fs_path: FileSystemPath) -> Vc { FileMeta::default().cell() } } #[turbo_tasks::value_impl] impl ValueToString for NullFileSystem { #[turbo_tasks::function] fn to_string(&self) -> Vc { Vc::cell(rcstr!("null")) } } pub async fn to_sys_path(mut path: FileSystemPath) -> Result> { loop { if let Some(fs) = Vc::try_resolve_downcast_type::(path.fs()).await? { path = fs.get_inner_fs_path(path).owned().await?; continue; } if let Some(fs) = Vc::try_resolve_downcast_type::(path.fs()).await? { let sys_path = fs.await?.to_sys_path(path).await?; return Ok(Some(sys_path)); } return Ok(None); } } #[turbo_tasks::function] async fn read_dir(path: FileSystemPath) -> Result> { let fs = path.fs().to_resolved().await?; match &*fs.raw_read_dir(path.clone()).await? { RawDirectoryContent::NotFound => Ok(DirectoryContent::not_found()), RawDirectoryContent::Entries(entries) => { let mut normalized_entries = AutoMap::new(); let dir_path = &path.path; for (name, entry) in entries { // Construct the path directly instead of going through `join`. // We do not need to normalize since the `name` is guaranteed to be a simple // path segment. let path = if dir_path.is_empty() { name.clone() } else { RcStr::from(format!("{dir_path}/{name}")) }; let entry_path = FileSystemPath::new_normalized(fs, path); let entry = match entry { RawDirectoryEntry::File => DirectoryEntry::File(entry_path), RawDirectoryEntry::Directory => DirectoryEntry::Directory(entry_path), RawDirectoryEntry::Symlink => DirectoryEntry::Symlink(entry_path), RawDirectoryEntry::Other => DirectoryEntry::Other(entry_path), RawDirectoryEntry::Error => DirectoryEntry::Error, }; normalized_entries.insert(name.clone(), entry); } Ok(DirectoryContent::new(normalized_entries)) } } } #[turbo_tasks::function] async fn get_type(path: FileSystemPath) -> Result> { if path.is_root() { return Ok(FileSystemEntryType::Directory.cell()); } let parent = path.parent(); let dir_content = parent.raw_read_dir().await?; match &*dir_content { RawDirectoryContent::NotFound => Ok(FileSystemEntryType::NotFound.cell()), RawDirectoryContent::Entries(entries) => { let (_, file_name) = path.split_file_name(); if let Some(entry) = entries.get(file_name) { Ok(FileSystemEntryType::from(entry).cell()) } else { Ok(FileSystemEntryType::NotFound.cell()) } } } } #[turbo_tasks::function] async fn realpath_with_links(path: FileSystemPath) -> Result> { let mut current_vc = path.clone(); let mut symlinks: IndexSet = IndexSet::new(); let mut visited: AutoSet = AutoSet::new(); // Pick some arbitrary symlink depth limit... similar to the ELOOP logic for realpath(3). // SYMLOOP_MAX is 40 for Linux: https://unix.stackexchange.com/q/721724 for _i in 0..40 { let current = current_vc.clone(); if current.is_root() { // fast path return Ok(RealPathResult { path: current_vc, symlinks: symlinks.into_iter().collect(), } .cell()); } if !visited.insert(current.path.clone()) { break; // we detected a cycle } // see if a parent segment of the path is a symlink and resolve that first let parent = current_vc.parent(); let parent_result = parent.realpath_with_links().owned().await?; let basename = current .path .rsplit_once('/') .map_or(current.path.as_str(), |(_, name)| name); if parent_result.path != parent { current_vc = parent_result.path.join(basename)?; } symlinks.extend(parent_result.symlinks); // use `get_type` before trying `read_link`, as there's a good chance of a cache hit on // `get_type`, and `read_link` isn't the common codepath. if !matches!(*current_vc.get_type().await?, FileSystemEntryType::Symlink) { return Ok(RealPathResult { path: current_vc, symlinks: symlinks.into_iter().collect(), // convert set to vec } .cell()); } if let LinkContent::Link { target, link_type } = &*current_vc.read_link().await? { symlinks.insert(current_vc.clone()); current_vc = if link_type.contains(LinkType::ABSOLUTE) { current_vc.root().owned().await? } else { parent_result.path } .join(target)?; } else { // get_type() and read_link() might disagree temporarily due to turbo-tasks // eventual consistency or if the file gets invalidated before the directory does return Ok(RealPathResult { path: current_vc, symlinks: symlinks.into_iter().collect(), // convert set to vec } .cell()); } } // Too many attempts or detected a cycle, we bailed out! // // TODO: There's no proper way to indicate an non-turbo-tasks error here, so just return the // original path and all the symlinks we followed. // // Returning the followed symlinks is still important, even if there is an error! Otherwise // we may never notice if the symlink loop is fixed. Ok(RealPathResult { path, symlinks: symlinks.into_iter().collect(), } .cell()) } pub fn register() { turbo_tasks::register(); include!(concat!(env!("OUT_DIR"), "/register.rs")); } #[cfg(test)] mod tests { use turbo_rcstr::rcstr; use super::*; #[test] fn test_get_relative_path_to() { assert_eq!(get_relative_path_to("a/b/c", "a/b/c").as_str(), "."); assert_eq!(get_relative_path_to("a/c/d", "a/b/c").as_str(), "../../b/c"); assert_eq!(get_relative_path_to("", "a/b/c").as_str(), "./a/b/c"); assert_eq!(get_relative_path_to("a/b/c", "").as_str(), "../../.."); assert_eq!( get_relative_path_to("a/b/c", "c/b/a").as_str(), "../../../c/b/a" ); assert_eq!( get_relative_path_to("file:///a/b/c", "file:///c/b/a").as_str(), "../../../c/b/a" ); } #[tokio::test] async fn with_extension() { crate::register(); turbo_tasks_testing::VcStorage::with(async { let fs = Vc::upcast::>(VirtualFileSystem::new()) .to_resolved() .await?; let path_txt = FileSystemPath::new_normalized(fs, rcstr!("foo/bar.txt")); let path_json = path_txt.with_extension("json"); assert_eq!(&*path_json.path, "foo/bar.json"); let path_no_ext = path_txt.with_extension(""); assert_eq!(&*path_no_ext.path, "foo/bar"); let path_new_ext = path_no_ext.with_extension("json"); assert_eq!(&*path_new_ext.path, "foo/bar.json"); let path_no_slash_txt = FileSystemPath::new_normalized(fs, rcstr!("bar.txt")); let path_no_slash_json = path_no_slash_txt.with_extension("json"); assert_eq!(path_no_slash_json.path.as_str(), "bar.json"); let path_no_slash_no_ext = path_no_slash_txt.with_extension(""); assert_eq!(path_no_slash_no_ext.path.as_str(), "bar"); let path_no_slash_new_ext = path_no_slash_no_ext.with_extension("json"); assert_eq!(path_no_slash_new_ext.path.as_str(), "bar.json"); anyhow::Ok(()) }) .await .unwrap() } #[tokio::test] async fn file_stem() { crate::register(); turbo_tasks_testing::VcStorage::with(async { let fs = Vc::upcast::>(VirtualFileSystem::new()) .to_resolved() .await?; let path = FileSystemPath::new_normalized(fs, rcstr!("")); assert_eq!(path.file_stem(), None); let path = FileSystemPath::new_normalized(fs, rcstr!("foo/bar.txt")); assert_eq!(path.file_stem(), Some("bar")); let path = FileSystemPath::new_normalized(fs, rcstr!("bar.txt")); assert_eq!(path.file_stem(), Some("bar")); let path = FileSystemPath::new_normalized(fs, rcstr!("foo/bar")); assert_eq!(path.file_stem(), Some("bar")); let path = FileSystemPath::new_normalized(fs, rcstr!("foo/.bar")); assert_eq!(path.file_stem(), Some(".bar")); anyhow::Ok(()) }) .await .unwrap() } }