Dataset Viewer
Auto-converted to Parquet Duplicate
repo
stringlengths
6
65
file_url
stringlengths
81
311
file_path
stringlengths
6
227
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 15:31:58
2026-01-04 20:25:31
truncated
bool
2 classes
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/config.rs
src/config.rs
use std::{path::PathBuf, sync::Arc, time::Duration}; use lscolors::LsColors; use regex::bytes::RegexSet; use crate::exec::CommandSet; use crate::filetypes::FileTypes; #[cfg(unix)] use crate::filter::OwnerFilter; use crate::filter::{SizeFilter, TimeFilter}; use crate::fmt::FormatTemplate; /// Configuration options for *fd*. pub struct Config { /// Whether the search is case-sensitive or case-insensitive. pub case_sensitive: bool, /// Whether to search within the full file path or just the base name (filename or directory /// name). pub search_full_path: bool, /// Whether to ignore hidden files and directories (or not). pub ignore_hidden: bool, /// Whether to respect `.fdignore` files or not. pub read_fdignore: bool, /// Whether to respect ignore files in parent directories or not. pub read_parent_ignore: bool, /// Whether to respect VCS ignore files (`.gitignore`, ..) or not. pub read_vcsignore: bool, /// Whether to require a `.git` directory to respect gitignore files. pub require_git_to_read_vcsignore: bool, /// Whether to respect the global ignore file or not. pub read_global_ignore: bool, /// Whether to follow symlinks or not. pub follow_links: bool, /// Whether to limit the search to starting file system or not. pub one_file_system: bool, /// Whether elements of output should be separated by a null character pub null_separator: bool, /// The maximum search depth, or `None` if no maximum search depth should be set. /// /// A depth of `1` includes all files under the current directory, a depth of `2` also includes /// all files under subdirectories of the current directory, etc. pub max_depth: Option<usize>, /// The minimum depth for reported entries, or `None`. pub min_depth: Option<usize>, /// Whether to stop traversing into matching directories. pub prune: bool, /// The number of threads to use. pub threads: usize, /// If true, the program doesn't print anything and will instead return an exit code of 0 /// if there's at least one match. Otherwise, the exit code will be 1. pub quiet: bool, /// Time to buffer results internally before streaming to the console. This is useful to /// provide a sorted output, in case the total execution time is shorter than /// `max_buffer_time`. pub max_buffer_time: Option<Duration>, /// `None` if the output should not be colorized. Otherwise, a `LsColors` instance that defines /// how to style different filetypes. pub ls_colors: Option<LsColors>, /// Whether or not we are writing to an interactive terminal #[cfg_attr(not(unix), allow(unused))] pub interactive_terminal: bool, /// The type of file to search for. If set to `None`, all file types are displayed. If /// set to `Some(..)`, only the types that are specified are shown. pub file_types: Option<FileTypes>, /// The extension to search for. Only entries matching the extension will be included. /// /// The value (if present) will be a lowercase string without leading dots. pub extensions: Option<RegexSet>, /// A format string to use to format results, similarly to exec pub format: Option<FormatTemplate>, /// If a value is supplied, each item found will be used to generate and execute commands. pub command: Option<Arc<CommandSet>>, /// Maximum number of search results to pass to each `command`. If zero, the number is /// unlimited. pub batch_size: usize, /// A list of glob patterns that should be excluded from the search. pub exclude_patterns: Vec<String>, /// A list of custom ignore files. pub ignore_files: Vec<PathBuf>, /// The given constraints on the size of returned files pub size_constraints: Vec<SizeFilter>, /// Constraints on last modification time of files pub time_constraints: Vec<TimeFilter>, #[cfg(unix)] /// User/group ownership constraint pub owner_constraint: Option<OwnerFilter>, /// Whether or not to display filesystem errors pub show_filesystem_errors: bool, /// The separator used to print file paths. pub path_separator: Option<String>, /// The actual separator, either the system default separator or `path_separator` pub actual_path_separator: String, /// The maximum number of search results pub max_results: Option<usize>, /// Whether or not to strip the './' prefix for search results pub strip_cwd_prefix: bool, /// Whether or not to use hyperlinks on paths pub hyperlink: bool, } impl Config { /// Check whether results are being printed. pub fn is_printing(&self) -> bool { self.command.is_none() } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/hyperlink.rs
src/hyperlink.rs
use crate::filesystem::absolute_path; use std::fmt::{self, Formatter, Write}; use std::path::{Path, PathBuf}; pub(crate) struct PathUrl(PathBuf); impl PathUrl { pub(crate) fn new(path: &Path) -> Option<PathUrl> { Some(PathUrl(absolute_path(path).ok()?)) } } impl fmt::Display for PathUrl { fn fmt(&self, f: &mut Formatter<'_>) -> fmt::Result { write!(f, "file://{}", host())?; let bytes = self.0.as_os_str().as_encoded_bytes(); for &byte in bytes.iter() { encode(f, byte)?; } Ok(()) } } fn encode(f: &mut Formatter, byte: u8) -> fmt::Result { // NOTE: // Most terminals can handle non-ascii unicode characters in a file url fine. But on some OSes (notably // windows), the encoded bytes of the path may not be valid UTF-8. Since we don't know if a // byte >= 128 is part of a valid UTF-8 encoding or not, we just percent encode any non-ascii // byte. // Percent encoding these bytes is probably safer anyway. match byte { b'0'..=b'9' | b'A'..=b'Z' | b'a'..=b'z' | b'/' | b':' | b'-' | b'.' | b'_' | b'~' => { f.write_char(byte.into()) } #[cfg(windows)] b'\\' => f.write_char('/'), _ => { write!(f, "%{byte:02X}") } } } #[cfg(unix)] fn host() -> &'static str { use std::sync::OnceLock; static HOSTNAME: OnceLock<String> = OnceLock::new(); HOSTNAME .get_or_init(|| { nix::unistd::gethostname() .ok() .and_then(|h| h.into_string().ok()) .unwrap_or_default() }) .as_ref() } #[cfg(not(unix))] const fn host() -> &'static str { "" } #[cfg(test)] mod test { use super::*; // This allows us to test the encoding without having to worry about the host, or absolute path struct Encoded(&'static str); impl fmt::Display for Encoded { fn fmt(&self, f: &mut Formatter<'_>) -> fmt::Result { for byte in self.0.bytes() { encode(f, byte)?; } Ok(()) } } #[test] fn test_unicode_encoding() { assert_eq!( Encoded("$*\x1bßé/βˆ«πŸ˜ƒ\x07").to_string(), "%24%2A%1B%C3%9F%C3%A9/%E2%88%AB%F0%9F%98%83%07", ); } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/filesystem.rs
src/filesystem.rs
use std::borrow::Cow; use std::env; use std::ffi::OsStr; use std::fs; use std::io; #[cfg(any(unix, target_os = "redox"))] use std::os::unix::fs::FileTypeExt; use std::path::{Path, PathBuf}; use normpath::PathExt; use crate::dir_entry; pub fn path_absolute_form(path: &Path) -> io::Result<PathBuf> { if path.is_absolute() { return Ok(path.to_path_buf()); } let path = path.strip_prefix(".").unwrap_or(path); env::current_dir().map(|path_buf| path_buf.join(path)) } pub fn absolute_path(path: &Path) -> io::Result<PathBuf> { let path_buf = path_absolute_form(path)?; #[cfg(windows)] let path_buf = Path::new( path_buf .as_path() .to_string_lossy() .trim_start_matches(r"\\?\"), ) .to_path_buf(); Ok(path_buf) } pub fn is_existing_directory(path: &Path) -> bool { // Note: we do not use `.exists()` here, as `.` always exists, even if // the CWD has been deleted. path.is_dir() && (path.file_name().is_some() || path.normalize().is_ok()) } pub fn is_empty(entry: &dir_entry::DirEntry) -> bool { if let Some(file_type) = entry.file_type() { if file_type.is_dir() { if let Ok(mut entries) = fs::read_dir(entry.path()) { entries.next().is_none() } else { false } } else if file_type.is_file() { entry.metadata().map(|m| m.len() == 0).unwrap_or(false) } else { false } } else { false } } #[cfg(any(unix, target_os = "redox"))] pub fn is_block_device(ft: fs::FileType) -> bool { ft.is_block_device() } #[cfg(windows)] pub fn is_block_device(_: fs::FileType) -> bool { false } #[cfg(any(unix, target_os = "redox"))] pub fn is_char_device(ft: fs::FileType) -> bool { ft.is_char_device() } #[cfg(windows)] pub fn is_char_device(_: fs::FileType) -> bool { false } #[cfg(any(unix, target_os = "redox"))] pub fn is_socket(ft: fs::FileType) -> bool { ft.is_socket() } #[cfg(windows)] pub fn is_socket(_: fs::FileType) -> bool { false } #[cfg(any(unix, target_os = "redox"))] pub fn is_pipe(ft: fs::FileType) -> bool { ft.is_fifo() } #[cfg(windows)] pub fn is_pipe(_: fs::FileType) -> bool { false } #[cfg(any(unix, target_os = "redox"))] pub fn osstr_to_bytes(input: &OsStr) -> Cow<'_, [u8]> { use std::os::unix::ffi::OsStrExt; Cow::Borrowed(input.as_bytes()) } #[cfg(windows)] pub fn osstr_to_bytes(input: &OsStr) -> Cow<'_, [u8]> { let string = input.to_string_lossy(); match string { Cow::Owned(string) => Cow::Owned(string.into_bytes()), Cow::Borrowed(string) => Cow::Borrowed(string.as_bytes()), } } /// Remove the `./` prefix from a path. pub fn strip_current_dir(path: &Path) -> &Path { path.strip_prefix(".").unwrap_or(path) } /// Default value for the path_separator, mainly for MSYS/MSYS2, which set the MSYSTEM /// environment variable, and we set fd's path separator to '/' rather than Rust's default of '\'. /// /// Returns Some to use a nonstandard path separator, or None to use rust's default on the target /// platform. pub fn default_path_separator() -> Option<String> { if cfg!(windows) { let msystem = env::var("MSYSTEM").ok()?; if !msystem.is_empty() { return Some("/".to_owned()); } } None } #[cfg(test)] mod tests { use super::strip_current_dir; use std::path::Path; #[test] fn strip_current_dir_basic() { assert_eq!(strip_current_dir(Path::new("./foo")), Path::new("foo")); assert_eq!(strip_current_dir(Path::new("foo")), Path::new("foo")); assert_eq!( strip_current_dir(Path::new("./foo/bar/baz")), Path::new("foo/bar/baz") ); assert_eq!( strip_current_dir(Path::new("foo/bar/baz")), Path::new("foo/bar/baz") ); } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/cli.rs
src/cli.rs
use std::num::NonZeroUsize; use std::path::{Path, PathBuf}; use std::time::Duration; use anyhow::anyhow; use clap::{ Arg, ArgAction, ArgGroup, ArgMatches, Command, Parser, ValueEnum, error::ErrorKind, value_parser, }; #[cfg(feature = "completions")] use clap_complete::Shell; use normpath::PathExt; use crate::error::print_error; use crate::exec::CommandSet; use crate::filesystem; #[cfg(unix)] use crate::filter::OwnerFilter; use crate::filter::SizeFilter; #[derive(Parser)] #[command( name = "fd", version, about = "A program to find entries in your filesystem", after_long_help = "Bugs can be reported on GitHub: https://github.com/sharkdp/fd/issues", max_term_width = 98, args_override_self = true, group(ArgGroup::new("execs").args(&["exec", "exec_batch", "list_details"]).conflicts_with_all(&[ "max_results", "quiet", "max_one_result"])), )] pub struct Opts { /// Include hidden directories and files in the search results (default: /// hidden files and directories are skipped). Files and directories are /// considered to be hidden if their name starts with a `.` sign (dot). /// Any files or directories that are ignored due to the rules described by /// --no-ignore are still ignored unless otherwise specified. /// The flag can be overridden with --no-hidden. #[arg( long, short = 'H', help = "Search hidden files and directories", long_help )] pub hidden: bool, /// Overrides --hidden #[arg(long, overrides_with = "hidden", hide = true, action = ArgAction::SetTrue)] no_hidden: (), /// Show search results from files and directories that would otherwise be /// ignored by '.gitignore', '.ignore', '.fdignore', or the global ignore file, /// The flag can be overridden with --ignore. #[arg( long, short = 'I', help = "Do not respect .(git|fd)ignore files", long_help )] pub no_ignore: bool, /// Overrides --no-ignore #[arg(long, overrides_with = "no_ignore", hide = true, action = ArgAction::SetTrue)] ignore: (), ///Show search results from files and directories that ///would otherwise be ignored by '.gitignore' files. ///The flag can be overridden with --ignore-vcs. #[arg( long, hide_short_help = true, help = "Do not respect .gitignore files", long_help )] pub no_ignore_vcs: bool, /// Overrides --no-ignore-vcs #[arg(long, overrides_with = "no_ignore_vcs", hide = true, action = ArgAction::SetTrue)] ignore_vcs: (), /// Do not require a git repository to respect gitignores. /// By default, fd will only respect global gitignore rules, .gitignore rules, /// and local exclude rules if fd detects that you are searching inside a /// git repository. This flag allows you to relax this restriction such that /// fd will respect all git related ignore rules regardless of whether you're /// searching in a git repository or not. /// /// /// This flag can be disabled with --require-git. #[arg( long, overrides_with = "require_git", hide_short_help = true, // same description as ripgrep's flag: ripgrep/crates/core/app.rs long_help )] pub no_require_git: bool, /// Overrides --no-require-git #[arg(long, overrides_with = "no_require_git", hide = true, action = ArgAction::SetTrue)] require_git: (), /// Show search results from files and directories that would otherwise be /// ignored by '.gitignore', '.ignore', or '.fdignore' files in parent directories. #[arg( long, hide_short_help = true, help = "Do not respect .(git|fd)ignore files in parent directories", long_help )] pub no_ignore_parent: bool, /// Do not respect the global ignore file #[arg(long, hide = true)] pub no_global_ignore_file: bool, /// Perform an unrestricted search, including ignored and hidden files. This is /// an alias for '--no-ignore --hidden'. #[arg(long = "unrestricted", short = 'u', overrides_with_all(&["ignore", "no_hidden"]), action(ArgAction::Count), hide_short_help = true, help = "Unrestricted search, alias for '--no-ignore --hidden'", long_help, )] rg_alias_hidden_ignore: u8, /// Case-sensitive search (default: smart case) #[arg( long, short = 's', overrides_with("ignore_case"), long_help = "Perform a case-sensitive search. By default, fd uses case-insensitive \ searches, unless the pattern contains an uppercase character (smart \ case)." )] pub case_sensitive: bool, /// Perform a case-insensitive search. By default, fd uses case-insensitive /// searches, unless the pattern contains an uppercase character (smart /// case). #[arg( long, short = 'i', overrides_with("case_sensitive"), help = "Case-insensitive search (default: smart case)", long_help )] pub ignore_case: bool, /// Perform a glob-based search instead of a regular expression search. #[arg( long, short = 'g', conflicts_with("fixed_strings"), help = "Glob-based search (default: regular expression)", long_help )] pub glob: bool, /// Perform a regular-expression based search (default). This can be used to /// override --glob. #[arg( long, overrides_with("glob"), hide_short_help = true, help = "Regular-expression based search (default)", long_help )] pub regex: bool, /// Treat the pattern as a literal string instead of a regular expression. Note /// that this also performs substring comparison. If you want to match on an /// exact filename, consider using '--glob'. #[arg( long, short = 'F', alias = "literal", hide_short_help = true, help = "Treat pattern as literal string stead of regex", long_help )] pub fixed_strings: bool, /// Add additional required search patterns, all of which must be matched. Multiple /// additional patterns can be specified. The patterns are regular /// expressions, unless '--glob' or '--fixed-strings' is used. #[arg( long = "and", value_name = "pattern", help = "Additional search patterns that need to be matched", long_help, hide_short_help = true, allow_hyphen_values = true )] pub exprs: Option<Vec<String>>, /// Shows the full path starting from the root as opposed to relative paths. /// The flag can be overridden with --relative-path. #[arg( long, short = 'a', help = "Show absolute instead of relative paths", long_help )] pub absolute_path: bool, /// Overrides --absolute-path #[arg(long, overrides_with = "absolute_path", hide = true, action = ArgAction::SetTrue)] relative_path: (), /// Use a detailed listing format like 'ls -l'. This is basically an alias /// for '--exec-batch ls -l' with some additional 'ls' options. This can be /// used to see more metadata, to show symlink targets and to achieve a /// deterministic sort order. #[arg( long, short = 'l', conflicts_with("absolute_path"), help = "Use a long listing format with file metadata", long_help )] pub list_details: bool, /// Follow symbolic links #[arg( long, short = 'L', alias = "dereference", long_help = "By default, fd does not descend into symlinked directories. Using this \ flag, symbolic links are also traversed. \ Flag can be overridden with --no-follow." )] pub follow: bool, /// Overrides --follow #[arg(long, overrides_with = "follow", hide = true, action = ArgAction::SetTrue)] no_follow: (), /// By default, the search pattern is only matched against the filename (or directory name). Using this flag, the pattern is matched against the full (absolute) path. Example: /// fd --glob -p '**/.git/config' #[arg( long, short = 'p', help = "Search full abs. path (default: filename only)", long_help, verbatim_doc_comment )] pub full_path: bool, /// Separate search results by the null character (instead of newlines). /// Useful for piping results to 'xargs'. #[arg( long = "print0", short = '0', conflicts_with("list_details"), hide_short_help = true, help = "Separate search results by the null character", long_help )] pub null_separator: bool, /// Limit the directory traversal to a given depth. By default, there is no /// limit on the search depth. #[arg( long, short = 'd', value_name = "depth", alias("maxdepth"), help = "Set maximum search depth (default: none)", long_help )] max_depth: Option<usize>, /// Only show search results starting at the given depth. /// See also: '--max-depth' and '--exact-depth' #[arg( long, value_name = "depth", hide_short_help = true, alias("mindepth"), help = "Only show search results starting at the given depth.", long_help )] min_depth: Option<usize>, /// Only show search results at the exact given depth. This is an alias for /// '--min-depth <depth> --max-depth <depth>'. #[arg(long, value_name = "depth", hide_short_help = true, conflicts_with_all(&["max_depth", "min_depth"]), help = "Only show search results at the exact given depth", long_help, )] exact_depth: Option<usize>, /// Exclude files/directories that match the given glob pattern. This /// overrides any other ignore logic. Multiple exclude patterns can be /// specified. /// /// Examples: /// {n} --exclude '*.pyc' /// {n} --exclude node_modules #[arg( long, short = 'E', value_name = "pattern", help = "Exclude entries that match the given glob pattern", long_help )] pub exclude: Vec<String>, /// Do not traverse into directories that match the search criteria. If /// you want to exclude specific directories, use the '--exclude=…' option. #[arg(long, hide_short_help = true, conflicts_with_all(&["size", "exact_depth"]), long_help, )] pub prune: bool, /// Filter the search by type: /// {n} 'f' or 'file': regular files /// {n} 'd' or 'dir' or 'directory': directories /// {n} 'l' or 'symlink': symbolic links /// {n} 's' or 'socket': socket /// {n} 'p' or 'pipe': named pipe (FIFO) /// {n} 'b' or 'block-device': block device /// {n} 'c' or 'char-device': character device /// {n}{n} 'x' or 'executable': executables /// {n} 'e' or 'empty': empty files or directories /// /// This option can be specified more than once to include multiple file types. /// Searching for '--type file --type symlink' will show both regular files as /// well as symlinks. Note that the 'executable' and 'empty' filters work differently: /// '--type executable' implies '--type file' by default. And '--type empty' searches /// for empty files and directories, unless either '--type file' or '--type directory' /// is specified in addition. /// /// Examples: /// {n} - Only search for files: /// {n} fd --type file … /// {n} fd -tf … /// {n} - Find both files and symlinks /// {n} fd --type file --type symlink … /// {n} fd -tf -tl … /// {n} - Find executable files: /// {n} fd --type executable /// {n} fd -tx /// {n} - Find empty files: /// {n} fd --type empty --type file /// {n} fd -te -tf /// {n} - Find empty directories: /// {n} fd --type empty --type directory /// {n} fd -te -td #[arg( long = "type", short = 't', value_name = "filetype", hide_possible_values = true, value_enum, help = "Filter by type: file (f), directory (d/dir), symlink (l), \ executable (x), empty (e), socket (s), pipe (p), \ char-device (c), block-device (b)", long_help )] pub filetype: Option<Vec<FileType>>, /// (Additionally) filter search results by their file extension. Multiple /// allowable file extensions can be specified. /// /// If you want to search for files without extension, /// you can use the regex '^[^.]+$' as a normal search pattern. #[arg( long = "extension", short = 'e', value_name = "ext", help = "Filter by file extension", long_help )] pub extensions: Option<Vec<String>>, /// Limit results based on the size of files using the format <+-><NUM><UNIT>. /// '+': file size must be greater than or equal to this /// '-': file size must be less than or equal to this /// /// If neither '+' nor '-' is specified, file size must be exactly equal to this. /// 'NUM': The numeric size (e.g. 500) /// 'UNIT': The units for NUM. They are not case-sensitive. /// Allowed unit values: /// 'b': bytes /// 'k': kilobytes (base ten, 10^3 = 1000 bytes) /// 'm': megabytes /// 'g': gigabytes /// 't': terabytes /// 'ki': kibibytes (base two, 2^10 = 1024 bytes) /// 'mi': mebibytes /// 'gi': gibibytes /// 'ti': tebibytes #[arg(long, short = 'S', value_parser = SizeFilter::from_string, allow_hyphen_values = true, verbatim_doc_comment, value_name = "size", help = "Limit results based on the size of files", long_help, verbatim_doc_comment, )] pub size: Vec<SizeFilter>, /// Filter results based on the file modification time. Files with modification times /// greater than the argument are returned. The argument can be provided /// as a specific point in time (YYYY-MM-DD HH:MM:SS or @timestamp) or as a duration (10h, 1d, 35min). /// If the time is not specified, it defaults to 00:00:00. /// '--change-newer-than', '--newer', or '--changed-after' can be used as aliases. /// /// Examples: /// {n} --changed-within 2weeks /// {n} --change-newer-than '2018-10-27 10:00:00' /// {n} --newer 2018-10-27 /// {n} --changed-after 1day #[arg( long, alias("change-newer-than"), alias("newer"), alias("changed-after"), value_name = "date|dur", help = "Filter by file modification time (newer than)", long_help )] pub changed_within: Option<String>, /// Filter results based on the file modification time. Files with modification times /// less than the argument are returned. The argument can be provided /// as a specific point in time (YYYY-MM-DD HH:MM:SS or @timestamp) or as a duration (10h, 1d, 35min). /// '--change-older-than' or '--older' can be used as aliases. /// /// Examples: /// {n} --changed-before '2018-10-27 10:00:00' /// {n} --change-older-than 2weeks /// {n} --older 2018-10-27 #[arg( long, alias("change-older-than"), alias("older"), value_name = "date|dur", help = "Filter by file modification time (older than)", long_help )] pub changed_before: Option<String>, /// Filter files by their user and/or group. /// Format: [(user|uid)][:(group|gid)]. Either side is optional. /// Precede either side with a '!' to exclude files instead. /// /// Examples: /// {n} --owner john /// {n} --owner :students /// {n} --owner '!john:students' #[cfg(unix)] #[arg(long, short = 'o', value_parser = OwnerFilter::from_string, value_name = "user:group", help = "Filter by owning user and/or group", long_help, )] pub owner: Option<OwnerFilter>, /// Instead of printing the file normally, print the format string with the following placeholders replaced: /// '{}': path (of the current search result) /// '{/}': basename /// '{//}': parent directory /// '{.}': path without file extension /// '{/.}': basename without file extension #[arg( long, value_name = "fmt", help = "Print results according to template", conflicts_with = "list_details" )] pub format: Option<String>, #[command(flatten)] pub exec: Exec, /// Maximum number of arguments to pass to the command given with -X. /// If the number of results is greater than the given size, /// the command given with -X is run again with remaining arguments. /// A batch size of zero means there is no limit (default), but note /// that batching might still happen due to OS restrictions on the /// maximum length of command lines. #[arg( long, value_name = "size", hide_short_help = true, requires("exec_batch"), value_parser = value_parser!(usize), default_value_t, help = "Max number of arguments to run as a batch size with -X", long_help, )] pub batch_size: usize, /// Add a custom ignore-file in '.gitignore' format. These files have a low precedence. #[arg( long, value_name = "path", hide_short_help = true, help = "Add a custom ignore-file in '.gitignore' format", long_help )] pub ignore_file: Vec<PathBuf>, /// Declare when to use color for the pattern match output #[arg( long, short = 'c', value_enum, default_value_t = ColorWhen::Auto, value_name = "when", help = "When to use colors", long_help, )] pub color: ColorWhen, /// Add a terminal hyperlink to a file:// url for each path in the output. /// /// Auto mode is used if no argument is given to this option. /// /// This doesn't do anything for --exec and --exec-batch. #[arg( long, alias = "hyper", value_name = "when", require_equals = true, value_enum, default_value_t = HyperlinkWhen::Never, default_missing_value = "auto", num_args = 0..=1, help = "Add hyperlinks to output paths" )] pub hyperlink: HyperlinkWhen, /// Set number of threads to use for searching & executing (default: number /// of available CPU cores) #[arg(long, short = 'j', value_name = "num", hide_short_help = true, value_parser = str::parse::<NonZeroUsize>)] pub threads: Option<NonZeroUsize>, /// Milliseconds to buffer before streaming search results to console /// /// Amount of time in milliseconds to buffer, before streaming the search /// results to the console. #[arg(long, hide = true, value_parser = parse_millis)] pub max_buffer_time: Option<Duration>, ///Limit the number of search results to 'count' and quit immediately. #[arg( long, value_name = "count", hide_short_help = true, overrides_with("max_one_result"), help = "Limit the number of search results", long_help )] max_results: Option<usize>, /// Limit the search to a single result and quit immediately. /// This is an alias for '--max-results=1'. #[arg( short = '1', hide_short_help = true, overrides_with("max_results"), help = "Limit search to a single result", long_help )] max_one_result: bool, /// When the flag is present, the program does not print anything and will /// return with an exit code of 0 if there is at least one match. Otherwise, the /// exit code will be 1. /// '--has-results' can be used as an alias. #[arg( long, short = 'q', alias = "has-results", hide_short_help = true, conflicts_with("max_results"), help = "Print nothing, exit code 0 if match found, 1 otherwise", long_help )] pub quiet: bool, /// Enable the display of filesystem errors for situations such as /// insufficient permissions or dead symlinks. #[arg( long, hide_short_help = true, help = "Show filesystem errors", long_help )] pub show_errors: bool, /// Change the current working directory of fd to the provided path. This /// means that search results will be shown with respect to the given base /// path. Note that relative paths which are passed to fd via the positional /// <path> argument or the '--search-path' option will also be resolved /// relative to this directory. #[arg( long, short = 'C', value_name = "path", hide_short_help = true, help = "Change current working directory", long_help )] pub base_directory: Option<PathBuf>, /// the search pattern which is either a regular expression (default) or a glob /// pattern (if --glob is used). If no pattern has been specified, every entry /// is considered a match. If your pattern starts with a dash (-), make sure to /// pass '--' first, or it will be considered as a flag (fd -- '-foo'). #[arg( default_value = "", hide_default_value = true, value_name = "pattern", help = "the search pattern (a regular expression, unless '--glob' is used; optional)", long_help )] pub pattern: String, /// Set the path separator to use when printing file paths. The default is /// the OS-specific separator ('/' on Unix, '\' on Windows). #[arg( long, value_name = "separator", hide_short_help = true, help = "Set path separator when printing file paths", long_help )] pub path_separator: Option<String>, /// The directory where the filesystem search is rooted (optional). If /// omitted, search the current working directory. #[arg(action = ArgAction::Append, value_name = "path", help = "the root directories for the filesystem search (optional)", long_help, )] path: Vec<PathBuf>, /// Provide paths to search as an alternative to the positional <path> /// argument. Changes the usage to `fd [OPTIONS] --search-path <path> /// --search-path <path2> [<pattern>]` #[arg( long, conflicts_with("path"), value_name = "search-path", hide_short_help = true, help = "Provides paths to search as an alternative to the positional <path> argument", long_help )] search_path: Vec<PathBuf>, /// By default, relative paths are prefixed with './' when -x/--exec, /// -X/--exec-batch, or -0/--print0 are given, to reduce the risk of a /// path starting with '-' being treated as a command line option. Use /// this flag to change this behavior. If this flag is used without a value, /// it is equivalent to passing "always". #[arg(long, conflicts_with_all(&["path", "search_path"]), value_name = "when", hide_short_help = true, require_equals = true, long_help)] strip_cwd_prefix: Option<Option<StripCwdWhen>>, /// By default, fd will traverse the file system tree as far as other options /// dictate. With this flag, fd ensures that it does not descend into a /// different file system than the one it started in. Comparable to the -mount /// or -xdev filters of find(1). #[cfg(any(unix, windows))] #[arg(long, aliases(&["mount", "xdev"]), hide_short_help = true, long_help)] pub one_file_system: bool, #[cfg(feature = "completions")] #[arg(long, hide = true, exclusive = true)] gen_completions: Option<Option<Shell>>, } impl Opts { pub fn search_paths(&self) -> anyhow::Result<Vec<PathBuf>> { // would it make sense to concatenate these? let paths = if !self.path.is_empty() { &self.path } else if !self.search_path.is_empty() { &self.search_path } else { let current_directory = Path::new("./"); ensure_current_directory_exists(current_directory)?; return Ok(vec![self.normalize_path(current_directory)]); }; Ok(paths .iter() .filter_map(|path| { if filesystem::is_existing_directory(path) { Some(self.normalize_path(path)) } else { print_error(format!( "Search path '{}' is not a directory.", path.to_string_lossy() )); None } }) .collect()) } fn normalize_path(&self, path: &Path) -> PathBuf { if self.absolute_path { filesystem::absolute_path(path.normalize().unwrap().as_path()).unwrap() } else if path == Path::new(".") { // Change "." to "./" as a workaround for https://github.com/BurntSushi/ripgrep/pull/2711 PathBuf::from("./") } else { path.to_path_buf() } } pub fn no_search_paths(&self) -> bool { self.path.is_empty() && self.search_path.is_empty() } #[inline] pub fn rg_alias_ignore(&self) -> bool { self.rg_alias_hidden_ignore > 0 } pub fn max_depth(&self) -> Option<usize> { self.max_depth.or(self.exact_depth) } pub fn min_depth(&self) -> Option<usize> { self.min_depth.or(self.exact_depth) } pub fn threads(&self) -> NonZeroUsize { self.threads.unwrap_or_else(default_num_threads) } pub fn max_results(&self) -> Option<usize> { self.max_results .filter(|&m| m > 0) .or_else(|| self.max_one_result.then_some(1)) } pub fn strip_cwd_prefix<P: FnOnce() -> bool>(&self, auto_pred: P) -> bool { use self::StripCwdWhen::*; self.no_search_paths() && match self.strip_cwd_prefix.map_or(Auto, |o| o.unwrap_or(Always)) { Auto => auto_pred(), Always => true, Never => false, } } #[cfg(feature = "completions")] pub fn gen_completions(&self) -> anyhow::Result<Option<Shell>> { self.gen_completions .map(|maybe_shell| match maybe_shell { Some(sh) => Ok(sh), None => { Shell::from_env().ok_or_else(|| anyhow!("Unable to get shell from environment")) } }) .transpose() } } /// Get the default number of threads to use, if not explicitly specified. fn default_num_threads() -> NonZeroUsize { // If we can't get the amount of parallelism for some reason, then // default to a single thread, because that is safe. let fallback = NonZeroUsize::MIN; // To limit startup overhead on massively parallel machines, don't use more // than 64 threads. let limit = NonZeroUsize::new(64).unwrap(); std::thread::available_parallelism() .unwrap_or(fallback) .min(limit) } #[derive(Copy, Clone, PartialEq, Eq, ValueEnum)] pub enum FileType { #[value(alias = "f")] File, #[value(alias = "d", alias = "dir")] Directory, #[value(alias = "l")] Symlink, #[value(alias = "b")] BlockDevice, #[value(alias = "c")] CharDevice, /// A file which is executable by the current effective user #[value(alias = "x")] Executable, #[value(alias = "e")] Empty, #[value(alias = "s")] Socket, #[value(alias = "p")] Pipe, } #[derive(Copy, Clone, PartialEq, Eq, Debug, ValueEnum)] pub enum ColorWhen { /// show colors if the output goes to an interactive console (default) Auto, /// always use colorized output Always, /// do not use colorized output Never, } #[derive(Copy, Clone, PartialEq, Eq, Debug, ValueEnum)] pub enum StripCwdWhen { /// Use the default behavior Auto, /// Always strip the ./ at the beginning of paths Always, /// Never strip the ./ Never, } #[derive(Copy, Clone, PartialEq, Eq, Debug, ValueEnum)] pub enum HyperlinkWhen { /// Use hyperlinks only if color is enabled Auto, /// Always use hyperlinks when printing file paths Always, /// Never use hyperlinks Never, } // there isn't a derive api for getting grouped values yet, // so we have to use hand-rolled parsing for exec and exec-batch pub struct Exec { pub command: Option<CommandSet>, } impl clap::FromArgMatches for Exec { fn from_arg_matches(matches: &ArgMatches) -> clap::error::Result<Self> { let command = matches .get_occurrences::<String>("exec") .map(CommandSet::new) .or_else(|| { matches .get_occurrences::<String>("exec_batch") .map(CommandSet::new_batch) }) .transpose() .map_err(|e| clap::Error::raw(ErrorKind::InvalidValue, e))?; Ok(Exec { command }) } fn update_from_arg_matches(&mut self, matches: &ArgMatches) -> clap::error::Result<()> { *self = Self::from_arg_matches(matches)?; Ok(()) } } impl clap::Args for Exec { fn augment_args(cmd: Command) -> Command { cmd.arg(Arg::new("exec") .action(ArgAction::Append) .long("exec") .short('x') .num_args(1..) .allow_hyphen_values(true) .value_terminator(";") .value_name("cmd") .conflicts_with("list_details") .help("Execute a command for each search result") .long_help( "Execute a command for each search result in parallel (use --threads=1 for sequential command execution). \ There is no guarantee of the order commands are executed in, and the order should not be depended upon. \ All positional arguments following --exec are considered to be arguments to the command - not to fd. \ It is therefore recommended to place the '-x'/'--exec' option last.\n\ The following placeholders are substituted before the command is executed:\n \ '{}': path (of the current search result)\n \ '{/}': basename\n \ '{//}': parent directory\n \ '{.}': path without file extension\n \ '{/.}': basename without file extension\n \ '{{': literal '{' (for escaping)\n \ '}}': literal '}' (for escaping)\n\n\ If no placeholder is present, an implicit \"{}\" at the end is assumed.\n\n\ Examples:\n\n \ - find all *.zip files and unzip them:\n\n \ fd -e zip -x unzip\n\n \ - find *.h and *.cpp files and run \"clang-format -i ..\" for each of them:\n\n \ fd -e h -e cpp -x clang-format -i\n\n \ - Convert all *.jpg files to *.png files:\n\n \ fd -e jpg -x convert {} {.}.png\ ", ), ) .arg( Arg::new("exec_batch") .action(ArgAction::Append) .long("exec-batch") .short('X') .num_args(1..) .allow_hyphen_values(true) .value_terminator(";") .value_name("cmd") .conflicts_with_all(["exec", "list_details"]) .help("Execute a command with all search results at once") .long_help( "Execute the given command once, with all search results as arguments.\n\ The order of the arguments is non-deterministic, and should not be relied upon.\n\ One of the following placeholders is substituted before the command is executed:\n \ '{}': path (of all search results)\n \ '{/}': basename\n \ '{//}': parent directory\n \ '{.}': path without file extension\n \
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
true
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/error.rs
src/error.rs
pub fn print_error(msg: impl Into<String>) { eprintln!("[fd error]: {}", msg.into()); }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/walk.rs
src/walk.rs
use std::borrow::Cow; use std::ffi::OsStr; use std::io::{self, Write}; use std::mem; use std::path::PathBuf; use std::sync::atomic::{AtomicBool, Ordering}; use std::sync::{Arc, Mutex, MutexGuard}; use std::thread; use std::time::{Duration, Instant}; use anyhow::{Result, anyhow}; use crossbeam_channel::{Receiver, RecvTimeoutError, SendError, Sender, bounded}; use etcetera::BaseStrategy; use ignore::overrides::{Override, OverrideBuilder}; use ignore::{WalkBuilder, WalkParallel, WalkState}; use regex::bytes::Regex; use crate::config::Config; use crate::dir_entry::DirEntry; use crate::error::print_error; use crate::exec; use crate::exit_codes::{ExitCode, merge_exitcodes}; use crate::filesystem; use crate::output; /// The receiver thread can either be buffering results or directly streaming to the console. #[derive(PartialEq)] enum ReceiverMode { /// Receiver is still buffering in order to sort the results, if the search finishes fast /// enough. Buffering, /// Receiver is directly printing results to the output. Streaming, } /// The Worker threads can result in a valid entry having PathBuf or an error. #[allow(clippy::large_enum_variant)] #[derive(Debug)] pub enum WorkerResult { // Errors should be rare, so it's probably better to allow large_enum_variant than // to box the Entry variant Entry(DirEntry), Error(ignore::Error), } /// A batch of WorkerResults to send over a channel. #[derive(Clone)] struct Batch { items: Arc<Mutex<Option<Vec<WorkerResult>>>>, } impl Batch { fn new() -> Self { Self { items: Arc::new(Mutex::new(Some(vec![]))), } } fn lock(&self) -> MutexGuard<'_, Option<Vec<WorkerResult>>> { self.items.lock().unwrap() } } impl IntoIterator for Batch { type Item = WorkerResult; type IntoIter = std::vec::IntoIter<WorkerResult>; fn into_iter(self) -> Self::IntoIter { self.lock().take().unwrap().into_iter() } } /// Wrapper that sends batches of items at once over a channel. struct BatchSender { batch: Batch, tx: Sender<Batch>, limit: usize, } impl BatchSender { fn new(tx: Sender<Batch>, limit: usize) -> Self { Self { batch: Batch::new(), tx, limit, } } /// Check if we need to flush a batch. fn needs_flush(&self, batch: Option<&Vec<WorkerResult>>) -> bool { match batch { // Limit the batch size to provide some backpressure Some(vec) => vec.len() >= self.limit, // Batch was already taken by the receiver, so make a new one None => true, } } /// Add an item to a batch. fn send(&mut self, item: WorkerResult) -> Result<(), SendError<()>> { let mut batch = self.batch.lock(); if self.needs_flush(batch.as_ref()) { drop(batch); self.batch = Batch::new(); batch = self.batch.lock(); } let items = batch.as_mut().unwrap(); items.push(item); if items.len() == 1 { // New batch, send it over the channel self.tx .send(self.batch.clone()) .map_err(|_| SendError(()))?; } Ok(()) } } /// Maximum size of the output buffer before flushing results to the console const MAX_BUFFER_LENGTH: usize = 1000; /// Default duration until output buffering switches to streaming. const DEFAULT_MAX_BUFFER_TIME: Duration = Duration::from_millis(100); /// Wrapper for the receiver thread's buffering behavior. struct ReceiverBuffer<'a, W> { /// The configuration. config: &'a Config, /// For shutting down the senders. quit_flag: &'a AtomicBool, /// The ^C notifier. interrupt_flag: &'a AtomicBool, /// Receiver for worker results. rx: Receiver<Batch>, /// Standard output. stdout: W, /// The current buffer mode. mode: ReceiverMode, /// The deadline to switch to streaming mode. deadline: Instant, /// The buffer of quickly received paths. buffer: Vec<DirEntry>, /// Result count. num_results: usize, } impl<'a, W: Write> ReceiverBuffer<'a, W> { /// Create a new receiver buffer. fn new(state: &'a WorkerState, rx: Receiver<Batch>, stdout: W) -> Self { let config = &state.config; let quit_flag = state.quit_flag.as_ref(); let interrupt_flag = state.interrupt_flag.as_ref(); let max_buffer_time = config.max_buffer_time.unwrap_or(DEFAULT_MAX_BUFFER_TIME); let deadline = Instant::now() + max_buffer_time; Self { config, quit_flag, interrupt_flag, rx, stdout, mode: ReceiverMode::Buffering, deadline, buffer: Vec::with_capacity(MAX_BUFFER_LENGTH), num_results: 0, } } /// Process results until finished. fn process(&mut self) -> ExitCode { loop { if let Err(ec) = self.poll() { self.quit_flag.store(true, Ordering::Relaxed); return ec; } } } /// Receive the next worker result. fn recv(&self) -> Result<Batch, RecvTimeoutError> { match self.mode { ReceiverMode::Buffering => { // Wait at most until we should switch to streaming self.rx.recv_deadline(self.deadline) } ReceiverMode::Streaming => { // Wait however long it takes for a result Ok(self.rx.recv()?) } } } /// Wait for a result or state change. fn poll(&mut self) -> Result<(), ExitCode> { match self.recv() { Ok(batch) => { for result in batch { match result { WorkerResult::Entry(dir_entry) => { if self.config.quiet { return Err(ExitCode::HasResults(true)); } match self.mode { ReceiverMode::Buffering => { self.buffer.push(dir_entry); if self.buffer.len() > MAX_BUFFER_LENGTH { self.stream()?; } } ReceiverMode::Streaming => { self.print(&dir_entry)?; } } self.num_results += 1; if let Some(max_results) = self.config.max_results && self.num_results >= max_results { return self.stop(); } } WorkerResult::Error(err) => { if self.config.show_filesystem_errors { print_error(err.to_string()); } } } } // If we don't have another batch ready, flush before waiting if self.mode == ReceiverMode::Streaming && self.rx.is_empty() { self.flush()?; } } Err(RecvTimeoutError::Timeout) => { self.stream()?; } Err(RecvTimeoutError::Disconnected) => { return self.stop(); } } Ok(()) } /// Output a path. fn print(&mut self, entry: &DirEntry) -> Result<(), ExitCode> { if let Err(e) = output::print_entry(&mut self.stdout, entry, self.config) && e.kind() != ::std::io::ErrorKind::BrokenPipe { print_error(format!("Could not write to output: {e}")); return Err(ExitCode::GeneralError); } if self.interrupt_flag.load(Ordering::Relaxed) { // Ignore any errors on flush, because we're about to exit anyway let _ = self.flush(); return Err(ExitCode::KilledBySigint); } Ok(()) } /// Switch ourselves into streaming mode. fn stream(&mut self) -> Result<(), ExitCode> { self.mode = ReceiverMode::Streaming; let buffer = mem::take(&mut self.buffer); for path in buffer { self.print(&path)?; } self.flush() } /// Stop looping. fn stop(&mut self) -> Result<(), ExitCode> { if self.mode == ReceiverMode::Buffering { self.buffer.sort(); self.stream()?; } if self.config.quiet { Err(ExitCode::HasResults(self.num_results > 0)) } else { Err(ExitCode::Success) } } /// Flush stdout if necessary. fn flush(&mut self) -> Result<(), ExitCode> { if self.stdout.flush().is_err() { // Probably a broken pipe. Exit gracefully. return Err(ExitCode::GeneralError); } Ok(()) } } /// State shared by the sender and receiver threads. struct WorkerState { /// The search patterns. patterns: Vec<Regex>, /// The command line configuration. config: Config, /// Flag for cleanly shutting down the parallel walk quit_flag: Arc<AtomicBool>, /// Flag specifically for quitting due to ^C interrupt_flag: Arc<AtomicBool>, } impl WorkerState { fn new(patterns: Vec<Regex>, config: Config) -> Self { let quit_flag = Arc::new(AtomicBool::new(false)); let interrupt_flag = Arc::new(AtomicBool::new(false)); Self { patterns, config, quit_flag, interrupt_flag, } } fn build_overrides(&self, paths: &[PathBuf]) -> Result<Override> { let first_path = &paths[0]; let config = &self.config; let mut builder = OverrideBuilder::new(first_path); for pattern in &config.exclude_patterns { builder .add(pattern) .map_err(|e| anyhow!("Malformed exclude pattern: {}", e))?; } builder .build() .map_err(|_| anyhow!("Mismatch in exclude patterns")) } fn build_walker(&self, paths: &[PathBuf]) -> Result<WalkParallel> { let first_path = &paths[0]; let config = &self.config; let overrides = self.build_overrides(paths)?; let mut builder = WalkBuilder::new(first_path); builder .hidden(config.ignore_hidden) .ignore(config.read_fdignore) .parents(config.read_parent_ignore && (config.read_fdignore || config.read_vcsignore)) .git_ignore(config.read_vcsignore) .git_global(config.read_vcsignore) .git_exclude(config.read_vcsignore) .require_git(config.require_git_to_read_vcsignore) .overrides(overrides) .follow_links(config.follow_links) // No need to check for supported platforms, option is unavailable on unsupported ones .same_file_system(config.one_file_system) .max_depth(config.max_depth); if config.read_fdignore { builder.add_custom_ignore_filename(".fdignore"); } if config.read_global_ignore && let Ok(basedirs) = etcetera::choose_base_strategy() { let global_ignore_file = basedirs.config_dir().join("fd").join("ignore"); if global_ignore_file.is_file() { let result = builder.add_ignore(global_ignore_file); match result { Some(ignore::Error::Partial(_)) => (), Some(err) => { print_error(format!("Malformed pattern in global ignore file. {err}.")); } None => (), } } } for ignore_file in &config.ignore_files { let result = builder.add_ignore(ignore_file); match result { Some(ignore::Error::Partial(_)) => (), Some(err) => { print_error(format!("Malformed pattern in custom ignore file. {err}.")); } None => (), } } for path in &paths[1..] { builder.add(path); } let walker = builder.threads(config.threads).build_parallel(); Ok(walker) } /// Run the receiver work, either on this thread or a pool of background /// threads (for --exec). fn receive(&self, rx: Receiver<Batch>) -> ExitCode { let config = &self.config; // This will be set to `Some` if the `--exec` argument was supplied. if let Some(ref cmd) = config.command { if cmd.in_batch_mode() { exec::batch(rx.into_iter().flatten(), cmd, config) } else { thread::scope(|scope| { // Each spawned job will store its thread handle in here. let threads = config.threads; let mut handles = Vec::with_capacity(threads); for _ in 0..threads { let rx = rx.clone(); // Spawn a job thread that will listen for and execute inputs. let handle = scope.spawn(|| exec::job(rx.into_iter().flatten(), cmd, config)); // Push the handle of the spawned thread into the vector for later joining. handles.push(handle); } let exit_codes = handles.into_iter().map(|handle| handle.join().unwrap()); merge_exitcodes(exit_codes) }) } } else { let stdout = io::stdout().lock(); let stdout = io::BufWriter::new(stdout); ReceiverBuffer::new(self, rx, stdout).process() } } /// Spawn the sender threads. fn spawn_senders(&self, walker: WalkParallel, tx: Sender<Batch>) { walker.run(|| { let patterns = &self.patterns; let config = &self.config; let quit_flag = self.quit_flag.as_ref(); let mut limit = 0x100; if let Some(cmd) = &config.command && !cmd.in_batch_mode() && config.threads > 1 { // Evenly distribute work between multiple receivers limit = 1; } let mut tx = BatchSender::new(tx.clone(), limit); Box::new(move |entry| { if quit_flag.load(Ordering::Relaxed) { return WalkState::Quit; } let entry = match entry { Ok(ref e) if e.depth() == 0 => { // Skip the root directory entry. return WalkState::Continue; } Ok(e) => DirEntry::normal(e), Err(ignore::Error::WithPath { path, err: inner_err, }) => match inner_err.as_ref() { ignore::Error::Io(io_error) if io_error.kind() == io::ErrorKind::NotFound && path .symlink_metadata() .ok() .is_some_and(|m| m.file_type().is_symlink()) => { DirEntry::broken_symlink(path) } _ => { return match tx.send(WorkerResult::Error(ignore::Error::WithPath { path, err: inner_err, })) { Ok(_) => WalkState::Continue, Err(_) => WalkState::Quit, }; } }, Err(err) => { return match tx.send(WorkerResult::Error(err)) { Ok(_) => WalkState::Continue, Err(_) => WalkState::Quit, }; } }; if let Some(min_depth) = config.min_depth && entry.depth().is_none_or(|d| d < min_depth) { return WalkState::Continue; } // Check the name first, since it doesn't require metadata let entry_path = entry.path(); let search_str: Cow<OsStr> = if config.search_full_path { let path_abs_buf = filesystem::path_absolute_form(entry_path) .expect("Retrieving absolute path succeeds"); Cow::Owned(path_abs_buf.as_os_str().to_os_string()) } else { match entry_path.file_name() { Some(filename) => Cow::Borrowed(filename), None => unreachable!( "Encountered file system entry without a file name. This should only \ happen for paths like 'foo/bar/..' or '/' which are not supposed to \ appear in a file system traversal." ), } }; if !patterns .iter() .all(|pat| pat.is_match(&filesystem::osstr_to_bytes(search_str.as_ref()))) { return WalkState::Continue; } // Filter out unwanted extensions. if let Some(ref exts_regex) = config.extensions { if let Some(path_str) = entry_path.file_name() { if !exts_regex.is_match(&filesystem::osstr_to_bytes(path_str)) { return WalkState::Continue; } } else { return WalkState::Continue; } } // Filter out unwanted file types. if let Some(ref file_types) = config.file_types && file_types.should_ignore(&entry) { return WalkState::Continue; } #[cfg(unix)] { if let Some(ref owner_constraint) = config.owner_constraint { if let Some(metadata) = entry.metadata() { if !owner_constraint.matches(metadata) { return WalkState::Continue; } } else { return WalkState::Continue; } } } // Filter out unwanted sizes if it is a file and we have been given size constraints. if !config.size_constraints.is_empty() { if entry_path.is_file() { if let Some(metadata) = entry.metadata() { let file_size = metadata.len(); if config .size_constraints .iter() .any(|sc| !sc.is_within(file_size)) { return WalkState::Continue; } } else { return WalkState::Continue; } } else { return WalkState::Continue; } } // Filter out unwanted modification times if !config.time_constraints.is_empty() { let mut matched = false; if let Some(metadata) = entry.metadata() && let Ok(modified) = metadata.modified() { matched = config .time_constraints .iter() .all(|tf| tf.applies_to(&modified)); } if !matched { return WalkState::Continue; } } if config.is_printing() && let Some(ls_colors) = &config.ls_colors { // Compute colors in parallel entry.style(ls_colors); } let send_result = tx.send(WorkerResult::Entry(entry)); if send_result.is_err() { return WalkState::Quit; } // Apply pruning. if config.prune { return WalkState::Skip; } WalkState::Continue }) }); } /// Perform the recursive scan. fn scan(&self, paths: &[PathBuf]) -> Result<ExitCode> { let config = &self.config; let walker = self.build_walker(paths)?; if config.ls_colors.is_some() && config.is_printing() { let quit_flag = Arc::clone(&self.quit_flag); let interrupt_flag = Arc::clone(&self.interrupt_flag); ctrlc::set_handler(move || { quit_flag.store(true, Ordering::Relaxed); if interrupt_flag.fetch_or(true, Ordering::Relaxed) { // Ctrl-C has been pressed twice, exit NOW ExitCode::KilledBySigint.exit(); } }) .unwrap(); } let (tx, rx) = bounded(2 * config.threads); let exit_code = thread::scope(|scope| { // Spawn the receiver thread(s) let receiver = scope.spawn(|| self.receive(rx)); // Spawn the sender threads. self.spawn_senders(walker, tx); receiver.join().unwrap() }); if self.interrupt_flag.load(Ordering::Relaxed) { Ok(ExitCode::KilledBySigint) } else { Ok(exit_code) } } } /// Recursively scan the given search path for files / pathnames matching the patterns. /// /// If the `--exec` argument was supplied, this will create a thread pool for executing /// jobs in parallel from a given command line and the discovered paths. Otherwise, each /// path will simply be written to standard output. pub fn scan(paths: &[PathBuf], patterns: Vec<Regex>, config: Config) -> Result<ExitCode> { WorkerState::new(patterns, config).scan(paths) }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/dir_entry.rs
src/dir_entry.rs
use std::cell::OnceCell; use std::ffi::OsString; use std::fs::{FileType, Metadata}; use std::path::{Path, PathBuf}; use lscolors::{Colorable, LsColors, Style}; use crate::config::Config; use crate::filesystem::strip_current_dir; #[derive(Debug)] enum DirEntryInner { Normal(ignore::DirEntry), BrokenSymlink(PathBuf), } #[derive(Debug)] pub struct DirEntry { inner: DirEntryInner, metadata: OnceCell<Option<Metadata>>, style: OnceCell<Option<Style>>, } impl DirEntry { #[inline] pub fn normal(e: ignore::DirEntry) -> Self { Self { inner: DirEntryInner::Normal(e), metadata: OnceCell::new(), style: OnceCell::new(), } } pub fn broken_symlink(path: PathBuf) -> Self { Self { inner: DirEntryInner::BrokenSymlink(path), metadata: OnceCell::new(), style: OnceCell::new(), } } pub fn path(&self) -> &Path { match &self.inner { DirEntryInner::Normal(e) => e.path(), DirEntryInner::BrokenSymlink(pathbuf) => pathbuf.as_path(), } } pub fn into_path(self) -> PathBuf { match self.inner { DirEntryInner::Normal(e) => e.into_path(), DirEntryInner::BrokenSymlink(p) => p, } } /// Returns the path as it should be presented to the user. pub fn stripped_path(&self, config: &Config) -> &Path { if config.strip_cwd_prefix { strip_current_dir(self.path()) } else { self.path() } } /// Returns the path as it should be presented to the user. pub fn into_stripped_path(self, config: &Config) -> PathBuf { if config.strip_cwd_prefix { self.stripped_path(config).to_path_buf() } else { self.into_path() } } pub fn file_type(&self) -> Option<FileType> { match &self.inner { DirEntryInner::Normal(e) => e.file_type(), DirEntryInner::BrokenSymlink(_) => self.metadata().map(|m| m.file_type()), } } pub fn metadata(&self) -> Option<&Metadata> { self.metadata .get_or_init(|| match &self.inner { DirEntryInner::Normal(e) => e.metadata().ok(), DirEntryInner::BrokenSymlink(path) => path.symlink_metadata().ok(), }) .as_ref() } pub fn depth(&self) -> Option<usize> { match &self.inner { DirEntryInner::Normal(e) => Some(e.depth()), DirEntryInner::BrokenSymlink(_) => None, } } pub fn style(&self, ls_colors: &LsColors) -> Option<&Style> { self.style .get_or_init(|| ls_colors.style_for(self).cloned()) .as_ref() } } impl PartialEq for DirEntry { #[inline] fn eq(&self, other: &Self) -> bool { self.path() == other.path() } } impl Eq for DirEntry {} impl PartialOrd for DirEntry { #[inline] fn partial_cmp(&self, other: &Self) -> Option<std::cmp::Ordering> { Some(self.cmp(other)) } } impl Ord for DirEntry { #[inline] fn cmp(&self, other: &Self) -> std::cmp::Ordering { self.path().cmp(other.path()) } } impl Colorable for DirEntry { fn path(&self) -> PathBuf { self.path().to_owned() } fn file_name(&self) -> OsString { let name = match &self.inner { DirEntryInner::Normal(e) => e.file_name(), DirEntryInner::BrokenSymlink(path) => { // Path::file_name() only works if the last component is Normal, // but we want it for all component types, so we open code it. // Copied from LsColors::style_for_path_with_metadata(). path.components() .next_back() .map(|c| c.as_os_str()) .unwrap_or_else(|| path.as_os_str()) } }; name.to_owned() } fn file_type(&self) -> Option<FileType> { self.file_type() } fn metadata(&self) -> Option<Metadata> { self.metadata().cloned() } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/exit_codes.rs
src/exit_codes.rs
use std::process; #[cfg(unix)] use nix::sys::signal::{SigHandler, Signal, raise, signal}; #[derive(Debug, Clone, Copy, PartialEq, Eq)] pub enum ExitCode { Success, HasResults(bool), GeneralError, KilledBySigint, } impl From<ExitCode> for i32 { fn from(code: ExitCode) -> Self { match code { ExitCode::Success => 0, ExitCode::HasResults(has_results) => !has_results as i32, ExitCode::GeneralError => 1, ExitCode::KilledBySigint => 130, } } } impl ExitCode { fn is_error(self) -> bool { i32::from(self) != 0 } /// Exit the process with the appropriate code. pub fn exit(self) -> ! { #[cfg(unix)] if self == ExitCode::KilledBySigint { // Get rid of the SIGINT handler, if present, and raise SIGINT unsafe { if signal(Signal::SIGINT, SigHandler::SigDfl).is_ok() { let _ = raise(Signal::SIGINT); } } } process::exit(self.into()) } } pub fn merge_exitcodes(results: impl IntoIterator<Item = ExitCode>) -> ExitCode { if results.into_iter().any(ExitCode::is_error) { return ExitCode::GeneralError; } ExitCode::Success } #[cfg(test)] mod tests { use super::*; #[test] fn success_when_no_results() { assert_eq!(merge_exitcodes([]), ExitCode::Success); } #[test] fn general_error_if_at_least_one_error() { assert_eq!( merge_exitcodes([ExitCode::GeneralError]), ExitCode::GeneralError ); assert_eq!( merge_exitcodes([ExitCode::KilledBySigint]), ExitCode::GeneralError ); assert_eq!( merge_exitcodes([ExitCode::KilledBySigint, ExitCode::Success]), ExitCode::GeneralError ); assert_eq!( merge_exitcodes([ExitCode::Success, ExitCode::GeneralError]), ExitCode::GeneralError ); assert_eq!( merge_exitcodes([ExitCode::GeneralError, ExitCode::KilledBySigint]), ExitCode::GeneralError ); } #[test] fn success_if_no_error() { assert_eq!(merge_exitcodes([ExitCode::Success]), ExitCode::Success); assert_eq!( merge_exitcodes([ExitCode::Success, ExitCode::Success]), ExitCode::Success ); } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/main.rs
src/main.rs
mod cli; mod config; mod dir_entry; mod error; mod exec; mod exit_codes; mod filesystem; mod filetypes; mod filter; mod fmt; mod hyperlink; mod output; mod regex_helper; mod walk; use std::env; use std::io::IsTerminal; use std::path::Path; use std::sync::Arc; use anyhow::{Context, Result, anyhow, bail}; use clap::{CommandFactory, Parser}; use globset::GlobBuilder; use lscolors::LsColors; use regex::bytes::{Regex, RegexBuilder, RegexSetBuilder}; use crate::cli::{ColorWhen, HyperlinkWhen, Opts}; use crate::config::Config; use crate::exec::CommandSet; use crate::exit_codes::ExitCode; use crate::filetypes::FileTypes; #[cfg(unix)] use crate::filter::OwnerFilter; use crate::filter::TimeFilter; use crate::regex_helper::{pattern_has_uppercase_char, pattern_matches_strings_with_leading_dot}; // We use jemalloc for performance reasons, see https://github.com/sharkdp/fd/pull/481 // FIXME: re-enable jemalloc on macOS, see comment in Cargo.toml file for more infos // This has to be kept in sync with the Cargo.toml file section that declares a // dependency on tikv-jemallocator. #[cfg(all( not(windows), not(target_os = "android"), not(target_os = "macos"), not(target_os = "freebsd"), not(target_os = "openbsd"), not(target_os = "illumos"), not(all(target_env = "musl", target_pointer_width = "32")), not(target_arch = "riscv64"), feature = "use-jemalloc" ))] #[global_allocator] static ALLOC: tikv_jemallocator::Jemalloc = tikv_jemallocator::Jemalloc; // vivid --color-mode 8-bit generate molokai const DEFAULT_LS_COLORS: &str = " ow=0:or=0;38;5;16;48;5;203:no=0:ex=1;38;5;203:cd=0;38;5;203;48;5;236:mi=0;38;5;16;48;5;203:*~=0;38;5;243:st=0:pi=0;38;5;16;48;5;81:fi=0:di=0;38;5;81:so=0;38;5;16;48;5;203:bd=0;38;5;81;48;5;236:tw=0:ln=0;38;5;203:*.m=0;38;5;48:*.o=0;38;5;243:*.z=4;38;5;203:*.a=1;38;5;203:*.r=0;38;5;48:*.c=0;38;5;48:*.d=0;38;5;48:*.t=0;38;5;48:*.h=0;38;5;48:*.p=0;38;5;48:*.cc=0;38;5;48:*.ll=0;38;5;48:*.jl=0;38;5;48:*css=0;38;5;48:*.md=0;38;5;185:*.gz=4;38;5;203:*.nb=0;38;5;48:*.mn=0;38;5;48:*.go=0;38;5;48:*.xz=4;38;5;203:*.so=1;38;5;203:*.rb=0;38;5;48:*.pm=0;38;5;48:*.bc=0;38;5;243:*.py=0;38;5;48:*.as=0;38;5;48:*.pl=0;38;5;48:*.rs=0;38;5;48:*.sh=0;38;5;48:*.7z=4;38;5;203:*.ps=0;38;5;186:*.cs=0;38;5;48:*.el=0;38;5;48:*.rm=0;38;5;208:*.hs=0;38;5;48:*.td=0;38;5;48:*.ui=0;38;5;149:*.ex=0;38;5;48:*.js=0;38;5;48:*.cp=0;38;5;48:*.cr=0;38;5;48:*.la=0;38;5;243:*.kt=0;38;5;48:*.ml=0;38;5;48:*.vb=0;38;5;48:*.gv=0;38;5;48:*.lo=0;38;5;243:*.hi=0;38;5;243:*.ts=0;38;5;48:*.ko=1;38;5;203:*.hh=0;38;5;48:*.pp=0;38;5;48:*.di=0;38;5;48:*.bz=4;38;5;203:*.fs=0;38;5;48:*.png=0;38;5;208:*.zsh=0;38;5;48:*.mpg=0;38;5;208:*.pid=0;38;5;243:*.xmp=0;38;5;149:*.iso=4;38;5;203:*.m4v=0;38;5;208:*.dot=0;38;5;48:*.ods=0;38;5;186:*.inc=0;38;5;48:*.sxw=0;38;5;186:*.aif=0;38;5;208:*.git=0;38;5;243:*.gvy=0;38;5;48:*.tbz=4;38;5;203:*.log=0;38;5;243:*.txt=0;38;5;185:*.ico=0;38;5;208:*.csx=0;38;5;48:*.vob=0;38;5;208:*.pgm=0;38;5;208:*.pps=0;38;5;186:*.ics=0;38;5;186:*.img=4;38;5;203:*.fon=0;38;5;208:*.hpp=0;38;5;48:*.bsh=0;38;5;48:*.sql=0;38;5;48:*TODO=1:*.php=0;38;5;48:*.pkg=4;38;5;203:*.ps1=0;38;5;48:*.csv=0;38;5;185:*.ilg=0;38;5;243:*.ini=0;38;5;149:*.pyc=0;38;5;243:*.psd=0;38;5;208:*.htc=0;38;5;48:*.swp=0;38;5;243:*.mli=0;38;5;48:*hgrc=0;38;5;149:*.bst=0;38;5;149:*.ipp=0;38;5;48:*.fsi=0;38;5;48:*.tcl=0;38;5;48:*.exs=0;38;5;48:*.out=0;38;5;243:*.jar=4;38;5;203:*.xls=0;38;5;186:*.ppm=0;38;5;208:*.apk=4;38;5;203:*.aux=0;38;5;243:*.rpm=4;38;5;203:*.dll=1;38;5;203:*.eps=0;38;5;208:*.exe=1;38;5;203:*.doc=0;38;5;186:*.wma=0;38;5;208:*.deb=4;38;5;203:*.pod=0;38;5;48:*.ind=0;38;5;243:*.nix=0;38;5;149:*.lua=0;38;5;48:*.epp=0;38;5;48:*.dpr=0;38;5;48:*.htm=0;38;5;185:*.ogg=0;38;5;208:*.bin=4;38;5;203:*.otf=0;38;5;208:*.yml=0;38;5;149:*.pro=0;38;5;149:*.cxx=0;38;5;48:*.tex=0;38;5;48:*.fnt=0;38;5;208:*.erl=0;38;5;48:*.sty=0;38;5;243:*.bag=4;38;5;203:*.rst=0;38;5;185:*.pdf=0;38;5;186:*.pbm=0;38;5;208:*.xcf=0;38;5;208:*.clj=0;38;5;48:*.gif=0;38;5;208:*.rar=4;38;5;203:*.elm=0;38;5;48:*.bib=0;38;5;149:*.tsx=0;38;5;48:*.dmg=4;38;5;203:*.tmp=0;38;5;243:*.bcf=0;38;5;243:*.mkv=0;38;5;208:*.svg=0;38;5;208:*.cpp=0;38;5;48:*.vim=0;38;5;48:*.bmp=0;38;5;208:*.ltx=0;38;5;48:*.fls=0;38;5;243:*.flv=0;38;5;208:*.wav=0;38;5;208:*.m4a=0;38;5;208:*.mid=0;38;5;208:*.hxx=0;38;5;48:*.pas=0;38;5;48:*.wmv=0;38;5;208:*.tif=0;38;5;208:*.kex=0;38;5;186:*.mp4=0;38;5;208:*.bak=0;38;5;243:*.xlr=0;38;5;186:*.dox=0;38;5;149:*.swf=0;38;5;208:*.tar=4;38;5;203:*.tgz=4;38;5;203:*.cfg=0;38;5;149:*.xml=0; 38;5;185:*.jpg=0;38;5;208:*.mir=0;38;5;48:*.sxi=0;38;5;186:*.bz2=4;38;5;203:*.odt=0;38;5;186:*.mov=0;38;5;208:*.toc=0;38;5;243:*.bat=1;38;5;203:*.asa=0;38;5;48:*.awk=0;38;5;48:*.sbt=0;38;5;48:*.vcd=4;38;5;203:*.kts=0;38;5;48:*.arj=4;38;5;203:*.blg=0;38;5;243:*.c++=0;38;5;48:*.odp=0;38;5;186:*.bbl=0;38;5;243:*.idx=0;38;5;243:*.com=1;38;5;203:*.mp3=0;38;5;208:*.avi=0;38;5;208:*.def=0;38;5;48:*.cgi=0;38;5;48:*.zip=4;38;5;203:*.ttf=0;38;5;208:*.ppt=0;38;5;186:*.tml=0;38;5;149:*.fsx=0;38;5;48:*.h++=0;38;5;48:*.rtf=0;38;5;186:*.inl=0;38;5;48:*.yaml=0;38;5;149:*.html=0;38;5;185:*.mpeg=0;38;5;208:*.java=0;38;5;48:*.hgrc=0;38;5;149:*.orig=0;38;5;243:*.conf=0;38;5;149:*.dart=0;38;5;48:*.psm1=0;38;5;48:*.rlib=0;38;5;243:*.fish=0;38;5;48:*.bash=0;38;5;48:*.make=0;38;5;149:*.docx=0;38;5;186:*.json=0;38;5;149:*.psd1=0;38;5;48:*.lisp=0;38;5;48:*.tbz2=4;38;5;203:*.diff=0;38;5;48:*.epub=0;38;5;186:*.xlsx=0;38;5;186:*.pptx=0;38;5;186:*.toml=0;38;5;149:*.h264=0;38;5;208:*.purs=0;38;5;48:*.flac=0;38;5;208:*.tiff=0;38;5;208:*.jpeg=0;38;5;208:*.lock=0;38;5;243:*.less=0;38;5;48:*.dyn_o=0;38;5;243:*.scala=0;38;5;48:*.mdown=0;38;5;185:*.shtml=0;38;5;185:*.class=0;38;5;243:*.cache=0;38;5;243:*.cmake=0;38;5;149:*passwd=0;38;5;149:*.swift=0;38;5;48:*shadow=0;38;5;149:*.xhtml=0;38;5;185:*.patch=0;38;5;48:*.cabal=0;38;5;48:*README=0;38;5;16;48;5;186:*.toast=4;38;5;203:*.ipynb=0;38;5;48:*COPYING=0;38;5;249:*.gradle=0;38;5;48:*.matlab=0;38;5;48:*.config=0;38;5;149:*LICENSE=0;38;5;249:*.dyn_hi=0;38;5;243:*.flake8=0;38;5;149:*.groovy=0;38;5;48:*INSTALL=0;38;5;16;48;5;186:*TODO.md=1:*.ignore=0;38;5;149:*Doxyfile=0;38;5;149:*TODO.txt=1:*setup.py=0;38;5;149:*Makefile=0;38;5;149:*.gemspec=0;38;5;149:*.desktop=0;38;5;149:*.rgignore=0;38;5;149:*.markdown=0;38;5;185:*COPYRIGHT=0;38;5;249:*configure=0;38;5;149:*.DS_Store=0;38;5;243:*.kdevelop=0;38;5;149:*.fdignore=0;38;5;149:*README.md=0;38;5;16;48;5;186:*.cmake.in=0;38;5;149:*SConscript=0;38;5;149:*CODEOWNERS=0;38;5;149:*.localized=0;38;5;243:*.gitignore=0;38;5;149:*Dockerfile=0;38;5;149:*.gitconfig=0;38;5;149:*INSTALL.md=0;38;5;16;48;5;186:*README.txt=0;38;5;16;48;5;186:*SConstruct=0;38;5;149:*.scons_opt=0;38;5;243:*.travis.yml=0;38;5;186:*.gitmodules=0;38;5;149:*.synctex.gz=0;38;5;243:*LICENSE-MIT=0;38;5;249:*MANIFEST.in=0;38;5;149:*Makefile.in=0;38;5;243:*Makefile.am=0;38;5;149:*INSTALL.txt=0;38;5;16;48;5;186:*configure.ac=0;38;5;149:*.applescript=0;38;5;48:*appveyor.yml=0;38;5;186:*.fdb_latexmk=0;38;5;243:*CONTRIBUTORS=0;38;5;16;48;5;186:*.clang-format=0;38;5;149:*LICENSE-APACHE=0;38;5;249:*CMakeLists.txt=0;38;5;149:*CMakeCache.txt=0;38;5;243:*.gitattributes=0;38;5;149:*CONTRIBUTORS.md=0;38;5;16;48;5;186:*.sconsign.dblite=0;38;5;243:*requirements.txt=0;38;5;149:*CONTRIBUTORS.txt=0;38;5;16;48;5;186:*package-lock.json=0;38;5;243:*.CFUserTextEncoding=0;38;5;243 "; fn main() { let result = run(); match result { Ok(exit_code) => { exit_code.exit(); } Err(err) => { eprintln!("[fd error]: {err:#}"); ExitCode::GeneralError.exit(); } } } fn run() -> Result<ExitCode> { let opts = Opts::parse(); #[cfg(feature = "completions")] if let Some(shell) = opts.gen_completions()? { return print_completions(shell); } set_working_dir(&opts)?; let search_paths = opts.search_paths()?; if search_paths.is_empty() { bail!("No valid search paths given."); } ensure_search_pattern_is_not_a_path(&opts)?; let pattern = &opts.pattern; let exprs = &opts.exprs; let empty = Vec::new(); let pattern_regexps = exprs .as_ref() .unwrap_or(&empty) .iter() .chain([pattern]) .map(|pat| build_pattern_regex(pat, &opts)) .collect::<Result<Vec<String>>>()?; let config = construct_config(opts, &pattern_regexps)?; ensure_use_hidden_option_for_leading_dot_pattern(&config, &pattern_regexps)?; let regexps = pattern_regexps .into_iter() .map(|pat| build_regex(pat, &config)) .collect::<Result<Vec<Regex>>>()?; walk::scan(&search_paths, regexps, config) } #[cfg(feature = "completions")] #[cold] fn print_completions(shell: clap_complete::Shell) -> Result<ExitCode> { // The program name is the first argument. let first_arg = env::args().next(); let program_name = first_arg .as_ref() .map(Path::new) .and_then(|path| path.file_stem()) .and_then(|file| file.to_str()) .unwrap_or("fd"); let mut cmd = Opts::command(); cmd.build(); clap_complete::generate(shell, &mut cmd, program_name, &mut std::io::stdout()); Ok(ExitCode::Success) } fn set_working_dir(opts: &Opts) -> Result<()> { if let Some(ref base_directory) = opts.base_directory { if !filesystem::is_existing_directory(base_directory) { return Err(anyhow!( "The '--base-directory' path '{}' is not a directory.", base_directory.to_string_lossy() )); } env::set_current_dir(base_directory).with_context(|| { format!( "Could not set '{}' as the current working directory", base_directory.to_string_lossy() ) })?; } Ok(()) } /// Detect if the user accidentally supplied a path instead of a search pattern fn ensure_search_pattern_is_not_a_path(opts: &Opts) -> Result<()> { if !opts.full_path && opts.pattern.contains(std::path::MAIN_SEPARATOR) && Path::new(&opts.pattern).is_dir() { Err(anyhow!( "The search pattern '{pattern}' contains a path-separation character ('{sep}') \ and will not lead to any search results.\n\n\ If you want to search for all files inside the '{pattern}' directory, use a match-all pattern:\n\n \ fd . '{pattern}'\n\n\ Instead, if you want your pattern to match the full file path, use:\n\n \ fd --full-path '{pattern}'", pattern = &opts.pattern, sep = std::path::MAIN_SEPARATOR, )) } else { Ok(()) } } fn build_pattern_regex(pattern: &str, opts: &Opts) -> Result<String> { Ok(if opts.glob && !pattern.is_empty() { let glob = GlobBuilder::new(pattern).literal_separator(true).build()?; glob.regex().to_owned() } else if opts.fixed_strings { // Treat pattern as literal string if '--fixed-strings' is used regex::escape(pattern) } else { String::from(pattern) }) } fn check_path_separator_length(path_separator: Option<&str>) -> Result<()> { match (cfg!(windows), path_separator) { (true, Some(sep)) if sep.len() > 1 => Err(anyhow!( "A path separator must be exactly one byte, but \ the given separator is {} bytes: '{}'.\n\ In some shells on Windows, '/' is automatically \ expanded. Try to use '//' instead.", sep.len(), sep )), _ => Ok(()), } } fn construct_config(mut opts: Opts, pattern_regexps: &[String]) -> Result<Config> { // The search will be case-sensitive if the command line flag is set or // if any of the patterns has an uppercase character (smart case). let case_sensitive = !opts.ignore_case && (opts.case_sensitive || pattern_regexps .iter() .any(|pat| pattern_has_uppercase_char(pat))); let path_separator = opts .path_separator .take() .or_else(filesystem::default_path_separator); let actual_path_separator = path_separator .clone() .unwrap_or_else(|| std::path::MAIN_SEPARATOR.to_string()); check_path_separator_length(path_separator.as_deref())?; let size_limits = std::mem::take(&mut opts.size); let time_constraints = extract_time_constraints(&opts)?; #[cfg(unix)] let owner_constraint: Option<OwnerFilter> = opts.owner.and_then(OwnerFilter::filter_ignore); #[cfg(windows)] let ansi_colors_support = nu_ansi_term::enable_ansi_support().is_ok() || std::env::var_os("TERM").is_some(); #[cfg(not(windows))] let ansi_colors_support = true; let interactive_terminal = std::io::stdout().is_terminal(); let colored_output = match opts.color { ColorWhen::Always => true, ColorWhen::Never => false, ColorWhen::Auto => { let no_color = env::var_os("NO_COLOR").is_some_and(|x| !x.is_empty()); ansi_colors_support && !no_color && interactive_terminal } }; let ls_colors = if colored_output { Some(LsColors::from_env().unwrap_or_else(|| LsColors::from_string(DEFAULT_LS_COLORS))) } else { None }; let hyperlink = match opts.hyperlink { HyperlinkWhen::Always => true, HyperlinkWhen::Never => false, HyperlinkWhen::Auto => colored_output, }; let command = extract_command(&mut opts, colored_output)?; let has_command = command.is_some(); Ok(Config { case_sensitive, search_full_path: opts.full_path, ignore_hidden: !(opts.hidden || opts.rg_alias_ignore()), read_fdignore: !(opts.no_ignore || opts.rg_alias_ignore()), read_vcsignore: !(opts.no_ignore || opts.rg_alias_ignore() || opts.no_ignore_vcs), require_git_to_read_vcsignore: !opts.no_require_git, read_parent_ignore: !opts.no_ignore_parent, read_global_ignore: !(opts.no_ignore || opts.rg_alias_ignore() || opts.no_global_ignore_file), follow_links: opts.follow, one_file_system: opts.one_file_system, null_separator: opts.null_separator, quiet: opts.quiet, max_depth: opts.max_depth(), min_depth: opts.min_depth(), prune: opts.prune, threads: opts.threads().get(), max_buffer_time: opts.max_buffer_time, ls_colors, hyperlink, interactive_terminal, file_types: opts.filetype.as_ref().map(|values| { use crate::cli::FileType::*; let mut file_types = FileTypes::default(); for value in values { match value { File => file_types.files = true, Directory => file_types.directories = true, Symlink => file_types.symlinks = true, Executable => { file_types.executables_only = true; file_types.files = true; } Empty => file_types.empty_only = true, BlockDevice => file_types.block_devices = true, CharDevice => file_types.char_devices = true, Socket => file_types.sockets = true, Pipe => file_types.pipes = true, } } // If only 'empty' was specified, search for both files and directories: if file_types.empty_only && !(file_types.files || file_types.directories) { file_types.files = true; file_types.directories = true; } file_types }), extensions: opts .extensions .as_ref() .map(|exts| { let patterns = exts .iter() .map(|e| e.trim_start_matches('.')) .map(|e| format!(r".\.{}$", regex::escape(e))); RegexSetBuilder::new(patterns) .case_insensitive(true) .build() }) .transpose()?, format: opts .format .as_deref() .map(crate::fmt::FormatTemplate::parse), command: command.map(Arc::new), batch_size: opts.batch_size, exclude_patterns: opts.exclude.iter().map(|p| String::from("!") + p).collect(), ignore_files: std::mem::take(&mut opts.ignore_file), size_constraints: size_limits, time_constraints, #[cfg(unix)] owner_constraint, show_filesystem_errors: opts.show_errors, path_separator, actual_path_separator, max_results: opts.max_results(), strip_cwd_prefix: opts.strip_cwd_prefix(|| !(opts.null_separator || has_command)), }) } fn extract_command(opts: &mut Opts, colored_output: bool) -> Result<Option<CommandSet>> { opts.exec .command .take() .map(Ok) .or_else(|| { if !opts.list_details { return None; } let res = determine_ls_command(colored_output) .map(|cmd| CommandSet::new_batch([cmd]).unwrap()); Some(res) }) .transpose() } fn determine_ls_command(colored_output: bool) -> Result<Vec<&'static str>> { #[allow(unused)] let gnu_ls = |command_name| { let color_arg = if colored_output { "--color=always" } else { "--color=never" }; // Note: we use short options here (instead of --long-options) to support more // platforms (like BusyBox). vec![ command_name, "-l", // long listing format "-h", // human readable file sizes "-d", // list directories themselves, not their contents color_arg, ] }; let cmd: Vec<&str> = if cfg!(unix) { if !cfg!(any( target_os = "macos", target_os = "dragonfly", target_os = "freebsd", target_os = "netbsd", target_os = "openbsd" )) { // Assume ls is GNU ls gnu_ls("ls") } else { // MacOS, DragonFlyBSD, FreeBSD use std::process::{Command, Stdio}; // Use GNU ls, if available (support for --color=auto, better LS_COLORS support) let gnu_ls_exists = Command::new("gls") .arg("--version") .stdout(Stdio::null()) .stderr(Stdio::null()) .status() .is_ok(); if gnu_ls_exists { gnu_ls("gls") } else { let mut cmd = vec![ "ls", // BSD version of ls "-l", // long listing format "-h", // '--human-readable' is not available, '-h' is "-d", // '--directory' is not available, but '-d' is ]; if !cfg!(any(target_os = "netbsd", target_os = "openbsd")) && colored_output { // -G is not available in NetBSD's and OpenBSD's ls cmd.push("-G"); } cmd } } } else if cfg!(windows) { use std::process::{Command, Stdio}; // Use GNU ls, if available let gnu_ls_exists = Command::new("ls") .arg("--version") .stdout(Stdio::null()) .stderr(Stdio::null()) .status() .is_ok(); if gnu_ls_exists { gnu_ls("ls") } else { return Err(anyhow!( "'fd --list-details' is not supported on Windows unless GNU 'ls' is installed." )); } } else { return Err(anyhow!( "'fd --list-details' is not supported on this platform." )); }; Ok(cmd) } fn extract_time_constraints(opts: &Opts) -> Result<Vec<TimeFilter>> { let mut time_constraints: Vec<TimeFilter> = Vec::new(); if let Some(ref t) = opts.changed_within { if let Some(f) = TimeFilter::after(t) { time_constraints.push(f); } else { return Err(anyhow!( "'{}' is not a valid date or duration. See 'fd --help'.", t )); } } if let Some(ref t) = opts.changed_before { if let Some(f) = TimeFilter::before(t) { time_constraints.push(f); } else { return Err(anyhow!( "'{}' is not a valid date or duration. See 'fd --help'.", t )); } } Ok(time_constraints) } fn ensure_use_hidden_option_for_leading_dot_pattern( config: &Config, pattern_regexps: &[String], ) -> Result<()> { if cfg!(unix) && config.ignore_hidden && pattern_regexps .iter() .any(|pat| pattern_matches_strings_with_leading_dot(pat)) { Err(anyhow!( "The pattern(s) seems to only match files with a leading dot, but hidden files are \ filtered by default. Consider adding -H/--hidden to search hidden files as well \ or adjust your search pattern(s)." )) } else { Ok(()) } } fn build_regex(pattern_regex: String, config: &Config) -> Result<regex::bytes::Regex> { RegexBuilder::new(&pattern_regex) .case_insensitive(!config.case_sensitive) .dot_matches_new_line(true) .build() .map_err(|e| { anyhow!( "{}\n\nNote: You can use the '--fixed-strings' option to search for a \ literal string instead of a regular expression. Alternatively, you can \ also use the '--glob' option to match on a glob pattern.", e ) }) }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/output.rs
src/output.rs
use std::borrow::Cow; use std::io::{self, Write}; use lscolors::{Indicator, LsColors, Style}; use crate::config::Config; use crate::dir_entry::DirEntry; use crate::fmt::FormatTemplate; use crate::hyperlink::PathUrl; fn replace_path_separator(path: &str, new_path_separator: &str) -> String { path.replace(std::path::MAIN_SEPARATOR, new_path_separator) } // TODO: this function is performance critical and can probably be optimized pub fn print_entry<W: Write>(stdout: &mut W, entry: &DirEntry, config: &Config) -> io::Result<()> { let mut has_hyperlink = false; if config.hyperlink && let Some(url) = PathUrl::new(entry.path()) { write!(stdout, "\x1B]8;;{url}\x1B\\")?; has_hyperlink = true; } if let Some(ref format) = config.format { print_entry_format(stdout, entry, config, format)?; } else if let Some(ref ls_colors) = config.ls_colors { print_entry_colorized(stdout, entry, config, ls_colors)?; } else { print_entry_uncolorized(stdout, entry, config)?; }; if has_hyperlink { write!(stdout, "\x1B]8;;\x1B\\")?; } if config.null_separator { write!(stdout, "\0") } else { writeln!(stdout) } } // Display a trailing slash if the path is a directory and the config option is enabled. // If the path_separator option is set, display that instead. // The trailing slash will not be colored. #[inline] fn print_trailing_slash<W: Write>( stdout: &mut W, entry: &DirEntry, config: &Config, style: Option<&Style>, ) -> io::Result<()> { if entry.file_type().is_some_and(|ft| ft.is_dir()) { write!( stdout, "{}", style .map(Style::to_nu_ansi_term_style) .unwrap_or_default() .paint(&config.actual_path_separator) )?; } Ok(()) } // TODO: this function is performance critical and can probably be optimized fn print_entry_format<W: Write>( stdout: &mut W, entry: &DirEntry, config: &Config, format: &FormatTemplate, ) -> io::Result<()> { let output = format.generate( entry.stripped_path(config), config.path_separator.as_deref(), ); // TODO: support writing raw bytes on unix? write!(stdout, "{}", output.to_string_lossy()) } // TODO: this function is performance critical and can probably be optimized fn print_entry_colorized<W: Write>( stdout: &mut W, entry: &DirEntry, config: &Config, ls_colors: &LsColors, ) -> io::Result<()> { // Split the path between the parent and the last component let mut offset = 0; let path = entry.stripped_path(config); let path_str = path.to_string_lossy(); if let Some(parent) = path.parent() { offset = parent.to_string_lossy().len(); for c in path_str[offset..].chars() { if std::path::is_separator(c) { offset += c.len_utf8(); } else { break; } } } if offset > 0 { let mut parent_str = Cow::from(&path_str[..offset]); if let Some(ref separator) = config.path_separator { *parent_str.to_mut() = replace_path_separator(&parent_str, separator); } let style = ls_colors .style_for_indicator(Indicator::Directory) .map(Style::to_nu_ansi_term_style) .unwrap_or_default(); write!(stdout, "{}", style.paint(parent_str))?; } let style = entry .style(ls_colors) .map(Style::to_nu_ansi_term_style) .unwrap_or_default(); write!(stdout, "{}", style.paint(&path_str[offset..]))?; print_trailing_slash( stdout, entry, config, ls_colors.style_for_indicator(Indicator::Directory), )?; Ok(()) } // TODO: this function is performance critical and can probably be optimized fn print_entry_uncolorized_base<W: Write>( stdout: &mut W, entry: &DirEntry, config: &Config, ) -> io::Result<()> { let path = entry.stripped_path(config); let mut path_string = path.to_string_lossy(); if let Some(ref separator) = config.path_separator { *path_string.to_mut() = replace_path_separator(&path_string, separator); } write!(stdout, "{path_string}")?; print_trailing_slash(stdout, entry, config, None) } #[cfg(not(unix))] fn print_entry_uncolorized<W: Write>( stdout: &mut W, entry: &DirEntry, config: &Config, ) -> io::Result<()> { print_entry_uncolorized_base(stdout, entry, config) } #[cfg(unix)] fn print_entry_uncolorized<W: Write>( stdout: &mut W, entry: &DirEntry, config: &Config, ) -> io::Result<()> { use std::os::unix::ffi::OsStrExt; if config.interactive_terminal || config.path_separator.is_some() { // Fall back to the base implementation print_entry_uncolorized_base(stdout, entry, config) } else { // Print path as raw bytes, allowing invalid UTF-8 filenames to be passed to other processes stdout.write_all(entry.stripped_path(config).as_os_str().as_bytes())?; print_trailing_slash(stdout, entry, config, None) } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/filetypes.rs
src/filetypes.rs
use crate::dir_entry; use crate::filesystem; use faccess::PathExt; /// Whether or not to show #[derive(Default)] pub struct FileTypes { pub files: bool, pub directories: bool, pub symlinks: bool, pub block_devices: bool, pub char_devices: bool, pub sockets: bool, pub pipes: bool, pub executables_only: bool, pub empty_only: bool, } impl FileTypes { pub fn should_ignore(&self, entry: &dir_entry::DirEntry) -> bool { if let Some(ref entry_type) = entry.file_type() { (!self.files && entry_type.is_file()) || (!self.directories && entry_type.is_dir()) || (!self.symlinks && entry_type.is_symlink()) || (!self.block_devices && filesystem::is_block_device(*entry_type)) || (!self.char_devices && filesystem::is_char_device(*entry_type)) || (!self.sockets && filesystem::is_socket(*entry_type)) || (!self.pipes && filesystem::is_pipe(*entry_type)) || (self.executables_only && !entry.path().executable()) || (self.empty_only && !filesystem::is_empty(entry)) || !(entry_type.is_file() || entry_type.is_dir() || entry_type.is_symlink() || filesystem::is_block_device(*entry_type) || filesystem::is_char_device(*entry_type) || filesystem::is_socket(*entry_type) || filesystem::is_pipe(*entry_type)) } else { true } } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/regex_helper.rs
src/regex_helper.rs
use regex_syntax::ParserBuilder; use regex_syntax::hir::Hir; /// Determine if a regex pattern contains a literal uppercase character. pub fn pattern_has_uppercase_char(pattern: &str) -> bool { let mut parser = ParserBuilder::new().utf8(false).build(); parser .parse(pattern) .map(|hir| hir_has_uppercase_char(&hir)) .unwrap_or(false) } /// Determine if a regex expression contains a literal uppercase character. fn hir_has_uppercase_char(hir: &Hir) -> bool { use regex_syntax::hir::*; match hir.kind() { HirKind::Literal(Literal(bytes)) => match std::str::from_utf8(bytes) { Ok(s) => s.chars().any(|c| c.is_uppercase()), Err(_) => bytes.iter().any(|b| char::from(*b).is_uppercase()), }, HirKind::Class(Class::Unicode(ranges)) => ranges .iter() .any(|r| r.start().is_uppercase() || r.end().is_uppercase()), HirKind::Class(Class::Bytes(ranges)) => ranges .iter() .any(|r| char::from(r.start()).is_uppercase() || char::from(r.end()).is_uppercase()), HirKind::Capture(Capture { sub, .. }) | HirKind::Repetition(Repetition { sub, .. }) => { hir_has_uppercase_char(sub) } HirKind::Concat(hirs) | HirKind::Alternation(hirs) => { hirs.iter().any(hir_has_uppercase_char) } _ => false, } } /// Determine if a regex pattern only matches strings starting with a literal dot (hidden files) pub fn pattern_matches_strings_with_leading_dot(pattern: &str) -> bool { let mut parser = ParserBuilder::new().utf8(false).build(); parser .parse(pattern) .map(|hir| hir_matches_strings_with_leading_dot(&hir)) .unwrap_or(false) } /// See above. fn hir_matches_strings_with_leading_dot(hir: &Hir) -> bool { use regex_syntax::hir::*; // Note: this only really detects the simplest case where a regex starts with // "^\\.", i.e. a start text anchor and a literal dot character. There are a lot // of other patterns that ONLY match hidden files, e.g. ^(\\.foo|\\.bar) which are // not (yet) detected by this algorithm. match hir.kind() { HirKind::Concat(hirs) => { let mut hirs = hirs.iter(); if let Some(hir) = hirs.next() { if hir.kind() != &HirKind::Look(Look::Start) { return false; } } else { return false; } if let Some(hir) = hirs.next() { match hir.kind() { HirKind::Literal(Literal(bytes)) => bytes.starts_with(b"."), _ => false, } } else { false } } _ => false, } } #[test] fn pattern_has_uppercase_char_simple() { assert!(pattern_has_uppercase_char("A")); assert!(pattern_has_uppercase_char("foo.EXE")); assert!(!pattern_has_uppercase_char("a")); assert!(!pattern_has_uppercase_char("foo.exe123")); } #[test] fn pattern_has_uppercase_char_advanced() { assert!(pattern_has_uppercase_char("foo.[a-zA-Z]")); assert!(!pattern_has_uppercase_char(r"\Acargo")); assert!(!pattern_has_uppercase_char(r"carg\x6F")); } #[test] fn matches_strings_with_leading_dot_simple() { assert!(pattern_matches_strings_with_leading_dot("^\\.gitignore")); assert!(!pattern_matches_strings_with_leading_dot("^.gitignore")); assert!(!pattern_matches_strings_with_leading_dot("\\.gitignore")); assert!(!pattern_matches_strings_with_leading_dot("^gitignore")); }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/filter/time.rs
src/filter/time.rs
use jiff::{Span, Timestamp, Zoned, civil::DateTime, tz::TimeZone}; use std::time::{Duration, SystemTime, UNIX_EPOCH}; /// Filter based on time ranges. #[derive(Debug, PartialEq, Eq)] pub enum TimeFilter { Before(SystemTime), After(SystemTime), } #[cfg(not(test))] fn now() -> Zoned { Zoned::now() } #[cfg(test)] thread_local! { static TESTTIME: std::cell::RefCell<Option<Zoned>> = None.into(); } /// This allows us to set a specific time when running tests #[cfg(test)] fn now() -> Zoned { TESTTIME.with_borrow(|reftime| reftime.as_ref().cloned().unwrap_or_else(Zoned::now)) } impl TimeFilter { fn from_str(s: &str) -> Option<SystemTime> { if let Ok(span) = s.parse::<Span>() { let datetime = now().checked_sub(span).ok()?; Some(datetime.into()) } else if let Ok(timestamp) = s.parse::<Timestamp>() { Some(timestamp.into()) } else if let Ok(datetime) = s.parse::<DateTime>() { Some( TimeZone::system() .to_ambiguous_zoned(datetime) .later() .ok()? .into(), ) } else { let timestamp_secs: u64 = s.strip_prefix('@')?.parse().ok()?; Some(UNIX_EPOCH + Duration::from_secs(timestamp_secs)) } } pub fn before(s: &str) -> Option<TimeFilter> { TimeFilter::from_str(s).map(TimeFilter::Before) } pub fn after(s: &str) -> Option<TimeFilter> { TimeFilter::from_str(s).map(TimeFilter::After) } pub fn applies_to(&self, t: &SystemTime) -> bool { match self { TimeFilter::Before(limit) => t < limit, TimeFilter::After(limit) => t > limit, } } } #[cfg(test)] mod tests { use super::*; use std::time::Duration; struct TestTime(SystemTime); impl TestTime { fn new(time: Zoned) -> Self { TESTTIME.with_borrow_mut(|t| *t = Some(time.clone())); TestTime(time.into()) } fn set(&mut self, time: Zoned) { TESTTIME.with_borrow_mut(|t| *t = Some(time.clone())); self.0 = time.into(); } fn timestamp(&self) -> SystemTime { self.0 } } impl Drop for TestTime { fn drop(&mut self) { // Stop using manually set times TESTTIME.with_borrow_mut(|t| *t = None); } } #[test] fn is_time_filter_applicable() { let local_tz = TimeZone::system(); let mut test_time = TestTime::new( local_tz .to_ambiguous_zoned("2010-10-10 10:10:10".parse::<DateTime>().unwrap()) .later() .unwrap(), ); let mut ref_time = test_time.timestamp(); assert!(TimeFilter::after("1min").unwrap().applies_to(&ref_time)); assert!(!TimeFilter::before("1min").unwrap().applies_to(&ref_time)); let t1m_ago = ref_time - Duration::from_secs(60); assert!(!TimeFilter::after("30sec").unwrap().applies_to(&t1m_ago)); assert!(TimeFilter::after("2min").unwrap().applies_to(&t1m_ago)); assert!(TimeFilter::before("30sec").unwrap().applies_to(&t1m_ago)); assert!(!TimeFilter::before("2min").unwrap().applies_to(&t1m_ago)); let t10s_before = "2010-10-10 10:10:00"; assert!( !TimeFilter::before(t10s_before) .unwrap() .applies_to(&ref_time) ); assert!( TimeFilter::before(t10s_before) .unwrap() .applies_to(&t1m_ago) ); assert!( TimeFilter::after(t10s_before) .unwrap() .applies_to(&ref_time) ); assert!(!TimeFilter::after(t10s_before).unwrap().applies_to(&t1m_ago)); let same_day = "2010-10-10"; assert!(!TimeFilter::before(same_day).unwrap().applies_to(&ref_time)); assert!(!TimeFilter::before(same_day).unwrap().applies_to(&t1m_ago)); assert!(TimeFilter::after(same_day).unwrap().applies_to(&ref_time)); assert!(TimeFilter::after(same_day).unwrap().applies_to(&t1m_ago)); test_time.set( "2010-10-10T10:10:10+00:00" .parse::<Timestamp>() .unwrap() .to_zoned(local_tz.clone()), ); ref_time = test_time.timestamp(); let t1m_ago = ref_time - Duration::from_secs(60); let t10s_before = "2010-10-10T10:10:00+00:00"; assert!( !TimeFilter::before(t10s_before) .unwrap() .applies_to(&ref_time) ); assert!( TimeFilter::before(t10s_before) .unwrap() .applies_to(&t1m_ago) ); assert!( TimeFilter::after(t10s_before) .unwrap() .applies_to(&ref_time) ); assert!(!TimeFilter::after(t10s_before).unwrap().applies_to(&t1m_ago)); let ref_timestamp = 1707723412u64; // Mon Feb 12 07:36:52 UTC 2024 test_time.set( "2024-02-12T07:36:52+00:00" .parse::<Timestamp>() .unwrap() .to_zoned(local_tz), ); ref_time = test_time.timestamp(); let t1m_ago = ref_time - Duration::from_secs(60); let t1s_later = ref_time + Duration::from_secs(1); // Timestamp only supported via '@' prefix assert!(TimeFilter::before(&ref_timestamp.to_string()).is_none()); assert!( TimeFilter::before(&format!("@{ref_timestamp}")) .unwrap() .applies_to(&t1m_ago) ); assert!( !TimeFilter::before(&format!("@{ref_timestamp}")) .unwrap() .applies_to(&t1s_later) ); assert!( !TimeFilter::after(&format!("@{ref_timestamp}")) .unwrap() .applies_to(&t1m_ago) ); assert!( TimeFilter::after(&format!("@{ref_timestamp}")) .unwrap() .applies_to(&t1s_later) ); } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/filter/mod.rs
src/filter/mod.rs
pub use self::size::SizeFilter; pub use self::time::TimeFilter; #[cfg(unix)] pub use self::owner::OwnerFilter; mod size; mod time; #[cfg(unix)] mod owner;
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/filter/owner.rs
src/filter/owner.rs
use anyhow::{Result, anyhow}; use nix::unistd::{Group, User}; use std::fs; #[derive(Clone, Copy, Debug, PartialEq, Eq)] pub struct OwnerFilter { uid: Check<u32>, gid: Check<u32>, } #[derive(Clone, Copy, Debug, PartialEq, Eq)] enum Check<T> { Equal(T), NotEq(T), Ignore, } impl OwnerFilter { const IGNORE: Self = OwnerFilter { uid: Check::Ignore, gid: Check::Ignore, }; /// Parses an owner constraint /// Returns an error if the string is invalid /// Returns Ok(None) when string is acceptable but a noop (such as "" or ":") pub fn from_string(input: &str) -> Result<Self> { let mut it = input.split(':'); let (fst, snd) = (it.next(), it.next()); if it.next().is_some() { return Err(anyhow!( "more than one ':' present in owner string '{}'. See 'fd --help'.", input )); } let uid = Check::parse(fst, |s| { if let Ok(uid) = s.parse() { Ok(uid) } else { User::from_name(s)? .map(|user| user.uid.as_raw()) .ok_or_else(|| anyhow!("'{}' is not a recognized user name", s)) } })?; let gid = Check::parse(snd, |s| { if let Ok(gid) = s.parse() { Ok(gid) } else { Group::from_name(s)? .map(|group| group.gid.as_raw()) .ok_or_else(|| anyhow!("'{}' is not a recognized group name", s)) } })?; Ok(OwnerFilter { uid, gid }) } /// If self is a no-op (ignore both uid and gid) then return `None`, otherwise wrap in a `Some` pub fn filter_ignore(self) -> Option<Self> { if self == Self::IGNORE { None } else { Some(self) } } pub fn matches(&self, md: &fs::Metadata) -> bool { use std::os::unix::fs::MetadataExt; self.uid.check(md.uid()) && self.gid.check(md.gid()) } } impl<T: PartialEq> Check<T> { fn check(&self, v: T) -> bool { match self { Check::Equal(x) => v == *x, Check::NotEq(x) => v != *x, Check::Ignore => true, } } fn parse<F>(s: Option<&str>, f: F) -> Result<Self> where F: Fn(&str) -> Result<T>, { let (s, equality) = match s { Some("") | None => return Ok(Check::Ignore), Some(s) if s.starts_with('!') => (&s[1..], false), Some(s) => (s, true), }; f(s).map(|x| { if equality { Check::Equal(x) } else { Check::NotEq(x) } }) } } #[cfg(test)] mod owner_parsing { use super::OwnerFilter; macro_rules! owner_tests { ($($name:ident: $value:expr => $result:pat,)*) => { $( #[test] fn $name() { let o = OwnerFilter::from_string($value); match o { $result => {}, _ => panic!("{:?} does not match {}", o, stringify!($result)), } } )* }; } use super::Check::*; owner_tests! { empty: "" => Ok(OwnerFilter::IGNORE), uid_only: "5" => Ok(OwnerFilter { uid: Equal(5), gid: Ignore }), uid_gid: "9:3" => Ok(OwnerFilter { uid: Equal(9), gid: Equal(3) }), gid_only: ":8" => Ok(OwnerFilter { uid: Ignore, gid: Equal(8) }), colon_only: ":" => Ok(OwnerFilter::IGNORE), trailing: "5:" => Ok(OwnerFilter { uid: Equal(5), gid: Ignore }), uid_negate: "!5" => Ok(OwnerFilter { uid: NotEq(5), gid: Ignore }), both_negate:"!4:!3" => Ok(OwnerFilter { uid: NotEq(4), gid: NotEq(3) }), uid_not_gid:"6:!8" => Ok(OwnerFilter { uid: Equal(6), gid: NotEq(8) }), more_colons:"3:5:" => Err(_), only_colons:"::" => Err(_), } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/filter/size.rs
src/filter/size.rs
use std::sync::OnceLock; use anyhow::anyhow; use regex::Regex; static SIZE_CAPTURES: OnceLock<Regex> = OnceLock::new(); #[derive(Clone, Copy, Debug, PartialEq, Eq)] pub enum SizeFilter { Max(u64), Min(u64), Equals(u64), } // SI prefixes (powers of 10) const KILO: u64 = 1000; const MEGA: u64 = KILO * 1000; const GIGA: u64 = MEGA * 1000; const TERA: u64 = GIGA * 1000; // Binary prefixes (powers of 2) const KIBI: u64 = 1024; const MEBI: u64 = KIBI * 1024; const GIBI: u64 = MEBI * 1024; const TEBI: u64 = GIBI * 1024; impl SizeFilter { pub fn from_string(s: &str) -> anyhow::Result<Self> { SizeFilter::parse_opt(s) .ok_or_else(|| anyhow!("'{}' is not a valid size constraint. See 'fd --help'.", s)) } fn parse_opt(s: &str) -> Option<Self> { let pattern = SIZE_CAPTURES.get_or_init(|| Regex::new(r"(?i)^([+-]?)(\d+)(b|[kmgt]i?b?)$").unwrap()); if !pattern.is_match(s) { return None; } let captures = pattern.captures(s)?; let limit_kind = captures.get(1).map_or("+", |m| m.as_str()); let quantity = captures .get(2) .and_then(|v| v.as_str().parse::<u64>().ok())?; let multiplier = match &captures.get(3).map_or("b", |m| m.as_str()).to_lowercase()[..] { v if v.starts_with("ki") => KIBI, v if v.starts_with('k') => KILO, v if v.starts_with("mi") => MEBI, v if v.starts_with('m') => MEGA, v if v.starts_with("gi") => GIBI, v if v.starts_with('g') => GIGA, v if v.starts_with("ti") => TEBI, v if v.starts_with('t') => TERA, "b" => 1, _ => return None, }; let size = quantity * multiplier; match limit_kind { "+" => Some(SizeFilter::Min(size)), "-" => Some(SizeFilter::Max(size)), "" => Some(SizeFilter::Equals(size)), _ => None, } } pub fn is_within(&self, size: u64) -> bool { match *self { SizeFilter::Max(limit) => size <= limit, SizeFilter::Min(limit) => size >= limit, SizeFilter::Equals(limit) => size == limit, } } } #[cfg(test)] mod tests { use super::*; macro_rules! gen_size_filter_parse_test { ($($name: ident: $val: expr,)*) => { $( #[test] fn $name() { let (txt, expected) = $val; let actual = SizeFilter::from_string(txt).unwrap(); assert_eq!(actual, expected); } )* }; } // Parsing and size conversion tests data. Ensure that each type gets properly interpreted. // Call with higher base values to ensure expected multiplication (only need a couple) gen_size_filter_parse_test! { byte_plus: ("+1b", SizeFilter::Min(1)), byte_plus_multiplier: ("+10b", SizeFilter::Min(10)), byte_minus: ("-1b", SizeFilter::Max(1)), kilo_plus: ("+1k", SizeFilter::Min(1000)), kilo_plus_suffix: ("+1kb", SizeFilter::Min(1000)), kilo_minus: ("-1k", SizeFilter::Max(1000)), kilo_minus_multiplier: ("-100k", SizeFilter::Max(100_000)), kilo_minus_suffix: ("-1kb", SizeFilter::Max(1000)), kilo_plus_upper: ("+1K", SizeFilter::Min(1000)), kilo_plus_suffix_upper: ("+1KB", SizeFilter::Min(1000)), kilo_minus_upper: ("-1K", SizeFilter::Max(1000)), kilo_minus_suffix_upper: ("-1Kb", SizeFilter::Max(1000)), kibi_plus: ("+1ki", SizeFilter::Min(1024)), kibi_plus_multiplier: ("+10ki", SizeFilter::Min(10_240)), kibi_plus_suffix: ("+1kib", SizeFilter::Min(1024)), kibi_minus: ("-1ki", SizeFilter::Max(1024)), kibi_minus_multiplier: ("-100ki", SizeFilter::Max(102_400)), kibi_minus_suffix: ("-1kib", SizeFilter::Max(1024)), kibi_plus_upper: ("+1KI", SizeFilter::Min(1024)), kibi_plus_suffix_upper: ("+1KiB", SizeFilter::Min(1024)), kibi_minus_upper: ("-1Ki", SizeFilter::Max(1024)), kibi_minus_suffix_upper: ("-1KIB", SizeFilter::Max(1024)), mega_plus: ("+1m", SizeFilter::Min(1_000_000)), mega_plus_suffix: ("+1mb", SizeFilter::Min(1_000_000)), mega_minus: ("-1m", SizeFilter::Max(1_000_000)), mega_minus_suffix: ("-1mb", SizeFilter::Max(1_000_000)), mega_plus_upper: ("+1M", SizeFilter::Min(1_000_000)), mega_plus_suffix_upper: ("+1MB", SizeFilter::Min(1_000_000)), mega_minus_upper: ("-1M", SizeFilter::Max(1_000_000)), mega_minus_suffix_upper: ("-1Mb", SizeFilter::Max(1_000_000)), mebi_plus: ("+1mi", SizeFilter::Min(1_048_576)), mebi_plus_suffix: ("+1mib", SizeFilter::Min(1_048_576)), mebi_minus: ("-1mi", SizeFilter::Max(1_048_576)), mebi_minus_suffix: ("-1mib", SizeFilter::Max(1_048_576)), mebi_plus_upper: ("+1MI", SizeFilter::Min(1_048_576)), mebi_plus_suffix_upper: ("+1MiB", SizeFilter::Min(1_048_576)), mebi_minus_upper: ("-1Mi", SizeFilter::Max(1_048_576)), mebi_minus_suffix_upper: ("-1MIB", SizeFilter::Max(1_048_576)), giga_plus: ("+1g", SizeFilter::Min(1_000_000_000)), giga_plus_suffix: ("+1gb", SizeFilter::Min(1_000_000_000)), giga_minus: ("-1g", SizeFilter::Max(1_000_000_000)), giga_minus_suffix: ("-1gb", SizeFilter::Max(1_000_000_000)), giga_plus_upper: ("+1G", SizeFilter::Min(1_000_000_000)), giga_plus_suffix_upper: ("+1GB", SizeFilter::Min(1_000_000_000)), giga_minus_upper: ("-1G", SizeFilter::Max(1_000_000_000)), giga_minus_suffix_upper: ("-1Gb", SizeFilter::Max(1_000_000_000)), gibi_plus: ("+1gi", SizeFilter::Min(1_073_741_824)), gibi_plus_suffix: ("+1gib", SizeFilter::Min(1_073_741_824)), gibi_minus: ("-1gi", SizeFilter::Max(1_073_741_824)), gibi_minus_suffix: ("-1gib", SizeFilter::Max(1_073_741_824)), gibi_plus_upper: ("+1GI", SizeFilter::Min(1_073_741_824)), gibi_plus_suffix_upper: ("+1GiB", SizeFilter::Min(1_073_741_824)), gibi_minus_upper: ("-1Gi", SizeFilter::Max(1_073_741_824)), gibi_minus_suffix_upper: ("-1GIB", SizeFilter::Max(1_073_741_824)), tera_plus: ("+1t", SizeFilter::Min(1_000_000_000_000)), tera_plus_suffix: ("+1tb", SizeFilter::Min(1_000_000_000_000)), tera_minus: ("-1t", SizeFilter::Max(1_000_000_000_000)), tera_minus_suffix: ("-1tb", SizeFilter::Max(1_000_000_000_000)), tera_plus_upper: ("+1T", SizeFilter::Min(1_000_000_000_000)), tera_plus_suffix_upper: ("+1TB", SizeFilter::Min(1_000_000_000_000)), tera_minus_upper: ("-1T", SizeFilter::Max(1_000_000_000_000)), tera_minus_suffix_upper: ("-1Tb", SizeFilter::Max(1_000_000_000_000)), tebi_plus: ("+1ti", SizeFilter::Min(1_099_511_627_776)), tebi_plus_suffix: ("+1tib", SizeFilter::Min(1_099_511_627_776)), tebi_minus: ("-1ti", SizeFilter::Max(1_099_511_627_776)), tebi_minus_suffix: ("-1tib", SizeFilter::Max(1_099_511_627_776)), tebi_plus_upper: ("+1TI", SizeFilter::Min(1_099_511_627_776)), tebi_plus_suffix_upper: ("+1TiB", SizeFilter::Min(1_099_511_627_776)), tebi_minus_upper: ("-1Ti", SizeFilter::Max(1_099_511_627_776)), tebi_minus_suffix_upper: ("-1TIB", SizeFilter::Max(1_099_511_627_776)), } /// Invalid parse testing macro_rules! gen_size_filter_failure { ($($name:ident: $value:expr,)*) => { $( #[test] fn $name() { let i = SizeFilter::from_string($value); assert!(i.is_err()); } )* }; } // Invalid parse data gen_size_filter_failure! { ensure_missing_number_returns_none: "+g", ensure_missing_unit_returns_none: "+18", ensure_bad_format_returns_none_1: "$10M", ensure_bad_format_returns_none_2: "badval", ensure_bad_format_returns_none_3: "9999", ensure_invalid_unit_returns_none_1: "+50a", ensure_invalid_unit_returns_none_2: "-10v", ensure_invalid_unit_returns_none_3: "+1Mv", ensure_bib_format_returns_none: "+1bib", ensure_bb_format_returns_none: "+1bb", } #[test] fn is_within_less_than() { let f = SizeFilter::from_string("-1k").unwrap(); assert!(f.is_within(999)); } #[test] fn is_within_less_than_equal() { let f = SizeFilter::from_string("-1k").unwrap(); assert!(f.is_within(1000)); } #[test] fn is_within_greater_than() { let f = SizeFilter::from_string("+1k").unwrap(); assert!(f.is_within(1001)); } #[test] fn is_within_greater_than_equal() { let f = SizeFilter::from_string("+1K").unwrap(); assert!(f.is_within(1000)); } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/exec/command.rs
src/exec/command.rs
use std::io; use std::io::Write; use argmax::Command; use crate::error::print_error; use crate::exit_codes::ExitCode; struct Outputs { stdout: Vec<u8>, stderr: Vec<u8>, } pub struct OutputBuffer { null_separator: bool, outputs: Vec<Outputs>, } impl OutputBuffer { pub fn new(null_separator: bool) -> Self { Self { null_separator, outputs: Vec::new(), } } fn push(&mut self, stdout: Vec<u8>, stderr: Vec<u8>) { self.outputs.push(Outputs { stdout, stderr }); } fn write(self) { // Avoid taking the lock if there is nothing to do. // If null_separator is true, then we still need to write the // null separator, because the output may have been written directly // to stdout if self.outputs.is_empty() && !self.null_separator { return; } let stdout = io::stdout(); let stderr = io::stderr(); // While we hold these locks, only this thread will be able // to write its outputs. let mut stdout = stdout.lock(); let mut stderr = stderr.lock(); for output in self.outputs.iter() { let _ = stdout.write_all(&output.stdout); let _ = stderr.write_all(&output.stderr); } if self.null_separator { // If null_separator is enabled, then we should write a \0 at the end // of the output for this entry let _ = stdout.write_all(b"\0"); } } } /// Executes a command. pub fn execute_commands<I: Iterator<Item = io::Result<Command>>>( cmds: I, mut output_buffer: OutputBuffer, enable_output_buffering: bool, ) -> ExitCode { for result in cmds { let mut cmd = match result { Ok(cmd) => cmd, Err(e) => return handle_cmd_error(None, e), }; // Spawn the supplied command. let output = if enable_output_buffering { cmd.output() } else { // If running on only one thread, don't buffer output // Allows for viewing and interacting with intermediate command output cmd.spawn().and_then(|c| c.wait_with_output()) }; // Then wait for the command to exit, if it was spawned. match output { Ok(output) => { if enable_output_buffering { output_buffer.push(output.stdout, output.stderr); } if output.status.code() != Some(0) { output_buffer.write(); return ExitCode::GeneralError; } } Err(why) => { output_buffer.write(); return handle_cmd_error(Some(&cmd), why); } } } output_buffer.write(); ExitCode::Success } pub fn handle_cmd_error(cmd: Option<&Command>, err: io::Error) -> ExitCode { match (cmd, err) { (Some(cmd), err) if err.kind() == io::ErrorKind::NotFound => { print_error(format!( "Command not found: {}", cmd.get_program().to_string_lossy() )); ExitCode::GeneralError } (_, err) => { print_error(format!("Problem while executing command: {err}")); ExitCode::GeneralError } } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/exec/mod.rs
src/exec/mod.rs
mod command; mod job; use std::ffi::OsString; use std::io; use std::iter; use std::path::{Path, PathBuf}; use std::process::Stdio; use anyhow::{Result, bail}; use argmax::Command; use crate::exec::command::OutputBuffer; use crate::exit_codes::{ExitCode, merge_exitcodes}; use crate::fmt::{FormatTemplate, Token}; use self::command::{execute_commands, handle_cmd_error}; pub use self::job::{batch, job}; /// Execution mode of the command #[derive(Debug, Clone, Copy, PartialEq, Eq)] pub enum ExecutionMode { /// Command is executed for each search result OneByOne, /// Command is run for a batch of results at once Batch, } #[derive(Debug, Clone, PartialEq)] pub struct CommandSet { mode: ExecutionMode, commands: Vec<CommandTemplate>, } impl CommandSet { pub fn new<I, T, S>(input: I) -> Result<CommandSet> where I: IntoIterator<Item = T>, T: IntoIterator<Item = S>, S: AsRef<str>, { Ok(CommandSet { mode: ExecutionMode::OneByOne, commands: input .into_iter() .map(CommandTemplate::new) .collect::<Result<_>>()?, }) } pub fn new_batch<I, T, S>(input: I) -> Result<CommandSet> where I: IntoIterator<Item = T>, T: IntoIterator<Item = S>, S: AsRef<str>, { Ok(CommandSet { mode: ExecutionMode::Batch, commands: input .into_iter() .map(|args| { let cmd = CommandTemplate::new(args)?; if cmd.number_of_tokens() > 1 { bail!("Only one placeholder allowed for batch commands"); } if cmd.args[0].has_tokens() { bail!("First argument of exec-batch is expected to be a fixed executable"); } Ok(cmd) }) .collect::<Result<Vec<_>>>()?, }) } pub fn in_batch_mode(&self) -> bool { self.mode == ExecutionMode::Batch } pub fn execute( &self, input: &Path, path_separator: Option<&str>, null_separator: bool, buffer_output: bool, ) -> ExitCode { let commands = self .commands .iter() .map(|c| c.generate(input, path_separator)); execute_commands(commands, OutputBuffer::new(null_separator), buffer_output) } pub fn execute_batch<I>(&self, paths: I, limit: usize, path_separator: Option<&str>) -> ExitCode where I: Iterator<Item = PathBuf>, { let builders: io::Result<Vec<_>> = self .commands .iter() .map(|c| CommandBuilder::new(c, limit)) .collect(); match builders { Ok(mut builders) => { for path in paths { for builder in &mut builders { if let Err(e) = builder.push(&path, path_separator) { return handle_cmd_error(Some(&builder.cmd), e); } } } for builder in &mut builders { if let Err(e) = builder.finish() { return handle_cmd_error(Some(&builder.cmd), e); } } merge_exitcodes(builders.iter().map(|b| b.exit_code())) } Err(e) => handle_cmd_error(None, e), } } } /// Represents a multi-exec command as it is built. #[derive(Debug)] struct CommandBuilder { pre_args: Vec<OsString>, path_arg: FormatTemplate, post_args: Vec<OsString>, cmd: Command, count: usize, limit: usize, exit_code: ExitCode, } impl CommandBuilder { fn new(template: &CommandTemplate, limit: usize) -> io::Result<Self> { let mut pre_args = vec![]; let mut path_arg = None; let mut post_args = vec![]; for arg in &template.args { if arg.has_tokens() { path_arg = Some(arg.clone()); } else if path_arg.is_none() { pre_args.push(arg.generate("", None)); } else { post_args.push(arg.generate("", None)); } } let cmd = Self::new_command(&pre_args)?; Ok(Self { pre_args, path_arg: path_arg.unwrap(), post_args, cmd, count: 0, limit, exit_code: ExitCode::Success, }) } fn new_command(pre_args: &[OsString]) -> io::Result<Command> { let mut cmd = Command::new(&pre_args[0]); cmd.stdin(Stdio::inherit()); cmd.stdout(Stdio::inherit()); cmd.stderr(Stdio::inherit()); cmd.try_args(&pre_args[1..])?; Ok(cmd) } fn push(&mut self, path: &Path, separator: Option<&str>) -> io::Result<()> { if self.limit > 0 && self.count >= self.limit { self.finish()?; } let arg = self.path_arg.generate(path, separator); if !self .cmd .args_would_fit(iter::once(&arg).chain(&self.post_args)) { self.finish()?; } self.cmd.try_arg(arg)?; self.count += 1; Ok(()) } fn finish(&mut self) -> io::Result<()> { if self.count > 0 { self.cmd.try_args(&self.post_args)?; if !self.cmd.status()?.success() { self.exit_code = ExitCode::GeneralError; } self.cmd = Self::new_command(&self.pre_args)?; self.count = 0; } Ok(()) } fn exit_code(&self) -> ExitCode { self.exit_code } } /// Represents a template that is utilized to generate command strings. /// /// The template is meant to be coupled with an input in order to generate a command. The /// `generate_and_execute()` method will be used to generate a command and execute it. #[derive(Debug, Clone, PartialEq)] struct CommandTemplate { args: Vec<FormatTemplate>, } impl CommandTemplate { fn new<I, S>(input: I) -> Result<CommandTemplate> where I: IntoIterator<Item = S>, S: AsRef<str>, { let mut args = Vec::new(); let mut has_placeholder = false; for arg in input { let arg = arg.as_ref(); let tmpl = FormatTemplate::parse(arg); has_placeholder |= tmpl.has_tokens(); args.push(tmpl); } // We need to check that we have at least one argument, because if not // it will try to execute each file and directory it finds. // // Sadly, clap can't currently handle this for us, see // https://github.com/clap-rs/clap/issues/3542 if args.is_empty() { bail!("No executable provided for --exec or --exec-batch"); } // If a placeholder token was not supplied, append one at the end of the command. if !has_placeholder { args.push(FormatTemplate::Tokens(vec![Token::Placeholder])); } Ok(CommandTemplate { args }) } fn number_of_tokens(&self) -> usize { self.args.iter().filter(|arg| arg.has_tokens()).count() } /// Generates and executes a command. /// /// Using the internal `args` field, and a supplied `input` variable, a `Command` will be /// build. fn generate(&self, input: &Path, path_separator: Option<&str>) -> io::Result<Command> { let mut cmd = Command::new(self.args[0].generate(input, path_separator)); for arg in &self.args[1..] { cmd.try_arg(arg.generate(input, path_separator))?; } Ok(cmd) } } #[cfg(test)] mod tests { use super::*; fn generate_str(template: &CommandTemplate, input: &str) -> Vec<String> { template .args .iter() .map(|arg| arg.generate(input, None).into_string().unwrap()) .collect() } #[test] fn tokens_with_placeholder() { assert_eq!( CommandSet::new(vec![vec![&"echo", &"${SHELL}:"]]).unwrap(), CommandSet { commands: vec![CommandTemplate { args: vec![ FormatTemplate::Text("echo".into()), FormatTemplate::Text("${SHELL}:".into()), FormatTemplate::Tokens(vec![Token::Placeholder]), ] }], mode: ExecutionMode::OneByOne, } ); } #[test] fn tokens_with_no_extension() { assert_eq!( CommandSet::new(vec![vec!["echo", "{.}"]]).unwrap(), CommandSet { commands: vec![CommandTemplate { args: vec![ FormatTemplate::Text("echo".into()), FormatTemplate::Tokens(vec![Token::NoExt]), ], }], mode: ExecutionMode::OneByOne, } ); } #[test] fn tokens_with_basename() { assert_eq!( CommandSet::new(vec![vec!["echo", "{/}"]]).unwrap(), CommandSet { commands: vec![CommandTemplate { args: vec![ FormatTemplate::Text("echo".into()), FormatTemplate::Tokens(vec![Token::Basename]), ], }], mode: ExecutionMode::OneByOne, } ); } #[test] fn tokens_with_parent() { assert_eq!( CommandSet::new(vec![vec!["echo", "{//}"]]).unwrap(), CommandSet { commands: vec![CommandTemplate { args: vec![ FormatTemplate::Text("echo".into()), FormatTemplate::Tokens(vec![Token::Parent]), ], }], mode: ExecutionMode::OneByOne, } ); } #[test] fn tokens_with_basename_no_extension() { assert_eq!( CommandSet::new(vec![vec!["echo", "{/.}"]]).unwrap(), CommandSet { commands: vec![CommandTemplate { args: vec![ FormatTemplate::Text("echo".into()), FormatTemplate::Tokens(vec![Token::BasenameNoExt]), ], }], mode: ExecutionMode::OneByOne, } ); } #[test] fn tokens_with_literal_braces() { let template = CommandTemplate::new(vec!["{{}}", "{{", "{.}}"]).unwrap(); assert_eq!( generate_str(&template, "foo"), vec!["{}", "{", "{.}", "foo"] ); } #[test] fn tokens_with_literal_braces_and_placeholder() { let template = CommandTemplate::new(vec!["{{{},end}"]).unwrap(); assert_eq!(generate_str(&template, "foo"), vec!["{foo,end}"]); } #[test] fn tokens_multiple() { assert_eq!( CommandSet::new(vec![vec!["cp", "{}", "{/.}.ext"]]).unwrap(), CommandSet { commands: vec![CommandTemplate { args: vec![ FormatTemplate::Text("cp".into()), FormatTemplate::Tokens(vec![Token::Placeholder]), FormatTemplate::Tokens(vec![ Token::BasenameNoExt, Token::Text(".ext".into()) ]), ], }], mode: ExecutionMode::OneByOne, } ); } #[test] fn tokens_single_batch() { assert_eq!( CommandSet::new_batch(vec![vec!["echo", "{.}"]]).unwrap(), CommandSet { commands: vec![CommandTemplate { args: vec![ FormatTemplate::Text("echo".into()), FormatTemplate::Tokens(vec![Token::NoExt]), ], }], mode: ExecutionMode::Batch, } ); } #[test] fn tokens_multiple_batch() { assert!(CommandSet::new_batch(vec![vec!["echo", "{.}", "{}"]]).is_err()); } #[test] fn template_no_args() { assert!(CommandTemplate::new::<Vec<_>, &'static str>(vec![]).is_err()); } #[test] fn command_set_no_args() { assert!(CommandSet::new(vec![vec!["echo"], vec![]]).is_err()); } #[test] fn generate_custom_path_separator() { let arg = FormatTemplate::Tokens(vec![Token::Placeholder]); macro_rules! check { ($input:expr, $expected:expr) => { assert_eq!(arg.generate($input, Some("#")), OsString::from($expected)); }; } check!("foo", "foo"); check!("foo/bar", "foo#bar"); check!("/foo/bar/baz", "#foo#bar#baz"); } #[cfg(windows)] #[test] fn generate_custom_path_separator_windows() { let arg = FormatTemplate::Tokens(vec![Token::Placeholder]); macro_rules! check { ($input:expr, $expected:expr) => { assert_eq!(arg.generate($input, Some("#")), OsString::from($expected)); }; } // path starting with a drive letter check!(r"C:\foo\bar", "C:#foo#bar"); // UNC path check!(r"\\server\share\path", "##server#share#path"); // Drive Relative path - no separator after the colon omits the RootDir path component. // This is uncommon, but valid check!(r"C:foo\bar", "C:foo#bar"); // forward slashes should get normalized and interpreted as separators check!("C:/foo/bar", "C:#foo#bar"); check!("C:foo/bar", "C:foo#bar"); // Rust does not interpret "//server/share" as a UNC path, but rather as a normal // absolute path that begins with RootDir, and the two slashes get combined together as // a single path separator during normalization. //check!("//server/share/path", "##server#share#path"); } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/exec/job.rs
src/exec/job.rs
use crate::config::Config; use crate::error::print_error; use crate::exit_codes::{ExitCode, merge_exitcodes}; use crate::walk::WorkerResult; use super::CommandSet; /// An event loop that listens for inputs from the `rx` receiver. Each received input will /// generate a command with the supplied command template. The generated command will then /// be executed, and this process will continue until the receiver's sender has closed. pub fn job( results: impl IntoIterator<Item = WorkerResult>, cmd: &CommandSet, config: &Config, ) -> ExitCode { // Output should be buffered when only running a single thread let buffer_output: bool = config.threads > 1; let mut ret = ExitCode::Success; for result in results { // Obtain the next result from the receiver, else if the channel // has closed, exit from the loop let dir_entry = match result { WorkerResult::Entry(dir_entry) => dir_entry, WorkerResult::Error(err) => { if config.show_filesystem_errors { print_error(err.to_string()); } continue; } }; // Generate a command, execute it and store its exit code. let code = cmd.execute( dir_entry.stripped_path(config), config.path_separator.as_deref(), config.null_separator, buffer_output, ); ret = merge_exitcodes([ret, code]); } // Returns error in case of any error. ret } pub fn batch( results: impl IntoIterator<Item = WorkerResult>, cmd: &CommandSet, config: &Config, ) -> ExitCode { let paths = results .into_iter() .filter_map(|worker_result| match worker_result { WorkerResult::Entry(dir_entry) => Some(dir_entry.into_stripped_path(config)), WorkerResult::Error(err) => { if config.show_filesystem_errors { print_error(err.to_string()); } None } }); cmd.execute_batch(paths, config.batch_size, config.path_separator.as_deref()) }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/fmt/mod.rs
src/fmt/mod.rs
mod input; use std::borrow::Cow; use std::ffi::{OsStr, OsString}; use std::fmt::{self, Display, Formatter}; use std::path::{Component, Path, Prefix}; use std::sync::OnceLock; use aho_corasick::AhoCorasick; use self::input::{basename, dirname, remove_extension}; /// Designates what should be written to a buffer /// /// Each `Token` contains either text, or a placeholder variant, which will be used to generate /// commands after all tokens for a given command template have been collected. #[derive(Clone, Debug, PartialEq, Eq)] pub enum Token { Placeholder, Basename, Parent, NoExt, BasenameNoExt, Text(String), } impl Display for Token { fn fmt(&self, f: &mut Formatter) -> fmt::Result { match *self { Token::Placeholder => f.write_str("{}")?, Token::Basename => f.write_str("{/}")?, Token::Parent => f.write_str("{//}")?, Token::NoExt => f.write_str("{.}")?, Token::BasenameNoExt => f.write_str("{/.}")?, Token::Text(ref string) => f.write_str(string)?, } Ok(()) } } /// A parsed format string /// /// This is either a collection of `Token`s including at least one placeholder variant, /// or a fixed text. #[derive(Clone, Debug, PartialEq)] pub enum FormatTemplate { Tokens(Vec<Token>), Text(String), } static PLACEHOLDERS: OnceLock<AhoCorasick> = OnceLock::new(); impl FormatTemplate { pub fn has_tokens(&self) -> bool { matches!(self, FormatTemplate::Tokens(_)) } pub fn parse(fmt: &str) -> Self { // NOTE: we assume that { and } have the same length const BRACE_LEN: usize = '{'.len_utf8(); let mut tokens = Vec::new(); let mut remaining = fmt; let mut buf = String::new(); let placeholders = PLACEHOLDERS.get_or_init(|| { AhoCorasick::new(["{{", "}}", "{}", "{/}", "{//}", "{.}", "{/.}"]).unwrap() }); while let Some(m) = placeholders.find(remaining) { match m.pattern().as_u32() { 0 | 1 => { // we found an escaped {{ or }}, so add // everything up to the first char to the buffer // then skip the second one. buf += &remaining[..m.start() + BRACE_LEN]; remaining = &remaining[m.end()..]; } id if !remaining[m.end()..].starts_with('}') => { buf += &remaining[..m.start()]; if !buf.is_empty() { tokens.push(Token::Text(std::mem::take(&mut buf))); } tokens.push(token_from_pattern_id(id)); remaining = &remaining[m.end()..]; } _ => { // We got a normal pattern, but the final "}" // is escaped, so add up to that to the buffer, then // skip the final } buf += &remaining[..m.end()]; remaining = &remaining[m.end() + BRACE_LEN..]; } } } // Add the rest of the string to the buffer, and add the final buffer to the tokens if !remaining.is_empty() { buf += remaining; } if tokens.is_empty() { // No placeholders were found, so just return the text return FormatTemplate::Text(buf); } // Add final text segment if !buf.is_empty() { tokens.push(Token::Text(buf)); } debug_assert!(!tokens.is_empty()); FormatTemplate::Tokens(tokens) } /// Generate a result string from this template. If path_separator is Some, then it will replace /// the path separator in all placeholder tokens. Fixed text and tokens are not affected by /// path separator substitution. pub fn generate(&self, path: impl AsRef<Path>, path_separator: Option<&str>) -> OsString { use Token::*; let path = path.as_ref(); match *self { Self::Tokens(ref tokens) => { let mut s = OsString::new(); for token in tokens { match token { Basename => s.push(Self::replace_separator(basename(path), path_separator)), BasenameNoExt => s.push(Self::replace_separator( &remove_extension(basename(path).as_ref()), path_separator, )), NoExt => s.push(Self::replace_separator( &remove_extension(path), path_separator, )), Parent => s.push(Self::replace_separator(&dirname(path), path_separator)), Placeholder => { s.push(Self::replace_separator(path.as_ref(), path_separator)) } Text(string) => s.push(string), } } s } Self::Text(ref text) => OsString::from(text), } } /// Replace the path separator in the input with the custom separator string. If path_separator /// is None, simply return a borrowed Cow<OsStr> of the input. Otherwise, the input is /// interpreted as a Path and its components are iterated through and re-joined into a new /// OsString. fn replace_separator<'a>(path: &'a OsStr, path_separator: Option<&str>) -> Cow<'a, OsStr> { // fast-path - no replacement necessary if path_separator.is_none() { return Cow::Borrowed(path); } let path_separator = path_separator.unwrap(); let mut out = OsString::with_capacity(path.len()); let mut components = Path::new(path).components().peekable(); while let Some(comp) = components.next() { match comp { // Absolute paths on Windows are tricky. A Prefix component is usually a drive // letter or UNC path, and is usually followed by RootDir. There are also // "verbatim" prefixes beginning with "\\?\" that skip normalization. We choose to // ignore verbatim path prefixes here because they're very rare, might be // impossible to reach here, and there's no good way to deal with them. If users // are doing something advanced involving verbatim windows paths, they can do their // own output filtering with a tool like sed. Component::Prefix(prefix) => { if let Prefix::UNC(server, share) = prefix.kind() { // Prefix::UNC is a parsed version of '\\server\share' out.push(path_separator); out.push(path_separator); out.push(server); out.push(path_separator); out.push(share); } else { // All other Windows prefix types are rendered as-is. This results in e.g. "C:" for // drive letters. DeviceNS and Verbatim* prefixes won't have backslashes converted, // but they're not returned by directories fd can search anyway so we don't worry // about them. out.push(comp.as_os_str()); } } // Root directory is always replaced with the custom separator. Component::RootDir => out.push(path_separator), // Everything else is joined normally, with a trailing separator if we're not last _ => { out.push(comp.as_os_str()); if components.peek().is_some() { out.push(path_separator); } } } } Cow::Owned(out) } } // Convert the id from an aho-corasick match to the // appropriate token fn token_from_pattern_id(id: u32) -> Token { use Token::*; match id { 2 => Placeholder, 3 => Basename, 4 => Parent, 5 => NoExt, 6 => BasenameNoExt, _ => unreachable!(), } } #[cfg(test)] mod fmt_tests { use super::*; use std::path::PathBuf; #[test] fn parse_no_placeholders() { let templ = FormatTemplate::parse("This string has no placeholders"); assert_eq!( templ, FormatTemplate::Text("This string has no placeholders".into()) ); } #[test] fn parse_only_brace_escapes() { let templ = FormatTemplate::parse("This string only has escapes like {{ and }}"); assert_eq!( templ, FormatTemplate::Text("This string only has escapes like { and }".into()) ); } #[test] fn all_placeholders() { use Token::*; let templ = FormatTemplate::parse( "{{path={} \ basename={/} \ parent={//} \ noExt={.} \ basenameNoExt={/.} \ }}", ); assert_eq!( templ, FormatTemplate::Tokens(vec![ Text("{path=".into()), Placeholder, Text(" basename=".into()), Basename, Text(" parent=".into()), Parent, Text(" noExt=".into()), NoExt, Text(" basenameNoExt=".into()), BasenameNoExt, Text(" }".into()), ]) ); let mut path = PathBuf::new(); path.push("a"); path.push("folder"); path.push("file.txt"); let expanded = templ.generate(&path, Some("/")).into_string().unwrap(); assert_eq!( expanded, "{path=a/folder/file.txt \ basename=file.txt \ parent=a/folder \ noExt=a/folder/file \ basenameNoExt=file }" ); } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
sharkdp/fd
https://github.com/sharkdp/fd/blob/5f95a781212e3efbc6d91ae50e0ca0ce0693da50/src/fmt/input.rs
src/fmt/input.rs
use std::ffi::{OsStr, OsString}; use std::path::{Path, PathBuf}; use crate::filesystem::strip_current_dir; /// Removes the parent component of the path pub fn basename(path: &Path) -> &OsStr { path.file_name().unwrap_or(path.as_os_str()) } /// Removes the extension from the path pub fn remove_extension(path: &Path) -> OsString { let dirname = dirname(path); let stem = path.file_stem().unwrap_or(path.as_os_str()); let path = PathBuf::from(dirname).join(stem); strip_current_dir(&path).to_owned().into_os_string() } /// Removes the basename from the path. pub fn dirname(path: &Path) -> OsString { path.parent() .map(|p| { if p == OsStr::new("") { OsString::from(".") } else { p.as_os_str().to_owned() } }) .unwrap_or_else(|| path.as_os_str().to_owned()) } #[cfg(test)] mod path_tests { use super::*; use std::path::MAIN_SEPARATOR_STR; fn correct(input: &str) -> String { input.replace('/', MAIN_SEPARATOR_STR) } macro_rules! func_tests { ($($name:ident: $func:ident for $input:expr => $output:expr)+) => { $( #[test] fn $name() { let input_path = PathBuf::from(&correct($input)); let output_string = OsString::from(correct($output)); assert_eq!($func(&input_path), output_string); } )+ } } func_tests! { remove_ext_simple: remove_extension for "foo.txt" => "foo" remove_ext_dir: remove_extension for "dir/foo.txt" => "dir/foo" hidden: remove_extension for ".foo" => ".foo" remove_ext_utf8: remove_extension for "πŸ’–.txt" => "πŸ’–" remove_ext_empty: remove_extension for "" => "" basename_simple: basename for "foo.txt" => "foo.txt" basename_dir: basename for "dir/foo.txt" => "foo.txt" basename_empty: basename for "" => "" basename_utf8_0: basename for "πŸ’–/foo.txt" => "foo.txt" basename_utf8_1: basename for "dir/πŸ’–.txt" => "πŸ’–.txt" dirname_simple: dirname for "foo.txt" => "." dirname_dir: dirname for "dir/foo.txt" => "dir" dirname_utf8_0: dirname for "πŸ’–/foo.txt" => "πŸ’–" dirname_utf8_1: dirname for "dir/πŸ’–.txt" => "dir" } #[test] #[cfg(windows)] fn dirname_root() { assert_eq!(dirname(&PathBuf::from("C:")), OsString::from("C:")); assert_eq!(dirname(&PathBuf::from("\\")), OsString::from("\\")); } #[test] #[cfg(not(windows))] fn dirname_root() { assert_eq!(dirname(&PathBuf::from("/")), OsString::from("/")); } }
rust
Apache-2.0
5f95a781212e3efbc6d91ae50e0ca0ce0693da50
2026-01-04T15:31:59.463143Z
false
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
45

Models trained or fine-tuned on Shuu12121/github-file-programs-dataset-rust