command
stringlengths 1
42
| description
stringlengths 29
182k
⌀ | name
stringlengths 7
64.9k
⌀ | synopsis
stringlengths 4
85.3k
⌀ | options
stringclasses 593
values | examples
stringclasses 455
values |
|---|---|---|---|---|---|
lpoptions
|
lpoptions displays or sets printer options and defaults. If no printer is specified using the -p option, the default printer is used as described in lp(1). If no -l, -o, or -r options are specified, the current options are reported on the standard output. Options set with the lpoptions command are used by the lp(1) and lpr(1) commands when submitting jobs. When run by the root user, lpoptions gets and sets default options and instances for all users in the /etc/cups/lpoptions file. Otherwise, the per-user defaults are managed in the ~/.cups/lpoptions file.
|
lpoptions - display or set printer options and defaults
|
lpoptions [ -E ] [ -h server[:port] ] -d destination[/instance] [ -l ] lpoptions [ -E ] [ -h server[:port] ] [ -p destination[/instance] ] -o option[=value] ... lpoptions [ -E ] [ -h server[:port] ] [ -p destination[/instance] ] -r option lpoptions [ -E ] [ -h server[:port] ] -x destination[/instance]
|
lpoptions supports the following options: -E Enables encryption when communicating with the CUPS server. -d destination[/instance] Sets the user default printer to destination. If instance is supplied then that particular instance is used. This option overrides the system default printer for the current user. -h server[:port] Uses an alternate server. -l Lists the printer specific options and their current settings. -o option[=value] Specifies a new option for the named destination. -p destination[/instance] Sets the destination and instance, if specified, for any options that follow. If the named instance does not exist then it is created. Destinations can only be created using the lpadmin(8) program. -r option Removes the specified option from the named destination. -x destination[/instance] Removes the options for the named destination and instance, if specified. If the named instance does not exist then this does nothing. Destinations can only be removed using the lpadmin(8) command. FILES ~/.cups/lpoptions - user defaults and instances created by non-root users. /etc/cups/lpoptions - system-wide defaults and instances created by the root user. CONFORMING TO The lpoptions command is unique to CUPS. SEE ALSO cancel(1), lp(1), lpadmin(8), lpr(1), lprm(1), CUPS Online Help (http://localhost:631/help) COPYRIGHT Copyright © 2007-2019 by Apple Inc. 26 April 2019 CUPS lpoptions(1)
| null |
tabs
|
The tabs utility displays a series of characters that clear the hardware terminal tab settings then initialises tab stops at specified positions, and optionally adjusts the margin. In the first synopsis form, the tab stops set depend on the command line options used, and may be one of the predefined formats or at regular intervals. In the second synopsis form, tab stops are set at positions n1, n2, etc. If a position is preceded by a ‘+’, it is relative to the previous position set. No more than 20 positions may be specified. If no tab stops are specified, the “standard” UNIX tab width of 8 is used. The options are as follows: -n Set a tab stop every n columns. If n is 0, the tab stops are cleared but no new ones are set. -a Assembler format (columns 1, 10, 16, 36, 72). -a2 Assembler format (columns 1, 10, 16, 40, 72). -c COBOL normal format (columns 1, 8, 12, 16, 20, 55) -c2 COBOL compact format (columns 1, 6, 10, 14, 49) -c3 COBOL compact format (columns 1, 6, 10, 14, 18, 22, 26, 30, 34, 38, 42, 46, 50, 54, 58, 62, 67). -f FORTRAN format (columns 1, 7, 11, 15, 19, 23). -p PL/1 format (columns 1, 5, 9, 13, 17, 21, 25, 29, 33, 37, 41, 45, 49, 53, 57, 61). -s SNOBOL format (columns 1, 10, 55). -u Assembler format (columns 1, 12, 20, 44). +m[n], +[n] Set an n character left margin, or 10 if n is omitted. -T type Output escape sequence for the terminal type type. ENVIRONMENT The LANG, LC_ALL, LC_CTYPE and TERM environment variables affect the execution of tabs as described in environ(7). The -T option overrides the setting of the TERM environment variable. If neither TERM nor the -T option are present, tabs will fail. EXIT STATUS The tabs utility exits 0 on success, and >0 if an error occurs. SEE ALSO expand(1), stty(1), tput(1), unexpand(1), termcap(5) STANDARDS The tabs utility conforms to IEEE Std 1003.1-2001 (“POSIX.1”). HISTORY /etc/tabs appeared in Version 1 AT&T UNIX which set tabs every 8 on Teletype model 37 and compatible printing terminals when printed. A tabs utility that worked on many terminals appeared in PWB UNIX. This implementation was introduced in FreeBSD 5.0. BUGS The current termcap(5) database does not define the ‘ML’ (set left soft margin) capability for any terminals. macOS 14.5 September 9, 2021 macOS 14.5
|
tabs – set terminal tabs
|
tabs [-n | -a | -a2 | -c | -c2 | -c3 | -f | -p | -s | -u] [+m[n]] [-T type] tabs [-T type] [+[n]] n1[,n2,...]
| null | null |
rview
|
Vim is a text editor that is upwards compatible to Vi. It can be used to edit all kinds of plain text. It is especially useful for editing programs. There are a lot of enhancements above Vi: multi level undo, multi windows and buffers, syntax highlighting, command line editing, filename completion, on-line help, visual selection, etc.. See ":help vi_diff.txt" for a summary of the differences between Vim and Vi. While running Vim a lot of help can be obtained from the on-line help system, with the ":help" command. See the ON-LINE HELP section below. Most often Vim is started to edit a single file with the command vim file More generally Vim is started with: vim [options] [filelist] If the filelist is missing, the editor will start with an empty buffer. Otherwise exactly one out of the following four may be used to choose one or more files to be edited. file .. A list of filenames. The first one will be the current file and read into the buffer. The cursor will be positioned on the first line of the buffer. You can get to the other files with the ":next" command. To edit a file that starts with a dash, precede the filelist with "--". - The file to edit is read from stdin. Commands are read from stderr, which should be a tty. -t {tag} The file to edit and the initial cursor position depends on a "tag", a sort of goto label. {tag} is looked up in the tags file, the associated file becomes the current file and the associated command is executed. Mostly this is used for C programs, in which case {tag} could be a function name. The effect is that the file containing that function becomes the current file and the cursor is positioned on the start of the function. See ":help tag-commands". -q [errorfile] Start in quickFix mode. The file [errorfile] is read and the first error is displayed. If [errorfile] is omitted, the filename is obtained from the 'errorfile' option (defaults to "AztecC.Err" for the Amiga, "errors.err" on other systems). Further errors can be jumped to with the ":cn" command. See ":help quickfix". Vim behaves differently, depending on the name of the command (the executable may still be the same file). vim The "normal" way, everything is default. ex Start in Ex mode. Go to Normal mode with the ":vi" command. Can also be done with the "-e" argument. view Start in read-only mode. You will be protected from writing the files. Can also be done with the "-R" argument. gvim gview The GUI version. Starts a new window. Can also be done with the "-g" argument. evim eview The GUI version in easy mode. Starts a new window. Can also be done with the "-y" argument. rvim rview rgvim rgview Like the above, but with restrictions. It will not be possible to start shell commands, or suspend Vim. Can also be done with the "-Z" argument.
|
vim - Vi IMproved, a programmer's text editor
|
vim [options] [file ..] vim [options] - vim [options] -t tag vim [options] -q [errorfile] ex view gvim gview evim eview rvim rview rgvim rgview
|
The options may be given in any order, before or after filenames. Options without an argument can be combined after a single dash. +[num] For the first file the cursor will be positioned on line "num". If "num" is missing, the cursor will be positioned on the last line. +/{pat} For the first file the cursor will be positioned in the line with the first occurrence of {pat}. See ":help search-pattern" for the available search patterns. +{command} -c {command} {command} will be executed after the first file has been read. {command} is interpreted as an Ex command. If the {command} contains spaces it must be enclosed in double quotes (this depends on the shell that is used). Example: vim "+set si" main.c Note: You can use up to 10 "+" or "-c" commands. -S {file} {file} will be sourced after the first file has been read. This is equivalent to -c "source {file}". {file} cannot start with '-'. If {file} is omitted "Session.vim" is used (only works when -S is the last argument). --cmd {command} Like using "-c", but the command is executed just before processing any vimrc file. You can use up to 10 of these commands, independently from "-c" commands. -A If Vim has been compiled with ARABIC support for editing right-to-left oriented files and Arabic keyboard mapping, this option starts Vim in Arabic mode, i.e. 'arabic' is set. Otherwise an error message is given and Vim aborts. -b Binary mode. A few options will be set that makes it possible to edit a binary or executable file. -C Compatible. Set the 'compatible' option. This will make Vim behave mostly like Vi, even though a .vimrc file exists. -d Start in diff mode. There should between two to eight file name arguments. Vim will open all the files and show differences between them. Works like vimdiff(1). -d {device}, -dev {device} Open {device} for use as a terminal. Only on the Amiga. Example: "-d con:20/30/600/150". -D Debugging. Go to debugging mode when executing the first command from a script. -e Start Vim in Ex mode, just like the executable was called "ex". -E Start Vim in improved Ex mode, just like the executable was called "exim". -f Foreground. For the GUI version, Vim will not fork and detach from the shell it was started in. On the Amiga, Vim is not restarted to open a new window. This option should be used when Vim is executed by a program that will wait for the edit session to finish (e.g. mail). On the Amiga the ":sh" and ":!" commands will not work. --nofork Foreground. For the GUI version, Vim will not fork and detach from the shell it was started in. -F If Vim has been compiled with FKMAP support for editing right-to-left oriented files and Farsi keyboard mapping, this option starts Vim in Farsi mode, i.e. 'fkmap' and 'rightleft' are set. Otherwise an error message is given and Vim aborts. -g If Vim has been compiled with GUI support, this option enables the GUI. If no GUI support was compiled in, an error message is given and Vim aborts. --gui-dialog-file {name} When using the GUI, instead of showing a dialog, write the title and message of the dialog to file {name}. The file is created or appended to. Only useful for testing, to avoid that the test gets stuck on a dialog that can't be seen. Without the GUI the argument is ignored. --help, -h, -? Give a bit of help about the command line arguments and options. After this Vim exits. -H If Vim has been compiled with RIGHTLEFT support for editing right-to-left oriented files and Hebrew keyboard mapping, this option starts Vim in Hebrew mode, i.e. 'hkmap' and 'rightleft' are set. Otherwise an error message is given and Vim aborts. -i {viminfo} Specifies the filename to use when reading or writing the viminfo file, instead of the default "~/.viminfo". This can also be used to skip the use of the .viminfo file, by giving the name "NONE". -L Same as -r. -l Lisp mode. Sets the 'lisp' and 'showmatch' options on. -m Modifying files is disabled. Resets the 'write' option. You can still modify the buffer, but writing a file is not possible. -M Modifications not allowed. The 'modifiable' and 'write' options will be unset, so that changes are not allowed and files can not be written. Note that these options can be set to enable making modifications. -N No-compatible mode. Resets the 'compatible' option. This will make Vim behave a bit better, but less Vi compatible, even though a .vimrc file does not exist. -n No swap file will be used. Recovery after a crash will be impossible. Handy if you want to edit a file on a very slow medium (e.g. floppy). Can also be done with ":set uc=0". Can be undone with ":set uc=200". -nb Become an editor server for NetBeans. See the docs for details. -o[N] Open N windows stacked. When N is omitted, open one window for each file. -O[N] Open N windows side by side. When N is omitted, open one window for each file. -p[N] Open N tab pages. When N is omitted, open one tab page for each file. -P {parent-title} Win32 GUI only: Specify the title of the parent application. When possible, Vim will run in an MDI window inside the application. {parent-title} must appear in the window title of the parent application. Make sure that it is specific enough. Note that the implementation is still primitive. It won't work with all applications and the menu doesn't work. -R Read-only mode. The 'readonly' option will be set. You can still edit the buffer, but will be prevented from accidentally overwriting a file. If you do want to overwrite a file, add an exclamation mark to the Ex command, as in ":w!". The -R option also implies the -n option (see above). The 'readonly' option can be reset with ":set noro". See ":help 'readonly'". -r List swap files, with information about using them for recovery. -r {file} Recovery mode. The swap file is used to recover a crashed editing session. The swap file is a file with the same filename as the text file with ".swp" appended. See ":help recovery". -s Silent mode. Only when started as "Ex" or when the "-e" option was given before the "-s" option. -s {scriptin} The script file {scriptin} is read. The characters in the file are interpreted as if you had typed them. The same can be done with the command ":source! {scriptin}". If the end of the file is reached before the editor exits, further characters are read from the keyboard. -T {terminal} Tells Vim the name of the terminal you are using. Only required when the automatic way doesn't work. Should be a terminal known to Vim (builtin) or defined in the termcap or terminfo file. --not-a-term Tells Vim that the user knows that the input and/or output is not connected to a terminal. This will avoid the warning and the two second delay that would happen. --ttyfail When stdin or stdout is not a a terminal (tty) then exit right away. -u {vimrc} Use the commands in the file {vimrc} for initializations. All the other initializations are skipped. Use this to edit a special kind of files. It can also be used to skip all initializations by giving the name "NONE". See ":help initialization" within vim for more details. -U {gvimrc} Use the commands in the file {gvimrc} for GUI initializations. All the other GUI initializations are skipped. It can also be used to skip all GUI initializations by giving the name "NONE". See ":help gui-init" within vim for more details. -V[N] Verbose. Give messages about which files are sourced and for reading and writing a viminfo file. The optional number N is the value for 'verbose'. Default is 10. -V[N]{filename} Like -V and set 'verbosefile' to {filename}. The result is that messages are not displayed but written to the file {filename}. {filename} must not start with a digit. --log {filename} If Vim has been compiled with eval and channel feature, start logging and write entries to {filename}. This works like calling ch_logfile({filename}, 'ao') very early during startup. -v Start Vim in Vi mode, just like the executable was called "vi". This only has effect when the executable is called "ex". -w{number} Set the 'window' option to {number}. -w {scriptout} All the characters that you type are recorded in the file {scriptout}, until you exit Vim. This is useful if you want to create a script file to be used with "vim -s" or ":source!". If the {scriptout} file exists, characters are appended. -W {scriptout} Like -w, but an existing file is overwritten. -x Use encryption when writing files. Will prompt for a crypt key. -X Don't connect to the X server. Shortens startup time in a terminal, but the window title and clipboard will not be used. -y Start Vim in easy mode, just like the executable was called "evim" or "eview". Makes Vim behave like a click-and-type editor. -Z Restricted mode. Works like the executable starts with "r". -- Denotes the end of the options. Arguments after this will be handled as a file name. This can be used to edit a filename that starts with a '-'. --clean Do not use any personal configuration (vimrc, plugins, etc.). Useful to see if a problem reproduces with a clean Vim setup. --echo-wid GTK GUI only: Echo the Window ID on stdout. --literal Take file name arguments literally, do not expand wildcards. This has no effect on Unix where the shell expands wildcards. --noplugin Skip loading plugins. Implied by -u NONE. --remote Connect to a Vim server and make it edit the files given in the rest of the arguments. If no server is found a warning is given and the files are edited in the current Vim. --remote-expr {expr} Connect to a Vim server, evaluate {expr} in it and print the result on stdout. --remote-send {keys} Connect to a Vim server and send {keys} to it. --remote-silent As --remote, but without the warning when no server is found. --remote-wait As --remote, but Vim does not exit until the files have been edited. --remote-wait-silent As --remote-wait, but without the warning when no server is found. --serverlist List the names of all Vim servers that can be found. --servername {name} Use {name} as the server name. Used for the current Vim, unless used with a --remote argument, then it's the name of the server to connect to. --socketid {id} GTK GUI only: Use the GtkPlug mechanism to run gvim in another window. --startuptime {file} During startup write timing messages to the file {fname}. --version Print version information and exit. --windowid {id} Win32 GUI only: Make gvim try to use the window {id} as a parent, so that it runs inside that window. ON-LINE HELP Type ":help" in Vim to get started. Type ":help subject" to get help on a specific subject. For example: ":help ZZ" to get help for the "ZZ" command. Use <Tab> and CTRL-D to complete subjects (":help cmdline-completion"). Tags are present to jump from one place to another (sort of hypertext links, see ":help"). All documentation files can be viewed in this way, for example ":help syntax.txt". FILES /usr/local/share/vim/vim??/doc/*.txt The Vim documentation files. Use ":help doc-file-list" to get the complete list. vim?? is short version number, like vim91 for Vim 9.1 /usr/local/share/vim/vim??/doc/tags The tags file used for finding information in the documentation files. /usr/local/share/vim/vim??/syntax/syntax.vim System wide syntax initializations. /usr/local/share/vim/vim??/syntax/*.vim Syntax files for various languages. /usr/local/share/vim/vimrc System wide Vim initializations. ~/.vimrc, ~/.vim/vimrc, $XDG_CONFIG_HOME/vim/vimrc Your personal Vim initializations (first one found is used). /usr/local/share/vim/gvimrc System wide gvim initializations. ~/.gvimrc, ~/.vim/gvimrc, $XDG_CONFIG_HOME/vim/gvimrc Your personal gvim initializations (first one found is used). /usr/local/share/vim/vim??/optwin.vim Script used for the ":options" command, a nice way to view and set options. /usr/local/share/vim/vim??/menu.vim System wide menu initializations for gvim. /usr/local/share/vim/vim??/bugreport.vim Script to generate a bug report. See ":help bugs". /usr/local/share/vim/vim??/filetype.vim Script to detect the type of a file by its name. See ":help 'filetype'". /usr/local/share/vim/vim??/scripts.vim Script to detect the type of a file by its contents. See ":help 'filetype'". /usr/local/share/vim/vim??/print/*.ps Files used for PostScript printing. For recent info read the VIM home page: <URL:http://www.vim.org/> SEE ALSO vimtutor(1) AUTHOR Most of Vim was made by Bram Moolenaar, with a lot of help from others. See ":help credits" in Vim. Vim is based on Stevie, worked on by: Tim Thompson, Tony Andrews and G.R. (Fred) Walter. Although hardly any of the original code remains. BUGS Probably. See ":help todo" for a list of known problems. Note that a number of things that may be regarded as bugs by some, are in fact caused by a too-faithful reproduction of Vi's behaviour. And if you think other things are bugs "because Vi does it differently", you should take a closer look at the vi_diff.txt file (or type :help vi_diff.txt when in Vim). Also have a look at the 'compatible' and 'cpoptions' options. 2024 Jun 04 VIM(1)
| null |
lsvfs
|
The lsvfs command lists information about the currently loaded virtual file system modules. When vfsname arguments are given, lsvfs lists information about the specified VFS modules. Otherwise, lsvfs lists all currently loaded modules. The information is as follows: Filesystem the name of the file system, as would be used in the type parameter to mount(2) and the -t option to mount(8) Refs the number of references to this VFS; i.e., the number of currently mounted file systems of this type Flags flag bits.
|
lsvfs – list installed virtual file systems
|
lsvfs [vfsname ...]
| null |
Show information about the ‘apfs’ and ‘devfs’ filesystems and check the number of mounts for the former: $ lsvfs apfs devfs Filesystem Refs Flags -------------------------------- ----- --------------- apfs 5 local devfs 1 $ mount -t apfs | wc -l 5 SEE ALSO mount(2), getvfsbyname(3), mount(8) HISTORY A lsvfs command appeared in FreeBSD 2.0. macOS 14.5 February 10, 2022 macOS 14.5
|
zgrep
|
The grep utility searches any given input files, selecting lines that match one or more patterns. By default, a pattern matches an input line if the regular expression (RE) in the pattern matches the input line without its trailing newline. An empty expression matches every line. Each input line that matches at least one of the patterns is written to the standard output. grep is used for simple patterns and basic regular expressions (BREs); egrep can handle extended regular expressions (EREs). See re_format(7) for more information on regular expressions. fgrep is quicker than both grep and egrep, but can only handle fixed patterns (i.e., it does not interpret regular expressions). Patterns may consist of one or more lines, allowing any of the pattern lines to match a portion of the input. zgrep, zegrep, and zfgrep act like grep, egrep, and fgrep, respectively, but accept input files compressed with the compress(1) or gzip(1) compression utilities. bzgrep, bzegrep, and bzfgrep act like grep, egrep, and fgrep, respectively, but accept input files compressed with the bzip2(1) compression utility. The following options are available: -A num, --after-context=num Print num lines of trailing context after each match. See also the -B and -C options. -a, --text Treat all files as ASCII text. Normally grep will simply print “Binary file ... matches” if files contain binary characters. Use of this option forces grep to output lines matching the specified pattern. -B num, --before-context=num Print num lines of leading context before each match. See also the -A and -C options. -b, --byte-offset The offset in bytes of a matched pattern is displayed in front of the respective matched line. -C num, --context=num Print num lines of leading and trailing context surrounding each match. See also the -A and -B options. -c, --count Only a count of selected lines is written to standard output. --colour=[when], --color=[when] Mark up the matching text with the expression stored in the GREP_COLOR environment variable. The possible values of when are “never”, “always” and “auto”. -D action, --devices=action Specify the demanded action for devices, FIFOs and sockets. The default action is “read”, which means, that they are read as if they were normal files. If the action is set to “skip”, devices are silently skipped. -d action, --directories=action Specify the demanded action for directories. It is “read” by default, which means that the directories are read in the same manner as normal files. Other possible values are “skip” to silently ignore the directories, and “recurse” to read them recursively, which has the same effect as the -R and -r option. -E, --extended-regexp Interpret pattern as an extended regular expression (i.e., force grep to behave as egrep). -e pattern, --regexp=pattern Specify a pattern used during the search of the input: an input line is selected if it matches any of the specified patterns. This option is most useful when multiple -e options are used to specify multiple patterns, or when a pattern begins with a dash (‘-’). --exclude pattern If specified, it excludes files matching the given filename pattern from the search. Note that --exclude and --include patterns are processed in the order given. If a name matches multiple patterns, the latest matching rule wins. If no --include pattern is specified, all files are searched that are not excluded. Patterns are matched to the full path specified, not only to the filename component. --exclude-dir pattern If -R is specified, it excludes directories matching the given filename pattern from the search. Note that --exclude-dir and --include-dir patterns are processed in the order given. If a name matches multiple patterns, the latest matching rule wins. If no --include-dir pattern is specified, all directories are searched that are not excluded. -F, --fixed-strings Interpret pattern as a set of fixed strings (i.e., force grep to behave as fgrep). -f file, --file=file Read one or more newline separated patterns from file. Empty pattern lines match every input line. Newlines are not considered part of a pattern. If file is empty, nothing is matched. -G, --basic-regexp Interpret pattern as a basic regular expression (i.e., force grep to behave as traditional grep). -H Always print filename headers with output lines. -h, --no-filename Never print filename headers (i.e., filenames) with output lines. --help Print a brief help message. -I Ignore binary files. This option is equivalent to the “--binary-files=without-match” option. -i, --ignore-case Perform case insensitive matching. By default, grep is case sensitive. --include pattern If specified, only files matching the given filename pattern are searched. Note that --include and --exclude patterns are processed in the order given. If a name matches multiple patterns, the latest matching rule wins. Patterns are matched to the full path specified, not only to the filename component. --include-dir pattern If -R is specified, only directories matching the given filename pattern are searched. Note that --include-dir and --exclude-dir patterns are processed in the order given. If a name matches multiple patterns, the latest matching rule wins. -J, --bz2decompress Decompress the bzip2(1) compressed file before looking for the text. -L, --files-without-match Only the names of files not containing selected lines are written to standard output. Pathnames are listed once per file searched. If the standard input is searched, the string “(standard input)” is written unless a --label is specified. -l, --files-with-matches Only the names of files containing selected lines are written to standard output. grep will only search a file until a match has been found, making searches potentially less expensive. Pathnames are listed once per file searched. If the standard input is searched, the string “(standard input)” is written unless a --label is specified. --label Label to use in place of “(standard input)” for a file name where a file name would normally be printed. This option applies to -H, -L, and -l. --mmap Use mmap(2) instead of read(2) to read input, which can result in better performance under some circumstances but can cause undefined behaviour. -M, --lzma Decompress the LZMA compressed file before looking for the text. -m num, --max-count=num Stop reading the file after num matches. -n, --line-number Each output line is preceded by its relative line number in the file, starting at line 1. The line number counter is reset for each file processed. This option is ignored if -c, -L, -l, or -q is specified. --null Prints a zero-byte after the file name. -O If -R is specified, follow symbolic links only if they were explicitly listed on the command line. The default is not to follow symbolic links. -o, --only-matching Prints only the matching part of the lines. -p If -R is specified, no symbolic links are followed. This is the default. -q, --quiet, --silent Quiet mode: suppress normal output. grep will only search a file until a match has been found, making searches potentially less expensive. -R, -r, --recursive Recursively search subdirectories listed. (i.e., force grep to behave as rgrep). -S If -R is specified, all symbolic links are followed. The default is not to follow symbolic links. -s, --no-messages Silent mode. Nonexistent and unreadable files are ignored (i.e., their error messages are suppressed). -U, --binary Search binary files, but do not attempt to print them. -u This option has no effect and is provided only for compatibility with GNU grep. -V, --version Display version information and exit. -v, --invert-match Selected lines are those not matching any of the specified patterns. -w, --word-regexp The expression is searched for as a word (as if surrounded by ‘[[:<:]]’ and ‘[[:>:]]’; see re_format(7)). This option has no effect if -x is also specified. -x, --line-regexp Only input lines selected against an entire fixed string or regular expression are considered to be matching lines. -y Equivalent to -i. Obsoleted. -z, --null-data Treat input and output data as sequences of lines terminated by a zero-byte instead of a newline. -X, --xz Decompress the xz(1) compressed file before looking for the text. -Z, --decompress Force grep to behave as zgrep. --binary-files=value Controls searching and printing of binary files. Options are: binary (default) Search binary files but do not print them. without-match Do not search binary files. text Treat all files as text. --line-buffered Force output to be line buffered. By default, output is line buffered when standard output is a terminal and block buffered otherwise. If no file arguments are specified, the standard input is used. Additionally, “-” may be used in place of a file name, anywhere that a file name is accepted, to read from standard input. This includes both -f and file arguments. ENVIRONMENT GREP_OPTIONS May be used to specify default options that will be placed at the beginning of the argument list. Backslash-escaping is not supported, unlike the behavior in GNU grep. EXIT STATUS The grep utility exits with one of the following values: 0 One or more lines were selected. 1 No lines were selected. >1 An error occurred.
|
grep, egrep, fgrep, rgrep, bzgrep, bzegrep, bzfgrep, zgrep, zegrep, zfgrep – file pattern searcher
|
grep [-abcdDEFGHhIiJLlMmnOopqRSsUVvwXxZz] [-A num] [-B num] [-C num] [-e pattern] [-f file] [--binary-files=value] [--color[=when]] [--colour[=when]] [--context=num] [--label] [--line-buffered] [--null] [pattern] [file ...]
| null |
- Find all occurrences of the pattern ‘patricia’ in a file: $ grep 'patricia' myfile - Same as above but looking only for complete words: $ grep -w 'patricia' myfile - Count occurrences of the exact pattern ‘FOO’ : $ grep -c FOO myfile - Same as above but ignoring case: $ grep -c -i FOO myfile - Find all occurrences of the pattern ‘.Pp’ at the beginning of a line: $ grep '^\.Pp' myfile The apostrophes ensure the entire expression is evaluated by grep instead of by the user's shell. The caret ‘^’ matches the null string at the beginning of a line, and the ‘\’ escapes the ‘.’, which would otherwise match any character. - Find all lines in a file which do not contain the words ‘foo’ or ‘bar’: $ grep -v -e 'foo' -e 'bar' myfile - Peruse the file ‘calendar’ looking for either 19, 20, or 25 using extended regular expressions: $ egrep '19|20|25' calendar - Show matching lines and the name of the ‘*.h’ files which contain the pattern ‘FIXME’. Do the search recursively from the /usr/src/sys/arm directory $ grep -H -R FIXME --include="*.h" /usr/src/sys/arm/ - Same as above but show only the name of the matching file: $ grep -l -R FIXME --include="*.h" /usr/src/sys/arm/ - Show lines containing the text ‘foo’. The matching part of the output is colored and every line is prefixed with the line number and the offset in the file for those lines that matched. $ grep -b --colour -n foo myfile - Show lines that match the extended regular expression patterns read from the standard input: $ echo -e 'Free\nBSD\nAll.*reserved' | grep -E -f - myfile - Show lines from the output of the pciconf(8) command matching the specified extended regular expression along with three lines of leading context and one line of trailing context: $ pciconf -lv | grep -B3 -A1 -E 'class.*=.*storage' - Suppress any output and use the exit status to show an appropriate message: $ grep -q foo myfile && echo File matches SEE ALSO bzip2(1), compress(1), ed(1), ex(1), gzip(1), sed(1), xz(1), zgrep(1), re_format(7) STANDARDS The grep utility is compliant with the IEEE Std 1003.1-2008 (“POSIX.1”) specification. The flags [-AaBbCDdGHhILmopRSUVw] are extensions to that specification, and the behaviour of the -f flag when used with an empty pattern file is left undefined. All long options are provided for compatibility with GNU versions of this utility. Historic versions of the grep utility also supported the flags [-ruy]. This implementation supports those options; however, their use is strongly discouraged. HISTORY The grep command first appeared in Version 6 AT&T UNIX. BUGS The grep utility does not normalize Unicode input, so a pattern containing composed characters will not match decomposed input, and vice versa. macOS 14.5 November 10, 2021 macOS 14.5
|
dbiprof
|
This tool is a command-line client for the DBI::ProfileData. It allows you to analyze the profile data file produced by DBI::ProfileDumper and produce various useful reports.
|
dbiprof - command-line client for DBI::ProfileData
|
See a report of the ten queries with the longest total runtime in the profile dump file prof1.out: dbiprof prof1.out See the top 10 most frequently run queries in the profile file dbi.prof (the default): dbiprof --sort count See the same report with 15 entries: dbiprof --sort count --number 15
|
This program accepts the following options: --number N Produce this many items in the report. Defaults to 10. If set to "all" then all results are shown. --sort field Sort results by the given field. Sorting by multiple fields isn't currently supported (patches welcome). The available sort fields are: total Sorts by total time run time across all runs. This is the default sort. longest Sorts by the longest single run. count Sorts by total number of runs. first Sorts by the time taken in the first run. shortest Sorts by the shortest single run. key1 Sorts by the value of the first element in the Path, which should be numeric. You can also sort by "key2" and "key3". --reverse Reverses the selected sort. For example, to see a report of the shortest overall time: dbiprof --sort total --reverse --match keyN=value Consider only items where the specified key matches the given value. Keys are numbered from 1. For example, let's say you used a DBI::Profile Path of: [ DBIprofile_Statement, DBIprofile_Methodname ] And called dbiprof as in: dbiprof --match key2=execute Your report would only show execute queries, leaving out prepares, fetches, etc. If the value given starts and ends with slashes ("/") then it will be treated as a regular expression. For example, to only include SELECT queries where key1 is the statement: dbiprof --match key1=/^SELECT/ By default the match expression is matched case-insensitively, but this can be changed with the --case-sensitive option. --exclude keyN=value Remove items for where the specified key matches the given value. For example, to exclude all prepare entries where key2 is the method name: dbiprof --exclude key2=prepare Like "--match", If the value given starts and ends with slashes ("/") then it will be treated as a regular expression. For example, to exclude UPDATE queries where key1 is the statement: dbiprof --match key1=/^UPDATE/ By default the exclude expression is matched case-insensitively, but this can be changed with the --case-sensitive option. --case-sensitive Using this option causes --match and --exclude to work case- sensitively. Defaults to off. --delete Sets the "DeleteFiles" option to DBI::ProfileData which causes the files to be deleted after reading. See DBI::ProfileData for more details. --dumpnodes Print the list of nodes in the form of a perl data structure. Use the "-sort" option if you want the list sorted. --version Print the dbiprof version number and exit. AUTHOR Sam Tregar <sam@tregar.com> COPYRIGHT AND LICENSE Copyright (C) 2002 Sam Tregar This program is free software; you can redistribute it and/or modify it under the same terms as Perl 5 itself. SEE ALSO DBI::ProfileDumper, DBI::Profile, DBI. perl v5.34.0 2024-04-13 DBIPROF(1)
| null |
zmore
|
zmore is a filter that allows the viewing of files compressed with Lempel-Ziv encoding. Such files generally have a “Z” or “gz” extension (both the compress(1) and gzip(1) formats are supported). Any flags that are specified are passed to the user's preferred PAGER (which is /usr/bin/more by default). zless is equivalent to zmore but uses less(1) as a pager instead of more(1). When multiple files are specified, zmore will pause at the end of each file and present the following prompt to the user: prev_file (END) - Next: next_file Where prev_file is the file that was just displayed and next_file is the next file to be displayed. The following keys are recognized at the prompt: e or q quit zmore. s skip the next file (or exit if the next file is the last). If no files are specified, zmore will read from the standard input. In this mode zmore will assume gzip(1) style compression since there is no suffix on which to make a decision. ENVIRONMENT PAGER Program used to display files. If unset, /usr/bin/more is used (zmore) or /usr/bin/less (zless). SEE ALSO compress(1), less(1), more(1) macOS 14.5 October 22, 2014 macOS 14.5
|
zmore, zless – view compressed files
|
zmore [flags] [file ...] zless [flags] [file ...]
| null | null |
pod2usage5.30
|
pod2usage will read the given input file looking for pod documentation and will print the corresponding usage message. If no input file is specified then standard input is read. pod2usage invokes the pod2usage() function in the Pod::Usage module. Please see "pod2usage()" in Pod::Usage. SEE ALSO Pod::Usage, pod2text(1) AUTHOR Please report bugs using <http://rt.cpan.org>. Brad Appleton <bradapp@enteract.com> Based on code for pod2text(1) written by Tom Christiansen <tchrist@mox.perl.com> perl v5.30.3 2024-04-13 POD2USAGE(1)
|
pod2usage - print usage messages from embedded pod docs in files
|
pod2usage [-help] [-man] [-exit exitval] [-output outfile] [-verbose level] [-pathlist dirlist] [-formatter module] [-utf8] file OPTIONS AND ARGUMENTS -help Print a brief help message and exit. -man Print this command's manual page and exit. -exit exitval The exit status value to return. -output outfile The output file to print to. If the special names "-" or ">&1" or ">&STDOUT" are used then standard output is used. If ">&2" or ">&STDERR" is used then standard error is used. -verbose level The desired level of verbosity to use: 1 : print SYNOPSIS only 2 : print SYNOPSIS sections and any OPTIONS/ARGUMENTS sections 3 : print the entire manpage (similar to running pod2text) -pathlist dirlist Specifies one or more directories to search for the input file if it was not supplied with an absolute path. Each directory path in the given list should be separated by a ':' on Unix (';' on MSWin32 and DOS). -formatter module Which text formatter to use. Default is Pod::Text, or for very old Perl versions Pod::PlainText. An alternative would be e.g. Pod::Text::Termcap. -utf8 This option assumes that the formatter (see above) understands the option "utf8". It turns on generation of utf8 output. file The pathname of a file containing pod documentation to be output in usage message format (defaults to standard input).
| null | null |
gcov
|
The llvm-cov tool shows code coverage information for programs that are instrumented to emit profile data. It can be used to work with gcov-style coverage or with clang's instrumentation based profiling. If the program is invoked with a base name of gcov, it will behave as if the llvm-cov gcov command were called. Otherwise, a command should be provided. COMMANDS • gcov • show • report • export GCOV COMMAND The llvm-cov gcov tool reads code coverage data files and displays the coverage information for a specified source file. It is compatible with the gcov tool from version 4.2 of GCC and may also be compatible with some later versions of gcov. To use llvm-cov gcov, you must first build an instrumented version of your application that collects coverage data as it runs. Compile with the -fprofile-arcs and -ftest-coverage options to add the instrumentation. (Alternatively, you can use the --coverage option, which includes both of those other options.) At the time you compile the instrumented code, a .gcno data file will be generated for each object file. These .gcno files contain half of the coverage data. The other half of the data comes from .gcda files that are generated when you run the instrumented program, with a separate .gcda file for each object file. Each time you run the program, the execution counts are summed into any existing .gcda files, so be sure to remove any old files if you do not want their contents to be included. By default, the .gcda files are written into the same directory as the object files, but you can override that by setting the GCOV_PREFIX and GCOV_PREFIX_STRIP environment variables. The GCOV_PREFIX_STRIP variable specifies a number of directory components to be removed from the start of the absolute path to the object file directory. After stripping those directories, the prefix from the GCOV_PREFIX variable is added. These environment variables allow you to run the instrumented program on a machine where the original object file directories are not accessible, but you will then need to copy the .gcda files back to the object file directories where llvm-cov gcov expects to find them. Once you have generated the coverage data files, run llvm-cov gcov for each main source file where you want to examine the coverage results. This should be run from the same directory where you previously ran the compiler. The results for the specified source file are written to a file named by appending a .gcov suffix. A separate output file is also created for each file included by the main source file, also with a .gcov suffix added. The basic content of an .gcov output file is a copy of the source file with an execution count and line number prepended to every line. The execution count is shown as - if a line does not contain any executable code. If a line contains code but that code was never executed, the count is displayed as #####. The llvm-cov show command shows line by line coverage of the binaries BIN,... using the profile data PROFILE. It can optionally be filtered to only show the coverage for the files listed in SOURCES. BIN may be an executable, object file, dynamic library, or archive (thin or otherwise). To use llvm-cov show, you need a program that is compiled with instrumentation to emit profile and coverage data. To build such a program with clang use the -fprofile-instr-generate and -fcoverage-mapping flags. If linking with the clang driver, pass -fprofile-instr-generate to the link stage to make sure the necessary runtime libraries are linked in. The coverage information is stored in the built executable or library itself, and this is what you should pass to llvm-cov show as a BIN argument. The profile data is generated by running this instrumented program normally. When the program exits it will write out a raw profile file, typically called default.profraw, which can be converted to a format that is suitable for the PROFILE argument using the llvm-profdata merge tool. The llvm-cov report command displays a summary of the coverage of the binaries BIN,... using the profile data PROFILE. It can optionally be filtered to only show the coverage for the files listed in SOURCES. BIN may be an executable, object file, dynamic library, or archive (thin or otherwise). If no source files are provided, a summary line is printed for each file in the coverage data. If any files are provided, summaries can be shown for each function in the listed files if the -show-functions option is enabled. For information on compiling programs for coverage and generating profile data, see SHOW COMMAND. The llvm-cov export command exports coverage data of the binaries BIN,... using the profile data PROFILE in either JSON or lcov trace file format. When exporting JSON, the regions, functions, branches, expansions, and summaries of the coverage data will be exported. When exporting an lcov trace file, the line-based coverage, branch coverage, and summaries will be exported. The exported data can optionally be filtered to only export the coverage for the files listed in SOURCES. For information on compiling programs for coverage and generating profile data, see SHOW COMMAND.
|
llvm-cov - emit coverage information
|
llvm-cov command [args...] llvm-cov gcov [options] SOURCEFILE llvm-cov show [options] -instr-profile PROFILE BIN [-object BIN,...] [[-object BIN]] [SOURCES] llvm-cov report [options] -instr-profile PROFILE BIN [-object BIN,...] [[-object BIN]] [SOURCES] llvm-cov export [options] -instr-profile PROFILE BIN [-object BIN,...] [[-object BIN]] [SOURCES]
|
-a, --all-blocks Display all basic blocks. If there are multiple blocks for a single line of source code, this option causes llvm-cov to show the count for each block instead of just one count for the entire line. -b, --branch-probabilities Display conditional branch probabilities and a summary of branch information. -c, --branch-counts Display branch counts instead of probabilities (requires -b). -m, --demangled-names Demangle function names. -f, --function-summaries Show a summary of coverage for each function instead of just one summary for an entire source file. --help Display available options (--help-hidden for more). -l, --long-file-names For coverage output of files included from the main source file, add the main file name followed by ## as a prefix to the output file names. This can be combined with the --preserve-paths option to use complete paths for both the main file and the included file. -n, --no-output Do not output any .gcov files. Summary information is still displayed. -o <DIR|FILE>, --object-directory=<DIR>, --object-file=<FILE> Find objects in DIR or based on FILE's path. If you specify a particular object file, the coverage data files are expected to have the same base name with .gcno and .gcda extensions. If you specify a directory, the files are expected in that directory with the same base name as the source file. -p, --preserve-paths Preserve path components when naming the coverage output files. In addition to the source file name, include the directories from the path to that file. The directories are separate by # characters, with . directories removed and .. directories replaced by ^ characters. When used with the --long-file-names option, this applies to both the main file name and the included file name. -r Only dump files with relative paths or absolute paths with the prefix specified by -s. -s <string> Source prefix to elide. -t, --stdout Print to stdout instead of producing .gcov files. -u, --unconditional-branches Include unconditional branches in the output for the --branch-probabilities option. -version Display the version of llvm-cov. -x, --hash-filenames Use md5 hash of file name when naming the coverage output files. The source file name will be suffixed by ## followed by MD5 hash calculated for it. EXIT STATUS llvm-cov gcov returns 1 if it cannot read input files. Otherwise, it exits with zero. SHOW COMMAND -show-branches=<VIEW> Show coverage for branch conditions in terms of either count or percentage. The supported views are: "count", "percent". -show-line-counts Show the execution counts for each line. Defaults to true, unless another -show option is used. -show-expansions Expand inclusions, such as preprocessor macros or textual inclusions, inline in the display of the source file. Defaults to false. -show-instantiations For source regions that are instantiated multiple times, such as templates in C++, show each instantiation separately as well as the combined summary. Defaults to true. -show-regions Show the execution counts for each region by displaying a caret that points to the character where the region starts. Defaults to false. -show-line-counts-or-regions Show the execution counts for each line if there is only one region on the line, but show the individual regions if there are multiple on the line. Defaults to false. -use-color Enable or disable color output. By default this is autodetected. -arch=[*NAMES*] Specify a list of architectures such that the Nth entry in the list corresponds to the Nth specified binary. If the covered object is a universal binary, this specifies the architecture to use. It is an error to specify an architecture that is not included in the universal binary or to use an architecture that does not match a non-universal binary. -name=<NAME> Show code coverage only for functions with the given name. -name-allowlist=<FILE> Show code coverage only for functions listed in the given file. Each line in the file should start with allowlist_fun:, immediately followed by the name of the function to accept. This name can be a wildcard expression. -name-regex=<PATTERN> Show code coverage only for functions that match the given regular expression. -ignore-filename-regex=<PATTERN> Skip source code files with file paths that match the given regular expression. -format=<FORMAT> Use the specified output format. The supported formats are: "text", "html". -tab-size=<TABSIZE> Replace tabs with <TABSIZE> spaces when preparing reports. Currently, this is only supported for the html format. -output-dir=PATH Specify a directory to write coverage reports into. If the directory does not exist, it is created. When used in function view mode (i.e when -name or -name-regex are used to select specific functions), the report is written to PATH/functions.EXTENSION. When used in file view mode, a report for each file is written to PATH/REL_PATH_TO_FILE.EXTENSION. -Xdemangler=<TOOL>|<TOOL-OPTION> Specify a symbol demangler. This can be used to make reports more human-readable. This option can be specified multiple times to supply arguments to the demangler (e.g -Xdemangler c++filt -Xdemangler -n for C++). The demangler is expected to read a newline-separated list of symbols from stdin and write a newline-separated list of the same length to stdout. -num-threads=N, -j=N Use N threads to write file reports (only applicable when -output-dir is specified). When N=0, llvm-cov auto-detects an appropriate number of threads to use. This is the default. -compilation-dir=<dir> Directory used as a base for relative coverage mapping paths. Only applicable when binaries have been compiled with one of -fcoverage-prefix-map -fcoverage-compilation-dir, or -ffile-compilation-dir. -line-coverage-gt=<N> Show code coverage only for functions with line coverage greater than the given threshold. -line-coverage-lt=<N> Show code coverage only for functions with line coverage less than the given threshold. -region-coverage-gt=<N> Show code coverage only for functions with region coverage greater than the given threshold. -region-coverage-lt=<N> Show code coverage only for functions with region coverage less than the given threshold. -path-equivalence=<from>,<to> Map the paths in the coverage data to local source file paths. This allows you to generate the coverage data on one machine, and then use llvm-cov on a different machine where you have the same files on a different path. -coverage-watermark=<high>,<low> Set high and low watermarks for coverage in html format output. This allows you to set the high and low watermark of coverage as desired, green when coverage >= high, red when coverage < low, and yellow otherwise. Both high and low should be between 0-100 and high > low. REPORT COMMAND -use-color[=VALUE] Enable or disable color output. By default this is autodetected. -arch=<name> If the covered binary is a universal binary, select the architecture to use. It is an error to specify an architecture that is not included in the universal binary or to use an architecture that does not match a non-universal binary. -show-region-summary Show statistics for all regions. Defaults to true. -show-branch-summary Show statistics for all branch conditions. Defaults to true. -show-functions Show coverage summaries for each function. Defaults to false. -show-instantiation-summary Show statistics for all function instantiations. Defaults to false. -ignore-filename-regex=<PATTERN> Skip source code files with file paths that match the given regular expression. -compilation-dir=<dir> Directory used as a base for relative coverage mapping paths. Only applicable when binaries have been compiled with one of -fcoverage-prefix-map -fcoverage-compilation-dir, or -ffile-compilation-dir. EXPORT COMMAND -arch=<name> If the covered binary is a universal binary, select the architecture to use. It is an error to specify an architecture that is not included in the universal binary or to use an architecture that does not match a non-universal binary. -format=<FORMAT> Use the specified output format. The supported formats are: "text" (JSON), "lcov". -summary-only Export only summary information for each file in the coverage data. This mode will not export coverage information for smaller units such as individual functions or regions. The result will contain the same information as produced by the llvm-cov report command, but presented in JSON or lcov format rather than text. -ignore-filename-regex=<PATTERN> Skip source code files with file paths that match the given regular expression. -skip-expansions Skip exporting macro expansion coverage data. -skip-functions Skip exporting per-function coverage data. -num-threads=N, -j=N Use N threads to export coverage data. When N=0, llvm-cov auto-detects an appropriate number of threads to use. This is the default. -compilation-dir=<dir> Directory used as a base for relative coverage mapping paths. Only applicable when binaries have been compiled with one of -fcoverage-prefix-map -fcoverage-compilation-dir, or -ffile-compilation-dir. AUTHOR Maintained by the LLVM Team (https://llvm.org/). COPYRIGHT 2003-2024, LLVM Project 11 2024-01-28 LLVM-COV(1)
| null |
parldyn5.34
| null | null | null | null | null |
package-stash-conflicts5.30
| null |
package-stash-conflicts VERSION version 0.38 SUPPORT Bugs may be submitted through the RT bug tracker <https://rt.cpan.org/Public/Dist/Display.html?Name=Package-Stash> (or bug-Package-Stash@rt.cpan.org <mailto:bug-Package-Stash@rt.cpan.org>). AUTHOR Jesse Luehrs <doy@tozt.net> COPYRIGHT AND LICENSE This software is copyright (c) 2018 by Jesse Luehrs. This is free software; you can redistribute it and/or modify it under the same terms as the Perl 5 programming language system itself. perl v5.30.3 2018-12-31 PACKAGE-STASH-CONFLICTS(1)
| null | null | null |
dbiprof5.30
|
This tool is a command-line client for the DBI::ProfileData. It allows you to analyze the profile data file produced by DBI::ProfileDumper and produce various useful reports.
|
dbiprof - command-line client for DBI::ProfileData
|
See a report of the ten queries with the longest total runtime in the profile dump file prof1.out: dbiprof prof1.out See the top 10 most frequently run queries in the profile file dbi.prof (the default): dbiprof --sort count See the same report with 15 entries: dbiprof --sort count --number 15
|
This program accepts the following options: --number N Produce this many items in the report. Defaults to 10. If set to "all" then all results are shown. --sort field Sort results by the given field. Sorting by multiple fields isn't currently supported (patches welcome). The available sort fields are: total Sorts by total time run time across all runs. This is the default sort. longest Sorts by the longest single run. count Sorts by total number of runs. first Sorts by the time taken in the first run. shortest Sorts by the shortest single run. key1 Sorts by the value of the first element in the Path, which should be numeric. You can also sort by "key2" and "key3". --reverse Reverses the selected sort. For example, to see a report of the shortest overall time: dbiprof --sort total --reverse --match keyN=value Consider only items where the specified key matches the given value. Keys are numbered from 1. For example, let's say you used a DBI::Profile Path of: [ DBIprofile_Statement, DBIprofile_Methodname ] And called dbiprof as in: dbiprof --match key2=execute Your report would only show execute queries, leaving out prepares, fetches, etc. If the value given starts and ends with slashes ("/") then it will be treated as a regular expression. For example, to only include SELECT queries where key1 is the statement: dbiprof --match key1=/^SELECT/ By default the match expression is matched case-insensitively, but this can be changed with the --case-sensitive option. --exclude keyN=value Remove items for where the specified key matches the given value. For example, to exclude all prepare entries where key2 is the method name: dbiprof --exclude key2=prepare Like "--match", If the value given starts and ends with slashes ("/") then it will be treated as a regular expression. For example, to exclude UPDATE queries where key1 is the statement: dbiprof --match key1=/^UPDATE/ By default the exclude expression is matched case-insensitively, but this can be changed with the --case-sensitive option. --case-sensitive Using this option causes --match and --exclude to work case- sensitively. Defaults to off. --delete Sets the "DeleteFiles" option to DBI::ProfileData which causes the files to be deleted after reading. See DBI::ProfileData for more details. --dumpnodes Print the list of nodes in the form of a perl data structure. Use the "-sort" option if you want the list sorted. --version Print the dbiprof version number and exit. AUTHOR Sam Tregar <sam@tregar.com> COPYRIGHT AND LICENSE Copyright (C) 2002 Sam Tregar This program is free software; you can redistribute it and/or modify it under the same terms as Perl 5 itself. SEE ALSO DBI::ProfileDumper, DBI::Profile, DBI. perl v5.30.3 2024-04-13 DBIPROF(1)
| null |
vim
|
Vim is a text editor that is upwards compatible to Vi. It can be used to edit all kinds of plain text. It is especially useful for editing programs. There are a lot of enhancements above Vi: multi level undo, multi windows and buffers, syntax highlighting, command line editing, filename completion, on-line help, visual selection, etc.. See ":help vi_diff.txt" for a summary of the differences between Vim and Vi. While running Vim a lot of help can be obtained from the on-line help system, with the ":help" command. See the ON-LINE HELP section below. Most often Vim is started to edit a single file with the command vim file More generally Vim is started with: vim [options] [filelist] If the filelist is missing, the editor will start with an empty buffer. Otherwise exactly one out of the following four may be used to choose one or more files to be edited. file .. A list of filenames. The first one will be the current file and read into the buffer. The cursor will be positioned on the first line of the buffer. You can get to the other files with the ":next" command. To edit a file that starts with a dash, precede the filelist with "--". - The file to edit is read from stdin. Commands are read from stderr, which should be a tty. -t {tag} The file to edit and the initial cursor position depends on a "tag", a sort of goto label. {tag} is looked up in the tags file, the associated file becomes the current file and the associated command is executed. Mostly this is used for C programs, in which case {tag} could be a function name. The effect is that the file containing that function becomes the current file and the cursor is positioned on the start of the function. See ":help tag-commands". -q [errorfile] Start in quickFix mode. The file [errorfile] is read and the first error is displayed. If [errorfile] is omitted, the filename is obtained from the 'errorfile' option (defaults to "AztecC.Err" for the Amiga, "errors.err" on other systems). Further errors can be jumped to with the ":cn" command. See ":help quickfix". Vim behaves differently, depending on the name of the command (the executable may still be the same file). vim The "normal" way, everything is default. ex Start in Ex mode. Go to Normal mode with the ":vi" command. Can also be done with the "-e" argument. view Start in read-only mode. You will be protected from writing the files. Can also be done with the "-R" argument. gvim gview The GUI version. Starts a new window. Can also be done with the "-g" argument. evim eview The GUI version in easy mode. Starts a new window. Can also be done with the "-y" argument. rvim rview rgvim rgview Like the above, but with restrictions. It will not be possible to start shell commands, or suspend Vim. Can also be done with the "-Z" argument.
|
vim - Vi IMproved, a programmer's text editor
|
vim [options] [file ..] vim [options] - vim [options] -t tag vim [options] -q [errorfile] ex view gvim gview evim eview rvim rview rgvim rgview
|
The options may be given in any order, before or after filenames. Options without an argument can be combined after a single dash. +[num] For the first file the cursor will be positioned on line "num". If "num" is missing, the cursor will be positioned on the last line. +/{pat} For the first file the cursor will be positioned in the line with the first occurrence of {pat}. See ":help search-pattern" for the available search patterns. +{command} -c {command} {command} will be executed after the first file has been read. {command} is interpreted as an Ex command. If the {command} contains spaces it must be enclosed in double quotes (this depends on the shell that is used). Example: vim "+set si" main.c Note: You can use up to 10 "+" or "-c" commands. -S {file} {file} will be sourced after the first file has been read. This is equivalent to -c "source {file}". {file} cannot start with '-'. If {file} is omitted "Session.vim" is used (only works when -S is the last argument). --cmd {command} Like using "-c", but the command is executed just before processing any vimrc file. You can use up to 10 of these commands, independently from "-c" commands. -A If Vim has been compiled with ARABIC support for editing right-to-left oriented files and Arabic keyboard mapping, this option starts Vim in Arabic mode, i.e. 'arabic' is set. Otherwise an error message is given and Vim aborts. -b Binary mode. A few options will be set that makes it possible to edit a binary or executable file. -C Compatible. Set the 'compatible' option. This will make Vim behave mostly like Vi, even though a .vimrc file exists. -d Start in diff mode. There should between two to eight file name arguments. Vim will open all the files and show differences between them. Works like vimdiff(1). -d {device}, -dev {device} Open {device} for use as a terminal. Only on the Amiga. Example: "-d con:20/30/600/150". -D Debugging. Go to debugging mode when executing the first command from a script. -e Start Vim in Ex mode, just like the executable was called "ex". -E Start Vim in improved Ex mode, just like the executable was called "exim". -f Foreground. For the GUI version, Vim will not fork and detach from the shell it was started in. On the Amiga, Vim is not restarted to open a new window. This option should be used when Vim is executed by a program that will wait for the edit session to finish (e.g. mail). On the Amiga the ":sh" and ":!" commands will not work. --nofork Foreground. For the GUI version, Vim will not fork and detach from the shell it was started in. -F If Vim has been compiled with FKMAP support for editing right-to-left oriented files and Farsi keyboard mapping, this option starts Vim in Farsi mode, i.e. 'fkmap' and 'rightleft' are set. Otherwise an error message is given and Vim aborts. -g If Vim has been compiled with GUI support, this option enables the GUI. If no GUI support was compiled in, an error message is given and Vim aborts. --gui-dialog-file {name} When using the GUI, instead of showing a dialog, write the title and message of the dialog to file {name}. The file is created or appended to. Only useful for testing, to avoid that the test gets stuck on a dialog that can't be seen. Without the GUI the argument is ignored. --help, -h, -? Give a bit of help about the command line arguments and options. After this Vim exits. -H If Vim has been compiled with RIGHTLEFT support for editing right-to-left oriented files and Hebrew keyboard mapping, this option starts Vim in Hebrew mode, i.e. 'hkmap' and 'rightleft' are set. Otherwise an error message is given and Vim aborts. -i {viminfo} Specifies the filename to use when reading or writing the viminfo file, instead of the default "~/.viminfo". This can also be used to skip the use of the .viminfo file, by giving the name "NONE". -L Same as -r. -l Lisp mode. Sets the 'lisp' and 'showmatch' options on. -m Modifying files is disabled. Resets the 'write' option. You can still modify the buffer, but writing a file is not possible. -M Modifications not allowed. The 'modifiable' and 'write' options will be unset, so that changes are not allowed and files can not be written. Note that these options can be set to enable making modifications. -N No-compatible mode. Resets the 'compatible' option. This will make Vim behave a bit better, but less Vi compatible, even though a .vimrc file does not exist. -n No swap file will be used. Recovery after a crash will be impossible. Handy if you want to edit a file on a very slow medium (e.g. floppy). Can also be done with ":set uc=0". Can be undone with ":set uc=200". -nb Become an editor server for NetBeans. See the docs for details. -o[N] Open N windows stacked. When N is omitted, open one window for each file. -O[N] Open N windows side by side. When N is omitted, open one window for each file. -p[N] Open N tab pages. When N is omitted, open one tab page for each file. -P {parent-title} Win32 GUI only: Specify the title of the parent application. When possible, Vim will run in an MDI window inside the application. {parent-title} must appear in the window title of the parent application. Make sure that it is specific enough. Note that the implementation is still primitive. It won't work with all applications and the menu doesn't work. -R Read-only mode. The 'readonly' option will be set. You can still edit the buffer, but will be prevented from accidentally overwriting a file. If you do want to overwrite a file, add an exclamation mark to the Ex command, as in ":w!". The -R option also implies the -n option (see above). The 'readonly' option can be reset with ":set noro". See ":help 'readonly'". -r List swap files, with information about using them for recovery. -r {file} Recovery mode. The swap file is used to recover a crashed editing session. The swap file is a file with the same filename as the text file with ".swp" appended. See ":help recovery". -s Silent mode. Only when started as "Ex" or when the "-e" option was given before the "-s" option. -s {scriptin} The script file {scriptin} is read. The characters in the file are interpreted as if you had typed them. The same can be done with the command ":source! {scriptin}". If the end of the file is reached before the editor exits, further characters are read from the keyboard. -T {terminal} Tells Vim the name of the terminal you are using. Only required when the automatic way doesn't work. Should be a terminal known to Vim (builtin) or defined in the termcap or terminfo file. --not-a-term Tells Vim that the user knows that the input and/or output is not connected to a terminal. This will avoid the warning and the two second delay that would happen. --ttyfail When stdin or stdout is not a a terminal (tty) then exit right away. -u {vimrc} Use the commands in the file {vimrc} for initializations. All the other initializations are skipped. Use this to edit a special kind of files. It can also be used to skip all initializations by giving the name "NONE". See ":help initialization" within vim for more details. -U {gvimrc} Use the commands in the file {gvimrc} for GUI initializations. All the other GUI initializations are skipped. It can also be used to skip all GUI initializations by giving the name "NONE". See ":help gui-init" within vim for more details. -V[N] Verbose. Give messages about which files are sourced and for reading and writing a viminfo file. The optional number N is the value for 'verbose'. Default is 10. -V[N]{filename} Like -V and set 'verbosefile' to {filename}. The result is that messages are not displayed but written to the file {filename}. {filename} must not start with a digit. --log {filename} If Vim has been compiled with eval and channel feature, start logging and write entries to {filename}. This works like calling ch_logfile({filename}, 'ao') very early during startup. -v Start Vim in Vi mode, just like the executable was called "vi". This only has effect when the executable is called "ex". -w{number} Set the 'window' option to {number}. -w {scriptout} All the characters that you type are recorded in the file {scriptout}, until you exit Vim. This is useful if you want to create a script file to be used with "vim -s" or ":source!". If the {scriptout} file exists, characters are appended. -W {scriptout} Like -w, but an existing file is overwritten. -x Use encryption when writing files. Will prompt for a crypt key. -X Don't connect to the X server. Shortens startup time in a terminal, but the window title and clipboard will not be used. -y Start Vim in easy mode, just like the executable was called "evim" or "eview". Makes Vim behave like a click-and-type editor. -Z Restricted mode. Works like the executable starts with "r". -- Denotes the end of the options. Arguments after this will be handled as a file name. This can be used to edit a filename that starts with a '-'. --clean Do not use any personal configuration (vimrc, plugins, etc.). Useful to see if a problem reproduces with a clean Vim setup. --echo-wid GTK GUI only: Echo the Window ID on stdout. --literal Take file name arguments literally, do not expand wildcards. This has no effect on Unix where the shell expands wildcards. --noplugin Skip loading plugins. Implied by -u NONE. --remote Connect to a Vim server and make it edit the files given in the rest of the arguments. If no server is found a warning is given and the files are edited in the current Vim. --remote-expr {expr} Connect to a Vim server, evaluate {expr} in it and print the result on stdout. --remote-send {keys} Connect to a Vim server and send {keys} to it. --remote-silent As --remote, but without the warning when no server is found. --remote-wait As --remote, but Vim does not exit until the files have been edited. --remote-wait-silent As --remote-wait, but without the warning when no server is found. --serverlist List the names of all Vim servers that can be found. --servername {name} Use {name} as the server name. Used for the current Vim, unless used with a --remote argument, then it's the name of the server to connect to. --socketid {id} GTK GUI only: Use the GtkPlug mechanism to run gvim in another window. --startuptime {file} During startup write timing messages to the file {fname}. --version Print version information and exit. --windowid {id} Win32 GUI only: Make gvim try to use the window {id} as a parent, so that it runs inside that window. ON-LINE HELP Type ":help" in Vim to get started. Type ":help subject" to get help on a specific subject. For example: ":help ZZ" to get help for the "ZZ" command. Use <Tab> and CTRL-D to complete subjects (":help cmdline-completion"). Tags are present to jump from one place to another (sort of hypertext links, see ":help"). All documentation files can be viewed in this way, for example ":help syntax.txt". FILES /usr/local/share/vim/vim??/doc/*.txt The Vim documentation files. Use ":help doc-file-list" to get the complete list. vim?? is short version number, like vim91 for Vim 9.1 /usr/local/share/vim/vim??/doc/tags The tags file used for finding information in the documentation files. /usr/local/share/vim/vim??/syntax/syntax.vim System wide syntax initializations. /usr/local/share/vim/vim??/syntax/*.vim Syntax files for various languages. /usr/local/share/vim/vimrc System wide Vim initializations. ~/.vimrc, ~/.vim/vimrc, $XDG_CONFIG_HOME/vim/vimrc Your personal Vim initializations (first one found is used). /usr/local/share/vim/gvimrc System wide gvim initializations. ~/.gvimrc, ~/.vim/gvimrc, $XDG_CONFIG_HOME/vim/gvimrc Your personal gvim initializations (first one found is used). /usr/local/share/vim/vim??/optwin.vim Script used for the ":options" command, a nice way to view and set options. /usr/local/share/vim/vim??/menu.vim System wide menu initializations for gvim. /usr/local/share/vim/vim??/bugreport.vim Script to generate a bug report. See ":help bugs". /usr/local/share/vim/vim??/filetype.vim Script to detect the type of a file by its name. See ":help 'filetype'". /usr/local/share/vim/vim??/scripts.vim Script to detect the type of a file by its contents. See ":help 'filetype'". /usr/local/share/vim/vim??/print/*.ps Files used for PostScript printing. For recent info read the VIM home page: <URL:http://www.vim.org/> SEE ALSO vimtutor(1) AUTHOR Most of Vim was made by Bram Moolenaar, with a lot of help from others. See ":help credits" in Vim. Vim is based on Stevie, worked on by: Tim Thompson, Tony Andrews and G.R. (Fred) Walter. Although hardly any of the original code remains. BUGS Probably. See ":help todo" for a list of known problems. Note that a number of things that may be regarded as bugs by some, are in fact caused by a too-faithful reproduction of Vi's behaviour. And if you think other things are bugs "because Vi does it differently", you should take a closer look at the vi_diff.txt file (or type :help vi_diff.txt when in Vim). Also have a look at the 'compatible' and 'cpoptions' options. 2024 Jun 04 VIM(1)
| null |
mandoc_soelim
|
soelim reads files lines by lines. If a line starts by: “.so anotherfile” it replace the line by processing “anotherfile”. Otherwise the line is printed to stdout. -C Recognise .so when not followed by a space character. -r Compatibility with GNU groff's soelim (does nothing). -t Compatibility with GNU groff's soelim (does nothing). -v Compatibility with GNU groff's soelim (does nothing). -I dir This option specify directories where soelim searches for files (both those on the command line and those named in “.so” directive.) This options may be specified multiple times. The directories will be searched in the order specified. The files are always searched first in the current directory. A file specified with an absolute path will be opened directly without performing a search. SEE ALSO mandoc(1) AUTHORS This version of the soelim utility was written by Baptiste Daroussin <bapt@freebsd.org>. macOS 14.5 July 4, 2017 macOS 14.5
|
soelim – interpret .so requests in manpages
|
soelim [-Crtv] [-I dir] [files ...]
| null | null |
xjc
| null | null | null | null | null |
bzmore
|
Bzmore is a filter which allows examination of compressed or plain text files one screenful at a time on a soft-copy terminal. bzmore works on files compressed with bzip2 and also on uncompressed files. If a file does not exist, bzmore looks for a file of the same name with the addition of a .bz2 suffix. Bzmore normally pauses after each screenful, printing --More-- at the bottom of the screen. If the user then types a carriage return, one more line is displayed. If the user hits a space, another screenful is displayed. Other possibilities are enumerated later. Bzmore looks in the file /etc/termcap to determine terminal characteristics, and to determine the default window size. On a terminal capable of displaying 24 lines, the default window size is 22 lines. Other sequences which may be typed when bzmore pauses, and their effects, are as follows (i is an optional integer argument, defaulting to 1) : i<space> display i more lines, (or another screenful if no argument is given) ^D display 11 more lines (a ``scroll''). If i is given, then the scroll size is set to i. d same as ^D (control-D) iz same as typing a space except that i, if present, becomes the new window size. Note that the window size reverts back to the default at the end of the current file. is skip i lines and print a screenful of lines if skip i screenfuls and print a screenful of lines q or Q quit reading the current file; go on to the next (if any) e or q When the prompt --More--(Next file: file) is printed, this command causes bzmore to exit. s When the prompt --More--(Next file: file) is printed, this command causes bzmore to skip the next file and continue. = Display the current line number. i/expr search for the i-th occurrence of the regular expression expr. If the pattern is not found, bzmore goes on to the next file (if any). Otherwise, a screenful is displayed, starting two lines before the place where the expression was found. The user's erase and kill characters may be used to edit the regular expression. Erasing back past the first column cancels the search command. in search for the i-th occurrence of the last regular expression entered. !command invoke a shell with command. The character `!' in "command" are replaced with the previous shell command. The sequence "\!" is replaced by "!". :q or :Q quit reading the current file; go on to the next (if any) (same as q or Q). . (dot) repeat the previous command. The commands take effect immediately, i.e., it is not necessary to type a carriage return. Up to the time when the command character itself is given, the user may hit the line kill character to cancel the numerical argument being formed. In addition, the user may hit the erase character to redisplay the --More-- message. At any time when output is being sent to the terminal, the user can hit the quit key (normally control-\). Bzmore will stop sending output, and will display the usual --More-- prompt. The user may then enter one of the above commands in the normal manner. Unfortunately, some output is lost when this is done, due to the fact that any characters waiting in the terminal's output queue are flushed when the quit signal occurs. The terminal is set to noecho mode by this program so that the output can be continuous. What you type will thus not show on your terminal, except for the / and ! commands. If the standard output is not a teletype, then bzmore acts just like bzcat, except that a header is printed before each file. FILES /etc/termcap Terminal data base SEE ALSO more(1), less(1), bzip2(1), bzdiff(1), bzgrep(1) BZMORE(1)
|
bzmore, bzless - file perusal filter for crt viewing of bzip2 compressed text
|
bzmore [ name ... ] bzless [ name ... ] NOTE In the following description, bzless and less can be used interchangeably with bzmore and more.
| null | null |
dyld_info
|
The dyld_info tool will display information that dyld uses from binaries. It supports fixups information encoded as opcode based or as chained fixups. If the tool is run with no options, it defaults to the common options (-platform -segments -dependents). The tool can be run with paths to dylibs that are in the dyld cache and not on disk. The options are as follows: -arch arch Only display the specified architecture. Other architectures in a universal image are ignored. -platform Display the platform (e.g. macOS, iOS) that the binary was built for. -segments Display all segments and sections with size info. -dependents Display all dylibs that the binary is linked against. -inits Display all static initializers and Objective-C +load methods. -exports Display all exported symbols. -imports Display all imported symbols. -objc Display the Objective-C classes and categories defined. -fixups Display a simple table of fixup locations and what dyld will set them to. -fixup_chains If the binary uses chained fixups, displays the format of the chains and the starts. -fixup_chain_details Display a table of all the raw fixup chain bits and their meanings. -symbolic_fixups Display the offset of all functions starts and their name. -symbolic_fixups Display all fixups, grouped by data symbol. -opcodes Display all the opcodes used to generate the rebase and binding information. -validate_only Display nothing if binary is valid, or information about what parts are malformed. SEE ALSO otool(1) nm(1) Darwin April 15, 2021 Darwin
|
dyld_info – Displays information used by dyld in programs and dylibs
|
dyld_info [-arch arch-name] [-platform] [-segments] [-dependents] [-inits] [-exports] [-imports] [-objc] [-fixups] [-fixup_chains] [-fixup_chain_details] [-function_starts] [-symbolic_fixups] [-opcodes] [-validate_only] file(s)
| null | null |
package-stash-conflicts
| null |
package-stash-conflicts VERSION version 0.38 SUPPORT Bugs may be submitted through the RT bug tracker <https://rt.cpan.org/Public/Dist/Display.html?Name=Package-Stash> (or bug-Package-Stash@rt.cpan.org <mailto:bug-Package-Stash@rt.cpan.org>). AUTHOR Jesse Luehrs <doy@tozt.net> COPYRIGHT AND LICENSE This software is copyright (c) 2018 by Jesse Luehrs. This is free software; you can redistribute it and/or modify it under the same terms as the Perl 5 programming language system itself. perl v5.34.0 2018-12-31 PACKAGE-STASH-CONFLICTS(1)
| null | null | null |
cancel
|
The cancel command cancels print jobs. If no destination or id is specified, the currently printing job on the default destination is canceled.
|
cancel - cancel jobs
|
cancel [ -E ] [ -U username ] [ -a ] [ -h hostname[:port] ] [ -u username ] [ -x ] [ id ] [ destination ] [ destination-id ]
|
The following options are recognized by cancel: -a Cancel all jobs on the named destination, or all jobs on all destinations if none is provided. -E Forces encryption when connecting to the server. -h hostname[:port] Specifies an alternate server. -U username Specifies the username to use when connecting to the server. -u username Cancels jobs owned by username. -x Deletes job data files in addition to canceling. CONFORMING TO Unlike the System V printing system, CUPS allows printer names to contain any printable character except SPACE, TAB, "/", or "#". Also, printer and class names are not case-sensitive.
|
Cancel the current print job: cancel Cancel job "myprinter-42": cancel myprinter-42 Cancel all jobs: cancel -a NOTES Administrators wishing to prevent unauthorized cancellation of jobs via the -u option should require authentication for Cancel-Jobs operations in cupsd.conf(5). SEE ALSO cupsd.conf(5), lp(1), lpmove(8), lpstat(1), CUPS Online Help (http://localhost:631/help) COPYRIGHT Copyright © 2007-2019 by Apple Inc. 26 April 2019 CUPS cancel(1)
|
apply
|
The apply utility runs the named command on each argument argument in turn. Character sequences of the form “%d” in command, where ‘d’ is a digit from 1 to 9, are replaced by the d´th following unused argument. In this case, the largest digit number of arguments are discarded for each execution of command. The options are as follows: -# Normally arguments are taken singly; the optional number # specifies the number of arguments to be passed to command. If the number is zero, command is run, without arguments, once for each argument. If any sequences of “%d” occur in command, the -# option is ignored. -a c The use of the character ‘%’ as a magic character may be changed with the -a option. -d Display the commands that would have been executed, but do not actually execute them. ENVIRONMENT The following environment variable affects the execution of apply: SHELL Pathname of shell to use. If this variable is not defined, the Bourne shell is used. FILES /bin/sh default shell
|
apply – apply a command to a set of arguments
|
apply [-a c] [-d] [-#] command argument ...
| null |
apply echo * is similar to ls(1); apply -2 cmp a1 b1 a2 b2 a3 b3 compares the `a' files to the `b' files; apply -0 who 1 2 3 4 5 runs who(1) 5 times; and apply ´ln %1 /usr/joe´ * links all files in the current directory to the directory /usr/joe. HISTORY The apply command appeared in 4.2BSD. AUTHORS Rob Pike BUGS Shell metacharacters in command may have bizarre effects; it is best to enclose complicated commands in single quotes (''). The apply utility does not recognize multibyte characters. macOS 14.5 January 28, 2022 macOS 14.5
|
bzgrep
|
Bzgrep is used to invoke the grep on bzip2-compressed files. All options specified are passed directly to grep. If no file is specified, then the standard input is decompressed if necessary and fed to grep. Otherwise the given files are uncompressed if necessary and fed to grep. If bzgrep is invoked as bzegrep or bzfgrep then egrep or fgrep is used instead of grep. If the GREP environment variable is set, bzgrep uses it as the grep program to be invoked. For example: for sh: GREP=fgrep bzgrep string files for csh: (setenv GREP fgrep; bzgrep string files) AUTHOR Charles Levert (charles@comm.polymtl.ca). Adapted to bzip2 by Philippe Troin <phil@fifi.org> for Debian GNU/Linux. SEE ALSO grep(1), egrep(1), fgrep(1), bzdiff(1), bzmore(1), bzless(1), bzip2(1) BZGREP(1)
|
bzgrep, bzfgrep, bzegrep - search possibly bzip2 compressed files for a regular expression
|
bzgrep [ grep_options ] [ -e ] pattern filename... bzegrep [ egrep_options ] [ -e ] pattern filename... bzfgrep [ fgrep_options ] [ -e ] pattern filename...
| null | null |
umask
|
Shell builtin commands are commands that can be executed within the running shell's process. Note that, in the case of csh(1) builtin commands, the command is executed in a subshell if it occurs as any component of a pipeline except the last. If a command specified to the shell contains a slash ‘/’, the shell will not execute a builtin command, even if the last component of the specified command matches the name of a builtin command. Thus, while specifying “echo” causes a builtin command to be executed under shells that support the echo builtin command, specifying “/bin/echo” or “./echo” does not. While some builtin commands may exist in more than one shell, their operation may be different under each shell which supports them. Below is a table which lists shell builtin commands, the standard shells that support them and whether they exist as standalone utilities. Only builtin commands for the csh(1) and sh(1) shells are listed here. Consult a shell's manual page for details on the operation of its builtin commands. Beware that the sh(1) manual page, at least, calls some of these commands “built-in commands” and some of them “reserved words”. Users of other shells may need to consult an info(1) page or other sources of documentation. Commands marked “No**” under External do exist externally, but are implemented as scripts using a builtin command of the same name. Command External csh(1) sh(1) ! No No Yes % No Yes No . No No Yes : No Yes Yes @ No Yes Yes [ Yes No Yes { No No Yes } No No Yes alias No** Yes Yes alloc No Yes No bg No** Yes Yes bind No No Yes bindkey No Yes No break No Yes Yes breaksw No Yes No builtin No No Yes builtins No Yes No case No Yes Yes cd No** Yes Yes chdir No Yes Yes command No** No Yes complete No Yes No continue No Yes Yes default No Yes No dirs No Yes No do No No Yes done No No Yes echo Yes Yes Yes echotc No Yes No elif No No Yes else No Yes Yes end No Yes No endif No Yes No endsw No Yes No esac No No Yes eval No Yes Yes exec No Yes Yes exit No Yes Yes export No No Yes false Yes No Yes fc No** No Yes fg No** Yes Yes filetest No Yes No fi No No Yes for No No Yes foreach No Yes No getopts No** No Yes glob No Yes No goto No Yes No hash No** No Yes hashstat No Yes No history No Yes No hup No Yes No if No Yes Yes jobid No No Yes jobs No** Yes Yes kill Yes Yes Yes limit No Yes No local No No Yes log No Yes No login Yes Yes No logout No Yes No ls-F No Yes No nice Yes Yes No nohup Yes Yes No notify No Yes No onintr No Yes No popd No Yes No printenv Yes Yes No printf Yes No Yes pushd No Yes No pwd Yes No Yes read No** No Yes readonly No No Yes rehash No Yes No repeat No Yes No return No No Yes sched No Yes No set No Yes Yes setenv No Yes No settc No Yes No setty No Yes No setvar No No Yes shift No Yes Yes source No Yes No stop No Yes No suspend No Yes No switch No Yes No telltc No Yes No test Yes No Yes then No No Yes time Yes Yes No times No No Yes trap No No Yes true Yes No Yes type No** No Yes ulimit No** No Yes umask No** Yes Yes unalias No** Yes Yes uncomplete No Yes No unhash No Yes No unlimit No Yes No unset No Yes Yes unsetenv No Yes No until No No Yes wait No** Yes Yes where No Yes No which Yes Yes No while No Yes Yes SEE ALSO csh(1), dash(1), echo(1), false(1), info(1), kill(1), login(1), nice(1), nohup(1), printenv(1), printf(1), pwd(1), sh(1), test(1), time(1), true(1), which(1), zsh(1) HISTORY The builtin manual page first appeared in FreeBSD 3.4. AUTHORS This manual page was written by Sheldon Hearn <sheldonh@FreeBSD.org>. macOS 14.5 December 21, 2010 macOS 14.5
|
builtin, !, %, ., :, @, [, {, }, alias, alloc, bg, bind, bindkey, break, breaksw, builtins, case, cd, chdir, command, complete, continue, default, dirs, do, done, echo, echotc, elif, else, end, endif, endsw, esac, eval, exec, exit, export, false, fc, fg, filetest, fi, for, foreach, getopts, glob, goto, hash, hashstat, history, hup, if, jobid, jobs, kill, limit, local, log, login, logout, ls-F, nice, nohup, notify, onintr, popd, printenv, printf, pushd, pwd, read, readonly, rehash, repeat, return, sched, set, setenv, settc, setty, setvar, shift, source, stop, suspend, switch, telltc, test, then, time, times, trap, true, type, ulimit, umask, unalias, uncomplete, unhash, unlimit, unset, unsetenv, until, wait, where, which, while – shell built-in commands
|
See the built-in command description in the appropriate shell manual page.
| null | null |
g++
| null | null | null | null | null |
libtool
|
The libtool command takes the specified input object files and creates a library for use with the link editor, ld(1). The library's name is specified by output (the argument to the -o flag). The input object files may be in any correct format that contains object files (``universal'' files, archives, object files). Libtool will not put any non-object input file into the output library (unlike ranlib, which allows this in the archives it operates on). When producing a ``universal'' file from objects of the same CPU type and differing CPU subtypes, libtool and ranlib create at most one library for each CPU type, rather than a separate library in a universal file for each of the unique pairings of CPU type and CPU subtype. Thus, the resulting CPU subtype for each library is the _ALL CPU subtype for that CPU type. This strategy strongly encourages the implementor of a library to create one library that chooses optimum code to run at run time, rather than at link time. Libtool can create either dynamically linked shared libraries, with -dynamic, or statically linked (archive) libraries, with -static. DYNAMICALLY LINKED SHARED LIBRARIES Dynamically linked libraries, unlike statically linked libraries, are Mach-O format files and not ar(5) format files. Dynamically linked libraries have two restrictions: No symbol may be defined in more than one object file and no common symbol can be used. To maximize sharing of a dynamically linked shared library the objects should be compiled with the -dynamic flag of cc(1) to produce indirect undefined references and position-independent code. To build a dynamically linked library, libtool, runs the link editor, ld(1), with -dylib once for each architecture present in the input objects and then lipo(1) to create a universal file if needed. ARCHIVE (or statically linked) LIBRARIES Libtool with -static is intended to replace ar(5) and ranlib. For backward compatibility, ranlib is still available, and it supports universal files. Ranlib adds or updates the table of contents to each archive so it can be linked by the link editor, ld(1). The table of contents is an archive member at the beginning of the archive that indicates which symbols are defined in which library members. Because ranlib rewrites the archive, sufficient temporary file space must be available in the file system that contains the current directory. Ranlib takes all correct forms of libraries (universal files containing archives, and simple archives) and updates the table of contents for all archives in the file. Ranlib also takes one common incorrect form of archive, an archive whose members are universal object files, adding or updating the table of contents and producing the library in correct form (a universal file containing multiple archives). The archive member name for a table of contents begins with ``__.SYMDEF''. Currently, there are two types of table of contents produced by libtool -static and ranlib and understood by the link editor, ld(1). These are explained below, under the -s and -a options.
|
libtool - create libraries ranlib - add or update the table of contents of archive libraries
|
libtool -static -o output [ -sacLTD ] [ - ] [ -arch_only arch_type ] [ -no_warning_for_no_symbols ] file... [-filelist listfile[,dirname]] libtool -dynamic -o output [ -install_name name ] [ -compatibility_version number ] [ -current_version number ] [ link editor flags ] [ -v ] [ -noall_load ] [ - ] [ -arch_only arch_type ] [ -V ] file... [-filelist listfile[,dirname]] ranlib [ -sactfqLT ] [ - ] archive...
|
The following options pertain to libtool only. @file Arguments beginning with @ are replaced by arguments read from the specified file, as an alternative to listing those arguments on the command line. The files simply contain libtool options and files separated by whitespace: spaces, tabs, and newlines. Characters can be escaped with a backslash (\), including whitespace characters and other backslashes. Also, arguments that include whitespace can be enclosed, wholly or in part, by single- or double-quote charcters. These files may contain @file references to additional files, although libtool will error on include cycles. If a file cannot be found, the original @file argument will remain in the argument list. -static Produce a statically linked (archive) library from the input files. This is the default. -dynamic Produce a dynamically linked shared library from the input files. -install_name name For a dynamic shared library, this specifies the file name the library will be installed in for programs that use it. If this is not specified the name specified by the -o output option will be used. -compatibility_version number For a dynamic shared library, this specifies the compatibility version number of the library. When a library is used the compatibility version is checked and if the user's version is greater that the library's version, an error message is printed and the using program exits. The format of number is X[.Y[.Z]] where X must be a positive non-zero number less than or equal to 65535, and .Y and .Z are optional and if present must be non- negative numbers less than or equal to 255. If this is not specified then it has a value of 0 and no checking is done when the library is used. -current_version number For dynamic shared library files this specifies the current version number of the library. The program using the library can obtain the current version of the library programmatically to determine exactly which version of the library it is using. The format of number is X[.Y[.Z]] where X must be a positive non-zero number less than or equal to 65535, and .Y and .Z are optional and if present must be non-negative numbers less than or equal to 255. If this is not specified then it has a value of 0. -noall_load For dynamic shared library files this specifies the the default behavior of loading all members of archives on the command line is not to be done. This option is used by the GNU compiler driver, cc(1), when used with it's -dynamiclib option. This is done to allow selective loading of the GNU's compiler's runtime support library, libcc_dynamic.a . link editor flags For a dynamic shared library the following ld(1) flags are accepted and passed through: -lx, -weak-lx, -search_paths_first -weak_library, -Ldir, -ysym, -usym, -initsym, -idefinition:indirect, -seg1addr, -segs_read_only_addr, -segs_read_write_addr, -seg_addr_table, -seg_addr_table_filename, -segprot, -segalign, -sectcreate, -sectorder, -sectorder_detail, -sectalign, -undefined, -read_only_relocs, -prebind, -prebind_all_twolevel_modules, -prebind_allow_overlap, -noprebind, -framework, -weak_framework, -umbrella, -allowable_client, -sub_umbrella, -sub_library, -F, -U, -Y, -Sn, -Si, -Sp, -S, -X, -x, -whyload, -all_load. -arch_errors_fatal, -dylib_file, -run_init_lazily, -final_output, -macosx_version_min, -multiply_defined, -multiply_defined_unused, -twolevel_namespace, -twolevel_namespace_hints, -flat_namespace, -nomultidefs, -headerpad, -headerpad_max_install_names, -weak_reference_mismatches, -M, -t, -no_arch_warnings, -single_module, -multi_module, -exported_symbols_list, -unexported_symbols_list, -m, -dead_strip, -no_dead_strip_inits_and_terms, -executable_path, -syslibroot, -no_uuid. See the ld(1) man page for details on these flags. The flag -image_base is a synonym for -seg1addr. -v Verbose mode, which prints the ld(1) commands and lipo(1) commands executed. -V Print the version of libtool. -filelist listfile[,dirname] The listfile contains a list of file names and is an alternative way of specifiying file names on the command line. The file names are listed one per line separated only by newlines (spaces and tabs are assumed to be part of the file name). If the optional directory name, dirname is specified then it is prepended to each name in the list file. -arch_only arch_type This option causes libtool to build a library only for the specified arch_type and ignores all other architectures in the input files. When building a dynamic library, if this is specified with a specific cpusubtype other than the family cpusubtype then libtool it does not use the ld(1) -force_cpusubtype_ALL flag and passes the -arch_only argument to ld(1) as the -arch flag so that the output is tagged with that cpusubtype. The following options pertain to the table of contents for an archive library, and apply to both libtool -static and ranlib: -s Produce the preferred type of table of contents, which results in faster link editing when linking with the archive. The order of the table of contents is sorted by symbol name. The library member name of this type of table of contents is ``__.SYMDEF SORTED''. This type of table of contents can only be produced when the library does not have multiple members that define the same symbol. This is the default. -a Produce the original type of table of contents, whose order is based on the order of the members in the archive. The library member name of this type of table of contents is ``__.SYMDEF''. This type of table of contents must be used when the library has multiple members that define the same symbol. -c Include common symbols as definitions with respect to the table of contents. This is seldom the intended behavior for linking from a library, as it forces the linking of a library member just because it uses an uninitialized global that is undefined at that point in the linking. This option is included only because this was the original behavior of ranlib. This option is not the default. -L Use the 4.4bsd archive extended format #1, which allows archive member names to be longer than 16 characters and have spaces in their names. This option is the default. -T Truncate archive member names to 16 characters and don't use the 4.4bsd extended format #1. This option is not the default. -f Warns when the output archive is universal and ar(1) will no longer be able to operate on it. -q Do nothing if a universal file would be created. -D When building a static library, set archive contents' user ids, group ids, dates, and file modes to reasonable defaults. This allows libraries created with identical input to be identical to each other, regardless of time of day, user, group, umask, and other aspects of the environment. For compatibility, the following ranlib option is accepted (but ignored): -t This option used to request that ranlib only ``touch'' the archives instead of modifying them. The option is now ignored, and the table of contents is rebuilt. Other options applying to both libtool and ranlib: - Treat all remaining arguments as names of files (or archives) and not as options. -no_warning_for_no_symbols Don't warn about file that have no symbols. -dependency_info path Write an Xcode dependency info file describing a successful build operation. This file describes the inputs directly or indirectly used to create the library or dylib. SEE ALSO ld(1), ar(1), otool(1), make(1), redo_prebinding(1), ar(5) BUGS With the way libraries used to be created, errors were possible if the library was modified with ar(1) and the table of contents was not updated by rerunning ranlib(1). So previously the link editor, ld(1), generated an error when the modification date of a library was more recent than the creation date of its table of contents. Unfortunately, this meant that you got the error even if you only copy the library. Since this error was found to be too much of a nuisance it was removed. So now it is possible again to get link errors if the library is modified and the table of contents is not updated. Apple Inc. June 23, 2020 LIBTOOL(1)
| null |
findrule5.30
|
"findrule" mostly borrows the interface from GNU find(1) to provide a command-line interface onto the File::Find::Rule heirarchy of modules. The syntax for expressions is the rule name, preceded by a dash, followed by an optional argument. If the argument is an opening parenthesis it is taken as a list of arguments, terminated by a closing parenthesis. Some examples: find -file -name ( foo bar ) files named "foo" or "bar", below the current directory. find -file -name foo -bar files named "foo", that have pubs (for this is what our ficticious "bar" clause specifies), below the current directory. find -file -name ( -bar ) files named "-bar", below the current directory. In this case if we'd have omitted the parenthesis it would have parsed as a call to name with no arguments, followed by a call to -bar. Supported switches I'm very slack. Please consult the File::Find::Rule manpage for now, and prepend - to the commands that you want. Extra bonus switches findrule automatically loads all of your installed File::Find::Rule::* extension modules, so check the documentation to see what those would be. AUTHOR Richard Clamp <richardc@unixbeard.net> from a suggestion by Tatsuhiko Miyagawa COPYRIGHT Copyright (C) 2002 Richard Clamp. All Rights Reserved. This program is free software; you can redistribute it and/or modify it under the same terms as Perl itself. SEE ALSO File::Find::Rule perl v5.30.3 2015-12-03 FINDRULE(1)
|
findrule - command line wrapper to File::Find::Rule USAGE findrule [path...] [expression]
| null | null | null |
ptar5.30
|
ptar is a small, tar look-alike program that uses the perl module Archive::Tar to extract, create and list tar archives.
|
ptar - a tar-like program written in perl
|
ptar -c [-v] [-z] [-C] [-f ARCHIVE_FILE | -] FILE FILE ... ptar -c [-v] [-z] [-C] [-T index | -] [-f ARCHIVE_FILE | -] ptar -x [-v] [-z] [-f ARCHIVE_FILE | -] ptar -t [-z] [-f ARCHIVE_FILE | -] ptar -h
|
c Create ARCHIVE_FILE or STDOUT (-) from FILE x Extract from ARCHIVE_FILE or STDIN (-) t List the contents of ARCHIVE_FILE or STDIN (-) f Name of the ARCHIVE_FILE to use. Default is './default.tar' z Read/Write zlib compressed ARCHIVE_FILE (not always available) v Print filenames as they are added or extracted from ARCHIVE_FILE h Prints this help message C CPAN mode - drop 022 from permissions T get names to create from file SEE ALSO tar(1), Archive::Tar. perl v5.30.3 2024-04-13 PTAR(1)
| null |
unalias
|
Shell builtin commands are commands that can be executed within the running shell's process. Note that, in the case of csh(1) builtin commands, the command is executed in a subshell if it occurs as any component of a pipeline except the last. If a command specified to the shell contains a slash ‘/’, the shell will not execute a builtin command, even if the last component of the specified command matches the name of a builtin command. Thus, while specifying “echo” causes a builtin command to be executed under shells that support the echo builtin command, specifying “/bin/echo” or “./echo” does not. While some builtin commands may exist in more than one shell, their operation may be different under each shell which supports them. Below is a table which lists shell builtin commands, the standard shells that support them and whether they exist as standalone utilities. Only builtin commands for the csh(1) and sh(1) shells are listed here. Consult a shell's manual page for details on the operation of its builtin commands. Beware that the sh(1) manual page, at least, calls some of these commands “built-in commands” and some of them “reserved words”. Users of other shells may need to consult an info(1) page or other sources of documentation. Commands marked “No**” under External do exist externally, but are implemented as scripts using a builtin command of the same name. Command External csh(1) sh(1) ! No No Yes % No Yes No . No No Yes : No Yes Yes @ No Yes Yes [ Yes No Yes { No No Yes } No No Yes alias No** Yes Yes alloc No Yes No bg No** Yes Yes bind No No Yes bindkey No Yes No break No Yes Yes breaksw No Yes No builtin No No Yes builtins No Yes No case No Yes Yes cd No** Yes Yes chdir No Yes Yes command No** No Yes complete No Yes No continue No Yes Yes default No Yes No dirs No Yes No do No No Yes done No No Yes echo Yes Yes Yes echotc No Yes No elif No No Yes else No Yes Yes end No Yes No endif No Yes No endsw No Yes No esac No No Yes eval No Yes Yes exec No Yes Yes exit No Yes Yes export No No Yes false Yes No Yes fc No** No Yes fg No** Yes Yes filetest No Yes No fi No No Yes for No No Yes foreach No Yes No getopts No** No Yes glob No Yes No goto No Yes No hash No** No Yes hashstat No Yes No history No Yes No hup No Yes No if No Yes Yes jobid No No Yes jobs No** Yes Yes kill Yes Yes Yes limit No Yes No local No No Yes log No Yes No login Yes Yes No logout No Yes No ls-F No Yes No nice Yes Yes No nohup Yes Yes No notify No Yes No onintr No Yes No popd No Yes No printenv Yes Yes No printf Yes No Yes pushd No Yes No pwd Yes No Yes read No** No Yes readonly No No Yes rehash No Yes No repeat No Yes No return No No Yes sched No Yes No set No Yes Yes setenv No Yes No settc No Yes No setty No Yes No setvar No No Yes shift No Yes Yes source No Yes No stop No Yes No suspend No Yes No switch No Yes No telltc No Yes No test Yes No Yes then No No Yes time Yes Yes No times No No Yes trap No No Yes true Yes No Yes type No** No Yes ulimit No** No Yes umask No** Yes Yes unalias No** Yes Yes uncomplete No Yes No unhash No Yes No unlimit No Yes No unset No Yes Yes unsetenv No Yes No until No No Yes wait No** Yes Yes where No Yes No which Yes Yes No while No Yes Yes SEE ALSO csh(1), dash(1), echo(1), false(1), info(1), kill(1), login(1), nice(1), nohup(1), printenv(1), printf(1), pwd(1), sh(1), test(1), time(1), true(1), which(1), zsh(1) HISTORY The builtin manual page first appeared in FreeBSD 3.4. AUTHORS This manual page was written by Sheldon Hearn <sheldonh@FreeBSD.org>. macOS 14.5 December 21, 2010 macOS 14.5
|
builtin, !, %, ., :, @, [, {, }, alias, alloc, bg, bind, bindkey, break, breaksw, builtins, case, cd, chdir, command, complete, continue, default, dirs, do, done, echo, echotc, elif, else, end, endif, endsw, esac, eval, exec, exit, export, false, fc, fg, filetest, fi, for, foreach, getopts, glob, goto, hash, hashstat, history, hup, if, jobid, jobs, kill, limit, local, log, login, logout, ls-F, nice, nohup, notify, onintr, popd, printenv, printf, pushd, pwd, read, readonly, rehash, repeat, return, sched, set, setenv, settc, setty, setvar, shift, source, stop, suspend, switch, telltc, test, then, time, times, trap, true, type, ulimit, umask, unalias, uncomplete, unhash, unlimit, unset, unsetenv, until, wait, where, which, while – shell built-in commands
|
See the built-in command description in the appropriate shell manual page.
| null | null |
h2xs5.30
|
h2xs builds a Perl extension from C header files. The extension will include functions which can be used to retrieve the value of any #define statement which was in the C header files. The module_name will be used for the name of the extension. If module_name is not supplied then the name of the first header file will be used, with the first character capitalized. If the extension might need extra libraries, they should be included here. The extension Makefile.PL will take care of checking whether the libraries actually exist and how they should be loaded. The extra libraries should be specified in the form -lm -lposix, etc, just as on the cc command line. By default, the Makefile.PL will search through the library path determined by Configure. That path can be augmented by including arguments of the form -L/another/library/path in the extra-libraries argument. In spite of its name, h2xs may also be used to create a skeleton pure Perl module. See the -X option.
|
h2xs - convert .h C header files to Perl extensions
|
h2xs [OPTIONS ...] [headerfile ... [extra_libraries]] h2xs -h|-?|--help
|
-A, --omit-autoload Omit all autoload facilities. This is the same as -c but also removes the "use AutoLoader" statement from the .pm file. -B, --beta-version Use an alpha/beta style version number. Causes version number to be "0.00_01" unless -v is specified. -C, --omit-changes Omits creation of the Changes file, and adds a HISTORY section to the POD template. -F, --cpp-flags=addflags Additional flags to specify to C preprocessor when scanning header for function declarations. Writes these options in the generated Makefile.PL too. -M, --func-mask=regular expression selects functions/macros to process. -O, --overwrite-ok Allows a pre-existing extension directory to be overwritten. -P, --omit-pod Omit the autogenerated stub POD section. -X, --omit-XS Omit the XS portion. Used to generate a skeleton pure Perl module. "-c" and "-f" are implicitly enabled. -a, --gen-accessors Generate an accessor method for each element of structs and unions. The generated methods are named after the element name; will return the current value of the element if called without additional arguments; and will set the element to the supplied value (and return the new value) if called with an additional argument. Embedded structures and unions are returned as a pointer rather than the complete structure, to facilitate chained calls. These methods all apply to the Ptr type for the structure; additionally two methods are constructed for the structure type itself, "_to_ptr" which returns a Ptr type pointing to the same structure, and a "new" method to construct and return a new structure, initialised to zeroes. -b, --compat-version=version Generates a .pm file which is backwards compatible with the specified perl version. For versions < 5.6.0, the changes are. - no use of 'our' (uses 'use vars' instead) - no 'use warnings' Specifying a compatibility version higher than the version of perl you are using to run h2xs will have no effect. If unspecified h2xs will default to compatibility with the version of perl you are using to run h2xs. -c, --omit-constant Omit "constant()" from the .xs file and corresponding specialised "AUTOLOAD" from the .pm file. -d, --debugging Turn on debugging messages. -e, --omit-enums=[regular expression] If regular expression is not given, skip all constants that are defined in a C enumeration. Otherwise skip only those constants that are defined in an enum whose name matches regular expression. Since regular expression is optional, make sure that this switch is followed by at least one other switch if you omit regular expression and have some pending arguments such as header-file names. This is ok: h2xs -e -n Module::Foo foo.h This is not ok: h2xs -n Module::Foo -e foo.h In the latter, foo.h is taken as regular expression. -f, --force Allows an extension to be created for a header even if that header is not found in standard include directories. -g, --global Include code for safely storing static data in the .xs file. Extensions that do no make use of static data can ignore this option. -h, -?, --help Print the usage, help and version for this h2xs and exit. -k, --omit-const-func For function arguments declared as "const", omit the const attribute in the generated XS code. -m, --gen-tied-var Experimental: for each variable declared in the header file(s), declare a perl variable of the same name magically tied to the C variable. -n, --name=module_name Specifies a name to be used for the extension, e.g., -n RPC::DCE -o, --opaque-re=regular expression Use "opaque" data type for the C types matched by the regular expression, even if these types are "typedef"-equivalent to types from typemaps. Should not be used without -x. This may be useful since, say, types which are "typedef"-equivalent to integers may represent OS-related handles, and one may want to work with these handles in OO-way, as in "$handle->do_something()". Use "-o ." if you want to handle all the "typedef"ed types as opaque types. The type-to-match is whitewashed (except for commas, which have no whitespace before them, and multiple "*" which have no whitespace between them). -p, --remove-prefix=prefix Specify a prefix which should be removed from the Perl function names, e.g., -p sec_rgy_ This sets up the XS PREFIX keyword and removes the prefix from functions that are autoloaded via the "constant()" mechanism. -s, --const-subs=sub1,sub2 Create a perl subroutine for the specified macros rather than autoload with the constant() subroutine. These macros are assumed to have a return type of char *, e.g., -s sec_rgy_wildcard_name,sec_rgy_wildcard_sid. -t, --default-type=type Specify the internal type that the constant() mechanism uses for macros. The default is IV (signed integer). Currently all macros found during the header scanning process will be assumed to have this type. Future versions of "h2xs" may gain the ability to make educated guesses. --use-new-tests When --compat-version (-b) is present the generated tests will use "Test::More" rather than "Test" which is the default for versions before 5.6.2. "Test::More" will be added to PREREQ_PM in the generated "Makefile.PL". --use-old-tests Will force the generation of test code that uses the older "Test" module. --skip-exporter Do not use "Exporter" and/or export any symbol. --skip-ppport Do not use "Devel::PPPort": no portability to older version. --skip-autoloader Do not use the module "AutoLoader"; but keep the constant() function and "sub AUTOLOAD" for constants. --skip-strict Do not use the pragma "strict". --skip-warnings Do not use the pragma "warnings". -v, --version=version Specify a version number for this extension. This version number is added to the templates. The default is 0.01, or 0.00_01 if "-B" is specified. The version specified should be numeric. -x, --autogen-xsubs Automatically generate XSUBs basing on function declarations in the header file. The package "C::Scan" should be installed. If this option is specified, the name of the header file may look like "NAME1,NAME2". In this case NAME1 is used instead of the specified string, but XSUBs are emitted only for the declarations included from file NAME2. Note that some types of arguments/return-values for functions may result in XSUB-declarations/typemap-entries which need hand- editing. Such may be objects which cannot be converted from/to a pointer (like "long long"), pointers to functions, or arrays. See also the section on "LIMITATIONS of -x".
|
# Default behavior, extension is Rusers h2xs rpcsvc/rusers # Same, but extension is RUSERS h2xs -n RUSERS rpcsvc/rusers # Extension is rpcsvc::rusers. Still finds <rpcsvc/rusers.h> h2xs rpcsvc::rusers # Extension is ONC::RPC. Still finds <rpcsvc/rusers.h> h2xs -n ONC::RPC rpcsvc/rusers # Without constant() or AUTOLOAD h2xs -c rpcsvc/rusers # Creates templates for an extension named RPC h2xs -cfn RPC # Extension is ONC::RPC. h2xs -cfn ONC::RPC # Extension is a pure Perl module with no XS code. h2xs -X My::Module # Extension is Lib::Foo which works at least with Perl5.005_03. # Constants are created for all #defines and enums h2xs can find # in foo.h. h2xs -b 5.5.3 -n Lib::Foo foo.h # Extension is Lib::Foo which works at least with Perl5.005_03. # Constants are created for all #defines but only for enums # whose names do not start with 'bar_'. h2xs -b 5.5.3 -e '^bar_' -n Lib::Foo foo.h # Makefile.PL will look for library -lrpc in # additional directory /opt/net/lib h2xs rpcsvc/rusers -L/opt/net/lib -lrpc # Extension is DCE::rgynbase # prefix "sec_rgy_" is dropped from perl function names h2xs -n DCE::rgynbase -p sec_rgy_ dce/rgynbase # Extension is DCE::rgynbase # prefix "sec_rgy_" is dropped from perl function names # subroutines are created for sec_rgy_wildcard_name and # sec_rgy_wildcard_sid h2xs -n DCE::rgynbase -p sec_rgy_ \ -s sec_rgy_wildcard_name,sec_rgy_wildcard_sid dce/rgynbase # Make XS without defines in perl.h, but with function declarations # visible from perl.h. Name of the extension is perl1. # When scanning perl.h, define -DEXT=extern -DdEXT= -DINIT(x)= # Extra backslashes below because the string is passed to shell. # Note that a directory with perl header files would # be added automatically to include path. h2xs -xAn perl1 -F "-DEXT=extern -DdEXT= -DINIT\(x\)=" perl.h # Same with function declaration in proto.h as visible from perl.h. h2xs -xAn perl2 perl.h,proto.h # Same but select only functions which match /^av_/ h2xs -M '^av_' -xAn perl2 perl.h,proto.h # Same but treat SV* etc as "opaque" types h2xs -o '^[S]V \*$' -M '^av_' -xAn perl2 perl.h,proto.h Extension based on .h and .c files Suppose that you have some C files implementing some functionality, and the corresponding header files. How to create an extension which makes this functionality accessible in Perl? The example below assumes that the header files are interface_simple.h and interface_hairy.h, and you want the perl module be named as "Ext::Ension". If you need some preprocessor directives and/or linking with external libraries, see the flags "-F", "-L" and "-l" in "OPTIONS". Find the directory name Start with a dummy run of h2xs: h2xs -Afn Ext::Ension The only purpose of this step is to create the needed directories, and let you know the names of these directories. From the output you can see that the directory for the extension is Ext/Ension. Copy C files Copy your header files and C files to this directory Ext/Ension. Create the extension Run h2xs, overwriting older autogenerated files: h2xs -Oxan Ext::Ension interface_simple.h interface_hairy.h h2xs looks for header files after changing to the extension directory, so it will find your header files OK. Archive and test As usual, run cd Ext/Ension perl Makefile.PL make dist make make test Hints It is important to do "make dist" as early as possible. This way you can easily merge(1) your changes to autogenerated files if you decide to edit your ".h" files and rerun h2xs. Do not forget to edit the documentation in the generated .pm file. Consider the autogenerated files as skeletons only, you may invent better interfaces than what h2xs could guess. Consider this section as a guideline only, some other options of h2xs may better suit your needs. ENVIRONMENT No environment variables are used. AUTHOR Larry Wall and others SEE ALSO perl, perlxstut, ExtUtils::MakeMaker, and AutoLoader. DIAGNOSTICS The usual warnings if it cannot read or write the files involved. LIMITATIONS of -x h2xs would not distinguish whether an argument to a C function which is of the form, say, "int *", is an input, output, or input/output parameter. In particular, argument declarations of the form int foo(n) int *n should be better rewritten as int foo(n) int &n if "n" is an input parameter. Additionally, h2xs has no facilities to intuit that a function int foo(addr,l) char *addr int l takes a pair of address and length of data at this address, so it is better to rewrite this function as int foo(sv) SV *addr PREINIT: STRLEN len; char *s; CODE: s = SvPV(sv,len); RETVAL = foo(s, len); OUTPUT: RETVAL or alternately static int my_foo(SV *sv) { STRLEN len; char *s = SvPV(sv,len); return foo(s, len); } MODULE = foo PACKAGE = foo PREFIX = my_ int foo(sv) SV *sv See perlxs and perlxstut for additional details. perl v5.30.3 2024-04-13 H2XS(1)
|
scandeps5.34.pl
| null | null | null | null | null |
mp2bug
|
The mp2bug utilty is used to generate a mod_perl bug report template which can be filled out and submitted to the mod_perl project. SEE ALSO httpd(8), perl(1) macOS 14.5 January 23, 2008 macOS 14.5
|
mp2bug – mod_perl bug report generator
|
mp2bug
| null | null |
viewdiagnostic
|
viewdiagnostic translates crash reports and other diagnostic reports into a textual representation for improved visualization. Tools that parse data from the report should use its as-written format instead. macOS 14.5 macOS 14.5
|
viewdiagnostic – Visualize diagnostic reports
|
viewdiagnostic path
| null | null |
mkfifo
|
The mkfifo utility creates the fifos requested, in the order specified. The options are as follows: -m Set the file permission bits of the created fifos to the specified mode, ignoring the umask(2) of the calling process. The mode argument takes any format that can be specified to the chmod(1) command. If a symbolic mode is specified, the op symbols ‘+’ (plus) and ‘-’ (hyphen) are interpreted relative to an assumed initial mode of “a=rw” (read and write permissions for all). If the -m option is not specified, fifos are created with mode 0666 modified by the umask(2) of the calling process. The mkfifo utility requires write permission in the parent directory. EXIT STATUS The mkfifo utility exits 0 on success, and >0 if an error occurs. SEE ALSO mkdir(1), rm(1), umask(1), mkfifo(2), mknod(2), umask(2), compat(5), mknod(8) STANDARDS The mkfifo utility is expected to be IEEE Std 1003.2 (“POSIX.2”) compliant. HISTORY The mkfifo command appeared in 4.4BSD. macOS 14.5 January 5, 1994 macOS 14.5
|
mkfifo – make fifos
|
mkfifo [-m mode] fifo_name ...
| null | null |
toe
|
With no options, toe lists all available terminal types by primary name with descriptions. File arguments specify the directories to be scanned; if no such arguments are given, your default terminfo directory is scanned. If you also specify the -h option, a directory header will be issued as each directory is entered. There are other options intended for use by terminfo file maintainers: -a report on all of the terminal databases which ncurses would search, rather than only the first one that it finds. If the -s is also given, toe adds a column to the report, showing (like conflict(1)) which entries which belong to a given terminal database. An "*" marks entries which differ, and "+" marks equivalent entries. -s sort the output by the entry names. -u file says to write a report to the standard output, listing dependencies in the given terminfo/termcap source file. The report condenses the `use' relation: each line consists of the primary name of a terminal that has use capabilities, followed by a colon, followed by the whitespace-separated primary names of all terminals which occur in those use capabilities, followed by a newline -U file says to write a report to the standard output, listing reverse dependencies in the given terminfo/termcap source file. The report reverses the `use' relation: each line consists of the primary name of a terminal that occurs in use capabilities, followed by a colon, followed by the whitespace-separated primary names of all terminals which depend on it, followed by a newline. -vn specifies that (verbose) output be written to standard error, showing toe's progress. The optional parameter n is a number from 1 to 10, interpreted as for tic(1M). -V reports the version of ncurses which was used in this program, and exits. FILES /usr/share/terminfo/?/* Compiled terminal description database. SEE ALSO tic(1M), infocmp(1M), captoinfo(1M), infotocap(1M), curses(3X), terminfo(5). This describes ncurses version 5.7 (patch 20081102). toe(1M)
|
toe - table of (terminfo) entries
|
toe [-v[n]] [-ahsuUV] file...
| null | null |
hexdump
|
The hexdump utility is a filter which displays the specified files, or the standard input, if no files are specified, in a user specified format. The options are as follows: -b One-byte octal display. Display the input offset in hexadecimal, followed by sixteen space-separated, three column, zero-filled, bytes of input data, in octal, per line. -c One-byte character display. Display the input offset in hexadecimal, followed by sixteen space-separated, three column, space-filled, characters of input data per line. -C Canonical hex+ASCII display. Display the input offset in hexadecimal, followed by sixteen space-separated, two column, hexadecimal bytes, followed by the same sixteen bytes in %_p format enclosed in ``|'' characters. -d Two-byte decimal display. Display the input offset in hexadecimal, followed by eight space-separated, five column, zero-filled, two-byte units of input data, in unsigned decimal, per line. -e format_string Specify a format string to be used for displaying data. -f format_file Specify a file that contains one or more newline separated format strings. Empty lines and lines whose first non-blank character is a hash mark (#) are ignored. -n length Interpret only length bytes of input. -o Two-byte octal display. Display the input offset in hexadecimal, followed by eight space-separated, six column, zero-filled, two byte quantities of input data, in octal, per line. -s offset Skip offset bytes from the beginning of the input. By default, offset is interpreted as a decimal number. With a leading 0x or 0X, offset is interpreted as a hexadecimal number, otherwise, with a leading 0, offset is interpreted as an octal number. Appending the character b, k, m, or g to offset causes it to be interpreted as a multiple of 512, 1024, 1048576, or 1073741824, respectively. -v Cause hexdump to display all input data. Without the -v option, any number of groups of output lines, which would be identical to the immediately preceding group of output lines (except for the input offsets), are replaced with a line comprised of a single asterisk. -x Two-byte hexadecimal display. Display the input offset in hexadecimal, followed by eight, space separated, four column, zero-filled, two-byte quantities of input data, in hexadecimal, per line. For each input file, hexdump sequentially copies the input to standard output, transforming the data according to the format strings specified by the -e and -f options, in the order that they were specified. Formats A format string contains any number of format units, separated by whitespace. A format unit contains up to three items: an iteration count, a byte count, and a format. The iteration count is an optional positive integer, which defaults to one. Each format is applied iteration count times. The byte count is an optional positive integer. If specified it defines the number of bytes to be interpreted by each iteration of the format. If an iteration count and/or a byte count is specified, a single slash must be placed after the iteration count and/or before the byte count to disambiguate them. Any whitespace before or after the slash is ignored. The format is required and must be surrounded by double quote (" ") marks. It is interpreted as a fprintf-style format string (see fprintf(3)), with the following exceptions: • An asterisk (*) may not be used as a field width or precision. • A byte count or field precision is required for each ``s'' conversion character (unlike the fprintf(3) default which prints the entire string if the precision is unspecified). • The conversion characters ``h'', ``l'', ``n'', ``p'' and ``q'' are not supported. • The single character escape sequences described in the C standard are supported: NUL \0 <alert character> \a <backspace> \b <form-feed> \f <newline> \n <carriage return> \r <tab> \t <vertical tab> \v The hexdump utility also supports the following additional conversion strings: _a[dox] Display the input offset, cumulative across input files, of the next byte to be displayed. The appended characters d, o, and x specify the display base as decimal, octal or hexadecimal respectively. _A[dox] Identical to the _a conversion string except that it is only performed once, when all of the input data has been processed. _c Output characters in the default character set. Nonprinting characters are displayed in three character, zero-padded octal, except for those representable by standard escape notation (see above), which are displayed as two character strings. _p Output characters in the default character set. Nonprinting characters are displayed as a single “.”. _u Output US ASCII characters, with the exception that control characters are displayed using the following, lower-case, names. Characters greater than 0xff, hexadecimal, are displayed as hexadecimal strings. 000 NUL 001 SOH 002 STX 003 ETX 004 EOT 005 ENQ 006 ACK 007 BEL 008 BS 009 HT 00A LF 00B VT 00C FF 00D CR 00E SO 00F SI 010 DLE 011 DC1 012 DC2 013 DC3 014 DC4 015 NAK 016 SYN 017 ETB 018 CAN 019 EM 01A SUB 01B ESC 01C FS 01D GS 01E RS 01F US 07F DEL The default and supported byte counts for the conversion characters are as follows: %_c, %_p, %_u, %c One byte counts only. %d, %i, %o, %u, %X, %x Four byte default, one, two and four byte counts supported. %E, %e, %f, %G, %g Eight byte default, four and twelve byte counts supported. The amount of data interpreted by each format string is the sum of the data required by each format unit, which is the iteration count times the byte count, or the iteration count times the number of bytes required by the format if the byte count is not specified. The input is manipulated in ``blocks'', where a block is defined as the largest amount of data specified by any format string. Format strings interpreting less than an input block's worth of data, whose last format unit both interprets some number of bytes and does not have a specified iteration count, have the iteration count incremented until the entire input block has been processed or there is not enough data remaining in the block to satisfy the format string. If, either as a result of user specification or hexdump modifying the iteration count as described above, an iteration count is greater than one, no trailing whitespace characters are output during the last iteration. It is an error to specify a byte count as well as multiple conversion characters or strings unless all but one of the conversion characters or strings is _a or _A. If, as a result of the specification of the -n option or end-of-file being reached, input data only partially satisfies a format string, the input block is zero-padded sufficiently to display all available data (i.e., any format units overlapping the end of data will display some number of the zero bytes). Further output by such format strings is replaced by an equivalent number of spaces. An equivalent number of spaces is defined as the number of spaces output by an s conversion character with the same field width and precision as the original conversion character or conversion string but with any “+”, “ ”, “#” conversion flag characters removed, and referencing a NULL string. If no format strings are specified, the default display is equivalent to specifying the -x option. EXIT STATUS The hexdump utility exits 0 on success, and >0 if an error occurs.
|
hexdump – ASCII, decimal, hexadecimal, octal dump
|
hexdump [-bcCdovx] [-e format_string] [-f format_file] [-n length] [-s offset] file ...
| null |
Dump input in canonical (hex+ASCII) form: $ echo "FreeBSD: The power to serve" | hexdump -C 00000000 46 72 65 65 42 53 44 3a 20 54 68 65 20 70 6f 77 |FreeBSD: The pow| 00000010 65 72 20 74 6f 20 73 65 72 76 65 0a |er to serve.| 0000001c Same as above but skipping the first 4 bytes of stdin and interpreting only 3 bytes of input: $ echo "FreeBSD: The power to serve" | hexdump -C -s 4 -n 3 00000004 42 53 44 |BSD| 00000007 Assuming a format file named format.txt with the following contents that specify a perusal format: "%06.6_ao " 12/1 "%3_u " "\t\t" "%_p " "\n" Dump input in canonical form using the format in format.txt: $ echo "FreeBSD" | hexdump -f format.txt -C 000000 F r e e B S D lf F r e e B S D . 00000000 46 72 65 65 42 53 44 0a |FreeBSD.| 00000008 Assuming a format file named format.txt with the following contents that simulate the -x option: "%07.7_Ax\n" "%07.7_ax " 8/2 "%04x " "\n" Dump input in canonical form using the format in format.txt: $ echo "FreeBSD: The power to serve" | hexdump -f format.txt -C 0000000 7246 6565 5342 3a44 5420 6568 7020 776f 00000000 46 72 65 65 42 53 44 3a 20 54 68 65 20 70 6f 77 |FreeBSD: The pow| 0000010 7265 7420 206f 6573 7672 0a65 00000010 65 72 20 74 6f 20 73 65 72 76 65 0a |er to serve.| 0000001c SEE ALSO od(1) macOS 14.5 June 29, 2020 macOS 14.5
|
Rez
|
Tools supporting Carbon development, including Rez, were deprecated with Xcode 6. The Rez tool compiles the resource fork of a file according to the textual description contained in the resource description files. These resource description files must contain both the type declarations and the resource definitions needed to compile the resources. This data can come directly from the resource description files, as in this example: Rez -F Carbon Carbon.r myResource.r The data can also come from other text files that are included in resource description files using #include and read directives. For example, myResource.r could use #include to include the Carbon.r file. INPUT Standard input, unless you specify one or more resource description files. OUTPUT Normally Rez writes the resource fork to the file Rez.out. You can specify a different output file by using the -o option. Errors and warnings are written to diagnostic output. The Rez tool does not write to standard output. ALIAS RESOLUTION This command resolves Finder aliases in all input and output file specifications. This includes input source files, listing files, output object files, paths specified with the -i and -s options, and paths specified within source code using INCLUDE or #include statements. This behavior may be changed for resource files by using the -noResolve option. STATUS Rez can return the following status codes: 0 no errors 1 error in parameters 2 syntax error in resource description file 3 I/O or program error Note If any errors are detected, Rez sets the output file's modification date to 0, which is January 1, 1904, 12:00 A.M. PARAMETERS rezFile1 [ rezFile2 ] ... Specifies one or more resource description files that contain type declarations and resource definitions. Typically this pairs a file containing only resource definitions with another containing only type declarations. The type declarations for the standard Macintosh resources are contained in the MacTypes.r file in the CarbonCore framework. You can also specify resource description files by using #include. For example, a file containing only resource definitions could include those containing the appropriate type declarations (for example, myResource.r could include <Carbon/MacTypes.r>). In addition, you can also include resource files that have already been compiled by Rez include directive. The -script option enables Rez to correctly process the 2- byte character sets for foreign-language script systems.
|
Rez - compiles resources (DEPRECATED)
|
Rez [ rezFile1 ] [ rezFile2 ] ... [ -align word | longword ] [ -a[ppend] ] [ -arch architecture ] [ -c[reator] creatorExpr ] [ -d[efine] macro [ = data ] ] [ -i directoryPath ] [ -F frameworkDirectoryPath ] [ -is[ysroot] sdkPath ] [ [ -m[odification] ] [ -noResolve [ output | include ] ] [ -o outputFile ] [ -ov ] [ -p[rogress] ] [ -rd ] [ -ro ] [ -s directoryPath ] [ -script Roman | Japanese | Korean | SimpChinese | TradChinese ] [ -t[ype] typeExpr ] [ -u[ndef] macro ] [ -useDF ]
|
-align word | longword Aligns resources along word or longword boundaries. This allows the Resource Manager to load these resources faster. The Rez tool ignores this option if you use -a[ppend]. -a[ppend] Appends new resources to the output file rather than replacing the output file. Note The Rez tool cannot append resources to a resource file that has its read-only bit set. It also cannot replace a resource that has its protected bit set unless you specify -ov. WARNING The Rez tool overwrites any existing resource of the same type and ID without any warning message. -arch architecture A synonym for -d __architecture__ If no -arch arguments are provided, Rez automatically adds one for the current system's architecture. -c[reator] creatorExpr Sets the output file creator. (The default value is '????'.) Note that creatorExpr is a Rez expression such as -c "3*200+5" If the creator begins with a letter and does not contain any spaces or special characters, you can simply type it in, as in this example: -c APPL Otherwise, you must format the creator as a numeric expression or as a literal expression such as -c " '@@@@' " -d[efine] macro [= data ] Defines the macro variable macro as having the value data. You can use this option more than once on a command line. macro Specifies the macro variable to be defined. data Specifies the value of macro. This is the same as writing #define macro [data] at the beginning of the input. If you do not specify data, Rez sets the value of data to the null string. Note that this still defines the macro. -i directoryPath Directs Rez to search this directory for #include files. You can specify this option more than once. The Rez tool searches directories in the order in which they appear on the command line. -F frameworkDirectoryPath Directs Rez to search this directory for frameworks referred to by framework-style #include references (e.g. <Carbon/Carbon.r>.) By default, Rez only searches the /System/Library/Frameworks directory; using this option you can specify other directories to be searched. You can specify this option more than once. The Rez tool searches directories in the order in which they appear on the command line. -is[ysroot] sdkPath Directs Rez to search for included files and frameworks in the designated SDK. If omitted, the system root ("/") is assumed. -m[odification] Does not change the output file's modification date. If an error occurs, the output file's modification date is set to zero, even if you use this option. A date of 0 means January 1, 1904, 12:00 A.M. -noResolve [output|include] Overrides the default alias resolution behavior by not resolving leaf aliases in the output resource file specification or in any included resource input files. output If output is specified, Rez overrides the default Finder alias resolution behavior by not resolving leaf aliases in the output resource file specification. This allows you to modify a Finder alias file directly. include If include is specified with -noResolve then Rez overrides the default Finder alias resolution behavior by not resolving leaf aliases in any included resource input files. This allows you to include resources directly from Finder alias files. -o outputFile Places output in the specified output file. The default output file is Rez.out. -ov Overrides the protected bit when replacing resources with -a[ppend]. -p[rogress] Writes version and progress information to diagnostic output. -rd Suppresses warning messages for redeclared resource types. -ro Sets the mapReadOnly flag in the resource map. -s directoryPath Directs Rez to search this directory for resource include files. You can specify this option more than once. The Rez tool searches directories in the order in which they appear on the command line. -script Roman | Japanese | Korean | SimpChinese | TradChinese Enables the recognition of any of several 2-byte character script systems to use when compiling and decompiling files. This option insures that 2-byte characters in strings are handled as indivisible entities. The default language is Roman and specifies 1-byte character sets. -t[ype] typeExpr Sets the type of the output file (the default is 'APPL'). Note that typeExpr is a Rez expression, such as -t "3*200+5" If the type begins with a letter and does not contain any spaces or special characters, you can simply type it in, as in this example: -t MPST Otherwise, you must format it as a numeric expression or literal expression, such as -t " '@@@@' " -u[ndef] macro Undefines the preset macro variable This is the same as writing #undef macro at the beginning of the input. This option can be repeated more than once on a command line. -useDF Reads and writes resource information from the files' data forks, instead of their resource forks.
|
The following command line generates a resource fork for the file Sample, based on the type declarations and resource definitions in Types.r and Sample.r: Rez Types.r Sample.r -o Sample SEE ALSO DeRez Mac OS X July 25, 2000 REZ(1)
|
cap_mkdb
|
Cap_mkdb builds a hashed database out of the getcap(3) logical database constructed by the concatenation of the specified files . The database is named by the basename of the first file argument and the string “.db”. The getcap(3) routines can access the database in this form much more quickly than they can the original text file(s). The ``tc'' capabilities of the records are expanded before the record is stored into the database. The options as as follows: -f outfile Specify a different database basename. -v Print out the number of capability records in the database. FORMAT Each record is stored in the database using two different types of keys. The first type is a key which consists of the first capability of the record (not including the trailing colon (``:'')) with a data field consisting of a special byte followed by the rest of the record. The special byte is either a 0 or 1, where a 0 means that the record is okay, and a 1 means that there was a ``tc'' capability in the record that couldn't be expanded. The second type is a key which consists of one of the names from the first capability of the record with a data field consisting a special byte followed by the the first capability of the record. The special byte is a 2. In normal operation names are looked up in the database, resulting in a key/data pair of the second type. The data field of this key/data pair is used to look up a key/data pair of the first type which has the real data associated with the name. RETURN VALUE The cap_mkdb utility exits 0 on success and >0 if an error occurs. SEE ALSO dbopen(3), getcap(3), termcap(5) macOS 14.5 June 6, 1993 macOS 14.5
|
cap_mkdb – create capability database
|
cap_mkdb [-v] [-f outfile] file1 [file2 ...]
| null | null |
ictool
| null | null | null | null | null |
cmp
|
Every cmd listed above is a (sub-)command of the openssl(1) application. It has its own detailed manual page at openssl-cmd(1). For example, to view the manual page for the openssl dgst command, type "man openssl-dgst".
|
asn1parse, ca, ciphers, cmp, cms, crl, crl2pkcs7, dgst, dhparam, dsa, dsaparam, ec, ecparam, enc, engine, errstr, gendsa, genpkey, genrsa, info, kdf, mac, nseq, ocsp, passwd, pkcs12, pkcs7, pkcs8, pkey, pkeyparam, pkeyutl, prime, rand, rehash, req, rsa, rsautl, s_client, s_server, s_time, sess_id, smime, speed, spkac, srp, storeutl, ts, verify, version, x509 - OpenSSL application commands
|
openssl cmd -help | [-option | -option arg] ... [arg] ...
|
Among others, every subcommand has a help option. -help Print out a usage message for the subcommand. SEE ALSO openssl(1), openssl-asn1parse(1), openssl-ca(1), openssl-ciphers(1), openssl-cmp(1), openssl-cms(1), openssl-crl(1), openssl-crl2pkcs7(1), openssl-dgst(1), openssl-dhparam(1), openssl-dsa(1), openssl-dsaparam(1), openssl-ec(1), openssl-ecparam(1), openssl-enc(1), openssl-engine(1), openssl-errstr(1), openssl-gendsa(1), openssl-genpkey(1), openssl-genrsa(1), openssl-info(1), openssl-kdf(1), openssl-mac(1), openssl-nseq(1), openssl-ocsp(1), openssl-passwd(1), openssl-pkcs12(1), openssl-pkcs7(1), openssl-pkcs8(1), openssl-pkey(1), openssl-pkeyparam(1), openssl-pkeyutl(1), openssl-prime(1), openssl-rand(1), openssl-rehash(1), openssl-req(1), openssl-rsa(1), openssl-rsautl(1), openssl-s_client(1), openssl-s_server(1), openssl-s_time(1), openssl-sess_id(1), openssl-smime(1), openssl-speed(1), openssl-spkac(1), openssl-srp(1), openssl-storeutl(1), openssl-ts(1), openssl-verify(1), openssl-version(1), openssl-x509(1), HISTORY Initially, the manual page entry for the "openssl cmd" command used to be available at cmd(1). Later, the alias openssl-cmd(1) was introduced, which made it easier to group the openssl commands using the apropos(1) command or the shell's tab completion. In order to reduce cluttering of the global manual page namespace, the manual page entries without the 'openssl-' prefix have been deprecated in OpenSSL 3.0 and will be removed in OpenSSL 4.0. COPYRIGHT Copyright 2019-2020 The OpenSSL Project Authors. All Rights Reserved. Licensed under the Apache License 2.0 (the "License"). You may not use this file except in compliance with the License. You can obtain a copy in the file LICENSE in the source distribution or at <https://www.openssl.org/source/license.html>. 3.3.1 2024-06-04 OPENSSL-CMDS(1ssl)
| null |
syscallbypid.d
|
This reports the number of each type of system call made by PID. This is useful to identify which process is causing the most system calls. This is based on a script from DExplorer. Since this uses DTrace, only users with root privileges can run this command.
|
syscallbypid.d - syscalls by process ID. Uses DTrace.
|
syscallbypid.d
| null |
This samples until Ctrl-C is hit. # syscallbypid.d FIELDS PID process ID CMD process name SYSCALL system call name COUNT number of system calls made in this sample DOCUMENTATION See the DTraceToolkit for further documentation under the Docs directory. The DTraceToolkit docs may include full worked examples with verbose descriptions explaining the output. EXIT syscallbypid.d will sample until Ctrl-C is hit. AUTHOR Brendan Gregg [Sydney, Australia] SEE ALSO procsystime(1M), dtrace(1M), truss(1) version 1.00 June 28, 2005 syscallbypid.d(1m)
|
snmpset
|
snmpset is an SNMP application that uses the SNMP SET request to set information on a network entity. One or more object identifiers (OIDs) must be given as arguments on the command line. A type and a value to be set must accompany each object identifier. Each variable name is given in the format specified in variables(5). The TYPE is a single character, one of: i INTEGER u UNSIGNED s STRING x HEX STRING d DECIMAL STRING n NULLOBJ o OBJID t TIMETICKS a IPADDRESS b BITS Most of these will use the obvious corresponding ASN.1 type. 's', 'x', 'd' and 'b' are all different ways of specifying an OCTET STRING value, and the 'u' unsigned type is also used for handling Gauge32 values. If you have the proper MIB file loaded, you can, in most cases, replace the type with an '=' sign. For an object of type OCTET STRING this will assume a string like the 's' type notation. For other types it will do "The Right Thing". For example: snmpset -c private -v 1 test-hub system.sysContact.0 s dpz@noc.rutgers.edu ip.ipforwarding.0 = 2 will set the variables sysContact.0 and ipForwarding.0: system.sysContact.0 = STRING: "dpz@noc.rutgers.edu" ip.ipForwarding.0 = INTEGER: not-forwarding(2) If the network entity has an error processing the request packet, an error packet will be returned and a message will be shown, helping to pinpoint in what way the request was malformed.
|
snmpset - communicates with a network entity using SNMP SET requests
|
snmpset [COMMON OPTIONS] AGENT OID TYPE VALUE [OID TYPE VALUE]...
|
snmpset takes the common options described in the snmpcmd(1) manual page. Note that snmpset REQUIRES an argument specifying the agent to query and at least one set of OID/type/value arguments, as described in there. SEE ALSO snmpcmd(1), variables(5). V5.6.2.1 19 Jun 2003 SNMPSET(1)
| null |
locate
|
The locate program searches a database for all pathnames which match the specified pattern. The database is recomputed periodically (usually weekly or daily), and contains the pathnames of all files which are publicly accessible. Shell globbing and quoting characters (“*”, “?”, “\”, “[” and “]”) may be used in pattern, although they will have to be escaped from the shell. Preceding any character with a backslash (“\”) eliminates any special meaning which it may have. The matching differs in that no characters must be matched explicitly, including slashes (“/”). As a special case, a pattern containing no globbing characters (“foo”) is matched as though it were “*foo*”. Historically, locate only stored characters between 32 and 127. The current implementation stores any character except newline (‘\n’) and NUL (‘\0’). The 8-bit character support does not waste extra space for plain ASCII file names. Characters less than 32 or greater than 127 are stored in 2 bytes. The following options are available: -0 Print pathnames separated by an ASCII NUL character (character code 0) instead of default NL (newline, character code 10). -S Print some statistics about the database and exit. -c Suppress normal output; instead print a count of matching file names. -d database Search in database instead of the default file name database. Multiple -d options are allowed. Each additional -d option adds the specified database to the list of databases to be searched. The option database may be a colon-separated list of databases. A single colon is a reference to the default database. $ locate -d $HOME/lib/mydb: foo will first search string “foo” in $HOME/lib/mydb and then in /var/db/locate.database. $ locate -d $HOME/lib/mydb::/cdrom/locate.database foo will first search string “foo” in $HOME/lib/mydb and then in /var/db/locate.database and then in /cdrom/locate.database. $ locate -d db1 -d db2 -d db3 pattern is the same as $ locate -d db1:db2:db3 pattern or $ locate -d db1:db2 -d db3 pattern If - is given as the database name, standard input will be read instead. For example, you can compress your database and use: $ zcat database.gz | locate -d - pattern This might be useful on machines with a fast CPU and little RAM and slow I/O. Note: you can only use one pattern for stdin. -i Ignore case distinctions in both the pattern and the database. -l number Limit output to number of file names and exit. -m Use mmap(2) instead of the stdio(3) library. This is the default behavior and is faster in most cases. -s Use the stdio(3) library instead of mmap(2). ENVIRONMENT LOCATE_PATH path to the locate database if set and not empty, ignored if the -d option was specified. FILES /var/db/locate.database locate database /usr/libexec/locate.updatedb Script to update the locate database /System/Library/LaunchDaemons/com.apple.locate.plist Job that starts the database rebuild SEE ALSO find(1), whereis(1), which(1), fnmatch(3), locate.updatedb(8) Woods, James A., “Finding Files Fast”, ;login, 8:1, pp. 8-10, 1983. HISTORY The locate command first appeared in 4.4BSD. Many new features were added in FreeBSD 2.2. BUGS The locate program may fail to list some files that are present, or may list files that have been removed from the system. This is because locate only reports files that are present in the database, which is typically only regenerated once a week by the /System/Library/LaunchDaemons/com.apple.locate.plist job. Use find(1) to locate files that are of a more transitory nature. The locate database is typically built by user “nobody” and the locate.updatedb(8) utility skips directories which are not readable for user “nobody”, group “nobody”, or world. For example, if your HOME directory is not world-readable, none of your files are in the database. The locate database is not byte order independent. It is not possible to share the databases between machines with different byte order. The current locate implementation understands databases in host byte order or network byte order if both architectures use the same integer size. So on a FreeBSD/i386 machine (little endian), you can read a locate database which was built on SunOS/sparc machine (big endian, net). The locate utility does not recognize multibyte characters. macOS 14.5 December 11, 2020 macOS 14.5
|
locate – find filenames quickly
|
locate [-0Scims] [-l limit] [-d database] pattern ...
| null | null |
trustcachectl
|
trustcachectl can load new trust caches into the kernel and query whether a given cd hash is present in a loaded trust cache. SUBCOMMANDS Note that cdhashes should be 40 hexadecimal characters. query cdhash Check if cdhash is in a loaded trust cache. load path Load the trust cache at path. This tool is only entitled to load the OTA update brain or global install assistant trust cache. macOS 14.5 September 19, 2022 macOS 14.5
|
trustcachectl – Load trust caches and query whether a cd hash is present in a loaded trust cache.
|
trustcachectl subcommand [arguments ...]
| null | null |
afconvert
|
Audio File Convert will convert a source audio file to a new audio file with the specified file and data types
|
afconvert – Audio File Convert
|
afconvert [option] src_audiofile dest_audiofile
|
-h print help text Darwin February 13, 2007 Darwin
| null |
afclip
| null | null | null | null | null |
tidy_changelog5.30
|
Takes a changelog file, parse it using CPAN::Changes and prints out the resulting output. If a file is not given, the program will see if there is one file in the current directory beginning by 'change' (case- insensitive) and, if so, assume it to be the changelog. ARGUMENTS --next If provided, assumes that there is a placeholder header for an upcoming next release. The placeholder token is given via --token. --token Regular expression to use to detect the token for an upcoming release if --next is used. If not explicitly given, defaults to "\{\{\$NEXT\}\}". --headers If given, only print out the release header lines, without any of the changes. --reverse Prints the releases in reverse order (from the oldest to latest). --check Only check if the changelog is formatted properly using the changes_file_ok function of Test::CPAN::Changes. --help This help perl v5.30.3 2014-10-10 TIDY_CHANGELOG(1)
|
tidy_changelog - command-line tool for CPAN::Changes
|
$ tidy_changelog Changelog
| null | null |
bzegrep
|
Bzgrep is used to invoke the grep on bzip2-compressed files. All options specified are passed directly to grep. If no file is specified, then the standard input is decompressed if necessary and fed to grep. Otherwise the given files are uncompressed if necessary and fed to grep. If bzgrep is invoked as bzegrep or bzfgrep then egrep or fgrep is used instead of grep. If the GREP environment variable is set, bzgrep uses it as the grep program to be invoked. For example: for sh: GREP=fgrep bzgrep string files for csh: (setenv GREP fgrep; bzgrep string files) AUTHOR Charles Levert (charles@comm.polymtl.ca). Adapted to bzip2 by Philippe Troin <phil@fifi.org> for Debian GNU/Linux. SEE ALSO grep(1), egrep(1), fgrep(1), bzdiff(1), bzmore(1), bzless(1), bzip2(1) BZGREP(1)
|
bzgrep, bzfgrep, bzegrep - search possibly bzip2 compressed files for a regular expression
|
bzgrep [ grep_options ] [ -e ] pattern filename... bzegrep [ egrep_options ] [ -e ] pattern filename... bzfgrep [ fgrep_options ] [ -e ] pattern filename...
| null | null |
instmodsh
|
A little interface to ExtUtils::Installed to examine installed modules, validate your packlists and even create a tarball from an installed module. SEE ALSO ExtUtils::Installed perl v5.38.2 2023-11-28 INSTMODSH(1)
|
instmodsh - A shell to examine installed modules
|
instmodsh
| null | null |
extcheck
| null | null | null | null | null |
javapackager
| null | null | null | null | null |
pl
|
pl can be used to check the syntax of old-style ASCII property list files. The options are: --input file Cause pl to use the specified file as input. If the input option is not specified, stdin is used. --output file Cause pl to use the specified file as output. If the output option is not specified, stdout is used. DIAGNOSTICS The pl command exits 0 on success, and <0 on failure. SEE ALSO plist(5) plutil(1) STANDARDS The pl command obeys no one's rules but its own. HISTORY The pl command first appeared in Mac OS X 10.5. Mac OS X October 23, 2007 Mac OS X
|
pl – ASCII property list utility
|
pl [--input file] [--output file]
| null | null |
compress
|
The compress utility reduces the size of files using adaptive Lempel-Ziv coding. Each file is renamed to the same name plus the extension .Z. A file argument with a .Z extension will be ignored except it will cause an error exit after other arguments are processed. If compression would not reduce the size of a file, the file is ignored. The uncompress utility restores compressed files to their original form, renaming the files by deleting the .Z extensions. A file specification need not include the file's .Z extension. If a file's name in its file system does not have a .Z extension, it will not be uncompressed and it will cause an error exit after other arguments are processed. If renaming the files would cause files to be overwritten and the standard input device is a terminal, the user is prompted (on the standard error output) for confirmation. If prompting is not possible or confirmation is not received, the files are not overwritten. As many of the modification time, access time, file flags, file mode, user ID, and group ID as allowed by permissions are retained in the new file. If no files are specified or a file argument is a single dash (‘-’), the standard input is compressed or uncompressed to the standard output. If either the input and output files are not regular files, the checks for reduction in size and file overwriting are not performed, the input file is not removed, and the attributes of the input file are not retained in the output file. The options are as follows: -b bits The code size (see below) is limited to bits, which must be in the range 9..16. The default is 16. -c Compressed or uncompressed output is written to the standard output. No files are modified. The -v option is ignored. Compression is attempted even if the results will be larger than the original. -f Files are overwritten without prompting for confirmation. Also, for compress, files are compressed even if they are not actually reduced in size. -v Print the percentage reduction of each file. Ignored by uncompress or if the -c option is also used. The compress utility uses a modified Lempel-Ziv algorithm. Common substrings in the file are first replaced by 9-bit codes 257 and up. When code 512 is reached, the algorithm switches to 10-bit codes and continues to use more bits until the limit specified by the -b option or its default is reached. After the limit is reached, compress periodically checks the compression ratio. If it is increasing, compress continues to use the existing code dictionary. However, if the compression ratio decreases, compress discards the table of substrings and rebuilds it from scratch. This allows the algorithm to adapt to the next "block" of the file. The -b option is unavailable for uncompress since the bits parameter specified during compression is encoded within the output, along with a magic number to ensure that neither decompression of random data nor recompression of compressed data is attempted. The amount of compression obtained depends on the size of the input, the number of bits per code, and the distribution of common substrings. Typically, text such as source code or English is reduced by 50-60%. Compression is generally much better than that achieved by Huffman coding (as used in the historical command pack), or adaptive Huffman coding (as used in the historical command compact), and takes less time to compute. If file is a soft or hard link compress will replace it with a compressed copy of the file pointed to by the link. The link's target file is left uncompressed. EXIT STATUS The compress and uncompress utilities exit 0 on success, and >0 if an error occurs. The compress utility exits 2 if attempting to compress a file would not reduce its size and the -f option was not specified and if no other error occurs.
|
compress, uncompress – compress and expand data
|
compress [-fv] [-b bits] [file ...] compress -c [-b bits] [file] uncompress [-fv] [file ...] uncompress -c [file ...]
| null |
Create a file test_file with a single line of text: echo "This is a test" > test_file Try to reduce the size of the file using a 10-bit code and show the exit status: $ compress -b 10 test_file $ echo $? 2 Try to compress the file and show compression percentage: $ compress -v test_file test_file: file would grow; left unmodified Same as above but forcing compression: $ compress -f -v test_file test_file.Z: 79% expansion Compress and uncompress the string ‘hello’ on the fly: $ echo "hello" | compress | uncompress hello SEE ALSO gunzip(1), gzexe(1), gzip(1), zcat(1), zmore(1), znew(1) Welch, Terry A., “A Technique for High Performance Data Compression”, IEEE Computer, 17:6, pp. 8-19, June, 1984. STANDARDS The compress and uncompress utilities conform to IEEE Std 1003.1-2001 (“POSIX.1”). HISTORY The compress command appeared in 4.3BSD. BUGS The program does not handle links well and has no link-handling options. Some of these might be considered otherwise-undocumented features. compress: If the utility does not compress a file because doing so would not reduce its size, and a file of the same name except with an .Z extension exists, the named file is not really ignored as stated above; it causes a prompt to confirm the overwriting of the file with the extension. If the operation is confirmed, that file is deleted. uncompress: If an empty file is compressed (using -f), the resulting .Z file is also empty. That seems right, but if uncompress is then used on that file, an error will occur. Both utilities: If a ‘-’ argument is used and the utility prompts the user, the standard input is taken as the user's reply to the prompt. Both utilities: If the specified file does not exist, but a similarly- named one with (for compress) or without (for uncompress) a .Z extension does exist, the utility will waste the user's time by not immediately emitting an error message about the missing file and continuing. Instead, it first asks for confirmation to overwrite the existing file and then does not overwrite it. macOS 14.5 March 4, 2021 macOS 14.5
|
fddist
|
This prints distributions for read and write events by file descriptor, by process. This can be used to determine which file descriptor a process is doing the most I/O with. Since this uses DTrace, only users with root privileges can run this command.
|
fddist - file descriptor usage distributions. Uses DTrace.
|
fddist [-r|-w]
|
-r reads only -w writes only
|
Sample both read and write activity, # fddist Sample reads only, # fddist -r FIELDS EXEC process name PID process ID value file descriptor count number of events BASED ON /usr/demo/dtrace/lquantize.d DOCUMENTATION DTrace Guide "Aggregations" chapter (docs.sun.com) See the DTraceToolkit for further documentation under the Docs directory. The DTraceToolkit docs may include full worked examples with verbose descriptions explaining the output. EXIT fddist will sample until Ctrl-C is hit. SEE ALSO dtrace(1M) version 0.70 June 8, 2005 fddist(1m)
|
db_upgrade
|
The db_upgrade utility upgrades the Berkeley DB version of one or more files and the databases they contain to the current release version. The options are as follows: -h Specify a home directory for the database environment; by default, the current working directory is used. -N Do not acquire shared region mutexes while running. Other problems, such as potentially fatal errors in Berkeley DB, will be ignored as well. This option is intended only for debugging errors, and should not be used under any other circumstances. -P Specify an environment password. Although Berkeley DB utilities overwrite password strings as soon as possible, be aware there may be a window of vulnerability on systems where unprivileged users can see command-line arguments or where utilities are not able to overwrite the memory containing the command-line arguments. -s This flag is only meaningful when upgrading databases from releases before the Berkeley DB 3.1 release. As part of the upgrade from the Berkeley DB 3.0 release to the 3.1 release, the on-disk format of duplicate data items changed. To correctly upgrade the format requires that applications specify whether duplicate data items in the database are sorted or not. Specifying the -s flag means that the duplicates are sorted; otherwise, they are assumed to be unsorted. Incorrectly specifying the value of this flag may lead to database corruption. Because the db_upgrade utility upgrades a physical file (including all the databases it contains), it is not possible to use db_upgrade to upgrade files where some of the databases it includes have sorted duplicate data items, and some of the databases it includes have unsorted duplicate data items. If the file does not have more than a single database, if the databases do not support duplicate data items, or if all the databases that support duplicate data items support the same style of duplicates (either sorted or unsorted), db_upgrade will work correctly as long as the -s flag is correctly specified. Otherwise, the file cannot be upgraded using db_upgrade, and must be upgraded manually using the db_dump and db_load utilities. -V Write the library version number to the standard output, and exit. It is important to realize that Berkeley DB database upgrades are done in place, and so are potentially destructive. This means that if the system crashes during the upgrade procedure, or if the upgrade procedure runs out of disk space, the databases may be left in an inconsistent and unrecoverable state. See Upgrading databases for more information. The db_upgrade utility may be used with a Berkeley DB environment (as described for the -h option, the environment variable DB_HOME, or because the utility was run in a directory containing a Berkeley DB environment). In order to avoid environment corruption when using a Berkeley DB environment, db_upgrade should always be given the chance to detach from the environment and exit gracefully. To cause db_upgrade to release all environment resources and exit cleanly, send it an interrupt signal (SIGINT). The db_upgrade utility exits 0 on success, and >0 if an error occurs. ENVIRONMENT DB_HOME If the -h option is not specified and the environment variable DB_HOME is set, it is used as the path of the database home, as described in DB_ENV->open. SEE ALSO db_archive(1), db_checkpoint(1), db_deadlock(1), db_dump(1), db_load(1), db_printlog(1), db_recover(1), db_stat(1), db_verify(1) Darwin December 3, 2003 Darwin
|
db_upgrade
|
db_upgrade [-NsV] [-h home] [-P password] file ...
| null | null |
diagnose-fu
| null | null | null | null | null |
yamlpp-load-dump
| null | null | null | null | null |
ctags
|
The ctags utility makes a tags file for ex(1) from the specified C, Pascal, Fortran, yacc(1), lex(1), and Lisp sources. A tags file gives the locations of specified objects in a group of files. Each line of the tags file contains the object name, the file in which it is defined, and a search pattern for the object definition, separated by white-space. Using the tags file, ex(1) can quickly locate these object definitions. Depending upon the options provided to ctags, objects will consist of subroutines, typedefs, defines, structs, enums and unions. The following options are available: -B Use backward searching patterns (?...?). -F Use forward searching patterns (/.../) (the default). -T Do not create tags for typedefs, structs, unions, and enums. -a Append to tags file. -d Create tags for #defines that do not take arguments; #defines that take arguments are tagged automatically. -f Place the tag descriptions in a file called tagsfile. The default behaviour is to place them in a file called tags. If tagsfile is “-”, the tags will be written to standard output instead. -u Update the specified files in the tags file, that is, all references to them are deleted, and the new values are appended to the file. This is ignored if the tags file does not exist or is not a regular file (e.g. -f- was used to write to standard output). Beware: this option is implemented in a way which is rather slow; it is usually faster to simply rebuild the tags file. -v An index of the form expected by vgrind(1) is produced on the standard output. This listing contains the object name, file name, and page number (assuming 64 line pages). Since the output will be sorted into lexicographic order, it may be desired to run the output through sort(1). Sample use: ctags -v files | sort -f > index vgrind -x index -w Suppress warning diagnostics. -x ctags produces a list of object names, the line number and file name on which each is defined, as well as the text of that line and prints this on the standard output. This is a simple index which can be printed out as an off-line readable function index. Files whose names end in .c or .h are assumed to be C source files and are searched for C style routine and macro definitions. Files whose names end in .y are assumed to be yacc(1) source files. Files whose names end in .l are assumed to be Lisp files if their first non-blank character is ‘;’, ‘(’, or ‘[’, otherwise, they are treated as lex(1) files. Other files are first examined to see if they contain any Pascal or Fortran routine definitions, and, if not, are searched for C style definitions. The tag “main” is treated specially in C programs. The tag formed is created by prepending ‘M’ to the name of the file, with the trailing .c and any leading pathname components removed. This makes use of ctags practical in directories with more than one program. The yacc(1) and lex(1) files each have a special tag. “yyparse” is the start of the second section of the yacc(1) file, and “yylex” is the start of the second section of the lex(1) file. FILES tags default output tags file EXIT STATUS The ctags utility exits with a value of 1 if an error occurred, 0 otherwise. Duplicate objects are not considered errors. COMPATIBILITY The -t option is a no-op for compatibility with previous versions of ctags that did not create tags for typedefs, enums, structs and unions by default. SEE ALSO ex(1), vi(1) STANDARDS The ctags utility conforms to IEEE Std 1003.1-2001 (“POSIX.1”). HISTORY The ctags utility appeared in 3.0BSD. BUGS Recognition of functions, subroutines and procedures for Fortran and Pascal is done in a very simpleminded way. No attempt is made to deal with block structure; if you have two Pascal procedures in different blocks with the same name you lose. The ctags utility does not understand about Pascal types. The method of deciding whether to look for C, Pascal or Fortran functions is a hack. The ctags utility relies on the input being well formed, and any syntactical errors will completely confuse it. It also finds some legal syntax confusing; for example, since it does not understand #ifdef's (incidentally, that is a feature, not a bug), any code with unbalanced braces inside #ifdef's will cause it to become somewhat disoriented. In a similar fashion, multiple line changes within a definition will cause it to enter the last line of the object, rather than the first, as the searching pattern. The last line of multiple line typedef's will similarly be noted. macOS 14.5 May 23, 2023 macOS 14.5
|
ctags – create a tags file
|
ctags [-BFTaduwvx] [-f tagsfile] file ...
| null | null |
zdiff
|
zcmp and zdiff are filters that invoke cmp(1) or diff(1) respectively to compare compressed files. Any options that are specified are passed to cmp(1) or diff(1). If only file1 is specified, it is compared against a file with the same name, but with the extension removed. When both file1 or file2 are specified, either file may be compressed. Extensions handled by gzip(1): • z, Z, • gz, • taz, • tgz. Extensions handled by bzip2(1): • bz, • bz2, • tbz, • tbz2. Extensions handled by xz(1): • lzma, • xz, • tlz, • txz. ENVIRONMENT TMPDIR Directory in which to place temporary files. If unset, /tmp is used. FILES /tmp/zcmp.XXXXXXXXXX Temporary file for zcmp. /tmp/zdiff.XXXXXXXXXX Temporary file for zdiff. SEE ALSO bzip2(1), cmp(1), diff(1), gzip(1), xz(1) CAVEATS zcmp and zdiff rely solely on the file extension to determine what is, or is not, a compressed file. Consequently, the following are not supported as arguments: - directories - device special files - filenames indicating the standard input (“-”) macOS 14.5 May 23, 2011 macOS 14.5
|
zcmp, zdiff – compare compressed files
|
zcmp [options] file [file2] zdiff [options] file [file2]
| null | null |
yamlpp-parse-emit
| null | null | null | null | null |
pathchk
|
The pathchk utility checks whether each of the specified pathname arguments is valid or portable. A diagnostic message is written for each argument that: • Is longer than PATH_MAX bytes. • Contains any component longer than NAME_MAX bytes. (The value of NAME_MAX depends on the underlying file system.) • Contains a directory component that is not searchable. It is not considered an error if a pathname argument contains a nonexistent component as long as a component by that name could be created. The options are as follows: -p Perform portability checks on the specified pathname arguments. Diagnostic messages will be written for each argument that: • Is longer than _POSIX_PATH_MAX (255) bytes. • Contains a component longer than _POSIX_NAME_MAX (14) bytes. • Contains any character not in the portable filename character set (that is, alphanumeric characters, ‘.’, ‘-’ and ‘_’). No component may start with the hyphen (‘-’) character. -P In addition to the default or -p checks, write a diagnostic for each argument that: • Is empty. • Contains a component that starts with a hyphen. EXIT STATUS The pathchk utility exits 0 on success, and >0 if an error occurs.
|
pathchk – check pathnames
|
pathchk [-pP] pathname ...
| null |
Check whether the names of files in the current directory are portable to other POSIX systems: find . -exec pathchk -p -- {} + SEE ALSO getconf(1), pathconf(2), stat(2) STANDARDS The pathchk utility conforms to IEEE Std 1003.1-2001 (“POSIX.1”). HISTORY A pathchk utility appeared in FreeBSD 5.0. macOS 14.5 May 1, 2010 macOS 14.5
|
mesg
|
The mesg utility is invoked by a user to control write access others have to a terminal device. Write access is allowed by default, and programs such as talk(1) and write(1) may display messages on the terminal. The first terminal device in the sequence of devices associated with standard input, standard output and standard error is affected. The following options are available: n Disallow messages. y Permit messages to be displayed. If no arguments are given, mesg displays the present message status to the standard output. EXIT STATUS The mesg utility exits with one of the following values: 0 Messages are allowed. 1 Messages are not allowed. >1 An error has occurred.
|
mesg – display (do not display) messages from other users
|
mesg [n | y]
| null |
Disallow messages from other users to the current terminal: mesg n Allow messages from other users to ttyp1 (assuming you are also logged in on that terminal): mesg y </dev/ttyp1 COMPATIBILITY Previous versions of the mesg utility wrote the message status to the standard error output and affected the terminal attached to standard error without first trying the standard input or output devices. SEE ALSO biff(1), talk(1), wall(1), write(1) STANDARDS The mesg utility conforms to IEEE Std 1003.1-2001 (“POSIX.1”). HISTORY A mesg command appeared in Version 1 AT&T UNIX. macOS 14.5 May 5, 2002 macOS 14.5
|
setregion
|
setregion is used by the DVD Player to set the disc region code for a DVD drive. Mac OS X December 5, 2006 Mac OS X
|
setregion
|
has no command line interface
| null | null |
perlivp
|
The perlivp program is set up at Perl source code build time to test the Perl version it was built under. It can be used after running: make install (or your platform's equivalent procedure) to verify that perl and its libraries have been installed correctly. A correct installation is verified by output that looks like: ok 1 ok 2 etc.
|
perlivp - Perl Installation Verification Procedure
|
perlivp [-p] [-v] [-h]
|
-h help Prints out a brief help message. -p print preface Gives a description of each test prior to performing it. -v verbose Gives more detailed information about each test, after it has been performed. Note that any failed tests ought to print out some extra information whether or not -v is thrown. DIAGNOSTICS • print "# Perl binary '$perlpath' does not appear executable.\n"; Likely to occur for a perl binary that was not properly installed. Correct by conducting a proper installation. • print "# Perl version '$]' installed, expected $ivp_VERSION.\n"; Likely to occur for a perl that was not properly installed. Correct by conducting a proper installation. • print "# Perl \@INC directory '$_' does not appear to exist.\n"; Likely to occur for a perl library tree that was not properly installed. Correct by conducting a proper installation. • print "# Needed module '$_' does not appear to be properly installed.\n"; One of the two modules that is used by perlivp was not present in the installation. This is a serious error since it adversely affects perlivp's ability to function. You may be able to correct this by performing a proper perl installation. • print "# Required module '$_' does not appear to be properly installed.\n"; An attempt to "eval "require $module"" failed, even though the list of extensions indicated that it should succeed. Correct by conducting a proper installation. • print "# Unnecessary module 'bLuRfle' appears to be installed.\n"; This test not coming out ok could indicate that you have in fact installed a bLuRfle.pm module or that the "eval " require \"$module_name.pm\"; "" test may give misleading results with your installation of perl. If yours is the latter case then please let the author know. • print "# file",+($#missing == 0) ? '' : 's'," missing from installation:\n"; One or more files turned up missing according to a run of "ExtUtils::Installed -> validate()" over your installation. Correct by conducting a proper installation. For further information on how to conduct a proper installation consult the INSTALL file that comes with the perl source and the README file for your platform. AUTHOR Peter Prymmer perl v5.38.2 2023-11-28 PERLIVP(1)
| null |
yamlpp-highlight5.30
| null | null | null | null | null |
xpath5.30
|
xpath uses the XML::XPath perl module to make XPath queries to any XML document. The XML::XPath module aims to comply exactly to the XPath specification at "http://www.w3.org/TR/xpath" and yet allows extensions to be added in the form of functions. The script takes any number of XPath pointers and tries to apply them to each XML document given on the command line. If no file arguments are given, the query is done using "STDIN" as an XML document. When multiple queries exist, the result of the last query is used as context for the next query and only the result of the last one is output. The context of the first query is always the root of the current document.
|
xpath - a script to query XPath statements in XML documents.
|
xpath [-s suffix] [-p prefix] [-n] [-q] -e query [-e query] ... [file] ...
|
-q Be quiet. Output only errors (and no separator) on stderr. -n Never use an external DTD, ie. instantiate the XML::Parser module with 'ParseParamEnt => 0'. -s suffix Place "suffix" at the end of each entry. Default is a linefeed. -p prefix Place "prefix" preceding each entry. Default is nothing. BUGS The author of this man page is not very fluant in english. Please, send him (fabien@tzone.org) any corrections concerning this text. SEE ALSO XML::XPath LICENSE AND COPYRIGHT This module is copyright 2000 AxKit.com Ltd. This is free software, and as such comes with NO WARRANTY. No dates are used in this module. You may distribute this module under the terms of either the Gnu GPL, or the Artistic License (the same terms as Perl itself). For support, please subscribe to the Perl-XML <http://listserv.activestate.com/mailman/listinfo/perl-xml> mailing list at the URL perl v5.30.3 2017-07-27 XPATH(1)
| null |
c++
|
clang is a C, C++, and Objective-C compiler which encompasses preprocessing, parsing, optimization, code generation, assembly, and linking. Depending on which high-level mode setting is passed, Clang will stop before doing a full link. While Clang is highly integrated, it is important to understand the stages of compilation, to understand how to invoke it. These stages are: Driver The clang executable is actually a small driver which controls the overall execution of other tools such as the compiler, assembler and linker. Typically you do not need to interact with the driver, but you transparently use it to run the other tools. Preprocessing This stage handles tokenization of the input source file, macro expansion, #include expansion and handling of other preprocessor directives. The output of this stage is typically called a ".i" (for C), ".ii" (for C++), ".mi" (for Objective-C), or ".mii" (for Objective-C++) file. Parsing and Semantic Analysis This stage parses the input file, translating preprocessor tokens into a parse tree. Once in the form of a parse tree, it applies semantic analysis to compute types for expressions as well and determine whether the code is well formed. This stage is responsible for generating most of the compiler warnings as well as parse errors. The output of this stage is an "Abstract Syntax Tree" (AST). Code Generation and Optimization This stage translates an AST into low-level intermediate code (known as "LLVM IR") and ultimately to machine code. This phase is responsible for optimizing the generated code and handling target-specific code generation. The output of this stage is typically called a ".s" file or "assembly" file. Clang also supports the use of an integrated assembler, in which the code generator produces object files directly. This avoids the overhead of generating the ".s" file and of calling the target assembler. Assembler This stage runs the target assembler to translate the output of the compiler into a target object file. The output of this stage is typically called a ".o" file or "object" file. Linker This stage runs the target linker to merge multiple object files into an executable or dynamic library. The output of this stage is typically called an "a.out", ".dylib" or ".so" file. Clang Static Analyzer The Clang Static Analyzer is a tool that scans source code to try to find bugs through code analysis. This tool uses many parts of Clang and is built into the same driver. Please see <https://clang-analyzer.llvm.org> for more details on how to use the static analyzer.
|
clang - the Clang C, C++, and Objective-C compiler
|
clang [options] filename ...
|
Stage Selection Options -E Run the preprocessor stage. -fsyntax-only Run the preprocessor, parser and semantic analysis stages. -S Run the previous stages as well as LLVM generation and optimization stages and target-specific code generation, producing an assembly file. -c Run all of the above, plus the assembler, generating a target ".o" object file. no stage selection option If no stage selection option is specified, all stages above are run, and the linker is run to combine the results into an executable or shared library. Language Selection and Mode Options -x <language> Treat subsequent input files as having type language. -std=<standard> Specify the language standard to compile for. Supported values for the C language are: c89 c90 iso9899:1990 ISO C 1990 iso9899:199409 ISO C 1990 with amendment 1 gnu89 gnu90 ISO C 1990 with GNU extensions c99 iso9899:1999 ISO C 1999 gnu99 ISO C 1999 with GNU extensions c11 iso9899:2011 ISO C 2011 gnu11 ISO C 2011 with GNU extensions c17 iso9899:2017 ISO C 2017 gnu17 ISO C 2017 with GNU extensions The default C language standard is gnu17, except on PS4, where it is gnu99. Supported values for the C++ language are: c++98 c++03 ISO C++ 1998 with amendments gnu++98 gnu++03 ISO C++ 1998 with amendments and GNU extensions c++11 ISO C++ 2011 with amendments gnu++11 ISO C++ 2011 with amendments and GNU extensions c++14 ISO C++ 2014 with amendments gnu++14 ISO C++ 2014 with amendments and GNU extensions c++17 ISO C++ 2017 with amendments gnu++17 ISO C++ 2017 with amendments and GNU extensions c++20 ISO C++ 2020 with amendments gnu++20 ISO C++ 2020 with amendments and GNU extensions c++2b Working draft for ISO C++ 2023 gnu++2b Working draft for ISO C++ 2023 with GNU extensions The default C++ language standard is gnu++98. Supported values for the OpenCL language are: cl1.0 OpenCL 1.0 cl1.1 OpenCL 1.1 cl1.2 OpenCL 1.2 cl2.0 OpenCL 2.0 The default OpenCL language standard is cl1.0. Supported values for the CUDA language are: cuda NVIDIA CUDA(tm) -stdlib=<library> Specify the C++ standard library to use; supported options are libstdc++ and libc++. If not specified, platform default will be used. -rtlib=<library> Specify the compiler runtime library to use; supported options are libgcc and compiler-rt. If not specified, platform default will be used. -ansi Same as -std=c89. -ObjC, -ObjC++ Treat source input files as Objective-C and Object-C++ inputs respectively. -trigraphs Enable trigraphs. -ffreestanding Indicate that the file should be compiled for a freestanding, not a hosted, environment. Note that it is assumed that a freestanding environment will additionally provide memcpy, memmove, memset and memcmp implementations, as these are needed for efficient codegen for many programs. -fno-builtin Disable special handling and optimizations of well-known library functions, like strlen() and malloc(). -fno-builtin-<function> Disable special handling and optimizations for the specific library function. For example, -fno-builtin-strlen removes any special handling for the strlen() library function. -fno-builtin-std-<function> Disable special handling and optimizations for the specific C++ standard library function in namespace std. For example, -fno-builtin-std-move_if_noexcept removes any special handling for the std::move_if_noexcept() library function. For C standard library functions that the C++ standard library also provides in namespace std, use -fno-builtin-<function> instead. -fmath-errno Indicate that math functions should be treated as updating errno. -fpascal-strings Enable support for Pascal-style strings with "\pfoo". -fms-extensions Enable support for Microsoft extensions. -fmsc-version= Set _MSC_VER. Defaults to 1300 on Windows. Not set otherwise. -fborland-extensions Enable support for Borland extensions. -fwritable-strings Make all string literals default to writable. This disables uniquing of strings and other optimizations. -flax-vector-conversions, -flax-vector-conversions=<kind>, -fno-lax-vector-conversions Allow loose type checking rules for implicit vector conversions. Possible values of <kind>: • none: allow no implicit conversions between vectors • integer: allow implicit bitcasts between integer vectors of the same overall bit-width • all: allow implicit bitcasts between any vectors of the same overall bit-width <kind> defaults to integer if unspecified. -fblocks Enable the "Blocks" language feature. -fobjc-abi-version=version Select the Objective-C ABI version to use. Available versions are 1 (legacy "fragile" ABI), 2 (non-fragile ABI 1), and 3 (non-fragile ABI 2). -fobjc-nonfragile-abi-version=<version> Select the Objective-C non-fragile ABI version to use by default. This will only be used as the Objective-C ABI when the non-fragile ABI is enabled (either via -fobjc-nonfragile-abi, or because it is the platform default). -fobjc-nonfragile-abi, -fno-objc-nonfragile-abi Enable use of the Objective-C non-fragile ABI. On platforms for which this is the default ABI, it can be disabled with -fno-objc-nonfragile-abi. Target Selection Options Clang fully supports cross compilation as an inherent part of its design. Depending on how your version of Clang is configured, it may have support for a number of cross compilers, or may only support a native target. -arch <architecture> Specify the architecture to build for (Mac OS X specific). -target <architecture> Specify the architecture to build for (all platforms). -mmacosx-version-min=<version> When building for macOS, specify the minimum version supported by your application. -miphoneos-version-min When building for iPhone OS, specify the minimum version supported by your application. --print-supported-cpus Print out a list of supported processors for the given target (specified through --target=<architecture> or -arch <architecture>). If no target is specified, the system default target will be used. -mcpu=?, -mtune=? Acts as an alias for --print-supported-cpus. -march=<cpu> Specify that Clang should generate code for a specific processor family member and later. For example, if you specify -march=i486, the compiler is allowed to generate instructions that are valid on i486 and later processors, but which may not exist on earlier ones. Code Generation Options -O0, -O1, -O2, -O3, -Ofast, -Os, -Oz, -Og, -O, -O4 Specify which optimization level to use: -O0 Means "no optimization": this level compiles the fastest and generates the most debuggable code. -O1 Somewhere between -O0 and -O2. -O2 Moderate level of optimization which enables most optimizations. -O3 Like -O2, except that it enables optimizations that take longer to perform or that may generate larger code (in an attempt to make the program run faster). -Ofast Enables all the optimizations from -O3 along with other aggressive optimizations that may violate strict compliance with language standards. -Os Like -O2 with extra optimizations to reduce code size. -Oz Like -Os (and thus -O2), but reduces code size further. -Og Like -O1. In future versions, this option might disable different optimizations in order to improve debuggability. -O Equivalent to -O1. -O4 and higher Currently equivalent to -O3 -g, -gline-tables-only, -gmodules Control debug information output. Note that Clang debug information works best at -O0. When more than one option starting with -g is specified, the last one wins: -g Generate debug information. -gline-tables-only Generate only line table debug information. This allows for symbolicated backtraces with inlining information, but does not include any information about variables, their locations or types. -gmodules Generate debug information that contains external references to types defined in Clang modules or precompiled headers instead of emitting redundant debug type information into every object file. This option transparently switches the Clang module format to object file containers that hold the Clang module together with the debug information. When compiling a program that uses Clang modules or precompiled headers, this option produces complete debug information with faster compile times and much smaller object files. This option should not be used when building static libraries for distribution to other machines because the debug info will contain references to the module cache on the machine the object files in the library were built on. -fstandalone-debug -fno-standalone-debug Clang supports a number of optimizations to reduce the size of debug information in the binary. They work based on the assumption that the debug type information can be spread out over multiple compilation units. For instance, Clang will not emit type definitions for types that are not needed by a module and could be replaced with a forward declaration. Further, Clang will only emit type info for a dynamic C++ class in the module that contains the vtable for the class. The -fstandalone-debug option turns off these optimizations. This is useful when working with 3rd-party libraries that don't come with debug information. This is the default on Darwin. Note that Clang will never emit type information for types that are not referenced at all by the program. -feliminate-unused-debug-types By default, Clang does not emit type information for types that are defined but not used in a program. To retain the debug info for these unused types, the negation -fno-eliminate-unused-debug-types can be used. -fexceptions Enable generation of unwind information. This allows exceptions to be thrown through Clang compiled stack frames. This is on by default in x86-64. -ftrapv Generate code to catch integer overflow errors. Signed integer overflow is undefined in C. With this flag, extra code is generated to detect this and abort when it happens. -fvisibility This flag sets the default visibility level. -fcommon, -fno-common This flag specifies that variables without initializers get common linkage. It can be disabled with -fno-common. -ftls-model=<model> Set the default thread-local storage (TLS) model to use for thread-local variables. Valid values are: "global-dynamic", "local-dynamic", "initial-exec" and "local-exec". The default is "global-dynamic". The default model can be overridden with the tls_model attribute. The compiler will try to choose a more efficient model if possible. -flto, -flto=full, -flto=thin, -emit-llvm Generate output files in LLVM formats, suitable for link time optimization. When used with -S this generates LLVM intermediate language assembly files, otherwise this generates LLVM bitcode format object files (which may be passed to the linker depending on the stage selection options). The default for -flto is "full", in which the LLVM bitcode is suitable for monolithic Link Time Optimization (LTO), where the linker merges all such modules into a single combined module for optimization. With "thin", ThinLTO compilation is invoked instead. NOTE: On Darwin, when using -flto along with -g and compiling and linking in separate steps, you also need to pass -Wl,-object_path_lto,<lto-filename>.o at the linking step to instruct the ld64 linker not to delete the temporary object file generated during Link Time Optimization (this flag is automatically passed to the linker by Clang if compilation and linking are done in a single step). This allows debugging the executable as well as generating the .dSYM bundle using dsymutil(1). Driver Options -### Print (but do not run) the commands to run for this compilation. --help Display available options. -Qunused-arguments Do not emit any warnings for unused driver arguments. -Wa,<args> Pass the comma separated arguments in args to the assembler. -Wl,<args> Pass the comma separated arguments in args to the linker. -Wp,<args> Pass the comma separated arguments in args to the preprocessor. -Xanalyzer <arg> Pass arg to the static analyzer. -Xassembler <arg> Pass arg to the assembler. -Xlinker <arg> Pass arg to the linker. -Xpreprocessor <arg> Pass arg to the preprocessor. -o <file> Write output to file. -print-file-name=<file> Print the full library path of file. -print-libgcc-file-name Print the library path for the currently used compiler runtime library ("libgcc.a" or "libclang_rt.builtins.*.a"). -print-prog-name=<name> Print the full program path of name. -print-search-dirs Print the paths used for finding libraries and programs. -save-temps Save intermediate compilation results. -save-stats, -save-stats=cwd, -save-stats=obj Save internal code generation (LLVM) statistics to a file in the current directory (-save-stats/"-save-stats=cwd") or the directory of the output file ("-save-state=obj"). -integrated-as, -no-integrated-as Used to enable and disable, respectively, the use of the integrated assembler. Whether the integrated assembler is on by default is target dependent. -time Time individual commands. -ftime-report Print timing summary of each stage of compilation. -v Show commands to run and use verbose output. Diagnostics Options -fshow-column, -fshow-source-location, -fcaret-diagnostics, -fdiagnostics-fixit-info, -fdiagnostics-parseable-fixits, -fdiagnostics-print-source-range-info, -fprint-source-range-info, -fdiagnostics-show-option, -fmessage-length These options control how Clang prints out information about diagnostics (errors and warnings). Please see the Clang User's Manual for more information. Preprocessor Options -D<macroname>=<value> Adds an implicit #define into the predefines buffer which is read before the source file is preprocessed. -U<macroname> Adds an implicit #undef into the predefines buffer which is read before the source file is preprocessed. -include <filename> Adds an implicit #include into the predefines buffer which is read before the source file is preprocessed. -I<directory> Add the specified directory to the search path for include files. -F<directory> Add the specified directory to the search path for framework include files. -nostdinc Do not search the standard system directories or compiler builtin directories for include files. -nostdlibinc Do not search the standard system directories for include files, but do search compiler builtin include directories. -nobuiltininc Do not search clang's builtin directory for include files. ENVIRONMENT TMPDIR, TEMP, TMP These environment variables are checked, in order, for the location to write temporary files used during the compilation process. CPATH If this environment variable is present, it is treated as a delimited list of paths to be added to the default system include path list. The delimiter is the platform dependent delimiter, as used in the PATH environment variable. Empty components in the environment variable are ignored. C_INCLUDE_PATH, OBJC_INCLUDE_PATH, CPLUS_INCLUDE_PATH, OBJCPLUS_INCLUDE_PATH These environment variables specify additional paths, as for CPATH, which are only used when processing the appropriate language. MACOSX_DEPLOYMENT_TARGET If -mmacosx-version-min is unspecified, the default deployment target is read from this environment variable. This option only affects Darwin targets. BUGS To report bugs, please visit <https://github.com/llvm/llvm-project/issues/>. Most bug reports should include preprocessed source files (use the -E option) and the full output of the compiler, along with information to reproduce. SEE ALSO as(1), ld(1) AUTHOR Maintained by the Clang / LLVM Team (<http://clang.llvm.org>) COPYRIGHT 2007-2024, The Clang Team 11 January 28, 2024 CLANG(1)
| null |
pod2readme
|
This utility will use Pod::Readme to extract a README file from a POD document. It works by extracting and filtering the POD, and then calling the appropriate filter program to convert the POD to another format.
|
pod2readme - Intelligently generate a README file from POD USAGE pod2readme [-cfho] [long options...] input-file [output-file] [target] Intelligently generate a README file from POD -t --target target type (default: 'readme') -f --format output format (default: 'text') -b --backup backup output file -o --output output filename (default based on target) -c --stdout output to stdout (console) -F --force only update if files are changed -h --help print usage and exit
|
pod2readme -f markdown lib/MyApp.pm
|
"--backup" By default, "pod2readme" will back up the output file. To disable this, use the "--no-backup" option. "--output" Specifies the name of the output file. If omitted, it will use the second command line argument, or default to the "--target" plus the corresponding extension of the "--format". For all intents, the default is README. If a format other than "text" is chosen, then the appropriate extension will be added, e.g. for "markdown", the default output file is README.md. "--target" The target of the filter, which defaults to "readme". "--format" The output format, which defaults to "text". Other supposed formats are "github", "html", "latex", "man", "markdown", "pod", "rtf", and "xhtml". You can also use "gfm" instead of "github". Similary you can use "md" for "markdown". "--stdout" If enabled, it will output to the console instead of "--output". "--force" By default, the README will be generated if the source files have been changed. Using "--force" will force the file to be updated. Note: POD format files will always be updated. "--help" Prints the usage and exits. SEE ALSO pod2text, pod2markdown. perl v5.34.0 2018-10-31 POD2README(1)
| null |
pp
|
pp creates standalone executables from Perl programs, using the compressed packager provided by PAR, and dependency detection heuristics offered by Module::ScanDeps. Source files are compressed verbatim without compilation. You may think of pp as "perlcc that works without hassle". :-) A GUI interface is also available as the tkpp command. It does not provide the compilation-step acceleration provided by perlcc (however, see -f below for byte-compiled, source-hiding techniques), but makes up for it with better reliability, smaller executable size, and full retrieval of original source code. When a single input program is specified, the resulting executable will behave identically as that program. However, when multiple programs are packaged, the produced executable will run the one that has the same basename as $0 (i.e. the filename used to invoke it). If nothing matches, it dies with the error "Can't open perl script "$0"".
|
pp - PAR Packager
|
pp [ -ABCEFILMPTSVXacdefghilmnoprsuvxz ] [ parfile | scriptfile ]...
|
Options are available in a short form and a long form. For example, the three lines below are all equivalent: % pp -o output.exe input.pl % pp --output output.exe input.pl % pp --output=output.exe input.pl Since the command lines can become sufficiently long to reach the limits imposed by some shells, it is possible to have pp read some of its options from one or more text files. The basic usage is to just include an argument starting with an 'at' (@) sigil. This argument will be interpreted as a file to read options from. Mixing ordinary options and @file options is possible. This is implemented using the Getopt::ArgvFile module, so read its documentation for advanced usage. -a, --addfile=FILE|DIR Add an extra file into the package. If the file is a directory, recursively add all files inside that directory, with links turned into actual files. By default, files are placed under "/" inside the package with their original names. You may override this by appending the target filename after a ";", like this: % pp -a "old_filename.txt;new_filename.txt" % pp -a "old_dirname;new_dirname" You may specify "-a" multiple times. -A, --addlist=FILE Read a list of file/directory names from FILE, adding them into the package. Each line in FILE is taken as an argument to -a above. You may specify "-A" multiple times. -B, --bundle Bundle core modules in the resulting package. This option is enabled by default, except when "-p" or "-P" is specified. Since PAR version 0.953, this also strips any local paths from the list of module search paths @INC before running the contained script. -C, --clean Clean up temporary files extracted from the application at runtime. By default, these files are cached in the temporary directory; this allows the program to start up faster next time. -c, --compile Run "perl -c inputfile" to determine additional run-time dependencies. -cd, --cachedeps=FILE Use FILE to cache detected dependencies. Creates FILE unless present. This will speed up the scanning process on subsequent runs. -d, --dependent Reduce the executable size by not including a copy of perl interpreter. Executables built this way will need a separate perl5x.dll or libperl.so to function correctly. This option is only available if perl is built as a shared library. -e, --eval=STRING Package a one-liner, much the same as "perl -e '...'" -E, --evalfeature=STRING Behaves just like "-e", except that it implicitly enables all optional features (in the main compilation unit) with Perl 5.10 and later. See feature. -x, --execute Run "perl inputfile" to determine additional run-time dependencies. Using this option, pp may be able to detect the use of modules that can't be determined by static analysis of "inputfile". Examples are stuff loaded by run-time loaders like Module::Runtime or "plugin" loaders like Module::Loader. Note that which modules are detected depends on which parts of your program are exercised when running "inputfile". E.g. if your program immediately terminates when run as "perl inputfile" because it lacks mandatory arguments, then this option will probably have no effect. You may use --xargs to supply arguments in this case. --xargs=STRING If -x is given, splits the "STRING" using the function "shellwords" from Text::ParseWords and passes the result as @ARGV when running "perl inputfile". -X, --exclude=MODULE Exclude the given module from the dependency search path and from the package. If the given file is a zip or par or par executable, all the files in the given file (except MANIFEST, META.yml and script/*) will be excluded and the output file will "use" the given file at runtime. -f, --filter=FILTER Filter source script(s) with a PAR::Filter subclass. You may specify multiple such filters. If you wish to hide the source code from casual prying, this will do: % pp -f Bleach source.pl If you are more serious about hiding your source code, you should have a look at Steve Hay's PAR::Filter::Crypto module. Make sure you understand the Filter::Crypto caveats! -g, --gui Build an executable that does not have a console window. This option is ignored on non-MSWin32 platforms or when "-p" is specified. -h, --help Show basic usage information. -I, --lib=DIR Add the given directory to the perl module search path. May be specified multiple times. -l, --link=FILE|LIBRARY Add the given shared library (a.k.a. shared object or DLL) into the packed file. Also accepts names under library paths; i.e. "-l ncurses" means the same thing as "-l libncurses.so" or "-l /usr/local/lib/libncurses.so" in most Unixes. May be specified multiple times. -L, --log=FILE Log the output of packaging to a file rather than to stdout. -F, --modfilter=FILTER[=REGEX], Filter included perl module(s) with a PAR::Filter subclass. You may specify multiple such filters. By default, the PodStrip filter is applied. In case that causes trouble, you can turn this off by setting the environment variable "PAR_VERBATIM" to 1. Since PAR 0.958, you can use an optional regular expression (REGEX above) to select the files in the archive which should be filtered. Example: pp -o foo.exe -F Bleach=warnings\.pm$ foo.pl This creates a binary executable foo.exe from foo.pl packaging all files as usual except for files ending in "warnings.pm" which are filtered with PAR::Filter::Bleach. -M, --module=MODULE Add the specified module into the package, along with its dependencies. The following variants may be used to add whole module namespaces: -M Foo::** Add every module in the "Foo" namespace except "Foo" itself, i.e. add "Foo::Bar", "Foo::Bar::Quux" etc up to any depth. -M Foo::* Add every module at level 1 in the "Foo" namespace, i.e. add "Foo::Bar", but neither "Foo::Bar::Quux" nor "Foo". -M Foo:: Shorthand for "-M Foo -M Foo:**": every module in the "Foo" namespace including "Foo" itself. Instead of a module name, MODULE may also be specified as a filename relative to the @INC path, i.e. "-M Module/ScanDeps.pm" means the same thing as "-M Module::ScanDeps". If MODULE has an extension that is not ".pm"/".ix"/".al", it will not be scanned for dependencies, and will be placed under "/" instead of "/lib/" inside the PAR file. This use is deprecated -- consider using the -a option instead. You may specify "-M" multiple times. -m, --multiarch Build a multi-architecture PAR file. Implies -p. -n, --noscan Skip the default static scanning altogether, using run-time dependencies from -c or -x exclusively. -N, --namespace=NAMESPACE Add all modules in the namespace into the package, along with their dependencies. If "NAMESPACE" is something like "Foo::Bar" then this will add all modules "Foo/Bar/Quux.pm", "Foo/Bar/Fred/Barnie.pm" etc that can be located in your module search path. It mimics the behaviour of "plugin" loaders like Module::Loader. This is different from using "-M Foo::Bar::", as the latter insists on adding "Foo/Bar.pm" which might not exist in the above "plugin" scenario. You may specify "-N" multiple times. -o, --output=FILE File name for the final packaged executable. -p, --par Create PAR archives only; do not package to a standalone binary. -P, --perlscript Create stand-alone perl script; do not package to a standalone binary. -r, --run Run the resulting packaged script after packaging it. --reusable EXPERIMENTAL Make the packaged executable reusable for running arbitrary, external Perl scripts as if they were part of the package: pp -o myapp --reusable someapp.pl ./myapp --par-options --reuse otherapp.pl The second line will run otherapp.pl instead of someapp.pl. -S, --save Do not delete generated PAR file after packaging. -s, --sign Cryptographically sign the generated PAR or binary file using Module::Signature. -T, --tempcache Set the program unique part of the cache directory name that is used if the program is run without -C. If not set, a hash of the executable is used. When the program is run, its contents are extracted to a temporary directory. On Unix systems, this is commonly /tmp/par-USER/cache-XXXXXXX. USER is replaced by the name of the user running the program, but "spelled" in hex. XXXXXXX is either a hash of the executable or the value passed to the "-T" or "--tempcache" switch. -u, --unicode Package Unicode support (essentially utf8_heavy.pl and everything below the directory unicore in your perl library). This option exists because it is impossible to detect using static analysis if your program needs Unicode support at runtime. (Note: If your program contains "use utf8" this does not imply it needs Unicode support. It merely says that your program is written in UTF-8.) If your packed program exits with an error message like Can't locate utf8_heavy.pl in @INC (@INC contains: ...) try to pack it with "-u" (or use "-x"). -v, --verbose[=NUMBER] Increase verbosity of output; NUMBER is an integer from 1 to 3, 3 being the most verbose. Defaults to 1 if specified without an argument. Alternatively, -vv sets verbose level to 2, and -vvv sets it to 3. -V, --version Display the version number and copyrights of this program. -z, --compress=NUMBER Set zip compression level; NUMBER is an integer from 0 to 9, 0 = no compression, 9 = max compression. Defaults to 6 if -z is not used. ENVIRONMENT PP_OPTS Command-line options (switches). Switches in this variable are taken as if they were on every pp command line. NOTES Here are some recipes showing how to utilize pp to bundle source.pl with all its dependencies, on target machines with different expected settings: Stone-alone setup: To make a stand-alone executable, suitable for running on a machine that doesn't have perl installed: % pp -o packed.exe source.pl # makes packed.exe # Now, deploy 'packed.exe' to target machine... $ packed.exe # run it Perl interpreter only, without core modules: To make a packed .pl file including core modules, suitable for running on a machine that has a perl interpreter, but where you want to be sure of the versions of the core modules that your program uses: % pp -B -P -o packed.pl source.pl # makes packed.pl # Now, deploy 'packed.pl' to target machine... $ perl packed.pl # run it Perl with core modules installed: To make a packed .pl file without core modules, relying on the target machine's perl interpreter and its core libraries. This produces a significantly smaller file than the previous version: % pp -P -o packed.pl source.pl # makes packed.pl # Now, deploy 'packed.pl' to target machine... $ perl packed.pl # run it Perl with PAR.pm and its dependencies installed: Make a separate archive and executable that uses the archive. This relies upon the perl interpreter and libraries on the target machine. % pp -p source.pl # makes source.par % echo "use PAR 'source.par';" > packed.pl; % cat source.pl >> packed.pl; # makes packed.pl # Now, deploy 'source.par' and 'packed.pl' to target machine... $ perl packed.pl # run it, perl + core modules required Note that even if your perl was built with a shared library, the 'Stand-alone executable' above will not need a separate perl5x.dll or libperl.so to function correctly. But even in this case, the underlying system libraries such as libc must be compatible between the host and target machines. Use "--dependent" if you are willing to ship the shared library with the application, which can significantly reduce the executable size. SEE ALSO tkpp, par.pl, parl, perlcc PAR, PAR::Packer, Module::ScanDeps Getopt::Long, Getopt::ArgvFile ACKNOWLEDGMENTS Simon Cozens, Tom Christiansen and Edward Peschko for writing perlcc; this program try to mimic its interface as close as possible, and copied liberally from their code. Jan Dubois for writing the exetype.pl utility, which has been partially adapted into the "-g" flag. Mattia Barbon for providing the "myldr" binary loader code. Jeff Goff for suggesting the name pp. AUTHORS Audrey Tang <cpan@audreyt.org>, Steffen Mueller <smueller@cpan.org> Roderich Schupp <rschupp@cpan.org> You can write to the mailing list at <par@perl.org>, or send an empty mail to <par-subscribe@perl.org> to participate in the discussion. Please submit bug reports to <bug-par-packer@rt.cpan.org>. COPYRIGHT Copyright 2002-2009 by Audrey Tang <cpan@audreyt.org>. Neither this program nor the associated parl program impose any licensing restrictions on files generated by their execution, in accordance with the 8th article of the Artistic License: "Aggregation of this Package with a commercial distribution is always permitted provided that the use of this Package is embedded; that is, when no overt attempt is made to make this Package's interfaces visible to the end user of the commercial distribution. Such use shall not be construed as a distribution of this Package." Therefore, you are absolutely free to place any license on the resulting executable, as long as the packed 3rd-party libraries are also available under the Artistic License. This program is free software; you can redistribute it and/or modify it under the same terms as Perl itself. See LICENSE. perl v5.34.0 2020-03-08 PP(1)
|
Note: When running on Microsoft Windows, the a.out below will be replaced by a.exe instead. % pp hello.pl # Pack 'hello.pl' into executable 'a.out' % pp -o hello hello.pl # Pack 'hello.pl' into executable 'hello' # (or 'hello.exe' on Win32) % pp -o foo foo.pl bar.pl # Pack 'foo.pl' and 'bar.pl' into 'foo' % ./foo # Run 'foo.pl' inside 'foo' % mv foo bar; ./bar # Run 'bar.pl' inside 'foo' % mv bar baz; ./baz # Error: Can't open perl script "baz" % pp -p file # Creates a PAR file, 'a.par' % pp -o hello a.par # Pack 'a.par' to executable 'hello' % pp -S -o hello file # Combine the two steps above % pp -p -o out.par file # Creates 'out.par' from 'file' % pp -B -p -o out.par file # same as above, but bundles core modules # and removes any local paths from @INC % pp -P -o out.pl file # Creates 'out.pl' from 'file' % pp -B -p -o out.pl file # same as above, but bundles core modules # and removes any local paths from @INC # (-B is assumed when making executables) % pp -e "print 123" # Pack a one-liner into 'a.out' % pp -p -e "print 123" # Creates a PAR file 'a.par' % pp -P -e "print 123" # Creates a perl script 'a.pl' % pp -c hello # Check dependencies from "perl -c hello" % pp -x hello # Check dependencies from "perl hello" % pp -n -x hello # same as above, but skips static scanning % pp -I /foo hello # Extra include paths % pp -M Foo::Bar hello # Extra modules in the include path % pp -M abbrev.pl hello # Extra libraries in the include path % pp -X Foo::Bar hello # Exclude modules % pp -a data.txt hello # Additional data files % pp -r hello # Pack 'hello' into 'a.out', runs 'a.out' % pp -r hello a b c # Pack 'hello' into 'a.out', runs 'a.out' # with arguments 'a b c' % pp hello --log=c # Pack 'hello' into 'a.out', logs # messages into 'c' # Pack 'hello' into a console-less 'out.exe' (Win32 only) % pp --gui -o out.exe hello % pp @file hello.pl # Pack 'hello.pl' but read _additional_ # options from file 'file'
|
vi
|
Vim is a text editor that is upwards compatible to Vi. It can be used to edit all kinds of plain text. It is especially useful for editing programs. There are a lot of enhancements above Vi: multi level undo, multi windows and buffers, syntax highlighting, command line editing, filename completion, on-line help, visual selection, etc.. See ":help vi_diff.txt" for a summary of the differences between Vim and Vi. While running Vim a lot of help can be obtained from the on-line help system, with the ":help" command. See the ON-LINE HELP section below. Most often Vim is started to edit a single file with the command vim file More generally Vim is started with: vim [options] [filelist] If the filelist is missing, the editor will start with an empty buffer. Otherwise exactly one out of the following four may be used to choose one or more files to be edited. file .. A list of filenames. The first one will be the current file and read into the buffer. The cursor will be positioned on the first line of the buffer. You can get to the other files with the ":next" command. To edit a file that starts with a dash, precede the filelist with "--". - The file to edit is read from stdin. Commands are read from stderr, which should be a tty. -t {tag} The file to edit and the initial cursor position depends on a "tag", a sort of goto label. {tag} is looked up in the tags file, the associated file becomes the current file and the associated command is executed. Mostly this is used for C programs, in which case {tag} could be a function name. The effect is that the file containing that function becomes the current file and the cursor is positioned on the start of the function. See ":help tag-commands". -q [errorfile] Start in quickFix mode. The file [errorfile] is read and the first error is displayed. If [errorfile] is omitted, the filename is obtained from the 'errorfile' option (defaults to "AztecC.Err" for the Amiga, "errors.err" on other systems). Further errors can be jumped to with the ":cn" command. See ":help quickfix". Vim behaves differently, depending on the name of the command (the executable may still be the same file). vim The "normal" way, everything is default. ex Start in Ex mode. Go to Normal mode with the ":vi" command. Can also be done with the "-e" argument. view Start in read-only mode. You will be protected from writing the files. Can also be done with the "-R" argument. gvim gview The GUI version. Starts a new window. Can also be done with the "-g" argument. evim eview The GUI version in easy mode. Starts a new window. Can also be done with the "-y" argument. rvim rview rgvim rgview Like the above, but with restrictions. It will not be possible to start shell commands, or suspend Vim. Can also be done with the "-Z" argument.
|
vim - Vi IMproved, a programmer's text editor
|
vim [options] [file ..] vim [options] - vim [options] -t tag vim [options] -q [errorfile] ex view gvim gview evim eview rvim rview rgvim rgview
|
The options may be given in any order, before or after filenames. Options without an argument can be combined after a single dash. +[num] For the first file the cursor will be positioned on line "num". If "num" is missing, the cursor will be positioned on the last line. +/{pat} For the first file the cursor will be positioned in the line with the first occurrence of {pat}. See ":help search-pattern" for the available search patterns. +{command} -c {command} {command} will be executed after the first file has been read. {command} is interpreted as an Ex command. If the {command} contains spaces it must be enclosed in double quotes (this depends on the shell that is used). Example: vim "+set si" main.c Note: You can use up to 10 "+" or "-c" commands. -S {file} {file} will be sourced after the first file has been read. This is equivalent to -c "source {file}". {file} cannot start with '-'. If {file} is omitted "Session.vim" is used (only works when -S is the last argument). --cmd {command} Like using "-c", but the command is executed just before processing any vimrc file. You can use up to 10 of these commands, independently from "-c" commands. -A If Vim has been compiled with ARABIC support for editing right-to-left oriented files and Arabic keyboard mapping, this option starts Vim in Arabic mode, i.e. 'arabic' is set. Otherwise an error message is given and Vim aborts. -b Binary mode. A few options will be set that makes it possible to edit a binary or executable file. -C Compatible. Set the 'compatible' option. This will make Vim behave mostly like Vi, even though a .vimrc file exists. -d Start in diff mode. There should between two to eight file name arguments. Vim will open all the files and show differences between them. Works like vimdiff(1). -d {device} Open {device} for use as a terminal. Only on the Amiga. Example: "-d con:20/30/600/150". -D Debugging. Go to debugging mode when executing the first command from a script. -e Start Vim in Ex mode, just like the executable was called "ex". -E Start Vim in improved Ex mode, just like the executable was called "exim". -f Foreground. For the GUI version, Vim will not fork and detach from the shell it was started in. On the Amiga, Vim is not restarted to open a new window. This option should be used when Vim is executed by a program that will wait for the edit session to finish (e.g. mail). On the Amiga the ":sh" and ":!" commands will not work. --nofork Foreground. For the GUI version, Vim will not fork and detach from the shell it was started in. -F If Vim has been compiled with FKMAP support for editing right-to-left oriented files and Farsi keyboard mapping, this option starts Vim in Farsi mode, i.e. 'fkmap' and 'rightleft' are set. Otherwise an error message is given and Vim aborts. -g If Vim has been compiled with GUI support, this option enables the GUI. If no GUI support was compiled in, an error message is given and Vim aborts. -h Give a bit of help about the command line arguments and options. After this Vim exits. -H If Vim has been compiled with RIGHTLEFT support for editing right-to-left oriented files and Hebrew keyboard mapping, this option starts Vim in Hebrew mode, i.e. 'hkmap' and 'rightleft' are set. Otherwise an error message is given and Vim aborts. -i {viminfo} Specifies the filename to use when reading or writing the viminfo file, instead of the default "~/.viminfo". This can also be used to skip the use of the .viminfo file, by giving the name "NONE". -L Same as -r. -l Lisp mode. Sets the 'lisp' and 'showmatch' options on. -m Modifying files is disabled. Resets the 'write' option. You can still modify the buffer, but writing a file is not possible. -M Modifications not allowed. The 'modifiable' and 'write' options will be unset, so that changes are not allowed and files can not be written. Note that these options can be set to enable making modifications. -N No-compatible mode. Resets the 'compatible' option. This will make Vim behave a bit better, but less Vi compatible, even though a .vimrc file does not exist. -n No swap file will be used. Recovery after a crash will be impossible. Handy if you want to edit a file on a very slow medium (e.g. floppy). Can also be done with ":set uc=0". Can be undone with ":set uc=200". -nb Become an editor server for NetBeans. See the docs for details. -o[N] Open N windows stacked. When N is omitted, open one window for each file. -O[N] Open N windows side by side. When N is omitted, open one window for each file. -p[N] Open N tab pages. When N is omitted, open one tab page for each file. -R Read-only mode. The 'readonly' option will be set. You can still edit the buffer, but will be prevented from accidentally overwriting a file. If you do want to overwrite a file, add an exclamation mark to the Ex command, as in ":w!". The -R option also implies the -n option (see above). The 'readonly' option can be reset with ":set noro". See ":help 'readonly'". -r List swap files, with information about using them for recovery. -r {file} Recovery mode. The swap file is used to recover a crashed editing session. The swap file is a file with the same filename as the text file with ".swp" appended. See ":help recovery". -s Silent mode. Only when started as "Ex" or when the "-e" option was given before the "-s" option. -s {scriptin} The script file {scriptin} is read. The characters in the file are interpreted as if you had typed them. The same can be done with the command ":source! {scriptin}". If the end of the file is reached before the editor exits, further characters are read from the keyboard. -T {terminal} Tells Vim the name of the terminal you are using. Only required when the automatic way doesn't work. Should be a terminal known to Vim (builtin) or defined in the termcap or terminfo file. -u {vimrc} Use the commands in the file {vimrc} for initializations. All the other initializations are skipped. Use this to edit a special kind of files. It can also be used to skip all initializations by giving the name "NONE". See ":help initialization" within vim for more details. -U {gvimrc} Use the commands in the file {gvimrc} for GUI initializations. All the other GUI initializations are skipped. It can also be used to skip all GUI initializations by giving the name "NONE". See ":help gui-init" within vim for more details. -V[N] Verbose. Give messages about which files are sourced and for reading and writing a viminfo file. The optional number N is the value for 'verbose'. Default is 10. -v Start Vim in Vi mode, just like the executable was called "vi". This only has effect when the executable is called "ex". -w {scriptout} All the characters that you type are recorded in the file {scriptout}, until you exit Vim. This is useful if you want to create a script file to be used with "vim -s" or ":source!". If the {scriptout} file exists, characters are appended. -W {scriptout} Like -w, but an existing file is overwritten. -x Use encryption when writing files. Will prompt for a crypt key. -X Don't connect to the X server. Shortens startup time in a terminal, but the window title and clipboard will not be used. -y Start Vim in easy mode, just like the executable was called "evim" or "eview". Makes Vim behave like a click-and-type editor. -Z Restricted mode. Works like the executable starts with "r". -- Denotes the end of the options. Arguments after this will be handled as a file name. This can be used to edit a filename that starts with a '-'. --clean Do not use any personal configuration (vimrc, plugins, etc.). Useful to see if a problem reproduces with a clean Vim setup. --echo-wid GTK GUI only: Echo the Window ID on stdout. --help Give a help message and exit, just like "-h". --literal Take file name arguments literally, do not expand wildcards. This has no effect on Unix where the shell expands wildcards. --noplugin Skip loading plugins. Implied by -u NONE. --remote Connect to a Vim server and make it edit the files given in the rest of the arguments. If no server is found a warning is given and the files are edited in the current Vim. --remote-expr {expr} Connect to a Vim server, evaluate {expr} in it and print the result on stdout. --remote-send {keys} Connect to a Vim server and send {keys} to it. --remote-silent As --remote, but without the warning when no server is found. --remote-wait As --remote, but Vim does not exit until the files have been edited. --remote-wait-silent As --remote-wait, but without the warning when no server is found. --serverlist List the names of all Vim servers that can be found. --servername {name} Use {name} as the server name. Used for the current Vim, unless used with a --remote argument, then it's the name of the server to connect to. --socketid {id} GTK GUI only: Use the GtkPlug mechanism to run gvim in another window. --startuptime {file} During startup write timing messages to the file {fname}. --version Print version information and exit. ON-LINE HELP Type ":help" in Vim to get started. Type ":help subject" to get help on a specific subject. For example: ":help ZZ" to get help for the "ZZ" command. Use <Tab> and CTRL-D to complete subjects (":help cmdline-completion"). Tags are present to jump from one place to another (sort of hypertext links, see ":help"). All documentation files can be viewed in this way, for example ":help syntax.txt". FILES /usr/local/lib/vim/doc/*.txt The Vim documentation files. Use ":help doc-file-list" to get the complete list. /usr/local/lib/vim/doc/tags The tags file used for finding information in the documentation files. /usr/local/lib/vim/syntax/syntax.vim System wide syntax initializations. /usr/local/lib/vim/syntax/*.vim Syntax files for various languages. /usr/local/lib/vim/vimrc System wide Vim initializations. ~/.vimrc Your personal Vim initializations. /usr/local/lib/vim/gvimrc System wide gvim initializations. ~/.gvimrc Your personal gvim initializations. /usr/local/lib/vim/optwin.vim Script used for the ":options" command, a nice way to view and set options. /usr/local/lib/vim/menu.vim System wide menu initializations for gvim. /usr/local/lib/vim/bugreport.vim Script to generate a bug report. See ":help bugs". /usr/local/lib/vim/filetype.vim Script to detect the type of a file by its name. See ":help 'filetype'". /usr/local/lib/vim/scripts.vim Script to detect the type of a file by its contents. See ":help 'filetype'". /usr/local/lib/vim/print/*.ps Files used for PostScript printing. For recent info read the VIM home page: <URL:http://www.vim.org/> SEE ALSO vimtutor(1) AUTHOR Most of Vim was made by Bram Moolenaar, with a lot of help from others. See ":help credits" in Vim. Vim is based on Stevie, worked on by: Tim Thompson, Tony Andrews and G.R. (Fred) Walter. Although hardly any of the original code remains. BUGS Probably. See ":help todo" for a list of known problems. Note that a number of things that may be regarded as bugs by some, are in fact caused by a too-faithful reproduction of Vi's behaviour. And if you think other things are bugs "because Vi does it differently", you should take a closer look at the vi_diff.txt file (or type :help vi_diff.txt when in Vim). Also have a look at the 'compatible' and 'cpoptions' options. 2021 Jun 13 VIM(1)
| null |
iofileb.d
|
This prints a summary of requested disk activity by pathname, providing totals of the I/O events in bytes. It is a companion to the iofile.d script - which prints in terms of I/O wait time, not bytes. I/O wait time is a better metric for understanding performance issues. Both disk and NFS I/O are measured. Since this uses DTrace, only users with root privileges can run this command.
|
iofileb.d - I/O bytes by file and process. Uses DTrace.
|
iofileb.d
| null |
Sample until Ctrl-C is hit then print report, # iofileb.d FIELDS PID process ID CMD process name KB kilobytes of disk I/O FILE file pathname DOCUMENTATION See the DTraceToolkit for further documentation under the Docs directory. The DTraceToolkit docs may include full worked examples with verbose descriptions explaining the output. EXIT iofileb.d will sample until Ctrl-C is hit. SEE ALSO iofile.d(1M), iosnoop(1M), dtrace(1M) version 1.00 February 20, 2006 iofileb.d(1m)
|
errinfo
|
errinfo snoops syscall failures and prints the errno value and description of the error number. This program can help determine if applications are silently failing, providing some details on the cause. Since this uses DTrace, only users with root privileges can run this command.
|
errinfo - print errno for syscall fails. Uses DTrace.
|
errinfo [-a|-A|-hsvZ] [-c command]
|
-c counts - print an aggregate style report containing a frequency count of errors -p PID examine this PID only -n name examine processes with ths name only (eg, "ls")
|
Default output, print errors as they occur, # errinfo Print a frequency count report, # errinfo -c Snoop errors as they occur for "ssh" processes, # errinfo -n ssh PP Snoop errors for PID 81 only, # errinfo -p 81 FIELDS EXEC Program name (truncated) SYSCALL System call name ERR Value of errno DESC Description of errno message DOCUMENTATION See the DTraceToolkit for further documentation under the Docs directory. The DTraceToolkit docs may include full worked examples with verbose descriptions explaining the output. EXIT errinfo will run forever until Ctrl-C is hit. FILES /usr/include/sys/errno.h Contains the full descriptions for the error numbers. AUTHOR Brendan Gregg [Sydney, Australia] SEE ALSO dtrace(1M), truss(1) version 1.10 May 14, 2005 errinfo(1m)
|
zipgrep
|
zipgrep will search files within a ZIP archive for lines matching the given string or pattern. zipgrep is a shell script and requires egrep(1) and unzip(1L) to function. Its output is identical to that of egrep(1). ARGUMENTS pattern The pattern to be located within a ZIP archive. Any string or regular expression accepted by egrep(1) may be used. file[.zip] Path of the ZIP archive. (Wildcard expressions for the ZIP archive name are not supported.) If the literal filename is not found, the suffix .zip is appended. Note that self-extracting ZIP files are supported, as with any other ZIP archive; just specify the .exe suffix (if any) explicitly. [file(s)] An optional list of archive members to be processed, separated by spaces. If no member files are specified, all members of the ZIP archive are searched. Regular expressions (wildcards) may be used to match multiple members: * matches a sequence of 0 or more characters ? matches exactly 1 character [...] matches any single character found inside the brackets; ranges are specified by a beginning character, a hyphen, and an ending character. If an exclamation point or a caret (`!' or `^') follows the left bracket, then the range of characters within the brackets is complemented (that is, anything except the characters inside the brackets is considered a match). (Be sure to quote any character that might otherwise be interpreted or modified by the operating system.) [-x xfile(s)] An optional list of archive members to be excluded from processing. Since wildcard characters match directory separators (`/'), this option may be used to exclude any files that are in subdirectories. For example, ``zipgrep grumpy foo *.[ch] -x */*'' would search for the string ``grumpy'' in all C source files in the main directory of the ``foo'' archive, but none in any subdirectories. Without the -x option, all C source files in all directories within the zipfile would be searched.
|
zipgrep - search files in a ZIP archive for lines matching a pattern
|
zipgrep [egrep_options] pattern file[.zip] [file(s) ...] [-x xfile(s) ...]
|
All options prior to the ZIP archive filename are passed to egrep(1). SEE ALSO egrep(1), unzip(1L), zip(1L), funzip(1L), zipcloak(1L), zipinfo(1L), zipnote(1L), zipsplit(1L) URL The Info-ZIP home page is currently at http://www.info-zip.org/pub/infozip/ or ftp://ftp.info-zip.org/pub/infozip/ . AUTHORS zipgrep was written by Jean-loup Gailly. Info-ZIP 20 April 2009 ZIPGREP(1L)
| null |
pod2usage
|
pod2usage will read the given input file looking for pod documentation and will print the corresponding usage message. If no input file is specified then standard input is read. pod2usage invokes the pod2usage() function in the Pod::Usage module. Please see "pod2usage()" in Pod::Usage. SEE ALSO Pod::Usage, pod2text, Pod::Text, Pod::Text::Termcap, perldoc AUTHOR Please report bugs using <http://rt.cpan.org>. Brad Appleton <bradapp@enteract.com> Based on code for pod2text(1) written by Tom Christiansen <tchrist@mox.perl.com> perl v5.38.2 2023-11-28 POD2USAGE(1)
|
pod2usage - print usage messages from embedded pod docs in files
|
pod2usage [-help] [-man] [-exit exitval] [-output outfile] [-verbose level] [-pathlist dirlist] [-formatter module] [-utf8] file OPTIONS AND ARGUMENTS -help Print a brief help message and exit. -man Print this command's manual page and exit. -exit exitval The exit status value to return. -output outfile The output file to print to. If the special names "-" or ">&1" or ">&STDOUT" are used then standard output is used. If ">&2" or ">&STDERR" is used then standard error is used. -verbose level The desired level of verbosity to use: 1 : print SYNOPSIS only 2 : print SYNOPSIS sections and any OPTIONS/ARGUMENTS sections 3 : print the entire manpage (similar to running pod2text) -pathlist dirlist Specifies one or more directories to search for the input file if it was not supplied with an absolute path. Each directory path in the given list should be separated by a ':' on Unix (';' on MSWin32 and DOS). -formatter module Which text formatter to use. Default is Pod::Text, or for very old Perl versions Pod::PlainText. An alternative would be e.g. Pod::Text::Termcap. -utf8 This option assumes that the formatter (see above) understands the option "utf8". It turns on generation of utf8 output. file The pathname of a file containing pod documentation to be output in usage message format. If omitted, standard input is read - but the output is then formatted with Pod::Text only - unless a specific formatter has been specified with -formatter.
| null | null |
alias
|
Shell builtin commands are commands that can be executed within the running shell's process. Note that, in the case of csh(1) builtin commands, the command is executed in a subshell if it occurs as any component of a pipeline except the last. If a command specified to the shell contains a slash ‘/’, the shell will not execute a builtin command, even if the last component of the specified command matches the name of a builtin command. Thus, while specifying “echo” causes a builtin command to be executed under shells that support the echo builtin command, specifying “/bin/echo” or “./echo” does not. While some builtin commands may exist in more than one shell, their operation may be different under each shell which supports them. Below is a table which lists shell builtin commands, the standard shells that support them and whether they exist as standalone utilities. Only builtin commands for the csh(1) and sh(1) shells are listed here. Consult a shell's manual page for details on the operation of its builtin commands. Beware that the sh(1) manual page, at least, calls some of these commands “built-in commands” and some of them “reserved words”. Users of other shells may need to consult an info(1) page or other sources of documentation. Commands marked “No**” under External do exist externally, but are implemented as scripts using a builtin command of the same name. Command External csh(1) sh(1) ! No No Yes % No Yes No . No No Yes : No Yes Yes @ No Yes Yes [ Yes No Yes { No No Yes } No No Yes alias No** Yes Yes alloc No Yes No bg No** Yes Yes bind No No Yes bindkey No Yes No break No Yes Yes breaksw No Yes No builtin No No Yes builtins No Yes No case No Yes Yes cd No** Yes Yes chdir No Yes Yes command No** No Yes complete No Yes No continue No Yes Yes default No Yes No dirs No Yes No do No No Yes done No No Yes echo Yes Yes Yes echotc No Yes No elif No No Yes else No Yes Yes end No Yes No endif No Yes No endsw No Yes No esac No No Yes eval No Yes Yes exec No Yes Yes exit No Yes Yes export No No Yes false Yes No Yes fc No** No Yes fg No** Yes Yes filetest No Yes No fi No No Yes for No No Yes foreach No Yes No getopts No** No Yes glob No Yes No goto No Yes No hash No** No Yes hashstat No Yes No history No Yes No hup No Yes No if No Yes Yes jobid No No Yes jobs No** Yes Yes kill Yes Yes Yes limit No Yes No local No No Yes log No Yes No login Yes Yes No logout No Yes No ls-F No Yes No nice Yes Yes No nohup Yes Yes No notify No Yes No onintr No Yes No popd No Yes No printenv Yes Yes No printf Yes No Yes pushd No Yes No pwd Yes No Yes read No** No Yes readonly No No Yes rehash No Yes No repeat No Yes No return No No Yes sched No Yes No set No Yes Yes setenv No Yes No settc No Yes No setty No Yes No setvar No No Yes shift No Yes Yes source No Yes No stop No Yes No suspend No Yes No switch No Yes No telltc No Yes No test Yes No Yes then No No Yes time Yes Yes No times No No Yes trap No No Yes true Yes No Yes type No** No Yes ulimit No** No Yes umask No** Yes Yes unalias No** Yes Yes uncomplete No Yes No unhash No Yes No unlimit No Yes No unset No Yes Yes unsetenv No Yes No until No No Yes wait No** Yes Yes where No Yes No which Yes Yes No while No Yes Yes SEE ALSO csh(1), dash(1), echo(1), false(1), info(1), kill(1), login(1), nice(1), nohup(1), printenv(1), printf(1), pwd(1), sh(1), test(1), time(1), true(1), which(1), zsh(1) HISTORY The builtin manual page first appeared in FreeBSD 3.4. AUTHORS This manual page was written by Sheldon Hearn <sheldonh@FreeBSD.org>. macOS 14.5 December 21, 2010 macOS 14.5
|
builtin, !, %, ., :, @, [, {, }, alias, alloc, bg, bind, bindkey, break, breaksw, builtins, case, cd, chdir, command, complete, continue, default, dirs, do, done, echo, echotc, elif, else, end, endif, endsw, esac, eval, exec, exit, export, false, fc, fg, filetest, fi, for, foreach, getopts, glob, goto, hash, hashstat, history, hup, if, jobid, jobs, kill, limit, local, log, login, logout, ls-F, nice, nohup, notify, onintr, popd, printenv, printf, pushd, pwd, read, readonly, rehash, repeat, return, sched, set, setenv, settc, setty, setvar, shift, source, stop, suspend, switch, telltc, test, then, time, times, trap, true, type, ulimit, umask, unalias, uncomplete, unhash, unlimit, unset, unsetenv, until, wait, where, which, while – shell built-in commands
|
See the built-in command description in the appropriate shell manual page.
| null | null |
wait
|
Shell builtin commands are commands that can be executed within the running shell's process. Note that, in the case of csh(1) builtin commands, the command is executed in a subshell if it occurs as any component of a pipeline except the last. If a command specified to the shell contains a slash ‘/’, the shell will not execute a builtin command, even if the last component of the specified command matches the name of a builtin command. Thus, while specifying “echo” causes a builtin command to be executed under shells that support the echo builtin command, specifying “/bin/echo” or “./echo” does not. While some builtin commands may exist in more than one shell, their operation may be different under each shell which supports them. Below is a table which lists shell builtin commands, the standard shells that support them and whether they exist as standalone utilities. Only builtin commands for the csh(1) and sh(1) shells are listed here. Consult a shell's manual page for details on the operation of its builtin commands. Beware that the sh(1) manual page, at least, calls some of these commands “built-in commands” and some of them “reserved words”. Users of other shells may need to consult an info(1) page or other sources of documentation. Commands marked “No**” under External do exist externally, but are implemented as scripts using a builtin command of the same name. Command External csh(1) sh(1) ! No No Yes % No Yes No . No No Yes : No Yes Yes @ No Yes Yes [ Yes No Yes { No No Yes } No No Yes alias No** Yes Yes alloc No Yes No bg No** Yes Yes bind No No Yes bindkey No Yes No break No Yes Yes breaksw No Yes No builtin No No Yes builtins No Yes No case No Yes Yes cd No** Yes Yes chdir No Yes Yes command No** No Yes complete No Yes No continue No Yes Yes default No Yes No dirs No Yes No do No No Yes done No No Yes echo Yes Yes Yes echotc No Yes No elif No No Yes else No Yes Yes end No Yes No endif No Yes No endsw No Yes No esac No No Yes eval No Yes Yes exec No Yes Yes exit No Yes Yes export No No Yes false Yes No Yes fc No** No Yes fg No** Yes Yes filetest No Yes No fi No No Yes for No No Yes foreach No Yes No getopts No** No Yes glob No Yes No goto No Yes No hash No** No Yes hashstat No Yes No history No Yes No hup No Yes No if No Yes Yes jobid No No Yes jobs No** Yes Yes kill Yes Yes Yes limit No Yes No local No No Yes log No Yes No login Yes Yes No logout No Yes No ls-F No Yes No nice Yes Yes No nohup Yes Yes No notify No Yes No onintr No Yes No popd No Yes No printenv Yes Yes No printf Yes No Yes pushd No Yes No pwd Yes No Yes read No** No Yes readonly No No Yes rehash No Yes No repeat No Yes No return No No Yes sched No Yes No set No Yes Yes setenv No Yes No settc No Yes No setty No Yes No setvar No No Yes shift No Yes Yes source No Yes No stop No Yes No suspend No Yes No switch No Yes No telltc No Yes No test Yes No Yes then No No Yes time Yes Yes No times No No Yes trap No No Yes true Yes No Yes type No** No Yes ulimit No** No Yes umask No** Yes Yes unalias No** Yes Yes uncomplete No Yes No unhash No Yes No unlimit No Yes No unset No Yes Yes unsetenv No Yes No until No No Yes wait No** Yes Yes where No Yes No which Yes Yes No while No Yes Yes SEE ALSO csh(1), dash(1), echo(1), false(1), info(1), kill(1), login(1), nice(1), nohup(1), printenv(1), printf(1), pwd(1), sh(1), test(1), time(1), true(1), which(1), zsh(1) HISTORY The builtin manual page first appeared in FreeBSD 3.4. AUTHORS This manual page was written by Sheldon Hearn <sheldonh@FreeBSD.org>. macOS 14.5 December 21, 2010 macOS 14.5
|
builtin, !, %, ., :, @, [, {, }, alias, alloc, bg, bind, bindkey, break, breaksw, builtins, case, cd, chdir, command, complete, continue, default, dirs, do, done, echo, echotc, elif, else, end, endif, endsw, esac, eval, exec, exit, export, false, fc, fg, filetest, fi, for, foreach, getopts, glob, goto, hash, hashstat, history, hup, if, jobid, jobs, kill, limit, local, log, login, logout, ls-F, nice, nohup, notify, onintr, popd, printenv, printf, pushd, pwd, read, readonly, rehash, repeat, return, sched, set, setenv, settc, setty, setvar, shift, source, stop, suspend, switch, telltc, test, then, time, times, trap, true, type, ulimit, umask, unalias, uncomplete, unhash, unlimit, unset, unsetenv, until, wait, where, which, while – shell built-in commands
|
See the built-in command description in the appropriate shell manual page.
| null | null |
safaridriver
| null | null | null | null | null |
bzdiff
|
Bzcmp and bzdiff are used to invoke the cmp or the diff program on bzip2 compressed files. All options specified are passed directly to cmp or diff. If only 1 file is specified, then the files compared are file1 and an uncompressed file1.bz2. If two files are specified, then they are uncompressed if necessary and fed to cmp or diff. The exit status from cmp or diff is preserved. SEE ALSO cmp(1), diff(1), bzmore(1), bzless(1), bzgrep(1), bzip2(1) BUGS Messages from the cmp or diff programs refer to temporary filenames instead of those specified. BZDIFF(1)
|
bzcmp, bzdiff - compare bzip2 compressed files
|
bzcmp [ cmp_options ] file1 [ file2 ] bzdiff [ diff_options ] file1 [ file2 ]
| null | null |
unpack200
| null | null | null | null | null |
corelist
|
See Module::CoreList for one.
|
corelist - a commandline frontend to Module::CoreList
|
corelist -v corelist [-a|-d] <ModuleName> | /<ModuleRegex>/ [<ModuleVersion>] ... corelist [-v <PerlVersion>] [ <ModuleName> | /<ModuleRegex>/ ] ... corelist [-r <PerlVersion>] ... corelist --utils [-d] <UtilityName> [<UtilityName>] ... corelist --utils -v <PerlVersion> corelist --feature <FeatureName> [<FeatureName>] ... corelist --diff PerlVersion PerlVersion corelist --upstream <ModuleName>
|
-a lists all versions of the given module (or the matching modules, in case you used a module regexp) in the perls Module::CoreList knows about. corelist -a Unicode Unicode was first released with perl v5.6.2 v5.6.2 3.0.1 v5.8.0 3.2.0 v5.8.1 4.0.0 v5.8.2 4.0.0 v5.8.3 4.0.0 v5.8.4 4.0.1 v5.8.5 4.0.1 v5.8.6 4.0.1 v5.8.7 4.1.0 v5.8.8 4.1.0 v5.8.9 5.1.0 v5.9.0 4.0.0 v5.9.1 4.0.0 v5.9.2 4.0.1 v5.9.3 4.1.0 v5.9.4 4.1.0 v5.9.5 5.0.0 v5.10.0 5.0.0 v5.10.1 5.1.0 v5.11.0 5.1.0 v5.11.1 5.1.0 v5.11.2 5.1.0 v5.11.3 5.2.0 v5.11.4 5.2.0 v5.11.5 5.2.0 v5.12.0 5.2.0 v5.12.1 5.2.0 v5.12.2 5.2.0 v5.12.3 5.2.0 v5.12.4 5.2.0 v5.13.0 5.2.0 v5.13.1 5.2.0 v5.13.2 5.2.0 v5.13.3 5.2.0 v5.13.4 5.2.0 v5.13.5 5.2.0 v5.13.6 5.2.0 v5.13.7 6.0.0 v5.13.8 6.0.0 v5.13.9 6.0.0 v5.13.10 6.0.0 v5.13.11 6.0.0 v5.14.0 6.0.0 v5.14.1 6.0.0 v5.15.0 6.0.0 -d finds the first perl version where a module has been released by date, and not by version number (as is the default). --diff Given two versions of perl, this prints a human-readable table of all module changes between the two. The output format may change in the future, and is meant for humans, not programs. For programs, use the Module::CoreList API. -? or -help help! help! help! to see more help, try --man. -man all of the help -v lists all of the perl release versions we got the CoreList for. If you pass a version argument (value of $], like 5.00503 or 5.008008), you get a list of all the modules and their respective versions. (If you have the "version" module, you can also use new- style version numbers, like 5.8.8.) In module filtering context, it can be used as Perl version filter. -r lists all of the perl releases and when they were released If you pass a perl version you get the release date for that version only. --utils lists the first version of perl each named utility program was released with May be used with -d to modify the first release criteria. If used with -v <version> then all utilities released with that version of perl are listed, and any utility programs named on the command line are ignored. --feature, -f lists the first version bundle of each named feature given --upstream, -u Shows if the given module is primarily maintained in perl core or on CPAN and bug tracker URL. As a special case, if you specify the module name "Unicode", you'll get the version number of the Unicode Character Database bundled with the requested perl versions.
|
$ corelist File::Spec File::Spec was first released with perl 5.005 $ corelist File::Spec 0.83 File::Spec 0.83 was released with perl 5.007003 $ corelist File::Spec 0.89 File::Spec 0.89 was not in CORE (or so I think) $ corelist File::Spec::Aliens File::Spec::Aliens was not in CORE (or so I think) $ corelist /IPC::Open/ IPC::Open2 was first released with perl 5 IPC::Open3 was first released with perl 5 $ corelist /MANIFEST/i ExtUtils::Manifest was first released with perl 5.001 $ corelist /Template/ /Template/ has no match in CORE (or so I think) $ corelist -v 5.8.8 B B 1.09_01 $ corelist -v 5.8.8 /^B::/ B::Asmdata 1.01 B::Assembler 0.07 B::Bblock 1.02_01 B::Bytecode 1.01_01 B::C 1.04_01 B::CC 1.00_01 B::Concise 0.66 B::Debug 1.02_01 B::Deparse 0.71 B::Disassembler 1.05 B::Lint 1.03 B::O 1.00 B::Showlex 1.02 B::Stackobj 1.00 B::Stash 1.00 B::Terse 1.03_01 B::Xref 1.01 COPYRIGHT Copyright (c) 2002-2007 by D.H. aka PodMaster Currently maintained by the perl 5 porters <perl5-porters@perl.org>. This program is distributed under the same terms as perl itself. See http://perl.org/ or http://cpan.org/ for more info on that. perl v5.38.2 2023-11-28 CORELIST(1)
|
expand
|
The expand utility processes the named files or the standard input writing the standard output with tabs changed into blanks. Backspace characters are preserved into the output and decrement the column count for tab calculations. The expand utility is useful for pre-processing character files (before sorting, looking at specific columns, etc.) that contain tabs. The unexpand utility puts tabs back into the data from the standard input or the named files and writes the result on the standard output. The following options are available: -a (unexpand only.) By default, only leading blanks and tabs are reconverted to maximal strings of tabs. If the -a option is given, then tabs are inserted whenever they would compress the resultant file by replacing two or more characters. -t -Sm tab1, tab2, ..., tabn Sm Set tab stops at column positions tab1, tab2, ..., tabn. If only a single number is given, tab stops are set that number of column positions apart instead of the default number of 8. ENVIRONMENT The LANG, LC_ALL and LC_CTYPE environment variables affect the execution of expand and unexpand as described in environ(7). EXIT STATUS The expand and unexpand utilities exit 0 on success, and >0 if an error occurs. STANDARDS The expand and unexpand utilities conform to IEEE Std 1003.1-2001 (“POSIX.1”). HISTORY The expand utility first appeared in 1BSD. macOS 14.5 June 6, 2015 macOS 14.5
|
expand, unexpand – expand tabs to spaces, and vice versa
|
expand [-t tab1,tab2,...,tabn] [file ...] unexpand [-a | -t tab1,tab2,...,tabn] [file ...]
| null | null |
strings
|
Strings looks for ASCII strings in a binary file or standard input. Strings is useful for identifying random object files and many other things. A string is any sequence of 4 (the default) or more printing characters [ending at, but not including, any other character or EOF]. Unless the - flag is given, strings looks in all sections of the object files except the (__TEXT,__text) section. If no files are specified standard input is read. The file arguments may be of the form libx.a(foo.o), to request information about only that object file and not the entire library. (Typically this argument must be quoted, ``libx.a(foo.o)'', to get it past the shell.) The options to strings(1) are: -a This option causes strings to look for strings in all sections of the object file (including the (__TEXT,__text) section. - This option causes strings to look for strings in all bytes of the files (the default for non-object files). -- This option causes strings to treat all the following arguments as files. -o Preceded each string by its offset in the file (in decimal). -t format Write each string preceded by its byte offset from the start of the file. The format shall be dependent on the single character used as the format option-argument: d The offset shall be written in decimal. o The offset shall be written in octal. x The offset shall be written in hexadecimal. -number The decimal number is used as the minimum string length rather than the default of 4. -n number Specify the minimum string length, where the number argument is a positive decimal integer. The default shall be 4. -arch arch_type Specifies the architecture, arch_type, of the file for strings(1) to operate on when the file is a universal file. (See arch(3) for the currently know arch_types.) The arch_type can be "all" to operate on all architectures in the file. SEE ALSO od(1) BUGS The algorithm for identifying strings is extremely primitive. Apple, Inc. June 7, 2016 STRINGS(1)
|
strings - find the printable strings in a object, or other binary, file
|
strings [ - ] [ -a ] [ -o ] [ -t format ] [ -number ] [ -n number ] [--] [file ...]
| null | null |
lwp-dump5.34
|
The lwp-dump program will get the resource identified by the URL and then dump the response object to STDOUT. This will display the headers returned and the initial part of the content, escaped so that it's safe to display even binary content. The escapes syntax used is the same as for Perl's double quoted strings. If there is no content the string "(no content)" is shown in its place. The following options are recognized: --agent string Override the user agent string passed to the server. --keep-client-headers LWP internally generate various "Client-*" headers that are stripped by lwp-dump in order to show the headers exactly as the server provided them. This option will suppress this. --max-length n How much of the content to show. The default is 512. Set this to 0 for unlimited. If the content is longer then the string is chopped at the limit and the string "...\n(### more bytes not shown)" appended. --method string Use the given method for the request instead of the default "GET". --parse-head By default lwp-dump will not try to initialize headers by looking at the head section of HTML documents. This option enables this. This corresponds to "parse_head" in LWP::UserAgent. --request Also dump the request sent. SEE ALSO lwp-request, LWP, "dump" in HTTP::Message perl v5.34.0 2020-04-14 LWP-DUMP(1)
|
lwp-dump - See what headers and content is returned for a URL
|
lwp-dump [ options ] URL
| null | null |
appletviewer
| null | null | null | null | null |
idlj
| null | null | null | null | null |
dbiproxy5.34
|
This tool is just a front end for the DBI::ProxyServer package. All it does is picking options from the command line and calling DBI::ProxyServer::main(). See DBI::ProxyServer for details. Available options include: --chroot=dir (UNIX only) After doing a bind(), change root directory to the given directory by doing a chroot(). This is useful for security, but it restricts the environment a lot. For example, you need to load DBI drivers in the config file or you have to create hard links to Unix sockets, if your drivers are using them. For example, with MySQL, a config file might contain the following lines: my $rootdir = '/var/dbiproxy'; my $unixsockdir = '/tmp'; my $unixsockfile = 'mysql.sock'; foreach $dir ($rootdir, "$rootdir$unixsockdir") { mkdir 0755, $dir; } link("$unixsockdir/$unixsockfile", "$rootdir$unixsockdir/$unixsockfile"); require DBD::mysql; { 'chroot' => $rootdir, ... } If you don't know chroot(), think of an FTP server where you can see a certain directory tree only after logging in. See also the --group and --user options. --configfile=file Config files are assumed to return a single hash ref that overrides the arguments of the new method. However, command line arguments in turn take precedence over the config file. See the "CONFIGURATION FILE" section in the DBI::ProxyServer documentation for details on the config file. --debug Turn debugging mode on. Mainly this asserts that logging messages of level "debug" are created. --facility=mode (UNIX only) Facility to use for Sys::Syslog. The default is daemon. --group=gid After doing a bind(), change the real and effective GID to the given. This is useful, if you want your server to bind to a privileged port (<1024), but don't want the server to execute as root. See also the --user option. GID's can be passed as group names or numeric values. --localaddr=ip By default a daemon is listening to any IP number that a machine has. This attribute allows one to restrict the server to the given IP number. --localport=port This attribute sets the port on which the daemon is listening. It must be given somehow, as there's no default. --logfile=file Be default logging messages will be written to the syslog (Unix) or to the event log (Windows NT). On other operating systems you need to specify a log file. The special value "STDERR" forces logging to stderr. See Net::Daemon::Log for details. --mode=modename The server can run in three different modes, depending on the environment. If you are running Perl 5.005 and did compile it for threads, then the server will create a new thread for each connection. The thread will execute the server's Run() method and then terminate. This mode is the default, you can force it with "--mode=threads". If threads are not available, but you have a working fork(), then the server will behave similar by creating a new process for each connection. This mode will be used automatically in the absence of threads or if you use the "--mode=fork" option. Finally there's a single-connection mode: If the server has accepted a connection, he will enter the Run() method. No other connections are accepted until the Run() method returns (if the client disconnects). This operation mode is useful if you have neither threads nor fork(), for example on the Macintosh. For debugging purposes you can force this mode with "--mode=single". --pidfile=file (UNIX only) If this option is present, a PID file will be created at the given location. Default is to not create a pidfile. --user=uid After doing a bind(), change the real and effective UID to the given. This is useful, if you want your server to bind to a privileged port (<1024), but don't want the server to execute as root. See also the --group and the --chroot options. UID's can be passed as group names or numeric values. --version Suppresses startup of the server; instead the version string will be printed and the program exits immediately. AUTHOR Copyright (c) 1997 Jochen Wiedmann Am Eisteich 9 72555 Metzingen Germany Email: joe@ispsoft.de Phone: +49 7123 14881 The DBI::ProxyServer module is free software; you can redistribute it and/or modify it under the same terms as Perl itself. In particular permission is granted to Tim Bunce for distributing this as a part of the DBI. SEE ALSO DBI::ProxyServer, DBD::Proxy, DBI perl v5.34.0 2024-04-13 DBIPROXY(1)
|
dbiproxy - A proxy server for the DBD::Proxy driver
|
dbiproxy <options> --localport=<port>
| null | null |
wall
|
The wall utility displays the contents of file or, by default, its standard input, on the terminals of all currently logged in users. Only the super-user can write on the terminals of users who have chosen to deny messages or are using a program which automatically denies messages. -g Send messages to users in this group. This option may be specified multiple times, and any user in any of the specified groups will receive the message. SEE ALSO mesg(1), talk(1), write(1), shutdown(8) HISTORY A wall command appeared in PWB UNIX. BUGS The sender's LC_CTYPE setting is used to determine which characters are safe to write to a terminal, not the receiver's (which wall has no way of knowing). macOS 14.5 February 24, 2012 macOS 14.5
|
wall – write a message to users
|
wall [-g group] [file]
| null | null |
brctl
|
brctl understands the following commands: diagnose [options] [⟨diagnosis-output-path⟩] diagnose and collect logs -M,--collect-mobile-documents[=<container>] (default: all containers) -s,--sysdiagnose Do not collect what's already part of sysdiagnose -n,--name=<name> Change the device name [<diagnosis-output-path>] Specifies the output path of the diagnosis; -n becomes useless. download ⟨path⟩ download a local copy of the document at this path evict ⟨path⟩ evict the local copy of the document at this path log [options] [⟨command⟩] -c,--color[={yes,no}] turn on or off color use -d,--path=<logs-dir> use <logs-dir> instead of default -H,--home=<home-dir> use this as the ~ prefix, to look for ~/L/ -f,--filter=<predicate> only show lines matching predicate -m,--multiline[={yes,no}] turn on or off multiple line logging -n=<number> number of initial lines to display -p,--page use paging -w,--wait wait for new logs continuously (syslog -w) -t,--shorten Shorten UUIDs, paths, etc -s,--digest Only print digest logs dump [options] [⟨container⟩] dump the CloudDocs database -o,--output=<file-path> redirect output to <file-path> -d,--database-path=<db-path> Use the database at <db-path> [<container>] the container to be dumped monitor [options] ⟨container⟩ use NSMetadataQuery to monitor the container -S,--scope=<scope> restrict the NSMDQ scope to DOCS, DATA, or BOTH versions [options] ⟨path⟩ [ALL|etags...] list the non-local versions of the document at this path. -a,--all List all non-local versions including those that are locally cached SEE ALSO bird(8) Mac OS X 22/04/14 Mac OS X
|
brctl – Manage the CloudDocs daemon
|
brctl ⟨command⟩ [command-options and arguments]
| null | null |
yapp5.34
|
yapp is a frontend to the Parse::Yapp module, which lets you compile Parse::Yapp grammar input files into Perl LALR(1) OO parser modules.
|
yapp - A perl frontend to the Parse::Yapp module SYNOPSYS yapp [options] grammar[.yp] yapp -V yapp -h
| null |
Options, as of today, are all optionals :-) -v Creates a file grammar.output describing your parser. It will show you a summary of conflicts, rules, the DFA (Deterministic Finite Automaton) states and overall usage of the parser. -s Create a standalone module in which the driver is included. Note that if you have more than one parser module called from a program, to have it standalone, you need this option only for one of your parser module. -n Disable source file line numbering embedded in your parser module. I don't know why one should need it, but it's there. -m module Gives your parser module the package name (or name space or module name or class name or whatever-you-call-it) of module. It defaults to grammar -o outfile The compiled output file will be named outfile for your parser module. It defaults to grammar.pm or, if you specified the option -m A::Module::Name (see below), to Name.pm, in the current working directory. -t filename The -t filename option allows you to specify a file which should be used as template for generating the parser output. The default is to use the internal template defined in Parse::Yapp::Output.pm. For how to write your own template and which substitutions are available, have a look to the module Parse::Yapp::Output.pm : it should be obvious. -b shebang If you work on systems that understand so called shebangs, and your generated parser is directly an executable script, you can specifie one with the -b option, ie: yapp -b '/usr/local/bin/perl -w' -o myscript.pl myscript.yp This will output a file called myscript.pl whose very first line is: #!/usr/local/bin/perl -w The argument is mandatory, but if you specify an empty string, the value of $Config{perlpath} will be used instead. grammar The input grammar file. If no suffix is given, and the file does not exists, an attempt to open the file with a suffix of .yp is tried before exiting. -V Display current version of Parse::Yapp and gracefully exits. -h Display the usage screen. BUGS None known now :-) AUTHOR William N. Braswell, Jr. <wbraswell_cpan@NOSPAM.nym.hush.com> (Remove "NOSPAM".) COPYRIGHT Copyright © 1998, 1999, 2000, 2001, Francois Desarmenien. Copyright © 2017 William N. Braswell, Jr. See Parse::Yapp(3) for legal use and distribution rights SEE ALSO Parse::Yapp(3) Perl(1) yacc(1) bison(1) perl v5.34.0 2017-08-04 YAPP(1)
| null |
gm4
|
Process macros in FILEs. If no FILE or if FILE is `-', standard input is read. Mandatory or optional arguments to long options are mandatory or optional for short options too. Operation modes: --help display this help and exit --version output version information and exit -E, --fatal-warnings stop execution after first warning -e, --interactive unbuffer output, ignore interrupts -P, --prefix-builtins force a `m4_' prefix to all builtins -Q, --quiet, --silent suppress some warnings for builtins Preprocessor features: -D, --define=NAME[=VALUE] define NAME has having VALUE, or empty -I, --include=DIRECTORY append DIRECTORY to include path -s, --synclines generate `#line NUM "FILE"' lines -U, --undefine=NAME undefine NAME Limits control: -G, --traditional suppress all GNU extensions -H, --hashsize=PRIME set symbol lookup hash table size [509] -L, --nesting-limit=NUMBER change artificial nesting limit [1024] Frozen state files: -F, --freeze-state=FILE produce a frozen state on FILE at end -R, --reload-state=FILE reload a frozen state from FILE at start Debugging: -d, --debug[=FLAGS] set debug level (no FLAGS implies `aeq') -l, --arglength=NUM restrict macro tracing size -o, --error-output=FILE redirect debug and trace output -t, --trace=NAME trace NAME when it will be defined FLAGS is any of: a show actual arguments c show before collect, after collect and after call e show expansion f say current input file name i show changes in input files l say current input line number p show results of path searches q quote values as necessary, with a or e flag t trace for all macro calls, not only traceon'ed x add a unique macro call id, useful with c flag V shorthand for all of the above flags If defined, the environment variable `M4PATH' is a colon-separated list of directories included after any specified by `-I'. Exit status is 0 for success, 1 for failure, 63 for frozen file version mismatch, or whatever value was passed to the m4exit macro. AUTHOR Written by Rene' Seindal. REPORTING BUGS Report bugs to <bug-m4@gnu.org>. COPYRIGHT Copyright © 2006 Free Software Foundation, Inc. This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. SEE ALSO The full documentation for M4 is maintained as a Texinfo manual. If the info and M4 programs are properly installed at your site, the command info m4 should give you access to the complete manual. FSF August 2006 M4(1)
|
M4 - macro processor
|
m4 [OPTION]... [FILE]...
| null | null |
db_hotbackup
| null | null | null | null | null |
auvaltool
|
AUValidation tests a specified AudioUnit for API and behavioural conformance. returns: OK: 0, malformed execution: 1, unit not conformant: -1
|
auval – AudioUnit validation
|
auval [-s TYPE] [-a] [-v|vt TYPE SUBT MANU [-w] [-de] [-dw]] [-f file]
|
-h print help text -a lists all available AudioUnits of any type -s TYPE lists all available AudioUnits of type 'TYPE' -v TYPE SUBT MANU opens the AudioUnit specified by the TYPE SUBT MANU component ID's and tests that unit. -vt TYPE MANU iterates through all of the AU's of specified TYPE and MANU -de execution is terminated when first error is encountered -dw execution is terminated when first warning is encountered -c continue validating when an error occurs in batch mode. -q quiet - does no printing except for errors or warnings -qp doesn't print parameter or Factory Presets information -o only runs a basic open and initialize test. good for debugging basic functionality -r N repeat the whole process of validation N times. good for catching open/init bugs. -w wait after finished - good for profiling memory usage see 'man leaks' -vers The version is printed to stdout. -versh The version is printed to stdout in hexadecimal. -f FILENAME Each line in the file should contain one complete command. Darwin February 13, 2006 Darwin
| null |
applesingle
|
applesingle, binhex, macbinary are implemented as a single tool with multiple names. All invocations support the three verbs encode, decode, and probe. If multiple files are passed to probe, the exit status will be non-zero only if all files contain data in the specified encoding.
|
applesingle, binhex, macbinary – encode and decode files
|
<tool> probe file ... <tool> [decode] [-c] [-fv] [-C dir] [-o outfile] [file ...] <tool> -h | -V applesingle encode [-cfv] [-s suf] [-C dir] [-o outfile] file ... binhex encode [-R] [-cfv] [-s suf] [-C dir] [-o outfile] file ... macbinary encode [-t 1-3] [-cfv] [-s suf] [-C dir] [-o outfile] file ...
|
-f, --force perform the operation even if the output file already exists -h, --help display version and usage, then quit -v, --verbose be verbose -V, --version display version, then quit -c, --pipe, --from-stdin, --to-stdout For decode, read encoded data from the standand input. For encode, write encoded data to the standard output. Currently, "plain" data must be written to and from specified filenames (see also mount_fdesc(8)). -C, --directory dir create output files in dir -o, --rename name Use name for output, overriding any stored or default name. For encode, the appropriate suffix will be added to name. -o implies only one file to be encoded or decoded. -s, --suffix .suf override the default suffix for the given encoding -R, --no-runlength-encoding don't use BinHex runlength compression when encoding -t, --type 1-3 Specify MacBinary encoding type. Type 1 is undesirable because it has neither a checksum nor a signature and is thus difficult to recognize. DIAGNOSTICS In general, the tool returns a non-zero exit status if it fails. Darwin 14 November 2005 Darwin
| null |
scandeps5.30.pl
| null | null | null | null | null |
clang
|
clang is a C, C++, and Objective-C compiler which encompasses preprocessing, parsing, optimization, code generation, assembly, and linking. Depending on which high-level mode setting is passed, Clang will stop before doing a full link. While Clang is highly integrated, it is important to understand the stages of compilation, to understand how to invoke it. These stages are: Driver The clang executable is actually a small driver which controls the overall execution of other tools such as the compiler, assembler and linker. Typically you do not need to interact with the driver, but you transparently use it to run the other tools. Preprocessing This stage handles tokenization of the input source file, macro expansion, #include expansion and handling of other preprocessor directives. The output of this stage is typically called a ".i" (for C), ".ii" (for C++), ".mi" (for Objective-C), or ".mii" (for Objective-C++) file. Parsing and Semantic Analysis This stage parses the input file, translating preprocessor tokens into a parse tree. Once in the form of a parse tree, it applies semantic analysis to compute types for expressions as well and determine whether the code is well formed. This stage is responsible for generating most of the compiler warnings as well as parse errors. The output of this stage is an "Abstract Syntax Tree" (AST). Code Generation and Optimization This stage translates an AST into low-level intermediate code (known as "LLVM IR") and ultimately to machine code. This phase is responsible for optimizing the generated code and handling target-specific code generation. The output of this stage is typically called a ".s" file or "assembly" file. Clang also supports the use of an integrated assembler, in which the code generator produces object files directly. This avoids the overhead of generating the ".s" file and of calling the target assembler. Assembler This stage runs the target assembler to translate the output of the compiler into a target object file. The output of this stage is typically called a ".o" file or "object" file. Linker This stage runs the target linker to merge multiple object files into an executable or dynamic library. The output of this stage is typically called an "a.out", ".dylib" or ".so" file. Clang Static Analyzer The Clang Static Analyzer is a tool that scans source code to try to find bugs through code analysis. This tool uses many parts of Clang and is built into the same driver. Please see <https://clang-analyzer.llvm.org> for more details on how to use the static analyzer.
|
clang - the Clang C, C++, and Objective-C compiler
|
clang [options] filename ...
|
Stage Selection Options -E Run the preprocessor stage. -fsyntax-only Run the preprocessor, parser and semantic analysis stages. -S Run the previous stages as well as LLVM generation and optimization stages and target-specific code generation, producing an assembly file. -c Run all of the above, plus the assembler, generating a target ".o" object file. no stage selection option If no stage selection option is specified, all stages above are run, and the linker is run to combine the results into an executable or shared library. Language Selection and Mode Options -x <language> Treat subsequent input files as having type language. -std=<standard> Specify the language standard to compile for. Supported values for the C language are: c89 c90 iso9899:1990 ISO C 1990 iso9899:199409 ISO C 1990 with amendment 1 gnu89 gnu90 ISO C 1990 with GNU extensions c99 iso9899:1999 ISO C 1999 gnu99 ISO C 1999 with GNU extensions c11 iso9899:2011 ISO C 2011 gnu11 ISO C 2011 with GNU extensions c17 iso9899:2017 ISO C 2017 gnu17 ISO C 2017 with GNU extensions The default C language standard is gnu17, except on PS4, where it is gnu99. Supported values for the C++ language are: c++98 c++03 ISO C++ 1998 with amendments gnu++98 gnu++03 ISO C++ 1998 with amendments and GNU extensions c++11 ISO C++ 2011 with amendments gnu++11 ISO C++ 2011 with amendments and GNU extensions c++14 ISO C++ 2014 with amendments gnu++14 ISO C++ 2014 with amendments and GNU extensions c++17 ISO C++ 2017 with amendments gnu++17 ISO C++ 2017 with amendments and GNU extensions c++20 ISO C++ 2020 with amendments gnu++20 ISO C++ 2020 with amendments and GNU extensions c++2b Working draft for ISO C++ 2023 gnu++2b Working draft for ISO C++ 2023 with GNU extensions The default C++ language standard is gnu++98. Supported values for the OpenCL language are: cl1.0 OpenCL 1.0 cl1.1 OpenCL 1.1 cl1.2 OpenCL 1.2 cl2.0 OpenCL 2.0 The default OpenCL language standard is cl1.0. Supported values for the CUDA language are: cuda NVIDIA CUDA(tm) -stdlib=<library> Specify the C++ standard library to use; supported options are libstdc++ and libc++. If not specified, platform default will be used. -rtlib=<library> Specify the compiler runtime library to use; supported options are libgcc and compiler-rt. If not specified, platform default will be used. -ansi Same as -std=c89. -ObjC, -ObjC++ Treat source input files as Objective-C and Object-C++ inputs respectively. -trigraphs Enable trigraphs. -ffreestanding Indicate that the file should be compiled for a freestanding, not a hosted, environment. Note that it is assumed that a freestanding environment will additionally provide memcpy, memmove, memset and memcmp implementations, as these are needed for efficient codegen for many programs. -fno-builtin Disable special handling and optimizations of well-known library functions, like strlen() and malloc(). -fno-builtin-<function> Disable special handling and optimizations for the specific library function. For example, -fno-builtin-strlen removes any special handling for the strlen() library function. -fno-builtin-std-<function> Disable special handling and optimizations for the specific C++ standard library function in namespace std. For example, -fno-builtin-std-move_if_noexcept removes any special handling for the std::move_if_noexcept() library function. For C standard library functions that the C++ standard library also provides in namespace std, use -fno-builtin-<function> instead. -fmath-errno Indicate that math functions should be treated as updating errno. -fpascal-strings Enable support for Pascal-style strings with "\pfoo". -fms-extensions Enable support for Microsoft extensions. -fmsc-version= Set _MSC_VER. Defaults to 1300 on Windows. Not set otherwise. -fborland-extensions Enable support for Borland extensions. -fwritable-strings Make all string literals default to writable. This disables uniquing of strings and other optimizations. -flax-vector-conversions, -flax-vector-conversions=<kind>, -fno-lax-vector-conversions Allow loose type checking rules for implicit vector conversions. Possible values of <kind>: • none: allow no implicit conversions between vectors • integer: allow implicit bitcasts between integer vectors of the same overall bit-width • all: allow implicit bitcasts between any vectors of the same overall bit-width <kind> defaults to integer if unspecified. -fblocks Enable the "Blocks" language feature. -fobjc-abi-version=version Select the Objective-C ABI version to use. Available versions are 1 (legacy "fragile" ABI), 2 (non-fragile ABI 1), and 3 (non-fragile ABI 2). -fobjc-nonfragile-abi-version=<version> Select the Objective-C non-fragile ABI version to use by default. This will only be used as the Objective-C ABI when the non-fragile ABI is enabled (either via -fobjc-nonfragile-abi, or because it is the platform default). -fobjc-nonfragile-abi, -fno-objc-nonfragile-abi Enable use of the Objective-C non-fragile ABI. On platforms for which this is the default ABI, it can be disabled with -fno-objc-nonfragile-abi. Target Selection Options Clang fully supports cross compilation as an inherent part of its design. Depending on how your version of Clang is configured, it may have support for a number of cross compilers, or may only support a native target. -arch <architecture> Specify the architecture to build for (Mac OS X specific). -target <architecture> Specify the architecture to build for (all platforms). -mmacosx-version-min=<version> When building for macOS, specify the minimum version supported by your application. -miphoneos-version-min When building for iPhone OS, specify the minimum version supported by your application. --print-supported-cpus Print out a list of supported processors for the given target (specified through --target=<architecture> or -arch <architecture>). If no target is specified, the system default target will be used. -mcpu=?, -mtune=? Acts as an alias for --print-supported-cpus. -march=<cpu> Specify that Clang should generate code for a specific processor family member and later. For example, if you specify -march=i486, the compiler is allowed to generate instructions that are valid on i486 and later processors, but which may not exist on earlier ones. Code Generation Options -O0, -O1, -O2, -O3, -Ofast, -Os, -Oz, -Og, -O, -O4 Specify which optimization level to use: -O0 Means "no optimization": this level compiles the fastest and generates the most debuggable code. -O1 Somewhere between -O0 and -O2. -O2 Moderate level of optimization which enables most optimizations. -O3 Like -O2, except that it enables optimizations that take longer to perform or that may generate larger code (in an attempt to make the program run faster). -Ofast Enables all the optimizations from -O3 along with other aggressive optimizations that may violate strict compliance with language standards. -Os Like -O2 with extra optimizations to reduce code size. -Oz Like -Os (and thus -O2), but reduces code size further. -Og Like -O1. In future versions, this option might disable different optimizations in order to improve debuggability. -O Equivalent to -O1. -O4 and higher Currently equivalent to -O3 -g, -gline-tables-only, -gmodules Control debug information output. Note that Clang debug information works best at -O0. When more than one option starting with -g is specified, the last one wins: -g Generate debug information. -gline-tables-only Generate only line table debug information. This allows for symbolicated backtraces with inlining information, but does not include any information about variables, their locations or types. -gmodules Generate debug information that contains external references to types defined in Clang modules or precompiled headers instead of emitting redundant debug type information into every object file. This option transparently switches the Clang module format to object file containers that hold the Clang module together with the debug information. When compiling a program that uses Clang modules or precompiled headers, this option produces complete debug information with faster compile times and much smaller object files. This option should not be used when building static libraries for distribution to other machines because the debug info will contain references to the module cache on the machine the object files in the library were built on. -fstandalone-debug -fno-standalone-debug Clang supports a number of optimizations to reduce the size of debug information in the binary. They work based on the assumption that the debug type information can be spread out over multiple compilation units. For instance, Clang will not emit type definitions for types that are not needed by a module and could be replaced with a forward declaration. Further, Clang will only emit type info for a dynamic C++ class in the module that contains the vtable for the class. The -fstandalone-debug option turns off these optimizations. This is useful when working with 3rd-party libraries that don't come with debug information. This is the default on Darwin. Note that Clang will never emit type information for types that are not referenced at all by the program. -feliminate-unused-debug-types By default, Clang does not emit type information for types that are defined but not used in a program. To retain the debug info for these unused types, the negation -fno-eliminate-unused-debug-types can be used. -fexceptions Enable generation of unwind information. This allows exceptions to be thrown through Clang compiled stack frames. This is on by default in x86-64. -ftrapv Generate code to catch integer overflow errors. Signed integer overflow is undefined in C. With this flag, extra code is generated to detect this and abort when it happens. -fvisibility This flag sets the default visibility level. -fcommon, -fno-common This flag specifies that variables without initializers get common linkage. It can be disabled with -fno-common. -ftls-model=<model> Set the default thread-local storage (TLS) model to use for thread-local variables. Valid values are: "global-dynamic", "local-dynamic", "initial-exec" and "local-exec". The default is "global-dynamic". The default model can be overridden with the tls_model attribute. The compiler will try to choose a more efficient model if possible. -flto, -flto=full, -flto=thin, -emit-llvm Generate output files in LLVM formats, suitable for link time optimization. When used with -S this generates LLVM intermediate language assembly files, otherwise this generates LLVM bitcode format object files (which may be passed to the linker depending on the stage selection options). The default for -flto is "full", in which the LLVM bitcode is suitable for monolithic Link Time Optimization (LTO), where the linker merges all such modules into a single combined module for optimization. With "thin", ThinLTO compilation is invoked instead. NOTE: On Darwin, when using -flto along with -g and compiling and linking in separate steps, you also need to pass -Wl,-object_path_lto,<lto-filename>.o at the linking step to instruct the ld64 linker not to delete the temporary object file generated during Link Time Optimization (this flag is automatically passed to the linker by Clang if compilation and linking are done in a single step). This allows debugging the executable as well as generating the .dSYM bundle using dsymutil(1). Driver Options -### Print (but do not run) the commands to run for this compilation. --help Display available options. -Qunused-arguments Do not emit any warnings for unused driver arguments. -Wa,<args> Pass the comma separated arguments in args to the assembler. -Wl,<args> Pass the comma separated arguments in args to the linker. -Wp,<args> Pass the comma separated arguments in args to the preprocessor. -Xanalyzer <arg> Pass arg to the static analyzer. -Xassembler <arg> Pass arg to the assembler. -Xlinker <arg> Pass arg to the linker. -Xpreprocessor <arg> Pass arg to the preprocessor. -o <file> Write output to file. -print-file-name=<file> Print the full library path of file. -print-libgcc-file-name Print the library path for the currently used compiler runtime library ("libgcc.a" or "libclang_rt.builtins.*.a"). -print-prog-name=<name> Print the full program path of name. -print-search-dirs Print the paths used for finding libraries and programs. -save-temps Save intermediate compilation results. -save-stats, -save-stats=cwd, -save-stats=obj Save internal code generation (LLVM) statistics to a file in the current directory (-save-stats/"-save-stats=cwd") or the directory of the output file ("-save-state=obj"). -integrated-as, -no-integrated-as Used to enable and disable, respectively, the use of the integrated assembler. Whether the integrated assembler is on by default is target dependent. -time Time individual commands. -ftime-report Print timing summary of each stage of compilation. -v Show commands to run and use verbose output. Diagnostics Options -fshow-column, -fshow-source-location, -fcaret-diagnostics, -fdiagnostics-fixit-info, -fdiagnostics-parseable-fixits, -fdiagnostics-print-source-range-info, -fprint-source-range-info, -fdiagnostics-show-option, -fmessage-length These options control how Clang prints out information about diagnostics (errors and warnings). Please see the Clang User's Manual for more information. Preprocessor Options -D<macroname>=<value> Adds an implicit #define into the predefines buffer which is read before the source file is preprocessed. -U<macroname> Adds an implicit #undef into the predefines buffer which is read before the source file is preprocessed. -include <filename> Adds an implicit #include into the predefines buffer which is read before the source file is preprocessed. -I<directory> Add the specified directory to the search path for include files. -F<directory> Add the specified directory to the search path for framework include files. -nostdinc Do not search the standard system directories or compiler builtin directories for include files. -nostdlibinc Do not search the standard system directories for include files, but do search compiler builtin include directories. -nobuiltininc Do not search clang's builtin directory for include files. ENVIRONMENT TMPDIR, TEMP, TMP These environment variables are checked, in order, for the location to write temporary files used during the compilation process. CPATH If this environment variable is present, it is treated as a delimited list of paths to be added to the default system include path list. The delimiter is the platform dependent delimiter, as used in the PATH environment variable. Empty components in the environment variable are ignored. C_INCLUDE_PATH, OBJC_INCLUDE_PATH, CPLUS_INCLUDE_PATH, OBJCPLUS_INCLUDE_PATH These environment variables specify additional paths, as for CPATH, which are only used when processing the appropriate language. MACOSX_DEPLOYMENT_TARGET If -mmacosx-version-min is unspecified, the default deployment target is read from this environment variable. This option only affects Darwin targets. BUGS To report bugs, please visit <https://github.com/llvm/llvm-project/issues/>. Most bug reports should include preprocessed source files (use the -E option) and the full output of the compiler, along with information to reproduce. SEE ALSO as(1), ld(1) AUTHOR Maintained by the Clang / LLVM Team (<http://clang.llvm.org>) COPYRIGHT 2007-2024, The Clang Team 11 January 28, 2024 CLANG(1)
| null |
ibtool
|
ibtool verifies, updates, and prints the contents of an Interface Builder document, generating its output in standard plist format. The tool follows a "read", "modify", "write", "print" order of operations.
|
ibtool - compiles, prints, updates, and verifies Interface Builder documents.
|
ibtool [options] document
|
Specifying Output: --write path Write the resulting Interface Builder document to the specified path. --output-format format By default, ibtool provides output in the form of an XML property list. Specifying binary1 will instruct ibtool to output a binary property list. Similarly, xml1 specifies an XML property list, and human-readable-text specifies human readable text. Compiling: --compile path Compile the input file and write it to the specified path. The compile option instructs ibtool to convert an Interface Builder document to a file that can be loaded at runtime. The compile option can be combined with --flatten NO to produce an output file that is both runnable, and editable. Additionally, --warnings, --errors, and --output-format are three other options that are commonly combined with --compile. --flatten boolean When combined with the --compile option and a value of NO, --flatten instructs ibtool to produce an output file that is both runnable, and editable. This option is typically used when preparing a product for localization. If no value is specified, the default flattening option is YES. --module moduleName Specifies the module name to include in applicable custom class names at runtime. The module name should match the PRODUCT_MODULE_NAME build setting of the target for which the Interface Builder document is being compiled. This option is required for compiling any document that contains an object whose custom class is a Swift class and whose customModuleProvider attribute is set to target. Postprocessing: --strip path Remove design-time content from the input Interface Builder NIB document and write the stripped output to path. Stripping a NIB reduces its size and makes it uneditable. Loading Bundles: --bundle bundlePath Load the bundle located at bundlePath. Localization: --previous-file previousFile Specifies the Interface Builder document from the previous development iteration for the purpose of incremental localizations. --incremental-file incrementalFile Specifies the Interface Builder document from which to take prior localizations during incremental localizations. incrementalFile will often be the translated version of the document specified with --previous-file. --localize-incremental Consolidate the structural and localization changes between the input document, the document specified by previousFile, and the document specified by incrementalFile, writing the resulting document to the path specified by --write. When specifying this option, one MUST provide a --previous-file and an --incremental-file. --reference-external-strings-file When combined with the --compile option, this flag indicates that the files in the Base.lproj locale folder should be compiled to reference the matching external strings files found in the other locales when loaded. Available on 10.8 and later. --companion-strings-file locale:stringsFile When combined with the --compile option, --companion-strings- file specifies the locale and input path of a strings file associated with the compiled document. ibtool will copy the strings file into the lproj directory for that locale. If ibtool renames the output compiled document, it will rename the output strings file as well. This option implies --reference-external- strings-file and should be repeated once for each associated strings file. The output path provided with --compile must be located in a directory named "Base.lproj". Importing: --import plist Apply the property values specified in plist to the object in the Interface Builder document with the matching object ID. plist should be a plist in the format produced by --export, only with the dictionary containing the key com.apple.ibtool.document.export removed from surrounding the object IDs. Note that setting a property in this manner may not actually modify the value of the property, or conversely may alter additional properties. Often, the input to this option is a modified version of the output from --export. Combine this option with --write in order to produce an Interface Builder document with the modified property values. --import-strings-file stringsFile Replace the document's localizable strings with the translations specified by stringsFile. stringsFile must be in the format generated by --generate-strings-file. --import-xliff XLIFFDocument Replace the document's localizable strings with the translations specified by XLIFFDocument. XLIFFDocument must be in the format generated by --export-xliff. Exporting: --export plist For each object in the Interface Builder document, export the union of the properties specified in plist for each class listed in plist of which the object is either a member, or a member of a descending class. plist should be a plist composed of a dictionary with class names for keys, and arrays of property key paths for values. The output is a plist composed of a dictionary under the key com.apple.ibtool.document.export, where each key is an object ID, and each value is a dictionary of property key paths to their corresponding value. If a key in the second level dictionary begins with a ´.´, it isn't a key path. Currently, the only special key is ".nilKeypaths", which identifies an array of key paths whose value was nil. An example of the plist format expected for the --export command follows; this sample will export the 'title' property of any NSCell in the given document. <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>NSCell</key> <array> <string>title</string> </array> </dict> </plist> Refactoring: --convert old-new Rename the class named 'old' to 'new'. If 'old' ends with an '*', then rename all classes beginning with 'old' to instead begin with 'new'. This Updates all outlet types, class names, custom classes, as well as actual instances. --upgrade Upgrade the input Interface Builder document to the latest document type. If the output path is not specified by --write, the upgraded document will be written to the input path. --remove-plugin-dependencies Remove any dependencies on Interface Builder 3 plug-ins so that the document may be used in Xcode, and write the resulting document to the path specified by --write. Instances of objects integrated by plug-ins are replaced with instances of classes integrated by system frameworks and the custom class is set to correspond to the old class. For each attribute set on the object, a corresponding User Defined Runtime Attribute is set on the resulting object where possible. Any unconverted key paths are reported. Conversion may result in changes to runtime behavior, thus converted files must be verified. Working with Auto Layout: --enable-auto-layout Enables Auto Layout in the given document and writes the resulting document to the path specified by --write. --update-frames Adjust the frames of misplaced or ambiguous objects to match their constraints. All misplaced and ambiguous objects in the output document specified by --write are affected by this option. Frames can only be updated in Xcode 5.0 documents that use Auto Layout. --update-constraints Adjust the constraint constants of misplaced objects so that the objects match their frames. All misplaced objects in the output document specified by --write are affected by this option. Constraints can only be updated in Xcode 5.0 documents that use Auto Layout. Printing File Contents: --export-strings-file stringsfile Extract the localizable strings from the Interface Builder document into stringsfile in the format expected by --import-strings-file. --export-xliff XLIFFDocument Extract the localizable strings from the Interface Builder document into XLIFFDocument as an XLIFF XML document in the format expected by --import-xliff. This argument can be combined with the --source-language and --target-language arguments in order to specify the resultant XLIFF document's source and target languages. --source-language sourceLanguage Specifies the source language of the resultant XLIFF document. This argument MUST be used in conjunction with the --export-xliff argument. If the --source-language argument is not specified, the language "en" is used by default. --target-language targetLanguage Specifies the target language of the resultant XLIFF document. This argument MUST be used in conjunction with the --export-xliff argument. If the --target-language argument is not specified, no target language is used. --warnings Include document warning messages in ibtool's plist output. Warnings will appear under the key com.apple.ibtool.document.warnings, with messages listed under the subkey message and warning types under the subkey type. --errors Include document error messages in ibtool's plist output. Errors will appear under the key com.apple.ibtool.document.errors, with messages listed under the subkey message and error types under the subkey type. --notices Include document notice messages in ibtool's plist output. Notices will appear under the key com.apple.ibtool.document.notices, with messages listed under the subkey message and error types under the subkey type. --localizable-strings Include the document's non-empty localizable strings in ibtool's plist output. The localizable strings are output as a dictionary under the key com.apple.ibtool.document.localizable-strings, where the keys are the object ids for objects with localizable strings, and the values are dictionaries of property names and their associated localizable strings. Empty strings are ignored. --localizable-stringarrays Include the document's localizable strings that are values of to-many relationships in ibtool's plist output. The to-many localizable strings are output as a dictionary under the key com.apple.ibtool.document.localizable-stringarrays, where the keys are the object ids for objects with to-many localizable strings, and the values are dictionaries of property names and their associated array of localizable strings. --localizable-geometry Include the document's localizable geometry in ibtool's plist output. The localizable geometry is output as a dictionary under the key com.apple.ibtool.document.localizable-geometry, where the keys are the object ids for objects with localizable geometry, and the values are dictionaries of property names and a plist representation of that geometry. --localizable-other Include the document's localizable attributes, other than strings and geometry, in ibtool's plist output. These attributes are output as a dictionary under the key com.apple.ibtool.document.localizable-other, where the keys are the object ids for objects with the localizable attributes, and the values are dictionaries of localizable property names and their associated values. --localizable-to-many-relationships Include the document's localizable to-many relationships in ibtool's plist output. These attributes are output as a dictionary under the key com.apple.ibtool.document.localizable-to-many-relationships, where the keys are the object ids for objects with the localizable attributes, and the values are dictionaries of localizable property names and their associated values. --localizable-all Include all the document's localizable attributes in ibtool's plist output. The localizable attributes are output as a dictionary under the key com.apple.ibtool.document.localizable-all, where the keys are the objects ids of the objects with localizable properties, and the values are the union of the properties returned by --localizable-strings, --localizable-stringarrays, --localizable-geometry, and --localizable-other. --localizable-to-many-relationships. --objects Include the document's objects in ibtool's plist output. The objects are output as a dictionary under the key com.apple.ibtool.document.objects, where the keys are the object ids, and the values are the associated object's properties. --hierarchy Include the document's hierarchy in ibtool's plist output. The hierarchy is output as a nested array of dictionaries under the key com.apple.ibtool.document.hierarchy with the subkeys objectID, label, name, children, and customClass. --connections Include the document's connections in ibtool's plist output. The connections are output as a dictionary under the key com.apple.ibtool.document.connections, where the keys are the object ids, and the values are dictionaries including the subkeys destination-id, destination-label, label, source-id, source-label, and type. --classes Include the document's classes in ibtool's plist output. The classes are output as a dictionary under the key com.apple.ibtool.document.classes, where the keys are the class names, and the values are dictionaries including the subkeys actions, outlets, superclass, and class. --version-history Print the version of IB/System Version used when the document was last saved. The version history information is output under the key com.apple.ibtool.document.version-history with the subkey interface-builder-version. --all Include all the information listed above in ibtool's plist output. Version Information: --version Print the version of ibtool. The version information is output under the key com.apple.ibtool.version with the subkeys bundle-version and short-bundle-version.
|
ibtool --export-strings-file file.strings file.nib ibtool will output a strings file containing all of file.nib's localizable strings. ibtool --previous-file orig.nib --incremental-file trans.nib --localize-incremental --write newTrans.nib mod.nib ibtool will leverage orig.nib against changes in mod.nib and localizations from trans.nib, and write the new translated nib. ibtool --previous-file orig.nib --incremental-file trans.nib --import-strings-file str.strings --localize-incremental mod.nib --write newTrans.nib ibtool will execute in the same manner as above, but will apply the strings from str.strings before writing the document. ibtool --warnings --errors --notices file.nib > alerts.plist alerts.plist will contain all warning, error, and notice messages from file.nib. ibtool --plugin path/to/some.plugin --localizable-geometry file.nib ibtool will load some.plugin and output a plist of all localizable geometry in file.nib. ibtool --convert oldName-newName file.nib ibtool will change all classes with the exact name "oldName" to be named "newName" instead. ibtool --convert Old'*'-New file ibtool will change all classes with names beginning with "Old", such as "OldName", "OldClassName", and "OldClass", to instead begin with "New". For example, the previously mentioned classes would be changed to "NewName", "NewClassName", and "NewClass", respectively. ibtool --export export.plist file.nib ibtool will export properties specified in the export.plist, and output a plist mapping Object IDs to property values. ibtool --export-xliff generated.xlf inputDocument.xib ibtool will create an XLIFF document, "generated.xlf", that contains all of the localizable strings from the input document "inputDocument.xib". As the source and target languages were not specified, English will be used as the source language, and the target language will not be included. ibtool --export-xliff generated.xlf --source-language es --target-language fr inputDocument.xib ibtool will create an XLIFF document, "generated.xlf", that contains all of the localizable strings from the input document "inputDocument.xib". The source language of "generated.xlf" is marked as Spanish, and the target language is marked as French. ibtool --import-xliff translation.xlf --write translated.xib inputDocument.xib ibtool will apply the translations found in the XLIFF document "translation.xlf" to the input document "inputDocument.xib", and write the resulting translated document to "translated.xib". SEE ALSO plist(5) Apple Inc. Mar 19 2015 ibtool(1)
|
tty
|
The tty utility writes the name of the terminal attached to standard input to standard output. The name that is written is the string returned by ttyname(3). If the standard input is not a terminal, the message “not a tty” is written. The options are as follows: -s Do not write the terminal name; only the exit status is affected when this option is specified. The -s option is deprecated in favor of the “test -t 0” command. EXIT STATUS The tty utility exits 0 if the standard input is a terminal, 1 if the standard input is not a terminal, and >1 if an error occurs. SEE ALSO test(1), ttyname(3) STANDARDS The tty utility is expected to be IEEE Std 1003.2 (“POSIX.2”) compatible. HISTORY A tty command appeared in Version 1 AT&T UNIX. macOS 14.5 June 6, 1993 macOS 14.5
|
tty – return user's terminal name
|
tty [-s]
| null | null |
demandoc
|
The demandoc utility emits only the text portions of well-formed mdoc(7) and man(7) UNIX manual files. By default, demandoc parses standard input and outputs only text nodes, preserving line and column position. Escape sequences are omitted from the output. Its arguments are as follows: -w Output a word list. This outputs each word of text on its own line. A "word", in this case, refers to whitespace-delimited terms beginning with at least two letters and not consisting of any escape sequences. Words have their leading and trailing punctuation (double-quotes, sentence punctuation, etc.) stripped. file ... The input files. If a document is not well-formed, it is skipped. The -i, -k, -m, and -p flags are silently discarded for calling compatibility with the historical deroff. EXIT STATUS The demandoc utility exits with one of the following values: 0 No errors occurred. 6 An operating system error occurred, for example memory exhaustion or an error accessing input files. Such errors cause demandoc to exit at once, possibly in the middle of parsing or formatting a file. The output databases are corrupt and should be removed .
|
demandoc – emit only text of UNIX manuals
|
demandoc [-w] [file ...]
| null |
The traditional usage of demandoc is for spell-checking manuals on BSD. This is accomplished as follows (assuming British spelling): $ demandoc -w file.1 | spell -b SEE ALSO mandoc(1), man(7), mdoc(7) HISTORY demandoc replaces the historical deroff utility for handling modern man(7) and mdoc(7) documents. AUTHORS The demandoc utility was written by Kristaps Dzonsons <kristaps@bsd.lv>. macOS 14.5 September 12, 2014 macOS 14.5
|
git-receive-pack
|
Invoked by git send-pack and updates the repository with the information fed from the remote end. This command is usually not invoked directly by the end user. The UI for the protocol is on the git send-pack side, and the program pair is meant to be used to push updates to remote repository. For pull operations, see git-fetch-pack(1). The command allows for creation and fast-forwarding of sha1 refs (heads/tags) on the remote end (strictly speaking, it is the local end git-receive-pack runs, but to the user who is sitting at the send-pack end, it is updating the remote. Confused?) There are other real-world examples of using update and post-update hooks found in the Documentation/howto directory. git-receive-pack honours the receive.denyNonFastForwards config option, which tells it if updates to a ref should be denied if they are not fast-forwards. A number of other receive.* config options are available to tweak its behavior, see git-config(1).
|
git-receive-pack - Receive what is pushed into the repository
|
git receive-pack <git-dir>
|
<git-dir> The repository to sync into. --http-backend-info-refs Used by git-http-backend(1) to serve up $GIT_URL/info/refs?service=git-receive-pack requests. See --http-backend-info-refs in git-upload-pack(1). PRE-RECEIVE HOOK Before any ref is updated, if $GIT_DIR/hooks/pre-receive file exists and is executable, it will be invoked once with no parameters. The standard input of the hook will be one line per ref to be updated: sha1-old SP sha1-new SP refname LF The refname value is relative to $GIT_DIR; e.g. for the master head this is "refs/heads/master". The two sha1 values before each refname are the object names for the refname before and after the update. Refs to be created will have sha1-old equal to 0{40}, while refs to be deleted will have sha1-new equal to 0{40}, otherwise sha1-old and sha1-new should be valid objects in the repository. When accepting a signed push (see git-push(1)), the signed push certificate is stored in a blob and an environment variable GIT_PUSH_CERT can be consulted for its object name. See the description of post-receive hook for an example. In addition, the certificate is verified using GPG and the result is exported with the following environment variables: GIT_PUSH_CERT_SIGNER The name and the e-mail address of the owner of the key that signed the push certificate. GIT_PUSH_CERT_KEY The GPG key ID of the key that signed the push certificate. GIT_PUSH_CERT_STATUS The status of GPG verification of the push certificate, using the same mnemonic as used in %G? format of git log family of commands (see git-log(1)). GIT_PUSH_CERT_NONCE The nonce string the process asked the signer to include in the push certificate. If this does not match the value recorded on the "nonce" header in the push certificate, it may indicate that the certificate is a valid one that is being replayed from a separate "git push" session. GIT_PUSH_CERT_NONCE_STATUS UNSOLICITED "git push --signed" sent a nonce when we did not ask it to send one. MISSING "git push --signed" did not send any nonce header. BAD "git push --signed" sent a bogus nonce. OK "git push --signed" sent the nonce we asked it to send. SLOP "git push --signed" sent a nonce different from what we asked it to send now, but in a previous session. See GIT_PUSH_CERT_NONCE_SLOP environment variable. GIT_PUSH_CERT_NONCE_SLOP "git push --signed" sent a nonce different from what we asked it to send now, but in a different session whose starting time is different by this many seconds from the current session. Only meaningful when GIT_PUSH_CERT_NONCE_STATUS says SLOP. Also read about receive.certNonceSlop variable in git-config(1). This hook is called before any refname is updated and before any fast-forward checks are performed. If the pre-receive hook exits with a non-zero exit status no updates will be performed, and the update, post-receive and post-update hooks will not be invoked either. This can be useful to quickly bail out if the update is not to be supported. See the notes on the quarantine environment below. UPDATE HOOK Before each ref is updated, if $GIT_DIR/hooks/update file exists and is executable, it is invoked once per ref, with three parameters: $GIT_DIR/hooks/update refname sha1-old sha1-new The refname parameter is relative to $GIT_DIR; e.g. for the master head this is "refs/heads/master". The two sha1 arguments are the object names for the refname before and after the update. Note that the hook is called before the refname is updated, so either sha1-old is 0{40} (meaning there is no such ref yet), or it should match what is recorded in refname. The hook should exit with non-zero status if it wants to disallow updating the named ref. Otherwise it should exit with zero. Successful execution (a zero exit status) of this hook does not ensure the ref will actually be updated, it is only a prerequisite. As such it is not a good idea to send notices (e.g. email) from this hook. Consider using the post-receive hook instead. POST-RECEIVE HOOK After all refs were updated (or attempted to be updated), if any ref update was successful, and if $GIT_DIR/hooks/post-receive file exists and is executable, it will be invoked once with no parameters. The standard input of the hook will be one line for each successfully updated ref: sha1-old SP sha1-new SP refname LF The refname value is relative to $GIT_DIR; e.g. for the master head this is "refs/heads/master". The two sha1 values before each refname are the object names for the refname before and after the update. Refs that were created will have sha1-old equal to 0{40}, while refs that were deleted will have sha1-new equal to 0{40}, otherwise sha1-old and sha1-new should be valid objects in the repository. The GIT_PUSH_CERT* environment variables can be inspected, just as in pre-receive hook, after accepting a signed push. Using this hook, it is easy to generate mails describing the updates to the repository. This example script sends one mail message per ref listing the commits pushed to the repository, and logs the push certificates of signed pushes with good signatures to a logger service: #!/bin/sh # mail out commit update information. while read oval nval ref do if expr "$oval" : '0*$' >/dev/null then echo "Created a new ref, with the following commits:" git rev-list --pretty "$nval" else echo "New commits:" git rev-list --pretty "$nval" "^$oval" fi | mail -s "Changes to ref $ref" commit-list@mydomain done # log signed push certificate, if any if test -n "${GIT_PUSH_CERT-}" && test ${GIT_PUSH_CERT_STATUS} = G then ( echo expected nonce is ${GIT_PUSH_NONCE} git cat-file blob ${GIT_PUSH_CERT} ) | mail -s "push certificate from $GIT_PUSH_CERT_SIGNER" push-log@mydomain fi exit 0 The exit code from this hook invocation is ignored, however a non-zero exit code will generate an error message. Note that it is possible for refname to not have sha1-new when this hook runs. This can easily occur if another user modifies the ref after it was updated by git-receive-pack, but before the hook was able to evaluate it. It is recommended that hooks rely on sha1-new rather than the current value of refname. POST-UPDATE HOOK After all other processing, if at least one ref was updated, and if $GIT_DIR/hooks/post-update file exists and is executable, then post-update will be called with the list of refs that have been updated. This can be used to implement any repository wide cleanup tasks. The exit code from this hook invocation is ignored; the only thing left for git-receive-pack to do at that point is to exit itself anyway. This hook can be used, for example, to run git update-server-info if the repository is packed and is served via a dumb transport. #!/bin/sh exec git update-server-info QUARANTINE ENVIRONMENT When receive-pack takes in objects, they are placed into a temporary "quarantine" directory within the $GIT_DIR/objects directory and migrated into the main object store only after the pre-receive hook has completed. If the push fails before then, the temporary directory is removed entirely. This has a few user-visible effects and caveats: 1. Pushes which fail due to problems with the incoming pack, missing objects, or due to the pre-receive hook will not leave any on-disk data. This is usually helpful to prevent repeated failed pushes from filling up your disk, but can make debugging more challenging. 2. Any objects created by the pre-receive hook will be created in the quarantine directory (and migrated only if it succeeds). 3. The pre-receive hook MUST NOT update any refs to point to quarantined objects. Other programs accessing the repository will not be able to see the objects (and if the pre-receive hook fails, those refs would become corrupted). For safety, any ref updates from within pre-receive are automatically rejected. SEE ALSO git-send-pack(1), gitnamespaces(7) GIT Part of the git(1) suite Git 2.41.0 2023-06-01 GIT-RECEIVE-PACK(1)
| null |
infotocap
|
infotocap looks in each given text file for terminfo descriptions. For each terminfo description found, an equivalent termcap description is written to standard output. Terminfo use capabilities are translated directly to termcap tc capabilities. -v print out tracing information on standard error as the program runs. -V print out the version of the program in use on standard error and exit. -1 cause the fields to print out one to a line. Otherwise, the fields will be printed several to a line to a maximum width of 60 characters. -w change the output to width characters. FILES /usr/share/terminfo Compiled terminal description database. NOTES This utility is actually a link to tic, running in -C mode. You can use other tic options such as -f and -x. SEE ALSO curses(3X), tic(1M), infocmp(1M), terminfo(5) This describes ncurses version 5.7 (patch 20081102). AUTHOR Eric S. Raymond <esr@snark.thyrsus.com> and Thomas E. Dickey <dickey@invisible-island.net> infotocap(1M)
|
infotocap - convert a terminfo description into a termcap description
|
infotocap [-vn width] [-V] [-1] [-w width] file . . .
| null | null |
pkill
|
The pgrep command searches the process table on the running system and prints the process IDs of all processes that match the criteria given on the command line. The pkill command searches the process table on the running system and signals all processes that match the criteria given on the command line. The following options are available: -F pidfile Restrict matches to a process whose PID is stored in the pidfile file. -G gid Restrict matches to processes with a real group ID in the comma-separated list gid. -I Request confirmation before attempting to signal each process. -L The pidfile file given for the -F option must be locked with the flock(2) syscall or created with pidfile(3). -P ppid Restrict matches to processes with a parent process ID in the comma-separated list ppid. -U uid Restrict matches to processes with a real user ID in the comma-separated list uid. -d delim Specify a delimiter to be printed between each process ID. The default is a newline. This option can only be used with the pgrep command. -a Include process ancestors in the match list. By default, the current pgrep or pkill process and all of its ancestors are excluded (unless -v is used). -f Match against full argument lists. The default is to match against process names. -g pgrp Restrict matches to processes with a process group ID in the comma-separated list pgrp. The value zero is taken to mean the process group ID of the running pgrep or pkill command. -i Ignore case distinctions in both the process table and the supplied pattern. -l Long output. For pgrep, print the process name in addition to the process ID for each matching process. If used in conjunction with -f, print the process ID and the full argument list for each matching process. For pkill, display the kill command used for each process killed. -n Select only the newest (most recently started) of the matching processes. -o Select only the oldest (least recently started) of the matching processes. -q For pgrep, Do not write anything to standard output. -t tty Restrict matches to processes associated with a terminal in the comma-separated list tty. Terminal names may be of the form ttyxx or the shortened form xx. A single dash (‘-’) matches processes not associated with a terminal. -u euid Restrict matches to processes with an effective user ID in the comma-separated list euid. -v Reverse the sense of the matching; display processes that do not match the given criteria. -x Require an exact match of the process name, or argument list if -f is given. The default is to match any substring. -signal A non-negative decimal number or symbolic signal name specifying the signal to be sent instead of the default TERM. This option is valid only when given as the first argument to pkill. If any pattern operands are specified, they are used as extended regular expressions to match the command name or full argument list of each process. Note that a running pgrep or pkill process will never consider itself as a potential match. EXIT STATUS The pgrep and pkill utilities return one of the following values upon exit: 0 One or more processes were matched. 1 No processes were matched. 2 Invalid options were specified on the command line. 3 An internal error occurred.
|
pgrep, pkill – find or signal processes by name
|
pgrep [-Lafilnoqvx] [-F pidfile] [-G gid] [-P ppid] [-U uid] [-d delim] [-g pgrp] [-t tty] [-u euid] pattern ... pkill [-signal] [-ILafilnovx] [-F pidfile] [-G gid] [-P ppid] [-U uid] [-g pgrp] [-t tty] [-u euid] pattern ...
| null |
Show the pid of the process holding the /tmp/.X0-lock pid file: $ pgrep -F /tmp/.X0-lock 1211 Show long output for firefox processes: $ pgrep -l firefox 1312 firefox 1309 firefox 1288 firefox 1280 firefox 1279 firefox 1278 firefox 1277 firefox 1264 firefox Same as above but just showing the pid of the most recent process: $ pgrep -n firefox 1312 Look for vim processes. Match against the full argument list: $ pgrep -f vim 44968 30790 Same as above but matching against the ‘list’ word and showing the full argument list: $ pgrep -f -l list 30790 vim list.txt Send SIGSTOP signal to processes that are an exact match: $ pkill -SIGSTOP -f -x "vim list.txt" Without -f names over 19 characters will silently fail: $ vim this_is_a_very_long_file_name & [1] 36689 $ [1]+ Stopped vim this_is_a_very_long_file_name $ pgrep "vim this" $ Same as above using the -f flag: $ pgrep -f "vim this" 36689 SEE ALSO kill(1), killall(1), ps(1), flock(2), kill(2), sigaction(2), pidfile(3), re_format(7) HISTORY The pkill and pgrep utilities first appeared in NetBSD 1.6. They are modelled after utilities of the same name that appeared in Sun Solaris 7. They made their first appearance in FreeBSD 5.3. AUTHORS Andrew Doran <ad@NetBSD.org> macOS 14.5 October 5, 2020 macOS 14.5
|
cups-config
|
The cups-config command allows application developers to determine the necessary command-line options for the compiler and linker, as well as the installation directories for filters, configuration files, and drivers. All values are reported to the standard output.
|
cups-config - get cups api, compiler, directory, and link information.
|
cups-config --api-version cups-config --build cups-config --cflags cups-config --datadir cups-config --help cups-config --ldflags cups-config [ --image ] [ --static ] --libs cups-config --serverbin cups-config --serverroot cups-config --version
|
The cups-config command accepts the following command-line options: --api-version Reports the current API version (major.minor). --build Reports a system-specific build number. --cflags Reports the necessary compiler options. --datadir Reports the default CUPS data directory. --help Reports the program usage message. --ldflags Reports the necessary linker options. --libs Reports the necessary libraries to link to. --serverbin Reports the default CUPS binary directory, where filters and backends are stored. --serverroot Reports the default CUPS configuration file directory. --static When used with --libs, reports the static libraries instead of the default (shared) libraries. --version Reports the full version number of the CUPS installation (major.minor.patch).
|
Show the currently installed version of CUPS: cups-config --version Compile a simple one-file CUPS filter: cc `cups-config --cflags --ldflags` -o filter filter.c \ `cups-config --libs` DEPRECATED OPTIONS The following options are deprecated but continue to work for backwards compatibility: --image Formerly used to add the CUPS imaging library to the list of libraries. SEE ALSO cups(1), CUPS Online Help (http://localhost:631/help) COPYRIGHT Copyright © 2007-2019 by Apple Inc. 26 April 2019 CUPS cups-config(1)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.