command
stringlengths
1
42
description
stringlengths
29
182k
name
stringlengths
7
64.9k
synopsis
stringlengths
4
85.3k
options
stringclasses
593 values
examples
stringclasses
455 values
lzmainfo
lzmainfo shows information stored in the .lzma file header. It reads the first 13 bytes from the specified file, decodes the header, and prints it to standard output in human readable format. If no files are given or file is -, standard input is read. Usually the most interesting information is the uncompressed size and the dictionary size. Uncompressed size can be shown only if the file is in the non-streamed .lzma format variant. The amount of memory required to decompress the file is a few dozen kilobytes plus the dictionary size. lzmainfo is included in XZ Utils primarily for backward compatibility with LZMA Utils. EXIT STATUS 0 All is good. 1 An error occurred. BUGS lzmainfo uses MB while the correct suffix would be MiB (2^20 bytes). This is to keep the output compatible with LZMA Utils. SEE ALSO xz(1) Tukaani 2013-06-30 LZMAINFO(1)
lzmainfo - show information stored in the .lzma file header
lzmainfo [--help] [--version] [file...]
null
null
unzstd
zstd is a fast lossless compression algorithm and data compression tool, with command line syntax similar to gzip(1) and xz(1). It is based on the LZ77 family, with further FSE & huff0 entropy stages. zstd offers highly configurable compression speed, from fast modes at > 200 MB/s per core, to strong modes with excellent compression ratios. It also features a very fast decoder, with speeds > 500 MB/s per core, which remains roughly stable at all compression settings. zstd command line syntax is generally similar to gzip, but features the following few differences: • Source files are preserved by default. It´s possible to remove them automatically by using the --rm command. • When compressing a single file, zstd displays progress notifications and result summary by default. Use -q to turn them off. • zstd displays a short help page when command line is an error. Use -q to turn it off. • zstd does not accept input from console, though it does accept stdin when it´s not the console. • zstd does not store the input´s filename or attributes, only its contents. zstd processes each file according to the selected operation mode. If no files are given or file is -, zstd reads from standard input and writes the processed data to standard output. zstd will refuse to write compressed data to standard output if it is a terminal: it will display an error message and skip the file. Similarly, zstd will refuse to read compressed data from standard input if it is a terminal. Unless --stdout or -o is specified, files are written to a new file whose name is derived from the source file name: • When compressing, the suffix .zst is appended to the source filename to get the target filename. • When decompressing, the .zst suffix is removed from the source filename to get the target filename Concatenation with .zst Files It is possible to concatenate multiple .zst files. zstd will decompress such agglomerated file as if it was a single .zst file.
zstd - zstd, zstdmt, unzstd, zstdcat - Compress or decompress .zst files
zstd [OPTIONS] [-|INPUT-FILE] [-o OUTPUT-FILE] zstdmt is equivalent to zstd -T0 unzstd is equivalent to zstd -d zstdcat is equivalent to zstd -dcf
Integer Suffixes and Special Values In most places where an integer argument is expected, an optional suffix is supported to easily indicate large integers. There must be no space between the integer and the suffix. KiB Multiply the integer by 1,024 (2^10). Ki, K, and KB are accepted as synonyms for KiB. MiB Multiply the integer by 1,048,576 (2^20). Mi, M, and MB are accepted as synonyms for MiB. Operation Mode If multiple operation mode options are given, the last one takes effect. -z, --compress Compress. This is the default operation mode when no operation mode option is specified and no other operation mode is implied from the command name (for example, unzstd implies --decompress). -d, --decompress, --uncompress Decompress. -t, --test Test the integrity of compressed files. This option is equivalent to --decompress --stdout > /dev/null, decompressed data is discarded and checksummed for errors. No files are created or removed. -b# Benchmark file(s) using compression level #. See BENCHMARK below for a description of this operation. --train FILES Use FILES as a training set to create a dictionary. The training set should contain a lot of small files (> 100). See DICTIONARY BUILDER below for a description of this operation. -l, --list Display information related to a zstd compressed file, such as size, ratio, and checksum. Some of these fields may not be available. This command´s output can be augmented with the -v modifier. Operation Modifiers • -#: selects # compression level [1-19] (default: 3). Higher compression levels generally produce higher compression ratio at the expense of speed and memory. A rough rule of thumb is that compression speed is expected to be divided by 2 every 2 levels. Technically, each level is mapped to a set of advanced parameters (that can also be modified individually, see below). Because the compressor´s behavior highly depends on the content to compress, there´s no guarantee of a smooth progression from one level to another. • --ultra: unlocks high compression levels 20+ (maximum 22), using a lot more memory. Note that decompression will also require more memory when using these levels. • --fast[=#]: switch to ultra-fast compression levels. If =# is not present, it defaults to 1. The higher the value, the faster the compression speed, at the cost of some compression ratio. This setting overwrites compression level if one was set previously. Similarly, if a compression level is set after --fast, it overrides it. • -T#, --threads=#: Compress using # working threads (default: 1). If # is 0, attempt to detect and use the number of physical CPU cores. In all cases, the nb of threads is capped to ZSTDMT_NBWORKERS_MAX, which is either 64 in 32-bit mode, or 256 for 64-bit environments. This modifier does nothing if zstd is compiled without multithread support. • --single-thread: Use a single thread for both I/O and compression. As compression is serialized with I/O, this can be slightly slower. Single-thread mode features significantly lower memory usage, which can be useful for systems with limited amount of memory, such as 32-bit systems. Note 1: this mode is the only available one when multithread support is disabled. Note 2: this mode is different from -T1, which spawns 1 compression thread in parallel with I/O. Final compressed result is also slightly different from -T1. • --auto-threads={physical,logical} (default: physical): When using a default amount of threads via -T0, choose the default based on the number of detected physical or logical cores. • --adapt[=min=#,max=#]: zstd will dynamically adapt compression level to perceived I/O conditions. Compression level adaptation can be observed live by using command -v. Adaptation can be constrained between supplied min and max levels. The feature works when combined with multi-threading and --long mode. It does not work with --single-thread. It sets window size to 8 MiB by default (can be changed manually, see wlog). Due to the chaotic nature of dynamic adaptation, compressed result is not reproducible. Note: at the time of this writing, --adapt can remain stuck at low speed when combined with multiple worker threads (>=2). • --long[=#]: enables long distance matching with # windowLog, if # is not present it defaults to 27. This increases the window size (windowLog) and memory usage for both the compressor and decompressor. This setting is designed to improve the compression ratio for files with long matches at a large distance. Note: If windowLog is set to larger than 27, --long=windowLog or --memory=windowSize needs to be passed to the decompressor. • -D DICT: use DICT as Dictionary to compress or decompress FILE(s) • --patch-from FILE: Specify the file to be used as a reference point for zstd´s diff engine. This is effectively dictionary compression with some convenient parameter selection, namely that windowSize > srcSize. Note: cannot use both this and -D together. Note: --long mode will be automatically activated if chainLog < fileLog (fileLog being the windowLog required to cover the whole file). You can also manually force it. Note: for all levels, you can use --patch-from in --single-thread mode to improve compression ratio at the cost of speed. Note: for level 19, you can get increased compression ratio at the cost of speed by specifying --zstd=targetLength= to be something large (i.e. 4096), and by setting a large --zstd=chainLog=. • --rsyncable: zstd will periodically synchronize the compression state to make the compressed file more rsync-friendly. There is a negligible impact to compression ratio, and a potential impact to compression speed, perceptible at higher speeds, for example when combining --rsyncable with many parallel worker threads. This feature does not work with --single-thread. You probably don´t want to use it with long range mode, since it will decrease the effectiveness of the synchronization points, but your mileage may vary. • -C, --[no-]check: add integrity check computed from uncompressed data (default: enabled) • --[no-]content-size: enable / disable whether or not the original size of the file is placed in the header of the compressed file. The default option is --content-size (meaning that the original size will be placed in the header). • --no-dictID: do not store dictionary ID within frame header (dictionary compression). The decoder will have to rely on implicit knowledge about which dictionary to use, it won´t be able to check if it´s correct. • -M#, --memory=#: Set a memory usage limit. By default, zstd uses 128 MiB for decompression as the maximum amount of memory the decompressor is allowed to use, but you can override this manually if need be in either direction (i.e. you can increase or decrease it). This is also used during compression when using with --patch-from=. In this case, this parameter overrides that maximum size allowed for a dictionary. (128 MiB). Additionally, this can be used to limit memory for dictionary training. This parameter overrides the default limit of 2 GiB. zstd will load training samples up to the memory limit and ignore the rest. • --stream-size=#: Sets the pledged source size of input coming from a stream. This value must be exact, as it will be included in the produced frame header. Incorrect stream sizes will cause an error. This information will be used to better optimize compression parameters, resulting in better and potentially faster compression, especially for smaller source sizes. • --size-hint=#: When handling input from a stream, zstd must guess how large the source size will be when optimizing compression parameters. If the stream size is relatively small, this guess may be a poor one, resulting in a higher compression ratio than expected. This feature allows for controlling the guess when needed. Exact guesses result in better compression ratios. Overestimates result in slightly degraded compression ratios, while underestimates may result in significant degradation. • --target-compressed-block-size=#: Attempt to produce compressed blocks of approximately this size. This will split larger blocks in order to approach this target. This feature is notably useful for improved latency, when the receiver can leverage receiving early incomplete data. This parameter defines a loose target: compressed blocks will target this size "on average", but individual blocks can still be larger or smaller. Enabling this feature can decrease compression speed by up to ~10% at level 1. Higher levels will see smaller relative speed regression, becoming invisible at higher settings. • -f, --force: disable input and output checks. Allows overwriting existing files, input from console, output to stdout, operating on links, block devices, etc. During decompression and when the output destination is stdout, pass-through unrecognized formats as-is. • -c, --stdout: write to standard output (even if it is the console); keep original files (disable --rm). • -o FILE: save result into FILE. Note that this operation is in conflict with -c. If both operations are present on the command line, the last expressed one wins. • --[no-]sparse: enable / disable sparse FS support, to make files with many zeroes smaller on disk. Creating sparse files may save disk space and speed up decompression by reducing the amount of disk I/O. default: enabled when output is into a file, and disabled when output is stdout. This setting overrides default and can force sparse mode over stdout. • --[no-]pass-through enable / disable passing through uncompressed files as-is. During decompression when pass-through is enabled, unrecognized formats will be copied as-is from the input to the output. By default, pass-through will occur when the output destination is stdout and the force (-f) option is set. • --rm: remove source file(s) after successful compression or decompression. This command is silently ignored if output is stdout. If used in combination with -o, triggers a confirmation prompt (which can be silenced with -f), as this is a destructive operation. • -k, --keep: keep source file(s) after successful compression or decompression. This is the default behavior. • -r: operate recursively on directories. It selects all files in the named directory and all its subdirectories. This can be useful both to reduce command line typing, and to circumvent shell expansion limitations, when there are a lot of files and naming breaks the maximum size of a command line. • --filelist FILE read a list of files to process as content from FILE. Format is compatible with ls output, with one file per line. • --output-dir-flat DIR: resulting files are stored into target DIR directory, instead of same directory as origin file. Be aware that this command can introduce name collision issues, if multiple files, from different directories, end up having the same name. Collision resolution ensures first file with a given name will be present in DIR, while in combination with -f, the last file will be present instead. • --output-dir-mirror DIR: similar to --output-dir-flat, the output files are stored underneath target DIR directory, but this option will replicate input directory hierarchy into output DIR. If input directory contains "..", the files in this directory will be ignored. If input directory is an absolute directory (i.e. "/var/tmp/abc"), it will be stored into the "output-dir/var/tmp/abc". If there are multiple input files or directories, name collision resolution will follow the same rules as --output-dir-flat. • --format=FORMAT: compress and decompress in other formats. If compiled with support, zstd can compress to or decompress from other compression algorithm formats. Possibly available options are zstd, gzip, xz, lzma, and lz4. If no such format is provided, zstd is the default. • -h/-H, --help: display help/long help and exit • -V, --version: display version number and immediately exit. note that, since it exits, flags specified after -V are effectively ignored. Advanced: -vV also displays supported formats. -vvV also displays POSIX support. -qV will only display the version number, suitable for machine reading. • -v, --verbose: verbose mode, display more information • -q, --quiet: suppress warnings, interactivity, and notifications. specify twice to suppress errors too. • --no-progress: do not display the progress bar, but keep all other messages. • --show-default-cparams: shows the default compression parameters that will be used for a particular input file, based on the provided compression level and the input size. If the provided file is not a regular file (e.g. a pipe), this flag will output the parameters used for inputs of unknown size. • --exclude-compressed: only compress files that are not already compressed. • --: All arguments after -- are treated as files gzip Operation Modifiers When invoked via a gzip symlink, zstd will support further options that intend to mimic the gzip behavior: -n, --no-name do not store the original filename and timestamps when compressing a file. This is the default behavior and hence a no-op. --best alias to the option -9. Environment Variables Employing environment variables to set parameters has security implications. Therefore, this avenue is intentionally limited. Only ZSTD_CLEVEL and ZSTD_NBTHREADS are currently supported. They set the default compression level and number of threads to use during compression, respectively. ZSTD_CLEVEL can be used to set the level between 1 and 19 (the "normal" range). If the value of ZSTD_CLEVEL is not a valid integer, it will be ignored with a warning message. ZSTD_CLEVEL just replaces the default compression level (3). ZSTD_NBTHREADS can be used to set the number of threads zstd will attempt to use during compression. If the value of ZSTD_NBTHREADS is not a valid unsigned integer, it will be ignored with a warning message. ZSTD_NBTHREADS has a default value of (1), and is capped at ZSTDMT_NBWORKERS_MAX==200. zstd must be compiled with multithread support for this variable to have any effect. They can both be overridden by corresponding command line arguments: -# for compression level and -T# for number of compression threads. ADVANCED COMPRESSION OPTIONS zstd provides 22 predefined regular compression levels plus the fast levels. A compression level is translated internally into multiple advanced parameters that control the behavior of the compressor (one can observe the result of this translation with --show-default-cparams). These advanced parameters can be overridden using advanced compression options. --zstd[=options]: The options are provided as a comma-separated list. You may specify only the options you want to change and the rest will be taken from the selected or default compression level. The list of available options: strategy=strat, strat=strat Specify a strategy used by a match finder. There are 9 strategies numbered from 1 to 9, from fastest to strongest: 1=ZSTD_fast, 2=ZSTD_dfast, 3=ZSTD_greedy, 4=ZSTD_lazy, 5=ZSTD_lazy2, 6=ZSTD_btlazy2, 7=ZSTD_btopt, 8=ZSTD_btultra, 9=ZSTD_btultra2. windowLog=wlog, wlog=wlog Specify the maximum number of bits for a match distance. The higher number of increases the chance to find a match which usually improves compression ratio. It also increases memory requirements for the compressor and decompressor. The minimum wlog is 10 (1 KiB) and the maximum is 30 (1 GiB) on 32-bit platforms and 31 (2 GiB) on 64-bit platforms. Note: If windowLog is set to larger than 27, --long=windowLog or --memory=windowSize needs to be passed to the decompressor. hashLog=hlog, hlog=hlog Specify the maximum number of bits for a hash table. Bigger hash tables cause fewer collisions which usually makes compression faster, but requires more memory during compression. The minimum hlog is 6 (64 entries / 256 B) and the maximum is 30 (1B entries / 4 GiB). chainLog=clog, clog=clog Specify the maximum number of bits for the secondary search structure, whose form depends on the selected strategy. Higher numbers of bits increases the chance to find a match which usually improves compression ratio. It also slows down compression speed and increases memory requirements for compression. This option is ignored for the ZSTD_fast strategy, which only has the primary hash table. The minimum clog is 6 (64 entries / 256 B) and the maximum is 29 (512M entries / 2 GiB) on 32-bit platforms and 30 (1B entries / 4 GiB) on 64-bit platforms. searchLog=slog, slog=slog Specify the maximum number of searches in a hash chain or a binary tree using logarithmic scale. More searches increases the chance to find a match which usually increases compression ratio but decreases compression speed. The minimum slog is 1 and the maximum is ´windowLog´ - 1. minMatch=mml, mml=mml Specify the minimum searched length of a match in a hash table. Larger search lengths usually decrease compression ratio but improve decompression speed. The minimum mml is 3 and the maximum is 7. targetLength=tlen, tlen=tlen The impact of this field vary depending on selected strategy. For ZSTD_btopt, ZSTD_btultra and ZSTD_btultra2, it specifies the minimum match length that causes match finder to stop searching. A larger targetLength usually improves compression ratio but decreases compression speed. For ZSTD_fast, it triggers ultra-fast mode when > 0. The value represents the amount of data skipped between match sampling. Impact is reversed: a larger targetLength increases compression speed but decreases compression ratio. For all other strategies, this field has no impact. The minimum tlen is 0 and the maximum is 128 KiB. overlapLog=ovlog, ovlog=ovlog Determine overlapSize, amount of data reloaded from previous job. This parameter is only available when multithreading is enabled. Reloading more data improves compression ratio, but decreases speed. The minimum ovlog is 0, and the maximum is 9. 1 means "no overlap", hence completely independent jobs. 9 means "full overlap", meaning up to windowSize is reloaded from previous job. Reducing ovlog by 1 reduces the reloaded amount by a factor 2. For example, 8 means "windowSize/2", and 6 means "windowSize/8". Value 0 is special and means "default": ovlog is automatically determined by zstd. In which case, ovlog will range from 6 to 9, depending on selected strat. ldmHashLog=lhlog, lhlog=lhlog Specify the maximum size for a hash table used for long distance matching. This option is ignored unless long distance matching is enabled. Bigger hash tables usually improve compression ratio at the expense of more memory during compression and a decrease in compression speed. The minimum lhlog is 6 and the maximum is 30 (default: 20). ldmMinMatch=lmml, lmml=lmml Specify the minimum searched length of a match for long distance matching. This option is ignored unless long distance matching is enabled. Larger/very small values usually decrease compression ratio. The minimum lmml is 4 and the maximum is 4096 (default: 64). ldmBucketSizeLog=lblog, lblog=lblog Specify the size of each bucket for the hash table used for long distance matching. This option is ignored unless long distance matching is enabled. Larger bucket sizes improve collision resolution but decrease compression speed. The minimum lblog is 1 and the maximum is 8 (default: 3). ldmHashRateLog=lhrlog, lhrlog=lhrlog Specify the frequency of inserting entries into the long distance matching hash table. This option is ignored unless long distance matching is enabled. Larger values will improve compression speed. Deviating far from the default value will likely result in a decrease in compression ratio. The default value is wlog - lhlog. Example The following parameters sets advanced compression options to something similar to predefined level 19 for files bigger than 256 KB: --zstd=wlog=23,clog=23,hlog=22,slog=6,mml=3,tlen=48,strat=6 -B#: Specify the size of each compression job. This parameter is only available when multi-threading is enabled. Each compression job is run in parallel, so this value indirectly impacts the nb of active threads. Default job size varies depending on compression level (generally 4 * windowSize). -B# makes it possible to manually select a custom size. Note that job size must respect a minimum value which is enforced transparently. This minimum is either 512 KB, or overlapSize, whichever is largest. Different job sizes will lead to non-identical compressed frames. DICTIONARY BUILDER zstd offers dictionary compression, which greatly improves efficiency on small files and messages. It´s possible to train zstd with a set of samples, the result of which is saved into a file called a dictionary. Then, during compression and decompression, reference the same dictionary, using command -D dictionaryFileName. Compression of small files similar to the sample set will be greatly improved. --train FILEs Use FILEs as training set to create a dictionary. The training set should ideally contain a lot of samples (> 100), and weight typically 100x the target dictionary size (for example, ~10 MB for a 100 KB dictionary). --train can be combined with -r to indicate a directory rather than listing all the files, which can be useful to circumvent shell expansion limits. Since dictionary compression is mostly effective for small files, the expectation is that the training set will only contain small files. In the case where some samples happen to be large, only the first 128 KiB of these samples will be used for training. --train supports multithreading if zstd is compiled with threading support (default). Additional advanced parameters can be specified with --train-fastcover. The legacy dictionary builder can be accessed with --train-legacy. The slower cover dictionary builder can be accessed with --train-cover. Default --train is equivalent to --train-fastcover=d=8,steps=4. -o FILE Dictionary saved into FILE (default name: dictionary). --maxdict=# Limit dictionary to specified size (default: 112640 bytes). As usual, quantities are expressed in bytes by default, and it´s possible to employ suffixes (like KB or MB) to specify larger values. -# Use # compression level during training (optional). Will generate statistics more tuned for selected compression level, resulting in a small compression ratio improvement for this level. -B# Split input files into blocks of size # (default: no split) -M#, --memory=# Limit the amount of sample data loaded for training (default: 2 GB). Note that the default (2 GB) is also the maximum. This parameter can be useful in situations where the training set size is not well controlled and could be potentially very large. Since speed of the training process is directly correlated to the size of the training sample set, a smaller sample set leads to faster training. In situations where the training set is larger than maximum memory, the CLI will randomly select samples among the available ones, up to the maximum allowed memory budget. This is meant to improve dictionary relevance by mitigating the potential impact of clustering, such as selecting only files from the beginning of a list sorted by modification date, or sorted by alphabetical order. The randomization process is deterministic, so training of the same list of files with the same parameters will lead to the creation of the same dictionary. --dictID=# A dictionary ID is a locally unique ID. The decoder will use this value to verify it is using the right dictionary. By default, zstd will create a 4-bytes random number ID. It´s possible to provide an explicit number ID instead. It´s up to the dictionary manager to not assign twice the same ID to 2 different dictionaries. Note that short numbers have an advantage: an ID < 256 will only need 1 byte in the compressed frame header, and an ID < 65536 will only need 2 bytes. This compares favorably to 4 bytes default. Note that RFC8878 reserves IDs less than 32768 and greater than or equal to 2^31, so they should not be used in public. --train-cover[=k#,d=#,steps=#,split=#,shrink[=#]] Select parameters for the default dictionary builder algorithm named cover. If d is not specified, then it tries d = 6 and d = 8. If k is not specified, then it tries steps values in the range [50, 2000]. If steps is not specified, then the default value of 40 is used. If split is not specified or split <= 0, then the default value of 100 is used. Requires that d <= k. If shrink flag is not used, then the default value for shrinkDict of 0 is used. If shrink is not specified, then the default value for shrinkDictMaxRegression of 1 is used. Selects segments of size k with highest score to put in the dictionary. The score of a segment is computed by the sum of the frequencies of all the subsegments of size d. Generally d should be in the range [6, 8], occasionally up to 16, but the algorithm will run faster with d <= 8. Good values for k vary widely based on the input data, but a safe range is [2 * d, 2000]. If split is 100, all input samples are used for both training and testing to find optimal d and k to build dictionary. Supports multithreading if zstd is compiled with threading support. Having shrink enabled takes a truncated dictionary of minimum size and doubles in size until compression ratio of the truncated dictionary is at most shrinkDictMaxRegression% worse than the compression ratio of the largest dictionary. Examples: zstd --train-cover FILEs zstd --train-cover=k=50,d=8 FILEs zstd --train-cover=d=8,steps=500 FILEs zstd --train-cover=k=50 FILEs zstd --train-cover=k=50,split=60 FILEs zstd --train-cover=shrink FILEs zstd --train-cover=shrink=2 FILEs --train-fastcover[=k#,d=#,f=#,steps=#,split=#,accel=#] Same as cover but with extra parameters f and accel and different default value of split If split is not specified, then it tries split = 75. If f is not specified, then it tries f = 20. Requires that 0 < f < 32. If accel is not specified, then it tries accel = 1. Requires that 0 < accel <= 10. Requires that d = 6 or d = 8. f is log of size of array that keeps track of frequency of subsegments of size d. The subsegment is hashed to an index in the range [0,2^f - 1]. It is possible that 2 different subsegments are hashed to the same index, and they are considered as the same subsegment when computing frequency. Using a higher f reduces collision but takes longer. Examples: zstd --train-fastcover FILEs zstd --train-fastcover=d=8,f=15,accel=2 FILEs --train-legacy[=selectivity=#] Use legacy dictionary builder algorithm with the given dictionary selectivity (default: 9). The smaller the selectivity value, the denser the dictionary, improving its efficiency but reducing its achievable maximum size. --train-legacy=s=# is also accepted. Examples: zstd --train-legacy FILEs zstd --train-legacy=selectivity=8 FILEs BENCHMARK The zstd CLI provides a benchmarking mode that can be used to easily find suitable compression parameters, or alternatively to benchmark a computer´s performance. Note that the results are highly dependent on the content being compressed. -b# benchmark file(s) using compression level # -e# benchmark file(s) using multiple compression levels, from -b# to -e# (inclusive) -d benchmark decompression speed only (requires providing an already zstd-compressed content) -i# minimum evaluation time, in seconds (default: 3s), benchmark mode only -B#, --block-size=# cut file(s) into independent chunks of size # (default: no chunking) --priority=rt set process priority to real-time (Windows) Output Format: CompressionLevel#Filename: InputSize -> OutputSize (CompressionRatio), CompressionSpeed, DecompressionSpeed Methodology: For both compression and decompression speed, the entire input is compressed/decompressed in-memory to measure speed. A run lasts at least 1 sec, so when files are small, they are compressed/decompressed several times per run, in order to improve measurement accuracy. SEE ALSO zstdgrep(1), zstdless(1), gzip(1), xz(1) The zstandard format is specified in Y. Collet, "Zstandard Compression and the ´application/zstd´ Media Type", https://www.ietf.org/rfc/rfc8878.txt, Internet RFC 8878 (February 2021). BUGS Report bugs at: https://github.com/facebook/zstd/issues AUTHOR Yann Collet zstd 1.5.6 March 2024 ZSTD(1)
null
grpc_csharp_plugin
null
null
null
null
null
bzfgrep
Bzgrep is used to invoke the grep on bzip2-compressed files. All options specified are passed directly to grep. If no file is specified, then the standard input is decompressed if necessary and fed to grep. Otherwise the given files are uncompressed if necessary and fed to grep. If bzgrep is invoked as bzegrep or bzfgrep then egrep or fgrep is used instead of grep. If the GREP environment variable is set, bzgrep uses it as the grep program to be invoked. For example: for sh: GREP=fgrep bzgrep string files for csh: (setenv GREP fgrep; bzgrep string files) AUTHOR Charles Levert (charles@comm.polymtl.ca). Adapted to bzip2 by Philippe Troin <phil@fifi.org> for Debian GNU/Linux. SEE ALSO grep(1), egrep(1), fgrep(1), bzdiff(1), bzmore(1), bzless(1), bzip2(1) BZGREP(1)
bzgrep, bzfgrep, bzegrep - search possibly bzip2 compressed files for a regular expression
bzgrep [ grep_options ] [ -e ] pattern filename... bzegrep [ egrep_options ] [ -e ] pattern filename... bzfgrep [ fgrep_options ] [ -e ] pattern filename...
null
null
anaconda
null
null
null
null
null
qmake
null
null
null
null
null
zfp
null
null
null
null
null
tabs
The tabs utility displays a series of characters that clear the hardware terminal tab settings then initialises tab stops at specified positions, and optionally adjusts the margin. In the first synopsis form, the tab stops set depend on the command line options used, and may be one of the predefined formats or at regular intervals. In the second synopsis form, tab stops are set at positions n1, n2, etc. If a position is preceded by a ‘+’, it is relative to the previous position set. No more than 20 positions may be specified. If no tab stops are specified, the “standard” UNIX tab width of 8 is used. The options are as follows: -n Set a tab stop every n columns. If n is 0, the tab stops are cleared but no new ones are set. -a Assembler format (columns 1, 10, 16, 36, 72). -a2 Assembler format (columns 1, 10, 16, 40, 72). -c COBOL normal format (columns 1, 8, 12, 16, 20, 55) -c2 COBOL compact format (columns 1, 6, 10, 14, 49) -c3 COBOL compact format (columns 1, 6, 10, 14, 18, 22, 26, 30, 34, 38, 42, 46, 50, 54, 58, 62, 67). -f FORTRAN format (columns 1, 7, 11, 15, 19, 23). -p PL/1 format (columns 1, 5, 9, 13, 17, 21, 25, 29, 33, 37, 41, 45, 49, 53, 57, 61). -s SNOBOL format (columns 1, 10, 55). -u Assembler format (columns 1, 12, 20, 44). +m[n], +[n] Set an n character left margin, or 10 if n is omitted. -T type Output escape sequence for the terminal type type. ENVIRONMENT The LANG, LC_ALL, LC_CTYPE and TERM environment variables affect the execution of tabs as described in environ(7). The -T option overrides the setting of the TERM environment variable. If neither TERM nor the -T option are present, tabs will fail. EXIT STATUS The tabs utility exits 0 on success, and >0 if an error occurs. SEE ALSO expand(1), stty(1), tput(1), unexpand(1), termcap(5) STANDARDS The tabs utility conforms to IEEE Std 1003.1-2001 (“POSIX.1”). HISTORY /etc/tabs appeared in Version 1 AT&T UNIX which set tabs every 8 on Teletype model 37 and compatible printing terminals when printed. A tabs utility that worked on many terminals appeared in PWB UNIX. This implementation was introduced in FreeBSD 5.0. BUGS The current termcap(5) database does not define the ‘ML’ (set left soft margin) capability for any terminals. macOS 14.5 September 9, 2021 macOS 14.5
tabs – set terminal tabs
tabs [-n | -a | -a2 | -c | -c2 | -c3 | -f | -p | -s | -u] [+m[n]] [-T type] tabs [-T type] [+[n]] n1[,n2,...]
null
null
msgconv
Converts a translation catalog to a different character encoding. Mandatory arguments to long options are mandatory for short options too. Input file location: INPUTFILE input PO file -D, --directory=DIRECTORY add DIRECTORY to list for input files search If no input file is given or if it is -, standard input is read. Output file location: -o, --output-file=FILE write output to specified file The results are written to standard output if no output file is specified or if it is -. Conversion target: -t, --to-code=NAME encoding for output The default encoding is the current locale's encoding. Input file syntax: -P, --properties-input input file is in Java .properties syntax --stringtable-input input file is in NeXTstep/GNUstep .strings syntax Output details: --color use colors and other text attributes always --color=WHEN use colors and other text attributes if WHEN. WHEN may be 'always', 'never', 'auto', or 'html'. --style=STYLEFILE specify CSS style rule file for --color -e, --no-escape do not use C escapes in output (default) -E, --escape use C escapes in output, no extended chars --force-po write PO file even if empty -i, --indent indented output style --no-location suppress '#: filename:line' lines -n, --add-location preserve '#: filename:line' lines (default) --strict strict Uniforum output style -p, --properties-output write out a Java .properties file --stringtable-output write out a NeXTstep/GNUstep .strings file -w, --width=NUMBER set output page width --no-wrap do not break long message lines, longer than the output page width, into several lines -s, --sort-output generate sorted output -F, --sort-by-file sort output by file location Informative output: -h, --help display this help and exit -V, --version output version information and exit AUTHOR Written by Bruno Haible. REPORTING BUGS Report bugs in the bug tracker at <https://savannah.gnu.org/projects/gettext> or by email to <bug-gettext@gnu.org>. COPYRIGHT Copyright © 2001-2023 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <https://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. SEE ALSO The full documentation for msgconv is maintained as a Texinfo manual. If the info and msgconv programs are properly installed at your site, the command info msgconv should give you access to the complete manual. GNU gettext-tools 0.22.5 February 2024 MSGCONV(1)
msgconv - character set conversion for message catalog
msgconv [OPTION] [INPUTFILE]
null
null
flask
null
null
null
null
null
qdoc
null
null
null
null
null
sip-sdist
null
null
null
null
null
grpc_node_plugin
null
null
null
null
null
gif2rgb
null
gif2rgb - convert images saved as GIF to 24-bit RGB triplets
gif2rgb [-v] [-1] [-c colors] [-s width height] [-o outfile] [-h] [gif-file]
-v Verbose mode (show progress). Enables printout of running scan lines. -1 Only one file in the format of RGBRGB... triplets (Each of R, G,B is a byte) is being read or written. This file size is 3 * Width * Height. If stdin is used for input or stdout for output, this option is implicitly applied. The default (if not `-1') is 3 files with the names OutFileName.R, OutFileName.G, OutFileName.B, each of which is Width * Height bytes. -c colors Specifies number of colors to use in RGB-to-GIF conversions, in bits per pixels, so '-c 8' actually specifies 256 colors (maximum and default). -s width height Sets RGB-to-GIF conversion mode and specifies the size of the image to read. -o specifies the name of the out file (see also `-1' above). -h Print one line of command line help, similar to Usage above. By default, convert a GIF input file to RGB triplets. If -s is specified, convert RGB input to a GIF. If no input file is given, gif2rgb will try to read data from stdin. AUTHOR Gershon Elber. GIFLIB 2 May 2012 GIF2RGB(1)
null
bzmore
Bzmore is a filter which allows examination of compressed or plain text files one screenful at a time on a soft-copy terminal. bzmore works on files compressed with bzip2 and also on uncompressed files. If a file does not exist, bzmore looks for a file of the same name with the addition of a .bz2 suffix. Bzmore normally pauses after each screenful, printing --More-- at the bottom of the screen. If the user then types a carriage return, one more line is displayed. If the user hits a space, another screenful is displayed. Other possibilities are enumerated later. Bzmore looks in the file /etc/termcap to determine terminal characteristics, and to determine the default window size. On a terminal capable of displaying 24 lines, the default window size is 22 lines. Other sequences which may be typed when bzmore pauses, and their effects, are as follows (i is an optional integer argument, defaulting to 1) : i<space> display i more lines, (or another screenful if no argument is given) ^D display 11 more lines (a ``scroll''). If i is given, then the scroll size is set to i. d same as ^D (control-D) iz same as typing a space except that i, if present, becomes the new window size. Note that the window size reverts back to the default at the end of the current file. is skip i lines and print a screenful of lines if skip i screenfuls and print a screenful of lines q or Q quit reading the current file; go on to the next (if any) e or q When the prompt --More--(Next file: file) is printed, this command causes bzmore to exit. s When the prompt --More--(Next file: file) is printed, this command causes bzmore to skip the next file and continue. = Display the current line number. i/expr search for the i-th occurrence of the regular expression expr. If the pattern is not found, bzmore goes on to the next file (if any). Otherwise, a screenful is displayed, starting two lines before the place where the expression was found. The user's erase and kill characters may be used to edit the regular expression. Erasing back past the first column cancels the search command. in search for the i-th occurrence of the last regular expression entered. !command invoke a shell with command. The character `!' in "command" are replaced with the previous shell command. The sequence "\!" is replaced by "!". :q or :Q quit reading the current file; go on to the next (if any) (same as q or Q). . (dot) repeat the previous command. The commands take effect immediately, i.e., it is not necessary to type a carriage return. Up to the time when the command character itself is given, the user may hit the line kill character to cancel the numerical argument being formed. In addition, the user may hit the erase character to redisplay the --More-- message. At any time when output is being sent to the terminal, the user can hit the quit key (normally control-\). Bzmore will stop sending output, and will display the usual --More-- prompt. The user may then enter one of the above commands in the normal manner. Unfortunately, some output is lost when this is done, due to the fact that any characters waiting in the terminal's output queue are flushed when the quit signal occurs. The terminal is set to noecho mode by this program so that the output can be continuous. What you type will thus not show on your terminal, except for the / and ! commands. If the standard output is not a teletype, then bzmore acts just like bzcat, except that a header is printed before each file. FILES /etc/termcap Terminal data base SEE ALSO more(1), less(1), bzip2(1), bzdiff(1), bzgrep(1) BZMORE(1)
bzmore, bzless - file perusal filter for crt viewing of bzip2 compressed text
bzmore [ name ... ] bzless [ name ... ] NOTE In the following description, bzless and less can be used interchangeably with bzmore and more.
null
null
streamlit.cmd
null
null
null
null
null
msggrep
Extracts all messages of a translation catalog that match a given pattern or belong to some given source files. Mandatory arguments to long options are mandatory for short options too. Input file location: INPUTFILE input PO file -D, --directory=DIRECTORY add DIRECTORY to list for input files search If no input file is given or if it is -, standard input is read. Output file location: -o, --output-file=FILE write output to specified file The results are written to standard output if no output file is specified or if it is -. Message selection: [-N SOURCEFILE]... [-M DOMAINNAME]... [-J MSGCTXT-PATTERN] [-K MSGID-PATTERN] [-T MSGSTR-PATTERN] [-C COMMENT-PATTERN] [-X EXTRACTED-COMMENT-PATTERN] A message is selected if it comes from one of the specified source files, or if it comes from one of the specified domains, or if -J is given and its context (msgctxt) matches MSGCTXT-PATTERN, or if -K is given and its key (msgid or msgid_plural) matches MSGID-PATTERN, or if -T is given and its translation (msgstr) matches MSGSTR-PATTERN, or if -C is given and the translator's comment matches COMMENT-PATTERN, or if -X is given and the extracted comment matches EXTRACTED-COMMENT-PATTERN. When more than one selection criterion is specified, the set of selected messages is the union of the selected messages of each criterion. MSGCTXT-PATTERN or MSGID-PATTERN or MSGSTR-PATTERN or COMMENT-PATTERN or EXTRACTED-COMMENT-PATTERN syntax: [-E | -F] [-e PATTERN | -f FILE]... PATTERNs are basic regular expressions by default, or extended regular expressions if -E is given, or fixed strings if -F is given. -N, --location=SOURCEFILE select messages extracted from SOURCEFILE -M, --domain=DOMAINNAME select messages belonging to domain DOMAINNAME -J, --msgctxt start of patterns for the msgctxt -K, --msgid start of patterns for the msgid -T, --msgstr start of patterns for the msgstr -C, --comment start of patterns for the translator's comment -X, --extracted-comment start of patterns for the extracted comment -E, --extended-regexp PATTERN is an extended regular expression -F, --fixed-strings PATTERN is a set of newline-separated strings -e, --regexp=PATTERN use PATTERN as a regular expression -f, --file=FILE obtain PATTERN from FILE -i, --ignore-case ignore case distinctions -v, --invert-match output only the messages that do not match any selection criterion Input file syntax: -P, --properties-input input file is in Java .properties syntax --stringtable-input input file is in NeXTstep/GNUstep .strings syntax Output details: --color use colors and other text attributes always --color=WHEN use colors and other text attributes if WHEN. WHEN may be 'always', 'never', 'auto', or 'html'. --style=STYLEFILE specify CSS style rule file for --color --no-escape do not use C escapes in output (default) --escape use C escapes in output, no extended chars --force-po write PO file even if empty --indent indented output style --no-location suppress '#: filename:line' lines -n, --add-location preserve '#: filename:line' lines (default) --strict strict Uniforum output style -p, --properties-output write out a Java .properties file --stringtable-output write out a NeXTstep/GNUstep .strings file -w, --width=NUMBER set output page width --no-wrap do not break long message lines, longer than the output page width, into several lines --sort-output generate sorted output --sort-by-file sort output by file location Informative output: -h, --help display this help and exit -V, --version output version information and exit AUTHOR Written by Bruno Haible. REPORTING BUGS Report bugs in the bug tracker at <https://savannah.gnu.org/projects/gettext> or by email to <bug-gettext@gnu.org>. COPYRIGHT Copyright © 2001-2023 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <https://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. SEE ALSO The full documentation for msggrep is maintained as a Texinfo manual. If the info and msggrep programs are properly installed at your site, the command info msggrep should give you access to the complete manual. GNU gettext-tools 0.22.5 February 2024 MSGGREP(1)
msggrep - pattern matching on message catalog
msggrep [OPTION] [INPUTFILE]
null
null
idle3.11
null
null
null
null
null
bzgrep
Bzgrep is used to invoke the grep on bzip2-compressed files. All options specified are passed directly to grep. If no file is specified, then the standard input is decompressed if necessary and fed to grep. Otherwise the given files are uncompressed if necessary and fed to grep. If bzgrep is invoked as bzegrep or bzfgrep then egrep or fgrep is used instead of grep. If the GREP environment variable is set, bzgrep uses it as the grep program to be invoked. For example: for sh: GREP=fgrep bzgrep string files for csh: (setenv GREP fgrep; bzgrep string files) AUTHOR Charles Levert (charles@comm.polymtl.ca). Adapted to bzip2 by Philippe Troin <phil@fifi.org> for Debian GNU/Linux. SEE ALSO grep(1), egrep(1), fgrep(1), bzdiff(1), bzmore(1), bzless(1), bzip2(1) BZGREP(1)
bzgrep, bzfgrep, bzegrep - search possibly bzip2 compressed files for a regular expression
bzgrep [ grep_options ] [ -e ] pattern filename... bzegrep [ egrep_options ] [ -e ] pattern filename... bzfgrep [ fgrep_options ] [ -e ] pattern filename...
null
null
libtool
The libtool command takes the specified input object files and creates a library for use with the link editor, ld(1). The library's name is specified by output (the argument to the -o flag). The input object files may be in any correct format that contains object files (``universal'' files, archives, object files). Libtool will not put any non-object input file into the output library (unlike ranlib, which allows this in the archives it operates on). When producing a ``universal'' file from objects of the same CPU type and differing CPU subtypes, libtool and ranlib create at most one library for each CPU type, rather than a separate library in a universal file for each of the unique pairings of CPU type and CPU subtype. Thus, the resulting CPU subtype for each library is the _ALL CPU subtype for that CPU type. This strategy strongly encourages the implementor of a library to create one library that chooses optimum code to run at run time, rather than at link time. Libtool can create either dynamically linked shared libraries, with -dynamic, or statically linked (archive) libraries, with -static. DYNAMICALLY LINKED SHARED LIBRARIES Dynamically linked libraries, unlike statically linked libraries, are Mach-O format files and not ar(5) format files. Dynamically linked libraries have two restrictions: No symbol may be defined in more than one object file and no common symbol can be used. To maximize sharing of a dynamically linked shared library the objects should be compiled with the -dynamic flag of cc(1) to produce indirect undefined references and position-independent code. To build a dynamically linked library, libtool, runs the link editor, ld(1), with -dylib once for each architecture present in the input objects and then lipo(1) to create a universal file if needed. ARCHIVE (or statically linked) LIBRARIES Libtool with -static is intended to replace ar(5) and ranlib. For backward compatibility, ranlib is still available, and it supports universal files. Ranlib adds or updates the table of contents to each archive so it can be linked by the link editor, ld(1). The table of contents is an archive member at the beginning of the archive that indicates which symbols are defined in which library members. Because ranlib rewrites the archive, sufficient temporary file space must be available in the file system that contains the current directory. Ranlib takes all correct forms of libraries (universal files containing archives, and simple archives) and updates the table of contents for all archives in the file. Ranlib also takes one common incorrect form of archive, an archive whose members are universal object files, adding or updating the table of contents and producing the library in correct form (a universal file containing multiple archives). The archive member name for a table of contents begins with ``__.SYMDEF''. Currently, there are two types of table of contents produced by libtool -static and ranlib and understood by the link editor, ld(1). These are explained below, under the -s and -a options.
libtool - create libraries ranlib - add or update the table of contents of archive libraries
libtool -static -o output [ -sacLTD ] [ - ] [ -arch_only arch_type ] [ -no_warning_for_no_symbols ] file... [-filelist listfile[,dirname]] libtool -dynamic -o output [ -install_name name ] [ -compatibility_version number ] [ -current_version number ] [ link editor flags ] [ -v ] [ -noall_load ] [ - ] [ -arch_only arch_type ] [ -V ] file... [-filelist listfile[,dirname]] ranlib [ -sactfqLT ] [ - ] archive...
The following options pertain to libtool only. @file Arguments beginning with @ are replaced by arguments read from the specified file, as an alternative to listing those arguments on the command line. The files simply contain libtool options and files separated by whitespace: spaces, tabs, and newlines. Characters can be escaped with a backslash (\), including whitespace characters and other backslashes. Also, arguments that include whitespace can be enclosed, wholly or in part, by single- or double-quote charcters. These files may contain @file references to additional files, although libtool will error on include cycles. If a file cannot be found, the original @file argument will remain in the argument list. -static Produce a statically linked (archive) library from the input files. This is the default. -dynamic Produce a dynamically linked shared library from the input files. -install_name name For a dynamic shared library, this specifies the file name the library will be installed in for programs that use it. If this is not specified the name specified by the -o output option will be used. -compatibility_version number For a dynamic shared library, this specifies the compatibility version number of the library. When a library is used the compatibility version is checked and if the user's version is greater that the library's version, an error message is printed and the using program exits. The format of number is X[.Y[.Z]] where X must be a positive non-zero number less than or equal to 65535, and .Y and .Z are optional and if present must be non- negative numbers less than or equal to 255. If this is not specified then it has a value of 0 and no checking is done when the library is used. -current_version number For dynamic shared library files this specifies the current version number of the library. The program using the library can obtain the current version of the library programmatically to determine exactly which version of the library it is using. The format of number is X[.Y[.Z]] where X must be a positive non-zero number less than or equal to 65535, and .Y and .Z are optional and if present must be non-negative numbers less than or equal to 255. If this is not specified then it has a value of 0. -noall_load For dynamic shared library files this specifies the the default behavior of loading all members of archives on the command line is not to be done. This option is used by the GNU compiler driver, cc(1), when used with it's -dynamiclib option. This is done to allow selective loading of the GNU's compiler's runtime support library, libcc_dynamic.a . link editor flags For a dynamic shared library the following ld(1) flags are accepted and passed through: -lx, -weak-lx, -search_paths_first -weak_library, -Ldir, -ysym, -usym, -initsym, -idefinition:indirect, -seg1addr, -segs_read_only_addr, -segs_read_write_addr, -seg_addr_table, -seg_addr_table_filename, -segprot, -segalign, -sectcreate, -sectorder, -sectorder_detail, -sectalign, -undefined, -read_only_relocs, -prebind, -prebind_all_twolevel_modules, -prebind_allow_overlap, -noprebind, -framework, -weak_framework, -umbrella, -allowable_client, -sub_umbrella, -sub_library, -F, -U, -Y, -Sn, -Si, -Sp, -S, -X, -x, -whyload, -all_load. -arch_errors_fatal, -dylib_file, -run_init_lazily, -final_output, -macosx_version_min, -multiply_defined, -multiply_defined_unused, -twolevel_namespace, -twolevel_namespace_hints, -flat_namespace, -nomultidefs, -headerpad, -headerpad_max_install_names, -weak_reference_mismatches, -M, -t, -no_arch_warnings, -single_module, -multi_module, -exported_symbols_list, -unexported_symbols_list, -m, -dead_strip, -no_dead_strip_inits_and_terms, -executable_path, -syslibroot, -no_uuid. See the ld(1) man page for details on these flags. The flag -image_base is a synonym for -seg1addr. -v Verbose mode, which prints the ld(1) commands and lipo(1) commands executed. -V Print the version of libtool. -filelist listfile[,dirname] The listfile contains a list of file names and is an alternative way of specifiying file names on the command line. The file names are listed one per line separated only by newlines (spaces and tabs are assumed to be part of the file name). If the optional directory name, dirname is specified then it is prepended to each name in the list file. -arch_only arch_type This option causes libtool to build a library only for the specified arch_type and ignores all other architectures in the input files. When building a dynamic library, if this is specified with a specific cpusubtype other than the family cpusubtype then libtool it does not use the ld(1) -force_cpusubtype_ALL flag and passes the -arch_only argument to ld(1) as the -arch flag so that the output is tagged with that cpusubtype. The following options pertain to the table of contents for an archive library, and apply to both libtool -static and ranlib: -s Produce the preferred type of table of contents, which results in faster link editing when linking with the archive. The order of the table of contents is sorted by symbol name. The library member name of this type of table of contents is ``__.SYMDEF SORTED''. This type of table of contents can only be produced when the library does not have multiple members that define the same symbol. This is the default. -a Produce the original type of table of contents, whose order is based on the order of the members in the archive. The library member name of this type of table of contents is ``__.SYMDEF''. This type of table of contents must be used when the library has multiple members that define the same symbol. -c Include common symbols as definitions with respect to the table of contents. This is seldom the intended behavior for linking from a library, as it forces the linking of a library member just because it uses an uninitialized global that is undefined at that point in the linking. This option is included only because this was the original behavior of ranlib. This option is not the default. -L Use the 4.4bsd archive extended format #1, which allows archive member names to be longer than 16 characters and have spaces in their names. This option is the default. -T Truncate archive member names to 16 characters and don't use the 4.4bsd extended format #1. This option is not the default. -f Warns when the output archive is universal and ar(1) will no longer be able to operate on it. -q Do nothing if a universal file would be created. -D When building a static library, set archive contents' user ids, group ids, dates, and file modes to reasonable defaults. This allows libraries created with identical input to be identical to each other, regardless of time of day, user, group, umask, and other aspects of the environment. For compatibility, the following ranlib option is accepted (but ignored): -t This option used to request that ranlib only ``touch'' the archives instead of modifying them. The option is now ignored, and the table of contents is rebuilt. Other options applying to both libtool and ranlib: - Treat all remaining arguments as names of files (or archives) and not as options. -no_warning_for_no_symbols Don't warn about file that have no symbols. -dependency_info path Write an Xcode dependency info file describing a successful build operation. This file describes the inputs directly or indirectly used to create the library or dylib. SEE ALSO ld(1), ar(1), otool(1), make(1), redo_prebinding(1), ar(5) BUGS With the way libraries used to be created, errors were possible if the library was modified with ar(1) and the table of contents was not updated by rerunning ranlib(1). So previously the link editor, ld(1), generated an error when the modification date of a library was more recent than the creation date of its table of contents. Unfortunately, this meant that you got the error even if you only copy the library. Since this error was found to be too much of a nuisance it was removed. So now it is possible again to get link errors if the library is modified and the table of contents is not updated. Apple Inc. June 23, 2020 LIBTOOL(1)
null
tracegen
null
null
null
null
null
qdarkstyle.example
null
null
null
null
null
conda-pack
null
null
null
null
null
canbusutil
null
null
null
null
null
makeconv
makeconv converts the ICU converter table convertertable into a binary file. The binary file has the same base name as convertertable but has a .cnv extension (instead of the typical .ucm extension of the convertertable file). This binary file can then be read directly by ICU, or used by pkgdata(1) for incorporation into a larger archive or library. The convertertable must be in the ICU ucm (Unicode Codepage Mapping) format in order to be understood by makeconv. The ICU ucm format is similar to the IBM NLTC upmap/tpmap/rpmap files. Comments in the convertertable are handled as follows. If a comment (starting with a `#' sign) that is after some text does contain the fallback indicator `|' then only the text starting with the `#' sign, and ending before the `|' sign, is ignored. Otherwise, or if the comment is the first thing on the line, the comment runs up to the end of the line. This special handling of comments is to accommodate the practice of putting fallback information in comments in the strict IBM NLTC ucmap format. Note that new converters will be automatically found by ICU after their installation in ICU's data directory. They do not need to be listed in the convrtrs.txt(5) converters aliases file in order to be available to applications using ICU. They do need to be listed there if one wants to give them aliases, or tags, though.
makeconv - compile a converter table
makeconv [ -h, -?, --help ] [ -c, --copyright ] [ -v, --verbose ] [ -d, --destdir destination ] convertertable ...
-h, -?, --help Print help about usage and exit. -c, --copyright Include a copyright notice in the binary data. -v, --verbose Display extra informative messages during execution. -d, --destdir destination Set the destination directory to destination. The default destination directory is specified by the environment variable ICU_DATA. CAVEATS If an existing converter table is changed and recompiled using makeconv, the resulting binary file must be packaged in the same way that it was packaged initially. For example, if converters were grouped together in an archive or a library with pkgdata(1), then the archive or library must be rebuilt with the new binary file. A standalone binary converter file will not take precedence over a packaged one. ENVIRONMENT ICU_DATA Specifies the directory containing ICU data. Defaults to ${prefix}/share/icu/68.1/. Some tools in ICU depend on the presence of the trailing slash. It is thus important to make sure that it is present if ICU_DATA is set. VERSION 68.1 COPYRIGHT Copyright (C) 2000 IBM, Inc. and others. SEE ALSO convrtrs.txt(5) pkgdata(1) ICU MANPAGE 16 April 2002 MAKECONV(1)
null
toe
With no options, toe lists all available terminal types by primary name with descriptions. File arguments specify the directories to be scanned; if no such arguments are given, your default terminfo directory is scanned. If you also specify the -h option, a directory header will be issued as each directory is entered. There are other options intended for use by terminfo file maintainers: -a report on all of the terminal databases which ncurses would search, rather than only the first one that it finds. If the -s is also given, toe adds a column to the report, showing (like conflict(1)) which entries which belong to a given terminal database. An "*" marks entries which differ, and "+" marks equivalent entries. -s sort the output by the entry names. -u file says to write a report to the standard output, listing dependencies in the given terminfo/termcap source file. The report condenses the `use' relation: each line consists of the primary name of a terminal that has use capabilities, followed by a colon, followed by the whitespace-separated primary names of all terminals which occur in those use capabilities, followed by a newline -U file says to write a report to the standard output, listing reverse dependencies in the given terminfo/termcap source file. The report reverses the `use' relation: each line consists of the primary name of a terminal that occurs in use capabilities, followed by a colon, followed by the whitespace-separated primary names of all terminals which depend on it, followed by a newline. -vn specifies that (verbose) output be written to standard error, showing toe's progress. The optional parameter n is a number from 1 to 10, interpreted as for tic(1M). -V reports the version of ncurses which was used in this program, and exits. FILES /usr/share/terminfo/?/* Compiled terminal description database. SEE ALSO tic(1M), infocmp(1M), captoinfo(1M), infotocap(1M), curses(3X), terminfo(5). This describes ncurses version 5.7 (patch 20081102). toe(1M)
toe - table of (terminfo) entries
toe [-v[n]] [-ahsuUV] file...
null
null
pcre2grep
pcre2grep searches files for character patterns, in the same way as other grep commands do, but it uses the PCRE2 regular expression library to support patterns that are compatible with the regular expressions of Perl 5. See pcre2syntax(3) for a quick-reference summary of pattern syntax, or pcre2pattern(3) for a full description of the syntax and semantics of the regular expressions that PCRE2 supports. Patterns, whether supplied on the command line or in a separate file, are given without delimiters. For example: pcre2grep Thursday /etc/motd If you attempt to use delimiters (for example, by surrounding a pattern with slashes, as is common in Perl scripts), they are interpreted as part of the pattern. Quotes can of course be used to delimit patterns on the command line because they are interpreted by the shell, and indeed quotes are required if a pattern contains white space or shell metacharacters. The first argument that follows any option settings is treated as the single pattern to be matched when neither -e nor -f is present. Conversely, when one or both of these options are used to specify patterns, all arguments are treated as path names. At least one of -e, -f, or an argument pattern must be provided. If no files are specified, pcre2grep reads the standard input. The standard input can also be referenced by a name consisting of a single hyphen. For example: pcre2grep some-pattern file1 - file3 By default, input files are searched line by line, so pattern assertions about the beginning and end of a subject string (^, $, \A, \Z, and \z) match at the beginning and end of each line. When a line matches a pattern, it is copied to the standard output, and if there is more than one file, the file name is output at the start of each line, followed by a colon. However, there are options that can change how pcre2grep behaves. For example, the -M option makes it possible to search for strings that span line boundaries. What defines a line boundary is controlled by the -N (--newline) option. The -h and -H options control whether or not file names are shown, and the -Z option changes the file name terminator to a zero byte. The amount of memory used for buffering files that are being scanned is controlled by parameters that can be set by the --buffer-size and --max-buffer-size options. The first of these sets the size of buffer that is obtained at the start of processing. If an input file contains very long lines, a larger buffer may be needed; this is handled by automatically extending the buffer, up to the limit specified by --max- buffer-size. The default values for these parameters can be set when pcre2grep is built; if nothing is specified, the defaults are set to 20KiB and 1MiB respectively. An error occurs if a line is too long and the buffer can no longer be expanded. The block of memory that is actually used is three times the "buffer size", to allow for buffering "before" and "after" lines. If the buffer size is too small, fewer than requested "before" and "after" lines may be output. When matching with a multiline pattern, the size of the buffer must be at least half of the maximum match expected or the pattern might fail to match. Patterns can be no longer than 8KiB or BUFSIZ bytes, whichever is the greater. BUFSIZ is defined in <stdio.h>. When there is more than one pattern (specified by the use of -e and/or -f), each pattern is applied to each line in the order in which they are defined, except that all the -e patterns are tried before the -f patterns. By default, as soon as one pattern matches a line, no further patterns are considered. However, if --colour (or --color) is used to colour the matching substrings, or if --only-matching, --file-offsets, --line- offsets, or --output is used to output only the part of the line that matched (either shown literally, or as an offset), the behaviour is different. In this situation, all the patterns are applied to the line. If there is more than one match, the one that begins nearest to the start of the subject is processed; if there is more than one match at that position, the one with the longest matching substring is processed; if the matching substrings are equal, the first match found is processed. Scanning with all the patterns resumes immediately following the match, so that later matches on the same line can be found. Note, however, that an overlapping match that starts in the middle of another match will not be processed. The above behaviour was changed at release 10.41 to be more compatible with GNU grep. In earlier releases, pcre2grep did not recognize matches from later patterns that were earlier in the subject. Patterns that can match an empty string are accepted, but empty string matches are never recognized. An example is the pattern "(super)?(man)?", in which all components are optional. This pattern finds all occurrences of both "super" and "man"; the output differs from matching with "super|man" when only the matching substrings are being shown. If the LC_ALL or LC_CTYPE environment variable is set, pcre2grep uses the value to set a locale when calling the PCRE2 library. The --locale option can be used to override this. SUPPORT FOR COMPRESSED FILES Compile-time options for pcre2grep can set it up to use libz or libbz2 for reading compressed files whose names end in .gz or .bz2, respectively. You can find out whether your pcre2grep binary has support for one or both of these file types by running it with the --help option. If the appropriate support is not present, all files are treated as plain text. The standard input is always so treated. If a file with a .gz or .bz2 extension is not in fact compressed, it is read as a plain text file. When input is from a compressed .gz or .bz2 file, the --line-buffered option is ignored. BINARY FILES By default, a file that contains a binary zero byte within the first 1024 bytes is identified as a binary file, and is processed specially. However, if the newline type is specified as NUL, that is, the line terminator is a binary zero, the test for a binary file is not applied. See the --binary-files option for a means of changing the way binary files are handled. BINARY ZEROS IN PATTERNS Patterns passed from the command line are strings that are terminated by a binary zero, so cannot contain internal zeros. However, patterns that are read from a file via the -f option may contain binary zeros.
pcre2grep - a grep with Perl-compatible regular expressions.
pcre2grep [options] [long options] [pattern] [path1 path2 ...]
The order in which some of the options appear can affect the output. For example, both the -H and -l options affect the printing of file names. Whichever comes later in the command line will be the one that takes effect. Similarly, except where noted below, if an option is given twice, the later setting is used. Numerical values for options may be followed by K or M, to signify multiplication by 1024 or 1024*1024 respectively. -- This terminates the list of options. It is useful if the next item on the command line starts with a hyphen but is not an option. This allows for the processing of patterns and file names that start with hyphens. -A number, --after-context=number Output up to number lines of context after each matching line. Fewer lines are output if the next match or the end of the file is reached, or if the processing buffer size has been set too small. If file names and/or line numbers are being output, a hyphen separator is used instead of a colon for the context lines (the -Z option can be used to change the file name terminator to a zero byte). A line containing "--" is output between each group of lines, unless they are in fact contiguous in the input file. The value of number is expected to be relatively small. When -c is used, -A is ignored. -a, --text Treat binary files as text. This is equivalent to --binary- files=text. --allow-lookaround-bsk PCRE2 now forbids the use of \K in lookarounds by default, in line with Perl. This option causes pcre2grep to set the PCRE2_EXTRA_ALLOW_LOOKAROUND_BSK option, which enables this somewhat dangerous usage. -B number, --before-context=number Output up to number lines of context before each matching line. Fewer lines are output if the previous match or the start of the file is within number lines, or if the processing buffer size has been set too small. If file names and/or line numbers are being output, a hyphen separator is used instead of a colon for the context lines (the -Z option can be used to change the file name terminator to a zero byte). A line containing "--" is output between each group of lines, unless they are in fact contiguous in the input file. The value of number is expected to be relatively small. When -c is used, -B is ignored. --binary-files=word Specify how binary files are to be processed. If the word is "binary" (the default), pattern matching is performed on binary files, but the only output is "Binary file <name> matches" when a match succeeds. If the word is "text", which is equivalent to the -a or --text option, binary files are processed in the same way as any other file. In this case, when a match succeeds, the output may be binary garbage, which can have nasty effects if sent to a terminal. If the word is "without-match", which is equivalent to the -I option, binary files are not processed at all; they are assumed not to be of interest and are skipped without causing any output or affecting the return code. --buffer-size=number Set the parameter that controls how much memory is obtained at the start of processing for buffering files that are being scanned. See also --max-buffer-size below. -C number, --context=number Output number lines of context both before and after each matching line. This is equivalent to setting both -A and -B to the same value. -c, --count Do not output lines from the files that are being scanned; instead output the number of lines that would have been shown, either because they matched, or, if -v is set, because they failed to match. By default, this count is exactly the same as the number of lines that would have been output, but if the -M (multiline) option is used (without -v), there may be more suppressed lines than the count (that is, the number of matches). If no lines are selected, the number zero is output. If several files are being scanned, a count is output for each of them and the -t option can be used to cause a total to be output at the end. However, if the --files-with-matches option is also used, only those files whose counts are greater than zero are listed. When -c is used, the -A, -B, and -C options are ignored. --colour, --color If this option is given without any data, it is equivalent to "--colour=auto". If data is required, it must be given in the same shell item, separated by an equals sign. --colour=value, --color=value This option specifies under what circumstances the parts of a line that matched a pattern should be coloured in the output. It is ignored if --file-offsets, --line-offsets, or --output is set. By default, output is not coloured. The value for the --colour option (which is optional, see above) may be "never", "always", or "auto". In the latter case, colouring happens only if the standard output is connected to a terminal. More resources are used when colouring is enabled, because pcre2grep has to search for all possible matches in a line, not just one, in order to colour them all. The colour that is used can be specified by setting one of the environment variables PCRE2GREP_COLOUR, PCRE2GREP_COLOR, PCREGREP_COLOUR, or PCREGREP_COLOR, which are checked in that order. If none of these are set, pcre2grep looks for GREP_COLORS or GREP_COLOR (in that order). The value of the variable should be a string of two numbers, separated by a semicolon, except in the case of GREP_COLORS, which must start with "ms=" or "mt=" followed by two semicolon-separated colours, terminated by the end of the string or by a colon. If GREP_COLORS does not start with "ms=" or "mt=" it is ignored, and GREP_COLOR is checked. If the string obtained from one of the above variables contains any characters other than semicolon or digits, the setting is ignored and the default colour is used. The string is copied directly into the control string for setting colour on a terminal, so it is your responsibility to ensure that the values make sense. If no relevant environment variable is set, the default is "1;31", which gives red. -D action, --devices=action If an input path is not a regular file or a directory, "action" specifies how it is to be processed. Valid values are "read" (the default) or "skip" (silently skip the path). -d action, --directories=action If an input path is a directory, "action" specifies how it is to be processed. Valid values are "read" (the default in non-Windows environments, for compatibility with GNU grep), "recurse" (equivalent to the -r option), or "skip" (silently skip the path, the default in Windows environments). In the "read" case, directories are read as if they were ordinary files. In some operating systems the effect of reading a directory like this is an immediate end-of-file; in others it may provoke an error. --depth-limit=number See --match-limit below. -E, --case-restrict When case distinctions are being ignored in Unicode mode, two ASCII letters (K and S) will by default match Unicode characters U+212A (Kelvin sign) and U+017F (long S) respectively, as well as their lower case ASCII counterparts. When this option is set, case equivalences are restricted such that no ASCII character matches a non-ASCII character, and vice versa. -e pattern, --regex=pattern, --regexp=pattern Specify a pattern to be matched. This option can be used multiple times in order to specify several patterns. It can also be used as a way of specifying a single pattern that starts with a hyphen. When -e is used, no argument pattern is taken from the command line; all arguments are treated as file names. There is no limit to the number of patterns. They are applied to each line in the order in which they are defined. If -f is used with -e, the command line patterns are matched first, followed by the patterns from the file(s), independent of the order in which these options are specified. --exclude=pattern Files (but not directories) whose names match the pattern are skipped without being processed. This applies to all files, whether listed on the command line, obtained from --file- list, or by scanning a directory. The pattern is a PCRE2 regular expression, and is matched against the final component of the file name, not the entire path. The -F, -w, and -x options do not apply to this pattern. The option may be given any number of times in order to specify multiple patterns. If a file name matches both an --include and an --exclude pattern, it is excluded. There is no short form for this option. --exclude-from=filename Treat each non-empty line of the file as the data for an --exclude option. What constitutes a newline when reading the file is the operating system's default. The --newline option has no effect on this option. This option may be given more than once in order to specify a number of files to read. --exclude-dir=pattern Directories whose names match the pattern are skipped without being processed, whatever the setting of the --recursive option. This applies to all directories, whether listed on the command line, obtained from --file-list, or by scanning a parent directory. The pattern is a PCRE2 regular expression, and is matched against the final component of the directory name, not the entire path. The -F, -w, and -x options do not apply to this pattern. The option may be given any number of times in order to specify more than one pattern. If a directory matches both --include-dir and --exclude-dir, it is excluded. There is no short form for this option. -F, --fixed-strings Interpret each data-matching pattern as a list of fixed strings, separated by newlines, instead of as a regular expression. What constitutes a newline for this purpose is controlled by the --newline option. The -w (match as a word) and -x (match whole line) options can be used with -F. They apply to each of the fixed strings. A line is selected if any of the fixed strings are found in it (subject to -w or -x, if present). This option applies only to the patterns that are matched against the contents of files; it does not apply to patterns specified by any of the --include or --exclude options. -f filename, --file=filename Read patterns from the file, one per line. As is the case with patterns on the command line, no delimiters should be used. What constitutes a newline when reading the file is the operating system's default interpretation of \n. The --newline option has no effect on this option. Trailing white space is removed from each line, and blank lines are ignored. An empty file contains no patterns and therefore matches nothing. Patterns read from a file in this way may contain binary zeros, which are treated as ordinary data characters. If this option is given more than once, all the specified files are read. A data line is output if any of the patterns match it. A file name can be given as "-" to refer to the standard input. When -f is used, patterns specified on the command line using -e may also be present; they are matched before the file's patterns. However, no pattern is taken from the command line; all arguments are treated as the names of paths to be searched. --file-list=filename Read a list of files and/or directories that are to be scanned from the given file, one per line. What constitutes a newline when reading the file is the operating system's default. Trailing white space is removed from each line, and blank lines are ignored. These paths are processed before any that are listed on the command line. The file name can be given as "-" to refer to the standard input. If --file and --file-list are both specified as "-", patterns are read first. This is useful only when the standard input is a terminal, from which further lines (the list of files) can be read after an end-of-file indication. If this option is given more than once, all the specified files are read. --file-offsets Instead of showing lines or parts of lines that match, show each match as an offset from the start of the file and a length, separated by a comma. In this mode, --colour has no effect, and no context is shown. That is, the -A, -B, and -C options are ignored. If there is more than one match in a line, each of them is shown separately. This option is mutually exclusive with --output, --line-offsets, and --only- matching. --group-separator=text Output this text string instead of two hyphens between groups of lines when -A, -B, or -C is in use. See also --no-group- separator. -H, --with-filename Force the inclusion of the file name at the start of output lines when searching a single file. The file name is not normally shown in this case. By default, for matching lines, the file name is followed by a colon; for context lines, a hyphen separator is used. The -Z option can be used to change the terminator to a zero byte. If a line number is also being output, it follows the file name. When the -M option causes a pattern to match more than one line, only the first is preceded by the file name. This option overrides any previous -h, -l, or -L options. -h, --no-filename Suppress the output file names when searching multiple files. File names are normally shown when multiple files are searched. By default, for matching lines, the file name is followed by a colon; for context lines, a hyphen separator is used. The -Z option can be used to change the terminator to a zero byte. If a line number is also being output, it follows the file name. This option overrides any previous -H, -L, or -l options. --heap-limit=number See --match-limit below. --help Output a help message, giving brief details of the command options and file type support, and then exit. Anything else on the command line is ignored. -I Ignore binary files. This is equivalent to --binary- files=without-match. -i, --ignore-case Ignore upper/lower case distinctions when pattern matching. This applies when matching path names for inclusion or exclusion as well as when matching lines in files. --include=pattern If any --include patterns are specified, the only files that are processed are those whose names match one of the patterns and do not match an --exclude pattern. This option does not affect directories, but it applies to all files, whether listed on the command line, obtained from --file-list, or by scanning a directory. The pattern is a PCRE2 regular expression, and is matched against the final component of the file name, not the entire path. The -F, -w, and -x options do not apply to this pattern. The option may be given any number of times. If a file name matches both an --include and an --exclude pattern, it is excluded. There is no short form for this option. --include-from=filename Treat each non-empty line of the file as the data for an --include option. What constitutes a newline for this purpose is the operating system's default. The --newline option has no effect on this option. This option may be given any number of times; all the files are read. --include-dir=pattern If any --include-dir patterns are specified, the only directories that are processed are those whose names match one of the patterns and do not match an --exclude-dir pattern. This applies to all directories, whether listed on the command line, obtained from --file-list, or by scanning a parent directory. The pattern is a PCRE2 regular expression, and is matched against the final component of the directory name, not the entire path. The -F, -w, and -x options do not apply to this pattern. The option may be given any number of times. If a directory matches both --include-dir and --exclude-dir, it is excluded. There is no short form for this option. -L, --files-without-match Instead of outputting lines from the files, just output the names of the files that do not contain any lines that would have been output. Each file name is output once, on a separate line by default, but if the -Z option is set, they are separated by zero bytes instead of newlines. This option overrides any previous -H, -h, or -l options. -l, --files-with-matches Instead of outputting lines from the files, just output the names of the files containing lines that would have been output. Each file name is output once, on a separate line, but if the -Z option is set, they are separated by zero bytes instead of newlines. Searching normally stops as soon as a matching line is found in a file. However, if the -c (count) option is also used, matching continues in order to obtain the correct count, and those files that have at least one match are listed along with their counts. Using this option with -c is a way of suppressing the listing of files with no matches that occurs with -c on its own. This option overrides any previous -H, -h, or -L options. --label=name This option supplies a name to be used for the standard input when file names are being output. If not supplied, "(standard input)" is used. There is no short form for this option. --line-buffered When this option is given, non-compressed input is read and processed line by line, and the output is flushed after each write. By default, input is read in large chunks, unless pcre2grep can determine that it is reading from a terminal, which is currently possible only in Unix-like environments or Windows. Output to terminal is normally automatically flushed by the operating system. This option can be useful when the input or output is attached to a pipe and you do not want pcre2grep to buffer up large amounts of data. However, its use will affect performance, and the -M (multiline) option ceases to work. When input is from a compressed .gz or .bz2 file, --line-buffered is ignored. --line-offsets Instead of showing lines or parts of lines that match, show each match as a line number, the offset from the start of the line, and a length. The line number is terminated by a colon (as usual; see the -n option), and the offset and length are separated by a comma. In this mode, --colour has no effect, and no context is shown. That is, the -A, -B, and -C options are ignored. If there is more than one match in a line, each of them is shown separately. This option is mutually exclusive with --output, --file-offsets, and --only-matching. --locale=locale-name This option specifies a locale to be used for pattern matching. It overrides the value in the LC_ALL or LC_CTYPE environment variables. If no locale is specified, the PCRE2 library's default (usually the "C" locale) is used. There is no short form for this option. -M, --multiline Allow patterns to match more than one line. When this option is set, the PCRE2 library is called in "multiline" mode, and a match is allowed to continue past the end of the initial line and onto one or more subsequent lines. Patterns used with -M may usefully contain literal newline characters and internal occurrences of ^ and $ characters, because in multiline mode these can match at internal newlines. Because pcre2grep is scanning multiple lines, the \Z and \z assertions match only at the end of the last line in the file. The \A assertion matches at the start of the first line of a match. This can be any line in the file; it is not anchored to the first line. The output for a successful match may consist of more than one line. The first line is the line in which the match started, and the last line is the line in which the match ended. If the matched string ends with a newline sequence, the output ends at the end of that line. If -v is set, none of the lines in a multi-line match are output. Once a match has been handled, scanning restarts at the beginning of the line after the one in which the match ended. The newline sequence that separates multiple lines must be matched as part of the pattern. For example, to find the phrase "regular expression" in a file where "regular" might be at the end of a line and "expression" at the start of the next line, you could use this command: pcre2grep -M 'regular\s+expression' <file> The \s escape sequence matches any white space character, including newlines, and is followed by + so as to match trailing white space on the first line as well as possibly handling a two-character newline sequence. There is a limit to the number of lines that can be matched, imposed by the way that pcre2grep buffers the input file as it scans it. With a sufficiently large processing buffer, this should not be a problem. The -M option does not work when input is read line by line (see --line-buffered.) -m number, --max-count=number Stop processing after finding number matching lines, or non- matching lines if -v is also set. Any trailing context lines are output after the final match. In multiline mode, each multiline match counts as just one line for this purpose. If this limit is reached when reading the standard input from a regular file, the file is left positioned just after the last matching line. If -c is also set, the count that is output is never greater than number. This option has no effect if used with -L, -l, or -q, or when just checking for a match in a binary file. --match-limit=number Processing some regular expression patterns may take a very long time to search for all possible matching strings. Others may require a very large amount of memory. There are three options that set resource limits for matching. The --match-limit option provides a means of limiting computing resource usage when processing patterns that are not going to match, but which have a very large number of possibilities in their search trees. The classic example is a pattern that uses nested unlimited repeats. Internally, PCRE2 has a counter that is incremented each time around its main processing loop. If the value set by --match-limit is reached, an error occurs. The --heap-limit option specifies, as a number of kibibytes (units of 1024 bytes), the maximum amount of heap memory that may be used for matching. The --depth-limit option limits the depth of nested backtracking points, which indirectly limits the amount of memory that is used. The amount of memory needed for each backtracking point depends on the number of capturing parentheses in the pattern, so the amount of memory that is used before this limit acts varies from pattern to pattern. This limit is of use only if it is set smaller than --match- limit. There are no short forms for these options. The default limits can be set when the PCRE2 library is compiled; if they are not specified, the defaults are very large and so effectively unlimited. --max-buffer-size=number This limits the expansion of the processing buffer, whose initial size can be set by --buffer-size. The maximum buffer size is silently forced to be no smaller than the starting buffer size. -N newline-type, --newline=newline-type Six different conventions for indicating the ends of lines in scanned files are supported. For example: pcre2grep -N CRLF 'some pattern' <file> The newline type may be specified in upper, lower, or mixed case. If the newline type is NUL, lines are separated by binary zero characters. The other types are the single- character sequences CR (carriage return) and LF (linefeed), the two-character sequence CRLF, an "anycrlf" type, which recognizes any of the preceding three types, and an "any" type, for which any Unicode line ending sequence is assumed to end a line. The Unicode sequences are the three just mentioned, plus VT (vertical tab, U+000B), FF (form feed, U+000C), NEL (next line, U+0085), LS (line separator, U+2028), and PS (paragraph separator, U+2029). When the PCRE2 library is built, a default line-ending sequence is specified. This is normally the standard sequence for the operating system. Unless otherwise specified by this option, pcre2grep uses the library's default. This option makes it possible to use pcre2grep to scan files that have come from other environments without having to modify their line endings. If the data that is being scanned does not agree with the convention set by this option, pcre2grep may behave in strange ways. Note that this option does not apply to files specified by the -f, --exclude-from, or --include-from options, which are expected to use the operating system's standard newline sequence. -n, --line-number Precede each output line by its line number in the file, followed by a colon for matching lines or a hyphen for context lines. If the file name is also being output, it precedes the line number. When the -M option causes a pattern to match more than one line, only the first is preceded by its line number. This option is forced if --line-offsets is used. --no-group-separator Do not output a separator between groups of lines when -A, -B, or -C is in use. The default is to output a line containing two hyphens. See also --group-separator. --no-jit If the PCRE2 library is built with support for just-in-time compiling (which speeds up matching), pcre2grep automatically makes use of this, unless it was explicitly disabled at build time. This option can be used to disable the use of JIT at run time. It is provided for testing and working around problems. It should never be needed in normal use. -O text, --output=text When there is a match, instead of outputting the line that matched, output just the text specified in this option, followed by an operating-system standard newline. In this mode, --colour has no effect, and no context is shown. That is, the -A, -B, and -C options are ignored. The --newline option has no effect on this option, which is mutually exclusive with --only-matching, --file-offsets, and --line- offsets. However, like --only-matching, if there is more than one match in a line, each of them causes a line of output. Escape sequences starting with a dollar character may be used to insert the contents of the matched part of the line and/or captured substrings into the text. $<digits> or ${<digits>} is replaced by the captured substring of the given decimal number; zero substitutes the whole match. If the number is greater than the number of capturing substrings, or if the capture is unset, the replacement is empty. $a is replaced by bell; $b by backspace; $e by escape; $f by form feed; $n by newline; $r by carriage return; $t by tab; $v by vertical tab. $o<digits> or $o{<digits>} is replaced by the character whose code point is the given octal number. In the first form, up to three octal digits are processed. When more digits are needed in Unicode mode to specify a wide character, the second form must be used. $x<digits> or $x{<digits>} is replaced by the character represented by the given hexadecimal number. In the first form, up to two hexadecimal digits are processed. When more digits are needed in Unicode mode to specify a wide character, the second form must be used. Any other character is substituted by itself. In particular, $$ is replaced by a single dollar. -o, --only-matching Show only the part of the line that matched a pattern instead of the whole line. In this mode, no context is shown. That is, the -A, -B, and -C options are ignored. If there is more than one match in a line, each of them is shown separately, on a separate line of output. If -o is combined with -v (invert the sense of the match to find non-matching lines), no output is generated, but the return code is set appropriately. If the matched portion of the line is empty, nothing is output unless the file name or line number are being printed, in which case they are shown on an otherwise empty line. This option is mutually exclusive with --output, --file-offsets and --line-offsets. -onumber, --only-matching=number Show only the part of the line that matched the capturing parentheses of the given number. Up to 50 capturing parentheses are supported by default. This limit can be changed via the --om-capture option. A pattern may contain any number of capturing parentheses, but only those whose number is within the limit can be accessed by -o. An error occurs if the number specified by -o is greater than the limit. -o0 is the same as -o without a number. Because these options can be given without an argument (see above), if an argument is present, it must be given in the same shell item, for example, -o3 or --only-matching=2. The comments given for the non-argument case above also apply to this option. If the specified capturing parentheses do not exist in the pattern, or were not set in the match, nothing is output unless the file name or line number are being output. If this option is given multiple times, multiple substrings are output for each match, in the order the options are given, and all on one line. For example, -o3 -o1 -o3 causes the substrings matched by capturing parentheses 3 and 1 and then 3 again to be output. By default, there is no separator (but see the next but one option). --om-capture=number Set the number of capturing parentheses that can be accessed by -o. The default is 50. --om-separator=text Specify a separating string for multiple occurrences of -o. The default is an empty string. Separating strings are never coloured. -P, --no-ucp Starting from release 10.43, when UTF/Unicode mode is specified with -u or -U, the PCRE2_UCP option is used by default. This means that the POSIX classes in patterns match more than just ASCII characters. For example, [:digit:] matches any Unicode decimal digit. The --no-ucp option suppresses PCRE2_UCP, thus restricting the POSIX classes to ASCII characters, as was the case in earlier releases. Note that there are now more fine-grained option settings within patterns that affect individual classes. For example, when in UCP mode, the sequence (?aP) restricts [:word:] to ASCII letters, while allowing \w to match Unicode letters and digits. -q, --quiet Work quietly, that is, display nothing except error messages. The exit status indicates whether or not any matches were found. -r, --recursive If any given path is a directory, recursively scan the files it contains, taking note of any --include and --exclude settings. By default, a directory is read as a normal file; in some operating systems this gives an immediate end-of- file. This option is a shorthand for setting the -d option to "recurse". --recursion-limit=number This is an obsolete synonym for --depth-limit. See --match- limit above for details. -s, --no-messages Suppress error messages about non-existent or unreadable files. Such files are quietly skipped. However, the return code is still 2, even if matches were found in other files. -t, --total-count This option is useful when scanning more than one file. If used on its own, -t suppresses all output except for a grand total number of matching lines (or non-matching lines if -v is used) in all the files. If -t is used with -c, a grand total is output except when the previous output is just one line. In other words, it is not output when just one file's count is listed. If file names are being output, the grand total is preceded by "TOTAL:". Otherwise, it appears as just another number. The -t option is ignored when used with -L (list files without matches), because the grand total would always be zero. -u, --utf Operate in UTF/Unicode mode. This option is available only if PCRE2 has been compiled with UTF-8 support. All patterns (including those for any --exclude and --include options) and all lines that are scanned must be valid strings of UTF-8 characters. If an invalid UTF-8 string is encountered, an error occurs. -U, --utf-allow-invalid As --utf, but in addition subject lines may contain invalid UTF-8 code unit sequences. These can never form part of any pattern match. Patterns themselves, however, must still be valid UTF-8 strings. This facility allows valid UTF-8 strings to be sought within arbitrary byte sequences in executable or other binary files. For more details about matching in non- valid UTF-8 strings, see the pcre2unicode(3) documentation. -V, --version Write the version numbers of pcre2grep and the PCRE2 library to the standard output and then exit. Anything else on the command line is ignored. -v, --invert-match Invert the sense of the match, so that lines which do not match any of the patterns are the ones that are found. When this option is set, options such as --only-matching and --output, which specify parts of a match that are to be output, are ignored. -w, --word-regex, --word-regexp Force the patterns only to match "words". That is, there must be a word boundary at the start and end of each matched string. This is equivalent to having "\b(?:" at the start of each pattern, and ")\b" at the end. This option applies only to the patterns that are matched against the contents of files; it does not apply to patterns specified by any of the --include or --exclude options. -x, --line-regex, --line-regexp Force the patterns to start matching only at the beginnings of lines, and in addition, require them to match entire lines. In multiline mode the match may be more than one line. This is equivalent to having "^(?:" at the start of each pattern and ")$" at the end. This option applies only to the patterns that are matched against the contents of files; it does not apply to patterns specified by any of the --include or --exclude options. -Z, --null Terminate files names in the regular output with a zero byte (the NUL character) instead of what would normally appear. This is useful when file names contain unusual characters such as colons, hyphens, or even newlines. The option does not apply to file names in error messages. ENVIRONMENT VARIABLES The environment variables LC_ALL and LC_CTYPE are examined, in that order, for a locale. The first one that is set is used. This can be overridden by the --locale option. If no locale is set, the PCRE2 library's default (usually the "C" locale) is used. NEWLINES The -N (--newline) option allows pcre2grep to scan files with newline conventions that differ from the default. This option affects only the way scanned files are processed. It does not affect the interpretation of files specified by the -f, --file-list, --exclude-from, or --include-from options. Any parts of the scanned input files that are written to the standard output are copied with whatever newline sequences they have in the input. However, if the final line of a file is output, and it does not end with a newline sequence, a newline sequence is added. If the newline setting is CR, LF, CRLF or NUL, that line ending is output; for the other settings (ANYCRLF or ANY) a single NL is used. The newline setting does not affect the way in which pcre2grep writes newlines in informational messages to the standard output and error streams. Under Windows, the standard output is set to be binary, so that "\r\n" at the ends of output lines that are copied from the input is not converted to "\r\r\n" by the C I/O library. This means that any messages written to the standard output must end with "\r\n". For all other operating systems, and for all messages to the standard error stream, "\n" is used. OPTIONS COMPATIBILITY WITH GNU GREP Many of the short and long forms of pcre2grep's options are the same as in the GNU grep program. Any long option of the form --xxx-regexp (GNU terminology) is also available as --xxx-regex (PCRE2 terminology). However, the --case-restrict, --depth-limit, -E, --file-list, --file- offsets, --heap-limit, --include-dir, --line-offsets, --locale, --match-limit, -M, --multiline, -N, --newline, --no-ucp, --om- separator, --output, -P, -u, --utf, -U, and --utf-allow-invalid options are specific to pcre2grep, as is the use of the --only-matching option with a capturing parentheses number. Although most of the common options work the same way, a few are different in pcre2grep. For example, the --include option's argument is a glob for GNU grep, but in pcre2grep it is a regular expression to which the -i option applies. If both the -c and -l options are given, GNU grep lists only file names, without counts, but pcre2grep gives the counts as well. OPTIONS WITH DATA There are four different ways in which an option with data can be specified. If a short form option is used, the data may follow immediately, or (with one exception) in the next command line item. For example: -f/some/file -f /some/file The exception is the -o option, which may appear with or without data. Because of this, if data is present, it must follow immediately in the same item, for example -o3. If a long form option is used, the data may appear in the same command line item, separated by an equals character, or (with two exceptions) it may appear in the next command line item. For example: --file=/some/file --file /some/file Note, however, that if you want to supply a file name beginning with ~ as data in a shell command, and have the shell expand ~ to a home directory, you must separate the file name from the option, because the shell does not treat ~ specially unless it is at the start of an item. The exceptions to the above are the --colour (or --color) and --only- matching options, for which the data is optional. If one of these options does have data, it must be given in the first form, using an equals character. Otherwise pcre2grep will assume that it has no data. USING PCRE2'S CALLOUT FACILITY pcre2grep has, by default, support for calling external programs or scripts or echoing specific strings during matching by making use of PCRE2's callout facility. However, this support can be completely or partially disabled when pcre2grep is built. You can find out whether your binary has support for callouts by running it with the --help option. If callout support is completely disabled, all callouts in patterns are ignored by pcre2grep. If the facility is partially disabled, calling external programs is not supported, and callouts that request it are ignored. A callout in a PCRE2 pattern is of the form (?C<arg>) where the argument is either a number or a quoted string (see the pcre2callout documentation for details). Numbered callouts are ignored by pcre2grep; only callouts with string arguments are useful. Echoing a specific string Starting the callout string with a pipe character invokes an echoing facility that avoids calling an external program or script. This facility is always available, provided that callouts were not completely disabled when pcre2grep was built. The rest of the callout string is processed as a zero-terminated string, which means it should not contain any internal binary zeros. It is written to the output, having first been passed through the same escape processing as text from the --output (-O) option (see above). However, $0 cannot be used to insert a matched substring because the match is still in progress. Instead, the single character '0' is inserted. Any syntax errors in the string (for example, a dollar not followed by another character) causes the callout to be ignored. No terminator is added to the output string, so if you want a newline, you must include it explicitly using the escape $n. For example: pcre2grep '(.)(..(.))(?C"|[$1] [$2] [$3]$n")' <some file> Matching continues normally after the string is output. If you want to see only the callout output but not any output from an actual match, you should end the pattern with (*FAIL). Calling external programs or scripts This facility can be independently disabled when pcre2grep is built. It is supported for Windows, where a call to _spawnvp() is used, for VMS, where lib$spawn() is used, and for any Unix-like environment where fork() and execv() are available. If the callout string does not start with a pipe (vertical bar) character, it is parsed into a list of substrings separated by pipe characters. The first substring must be an executable name, with the following substrings specifying arguments: executable_name|arg1|arg2|... Any substring (including the executable name) may contain escape sequences started by a dollar character. These are the same as for the --output (-O) option documented above, except that $0 cannot insert the matched string because the match is still in progress. Instead, the character '0' is inserted. If you need a literal dollar or pipe character in any substring, use $$ or $| respectively. Here is an example: echo -e "abcde\n12345" | pcre2grep \ '(?x)(.)(..(.)) (?C"/bin/echo|Arg1: [$1] [$2] [$3]|Arg2: $|${1}$| ($4)")()' - Output: Arg1: [a] [bcd] [d] Arg2: |a| () abcde Arg1: [1] [234] [4] Arg2: |1| () 12345 The parameters for the system call that is used to run the program or script are zero-terminated strings. This means that binary zero characters in the callout argument will cause premature termination of their substrings, and therefore should not be present. Any syntax errors in the string (for example, a dollar not followed by another character) causes the callout to be ignored. If running the program fails for any reason (including the non-existence of the executable), a local matching failure occurs and the matcher backtracks in the normal way. MATCHING ERRORS It is possible to supply a regular expression that takes a very long time to fail to match certain lines. Such patterns normally involve nested indefinite repeats, for example: (a+)*\d when matched against a line of a's with no final digit. The PCRE2 matching function has a resource limit that causes it to abort in these circumstances. If this happens, pcre2grep outputs an error message and the line that caused the problem to the standard error stream. If there are more than 20 such errors, pcre2grep gives up. The --match-limit option of pcre2grep can be used to set the overall resource limit. There are also other limits that affect the amount of memory used during matching; see the discussion of --heap-limit and --depth-limit above. DIAGNOSTICS Exit status is 0 if any matches were found, 1 if no matches were found, and 2 for syntax errors, overlong lines, non-existent or inaccessible files (even if matches were found in other files) or too many matching errors. Using the -s option to suppress error messages about inaccessible files does not affect the return code. When run under VMS, the return code is placed in the symbol PCRE2GREP_RC because VMS does not distinguish between exit(0) and exit(1). SEE ALSO pcre2pattern(3), pcre2syntax(3), pcre2callout(3), pcre2unicode(3). AUTHOR Philip Hazel Retired from University Computing Service Cambridge, England. REVISION Last updated: 22 December 2023 Copyright (c) 1997-2023 University of Cambridge. PCRE2 10.43 22 December 2023 PCRE2GREP(1)
null
autopoint
Copies standard gettext infrastructure files into a source package.
autopoint - copies standard gettext infrastructure
autopoint [OPTION]...
--help print this help and exit --version print version information and exit -f, --force force overwriting of files that already exist -n, --dry-run print modifications but don't perform them AUTHOR Written by Bruno Haible REPORTING BUGS Report bugs in the bug tracker at <https://savannah.gnu.org/projects/gettext> or by email to <bug-gettext@gnu.org>. COPYRIGHT Copyright © 2002-2023 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <https://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. SEE ALSO The full documentation for autopoint is maintained as a Texinfo manual. If the info and autopoint programs are properly installed at your site, the command info autopoint should give you access to the complete manual. GNU gettext-tools 0.22.5 February 2024 AUTOPOINT(1)
null
machocheck
null
null
null
null
null
macdeployqt
null
null
null
null
null
qdarkstyle
null
null
null
null
null
gifclrmp
A program to modify GIF image colormaps. Any local colormap in a GIF file can be modified at a time, or the global screen one.
gifclrmp - extract colormaps from GIF images
gifclrmp [-v] [-s] [-l mapfile] [-t trans] [-g gamma] [-i image] [-h] [gif-file]
-v Verbose mode (show progress). Enables printout of running scan lines. -s Select the global screen color map. -l mapfile Load color map from this file instead of selected color map. -t trans Change color index values. The change is made to both the selected color table and the raster bits of the selected image. A translation file is a list of pairs of `before' and `after' index values. At present, the `before' index values must be in ascending order starting from 0. -g gamma Apply gamma correction to selected color map. -i image Select the color map of the numbered image. -h Print one command line help, similar to Usage above. If no GIF file is given, gifclip will try to read a GIF file from stdin. NOTES • The default operation is to dump out the selected color map in text format. • The file to load/dump is simply one color map entry per line. Each such entry line has four integers: "ColorIndex Red Green Blue", where color index is in ascending order starting from 1. AUTHOR Gershon Elber. GIFLIB 2 May 2012 GIFCLRMP(1)
null
mamba-package
null
null
null
null
null
pyftsubset
null
null
null
null
null
mysqlbinlog
The server's binary log consists of files containing “events” that describe modifications to database contents. The server writes these files in binary format. To display their contents in text format, use the mysqlbinlog utility. You can also use mysqlbinlog to display the contents of relay log files written by a replica server in a replication setup because relay logs have the same format as binary logs. The binary log and relay log are discussed further in Section 5.4.4, “The Binary Log”, and Section 17.2.4, “Relay Log and Replication Metadata Repositories”. Invoke mysqlbinlog like this: mysqlbinlog [options] log_file ... For example, to display the contents of the binary log file named binlog.000003, use this command: mysqlbinlog binlog.000003 The output includes events contained in binlog.000003. For statement-based logging, event information includes the SQL statement, the ID of the server on which it was executed, the timestamp when the statement was executed, how much time it took, and so forth. For row-based logging, the event indicates a row change rather than an SQL statement. See Section 17.2.1, “Replication Formats”, for information about logging modes. Events are preceded by header comments that provide additional information. For example: # at 141 #100309 9:28:36 server id 123 end_log_pos 245 Query thread_id=3350 exec_time=11 error_code=0 In the first line, the number following at indicates the file offset, or starting position, of the event in the binary log file. The second line starts with a date and time indicating when the statement started on the server where the event originated. For replication, this timestamp is propagated to replica servers. server id is the server_id value of the server where the event originated. end_log_pos indicates where the next event starts (that is, it is the end position of the current event + 1). thread_id indicates which thread executed the event. exec_time is the time spent executing the event, on a replication source server. On a replica, it is the difference of the end execution time on the replica minus the beginning execution time on the source. The difference serves as an indicator of how much replication lags behind the source. error_code indicates the result from executing the event. Zero means that no error occurred. Note When using event groups, the file offsets of events may be grouped together and the comments of events may be grouped together. Do not mistake these grouped events for blank file offsets. The output from mysqlbinlog can be re-executed (for example, by using it as input to mysql) to redo the statements in the log. This is useful for recovery operations after an unexpected server exit. For other usage examples, see the discussion later in this section and in Section 7.5, “Point-in-Time (Incremental) Recovery”. To execute the internal-use BINLOG statements used by mysqlbinlog, the user requires the BINLOG_ADMIN privilege (or the deprecated SUPER privilege), or the REPLICATION_APPLIER privilege plus the appropriate privileges to execute each log event. You can use mysqlbinlog to read binary log files directly and apply them to the local MySQL server. You can also read binary logs from a remote server by using the --read-from-remote-server option. To read remote binary logs, the connection parameter options can be given to indicate how to connect to the server. These options are --host, --password, --port, --protocol, --socket, and --user. When binary log files have been encrypted, which can be done from MySQL 8.0.14 onwards, mysqlbinlog cannot read them directly, but can read them from the server using the --read-from-remote-server option. Binary log files are encrypted when the server's binlog_encryption system variable is set to ON. The SHOW BINARY LOGS statement shows whether a particular binary log file is encrypted or unencrypted. Encrypted and unencrypted binary log files can also be distinguished using the magic number at the start of the file header for encrypted log files (0xFD62696E), which differs from that used for unencrypted log files (0xFE62696E). Note that from MySQL 8.0.14, mysqlbinlog returns a suitable error if you attempt to read an encrypted binary log file directly, but older versions of mysqlbinlog do not recognise the file as a binary log file at all. For more information on binary log encryption, see Section 17.3.2, “Encrypting Binary Log Files and Relay Log Files”. When binary log transaction payloads have been compressed, which can be done from MySQL 8.0.20 onwards, mysqlbinlog versions from that release on automatically decompress and decode the transaction payloads, and print them as they would uncompressed events. Older versions of mysqlbinlog cannot read compressed transaction payloads. When the server's binlog_transaction_compression system variable is set to ON, transaction payloads are compressed and then written to the server's binary log file as a single event (a Transaction_payload_event). With the --verbose option, mysqlbinlog adds comments stating the compression algorithm used, the compressed payload size that was originally received, and the resulting payload size after decompression. Note The end position (end_log_pos) that mysqlbinlog states for an individual event that was part of a compressed transaction payload is the same as the end position of the original compressed payload. Multiple decompressed events can therefore have the same end position. mysqlbinlog's own connection compression does less if transaction payloads are already compressed, but still operates on uncompressed transactions and headers. For more information on binary log transaction compression, see Section 5.4.4.5, “Binary Log Transaction Compression”. When running mysqlbinlog against a large binary log, be careful that the filesystem has enough space for the resulting files. To configure the directory that mysqlbinlog uses for temporary files, use the TMPDIR environment variable. mysqlbinlog sets the value of pseudo_replica_mode to true before executing any SQL statements. This system variable affects the handling of XA transactions, the original_commit_timestamp replication delay timestamp and the original_server_version system variable, and unsupported SQL modes. mysqlbinlog supports the following options, which can be specified on the command line or in the [mysqlbinlog] and [client] groups of an option file. For information about option files used by MySQL programs, see Section 4.2.2.2, “Using Option Files”. • --help, -? ┌────────────────────┬────────┐ │Command-Line Format │ --help │ └────────────────────┴────────┘ Display a help message and exit. • --base64-output=value ┌────────────────────┬────────────────────────┐ │Command-Line Format │ --base64-output=value │ ├────────────────────┼────────────────────────┤ │Type │ String │ ├────────────────────┼────────────────────────┤ │Default Value │ AUTO │ ├────────────────────┼────────────────────────┤ │Valid Values │ AUTO NEVER │ │ │ DECODE-ROWS │ └────────────────────┴────────────────────────┘ This option determines when events should be displayed encoded as base-64 strings using BINLOG statements. The option has these permissible values (not case-sensitive): • AUTO ("automatic") or UNSPEC ("unspecified") displays BINLOG statements automatically when necessary (that is, for format description events and row events). If no --base64-output option is given, the effect is the same as --base64-output=AUTO. Note Automatic BINLOG display is the only safe behavior if you intend to use the output of mysqlbinlog to re-execute binary log file contents. The other option values are intended only for debugging or testing purposes because they may produce output that does not include all events in executable form. • NEVER causes BINLOG statements not to be displayed. mysqlbinlog exits with an error if a row event is found that must be displayed using BINLOG. • DECODE-ROWS specifies to mysqlbinlog that you intend for row events to be decoded and displayed as commented SQL statements by also specifying the --verbose option. Like NEVER, DECODE-ROWS suppresses display of BINLOG statements, but unlike NEVER, it does not exit with an error if a row event is found. For examples that show the effect of --base64-output and --verbose on row event output, see the section called “MYSQLBINLOG ROW EVENT DISPLAY”. • --bind-address=ip_address ┌────────────────────┬───────────────────────────┐ │Command-Line Format │ --bind-address=ip_address │ └────────────────────┴───────────────────────────┘ On a computer having multiple network interfaces, use this option to select which interface to use for connecting to the MySQL server. • --binlog-row-event-max-size=N ┌────────────────────┬───────────────────────────────┐ │Command-Line Format │ --binlog-row-event-max-size=# │ ├────────────────────┼───────────────────────────────┤ │Type │ Numeric │ ├────────────────────┼───────────────────────────────┤ │Default Value │ 4294967040 │ ├────────────────────┼───────────────────────────────┤ │Minimum Value │ 256 │ ├────────────────────┼───────────────────────────────┤ │Maximum Value │ 18446744073709547520 │ └────────────────────┴───────────────────────────────┘ Specify the maximum size of a row-based binary log event, in bytes. Rows are grouped into events smaller than this size if possible. The value should be a multiple of 256. The default is 4GB. • --character-sets-dir=dir_name ┌────────────────────┬───────────────────────────────┐ │Command-Line Format │ --character-sets-dir=dir_name │ ├────────────────────┼───────────────────────────────┤ │Type │ Directory name │ └────────────────────┴───────────────────────────────┘ The directory where character sets are installed. See Section 10.15, “Character Set Configuration”. • --compress ┌────────────────────┬───────────────────────┐ │Command-Line Format │ --compress[={OFF|ON}] │ ├────────────────────┼───────────────────────┤ │Deprecated │ Yes │ ├────────────────────┼───────────────────────┤ │Type │ Boolean │ ├────────────────────┼───────────────────────┤ │Default Value │ OFF │ └────────────────────┴───────────────────────┘ Compress all information sent between the client and the server if possible. See Section 4.2.8, “Connection Compression Control”. This option was added in MySQL 8.0.17. As of MySQL 8.0.18 it is deprecated. Expect it to be removed in a future version of MySQL. See the section called “Configuring Legacy Connection Compression”. • --compression-algorithms=value ┌────────────────────┬────────────────────────────────┐ │Command-Line Format │ --compression-algorithms=value │ ├────────────────────┼────────────────────────────────┤ │Type │ Set │ ├────────────────────┼────────────────────────────────┤ │Default Value │ uncompressed │ ├────────────────────┼────────────────────────────────┤ │Valid Values │ zlib zstd uncompressed │ └────────────────────┴────────────────────────────────┘ The permitted compression algorithms for connections to the server. The available algorithms are the same as for the protocol_compression_algorithms system variable. The default value is uncompressed. For more information, see Section 4.2.8, “Connection Compression Control”. • --connection-server-id=server_id ┌────────────────────┬───────────────────────────┐ │Command-Line Format │ --connection-server-id=#] │ ├────────────────────┼───────────────────────────┤ │Type │ Integer │ ├────────────────────┼───────────────────────────┤ │Default Value │ 0 (1) │ ├────────────────────┼───────────────────────────┤ │Minimum Value │ 0 (1) │ ├────────────────────┼───────────────────────────┤ │Maximum Value │ 4294967295 │ └────────────────────┴───────────────────────────┘ --connection-server-id specifies the server ID that mysqlbinlog reports when it connects to the server. It can be used to avoid a conflict with the ID of a replica server or another mysqlbinlog process. If the --read-from-remote-server option is specified, mysqlbinlog reports a server ID of 0, which tells the server to disconnect after sending the last log file (nonblocking behavior). If the --stop-never option is also specified to maintain the connection to the server, mysqlbinlog reports a server ID of 1 by default instead of 0, and --connection-server-id can be used to replace that server ID if required. See the section called “SPECIFYING THE MYSQLBINLOG SERVER ID”. • --database=db_name, -d db_name ┌────────────────────┬────────────────────┐ │Command-Line Format │ --database=db_name │ ├────────────────────┼────────────────────┤ │Type │ String │ └────────────────────┴────────────────────┘ This option causes mysqlbinlog to output entries from the binary log (local log only) that occur while db_name is been selected as the default database by USE. The --database option for mysqlbinlog is similar to the --binlog-do-db option for mysqld, but can be used to specify only one database. If --database is given multiple times, only the last instance is used. The effects of this option depend on whether the statement-based or row-based logging format is in use, in the same way that the effects of --binlog-do-db depend on whether statement-based or row-based logging is in use. Statement-based logging. The --database option works as follows: • While db_name is the default database, statements are output whether they modify tables in db_name or a different database. • Unless db_name is selected as the default database, statements are not output, even if they modify tables in db_name. • There is an exception for CREATE DATABASE, ALTER DATABASE, and DROP DATABASE. The database being created, altered, or dropped is considered to be the default database when determining whether to output the statement. Suppose that the binary log was created by executing these statements using statement-based-logging: INSERT INTO test.t1 (i) VALUES(100); INSERT INTO db2.t2 (j) VALUES(200); USE test; INSERT INTO test.t1 (i) VALUES(101); INSERT INTO t1 (i) VALUES(102); INSERT INTO db2.t2 (j) VALUES(201); USE db2; INSERT INTO test.t1 (i) VALUES(103); INSERT INTO db2.t2 (j) VALUES(202); INSERT INTO t2 (j) VALUES(203); mysqlbinlog --database=test does not output the first two INSERT statements because there is no default database. It outputs the three INSERT statements following USE test, but not the three INSERT statements following USE db2. mysqlbinlog --database=db2 does not output the first two INSERT statements because there is no default database. It does not output the three INSERT statements following USE test, but does output the three INSERT statements following USE db2. Row-based logging. mysqlbinlog outputs only entries that change tables belonging to db_name. The default database has no effect on this. Suppose that the binary log just described was created using row-based logging rather than statement-based logging. mysqlbinlog --database=test outputs only those entries that modify t1 in the test database, regardless of whether USE was issued or what the default database is. If a server is running with binlog_format set to MIXED and you want it to be possible to use mysqlbinlog with the --database option, you must ensure that tables that are modified are in the database selected by USE. (In particular, no cross-database updates should be used.) When used together with the --rewrite-db option, the --rewrite-db option is applied first; then the --database option is applied, using the rewritten database name. The order in which the options are provided makes no difference in this regard. • --debug[=debug_options], -# [debug_options] ┌────────────────────┬──────────────────────────────┐ │Command-Line Format │ --debug[=debug_options] │ ├────────────────────┼──────────────────────────────┤ │Type │ String │ ├────────────────────┼──────────────────────────────┤ │Default Value │ d:t:o,/tmp/mysqlbinlog.trace │ └────────────────────┴──────────────────────────────┘ Write a debugging log. A typical debug_options string is d:t:o,file_name. The default is d:t:o,/tmp/mysqlbinlog.trace. This option is available only if MySQL was built using WITH_DEBUG. MySQL release binaries provided by Oracle are not built using this option. • --debug-check ┌────────────────────┬───────────────┐ │Command-Line Format │ --debug-check │ ├────────────────────┼───────────────┤ │Type │ Boolean │ ├────────────────────┼───────────────┤ │Default Value │ FALSE │ └────────────────────┴───────────────┘ Print some debugging information when the program exits. This option is available only if MySQL was built using WITH_DEBUG. MySQL release binaries provided by Oracle are not built using this option. • --debug-info ┌────────────────────┬──────────────┐ │Command-Line Format │ --debug-info │ ├────────────────────┼──────────────┤ │Type │ Boolean │ ├────────────────────┼──────────────┤ │Default Value │ FALSE │ └────────────────────┴──────────────┘ Print debugging information and memory and CPU usage statistics when the program exits. This option is available only if MySQL was built using WITH_DEBUG. MySQL release binaries provided by Oracle are not built using this option. • --default-auth=plugin ┌────────────────────┬───────────────────────┐ │Command-Line Format │ --default-auth=plugin │ ├────────────────────┼───────────────────────┤ │Type │ String │ └────────────────────┴───────────────────────┘ A hint about which client-side authentication plugin to use. See Section 6.2.17, “Pluggable Authentication”. • --defaults-extra-file=file_name ┌────────────────────┬─────────────────────────────────┐ │Command-Line Format │ --defaults-extra-file=file_name │ ├────────────────────┼─────────────────────────────────┤ │Type │ File name │ └────────────────────┴─────────────────────────────────┘ Read this option file after the global option file but (on Unix) before the user option file. If the file does not exist or is otherwise inaccessible, an error occurs. If file_name is not an absolute path name, it is interpreted relative to the current directory. For additional information about this and other option-file options, see Section 4.2.2.3, “Command-Line Options that Affect Option-File Handling”. • --defaults-file=file_name ┌────────────────────┬───────────────────────────┐ │Command-Line Format │ --defaults-file=file_name │ ├────────────────────┼───────────────────────────┤ │Type │ File name │ └────────────────────┴───────────────────────────┘ Use only the given option file. If the file does not exist or is otherwise inaccessible, an error occurs. If file_name is not an absolute path name, it is interpreted relative to the current directory. Exception: Even with --defaults-file, client programs read .mylogin.cnf. For additional information about this and other option-file options, see Section 4.2.2.3, “Command-Line Options that Affect Option-File Handling”. • --defaults-group-suffix=str ┌────────────────────┬─────────────────────────────┐ │Command-Line Format │ --defaults-group-suffix=str │ ├────────────────────┼─────────────────────────────┤ │Type │ String │ └────────────────────┴─────────────────────────────┘ Read not only the usual option groups, but also groups with the usual names and a suffix of str. For example, mysqlbinlog normally reads the [client] and [mysqlbinlog] groups. If this option is given as --defaults-group-suffix=_other, mysqlbinlog also reads the [client_other] and [mysqlbinlog_other] groups. For additional information about this and other option-file options, see Section 4.2.2.3, “Command-Line Options that Affect Option-File Handling”. • --disable-log-bin, -D ┌────────────────────┬───────────────────┐ │Command-Line Format │ --disable-log-bin │ └────────────────────┴───────────────────┘ Disable binary logging. This is useful for avoiding an endless loop if you use the --to-last-log option and are sending the output to the same MySQL server. This option also is useful when restoring after an unexpected exit to avoid duplication of the statements you have logged. This option causes mysqlbinlog to include a SET sql_log_bin = 0 statement in its output to disable binary logging of the remaining output. Manipulating the session value of the sql_log_bin system variable is a restricted operation, so this option requires that you have privileges sufficient to set restricted session variables. See Section 5.1.9.1, “System Variable Privileges”. • --exclude-gtids=gtid_set ┌────────────────────┬──────────────────────────┐ │Command-Line Format │ --exclude-gtids=gtid_set │ ├────────────────────┼──────────────────────────┤ │Type │ String │ ├────────────────────┼──────────────────────────┤ │Default Value │ │ └────────────────────┴──────────────────────────┘ Do not display any of the groups listed in the gtid_set. • --force-if-open, -F ┌────────────────────┬─────────────────┐ │Command-Line Format │ --force-if-open │ └────────────────────┴─────────────────┘ Read binary log files even if they are open or were not closed properly (IN_USE flag is set); do not fail if the file ends with a truncated event. The IN_USE flag is set only for the binary log that is currently written by the server; if the server has crashed, the flag remains set until the server is started up again and recovers the binary log. Without this option, mysqlbinlog refuses to process a file with this flag set. Since the server may be in the process of writing the file, truncation of the last event is considered normal. • --force-read, -f ┌────────────────────┬──────────────┐ │Command-Line Format │ --force-read │ └────────────────────┴──────────────┘ With this option, if mysqlbinlog reads a binary log event that it does not recognize, it prints a warning, ignores the event, and continues. Without this option, mysqlbinlog stops if it reads such an event. • --get-server-public-key ┌────────────────────┬─────────────────────────┐ │Command-Line Format │ --get-server-public-key │ ├────────────────────┼─────────────────────────┤ │Type │ Boolean │ └────────────────────┴─────────────────────────┘ Request from the server the public key required for RSA key pair-based password exchange. This option applies to clients that authenticate with the caching_sha2_password authentication plugin. For that plugin, the server does not send the public key unless requested. This option is ignored for accounts that do not authenticate with that plugin. It is also ignored if RSA-based password exchange is not used, as is the case when the client connects to the server using a secure connection. If --server-public-key-path=file_name is given and specifies a valid public key file, it takes precedence over --get-server-public-key. For information about the caching_sha2_password plugin, see Section 6.4.1.2, “Caching SHA-2 Pluggable Authentication”. • --hexdump, -H ┌────────────────────┬───────────┐ │Command-Line Format │ --hexdump │ └────────────────────┴───────────┘ Display a hex dump of the log in comments, as described in the section called “MYSQLBINLOG HEX DUMP FORMAT”. The hex output can be helpful for replication debugging. • --host=host_name, -h host_name ┌────────────────────┬──────────────────┐ │Command-Line Format │ --host=host_name │ ├────────────────────┼──────────────────┤ │Type │ String │ ├────────────────────┼──────────────────┤ │Default Value │ localhost │ └────────────────────┴──────────────────┘ Get the binary log from the MySQL server on the given host. • --idempotent ┌────────────────────┬──────────────┐ │Command-Line Format │ --idempotent │ ├────────────────────┼──────────────┤ │Type │ Boolean │ ├────────────────────┼──────────────┤ │Default Value │ true │ └────────────────────┴──────────────┘ Tell the MySQL Server to use idempotent mode while processing updates; this causes suppression of any duplicate-key or key-not-found errors that the server encounters in the current session while processing updates. This option may prove useful whenever it is desirable or necessary to replay one or more binary logs to a MySQL Server which may not contain all of the data to which the logs refer. The scope of effect for this option includes the current mysqlbinlog client and session only. • --include-gtids=gtid_set ┌────────────────────┬──────────────────────────┐ │Command-Line Format │ --include-gtids=gtid_set │ ├────────────────────┼──────────────────────────┤ │Type │ String │ ├────────────────────┼──────────────────────────┤ │Default Value │ │ └────────────────────┴──────────────────────────┘ Display only the groups listed in the gtid_set. • --local-load=dir_name, -l dir_name ┌────────────────────┬───────────────────────┐ │Command-Line Format │ --local-load=dir_name │ ├────────────────────┼───────────────────────┤ │Type │ Directory name │ └────────────────────┴───────────────────────┘ For data loading operations corresponding to LOAD DATA statements, mysqlbinlog extracts the files from the binary log events, writes them as temporary files to the local file system, and writes LOAD DATA LOCAL statements to cause the files to be loaded. By default, mysqlbinlog writes these temporary files to an operating system-specific directory. The --local-load option can be used to explicitly specify the directory where mysqlbinlog should prepare local temporary files. Because other processes can write files to the default system-specific directory, it is advisable to specify the --local-load option to mysqlbinlog to designate a different directory for data files, and then designate that same directory by specifying the --load-data-local-dir option to mysql when processing the output from mysqlbinlog. For example: mysqlbinlog --local-load=/my/local/data ... | mysql --load-data-local-dir=/my/local/data ... Important These temporary files are not automatically removed by mysqlbinlog or any other MySQL program. • --login-path=name ┌────────────────────┬───────────────────┐ │Command-Line Format │ --login-path=name │ ├────────────────────┼───────────────────┤ │Type │ String │ └────────────────────┴───────────────────┘ Read options from the named login path in the .mylogin.cnf login path file. A “login path” is an option group containing options that specify which MySQL server to connect to and which account to authenticate as. To create or modify a login path file, use the mysql_config_editor utility. See mysql_config_editor(1). For additional information about this and other option-file options, see Section 4.2.2.3, “Command-Line Options that Affect Option-File Handling”. • --no-login-paths ┌────────────────────┬──────────────────┐ │Command-Line Format │ --no-login-paths │ └────────────────────┴──────────────────┘ Skips reading options from the login path file. See --login-path for related information. For additional information about this and other option-file options, see Section 4.2.2.3, “Command-Line Options that Affect Option-File Handling”. • --no-defaults ┌────────────────────┬───────────────┐ │Command-Line Format │ --no-defaults │ └────────────────────┴───────────────┘ Do not read any option files. If program startup fails due to reading unknown options from an option file, --no-defaults can be used to prevent them from being read. The exception is that the .mylogin.cnf file is read in all cases, if it exists. This permits passwords to be specified in a safer way than on the command line even when --no-defaults is used. To create .mylogin.cnf, use the mysql_config_editor utility. See mysql_config_editor(1). For additional information about this and other option-file options, see Section 4.2.2.3, “Command-Line Options that Affect Option-File Handling”. • --offset=N, -o N ┌────────────────────┬────────────┐ │Command-Line Format │ --offset=# │ ├────────────────────┼────────────┤ │Type │ Numeric │ └────────────────────┴────────────┘ Skip the first N entries in the log. • --open-files-limit=N ┌────────────────────┬──────────────────────┐ │Command-Line Format │ --open-files-limit=# │ ├────────────────────┼──────────────────────┤ │Type │ Numeric │ ├────────────────────┼──────────────────────┤ │Default Value │ 8 │ ├────────────────────┼──────────────────────┤ │Minimum Value │ 1 │ ├────────────────────┼──────────────────────┤ │Maximum Value │ [platform dependent] │ └────────────────────┴──────────────────────┘ Specify the number of open file descriptors to reserve. • --password[=password], -p[password] ┌────────────────────┬───────────────────────┐ │Command-Line Format │ --password[=password] │ ├────────────────────┼───────────────────────┤ │Type │ String │ └────────────────────┴───────────────────────┘ The password of the MySQL account used for connecting to the server. The password value is optional. If not given, mysqlbinlog prompts for one. If given, there must be no space between --password= or -p and the password following it. If no password option is specified, the default is to send no password. Specifying a password on the command line should be considered insecure. To avoid giving the password on the command line, use an option file. See Section 6.1.2.1, “End-User Guidelines for Password Security”. To explicitly specify that there is no password and that mysqlbinlog should not prompt for one, use the --skip-password option. • --plugin-dir=dir_name ┌────────────────────┬───────────────────────┐ │Command-Line Format │ --plugin-dir=dir_name │ ├────────────────────┼───────────────────────┤ │Type │ Directory name │ └────────────────────┴───────────────────────┘ The directory in which to look for plugins. Specify this option if the --default-auth option is used to specify an authentication plugin but mysqlbinlog does not find it. See Section 6.2.17, “Pluggable Authentication”. • --port=port_num, -P port_num ┌────────────────────┬─────────────────┐ │Command-Line Format │ --port=port_num │ ├────────────────────┼─────────────────┤ │Type │ Numeric │ ├────────────────────┼─────────────────┤ │Default Value │ 3306 │ └────────────────────┴─────────────────┘ The TCP/IP port number to use for connecting to a remote server. • --print-defaults ┌────────────────────┬──────────────────┐ │Command-Line Format │ --print-defaults │ └────────────────────┴──────────────────┘ Print the program name and all options that it gets from option files. For additional information about this and other option-file options, see Section 4.2.2.3, “Command-Line Options that Affect Option-File Handling”. • --print-table-metadata ┌────────────────────┬────────────────────────┐ │Command-Line Format │ --print-table-metadata │ └────────────────────┴────────────────────────┘ Print table related metadata from the binary log. Configure the amount of table related metadata binary logged using binlog-row-metadata. • --protocol={TCP|SOCKET|PIPE|MEMORY} ┌────────────────────┬────────────────────────┐ │Command-Line Format │ --protocol=type │ ├────────────────────┼────────────────────────┤ │Type │ String │ ├────────────────────┼────────────────────────┤ │Default Value │ [see text] │ ├────────────────────┼────────────────────────┤ │Valid Values │ TCP SOCKET PIPE │ │ │ MEMORY │ └────────────────────┴────────────────────────┘ The transport protocol to use for connecting to the server. It is useful when the other connection parameters normally result in use of a protocol other than the one you want. For details on the permissible values, see Section 4.2.7, “Connection Transport Protocols”. • --raw ┌────────────────────┬─────────┐ │Command-Line Format │ --raw │ ├────────────────────┼─────────┤ │Type │ Boolean │ ├────────────────────┼─────────┤ │Default Value │ FALSE │ └────────────────────┴─────────┘ By default, mysqlbinlog reads binary log files and writes events in text format. The --raw option tells mysqlbinlog to write them in their original binary format. Its use requires that --read-from-remote-server also be used because the files are requested from a server. mysqlbinlog writes one output file for each file read from the server. The --raw option can be used to make a backup of a server's binary log. With the --stop-never option, the backup is “live” because mysqlbinlog stays connected to the server. By default, output files are written in the current directory with the same names as the original log files. Output file names can be modified using the --result-file option. For more information, see the section called “USING MYSQLBINLOG TO BACK UP BINARY LOG FILES”. • --read-from-remote-source=type ┌────────────────────┬────────────────────────────────┐ │Command-Line Format │ --read-from-remote-source=type │ └────────────────────┴────────────────────────────────┘ This option reads binary logs from a MySQL server with the COM_BINLOG_DUMP or COM_BINLOG_DUMP_GTID commands by setting the option value to either BINLOG-DUMP-NON-GTIDS or BINLOG-DUMP-GTIDS, respectively. If --read-from-remote-source=BINLOG-DUMP-GTIDS is combined with --exclude-gtids, transactions can be filtered out on the source, avoiding unnecessary network traffic. The connection parameter options are used with these options or the --read-from-remote-server option. These options are --host, --password, --port, --protocol, --socket, and --user. If none of the remote options is specified, the connection parameter options are ignored. The REPLICATION SLAVE privilege is required to use these options. • --read-from-remote-master=type ┌────────────────────┬────────────────────────────────┐ │Command-Line Format │ --read-from-remote-master=type │ ├────────────────────┼────────────────────────────────┤ │Deprecated │ Yes │ └────────────────────┴────────────────────────────────┘ Deprecated synonym for --read-from-remote-source. • --read-from-remote-server=file_name, -R ┌────────────────────┬─────────────────────────────────────┐ │Command-Line Format │ --read-from-remote-server=file_name │ └────────────────────┴─────────────────────────────────────┘ Read the binary log from a MySQL server rather than reading a local log file. This option requires that the remote server be running. It works only for binary log files on the remote server, not relay log files, and takes only the binary log file name (including the numeric suffix) as its argument, while ignoring any path. The connection parameter options are used with this option or the --read-from-remote-source option. These options are --host, --password, --port, --protocol, --socket, and --user. If neither of the remote options is specified, the connection parameter options are ignored. The REPLICATION SLAVE privilege is required to use this option. This option is like --read-from-remote-source=BINLOG-DUMP-NON-GTIDS. • --result-file=name, -r name ┌────────────────────┬────────────────────┐ │Command-Line Format │ --result-file=name │ └────────────────────┴────────────────────┘ Without the --raw option, this option indicates the file to which mysqlbinlog writes text output. With --raw, mysqlbinlog writes one binary output file for each log file transferred from the server, writing them by default in the current directory using the same names as the original log file. In this case, the --result-file option value is treated as a prefix that modifies output file names. • --require-row-format ┌────────────────────┬──────────────────────┐ │Command-Line Format │ --require-row-format │ ├────────────────────┼──────────────────────┤ │Type │ Boolean │ ├────────────────────┼──────────────────────┤ │Default Value │ false │ └────────────────────┴──────────────────────┘ Require row-based binary logging format for events. This option enforces row-based replication events for mysqlbinlog output. The stream of events produced with this option would be accepted by a replication channel that is secured using the REQUIRE_ROW_FORMAT option of the CHANGE REPLICATION SOURCE TO statement. binlog_format=ROW must be set on the server where the binary log was written. When you specify this option, mysqlbinlog stops with an error message if it encounters any events that are disallowed under the REQUIRE_ROW_FORMAT restrictions, including LOAD DATA INFILE instructions, creating or dropping temporary tables, INTVAR, RAND, or USER_VAR events, and non-row-based events within a DML transaction. mysqlbinlog also prints a SET @@session.require_row_format statement at the start of its output to apply the restrictions when the output is executed, and does not print the SET @@session.pseudo_thread_id statement. • --rewrite-db='from_name->to_name' ┌────────────────────┬─────────────────────────────────┐ │Command-Line Format │ --rewrite-db='oldname->newname' │ ├────────────────────┼─────────────────────────────────┤ │Type │ String │ ├────────────────────┼─────────────────────────────────┤ │Default Value │ [none] │ └────────────────────┴─────────────────────────────────┘ When reading from a row-based or statement-based log, rewrite all occurrences of from_name to to_name. Rewriting is done on the rows, for row-based logs, as well as on the USE clauses, for statement-based logs. Warning Statements in which table names are qualified with database names are not rewritten to use the new name when using this option. The rewrite rule employed as a value for this option is a string having the form 'from_name->to_name', as shown previously, and for this reason must be enclosed by quotation marks. To employ multiple rewrite rules, specify the option multiple times, as shown here: mysqlbinlog --rewrite-db='dbcurrent->dbold' --rewrite-db='dbtest->dbcurrent' \ binlog.00001 > /tmp/statements.sql When used together with the --database option, the --rewrite-db option is applied first; then --database option is applied, using the rewritten database name. The order in which the options are provided makes no difference in this regard. This means that, for example, if mysqlbinlog is started with --rewrite-db='mydb->yourdb' --database=yourdb, then all updates to any tables in databases mydb and yourdb are included in the output. On the other hand, if it is started with --rewrite-db='mydb->yourdb' --database=mydb, then mysqlbinlog outputs no statements at all: since all updates to mydb are first rewritten as updates to yourdb before applying the --database option, there remain no updates that match --database=mydb. • --server-id=id ┌────────────────────┬────────────────┐ │Command-Line Format │ --server-id=id │ ├────────────────────┼────────────────┤ │Type │ Numeric │ └────────────────────┴────────────────┘ Display only those events created by the server having the given server ID. • --server-id-bits=N ┌────────────────────┬────────────────────┐ │Command-Line Format │ --server-id-bits=# │ ├────────────────────┼────────────────────┤ │Type │ Numeric │ ├────────────────────┼────────────────────┤ │Default Value │ 32 │ ├────────────────────┼────────────────────┤ │Minimum Value │ 7 │ ├────────────────────┼────────────────────┤ │Maximum Value │ 32 │ └────────────────────┴────────────────────┘ Use only the first N bits of the server_id to identify the server. If the binary log was written by a mysqld with server-id-bits set to less than 32 and user data stored in the most significant bit, running mysqlbinlog with --server-id-bits set to 32 enables this data to be seen. This option is supported only by the version of mysqlbinlog supplied with the NDB Cluster distribution, or built with NDB Cluster support. • --server-public-key-path=file_name ┌────────────────────┬────────────────────────────────────┐ │Command-Line Format │ --server-public-key-path=file_name │ ├────────────────────┼────────────────────────────────────┤ │Type │ File name │ └────────────────────┴────────────────────────────────────┘ The path name to a file in PEM format containing a client-side copy of the public key required by the server for RSA key pair-based password exchange. This option applies to clients that authenticate with the sha256_password or caching_sha2_password authentication plugin. This option is ignored for accounts that do not authenticate with one of those plugins. It is also ignored if RSA-based password exchange is not used, as is the case when the client connects to the server using a secure connection. If --server-public-key-path=file_name is given and specifies a valid public key file, it takes precedence over --get-server-public-key. For sha256_password, this option applies only if MySQL was built using OpenSSL. For information about the sha256_password and caching_sha2_password plugins, see Section 6.4.1.3, “SHA-256 Pluggable Authentication”, and Section 6.4.1.2, “Caching SHA-2 Pluggable Authentication”. • --set-charset=charset_name ┌────────────────────┬────────────────────────────┐ │Command-Line Format │ --set-charset=charset_name │ ├────────────────────┼────────────────────────────┤ │Type │ String │ └────────────────────┴────────────────────────────┘ Add a SET NAMES charset_name statement to the output to specify the character set to be used for processing log files. • --shared-memory-base-name=name ┌────────────────────┬────────────────────────────────┐ │Command-Line Format │ --shared-memory-base-name=name │ ├────────────────────┼────────────────────────────────┤ │Platform Specific │ Windows │ └────────────────────┴────────────────────────────────┘ On Windows, the shared-memory name to use for connections made using shared memory to a local server. The default value is MYSQL. The shared-memory name is case-sensitive. This option applies only if the server was started with the shared_memory system variable enabled to support shared-memory connections. • --short-form, -s ┌────────────────────┬──────────────┐ │Command-Line Format │ --short-form │ └────────────────────┴──────────────┘ Display only the statements contained in the log, without any extra information or row-based events. This is for testing only, and should not be used in production systems. It is deprecated, and you should expect it to be removed in a future release. • --skip-gtids[=(true|false)] ┌────────────────────┬───────────────────────────┐ │Command-Line Format │ --skip-gtids[=true|false] │ ├────────────────────┼───────────────────────────┤ │Type │ Boolean │ ├────────────────────┼───────────────────────────┤ │Default Value │ false │ └────────────────────┴───────────────────────────┘ Do not include the GTIDs from the binary log files in the output dump file. For example: mysqlbinlog --skip-gtids binlog.000001 > /tmp/dump.sql mysql -u root -p -e "source /tmp/dump.sql" You should not normally use this option in production or in recovery, except in the specific, and rare, scenarios where the GTIDs are actively unwanted. For example, an administrator might want to duplicate selected transactions (such as table definitions) from a deployment to another, unrelated, deployment that will not replicate to or from the original. In that scenario, --skip-gtids can be used to enable the administrator to apply the transactions as if they were new, and ensure that the deployments remain unrelated. However, you should only use this option if the inclusion of the GTIDs causes a known issue for your use case. • --socket=path, -S path ┌────────────────────┬────────────────────────────────┐ │Command-Line Format │ --socket={file_name|pipe_name} │ ├────────────────────┼────────────────────────────────┤ │Type │ String │ └────────────────────┴────────────────────────────────┘ For connections to localhost, the Unix socket file to use, or, on Windows, the name of the named pipe to use. On Windows, this option applies only if the server was started with the named_pipe system variable enabled to support named-pipe connections. In addition, the user making the connection must be a member of the Windows group specified by the named_pipe_full_access_group system variable. • --ssl* Options that begin with --ssl specify whether to connect to the server using encryption and indicate where to find SSL keys and certificates. See the section called “Command Options for Encrypted Connections”. • --ssl-fips-mode={OFF|ON|STRICT} ┌────────────────────┬─────────────────────────────────┐ │Command-Line Format │ --ssl-fips-mode={OFF|ON|STRICT} │ ├────────────────────┼─────────────────────────────────┤ │Deprecated │ Yes │ ├────────────────────┼─────────────────────────────────┤ │Type │ Enumeration │ ├────────────────────┼─────────────────────────────────┤ │Default Value │ OFF │ ├────────────────────┼─────────────────────────────────┤ │Valid Values │ OFF ON STRICT │ └────────────────────┴─────────────────────────────────┘ Controls whether to enable FIPS mode on the client side. The --ssl-fips-mode option differs from other --ssl-xxx options in that it is not used to establish encrypted connections, but rather to affect which cryptographic operations to permit. See Section 6.8, “FIPS Support”. These --ssl-fips-mode values are permitted: • OFF: Disable FIPS mode. • ON: Enable FIPS mode. • STRICT: Enable “strict” FIPS mode. Note If the OpenSSL FIPS Object Module is not available, the only permitted value for --ssl-fips-mode is OFF. In this case, setting --ssl-fips-mode to ON or STRICT causes the client to produce a warning at startup and to operate in non-FIPS mode. This option is deprecated. Expect it to be removed in a future version of MySQL. • --start-datetime=datetime ┌────────────────────┬───────────────────────────┐ │Command-Line Format │ --start-datetime=datetime │ ├────────────────────┼───────────────────────────┤ │Type │ Datetime │ └────────────────────┴───────────────────────────┘ Start reading the binary log at the first event having a timestamp equal to or later than the datetime argument. The datetime value is relative to the local time zone on the machine where you run mysqlbinlog. The value should be in a format accepted for the DATETIME or TIMESTAMP data types. For example: mysqlbinlog --start-datetime="2005-12-25 11:25:56" binlog.000003 This option is useful for point-in-time recovery. See Section 7.5, “Point-in-Time (Incremental) Recovery”. • --start-position=N, -j N ┌────────────────────┬────────────────────┐ │Command-Line Format │ --start-position=# │ ├────────────────────┼────────────────────┤ │Type │ Numeric │ └────────────────────┴────────────────────┘ Start decoding the binary log at the log position N, including in the output any events that begin at position N or after. The position is a byte point in the log file, not an event counter; it needs to point to the starting position of an event to generate useful output. This option applies to the first log file named on the command line. Prior to MySQL 8.0.33, the maximum value supported for this option was 4294967295 (232-1). In MySQL 8.0.33 and later, it is 18446744073709551616 (264-1), unless --read-from-remote-server or --read-from-remote-source is also used, in which case the maximum is 4294967295. This option is useful for point-in-time recovery. See Section 7.5, “Point-in-Time (Incremental) Recovery”. • --stop-datetime=datetime ┌────────────────────┬──────────────────────────┐ │Command-Line Format │ --stop-datetime=datetime │ └────────────────────┴──────────────────────────┘ Stop reading the binary log at the first event having a timestamp equal to or later than the datetime argument. See the description of the --start-datetime option for information about the datetime value. This option is useful for point-in-time recovery. See Section 7.5, “Point-in-Time (Incremental) Recovery”. • --stop-never ┌────────────────────┬──────────────┐ │Command-Line Format │ --stop-never │ ├────────────────────┼──────────────┤ │Type │ Boolean │ ├────────────────────┼──────────────┤ │Default Value │ FALSE │ └────────────────────┴──────────────┘ This option is used with --read-from-remote-server. It tells mysqlbinlog to remain connected to the server. Otherwise mysqlbinlog exits when the last log file has been transferred from the server. --stop-never implies --to-last-log, so only the first log file to transfer need be named on the command line. --stop-never is commonly used with --raw to make a live binary log backup, but also can be used without --raw to maintain a continuous text display of log events as the server generates them. With --stop-never, by default, mysqlbinlog reports a server ID of 1 when it connects to the server. Use --connection-server-id to explicitly specify an alternative ID to report. It can be used to avoid a conflict with the ID of a replica server or another mysqlbinlog process. See the section called “SPECIFYING THE MYSQLBINLOG SERVER ID”. • --stop-never-slave-server-id=id ┌────────────────────┬────────────────────────────────┐ │Command-Line Format │ --stop-never-slave-server-id=# │ ├────────────────────┼────────────────────────────────┤ │Type │ Numeric │ ├────────────────────┼────────────────────────────────┤ │Default Value │ 65535 │ ├────────────────────┼────────────────────────────────┤ │Minimum Value │ 1 │ └────────────────────┴────────────────────────────────┘ This option is deprecated; expect it to be removed in a future release. Use the --connection-server-id option instead to specify a server ID for mysqlbinlog to report. • --stop-position=N ┌────────────────────┬───────────────────┐ │Command-Line Format │ --stop-position=# │ ├────────────────────┼───────────────────┤ │Type │ Numeric │ └────────────────────┴───────────────────┘ Stop decoding the binary log at the log position N, excluding from the output any events that begin at position N or after. The position is a byte point in the log file, not an event counter; it needs to point to a spot after the starting position of the last event you want to include in the output. The event starting before position N and finishing at or after the position is the last event to be processed. This option applies to the last log file named on the command line. This option is useful for point-in-time recovery. See Section 7.5, “Point-in-Time (Incremental) Recovery”. • --tls-ciphersuites=ciphersuite_list ┌────────────────────┬─────────────────────────────────────┐ │Command-Line Format │ --tls-ciphersuites=ciphersuite_list │ ├────────────────────┼─────────────────────────────────────┤ │Type │ String │ └────────────────────┴─────────────────────────────────────┘ The permissible ciphersuites for encrypted connections that use TLSv1.3. The value is a list of one or more colon-separated ciphersuite names. The ciphersuites that can be named for this option depend on the SSL library used to compile MySQL. For details, see Section 6.3.2, “Encrypted Connection TLS Protocols and Ciphers”. • --tls-sni-servername=server_name ┌────────────────────┬──────────────────────────────────┐ │Command-Line Format │ --tls-sni-servername=server_name │ ├────────────────────┼──────────────────────────────────┤ │Type │ String │ └────────────────────┴──────────────────────────────────┘ When specified, the name is passed to the libmysqlclient C API library using the MYSQL_OPT_TLS_SNI_SERVERNAME option of mysql_options(). The server name is not case-sensitive. To show which server name the client specified for the current session, if any, check the Tls_sni_server_name status variable. Server Name Indication (SNI) is an extension to the TLS protocol (OpenSSL must be compiled using TLS extensions for this option to function). The MySQL implementation of SNI represents the client-side only. • --tls-version=protocol_list ┌────────────────────┬───────────────────────────────┐ │Command-Line Format │ --tls-version=protocol_list │ ├────────────────────┼───────────────────────────────┤ │Type │ String │ ├────────────────────┼───────────────────────────────┤ │Default Value │ TLSv1,TLSv1.1,TLSv1.2,TLSv1.3 │ │ │ (OpenSSL 1.1.1 or higher) │ │ │ TLSv1,TLSv1.1,TLSv1.2 │ │ │ (otherwise) │ └────────────────────┴───────────────────────────────┘ The permissible TLS protocols for encrypted connections. The value is a list of one or more comma-separated protocol names. The protocols that can be named for this option depend on the SSL library used to compile MySQL. For details, see Section 6.3.2, “Encrypted Connection TLS Protocols and Ciphers”. • --to-last-log, -t ┌────────────────────┬───────────────┐ │Command-Line Format │ --to-last-log │ └────────────────────┴───────────────┘ Do not stop at the end of the requested binary log from a MySQL server, but rather continue printing until the end of the last binary log. If you send the output to the same MySQL server, this may lead to an endless loop. This option requires --read-from-remote-server. • --user=user_name, -u user_name ┌────────────────────┬───────────────────┐ │Command-Line Format │ --user=user_name, │ ├────────────────────┼───────────────────┤ │Type │ String │ └────────────────────┴───────────────────┘ The user name of the MySQL account to use when connecting to a remote server. If you are using the Rewriter plugin with MySQL 8.0.31 or later, you should grant this user the SKIP_QUERY_REWRITE privilege. • --verbose, -v ┌────────────────────┬───────────┐ │Command-Line Format │ --verbose │ └────────────────────┴───────────┘ Reconstruct row events and display them as commented SQL statements, with table partition information where applicable. If this option is given twice (by passing in either "-vv" or "--verbose --verbose"), the output includes comments to indicate column data types and some metadata, and informational log events such as row query log events if the binlog_rows_query_log_events system variable is set to TRUE. For examples that show the effect of --base64-output and --verbose on row event output, see the section called “MYSQLBINLOG ROW EVENT DISPLAY”. • --verify-binlog-checksum, -c ┌────────────────────┬──────────────────────────┐ │Command-Line Format │ --verify-binlog-checksum │ └────────────────────┴──────────────────────────┘ Verify checksums in binary log files. • --version, -V ┌────────────────────┬───────────┐ │Command-Line Format │ --version │ └────────────────────┴───────────┘ Display version information and exit. • --zstd-compression-level=level ┌────────────────────┬────────────────────────────┐ │Command-Line Format │ --zstd-compression-level=# │ ├────────────────────┼────────────────────────────┤ │Type │ Integer │ └────────────────────┴────────────────────────────┘ The compression level to use for connections to the server that use the zstd compression algorithm. The permitted levels are from 1 to 22, with larger values indicating increasing levels of compression. The default zstd compression level is 3. The compression level setting has no effect on connections that do not use zstd compression. For more information, see Section 4.2.8, “Connection Compression Control”. You can pipe the output of mysqlbinlog into the mysql client to execute the events contained in the binary log. This technique is used to recover from an unexpected exit when you have an old backup (see Section 7.5, “Point-in-Time (Incremental) Recovery”). For example: mysqlbinlog binlog.000001 | mysql -u root -p Or: mysqlbinlog binlog.[0-9]* | mysql -u root -p If the statements produced by mysqlbinlog may contain BLOB values, these may cause problems when mysql processes them. In this case, invoke mysql with the --binary-mode option. You can also redirect the output of mysqlbinlog to a text file instead, if you need to modify the statement log first (for example, to remove statements that you do not want to execute for some reason). After editing the file, execute the statements that it contains by using it as input to the mysql program: mysqlbinlog binlog.000001 > tmpfile ... edit tmpfile ... mysql -u root -p < tmpfile When mysqlbinlog is invoked with the --start-position option, it displays only those events with an offset in the binary log greater than or equal to a given position (the given position must match the start of one event). It also has options to stop and start when it sees an event with a given date and time. This enables you to perform point-in-time recovery using the --stop-datetime option (to be able to say, for example, “roll forward my databases to how they were today at 10:30 a.m.”). Processing multiple files. If you have more than one binary log to execute on the MySQL server, the safe method is to process them all using a single connection to the server. Here is an example that demonstrates what may be unsafe: mysqlbinlog binlog.000001 | mysql -u root -p # DANGER!! mysqlbinlog binlog.000002 | mysql -u root -p # DANGER!! Processing binary logs this way using multiple connections to the server causes problems if the first log file contains a CREATE TEMPORARY TABLE statement and the second log contains a statement that uses the temporary table. When the first mysql process terminates, the server drops the temporary table. When the second mysql process attempts to use the table, the server reports “unknown table.” To avoid problems like this, use a single mysql process to execute the contents of all binary logs that you want to process. Here is one way to do so: mysqlbinlog binlog.000001 binlog.000002 | mysql -u root -p Another approach is to write all the logs to a single file and then process the file: mysqlbinlog binlog.000001 > /tmp/statements.sql mysqlbinlog binlog.000002 >> /tmp/statements.sql mysql -u root -p -e "source /tmp/statements.sql" From MySQL 8.0.12, you can also supply multiple binary log files to mysqlbinlog as streamed input using a shell pipe. An archive of compressed binary log files can be decompressed and provided directly to mysqlbinlog. In this example, binlog-files_1.gz contains multiple binary log files for processing. The pipeline extracts the contents of binlog-files_1.gz, pipes the binary log files to mysqlbinlog as standard input, and pipes the output of mysqlbinlog into the mysql client for execution: gzip -cd binlog-files_1.gz | ./mysqlbinlog - | ./mysql -uroot -p You can specify more than one archive file, for example: gzip -cd binlog-files_1.gz binlog-files_2.gz | ./mysqlbinlog - | ./mysql -uroot -p For streamed input, do not use --stop-position, because mysqlbinlog cannot identify the last log file to apply this option. LOAD DATA operations. mysqlbinlog can produce output that reproduces a LOAD DATA operation without the original data file. mysqlbinlog copies the data to a temporary file and writes a LOAD DATA LOCAL statement that refers to the file. The default location of the directory where these files are written is system-specific. To specify a directory explicitly, use the --local-load option. Because mysqlbinlog converts LOAD DATA statements to LOAD DATA LOCAL statements (that is, it adds LOCAL), both the client and the server that you use to process the statements must be configured with the LOCAL capability enabled. See Section 6.1.6, “Security Considerations for LOAD DATA LOCAL”. Warning The temporary files created for LOAD DATA LOCAL statements are not automatically deleted because they are needed until you actually execute those statements. You should delete the temporary files yourself after you no longer need the statement log. The files can be found in the temporary file directory and have names like original_file_name-#-#. MYSQLBINLOG HEX DUMP FORMAT The --hexdump option causes mysqlbinlog to produce a hex dump of the binary log contents: mysqlbinlog --hexdump source-bin.000001 The hex output consists of comment lines beginning with #, so the output might look like this for the preceding command: /*!40019 SET @@SESSION.max_insert_delayed_threads=0*/; /*!50003 SET @OLD_COMPLETION_TYPE=@@COMPLETION_TYPE,COMPLETION_TYPE=0*/; # at 4 #051024 17:24:13 server id 1 end_log_pos 98 # Position Timestamp Type Master ID Size Master Pos Flags # 00000004 9d fc 5c 43 0f 01 00 00 00 5e 00 00 00 62 00 00 00 00 00 # 00000017 04 00 35 2e 30 2e 31 35 2d 64 65 62 75 67 2d 6c |..5.0.15.debug.l| # 00000027 6f 67 00 00 00 00 00 00 00 00 00 00 00 00 00 00 |og..............| # 00000037 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 |................| # 00000047 00 00 00 00 9d fc 5c 43 13 38 0d 00 08 00 12 00 |.......C.8......| # 00000057 04 04 04 04 12 00 00 4b 00 04 1a |.......K...| # Start: binlog v 4, server v 5.0.15-debug-log created 051024 17:24:13 # at startup ROLLBACK; Hex dump output currently contains the elements in the following list. This format is subject to change. For more information about binary log format, see MySQL Internals: The Binary Log[1]. • Position: The byte position within the log file. • Timestamp: The event timestamp. In the example shown, '9d fc 5c 43' is the representation of '051024 17:24:13' in hexadecimal. • Type: The event type code. • Master ID: The server ID of the replication source server that created the event. • Size: The size in bytes of the event. • Master Pos: The position of the next event in the original source's binary log file. • Flags: Event flag values. MYSQLBINLOG ROW EVENT DISPLAY The following examples illustrate how mysqlbinlog displays row events that specify data modifications. These correspond to events with the WRITE_ROWS_EVENT, UPDATE_ROWS_EVENT, and DELETE_ROWS_EVENT type codes. The --base64-output=DECODE-ROWS and --verbose options may be used to affect row event output. Suppose that the server is using row-based binary logging and that you execute the following sequence of statements: CREATE TABLE t ( id INT NOT NULL, name VARCHAR(20) NOT NULL, date DATE NULL ) ENGINE = InnoDB; START TRANSACTION; INSERT INTO t VALUES(1, 'apple', NULL); UPDATE t SET name = 'pear', date = '2009-01-01' WHERE id = 1; DELETE FROM t WHERE id = 1; COMMIT; By default, mysqlbinlog displays row events encoded as base-64 strings using BINLOG statements. Omitting extraneous lines, the output for the row events produced by the preceding statement sequence looks like this: $> mysqlbinlog log_file ... # at 218 #080828 15:03:08 server id 1 end_log_pos 258 Write_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAANoAAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBcBAAAAKAAAAAIBAAAQABEAAAAAAAEAA//8AQAAAAVhcHBsZQ== '/*!*/; ... # at 302 #080828 15:03:08 server id 1 end_log_pos 356 Update_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAAC4BAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBgBAAAANgAAAGQBAAAQABEAAAAAAAEAA////AEAAAAFYXBwbGX4AQAAAARwZWFyIbIP '/*!*/; ... # at 400 #080828 15:03:08 server id 1 end_log_pos 442 Delete_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAAJABAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBkBAAAAKgAAALoBAAAQABEAAAAAAAEAA//4AQAAAARwZWFyIbIP '/*!*/; To see the row events as comments in the form of “pseudo-SQL” statements, run mysqlbinlog with the --verbose or -v option. This output level also shows table partition information where applicable. The output contains lines beginning with ###: $> mysqlbinlog -v log_file ... # at 218 #080828 15:03:08 server id 1 end_log_pos 258 Write_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAANoAAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBcBAAAAKAAAAAIBAAAQABEAAAAAAAEAA//8AQAAAAVhcHBsZQ== '/*!*/; ### INSERT INTO test.t ### SET ### @1=1 ### @2='apple' ### @3=NULL ... # at 302 #080828 15:03:08 server id 1 end_log_pos 356 Update_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAAC4BAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBgBAAAANgAAAGQBAAAQABEAAAAAAAEAA////AEAAAAFYXBwbGX4AQAAAARwZWFyIbIP '/*!*/; ### UPDATE test.t ### WHERE ### @1=1 ### @2='apple' ### @3=NULL ### SET ### @1=1 ### @2='pear' ### @3='2009:01:01' ... # at 400 #080828 15:03:08 server id 1 end_log_pos 442 Delete_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAAJABAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBkBAAAAKgAAALoBAAAQABEAAAAAAAEAA//4AQAAAARwZWFyIbIP '/*!*/; ### DELETE FROM test.t ### WHERE ### @1=1 ### @2='pear' ### @3='2009:01:01' Specify --verbose or -v twice to also display data types and some metadata for each column, and informational log events such as row query log events if the binlog_rows_query_log_events system variable is set to TRUE. The output contains an additional comment following each column change: $> mysqlbinlog -vv log_file ... # at 218 #080828 15:03:08 server id 1 end_log_pos 258 Write_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAANoAAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBcBAAAAKAAAAAIBAAAQABEAAAAAAAEAA//8AQAAAAVhcHBsZQ== '/*!*/; ### INSERT INTO test.t ### SET ### @1=1 /* INT meta=0 nullable=0 is_null=0 */ ### @2='apple' /* VARSTRING(20) meta=20 nullable=0 is_null=0 */ ### @3=NULL /* VARSTRING(20) meta=0 nullable=1 is_null=1 */ ... # at 302 #080828 15:03:08 server id 1 end_log_pos 356 Update_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAAC4BAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBgBAAAANgAAAGQBAAAQABEAAAAAAAEAA////AEAAAAFYXBwbGX4AQAAAARwZWFyIbIP '/*!*/; ### UPDATE test.t ### WHERE ### @1=1 /* INT meta=0 nullable=0 is_null=0 */ ### @2='apple' /* VARSTRING(20) meta=20 nullable=0 is_null=0 */ ### @3=NULL /* VARSTRING(20) meta=0 nullable=1 is_null=1 */ ### SET ### @1=1 /* INT meta=0 nullable=0 is_null=0 */ ### @2='pear' /* VARSTRING(20) meta=20 nullable=0 is_null=0 */ ### @3='2009:01:01' /* DATE meta=0 nullable=1 is_null=0 */ ... # at 400 #080828 15:03:08 server id 1 end_log_pos 442 Delete_rows: table id 17 flags: STMT_END_F BINLOG ' fAS3SBMBAAAALAAAAJABAAAAABEAAAAAAAAABHRlc3QAAXQAAwMPCgIUAAQ= fAS3SBkBAAAAKgAAALoBAAAQABEAAAAAAAEAA//4AQAAAARwZWFyIbIP '/*!*/; ### DELETE FROM test.t ### WHERE ### @1=1 /* INT meta=0 nullable=0 is_null=0 */ ### @2='pear' /* VARSTRING(20) meta=20 nullable=0 is_null=0 */ ### @3='2009:01:01' /* DATE meta=0 nullable=1 is_null=0 */ You can tell mysqlbinlog to suppress the BINLOG statements for row events by using the --base64-output=DECODE-ROWS option. This is similar to --base64-output=NEVER but does not exit with an error if a row event is found. The combination of --base64-output=DECODE-ROWS and --verbose provides a convenient way to see row events only as SQL statements: $> mysqlbinlog -v --base64-output=DECODE-ROWS log_file ... # at 218 #080828 15:03:08 server id 1 end_log_pos 258 Write_rows: table id 17 flags: STMT_END_F ### INSERT INTO test.t ### SET ### @1=1 ### @2='apple' ### @3=NULL ... # at 302 #080828 15:03:08 server id 1 end_log_pos 356 Update_rows: table id 17 flags: STMT_END_F ### UPDATE test.t ### WHERE ### @1=1 ### @2='apple' ### @3=NULL ### SET ### @1=1 ### @2='pear' ### @3='2009:01:01' ... # at 400 #080828 15:03:08 server id 1 end_log_pos 442 Delete_rows: table id 17 flags: STMT_END_F ### DELETE FROM test.t ### WHERE ### @1=1 ### @2='pear' ### @3='2009:01:01' Note You should not suppress BINLOG statements if you intend to re-execute mysqlbinlog output. The SQL statements produced by --verbose for row events are much more readable than the corresponding BINLOG statements. However, they do not correspond exactly to the original SQL statements that generated the events. The following limitations apply: • The original column names are lost and replaced by @N, where N is a column number. • Character set information is not available in the binary log, which affects string column display: • There is no distinction made between corresponding binary and nonbinary string types (BINARY and CHAR, VARBINARY and VARCHAR, BLOB and TEXT). The output uses a data type of STRING for fixed-length strings and VARSTRING for variable-length strings. • For multibyte character sets, the maximum number of bytes per character is not present in the binary log, so the length for string types is displayed in bytes rather than in characters. For example, STRING(4) is used as the data type for values from either of these column types: CHAR(4) CHARACTER SET latin1 CHAR(2) CHARACTER SET ucs2 • Due to the storage format for events of type UPDATE_ROWS_EVENT, UPDATE statements are displayed with the WHERE clause preceding the SET clause. Proper interpretation of row events requires the information from the format description event at the beginning of the binary log. Because mysqlbinlog does not know in advance whether the rest of the log contains row events, by default it displays the format description event using a BINLOG statement in the initial part of the output. If the binary log is known not to contain any events requiring a BINLOG statement (that is, no row events), the --base64-output=NEVER option can be used to prevent this header from being written. USING MYSQLBINLOG TO BACK UP BINARY LOG FILES By default, mysqlbinlog reads binary log files and displays their contents in text format. This enables you to examine events within the files more easily and to re-execute them (for example, by using the output as input to mysql). mysqlbinlog can read log files directly from the local file system, or, with the --read-from-remote-server option, it can connect to a server and request binary log contents from that server. mysqlbinlog writes text output to its standard output, or to the file named as the value of the --result-file=file_name option if that option is given. • mysqlbinlog Backup Capabilities • mysqlbinlog Backup Options • Static and Live Backups • Output File Naming • Example: mysqldump + mysqlbinlog for Backup and Restore • mysqlbinlog Backup Restrictions mysqlbinlog Backup Capabilities mysqlbinlog can read binary log files and write new files containing the same content—that is, in binary format rather than text format. This capability enables you to easily back up a binary log in its original format. mysqlbinlog can make a static backup, backing up a set of log files and stopping when the end of the last file is reached. It can also make a continuous (“live”) backup, staying connected to the server when it reaches the end of the last log file and continuing to copy new events as they are generated. In continuous-backup operation, mysqlbinlog runs until the connection ends (for example, when the server exits) or mysqlbinlog is forcibly terminated. When the connection ends, mysqlbinlog does not wait and retry the connection, unlike a replica server. To continue a live backup after the server has been restarted, you must also restart mysqlbinlog. Important mysqlbinlog can back up both encrypted and unencrypted binary log files . However, copies of encrypted binary log files that are generated using mysqlbinlog are stored in an unencrypted format. mysqlbinlog Backup Options Binary log backup requires that you invoke mysqlbinlog with two options at minimum: • The --read-from-remote-server (or -R) option tells mysqlbinlog to connect to a server and request its binary log. (This is similar to a replica server connecting to its replication source server.) • The --raw option tells mysqlbinlog to write raw (binary) output, not text output. Along with --read-from-remote-server, it is common to specify other options: --host indicates where the server is running, and you may also need to specify connection options such as --user and --password. Several other options are useful in conjunction with --raw: • --stop-never: Stay connected to the server after reaching the end of the last log file and continue to read new events. • --connection-server-id=id: The server ID that mysqlbinlog reports when it connects to a server. When --stop-never is used, the default reported server ID is 1. If this causes a conflict with the ID of a replica server or another mysqlbinlog process, use --connection-server-id to specify an alternative server ID. See the section called “SPECIFYING THE MYSQLBINLOG SERVER ID”. • --result-file: A prefix for output file names, as described later. Static and Live Backups To back up a server's binary log files with mysqlbinlog, you must specify file names that actually exist on the server. If you do not know the names, connect to the server and use the SHOW BINARY LOGS statement to see the current names. Suppose that the statement produces this output: mysql> SHOW BINARY LOGS; +---------------+-----------+-----------+ | Log_name | File_size | Encrypted | +---------------+-----------+-----------+ | binlog.000130 | 27459 | No | | binlog.000131 | 13719 | No | | binlog.000132 | 43268 | No | +---------------+-----------+-----------+ With that information, you can use mysqlbinlog to back up the binary log to the current directory as follows (enter each command on a single line): • To make a static backup of binlog.000130 through binlog.000132, use either of these commands: mysqlbinlog --read-from-remote-server --host=host_name --raw binlog.000130 binlog.000131 binlog.000132 mysqlbinlog --read-from-remote-server --host=host_name --raw --to-last-log binlog.000130 The first command specifies every file name explicitly. The second names only the first file and uses --to-last-log to read through the last. A difference between these commands is that if the server happens to open binlog.000133 before mysqlbinlog reaches the end of binlog.000132, the first command does not read it, but the second command does. • To make a live backup in which mysqlbinlog starts with binlog.000130 to copy existing log files, then stays connected to copy new events as the server generates them: mysqlbinlog --read-from-remote-server --host=host_name --raw --stop-never binlog.000130 With --stop-never, it is not necessary to specify --to-last-log to read to the last log file because that option is implied. Output File Naming Without --raw, mysqlbinlog produces text output and the --result-file option, if given, specifies the name of the single file to which all output is written. With --raw, mysqlbinlog writes one binary output file for each log file transferred from the server. By default, mysqlbinlog writes the files in the current directory with the same names as the original log files. To modify the output file names, use the --result-file option. In conjunction with --raw, the --result-file option value is treated as a prefix that modifies the output file names. Suppose that a server currently has binary log files named binlog.000999 and up. If you use mysqlbinlog --raw to back up the files, the --result-file option produces output file names as shown in the following table. You can write the files to a specific directory by beginning the --result-file value with the directory path. If the --result-file value consists only of a directory name, the value must end with the pathname separator character. Output files are overwritten if they exist. ┌─────────────────────┬────────────────────────────┐ │--result-file Option │ Output File Names │ ├─────────────────────┼────────────────────────────┤ │--result-file=x │ xbinlog.000999 and up │ ├─────────────────────┼────────────────────────────┤ │--result-file=/tmp/ │ /tmp/binlog.000999 and up │ ├─────────────────────┼────────────────────────────┤ │--result-file=/tmp/x │ /tmp/xbinlog.000999 and up │ └─────────────────────┴────────────────────────────┘ Example: mysqldump + mysqlbinlog for Backup and Restore The following example describes a simple scenario that shows how to use mysqldump and mysqlbinlog together to back up a server's data and binary log, and how to use the backup to restore the server if data loss occurs. The example assumes that the server is running on host host_name and its first binary log file is named binlog.000999. Enter each command on a single line. Use mysqlbinlog to make a continuous backup of the binary log: mysqlbinlog --read-from-remote-server --host=host_name --raw --stop-never binlog.000999 Use mysqldump to create a dump file as a snapshot of the server's data. Use --all-databases, --events, and --routines to back up all data, and --source-data=2 to include the current binary log coordinates in the dump file. mysqldump --host=host_name --all-databases --events --routines --source-data=2> dump_file Execute the mysqldump command periodically to create newer snapshots as desired. If data loss occurs (for example, if the server unexpectedly exits), use the most recent dump file to restore the data: mysql --host=host_name -u root -p < dump_file Then use the binary log backup to re-execute events that were written after the coordinates listed in the dump file. Suppose that the coordinates in the file look like this: -- CHANGE REPLICATION SOURCE TO SOURCE_LOG_FILE='binlog.001002', SOURCE_LOG_POS=27284; If the most recent backed-up log file is named binlog.001004, re-execute the log events like this: mysqlbinlog --start-position=27284 binlog.001002 binlog.001003 binlog.001004 | mysql --host=host_name -u root -p You might find it easier to copy the backup files (dump file and binary log files) to the server host to make it easier to perform the restore operation, or if MySQL does not allow remote root access. mysqlbinlog Backup Restrictions Binary log backups with mysqlbinlog are subject to these restrictions: • mysqlbinlog does not automatically reconnect to the MySQL server if the connection is lost (for example, if a server restart occurs or there is a network outage). • The delay for a backup is similar to the delay for a replica server. SPECIFYING THE MYSQLBINLOG SERVER ID When invoked with the --read-from-remote-server option, mysqlbinlog connects to a MySQL server, specifies a server ID to identify itself, and requests binary log files from the server. You can use mysqlbinlog to request log files from a server in several ways: • Specify an explicitly named set of files: For each file, mysqlbinlog connects and issues a Binlog dump command. The server sends the file and disconnects. There is one connection per file. • Specify the beginning file and --to-last-log: mysqlbinlog connects and issues a Binlog dump command for all files. The server sends all files and disconnects. • Specify the beginning file and --stop-never (which implies --to-last-log): mysqlbinlog connects and issues a Binlog dump command for all files. The server sends all files, but does not disconnect after sending the last one. With --read-from-remote-server only, mysqlbinlog connects using a server ID of 0, which tells the server to disconnect after sending the last requested log file. With --read-from-remote-server and --stop-never, mysqlbinlog connects using a nonzero server ID, so the server does not disconnect after sending the last log file. The server ID is 1 by default, but this can be changed with --connection-server-id. Thus, for the first two ways of requesting files, the server disconnects because mysqlbinlog specifies a server ID of 0. It does not disconnect if --stop-never is given because mysqlbinlog specifies a nonzero server ID. COPYRIGHT Copyright © 1997, 2023, Oracle and/or its affiliates. This documentation is free software; you can redistribute it and/or modify it only under the terms of the GNU General Public License as published by the Free Software Foundation; version 2 of the License. This documentation is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with the program; if not, write to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA or see http://www.gnu.org/licenses/. NOTES 1. MySQL Internals: The Binary Log https://dev.mysql.com/doc/internals/en/binary-log.html SEE ALSO For more information, please refer to the MySQL Reference Manual, which may already be installed locally and which is also available online at http://dev.mysql.com/doc/. AUTHOR Oracle Corporation (http://dev.mysql.com/). MySQL 8.3 11/23/2023 MYSQLBINLOG(1)
mysqlbinlog - utility for processing binary log files
mysqlbinlog [options] log_file ...
null
null
funpack
null
null
null
null
null
pyrsa-sign
null
null
null
null
null
rcc
null
null
null
null
null
onnxruntime_test
null
null
null
null
null
dask-ssh
null
null
null
null
null
marker_chunk_convert
null
null
null
null
null
iusql
isql and iusql are command-line tools allowing users to execute SQL interactively or in batches. The tools provide several useful features, including an option to generate output wrapped in an HTML table. iusql is the same as isql but includes built-in Unicode support. Some data sources only work with iusql. An important difference between the two tools is that isql connects using SQLConnect and iusql connects using SQLDriverConnect. ARGUMENTS DSN The Data Source Name (DSN) used to connect to the SQL database. unixODBC looks for the specified DSN in /etc/odbc.ini and $HOME/.odbc.ini, with the latter taking precedence. When searching the configuration files, unixODBC looks for a bare name. If the DSN begins with a semicolon, it is treated as a connection string. The connection string can contain a DSN and/or other semicolon-separated parameters. USER Specifies the database user or role under which the connection should be made. This parameter overrides any UID specified in the data source configuration files. PASSWORD Password required to access the database for the specified USER. This parameter overrides any PASSWORD specified in the data source configuration files. When using iusql, passwords containing semicolons should be escaped with braces (curly brackets) and terminated with a semicolon. Refer to the Examples section below for syntax. "ConnectionString" A connection string starting with DSN=, DRIVER= or FILEDSN= will be passed unchanged to SQLDriverConnect. This option allows for the use of more complex syntax in a connection string than would otherwise be possible by just using DSN, UID and PWD. It also (and this was the main reason for its inclusion) allows passwords to contain semicolons without having to add complex escape syntax to the existing code.
isql, iusql - unixODBC interactive SQL command-line tools
isql DSN [USER [PASSWORD]] [options] iusql DSN [USER [PASSWORD]] [options] iusql "ConnectionString" [options]
-b Run 'isql' in non-interactive batch mode. In this mode, 'isql' processes from standard input, expecting one SQL command per line. -dDELIMITER Delimit columns with the specified delimiter. -xHEX Delimit columns with the character represented in hexadecimal by HEX. The hexadecimal code must be in the format 0xNN (e.g. 0x09 for the TAB character). -w Format the result as an HTML table. -c Output the names of the columns on the first row. This option can only be used with the -d or -x options. -mNUM Limit the column display width to NUM characters. -lLOCALE Set the character locale to LOCALE. -q Wrap the character fields in double quotes. -3 Use calls from ODBC version 3. -n Process multiple lines of SQL, terminated with the GO command. -e Use SQLExecDirect instead of Prepare. -k Use SQLDriverConnect. -v Enable verbose mode, fully describing all errors. This option is useful for debugging. --version Display the program version. -LNUM Set the maximum number of characters displayed from a character field to NUM. The default value is 300 characters. COMMANDS This section briefly describes some isql and iusql run-time commands. help List all tables in the database. help table List all columns in the table. help help List all help options.
A bare DSN: $ iusql WebDB MyID MyPWD -w -b < My.sql Connects to the WebDB DSN as user MyID with password MyPWD, then executes the commands in the My.sql file and returns the results wrapped in an HTML table. Each line in My.sql must only contain one SQL command, except for the last line, which must be blank (unless the -n option is specified). A DSN in a connection string: Note the leading semicolon on the connection string. $ iusql ";DSN=WebDB" MyID MyPWD -w -b < My.sql Options in the DSN may be overridden in the connection string: $ iusql ";DSN=WebDB;Driver=PostgreSQL ODBC;UID=MyID;PASSWORD=secret;Debug=1;CommLog=1" -v A string DSN: A string DSN may be provided in its entirety, with no file DSN reference at all: $ iusql ";Driver=PostgreSQL Unicode;UID=MyID;PASSWORD=secret" -v A password containing a semicolon (iusql): $ iusql WebDB MyID '{My;PWD};' $ iusql 'DSN=WebDB;UID=MyID;PWD={My;PWD};' TROUBLESHOOTING Cryptic error messages Re-run isql or iusql with the -v flag to get more information from errors, and/or enable Trace mode in odbcinst.ini. Missing driver definition Check that the driver name specified by the Driver entry in the odbc.ini data-source definition is present in odbcinst.ini and exactly matches the odbcinst.ini [section name]. Unloadable or incompatible driver If the ODBC driver is properly specified for the data source, it is possible that the driver is not loadable. Check for mix-ups between Unicode and ANSI drivers, and verify the driver paths in the odbcinst.ini [section name]. Unicode data sources with ANSI clients Some data sources are Unicode-only and require the use of iusql. If isql reports [IM002][unixODBC][Driver Manager]Data source name not found and no default driver specified [ISQL]ERROR: Could not SQLConnect but the data source and driver required are listed by odbcinst -q -d and odbcinst -q -s then try iusql. FILES /etc/odbc.ini Configuration file containing system-wide Data Source Name (DSN) definitions. See odbc.ini(5) for more information. $HOME/.odbc.ini Configuration file containing user-specific Data Source Name (DSN) definitions. See odbc.ini(5) for more information. SEE ALSO unixODBC(7), odbcinst(1), odbc.ini(5) "The unixODBC Administrator Manual (HTML)" AUTHORS The authors of unixODBC are Peter Harvey <pharvey@codebydesign.com> and Nick Gorham <nick@lurcher.org>. For a full list of contributors, refer to the AUTHORS file. COPYRIGHT unixODBC is licensed under the GNU Lesser General Public License. For details about the license, see the COPYING file. version 2.3.12 Thu 14 Jan 2021 isql(1)
tificc
lcms is a standalone CMM engine, which deals with the color management. It implements a fast transformation between ICC profiles. tificc is a little cms ICC profile applier for TIFF.
tificc - little cms ICC profile applier for TIFF.
tificc [options] input.tif output.tif
-a Handle channels > 4 as alpha. -b Black point compensation. -c NUM Precalculates transform (0=Off, 1=Normal, 2=Hi-res, 3=LoRes) [defaults to 1]. -d NUM Observer adaptation state (abs.col. only), (0..1.0, float value) [defaults to 0.0]. -e Embed destination profile. -g Marks out-of-gamut colors on softproof. -h NUM Show summary of options and examples (0=help, 1=Examples, 2=Built-in profiles, 3=Contact information) -i profile Input profile (defaults to sRGB). -k inklimit Ink-limiting in % (CMYK only), (0..400.0, float value) [default 400.0]. -l profile Transform by device-link profile. -m TODO: check if values outside 0..3 are possible SoftProof intent [defaults to 0]. -n Ignore embedded profile on input. -o profile Output profile (defaults to sRGB). -p profile Soft proof profile. -s newprofile Save embedded profile as newprofile. -t NUM Rendering intent 0=Perceptual [default] 1=Relative colorimetric 2=Saturation 3=Absolute colorimetric 10=Perceptual preserving black ink 11=Relative colorimetric preserving black ink 12=Saturation preserving black ink 13=Perceptual preserving black plane 14=Relative colorimetric preserving black plane 15=Saturation preserving black plane -v Verbose. -w NUM Output depth (8, 16 or 32). Use 32 for floating-point. BUILT-IN PROFILES *Lab2 -- D50-based v2 CIEL*a*b *Lab4 -- D50-based v4 CIEL*a*b *Lab -- D50-based v4 CIEL*a*b *XYZ -- CIE XYZ (PCS) *sRGB -- sRGB color space *Gray22 - Monochrome of Gamma 2.2 *Gray30 - Monochrome of Gamma 3.0 *null - Monochrome black for all input *Lin2222- CMYK linearization of gamma 2.2 on each channel
To color correct from scanner to sRGB: tificc -iscanner.icm in.tif out.tif To convert from monitor1 to monitor2: tificc -imon1.icm -omon2.icm in.tif out.tif To make a CMYK separation: tificc -oprinter.icm inrgb.tif outcmyk.tif To recover sRGB from a CMYK separation: tificc -iprinter.icm incmyk.tif outrgb.tif To convert from CIELab TIFF to sRGB tificc -i*Lab in.tif out.tif NOTES For suggestions, comments, bug reports etc. send mail to info@littlecms.com. SEE ALSO jpgicc(1), linkicc(1), psicc(1), transicc(1) AUTHOR This manual page was originally written by Shiju p. Nair <shiju.p@gmail.com>, for the Debian project. Modified by Marti Maria to reflect further changes. October 23, 2004 TIFICC(1)
unstructured-ingest
null
null
null
null
null
markdown_py
null
null
null
null
null
bzegrep
Bzgrep is used to invoke the grep on bzip2-compressed files. All options specified are passed directly to grep. If no file is specified, then the standard input is decompressed if necessary and fed to grep. Otherwise the given files are uncompressed if necessary and fed to grep. If bzgrep is invoked as bzegrep or bzfgrep then egrep or fgrep is used instead of grep. If the GREP environment variable is set, bzgrep uses it as the grep program to be invoked. For example: for sh: GREP=fgrep bzgrep string files for csh: (setenv GREP fgrep; bzgrep string files) AUTHOR Charles Levert (charles@comm.polymtl.ca). Adapted to bzip2 by Philippe Troin <phil@fifi.org> for Debian GNU/Linux. SEE ALSO grep(1), egrep(1), fgrep(1), bzdiff(1), bzmore(1), bzless(1), bzip2(1) BZGREP(1)
bzgrep, bzfgrep, bzegrep - search possibly bzip2 compressed files for a regular expression
bzgrep [ grep_options ] [ -e ] pattern filename... bzegrep [ egrep_options ] [ -e ] pattern filename... bzfgrep [ fgrep_options ] [ -e ] pattern filename...
null
null
jupyter-run
null
null
null
null
null
gst-inspect-1.0
gst-inspect-1.0 is a tool that prints out information on available GStreamer plugins, information about a particular plugin, or information about a particular element. When executed with no PLUGIN or ELEMENT argument, gst-inspect-1.0 will print a list of all plugins and elements together with a summary. When executed with a PLUGIN or ELEMENT argument, gst-inspect-1.0 will print information about that plug-in or element.
gst-inspect-1.0 - print info about a GStreamer plugin or element
gst-inspect-1.0 [OPTION...] [PLUGIN|ELEMENT]
gst-inspect-1.0 accepts the following arguments and options: PLUGIN Name of a plugin ELEMENT Name of an element --help Print help synopsis and available FLAGS --gst-info-mask=FLAGS GStreamer info flags to set (list with --help) -a, --print-all Print all plugins and elements -b, --print-blacklist Print list of blacklisted files --plugin List the plugin contents --types=Element/Types Allow inspecting only elements that match all the element types filtered in this slash (´/´) separated list of element types. Those types correspond to what is also called ´klass´ which is a string describing the type of element, like ´Decoder´, ´Audio´, ´Encoder´ etc... This options implies that only elements will be printed (not typefind functions or tracers). --exists Check if the specified element or plugin exists --atleast-version When checking if an element or plugin exists, also check that its version is at least the version specified -u, --uri-handlers Print supported URI schemes, with the elements that implement them --print-plugin-auto-install-info Print a machine-parsable list of features the specified plugin provides. Useful in connection with external automatic plugin installation mechanisms. --gst-debug-mask=FLAGS GStreamer debugging flags to set (list with --help) --gst-mask=FLAGS GStreamer info and debugging flags to set (list with --help) --gst-plugin-spew GStreamer info flags to set Enable printout of errors while loading GStreamer plugins --gst-plugin-path=PATH Add directories separated with ':' to the plugin search path SEE ALSO gst-launch-1.0(1), gst-typefind-1.0(1) AUTHOR The GStreamer team at http://gstreamer.freedesktop.org/ December 2005 GStreamer(1)
null
pyjson5
null
null
null
null
null
nghttp
null
null
null
null
null
conda-build
null
null
null
null
null
h5copy
null
null
null
null
null
c_rehash
This command is generally equivalent to the external script c_rehash, except for minor differences noted below. openssl rehash scans directories and calculates a hash value of each .pem, .crt, .cer, or .crl file in the specified directory list and creates symbolic links for each file, where the name of the link is the hash value. (If the platform does not support symbolic links, a copy is made.) This command is useful as many programs that use OpenSSL require directories to be set up like this in order to find certificates. If any directories are named on the command line, then those are processed in turn. If not, then the SSL_CERT_DIR environment variable is consulted; this should be a colon-separated list of directories, like the Unix PATH variable. If that is not set then the default directory (installation-specific but often /usr/local/ssl/certs) is processed. In order for a directory to be processed, the user must have write permissions on that directory, otherwise an error will be generated. The links created are of the form HHHHHHHH.D, where each H is a hexadecimal character and D is a single decimal digit. When a directory is processed, all links in it that have a name in that syntax are first removed, even if they are being used for some other purpose. To skip the removal step, use the -n flag. Hashes for CRL's look similar except the letter r appears after the period, like this: HHHHHHHH.rD. Multiple objects may have the same hash; they will be indicated by incrementing the D value. Duplicates are found by comparing the full SHA-1 fingerprint. A warning will be displayed if a duplicate is found. A warning will also be displayed if there are files that cannot be parsed as either a certificate or a CRL or if more than one such object appears in the file. Script Configuration The c_rehash script uses the openssl program to compute the hashes and fingerprints. If not found in the user's PATH, then set the OPENSSL environment variable to the full pathname. Any program can be used, it will be invoked as follows for either a certificate or CRL: $OPENSSL x509 -hash -fingerprint -noout -in FILENAME $OPENSSL crl -hash -fingerprint -noout -in FILENAME where FILENAME is the filename. It must output the hash of the file on the first line, and the fingerprint on the second, optionally prefixed with some text and an equals sign.
openssl-rehash, c_rehash - Create symbolic links to files named by the hash values
openssl rehash [-h] [-help] [-old] [-compat] [-n] [-v] [-provider name] [-provider-path path] [-propquery propq] [directory] ... c_rehash [-h] [-help] [-old] [-n] [-v] [-provider name] [-provider-path path] [-propquery propq] [directory] ...
-help -h Display a brief usage message. -old Use old-style hashing (MD5, as opposed to SHA-1) for generating links to be used for releases before 1.0.0. Note that current versions will not use the old style. -n Do not remove existing links. This is needed when keeping new and old-style links in the same directory. -compat Generate links for both old-style (MD5) and new-style (SHA1) hashing. This allows releases before 1.0.0 to use these links along-side newer releases. -v Print messages about old links removed and new links created. By default, this command only lists each directory as it is processed. -provider name -provider-path path -propquery propq See "Provider Options" in openssl(1), provider(7), and property(7). ENVIRONMENT OPENSSL The path to an executable to use to generate hashes and fingerprints (see above). SSL_CERT_DIR Colon separated list of directories to operate on. Ignored if directories are listed on the command line. SEE ALSO openssl(1), openssl-crl(1), openssl-x509(1) COPYRIGHT Copyright 2015-2020 The OpenSSL Project Authors. All Rights Reserved. Licensed under the Apache License 2.0 (the "License"). You may not use this file except in compliance with the License. You can obtain a copy in the file LICENSE in the source distribution or at <https://www.openssl.org/source/license.html>. 3.3.1 2024-06-04 OPENSSL-REHASH(1ssl)
null
k5srvutil
k5srvutil allows an administrator to list keys currently in a keytab, to obtain new keys for a principal currently in a keytab, or to delete non-current keys from a keytab. operation must be one of the following: list Lists the keys in a keytab, showing version number and principal name. change Uses the kadmin protocol to update the keys in the Kerberos database to new randomly-generated keys, and updates the keys in the keytab to match. If a key's version number doesn't match the version number stored in the Kerberos server's database, then the operation will fail. If the -i flag is given, k5srvutil will prompt for confirmation before changing each key. If the -k option is given, the old and new keys will be displayed. Ordinarily, keys will be generated with the default encryption types and key salts. This can be overridden with the -e option. Old keys are retained in the keytab so that existing tickets continue to work, but delold should be used after such tickets expire, to prevent attacks against the old keys. delold Deletes keys that are not the most recent version from the keytab. This operation should be used some time after a change operation to remove old keys, after existing tickets issued for the service have expired. If the -i flag is given, then k5srvutil will prompt for confirmation for each principal. delete Deletes particular keys in the keytab, interactively prompting for each key. In all cases, the default keytab is used unless this is overridden by the -f option. k5srvutil uses the kadmin(1) program to edit the keytab in place. ENVIRONMENT See kerberos(7) for a description of Kerberos environment variables. SEE ALSO kadmin(1), ktutil(1), kerberos(7) AUTHOR MIT COPYRIGHT 1985-2022, MIT 1.20.1 K5SRVUTIL(1)
k5srvutil - host key table (keytab) manipulation utility
k5srvutil operation [-i] [-f filename] [-e keysalts]
null
null
bjam
null
null
null
null
null
xzdec
xzdec is a liblzma-based decompression-only tool for .xz (and only .xz) files. xzdec is intended to work as a drop-in replacement for xz(1) in the most common situations where a script has been written to use xz --decompress --stdout (and possibly a few other commonly used options) to decompress .xz files. lzmadec is identical to xzdec except that lzmadec supports .lzma files instead of .xz files. To reduce the size of the executable, xzdec doesn't support multithreading or localization, and doesn't read options from XZ_DEFAULTS and XZ_OPT environment variables. xzdec doesn't support displaying intermediate progress information: sending SIGINFO to xzdec does nothing, but sending SIGUSR1 terminates the process instead of displaying progress information.
xzdec, lzmadec - Small .xz and .lzma decompressors
xzdec [option...] [file...] lzmadec [option...] [file...]
-d, --decompress, --uncompress Ignored for xz(1) compatibility. xzdec supports only decompression. -k, --keep Ignored for xz(1) compatibility. xzdec never creates or removes any files. -c, --stdout, --to-stdout Ignored for xz(1) compatibility. xzdec always writes the decompressed data to standard output. -q, --quiet Specifying this once does nothing since xzdec never displays any warnings or notices. Specify this twice to suppress errors. -Q, --no-warn Ignored for xz(1) compatibility. xzdec never uses the exit status 2. -h, --help Display a help message and exit successfully. -V, --version Display the version number of xzdec and liblzma. EXIT STATUS 0 All was good. 1 An error occurred. xzdec doesn't have any warning messages like xz(1) has, thus the exit status 2 is not used by xzdec. NOTES Use xz(1) instead of xzdec or lzmadec for normal everyday use. xzdec or lzmadec are meant only for situations where it is important to have a smaller decompressor than the full-featured xz(1). xzdec and lzmadec are not really that small. The size can be reduced further by dropping features from liblzma at compile time, but that shouldn't usually be done for executables distributed in typical non- embedded operating system distributions. If you need a truly small .xz decompressor, consider using XZ Embedded. SEE ALSO xz(1) XZ Embedded: <https://tukaani.org/xz/embedded.html> Tukaani 2024-04-08 XZDEC(1)
null
bokeh
null
null
null
null
null
tiffsplit
tiffsplit takes a multi-directory (page) TIFF file and creates one or more single-directory (page) TIFF files from it. The output files are given names created by concatenating a prefix, a lexically ordered suffix in the range [aaa--zzz], the suffix .tif (e.g. xaaa.tif, xaab.tif, …, xzzz.tif). If a prefix is not specified on the command line, the default prefix of x is used.
tiffsplit - split a multi-image TIFF into single-image TIFF files
tiffsplit src.tif [ prefix ]
None. EXIT STATUS tiffsplit exits with one of the following values: 0: Success 1: An error occurred either reading the input or writing results. BUGS Only a select set of "known tags" are copied when splitting. SEE ALSO tiffcp (1), tiffinfo (1), libtiff (3tiff), AUTHOR LibTIFF contributors COPYRIGHT 1988-2022, LibTIFF contributors 4.6 September 8, 2023 TIFFSPLIT(1)
null
bzdiff
Bzcmp and bzdiff are used to invoke the cmp or the diff program on bzip2 compressed files. All options specified are passed directly to cmp or diff. If only 1 file is specified, then the files compared are file1 and an uncompressed file1.bz2. If two files are specified, then they are uncompressed if necessary and fed to cmp or diff. The exit status from cmp or diff is preserved. SEE ALSO cmp(1), diff(1), bzmore(1), bzless(1), bzgrep(1), bzip2(1) BUGS Messages from the cmp or diff programs refer to temporary filenames instead of those specified. BZDIFF(1)
bzcmp, bzdiff - compare bzip2 compressed files
bzcmp [ cmp_options ] file1 [ file2 ] bzdiff [ diff_options ] file1 [ file2 ]
null
null
dltest
dltest is a simple program that determines whether a symbol appears in a (shared object) library. The name of the library must be prefixed by a full system path. If no parameters are provided, dltest prints a short help message.
dltest - A simple library symbol test program
dltest [ library symbol ]
null
To determine if the symbol printf is found in libc-2.18.so, run: $ dltest /usr/lib/libc-2.18.so printf AUTHORS The authors of unixODBC are Peter Harvey <pharvey@codebydesign.com> and Nick Gorham <nick@lurcher.org>. For a full list of contributors, refer to the AUTHORS file. COPYRIGHT unixODBC is licensed under the GNU Lesser General Public License. For details about the license, see the COPYING file. version 2.3.12 Thu 07 Jan 2021 dltest(1)
sip-module
null
null
null
null
null
repo2solv
The repo2solv tool converts repository metadata in the directory DIR into a solv file written to standard output. Note that repo2solv does not verify signatures or checksum, it is expected that this is done by the tool that downloads the metadata. If no metadata is detected, repo2solv assumes the "plaindir" format and generates the solv file from all rpm files it finds. -o OUTFILE Write the solv file to OUTFILE instead of stdout. -R Also recurse into subdirectories in "plaindir" mode. -F Put the complete filelist in the output. The default is to just include the "importent" parts of the file list, except for "plaindir" mode, which always includes all files. -C Add changelog entires to the output. -A Add appdata pseudo packages to the output. This is an experimental feature. -X Autoexpand SUSE pattern and product provides into packages. SEE ALSO dumpsolv(1) AUTHOR Michael Schroeder <mls@suse.de> libsolv 10/22/2018 REPO2SOLV(1)
repo2solv - convert repository metadata into a solv file
repo2solv [OPTIONS] DIR
null
null
orc-statistics
null
null
null
null
null
mysql_config_editor
The mysql_config_editor utility enables you to store authentication credentials in an obfuscated login path file named .mylogin.cnf. The file location is the %APPDATA%\MySQL directory on Windows and the current user's home directory on non-Windows systems. The file can be read later by MySQL client programs to obtain authentication credentials for connecting to MySQL Server. The unobfuscated format of the .mylogin.cnf login path file consists of option groups, similar to other option files. Each option group in .mylogin.cnf is called a “login path,” which is a group that permits only certain options: host, user, password, port and socket. Think of a login path option group as a set of options that specify which MySQL server to connect to and which account to authenticate as. Here is an unobfuscated example: [client] user = mydefaultname password = mydefaultpass host = 127.0.0.1 [mypath] user = myothername password = myotherpass host = localhost When you invoke a client program to connect to the server, the client uses .mylogin.cnf in conjunction with other option files. Its precedence is higher than other option files, but less than options specified explicitly on the client command line. For information about the order in which option files are used, see Section 4.2.2.2, “Using Option Files”. To specify an alternate login path file name, set the MYSQL_TEST_LOGIN_FILE environment variable. This variable is recognized by mysql_config_editor, by standard MySQL clients (mysql, mysqladmin, and so forth), and by the mysql-test-run.pl testing utility. Programs use groups in the login path file as follows: • mysql_config_editor operates on the client login path by default if you specify no --login-path=name option to indicate explicitly which login path to use. • Without a --login-path option, client programs read the same option groups from the login path file that they read from other option files. Consider this command: mysql By default, the mysql client reads the [client] and [mysql] groups from other option files, so it reads them from the login path file as well. • With a --login-path option, client programs additionally read the named login path from the login path file. The option groups read from other option files remain the same. Consider this command: mysql --login-path=mypath The mysql client reads [client] and [mysql] from other option files, and [client], [mysql], and [mypath] from the login path file. • Client programs read the login path file even when the --no-defaults option is used, unless --no-login-paths is set. This permits passwords to be specified in a safer way than on the command line even if --no-defaults is present. mysql_config_editor obfuscates the .mylogin.cnf file so it cannot be read as cleartext, and its contents when unobfuscated by client programs are used only in memory. In this way, passwords can be stored in a file in non-cleartext format and used later without ever needing to be exposed on the command line or in an environment variable. mysql_config_editor provides a print command for displaying the login path file contents, but even in this case, password values are masked so as never to appear in a way that other users can see them. The obfuscation used by mysql_config_editor prevents passwords from appearing in .mylogin.cnf as cleartext and provides a measure of security by preventing inadvertent password exposure. For example, if you display a regular unobfuscated my.cnf option file on the screen, any passwords it contains are visible for anyone to see. With .mylogin.cnf, that is not true, but the obfuscation used is not likely to deter a determined attacker and you should not consider it unbreakable. A user who can gain system administration privileges on your machine to access your files could unobfuscate the .mylogin.cnf file with some effort. The login path file must be readable and writable to the current user, and inaccessible to other users. Otherwise, mysql_config_editor ignores it, and client programs do not use it, either. Invoke mysql_config_editor like this: mysql_config_editor [program_options] command [command_options] If the login path file does not exist, mysql_config_editor creates it. Command arguments are given as follows: • program_options consists of general mysql_config_editor options. • command indicates what action to perform on the .mylogin.cnf login path file. For example, set writes a login path to the file, remove removes a login path, and print displays login path contents. • command_options indicates any additional options specific to the command, such as the login path name and the values to use in the login path. The position of the command name within the set of program arguments is significant. For example, these command lines have the same arguments, but produce different results: mysql_config_editor --help set mysql_config_editor set --help The first command line displays a general mysql_config_editor help message, and ignores the set command. The second command line displays a help message specific to the set command. Suppose that you want to establish a client login path that defines your default connection parameters, and an additional login path named remote for connecting to the MySQL server the host remote.example.com. You want to log in as follows: • By default, to the local server with a user name and password of localuser and localpass • To the remote server with a user name and password of remoteuser and remotepass To set up the login paths in the .mylogin.cnf file, use the following set commands. Enter each command on a single line, and enter the appropriate passwords when prompted: $> mysql_config_editor set --login-path=client --host=localhost --user=localuser --password Enter password: enter password "localpass" here $> mysql_config_editor set --login-path=remote --host=remote.example.com --user=remoteuser --password Enter password: enter password "remotepass" here mysql_config_editor uses the client login path by default, so the --login-path=client option can be omitted from the first command without changing its effect. To see what mysql_config_editor writes to the .mylogin.cnf file, use the print command: $> mysql_config_editor print --all [client] user = localuser password = ***** host = localhost [remote] user = remoteuser password = ***** host = remote.example.com The print command displays each login path as a set of lines beginning with a group header indicating the login path name in square brackets, followed by the option values for the login path. Password values are masked and do not appear as cleartext. If you do not specify --all to display all login paths or --login-path=name to display a named login path, the print command displays the client login path by default, if there is one. As shown by the preceding example, the login path file can contain multiple login paths. In this way, mysql_config_editor makes it easy to set up multiple “personalities” for connecting to different MySQL servers, or for connecting to a given server using different accounts. Any of these can be selected by name later using the --login-path option when you invoke a client program. For example, to connect to the remote server, use this command: mysql --login-path=remote Here, mysql reads the [client] and [mysql] option groups from other option files, and the [client], [mysql], and [remote] groups from the login path file. To connect to the local server, use this command: mysql --login-path=client Because mysql reads the client and mysql login paths by default, the --login-path option does not add anything in this case. That command is equivalent to this one: mysql Options read from the login path file take precedence over options read from other option files. Options read from login path groups appearing later in the login path file take precedence over options read from groups appearing earlier in the file. mysql_config_editor adds login paths to the login path file in the order you create them, so you should create more general login paths first and more specific paths later. If you need to move a login path within the file, you can remove it, then recreate it to add it to the end. For example, a client login path is more general because it is read by all client programs, whereas a mysqldump login path is read only by mysqldump. Options specified later override options specified earlier, so putting the login paths in the order client, mysqldump enables mysqldump-specific options to override client options. When you use the set command with mysql_config_editor to create a login path, you need not specify all possible option values (host name, user name, password, port, socket). Only those values given are written to the path. Any missing values required later can be specified when you invoke a client path to connect to the MySQL server, either in other option files or on the command line. Any options specified on the command line override those specified in the login path file or other option files. For example, if the credentials in the remote login path also apply for the host remote2.example.com, connect to the server on that host like this: mysql --login-path=remote --host=remote2.example.com mysql_config_editor General Options mysql_config_editor supports the following general options, which may be used preceding any command named on the command line. For descriptions of command-specific options, see mysql_config_editor Commands and Command-Specific Options. • --help, -? ┌────────────────────┬────────┐ │Command-Line Format │ --help │ └────────────────────┴────────┘ Display a general help message and exit. To see a command-specific help message, invoke mysql_config_editor as follows, where command is a command other than help: mysql_config_editor command --help • --debug[=debug_options], -# debug_options ┌────────────────────┬─────────────────────────┐ │Command-Line Format │ --debug[=debug_options] │ ├────────────────────┼─────────────────────────┤ │Type │ String │ ├────────────────────┼─────────────────────────┤ │Default Value │ d:t:o │ └────────────────────┴─────────────────────────┘ Write a debugging log. A typical debug_options string is d:t:o,file_name. The default is d:t:o,/tmp/mysql_config_editor.trace. This option is available only if MySQL was built using WITH_DEBUG. MySQL release binaries provided by Oracle are not built using this option. • --verbose, -v ┌────────────────────┬───────────┐ │Command-Line Format │ --verbose │ └────────────────────┴───────────┘ Verbose mode. Print more information about what the program does. This option may be helpful in diagnosing problems if an operation does not have the effect you expect. • --version, -V ┌────────────────────┬───────────┐ │Command-Line Format │ --version │ └────────────────────┴───────────┘ Display version information and exit. mysql_config_editor Commands and Command-Specific Options This section describes the permitted mysql_config_editor commands, and, for each one, the command-specific options permitted following the command name on the command line. In addition, mysql_config_editor supports general options that can be used preceding any command. For descriptions of these options, see mysql_config_editor General Options. mysql_config_editor supports these commands: • help Display a general help message and exit. This command takes no following options. To see a command-specific help message, invoke mysql_config_editor as follows, where command is a command other than help: mysql_config_editor command --help • print [options] Print the contents of the login path file in unobfuscated form, with the exception that passwords are displayed as *****. The default login path name is client if no login path is named. If both --all and --login-path are given, --all takes precedence. The print command permits these options following the command name: • --help, -? Display a help message for the print command and exit. To see a general help message, use mysql_config_editor --help. • --all Print the contents of all login paths in the login path file. • --login-path=name, -G name Print the contents of the named login path. • remove [options] Remove a login path from the login path file, or modify a login path by removing options from it. This command removes from the login path only such options as are specified with the --host, --password, --port, --socket, and --user options. If none of those options are given, remove removes the entire login path. For example, this command removes only the user option from the mypath login path rather than the entire mypath login path: mysql_config_editor remove --login-path=mypath --user This command removes the entire mypath login path: mysql_config_editor remove --login-path=mypath The remove command permits these options following the command name: • --help, -? Display a help message for the remove command and exit. To see a general help message, use mysql_config_editor --help. • --host, -h Remove the host name from the login path. • --login-path=name, -G name The login path to remove or modify. The default login path name is client if this option is not given. • --password, -p Remove the password from the login path. • --port, -P Remove the TCP/IP port number from the login path. • --socket, -S Remove the Unix socket file name from the login path. • --user, -u Remove the user name from the login path. • --warn, -w Warn and prompt the user for confirmation if the command attempts to remove the default login path (client) and --login-path=client was not specified. This option is enabled by default; use --skip-warn to disable it. • reset [options] Empty the contents of the login path file. The reset command permits these options following the command name: • --help, -? Display a help message for the reset command and exit. To see a general help message, use mysql_config_editor --help. • set [options] Write a login path to the login path file. This command writes to the login path only such options as are specified with the --host, --password, --port, --socket, and --user options. If none of those options are given, mysql_config_editor writes the login path as an empty group. The set command permits these options following the command name: • --help, -? Display a help message for the set command and exit. To see a general help message, use mysql_config_editor --help. • --host=host_name, -h host_name The host name to write to the login path. • --login-path=name, -G name The login path to create. The default login path name is client if this option is not given. • --password, -p Prompt for a password to write to the login path. After mysql_config_editor displays the prompt, type the password and press Enter. To prevent other users from seeing the password, mysql_config_editor does not echo it. To specify an empty password, press Enter at the password prompt. The resulting login path written to the login path file includes a line like this: password = • --port=port_num, -P port_num The TCP/IP port number to write to the login path. • --socket=file_name, -S file_name The Unix socket file name to write to the login path. • --user=user_name, -u user_name The user name to write to the login path. • --warn, -w Warn and prompt the user for confirmation if the command attempts to overwrite an existing login path. This option is enabled by default; use --skip-warn to disable it. COPYRIGHT Copyright © 1997, 2023, Oracle and/or its affiliates. This documentation is free software; you can redistribute it and/or modify it only under the terms of the GNU General Public License as published by the Free Software Foundation; version 2 of the License. This documentation is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with the program; if not, write to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA or see http://www.gnu.org/licenses/. SEE ALSO For more information, please refer to the MySQL Reference Manual, which may already be installed locally and which is also available online at http://dev.mysql.com/doc/. AUTHOR Oracle Corporation (http://dev.mysql.com/). MySQL 8.3 11/23/2023 MYSQL_CONFIG_EDITOR(1)
mysql_config_editor - configure authentication information for connecting to MySQL server
mysql_config_editor options command
null
null
jupyter-labextension
null
null
null
null
null
strings
Strings looks for ASCII strings in a binary file or standard input. Strings is useful for identifying random object files and many other things. A string is any sequence of 4 (the default) or more printing characters [ending at, but not including, any other character or EOF]. Unless the - flag is given, strings looks in all sections of the object files except the (__TEXT,__text) section. If no files are specified standard input is read. The file arguments may be of the form libx.a(foo.o), to request information about only that object file and not the entire library. (Typically this argument must be quoted, ``libx.a(foo.o)'', to get it past the shell.) The options to strings(1) are: -a This option causes strings to look for strings in all sections of the object file (including the (__TEXT,__text) section. - This option causes strings to look for strings in all bytes of the files (the default for non-object files). -- This option causes strings to treat all the following arguments as files. -o Preceded each string by its offset in the file (in decimal). -t format Write each string preceded by its byte offset from the start of the file. The format shall be dependent on the single character used as the format option-argument: d The offset shall be written in decimal. o The offset shall be written in octal. x The offset shall be written in hexadecimal. -number The decimal number is used as the minimum string length rather than the default of 4. -n number Specify the minimum string length, where the number argument is a positive decimal integer. The default shall be 4. -arch arch_type Specifies the architecture, arch_type, of the file for strings(1) to operate on when the file is a universal file. (See arch(3) for the currently know arch_types.) The arch_type can be "all" to operate on all architectures in the file. SEE ALSO od(1) BUGS The algorithm for identifying strings is extremely primitive. Apple, Inc. June 7, 2016 STRINGS(1)
strings - find the printable strings in a object, or other binary, file
strings [ - ] [ -a ] [ -o ] [ -t format ] [ -number ] [ -n number ] [--] [file ...]
null
null
langsmith
null
null
null
null
null
lrelease-pro
null
null
null
null
null
h52gif
null
null
null
null
null
yapf-diff
null
null
null
null
null
fax2tiff
null
null
null
null
null
ptrepack
null
null
null
null
null
jupyter-nbclassic-serverextension
null
null
null
null
null
pylint
null
null
null
null
null
linkicc
lcms is a standalone CMM engine, which deals with the color management. It implements a fast transformation between ICC profiles. linkicc is a little cms device link generator. Links two or more profiles into a single devicelink profile. Colorspaces must be paired except Lab/XYZ, that can be interchanged.
linkicc - little cms device link generator.
linkicc [options] profiles
-a NUM Observer adaptation state (abs.col. only), (0..1.0, float value) [defaults to 1.0]. -b Black point compensation. -c precision Precision (0=LowRes, 1=Normal, 2=Hi-res) [defaults to 1]. -d description Description text (quotes can be used). -h NUM Show summary of options and examples (0=help, 1=Built-in profiles, 2=Examples, 3=Contact information) -k inklimit Ink-limiting in % (CMYK only), (0..400.0, float value) [default 400.0]. -l Use linearization curves (may affect accuracy). -n gridpoints Alternate way to set precision, number of CLUT points. -o profile Output devicelink profile [defaults to 'devicelink.icm']. -r profileversion Profile version. (CAUTION: may change the profile implementation), (2.0..4.3, float value) [defaults to 4.3]. -t NUM Rendering intent 0=Perceptual [default] 1=Relative colorimetric 2=Saturation 3=Absolute colorimetric 10=Perceptual preserving black ink 11=Relative colorimetric preserving black ink 12=Saturation preserving black ink 13=Perceptual preserving black plane 14=Relative colorimetric preserving black plane 15=Saturation preserving black plane -v verbosity Verbosity level, (0=None, 1=Normal, 2=High, 3=Very High) [defaults to 0]. -x Creatively, guess deviceclass of resulting profile. -y copyright Copyright notice (quotes can be used) ["No copyright, use freely"]. -8 Creates 8-bit devicelink. BUILT-IN PROFILES *Lab2 -- D50-based v2 CIEL*a*b *Lab4 -- D50-based v4 CIEL*a*b *Lab -- D50-based v4 CIEL*a*b *XYZ -- CIE XYZ (PCS) *sRGB -- sRGB color space *Gray22 - Monochrome of Gamma 2.2 *Gray30 - Monochrome of Gamma 3.0 *null - Monochrome black for all input *Lin2222- CMYK linearization of gamma 2.2 on each channel
To create 'devicelink.icm' from a.icc to b.icc: linkicc a.icc b.icc To create 'out.icc' from sRGB to cmyk.icc: linkicc -o out.icc *sRGB cmyk.icc To create a sRGB input profile working in Lab: linkicc -x -o sRGBLab.icc *sRGB *Lab To create a XYZ -> sRGB output profile: linkicc -x -o sRGBLab.icc *XYZ *sRGB To create a abstract profile doing softproof for cmyk.icc: linkicc -t1 -x -o softproof.icc *Lab cmyk.icc cmyk.icc *Lab To create a 'grayer' sRGB input profile: linkicc -x -o grayer.icc *sRGB gray.icc gray.icc *Lab To embed ink limiting into a cmyk output profile: linkicc -x -o cmyklimited.icc -k 250 cmyk.icc *Lab NOTES For suggestions, comments, bug reports etc. send mail to info@littlecms.com. SEE ALSO jpgicc(1), psicc(1), tificc(1), transicc(1) AUTHOR This manual page was written by Shiju p. Nair <shiju.p@gmail.com>, for the Debian project. September 30, 2004 LINKICC(1)
rst2xetex.py
null
null
null
null
null
msgcomm
Find messages which are common to two or more of the specified PO files. By using the --more-than option, greater commonality may be requested before messages are printed. Conversely, the --less-than option may be used to specify less commonality before messages are printed (i.e. --less-than=2 will only print the unique messages). Translations, comments and extracted comments will be preserved, but only from the first PO file to define them. File positions from all PO files will be cumulated. Mandatory arguments to long options are mandatory for short options too. Input file location: INPUTFILE ... input files -f, --files-from=FILE get list of input files from FILE -D, --directory=DIRECTORY add DIRECTORY to list for input files search If input file is -, standard input is read. Output file location: -o, --output-file=FILE write output to specified file The results are written to standard output if no output file is specified or if it is -. Message selection: -<, --less-than=NUMBER print messages with less than this many definitions, defaults to infinite if not set ->, --more-than=NUMBER print messages with more than this many definitions, defaults to 1 if not set -u, --unique shorthand for --less-than=2, requests that only unique messages be printed Input file syntax: -P, --properties-input input files are in Java .properties syntax --stringtable-input input files are in NeXTstep/GNUstep .strings syntax Output details: --color use colors and other text attributes always --color=WHEN use colors and other text attributes if WHEN. WHEN may be 'always', 'never', 'auto', or 'html'. --style=STYLEFILE specify CSS style rule file for --color -e, --no-escape do not use C escapes in output (default) -E, --escape use C escapes in output, no extended chars --force-po write PO file even if empty -i, --indent write the .po file using indented style --no-location do not write '#: filename:line' lines -n, --add-location generate '#: filename:line' lines (default) --strict write out strict Uniforum conforming .po file -p, --properties-output write out a Java .properties file --stringtable-output write out a NeXTstep/GNUstep .strings file -w, --width=NUMBER set output page width --no-wrap do not break long message lines, longer than the output page width, into several lines -s, --sort-output generate sorted output -F, --sort-by-file sort output by file location --omit-header don't write header with 'msgid ""' entry Informative output: -h, --help display this help and exit -V, --version output version information and exit AUTHOR Written by Peter Miller. REPORTING BUGS Report bugs in the bug tracker at <https://savannah.gnu.org/projects/gettext> or by email to <bug-gettext@gnu.org>. COPYRIGHT Copyright © 1995-2023 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <https://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. SEE ALSO The full documentation for msgcomm is maintained as a Texinfo manual. If the info and msgcomm programs are properly installed at your site, the command info msgcomm should give you access to the complete manual. GNU gettext-tools 0.22.5 February 2024 MSGCOMM(1)
msgcomm - match two message catalogs
msgcomm [OPTION] [INPUTFILE]...
null
null
ktutil
The ktutil command invokes a command interface from which an administrator can read, write, or edit entries in a keytab. (Kerberos V4 srvtab files are no longer supported.) COMMANDS list list [-t] [-k] [-e] Displays the current keylist. If -t, -k, and/or -e are specified, also display the timestamp, key contents, or enctype (respectively). Alias: l read_kt read_kt keytab Read the Kerberos V5 keytab file keytab into the current keylist. Alias: rkt write_kt write_kt keytab Write the current keylist into the Kerberos V5 keytab file keytab. Alias: wkt clear_list clear_list Clear the current keylist. Alias: clear delete_entry delete_entry slot Delete the entry in slot number slot from the current keylist. Alias: delent add_entry add_entry {-key|-password} -p principal -k kvno [-e enctype] [-f|-s salt] Add principal to keylist using key or password. If the -f flag is specified, salt information will be fetched from the KDC; in this case the -e flag may be omitted, or it may be supplied to force a particular enctype. If the -f flag is not specified, the -e flag must be specified, and the default salt will be used unless overridden with the -s option. Alias: addent list_requests list_requests Displays a listing of available commands. Aliases: lr, ? quit quit Quits ktutil. Aliases: exit, q EXAMPLE ktutil: add_entry -password -p alice@BLEEP.COM -k 1 -e aes128-cts-hmac-sha1-96 Password for alice@BLEEP.COM: ktutil: add_entry -password -p alice@BLEEP.COM -k 1 -e aes256-cts-hmac-sha1-96 Password for alice@BLEEP.COM: ktutil: write_kt alice.keytab ktutil: ENVIRONMENT See kerberos(7) for a description of Kerberos environment variables. SEE ALSO kadmin(1), kdb5_util(8), kerberos(7) AUTHOR MIT COPYRIGHT 1985-2022, MIT 1.20.1 KTUTIL(1)
ktutil - Kerberos keytab file maintenance utility
ktutil
null
null
pycodestyle
null
null
null
null
null
infotocap
infotocap looks in each given text file for terminfo descriptions. For each terminfo description found, an equivalent termcap description is written to standard output. Terminfo use capabilities are translated directly to termcap tc capabilities. -v print out tracing information on standard error as the program runs. -V print out the version of the program in use on standard error and exit. -1 cause the fields to print out one to a line. Otherwise, the fields will be printed several to a line to a maximum width of 60 characters. -w change the output to width characters. FILES /usr/share/terminfo Compiled terminal description database. NOTES This utility is actually a link to tic, running in -C mode. You can use other tic options such as -f and -x. SEE ALSO curses(3X), tic(1M), infocmp(1M), terminfo(5) This describes ncurses version 5.7 (patch 20081102). AUTHOR Eric S. Raymond <esr@snark.thyrsus.com> and Thomas E. Dickey <dickey@invisible-island.net> infotocap(1M)
infotocap - convert a terminfo description into a termcap description
infotocap [-vn width] [-V] [-1] [-w width] file . . .
null
null
h5perf_serial
null
null
null
null
null
imagecodecs
null
null
null
null
null
opentelemetry-instrument
null
null
null
null
null
cph
null
null
null
null
null
gencnval
gencnval converts the ICU aliases file converterfile into the binary file cnvalias.icu. This binary file can then be read directly by ICU, or used by pkgdata(1) for incorporation into a larger archive or library. If converterfile is not provided, the default ICU convrtrs.txt file is used.
gencnval - compile the converters aliases file
gencnval [ -h, -?, --help ] [ -v, --verbose ] [ -c, --copyright ] [ -s, --sourcedir source ] [ -d, --destdir destination ] [ converterfile ]
-h, -?, --help Print help about usage and exit. -v, --verbose Display verbose output. This information can include information about conflicting aliases and the converters the aliases resolve to. -c, --copyright Include a copyright notice in the binary data. -s, --sourcedir source Set the source directory to source. The default source directory is specified by the environment variable ICU_DATA. -d, --destdir destination Set the destination directory to destination. The default destination directory is specified by the environment variable ICU_DATA. ENVIRONMENT ICU_DATA Specifies the directory containing ICU data. Defaults to ${prefix}/share/icu/68.1/. Some tools in ICU depend on the presence of the trailing slash. It is thus important to make sure that it is present if ICU_DATA is set. FILES icu/source/data/mappings/convrtrs.txt Description of ICU's converters and their aliases. This data file is not normally installed, and it is available as a part of ICU source code. VERSION 68.1 COPYRIGHT Copyright (C) 2000-2004 IBM, Inc. and others. SEE ALSO pkgdata(1) ICU MANPAGE 11 March 2004 GENCNVAL(1)
null
opj_compress
null
opj_compress - This program reads in an image of a certain type and converts it to a jpeg2000 file. It is part of the OpenJPEG library. Valid input image extensions are .bmp, .pgm, .pgx, .png, .pnm, .ppm, .raw, .tga, .tif . For PNG resp. TIF it needs libpng resp. libtiff . Valid output image extensions are .j2k, .jp2
opj_compress -i infile.bmp -o outfile.j2k opj_compress -ImgDir directory_name -OutFor jp2 opj_compress -h Print a help message and exit. See JPWL OPTIONS for special options
-b n,n (Size of code block (e.g. -b 32,32). Default: 64 x 64) -c n (Size of precinct (e.g. -c 128,128). Default: 2^15 x 2^15) -cinema2K fps Digital Cinema 2K profile compliant codestream. Valid fps values are 24 or 48. -cinema4K Digital Cinema 4K profile compliant codestream. Does not need an fps: default is 24 fps. -d X,Y (Offset of image origin (e.g. -d 150,300)) -h Print a help message and exit. -i name (input file name) -n n (Number of resolutions. Default: 6) -o name (output file name) -p name Progression order. name can be one out of:LRCP, RLCP, RPCL, PCRL, CPRL. Default: LRCP. -q n different psnr for successive layers Note: (options -r and -q cannot be used together) -r n different compression ratio(s) for successive layers. The rate specified for each quality level is the desired compression factor. Note: (options -r and -q cannot be used together) -s X,Y sub-sampling factor (e.g. -s 2,2). Default: No sub-sampling in x or y direction. Remark: sub-sampling bigger than 2 can produce errors. -t W,H (Size of tile (e.g. -t 512,512) ) -x name (Create index file and fill it. Default: no index file) -EPH (Write EPH marker after each header packet. Default:no EPH) -F rawWidth,rawHeight,rawComp,rawBitDepth,s_or_u characteristics of the raw input image -I (Use the irreversible DWT 9-7. Default: Reversible DWT 5-3) -ImgDir directory_name (directory containing input files) -M n mode switch with values: 1, 2, 4, 8, 16, 32. Default:No mode switch activated. Meaning: BYPASS(1) RESET(2) RESTART(4) VSC(8) ERTERM(16) SEGMARK(32) Values can be added: RESTART(4) + RESET(2) + SEGMARK(32) = -M 38 -OutFor ext (extension for output files) -POC TtileNr=resolutionStart, componentStart, layerEnd, resolutionEnd, componentEnd, progressionOrder (see Examples) -ROI c=n,U=n quantization indices upshifted for component c (0 or 1 or 2) with a value of U (>= 0 and <= 37) e.g. -ROI c=0,U=25 -SOP (Write SOP marker before each packet. Default: No SOP marker in the codestream.) -T X,Y (Offset of the origin of the tiles (e.g. -T 100,75) ) -W (see JPWL OPTIONS) JPWL OPTIONS Options usable only if the library has been compiled with -DUSE_JPWL -W h<tilepart><=type>, s<tilepart><=method>, a=<addr>, z=<size>, g=<range>, p<tilepart:pack><=type> h selects the header error protection (EPB): type can be [0=none 1,absent=predefined 16=CRC-16 32=CRC-32 37-128=RS] if tilepart is absent, it is for main and tile headers if tilepart is present, it applies from that tile onwards, up to the next h<> spec, or to the last tilepart in the codestream (max. 16 specs) p selects the packet error protection (EEP/UEP with EPBs) to be applied to raw data: type can be [0=none 1,absent=predefined 16=CRC-16 32=CRC-32 37-128=RS] if tilepart:pack is absent, it is from tile 0, packet 0 if tilepart:pack is present, it applies from that tile and that packet onwards, up to the next packet spec or to the last packet in the last tilepart in the stream (max. 16 specs) s enables sensitivity data insertion (ESD): method can be [-1=NO ESD 0=RELATIVE ERROR 1=MSE 2=MSE REDUCTION 3=PSNR 4=PSNR INCREMENT 5=MAXERR 6=TSE 7=RESERVED] if tilepart is absent, it is for main header only if tilepart is present, it applies from that tile onwards, up to the next s<> spec, or to the last tilepart in the codestream (max. 16 specs) g determines the addressing mode: range can be [0=PACKET 1=BYTE RANGE 2=PACKET RANGE] a determines the size of data addressing: addr can be 2/4 bytes (small/large codestreams). If not set, auto-mode z determines the size of sensitivity values: size can be 1/2 bytes, for the transformed pseudo-floating point value
opj_compress -i file.bmp -o file.j2k -r 20,10,1 (compress 20x, then 10x, then lossless). opj_compress -i file.ppm -o file.j2k -q 30,40,50 opj_compress -i file.pgx -o file.j2k -POC T1=0,0,1,5,3,CPRL opj_compress -i lena.raw -o lena.j2k -F 512,512,3,8,u AUTHORS Copyright (c) 2002-2014, Universite catholique de Louvain (UCL), Belgium Copyright (c) 2002-2014, Professor Benoit Macq Copyright (c) 2001-2003, David Janssens Copyright (c) 2002-2003, Yannick Verschueren Copyright (c) 2003-2007, Francois-Olivier Devaux and Antonin Descampe Copyright (c) 2005, Herve Drolon, FreeImage Team Copyright (c) 2006-2007, Parvatha Elangovan SEE ALSO opj_decompress(1) opj_dump(1) opj_compress Version 2.1.1 opj_compress(1)
conda-debug
null
null
null
null
null
texify_gui
null
null
null
null
null
qmltestrunner
null
null
null
null
null
xmllint
The xmllint program parses one or more XML files, specified on the command line as XML-FILE (or the standard input if the filename provided is - ). It prints various types of output, depending upon the options selected. It is useful for detecting errors both in XML code and in the XML parser itself. xmllint is included in libxml(3).
xmllint - command line XML tool
xmllint [--version | --debug | --shell | --xpath "XPath_expression" | --debugent | --copy | --recover | --noent | --noout | --nonet | --path "PATH(S)" | --load-trace | --htmlout | --nowrap | --valid | --postvalid | --dtdvalid URL | --dtdvalidfpi FPI | --timing | --output FILE | --repeat | --insert | --compress | --html | --xmlout | --push | --memory | --maxmem NBBYTES | --nowarning | --noblanks | --nocdata | --format | --encode ENCODING | --dropdtd | --nsclean | --testIO | --catalogs | --nocatalogs | --auto | --xinclude | --noxincludenode | --loaddtd | --dtdattr | --stream | --walker | --pattern PATTERNVALUE | --chkregister | --relaxng SCHEMA | --schema SCHEMA | --c14n] {XML-FILE(S)... | -} xmllint --help
xmllint accepts the following options (in alphabetical order): --auto Generate a small document for testing purposes. --catalogs Use the SGML catalog(s) from SGML_CATALOG_FILES. Otherwise XML catalogs starting from /etc/xml/catalog are used by default. --chkregister Turn on node registration. Useful for developers testing libxml(3) node tracking code. --compress Turn on gzip(1) compression of output. --copy Test the internal copy implementation. --c14n Use the W3C XML Canonicalisation (C14N) to serialize the result of parsing to stdout. It keeps comments in the result. --dtdvalid URL Use the DTD specified by an URL for validation. --dtdvalidfpi FPI Use the DTD specified by a Formal Public Identifier FPI for validation, note that this will require a catalog exporting that Formal Public Identifier to work. --debug Parse a file and output an annotated tree of the in-memory version of the document. --debugent Debug the entities defined in the document. --dropdtd Remove DTD from output. --dtdattr Fetch external DTD and populate the tree with inherited attributes. --encode ENCODING Output in the given encoding. Note that this works for full document not fragments or result from XPath queries. --format Reformat and reindent the output. The XMLLINT_INDENT environment variable controls the indentation. The default value is two spaces " "). --help Print out a short usage summary for xmllint. --html Use the HTML parser. --htmlout Output results as an HTML file. This causes xmllint to output the necessary HTML tags surrounding the result tree output so the results can be displayed/viewed in a browser. --insert Test for valid insertions. --loaddtd Fetch an external DTD. --load-trace Display all the documents loaded during the processing to stderr. --maxmem NNBYTES Test the parser memory support. NNBYTES is the maximum number of bytes the library is allowed to allocate. This can also be used to make sure batch processing of XML files will not exhaust the virtual memory of the server running them. --memory Parse from memory. --noblanks Drop ignorable blank spaces. --nocatalogs Do not use any catalogs. --nocdata Substitute CDATA section by equivalent text nodes. --noent Substitute entity values for entity references. By default, xmllint leaves entity references in place. --nonet Do not use the Internet to fetch DTDs or entities. --noout Suppress output. By default, xmllint outputs the result tree. --nowarning Do not emit warnings from the parser and/or validator. --nowrap Do not output HTML doc wrapper. --noxincludenode Do XInclude processing but do not generate XInclude start and end nodes. --nsclean Remove redundant namespace declarations. --output FILE Define a file path where xmllint will save the result of parsing. Usually the programs build a tree and save it on stdout, with this option the result XML instance will be saved onto a file. --path "PATH(S)" Use the (space- or colon-separated) list of filesystem paths specified by PATHS to load DTDs or entities. Enclose space-separated lists by quotation marks. --pattern PATTERNVALUE Used to exercise the pattern recognition engine, which can be used with the reader interface to the parser. It allows to select some nodes in the document based on an XPath (subset) expression. Used for debugging. --postvalid Validate after parsing has completed. --push Use the push mode of the parser. --recover Output any parsable portions of an invalid document. --relaxng SCHEMA Use RelaxNG file named SCHEMA for validation. --repeat Repeat 100 times, for timing or profiling. --schema SCHEMA Use a W3C XML Schema file named SCHEMA for validation. --shell Run a navigating shell. Details on available commands in shell mode are below (see the section called “SHELL COMMANDS”). --xpath "XPath_expression" Run an XPath expression given as argument and print the result. In case of a nodeset result, each node in the node set is serialized in full in the output. In case of an empty node set the "XPath set is empty" result will be shown and an error exit code will be returned. --stream Use streaming API - useful when used in combination with --relaxng or --valid options for validation of files that are too large to be held in memory. --testIO Test user input/output support. --timing Output information about the time it takes xmllint to perform the various steps. --valid Determine if the document is a valid instance of the included Document Type Definition (DTD). A DTD to be validated against also can be specified at the command line using the --dtdvalid option. By default, xmllint also checks to determine if the document is well-formed. --version Display the version of libxml(3) used. --walker Test the walker module, which is a reader interface but for a document tree, instead of using the reader API on an unparsed document it works on an existing in-memory tree. Used for debugging. --xinclude Do XInclude processing. --xmlout Used in conjunction with --html. Usually when HTML is parsed the document is saved with the HTML serializer. But with this option the resulting document is saved with the XML serializer. This is primarily used to generate XHTML from HTML input. SHELL COMMANDS xmllint offers an interactive shell mode invoked with the --shell command. Available commands in shell mode include (in alphabetical order): base Display XML base of the node. bye Leave the shell. cat NODE Display the given node or the current one. cd PATH Change the current node to the given path (if unique) or root if no argument is given. dir PATH Dumps information about the node (namespace, attributes, content). du PATH Show the structure of the subtree under the given path or the current node. exit Leave the shell. help Show this help. free Display memory usage. load FILENAME Load a new document with the given filename. ls PATH List contents of the given path or the current directory. pwd Display the path to the current node. quit Leave the shell. save FILENAME Save the current document to the given filename or to the original name. validate Check the document for errors. write FILENAME Write the current node to the given filename. ENVIRONMENT SGML_CATALOG_FILES SGML catalog behavior can be changed by redirecting queries to the user's own set of catalogs. This can be done by setting the SGML_CATALOG_FILES environment variable to a list of catalogs. An empty one should deactivate loading the default /etc/sgml/catalog catalog. XML_CATALOG_FILES XML catalog behavior can be changed by redirecting queries to the user's own set of catalogs. This can be done by setting the XML_CATALOG_FILES environment variable to a space-separated list of catalogs. Use percent-encoding to escape spaces or other characters. An empty variable should deactivate loading the default /etc/xml/catalog catalog. XML_DEBUG_CATALOG Setting the environment variable XML_DEBUG_CATALOG to non-zero using the export command outputs debugging information related to catalog operations. XMLLINT_INDENT Setting the environment variable XMLLINT_INDENT controls the indentation. The default value is two spaces " ". DIAGNOSTICS xmllint return codes provide information that can be used when calling it from scripts. 0 No error 1 Unclassified 2 Error in DTD 3 Validation error 4 Validation error 5 Error in schema compilation 6 Error writing output 7 Error in pattern (generated when --pattern option is used) 8 Error in Reader registration (generated when --chkregister option is used) 9 Out of memory error 10 XPath evaluation error SEE ALSO libxml(3) More information can be found at • libxml(3) web page https://gitlab.gnome.org/GNOME/libxml2 AUTHORS John Fleck <jfleck@inkstain.net> Author. Ziying Sherwin <sherwin@nlm.nih.gov> Author. Heiko Rupp <hwr@pilhuhn.de> Author. COPYRIGHT Copyright © 2001, 2004 libxml2 02/19/2022 XMLLINT(1)
null
mysql_plugin
Note mysql_plugin is deprecated as of MySQL 5.7.11 and removed in MySQL 8.0. Alternatives include loading plugins at server startup using the --plugin-load or --plugin-load-add option, or at runtime using the INSTALL PLUGIN statement. The mysql_plugin utility enables MySQL administrators to manage which plugins a MySQL server loads. It provides an alternative to manually specifying the --plugin-load option at server startup or using the INSTALL PLUGIN and UNINSTALL PLUGIN statements at runtime. Depending on whether mysql_plugin is invoked to enable or disable plugins, it inserts or deletes rows in the mysql.plugin table that serves as a plugin registry. (To perform this operation, mysql_plugin invokes the MySQL server in bootstrap mode. This means that the server must not already be running.) For normal server startups, the server loads and enables plugins listed in mysql.plugin automatically. For additional control over plugin activation, use --plugin_name options named for specific plugins, as described in Section 5.5.1, “Installing and Uninstalling Plugins”. Each invocation of mysql_plugin reads a configuration file to determine how to configure the plugins contained in a single plugin library file. To invoke mysql_plugin, use this syntax: mysql_plugin [options] plugin {ENABLE|DISABLE} plugin is the name of the plugin to configure. ENABLE or DISABLE (not case-sensitive) specify whether to enable or disable components of the plugin library named in the configuration file. The order of the plugin and ENABLE or DISABLE arguments does not matter. For example, to configure components of a plugin library file named myplugins.so on Linux or myplugins.dll on Windows, specify a plugin value of myplugins. Suppose that this plugin library contains three plugins, plugin1, plugin2, and plugin3, all of which should be configured under mysql_plugin control. By convention, configuration files have a suffix of .ini and the same base name as the plugin library, so the default configuration file name for this plugin library is myplugins.ini. The configuration file contents look like this: myplugins plugin1 plugin2 plugin3 The first line in the myplugins.ini file is the name of the library file, without any extension such as .so or .dll. The remaining lines are the names of the components to be enabled or disabled. Each value in the file should be on a separate line. Lines on which the first character is '#' are taken as comments and ignored. To enable the plugins listed in the configuration file, invoke mysql_plugin this way: shell> mysql_plugin myplugins ENABLE To disable the plugins, use DISABLE rather than ENABLE. An error occurs if mysql_plugin cannot find the configuration file or plugin library file, or if mysql_plugin cannot start the MySQL server. mysql_plugin supports the following options, which can be specified on the command line or in the [mysqld] group of any option file. For options specified in a [mysqld] group, mysql_plugin recognizes the --basedir, --datadir, and --plugin-dir options and ignores others. For information about option files used by MySQL programs, see Section 4.2.6, “Using Option Files”. • --help, -? Display a help message and exit. • --basedir=dir_name, -b dir_name The server base directory. • --datadir=dir_name, -d dir_name The server data directory. • --my-print-defaults=file_name, -b file_name The path to the my_print_defaults program. • --mysqld=file_name, -b file_name The path to the mysqld server. • --no-defaults, -p Do not read values from the configuration file. This option enables an administrator to skip reading defaults from the configuration file. With mysql_plugin, this option need not be given first on the command line, unlike most other MySQL programs that support --no-defaults. • --plugin-dir=dir_name, -p dir_name The server plugin directory. • --plugin-ini=file_name, -i file_name The mysql_plugin configuration file. Relative path names are interpreted relative to the current directory. If this option is not given, the default is plugin.ini in the plugin directory, where plugin is the plugin argument on the command line. • --print-defaults, -P Display the default values from the configuration file. This option causes mysql_plugin to print the defaults for --basedir, --datadir, and --plugin-dir if they are found in the configuration file. If no value for a variable is found, nothing is shown. With mysql_plugin, this option need not be given first on the command line, unlike most other MySQL programs that support --print-defaults. • --verbose, -v Verbose mode. Print more information about what the program does. This option can be used multiple times to increase the amount of information. • --version, -V Display version information and exit. COPYRIGHT Copyright © 1997, 2018, Oracle and/or its affiliates. All rights reserved. This documentation is free software; you can redistribute it and/or modify it only under the terms of the GNU General Public License as published by the Free Software Foundation; version 2 of the License. This documentation is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with the program; if not, write to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA or see http://www.gnu.org/licenses/. SEE ALSO For more information, please refer to the MySQL Reference Manual, which may already be installed locally and which is also available online at http://dev.mysql.com/doc/. AUTHOR Oracle Corporation (http://dev.mysql.com/). MySQL 5.7 10/03/2018 MYSQL_PLUGIN(1)
mysql_plugin - configure MySQL server plugins
mysql_plugin [options] plugin {ENABLE|DISABLE}
null
null
mysql_tzinfo_to_sql
The mysql_tzinfo_to_sql program loads the time zone tables in the mysql database. It is used on systems that have a zoneinfo database (the set of files describing time zones). Examples of such systems are Linux, FreeBSD, Solaris, and macOS. One likely location for these files is the /usr/share/zoneinfo directory (/usr/share/lib/zoneinfo on Solaris). If your system does not have a zoneinfo database, you can use the downloadable package described in Section 5.1.15, “MySQL Server Time Zone Support”. mysql_tzinfo_to_sql can be invoked several ways: mysql_tzinfo_to_sql tz_dir mysql_tzinfo_to_sql tz_file tz_name mysql_tzinfo_to_sql --leap tz_file For the first invocation syntax, pass the zoneinfo directory path name to mysql_tzinfo_to_sql and send the output into the mysql program. For example: mysql_tzinfo_to_sql /usr/share/zoneinfo | mysql -u root mysql mysql_tzinfo_to_sql reads your system's time zone files and generates SQL statements from them. mysql processes those statements to load the time zone tables. The second syntax causes mysql_tzinfo_to_sql to load a single time zone file tz_file that corresponds to a time zone name tz_name: mysql_tzinfo_to_sql tz_file tz_name | mysql -u root mysql If your time zone needs to account for leap seconds, invoke mysql_tzinfo_to_sql using the third syntax, which initializes the leap second information. tz_file is the name of your time zone file: mysql_tzinfo_to_sql --leap tz_file | mysql -u root mysql After running mysql_tzinfo_to_sql, it is best to restart the server so that it does not continue to use any previously cached time zone data. COPYRIGHT Copyright © 1997, 2023, Oracle and/or its affiliates. This documentation is free software; you can redistribute it and/or modify it only under the terms of the GNU General Public License as published by the Free Software Foundation; version 2 of the License. This documentation is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with the program; if not, write to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA or see http://www.gnu.org/licenses/. SEE ALSO For more information, please refer to the MySQL Reference Manual, which may already be installed locally and which is also available online at http://dev.mysql.com/doc/. AUTHOR Oracle Corporation (http://dev.mysql.com/). MySQL 8.3 11/23/2023 MYSQL_TZINFO_TO_SQL(1)
mysql_tzinfo_to_sql - load the time zone tables
mysql_tzinfo_to_sql arguments
null
null
streamlit
null
null
null
null
null
acountry
Print the country where HOST (an IPv4 address or hostname) is located, using the countries.nerd.dk DNS domain to identify the country. This utility comes with the c-ares asynchronous resolver library.
acountry - print the country where an IPv4 address or host is located
acountry [OPTION]... HOST...
-d Print some extra debugging output. -h, -? Display this help and exit. -v Be more verbose. Print extra information. REPORTING BUGS Report bugs to the c-ares mailing list: https://lists.haxx.se/listinfo/c-ares SEE ALSO adig(1), ahost(1). The DNSBL countries.nerd.dk http://countries.nerd.dk/ COPYRIGHT This utility is based on code/ideas contained in sofware written by Greg Hudson (ares) carrying the following notice: Copyright 1998 by the Massachusetts Institute of Technology. Permission to use, copy, modify, and distribute this software and its documentation for any purpose and without fee is hereby granted, provided that the above copyright notice appear in all copies and that both that copyright notice and this permission notice appear in supporting documentation, and that the name of M.I.T. not be used in advertising or publicity pertaining to distribution of the software without specific, written prior permission. M.I.T. makes no representations about the suitability of this software for any purpose. It is provided "as is" without express or implied warranty. No further copyright claims are being made by the author(s) of this utility. AUTHOR Gisle Vanem c-ares utilities April 2011 ACOUNTRY(1)
null
jupyter-nbextension
null
null
null
null
null
dotenv
null
null
null
null
null
gtester
null
null
null
null
null
arm64-apple-darwin20.0.0-pagestuff
pagestuff displays information about the specified logical pages of a file conforming to the Mach-O executable format. For each specified page of code, symbols (function and static data structure names) are displayed. The options to pagestuff(1) are: -arch arch_type Specifies the architecture, arch_type, of the file for pagestuff(1) to operate on when the file is a universal file. (See arch(3) for the currently know arch_types.) When this option is used the page numbers are the logical page numbers starting at zero for the offset that architecture starts in a universal file. -a Displays all pages. All other arguments are ignored. -p Print a list of the sections of the specified Mach-O file, with offsets and lengths. All other arguments are ignored. Note that the size(1) tool given arguments "-m -l -x" displays a much more concise listing. SEE ALSO Mach-O(5), size(1) Apple, Inc. February 10, 2017 PAGESTUFF(1)
pagestuff - Mach-O file page analysis tool
pagestuff file [-arch arch_flag] [[-a] [-p] | [pagenumber...]]
null
null
uuclient
null
null
null
null
null
.anaconda-navigator-pre-unlink.sh
null
null
null
null
null