mydumper Usage¶
Synopsis¶
mydumper [OPTIONS]
Description¶
mydumper is a tool used for backing up MySQL database servers much faster than the mysqldump tool distributed with MySQL. It also has the capability to retrieve the binary logs from the remote server at the same time as the dump itself. The advantages of mydumper are:
Parallelism (hence, speed) and performance (avoids expensive character set conversion routines, efficient code overall)
Easier to manage output (separate files for tables, dump metadata, etc, easy to view/parse data)
Consistency - maintains snapshot across all threads, provides accurate master and slave log positions, etc
Manageability - supports PCRE for specifying database and tables inclusions and exclusions
Options¶
The mydumper tool has several available options:
Connection Options¶
- -h, --host¶
The host to connect to
- -u, --user¶
Username with the necessary privileges
- -p, --password¶
User password
- -a, --ask-password¶
Prompt For User password
- -P, --port¶
TCP/IP port to connect to
- -S, --socket¶
UNIX domain socket file to use for connection
- --protocol¶
The protocol to use for connection (tcp, socket)
- -C, --compress-protocol¶
Use compression on the MySQL connection
- --ssl¶
Connect using SSL
- --ssl-mode¶
Desired security state of the connection to the server: REQUIRED, VERIFY_IDENTITY
- --key¶
The path name to the key file
- --cert¶
The path name to the certificate file
- --ca¶
The path name to the certificate authority file
- --capath¶
The path name to a directory that contains trusted SSL CA certificates in PEM format
- --cipher¶
A list of permissible ciphers to use for SSL encryption
- --tls-version¶
Which protocols the server permits for encrypted connections
Filter Options¶
- -x, --regex¶
Regular expression for ‘db.table’ matching
- -B, --database¶
Comma delimited list of databases to dump
- -i, --ignore-engines¶
Comma delimited list of storage engines to ignore
- --where¶
Dump only selected records.
- -U, --updated-since¶
Use Update_time to dump only tables updated in the last U days
- --partition-regex¶
Regex to filter by partition name.
- -O, --omit-from-file¶
File containing a list of database.table entries to skip, one per line (skips before applying regex option)
- -T, --tables-list¶
Comma delimited table list to dump (does not exclude regex option). Table name must include database name. For instance: test.t1,test.t2
Lock Options¶
- -z, --tidb-snapshot¶
Snapshot to use for TiDB
- -k, --no-locks¶
Do not execute the temporary shared read lock.
WARNING: This will cause inconsistent backups
- --use-savepoints¶
Use savepoints to reduce metadata locking issues, needs SUPER privilege
- --no-backup-locks¶
Do not use Percona backup locks
- --lock-all-tables¶
Use LOCK TABLE for all, instead of FTWRL
- --less-locking¶
Minimize locking time on InnoDB tables.
- --trx-consistency-only¶
Transactional consistency only
- --skip-ddl-locks¶
Do not send DDL locks when possible
PMM Options¶
- --pmm-path¶
which default value will be /usr/local/percona/pmm2/collectors/textfile-collector/high-resolution
- --pmm-resolution¶
which default will be high
Exec Options¶
- --exec-threads¶
Amount of threads to use with –exec
- --exec¶
Command to execute using the file as parameter
- --exec-per-thread¶
Set the command that will receive by STDIN and write in the STDOUT into the output file
- --exec-per-thread-extension¶
Set the extension for the STDOUT file when –exec-per-thread is used
- --long-query-retries¶
Retry checking for long queries, default 0 (do not retry)
- --long-query-retry-interval¶
Time to wait before retrying the long query check in seconds, default 60
- -l, --long-query-guard¶
Set long query timer in seconds, default 60
- -K, --kill-long-queries¶
Kill long running queries (instead of aborting)
Job Options¶
- --max-threads-per-table¶
Maximum number of threads per table to use
- --char-deep¶
Defines the amount of characters to use when the primary key is a string
- --char-chunk¶
Defines in how many pieces should split the table. By default we use the amount of threads
- -r, --rows¶
Spliting tables into chunks of this many rows. It can be MIN:START_AT:MAX. MAX can be 0 which means that there is no limit. It will double the chunk size if query takes less than 1 second and half of the size if it is more than 2 seconds
- --split-partitions¶
Dump partitions into separate files. This options overrides the –rows option for partitioned tables.
Checksum Options¶
- -M, --checksum-all¶
Dump checksums for all elements
- --data-checksums¶
Dump table checksums with the data
- --schema-checksums¶
Dump schema table and view creation checksums
- --routine-checksums¶
Dump triggers, functions and routines checksums
Objects Options¶
- -m, --no-schemas¶
Do not dump table schemas with the data and triggers
- -Y, --all-tablespaces¶
Dump all the tablespaces.
- -d, --no-data¶
Do not dump table data
- -G, --triggers¶
Dump triggers. By default, it do not dump triggers
- -E, --events¶
Dump events. By default, it do not dump events
- -R, --routines¶
Dump stored procedures and functions. By default, it do not dump stored procedures nor functions
- --skip-constraints¶
Remove the constraints from the CREATE TABLE statement. By default, the statement is not modified
- --skip-indexes¶
Remove the indexes from the CREATE TABLE statement. By default, the statement is not modified
- --views-as-tables¶
Export VIEWs as they were tables
- -W, --no-views¶
Do not dump VIEWs
Statement Options¶
- --load-data¶
Instead of creating INSERT INTO statements, it creates LOAD DATA statements and .dat files. This option will be deprecated on future releases use –format
- --csv¶
Automatically enables –load-data and set variables to export in CSV format. This option will be deprecated on future releases use –format
- --format¶
Set the output format which can be INSERT, LOAD_DATA, CSV or CLICKHOUSE. Default: INSERT
- --include-header¶
When –load-data or –csv is used, it will include the header with the column name
- --fields-terminated-by¶
Defines the character that is written between fields
- --fields-enclosed-by¶
Defines the character to enclose fields. Default: “
- --fields-escaped-by¶
Single character that is going to be used to escape characters in theLOAD DATA stament, default: ‘’
- --lines-starting-by¶
Adds the string at the begining of each row. When –load-data is usedit is added to the LOAD DATA statement. Its affects INSERT INTO statementsalso when it is used.
- --lines-terminated-by¶
Adds the string at the end of each row. When –load-data is used it isadded to the LOAD DATA statement. Its affects INSERT INTO statementsalso when it is used.
- --statement-terminated-by¶
This might never be used, unless you know what are you doing
- -N, --insert-ignore¶
Dump rows with INSERT IGNORE
- --replace¶
Dump rows with REPLACE
- --complete-insert¶
Use complete INSERT statements that include column names
- --hex-blob¶
Dump binary columns using hexadecimal notation
- --skip-definer¶
Removes DEFINER from the CREATE statement. By default, statements are not modified
- -s, --statement-size¶
Attempted size of INSERT statement in bytes, default 1000000
- --tz-utc¶
SET TIME_ZONE=’+00:00’ at top of dump to allow dumping of TIMESTAMP data when a server has data in different time zones or data is being moved between servers with different time zones, defaults to on use –skip-tz-utc to disable.
- --skip-tz-utc¶
Doesn’t add SET TIMEZONE on the backup files
- --set-names¶
Sets the names, use it at your own risk, default binary
Extra Options¶
- -F, --chunk-filesize¶
Split data files into pieces of this size in MB. Useful for myloader multi-threading.
- --exit-if-broken-table-found¶
Exits if a broken table has been found
- --success-on-1146¶
Not increment error count and Warning instead of Critical in case of table doesn’t exist
- -e, --build-empty-files¶
Build dump files even if no data available from table
- --no-check-generated-fields¶
Queries related to generated fields are not going to be executed.It will lead to restoration issues if you have generated columns
- --order-by-primary¶
Sort the data by Primary Key or Unique key if no primary key exists
- --compact¶
Give less verbose output. Disables header/footer constructs.
- -c, --compress¶
Compress output files using: /usr/bin/gzip and /usr/bin/zstd. Options: GZIP and ZSTD. Default: GZIP
- --use-defer¶
Use defer integer sharding until all non-integer PK tables processed (saves RSS for huge quantities of tables)
- --check-row-count¶
Run SELECT COUNT(*) and fail mydumper if dumped row count is different
- --source-data¶
It will include the options in the metadata file, to allow myloader to establish replication
Daemon Options¶
- -D, --daemon¶
Enable daemon mode
- -I, --snapshot-interval¶
Interval between each dump snapshot (in minutes), requires –daemon, default 60
- -X, --snapshot-count¶
number of snapshots, default 2
Application Options:¶
- -?, --help¶
Show help options
- -o, --outputdir¶
Directory to output files to
- --clear¶
Clear output directory before dumping
- --dirty¶
Overwrite output directory without clearing (beware of leftower chunks)
- --stream¶
It will stream over STDOUT once the files has been written. Since v0.12.7-1, accepts NO_DELETE, NO_STREAM_AND_NO_DELETE and TRADITIONAL which is the default value and used if no parameter is given
- -L, --logfile¶
Log file name to use, by default stdout is used
- --disk-limits¶
Set the limit to pause and resume if determines there is no enough disk space.Accepts values like: ‘<resume>:<pause>’ in MB.For instance: 100:500 will pause when there is only 100MB free and willresume if 500MB are available
- -t, --threads¶
Number of threads to use, 0 means to use number of CPUs. Default: 4
- -V, --version¶
Show the program version and exit
- -v, --verbose¶
Verbosity of output, 0 = silent, 1 = errors, 2 = warnings, 3 = info, default 2
- --debug¶
Turn on debugging output (automatically sets verbosity to 3)
- --defaults-file¶
Use a specific defaults file. Default: /etc/mydumper.cnf
- --defaults-extra-file¶
Use an additional defaults file. This is loaded after –defaults-file, replacing previous defined values
- --source-control-command¶
Instruct the proper commands to execute depending where are configuring the replication. Options: TRADITIONAL, AWS