MySQL Parallel Dump

Multi threaded mysqldump is not an utopia any more. mysqlpdump can dump all your tables and databases in parallel so it can be much faster in systems with multiple cpu’s.

It stores each table in a different file by default. It can also generate the dump to stdout although this is not recommended because it can use all the memory in your system if your tables are big.


I saw an interesting post on MySQL Performance Blog with some suggestions to improve mysqldump.

Here is my effort to implement some of that suggestions.




Simplest usage (will save a file for each table): -u root -p password

Save compressed files (gzip) to /tmp/dumps and pass “–skip-opt” to mysqldump -u root -p password -d /tmp/dumps/ -g -P "--skip-opt"

Output to stdout and use 20 threads: -u root -p password -stdout -t 20

Be more “verbose”: -u root -p password -v

Exclude “mysql” and “test” table from dumping: -u root -p password -e mysql -e test

Only dump “mysql” table: -u root -p password -i mysql



  • 0.5
    • Compress 00_master_data.sql file if specified
    • bugfix: when it’s called without a terminal or a logged user, it uses “nobody”.
    • bugfix: destination now works with 00_master_data.sql
  • 0.4
    • Made it compatible with python 2.4
    • Can include and exclude specified databases.
  • 0.3
    • Fixed a bug that prevented the tables of being dumped because of a lock
    • Added –master-data option to write “CHANGE MASTER TO ” statement
  • 0.2
    • Store dumps to files directly instead to stdout
    • Can compress files
    • Dump each table in its own file
    • Can pass parameters directly to mysqldump
  • 0.1
    • First version


mysqlpdump uses GNU/GPL License.

16 thoughts on “mysqlpdump

  1. Devis: That’s true. Python 2.5 is necessary because I’m using some functions in the Queue module that are only available in 2.5.

  2. Hi fr3nd, it’s pity because Debian will release 2.5 in 2070… So for the moment I cannot use this script, hope for the future… Do you think it is possible to remove the 2.5 instructions, like queue.join ?
    Anyway, thank you!

  3. Devis, I’ll try it… I just need to find some safe way to substitute .join()

    Let me check it.

  4. Devis: I made it compatible with python 2.4. It was easier than I thought actually. I’ve also added two more options to specify what databases to dump and what databases not to dump. See usage for examples.

  5. Wonderful, it works! 10M Kudos to you!
    Surely I will follow the script evolution and if I can help just write me :-)
    Thank you!

  6. Hi fr3nd I’ve put mysqlpdump in a cron job but I am having little problems:

    Traceback (most recent call last):
    File “”, line 241, in ?
    File “”, line 187, in main
    parser.add_option(“-u”, “–user”, action=”store”, dest=”user”, type=”string”, default=os.getlogin(), help=”User for login.”)
    OSError: [Errno 25] Inappropriate ioctl for device

    It is caused by “os.getlogin()” in combination with cron, as there isn’t any terminal or logged user.
    For the moment I’ve substitued this with ‘root’ and it works for me.

  7. This sounds very interesting and promising; but I think it will not be possible to create a consistent dump that way? Or how should locking be handled?

  8. Thank you !

    Too bad MySQL does not have support for several connections sharing same transactional context so the only way you can get consistent backup is setting low priority updates and locking all the tables. As you have multiple threads you may have no luck locking all of them otherwise.

  9. This tool automatically locks all tables before starting the dump, so there is no need to lock them manually. It’ll create a consistent dump.

    About the restore tool, I don’t think it’s needed… Restoring is as easy as doing:

    cat *.sql | mysql

  10. surts referiat a la segona edicio del “High Performance MySQL” pero veient qui et comenta més amunt, no m’extranya

  11. Que bò, havia llegit alguns capitols del llibre, pero no tot… suposo que hauré de fer-li una repassada més intensa :)
    Merci per avisar

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.