-
-
Save jasny/1608062 to your computer and use it in GitHub Desktop.
#!/bin/bash | |
#### | |
# Split MySQL dump SQL file into one file per table | |
# based on http://blog.tty.nl/2011/12/28/splitting-a-database-dump | |
#### | |
if [ $# -lt 1 ] ; then | |
echo "USAGE $0 DUMP_FILE [TABLE]" | |
exit | |
fi | |
if [ $# -ge 2 ] ; then | |
csplit -s -ftable $1 "/-- Table structure for table/" "%-- Table structure for table \`$2\`%" "/-- Table structure for table/" "%40103 SET TIME_ZONE=@OLD_TIME_ZONE%1" | |
else | |
csplit -s -ftable $1 "/-- Table structure for table/" {*} | |
fi | |
[ $? -eq 0 ] || exit | |
mv table00 head | |
FILE=`ls -1 table* | tail -n 1` | |
if [ $# -ge 2 ] ; then | |
mv $FILE foot | |
else | |
csplit -b '%d' -s -f$FILE $FILE "/40103 SET TIME_ZONE=@OLD_TIME_ZONE/" {*} | |
mv ${FILE}1 foot | |
fi | |
for FILE in `ls -1 table*`; do | |
NAME=`head -n1 $FILE | cut -d$'\x60' -f2` | |
cat head $FILE foot > "$NAME.sql" | |
done | |
rm head foot table* |
The foot
part doesn't work if you have over 100 tables. The ls
has to be sorted numerically with -v
so you get the correct last file (otherwise it sees table99
or table999
as the last file):
FILE=`ls -1v table* | tail -n 1`
This script doesn't work if there is more than 100 tables. If there is more than 100 tables, tail -n 1
doesn't get the last files (as ls
doesn't sort files by number : table101 is before table29). To fix it, add -n4
in csplit
to get the number formatted to 4 number and change line #21 to mv table0000 head
.
Hello There, I took a different approach and have written a node module (installable as a cli command) that splits a 16G dump file containing more than 150 tables in less than 2 minutes (on my machine).
If this is of your interest please take a look at mysqldumpsplit
Take note that the script only works if the dump file does NOT contain more than one database.
All generated .sql-files will USE
the first database in the dump file. Furthermore, if several databases contain tables with the same name, the files will overwrite each other.
Hi there,
the SQLDumpSplitter3 should do the trick here, too. :)
https://philiplb.de/sqldumpsplitter3/
SQLDumpSplitter3 caused issues for me with the dump and it also cannot split by table. The ls -1v table* | tail -n 1
fix from @tmirks was an awesome patch. 👍
Hello There, I took a different approach and have written a node module (installable as a cli command) that splits a 16G dump file containing more than 150 tables in less than 2 minutes (on my machine).
If this is of your interest please take a look at mysqldumpsplit
I'll vouch for this. I just used it on a 2 GiB, 470 table dumpfile. 10.3 seconds, no errors, works fine.
Thank you vekexasia!
Any tool that could export all tables from a specific database inside of a multi database dump would be absolutely amazing. If anyone knows of such a tool please share ;)
This script doesn't work if there is more than 100 tables. If there is more than 100 tables,
tail -n 1
doesn't get the last files (asls
doesn't sort files by number : table101 is before table29). To fix it, add-n4
incsplit
to get the number formatted to 4 number and change line #21 tomv table0000 head
.
which csplit command? there are several :(
there is a simple fix to those saying issue with more than a 100 files: -
- add
-n5
in bothcsplit
commands to allow upto 5 digits zero-spaced in the name. - change
mv table00 head
tomv table00000 head
.
Now this will handle upto 100000
tables starting from 00000
to 99999
.
Upgrade as required!
Bonus: add rm $FILE
after cat head $FILE foot
to remove the old file as well so as to save some space if there are too many files. My scenario had over 2500 tables with the base sql file at 130GB!
Bonus 2 hours later: for the really lazy ones, heres a link to my version of this script including output and timing
#https://gist.github.com/m-faraz/4e32cbae283e3a9454118cd1d281f8cf
Worked perfect for me on a 2.5gb file. Took less than 10 seconds.
csplit: *}: bad repetition count
csplit -s -ftable $1 "$START" {**}
->
csplit -s -ftable $1 "$START" {9999999}