Today, I have encountered problems during fetching large MySQL table. Simply, the server was getting out of memory due to large table size.
I knew, you can fetch subset of data from MySQL table using LIMIT. Standard `LIMIT N` returns first N rows of your table.
However there is second way of calling LIMIT, `LIMIT S, N`; where S is the starting row and N is total number of rows to be returned. Thus, you can easily plug it into FOR LOOP to fetch from large MySQL table in for example 1 million rows batches:
#get row count
c=`mysql -u$user -p$pass $db -N -A -e"select count(*) from $t"`
echo `date` $t "with $c rows";
#write table header
echo -e $header | gzip > $t.txt.gz
#get table in batch
for i in `seq 1 $step $c`; do
cmd="select * from $t limit $i, $step"
echo -en "r$i / $c"
mysql -u$user -p$pass $db -A -N -e"$cmd" | gzip >> $t.txt.gz
Hope, someone will find it useful:)