I have the following update procedure that update mySQL DB over the internet between source Linux Centos (local machine on my net behind a DMZ with real IP A.B.C.D) and target Linux fedora (web server www.myweb.com) every day on a specific time 18:00 through a crontab on my source linux server server(source) ---DMZ---ASA---Router-----Internet----HostingCompany---Myweb(target) [root at source]# mysql -u updatex -p -h www.myweb.com test < sample.SQL [root at source]$ mysql -u updatex -p -h www.myweb.com test < sample.SQL Enter password: ***** CURTIME() 19:41:44 CURTIME() 19:50:09 [root at source]$ mysql -u updatex -p -h www.myweb.com test < sample.SQL Enter password:***** CURTIME() 08:26:08 CURTIME() 08:26:34 I did the above procedure multiple times in different times in the day. the duration of this procedure takes from 22sec to 10min see above...., before a while it was running constant with duration of 30sec. I checked with my ISP, hosting company and network nothing been changed from the structure/configuration..... [root at source]# lsof -i -P | grep 3306 mysqld 3806 mysql 11u IPv4 10926 TCP *:3306 (LISTEN) mysql 15150 user 3u IPv4 297528 TCP 192.168.10.5:8376->www.myweb.com:3306 (ESTABLISHED) [root at target]# netstat -a |grep mysql tcp 0 0 *:mysql *:* LISTEN tcp 0 0 www.myweb.:mysql A.B.C.D:8366 TIME_WAIT tcp 0 11 www.myweb.:mysql A.B.C.D:8372 ESTABLISHED also i attached tcp connection between the nodes as above from source and target, can any one help why i have this behavior and how can i fix the delay, thinking doing QoS or clean up and remoteexcution at that time ... Thanks