Narayanan Subramaniam
2006-Apr-09  10:59 UTC
[Samba] Samba Storage Server getting extremely slow!!!
Dear all
 I am running a clustering setup with Linux samba Server as output
file server on RHEL 3 workstation . All the clustering output data is 
written to this Linux Samba server which is shared to more than 40
 machines in the network.
All the machines are Windows XP, logging to a windows 2000 domain
server.
The linux samba server is a workgroup member of this windows 2000
server. 
The complete network is Gigabit ethernet . All the cluster nodes acts
as smb clients and mount the output data shares  using samba. 
  The storage server is an IBM Server connected to an Fibre channel
external storage of 3 TB. The storage server is almost used up but
Still some 4 to 5 GB is free in each partition.
 Now the problem is that the storage server is getting very slow
and data writing has become an extremely difficult task.  
  Following the output of log messages of /var/log/samba/smbd.log
 
------------------------------------------------------------------------  
Mar 29 15:22:29 storage winbindd[1305]: [2006/03/29 15:22:29, 0]
param/loadparm.c:map_parameter(2462)
Mar 29 15:22:29 storage winbindd[1305]: Unknown parameter encountered:
"revalidate"
Mar 29 15:22:29 storage winbindd[1305]: [2006/03/29 15:22:29, 0]
param/loadparm.c:lp_do_parameter(3144)
Mar 29 15:22:29 storage winbindd[1305]: Ignoring unknown parameter
"revalidate"
Mar 29 15:22:29 storage winbindd[1305]: [2006/03/29 15:22:29, 0]
nsswitch/winbindd_util.c:winbindd_param_init(555)
Mar 29 15:22:29 storage winbindd[1305]: winbindd: idmap uid range
missing or invalid
Mar 29 15:22:29 storage winbindd[1305]: [2006/03/29 15:22:29, 0]
nsswitch/winbindd_util.c:winbindd_param_init(556)
Mar 29 15:22:29 storage winbindd[1305]: winbindd: cannot continue,
exiting.
Mar 29 15:22:29 storage smb: winbindd startup succeeded
Mar 29 15:25:10 storage smbd[1319]: [2006/03/29 15:25:10, 0]
lib/util_sock.c:write_socket_data(430)
Mar 29 15:25:10 storage smbd[1319]: write_socket_data: write failure.
Error = Connection reset by peer
Mar 29 15:25:10 storage smbd[1319]: [2006/03/29 15:25:10, 0]
lib/util_sock.c:write_socket(455)
Mar 29 15:25:10 storage smbd[1319]: write_socket: Error writing 4 bytes
to socket 25: ERRNO = Connection reset by peer
Mar 29 15:25:10 storage smbd[1319]: [2006/03/29 15:25:10, 0]
lib/util_sock.c:send_smb(647)
Mar 29 15:25:10 storage smbd[1319]: Error writing 4 bytes to client. -1.
(Connection reset by peer)
Mar 29 15:25:13 storage smbd[1323]: [2006/03/29 15:25:13, 0]
lib/util_sock.c:get_peer_addr(1150)
Mar 29 15:25:13 storage smbd[1323]: getpeername failed. Error was
Transport endpoint is not connected
Mar 29 15:25:13 storage smbd[1323]: [2006/03/29 15:25:13, 0]
lib/util_sock.c:write_socket_data(430)
Mar 29 15:25:13 storage smbd[1323]: write_socket_data: write failure.
Error = 
----------------------------------------------------------------------------
Following is the output of df -h
----------------------------------
Filesystem            Size  Used Avail Use% Mounted on
/dev/sda6             2.9G  1.9G  936M  67% /
/dev/sda1              99M   15M   79M  17% /boot
none                  504M     0  504M   0% /dev/shm
/dev/sda3             3.9G  2.4G  1.3G  66% /usr
/dev/sda5             2.9G  189M  2.6G   7% /var
/dev/sdb1             596G  510G   56G  91% /share/share1
/dev/sdb5             459G  425G   11G  98% /share/share2
/dev/sdb6             230G  205G   13G  95% /share/share3
/dev/sdb7             547G  516G  3.4G 100% /share/backup
Following is the output of ps aux - Here we can see that lot of
pids that are created on March 29 and 30 till April 4 are still there. 
--------------------------------------
root      1298  0.0  0.0  9704  800 ?        S    Mar29   0:01 smbd -D
root      1306  0.0  0.0  9700  472 ?        S    Mar29   0:00 smbd -D
10052     1312  0.0  0.1 10260 1360 ?        S    Mar29   0:45 smbd -D
10052     1313  0.0  0.1 10268 1364 ?        S    Mar29   0:41 smbd -D
root      1315  0.0  0.1 10328 1352 ?        S    Mar29   1:17 smbd -D
root      1316  0.0  0.1 10692 1840 ?        S    Mar29   1:05 smbd -D
root      1338  0.0  0.1 10316 1156 ?        S    Mar29   0:02 smbd -D
10052     1418  0.0  0.1 10328 1332 ?        S    Mar29   0:40 smbd -D
10052     1419  0.0  0.1 10392 1384 ?        S    Mar29   0:43 smbd -D
10052     1420  0.0  0.1 10384 1316 ?        S    Mar29   0:40 smbd -D
10052     1421  0.0  0.1 10384 1316 ?        S    Mar29   0:40 smbd -D
10052     1422  0.0  0.1 10376 1292 ?        S    Mar29   0:46 smbd -D
10052     1703  0.0  0.1 10412 1504 ?        S    Mar29   0:41 smbd -D
10052     1746  0.0  0.1 10460 1676 ?        S    Mar29   0:39 smbd -D
10052     1747  0.0  0.1 10460 1648 ?        S    Mar29   0:37 smbd -D
root      2360  0.0  0.7 16580 7512 ?        S    Mar29   3:43 smbd -D
root      4145  0.0  0.1 10516 1448 ?        S    Mar30   1:05 smbd -D
root     13913  0.0  0.1 11096 2024 ?        S    Apr02   0:06 smbd -D
root     14339  0.0  0.2 11140 2152 ?        S    Apr03   0:51 smbd -D
root     15549  0.0  0.3 12452 3468 ?        S    Apr04   1:05 smbd -D
root     15559  0.0  0.1 10660 1520 ?        S    Apr04   0:02 smbd -D
root     16166  0.0  0.3 12332 3204 ?        S    08:53   0:31 smbd -D
root     16214  0.1  0.3 12408 3636 ?        S    09:04   0:58 smbd -D
10052    16290  0.0  0.2 10696 2508 ?        R    09:43   0:07 smbd -D
root     16325  0.0  0.1 11088 2020 ?        S    10:02   0:14 smbd -D
root     16341  0.2  0.3 12364 3624 ?        S    10:07   1:09 smbd -D
root     16384  0.1  0.3 12844 3948 ?        S    10:27   0:59 smbd -D
root     16428  0.0  0.2 11528 2484 ?        S    11:06   0:15 smbd -D
root     16430  0.0  0.1 10432 1256 ?        S    11:19   0:00 smbd -D
root     16491  0.0  0.1 11076 2040 ?        S    13:49   0:17 smbd -D
root     16511  0.0  0.1 10320 1124 ?        S    14:11   0:00 smbd -D
root     16540  0.0  0.1 10564 1500 ?        S    14:29   0:01 smbd -D
root     16544  0.0  0.1 10440 1176 ?        S    14:41   0:01 smbd -D
root     16666  0.0  0.1 10448 1300 ?        S    16:43   0:01 smbd -D
root     16713  0.1  0.7 17100 8124 ?        S    17:25   0:09 smbd -D
10052    16740  1.6  0.1 10448 1708 ?        S    17:42   1:18 smbd -D
root     16761  0.0  0.1 10560 1404 ?        S    18:05   0:01 smbd -D
10052    16778  0.0  0.1 10448 1452 ?        S    18:17   0:01 smbd -D
10052    16788  0.9  0.1 10452 1672 ?        S    18:29   0:18 smbd -D
10052    16791  0.2  0.1 10436 1616 ?        S    18:30   0:04 smbd -D
root     16792  0.2  0.1 10788 1900 ?        S    18:30   0:05 smbd -D
10052    16850  0.1  0.1 10436 1604 ?        S    18:32   0:03 smbd -D
10052    16852  0.0  0.1 10320 1508 ?        S    18:32   0:00 smbd -D
root     16855  0.1  0.1 10604 1832 ?        S    18:32   0:03 smbd -D
10052    16857  0.2  0.1 10436 1604 ?        S    18:33   0:03 smbd -D
10052    16868  1.2  0.1 10452 1692 ?        S    18:35   0:19 smbd -D
10052    16873  0.3  0.2 10816 2372 ?        S    18:36   0:06 smbd -D
root     16883  0.0  0.1 10204 1092 ?        S    18:38   0:00 smbd -D
root     16886  0.0  0.1 10320 1120 ?        S    18:38   0:00 smbd -D
root     16889  0.3  0.1 10784 1924 ?        S    18:38   0:05 smbd -D
root     16890  0.0  0.1 10320 1216 ?        S    18:38   0:00 smbd -D
10052    16941  0.9  0.1 10448 1692 ?        S    18:39   0:13 smbd -D
10052    16942  0.6  0.1 10448 1668 ?        S    18:39   0:10 smbd -D
10052    16943  1.2  0.1 10448 1684 ?        S    18:39   0:17 smbd -D
10052    16946  1.0  0.1 10452 1656 ?        S    18:40   0:15 smbd -D
10052    16947  0.1  0.1 10436 1612 ?        S    18:40   0:02 smbd -D
10052    16949  0.4  0.2 10848 2764 ?        S    18:41   0:05 smbd -D
10052    17013  0.4  0.2 10468 2636 ?        S    19:01   0:00 smbd -D
root     17021  0.0  0.0  3672  636 pts/0    D    19:03   0:00 grep smbd
Following is the output of service smb status
--------------------------------------------------
smbd (pid 17013 16949 16947 16946 16943 16942 16941 16890 16889 16886
16883 16873 16868 16857 16855 16852 
16850 16792 16791 16788 16778 16761 16740 16713 16666 16544 16540 16511
16491 16430 16428 16384 16341 16325 
16290 16214 16166 15559 15549 14339 13913 4145 2360 1747 1746 1703 1422
1421 1420 1419 1418 1338 1316 1315 
1313 1312 1306 1298) is running...
nmbd (pid 1302) is running...
winbindd (pid 1308 1307) is running...
Following is the output of TOP
---------------------------------
 19:02:50  up 7 days,  4:15,  2 users,  load average: 6.61, 8.09, 9.57
117 processes: 116 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:  cpu    user    nice  system    irq  softirq  iowait    idle
           total    2.9%    0.0%   12.7%   9.8%    10.7%   63.7%    0.0%
Mem:  1030408k av, 1021836k used,    8572k free,       0k shrd,   27200k
buff
                    776876k actv,   96396k in_d,   13936k in_c
Swap: 4192956k av,    4504k used, 4188452k free                  876572k
cached
  PID USER     PRI  NI  SIZE  RSS SHARE STAT %CPU %MEM   TIME CPU
COMMAND
16946 clustadm  16   0  1624 1624  1036 S     9.8  0.1   0:14   0 smbd
16943 clustadm  15   0  1664 1664  1044 D     3.9  0.1   0:17   0 smbd
16740 clustadm  15   0  1696 1696  1056 D     0.9  0.1   1:18   0 smbd
17019 root      20   0  1220 1220   904 R     0.9  0.1   0:00   0 top
    1 root      15   0   500  500   448 S     0.0  0.0   0:03   0 init
    2 root      15   0     0    0     0 SW    0.0  0.0   0:07   0
keventd
    3 root      34  19     0    0     0 SWN   0.0  0.0   0:00   0
ksoftirqd/0
    6 root      15   0     0    0     0 SW    0.0  0.0   0:04   0
bdflush
    4 root      15   0     0    0     0 SW    0.0  0.0   5:54   0 kswapd
    5 root      15   0     0    0     0 SW    0.0  0.0   0:21   0 kscand
    7 root      15   0     0    0     0 SW    0.0  0.0   0:00   0
kupdated
    8 root      25   0     0    0     0 SW    0.0  0.0   0:00   0
mdrecoveryd
   14 root      15   0     0    0     0 SW    0.0  0.0   0:00   0
ahd_dv_0
   15 root      15   0     0    0     0 SW    0.0  0.0   0:00   0
ahd_dv_1
   16 root      25   0     0    0     0 SW    0.0  0.0   0:00   0
scsi_eh_0
   17 root      25   0     0    0     0 SW    0.0  0.0   0:00   0
scsi_eh_1
   20 root      21   0     0    0     0 SW    0.0  0.0   0:00   0
qla2300_dpc2
   21 root      20   0     0    0     0 SW    0.0  0.0   0:00   0
qla2300_dpc3
   22 root      20   0     0    0     0 SW    0.0  0.0   0:00   0
scsi_eh_2
   23 root      20   0     0    0     0 SW    0.0  0.0   0:00   0
scsi_eh_3
   26 root      15   0     0    0     0 SW    0.0  0.0   0:04   0
kjournald
   98 root      25   0     0    0     0 SW    0.0  0.0   0:00   0 khubd
  295 root      15   0     0    0     0 SW    0.0  0.0   0:00   0
kjournald
  296 root      15   0     0    0     0 SW    0.0  0.0   0:00   0
kjournald
  297 root      15   0     0    0     0 SW    0.0  0.0   0:00   0
kjournald
  298 root      15   0     0    0     0 DW    0.0  0.0   4:34   0
kjournald
  299 root      15   0     0    0     0 SW    0.0  0.0   0:09   0
kjournald
  300 root      15   0     0    0     0 SW    0.0  0.0   0:12   0
kjournald
  301 root      15   0     0    0     0 SW    0.0  0.0   0:04   0
kjournald
Inorder to fine tune samba ,i changed the following line in smb.conf ,
 but still the problem continues 
/etc/samba/smb.conf
name resolve order = lmhosts wins bcast
Please advise me how to solve this problem
Thanks for any help
Narayanan S
-- 
This message has been scanned for viruses and
dangerous content by Techfocuz Communicator 2.6  (http://www.focuzinfotech.com),
and is
believed to be clean.
I would say that, generally, the speed of writes under most OS and filesystems becomes slower the less space is free on the device. This is a direct result of the OS doing more work to find free space on the device. I think your first step will be ruling out that performance bottleneck before looking at samba for more tuning. If you are still having trouble it would be important to know you filesystems as well as your smb.conf to see how they might be affecting performance as well. Cheers, Eric On 4/9/06, Narayanan Subramaniam <narayanan.s@focuzinfotech.com> wrote:> > Dear all > > I am running a clustering setup with Linux samba Server as output > file server on RHEL 3 workstation . All the clustering output data is > written to this Linux Samba server which is shared to more than 40 > machines in the network. > > All the machines are Windows XP, logging to a windows 2000 domain > server. > The linux samba server is a workgroup member of this windows 2000 > server. > The complete network is Gigabit ethernet . All the cluster nodes acts > as smb clients and mount the output data shares using samba. > > The storage server is an IBM Server connected to an Fibre channel > external storage of 3 TB. The storage server is almost used up but > Still some 4 to 5 GB is free in each partition. > > Now the problem is that the storage server is getting very slow > and data writing has become an extremely difficult task. > > > Following the output of log messages of /var/log/samba/smbd.log > > > ------------------------------------------------------------------------ > Mar 29 15:22:29 storage winbindd[1305]: [2006/03/29 15:22:29, 0] > param/loadparm.c:map_parameter(2462) > > Mar 29 15:22:29 storage winbindd[1305]: Unknown parameter encountered: > "revalidate" > > Mar 29 15:22:29 storage winbindd[1305]: [2006/03/29 15:22:29, 0] > param/loadparm.c:lp_do_parameter(3144) > > Mar 29 15:22:29 storage winbindd[1305]: Ignoring unknown parameter > "revalidate" > > Mar 29 15:22:29 storage winbindd[1305]: [2006/03/29 15:22:29, 0] > nsswitch/winbindd_util.c:winbindd_param_init(555) > > Mar 29 15:22:29 storage winbindd[1305]: winbindd: idmap uid range > missing or invalid > > Mar 29 15:22:29 storage winbindd[1305]: [2006/03/29 15:22:29, 0] > nsswitch/winbindd_util.c:winbindd_param_init(556) > > Mar 29 15:22:29 storage winbindd[1305]: winbindd: cannot continue, > exiting. > > Mar 29 15:22:29 storage smb: winbindd startup succeeded > > Mar 29 15:25:10 storage smbd[1319]: [2006/03/29 15:25:10, 0] > lib/util_sock.c:write_socket_data(430) > > Mar 29 15:25:10 storage smbd[1319]: write_socket_data: write failure. > Error = Connection reset by peer > > Mar 29 15:25:10 storage smbd[1319]: [2006/03/29 15:25:10, 0] > lib/util_sock.c:write_socket(455) > > Mar 29 15:25:10 storage smbd[1319]: write_socket: Error writing 4 bytes > to socket 25: ERRNO = Connection reset by peer > > Mar 29 15:25:10 storage smbd[1319]: [2006/03/29 15:25:10, 0] > lib/util_sock.c:send_smb(647) > > Mar 29 15:25:10 storage smbd[1319]: Error writing 4 bytes to client. -1. > (Connection reset by peer) > > Mar 29 15:25:13 storage smbd[1323]: [2006/03/29 15:25:13, 0] > lib/util_sock.c:get_peer_addr(1150) > > Mar 29 15:25:13 storage smbd[1323]: getpeername failed. Error was > Transport endpoint is not connected > > Mar 29 15:25:13 storage smbd[1323]: [2006/03/29 15:25:13, 0] > lib/util_sock.c:write_socket_data(430) > > Mar 29 15:25:13 storage smbd[1323]: write_socket_data: write failure. > Error > > > ---------------------------------------------------------------------------- > > Following is the output of df -h > ---------------------------------- > Filesystem Size Used Avail Use% Mounted on > /dev/sda6 2.9G 1.9G 936M 67% / > /dev/sda1 99M 15M 79M 17% /boot > none 504M 0 504M 0% /dev/shm > /dev/sda3 3.9G 2.4G 1.3G 66% /usr > /dev/sda5 2.9G 189M 2.6G 7% /var > /dev/sdb1 596G 510G 56G 91% /share/share1 > /dev/sdb5 459G 425G 11G 98% /share/share2 > /dev/sdb6 230G 205G 13G 95% /share/share3 > /dev/sdb7 547G 516G 3.4G 100% /share/backup > > > Following is the output of ps aux - Here we can see that lot of > pids that are created on March 29 and 30 till April 4 are still there. > -------------------------------------- > > > root 1298 0.0 0.0 9704 800 ? S Mar29 0:01 smbd -D > root 1306 0.0 0.0 9700 472 ? S Mar29 0:00 smbd -D > 10052 1312 0.0 0.1 10260 1360 ? S Mar29 0:45 smbd -D > 10052 1313 0.0 0.1 10268 1364 ? S Mar29 0:41 smbd -D > root 1315 0.0 0.1 10328 1352 ? S Mar29 1:17 smbd -D > root 1316 0.0 0.1 10692 1840 ? S Mar29 1:05 smbd -D > root 1338 0.0 0.1 10316 1156 ? S Mar29 0:02 smbd -D > 10052 1418 0.0 0.1 10328 1332 ? S Mar29 0:40 smbd -D > 10052 1419 0.0 0.1 10392 1384 ? S Mar29 0:43 smbd -D > 10052 1420 0.0 0.1 10384 1316 ? S Mar29 0:40 smbd -D > 10052 1421 0.0 0.1 10384 1316 ? S Mar29 0:40 smbd -D > 10052 1422 0.0 0.1 10376 1292 ? S Mar29 0:46 smbd -D > 10052 1703 0.0 0.1 10412 1504 ? S Mar29 0:41 smbd -D > 10052 1746 0.0 0.1 10460 1676 ? S Mar29 0:39 smbd -D > 10052 1747 0.0 0.1 10460 1648 ? S Mar29 0:37 smbd -D > root 2360 0.0 0.7 16580 7512 ? S Mar29 3:43 smbd -D > root 4145 0.0 0.1 10516 1448 ? S Mar30 1:05 smbd -D > root 13913 0.0 0.1 11096 2024 ? S Apr02 0:06 smbd -D > root 14339 0.0 0.2 11140 2152 ? S Apr03 0:51 smbd -D > root 15549 0.0 0.3 12452 3468 ? S Apr04 1:05 smbd -D > root 15559 0.0 0.1 10660 1520 ? S Apr04 0:02 smbd -D > root 16166 0.0 0.3 12332 3204 ? S 08:53 0:31 smbd -D > root 16214 0.1 0.3 12408 3636 ? S 09:04 0:58 smbd -D > 10052 16290 0.0 0.2 10696 2508 ? R 09:43 0:07 smbd -D > root 16325 0.0 0.1 11088 2020 ? S 10:02 0:14 smbd -D > root 16341 0.2 0.3 12364 3624 ? S 10:07 1:09 smbd -D > root 16384 0.1 0.3 12844 3948 ? S 10:27 0:59 smbd -D > root 16428 0.0 0.2 11528 2484 ? S 11:06 0:15 smbd -D > root 16430 0.0 0.1 10432 1256 ? S 11:19 0:00 smbd -D > root 16491 0.0 0.1 11076 2040 ? S 13:49 0:17 smbd -D > root 16511 0.0 0.1 10320 1124 ? S 14:11 0:00 smbd -D > root 16540 0.0 0.1 10564 1500 ? S 14:29 0:01 smbd -D > root 16544 0.0 0.1 10440 1176 ? S 14:41 0:01 smbd -D > root 16666 0.0 0.1 10448 1300 ? S 16:43 0:01 smbd -D > root 16713 0.1 0.7 17100 8124 ? S 17:25 0:09 smbd -D > 10052 16740 1.6 0.1 10448 1708 ? S 17:42 1:18 smbd -D > root 16761 0.0 0.1 10560 1404 ? S 18:05 0:01 smbd -D > 10052 16778 0.0 0.1 10448 1452 ? S 18:17 0:01 smbd -D > 10052 16788 0.9 0.1 10452 1672 ? S 18:29 0:18 smbd -D > 10052 16791 0.2 0.1 10436 1616 ? S 18:30 0:04 smbd -D > root 16792 0.2 0.1 10788 1900 ? S 18:30 0:05 smbd -D > 10052 16850 0.1 0.1 10436 1604 ? S 18:32 0:03 smbd -D > 10052 16852 0.0 0.1 10320 1508 ? S 18:32 0:00 smbd -D > root 16855 0.1 0.1 10604 1832 ? S 18:32 0:03 smbd -D > 10052 16857 0.2 0.1 10436 1604 ? S 18:33 0:03 smbd -D > 10052 16868 1.2 0.1 10452 1692 ? S 18:35 0:19 smbd -D > 10052 16873 0.3 0.2 10816 2372 ? S 18:36 0:06 smbd -D > root 16883 0.0 0.1 10204 1092 ? S 18:38 0:00 smbd -D > root 16886 0.0 0.1 10320 1120 ? S 18:38 0:00 smbd -D > root 16889 0.3 0.1 10784 1924 ? S 18:38 0:05 smbd -D > root 16890 0.0 0.1 10320 1216 ? S 18:38 0:00 smbd -D > 10052 16941 0.9 0.1 10448 1692 ? S 18:39 0:13 smbd -D > 10052 16942 0.6 0.1 10448 1668 ? S 18:39 0:10 smbd -D > 10052 16943 1.2 0.1 10448 1684 ? S 18:39 0:17 smbd -D > 10052 16946 1.0 0.1 10452 1656 ? S 18:40 0:15 smbd -D > 10052 16947 0.1 0.1 10436 1612 ? S 18:40 0:02 smbd -D > 10052 16949 0.4 0.2 10848 2764 ? S 18:41 0:05 smbd -D > 10052 17013 0.4 0.2 10468 2636 ? S 19:01 0:00 smbd -D > root 17021 0.0 0.0 3672 636 pts/0 D 19:03 0:00 grep smbd > > > Following is the output of service smb status > -------------------------------------------------- > > smbd (pid 17013 16949 16947 16946 16943 16942 16941 16890 16889 16886 > 16883 16873 16868 16857 16855 16852 > 16850 16792 16791 16788 16778 16761 16740 16713 16666 16544 16540 16511 > 16491 16430 16428 16384 16341 16325 > 16290 16214 16166 15559 15549 14339 13913 4145 2360 1747 1746 1703 1422 > 1421 1420 1419 1418 1338 1316 1315 > 1313 1312 1306 1298) is running... > nmbd (pid 1302) is running... > winbindd (pid 1308 1307) is running... > > > Following is the output of TOP > --------------------------------- > > 19:02:50 up 7 days, 4:15, 2 users, load average: 6.61, 8.09, 9.57 > 117 processes: 116 sleeping, 1 running, 0 zombie, 0 stopped > CPU states: cpu user nice system irq softirq iowait idle > total 2.9% 0.0% 12.7% 9.8% 10.7% 63.7% 0.0% > Mem: 1030408k av, 1021836k used, 8572k free, 0k shrd, 27200k > buff > 776876k actv, 96396k in_d, 13936k in_c > Swap: 4192956k av, 4504k used, 4188452k free 876572k > cached > > PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME CPU > COMMAND > 16946 clustadm 16 0 1624 1624 1036 S 9.8 0.1 0:14 0 smbd > 16943 clustadm 15 0 1664 1664 1044 D 3.9 0.1 0:17 0 smbd > 16740 clustadm 15 0 1696 1696 1056 D 0.9 0.1 1:18 0 smbd > 17019 root 20 0 1220 1220 904 R 0.9 0.1 0:00 0 top > 1 root 15 0 500 500 448 S 0.0 0.0 0:03 0 init > 2 root 15 0 0 0 0 SW 0.0 0.0 0:07 0 > keventd > 3 root 34 19 0 0 0 SWN 0.0 0.0 0:00 0 > ksoftirqd/0 > 6 root 15 0 0 0 0 SW 0.0 0.0 0:04 0 > bdflush > 4 root 15 0 0 0 0 SW 0.0 0.0 5:54 0 kswapd > 5 root 15 0 0 0 0 SW 0.0 0.0 0:21 0 kscand > 7 root 15 0 0 0 0 SW 0.0 0.0 0:00 0 > kupdated > 8 root 25 0 0 0 0 SW 0.0 0.0 0:00 0 > mdrecoveryd > 14 root 15 0 0 0 0 SW 0.0 0.0 0:00 0 > ahd_dv_0 > 15 root 15 0 0 0 0 SW 0.0 0.0 0:00 0 > ahd_dv_1 > 16 root 25 0 0 0 0 SW 0.0 0.0 0:00 0 > scsi_eh_0 > 17 root 25 0 0 0 0 SW 0.0 0.0 0:00 0 > scsi_eh_1 > 20 root 21 0 0 0 0 SW 0.0 0.0 0:00 0 > qla2300_dpc2 > 21 root 20 0 0 0 0 SW 0.0 0.0 0:00 0 > qla2300_dpc3 > 22 root 20 0 0 0 0 SW 0.0 0.0 0:00 0 > scsi_eh_2 > 23 root 20 0 0 0 0 SW 0.0 0.0 0:00 0 > scsi_eh_3 > 26 root 15 0 0 0 0 SW 0.0 0.0 0:04 0 > kjournald > 98 root 25 0 0 0 0 SW 0.0 0.0 0:00 0 khubd > 295 root 15 0 0 0 0 SW 0.0 0.0 0:00 0 > kjournald > 296 root 15 0 0 0 0 SW 0.0 0.0 0:00 0 > kjournald > 297 root 15 0 0 0 0 SW 0.0 0.0 0:00 0 > kjournald > 298 root 15 0 0 0 0 DW 0.0 0.0 4:34 0 > kjournald > 299 root 15 0 0 0 0 SW 0.0 0.0 0:09 0 > kjournald > 300 root 15 0 0 0 0 SW 0.0 0.0 0:12 0 > kjournald > 301 root 15 0 0 0 0 SW 0.0 0.0 0:04 0 > kjournald > > > > Inorder to fine tune samba ,i changed the following line in smb.conf , > but still the problem continues > > /etc/samba/smb.conf > > name resolve order = lmhosts wins bcast > > > > Please advise me how to solve this problem > > > Thanks for any help > > > Narayanan S > > > -- > This message has been scanned for viruses and > dangerous content by Techfocuz Communicator 2.6 ( > http://www.focuzinfotech.com), and is > believed to be clean. > > -- > To unsubscribe from this list go to the following URL and read the > instructions: https://lists.samba.org/mailman/listinfo/samba >
Narayanan Subramaniam
2006-Apr-10  04:49 UTC
[Samba] Samba Storage Server getting extremely slow!!!
Dear Eric Warnke
   Thanks for your valuable suggestions . I am freeing up the space
in storage partitions and i will be then monitoring the performance.
I will get back to you with the status. 
Following is the smb.conf  i am having . 
[global]
   workgroup = CLUST1
   netbios name = cluststorage
   server string = Cluster Storage Server
   printcap name = /etc/printcap
   load printers = yes
   log file = /var/log/samba/log.smbd
   max log size = 50
   security = user
  encrypt passwords = yes
 smb passwd file = /etc/samba/smbpasswd
  unix password sync = no
  passwd program = /usr/bin/passwd %u
  passwd chat = *New*UNIX*password* %n\n *ReType*new*UNIX*password* %n\n
*passwd:*all*authentication*tokens*updated*successfully*
   socket options = TCP_NODELAY SO_RCVBUF=8192 SO_SNDBUF=8192
   local master = no
   domain master = no 
   preferred master = no
   
dead time = 100 
   wins server = 192.168.0.254
   dns proxy = no 
[project1]
	writeable = yes
	locking = no
	only user = yes
	path = /storage/project1
	create mask = 2666
	directory mask = 2777
	comment = Project1 
	valid users = sam,john,root,@users,clustnode
[backup]
	create mask = 2666
	comment = Project Backup
	directory mask = 2777
	browseable = yes
	writeable = yes
	valid users = clusteradmin,administrator
	path = /storage/probackup
[project2]
   comment = Project 2  
   path = /storage/project2
   public = no
   valid users = @users
   browseable = yes
   writable = yes
 create mask = 2666
    directory mask = 2777
[Libraries]
   comment = Libraries
   path = /storage/libraries
   public = no
   valid users = @users
   browseable = yes
   writable = yes
 create mask = 2666
    directory mask = 2777
 [clusterlogs]
   comment = cluster logs
   path    = /storage/clusterlogs
   public  = no
   valid users = @users,clusternode
   browseable  = yes
   writeable   = yes
 create mask = 2666
    directory mask = 2777
Here the dead time is set as 100. 
Regds
Narayanan S
On Sun, 2006-04-09 at 10:42 -0400, Eric Warnke wrote:> I would say that, generally, the speed of writes under most OS and
> filesystems becomes slower the less space is free on the device.  This
> is a direct result of the OS doing more work to find free space on the
> device.  I think your first step will be ruling out that performance
> bottleneck before looking at samba for more tuning. 
> 
> If you are still having trouble it would be important to know you
> filesystems as well as your smb.conf to see how they might be
> affecting performance as well.
> 
> Cheers,
> Eric
> 
-- 
This message has been scanned for viruses and
dangerous content by Techfocuz Communicator 2.6  (http://www.focuzinfotech.com),
and is
believed to be clean.