similar to: Throughput Constraints

Displaying 20 results from an estimated 10000 matches similar to: "Throughput Constraints"

2010 May 29
1
IFB0 throughput 3-4% lower than expected
I have two boxes for the purpose of testing traffic control and my knowledge thereof (which is at the inkling stage). The boxes are connected by 100Mbit ethernet cards via a switch. For egress traffic via eth0 I achieve a throughput that is close to the specified CEILing, particularly for values above 1mbit. Ingress traffic does not seem so well behaved. Above about 1mbit rates achieved are
2010 Mar 17
3
tinc throughput on alix machines
Hi, I've setup a wlan ap using a an alix 2d2 (AMD Geode LX 500 Mhz + AES Offload) and am now trying to encapsulate the wlan traffic into a vpn (tap bridge). Though even when disabling crypto and digest (there is a crypto engine onboard can be setup to handle tinc encryption), the best throughput I get is aroung 2,9 MByte per second (wget+http) (even if the client is run locally). The network
2008 Mar 25
2
Expected transfer rate of samba, SATA over gigabit to SATA
What is a good transfer rate to expect when writing from Windows XP/Vista to a samba network share in the following situation? * New client. Intel Core 2 Duo 6600 with 4GB memory, and Intel gigabit NIC. * New server. HP XW8400 with (currently) a single 5335 Xeon with 4GB memory, and integrated Intel gigabit. Running samba 3.0.25b, as included with CentOS (RHEL). * Netgear GS116 (gigabit)
2005 May 13
4
Gigabit Throughput too low
Hi I was wondering if you ever got better performance out of your Gigabit/IDE/Fc2? I am facing a similar situation. I am running FC2 with Samba 3.x My problem lies in not that I am limited to 10 MBytes per second sustained. I think it's related to this pdflush and how it's buffers are setup. (I have been doing some research and before 2.6 kernels bdflush was the method that was used and
2011 Jan 11
1
Bonding performance question
I have a Dell server with four bonded, gigabit interfaces. Bonding mode is 802.3ad, xmit_hash_policy=layer3+4. When testing this setup with iperf, I never get more than a total of about 3Gbps throughput. Is there anything to tweak to get better throughput? Or am I running into other limits (e.g. was reading about tcp retransmit limits for mode 0). The iperf test was run with iperf -s on the
2013 Jul 30
3
SMB throughput inquiry, Jeremy, and James' bow tie
I went to the site to subscribe again and ended up watching some of Jeremy's Google interviews. I particularly enjoyed the interview with James and the bow tie lesson at the end. :) So anyway, I recently upgraded my home network to end-to-end GbE. My clients are Windows XP SP3 w/hot fixes, and my Samba server is 3.5.6 atop vanilla kernel.org Linux 3.2.6 and Debian 6.0.6. With FDX fast
2008 Jan 25
2
rsyncd performance when handling multiple clients in parallel
Hi: I use rsync to transfer multiple files from several clients to a server in parallel. I am wondering how many concurrent connections the server should handle to maximize the throughput (number of bytes written to server). In an extreme case, if only one connection allowed, the disk IO speed of server will not be fully utilized. On the other hand, if the server allows too many connections, the
2014 Sep 06
2
Decent Performance
Hi all, I am trying to get some decent speed out of a vanilla Debian (Proxmox) box and am running out of ideas. Basically I can exchange data at 110MB/sec (raw, nfs, ftp) but not with Samba shares. I tried almost every samba tuning tip I could find to no avail. The only optimisation left to smb.conf at the moment is debug level = 0. As one can see the robocopy throughput is merely half of what
2013 Sep 06
1
Gluster native client very slow
Hello, I'm testing a two nodes glusterfs distributed cluster (version 3.3.1-1) on Debian 7. The two nodes write on the same iscsi volume on a SAN. When I try to write an 1G file with dd, I have the following results : NFS : 107 Mbytes/s Gluster client : 8 Mbytes/sec My /etc/fstab on the client : /etc/glusterfs/cms.vol /data/cms glusterfs defaults 0 0 I'd like to use the gluster
2010 Jan 17
1
How to improve the throughput
Dear I am very new to tinc and would like to ask any advice regarding tinc's throughput. I have been using OpenVPN for several years and found tinc as a simpler/scaler VPN using TUN/TAP device. I have tried to use tinc on WindowsXP and measured throughput in contrast to OpenVPN. My test revealed that tinc is faster in PING response (almost half of OpenVPN). However, when I tested with
2011 Jul 25
3
gluster client performance
Hi- I'm new to Gluster, but am trying to get it set up on a new compute cluster we're building. We picked Gluster for one of our cluster file systems (we're also using Lustre for fast scratch space), but the Gluster performance has been so bad that I think maybe we have a configuration problem -- perhaps we're missing a tuning parameter that would help, but I can't find
2006 Dec 08
22
ZFS Usage in Warehousing (lengthy intro)
Dear all, we''re currently looking forward to restructure our hardware environment for our datawarehousing product/suite/solution/whatever. We''re currently running the database side on various SF V440''s attached via dual FC to our SAN backend (EMC DMX3) with UFS. The storage system is (obviously in a SAN) shared between many systems. Performance is mediocre in terms
2011 Mar 25
1
Samba Tuning to increase Throughput
Hi All, I have gone through threads related to throughput issue in this list. Found few similar issue, but could not get the solution. So looking for some advice from group. I am trying to use the samba to access a USB disk connected to our evaluation board which has xtensa core running at 400 MHz. Samba 3.5.x is running on the board. We are getting below throughput as tested with the
2012 Sep 10
1
Issue with domU RX throughput
Hi, I am encountering a weird issue, where domU RX throughput (of traffic from another machine) is very low. But domU RX throughput from dom0 is fine, and dom0 RX from another machine is fine. iperf reported the following throughputs: dom0 <- outside -- 942Mbps domU <- dom0 -- 3467Mbps domU <- outside -- 87Mbps The NIC on the machine is 1GE. What puzzles me is, some instances of domU
2017 Aug 13
0
throughput question
Hi everybody I have a some question about throughput for glusterfs volume. I have 3 server for glusterfs, each have one brick and 1GbE for their network, I have made distributed replica 3 volume with these 3 bricks. the network between the clients and the servers are 1GbE. refer to this link: https://s3.amazonaws.com/aws001/guided_trek/Perform ance_in_a_Gluster_Systemv6F.pdf I have setup
2008 Jun 24
5
Reg: Throughput b/w domU & dom0
Hi all, I used netperf to measure throughput between dom0 & domU. The throughput between dom0 -> domU was 256.00 Mb/sec domU -> dom0 was 401.15 Mb/sec. The throughput between dom0 & domU seems to be very asymmetry. To my surprise the throughput between domU -> dom0 is more. The value which I specified are consistent values. Is
2010 Mar 02
9
Filebench Performance is weird
Greeting All I am using Filebench benchmark in an "Interactive mode" to test ZFS performance with randomread wordload. My Filebench setting & run results are as follwos ------------------------------------------------------------------------------------------ filebench> set $filesize=5g filebench> set $dir=/hdd/fs32k filebench> set $iosize=32k filebench> set
2009 Sep 04
3
bridge throughput problem
I have set-up xen on my Intel quad core server. Now running different experiment to measure network throughput in virtualized environment. these are some of the results: Netperf-4.5 results for inter-domain communication. Sr.No.              Client               Server               Time(sec)        Throughput(Mbps)    1                   Guest-1              Dom0                   
2006 Jun 21
1
Expected network throughput
Hi, I have just started to work with Xen and have a question regarding the expected network throughput. Here is my configuration: Processor: 2.8 GHz Intel Celeron (Socket 775) Motherboard: Gigabyte 8I865GVMF-775 Memory: 1.5 GB Basic system: Kubuntu 6.06 Dapper Drake Xen version: 3.02 (Latest 3.0 stable download) I get the following iperf results: Src Dest Throughput Dom0 Dom0
2010 Nov 24
1
slow network throughput, how to improve?
would like some input on this one please. Two CentOS 5.5 XEN servers, with 1GB NIC's, connected to a 1GB switch transfer files to each other at about 30MB/s between each other. Both servers have the following setup: CentOS 5.5 x64 XEN 1GB NIC's 7200rpm SATA HDD's The hardware configuration can't change, I need to use these servers as they are. They are both used in production