similar to: Transport Endpoint Not connected while running sysbench on Gluster Volume

Displaying 20 results from an estimated 200 matches similar to: "Transport Endpoint Not connected while running sysbench on Gluster Volume"

2017 Jun 13
2
Transport Endpoint Not connected while running sysbench on Gluster Volume
I'm having a hard time trying to get a gluster volume up and running. I have setup other gluster volumes on other systems without much problems but this one is killing me. The gluster vol was created with the command: gluster volume create mariadb_gluster_volume laeft-dccdb01p:/export/mariadb/brick I had to lower frame-timeout since the system would become unresponsive until the frame failed
2017 Jun 14
0
Transport Endpoint Not connected while running sysbench on Gluster Volume
Also, this is the profile output of this Volume: gluster> volume profile mariadb_gluster_volume info cumulative Brick: laeft-dccdb01p.core.epay.us.loc:/export/mariadb_backup/brick ------------------------------------------------------------------- Cumulative Stats: Block Size: 16384b+ 32768b+ 65536b+ No. of Reads: 0 0 0
2017 Aug 22
0
Performance testing with sysbench...
Hi all, I'm doing some performance test... If I test a simple sequential write using dd I get a thorughput of about 550 Mb/s. When I do a sequential write test using sysbench this drops to about 200. Is this due to the way sysbench tests? Or has in this case the performance of sysbench itself become the bottleneck? Krist -- Vriendelijke Groet | Best Regards | Freundliche Gr??e |
2009 Jul 15
2
Crypto in 7000 Family
Hi Darren, I found a presentation (Data at Rest: ZFS & Lofi Crypto) with your information. Do yo have any information about the release date for encryption in ZFS, or encryption related in our 7000 Unified Storage Family. Thanks in advance for any help and information that you can send to me. -- <http://www.sun.com> * H. Alejandro Guevara Castro * Storage Practice Solution
2013 Sep 29
9
DomU vs Dom0 performance.
Hi, I have been doing some diskIO bench-marking of dom0 and domU (HVM). I ran into an issue where domU performed better than dom0. So I ran few experiments to check if it is just diskIO performance. I have an archlinux (kernel 3.5.0) + xen 4.2.2) installed on a Intel Core i7 Q720 machine. I have also installed archlinux (kernel 3.5.0) in domU running on this machine. The domU runs with 8 vcpus.
2012 Apr 17
2
Kernel bug in BTRFS (kernel 3.3.0)
Hi, Doing some extensive benchmarks on BTRFS, I encountered a kernel bug in BTRFS (as reported in dmesg) Maybe the information below can help you making btrfs better. Situation Doing an intensive sequential write on a SAS 3TB disk drive (SEAGATE ST33000652SS) with 128 threads with Sysbench. Device is connected through an HBA. Blocksize was 256k ; Kernel is 3.3.0 (x86_64) ; Btrfs is version
2009 Mar 05
1
[PATCH] OCFS2: Pagecache usage optimization on OCFS2
Hi. I introduced "is_partially_uptodate" aops for OCFS2. A page can have multiple buffers and even if a page is not uptodate, some buffers can be uptodate on pagesize != blocksize environment. This aops checks that all buffers which correspond to a part of a file that we want to read are uptodate. If so, we do not have to issue actual read IO to HDD even if a page is not uptodate
2020 Jun 15
1
very low performance of Xen guests
On 6/15/20 2:46 PM, Stephen John Smoogen wrote: > > > On Sun, 14 Jun 2020 at 14:49, Manuel Wolfshant > <wolfy at nobugconsulting.ro <mailto:wolfy at nobugconsulting.ro>> wrote: > > Hello > > > ??? For the past months I've been testing upgrading my Xen hosts > to CentOS 7 and I face an issue for which I need your help to solve. > >
2020 Jun 14
4
very low performance of Xen guests
Hello ??? For the past months I've been testing upgrading my Xen hosts to CentOS 7 and I face an issue for which I need your help to solve. ??? The testing machines are IBM blades, model H21 and H21XM. Initial tests were performed on the H21 with 16 GB RAM; during the last 6=7 weeks I've been using the H21XM with 64 GB. In all cases the guests were fully updated CentOS 7 --
2013 Nov 19
6
[PATCH] Btrfs: fix very slow inode eviction and fs unmount
The inode eviction can be very slow, because during eviction we tell the VFS to truncate all of the inode''s pages. This results in calls to btrfs_invalidatepage() which in turn does calls to lock_extent_bits() and clear_extent_bit(). These calls result in too many merges and splits of extent_state structures, which consume a lot of time and cpu when the inode has many pages. In some
2017 Feb 14
6
[PATCH v2 0/3] x86/vdso: Add Hyper-V TSC page clocksource support
Hi, while we're still waiting for a definitive ACK from Microsoft that the algorithm is good for SMP case (as we can't prevent the code in vdso from migrating between CPUs) I'd like to send v2 with some modifications to keep the discussion going. Changes since v1: - Document the TSC page reading protocol [Thomas Gleixner]. - Separate the TSC page reading code from
2017 Feb 14
6
[PATCH v2 0/3] x86/vdso: Add Hyper-V TSC page clocksource support
Hi, while we're still waiting for a definitive ACK from Microsoft that the algorithm is good for SMP case (as we can't prevent the code in vdso from migrating between CPUs) I'd like to send v2 with some modifications to keep the discussion going. Changes since v1: - Document the TSC page reading protocol [Thomas Gleixner]. - Separate the TSC page reading code from
2009 Jun 08
1
[PATCH] Btrfs: fdatasync should skip metadata writeout
Hi. In btrfs, fdatasync and fsync are identical. I think fdatasync should skip committing transaction when inode->i_state is set just I_DIRTY_SYNC and this indicates only atime or/and mtime updates. Following patch improves fdatasync throughput. #sysbench --num-threads=16 --max-requests=10000 --test=fileio --file-block-size=4K --file-total-size=16G --file-test-mode=rndwr
2009 Jan 28
0
smp_tlb_shootdown bottleneck?
Hi. Sometimes I see much contention in smp_tlb_shootdown while running sysbench: sysbench --test=fileio --num-threads=8 --file-test-mode=rndrd --file-total-size=3G run kern.smp.cpus: 8 FreeBSD 7.1-R CPU: 0.8% user, 0.0% nice, 93.8% system, 0.0% interrupt, 5.4% idle Mem: 11M Active, 2873M Inact, 282M Wired, 8K Cache, 214M Buf, 765M Free Swap: 4096M Total, 4096M Free PID USERNAME PRI NICE
2012 Dec 03
17
[PATCH 0 of 3] xen: sched_credit: fix tickling and add some tracing
Hello, This small series deals with some weirdness in the mechanism with which the credit scheduler choses what PCPU to tickle upon a VCPU wake-up. Details are available in the changelog of the first patch. The new approach has been extensively benchmarked and proved itself either beneficial or harmless. That means it does not introduce any significant amount of overhead and/or performances
2005 Apr 29
1
Can't login samba domain from xp/2k
Hello everybody. I am having troubles to log on a samba domain from winxp and win 2000 workstations. I patched the registry with the requiresignorseal=0, changed the local and group policies (following various comments found on the web), and added the machine account for samba (obviously, /etc/passwd too). I tried with samba 3.0.7, 3.0.10 and 3.0.13, Winxp with and without sp2, and win200
2009 Dec 10
4
Linux router with CentOS
Hello everybody. I'm wondering here if is it possible to setup a CentOS machine as a router for two Internet connections in a LAN. This _router_ would work as the gateway for the workstations using DHCPD. The purpose of this is to optimize the broadband "joining" both connections, and given the case, do not lose the Internet access. ?is this possible? ?too much complicated?
2020 Jun 15
0
very low performance of Xen guests
On Sun, 14 Jun 2020 at 14:49, Manuel Wolfshant <wolfy at nobugconsulting.ro> wrote: > Hello > > > For the past months I've been testing upgrading my Xen hosts to CentOS > 7 and I face an issue for which I need your help to solve. > > The testing machines are IBM blades, model H21 and H21XM. Initial > tests were performed on the H21 with 16 GB RAM; during
2010 May 05
6
Benchmark Disk IO
What is the best way to benchmark disk IO? I'm looking to move one of my servers, which is rather IO intense. But not without first benchmarking the current and new disk array, To make sure this isn't a full waste of time. thanks
2017 Feb 08
3
[PATCH RFC 0/2] x86/vdso: Add Hyper-V TSC page clocksource support
Hi, Hyper-V TSC page clocksource is suitable for vDSO, however, the protocol defined by the hypervisor is different from VCLOCK_PVCLOCK. I implemented the required support re-using pvclock_page VVAR. Simple sysbench test shows the following results: Before: # time sysbench --test=memory --max-requests=500000 run ... real 1m22.618s user 0m50.193s sys 0m32.268s After: # time sysbench