similar to: DO NOT REPLY [Bug 5167] New: The size of the transferred part of a file exceeds 2 Gb.

Displaying 20 results from an estimated 900 matches similar to: "DO NOT REPLY [Bug 5167] New: The size of the transferred part of a file exceeds 2 Gb."

2005 Oct 05
2
Define variable in sip.conf
I'm looking for a way to transmit a user specific variable to my dialplan If we use the example of the hair color, I was thinking of having something like: [bob] context=users host=dynamic secret=password type=friend username=bob hair=brown [lary] context=users host=dynamic secret=password type=friend username=lary hair=black And be able to access from the dialplan: [users] Exten =>
2003 Oct 01
0
[releng_4 tinderbox] failure on i386/i386
TB --- 2003-10-02 04:44:07 - starting RELENG_4 tinderbox run for i386/i386 TB --- 2003-10-02 04:44:07 - checking out the source tree TB --- cd /home/des/tinderbox/RELENG_4/i386/i386 TB --- /usr/bin/cvs -f -R -q -d/home/ncvs update -Pd -rRELENG_4 src TB --- 2003-10-02 04:49:12 - building world TB --- cd /home/des/tinderbox/RELENG_4/i386/i386/src TB --- /usr/bin/make -B buildworld >>>
2003 Oct 01
0
[releng_4 tinderbox] failure on alpha/alpha
TB --- 2003-10-02 04:00:01 - starting RELENG_4 tinderbox run for alpha/alpha TB --- 2003-10-02 04:00:01 - checking out the source tree TB --- cd /home/des/tinderbox/RELENG_4/alpha/alpha TB --- /usr/bin/cvs -f -R -q -d/home/ncvs update -Pd -rRELENG_4 src TB --- 2003-10-02 04:07:38 - building world TB --- cd /home/des/tinderbox/RELENG_4/alpha/alpha/src TB --- /usr/bin/make -B buildworld >>>
2014 Jun 27
1
geo-replication status faulty
Venky Shankar, can you follow up on these questions? I too have this issue and cannot resolve the reference to '/nonexistent/gsyncd'. As Steve mentions, the nonexistent reference in the logs looks like the culprit especially seeing that the ssh command trying to be run is printed on an earlier line with the incorrect remote path. I have followed the configuration steps as documented in
2012 Apr 12
1
CentOS 6.2 anaconda bug?
I have a kickstart file with the following partitioning directives: part /boot --fstype ext3 --onpart=sda1 part pv.100000 --onpart=sda2 --noformat volgroup vol0 pv.100000 --noformat logvol / --vgname=vol0 --name=lvol1 --useexisting --fstype=ext4 logvol /tmp --vgname=vol0 --name=lvol2 --useexisting --fstype=ext4 logvol swap --vgname=vol0 --name=lvol3 --useexisting logvol /data --vgname=vol0
2006 May 30
1
Cannot remove Maildir folder
Hi, I have been using dovecot 1.0 beta 8 on Debian Sarge for a couple of days and I experienced some problems when removing folders.. As an IMAP client I use Thunderbird 1.5 and IMAP folders are on NFS. Here are the steps to re-produce the problem: 1. Run Thunderbird as usual 2. Create folder 3. Remove folder - In my case folder is not removed but moved to Trash (Thunderbird setting) 4. I go to
2017 Dec 18
2
Upgrading from Gluster 3.8 to 3.12
Hi, I have a cluster of 10 servers all running Fedora 24 along with Gluster 3.8. I'm planning on doing rolling upgrades to Fedora 27 with Gluster 3.12. I saw the documentation and did some testing but I would like to run my plan through some (more?) educated minds. The current setup is: Volume Name: vol0 Distributed-Replicate Number of Bricks: 2 x (2 + 1) = 6 Bricks: Brick1:
2017 Dec 19
0
Upgrading from Gluster 3.8 to 3.12
On Tue, Dec 19, 2017 at 1:10 AM, Ziemowit Pierzycki <ziemowit at pierzycki.com> wrote: > Hi, > > I have a cluster of 10 servers all running Fedora 24 along with > Gluster 3.8. I'm planning on doing rolling upgrades to Fedora 27 with > Gluster 3.12. I saw the documentation and did some testing but I > would like to run my plan through some (more?) educated minds. >
2013 Jul 10
1
Re: Libvirt and Glusterfs
On 07/09/2013 08:18 PM, Olivier Mauras wrote: > On 2013-07-09 09:40, Vijay Bellur wrote: > >>> Hi, I'm trying to use qemu native glusterfs integration with libvirt. >>> It's all working well from the qemu side, but libvirt fails to start >>> a domain with a gluster drive or attach a drive. I have exactly the >>> same error as this person:
2013 Jul 07
0
Libvirt and Glusterfs
Hi, I'm trying to use qemu native glusterfs integration with libvirt. It's all working well from the qemu side, but libvirt fails to start a domain with a gluster drive or attach a drive. I have exactly the same error as this person: https://www.redhat.com/archives/libvirt-users/2013-April/msg00204.html [1] I use qemu 1.5.1 with glusterfs 3.4 beta 4 and libvirt 1.0.6. [root@bbox ~]#
2013 Jul 09
0
Re: Libvirt and Glusterfs
On 2013-07-09 09:40, Vijay Bellur wrote: >> Hi, I'm trying to use qemu native glusterfs integration with libvirt. It's all working well from the qemu side, but libvirt fails to start a domain with a gluster drive or attach a drive. I have exactly the same error as this person: https://www.redhat.com/archives/libvirt-users/2013-April/msg00204.html [1] I use qemu 1.5.1 with glusterfs
2013 Jul 09
2
Re: Libvirt and Glusterfs
> Hi, > > I'm trying to use qemu native glusterfs integration with libvirt. It's > all working well from the qemu side, but libvirt fails to start a domain > with a gluster drive or attach a drive. > I have exactly the same error as this person: > https://www.redhat.com/archives/libvirt-users/2013-April/msg00204.html > > I use qemu 1.5.1 with glusterfs 3.4 beta 4
2008 Jul 30
1
Unable to write video dvd using k3b
Hi, I download dvd movie files from torrent.It is like below -rw-r--r-- 1 root root 1048614912 2008-07-24 16:49 VTS_01_1.VOB -rw-r--r-- 1 root root 538210304 2008-07-24 16:52 VTS_01_2.VOB File command on above giving the below ouput. file VTS_01_1.VOB VTS_01_1.VOB: MPEG sequence, v2, program multiplex file VTS_01_2.VOB VTS_01_2.VOB: MPEG sequence, v2, program multiplex When i opened any
2018 Mar 01
0
df reports wrong full capacity for distributed volumes (Glusterfs 3.12.6-1)
I'm sorry for my last incomplete message. Below the output of both volumes: [root at stor1t ~]# gluster volume rebalance volumedisk1 status Node Rebalanced-files size scanned failures skipped status run time in h:m:s --------- ----------- ----------- ----------- -----------
2018 Mar 01
0
df reports wrong full capacity for distributed volumes (Glusterfs 3.12.6-1)
Hi Jose, On 28 February 2018 at 22:31, Jose V. Carri?n <jocarbur at gmail.com> wrote: > Hi Nithya, > > My initial setup was composed of 2 similar nodes: stor1data and stor2data. > A month ago I expanded both volumes with a new node: stor3data (2 bricks > per volume). > Of course, then to add the new peer with the bricks I did the 'balance > force' operation.
2018 Feb 28
0
df reports wrong full capacity for distributed volumes (Glusterfs 3.12.6-1)
Hi Jose, On 28 February 2018 at 18:28, Jose V. Carri?n <jocarbur at gmail.com> wrote: > Hi Nithya, > > I applied the workarround for this bug and now df shows the right size: > > That is good to hear. > [root at stor1 ~]# df -h > Filesystem Size Used Avail Use% Mounted on > /dev/sdb1 26T 1,1T 25T 4% /mnt/glusterfs/vol0 > /dev/sdc1
2018 Feb 28
2
df reports wrong full capacity for distributed volumes (Glusterfs 3.12.6-1)
Hi Nithya, My initial setup was composed of 2 similar nodes: stor1data and stor2data. A month ago I expanded both volumes with a new node: stor3data (2 bricks per volume). Of course, then to add the new peer with the bricks I did the 'balance force' operation. This task finished successfully (you can see info below) and number of files on the 3 nodes were very similar . For volumedisk1 I
2018 Mar 01
2
df reports wrong full capacity for distributed volumes (Glusterfs 3.12.6-1)
Hi Nithya, Below the output of both volumes: [root at stor1t ~]# gluster volume rebalance volumedisk1 status Node Rebalanced-files size scanned failures skipped status run time in h:m:s --------- ----------- ----------- ----------- ----------- ----------- ------------
2011 Apr 08
1
Kickstart and lvm
I've written some %pre code to grab a few files off a logical volume, if it exits, before the disk gets formatted, but can't get get it to work correctly. Essentially: %pre ... lvm vgscan lvm vgchange -a y ... if [ -d /dev/vol0 ]; then # do stuff fi lvm vgremove -f vol0 The problem is that /dev/vol0 does not exist after lvm vgchange. I added a few debug statements (lvm pvdisplay, lvm
2014 Apr 15
0
Botched kernel update
I have a 6.5 box that failed to boot after a kernel update. The reason is the first arg on the kernel line: title CentOS (2.6.32-431.11.2.el6.x86_64) root (hd0,0) kernel /tboot.gz ro root=/dev/mapper/vol0-lvol1 intel_iommu=on rd_NO_LUKS LANG=en_US.UTF-8 rd_NO_MD SYSFONT=latarcyrheb-sun16 crashkernel=auto rd_NO_DM KEYBOARDTYPE=pc KEYTABLE=us rd_LVM_LV=vol0/lv ol3