similar to: Strange network issue; Guest/DomU outgoing traffic

Displaying 20 results from an estimated 20000 matches similar to: "Strange network issue; Guest/DomU outgoing traffic"

2008 Sep 03
6
eth0 on dom0 not working on a bridged conf
Hi, I''ve installed xen 3.0.3 from packages (xen-linux-system) on a debian etch, and i''ve configured it with network-bridge script in the default way (netdev=eth0, bridge=xenbr0, etc...) which is ok for me. the problem I have: I cannot ping any outer machine from dom0 (nor any outer machine can ping me). It gives me a "Destination Host Unreachable" message that
2009 Feb 04
1
AW: Workaround for pcifront issues
Thanks Anna, this was also quite interesting for me. Did you mean that I just have to eliminate the else branch or raise arround "All devices behind the uppermost PCI/PCI-X bridge must be co-assigned to the same guest." or do I have to eliminate more in order to prevent an FLR (whatever that is). For my board it''s definitely necessary to move PCI devices in different guests,
2008 Feb 01
11
xenoprof error
Hi , I am getting following error when i try to run xenoprof /usr/local/bin/opcontrol: line 1082: echo: write error: Invalid argument Using 2.6+ OProfile kernel interface. Reading module info. Using log file /var/lib/oprofile/oprofiled.log Daemon started. what is this write error ? Any help is really appreciated. -John _______________________________________________ Xen-users mailing list
2008 Feb 01
11
xenoprof error
Hi , I am getting following error when i try to run xenoprof /usr/local/bin/opcontrol: line 1082: echo: write error: Invalid argument Using 2.6+ OProfile kernel interface. Reading module info. Using log file /var/lib/oprofile/oprofiled.log Daemon started. what is this write error ? Any help is really appreciated. -John _______________________________________________ Xen-users mailing list
2010 Jun 11
10
Slow TCP performance between Windows Vista and Xen PV-on-HVM guest
I am running a Xen HVM guest with netfront PV drivers. This is running SLES10 SP3 inside the guest. The Dom0 is also SLES10 SP3. Now I am trying to communicate from that HVM guest to a Windows Visa or also Windows 7 machine and I am getting really poor TCP performance. When tracing on the network traffic, I can see that no packets are dropped or missing or anything, but what happens is that the
2009 Jun 09
5
Intel Quad NIC made visible in guest -> system crash
Hi! we have two Intel Quad Nic 82576, PCI ID 8086:10E8 and use the igb driver 1.3.19.3 on Debian 5.0.1. I used the pciback.hide XEN kernel parameter and made on of the NIC''s interfaces available in a DomU. Now, when I am starting the VM, the system crashes (log attached) I also tried to boot with the irqpoll option - the Interrupt Disabled message still appears when I try to start the
2008 Apr 21
7
Weird network performance behaviour?
I'm recording TCP network throughput performance between Dom0 IP stack and all my guest domains (currently 6 running simultaneously). All guests are running simultaneously, but I only ever transmit between Dom0 and one of the DomUs while the other five are idle. When running a netperf TCP_STREAM test (netperf on Dom0, netserver on DomU) I record the following performance numbers, CPU % as
2009 Aug 13
4
[Bridge] [PATCH] net/bridge: Add 'hairpin' port forwarding mode
This patch adds a 'hairpin' (also called 'reflective relay') mode port configuration to the Linux Ethernet bridge kernel module. A bridge supporting hairpin forwarding mode can send frames back out through the port the frame was received on. Hairpin mode is required to support basic VEPA (Virtual Ethernet Port Aggregator) capabilities. You can find additional information on VEPA
2009 Aug 13
4
[Bridge] [PATCH] net/bridge: Add 'hairpin' port forwarding mode
This patch adds a 'hairpin' (also called 'reflective relay') mode port configuration to the Linux Ethernet bridge kernel module. A bridge supporting hairpin forwarding mode can send frames back out through the port the frame was received on. Hairpin mode is required to support basic VEPA (Virtual Ethernet Port Aggregator) capabilities. You can find additional information on VEPA
2009 Aug 13
4
[Bridge] [PATCH] net/bridge: Add 'hairpin' port forwarding mode
This patch adds a 'hairpin' (also called 'reflective relay') mode port configuration to the Linux Ethernet bridge kernel module. A bridge supporting hairpin forwarding mode can send frames back out through the port the frame was received on. Hairpin mode is required to support basic VEPA (Virtual Ethernet Port Aggregator) capabilities. You can find additional information on VEPA
2007 Sep 07
2
Dom0 cannot see network when bridge is enabled
I''ve seen a lot of threads w/ similar problems, but none have posted a resolution. I am using Debian 4.0r1 (Etch). I was using the xen packages from stable, but have tried w/ testing as well and the problem persists. http://pastie.caboo.se/95144 Host is 10.0.0.20 on network 10.0.0.0/24. Dom0 is thus 10.0.0.20 DomU is 10.0.0.30 When the bridge is enabled, DomU can ping everything.
2010 Jun 02
14
ARP problems with xen 4.0 with pvops kernel
Hello, Finally I managed to get a xen 4.0 working on ubuntu 10.04 with pvops kernel and libvirt. However I am having some problems with networking... after initial installation with netinstall image in hvm mode, when I transform the vm in xen pv (via pygrub with the current ubuntu kernel), networking startEd to act weird... Basically I''m not using a network script from xen. I define a
2010 Jun 02
14
ARP problems with xen 4.0 with pvops kernel
Hello, Finally I managed to get a xen 4.0 working on ubuntu 10.04 with pvops kernel and libvirt. However I am having some problems with networking... after initial installation with netinstall image in hvm mode, when I transform the vm in xen pv (via pygrub with the current ubuntu kernel), networking startEd to act weird... Basically I''m not using a network script from xen. I define a
2010 Jun 14
4
Promiscuous mode
Hi Everyone, In order to prevent DomU from entering promiscuous mode, is it just a matter of adding these 2 rules when the vif is created? # Accept packets leaving the bridge going to the domU only if # the destination IP for that packet matches an authorized IPv4 # address for that domU. iptables -A FORWARD -m physdev --physdev-out vif1.0 \ --destination 216.146.46.43 -j ACCEPT
2006 Feb 10
10
Seeing the IP addresses of domUs from dom0
Using a default xen3.0.1 setup for dom0(rhel4 distro) and domUs (gentoo, rhel4, centos4.2 distros, ttylinux) with the domUs all having dhcp turned on, is there an easy way to tell from within the dom0 what ip addresses were assigned to the domUs after bootup? Toby Ford USi _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com
2010 Apr 27
3
High network latency on first packet
Hi all, My setup is Debian testing dom0/domUs with a 2.6.32 pvops kernel from Debian unstable. Hypervisor is 3.4.2 from Debian testing. I use network- and vif- route with a default route in domU pointing to the nic because multiple IPs with bridge would trigger port shutdown on the switch (only one mac-address allowed per port). A subnet is routed to the dom0, which then knows which addresses the
2007 Sep 25
4
No eth0 device found in domU
Hi, I tried already a long time to enable network in my domU but I always fail with: # ifconfig eth0 192.168.0.1 up SIOCSIFADDR: No such device SIOCGIFFLAGS: No such device Do I need to load a kernel module? I use OpenSuse 10.2 with kernel 2.6.18.8-0.5-xen (for both domU and dom0). My config: kernel = "/boot/vmlinuz-xen" ramdisk = "/home/jens/initrd.gz" memory = 1024 name
2010 Nov 16
4
problem after migration
Hi, I believe I can post this here. I have a domU ( Redhat 6) running in Paravirt mode. When I migrate this domU to another host, looks like the domU is freezing in the other host. When I take it back to original host ,it starts to run normally and this is what can see in dmesg of domU : suspending xenstore... trying to map vcpu_info 0 at ffff88000434b020, mfn b67660, offset 32 cpu 0 using
2009 Aug 07
3
[Bridge] [evb] RE: [PATCH][RFC] net/bridge: add basic VEPA support
Paul, I also think that bridge may not be the right place for VEPA, but rather a simpler sw/hw mux Although the VEPA support may reside in multiple places (I.e. also in the bridge) As Arnd pointed out Or already added an extension to qemu that allow direct guest virtual NIC mapping to an interface device (vs using tap), this was done specifically to address VEPA, and result in much faster
2009 Aug 07
3
[Bridge] [evb] RE: [PATCH][RFC] net/bridge: add basic VEPA support
Paul, I also think that bridge may not be the right place for VEPA, but rather a simpler sw/hw mux Although the VEPA support may reside in multiple places (I.e. also in the bridge) As Arnd pointed out Or already added an extension to qemu that allow direct guest virtual NIC mapping to an interface device (vs using tap), this was done specifically to address VEPA, and result in much faster