search for: nodea

Displaying 20 results from an estimated 23 matches for "nodea".

Did you mean: node
2017 May 22
6
Advertising a Public IP address
...I read it someplace sometime, but it evades me right now. It's also an opportunity to say hello to the list and many thanks for writing and supporting tinc vpn! We make great use of it at rhizomatica. So, Let's take this example setup. I have two tinc nodes (A and B) behind a firewall NodeA and NodeB have 192.168.1.2 and 192.168.1.3 assigned on an internal LAN, and they both have different public IP addresses forwarded to them, port 655 udp/tcp The rest of the nodes C-Z are spread out around the internet. NodeA is our "master" server with all the keys for all nodes, so eve...
2017 May 22
0
Advertising a Public IP address
...hus far, unless you have NodeB’s public key in Nodes C-Z (with NodeB’s public/reachable IP configured in Nodes C-Z) *and* you have the PublicKeys for NodesC-z configured in NodeB, there is no way that NodesC-Z will be able to establish a connection directly with NodeB, and it’ll have to forward via NodeA (or any other node that do have a direct connection established with NodeB using it’s keys etc) Understand it like this: for any two nodes to have a *direct* connection, they need to share the other’s Public Key to properly authenticate each other. It is a function of the security choices for TINC...
2009 Mar 22
2
GraphDumpFile - Feature request
...le parameter in tinc -- it certainly beats sending USR2 to tincd, and ploughing through syslog to figure out which nodes are connected. Could I request that GraphDumpFile dumps a little more data to the file, namely, the subnet; changing the output file from (for example): digraph { nodeA [label = "nodeA"]; nodeB [label = "nodeB"]; } ...to this... digraph { nodeA [label = "nodeA\n172.16.0.0/16"]; nodeB [label = "nodeB\n172.17.1.0/24"]; } Thanks, Peter
2004 Jan 19
1
daemon mode - local to server sync failing
I have an rsync server running on nodeA and a client, nodeB. I am attempting to do a sync/copy from the local machine, nodeB, to the rsync server, nodeA. The following line from the rsync man page makes it look like this is relatively simple: for copying from the local machine to a remote rsync server. This is invoked when the destinat...
2017 Jul 25
0
[Questions] About small files performance
Dear all Recently, i did some work to test small files performance for gnfsv3 transport. Following is my scenario. #####environment##### ==2 cluster nodes(nodeA/nodeB)== each is equipped with E5-2650*2, 128G memory and 10GB*2 netcard nodeA: 10.254.3.77 10.128.3.77 nodeB: 10.254.3.78 10.128.3.78 ==2 stress nodes(clientA/clientB)== each is equipped with E5-2650*2, 128G memory and 10GB*2 netcard clientA: 10.254.3.75 clientB: 10.254.3.76 1) 10.254.3.* is f...
2017 May 22
0
Advertising a Public IP address
On Mon, May 22, 2017 at 11:03:30AM +0200, Keith Whyte wrote: > I have two tinc nodes (A and B) behind a firewall > > NodeA and NodeB have 192.168.1.2 and 192.168.1.3 assigned on an internal > LAN, and they both have different public IP addresses forwarded to them, > port 655 udp/tcp > > I would like any given node in the C-Z group to be able to find Node B > on it's public IP and therefore not forwa...
2013 Jun 20
2
[PATCH] ocfs2: llseek requires to ocfs2 inode lock for the file in SEEK_END
llseek requires ocfs2 inode lock for updating the file size in SEEK_END. because the file size maybe update on another node. if it not . after call llseek in SEEK_END. the position is old. this bug can be reproduce the following scenario: at first ,we dd a test fileA,the file size is 10k. on NodeA: --------- 1) open the test fileA, lseek the end of file. and print the position. 2) close the test fileA on NodeB: 1) open the test fileA, append the 5k data to test FileA. 2) lseek the end of file. and print the position. 3) close file. at first we run the test program1 on NodeA , the result is...
2013 Jul 26
0
[PATCH] ocfs2: dlm_request_all_locks() should deal with the status sent from target node
dlm_request_all_locks() should deal with the status sent from target node if DLM_LOCK_REQUEST_MSG is sent successfully, or recovery master will fall into endless loop, waiting for other nodes to send locks and DLM_RECO_DATA_DONE_MSG to me. NodeA NodeB selected as recovery master dlm_remaster_locks() ->dlm_request_all_locks() send DLM_LOCK_REQUEST_MSG to nodeA...
2013 Aug 27
0
[patch 07/22] ocfs2: dlm_request_all_locks() should deal with the status sent from target node
...should deal with the status sent from target node dlm_request_all_locks() should deal with the status sent from target node if DLM_LOCK_REQUEST_MSG is sent successfully, or recovery master will fall into endless loop, waiting for other nodes to send locks and DLM_RECO_DATA_DONE_MSG to me. NodeA NodeB selected as recovery master dlm_remaster_locks() ->dlm_request_all_locks() send DLM_LOCK_REQUEST_MSG to nodeA...
2018 Apr 09
0
nm-bridge & nm-team and no connection.
hi guys I have a kvm quest & two nodes. a kvm guest on nodeA and IPaddr2 on nodeB(at this time) nodeA & nodeB comprise a HA cluster now... - nodeA can ping IPaddr2 IP - nodeB can ping kvm guest IP - kvm guest can get to nodeB's IP but* Only a bit non-common bit in my setup is: - both nodeA & nodeB be interfaces which are relevant to the subnet...
2015 May 21
2
IPv6 subnet routing
I have 2 nodes nodeA and nodeB I'm using tinc 1.1pre11 -- nodeA(fd80:2015:2105:abcd::1) : $ ip -6 route fd80:2015:2105:abcd::1 dev tun0 proto kernel metric 256 fd80:2015:2105:adcd::/64 dev tun0 metric 1024 fe80::/64 dev eth0 proto kernel metric 256 $ ping6 fd80:2015:2105:abcd::1 PING fd80:2015:2105:abcd::1(...
2013 Jun 27
0
[PATCH V3] ocfs2: llseek requires to ocfs2 inode lock for the file in SEEK_END
llseek requires ocfs2 inode lock for updating the file size in SEEK_END. because the file size maybe update on another node. if it not . after call llseek in SEEK_END. the position is old. this bug can be reproduce the following scenario: at first ,we dd a test fileA,the file size is 10k. on NodeA: --------- 1) open the test fileA, lseek the end of file. and print the position. 2) close the test fileA on NodeB: 1) open the test fileA, append the 5k data to test FileA. 2) lseek the end of file. and print the position. 3) close file. at first we run the test program1 on NodeA , the result is...
2011 Jun 21
1
wonder about redirect-gateway subnet/switch-mode
...is too short to understand. is VPN_GATEWAY set by tinc in switch mode? and why is "In switch mode, the gateway routes are necessary" mentioned here suddenly? afaik, Subnet is announce by node itself to other nodes. that mean i cannot set, for example. subnet=1.2.3.4/32 in hosts/X within nodeA, while subnet=5.6.7.8/32 in hosts/X within nodeB so, is it correct that i cannot have 0.0.0.0/0 for *two* gateway node as in http://www.tinc-vpn.org/examples/redirect-gateway/ ? mode=switch/hub instead of router have to be used in this case ?
2015 May 21
0
IPv6 subnet routing
You have a typo in your routes: fd80:2015:2105:adcd::/6 abcd vs. adcd On 21 May 2015, at 19:35, Martin <martinmoen at gmail.com<mailto:martinmoen at gmail.com>> wrote: I have 2 nodes nodeA and nodeB I'm using tinc 1.1pre11 -- nodeA(fd80:2015:2105:abcd::1) : $ ip -6 route fd80:2015:2105:abcd::1 dev tun0 proto kernel metric 256 fd80:2015:2105:adcd::/64 dev tun0 metric 1024 fe80::/64 dev eth0 proto kernel metric 256 $ ping6 fd80:2015:2105:abcd::1 PING fd80:2015:2105:abcd::1(...
2011 Jun 08
2
[Gluster3.2@Grid5000] 128 nodes failure and rr scheduler question
Hello, I'm driving some experiments on grid'5000 with GlusterFS 3.2 and, as a first point, i've been unable to start a volume featuring 128bricks (64 ok) Then, due to the round-robin scheduler, as the number of nodes increase (every node is also a brick), the performance of an application on an individual node decrease! So my question is : how to STOP the round-robin distribution
2015 May 21
2
IPv6 subnet routing
I've been trying out IPv6 networking with tinc and noticed that it will not route smaller segments than /48 If I try to run ip -6 route add aaaa:bbbb:cccc:dddd::/64 dev tun0 I get network unreachable However if I run ip -6 route add aaaa:bbbb:cccc::/48 dev tun0 it works fine Is this is a limitation in tinc or the kernel network stack itself? -------------- next part -------------- An HTML
2013 Sep 06
21
[PATCH v2 0/5] xl: allow for node-wise specification of vcpu pinning
Hi all, This is the second take of a patch that I submitted some time ago for allowing specifying vcpu pinning taking NUMA nodes into account. IOW, something like this: * "nodes:0-3": all pCPUs of nodes 0,1,2,3;  * "nodes:0-3,^node:2": all pCPUS of nodes 0,1,3;  * "1,nodes:1-2,^6": pCPU 1 plus all pCPUs of nodes 1,2    but not pCPU 6; v1 was a single patch, this is
2017 May 22
0
Advertising a Public IP address
On Mon, May 22, 2017 at 4:03 AM, Keith Whyte <keith at rhizomatica.org> wrote: > Is there a way to force NodeA or NodeB to "advertise" it's public IP to > the rest of the tinc network, or did I miss something really obvious? the config files can use DNS names, right? Have the nodes that you want to advertise their public IP addresses sign up with Dyndns or similar, presto done. -- every...
2004 Jul 18
7
Resize ocfs....?
I've tried several times to resize one of my ocfs volumes[see emcpowere1 below] I'm on the latest [prod] version of ocfs/ocfs-tools as of 3pm EST on 18JUL04. Per instructions... Take down db, unmount all ocfs drives, use ' tuneocfs -F -S 100G /dev/emcpowere1 ' Supposedly this should work, but I get get.... The size specified, 100G, is larger than the device size, 59G. Aborting.
2004 Oct 20
11
Shorewall, Freeswan and SuSE 9.1
I have been using shorewall and freeswan successfully for 3 or more years now. But they have all been using the Linux 2.4 kernel. My current configuration is (as the title suggests) using SuSE 9.1 which has a 2.6.5 kernel and freeswan 2.0.4 built-in. After much reading and a lot of trial and error, I did get this combination to work with Shorewall 2.0.9. It is happily talking to an older Mandrake