[one-users] newbie how-to configure network to connect to the vm

Valentin Bud valentin.bud at gmail.com
Thu Nov 20 09:36:11 PST 2014


Hi Nikolaj,

Thomas deservers all the credit, he is wise enough to ask the right
questions.
After this was just a matter of text scraping :).

Best
Valentin


On Thu, Nov 20, 2014 at 7:23 PM, Nikolaj Majorov <nikolaj at majorov.biz>
wrote:

> Hi,
> many thanks Valentin ! Thanks Thomas !
>
> Valentin  you are right ! changing prefix , allow me to connect to the
> guest vm…  Cool !
>
> regards,
>
> Nikolaj
>
> On 20.11.2014, at 17:56, Valentin Bud <valentin.bud at gmail.com> wrote:
>
> Hello Nikolaj,
>
> I think the problem is that your AR uses a wrong network prefix. virbr0 is
> configured
> with 192.168.122.0/24 prefix and your definition in the AR is IP =
> 192.168.0.100.
>
> Best,
> Valentin
>
> On Thu, Nov 20, 2014 at 6:20 PM, Nikolaj Majorov <nikolaj at majorov.biz>
> wrote:
>
>> hi Thomas,
>> thanks for help.
>>
>>
>> On the open nebula host server :
>>
>> ——————————————-
>> [root at CentOS-70-64-minimal ~]# ip a
>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>>     inet 127.0.0.1/8 scope host lo
>>        valid_lft forever preferred_lft forever
>>     inet6 ::1/128 scope host
>>        valid_lft forever preferred_lft forever
>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> master br0 state UP qlen 1000
>>     link/ether 6c:62:6d:d9:09:05 brd ff:ff:ff:ff:ff:ff
>>     inet6 fe80::6e62:6dff:fed9:905/64 scope link
>>        valid_lft forever preferred_lft forever
>> 3: br0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP
>>     link/ether 6c:62:6d:d9:09:05 brd ff:ff:ff:ff:ff:ff
>>     inet 46.4.99.35 peer 46.4.99.33/32 brd 46.4.99.35 scope global br0
>>        valid_lft forever preferred_lft forever
>>     inet6 fe80::6e62:6dff:fed9:905/64 scope link
>>        valid_lft forever preferred_lft forever
>> 4: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
>> UP
>>     link/ether fe:00:c0:a8:00:64 brd ff:ff:ff:ff:ff:ff
>>     inet 192.168.122.1/24 brd 192.168.122.255 scope global virbr0
>>        valid_lft forever preferred_lft forever
>> 9: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> master virbr0 state UNKNOWN qlen 500
>>     link/ether fe:00:c0:a8:00:64 brd ff:ff:ff:ff:ff:ff
>>     inet6 fe80::fc00:c0ff:fea8:64/64 scope link
>>        valid_lft forever preferred_lft forever
>> 10: vnet1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> master virbr0 state UNKNOWN qlen 500
>>     link/ether fe:00:c0:a8:00:65 brd ff:ff:ff:ff:ff:ff
>>     inet6 fe80::fc00:c0ff:fea8:65/64 scope link
>>        valid_lft forever preferred_lft forever
>> 11: vnet2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> master virbr0 state UNKNOWN qlen 500
>>     link/ether fe:00:c0:a8:00:66 brd ff:ff:ff:ff:ff:ff
>>     inet6 fe80::fc00:c0ff:fea8:66/64 scope link
>>        valid_lft forever preferred_lft forever
>>
>>
>> [root at CentOS-70-64-minimal ~]# ip r
>> default via 46.4.99.33 dev br0
>> 46.4.99.33 dev br0  proto kernel  scope link  src 46.4.99.35
>> 169.254.0.0/16 dev br0  scope link  metric 1003
>> 192.168.0.102 dev virbr0  scope link
>> 192.168.122.0/24 dev virbr0  proto kernel  scope link  src 192.168.122.1
>>
>> [root at CentOS-70-64-minimal ~]# ip neigh show
>> 192.168.0.102 dev virbr0 lladdr 02:00:c0:a8:00:66 STALE
>> 192.168.0.101 dev br0 lladdr 02:00:c0:a8:00:65 STALE
>> 192.168.122.56 dev virbr0 lladdr 02:00:c0:a8:00:64 STALE
>> 192.168.0.100 dev virbr0  FAILED
>> 192.168.0.100 dev br0  FAILED
>> 46.4.99.33 dev br0 lladdr 00:26:88:75:e6:88 REACHABLE
>>
>> [root at CentOS-70-64-minimal ~]# iptables -nvL
>> Chain INPUT (policy ACCEPT 767K packets, 2018M bytes)
>>  pkts bytes target     prot opt in     out     source
>> destination
>>     0     0 ACCEPT     udp  --  virbr0 *       0.0.0.0/0
>> 0.0.0.0/0            udp dpt:53
>>     0     0 ACCEPT     tcp  --  virbr0 *       0.0.0.0/0
>> 0.0.0.0/0            tcp dpt:53
>>    17  5576 ACCEPT     udp  --  virbr0 *       0.0.0.0/0
>> 0.0.0.0/0            udp dpt:67
>>     0     0 ACCEPT     tcp  --  virbr0 *       0.0.0.0/0
>> 0.0.0.0/0            tcp dpt:67
>>
>> Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
>>  pkts bytes target     prot opt in     out     source
>> destination
>>     0     0 ACCEPT     all  --  *      virbr0  0.0.0.0/0
>> 192.168.122.0/24     ctstate RELATED,ESTABLISHED
>>     0     0 ACCEPT     all  --  virbr0 *       192.168.122.0/24
>> 0.0.0.0/0
>>     0     0 ACCEPT     all  --  virbr0 virbr0  0.0.0.0/0
>> 0.0.0.0/0
>>     0     0 REJECT     all  --  *      virbr0  0.0.0.0/0
>> 0.0.0.0/0            reject-with icmp-port-unreachable
>>     0     0 REJECT     all  --  virbr0 *       0.0.0.0/0
>> 0.0.0.0/0            reject-with icmp-port-unreachable
>>
>> Chain OUTPUT (policy ACCEPT 685K packets, 270M bytes)
>>  pkts bytes target     prot opt in     out     source
>> destination
>>
>> but the iptables seems not running:
>>
>> [root at CentOS-70-64-minimal ~]# service iptables status
>> Redirecting to /bin/systemctl status  iptables.service
>> iptables.service - IPv4 firewall with iptables
>>    Loaded: loaded (/usr/lib/systemd/system/iptables.service; enabled)
>>    Active: inactive (dead)
>>
>> Nov 20 16:38:39 CentOS-70-64-minimal systemd[1]: Stopped IPv4 firewall
>> with iptables.
>>
>>
>> ————————
>>
>>
>>
>> from within  vm (can connect only  to one with vnc so can’t simple cut
>> and paste):
>>
>> +++++++++++++++++++++++++++++++++++++++++++++++++++
>>
>> #ifconfig -a
>> eth0: Link…
>>          inet addr:192.168.0.102 Bcast:192.168.0.255 Mask: 255.255.255.0
>>
>>
>> # route -n
>>  Kernel IP routing table
>> Destingation G
>>
>> Destingation            Gateway          Genmask                 Flags
>> Metric Ref    Use Iface
>> 192.168.0.0             0.0.0.0               255.255.255.0         U
>>   0           0        0    eth0
>> 0.0.0.0                     192.168.0.1       0.0.0.0
>> UG    0           0        0    eth0
>>
>> iptables ist not running….
>>
>> +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>
>>
>>
>>
>> On 20.11.2014, at 16:50, Thomas Stein <himbeere at meine-oma.de> wrote:
>>
>>
>> Please post the output of
>>
>> ip a
>> ip r
>> ip neigh show
>>
>> iptables -nvL
>>
>> from the hardwarenode the vm is running on and from within the vm.
>>
>> cheers
>> t.
>>
>>
>> On 20. November 2014 16:44:29 MEZ, Nikolaj Majorov <nikolaj at majorov.biz>
>> wrote:
>>
>> HI,
>> unfortunately I can’t connect … :-(
>>
>> Here is my new configuration:
>>
>> $cat mynetwork_virbr0.one
>> NAME = "2private"
>>
>> BRIDGE = virbr0
>>
>> AR = [
>>   TYPE = IP4,
>>   IP = 192.168.0.100,
>>   SIZE = 3
>> ]
>>
>>
>>
>> $ tracepath -n 192.168.0.101
>> 1:  46.4.99.35                                            0.052ms pmtu
>> 1500
>> 1:  46.4.99.33                                            0.901ms
>> 1:  46.4.99.33                                            0.956ms
>> 2:  213.239.224.193                                       0.305ms
>> 3:  213.239.245.125                                       0.453ms
>> 4:  213.239.245.221                                       3.034ms
>> 5:  no reply
>> 6:  no reply
>> 7:  no reply
>> 8:  no reply
>>
>>
>> On 20.11.2014, at 15:00, Thomas Stein <himbeere at meine-oma.de> wrote:
>>
>> On Thursday 20 November 2014 11:57:17 Nikolaj Majorov wrote:
>>
>> Hello,
>>
>>
>> I install open nebula 4.8 on centos 7 as described in quickstarts
>>
>> and
>>
>> confiture my private network as it was showed in example:
>>
>> NAME = "private"
>>
>> BRIDGE = br0
>>
>> AR = [
>>   TYPE = IP4,
>>   IP = 192.168.0.100,
>>   SIZE = 3
>> ]
>>
>>
>> I suppose you using virbr0 as bridge device. Can you try this?
>>
>> t.
>>
>> Starting vm I have got IP 192.168.0.100, but  I can’t connect to the
>>
>> VM over
>>
>> ssh and even ping it. What should I do with network to access the VM
>>
>> ( add
>>
>> route/gateway  ) ? I’m really newbie in the networking  , so please
>>
>> give
>>
>> some hints ..
>>
>>
>> So my network configuration is :
>>
>> ifconfig -a
>>
>> br0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1500
>>       inet 46.4.99.35  netmask 255.255.255.255  broadcast
>>
>> 46.4.99.35
>>
>>       inet6 fe80::6e62:6dff:fed9:905  prefixlen 64  scopeid
>>
>> 0x20<link>
>>
>>       ether 6c:62:6d:d9:09:05  txqueuelen 0  (Ethernet)
>>       RX packets 522518  bytes 1880517877 (1.7 GiB)
>>       RX errors 0  dropped 0  overruns 0  frame 0
>>       TX packets 444071  bytes 51737672 (49.3 MiB)
>>       TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
>>
>> eth0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1500
>>       inet6 fe80::6e62:6dff:fed9:905  prefixlen 64  scopeid
>>
>> 0x20<link>
>>
>>       ether 6c:62:6d:d9:09:05  txqueuelen 1000  (Ethernet)
>>       RX packets 1498963  bytes 1952763267 (1.8 GiB)
>>       RX errors 0  dropped 0  overruns 0  frame 0
>>       TX packets 444309  bytes 51763060 (49.3 MiB)
>>       TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
>>
>> lo: flags=73<UP,LOOPBACK,RUNNING>  mtu 65536
>>       inet 127.0.0.1  netmask 255.0.0.0
>>       inet6 ::1  prefixlen 128  scopeid 0x10<host>
>>       loop  txqueuelen 0  (Local Loopback)
>>       RX packets 149945  bytes 166430929 (158.7 MiB)
>>       RX errors 0  dropped 0  overruns 0  frame 0
>>       TX packets 149945  bytes 166430929 (158.7 MiB)
>>       TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
>>
>> virbr0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1500
>>       inet 192.168.122.1  netmask 255.255.255.0  broadcast
>>
>> 192.168.122.255
>>
>> ether fe:00:c0:a8:7a:03  txqueuelen 0  (Ethernet)
>>       RX packets 18  bytes 3711 (3.6 KiB)
>>       RX errors 0  dropped 0  overruns 0  frame 0
>>       TX packets 69  bytes 3818 (3.7 KiB)
>>       TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
>>
>> vnet2: flags=4163<UP,BROADCAST,RUNNING,MULTICAST>  mtu 1500
>>       inet6 fe80::fc00:c0ff:fea8:7a03  prefixlen 64  scopeid
>>
>> 0x20<link>
>>
>>       ether fe:00:c0:a8:7a:03  txqueuelen 500  (Ethernet)
>>       RX packets 25  bytes 4805 (4.6 KiB)
>>       RX errors 0  dropped 0  overruns 0  frame 0
>>       TX packets 1303  bytes 68138 (66.5 KiB)
>>       TX errors 0  dropped 0 overruns 0  carrier 0  collisions 0
>>
>>
>> and routing looks like this:
>>
>> $ route -n
>> Kernel IP routing table
>> Destination     Gateway         Genmask         Flags Metric Ref
>>
>> Use
>>
>> Iface 0.0.0.0         46.4.99.33      0.0.0.0         UG    0      0
>>
>>
>> 0 br0 46.4.99.33      0.0.0.0         255.255.255.255 UH    0      0
>>
>>
>> 0 br0 169.254.0.0     0.0.0.0         255.255.0.0     U     1003   0
>>
>>
>> 0 br0 192.168.122.0   0.0.0.0         255.255.255.0   U     0      0
>>
>>
>> 0 virbr0
>>
>> Many thanks for any hint !
>>
>>
>> Kind regards ,
>>
>> Nikolaj Majorov
>> nikolaj at majorov.biz
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at lists.opennebula.org <mailto:Users at lists.opennebula.org
>> <Users at lists.opennebula.org>>
>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>
>> <http://lists.opennebula.org/listinfo.cgi/users-opennebula.org>
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at lists.opennebula.org <mailto:Users at lists.opennebula.org
>> <Users at lists.opennebula.org>>
>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>
>> <http://lists.opennebula.org/listinfo.cgi/users-opennebula.org>
>>
>>
>> ------------------------------------------------------------------------
>>
>> _______________________________________________
>> Users mailing list
>> Users at lists.opennebula.org
>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at lists.opennebula.org
>> http://lists.opennebula.org/listinfo.cgi/users-opennebula.org
>>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.opennebula.org/pipermail/users-opennebula.org/attachments/20141120/89ac8cd1/attachment-0001.htm>


More information about the Users mailing list