[one-users] FlatDHCP networking problem
Steven Timm
timm at fnal.gov
Thu Apr 18 08:05:47 PDT 2013
You're using the wrong cloud.. You should be using OpenNebula
and not OpenStack and then you would not be having these OpenStack
problems. Probably better to get help on OpenStack on the OpenSTack list.
Steve Timm
On Thu, 18 Apr 2013, Javier Alvarez wrote:
> Hello all,
>
> Here it is my situation:
>
> I am trying to install Essex on a small cluster (3 nodes) running Debian.
> There is a front-end node that has a public IP and then there are 2 compute
> nodes in a LAN. I cannot run nova-network on the front-end node because it
> is overwritting the iptables there and some other services start to
> misbehave, so I am trying a multi-host solution with nova-network running in
> each compute node.
>
> The nova.conf I'm using in both compute nodes is the following:
>
> [DEFAULT]
> logdir=/var/log/nova
> state_path=/var/lib/nova
> lock_path=/var/lock/nova
> root_helper=sudo nova-rootwrap
> auth_strategy=keystone
> iscsi_helper=tgtadm
> sql_connection=mysql://nova-common:password@172.16.8.1/nova
> connection_type=libvirt
> libvirt_type=kvm
> my_ip=172.16.8.22
> rabbit_host=172.16.8.1
> glance_host=172.16.8.1
> image_service=nova.image.glance.GlanceImageService
> network_manager=nova.network.manager.FlatDHCPManager
> fixed_range=192.168.100.0/24
> flat_interface=eth1
> public_interface=eth0
> flat_network_bridge=br100
> flat_network_dhcp_start=192.168.100.2
> network_size=256
> dhcpbridge_flagfile=/etc/nova/nova.conf
> dhcpbridge=/usr/bin/nova-dhcpbridge
> multi_host=True
> send_arp_for_ha=true
>
> I have created a network with:
>
> nova-manage network create private --fixed_range_v4=192.168.100.0/24
> --multi_host=T --bridge_interface=br100
>
> And I have set up eth1 with no IP and running in promisc mode. When I launch
> an instance, ifconfig outputs the following:
>
>
> br100 Link encap:Ethernet HWaddr 68:b5:99:c2:7b:a7
> inet addr:192.168.100.3 Bcast:192.168.100.255 Mask:255.255.255.0
> inet6 addr: fe80::7033:eeff:fe29:81ae/64 Scope:Link
> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
> RX packets:0 errors:0 dropped:0 overruns:0 frame:0
> TX packets:1 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:0
> RX bytes:0 (0.0 B) TX bytes:90 (90.0 B)
>
> eth0 Link encap:Ethernet HWaddr 68:b5:99:c2:7b:a6
> inet addr:172.16.8.22 Bcast:172.16.8.255 Mask:255.255.255.0
> inet6 addr: fe80::6ab5:99ff:fec2:7ba6/64 Scope:Link
> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
> RX packets:4432580 errors:0 dropped:0 overruns:0 frame:0
> TX packets:4484811 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:1000
> RX bytes:457880509 (436.6 MiB) TX bytes:398588034 (380.1 MiB)
> Memory:fe860000-fe880000
>
> eth1 Link encap:Ethernet HWaddr 68:b5:99:c2:7b:a7
> UP BROADCAST PROMISC MULTICAST MTU:1500 Metric:1
> RX packets:0 errors:0 dropped:0 overruns:0 frame:0
> TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:1000
> RX bytes:0 (0.0 B) TX bytes:0 (0.0 B)
> Memory:fe8e0000-fe900000
>
> lo Link encap:Local Loopback
> inet addr:127.0.0.1 Mask:255.0.0.0
> inet6 addr: ::1/128 Scope:Host
> UP LOOPBACK RUNNING MTU:16436 Metric:1
> RX packets:52577 errors:0 dropped:0 overruns:0 frame:0
> TX packets:52577 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:0
> RX bytes:2737820 (2.6 MiB) TX bytes:2737820 (2.6 MiB)
>
> vnet0 Link encap:Ethernet HWaddr fe:16:3e:2d:40:3b
> inet6 addr: fe80::fc16:3eff:fe2d:403b/64 Scope:Link
> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
> RX packets:0 errors:0 dropped:0 overruns:0 frame:0
> TX packets:5 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:500
> RX bytes:0 (0.0 B) TX bytes:370 (370.0 B)
>
> And brctl show:
>
> bridge name bridge id STP enabled interfaces
> br100 8000.68b599c27ba7 no eth1
>
?? ? vnet0
>
> Which looks fine to me. However, the VM log shows that it is unable to get
> an IP through DHCP (despite dnsmasq is running):
>
> Starting network...
> udhcpc (v1.18.5) started
> Sending discover...
> Sending discover...
> Sending discover...
> No lease, failing
> WARN: /etc/rc3.d/S40-network failed
>
> What am I doing wrong? Any help would be very much appreciated.
>
> Thanks in advance,
>
> Javier
>
> --
> Javier Álvarez Cid-Fuentes
> Grid Computing and Clusters Group
> Barcelona Supercomputing Center (BSC-CNS)
> Tel. (+34) 93 413 72 46
>
>
> WARNING / LEGAL TEXT: This message is intended only for the use of the
> individual or entity to which it is addressed and may contain information
> which is privileged, confidential, proprietary, or exempt from disclosure
> under applicable law. If you are not the intended recipient or the person
> responsible for delivering the message to the intended recipient, you are
> strictly prohibited from disclosing, distributing, copying, or in any way
> using this message. If you have received this communication in error, please
> notify the sender and destroy and delete any copies you may have received.
>
> http://www.bsc.es/disclaimer
>
>
------------------------------------------------------------------
Steven C. Timm, Ph.D (630) 840-8525
timm at fnal.gov http://home.fnal.gov/~timm/
Fermilab Computing Division, Scientific Computing Facilities,
Grid Facilities Department, FermiGrid Services Group, Group Leader.
Lead of FermiCloud project.
More information about the Users
mailing list