In my I case I replicated the environment containing 3 VMs where MAAS was installed on the first VM (kvm01) and other two VMs (kvm02, kvm03) were provisioned by the MAAS. I also defined few networks with IP ranges: deploy_network 192.168.25.0/24, control_network 192.168.26.0/24, tenant_network 192.168.27.0/24.
Here is the network diagram of the lab:
You can see the commands I used and descriptive video...
It's better to see the video with lab description:
Here are the commands I used:
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# Install Vagrant and virsh command | |
dnf install -y libvirt-client vagrant-hostmanager vagrant-libvirt | |
# Allow to manage VMs via libvirt remotely (using TCP connection) | |
cat >> /etc/libvirt/libvirtd.conf << EOF | |
listen_tls = 0 | |
listen_tcp = 1 | |
listen_addr = "0.0.0.0" | |
auth_tcp = "none" | |
EOF | |
echo 'LIBVIRTD_ARGS="--listen --config /etc/libvirt/libvirtd.conf"' >> /etc/sysconfig/libvirtd | |
service libvirtd restart | |
# Generate ssh key used for accessing the VMs | |
ssh-keygen -P "" -f /root/.ssh/id_rsa -C "admin@example.com" | |
# Check the VMs | |
VIRSH_VMS=$(virsh list | awk '/_kvm/ { print $2 }') | |
for VIRSH_VM in $VIRSH_VMS; do echo "*** $VIRSH_VM"; virsh dumpxml $VIRSH_VM | grep 'mac address' | sort; done | |
# Check the subnets | |
virsh net-list --all | grep network | |
VIRSH_NETWORKS=$(virsh net-list | awk '/network|vagrant/ { print $1 }') | |
for VIRSH_NETWORK in $VIRSH_NETWORKS; do echo "*** $VIRSH_NETWORK"; virsh net-dumpxml $VIRSH_NETWORK; done | |
# Create Vagrantfile | |
mkdir /var/tmp/test | |
cat > /var/tmp/test/Vagrantfile << \EOF | |
box_image = "peru/my_ubuntu-16.04-server-amd64" | |
node_count = 3 | |
ssh_pub_key = File.readlines("#{Dir.home}/.ssh/id_rsa.pub").first.strip | |
Vagrant.configure(2) do |config| | |
config.hostmanager.enabled = true | |
config.hostmanager.manage_host = true | |
config.hostmanager.manage_guest = false | |
config.vm.synced_folder ".", "/vagrant", :disabled => true | |
config.vm.provider :libvirt do |domain| | |
domain.cpus = 2 | |
domain.cpu_mode = "host-passthrough" | |
domain.memory = 1536 | |
domain.machine_virtual_size = 150 | |
domain.disk_bus = "virtio" | |
domain.nic_model_type = "virtio" | |
domain.video_type = "qxl" | |
domain.graphics_type = "spice" | |
domain.management_network_address = "192.168.100.0/24" | |
end | |
(1..node_count).each do |i| | |
config.vm.define "kvm0#{i}" do |config| | |
config.vm.hostname = "kvm0#{i}" | |
config.hostmanager.ignore_private_ip = true | |
if i == 1 | |
config.vm.box = box_image | |
config.vm.network :forwarded_port, guest: 80, host: 80, host_ip: "*" | |
end | |
config.vm.provider :libvirt do |domain| | |
domain.storage :file, :size => '100G', :type => 'qcow2' | |
if i == 1 | |
domain.memory = 4096 | |
else | |
boot_network = {'network' => 'deploy_network'} | |
domain.boot boot_network | |
#domain.loader = "/usr/share/edk2/ovmf/OVMF_CODE.fd" | |
end | |
end | |
config.vm.network :private_network, ip: "192.168.25.#{i+10}", auto_config: false, libvirt__network_name: "deploy_network", libvirt__dhcp_enabled: false, libvirt__forward_mode: "nat", :mac => "52:54:00:00:25:#{i+10}" | |
config.vm.network :private_network, ip: "192.168.26.#{i+10}", auto_config: false, libvirt__network_name: "control_network", libvirt__dhcp_enabled: false, libvirt__forward_mode: "nat", :mac => "52:54:00:00:26:#{i+10}" | |
config.vm.network :private_network, ip: "192.168.26.#{i+20}", auto_config: false, libvirt__network_name: "control_network", libvirt__dhcp_enabled: false, libvirt__forward_mode: "nat", :mac => "52:54:00:00:26:#{i+20}" | |
config.vm.network :private_network, ip: "192.168.27.#{i+10}", auto_config: false, libvirt__network_name: "tenant_network", libvirt__dhcp_enabled: false, libvirt__forward_mode: "nat", :mac => "52:54:00:00:27:#{i+10}" | |
config.vm.network :private_network, ip: "192.168.27.#{i+20}", auto_config: false, libvirt__network_name: "tenant_network", libvirt__dhcp_enabled: false, libvirt__forward_mode: "nat", :mac => "52:54:00:00:27:#{i+20}" | |
end | |
end | |
config.vm.provision 'shell', inline: "install -m 0700 -d /root/.ssh/; echo #{ssh_pub_key} >> /root/.ssh/authorized_keys; chmod 0600 /root/.ssh/authorized_keys" | |
config.vm.provision 'shell', inline: "echo #{ssh_pub_key} >> /home/vagrant/.ssh/authorized_keys", privileged: false | |
end | |
# Deploy network subnet: 192.168.25.0/24, GW: 192.168.125.1 | |
# Control network subnet: 192.168.26.0/24, GW: 192.168.126.1 | |
# Tenant network subnet: 192.168.27.0/24, GW: 192.168.127.1 | |
EOF | |
cd /var/tmp/test | |
# Start all VMs + create networking | |
vagrant up kvm01 | |
vagrant up | |
# Check the VMs - all should be running | |
VIRSH_VMS=$(virsh list | awk '/_kvm/ { print $2 }') | |
for VIRSH_VM in $VIRSH_VMS; do echo "*** $VIRSH_VM"; virsh dumpxml $VIRSH_VM | grep 'mac address' | sort; done | |
# Check the subnets | |
virsh net-list --all | grep network | |
VIRSH_NETWORKS=$(virsh net-list | awk '/network|vagrant/ { print $1 }') | |
for VIRSH_NETWORK in $VIRSH_NETWORKS; do echo "*** $VIRSH_NETWORK"; virsh net-dumpxml $VIRSH_NETWORK; done | |
# Check the DHCP lease file - there should be only kvm01 | |
DEVICE=$(virsh net-dumpxml vagrant-libvirt | awk -F\' '/bridge/ { print $2 }') | |
cat /var/lib/libvirt/dnsmasq/${DEVICE}.status | |
# SSH to the first node where MAAS will be installed | |
ssh -o StrictHostKeyChecking=no kvm01 | |
# Check IPs / NICS | |
ip a | |
# Configure the NICs | |
apt update; apt install -y ifenslave | |
cat >> /etc/network/interfaces << EOF | |
auto eth1 | |
iface eth1 inet static | |
address 192.168.25.11 | |
netmask 255.255.255.0 | |
auto eth2 | |
iface eth2 inet manual | |
bond-master bond0 | |
auto eth3 | |
iface eth3 inet manual | |
bond-master bond0 | |
auto bond0 | |
iface bond0 inet static | |
address 192.168.26.11 | |
netmask 255.255.255.0 | |
bond-slaves eth2 eth3 | |
bond_mode active-backup | |
auto eth4 | |
iface eth4 inet manual | |
bond-master bond1 | |
auto eth5 | |
iface eth5 inet manual | |
bond-master bond1 | |
auto bond1 | |
iface bond1 inet static | |
address 192.168.27.11 | |
netmask 255.255.255.0 | |
bond-slaves eth4 eth5 | |
bond_mode active-backup | |
EOF | |
service networking restart | |
# Install MAAS | |
apt install -y jq libvirt-bin maas | |
# Configure Region controller to point to the right MAAS IP address | |
maas-region local_config_set --maas-url http://192.168.25.11:5240/MAAS | |
systemctl restart maas-regiond | |
# Register a rack controller with the MAAS | |
maas-rack register --url http://192.168.25.11:5240/MAAS --secret `cat /var/lib/maas/secret` | |
# Create administrator (MAAS "superuser") | |
maas createadmin --username=admin --email=admin@example.com --password admin123 | |
# Export API key that was generated when the MAAS account was created | |
maas-region apikey --username=admin > /root/api_key | |
# Create a short script which will help you to login to MAAS quickly | |
cat > /root/maas-login.sh << \EOF | |
#!/bin/sh | |
# Change these 3 values as required | |
PROFILE=admin | |
API_KEY_FILE=/root/api_key | |
API_SERVER=localhost | |
MAAS_URL=http://$API_SERVER/MAAS/api/2.0 | |
maas login $PROFILE $MAAS_URL - < $API_KEY_FILE | |
EOF | |
chmod a+x /root/maas-login.sh | |
# Login to MAAS | |
/root/maas-login.sh | |
# Generate SSH key which will be imported in the next command | |
ssh-keygen -P "" -f /root/.ssh/id_rsa -C "admin@example.com" | |
# Import the admin SSH key | |
maas admin sshkeys create "key=`cat /root/.ssh/id_rsa.pub`" | |
# Turn OFF all VMs except the first one running MAAS | |
# This will also test if the libvirtd daemon is properly configured allowing MAAS to manage the VMs | |
virsh -c qemu+tcp://192.168.25.1/system destroy test_kvm02 | |
virsh -c qemu+tcp://192.168.25.1/system destroy test_kvm03 | |
virsh -c qemu+tcp://192.168.25.1/system list --all | |
SUBNET_CIDR="192.168.25.0/24" | |
SUBNET_PREFIX=$(echo $SUBNET_CIDR | sed -r 's/(([0-9]{1,3}\.){2}.[0-9]{1,3}).*/\1/') | |
PRIMARY_RACK_CONTROLLER=$(maas admin rack-controllers read | jq -r '.[0].system_id') | |
VLAN_FABRIC_ID=`maas admin subnet read $SUBNET_CIDR | jq '.vlan.fabric_id'` | |
VLAN_VID=`maas admin subnets read | jq -r ".[] | select(.cidr==\"$SUBNET_CIDR\")".vlan.vid` | |
# Add default gateway for 192.168.25.0/24 | |
maas admin subnet update cidr:${SUBNET_CIDR} gateway_ip=${SUBNET_PREFIX}.1 | |
# Enable DHCP on the subnet 192.168.25.0/24 and reserve dynamic IP range (192.168.25.200 - 192.168.25.250) - used for commisioning | |
maas admin ipranges create type=dynamic start_ip=${SUBNET_PREFIX}.200 end_ip=${SUBNET_PREFIX}.250 comment='This is a reserved dynamic range' | |
maas admin vlan update $VLAN_FABRIC_ID $VLAN_VID dhcp_on=True primary_rack=$PRIMARY_RACK_CONTROLLER | |
# Define node by specifying the libvirt VM management and start comissioning | |
for INDEX in {2..3}; do | |
MAC="52:54:00:00:25:1${INDEX}" | |
maas admin machines create power_parameters_power_address=qemu+tcp://192.168.25.1/system hostname=kvm0${INDEX} power_type=virsh power_parameters_power_id=test_kvm0${INDEX} architecture=amd64/generic mac_addresses=$MAC | |
done | |
# All the machines should be in commisioning state right now | |
virsh -c qemu+tcp://192.168.25.1/system list --all | |
# Set static IPs, create bonds and deploy the Ubuntu Xenial | |
for INDEX in {2..3}; do | |
MAC_1="52:54:00:00:25:1${INDEX}" | |
MAC_2="52:54:00:00:26:1${INDEX}" | |
MAC_3="52:54:00:00:26:2${INDEX}" | |
MAC_4="52:54:00:00:27:1${INDEX}" | |
MAC_5="52:54:00:00:27:2${INDEX}" | |
IP_1="192.168.25.1${INDEX}" | |
SUBNET_CIDR_1="192.168.25.0/24" | |
IP_2="192.168.26.1${INDEX}" | |
SUBNET_CIDR_2="192.168.26.0/24" | |
IP_3="192.168.27.1${INDEX}" | |
SUBNET_CIDR_3="192.168.27.0/24" | |
maas admin nodes read mac_address=$MAC_1 > /tmp/maas_nodes_read | |
SYSTEM_ID=`jq -r ".[].system_id" /tmp/maas_nodes_read` | |
INTERFACE_ID_1=`jq -r ".[].interface_set[] | select(.mac_address==\"$MAC_1\").id" /tmp/maas_nodes_read` | |
INTERFACE_ID_2=`jq -r ".[].interface_set[] | select(.mac_address==\"$MAC_2\").id" /tmp/maas_nodes_read` | |
INTERFACE_ID_3=`jq -r ".[].interface_set[] | select(.mac_address==\"$MAC_3\").id" /tmp/maas_nodes_read` | |
INTERFACE_ID_4=`jq -r ".[].interface_set[] | select(.mac_address==\"$MAC_4\").id" /tmp/maas_nodes_read` | |
INTERFACE_ID_5=`jq -r ".[].interface_set[] | select(.mac_address==\"$MAC_5\").id" /tmp/maas_nodes_read` | |
# Remove the "Auto assign" IP address and set static instead | |
# https://askubuntu.com/questions/942412/how-do-you-statically-asign-an-ip-to-a-commissioned-machine-in-maas | |
OLD_LINK_ID=`jq ".[].interface_set[] | select(.id==$INTERFACE_ID_1).links[].id" /tmp/maas_nodes_read` | |
maas admin interface unlink-subnet $SYSTEM_ID $INTERFACE_ID_1 id=$OLD_LINK_ID | |
maas admin interface link-subnet $SYSTEM_ID $INTERFACE_ID_1 mode=STATIC subnet="cidr:$SUBNET_CIDR_1" ip_address=$IP_1 default_gateway=true | |
# Create bond interfaces | |
maas admin interfaces create-bond $SYSTEM_ID name=bond0 parents=$INTERFACE_ID_2 mac_address=$MAC_2 parents=$INTERFACE_ID_3 bond_mode=active-backup | |
maas admin interfaces create-bond $SYSTEM_ID name=bond1 parents=$INTERFACE_ID_4 mac_address=$MAC_4 parents=$INTERFACE_ID_5 bond_mode=active-backup | |
# Regenerate /tmp/maas_nodes_read - now with the bond interfaces | |
maas admin nodes read mac_address=$MAC_1 > /tmp/maas_nodes_read | |
BOND0_ID=`jq -r ".[].interface_set[] | select(.name==\"bond0\").id" /tmp/maas_nodes_read` | |
BOND1_ID=`jq -r ".[].interface_set[] | select(.name==\"bond1\").id" /tmp/maas_nodes_read` | |
# Assign proper fabric and IP to the bond0 | |
FABRIC_VLAN_ID=`maas admin subnets read | jq ".[] | select(.cidr==\"$SUBNET_CIDR_2\").vlan.id"` | |
maas admin interface update $SYSTEM_ID $BOND0_ID vlan=$FABRIC_VLAN_ID | |
maas admin interface link-subnet $SYSTEM_ID $BOND0_ID mode=STATIC subnet="cidr:$SUBNET_CIDR_2" ip_address=$IP_2 | |
# Assign proper fabric and IP to the bond1 | |
FABRIC_VLAN_ID=`maas admin subnets read | jq ".[] | select(.cidr==\"$SUBNET_CIDR_3\").vlan.id"` | |
maas admin interface update $SYSTEM_ID $BOND1_ID vlan=$FABRIC_VLAN_ID | |
maas admin interface link-subnet $SYSTEM_ID $BOND1_ID mode=STATIC subnet="cidr:$SUBNET_CIDR_3" ip_address=$IP_3 | |
# Deploy server | |
maas admin machine deploy $SYSTEM_ID | |
done | |
# All machines should be installed + deployed... | |
virsh -c qemu+tcp://192.168.25.1/system list --all | |
ssh ubuntu@192.168.25.12 -o StrictHostKeyChecking=no | |
ssh ubuntu@192.168.25.13 -o StrictHostKeyChecking=no |
Asciinema if needed:
I hope it's helpful for somebody...
Enjoy ;-)
Very neat - and yes, this was helpful for me - thank you for sharing
ReplyDeleteis there something special in your box ? because i get this error:
ReplyDeleteBox 'peru/my_ubuntu-16.04-server-amd64' could not be found....
Nothing special there...
ReplyDeleteYou can use any generic "Ubuntu image" or peru/my_ubuntu-18.04-server-amd64.
Excellent work. Thank you sir
ReplyDeleteAmazing, super useful
ReplyDeletehi ravi, i checked, its not peru/my_ubuntu-16.04..image, but peru/ubuntu-16.04...image
ReplyDeleteyou can try