The Vagrant scripts deploy OpenStack with Open Virtual Network (OVN) using four nodes (five if you use the optional ovn-vtep node) to implement a minimal variant of the reference architecture:
ovn-northd services.networking-ovn ML2
driver, and the dashboard. In addition, the controller node is configured
as an NFS server to support instance live migration between the two
compute nodes.ovn-controller service for OVN, metadata agents for the
Networking service, and OVS services. In addition, the compute nodes are
configured as NFS clients to support instance live migration between them.During deployment, Vagrant creates three VirtualBox networks:
eth0 network interface.eth1 network
interface.eth2 network interface.The default configuration requires approximately 12 GB of RAM and supports
launching approximately four OpenStack instances using the m1.tiny
flavor. You can change the amount of resources for each VM in the
instances.yml file.
Follow the pre-requisites described in Vagrant prerequisites
Clone the networking-ovn repository locally and change to the
vagrant/sparse directory:
$ git clone https://git.openstack.org/openstack/networking-ovn.git
$ cd networking-ovn/vagrant/sparse
If necessary, adjust any configuration in the instances.yml file.
If you change any IP addresses or networks, avoid conflicts with the host.
For evaluating large MTUs, adjust the mtu option. You must also
change the MTU on the equivalent vboxnet interfaces on the host
to the same value after Vagrant creates them. For example:
# ip link set dev vboxnet0 mtu 9000
# ip link set dev vboxnet1 mtu 9000
Launch the VMs and grab some coffee:
$ vagrant up
After the process completes, you can use the vagrant status command
to determine the VM status:
$ vagrant status
Current machine states:
ovn-db running (virtualbox)
ovn-controller running (virtualbox)
ovn-vtep running (virtualbox)
ovn-compute1 running (virtualbox)
ovn-compute2 running (virtualbox)
You can access the VMs using the following commands:
$ vagrant ssh ovn-db
$ vagrant ssh ovn-controller
$ vagrant ssh ovn-vtep
$ vagrant ssh ovn-compute1
$ vagrant ssh ovn-compute2
rootaccount is vagrant. Since ovn-controller is set as the primary
in the Vagrantfile, the command vagrant ssh (without specifying
the name) will connect ssh to that virtual machine.
Access OpenStack services via command-line tools on the ovn-controller
node or via the dashboard from the host by pointing a web browser at the
IP address of the ovn-controller node.
admin and demo,both using password password.
After completing your tasks, you can destroy the VMs:
$ vagrant destroy
Except where otherwise noted, this document is licensed under Creative Commons Attribution 3.0 License. See all OpenStack Legal Documents.