ONES-GA 2.0
Release
  • ONES GA v2.0
    • Open Networking Enterprise Suite (ONES)
      • ONES Application
        • ONES Telemetry Collector(s) and Visibility
        • ONES Rule Engine
        • ONES Orchestration
        • ONES Supportability
        • ONES Security
    • What's new?
    • Getting Started
      • Supported Switch Platforms and NOS
      • Scalability
      • Subscription
      • Download ONES
      • ONES Installation
        • Installation Pre-requisites
        • Download ONES Package
        • Installing ONES Application
        • Installing ONES Agents
          • ONES Telemetry Agent Installation
          • ONES Orchestration Agent Installation
          • Agent Less Telemetry
            • Cumulus(NVUE API)
            • Arista EOS (OpenConfig)
        • Adding New Controller
      • VM Deployment
        • VMware ONES Deployment
        • KVM ONES Deployment
        • Upgrade VM
    • ONES Web GUI Administration
      • Login Page
      • Adding Devices
      • Monitor
      • Inventory
        • Devices
        • Configurations
      • Rule Engine
        • Slack Channel Integration
        • Zendesk Support Integration
        • Rules Type
          • Add Rules: Entity
          • Add Rules: Entity by Properties
        • Alerts
      • Analytics
      • Settings
      • Integrations
        • Slack Channel Integration
        • Zendesk Support Integration
      • Accounts
      • ONES Orchestration
        • VXLAN-Symmetric
          • VXLAN-Symmetric-SAG-no-mclag-vrf
          • VXLAN-Symmetric-SAG-mclag-vrf
          • VXLAN-Symmetric-SVI-no-mclag-vrf
          • VXLAN-Symmetric-SVI-mclag-vrf
        • VXLAN-SVI
          • VXLAN-SVI-no-mclag-IPv6-ebgp-IPv6-SVI
          • VXLAN-SVI-no-mclag-ibgp-IPv4-Underlay
          • VXLAN-SVI-no-mclag-ebgp-IPv6-SAG
          • VXLAN-SVI-no-mclag-ebgp-IPv4-Underlay
          • VXLAN-svi-no-mclag-ebgp-BGPU-Underlay
        • VXLAN-SAG
          • VXLAN-SAG-no-mclag-ibgp-no-host
          • VXLAN-SAG-no-mclag-ibgp-BGPU-Underlay
          • VXLAN-SAG-no-mclag-ibgp-BGPU-Underlay-HostPO
          • VXLAN-SAG-no-mclag-ebgp-no-host
        • VXLAN-MGLAG
          • VXLAN-MCLAG-ibgp-sag
          • VXLAN-MCLAG-ebgp-svi.yaml
          • VXLAN-MCLAG-ipv6-ibgp-sag
          • VXLAN-MCLAG-ebgp-BGPU-IPv6SAG
          • VXLAN-MCLAG-ibgp-BGPU-IPv6svi
          • VXLAN-MCLAG-ipv4-ibgp-svi
          • VXLAN-MCLAG-ipv6-ebgp-sag
        • VXLAN-Asymmetric
          • VXLAN-Asymmetric-SAG-no-mclag
          • VXLAN-Asymmetric-SAG-mclag
          • VXLAN-Asymmetric-SVI-no-mclag
          • VXLAN-Asymmetric-SVI-mclag
        • MH-VXLAN
          • MH-VXLAN-ibgp-sag
          • MH-VXLAN-ipv6-ibgp-sag
          • MH-VXLAN-Asymmetric-SAG
          • MH-VXLAN-ebgp-BGPU-IPv6SAG
          • MH-VXLAN-ipv6-ebgp-sag
          • MH-VXLAN-Symmetric-2-SAG-vrf
        • MCLAG
          • MCLAG-IPv4-SVI-AccessHosts
          • MCLAG-ibgp-BGPU-Underlay-IPv4SVI-AccessHosts
          • MCLAG-ibgp-IPv6-Underlay-IPv6SVI-AccessHosts
          • MCLAG-IPv4-Underlay-IPv6SVI-AccessHosts
          • MCLAG-IPv6-Underlay-IPv6SVI-AccessHosts
        • L3-MCLAG
          • L3-MCLAG
          • L3-MCLAG-bgpU-combinedLink
          • L3-MCLAG-bgp-combinedLink
        • L2-LS
          • L2LS-EC-L2
          • L2LS-EC-L3-ipv4
          • L2LS-EC-L3-ipv6
        • iBGP
          • i-BGP-IPv6-CLOS-IPv6-host
          • i-BGP-IPv6-CLOS-IPv4-svi
          • i-BGP-IPv4-CLOS-L2PO-host
          • i-BGP-IPv4-CLOS-L2-host
          • i-BGP-IPv4-CLOS-IPv6-svi
          • i-BGP-IPv4-CLOS-IPv4-PO-host
          • i-BGP-BGP-U-CLOS-IPv4-svi
          • i-BGP-BGP-U-CLOS-IPv4-host
        • e-BGP
          • e-BGP-IPv6-CLOS-IPv6-host
          • e-BGP-IPv6-CLOS-IPv4-svi
          • e-BGP-IPv4-CLOS-L2-host
          • e-BGP-IPv4-CLOS-IPv6-svi
          • e-BGP-BGP-U-CLOS-IPv6-svi
          • e-BGP-BGP-U-CLOS-IPv6-host
        • BGP-PO
          • BGP-PO-SLFOW-LeafOnly-EC
          • BGP-PO-SLFOW-LeafOnly-EC-incr1
          • BGP-PO-MCLAG-SLFOW-LeafOnly-EC
        • LeafOnly-EC
        • IPCLOS-1-IPv4-SVI-AccessHosts
        • DHCP-IPCLOS-IPv4-SVI-AccessHosts
    • Common Issues and Troubleshooting Steps
      • ONES Installation
        • Storage issue
      • ONES Uninstallation
        • ONES-Application
        • Telemetry Agent
        • Orchestration Agent
      • Permission Issues
        • Application Installation
      • gNMI Troubleshooting on non-SONiC Switches
    • How to contact Aviz Networks Support?
    • Backup and Recovery
Powered by GitBook
On this page
  • QCOW Deployment
  • Download Qcow2 Package
  • Create the VM using GUI App virt-manager
  • Create the VM using QEMU (XML configuration)
  • Credentials to access ONES QCOW2 VM
  • OVA Integration
Export as PDF
  1. ONES GA v2.0
  2. Getting Started
  3. VM Deployment

KVM ONES Deployment

PreviousVMware ONES DeploymentNextUpgrade VM

QCOW Deployment

Download Qcow2 Package

  • Work with Aviz Sales/Support contact to create an account on Aviz Networks Support Portal

  • Login to with your account credentials

  • Click on the Downloads section, under ONES, click to download ONES Release 2.0

  • Copy ONES Release 2.0 package (qcow2) to KVM Hypervisor Server

Create the VM using GUI App virt-manager

If your host server has Ubuntu Desktop and virt-manager installed you can use it to deploy the VM. Make sure you can start the Virtual Machine Manager and that it connects successfully to the local hypervisor.

Creating a VM with virt-manager is very straightforward, Use the following steps to deploy the ONES-Application

  • File -> New Virtual Machine -> Import existing disk image -> Forward

Now the ONES Application is ready to use

Create the VM using QEMU (XML configuration)

  • Create an XML configuration file from the following template using vi

vi ones.xml
<domain type='kvm' xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'>
  <name>ONES_VM01</name>
  <memory unit='KiB'>4194304</memory>
  <currentMemory unit='KiB'>4194304</currentMemory>
  <vcpu placement='static'>4</vcpu>
  <resource>
    <partition>/machine</partition>
  </resource>
  <os>
    <type arch='x86_64' machine='pc-i440fx-1.5'>hvm</type>
    <boot dev='hd'/>
  </os>
  <features>
    <acpi/>
    <apic/>
  </features>
  <clock offset='utc'/>
  <on_poweroff>destroy</on_poweroff>
  <on_reboot>restart</on_reboot>
  <on_crash>restart</on_crash>
  <devices>
    <emulator>/usr/bin/qemu-system-x86_64</emulator>
    <disk type='file' device='disk'>
      <driver name='qemu' type='qcow2' cache='writeback'/>
      <source file='/home/oper/taas_vm/ones.qcow2' />
      <target bus='virtio' dev='vda'/>
    </disk>
    <serial type='pty'>
      <source path='/dev/pts/3'/>
      <target port='0'/>
    </serial>
    <!-- Management interface eth0 -->
    <interface type='network'>
	<model type='e1000' />
        <source network='br0'/>
        <address type='pci' domain='0x0000' bus='0x00' slot='0x00' function='0x0'/>
    </interface>
   <controller type='usb' index='0'/>
    <memballoon model='virtio'>
      <alias name='balloon0'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x04' function='0x0'/>
    </memballoon>
  </devices>
</domain>
User Inputs in XML File
Line #2   The Name of the VM
Line #3   The amount of System Memory for the VM
Line #4   The amount of System Memory for the VM
Line #5   The number of vCPU Core for the VM
Line #25  The Path to the qcow2 VM image file
Line #35  The name of the Linux bridge on the host machine
  • Create a Linux bridge configuration file (bridged-network.xml) for libvirt from the following template

vi bridged-network.xml
<network>
    <name>br0</name>
    <forward mode="bridge" />
    <bridge name="br0" />
</network>
User Inputs in XML File
Line #4 the name of the Linux bridge on the host machine
  • Define the Linux bridge for the VM

#Execute the below command to attach the VM to the Linux Bridge 
sonic@sonic-39:~$ virsh net-define bridged-network.xml
sonic@sonic-39:~$ virsh net-start br0
sonic@sonic-39:~$ virsh net-autostart br0
sonic@sonic-39:~$ virsh net-list
 Name                 State      Autostart     Persistent
----------------------------------------------------------
 br0                  active     yes           yes

sonic@sonic-39:~$ 
  • Start the VM

virsh create <VM XML configuration file>

#sonic@sonic-39:~$ virsh create ones.xml 
#Domain ONES_VM01 created from ones.xml
#sonic@sonic-39:~$ 

If you see a permission error run the virsh command with sudo may fix the issue

  • Check the VM status

sonic@sonic-39:~$ virsh list
 Id    Name                           State
----------------------------------------------------
 8     ONES_VM01                      running
sonic@sonic-39:~$ 

Credentials to access ONES QCOW2 VM

Username: aviz
Password: Aviz@123

OVA Integration

https://support.aviznetworks.com