Difference between revisions of "FAQ"

From Proxmox VE
Jump to: navigation, search
m (Troubleshooting)
 
(22 intermediate revisions by 8 users not shown)
Line 1: Line 1:
tbd, if possible, detailed answers should linked to internal wiki pages
+
<!--PVE_IMPORT_START_MARKER-->
 
+
<!-- Do not edit - this is autogenerated content -->
==General==
+
{{#pvedocs:pve-faq-plain.html}}
===What is a container, CT, VE, Virtual Private Server, VPS?===
+
[[Category:Reference Documentation]]
:See [[Container and Full Virtualization]]
+
<pvehide>
 
+
New FAQs are appended to the bottom of this section.
===What is a KVM guest (KVM VM)?===
+
What distribution is Proxmox VE based on?
:A KVM guest or KVM VM is a guest system running virtualized under Proxmox VE with KVM.
+
Proxmox VE is based on Debian GNU/Linux
 
+
What license does the Proxmox VE project use?
===What is a Virtual Appliance?===
+
Proxmox VE code is licensed under the GNU Affero General Public License,
:See [[Overview]]
+
version 3.
 
+
Will Proxmox VE run on a 32bit processor?
===What distribution is Proxmox VE based on?===
+
Proxmox VE works only on 64-bit CPUs (AMD or Intel). There is no plan
:Proxmox VE is based on [http://www.debian.org Debian GNU/Linux], Proxmox VE Kernel is based on RHEL6 Kernel with OpenVZ patches
+
for 32-bit for the platform.
 
+
VMs and Containers can be both 32-bit and/or 64-bit.
==Installation and upgrade==
+
Does my CPU support virtualization?
 
+
To check if your CPU is virtualization compatible, check for the vmx
===Where can I find installation instructions?===
+
or svm tag in this command output:
:See [[Installation]]
+
egrep '(vmx|svm)' /proc/cpuinfo
 
+
Supported Intel CPUs
===Proxmox VE command line tools===
+
64-bit processors with
:See [[Command line tools]]
+
Intel
 
+
Virtualization Technology (Intel VT-x) support. (List of processors with Intel VT and 64-bit)
==Hardware==
+
Supported AMD CPUs
===CPU===
+
64-bit processors with
====Will Proxmox VE run on a 32bit processor?====
+
AMD
Proxmox VE works only on 64-bit CPU´s (AMD or Intel). There is no plan for 32-bit for the platform.  
+
Virtualization Technology (AMD-V) support.
 
+
What is a container, CT, VE, Virtual Private Server, VPS?
There are, however,  unofficial (and unsupported) instructions for manually installing Proxmox on 32-bit systems:
+
Operating-system-level virtualization is a server-virtualization
* Proxmox 2.0 on Squeeze [[Install Proxmox VE on Debian Squeeze on 32-Bit Processor]]
+
method where the kernel of an operating system allows for multiple
* Proxmox 1.4 on Lenny [[Install Proxmox VE on Debian Lenny on 32-Bit Processor]]
+
isolated user-space instances, instead of just one. We call such
 
+
instances containers. As containers use the host&#8217;s kernel they are
===Supported CPU chips===
+
limited to Linux guests.
To check if your CPU is virtualization compatible, check for the "vmx" or "svm" tag in this command output:
+
What is a QEMU/KVM guest (or VM)?
egrep '(vmx|svm)' /proc/cpuinfo
+
A QEMU/KVM guest (or VM) is a guest system running virtualized under
 
+
Proxmox VE using QEMU and the Linux KVM kernel module.
====Intel====
+
What is QEMU?
64-bit processors with [http://en.wikipedia.org/wiki/Virtualization_Technology#Intel_virtualization_.28VT-x.29 Intel Virtualization Technology (Intel VT-x)] support
+
QEMU is a generic and open source machine emulator and
 
+
virtualizer. QEMU uses the Linux KVM kernel module to achieve near
[http://ark.intel.com/search/advanced/?s=t&VTX=true&InstructionSet=64-bit List of processors with Intel VT and 64-bit]
+
native performance by executing the guest code directly on the host
 
+
CPU.
====AMD====
+
It is not limited to Linux guests but allows arbitrary operating systems
64-bit processors with [http://en.wikipedia.org/wiki/Virtualization_Technology#AMD_virtualization_.28AMD-V.29 AMD Virtualization Technology (AMD-V)] support
+
to run.
 
+
How long will my Proxmox VE version be supported?
==Networking==
+
Proxmox VE versions are supported at least as long as the corresponding
 
+
Debian Version is
===How do I configure bridged networking in an OpenVZ Ubuntu/Debian container?===
+
oldstable. Proxmox VE uses a
<ol>
+
rolling release model and using the latest stable version is always
<li>In the web gui under Virtual Machine configuration go to the network tab.
+
recommended.
<li>Remove the ip address for venet and save. (Bridged Ethernet Devices will appear)
+
Proxmox VE Version
<li>SSH into your host system and enter the container you want set bridged networking for:
+
  Debian Version
# vzctl enter <VMID>
+
  First Release
<li>edit  /etc/network/interfaces using using the following format and save. (replace with settings for your network)
+
  Debian EOL
<pre>
+
  Proxmox EOL
  auto lo
+
Proxmox VE 6.x
  iface lo inet loopback
+
Debian 10 (Buster)
   
+
2019-07
  auto eth0
+
tba
iface eth0 inet static
+
tba
        address 10.0.0.17
+
Proxmox VE 5.x
        netmask 255.255.255.0
+
Debian 9 (Stretch)
        network 10.0.0.0
+
2017-07
        broadcast 10.0.0.255
+
2020-07
        gateway 10.0.0.10
+
2020-07
</pre>
+
Proxmox VE 4.x
<li>Shutdown the container.  
+
Debian 8 (Jessie)
<li>Go back to web gui and under "Bridged Ethernet Devices"  configure eth0 to vmbr0  and save. (a mac address will be automatically assigned)
+
2015-10
<li>Start the container.
+
2018-06
</ol>
+
2018-06
Finally check that networking is working by entering the guest and viewing the results of ifconfig
+
Proxmox VE 3.x
 
+
Debian 7 (Wheezy)
*In a Centos/RHEL container, check the gateway device is set correctly.
+
2013-05
edit /etc/sysconfig/network
+
2016-04
<pre>
+
2017-02
NETWORKING="yes"
+
Proxmox VE 2.x
#GATEWAYDEV="venet0"              # comment this and add line below
+
Debian 6 (Squeeze)
GATEWAYDEV="eth0"
+
2012-04
HOSTNAME="hostname"    # should be set by proxmox
+
2014-05
GATEWAY=123.123.123.123        # CHANGE (and remove from ifcfg-eth0)
+
2014-05
</pre>
+
Proxmox VE 1.x
 
+
Debian 5 (Lenny)
==Virtualization==
+
2008-10
===Why do you recommend 32-bit guests over 64 bit guests?===
+
2012-03
:64-bit makes sense only if you need greater than 4GB of memory.
+
2013-01
:32-bit guests use less memory in certain situations
+
How can I upgrade Proxmox VE to the next release?
::e.g. a standard installation of apache2 on 64 bit containers consumes much more memory than on 32 bit.
+
Minor version upgrades, for example upgrading from Proxmox VE in version 5.1
 
+
to 5.2, can be done just like any normal update, either through the Web
 
+
GUI Node &#8594; Updates panel or through the CLI with:
== Troubleshooting ==
+
apt update
See [[Troubleshooting]] page.
+
apt full-upgrade
 +
Always ensure you correctly setup the
 +
package repositories and only
 +
continue with the actual upgrade if apt update did not hit any error.
 +
Major version upgrades, for example going from Proxmox VE 4.4 to 5.0, are
 +
also supported. They must be carefully planned and tested and should
 +
never be started without having a current backup ready.
 +
Although the specific upgrade steps depend on your respective setup, we
 +
provide general instructions and advice of how a upgrade should be
 +
performed:
 +
Upgrade from Proxmox VE 5.x to 6.0
 +
Upgrade from Proxmox VE 4.x to 5.0
 +
Upgrade from Proxmox VE 3.x to 4.0
 +
LXC vs LXD vs Proxmox Containers vs Docker
 +
LXC is a userspace interface for the Linux kernel containment
 +
features. Through a powerful API and simple tools, it lets Linux users
 +
easily create and manage system containers. LXC, as well as the former
 +
OpenVZ, aims at system virtualization, i.e. allows you to run a
 +
complete OS inside a container, where you log in as ssh, add users,
 +
run apache, etc&#8230;
 +
LXD is building on top of LXC to provide a new, better user
 +
experience. Under the hood, LXD uses LXC through liblxc and its Go
 +
binding to create and manage the containers. It&#8217;s basically an
 +
alternative to LXC&#8217;s tools and distribution template system with the
 +
added features that come from being controllable over the network.
 +
Proxmox Containers also aims at system virtualization, and thus uses
 +
LXC as the basis of its own container offer. The Proxmox Container
 +
Toolkit is called pct, and is tightly coupled with Proxmox VE. That means
 +
that it is aware of the cluster setup, and it can use the same network
 +
and storage resources as fully virtualized VMs. You can even use the
 +
Proxmox VE firewall, create and restore backups, or manage containers using
 +
the HA framework. Everything can be controlled over the network using
 +
the Proxmox VE API.
 +
Docker aims at running a single application running in a contained
 +
environment. Hence you&#8217;re managing a docker instance from the host with the
 +
docker toolkit. It is not recommended to run docker directly on your
 +
Proxmox VE host.
 +
You can however perfectly install and use docker inside a Proxmox Qemu
 +
VM, and thus getting the benefit of software containerization with the very
 +
strong isolation that VMs provide.
 +
</pvehide>
 +
<!--PVE_IMPORT_END_MARKER-->

Latest revision as of 11:23, 16 July 2019

Note New FAQs are appended to the bottom of this section.
  1. What distribution is Proxmox VE based on?

    Proxmox VE is based on Debian GNU/Linux

  2. What license does the Proxmox VE project use?

    Proxmox VE code is licensed under the GNU Affero General Public License, version 3.

  3. Will Proxmox VE run on a 32bit processor?

    Proxmox VE works only on 64-bit CPUs (AMD or Intel). There is no plan for 32-bit for the platform.

    Note VMs and Containers can be both 32-bit and/or 64-bit.
  4. Does my CPU support virtualization?

    To check if your CPU is virtualization compatible, check for the vmx or svm tag in this command output:

    egrep '(vmx|svm)' /proc/cpuinfo
  5. Supported Intel CPUs

    64-bit processors with Intel Virtualization Technology (Intel VT-x) support. (List of processors with Intel VT and 64-bit)

  6. Supported AMD CPUs

    64-bit processors with AMD Virtualization Technology (AMD-V) support.

  7. What is a container, CT, VE, Virtual Private Server, VPS?

    Operating-system-level virtualization is a server-virtualization method where the kernel of an operating system allows for multiple isolated user-space instances, instead of just one. We call such instances containers. As containers use the host’s kernel they are limited to Linux guests.

  8. What is a QEMU/KVM guest (or VM)?

    A QEMU/KVM guest (or VM) is a guest system running virtualized under Proxmox VE using QEMU and the Linux KVM kernel module.

  9. What is QEMU?

    QEMU is a generic and open source machine emulator and virtualizer. QEMU uses the Linux KVM kernel module to achieve near native performance by executing the guest code directly on the host CPU. It is not limited to Linux guests but allows arbitrary operating systems to run.

  10. How long will my Proxmox VE version be supported?

    Proxmox VE versions are supported at least as long as the corresponding Debian Version is oldstable. Proxmox VE uses a rolling release model and using the latest stable version is always recommended.

    Proxmox VE Version Debian Version First Release Debian EOL Proxmox EOL

    Proxmox VE 6.x

    Debian 10 (Buster)

    2019-07

    tba

    tba

    Proxmox VE 5.x

    Debian 9 (Stretch)

    2017-07

    2020-07

    2020-07

    Proxmox VE 4.x

    Debian 8 (Jessie)

    2015-10

    2018-06

    2018-06

    Proxmox VE 3.x

    Debian 7 (Wheezy)

    2013-05

    2016-04

    2017-02

    Proxmox VE 2.x

    Debian 6 (Squeeze)

    2012-04

    2014-05

    2014-05

    Proxmox VE 1.x

    Debian 5 (Lenny)

    2008-10

    2012-03

    2013-01

  11. How can I upgrade Proxmox VE to the next release?

    Minor version upgrades, for example upgrading from Proxmox VE in version 5.1 to 5.2, can be done just like any normal update, either through the Web GUI Node → Updates panel or through the CLI with:

    apt update
    apt full-upgrade
    Note Always ensure you correctly setup the package repositories and only continue with the actual upgrade if apt update did not hit any error.

    Major version upgrades, for example going from Proxmox VE 4.4 to 5.0, are also supported. They must be carefully planned and tested and should never be started without having a current backup ready. Although the specific upgrade steps depend on your respective setup, we provide general instructions and advice of how a upgrade should be performed:

  12. LXC vs LXD vs Proxmox Containers vs Docker

    LXC is a userspace interface for the Linux kernel containment features. Through a powerful API and simple tools, it lets Linux users easily create and manage system containers. LXC, as well as the former OpenVZ, aims at system virtualization, i.e. allows you to run a complete OS inside a container, where you log in as ssh, add users, run apache, etc…

    LXD is building on top of LXC to provide a new, better user experience. Under the hood, LXD uses LXC through liblxc and its Go binding to create and manage the containers. It’s basically an alternative to LXC’s tools and distribution template system with the added features that come from being controllable over the network.

    Proxmox Containers also aims at system virtualization, and thus uses LXC as the basis of its own container offer. The Proxmox Container Toolkit is called pct, and is tightly coupled with Proxmox VE. That means that it is aware of the cluster setup, and it can use the same network and storage resources as fully virtualized VMs. You can even use the Proxmox VE firewall, create and restore backups, or manage containers using the HA framework. Everything can be controlled over the network using the Proxmox VE API.

    Docker aims at running a single application running in a contained environment. Hence you’re managing a docker instance from the host with the docker toolkit. It is not recommended to run docker directly on your Proxmox VE host.

    Note You can however perfectly install and use docker inside a Proxmox Qemu VM, and thus getting the benefit of software containerization with the very strong isolation that VMs provide.