]>
Commit | Line | Data |
---|---|---|
e71b5d0d DM |
1 | Installing Proxmox VE |
2 | --------------------- | |
3 | include::attributes.txt[] | |
d91f8c1e DM |
4 | |
5 | {pve} ships as a set of Debian packages, so you can simply install it | |
6 | on top of a normal Debian installation. After configuring the | |
7 | repositories, you need to run: | |
8 | ||
9 | [source,bash] | |
10 | ---- | |
11 | apt-get update | |
12 | apt-get install proxmox-ve | |
13 | ---- | |
14 | ||
15 | While this looks easy, it presumes that you have correctly installed | |
16 | the base system, and you know how you want to configure and use the | |
17 | local storage. Network configuration is also completely up to you. | |
18 | ||
19 | In general, this is not trivial, especially when you use LVM or | |
20 | ZFS. This is why we provide an installation CD-ROM for {pve}. That | |
21 | installer just ask you a few questions, then partitions the local | |
22 | disk(s), installs all required packages, and configures the system | |
23 | including a basic network setup. You can get a fully functional system | |
24 | within a few minutes, including the following: | |
25 | ||
26 | * Complete operating system (Debian Linux, 64-bit) | |
27 | * Partition the hard drive with ext4 (alternative ext3 or xfs) or ZFS | |
28 | * {pve} Kernel with LXC and KVM support | |
29 | * Complete toolset | |
30 | * Web based management interface | |
31 | ||
32 | NOTE: By default, the complete server is used and all existing data is | |
33 | removed. | |
34 | ||
35 | Using the {pve} Installation CD-ROM | |
36 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
37 | ||
38 | Please insert the installation CD-ROM, then boot from that | |
39 | drive. Immediately afterwards you can choose the following menu | |
40 | options: | |
41 | ||
42 | Install Proxmox VE:: | |
43 | ||
44 | Start normal installation. | |
45 | ||
46 | Install Proxmox VE (Debug mode):: | |
47 | ||
48 | Start installation in debug mode. It opens a shell console at several | |
49 | installation steps, so that you can debug things if something goes | |
50 | wrong. Please press `CTRL-D` to exit those debug consoles and continue | |
51 | installation. This option is mostly for developers and not meant for | |
52 | general use. | |
53 | ||
54 | Rescue Boot:: | |
55 | ||
56 | This option allows you to boot an existing installation. It searches | |
57 | all attached hard disks, and if it finds an existing installation, | |
58 | boots directly into that disk using the existing Linux kernel. This | |
59 | can be useful if there are problems with the boot block (grub), or the | |
60 | BIOS is unable to read the boot block from the disk. | |
61 | ||
62 | Test Memory:: | |
63 | ||
64 | Runs 'memtest86+'. This is useful to check if your memory if | |
65 | functional and error free. | |
66 | ||
67 | You normally select *Install Proxmox VE* to start the installation. | |
68 | After that you get prompted to select the target hard disk(s). The | |
8e4bb261 | 69 | `Options` button lets you select the target file system, which |
d91f8c1e DM |
70 | defaults to `ext4`. The installer uses LVM if you select 'ext3', |
71 | 'ext4' or 'xfs' as file system, and offers additional option to | |
72 | restrict LVM space (see <<advanced_lvm_options,below>>) | |
73 | ||
74 | If you have more than one disk, you can also use ZFS as file system. | |
75 | ZFS supports several software RAID levels, so this is specially useful | |
76 | if you do not have a hardware RAID controller. The `Options` button | |
77 | lets you select the ZFS RAID level, and you can choose disks there. | |
78 | ||
79 | The next pages just asks for basic configuration options like time | |
80 | zone and keyboard layout. You also need to specify your email address | |
81 | and select a superuser password. | |
82 | ||
83 | The last step is the network configuration. Please note that you can | |
84 | use either IPv4 or IPv6 here, but not both. If you want to configure a | |
85 | dual stack node, you can easily do that after installation. | |
86 | ||
87 | If you press `Next` now, installation starts to format disks, and | |
88 | copies packages to the target. Please wait until that is finished, | |
89 | then reboot the server. | |
90 | ||
91 | Further configuration is done via the Proxmox web interface. Just | |
92 | point your browser to the IP address given during installation | |
93 | (https://youripaddress:8006). {pve} is tested for IE9, Firefox 10 | |
94 | and higher, and Google Chrome. | |
95 | ||
96 | ||
97 | [[advanced_lvm_options]] | |
98 | Advanced LVM configuration options | |
99 | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | |
100 | ||
101 | The installer creates a Volume Group (VG) called `pve`, and additional | |
102 | Logical Volumes (LVs) called `root`, `data` and `swap`. The size of | |
103 | those volumes can be controlled with: | |
104 | ||
105 | `hdsize`:: | |
106 | ||
107 | Defines the total HD size to be used. This way you can save free | |
108 | space on the HD for further partitioning (i.e. for an additional PV | |
109 | and VG on the same hard disk that can be used for LVM storage). | |
110 | ||
111 | `swapsize`:: | |
112 | ||
113 | To define the size of the `swap` volume. Default is the same size as | |
114 | installed RAM, with 4GB minimum and `hdsize/8` as maximum. | |
115 | ||
116 | `maxroot`:: | |
117 | ||
118 | The `root` volume size. The `root` volume stores the whole operation | |
119 | system. | |
120 | ||
121 | `maxvz`:: | |
122 | ||
123 | Define the size of the `data` volume, which is mounted at | |
124 | '/var/lib/vz'. | |
125 | ||
126 | `minfree`:: | |
127 | ||
128 | To define the amount of free space left in LVM volume group `pve`. | |
129 | 16GB is the default if storage available > 128GB, `hdsize/8` otherwise. | |
130 | + | |
131 | NOTE: LVM requires free space in the VG for snapshot creation (not | |
132 | required for lvmthin snapshots). | |
133 | ||
134 | ||
135 | ZFS Performance Tips | |
136 | ^^^^^^^^^^^^^^^^^^^^ | |
137 | ||
138 | ZFS uses a lot of memory, so it is best to add additional 8-16GB RAM | |
139 | if you want to use ZFS. | |
140 | ||
141 | ZFS also provides the feature to use a fast SSD drive as write cache. The | |
142 | write cache is called the ZFS Intent Log (ZIL). You can add that after | |
143 | installation using the following command: | |
144 | ||
145 | zpool add <pool-name> log </dev/path_to_fast_ssd> | |
146 |