by theforeman

theforeman / forklift

Helpful deployment scripts for Foreman and Katello

133 Stars 169 Forks Last release: Not found GNU General Public License v3.0 1.4K Commits 1 Releases

Available items

No Items, yet!

The developer of this repository has not created any items for sale yet. Need a bug fixed? Help with integration? A different license? Create a request here:

Forklift: lifting Foreman into your environment

Forklift provides tools to create Foreman/Katello environments for development, testing and production configurations. Whether you are a developer wanting an environment to write code, or a user wanting to spin up the latest production environment for deployment or evaluation Forklift has you covered.

Using Forklift


  • Vagrant - 2.2+ - Both the VirtualBox and Libvirt providers are tested
  • Ansible - 2.7+
  • Vagrant Libvirt provider plugin (if using Libvirt)
  • Virtualization enabled in BIOS

See Installing Vagrant for installation instructions.


This will walk through the simplest path of spinning up a production test environment of a bleeding edge nightly installation assuming Vagrant and Libvirt are installed and configured.

git clone
cd forklift
vagrant up centos7-foreman-nightly

The same can be quickly done for a development environment where GITHUB_NICK is your GitHub username:

git clone
cd forklift
cp vagrant/boxes.d/99-local.yaml.example vagrant/boxes.d/99-local.yaml
sed -i.bak "s//GITHUB_NICK/g" vagrant/boxes.d/99-local.yaml
vagrant up centos7-katello-devel

In case using vagrant is not desired, ansible playbooks and roles from this repo can be used separately. This is useful if an existing host should be used for the installation, e.g. a beaker machine. In order to deploy the devel environment on host, the following needs to be done:

on machine, where the dev env should be deployed

useradd vagrant
echo "vagrant   ALL=(ALL)   NOPASSWD: ALL" >> /etc/sudoers`

in forklift checkout

echo "[devel]\" > inventories/local_inventory
ansible-playbook --private-key=~/.ssh/id_rsa --user root --inventory inventories/local_inventory --extra-vars katello_devel_github_username=katello playbooks/devel.yml

In an example above, ansible was instructed to use specific private key (overriding the value from ansible.cfg), root user was set as ssh user and playbook variable was set, so that checkout will be made from katello user.

Other playbooks from playbooks/ directory can be used similarly, though some might need more variables and investigating their parameters is recommended first.

More thorough guides can be found in the docs folder.


By default

deploys Foreman with
as username and password, please change this on production installs (either after the install, or by setting
during the initial deployment).

Poor man's DNS a.k.a /etc/hosts

For the multi-host setup, one of the easiest way of making the name resolution working with vagrant is using vagrant-hostmanager. Forklift supports this plugin by default. The only thing one needs to do is install the vagrant-hostmanager plugin:

vagrant plugin install vagrant-hostmanager

By default, the boxes are set with

If you're using NetworkManager, this advanced DNS configuration allows completely automated dns resolution using dnsmasq from host to guest and guest to guest.

You can disable hostmanager in

by setting

When using a briged network or with multiple network interfaces, the hostmanager would resolve the internal ip from the ssh-info, you can choose an ip of a particular network interface in

by setting
option with the device name.
hostmanager_ip_resolver_device: 'eth1'

Adding Custom Boxes

Sometimes you want to spin up the same box type (e.g. centos7-katello-devel) from within the forklift directory. While this can be added to the Vagrantfile directly, updates to the forklift repository could wipe out your local changes. To help with this, you can define a custom box re-using the configuration within the Vagrantfile. To do so, create a

file in vagrant/boxes.d/. For example, to create a custom box on CentOS 7 with nightly and run the installers reset command:
  box: centos7
    playbook: playbooks/katello.yml
      katello_repositories_environment: staging
    verbose: vvv


| Option | Description | |:-------------------------|:----------------------------------------------------------------------| | box | the ':name' one of the defined boxes in the Vagrantfile | | bridged | deploy on Libvirt with a bridged networking configuration, value of this parameter should be the interface of the host (e.g. em1) | | memory | set the amount of memory (in megabytes) this box will consume | | cpus | set the number of cpus this box will use | | hostname | hostname to set on the box | | networks | custom networks to use in addition to the management network | | disksize | specify the size (in gigabytes) of the box's virtual disk. This only sets the virtual disk size, so you will still need to resize partitions and filesystems manually. | | adddisks | (libvirt provider only) specify additional libvirt volumes | | ansible | updates the Ansible provisioner configuration including the playbook to be ran or any variables to set | | libvirtoptions | sets Libvirt specific options, see

for possible options | | virtualbox
options | sets VirtualBox specific options | | rackspaceoptions | sets Rackspace specific options | | openstackoptions | sets OpenStack specific options | | googleoptions | sets Google specific options | | domain | forklift uses short name of your host + '' as domain name for your boxes. You can use this option to override it. | | sshfs | if you have vagrant-sshfs plugin, you can use sshfs to share folders between your host and guest. See an example below for details. | | nfs | share folders between host and guest. See an example below for details. | | autostart | set to true to automatically start when using 'vagrant up' | | primary | set the machine to be the default target of vagrant commands such as 'vagrant ssh' | | libvirtqemuusesession | Use qemu session instead of system |

Entirely new boxes can be created that do not orginate from a box defined within the Vagrantfile. For example, if you had access to a RHEL Vagrant box:

  box_name: rhel7
  shell: 'echo TEST'
  pty: true

Example with custom networking, static IP on custom libvirt network:

  box: centos7
    - type: 'private_network'
        libvirt__network_name: lab-private
        libvirt__iface_name: vnet2

Example with custom libvirt management network:

  box: centos7

Example with openstack provider: You will need to install vagrant openstack provider. For more information click here. Do not forget to set openstack API credentials. To use openstack provider as default look here.

  image_name: 'Centos7'
  username: 'centos'  #root by default
  hostname: 'john-doe'
  openstack_flavor: 'm1.medium'
  sync_type: 'disabled'

Using SSHFS to share folders

You will need to install vagrant-sshfs plugin. Make sure your host actually has sshfs installed. Example with sshfs mounting folder from guest to host:

  box: centos7
    host_path: '/some/host/path'
    guest_path: '/some/guest/path'
    reverse: True

If you want to mount in the opposite direction, just change

or remove it entirely.

Additonal options may be specified with using

  box: centos7
    host_path: '/some/host/path'
    guest_path: '/some/guest/path'
    options: '-o allow_other'

Example with an additional disk (libvirt volume) presented as /dev/vdb in the vm:

  box: centos7
    - size: 100GiB
      device: vdb
      type: qcow2

Using NFS to share folders

An alternative to SSHFS is to share the folders with NFS. It is slightly more work than SSHFS. See the Fedora developer documentation for information about how to configure an NFS server for Vagrant.

Then create your box:

  box: centos7
    host_path: '/some/host/path'
    guest_path: '/some/guest/path'

Customize Deployment Settings

Some settings can be customized for the entirety of the deployment by copying

and add, remove or updating:
  • memory: Memory to give boxes by default unless specified by a box
  • cpus: Number of CPUs to give boxes by default unless specified by a box
  • scale_memory: Factor to multiply memory of boxes that specify an own value
  • scale_cpus: Factor to multiply CPUs of boxes that specify an own value
  • sync_type: type of sync to use for transfer to the Vagrant box
  • mount_options: options for the vagrant-cachier plugin
  • domain: domain for your hosts, you can override this per-box by configuring your box with a domain directly
  • libvirtoptions, virtualboxoptions, rackspaceoptions, openstackoptions, google_options: custom options for the various providers

Customize Available Boxes

The list of available boxes can be customized by setting an exclude list in

. This allows faster
vagrant status
calls as well as reducing the the scope of boxes a user sees to tailor to their use cases. To specify boxes to exclude add the following to
, for example, to remove fips, fedora and any Foreman 1.2X boxes from view:
    - "katello" # exclude any box containing "katello"
    - "ubuntu1604-foreman-1\\.24" # exclude only the box "ubuntu1604-foreman-1.24". Notice the escaped '.' character to match the specific character instead of any single character
    - "^centos7-fips" # exclude any box that starts with "centos7-fips"
    - "foreman-1\\.(?:[2][0-3])" # exclude any foreman-1.20 to foreman-1.23 version box

Post Install Playbooks

Boxes can be further customized by declaring Ansible playbooks to be run during provisioning. One or more playbooks can be specified and will be executed sequentially. An ignored directory can be used to put playbooks into 'user_playbooks' without worrying about adding them during a git commit.

Ansible roles may also be installed directly using the

command. These roles will be installed at

and will be ignored by git. You may also specify roles in a
, which you can use to install all desired roles with
ansible-galaxy install -r requirements.yml
  box: centos7-katello-nightly
      - 'user_playbooks/vim.yml'
      - 'user_playbooks/zsh.yml'

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.