Virtual environment / ansible core
Setup SSH connection details for all hosts
Shared connection to 10.11.12.13
title Converts first character of every word to uppercase.
ansible_os_family and ansible_distribution
ansible_distribution_major_version
Dynamic variables with set_fact
Extract output from a command in a loop
Loop over comma separated list
authorized_key (SSH authorized keys)
Cluster nodes (lineinfile in a loop)
Write a file on the controller
Check support matrix to determine whether you have a supported Python version.
Install Python venv module, e.g. system-wide for Debian/Ubuntu:
$ sudo apt install python3-venv
Create virtual environment for Ansible core:
$ python3 -m venv ansible2.18
Activate virtual environment:
$ source ansible2.18/bin/activate (ansible2.18) $
Install Ansible core:
$ python3 -m pip install ansible-core==2.18 ...
The only inevitable requirements for the target are a SSH server and Python.
Modern distributions are using Python 3, therefore we recommend to force Ansible using it:
# group_vars/all.yml ansible_python_interpreter: /usr/bin/python3
Otherwise Ansible uses Python 2 if available on the system.
This applies to:
Debian 9 (Stretch) and later
CentOS 8
Install Python3:
~ apk update ~ apk add --quiet python3
Synchronizing package databases:
$ sudo pacman -Fy
Install Python3:
$ sudo pacman -S python3
On a Debian host with minimal setup you can install Python as follows:
% apt-get update % apt-get install python3
It is also useful to install python3-apt on Debian hosts. This allows you to run the playbook in check mode on a pristine system.
Installing packages requires the equery program:
$ sudo emerge app-portage/gentoolkit
Install Python3:
~# pkg install python3
Adjust ansible_python_interpreter variable:
ansible_python_interpreter: /usr/local/bin/python3
RedHat 8 and derivatives like CentOS and Alma Linux OS don't come with Python pre-installed,
~ dnf update ~ dnf install -y python3
In the file group_vars/all:
--- ansible_connection: ssh ansible_user: ansible
IP address of the target host
Ansible user on target host
Connection method (defaults to ssh)
Private key file on the controller
Example:
test_vml ansible_host=192.168.6.66 ansible_user=ansible ansible_ssh_private_key_file=/home/mynames/keys/example.pem
The ansible_user needs to be in the group sudo on Ubuntu host.
Oct 9 17:59:25 mytestbox sudo: nevairbe : command not allowed ; TTY=pts/1 ; PWD=/home/nevairbe ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-xrqnygkastkbrykkzzvttjbhvzajdltm ; /usr/bin/python3 /home/nevairbe/.ansible/tmp/ansible-tmp-1696867164.8682141-9644-176952864770288/AnsiballZ_setup.py
The Check mode is very useful to simulate what will be applied to the host without actually changing anything.
However, some tasks don't work in check mode. You can choose either to ignore errors in check mode or skip these tasks altogether. Please note that the command module automatically skip its tasks in check mode.
Ignore errors in check mode:
- name: Check if the elasticsearch package is installed
shell: dpkg-query -W -f'${Status}' elasticsearch
ignore_errors: "{{ ansible_check_mode }}"
Skip tasks in check mode:
- name: Check if the elasticsearch package is installed
shell: dpkg-query -W -f'${Status}' elasticsearch
when: not ansible_check_mode
The default output for errors is hard to read, thus it is recommended to configure the YAML callback plugin in ansible.cfg:
[defaults] ... # Use the YAML callback plugin. stdout_callback = yaml # Use the stdout_callback when running ad-hoc commands. bin_ansible_callbacks = True
Thanks to Jeff Geerling for his blog post about this feature.
This callback can also be turned out by an environment variable:
export ANSIBLE_STDOUT_CALLBACK=yaml
Reference: https://docs.ansible.com/ansible/latest/plugins/callback.html
The following characters are invalid group names: spaces, hyphens, and preceding numbers.
A task calls an Ansible module and can be modified by task keywords, e.g. conditions or delegations.
Create directory:
- name: Create DKIM directory
file:
state: directory
path: "/etc/exim4/dkim"
owner: root
Copy file:
- name: Install DKIM domain setup script
copy:
src: "dkimsetup"
dest: "/usr/local/sbin/dkimsetup"
owner: root
mode: 0755
Create file from template:
- name: Add DKIM settings
template:
src: "dkim_macros.j2"
dest: "/etc/exim4/conf.d/main/00_dkim_macros"
owner: root
Add or change line in a file:
- name: Register Foo's certificate chain in LDAP configuration file
lineinfile:
dest: /etc/openldap/ldap.conf
regex: '^TLS_CACERT\s+'
line: 'TLS_CACERT /etc/ssl/foo-chain.pem'
You can install distribution packages with the generic package module or with the distribution specific module.
Install Git package:
- name: Install Git package
package:
name: git
Example for creating a cronjob:
- name: Create cronjob
cron:
name: Mrad backup
minute: 12
hour: 22
user: root
job: "/usr/local/bin/mrad /etc/mrad.cfg"
cron_file: mrad
This creates a file mrad in the directory /etc/cron.d as per the cron_file attribute.
The contents are:
#Ansible: Mrad backup 12 22 * * * root /usr/local/bin/mrad /etc/mrad.cfg
For configuration and other text files you can use either lineinfile or blockinfile module.
There also other modules for specific formats, e.g. the xml module for XML files.
- name: Install sensor package for Ubuntu
apt:
name: sensor
environment:
SERVER_ADDRESS: "{{ cyber_server_address }}"
CUSTOMER_ID: "{{ cyber_customer_id }"
Jinja2 templates are used all over space. First in mind are template files for the template task, but they can used in other places like in when conditions.
Jinja has nice features, e.g. template inheritance.
You might consider to add a header line, which makes system administrators aware of the origin on the file, e.g.:
{{ ansible_managed | comment }}
Default output is:
# Ansible managed
Reference: https://docs.ansible.com/ansible/latest/user_guide/playbooks_filters.html
In addition to the Ansible filters you can also use Jinja filters.
This filter returns the value 443 if the variable https_port is not defined.
"{{ https_port | default(443) }}"
If you use the default filter with the special variable omit, the parameter would not be passed to the ansible module. That allows us to gracefully fall back to the default of the module parameter.
- name: Create PostgreSQL database
postgresql_db:
name: "{{ dbname }}"
owner: "{{ dbuser | default(omit) }}"
Alias for this filter: d
Returns the file name from a file path. Useful with with_fileglob.
Returns a file path without the file name.
{{ '/etc/fstab' | dirname }}
/etc
}}
**** *reverse* (Jinja filter)
#filterreverse
This reverses a string, a list or any other Python object.
*** List filters
#listfilters
See also how to use [[#pythonlists][Python syntax]] to manipulate lists.
**** *first*
#filterfirst
Returns first element of a list.
**** *join*
#filterjoin
Joins list members into a string:
{{{
{{ monitoring_server_ips | join(' ') }}
This is filter is also useful to print the output from a command into a file, line by line:
- copy:
content: "{{ grep_results.stdout_lines | join('\n') }}"
dest: /tmp/grep.txt
Counts the members of the list which is convenient in a condition:
when: mylist | length
The map filter has a number of different usages.
It can be used to apply any of the string filters to all elements of the list, e.g.
"{{ services.split(',') | map('trim') | list }}"
Extract
If you have a list of dictionaries and you want a single value from each dictionary, apply attribute map:
{{ nginx_vhosts | map(attribute='server_name') | list }}
This filter selects any elements with matches a given condition:
{{ omd_versions.stdout_lines | select('search', '\\s+\\(default\\)$') }}
The difference between select('search') and select('match') is that the latter requires the whole element to match.
This select all entries from a list of dictionaries where one attribute from the dictionary fulfils a certain condition.
{{ rest_api_fields | selectattr('type', 'equalto', 'string') | list }}
This filter removes elements from a list of strings.
In the following example we remove all filenames which contain the string junk:
{{ filenames | reject('search', 'junk') | list }}
This filter sorts a list, which can be useful to achieve a stable output. For example, the list of hosts in a group doesn't maintain an order.
{% for host in groups['all'] | sort %}
{{ hostvars[host]['inventory_hostname_short'] }} {{ hostvars[host['ansible_facts']['eth0']['ipv4']['address'] }}
{% endfor %}
For list consisting of dictionaries, you can also sort by an attribute.
Here we are taking the list of files registered by the find module and sorting them by their modification time (oldest first):
{{ find_result.files | sort(attribute='mtime') | list }}
You can also sort in reverse order (newest first):
{{ find_result.files | sort(attribute='mtime',reverse=True) | list }}
Reduces a list to unique items by omitting duplicate ones.
Turns a dictionary in a list of dictionaries. Each entry in the new list has a key attribute with the key in the orginal dictionary and a value attribute with value from the original dictionary.
This is handy to loop over a dictionary:
{% for vg in ansible_lvm.vgs | dict2items %}
{{ vg.key }} {{ vg.value.free_g }} GB
{% endfor %}
Uppercases first character of the string.
{{ "bullseye" | capitalize }}
# => 'Bullseye'
See filter title for uppercasing the first character of every word.
Turns string into a comment. This is especially useful for multi-line strings. By default the filter uses the # sign.
Converts whole string to lowercase.
Encrypts string with given method:
{{ user_password | password_hash('bcrypt') }}
This filter can be used to extract strings, e.g. a form value from HTML retrieved by the uri module.
Form element:
<input type="hidden" name="csrftoken" value="db788e6feb8a4927db84d6c0da1dfe67" />
Filter to extract the value:
"{{ sympa_response.content | regex_findall('name=\"csrftoken\" value=\"(.*)\"') | first }}"
Result: db788e6feb8a4927db84d6c0da1dfe67
Replaces match of regular expression.
If you want to replace a multiline string, e.g. a comment:
/* Get rid of me */ Keep me
You can't use .* here as the dot doesn't match the newline.
It works if you specify a character class with whitespace and non whitespace:
regex_replace('/\\*([\\s\\S]*)\\*/', '')
Compares string with a regular expression and returns the match.
For example to get the numeric part from the hostname of the target:
- hosts: fedora33-text-box
tasks:
debug:
msg: "{{ inventory_hostname | regex_search ('[0-9]+') }}"
This results in the following output:
TASK [debug] ******************************************************************* ok: [fedora33-test-box] => msg: '33'
You can also capture part of the match:
- name: Determine version of RPM package to install on the server
set_fact:
rpm_package_version: "{{ rhn_ssl_tool.stdout | regex_search(regexp,'\\1') | first }}"
vars:
regexp: 'rhn-org-httpd-ssl-key-pair-spacewalk7-(.*).noarch.rpm'
{{ "foo bar" | title }}
# => 'Foo Bar'
Removes leading and trailing whitespace.
Converts whole string to uppercase.
Calculate 1 GB:
1024 | pow(3)
Rounds a number with or without decimal points:
10.5 | round => 11 10.444 | round(2) => 10.44
As a JSON query can result in multiple matches, the json_query filter returns a list.
Uses JSON Matching Expression paths (JMESPATH).
It requires the jmespath Python library to be installed on the controller.
Converts to integer type. Useful for numerical comparison between a variable and an integer:
ansible_distribution_major_version|int >= 10
Converts to string type.
- name: Configure IPv6
ansible.posix.sysctl:
name: "{{ item }}"
value: "{{ ipv6_disabled | string }}"
loop:
- net.ipv6.conf.default.disable_ipv6
- net.ipv6.conf.lo.disable_ipv6
- net.ipv6.conf.all.disable_ipv6
Reference for tests builtin into Jinja: https://jinja.palletsprojects.com/en/2.11.x/templates/#list-of-builtin-tests
Whether a variable is defined or not.
Whether a number is divisble without fraction. Can be useful in loops to create batches:
{% for email in addresses %}{{ email }}{% if loop.index is divisibleby 10 %}{{ '\n -}}{% else %};{% endif %}{% endfor %}
Check if variable internet_access is true:
- name: Ensure that APT cache is up-to-date
apt:
update_cache: true
cache_valid_time: 14400
when: internet_access is sameas true
Determines if a task was successful by looking at a variable registered by the task. Useful in combination with the until task keyword.
Don't use curly braces in the conditions.
Regular expressions are used by the regex_findall and regex_replace filters.
Facts are information gathered by Ansible, usually by running a setup task before executing other tasks.
TASK [Gathering Facts] ********************************************************* ok: [buster-test-box]
Facts are accessible as variables.
Magic variables are automatically set by the Ansible but can't be overridden.
Commonly used magic variables are:
map with all groups and the corresponding hosts
full hostname as specified in the inventory, e.g. foo.example.com
short version of inventory_hostname, e.g. foo
useful to locate resources relative to the inventory directory
The complete list is available in the Ansible documentation.
A list of common OS families and their distributions:
Debian, Ubuntu, Kali
RedHat, CentOS, Fedora
SLES, OpenSUSE
FreeBSD
Gentoo
Alpine
Archlinux
As code names for releases are used only by Debian based distributions, this variable makes only sense for distributions in the Debian ansible_os_family.
bookworm
bullseye
buster
stretch
jammy
focal
bionic
xenial
This is the distribution version number, e.g 10.5 for Debian buster.
8
9
10
16
18
20
31
15
2
2020 (current year)
You can use this variable to install a Debian package only on releases that comes with that package:
- name: Install certbot package(s)
apt:
name:
- certbot
when:
- ansible_distribution == 'Debian'
- ansible_distribution_major_version|int >= 10
This shows you the main architecture for your OS.
| Family | Architecture |
|---|---|
| Alpine | x86_64 |
| Archlinux |
x86_64
|
| Debian |
x86_64
|
| FreeBSD |
amd64
|
| Gentoo | x86_64 |
| RedHat |
x86_64
|
| Suse |
x86_64
|
This shows you the kernel version, e.g. 5.10.0-9-amd64.
The machine id (UUID) for the target, which is stored in /etc/machine_id,
e.g. 1ff77447cd174f4a9d7e37aed637c388.
Service manager on the target. Most modern distributions are using systemd.
| Distribution | Service manager | Using it since |
| Alpine | service | |
| Archlinux | systemd | |
| CentOS | systemd | CentOS 7 |
| Debian | systemd | Debian 8 |
| Fedora | systemd | |
| FreeBSD | bsdinit | |
| Gentoo | openrc | |
| Suse | systemd |
ansible_facts['virtualization_role'] has the value guest for virtual
machines.
ansible_memory_mb provides memory information in Megabytes:
nocache: free: 828 used: 1154 real: free: 82 total: 1982 used: 1900 swap: cached: 0 free: 0 total: 0 used: 0
ansible_interfaces are a list of the network interfaces on the target host.
Example contents:
ansible_interfaces: - lo - enp2s0 - wlp0s20f3
ansible_default_ipv4 points to the interface used for the default route. In most cases it can be used to determine the "main IP address" for a server.
Example contents of the ansible_default_ipv4 variable:
ansible_default_ipv4:
address: 10.0.2.15
alias: eth0
broadcast: 10.0.2.255
gateway: 10.0.2.2
interface: eth0
macaddress: 08:00:27:8d:c0:4d
mtu: 1500
netmask: 255.255.255.0
network: 10.0.2.0
type: ether
The variable ansible_all_ipv4_addresses is list of all IPv4 addresses on the target
host with the exception of addresses from the loopback interface
(127.0.0.0/8 address block).
Example contents:
["192.168.2.130", "141.57.69.174", "192.168.2.112"]
In order to determine whether the host uses Systemd you can check the *ansible_service_mgr" variable.
user name used for connecting to the target host
List of tags passed to ansible-playbook. Using --tags=certbot,nginx
results in:
ansible_run_tags: - nginx - certbot
Without --tags the list contains one element all:
ansible_run_tags: - all
The set_fact module allows you to generate variables from other variables (from inventory, register, ...).
It can be a challenge to find out how to use set_fact based on a previous command in a loop, but it is quiet simple if you use the sum filter:
- set_fact:
modules_list:
"{{ command_output.results | sum(attribute='stdout_lines', start=[]) }}"
You can access variables for other hosts from the current inventory through the hostvars array.
You can apply standard Python methods to variables.
Useful methods are join, split, startswith and endswith.
You can use a regular expression instead of startswith or endswith, but these methods improve the readablility.
In the following we check if the listen address is a Unix socket. If not we configure the allowed clients for TCP address.
{% if item.pool_listen.startswith('/') == false %}
listen.allowed_clients = {{ item.pool_listen_allowed_clients | default('127.0.0.1', true) }}
{% endif %}
See also: https://docs.python.org/3/library/stdtypes.html#string-methods.
Slicing can be used for strings and lists with the same syntax
[start:end]. The first element is 0, so [1:] removes the first character
from a string respective the first element of an array.
Returns the string represented by the variable backup_base_directory with the first character removed:
{{ backup_base_directory[1:] }}
Removes subdomain to use as cookie domain:
cookie_domain: ".{{ app_domain.split('.')[1:] | join('.') }}"
Lists can be combined with the + operator:
- set_fact: Combine two arrays
myarray: "{{ myarray + youarray }}"
This can also used to add element(s) to a list, just wrap
them in a new list []:
- set_fact: Add element to array
myarray: "{{ myarray + ['element'] }}"
This is good example to prevent variable nesting while inside a Jinja template:
Environment="FCGI_CHILDREN={{ sympa.web.get(unit_name + '_procs') }}"
Lookups can be used for file contents, templates, environment variable and various other things.
Lookups are always executed on the controller.
Ansible will lookup the file in multiple directories if you are using a relative path. Run the playbook with -vvvvv to see the paths to these directories.
In this case the APT key is stored in the files directory of the role.
- name: Ensure key for Docker repository is imported
apt_key:
data: "{{ lookup('ansible.builtin.file', 'docker_apt_pgp.asc') }}"
id: 0EBFCD88
state: present
The file lookup removes newlines at the end of the file which breaks SSH private keys (https://github.com/ansible/ansible/issues/30829).
Here comes a neat trick to avoid it by utilizing YAML syntax:
- name: Add SSH private key
copy:
content: |
{{ lookup(''ansible.builtin.file', users_inventory_dir + '/files/ssh-keys/' + ssh_id ) }}
dest: "/home/{{ item.username }}/.ssh/id_rsa"
owner: "{{ item.username }}"
mode: 0600
no_log: true
If your file is located in the inventory, you can use the inventory_dir variable inside the lookup:
"{{ lookup(''ansible.builtin.file', inventory_dir + '/../files/ssh-keys/checkmk-ssh.pub') }}"
Note: you can use with_file as alternative to the file lookup, see authorized_key example.
Read a value from a vaulted YAML file:
"{% set secrets = lookup('ansible.builtin.unvault', 'vault/' + lookup('env', 'USER') + '-secrets.yml') | from_yaml %}{{ secrets.password }}"
This example shows how to determine the user on the controller:
"{{ lookup('ansible.builtin.env', 'USER') }}"
Query DNS A record for example.com:
"{{ lookup('community.general.dig', 'example.com.') }}"
Ouery DNS PTR record for example.com:
"{{ lookup(community.general.dig,'96.7.128.198/PTR') }}"
Specific name servers can be used as follows:
"{{ lookup('community.general.dig', 'example.com.', '@199.43.135.53,199.43.133.53' )}}"
Needs dnspython Python module.
DNS TXT records are used for various things. For example common email authentication methods like SPF, DKIM and DMARC are using TXT records.
You can use the dnstxt lookup plugin to check the DKIM record for the domain linuxia.de and selector mail:
- name: Check whether DKIM DNS entry exists
debug:
msg: "{{lookup('community.general.dnstxt', 'mail._domainkey.linuxia.de')}}"
The output looks like:
msg: v=DKIM1; p=MIGfMA0GCSqGSIb3DQEBAQUAA4GNADCBiQKBgQDywbZfxszCUIV3WsMWChd+8iergHXcsWNR/vPtc4gwtbR+3xWd9WNQzGuwhiCBjJcrC79yYnJYhw5vBvVY3YFgV4gD/V3gwWN4NqIb/LoNIgsqLZ8wy+cNnZJnT46K20Dmge1LAV5uxo9Mjq7xtbqTQ1ELI51rTCunos2noZ0JRQIDAQAB
Needs dnspython Python module.
Conditions determine whether a task is executed.
Arguments to conditions are automatically passed through templating. Don't wrap them in curly braces.
Examples:
Whether the key owner exist in the item dict:
when: "'owner' in item"
Whether list is not empty:
when: mylist | length
The traditional way for loops in Ansible using with_items, with_dict, with_subelements etc. has been replaced by loop: https://docs.ansible.com/ansible/latest/user_guide/playbooks_loops.html#migrating-from-with-x-to-loop. This has been introduced with Ansible 2.5. Of course you can still use the with_ loops.
Documentation for with_subelements: https://docs.ansible.com/ansible/latest/plugins/lookup/subelements.html
Data:
users:
- username: backuptransfer
ssh_keys:
- backup
ssh_keys_to_remove: []
Task:
- name: Add SSH keys
authorized_key:
user: "{{ item.0.username }}"
key: "{{ lookup('file', inventory_dir + '/files/ssh-keys/' + item.1 + '-ssh.pub') }}"
state: present
with_subelements:
- "{{ users }}"
- ssh_keys
loop: "{{ services.split(',') | map('trim') | list }}"
Handlers are only running if one of the triggering tasks is in the changed status. Also they are executed at the end of the playbook. So if multiple tasks triggering a handler, it is only run once.
A typical task executed by a handler is the restart of a service:
- name: Restart Exim4
service:
name: exim4
state: restarted
It is also possible to run handlers in the middle of playbook:
- name: Flush handlers meta: flush_handlers
With the ad-hoc mode you can execute arbitrary commands on the target hosts.
You can execute commands with ansible -a for all hosts in a group, e.g a simple call of the id program, which doesn't need superuser permissions:
$ ansible shopserver -i production -a id www1.linuxia.de | CHANGED | rc=0 >> uid=1001(ansible) gid=1001(ansible) Gruppen=1001(ansible),1002(sysadmin)
Add --become for commands which need to be executed as superuser:
$ ansible shopserver -i production --become -a "grep Failed.password /var/log/auth.log"
Use the pseudo group all to execute the command on all hosts in the given inventory.
You can run tasks for certain tags respective skip them.
ansible-playbook --tags=letsencrypt site.yml
ansible-playbook --skip-tags=letsencrypt site.yml
List all tasks which would be executed by the playbook site.yml.
ansible-playbook --tags=letsencrypt --list-tasks site.yml
List tasks tagged with letsencrypt which would be executed by the playbook site.yml.
ansible-playbook --tags=letsencrypt --list-tasks site.yml
Limit the execution of the playbook to one or more host groups:
ansible-playbook -i prod --limit myshop,mydb plays/ecommerce.yml
The location of the Vault password file can be set by:
–vault-password-file
ANSIBLE_VAULT_PASSWORD_FILE
vault_password_file
$ ansible-vault encrypt repo-git-id Encryption successful
$ ansible-vault encrypt_string --stdin-name mysql_root_password ...
The result looks as follows and you can put that as is into a YAML inventory file:
mysql_root_password: !vault | $ANSIBLE_VAULT;1.1;AES256 62373730376636323437633965356564366233326266663238343765353634646362323234663664 6239663063343861656234663035323765386366333632630a336363396562653437366339316538 37383639366532633839313961383465646164626638316238386239303264303463633936316539 3237356539366266640a306363393136656636346435303862653335333564326638336539313161 3134
Unfortunately Ansible doesn't offer a command to decrypt single variables in a YAML file.
Here the yq command comes to rescue.
Sample yq installation on Debian:
apt-get install jq pip install yq
Now you can select the variable with yp and output it as raw string instead of json (-r option):
yq -r .mysql_root_password production/production.yml
This output can be piped into the ansible-vault command as follows:
yq -r .mysql_root_password production/production.yml | ansible-vault decrypt
This also works for variables on deeper levels of the YAML file. Please note that some keys
with "special characters" like . and - needs to be quoted.
yq -r '."nginx-servers".hosts."example.linuxia.de".htpasswd_credentials[0].password' production/production.yml | ansible-vault decrypt
Check connection first:
$ ansible -m ping -i live DEBIAN
Exclude hardware from facts:
$ ansible -m setup -i live DEBIAN -a 'gather_subset=!hardware'
Ansible Docker Reference: https://docs.ansible.com/ansible/latest/scenario_guides/guide_docker.html
Ansible provides the fact ansible_selinux with information about the SELinux status. It is a dict with the fields config_mode, mode, policyvers, status and type.
Typical values are:
| Distribution / Release | Status | Mode | Type |
|---|---|---|---|
| Debian | disabled | - | - |
| Ubuntu | disabled | - | - |
| Centos | enabled | permissive | targeted |
| Fedora | enabled | enforcing | targeted |
However this fact requires installation of a SELinux Python library. For recent releases these are libselinux-python3 for RedHat OS family and python3-selinux for the Debian OS family.
You can refresh the ansible_selinux fact after the installation of the library with the setup module:
- name: Refresh SELinux fact
setup:
filter: 'ansible_selinux'
- name: Enable fcgid module
apache2_module:
name: fcgid
state: present
The apt module is used to install APT packages from repositories and from files.
To install multiple packages, pass a list to name:
- name: Install Perl packages needed by mrad
apt:
name:
- dar
- libappconfig-perl
- libdate-manip-perl
- libmime-lite-perl
- libfilesys-df-perl
Installing from a file requires to transfer the file to the target first:
- name: Copy Sympa package to target
copy:
src: "{{ sympa.package_file }}"
dest: /var/cache/apt/archives
Now you can install the package with the help of the deb parameter:
- name: Install Sympa package from a file
apt:
deb: /home/racke/sympa-community/sympa-6.2.48~dfsg-1_amd64.deb
Install a Debian package only on releases that comes with that package:
- name: Install certbot package(s)
apt:
name:
- certbot
when:
- ansible_distribution == 'Debian'
- ansible_distribution_major_version|int >= 10
Purging a package is triggered by the purge parameter:
- name: Purge Sympa package
apt:
name: sympa
state: absent
purge: yes
To update the APT cache without installing a package:
- name: Update APT cache
apt:
update_cache: yes
when: ansible_os_family == 'Debian'
The apt_key module manages APT keys for authenticating packages.
Example:
- name: Install APT key for MySQL repository
apt_key:
data: "{{ lookup('file', 'mysql-apt-repo.key') }}"
state: present
This requires gpg installed on the target machine:
- name: Install GnuPG
apt:
name:
- gpg
Add Ubuntu PPA repository:
- name: Add Linbit DRBD PPA repository
apt_repository:
repo: ppa:linbit/linbit-drbd9-stack
In most cases you will also need to update the APT cache when adding the repository.
Otherwise you wouldn't be able to install packages from the repository.
- name: Add Linbit DRBD PPA repository
apt_repository:
repo: ppa:linbit/linbit-drbd9-stack
codename: focal
register: drbd_linbit_apt
- name: Update APT cache
apt:
update_cache: yes
when: drbd_linbit_apt.changed
You can also select the filename for the APT sources file configuring the repository:
- name: Install apt repository for Gitlab
apt_repository:
repo: "deb https://packages.gitlab.com/gitlab/gitlab-ce/ubuntu/ focal main"
filename: gitlab-repo
The resulting file name would be /etc/apt/sources.list.d/gitlab-repo.list.
The assert module checks whether given conditions are met. It fails if not.
- name: Ensure that role is executed on Debian Jessie or Debian Stretch
assert:
that:
- ansible_distribution == 'Debian'
- ansible_distribution_major_version == '8' or ansible_distribution_major_version == '9'
msg: "Debian Jessie or Debian Stretch required for docker role"
The authorized_key module allows you to add or remove keys from user accounts.
- name: "Add SSH keys"
authorized_key:
user: interch
state: present
key: "{{ item }}"
with_file:
- racke-ssh.pub
- linuxia-ssh.pub
The blockinfile module is similar to the lineinfile module, but allows you to manipulate multiple lines in a file.
- name: Ensure that SSH port is in user's config
blockinfile:
path: "/home/{{ username }}/.ssh/config"
block: |
Host *.example.com
Port 7494
create: yes
mode: 0644
become_user: "{{ username }}"
If the file doesn't exist already, it be will created (because of create set to yes). In that case the content of the SSH config file would look like:
# BEGIN ANSIBLE MANAGED BLOCK Host *.example.com Port 7494 # END ANSIBLE MANAGED BLOCK
If you want to put multiple entries in a single file (either through separate tasks or by running the blockinfile task in a loop), you need to add an unique marker to the task, e.g.:
- name: Ensure that SSH port is in user's config
blockinfile:
path: "/home/{{ username }}/.ssh/config"
block: |
Host *.example.com
Port 7494
marker: "# {mark} ANSIBLE MANAGED BLOCK FOR EXAMPLE.COM"
create: yes
mode: 0644
become_user: "{{ username }}"
Now the output would look like:
# BEGIN ANSIBLE MANAGED BLOCK FOR EXAMPLE.COM Host *.example.com Port 7494 # END ANSIBLE MANAGED BLOCK FOR EXAMPLE.COM
The copy module copies files to the target.
The cpanm module allows you to manage Perl modules.
- name: Install Dancer2 module with cpanm
cpanm:
name: Dancer2
You can also specify a minimum version:
- name: Install Dancer2 module with cpanm
cpanm:
name: Dancer2
version: '0.301000'
It is not possible to use a list of Perl modules in the name parameter, so you need to run the task in a loop:
- name: Install Dancer2 plugins with cpanm
cpanm:
name: "{{ item }}"
loop:
- Dancer2::Plugin::Email
- Dancer2::Plugin::Auth::Extensible
- Dancer2::Plugin::GraphQL
The debconf module allows you to preseed values for the debconf configuration.
- name: Preseed Debconf values for Sympa (wwsympa_url)
debconf:
name: sympa
question: wwsympa/wwsympa_url
value: "https://{{ common.web.domain }}/sympa"
vtype: string
Start Elasticsearch container with image from GitHub:
- name: Ensure that Elasticsearch container is running
docker_container:
name: "esdemo"
image: "elasticsearch:7.10.1"
state: 'started'
restart_policy: always
You can also use registries other than Docker hub for the images, e.g. the Elasticsearch registry at https://www.docker.elastic.co/r/elasticsearch:
- name: Ensure that Open Source Elasticsearch container is running
docker_container:
name: "esdemo"
image: "docker.elastic.co/elasticsearch/elasticsearch-oss:7.10.1-amd64"
state: 'started'
restart_policy: always
The fetch module retrieves a file from a remote host.
This module doesn't support check mode. Recursive fetching is not supported.
Example:
- name: Upload ~/.gnupg archive to Ansible Controller
fetch:
src: '{{ reprepro_home + "/" + reprepro_gpg_snapshot_name }}'
dest: '{{ reprepro_gpg_snapshot_path + "/" + reprepro_gpg_snapshot_name }}'
flat: True
The file module manages files and directories on the target.
Create a directory:
- name: Ensure that installation directory exists
file:
state: directory
path: /usr/local/sympa
owner: sympa
group: sympa
mode: 0755
The find module locates files and directories. It supports a subset of the functionality of the Unix find command.
Example:
- name: Get list of available Apache modules
find:
file_type: 'file'
paths: '/etc/apache2/mods-available/'
patterns: '*.load'
register: apache_mods_available
You can specify multiple paths and patterns by passing a list to these parameters.
In order to extract all file paths from the result use the map filter:
- name: Paths for Apache modules
set_fact:
apache_mods_paths: "{{ apache_mods_available.files | map(attribute='path') | list }}"
This gives you a list of full paths:
- /etc/apache2/mods-available/fcgid.conf - /etc/apache2/mods-available/fcgid.load - /etc/apache2/mods-available/http2.conf - /etc/apache2/mods-available/http2.load
In order to get only the filenames, apply the basename filter as well:
- name: Filenames for Apache modules
set_fact:
apache_mods_files: "{{ apache_mods_available.files | map(attribute='path') | map('basename') | list }}"
Now the resulting list is:
- fcgid.conf - fcgid.load - http2.conf - http2.load
The get_url module download files to the target.
- name: Download Debian 10 image for KVM
get_url:
url: https://cdimage.debian.org/cdimage/openstack/current-10/debian-10-openstack-amd64.qcow2
checksum: sha256:85c43e90a13f5c1021afd07f843ace498b4bca4ff71b8e5c50d70e2566a304aa
dest: /var/lib/libvirt/images/debian10.qcow2
Determine the home directory of the remote user:
- name: Retrieve account information for remote user
getent:
database: passwd
key: "{{ ansible_ssh_user }}"
split: ":"
- name: Set fact for home directory
set_fact:
user_home: "{{ getent_passwd[ansible_ssh_user][4] }}"
The git module clones a Git repository.
Notable parameters are:
if set, creates a shallow clone
Updates Git configuration files, e.g user name and email:
- name: Configure Git username for etckeeper commits
git_config:
name: user.name
scope: global
value: "{{ git_user_name }}"
- name: Configure Git email for etckeeper commits
git_config:
name: user.email
scope: global
value: "{{ git_user_email }}"
The scope setting might be a bit confusing - global refers to the user's global configuration ~/.gitconfig, while system refers to /etc/gitconfig.
Creates a user group.
- name: Create unix group for Sympa
group:
name: sympa
The hostname module sets the hostname of the target system, e.g:
- name: Set hostname
hostname:
name: "{{ inventory_hostname_short }}"
The variable inventory_hostname_short contains the first part of the fully qualified domain name (FQDN), e.g. foo for foo.linuxia.de.
This adjusts the hostname in /etc/hostname and creates or updates
/etc/machine-info:
$ cat /etc/hostname foo $ cat /etc/machine-info PRETTY_HOSTNAME=foo
The import_role module loads the given role similar to roles specified with the roles: keyword.
With import_role Ansible checks first whether the role is available before executing the playbook.
The include_role loads a role dynamically.
Instead of executing the role as usual you can run a specific task file:
- name: Run task file dkim-key-pair.yml from exim-dkim role
import_role:
name: exim-dkim
tasks_from: dkim-key-pair.yml
The import_tasks module allows you to import tasks from another task file than main.yml:
- name: Install systemd timers import_tasks: timers.yml
import_tasks can not run in a loop.
This is useful to group related tasks and keep the main task file lean.
In addition you can restrict the scope to a subset of the targets, e.g. based on the distribution.
- name: Initialize PostgreSQL cluster import_tasks: initialize.yml when: ansible_os_family in ['Alpine', 'RedHat', 'FreeBSD', 'Suse']
Ansible provides the known_hosts module for adding or removing host keys from the ~/.ssh/known_hosts file.
- name: Add Git remote to known hosts
known_hosts:
name: 'git.linuxia.de'
key: '[git.linuxia.de] ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBG5jcaKO5t0A2NOxIPFFoBz6tKFUOZygTGZeBEpH9ib3wO98sWAdi8QtjKY6WpMIe5Z7ZvKC+awsvaBmzEfUVDE='
hash_host: yes
The ECDSA host key is located in the file /etc/ssh/ssh_host_ecdsa_key.pub on the server. Don't use the RSA host key, which might not accepted.
Another use case is a backup server which pulls backups from a number of clients with SSH.
The SSH public key of any client needs to be added to the ~/.ssh/known_hosts file of the server:
- name: Add known hosts entries for backup clients
known_hosts:
name: "{{ item }}"
key: "{{ item }}: {{ hostvars[item].ssh_host_key }}"
hash_host: yes
with_items: "{{ groups['backup_clients'] }}"
when:
- "'backup_servers' in group_names"
The lineinfile module add or updates a particular line in a text file.
It isn't applicable to the following use cases:
Replacing multiple occurrences of the same line (see replace)
Replacing a block with multiple lines blockinfile (see blockinfile)
Manipulating files with structured data such as XML and JSON
- name: Configure proxy (Debian, Ubuntu)
lineinfile:
dest: /etc/environment
regexp: '(?i)^no_proxy='
line: "no_proxy=example.com,localhost"
when: ansible_os_family == 'Debian'
This is an example where we add the report email address from the variable unattended_upgrades_report_email to the configuration of the unattended-upgrades Debian package:
- name: Add email address for reports on upgrades
lineinfile:
dest: "/etc/apt/apt.conf.d/50unattended-upgrades"
regexp: '^(//)?Unattended-Upgrade::Mail\s+"(.*)";'
line: "Unattended-Upgrade::Mail \"{{ unattended_upgrades_report_email }}\";"
The orignal line after installation looks like that:
//Unattended-Upgrade::Mail "root";
So the regular expression needs to cover both the option with or without the // at the beginning of the line.
- name: Adjust PHP FPM configuration
lineinfile:
path: "/etc/php-fpm.d/www.conf"
line: "catch_workers_output = yes"
state: present
regexp: "^;?catch_workers_output"
This makes sure that the line catch_workers_output = yes appears in the configuration file. It replaces an existing line by regular expression, so it would work whether the configuration directive is commented out or not:
catch_workers_output = no
;catch_workers_output = yes
In some cases it may be warranted to add multiple lines to a file by using lineinfile in a loop.
Inserting a list of cluster nodes into /etc/hosts could be such a case:
- name: Ensure that the node hostnames can be resolved (essential for DRBD/OCFS2 clustering)
lineinfile:
path: /etc/hosts
line: '{{ hostvars[item].ansible_default_ipv4.address }} {{ item }}'
regexp: '^{{ hostvars[item].ansible_default_ipv4.address | regex_escape }}'
insertbefore: '^$'
firstmatch: yes
loop: "{{ groups['mycluster'] }}"
The combination of insertbefore: '^$' and firstmatch: yes instructs
lineinfile to add the new lines before the first empty line in /etc/hosts.
The pristine file for host debian after initial installation may look like that:
127.0.0.1 localhost 127.0.1.1 debian.localdomain debian # The following lines are desirable for IPv6 capable hosts ::1 localhost ip6-localhost ip6-loopback ff02::1 ip6-allnodes ff02::2 ip6-allrouters
And after you added node1, node2, node3 from the group mycluster:
127.0.0.1 localhost 127.0.1.1 debian.localdomain debian 10.1.118.11 node1 10.1.118.12 node2 10.1.118.13 node3 # The following lines are desirable for IPv6 capable hosts ::1 localhost ip6-localhost ip6-loopback ff02::1 ip6-allnodes ff02::2 ip6-allrouters
- name: Update Python interpreter
ansible.builtin.lineinfile:
path: "{{ playbook_dir }}/host_vars/{{ inventory_hostname }}.yml"
regex: '^ansible_python_interpreter:'
line: "ansible_python_interpreter: /usr/bin/python3"
create: yes
delegate_to: localhost
become: false
lineinfile uses root user by default. If you want to write files as the user
on the controller change that with become: false.
Debian only enables the basic locales by default. You can use the locale_gen module to add needed locales:
- name: Ensure that German locale exists
locale_gen:
name: de_DE.UTF-8
state: present
Ansible provides the mysql_user module and the mysql_db module for creating MySQL roles and databases.
Create database:
- name: Create database for Sympa
mysql_db:
name: sympa
state: present
Create user for the Sympa database:
- name: Create MySQL user for Sympa
mysql_user:
name: sympa
priv: 'sympa.*:ALL'
password: nevairbe
state: present
Create remote user for replicating the Sympa database:
- name: Add remote replication user
mysql_user:
name: replicator
host: '10.0.2.20'
password: nevairbe
priv: "sympa:*:REPLICATION SLAVE"
state: present
Dump database:
- name: Create a dump from Sympa database
mysql_db:
name: sympa
state: dump
target: "/home/backup/sympa-{{ ansible_date_time.date }}.sql"
Please note that the dump file is world readable, so you need to protect sensitive data with restricted permissions for the dump directory.
- name: Update robots
community.mysql.mysql_query:
login_db: sympa
query: 'UPDATE list_table SET robot_list = "example.org"'
Determine MySQL slave status.
- name: Get slave status
mysql_replication:
mode: getslave
register: mysql_slave_info
- set_fact:
mysql_slave_health: |-
{% if mysql_slave_info.Slave_IO_Running == 'Yes' and mysql_slave_info.Slave_SQL_Running == 'Yes' %}UP{% else %}DOWN{% endif %}
The package module is a generic package manager for supported distributions.
Simple tasks like installing Git can be accomplished with it, but software might be packaged with different names in the distributions.
- name: Ensure that Git is installed
package:
name:
- git
Specific modules by distributions are:
| Distributions | Ansible module |
| Alpine | apk |
| ArchLinux | pacman |
| Debian | apt |
| Gentoo | portage |
| RedHat | yum |
| Suse/SLES | zypper |
| Ubuntu | apt |
This module changes the Linux PAM limits in the file
/etc/security/limits.conf (or the file specified by the dest parameter).
To impose soft and hard limits for the number of open files to user foo:
- name: ulimits - Add soft value of maximum open files to user foo
pam_limits:
domain: foo
limit_type: soft
limit_item: nofile
value: "1024"
- name: ulimits - Add hard value of maximum open files to user foo
pam_limits:
domain: foo
limit_type: hard
limit_item: nofile
value: "2048"
Please note that this module doesn't indicate possible changes in check mode.
Available limit types are:
| core | limits the core file size (KB) |
| data | max data size (KB) |
| fsize | maximum filesize (KB) |
| memlock | max locked-in-memory address space (KB) |
| nofile | max number of open file descriptors |
| rss | max resident set size (KB) |
| stack | max stack size (KB) |
| cpu | max CPU time (MIN) |
| nproc | max number of processes |
| as | address space limit (KB) |
| maxlogins | max number of logins for this user |
| maxsyslogins | max number of logins on the system |
| priority | the priority to run user process with |
| locks | max number of file locks the user can hold |
| sigpending | max number of pending signals |
| msgqueue | max memory used by POSIX message queues (bytes) |
| nice | max nice priority allowed to raise to values: [-20, 19] |
| rtprio | max realtime priority |
| chroot | change root to directory (Debian-specific) |
The pip module install and removes Python packages.
- name: Ensure that SSLyze is installed
pip:
name: sslyze
The portage module manages packages on Gentoo through the emerge command.
In general the "packages" are compiled from sources, so it can take a long time to install a Gentoo package.
- name: Ensure that Git is installed
portage:
name:
- git
Ansible provides the postgresql_user module and the postgresql_db module for creating PostgreSQL roles and databases.
Creating a role can be as simple as in the following example:
- name: Create PostgreSQL role for Sympa
postgresql_user:
name: sympa
Now you can create a database for this role:
- name: Create PostgreSQL database for Sympa
postgresql_db:
name: sympa
encoding: UTF-8
lc_collate: en_US.UTF-8
lc_ctype: en_US.UTF-8
template: template0
owner: sympa
state: present
You can ensure that a column exists in a table with using this module:
- name: Add unsubscribe_link field to subscriber table
postgresql_query:
db: sympa
query: 'ALTER TABLE subscriber_table ADD COLUMN IF NOT EXISTS
unsubscribe_link text'
become: true
become_user: postgres
vars:
ansible_ssh_pipelining: true
The reboot module reboots the target host.
The output for a successful reboot looks like that:
ok: [buster-test-box] =>
msg:
changed: true
elapsed: 30
failed: false
rebooted: true
The replace module replace all instances of a pattern within a file.
E.g. if you want to prepend every line in a file with #:
- replace:
path: "/home/racke/example.txt"
regexp: '^(.*)$'
replace: '# \1'
The service module start and stops services.
On hosts with systemd, the task is delegated to the systemd module.
Services are not always enabled and started when you install the corresponding package, so add a task to ensure that this is the case:
- name: Ensure that rsyslog is installed
package:
name: rsyslog
- name: Ensure that rsyslog is enabled and running
service:
name: rsyslog
state: started
enabled: yes
Some services have stop scripts that terminate immediately even before the actual process is vanished. This can prevent a successful start when you restart the service.
To alleviate this problem you can add a value for the sleep parameter:
- name: restart mysql
service:
name: {{ mysql_daemon }}
state: restarted
sleep: 5
Disable SELinux:
- name: Disable SELinux (Fedora)
selinux:
state: disabled
when: ansible_distribution == 'Fedora'
The setup module allows you to gather facts on the remote hosts. This is done automatically at the beginning of a play unless you disable it with setting gather_facts to False.
In some cases you might need to refresh the facts. You can find one example in the SELinux section.
Another one follows:
# Vagrant box debian/buster64 has "buster/sid" in /etc/debian_version
- name: Ensure that base-files package is up-to-date for the Debian buster assertion
apt:
name:
- base-files
state: latest
register: basefiles
- name: Reread facts
setup:
when:
basefiles.changed
Installing base-files changes the content of /etc/debian_version from buster/sid to 10.0,
but this is not reflected in the ansible_distribution_major_version variable.
Without refreshing the facts the following assertion would fail in a fresh buster VM:
- name: Role requires Debian Buster
assert:
that:
- ansible_distribution == 'Debian'
- ansible_distribution_major_version == '10'
fail_msg: "Distribution {{ ansible_distribution }}, major version: {{ ansible_distribution_major_version }}"
You sometimes need to find out the status of the services, especially in a heterogeneous environment. The service_facts sets a dict with all existing services, whether they are running or not.
So you can use that to determine whether the service exists at all and and is in a certain state:
- name: Populate service facts
ansible.builtin.service_facts:
- name: Adjust firewalld settings
import_role:
name: firewalld
when:
- "'firewalld.service' in ansible_facts.services"
- ansible_facts.services['firewalld.service']['state'] == 'running'
Reference: synchronize module
This can be used to synchronize backups from server A (sync_src_server) to server B (sync_dest_server). Both need to be part of the inventory:
- name: Synchronize
synchronize:
mode: push
src: "/data/backups/"
dest: "/srv/backups/server-a/"
delegate_to: "{{ sync_src_server }}"
when: inventory_hostname == sync_dest_server
Ensure that rsync is installed on both hosts:
- name: Ensure that rsync is installed on both hosts
package:
name: rsync
when: inventory_hostname in [sync_src_server, sync_dest_server]
The systemd module is the similar to the service module, but is specifically used to manage systemd services.
After changing an unit file, let systemd know about it:
- name: Reload systemd daemon on unit file changes
systemd:
daemon_reload: true
This is the equivalent to the systemctl daemon-reload command.
The template module generates files from templates.
The template needs to exists on the Ansible controller. There is remote_src parameter as for the copy module.
- name: Set timezone to Europe/Berlin
timezone:
name: Europe/Berlin
Manages ufw firewall rules.
- name: Open firewall port for IMAP/TLS
ufw:
rule: allow
port: '993'
proto: tcp
The unarchive module unpacks an archive.
Please make sure that the file permissions are correct inside the archive. Although there is a mode parameter, it is pretty much useless for software archives as the same mode is applied to files and directories.
As the unzip binary is not always installed by default (e.g. Debian), make sure it is installed first when dealing with zip archives.
For example:
- name: Install unzip
apt:
name:
- unzip
when:
- software_archive is match('.*\.zip')
Alpine Linux uses tar provided by busybox, which is not sufficient to unpack tar archives with this module.
Install GNU tar on Alpine Linux:
- name: Install GNU tar on Alpine Linux
apk:
name: tar
FreeBSD has a similar problem:
fatal: [freebsd-test-box]: FAILED! => changed=false msg: Failed to find handler for "/usr/local/src/sympa/sympa-6.2.59b.1.tar.gz". Make sure the required command to extract the file is installed. Command "/usr/bin/unzip" could not handle archive. Command "/usr/bin/tar" detected as tar type bsd. GNU tar required.
Install GNU tar on FreeBSD:
- name: Install GNU tar on FreeBSD
pkgng:
name: gtar
The uri module interacts with web services.
Example:
- name: Create schema fields
uri:
url: "http://localhost:{{ solr_port }}/solr/{{ solr_core }}/schema"
method: "POST"
body_format: json
body: "{{ { 'add-field' : solr_rest_fields} | to_json }}"
It can also be used to test correctness of web server configuration:
- name: Test redirection of / to /sympa
uri:
url: "https://lists.linuxia.de"
register: sympa_redirection
failed_when: not sympa_redirection.redirected or sympa_redirection.url != "https://lists.linuxia.de/sympa"
HTTP headers can be added as dict:
- name: Trigger list overview
uri:
url: "https://lists.linuxia.de/sympa/lists"
method: "GET"
headers:
referer: "https://lists.linuxia.de/sympa"
The user module creates users on the target system.
Create an user:
- name: Create unix user for Sympa
user:
name: sympa
group: sympa
shell: /bin/bash
createhome: yes
password_lock: yes
The xml module manipulates XML files and strings.
The zypper module manages packages on Suse distributions (OpenSUSE and SLES).
The syntax is similar to the package module.
The zypper_repository module manages repositories on Suse distributions (OpenSUSE and SLES).
The location of the parent directory for the roles can be configured through the roles_path variable in the Ansible configuration file:
[defaults] roles_path = roles
ansible-galaxy commandInstall role from Github repository:
$ ansible-galaxy roles install git+https://github.com/racke/ansible-role-clamav.git Starting galaxy role install process - extracting ansible-role-clamav to /home/racke/provisioning/linuxia-inventory/roles/ansible-role-clamav - ansible-role-clamav was installed successfully
Roles from Ansible Galaxy can be integrated as follows:
Eintrag in roles/requirements.yml z.B.
- src: geerlingguy.memcached version: "1.0.8"
Ansible Galaxy Installation:
ansible-galaxy install -p roles -r roles/requirements.yml
Git commit, e.g.:
git add roles/requirements.yml roles/geerlingguy.memcached git commit -m "Add external role for installing memcached."
Git tag, e.g.:
git tag -a -s -m "Role geerlingguy.memcached, version 1.0.8" ROLE_GEERLINGGUY_MEMCACHED_1_0_8
If you want to upgrade that role later, you need to bump up the version in roles/requirements.yml and rerun the ansible-galaxy command with the -f or --force flag:
ansible-galaxy install -f -p roles -r roles/requirements.yml
It is quite useful to wrap external roles into your own role with the import_role module. The following role demonstrates that:
Ensure that the default password has been overridden
Import the role with variables specific for your servers
Execute additional tasks (here we install backup package)
- name: Assert that default root password has been changed
assert:
that: "mysql_root_password != 'root'"
fail_msg: 'Default value for mysql_root_password'
- name: Run external MySQL role
import_role:
name: geerlingguy.mysql
vars:
mysql_packages:
- mariadb-server
- mariadb-client
- name: Install automysqlbackup
apt:
name: automysqlbackup
You can also check in the wrapping role that the tasks in the imported role are not executed if a certain condition is met. For an example, it doesn't make sense to install htpasswd when you don't have credentials configured:
- name: Run external HTTP basic authentication role
import_role:
name: geerlingguy.htpasswd
when:
- htpasswd_credentials
Note: the role will be still imported, but all tasks are going to be skipped.
If you need the variables from a role without actually executing the role, you can use the following workaround:
roles:
# including sympa role for defaults, but skipping it
- role: sympa
when: false
With this role we install a check_mk agent on the system, which runs as systemd service instead of as xinetd.
The agent is accessed remotely by SSH from the monitoring host, specified by the variables omd_url and omd_version.
roles/monitoring/tasks/main.yml
---
- name: Install check-mk-agent
apt:
deb: "{{ omd_url }}/check_mk/agents/check-mk-agent_{{ omd_version }}-1_all.deb"
tags:
- monitoring
- name: Enable and start systemd socket
systemd:
name: "check_mk.socket"
enabled: yes
state: started
tags:
- monitoring
- name: Add SSH key for monitoring
authorized_key:
user: root
key: "{{ lookup('file', 'checkmk-ssh.pub') }}"
key_options: 'command="/usr/bin/check_mk_agent"'
state: present
Sample playbook:
---
- hosts: all
vars:
omd_version: "1.5.0p7"
omd_url: "https://monitor.linuxia.de/omd"
roles:
- monitoring