fuel-web/docs/ci/overview.rst
Jeremy Stanley 0d140fac69 Replace ci.o.o links with docs.o.o/infra
The http://ci.openstack.org/ documentation site has been deprecated,
replaced by redirects to corresponding paths within
http://docs.openstack.org/infra/ where other Project Infrastructure
documentation already resides.

Change-Id: I5531741211dbd5917ec9ffeb19acd5210f968de7
2015-05-14 21:38:20 +00:00

191 lines
6.5 KiB
ReStructuredText

Fuel Infrastructure
===================
Overview
--------
Fuel Infrastructure is the set of systems (servers and services) which provide
the following functionality:
* automatic tests for every patchset commited to Fuel Gerrit repositories,
* Fuel nightly builds,
* regular integration tests,
* custom builds and custom tests,
* release management and publishing,
* small helper subsystems like common ZNC-bouncer, status pages and so on.
Fuel Infrastructure servers are managed by Puppet from one Puppet Master node.
To add new server to the infrastructure you can either take any server with base
Ubuntu 14.04 installed and connect it to the Puppet Master via puppet agent, or
you can first set up the PXE-server with PXETool :ref:`pxe-tool` and then run server
provisioning in automated way.
Puppet
------
.. _pxe-tool:
Puppet Master
~~~~~~~~~~~~~
Puppet deployment with Fuel Infra manifests requires Puppet Master.
To install Puppet Master to the brand new server:
#. Get required repository with Puppet configuration:
::
git clone ssh://(gerrit_user_name)@review.fuel-infra.org:29418/fuel-infra/puppet-manifests
#. Create a script to update puppet server to current version:
::
#!/bin/bash
REPO='/home/user/puppet-manifests'
SERVER='pxetool'
# sync repo files
rsync -av --delete $REPO/modules/ root@$SERVER:/etc/puppet/modules/
rsync -av --delete $REPO/manifests/ root@$SERVER:/etc/puppet/manifests/
rsync -av --delete $REPO/bin/ root@$SERVER:/etc/puppet/bin/
rsync -av --delete $REPO/hiera/ root@$SERVER:/etc/puppet/hiera/
# create symlinks
ssh root@$SERVER ln -s /etc/puppet/hiera/common-example.yaml /var/lib/hiera/common.yaml 2> /dev/null
#. Install base Ubuntu 14.04 with SSH server running (set hostname to pxetool.test.local).
Download and install Puppet Agent package.
#. Run previously created script on your workstation to push configuration and scripts
to new server. Run /etc/puppet/bin/install_puppet_master.sh as root on new server.
The last script does the following:
* upgrades all packages on the system
* installs required modules
* installs puppet and Puppet Master packages
* runs puppet apply to setup Puppet Master
* runs puppet agent to do a second pass and verify installation is usable
.. note:: Puppet manifests take data (passwords, keys or configuration
parameters) from hiera configuration. To work with our predefined test data
script links ``hiera/common-example.yaml`` file to
``/var/lib/hiera/common.yaml``. This step must be done before running
``bin/install_puppet_master.sh``.
Once done, you can start deploying other nodes.
Nodes and Roles
~~~~~~~~~~~~~~~
Currently, we define server roles via our hiera configuration (ssh://review.fuel-infra.org:29418/fuel-infra/puppet-manifests) using facter value ``ROLE``. Several roles are defined in ``hiera/roles``:
* anon_stat.yaml
* gerrit.yaml
* glusterfs_testing.yaml
* jenkins_master.yaml
* jenkins_slave.yaml
* lab.yaml
* mongo_testing.yaml
* nailgun_demo.yaml
* puppetmaster.yaml
* seed.yaml
* tools.yaml
* tracker.yaml
* web.yaml
* zbxproxy.yaml
* zbxserver.yaml
* znc.yaml
The most of roles are self explainable.
Generic Node Installation
~~~~~~~~~~~~~~~~~~~~~~~~~
Follow these steps to deploy chosen node:
* install base Ubuntu 14.04
* install puppetlabs agent package
* append to ``/etc/hosts`` - ``<Puppet Master IP> puppet pxetool.test.local``
* run ``FACTER_ROLE=<role name> puppet agent --test`` to apply configuration
Jenkins
-------
Our Jenkins instances are configured to run in master-slave mode. We have
Jenkins master instance on a virtual machine and a number of hardware nodes
working as Jenkins slaves.
Jenkins slaves setup
~~~~~~~~~~~~~~~~~~~~
There are several ways to setup Jenkins master-slave connection, and we use two
of them. The first one is organized simply by putting Jenkins master SSH-key in
authorized_keys file for jenkins user on a slave machine. Then you go to the
Jenkins Web UI and create node manually by specifying node IP address. Jenkins
master connects to the slave via SSH, downloads slave.jar file and runs jenkins
process on a slave.
The second approach requires more configuration steps to take, but allows you to
create slave node automatically from a slave node itself. To use it you need:
* install Swarm Plugin on Jenkins master
* create Jenkins user with ability to create nodes
* install jenkins-swarm-slave package on the slave
* configure the slave to use the mentioned Jenkins user
* run jenkins-swarm-slave service on the slave
Service will automatically connect to Jenkins master and create a node with proper
name and IP address.
Though this approach seems to be complicated, it is quite easy to implement it
with Puppet, as we do in jenkins::slave Puppet class (defined in
puppet-manifests/modules/jenkins/manifests/slave.pp).
If you use Gerrit slave with HTTPs support (default hiera value), please also
include jenkins::swarm_slave as it will trust Jenkins Master certificate on
Node side.
The downside of the swarm slave plugin is that every time you reboot Jenkins
master instance, slaves are recreated and, therefore, lose all the labels
assigned to them via Jenkins WebUI.
Gerrit
------
Although fuel-* repositories are hosted by the `OpenStack Gerrit <http://review.openstack.org>`_,
we use additional Gerrit instance to host OpenStack packages, internal projects and all the code
related to Infrastructure itself.
Our Gerrit instance is installed and configured by Puppet, including specifying
the exact Java WAR file that is used(link). To manage Gerrit instance we use
`Jeepyb <http://docs.openstack.org/infra/system-config/jeepyb.html>`_ - the tool written by Openstack Infra
team, which allows to store projects configuration in YAML format.
To use Jeepyb with gerrit you need to create "projects.yaml" configuration file,
where for each project you add the following information:
* project name
* project description
* project ACL
* project upstream
If "upstream" option is specified, Jeepyb will automaticaly import the upstream
repository to this new project. To apply the configuration, use "manage-projects" command.
Every project has ACL file. One ACL file can be reused in several projects. In
ACL file, access rights are defined based on the Gerrit user groups.
For example, in this file you can allow certain group to use the Code-Review
+/-2 marks.
In our gerrit, we have some global projects - <projects>/. The Core Reviewers
for these projects are <one-core-group>.
Contributing
~~~~~~~~~~~~
Feedback
~~~~~~~~