collect-logs: run in two stages

We should run and publish a minimal sets of files even without any host,
when we failed the run before inventory generation.

This change separates the collection step that runs on all hosts except
localhost, and the rest running on localhost. Running on localhost
always succeeds, even with an empty inventory.

Also add a log environment file for local collect-logs.sh runs that
does not upload logs.

Change-Id: I48d07d42be879026fb80afd73835484770006f85
This commit is contained in:
Attila Darazs 2016-12-02 16:12:24 +01:00
parent 1a3897c326
commit 2bb8b177bc
6 changed files with 106 additions and 97 deletions

View File

@ -0,0 +1,6 @@
---
# this file is intentionally left blank, it can be used to pass the "local"
# parameter to the collect-logs.sh CI script, resulting in using the default
# settings, collecting logs locally but not attempting to upload them anywhere
# we need at least one dummy value here
artcl_placeholder: true

View File

@ -1,7 +1,11 @@
---
- name: Collect logs, create docs, publish
- name: Collect logs
hosts: all:!localhost
gather_facts: no
roles:
- role: collect-logs
- collect-logs
- name: Create docs, publish logs
hosts: localhost
roles:
- {role: collect-logs, artcl_collect: false }

View File

@ -1,12 +1,13 @@
---
- name: Ensure required rpms for logging are installed
- become: true
ignore_errors: true
block:
- name: Ensure required rpms for logging are installed
yum: name={{ item }} state=present
with_flattened:
- gzip
- tar
- block:
- name: Prepare directory with extra logs
file: dest=/var/log/extra state=directory
@ -138,20 +139,20 @@
file:
path: "/tmp/{{ inventory_hostname }}"
state: absent
become: yes
become_user: root
ignore_errors: true
- name: Set default collect list
set_fact: collect_list="{{ artcl_collect_list | join(' ') }}"
set_fact:
collect_list: "{{ artcl_collect_list | join(' ') }}"
- name: Override collect list
set_fact: collect_list="{{ artcl_collect_override[inventory_hostname] | join(' ') }}"
when: artcl_collect_override is defined and artcl_collect_override[inventory_hostname] is defined
set_fact:
collect_list: "{{ artcl_collect_override[inventory_hostname] | join(' ') }}"
when:
- artcl_collect_override is defined
- artcl_collect_override[inventory_hostname] is defined
- name: Gather the logs to /tmp
become: yes
become_user: root
shell: >
mkdir -p /tmp/{{ inventory_hostname }};
for F in $(ls -d1 /var/log/rpm.list /var/log/extra {{ collect_list }}); do
@ -197,21 +198,15 @@
state: absent
ignore_errors: true
- block:
- delegate_to: localhost
when: artcl_gzip_only|bool
block:
- name: Extract the logs
shell: >
chdir={{ artcl_collect_dir }}
tar xf {{ inventory_hostname }}.tar;
when: artcl_gzip_only|bool
- name: delete the tar file after extraction
file:
path: "{{ artcl_collect_dir }}/{{ inventory_hostname }}.tar"
state: absent
when: artcl_gzip_only|bool
- name: fetch and gzip the console log
shell: >
curl {{ lookup('env', 'BUILD_URL') }}/consoleText | gzip > {{ artcl_collect_dir }}/console.txt.gz
when: artcl_publish|bool and "{{ lookup('env', 'BUILD_URL') }}" != ""
delegate_to: localhost

View File

@ -1,19 +1,17 @@
---
- block:
- name: Ensure required python packages are installed
- name: Ensure required python packages are installed
pip:
requirements: "{{ local_working_dir }}/usr/local/share/ansible/roles/collect-logs/doc-requirements.txt"
- name: Unarchive shell scripts
- name: Unarchive shell scripts
shell: >
gunzip "{{ artcl_collect_dir }}/undercloud/home/stack/{{ item }}.sh.gz";
with_items: "{{ artcl_create_docs_payload.included_deployment_scripts }}"
ignore_errors: yes
when: artcl_gzip_only|bool
- name: Generate rST docs from scripts and move to Sphinx src dir
- name: Generate rST docs from scripts and move to Sphinx src dir
shell: >
awk -f "{{ local_working_dir }}/usr/local/share/ansible/roles/collect-logs/scripts/doc_extrapolation.awk" \
"{{ artcl_collect_dir }}/undercloud/home/stack/{{ item }}.sh" > \
@ -21,23 +19,19 @@
with_items: "{{ artcl_create_docs_payload.included_deployment_scripts }}"
ignore_errors: yes
- name: Fetch static rST docs to include in output docs
- name: Fetch static rST docs to include in output docs
shell: >
cp "{{ artcl_docs_source_dir }}/../static/{{ item }}.rst" "{{ artcl_docs_source_dir }}"
with_items: "{{ artcl_create_docs_payload.included_static_docs }}"
ignore_errors: yes
- name: Generate fresh index.rst for Sphinx
- name: Generate fresh index.rst for Sphinx
template:
src: index.rst.j2
dest: "{{ artcl_docs_source_dir }}/index.rst"
force: yes
- name: Build docs with Sphinx
- name: Build docs with Sphinx
shell: >
sphinx-build -b html "{{ artcl_docs_source_dir }}" \
"{{ artcl_docs_build_dir }}" &> "{{ local_working_dir }}/sphinx_build.log"
delegate_to: localhost
run_once: true
when: artcl_gen_docs|bool

View File

@ -5,8 +5,12 @@
- name: Generate docs
include: create-docs.yml
when: artcl_gen_docs|bool
when:
- artcl_gen_docs|bool
- not artcl_collect|bool
- name: Publish logs
include: publish.yml
when: artcl_publish|bool
when:
- artcl_publish|bool
- not artcl_collect|bool

View File

@ -1,15 +1,21 @@
---
- name: Ensure the collection directory exists
file:
path: "{{ artcl_collect_dir }}"
state: directory
- run_once: true
delegate_to: localhost
block:
- name: upload to the artifact server using pubkey auth
- name: fetch and gzip the console log
shell: >
curl {{ lookup('env', 'BUILD_URL') }}/consoleText | gzip > {{ artcl_collect_dir }}/console.txt.gz
when: lookup('env', 'BUILD_URL') != ""
- name: upload to the artifact server using pubkey auth
command: rsync -av --quiet -e "ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null" {{ artcl_collect_dir }}/ {{ artcl_rsync_path }}/{{ lookup('env', 'BUILD_TAG') }}
retries: 5
delay: 60
when: artcl_use_rsync|bool and not artcl_rsync_use_daemon|bool
- name: upload to the artifact server using password auth
- name: upload to the artifact server using password auth
environment:
RSYNC_PASSWORD: "{{ lookup('env', 'RSYNC_PASSWORD') }}"
command: rsync -av --quiet {{ artcl_collect_dir }}/ {{ artcl_rsync_path }}/{{ lookup('env', 'BUILD_TAG') }}
@ -17,7 +23,7 @@
delay: 60
when: artcl_use_rsync|bool and artcl_rsync_use_daemon|bool
- name: upload to swift based artifact server
- name: upload to swift based artifact server
shell: swift upload --quiet --header "X-Delete-After:{{ artcl_swift_delete_after }}" {{ artcl_swift_container }}/{{ lookup('env', 'BUILD_TAG') }} *
args:
chdir: "{{ artcl_collect_dir }}"
@ -29,11 +35,11 @@
OS_TENANT_NAME: "{{ artcl_swift_tenant_name }}"
when: artcl_use_swift|bool
- name: use zuul_swift_upload.py to publish the files
- name: use zuul_swift_upload.py to publish the files
command: "{{ artcl_zuul_swift_upload_path }}/zuul_swift_upload.py --name {{ artcl_swift_container }} --delete-after {{ artcl_swift_delete_after }} {{ artcl_collect_dir }}"
when: artcl_use_zuul_swift_upload|bool
- name: create the artifact location redirect file
- name: create the artifact location redirect file
template:
src: full_logs.html.j2
dest: "{{ artcl_collect_dir }}/full_logs.html"