OSIC appears to not be able to create nodes right now, pull it from
the pool to let patch merges flow again.
Change-Id: Ia14b53b9e0ccd3b2e0ba97b0b6d156f4dea96fe4
As part of resource restructuring within OSIC, we're ceasing use of
512 nodes with their "disk" flavor, and continuing to use 256 each
of their "s3500" and "s3700" SSD-backed flavors. Independent of
this, they will be adjusting these flavors from 8vCPU to 4vCPU so
that we can bring our peak utilization in line with their current
availability.
This is expected to be a temporary measure, until they're able to
complete deployment of a dedicated 35-host OpenStack environment
where we can start adjusting overcommit ratios (hopefully driving
our cap back up again somewhat).
Change-Id: I0244362a61264bb36e5c1e1562684a0136944e49
If you specify a CUSTOM_PROJECTS_LIST_URL that doesn't exist you get a
message but the script continues. Since you clearly meant to specify
something we should bail so the user can correct the problem.
Additionally, add some documentation to the build-image.sh script to
describe how to use this to cut down build times considerably for
initial testing purposes.
Change-Id: I02d1cd341f793471b1ad4374e98031ea53db7f49
Because we want to share infracloud-chocolate between nodepool.o.o and
nl01.o.o, we need to use the new 'nodepool-id' for the provider.
Otherwise, nodepool.o.o will proceed to delete the instances that
nl01.o.o launches.
Change-Id: Ic5d1f0ebf7eb92c841ec09502e5134d89c8f6e97
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
stackviz: In general newer openSUSE distributions switched to nodejs6 LTS,
but Leap 42.2 is still on 4.x.
zuul-worker: No need for libselinux enablement in openSUSE
puppet: cleanup the puppet repo that got added in the install_puppet.sh
code part
Change-Id: If7bf3c799ebb74aeb36c4b1b607b0454993f0ed6
Now that the Pike PTG is over, bring infracloud-chocolate back online
for production. We do this, to avoid nodepool-builder issues (we are
not deleting diskimages from disk).
Moving forward, we'll need to update nodepool to properly support 2
nodepool launchers using a shared tenant.
Change-Id: I7fd90939db9fcb0edc1afc4f62248801f3c20ca5
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
Because we are using the 'zuul' user to SSH into our zuul workers, we
need to add it to configure_mirror.sh and still support the jenkins
user.
Change-Id: I79eff670f56d06c49a16e1330a73889e7b8482f1
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
Disable our ready-scripts, as we have a permission issue with the zuul
user.
Replace jenkins user with zuul user.
Switch our infracloud provider back to the original, otherwise we
cannot find an image to launch.
Change-Id: Ib3d6cf45308be3efdf2aa6c855b26c85b1db3804
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
We'll be using chocolate for nodepool+zuulv3 dev during ptg.
We don't want to just set max-servers to 0 because nodepool will
still attempt to manage and clean up servers and get in the way
of our v3 deployment.
Change-Id: If090190b541922ece7e67fa9107e521c050fa484
osic-cloud1 is back to normal.
This reverts commit bc22cb207d5e32895c3a76098be8ef3f525962a0.
Change-Id: I48ed5449431c9d251722d8c1f5152b0c9c8fad4a
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
We are seeing a large increase in failures with osic-cloud1, for the
moment lets disable the cloud so our other providers can absorb the
load.
Once we know the issue (looks to be networking) has been resolved, we
can revert this patch.
Change-Id: If875faa3ac7aa71020ed16f28ee702fc00d67ce2
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
Currently, we are seeing a large influx of data being uploaded to
logs.o.o. As a result, we are decreasing our retention policy on
logs.o.o but cannot say a head of the curve.
Until we have reclaimed HDD space, disable tripleo-test-cloud-rh1.
Change-Id: I4d7dc07552662ddd72aef74cef9650e6e06f65d3
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
The move to 65 has gone smoothly, and we should have capacity for
even more jobs. We could probably go straight to 75 (which is the
most we should run with the current te-broker setup), but larger
jumps in the number of testenvs have caused problems in the past so
I'd prefer to stick with 5 at a time. Note that 5 extra testenvs
will use more than 150 GB of memory, so this increase is not as
small as it looks.
Change-Id: If83b507ba6ca97fc850d3a9caf570c4950949914
Now that we have migrated all our jobs to fedora-25, we can remove
fedora-24 from project-config.
Change-Id: I3a58c785c12f779060a6dca5cd40ddcc46429971
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
We've made a number of changes recently in tripleo-ci so we're
making more efficient use of the rh1 cloud and should now have the
capacity to run more jobs at once.
Change-Id: I99cb92d7ada4b89bcac62dfc24846c04b03eb9f5
There is an issue when we restart zuul, that gearman information about
said nodes is lost. Revert for now, and hopefully address this in
zuulv3.
This reverts commit 23495050235dd80331e256782f6db6a85e577e22.
Change-Id: I32d331f065be041cd11c71c11702be93b3d42831
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
The people who manage our DNS broke it. Sadly we need to disable the
environment until we can get whatever they did undone.
Change-Id: I2427c84e7f0d980d6fb8aeba5f25be7c693319b2
Signed-off-by: Kevin Carter <kevin.carter@rackspace.com>
Now that fedora-25 is online, we can start using the new AFS mirror for
fedora. Keep in mind, we didn't mirror fedora-24, since fedora-25 will
be replacing it.
Once fedora-24 has been removed from nodepool.yaml, we can remove our
conditional check.
Change-Id: I1bd510c6a568b71b604c27d6cae16d2654063d0c
Depends-On: I762e55fb7fe0c5ebfb7810545ddb4d5ccfedf2ff
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
This is cosmetic change for nodepool. Diskimages will now be uploaded
as ubuntu-xenial-1484177984 instead of
template-ubuntu-xenial-1484177984.
Change-Id: Ib4ced36ff8ae5a7e99ce5ecedac16b2dd36fa6a3
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
Bring fedora-25 online with a 1 min-ready node. This is needed to
start the migration from fedora-24 to fedora-25.
Change-Id: I99b852d3f440814fcf2f2a7518c413cdeeeb4106
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
Switch back to master branch, until we are ready to run zuulv3 branch
again in production.
Change-Id: Id93a3c64282a31545a66dc1da740596d3db3909f
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
While we have jobs using these labels today, they are still used very
little. Rather then having 6 nodes idle for days at a time, only
launch them when jobs are needed.
On average it takes about 30seconds for nodes to come online these
days.
Change-Id: I8e74b62f3556823e17468a4e2d22d42490fc44e6
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
Our AFS mirror is back online. It looks like the compute server lost
network access on Jan 7, the solution was to reboot the server in
ironic.
This reverts commit 06ebd8a99e56ca586059a42f738b6b9fef921922.
Change-Id: I45587358e40d15bd8bbd660b0093096bdfb95c0c
Signed-off-by: Paul Belanger <pabelanger@redhat.com>
The region is currently failing tests since the mirror is down. Disable
cloud until everything is fixed.
Change-Id: I22a619e62067b47900bb6db7d01eedfdceaad3cf
TripleO Architectures allows complex scenarios where we can choose which
services (run with Pacemaker in HA) live on which node.
http://specs.openstack.org/openstack/tripleo-specs/specs/ocata/composable-ha-architecture.html
To test this architecture, we need a 4-nodes deployment where:
- 1 node is for undercloud
- 3 nodes are for Pacemaker cluster with composable services
scenario005 will deploy this complex architecture and makes sure
OpenStack Cloud is working well, like we do with other scenarios.
This patch aims to:
- introduce 4nodes scenario in OpenStack Nodepool.
- introduce scenario005, in experimental pipeline
- configure scenario005 to run only when needed:
- check experimental
- only starting from ocata
- only when we patch pacemaker files in puppet-tripleo and
tripleo-heat-templates.
Change-Id: Ic1d05ba7d572a291d2c77d85473394564699078c
We have had some job runs overrun the journald ring buffer which is used
by default resulting in losing older logs during the job runs. Update
the journald configuration to persistently store the journal so that we
can reliably retrieve those logs when jobs complete.
Change-Id: I5626ce76878287be220a8803f9dfe9a9da950d5b