autocloud: fare well autocloud, you served long and well...

Signed-off-by: Kevin Fenzi <kevin@scrye.com>
This commit is contained in:
Kevin Fenzi 2019-12-02 23:01:34 +00:00 committed by Pierre-Yves Chibon
parent 7451296458
commit 779fa01877
49 changed files with 6 additions and 1650 deletions

View file

@ -1,144 +0,0 @@
# -*- coding: utf-8 -*-
from datetime import datetime
import requests
import fedmsg.consumers
import fedfind.release
from sqlalchemy import exc
import autocloud
from autocloud.models import init_model, ComposeDetails, ComposeJobDetails
from autocloud.producer import publish_to_fedmsg
from autocloud.utils import is_valid_image, produce_jobs
import logging
log = logging.getLogger("fedmsg")
DEBUG = autocloud.DEBUG
class AutoCloudConsumer(fedmsg.consumers.FedmsgConsumer):
"""
Fedmsg consumer for Autocloud
"""
if DEBUG:
topic = [
'org.fedoraproject.dev.__main__.pungi.compose.status.change'
]
else:
topic = [
'org.fedoraproject.prod.pungi.compose.status.change'
]
config_key = 'autocloud.consumer.enabled'
def __init__(self, *args, **kwargs):
self.supported_archs = [arch for arch, _ in ComposeJobDetails.ARCH_TYPES]
log.info("Autocloud Consumer is ready for action.")
super(AutoCloudConsumer, self).__init__(*args, **kwargs)
def consume(self, msg):
""" This is called when we receive a message matching the topic. """
log.info('Received %r %r' % (msg['topic'], msg['body']['msg_id']))
STATUS_F = ('FINISHED_INCOMPLETE', 'FINISHED',)
images = []
compose_db_update = False
msg_body = msg['body']
status = msg_body['msg']['status']
compose_images_json = None
# Till F27, both cloud-base and atomic images were available
# under variant CloudImages. With F28 and onward releases,
# cloud-base image compose moved to cloud variant and atomic images
# moved under atomic variant.
prev_rel = ['26', '27']
if msg_body['msg']['release_version'] in prev_rel:
VARIANTS_F = ('CloudImages',)
else:
VARIANTS_F = ('AtomicHost', 'Cloud')
if status in STATUS_F:
location = msg_body['msg']['location']
json_metadata = '{}/metadata/images.json'.format(location)
resp = requests.get(json_metadata)
compose_images_json = getattr(resp, 'json', False)
if compose_images_json is not None:
compose_images_json = compose_images_json()
compose_images = compose_images_json['payload']['images']
compose_details = compose_images_json['payload']['compose']
compose_images = dict((variant, compose_images[variant])
for variant in VARIANTS_F
if variant in compose_images)
compose_id = compose_details['id']
rel = fedfind.release.get_release(cid=compose_id)
release = rel.release
compose_details.update({'release': release})
compose_images_variants = [variant for variant in VARIANTS_F
if variant in compose_images]
for variant in compose_images_variants:
compose_image = compose_images[variant]
for arch, payload in compose_image.iteritems():
if arch not in self.supported_archs:
continue
for item in payload:
relative_path = item['path']
if not is_valid_image(relative_path):
continue
absolute_path = '{}/{}'.format(location, relative_path)
item.update({
'compose': compose_details,
'absolute_path': absolute_path,
})
images.append(item)
compose_db_update = True
if compose_db_update:
session = init_model()
compose_date = datetime.strptime(compose_details['date'], '%Y%m%d')
try:
cd = ComposeDetails(
date=compose_date,
compose_id=compose_details['id'],
respin=compose_details['respin'],
type=compose_details['type'],
status=u'q',
location=location,
)
session.add(cd)
session.commit()
compose_details.update({
'status': 'queued',
'compose_job_id': cd.id,
})
publish_to_fedmsg(topic='compose.queued',
**compose_details)
except exc.IntegrityError:
session.rollback()
cd = session.query(ComposeDetails).filter_by(
compose_id=compose_details['id']).first()
log.info('Compose already exists %s: %s' % (
compose_details['id'],
cd.id
))
session.close()
num_images = len(images)
for pos, image in enumerate(images):
image.update({'pos': (pos+1, num_images)})
produce_jobs(images)

View file

@ -154,9 +154,6 @@
- name: restart cinder volume - name: restart cinder volume
service: name=openstack-cinder-volume state=restarted service: name=openstack-cinder-volume state=restarted
- name: restart autocloud
service: name=autocloud state=restarted
- name: restart mirrorlist-server - name: restart mirrorlist-server
service: name=mirrorlist-server state=restarted service: name=mirrorlist-server state=restarted

View file

@ -1,69 +0,0 @@
---
# Define resources for this group of hosts here.
lvm_size: 20000
mem_size: 6144
num_cpus: 4
nrpe_procs_warn: 900
nrpe_procs_crit: 1000
# for systems that do not match the above - specify the same parameter in
# the host_vars/$hostname file
tcp_ports: [
# These four ports are for outbound fedmsg
3000, 3001, 3002, 3003,
]
fas_client_groups: sysadmin-noc,sysadmin-fedimg,sysadmin-releng,sysadmin-veteran
sudoers: "{{ private }}/files/sudo/autocloud-backend"
# These are hw boxes and don't use the NM ifconfig setup
ansible_ifcfg_blacklist: true
# These people get told when something goes wrong.
fedmsg_error_recipients:
- sysadmin-fedimg-members@fedoraproject.org
# These are consumed by a task in roles/fedmsg/base/main.yml
fedmsg_certs:
- service: shell
owner: root
group: sysadmin
can_send:
- logger.log
- service: autocloud
owner: root
group: fedmsg
can_send:
- autocloud.image
- autocloud.image.running
- autocloud.image.success
- autocloud.image.failed
- autocloud.image.queued
- autocloud.compose
- autocloud.compose.queued
- autocloud.compose.running
- autocloud.compose.complete
# For the MOTD
csi_security_category: Moderate
csi_primary_contact: Cloudmeisters - sysadmin-fedimg-members@fedoraproject.org
csi_purpose: Run the autocloud testing backend
csi_relationship: |
This runs one of the two autocloud testing backends
This host relies on:
- A postgres db server running on db01. The db is called 'autocloud'.
The two autocloud-backend nodes populate that database with results.
- fedmsg messages produced by koji indicating that new images were built.
- queries to koji directly to download that content.
Running locally we have:
- fedmsg-hub which receives the initial notification from koji via fedmsg.
- redis, which stores a work queue populated by fedmsg
- the autocloud daemon, which pulls work from the redis queue and uses tunir
to test images in vagrant.
There's a cronjob that kills vagrant every 10 minutes if it got stuck on the
last test.

View file

@ -1,60 +0,0 @@
---
# Define resources for this group of hosts here.
lvm_size: 20000
mem_size: 6144
num_cpus: 4
# for systems that do not match the above - specify the same parameter in
# the host_vars/$hostname file
tcp_ports: [
# These four ports are for outbound fedmsg
3000, 3001, 3002, 3003,
]
fas_client_groups: sysadmin-noc,sysadmin-fedimg,sysadmin-releng,sysadmin-veteran
fedmsg_debug_loopback: True
# These people get told when something goes wrong.
fedmsg_error_recipients:
- sysadmin-fedimg-members@fedoraproject.org
# These are consumed by a task in roles/fedmsg/base/main.yml
fedmsg_certs:
- service: shell
owner: root
group: sysadmin
can_send:
- logger.log
- service: autocloud
owner: root
group: fedmsg
can_send:
- autocloud.image
- autocloud.image.running
- autocloud.image.success
- autocloud.image.failed
- autocloud.image.queued
# For the MOTD
csi_security_category: Moderate
csi_primary_contact: Cloudmeisters - sysadmin-fedimg-members@fedoraproject.org
csi_purpose: Run the autocloud testing backend
csi_relationship: |
This runs one of the two autocloud testing backends
This host relies on:
- A postgres db server running on db01. The db is called 'autocloud'.
The two autocloud-backend nodes populate that database with results.
- fedmsg messages produced by koji indicating that new images were built.
- queries to koji directly to download that content.
Running locally we have:
- fedmsg-hub which receives the initial notification from koji via fedmsg.
- redis, which stores a work queue populated by fedmsg
- the autocloud daemon, which pulls work from the redis queue and uses tunir
to test images in vagrant.
There's a cronjob that kills vagrant every 10 minutes if it got stuck on the
last test.

View file

@ -1,30 +0,0 @@
---
# Define resources for this group of hosts here.
lvm_size: 20000
mem_size: 2048
num_cpus: 2
# for systems that do not match the above - specify the same parameter in
# the host_vars/$hostname file
# This doesn't actually produce fedmsg messages, so no need for this var.
#wsgi_fedmsg_service: autocloud
wsgi_procs: 2
wsgi_threads: 2
tcp_ports: [ 80 ]
fas_client_groups: sysadmin-noc,sysadmin-fedimg,sysadmin-releng,sysadmin-veteran
# For the MOTD
csi_security_category: Moderate
csi_primary_contact: Cloudmeisters - sysadmin-fedimg-members@fedoraproject.org
csi_purpose: Run the readonly dashboard for autocloud results
csi_relationship: |
This only runs the mod_wsgi app for autocloud results.
https://apps.fedoraproject.org/autocloud
- This host relies on:
- A postgres db server running on db01. The db is called 'autocloud'.
- That db gets populated with results by daemons running on the
autocloud-backend nodes.

View file

@ -1,30 +0,0 @@
---
# Define resources for this group of hosts here.
lvm_size: 20000
mem_size: 2048
num_cpus: 1
# for systems that do not match the above - specify the same parameter in
# the host_vars/$hostname file
# This doesn't actually produce fedmsg messages, so no need for this var.
#wsgi_fedmsg_service: autocloud
wsgi_procs: 2
wsgi_threads: 2
tcp_ports: [ 80 ]
fas_client_groups: sysadmin-noc,sysadmin-fedimg,sysadmin-releng,sysadmin-veteran
# For the MOTD
csi_security_category: Moderate
csi_primary_contact: Cloudmeisters - sysadmin-fedimg-members@fedoraproject.org
csi_purpose: Run the readonly dashboard for autocloud results
csi_relationship: |
This only runs the mod_wsgi app for autocloud results.
https://apps.fedoraproject.org/autocloud
- This host relies on:
- A postgres db server running on db01. The db is called 'autocloud'.
- That db gets populated with results by daemons running on the
autocloud-backend nodes.

View file

@ -60,8 +60,6 @@ buildvmhost-01.phx2.fedoraproject.org
buildvmhost-02.phx2.fedoraproject.org buildvmhost-02.phx2.fedoraproject.org
buildvmhost-03.phx2.fedoraproject.org buildvmhost-03.phx2.fedoraproject.org
buildvmhost-04.phx2.fedoraproject.org buildvmhost-04.phx2.fedoraproject.org
autocloud-backend-libvirt2.phx2.fedoraproject.org
autocloud-backend-vbox2.phx2.fedoraproject.org
qa01.qa.fedoraproject.org qa01.qa.fedoraproject.org
qa02.qa.fedoraproject.org qa02.qa.fedoraproject.org
qa09.qa.fedoraproject.org qa09.qa.fedoraproject.org
@ -136,7 +134,6 @@ buildhw-aarch64-05.arm.fedoraproject.org
buildhw-aarch64-06.arm.fedoraproject.org buildhw-aarch64-06.arm.fedoraproject.org
buildhw-aarch64-07.arm.fedoraproject.org buildhw-aarch64-07.arm.fedoraproject.org
buildhw-aarch64-08.arm.fedoraproject.org buildhw-aarch64-08.arm.fedoraproject.org
autocloud-backend-aarch64.arm.fedoraproject.org
[storinator] [storinator]
storinator01.fedorainfracloud.org storinator01.fedorainfracloud.org

View file

@ -1,16 +0,0 @@
---
# this box is not mission critical
freezes: false
# general configs
nrpe_procs_warn: 900
nrpe_procs_crit: 1000
nm: 255.255.255.0
dns: 10.5.126.21
eth0_ip: 10.5.78.80
gw: 10.5.78.254
datacenter: phx2
autocloud_specialization: aarch64

View file

@ -1,22 +0,0 @@
---
# this box is not mission critical
freezes: false
# this box mounts a large share from the netapp to store combined http
# logs from the proxies.
nfs_mount_opts: "rw,hard,bg,intr,noatime,nodev,nosuid,nfsvers=3"
# general configs
nrpe_procs_warn: 900
nrpe_procs_crit: 1000
nm: 255.255.255.0
gw: 10.5.126.254
dns: 10.5.126.21
eth0_ip: 10.5.126.123
eth1_ip: 10.5.127.198
datacenter: phx2
autocloud_specialization: libvirt

View file

@ -1,25 +0,0 @@
---
# this box is not mission critical
freezes: false
# this box mounts a large share from the netapp to store combined http
# logs from the proxies.
nfs_mount_opts: "rw,hard,bg,intr,noatime,nodev,nosuid,nfsvers=3"
# general configs
nrpe_procs_warn: 900
nrpe_procs_crit: 1000
nm: 255.255.255.0
gw: 10.5.125.254
dns: 10.5.126.21
eth0_ip: 10.5.125.178
eth1_ip: 10.5.127.179
datacenter: phx2
autocloud_specialization: libvirt
nagios_Check_Services:
raid: true

View file

@ -1,22 +0,0 @@
---
# this box is not mission critical
freezes: false
# this box mounts a large share from the netapp to store combined http
# logs from the proxies.
nfs_mount_opts: "rw,hard,bg,intr,noatime,nodev,nosuid,nfsvers=3"
# general configs
nrpe_procs_warn: 900
nrpe_procs_crit: 1000
nm: 255.255.255.0
gw: 10.5.126.254
dns: 10.5.126.21
eth0_ip: 10.5.126.122
eth1_ip: 10.5.127.194
datacenter: phx2
autocloud_specialization: virtualbox

View file

@ -1,25 +0,0 @@
---
# this box is not mission critical
freezes: false
# this box mounts a large share from the netapp to store combined http
# logs from the proxies.
nfs_mount_opts: "rw,hard,bg,intr,noatime,nodev,nosuid,nfsvers=3"
# general configs
nrpe_procs_warn: 900
nrpe_procs_crit: 1000
nm: 255.255.255.0
gw: 10.5.125.254
dns: 10.5.126.21
eth0_ip: 10.5.125.179
eth1_ip: 10.5.127.178
datacenter: phx2
autocloud_specialization: virtualbox
nagios_Check_Services:
raid: true

View file

@ -1,16 +0,0 @@
---
nm: 255.255.255.0
gw: 10.5.128.254
dns: 10.5.126.21
ks_url: http://10.5.126.23/repo/rhel/ks/kvm-fedora-27
ks_repo: http://10.5.126.23/pub/fedora/linux/releases/27/Server/x86_64/os/
eth0_ip: 10.5.128.64
volgroup: /dev/vg_guests
vmhost: virthost04.stg.phx2.fedoraproject.org
datacenter: phx2
autocloud_specialization: libvirt

View file

@ -1,16 +0,0 @@
---
nm: 255.255.255.0
gw: 10.5.128.254
dns: 10.5.126.21
ks_url: http://10.5.126.23/repo/rhel/ks/kvm-fedora-27
ks_repo: http://10.5.126.23/pub/fedora/linux/releases/27/Server/x86_64/os/
eth0_ip: 10.5.128.65
volgroup: /dev/vg_virthost16
vmhost: virthost05.stg.phx2.fedoraproject.org
datacenter: phx2
autocloud_specialization: virtualbox

View file

@ -1,14 +0,0 @@
---
nm: 255.255.255.0
gw: 10.5.126.254
dns: 10.5.126.21
ks_url: http://10.5.126.23/repo/rhel/ks/kvm-fedora-27
ks_repo: http://10.5.126.23/pub/fedora/linux/releases/27/Server/x86_64/os/
eth0_ip: 10.5.126.117
volgroup: /dev/vg_guests
vmhost: virthost01.phx2.fedoraproject.org
datacenter: phx2

View file

@ -1,14 +0,0 @@
---
nm: 255.255.255.0
gw: 10.5.128.254
dns: 10.5.126.21
ks_url: http://10.5.126.23/repo/rhel/ks/kvm-fedora-27
ks_repo: http://10.5.126.23/pub/fedora/linux/releases/27/Server/x86_64/os/
eth0_ip: 10.5.128.66
volgroup: /dev/vg_guests
vmhost: virthost01.stg.phx2.fedoraproject.org
datacenter: phx2

View file

@ -1,14 +0,0 @@
---
nm: 255.255.255.0
gw: 10.5.126.254
dns: 10.5.126.21
ks_url: http://10.5.126.23/repo/rhel/ks/kvm-fedora-27
ks_repo: http://10.5.126.23/pub/fedora/linux/releases/27/Server/x86_64/os/
eth0_ip: 10.5.126.118
volgroup: /dev/vg_guests
vmhost: virthost19.phx2.fedoraproject.org
datacenter: phx2

View file

@ -1,14 +0,0 @@
---
nm: 255.255.255.0
gw: 10.5.128.254
dns: 10.5.126.21
ks_url: http://10.5.126.23/repo/rhel/ks/kvm-fedora-27
ks_repo: http://10.5.126.23/pub/fedora/linux/releases/27/Server/x86_64/os/
eth0_ip: 10.5.128.67
volgroup: /dev/vg_guests
vmhost: virthost03.stg.phx2.fedoraproject.org
datacenter: phx2

View file

@ -14,7 +14,6 @@ ks_repo: http://10.5.126.23/repo/rhel/RHEL7-x86_64/
databases: databases:
- askfedora - askfedora
- anitya - anitya
- autocloud
- blockerbugs - blockerbugs
- bodhi - bodhi
- bodhi2 - bodhi2
@ -39,7 +38,6 @@ databases:
dbs_to_backup: dbs_to_backup:
- askfedora - askfedora
- anitya - anitya
- autocloud
- blockerbugs - blockerbugs
- bodhi - bodhi
- bodhi2 - bodhi2

View file

@ -221,38 +221,6 @@ sign-vault01.stg.phx2.fedoraproject.org
#sign-vault05.phx2.fedoraproject.org #sign-vault05.phx2.fedoraproject.org
#sign-vault06.phx2.fedoraproject.org #sign-vault06.phx2.fedoraproject.org
[autocloud_web]
autocloud-web01.phx2.fedoraproject.org
autocloud-web02.phx2.fedoraproject.org
[autocloud_web_stg]
autocloud-web01.stg.phx2.fedoraproject.org
autocloud-web02.stg.phx2.fedoraproject.org
[autocloud_backend:children]
autocloud_backend_libvirt
autocloud_backend_vbox
autocloud_backend_aarch64
[autocloud_backend_libvirt]
autocloud-backend-libvirt2.phx2.fedoraproject.org
[autocloud_backend_vbox]
autocloud-backend-vbox2.phx2.fedoraproject.org
[autocloud_backend_aarch64]
autocloud-backend-aarch64.arm.fedoraproject.org
[autocloud_backend_stg:children]
autocloud_backend_libvirt_stg
autocloud_backend_vbox_stg
[autocloud_backend_libvirt_stg]
autocloud-backend01.stg.phx2.fedoraproject.org
[autocloud_backend_vbox_stg]
autocloud-backend02.stg.phx2.fedoraproject.org
[autosign] [autosign]
autosign01.phx2.fedoraproject.org autosign01.phx2.fedoraproject.org
@ -685,10 +653,6 @@ smtp-mm-cc-rdu01.fedoraproject.org
# All staging hosts should be in this group too. # All staging hosts should be in this group too.
# #
[staging] [staging]
autocloud-backend01.stg.phx2.fedoraproject.org
autocloud-backend02.stg.phx2.fedoraproject.org
autocloud-web01.stg.phx2.fedoraproject.org
autocloud-web02.stg.phx2.fedoraproject.org
autosign01.stg.phx2.fedoraproject.org autosign01.stg.phx2.fedoraproject.org
badges-backend01.stg.phx2.fedoraproject.org badges-backend01.stg.phx2.fedoraproject.org
badges-web01.stg.phx2.fedoraproject.org badges-web01.stg.phx2.fedoraproject.org
@ -919,7 +883,6 @@ openqa-stg01.qa.fedoraproject.org
# assorted categories of fedmsg services, for convenience # assorted categories of fedmsg services, for convenience
[fedmsg_hubs:children] [fedmsg_hubs:children]
autocloud_backend
badges_backend badges_backend
busgateway busgateway
fedimg fedimg
@ -929,7 +892,6 @@ pdc_backend
pkgs pkgs
[fedmsg_hubs_stg:children] [fedmsg_hubs_stg:children]
autocloud_backend_stg
badges_backend_stg badges_backend_stg
busgateway_stg busgateway_stg
fedimg_stg fedimg_stg

View file

@ -13,8 +13,6 @@
## group playbooks ## group playbooks
## ##
- import_playbook: /srv/web/infra/ansible/playbooks/groups/arm-qa.yml - import_playbook: /srv/web/infra/ansible/playbooks/groups/arm-qa.yml
- import_playbook: /srv/web/infra/ansible/playbooks/groups/autocloud-backend.yml
- import_playbook: /srv/web/infra/ansible/playbooks/groups/autocloud-web.yml
- import_playbook: /srv/web/infra/ansible/playbooks/groups/backup-server.yml - import_playbook: /srv/web/infra/ansible/playbooks/groups/backup-server.yml
- import_playbook: /srv/web/infra/ansible/playbooks/groups/badges-backend.yml - import_playbook: /srv/web/infra/ansible/playbooks/groups/badges-backend.yml
- import_playbook: /srv/web/infra/ansible/playbooks/groups/badges-web.yml - import_playbook: /srv/web/infra/ansible/playbooks/groups/badges-web.yml

View file

@ -1,50 +0,0 @@
- import_playbook: "/srv/web/infra/ansible/playbooks/include/virt-create.yml myhosts=autocloud_backend_stg"
- name: dole out the generic configuration
hosts: autocloud_backend:autocloud_backend_stg
user: root
gather_facts: True
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
pre_tasks:
- include_vars: dir=/srv/web/infra/ansible/vars/all/ ignore_files=README
- import_tasks: "{{ tasks_path }}/yumrepos.yml"
roles:
- base
- rkhunter
- hosts
- fas_client
- nagios_client
- collectd/base
- fedmsg/base
- sudo
tasks:
- import_tasks: "{{ tasks_path }}/2fa_client.yml"
- import_tasks: "{{ tasks_path }}/motd.yml"
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
- name: dole out the service-specific config
hosts: autocloud_backend:autocloud_backend_stg
user: root
gather_facts: True
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
roles:
- redis
- fedmsg/hub
- autocloud/backend
- role: collectd/fedmsg-service
process: fedmsg-hub

View file

@ -1,49 +0,0 @@
- import_playbook: "/srv/web/infra/ansible/playbooks/include/virt-create.yml myhosts=autocloud_web:autocloud_web_stg"
- name: make the box be real
hosts: autocloud_web:autocloud_web_stg
user: root
gather_facts: True
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
pre_tasks:
- import_tasks: "{{ tasks_path }}/yumrepos.yml"
roles:
- base
- rkhunter
- nagios_client
- hosts
- fas_client
- collectd/base
- mod_wsgi
- fedmsg/base
- sudo
- role: openvpn/client
when: env != "staging"
tasks:
- import_tasks: "{{ tasks_path }}/2fa_client.yml"
- import_tasks: "{{ tasks_path }}/motd.yml"
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
- name: dole out the app-specific configuration
hosts: autocloud_web:autocloud_web_stg
user: root
gather_facts: True
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
roles:
- autocloud/frontend

View file

@ -377,13 +377,6 @@
http_not_https_yes_this_is_insecure_and_i_feel_bad: true http_not_https_yes_this_is_insecure_and_i_feel_bad: true
when: env == "staging" when: env == "staging"
- role: httpd/reverseproxy
website: apps.fedoraproject.org
destname: autocloud
localpath: /autocloud
remotepath: /autocloud
proxyurl: http://localhost:10041
- role: httpd/reverseproxy - role: httpd/reverseproxy
website: pdc.fedoraproject.org website: pdc.fedoraproject.org
destname: pdc destname: pdc

View file

@ -1,166 +0,0 @@
- name: push packages out
hosts: autocloud_backend:autocloud_backend_stg:autocloud_web:autocloud_web_stg
user: root
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
vars:
testing: False
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
tasks:
- name: clean all metadata
command: dnf clean all
check_mode: no
- name: update autocloud packages from main repo
package: name="autocloud*" state=latest
when: not testing
- name: update autocloud packages from testing repo
dnf: name="autocloud*" state=latest enablerepo=infrastructure-tags-stg
when: testing
- name: update tunir
hosts: autocloud_backend:autocloud_backend_stg
user: root
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
vars:
testing: False
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
tasks:
- name: update tunir packages from main repo
package: name="tunir" state=latest
when: not testing
- name: update tunir packages from testing repo
dnf: name="tunir" state=latest enablerepo=infrastructure-tags-stg
when: testing
- name: update fedfind
hosts: autocloud_backend:autocloud_backend_stg
user: root
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
vars:
testing: False
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
tasks:
- name: update fedfind packages from main repo
package: name="fedfind" state=latest
when: not testing
- name: update fedfind packages from testing repo
dnf: name="fedfind" state=latest enablerepo=infrastructure-tags-stg
when: testing
- name: update python2-fedfind
hosts: autocloud_backend:autocloud_backend_stg
user: root
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
vars:
testing: False
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
tasks:
- name: update fedfind packages from main repo
package: name="python2-fedfind" state=latest
when: not testing
- name: update fedfind packages from testing repo
dnf: name="python2-fedfind" state=latest enablerepo=infrastructure-tags-stg
when: testing
- name: verify the frontend and stop it
hosts: autocloud_web:autocloud_web_stg
user: root
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
pre_tasks:
- name: tell nagios to shush w.r.t. the frontend
nagios: action=downtime minutes=15 service=host host={{ inventory_hostname_short }}{{ env_suffix }}
delegate_to: noc01.phx2.fedoraproject.org
ignore_errors: true
roles:
- autocloud/frontend
post_tasks:
- service: name="httpd" state=stopped
- name: verify the backends, stop them, and then upgrade the db
hosts: autocloud_backend:autocloud_backend_stg
user: root
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
pre_tasks:
- name: tell nagios to shush w.r.t. the backend
nagios: action=downtime minutes=15 service=host host={{ inventory_hostname_short }}{{ env_suffix }}
delegate_to: noc01.phx2.fedoraproject.org
ignore_errors: true
roles:
- autocloud/backend
tasks:
- name: Stop the fedmsg-hub backend
service: name="fedmsg-hub" state=stopped
- name: Stop the autocloud backend
service: name="autocloud" state=stopped
# There is no alembic upgrade scripts yet...
#- name: Upgrade the database
# command: /usr/bin/alembic -c /usr/share/autocloud/alembic.ini upgrade head
# args:
# chdir: /usr/share/autocloud/
# when: inventory_hostname.startswith('autocloud-backend-libvirt')
- name: And... start the autocloud backend again
service: name="autocloud" state=started
- name: And... start the fedmsg-hub backend again
service: name="fedmsg-hub" state=started
post_tasks:
- name: tell nagios to unshush w.r.t. the backend
nagios: action=unsilence service=host host={{ inventory_hostname_short }}{{ env_suffix }}
delegate_to: noc01.phx2.fedoraproject.org
ignore_errors: true
- name: restart the frontend
hosts: autocloud_web:autocloud_web_stg
user: root
vars_files:
- /srv/web/infra/ansible/vars/global.yml
- "/srv/private/ansible/vars.yml"
- /srv/web/infra/ansible/vars/{{ ansible_distribution }}.yml
handlers:
- import_tasks: "{{ handlers_path }}/restart_services.yml"
tasks:
- service: name="httpd" state=started
post_tasks:
- name: tell nagios to unshush w.r.t. the frontend
nagios: action=unsilence service=host host={{ inventory_hostname_short }}{{ env_suffix }}
delegate_to: noc01.phx2.fedoraproject.org
ignore_errors: true

View file

@ -1,35 +0,0 @@
sudo sh -c 'echo -e "\n10.5.126.23 infrastructure.fedoraproject.org\n10.5.125.44 pkgs.fedoraproject.org pkgs\n10.5.126.51 download.fedora.redhat.com\n10.5.125.63 koji.fedoraproject.org\n10.5.126.51 mirrors.fedoraproject.org\n10.5.125.36 kojipkgs.fedoraproject.org" >> /etc/hosts'
curl -O http://infrastructure.fedoraproject.org/infra/autocloud/tunirtests.tar.gz
tar -xzvf tunirtests.tar.gz
sudo python3 -m unittest tunirtests.atomictests.TestAtomic01Status -v
## sudo python3 -m unittest tunirtests.nongatingtests.TunirNonGatingtests -v
## sudo python3 -m unittest tunirtests.nongatingtests.TunirNonGatingtestBzip2 -v
## sudo python3 -m unittest tunirtests.nongatingtests.TunirNonGatingtestsCpio -v
## sudo python3 -m unittest tunirtests.nongatingtests.TunirNonGatingtestDiffutills -v
## sudo python3 -m unittest tunirtests.nongatingtests.TunirNonGatingtestaudit -v
## sudo python3 -m unittest tunirtests.selinux.TestSELinux -v
## sudo python3 -m unittest tunirtests.sshkeygentest.sshkeygenTest -v
## sudo python3 -m unittest tunirtests.testumountroot.TestUmountRoot -v
sudo python3 -m unittest tunirtests.cloudtests.TestBase -v
sudo python3 -m unittest tunirtests.cloudtests.TestCloudtmp -v
sudo python3 -m unittest tunirtests.cloudtests.Testtmpmount -v
sudo python3 -m unittest tunirtests.cloudtests.Testnetname -v
## python3 -m unittest tunirtests.cloudtests.TestJournalWritten -v
sudo python3 -m unittest tunirtests.cloudservice.TestServiceStop -v
sudo python3 -m unittest tunirtests.cloudservice.TestServiceDisable -v
@@ sudo reboot
POLL
## sudo python3 -m unittest tunirtests.testreboot.TestReboot -v
sudo python3 -m unittest tunirtests.cloudservice.TestServiceManipulation -v
## python3 -m unittest tunirtests.cloudtests.TestJournalWrittenAfterReboot -v
@@ sudo reboot
POLL
sudo python3 -m unittest tunirtests.cloudservice.TestServiceAfter -v
sudo python3 -m unittest tunirtests.atomictests.TestDockerInstalled -v
sudo python3 -m unittest tunirtests.atomictests.TestDockerStorageSetup -v
sudo python3 -m unittest tunirtests.atomictests.TestAtomicFirstBootRun -v
sudo python3 -m unittest tunirtests.atomictests.TestAtomicCommand -v
sudo python3 -m unittest tunirtests.atomictests.TestAtomicDockerImage -v
sudo python3 -m unittest tunirtests.atomictests.TestRootMount -v
sudo python3 -m unittest tunirtests.atomictests.Testreadonlymount -v
sudo python3 -m unittest tunirtests.atomictests.TestDockerDaemon -v

View file

@ -1 +0,0 @@
*/10 * * * * root /sbin/kill_vagrant > /dev/null 2>&1

View file

@ -1,6 +0,0 @@
[virtualbox]
name=Fedora $releasever - $basearch - VirtualBox
baseurl=http://download.virtualbox.org/virtualbox/rpm/fedora/$releasever/$basearch
enabled=1
gpgcheck=1
gpgkey=https://www.virtualbox.org/download/oracle_vbox.asc

View file

@ -1,173 +0,0 @@
---
# Configuration for the Fedora Notifications webapp
- name: install needed packages
package:
state: present
name:
- autocloud-common
- autocloud-backend
- tunir
- fedfind
- python2-fedfind
- python-sqlalchemy
- python-sqlalchemy-utils
- libsemanage-python
- libselinux-python
notify:
- restart fedmsg-hub
- restart autocloud
tags:
- autocloud
- autocloud/backend
- name: Check for the existance of a 'tunirports' redis entry
command: redis-cli --scan --pattern tunirports
changed_when: False
register: tunirports
check_mode: no
tags:
- autocloud
- autocloud/backend
- name: copy autocloud backend configuration
template: >
src={{ item }} dest=/etc/autocloud/{{ item }}
owner=root group=fedmsg mode=0640
with_items:
- autocloud.cfg
notify:
- restart fedmsg-hub
- restart autocloud
tags:
- autocloud
- autocloud/backend
- name: copy autocloud fedmsg configuration
template: >
src={{ item }} dest=/etc/fedmsg.d/{{ item }}
owner=root group=fedmsg mode=0640
with_items:
- autocloud.py
notify:
- restart fedmsg-hub
- restart autocloud
tags:
- autocloud
- autocloud/backend
- name: install libvirt and vagrant-libvirt for the libvirt host
package:
state: present
name:
- libvirt
- vagrant-libvirt
when: autocloud_specialization == 'libvirt'
notify:
- restart fedmsg-hub
- restart autocloud
tags:
- autocloud
- autocloud/backend
- name: install libvirt for the aarch64 host
package:
state: present
name:
- libvirt
when: autocloud_specialization == 'aarch64'
notify:
- restart fedmsg-hub
- restart autocloud
tags:
- autocloud
- autocloud/backend
- name: Setup virtualbox requirements where needed
copy: src=virtualbox.repo dest=/etc/yum.repos.d/virtualbox.repo
when: autocloud_specialization == 'virtualbox'
tags:
- autocloud
- autocloud/backend
- name: install virtualbox basics where needed
package:
state: present
name:
- VirtualBox-4.3
- kernel-devel
- gcc
- vagrant
register: virtualbox_installed
check_mode: no
when: autocloud_specialization == 'virtualbox' and inventory_hostname == 'autocloud-libvirt-vbox.phx2.fedoraproject.org'
notify:
- restart fedmsg-hub
- restart autocloud
tags:
- autocloud
- autocloud/backend
- name: install virtualbox basics where needed
package:
state: present
name:
- VirtualBox-5.1
- kernel-devel
- gcc
- vagrant
register: virtualbox_installed
check_mode: no
when: autocloud_specialization == 'virtualbox' and inventory_hostname == 'autocloud-libvirt-vbox2.phx2.fedoraproject.org'
notify:
- restart fedmsg-hub
- restart autocloud
tags:
- autocloud
- autocloud/backend
- name: setup vbox if it was just installed.
command: /etc/init.d/vboxdrv setup
when: virtualbox_installed is changed
notify:
- restart fedmsg-hub
- restart autocloud
tags:
- autocloud
- autocloud/backend
- name: Copy the correct tunir job details
copy: src=fedora.txt dest=/etc/autocloud/fedora.txt
tags:
- autocloud
- autocloud/backend
- autocloud/backend/tests
- name: Set up a cronjob to run kill_vagrant every so often..
copy: src=kill_vagrant.cron dest=/etc/cron.d/kill_vagrant.cron
tags:
- cron
- autocloud
- autocloud/backend
- name: start autocloud
service: name=autocloud enabled=yes state=started
tags:
- autocloud
- autocloud/backend
#
# Install hotfix to ignore new architectures
# See PR - https://github.com/kushaldas/autocloud/pull/62/
#
- name: hotfix - copy over consumer files
copy: src='{{ files }}/{{ item.src }}' dest={{ item.dest }}
with_items:
- { src: 'hotfix/autocloud/consumer.py', dest: '/usr/lib/python2.7/site-packages/autocloud' }
notify:
- restart fedmsg-hub
tags:
- autocloud
- hotfix

View file

@ -1,27 +0,0 @@
[autocloud]
{% if env == 'staging' %}
koji_server_url = http://koji.stg.fedoraproject.org/kojihub/
base_koji_task_url = https://kojipkgs.stg.fedoraproject.org//work/
{% else %}
koji_server_url = http://koji.fedoraproject.org/kojihub/
base_koji_task_url = https://kojipkgs.fedoraproject.org//work/
{% endif %}
debug = false
# These are required, but are ignored.
host = 0.0.0.0
port = 5000
{% if autocloud_specialization == 'virtualbox' %}
virtualbox = true
{% else %}
# Presumably, this host is doing libvirt instead of virtualbox.
virtualbox = false
{% endif %}
[sqlalchemy]
{% if env == 'staging' %}
uri = postgres://autocloud:{{autocloud_db_password_stg}}@db01.stg/autocloud
{% else %}
uri = postgres://autocloud:{{autocloud_db_password}}@db01/autocloud
{% endif %}

View file

@ -1,21 +0,0 @@
config = {
# Consumer stuff
"autocloud.consumer.enabled": True,
{% if env == 'staging' %}
"autocloud.sqlalchemy.uri": "postgres://autocloud:{{autocloud_db_password_stg}}@db01.stg/autocloud",
{% else %}
"autocloud.sqlalchemy.uri": "postgres://autocloud:{{autocloud_db_password}}@db01/autocloud",
{% endif %}
# Turn on logging for autocloud
"logging": dict(
loggers=dict(
autocloud={
"level": "DEBUG",
"propagate": False,
"handlers": ["console"],
},
),
),
}

View file

@ -1,69 +0,0 @@
---
# Configuration for the Fedora Notifications webapp
- name: install needed packages
package:
state: present
name:
- autocloud-common
- autocloud-web
- python-sqlalchemy
- python-sqlalchemy-utils
- python2-flask-restless
- libsemanage-python
notify:
- restart apache
tags:
- autocloud
- autocloud/frontend
- name: copy autocloud dashboard configuration
template: >
src={{ item }} dest=/etc/autocloud/{{ item }}
owner=apache group=apache mode=0600
with_items:
- autocloud.cfg
notify:
- restart apache
tags:
- autocloud
- autocloud/frontend
- name: copy autocloud httpd config
template: >
src=autocloud.conf dest=/etc/httpd/conf.d/autocloud.conf
owner=apache group=apache mode=0644
notify:
- restart apache
tags:
- autocloud
- autocloud/frontend
- name: setup symlink to fedora theme
file: >
src=/usr/share/autocloud/static/bootstrap-3.3.4-fedora
dest=/usr/share/autocloud/static/bootstrap
state=link
tags:
- autocloud
- autocloud/frontend
- name: apply selinux type to static files
file: >
dest=/usr/share/autocloud/static
setype=httpd_sys_content_t
state=directory
recurse=yes
tags:
- autocloud
- autocloud/frontend
- selinux
- name: ensure selinux lets httpd talk to postgres.
seboolean: name={{item}} state=yes persistent=yes
with_items:
- httpd_can_network_connect_db
tags:
- autocloud
- autocloud/frontend
- selinux

View file

@ -1,22 +0,0 @@
[autocloud]
{% if env == 'staging' %}
koji_server_url = http://koji.fedoraproject.org/kojihub/
base_koji_task_url = https://kojipkgs.fedoraproject.org//work/
{% else %}
koji_server_url = http://koji.stg.fedoraproject.org/kojihub/
base_koji_task_url = https://kojipkgs.stg.fedoraproject.org//work/
{% endif %}
virtualbox = false
debug = false
# These are required, but are ignored when run under mod_wsgi.
host = 0.0.0.0
port = 5000
[sqlalchemy]
{% if env == 'staging' %}
uri = postgres://autocloud:{{autocloud_db_password_stg}}@db01.stg/autocloud
{% else %}
uri = postgres://autocloud:{{autocloud_db_password}}@db01/autocloud
{% endif %}

View file

@ -1,24 +0,0 @@
LoadModule wsgi_module modules/mod_wsgi.so
Alias /autocloud/static /usr/share/autocloud/static
WSGIPythonEggs /var/cache/autocloud/.python-eggs
WSGIDaemonProcess autocloud user=apache group=apache maximum-requests=50000 display-name=autocloud processes=3 threads=4 inactivity-timeout=300
WSGISocketPrefix run/wsgi
WSGIRestrictStdout Off
WSGIRestrictSignal Off
WSGIPythonOptimize 1
WSGIScriptAlias /autocloud /usr/share/autocloud/autocloud.wsgi
<Location /autocloud>
WSGIProcessGroup autocloud
Require all granted
</Location>
<Directory /usr/share/autocloud/>
Order deny,allow
Allow from all
Require all granted
</Directory>

View file

@ -184,15 +184,6 @@ syncHttpLogs {{host}}
syncHttpLogs {{host}} syncHttpLogs {{host}}
{% endfor %} {% endfor %}
## sync up autocloud
{% for host in groups['autocloud_web'] %}
syncHttpLogs {{host}}
{% endfor %}
{% for host in groups['autocloud_web_stg'] %}
syncHttpLogs {{host}}
{% endfor %}
## sync up notifs ## sync up notifs
{% for host in groups['notifs_web'] %} {% for host in groups['notifs_web'] %}
syncHttpLogs {{host}} syncHttpLogs {{host}}

View file

@ -1,310 +0,0 @@
server-identifier dhcp01.phx2.fedoraproject.org;
ddns-update-style none;
subnet 10.5.125.0 netmask 255.255.255.0 {
allow booting;
allow bootp;
option domain-name "phx2.fedoraproject.org";
option domain-name-servers 10.5.126.21, 10.5.126.22;
option routers 10.5.125.254;
# option vendor-class-identifier "PXEClient";
# option vendor-encapsulated-options 09:0f:80:00:0c:4e:65:74:77:6f:72:6b:20:62:6f:6f:74:0a:07:00:50:72:6f:6d:70:74:06:01:02:08:03:80:00:00:47:04:80:00:00:00:ff;
host unknown00 {
hardware ethernet 00:21:5e:5b:19:96;
fixed-address 10.5.125.33;
next-server 10.5.125.43;
option host-name "unknown00.phx2.fedoraproject.org";
# filename "yaboot";
}
host sign-vault03 {
hardware ethernet fc:99:47:49:43:84;
fixed-address 10.5.125.73;
option host-name "sign-vault03";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host sign-vault04 {
hardware ethernet fc:99:47:49:8d:fc;
fixed-address 10.5.125.74;
option host-name "sign-vault04";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host bkernel03 {
hardware ethernet D0:94:66:45:8C:0F;
fixed-address 10.5.125.81;
option host-name "bkernel03";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host bkernel04 {
hardware ethernet D0:94:66:45:A7:E4;
fixed-address 10.5.125.82;
option host-name "bkernel04";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host sign-vault05 {
hardware ethernet D0:94:66:45:87:C1;
fixed-address 10.5.125.83;
option host-name "sign-vault05";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host sign-vault06 {
hardware ethernet D0:94:66:45:A1:62;
fixed-address 10.5.125.84;
option host-name "sign-vault06";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host bvirthost12 {
hardware ethernet 24:6E:96:B1:61:C4;
fixed-address 10.5.125.10;
option host-name "bvirthost12";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host bvirthost13 {
hardware ethernet 24:6E:96:B1:5E:B4;
fixed-address 10.5.125.11;
option host-name "bvirthost13";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host bvirthost14 {
hardware ethernet 24:6E:96:B1:56:24;
fixed-address 10.5.125.12;
option host-name "bvirthost14";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host bvirthost15 {
hardware ethernet 24:6E:96:B0:E0:7C;
fixed-address 10.5.125.13;
option host-name "bvirthost15";
next-server 10.5.126.41;
filename "pxelinux.0";
}
host sign-vault01 {
hardware ethernet 00:1a:64:67:a3:38;
fixed-address 10.5.125.70;
next-server 10.5.126.41;
option host-name "sign-vault01";
filename "pxelinux.0";
}
host bvirthost01 {
hardware ethernet e4:1f:13:6a:c5:58;
fixed-address 10.5.125.124;
next-server 10.5.125.43;
option host-name "bvirthost01";
filename "pxelinux.0";
}
host junk01 {
hardware ethernet E4:1F:13:BA:E8:28;
fixed-address 10.5.125.127;
next-server 10.5.125.43;
option host-name "junk01";
filename "pxelinux.0";
}
host bvirthost04 {
hardware ethernet 18:66:DA:F7:61:58;
fixed-address 10.5.125.76;
next-server 10.5.126.41;
option host-name "bvirthost04";
filename "pxelinux.0";
}
host bvirthost07 {
hardware ethernet 90:B1:1C:32:7E:8E;
fixed-address 10.5.125.122;
next-server 10.5.126.41;
option host-name "bvirthost07";
filename "pxelinux.0";
}
host buildhw-01 {
hardware ethernet 14:9E:CF:61:9E:61;
fixed-address 10.5.125.164;
next-server 10.5.125.43;
option host-name "buildhw-01.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildhw-02 {
hardware ethernet 14:9E:CF:61:9F:4F;
fixed-address 10.5.125.165;
next-server 10.5.125.43;
option host-name "buildhw-02.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildhw-03 {
hardware ethernet 14:9E:CF:61:9E:C9;
fixed-address 10.5.125.166;
next-server 10.5.125.43;
option host-name "buildhw-03.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildhw-04 {
hardware ethernet 14:9E:CF:61:A5:07;
fixed-address 10.5.125.167;
next-server 10.5.125.43;
option host-name "buildhw-04.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildhw-05 {
hardware ethernet 14:9E:CF:61:9E:7B;
fixed-address 10.5.125.168;
next-server 10.5.125.43;
option host-name "buildhw-05.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildhw-06 {
hardware ethernet 14:9E:CF:61:A0:BD;
fixed-address 10.5.125.169;
next-server 10.5.125.43;
option host-name "buildhw-06.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildhw-07 {
hardware ethernet 14:9E:CF:61:9E:E3;
fixed-address 10.5.125.170;
next-server 10.5.125.43;
option host-name "buildhw-07.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildhw-08 {
hardware ethernet 14:9E:CF:61:A6:75;
fixed-address 10.5.125.171;
next-server 10.5.125.43;
option host-name "buildhw-08.phx2.fedoraproject.org";
filename "pxelinux.0";
}
#
host buildhw-09 {
hardware ethernet F8:CA:B8:F7:26:E1;
fixed-address 10.5.125.172;
next-server 10.5.125.43;
option host-name "buildhw-09.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildhw-10 {
hardware ethernet F8:CA:B8:F7:27:CF;
fixed-address 10.5.125.173;
next-server 10.5.125.43;
option host-name "buildhw-10.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildvmhost-01 {
hardware ethernet F8:CA:B8:F7:27:49;
fixed-address 10.5.125.174;
next-server 10.5.125.43;
option host-name "buildvmhost-01.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildvmhost-02 {
hardware ethernet F8:CA:B8:F7:2D:87;
fixed-address 10.5.125.175;
next-server 10.5.125.43;
option host-name "buildvmhost-02.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildvmhost-03 {
hardware ethernet F8:CA:B8:F7:26:FB;
fixed-address 10.5.125.176;
next-server 10.5.125.43;
option host-name "buildvmhost-03.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host buildvmhost-04 {
hardware ethernet F8:CA:B8:F7:29:3D;
fixed-address 10.5.125.177;
next-server 10.5.125.43;
option host-name "buildvmhost-04.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host autocloud-backend-libvirt2 {
hardware ethernet F8:CA:B8:F7:27:63;
fixed-address 10.5.125.178;
next-server 10.5.125.43;
option host-name "autocloud-backend-libvirt2.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host autocloud-backend-vbox2 {
hardware ethernet F8:CA:B8:F7:2E:F5;
fixed-address 10.5.125.179;
next-server 10.5.125.43;
option host-name "autocloud-backend-vbox2.phx2.fedoraproject.org";
filename "pxelinux.0";
}
host bvirthost05 {
hardware ethernet 80:18:44:DE:4D:FC;
fixed-address 10.5.125.121;
next-server 10.5.125.43;
option host-name "bvirthost05.phx2.fedoraproject.org";
filename "pxelinux.0";
}
group macs {
allow booting;
allow bootp;
option domain-name "phx2.fedoraproject.org";
option domain-name-servers 10.5.126.21, 10.5.126.22;
option routers 10.5.125.254;
filename "yaboot";
authoritative;
option dhcp-max-message-size 576;
option dhcp-parameter-request-list = concat(
option dhcp-parameter-request-list,
dc, dd, e6, e8, e9, ea, eb, ec, ed, ee); # mac options
host ppc02 {
hardware ethernet 26:49:30:a8:7f:04;
fixed-address 10.5.125.30;
next-server 10.5.126.41;
option host-name "ppc02.phx2.fedoraproject.org";
filename "yaboot";
}
}
# range 10.5.125.170 10.5.125.189;
# next-server 10.5.125.43;
# filename "pxelinux.0";
}

View file

@ -256,20 +256,20 @@ subnet 10.5.125.0 netmask 255.255.255.0 {
filename "uefi/grubx64.efi"; filename "uefi/grubx64.efi";
} }
host autocloud-backend-libvirt2 { host buildhw-11 {
hardware ethernet F8:CA:B8:F7:27:63; hardware ethernet F8:CA:B8:F7:27:63;
fixed-address 10.5.125.178; fixed-address 10.5.125.178;
next-server 10.5.125.43; next-server 10.5.125.43;
option host-name "autocloud-backend-libvirt2.phx2.fedoraproject.org"; option host-name "buildhw-11.phx2.fedoraproject.org";
filename "pxelinux.0"; filename "uefi/grubx64.efi";
} }
host autocloud-backend-vbox2 { host buildhw-12 {
hardware ethernet F8:CA:B8:F7:2E:F5; hardware ethernet F8:CA:B8:F7:2E:F5;
fixed-address 10.5.125.179; fixed-address 10.5.125.179;
next-server 10.5.125.43; next-server 10.5.125.43;
option host-name "autocloud-backend-vbox2.phx2.fedoraproject.org"; option host-name "buildhw-12.phx2.fedoraproject.org";
filename "pxelinux.0"; filename "uefi/grubx64.efi";
} }
group macs { group macs {

View file

@ -115,7 +115,6 @@
- endpoints-fedimg.py - endpoints-fedimg.py
- endpoints-bugzilla2fedmsg.py - endpoints-bugzilla2fedmsg.py
- endpoints-bodhi.py - endpoints-bodhi.py
- endpoints-autocloud.py
- endpoints-odcs-backend.py - endpoints-odcs-backend.py
- relay.py - relay.py
- logging.py - logging.py

View file

@ -1,17 +0,0 @@
{% if env == 'staging' %}
suffix = 'stg.phx2.fedoraproject.org'
machines = ['autocloud-backend01', 'autocloud-backend02']
{% else %}
suffix = 'phx2.fedoraproject.org'
machines = ['autocloud-backend-libvirt2', 'autocloud-backend-vbox2']
{% endif %}
endpoints = {}
for machine in machines:
endpoint = [
"tcp://%s.%s:30%0.2i" % (machine, suffix, i)
for i in range(4)
]
endpoints['autocloud.%s' % machine] = endpoint
config = dict(endpoints=endpoints)

View file

@ -212,16 +212,6 @@ backend kerneltest-backend
server kerneltest01 kerneltest01:80 check inter 10s rise 1 fall 2 server kerneltest01 kerneltest01:80 check inter 10s rise 1 fall 2
option httpchk GET /kerneltest option httpchk GET /kerneltest
frontend autocloud-frontend
bind 0.0.0.0:10041
default_backend autocloud-backend
backend autocloud-backend
balance hdr(appserver)
server autocloud-web01 autocloud-web01:80 check inter 10s rise 1 fall 2
server autocloud-web02 autocloud-web02:80 check inter 10s rise 1 fall 2
option httpchk GET /autocloud/
frontend openqa-frontend frontend openqa-frontend
bind 0.0.0.0:10044 bind 0.0.0.0:10044
default_backend openqa-backend default_backend openqa-backend

View file

@ -131,7 +131,6 @@
- check_fedmsg_gateway_proc.cfg - check_fedmsg_gateway_proc.cfg
- check_fedmsg_composer_proc.cfg - check_fedmsg_composer_proc.cfg
- check_redis_proc.cfg - check_redis_proc.cfg
- check_autocloud_proc.cfg
- check_fedmsg_consumers.cfg - check_fedmsg_consumers.cfg
- check_supybot_fedmsg_plugin.cfg - check_supybot_fedmsg_plugin.cfg
- check_datanommer_history.cfg - check_datanommer_history.cfg

View file

@ -1 +0,0 @@
command[check_autocloud_proc]=/usr/lib64/nagios/plugins/check_procs -c 1:1 -C 'python' -a 'autocloud_job.py' -u root

View file

@ -15,7 +15,6 @@
# 3 months -> 7884000 # 3 months -> 7884000
command[check_datanommer_anitya]={{libdir}}/nagios/plugins/check_datanommer_timesince.py anitya 604800 1814400 command[check_datanommer_anitya]={{libdir}}/nagios/plugins/check_datanommer_timesince.py anitya 604800 1814400
command[check_datanommer_ansible]={{libdir}}/nagios/plugins/check_datanommer_timesince.py ansible 432000 604800 command[check_datanommer_ansible]={{libdir}}/nagios/plugins/check_datanommer_timesince.py ansible 432000 604800
command[check_datanommer_autocloud]={{libdir}}/nagios/plugins/check_datanommer_timesince.py autocloud 259200 1814400
command[check_datanommer_bodhi]={{libdir}}/nagios/plugins/check_datanommer_timesince.py bodhi 86400 604800 command[check_datanommer_bodhi]={{libdir}}/nagios/plugins/check_datanommer_timesince.py bodhi 86400 604800
command[check_datanommer_bodhi_composes]={{libdir}}/nagios/plugins/check_datanommer_timesince.py org.fedoraproject.prod.bodhi.compose.start 86400 90000 command[check_datanommer_bodhi_composes]={{libdir}}/nagios/plugins/check_datanommer_timesince.py org.fedoraproject.prod.bodhi.compose.start 86400 90000
command[check_datanommer_buildsys]={{libdir}}/nagios/plugins/check_datanommer_timesince.py buildsys 14400 86400 command[check_datanommer_buildsys]={{libdir}}/nagios/plugins/check_datanommer_timesince.py buildsys 14400 86400

View file

@ -10,7 +10,6 @@ command[check_fedmsg_cp_notifs_backend]={{libdir}}/nagios/plugins/check_fedmsg_p
command[check_fedmsg_cp_bugzilla2fedmsg]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py moksha-hub BugzillaConsumer MonitoringProducer command[check_fedmsg_cp_bugzilla2fedmsg]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py moksha-hub BugzillaConsumer MonitoringProducer
command[check_fedmsg_cp_fedimg_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub FedimgConsumer MonitoringProducer command[check_fedmsg_cp_fedimg_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub FedimgConsumer MonitoringProducer
command[check_fedmsg_cp_hotness_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub BugzillaTicketFiler MonitoringProducer command[check_fedmsg_cp_hotness_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub BugzillaTicketFiler MonitoringProducer
command[check_fedmsg_cp_autocloud_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub AutoCloudConsumer MonitoringProducer
command[check_fedmsg_cp_packages_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub CacheInvalidator MonitoringProducer command[check_fedmsg_cp_packages_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub CacheInvalidator MonitoringProducer
command[check_fedmsg_cp_pdc_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub PDCUpdater MonitoringProducer command[check_fedmsg_cp_pdc_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub PDCUpdater MonitoringProducer
command[check_fedmsg_cp_mbs_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub MBSConsumer MBSProducer MonitoringProducer command[check_fedmsg_cp_mbs_backend]={{libdir}}/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub MBSConsumer MBSProducer MonitoringProducer
@ -26,7 +25,6 @@ command[check_fedmsg_cexceptions_notifs_backend]={{libdir}}/nagios/plugins/check
command[check_fedmsg_cexceptions_bugzilla2fedmsg]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py moksha-hub BugzillaConsumer 1 10 command[check_fedmsg_cexceptions_bugzilla2fedmsg]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py moksha-hub BugzillaConsumer 1 10
command[check_fedmsg_cexceptions_fedimg_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub FedimgConsumer 1 10 command[check_fedmsg_cexceptions_fedimg_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub FedimgConsumer 1 10
command[check_fedmsg_cexceptions_hotness_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub BugzillaTicketFiler 1 10 command[check_fedmsg_cexceptions_hotness_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub BugzillaTicketFiler 1 10
command[check_fedmsg_cexceptions_autocloud_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub AutoCloudConsumer 1 10
command[check_fedmsg_cexceptions_packages_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub CacheInvalidator 1 10 command[check_fedmsg_cexceptions_packages_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub CacheInvalidator 1 10
command[check_fedmsg_cexceptions_pdc_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub PDCUpdater 1 10 command[check_fedmsg_cexceptions_pdc_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub PDCUpdater 1 10
command[check_fedmsg_cexceptions_mbs_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub MBSConsumer 1 10 command[check_fedmsg_cexceptions_mbs_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub MBSConsumer 1 10
@ -42,7 +40,6 @@ command[check_fedmsg_cbacklog_notifs_backend]={{libdir}}/nagios/plugins/check_fe
command[check_fedmsg_cbacklog_bugzilla2fedmsg]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py moksha-hub BugzillaConsumer 10 100 command[check_fedmsg_cbacklog_bugzilla2fedmsg]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py moksha-hub BugzillaConsumer 10 100
command[check_fedmsg_cbacklog_fedimg_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub FedimgConsumer 2000 5000 command[check_fedmsg_cbacklog_fedimg_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub FedimgConsumer 2000 5000
command[check_fedmsg_cbacklog_hotness_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub BugzillaTicketFiler 1000 5000 command[check_fedmsg_cbacklog_hotness_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub BugzillaTicketFiler 1000 5000
command[check_fedmsg_cbacklog_autocloud_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub AutoCloudConsumer 100 500
command[check_fedmsg_cbacklog_packages_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub CacheInvalidator 30000 40000 command[check_fedmsg_cbacklog_packages_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub CacheInvalidator 30000 40000
command[check_fedmsg_cbacklog_pdc_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub PDCUpdater 10000 20000 command[check_fedmsg_cbacklog_pdc_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub PDCUpdater 10000 20000
command[check_fedmsg_cbacklog_mbs_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub MBSConsumer 10000 20000 command[check_fedmsg_cbacklog_mbs_backend]={{libdir}}/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub MBSConsumer 10000 20000

View file

@ -1,13 +0,0 @@
define service {
hostgroup_name autocloud_backend
service_description Check for autocloud proc
check_command check_by_nrpe!check_autocloud_proc
use defaulttemplate
}
define service {
hostgroup_name autocloud_backend
service_description Check for redis proc
check_command check_by_nrpe!check_redis_proc
use defaulttemplate
}

View file

@ -38,35 +38,6 @@ define service {
use defaulttemplate use defaulttemplate
} }
define service {
hostgroup_name autocloud_backend
service_description Check for fedmsg-hub proc
check_command check_by_nrpe!check_fedmsg_hub_proc
use defaulttemplate
}
define service {
hostgroup_name autocloud_backend
service_description Check fedmsg consumers and producers hub
check_command check_by_nrpe!check_fedmsg_cp_autocloud_backend
use defaulttemplate
}
define service {
hostgroup_name autocloud_backend
service_description Check fedmsg-hub consumers exceptions
check_command check_by_nrpe!check_fedmsg_cexceptions_autocloud_backend
use defaulttemplate
}
define service {
hostgroup_name autocloud_backend
service_description Check fedmsg-hub consumers backlog
check_command check_by_nrpe!check_fedmsg_cbacklog_autocloud_backend
use defaulttemplate
}
define service { define service {
host_name busgateway01.phx2.fedoraproject.org host_name busgateway01.phx2.fedoraproject.org
service_description Check for fedmsg-hub proc service_description Check for fedmsg-hub proc
@ -269,12 +240,6 @@ define service {
check_command check_by_nrpe!check_datanommer_fmn check_command check_by_nrpe!check_datanommer_fmn
use defaulttemplate use defaulttemplate
} }
define service {
host_name busgateway01.phx2.fedoraproject.org
service_description Check datanommer for recent autocloud messages
check_command check_by_nrpe!check_datanommer_autocloud
use defaulttemplate
}
define service { define service {
host_name busgateway01.phx2.fedoraproject.org host_name busgateway01.phx2.fedoraproject.org
service_description Check datanommer for recent atomic compose service_description Check datanommer for recent atomic compose

View file

@ -335,7 +335,6 @@ command[check_fedmsg_odcs_celery_proc]=/usr/lib64/nagios/plugins/check_procs -c
command[check_supybot_fedmsg_plugin]=/usr/lib64/nagios/plugins/check_supybot_plugin -t fedmsg command[check_supybot_fedmsg_plugin]=/usr/lib64/nagios/plugins/check_supybot_plugin -t fedmsg
command[check_haproxy_conns]=/usr/lib64/nagios/plugins/check_haproxy_conns.py command[check_haproxy_conns]=/usr/lib64/nagios/plugins/check_haproxy_conns.py
command[check_redis_proc]=/usr/lib64/nagios/plugins/check_procs -c 1:1 -C 'redis-server' -u redis command[check_redis_proc]=/usr/lib64/nagios/plugins/check_procs -c 1:1 -C 'redis-server' -u redis
command[check_autocloud_proc]=/usr/lib64/nagios/plugins/check_procs -c 1:1 -C 'python' -a 'autocloud_job.py' -u root
command[check_openvpn_link]=/usr/lib64/nagios/plugins/check_ping -H 192.168.1.41 -w 375.0,20% -c 500,60% command[check_openvpn_link]=/usr/lib64/nagios/plugins/check_ping -H 192.168.1.41 -w 375.0,20% -c 500,60%
command[check_memcache]=/usr/lib64/nagios/plugins/check_procs -c 1:1 -a '/usr/bin/memcached' -u memcached command[check_memcache]=/usr/lib64/nagios/plugins/check_procs -c 1:1 -a '/usr/bin/memcached' -u memcached
command[check_memcache_connect]=/usr/lib64/nagios/plugins/check_memcache_connect command[check_memcache_connect]=/usr/lib64/nagios/plugins/check_memcache_connect
@ -356,7 +355,6 @@ command[check_mailman_api]=/usr/lib64/nagios/plugins/check_http -H localhost -p
# 3 months -> 7884000 # 3 months -> 7884000
command[check_datanommer_anitya]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py anitya 604800 1814400 command[check_datanommer_anitya]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py anitya 604800 1814400
command[check_datanommer_ansible]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py ansible 432000 604800 command[check_datanommer_ansible]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py ansible 432000 604800
command[check_datanommer_autocloud]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py autocloud 259200 1814400
command[check_datanommer_bodhi]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py bodhi 86400 604800 command[check_datanommer_bodhi]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py bodhi 86400 604800
command[check_datanommer_bodhi_composes]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py org.fedoraproject.prod.bodhi.compose.start 86400 90000 command[check_datanommer_bodhi_composes]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py org.fedoraproject.prod.bodhi.compose.start 86400 90000
command[check_datanommer_buildsys]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py buildsys 14400 86400 command[check_datanommer_buildsys]=/usr/lib64/nagios/plugins/check_datanommer_timesince.py buildsys 14400 86400
@ -395,7 +393,6 @@ command[check_fedmsg_cp_notifs_backend]=/usr/lib64/nagios/plugins/check_fedmsg_p
command[check_fedmsg_cp_bugzilla2fedmsg]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py moksha-hub BugzillaConsumer MonitoringProducer command[check_fedmsg_cp_bugzilla2fedmsg]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py moksha-hub BugzillaConsumer MonitoringProducer
command[check_fedmsg_cp_fedimg_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub FedimgConsumer MonitoringProducer command[check_fedmsg_cp_fedimg_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub FedimgConsumer MonitoringProducer
command[check_fedmsg_cp_hotness_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub BugzillaTicketFiler MonitoringProducer command[check_fedmsg_cp_hotness_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub BugzillaTicketFiler MonitoringProducer
command[check_fedmsg_cp_autocloud_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub AutoCloudConsumer MonitoringProducer
command[check_fedmsg_cp_packages_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub CacheInvalidator MonitoringProducer command[check_fedmsg_cp_packages_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub CacheInvalidator MonitoringProducer
command[check_fedmsg_cp_pdc_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub PDCUpdater MonitoringProducer command[check_fedmsg_cp_pdc_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub PDCUpdater MonitoringProducer
command[check_fedmsg_cp_mbs_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub MBSConsumer MBSProducer MonitoringProducer command[check_fedmsg_cp_mbs_backend]=/usr/lib64/nagios/plugins/check_fedmsg_producers_consumers.py fedmsg-hub MBSConsumer MBSProducer MonitoringProducer
@ -410,7 +407,6 @@ command[check_fedmsg_cexceptions_notifs_backend]=/usr/lib64/nagios/plugins/check
command[check_fedmsg_cexceptions_bugzilla2fedmsg]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py moksha-hub BugzillaConsumer 1 10 command[check_fedmsg_cexceptions_bugzilla2fedmsg]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py moksha-hub BugzillaConsumer 1 10
command[check_fedmsg_cexceptions_fedimg_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub FedimgConsumer 1 10 command[check_fedmsg_cexceptions_fedimg_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub FedimgConsumer 1 10
command[check_fedmsg_cexceptions_hotness_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub BugzillaTicketFiler 1 10 command[check_fedmsg_cexceptions_hotness_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub BugzillaTicketFiler 1 10
command[check_fedmsg_cexceptions_autocloud_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub AutoCloudConsumer 1 10
command[check_fedmsg_cexceptions_packages_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub CacheInvalidator 1 10 command[check_fedmsg_cexceptions_packages_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub CacheInvalidator 1 10
command[check_fedmsg_cexceptions_pdc_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub PDCUpdater 1 10 command[check_fedmsg_cexceptions_pdc_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub PDCUpdater 1 10
command[check_fedmsg_cexceptions_mbs_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub MBSConsumer 1 10 command[check_fedmsg_cexceptions_mbs_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_exceptions.py fedmsg-hub MBSConsumer 1 10
@ -425,7 +421,6 @@ command[check_fedmsg_cbacklog_notifs_backend]=/usr/lib64/nagios/plugins/check_fe
command[check_fedmsg_cbacklog_bugzilla2fedmsg]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py moksha-hub BugzillaConsumer 10 100 command[check_fedmsg_cbacklog_bugzilla2fedmsg]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py moksha-hub BugzillaConsumer 10 100
command[check_fedmsg_cbacklog_fedimg_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub FedimgConsumer 2000 5000 command[check_fedmsg_cbacklog_fedimg_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub FedimgConsumer 2000 5000
command[check_fedmsg_cbacklog_hotness_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub BugzillaTicketFiler 100 500 command[check_fedmsg_cbacklog_hotness_backend]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub BugzillaTicketFiler 100 500
command[check_fedmsg_cbacklog_autocloud_backend_hub]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub AutoCloudConsumer 500 1000
command[check_fedmsg_cbacklog_packages_backend_hub]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub CacheInvalidator 30000 40000 command[check_fedmsg_cbacklog_packages_backend_hub]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub CacheInvalidator 30000 40000
command[check_fedmsg_cbacklog_pdc_backend_hub]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub PDCUpdater 10000 20000 command[check_fedmsg_cbacklog_pdc_backend_hub]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub PDCUpdater 10000 20000
command[check_fedmsg_cbacklog_mbs_backend_hub]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub MBSConsumer 1000 2000 command[check_fedmsg_cbacklog_mbs_backend_hub]=/usr/lib64/nagios/plugins/check_fedmsg_consumer_backlog.py fedmsg-hub MBSConsumer 1000 2000

View file

@ -185,7 +185,6 @@
- name: Copy /etc/nagios/services (PHX2) - name: Copy /etc/nagios/services (PHX2)
copy: src=nagios/services/{{ item }} dest=/etc/nagios/services/{{ item }} copy: src=nagios/services/{{ item }} dest=/etc/nagios/services/{{ item }}
with_items: with_items:
- autocloud.cfg
- basset.cfg - basset.cfg
- copr.cfg - copr.cfg
- db_backups.cfg - db_backups.cfg

View file

@ -179,14 +179,6 @@ define service {
## ##
## Other Frontend Websites ## Other Frontend Websites
define service {
hostgroup_name autocloud_web
service_description http-autocloud-internal
check_command check_website!localhost!/autocloud/jobs/!Output
max_check_attempts 8
use internalwebsitetemplate
}
define service { define service {
hostgroup_name badges_web hostgroup_name badges_web
service_description http-badges.fedoraproject.org-tahrir service_description http-badges.fedoraproject.org-tahrir