First stab at the new builder.
This commit is contained in:
parent
335733ad68
commit
72559fb948
6 changed files with 218 additions and 526 deletions
|
@ -1,26 +0,0 @@
|
||||||
FROM lopter/raring-base
|
|
||||||
MAINTAINER jake@devtable.com
|
|
||||||
|
|
||||||
RUN echo deb http://archive.ubuntu.com/ubuntu precise universe > /etc/apt/sources.list.d/universe.list
|
|
||||||
RUN apt-get update -qq
|
|
||||||
RUN apt-get install -qqy iptables ca-certificates lxc python-virtualenv git python-dev xz-utils aufs-tools
|
|
||||||
|
|
||||||
# This will use the latest public release. To use your own, comment it out...
|
|
||||||
ADD https://get.docker.io/builds/Linux/x86_64/docker-latest /usr/local/bin/docker
|
|
||||||
# ...then uncomment the following line, and copy your docker binary to current dir.
|
|
||||||
#ADD ./docker /usr/local/bin/docker
|
|
||||||
|
|
||||||
# Install the files
|
|
||||||
ADD ./startserver /usr/local/bin/startserver
|
|
||||||
ADD ./buildserver.py ./buildserver.py
|
|
||||||
ADD ./requirements.txt ./requirements.txt
|
|
||||||
|
|
||||||
RUN chmod +x /usr/local/bin/docker /usr/local/bin/startserver
|
|
||||||
|
|
||||||
RUN virtualenv --distribute venv
|
|
||||||
RUN venv/bin/pip install -r requirements.txt
|
|
||||||
|
|
||||||
VOLUME /var/lib/docker
|
|
||||||
|
|
||||||
EXPOSE 5002
|
|
||||||
CMD startserver
|
|
|
@ -1,13 +0,0 @@
|
||||||
To build:
|
|
||||||
|
|
||||||
```
|
|
||||||
sudo docker build -t quay.io/quay/buildserver .
|
|
||||||
sudo docker push quay.io/quay/buildserver
|
|
||||||
```
|
|
||||||
|
|
||||||
To run:
|
|
||||||
|
|
||||||
```
|
|
||||||
sudo docker pull quay.io/quay/buildserver
|
|
||||||
sudo docker run -d -privileged -lxc-conf="lxc.aa_profile=unconfined" quay.io/quay/buildserver
|
|
||||||
```
|
|
|
@ -1,214 +0,0 @@
|
||||||
import docker
|
|
||||||
import logging
|
|
||||||
import shutil
|
|
||||||
import os
|
|
||||||
import re
|
|
||||||
import requests
|
|
||||||
import json
|
|
||||||
|
|
||||||
from flask import Flask, jsonify, abort, make_response
|
|
||||||
from zipfile import ZipFile
|
|
||||||
from tempfile import TemporaryFile, mkdtemp
|
|
||||||
from multiprocessing.pool import ThreadPool
|
|
||||||
from base64 import b64encode
|
|
||||||
|
|
||||||
|
|
||||||
BUFFER_SIZE = 8 * 1024
|
|
||||||
LOG_FORMAT = '%(asctime)-15s - %(levelname)s - %(pathname)s - ' + \
|
|
||||||
'%(funcName)s - %(message)s'
|
|
||||||
|
|
||||||
app = Flask(__name__)
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def count_steps(dockerfile_path):
|
|
||||||
with open(dockerfile_path, 'r') as dockerfileobj:
|
|
||||||
steps = 0
|
|
||||||
for line in dockerfileobj.readlines():
|
|
||||||
stripped = line.strip()
|
|
||||||
if stripped and stripped[0] is not '#':
|
|
||||||
steps += 1
|
|
||||||
return steps
|
|
||||||
|
|
||||||
|
|
||||||
def prepare_zip(request_file):
|
|
||||||
build_dir = mkdtemp(prefix='docker-build-')
|
|
||||||
|
|
||||||
# Save the zip file to temp somewhere
|
|
||||||
with TemporaryFile() as zip_file:
|
|
||||||
zip_file.write(request_file.content)
|
|
||||||
to_extract = ZipFile(zip_file)
|
|
||||||
to_extract.extractall(build_dir)
|
|
||||||
|
|
||||||
return build_dir
|
|
||||||
|
|
||||||
|
|
||||||
def prepare_dockerfile(request_file):
|
|
||||||
build_dir = mkdtemp(prefix='docker-build-')
|
|
||||||
dockerfile_path = os.path.join(build_dir, "Dockerfile")
|
|
||||||
with open(dockerfile_path, 'w') as dockerfile:
|
|
||||||
dockerfile.write(request_file.content)
|
|
||||||
|
|
||||||
return build_dir
|
|
||||||
|
|
||||||
|
|
||||||
def total_completion(statuses, total_images):
|
|
||||||
percentage_with_sizes = float(len(statuses.values()))/total_images
|
|
||||||
sent_bytes = sum([status[u'current'] for status in statuses.values()])
|
|
||||||
total_bytes = sum([status[u'total'] for status in statuses.values()])
|
|
||||||
return float(sent_bytes)/total_bytes*percentage_with_sizes
|
|
||||||
|
|
||||||
|
|
||||||
def build_image(build_dir, tag_name, num_steps, result_object):
|
|
||||||
try:
|
|
||||||
logger.debug('Starting build.')
|
|
||||||
docker_cl = docker.Client(timeout=1200)
|
|
||||||
result_object['status'] = 'building'
|
|
||||||
build_status = docker_cl.build(path=build_dir, tag=tag_name, stream=True)
|
|
||||||
|
|
||||||
current_step = 0
|
|
||||||
built_image = None
|
|
||||||
for status in build_status:
|
|
||||||
# logger.debug('Status: %s', str(status))
|
|
||||||
step_increment = re.search(r'Step ([0-9]+) :', status)
|
|
||||||
if step_increment:
|
|
||||||
current_step = int(step_increment.group(1))
|
|
||||||
logger.debug('Step now: %s/%s' % (current_step, num_steps))
|
|
||||||
result_object['current_command'] = current_step
|
|
||||||
continue
|
|
||||||
|
|
||||||
complete = re.match(r'Successfully built ([a-z0-9]+)$', status)
|
|
||||||
if complete:
|
|
||||||
built_image = complete.group(1)
|
|
||||||
logger.debug('Final image ID is: %s' % built_image)
|
|
||||||
continue
|
|
||||||
|
|
||||||
shutil.rmtree(build_dir)
|
|
||||||
|
|
||||||
# Get the image count
|
|
||||||
if not built_image:
|
|
||||||
result_object['status'] = 'error'
|
|
||||||
result_object['message'] = 'Unable to build dockerfile.'
|
|
||||||
return
|
|
||||||
|
|
||||||
history = json.loads(docker_cl.history(built_image))
|
|
||||||
num_images = len(history)
|
|
||||||
result_object['total_images'] = num_images
|
|
||||||
|
|
||||||
result_object['status'] = 'pushing'
|
|
||||||
logger.debug('Pushing to tag name: %s' % tag_name)
|
|
||||||
resp = docker_cl.push(tag_name, stream=True)
|
|
||||||
|
|
||||||
for status_str in resp:
|
|
||||||
status = json.loads(status_str)
|
|
||||||
logger.debug('Status: %s', status_str)
|
|
||||||
if u'status' in status:
|
|
||||||
status_msg = status[u'status']
|
|
||||||
|
|
||||||
if status_msg == 'Pushing':
|
|
||||||
if u'progressDetail' in status and u'id' in status:
|
|
||||||
image_id = status[u'id']
|
|
||||||
detail = status[u'progressDetail']
|
|
||||||
|
|
||||||
if u'current' in detail and 'total' in detail:
|
|
||||||
images = result_object['image_completion']
|
|
||||||
|
|
||||||
images[image_id] = detail
|
|
||||||
result_object['push_completion'] = total_completion(images,
|
|
||||||
num_images)
|
|
||||||
|
|
||||||
elif u'errorDetail' in status:
|
|
||||||
result_object['status'] = 'error'
|
|
||||||
if u'message' in status[u'errorDetail']:
|
|
||||||
result_object['message'] = str(status[u'errorDetail'][u'message'])
|
|
||||||
return
|
|
||||||
|
|
||||||
result_object['status'] = 'complete'
|
|
||||||
except Exception as e:
|
|
||||||
logger.exception('Exception when processing request.')
|
|
||||||
result_object['status'] = 'error'
|
|
||||||
result_object['message'] = str(e.message)
|
|
||||||
|
|
||||||
|
|
||||||
MIME_PROCESSORS = {
|
|
||||||
'application/zip': prepare_zip,
|
|
||||||
'text/plain': prepare_dockerfile,
|
|
||||||
'application/octet-stream': prepare_dockerfile,
|
|
||||||
}
|
|
||||||
|
|
||||||
# If this format it should also be changed in the api method get_repo_builds
|
|
||||||
build = {
|
|
||||||
'total_commands': None,
|
|
||||||
'current_command': None,
|
|
||||||
'push_completion': 0.0,
|
|
||||||
'status': 'waiting',
|
|
||||||
'message': None,
|
|
||||||
'image_completion': {},
|
|
||||||
}
|
|
||||||
pool = ThreadPool(1)
|
|
||||||
|
|
||||||
|
|
||||||
@app.before_first_request
|
|
||||||
def start_build():
|
|
||||||
resource_url = os.environ['RESOURCE_URL']
|
|
||||||
tag_name = os.environ['TAG']
|
|
||||||
acccess_token = os.environ['TOKEN']
|
|
||||||
|
|
||||||
logger.debug('Starting job with resource url: %s tag: %s and token: %s' %
|
|
||||||
(resource_url, tag_name, acccess_token))
|
|
||||||
|
|
||||||
# Save the token
|
|
||||||
host = re.match(r'([a-z0-9.:]+)/.+/.+$', tag_name)
|
|
||||||
if host:
|
|
||||||
docker_endpoint = 'http://%s/v1/' % host.group(1)
|
|
||||||
dockercfg_path = os.path.join(os.environ.get('HOME', '.'), '.dockercfg')
|
|
||||||
token = b64encode('$token:%s' % acccess_token)
|
|
||||||
with open(dockercfg_path, 'w') as dockercfg:
|
|
||||||
payload = {
|
|
||||||
docker_endpoint: {
|
|
||||||
'auth': token,
|
|
||||||
'email': '',
|
|
||||||
}
|
|
||||||
}
|
|
||||||
dockercfg.write(json.dumps(payload))
|
|
||||||
|
|
||||||
else:
|
|
||||||
raise Exception('Invalid tag name: %s' % tag_name)
|
|
||||||
|
|
||||||
docker_resource = requests.get(resource_url)
|
|
||||||
c_type = docker_resource.headers['content-type']
|
|
||||||
|
|
||||||
logger.info('Request to build file of type: %s with tag: %s' %
|
|
||||||
(c_type, tag_name))
|
|
||||||
|
|
||||||
if c_type not in MIME_PROCESSORS:
|
|
||||||
raise Exception('Invalid dockerfile content type: %s' % c_type)
|
|
||||||
|
|
||||||
build_dir = MIME_PROCESSORS[c_type](docker_resource)
|
|
||||||
|
|
||||||
dockerfile_path = os.path.join(build_dir, "Dockerfile")
|
|
||||||
num_steps = count_steps(dockerfile_path)
|
|
||||||
logger.debug('Dockerfile had %s steps' % num_steps)
|
|
||||||
|
|
||||||
logger.info('Sending job to builder pool.')
|
|
||||||
build['total_commands'] = num_steps
|
|
||||||
|
|
||||||
pool.apply_async(build_image, [build_dir, tag_name, num_steps,
|
|
||||||
build])
|
|
||||||
|
|
||||||
|
|
||||||
@app.route('/build/', methods=['GET'])
|
|
||||||
def get_status():
|
|
||||||
if build:
|
|
||||||
return jsonify(build)
|
|
||||||
abort(404)
|
|
||||||
|
|
||||||
|
|
||||||
@app.route('/status/', methods=['GET'])
|
|
||||||
def health_check():
|
|
||||||
return make_response('Running')
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
|
||||||
logging.basicConfig(level=logging.DEBUG, format=LOG_FORMAT)
|
|
||||||
app.run(host='0.0.0.0', port=5002, threaded=True)
|
|
|
@ -1,5 +0,0 @@
|
||||||
mock==1.0.1
|
|
||||||
requests==1.2.3
|
|
||||||
six==1.3.0
|
|
||||||
flask==0.10.1
|
|
||||||
-e git+git://github.com/DevTable/docker-py.git#egg=docker-py
|
|
|
@ -1,48 +0,0 @@
|
||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# First, make sure that cgroups are mounted correctly.
|
|
||||||
CGROUP=/sys/fs/cgroup
|
|
||||||
|
|
||||||
[ -d $CGROUP ] ||
|
|
||||||
mkdir $CGROUP
|
|
||||||
|
|
||||||
mountpoint -q $CGROUP ||
|
|
||||||
mount -n -t tmpfs -o uid=0,gid=0,mode=0755 cgroup $CGROUP || {
|
|
||||||
echo "Could not make a tmpfs mount. Did you use -privileged?"
|
|
||||||
exit 1
|
|
||||||
}
|
|
||||||
|
|
||||||
# Mount the cgroup hierarchies exactly as they are in the parent system.
|
|
||||||
for SUBSYS in $(cut -d: -f2 /proc/1/cgroup)
|
|
||||||
do
|
|
||||||
[ -d $CGROUP/$SUBSYS ] || mkdir $CGROUP/$SUBSYS
|
|
||||||
mountpoint -q $CGROUP/$SUBSYS ||
|
|
||||||
mount -n -t cgroup -o $SUBSYS cgroup $CGROUP/$SUBSYS
|
|
||||||
done
|
|
||||||
|
|
||||||
# Note: as I write those lines, the LXC userland tools cannot setup
|
|
||||||
# a "sub-container" properly if the "devices" cgroup is not in its
|
|
||||||
# own hierarchy. Let's detect this and issue a warning.
|
|
||||||
grep -q :devices: /proc/1/cgroup ||
|
|
||||||
echo "WARNING: the 'devices' cgroup should be in its own hierarchy."
|
|
||||||
grep -qw devices /proc/1/cgroup ||
|
|
||||||
echo "WARNING: it looks like the 'devices' cgroup is not mounted."
|
|
||||||
|
|
||||||
# Now, close extraneous file descriptors.
|
|
||||||
pushd /proc/self/fd
|
|
||||||
for FD in *
|
|
||||||
do
|
|
||||||
case "$FD" in
|
|
||||||
# Keep stdin/stdout/stderr
|
|
||||||
[012])
|
|
||||||
;;
|
|
||||||
# Nuke everything else
|
|
||||||
*)
|
|
||||||
eval exec "$FD>&-"
|
|
||||||
;;
|
|
||||||
esac
|
|
||||||
done
|
|
||||||
popd
|
|
||||||
|
|
||||||
docker -d &
|
|
||||||
exec venv/bin/python buildserver.py
|
|
|
@ -1,20 +1,19 @@
|
||||||
import logging
|
import logging
|
||||||
import json
|
|
||||||
import daemon
|
import daemon
|
||||||
import time
|
|
||||||
import argparse
|
import argparse
|
||||||
import digitalocean
|
|
||||||
import requests
|
|
||||||
import os
|
import os
|
||||||
|
import requests
|
||||||
|
import re
|
||||||
|
import json
|
||||||
|
import shutil
|
||||||
|
|
||||||
from apscheduler.scheduler import Scheduler
|
from docker import Client, APIError
|
||||||
from multiprocessing.pool import ThreadPool
|
from tempfile import TemporaryFile, mkdtemp
|
||||||
from base64 import b64encode
|
from zipfile import ZipFile
|
||||||
from requests.exceptions import ConnectionError
|
|
||||||
|
|
||||||
from data.queue import dockerfile_build_queue
|
from data.queue import dockerfile_build_queue
|
||||||
from data import model
|
from data import model
|
||||||
from data.database import db as db_connection
|
from workers.worker import Worker
|
||||||
from app import app
|
from app import app
|
||||||
|
|
||||||
|
|
||||||
|
@ -26,233 +25,231 @@ formatter = logging.Formatter(FORMAT)
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
BUILD_SERVER_CMD = ('docker run -d -p 5002:5002 ' +
|
|
||||||
'-lxc-conf="lxc.aa_profile=unconfined" ' +
|
|
||||||
'-privileged -e \'RESOURCE_URL=%s\' -e \'TAG=%s\' ' +
|
|
||||||
'-e \'TOKEN=%s\' quay.io/quay/buildserver')
|
|
||||||
|
|
||||||
|
class DockerfileBuildWorker(Worker):
|
||||||
|
def __init__(self, *vargs, **kwargs):
|
||||||
|
super(DockerfileBuildWorker, self).__init__(*vargs, **kwargs)
|
||||||
|
|
||||||
def retry_command(to_call, args=[], kwargs={}, retries=5, period=5):
|
self._mime_processors = {
|
||||||
try:
|
'application/zip': DockerfileBuildWorker.__prepare_zip,
|
||||||
return to_call(*args, **kwargs)
|
'text/plain': DockerfileBuildWorker.__prepare_dockerfile,
|
||||||
except Exception as ex:
|
'application/octet-stream': DockerfileBuildWorker.__prepare_dockerfile,
|
||||||
if retries:
|
|
||||||
logger.debug('Retrying command after %ss' % period)
|
|
||||||
time.sleep(period)
|
|
||||||
return retry_command(to_call, args, kwargs, retries-1, period)
|
|
||||||
raise ex
|
|
||||||
|
|
||||||
|
|
||||||
def get_status(url):
|
|
||||||
return retry_command(requests.get, [url]).json()['status']
|
|
||||||
|
|
||||||
|
|
||||||
def babysit_builder(request):
|
|
||||||
""" Spin up a build node and ask it to build our job. Retryable errors
|
|
||||||
should return False, while fatal errors should return True.
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
logger.debug('Starting work item: %s' % request)
|
|
||||||
repository_build = model.get_repository_build(request['build_id'])
|
|
||||||
logger.debug('Request details: %s' % repository_build)
|
|
||||||
|
|
||||||
# Initialize digital ocean API
|
|
||||||
do_client_id = app.config['DO_CLIENT_ID']
|
|
||||||
do_api_key = app.config['DO_CLIENT_SECRET']
|
|
||||||
manager = digitalocean.Manager(client_id=do_client_id, api_key=do_api_key)
|
|
||||||
|
|
||||||
# check if there is already a DO node for this build, if so clean it up
|
|
||||||
old_id = repository_build.build_node_id
|
|
||||||
if old_id:
|
|
||||||
logger.debug('Cleaning up old DO node: %s' % old_id)
|
|
||||||
old_droplet = digitalocean.Droplet(id=old_id, client_id=do_client_id,
|
|
||||||
api_key=do_api_key)
|
|
||||||
retry_command(old_droplet.destroy)
|
|
||||||
|
|
||||||
# Pick the region for the new droplet
|
|
||||||
allowed_regions = app.config['DO_ALLOWED_REGIONS']
|
|
||||||
regions = retry_command(manager.get_all_regions)
|
|
||||||
available_regions = {region.id for region in regions}
|
|
||||||
regions = available_regions.intersection(allowed_regions)
|
|
||||||
if not regions:
|
|
||||||
logger.error('No droplets in our allowed regtions, available: %s' %
|
|
||||||
available_regions)
|
|
||||||
return False
|
|
||||||
|
|
||||||
# start the DO node
|
|
||||||
name = 'dockerfile-build-%s' % repository_build.id
|
|
||||||
logger.debug('Starting DO node: %s' % name)
|
|
||||||
droplet = digitalocean.Droplet(client_id=do_client_id,
|
|
||||||
api_key=do_api_key,
|
|
||||||
name=name,
|
|
||||||
region_id=regions.pop(),
|
|
||||||
image_id=app.config['DO_DOCKER_IMAGE'],
|
|
||||||
size_id=66, # 512MB,
|
|
||||||
backup_active=False)
|
|
||||||
retry_command(droplet.create, [],
|
|
||||||
{'ssh_key_ids': [app.config['DO_SSH_KEY_ID']]})
|
|
||||||
repository_build.build_node_id = droplet.id
|
|
||||||
repository_build.phase = 'starting'
|
|
||||||
repository_build.save()
|
|
||||||
|
|
||||||
logger.debug('Waiting for DO node to be available.')
|
|
||||||
|
|
||||||
startup = retry_command(droplet.get_events)[0]
|
|
||||||
while not startup.percentage or int(startup.percentage) != 100:
|
|
||||||
logger.debug('Droplet startup percentage: %s' % startup.percentage)
|
|
||||||
time.sleep(5)
|
|
||||||
retry_command(startup.load)
|
|
||||||
|
|
||||||
retry_command(droplet.load)
|
|
||||||
logger.debug('Droplet started at ip address: %s' % droplet.ip_address)
|
|
||||||
|
|
||||||
# connect to it with ssh
|
|
||||||
repository_build.phase = 'initializing'
|
|
||||||
repository_build.save()
|
|
||||||
|
|
||||||
# We wait until here to import paramiko because otherwise it doesn't work
|
|
||||||
# under the daemon context.
|
|
||||||
import paramiko
|
|
||||||
ssh_client = paramiko.SSHClient()
|
|
||||||
ssh_client.set_missing_host_key_policy(paramiko.AutoAddPolicy())
|
|
||||||
|
|
||||||
logger.debug('Connecting to droplet through ssh at ip: %s' %
|
|
||||||
droplet.ip_address)
|
|
||||||
retry_command(ssh_client.connect, [droplet.ip_address, 22, 'root'],
|
|
||||||
{'look_for_keys': False, 'timeout': 10.0,
|
|
||||||
'key_filename': app.config['DO_SSH_PRIVATE_KEY_FILENAME']})
|
|
||||||
|
|
||||||
# Load the node with the pull token
|
|
||||||
token = app.config['BUILD_NODE_PULL_TOKEN']
|
|
||||||
basicauth = b64encode('%s:%s' % ('$token', token))
|
|
||||||
auth_object = {
|
|
||||||
'https://quay.io/v1/': {
|
|
||||||
'auth': basicauth,
|
|
||||||
'email': '',
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
create_auth_cmd = 'echo \'%s\' > .dockercfg' % json.dumps(auth_object)
|
@staticmethod
|
||||||
ssh_client.exec_command(create_auth_cmd)
|
def __count_steps(dockerfile_path):
|
||||||
|
with open(dockerfile_path, 'r') as dockerfileobj:
|
||||||
|
steps = 0
|
||||||
|
for line in dockerfileobj.readlines():
|
||||||
|
stripped = line.strip()
|
||||||
|
if stripped and stripped[0] is not '#':
|
||||||
|
steps += 1
|
||||||
|
return steps
|
||||||
|
|
||||||
# Pull and run the buildserver
|
@staticmethod
|
||||||
pull_cmd = 'docker pull quay.io/quay/buildserver'
|
def __prepare_zip(request_file):
|
||||||
_, stdout, _ = ssh_client.exec_command(pull_cmd)
|
build_dir = mkdtemp(prefix='docker-build-')
|
||||||
pull_status = stdout.channel.recv_exit_status()
|
|
||||||
|
|
||||||
if pull_status != 0:
|
# Save the zip file to temp somewhere
|
||||||
logger.error('Pull command failed for host: %s' % droplet.ip_address)
|
with TemporaryFile() as zip_file:
|
||||||
return False
|
zip_file.write(request_file.content)
|
||||||
else:
|
to_extract = ZipFile(zip_file)
|
||||||
logger.debug('Pull status was: %s' % pull_status)
|
to_extract.extractall(build_dir)
|
||||||
|
|
||||||
# Remove the credentials we used to pull so crafty users cant steal them
|
return build_dir
|
||||||
remove_auth_cmd = 'rm .dockercfg'
|
|
||||||
ssh_client.exec_command(remove_auth_cmd)
|
@staticmethod
|
||||||
|
def __prepare_dockerfile(request_file):
|
||||||
|
build_dir = mkdtemp(prefix='docker-build-')
|
||||||
|
dockerfile_path = os.path.join(build_dir, "Dockerfile")
|
||||||
|
with open(dockerfile_path, 'w') as dockerfile:
|
||||||
|
dockerfile.write(request_file.content)
|
||||||
|
|
||||||
|
return build_dir
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def __total_completion(statuses, total_images):
|
||||||
|
percentage_with_sizes = float(len(statuses.values()))/total_images
|
||||||
|
sent_bytes = sum([status[u'current'] for status in statuses.values()])
|
||||||
|
total_bytes = sum([status[u'total'] for status in statuses.values()])
|
||||||
|
return float(sent_bytes)/total_bytes*percentage_with_sizes
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def __build_image(build_dir, tag_name, num_steps, result_object):
|
||||||
|
try:
|
||||||
|
logger.debug('Starting build.')
|
||||||
|
docker_cl = Client(timeout=1200)
|
||||||
|
|
||||||
|
result_object['status'] = 'building'
|
||||||
|
build_status = docker_cl.build(path=build_dir, tag=tag_name, stream=True)
|
||||||
|
|
||||||
|
current_step = 0
|
||||||
|
built_image = None
|
||||||
|
for status in build_status:
|
||||||
|
# logger.debug('Status: %s', str(status))
|
||||||
|
step_increment = re.search(r'Step ([0-9]+) :', status)
|
||||||
|
if step_increment:
|
||||||
|
current_step = int(step_increment.group(1))
|
||||||
|
logger.debug('Step now: %s/%s' % (current_step, num_steps))
|
||||||
|
result_object['current_command'] = current_step
|
||||||
|
continue
|
||||||
|
|
||||||
|
complete = re.match(r'Successfully built ([a-z0-9]+)$', status)
|
||||||
|
if complete:
|
||||||
|
built_image = complete.group(1)
|
||||||
|
logger.debug('Final image ID is: %s' % built_image)
|
||||||
|
continue
|
||||||
|
|
||||||
|
shutil.rmtree(build_dir)
|
||||||
|
|
||||||
|
# Get the image count
|
||||||
|
if not built_image:
|
||||||
|
result_object['status'] = 'error'
|
||||||
|
result_object['message'] = 'Unable to build dockerfile.'
|
||||||
|
return
|
||||||
|
|
||||||
|
return built_image
|
||||||
|
except Exception as exc:
|
||||||
|
logger.exception('Exception when processing request.')
|
||||||
|
result_object['status'] = 'error'
|
||||||
|
result_object['message'] = str(exc.message)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def __push_image(built_image, token, tag_name, result_object):
|
||||||
|
try:
|
||||||
|
# Login to the registry
|
||||||
|
host = re.match(r'([a-z0-9.:]+)/.+/.+$', tag_name)
|
||||||
|
if not host:
|
||||||
|
raise Exception('Invalid tag name: %s' % tag_name)
|
||||||
|
|
||||||
|
registry_endpoint = 'http://%s/v1/' % host.group(1)
|
||||||
|
|
||||||
|
docker_cl = Client(timeout=1200)
|
||||||
|
logger.debug('Attempting login to registry: %s' % registry_endpoint)
|
||||||
|
docker_cl.login('$token', token, registry=registry_endpoint)
|
||||||
|
|
||||||
|
history = json.loads(docker_cl.history(built_image))
|
||||||
|
num_images = len(history)
|
||||||
|
result_object['total_images'] = num_images
|
||||||
|
|
||||||
|
result_object['status'] = 'pushing'
|
||||||
|
logger.debug('Pushing to tag name: %s' % tag_name)
|
||||||
|
resp = docker_cl.push(tag_name, stream=True)
|
||||||
|
|
||||||
|
for status_str in resp:
|
||||||
|
status = json.loads(status_str)
|
||||||
|
logger.debug('Status: %s', status_str)
|
||||||
|
if u'status' in status:
|
||||||
|
status_msg = status[u'status']
|
||||||
|
|
||||||
|
if status_msg == 'Pushing':
|
||||||
|
if u'progressDetail' in status and u'id' in status:
|
||||||
|
image_id = status[u'id']
|
||||||
|
detail = status[u'progressDetail']
|
||||||
|
|
||||||
|
if u'current' in detail and 'total' in detail:
|
||||||
|
images = result_object['image_completion']
|
||||||
|
|
||||||
|
images[image_id] = detail
|
||||||
|
result_object['push_completion'] = \
|
||||||
|
DockerfileBuildWorker.__total_completion(images, num_images)
|
||||||
|
|
||||||
|
elif u'errorDetail' in status:
|
||||||
|
result_object['status'] = 'error'
|
||||||
|
if u'message' in status[u'errorDetail']:
|
||||||
|
result_object['message'] = str(status[u'errorDetail'][u'message'])
|
||||||
|
return
|
||||||
|
|
||||||
|
result_object['status'] = 'complete'
|
||||||
|
except Exception as exc:
|
||||||
|
logger.exception('Exception when processing request.')
|
||||||
|
result_object['status'] = 'error'
|
||||||
|
result_object['message'] = str(exc.message)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def __cleanup():
|
||||||
|
docker_cl = Client(timeout=1200)
|
||||||
|
|
||||||
|
# First clean up any containers that might be holding the images
|
||||||
|
for running in docker_cl.containers(quiet=True):
|
||||||
|
docker_cl.kill(running['Id'])
|
||||||
|
|
||||||
|
# Next, remove all of the containers (which should all now be killed)
|
||||||
|
for container in docker_cl.containers(all=True, quiet=True):
|
||||||
|
docker_cl.remove_container(container['Id'])
|
||||||
|
|
||||||
|
# Iterate all of the images and remove the ones that the public registry
|
||||||
|
# doesn't know about, this should preserve base images.
|
||||||
|
images_to_remove = set()
|
||||||
|
repos = set()
|
||||||
|
for image in docker_cl.images():
|
||||||
|
images_to_remove.add(image['Id'])
|
||||||
|
repos.add(image['Repository'])
|
||||||
|
|
||||||
|
for repo in repos:
|
||||||
|
repo_url = 'https://index.docker.io/v1/repositories/%s/images' % repo
|
||||||
|
repo_info = requests.get(repo_url)
|
||||||
|
if repo_info.status_code / 100 == 2:
|
||||||
|
for repo_image in repo_info.json():
|
||||||
|
if repo_image['id'] in images_to_remove:
|
||||||
|
logger.debug('Image was deemed public: %s' % repo_image['id'])
|
||||||
|
images_to_remove.remove(repo_image['id'])
|
||||||
|
|
||||||
|
for to_remove in images_to_remove:
|
||||||
|
logger.debug('Removing private image: %s' % to_remove)
|
||||||
|
try:
|
||||||
|
docker_cl.remove_image(to_remove)
|
||||||
|
except APIError:
|
||||||
|
# Sometimes an upstream image removed this one
|
||||||
|
pass
|
||||||
|
|
||||||
|
# Verify that our images were actually removed
|
||||||
|
for image in docker_cl.images():
|
||||||
|
if image['Id'] in images_to_remove:
|
||||||
|
raise RuntimeError('Image was not removed: %s' % image['Id'])
|
||||||
|
|
||||||
|
def process_queue_item(self, job_details):
|
||||||
|
repository_build = model.get_repository_build(job_details['build_id'])
|
||||||
|
|
||||||
# Prepare the signed resource url the build node can fetch the job from
|
|
||||||
user_files = app.config['USERFILES']
|
user_files = app.config['USERFILES']
|
||||||
resource_url = user_files.get_file_url(repository_build.resource_key)
|
resource_url = user_files.get_file_url(repository_build.resource_key)
|
||||||
|
tag_name = repository_build.tag
|
||||||
|
access_token = repository_build.access_token.code
|
||||||
|
|
||||||
# Start the build server
|
feedback = {
|
||||||
start_cmd = BUILD_SERVER_CMD % (resource_url, repository_build.tag,
|
'total_commands': None,
|
||||||
repository_build.access_token.code)
|
'current_command': None,
|
||||||
logger.debug('Sending build server request with command: %s' % start_cmd)
|
'push_completion': 0.0,
|
||||||
ssh_client.exec_command(start_cmd)
|
'status': 'waiting',
|
||||||
|
'message': None,
|
||||||
|
'image_completion': {},
|
||||||
|
}
|
||||||
|
|
||||||
status_endpoint = 'http://%s:5002/build/' % droplet.ip_address
|
logger.debug('Starting job with resource url: %s tag: %s and token: %s' %
|
||||||
# wait for the server to be ready
|
(resource_url, tag_name, access_token))
|
||||||
logger.debug('Waiting for buildserver to be ready')
|
|
||||||
retry_command(requests.get, [status_endpoint])
|
|
||||||
|
|
||||||
# wait for the job to be complete
|
docker_resource = requests.get(resource_url)
|
||||||
repository_build.phase = 'building'
|
c_type = docker_resource.headers['content-type']
|
||||||
repository_build.status_url = status_endpoint
|
|
||||||
repository_build.save()
|
|
||||||
|
|
||||||
logger.debug('Waiting for job to be complete')
|
logger.info('Request to build file of type: %s with tag: %s' %
|
||||||
status = get_status(status_endpoint)
|
(c_type, tag_name))
|
||||||
while status != 'error' and status != 'complete':
|
|
||||||
logger.debug('Job status is: %s' % status)
|
|
||||||
time.sleep(5)
|
|
||||||
status = get_status(status_endpoint)
|
|
||||||
|
|
||||||
logger.debug('Job complete with status: %s' % status)
|
if c_type not in self._mime_processors:
|
||||||
if status == 'error':
|
raise Exception('Invalid dockerfile content type: %s' % c_type)
|
||||||
error_message = requests.get(status_endpoint).json()['message']
|
|
||||||
logger.warning('Job error: %s' % error_message)
|
|
||||||
repository_build.phase = 'error'
|
|
||||||
else:
|
|
||||||
repository_build.phase = 'complete'
|
|
||||||
|
|
||||||
# clean up the DO node
|
build_dir = self._mime_processors[c_type](docker_resource)
|
||||||
logger.debug('Cleaning up DO node.')
|
|
||||||
retry_command(droplet.destroy)
|
|
||||||
|
|
||||||
repository_build.status_url = None
|
dockerfile_path = os.path.join(build_dir, "Dockerfile")
|
||||||
repository_build.build_node_id = None
|
num_steps = DockerfileBuildWorker.__count_steps(dockerfile_path)
|
||||||
repository_build.save()
|
logger.debug('Dockerfile had %s steps' % num_steps)
|
||||||
|
|
||||||
return True
|
built_image = DockerfileBuildWorker.__build_image(build_dir, tag_name,
|
||||||
|
num_steps, feedback)
|
||||||
|
|
||||||
except Exception as outer_ex:
|
DockerfileBuildWorker.__push_image(built_image, access_token, tag_name,
|
||||||
# We don't really know what these are, but they are probably retryable
|
feedback)
|
||||||
logger.exception('Exception processing job: %s' % outer_ex.message)
|
|
||||||
return False
|
|
||||||
|
|
||||||
finally:
|
DockerfileBuildWorker.__cleanup()
|
||||||
if not db_connection.is_closed():
|
|
||||||
logger.debug('Closing thread db connection.')
|
|
||||||
db_connection.close()
|
|
||||||
|
|
||||||
|
|
||||||
def process_work_items(pool):
|
|
||||||
logger.debug('Getting work item from queue.')
|
|
||||||
|
|
||||||
item = dockerfile_build_queue.get(processing_time=60*60) # allow 1 hr
|
|
||||||
|
|
||||||
while item:
|
|
||||||
logger.debug('Queue gave us some work: %s' % item.body)
|
|
||||||
|
|
||||||
request = json.loads(item.body)
|
|
||||||
|
|
||||||
def build_callback(item):
|
|
||||||
local_item = item
|
|
||||||
def complete_callback(completed):
|
|
||||||
if completed:
|
|
||||||
logger.debug('Queue item completed successfully, will be removed.')
|
|
||||||
dockerfile_build_queue.complete(local_item)
|
|
||||||
else:
|
|
||||||
# We have a retryable error, add the job back to the queue
|
|
||||||
logger.debug('Queue item incomplete, will be retryed.')
|
|
||||||
dockerfile_build_queue.incomplete(local_item)
|
|
||||||
|
|
||||||
return complete_callback
|
|
||||||
|
|
||||||
logger.debug('Sending work item to thread pool: %s' % pool)
|
|
||||||
pool.apply_async(babysit_builder, [request],
|
|
||||||
callback=build_callback(item))
|
|
||||||
|
|
||||||
item = dockerfile_build_queue.get()
|
|
||||||
|
|
||||||
logger.debug('No more work.')
|
|
||||||
|
|
||||||
if not db_connection.is_closed():
|
|
||||||
logger.debug('Closing thread db connection.')
|
|
||||||
db_connection.close()
|
|
||||||
|
|
||||||
|
|
||||||
def start_worker():
|
|
||||||
pool = ThreadPool(3)
|
|
||||||
logger.debug('Scheduling worker.')
|
|
||||||
|
|
||||||
sched = Scheduler()
|
|
||||||
sched.start()
|
|
||||||
|
|
||||||
sched.add_interval_job(process_work_items, args=[pool], seconds=30)
|
|
||||||
|
|
||||||
while True:
|
|
||||||
time.sleep(60 * 60 * 24) # sleep one day, basically forever
|
|
||||||
|
|
||||||
|
|
||||||
desc = 'Worker daemon to monitor dockerfile build'
|
desc = 'Worker daemon to monitor dockerfile build'
|
||||||
|
@ -264,16 +261,17 @@ parser.add_argument('--log', default='dockerfilebuild.log',
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
|
|
||||||
|
worker = DockerfileBuildWorker(dockerfile_build_queue)
|
||||||
|
|
||||||
if args.D:
|
if args.D:
|
||||||
handler = logging.FileHandler(args.log)
|
handler = logging.FileHandler(args.log)
|
||||||
handler.setFormatter(formatter)
|
handler.setFormatter(formatter)
|
||||||
root_logger.addHandler(handler)
|
root_logger.addHandler(handler)
|
||||||
with daemon.DaemonContext(files_preserve=[handler.stream],
|
with daemon.DaemonContext(files_preserve=[handler.stream]):
|
||||||
working_directory=os.getcwd()):
|
worker.start()
|
||||||
start_worker()
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
handler = logging.StreamHandler()
|
handler = logging.StreamHandler()
|
||||||
handler.setFormatter(formatter)
|
handler.setFormatter(formatter)
|
||||||
root_logger.addHandler(handler)
|
root_logger.addHandler(handler)
|
||||||
start_worker()
|
worker.start()
|
Reference in a new issue