From 9d898bca65192e493516bf0dc36eb3c885034ad2 Mon Sep 17 00:00:00 2001 From: Joseph Schorr Date: Mon, 6 Oct 2014 12:17:05 -0400 Subject: [PATCH 1/7] Make Hipchat token messaging more clear --- static/js/app.js | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/static/js/app.js b/static/js/app.js index 823de0ad0..2190a13de 100644 --- a/static/js/app.js +++ b/static/js/app.js @@ -1390,7 +1390,8 @@ quayApp = angular.module('quay', quayDependencies, function($provide, cfpLoading { 'name': 'notification_token', 'type': 'string', - 'title': 'Notification Token' + 'title': 'Room Notification Token', + 'help_url': 'https://hipchat.com/rooms/tokens/{room_id}' } ] }, From c4266140e28356d50104906ef196664dbb08aa6d Mon Sep 17 00:00:00 2001 From: Jake Moshenko Date: Mon, 6 Oct 2014 16:40:44 -0400 Subject: [PATCH 2/7] Fix all of the upgrades and downgrades to work on both mysql and postgres. --- ...5b38649_remove_fields_from_image_table_that_.py | 10 +++++----- ...fe1194671_backfill_the_namespace_user_fields.py | 4 ++-- ...a1087b007d_allow_the_namespace_column_to_be_.py | 3 ++- .../f42b0ea7a4d_remove_the_old_webhooks_table.py | 14 ++++++-------- 4 files changed, 15 insertions(+), 16 deletions(-) diff --git a/data/migrations/versions/201d55b38649_remove_fields_from_image_table_that_.py b/data/migrations/versions/201d55b38649_remove_fields_from_image_table_that_.py index d50c3a592..8185c1118 100644 --- a/data/migrations/versions/201d55b38649_remove_fields_from_image_table_that_.py +++ b/data/migrations/versions/201d55b38649_remove_fields_from_image_table_that_.py @@ -44,11 +44,11 @@ def downgrade(tables): op.create_index('notificationkind_name', 'notificationkind', ['name'], unique=False) op.drop_index('logentrykind_name', table_name='logentrykind') op.create_index('logentrykind_name', 'logentrykind', ['name'], unique=False) - op.add_column('image', sa.Column('created', mysql.DATETIME(), nullable=True)) - op.add_column('image', sa.Column('command', mysql.LONGTEXT(), nullable=True)) - op.add_column('image', sa.Column('image_size', mysql.BIGINT(display_width=20), nullable=True)) - op.add_column('image', sa.Column('checksum', mysql.VARCHAR(length=255), nullable=True)) - op.add_column('image', sa.Column('comment', mysql.LONGTEXT(), nullable=True)) + op.add_column('image', sa.Column('created', sa.DateTime(), nullable=True)) + op.add_column('image', sa.Column('command', sa.Text(), nullable=True)) + op.add_column('image', sa.Column('image_size', sa.BigInteger(), nullable=True)) + op.add_column('image', sa.Column('checksum', sa.String(length=255), nullable=True)) + op.add_column('image', sa.Column('comment', sa.Text(), nullable=True)) op.drop_index('buildtriggerservice_name', table_name='buildtriggerservice') op.create_index('buildtriggerservice_name', 'buildtriggerservice', ['name'], unique=False) ### end Alembic commands ### diff --git a/data/migrations/versions/3f4fe1194671_backfill_the_namespace_user_fields.py b/data/migrations/versions/3f4fe1194671_backfill_the_namespace_user_fields.py index 6f40f4fc0..4a1e2fe9d 100644 --- a/data/migrations/versions/3f4fe1194671_backfill_the_namespace_user_fields.py +++ b/data/migrations/versions/3f4fe1194671_backfill_the_namespace_user_fields.py @@ -16,8 +16,8 @@ import sqlalchemy as sa def upgrade(tables): conn = op.get_bind() - conn.execute('update repository set namespace_user_id = (select id from user where user.username = repository.namespace) where namespace_user_id is NULL') - + user_table_name_escaped = conn.dialect.identifier_preparer.format_table(tables['user']) + conn.execute('update repository set namespace_user_id = (select id from {0} where {0}.username = repository.namespace) where namespace_user_id is NULL'.format(user_table_name_escaped)) op.create_index('repository_namespace_user_id_name', 'repository', ['namespace_user_id', 'name'], unique=True) diff --git a/data/migrations/versions/9a1087b007d_allow_the_namespace_column_to_be_.py b/data/migrations/versions/9a1087b007d_allow_the_namespace_column_to_be_.py index 9b63ae190..a0726bf3b 100644 --- a/data/migrations/versions/9a1087b007d_allow_the_namespace_column_to_be_.py +++ b/data/migrations/versions/9a1087b007d_allow_the_namespace_column_to_be_.py @@ -22,7 +22,8 @@ def upgrade(tables): def downgrade(tables): conn = op.get_bind() - conn.execute('update repository set namespace = (select username from user where user.id = repository.namespace_user_id) where namespace is NULL') + user_table_name_escaped = conn.dialect.identifier_preparer.format_table(tables['user']) + conn.execute('update repository set namespace = (select username from {0} where {0}.id = repository.namespace_user_id) where namespace is NULL'.format(user_table_name_escaped)) op.create_index('repository_namespace_name', 'repository', ['namespace', 'name'], unique=True) op.alter_column('repository', 'namespace', nullable=False, existing_type=sa.String(length=255)) diff --git a/data/migrations/versions/f42b0ea7a4d_remove_the_old_webhooks_table.py b/data/migrations/versions/f42b0ea7a4d_remove_the_old_webhooks_table.py index 9ceab4218..5b3f6c812 100644 --- a/data/migrations/versions/f42b0ea7a4d_remove_the_old_webhooks_table.py +++ b/data/migrations/versions/f42b0ea7a4d_remove_the_old_webhooks_table.py @@ -23,13 +23,11 @@ def upgrade(tables): def downgrade(tables): ### commands auto generated by Alembic - please adjust! ### op.create_table('webhook', - sa.Column('id', mysql.INTEGER(display_width=11), nullable=False), - sa.Column('public_id', mysql.VARCHAR(length=255), nullable=False), - sa.Column('repository_id', mysql.INTEGER(display_width=11), autoincrement=False, nullable=False), - sa.Column('parameters', mysql.LONGTEXT(), nullable=False), - sa.ForeignKeyConstraint(['repository_id'], [u'repository.id'], name=u'fk_webhook_repository_repository_id'), - sa.PrimaryKeyConstraint('id'), - mysql_default_charset=u'latin1', - mysql_engine=u'InnoDB' + sa.Column('id', sa.Integer(), nullable=False), + sa.Column('public_id', sa.String(length=255), nullable=False), + sa.Column('repository_id', sa.Integer(), nullable=False), + sa.Column('parameters', sa.Text(), nullable=False), + sa.ForeignKeyConstraint(['repository_id'], ['repository.id'], ), + sa.PrimaryKeyConstraint('id') ) ### end Alembic commands ### From 773c9ac0eeaae0f30dbcc1224d0713089302b6e4 Mon Sep 17 00:00:00 2001 From: Joseph Schorr Date: Mon, 6 Oct 2014 16:56:43 -0400 Subject: [PATCH 3/7] Clarify the language around default permissions --- static/directives/prototype-manager.html | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/static/directives/prototype-manager.html b/static/directives/prototype-manager.html index 7143b13b3..35e5b9c0c 100644 --- a/static/directives/prototype-manager.html +++ b/static/directives/prototype-manager.html @@ -3,7 +3,7 @@
- Default permissions provide a means of specifying additional permissions that should be granted automatically to a repository. + Default permissions provide a means of specifying additional permissions that should be granted automatically to a repository when it is created.
From 153dbc3f92a280f5d9165cdb08fa4de019d46b4a Mon Sep 17 00:00:00 2001 From: Jake Moshenko Date: Mon, 6 Oct 2014 17:15:45 -0400 Subject: [PATCH 4/7] Select random records to use for the backfill script for uncompressed sizes, can now be parallelized. --- tools/uncompressedsize.py | 73 +++++++++++++++++++++------------------ 1 file changed, 39 insertions(+), 34 deletions(-) diff --git a/tools/uncompressedsize.py b/tools/uncompressedsize.py index 608446bae..1182f0838 100644 --- a/tools/uncompressedsize.py +++ b/tools/uncompressedsize.py @@ -4,7 +4,7 @@ import zlib from data import model from data.database import ImageStorage from app import app, storage as store -from data.database import db +from data.database import db, db_random_func from util.gzipstream import ZLIB_GZIP_WINDOW @@ -17,45 +17,50 @@ CHUNK_SIZE = 5 * 1024 * 1024 def backfill_sizes_from_data(): while True: # Load the record from the DB. - try: - record = (ImageStorage - .select(ImageStorage.uuid) - .where(ImageStorage.uncompressed_size >> None, ImageStorage.uploading == False) - .get()) - except ImageStorage.DoesNotExist: + batch_ids = list(ImageStorage + .select(ImageStorage.uuid) + .where(ImageStorage.uncompressed_size >> None, + ImageStorage.uploading == False) + .limit(100) + .order_by(db_random_func())) + if len(batch_ids) == 0: # We're done! return - uuid = record.uuid + for record in batch_ids: + uuid = record.uuid - with_locations = model.get_storage_by_uuid(uuid) - - # Read the layer from backing storage and calculate the uncompressed size. - logger.debug('Loading data: %s (%s bytes)', uuid, with_locations.image_size) - decompressor = zlib.decompressobj(ZLIB_GZIP_WINDOW) - - uncompressed_size = 0 - with store.stream_read_file(with_locations.locations, store.image_layer_path(uuid)) as stream: - while True: - current_data = stream.read(CHUNK_SIZE) - if len(current_data) == 0: - break - - uncompressed_size += len(decompressor.decompress(current_data)) - - # Write the size to the image storage. We do so under a transaction AFTER checking to - # make sure the image storage still exists and has not changed. - logger.debug('Writing entry: %s. Size: %s', uuid, uncompressed_size) - with app.config['DB_TRANSACTION_FACTORY'](db): - try: - current_record = model.get_storage_by_uuid(uuid) - except model.InvalidImageException: - logger.warning('Storage with uuid no longer exists: %s', uuid) + with_locations = model.get_storage_by_uuid(uuid) + if with_locations.uncompressed_size is not None: + logger.debug('Somebody else already filled this in for us: %s', uuid) continue - if not current_record.uploading and current_record.uncompressed_size == None: - current_record.uncompressed_size = uncompressed_size - current_record.save() + # Read the layer from backing storage and calculate the uncompressed size. + logger.debug('Loading data: %s (%s bytes)', uuid, with_locations.image_size) + decompressor = zlib.decompressobj(ZLIB_GZIP_WINDOW) + + uncompressed_size = 0 + with store.stream_read_file(with_locations.locations, store.image_layer_path(uuid)) as stream: + while True: + current_data = stream.read(CHUNK_SIZE) + if len(current_data) == 0: + break + + uncompressed_size += len(decompressor.decompress(current_data)) + + # Write the size to the image storage. We do so under a transaction AFTER checking to + # make sure the image storage still exists and has not changed. + logger.debug('Writing entry: %s. Size: %s', uuid, uncompressed_size) + with app.config['DB_TRANSACTION_FACTORY'](db): + try: + current_record = model.get_storage_by_uuid(uuid) + except model.InvalidImageException: + logger.warning('Storage with uuid no longer exists: %s', uuid) + continue + + if not current_record.uploading and current_record.uncompressed_size == None: + current_record.uncompressed_size = uncompressed_size + current_record.save() if __name__ == "__main__": From 4ad592e7ce837a023e38d1c1d9f8636bfd59b3b6 Mon Sep 17 00:00:00 2001 From: Jake Moshenko Date: Mon, 6 Oct 2014 18:44:37 -0400 Subject: [PATCH 5/7] Add an index to the image storage uuid to improve performance. --- data/database.py | 2 +- ..._add_an_index_to_the_uuid_in_the_image_.py | 22 +++++++++++++++++++ 2 files changed, 23 insertions(+), 1 deletion(-) create mode 100644 data/migrations/versions/b1d41e2071b_add_an_index_to_the_uuid_in_the_image_.py diff --git a/data/database.py b/data/database.py index aba62b88b..f7c4b8931 100644 --- a/data/database.py +++ b/data/database.py @@ -257,7 +257,7 @@ class EmailConfirmation(BaseModel): class ImageStorage(BaseModel): - uuid = CharField(default=uuid_generator) + uuid = CharField(default=uuid_generator, index=True) checksum = CharField(null=True) created = DateTimeField(null=True) comment = TextField(null=True) diff --git a/data/migrations/versions/b1d41e2071b_add_an_index_to_the_uuid_in_the_image_.py b/data/migrations/versions/b1d41e2071b_add_an_index_to_the_uuid_in_the_image_.py new file mode 100644 index 000000000..71a9df794 --- /dev/null +++ b/data/migrations/versions/b1d41e2071b_add_an_index_to_the_uuid_in_the_image_.py @@ -0,0 +1,22 @@ +"""Add an index to the uuid in the image storage table. + +Revision ID: b1d41e2071b +Revises: 9a1087b007d +Create Date: 2014-10-06 18:42:10.021235 + +""" + +# revision identifiers, used by Alembic. +revision = 'b1d41e2071b' +down_revision = '9a1087b007d' + +from alembic import op +import sqlalchemy as sa + + +def upgrade(tables): + op.create_index('imagestorage_uuid', 'imagestorage', ['uuid'], unique=True) + + +def downgrade(tables): + op.drop_index('imagestorage_uuid', table_name='imagestorage') From 55460cdcbaeb03f5a46eaf749fdc026c5bc86d87 Mon Sep 17 00:00:00 2001 From: Jake Moshenko Date: Tue, 7 Oct 2014 10:22:02 -0400 Subject: [PATCH 6/7] Better error handling in the uncompressedsize script. --- tools/uncompressedsize.py | 61 ++++++++++++++++++++------------------- 1 file changed, 32 insertions(+), 29 deletions(-) diff --git a/tools/uncompressedsize.py b/tools/uncompressedsize.py index 1182f0838..0eeaddaae 100644 --- a/tools/uncompressedsize.py +++ b/tools/uncompressedsize.py @@ -30,41 +30,44 @@ def backfill_sizes_from_data(): for record in batch_ids: uuid = record.uuid - with_locations = model.get_storage_by_uuid(uuid) - if with_locations.uncompressed_size is not None: - logger.debug('Somebody else already filled this in for us: %s', uuid) - continue - - # Read the layer from backing storage and calculate the uncompressed size. - logger.debug('Loading data: %s (%s bytes)', uuid, with_locations.image_size) - decompressor = zlib.decompressobj(ZLIB_GZIP_WINDOW) - - uncompressed_size = 0 - with store.stream_read_file(with_locations.locations, store.image_layer_path(uuid)) as stream: - while True: - current_data = stream.read(CHUNK_SIZE) - if len(current_data) == 0: - break - - uncompressed_size += len(decompressor.decompress(current_data)) - - # Write the size to the image storage. We do so under a transaction AFTER checking to - # make sure the image storage still exists and has not changed. - logger.debug('Writing entry: %s. Size: %s', uuid, uncompressed_size) - with app.config['DB_TRANSACTION_FACTORY'](db): - try: - current_record = model.get_storage_by_uuid(uuid) - except model.InvalidImageException: - logger.warning('Storage with uuid no longer exists: %s', uuid) + try: + with_locs = model.get_storage_by_uuid(uuid) + if with_locs.uncompressed_size is not None: + logger.debug('Somebody else already filled this in for us: %s', uuid) continue - if not current_record.uploading and current_record.uncompressed_size == None: - current_record.uncompressed_size = uncompressed_size - current_record.save() + # Read the layer from backing storage and calculate the uncompressed size. + logger.debug('Loading data: %s (%s bytes)', uuid, with_locs.image_size) + decompressor = zlib.decompressobj(ZLIB_GZIP_WINDOW) + uncompressed_size = 0 + with store.stream_read_file(with_locs.locations, store.image_layer_path(uuid)) as stream: + while True: + current_data = stream.read(CHUNK_SIZE) + if len(current_data) == 0: + break + + uncompressed_size += len(decompressor.decompress(current_data)) + + # Write the size to the image storage. We do so under a transaction AFTER checking to + # make sure the image storage still exists and has not changed. + logger.debug('Writing entry: %s. Size: %s', uuid, uncompressed_size) + with app.config['DB_TRANSACTION_FACTORY'](db): + current_record = model.get_storage_by_uuid(uuid) + + if not current_record.uploading and current_record.uncompressed_size == None: + current_record.uncompressed_size = uncompressed_size + current_record.save() + else: + logger.debug('Somebody else already filled this in for us, after we did the work: %s', + uuid) + + except model.InvalidImageException: + logger.warning('Storage with uuid no longer exists: %s', uuid) if __name__ == "__main__": logging.basicConfig(level=logging.DEBUG) logging.getLogger('boto').setLevel(logging.CRITICAL) + logging.getLogger('peewee').setLevel(logging.CRITICAL) backfill_sizes_from_data() From 1850a187d55839c982f5d9b13566224298131eb5 Mon Sep 17 00:00:00 2001 From: Jake Moshenko Date: Tue, 7 Oct 2014 13:13:03 -0400 Subject: [PATCH 7/7] Skip images with memory errors on the backfill. --- tools/uncompressedsize.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/tools/uncompressedsize.py b/tools/uncompressedsize.py index 0eeaddaae..53bb74126 100644 --- a/tools/uncompressedsize.py +++ b/tools/uncompressedsize.py @@ -64,6 +64,8 @@ def backfill_sizes_from_data(): except model.InvalidImageException: logger.warning('Storage with uuid no longer exists: %s', uuid) + except MemoryError: + logger.warning('MemoryError on %s', uuid) if __name__ == "__main__": logging.basicConfig(level=logging.DEBUG)