refactored DockerfileServiceImpl to return promise instead of callbacks
This commit is contained in:
commit
4e913f106d
34 changed files with 299 additions and 490 deletions
1
.gitignore
vendored
1
.gitignore
vendored
|
@ -22,3 +22,4 @@ htmlcov
|
||||||
.tox
|
.tox
|
||||||
.cache
|
.cache
|
||||||
.npm-debug.log
|
.npm-debug.log
|
||||||
|
Dockerfile-e
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
import datetime
|
import datetime
|
||||||
|
import os
|
||||||
import time
|
import time
|
||||||
import logging
|
import logging
|
||||||
import json
|
import json
|
||||||
|
@ -122,9 +123,21 @@ class BuildComponent(BaseComponent):
|
||||||
# base_image: The image name and credentials to use to conduct the base image pull.
|
# base_image: The image name and credentials to use to conduct the base image pull.
|
||||||
# username: The username for pulling the base image (if any).
|
# username: The username for pulling the base image (if any).
|
||||||
# password: The password for pulling the base image (if any).
|
# password: The password for pulling the base image (if any).
|
||||||
|
|
||||||
|
subdir, dockerfile_name = os.path.split(build_config.get('build_subdir', '/Dockerfile'))
|
||||||
|
|
||||||
|
# HACK HACK HACK HACK HACK HACK HACK
|
||||||
|
# TODO: FIX THIS in the database and then turn the split back on.
|
||||||
|
if dockerfile_name.find('Dockerfile') < 0:
|
||||||
|
# This is a *HACK* for the broken path handling. To be fixed ASAP.
|
||||||
|
subdir = build_config.get('build_subdir') or '/'
|
||||||
|
dockerfile_name = 'Dockerfile'
|
||||||
|
# /HACK HACK HACK HACK HACK HACK HACK
|
||||||
|
|
||||||
build_arguments = {
|
build_arguments = {
|
||||||
'build_package': build_job.get_build_package_url(self.user_files),
|
'build_package': build_job.get_build_package_url(self.user_files),
|
||||||
'sub_directory': build_config.get('build_subdir', ''),
|
'sub_directory': subdir,
|
||||||
|
'dockerfile_name': dockerfile_name,
|
||||||
'repository': repository_name,
|
'repository': repository_name,
|
||||||
'registry': self.registry_hostname,
|
'registry': self.registry_hostname,
|
||||||
'pull_token': build_job.repo_build.access_token.code,
|
'pull_token': build_job.repo_build.access_token.code,
|
||||||
|
|
|
@ -276,6 +276,7 @@ class PopenExecutor(BuilderExecutor):
|
||||||
'DOCKER_TLS_VERIFY': os.environ.get('DOCKER_TLS_VERIFY', ''),
|
'DOCKER_TLS_VERIFY': os.environ.get('DOCKER_TLS_VERIFY', ''),
|
||||||
'DOCKER_CERT_PATH': os.environ.get('DOCKER_CERT_PATH', ''),
|
'DOCKER_CERT_PATH': os.environ.get('DOCKER_CERT_PATH', ''),
|
||||||
'DOCKER_HOST': os.environ.get('DOCKER_HOST', ''),
|
'DOCKER_HOST': os.environ.get('DOCKER_HOST', ''),
|
||||||
|
'PATH': "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
|
||||||
}
|
}
|
||||||
|
|
||||||
logpipe = LogPipe(logging.INFO)
|
logpipe = LogPipe(logging.INFO)
|
||||||
|
|
|
@ -258,6 +258,11 @@ class BuildTriggerHandler(object):
|
||||||
can be called in a loop, so it should be as fast as possible. """
|
can be called in a loop, so it should be as fast as possible. """
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def path_is_dockerfile(cls, file_name):
|
||||||
|
""" Returns whether the file is named Dockerfile or follows the convention <name>.Dockerfile"""
|
||||||
|
return file_name.endswith(".Dockerfile") or u"Dockerfile" == file_name
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def service_name(cls):
|
def service_name(cls):
|
||||||
"""
|
"""
|
||||||
|
@ -285,14 +290,10 @@ class BuildTriggerHandler(object):
|
||||||
def get_dockerfile_path(self):
|
def get_dockerfile_path(self):
|
||||||
""" Returns the normalized path to the Dockerfile found in the subdirectory
|
""" Returns the normalized path to the Dockerfile found in the subdirectory
|
||||||
in the config. """
|
in the config. """
|
||||||
subdirectory = self.config.get('subdir', '')
|
dockerfile_path = self.config.get('subdir') or 'Dockerfile'
|
||||||
if subdirectory == '/':
|
if dockerfile_path[0] == '/':
|
||||||
subdirectory = ''
|
dockerfile_path = dockerfile_path[1:]
|
||||||
else:
|
return dockerfile_path
|
||||||
if not subdirectory.endswith('/'):
|
|
||||||
subdirectory = subdirectory + '/'
|
|
||||||
|
|
||||||
return subdirectory + 'Dockerfile'
|
|
||||||
|
|
||||||
def prepare_build(self, metadata, is_manual=False):
|
def prepare_build(self, metadata, is_manual=False):
|
||||||
# Ensure that the metadata meets the scheme.
|
# Ensure that the metadata meets the scheme.
|
||||||
|
|
|
@ -1,23 +1,21 @@
|
||||||
import logging
|
import logging
|
||||||
|
import os
|
||||||
import re
|
import re
|
||||||
|
|
||||||
from calendar import timegm
|
from calendar import timegm
|
||||||
|
|
||||||
import dateutil.parser
|
import dateutil.parser
|
||||||
|
from bitbucket import BitBucket
|
||||||
from jsonschema import validate
|
from jsonschema import validate
|
||||||
|
|
||||||
|
from app import app, get_app_url
|
||||||
|
from buildtrigger.basehandler import BuildTriggerHandler
|
||||||
from buildtrigger.triggerutil import (RepositoryReadException, TriggerActivationException,
|
from buildtrigger.triggerutil import (RepositoryReadException, TriggerActivationException,
|
||||||
TriggerDeactivationException, TriggerStartException,
|
TriggerDeactivationException, TriggerStartException,
|
||||||
InvalidPayloadException, TriggerProviderException,
|
InvalidPayloadException, TriggerProviderException,
|
||||||
determine_build_ref, raise_if_skipped_build,
|
determine_build_ref, raise_if_skipped_build,
|
||||||
find_matching_branches)
|
find_matching_branches)
|
||||||
|
|
||||||
from buildtrigger.basehandler import BuildTriggerHandler
|
|
||||||
|
|
||||||
from app import app, get_app_url
|
|
||||||
from bitbucket import BitBucket
|
|
||||||
from util.security.ssh import generate_ssh_keypair
|
|
||||||
from util.dict_wrappers import JSONPathDict, SafeDictSetter
|
from util.dict_wrappers import JSONPathDict, SafeDictSetter
|
||||||
|
from util.security.ssh import generate_ssh_keypair
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
@ -455,10 +453,7 @@ class BitbucketBuildTrigger(BuildTriggerHandler):
|
||||||
raise RepositoryReadException(err_msg)
|
raise RepositoryReadException(err_msg)
|
||||||
|
|
||||||
files = set([f['path'] for f in data['files']])
|
files = set([f['path'] for f in data['files']])
|
||||||
if 'Dockerfile' in files:
|
return ["/" + file_path for file_path in files if self.path_is_dockerfile(os.path.basename(file_path))]
|
||||||
return ['']
|
|
||||||
|
|
||||||
return []
|
|
||||||
|
|
||||||
def load_dockerfile_contents(self):
|
def load_dockerfile_contents(self):
|
||||||
repository = self._get_repository_client()
|
repository = self._get_repository_client()
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
import logging
|
import logging
|
||||||
import os.path
|
import os.path
|
||||||
import base64
|
import base64
|
||||||
|
import re
|
||||||
|
|
||||||
from calendar import timegm
|
from calendar import timegm
|
||||||
from functools import wraps
|
from functools import wraps
|
||||||
|
@ -348,9 +349,8 @@ class GithubBuildTrigger(BuildTriggerHandler):
|
||||||
default_commit = repo.get_branch(branches[0]).commit
|
default_commit = repo.get_branch(branches[0]).commit
|
||||||
commit_tree = repo.get_git_tree(default_commit.sha, recursive=True)
|
commit_tree = repo.get_git_tree(default_commit.sha, recursive=True)
|
||||||
|
|
||||||
return [os.path.dirname(elem.path) for elem in commit_tree.tree
|
return [elem.path for elem in commit_tree.tree
|
||||||
if (elem.type == u'blob' and
|
if (elem.type == u'blob' and self.path_is_dockerfile(os.path.basename(elem.path)))]
|
||||||
os.path.basename(elem.path) == u'Dockerfile')]
|
|
||||||
except GithubException as ghe:
|
except GithubException as ghe:
|
||||||
message = ghe.data.get('message', 'Unable to list contents of repository: %s' % source)
|
message = ghe.data.get('message', 'Unable to list contents of repository: %s' % source)
|
||||||
if message == 'Branch not found':
|
if message == 'Branch not found':
|
||||||
|
@ -371,6 +371,9 @@ class GithubBuildTrigger(BuildTriggerHandler):
|
||||||
raise RepositoryReadException(message)
|
raise RepositoryReadException(message)
|
||||||
|
|
||||||
path = self.get_dockerfile_path()
|
path = self.get_dockerfile_path()
|
||||||
|
if not path:
|
||||||
|
return None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
file_info = repo.get_file_contents(path)
|
file_info = repo.get_file_contents(path)
|
||||||
except GithubException as ghe:
|
except GithubException as ghe:
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
import logging
|
import logging
|
||||||
|
import os
|
||||||
|
|
||||||
from calendar import timegm
|
from calendar import timegm
|
||||||
from functools import wraps
|
from functools import wraps
|
||||||
|
@ -341,11 +342,7 @@ class GitLabBuildTrigger(BuildTriggerHandler):
|
||||||
msg = 'Unable to find GitLab repository tree for source: %s' % new_build_source
|
msg = 'Unable to find GitLab repository tree for source: %s' % new_build_source
|
||||||
raise RepositoryReadException(msg)
|
raise RepositoryReadException(msg)
|
||||||
|
|
||||||
for node in repo_tree:
|
return ["/"+node['name'] for node in repo_tree if self.path_is_dockerfile(node['name'])]
|
||||||
if node['name'] == 'Dockerfile':
|
|
||||||
return ['']
|
|
||||||
|
|
||||||
return []
|
|
||||||
|
|
||||||
@_catch_timeouts
|
@_catch_timeouts
|
||||||
def load_dockerfile_contents(self):
|
def load_dockerfile_contents(self):
|
||||||
|
|
|
@ -23,7 +23,7 @@ def get_repo_path_contents(path, revision):
|
||||||
return (True, data, None)
|
return (True, data, None)
|
||||||
|
|
||||||
def get_raw_path_contents(path, revision):
|
def get_raw_path_contents(path, revision):
|
||||||
if path == '/Dockerfile':
|
if path == 'Dockerfile':
|
||||||
return (True, 'hello world', None)
|
return (True, 'hello world', None)
|
||||||
|
|
||||||
if path == 'somesubdir/Dockerfile':
|
if path == 'somesubdir/Dockerfile':
|
||||||
|
|
|
@ -119,7 +119,7 @@ def get_mock_github():
|
||||||
return [master, otherbranch]
|
return [master, otherbranch]
|
||||||
|
|
||||||
def get_file_contents_mock(filepath):
|
def get_file_contents_mock(filepath):
|
||||||
if filepath == '/Dockerfile':
|
if filepath == 'Dockerfile':
|
||||||
m = Mock()
|
m = Mock()
|
||||||
m.content = 'hello world'
|
m.content = 'hello world'
|
||||||
return m
|
return m
|
||||||
|
|
|
@ -151,7 +151,7 @@ def gettag_mock(repo_id, tag):
|
||||||
}
|
}
|
||||||
|
|
||||||
def getrawfile_mock(repo_id, branch_name, path):
|
def getrawfile_mock(repo_id, branch_name, path):
|
||||||
if path == '/Dockerfile':
|
if path == 'Dockerfile':
|
||||||
return 'hello world'
|
return 'hello world'
|
||||||
|
|
||||||
if path == 'somesubdir/Dockerfile':
|
if path == 'somesubdir/Dockerfile':
|
||||||
|
|
15
buildtrigger/test/test_basehandler.py
Normal file
15
buildtrigger/test/test_basehandler.py
Normal file
|
@ -0,0 +1,15 @@
|
||||||
|
import pytest
|
||||||
|
|
||||||
|
from buildtrigger.basehandler import BuildTriggerHandler
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.parametrize('input,output', [
|
||||||
|
("Dockerfile", True),
|
||||||
|
("server.Dockerfile", True),
|
||||||
|
(u"Dockerfile", True),
|
||||||
|
(u"server.Dockerfile", True),
|
||||||
|
("bad file name", False),
|
||||||
|
(u"bad file name", False),
|
||||||
|
])
|
||||||
|
def test_path_is_dockerfile(input, output):
|
||||||
|
assert BuildTriggerHandler.path_is_dockerfile(input) == output
|
|
@ -13,12 +13,12 @@ def bitbucket_trigger():
|
||||||
|
|
||||||
|
|
||||||
def test_list_build_subdirs(bitbucket_trigger):
|
def test_list_build_subdirs(bitbucket_trigger):
|
||||||
assert bitbucket_trigger.list_build_subdirs() == ['']
|
assert bitbucket_trigger.list_build_subdirs() == ["/Dockerfile"]
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('subdir, contents', [
|
@pytest.mark.parametrize('subdir, contents', [
|
||||||
('', 'hello world'),
|
('/Dockerfile', 'hello world'),
|
||||||
('somesubdir', 'hi universe'),
|
('somesubdir/Dockerfile', 'hi universe'),
|
||||||
('unknownpath', None),
|
('unknownpath', None),
|
||||||
])
|
])
|
||||||
def test_load_dockerfile_contents(subdir, contents):
|
def test_load_dockerfile_contents(subdir, contents):
|
||||||
|
|
|
@ -68,8 +68,8 @@ def test_handle_trigger_request(github_trigger, payload, expected_error, expecte
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('subdir, contents', [
|
@pytest.mark.parametrize('subdir, contents', [
|
||||||
('', 'hello world'),
|
('/Dockerfile', 'hello world'),
|
||||||
('somesubdir', 'hi universe'),
|
('somesubdir/Dockerfile', 'hi universe'),
|
||||||
('unknownpath', None),
|
('unknownpath', None),
|
||||||
])
|
])
|
||||||
def test_load_dockerfile_contents(subdir, contents):
|
def test_load_dockerfile_contents(subdir, contents):
|
||||||
|
@ -86,4 +86,4 @@ def test_lookup_user(username, expected_response, github_trigger):
|
||||||
|
|
||||||
|
|
||||||
def test_list_build_subdirs(github_trigger):
|
def test_list_build_subdirs(github_trigger):
|
||||||
assert github_trigger.list_build_subdirs() == ['', 'somesubdir']
|
assert github_trigger.list_build_subdirs() == ['Dockerfile', 'somesubdir/Dockerfile']
|
||||||
|
|
|
@ -13,12 +13,12 @@ def gitlab_trigger():
|
||||||
|
|
||||||
|
|
||||||
def test_list_build_subdirs(gitlab_trigger):
|
def test_list_build_subdirs(gitlab_trigger):
|
||||||
assert gitlab_trigger.list_build_subdirs() == ['']
|
assert gitlab_trigger.list_build_subdirs() == ['/Dockerfile']
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('subdir, contents', [
|
@pytest.mark.parametrize('subdir, contents', [
|
||||||
('', 'hello world'),
|
('/Dockerfile', 'hello world'),
|
||||||
('somesubdir', 'hi universe'),
|
('somesubdir/Dockerfile', 'hi universe'),
|
||||||
('unknownpath', None),
|
('unknownpath', None),
|
||||||
])
|
])
|
||||||
def test_load_dockerfile_contents(subdir, contents):
|
def test_load_dockerfile_contents(subdir, contents):
|
||||||
|
|
|
@ -9,6 +9,5 @@ run:
|
||||||
- |
|
- |
|
||||||
set -eux
|
set -eux
|
||||||
cd quay-pull-request
|
cd quay-pull-request
|
||||||
npm install
|
yarn install --ignore-engines
|
||||||
npm link typescript
|
yarn test
|
||||||
npm test
|
|
||||||
|
|
|
@ -2,7 +2,7 @@ import logging
|
||||||
|
|
||||||
from uuid import uuid4
|
from uuid import uuid4
|
||||||
|
|
||||||
from peewee import IntegrityError
|
from peewee import IntegrityError, JOIN_LEFT_OUTER, fn
|
||||||
from data.model import (image, db_transaction, DataModelException, _basequery,
|
from data.model import (image, db_transaction, DataModelException, _basequery,
|
||||||
InvalidManifestException, TagAlreadyCreatedException, StaleTagException)
|
InvalidManifestException, TagAlreadyCreatedException, StaleTagException)
|
||||||
from data.database import (RepositoryTag, Repository, Image, ImageStorage, Namespace, TagManifest,
|
from data.database import (RepositoryTag, Repository, Image, ImageStorage, Namespace, TagManifest,
|
||||||
|
@ -13,6 +13,40 @@ from data.database import (RepositoryTag, Repository, Image, ImageStorage, Names
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def get_max_id_for_sec_scan():
|
||||||
|
""" Gets the maximum id for security scanning """
|
||||||
|
return RepositoryTag.select(fn.Max(RepositoryTag.id)).scalar()
|
||||||
|
|
||||||
|
|
||||||
|
def get_min_id_for_sec_scan(version):
|
||||||
|
""" Gets the minimum id for a security scanning """
|
||||||
|
return (RepositoryTag
|
||||||
|
.select(fn.Min(RepositoryTag.id))
|
||||||
|
.join(Image)
|
||||||
|
.where(Image.security_indexed_engine < version)
|
||||||
|
.scalar())
|
||||||
|
|
||||||
|
|
||||||
|
def get_tag_pk_field():
|
||||||
|
""" Returns the primary key for Image DB model """
|
||||||
|
return RepositoryTag.id
|
||||||
|
|
||||||
|
|
||||||
|
def get_tags_images_eligible_for_scan(clair_version):
|
||||||
|
Parent = Image.alias()
|
||||||
|
ParentImageStorage = ImageStorage.alias()
|
||||||
|
|
||||||
|
return _tag_alive(RepositoryTag
|
||||||
|
.select(Image, ImageStorage, Parent, ParentImageStorage, RepositoryTag)
|
||||||
|
.join(Image, on=(RepositoryTag.image == Image.id))
|
||||||
|
.join(ImageStorage, on=(Image.storage == ImageStorage.id))
|
||||||
|
.switch(Image)
|
||||||
|
.join(Parent, JOIN_LEFT_OUTER, on=(Image.parent == Parent.id))
|
||||||
|
.join(ParentImageStorage, JOIN_LEFT_OUTER, on=(ParentImageStorage.id == Parent.storage))
|
||||||
|
.where(RepositoryTag.hidden == False)
|
||||||
|
.where(Image.security_indexed_engine < clair_version))
|
||||||
|
|
||||||
|
|
||||||
def _tag_alive(query, now_ts=None):
|
def _tag_alive(query, now_ts=None):
|
||||||
if now_ts is None:
|
if now_ts is None:
|
||||||
now_ts = get_epoch_timestamp()
|
now_ts = get_epoch_timestamp()
|
||||||
|
|
|
@ -404,7 +404,7 @@ class ActivateBuildTrigger(RepositoryParamResource):
|
||||||
'description': '(Custom Only) If specified, the ref/SHA1 used to checkout a git repository.'
|
'description': '(Custom Only) If specified, the ref/SHA1 used to checkout a git repository.'
|
||||||
},
|
},
|
||||||
'refs': {
|
'refs': {
|
||||||
'type': 'object',
|
'type': ['object', 'null'],
|
||||||
'description': '(SCM Only) If specified, the ref to build.'
|
'description': '(SCM Only) If specified, the ref to build.'
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
|
15
package.json
15
package.json
|
@ -15,11 +15,11 @@
|
||||||
},
|
},
|
||||||
"homepage": "https://github.com/coreos-inc/quay#readme",
|
"homepage": "https://github.com/coreos-inc/quay#readme",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"angular": "1.5.3",
|
"angular": "1.6.2",
|
||||||
"angular-animate": "^1.5.3",
|
"angular-animate": "1.6.2",
|
||||||
"angular-cookies": "^1.5.3",
|
"angular-cookies": "1.6.2",
|
||||||
"angular-route": "^1.5.3",
|
"angular-route": "1.6.2",
|
||||||
"angular-sanitize": "^1.5.3",
|
"angular-sanitize": "1.6.2",
|
||||||
"bootbox": "^4.1.0",
|
"bootbox": "^4.1.0",
|
||||||
"bootstrap": "^3.3.2",
|
"bootstrap": "^3.3.2",
|
||||||
"bootstrap-datepicker": "^1.6.4",
|
"bootstrap-datepicker": "^1.6.4",
|
||||||
|
@ -35,15 +35,16 @@
|
||||||
"underscore": "^1.5.2"
|
"underscore": "^1.5.2"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@types/angular": "1.5.16",
|
"@types/angular": "1.6.2",
|
||||||
"@types/angular-mocks": "^1.5.8",
|
"@types/angular-mocks": "^1.5.8",
|
||||||
"@types/angular-route": "^1.3.3",
|
"@types/angular-route": "^1.3.3",
|
||||||
"@types/angular-sanitize": "^1.3.4",
|
"@types/angular-sanitize": "^1.3.4",
|
||||||
"@types/es6-shim": "^0.31.32",
|
"@types/es6-shim": "^0.31.32",
|
||||||
"@types/jasmine": "^2.5.41",
|
"@types/jasmine": "^2.5.41",
|
||||||
|
"@types/jquery": "^2.0.40",
|
||||||
"@types/react": "0.14.39",
|
"@types/react": "0.14.39",
|
||||||
"@types/react-dom": "0.14.17",
|
"@types/react-dom": "0.14.17",
|
||||||
"angular-mocks": "^1.5.3",
|
"angular-mocks": "1.6.2",
|
||||||
"angular-ts-decorators": "0.0.19",
|
"angular-ts-decorators": "0.0.19",
|
||||||
"css-loader": "0.25.0",
|
"css-loader": "0.25.0",
|
||||||
"jasmine-core": "^2.5.2",
|
"jasmine-core": "^2.5.2",
|
||||||
|
|
|
@ -30,6 +30,12 @@
|
||||||
height: 28px;
|
height: 28px;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@media (max-width: 768px) {
|
||||||
|
.manage-trigger-control .co-top-bar {
|
||||||
|
margin-bottom: 80px;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
.manage-trigger-control .namespace-avatar {
|
.manage-trigger-control .namespace-avatar {
|
||||||
margin-left: 2px;
|
margin-left: 2px;
|
||||||
margin-right: 2px;
|
margin-right: 2px;
|
||||||
|
|
|
@ -26,7 +26,8 @@
|
||||||
icon-key="kind"
|
icon-key="kind"
|
||||||
icon-map="field.iconMap"
|
icon-map="field.iconMap"
|
||||||
items="fieldOptions[field.name]"
|
items="fieldOptions[field.name]"
|
||||||
ng-show="fieldOptions[field.name]"></div>
|
ng-show="fieldOptions[field.name]"
|
||||||
|
clear-value="counter"></div>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Option -->
|
<!-- Option -->
|
||||||
|
|
|
@ -39,28 +39,7 @@ angular.module('quay').directive('dockerfileBuildForm', function () {
|
||||||
$scope.state = 'checking';
|
$scope.state = 'checking';
|
||||||
$scope.selectedFiles = files;
|
$scope.selectedFiles = files;
|
||||||
|
|
||||||
// FIXME: Remove this
|
DockerfileService.getDockerfile(files[0])
|
||||||
// DockerfileService.getDockerfile(files[0], function(df) {
|
|
||||||
// var baseImage = df.getRegistryBaseImage();
|
|
||||||
// if (baseImage) {
|
|
||||||
// checkPrivateImage(baseImage);
|
|
||||||
// } else {
|
|
||||||
// $scope.state = 'ready';
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// $scope.$apply(function() {
|
|
||||||
// opt_callback && opt_callback(true, 'Dockerfile found and valid')
|
|
||||||
// });
|
|
||||||
// }, function(msg) {
|
|
||||||
// $scope.state = 'empty';
|
|
||||||
// $scope.privateBaseRepository = null;
|
|
||||||
//
|
|
||||||
// $scope.$apply(function() {
|
|
||||||
// opt_callback && opt_callback(false, msg || 'Could not find valid Dockerfile');
|
|
||||||
// });
|
|
||||||
// });
|
|
||||||
|
|
||||||
DockerfileService.extractDockerfile(files[0])
|
|
||||||
.then(function(dockerfileInfo) {
|
.then(function(dockerfileInfo) {
|
||||||
var baseImage = dockerfileInfo.getRegistryBaseImage();
|
var baseImage = dockerfileInfo.getRegistryBaseImage();
|
||||||
if (baseImage) {
|
if (baseImage) {
|
||||||
|
|
|
@ -25,7 +25,7 @@
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<table class="co-table" style="margin-top: 20px;">
|
<table class="co-table">
|
||||||
<thead>
|
<thead>
|
||||||
<td class="checkbox-col"></td>
|
<td class="checkbox-col"></td>
|
||||||
<td ng-class="$ctrl.TableService.tablePredicateClass('id', $ctrl.local.namespaceOptions.predicate, $ctrl.local.namespaceOptions.reverse)">
|
<td ng-class="$ctrl.TableService.tablePredicateClass('id', $ctrl.local.namespaceOptions.predicate, $ctrl.local.namespaceOptions.reverse)">
|
||||||
|
@ -39,7 +39,7 @@
|
||||||
|
|
||||||
<tr class="co-checkable-row"
|
<tr class="co-checkable-row"
|
||||||
ng-repeat="namespace in $ctrl.local.orderedNamespaces.visibleEntries | slice:($ctrl.namespacesPerPage * $ctrl.local.namespaceOptions.page):($ctrl.namespacesPerPage * ($ctrl.local.namespaceOptions.page + 1))"
|
ng-repeat="namespace in $ctrl.local.orderedNamespaces.visibleEntries | slice:($ctrl.namespacesPerPage * $ctrl.local.namespaceOptions.page):($ctrl.namespacesPerPage * ($ctrl.local.namespaceOptions.page + 1))"
|
||||||
ng-class="$ctrl.local.selectedNamespace == $ctrl.namespace ? 'checked' : ''"
|
ng-class="$ctrl.local.selectedNamespace == namespace ? 'checked' : ''"
|
||||||
bindonce>
|
bindonce>
|
||||||
<td>
|
<td>
|
||||||
<input type="radio"
|
<input type="radio"
|
||||||
|
@ -290,19 +290,22 @@
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Warning -->
|
<!-- Warning -->
|
||||||
<div class="col-lg-7 col-md-7 col-sm-12 main-col" ng-if="$ctrl.local.triggerAnalysis.status == 'warning'">
|
<div class="col-lg-7 col-md-7 col-sm-12 main-col"
|
||||||
|
ng-if="$ctrl.local.triggerAnalysis.status == 'warning'">
|
||||||
<h3 class="warning"><i class="fa fa-exclamation-triangle"></i> Verification Warning</h3>
|
<h3 class="warning"><i class="fa fa-exclamation-triangle"></i> Verification Warning</h3>
|
||||||
{{ $ctrl.local.triggerAnalysis.message }}
|
{{ $ctrl.local.triggerAnalysis.message }}
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Public base -->
|
<!-- Public base -->
|
||||||
<div class="col-lg-7 col-md-7 col-sm-12 main-col" ng-if="$ctrl.local.triggerAnalysis.status == 'publicbase'">
|
<div class="col-lg-7 col-md-7 col-sm-12 main-col"
|
||||||
|
ng-if="$ctrl.local.triggerAnalysis.status == 'publicbase'">
|
||||||
<h3 class="success"><i class="fa fa-check-circle"></i> Ready to go!</h3>
|
<h3 class="success"><i class="fa fa-check-circle"></i> Ready to go!</h3>
|
||||||
<strong>Click "Create Trigger" to complete setup of this build trigger</strong>
|
<strong>Click "Create Trigger" to complete setup of this build trigger</strong>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Requires robot and is not admin -->
|
<!-- Requires robot and is not admin -->
|
||||||
<div class="col-lg-7 col-md-7 col-sm-12 main-col" ng-if="$ctrl.local.triggerAnalysis.status == 'requiresrobot' && !$ctrl.local.triggerAnalysis.is_admin">
|
<div class="col-lg-7 col-md-7 col-sm-12 main-col"
|
||||||
|
ng-if="$ctrl.local.triggerAnalysis.status == 'requiresrobot' && !$ctrl.local.triggerAnalysis.is_admin">
|
||||||
<h3>Robot Account Required</h3>
|
<h3>Robot Account Required</h3>
|
||||||
<p>The selected Dockerfile in the selected repository depends upon a private base image</p>
|
<p>The selected Dockerfile in the selected repository depends upon a private base image</p>
|
||||||
<p>A robot account with access to the base image is required to setup this trigger, but you are not the administrator of this namespace.</p>
|
<p>A robot account with access to the base image is required to setup this trigger, but you are not the administrator of this namespace.</p>
|
||||||
|
@ -310,7 +313,8 @@
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<!-- Requires robot and is admin -->
|
<!-- Requires robot and is admin -->
|
||||||
<div class="col-lg-7 col-md-7 col-sm-12 main-col" ng-if="$ctrl.local.triggerAnalysis.status == 'requiresrobot' && $ctrl.local.triggerAnalysis.is_admin">
|
<div class="col-lg-7 col-md-7 col-sm-12 main-col"
|
||||||
|
ng-if="$ctrl.local.triggerAnalysis.status == 'requiresrobot' && $ctrl.local.triggerAnalysis.is_admin">
|
||||||
<h3>Select Robot Account</h3>
|
<h3>Select Robot Account</h3>
|
||||||
<strong>
|
<strong>
|
||||||
The selected Dockerfile in the selected repository depends upon a private base image. Select a robot account with access:
|
The selected Dockerfile in the selected repository depends upon a private base image. Select a robot account with access:
|
||||||
|
@ -338,7 +342,7 @@
|
||||||
</thead>
|
</thead>
|
||||||
|
|
||||||
<tr class="co-checkable-row"
|
<tr class="co-checkable-row"
|
||||||
ng-repeat="robot in $ctrl.local.orderedRobotAccounts.visibleEntries | slice:($ctrl.robotsPerPage * $ctrl.local.namespaceOptions.page):(robotsPerPage * ($ctrl.local.robotOptions.page + 1))"
|
ng-repeat="robot in $ctrl.local.orderedRobotAccounts.visibleEntries | slice:($ctrl.robotsPerPage * $ctrl.local.namespaceOptions.page):($ctrl.robotsPerPage * ($ctrl.local.robotOptions.page + 1))"
|
||||||
ng-class="$ctrl.local.robotAccount == robot ? 'checked' : ''"
|
ng-class="$ctrl.local.robotAccount == robot ? 'checked' : ''"
|
||||||
bindonce>
|
bindonce>
|
||||||
<td>
|
<td>
|
||||||
|
@ -355,8 +359,8 @@
|
||||||
</td>
|
</td>
|
||||||
</tr>
|
</tr>
|
||||||
</table>
|
</table>
|
||||||
<div class="empty" ng-if="$ctrl.local.triggerAnalysis.robots.length && !$ctrl.local.orderedRobotAccounts.entries.length"
|
<div class="empty" style="margin-top: 20px;"
|
||||||
style="margin-top: 20px;">
|
ng-if="$ctrl.local.triggerAnalysis.robots.length && !$ctrl.local.orderedRobotAccounts.entries.length">
|
||||||
<div class="empty-primary-msg">No matching robot accounts found.</div>
|
<div class="empty-primary-msg">No matching robot accounts found.</div>
|
||||||
<div class="empty-secondary-msg">Try expanding your filtering terms.</div>
|
<div class="empty-secondary-msg">Try expanding your filtering terms.</div>
|
||||||
</div>
|
</div>
|
||||||
|
|
|
@ -62,14 +62,8 @@ export class ManageTriggerGithostComponent implements ng.IComponentController {
|
||||||
|
|
||||||
public $onInit(): void {
|
public $onInit(): void {
|
||||||
// TODO: Replace $scope.$watch with @Output methods for child component mutations or $onChanges for parent mutations
|
// TODO: Replace $scope.$watch with @Output methods for child component mutations or $onChanges for parent mutations
|
||||||
this.$scope.$watch(() => this.trigger, (trigger) => {
|
this.$scope.$watch(() => this.trigger, this.initialSetup.bind(this));
|
||||||
if (trigger && this.repository) {
|
this.$scope.$watch(() => this.repository, this.initialSetup.bind(this));
|
||||||
this.config = trigger['config'] || {};
|
|
||||||
this.namespaceTitle = 'organization';
|
|
||||||
this.local.selectedNamespace = null;
|
|
||||||
this.loadNamespaces();
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
this.$scope.$watch(() => this.local.selectedNamespace, (namespace) => {
|
this.$scope.$watch(() => this.local.selectedNamespace, (namespace) => {
|
||||||
if (namespace) {
|
if (namespace) {
|
||||||
|
@ -102,6 +96,20 @@ export class ManageTriggerGithostComponent implements ng.IComponentController {
|
||||||
this.$scope.$watch(() => this.local.robotOptions.filter, this.buildOrderedRobotAccounts);
|
this.$scope.$watch(() => this.local.robotOptions.filter, this.buildOrderedRobotAccounts);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private initialSetup(): void {
|
||||||
|
if (!this.repository || !this.trigger) { return; }
|
||||||
|
|
||||||
|
if (this.namespaceTitle) {
|
||||||
|
// Already setup.
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
this.config = this.trigger['config'] || {};
|
||||||
|
this.namespaceTitle = 'organization';
|
||||||
|
this.local.selectedNamespace = null;
|
||||||
|
this.loadNamespaces();
|
||||||
|
}
|
||||||
|
|
||||||
public getTriggerIcon(): any {
|
public getTriggerIcon(): any {
|
||||||
return this.TriggerService.getIcon(this.trigger.service);
|
return this.TriggerService.getIcon(this.trigger.service);
|
||||||
}
|
}
|
||||||
|
|
|
@ -57,6 +57,8 @@ angular.module('quay').directive('manualTriggerBuildDialog', function () {
|
||||||
$scope.fieldOptions[parameter['name']] = resp['values'];
|
$scope.fieldOptions[parameter['name']] = resp['values'];
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
delete $scope.parameters[parameter['name']];
|
||||||
}
|
}
|
||||||
|
|
||||||
$scope.runParameters = parameters;
|
$scope.runParameters = parameters;
|
||||||
|
|
|
@ -49,7 +49,7 @@ export class quay {
|
||||||
// TODO: Make injected values into services and move to NgModule.providers, as constants are not supported in Angular 2
|
// TODO: Make injected values into services and move to NgModule.providers, as constants are not supported in Angular 2
|
||||||
angular
|
angular
|
||||||
.module(quay.name)
|
.module(quay.name)
|
||||||
.factory("FileReaderFactory", () => () => new FileReader())
|
.factory("fileReaderFactory", () => () => new FileReader())
|
||||||
.constant('NAME_PATTERNS', NAME_PATTERNS)
|
.constant('NAME_PATTERNS', NAME_PATTERNS)
|
||||||
.constant('INJECTED_CONFIG', INJECTED_CONFIG)
|
.constant('INJECTED_CONFIG', INJECTED_CONFIG)
|
||||||
.constant('INJECTED_FEATURES', INJECTED_FEATURES)
|
.constant('INJECTED_FEATURES', INJECTED_FEATURES)
|
||||||
|
|
|
@ -76,14 +76,15 @@ angular.module('quay').factory('DataFileService', [function() {
|
||||||
return parts.join('/');
|
return parts.join('/');
|
||||||
};
|
};
|
||||||
|
|
||||||
var handler = new Untar(new Uint8Array(buf));
|
try {
|
||||||
handler.process(function(status, read, files, err) {
|
var handler = new Untar(new Uint8Array(buf));
|
||||||
switch (status) {
|
handler.process(function(status, read, files, err) {
|
||||||
case 'error':
|
switch (status) {
|
||||||
failure(err);
|
case 'error':
|
||||||
break;
|
failure(err);
|
||||||
|
break;
|
||||||
|
|
||||||
case 'done':
|
case 'done':
|
||||||
var processed = [];
|
var processed = [];
|
||||||
for (var i = 0; i < files.length; ++i) {
|
for (var i = 0; i < files.length; ++i) {
|
||||||
var currentFile = files[i];
|
var currentFile = files[i];
|
||||||
|
@ -104,8 +105,12 @@ angular.module('quay').factory('DataFileService', [function() {
|
||||||
}
|
}
|
||||||
success(processed);
|
success(processed);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
} catch (e) {
|
||||||
|
failure();
|
||||||
|
}
|
||||||
|
|
||||||
};
|
};
|
||||||
|
|
||||||
dataFileService.blobToString = function(blob, callback) {
|
dataFileService.blobToString = function(blob, callback) {
|
||||||
|
|
|
@ -1,127 +0,0 @@
|
||||||
/**
|
|
||||||
* Service which provides helper methods for extracting information out from a Dockerfile
|
|
||||||
* or an archive containing a Dockerfile.
|
|
||||||
*/
|
|
||||||
angular.module('quay').factory('DockerfileServiceOld', ['DataFileService', 'Config', function(DataFileService, Config) {
|
|
||||||
var dockerfileService = {};
|
|
||||||
|
|
||||||
function DockerfileInfo(contents) {
|
|
||||||
this.contents = contents;
|
|
||||||
}
|
|
||||||
|
|
||||||
DockerfileInfo.prototype.getRegistryBaseImage = function() {
|
|
||||||
var baseImage = this.getBaseImage();
|
|
||||||
if (!baseImage) {
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (baseImage.indexOf(Config.getDomain() + '/') != 0) {
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|
||||||
return baseImage.substring(Config.getDomain().length + 1);
|
|
||||||
};
|
|
||||||
|
|
||||||
DockerfileInfo.prototype.getBaseImage = function() {
|
|
||||||
var imageAndTag = this.getBaseImageAndTag();
|
|
||||||
if (!imageAndTag) {
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Note, we have to handle a few different cases here:
|
|
||||||
// 1) someimage
|
|
||||||
// 2) someimage:tag
|
|
||||||
// 3) host:port/someimage
|
|
||||||
// 4) host:port/someimage:tag
|
|
||||||
var lastIndex = imageAndTag.lastIndexOf(':');
|
|
||||||
if (lastIndex < 0) {
|
|
||||||
return imageAndTag;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Otherwise, check if there is a / in the portion after the split point. If so,
|
|
||||||
// then the latter is part of the path (and not a tag).
|
|
||||||
var afterColon = imageAndTag.substring(lastIndex + 1);
|
|
||||||
if (afterColon.indexOf('/') >= 0) {
|
|
||||||
return imageAndTag;
|
|
||||||
}
|
|
||||||
|
|
||||||
return imageAndTag.substring(0, lastIndex);
|
|
||||||
};
|
|
||||||
|
|
||||||
DockerfileInfo.prototype.getBaseImageAndTag = function() {
|
|
||||||
var fromIndex = this.contents.indexOf('FROM ');
|
|
||||||
if (fromIndex < 0) {
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|
||||||
var newline = this.contents.indexOf('\n', fromIndex);
|
|
||||||
if (newline < 0) {
|
|
||||||
newline = this.contents.length;
|
|
||||||
}
|
|
||||||
|
|
||||||
return $.trim(this.contents.substring(fromIndex + 'FROM '.length, newline));
|
|
||||||
};
|
|
||||||
|
|
||||||
DockerfileInfo.forData = function(contents) {
|
|
||||||
if (contents.indexOf('FROM ') < 0) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
return new DockerfileInfo(contents);
|
|
||||||
};
|
|
||||||
|
|
||||||
var processFiles = function(files, dataArray, success, failure) {
|
|
||||||
// The files array will be empty if the submitted file was not an archive. We therefore
|
|
||||||
// treat it as a single Dockerfile.
|
|
||||||
if (files.length == 0) {
|
|
||||||
DataFileService.arrayToString(dataArray, function(c) {
|
|
||||||
var result = DockerfileInfo.forData(c);
|
|
||||||
if (!result) {
|
|
||||||
failure('File chosen is not a valid Dockerfile');
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
success(result);
|
|
||||||
});
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
var found = false;
|
|
||||||
files.forEach(function(file) {
|
|
||||||
if (file['name'] == 'Dockerfile') {
|
|
||||||
DataFileService.blobToString(file.toBlob(), function(c) {
|
|
||||||
var result = DockerfileInfo.forData(c);
|
|
||||||
if (!result) {
|
|
||||||
failure('Dockerfile inside archive is not a valid Dockerfile');
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
success(result);
|
|
||||||
});
|
|
||||||
found = true;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
if (!found) {
|
|
||||||
failure('No Dockerfile found in root of archive');
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
dockerfileService.getDockerfile = function(file, success, failure) {
|
|
||||||
var reader = new FileReader();
|
|
||||||
reader.onload = function(e) {
|
|
||||||
var dataArray = reader.result;
|
|
||||||
DataFileService.readDataArrayAsPossibleArchive(dataArray, function(files) {
|
|
||||||
processFiles(files, dataArray, success, failure);
|
|
||||||
}, function() {
|
|
||||||
// Not an archive. Read directly as a single file.
|
|
||||||
processFiles([], dataArray, success, failure);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
|
|
||||||
reader.onerror = failure;
|
|
||||||
reader.readAsArrayBuffer(file);
|
|
||||||
};
|
|
||||||
|
|
||||||
return dockerfileService;
|
|
||||||
}]);
|
|
|
@ -52,157 +52,7 @@ describe("DockerfileServiceImpl", () => {
|
||||||
});
|
});
|
||||||
|
|
||||||
it("calls datafile service to read given file as possible archive file", (done) => {
|
it("calls datafile service to read given file as possible archive file", (done) => {
|
||||||
dockerfileServiceImpl.getDockerfile(file,
|
dockerfileServiceImpl.getDockerfile(file)
|
||||||
(dockerfile: DockerfileInfoImpl) => {
|
|
||||||
expect(readAsFileBufferSpy.calls.argsFor(0)[0]).toEqual(file);
|
|
||||||
expect(dataFileServiceMock.readDataArrayAsPossibleArchive).toHaveBeenCalled();
|
|
||||||
done();
|
|
||||||
},
|
|
||||||
(error: Event | string) => {
|
|
||||||
fail("Should not invoke failure callback");
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it("calls datafile service to convert file to string if given file is not an archive", (done) => {
|
|
||||||
dockerfileServiceImpl.getDockerfile(file,
|
|
||||||
(dockerfile: DockerfileInfoImpl) => {
|
|
||||||
expect(dataFileServiceMock.arrayToString.calls.argsFor(0)[0]).toEqual(file);
|
|
||||||
done();
|
|
||||||
},
|
|
||||||
(error: Event | string) => {
|
|
||||||
fail("Should not invoke success callback");
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it("calls failure callback if given non-archive file that is not a valid Dockerfile", (done) => {
|
|
||||||
forDataSpy.and.returnValue(null);
|
|
||||||
dockerfileServiceImpl.getDockerfile(file,
|
|
||||||
(dockerfile: DockerfileInfoImpl) => {
|
|
||||||
fail("Should not invoke success callback");
|
|
||||||
done();
|
|
||||||
},
|
|
||||||
(error: Event | string) => {
|
|
||||||
expect(error).toEqual('File chosen is not a valid Dockerfile');
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it("calls success callback with new DockerfileInfoImpl instance if given valid Dockerfile", (done) => {
|
|
||||||
dockerfileServiceImpl.getDockerfile(file,
|
|
||||||
(dockerfile: DockerfileInfoImpl) => {
|
|
||||||
expect(dockerfile).toBeDefined();
|
|
||||||
done();
|
|
||||||
},
|
|
||||||
(error: Event | string) => {
|
|
||||||
fail('Should not invoke failure callback');
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it("calls failure callback if given archive file with no Dockerfile present in root directory", (done) => {
|
|
||||||
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
|
||||||
success(invalidArchiveFile);
|
|
||||||
});
|
|
||||||
|
|
||||||
dockerfileServiceImpl.getDockerfile(file,
|
|
||||||
(dockerfile: DockerfileInfoImpl) => {
|
|
||||||
fail("Should not invoke success callback");
|
|
||||||
done();
|
|
||||||
},
|
|
||||||
(error: Event | string) => {
|
|
||||||
expect(error).toEqual('No Dockerfile found in root of archive');
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it("calls datafile service to convert blob to string if given file is an archive", (done) => {
|
|
||||||
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
|
||||||
success(validArchiveFile);
|
|
||||||
});
|
|
||||||
|
|
||||||
dockerfileServiceImpl.getDockerfile(file,
|
|
||||||
(dockerfile: DockerfileInfoImpl) => {
|
|
||||||
expect(validArchiveFile[0].toBlob).toHaveBeenCalled();
|
|
||||||
expect(dataFileServiceMock.blobToString.calls.argsFor(0)[0]).toEqual(validArchiveFile[0].toBlob());
|
|
||||||
done();
|
|
||||||
},
|
|
||||||
(error: Event | string) => {
|
|
||||||
fail("Should not invoke success callback");
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it("calls failure callback if given archive file with invalid Dockerfile", (done) => {
|
|
||||||
forDataSpy.and.returnValue(null);
|
|
||||||
invalidArchiveFile[0].name = 'Dockerfile';
|
|
||||||
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
|
||||||
success(invalidArchiveFile);
|
|
||||||
});
|
|
||||||
|
|
||||||
dockerfileServiceImpl.getDockerfile(file,
|
|
||||||
(dockerfile: DockerfileInfoImpl) => {
|
|
||||||
fail("Should not invoke success callback");
|
|
||||||
done();
|
|
||||||
},
|
|
||||||
(error: Event | string) => {
|
|
||||||
expect(error).toEqual('Dockerfile inside archive is not a valid Dockerfile');
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it("calls success callback with new DockerfileInfoImpl instance if given archive with valid Dockerfile", (done) => {
|
|
||||||
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
|
||||||
success(validArchiveFile);
|
|
||||||
});
|
|
||||||
|
|
||||||
dockerfileServiceImpl.getDockerfile(file,
|
|
||||||
(dockerfile: DockerfileInfoImpl) => {
|
|
||||||
expect(dockerfile).toBeDefined();
|
|
||||||
done();
|
|
||||||
},
|
|
||||||
(error: Event | string) => {
|
|
||||||
fail('Should not invoke failure callback');
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe("extractDockerfile", () => {
|
|
||||||
var file: any;
|
|
||||||
var invalidArchiveFile: any[];
|
|
||||||
var validArchiveFile: any[];
|
|
||||||
var readAsFileBufferSpy: Spy;
|
|
||||||
var forDataSpy: Spy;
|
|
||||||
|
|
||||||
beforeEach(() => {
|
|
||||||
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
|
||||||
failure([]);
|
|
||||||
});
|
|
||||||
|
|
||||||
dataFileServiceMock.arrayToString.and.callFake((buf, callback) => {
|
|
||||||
var contents: string = "";
|
|
||||||
callback(contents);
|
|
||||||
});
|
|
||||||
|
|
||||||
dataFileServiceMock.blobToString.and.callFake((blob, callback) => {
|
|
||||||
callback(blob.toString());
|
|
||||||
});
|
|
||||||
|
|
||||||
forDataSpy = spyOn(DockerfileInfoImpl, "forData").and.returnValue(new DockerfileInfoImpl(file, configMock));
|
|
||||||
readAsFileBufferSpy = spyOn(fileReaderMock, "readAsArrayBuffer").and.callFake(() => {
|
|
||||||
var event: any = {target: {result: file}};
|
|
||||||
fileReaderMock.onload(event);
|
|
||||||
});
|
|
||||||
|
|
||||||
file = "FROM quay.io/coreos/nginx:latest";
|
|
||||||
validArchiveFile = [{name: 'Dockerfile', toBlob: jasmine.createSpy('toBlobSpy').and.returnValue(file)}];
|
|
||||||
invalidArchiveFile = [{name: 'main.exe', toBlob: jasmine.createSpy('toBlobSpy').and.returnValue("")}];
|
|
||||||
});
|
|
||||||
|
|
||||||
it("calls datafile service to read given file as possible archive file", (done) => {
|
|
||||||
dockerfileServiceImpl.extractDockerfile(file)
|
|
||||||
.then((dockerfile: DockerfileInfoImpl) => {
|
.then((dockerfile: DockerfileInfoImpl) => {
|
||||||
expect(readAsFileBufferSpy.calls.argsFor(0)[0]).toEqual(file);
|
expect(readAsFileBufferSpy.calls.argsFor(0)[0]).toEqual(file);
|
||||||
expect(dataFileServiceMock.readDataArrayAsPossibleArchive).toHaveBeenCalled();
|
expect(dataFileServiceMock.readDataArrayAsPossibleArchive).toHaveBeenCalled();
|
||||||
|
@ -215,31 +65,107 @@ describe("DockerfileServiceImpl", () => {
|
||||||
});
|
});
|
||||||
|
|
||||||
it("calls datafile service to convert file to string if given file is not an archive", (done) => {
|
it("calls datafile service to convert file to string if given file is not an archive", (done) => {
|
||||||
done();
|
dockerfileServiceImpl.getDockerfile(file)
|
||||||
|
.then((dockerfile: DockerfileInfoImpl) => {
|
||||||
|
expect(dataFileServiceMock.arrayToString.calls.argsFor(0)[0]).toEqual(file);
|
||||||
|
done();
|
||||||
|
})
|
||||||
|
.catch((error: string) => {
|
||||||
|
fail('Promise should be resolved');
|
||||||
|
done();
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it("returns rejected promise if given non-archive file that is not a valid Dockerfile", (done) => {
|
it("returns rejected promise if given non-archive file that is not a valid Dockerfile", (done) => {
|
||||||
done();
|
forDataSpy.and.returnValue(null);
|
||||||
|
dockerfileServiceImpl.getDockerfile(file)
|
||||||
|
.then((dockerfile: DockerfileInfoImpl) => {
|
||||||
|
fail("Promise should be rejected");
|
||||||
|
done();
|
||||||
|
})
|
||||||
|
.catch((error: string) => {
|
||||||
|
expect(error).toEqual('File chosen is not a valid Dockerfile');
|
||||||
|
done();
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it("returns resolved promise with new DockerfileInfoImpl instance if given valid Dockerfile", (done) => {
|
it("returns resolved promise with new DockerfileInfoImpl instance if given valid Dockerfile", (done) => {
|
||||||
done();
|
dockerfileServiceImpl.getDockerfile(file)
|
||||||
|
.then((dockerfile: DockerfileInfoImpl) => {
|
||||||
|
expect(dockerfile).toBeDefined();
|
||||||
|
done();
|
||||||
|
})
|
||||||
|
.catch((error: string) => {
|
||||||
|
fail('Promise should be resolved');
|
||||||
|
done();
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it("returns rejected promise if given archive file with no Dockerfile present in root directory", (done) => {
|
it("returns rejected promise if given archive file with no Dockerfile present in root directory", (done) => {
|
||||||
done();
|
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
||||||
|
success(invalidArchiveFile);
|
||||||
|
});
|
||||||
|
|
||||||
|
dockerfileServiceImpl.getDockerfile(file)
|
||||||
|
.then((dockerfile: DockerfileInfoImpl) => {
|
||||||
|
fail('Promise should be rejected');
|
||||||
|
done();
|
||||||
|
})
|
||||||
|
.catch((error: string) => {
|
||||||
|
expect(error).toEqual('No Dockerfile found in root of archive');
|
||||||
|
done();
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it("calls datafile service to convert blob to string if given file is an archive", (done) => {
|
it("calls datafile service to convert blob to string if given file is an archive", (done) => {
|
||||||
done();
|
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
||||||
|
success(validArchiveFile);
|
||||||
|
});
|
||||||
|
|
||||||
|
dockerfileServiceImpl.getDockerfile(file)
|
||||||
|
.then((dockerfile: DockerfileInfoImpl) => {
|
||||||
|
expect(validArchiveFile[0].toBlob).toHaveBeenCalled();
|
||||||
|
expect(dataFileServiceMock.blobToString.calls.argsFor(0)[0]).toEqual(validArchiveFile[0].toBlob());
|
||||||
|
done();
|
||||||
|
})
|
||||||
|
.catch((error: string) => {
|
||||||
|
fail('Promise should be resolved');
|
||||||
|
done();
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it("returns rejected promise if given archive file with invalid Dockerfile", (done) => {
|
it("returns rejected promise if given archive file with invalid Dockerfile", (done) => {
|
||||||
done();
|
forDataSpy.and.returnValue(null);
|
||||||
|
invalidArchiveFile[0].name = 'Dockerfile';
|
||||||
|
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
||||||
|
success(invalidArchiveFile);
|
||||||
|
});
|
||||||
|
|
||||||
|
dockerfileServiceImpl.getDockerfile(file)
|
||||||
|
.then((dockerfile: DockerfileInfoImpl) => {
|
||||||
|
fail('Promise should be rejected');
|
||||||
|
done();
|
||||||
|
})
|
||||||
|
.catch((error: string) => {
|
||||||
|
expect(error).toEqual('Dockerfile inside archive is not a valid Dockerfile');
|
||||||
|
done();
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it("returns resolved promise of new DockerfileInfoImpl instance if given archive with valid Dockerfile", (done) => {
|
it("returns resolved promise of new DockerfileInfoImpl instance if given archive with valid Dockerfile", (done) => {
|
||||||
done();
|
dataFileServiceMock.readDataArrayAsPossibleArchive.and.callFake((buf, success, failure) => {
|
||||||
|
success(validArchiveFile);
|
||||||
|
});
|
||||||
|
|
||||||
|
dockerfileServiceImpl.getDockerfile(file)
|
||||||
|
.then((dockerfile: DockerfileInfoImpl) => {
|
||||||
|
expect(dockerfile).toBeDefined();
|
||||||
|
done();
|
||||||
|
})
|
||||||
|
.catch((error: string) => {
|
||||||
|
fail('Promise should be resolved');
|
||||||
|
done();
|
||||||
|
});
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
|
@ -5,21 +5,30 @@ import { Injectable } from 'angular-ts-decorators';
|
||||||
@Injectable(DockerfileService.name)
|
@Injectable(DockerfileService.name)
|
||||||
export class DockerfileServiceImpl implements DockerfileService {
|
export class DockerfileServiceImpl implements DockerfileService {
|
||||||
|
|
||||||
constructor(private DataFileService: any, private Config: any, private FileReaderFactory: () => FileReader) {
|
constructor(private DataFileService: any,
|
||||||
|
private Config: any,
|
||||||
|
private fileReaderFactory: () => FileReader) {
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
public extractDockerfile(file: any): Promise<DockerfileInfoImpl | string> {
|
public getDockerfile(file: any): Promise<DockerfileInfoImpl | string> {
|
||||||
return new Promise((resolve, reject) => {
|
return new Promise((resolve, reject) => {
|
||||||
var reader: FileReader = this.FileReaderFactory();
|
var reader: FileReader = this.fileReaderFactory();
|
||||||
reader.onload = (event: any) => {
|
reader.onload = (event: any) => {
|
||||||
|
// FIXME: Debugging
|
||||||
|
console.log(event.target.result);
|
||||||
|
|
||||||
this.DataFileService.readDataArrayAsPossibleArchive(event.target.result,
|
this.DataFileService.readDataArrayAsPossibleArchive(event.target.result,
|
||||||
(files: any[]) => {
|
(files: any[]) => {
|
||||||
this.processFiles1(files);
|
this.processFiles(files)
|
||||||
|
.then((dockerfileInfo: DockerfileInfoImpl) => resolve(dockerfileInfo))
|
||||||
|
.catch((error: string) => reject(error));
|
||||||
},
|
},
|
||||||
() => {
|
() => {
|
||||||
// Not an archive. Read directly as a single file.
|
// Not an archive. Read directly as a single file.
|
||||||
this.processFile1(event.target.result);
|
this.processFile(event.target.result)
|
||||||
|
.then((dockerfileInfo: DockerfileInfoImpl) => resolve(dockerfileInfo))
|
||||||
|
.catch((error: string) => reject(error));
|
||||||
});
|
});
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -28,40 +37,7 @@ export class DockerfileServiceImpl implements DockerfileService {
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
public getDockerfile(file: any,
|
private processFile(dataArray: any): Promise<DockerfileInfoImpl | string> {
|
||||||
success: (dockerfile: DockerfileInfoImpl) => void,
|
|
||||||
failure: (error: Event | string) => void): void {
|
|
||||||
var reader: FileReader = this.FileReaderFactory();
|
|
||||||
reader.onload = (event: any) => {
|
|
||||||
this.DataFileService.readDataArrayAsPossibleArchive(event.target.result,
|
|
||||||
(files: any[]) => {
|
|
||||||
this.processFiles(files, success, failure);
|
|
||||||
},
|
|
||||||
() => {
|
|
||||||
// Not an archive. Read directly as a single file.
|
|
||||||
this.processFile(event.target.result, success, failure);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
|
|
||||||
reader.onerror = failure;
|
|
||||||
reader.readAsArrayBuffer(file);
|
|
||||||
}
|
|
||||||
|
|
||||||
private processFile(dataArray: any,
|
|
||||||
success: (dockerfile: DockerfileInfoImpl) => void,
|
|
||||||
failure: (error: ErrorEvent | string) => void): void {
|
|
||||||
this.DataFileService.arrayToString(dataArray, (contents: string) => {
|
|
||||||
var result: DockerfileInfoImpl | null = DockerfileInfoImpl.forData(contents, this.Config);
|
|
||||||
if (result == null) {
|
|
||||||
failure('File chosen is not a valid Dockerfile');
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
success(result);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
private processFile1(dataArray: any): Promise<DockerfileInfoImpl | string> {
|
|
||||||
return new Promise((resolve, reject) => {
|
return new Promise((resolve, reject) => {
|
||||||
this.DataFileService.arrayToString(dataArray, (contents: string) => {
|
this.DataFileService.arrayToString(dataArray, (contents: string) => {
|
||||||
var result: DockerfileInfoImpl | null = DockerfileInfoImpl.forData(contents, this.Config);
|
var result: DockerfileInfoImpl | null = DockerfileInfoImpl.forData(contents, this.Config);
|
||||||
|
@ -75,31 +51,7 @@ export class DockerfileServiceImpl implements DockerfileService {
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
private processFiles(files: any[],
|
private processFiles(files: any[]): Promise<DockerfileInfoImpl | string> {
|
||||||
success: (dockerfile: DockerfileInfoImpl) => void,
|
|
||||||
failure: (error: ErrorEvent | string) => void): void {
|
|
||||||
var found: boolean = false;
|
|
||||||
files.forEach((file) => {
|
|
||||||
if (file['name'] == 'Dockerfile') {
|
|
||||||
this.DataFileService.blobToString(file.toBlob(), (contents: string) => {
|
|
||||||
var result: DockerfileInfoImpl | null = DockerfileInfoImpl.forData(contents, this.Config);
|
|
||||||
if (result == null) {
|
|
||||||
failure('Dockerfile inside archive is not a valid Dockerfile');
|
|
||||||
}
|
|
||||||
else {
|
|
||||||
success(result);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
found = true;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
if (!found) {
|
|
||||||
failure('No Dockerfile found in root of archive');
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private processFiles1(files: any[]): Promise<DockerfileInfoImpl | string> {
|
|
||||||
return new Promise((resolve, reject) => {
|
return new Promise((resolve, reject) => {
|
||||||
var found: boolean = false;
|
var found: boolean = false;
|
||||||
files.forEach((file) => {
|
files.forEach((file) => {
|
||||||
|
|
|
@ -7,19 +7,9 @@ export abstract class DockerfileService {
|
||||||
/**
|
/**
|
||||||
* Retrieve Dockerfile from given file.
|
* Retrieve Dockerfile from given file.
|
||||||
* @param file Dockerfile or archive file containing Dockerfile.
|
* @param file Dockerfile or archive file containing Dockerfile.
|
||||||
* @param success Success callback with retrieved Dockerfile as parameter.
|
* @return promise Promise which resolves to new DockerfileInfo instance or rejects with error message.
|
||||||
* @param failure Failure callback with failure message as parameter.
|
|
||||||
*/
|
*/
|
||||||
public abstract getDockerfile(file: any,
|
public abstract getDockerfile(file: any): Promise<DockerfileInfo | string>;
|
||||||
success: (dockerfile: DockerfileInfo) => void,
|
|
||||||
failure: (error: ErrorEvent | string) => void): void;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Retrieve Dockerfile from given file.
|
|
||||||
* @param file Dockerfile or archive file containing Dockerfile.
|
|
||||||
* @return promise Promise resolving to new DockerfileInfo instance or rejecting to error message.
|
|
||||||
*/
|
|
||||||
public abstract extractDockerfile(file: any): Promise<DockerfileInfo | string>;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -208,10 +208,6 @@ angular.module('quay').factory('TriggerService', ['UtilService', '$sanitize', 'K
|
||||||
return '//Dockerfile';
|
return '//Dockerfile';
|
||||||
}
|
}
|
||||||
|
|
||||||
if (subdirectory[subdirectory.length - 1] != '/') {
|
|
||||||
subdirectory = subdirectory + '/';
|
|
||||||
}
|
|
||||||
|
|
||||||
return '//' + subdirectory.replace(new RegExp('(^\/+|\/+$)'), '') + 'Dockerfile';
|
return '//' + subdirectory.replace(new RegExp('(^\/+|\/+$)'), '') + 'Dockerfile';
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -4059,11 +4059,18 @@ class TestBuildTriggers(ApiTestCase):
|
||||||
self.assertEquals('bar', py_json.loads(build_obj.job_config)['trigger_metadata']['foo'])
|
self.assertEquals('bar', py_json.loads(build_obj.job_config)['trigger_metadata']['foo'])
|
||||||
|
|
||||||
# Start another manual build, with a ref.
|
# Start another manual build, with a ref.
|
||||||
start_json = self.postJsonResponse(ActivateBuildTrigger,
|
self.postJsonResponse(ActivateBuildTrigger,
|
||||||
params=dict(repository=ADMIN_ACCESS_USER + '/simple',
|
params=dict(repository=ADMIN_ACCESS_USER + '/simple',
|
||||||
trigger_uuid=trigger.uuid),
|
trigger_uuid=trigger.uuid),
|
||||||
data=dict(refs={'kind': 'branch', 'name': 'foobar'}),
|
data=dict(refs={'kind': 'branch', 'name': 'foobar'}),
|
||||||
expected_code=201)
|
expected_code=201)
|
||||||
|
|
||||||
|
# Start another manual build with a null ref.
|
||||||
|
self.postJsonResponse(ActivateBuildTrigger,
|
||||||
|
params=dict(repository=ADMIN_ACCESS_USER + '/simple',
|
||||||
|
trigger_uuid=trigger.uuid),
|
||||||
|
data=dict(refs=None),
|
||||||
|
expected_code=201)
|
||||||
|
|
||||||
def test_invalid_robot_account(self):
|
def test_invalid_robot_account(self):
|
||||||
self.login(ADMIN_ACCESS_USER)
|
self.login(ADMIN_ACCESS_USER)
|
||||||
|
|
|
@ -6,8 +6,8 @@ import features
|
||||||
from app import app, secscan_api, prometheus
|
from app import app, secscan_api, prometheus
|
||||||
from workers.worker import Worker
|
from workers.worker import Worker
|
||||||
from data.database import UseThenDisconnect
|
from data.database import UseThenDisconnect
|
||||||
from data.model.image import (get_images_eligible_for_scan, get_max_id_for_sec_scan,
|
from data.model.tag import (get_tags_images_eligible_for_scan, get_tag_pk_field,
|
||||||
get_min_id_for_sec_scan, get_image_id)
|
get_max_id_for_sec_scan, get_min_id_for_sec_scan)
|
||||||
from util.secscan.api import SecurityConfigValidator
|
from util.secscan.api import SecurityConfigValidator
|
||||||
from util.secscan.analyzer import LayerAnalyzer, PreemptedException
|
from util.secscan.analyzer import LayerAnalyzer, PreemptedException
|
||||||
from util.migrate.allocator import yield_random_entries
|
from util.migrate.allocator import yield_random_entries
|
||||||
|
@ -43,7 +43,7 @@ class SecurityWorker(Worker):
|
||||||
|
|
||||||
def _index_images(self):
|
def _index_images(self):
|
||||||
def batch_query():
|
def batch_query():
|
||||||
return get_images_eligible_for_scan(self._target_version)
|
return get_tags_images_eligible_for_scan(self._target_version)
|
||||||
|
|
||||||
# Get the ID of the last image we can analyze. Will be None if there are no images in the
|
# Get the ID of the last image we can analyze. Will be None if there are no images in the
|
||||||
# database.
|
# database.
|
||||||
|
@ -56,14 +56,14 @@ class SecurityWorker(Worker):
|
||||||
with UseThenDisconnect(app.config):
|
with UseThenDisconnect(app.config):
|
||||||
to_scan_generator = yield_random_entries(
|
to_scan_generator = yield_random_entries(
|
||||||
batch_query,
|
batch_query,
|
||||||
get_image_id(),
|
get_tag_pk_field(),
|
||||||
BATCH_SIZE,
|
BATCH_SIZE,
|
||||||
max_id,
|
max_id,
|
||||||
self._min_id,
|
self._min_id,
|
||||||
)
|
)
|
||||||
for candidate, abt, num_remaining in to_scan_generator:
|
for candidate, abt, num_remaining in to_scan_generator:
|
||||||
try:
|
try:
|
||||||
self._analyzer.analyze_recursively(candidate)
|
self._analyzer.analyze_recursively(candidate.image)
|
||||||
except PreemptedException:
|
except PreemptedException:
|
||||||
logger.info('Another worker pre-empted us for layer: %s', candidate.id)
|
logger.info('Another worker pre-empted us for layer: %s', candidate.id)
|
||||||
abt.set()
|
abt.set()
|
||||||
|
|
24
yarn.lock
24
yarn.lock
|
@ -20,9 +20,9 @@
|
||||||
dependencies:
|
dependencies:
|
||||||
"@types/angular" "*"
|
"@types/angular" "*"
|
||||||
|
|
||||||
"@types/angular@*", "@types/angular@1.5.16":
|
"@types/angular@*", "@types/angular@1.6.2":
|
||||||
version "1.5.16"
|
version "1.6.2"
|
||||||
resolved "https://registry.yarnpkg.com/@types/angular/-/angular-1.5.16.tgz#02a56754b50dbf9209266b4339031a54317702d9"
|
resolved "https://registry.yarnpkg.com/@types/angular/-/angular-1.6.2.tgz#a5c323ea5d4426ad18984cc8167fa091f7c8201b"
|
||||||
dependencies:
|
dependencies:
|
||||||
"@types/jquery" "*"
|
"@types/jquery" "*"
|
||||||
|
|
||||||
|
@ -36,7 +36,7 @@
|
||||||
dependencies:
|
dependencies:
|
||||||
typescript ">=2.1.4"
|
typescript ">=2.1.4"
|
||||||
|
|
||||||
"@types/jquery@*":
|
"@types/jquery@*", "@types/jquery@^2.0.40":
|
||||||
version "2.0.40"
|
version "2.0.40"
|
||||||
resolved "https://registry.yarnpkg.com/@types/jquery/-/jquery-2.0.40.tgz#acdd69e29b74cdec15dc3c074bcb064bc1b87213"
|
resolved "https://registry.yarnpkg.com/@types/jquery/-/jquery-2.0.40.tgz#acdd69e29b74cdec15dc3c074bcb064bc1b87213"
|
||||||
|
|
||||||
|
@ -109,23 +109,23 @@ amdefine@>=0.0.4:
|
||||||
version "1.0.1"
|
version "1.0.1"
|
||||||
resolved "https://registry.yarnpkg.com/amdefine/-/amdefine-1.0.1.tgz#4a5282ac164729e93619bcfd3ad151f817ce91f5"
|
resolved "https://registry.yarnpkg.com/amdefine/-/amdefine-1.0.1.tgz#4a5282ac164729e93619bcfd3ad151f817ce91f5"
|
||||||
|
|
||||||
angular-animate@^1.5.3:
|
angular-animate@1.6.2:
|
||||||
version "1.6.2"
|
version "1.6.2"
|
||||||
resolved "https://registry.yarnpkg.com/angular-animate/-/angular-animate-1.6.2.tgz#def2a8b9ede53b4b6e234c25f5c64e4b4385df15"
|
resolved "https://registry.yarnpkg.com/angular-animate/-/angular-animate-1.6.2.tgz#def2a8b9ede53b4b6e234c25f5c64e4b4385df15"
|
||||||
|
|
||||||
angular-cookies@^1.5.3:
|
angular-cookies@1.6.2:
|
||||||
version "1.6.2"
|
version "1.6.2"
|
||||||
resolved "https://registry.yarnpkg.com/angular-cookies/-/angular-cookies-1.6.2.tgz#ffb69f35f84d1efe71addac20a2905476b658884"
|
resolved "https://registry.yarnpkg.com/angular-cookies/-/angular-cookies-1.6.2.tgz#ffb69f35f84d1efe71addac20a2905476b658884"
|
||||||
|
|
||||||
angular-mocks@^1.5.3:
|
angular-mocks@1.6.2:
|
||||||
version "1.6.2"
|
version "1.6.2"
|
||||||
resolved "https://registry.yarnpkg.com/angular-mocks/-/angular-mocks-1.6.2.tgz#fbb28208e74d3512769afdb8771f5cc5a99f9128"
|
resolved "https://registry.yarnpkg.com/angular-mocks/-/angular-mocks-1.6.2.tgz#fbb28208e74d3512769afdb8771f5cc5a99f9128"
|
||||||
|
|
||||||
angular-route@^1.5.3:
|
angular-route@1.6.2:
|
||||||
version "1.6.2"
|
version "1.6.2"
|
||||||
resolved "https://registry.yarnpkg.com/angular-route/-/angular-route-1.6.2.tgz#95a349de2e73674f3dd783bb21e8d7b3fc526312"
|
resolved "https://registry.yarnpkg.com/angular-route/-/angular-route-1.6.2.tgz#95a349de2e73674f3dd783bb21e8d7b3fc526312"
|
||||||
|
|
||||||
angular-sanitize@^1.5.3:
|
angular-sanitize@1.6.2:
|
||||||
version "1.6.2"
|
version "1.6.2"
|
||||||
resolved "https://registry.yarnpkg.com/angular-sanitize/-/angular-sanitize-1.6.2.tgz#8a327c1acb2c14f50da5b5cad5ea452750a1a375"
|
resolved "https://registry.yarnpkg.com/angular-sanitize/-/angular-sanitize-1.6.2.tgz#8a327c1acb2c14f50da5b5cad5ea452750a1a375"
|
||||||
|
|
||||||
|
@ -135,9 +135,9 @@ angular-ts-decorators@0.0.19:
|
||||||
dependencies:
|
dependencies:
|
||||||
reflect-metadata "^0.1.8"
|
reflect-metadata "^0.1.8"
|
||||||
|
|
||||||
angular@1.5.3:
|
angular@1.6.2:
|
||||||
version "1.5.3"
|
version "1.6.2"
|
||||||
resolved "https://registry.yarnpkg.com/angular/-/angular-1.5.3.tgz#37c2f198ae76c2d6f3717a4ecef1cddcb048af79"
|
resolved "https://registry.yarnpkg.com/angular/-/angular-1.6.2.tgz#d0b677242ac4bf9ae81424297c6320973af4bb5a"
|
||||||
|
|
||||||
ansi-align@^1.1.0:
|
ansi-align@^1.1.0:
|
||||||
version "1.1.0"
|
version "1.1.0"
|
||||||
|
|
Reference in a new issue