Compare commits

...

30 Commits

Author SHA1 Message Date
Daniel Gyulai b0d361f007 repo image build + bump version 3 years ago
Daniel Gyulai 5cef27298c Remove useless import from pypirepo 3 years ago
Daniel Gyulai ebf8d74843 Fix #33 3 years ago
Daniel Gyulai 9bf90060fc Merge pull request '0.0.12' (#32) from 0.0.12 into master 3 years ago
Daniel Gyulai afb5ff4d97 bugfix 3 years ago
Daniel Gyulai b545e84fe5 pep8 3 years ago
Daniel Gyulai 2515b1e0f6 mInOr fIxEs aNd iMpRoVeMeNtS 3 years ago
Daniel Gyulai 5a7b63f9f8 Bump version 3 years ago
Daniel Gyulai 07eed2de3f Bump version to 0.0.11 3 years ago
Daniel Gyulai fec9263f5e Added install test 3 years ago
Daniel Gyulai 65a9cb504b Merge pull request '0.0.10' (#31) from 0.0.10 into master 3 years ago
Daniel Gyulai 83dfa63035 #23 3 years ago
Daniel Gyulai 26ccb6985e Close #30 3 years ago
Daniel Gyulai bfb997ae95 Grabber support in gen_dict 3 years ago
Daniel Gyulai 472ecc89bc Closes #27 3 years ago
Daniel Gyulai 24cdacfcb0 Close #28 3 years ago
Daniel Gyulai b75f2358b8 Fix #25 3 years ago
Daniel Gyulai 20d0a45a34 Version bump 3 years ago
Daniel Gyulai d90c5b7659 5_DockerRunner (#26) 3 years ago
Daniel Gyulai 93bc4a7d07 Merge pull request 'Use python logging' (#21) from 17_loglib into 0.0.10 3 years ago
Daniel Gyulai 51c78028f1 Use python logging 3 years ago
Daniel Gyulai 66bfdbcd2c 0.0.9 (#20) 3 years ago
Daniel Gyulai 9e4b61225c 7_runner-interface (#12) 3 years ago
Daniel Gyulai 10af87bc19 output hotfix 3 years ago
Daniel Gyulai 0585715506 Change detection via git 3 years ago
Daniel Gyulai 4cdc162d50 Verbose switch added 3 years ago
Daniel Gyulai ff431b4d0b Bump version 3 years ago
Daniel Gyulai f4492c0e45 Fix charsh with missing env in config 3 years ago
Daniel Gyulai fdc5982acc Update 'README.md' 3 years ago
Daniel Gyulai 7e6c1fa786 9-hotfix - Fix #9 (#10) 3 years ago
  1. 11
      .devcontainer/Dockerfile
  2. 32
      .devcontainer/devcontainer.json
  3. 71
      .drone.yml
  4. 282
      .gitignore
  5. 18
      LICENSE
  6. 17
      README.md
  7. 30
      alice-ci/README.md
  8. 10
      alice-ci/pyproject.toml
  9. 55
      alice-ci/setup.cfg
  10. 14
      alice-ci/src/alice/__init__.py
  11. 7
      alice-ci/src/alice/__main__.py
  12. 121
      alice-ci/src/alice/cli.py
  13. 51
      alice-ci/src/alice/config.py
  14. 101
      alice-ci/src/alice/configparser.py
  15. 20
      alice-ci/src/alice/exceptions.py
  16. 91
      alice-ci/src/alice/runnerfactory.py
  17. 4
      alice-ci/src/alice/runners/__init__.py
  18. 242
      alice-ci/src/alice/runners/dockerrunner.py
  19. 168
      alice-ci/src/alice/runners/pypirepo.py
  20. 345
      alice-ci/src/alice/runners/pypirunner.py
  21. 198
      alice-ci/src/alice/runners/pythonrunner.py
  22. 154
      alice-ci/src/alice/runners/pyutils.py
  23. 78
      alice-ci/src/alice/utils.py
  24. 127
      ci-examples/full.yaml
  25. 9
      ci-examples/images/hello/Dockerfile
  26. 2
      ci-examples/images/hello/hello.py
  27. 18
      ci-examples/python1.yaml
  28. 46
      docs/examples.md
  29. 58
      docs/runners.md
  30. 26
      docs/runners/docker.md
  31. 78
      docs/syntax.md

11
.devcontainer/Dockerfile

@ -0,0 +1,11 @@
# See here for image contents: https://github.com/microsoft/vscode-dev-containers/tree/v0.231.6/containers/ubuntu/.devcontainer/base.Dockerfile
# [Choice] Ubuntu version (use hirsuite or bionic on local arm64/Apple Silicon): hirsute, focal, bionic
ARG VARIANT="hirsute"
FROM mcr.microsoft.com/vscode/devcontainers/base:0-${VARIANT}
# [Optional] Uncomment this section to install additional OS packages.
# RUN apt-get update && export DEBIAN_FRONTEND=noninteractive \
# && apt-get -y install --no-install-recommends <your-package-list-here>

32
.devcontainer/devcontainer.json

@ -0,0 +1,32 @@
// For format details, see https://aka.ms/devcontainer.json. For config options, see the README at:
// https://github.com/microsoft/vscode-dev-containers/tree/v0.231.6/containers/ubuntu
{
"name": "Ubuntu",
"build": {
"dockerfile": "Dockerfile",
// Update 'VARIANT' to pick an Ubuntu version: hirsute, focal, bionic
// Use hirsute or bionic on local arm64/Apple Silicon.
"args": { "VARIANT": "focal" }
},
// Set *default* container specific settings.json values on container create.
"settings": {},
// Add the IDs of extensions you want installed when the container is created.
"extensions": [],
// Use 'forwardPorts' to make a list of ports inside the container available locally.
// "forwardPorts": [],
// Use 'postCreateCommand' to run commands after the container is created.
// "postCreateCommand": "uname -a",
// Comment out to connect as root instead. More info: https://aka.ms/vscode-remote/containers/non-root.
"remoteUser": "vscode",
"features": {
"docker-from-docker": "20.10",
"git": "latest",
"python": "3.10"
}
}

71
.drone.yml

@ -1,32 +1,41 @@
kind: pipeline
type: docker
name: default
steps:
- name: static-test
image: alpine/flake8
commands:
- python3 -m flake8 --ignore E501,W503
- name: build
image: python
commands:
- python3 -m pip install build
- python3 -m build alice-ci
- name: publish
image: python
environment:
TWINE_PASSWORD:
from_secret: pypi_username
TWINE_USERNAME:
from_secret: pypi_password
commands:
- python3 -m pip install twine
- python3 -m twine upload --verbose alice-ci/dist/*
when:
branch:
- master
event:
exclude:
kind: pipeline
type: docker
name: default
steps:
- name: static-test
image: alpine/flake8
commands:
- python3 -m flake8 --ignore E501,W503 alice-ci/src
- name: build
image: python
commands:
- python3 -m pip install build
- python3 -m build alice-ci
- name: test
image: python
environment:
PYPIUSER: USER
PYPIPASS: PASS
commands:
- python3 -m pip install alice-ci/dist/alice_ci*.whl
- alice -i ci-examples/full.yaml -vv
- name: publish
image: python
environment:
TWINE_PASSWORD:
from_secret: pypi_username
TWINE_USERNAME:
from_secret: pypi_password
commands:
- python3 -m pip install twine
- python3 -m twine upload --verbose alice-ci/dist/*
when:
branch:
- master
event:
exclude:
- pull_request

282
.gitignore

@ -1,140 +1,142 @@
# ---> Python
# Byte-compiled / optimized / DLL files
__pycache__/
*.py[cod]
*$py.class
# C extensions
*.so
# Distribution / packaging
.Python
build/
develop-eggs/
dist/
downloads/
eggs/
.eggs/
lib/
lib64/
parts/
sdist/
var/
wheels/
share/python-wheels/
*.egg-info/
.installed.cfg
*.egg
MANIFEST
# PyInstaller
# Usually these files are written by a python script from a template
# before PyInstaller builds the exe, so as to inject date/other infos into it.
*.manifest
*.spec
# Installer logs
pip-log.txt
pip-delete-this-directory.txt
# Unit test / coverage reports
htmlcov/
.tox/
.nox/
.coverage
.coverage.*
.cache
nosetests.xml
coverage.xml
*.cover
*.py,cover
.hypothesis/
.pytest_cache/
cover/
# Translations
*.mo
*.pot
# Django stuff:
*.log
local_settings.py
db.sqlite3
db.sqlite3-journal
# Flask stuff:
instance/
.webassets-cache
# Scrapy stuff:
.scrapy
# Sphinx documentation
docs/_build/
# PyBuilder
.pybuilder/
target/
# Jupyter Notebook
.ipynb_checkpoints
# IPython
profile_default/
ipython_config.py
# pyenv
# For a library or package, you might want to ignore these files since the code is
# intended to run in multiple environments; otherwise, check them in:
# .python-version
# pipenv
# According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
# However, in case of collaboration, if having platform-specific dependencies or dependencies
# having no cross-platform support, pipenv may install dependencies that don't work, or not
# install all needed dependencies.
#Pipfile.lock
# PEP 582; used by e.g. github.com/David-OConnor/pyflow
__pypackages__/
# Celery stuff
celerybeat-schedule
celerybeat.pid
# SageMath parsed files
*.sage.py
# Environments
.env
.venv
env/
venv/
ENV/
env.bak/
venv.bak/
# Spyder project settings
.spyderproject
.spyproject
# Rope project settings
.ropeproject
# mkdocs documentation
/site
# mypy
.mypy_cache/
.dmypy.json
dmypy.json
# Pyre type checker
.pyre/
# pytype static type analyzer
.pytype/
# Cython debug symbols
cython_debug/
# ---> Python
# Byte-compiled / optimized / DLL files
__pycache__/
*.py[cod]
*$py.class
# C extensions
*.so
# Distribution / packaging
.Python
build/
develop-eggs/
dist/
downloads/
eggs/
.eggs/
lib/
lib64/
parts/
sdist/
var/
wheels/
share/python-wheels/
*.egg-info/
.installed.cfg
*.egg
MANIFEST
# PyInstaller
# Usually these files are written by a python script from a template
# before PyInstaller builds the exe, so as to inject date/other infos into it.
*.manifest
*.spec
# Installer logs
pip-log.txt
pip-delete-this-directory.txt
# Unit test / coverage reports
htmlcov/
.tox/
.nox/
.coverage
.coverage.*
.cache
nosetests.xml
coverage.xml
*.cover
*.py,cover
.hypothesis/
.pytest_cache/
cover/
# Translations
*.mo
*.pot
# Django stuff:
*.log
local_settings.py
db.sqlite3
db.sqlite3-journal
# Flask stuff:
instance/
.webassets-cache
# Scrapy stuff:
.scrapy
# Sphinx documentation
docs/_build/
# PyBuilder
.pybuilder/
target/
# Jupyter Notebook
.ipynb_checkpoints
# IPython
profile_default/
ipython_config.py
# pyenv
# For a library or package, you might want to ignore these files since the code is
# intended to run in multiple environments; otherwise, check them in:
# .python-version
# pipenv
# According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
# However, in case of collaboration, if having platform-specific dependencies or dependencies
# having no cross-platform support, pipenv may install dependencies that don't work, or not
# install all needed dependencies.
#Pipfile.lock
# PEP 582; used by e.g. github.com/David-OConnor/pyflow
__pypackages__/
# Celery stuff
celerybeat-schedule
celerybeat.pid
# SageMath parsed files
*.sage.py
# Environments
.env
.venv
env/
venv/
ENV/
env.bak/
venv.bak/
# Spyder project settings
.spyderproject
.spyproject
# Rope project settings
.ropeproject
# mkdocs documentation
/site
# mypy
.mypy_cache/
.dmypy.json
dmypy.json
# Pyre type checker
.pyre/
# pytype static type analyzer
.pytype/
# Cython debug symbols
cython_debug/
# persistency dir
.alice

18
LICENSE

@ -1,9 +1,9 @@
MIT License
Copyright (c) 2022 Daniel Gyulai
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
MIT License
Copyright (c) 2022 Daniel Gyulai
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.

17
README.md

@ -1,7 +1,12 @@
# alice
CI framework with support for local running.
* [Basic usage](alice-ci/README.md)
* [Runners](docs/runners.md)
# alice
CI framework with support for local running.
Main repo [here](https://git.gyulai.cloud/gyulaid/alice).
[![Build Status](https://ci.gyulai.cloud/api/badges/gyulaid/alice/status.svg)](https://ci.gyulai.cloud/gyulaid/alice)
[![PyPI version](https://badge.fury.io/py/alice-ci.svg)](https://badge.fury.io/py/alice-ci)
* [Basic usage](alice-ci/README.md)
* [Runners](docs/runners.md)
* [CI syntax](docs/syntax.md)

30
alice-ci/README.md

@ -1,16 +1,16 @@
# Alice-CI
Continous Integration framework with the goal of using the exact same code in CI and local env. Steps can be defined in yaml files, for syntax see the docs. Runs on LInux and Windows, Mac should work too, but not yet tested.
## Usage
Install with pip:
```
pythom3 -m pip install alice-ci
```
To run:
```
pythom3 -m alice [-i <ci.yaml>] STEPS
# Alice-CI
Continous Integration framework with the goal of using the exact same code in CI and local env. Steps can be defined in yaml files, for syntax see the docs. Runs on LInux and Windows, Mac should work too, but not yet tested.
## Usage
Install with pip:
```
pythom3 -m pip install alice-ci
```
To run:
```
pythom3 -m alice [-i <ci.yaml>] STEPS
```

10
alice-ci/pyproject.toml

@ -1,6 +1,6 @@
[build-system]
requires = [
"setuptools>=42",
"wheel"
]
[build-system]
requires = [
"setuptools>=42",
"wheel"
]
build-backend = "setuptools.build_meta"

55
alice-ci/setup.cfg

@ -1,26 +1,29 @@
[metadata]
name = alice-ci
version = 0.0.2
author = Daniel Gyulai
description = Alice CI framework
long_description = file: README.md
long_description_content_type = text/markdown
url = https://git.gyulai.cloud/gyulaid/alice
project_urls =
Bug Tracker = https://git.gyulai.cloud/gyulaid/alice/issues
classifiers =
Programming Language :: Python :: 3
License :: OSI Approved :: MIT License
Operating System :: OS Independent
[options]
package_dir =
= src
packages = find:
python_requires = >=3.6
install_requires =
PyYAML==6.0
virtualenv==20.14.0
[options.packages.find]
where = src
[metadata]
name = alice-ci
version = 0.0.14
author = Daniel Gyulai
description = Alice CI framework
long_description = file: README.md
long_description_content_type = text/markdown
url = https://git.gyulai.cloud/gyulaid/alice
project_urls =
Bug Tracker = https://git.gyulai.cloud/gyulaid/alice/issues
classifiers =
Programming Language :: Python :: 3
License :: OSI Approved :: MIT License
Operating System :: OS Independent
[options]
package_dir =
= src
packages =
alice
alice.runners
python_requires = >=3.6
install_requires =
PyYAML
docker
[options.entry_points]
console_scripts =
alice = alice.cli:main

14
alice-ci/src/alice/__init__.py

@ -1,5 +1,9 @@
# flake8: noqa F401
from .cli import App
from .jobparser import Job, JobParser
from .exceptions import NonZeroRetcode
from .pythonrunner import PythonRunner
# flake8: noqa F401
from .configparser import ConfigParser
from .exceptions import NonZeroRetcode
from .runnerfactory import Factory
from .exceptions import NonZeroRetcode
from .exceptions import RunnerError
from .exceptions import ConfigException
name = "alice"

7
alice-ci/src/alice/__main__.py

@ -1,3 +1,4 @@
from cli import main
main()
from .cli import main
if __name__ == '__main__':
main()

121
alice-ci/src/alice/cli.py

@ -1,59 +1,62 @@
import os
import argparse
from utils import ConfigParser
from runnerfactory import Factory
from exceptions import ConfigException, NonZeroRetcode, RunnerError
def gen_env(self, param_list):
env_vars = {}
for item in param_list:
item_parts = item.split("=")
if len(item_parts) == 2:
env_vars[item_parts[0]] = item_parts[1]
else:
raise ConfigException(f"Invalid parameter: {item}")
return env_vars
def parse_jobs(args):
try:
factory = Factory()
if len(args.env) > 0:
factory.update_runners({"env": gen_env(args.env)})
jobParser = ConfigParser(args.input, factory)
print("Begin pipeline steps...")
for step in args.steps:
if step in jobParser.jobs:
jobParser.execute_job(step)
print(f"[Step] {step}: SUCCESS")
else:
print(f"Step {step} not found in {args.input}")
exit(1)
except ConfigException as e:
print(f"Configuration error-> {e}")
exit(1)
except NonZeroRetcode:
print("FAILED")
exit(1)
except RunnerError as e:
print(f"RunnerError-> {e}")
def main():
parser = argparse.ArgumentParser(prog="alice")
parser.add_argument("steps", nargs='+')
parser.add_argument("-i", "--input", default="alice-ci.yaml")
parser.add_argument("-e", "--env", nargs='*', default=[])
parser.add_argument("-a", "--addrunner", nargs='*', default=[])
args = parser.parse_args()
if not os.path.isfile(args.input):
print(f"No such file: {args.input}")
exit(1)
parse_jobs(args)
if __name__ == "__main__":
main()
import logging
import os
import argparse
from .configparser import ConfigParser
from .exceptions import ConfigException, NonZeroRetcode, RunnerError
def gen_env(param_list):
env_vars = {}
for item in param_list:
item_parts = item.split("=")
if len(item_parts) == 2:
env_vars[item_parts[0]] = item_parts[1]
else:
raise ConfigException(f"Invalid parameter: {item}")
return env_vars
def parse_jobs(args):
try:
if len(args.env) > 0:
envs = gen_env(args.env)
logging.debug(f"[Alice] Env vars from CLI: {envs}")
jobParser = ConfigParser(args.input, gen_env(args.env))
for step in args.steps:
jobParser.execute(step)
except ConfigException as e:
print(f"Configuration error-> {e}")
exit(1)
except NonZeroRetcode:
print("[Alice] FAILED")
exit(1)
except RunnerError as e:
print(f"RunnerError-> {e}")
exit(1)
def main():
parser = argparse.ArgumentParser(prog="alice")
parser.add_argument("steps", nargs='*', default=["default"])
parser.add_argument("-i", "--input", default=".alice-ci.yaml")
parser.add_argument("-e", "--env", nargs='*', default=[])
parser.add_argument("-a", "--addrunner", nargs='*', default=[])
parser.add_argument('--verbose', '-v', action='count', default=0)
args = parser.parse_args()
loglevel = 30 - ((10 * args.verbose) if args.verbose > 0 else 0)
logging.basicConfig(level=loglevel, format='%(message)s')
if not os.path.isfile(args.input):
print(f"No such file: {args.input}")
exit(1)
persistency_path = os.path.join(os.getcwd(), ".alice")
if not os.path.isdir(persistency_path):
os.mkdir(persistency_path)
parse_jobs(args)
if __name__ == "__main__":
main()

51
alice-ci/src/alice/config.py

@ -0,0 +1,51 @@
import logging
import os
from .exceptions import ConfigException
class ConfigHolder:
__instance = None
file_name = os.path.join(os.getcwd(), ".alice", "vars")
@staticmethod
def getInstance():
""" Static access method. """
if ConfigHolder.__instance is None:
ConfigHolder()
return ConfigHolder.__instance
def __init__(self):
""" Virtually private constructor. """
if ConfigHolder.__instance is not None:
raise Exception("This class is a singleton!")
else:
ConfigHolder.__instance = self
config = os.path.abspath(os.path.join(os.getcwd(), self.file_name))
self.vars = {}
if os.path.isfile(config):
with open(config) as f:
for _line in f:
line = _line.strip()
items = line.split("=")
if len(items) > 1:
self.vars[items[0]] = line.replace(f"{items[0]}=", "")
logging.debug(f"Loaded from {self.file_name}: {self.vars}")
def get(self, key):
try:
return self.vars[key]
except KeyError:
raise ConfigException(f"{key} not defined in .conf!")
def set(self, key, value):
self.vars[key] = value
self.commit()
def soft_set(self, key, value):
self.vars[key] = value
def commit(self):
with open(self.file_name, 'w') as f:
for k, v in self.vars.items():
f.write(f"{k}={v if v is not None else ''}\n")

101
alice-ci/src/alice/configparser.py

@ -0,0 +1,101 @@
import logging
from os import getcwd, path, environ
import subprocess
import yaml
from .exceptions import ConfigException
from .runnerfactory import Factory
class ConfigParser:
def __init__(self, file_path, cli_env_vars) -> None:
with open(file_path) as f:
self.config = yaml.safe_load(f)
self.factory = Factory(self.__gen_globals(cli_env_vars), self.config.get("runners", {}))
self.jobs = self.__get_jobs()
self.pipelines = self.config.get("pipelines", {})
# Initialize env and workdir if not present in global
def __gen_globals(self, cli_vars):
env_vars = environ.copy()
env_vars.update(cli_vars)
globals = {
"env": env_vars,
"workdir": getcwd()
}
if "runners" in self.config:
if "global" in self.config["runners"]:
if "env" in self.config["runners"]["global"]:
for var in self.config["runners"]["global"]["env"]:
globals["env"][var["name"]] = var["value"]
if "workdir" in self.config["runners"]["global"]:
globals["workdir"] = self.config["runners"]["global"]["workdir"]
logging.debug(f"[Alice] Configured globals: {globals}")
return globals
def __get_jobs(self):
if "jobs" in self.config:
jobs = {}
for job_spec in self.config["jobs"]:
name = job_spec["name"]
if name in jobs:
raise ConfigException(f"Job with name {name} already exists!")
jobs[name] = job_spec
logging.info(f"[Alice] Parsed jobs: {', '.join(jobs.keys())}")
return jobs
else:
raise ConfigException("No jobs defined in config")
def __is_changed(self, changes):
try:
target = changes["branch"]
paths = []
for _path in changes["paths"]:
paths.append(path.abspath(_path))
# TODO: Error handling
command = ["git", "diff", "--name-only", target]
with subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE) as p:
p.wait()
for line in p.stdout:
change_path = path.abspath(line.decode("UTF-8").strip())
for _path in paths:
spec_path = path.abspath(_path)
if change_path.startswith(spec_path):
logging.info(f"[Alice] Modified file: {change_path}")
logging.info(f"[Alice] Path match: {_path}")
return True
except KeyError:
raise ConfigException(f"Invalid 'changes' config: {changes}")
return False
def execute(self, task_name):
if task_name in self.jobs:
self.execute_job(task_name)
elif task_name in self.pipelines:
self.execute_pipeline(task_name)
else:
raise ConfigException(f"No such job or pipeline: {task_name}")
def execute_pipeline(self, pipeline_name):
if pipeline_name in self.pipelines:
print(f"[Alice][Pipeline] {pipeline_name}: Start")
for task in self.pipelines[pipeline_name]:
self.execute(task)
print(f"[Alice][Pipeline] {pipeline_name}: Success")
def execute_job(self, job_name):
if job_name in self.jobs:
print(f"[Alice][Job] {job_name}: Start")
job_spec = self.jobs[job_name]
should_run = True
if "changes" in job_spec:
should_run = self.__is_changed(job_spec["changes"])
if should_run:
runner = self.factory.get_runner(job_spec["type"])
runner.run(job_spec)
status = "SUCCESS"
else:
status = "SKIP, no change detected"
print(f"[Alice][Job] {job_name}: {status}")

20
alice-ci/src/alice/exceptions.py

@ -1,10 +1,10 @@
class NonZeroRetcode(Exception):
pass
class RunnerError(Exception):
pass
class ConfigException(Exception):
pass
class NonZeroRetcode(Exception):
pass
class RunnerError(Exception):
pass
class ConfigException(Exception):
pass

91
alice-ci/src/alice/runnerfactory.py

@ -1,34 +1,57 @@
from runners.pythonrunner import PythonRunner
from os import getcwd
class Factory():
def __init__(self) -> None:
self.runnertypes = self.__load_runners()
self.runners = {}
self.workdir = getcwd()
self.globals = {}
def __load_runners(self):
# TODO: Runners can be imported via cli too
# module = __import__("module_file")
# my_class = getattr(module, "class_name")
return {"python": PythonRunner}
def set_globals(self, globals):
self.globals = globals
def update_globals(self, update):
if "env" in update:
self.globals["env"].update(update["env"])
def update_runners(self, config):
for runnertype, runnerconfig in config.items():
if runnertype != "global":
self.get_runner(runnertype).update_config(runnerconfig)
def get_runner(self, runnertype):
if runnertype not in self.runners:
self.runners[runnertype] = self.runnertypes[runnertype](self.workdir, self.globals)
return self.runners[runnertype]
import logging
from os.path import join, abspath
from .runners.pythonrunner import PythonRunner
from .runners.pypirunner import PyPiRunner
from .runners.dockerrunner import DockerRunner
from .runners.pypirepo import PypiRepoRunner
from .exceptions import ConfigException
class Factory():
def __init__(self, globals, runner_configs) -> None:
self.globals = globals
self.runner_configs = {}
self.runnertypes = {}
self.runners = {}
self.__load_runners()
self.__gen_runner_configs(runner_configs)
def __load_runners(self):
# TODO: Runners can be imported via cli too
# https://git.gyulai.cloud/gyulaid/alice/issues/4
# module = __import__("module_file")
# my_class = getattr(module, "class_name")
self.runnertypes = {"python": PythonRunner,
"pypi": PyPiRunner,
"docker": DockerRunner,
"pypirepo": PypiRepoRunner}
logging.info(f"[Alice] Available runners: {'|'.join(self.runnertypes.keys())}")
def __gen_runner_configs(self, config):
for runnertype, runnerconfig in config.items():
if runnertype != "global":
logging.info(f"[Alice] Global config found for runner {runnertype}")
config = self.globals.copy()
for key, value in runnerconfig.items():
if key == "env":
for env_var in value:
config["env"][env_var["name"]] = env_var["value"]
elif key == "workdir":
config["workdir"] = abspath(join(config["workdir"], value))
else:
config[key] = value
self.runner_configs[runnertype] = config
logging.debug(f"[Alice] Globals for {runnertype}: {runnerconfig}")
def get_runner(self, runnertype):
if runnertype not in self.runners:
if runnertype in self.runnertypes:
logging.info(f"[Alice] Initializing runner: {runnertype}")
# If there is a runner specific config, use that, else global
config = self.runner_configs.get(runnertype, self.globals.copy())
self.runners[runnertype] = self.runnertypes[runnertype](config)
else:
raise ConfigException(f"Invalid runner type: {runnertype}")
return self.runners[runnertype]

4
alice-ci/src/alice/runners/__init__.py

@ -0,0 +1,4 @@
# flake8: noqa F401
from .pythonrunner import PythonRunner
from .pypirunner import PyPiRunner
from .dockerrunner import DockerRunner

242
alice-ci/src/alice/runners/dockerrunner.py

@ -1 +1,241 @@
# TODO Implement
from enum import Enum
import json
import logging
from os import path, getcwd
import docker
from .pyutils import grab_from, gen_dict
from ..exceptions import ConfigException, NonZeroRetcode, RunnerError
class ImageSource(Enum):
NONE = 1
BUILD = 2
PULL = 3
def get_user(config, default):
if "credentials" in config:
if "username" in config["credentials"]:
data = config["credentials"]["username"]
if isinstance(data, str):
return data
else:
return grab_from(data)
return default
def get_pass(config, default):
if "credentials" in config:
if "password" in config["credentials"]:
data = config["credentials"]["password"]
if isinstance(data, str):
return data
else:
return grab_from(data)
return default
def get_provider(config, default, default_type):
if "image" in config:
build = False
pull = False
candidate_type = default_type
if "build" in config["image"]:
build = True
if default_type == ImageSource.BUILD:
candidate = default.copy(config["image"]["build"])
else:
candidate = Builder(config["image"]["build"])
candidate_type = ImageSource.BUILD
elif "pull" in config["image"]:
pull = True
if default_type == ImageSource.PULL:
candidate = default.copy(config["image"]["pull"])
else:
candidate = Puller(config["image"]["pull"])
candidate_type = ImageSource.PULL
if build and pull:
raise ConfigException("[DockerRunner] Can't build and pull the same image!")
return candidate, candidate_type
return default, default_type
class Tagger:
def __init__(self, config={}) -> None:
self.name = config.get("name", None)
self.username = get_user(config, None)
self.password = get_pass(config, None)
self.publish = config.get("publish", False)
def copy(self, job_config):
t = Tagger()
t.name = job_config.get("name", self.name)
t.username = get_user(job_config, self.username)
t.password = get_pass(job_config, self.password)
t.publish = job_config.get("publish", self.publish)
return t
def __str__(self) -> str:
data = {
"name": self.name,
"publish": self.publish,
"credentials": {
"username": self.username,
"password": self.password
}
}
return f"{data}"
def handle(self, client, image):
if self.name is not None:
if self.name not in image.tags and f"{self.name}:latest" not in image.tags:
print(f"[DockerRunner] Tagging {image.tags[0]} as {self.name}")
image.tag(self.name)
if self.publish:
print(f"[DockerRunner] Pushing {self.name}")
client.push(self.name)
class Builder():
def __init__(self, config) -> None:
self.dir = path.abspath(config.get("dir", getcwd()))
self.dockerfile = config.get("dockerfile", None)
self.name = config.get("name", None)
self.args = gen_dict(config.get("args", []))
def copy(self, job_config):
b = Builder({})
b.dir = path.abspath(path.join(self.dir, job_config.get("dir", ".")))
b.dockerfile = job_config.get("dockerfile", self.dockerfile)
b.name = job_config.get("name", self.name)
b.args = self.args.copy().update(gen_dict(job_config.get("args", [])))
return b
def __str__(self) -> str:
data = {
"type": "builder",
"dir": self.dir,
"dockerfile": self.dockerfile,
"name": self.name,
"args": self.args
}
return json.dumps(data)
def prepare(self, client):
print(f"[DockerRunner] Building image {self.name}")
if self.dockerfile is None:
self.dockerfile = "Dockerfile"
try:
image, log = client.images.build(path=self.dir,
dockerfile=self.dockerfile,
tag=self.name,
buildargs=self.args,
labels={"builder": "alice-ci"})
for i in log:
logging.debug(i)
return image
except docker.errors.BuildError as e:
raise RunnerError(f"[DockerRunner] Build failed: {e}")
except docker.errors.APIError as e:
raise RunnerError(f"[DockerRunner] Error: {e}")
class Puller():
def __init__(self, config={}) -> None:
self.name = config.get("name", None)
self.username = get_user(config, None)
self.password = get_pass(config, None)
def copy(self, job_config={}):
p = Puller()
p.name = job_config.get("name", self.name)
p.username = get_user(job_config, self.username)
p.password = get_pass(job_config, self.password)
def __str__(self) -> str:
data = {
"name": self.name,
"credentials": {
"username": self.username,
"password": self.password
}
}
return f"{data}"
def prepare(self, client):
print(f"[DockerRunner] Pulling image {self.name}")
return client.images.pull(self.name)
class DockerConfig:
def __init__(self, config={}) -> None:
self.username = get_user(config, None)
self.password = get_pass(config, None)
self.image_provider, self.provider_type = get_provider(config, None, ImageSource.NONE)
self.tagger = Tagger(config.get("tag", {}))
self.commands = config.get("commands", [])
self.env = config.get("env", {})
def copy(self, job_config={}):
d = DockerConfig()
d.username = get_user(job_config, self.username)
d.password = get_pass(job_config, self.password)
d.image_provider, d.provider_type = get_provider(job_config, self.image_provider, self.provider_type)
d.tagger = self.tagger.copy(job_config.get("tag", {}))
d.commands = self.commands.copy() + job_config.get("commands", [])
d.env = self.env.copy()
d.env.update(gen_dict(job_config.get("env", [])))
return d
def __str__(self) -> str:
data = {
"credentials": {
"username": {self.username},
"password": {self.password}
},
"image": self.image_provider.__str__(),
"commands": self.commands,
"tag": self.tagger.__str__()
}
return f"{data}"
class DockerRunner():
def __init__(self, config) -> None:
logging.info("[DockerRunner] Initializing")
self.config = DockerConfig(config)
self.client = docker.from_env()
def run(self, job_spec):
job_config = self.config.copy(job_spec)
logging.debug(f"[DockerRunner] Job config: {job_config.__str__()}")
if job_config.image_provider is None:
raise RunnerError("[DockerRunner] No image provider configured!")
image = job_config.image_provider.prepare(self.client)
logging.info(f"[DockerRunner] Image: {image.tags} ({image.id})")
if len(job_config.commands) > 0:
if "PATH" in job_config.env:
del job_config.env["PATH"]
container = self.client.containers.run(image=image.id,
entrypoint=["sleep", "infinity"],
detach=True,
auto_remove=True)
try:
for i in job_config.commands:
command = ["/bin/sh", "-c", i]
logging.debug(f"[DockerRunner] Command array: {command}")
code, output = container.exec_run(cmd=command,
environment=job_config.env)
for line in output.decode("UTF-8").splitlines():
print(f"[{job_spec['name']}] {line}")
if code != 0:
raise NonZeroRetcode(f"Command {i} returned code {code}")
finally:
if container is not None:
container.stop()
job_config.tagger.handle(self.client, image)

168
alice-ci/src/alice/runners/pypirepo.py

@ -0,0 +1,168 @@
import logging
import subprocess
import docker
from os.path import join, isdir
from os import getcwd, mkdir
import os
import requests
import platform
import time
from ..exceptions import RunnerError
from ..config import ConfigHolder
pipconf = """[global]
index-url = URL
trusted-host = BASE
pypi.org
extra-index-url= http://pypi.org/simple"""
class RepoConfig:
def __init__(self, config={}) -> None:
self.port = config.get("port", 8888)
self.enabled = config.get("enabled", True)
self.container_name = config.get("container_name", "alice-pypiserver")
def copy(self, job_config):
r = RepoConfig()
r.container_name = job_config.get("container_name", self.container_name)
r.enabled = job_config.get("enabled", self.enabled)
r.port = job_config.get("port", self.port)
return r
class PypiRepoRunner:
def __init__(self, config) -> None:
logging.info("[PyPiRepo] Initializing")
self.config = RepoConfig(config)
self.client = docker.from_env()
self.user = "alice"
self.passwd = "alice"
self.htpasswd = 'alice:{SHA}UisnajVr3zkBPfq+os1D4UHsyeg='
def get_image(self):
# TODO: remove when resolved:
# Official Docker image support for ARM?
# https://github.com/pypiserver/pypiserver/issues/364
pypiserver = "https://github.com/pypiserver/pypiserver.git"
if platform.machine() == "aarch64":
tag = "alice.localhost/pypiserver:arm"
try:
self.client.images.get(tag)
return tag
except docker.errors.ImageNotFound:
print("[PyPiRepo] Building PyPiServer ARM image, this could take a while")
workdir = join(getcwd(), ".alice", "pypirepo", "source")
if not os.path.isdir(workdir):
os.mkdir(workdir)
git_command = ["git", "clone", pypiserver, "--branch=v1.3.2"]
output = []
with subprocess.Popen(git_command, cwd=workdir, stdout=subprocess.PIPE) as p:
for line in p.stdout:
output.append(line.decode('utf8').strip())
p.wait()
if p.returncode != 0:
print("\n".join(output))
raise(RunnerError("[PyPiRepo] Could not fetch pypiserver source"))
source_path = os.path.join(workdir, "pypiserver")
self.client.images.build(path=source_path, tag=tag)
return tag
else:
return "pypiserver/pypiserver:latest"
def run(self, job_spec):
job_config = self.config.copy(job_spec)
docker_host_ip = None
for network in self.client.networks.list():
if network.name == "bridge":
try:
docker_host_ip = network.attrs["IPAM"]["Config"][0]["Gateway"]
except KeyError:
docker_host_ip = network.attrs["IPAM"]["Config"][0]["Subnet"].replace(".0/16", ".1")
if docker_host_ip is None:
raise RunnerError("Unable to determine Docker host IP")
if job_config.enabled:
try:
c = self.client.containers.get(job_config.container_name)
print(f"[PyPiRepo] {job_config.container_name} already running")
except docker.errors.NotFound:
persistency_dir = join(getcwd(), ".alice", "pypirepo")
if not isdir(persistency_dir):
mkdir(persistency_dir)
package_dir = join(persistency_dir, "packages")
if not isdir(package_dir):
mkdir(package_dir)
htpasswd_file = join(persistency_dir, ".htpasswd")
with open(htpasswd_file, 'w') as f:
f.write(self.htpasswd)
c = self.client.containers.run(
name=job_config.container_name,
image=self.get_image(),
detach=True,
labels={"app": "alice"},
command=["--overwrite", "-P", ".htpasswd", "packages"],
ports={"8080/tcp": job_config.port},
volumes={
package_dir: {
"bind": "/data/packages",
"mode": "rw"
},
htpasswd_file: {
"bind": "/data/.htpasswd",
"mode": "ro"
}
},
restart_policy={
"Name": "unless-stopped"
}
)
print(f"[PyPiRepo] Started {job_config.container_name}")
c.reload()
logging.info(f"[PyPiRepo] {job_config.container_name} : {c.status}")
if c.status != "running":
raise RunnerError(f"[PyPiRepo] Repo container unstable: {c.status}")
uri = f"http://localhost:{job_config.port}"
unreachable = True
attempts = 0
while unreachable and attempts < 5:
attempts += 1
try:
requests.get(uri)
unreachable = False
except Exception as e:
logging.info(f"[PyPiRepo] {attempts} - Repo at {uri} is unavailable: {e}")
time.sleep(2)
if unreachable:
raise RunnerError(f"[PyPiRepo] Repo unreachable")
cfgh = ConfigHolder.getInstance()
cfgh.soft_set("PYPI_USER", self.user)
cfgh.soft_set("PYPI_PASS", self.passwd)
cfgh.soft_set("PYPI_REPO", uri)
cfgh.soft_set("DOCKER_PYPI_USER", self.user)
cfgh.soft_set("DOCKER_PYPI_PASS", self.passwd)
cfgh.soft_set("DOCKER_PYPI_REPO", f"http://{docker_host_ip}:{job_config.port}")
cfgh.commit()
venv = join(os.getcwd(), "venv")
if os.path.isdir(venv):
netloc = f"localhost:{job_config.port}"
url = f"http://{self.user}:{self.passwd}@{netloc}"
conf = pipconf.replace("URL", url).replace("BASE", netloc)
if os.name == "nt": # Windows
filename = join(venv, "pip.ini")
else: # Linux & Mac
filename = join(venv, "pip.conf")
with open(filename, 'w') as f:
f.write(conf)
print(f"[PyPiRepo] pip conf written to {filename}")

345
alice-ci/src/alice/runners/pypirunner.py

@ -0,0 +1,345 @@
from distutils.command.config import config
from distutils.log import debug
import json
import logging
from ntpath import join
import os
import re
import subprocess
import sys
from pkg_resources import parse_version
from requests import get
from requests.auth import HTTPBasicAuth
from os import environ, path
from html.parser import HTMLParser
from alice.runners.pyutils import PackageManager, glob, grab_from
from alice.exceptions import ConfigException, RunnerError
import hashlib
from pathlib import Path
def md5_update_from_file(filename, hash):
assert Path(filename).is_file()
with open(str(filename), "rb") as f:
for chunk in iter(lambda: f.read(4096), b""):
hash.update(chunk)
return hash
def md5_file(filename):
return md5_update_from_file(filename, hashlib.md5()).hexdigest()
def md5_update_from_dir(directory, hash, exclude_dirs, exclude_extensions, exclude_dirs_wildcard):
assert Path(directory).is_dir()
for _path in os.listdir(directory):
path = os.path.join(directory, _path)
if os.path.isfile(path) :
hash.update(_path.encode())
logging.debug(f"[PyPiRunner][Hash] File: {path}")
hash = md5_update_from_file(path, hash)
elif os.path.isdir(path):
skip = False
for name in exclude_dirs:
if name in os.path.basename(_path):
skip = True
if not skip:
hash = md5_update_from_dir(path, hash, exclude_dirs, exclude_extensions, exclude_dirs_wildcard)
return hash
def md5_dir(directory, exclude_dirs=[], exclude_extensions=[], exclude_dirs_wildcard=[]):
return md5_update_from_dir(directory, hashlib.sha1(), exclude_dirs, exclude_extensions, exclude_dirs_wildcard).hexdigest()
def get_uri(config, default):
url = config.get("repo", {}).get("uri", default)
if url is not None:
if not isinstance(url, str):
url = grab_from(url)
if not re.match('(?:http|ftp|https)://', url):
url = f"https://{url}"
return url
def get_user(config, default):
if "repo" in config:
if "username" in config["repo"]:
data = config["repo"]["username"]
if isinstance(data, str):
return data
else:
return grab_from(data)
return default
def get_pass(config, default):
if "repo" in config:
if "password" in config["repo"]:
data = config["repo"]["password"]
if isinstance(data, str):
return data
else:
return grab_from(data)
return default
class SimpleRepoParser(HTMLParser):
def __init__(self):
HTMLParser.__init__(self)
self.packages = []
def handle_data(self, data):
re_groups = re.findall("(\d*\.\d*\.\d*)", data)
if len(re_groups) == 1:
file_version = re_groups[0]
if file_version not in self.packages:
self.packages.append(file_version)
# Parses and stores the config from yaml
class PypiConfig:
def __init__(self, config={}) -> None:
self.workdir = path.abspath(config.get("workdir", "."))
self.repo_uri = get_uri(config, None)
self.repo_user = get_user(config, None)
self.repo_pass = get_pass(config, None)
self.packages = set(config.get("packages", []))
self.upload = config.get("upload", False)
self.fail_if_exists = config.get("fail_if_exists", False)
# returns a PyPiConfig with merged values
def copy(self, job_config={}):
p = PypiConfig()
p.workdir = path.abspath(path.join(self.workdir, job_config.get("workdir", ".")))
p.repo_uri = get_uri(job_config, self.repo_uri)
p.repo_user = get_user(job_config, self.repo_user)
p.repo_pass = get_pass(job_config, self.repo_pass)
job_pkg_set = set(job_config["packages"])
job_pkg_set.update(self.packages)
p.packages = job_pkg_set
p.upload = job_config.get("upload", self.upload)
p.fail_if_exists = job_config.get("fail_if_exists", self.fail_if_exists)
return p
# TODO: Refactor to something sensible, more flexible
class PackageMeta:
def __init__(self):
self.conf_dir = path.join(os.getcwd(), ".alice", "pypirunner")
self.metafile = path.join(self.conf_dir, "packagemeta.json")
if not path.isdir(self.conf_dir):
os.mkdir(self.conf_dir)
if path.isfile(self.metafile):
with open(self.metafile) as f:
self.metadata = json.load(f)
else:
self.metadata = {}
self.__save()
def __save(self):
with open(self.metafile, 'w') as f:
json.dump(self.metadata, f)
def get(self, package, key):
return self.metadata.get(package, {}).get(key, "")
def set(self, package, key, value):
if package not in self.metadata:
self.metadata[package] = {}
self.metadata[package][key] = value
self.__save()
# TODO: consider "--skip-existing" flag for twine
class PyPiRunner():
def __init__(self, config) -> None:
logging.info("[PyPiRunner] Initializing")
self.workdir = config["workdir"]
self.config = PypiConfig(config)
self.metadata = PackageMeta()
def __versions(self, config, pkg_name):
repo = config.repo_uri
if repo is None:
repo = "https://pypi.python.org/pypi"
if config.repo_pass is not None and config.repo_user is not None:
logging.info(f"[PyPiRunner][Versions] Set auth headers from config")
logging.debug(f"[PyPiRunner][Versions] Auth: {config.repo_user}:{config.repo_pass}")
auth = HTTPBasicAuth(config.repo_user, config.repo_pass)
else:
logging.info(f"[PyPiRunner][Versions] No auth headers in config, skip")
logging.debug(f"[PyPiRunner][Versions] Auth: {config.repo_user}:{config.repo_pass}")
auth = None
try:
if repo.endswith("pypi"):
url = f'{repo}/{pkg_name}/json'
logging.info(f"[PyPiRunner][Versions] Trying JSON API at {url}")
response = get(url, auth=auth)
if response.status_code == 200:
releases = json.loads(response.text)["releases"]
return sorted(releases, key=parse_version, reverse=True)
else:
logging.info(f"[PyPiRunner][Versions] JSON failed: [{response.status_code}]")
logging.debug(response.text)
repo = f"{repo}/simple"
url = f"{repo}/{pkg_name}"
logging.info(f"[PyPiRunner][Versions] Trying Simple API at {url}")
response = get(url, auth=auth)
if response.status_code == 200:
parser = SimpleRepoParser()
parser.feed(response.text)
return sorted(parser.packages, key=parse_version, reverse=True)
if response.status_code == 404:
return []
else:
logging.info(f"[PyPiRunner][Versions] Simple failed: [{response.status_code}]")
logging.debug(response.text)
raise Exception("Failed to fetch available versions")
except Exception as e:
raise RunnerError(f"{url}: {e}")
def build(self, config, package):
print(f"[PyPiRunner] Building {package}")
pkg_path = path.join(config.workdir, package)
if not path.isdir(pkg_path):
raise ConfigException(f"Path does not exists: {pkg_path}")
PackageManager.getInstance().ensure("build")
command = [sys.executable, "-m", "build", package]
if logging.root.isEnabledFor(logging.DEBUG):
with subprocess.Popen(command, cwd=config.workdir) as p:
p.wait()
if p.returncode != 0:
raise RunnerError(f"[PyPiRunner] Failed to build {package}")
else:
with subprocess.Popen(command, cwd=config.workdir, stdout=subprocess.PIPE) as p:
p.wait()
if p.returncode != 0:
raise RunnerError(f"[PyPiRunner] Failed to build {package}")
print(f"[PyPiRunner] Package {package} built")
def find_unuploaded(self, config, file_list, pkg_name):
versions = self.__versions(config, pkg_name)
unuploaded = []
for file in file_list:
# flake8: noqa W605
re_groups = re.findall("(\d*\.\d*\.\d*)", file)
if len(re_groups) < 1:
raise RunnerError(f"Unable to determine version of file {file}")
file_version = re_groups[0]
if file_version not in versions:
unuploaded.append(file)
else:
print(f"[PyPiRunner] File already uploaded: {os.path.basename(file)}")
print(f"[PyPiRunner] Packages to publish: {', '.join(unuploaded) if len(unuploaded) > 1 else 'None'}")
return unuploaded
def upload_command(self, config, package, _command, to_upload):
unregistered = False
command = _command + to_upload
with subprocess.Popen(command, cwd=config.workdir, stdout=subprocess.PIPE, stderr=subprocess.PIPE) as p:
p.wait()
if p.returncode != 0:
for line in map(lambda x: x.decode('utf8').strip(), p.stderr):
if "405 Method Not Allowed" in line:
unregistered = True
if not unregistered:
print("STDOUT:")
sys.stdout.buffer.write(p.stdout.read())
print("STDERR:")
sys.stdout.buffer.write(p.stderr.read())
raise RunnerError(f"[PyPiRunner] Failed to upload {package} ({p.returncode})")
if unregistered:
print("[PyPiRunner] Registering package")
register_command = [sys.executable, "-m", "twine", "register", "--verbose", "--non-interactive"]
if config.repo_uri is not None:
register_command.append("--repository-url")
register_command.append(config.repo_uri)
if config.repo_user is not None and config.repo_pass is not None:
register_command.append("-u")
register_command.append(config.repo_user)
register_command.append("-p")
register_command.append(config.repo_pass)
register_command.append(to_upload[0])
with subprocess.Popen(register_command, cwd=config.workdir, stdout=subprocess.PIPE, stderr=subprocess.PIPE) as p:
p.wait()
if p.returncode != 0:
print("STDOUT:")
sys.stdout.buffer.write(p.stdout.read())
print("STDERR:")
sys.stdout.buffer.write(p.stderr.read())
raise RunnerError(f"[PyPiRunner] Failed to register {package} ({p.returncode})")
self.upload_command(config, package, _command, to_upload)
def upload(self, config, package, current_version):
print(f"[PyPiRunner] Uploading {package}")
PackageManager.getInstance().ensure("twine")
command = [sys.executable, "-m", "twine", "upload", "--verbose", "--non-interactive"]
if config.repo_uri is not None:
command.append("--repository-url")
command.append(config.repo_uri)
if config.repo_user is not None and config.repo_pass is not None:
command.append("-u")
command.append(config.repo_user)
command.append("-p")
command.append(config.repo_pass)
else:
raise RunnerError("[PyPiRunner] Can't upload without credentials!")
dist_path = os.path.abspath(os.path.join(config.workdir, package, "dist"))
_files = glob(os.path.join(dist_path, "*"), config.workdir)
files = []
for file in _files:
if current_version in os.path.basename(file):
files.append(file)
print(f"[PyPiRunner] Found: {file}")
else:
logging.info(f"[PyPiRunner] Dropped: {file} doesn't match current version: {current_version}")
to_upload = self.find_unuploaded(config, files, package)
if len(to_upload) == 0:
return
#command += to_upload
self.upload_command(config, package, command, to_upload)
print(f"[PyPiRunner] Uploaded {package}")
def package_version(self, config, package):
cfg_path = path.join(config.workdir, package, "setup.cfg")
with open(cfg_path) as f:
for line in f:
if line.startswith("version"):
re_groups = re.findall("(\d*\.\d*\.\d*)", line)
if len(re_groups) < 1:
raise RunnerError(f"Unable to determine version of package: |{line}|")
return re_groups[0]
def run(self, job_spec):
job_config = self.config.copy(job_spec)
for package in job_config.packages:
pkg_dir = path.join(job_config.workdir, package)
pkg_hash = md5_dir(pkg_dir, exclude_dirs=["pycache", "pytest_cache", "build", "dist", "egg-info"])
logging.debug(f"[PyPiRunner] {package} hash: {pkg_hash}")
pkg_version = self.package_version(job_config, package)
logging.debug(f"[PyPiRunner] {package} local version: {pkg_version}")
repo_versions = self.__versions(job_config, package)
logging.debug(f"[PyPiRunner] {package} remote version: {repo_versions}")
if pkg_version not in repo_versions:
print(f"[PyPiRunner] {package} not found in repo")
self.build(job_config, package)
self.metadata.set(package, pkg_version, pkg_hash)
else:
if pkg_hash != self.metadata.get(package, pkg_version):
self.build(job_config, package)
self.metadata.set(package, pkg_version, pkg_hash)
else:
print(f"[PyPiRunner] {package} Unchanged since last build")
if job_config.upload:
self.upload(job_config, package, pkg_version)
else:
print(f"[PyPiRunner] Upload disabled, skipping")

198
alice-ci/src/alice/runners/pythonrunner.py

@ -1,96 +1,102 @@
import subprocess
import os
import sys
import shlex
from exceptions import NonZeroRetcode, RunnerError, ConfigException
# same venv across all runs!
class PythonRunner():
def __init__(self, workdir, defaults) -> None:
self.workdir = workdir
self.virtual_dir = os.path.abspath(os.path.join(workdir, "venv"))
self.config = defaults
self.env_vars = os.environ.copy()
for env_var in defaults["env"]:
self.env_vars[env_var["name"]] = env_var["value"]
self.__init_venv()
def __init_venv(self):
if os.name == "nt": # Windows
self.vpython = os.path.join(self.virtual_dir, "Scripts", "python.exe")
else: # Linux & Mac
self.vpython = os.path.join(self.virtual_dir, "bin", "python3")
if not os.path.exists(self.vpython):
with subprocess.Popen([sys.executable, "-m", "virtualenv", self.virtual_dir],
stdout=subprocess.PIPE, stderr=subprocess.PIPE) as p:
p.wait()
if p.returncode != 0:
sys.stdout.buffer.write(p.stderr.read())
raise RunnerError("PythonRunner: Could not create virtualenv")
else:
print(f"PythonRunner: Virtualenv initialized at {self.virtual_dir}")
else:
print(f"PythonRunner: Found virtualenv at {self.virtual_dir}")
# Stores common defaults for all jobs - all types!
# Also - dependency install by config is only allowed in this step
def update_config(self, config):
if "dependencies" in config:
for dependency in config["dependencies"]:
# TODO: Check what happens with fixed version
with subprocess.Popen([self.vpython, "-m", "pip", "install", dependency, "--upgrade"], stdout=subprocess.PIPE, stderr=subprocess.PIPE) as p:
p.wait()
if p.returncode != 0:
sys.stdout.buffer.write(p.stderr.read())
raise(RunnerError(f"PythonRunner: Could not install dependency: {dependency} ({p.returncode})"))
for env_var in config["env"]:
self.env_vars[env_var["name"]] = env_var["value"]
if "workdir" in config and config["workdir"] is not None:
self.workdir = os.path.join(self.workdir, config["workdir"])
def __ghetto_glob(self, command):
new_command = []
for item in command:
if "*" in item:
dir = os.path.abspath(os.path.dirname(item))
base_name = os.path.basename(item)
if os.path.isdir(dir):
item_parts = base_name.split("*")
print(item_parts)
for file in os.listdir(dir):
# TODO: Fix ordering! A*B = B*A = AB*
if item_parts[0] in file and item_parts[1] in file:
new_command.append(os.path.join(dir, file))
else:
new_command.append(item)
return new_command
# Executes the given job in the one and only venv
# parameter shall be the raw jobscpec
def run(self, job_spec):
if "workdir" in job_spec:
pwd = os.path.abspath(os.path.join(self.workdir, job_spec["workdir"]))
else:
pwd = self.workdir
run_env = self.env_vars.copy()
if "env" in job_spec:
for env_var in job_spec["env"]:
run_env[env_var["name"]] = env_var["value"]
if "commands" in job_spec:
commands = job_spec["commands"]
for command in commands:
# TODO: only split if command is not an array
run_command = self.__ghetto_glob(shlex.split(command))
if os.path.isdir(pwd):
with subprocess.Popen([self.vpython] + run_command, cwd=pwd, env=run_env) as p:
p.wait()
if p.returncode != 0:
raise NonZeroRetcode(f"Command {command} returned code {p.returncode}")
else:
raise RunnerError(f"PythonRunner: Invalid path for shell command: {pwd}")
else:
raise ConfigException(f"PythonRunner: No commands specified in step {job_spec['name']}")
import logging
import subprocess
import os
import sys
import shlex
from ..exceptions import NonZeroRetcode, RunnerError, ConfigException
from .pyutils import PackageManager, glob_command, grab_from
# TODO: Handle config like PyPiConfig
class PythonRunner:
def __init__(self, config) -> None:
logging.info("[PythonRunner] Initializing")
self.workdir = config["workdir"]
self.virtual_dir = os.path.abspath(os.path.join(self.workdir, "venv"))
self.config = config
PackageManager.getInstance().ensure("virtualenv")
self.__init_venv()
# TODO: Detect if the prev venv is the same OS type
def __init_venv(self):
if os.name == "nt": # Windows
self.vpython = os.path.join(self.virtual_dir, "Scripts", "python.exe")
else: # Linux & Mac
self.vpython = os.path.join(self.virtual_dir, "bin", "python")
if not os.path.exists(self.vpython):
logging.debug(f"[PythonRunner] Venv not found at {self.vpython}")
logging.info("[PythonRunner] Initializing venv")
output = []
with subprocess.Popen([sys.executable, "-m", "virtualenv", self.virtual_dir],
stdout=subprocess.PIPE) as p:
p.wait()
for line in p.stdout:
output.append(line.decode('utf8').strip())
if p.returncode != 0:
print("\n".join(output))
raise RunnerError("[PythonRunner] Could not create virtualenv")
else:
logging.info(f"[PythonRunner] Virtualenv initialized at {self.virtual_dir}")
else:
logging.info(f"[PythonRunner] Found virtualenv at {self.virtual_dir}")
dependencies = self.config.get("dependencies", [])
if len(dependencies) > 0:
logging.info(f"[PythonRunner] Ensuring dependencies: {', '.join(dependencies)}")
command = [self.vpython, "-m", "pip", "install"] + dependencies
if logging.root.isEnabledFor(logging.DEBUG):
with subprocess.Popen(command) as p:
p.wait()
if p.returncode != 0:
raise(RunnerError(f"[PythonRunner] Could not install dependencies: {dependencies} ({p.returncode})"))
else:
output = []
with subprocess.Popen(command, stdout=subprocess.PIPE) as p:
for line in p.stdout:
output.append(line.decode('utf8').strip())
p.wait()
if p.returncode != 0:
print("\n".join(output))
raise(RunnerError(f"[PythonRunner] Could not install dependencies: {dependencies} ({p.returncode})"))
logging.info("[PythonRunner] Installation done")
# Executes the given job in the one and only venv
# parameter is the raw jobscpec
def run(self, job_spec):
if "workdir" in job_spec:
pwd = os.path.abspath(os.path.join(self.workdir, job_spec["workdir"]))
else:
pwd = self.workdir
run_env = {}
for k, v in self.config["env"].items():
if isinstance(v, str):
run_env[k] = v
else:
run_env[k] = grab_from(v)
if "env" in job_spec:
for env_var in job_spec["env"]:
if isinstance(env_var["value"], str):
run_env[env_var["name"]] = env_var["value"]
else:
run_env[env_var["name"]] = grab_from(env_var["value"])
if "commands" in job_spec:
commands = job_spec["commands"]
for command in commands:
logging.debug(f"[PythonRunner] Raw command: {command}")
# TODO: only split if command is not an array
if "*" in command:
run_command = glob_command(shlex.split(command), pwd)
else:
run_command = shlex.split(command)
logging.info(f"[PythonRunner] Command to execute: {run_command}")
logging.debug(f"[PythonRunner] Workdir: {pwd}")
if os.path.isdir(pwd):
with subprocess.Popen([self.vpython] + run_command, cwd=pwd, env=run_env) as p:
p.wait()
if p.returncode != 0:
raise NonZeroRetcode(f"Command {command} returned code {p.returncode}")
else:
raise RunnerError(f"[PythonRunner] Invalid path for shell command: {pwd}")
else:
raise ConfigException(f"[PythonRunner] No commands specified in step {job_spec['name']}")

154
alice-ci/src/alice/runners/pyutils.py

@ -0,0 +1,154 @@
import logging
import os
import subprocess
import sys
from pkg_resources import parse_version
import re
from ..exceptions import RunnerError, ConfigException
from ..config import ConfigHolder
class PackageManager:
__instance = None
@staticmethod
def getInstance():
""" Static access method. """
if PackageManager.__instance is None:
PackageManager()
return PackageManager.__instance
def __init__(self):
""" Virtually private constructor. """
if PackageManager.__instance is not None:
raise Exception("This class is a singleton!")
else:
PackageManager.__instance = self
self.package_list = self.__get_packages()
def __get_packages(self):
packages = {}
with subprocess.Popen([sys.executable, "-m", "pip", "freeze"],
stdout=subprocess.PIPE, stderr=subprocess.PIPE) as p:
p.wait()
installed = list(map(lambda x: x.decode("UTF-8").split("=="), filter(lambda x: b'==' in x, p.stdout.read().splitlines())))
for name, version in installed:
packages[name] = parse_version(version)
logging.debug(f"[PackageManager] Picked up packages: {packages}")
return packages
def ensure_more(self, package_list, executable=sys.executable):
to_install = list(filter(lambda x: not self.__has_package(x), package_list))
if len(to_install) > 0:
command = [executable, "-m", "pip", "install"] + to_install
with subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE) as p:
p.wait()
if p.returncode != 0:
sys.stdout.buffer.write(p.stderr.read())
raise(RunnerError(f"[PackageManager] Could not install dependencies ({p.returncode})"))
self.package_list = self.__get_packages()
# Assumption: there are more hits in the long run, than misses
def ensure(self, package_string, executable=sys.executable):
if not self.__has_package(package_string):
logging.info(f"[PackageManager] Installing {package_string}")
command = [executable, "-m", "pip", "install", package_string]
with subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE) as p:
p.wait()
if p.returncode != 0:
sys.stdout.buffer.write(p.stderr.read())
raise(RunnerError(f"[PackageManager] Could not install dependencies ({p.returncode})"))
self.package_list = self.__get_packages()
else:
logging.info(f"[PackageManager] {package_string} already installed")
def __has_package(self, package_string):
package_data = re.split("==|>|>=|<|<=", package_string)
# check in cache
if package_data[0] in self.package_list:
# check if version is needed
if len(package_data) == 2:
required_version = parse_version(package_data[1])
installed_version = self.package_list[package_data[0]]
comparator = package_string.replace(package_data[0], "").replace(package_data[1], "")
if comparator == "==":
return required_version == installed_version
elif comparator == ">":
return installed_version > required_version
elif comparator == ">=":
return installed_version >= required_version
elif comparator == "<":
return installed_version < required_version
elif comparator == "<=":
return installed_version <= required_version
else:
raise ConfigException(f"Illegal comparator found: {comparator}")
else:
return True
return False
def glob(item, workdir):
new_command = []
if "*" in item:
logging.debug(f"[Globbing] Found item: [{item}]")
dir = os.path.abspath(os.path.join(workdir, os.path.dirname(item)))
base_name = os.path.basename(item)
if os.path.isdir(dir):
item_parts = base_name.split("*")
for file in os.listdir(dir):
# TODO: Fix ordering! A*B = B*A = AB*
if item_parts[0] in file and item_parts[1] in file:
new_item = os.path.join(dir, file)
logging.debug(f"[Globbing] Substitute: {new_item}")
new_command.append(new_item)
else:
raise ConfigException(f"[Globbing] Dir not exists: {dir}")
return new_command
else:
return [item]
def glob_command(command, workdir):
logging.debug(f"[Globbing] Starting command: {' '.join(command)}")
new_command = []
for item in command:
new_command += glob(item, workdir)
return new_command
def grab_from(target):
if "from_env" in target:
try:
return os.environ[target["from_env"]]
except KeyError:
raise ConfigException(f"Env var unset: {target['from_env']}")
elif "from_cfg" in target:
value = ConfigHolder.getInstance().get(target["from_cfg"])
if len(value) == 0:
value = None
return value
else:
raise ConfigException(f"Unsupported grabber: {target}")
def gen_dict(list_of_dicts):
"""
Generates a dictionary from a list of dictionaries composed of
'name' and 'value' keys.
[{'name': 'a', 'value': 'b'}] => {'a': 'b'}
"""
return_dict = {}
for _dict in list_of_dicts:
try:
if isinstance(_dict["value"], str):
return_dict[_dict["name"]] = _dict["value"]
else:
return_dict[_dict["name"]] = grab_from(_dict["value"])
except KeyError:
raise ConfigException(f"Invalid dict item: {_dict}")
return return_dict

78
alice-ci/src/alice/utils.py

@ -1,78 +0,0 @@
import yaml
from runners.pythonrunner import PythonRunner
from exceptions import NonZeroRetcode, ConfigException
class DummyRunner():
def __init__(self, type) -> None:
self.type = type
def run(self, command, workdir=None, env=None):
raise Exception(f"Invalid runner type in config: {self.type}")
class Job():
def __init__(self, type, repoDir, vpython, workspace, env={}) -> None:
self.runner = self.__get_runner(type, repoDir, vpython)
self.commands = []
self.workspace = workspace
self.env = env
def __get_runner(self, type, repoDir, vpython):
if type == "python":
return PythonRunner(repoDir, vpython)
else:
return DummyRunner(type)
def run_commands(self, _env={}):
try:
if self.env is None:
env = _env.copy()
else:
env = self.env.copy()
env.update(_env)
for command in self.commands:
self.runner.run(command, self.workspace, env)
except NonZeroRetcode as n:
print(n)
exit(1)
class ConfigParser:
def __init__(self, file_path, factory) -> None:
with open(file_path) as f:
self.config = yaml.safe_load(f)
self.factory = factory
if "runners" in self.config:
if "global" in self.config["runners"]:
self.factory.set_globals(self.__gen_globals())
self.factory.update_runners(self.config["runners"])
self.jobs = self.__get_jobs()
# Initialize env, workdir if not present
def __gen_globals(self):
globals = self.config["runners"]["global"]
if "env" not in globals:
globals["env"] = []
if "workdir" not in globals:
globals["workdir"] = None
return globals
def __get_jobs(self):
if "jobs" in self.config:
jobs = {}
for job_spec in self.config["jobs"]:
name = job_spec["name"]
if name in jobs:
raise ConfigException(f"Job with name {name} already exists!")
jobs[name] = job_spec
return jobs
else:
raise ConfigException("No jobs defined in config")
def execute_job(self, job_name):
if job_name in self.jobs:
# Pass the job_spec to a runner
runner = self.factory.get_runner(self.jobs[job_name]["type"])
runner.run(self.jobs[job_name])

127
ci-examples/full.yaml

@ -1,29 +1,98 @@
runners:
global:
env:
- name: A
value: A
- name: B
value: B
- name: C
value: C
workdir: packages
python:
env:
- name: A
value: D
dependencies:
- flake8
- build
jobs:
- name: env
type: python
env:
- name: B
value: E
commands:
- "-c \"import os; print(os.environ)\""
- name: lint
workdir: alice-ci
commands:
- "-m flake8 --ignore E501"
runners:
global:
env:
- name: A
value: A
- name: B
value: B
- name: C
value: C
workdir: .
python:
env:
- name: A
value: D
dependencies:
- flake8
- build
docker:
credentials:
username: D
password: D
jobs:
- name: env
type: python
changes:
branch: origin/master
paths:
- "docs"
- "alice-ci"
env:
- name: B
value: E
commands:
- "-c \"from os import environ; assert environ['A'] == 'D'; assert environ['B'] == 'E'; assert environ['C'] == 'C'; print('Assertions passed')\""
- name: lint
type: python
workdir: alice-ci/src
commands:
- "-m flake8 --ignore E501"
- name: pkg
type: pypi
workdir: .
upload: false
fail_if_exists: false # TODO: currently unused
repo:
uri: example.com
username:
from_env: PYPIUSER
password:
from_env: PYPIPASS
packages:
- alice-ci
- name: "image"
type: docker
credentials:
username: A
#password: B
image:
build:
dir: ci-examples/images/hello
#dockerfile: ci-examples/images/hello/Dockerfile
dockerfile: Dockerfile
name: "sssss"
args:
- name: CIPASS
value: NONE
#pull:
#name: python:latest
#credentials:
#username: PASS
#password: WORD
env:
- name: VAR
value: CHAR
commands:
- which python3
- /usr/bin/python3 --version
- date
- env
tag:
publish: false
name: repo.example.com/test/na
credentials:
username: B
password: B
- name: pypi_init
type: pypirepo
enabled: true
port: 8888
container_name: pypiserver
pipelines:
default:
- lint
- env
- pkg

9
ci-examples/images/hello/Dockerfile

@ -0,0 +1,9 @@
FROM ubuntu:latest
RUN apt update && apt install -y python3
ADD hello.py /opt/hello.py
#ENTRYPOINT [ "/bin/sh", "-c" ]
#CMD ["/usr/local/python/bin/python3", "/opt/hello.py"]

2
ci-examples/images/hello/hello.py

@ -0,0 +1,2 @@
if __name__ == "__main__":
print("Hi Mom!")

18
ci-examples/python1.yaml

@ -0,0 +1,18 @@
runners:
python:
dependencies:
- flake8
- build
- twine
jobs:
- name: selfcheck
type: python
workdir: ci
commands:
- "-m flake8 --ignore E501 --exclude venv"
- name: lint
type: python
workdir: alice-ci/src
commands:
- "-m flake8 --ignore E501"

46
docs/examples.md

@ -1,24 +1,24 @@
# alice-ci.yaml examples
## Python lint
Installes flake8 package in a virtual elvironment, then lints the contents of the packages directory in the current working dir.
```
runners:
python:
dependencies:
- name: flake8
jobs:
- name: lint
type: python
workdir: packages
commands:
- "-m flake8"
```
To run this job:
```
pythom3 -m alice lint
# alice-ci.yaml examples
## Python lint
Installes flake8 package in a virtual elvironment, then lints the contents of the packages directory in the current working dir.
```
runners:
python:
dependencies:
- name: flake8
jobs:
- name: lint
type: python
workdir: packages
commands:
- "-m flake8"
```
To run this job:
```
pythom3 -m alice lint
```

58
docs/runners.md

@ -1,8 +1,50 @@
# Runners
Runners are responsible to execute a list of commands in a set environment defined in the CI yaml file.
## List of runners
* Python - executes python commands in a virtual environment
* Docker - executes each job in a separate Docker container - unimplemented
# Runners
Runners are responsible to execute a list of commands in a set environment defined in the CI yaml file.
## List of runners
* Python - executes python commands in a virtual environment
* Docker - executes each job in a separate Docker container - unimplemented
## Import schema
What you need to do to make Alice recognise and import your custom Runners
TODO
## Runner API
Each runner has to support the following functions:
### __init__(params, config)
* params: dict of runtime variables for the program itself.
* config: Runner config data, aplies to all jobs
#### Params
Currently the only param used is the dict is "verbose", whichis a boolean. The intended purpose is to show debug output if set to True.
#### config
Dict. Has two fix keys, `env` and `workdir`. Env is the environment variables of the host, expanded by CLI parameters, expanded by global config values from yaml. A key defined in the yaml overwrites the value copied from the host. Workdir is similar, can be assigned at CI yaml level as global, but runners may overwrite it.
Order:
By default: os.cwd()
if overwritten in global
------------------------------- Below this level is the runner's responsibility
if owerwritten in runner config
if overwritten in job
Runner shall receive the current working directory, unless stated otherwise in global config.
The expected behaviour of "overwriting" the workdir is adding the new directory name to the existing path. For example:
* cwd = /root
* global defines workdir as "projects"
* runner config defines workdir as "python"
* job defines workdir as "alice"
In the case above, the actual working directory of the running job shall be `/root/projects/python/alice`.
### run(job_spec)
This function executes one job attributed ith the type of the runner called. As the only hard requirement for Alice is the "type" field in a job (or the optional "changes"), everything else is handled by the runner.

26
docs/runners/docker.md

@ -0,0 +1,26 @@
# Schema
```
name: ""
type: docker
credentials: - global ...ish
username
password
image: - to use, pull, run
build:
dir:
dockerfile:
name: - defaults to step name
args:
- name:
- value:
pull: - pulls, current working image - mutually exclusive with build
name:
credentials: - optional
command: - overwrite, not append
- ...
tag:
publish: true
name: - published name with repo and everything
credentials:
```

78
docs/syntax.md

@ -1,39 +1,39 @@
# alice-ci.yaml
This yaml file defines the job steps executed by Alice. The jobs are called by names for each passed parameter on CLI. For example the following command searches for a job called lint defined in the `alice-ci.yaml` file in the current working directory, then runs it.
```
pythom3 -m alice lint
```
[Example configs](examples.md)
## runners
Contains global configuration for various runners. Currently the only supported runner is `python`.
### Python
#### Dependencies
List of dependencies installed in the virtual environment. Each dependency has a `name` and an `import_name`, as Alice checks the availability of each package by trying to import `import_name`, and if it fails, calls pip to install `name`.
## jobs
List of jobs. Each job has a mandatory name, type and a list of commands, optional parameter is workdir.
### name
Mandatory value, string. Has to be unique in the current file.
### type
Job type, selects the runner executing the commands. Currently the only supported type is `python`.
### comands
List of strings, each executed one by one from top to bottom in the current context.
### workdir
Optional, defines Working directory relative to PWD. The default working directory is the current directory.
# alice-ci.yaml
This yaml file defines the job steps executed by Alice. The jobs are called by names for each passed parameter on CLI. For example the following command searches for a job called lint defined in the `alice-ci.yaml` file in the current working directory, then runs it.
```
pythom3 -m alice lint
```
[Example configs](examples.md)
## runners
Contains global configuration for various runners. Currently the only supported runner is `python`.
### Python
#### Dependencies
List of dependencies installed in the virtual environment. Each dependency has a `name` and an `import_name`, as Alice checks the availability of each package by trying to import `import_name`, and if it fails, calls pip to install `name`.
## jobs
List of jobs. Each job has a mandatory name, type and a list of commands, optional parameter is workdir.
### name
Mandatory value, string. Has to be unique in the current file.
### type
Job type, selects the runner executing the commands. Currently the only supported type is `python`.
### comands
List of strings, each executed one by one from top to bottom in the current context.
### workdir
Optional, defines Working directory relative to PWD. The default working directory is the current directory.

Loading…
Cancel
Save