mirror of
https://github.com/actions/runner-container-hooks.git
synced 2025-12-17 10:16:44 +00:00
Compare commits
73 Commits
nikola-jok
...
v0.2.0
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
17d2b3b850 | ||
|
|
ea011028f5 | ||
|
|
eaae191ebb | ||
|
|
418d484160 | ||
|
|
ce3c55d086 | ||
|
|
d988d965c5 | ||
|
|
23cc6dda6f | ||
|
|
8986035ca8 | ||
|
|
e975289683 | ||
|
|
a555151eef | ||
|
|
16eb238caa | ||
|
|
8e06496e34 | ||
|
|
e2033b29c7 | ||
|
|
eb47baaf5e | ||
|
|
20c19dae27 | ||
|
|
4307828719 | ||
|
|
5c6995dba1 | ||
|
|
bb1a033ed7 | ||
|
|
898063bddd | ||
|
|
266b8edb99 | ||
|
|
47cbf5a0d7 | ||
|
|
de4553f25a | ||
|
|
8ea57170d8 | ||
|
|
643bf36fd8 | ||
|
|
de59bd8716 | ||
|
|
d3ec1c0040 | ||
|
|
3e04b45585 | ||
|
|
2b386f7cbd | ||
|
|
bf362ba0dd | ||
|
|
7ae8942b3d | ||
|
|
347e68d3c9 | ||
|
|
7c4e0f8d51 | ||
|
|
cd310988c9 | ||
|
|
1bfc52f466 | ||
|
|
2aa6f9d9c8 | ||
|
|
3d0ca83d2d | ||
|
|
5daaae120b | ||
|
|
df448fbbb0 | ||
|
|
ee2554e2c0 | ||
|
|
f764d18c4c | ||
|
|
55761eab39 | ||
|
|
51bd8b62a4 | ||
|
|
150bc0503a | ||
|
|
9ce39e5a60 | ||
|
|
8351f842bd | ||
|
|
bf3707d7e0 | ||
|
|
88b7b19db7 | ||
|
|
dd5dfb3e48 | ||
|
|
84a57de2e3 | ||
|
|
fa680b2073 | ||
|
|
02f0b322a0 | ||
|
|
ecb9376000 | ||
|
|
cc90cd2361 | ||
|
|
152c4e1cc8 | ||
|
|
d0e094649e | ||
|
|
3ba45d3d7e | ||
|
|
58ebf56ad3 | ||
|
|
e928fa3252 | ||
|
|
ddf09ad7bd | ||
|
|
689a74e352 | ||
|
|
55c9198ada | ||
|
|
bd7e053180 | ||
|
|
0ebccbd8c6 | ||
|
|
ec8131abb7 | ||
|
|
7010d21bff | ||
|
|
c65ec28bbb | ||
|
|
c2f9b10f4d | ||
|
|
1e49f4ba5b | ||
|
|
171956673c | ||
|
|
3ab4ae20f9 | ||
|
|
b0cf60b678 | ||
|
|
5ec2edbe11 | ||
|
|
4b7efe88ef |
7
.github/workflows/build.yaml
vendored
7
.github/workflows/build.yaml
vendored
@@ -11,6 +11,11 @@ jobs:
|
|||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v3
|
||||||
|
- run: sed -i "s|{{PATHTOREPO}}|$(pwd)|" packages/k8s/tests/test-kind.yaml
|
||||||
|
name: Setup kind cluster yaml config
|
||||||
|
- uses: helm/kind-action@v1.2.0
|
||||||
|
with:
|
||||||
|
config: packages/k8s/tests/test-kind.yaml
|
||||||
- run: npm install
|
- run: npm install
|
||||||
name: Install dependencies
|
name: Install dependencies
|
||||||
- run: npm run bootstrap
|
- run: npm run bootstrap
|
||||||
@@ -21,6 +26,6 @@ jobs:
|
|||||||
- name: Check linter
|
- name: Check linter
|
||||||
run: |
|
run: |
|
||||||
npm run lint
|
npm run lint
|
||||||
git diff --exit-code
|
git diff --exit-code -- ':!packages/k8s/tests/test-kind.yaml'
|
||||||
- name: Run tests
|
- name: Run tests
|
||||||
run: npm run test
|
run: npm run test
|
||||||
3
.gitignore
vendored
3
.gitignore
vendored
@@ -1,4 +1,5 @@
|
|||||||
node_modules/
|
node_modules/
|
||||||
lib/
|
lib/
|
||||||
dist/
|
dist/
|
||||||
**/tests/_temp/**
|
**/tests/_temp/**
|
||||||
|
packages/k8s/tests/test-kind.yaml
|
||||||
@@ -1 +1 @@
|
|||||||
* @actions/actions-runtime
|
* @actions/actions-runtime @actions/runner-akvelon
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ You'll need a runner compatible with hooks, a repository with container workflow
|
|||||||
- You'll need a runner compatible with hooks, a repository with container workflows to which you can register the runner and the hooks from this repository.
|
- You'll need a runner compatible with hooks, a repository with container workflows to which you can register the runner and the hooks from this repository.
|
||||||
- See [the runner contributing.md](../../github/CONTRIBUTING.MD) for how to get started with runner development.
|
- See [the runner contributing.md](../../github/CONTRIBUTING.MD) for how to get started with runner development.
|
||||||
- Build your hook using `npm run build`
|
- Build your hook using `npm run build`
|
||||||
- Enable the hooks by setting `ACTIONS_RUNNER_CONTAINER_HOOK=./packages/{libraryname}/dist/index.js` file generated by [ncc](https://github.com/vercel/ncc)
|
- Enable the hooks by setting `ACTIONS_RUNNER_CONTAINER_HOOKS=./packages/{libraryname}/dist/index.js` file generated by [ncc](https://github.com/vercel/ncc)
|
||||||
- Configure your self hosted runner against the a repository you have admin access
|
- Configure your self hosted runner against the a repository you have admin access
|
||||||
- Run a workflow with a container job, for example
|
- Run a workflow with a container job, for example
|
||||||
```
|
```
|
||||||
|
|||||||
3
examples/example-script.sh
Normal file
3
examples/example-script.sh
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
echo "Hello World"
|
||||||
@@ -5,7 +5,7 @@
|
|||||||
"args": {
|
"args": {
|
||||||
"container": {
|
"container": {
|
||||||
"image": "node:14.16",
|
"image": "node:14.16",
|
||||||
"workingDirectory": "/__w/thboop-test2/thboop-test2",
|
"workingDirectory": "/__w/repo/repo",
|
||||||
"createOptions": "--cpus 1",
|
"createOptions": "--cpus 1",
|
||||||
"environmentVariables": {
|
"environmentVariables": {
|
||||||
"NODE_ENV": "development"
|
"NODE_ENV": "development"
|
||||||
@@ -24,37 +24,37 @@
|
|||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work",
|
||||||
"targetVolumePath": "/__w",
|
"targetVolumePath": "/__w",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/externals",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/externals",
|
||||||
"targetVolumePath": "/__e",
|
"targetVolumePath": "/__e",
|
||||||
"readOnly": true
|
"readOnly": true
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_temp",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_temp",
|
||||||
"targetVolumePath": "/__w/_temp",
|
"targetVolumePath": "/__w/_temp",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_actions",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_actions",
|
||||||
"targetVolumePath": "/__w/_actions",
|
"targetVolumePath": "/__w/_actions",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_tool",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_tool",
|
||||||
"targetVolumePath": "/__w/_tool",
|
"targetVolumePath": "/__w/_tool",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_temp/_github_home",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_temp/_github_home",
|
||||||
"targetVolumePath": "/github/home",
|
"targetVolumePath": "/github/home",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_temp/_github_workflow",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_temp/_github_workflow",
|
||||||
"targetVolumePath": "/github/workflow",
|
"targetVolumePath": "/github/workflow",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -12,11 +12,11 @@
|
|||||||
"image": "node:14.16",
|
"image": "node:14.16",
|
||||||
"dockerfile": null,
|
"dockerfile": null,
|
||||||
"entryPointArgs": [
|
"entryPointArgs": [
|
||||||
"-c",
|
"-e",
|
||||||
"echo \"hello world2\""
|
"example-script.sh"
|
||||||
],
|
],
|
||||||
"entryPoint": "bash",
|
"entryPoint": "bash",
|
||||||
"workingDirectory": "/__w/thboop-test2/thboop-test2",
|
"workingDirectory": "/__w/repo/repo",
|
||||||
"createOptions": "--cpus 1",
|
"createOptions": "--cpus 1",
|
||||||
"environmentVariables": {
|
"environmentVariables": {
|
||||||
"NODE_ENV": "development"
|
"NODE_ENV": "development"
|
||||||
@@ -34,27 +34,27 @@
|
|||||||
],
|
],
|
||||||
"systemMountVolumes": [
|
"systemMountVolumes": [
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work",
|
||||||
"targetVolumePath": "/__w",
|
"targetVolumePath": "/__w",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/externals",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/externals",
|
||||||
"targetVolumePath": "/__e",
|
"targetVolumePath": "/__e",
|
||||||
"readOnly": true
|
"readOnly": true
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_temp",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_temp",
|
||||||
"targetVolumePath": "/__w/_temp",
|
"targetVolumePath": "/__w/_temp",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_actions",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_actions",
|
||||||
"targetVolumePath": "/__w/_actions",
|
"targetVolumePath": "/__w/_actions",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_tool",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_tool",
|
||||||
"targetVolumePath": "/__w/_tool",
|
"targetVolumePath": "/__w/_tool",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
@@ -64,7 +64,7 @@
|
|||||||
"readOnly": false
|
"readOnly": false
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"sourceVolumePath": "//Users/thomas/git/runner/_layout/_work/_temp/_github_workflow",
|
"sourceVolumePath": "/Users/thomas/git/runner/_layout/_work/_temp/_github_workflow",
|
||||||
"targetVolumePath": "/github/workflow",
|
"targetVolumePath": "/github/workflow",
|
||||||
"readOnly": false
|
"readOnly": false
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -10,8 +10,8 @@
|
|||||||
},
|
},
|
||||||
"args": {
|
"args": {
|
||||||
"entryPointArgs": [
|
"entryPointArgs": [
|
||||||
"-c",
|
"-e",
|
||||||
"echo \"hello world\""
|
"example-script.sh"
|
||||||
],
|
],
|
||||||
"entryPoint": "bash",
|
"entryPoint": "bash",
|
||||||
"environmentVariables": {
|
"environmentVariables": {
|
||||||
@@ -21,6 +21,6 @@
|
|||||||
"/foo/bar",
|
"/foo/bar",
|
||||||
"bar/foo"
|
"bar/foo"
|
||||||
],
|
],
|
||||||
"workingDirectory": "/__w/thboop-test2/thboop-test2"
|
"workingDirectory": "/__w/repo/repo"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
|||||||
{
|
{
|
||||||
"name": "hooks",
|
"name": "hooks",
|
||||||
"version": "0.1.0",
|
"version": "0.1.3",
|
||||||
"lockfileVersion": 2,
|
"lockfileVersion": 2,
|
||||||
"requires": true,
|
"requires": true,
|
||||||
"packages": {
|
"packages": {
|
||||||
"": {
|
"": {
|
||||||
"name": "hooks",
|
"name": "hooks",
|
||||||
"version": "0.1.0",
|
"version": "0.1.3",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@types/jest": "^27.5.1",
|
"@types/jest": "^27.5.1",
|
||||||
|
|||||||
@@ -1,13 +1,13 @@
|
|||||||
{
|
{
|
||||||
"name": "hooks",
|
"name": "hooks",
|
||||||
"version": "0.1.0",
|
"version": "0.2.0",
|
||||||
"description": "Three projects are included - k8s: a kubernetes hook implementation that spins up pods dynamically to run a job - docker: A hook implementation of the runner's docker implementation - A hook lib, which contains shared typescript definitions and utilities that the other packages consume",
|
"description": "Three projects are included - k8s: a kubernetes hook implementation that spins up pods dynamically to run a job - docker: A hook implementation of the runner's docker implementation - A hook lib, which contains shared typescript definitions and utilities that the other packages consume",
|
||||||
"main": "",
|
"main": "",
|
||||||
"directories": {
|
"directories": {
|
||||||
"doc": "docs"
|
"doc": "docs"
|
||||||
},
|
},
|
||||||
"scripts": {
|
"scripts": {
|
||||||
"test": "npm run test --prefix packages/docker",
|
"test": "npm run test --prefix packages/docker && npm run test --prefix packages/k8s",
|
||||||
"bootstrap": "npm install --prefix packages/hooklib && npm install --prefix packages/k8s && npm install --prefix packages/docker",
|
"bootstrap": "npm install --prefix packages/hooklib && npm install --prefix packages/k8s && npm install --prefix packages/docker",
|
||||||
"format": "prettier --write '**/*.ts'",
|
"format": "prettier --write '**/*.ts'",
|
||||||
"format-check": "prettier --check '**/*.ts'",
|
"format-check": "prettier --check '**/*.ts'",
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
jest.setTimeout(90000)
|
jest.setTimeout(500000)
|
||||||
|
|||||||
40
packages/docker/package-lock.json
generated
40
packages/docker/package-lock.json
generated
@@ -9,7 +9,7 @@
|
|||||||
"version": "0.1.0",
|
"version": "0.1.0",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": "^1.6.0",
|
"@actions/core": "^1.9.1",
|
||||||
"@actions/exec": "^1.1.1",
|
"@actions/exec": "^1.1.1",
|
||||||
"hooklib": "file:../hooklib",
|
"hooklib": "file:../hooklib",
|
||||||
"uuid": "^8.3.2"
|
"uuid": "^8.3.2"
|
||||||
@@ -30,7 +30,7 @@
|
|||||||
"version": "0.1.0",
|
"version": "0.1.0",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": "^1.6.0"
|
"@actions/core": "^1.9.1"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@types/node": "^17.0.23",
|
"@types/node": "^17.0.23",
|
||||||
@@ -43,11 +43,12 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@actions/core": {
|
"node_modules/@actions/core": {
|
||||||
"version": "1.6.0",
|
"version": "1.9.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.6.0.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.9.1.tgz",
|
||||||
"integrity": "sha512-NB1UAZomZlCV/LmJqkLhNTqtKfFXJZAUPcfl/zqG7EfsQdeUJtaWO98SGbuQ3pydJ3fHl2CvI/51OKYlCYYcaw==",
|
"integrity": "sha512-5ad+U2YGrmmiw6du20AQW5XuWo7UKN2052FjSV7MX+Wfjf8sCqcsZe62NfgHys4QI4/Y+vQvLKYL8jWtA1ZBTA==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/http-client": "^1.0.11"
|
"@actions/http-client": "^2.0.1",
|
||||||
|
"uuid": "^8.3.2"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@actions/exec": {
|
"node_modules/@actions/exec": {
|
||||||
@@ -59,11 +60,11 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@actions/http-client": {
|
"node_modules/@actions/http-client": {
|
||||||
"version": "1.0.11",
|
"version": "2.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-1.0.11.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-2.0.1.tgz",
|
||||||
"integrity": "sha512-VRYHGQV1rqnROJqdMvGUbY/Kn8vriQe/F9HR2AlYHzmKuM/p3kjNuXhmdBfcVgsvRWTz5C5XW5xvndZrVBuAYg==",
|
"integrity": "sha512-PIXiMVtz6VvyaRsGY268qvj57hXQEpsYogYOu2nrQhlf+XCGmZstmuZBbAybUl1nQGnvS1k1eEsQ69ZoD7xlSw==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"tunnel": "0.0.6"
|
"tunnel": "^0.0.6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@actions/io": {
|
"node_modules/@actions/io": {
|
||||||
@@ -5279,11 +5280,12 @@
|
|||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": {
|
"@actions/core": {
|
||||||
"version": "1.6.0",
|
"version": "1.9.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.6.0.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.9.1.tgz",
|
||||||
"integrity": "sha512-NB1UAZomZlCV/LmJqkLhNTqtKfFXJZAUPcfl/zqG7EfsQdeUJtaWO98SGbuQ3pydJ3fHl2CvI/51OKYlCYYcaw==",
|
"integrity": "sha512-5ad+U2YGrmmiw6du20AQW5XuWo7UKN2052FjSV7MX+Wfjf8sCqcsZe62NfgHys4QI4/Y+vQvLKYL8jWtA1ZBTA==",
|
||||||
"requires": {
|
"requires": {
|
||||||
"@actions/http-client": "^1.0.11"
|
"@actions/http-client": "^2.0.1",
|
||||||
|
"uuid": "^8.3.2"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"@actions/exec": {
|
"@actions/exec": {
|
||||||
@@ -5295,11 +5297,11 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"@actions/http-client": {
|
"@actions/http-client": {
|
||||||
"version": "1.0.11",
|
"version": "2.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-1.0.11.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-2.0.1.tgz",
|
||||||
"integrity": "sha512-VRYHGQV1rqnROJqdMvGUbY/Kn8vriQe/F9HR2AlYHzmKuM/p3kjNuXhmdBfcVgsvRWTz5C5XW5xvndZrVBuAYg==",
|
"integrity": "sha512-PIXiMVtz6VvyaRsGY268qvj57hXQEpsYogYOu2nrQhlf+XCGmZstmuZBbAybUl1nQGnvS1k1eEsQ69ZoD7xlSw==",
|
||||||
"requires": {
|
"requires": {
|
||||||
"tunnel": "0.0.6"
|
"tunnel": "^0.0.6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"@actions/io": {
|
"@actions/io": {
|
||||||
@@ -7376,7 +7378,7 @@
|
|||||||
"hooklib": {
|
"hooklib": {
|
||||||
"version": "file:../hooklib",
|
"version": "file:../hooklib",
|
||||||
"requires": {
|
"requires": {
|
||||||
"@actions/core": "^1.6.0",
|
"@actions/core": "^1.9.1",
|
||||||
"@types/node": "^17.0.23",
|
"@types/node": "^17.0.23",
|
||||||
"@typescript-eslint/parser": "^5.18.0",
|
"@typescript-eslint/parser": "^5.18.0",
|
||||||
"@zeit/ncc": "^0.22.3",
|
"@zeit/ncc": "^0.22.3",
|
||||||
|
|||||||
@@ -10,7 +10,7 @@
|
|||||||
"author": "",
|
"author": "",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": "^1.6.0",
|
"@actions/core": "^1.9.1",
|
||||||
"@actions/exec": "^1.1.1",
|
"@actions/exec": "^1.1.1",
|
||||||
"hooklib": "file:../hooklib",
|
"hooklib": "file:../hooklib",
|
||||||
"uuid": "^8.3.2"
|
"uuid": "^8.3.2"
|
||||||
|
|||||||
@@ -2,12 +2,11 @@ import * as core from '@actions/core'
|
|||||||
import * as fs from 'fs'
|
import * as fs from 'fs'
|
||||||
import {
|
import {
|
||||||
ContainerInfo,
|
ContainerInfo,
|
||||||
JobContainerInfo,
|
Registry,
|
||||||
RunContainerStepArgs,
|
RunContainerStepArgs,
|
||||||
ServiceContainerInfo,
|
ServiceContainerInfo
|
||||||
StepContainerInfo
|
|
||||||
} from 'hooklib/lib'
|
} from 'hooklib/lib'
|
||||||
import path from 'path'
|
import * as path from 'path'
|
||||||
import { env } from 'process'
|
import { env } from 'process'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
import { runDockerCommand, RunDockerCommandOptions } from '../utils'
|
import { runDockerCommand, RunDockerCommandOptions } from '../utils'
|
||||||
@@ -43,19 +42,15 @@ export async function createContainer(
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (args.environmentVariables) {
|
if (args.environmentVariables) {
|
||||||
for (const [key, value] of Object.entries(args.environmentVariables)) {
|
for (const [key] of Object.entries(args.environmentVariables)) {
|
||||||
dockerArgs.push('-e')
|
dockerArgs.push('-e')
|
||||||
if (!value) {
|
dockerArgs.push(key)
|
||||||
dockerArgs.push(`"${key}"`)
|
|
||||||
} else {
|
|
||||||
dockerArgs.push(`"${key}=${value}"`)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const mountVolumes = [
|
const mountVolumes = [
|
||||||
...(args.userMountVolumes || []),
|
...(args.userMountVolumes || []),
|
||||||
...((args as JobContainerInfo | StepContainerInfo).systemMountVolumes || [])
|
...(args.systemMountVolumes || [])
|
||||||
]
|
]
|
||||||
for (const mountVolume of mountVolumes) {
|
for (const mountVolume of mountVolumes) {
|
||||||
dockerArgs.push(
|
dockerArgs.push(
|
||||||
@@ -74,7 +69,9 @@ export async function createContainer(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const id = (await runDockerCommand(dockerArgs)).trim()
|
const id = (
|
||||||
|
await runDockerCommand(dockerArgs, { env: args.environmentVariables })
|
||||||
|
).trim()
|
||||||
if (!id) {
|
if (!id) {
|
||||||
throw new Error('Could not read id from docker command')
|
throw new Error('Could not read id from docker command')
|
||||||
}
|
}
|
||||||
@@ -146,17 +143,41 @@ export async function containerBuild(
|
|||||||
args: RunContainerStepArgs,
|
args: RunContainerStepArgs,
|
||||||
tag: string
|
tag: string
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const context = path.dirname(`${env.GITHUB_WORKSPACE}/${args.dockerfile}`)
|
if (!args.dockerfile) {
|
||||||
|
throw new Error("Container build expects 'args.dockerfile' to be set")
|
||||||
|
}
|
||||||
|
|
||||||
const dockerArgs: string[] = ['build']
|
const dockerArgs: string[] = ['build']
|
||||||
dockerArgs.push('-t', tag)
|
dockerArgs.push('-t', tag)
|
||||||
dockerArgs.push('-f', `${env.GITHUB_WORKSPACE}/${args.dockerfile}`)
|
dockerArgs.push('-f', args.dockerfile)
|
||||||
dockerArgs.push(context)
|
dockerArgs.push(getBuildContext(args.dockerfile))
|
||||||
// TODO: figure out build working directory
|
|
||||||
await runDockerCommand(dockerArgs, {
|
await runDockerCommand(dockerArgs, {
|
||||||
workingDir: args['buildWorkingDirectory']
|
workingDir: getWorkingDir(args.dockerfile)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function getBuildContext(dockerfilePath: string): string {
|
||||||
|
return path.dirname(dockerfilePath)
|
||||||
|
}
|
||||||
|
|
||||||
|
function getWorkingDir(dockerfilePath: string): string {
|
||||||
|
const workspace = env.GITHUB_WORKSPACE as string
|
||||||
|
let workingDir = workspace
|
||||||
|
if (!dockerfilePath?.includes(workspace)) {
|
||||||
|
// This is container action
|
||||||
|
const pathSplit = dockerfilePath.split('/')
|
||||||
|
const actionIndex = pathSplit?.findIndex(d => d === '_actions')
|
||||||
|
if (actionIndex) {
|
||||||
|
const actionSubdirectoryDepth = 3 // handle + repo + [branch | tag]
|
||||||
|
pathSplit.splice(actionIndex + actionSubdirectoryDepth + 1)
|
||||||
|
workingDir = pathSplit.join('/')
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return workingDir
|
||||||
|
}
|
||||||
|
|
||||||
export async function containerLogs(id: string): Promise<void> {
|
export async function containerLogs(id: string): Promise<void> {
|
||||||
const dockerArgs: string[] = ['logs']
|
const dockerArgs: string[] = ['logs']
|
||||||
dockerArgs.push('--details')
|
dockerArgs.push('--details')
|
||||||
@@ -171,6 +192,18 @@ export async function containerNetworkRemove(network: string): Promise<void> {
|
|||||||
await runDockerCommand(dockerArgs)
|
await runDockerCommand(dockerArgs)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export async function containerNetworkPrune(): Promise<void> {
|
||||||
|
const dockerArgs = [
|
||||||
|
'network',
|
||||||
|
'prune',
|
||||||
|
'--force',
|
||||||
|
'--filter',
|
||||||
|
`label=${getRunnerLabel()}`
|
||||||
|
]
|
||||||
|
|
||||||
|
await runDockerCommand(dockerArgs)
|
||||||
|
}
|
||||||
|
|
||||||
export async function containerPrune(): Promise<void> {
|
export async function containerPrune(): Promise<void> {
|
||||||
const dockerPSArgs: string[] = [
|
const dockerPSArgs: string[] = [
|
||||||
'ps',
|
'ps',
|
||||||
@@ -238,22 +271,36 @@ export async function healthCheck({
|
|||||||
export async function containerPorts(id: string): Promise<string[]> {
|
export async function containerPorts(id: string): Promise<string[]> {
|
||||||
const dockerArgs = ['port', id]
|
const dockerArgs = ['port', id]
|
||||||
const portMappings = (await runDockerCommand(dockerArgs)).trim()
|
const portMappings = (await runDockerCommand(dockerArgs)).trim()
|
||||||
return portMappings.split('\n')
|
return portMappings.split('\n').filter(p => !!p)
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function registryLogin(args): Promise<string> {
|
export async function getContainerEnvValue(
|
||||||
if (!args.registry) {
|
id: string,
|
||||||
|
name: string
|
||||||
|
): Promise<string> {
|
||||||
|
const dockerArgs = [
|
||||||
|
'inspect',
|
||||||
|
`--format='{{range $index, $value := .Config.Env}}{{if eq (index (split $value "=") 0) "${name}"}}{{index (split $value "=") 1}}{{end}}{{end}}'`,
|
||||||
|
id
|
||||||
|
]
|
||||||
|
const value = (await runDockerCommand(dockerArgs)).trim()
|
||||||
|
const lines = value.split('\n')
|
||||||
|
return lines.length ? lines[0].replace(/^'/, '').replace(/'$/, '') : ''
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function registryLogin(registry?: Registry): Promise<string> {
|
||||||
|
if (!registry) {
|
||||||
return ''
|
return ''
|
||||||
}
|
}
|
||||||
const credentials = {
|
const credentials = {
|
||||||
username: args.registry.username,
|
username: registry.username,
|
||||||
password: args.registry.password
|
password: registry.password
|
||||||
}
|
}
|
||||||
|
|
||||||
const configLocation = `${env.RUNNER_TEMP}/.docker_${uuidv4()}`
|
const configLocation = `${env.RUNNER_TEMP}/.docker_${uuidv4()}`
|
||||||
fs.mkdirSync(configLocation)
|
fs.mkdirSync(configLocation)
|
||||||
try {
|
try {
|
||||||
await dockerLogin(configLocation, args.registry.serverUrl, credentials)
|
await dockerLogin(configLocation, registry.serverUrl, credentials)
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
fs.rmdirSync(configLocation, { recursive: true })
|
fs.rmdirSync(configLocation, { recursive: true })
|
||||||
throw error
|
throw error
|
||||||
@@ -271,7 +318,7 @@ export async function registryLogout(configLocation: string): Promise<void> {
|
|||||||
async function dockerLogin(
|
async function dockerLogin(
|
||||||
configLocation: string,
|
configLocation: string,
|
||||||
registry: string,
|
registry: string,
|
||||||
credentials: { username: string; password: string }
|
credentials: { username?: string; password?: string }
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const credentialsArgs =
|
const credentialsArgs =
|
||||||
credentials.username && credentials.password
|
credentials.username && credentials.password
|
||||||
@@ -307,30 +354,36 @@ export async function containerExecStep(
|
|||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const dockerArgs: string[] = ['exec', '-i']
|
const dockerArgs: string[] = ['exec', '-i']
|
||||||
dockerArgs.push(`--workdir=${args.workingDirectory}`)
|
dockerArgs.push(`--workdir=${args.workingDirectory}`)
|
||||||
for (const [key, value] of Object.entries(args['environmentVariables'])) {
|
for (const [key] of Object.entries(args['environmentVariables'])) {
|
||||||
dockerArgs.push('-e')
|
dockerArgs.push('-e')
|
||||||
if (!value) {
|
dockerArgs.push(key)
|
||||||
dockerArgs.push(`"${key}"`)
|
|
||||||
} else {
|
|
||||||
dockerArgs.push(`"${key}=${value}"`)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Todo figure out prepend path and update it here
|
if (args.prependPath?.length) {
|
||||||
// (we need to pass path in as -e Path={fullpath}) where {fullpath is the prepend path added to the current containers path}
|
// TODO: remove compatibility with typeof prependPath === 'string' as we bump to next major version, the hooks will lose PrependPath compat with runners 2.293.0 and older
|
||||||
|
const prependPath =
|
||||||
|
typeof args.prependPath === 'string'
|
||||||
|
? args.prependPath
|
||||||
|
: args.prependPath.join(':')
|
||||||
|
|
||||||
|
dockerArgs.push(
|
||||||
|
'-e',
|
||||||
|
`PATH=${prependPath}:${await getContainerEnvValue(containerId, 'PATH')}`
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
dockerArgs.push(containerId)
|
dockerArgs.push(containerId)
|
||||||
dockerArgs.push(args.entryPoint)
|
dockerArgs.push(args.entryPoint)
|
||||||
for (const entryPointArg of args.entryPointArgs) {
|
for (const entryPointArg of args.entryPointArgs) {
|
||||||
dockerArgs.push(entryPointArg)
|
dockerArgs.push(entryPointArg)
|
||||||
}
|
}
|
||||||
await runDockerCommand(dockerArgs)
|
await runDockerCommand(dockerArgs, { env: args.environmentVariables })
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function containerRun(
|
export async function containerRun(
|
||||||
args: RunContainerStepArgs,
|
args: RunContainerStepArgs,
|
||||||
name: string,
|
name: string,
|
||||||
network: string
|
network?: string
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
if (!args.image) {
|
if (!args.image) {
|
||||||
throw new Error('expected image to be set')
|
throw new Error('expected image to be set')
|
||||||
@@ -340,15 +393,15 @@ export async function containerRun(
|
|||||||
dockerArgs.push('--name', name)
|
dockerArgs.push('--name', name)
|
||||||
dockerArgs.push(`--workdir=${args.workingDirectory}`)
|
dockerArgs.push(`--workdir=${args.workingDirectory}`)
|
||||||
dockerArgs.push(`--label=${getRunnerLabel()}`)
|
dockerArgs.push(`--label=${getRunnerLabel()}`)
|
||||||
dockerArgs.push(`--network=${network}`)
|
if (network) {
|
||||||
|
dockerArgs.push(`--network=${network}`)
|
||||||
|
}
|
||||||
|
|
||||||
if (args.createOptions) {
|
if (args.createOptions) {
|
||||||
dockerArgs.push(...args.createOptions.split(' '))
|
dockerArgs.push(...args.createOptions.split(' '))
|
||||||
}
|
}
|
||||||
if (args.environmentVariables) {
|
if (args.environmentVariables) {
|
||||||
for (const [key, value] of Object.entries(args.environmentVariables)) {
|
for (const [key] of Object.entries(args.environmentVariables)) {
|
||||||
// Pass in this way to avoid printing secrets
|
|
||||||
env[key] = value ?? undefined
|
|
||||||
dockerArgs.push('-e')
|
dockerArgs.push('-e')
|
||||||
dockerArgs.push(key)
|
dockerArgs.push(key)
|
||||||
}
|
}
|
||||||
@@ -374,11 +427,14 @@ export async function containerRun(
|
|||||||
dockerArgs.push(args.image)
|
dockerArgs.push(args.image)
|
||||||
if (args.entryPointArgs) {
|
if (args.entryPointArgs) {
|
||||||
for (const entryPointArg of args.entryPointArgs) {
|
for (const entryPointArg of args.entryPointArgs) {
|
||||||
|
if (!entryPointArg) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
dockerArgs.push(entryPointArg)
|
dockerArgs.push(entryPointArg)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
await runDockerCommand(dockerArgs)
|
await runDockerCommand(dockerArgs, { env: args.environmentVariables })
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function isContainerAlpine(containerId: string): Promise<boolean> {
|
export async function isContainerAlpine(containerId: string): Promise<boolean> {
|
||||||
|
|||||||
@@ -1,21 +1,9 @@
|
|||||||
import {
|
import {
|
||||||
containerRemove,
|
containerNetworkPrune,
|
||||||
containerNetworkRemove
|
containerPrune
|
||||||
} from '../dockerCommands/container'
|
} from '../dockerCommands/container'
|
||||||
|
|
||||||
// eslint-disable-next-line @typescript-eslint/no-unused-vars
|
export async function cleanupJob(): Promise<void> {
|
||||||
export async function cleanupJob(args, state, responseFile): Promise<void> {
|
await containerPrune()
|
||||||
const containerIds: string[] = []
|
await containerNetworkPrune()
|
||||||
if (state?.container) {
|
|
||||||
containerIds.push(state.container)
|
|
||||||
}
|
|
||||||
if (state?.services) {
|
|
||||||
containerIds.push(state.services)
|
|
||||||
}
|
|
||||||
if (containerIds.length > 0) {
|
|
||||||
await containerRemove(containerIds)
|
|
||||||
}
|
|
||||||
if (state.network) {
|
|
||||||
await containerNetworkRemove(state.network)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -48,6 +48,7 @@ export async function prepareJob(
|
|||||||
} finally {
|
} finally {
|
||||||
await registryLogout(configLocation)
|
await registryLogout(configLocation)
|
||||||
}
|
}
|
||||||
|
|
||||||
containerMetadata = await createContainer(
|
containerMetadata = await createContainer(
|
||||||
container,
|
container,
|
||||||
generateContainerName(container.image),
|
generateContainerName(container.image),
|
||||||
@@ -78,6 +79,7 @@ export async function prepareJob(
|
|||||||
generateContainerName(service.image),
|
generateContainerName(service.image),
|
||||||
networkName
|
networkName
|
||||||
)
|
)
|
||||||
|
|
||||||
servicesMetadata.push(response)
|
servicesMetadata.push(response)
|
||||||
await containerStart(response.id)
|
await containerStart(response.id)
|
||||||
}
|
}
|
||||||
@@ -94,7 +96,10 @@ export async function prepareJob(
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
const isAlpine = await isContainerAlpine(containerMetadata!.id)
|
let isAlpine = false
|
||||||
|
if (containerMetadata?.id) {
|
||||||
|
isAlpine = await isContainerAlpine(containerMetadata.id)
|
||||||
|
}
|
||||||
|
|
||||||
if (containerMetadata?.id) {
|
if (containerMetadata?.id) {
|
||||||
containerMetadata.ports = await containerPorts(containerMetadata.id)
|
containerMetadata.ports = await containerPorts(containerMetadata.id)
|
||||||
@@ -105,7 +110,10 @@ export async function prepareJob(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
const healthChecks: Promise<void>[] = [healthCheck(containerMetadata!)]
|
const healthChecks: Promise<void>[] = []
|
||||||
|
if (containerMetadata) {
|
||||||
|
healthChecks.push(healthCheck(containerMetadata))
|
||||||
|
}
|
||||||
for (const service of servicesMetadata) {
|
for (const service of servicesMetadata) {
|
||||||
healthChecks.push(healthCheck(service))
|
healthChecks.push(healthCheck(service))
|
||||||
}
|
}
|
||||||
@@ -133,7 +141,6 @@ function generateResponseFile(
|
|||||||
servicesMetadata?: ContainerMetadata[],
|
servicesMetadata?: ContainerMetadata[],
|
||||||
isAlpine = false
|
isAlpine = false
|
||||||
): void {
|
): void {
|
||||||
// todo figure out if we are alpine
|
|
||||||
const response = {
|
const response = {
|
||||||
state: { network: networkName },
|
state: { network: networkName },
|
||||||
context: {},
|
context: {},
|
||||||
@@ -186,15 +193,15 @@ function transformDockerPortsToContextPorts(
|
|||||||
meta: ContainerMetadata
|
meta: ContainerMetadata
|
||||||
): ContextPorts {
|
): ContextPorts {
|
||||||
// ex: '80/tcp -> 0.0.0.0:80'
|
// ex: '80/tcp -> 0.0.0.0:80'
|
||||||
const re = /^(\d+)\/(\w+)? -> (.*):(\d+)$/
|
const re = /^(\d+)(\/\w+)? -> (.*):(\d+)$/
|
||||||
const contextPorts: ContextPorts = {}
|
const contextPorts: ContextPorts = {}
|
||||||
|
|
||||||
if (meta.ports) {
|
if (meta.ports?.length) {
|
||||||
for (const port of meta.ports) {
|
for (const port of meta.ports) {
|
||||||
const matches = port.match(re)
|
const matches = port.match(re)
|
||||||
if (!matches) {
|
if (!matches) {
|
||||||
throw new Error(
|
throw new Error(
|
||||||
'Container ports could not match the regex: "^(\\d+)\\/(\\w+)? -> (.*):(\\d+)$"'
|
'Container ports could not match the regex: "^(\\d+)(\\/\\w+)? -> (.*):(\\d+)$"'
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
contextPorts[matches[1]] = matches[matches.length - 1]
|
contextPorts[matches[1]] = matches[matches.length - 1]
|
||||||
|
|||||||
@@ -1,13 +1,12 @@
|
|||||||
|
import { RunContainerStepArgs } from 'hooklib/lib'
|
||||||
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
import {
|
import {
|
||||||
containerBuild,
|
containerBuild,
|
||||||
registryLogin,
|
|
||||||
registryLogout,
|
|
||||||
containerPull,
|
containerPull,
|
||||||
containerRun
|
containerRun,
|
||||||
|
registryLogin,
|
||||||
|
registryLogout
|
||||||
} from '../dockerCommands'
|
} from '../dockerCommands'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
|
||||||
import * as core from '@actions/core'
|
|
||||||
import { RunContainerStepArgs } from 'hooklib/lib'
|
|
||||||
import { getRunnerLabel } from '../dockerCommands/constants'
|
import { getRunnerLabel } from '../dockerCommands/constants'
|
||||||
|
|
||||||
export async function runContainerStep(
|
export async function runContainerStep(
|
||||||
@@ -15,23 +14,23 @@ export async function runContainerStep(
|
|||||||
state
|
state
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const tag = generateBuildTag() // for docker build
|
const tag = generateBuildTag() // for docker build
|
||||||
if (!args.image) {
|
if (args.image) {
|
||||||
core.error('expected an image')
|
const configLocation = await registryLogin(args.registry)
|
||||||
} else {
|
try {
|
||||||
if (args.dockerfile) {
|
await containerPull(args.image, configLocation)
|
||||||
await containerBuild(args, tag)
|
} finally {
|
||||||
args.image = tag
|
await registryLogout(configLocation)
|
||||||
} else {
|
|
||||||
const configLocation = await registryLogin(args)
|
|
||||||
try {
|
|
||||||
await containerPull(args.image, configLocation)
|
|
||||||
} finally {
|
|
||||||
await registryLogout(configLocation)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
} else if (args.dockerfile) {
|
||||||
|
await containerBuild(args, tag)
|
||||||
|
args.image = tag
|
||||||
|
} else {
|
||||||
|
throw new Error(
|
||||||
|
'run container step should have image or dockerfile fields specified'
|
||||||
|
)
|
||||||
}
|
}
|
||||||
// container will get pruned at the end of the job based on the label, no need to cleanup here
|
// container will get pruned at the end of the job based on the label, no need to cleanup here
|
||||||
await containerRun(args, tag.split(':')[1], state.network)
|
await containerRun(args, tag.split(':')[1], state?.network)
|
||||||
}
|
}
|
||||||
|
|
||||||
function generateBuildTag(): string {
|
function generateBuildTag(): string {
|
||||||
|
|||||||
@@ -13,6 +13,7 @@ import {
|
|||||||
runContainerStep,
|
runContainerStep,
|
||||||
runScriptStep
|
runScriptStep
|
||||||
} from './hooks'
|
} from './hooks'
|
||||||
|
import { checkEnvironment } from './utils'
|
||||||
|
|
||||||
async function run(): Promise<void> {
|
async function run(): Promise<void> {
|
||||||
const input = await getInputFromStdin()
|
const input = await getInputFromStdin()
|
||||||
@@ -23,12 +24,13 @@ async function run(): Promise<void> {
|
|||||||
const state = input['state']
|
const state = input['state']
|
||||||
|
|
||||||
try {
|
try {
|
||||||
|
checkEnvironment()
|
||||||
switch (command) {
|
switch (command) {
|
||||||
case Command.PrepareJob:
|
case Command.PrepareJob:
|
||||||
await prepareJob(args as PrepareJobArgs, responseFile)
|
await prepareJob(args as PrepareJobArgs, responseFile)
|
||||||
return exit(0)
|
return exit(0)
|
||||||
case Command.CleanupJob:
|
case Command.CleanupJob:
|
||||||
await cleanupJob(null, state, null)
|
await cleanupJob()
|
||||||
return exit(0)
|
return exit(0)
|
||||||
case Command.RunScriptStep:
|
case Command.RunScriptStep:
|
||||||
await runScriptStep(args as RunScriptStepArgs, state)
|
await runScriptStep(args as RunScriptStepArgs, state)
|
||||||
|
|||||||
@@ -2,18 +2,21 @@
|
|||||||
/* eslint-disable @typescript-eslint/no-require-imports */
|
/* eslint-disable @typescript-eslint/no-require-imports */
|
||||||
/* eslint-disable import/no-commonjs */
|
/* eslint-disable import/no-commonjs */
|
||||||
import * as core from '@actions/core'
|
import * as core from '@actions/core'
|
||||||
|
import { env } from 'process'
|
||||||
// Import this way otherwise typescript has errors
|
// Import this way otherwise typescript has errors
|
||||||
const exec = require('@actions/exec')
|
const exec = require('@actions/exec')
|
||||||
|
|
||||||
export interface RunDockerCommandOptions {
|
export interface RunDockerCommandOptions {
|
||||||
workingDir?: string
|
workingDir?: string
|
||||||
input?: Buffer
|
input?: Buffer
|
||||||
|
env?: { [key: string]: string }
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function runDockerCommand(
|
export async function runDockerCommand(
|
||||||
args: string[],
|
args: string[],
|
||||||
options?: RunDockerCommandOptions
|
options?: RunDockerCommandOptions
|
||||||
): Promise<string> {
|
): Promise<string> {
|
||||||
|
options = optionsWithDockerEnvs(options)
|
||||||
const pipes = await exec.getExecOutput('docker', args, options)
|
const pipes = await exec.getExecOutput('docker', args, options)
|
||||||
if (pipes.exitCode !== 0) {
|
if (pipes.exitCode !== 0) {
|
||||||
core.error(`Docker failed with exit code ${pipes.exitCode}`)
|
core.error(`Docker failed with exit code ${pipes.exitCode}`)
|
||||||
@@ -22,6 +25,45 @@ export async function runDockerCommand(
|
|||||||
return Promise.resolve(pipes.stdout)
|
return Promise.resolve(pipes.stdout)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function optionsWithDockerEnvs(
|
||||||
|
options?: RunDockerCommandOptions
|
||||||
|
): RunDockerCommandOptions | undefined {
|
||||||
|
// From https://docs.docker.com/engine/reference/commandline/cli/#environment-variables
|
||||||
|
const dockerCliEnvs = new Set([
|
||||||
|
'DOCKER_API_VERSION',
|
||||||
|
'DOCKER_CERT_PATH',
|
||||||
|
'DOCKER_CONFIG',
|
||||||
|
'DOCKER_CONTENT_TRUST_SERVER',
|
||||||
|
'DOCKER_CONTENT_TRUST',
|
||||||
|
'DOCKER_CONTEXT',
|
||||||
|
'DOCKER_DEFAULT_PLATFORM',
|
||||||
|
'DOCKER_HIDE_LEGACY_COMMANDS',
|
||||||
|
'DOCKER_HOST',
|
||||||
|
'DOCKER_STACK_ORCHESTRATOR',
|
||||||
|
'DOCKER_TLS_VERIFY',
|
||||||
|
'BUILDKIT_PROGRESS'
|
||||||
|
])
|
||||||
|
const dockerEnvs = {}
|
||||||
|
for (const key in process.env) {
|
||||||
|
if (dockerCliEnvs.has(key)) {
|
||||||
|
dockerEnvs[key] = process.env[key]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const newOptions = {
|
||||||
|
workingDir: options?.workingDir,
|
||||||
|
input: options?.input,
|
||||||
|
env: options?.env || {}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set docker envs or overwrite provided ones
|
||||||
|
for (const [key, value] of Object.entries(dockerEnvs)) {
|
||||||
|
newOptions.env[key] = value as string
|
||||||
|
}
|
||||||
|
|
||||||
|
return newOptions
|
||||||
|
}
|
||||||
|
|
||||||
export function sanitize(val: string): string {
|
export function sanitize(val: string): string {
|
||||||
if (!val || typeof val !== 'string') {
|
if (!val || typeof val !== 'string') {
|
||||||
return ''
|
return ''
|
||||||
@@ -42,6 +84,12 @@ export function sanitize(val: string): string {
|
|||||||
return newNameBuilder.join('')
|
return newNameBuilder.join('')
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function checkEnvironment(): void {
|
||||||
|
if (!env.GITHUB_WORKSPACE) {
|
||||||
|
throw new Error('GITHUB_WORKSPACE is not set')
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// isAlpha accepts single character and checks if
|
// isAlpha accepts single character and checks if
|
||||||
// that character is [a-zA-Z]
|
// that character is [a-zA-Z]
|
||||||
function isAlpha(val: string): boolean {
|
function isAlpha(val: string): boolean {
|
||||||
|
|||||||
@@ -1,62 +1,33 @@
|
|||||||
import { prepareJob, cleanupJob } from '../src/hooks'
|
import { PrepareJobArgs } from 'hooklib/lib'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
import { cleanupJob, prepareJob } from '../src/hooks'
|
||||||
import * as fs from 'fs'
|
|
||||||
import * as path from 'path'
|
|
||||||
import TestSetup from './test-setup'
|
import TestSetup from './test-setup'
|
||||||
|
|
||||||
const prepareJobInputPath = path.resolve(
|
|
||||||
`${__dirname}/../../../examples/prepare-job.json`
|
|
||||||
)
|
|
||||||
|
|
||||||
const tmpOutputDir = `${__dirname}/${uuidv4()}`
|
|
||||||
|
|
||||||
let prepareJobOutputPath: string
|
|
||||||
let prepareJobData: any
|
|
||||||
|
|
||||||
let testSetup: TestSetup
|
let testSetup: TestSetup
|
||||||
|
|
||||||
jest.useRealTimers()
|
jest.useRealTimers()
|
||||||
|
|
||||||
describe('cleanup job', () => {
|
describe('cleanup job', () => {
|
||||||
beforeAll(() => {
|
|
||||||
fs.mkdirSync(tmpOutputDir, { recursive: true })
|
|
||||||
})
|
|
||||||
|
|
||||||
afterAll(() => {
|
|
||||||
fs.rmSync(tmpOutputDir, { recursive: true })
|
|
||||||
})
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
const prepareJobRawData = fs.readFileSync(prepareJobInputPath, 'utf8')
|
|
||||||
prepareJobData = JSON.parse(prepareJobRawData.toString())
|
|
||||||
|
|
||||||
prepareJobOutputPath = `${tmpOutputDir}/prepare-job-output-${uuidv4()}.json`
|
|
||||||
fs.writeFileSync(prepareJobOutputPath, '')
|
|
||||||
|
|
||||||
testSetup = new TestSetup()
|
testSetup = new TestSetup()
|
||||||
testSetup.initialize()
|
testSetup.initialize()
|
||||||
|
|
||||||
prepareJobData.args.container.userMountVolumes = testSetup.userMountVolumes
|
const prepareJobDefinition = testSetup.getPrepareJobDefinition()
|
||||||
prepareJobData.args.container.systemMountVolumes =
|
|
||||||
testSetup.systemMountVolumes
|
|
||||||
prepareJobData.args.container.workingDirectory = testSetup.workingDirectory
|
|
||||||
|
|
||||||
await prepareJob(prepareJobData.args, prepareJobOutputPath)
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
|
'prepare-job-output.json'
|
||||||
|
)
|
||||||
|
|
||||||
|
await prepareJob(
|
||||||
|
prepareJobDefinition.args as PrepareJobArgs,
|
||||||
|
prepareJobOutput
|
||||||
|
)
|
||||||
})
|
})
|
||||||
|
|
||||||
afterEach(() => {
|
afterEach(() => {
|
||||||
fs.rmSync(prepareJobOutputPath, { force: true })
|
|
||||||
testSetup.teardown()
|
testSetup.teardown()
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should cleanup successfully', async () => {
|
it('should cleanup successfully', async () => {
|
||||||
const prepareJobOutputContent = fs.readFileSync(
|
await expect(cleanupJob()).resolves.not.toThrow()
|
||||||
prepareJobOutputPath,
|
|
||||||
'utf-8'
|
|
||||||
)
|
|
||||||
const parsedPrepareJobOutput = JSON.parse(prepareJobOutputContent)
|
|
||||||
await expect(
|
|
||||||
cleanupJob(prepareJobData.args, parsedPrepareJobOutput.state, null)
|
|
||||||
).resolves.not.toThrow()
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
27
packages/docker/tests/container-build-test.ts
Normal file
27
packages/docker/tests/container-build-test.ts
Normal file
@@ -0,0 +1,27 @@
|
|||||||
|
import { containerBuild } from '../src/dockerCommands'
|
||||||
|
import TestSetup from './test-setup'
|
||||||
|
|
||||||
|
let testSetup
|
||||||
|
let runContainerStepDefinition
|
||||||
|
|
||||||
|
describe('container build', () => {
|
||||||
|
beforeEach(() => {
|
||||||
|
testSetup = new TestSetup()
|
||||||
|
testSetup.initialize()
|
||||||
|
|
||||||
|
runContainerStepDefinition = testSetup.getRunContainerStepDefinition()
|
||||||
|
})
|
||||||
|
|
||||||
|
afterEach(() => {
|
||||||
|
testSetup.teardown()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should build container', async () => {
|
||||||
|
runContainerStepDefinition.image = ''
|
||||||
|
const actionPath = testSetup.initializeDockerAction()
|
||||||
|
runContainerStepDefinition.dockerfile = `${actionPath}/Dockerfile`
|
||||||
|
await expect(
|
||||||
|
containerBuild(runContainerStepDefinition, 'example-test-tag')
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
|
})
|
||||||
@@ -4,7 +4,7 @@ jest.useRealTimers()
|
|||||||
|
|
||||||
describe('container pull', () => {
|
describe('container pull', () => {
|
||||||
it('should fail', async () => {
|
it('should fail', async () => {
|
||||||
const arg = { image: 'doesNotExist' }
|
const arg = { image: 'does-not-exist' }
|
||||||
await expect(containerPull(arg.image, '')).rejects.toThrow()
|
await expect(containerPull(arg.image, '')).rejects.toThrow()
|
||||||
})
|
})
|
||||||
it('should succeed', async () => {
|
it('should succeed', async () => {
|
||||||
|
|||||||
@@ -1,102 +1,72 @@
|
|||||||
import {
|
|
||||||
prepareJob,
|
|
||||||
cleanupJob,
|
|
||||||
runScriptStep,
|
|
||||||
runContainerStep
|
|
||||||
} from '../src/hooks'
|
|
||||||
import * as fs from 'fs'
|
import * as fs from 'fs'
|
||||||
import * as path from 'path'
|
import {
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
cleanupJob,
|
||||||
|
prepareJob,
|
||||||
|
runContainerStep,
|
||||||
|
runScriptStep
|
||||||
|
} from '../src/hooks'
|
||||||
import TestSetup from './test-setup'
|
import TestSetup from './test-setup'
|
||||||
|
|
||||||
const prepareJobJson = fs.readFileSync(
|
let definitions
|
||||||
path.resolve(__dirname + '/../../../examples/prepare-job.json'),
|
|
||||||
'utf8'
|
|
||||||
)
|
|
||||||
|
|
||||||
const containerStepJson = fs.readFileSync(
|
|
||||||
path.resolve(__dirname + '/../../../examples/run-container-step.json'),
|
|
||||||
'utf8'
|
|
||||||
)
|
|
||||||
|
|
||||||
const tmpOutputDir = `${__dirname}/_temp/${uuidv4()}`
|
|
||||||
|
|
||||||
let prepareJobData: any
|
|
||||||
let scriptStepJson: any
|
|
||||||
let scriptStepData: any
|
|
||||||
let containerStepData: any
|
|
||||||
|
|
||||||
let prepareJobOutputFilePath: string
|
|
||||||
|
|
||||||
let testSetup: TestSetup
|
let testSetup: TestSetup
|
||||||
|
|
||||||
describe('e2e', () => {
|
describe('e2e', () => {
|
||||||
beforeAll(() => {
|
|
||||||
fs.mkdirSync(tmpOutputDir, { recursive: true })
|
|
||||||
})
|
|
||||||
|
|
||||||
afterAll(() => {
|
|
||||||
fs.rmSync(tmpOutputDir, { recursive: true })
|
|
||||||
})
|
|
||||||
|
|
||||||
beforeEach(() => {
|
beforeEach(() => {
|
||||||
// init dirs
|
|
||||||
testSetup = new TestSetup()
|
testSetup = new TestSetup()
|
||||||
testSetup.initialize()
|
testSetup.initialize()
|
||||||
|
|
||||||
prepareJobData = JSON.parse(prepareJobJson)
|
definitions = {
|
||||||
prepareJobData.args.container.userMountVolumes = testSetup.userMountVolumes
|
prepareJob: testSetup.getPrepareJobDefinition(),
|
||||||
prepareJobData.args.container.systemMountVolumes =
|
runScriptStep: testSetup.getRunScriptStepDefinition(),
|
||||||
testSetup.systemMountVolumes
|
runContainerStep: testSetup.getRunContainerStepDefinition()
|
||||||
prepareJobData.args.container.workingDirectory = testSetup.workingDirectory
|
}
|
||||||
|
|
||||||
scriptStepJson = fs.readFileSync(
|
|
||||||
path.resolve(__dirname + '/../../../examples/run-script-step.json'),
|
|
||||||
'utf8'
|
|
||||||
)
|
|
||||||
scriptStepData = JSON.parse(scriptStepJson)
|
|
||||||
scriptStepData.args.workingDirectory = testSetup.workingDirectory
|
|
||||||
|
|
||||||
containerStepData = JSON.parse(containerStepJson)
|
|
||||||
containerStepData.args.workingDirectory = testSetup.workingDirectory
|
|
||||||
containerStepData.args.userMountVolumes = testSetup.userMountVolumes
|
|
||||||
containerStepData.args.systemMountVolumes = testSetup.systemMountVolumes
|
|
||||||
|
|
||||||
prepareJobOutputFilePath = `${tmpOutputDir}/prepare-job-output-${uuidv4()}.json`
|
|
||||||
fs.writeFileSync(prepareJobOutputFilePath, '')
|
|
||||||
})
|
})
|
||||||
|
|
||||||
afterEach(() => {
|
afterEach(() => {
|
||||||
fs.rmSync(prepareJobOutputFilePath, { force: true })
|
|
||||||
testSetup.teardown()
|
testSetup.teardown()
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should prepare job, then run script step, then run container step then cleanup', async () => {
|
it('should prepare job, then run script step, then run container step then cleanup', async () => {
|
||||||
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
|
'prepare-job-output.json'
|
||||||
|
)
|
||||||
|
|
||||||
await expect(
|
await expect(
|
||||||
prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
prepareJob(definitions.prepareJob.args, prepareJobOutput)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
let rawState = fs.readFileSync(prepareJobOutputFilePath, 'utf-8')
|
|
||||||
|
let rawState = fs.readFileSync(prepareJobOutput, 'utf-8')
|
||||||
let resp = JSON.parse(rawState)
|
let resp = JSON.parse(rawState)
|
||||||
|
|
||||||
await expect(
|
await expect(
|
||||||
runScriptStep(scriptStepData.args, resp.state)
|
runScriptStep(definitions.runScriptStep.args, resp.state)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
|
|
||||||
await expect(
|
await expect(
|
||||||
runContainerStep(containerStepData.args, resp.state)
|
runContainerStep(definitions.runContainerStep.args, resp.state)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
await expect(cleanupJob(resp, resp.state, null)).resolves.not.toThrow()
|
|
||||||
|
await expect(cleanupJob()).resolves.not.toThrow()
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should prepare job, then run script step, then run container step with Dockerfile then cleanup', async () => {
|
it('should prepare job, then run script step, then run container step with Dockerfile then cleanup', async () => {
|
||||||
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
|
'prepare-job-output.json'
|
||||||
|
)
|
||||||
|
|
||||||
await expect(
|
await expect(
|
||||||
prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
prepareJob(definitions.prepareJob.args, prepareJobOutput)
|
||||||
).resolves.not.toThrow()
|
|
||||||
let rawState = fs.readFileSync(prepareJobOutputFilePath, 'utf-8')
|
|
||||||
let resp = JSON.parse(rawState)
|
|
||||||
await expect(
|
|
||||||
runScriptStep(scriptStepData.args, resp.state)
|
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
|
|
||||||
const dockerfilePath = `${tmpOutputDir}/Dockerfile`
|
let rawState = fs.readFileSync(prepareJobOutput, 'utf-8')
|
||||||
|
let resp = JSON.parse(rawState)
|
||||||
|
|
||||||
|
await expect(
|
||||||
|
runScriptStep(definitions.runScriptStep.args, resp.state)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
|
||||||
|
const dockerfilePath = `${testSetup.workingDirectory}/Dockerfile`
|
||||||
fs.writeFileSync(
|
fs.writeFileSync(
|
||||||
dockerfilePath,
|
dockerfilePath,
|
||||||
`FROM ubuntu:latest
|
`FROM ubuntu:latest
|
||||||
@@ -104,14 +74,17 @@ ENV TEST=test
|
|||||||
ENTRYPOINT [ "tail", "-f", "/dev/null" ]
|
ENTRYPOINT [ "tail", "-f", "/dev/null" ]
|
||||||
`
|
`
|
||||||
)
|
)
|
||||||
const containerStepDataCopy = JSON.parse(JSON.stringify(containerStepData))
|
|
||||||
process.env.GITHUB_WORKSPACE = tmpOutputDir
|
const containerStepDataCopy = JSON.parse(
|
||||||
|
JSON.stringify(definitions.runContainerStep)
|
||||||
|
)
|
||||||
|
|
||||||
containerStepDataCopy.args.dockerfile = 'Dockerfile'
|
containerStepDataCopy.args.dockerfile = 'Dockerfile'
|
||||||
containerStepDataCopy.args.context = '.'
|
|
||||||
console.log(containerStepDataCopy.args)
|
|
||||||
await expect(
|
await expect(
|
||||||
runContainerStep(containerStepDataCopy.args, resp.state)
|
runContainerStep(containerStepDataCopy.args, resp.state)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
await expect(cleanupJob(resp, resp.state, null)).resolves.not.toThrow()
|
|
||||||
|
await expect(cleanupJob()).resolves.not.toThrow()
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -1,40 +1,18 @@
|
|||||||
import * as fs from 'fs'
|
import * as fs from 'fs'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
|
||||||
import { prepareJob } from '../src/hooks'
|
import { prepareJob } from '../src/hooks'
|
||||||
import TestSetup from './test-setup'
|
import TestSetup from './test-setup'
|
||||||
|
|
||||||
jest.useRealTimers()
|
jest.useRealTimers()
|
||||||
|
|
||||||
let prepareJobOutputPath: string
|
let prepareJobDefinition
|
||||||
let prepareJobData: any
|
|
||||||
const tmpOutputDir = `${__dirname}/_temp/${uuidv4()}`
|
|
||||||
const prepareJobInputPath = `${__dirname}/../../../examples/prepare-job.json`
|
|
||||||
|
|
||||||
let testSetup: TestSetup
|
let testSetup: TestSetup
|
||||||
|
|
||||||
describe('prepare job', () => {
|
describe('prepare job', () => {
|
||||||
beforeAll(() => {
|
beforeEach(() => {
|
||||||
fs.mkdirSync(tmpOutputDir, { recursive: true })
|
|
||||||
})
|
|
||||||
|
|
||||||
afterAll(() => {
|
|
||||||
fs.rmSync(tmpOutputDir, { recursive: true })
|
|
||||||
})
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
testSetup = new TestSetup()
|
testSetup = new TestSetup()
|
||||||
testSetup.initialize()
|
testSetup.initialize()
|
||||||
|
prepareJobDefinition = testSetup.getPrepareJobDefinition()
|
||||||
let prepareJobRawData = fs.readFileSync(prepareJobInputPath, 'utf8')
|
|
||||||
prepareJobData = JSON.parse(prepareJobRawData.toString())
|
|
||||||
|
|
||||||
prepareJobData.args.container.userMountVolumes = testSetup.userMountVolumes
|
|
||||||
prepareJobData.args.container.systemMountVolumes =
|
|
||||||
testSetup.systemMountVolumes
|
|
||||||
prepareJobData.args.container.workingDirectory = testSetup.workingDirectory
|
|
||||||
|
|
||||||
prepareJobOutputPath = `${tmpOutputDir}/prepare-job-output-${uuidv4()}.json`
|
|
||||||
fs.writeFileSync(prepareJobOutputPath, '')
|
|
||||||
})
|
})
|
||||||
|
|
||||||
afterEach(() => {
|
afterEach(() => {
|
||||||
@@ -42,38 +20,68 @@ describe('prepare job', () => {
|
|||||||
})
|
})
|
||||||
|
|
||||||
it('should not throw', async () => {
|
it('should not throw', async () => {
|
||||||
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
|
'prepare-job-output.json'
|
||||||
|
)
|
||||||
await expect(
|
await expect(
|
||||||
prepareJob(prepareJobData.args, prepareJobOutputPath)
|
prepareJob(prepareJobDefinition.args, prepareJobOutput)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
|
|
||||||
expect(() => fs.readFileSync(prepareJobOutputPath, 'utf-8')).not.toThrow()
|
expect(() => fs.readFileSync(prepareJobOutput, 'utf-8')).not.toThrow()
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should have JSON output written to a file', async () => {
|
it('should have JSON output written to a file', async () => {
|
||||||
await prepareJob(prepareJobData.args, prepareJobOutputPath)
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
const prepareJobOutputContent = fs.readFileSync(
|
'prepare-job-output.json'
|
||||||
prepareJobOutputPath,
|
|
||||||
'utf-8'
|
|
||||||
)
|
)
|
||||||
|
await prepareJob(prepareJobDefinition.args, prepareJobOutput)
|
||||||
|
const prepareJobOutputContent = fs.readFileSync(prepareJobOutput, 'utf-8')
|
||||||
expect(() => JSON.parse(prepareJobOutputContent)).not.toThrow()
|
expect(() => JSON.parse(prepareJobOutputContent)).not.toThrow()
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should have context written to a file', async () => {
|
it('should have context written to a file', async () => {
|
||||||
await prepareJob(prepareJobData.args, prepareJobOutputPath)
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
const prepareJobOutputContent = fs.readFileSync(
|
'prepare-job-output.json'
|
||||||
prepareJobOutputPath,
|
)
|
||||||
'utf-8'
|
await prepareJob(prepareJobDefinition.args, prepareJobOutput)
|
||||||
|
const parsedPrepareJobOutput = JSON.parse(
|
||||||
|
fs.readFileSync(prepareJobOutput, 'utf-8')
|
||||||
)
|
)
|
||||||
const parsedPrepareJobOutput = JSON.parse(prepareJobOutputContent)
|
|
||||||
expect(parsedPrepareJobOutput.context).toBeDefined()
|
expect(parsedPrepareJobOutput.context).toBeDefined()
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should have container ids written to file', async () => {
|
it('should have isAlpine field set correctly', async () => {
|
||||||
await prepareJob(prepareJobData.args, prepareJobOutputPath)
|
let prepareJobOutput = testSetup.createOutputFile(
|
||||||
const prepareJobOutputContent = fs.readFileSync(
|
'prepare-job-output-alpine.json'
|
||||||
prepareJobOutputPath,
|
|
||||||
'utf-8'
|
|
||||||
)
|
)
|
||||||
|
const prepareJobArgsClone = JSON.parse(
|
||||||
|
JSON.stringify(prepareJobDefinition.args)
|
||||||
|
)
|
||||||
|
prepareJobArgsClone.container.image = 'alpine:latest'
|
||||||
|
await prepareJob(prepareJobArgsClone, prepareJobOutput)
|
||||||
|
|
||||||
|
let parsedPrepareJobOutput = JSON.parse(
|
||||||
|
fs.readFileSync(prepareJobOutput, 'utf-8')
|
||||||
|
)
|
||||||
|
expect(parsedPrepareJobOutput.isAlpine).toBe(true)
|
||||||
|
|
||||||
|
prepareJobOutput = testSetup.createOutputFile(
|
||||||
|
'prepare-job-output-ubuntu.json'
|
||||||
|
)
|
||||||
|
prepareJobArgsClone.container.image = 'ubuntu:latest'
|
||||||
|
await prepareJob(prepareJobArgsClone, prepareJobOutput)
|
||||||
|
parsedPrepareJobOutput = JSON.parse(
|
||||||
|
fs.readFileSync(prepareJobOutput, 'utf-8')
|
||||||
|
)
|
||||||
|
expect(parsedPrepareJobOutput.isAlpine).toBe(false)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should have container ids written to file', async () => {
|
||||||
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
|
'prepare-job-output.json'
|
||||||
|
)
|
||||||
|
await prepareJob(prepareJobDefinition.args, prepareJobOutput)
|
||||||
|
const prepareJobOutputContent = fs.readFileSync(prepareJobOutput, 'utf-8')
|
||||||
const parsedPrepareJobOutput = JSON.parse(prepareJobOutputContent)
|
const parsedPrepareJobOutput = JSON.parse(prepareJobOutputContent)
|
||||||
|
|
||||||
expect(parsedPrepareJobOutput.context.container.id).toBeDefined()
|
expect(parsedPrepareJobOutput.context.container.id).toBeDefined()
|
||||||
@@ -82,11 +90,11 @@ describe('prepare job', () => {
|
|||||||
})
|
})
|
||||||
|
|
||||||
it('should have ports for context written in form [containerPort]:[hostPort]', async () => {
|
it('should have ports for context written in form [containerPort]:[hostPort]', async () => {
|
||||||
await prepareJob(prepareJobData.args, prepareJobOutputPath)
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
const prepareJobOutputContent = fs.readFileSync(
|
'prepare-job-output.json'
|
||||||
prepareJobOutputPath,
|
|
||||||
'utf-8'
|
|
||||||
)
|
)
|
||||||
|
await prepareJob(prepareJobDefinition.args, prepareJobOutput)
|
||||||
|
const prepareJobOutputContent = fs.readFileSync(prepareJobOutput, 'utf-8')
|
||||||
const parsedPrepareJobOutput = JSON.parse(prepareJobOutputContent)
|
const parsedPrepareJobOutput = JSON.parse(prepareJobOutputContent)
|
||||||
|
|
||||||
const mainContainerPorts = parsedPrepareJobOutput.context.container.ports
|
const mainContainerPorts = parsedPrepareJobOutput.context.container.ports
|
||||||
@@ -100,4 +108,14 @@ describe('prepare job', () => {
|
|||||||
expect(redisServicePorts['80']).toBe('8080')
|
expect(redisServicePorts['80']).toBe('8080')
|
||||||
expect(redisServicePorts['8080']).toBe('8088')
|
expect(redisServicePorts['8080']).toBe('8088')
|
||||||
})
|
})
|
||||||
|
|
||||||
|
it('should run prepare job without job container without exception', async () => {
|
||||||
|
prepareJobDefinition.args.container = null
|
||||||
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
|
'prepare-job-output.json'
|
||||||
|
)
|
||||||
|
await expect(
|
||||||
|
prepareJob(prepareJobDefinition.args, prepareJobOutput)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
63
packages/docker/tests/run-script-step-test.ts
Normal file
63
packages/docker/tests/run-script-step-test.ts
Normal file
@@ -0,0 +1,63 @@
|
|||||||
|
import * as fs from 'fs'
|
||||||
|
import { PrepareJobResponse } from 'hooklib/lib'
|
||||||
|
import { prepareJob, runScriptStep } from '../src/hooks'
|
||||||
|
import TestSetup from './test-setup'
|
||||||
|
|
||||||
|
jest.useRealTimers()
|
||||||
|
|
||||||
|
let testSetup: TestSetup
|
||||||
|
|
||||||
|
let definitions
|
||||||
|
|
||||||
|
let prepareJobResponse: PrepareJobResponse
|
||||||
|
|
||||||
|
describe('run script step', () => {
|
||||||
|
beforeEach(async () => {
|
||||||
|
testSetup = new TestSetup()
|
||||||
|
testSetup.initialize()
|
||||||
|
|
||||||
|
definitions = {
|
||||||
|
prepareJob: testSetup.getPrepareJobDefinition(),
|
||||||
|
runScriptStep: testSetup.getRunScriptStepDefinition()
|
||||||
|
}
|
||||||
|
|
||||||
|
const prepareJobOutput = testSetup.createOutputFile(
|
||||||
|
'prepare-job-output.json'
|
||||||
|
)
|
||||||
|
await prepareJob(definitions.prepareJob.args, prepareJobOutput)
|
||||||
|
|
||||||
|
prepareJobResponse = JSON.parse(fs.readFileSync(prepareJobOutput, 'utf-8'))
|
||||||
|
})
|
||||||
|
|
||||||
|
it('Should run script step without exceptions', async () => {
|
||||||
|
await expect(
|
||||||
|
runScriptStep(definitions.runScriptStep.args, prepareJobResponse.state)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('Should have path variable changed in container with prepend path string', async () => {
|
||||||
|
definitions.runScriptStep.args.prependPath = '/some/path'
|
||||||
|
definitions.runScriptStep.args.entryPoint = '/bin/bash'
|
||||||
|
definitions.runScriptStep.args.entryPointArgs = [
|
||||||
|
'-c',
|
||||||
|
`if [[ ! $(env | grep "^PATH=") = "PATH=${definitions.runScriptStep.args.prependPath}:"* ]]; then exit 1; fi`
|
||||||
|
]
|
||||||
|
await expect(
|
||||||
|
runScriptStep(definitions.runScriptStep.args, prepareJobResponse.state)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('Should have path variable changed in container with prepend path string array', async () => {
|
||||||
|
definitions.runScriptStep.args.prependPath = ['/some/other/path']
|
||||||
|
definitions.runScriptStep.args.entryPoint = '/bin/bash'
|
||||||
|
definitions.runScriptStep.args.entryPointArgs = [
|
||||||
|
'-c',
|
||||||
|
`if [[ ! $(env | grep "^PATH=") = "PATH=${definitions.runScriptStep.args.prependPath.join(
|
||||||
|
':'
|
||||||
|
)}:"* ]]; then exit 1; fi`
|
||||||
|
]
|
||||||
|
await expect(
|
||||||
|
runScriptStep(definitions.runScriptStep.args, prepareJobResponse.state)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
|
})
|
||||||
@@ -1,11 +1,15 @@
|
|||||||
import * as fs from 'fs'
|
import * as fs from 'fs'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
|
||||||
import { env } from 'process'
|
|
||||||
import { Mount } from 'hooklib'
|
import { Mount } from 'hooklib'
|
||||||
|
import { HookData } from 'hooklib/lib'
|
||||||
|
import * as path from 'path'
|
||||||
|
import { env } from 'process'
|
||||||
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
|
|
||||||
export default class TestSetup {
|
export default class TestSetup {
|
||||||
private testdir: string
|
private testdir: string
|
||||||
private runnerMockDir: string
|
private runnerMockDir: string
|
||||||
|
readonly runnerOutputDir: string
|
||||||
|
|
||||||
private runnerMockSubdirs = {
|
private runnerMockSubdirs = {
|
||||||
work: '_work',
|
work: '_work',
|
||||||
externals: 'externals',
|
externals: 'externals',
|
||||||
@@ -16,15 +20,16 @@ export default class TestSetup {
|
|||||||
githubWorkflow: '_work/_temp/_github_workflow'
|
githubWorkflow: '_work/_temp/_github_workflow'
|
||||||
}
|
}
|
||||||
|
|
||||||
private readonly projectName = 'example'
|
private readonly projectName = 'repo'
|
||||||
|
|
||||||
constructor() {
|
constructor() {
|
||||||
this.testdir = `${__dirname}/_temp/${uuidv4()}`
|
this.testdir = `${__dirname}/_temp/${uuidv4()}`
|
||||||
this.runnerMockDir = `${this.testdir}/runner/_layout`
|
this.runnerMockDir = `${this.testdir}/runner/_layout`
|
||||||
|
this.runnerOutputDir = `${this.testdir}/outputs`
|
||||||
}
|
}
|
||||||
|
|
||||||
private get allTestDirectories() {
|
private get allTestDirectories() {
|
||||||
const resp = [this.testdir, this.runnerMockDir]
|
const resp = [this.testdir, this.runnerMockDir, this.runnerOutputDir]
|
||||||
|
|
||||||
for (const [key, value] of Object.entries(this.runnerMockSubdirs)) {
|
for (const [key, value] of Object.entries(this.runnerMockSubdirs)) {
|
||||||
resp.push(`${this.runnerMockDir}/${value}`)
|
resp.push(`${this.runnerMockDir}/${value}`)
|
||||||
@@ -38,30 +43,27 @@ export default class TestSetup {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public initialize(): void {
|
public initialize(): void {
|
||||||
for (const dir of this.allTestDirectories) {
|
env['GITHUB_WORKSPACE'] = this.workingDirectory
|
||||||
fs.mkdirSync(dir, { recursive: true })
|
|
||||||
}
|
|
||||||
env['RUNNER_NAME'] = 'test'
|
env['RUNNER_NAME'] = 'test'
|
||||||
env[
|
env[
|
||||||
'RUNNER_TEMP'
|
'RUNNER_TEMP'
|
||||||
] = `${this.runnerMockDir}/${this.runnerMockSubdirs.workTemp}`
|
] = `${this.runnerMockDir}/${this.runnerMockSubdirs.workTemp}`
|
||||||
|
|
||||||
|
for (const dir of this.allTestDirectories) {
|
||||||
|
fs.mkdirSync(dir, { recursive: true })
|
||||||
|
}
|
||||||
|
|
||||||
|
fs.copyFileSync(
|
||||||
|
path.resolve(`${__dirname}/../../../examples/example-script.sh`),
|
||||||
|
`${env.RUNNER_TEMP}/example-script.sh`
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
public teardown(): void {
|
public teardown(): void {
|
||||||
fs.rmdirSync(this.testdir, { recursive: true })
|
fs.rmdirSync(this.testdir, { recursive: true })
|
||||||
}
|
}
|
||||||
|
|
||||||
public get userMountVolumes(): Mount[] {
|
private get systemMountVolumes(): Mount[] {
|
||||||
return [
|
|
||||||
{
|
|
||||||
sourceVolumePath: 'my_docker_volume',
|
|
||||||
targetVolumePath: '/volume_mount',
|
|
||||||
readOnly: false
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
|
|
||||||
public get systemMountVolumes(): Mount[] {
|
|
||||||
return [
|
return [
|
||||||
{
|
{
|
||||||
sourceVolumePath: '/var/run/docker.sock',
|
sourceVolumePath: '/var/run/docker.sock',
|
||||||
@@ -106,7 +108,89 @@ export default class TestSetup {
|
|||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public createOutputFile(name: string): string {
|
||||||
|
let filePath = path.join(this.runnerOutputDir, name || `${uuidv4()}.json`)
|
||||||
|
fs.writeFileSync(filePath, '')
|
||||||
|
return filePath
|
||||||
|
}
|
||||||
|
|
||||||
public get workingDirectory(): string {
|
public get workingDirectory(): string {
|
||||||
|
return `${this.runnerMockDir}/_work/${this.projectName}/${this.projectName}`
|
||||||
|
}
|
||||||
|
|
||||||
|
public get containerWorkingDirectory(): string {
|
||||||
return `/__w/${this.projectName}/${this.projectName}`
|
return `/__w/${this.projectName}/${this.projectName}`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public initializeDockerAction(): string {
|
||||||
|
const actionPath = `${this.testdir}/_actions/example-handle/example-repo/example-branch/mock-directory`
|
||||||
|
fs.mkdirSync(actionPath, { recursive: true })
|
||||||
|
this.writeDockerfile(actionPath)
|
||||||
|
this.writeEntrypoint(actionPath)
|
||||||
|
return actionPath
|
||||||
|
}
|
||||||
|
|
||||||
|
private writeDockerfile(actionPath: string) {
|
||||||
|
const content = `FROM alpine:3.10
|
||||||
|
COPY entrypoint.sh /entrypoint.sh
|
||||||
|
ENTRYPOINT ["/entrypoint.sh"]`
|
||||||
|
fs.writeFileSync(`${actionPath}/Dockerfile`, content)
|
||||||
|
}
|
||||||
|
|
||||||
|
private writeEntrypoint(actionPath) {
|
||||||
|
const content = `#!/bin/sh -l
|
||||||
|
echo "Hello $1"
|
||||||
|
time=$(date)
|
||||||
|
echo "::set-output name=time::$time"`
|
||||||
|
const entryPointPath = `${actionPath}/entrypoint.sh`
|
||||||
|
fs.writeFileSync(entryPointPath, content)
|
||||||
|
fs.chmodSync(entryPointPath, 0o755)
|
||||||
|
}
|
||||||
|
|
||||||
|
public getPrepareJobDefinition(): HookData {
|
||||||
|
const prepareJob = JSON.parse(
|
||||||
|
fs.readFileSync(
|
||||||
|
path.resolve(__dirname + '/../../../examples/prepare-job.json'),
|
||||||
|
'utf8'
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
prepareJob.args.container.systemMountVolumes = this.systemMountVolumes
|
||||||
|
prepareJob.args.container.workingDirectory = this.workingDirectory
|
||||||
|
prepareJob.args.container.userMountVolumes = undefined
|
||||||
|
prepareJob.args.container.registry = null
|
||||||
|
prepareJob.args.services.forEach(s => {
|
||||||
|
s.registry = null
|
||||||
|
})
|
||||||
|
|
||||||
|
return prepareJob
|
||||||
|
}
|
||||||
|
|
||||||
|
public getRunScriptStepDefinition(): HookData {
|
||||||
|
const runScriptStep = JSON.parse(
|
||||||
|
fs.readFileSync(
|
||||||
|
path.resolve(__dirname + '/../../../examples/run-script-step.json'),
|
||||||
|
'utf8'
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
runScriptStep.args.entryPointArgs[1] = `/__w/_temp/example-script.sh`
|
||||||
|
return runScriptStep
|
||||||
|
}
|
||||||
|
|
||||||
|
public getRunContainerStepDefinition(): HookData {
|
||||||
|
const runContainerStep = JSON.parse(
|
||||||
|
fs.readFileSync(
|
||||||
|
path.resolve(__dirname + '/../../../examples/run-container-step.json'),
|
||||||
|
'utf8'
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
runContainerStep.args.entryPointArgs[1] = `/__w/_temp/example-script.sh`
|
||||||
|
runContainerStep.args.systemMountVolumes = this.systemMountVolumes
|
||||||
|
runContainerStep.args.workingDirectory = this.workingDirectory
|
||||||
|
runContainerStep.args.userMountVolumes = undefined
|
||||||
|
runContainerStep.args.registry = null
|
||||||
|
return runContainerStep
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
import { sanitize } from '../src/utils'
|
import { optionsWithDockerEnvs, sanitize } from '../src/utils'
|
||||||
|
|
||||||
describe('Utilities', () => {
|
describe('Utilities', () => {
|
||||||
it('should return sanitized image name', () => {
|
it('should return sanitized image name', () => {
|
||||||
@@ -9,4 +9,41 @@ describe('Utilities', () => {
|
|||||||
const validStr = 'teststr8_one'
|
const validStr = 'teststr8_one'
|
||||||
expect(sanitize(validStr)).toBe(validStr)
|
expect(sanitize(validStr)).toBe(validStr)
|
||||||
})
|
})
|
||||||
|
|
||||||
|
describe('with docker options', () => {
|
||||||
|
it('should augment options with docker environment variables', () => {
|
||||||
|
process.env.DOCKER_HOST = 'unix:///run/user/1001/docker.sock'
|
||||||
|
process.env.DOCKER_NOTEXIST = 'notexist'
|
||||||
|
|
||||||
|
const optionDefinitions: any = [
|
||||||
|
undefined,
|
||||||
|
{},
|
||||||
|
{ env: {} },
|
||||||
|
{ env: { DOCKER_HOST: 'unix://var/run/docker.sock' } }
|
||||||
|
]
|
||||||
|
for (const opt of optionDefinitions) {
|
||||||
|
let options = optionsWithDockerEnvs(opt)
|
||||||
|
expect(options).toBeDefined()
|
||||||
|
expect(options?.env).toBeDefined()
|
||||||
|
expect(options?.env?.DOCKER_HOST).toBe(process.env.DOCKER_HOST)
|
||||||
|
expect(options?.env?.DOCKER_NOTEXIST).toBeUndefined()
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should not overwrite other options', () => {
|
||||||
|
process.env.DOCKER_HOST = 'unix:///run/user/1001/docker.sock'
|
||||||
|
const opt = {
|
||||||
|
workingDir: 'test',
|
||||||
|
input: Buffer.from('test')
|
||||||
|
}
|
||||||
|
|
||||||
|
const options = optionsWithDockerEnvs(opt)
|
||||||
|
expect(options).toBeDefined()
|
||||||
|
expect(options?.workingDir).toBe(opt.workingDir)
|
||||||
|
expect(options?.input).toBe(opt.input)
|
||||||
|
expect(options?.env).toStrictEqual({
|
||||||
|
DOCKER_HOST: process.env.DOCKER_HOST
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
49
packages/hooklib/package-lock.json
generated
49
packages/hooklib/package-lock.json
generated
@@ -9,7 +9,7 @@
|
|||||||
"version": "0.1.0",
|
"version": "0.1.0",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": "^1.6.0"
|
"@actions/core": "^1.9.1"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@types/node": "^17.0.23",
|
"@types/node": "^17.0.23",
|
||||||
@@ -22,19 +22,20 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@actions/core": {
|
"node_modules/@actions/core": {
|
||||||
"version": "1.6.0",
|
"version": "1.9.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.6.0.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.9.1.tgz",
|
||||||
"integrity": "sha512-NB1UAZomZlCV/LmJqkLhNTqtKfFXJZAUPcfl/zqG7EfsQdeUJtaWO98SGbuQ3pydJ3fHl2CvI/51OKYlCYYcaw==",
|
"integrity": "sha512-5ad+U2YGrmmiw6du20AQW5XuWo7UKN2052FjSV7MX+Wfjf8sCqcsZe62NfgHys4QI4/Y+vQvLKYL8jWtA1ZBTA==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/http-client": "^1.0.11"
|
"@actions/http-client": "^2.0.1",
|
||||||
|
"uuid": "^8.3.2"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@actions/http-client": {
|
"node_modules/@actions/http-client": {
|
||||||
"version": "1.0.11",
|
"version": "2.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-1.0.11.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-2.0.1.tgz",
|
||||||
"integrity": "sha512-VRYHGQV1rqnROJqdMvGUbY/Kn8vriQe/F9HR2AlYHzmKuM/p3kjNuXhmdBfcVgsvRWTz5C5XW5xvndZrVBuAYg==",
|
"integrity": "sha512-PIXiMVtz6VvyaRsGY268qvj57hXQEpsYogYOu2nrQhlf+XCGmZstmuZBbAybUl1nQGnvS1k1eEsQ69ZoD7xlSw==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"tunnel": "0.0.6"
|
"tunnel": "^0.0.6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@eslint/eslintrc": {
|
"node_modules/@eslint/eslintrc": {
|
||||||
@@ -2485,6 +2486,14 @@
|
|||||||
"punycode": "^2.1.0"
|
"punycode": "^2.1.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"node_modules/uuid": {
|
||||||
|
"version": "8.3.2",
|
||||||
|
"resolved": "https://registry.npmjs.org/uuid/-/uuid-8.3.2.tgz",
|
||||||
|
"integrity": "sha512-+NYs2QeMWy+GWFOEm9xnn6HCDp0l7QBD7ml8zLUmJ+93Q5NF0NocErnwkTkXVFNiX3/fpC6afS8Dhb/gz7R7eg==",
|
||||||
|
"bin": {
|
||||||
|
"uuid": "dist/bin/uuid"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/v8-compile-cache": {
|
"node_modules/v8-compile-cache": {
|
||||||
"version": "2.3.0",
|
"version": "2.3.0",
|
||||||
"resolved": "https://registry.npmjs.org/v8-compile-cache/-/v8-compile-cache-2.3.0.tgz",
|
"resolved": "https://registry.npmjs.org/v8-compile-cache/-/v8-compile-cache-2.3.0.tgz",
|
||||||
@@ -2546,19 +2555,20 @@
|
|||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": {
|
"@actions/core": {
|
||||||
"version": "1.6.0",
|
"version": "1.9.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.6.0.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.9.1.tgz",
|
||||||
"integrity": "sha512-NB1UAZomZlCV/LmJqkLhNTqtKfFXJZAUPcfl/zqG7EfsQdeUJtaWO98SGbuQ3pydJ3fHl2CvI/51OKYlCYYcaw==",
|
"integrity": "sha512-5ad+U2YGrmmiw6du20AQW5XuWo7UKN2052FjSV7MX+Wfjf8sCqcsZe62NfgHys4QI4/Y+vQvLKYL8jWtA1ZBTA==",
|
||||||
"requires": {
|
"requires": {
|
||||||
"@actions/http-client": "^1.0.11"
|
"@actions/http-client": "^2.0.1",
|
||||||
|
"uuid": "^8.3.2"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"@actions/http-client": {
|
"@actions/http-client": {
|
||||||
"version": "1.0.11",
|
"version": "2.0.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-1.0.11.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-2.0.1.tgz",
|
||||||
"integrity": "sha512-VRYHGQV1rqnROJqdMvGUbY/Kn8vriQe/F9HR2AlYHzmKuM/p3kjNuXhmdBfcVgsvRWTz5C5XW5xvndZrVBuAYg==",
|
"integrity": "sha512-PIXiMVtz6VvyaRsGY268qvj57hXQEpsYogYOu2nrQhlf+XCGmZstmuZBbAybUl1nQGnvS1k1eEsQ69ZoD7xlSw==",
|
||||||
"requires": {
|
"requires": {
|
||||||
"tunnel": "0.0.6"
|
"tunnel": "^0.0.6"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"@eslint/eslintrc": {
|
"@eslint/eslintrc": {
|
||||||
@@ -4300,6 +4310,11 @@
|
|||||||
"punycode": "^2.1.0"
|
"punycode": "^2.1.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"uuid": {
|
||||||
|
"version": "8.3.2",
|
||||||
|
"resolved": "https://registry.npmjs.org/uuid/-/uuid-8.3.2.tgz",
|
||||||
|
"integrity": "sha512-+NYs2QeMWy+GWFOEm9xnn6HCDp0l7QBD7ml8zLUmJ+93Q5NF0NocErnwkTkXVFNiX3/fpC6afS8Dhb/gz7R7eg=="
|
||||||
|
},
|
||||||
"v8-compile-cache": {
|
"v8-compile-cache": {
|
||||||
"version": "2.3.0",
|
"version": "2.3.0",
|
||||||
"resolved": "https://registry.npmjs.org/v8-compile-cache/-/v8-compile-cache-2.3.0.tgz",
|
"resolved": "https://registry.npmjs.org/v8-compile-cache/-/v8-compile-cache-2.3.0.tgz",
|
||||||
|
|||||||
@@ -23,6 +23,6 @@
|
|||||||
"typescript": "^4.6.3"
|
"typescript": "^4.6.3"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": "^1.6.0"
|
"@actions/core": "^1.9.1"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -34,6 +34,7 @@ export interface ContainerInfo {
|
|||||||
createOptions?: string
|
createOptions?: string
|
||||||
environmentVariables?: { [key: string]: string }
|
environmentVariables?: { [key: string]: string }
|
||||||
userMountVolumes?: Mount[]
|
userMountVolumes?: Mount[]
|
||||||
|
systemMountVolumes?: Mount[]
|
||||||
registry?: Registry
|
registry?: Registry
|
||||||
portMappings?: string[]
|
portMappings?: string[]
|
||||||
}
|
}
|
||||||
@@ -73,14 +74,6 @@ export enum Protocol {
|
|||||||
UDP = 'udp'
|
UDP = 'udp'
|
||||||
}
|
}
|
||||||
|
|
||||||
export enum PodPhase {
|
|
||||||
PENDING = 'Pending',
|
|
||||||
RUNNING = 'Running',
|
|
||||||
SUCCEEDED = 'Succeded',
|
|
||||||
FAILED = 'Failed',
|
|
||||||
UNKNOWN = 'Unknown'
|
|
||||||
}
|
|
||||||
|
|
||||||
export interface PrepareJobResponse {
|
export interface PrepareJobResponse {
|
||||||
state?: object
|
state?: object
|
||||||
context?: ContainerContext
|
context?: ContainerContext
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
import * as core from '@actions/core'
|
|
||||||
import * as events from 'events'
|
import * as events from 'events'
|
||||||
import * as fs from 'fs'
|
import * as fs from 'fs'
|
||||||
import * as os from 'os'
|
import * as os from 'os'
|
||||||
@@ -13,7 +12,6 @@ export async function getInputFromStdin(): Promise<HookData> {
|
|||||||
})
|
})
|
||||||
|
|
||||||
rl.on('line', line => {
|
rl.on('line', line => {
|
||||||
core.debug(`Line from STDIN: ${line}`)
|
|
||||||
input = line
|
input = line
|
||||||
})
|
})
|
||||||
await events.default.once(rl, 'close')
|
await events.default.once(rl, 'close')
|
||||||
|
|||||||
@@ -6,7 +6,40 @@ This implementation provides a way to dynamically spin up jobs to run container
|
|||||||
## Pre-requisites
|
## Pre-requisites
|
||||||
Some things are expected to be set when using these hooks
|
Some things are expected to be set when using these hooks
|
||||||
- The runner itself should be running in a pod, with a service account with the following permissions
|
- The runner itself should be running in a pod, with a service account with the following permissions
|
||||||
- The `ACTIONS_RUNNER_REQUIRE_JOB_CONTAINER=true` should be set to true
|
```
|
||||||
|
apiVersion: rbac.authorization.k8s.io/v1
|
||||||
|
kind: Role
|
||||||
|
metadata:
|
||||||
|
namespace: default
|
||||||
|
name: runner-role
|
||||||
|
rules:
|
||||||
|
- apiGroups: [""]
|
||||||
|
resources: ["pods"]
|
||||||
|
verbs: ["get", "list", "create", "delete"]
|
||||||
|
- apiGroups: [""]
|
||||||
|
resources: ["pods/exec"]
|
||||||
|
verbs: ["get", "create"]
|
||||||
|
- apiGroups: [""]
|
||||||
|
resources: ["pods/log"]
|
||||||
|
verbs: ["get", "list", "watch",]
|
||||||
|
- apiGroups: ["batch"]
|
||||||
|
resources: ["jobs"]
|
||||||
|
verbs: ["get", "list", "create", "delete"]
|
||||||
|
- apiGroups: [""]
|
||||||
|
resources: ["secrets"]
|
||||||
|
verbs: ["get", "list", "create", "delete"]
|
||||||
|
```
|
||||||
- The `ACTIONS_RUNNER_POD_NAME` env should be set to the name of the pod
|
- The `ACTIONS_RUNNER_POD_NAME` env should be set to the name of the pod
|
||||||
|
- The `ACTIONS_RUNNER_REQUIRE_JOB_CONTAINER` env should be set to true to prevent the runner from running any jobs outside of a container
|
||||||
- The runner pod should map a persistent volume claim into the `_work` directory
|
- The runner pod should map a persistent volume claim into the `_work` directory
|
||||||
- The `ACTIONS_RUNNER_CLAIM_NAME` should be set to the persistent volume claim that contains the runner's working directory
|
- The `ACTIONS_RUNNER_CLAIM_NAME` env should be set to the persistent volume claim that contains the runner's working directory, otherwise it defaults to `${ACTIONS_RUNNER_POD_NAME}-work`
|
||||||
|
- Some actions runner env's are expected to be set. These are set automatically by the runner.
|
||||||
|
- `RUNNER_WORKSPACE` is expected to be set to the workspace of the runner
|
||||||
|
- `GITHUB_WORKSPACE` is expected to be set to the workspace of the job
|
||||||
|
|
||||||
|
|
||||||
|
## Limitations
|
||||||
|
- A [job containers](https://docs.github.com/en/actions/using-jobs/running-jobs-in-a-container) will be required for all jobs
|
||||||
|
- Building container actions from a dockerfile is not supported at this time
|
||||||
|
- Container actions will not have access to the services network or job container network
|
||||||
|
- Docker [create options](https://docs.github.com/en/actions/using-workflows/workflow-syntax-for-github-actions#jobsjob_idcontaineroptions) are not supported
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
jest.setTimeout(90000)
|
jest.setTimeout(500000)
|
||||||
|
|||||||
51
packages/k8s/package-lock.json
generated
51
packages/k8s/package-lock.json
generated
@@ -9,7 +9,7 @@
|
|||||||
"version": "0.1.0",
|
"version": "0.1.0",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": "^1.6.0",
|
"@actions/core": "^1.9.1",
|
||||||
"@actions/exec": "^1.1.1",
|
"@actions/exec": "^1.1.1",
|
||||||
"@actions/io": "^1.1.2",
|
"@actions/io": "^1.1.2",
|
||||||
"@kubernetes/client-node": "^0.16.3",
|
"@kubernetes/client-node": "^0.16.3",
|
||||||
@@ -28,7 +28,7 @@
|
|||||||
"version": "0.1.0",
|
"version": "0.1.0",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": "^1.6.0"
|
"@actions/core": "^1.9.1"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@types/node": "^17.0.23",
|
"@types/node": "^17.0.23",
|
||||||
@@ -41,11 +41,20 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@actions/core": {
|
"node_modules/@actions/core": {
|
||||||
"version": "1.8.2",
|
"version": "1.9.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.8.2.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.9.1.tgz",
|
||||||
"integrity": "sha512-FXcBL7nyik8K5ODeCKlxi+vts7torOkoDAKfeh61EAkAy1HAvwn9uVzZBY0f15YcQTcZZ2/iSGBFHEuioZWfDA==",
|
"integrity": "sha512-5ad+U2YGrmmiw6du20AQW5XuWo7UKN2052FjSV7MX+Wfjf8sCqcsZe62NfgHys4QI4/Y+vQvLKYL8jWtA1ZBTA==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/http-client": "^2.0.1"
|
"@actions/http-client": "^2.0.1",
|
||||||
|
"uuid": "^8.3.2"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"node_modules/@actions/core/node_modules/uuid": {
|
||||||
|
"version": "8.3.2",
|
||||||
|
"resolved": "https://registry.npmjs.org/uuid/-/uuid-8.3.2.tgz",
|
||||||
|
"integrity": "sha512-+NYs2QeMWy+GWFOEm9xnn6HCDp0l7QBD7ml8zLUmJ+93Q5NF0NocErnwkTkXVFNiX3/fpC6afS8Dhb/gz7R7eg==",
|
||||||
|
"bin": {
|
||||||
|
"uuid": "dist/bin/uuid"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@actions/exec": {
|
"node_modules/@actions/exec": {
|
||||||
@@ -3428,9 +3437,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/jose": {
|
"node_modules/jose": {
|
||||||
"version": "2.0.5",
|
"version": "2.0.6",
|
||||||
"resolved": "https://registry.npmjs.org/jose/-/jose-2.0.5.tgz",
|
"resolved": "https://registry.npmjs.org/jose/-/jose-2.0.6.tgz",
|
||||||
"integrity": "sha512-BAiDNeDKTMgk4tvD0BbxJ8xHEHBZgpeRZ1zGPPsitSyMgjoMWiLGYAE7H7NpP5h0lPppQajQs871E8NHUrzVPA==",
|
"integrity": "sha512-FVoPY7SflDodE4lknJmbAHSUjLCzE2H1F6MS0RYKMQ8SR+lNccpMf8R4eqkNYyyUjR5qZReOzZo5C5YiHOCjjg==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@panva/asn1.js": "^1.0.0"
|
"@panva/asn1.js": "^1.0.0"
|
||||||
},
|
},
|
||||||
@@ -5145,11 +5154,19 @@
|
|||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": {
|
"@actions/core": {
|
||||||
"version": "1.8.2",
|
"version": "1.9.1",
|
||||||
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.8.2.tgz",
|
"resolved": "https://registry.npmjs.org/@actions/core/-/core-1.9.1.tgz",
|
||||||
"integrity": "sha512-FXcBL7nyik8K5ODeCKlxi+vts7torOkoDAKfeh61EAkAy1HAvwn9uVzZBY0f15YcQTcZZ2/iSGBFHEuioZWfDA==",
|
"integrity": "sha512-5ad+U2YGrmmiw6du20AQW5XuWo7UKN2052FjSV7MX+Wfjf8sCqcsZe62NfgHys4QI4/Y+vQvLKYL8jWtA1ZBTA==",
|
||||||
"requires": {
|
"requires": {
|
||||||
"@actions/http-client": "^2.0.1"
|
"@actions/http-client": "^2.0.1",
|
||||||
|
"uuid": "^8.3.2"
|
||||||
|
},
|
||||||
|
"dependencies": {
|
||||||
|
"uuid": {
|
||||||
|
"version": "8.3.2",
|
||||||
|
"resolved": "https://registry.npmjs.org/uuid/-/uuid-8.3.2.tgz",
|
||||||
|
"integrity": "sha512-+NYs2QeMWy+GWFOEm9xnn6HCDp0l7QBD7ml8zLUmJ+93Q5NF0NocErnwkTkXVFNiX3/fpC6afS8Dhb/gz7R7eg=="
|
||||||
|
}
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"@actions/exec": {
|
"@actions/exec": {
|
||||||
@@ -7074,7 +7091,7 @@
|
|||||||
"hooklib": {
|
"hooklib": {
|
||||||
"version": "file:../hooklib",
|
"version": "file:../hooklib",
|
||||||
"requires": {
|
"requires": {
|
||||||
"@actions/core": "^1.6.0",
|
"@actions/core": "^1.9.1",
|
||||||
"@types/node": "^17.0.23",
|
"@types/node": "^17.0.23",
|
||||||
"@typescript-eslint/parser": "^5.18.0",
|
"@typescript-eslint/parser": "^5.18.0",
|
||||||
"@zeit/ncc": "^0.22.3",
|
"@zeit/ncc": "^0.22.3",
|
||||||
@@ -7804,9 +7821,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"jose": {
|
"jose": {
|
||||||
"version": "2.0.5",
|
"version": "2.0.6",
|
||||||
"resolved": "https://registry.npmjs.org/jose/-/jose-2.0.5.tgz",
|
"resolved": "https://registry.npmjs.org/jose/-/jose-2.0.6.tgz",
|
||||||
"integrity": "sha512-BAiDNeDKTMgk4tvD0BbxJ8xHEHBZgpeRZ1zGPPsitSyMgjoMWiLGYAE7H7NpP5h0lPppQajQs871E8NHUrzVPA==",
|
"integrity": "sha512-FVoPY7SflDodE4lknJmbAHSUjLCzE2H1F6MS0RYKMQ8SR+lNccpMf8R4eqkNYyyUjR5qZReOzZo5C5YiHOCjjg==",
|
||||||
"requires": {
|
"requires": {
|
||||||
"@panva/asn1.js": "^1.0.0"
|
"@panva/asn1.js": "^1.0.0"
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -13,7 +13,7 @@
|
|||||||
"author": "",
|
"author": "",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@actions/core": "^1.6.0",
|
"@actions/core": "^1.9.1",
|
||||||
"@actions/exec": "^1.1.1",
|
"@actions/exec": "^1.1.1",
|
||||||
"@actions/io": "^1.1.2",
|
"@actions/io": "^1.1.2",
|
||||||
"@kubernetes/client-node": "^0.16.3",
|
"@kubernetes/client-node": "^0.16.3",
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
import { podPrune } from '../k8s'
|
import { prunePods, pruneSecrets } from '../k8s'
|
||||||
|
|
||||||
export async function cleanupJob(): Promise<void> {
|
export async function cleanupJob(): Promise<void> {
|
||||||
await podPrune()
|
await Promise.all([prunePods(), pruneSecrets()])
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -20,28 +20,33 @@ export function getJobPodName(): string {
|
|||||||
export function getStepPodName(): string {
|
export function getStepPodName(): string {
|
||||||
return `${getRunnerPodName().substring(
|
return `${getRunnerPodName().substring(
|
||||||
0,
|
0,
|
||||||
MAX_POD_NAME_LENGTH - ('-step'.length + STEP_POD_NAME_SUFFIX_LENGTH)
|
MAX_POD_NAME_LENGTH - ('-step-'.length + STEP_POD_NAME_SUFFIX_LENGTH)
|
||||||
)}-step-${uuidv4().substring(0, STEP_POD_NAME_SUFFIX_LENGTH)}`
|
)}-step-${uuidv4().substring(0, STEP_POD_NAME_SUFFIX_LENGTH)}`
|
||||||
}
|
}
|
||||||
|
|
||||||
export function getVolumeClaimName(): string {
|
export function getVolumeClaimName(): string {
|
||||||
const name = process.env.ACTIONS_RUNNER_CLAIM_NAME
|
const name = process.env.ACTIONS_RUNNER_CLAIM_NAME
|
||||||
if (!name) {
|
if (!name) {
|
||||||
throw new Error(
|
return `${getRunnerPodName()}-work`
|
||||||
"'ACTIONS_RUNNER_CLAIM_NAME' is required, please contact your self hosted runner administrator"
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
return name
|
return name
|
||||||
}
|
}
|
||||||
|
|
||||||
const MAX_POD_NAME_LENGTH = 63
|
export function getSecretName(): string {
|
||||||
const STEP_POD_NAME_SUFFIX_LENGTH = 8
|
return `${getRunnerPodName().substring(
|
||||||
|
0,
|
||||||
|
MAX_POD_NAME_LENGTH - ('-secret-'.length + STEP_POD_NAME_SUFFIX_LENGTH)
|
||||||
|
)}-secret-${uuidv4().substring(0, STEP_POD_NAME_SUFFIX_LENGTH)}`
|
||||||
|
}
|
||||||
|
|
||||||
|
export const MAX_POD_NAME_LENGTH = 63
|
||||||
|
export const STEP_POD_NAME_SUFFIX_LENGTH = 8
|
||||||
export const JOB_CONTAINER_NAME = 'job'
|
export const JOB_CONTAINER_NAME = 'job'
|
||||||
|
|
||||||
export class RunnerInstanceLabel {
|
export class RunnerInstanceLabel {
|
||||||
runnerhook: string
|
private podName: string
|
||||||
constructor() {
|
constructor() {
|
||||||
this.runnerhook = process.env.ACTIONS_RUNNER_POD_NAME as string
|
this.podName = getRunnerPodName()
|
||||||
}
|
}
|
||||||
|
|
||||||
get key(): string {
|
get key(): string {
|
||||||
@@ -49,10 +54,10 @@ export class RunnerInstanceLabel {
|
|||||||
}
|
}
|
||||||
|
|
||||||
get value(): string {
|
get value(): string {
|
||||||
return this.runnerhook
|
return this.podName
|
||||||
}
|
}
|
||||||
|
|
||||||
toString(): string {
|
toString(): string {
|
||||||
return `runner-pod=${this.runnerhook}`
|
return `runner-pod=${this.podName}`
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,27 +1,20 @@
|
|||||||
import * as core from '@actions/core'
|
import * as core from '@actions/core'
|
||||||
import * as io from '@actions/io'
|
import * as io from '@actions/io'
|
||||||
import * as k8s from '@kubernetes/client-node'
|
import * as k8s from '@kubernetes/client-node'
|
||||||
import {
|
import { ContextPorts, prepareJobArgs, writeToResponseFile } from 'hooklib'
|
||||||
ContextPorts,
|
|
||||||
PodPhase,
|
|
||||||
prepareJobArgs,
|
|
||||||
writeToResponseFile
|
|
||||||
} from 'hooklib'
|
|
||||||
import path from 'path'
|
import path from 'path'
|
||||||
import {
|
import {
|
||||||
containerPorts,
|
containerPorts,
|
||||||
createPod,
|
createPod,
|
||||||
isAuthPermissionsOK,
|
|
||||||
isPodContainerAlpine,
|
isPodContainerAlpine,
|
||||||
namespace,
|
prunePods,
|
||||||
podPrune,
|
|
||||||
requiredPermissions,
|
|
||||||
waitForPodPhases
|
waitForPodPhases
|
||||||
} from '../k8s'
|
} from '../k8s'
|
||||||
import {
|
import {
|
||||||
containerVolumes,
|
containerVolumes,
|
||||||
DEFAULT_CONTAINER_ENTRY_POINT,
|
DEFAULT_CONTAINER_ENTRY_POINT,
|
||||||
DEFAULT_CONTAINER_ENTRY_POINT_ARGS
|
DEFAULT_CONTAINER_ENTRY_POINT_ARGS,
|
||||||
|
PodPhase
|
||||||
} from '../k8s/utils'
|
} from '../k8s/utils'
|
||||||
import { JOB_CONTAINER_NAME } from './constants'
|
import { JOB_CONTAINER_NAME } from './constants'
|
||||||
|
|
||||||
@@ -29,25 +22,22 @@ export async function prepareJob(
|
|||||||
args: prepareJobArgs,
|
args: prepareJobArgs,
|
||||||
responseFile
|
responseFile
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
await podPrune()
|
if (!args.container) {
|
||||||
if (!(await isAuthPermissionsOK())) {
|
throw new Error('Job Container is required.')
|
||||||
throw new Error(
|
|
||||||
`The Service account needs the following permissions ${JSON.stringify(
|
|
||||||
requiredPermissions
|
|
||||||
)} on the pod resource in the '${namespace}' namespace. Please contact your self hosted runner administrator.`
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
await prunePods()
|
||||||
await copyExternalsToRoot()
|
await copyExternalsToRoot()
|
||||||
let container: k8s.V1Container | undefined = undefined
|
let container: k8s.V1Container | undefined = undefined
|
||||||
if (args.container?.image) {
|
if (args.container?.image) {
|
||||||
core.info(`Using image '${args.container.image}' for job image`)
|
core.debug(`Using image '${args.container.image}' for job image`)
|
||||||
container = createPodSpec(args.container, JOB_CONTAINER_NAME, true)
|
container = createPodSpec(args.container, JOB_CONTAINER_NAME, true)
|
||||||
}
|
}
|
||||||
|
|
||||||
let services: k8s.V1Container[] = []
|
let services: k8s.V1Container[] = []
|
||||||
if (args.services?.length) {
|
if (args.services?.length) {
|
||||||
services = args.services.map(service => {
|
services = args.services.map(service => {
|
||||||
core.info(`Adding service '${service.image}' to pod definition`)
|
core.debug(`Adding service '${service.image}' to pod definition`)
|
||||||
return createPodSpec(service, service.image.split(':')[0])
|
return createPodSpec(service, service.image.split(':')[0])
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@@ -56,15 +46,18 @@ export async function prepareJob(
|
|||||||
}
|
}
|
||||||
let createdPod: k8s.V1Pod | undefined = undefined
|
let createdPod: k8s.V1Pod | undefined = undefined
|
||||||
try {
|
try {
|
||||||
createdPod = await createPod(container, services, args.registry)
|
createdPod = await createPod(container, services, args.container.registry)
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
await podPrune()
|
await prunePods()
|
||||||
throw new Error(`failed to create job pod: ${err}`)
|
throw new Error(`failed to create job pod: ${err}`)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!createdPod?.metadata?.name) {
|
if (!createdPod?.metadata?.name) {
|
||||||
throw new Error('created pod should have metadata.name')
|
throw new Error('created pod should have metadata.name')
|
||||||
}
|
}
|
||||||
|
core.debug(
|
||||||
|
`Job pod created, waiting for it to come online ${createdPod?.metadata?.name}`
|
||||||
|
)
|
||||||
|
|
||||||
try {
|
try {
|
||||||
await waitForPodPhases(
|
await waitForPodPhases(
|
||||||
@@ -73,11 +66,11 @@ export async function prepareJob(
|
|||||||
new Set([PodPhase.PENDING])
|
new Set([PodPhase.PENDING])
|
||||||
)
|
)
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
await podPrune()
|
await prunePods()
|
||||||
throw new Error(`Pod failed to come online with error: ${err}`)
|
throw new Error(`Pod failed to come online with error: ${err}`)
|
||||||
}
|
}
|
||||||
|
|
||||||
core.info('Pod is ready for traffic')
|
core.debug('Job pod is ready for traffic')
|
||||||
|
|
||||||
let isAlpine = false
|
let isAlpine = false
|
||||||
try {
|
try {
|
||||||
@@ -88,7 +81,7 @@ export async function prepareJob(
|
|||||||
} catch (err) {
|
} catch (err) {
|
||||||
throw new Error(`Failed to determine if the pod is alpine: ${err}`)
|
throw new Error(`Failed to determine if the pod is alpine: ${err}`)
|
||||||
}
|
}
|
||||||
|
core.debug(`Setting isAlpine to ${isAlpine}`)
|
||||||
generateResponseFile(responseFile, createdPod, isAlpine)
|
generateResponseFile(responseFile, createdPod, isAlpine)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -97,8 +90,13 @@ function generateResponseFile(
|
|||||||
appPod: k8s.V1Pod,
|
appPod: k8s.V1Pod,
|
||||||
isAlpine
|
isAlpine
|
||||||
): void {
|
): void {
|
||||||
|
if (!appPod.metadata?.name) {
|
||||||
|
throw new Error('app pod must have metadata.name specified')
|
||||||
|
}
|
||||||
const response = {
|
const response = {
|
||||||
state: {},
|
state: {
|
||||||
|
jobPod: appPod.metadata.name
|
||||||
|
},
|
||||||
context: {},
|
context: {},
|
||||||
isAlpine
|
isAlpine
|
||||||
}
|
}
|
||||||
@@ -160,14 +158,11 @@ function createPodSpec(
|
|||||||
name: string,
|
name: string,
|
||||||
jobContainer = false
|
jobContainer = false
|
||||||
): k8s.V1Container {
|
): k8s.V1Container {
|
||||||
core.info(JSON.stringify(container))
|
if (!container.entryPoint && jobContainer) {
|
||||||
if (!container.entryPointArgs) {
|
container.entryPoint = DEFAULT_CONTAINER_ENTRY_POINT
|
||||||
container.entryPointArgs = DEFAULT_CONTAINER_ENTRY_POINT_ARGS
|
container.entryPointArgs = DEFAULT_CONTAINER_ENTRY_POINT_ARGS
|
||||||
}
|
}
|
||||||
container.entryPointArgs = DEFAULT_CONTAINER_ENTRY_POINT_ARGS
|
|
||||||
if (!container.entryPoint) {
|
|
||||||
container.entryPoint = DEFAULT_CONTAINER_ENTRY_POINT
|
|
||||||
}
|
|
||||||
const podContainer = {
|
const podContainer = {
|
||||||
name,
|
name,
|
||||||
image: container.image,
|
image: container.image,
|
||||||
|
|||||||
@@ -1,22 +1,39 @@
|
|||||||
import * as k8s from '@kubernetes/client-node'
|
|
||||||
import * as core from '@actions/core'
|
import * as core from '@actions/core'
|
||||||
import { PodPhase } from 'hooklib'
|
import * as k8s from '@kubernetes/client-node'
|
||||||
|
import { RunContainerStepArgs } from 'hooklib'
|
||||||
import {
|
import {
|
||||||
createJob,
|
createJob,
|
||||||
|
createSecretForEnvs,
|
||||||
getContainerJobPodName,
|
getContainerJobPodName,
|
||||||
getPodLogs,
|
getPodLogs,
|
||||||
getPodStatus,
|
getPodStatus,
|
||||||
waitForJobToComplete,
|
waitForJobToComplete,
|
||||||
waitForPodPhases
|
waitForPodPhases
|
||||||
} from '../k8s'
|
} from '../k8s'
|
||||||
|
import {
|
||||||
|
containerVolumes,
|
||||||
|
DEFAULT_CONTAINER_ENTRY_POINT,
|
||||||
|
DEFAULT_CONTAINER_ENTRY_POINT_ARGS,
|
||||||
|
PodPhase,
|
||||||
|
writeEntryPointScript
|
||||||
|
} from '../k8s/utils'
|
||||||
import { JOB_CONTAINER_NAME } from './constants'
|
import { JOB_CONTAINER_NAME } from './constants'
|
||||||
import { containerVolumes } from '../k8s/utils'
|
|
||||||
|
|
||||||
export async function runContainerStep(stepContainer): Promise<number> {
|
export async function runContainerStep(
|
||||||
|
stepContainer: RunContainerStepArgs
|
||||||
|
): Promise<number> {
|
||||||
if (stepContainer.dockerfile) {
|
if (stepContainer.dockerfile) {
|
||||||
throw new Error('Building container actions is not currently supported')
|
throw new Error('Building container actions is not currently supported')
|
||||||
}
|
}
|
||||||
const container = createPodSpec(stepContainer)
|
|
||||||
|
let secretName: string | undefined = undefined
|
||||||
|
if (stepContainer.environmentVariables) {
|
||||||
|
secretName = await createSecretForEnvs(stepContainer.environmentVariables)
|
||||||
|
}
|
||||||
|
|
||||||
|
core.debug(`Created secret ${secretName} for container job envs`)
|
||||||
|
const container = createPodSpec(stepContainer, secretName)
|
||||||
|
|
||||||
const job = await createJob(container)
|
const job = await createJob(container)
|
||||||
if (!job.metadata?.name) {
|
if (!job.metadata?.name) {
|
||||||
throw new Error(
|
throw new Error(
|
||||||
@@ -25,45 +42,69 @@ export async function runContainerStep(stepContainer): Promise<number> {
|
|||||||
)} to have correctly set the metadata.name`
|
)} to have correctly set the metadata.name`
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
core.debug(`Job created, waiting for pod to start: ${job.metadata?.name}`)
|
||||||
|
|
||||||
const podName = await getContainerJobPodName(job.metadata.name)
|
const podName = await getContainerJobPodName(job.metadata.name)
|
||||||
await waitForPodPhases(
|
await waitForPodPhases(
|
||||||
podName,
|
podName,
|
||||||
new Set([PodPhase.COMPLETED, PodPhase.RUNNING]),
|
new Set([PodPhase.COMPLETED, PodPhase.RUNNING, PodPhase.SUCCEEDED]),
|
||||||
new Set([PodPhase.PENDING])
|
new Set([PodPhase.PENDING, PodPhase.UNKNOWN])
|
||||||
)
|
)
|
||||||
|
core.debug('Container step is running or complete, pulling logs')
|
||||||
|
|
||||||
await getPodLogs(podName, JOB_CONTAINER_NAME)
|
await getPodLogs(podName, JOB_CONTAINER_NAME)
|
||||||
|
|
||||||
|
core.debug('Waiting for container job to complete')
|
||||||
await waitForJobToComplete(job.metadata.name)
|
await waitForJobToComplete(job.metadata.name)
|
||||||
// pod has failed so pull the status code from the container
|
// pod has failed so pull the status code from the container
|
||||||
const status = await getPodStatus(podName)
|
const status = await getPodStatus(podName)
|
||||||
if (!status?.containerStatuses?.length) {
|
if (status?.phase === 'Succeeded') {
|
||||||
core.warning(`Can't determine container status`)
|
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
if (!status?.containerStatuses?.length) {
|
||||||
|
core.error(
|
||||||
|
`Can't determine container status from response: ${JSON.stringify(
|
||||||
|
status
|
||||||
|
)}`
|
||||||
|
)
|
||||||
|
return 1
|
||||||
|
}
|
||||||
const exitCode =
|
const exitCode =
|
||||||
status.containerStatuses[status.containerStatuses.length - 1].state
|
status.containerStatuses[status.containerStatuses.length - 1].state
|
||||||
?.terminated?.exitCode
|
?.terminated?.exitCode
|
||||||
return Number(exitCode) || 0
|
return Number(exitCode) || 1
|
||||||
}
|
}
|
||||||
|
|
||||||
function createPodSpec(container): k8s.V1Container {
|
function createPodSpec(
|
||||||
|
container: RunContainerStepArgs,
|
||||||
|
secretName?: string
|
||||||
|
): k8s.V1Container {
|
||||||
const podContainer = new k8s.V1Container()
|
const podContainer = new k8s.V1Container()
|
||||||
podContainer.name = JOB_CONTAINER_NAME
|
podContainer.name = JOB_CONTAINER_NAME
|
||||||
podContainer.image = container.image
|
podContainer.image = container.image
|
||||||
if (container.entryPoint) {
|
|
||||||
podContainer.command = [container.entryPoint, ...container.entryPointArgs]
|
|
||||||
}
|
|
||||||
|
|
||||||
podContainer.env = []
|
const { entryPoint, entryPointArgs } = container
|
||||||
for (const [key, value] of Object.entries(
|
container.entryPoint = 'sh'
|
||||||
container['environmentVariables']
|
|
||||||
)) {
|
const { containerPath } = writeEntryPointScript(
|
||||||
if (value && key !== 'HOME') {
|
container.workingDirectory,
|
||||||
podContainer.env.push({ name: key, value: value as string })
|
entryPoint || DEFAULT_CONTAINER_ENTRY_POINT,
|
||||||
}
|
entryPoint ? entryPointArgs || [] : DEFAULT_CONTAINER_ENTRY_POINT_ARGS
|
||||||
|
)
|
||||||
|
container.entryPointArgs = ['-e', containerPath]
|
||||||
|
podContainer.command = [container.entryPoint, ...container.entryPointArgs]
|
||||||
|
|
||||||
|
if (secretName) {
|
||||||
|
podContainer.envFrom = [
|
||||||
|
{
|
||||||
|
secretRef: {
|
||||||
|
name: secretName,
|
||||||
|
optional: false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
]
|
||||||
}
|
}
|
||||||
podContainer.volumeMounts = containerVolumes()
|
podContainer.volumeMounts = containerVolumes(undefined, false, true)
|
||||||
|
|
||||||
return podContainer
|
return podContainer
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,8 @@
|
|||||||
/* eslint-disable @typescript-eslint/no-unused-vars */
|
/* eslint-disable @typescript-eslint/no-unused-vars */
|
||||||
|
import * as fs from 'fs'
|
||||||
import { RunScriptStepArgs } from 'hooklib'
|
import { RunScriptStepArgs } from 'hooklib'
|
||||||
import { execPodStep } from '../k8s'
|
import { execPodStep } from '../k8s'
|
||||||
|
import { writeEntryPointScript } from '../k8s/utils'
|
||||||
import { JOB_CONTAINER_NAME } from './constants'
|
import { JOB_CONTAINER_NAME } from './constants'
|
||||||
|
|
||||||
export async function runScriptStep(
|
export async function runScriptStep(
|
||||||
@@ -8,31 +10,26 @@ export async function runScriptStep(
|
|||||||
state,
|
state,
|
||||||
responseFile
|
responseFile
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const cb = new CommandsBuilder(
|
const { entryPoint, entryPointArgs, environmentVariables } = args
|
||||||
args.entryPoint,
|
const { containerPath, runnerPath } = writeEntryPointScript(
|
||||||
args.entryPointArgs,
|
args.workingDirectory,
|
||||||
args.environmentVariables
|
entryPoint,
|
||||||
|
entryPointArgs,
|
||||||
|
args.prependPath,
|
||||||
|
environmentVariables
|
||||||
)
|
)
|
||||||
await execPodStep(cb.command, state.jobPod, JOB_CONTAINER_NAME)
|
|
||||||
}
|
|
||||||
|
|
||||||
class CommandsBuilder {
|
args.entryPoint = 'sh'
|
||||||
constructor(
|
args.entryPointArgs = ['-e', containerPath]
|
||||||
private entryPoint: string,
|
try {
|
||||||
private entryPointArgs: string[],
|
await execPodStep(
|
||||||
private environmentVariables: { [key: string]: string }
|
[args.entryPoint, ...args.entryPointArgs],
|
||||||
) {}
|
state.jobPod,
|
||||||
|
JOB_CONTAINER_NAME
|
||||||
get command(): string[] {
|
)
|
||||||
const envCommands: string[] = []
|
} catch (err) {
|
||||||
if (
|
throw new Error(`failed to run script step: ${err}`)
|
||||||
this.environmentVariables &&
|
} finally {
|
||||||
Object.entries(this.environmentVariables).length
|
fs.rmSync(runnerPath)
|
||||||
) {
|
|
||||||
for (const [key, value] of Object.entries(this.environmentVariables)) {
|
|
||||||
envCommands.push(`${key}=${value}`)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return ['env', ...envCommands, this.entryPoint, ...this.entryPointArgs]
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
import * as core from '@actions/core'
|
||||||
import { Command, getInputFromStdin, prepareJobArgs } from 'hooklib'
|
import { Command, getInputFromStdin, prepareJobArgs } from 'hooklib'
|
||||||
import {
|
import {
|
||||||
cleanupJob,
|
cleanupJob,
|
||||||
@@ -5,6 +6,7 @@ import {
|
|||||||
runContainerStep,
|
runContainerStep,
|
||||||
runScriptStep
|
runScriptStep
|
||||||
} from './hooks'
|
} from './hooks'
|
||||||
|
import { isAuthPermissionsOK, namespace, requiredPermissions } from './k8s'
|
||||||
|
|
||||||
async function run(): Promise<void> {
|
async function run(): Promise<void> {
|
||||||
const input = await getInputFromStdin()
|
const input = await getInputFromStdin()
|
||||||
@@ -16,6 +18,13 @@ async function run(): Promise<void> {
|
|||||||
|
|
||||||
let exitCode = 0
|
let exitCode = 0
|
||||||
try {
|
try {
|
||||||
|
if (!(await isAuthPermissionsOK())) {
|
||||||
|
throw new Error(
|
||||||
|
`The Service account needs the following permissions ${JSON.stringify(
|
||||||
|
requiredPermissions
|
||||||
|
)} on the pod resource in the '${namespace()}' namespace. Please contact your self hosted runner administrator.`
|
||||||
|
)
|
||||||
|
}
|
||||||
switch (command) {
|
switch (command) {
|
||||||
case Command.PrepareJob:
|
case Command.PrepareJob:
|
||||||
await prepareJob(args as prepareJobArgs, responseFile)
|
await prepareJob(args as prepareJobArgs, responseFile)
|
||||||
@@ -34,8 +43,7 @@ async function run(): Promise<void> {
|
|||||||
throw new Error(`Command not recognized: ${command}`)
|
throw new Error(`Command not recognized: ${command}`)
|
||||||
}
|
}
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
// eslint-disable-next-line no-console
|
core.error(error as Error)
|
||||||
console.log(error)
|
|
||||||
exitCode = 1
|
exitCode = 1
|
||||||
}
|
}
|
||||||
process.exitCode = exitCode
|
process.exitCode = exitCode
|
||||||
|
|||||||
@@ -1,13 +1,16 @@
|
|||||||
|
import * as core from '@actions/core'
|
||||||
import * as k8s from '@kubernetes/client-node'
|
import * as k8s from '@kubernetes/client-node'
|
||||||
import { ContainerInfo, PodPhase, Registry } from 'hooklib'
|
import { ContainerInfo, Registry } from 'hooklib'
|
||||||
import * as stream from 'stream'
|
import * as stream from 'stream'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
|
||||||
import {
|
import {
|
||||||
getJobPodName,
|
getJobPodName,
|
||||||
getRunnerPodName,
|
getRunnerPodName,
|
||||||
|
getSecretName,
|
||||||
|
getStepPodName,
|
||||||
getVolumeClaimName,
|
getVolumeClaimName,
|
||||||
RunnerInstanceLabel
|
RunnerInstanceLabel
|
||||||
} from '../hooks/constants'
|
} from '../hooks/constants'
|
||||||
|
import { PodPhase } from './utils'
|
||||||
|
|
||||||
const kc = new k8s.KubeConfig()
|
const kc = new k8s.KubeConfig()
|
||||||
|
|
||||||
@@ -43,16 +46,15 @@ export const requiredPermissions = [
|
|||||||
verbs: ['get', 'list', 'create', 'delete'],
|
verbs: ['get', 'list', 'create', 'delete'],
|
||||||
resource: 'jobs',
|
resource: 'jobs',
|
||||||
subresource: ''
|
subresource: ''
|
||||||
|
},
|
||||||
|
{
|
||||||
|
group: '',
|
||||||
|
verbs: ['create', 'delete', 'get', 'list'],
|
||||||
|
resource: 'secrets',
|
||||||
|
subresource: ''
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
|
|
||||||
const secretPermission = {
|
|
||||||
group: '',
|
|
||||||
verbs: ['get', 'list', 'create', 'delete'],
|
|
||||||
resource: 'secrets',
|
|
||||||
subresource: ''
|
|
||||||
}
|
|
||||||
|
|
||||||
export async function createPod(
|
export async function createPod(
|
||||||
jobContainer?: k8s.V1Container,
|
jobContainer?: k8s.V1Container,
|
||||||
services?: k8s.V1Container[],
|
services?: k8s.V1Container[],
|
||||||
@@ -92,19 +94,13 @@ export async function createPod(
|
|||||||
]
|
]
|
||||||
|
|
||||||
if (registry) {
|
if (registry) {
|
||||||
if (await isSecretsAuthOK()) {
|
const secret = await createDockerSecret(registry)
|
||||||
const secret = await createDockerSecret(registry)
|
if (!secret?.metadata?.name) {
|
||||||
if (!secret?.metadata?.name) {
|
throw new Error(`created secret does not have secret.metadata.name`)
|
||||||
throw new Error(`created secret does not have secret.metadata.name`)
|
|
||||||
}
|
|
||||||
const secretReference = new k8s.V1LocalObjectReference()
|
|
||||||
secretReference.name = secret.metadata.name
|
|
||||||
appPod.spec.imagePullSecrets = [secretReference]
|
|
||||||
} else {
|
|
||||||
throw new Error(
|
|
||||||
`Pulls from private registry is not allowed. Please contact your self hosted runner administrator. Service account needs permissions for ${secretPermission.verbs} in resource ${secretPermission.resource}`
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
const secretReference = new k8s.V1LocalObjectReference()
|
||||||
|
secretReference.name = secret.metadata.name
|
||||||
|
appPod.spec.imagePullSecrets = [secretReference]
|
||||||
}
|
}
|
||||||
|
|
||||||
const { body } = await k8sApi.createNamespacedPod(namespace(), appPod)
|
const { body } = await k8sApi.createNamespacedPod(namespace(), appPod)
|
||||||
@@ -114,13 +110,14 @@ export async function createPod(
|
|||||||
export async function createJob(
|
export async function createJob(
|
||||||
container: k8s.V1Container
|
container: k8s.V1Container
|
||||||
): Promise<k8s.V1Job> {
|
): Promise<k8s.V1Job> {
|
||||||
const job = new k8s.V1Job()
|
const runnerInstanceLabel = new RunnerInstanceLabel()
|
||||||
|
|
||||||
|
const job = new k8s.V1Job()
|
||||||
job.apiVersion = 'batch/v1'
|
job.apiVersion = 'batch/v1'
|
||||||
job.kind = 'Job'
|
job.kind = 'Job'
|
||||||
job.metadata = new k8s.V1ObjectMeta()
|
job.metadata = new k8s.V1ObjectMeta()
|
||||||
job.metadata.name = getJobPodName()
|
job.metadata.name = getStepPodName()
|
||||||
job.metadata.labels = { 'runner-pod': getRunnerPodName() }
|
job.metadata.labels = { [runnerInstanceLabel.key]: runnerInstanceLabel.value }
|
||||||
|
|
||||||
job.spec = new k8s.V1JobSpec()
|
job.spec = new k8s.V1JobSpec()
|
||||||
job.spec.ttlSecondsAfterFinished = 300
|
job.spec.ttlSecondsAfterFinished = 300
|
||||||
@@ -132,7 +129,7 @@ export async function createJob(
|
|||||||
job.spec.template.spec.restartPolicy = 'Never'
|
job.spec.template.spec.restartPolicy = 'Never'
|
||||||
job.spec.template.spec.nodeName = await getCurrentNodeName()
|
job.spec.template.spec.nodeName = await getCurrentNodeName()
|
||||||
|
|
||||||
const claimName = `${runnerName()}-work`
|
const claimName = getVolumeClaimName()
|
||||||
job.spec.template.spec.volumes = [
|
job.spec.template.spec.volumes = [
|
||||||
{
|
{
|
||||||
name: 'work',
|
name: 'work',
|
||||||
@@ -173,7 +170,13 @@ export async function getContainerJobPodName(jobName: string): Promise<string> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export async function deletePod(podName: string): Promise<void> {
|
export async function deletePod(podName: string): Promise<void> {
|
||||||
await k8sApi.deleteNamespacedPod(podName, namespace())
|
await k8sApi.deleteNamespacedPod(
|
||||||
|
podName,
|
||||||
|
namespace(),
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
0
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function execPodStep(
|
export async function execPodStep(
|
||||||
@@ -182,36 +185,32 @@ export async function execPodStep(
|
|||||||
containerName: string,
|
containerName: string,
|
||||||
stdin?: stream.Readable
|
stdin?: stream.Readable
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
// TODO, we need to add the path from `prependPath` to the PATH variable. How can we do that? Maybe another exec before running this one?
|
|
||||||
// Maybe something like, get the current path, if these entries aren't in it, add them, then set the current path to that?
|
|
||||||
|
|
||||||
// TODO: how do we set working directory? There doesn't seem to be an easy way to do it. Should we cd then execute our bash script?
|
|
||||||
const exec = new k8s.Exec(kc)
|
const exec = new k8s.Exec(kc)
|
||||||
return new Promise(async function (resolve, reject) {
|
await new Promise(async function (resolve, reject) {
|
||||||
try {
|
await exec.exec(
|
||||||
await exec.exec(
|
namespace(),
|
||||||
namespace(),
|
podName,
|
||||||
podName,
|
containerName,
|
||||||
containerName,
|
command,
|
||||||
command,
|
process.stdout,
|
||||||
process.stdout,
|
process.stderr,
|
||||||
process.stderr,
|
stdin ?? null,
|
||||||
stdin ?? null,
|
false /* tty */,
|
||||||
false /* tty */,
|
resp => {
|
||||||
resp => {
|
// kube.exec returns an error if exit code is not 0, but we can't actually get the exit code
|
||||||
// kube.exec returns an error if exit code is not 0, but we can't actually get the exit code
|
if (resp.status === 'Success') {
|
||||||
if (resp.status === 'Success') {
|
resolve(resp.code)
|
||||||
resolve()
|
} else {
|
||||||
} else {
|
core.debug(
|
||||||
reject(
|
JSON.stringify({
|
||||||
JSON.stringify({ message: resp?.message, details: resp?.details })
|
message: resp?.message,
|
||||||
)
|
details: resp?.details
|
||||||
}
|
})
|
||||||
|
)
|
||||||
|
reject(resp?.message)
|
||||||
}
|
}
|
||||||
)
|
}
|
||||||
} catch (error) {
|
)
|
||||||
reject(error)
|
|
||||||
}
|
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -234,46 +233,100 @@ export async function createDockerSecret(
|
|||||||
): Promise<k8s.V1Secret> {
|
): Promise<k8s.V1Secret> {
|
||||||
const authContent = {
|
const authContent = {
|
||||||
auths: {
|
auths: {
|
||||||
[registry.serverUrl]: {
|
[registry.serverUrl || 'https://index.docker.io/v1/']: {
|
||||||
username: registry.username,
|
username: registry.username,
|
||||||
password: registry.password,
|
password: registry.password,
|
||||||
auth: Buffer.from(
|
auth: Buffer.from(`${registry.username}:${registry.password}`).toString(
|
||||||
`${registry.username}:${registry.password}`,
|
|
||||||
'base64'
|
'base64'
|
||||||
).toString()
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
const secretName = generateSecretName()
|
|
||||||
|
const runnerInstanceLabel = new RunnerInstanceLabel()
|
||||||
|
|
||||||
|
const secretName = getSecretName()
|
||||||
const secret = new k8s.V1Secret()
|
const secret = new k8s.V1Secret()
|
||||||
secret.immutable = true
|
secret.immutable = true
|
||||||
secret.apiVersion = 'v1'
|
secret.apiVersion = 'v1'
|
||||||
secret.metadata = new k8s.V1ObjectMeta()
|
secret.metadata = new k8s.V1ObjectMeta()
|
||||||
secret.metadata.name = secretName
|
secret.metadata.name = secretName
|
||||||
|
secret.metadata.namespace = namespace()
|
||||||
|
secret.metadata.labels = {
|
||||||
|
[runnerInstanceLabel.key]: runnerInstanceLabel.value
|
||||||
|
}
|
||||||
|
secret.type = 'kubernetes.io/dockerconfigjson'
|
||||||
secret.kind = 'Secret'
|
secret.kind = 'Secret'
|
||||||
secret.data = {
|
secret.data = {
|
||||||
'.dockerconfigjson': Buffer.from(
|
'.dockerconfigjson': Buffer.from(JSON.stringify(authContent)).toString(
|
||||||
JSON.stringify(authContent),
|
|
||||||
'base64'
|
'base64'
|
||||||
).toString()
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
const { body } = await k8sApi.createNamespacedSecret(namespace(), secret)
|
const { body } = await k8sApi.createNamespacedSecret(namespace(), secret)
|
||||||
return body
|
return body
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export async function createSecretForEnvs(envs: {
|
||||||
|
[key: string]: string
|
||||||
|
}): Promise<string> {
|
||||||
|
const runnerInstanceLabel = new RunnerInstanceLabel()
|
||||||
|
|
||||||
|
const secret = new k8s.V1Secret()
|
||||||
|
const secretName = getSecretName()
|
||||||
|
secret.immutable = true
|
||||||
|
secret.apiVersion = 'v1'
|
||||||
|
secret.metadata = new k8s.V1ObjectMeta()
|
||||||
|
secret.metadata.name = secretName
|
||||||
|
|
||||||
|
secret.metadata.labels = {
|
||||||
|
[runnerInstanceLabel.key]: runnerInstanceLabel.value
|
||||||
|
}
|
||||||
|
secret.kind = 'Secret'
|
||||||
|
secret.data = {}
|
||||||
|
for (const [key, value] of Object.entries(envs)) {
|
||||||
|
secret.data[key] = Buffer.from(value).toString('base64')
|
||||||
|
}
|
||||||
|
|
||||||
|
await k8sApi.createNamespacedSecret(namespace(), secret)
|
||||||
|
return secretName
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function deleteSecret(secretName: string): Promise<void> {
|
||||||
|
await k8sApi.deleteNamespacedSecret(secretName, namespace())
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function pruneSecrets(): Promise<void> {
|
||||||
|
const secretList = await k8sApi.listNamespacedSecret(
|
||||||
|
namespace(),
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
new RunnerInstanceLabel().toString()
|
||||||
|
)
|
||||||
|
if (!secretList.body.items.length) {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
await Promise.all(
|
||||||
|
secretList.body.items.map(
|
||||||
|
secret => secret.metadata?.name && deleteSecret(secret.metadata.name)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
export async function waitForPodPhases(
|
export async function waitForPodPhases(
|
||||||
podName: string,
|
podName: string,
|
||||||
awaitingPhases: Set<PodPhase>,
|
awaitingPhases: Set<PodPhase>,
|
||||||
backOffPhases: Set<PodPhase>,
|
backOffPhases: Set<PodPhase>,
|
||||||
maxTimeSeconds = 45 * 60 // 45 min
|
maxTimeSeconds = 10 * 60 // 10 min
|
||||||
): Promise<void> {
|
): Promise<void> {
|
||||||
const backOffManager = new BackOffManager(maxTimeSeconds)
|
const backOffManager = new BackOffManager(maxTimeSeconds)
|
||||||
let phase: PodPhase = PodPhase.UNKNOWN
|
let phase: PodPhase = PodPhase.UNKNOWN
|
||||||
try {
|
try {
|
||||||
while (true) {
|
while (true) {
|
||||||
phase = await getPodPhase(podName)
|
phase = await getPodPhase(podName)
|
||||||
|
|
||||||
if (awaitingPhases.has(phase)) {
|
if (awaitingPhases.has(phase)) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
@@ -304,7 +357,7 @@ async function getPodPhase(podName: string): Promise<PodPhase> {
|
|||||||
if (!pod.status?.phase || !podPhaseLookup.has(pod.status.phase)) {
|
if (!pod.status?.phase || !podPhaseLookup.has(pod.status.phase)) {
|
||||||
return PodPhase.UNKNOWN
|
return PodPhase.UNKNOWN
|
||||||
}
|
}
|
||||||
return pod.status?.phase
|
return pod.status?.phase as PodPhase
|
||||||
}
|
}
|
||||||
|
|
||||||
async function isJobSucceeded(jobName: string): Promise<boolean> {
|
async function isJobSucceeded(jobName: string): Promise<boolean> {
|
||||||
@@ -328,7 +381,7 @@ export async function getPodLogs(
|
|||||||
})
|
})
|
||||||
|
|
||||||
logStream.on('error', err => {
|
logStream.on('error', err => {
|
||||||
process.stderr.write(JSON.stringify(err))
|
process.stderr.write(err.message)
|
||||||
})
|
})
|
||||||
|
|
||||||
const r = await log.log(namespace(), podName, containerName, logStream, {
|
const r = await log.log(namespace(), podName, containerName, logStream, {
|
||||||
@@ -340,7 +393,7 @@ export async function getPodLogs(
|
|||||||
await new Promise(resolve => r.on('close', () => resolve(null)))
|
await new Promise(resolve => r.on('close', () => resolve(null)))
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function podPrune(): Promise<void> {
|
export async function prunePods(): Promise<void> {
|
||||||
const podList = await k8sApi.listNamespacedPod(
|
const podList = await k8sApi.listNamespacedPod(
|
||||||
namespace(),
|
namespace(),
|
||||||
undefined,
|
undefined,
|
||||||
@@ -389,26 +442,6 @@ export async function isAuthPermissionsOK(): Promise<boolean> {
|
|||||||
return responses.every(resp => resp.body.status?.allowed)
|
return responses.every(resp => resp.body.status?.allowed)
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function isSecretsAuthOK(): Promise<boolean> {
|
|
||||||
const sar = new k8s.V1SelfSubjectAccessReview()
|
|
||||||
const asyncs: Promise<{
|
|
||||||
response: unknown
|
|
||||||
body: k8s.V1SelfSubjectAccessReview
|
|
||||||
}>[] = []
|
|
||||||
for (const verb of secretPermission.verbs) {
|
|
||||||
sar.spec = new k8s.V1SelfSubjectAccessReviewSpec()
|
|
||||||
sar.spec.resourceAttributes = new k8s.V1ResourceAttributes()
|
|
||||||
sar.spec.resourceAttributes.verb = verb
|
|
||||||
sar.spec.resourceAttributes.namespace = namespace()
|
|
||||||
sar.spec.resourceAttributes.group = secretPermission.group
|
|
||||||
sar.spec.resourceAttributes.resource = secretPermission.resource
|
|
||||||
sar.spec.resourceAttributes.subresource = secretPermission.subresource
|
|
||||||
asyncs.push(k8sAuthorizationV1Api.createSelfSubjectAccessReview(sar))
|
|
||||||
}
|
|
||||||
const responses = await Promise.all(asyncs)
|
|
||||||
return responses.every(resp => resp.body.status?.allowed)
|
|
||||||
}
|
|
||||||
|
|
||||||
export async function isPodContainerAlpine(
|
export async function isPodContainerAlpine(
|
||||||
podName: string,
|
podName: string,
|
||||||
containerName: string
|
containerName: string
|
||||||
@@ -454,20 +487,6 @@ export function namespace(): string {
|
|||||||
return context.namespace
|
return context.namespace
|
||||||
}
|
}
|
||||||
|
|
||||||
function generateSecretName(): string {
|
|
||||||
return `github-secret-${uuidv4()}`
|
|
||||||
}
|
|
||||||
|
|
||||||
function runnerName(): string {
|
|
||||||
const name = process.env.ACTIONS_RUNNER_POD_NAME
|
|
||||||
if (!name) {
|
|
||||||
throw new Error(
|
|
||||||
'Failed to determine runner name. "ACTIONS_RUNNER_POD_NAME" env variables should be set.'
|
|
||||||
)
|
|
||||||
}
|
|
||||||
return name
|
|
||||||
}
|
|
||||||
|
|
||||||
class BackOffManager {
|
class BackOffManager {
|
||||||
private backOffSeconds = 1
|
private backOffSeconds = 1
|
||||||
totalTime = 0
|
totalTime = 0
|
||||||
@@ -497,27 +516,37 @@ class BackOffManager {
|
|||||||
export function containerPorts(
|
export function containerPorts(
|
||||||
container: ContainerInfo
|
container: ContainerInfo
|
||||||
): k8s.V1ContainerPort[] {
|
): k8s.V1ContainerPort[] {
|
||||||
// 8080:8080/tcp
|
|
||||||
const portFormat = /(\d{1,5})(:(\d{1,5}))?(\/(tcp|udp))?/
|
|
||||||
|
|
||||||
const ports: k8s.V1ContainerPort[] = []
|
const ports: k8s.V1ContainerPort[] = []
|
||||||
for (const portDefinition of container.portMappings) {
|
for (const portDefinition of container.portMappings) {
|
||||||
const submatches = portFormat.exec(portDefinition)
|
const portProtoSplit = portDefinition.split('/')
|
||||||
if (!submatches) {
|
if (portProtoSplit.length > 2) {
|
||||||
throw new Error(
|
throw new Error(`Unexpected port format: ${portDefinition}`)
|
||||||
`Port definition "${portDefinition}" is in incorrect format`
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const port = new k8s.V1ContainerPort()
|
const port = new k8s.V1ContainerPort()
|
||||||
port.hostPort = Number(submatches[1])
|
port.protocol =
|
||||||
if (submatches[3]) {
|
portProtoSplit.length === 2 ? portProtoSplit[1].toUpperCase() : 'TCP'
|
||||||
port.containerPort = Number(submatches[3])
|
|
||||||
|
const portSplit = portProtoSplit[0].split(':')
|
||||||
|
if (portSplit.length > 2) {
|
||||||
|
throw new Error('ports should have at most one ":" separator')
|
||||||
}
|
}
|
||||||
if (submatches[5]) {
|
|
||||||
port.protocol = submatches[5].toUpperCase()
|
const parsePort = (p: string): number => {
|
||||||
|
const num = Number(p)
|
||||||
|
if (!Number.isInteger(num) || num < 1 || num > 65535) {
|
||||||
|
throw new Error(`invalid container port: ${p}`)
|
||||||
|
}
|
||||||
|
return num
|
||||||
|
}
|
||||||
|
|
||||||
|
if (portSplit.length === 1) {
|
||||||
|
port.containerPort = parsePort(portSplit[0])
|
||||||
} else {
|
} else {
|
||||||
port.protocol = 'TCP'
|
port.hostPort = parsePort(portSplit[0])
|
||||||
|
port.containerPort = parsePort(portSplit[1])
|
||||||
}
|
}
|
||||||
|
|
||||||
ports.push(port)
|
ports.push(port)
|
||||||
}
|
}
|
||||||
return ports
|
return ports
|
||||||
|
|||||||
@@ -1,6 +1,8 @@
|
|||||||
import * as k8s from '@kubernetes/client-node'
|
import * as k8s from '@kubernetes/client-node'
|
||||||
|
import * as fs from 'fs'
|
||||||
import { Mount } from 'hooklib'
|
import { Mount } from 'hooklib'
|
||||||
import * as path from 'path'
|
import * as path from 'path'
|
||||||
|
import { v1 as uuidv4 } from 'uuid'
|
||||||
import { POD_VOLUME_NAME } from './index'
|
import { POD_VOLUME_NAME } from './index'
|
||||||
|
|
||||||
export const DEFAULT_CONTAINER_ENTRY_POINT_ARGS = [`-f`, `/dev/null`]
|
export const DEFAULT_CONTAINER_ENTRY_POINT_ARGS = [`-f`, `/dev/null`]
|
||||||
@@ -8,7 +10,8 @@ export const DEFAULT_CONTAINER_ENTRY_POINT = 'tail'
|
|||||||
|
|
||||||
export function containerVolumes(
|
export function containerVolumes(
|
||||||
userMountVolumes: Mount[] = [],
|
userMountVolumes: Mount[] = [],
|
||||||
jobContainer = true
|
jobContainer = true,
|
||||||
|
containerAction = false
|
||||||
): k8s.V1VolumeMount[] {
|
): k8s.V1VolumeMount[] {
|
||||||
const mounts: k8s.V1VolumeMount[] = [
|
const mounts: k8s.V1VolumeMount[] = [
|
||||||
{
|
{
|
||||||
@@ -17,6 +20,25 @@ export function containerVolumes(
|
|||||||
}
|
}
|
||||||
]
|
]
|
||||||
|
|
||||||
|
const workspacePath = process.env.GITHUB_WORKSPACE as string
|
||||||
|
if (containerAction) {
|
||||||
|
const i = workspacePath.lastIndexOf('_work/')
|
||||||
|
const workspaceRelativePath = workspacePath.slice(i + '_work/'.length)
|
||||||
|
mounts.push(
|
||||||
|
{
|
||||||
|
name: POD_VOLUME_NAME,
|
||||||
|
mountPath: '/github/workspace',
|
||||||
|
subPath: workspaceRelativePath
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: POD_VOLUME_NAME,
|
||||||
|
mountPath: '/github/file_commands',
|
||||||
|
subPath: '_temp/_runner_file_commands'
|
||||||
|
}
|
||||||
|
)
|
||||||
|
return mounts
|
||||||
|
}
|
||||||
|
|
||||||
if (!jobContainer) {
|
if (!jobContainer) {
|
||||||
return mounts
|
return mounts
|
||||||
}
|
}
|
||||||
@@ -44,14 +66,20 @@ export function containerVolumes(
|
|||||||
}
|
}
|
||||||
|
|
||||||
for (const userVolume of userMountVolumes) {
|
for (const userVolume of userMountVolumes) {
|
||||||
const sourceVolumePath = `${
|
let sourceVolumePath = ''
|
||||||
path.isAbsolute(userVolume.sourceVolumePath)
|
if (path.isAbsolute(userVolume.sourceVolumePath)) {
|
||||||
? userVolume.sourceVolumePath
|
if (!userVolume.sourceVolumePath.startsWith(workspacePath)) {
|
||||||
: path.join(
|
throw new Error(
|
||||||
process.env.GITHUB_WORKSPACE as string,
|
'Volume mounts outside of the work folder are not supported'
|
||||||
userVolume.sourceVolumePath
|
)
|
||||||
)
|
}
|
||||||
}`
|
// source volume path should be relative path
|
||||||
|
sourceVolumePath = userVolume.sourceVolumePath.slice(
|
||||||
|
workspacePath.length + 1
|
||||||
|
)
|
||||||
|
} else {
|
||||||
|
sourceVolumePath = userVolume.sourceVolumePath
|
||||||
|
}
|
||||||
|
|
||||||
mounts.push({
|
mounts.push({
|
||||||
name: POD_VOLUME_NAME,
|
name: POD_VOLUME_NAME,
|
||||||
@@ -63,3 +91,57 @@ export function containerVolumes(
|
|||||||
|
|
||||||
return mounts
|
return mounts
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function writeEntryPointScript(
|
||||||
|
workingDirectory: string,
|
||||||
|
entryPoint: string,
|
||||||
|
entryPointArgs?: string[],
|
||||||
|
prependPath?: string[],
|
||||||
|
environmentVariables?: { [key: string]: string }
|
||||||
|
): { containerPath: string; runnerPath: string } {
|
||||||
|
let exportPath = ''
|
||||||
|
if (prependPath?.length) {
|
||||||
|
// TODO: remove compatibility with typeof prependPath === 'string' as we bump to next major version, the hooks will lose PrependPath compat with runners 2.293.0 and older
|
||||||
|
const prepend =
|
||||||
|
typeof prependPath === 'string' ? prependPath : prependPath.join(':')
|
||||||
|
exportPath = `export PATH=${prepend}:$PATH`
|
||||||
|
}
|
||||||
|
let environmentPrefix = ''
|
||||||
|
|
||||||
|
if (environmentVariables && Object.entries(environmentVariables).length) {
|
||||||
|
const envBuffer: string[] = []
|
||||||
|
for (const [key, value] of Object.entries(environmentVariables)) {
|
||||||
|
envBuffer.push(
|
||||||
|
`"${key}=${value
|
||||||
|
.replace(/\\/g, '\\\\')
|
||||||
|
.replace(/"/g, '\\"')
|
||||||
|
.replace(/=/g, '\\=')}"`
|
||||||
|
)
|
||||||
|
}
|
||||||
|
environmentPrefix = `env ${envBuffer.join(' ')} `
|
||||||
|
}
|
||||||
|
|
||||||
|
const content = `#!/bin/sh -l
|
||||||
|
${exportPath}
|
||||||
|
cd ${workingDirectory} && \
|
||||||
|
exec ${environmentPrefix} ${entryPoint} ${
|
||||||
|
entryPointArgs?.length ? entryPointArgs.join(' ') : ''
|
||||||
|
}
|
||||||
|
`
|
||||||
|
const filename = `${uuidv4()}.sh`
|
||||||
|
const entryPointPath = `${process.env.RUNNER_TEMP}/${filename}`
|
||||||
|
fs.writeFileSync(entryPointPath, content)
|
||||||
|
return {
|
||||||
|
containerPath: `/__w/_temp/${filename}`,
|
||||||
|
runnerPath: entryPointPath
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export enum PodPhase {
|
||||||
|
PENDING = 'Pending',
|
||||||
|
RUNNING = 'Running',
|
||||||
|
SUCCEEDED = 'Succeeded',
|
||||||
|
FAILED = 'Failed',
|
||||||
|
UNKNOWN = 'Unknown',
|
||||||
|
COMPLETED = 'Completed'
|
||||||
|
}
|
||||||
|
|||||||
@@ -1,31 +1,65 @@
|
|||||||
import * as path from 'path'
|
import * as k8s from '@kubernetes/client-node'
|
||||||
import * as fs from 'fs'
|
import { cleanupJob, prepareJob } from '../src/hooks'
|
||||||
import { prepareJob, cleanupJob } from '../src/hooks'
|
import { RunnerInstanceLabel } from '../src/hooks/constants'
|
||||||
import { TestTempOutput } from './test-setup'
|
import { namespace } from '../src/k8s'
|
||||||
|
import { TestHelper } from './test-setup'
|
||||||
|
|
||||||
let testTempOutput: TestTempOutput
|
let testHelper: TestHelper
|
||||||
|
|
||||||
const prepareJobJsonPath = path.resolve(
|
|
||||||
`${__dirname}/../../../examples/prepare-job.json`
|
|
||||||
)
|
|
||||||
|
|
||||||
let prepareJobOutputFilePath: string
|
|
||||||
|
|
||||||
describe('Cleanup Job', () => {
|
describe('Cleanup Job', () => {
|
||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
const prepareJobJson = fs.readFileSync(prepareJobJsonPath)
|
testHelper = new TestHelper()
|
||||||
let prepareJobData = JSON.parse(prepareJobJson.toString())
|
await testHelper.initialize()
|
||||||
|
let prepareJobData = testHelper.getPrepareJobDefinition()
|
||||||
testTempOutput = new TestTempOutput()
|
const prepareJobOutputFilePath = testHelper.createFile(
|
||||||
testTempOutput.initialize()
|
|
||||||
prepareJobOutputFilePath = testTempOutput.createFile(
|
|
||||||
'prepare-job-output.json'
|
'prepare-job-output.json'
|
||||||
)
|
)
|
||||||
await prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
await prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
||||||
})
|
})
|
||||||
|
|
||||||
|
afterEach(async () => {
|
||||||
|
await testHelper.cleanup()
|
||||||
|
})
|
||||||
|
|
||||||
it('should not throw', async () => {
|
it('should not throw', async () => {
|
||||||
const outputJson = fs.readFileSync(prepareJobOutputFilePath)
|
|
||||||
const outputData = JSON.parse(outputJson.toString())
|
|
||||||
await expect(cleanupJob()).resolves.not.toThrow()
|
await expect(cleanupJob()).resolves.not.toThrow()
|
||||||
})
|
})
|
||||||
|
|
||||||
|
it('should have no runner linked pods running', async () => {
|
||||||
|
await cleanupJob()
|
||||||
|
const kc = new k8s.KubeConfig()
|
||||||
|
|
||||||
|
kc.loadFromDefault()
|
||||||
|
const k8sApi = kc.makeApiClient(k8s.CoreV1Api)
|
||||||
|
|
||||||
|
const podList = await k8sApi.listNamespacedPod(
|
||||||
|
namespace(),
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
new RunnerInstanceLabel().toString()
|
||||||
|
)
|
||||||
|
|
||||||
|
expect(podList.body.items.length).toBe(0)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should have no runner linked secrets', async () => {
|
||||||
|
await cleanupJob()
|
||||||
|
const kc = new k8s.KubeConfig()
|
||||||
|
|
||||||
|
kc.loadFromDefault()
|
||||||
|
const k8sApi = kc.makeApiClient(k8s.CoreV1Api)
|
||||||
|
|
||||||
|
const secretList = await k8sApi.listNamespacedSecret(
|
||||||
|
namespace(),
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
new RunnerInstanceLabel().toString()
|
||||||
|
)
|
||||||
|
|
||||||
|
expect(secretList.body.items.length).toBe(0)
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
182
packages/k8s/tests/constants-test.ts
Normal file
182
packages/k8s/tests/constants-test.ts
Normal file
@@ -0,0 +1,182 @@
|
|||||||
|
import {
|
||||||
|
getJobPodName,
|
||||||
|
getRunnerPodName,
|
||||||
|
getSecretName,
|
||||||
|
getStepPodName,
|
||||||
|
getVolumeClaimName,
|
||||||
|
JOB_CONTAINER_NAME,
|
||||||
|
MAX_POD_NAME_LENGTH,
|
||||||
|
RunnerInstanceLabel,
|
||||||
|
STEP_POD_NAME_SUFFIX_LENGTH
|
||||||
|
} from '../src/hooks/constants'
|
||||||
|
|
||||||
|
describe('constants', () => {
|
||||||
|
describe('runner instance label', () => {
|
||||||
|
beforeEach(() => {
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = 'example'
|
||||||
|
})
|
||||||
|
it('should throw if ACTIONS_RUNNER_POD_NAME env is not set', () => {
|
||||||
|
delete process.env.ACTIONS_RUNNER_POD_NAME
|
||||||
|
expect(() => new RunnerInstanceLabel()).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should have key truthy', () => {
|
||||||
|
const runnerInstanceLabel = new RunnerInstanceLabel()
|
||||||
|
expect(typeof runnerInstanceLabel.key).toBe('string')
|
||||||
|
expect(runnerInstanceLabel.key).toBeTruthy()
|
||||||
|
expect(runnerInstanceLabel.key.length).toBeGreaterThan(0)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should have value as runner pod name', () => {
|
||||||
|
const name = process.env.ACTIONS_RUNNER_POD_NAME as string
|
||||||
|
const runnerInstanceLabel = new RunnerInstanceLabel()
|
||||||
|
expect(typeof runnerInstanceLabel.value).toBe('string')
|
||||||
|
expect(runnerInstanceLabel.value).toBe(name)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should have toString combination of key and value', () => {
|
||||||
|
const runnerInstanceLabel = new RunnerInstanceLabel()
|
||||||
|
expect(runnerInstanceLabel.toString()).toBe(
|
||||||
|
`${runnerInstanceLabel.key}=${runnerInstanceLabel.value}`
|
||||||
|
)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('getRunnerPodName', () => {
|
||||||
|
it('should throw if ACTIONS_RUNNER_POD_NAME env is not set', () => {
|
||||||
|
delete process.env.ACTIONS_RUNNER_POD_NAME
|
||||||
|
expect(() => getRunnerPodName()).toThrow()
|
||||||
|
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = ''
|
||||||
|
expect(() => getRunnerPodName()).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should return corrent ACTIONS_RUNNER_POD_NAME name', () => {
|
||||||
|
const name = 'example'
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = name
|
||||||
|
expect(getRunnerPodName()).toBe(name)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('getJobPodName', () => {
|
||||||
|
it('should throw on getJobPodName if ACTIONS_RUNNER_POD_NAME env is not set', () => {
|
||||||
|
delete process.env.ACTIONS_RUNNER_POD_NAME
|
||||||
|
expect(() => getJobPodName()).toThrow()
|
||||||
|
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = ''
|
||||||
|
expect(() => getRunnerPodName()).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should contain suffix -workflow', () => {
|
||||||
|
const tableTests = [
|
||||||
|
{
|
||||||
|
podName: 'test',
|
||||||
|
expect: 'test-workflow'
|
||||||
|
},
|
||||||
|
{
|
||||||
|
// podName.length == 63
|
||||||
|
podName:
|
||||||
|
'abcdaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa',
|
||||||
|
expect:
|
||||||
|
'abcdaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa-workflow'
|
||||||
|
}
|
||||||
|
]
|
||||||
|
|
||||||
|
for (const tt of tableTests) {
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = tt.podName
|
||||||
|
const actual = getJobPodName()
|
||||||
|
expect(actual).toBe(tt.expect)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('getVolumeClaimName', () => {
|
||||||
|
it('should throw if ACTIONS_RUNNER_POD_NAME env is not set', () => {
|
||||||
|
delete process.env.ACTIONS_RUNNER_CLAIM_NAME
|
||||||
|
delete process.env.ACTIONS_RUNNER_POD_NAME
|
||||||
|
expect(() => getVolumeClaimName()).toThrow()
|
||||||
|
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = ''
|
||||||
|
expect(() => getVolumeClaimName()).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should return ACTIONS_RUNNER_CLAIM_NAME env if set', () => {
|
||||||
|
const claimName = 'testclaim'
|
||||||
|
process.env.ACTIONS_RUNNER_CLAIM_NAME = claimName
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = 'example'
|
||||||
|
expect(getVolumeClaimName()).toBe(claimName)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should contain suffix -work if ACTIONS_RUNNER_CLAIM_NAME is not set', () => {
|
||||||
|
delete process.env.ACTIONS_RUNNER_CLAIM_NAME
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = 'example'
|
||||||
|
expect(getVolumeClaimName()).toBe('example-work')
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('getSecretName', () => {
|
||||||
|
it('should throw if ACTIONS_RUNNER_POD_NAME env is not set', () => {
|
||||||
|
delete process.env.ACTIONS_RUNNER_POD_NAME
|
||||||
|
expect(() => getSecretName()).toThrow()
|
||||||
|
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = ''
|
||||||
|
expect(() => getSecretName()).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should contain suffix -secret- and name trimmed', () => {
|
||||||
|
const podNames = [
|
||||||
|
'test',
|
||||||
|
'abcdaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'
|
||||||
|
]
|
||||||
|
|
||||||
|
for (const podName of podNames) {
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = podName
|
||||||
|
const actual = getSecretName()
|
||||||
|
const re = new RegExp(
|
||||||
|
`${podName.substring(
|
||||||
|
MAX_POD_NAME_LENGTH -
|
||||||
|
'-secret-'.length -
|
||||||
|
STEP_POD_NAME_SUFFIX_LENGTH
|
||||||
|
)}-secret-[a-z0-9]{8,}`
|
||||||
|
)
|
||||||
|
expect(actual).toMatch(re)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('getStepPodName', () => {
|
||||||
|
it('should throw if ACTIONS_RUNNER_POD_NAME env is not set', () => {
|
||||||
|
delete process.env.ACTIONS_RUNNER_POD_NAME
|
||||||
|
expect(() => getStepPodName()).toThrow()
|
||||||
|
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = ''
|
||||||
|
expect(() => getStepPodName()).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should contain suffix -step- and name trimmed', () => {
|
||||||
|
const podNames = [
|
||||||
|
'test',
|
||||||
|
'abcdaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'
|
||||||
|
]
|
||||||
|
|
||||||
|
for (const podName of podNames) {
|
||||||
|
process.env.ACTIONS_RUNNER_POD_NAME = podName
|
||||||
|
const actual = getStepPodName()
|
||||||
|
const re = new RegExp(
|
||||||
|
`${podName.substring(
|
||||||
|
MAX_POD_NAME_LENGTH - '-step-'.length - STEP_POD_NAME_SUFFIX_LENGTH
|
||||||
|
)}-step-[a-z0-9]{8,}`
|
||||||
|
)
|
||||||
|
expect(actual).toMatch(re)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('const values', () => {
|
||||||
|
it('should have constants set', () => {
|
||||||
|
expect(JOB_CONTAINER_NAME).toBeTruthy()
|
||||||
|
expect(MAX_POD_NAME_LENGTH).toBeGreaterThan(0)
|
||||||
|
expect(STEP_POD_NAME_SUFFIX_LENGTH).toBeGreaterThan(0)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
||||||
@@ -1,51 +1,36 @@
|
|||||||
import * as fs from 'fs'
|
import * as fs from 'fs'
|
||||||
import * as path from 'path'
|
|
||||||
import {
|
import {
|
||||||
cleanupJob,
|
cleanupJob,
|
||||||
prepareJob,
|
prepareJob,
|
||||||
runContainerStep,
|
runContainerStep,
|
||||||
runScriptStep
|
runScriptStep
|
||||||
} from '../src/hooks'
|
} from '../src/hooks'
|
||||||
import { TestTempOutput } from './test-setup'
|
import { TestHelper } from './test-setup'
|
||||||
|
|
||||||
jest.useRealTimers()
|
jest.useRealTimers()
|
||||||
|
|
||||||
let testTempOutput: TestTempOutput
|
let testHelper: TestHelper
|
||||||
|
|
||||||
const prepareJobJsonPath = path.resolve(
|
|
||||||
`${__dirname}/../../../../examples/prepare-job.json`
|
|
||||||
)
|
|
||||||
const runScriptStepJsonPath = path.resolve(
|
|
||||||
`${__dirname}/../../../../examples/run-script-step.json`
|
|
||||||
)
|
|
||||||
let runContainerStepJsonPath = path.resolve(
|
|
||||||
`${__dirname}/../../../../examples/run-container-step.json`
|
|
||||||
)
|
|
||||||
|
|
||||||
let prepareJobData: any
|
let prepareJobData: any
|
||||||
|
|
||||||
let prepareJobOutputFilePath: string
|
let prepareJobOutputFilePath: string
|
||||||
describe('e2e', () => {
|
describe('e2e', () => {
|
||||||
beforeEach(() => {
|
beforeEach(async () => {
|
||||||
const prepareJobJson = fs.readFileSync(prepareJobJsonPath)
|
testHelper = new TestHelper()
|
||||||
prepareJobData = JSON.parse(prepareJobJson.toString())
|
await testHelper.initialize()
|
||||||
|
|
||||||
testTempOutput = new TestTempOutput()
|
prepareJobData = testHelper.getPrepareJobDefinition()
|
||||||
testTempOutput.initialize()
|
prepareJobOutputFilePath = testHelper.createFile('prepare-job-output.json')
|
||||||
prepareJobOutputFilePath = testTempOutput.createFile(
|
|
||||||
'prepare-job-output.json'
|
|
||||||
)
|
|
||||||
})
|
})
|
||||||
afterEach(async () => {
|
afterEach(async () => {
|
||||||
testTempOutput.cleanup()
|
await testHelper.cleanup()
|
||||||
})
|
})
|
||||||
it('should prepare job, run script step, run container step then cleanup without errors', async () => {
|
it('should prepare job, run script step, run container step then cleanup without errors', async () => {
|
||||||
await expect(
|
await expect(
|
||||||
prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
|
|
||||||
const scriptStepContent = fs.readFileSync(runScriptStepJsonPath)
|
const scriptStepData = testHelper.getRunScriptStepDefinition()
|
||||||
const scriptStepData = JSON.parse(scriptStepContent.toString())
|
|
||||||
|
|
||||||
const prepareJobOutputJson = fs.readFileSync(prepareJobOutputFilePath)
|
const prepareJobOutputJson = fs.readFileSync(prepareJobOutputFilePath)
|
||||||
const prepareJobOutputData = JSON.parse(prepareJobOutputJson.toString())
|
const prepareJobOutputData = JSON.parse(prepareJobOutputJson.toString())
|
||||||
@@ -54,8 +39,7 @@ describe('e2e', () => {
|
|||||||
runScriptStep(scriptStepData.args, prepareJobOutputData.state, null)
|
runScriptStep(scriptStepData.args, prepareJobOutputData.state, null)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
|
|
||||||
const runContainerStepContent = fs.readFileSync(runContainerStepJsonPath)
|
const runContainerStepData = testHelper.getRunContainerStepDefinition()
|
||||||
const runContainerStepData = JSON.parse(runContainerStepContent.toString())
|
|
||||||
|
|
||||||
await expect(
|
await expect(
|
||||||
runContainerStep(runContainerStepData.args)
|
runContainerStep(runContainerStepData.args)
|
||||||
|
|||||||
224
packages/k8s/tests/k8s-utils-test.ts
Normal file
224
packages/k8s/tests/k8s-utils-test.ts
Normal file
@@ -0,0 +1,224 @@
|
|||||||
|
import * as fs from 'fs'
|
||||||
|
import { containerPorts, POD_VOLUME_NAME } from '../src/k8s'
|
||||||
|
import { containerVolumes, writeEntryPointScript } from '../src/k8s/utils'
|
||||||
|
import { TestHelper } from './test-setup'
|
||||||
|
|
||||||
|
let testHelper: TestHelper
|
||||||
|
|
||||||
|
describe('k8s utils', () => {
|
||||||
|
describe('write entrypoint', () => {
|
||||||
|
beforeEach(async () => {
|
||||||
|
testHelper = new TestHelper()
|
||||||
|
await testHelper.initialize()
|
||||||
|
})
|
||||||
|
|
||||||
|
afterEach(async () => {
|
||||||
|
await testHelper.cleanup()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should not throw', () => {
|
||||||
|
expect(() =>
|
||||||
|
writeEntryPointScript(
|
||||||
|
'/test',
|
||||||
|
'sh',
|
||||||
|
['-e', 'script.sh'],
|
||||||
|
['/prepend/path'],
|
||||||
|
{
|
||||||
|
SOME_ENV: 'SOME_VALUE'
|
||||||
|
}
|
||||||
|
)
|
||||||
|
).not.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should throw if RUNNER_TEMP is not set', () => {
|
||||||
|
delete process.env.RUNNER_TEMP
|
||||||
|
expect(() =>
|
||||||
|
writeEntryPointScript(
|
||||||
|
'/test',
|
||||||
|
'sh',
|
||||||
|
['-e', 'script.sh'],
|
||||||
|
['/prepend/path'],
|
||||||
|
{
|
||||||
|
SOME_ENV: 'SOME_VALUE'
|
||||||
|
}
|
||||||
|
)
|
||||||
|
).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should return object with containerPath and runnerPath', () => {
|
||||||
|
const { containerPath, runnerPath } = writeEntryPointScript(
|
||||||
|
'/test',
|
||||||
|
'sh',
|
||||||
|
['-e', 'script.sh'],
|
||||||
|
['/prepend/path'],
|
||||||
|
{
|
||||||
|
SOME_ENV: 'SOME_VALUE'
|
||||||
|
}
|
||||||
|
)
|
||||||
|
expect(containerPath).toMatch(/\/__w\/_temp\/.*\.sh/)
|
||||||
|
const re = new RegExp(`${process.env.RUNNER_TEMP}/.*\\.sh`)
|
||||||
|
expect(runnerPath).toMatch(re)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should write entrypoint path and the file should exist', () => {
|
||||||
|
const { runnerPath } = writeEntryPointScript(
|
||||||
|
'/test',
|
||||||
|
'sh',
|
||||||
|
['-e', 'script.sh'],
|
||||||
|
['/prepend/path'],
|
||||||
|
{
|
||||||
|
SOME_ENV: 'SOME_VALUE'
|
||||||
|
}
|
||||||
|
)
|
||||||
|
expect(fs.existsSync(runnerPath)).toBe(true)
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
describe('container volumes', () => {
|
||||||
|
beforeEach(async () => {
|
||||||
|
testHelper = new TestHelper()
|
||||||
|
await testHelper.initialize()
|
||||||
|
})
|
||||||
|
|
||||||
|
afterEach(async () => {
|
||||||
|
await testHelper.cleanup()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should throw if container action and GITHUB_WORKSPACE env is not set', () => {
|
||||||
|
delete process.env.GITHUB_WORKSPACE
|
||||||
|
expect(() => containerVolumes([], true, true)).toThrow()
|
||||||
|
expect(() => containerVolumes([], false, true)).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should always have work mount', () => {
|
||||||
|
let volumes = containerVolumes([], true, true)
|
||||||
|
expect(volumes.find(e => e.mountPath === '/__w')).toBeTruthy()
|
||||||
|
volumes = containerVolumes([], true, false)
|
||||||
|
expect(volumes.find(e => e.mountPath === '/__w')).toBeTruthy()
|
||||||
|
volumes = containerVolumes([], false, true)
|
||||||
|
expect(volumes.find(e => e.mountPath === '/__w')).toBeTruthy()
|
||||||
|
volumes = containerVolumes([], false, false)
|
||||||
|
expect(volumes.find(e => e.mountPath === '/__w')).toBeTruthy()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should have container action volumes', () => {
|
||||||
|
let volumes = containerVolumes([], true, true)
|
||||||
|
let workspace = volumes.find(e => e.mountPath === '/github/workspace')
|
||||||
|
let fileCommands = volumes.find(
|
||||||
|
e => e.mountPath === '/github/file_commands'
|
||||||
|
)
|
||||||
|
expect(workspace).toBeTruthy()
|
||||||
|
expect(workspace?.subPath).toBe('repo/repo')
|
||||||
|
expect(fileCommands).toBeTruthy()
|
||||||
|
expect(fileCommands?.subPath).toBe('_temp/_runner_file_commands')
|
||||||
|
|
||||||
|
volumes = containerVolumes([], false, true)
|
||||||
|
workspace = volumes.find(e => e.mountPath === '/github/workspace')
|
||||||
|
fileCommands = volumes.find(e => e.mountPath === '/github/file_commands')
|
||||||
|
expect(workspace).toBeTruthy()
|
||||||
|
expect(workspace?.subPath).toBe('repo/repo')
|
||||||
|
expect(fileCommands).toBeTruthy()
|
||||||
|
expect(fileCommands?.subPath).toBe('_temp/_runner_file_commands')
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should have externals, github home and github workflow mounts if job container', () => {
|
||||||
|
const volumes = containerVolumes()
|
||||||
|
expect(volumes.find(e => e.mountPath === '/__e')).toBeTruthy()
|
||||||
|
expect(volumes.find(e => e.mountPath === '/github/home')).toBeTruthy()
|
||||||
|
expect(volumes.find(e => e.mountPath === '/github/workflow')).toBeTruthy()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should throw if user volume source volume path is not in workspace', () => {
|
||||||
|
expect(() =>
|
||||||
|
containerVolumes(
|
||||||
|
[
|
||||||
|
{
|
||||||
|
sourceVolumePath: '/outside/of/workdir'
|
||||||
|
}
|
||||||
|
],
|
||||||
|
true,
|
||||||
|
false
|
||||||
|
)
|
||||||
|
).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it(`all volumes should have name ${POD_VOLUME_NAME}`, () => {
|
||||||
|
let volumes = containerVolumes([], true, true)
|
||||||
|
expect(volumes.every(e => e.name === POD_VOLUME_NAME)).toBeTruthy()
|
||||||
|
volumes = containerVolumes([], true, false)
|
||||||
|
expect(volumes.every(e => e.name === POD_VOLUME_NAME)).toBeTruthy()
|
||||||
|
volumes = containerVolumes([], false, true)
|
||||||
|
expect(volumes.every(e => e.name === POD_VOLUME_NAME)).toBeTruthy()
|
||||||
|
volumes = containerVolumes([], false, false)
|
||||||
|
expect(volumes.every(e => e.name === POD_VOLUME_NAME)).toBeTruthy()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should parse container ports', () => {
|
||||||
|
const tt = [
|
||||||
|
{
|
||||||
|
spec: '8080:80',
|
||||||
|
want: {
|
||||||
|
containerPort: 80,
|
||||||
|
hostPort: 8080,
|
||||||
|
protocol: 'TCP'
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
spec: '8080:80/udp',
|
||||||
|
want: {
|
||||||
|
containerPort: 80,
|
||||||
|
hostPort: 8080,
|
||||||
|
protocol: 'UDP'
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
spec: '8080/udp',
|
||||||
|
want: {
|
||||||
|
containerPort: 8080,
|
||||||
|
hostPort: undefined,
|
||||||
|
protocol: 'UDP'
|
||||||
|
}
|
||||||
|
},
|
||||||
|
{
|
||||||
|
spec: '8080',
|
||||||
|
want: {
|
||||||
|
containerPort: 8080,
|
||||||
|
hostPort: undefined,
|
||||||
|
protocol: 'TCP'
|
||||||
|
}
|
||||||
|
}
|
||||||
|
]
|
||||||
|
|
||||||
|
for (const tc of tt) {
|
||||||
|
const got = containerPorts({ portMappings: [tc.spec] })
|
||||||
|
for (const [key, value] of Object.entries(tc.want)) {
|
||||||
|
expect(got[0][key]).toBe(value)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should throw when ports are out of range (0, 65536)', () => {
|
||||||
|
expect(() => containerPorts({ portMappings: ['65536'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['0'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['65536/udp'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['0/udp'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['1:65536'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['65536:1'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['1:65536/tcp'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['65536:1/tcp'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['1:'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: [':1'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['1:/tcp'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: [':1/tcp'] })).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should throw on multi ":" splits', () => {
|
||||||
|
expect(() => containerPorts({ portMappings: ['1:1:1'] })).toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should throw on multi "/" splits', () => {
|
||||||
|
expect(() => containerPorts({ portMappings: ['1:1/tcp/udp'] })).toThrow()
|
||||||
|
expect(() => containerPorts({ portMappings: ['1/tcp/udp'] })).toThrow()
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
||||||
@@ -2,35 +2,26 @@ import * as fs from 'fs'
|
|||||||
import * as path from 'path'
|
import * as path from 'path'
|
||||||
import { cleanupJob } from '../src/hooks'
|
import { cleanupJob } from '../src/hooks'
|
||||||
import { prepareJob } from '../src/hooks/prepare-job'
|
import { prepareJob } from '../src/hooks/prepare-job'
|
||||||
import { TestTempOutput } from './test-setup'
|
import { TestHelper } from './test-setup'
|
||||||
|
|
||||||
jest.useRealTimers()
|
jest.useRealTimers()
|
||||||
|
|
||||||
let testTempOutput: TestTempOutput
|
let testHelper: TestHelper
|
||||||
|
|
||||||
const prepareJobJsonPath = path.resolve(
|
|
||||||
`${__dirname}/../../../examples/prepare-job.json`
|
|
||||||
)
|
|
||||||
let prepareJobData: any
|
let prepareJobData: any
|
||||||
|
|
||||||
let prepareJobOutputFilePath: string
|
let prepareJobOutputFilePath: string
|
||||||
|
|
||||||
describe('Prepare job', () => {
|
describe('Prepare job', () => {
|
||||||
beforeEach(() => {
|
beforeEach(async () => {
|
||||||
const prepareJobJson = fs.readFileSync(prepareJobJsonPath)
|
testHelper = new TestHelper()
|
||||||
prepareJobData = JSON.parse(prepareJobJson.toString())
|
await testHelper.initialize()
|
||||||
|
prepareJobData = testHelper.getPrepareJobDefinition()
|
||||||
testTempOutput = new TestTempOutput()
|
prepareJobOutputFilePath = testHelper.createFile('prepare-job-output.json')
|
||||||
testTempOutput.initialize()
|
|
||||||
prepareJobOutputFilePath = testTempOutput.createFile(
|
|
||||||
'prepare-job-output.json'
|
|
||||||
)
|
|
||||||
})
|
})
|
||||||
afterEach(async () => {
|
afterEach(async () => {
|
||||||
const outputJson = fs.readFileSync(prepareJobOutputFilePath)
|
|
||||||
const outputData = JSON.parse(outputJson.toString())
|
|
||||||
await cleanupJob()
|
await cleanupJob()
|
||||||
testTempOutput.cleanup()
|
await testHelper.cleanup()
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should not throw exception', async () => {
|
it('should not throw exception', async () => {
|
||||||
@@ -44,4 +35,40 @@ describe('Prepare job', () => {
|
|||||||
const content = fs.readFileSync(prepareJobOutputFilePath)
|
const content = fs.readFileSync(prepareJobOutputFilePath)
|
||||||
expect(() => JSON.parse(content.toString())).not.toThrow()
|
expect(() => JSON.parse(content.toString())).not.toThrow()
|
||||||
})
|
})
|
||||||
|
|
||||||
|
it('should prepare job with absolute path for userVolumeMount', async () => {
|
||||||
|
prepareJobData.args.container.userMountVolumes = [
|
||||||
|
{
|
||||||
|
sourceVolumePath: path.join(
|
||||||
|
process.env.GITHUB_WORKSPACE as string,
|
||||||
|
'/myvolume'
|
||||||
|
),
|
||||||
|
targetVolumePath: '/volume_mount',
|
||||||
|
readOnly: false
|
||||||
|
}
|
||||||
|
]
|
||||||
|
await expect(
|
||||||
|
prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should throw an exception if the user volume mount is absolute path outside of GITHUB_WORKSPACE', async () => {
|
||||||
|
prepareJobData.args.container.userMountVolumes = [
|
||||||
|
{
|
||||||
|
sourceVolumePath: '/somewhere/not/in/gh-workspace',
|
||||||
|
targetVolumePath: '/containermount',
|
||||||
|
readOnly: false
|
||||||
|
}
|
||||||
|
]
|
||||||
|
await expect(
|
||||||
|
prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
||||||
|
).rejects.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should not run prepare job without the job container', async () => {
|
||||||
|
prepareJobData.args.container = undefined
|
||||||
|
await expect(
|
||||||
|
prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
||||||
|
).rejects.toThrow()
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -1,25 +1,39 @@
|
|||||||
import { TestTempOutput } from './test-setup'
|
|
||||||
import * as path from 'path'
|
|
||||||
import { runContainerStep } from '../src/hooks'
|
import { runContainerStep } from '../src/hooks'
|
||||||
import * as fs from 'fs'
|
import { TestHelper } from './test-setup'
|
||||||
|
|
||||||
jest.useRealTimers()
|
jest.useRealTimers()
|
||||||
|
|
||||||
let testTempOutput: TestTempOutput
|
let testHelper: TestHelper
|
||||||
|
|
||||||
let runContainerStepJsonPath = path.resolve(
|
|
||||||
`${__dirname}/../../../examples/run-container-step.json`
|
|
||||||
)
|
|
||||||
|
|
||||||
let runContainerStepData: any
|
let runContainerStepData: any
|
||||||
|
|
||||||
describe('Run container step', () => {
|
describe('Run container step', () => {
|
||||||
beforeAll(() => {
|
beforeEach(async () => {
|
||||||
const content = fs.readFileSync(runContainerStepJsonPath)
|
testHelper = new TestHelper()
|
||||||
runContainerStepData = JSON.parse(content.toString())
|
await testHelper.initialize()
|
||||||
process.env.RUNNER_NAME = 'testjob'
|
runContainerStepData = testHelper.getRunContainerStepDefinition()
|
||||||
})
|
})
|
||||||
|
|
||||||
|
afterEach(async () => {
|
||||||
|
await testHelper.cleanup()
|
||||||
|
})
|
||||||
|
|
||||||
it('should not throw', async () => {
|
it('should not throw', async () => {
|
||||||
|
const exitCode = await runContainerStep(runContainerStepData.args)
|
||||||
|
expect(exitCode).toBe(0)
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should fail if the working directory does not exist', async () => {
|
||||||
|
runContainerStepData.args.workingDirectory = '/foo/bar'
|
||||||
|
await expect(runContainerStep(runContainerStepData.args)).rejects.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should shold have env variables available', async () => {
|
||||||
|
runContainerStepData.args.entryPoint = 'bash'
|
||||||
|
runContainerStepData.args.entryPointArgs = [
|
||||||
|
'-c',
|
||||||
|
"'if [[ -z $NODE_ENV ]]; then exit 1; fi'"
|
||||||
|
]
|
||||||
await expect(
|
await expect(
|
||||||
runContainerStep(runContainerStepData.args)
|
runContainerStep(runContainerStepData.args)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
|
|||||||
@@ -1,31 +1,26 @@
|
|||||||
import { prepareJob, cleanupJob, runScriptStep } from '../src/hooks'
|
|
||||||
import { TestTempOutput } from './test-setup'
|
|
||||||
import * as path from 'path'
|
|
||||||
import * as fs from 'fs'
|
import * as fs from 'fs'
|
||||||
|
import { cleanupJob, prepareJob, runScriptStep } from '../src/hooks'
|
||||||
|
import { TestHelper } from './test-setup'
|
||||||
|
|
||||||
jest.useRealTimers()
|
jest.useRealTimers()
|
||||||
|
|
||||||
let testTempOutput: TestTempOutput
|
let testHelper: TestHelper
|
||||||
|
|
||||||
const prepareJobJsonPath = path.resolve(
|
|
||||||
`${__dirname}/../../../examples/prepare-job.json`
|
|
||||||
)
|
|
||||||
let prepareJobData: any
|
|
||||||
|
|
||||||
let prepareJobOutputFilePath: string
|
|
||||||
let prepareJobOutputData: any
|
let prepareJobOutputData: any
|
||||||
|
|
||||||
|
let runScriptStepDefinition
|
||||||
|
|
||||||
describe('Run script step', () => {
|
describe('Run script step', () => {
|
||||||
beforeEach(async () => {
|
beforeEach(async () => {
|
||||||
const prepareJobJson = fs.readFileSync(prepareJobJsonPath)
|
testHelper = new TestHelper()
|
||||||
prepareJobData = JSON.parse(prepareJobJson.toString())
|
await testHelper.initialize()
|
||||||
console.log(prepareJobData)
|
const prepareJobOutputFilePath = testHelper.createFile(
|
||||||
|
|
||||||
testTempOutput = new TestTempOutput()
|
|
||||||
testTempOutput.initialize()
|
|
||||||
prepareJobOutputFilePath = testTempOutput.createFile(
|
|
||||||
'prepare-job-output.json'
|
'prepare-job-output.json'
|
||||||
)
|
)
|
||||||
|
|
||||||
|
const prepareJobData = testHelper.getPrepareJobDefinition()
|
||||||
|
runScriptStepDefinition = testHelper.getRunScriptStepDefinition()
|
||||||
|
|
||||||
await prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
await prepareJob(prepareJobData.args, prepareJobOutputFilePath)
|
||||||
const outputContent = fs.readFileSync(prepareJobOutputFilePath)
|
const outputContent = fs.readFileSync(prepareJobOutputFilePath)
|
||||||
prepareJobOutputData = JSON.parse(outputContent.toString())
|
prepareJobOutputData = JSON.parse(outputContent.toString())
|
||||||
@@ -33,7 +28,7 @@ describe('Run script step', () => {
|
|||||||
|
|
||||||
afterEach(async () => {
|
afterEach(async () => {
|
||||||
await cleanupJob()
|
await cleanupJob()
|
||||||
testTempOutput.cleanup()
|
await testHelper.cleanup()
|
||||||
})
|
})
|
||||||
|
|
||||||
// NOTE: To use this test, do kubectl apply -f podspec.yaml (from podspec examples)
|
// NOTE: To use this test, do kubectl apply -f podspec.yaml (from podspec examples)
|
||||||
@@ -41,21 +36,75 @@ describe('Run script step', () => {
|
|||||||
// npm run test run-script-step
|
// npm run test run-script-step
|
||||||
|
|
||||||
it('should not throw an exception', async () => {
|
it('should not throw an exception', async () => {
|
||||||
const args = {
|
|
||||||
entryPointArgs: ['echo "test"'],
|
|
||||||
entryPoint: '/bin/bash',
|
|
||||||
environmentVariables: {
|
|
||||||
NODE_ENV: 'development'
|
|
||||||
},
|
|
||||||
prependPath: ['/foo/bar', 'bar/foo'],
|
|
||||||
workingDirectory: '/__w/thboop-test2/thboop-test2'
|
|
||||||
}
|
|
||||||
const state = {
|
|
||||||
jobPod: prepareJobOutputData.state.jobPod
|
|
||||||
}
|
|
||||||
const responseFile = null
|
|
||||||
await expect(
|
await expect(
|
||||||
runScriptStep(args, state, responseFile)
|
runScriptStep(
|
||||||
|
runScriptStepDefinition.args,
|
||||||
|
prepareJobOutputData.state,
|
||||||
|
null
|
||||||
|
)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should fail if the working directory does not exist', async () => {
|
||||||
|
runScriptStepDefinition.args.workingDirectory = '/foo/bar'
|
||||||
|
await expect(
|
||||||
|
runScriptStep(
|
||||||
|
runScriptStepDefinition.args,
|
||||||
|
prepareJobOutputData.state,
|
||||||
|
null
|
||||||
|
)
|
||||||
|
).rejects.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('should shold have env variables available', async () => {
|
||||||
|
runScriptStepDefinition.args.entryPoint = 'bash'
|
||||||
|
|
||||||
|
runScriptStepDefinition.args.entryPointArgs = [
|
||||||
|
'-c',
|
||||||
|
"'if [[ -z $NODE_ENV ]]; then exit 1; fi'"
|
||||||
|
]
|
||||||
|
await expect(
|
||||||
|
runScriptStep(
|
||||||
|
runScriptStepDefinition.args,
|
||||||
|
prepareJobOutputData.state,
|
||||||
|
null
|
||||||
|
)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('Should have path variable changed in container with prepend path string', async () => {
|
||||||
|
runScriptStepDefinition.args.prependPath = '/some/path'
|
||||||
|
runScriptStepDefinition.args.entryPoint = '/bin/bash'
|
||||||
|
runScriptStepDefinition.args.entryPointArgs = [
|
||||||
|
'-c',
|
||||||
|
`'if [[ ! $(env | grep "^PATH=") = "PATH=${runScriptStepDefinition.args.prependPath}:"* ]]; then exit 1; fi'`
|
||||||
|
]
|
||||||
|
|
||||||
|
await expect(
|
||||||
|
runScriptStep(
|
||||||
|
runScriptStepDefinition.args,
|
||||||
|
prepareJobOutputData.state,
|
||||||
|
null
|
||||||
|
)
|
||||||
|
).resolves.not.toThrow()
|
||||||
|
})
|
||||||
|
|
||||||
|
it('Should have path variable changed in container with prepend path string array', async () => {
|
||||||
|
runScriptStepDefinition.args.prependPath = ['/some/other/path']
|
||||||
|
runScriptStepDefinition.args.entryPoint = '/bin/bash'
|
||||||
|
runScriptStepDefinition.args.entryPointArgs = [
|
||||||
|
'-c',
|
||||||
|
`'if [[ ! $(env | grep "^PATH=") = "PATH=${runScriptStepDefinition.args.prependPath.join(
|
||||||
|
':'
|
||||||
|
)}:"* ]]; then exit 1; fi'`
|
||||||
|
]
|
||||||
|
|
||||||
|
await expect(
|
||||||
|
runScriptStep(
|
||||||
|
runScriptStepDefinition.args,
|
||||||
|
prepareJobOutputData.state,
|
||||||
|
null
|
||||||
|
)
|
||||||
).resolves.not.toThrow()
|
).resolves.not.toThrow()
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
18
packages/k8s/tests/test-kind.yaml
Normal file
18
packages/k8s/tests/test-kind.yaml
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
kind: Cluster
|
||||||
|
apiVersion: kind.x-k8s.io/v1alpha4
|
||||||
|
nodes:
|
||||||
|
- role: control-plane
|
||||||
|
# add a mount from /path/to/my/files on the host to /files on the node
|
||||||
|
extraMounts:
|
||||||
|
- hostPath: {{PATHTOREPO}}
|
||||||
|
containerPath: {{PATHTOREPO}}
|
||||||
|
# optional: if set, the mount is read-only.
|
||||||
|
# default false
|
||||||
|
readOnly: false
|
||||||
|
# optional: if set, the mount needs SELinux relabeling.
|
||||||
|
# default false
|
||||||
|
selinuxRelabel: false
|
||||||
|
# optional: set propagation mode (None, HostToContainer or Bidirectional)
|
||||||
|
# see https://kubernetes.io/docs/concepts/storage/volumes/#mount-propagation
|
||||||
|
# default None
|
||||||
|
propagation: None
|
||||||
@@ -1,20 +1,80 @@
|
|||||||
|
import * as k8s from '@kubernetes/client-node'
|
||||||
import * as fs from 'fs'
|
import * as fs from 'fs'
|
||||||
|
import { HookData } from 'hooklib/lib'
|
||||||
|
import * as path from 'path'
|
||||||
import { v4 as uuidv4 } from 'uuid'
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
|
|
||||||
export class TestTempOutput {
|
const kc = new k8s.KubeConfig()
|
||||||
|
|
||||||
|
kc.loadFromDefault()
|
||||||
|
|
||||||
|
const k8sApi = kc.makeApiClient(k8s.CoreV1Api)
|
||||||
|
const k8sStorageApi = kc.makeApiClient(k8s.StorageV1Api)
|
||||||
|
|
||||||
|
export class TestHelper {
|
||||||
private tempDirPath: string
|
private tempDirPath: string
|
||||||
|
private podName: string
|
||||||
constructor() {
|
constructor() {
|
||||||
this.tempDirPath = `${__dirname}/_temp/${uuidv4()}`
|
this.tempDirPath = `${__dirname}/_temp/runner`
|
||||||
|
this.podName = uuidv4().replace(/-/g, '')
|
||||||
}
|
}
|
||||||
|
|
||||||
public initialize(): void {
|
public async initialize(): Promise<void> {
|
||||||
fs.mkdirSync(this.tempDirPath, { recursive: true })
|
process.env['ACTIONS_RUNNER_POD_NAME'] = `${this.podName}`
|
||||||
|
process.env['RUNNER_WORKSPACE'] = `${this.tempDirPath}/_work/repo`
|
||||||
|
process.env['RUNNER_TEMP'] = `${this.tempDirPath}/_work/_temp`
|
||||||
|
process.env['GITHUB_WORKSPACE'] = `${this.tempDirPath}/_work/repo/repo`
|
||||||
|
process.env['ACTIONS_RUNNER_KUBERNETES_NAMESPACE'] = 'default'
|
||||||
|
|
||||||
|
fs.mkdirSync(`${this.tempDirPath}/_work/repo/repo`, { recursive: true })
|
||||||
|
fs.mkdirSync(`${this.tempDirPath}/externals`, { recursive: true })
|
||||||
|
fs.mkdirSync(process.env.RUNNER_TEMP, { recursive: true })
|
||||||
|
|
||||||
|
fs.copyFileSync(
|
||||||
|
path.resolve(`${__dirname}/../../../examples/example-script.sh`),
|
||||||
|
`${process.env.RUNNER_TEMP}/example-script.sh`
|
||||||
|
)
|
||||||
|
|
||||||
|
await this.cleanupK8sResources()
|
||||||
|
try {
|
||||||
|
await this.createTestVolume()
|
||||||
|
await this.createTestJobPod()
|
||||||
|
} catch (e) {
|
||||||
|
console.log(e)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public cleanup(): void {
|
public async cleanup(): Promise<void> {
|
||||||
fs.rmSync(this.tempDirPath, { recursive: true })
|
try {
|
||||||
|
await this.cleanupK8sResources()
|
||||||
|
fs.rmSync(this.tempDirPath, { recursive: true })
|
||||||
|
} catch {}
|
||||||
|
}
|
||||||
|
public async cleanupK8sResources() {
|
||||||
|
await k8sApi
|
||||||
|
.deleteNamespacedPersistentVolumeClaim(
|
||||||
|
`${this.podName}-work`,
|
||||||
|
'default',
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
0
|
||||||
|
)
|
||||||
|
.catch(e => {})
|
||||||
|
await k8sApi.deletePersistentVolume(`${this.podName}-pv`).catch(e => {})
|
||||||
|
await k8sStorageApi.deleteStorageClass('local-storage').catch(e => {})
|
||||||
|
await k8sApi
|
||||||
|
.deleteNamespacedPod(this.podName, 'default', undefined, undefined, 0)
|
||||||
|
.catch(e => {})
|
||||||
|
await k8sApi
|
||||||
|
.deleteNamespacedPod(
|
||||||
|
`${this.podName}-workflow`,
|
||||||
|
'default',
|
||||||
|
undefined,
|
||||||
|
undefined,
|
||||||
|
0
|
||||||
|
)
|
||||||
|
.catch(e => {})
|
||||||
}
|
}
|
||||||
|
|
||||||
public createFile(fileName?: string): string {
|
public createFile(fileName?: string): string {
|
||||||
const filePath = `${this.tempDirPath}/${fileName || uuidv4()}`
|
const filePath = `${this.tempDirPath}/${fileName || uuidv4()}`
|
||||||
fs.writeFileSync(filePath, '')
|
fs.writeFileSync(filePath, '')
|
||||||
@@ -25,4 +85,112 @@ export class TestTempOutput {
|
|||||||
const filePath = `${this.tempDirPath}/${fileName}`
|
const filePath = `${this.tempDirPath}/${fileName}`
|
||||||
fs.rmSync(filePath)
|
fs.rmSync(filePath)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public async createTestJobPod() {
|
||||||
|
const container = {
|
||||||
|
name: 'nginx',
|
||||||
|
image: 'nginx:latest',
|
||||||
|
imagePullPolicy: 'IfNotPresent'
|
||||||
|
} as k8s.V1Container
|
||||||
|
|
||||||
|
const pod: k8s.V1Pod = {
|
||||||
|
metadata: {
|
||||||
|
name: this.podName
|
||||||
|
},
|
||||||
|
spec: {
|
||||||
|
restartPolicy: 'Never',
|
||||||
|
containers: [container]
|
||||||
|
}
|
||||||
|
} as k8s.V1Pod
|
||||||
|
await k8sApi.createNamespacedPod('default', pod)
|
||||||
|
}
|
||||||
|
|
||||||
|
public async createTestVolume() {
|
||||||
|
var sc: k8s.V1StorageClass = {
|
||||||
|
metadata: {
|
||||||
|
name: 'local-storage'
|
||||||
|
},
|
||||||
|
provisioner: 'kubernetes.io/no-provisioner',
|
||||||
|
volumeBindingMode: 'Immediate'
|
||||||
|
}
|
||||||
|
await k8sStorageApi.createStorageClass(sc)
|
||||||
|
|
||||||
|
var volume: k8s.V1PersistentVolume = {
|
||||||
|
metadata: {
|
||||||
|
name: `${this.podName}-pv`
|
||||||
|
},
|
||||||
|
spec: {
|
||||||
|
storageClassName: 'local-storage',
|
||||||
|
capacity: {
|
||||||
|
storage: '2Gi'
|
||||||
|
},
|
||||||
|
volumeMode: 'Filesystem',
|
||||||
|
accessModes: ['ReadWriteOnce'],
|
||||||
|
hostPath: {
|
||||||
|
path: `${this.tempDirPath}/_work`
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
await k8sApi.createPersistentVolume(volume)
|
||||||
|
var volumeClaim: k8s.V1PersistentVolumeClaim = {
|
||||||
|
metadata: {
|
||||||
|
name: `${this.podName}-work`
|
||||||
|
},
|
||||||
|
spec: {
|
||||||
|
accessModes: ['ReadWriteOnce'],
|
||||||
|
volumeMode: 'Filesystem',
|
||||||
|
storageClassName: 'local-storage',
|
||||||
|
volumeName: `${this.podName}-pv`,
|
||||||
|
resources: {
|
||||||
|
requests: {
|
||||||
|
storage: '1Gi'
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
await k8sApi.createNamespacedPersistentVolumeClaim('default', volumeClaim)
|
||||||
|
}
|
||||||
|
|
||||||
|
public getPrepareJobDefinition(): HookData {
|
||||||
|
const prepareJob = JSON.parse(
|
||||||
|
fs.readFileSync(
|
||||||
|
path.resolve(__dirname + '/../../../examples/prepare-job.json'),
|
||||||
|
'utf8'
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
prepareJob.args.container.userMountVolumes = undefined
|
||||||
|
prepareJob.args.container.registry = null
|
||||||
|
prepareJob.args.services.forEach(s => {
|
||||||
|
s.registry = null
|
||||||
|
})
|
||||||
|
|
||||||
|
return prepareJob
|
||||||
|
}
|
||||||
|
|
||||||
|
public getRunScriptStepDefinition(): HookData {
|
||||||
|
const runScriptStep = JSON.parse(
|
||||||
|
fs.readFileSync(
|
||||||
|
path.resolve(__dirname + '/../../../examples/run-script-step.json'),
|
||||||
|
'utf8'
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
runScriptStep.args.entryPointArgs[1] = `/__w/_temp/example-script.sh`
|
||||||
|
return runScriptStep
|
||||||
|
}
|
||||||
|
|
||||||
|
public getRunContainerStepDefinition(): HookData {
|
||||||
|
const runContainerStep = JSON.parse(
|
||||||
|
fs.readFileSync(
|
||||||
|
path.resolve(__dirname + '/../../../examples/run-container-step.json'),
|
||||||
|
'utf8'
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
runContainerStep.args.entryPointArgs[1] = `/__w/_temp/example-script.sh`
|
||||||
|
runContainerStep.args.userMountVolumes = undefined
|
||||||
|
runContainerStep.args.registry = null
|
||||||
|
return runContainerStep
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +1,9 @@
|
|||||||
## Features
|
## Features
|
||||||
- Initial Release
|
- Always use the Docker related ENVs from the host machine instead of ENVs from the runner job [#40]
|
||||||
|
- Use user defined entrypoints for service containers (instead of `tail -f /dev/null`)
|
||||||
|
|
||||||
## Bugs
|
## Bugs
|
||||||
|
- Fixed substring issue with /github/workspace and /github/file_commands [#35]
|
||||||
|
- Fixed issue related to setting hostPort and containerPort when formatting is not recognized by k8s default [#38]
|
||||||
|
|
||||||
|
<!-- ## Misc
|
||||||
## Misc
|
|
||||||
|
|||||||
Reference in New Issue
Block a user