Compare commits

..

67 Commits

Author SHA1 Message Date
LinuxServer-CI
e052b086ce Bot Updating Package Versions 2019-06-13 19:23:28 -04:00
LinuxServer-CI
118f21be42 Bot Updating Templated Files 2019-06-12 18:36:45 -04:00
LinuxServer-CI
0f8ca85398 Bot Updating Templated Files 2019-06-12 23:35:45 +01:00
Ryan Kuba
e7ffd7fb7f Extract Picons on boot Release-4.2 (#128)
* reworking logic to untar on first boot and reducing sample size for chown logic as it is not recursive

* bad path in arm variants
2019-03-28 17:18:44 +01:00
LinuxServer-CI
6eb29390f0 Bot Updating Package Versions 2019-03-28 07:53:07 +00:00
Ryan Kuba
c064466ce4 #72 fix chowning Release 4.2 (#126)
* making mods for release-4.2 branch

* typo
2019-03-28 08:03:35 +01:00
LinuxServer-CI
b6a321aee6 Bot Updating Package Versions 2019-03-25 02:02:20 +01:00
Ryan Kuba
3d34a9cf03 Merge pull request #124 from thelamer/release-4.2
Switching to new Base images, shift to arm32v7 tag.
2019-03-24 17:10:13 -07:00
thelamer
e0a76e29e2 Switching to new Base images, shift to arm32v7 tag. 2019-03-23 22:18:13 -07:00
LinuxServer-CI
13f64cbc4b Bot Updating Package Versions 2019-03-22 00:31:34 +00:00
LinuxServer-CI
2c696d51b0 Bot Updating Package Versions 2019-03-07 19:02:51 -05:00
LinuxServer-CI
780b22728e Bot Updating Templated Files 2019-03-07 18:14:52 -05:00
aptalca
9417e1f9c4 Merge pull request #121 from thelamer/release-4.2
libiconv shifted to community repos- release 4.2
2019-03-07 18:14:02 -05:00
thelamer
5fb3338564 libiconv shifted to community repos 2019-03-07 14:12:48 -08:00
LinuxServer-CI
89322d5956 Bot Updating Package Versions 2019-03-02 00:28:07 +00:00
LinuxServer-CI
267e3ef786 Bot Updating Templated Files 2019-03-01 18:35:41 -05:00
saarg
5de6faf27f Merge pull request #114 from thelamer/release-4.2
switching back to clone and checkout release-4.2
2019-02-17 22:54:37 +01:00
thelamer
dab37e4ecd bumping to xmltv 0.5.70 and pulling it from github instead of sourceforge 2019-02-17 12:42:34 -08:00
thelamer
6ab17ce8b7 adding jq package for commit parsing on local build 2019-02-17 12:15:38 -08:00
thelamer
81c71ee54d switching back to clone and checkout 2019-02-17 10:41:49 -08:00
saarg
a5fe08ef84 Merge pull request #111 from thelamer/release-4.2
Fix for picons Release 4.2
2019-02-16 21:33:35 +01:00
thelamer
65b2f3c3c7 making requested changes to default config 2019-02-16 09:23:06 -08:00
thelamer
7d94cc4b64 forgot schema value 2019-02-15 13:03:14 -08:00
thelamer
e834619ec3 fixing paths for initial config 2019-02-15 12:58:34 -08:00
thelamer
ab115449e7 need to untar picons for htem to be read by the app 2019-02-15 11:45:41 -08:00
thelamer
322e35bb00 adding picons path to config if this is a first boot 2019-02-14 22:19:37 -08:00
j0nnymoe
6dfbb30c0a Merge pull request #108 from thelamer/release-4.2
adding dynamic readme
2019-02-12 14:34:26 +00:00
thelamer
94b0bc34c7 adding dynamic readme 2019-02-11 21:07:25 -08:00
LinuxServer-CI
2c89a24550 Bot Updating Package Versions 2019-02-11 22:20:05 +00:00
LinuxServer-CI
4d2207db61 Bot Updating Templated Files 2019-02-11 21:42:22 +00:00
aptalca
27ac3525fb Merge pull request #105 from linuxserver/pipeline-4.2
adding multi arch and pipeline logic, pulling picons from object storage
2019-02-11 16:41:43 -05:00
thelamer
946c4fd8a9 adding multi arch and pipeline logic, pulling picons from object storage 2019-01-17 22:46:45 -08:00
aptalca
bf0911b6f5 Merge pull request #95 from linuxserver/release-4.2_rebase_alpine_3.8
rebase to alpine 3.8 and use buildstage type build
2018-09-12 15:31:15 -04:00
sparklyballs
1a9770d227 rebase to alpine 3.8 2018-09-12 15:46:35 +01:00
sparklyballs
c0a3661442 rebase to alpine 3.8 and use buildstage type build 2018-09-03 12:38:00 +01:00
saarg
ba0a26bfb1 Merge pull request #85 from linuxserver/xmltv_4.2
Add JSON::XS Perl package for grab_tv_huro
2018-04-21 21:00:35 +02:00
tobbenb
ac7dc7b478 Add JSON::XS Perl package for grab_tv_huro 2018-04-21 16:41:11 +02:00
sparklyballs
24d94ed4ad Merge pull request #79 from linuxserver/release-4.2_use_sourceforge_master_xmltv
use sourceforge master rather than mirror for xmltv
2018-03-04 11:26:02 +00:00
sparklyballs
fb45df2646 use sourceforge master rather than mirror for xmltv 2018-03-04 09:39:23 +00:00
sparklyballs
57893d0451 Merge pull request #76 from linuxserver/pngquant
Add missing pngquant package
2018-02-23 13:06:20 +00:00
tobbenb
692ee67cf4 Add missing pngquant package 2018-02-22 21:57:43 +01:00
saarg
cacfb82b62 Merge pull request #75 from linuxserver/iconv_4.2
Fix wrong version of Iconv and add build options
2018-02-21 23:41:34 +01:00
tobbenb
e3b56a32cc Remove old libiconv patch 2018-02-21 23:01:55 +01:00
tobbenb
b487342878 Sparklyfi Dockerfile 2018-02-21 21:57:19 +01:00
tobbenb
0a365513dd Update README 2018-02-21 21:54:23 +01:00
tobbenb
1075ff6c38 Add build options from CvH 2018-02-21 21:52:32 +01:00
tobbenb
76d27bb36d Fix iconv runtime crash 2018-02-21 21:50:56 +01:00
sparklyballs
2f197666ac Merge pull request #66 from linuxserver/release-4.2_deprecate_cpu_core
deprecate cpu_core routine lack of scaling
2018-01-04 20:45:39 +00:00
sparklyballs
34b9ebf98f deprecate cpu_core routine lack of scaling 2018-01-04 11:49:17 +00:00
sparklyballs
e1e7730dcc Merge pull request #64 from linuxserver/release-4.2_3.7
bump alpine 3.7, fix continuation lines, linting fixes
2017-12-11 14:44:51 +00:00
sparklyballs
8269eee493 bump alpine 3.7, fix continuation lines, linting fixes 2017-12-11 12:42:03 +00:00
sparklyballs
4fa1634986 Update README.md 2017-12-09 00:40:09 +00:00
sparklyballs
5b5dee7f67 dot resolves weird github/jenkins syncing commits 2017-12-02 00:08:48 +00:00
aptalca
56c3b9a524 Merge pull request #61 from linuxserver/release-4.2_config_patches
increased uniformity across all archs
2017-07-13 20:28:00 -04:00
sparklyballs
a815f66aa3 increased uniformity across all archs 2017-07-13 20:21:04 +01:00
sparklyballs
bcaaad93e4 update of README with badges for new ci 2017-07-11 09:38:21 +01:00
sparklyballs
1cd1bf3fc5 Merge pull request #59 from linuxserver/readme-release
Update README with full path for comskip
2017-07-09 12:25:22 +01:00
tobbenb
b210ad4793 Update README with full path for comskip 2017-07-08 12:28:27 +02:00
sparklyballs
4af4d02187 Merge pull request #57 from linuxserver/dockerfile_tidy
tidy up Dockerfile and install runtime packages all in one
2017-07-03 11:12:17 +01:00
sparklyballs
c340938778 tidy up Dockerfile and install runtime packages all in one 2017-07-03 05:42:55 +01:00
tobbenb
cce4adf7e4 Move to one branch for all 4.2 releases 2017-07-02 16:40:53 +02:00
saarg
ee18601f12 Merge pull request #55 from linuxserver/stable-4.2.1_alpine3.6
bump to alpine 3.6
2017-05-28 16:12:22 +02:00
sparklyballs
68cc40f86e bump to alpine 3.6 2017-05-28 02:29:34 +01:00
sparklyballs
9b1f7b3303 Merge pull request #51 from linuxserver/4.2.1
Update to tvheadend 4.2.1 stable
2017-05-01 13:56:26 +01:00
sparklyballs
01af20a3df Update README.md 2017-05-01 12:37:09 +01:00
tobbenb
f65d09bea2 Cleaning up after sparklys cut and paste double 2017-05-01 13:17:08 +02:00
tobbenb
aab4fb1b35 Update to tvheadend 4.2.1 stable 2017-05-01 12:55:41 +02:00
26 changed files with 561 additions and 1318 deletions

View File

@@ -1,122 +0,0 @@
# Contributing to tvheadend
## Gotchas
* While contributing make sure to make all your changes before creating a Pull Request, as our pipeline builds each commit after the PR is open.
* Read, and fill the Pull Request template
* If this is a fix for a typo in code or documentation in the README please file an issue
* If the PR is addressing an existing issue include, closes #\<issue number>, in the body of the PR commit message
* If you want to discuss changes, you can also bring it up in [#dev-talk](https://discordapp.com/channels/354974912613449730/757585807061155840) in our [Discord server](https://discord.gg/YWrKVTn)
## Common files
| File | Use case |
| :----: | --- |
| `Dockerfile` | Dockerfile used to build amd64 images |
| `Dockerfile.aarch64` | Dockerfile used to build 64bit ARM architectures |
| `Dockerfile.armhf` | Dockerfile used to build 32bit ARM architectures |
| `Jenkinsfile` | This file is a product of our builder and should not be edited directly. This is used to build the image |
| `jenkins-vars.yml` | This file is used to generate the `Jenkinsfile` mentioned above, it only affects the build-process |
| `package_versions.txt` | This file is generated as a part of the build-process and should not be edited directly. It lists all the installed packages and their versions |
| `README.md` | This file is a product of our builder and should not be edited directly. This displays the readme for the repository and image registries |
| `readme-vars.yml` | This file is used to generate the `README.md` |
## Readme
If you would like to change our readme, please __**do not**__ directly edit the readme, as it is auto-generated on each commit.
Instead edit the [readme-vars.yml](https://github.com/linuxserver/docker-tvheadend/edit/master/readme-vars.yml).
These variables are used in a template for our [Jenkins Builder](https://github.com/linuxserver/docker-jenkins-builder) as part of an ansible play.
Most of these variables are also carried over to [docs.linuxserver.io](https://docs.linuxserver.io/images/docker-tvheadend)
### Fixing typos or clarify the text in the readme
There are variables for multiple parts of the readme, the most common ones are:
| Variable | Description |
| :----: | --- |
| `project_blurb` | This is the short excerpt shown above the project logo. |
| `app_setup_block` | This is the text that shows up under "Application Setup" if enabled |
### Parameters
The compose and run examples are also generated from these variables.
We have a [reference file](https://github.com/linuxserver/docker-jenkins-builder/blob/master/vars/_container-vars-blank) in our Jenkins Builder.
These are prefixed with `param_` for required parameters, or `opt_param` for optional parameters, except for `cap_add`.
Remember to enable param, if currently disabled. This differs between parameters, and can be seen in the reference file.
Devices, environment variables, ports and volumes expects its variables in a certain way.
### Devices
```yml
param_devices:
- { device_path: "/dev/dri", device_host_path: "/dev/dri", desc: "For hardware transcoding" }
opt_param_devices:
- { device_path: "/dev/dri", device_host_path: "/dev/dri", desc: "For hardware transcoding" }
```
### Environment variables
```yml
param_env_vars:
- { env_var: "TZ", env_value: "Europe/London", desc: "Specify a timezone to use EG Europe/London." }
opt_param_env_vars:
- { env_var: "VERSION", env_value: "latest", desc: "Supported values are LATEST, PLEXPASS or a specific version number." }
```
### Ports
```yml
param_ports:
- { external_port: "80", internal_port: "80", port_desc: "Application WebUI" }
opt_param_ports:
- { external_port: "80", internal_port: "80", port_desc: "Application WebUI" }
```
### Volumes
```yml
param_volumes:
- { vol_path: "/config", vol_host_path: "</path/to/appdata/config>", desc: "Configuration files." }
opt_param_volumes:
- { vol_path: "/config", vol_host_path: "</path/to/appdata/config>", desc: "Configuration files." }
```
### Testing template changes
After you make any changes to the templates, you can use our [Jenkins Builder](https://github.com/linuxserver/docker-jenkins-builder) to have the files updated from the modified templates. Please use the command found under `Running Locally` [on this page](https://github.com/linuxserver/docker-jenkins-builder/blob/master/README.md) to generate them prior to submitting a PR.
## Dockerfiles
We use multiple Dockerfiles in our repos, this is because sometimes some CPU architectures needs different packages to work.
If you are proposing additional packages to be added, ensure that you added the packages to all the Dockerfiles in alphabetical order.
### Testing your changes
```
git clone https://github.com/linuxserver/docker-tvheadend.git
cd docker-tvheadend
docker build \
--no-cache \
--pull \
-t linuxserver/tvheadend:latest .
```
The ARM variants can be built on x86_64 hardware using `multiarch/qemu-user-static`
```
docker run --rm --privileged multiarch/qemu-user-static:register --reset
```
Once registered you can define the dockerfile to use with `-f Dockerfile.aarch64`.
## Update the chagelog
If you are modifying the Dockerfiles or any of the startup scripts in [root](https://github.com/linuxserver/docker-tvheadend/tree/master/root), add an entry to the changelog
```yml
changelogs:
- { date: "DD.MM.YY:", desc: "Added some love to templates" }
```

2
.github/FUNDING.yml vendored
View File

@@ -1,2 +0,0 @@
github: linuxserver
open_collective: linuxserver

21
.github/ISSUE_TEMPLATE.md vendored Normal file
View File

@@ -0,0 +1,21 @@
<!--- Provide a general summary of the issue in the Title above -->
[linuxserverurl]: https://linuxserver.io
[![linuxserver.io](https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/linuxserver_medium.png)][linuxserverurl]
<!--- If you have an issue with the project, please provide us with the following information -->
<!--- Host OS -->
<!--- Command line users, your run/create command, GUI/Unraid users, a screenshot of your template settings. -->
<!--- Docker log output, docker log <container-name> -->
<!--- Mention if you're using symlinks on any of the volume mounts. -->
<!--- If you have a suggestion or fix for the project, please provide us with the following information -->
<!--- What you think your suggestion brings to the project, or fixes with the project -->
<!--- If it's a fix, would it be better suited as a Pull request to the repo ? -->
## Thanks, team linuxserver.io

View File

@@ -1,13 +0,0 @@
blank_issues_enabled: false
contact_links:
- name: Discord chat support
url: https://discord.gg/YWrKVTn
about: Realtime support / chat with the community and the team.
- name: Discourse discussion forum
url: https://discourse.linuxserver.io
about: Post on our community forum.
- name: Documentation
url: https://docs.linuxserver.io/images/docker-tvheadend
about: Documentation - information about all of our containers.

View File

@@ -1,40 +0,0 @@
---
name: Bug report
about: Create a report to help us improve
---
[linuxserverurl]: https://linuxserver.io
[![linuxserver.io](https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/linuxserver_medium.png)][linuxserverurl]
<!--- If you are new to Docker or this application our issue tracker is **ONLY** used for reporting bugs or requesting features. Please use [our discord server](https://discord.gg/YWrKVTn) for general support. --->
<!--- Provide a general summary of the bug in the Title above -->
------------------------------
## Expected Behavior
<!--- Tell us what should happen -->
## Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
## Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1.
2.
3.
4.
## Environment
**OS:**
**CPU architecture:** x86_64/arm32/arm64
**How docker service was installed:**
<!--- ie. from the official docker repo, from the distro repo, nas OS provided, etc. -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
## Command used to create docker container (run/create/compose/screenshot)
<!--- Provide your docker create/run command or compose yaml snippet, or a screenshot of settings if using a gui to create the container -->
## Docker logs
<!--- Provide a full docker log, output of "docker logs tvheadend" -->

View File

@@ -1,25 +0,0 @@
---
name: Feature request
about: Suggest an idea for this project
---
[linuxserverurl]: https://linuxserver.io
[![linuxserver.io](https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/linuxserver_medium.png)][linuxserverurl]
<!--- If you are new to Docker or this application our issue tracker is **ONLY** used for reporting bugs or requesting features. Please use [our discord server](https://discord.gg/YWrKVTn) for general support. --->
<!--- If this acts as a feature request please ask yourself if this modification is something the whole userbase will benefit from --->
<!--- If this is a specific change for corner case functionality or plugins please look at making a Docker Mod or local script https://blog.linuxserver.io/2019/09/14/customizing-our-containers/ -->
<!--- Provide a general summary of the request in the Title above -->
------------------------------
## Desired Behavior
<!--- Tell us what should happen -->
## Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
## Alternatives Considered
<!--- Tell us what other options you have tried or considered -->

View File

@@ -2,42 +2,14 @@
[linuxserverurl]: https://linuxserver.io [linuxserverurl]: https://linuxserver.io
[![linuxserver.io](https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/linuxserver_medium.png)][linuxserverurl] [![linuxserver.io](https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/linuxserver_medium.png)][linuxserverurl]
<!--- Before submitting a pull request please check the following --> <!--- Before submitting a pull request please check the following -->
<!--- If this is a fix for a typo in code or documentation in the README please file an issue and let us sort it out we do not need a PR --> <!--- That you have made a branch in your fork, we'd rather not merge from your master -->
<!--- Ask yourself if this modification is something the whole userbase will benefit from, if this is a specific change for corner case functionality or plugins please look at making a Docker Mod or local script https://blog.linuxserver.io/2019/09/14/customizing-our-containers/ -->
<!--- That if the PR is addressing an existing issue include, closes #<issue number> , in the body of the PR commit message --> <!--- That if the PR is addressing an existing issue include, closes #<issue number> , in the body of the PR commit message -->
<!--- You have included links to any files / patches etc your PR may be using in the body of the PR commit message --> <!--- You have included links to any files / patches etc your PR may be using in the body of the PR commit message -->
<!--- We maintain a changelog of major revisions to the container at the end of readme-vars.yml in the root of this repository, please add your changes there if appropriate --> <!--- -->
## Thanks, team linuxserver.io
<!--- Coding guidelines: -->
<!--- 1. Installed packages in the Dockerfiles should be in alphabetical order -->
<!--- 2. Changes to Dockerfile should be replicated in Dockerfile.armhf and Dockerfile.aarch64 if applicable -->
<!--- 3. Indentation style (tabs vs 4 spaces vs 1 space) should match the rest of the document -->
<!--- 4. Readme is auto generated from readme-vars.yml, make your changes there -->
------------------------------
- [ ] I have read the [contributing](https://github.com/linuxserver/docker-tvheadend/blob/master/.github/CONTRIBUTING.md) guideline and understand that I have made the correct modifications
------------------------------
<!--- We welcome all PRs though this doesnt guarantee it will be accepted. -->
## Description:
<!--- Describe your changes in detail -->
## Benefits of this PR and context:
<!--- Please explain why we should accept this PR. If this fixes an outstanding bug, please reference the issue # -->
## How Has This Been Tested?
<!--- Please describe in detail how you tested your changes. -->
<!--- Include details of your testing environment, and the tests you ran to -->
<!--- see how your change affects other areas of the code, etc. -->
## Source / References:
<!--- Please include any forum posts/github links relevant to the PR -->

View File

@@ -1,92 +0,0 @@
name: External Trigger Main
on:
workflow_dispatch:
jobs:
external-trigger-master:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2.3.3
- name: External Trigger
if: github.ref == 'refs/heads/master'
run: |
if [ -n "${{ secrets.PAUSE_EXTERNAL_TRIGGER_TVHEADEND_MASTER }}" ]; then
echo "**** Github secret PAUSE_EXTERNAL_TRIGGER_TVHEADEND_MASTER is set; skipping trigger. ****"
exit 0
fi
echo "**** External trigger running off of master branch. To disable this trigger, set a Github secret named \"PAUSE_EXTERNAL_TRIGGER_TVHEADEND_MASTER\". ****"
echo "**** Retrieving external version ****"
EXT_RELEASE=$(curl -u "${{ secrets.CR_USER }}:${{ secrets.CR_PAT }}" -sX GET "https://api.github.com/repos/tvheadend/tvheadend/commits/master" | jq -r '. | .sha' | cut -c1-8)
if [ -z "${EXT_RELEASE}" ] || [ "${EXT_RELEASE}" == "null" ]; then
echo "**** Can't retrieve external version, exiting ****"
FAILURE_REASON="Can't retrieve external version for tvheadend branch master"
GHA_TRIGGER_URL="https://github.com/linuxserver/docker-tvheadend/actions/runs/${{ github.run_id }}"
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 16711680,
"description": "**Trigger Failed** \n**Reason:** '"${FAILURE_REASON}"' \n**Trigger URL:** '"${GHA_TRIGGER_URL}"' \n"}],
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
exit 1
fi
EXT_RELEASE=$(echo ${EXT_RELEASE} | sed 's/[~,%@+;:/]//g')
echo "**** External version: ${EXT_RELEASE} ****"
echo "**** Retrieving last pushed version ****"
image="linuxserver/tvheadend"
tag="latest"
token=$(curl -sX GET \
"https://ghcr.io/token?scope=repository%3Alinuxserver%2Ftvheadend%3Apull" \
| jq -r '.token')
multidigest=$(curl -s \
--header "Accept: application/vnd.docker.distribution.manifest.v2+json" \
--header "Authorization: Bearer ${token}" \
"https://ghcr.io/v2/${image}/manifests/${tag}" \
| jq -r 'first(.manifests[].digest)')
digest=$(curl -s \
--header "Accept: application/vnd.docker.distribution.manifest.v2+json" \
--header "Authorization: Bearer ${token}" \
"https://ghcr.io/v2/${image}/manifests/${multidigest}" \
| jq -r '.config.digest')
image_info=$(curl -sL \
--header "Authorization: Bearer ${token}" \
"https://ghcr.io/v2/${image}/blobs/${digest}" \
| jq -r '.container_config')
IMAGE_RELEASE=$(echo ${image_info} | jq -r '.Labels.build_version' | awk '{print $3}')
IMAGE_VERSION=$(echo ${IMAGE_RELEASE} | awk -F'-ls' '{print $1}')
if [ -z "${IMAGE_VERSION}" ]; then
echo "**** Can't retrieve last pushed version, exiting ****"
FAILURE_REASON="Can't retrieve last pushed version for tvheadend tag latest"
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 16711680,
"description": "**Trigger Failed** \n**Reason:** '"${FAILURE_REASON}"' \n"}],
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
exit 1
fi
echo "**** Last pushed version: ${IMAGE_VERSION} ****"
if [ "${EXT_RELEASE}" == "${IMAGE_VERSION}" ]; then
echo "**** Version ${EXT_RELEASE} already pushed, exiting ****"
exit 0
elif [ $(curl -s https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-tvheadend/job/master/lastBuild/api/json | jq -r '.building') == "true" ]; then
echo "**** New version ${EXT_RELEASE} found; but there already seems to be an active build on Jenkins; exiting ****"
exit 0
else
echo "**** New version ${EXT_RELEASE} found; old version was ${IMAGE_VERSION}. Triggering new build ****"
response=$(curl -iX POST \
https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-tvheadend/job/master/buildWithParameters?PACKAGE_CHECK=false \
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} | grep -i location | sed "s|^[L|l]ocation: \(.*\)|\1|")
echo "**** Jenkins job queue url: ${response%$'\r'} ****"
echo "**** Sleeping 10 seconds until job starts ****"
sleep 10
buildurl=$(curl -s "${response%$'\r'}api/json" | jq -r '.executable.url')
buildurl="${buildurl%$'\r'}"
echo "**** Jenkins job build url: ${buildurl} ****"
echo "**** Attempting to change the Jenkins job description ****"
curl -iX POST \
"${buildurl}submitDescription" \
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} \
--data-urlencode "description=GHA external trigger https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }}" \
--data-urlencode "Submit=Submit"
echo "**** Notifying Discord ****"
TRIGGER_REASON="A version change was detected for tvheadend tag latest. Old version:${IMAGE_VERSION} New version:${EXT_RELEASE}"
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 9802903,
"description": "**Build Triggered** \n**Reason:** '"${TRIGGER_REASON}"' \n**Build URL:** '"${buildurl}display/redirect"' \n"}],
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}
fi

View File

@@ -1,43 +0,0 @@
name: External Trigger Scheduler
on:
schedule:
- cron: '03 * * * *'
workflow_dispatch:
jobs:
external-trigger-scheduler:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2.3.3
with:
fetch-depth: '0'
- name: External Trigger Scheduler
run: |
echo "**** Branches found: ****"
git for-each-ref --format='%(refname:short)' refs/remotes
echo "**** Pulling the yq docker image ****"
docker pull ghcr.io/linuxserver/yq
for br in $(git for-each-ref --format='%(refname:short)' refs/remotes)
do
br=$(echo "$br" | sed 's|origin/||g')
echo "**** Evaluating branch ${br} ****"
ls_branch=$(curl -sX GET https://raw.githubusercontent.com/linuxserver/docker-tvheadend/${br}/jenkins-vars.yml \
| docker run --rm -i --entrypoint yq ghcr.io/linuxserver/yq -r .ls_branch)
if [ "$br" == "$ls_branch" ]; then
echo "**** Branch ${br} appears to be live; checking workflow. ****"
if curl -sfX GET https://raw.githubusercontent.com/linuxserver/docker-tvheadend/${br}/.github/workflows/external_trigger.yml > /dev/null 2>&1; then
echo "**** Workflow exists. Triggering external trigger workflow for branch ${br} ****."
curl -iX POST \
-H "Authorization: token ${{ secrets.CR_PAT }}" \
-H "Accept: application/vnd.github.v3+json" \
-d "{\"ref\":\"refs/heads/${br}\"}" \
https://api.github.com/repos/linuxserver/docker-tvheadend/actions/workflows/external_trigger.yml/dispatches
else
echo "**** Workflow doesn't exist; skipping trigger. ****"
fi
else
echo "**** ${br} appears to be a dev branch; skipping trigger. ****"
fi
done

View File

@@ -1,13 +0,0 @@
name: Greetings
on: [pull_request_target, issues]
jobs:
greeting:
runs-on: ubuntu-latest
steps:
- uses: actions/first-interaction@v1
with:
issue-message: 'Thanks for opening your first issue here! Be sure to follow the [bug](https://github.com/linuxserver/docker-tvheadend/blob/master/.github/ISSUE_TEMPLATE/issue.bug.md) or [feature](https://github.com/linuxserver/docker-tvheadend/blob/master/.github/ISSUE_TEMPLATE/issue.feature.md) issue templates!'
pr-message: 'Thanks for opening this pull request! Be sure to follow the [pull request template](https://github.com/linuxserver/docker-tvheadend/blob/master/.github/PULL_REQUEST_TEMPLATE.md)!'
repo-token: ${{ secrets.GITHUB_TOKEN }}

View File

@@ -1,38 +0,0 @@
name: Package Trigger Main
on:
workflow_dispatch:
jobs:
package-trigger-master:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2.3.3
- name: Package Trigger
if: github.ref == 'refs/heads/master'
run: |
if [ -n "${{ secrets.PAUSE_PACKAGE_TRIGGER_TVHEADEND_MASTER }}" ]; then
echo "**** Github secret PAUSE_PACKAGE_TRIGGER_TVHEADEND_MASTER is set; skipping trigger. ****"
exit 0
fi
if [ $(curl -s https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-tvheadend/job/master/lastBuild/api/json | jq -r '.building') == "true" ]; then
echo "**** There already seems to be an active build on Jenkins; skipping package trigger ****"
exit 0
fi
echo "**** Package trigger running off of master branch. To disable, set a Github secret named \"PAUSE_PACKAGE_TRIGGER_TVHEADEND_MASTER\". ****"
response=$(curl -iX POST \
https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-tvheadend/job/master/buildWithParameters?PACKAGE_CHECK=true \
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} | grep -i location | sed "s|^[L|l]ocation: \(.*\)|\1|")
echo "**** Jenkins job queue url: ${response%$'\r'} ****"
echo "**** Sleeping 10 seconds until job starts ****"
sleep 10
buildurl=$(curl -s "${response%$'\r'}api/json" | jq -r '.executable.url')
buildurl="${buildurl%$'\r'}"
echo "**** Jenkins job build url: ${buildurl} ****"
echo "**** Attempting to change the Jenkins job description ****"
curl -iX POST \
"${buildurl}submitDescription" \
--user ${{ secrets.JENKINS_USER }}:${{ secrets.JENKINS_TOKEN }} \
--data-urlencode "description=GHA package trigger https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }}" \
--data-urlencode "Submit=Submit"

View File

@@ -1,50 +0,0 @@
name: Package Trigger Scheduler
on:
schedule:
- cron: '52 14 * * 0'
workflow_dispatch:
jobs:
package-trigger-scheduler:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2.3.3
with:
fetch-depth: '0'
- name: Package Trigger Scheduler
run: |
echo "**** Branches found: ****"
git for-each-ref --format='%(refname:short)' refs/remotes
echo "**** Pulling the yq docker image ****"
docker pull ghcr.io/linuxserver/yq
for br in $(git for-each-ref --format='%(refname:short)' refs/remotes)
do
br=$(echo "$br" | sed 's|origin/||g')
echo "**** Evaluating branch ${br} ****"
ls_branch=$(curl -sX GET https://raw.githubusercontent.com/linuxserver/docker-tvheadend/${br}/jenkins-vars.yml \
| docker run --rm -i --entrypoint yq ghcr.io/linuxserver/yq -r .ls_branch)
if [ "${br}" == "${ls_branch}" ]; then
echo "**** Branch ${br} appears to be live; checking workflow. ****"
if curl -sfX GET https://raw.githubusercontent.com/linuxserver/docker-tvheadend/${br}/.github/workflows/package_trigger.yml > /dev/null 2>&1; then
echo "**** Workflow exists. Triggering package trigger workflow for branch ${br}. ****"
triggered_branches="${triggered_branches}${br} "
curl -iX POST \
-H "Authorization: token ${{ secrets.CR_PAT }}" \
-H "Accept: application/vnd.github.v3+json" \
-d "{\"ref\":\"refs/heads/${br}\"}" \
https://api.github.com/repos/linuxserver/docker-tvheadend/actions/workflows/package_trigger.yml/dispatches
sleep 30
else
echo "**** Workflow doesn't exist; skipping trigger. ****"
fi
else
echo "**** ${br} appears to be a dev branch; skipping trigger. ****"
fi
done
echo "**** Package check build(s) triggered for branch(es): ${triggered_branches} ****"
echo "**** Notifying Discord ****"
curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://cdn.discordapp.com/avatars/354986384542662657/df91181b3f1cf0ef1592fbe18e0962d7.png","embeds": [{"color": 9802903,
"description": "**Package Check Build(s) Triggered for tvheadend** \n**Branch(es):** '"${triggered_branches}"' \n**Build URL:** '"https://ci.linuxserver.io/blue/organizations/jenkins/Docker-Pipeline-Builders%2Fdocker-tvheadend/activity/"' \n"}],
"username": "Github Actions"}' ${{ secrets.DISCORD_WEBHOOK }}

View File

@@ -1,23 +0,0 @@
name: Mark stale issues and pull requests
on:
schedule:
- cron: "30 1 * * *"
jobs:
stale:
runs-on: ubuntu-latest
steps:
- uses: actions/stale@v1
with:
stale-issue-message: "This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions."
stale-pr-message: "This pull request has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions."
stale-issue-label: 'no-issue-activity'
stale-pr-label: 'no-pr-activity'
days-before-stale: 30
days-before-close: 365
exempt-issue-labels: 'awaiting-approval,work-in-progress'
exempt-pr-labels: 'awaiting-approval,work-in-progress'
repo-token: ${{ secrets.GITHUB_TOKEN }}

1
.gitignore vendored
View File

@@ -44,4 +44,3 @@ Temporary Items
# Picons file # Picons file
picons.tar.bz2 picons.tar.bz2
.jenkins-external

View File

@@ -1,13 +1,14 @@
FROM ghcr.io/linuxserver/baseimage-alpine:3.12 as buildstage FROM lsiobase/alpine:3.9 as buildstage
############## build stage ############## ############## build stage ##############
# package versions # package versions
ARG ARGTABLE_VER="2.13" ARG ARGTABLE_VER="2.13"
ARG XMLTV_VER="v0.6.3" ARG TVH_VER="release/4.2"
ARG XMLTV_VER="v0.5.70"
ARG TVHEADEND_COMMIT
# environment settings # environment settings
ARG TZ="Europe/Oslo" ARG TZ="Europe/Oslo"
ARG TVHEADEND_COMMIT
ENV HOME="/config" ENV HOME="/config"
# copy patches # copy patches
@@ -29,21 +30,18 @@ RUN \
gcc \ gcc \
gettext-dev \ gettext-dev \
git \ git \
gnu-libiconv-dev \
gzip \ gzip \
jq \ jq \
libcurl \
libdvbcsa-dev \
libgcrypt-dev \ libgcrypt-dev \
libhdhomerun-dev \ libhdhomerun-dev \
libressl-dev \
libtool \ libtool \
libva-dev \
libvpx-dev \ libvpx-dev \
libxml2-dev \ libxml2-dev \
libxslt-dev \ libxslt-dev \
linux-headers \ linux-headers \
make \ make \
openssl-dev \ mercurial \
opus-dev \ opus-dev \
patch \ patch \
pcre2-dev \ pcre2-dev \
@@ -75,12 +73,10 @@ RUN \
perl-io-socket-ssl \ perl-io-socket-ssl \
perl-io-stringy \ perl-io-stringy \
perl-json \ perl-json \
perl-json-xs \
perl-libwww \ perl-libwww \
perl-lingua-en-numbers-ordinate \ perl-lingua-en-numbers-ordinate \
perl-lingua-preferred \ perl-lingua-preferred \
perl-list-moreutils \ perl-list-moreutils \
perl-lwp-useragent-determined \
perl-module-build \ perl-module-build \
perl-module-pluggable \ perl-module-pluggable \
perl-net-ssleay \ perl-net-ssleay \
@@ -103,14 +99,17 @@ RUN \
perl-xml-writer \ perl-xml-writer \
pkgconf \ pkgconf \
pngquant \ pngquant \
python2 \ python \
sdl-dev \ sdl-dev \
tar \ tar \
uriparser-dev \ uriparser-dev \
wget \ wget \
x264-dev \ x264-dev \
x265-dev \ x265-dev \
zlib-dev zlib-dev && \
apk add --no-cache \
--repository http://dl-cdn.alpinelinux.org/alpine/edge/community \
gnu-libiconv-dev
RUN \ RUN \
echo "**** remove musl iconv.h and replace with gnu-iconv.h ****" && \ echo "**** remove musl iconv.h and replace with gnu-iconv.h ****" && \
@@ -119,7 +118,7 @@ RUN \
RUN \ RUN \
echo "**** install perl modules for xmltv ****" && \ echo "**** install perl modules for xmltv ****" && \
curl -s -L https://cpanmin.us | perl - App::cpanminus && \ curl -L http://cpanmin.us | perl - App::cpanminus && \
cpanm --installdeps /tmp/patches cpanm --installdeps /tmp/patches
RUN \ RUN \
@@ -140,44 +139,41 @@ RUN \
make test && \ make test && \
make DESTDIR=/tmp/xmltv-build install make DESTDIR=/tmp/xmltv-build install
RUN \
echo "**** compile dvb-apps ****" && \
hg clone http://linuxtv.org/hg/dvb-apps /tmp/dvb-apps && \
cd /tmp/dvb-apps && \
make -C lib && \
make -C lib DESTDIR=/tmp/dvbapps-build install && \
cp -pr /tmp/dvbapps-build/usr/* /usr/
RUN \ RUN \
echo "**** compile tvheadend ****" && \ echo "**** compile tvheadend ****" && \
if [ -z ${TVHEADEND_COMMIT+x} ]; then \ if [ -z ${TVHEADEND_COMMIT+x} ]; then \
TVHEADEND_COMMIT=$(curl -sX GET https://api.github.com/repos/tvheadend/tvheadend/commits/master \ TVHEADEND_COMMIT=$(curl -sX GET https://api.github.com/repos/tvheadend/tvheadend/commits/${TVH_VER} \
| jq -r '. | .sha'); \ | jq -r '. | .sha'); \
fi && \ fi && \
mkdir -p \ mkdir -p \
/tmp/tvheadend && \ /tmp/tvheadend && \
git clone https://github.com/tvheadend/tvheadend.git /tmp/tvheadend && \ git clone https://github.com/tvheadend/tvheadend.git /tmp/tvheadend && \
cd /tmp/tvheadend && \ cd /tmp/tvheadend && \
git checkout ${TVHEADEND_COMMIT} && \ git checkout ${TVHEADEND_COMMIT} && \
./configure \ ./configure \
`#Encoding` \ --disable-avahi \
--disable-bintray_cache \
--disable-dbus_1 \
--disable-ffmpeg_static \ --disable-ffmpeg_static \
--disable-hdhomerun_static \
--disable-libfdkaac_static \ --disable-libfdkaac_static \
--disable-libmfx_static \
--disable-libtheora_static \ --disable-libtheora_static \
--disable-libopus_static \
--disable-libvorbis_static \ --disable-libvorbis_static \
--disable-libvpx_static \ --disable-libvpx_static \
--disable-libx264_static \ --disable-libx264_static \
--disable-libx265_static \ --disable-libx265_static \
--disable-libfdkaac \
--enable-libopus \
--enable-libvorbis \
--enable-libvpx \
--enable-libx264 \
--enable-libx265 \
\
`#Options` \
--disable-avahi \
--disable-dbus_1 \
--disable-bintray_cache \
--disable-hdhomerun_static \
--enable-hdhomerun_client \ --enable-hdhomerun_client \
--enable-libav \ --enable-libav \
--enable-pngquant \ --enable-pngquant \
--enable-trace \
--enable-vaapi \
--infodir=/usr/share/info \ --infodir=/usr/share/info \
--localstatedir=/var \ --localstatedir=/var \
--mandir=/usr/share/man \ --mandir=/usr/share/man \
@@ -191,7 +187,7 @@ RUN \
ARGTABLE_VER1="${ARGTABLE_VER//./-}" && \ ARGTABLE_VER1="${ARGTABLE_VER//./-}" && \
mkdir -p \ mkdir -p \
/tmp/argtable && \ /tmp/argtable && \
curl -s -o \ curl -o \
/tmp/argtable-src.tar.gz -L \ /tmp/argtable-src.tar.gz -L \
"https://sourceforge.net/projects/argtable/files/argtable/argtable-${ARGTABLE_VER}/argtable${ARGTABLE_VER1}.tar.gz" && \ "https://sourceforge.net/projects/argtable/files/argtable/argtable-${ARGTABLE_VER}/argtable${ARGTABLE_VER1}.tar.gz" && \
tar xf \ tar xf \
@@ -219,7 +215,7 @@ RUN \
make DESTDIR=/tmp/comskip-build install make DESTDIR=/tmp/comskip-build install
############## runtime stage ############## ############## runtime stage ##############
FROM ghcr.io/linuxserver/baseimage-alpine:3.12 FROM lsiobase/alpine:3.9
# set version label # set version label
ARG BUILD_DATE ARG BUILD_DATE
@@ -238,21 +234,16 @@ RUN \
curl \ curl \
ffmpeg \ ffmpeg \
ffmpeg-libs \ ffmpeg-libs \
gnu-libiconv \
gzip \ gzip \
libcrypto1.1 \ libcrypto1.1 \
libcurl \ libcurl \
libdvbcsa \
libhdhomerun-libs \ libhdhomerun-libs \
libressl \
libssl1.1 \ libssl1.1 \
libva \
libva-intel-driver \
mesa-dri-ati \
libvpx \ libvpx \
libxml2 \ libxml2 \
libxslt \ libxslt \
linux-headers \ linux-headers \
openssl \
opus \ opus \
pcre2 \ pcre2 \
perl \ perl \
@@ -283,12 +274,10 @@ RUN \
perl-io-socket-ssl \ perl-io-socket-ssl \
perl-io-stringy \ perl-io-stringy \
perl-json \ perl-json \
perl-json-xs \
perl-libwww \ perl-libwww \
perl-lingua-en-numbers-ordinate \ perl-lingua-en-numbers-ordinate \
perl-lingua-preferred \ perl-lingua-preferred \
perl-list-moreutils \ perl-list-moreutils \
perl-lwp-useragent-determined \
perl-module-build \ perl-module-build \
perl-module-pluggable \ perl-module-pluggable \
perl-net-ssleay \ perl-net-ssleay \
@@ -309,23 +298,26 @@ RUN \
perl-xml-treepp \ perl-xml-treepp \
perl-xml-twig \ perl-xml-twig \
perl-xml-writer \ perl-xml-writer \
py3-requests \ python \
python2 \
tar \ tar \
uriparser \ uriparser \
wget \ wget \
x264 \ x264 \
x265 \ x265 \
zlib && \ zlib && \
apk add --no-cache \
--repository http://dl-cdn.alpinelinux.org/alpine/edge/community \
gnu-libiconv && \
echo "**** Add Picons ****" && \ echo "**** Add Picons ****" && \
mkdir -p /picons && \ mkdir -p /picons && \
curl -s -o \ curl -o \
/picons.tar.bz2 -L \ /picons.tar.bz2 -L \
https://lsio-ci.ams3.digitaloceanspaces.com/picons/picons.tar.bz2 https://lsio-ci.ams3.digitaloceanspaces.com/picons/picons.tar.bz2
# copy local files and buildstage artifacts # copy local files and buildstage artifacts
COPY --from=buildstage /tmp/argtable-build/usr/ /usr/ COPY --from=buildstage /tmp/argtable-build/usr/ /usr/
COPY --from=buildstage /tmp/comskip-build/usr/ /usr/ COPY --from=buildstage /tmp/comskip-build/usr/ /usr/
COPY --from=buildstage /tmp/dvbapps-build/usr/ /usr/
COPY --from=buildstage /tmp/tvheadend-build/usr/ /usr/ COPY --from=buildstage /tmp/tvheadend-build/usr/ /usr/
COPY --from=buildstage /tmp/xmltv-build/usr/ /usr/ COPY --from=buildstage /tmp/xmltv-build/usr/ /usr/
COPY --from=buildstage /usr/local/share/man/ /usr/local/share/man/ COPY --from=buildstage /usr/local/share/man/ /usr/local/share/man/
@@ -334,4 +326,4 @@ COPY root/ /
# ports and volumes # ports and volumes
EXPOSE 9981 9982 EXPOSE 9981 9982
VOLUME /config VOLUME /config /recordings

View File

@@ -1,13 +1,14 @@
FROM ghcr.io/linuxserver/baseimage-alpine:arm64v8-3.12 as buildstage FROM lsiobase/alpine:arm64v8-3.9 as buildstage
############## build stage ############## ############## build stage ##############
# package versions # package versions
ARG ARGTABLE_VER="2.13" ARG ARGTABLE_VER="2.13"
ARG XMLTV_VER="v0.6.3" ARG TVH_VER="release/4.2"
ARG XMLTV_VER="v0.5.70"
ARG TVHEADEND_COMMIT
# environment settings # environment settings
ARG TZ="Europe/Oslo" ARG TZ="Europe/Oslo"
ARG TVHEADEND_COMMIT
ENV HOME="/config" ENV HOME="/config"
# copy patches # copy patches
@@ -29,20 +30,18 @@ RUN \
gcc \ gcc \
gettext-dev \ gettext-dev \
git \ git \
gnu-libiconv-dev \
gzip \ gzip \
jq \ jq \
libcurl \
libdvbcsa-dev \
libgcrypt-dev \ libgcrypt-dev \
libhdhomerun-dev \ libhdhomerun-dev \
libressl-dev \
libtool \ libtool \
libvpx-dev \ libvpx-dev \
libxml2-dev \ libxml2-dev \
libxslt-dev \ libxslt-dev \
linux-headers \ linux-headers \
make \ make \
openssl-dev \ mercurial \
opus-dev \ opus-dev \
patch \ patch \
pcre2-dev \ pcre2-dev \
@@ -74,12 +73,10 @@ RUN \
perl-io-socket-ssl \ perl-io-socket-ssl \
perl-io-stringy \ perl-io-stringy \
perl-json \ perl-json \
perl-json-xs \
perl-libwww \ perl-libwww \
perl-lingua-en-numbers-ordinate \ perl-lingua-en-numbers-ordinate \
perl-lingua-preferred \ perl-lingua-preferred \
perl-list-moreutils \ perl-list-moreutils \
perl-lwp-useragent-determined \
perl-module-build \ perl-module-build \
perl-module-pluggable \ perl-module-pluggable \
perl-net-ssleay \ perl-net-ssleay \
@@ -102,14 +99,17 @@ RUN \
perl-xml-writer \ perl-xml-writer \
pkgconf \ pkgconf \
pngquant \ pngquant \
python2 \ python \
sdl-dev \ sdl-dev \
tar \ tar \
uriparser-dev \ uriparser-dev \
wget \ wget \
x264-dev \ x264-dev \
x265-dev \ x265-dev \
zlib-dev zlib-dev && \
apk add --no-cache \
--repository http://dl-cdn.alpinelinux.org/alpine/edge/community \
gnu-libiconv-dev
RUN \ RUN \
echo "**** remove musl iconv.h and replace with gnu-iconv.h ****" && \ echo "**** remove musl iconv.h and replace with gnu-iconv.h ****" && \
@@ -118,7 +118,7 @@ RUN \
RUN \ RUN \
echo "**** install perl modules for xmltv ****" && \ echo "**** install perl modules for xmltv ****" && \
curl -L https://cpanmin.us | perl - App::cpanminus && \ curl -L http://cpanmin.us | perl - App::cpanminus && \
cpanm --installdeps /tmp/patches cpanm --installdeps /tmp/patches
RUN \ RUN \
@@ -139,11 +139,19 @@ RUN \
make test && \ make test && \
make DESTDIR=/tmp/xmltv-build install make DESTDIR=/tmp/xmltv-build install
RUN \
echo "**** compile dvb-apps ****" && \
hg clone http://linuxtv.org/hg/dvb-apps /tmp/dvb-apps && \
cd /tmp/dvb-apps && \
make -C lib && \
make -C lib DESTDIR=/tmp/dvbapps-build install && \
cp -pr /tmp/dvbapps-build/usr/* /usr/
RUN \ RUN \
echo "**** compile tvheadend ****" && \ echo "**** compile tvheadend ****" && \
if [ -z ${TVHEADEND_COMMIT+x} ]; then \ if [ -z ${TVHEADEND_COMMIT+x} ]; then \
TVHEADEND_COMMIT=$(curl -sX GET https://api.github.com/repos/tvheadend/tvheadend/commits/master \ TVHEADEND_COMMIT=$(curl -sX GET https://api.github.com/repos/tvheadend/tvheadend/commits/${TVH_VER} \
| jq -r '. | .sha'); \ | jq -r '. | .sha'); \
fi && \ fi && \
mkdir -p \ mkdir -p \
/tmp/tvheadend && \ /tmp/tvheadend && \
@@ -151,31 +159,21 @@ RUN \
cd /tmp/tvheadend && \ cd /tmp/tvheadend && \
git checkout ${TVHEADEND_COMMIT} && \ git checkout ${TVHEADEND_COMMIT} && \
./configure \ ./configure \
`#Encoding` \ --disable-avahi \
--disable-bintray_cache \
--disable-dbus_1 \
--disable-ffmpeg_static \ --disable-ffmpeg_static \
--disable-hdhomerun_static \
--disable-libfdkaac_static \ --disable-libfdkaac_static \
--disable-libmfx_static \
--disable-libtheora_static \ --disable-libtheora_static \
--disable-libopus_static \
--disable-libvorbis_static \ --disable-libvorbis_static \
--disable-libvpx_static \ --disable-libvpx_static \
--disable-libx264_static \ --disable-libx264_static \
--disable-libx265_static \ --disable-libx265_static \
--disable-libfdkaac \
--enable-libopus \
--enable-libvorbis \
--enable-libvpx \
--enable-libx264 \
--enable-libx265 \
\
`#Options` \
--disable-avahi \
--disable-dbus_1 \
--disable-bintray_cache \
--disable-hdhomerun_static \
--enable-hdhomerun_client \ --enable-hdhomerun_client \
--enable-libav \ --enable-libav \
--enable-pngquant \ --enable-pngquant \
--enable-trace \
--infodir=/usr/share/info \ --infodir=/usr/share/info \
--localstatedir=/var \ --localstatedir=/var \
--mandir=/usr/share/man \ --mandir=/usr/share/man \
@@ -217,7 +215,7 @@ RUN \
make DESTDIR=/tmp/comskip-build install make DESTDIR=/tmp/comskip-build install
############## runtime stage ############## ############## runtime stage ##############
FROM ghcr.io/linuxserver/baseimage-alpine:arm64v8-3.12 FROM lsiobase/alpine:arm64v8-3.9
# set version label # set version label
ARG BUILD_DATE ARG BUILD_DATE
@@ -236,18 +234,16 @@ RUN \
curl \ curl \
ffmpeg \ ffmpeg \
ffmpeg-libs \ ffmpeg-libs \
gnu-libiconv \
gzip \ gzip \
libcrypto1.1 \ libcrypto1.1 \
libcurl \ libcurl \
libdvbcsa \
libhdhomerun-libs \ libhdhomerun-libs \
libressl \
libssl1.1 \ libssl1.1 \
libvpx \ libvpx \
libxml2 \ libxml2 \
libxslt \ libxslt \
linux-headers \ linux-headers \
openssl \
opus \ opus \
pcre2 \ pcre2 \
perl \ perl \
@@ -278,12 +274,10 @@ RUN \
perl-io-socket-ssl \ perl-io-socket-ssl \
perl-io-stringy \ perl-io-stringy \
perl-json \ perl-json \
perl-json-xs \
perl-libwww \ perl-libwww \
perl-lingua-en-numbers-ordinate \ perl-lingua-en-numbers-ordinate \
perl-lingua-preferred \ perl-lingua-preferred \
perl-list-moreutils \ perl-list-moreutils \
perl-lwp-useragent-determined \
perl-module-build \ perl-module-build \
perl-module-pluggable \ perl-module-pluggable \
perl-net-ssleay \ perl-net-ssleay \
@@ -304,14 +298,16 @@ RUN \
perl-xml-treepp \ perl-xml-treepp \
perl-xml-twig \ perl-xml-twig \
perl-xml-writer \ perl-xml-writer \
py3-requests \ python \
python2 \
tar \ tar \
uriparser \ uriparser \
wget \ wget \
x264 \ x264 \
x265 \ x265 \
zlib && \ zlib && \
apk add --no-cache \
--repository http://dl-cdn.alpinelinux.org/alpine/edge/community \
gnu-libiconv && \
echo "**** Add Picons ****" && \ echo "**** Add Picons ****" && \
mkdir -p /picons && \ mkdir -p /picons && \
curl -o \ curl -o \
@@ -321,6 +317,7 @@ RUN \
# copy local files and buildstage artifacts # copy local files and buildstage artifacts
COPY --from=buildstage /tmp/argtable-build/usr/ /usr/ COPY --from=buildstage /tmp/argtable-build/usr/ /usr/
COPY --from=buildstage /tmp/comskip-build/usr/ /usr/ COPY --from=buildstage /tmp/comskip-build/usr/ /usr/
COPY --from=buildstage /tmp/dvbapps-build/usr/ /usr/
COPY --from=buildstage /tmp/tvheadend-build/usr/ /usr/ COPY --from=buildstage /tmp/tvheadend-build/usr/ /usr/
COPY --from=buildstage /tmp/xmltv-build/usr/ /usr/ COPY --from=buildstage /tmp/xmltv-build/usr/ /usr/
COPY --from=buildstage /usr/local/share/man/ /usr/local/share/man/ COPY --from=buildstage /usr/local/share/man/ /usr/local/share/man/
@@ -329,4 +326,4 @@ COPY root/ /
# ports and volumes # ports and volumes
EXPOSE 9981 9982 EXPOSE 9981 9982
VOLUME /config VOLUME /config /recordings

View File

@@ -1,13 +1,14 @@
FROM ghcr.io/linuxserver/baseimage-alpine:arm32v7-3.12 as buildstage FROM lsiobase/alpine:arm32v7-3.9 as buildstage
############## build stage ############## ############## build stage ##############
# package versions # package versions
ARG ARGTABLE_VER="2.13" ARG ARGTABLE_VER="2.13"
ARG XMLTV_VER="v0.6.3" ARG TVH_VER="release/4.2"
ARG XMLTV_VER="v0.5.70"
ARG TVHEADEND_COMMIT
# environment settings # environment settings
ARG TZ="Europe/Oslo" ARG TZ="Europe/Oslo"
ARG TVHEADEND_COMMIT
ENV HOME="/config" ENV HOME="/config"
# copy patches # copy patches
@@ -29,20 +30,18 @@ RUN \
gcc \ gcc \
gettext-dev \ gettext-dev \
git \ git \
gnu-libiconv-dev \
gzip \ gzip \
jq \ jq \
libcurl \
libdvbcsa-dev \
libgcrypt-dev \ libgcrypt-dev \
libhdhomerun-dev \ libhdhomerun-dev \
libressl-dev \
libtool \ libtool \
libvpx-dev \ libvpx-dev \
libxml2-dev \ libxml2-dev \
libxslt-dev \ libxslt-dev \
linux-headers \ linux-headers \
make \ make \
openssl-dev \ mercurial \
opus-dev \ opus-dev \
patch \ patch \
pcre2-dev \ pcre2-dev \
@@ -74,12 +73,10 @@ RUN \
perl-io-socket-ssl \ perl-io-socket-ssl \
perl-io-stringy \ perl-io-stringy \
perl-json \ perl-json \
perl-json-xs \
perl-libwww \ perl-libwww \
perl-lingua-en-numbers-ordinate \ perl-lingua-en-numbers-ordinate \
perl-lingua-preferred \ perl-lingua-preferred \
perl-list-moreutils \ perl-list-moreutils \
perl-lwp-useragent-determined \
perl-module-build \ perl-module-build \
perl-module-pluggable \ perl-module-pluggable \
perl-net-ssleay \ perl-net-ssleay \
@@ -102,14 +99,17 @@ RUN \
perl-xml-writer \ perl-xml-writer \
pkgconf \ pkgconf \
pngquant \ pngquant \
python2 \ python \
sdl-dev \ sdl-dev \
tar \ tar \
uriparser-dev \ uriparser-dev \
wget \ wget \
x264-dev \ x264-dev \
x265-dev \ x265-dev \
zlib-dev zlib-dev && \
apk add --no-cache \
--repository http://dl-cdn.alpinelinux.org/alpine/edge/community \
gnu-libiconv-dev
RUN \ RUN \
echo "**** remove musl iconv.h and replace with gnu-iconv.h ****" && \ echo "**** remove musl iconv.h and replace with gnu-iconv.h ****" && \
@@ -118,7 +118,7 @@ RUN \
RUN \ RUN \
echo "**** install perl modules for xmltv ****" && \ echo "**** install perl modules for xmltv ****" && \
curl -L https://cpanmin.us | perl - App::cpanminus && \ curl -L http://cpanmin.us | perl - App::cpanminus && \
cpanm --installdeps /tmp/patches cpanm --installdeps /tmp/patches
RUN \ RUN \
@@ -139,11 +139,19 @@ RUN \
make test && \ make test && \
make DESTDIR=/tmp/xmltv-build install make DESTDIR=/tmp/xmltv-build install
RUN \
echo "**** compile dvb-apps ****" && \
hg clone http://linuxtv.org/hg/dvb-apps /tmp/dvb-apps && \
cd /tmp/dvb-apps && \
make -C lib && \
make -C lib DESTDIR=/tmp/dvbapps-build install && \
cp -pr /tmp/dvbapps-build/usr/* /usr/
RUN \ RUN \
echo "**** compile tvheadend ****" && \ echo "**** compile tvheadend ****" && \
if [ -z ${TVHEADEND_COMMIT+x} ]; then \ if [ -z ${TVHEADEND_COMMIT+x} ]; then \
TVHEADEND_COMMIT=$(curl -sX GET https://api.github.com/repos/tvheadend/tvheadend/commits/master \ TVHEADEND_COMMIT=$(curl -sX GET https://api.github.com/repos/tvheadend/tvheadend/commits/${TVH_VER} \
| jq -r '. | .sha'); \ | jq -r '. | .sha'); \
fi && \ fi && \
mkdir -p \ mkdir -p \
/tmp/tvheadend && \ /tmp/tvheadend && \
@@ -151,31 +159,21 @@ RUN \
cd /tmp/tvheadend && \ cd /tmp/tvheadend && \
git checkout ${TVHEADEND_COMMIT} && \ git checkout ${TVHEADEND_COMMIT} && \
./configure \ ./configure \
`#Encoding` \ --disable-avahi \
--disable-bintray_cache \
--disable-dbus_1 \
--disable-ffmpeg_static \ --disable-ffmpeg_static \
--disable-hdhomerun_static \
--disable-libfdkaac_static \ --disable-libfdkaac_static \
--disable-libmfx_static \
--disable-libtheora_static \ --disable-libtheora_static \
--disable-libopus_static \
--disable-libvorbis_static \ --disable-libvorbis_static \
--disable-libvpx_static \ --disable-libvpx_static \
--disable-libx264_static \ --disable-libx264_static \
--disable-libx265_static \ --disable-libx265_static \
--disable-libfdkaac \
--enable-libopus \
--enable-libvorbis \
--enable-libvpx \
--enable-libx264 \
--enable-libx265 \
\
`#Options` \
--disable-avahi \
--disable-dbus_1 \
--disable-bintray_cache \
--disable-hdhomerun_static \
--enable-hdhomerun_client \ --enable-hdhomerun_client \
--enable-libav \ --enable-libav \
--enable-pngquant \ --enable-pngquant \
--enable-trace \
--infodir=/usr/share/info \ --infodir=/usr/share/info \
--localstatedir=/var \ --localstatedir=/var \
--mandir=/usr/share/man \ --mandir=/usr/share/man \
@@ -217,7 +215,7 @@ RUN \
make DESTDIR=/tmp/comskip-build install make DESTDIR=/tmp/comskip-build install
############## runtime stage ############## ############## runtime stage ##############
FROM ghcr.io/linuxserver/baseimage-alpine:arm32v7-3.12 FROM lsiobase/alpine:arm32v7-3.9
# set version label # set version label
ARG BUILD_DATE ARG BUILD_DATE
@@ -236,18 +234,16 @@ RUN \
curl \ curl \
ffmpeg \ ffmpeg \
ffmpeg-libs \ ffmpeg-libs \
gnu-libiconv \
gzip \ gzip \
libcrypto1.1 \ libcrypto1.1 \
libcurl \ libcurl \
libdvbcsa \
libhdhomerun-libs \ libhdhomerun-libs \
libressl \
libssl1.1 \ libssl1.1 \
libvpx \ libvpx \
libxml2 \ libxml2 \
libxslt \ libxslt \
linux-headers \ linux-headers \
openssl \
opus \ opus \
pcre2 \ pcre2 \
perl \ perl \
@@ -278,12 +274,10 @@ RUN \
perl-io-socket-ssl \ perl-io-socket-ssl \
perl-io-stringy \ perl-io-stringy \
perl-json \ perl-json \
perl-json-xs \
perl-libwww \ perl-libwww \
perl-lingua-en-numbers-ordinate \ perl-lingua-en-numbers-ordinate \
perl-lingua-preferred \ perl-lingua-preferred \
perl-list-moreutils \ perl-list-moreutils \
perl-lwp-useragent-determined \
perl-module-build \ perl-module-build \
perl-module-pluggable \ perl-module-pluggable \
perl-net-ssleay \ perl-net-ssleay \
@@ -304,14 +298,16 @@ RUN \
perl-xml-treepp \ perl-xml-treepp \
perl-xml-twig \ perl-xml-twig \
perl-xml-writer \ perl-xml-writer \
py3-requests \ python \
python2 \
tar \ tar \
uriparser \ uriparser \
wget \ wget \
x264 \ x264 \
x265 \ x265 \
zlib && \ zlib && \
apk add --no-cache \
--repository http://dl-cdn.alpinelinux.org/alpine/edge/community \
gnu-libiconv && \
echo "**** Add Picons ****" && \ echo "**** Add Picons ****" && \
mkdir -p /picons && \ mkdir -p /picons && \
curl -o \ curl -o \
@@ -321,6 +317,7 @@ RUN \
# copy local files and buildstage artifacts # copy local files and buildstage artifacts
COPY --from=buildstage /tmp/argtable-build/usr/ /usr/ COPY --from=buildstage /tmp/argtable-build/usr/ /usr/
COPY --from=buildstage /tmp/comskip-build/usr/ /usr/ COPY --from=buildstage /tmp/comskip-build/usr/ /usr/
COPY --from=buildstage /tmp/dvbapps-build/usr/ /usr/
COPY --from=buildstage /tmp/tvheadend-build/usr/ /usr/ COPY --from=buildstage /tmp/tvheadend-build/usr/ /usr/
COPY --from=buildstage /tmp/xmltv-build/usr/ /usr/ COPY --from=buildstage /tmp/xmltv-build/usr/ /usr/
COPY --from=buildstage /usr/local/share/man/ /usr/local/share/man/ COPY --from=buildstage /usr/local/share/man/ /usr/local/share/man/
@@ -329,4 +326,4 @@ COPY root/ /
# ports and volumes # ports and volumes
EXPOSE 9981 9982 EXPOSE 9981 9982
VOLUME /config VOLUME /config /recordings

450
Jenkinsfile vendored
View File

@@ -2,10 +2,6 @@ pipeline {
agent { agent {
label 'X86-64-MULTI' label 'X86-64-MULTI'
} }
options {
buildDiscarder(logRotator(numToKeepStr: '10', daysToKeepStr: '60'))
parallelsAlwaysFailFast()
}
// Input to determine if this is a package check // Input to determine if this is a package check
parameters { parameters {
string(defaultValue: 'false', description: 'package check run', name: 'PACKAGE_CHECK') string(defaultValue: 'false', description: 'package check run', name: 'PACKAGE_CHECK')
@@ -14,9 +10,7 @@ pipeline {
environment { environment {
BUILDS_DISCORD=credentials('build_webhook_url') BUILDS_DISCORD=credentials('build_webhook_url')
GITHUB_TOKEN=credentials('498b4638-2d02-4ce5-832d-8a57d01d97ab') GITHUB_TOKEN=credentials('498b4638-2d02-4ce5-832d-8a57d01d97ab')
GITLAB_TOKEN=credentials('b6f0f1dd-6952-4cf6-95d1-9c06380283f0') EXT_GIT_BRANCH = 'release/4.2'
GITLAB_NAMESPACE=credentials('gitlab-namespace-id')
EXT_GIT_BRANCH = 'master'
EXT_USER = 'tvheadend' EXT_USER = 'tvheadend'
EXT_REPO = 'tvheadend' EXT_REPO = 'tvheadend'
BUILD_VERSION_ARG = 'TVHEADEND_COMMIT' BUILD_VERSION_ARG = 'TVHEADEND_COMMIT'
@@ -44,7 +38,7 @@ pipeline {
script{ script{
env.EXIT_STATUS = '' env.EXIT_STATUS = ''
env.LS_RELEASE = sh( env.LS_RELEASE = sh(
script: '''docker run --rm ghcr.io/linuxserver/alexeiled-skopeo sh -c 'skopeo inspect docker://docker.io/'${DOCKERHUB_IMAGE}':latest 2>/dev/null' | jq -r '.Labels.build_version' | awk '{print $3}' | grep '\\-ls' || : ''', script: '''docker run --rm alexeiled/skopeo sh -c 'skopeo inspect docker://docker.io/'${DOCKERHUB_IMAGE}':release-4.2 2>/dev/null' | jq -r '.Labels.build_version' | awk '{print $3}' | grep '\\-ls' || : ''',
returnStdout: true).trim() returnStdout: true).trim()
env.LS_RELEASE_NOTES = sh( env.LS_RELEASE_NOTES = sh(
script: '''cat readme-vars.yml | awk -F \\" '/date: "[0-9][0-9].[0-9][0-9].[0-9][0-9]:/ {print $4;exit;}' | sed -E ':a;N;$!ba;s/\\r{0,1}\\n/\\\\n/g' ''', script: '''cat readme-vars.yml | awk -F \\" '/date: "[0-9][0-9].[0-9][0-9].[0-9][0-9]:/ {print $4;exit;}' | sed -E ':a;N;$!ba;s/\\r{0,1}\\n/\\\\n/g' ''',
@@ -58,7 +52,14 @@ pipeline {
env.CODE_URL = 'https://github.com/' + env.LS_USER + '/' + env.LS_REPO + '/commit/' + env.GIT_COMMIT env.CODE_URL = 'https://github.com/' + env.LS_USER + '/' + env.LS_REPO + '/commit/' + env.GIT_COMMIT
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.DOCKERHUB_IMAGE + '/tags/' env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.DOCKERHUB_IMAGE + '/tags/'
env.PULL_REQUEST = env.CHANGE_ID env.PULL_REQUEST = env.CHANGE_ID
env.TEMPLATED_FILES = 'Jenkinsfile README.md LICENSE ./.github/CONTRIBUTING.md ./.github/FUNDING.yml ./.github/ISSUE_TEMPLATE/config.yml ./.github/ISSUE_TEMPLATE/issue.bug.md ./.github/ISSUE_TEMPLATE/issue.feature.md ./.github/PULL_REQUEST_TEMPLATE.md ./.github/workflows/greetings.yml ./.github/workflows/stale.yml ./.github/workflows/package_trigger.yml ./.github/workflows/package_trigger_scheduler.yml ./.github/workflows/external_trigger.yml ./.github/workflows/external_trigger_scheduler.yml' env.LICENSE_TAG = sh(
script: '''#!/bin/bash
if [ -e LICENSE ] ; then
cat LICENSE | md5sum | cut -c1-8
else
echo none
fi''',
returnStdout: true).trim()
} }
script{ script{
env.LS_RELEASE_NUMBER = sh( env.LS_RELEASE_NUMBER = sh(
@@ -129,47 +130,39 @@ pipeline {
} }
} }
} }
// If this is a master build use live docker endpoints // If this is a release-4.2 build use live docker endpoints
stage("Set ENV live build"){ stage("Set ENV live build"){
when { when {
branch "master" branch "release-4.2"
environment name: 'CHANGE_ID', value: '' environment name: 'CHANGE_ID', value: ''
} }
steps { steps {
script{ script{
env.IMAGE = env.DOCKERHUB_IMAGE env.IMAGE = env.DOCKERHUB_IMAGE
env.GITHUBIMAGE = 'ghcr.io/' + env.LS_USER + '/' + env.CONTAINER_NAME
env.GITLABIMAGE = 'registry.gitlab.com/linuxserver.io/' + env.LS_REPO + '/' + env.CONTAINER_NAME
if (env.MULTIARCH == 'true') { if (env.MULTIARCH == 'true') {
env.CI_TAGS = 'amd64-' + env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER + '|arm32v7-' + env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER + '|arm64v8-' + env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER env.CI_TAGS = 'amd64-' + env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER + '|arm32v7-' + env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER + '|arm64v8-' + env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
} else { } else {
env.CI_TAGS = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER env.CI_TAGS = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
} }
env.VERSION_TAG = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
env.META_TAG = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER env.META_TAG = env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
} }
} }
} }
// If this is a dev build use dev docker endpoints // If this is a dev build use dev docker endpoints
stage("Set ENV dev build"){ stage("Set ENV dev build"){
when { when {
not {branch "master"} not {branch "release-4.2"}
environment name: 'CHANGE_ID', value: '' environment name: 'CHANGE_ID', value: ''
} }
steps { steps {
script{ script{
env.IMAGE = env.DEV_DOCKERHUB_IMAGE env.IMAGE = env.DEV_DOCKERHUB_IMAGE
env.GITHUBIMAGE = 'ghcr.io/' + env.LS_USER + '/lsiodev-' + env.CONTAINER_NAME
env.GITLABIMAGE = 'registry.gitlab.com/linuxserver.io/' + env.LS_REPO + '/lsiodev-' + env.CONTAINER_NAME
if (env.MULTIARCH == 'true') { if (env.MULTIARCH == 'true') {
env.CI_TAGS = 'amd64-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA + '|arm32v7-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA + '|arm64v8-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA env.CI_TAGS = 'amd64-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA + '|arm32v7-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA + '|arm64v8-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA
} else { } else {
env.CI_TAGS = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA env.CI_TAGS = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA
} }
env.VERSION_TAG = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA
env.META_TAG = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA env.META_TAG = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-dev-' + env.COMMIT_SHA
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.DEV_DOCKERHUB_IMAGE + '/tags/' env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.DEV_DOCKERHUB_IMAGE + '/tags/'
} }
} }
@@ -182,16 +175,12 @@ pipeline {
steps { steps {
script{ script{
env.IMAGE = env.PR_DOCKERHUB_IMAGE env.IMAGE = env.PR_DOCKERHUB_IMAGE
env.GITHUBIMAGE = 'ghcr.io/' + env.LS_USER + '/lspipepr-' + env.CONTAINER_NAME
env.GITLABIMAGE = 'registry.gitlab.com/linuxserver.io/' + env.LS_REPO + '/lspipepr-' + env.CONTAINER_NAME
if (env.MULTIARCH == 'true') { if (env.MULTIARCH == 'true') {
env.CI_TAGS = 'amd64-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST + '|arm32v7-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST + '|arm64v8-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST env.CI_TAGS = 'amd64-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST + '|arm32v7-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST + '|arm64v8-' + env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST
} else { } else {
env.CI_TAGS = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST env.CI_TAGS = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST
} }
env.VERSION_TAG = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST
env.META_TAG = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST env.META_TAG = env.EXT_RELEASE_CLEAN + '-pkg-' + env.PACKAGE_TAG + '-pr-' + env.PULL_REQUEST
env.EXT_RELEASE_TAG = 'version-' + env.EXT_RELEASE_CLEAN
env.CODE_URL = 'https://github.com/' + env.LS_USER + '/' + env.LS_REPO + '/pull/' + env.PULL_REQUEST env.CODE_URL = 'https://github.com/' + env.LS_USER + '/' + env.LS_REPO + '/pull/' + env.PULL_REQUEST
env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.PR_DOCKERHUB_IMAGE + '/tags/' env.DOCKERHUB_LINK = 'https://hub.docker.com/r/' + env.PR_DOCKERHUB_IMAGE + '/tags/'
} }
@@ -204,24 +193,24 @@ pipeline {
} }
steps { steps {
withCredentials([ withCredentials([
string(credentialsId: 'ci-tests-s3-key-id', variable: 'S3_KEY'), string(credentialsId: 'spaces-key', variable: 'DO_KEY'),
string(credentialsId: 'ci-tests-s3-secret-access-key', variable: 'S3_SECRET') string(credentialsId: 'spaces-secret', variable: 'DO_SECRET')
]) { ]) {
script{ script{
env.SHELLCHECK_URL = 'https://ci-tests.linuxserver.io/' + env.IMAGE + '/' + env.META_TAG + '/shellcheck-result.xml' env.SHELLCHECK_URL = 'https://lsio-ci.ams3.digitaloceanspaces.com/' + env.IMAGE + '/' + env.META_TAG + '/shellcheck-result.xml'
} }
sh '''curl -sL https://raw.githubusercontent.com/linuxserver/docker-shellcheck/master/checkrun.sh | /bin/bash''' sh '''curl -sL https://raw.githubusercontent.com/linuxserver/docker-shellcheck/master/checkrun.sh | /bin/bash'''
sh '''#! /bin/bash sh '''#! /bin/bash
set -e set -e
docker pull ghcr.io/linuxserver/lsiodev-spaces-file-upload:latest docker pull lsiodev/spaces-file-upload:latest
docker run --rm \ docker run --rm \
-e DESTINATION=\"${IMAGE}/${META_TAG}/shellcheck-result.xml\" \ -e DESTINATION=\"${IMAGE}/${META_TAG}/shellcheck-result.xml\" \
-e FILE_NAME="shellcheck-result.xml" \ -e FILE_NAME="shellcheck-result.xml" \
-e MIMETYPE="text/xml" \ -e MIMETYPE="text/xml" \
-v ${WORKSPACE}:/mnt \ -v ${WORKSPACE}:/mnt \
-e SECRET_KEY=\"${S3_SECRET}\" \ -e SECRET_KEY=\"${DO_SECRET}\" \
-e ACCESS_KEY=\"${S3_KEY}\" \ -e ACCESS_KEY=\"${DO_KEY}\" \
-t ghcr.io/linuxserver/lsiodev-spaces-file-upload:latest \ -t lsiodev/spaces-file-upload:latest \
python /upload.py''' python /upload.py'''
} }
} }
@@ -229,7 +218,7 @@ pipeline {
// Use helper containers to render templated files // Use helper containers to render templated files
stage('Update-Templates') { stage('Update-Templates') {
when { when {
branch "master" branch "release-4.2"
environment name: 'CHANGE_ID', value: '' environment name: 'CHANGE_ID', value: ''
expression { expression {
env.CONTAINER_NAME != null env.CONTAINER_NAME != null
@@ -239,101 +228,36 @@ pipeline {
sh '''#! /bin/bash sh '''#! /bin/bash
set -e set -e
TEMPDIR=$(mktemp -d) TEMPDIR=$(mktemp -d)
docker pull ghcr.io/linuxserver/jenkins-builder:latest docker pull linuxserver/jenkins-builder:latest
docker run --rm -e CONTAINER_NAME=${CONTAINER_NAME} -e GITHUB_BRANCH=master -v ${TEMPDIR}:/ansible/jenkins ghcr.io/linuxserver/jenkins-builder:latest docker run --rm -e CONTAINER_NAME=${CONTAINER_NAME} -e GITHUB_BRANCH=release-4.2 -v ${TEMPDIR}:/ansible/jenkins linuxserver/jenkins-builder:latest
# Stage 1 - Jenkinsfile update docker pull linuxserver/doc-builder:latest
if [[ "$(md5sum Jenkinsfile | awk '{ print $1 }')" != "$(md5sum ${TEMPDIR}/docker-${CONTAINER_NAME}/Jenkinsfile | awk '{ print $1 }')" ]]; then docker run --rm -e CONTAINER_NAME=${CONTAINER_NAME} -e GITHUB_BRANCH=release-4.2 -v ${TEMPDIR}:/ansible/readme linuxserver/doc-builder:latest
if [ "$(md5sum ${TEMPDIR}/${LS_REPO}/Jenkinsfile | awk '{ print $1 }')" != "$(md5sum Jenkinsfile | awk '{ print $1 }')" ] || \
[ "$(md5sum ${TEMPDIR}/${CONTAINER_NAME}/README.md | awk '{ print $1 }')" != "$(md5sum README.md | awk '{ print $1 }')" ] || \
[ "$(cat ${TEMPDIR}/${LS_REPO}/LICENSE | md5sum | cut -c1-8)" != "${LICENSE_TAG}" ]; then
mkdir -p ${TEMPDIR}/repo mkdir -p ${TEMPDIR}/repo
git clone https://github.com/${LS_USER}/${LS_REPO}.git ${TEMPDIR}/repo/${LS_REPO} git clone https://github.com/${LS_USER}/${LS_REPO}.git ${TEMPDIR}/repo/${LS_REPO}
cd ${TEMPDIR}/repo/${LS_REPO} git --git-dir ${TEMPDIR}/repo/${LS_REPO}/.git checkout -f release-4.2
git checkout -f master cp ${TEMPDIR}/${CONTAINER_NAME}/README.md ${TEMPDIR}/repo/${LS_REPO}/
cp ${TEMPDIR}/docker-${CONTAINER_NAME}/Jenkinsfile ${TEMPDIR}/repo/${LS_REPO}/ cp ${TEMPDIR}/docker-${CONTAINER_NAME}/Jenkinsfile ${TEMPDIR}/repo/${LS_REPO}/
git add Jenkinsfile cp ${TEMPDIR}/docker-${CONTAINER_NAME}/LICENSE ${TEMPDIR}/repo/${LS_REPO}/
git commit -m 'Bot Updating Templated Files'
git push https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/${LS_USER}/${LS_REPO}.git --all
echo "true" > /tmp/${COMMIT_SHA}-${BUILD_NUMBER}
echo "Updating Jenkinsfile"
rm -Rf ${TEMPDIR}
exit 0
else
echo "Jenkinsfile is up to date."
fi
# Stage 2 - Delete old templates
OLD_TEMPLATES=".github/ISSUE_TEMPLATE.md"
for i in ${OLD_TEMPLATES}; do
if [[ -f "${i}" ]]; then
TEMPLATES_TO_DELETE="${i} ${TEMPLATES_TO_DELETE}"
fi
done
if [[ -n "${TEMPLATES_TO_DELETE}" ]]; then
mkdir -p ${TEMPDIR}/repo
git clone https://github.com/${LS_USER}/${LS_REPO}.git ${TEMPDIR}/repo/${LS_REPO}
cd ${TEMPDIR}/repo/${LS_REPO}
git checkout -f master
cd ${TEMPDIR}/docker-${CONTAINER_NAME}
for i in ${TEMPLATES_TO_DELETE}; do
git rm "${i}"
done
git commit -m 'Bot Updating Templated Files'
git push https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/${LS_USER}/${LS_REPO}.git --all
echo "true" > /tmp/${COMMIT_SHA}-${BUILD_NUMBER}
echo "Deleting old templates"
rm -Rf ${TEMPDIR}
exit 0
else
echo "No templates to delete"
fi
# Stage 3 - Update templates
CURRENTHASH=$(grep -hs ^ ${TEMPLATED_FILES} | md5sum | cut -c1-8)
cd ${TEMPDIR}/docker-${CONTAINER_NAME}
NEWHASH=$(grep -hs ^ ${TEMPLATED_FILES} | md5sum | cut -c1-8)
if [[ "${CURRENTHASH}" != "${NEWHASH}" ]] || ! grep -q '.jenkins-external' "${WORKSPACE}/.gitignore" 2>/dev/null; then
mkdir -p ${TEMPDIR}/repo
git clone https://github.com/${LS_USER}/${LS_REPO}.git ${TEMPDIR}/repo/${LS_REPO}
cd ${TEMPDIR}/repo/${LS_REPO}
git checkout -f master
cd ${TEMPDIR}/docker-${CONTAINER_NAME}
mkdir -p ${TEMPDIR}/repo/${LS_REPO}/.github/workflows
mkdir -p ${TEMPDIR}/repo/${LS_REPO}/.github/ISSUE_TEMPLATE
cp --parents ${TEMPLATED_FILES} ${TEMPDIR}/repo/${LS_REPO}/ || :
cd ${TEMPDIR}/repo/${LS_REPO}/ cd ${TEMPDIR}/repo/${LS_REPO}/
if ! grep -q '.jenkins-external' .gitignore 2>/dev/null; then git --git-dir ${TEMPDIR}/repo/${LS_REPO}/.git add Jenkinsfile README.md LICENSE
echo ".jenkins-external" >> .gitignore git --git-dir ${TEMPDIR}/repo/${LS_REPO}/.git commit -m 'Bot Updating Templated Files'
git add .gitignore git --git-dir ${TEMPDIR}/repo/${LS_REPO}/.git push https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/${LS_USER}/${LS_REPO}.git --all
fi
git add ${TEMPLATED_FILES}
git commit -m 'Bot Updating Templated Files'
git push https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/${LS_USER}/${LS_REPO}.git --all
echo "true" > /tmp/${COMMIT_SHA}-${BUILD_NUMBER} echo "true" > /tmp/${COMMIT_SHA}-${BUILD_NUMBER}
else else
echo "false" > /tmp/${COMMIT_SHA}-${BUILD_NUMBER} echo "false" > /tmp/${COMMIT_SHA}-${BUILD_NUMBER}
fi fi
mkdir -p ${TEMPDIR}/gitbook mkdir -p ${TEMPDIR}/gitbook
git clone https://github.com/linuxserver/docker-documentation.git ${TEMPDIR}/gitbook/docker-documentation git clone https://github.com/linuxserver/docker-documentation.git ${TEMPDIR}/gitbook/docker-documentation
if [[ ("${BRANCH_NAME}" == "master") || ("${BRANCH_NAME}" == "main") ]] && [[ (! -f ${TEMPDIR}/gitbook/docker-documentation/images/docker-${CONTAINER_NAME}.md) || ("$(md5sum ${TEMPDIR}/gitbook/docker-documentation/images/docker-${CONTAINER_NAME}.md | awk '{ print $1 }')" != "$(md5sum ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/docker-${CONTAINER_NAME}.md | awk '{ print $1 }')") ]]; then if [[ "${BRANCH_NAME}" == "master" ]] && [[ (! -f ${TEMPDIR}/gitbook/docker-documentation/images/docker-${CONTAINER_NAME}.md) || ("$(md5sum ${TEMPDIR}/gitbook/docker-documentation/images/docker-${CONTAINER_NAME}.md | awk '{ print $1 }')" != "$(md5sum ${TEMPDIR}/${CONTAINER_NAME}/docker-${CONTAINER_NAME}.md | awk '{ print $1 }')") ]]; then
cp ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/docker-${CONTAINER_NAME}.md ${TEMPDIR}/gitbook/docker-documentation/images/ cp ${TEMPDIR}/${CONTAINER_NAME}/docker-${CONTAINER_NAME}.md ${TEMPDIR}/gitbook/docker-documentation/images/
cd ${TEMPDIR}/gitbook/docker-documentation/ cd ${TEMPDIR}/gitbook/docker-documentation/
git add images/docker-${CONTAINER_NAME}.md git add images/docker-${CONTAINER_NAME}.md
git commit -m 'Bot Updating Documentation' git commit -m 'Bot Updating Templated Files'
git push https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/linuxserver/docker-documentation.git --all git push https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/linuxserver/docker-documentation.git --all
fi fi
mkdir -p ${TEMPDIR}/unraid
git clone https://github.com/linuxserver/docker-templates.git ${TEMPDIR}/unraid/docker-templates
git clone https://github.com/linuxserver/templates.git ${TEMPDIR}/unraid/templates
if [[ -f ${TEMPDIR}/unraid/docker-templates/linuxserver.io/img/${CONTAINER_NAME}-icon.png ]]; then
sed -i "s|master/linuxserver.io/img/linuxserver-ls-logo.png|master/linuxserver.io/img/${CONTAINER_NAME}-icon.png|" ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/${CONTAINER_NAME}.xml
fi
if [[ ("${BRANCH_NAME}" == "master") || ("${BRANCH_NAME}" == "main") ]] && [[ (! -f ${TEMPDIR}/unraid/templates/unraid/${CONTAINER_NAME}.xml) || ("$(md5sum ${TEMPDIR}/unraid/templates/unraid/${CONTAINER_NAME}.xml | awk '{ print $1 }')" != "$(md5sum ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/${CONTAINER_NAME}.xml | awk '{ print $1 }')") ]]; then
if grep -wq "${CONTAINER_NAME}" ${TEMPDIR}/unraid/templates/unraid/ignore.list; then
echo "Image is on the ignore list, skipping Unraid template upload"
else
cp ${TEMPDIR}/docker-${CONTAINER_NAME}/.jenkins-external/${CONTAINER_NAME}.xml ${TEMPDIR}/unraid/templates/unraid/
cd ${TEMPDIR}/unraid/templates/
git add unraid/${CONTAINER_NAME}.xml
git commit -m 'Bot Updating Unraid Template'
git push https://LinuxServer-CI:${GITHUB_TOKEN}@github.com/linuxserver/templates.git --all
fi
fi
rm -Rf ${TEMPDIR}''' rm -Rf ${TEMPDIR}'''
script{ script{
env.FILES_UPDATED = sh( env.FILES_UPDATED = sh(
@@ -345,7 +269,7 @@ pipeline {
// Exit the build if the Templated files were just updated // Exit the build if the Templated files were just updated
stage('Template-exit') { stage('Template-exit') {
when { when {
branch "master" branch "release-4.2"
environment name: 'CHANGE_ID', value: '' environment name: 'CHANGE_ID', value: ''
environment name: 'FILES_UPDATED', value: 'true' environment name: 'FILES_UPDATED', value: 'true'
expression { expression {
@@ -358,26 +282,6 @@ pipeline {
} }
} }
} }
/* #######################
GitLab Mirroring
####################### */
// Ping into Gitlab to mirror this repo and have a registry endpoint
stage("GitLab Mirror"){
when {
environment name: 'EXIT_STATUS', value: ''
}
steps{
sh '''curl -H "Content-Type: application/json" -H "Private-Token: ${GITLAB_TOKEN}" -X POST https://gitlab.com/api/v4/projects \
-d '{"namespace_id":'${GITLAB_NAMESPACE}',\
"name":"'${LS_REPO}'",
"mirror":true,\
"import_url":"https://github.com/linuxserver/'${LS_REPO}'.git",\
"issues_access_level":"disabled",\
"merge_requests_access_level":"disabled",\
"repository_access_level":"enabled",\
"visibility":"public"}' '''
}
}
/* ############### /* ###############
Build Container Build Container
############### */ ############### */
@@ -388,9 +292,8 @@ pipeline {
environment name: 'EXIT_STATUS', value: '' environment name: 'EXIT_STATUS', value: ''
} }
steps { steps {
echo "Running on node: ${NODE_NAME}"
sh "docker build --no-cache --pull -t ${IMAGE}:${META_TAG} \ sh "docker build --no-cache --pull -t ${IMAGE}:${META_TAG} \
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ." --build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${META_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
} }
} }
// Build MultiArch Docker containers for push to LS Repo // Build MultiArch Docker containers for push to LS Repo
@@ -402,9 +305,8 @@ pipeline {
parallel { parallel {
stage('Build X86') { stage('Build X86') {
steps { steps {
echo "Running on node: ${NODE_NAME}"
sh "docker build --no-cache --pull -t ${IMAGE}:amd64-${META_TAG} \ sh "docker build --no-cache --pull -t ${IMAGE}:amd64-${META_TAG} \
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ." --build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${META_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
} }
} }
stage('Build ARMHF') { stage('Build ARMHF') {
@@ -412,20 +314,26 @@ pipeline {
label 'ARMHF' label 'ARMHF'
} }
steps { steps {
echo "Running on node: ${NODE_NAME}" withCredentials([
echo 'Logging into Github' [
sh '''#! /bin/bash $class: 'UsernamePasswordMultiBinding',
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin credentialsId: '3f9ba4d5-100d-45b0-a3c4-633fd6061207',
''' usernameVariable: 'DOCKERUSER',
sh "docker build --no-cache --pull -f Dockerfile.armhf -t ${IMAGE}:arm32v7-${META_TAG} \ passwordVariable: 'DOCKERPASS'
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ." ]
sh "docker tag ${IMAGE}:arm32v7-${META_TAG} ghcr.io/linuxserver/lsiodev-buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER}" ]) {
retry(5) { echo 'Logging into DockerHub'
sh "docker push ghcr.io/linuxserver/lsiodev-buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER}" sh '''#! /bin/bash
echo $DOCKERPASS | docker login -u $DOCKERUSER --password-stdin
'''
sh "docker build --no-cache --pull -f Dockerfile.armhf -t ${IMAGE}:arm32v7-${META_TAG} \
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${META_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
sh "docker tag ${IMAGE}:arm32v7-${META_TAG} lsiodev/buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER}"
sh "docker push lsiodev/buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER}"
sh '''docker rmi \
${IMAGE}:arm32v7-${META_TAG} \
lsiodev/buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} || :'''
} }
sh '''docker rmi \
${IMAGE}:arm32v7-${META_TAG} \
ghcr.io/linuxserver/lsiodev-buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} || :'''
} }
} }
stage('Build ARM64') { stage('Build ARM64') {
@@ -433,20 +341,26 @@ pipeline {
label 'ARM64' label 'ARM64'
} }
steps { steps {
echo "Running on node: ${NODE_NAME}" withCredentials([
echo 'Logging into Github' [
sh '''#! /bin/bash $class: 'UsernamePasswordMultiBinding',
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin credentialsId: '3f9ba4d5-100d-45b0-a3c4-633fd6061207',
''' usernameVariable: 'DOCKERUSER',
sh "docker build --no-cache --pull -f Dockerfile.aarch64 -t ${IMAGE}:arm64v8-${META_TAG} \ passwordVariable: 'DOCKERPASS'
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${VERSION_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ." ]
sh "docker tag ${IMAGE}:arm64v8-${META_TAG} ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}" ]) {
retry(5) { echo 'Logging into DockerHub'
sh "docker push ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}" sh '''#! /bin/bash
echo $DOCKERPASS | docker login -u $DOCKERUSER --password-stdin
'''
sh "docker build --no-cache --pull -f Dockerfile.aarch64 -t ${IMAGE}:arm64v8-${META_TAG} \
--build-arg ${BUILD_VERSION_ARG}=${EXT_RELEASE} --build-arg VERSION=\"${META_TAG}\" --build-arg BUILD_DATE=${GITHUB_DATE} ."
sh "docker tag ${IMAGE}:arm64v8-${META_TAG} lsiodev/buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}"
sh "docker push lsiodev/buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}"
sh '''docker rmi \
${IMAGE}:arm64v8-${META_TAG} \
lsiodev/buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} || :'''
} }
sh '''docker rmi \
${IMAGE}:arm64v8-${META_TAG} \
ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} || :'''
} }
} }
} }
@@ -454,7 +368,7 @@ pipeline {
// Take the image we just built and dump package versions for comparison // Take the image we just built and dump package versions for comparison
stage('Update-packages') { stage('Update-packages') {
when { when {
branch "master" branch "release-4.2"
environment name: 'CHANGE_ID', value: '' environment name: 'CHANGE_ID', value: ''
environment name: 'EXIT_STATUS', value: '' environment name: 'EXIT_STATUS', value: ''
} }
@@ -482,7 +396,7 @@ pipeline {
echo "Package tag sha from current packages in buit container is ${NEW_PACKAGE_TAG} comparing to old ${PACKAGE_TAG} from github" echo "Package tag sha from current packages in buit container is ${NEW_PACKAGE_TAG} comparing to old ${PACKAGE_TAG} from github"
if [ "${NEW_PACKAGE_TAG}" != "${PACKAGE_TAG}" ]; then if [ "${NEW_PACKAGE_TAG}" != "${PACKAGE_TAG}" ]; then
git clone https://github.com/${LS_USER}/${LS_REPO}.git ${TEMPDIR}/${LS_REPO} git clone https://github.com/${LS_USER}/${LS_REPO}.git ${TEMPDIR}/${LS_REPO}
git --git-dir ${TEMPDIR}/${LS_REPO}/.git checkout -f master git --git-dir ${TEMPDIR}/${LS_REPO}/.git checkout -f release-4.2
cp ${TEMPDIR}/package_versions.txt ${TEMPDIR}/${LS_REPO}/ cp ${TEMPDIR}/package_versions.txt ${TEMPDIR}/${LS_REPO}/
cd ${TEMPDIR}/${LS_REPO}/ cd ${TEMPDIR}/${LS_REPO}/
wait wait
@@ -506,19 +420,12 @@ pipeline {
// Exit the build if the package file was just updated // Exit the build if the package file was just updated
stage('PACKAGE-exit') { stage('PACKAGE-exit') {
when { when {
branch "master" branch "release-4.2"
environment name: 'CHANGE_ID', value: '' environment name: 'CHANGE_ID', value: ''
environment name: 'PACKAGE_UPDATED', value: 'true' environment name: 'PACKAGE_UPDATED', value: 'true'
environment name: 'EXIT_STATUS', value: '' environment name: 'EXIT_STATUS', value: ''
} }
steps { steps {
sh '''#! /bin/bash
echo "Packages were updated. Cleaning up the image and exiting."
if [ "${MULTIARCH}" == "true" ]; then
docker rmi ${IMAGE}:amd64-${META_TAG}
else
docker rmi ${IMAGE}:${META_TAG}
fi'''
script{ script{
env.EXIT_STATUS = 'ABORTED' env.EXIT_STATUS = 'ABORTED'
} }
@@ -527,7 +434,7 @@ pipeline {
// Exit the build if this is just a package check and there are no changes to push // Exit the build if this is just a package check and there are no changes to push
stage('PACKAGECHECK-exit') { stage('PACKAGECHECK-exit') {
when { when {
branch "master" branch "release-4.2"
environment name: 'CHANGE_ID', value: '' environment name: 'CHANGE_ID', value: ''
environment name: 'PACKAGE_UPDATED', value: 'false' environment name: 'PACKAGE_UPDATED', value: 'false'
environment name: 'EXIT_STATUS', value: '' environment name: 'EXIT_STATUS', value: ''
@@ -536,13 +443,6 @@ pipeline {
} }
} }
steps { steps {
sh '''#! /bin/bash
echo "There are no package updates. Cleaning up the image and exiting."
if [ "${MULTIARCH}" == "true" ]; then
docker rmi ${IMAGE}:amd64-${META_TAG}
else
docker rmi ${IMAGE}:${META_TAG}
fi'''
script{ script{
env.EXIT_STATUS = 'ABORTED' env.EXIT_STATUS = 'ABORTED'
} }
@@ -559,20 +459,20 @@ pipeline {
} }
steps { steps {
withCredentials([ withCredentials([
string(credentialsId: 'ci-tests-s3-key-id', variable: 'S3_KEY'), string(credentialsId: 'spaces-key', variable: 'DO_KEY'),
string(credentialsId: 'ci-tests-s3-secret-access-key ', variable: 'S3_SECRET') string(credentialsId: 'spaces-secret', variable: 'DO_SECRET')
]) { ]) {
script{ script{
env.CI_URL = 'https://ci-tests.linuxserver.io/' + env.IMAGE + '/' + env.META_TAG + '/index.html' env.CI_URL = 'https://lsio-ci.ams3.digitaloceanspaces.com/' + env.IMAGE + '/' + env.META_TAG + '/index.html'
} }
sh '''#! /bin/bash sh '''#! /bin/bash
set -e set -e
docker pull ghcr.io/linuxserver/ci:latest docker pull lsiodev/ci:latest
if [ "${MULTIARCH}" == "true" ]; then if [ "${MULTIARCH}" == "true" ]; then
docker pull ghcr.io/linuxserver/lsiodev-buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} docker pull lsiodev/buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER}
docker pull ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} docker pull lsiodev/buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}
docker tag ghcr.io/linuxserver/lsiodev-buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm32v7-${META_TAG} docker tag lsiodev/buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm32v7-${META_TAG}
docker tag ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm64v8-${META_TAG} docker tag lsiodev/buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm64v8-${META_TAG}
fi fi
docker run --rm \ docker run --rm \
--shm-size=1gb \ --shm-size=1gb \
@@ -584,15 +484,15 @@ pipeline {
-e PORT=\"${CI_PORT}\" \ -e PORT=\"${CI_PORT}\" \
-e SSL=\"${CI_SSL}\" \ -e SSL=\"${CI_SSL}\" \
-e BASE=\"${DIST_IMAGE}\" \ -e BASE=\"${DIST_IMAGE}\" \
-e SECRET_KEY=\"${S3_SECRET}\" \ -e SECRET_KEY=\"${DO_SECRET}\" \
-e ACCESS_KEY=\"${S3_KEY}\" \ -e ACCESS_KEY=\"${DO_KEY}\" \
-e DOCKER_ENV=\"${CI_DOCKERENV}\" \ -e DOCKER_ENV=\"${CI_DOCKERENV}\" \
-e WEB_SCREENSHOT=\"${CI_WEB}\" \ -e WEB_SCREENSHOT=\"${CI_WEB}\" \
-e WEB_AUTH=\"${CI_AUTH}\" \ -e WEB_AUTH=\"${CI_AUTH}\" \
-e WEB_PATH=\"${CI_WEBPATH}\" \ -e WEB_PATH=\"${CI_WEBPATH}\" \
-e DO_REGION="ams3" \ -e DO_REGION="ams3" \
-e DO_BUCKET="lsio-ci" \ -e DO_BUCKET="lsio-ci" \
-t ghcr.io/linuxserver/ci:latest \ -t lsiodev/ci:latest \
python /ci/ci.py''' python /ci/ci.py'''
} }
} }
@@ -615,30 +515,17 @@ pipeline {
passwordVariable: 'DOCKERPASS' passwordVariable: 'DOCKERPASS'
] ]
]) { ]) {
retry(5) { echo 'Logging into DockerHub'
sh '''#! /bin/bash
set -e
echo $DOCKERPASS | docker login -u $DOCKERUSER --password-stdin
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin
echo $GITLAB_TOKEN | docker login registry.gitlab.com -u LinuxServer.io --password-stdin
for PUSHIMAGE in "${GITHUBIMAGE}" "${GITLABIMAGE}" "${IMAGE}"; do
docker tag ${IMAGE}:${META_TAG} ${PUSHIMAGE}:${META_TAG}
docker tag ${PUSHIMAGE}:${META_TAG} ${PUSHIMAGE}:latest
docker tag ${PUSHIMAGE}:${META_TAG} ${PUSHIMAGE}:${EXT_RELEASE_TAG}
docker push ${PUSHIMAGE}:latest
docker push ${PUSHIMAGE}:${META_TAG}
docker push ${PUSHIMAGE}:${EXT_RELEASE_TAG}
done
'''
}
sh '''#! /bin/bash sh '''#! /bin/bash
for DELETEIMAGE in "${GITHUBIMAGE}" "${GITLABIMAGE}" "${IMAGE}"; do echo $DOCKERPASS | docker login -u $DOCKERUSER --password-stdin
docker rmi \
${DELETEIMAGE}:${META_TAG} \
${DELETEIMAGE}:${EXT_RELEASE_TAG} \
${DELETEIMAGE}:latest || :
done
''' '''
sh "docker tag ${IMAGE}:${META_TAG} ${IMAGE}:release-4.2"
sh "docker push ${IMAGE}:release-4.2"
sh "docker push ${IMAGE}:${META_TAG}"
sh '''docker rmi \
${IMAGE}:${META_TAG} \
${IMAGE}:release-4.2 || :'''
} }
} }
} }
@@ -657,79 +544,51 @@ pipeline {
passwordVariable: 'DOCKERPASS' passwordVariable: 'DOCKERPASS'
] ]
]) { ]) {
retry(5) {
sh '''#! /bin/bash
set -e
echo $DOCKERPASS | docker login -u $DOCKERUSER --password-stdin
echo $GITHUB_TOKEN | docker login ghcr.io -u LinuxServer-CI --password-stdin
echo $GITLAB_TOKEN | docker login registry.gitlab.com -u LinuxServer.io --password-stdin
if [ "${CI}" == "false" ]; then
docker pull ghcr.io/linuxserver/lsiodev-buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER}
docker pull ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}
docker tag ghcr.io/linuxserver/lsiodev-buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm32v7-${META_TAG}
docker tag ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm64v8-${META_TAG}
fi
for MANIFESTIMAGE in "${IMAGE}" "${GITLABIMAGE}" "${GITHUBIMAGE}"; do
docker tag ${IMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:amd64-${META_TAG}
docker tag ${IMAGE}:arm32v7-${META_TAG} ${MANIFESTIMAGE}:arm32v7-${META_TAG}
docker tag ${IMAGE}:arm64v8-${META_TAG} ${MANIFESTIMAGE}:arm64v8-${META_TAG}
docker tag ${MANIFESTIMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:amd64-latest
docker tag ${MANIFESTIMAGE}:arm32v7-${META_TAG} ${MANIFESTIMAGE}:arm32v7-latest
docker tag ${MANIFESTIMAGE}:arm64v8-${META_TAG} ${MANIFESTIMAGE}:arm64v8-latest
docker tag ${MANIFESTIMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:amd64-${EXT_RELEASE_TAG}
docker tag ${MANIFESTIMAGE}:arm32v7-${META_TAG} ${MANIFESTIMAGE}:arm32v7-${EXT_RELEASE_TAG}
docker tag ${MANIFESTIMAGE}:arm64v8-${META_TAG} ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG}
docker push ${MANIFESTIMAGE}:amd64-${META_TAG}
docker push ${MANIFESTIMAGE}:arm32v7-${META_TAG}
docker push ${MANIFESTIMAGE}:arm64v8-${META_TAG}
docker push ${MANIFESTIMAGE}:amd64-latest
docker push ${MANIFESTIMAGE}:arm32v7-latest
docker push ${MANIFESTIMAGE}:arm64v8-latest
docker push ${MANIFESTIMAGE}:amd64-${EXT_RELEASE_TAG}
docker push ${MANIFESTIMAGE}:arm32v7-${EXT_RELEASE_TAG}
docker push ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG}
docker manifest push --purge ${MANIFESTIMAGE}:latest || :
docker manifest create ${MANIFESTIMAGE}:latest ${MANIFESTIMAGE}:amd64-latest ${MANIFESTIMAGE}:arm32v7-latest ${MANIFESTIMAGE}:arm64v8-latest
docker manifest annotate ${MANIFESTIMAGE}:latest ${MANIFESTIMAGE}:arm32v7-latest --os linux --arch arm
docker manifest annotate ${MANIFESTIMAGE}:latest ${MANIFESTIMAGE}:arm64v8-latest --os linux --arch arm64 --variant v8
docker manifest push --purge ${MANIFESTIMAGE}:${META_TAG} || :
docker manifest create ${MANIFESTIMAGE}:${META_TAG} ${MANIFESTIMAGE}:amd64-${META_TAG} ${MANIFESTIMAGE}:arm32v7-${META_TAG} ${MANIFESTIMAGE}:arm64v8-${META_TAG}
docker manifest annotate ${MANIFESTIMAGE}:${META_TAG} ${MANIFESTIMAGE}:arm32v7-${META_TAG} --os linux --arch arm
docker manifest annotate ${MANIFESTIMAGE}:${META_TAG} ${MANIFESTIMAGE}:arm64v8-${META_TAG} --os linux --arch arm64 --variant v8
docker manifest push --purge ${MANIFESTIMAGE}:${EXT_RELEASE_TAG} || :
docker manifest create ${MANIFESTIMAGE}:${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:amd64-${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:arm32v7-${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG}
docker manifest annotate ${MANIFESTIMAGE}:${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:arm32v7-${EXT_RELEASE_TAG} --os linux --arch arm
docker manifest annotate ${MANIFESTIMAGE}:${EXT_RELEASE_TAG} ${MANIFESTIMAGE}:arm64v8-${EXT_RELEASE_TAG} --os linux --arch arm64 --variant v8
docker manifest push --purge ${MANIFESTIMAGE}:latest
docker manifest push --purge ${MANIFESTIMAGE}:${META_TAG}
docker manifest push --purge ${MANIFESTIMAGE}:${EXT_RELEASE_TAG}
done
'''
}
sh '''#! /bin/bash sh '''#! /bin/bash
for DELETEIMAGE in "${GITHUBIMAGE}" "${GITLABIMAGE}" "${IMAGE}"; do echo $DOCKERPASS | docker login -u $DOCKERUSER --password-stdin
docker rmi \
${DELETEIMAGE}:amd64-${META_TAG} \
${DELETEIMAGE}:amd64-latest \
${DELETEIMAGE}:amd64-${EXT_RELEASE_TAG} \
${DELETEIMAGE}:arm32v7-${META_TAG} \
${DELETEIMAGE}:arm32v7-latest \
${DELETEIMAGE}:arm32v7-${EXT_RELEASE_TAG} \
${DELETEIMAGE}:arm64v8-${META_TAG} \
${DELETEIMAGE}:arm64v8-latest \
${DELETEIMAGE}:arm64v8-${EXT_RELEASE_TAG} || :
done
docker rmi \
ghcr.io/linuxserver/lsiodev-buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} \
ghcr.io/linuxserver/lsiodev-buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} || :
''' '''
sh '''#! /bin/bash
if [ "${CI}" == "false" ]; then
docker pull lsiodev/buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER}
docker pull lsiodev/buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER}
docker tag lsiodev/buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm32v7-${META_TAG}
docker tag lsiodev/buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} ${IMAGE}:arm64v8-${META_TAG}
fi'''
sh "docker tag ${IMAGE}:amd64-${META_TAG} ${IMAGE}:amd64-release-4.2"
sh "docker tag ${IMAGE}:arm32v7-${META_TAG} ${IMAGE}:arm32v7-release-4.2"
sh "docker tag ${IMAGE}:arm64v8-${META_TAG} ${IMAGE}:arm64v8-release-4.2"
sh "docker push ${IMAGE}:amd64-${META_TAG}"
sh "docker push ${IMAGE}:arm32v7-${META_TAG}"
sh "docker push ${IMAGE}:arm64v8-${META_TAG}"
sh "docker push ${IMAGE}:amd64-release-4.2"
sh "docker push ${IMAGE}:arm32v7-release-4.2"
sh "docker push ${IMAGE}:arm64v8-release-4.2"
sh "docker manifest push --purge ${IMAGE}:release-4.2 || :"
sh "docker manifest create ${IMAGE}:release-4.2 ${IMAGE}:amd64-release-4.2 ${IMAGE}:arm32v7-release-4.2 ${IMAGE}:arm64v8-release-4.2"
sh "docker manifest annotate ${IMAGE}:release-4.2 ${IMAGE}:arm32v7-release-4.2 --os linux --arch arm"
sh "docker manifest annotate ${IMAGE}:release-4.2 ${IMAGE}:arm64v8-release-4.2 --os linux --arch arm64 --variant v8"
sh "docker manifest push --purge ${IMAGE}:${META_TAG} || :"
sh "docker manifest create ${IMAGE}:${META_TAG} ${IMAGE}:amd64-${META_TAG} ${IMAGE}:arm32v7-${META_TAG} ${IMAGE}:arm64v8-${META_TAG}"
sh "docker manifest annotate ${IMAGE}:${META_TAG} ${IMAGE}:arm32v7-${META_TAG} --os linux --arch arm"
sh "docker manifest annotate ${IMAGE}:${META_TAG} ${IMAGE}:arm64v8-${META_TAG} --os linux --arch arm64 --variant v8"
sh "docker manifest push --purge ${IMAGE}:release-4.2"
sh "docker manifest push --purge ${IMAGE}:${META_TAG}"
sh '''docker rmi \
${IMAGE}:amd64-${META_TAG} \
${IMAGE}:amd64-release-4.2 \
${IMAGE}:arm32v7-${META_TAG} \
${IMAGE}:arm32v7-release-4.2 \
${IMAGE}:arm64v8-${META_TAG} \
${IMAGE}:arm64v8-release-4.2 \
lsiodev/buildcache:arm32v7-${COMMIT_SHA}-${BUILD_NUMBER} \
lsiodev/buildcache:arm64v8-${COMMIT_SHA}-${BUILD_NUMBER} || :'''
} }
} }
} }
// If this is a public release tag it in the LS Github // If this is a public release tag it in the LS Github
stage('Github-Tag-Push-Release') { stage('Github-Tag-Push-Release') {
when { when {
branch "master" branch "release-4.2"
expression { expression {
env.LS_RELEASE != env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER env.LS_RELEASE != env.EXT_RELEASE_CLEAN + '-ls' + env.LS_TAG_NUMBER
} }
@@ -737,21 +596,21 @@ pipeline {
environment name: 'EXIT_STATUS', value: '' environment name: 'EXIT_STATUS', value: ''
} }
steps { steps {
echo "Pushing New tag for current commit ${META_TAG}" echo "Pushing New tag for current commit ${EXT_RELEASE_CLEAN}-ls${LS_TAG_NUMBER}"
sh '''curl -H "Authorization: token ${GITHUB_TOKEN}" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/git/tags \ sh '''curl -H "Authorization: token ${GITHUB_TOKEN}" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/git/tags \
-d '{"tag":"'${META_TAG}'",\ -d '{"tag":"'${EXT_RELEASE_CLEAN}'-ls'${LS_TAG_NUMBER}'",\
"object": "'${COMMIT_SHA}'",\ "object": "'${COMMIT_SHA}'",\
"message": "Tagging Release '${EXT_RELEASE_CLEAN}'-ls'${LS_TAG_NUMBER}' to master",\ "message": "Tagging Release '${EXT_RELEASE_CLEAN}'-ls'${LS_TAG_NUMBER}' to release-4.2",\
"type": "commit",\ "type": "commit",\
"tagger": {"name": "LinuxServer Jenkins","email": "jenkins@linuxserver.io","date": "'${GITHUB_DATE}'"}}' ''' "tagger": {"name": "LinuxServer Jenkins","email": "jenkins@linuxserver.io","date": "'${GITHUB_DATE}'"}}' '''
echo "Pushing New release for Tag" echo "Pushing New release for Tag"
sh '''#! /bin/bash sh '''#! /bin/bash
curl -s https://api.github.com/repos/${EXT_USER}/${EXT_REPO}/commits/${EXT_GIT_BRANCH} | jq '. | .commit.message' | sed 's:^.\\(.*\\).$:\\1:' > releasebody.json curl -s https://api.github.com/repos/${EXT_USER}/${EXT_REPO}/commits/${EXT_GIT_BRANCH} | jq '. | .commit.message' | sed 's:^.\\(.*\\).$:\\1:' > releasebody.json
echo '{"tag_name":"'${META_TAG}'",\ echo '{"tag_name":"'${EXT_RELEASE_CLEAN}'-ls'${LS_TAG_NUMBER}'",\
"target_commitish": "master",\ "target_commitish": "release-4.2",\
"name": "'${META_TAG}'",\ "name": "'${EXT_RELEASE_CLEAN}'-ls'${LS_TAG_NUMBER}'",\
"body": "**LinuxServer Changes:**\\n\\n'${LS_RELEASE_NOTES}'\\n**'${EXT_REPO}' Changes:**\\n\\n' > start "body": "**LinuxServer Changes:**\\n\\n'${LS_RELEASE_NOTES}'\\n**'${EXT_REPO}' Changes:**\\n\\n' > start
printf '","draft": false,"prerelease": false}' >> releasebody.json printf '","draft": false,"prerelease": true}' >> releasebody.json
paste -d'\\0' start releasebody.json > releasebody.json.done paste -d'\\0' start releasebody.json > releasebody.json.done
curl -H "Authorization: token ${GITHUB_TOKEN}" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/releases -d @releasebody.json.done''' curl -H "Authorization: token ${GITHUB_TOKEN}" -X POST https://api.github.com/repos/${LS_USER}/${LS_REPO}/releases -d @releasebody.json.done'''
} }
@@ -772,20 +631,14 @@ pipeline {
] ]
]) { ]) {
sh '''#! /bin/bash sh '''#! /bin/bash
set -e docker pull lsiodev/readme-sync
TEMPDIR=$(mktemp -d)
docker pull ghcr.io/linuxserver/jenkins-builder:latest
docker run --rm -e CONTAINER_NAME=${CONTAINER_NAME} -e GITHUB_BRANCH="${BRANCH_NAME}" -v ${TEMPDIR}:/ansible/jenkins ghcr.io/linuxserver/jenkins-builder:latest
docker pull ghcr.io/linuxserver/readme-sync
docker run --rm=true \ docker run --rm=true \
-e DOCKERHUB_USERNAME=$DOCKERUSER \ -e DOCKERHUB_USERNAME=$DOCKERUSER \
-e DOCKERHUB_PASSWORD=$DOCKERPASS \ -e DOCKERHUB_PASSWORD=$DOCKERPASS \
-e GIT_REPOSITORY=${LS_USER}/${LS_REPO} \ -e GIT_REPOSITORY=${LS_USER}/${LS_REPO} \
-e DOCKER_REPOSITORY=${IMAGE} \ -e DOCKER_REPOSITORY=${IMAGE} \
-e GIT_BRANCH=master \ -e GIT_BRANCH=master \
-v ${TEMPDIR}/docker-${CONTAINER_NAME}:/mnt \ lsiodev/readme-sync bash -c 'node sync' '''
ghcr.io/linuxserver/readme-sync bash -c 'node sync'
rm -Rf ${TEMPDIR} '''
} }
} }
} }
@@ -812,19 +665,16 @@ pipeline {
sh 'echo "build aborted"' sh 'echo "build aborted"'
} }
else if (currentBuild.currentResult == "SUCCESS"){ else if (currentBuild.currentResult == "SUCCESS"){
sh ''' curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://wiki.jenkins-ci.org/download/attachments/2916393/headshot.png","embeds": [{"color": 1681177,\ sh ''' curl -X POST --data '{"avatar_url": "https://wiki.jenkins-ci.org/download/attachments/2916393/headshot.png","embeds": [{"color": 1681177,\
"description": "**Build:** '${BUILD_NUMBER}'\\n**CI Results:** '${CI_URL}'\\n**ShellCheck Results:** '${SHELLCHECK_URL}'\\n**Status:** Success\\n**Job:** '${RUN_DISPLAY_URL}'\\n**Change:** '${CODE_URL}'\\n**External Release:**: '${RELEASE_LINK}'\\n**DockerHub:** '${DOCKERHUB_LINK}'\\n"}],\ "description": "**Build:** '${BUILD_NUMBER}'\\n**CI Results:** '${CI_URL}'\\n**ShellCheck Results:** '${SHELLCHECK_URL}'\\n**Status:** Success\\n**Job:** '${RUN_DISPLAY_URL}'\\n**Change:** '${CODE_URL}'\\n**External Release:**: '${RELEASE_LINK}'\\n**DockerHub:** '${DOCKERHUB_LINK}'\\n"}],\
"username": "Jenkins"}' ${BUILDS_DISCORD} ''' "username": "Jenkins"}' ${BUILDS_DISCORD} '''
} }
else { else {
sh ''' curl -X POST -H "Content-Type: application/json" --data '{"avatar_url": "https://wiki.jenkins-ci.org/download/attachments/2916393/headshot.png","embeds": [{"color": 16711680,\ sh ''' curl -X POST --data '{"avatar_url": "https://wiki.jenkins-ci.org/download/attachments/2916393/headshot.png","embeds": [{"color": 16711680,\
"description": "**Build:** '${BUILD_NUMBER}'\\n**CI Results:** '${CI_URL}'\\n**ShellCheck Results:** '${SHELLCHECK_URL}'\\n**Status:** failure\\n**Job:** '${RUN_DISPLAY_URL}'\\n**Change:** '${CODE_URL}'\\n**External Release:**: '${RELEASE_LINK}'\\n**DockerHub:** '${DOCKERHUB_LINK}'\\n"}],\ "description": "**Build:** '${BUILD_NUMBER}'\\n**CI Results:** '${CI_URL}'\\n**ShellCheck Results:** '${SHELLCHECK_URL}'\\n**Status:** failure\\n**Job:** '${RUN_DISPLAY_URL}'\\n**Change:** '${CODE_URL}'\\n**External Release:**: '${RELEASE_LINK}'\\n**DockerHub:** '${DOCKERHUB_LINK}'\\n"}],\
"username": "Jenkins"}' ${BUILDS_DISCORD} ''' "username": "Jenkins"}' ${BUILDS_DISCORD} '''
} }
} }
} }
cleanup {
cleanWs()
}
} }
} }

168
README.md
View File

@@ -1,16 +1,6 @@
<!-- DO NOT EDIT THIS FILE MANUALLY -->
<!-- Please read the CONTRIBUTING.md -->
[![linuxserver.io](https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/linuxserver_medium.png)](https://linuxserver.io) [![linuxserver.io](https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/linuxserver_medium.png)](https://linuxserver.io)
[![Blog](https://img.shields.io/static/v1.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=linuxserver.io&message=Blog)](https://blog.linuxserver.io "all the things you can do with our containers including How-To guides, opinions and much more!") The [LinuxServer.io](https://linuxserver.io) team brings you another container release featuring :-
[![Discord](https://img.shields.io/discord/354974912613449730.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=Discord&logo=discord)](https://discord.gg/YWrKVTn "realtime support / chat with the community and the team.")
[![Discourse](https://img.shields.io/discourse/https/discourse.linuxserver.io/topics.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&logo=discourse)](https://discourse.linuxserver.io "post on our community forum.")
[![Fleet](https://img.shields.io/static/v1.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=linuxserver.io&message=Fleet)](https://fleet.linuxserver.io "an online web interface which displays all of our maintained images.")
[![GitHub](https://img.shields.io/static/v1.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=linuxserver.io&message=GitHub&logo=github)](https://github.com/linuxserver "view the source for all of our repositories.")
[![Open Collective](https://img.shields.io/opencollective/all/linuxserver.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=Supporters&logo=open%20collective)](https://opencollective.com/linuxserver "please consider helping us by either donating or contributing to our budget")
The [LinuxServer.io](https://linuxserver.io) team brings you another container release featuring:
* regular and timely application updates * regular and timely application updates
* easy user mappings (PGID, PUID) * easy user mappings (PGID, PUID)
@@ -19,24 +9,18 @@ The [LinuxServer.io](https://linuxserver.io) team brings you another container r
* regular security updates * regular security updates
Find us at: Find us at:
* [Blog](https://blog.linuxserver.io) - all the things you can do with our containers including How-To guides, opinions and much more!
* [Discord](https://discord.gg/YWrKVTn) - realtime support / chat with the community and the team. * [Discord](https://discord.gg/YWrKVTn) - realtime support / chat with the community and the team.
* [Discourse](https://discourse.linuxserver.io) - post on our community forum. * [IRC](https://irc.linuxserver.io) - on freenode at `#linuxserver.io`. Our primary support channel is Discord.
* [Fleet](https://fleet.linuxserver.io) - an online web interface which displays all of our maintained images. * [Blog](https://blog.linuxserver.io) - all the things you can do with our containers including How-To guides, opinions and much more!
* [GitHub](https://github.com/linuxserver) - view the source for all of our repositories.
* [Open Collective](https://opencollective.com/linuxserver) - please consider helping us by either donating or contributing to our budget
# [linuxserver/tvheadend](https://github.com/linuxserver/docker-tvheadend) # [linuxserver/tvheadend](https://github.com/linuxserver/docker-tvheadend)
[![](https://img.shields.io/discord/354974912613449730.svg?logo=discord&label=LSIO%20Discord&style=flat-square)](https://discord.gg/YWrKVTn)
[![GitHub Stars](https://img.shields.io/github/stars/linuxserver/docker-tvheadend.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&logo=github)](https://github.com/linuxserver/docker-tvheadend) [![](https://images.microbadger.com/badges/version/linuxserver/tvheadend.svg)](https://microbadger.com/images/linuxserver/tvheadend "Get your own version badge on microbadger.com")
[![GitHub Release](https://img.shields.io/github/release/linuxserver/docker-tvheadend.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&logo=github)](https://github.com/linuxserver/docker-tvheadend/releases) [![](https://images.microbadger.com/badges/image/linuxserver/tvheadend.svg)](https://microbadger.com/images/linuxserver/tvheadend "Get your own version badge on microbadger.com")
[![GitHub Package Repository](https://img.shields.io/static/v1.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=linuxserver.io&message=GitHub%20Package&logo=github)](https://github.com/linuxserver/docker-tvheadend/packages) ![Docker Pulls](https://img.shields.io/docker/pulls/linuxserver/tvheadend.svg)
[![GitLab Container Registry](https://img.shields.io/static/v1.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=linuxserver.io&message=GitLab%20Registry&logo=gitlab)](https://gitlab.com/linuxserver.io/docker-tvheadend/container_registry) ![Docker Stars](https://img.shields.io/docker/stars/linuxserver/tvheadend.svg)
[![MicroBadger Layers](https://img.shields.io/microbadger/layers/linuxserver/tvheadend.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge)](https://microbadger.com/images/linuxserver/tvheadend "Get your own version badge on microbadger.com") [![Build Status](https://ci.linuxserver.io/buildStatus/icon?job=Docker-Pipeline-Builders/docker-tvheadend/master)](https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-tvheadend/job/master/)
[![Docker Pulls](https://img.shields.io/docker/pulls/linuxserver/tvheadend.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=pulls&logo=docker)](https://hub.docker.com/r/linuxserver/tvheadend) [![](https://lsio-ci.ams3.digitaloceanspaces.com/linuxserver/tvheadend/latest/badge.svg)](https://lsio-ci.ams3.digitaloceanspaces.com/linuxserver/tvheadend/latest/index.html)
[![Docker Stars](https://img.shields.io/docker/stars/linuxserver/tvheadend.svg?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=stars&logo=docker)](https://hub.docker.com/r/linuxserver/tvheadend)
[![Jenkins Build](https://img.shields.io/jenkins/build?labelColor=555555&logoColor=ffffff&style=for-the-badge&jobUrl=https%3A%2F%2Fci.linuxserver.io%2Fjob%2FDocker-Pipeline-Builders%2Fjob%2Fdocker-tvheadend%2Fjob%2Fmaster%2F&logo=jenkins)](https://ci.linuxserver.io/job/Docker-Pipeline-Builders/job/docker-tvheadend/job/master/)
[![LSIO CI](https://img.shields.io/badge/dynamic/yaml?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=CI&query=CI&url=https%3A%2F%2Fci-tests.linuxserver.io%2Flinuxserver%2Ftvheadend%2Flatest%2Fci-status.yml)](https://ci-tests.linuxserver.io/linuxserver/tvheadend/latest/index.html)
[Tvheadend](https://www.tvheadend.org/) works as a proxy server: is a TV streaming server and recorder for Linux, FreeBSD and Android supporting DVB-S, DVB-S2, DVB-C, DVB-T, ATSC, ISDB-T, IPTV, SAT>IP and HDHomeRun as input sources. [Tvheadend](https://www.tvheadend.org/) works as a proxy server: is a TV streaming server and recorder for Linux, FreeBSD and Android supporting DVB-S, DVB-S2, DVB-C, DVB-T, ATSC, ISDB-T, IPTV, SAT>IP and HDHomeRun as input sources.
Tvheadend offers the HTTP (VLC, MPlayer), HTSP (Kodi, Movian) and SAT>IP streaming. Tvheadend offers the HTTP (VLC, MPlayer), HTSP (Kodi, Movian) and SAT>IP streaming.
@@ -47,9 +31,9 @@ Multiple EPG sources are supported (over-the-air DVB and ATSC including OpenTV D
## Supported Architectures ## Supported Architectures
Our images support multiple architectures such as `x86-64`, `arm64` and `armhf`. We utilise the docker manifest for multi-platform awareness. More information is available from docker [here](https://github.com/docker/distribution/blob/master/docs/spec/manifest-v2-2.md#manifest-list) and our announcement [here](https://blog.linuxserver.io/2019/02/21/the-lsio-pipeline-project/). Our images support multiple architectures such as `x86-64`, `arm64` and `armhf`. We utilise the docker manifest for multi-platform awareness. More information is available from docker [here](https://github.com/docker/distribution/blob/master/docs/spec/manifest-v2-2.md#manifest-list) and our announcement [here](https://blog.linuxserver.io/2019/02/21/the-lsio-pipeline-project/).
Simply pulling `ghcr.io/linuxserver/tvheadend` should retrieve the correct image for your arch, but you can also pull specific arch images via tags. Simply pulling `linuxserver/tvheadend` should retrieve the correct image for your arch, but you can also pull specific arch images via tags.
The architectures supported by this image are: The architectures supported by this image are:
@@ -74,16 +58,40 @@ This image provides various versions that are available via tags. `latest` tag u
Here are some example snippets to help you get started creating a container. Here are some example snippets to help you get started creating a container.
### docker-compose ([recommended](https://docs.linuxserver.io/general/docker-compose)) ### docker
```
docker create \
--name=tvheadend \
-e PUID=1000 \
-e PGID=1000 \
-e TZ=Europe/London \
-e RUN_OPTS=<run options here> `#optional` \
-p 9981:9981 \
-p 9982:9982 \
-v <path to data>:/config \
-v <path to recordings>:/recordings \
--device /dev/dri:/dev/dri `#optional` \
--device /dev/drb:/dev/dvb `#optional` \
--restart unless-stopped \
linuxserver/tvheadend
```
#### Host vs. Bridge
If you use IPTV, SAT>IP or HDHomeRun, you need to create the container with --net=host and remove the -p flags. This is because of a limitation in docker and multicast.
### docker-compose
Compatible with docker-compose v2 schemas. Compatible with docker-compose v2 schemas.
```yaml ```
--- ---
version: "2.1" version: "2"
services: services:
tvheadend: tvheadend:
image: ghcr.io/linuxserver/tvheadend image: linuxserver/tvheadend
container_name: tvheadend container_name: tvheadend
environment: environment:
- PUID=1000 - PUID=1000
@@ -98,35 +106,10 @@ services:
- 9982:9982 - 9982:9982
devices: devices:
- /dev/dri:/dev/dri #optional - /dev/dri:/dev/dri #optional
- /dev/dvb:/dev/dvb #optional - /dev/drb:/dev/dvb #optional
restart: unless-stopped restart: unless-stopped
``` ```
### docker cli
```
docker run -d \
--name=tvheadend \
-e PUID=1000 \
-e PGID=1000 \
-e TZ=Europe/London \
-e RUN_OPTS=<run options here> `#optional` \
-p 9981:9981 \
-p 9982:9982 \
-v <path to data>:/config \
-v <path to recordings>:/recordings \
--device /dev/dri:/dev/dri `#optional` \
--device /dev/dvb:/dev/dvb `#optional` \
--restart unless-stopped \
ghcr.io/linuxserver/tvheadend
```
#### Host vs. Bridge
If you use IPTV, SAT>IP or HDHomeRun, you need to create the container with --net=host and remove the -p flags. This is because to work with these services Tvheadend requires a multicast address of `239.255.255.250` and a UDP port of `1900` which at this time is not possible with docker bridge mode.
If you have other host services which also use multicast such as SSDP/DLNA/Emby you may experience stabilty problems. These can be solved by giving tvheadend its own IP using macavlan.
## Parameters ## Parameters
Container images are configured using parameters passed at runtime (such as those above). These parameters are separated by a colon and indicate `<external>:<internal>` respectively. For example, `-p 8080:80` would expose port `80` from inside the container to be accessible from the host's IP on port `8080` outside the container. Container images are configured using parameters passed at runtime (such as those above). These parameters are separated by a colon and indicate `<external>:<internal>` respectively. For example, `-p 8080:80` would expose port `80` from inside the container to be accessible from the host's IP on port `8080` outside the container.
@@ -144,23 +127,6 @@ Container images are configured using parameters passed at runtime (such as thos
| `--device /dev/dri` | Only needed if you want to use your AMD/Intel GPU for hardware accelerated video encoding (vaapi). | | `--device /dev/dri` | Only needed if you want to use your AMD/Intel GPU for hardware accelerated video encoding (vaapi). |
| `--device /dev/dvb` | Only needed if you want to pass through a DVB card to the container. If you use IPTV or HDHomeRun you can leave it out. | | `--device /dev/dvb` | Only needed if you want to pass through a DVB card to the container. If you use IPTV or HDHomeRun you can leave it out. |
## Environment variables from files (Docker secrets)
You can set any environment variable from a file by using a special prepend `FILE__`.
As an example:
```
-e FILE__PASSWORD=/run/secrets/mysecretpassword
```
Will set the environment variable `PASSWORD` based on the contents of the `/run/secrets/mysecretpassword` file.
## Umask for running applications
For all of our images we provide the ability to override the default umask settings for services started within the containers using the optional `-e UMASK=022` setting.
Keep in mind umask is not chmod it subtracts from permissions based on it's value it does not add. Please read up [here](https://en.wikipedia.org/wiki/Umask) before asking for support.
## User / Group Identifiers ## User / Group Identifiers
When using volumes (`-v` flags) permissions issues can arise between the host OS and the container, we avoid this issue by allowing you to specify the user `PUID` and group `PGID`. When using volumes (`-v` flags) permissions issues can arise between the host OS and the container, we avoid this issue by allowing you to specify the user `PUID` and group `PGID`.
@@ -229,7 +195,7 @@ For advanced setup of tvheadend, go to [Tvheadend][appurl]
**Picons** **Picons**
We have added all the picons from [picons.eu](https://picons.eu/) in the folder /picons. To enable the use of these picons, add the path to the Channel icon path in Configuration --> General --> Base. We have added all the picons from [picons.xyz](https://picons.xyz/) in the folder /picons. To enable the use of these picons, add the path to the Channel icon path in Configuration --> General --> Base.
You need to enable minimum advanced view level to see the picons options. You need to enable minimum advanced view level to see the picons options.
## Additional runtime parameters ## Additional runtime parameters
@@ -237,26 +203,29 @@ You need to enable minimum advanced view level to see the picons options.
In some cases it might be necessary to start tvheadend with additional parameters, for example to enable debugging or specify webroot for reverse proxy. Be sure to have the right parameters set, as adding the wrong once might lead to the container not starting correctly. In some cases it might be necessary to start tvheadend with additional parameters, for example to enable debugging or specify webroot for reverse proxy. Be sure to have the right parameters set, as adding the wrong once might lead to the container not starting correctly.
## Docker Mods
[![Docker Mods](https://img.shields.io/badge/dynamic/yaml?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=tvheadend&query=%24.mods%5B%27tvheadend%27%5D.mod_count&url=https%3A%2F%2Fraw.githubusercontent.com%2Flinuxserver%2Fdocker-mods%2Fmaster%2Fmod-list.yml)](https://mods.linuxserver.io/?mod=tvheadend "view available mods for this container.") [![Docker Universal Mods](https://img.shields.io/badge/dynamic/yaml?color=94398d&labelColor=555555&logoColor=ffffff&style=for-the-badge&label=universal&query=%24.mods%5B%27universal%27%5D.mod_count&url=https%3A%2F%2Fraw.githubusercontent.com%2Flinuxserver%2Fdocker-mods%2Fmaster%2Fmod-list.yml)](https://mods.linuxserver.io/?mod=universal "view available universal mods.")
We publish various [Docker Mods](https://github.com/linuxserver/docker-mods) to enable additional functionality within the containers. The list of Mods available for this image (if any) as well as universal mods that can be applied to any one of our images can be accessed via the dynamic badges above.
## Support Info ## Support Info
* Shell access whilst the container is running: `docker exec -it tvheadend /bin/bash` * Shell access whilst the container is running: `docker exec -it tvheadend /bin/bash`
* To monitor the logs of the container in realtime: `docker logs -f tvheadend` * To monitor the logs of the container in realtime: `docker logs -f tvheadend`
* container version number * container version number
* `docker inspect -f '{{ index .Config.Labels "build_version" }}' tvheadend` * `docker inspect -f '{{ index .Config.Labels "build_version" }}' tvheadend`
* image version number * image version number
* `docker inspect -f '{{ index .Config.Labels "build_version" }}' ghcr.io/linuxserver/tvheadend` * `docker inspect -f '{{ index .Config.Labels "build_version" }}' linuxserver/tvheadend`
## Updating Info ## Updating Info
Most of our images are static, versioned, and require an image update and container recreation to update the app inside. With some exceptions (ie. nextcloud, plex), we do not recommend or support updating apps inside the container. Please consult the [Application Setup](#application-setup) section above to see if it is recommended for the image. Most of our images are static, versioned, and require an image update and container recreation to update the app inside. With some exceptions (ie. nextcloud, plex), we do not recommend or support updating apps inside the container. Please consult the [Application Setup](#application-setup) section above to see if it is recommended for the image.
Below are the instructions for updating containers: Below are the instructions for updating containers:
### Via Docker Run/Create
* Update the image: `docker pull linuxserver/tvheadend`
* Stop the running container: `docker stop tvheadend`
* Delete the container: `docker rm tvheadend`
* Recreate a new container with the same docker create parameters as instructed above (if mapped correctly to a host folder, your `/config` folder and settings will be preserved)
* Start the new container: `docker start tvheadend`
* You can also remove the old dangling images: `docker image prune`
### Via Docker Compose ### Via Docker Compose
* Update all images: `docker-compose pull` * Update all images: `docker-compose pull`
@@ -265,14 +234,7 @@ Below are the instructions for updating containers:
* or update a single container: `docker-compose up -d tvheadend` * or update a single container: `docker-compose up -d tvheadend`
* You can also remove the old dangling images: `docker image prune` * You can also remove the old dangling images: `docker image prune`
### Via Docker Run ### Via Watchtower auto-updater (especially useful if you don't remember the original parameters)
* Update the image: `docker pull ghcr.io/linuxserver/tvheadend`
* Stop the running container: `docker stop tvheadend`
* Delete the container: `docker rm tvheadend`
* Recreate a new container with the same docker run parameters as instructed above (if mapped correctly to a host folder, your `/config` folder and settings will be preserved)
* You can also remove the old dangling images: `docker image prune`
### Via Watchtower auto-updater (only use if you don't remember the original parameters)
* Pull the latest image at its tag and replace it with the same env variables in one run: * Pull the latest image at its tag and replace it with the same env variables in one run:
``` ```
docker run --rm \ docker run --rm \
@@ -282,21 +244,16 @@ Below are the instructions for updating containers:
``` ```
* You can also remove the old dangling images: `docker image prune` * You can also remove the old dangling images: `docker image prune`
**Note:** We do not endorse the use of Watchtower as a solution to automated updates of existing Docker containers. In fact we generally discourage automated updates. However, this is a useful tool for one-time manual updates of containers where you have forgotten the original parameters. In the long term, we highly recommend using [Docker Compose](https://docs.linuxserver.io/general/docker-compose).
### Image Update Notifications - Diun (Docker Image Update Notifier)
* We recommend [Diun](https://crazymax.dev/diun/) for update notifications. Other tools that automatically update containers unattended are not recommended or supported.
## Building locally ## Building locally
If you want to make local modifications to these images for development purposes or just to customize the logic: If you want to make local modifications to these images for development purposes or just to customize the logic:
``` ```
git clone https://github.com/linuxserver/docker-tvheadend.git git clone https://github.com/linuxserver/docker-tvheadend.git
cd docker-tvheadend cd docker-tvheadend
docker build \ docker build \
--no-cache \ --no-cache \
--pull \ --pull \
-t ghcr.io/linuxserver/tvheadend:latest . -t linuxserver/tvheadend:latest .
``` ```
The ARM variants can be built on x86_64 hardware using `multiarch/qemu-user-static` The ARM variants can be built on x86_64 hardware using `multiarch/qemu-user-static`
@@ -308,17 +265,8 @@ Once registered you can define the dockerfile to use with `-f Dockerfile.aarch64
## Versions ## Versions
* **02.06.20:** - Update to Alpine 3.12.
* **27.12.19:** - Add requests and perl-json-xs package.
* **27.12.19:** - Update to Alpine 3.11.
* **02.10.19:** - Improve permission fixing on render & dvb devices.
* **18.08.19:** - Add AMD drivers.
* **02.08.19:** - Attempt to automatically fix permissions on /dev/dri and /dev/dvb.
* **28.06.19:** - Rebasing to alpine 3.10.
* **27.03.19:** - Rebase to Alpine 3.9, fix init logic to only chown once. * **27.03.19:** - Rebase to Alpine 3.9, fix init logic to only chown once.
* **23.03.19:** - Switching to new Base images, shift to arm32v7 tag. * **23.03.19:** - Switching to new Base images, shift to arm32v7 tag.
* **01.03.19:** - Bump xmltv to 0.6.1.
* **28.02.19:** - add perl-lwp-useragent-determined.
* **17.02.19:** - Bump xmltv to 5.70, ensure version tagging works by cloning tvheadend. * **17.02.19:** - Bump xmltv to 5.70, ensure version tagging works by cloning tvheadend.
* **14.02.19:** - Add picons path to config. * **14.02.19:** - Add picons path to config.
* **15.01.19:** - Add pipeline logic and multi arch. * **15.01.19:** - Add pipeline logic and multi arch.

View File

@@ -3,11 +3,11 @@
# jenkins variables # jenkins variables
project_name: docker-tvheadend project_name: docker-tvheadend
external_type: github_commit external_type: github_commit
release_type: stable release_type: prerelease
release_tag: latest release_tag: release-4.2
ls_branch: master ls_branch: release-4.2
repo_vars: repo_vars:
- EXT_GIT_BRANCH = 'master' - EXT_GIT_BRANCH = 'release/4.2'
- EXT_USER = 'tvheadend' - EXT_USER = 'tvheadend'
- EXT_REPO = 'tvheadend' - EXT_REPO = 'tvheadend'
- BUILD_VERSION_ARG = 'TVHEADEND_COMMIT' - BUILD_VERSION_ARG = 'TVHEADEND_COMMIT'

View File

@@ -1,255 +1,218 @@
alpine-baselayout-3.2.0-r7 alpine-baselayout-3.1.0-r3
alpine-keys-2.2-r0 alpine-keys-2.1-r1
alsa-lib-1.2.2-r0 alsa-lib-1.1.8-r0
aom-libs-1.0.0-r1 apk-tools-2.10.3-r1
apk-tools-2.10.5-r1 bash-4.4.19-r1
bash-5.0.17-r0 bsd-compat-headers-0.7.1-r0
brotli-libs-1.0.9-r1 busybox-1.29.3-r10
bsd-compat-headers-0.7.2-r3 bzip2-1.0.6-r6
busybox-1.31.1-r19 ca-certificates-20190108-r0
bzip2-1.0.8-r1 ca-certificates-cacert-20190108-r0
ca-certificates-20191127-r4 coreutils-8.30-r0
ca-certificates-bundle-20191127-r4 curl-7.64.0-r2
coreutils-8.32-r0 expat-2.2.6-r0
curl-7.69.1-r3 ffmpeg-4.0.2-r0
expat-2.2.9-r1 ffmpeg-libs-4.0.2-r0
ffmpeg-4.3.1-r0 fontconfig-2.13.1-r0
ffmpeg-libs-4.3.1-r0 freetype-2.9.1-r2
fontconfig-2.13.1-r2 fribidi-1.0.5-r0
freetype-2.10.4-r0
fribidi-1.0.9-r0
fts-1.2.7-r1
gdbm-1.13-r1 gdbm-1.13-r1
gmp-6.2.0-r0 gmp-6.1.2-r1
gnu-libiconv-1.15-r2 gnu-libiconv-1.15-r2
gnutls-3.6.15-r0 gnutls-3.6.7-r0
gzip-1.10-r0 gzip-1.10-r0
lame-3.100-r0 lame-3.100-r0
libacl-2.2.53-r0 libacl-2.2.52-r5
libass-0.14.0-r0 libass-0.14.0-r0
libattr-2.4.48-r0 libattr-2.4.47-r7
libbsd-0.10.0-r0 libbsd-0.8.6-r2
libbz2-1.0.8-r1 libbz2-1.0.6-r6
libc-utils-0.7.2-r3 libc-utils-0.7.1-r0
libcrypto1.1-1.1.1j-r0 libcrypto1.1-1.1.1b-r1
libcurl-7.69.1-r3 libcurl-7.64.0-r2
libdav1d-0.7.0-r0 libdrm-2.4.96-r0
libdrm-2.4.102-r0 libffi-3.2.1-r6
libdvbcsa-1.1.0-r1 libgcc-8.3.0-r0
libelf-0.179-r0 libgcrypt-1.8.4-r0
libffi-3.3-r2 libgpg-error-1.33-r0
libgcc-9.3.0-r2 libhdhomerun-libs-20161117-r1
libgcrypt-1.8.5-r0
libgpg-error-1.37-r0
libhdhomerun-libs-20200225-r0
libidn-1.35-r0 libidn-1.35-r0
libidn2-2.3.0-r0 libjpeg-turbo-1.5.3-r4
libintl-0.20.2-r0 libogg-1.3.3-r2
libjpeg-turbo-2.0.5-r0 libpciaccess-0.14-r0
libogg-1.3.4-r0 libpng-1.6.37-r0
libpciaccess-0.16-r0 libressl-2.7.5-r0
libpng-1.6.37-r1 libressl2.7-libcrypto-2.7.5-r0
libproc-3.3.16-r0 libressl2.7-libssl-2.7.5-r0
libssh-0.9.4-r1 libressl2.7-libtls-2.7.5-r0
libssl1.1-1.1.1j-r0 libssh2-1.8.2-r0
libstdc++-9.3.0-r2 libssl1.1-1.1.1b-r1
libtasn1-4.16.0-r1 libstdc++-8.3.0-r0
libtheora-1.1.1-r14 libtasn1-4.13-r0
libtls-standalone-2.9.1-r1 libtheora-1.1.1-r13
libtls-standalone-2.7.4-r6
libunistring-0.9.10-r0 libunistring-0.9.10-r0
liburiparser-0.9.3-r1 libuuid-2.33-r0
libuuid-2.35.2-r0 libva-2.2.0-r0
libva-2.7.1-r0 libvdpau-1.1.1-r2
libva-intel-driver-2.4.0-r1
libvdpau-1.4-r0
libvorbis-1.3.6-r2 libvorbis-1.3.6-r2
libvpx-1.8.2-r0 libvpx-1.6.1-r1
libx11-1.6.12-r0 libx11-1.6.7-r0
libxau-1.0.9-r0 libxau-1.0.8-r3
libxcb-1.14-r1 libxcb-1.13-r2
libxdmcp-1.1.3-r0 libxdmcp-1.1.2-r5
libxext-1.3.4-r0 libxext-1.3.3-r3
libxfixes-5.0.3-r2 libxfixes-5.0.3-r2
libxml2-2.9.10-r5 libxml2-2.9.9-r1
libxshmfence-1.3-r1 libxslt-1.1.33-r1
libxslt-1.1.34-r0 linux-headers-4.18.13-r1
linux-headers-5.4.5-r1 linux-pam-1.3.0-r0
linux-pam-1.3.1-r4 musl-1.1.20-r4
llvm10-libs-10.0.0-r2 musl-utils-1.1.20-r4
mesa-20.0.7-r0 ncurses-libs-6.1_p20190105-r0
mesa-dri-ati-20.0.7-r0 ncurses-terminfo-6.1_p20190105-r0
mesa-dri-classic-20.0.7-r0 ncurses-terminfo-base-6.1_p20190105-r0
mesa-dri-gallium-20.0.7-r0 nettle-3.4.1-r0
mesa-glapi-20.0.7-r0 nghttp2-libs-1.35.1-r0
mesa-va-gallium-20.0.7-r0 opus-1.3-r0
mesa-vdpau-gallium-20.0.7-r0 p11-kit-0.23.14-r0
musl-1.1.24-r10 pcre2-10.32-r1
musl-utils-1.1.24-r10 perl-5.26.3-r0
ncurses-libs-6.2_p20200523-r0 perl-archive-zip-1.64-r0
ncurses-terminfo-base-6.2_p20200523-r0 perl-b-hooks-endofscope-0.24-r0
nettle-3.5.1-r1 perl-boolean-0.46-r0
nghttp2-libs-1.41.0-r0 perl-capture-tiny-0.48-r0
openssl-1.1.1j-r0 perl-cgi-4.40-r0
opus-1.3.1-r0 perl-class-data-inheritable-0.08-r0
p11-kit-0.23.22-r0 perl-class-inspector-1.32-r0
pcre2-10.35-r0 perl-class-methodmaker-2.24-r2
perl-5.30.3-r0 perl-class-singleton-1.5-r0
perl-archive-zip-1.68-r0 perl-compress-raw-bzip2-2.084-r0
perl-b-hooks-endofscope-0.24-r1 perl-compress-raw-zlib-2.084-r0
perl-boolean-0.46-r1
perl-canary-stability-2013-r0
perl-capture-tiny-0.48-r1
perl-cgi-4.47-r0
perl-class-data-inheritable-0.08-r1
perl-class-inspector-1.36-r1
perl-class-methodmaker-2.24-r6
perl-class-singleton-1.5-r1
perl-common-sense-3.75-r0
perl-compress-raw-bzip2-2.093-r1
perl-compress-raw-zlib-2.093-r1
perl-cpan-meta-check-0.014-r1 perl-cpan-meta-check-0.014-r1
perl-data-dumper-2.173-r3 perl-data-dumper-2.173-r0
perl-date-manip-6.81-r0 perl-date-manip-6.75-r1
perl-datetime-1.52-r0 perl-datetime-1.44-r0
perl-datetime-format-strptime-1.77-r0 perl-datetime-format-strptime-1.74-r0
perl-datetime-locale-1.25-r0 perl-datetime-locale-1.22-r0
perl-datetime-timezone-2.47-r0 perl-datetime-timezone-2.19-r0
perl-dbd-sqlite-1.64-r2 perl-dbd-sqlite-1.62-r0
perl-dbi-1.643-r0 perl-dbi-1.642-r0
perl-devel-stacktrace-2.04-r1 perl-devel-stacktrace-2.03-r0
perl-devel-symdump-2.18-r1 perl-devel-symdump-2.18-r0
perl-digest-sha1-2.13-r12 perl-digest-sha1-2.13-r9
perl-dist-checkconflicts-0.11-r1 perl-dist-checkconflicts-0.11-r1
perl-doc-5.30.3-r0 perl-doc-5.26.3-r0
perl-encode-locale-1.05-r2 perl-encode-locale-1.05-r1
perl-eval-closure-0.14-r1 perl-eval-closure-0.14-r0
perl-exception-class-1.44-r1 perl-exception-class-1.44-r0
perl-exporter-tiny-1.002002-r0 perl-exporter-tiny-1.002001-r0
perl-file-listing-6.04-r2 perl-file-listing-6.04-r1
perl-file-sharedir-1.116-r1 perl-file-sharedir-1.116-r0
perl-file-sharedir-install-0.13-r1 perl-file-sharedir-install-0.13-r0
perl-file-slurp-9999.30-r0 perl-file-slurp-9999.25-r0
perl-file-temp-0.2309-r1 perl-file-temp-0.2308-r0
perl-file-which-1.23-r1 perl-file-which-1.22-r0
perl-getopt-long-2.51-r1 perl-getopt-long-2.50-r0
perl-html-parser-3.72-r5 perl-html-parser-3.72-r2
perl-html-tagset-3.20-r2 perl-html-tagset-3.20-r1
perl-html-tree-5.07-r1 perl-html-tree-5.07-r0
perl-http-cookies-6.08-r0 perl-http-cookies-6.04-r0
perl-http-daemon-6.10-r0 perl-http-daemon-6.01-r1
perl-http-date-6.05-r0 perl-http-date-6.02-r1
perl-http-message-6.24-r0 perl-http-message-6.18-r0
perl-http-negotiate-6.01-r2 perl-http-negotiate-6.01-r1
perl-importer-0.025-r0 perl-importer-0.025-r0
perl-io-1.42-r0 perl-io-1.25-r4
perl-io-compress-2.093-r0 perl-io-compress-2.084-r0
perl-io-html-1.001-r2 perl-io-html-1.001-r1
perl-io-socket-ssl-2.068-r0 perl-io-socket-ssl-2.060-r0
perl-io-stringy-2.113-r0 perl-io-stringy-2.111-r1
perl-ipc-run3-0.048-r1 perl-ipc-run3-0.048-r0
perl-json-4.02-r0 perl-json-2.97000-r0
perl-json-xs-4.02-r1 perl-libwww-6.36-r0
perl-libwww-6.44-r0 perl-lingua-en-numbers-ordinate-1.04-r1
perl-lingua-en-numbers-ordinate-1.04-r2 perl-lingua-preferred-0.2.4-r1
perl-lingua-preferred-0.2.4-r2 perl-list-moreutils-0.419-r1
perl-list-moreutils-0.428-r0 perl-lwp-mediatypes-6.02-r1
perl-list-moreutils-xs-0.428-r3 perl-module-build-0.4224-r0
perl-lwp-mediatypes-6.04-r0 perl-module-implementation-0.09-r1
perl-lwp-useragent-determined-1.07-r1 perl-module-pluggable-5.2-r0
perl-module-build-0.4231-r0 perl-module-runtime-0.016-r1
perl-module-implementation-0.09-r2 perl-mro-compat-0.13-r0
perl-module-metadata-1.000037-r0 perl-namespace-autoclean-0.28-r0
perl-module-pluggable-5.2-r1 perl-namespace-clean-0.27-r0
perl-module-runtime-0.016-r3 perl-net-http-6.09-r1
perl-mro-compat-0.13-r1 perl-net-libidn-0.12-r5
perl-namespace-autoclean-0.29-r1 perl-net-ssleay-1.85-r4
perl-namespace-clean-0.27-r1 perl-package-deprecationmanager-0.17-r0
perl-net-http-6.19-r0 perl-package-stash-0.37-r0
perl-net-libidn-0.12-r8 perl-package-stash-xs-0.28-r4
perl-net-ssleay-1.88-r0 perl-params-classify-0.015-r0
perl-package-deprecationmanager-0.17-r1 perl-params-util-1.07-r5
perl-package-stash-0.38-r2 perl-params-validate-1.29-r0
perl-package-stash-xs-0.29-r3 perl-params-validationcompiler-0.27-r0
perl-params-util-1.07-r8 perl-parse-recdescent-1.967015-r1
perl-params-validate-1.29-r3 perl-path-class-0.37-r0
perl-params-validationcompiler-0.30-r0 perl-pod-coverage-0.23-r0
perl-parse-recdescent-1.967015-r2 perl-role-tiny-2.000006-r0
perl-path-class-0.37-r1 perl-scalar-list-utils-1.50-r0
perl-pod-coverage-0.23-r1 perl-scope-guard-0.21-r0
perl-role-tiny-2.001004-r0 perl-specio-0.42-r0
perl-scalar-list-utils-1.55-r0 perl-sub-exporter-progressive-0.001013-r0
perl-scope-guard-0.21-r1 perl-sub-identify-0.14-r1
perl-specio-0.46-r0 perl-sub-info-0.002-r0
perl-sub-exporter-progressive-0.001013-r1
perl-sub-identify-0.14-r4
perl-sub-info-0.002-r1
perl-sub-install-0.928-r1 perl-sub-install-0.928-r1
perl-sub-name-0.26-r1 perl-sub-name-0.21-r1
perl-sub-quote-2.006006-r0 perl-sub-quote-2.004000-r0
perl-sub-uplevel-0.2800-r1 perl-sub-uplevel-0.2800-r0
perl-term-progressbar-2.22-r1 perl-term-progressbar-2.22-r0
perl-term-readkey-2.38-r3 perl-term-readkey-2.37-r1
perl-term-table-0.015-r0 perl-term-table-0.012-r0
perl-test-exception-0.43-r1 perl-test-exception-0.43-r0
perl-test-fatal-0.014-r2 perl-test-fatal-0.014-r1
perl-test-needs-0.002006-r0 perl-test-needs-0.002005-r1
perl-test-pod-1.52-r1 perl-test-pod-1.52-r0
perl-test-requires-0.11-r0 perl-test-requires-0.10-r0
perl-test-simple-1.302175-r0 perl-test-simple-1.302141-r0
perl-test-taint-1.08-r2 perl-test-taint-1.06-r5
perl-test-warnings-0.030-r0 perl-test-warnings-0.026-r0
perl-test-without-module-0.20-r1 perl-test-without-module-0.20-r0
perl-test2-plugin-nowarnings-0.08-r1 perl-test2-plugin-nowarnings-0.06-r0
perl-test2-suite-0.000129-r0 perl-test2-suite-0.000114-r0
perl-timedate-2.33-r0 perl-timedate-2.30-r1
perl-try-tiny-0.30-r1 perl-try-tiny-0.30-r0
perl-types-serialiser-1.0-r2 perl-unicode-string-2.09-r6
perl-unicode-string-2.10-r1 perl-uri-1.74-r0
perl-uri-1.76-r0 perl-variable-magic-0.62-r0
perl-variable-magic-0.62-r3 perl-www-robotrules-6.02-r1
perl-www-robotrules-6.02-r2 perl-xml-libxml-2.0132-r1
perl-xml-libxml-2.0205-r0 perl-xml-libxslt-1.96-r0
perl-xml-libxslt-1.99-r0 perl-xml-namespacesupport-1.12-r0
perl-xml-namespacesupport-1.12-r1 perl-xml-parser-2.44-r4
perl-xml-parser-2.46-r0 perl-xml-sax-0.99-r2
perl-xml-sax-1.02-r0 perl-xml-sax-base-1.09-r0
perl-xml-sax-base-1.09-r1 perl-xml-treepp-0.43-r1
perl-xml-treepp-0.43-r2 perl-xml-twig-3.52-r1
perl-xml-twig-3.52-r2 perl-xml-writer-0.625-r2
perl-xml-writer-0.625-r3 perl-yaml-syck-1.31-r0
perl-yaml-syck-1.32-r0 python2-2.7.16-r1
procps-3.3.16-r0 readline-7.0.003-r1
py3-appdirs-1.4.4-r1 scanelf-1.2.3-r0
py3-certifi-2020.4.5.1-r0 sdl2-2.0.9-r0
py3-chardet-3.0.4-r4 shadow-4.5-r2
py3-idna-2.9-r0 sqlite-libs-3.28.0-r0
py3-ordered-set-4.0.1-r0 ssl_client-1.29.3-r10
py3-packaging-20.4-r0 tar-1.32-r0
py3-parsing-2.4.7-r0 tzdata-2019a-r0
py3-requests-2.23.0-r0 uriparser-0.9.1-r0
py3-setuptools-47.0.0-r0 v4l-utils-libs-1.16.2-r0
py3-six-1.15.0-r0 wget-1.20.3-r0
py3-urllib3-1.25.9-r0 x264-20180304-r1
python2-2.7.18-r0 x264-libs-20180304-r1
python3-3.8.5-r1 x265-2.8-r1
readline-8.0.4-r0 x265-libs-2.8-r1
scanelf-1.2.6-r0 xvidcore-1.3.4-r1
sdl2-2.0.12-r1 xz-5.2.4-r0
shadow-4.8.1-r0 xz-libs-5.2.4-r0
sqlite-libs-3.32.1-r0 zlib-1.2.11-r1
ssl_client-1.31.1-r19
tar-1.32-r1
tzdata-2021a-r0
uriparser-0.9.3-r1
v4l-utils-libs-1.20.0-r0
wayland-libs-client-1.18.0-r4
wget-1.20.3-r1
x264-20191119-r1
x264-libs-20191119-r1
x265-3.3-r1
x265-libs-3.3-r1
xvidcore-1.3.7-r1
xz-5.2.5-r0
xz-libs-5.2.5-r0
zlib-1.2.11-r3
zstd-libs-1.4.5-r0

View File

@@ -11,6 +11,7 @@ requires 'HTML::TableExtract' ;
requires 'HTTP::Cache::Transparent' => ">= 1.3, != 1.4" ; requires 'HTTP::Cache::Transparent' => ">= 1.3, != 1.4" ;
requires 'inc' ; requires 'inc' ;
requires 'JSON::PP' ; requires 'JSON::PP' ;
requires 'JSON::XS' ;
requires 'LWP::Simple' ; requires 'LWP::Simple' ;
requires 'LWP::UserAgent' ; requires 'LWP::UserAgent' ;
requires 'PerlIO::gzip' ; requires 'PerlIO::gzip' ;

View File

@@ -53,15 +53,14 @@ opt_param_usage_include_ports: false
opt_param_device_map: true opt_param_device_map: true
opt_param_devices: opt_param_devices:
- { device_path: "/dev/dri", device_host_path: "/dev/dri", desc: "Only needed if you want to use your AMD/Intel GPU for hardware accelerated video encoding (vaapi)." } - { device_path: "/dev/dri", device_host_path: "/dev/dri", desc: "Only needed if you want to use your AMD/Intel GPU for hardware accelerated video encoding (vaapi)." }
- { device_path: "/dev/dvb", device_host_path: "/dev/dvb", desc: "Only needed if you want to pass through a DVB card to the container. If you use IPTV or HDHomeRun you can leave it out." } - { device_path: "/dev/dvb", device_host_path: "/dev/drb", desc: "Only needed if you want to pass through a DVB card to the container. If you use IPTV or HDHomeRun you can leave it out." }
opt_cap_add_param: false opt_cap_add_param: false
optional_block_1: true optional_block_1: true
optional_block_1_items: optional_block_1_items:
- | - |
#### Host vs. Bridge #### Host vs. Bridge
If you use IPTV, SAT>IP or HDHomeRun, you need to create the container with --net=host and remove the -p flags. This is because to work with these services Tvheadend requires a multicast address of `239.255.255.250` and a UDP port of `1900` which at this time is not possible with docker bridge mode. If you use IPTV, SAT>IP or HDHomeRun, you need to create the container with --net=host and remove the -p flags. This is because of a limitation in docker and multicast.
If you have other host services which also use multicast such as SSDP/DLNA/Emby you may experience stabilty problems. These can be solved by giving tvheadend its own IP using macavlan.
# application setup block # application setup block
app_setup_block_enabled: true app_setup_block_enabled: true
@@ -117,7 +116,7 @@ app_setup_block: |
**Picons** **Picons**
We have added all the picons from [picons.eu](https://picons.eu/) in the folder /picons. To enable the use of these picons, add the path to the Channel icon path in Configuration --> General --> Base. We have added all the picons from [picons.xyz](https://picons.xyz/) in the folder /picons. To enable the use of these picons, add the path to the Channel icon path in Configuration --> General --> Base.
You need to enable minimum advanced view level to see the picons options. You need to enable minimum advanced view level to see the picons options.
## Additional runtime parameters ## Additional runtime parameters
@@ -126,17 +125,8 @@ app_setup_block: |
# changelog # changelog
changelogs: changelogs:
- { date: "02.06.20:", desc: "Update to Alpine 3.12." }
- { date: "27.12.19:", desc: "Add requests and perl-json-xs package." }
- { date: "27.12.19:", desc: "Update to Alpine 3.11." }
- { date: "02.10.19:", desc: "Improve permission fixing on render & dvb devices." }
- { date: "18.08.19:", desc: "Add AMD drivers." }
- { date: "02.08.19:", desc: "Attempt to automatically fix permissions on /dev/dri and /dev/dvb." }
- { date: "28.06.19:", desc: "Rebasing to alpine 3.10." }
- { date: "27.03.19:", desc: "Rebase to Alpine 3.9, fix init logic to only chown once." } - { date: "27.03.19:", desc: "Rebase to Alpine 3.9, fix init logic to only chown once." }
- { date: "23.03.19:", desc: "Switching to new Base images, shift to arm32v7 tag." } - { date: "23.03.19:", desc: "Switching to new Base images, shift to arm32v7 tag." }
- { date: "01.03.19:", desc: "Bump xmltv to 0.6.1." }
- { date: "28.02.19:", desc: "add perl-lwp-useragent-determined." }
- { date: "17.02.19:", desc: "Bump xmltv to 5.70, ensure version tagging works by cloning tvheadend." } - { date: "17.02.19:", desc: "Bump xmltv to 5.70, ensure version tagging works by cloning tvheadend." }
- { date: "14.02.19:", desc: "Add picons path to config." } - { date: "14.02.19:", desc: "Add picons path to config." }
- { date: "15.01.19:", desc: "Add pipeline logic and multi arch." } - { date: "15.01.19:", desc: "Add pipeline logic and multi arch." }

View File

@@ -1,7 +1,7 @@
{ {
"prefer_picon": true, "prefer_picon": true,
"chiconpath": "file:///picons/%C.png", "chiconpath": "file:///picons/%C.png",
"piconpath": "file:///picons/", "piconpath": "file:///picons/",
"piconscheme": 0, "piconscheme": 0,
"chiconscheme": 2 "chiconscheme": 2
} }

View File

@@ -1,26 +0,0 @@
#!/usr/bin/with-contenv bash
FILES=$(find /dev/dri /dev/dvb -type c -print 2>/dev/null)
for i in $FILES
do
VIDEO_GID=$(stat -c '%g' "$i")
if id -G abc | grep -qw "$VIDEO_GID"; then
touch /groupadd
else
if [ ! "${VIDEO_GID}" == '0' ]; then
VIDEO_NAME=$(getent group "${VIDEO_GID}" | awk -F: '{print $1}')
if [ -z "${VIDEO_NAME}" ]; then
VIDEO_NAME="video$(head /dev/urandom | tr -dc 'a-z0-9' | head -c8)"
groupadd "$VIDEO_NAME"
groupmod -g "$VIDEO_GID" "$VIDEO_NAME"
fi
usermod -a -G "$VIDEO_NAME" abc
touch /groupadd
fi
fi
done
if [ -n "${FILES}" ] && [ ! -f "/groupadd" ]; then
usermod -a -G root abc
fi

0
root/etc/services.d/tvheadend/run Executable file → Normal file
View File