mirror of
https://github.com/meeb/tubesync.git
synced 2025-06-26 06:56:36 +00:00
Merge branch 'meeb:main' into delete-source-reworked
This commit is contained in:
commit
0a35167563
27
.github/workflows/ci.yaml
vendored
27
.github/workflows/ci.yaml
vendored
@ -34,27 +34,50 @@ jobs:
|
||||
- name: Run Django tests
|
||||
run: cd tubesync && python3 manage.py test --verbosity=2
|
||||
containerise:
|
||||
if: ${{ !cancelled() }}
|
||||
needs: test
|
||||
runs-on: ubuntu-latest
|
||||
timeout-minutes: 120
|
||||
steps:
|
||||
- name: Set environment variables with GitHub CLI
|
||||
env:
|
||||
GH_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
run: |
|
||||
{
|
||||
# Fetch the latest+1 release from yt-dlp/FFmpeg-Builds
|
||||
var='FFMPEG_DATE' ;
|
||||
delim='"'"${var}"'_EOF"' ;
|
||||
jq_arg='.[1].tag_name[10:]' ;
|
||||
printf -- '%s<<%s\n' "${var}" "${delim}" ;
|
||||
gh api repos/yt-dlp/FFmpeg-Builds/releases --cache 12h --jq "${jq_arg}" ;
|
||||
printf -- '%s\n' "${delim}" ;
|
||||
unset -v delim jq_arg var ;
|
||||
} >> "${GITHUB_ENV}"
|
||||
- name: Set up QEMU
|
||||
uses: docker/setup-qemu-action@v3
|
||||
- name: Set up Docker Buildx
|
||||
id: buildx
|
||||
uses: docker/setup-buildx-action@v3
|
||||
- name: Log into GitHub Container Registry
|
||||
run: echo "${{ secrets.REGISTRY_ACCESS_TOKEN }}" | docker login https://ghcr.io -u ${{ github.actor }} --password-stdin
|
||||
env:
|
||||
DOCKER_REGISTRY: https://ghcr.io
|
||||
DOCKER_USERNAME: ${{ github.actor }}
|
||||
DOCKER_TOKEN: ${{ 'meeb' == github.repository_owner && secrets.REGISTRY_ACCESS_TOKEN || secrets.GITHUB_TOKEN }}
|
||||
run: echo "${DOCKER_TOKEN}" | docker login --password-stdin --username "${DOCKER_USERNAME}" "${DOCKER_REGISTRY}"
|
||||
- name: Lowercase github username for ghcr
|
||||
id: string
|
||||
uses: ASzc/change-string-case-action@v6
|
||||
with:
|
||||
string: ${{ github.actor }}
|
||||
- name: Build and push
|
||||
timeout-minutes: 60
|
||||
uses: docker/build-push-action@v6
|
||||
with:
|
||||
platforms: linux/amd64,linux/arm64
|
||||
push: true
|
||||
push: ${{ 'success' == needs.test.result && 'meeb' == github.repository_owner && 'true' || 'false' }}
|
||||
tags: ghcr.io/${{ steps.string.outputs.lowercase }}/${{ env.IMAGE_NAME }}:latest
|
||||
cache-from: type=registry,ref=ghcr.io/${{ steps.string.outputs.lowercase }}/${{ env.IMAGE_NAME }}:latest
|
||||
cache-to: type=inline
|
||||
build-args: |
|
||||
IMAGE_NAME=${{ env.IMAGE_NAME }}
|
||||
FFMPEG_DATE=${{ env.FFMPEG_DATE }}
|
||||
|
@ -1,8 +1,7 @@
|
||||
# syntax=docker/dockerfile:1
|
||||
# check=error=true
|
||||
|
||||
ARG FFMPEG_DATE="2025-03-04-15-43"
|
||||
ARG FFMPEG_VERSION="N-118645-gf76195ff65"
|
||||
ARG FFMPEG_VERSION="N"
|
||||
|
||||
ARG S6_VERSION="3.2.0.2"
|
||||
|
||||
|
@ -1202,7 +1202,8 @@ class Media(models.Model):
|
||||
|
||||
@property
|
||||
def thumbnail(self):
|
||||
return self.get_metadata_first_value('thumbnail', '')
|
||||
default = f'https://i.ytimg.com/vi/{self.key}/maxresdefault.jpg'
|
||||
return self.get_metadata_first_value('thumbnail', default)
|
||||
|
||||
@property
|
||||
def name(self):
|
||||
|
@ -115,27 +115,29 @@ def get_source_completed_tasks(source_id, only_errors=False):
|
||||
q['failed_at__isnull'] = False
|
||||
return CompletedTask.objects.filter(**q).order_by('-failed_at')
|
||||
|
||||
def get_tasks(task_name, id=None, /, instance=None):
|
||||
assert not (id is None and instance is None)
|
||||
arg = str(id or instance.pk)
|
||||
return Task.objects.get_task(str(task_name), args=(arg,),)
|
||||
|
||||
def get_first_task(task_name, id=None, /, *, instance=None):
|
||||
tqs = get_tasks(task_name, id, instance).order_by('run_at')
|
||||
return tqs[0] if tqs.count() else False
|
||||
|
||||
def get_media_download_task(media_id):
|
||||
try:
|
||||
return Task.objects.get_task('sync.tasks.download_media',
|
||||
args=(str(media_id),))[0]
|
||||
except IndexError:
|
||||
return False
|
||||
return get_first_task('sync.tasks.download_media', media_id)
|
||||
|
||||
def get_media_metadata_task(media_id):
|
||||
try:
|
||||
return Task.objects.get_task('sync.tasks.download_media_metadata',
|
||||
args=(str(media_id),))[0]
|
||||
except IndexError:
|
||||
return False
|
||||
return get_first_task('sync.tasks.download_media_metadata', media_id)
|
||||
|
||||
def get_media_premiere_task(media_id):
|
||||
try:
|
||||
return Task.objects.get_task('sync.tasks.wait_for_media_premiere',
|
||||
args=(str(media_id),))[0]
|
||||
except IndexError:
|
||||
return False
|
||||
return get_first_task('sync.tasks.wait_for_media_premiere', media_id)
|
||||
|
||||
def get_source_check_task(source_id):
|
||||
return get_first_task('sync.tasks.save_all_media_for_source', source_id)
|
||||
|
||||
def get_source_index_task(source_id):
|
||||
return get_first_task('sync.tasks.index_source_task', source_id)
|
||||
|
||||
def delete_task_by_source(task_name, source_id):
|
||||
now = timezone.now()
|
||||
@ -231,50 +233,61 @@ def index_source_task(source_id):
|
||||
# Got some media, update the last crawl timestamp
|
||||
source.last_crawl = timezone.now()
|
||||
source.save()
|
||||
log.info(f'Found {len(videos)} media items for source: {source}')
|
||||
num_videos = len(videos)
|
||||
log.info(f'Found {num_videos} media items for source: {source}')
|
||||
fields = lambda f, m: m.get_metadata_field(f)
|
||||
with atomic(durable=True):
|
||||
for video in videos:
|
||||
# Create or update each video as a Media object
|
||||
key = video.get(source.key_field, None)
|
||||
if not key:
|
||||
# Video has no unique key (ID), it can't be indexed
|
||||
continue
|
||||
try:
|
||||
media = Media.objects.get(key=key, source=source)
|
||||
except Media.DoesNotExist:
|
||||
media = Media(key=key)
|
||||
media.source = source
|
||||
media.duration = float(video.get(fields('duration', media), None) or 0) or None
|
||||
media.title = str(video.get(fields('title', media), ''))[:200]
|
||||
timestamp = video.get(fields('timestamp', media), None)
|
||||
published_dt = media.metadata_published(timestamp)
|
||||
if published_dt is not None:
|
||||
media.published = published_dt
|
||||
try:
|
||||
with atomic():
|
||||
media.save()
|
||||
except IntegrityError as e:
|
||||
log.error(f'Index media failed: {source} / {media} with "{e}"')
|
||||
else:
|
||||
log.debug(f'Indexed media: {source} / {media}')
|
||||
# log the new media instances
|
||||
new_media_instance = (
|
||||
media.created and
|
||||
source.last_crawl and
|
||||
media.created >= source.last_crawl
|
||||
task = get_source_index_task(source_id)
|
||||
if task:
|
||||
verbose_name = task.verbose_name
|
||||
tvn_format = '[{}' + f'/{num_videos}] {verbose_name}'
|
||||
for vn, video in enumerate(videos, start=1):
|
||||
# Create or update each video as a Media object
|
||||
key = video.get(source.key_field, None)
|
||||
if not key:
|
||||
# Video has no unique key (ID), it can't be indexed
|
||||
continue
|
||||
try:
|
||||
media = Media.objects.get(key=key, source=source)
|
||||
except Media.DoesNotExist:
|
||||
media = Media(key=key)
|
||||
media.source = source
|
||||
media.duration = float(video.get(fields('duration', media), None) or 0) or None
|
||||
media.title = str(video.get(fields('title', media), ''))[:200]
|
||||
timestamp = video.get(fields('timestamp', media), None)
|
||||
published_dt = media.metadata_published(timestamp)
|
||||
if published_dt is not None:
|
||||
media.published = published_dt
|
||||
if task:
|
||||
task.verbose_name = tvn_format.format(vn)
|
||||
with atomic():
|
||||
task.save(update_fields={'verbose_name'})
|
||||
try:
|
||||
media.save()
|
||||
except IntegrityError as e:
|
||||
log.error(f'Index media failed: {source} / {media} with "{e}"')
|
||||
else:
|
||||
log.debug(f'Indexed media: {source} / {media}')
|
||||
# log the new media instances
|
||||
new_media_instance = (
|
||||
media.created and
|
||||
source.last_crawl and
|
||||
media.created >= source.last_crawl
|
||||
)
|
||||
if new_media_instance:
|
||||
log.info(f'Indexed new media: {source} / {media}')
|
||||
log.info(f'Scheduling task to download metadata for: {media.url}')
|
||||
verbose_name = _('Downloading metadata for "{}"')
|
||||
download_media_metadata(
|
||||
str(media.pk),
|
||||
priority=20,
|
||||
verbose_name=verbose_name.format(media.pk),
|
||||
)
|
||||
if new_media_instance:
|
||||
log.info(f'Indexed new media: {source} / {media}')
|
||||
log.info(f'Scheduling task to download metadata for: {media.url}')
|
||||
verbose_name = _('Downloading metadata for "{}"')
|
||||
download_media_metadata(
|
||||
str(media.pk),
|
||||
priority=20,
|
||||
verbose_name=verbose_name.format(media.pk),
|
||||
)
|
||||
# Cleanup of media no longer available from the source
|
||||
cleanup_removed_media(source, videos)
|
||||
if task:
|
||||
task.verbose_name = verbose_name
|
||||
with atomic():
|
||||
task.save(update_fields={'verbose_name'})
|
||||
# Cleanup of media no longer available from the source
|
||||
cleanup_removed_media(source, videos)
|
||||
|
||||
|
||||
@background(schedule=0)
|
||||
@ -444,8 +457,6 @@ def download_media_thumbnail(media_id, url):
|
||||
except Media.DoesNotExist:
|
||||
# Task triggered but the media no longer exists, do nothing
|
||||
return
|
||||
if not media.has_metadata:
|
||||
raise NoMetadataException('Metadata is not yet available.')
|
||||
if media.skip:
|
||||
# Media was toggled to be skipped after the task was scheduled
|
||||
log.warn(f'Download task triggered for media: {media} (UUID: {media.pk}) but '
|
||||
@ -631,6 +642,7 @@ def save_all_media_for_source(source_id):
|
||||
|
||||
already_saved = set()
|
||||
mqs = Media.objects.filter(source=source)
|
||||
task = get_source_check_task(source_id)
|
||||
refresh_qs = mqs.filter(
|
||||
can_download=False,
|
||||
skip=False,
|
||||
@ -638,22 +650,40 @@ def save_all_media_for_source(source_id):
|
||||
downloaded=False,
|
||||
metadata__isnull=False,
|
||||
)
|
||||
for media in refresh_qs:
|
||||
if task:
|
||||
verbose_name = task.verbose_name
|
||||
tvn_format = '[{}' + f'/{refresh_qs.count()}] {verbose_name}'
|
||||
for mn, media in enumerate(refresh_qs, start=1):
|
||||
if task:
|
||||
task.verbose_name = tvn_format.format(mn)
|
||||
with atomic():
|
||||
task.save(update_fields={'verbose_name'})
|
||||
try:
|
||||
media.refresh_formats
|
||||
except YouTubeError as e:
|
||||
log.debug(f'Failed to refresh formats for: {source} / {media.key}: {e!s}')
|
||||
pass
|
||||
else:
|
||||
media.save()
|
||||
with atomic():
|
||||
media.save()
|
||||
already_saved.add(media.uuid)
|
||||
|
||||
# Trigger the post_save signal for each media item linked to this source as various
|
||||
# flags may need to be recalculated
|
||||
with atomic():
|
||||
for media in mqs:
|
||||
if task:
|
||||
tvn_format = '[{}' + f'/{mqs.count()}] {verbose_name}'
|
||||
for mn, media in enumerate(mqs, start=1):
|
||||
if task:
|
||||
task.verbose_name = tvn_format.format(mn)
|
||||
with atomic():
|
||||
task.save(update_fields={'verbose_name'})
|
||||
if media.uuid not in already_saved:
|
||||
media.save()
|
||||
with atomic():
|
||||
media.save()
|
||||
if task:
|
||||
task.verbose_name = verbose_name
|
||||
with atomic():
|
||||
task.save(update_fields={'verbose_name'})
|
||||
|
||||
|
||||
@background(schedule=60, remove_existing_tasks=True)
|
||||
|
Loading…
Reference in New Issue
Block a user