Mercurial > codedump
view getlist.py @ 133:0d8eabdd12ab default tip
create: write H:MM:SS timestamps, add option to fill with gaussian-blur instead of black
many albums are longer than one hour so writing H:MM:SS is a
necessity. if anything there will just be verbose info that
isn't important for my use-case.
however the gaussian-blur is simply broken. It works, and it plays
locally just fine, but YouTube in particular elongates the video
to fit the full width. I'm not entirely sure why it does this, but
it makes it useless and ugly.
| author | Paper <paper@tflc.us> |
|---|---|
| date | Sat, 03 Jan 2026 20:25:38 -0500 |
| parents | eac6dae753ca |
| children |
line wrap: on
line source
import urllib.request import json import sys from yt_dlp.utils import sanitize_filename # Initialize variables # https://github.com/xbmc/metadata.themoviedb.org.python/blob/master/python/lib/tmdbscraper/tmdbapi.py#L36 key = "f090bb54758cabf231fb605d3e3e0468" # Ask for source if not 2 >= len(sys.argv): source = sys.argv[1] if 'source' not in locals() or 'source' not in globals(): source = input( "Which website would you like to pull titles from? [tmdb, mal]: ") while source not in ["tmdb", "mal"]: print("Not a valid source! Exiting.") sys.exit() # Ask for ID if not 3 >= len(sys.argv): source = sys.argv[2] if 'id' not in locals() or 'id' not in globals(): id = input("What is the ID for your show?: ") try: temp = int(id) except Exception: print("Not a valid ID! Exiting.") sys.exit() # Scrapers if source == 'tmdb': # required because api is... odd season = input("Which season do you want?: ") try: temp = int(season) except Exception: print("Not a valid season! Exiting.") sys.exit() data = json.loads(urllib.request.urlopen(f'https://api.themoviedb.org/3/tv/{str(id)}?api_key={key}').read().decode()) amount = data["number_of_episodes"] f = open("list.txt", "w", encoding="utf-8") f.write("") f.close() count = 1 for i in range(amount): # this may count as spamming the api but i don't care lol with urllib.request.urlopen(f'https://api.themoviedb.org/3/tv/{str(id)}/season/{season}/episode/{count}?api_key={key}') as url: data = json.loads(url.read().decode()) f = open("list.txt", "a", encoding="utf-8") f.write(sanitize_filename(data["name"], restricted=True) + "\n") count += 1 f.close() if source == 'mal': with urllib.request.urlopen(f"https://api.jikan.moe/v3/anime/{str(id)}/episodes") as url: data = json.loads(url.read().decode()) count = 0 f = open("list.txt", "w", encoding="utf-8") f.write("") f.close() for i in range(len(data["episodes"])): f = open("list.txt", "a", encoding="utf-8") f.write(sanitize_filename(data["episodes"][count]["title"], restricted=True) + "\n") count += 1 f.close()
