I wake up to find the next morning I could only download a 136 of his videos because it dc :(
Backing up YouTube channel
Other urls found in this thread:
youtube.com
twitter.com
try youtube dl gui
-i
this op last time i scraped defcon vids it's works great
...
You know what you can insult me for not looking at the manual all you want but I'm not gonna download anymore because I just saw the folder size 72GB yeah I know hes dead and all but im not backing up any more
136 videos of playing doom and rainbow 6 is were it stays
If i keep down loading shit at that rate i wouldn't have space for all of my neko porn
it wouldn't've helped you to look at the manual because you still wouldn't've expected this problem.
You will next time.
Yeah I probably should have done that and I've also probably should have logged in as well Think there is an option that prevents you from being kicked by a logging in i think
gui is bloat
Just because you can't use it right u faggot
Reminded me of the hell that was downloading a channel with 500+ videos. But yeah, as said, you just need to pass the -i flag which will ignore the errors and continue to download other videos.
Using it from within Python gives you more control but if you don't want to go that deep, then you could use something like this:CHANNEL='UCD6VugMZKRhSyzWEWA9W2fg'youtube-dl \ --no-progress --no-warnings \ -i --ignore-config --prefer-ffmpeg \ -R 'infinite' --fragment-retries 'infinite' \ --abort-on-unavailable-fragment --geo-bypass --no-check-certificate \ --match-filter "channel_id = '${CHANNEL}'" --playlist-reverse \ -w -o '%(upload_date)s.%(id)s/%(id)s.%(ext)s' -f 'bestvideo+bestaudio/best' \ --download-archive 'ARCHIVE' --merge-output-format 'mkv' \ --no-continue --write-info-json --write-thumbnail --all-subs \ -- "youtube.com
The result would look something like this:.├── 20190127.0FW23bamIZI│ ├── 0FW23bamIZI.info.json│ ├── 0FW23bamIZI.jpg│ └── 0FW23bamIZI.mkv├── 20190226.wXo24imR_54│ ├── wXo24imR_54.info.json│ ├── wXo24imR_54.jpg│ └── wXo24imR_54.mkv├── 20190317.URJ_qSXruW0│ ├── URJ_qSXruW0.info.json│ ├── URJ_qSXruW0.jpg│ └── URJ_qSXruW0.mkv└── ARCHIVEThe only problem this way of doing things has is that when scanning the playlists for videos that belong to this channel, youtube-dl has to first download the video page and check the metadata which is slow and a waste of traffic. On top of that, there's no internal mechanism to prevent youtube-dl from analyzing already rejected videos. What you can do, is get all the rejected videos from youtube-dl.log and keep them in the ARCHIVE file so they'll be ignored entirely when you update the archive.
They all have the same message:[youtube] 8UD50tPFCYo: Downloading webpage[youtube] 8UD50tPFCYo: Downloading video info webpage[download] Mario Tennis does not pass filter channel_id = 'UCPcIwIn5WO6_o_vXF8SXx3w', skipping ..You can also do the same thing with the videos that weren't downloaded due to copyright errors or some other shit. Just find them in the log, and write in ARCHIVE before reattempting to archive the channel.
Personally, I just kept an EXCLUDE file with all the videos I don't need and each time I needed to update a channel, before actually starting youtube-dl, I was switching to the directory and doing something like this:jq -r '"youtube \(.id)"' */*.info.json | cat - EXCLUDE > ARCHIVEThis is how I used to do it before switching to Python.
Forgot to mention, YouTube channels have an RSS feed (youtube.com
Not all of them have it activated, but most.
I feel like you could get around the duplicate video issue by simply doing a run grabbing video URLs only, feeding them into a file, and then filtering out duplicates. From there, do a second run using the sorted file as input.
If there's no duplicates (ie, you're not searching multiple playlists and sources) it will run a little longer, but it will save you the hassle of re-downloading the webpage multiple times if there is a possibility of duplicates.
/thread
HAPAS ARE SUPERIOR TO WHITES
HAPAS ARE SUPERIOR TO WHITES
HAPAS ARE SUPERIOR TO WHITES
HAPAS ARE SUPERIOR TO WHITES
This is a good thing.
DAILY REMINDER THAT THE MODS ARE COMPROMISED MOSSAD/CIA SHILLS
Kike mods are trying to get TOR banned. Don't let them get away with it!
Kike mods are trying to get TOR banned. Don't let them get away with it!
Mods are kikes. Why are they deleting my posts?
LOL, I wish I thought of doing that!
Yeah, right, and the moon is made of cheese.