- Joined
- Nov 14, 2012
Follow along with the video below to see how to install our site as a web app on your home screen.
Note: This feature may not be available in some browsers.
Look up Nitter and Bibliogram.@Null have you gone through the thread and found tools for archiving twitter and Instagram for the OP? They're fickle and in demand. I don't know of a way to archive Twitter and it would very useful
There is tweetsave for individual tweets, you paste the link and the tweet gets saved to several archives. They also have a firefox plugin.@Null have you gone through the thread and found tools for archiving twitter and Instagram for the OP? They're fickle and in demand. I don't know of a way to archive Twitter and it would very useful
--no-upload
. You get all videos of a user with tikup --no-upload --folder name_of_your_archive_folder tiktok_username
. I never upload because the content isn't worth the storage it would take up. Laughing about it here is enough.I already have archival tools for these (internet archive and yt-dlp). I should specify archiving whole accounts.There is tweetsave for individual tweets, you paste the link and the tweet gets saved to several archives. They also have a firefox plugin.
For tiktok there is tikup which uploads videos to the internet archive unless you turn it off with--no-upload
. You get all videos of a user withtikup --no-upload --folder name_of_your_archive_folder tiktok_username
. I never upload because the content isn't worth the storage it would take up. Laughing about it here is enough.
Just checked, it has a slightly different menu when you go to download something, but like with Youtube-dl it still needs ffmpeg to merge files together properly. It's wicked fast compared to Youtube-dl as well. If you already have ffmpeg, just paste it to your new yt-dlp folder.Youtube-dl has not updated in 4 months because Susan went after them hard. Use yt-dlp now.
![]()
Releases · yt-dlp/yt-dlp
A youtube-dl fork with additional features and fixes - yt-dlp/yt-dlpgithub.com
for what program? mkv is a container format that contains encoded streams and ts files are just mpeg chunksWhy are only webm and mp4 supported? What about mkv? What about ts?
I'm talking about the farms. Mkv and ts are not supported as attachments, which is some niggerkike shit due to both not being uncommon.for what program? mkv is a container format that contains encoded streams and ts files are just mpeg chunks
You can use archive.today to archive individual posts on twitter and instagram if I recall..@Null have you gone through the thread and found tools for archiving twitter and Instagram for the OP? They're fickle and in demand. I don't know of a way to archive Twitter and it would very useful
also for those who like to experiment technically, ffmpeg can be tweaked based on compression quality and subjective measures of video quality. HEVC/x265 and VBR might offer better quality if it is supported by playback. I haven't messed around with it but x264 should be compatible with everything
Try the browser extension "SingleFile"I tried to archive a private group FB page where the mod is calling me a 'disgusting homophobe' but all it saves is the banner and page rules, guessing because the page is private.
Anyone got a workaround to help out an old boomer?
Edited to add: I have posted in A&H without archiving and evaded the promised strangling so far. You would be helping me out there. I know inmates post newspaper articles from behind a paywall which may be the same situation here.
Thanks for that. Seems Safari has its own web archiving tool which I found Googling SingleFile.Try the browser extension "SingleFile"
I'm trying to archive tweets from a Twitter username but they have 98.2K Tweets. It seems like Twitter changed their coding because Twint isn't working anymore. Is there any other scraper I could use?
I already have archival tools for these (internet archive and yt-dlp). I should specify archiving whole accounts.
torsocks snscrape twitter-user username | tee -a username-fullurls.txt
./iacheck.sh username-fullurls.txt
to generate toarchive-username.txt#!/bin/bash
#
# IA Wayback Machine archive check
# Usage: ./iacheck.sh listofurls
echo ["$(date +"%Y-%m-%d %T")"] Script Started
echo ["$(date +"%Y-%m-%d %T")"] Total urls to check: "$(wc -l "$1" | awk '{ print $1 }')"
echo ["$(date +"%Y-%m-%d %T")"] Output to file: toarchive-"$1"
while read -r line; do
curl -4 --tcp-fastopen --compressed --tr-encoding -L -s http://web.archive.org/wayback/available?url="${line}" | jq -r '@text' | grep {} | jq -r '.[]' | head -n 1 | tee -a toarchive-"$1"
sleep 1
done <"$1"
./spn.sh -n -q -p '4' -c '-4 --http2-prior-knowledge --compressed --tr-encoding' -d '&skip_first_archive=1' toarchive-username.txt
--since DATE
option:torsocks snscrape --since 2022-01-25 twitter-user username > username-01-25.txt
./twb.sh username DATE
, with DATE being the date you want to view tweets since. Change $HOME/Desktop/test/
to your preferred path.twarc2 configure
and enter the bearer token from the twitter dev site.torsocks snscrape twitter-user username | tee -a username-fullurls.txt
awk -F/ '{print $6}' username-fullurls.txt > username-tweetids.txt
twarc2 hydrate username-tweetids.txt username-hydrated.jsonl
jq -c 'del(.__twarc)' username-hydrated.jsonl > "tmp" && mv -f "tmp" username-hydrated.jsonl
twarc2 csv username-hydrated.jsonl username-csv.csv
./iacheck.sh username-fullurls.txt
to generate toarchive-username.txt:#!/bin/bash
#
# IA Wayback Machine archive check
# Usage: ./iacheck.sh listofurls
echo ["$(date +"%Y-%m-%d %T")"] Script Started
echo ["$(date +"%Y-%m-%d %T")"] Total urls to check: "$(wc -l "$1" | awk '{ print $1 }')"
echo ["$(date +"%Y-%m-%d %T")"] Output to file: toarchive-"$1"
while read -r line; do
curl -4 --tcp-fastopen --compressed --tr-encoding -L -s http://web.archive.org/wayback/available?url="${line}" | jq -r '@text' | grep {} | jq -r '.[]' | head -n 1 | tee -a toarchive-"$1"
sleep 1
done <"$1"
./spn.sh -n -q -p '4' -c '-4 --http2-prior-knowledge --compressed --tr-encoding' -d '&skip_first_archive=1' toarchive-username.txt
torsocks snscrape --jsonl twitter-user username | tee -a username-raw_tweets.jsonl