✨ 𝘼 𝙏𝙚𝙡𝙚𝙜𝙧𝙖𝙢 𝙈𝙞𝙧𝙧𝙤𝙧/𝙇𝙚𝙚𝙘𝙝 𝘽𝙤𝙩 𝙫 2.0 𝘽𝙮 𝙎𝙥𝙖𝙧𝙠𝙭𝘾𝙡𝙤𝙪𝙙 𝙂𝙧𝙤𝙪𝙥 ✨
SparkXcloud-Gdrive-MirrorBot is a multipurpose Telegram Bot writen in Python for mirroring files on the Internet to our beloved Google Drive.
In each single file there is a major change from base code, it's almost totaly different. Here some of features and fixes that I remember.
root
or TeamDrive ID, folder ids will be listed with non-recursive method)
amd64, arm64/v8, arm/v7, s390x
UPSTREAM_REPO
BOT_TOKEN
: The Telegram Bot Token that you got from @BotFather
GDRIVE_FOLDER_ID
: This is the Folder/TeamDrive ID of the Google Drive Folder to which you want to upload all the mirrors.
OWNER_ID
: The Telegram User ID (not username) of the Owner of the bot. Int
DOWNLOAD_DIR
: The path to the local folder where the downloads should be downloaded to.
DOWNLOAD_STATUS_UPDATE_INTERVAL
: Time in seconds after which the progress/status message will be updated. Recommended 10
seconds at least. Int
AUTO_DELETE_MESSAGE_DURATION
: Interval of time (in seconds), after which the bot deletes it's message and command message which is expected to be viewed instantly. NOTE: Set to -1
to disable auto message deletion. Int
IS_TEAM_DRIVE
: Set True
if uploading to TeamDrive. Default is False
. Bool
TELEGRAM_API
: This is to authenticate your Telegram account for downloading Telegram files. You can get this from https://my.telegram.org. Int
TELEGRAM_HASH
: This is to authenticate your Telegram account for downloading Telegram files. You can get this from https://my.telegram.org.
DATABASE_URL
: Your SQL Database URL. Follow this Generate Database to generate database. Data will be saved in Database: auth and sudo users, leech settings including thumbnails for each user, rss data and incomplete tasks. NOTE: If deploying on heroku and using heroku postgresql delete this variable from config.env file. DATABASE_URL will be grabbed from heroku variables.
AUTHORIZED_CHATS
: Fill user_id and chat_id of groups/users you want to authorize. Separate them by space.
SUDO_USERS
: Fill user_id of users whom you want to give sudo permission. Separate them by space.
IGNORE_PENDING_REQUESTS
: Ignore pending requests after restart. Default is False
. Bool
USE_SERVICE_ACCOUNTS
: Whether to use Service Accounts or not. For this to work see Using Service Accounts section. Default is False
. Bool
INDEX_URL
: Refer to https://gitlab.com/ParveenBhadooOfficial/Google-Drive-Index.
STATUS_LIMIT
: Limit the no. of tasks shown in status message with buttons. NOTE: Recommended limit is 4
tasks.
STOP_DUPLICATE
: Bot will check file in Drive, if it is present in Drive, downloading or cloning will be stopped. (NOTE: File will be checked using filename not file hash, so this feature is not perfect yet). Default is False
. Bool
CMD_INDEX
: commands index number. This number will added at the end all commands.
UPTOBOX_TOKEN
: Uptobox token to mirror uptobox links. Get it from Uptobox Premium Account.
TORRENT_TIMEOUT
: Timeout of dead torrents downloading with qBittorrent and Aria2c in seconds.
EXTENTION_FILTER
: File extentions that won't upload/clone. Separate them by space.
INCOMPLETE_TASK_NOTIFIER
: Get incomplete task messages after restart. Require database and superGroup. Default is False
. Bool
UPSTREAM_REPO
: Your github repository link, if your repo is private add https://username:{githubtoken}@github.com/{username}/{reponame}
format. Get token from Github settings. So you can update your bot from filled repository on each restart. NOTE: Any change in docker or requirements you need to deploy/build again with updated repo to take effect. DON'T delete .gitignore file.
UPSTREAM_BRANCH
: Upstream branch for update. Default is master
.
USER_SESSION_STRING
: To download/upload from your telegram account. If you own premium account. To generate session string use this command python3 generate_string_session.py
after mounting repo folder for sure. NOTE: You can't use bot with private message, use it with superGroup.
LEECH_SPLIT_SIZE
: Size of split in bytes. Default is 2GB
. Default is 4GB
if your account is premium.
AS_DOCUMENT
: Default type of Telegram file upload. Default is False
mean as media. Bool
EQUAL_SPLITS
: Split files larger than LEECH_SPLIT_SIZE into equal parts size (Not working with zip cmd). Default is False
. Bool
CUSTOM_FILENAME
: Add custom word to leeched file name.
DUMP_CHAT
: Chat ID. Upload files to specific chat. str
. NOTE: Only available for superGroup/channel. Add -100
before channel/superGroup id. In short don't add bot id or your id!
BASE_URL_OF_BOT
: Valid BASE URL where the bot is deployed to use qbittorrent web selection. Format of URL should be http://myip
, where myip
is the IP/Domain(public) of your bot or if you have chosen port other than 80
so write it in this format http://myip:port
(http
and not https
). This Var is optional on VPS and required for Heroku specially to avoid app sleeping/idling. For Heroku fill https://yourappname.herokuapp.com
. Still got idling? You can use http://cron-job.org to ping your Heroku app.
SERVER_PORT
: Only For VPS even if IS_VPS
is False
, which is the BASE_URL_OF_BOT Port.
WEB_PINCODE
: If empty or False
means no more pincode required while qbit web selection. Bool
MaxConnecs
and decrease AsyncIOThreadsCount
in qbittorrent config.
RSS_DELAY
: Time in seconds for rss refresh interval. Recommended 900
second at least. Default is 900
in sec.
RSS_COMMAND
: Choose command for the desired action.
RSS_CHAT_ID
: Chat ID where rss links will be sent. If using channel then add channel id. Add -100
before channel id.
RSS_USER_SESSION_STRING
: To send rss links from your telegram account. Instead of adding bot to channel then linking the channel to group to get rss link since bot will not read command from itself or other bot. To generate session string use this command python3 generate_string_session.py
after mounting repo folder for sure. NOTE: Don't use same session string as USER_SESSION_STRING
.
DATABASE_URL
and RSS_CHAT_ID
is required, otherwise all rss commands will not work. You must use bot in group. You can add the bot to a channel and link this channel to group so messages sent by bot to channel will be forwarded to group without using RSS_USER_STRING_SESSION
.ACCOUNTS_ZIP_URL
: Only if you want to load your Service Account externally from an Index Link or by any direct download link NOT webpage link. Archive the accounts folder to ZIP file. Fill this with the direct download link of zip file. If index need authentication so add direct download as shown below:
https://username:[email protected]/...
TOKEN_PICKLE_URL
: Only if you want to load your token.pickle externally from an Index Link. Fill this with the direct link of that file.
MULTI_SEARCH_URL
: Check drive_folder
setup. Write drive_folder file here. Open the raw file of that gist, it's URL will be your required variable. Should be in this form after removing commit id: https://gist.githubusercontent.com/username/gist-id/raw/drive_folder
YT_COOKIES_URL
: Youtube authentication cookies. Check setup Here. Use gist raw link and remove commit id from the link, so you can edit it from gists only.
NETRC_URL
: To create .netrc file contains authentication for aria2c and yt-dlp. Use gist raw link and remove commit id from the link, so you can edit it from gists only. NOTE: After editing .nterc you need to restart the docker or if deployed on heroku so restart dyno in case your edits related to aria2c authentication.
MEGA_API_KEY
: Mega.nz API key to mirror mega.nz links. Get it from Mega SDK Page
MEGA_EMAIL_ID
: E-Mail ID used to sign up on mega.nz for using premium account.
MEGA_PASSWORD
: Password for mega.nz account.
SHORTENER_API
: Fill your Shortener API key.
SHORTENER
: Shortener URL.
exe.io, gplinks.in, shrinkme.io, urlshortx.com, shortzon.com, bit.ly, shorte.st, linkvertise.com , ouo.io, adfoc.us, cutt.ly
CRYPT
: Cookie for gdtot google drive link generator. Paste below code in browser or make below code as bookmark in browser to run code easily.
Login/Register to gdtot.
Copy this script and paste it in browser address bar.
javascript:
exists or not, if not so write it as shown below.
javascript:(function () {
const input = document.createElement('input');
COOKIE = JSON.parse(JSON.stringify({cookie : document.cookie}));
input.value = COOKIE['cookie'].split('crypt=')[1];
document.body.appendChild(input);
input.focus();
input.select();
var result = document.execCommand('copy');
document.body.removeChild(input);
if(result)
alert('Crypt copied to clipboard');
else
prompt('Failed to copy Crypt. Manually copy below Crypt\n\n', input.value);
})();
NGxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxWdSVT0%3D
TORRENT_DIRECT_LIMIT
: To limit the Torrent/Direct mirror size. Don't add unit. Default unit is GB
.
ZIP_UNZIP_LIMIT
: To limit the size of zip and unzip commands. Don't add unit. Default unit is GB
.
CLONE_LIMIT
: To limit the size of Google Drive folder/file which you can clone. Don't add unit. Default unit is GB
.
MEGA_LIMIT
: To limit the size of Mega download. Don't add unit. Default unit is GB
.
STORAGE_THRESHOLD
: To leave specific storage free and any download will lead to leave free storage less than this value will be cancelled. Don't add unit. Default unit is GB
.
VIEW_LINK
: View Link button to open file Index Link in browser instead of direct download link, you can figure out if it's compatible with your Index code or not, open any video from you Index and check if its URL ends with ?a=view
. Compatible with BhadooIndex Code. Default is False
. Bool
Three buttons are already added including Drive Link, Index Link, and View Link, you can add extra buttons, if you don't know what are the below entries, simply leave them empty.
BUTTON_FOUR_NAME
:BUTTON_FOUR_URL
:BUTTON_FIVE_NAME
:BUTTON_FIVE_URL
:BUTTON_SIX_NAME
:BUTTON_SIX_URL
:SEARCH_API_LINK
: Search api app link. Get your api from deploying this repository.
1337x, Piratebay, Nyaasi, Torlock, Torrent Galaxy, Zooqle, Kickass, Bitsearch, MagnetDL, Libgen, YTS, Limetorrent, TorrentFunk, Glodls, TorrentProject and YourBittorrent
SEARCH_LIMIT
: Search limit for search api, limit for each site and not overall result limit. Default is zero (Default api limit for each site).
SEARCH_PLUGINS
: List of qBittorrent search plugins (github raw links). I have added some plugins, you can remove/add plugins as you want. Main Source: qBittorrent Search Plugins (Official/Unofficial).
All commands can also be called using dot(.) instead of slash(/). For e.x: .mirror <url>
or .m <url>
All commands except list
can have the bot's username appended to them. See COMMANDS_USE_BOT_NAME
under constants description. This is useful if you have multiple instances of this bot in the same group.
While creating a Telegram bot in the pre-installation section below, you might want to add the above commands to your new bot by using /setcommand
in BotFather, make sure all the commands are in lower case. This will cause a list of available bot commands to pop up in chats when you type /
, and you can long press one of them to select it instead of typing out the entire command.
nyaa.si, sukebei, 1337x, piratebay,
tgx, yts, eztv, torlock, rarbg
>mediafire, letsupload.io, hxfile.co, anonfiles.com, bayfiles.com, antfiles, fembed.com, fembed.net, femax20.com, layarkacaxxi.icu, fcdn.stream, sbplay.org, naniplay.com, naniplay.nanime.in, naniplay.nanime.biz, sbembed.com, streamtape.com, streamsb.net, feurl.com, upload.ee, pixeldrain.com, racaty.net, 1fichier.com, 1drv.ms (Only works for file not folder or business account), uptobox.com and solidfiles.com
ZIP, RAR, TAR, 7z, ISO, WIM, CAB, GZIP, BZIP2,
APM, ARJ, CHM, CPIO, CramFS, DEB, DMG, FAT,
HFS, LZH, LZMA, LZMA2, MBR, MSI, MSLZ, NSIS,
NTFS, RPM, SquashFS, UDF, VHD, XAR, Z.
Deploying is pretty much straight forward and is divided into several steps as follows:
git clone https://github.com/Spark-X-Cloud/SparkXcloud-Gdrive-MirrorBot/
cd mirrorbot
sudo apt install python3 python3-pip
Install Docker by following the official Docker docs
sudo pacman -S docker python
pip3 install -r requirements-cli.txt
d:ratio:time
perfix along with leech or mirror cmd.d
perfix alone will lead to use global options for aria2c or qbittorrent.MaxRatio
and GlobalMaxSeedingMinutes
in qbittorrent.conf, -1
means no limit, but you can cancel manually.
MaxRatioAction
.--seed-ratio
(0 means no limit) and --seed-time
(0 means no seed) in aria.sh.For Service Account to work, you must set
USE_SERVICE_ACCOUNTS
= "True" in config file or environment variables. NOTE: Using Service Accounts is only recommended while uploading to a Team Drive.
Let us create only the Service Accounts that we need.
Warning: Abuse of this feature is not the aim of this project and we do NOT recommend that you make a lot of projects, just one project and 100 SAs allow you plenty of use, its also possible that over abuse might get your projects banned by Google.
NOTE: If you have created SAs in past from this script, you can also just re download the keys by running:
python3 gen_sa_accounts.py --download-keys $PROJECTID
NOTE: 1 Service Account can upload/copy around 750 GB a day, 1 project can make 100 Service Accounts so you can upload 75 TB a day or clone 2 TB from each file creator (uploader email).
*Choose one of these methods
python3 gen_sa_accounts.py --list-projects
python3 gen_sa_accounts.py --enable-services $PROJECTID
python3 gen_sa_accounts.py --create-sas $PROJECTID
python3 gen_sa_accounts.py --download-keys $PROJECTID
python3 gen_sa_accounts.py --quick-setup 1 --new-only
A folder named accounts will be created which will contain keys for the Service Accounts.
Choose one of these methods
cd accounts
For Windows using PowerShell
$emails = Get-ChildItem .\**.json |Get-Content -Raw |ConvertFrom-Json |Select -ExpandProperty client_email >>emails.txt
For Linux
grep -oPh '"client_email": "\K[^"]+' *.json > emails.txt
cd ..
Then add emails from emails.txt to Google Group, after that add this Google Group to your Shared Drive and promote it to manager and delete email.txt file from accounts folder
python3 add_to_team_drive.py -d SharedTeamDriveSrcID
1. Using Railway
Provision PostgreSQL
PostgresSQL
Connect
column
Postgres Connection URL
and fill DATABASE_URL
variable with it
2. Using Heroku PostgreSQL
3. Using ElephantSQL
Create New Instance
Select Region
Review
Create instance
DATABASE_URL
variable with it
mirror - or /m Mirror
zipmirror - or /zm Mirror and upload as zip
unzipmirror - or /uzm Mirror and extract files
qbmirror - or /qm Mirror torrent using qBittorrent
qbzipmirror - or /qzm Mirror torrent and upload as zip using qb
qbunzipmirror - or /quzm Mirror torrent and extract files using qb
leech - or /l Leech
zipleech - or /zl Leech and upload as zip
unzipleech - or /uzl Leech and extract files
qbleech - or /ql Leech torrent using qBittorrent
qbzipleech - or /qzl Leech torrent and upload as zip using qb
qbunzipleech - or /quzl Leech torrent and extract using qb
clone - Copy file/folder to Drive
count - Count file/folder of Drive
ytdl - or /y Mirror yt-dlp supported link
ytdlzip - or /yz Mirror yt-dlp supported link as zip
ytdlleech - or /yl Leech through yt-dlp supported link
ytdlzipleech - or /yzl Leech yt-dlp support link as zip
leechset - Leech settings
setthumb - Set thumbnail
status - Get Mirror Status message
btsel - select files from torrent
rsslist - or /rl List all subscribed rss feed info
rssget - or /rg Get specific No. of links from specific rss feed
rsssub - or /rs Subscribe new rss feed
rssunsub - or /rus Unsubscribe rss feed by title
rssset - or /rst Rss Settings
list - Search files in Drive
search - Search for torrents with API
cancel - Cancel a task
cancelall - Cancel all tasks
del - Delete file/folder from Drive
log - Get the Bot Log
shell - Run commands in Shell
restart - Restart the Bot
stats - Bot Usage Stats
ping - Ping the Bot
help - All cmds with description
UPSTREAM_REPO
variable can be used for edit/add any file in repository.
UPSTREAM_REPO
private one in case you want to grab all files including private files.
UPSTREAM_REPO
and your private files in this private repository, so your private files will be overwritten from this repository. Also if you are using URL variables like TOKEN_PICKLE_URL
then all files from those variables will override the private files that added before deploying or from private UPSTREAM_REPO
.
UPSTREAM_REPO
with the official repository link, then be carefull incase any change in requirements.txt your bot will not start after restart. In this case you need to deploy again with updated code to install the new requirements or simply by changing the UPSTREAM_REPO
to you fork link with that old updates.
UPSTREAM_REPO
with your fork link be carefull also if you fetched the commits from the official repository.
UPSTREAM_REPO
will take affect only after restart.
UPSTREAM_BRANCH
don't ever fill heroku here.
d:ratio:time
perfix along with leech or mirror cmd.
d
perfix alone will lead to use global options for aria2c or qbittorrent.
MaxRatio
and GlobalMaxSeedingMinutes
in qbittorrent.conf, -1
means no limit, but you can cancel manually.
MaxRatioAction
.--seed-ratio
(0 means no limit) and --seed-time
(0 means no seed) in aria.sh.
pip install google-api-python-client google-auth-httplib2 google-auth-oauthlib
python3 generate_drive_token.py
sudo dockerd
docker build . --rm --force-rm --compress --no-cache=true --pull --file Dockerfile -t mirrorbot
sudo docker run mirrorbot
NOTE: If you want to use port other than 80, change it in docker-compose.yml also.
sudo apt install docker-compose
sudo docker-compose up
sudo docker-compose up --build
sudo docker-compose stop
sudo docker-compose start
For Service Account to work, you must set USE_SERVICE_ACCOUNTS="True" in config file or environment variables, Many thanks to AutoRClone for the scripts. NOTE: Using Service Accounts is only recommended while uploading to a Team Drive.
Let us create only the Service Accounts that we need. Warning: abuse of this feature is not the aim of this project and we do NOT recommend that you make a lot of projects, just one project and 100 SAs allow you plenty of use, its also possible that over abuse might get your projects banned by Google.
NOTE: 1 Service Account can copy around 750gb a day, 1 project can make 100 Service Accounts so that's 75tb a day, for most users this should easily suffice.
python3 gen_sa_accounts.py --quick-setup 1 --new-only
A folder named accounts will be created which will contain keys for the Service Accounts.
Or you can create Service Accounts to current project, no need to create new one
python3 gen_sa_accounts.py --list-projects
python3 gen_sa_accounts.py --enable-services $PROJECTID
python3 gen_sa_accounts.py --create-sas $PROJECTID
python3 gen_sa_accounts.py --download-keys $PROJECTID
If you want to add Service Accounts to Google Group, follow these steps
cd accounts
grep -oPh '"client_email": "\K[^"]+' *.json > emails.txt
cd -
Then add emails from emails.txt to Google Group, after that add Google Group to your Shared Drive and promote it to manager.
NOTE: If you have created SAs in past from this script, you can also just re download the keys by running:
python3 gen_sa_accounts.py --download-keys project_id
python3 add_to_team_drive.py -d SharedTeamDriveSrcID
For using your premium accounts in Youtube-dl or for protected Index Links, edit the netrc file according to following format:
machine host login username password my_youtube_password
For Index Link with only password without username, even http auth will not work, so this is the solution.
machine example.workers.dev password index_password
Where host is the name of extractor (eg. Youtube, Twitch). Multiple accounts of different hosts can be added each separated by a new line.
If you feel like showing your appreciation for this project, then how about buying me a coffee.