This Telegram Bot, based on python-aria-mirror-bot, has undergone substantial modifications and is designed for efficiently mirroring or leeching files from the Internet to various destinations, including Google Drive, Telegram, or any rclone-supported cloud. It is built using asynchronous programming in Python.
root
or TeamDrive ID, folder ids will be listed with a non-recursive method). Based
on Sreeraj searchX-bot. (task option)ZIP, RAR, TAR, 7z, ISO, WIM, CAB, GZIP, BZIP2, APM, ARJ, CHM, CPIO, CramFS, DEB, DMG, FAT, HFS, LZH, LZMA, LZMA2,MBR, MSI, MSLZ, NSIS, NTFS, RPM, SquashFS, UDF, VHD, XAR, Z, TAR.XZ, CBZ
amd64, arm64/v8, arm/v7
UPSTREAM_REPO
mediafire (file/folders), hxfile.co (need cookies txt with name) [hxfile.txt], streamtape.com, streamsb.net, streamhub.ink, streamvid.net, doodstream.com, feurl.com, upload.ee, pixeldrain.com, racaty.net, 1fichier.com, 1drv.ms (Only works for file not folder or business account), filelions.com, streamwish.com, send.cm (file/folders), solidfiles.com, linkbox.to (file/folders), shrdsk.me ( sharedisk.io), akmfiles.com, wetransfer.com, pcloud.link, gofile.io (file/folders), easyupload.io, mdisk.me (with ytdl), tmpsend.com, qiwi.gg, berkasdrive.com, mp4upload.com, terabox.com (videos only file/folders).
git clone https://github.com/anasty17/mirror-leech-telegram-bot mirrorbot/ && cd mirrorbot
sudo apt install python3 python3-pip
Install Docker by following the official Docker docs
sudo pacman -S docker python
pip3 install -r requirements-cli.txt
cp config_sample.env config.env
_____REMOVE_THIS_LINE_____=True
Fill up rest of the fields. Meaning of each field is discussed below. NOTE: All values must be filled between
quotes, even if it's Int
, Bool
or List
.
1. Required Fields
BOT_TOKEN
: The Telegram Bot Token that you got from @BotFather. Str
OWNER_ID
: The Telegram User ID (not username) of the Owner of the bot. Int
TELEGRAM_API
: This is to authenticate your Telegram account for downloading Telegram files. You can get this
from https://my.telegram.org. Int
TELEGRAM_HASH
: This is to authenticate your Telegram account for downloading Telegram files. You can get this
from https://my.telegram.org. Str
2. Optional Fields
USER_SESSION_STRING
: To download/upload from your telegram account if user is PREMIUM
and to send rss. To generate
session string use this command python3 generate_string_session.py
after mounting repo folder for sure. Str
. *
NOTE*: You can't use bot with private message. Use it with superGroup.DATABASE_URL
: Your Mongo Database URL (Connection string). Follow
this Generate Database to
generate database. Data will be saved in Database: bot settings, users settings, rss data and incomplete tasks. NOTE: You can always edit all settings that saved in database from the official site -> (Browse collections). Str
DOWNLOAD_DIR
: The path to the vps local folder where the downloads should be downloaded to. Str
CMD_SUFFIX
: Commands index number. This number will added at the end all commands. Str
|Int
AUTHORIZED_CHATS
: Fill user_id and chat_id of groups/users you want to authorize. To auth only specific topic(s) write it in this format chat_id|thread_id
Ex:-100XXXXXXXXXXX|10 or Ex:-100XXXXXXXXXXX|10|12. Separate them by space. Int
SUDO_USERS
: Fill user_id of users whom you want to give sudo permission. Separate them by space. Int
DEFAULT_UPLOAD
: Whether rc
to upload to RCLONE_PATH
or gd
to upload to GDRIVE_ID
. Default is rc
. Read
More HERE.Str
STATUS_UPDATE_INTERVAL
: Time in seconds after which the progress/status message will be updated. Recommended 10
seconds at least. Int
STATUS_LIMIT
: Limit the no. of tasks shown in status message with buttons. Default is 10
. NOTE: Recommended
limit is 4
tasks. Int
EXTENSION_FILTER
: File extensions that won't upload/clone. Separate them by space. Str
INCOMPLETE_TASK_NOTIFIER
: Get incomplete task messages after restart. Require database and superGroup. Default
is False
. Bool
FILELION_API
: Filelion api key to mirror Filelion links. Get it
from Filelion. str
STREAMWISH_API
: Streamwish api key to mirror Streamwish links. Get it
from Streamwish. str
YT_DLP_OPTIONS
: Default yt-dlp options. Check all possible
options HERE or use
this script to convert cli arguments to api options. Format: key:value|key:
value|key:value. Add ^
before integer or float, some numbers must be numeric and some string. str
USE_SERVICE_ACCOUNTS
: Whether to use Service Accounts or not, with google-api-python-client. For this to work
see Using Service Accounts
section below. Default is False
. Bool
FFMPEG_CMDS
: list of lists of ffmpeg commands. You can set multiple ffmpeg commands for all files before upload. Don't write ffmpeg at beginning, start directly with the arguments. list
-del
to the list(s) which you want from the bot to delete the original files after command run complete!-del
will delete the original media after complete run of the cmd.NAME_SUBSTITUTE
: Add word/letter/character/sentense/pattern to remove or replace with other words with sensitive case or without. Notes:
BACKSLASH
, those are the characters: ^$.|?*+()[]{}-
3. GDrive Tools
GDRIVE_ID
: This is the Folder/TeamDrive ID of the Google Drive OR root
to which you want to upload all the mirrors
using google-api-python-client. Str
IS_TEAM_DRIVE
: Set True
if uploading to TeamDrive using google-api-python-client. Default is False
. Bool
INDEX_URL
: Refer to https://gitlab.com/ParveenBhadooOfficial/Google-Drive-Index. Str
STOP_DUPLICATE
: Bot will check file/folder name in Drive incase uploading to GDRIVE_ID
. If it's present in Drive
then downloading or cloning will be stopped. (NOTE: Item will be checked using name and not hash, so this feature
is not perfect yet). Default is False
. Bool
4. Rclone
RCLONE_PATH
: Default rclone path to which you want to upload all the files/folders using rclone. Str
RCLONE_FLAGS
: key:value|key|key|key:value . Check here all RcloneFlags. Str
RCLONE_SERVE_URL
: Valid URL where the bot is deployed to use rclone serve. Format of URL should be http://myip
,
where myip
is the IP/Domain(public) of your bot or if you have chosen port other than 80
so write it in this
format http://myip:port
(http
and not https
). Str
RCLONE_SERVE_PORT
: Which is the RCLONE_SERVE_URL Port. Default is 8080
. Int
RCLONE_SERVE_USER
: Username for rclone serve authentication. Str
RCLONE_SERVE_PASS
: Password for rclone serve authentication. Str
5. Update
UPSTREAM_REPO
: Your github repository link, if your repo is private
add https://username:{githubtoken}@github.com/{username}/{reponame}
format. Get token
from Github settings. So you can update your bot from filled repository on each
restart. Str
.
UPSTREAM_BRANCH
: Upstream branch for update. Default is master
. Str
6. Leech
LEECH_SPLIT_SIZE
: Size of split in bytes. Default is 2GB
. Default is 4GB
if your account is premium. Int
AS_DOCUMENT
: Default type of Telegram file upload. Default is False
mean as media. Bool
EQUAL_SPLITS
: Split files larger than LEECH_SPLIT_SIZE into equal parts size (Not working with zip cmd). Default
is False
. Bool
MEDIA_GROUP
: View Uploaded splitted file parts in media group. Default is False
. Bool
.USER_TRANSMISSION
: Upload/Download by user session. Only in superChat. Default is False
. Bool
MIXED_LEECH
: Upload by user and bot session with respect to file size. Only in superChat. Default is False
. Bool
LEECH_FILENAME_PREFIX
: Add custom word to leeched file name. Str
LEECH_DUMP_CHAT
: ID or USERNAME or PM(private message) to where files would be uploaded. Int
|Str
. Add -100
before channel/superGroup id.THUMBNAIL_LAYOUT
: Thumbnail layout (widthxheight, 2x2, 3x3, 2x4, 4x4, ...) of how many photo arranged for the thumbnail.Str
7. qBittorrent/Aria2c/Sabnzbd
TORRENT_TIMEOUT
: Timeout of dead torrents downloading with qBittorrent and Aria2c in seconds. Int
BASE_URL
: Valid BASE URL where the bot is deployed to use torrent/nzb web files selection. Format of URL should
be http://myip
, where myip
is the IP/Domain(public) of your bot or if you have chosen port other than 80
so
write it in this format http://myip:port
(http
and not https
). Str
BASE_URL_PORT
: Which is the BASE_URL Port. Default is 80
. Int
WEB_PINCODE
: Whether to ask for pincode before selecting files from torrent in web or not. Default
is False
. Bool
.
MaxConnections
,
decrease AsyncIOThreadsCount
, set limit of DiskWriteCacheSize
to 32
and decrease MemoryWorkingSetLimit
from qbittorrent.conf or bsetting command.8. JDownloader
JD_EMAIL
: jdownloader email sign up on JDownloaderJD_PASS
: jdownloader password
9. Sabnzbd
USENET_SERVERS
: list of dictionaries, you can add as much as you want and there is a button for servers in sabnzbd settings to edit current servers and add new servers.
[{'name': 'main', 'host': '', 'port': 563, 'timeout': 60, 'username': '', 'password': '', 'connections': 8, 'ssl': 1, 'ssl_verify': 2, 'ssl_ciphers': '', 'enable': 1, 'required': 0, 'optional': 0, 'retention': 0, 'send_group': 0, 'priority': 0}]
READ THIS FOR MORE INFORMATION
Open port 8070 in your vps to access full web interface from any device. Use it like http://ip:8070/sabnzbd/. username: mltb, password: mltbmltb
10. RSS
RSS_DELAY
: Time in seconds for rss refresh interval. Recommended 600
second at least. Default is 600
in
sec. Int
RSS_CHAT
: Chat ID or USERNAME or ID|TOPIC_ID or USERNAME|TOPIC_ID
where rss links will be sent. If you want message to be sent to the channel then add channel id. Add -100
before channel id. Int
|Str
RSS_CHAT
is required, otherwise monitor will not work. You must use USER_STRING_SESSION
--OR--
CHANNEL. If using channel then bot should be added in both channel and group(linked to channel) and RSS_CHAT
is the channel id, so messages sent by the bot to channel will be forwarded to group. Otherwise
with USER_STRING_SESSION
add group id for RSS_CHAT
. If DATABASE_URL
not added you will miss the feeds while
bot offline.11. Queue System
QUEUE_ALL
: Number of parallel tasks of downloads and uploads. For example if 20 task added and QUEUE_ALL
is 8
,
then the summation of uploading and downloading tasks are 8 and the rest in queue. Int
. NOTE: if you want to
fill QUEUE_DOWNLOAD
or QUEUE_UPLOAD
, then QUEUE_ALL
value must be greater than or equal to the greatest one and
less than or equal to summation of QUEUE_UPLOAD
and QUEUE_DOWNLOAD
.QUEUE_DOWNLOAD
: Number of all parallel downloading tasks. Int
QUEUE_UPLOAD
: Number of all parallel uploading tasks. Int
12. Torrent Search
SEARCH_API_LINK
: Search api app link. Get your api from deploying
this repository. Str
1337x, Piratebay, Nyaasi, Torlock, Torrent Galaxy, Zooqle, Kickass, Bitsearch, MagnetDL, Libgen, YTS, Limetorrent, TorrentFunk, Glodls, TorrentProject and YourBittorrent
SEARCH_LIMIT
: Search limit for search api, limit for each site and not overall result limit. Default is zero (
Default api limit for each site). Int
SEARCH_PLUGINS
: List of qBittorrent search plugins (github raw links). I have added some plugins, you can remove/add
plugins as you want. Main
Source: qBittorrent Search Plugins (Official/Unofficial). List
Make sure you still mount the app folder and installed the docker from official documentation.
sudo dockerd
sudo docker build . -t mltb
sudo docker run -p 80:80 -p 8080:8080 mltb
sudo docker ps
sudo docker stop id
NOTE: If you want to use ports other than 80 and 8080 for torrent file selection and rclone serve respectively, change it in docker-compose.yml also.
sudo apt install docker-compose-plugin
sudo docker compose up
sudo docker compose up --build
sudo docker compose stop
sudo docker compose start
sudo docker compose logs --follow
IMPORTANT NOTES:
Set BASE_URL_PORT
and RCLONE_SERVE_PORT
variables to any port you want to use. Default is 80
and 8080
respectively.
Check the number of processing units of your machine with nproc
cmd and times it by 4, then
edit AsyncIOThreadsCount
in qBittorrent.conf.
mirror - or /m Mirror
qbmirror - or /qm Mirror torrent using qBittorrent
jdmirror - or /jm Mirror using jdownloader
nzbmirror - or /nm Mirror using sabnzbd
ytdl - or /y Mirror yt-dlp supported links
leech - or /l Upload to telegram
qbleech - or /ql Leech torrent using qBittorrent
jdleech - or /jl Leech using jdownloader
nzbleech - or /nl Leech using sabnzbd
ytdlleech - or /yl Leech yt-dlp supported links
clone - Copy file/folder to Drive
count - Count file/folder from GDrive
usetting - or /us User settings
bsetting - or /bs Bot settings
status - Get Mirror Status message
sel - Select files from torrent
rss - Rss menu
list - Search files in Drive
search - Search for torrents with API
cancel - or /c Cancel a task
cancelall - Cancel all tasks
forcestart - or /fs to start task from queue
del - Delete file/folder from GDrive
log - Get the Bot Log
shell - Run commands in Shell
aexec - Execute async function
exec - Execute sync function
restart - Restart the Bot
stats - Bot Usage Stats
ping - Ping the Bot
help - All cmds with description
NOTES
Termux
.generate_drive_token.py
in local browser.pip3 install google-api-python-client google-auth-httplib2 google-auth-oauthlib
python3 generate_drive_token.py
rclone config
command.RCLONE_PATH
is like GDRIVE_ID
a default path for mirror. In additional to those variables DEFAULT_UPLOAD
to
choose the default tool whether it's rclone or google-api-python-client.DEFAULT_UPLOAD
= 'rc' then you must fill RCLONE_PATH
with path as default one or with rcl
to select
destination path on each new task.DEFAULT_UPLOAD
= 'gd' then you must fill GDRIVE_ID
with folder/TD id.rclone/{user_id}.conf
then RCLONE_PATH
must start with mrcc:
.mrcc:
at the beginning.UPSTREAM_REPO
variable can be used for edit/add any file in repository.UPSTREAM_REPO
private one in case you want to grab all files including private files.UPSTREAM_REPO
and your private files in this
private repository, so your private files will be overwritten from this repository. Also if you are using database for
private files, then all files from database will override the private files that added before deploying or from
private UPSTREAM_REPO
.UPSTREAM_REPO
with the official repository link, then be carefull incase any change in
requirements.txt your bot will not start after restart. In this case you need to deploy again with updated code to
install the new requirements or simply by changing the UPSTREAM_REPO
to you fork link with that old updates.UPSTREAM_REPO
with your fork link be carefull also if you fetched the commits from the
official repository.UPSTREAM_REPO
will take affect only after restart.-d
argument alone will lead to use global options for aria2c or qbittorrent.GlobalMaxRatio
and GlobalMaxSeedingMinutes
in qbittorrent.conf, -1
means no limit, but you can
cancel manually.
MaxRatioAction
.--seed-ratio
(0 means no limit) and --seed-time
(0 means no seed) in aria.sh.For Service Account to work, you must set
USE_SERVICE_ACCOUNTS
= "True" in config file or environment variables. NOTE: Using Service Accounts is only recommended while uploading to a Team Drive.
Let us create only the Service Accounts that we need.
Warning: Abuse of this feature is not the aim of this project and we do NOT recommend that you make a lot of projects, just one project and 100 SAs allow you plenty of use, its also possible that over abuse might get your projects banned by Google.
NOTE: If you have created SAs in past from this script, you can also just re download the keys by running:
python3 gen_sa_accounts.py --download-keys $PROJECTID
NOTE: 1 Service Account can upload/copy around 750 GB a day, 1 project can make 100 Service Accounts so you can upload 75 TB a day.
NOTE: All people can copy
2TB/DAY
from each file creator (uploader account), so if you got erroruserRateLimitExceeded
that doesn't mean your limit exceeded but file creator limit have been exceeded which is2TB/DAY
.
Choose one of these methods
python3 gen_sa_accounts.py --list-projects
python3 gen_sa_accounts.py --enable-services $PROJECTID
python3 gen_sa_accounts.py --create-sas $PROJECTID
python3 gen_sa_accounts.py --download-keys $PROJECTID
python3 gen_sa_accounts.py --quick-setup 1 --new-only
A folder named accounts will be created which will contain keys for the Service Accounts.
Choose one of these methods
cd accounts
For Windows using PowerShell
$emails = Get-ChildItem .**.json |Get-Content -Raw |ConvertFrom-Json |Select -ExpandProperty client_email >>emails.txt
For Linux
grep -oPh '"client_email": "K[^"]+' *.json > emails.txt
cd ..
Then add emails from emails.txt to Google Group, after that add this Google Group to your Shared Drive and promote it to manager and delete email.txt file from accounts folder
python3 add_to_team_drive.py -d SharedTeamDriveSrcID
https://mongodb.com/
and sign-up.Database
under Deployment
Header, your created cluster will be there.Allow Access From Anywhere
and press on Add IP Address
without editing the ip, then
create user.Choose a connection
, then press on Connect your application
. Choose Driver
*
python* and version
3.12 or later.connection string
and replace
with the password of your user, then press close.To use list from multi TD/folder. Run driveid.py in your terminal and follow it. It will generate list_drives.txt
file or u can simply create list_drives.txt
file in working directory and fill it, check below format:
DriveName folderID/tdID or `root` IndexLink(if available)
DriveName folderID/tdID or `root` IndexLink(if available)
Example:
TD1 root https://example.dev
TD2 0AO1JDB1t3i5jUk9PVA https://example.dev
For using your premium accounts in yt-dlp or for protected Index Links, create .netrc file according to following format:
Note: Create .netrc and not netrc, this file will be hidden, so view hidden files to edit it after creation.
Format:
machine host login username password my_password
Example:
machine instagram login anas.tayyar password mypassword
Instagram Note: You must login even if you want to download public posts and after first try you must confirm that this was you logged in from different ip(you can confirm from phone app).
Youtube Note: For youtube
authentication
use cookies.txt file.
Using Aria2c you can also use built in feature from bot with or without username. Here example for index link without username.
machine example.workers.dev password index_password
Where host is the name of extractor (eg. instagram, Twitch). Multiple accounts of different hosts can be added each separated by a new line.
If you feel like showing your appreciation for this project, then how about buying me a coffee.
Binance ID:
52187862
USDT Address:
TEzjjfkxLKQqndpsdpkA7jgiX7QQCL5p4f
Network:
TRC20
TRX Address:
TEzjjfkxLKQqndpsdpkA7jgiX7QQCL5p4f
Network:
TRC20
BTC Address:
17dkvxjqdc3yiaTs6dpjUB1TjV3tD7ScWe
ETH Address:
0xf798a8a1c72d593e16d8f3bb619ebd1a093c7309