An Rclone Mirror-Leech Telegram Bot to transfer to and from many clouds. Based on mirror-leech-telegram-bot with rclone support added, and other features and changes from base code.
- Copy file/folder from cloud to cloud
- Leech file/folder from cloud to Telegram
- Mirror Link/Torrent/Magnets/Mega/Telegram-Files to cloud
- Mirror from Telegram to multiple clouds at the same time
- Telegram Navigation Button Menus to interact with cloud
- File Manager: size, mkdir, delete, dedupe and rename
- Service Accounts support with automatic switching
- Create cloud index as http or webdav webserver
- Sync between clouds (not folders)
- Search files on cloud
- Clean cloud trash
- View cloud storage info
- Send rclone config file from bot
- Renaming menu for Telegram files
- Index support (rclone index for all remotes)
- Mirror and Leech files in batch from Telegram private/restricted channels
- Mirror and Leech links in batch from txt file
- Multizip mirror & leech
- Extract and zip link/file from Telegram to cloud
- Extract and zip folder/file from cloud to Telegram
- Mirror to local
- Queue system for mirror
- Refactor of the whole code to use only pyrogram with asyncio
mirror - Mirror to selected cloud
unzipmirror - Mirror and extract to cloud
zipmirror - Mirror and zip to cloud
multizipmirror - Mirror and zip multiple files to cloud
mirrorbatch - Mirror Telegram files and links in batch to cloud
cloudselect - Select cloud/folder for mirror
leech - Leech from cloud/link to Telegram
unzipleech - Leech and extract to Telegram
zipleech - Leech and zip to Telegram
multizipleech - Leech and zip multiple files to Telegram
leechbatch - Leech Telegram files/links in batch to Telegram
ytdl - Mirror ytdlp supported link
ytdlzip- Mirror and zip ytdlp supported link
ytdlleech - Leech yt-dlp supported link
ytdlzipleech - Leech and zip yt-dlp supported link
botfiles - Bot configuration files
myfiles - Rclone File Manager
copy - Copy from cloud to cloud
clone - Clone gdrive link file/folder
usetting - User settings
ownsetting - Owner settings
rss - Rss feed
cleanup - Clean cloud trash
cancelall - Cancel all tasks
storage - Cloud details
serve - Serve cloud as web index
sync - Sync two clouds
search - Search for torrents
status - Status message of tasks
stats - Bot stats
shell - Run cmds in shell
log - Bot log
ping - Ping bot
restart - Restart bot
- Installing requirements
-
Clone repo:
git clone https://github.com/Sam-Max/rclone-mirror-leech-telegram-bot rclonetgbot/ && cd rclonetgbot
-
For Debian based distros
sudo apt install python3 python3-pip
Install Docker by following the official Docker docs
- For Arch and it's derivatives:
sudo pacman -S docker python
- Install dependencies for running setup scripts:
pip3 install -r requirements-cli.txt
- Set up config file
-
cp config_sample.env config.env
-
Fill up fields: All values must be filled between quotes, even if
Int
,Bool
orList
.-
Mandatory Fields:
API_ID
: get this from https://my.telegram.org.Int
API_HASH
: get this from https://my.telegram.org.Str
BOT_TOKEN
: The Telegram Bot Token (get from @BotFather).Str
OWNER_ID
: your Telegram User ID (not username) of the owner of the bot.Int
-
Optional Fields:
DOWNLOAD_DIR
: The path to the local folder where the downloads will go.Str
SUDO_USERS
: Fill user_id of users whom you want to give sudo permission separated by spaces.Str
ALLOWED_CHATS
: list of IDs of allowed chats who can use this bot separated by spacesStr
AUTO_MIRROR
: For auto mirroring files sent to the bot. NOTE: If you add bot to group(not channel), you can also use this feature. Default isFalse
.Bool
DATABASE_URL
: Your Mongo Database URL (Connection string). Data will be saved in Database (auth and sudo users, owner and user setting, etc). NOTE: You can always edit all settings saved in database from mongodb site -> (browse collections).Str
CMD_INDEX
: index number that will be added at the end of all commands.Str
BASE_URL
: Ip (public/domain) where bot is running for local mirrors web listing/downloads. Format of URL should be http://myip, where myip is the IP/Domain(public).Str
SERVER_PORT
: Port for local mirrors web listing/downloads. Default to81
. `IntGD_INDEX_URL
: Refer to https://gitlab.com/ParveenBhadooOfficial/Google-Drive-Index.Str
VIEW_LINK
: View Link button to open file Google Drive Index Link in browser instead of direct download link, you can figure out if it's compatible with your Index code or not, open any video from you Index and check if its URL ends with?a=view
. Compatible with BhadooIndex Code. Default isFalse
.Bool
STATUS_LIMIT
: No. of tasks shown in status message with buttons. NOTE: Recommended limit is4
tasks.Int
TORRENT_TIMEOUT
: Timeout of dead torrents downloading with qBittorrentAUTO_DELETE_MESSAGE_DURATION
: Interval of time (in seconds), after which the bot deletes it's message and command message. Set to-1
to disable auto message deletion.Int
PARALLEL_TASKS
: Number of parallel tasks for queue system.Int
-
UPDATE
UPSTREAM_REPO
: if your repo is private add your github repo link with format:https://username:{githubtoken}@github.com/{username}/{reponame}
, so you can update your app from private repository on each restart. Get token from Github settings.Str
UPSTREAM_BRANCH
: Upstream branch for update. Default ismaster
.Str
NOTE: If any change in docker or requirements you will need to deploy/build again with updated repo for changes to apply.
-
RCLONE
DEFAULT_OWNER_REMOTE
: to set default remote from your rclone config for mirroring. (only for owner).Str
DEFAULT_GLOBAL_REMOTE
: to set default remote from global rclone config for mirroring. Use this whenMULTI_RCLONE_CONFIG
isFalse
.Str
MULTI_RCLONE_CONFIG
: set toTrue
for allowing each user to use their own rclone config. Default to False.Bool
REMOTE_SELECTION
: set toTrue
to activate selection of cloud each time using mirror command. Default toFalse
.Bool
MULTI_REMOTE_UP
= set toTrue
for allowing upload to multiple clouds at the same time.Bool
. (only for owner)USE_SERVICE_ACCOUNTS
: set toTrue
for enabling SA for rclone copy. Default to False.Bool
.SERVICE_ACCOUNTS_REMOTE
= To set shared drive remote name from your rclone config file that is using SA.Str
. Note: teamdrive remote must have team_drive field with id.Str
SERVER_SIDE
= set toTrue
for enabling rclone server side copy. Default to False. NOTE: if you get errors while copy set this toFalse
.Bool
RC_INDEX_URL
: Ip (public/domain) where bot is running for rclone index. Format of URL should be http://myip, where myip is the IP/Domain(public).Str
RC_INDEX_PORT
: Port to use. Default to8080
.Str
RC_INDEX_USER
: Custom user. Default toadmin
.Str
RC_INDEX_PASS
: Custom password. Default toadmin
.Str
-
CLONE
GDRIVE_FOLDER_ID
: Folder/TeamDrive ID of the Google Drive Folder orroot
to which you want to clone. Required forGoogle Drive
.Int
IS_TEAM_DRIVE
: SetTrue
if TeamDrive. Default isFalse
.Bool
EXTENSION_FILTER
: File extensions that won't clone. Separate them by space.Str
Notes: Must add token.pickle file directly to root for cloning to work. You can use /config command to add from bot.
-
LEECH
LEECH_SPLIT_SIZE
: Telegram upload limit in bytes, to automatically slice the file bigger that this size into small parts to upload to Telegram. Default is2GB
for non premium account or4GB
if your account is premium.Int
EQUAL_SPLITS
: Split files larger than LEECH_SPLIT_SIZE into equal parts size (not working with zip cmd). Default isFalse
.Bool
USER_SESSION_STRING
: Pyrogram session string for batch commands and for telegram premium upload. To generate string session use this commandpython3 session_generator.py
on command line on your pc from repository folder. NOTE: When using string session, you have to use withLEECH_LOG
. You can also use batch commands without string session, but you can't save messages from private/restricted telegram channels.Str
LEECH_LOG
: Chat ID. Upload files to specific chat/chats. Add chats separated by spaces.Str
NOTE: Only available for superGroup/channel. Add-100
before channel/supergroup id. Add bot in that channel/group as admin if using without string session.BOT_PM
: set toTrue
if you want to send leeched files in user's PM. Default to False.Bool
-
MEGA
MEGA_API_KEY
: Mega.nz API key to mirror mega.nz links. Get it from Mega SDK Page.Str
MEGA_EMAIL_ID
: E-Mail ID used to sign up on mega.nz for using premium account.Str
MEGA_PASSWORD
: Password for mega.nz account.Str
-
RSS
RSS_DELAY
: Time in seconds for rss refresh interval. Default is900
in sec.Int
RSS_COMMAND
: Choose command for the desired action.Str
RSS_CHAT_ID
: Chat ID where rss links will be sent. Add-100
before channel id.Str
RSS_USER_SESSION_STRING
: To send rss links from your telegram account. To generate session string use this commandpython3 generate_string_session.py
.Str
. NOTE: Don't use same session string asUSER_SESSION_STRING
.- RSS NOTE:
DATABASE_URL
andRSS_CHAT_ID
are required, otherwise rss commands will not work. You must use bot in group. You can also add the bot to a channel and link this channel to group so messages sent by bot to channel will be forwarded to group without usingRSS_USER_STRING_SESSION
.
-
QBITTORRENT
QB_BASE_URL
: Valid BASE URL where the bot is deployed to use qbittorrent web selection and local mirror. Format of URL should be http://myip, where myip is the IP/Domain(public). If you have chosen port other than 80 so write it in this format http://myip:port (http and not https).Str
QB_SERVER_PORT
: Port. Default to80
.Int
WEB_PINCODE
: If empty or False means no pincode required while torrent file web selection. Bool Qbittorrent NOTE: If your facing ram exceeded issue then set limit for MaxConnecs, decrease AsyncIOThreadsCount in qbittorrent config and set limit of DiskWriteCacheSize to 32.Int
-
TORRENT SEARCH
SEARCH_API_LINK
: Search api app link. Get your api from deploying this repository.Str
SEARCH_LIMIT
: Search limit for search api, limit for each site. Default is zero.Int
SEARCH_PLUGINS
: List of qBittorrent search plugins (github raw links). Add/Delete plugins as you wish. Main Source: qBittorrent Search Plugins (Official/Unofficial).List
-
- Deploying with Docker
-
Start Docker daemon (skip if already running)
sudo dockerd
-
Build Docker image:
sudo docker build . -t rcmltb
-
Run the image:
sudo docker run -p 80:80 -p 81:81 -p 8080:8080 rcmltb
-
To stop the image:
sudo docker ps sudo docker stop id
-
To clear the container:
sudo docker container prune
-
To delete the images:
sudo docker image prune -a
- Deploying using docker-compose
NOTE: If you want to use port other than 80, change it in docker-compose.yml
sudo apt install docker-compose
- Build and run Docker image:
sudo docker-compose up
- After editing files with nano for example (nano start.sh):
sudo docker-compose up --build
- To stop the image:
sudo docker-compose stop
- To run the image:
sudo docker-compose start
- Go to
https://mongodb.com/
and sign-up. - Create Shared Cluster.
- Press on
Database
underDeployment
Header, your created cluster will be there. - Press on connect, choose
Allow Acces From Anywhere
and press onAdd IP Address
without editing the ip, then create user. - After creating user press on
Choose a connection
, then press onConnect your application
. ChooseDriver
python andversion
3.6 or later. - Copy your
connection string
and replace<password>
with the password of your user, then press close.
Check this youtube video (not mine, credits to author):
Notes:
- When you create rclone.conf file add at least two accounts if you want to copy from cloud to cloud.
- For those on android phone, you can use RCX app app to create rclone.conf file. Use "Export rclone config" option in app menu to get config file.
- Rclone supported providers:
1Fichier, Amazon Drive, Amazon S3, Backblaze B2, Box, Ceph, DigitalOcean Spaces, Dreamhost, Dropbox, Enterprise File Fabric, FTP, GetSky, Google Cloud Storage, Google Drive, Google Photos, HDFS, HTTP, Hubic, IBM COS S3, Koofr, Mail.ru Cloud, Mega, Microsoft Azure Blob Storage, Microsoft OneDrive, Nextcloud, OVH, OpenDrive, Oracle Cloud Storage, ownCloud, pCloud, premiumize.me, put.io, Scaleway, Seafile, SFTP, WebDAV, Yandex Disk, etc. Check all providers on official site: Click here.
NOTES
- You need OS with a browser.
- Windows users should install python3 and pip. You can find how to install and use them from google.
- You can ONLY open the generated link from
generate_drive_token.py
in local browser.
- Visit the Google Cloud Console
- Go to the OAuth Consent tab, fill it, and save.
- Go to the Credentials tab and click Create Credentials -> OAuth Client ID
- Choose Desktop and Create.
- Publish your OAuth consent screen App to prevent token.pickle from expire
- Use the download button to download your credentials.
- Move that file to the root of rclone-tg-bot, and rename it to credentials.json
- Visit Google API page
- Search for Google Drive Api and enable it
- Finally, run the script to generate token.pickle file for Google Drive:
pip3 install google-api-python-client google-auth-httplib2 google-auth-oauthlib
python3 generate_drive_token.py
NOTE: Using Service Accounts is only recommended for Team Drive.
Warning: Abuse of this feature is not the aim of this project and we do not recommend that you make a lot of projects, just one project and 100 SAs allow you plenty of use, its also possible that over abuse might get your projects banned by Google.
NOTE: If you have created SAs in past from this script, you can also just re download the keys by running:
python3 gen_sa_accounts.py --download-keys $PROJECTID
NOTE: 1 Service Account can copy around 750 GB a day, 1 project can make 100 Service Accounts so you can copy 75 TB a day.
Choose one of these methods
- List your projects ids
python3 gen_sa_accounts.py --list-projects
- Enable services automatically by this command
python3 gen_sa_accounts.py --enable-services $PROJECTID
- Create Sevice Accounts to current project
python3 gen_sa_accounts.py --create-sas $PROJECTID
- Download Sevice Accounts as accounts folder
python3 gen_sa_accounts.py --download-keys $PROJECTID
python3 gen_sa_accounts.py --quick-setup 1 --new-only
A folder named accounts will be created which will contain keys for the Service Accounts.
Choose one of these methods
- Mount accounts folder
cd accounts
- Grab emails form all accounts to emails.txt file that would be created in accounts folder
For Windows using PowerShell
$emails = Get-ChildItem .\**.json |Get-Content -Raw |ConvertFrom-Json |Select -ExpandProperty client_email >>emails.txt
For Linux
grep -oPh '"client_email": "\K[^"]+' *.json > emails.txt
- Unmount acounts folder
cd ..
Then add emails from emails.txt to Google Group, after that add this Google Group to your Shared Drive and promote it to manager and delete email.txt file from accounts folder
- Run:
python3 add_to_team_drive.py -d SharedTeamDriveSrcID
For using your premium accounts in yt-dlp or for protected Index Links, create .netrc and not netrc, this file will be hidden, so view hidden files to edit it after creation. Use following format on file:
Format:
machine host login username password my_password
Example:
machine example.workers.dev login index_username password index_password
Note: Using aria2c you can also use without username. Multiple accounts of different hosts can be added each separated by a new line.
Youtube Note: For youtube
authentication use cookies.txt file.