diff options
Diffstat (limited to 'binance-scraping-bot.py')
-rw-r--r-- | binance-scraping-bot.py | 85 |
1 files changed, 85 insertions, 0 deletions
diff --git a/binance-scraping-bot.py b/binance-scraping-bot.py new file mode 100644 index 0000000..f96b6d0 --- /dev/null +++ b/binance-scraping-bot.py @@ -0,0 +1,85 @@ +import logging, os, time, tweepy +from telegram.ext import Updater, CommandHandler, MessageHandler, Filters +import bs4 as bs +import urllib.request + +# Create an environment variable and get the token +TG_TOKEN = os.environ.get('TG_TOKEN') + +# Enable logging +logging.basicConfig(format='%(asctime)s - %(name)s - %(levelname)s - %(message)s', level=logging.INFO) +logger = logging.getLogger(__name__) + +# Get the tokens +consumer_key = os.environ.get('TW_CONSUMER_KEY') +consumer_secret = os.environ.get('TW_CONSUMER_SECRET') +access_token = os.environ.get('TW_ACCESS_TOKEN') +access_token_secret = os.environ.get('TW_ACCESS_TOKEN_SECRET') + +# authentication of consumer key and secret +auth = tweepy.OAuthHandler(consumer_key, consumer_secret) + +# authentication of access token and secret +auth.set_access_token(access_token, access_token_secret) +api = tweepy.API(auth) + +# Binance extract function +def extract_binance(main_webpage, key_words): + final_item, final_list = [], [] + sauce = urllib.request.urlopen(main_webpage).read() + soup = bs.BeautifulSoup(sauce, 'lxml') + list = soup.find_all('li', class_ = 'article-list-item') + for article in list: + article_text = article.get_text().replace('\n', '') + for item in key_words: + if item in article_text: + final_item.append(article_text) + final_item.append('https://www.binance.com' + article.find('a').get('href')) + final_list.append(final_item) + final_item = [] # Reset once is in the final_list to not get duplicates + return final_list + +# Telegram function +def tg_call(update, context): + # Send a message when the command /start is triggered + update.message.reply_text("Hello mate! Let me start checking") + + # Create two empty list for storing and comparing urls + old_urls, news_urls = [], [] + + # Create a bag of key words for getting matches + key_words = ['List', 'list', 'Token Sale', 'Open Trading', 'open trading'] + + # Create the first pass + main_webpage = 'https://www.binance.com/en/support/categories/115000056351' + old_urls = extract_binance(main_webpage, key_words) + + # Loop pass - Watchdog mode + while True: + new_urls = extract_binance(main_webpage, key_words) + for item in new_urls: + if item not in old_urls: + msg = item[0] + '\n' + item[1] + api.update_status(msg) # Twitter + update.message.reply_text(msg) # Telegram + update.message.reply_text('Done for now. Time to go to sleep mate!') + time.sleep(900) # Sleep for 15 min + +# Main function +def main(): + # Create the updater + updater = Updater(TG_TOKEN, use_context=True) + + # Get the dispatcher to register handlers + dp = updater.dispatcher + + # Start the loop with /start + dp.add_handler(CommandHandler("start", tg_call)) + + # Start the Bot + updater.start_polling() + updater.idle() # killall python3.7 to kill the app + +# Start +if __name__ == '__main__': + main()
\ No newline at end of file |