from selenium import webdriver from flask import Flask, request, redirect import concurrent.futures import re from collections import OrderedDict import time app = Flask(__name__) # Define the base URL for scraping base_url = "https://instagram.com" # Replace with your actual base URL # Initialize WebDriver globally user_agent = 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.50 Safari/537.36' options = webdriver.ChromeOptions() options.add_argument('--headless') options.add_argument('--disable-gpu') options.add_argument('--window-size=1920,1080') options.add_argument('--no-sandbox') options.add_argument(f'user-agent={user_agent}') browser = webdriver.Chrome(executable_path="/usr/bin/chromedriver", options=options) # Define the maximum cache size and duration in seconds (4 hours) MAX_CACHE_SIZE = 50 CACHE_DURATION = 4 * 60 * 60 # 4 hours in seconds cache = OrderedDict(maxlen=MAX_CACHE_SIZE) # Validate query, modify this regex as needed VALID_QUERY_REGEX = re.compile(r'^[\w\-\.\/]+$') # Function to handle web scraping using Selenium def get_video_source(query_string): try: browser.delete_all_cookies() query_string = "/" + query_string url = f"{base_url}{query_string}" # Combine base URL and video ID browser.get(url) # Replace sleep with explicit wait if possible browser.implicitly_wait(4) # Locate the video element using your specific xpath video_element = browser.find_element_by_xpath( "/html/body/div[2]/div/div/div[2]/div/div/div[1]/section/main/div[1]/div/article/div/div[1]/div/div/div/div/div/div/div/video" ) # Get the video source and return it video_source = video_element.get_attribute("src") return video_source except Exception as e: # Handle exceptions and return a default URL or re-raise the exception return base_url @app.route("/", methods=["GET"]) # Route for empty query string def handle_empty_query(): return redirect("https://github.com/gabrielkheisa/instagram-downloader") @app.route("/", methods=["GET"]) def get_video_source_server(query_string): global cache # Ensure we reference the global cache variable print(query_string) if len(query_string) > 30: return '', 204 if not VALID_QUERY_REGEX.match(query_string): return "Invalid link", 400 # Clean up entries older than 4 hours current_time = time.time() keys_to_remove = [] for key in list(cache.keys()): value = cache[key] if isinstance(value, dict) and "timestamp" in value: timestamp = value["timestamp"] if current_time - timestamp >= CACHE_DURATION: keys_to_remove.append(key) for key in keys_to_remove: cache.pop(key, None) if query_string in cache: # Move the existing entry to the front of the cache and update its timestamp video_source = cache.pop(query_string) video_source["timestamp"] = time.time() cache[query_string] = video_source return redirect(video_source["url"]) # Create a ThreadPoolExecutor for parallel execution with a timeout of 8 seconds with concurrent.futures.ThreadPoolExecutor() as executor: future = executor.submit(get_video_source, query_string) try: video_source = future.result(timeout=8) # Timeout set to 8 seconds # Add the new entry to the cache with a timestamp cache[query_string] = {"url": video_source, "timestamp": time.time()} return redirect(video_source) except concurrent.futures.TimeoutError: return redirect(base_url) # Handle timeout - return a default URL or handle as needed if __name__ == "__main__": app.run(debug=False, port=8080, host="0.0.0.0")