Automate Scrubbing your Facebook Presence
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 

116 lines
3.1 KiB

from .archive import archiver
from ..types import Page
from .common import SELENIUM_EXCEPTIONS, logger, click_button, wait_xpath, force_mobile, scroll_to, scroll_until_element_exists
from selenium.webdriver.common.by import By
from selenium.webdriver.support import expected_conditions as EC
from selenium.webdriver.support.ui import WebDriverWait
LOG = logger(__name__)
def load_likes(driver, profile_url):
"""
Loads the page that lists all pages you like
Args:
driver: seleniumrequests.Chrome Driver instance
Returns:
None
"""
likes_link_xpath = "//div[normalize-space(text())='Likes']/../..//a[contains(@href, 'app_section')]"
all_likes_link_xpath = "//div[normalize-space(text())='All Likes']/../../../..//a[contains(@href, 'app_collection')]"
driver.get(force_mobile("{0}/about".format(profile_url)))
scroll_until_element_exists(driver, "//div[text()='Likes']")
likes_link_el = driver.find_element_by_xpath(likes_link_xpath)
driver.get(likes_link_el.get_attribute("href"))
wait_xpath(driver, all_likes_link_xpath)
all_likes_link_el = driver.find_element_by_xpath(all_likes_link_xpath)
all_likes_link = all_likes_link_el.get_attribute("href")
driver.get(all_likes_link)
def get_page_links(driver):
"""
Gets all of the links to the pages you like
Args:
driver: seleniumrequests.Chrome Driver instance
Returns:
List of URLs to pages
"""
pages = driver.find_elements_by_xpath("//header/..//a")
return [page.get_attribute("href").replace("www", "mobile") for page in pages]
def unlike_page(driver, url, archive=None):
"""
Unlikes a page given the URL to it
Args:
driver: seleniumrequests.Chrome Driver instance
url: url string pointing to a page
archive: archiver instance
Returns:
None
"""
driver.get(url)
print("Unliking {0}".format(url))
wait = WebDriverWait(driver, 5)
try:
wait.until(
EC.presence_of_element_located((By.XPATH, "//*[text()='Liked']"))
)
except SELENIUM_EXCEPTIONS:
# Something went wrong with this page, so skip it
return
driver.find_element_by_xpath("//*[text()='Liked']/../../../..").click()
wait.until(
EC.presence_of_element_located((By.XPATH, "//*[text()='Unlike']"))
)
# There should be an "Unlike" button now, click it
driver.find_element_by_xpath("//*[text()='Unlike']/..").click()
if archive:
archive(Page(name=url))
def unlike_pages(driver, profile_url):
"""
Unlike all pages
Args:
driver: seleniumrequests.Chrome Driver instance
Returns:
None
"""
with archiver("likes") as archive_likes:
load_likes(driver, profile_url)
urls = get_page_links(driver)
while urls:
for url in urls:
unlike_page(driver, url, archive=archive_likes.archive)
try:
load_likes(driver, profile_url)
urls = get_page_links(driver)
except SELENIUM_EXCEPTIONS:
# We're done
break