# -*- coding: utf-8 -*- import time from flask import url_for from .util import live_server_setup, wait_for_all_checks from ..html_tools import * def test_setup(live_server): live_server_setup(live_server) def set_original_response(): test_return_data = """ Some initial text

Which is across multiple lines


So let's see what happens.
Some text thats the same
Some text that will change
""" with open("test-datastore/endpoint-content.txt", "w") as f: f.write(test_return_data) return None def set_modified_response(): test_return_data = """ Some initial text

Which is across multiple lines


So let's see what happens. THIS CHANGES AND SHOULDNT TRIGGER A CHANGE
Some text thats the same
Some new text
""" with open("test-datastore/endpoint-content.txt", "w") as f: f.write(test_return_data) return None # Handle utf-8 charset replies https://github.com/dgtlmoon/changedetection.io/pull/613 def test_check_xpath_filter_utf8(client, live_server, measure_memory_usage): filter = '//item/*[self::description]' d = ''' rpilocator.com https://rpilocator.com Find Raspberry Pi Computers in Stock Thu, 19 May 2022 23:27:30 GMT https://rpilocator.com/favicon.png rpilocator.com https://rpilocator.com/ 32 32 Stock Alert (UK): RPi CM4 - 1GB RAM, No MMC, No Wifi is In Stock at Pimoroni Stock Alert (UK): RPi CM4 - 1GB RAM, No MMC, No Wifi is In Stock at Pimoroni https://rpilocator.com?vendor=pimoroni&utm_source=feed&utm_medium=rss pimoroni UK CM4 F9FAB0D9-DF6F-40C8-8DEE5FC0646BB722 Thu, 19 May 2022 14:32:32 GMT ''' with open("test-datastore/endpoint-content.txt", "w") as f: f.write(d) # Add our URL to the import page test_url = url_for('test_endpoint', _external=True, content_type="application/rss+xml;charset=UTF-8") res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": filter, "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"Updated watch." in res.data wait_for_all_checks(client) res = client.get(url_for("index")) assert b'Unicode strings with encoding declaration are not supported.' not in res.data res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data # Handle utf-8 charset replies https://github.com/dgtlmoon/changedetection.io/pull/613 def test_check_xpath_text_function_utf8(client, live_server, measure_memory_usage): filter = '//item/title/text()' d = ''' rpilocator.com https://rpilocator.com Find Raspberry Pi Computers in Stock Thu, 19 May 2022 23:27:30 GMT https://rpilocator.com/favicon.png rpilocator.com https://rpilocator.com/ 32 32 Stock Alert (UK): RPi CM4 something else unrelated Stock Alert (UK): Big monitor something else unrelated ''' with open("test-datastore/endpoint-content.txt", "w") as f: f.write(d) # Add our URL to the import page test_url = url_for('test_endpoint', _external=True, content_type="application/rss+xml;charset=UTF-8") res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": filter, "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"Updated watch." in res.data wait_for_all_checks(client) res = client.get(url_for("index")) assert b'Unicode strings with encoding declaration are not supported.' not in res.data # The service should echo back the request headers res = client.get( url_for("preview_page", uuid="first"), follow_redirects=True ) assert b'Stock Alert (UK): RPi CM4' in res.data assert b'Stock Alert (UK): Big monitor' in res.data res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data def test_check_markup_xpath_filter_restriction(client, live_server, measure_memory_usage): xpath_filter = "//*[contains(@class, 'sametext')]" set_original_response() # Add our URL to the import page test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data # Give the thread time to pick it up wait_for_all_checks(client) # Goto the edit page, add our ignore text # Add our URL to the import page res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": xpath_filter, "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"Updated watch." in res.data # Give the thread time to pick it up wait_for_all_checks(client) # view it/reset state back to viewed client.get(url_for("diff_history_page", uuid="first"), follow_redirects=True) # Make a change set_modified_response() # Trigger a check client.get(url_for("form_watch_checknow"), follow_redirects=True) # Give the thread time to pick it up wait_for_all_checks(client) res = client.get(url_for("index")) assert b'unviewed' not in res.data res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data def test_xpath_validation(client, live_server, measure_memory_usage): # Add our URL to the import page test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": "/something horrible", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"is not a valid XPath expression" in res.data res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data def test_xpath23_prefix_validation(client, live_server, measure_memory_usage): # Add our URL to the import page test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": "xpath:/something horrible", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"is not a valid XPath expression" in res.data res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data def test_xpath1_lxml(client, live_server, measure_memory_usage): #live_server_setup(live_server) d = ''' rpilocator.com https://rpilocator.com Find Raspberry Pi Computers in Stock Thu, 19 May 2022 23:27:30 GMT https://rpilocator.com/favicon.png rpilocator.com https://rpilocator.com/ 32 32 Stock Alert (UK): RPi CM4 something else unrelated Stock Alert (UK): Big monitorěěěě something else unrelated '''.encode('utf-8') with open("test-datastore/endpoint-content.txt", "wb") as f: f.write(d) test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": "xpath1://title/text()", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) ##### #2312 wait_for_all_checks(client) res = client.get(url_for("index")) assert b'_ElementStringResult' not in res.data # tested with 5.1.1 when it was removed and 5.1.0 assert b'Exception' not in res.data res = client.get( url_for("preview_page", uuid="first"), follow_redirects=True ) assert b"rpilocator.com" in res.data # in selector assert "Stock Alert (UK): Big monitorěěěě".encode('utf-8') in res.data # not in selector ##### def test_xpath1_validation(client, live_server, measure_memory_usage): # Add our URL to the import page test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": "xpath1:/something horrible", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"is not a valid XPath expression" in res.data res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data # actually only really used by the distll.io importer, but could be handy too def test_check_with_prefix_include_filters(client, live_server, measure_memory_usage): res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data set_original_response() wait_for_all_checks(client) # Add our URL to the import page test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": "xpath://*[contains(@class, 'sametext')]", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"Updated watch." in res.data wait_for_all_checks(client) res = client.get( url_for("preview_page", uuid="first"), follow_redirects=True ) assert b"Some text thats the same" in res.data # in selector assert b"Some text that will change" not in res.data # not in selector client.get(url_for("form_delete", uuid="all"), follow_redirects=True) def test_various_rules(client, live_server, measure_memory_usage): # Just check these don't error # live_server_setup(live_server) with open("test-datastore/endpoint-content.txt", "w") as f: f.write(""" Some initial text

Which is across multiple lines


So let's see what happens.
Some text thats the same
Some text that will change
some linky another some linky """) test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) for r in ['//div', '//a', 'xpath://div', 'xpath://a']: res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": r, "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) wait_for_all_checks(client) assert b"Updated watch." in res.data res = client.get(url_for("index")) assert b'fetch-error' not in res.data, f"Should not see errors after '{r} filter" res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data def test_xpath_20(client, live_server, measure_memory_usage): test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) set_original_response() test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": "//*[contains(@class, 'sametext')]|//*[contains(@class, 'changetext')]", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"Updated watch." in res.data wait_for_all_checks(client) res = client.get( url_for("preview_page", uuid="first"), follow_redirects=True ) assert b"Some text thats the same" in res.data # in selector assert b"Some text that will change" in res.data # in selector client.get(url_for("form_delete", uuid="all"), follow_redirects=True) def test_xpath_20_function_count(client, live_server, measure_memory_usage): set_original_response() # Add our URL to the import page test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": "xpath:count(//div) * 123456789987654321", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"Updated watch." in res.data wait_for_all_checks(client) res = client.get( url_for("preview_page", uuid="first"), follow_redirects=True ) assert b"246913579975308642" in res.data # in selector client.get(url_for("form_delete", uuid="all"), follow_redirects=True) def test_xpath_20_function_count2(client, live_server, measure_memory_usage): set_original_response() # Add our URL to the import page test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={"include_filters": "/html/body/count(div) * 123456789987654321", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"Updated watch." in res.data wait_for_all_checks(client) res = client.get( url_for("preview_page", uuid="first"), follow_redirects=True ) assert b"246913579975308642" in res.data # in selector client.get(url_for("form_delete", uuid="all"), follow_redirects=True) def test_xpath_20_function_string_join_matches(client, live_server, measure_memory_usage): set_original_response() # Add our URL to the import page test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) res = client.post( url_for("edit_page", uuid="first"), data={ "include_filters": "xpath:string-join(//*[contains(@class, 'sametext')]|//*[matches(@class, 'changetext')], 'specialconjunction')", "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"}, follow_redirects=True ) assert b"Updated watch." in res.data wait_for_all_checks(client) res = client.get( url_for("preview_page", uuid="first"), follow_redirects=True ) assert b"Some text thats the samespecialconjunctionSome text that will change" in res.data # in selector client.get(url_for("form_delete", uuid="all"), follow_redirects=True)