|
|
@ -2,7 +2,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
import time
|
|
|
|
import time
|
|
|
|
from flask import url_for
|
|
|
|
from flask import url_for
|
|
|
|
from . util import live_server_setup
|
|
|
|
from .util import live_server_setup, wait_for_all_checks
|
|
|
|
|
|
|
|
|
|
|
|
from ..html_tools import *
|
|
|
|
from ..html_tools import *
|
|
|
|
|
|
|
|
|
|
|
@ -86,14 +86,14 @@ def test_check_xpath_filter_utf8(client, live_server):
|
|
|
|
follow_redirects=True
|
|
|
|
follow_redirects=True
|
|
|
|
)
|
|
|
|
)
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
time.sleep(1)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
res = client.post(
|
|
|
|
res = client.post(
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
|
data={"include_filters": filter, "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"},
|
|
|
|
data={"include_filters": filter, "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"},
|
|
|
|
follow_redirects=True
|
|
|
|
follow_redirects=True
|
|
|
|
)
|
|
|
|
)
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
time.sleep(3)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
res = client.get(url_for("index"))
|
|
|
|
res = client.get(url_for("index"))
|
|
|
|
assert b'Unicode strings with encoding declaration are not supported.' not in res.data
|
|
|
|
assert b'Unicode strings with encoding declaration are not supported.' not in res.data
|
|
|
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
|
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
|
@ -140,14 +140,14 @@ def test_check_xpath_text_function_utf8(client, live_server):
|
|
|
|
follow_redirects=True
|
|
|
|
follow_redirects=True
|
|
|
|
)
|
|
|
|
)
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
time.sleep(1)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
res = client.post(
|
|
|
|
res = client.post(
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
|
data={"include_filters": filter, "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"},
|
|
|
|
data={"include_filters": filter, "url": test_url, "tags": "", "headers": "", 'fetch_backend': "html_requests"},
|
|
|
|
follow_redirects=True
|
|
|
|
follow_redirects=True
|
|
|
|
)
|
|
|
|
)
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
time.sleep(3)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
res = client.get(url_for("index"))
|
|
|
|
res = client.get(url_for("index"))
|
|
|
|
assert b'Unicode strings with encoding declaration are not supported.' not in res.data
|
|
|
|
assert b'Unicode strings with encoding declaration are not supported.' not in res.data
|
|
|
|
|
|
|
|
|
|
|
@ -164,7 +164,6 @@ def test_check_xpath_text_function_utf8(client, live_server):
|
|
|
|
assert b'Deleted' in res.data
|
|
|
|
assert b'Deleted' in res.data
|
|
|
|
|
|
|
|
|
|
|
|
def test_check_markup_xpath_filter_restriction(client, live_server):
|
|
|
|
def test_check_markup_xpath_filter_restriction(client, live_server):
|
|
|
|
sleep_time_for_fetch_thread = 3
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
xpath_filter = "//*[contains(@class, 'sametext')]"
|
|
|
|
xpath_filter = "//*[contains(@class, 'sametext')]"
|
|
|
|
|
|
|
|
|
|
|
@ -183,7 +182,7 @@ def test_check_markup_xpath_filter_restriction(client, live_server):
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
|
|
|
|
|
|
|
|
# Give the thread time to pick it up
|
|
|
|
# Give the thread time to pick it up
|
|
|
|
time.sleep(sleep_time_for_fetch_thread)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
|
|
|
|
|
|
|
|
# Goto the edit page, add our ignore text
|
|
|
|
# Goto the edit page, add our ignore text
|
|
|
|
# Add our URL to the import page
|
|
|
|
# Add our URL to the import page
|
|
|
@ -195,7 +194,7 @@ def test_check_markup_xpath_filter_restriction(client, live_server):
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
|
|
|
|
|
|
|
|
# Give the thread time to pick it up
|
|
|
|
# Give the thread time to pick it up
|
|
|
|
time.sleep(sleep_time_for_fetch_thread)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
|
|
|
|
|
|
|
|
# view it/reset state back to viewed
|
|
|
|
# view it/reset state back to viewed
|
|
|
|
client.get(url_for("diff_history_page", uuid="first"), follow_redirects=True)
|
|
|
|
client.get(url_for("diff_history_page", uuid="first"), follow_redirects=True)
|
|
|
@ -206,7 +205,7 @@ def test_check_markup_xpath_filter_restriction(client, live_server):
|
|
|
|
# Trigger a check
|
|
|
|
# Trigger a check
|
|
|
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
|
|
|
client.get(url_for("form_watch_checknow"), follow_redirects=True)
|
|
|
|
# Give the thread time to pick it up
|
|
|
|
# Give the thread time to pick it up
|
|
|
|
time.sleep(sleep_time_for_fetch_thread)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
|
|
|
|
|
|
|
|
res = client.get(url_for("index"))
|
|
|
|
res = client.get(url_for("index"))
|
|
|
|
assert b'unviewed' not in res.data
|
|
|
|
assert b'unviewed' not in res.data
|
|
|
@ -216,9 +215,6 @@ def test_check_markup_xpath_filter_restriction(client, live_server):
|
|
|
|
|
|
|
|
|
|
|
|
def test_xpath_validation(client, live_server):
|
|
|
|
def test_xpath_validation(client, live_server):
|
|
|
|
|
|
|
|
|
|
|
|
# Give the endpoint time to spin up
|
|
|
|
|
|
|
|
time.sleep(1)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# Add our URL to the import page
|
|
|
|
# Add our URL to the import page
|
|
|
|
test_url = url_for('test_endpoint', _external=True)
|
|
|
|
test_url = url_for('test_endpoint', _external=True)
|
|
|
|
res = client.post(
|
|
|
|
res = client.post(
|
|
|
@ -227,7 +223,7 @@ def test_xpath_validation(client, live_server):
|
|
|
|
follow_redirects=True
|
|
|
|
follow_redirects=True
|
|
|
|
)
|
|
|
|
)
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
time.sleep(2)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
|
|
|
|
|
|
|
|
res = client.post(
|
|
|
|
res = client.post(
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
@ -244,11 +240,8 @@ def test_check_with_prefix_include_filters(client, live_server):
|
|
|
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
|
|
res = client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
|
|
assert b'Deleted' in res.data
|
|
|
|
assert b'Deleted' in res.data
|
|
|
|
|
|
|
|
|
|
|
|
# Give the endpoint time to spin up
|
|
|
|
|
|
|
|
time.sleep(1)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
set_original_response()
|
|
|
|
set_original_response()
|
|
|
|
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
# Add our URL to the import page
|
|
|
|
# Add our URL to the import page
|
|
|
|
test_url = url_for('test_endpoint', _external=True)
|
|
|
|
test_url = url_for('test_endpoint', _external=True)
|
|
|
|
res = client.post(
|
|
|
|
res = client.post(
|
|
|
@ -257,7 +250,7 @@ def test_check_with_prefix_include_filters(client, live_server):
|
|
|
|
follow_redirects=True
|
|
|
|
follow_redirects=True
|
|
|
|
)
|
|
|
|
)
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
time.sleep(3)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
|
|
|
|
|
|
|
|
res = client.post(
|
|
|
|
res = client.post(
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
@ -266,7 +259,7 @@ def test_check_with_prefix_include_filters(client, live_server):
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
time.sleep(3)
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
|
|
|
|
|
|
|
|
res = client.get(
|
|
|
|
res = client.get(
|
|
|
|
url_for("preview_page", uuid="first"),
|
|
|
|
url_for("preview_page", uuid="first"),
|
|
|
@ -277,3 +270,46 @@ def test_check_with_prefix_include_filters(client, live_server):
|
|
|
|
assert b"Some text that will change" not in res.data #not in selector
|
|
|
|
assert b"Some text that will change" not in res.data #not in selector
|
|
|
|
|
|
|
|
|
|
|
|
client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
|
|
client.get(url_for("form_delete", uuid="all"), follow_redirects=True)
|
|
|
|
|
|
|
|
def test_various_rules(client, live_server):
|
|
|
|
|
|
|
|
# Just check these don't error
|
|
|
|
|
|
|
|
#live_server_setup(live_server)
|
|
|
|
|
|
|
|
with open("test-datastore/endpoint-content.txt", "w") as f:
|
|
|
|
|
|
|
|
f.write("""<html>
|
|
|
|
|
|
|
|
<body>
|
|
|
|
|
|
|
|
Some initial text<br>
|
|
|
|
|
|
|
|
<p>Which is across multiple lines</p>
|
|
|
|
|
|
|
|
<br>
|
|
|
|
|
|
|
|
So let's see what happens. <br>
|
|
|
|
|
|
|
|
<div class="sametext">Some text thats the same</div>
|
|
|
|
|
|
|
|
<div class="changetext">Some text that will change</div>
|
|
|
|
|
|
|
|
<a href=''>some linky </a>
|
|
|
|
|
|
|
|
<a href=''>another some linky </a>
|
|
|
|
|
|
|
|
<!-- related to https://github.com/dgtlmoon/changedetection.io/pull/1774 -->
|
|
|
|
|
|
|
|
<input type="email" id="email" />
|
|
|
|
|
|
|
|
</body>
|
|
|
|
|
|
|
|
</html>
|
|
|
|
|
|
|
|
""")
|
|
|
|
|
|
|
|
test_url = url_for('test_endpoint', _external=True)
|
|
|
|
|
|
|
|
res = client.post(
|
|
|
|
|
|
|
|
url_for("import_page"),
|
|
|
|
|
|
|
|
data={"urls": test_url},
|
|
|
|
|
|
|
|
follow_redirects=True
|
|
|
|
|
|
|
|
)
|
|
|
|
|
|
|
|
assert b"1 Imported" in res.data
|
|
|
|
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for r in ['//div', '//a', 'xpath://div', 'xpath://a']:
|
|
|
|
|
|
|
|
res = client.post(
|
|
|
|
|
|
|
|
url_for("edit_page", uuid="first"),
|
|
|
|
|
|
|
|
data={"include_filters": r,
|
|
|
|
|
|
|
|
"url": test_url,
|
|
|
|
|
|
|
|
"tags": "",
|
|
|
|
|
|
|
|
"headers": "",
|
|
|
|
|
|
|
|
'fetch_backend': "html_requests"},
|
|
|
|
|
|
|
|
follow_redirects=True
|
|
|
|
|
|
|
|
)
|
|
|
|
|
|
|
|
wait_for_all_checks(client)
|
|
|
|
|
|
|
|
assert b"Updated watch." in res.data
|
|
|
|
|
|
|
|
res = client.get(url_for("index"))
|
|
|
|
|
|
|
|
assert b'fetch-error' not in res.data, f"Should not see errors after '{r} filter"
|
|
|
|