1
0

Compare commits

...

3 Commits

Author SHA1 Message Date
1cda90c0bc
Fixed tests
All checks were successful
Python Ruff Code Quality / ruff (push) Successful in 22s
Run Python tests (through Pytest) / Test (push) Successful in 25s
Verify Python project can be installed, loaded and have version checked / Test (push) Successful in 22s
2024-11-01 21:45:28 +01:00
9e76e5f0ee
ruff 2024-11-01 21:43:46 +01:00
a9e2cb3ccc
Wikidata client and property 2024-11-01 21:42:16 +01:00
6 changed files with 62 additions and 4 deletions

View File

@ -1 +1,2 @@
pytest
wikidata

View File

@ -14,7 +14,14 @@ from enforce_typing import enforce_types
from socials_util._version import __version__
__all__ = ['__version__', 'SocialSiteId', 'SocialLink', 'WikidataInfo', 'AGGERAGOR_SOCIALS', 'determine_social_from_url']
__all__ = [
'__version__',
'SocialSiteId',
'SocialLink',
'WikidataInfo',
'AGGERAGOR_SOCIALS',
'determine_social_from_url',
]
class SocialSiteId(aenum.Enum):

6
test/test_data.py Normal file
View File

@ -0,0 +1,6 @@
import socials_util
def test_is_aggregator():
assert not socials_util.SocialSiteId.PAGE_WATCH.is_aggregator()
assert not socials_util.SocialSiteId.FALKON_PROFILE_BOOKMARKS.is_aggregator()

View File

@ -1,8 +1,14 @@
import aenum
import pytest
import socials_util
def test_re_social_path_validation():
with pytest.raises(ValueError, match='Redundant www: www.example.org'):
socials_util.re_social_path('www.example.org')
def test_extension():
my_secret_site = aenum.extend_enum(socials_util.SocialSiteId, 'MY_SECRET_SITE', 666)
assert my_secret_site

View File

@ -1,8 +1,11 @@
import urllib.parse
import pytest
import socials_util
from socials_util import SocialLink, SocialSiteId, determine_social_from_url
PARSABLE_SOCIAL_IDS_COMBINED: list[tuple[str, object, str]] = [
PARSABLE_SOCIAL_IDS_COMBINED: list[tuple[str, object, str | None]] = [
# Tumblr formats
('https://triviallytrue.tumblr.com/', SocialSiteId.TUMBLR, 'triviallytrue'),
('https://www.triviallytrue.tumblr.com/', SocialSiteId.TUMBLR, 'triviallytrue'),
@ -186,6 +189,11 @@ PARSABLE_SOCIAL_IDS_COMBINED: list[tuple[str, object, str]] = [
# Cara
('https://cara.app/simzart', SocialSiteId.CARA_PROFILE, 'simzart'),
('https://cara.app/simzart/all', SocialSiteId.CARA_PROFILE, 'simzart'),
# Mastodon
('https://idlethumbs.social/@testtest', SocialSiteId.MASTODON_PAGE, None),
('https://mastodon.example.org/testtest', SocialSiteId.MASTODON_PAGE, None),
# Feeds
('https://example.org/main.atom', SocialSiteId.RSS_FEED, None),
]
NOT_PARSABLE = [
@ -198,10 +206,12 @@ NOT_PARSABLE = [
@pytest.mark.parametrize(
('url','expected_social_site_id','expected_social_id'),
('url', 'expected_social_site_id', 'expected_social_id'),
PARSABLE_SOCIAL_IDS_COMBINED,
)
def test_parse_social_ids(url: str, expected_social_site_id: SocialSiteId, expected_social_id: str) -> None:
def test_parse_social_ids(
url: str, expected_social_site_id: SocialSiteId, expected_social_id: str,
) -> None:
social_link: SocialLink | None = determine_social_from_url(url)
assert social_link is not None, url
assert (social_link.social_id, social_link.social_site_id) == (
@ -213,3 +223,20 @@ def test_parse_social_ids(url: str, expected_social_site_id: SocialSiteId, expec
@pytest.mark.parametrize('url', NOT_PARSABLE)
def test_not_parsable(url: str) -> None:
assert determine_social_from_url(url) is None
def test_wrong_parse_type() -> None:
with pytest.raises(TypeError):
assert socials_util.to_parse_result(None)
def test_from_parse_result() -> None:
urlresult = urllib.parse.urlparse(
'https://old.reddit.com/user/Harpsibored/submitted/',
)
assert socials_util.to_parse_result(urlresult) is urlresult
def test_determine_social_from_url_internally() -> None:
with pytest.raises(TypeError):
assert socials_util.determine_social_from_url_internally(None)

View File

@ -0,0 +1,11 @@
import wikidata.client
import socials_util
def test_wikidata_properties():
wikidata_client = wikidata.client.Client()
wikidata_property = socials_util.SocialSiteId.RSS_FEED.wikidata_property(
wikidata_client,
)
assert wikidata_property is not None