1
0
This commit is contained in:
Jon Michael Aanes 2024-11-01 21:43:46 +01:00
parent a9e2cb3ccc
commit 9e76e5f0ee
Signed by: Jmaa
SSH Key Fingerprint: SHA256:Ab0GfHGCblESJx7JRE4fj4bFy/KRpeLhi41y4pF3sNA
4 changed files with 27 additions and 11 deletions

View File

@ -14,7 +14,14 @@ from enforce_typing import enforce_types
from socials_util._version import __version__
__all__ = ['__version__', 'SocialSiteId', 'SocialLink', 'WikidataInfo', 'AGGERAGOR_SOCIALS', 'determine_social_from_url']
__all__ = [
'__version__',
'SocialSiteId',
'SocialLink',
'WikidataInfo',
'AGGERAGOR_SOCIALS',
'determine_social_from_url',
]
class SocialSiteId(aenum.Enum):

View File

@ -1,5 +1,6 @@
import socials_util
def test_is_aggregator():
assert not socials_util.SocialSiteId.PAGE_WATCH.is_aggregator()
assert not socials_util.SocialSiteId.FALKON_PROFILE_BOOKMARKS.is_aggregator()

View File

@ -1,6 +1,7 @@
import urllib.parse
import pytest
import urllib.parse
import socials_util
from socials_util import SocialLink, SocialSiteId, determine_social_from_url
@ -188,11 +189,9 @@ PARSABLE_SOCIAL_IDS_COMBINED: list[tuple[str, object, str | None]] = [
# Cara
('https://cara.app/simzart', SocialSiteId.CARA_PROFILE, 'simzart'),
('https://cara.app/simzart/all', SocialSiteId.CARA_PROFILE, 'simzart'),
# Mastodon
('https://idlethumbs.social/@testtest', SocialSiteId.MASTODON_PAGE, None),
('https://mastodon.example.org/testtest', SocialSiteId.MASTODON_PAGE, None),
# Feeds
('https://example.org/main.atom', SocialSiteId.RSS_FEED, None),
]
@ -210,7 +209,9 @@ NOT_PARSABLE = [
('url', 'expected_social_site_id', 'expected_social_id'),
PARSABLE_SOCIAL_IDS_COMBINED,
)
def test_parse_social_ids(url: str, expected_social_site_id: SocialSiteId, expected_social_id: str) -> None:
def test_parse_social_ids(
url: str, expected_social_site_id: SocialSiteId, expected_social_id: str,
) -> None:
social_link: SocialLink | None = determine_social_from_url(url)
assert social_link is not None, url
assert (social_link.social_id, social_link.social_site_id) == (
@ -223,14 +224,19 @@ def test_parse_social_ids(url: str, expected_social_site_id: SocialSiteId, expec
def test_not_parsable(url: str) -> None:
assert determine_social_from_url(url) is None
def test_wrong_parse_type() -> None:
with pytest.raises(TypeError):
assert socials_util.to_parse_result(None)
def test_from_parse_result() -> None:
urlresult = urllib.parse.urlparse('https://old.reddit.com/user/Harpsibored/submitted/')
urlresult = urllib.parse.urlparse(
'https://old.reddit.com/user/Harpsibored/submitted/',
)
assert socials_util.to_parse_result(urlresult) is urlresult
def test_determine_social_from_url_internally() -> None:
with pytest.raises(TypeError):
assert socials_util.determine_social_from_url_internally(None)

View File

@ -1,9 +1,11 @@
import pytest
import socials_util
import wikidata
import socials_util
def test_wikidata_properties():
wikidata_client = wikidata.client.Client()
wikidata_property = socials_util.SocialSiteId.RSS_FEED.wikidata_property(wikidata_client)
wikidata_property = socials_util.SocialSiteId.RSS_FEED.wikidata_property(
wikidata_client,
)
assert wikidata_property is not None