1
0

Compare commits

..

No commits in common. "1cda90c0bcc36c3436befbf372aad8ccbfa64ab0" and "431962c02f2bdd718b767325277baa8c914f7124" have entirely different histories.

6 changed files with 4 additions and 62 deletions

View File

@ -1,2 +1 @@
pytest
wikidata

View File

@ -14,14 +14,7 @@ from enforce_typing import enforce_types
from socials_util._version import __version__
__all__ = [
'__version__',
'SocialSiteId',
'SocialLink',
'WikidataInfo',
'AGGERAGOR_SOCIALS',
'determine_social_from_url',
]
__all__ = ['__version__', 'SocialSiteId', 'SocialLink', 'WikidataInfo', 'AGGERAGOR_SOCIALS', 'determine_social_from_url']
class SocialSiteId(aenum.Enum):

View File

@ -1,6 +0,0 @@
import socials_util
def test_is_aggregator():
assert not socials_util.SocialSiteId.PAGE_WATCH.is_aggregator()
assert not socials_util.SocialSiteId.FALKON_PROFILE_BOOKMARKS.is_aggregator()

View File

@ -1,14 +1,8 @@
import aenum
import pytest
import socials_util
def test_re_social_path_validation():
with pytest.raises(ValueError, match='Redundant www: www.example.org'):
socials_util.re_social_path('www.example.org')
def test_extension():
my_secret_site = aenum.extend_enum(socials_util.SocialSiteId, 'MY_SECRET_SITE', 666)
assert my_secret_site

View File

@ -1,11 +1,8 @@
import urllib.parse
import pytest
import socials_util
from socials_util import SocialLink, SocialSiteId, determine_social_from_url
PARSABLE_SOCIAL_IDS_COMBINED: list[tuple[str, object, str | None]] = [
PARSABLE_SOCIAL_IDS_COMBINED: list[tuple[str, object, str]] = [
# Tumblr formats
('https://triviallytrue.tumblr.com/', SocialSiteId.TUMBLR, 'triviallytrue'),
('https://www.triviallytrue.tumblr.com/', SocialSiteId.TUMBLR, 'triviallytrue'),
@ -189,11 +186,6 @@ PARSABLE_SOCIAL_IDS_COMBINED: list[tuple[str, object, str | None]] = [
# Cara
('https://cara.app/simzart', SocialSiteId.CARA_PROFILE, 'simzart'),
('https://cara.app/simzart/all', SocialSiteId.CARA_PROFILE, 'simzart'),
# Mastodon
('https://idlethumbs.social/@testtest', SocialSiteId.MASTODON_PAGE, None),
('https://mastodon.example.org/testtest', SocialSiteId.MASTODON_PAGE, None),
# Feeds
('https://example.org/main.atom', SocialSiteId.RSS_FEED, None),
]
NOT_PARSABLE = [
@ -209,9 +201,7 @@ NOT_PARSABLE = [
('url','expected_social_site_id','expected_social_id'),
PARSABLE_SOCIAL_IDS_COMBINED,
)
def test_parse_social_ids(
url: str, expected_social_site_id: SocialSiteId, expected_social_id: str,
) -> None:
def test_parse_social_ids(url: str, expected_social_site_id: SocialSiteId, expected_social_id: str) -> None:
social_link: SocialLink | None = determine_social_from_url(url)
assert social_link is not None, url
assert (social_link.social_id, social_link.social_site_id) == (
@ -223,20 +213,3 @@ def test_parse_social_ids(
@pytest.mark.parametrize('url', NOT_PARSABLE)
def test_not_parsable(url: str) -> None:
assert determine_social_from_url(url) is None
def test_wrong_parse_type() -> None:
with pytest.raises(TypeError):
assert socials_util.to_parse_result(None)
def test_from_parse_result() -> None:
urlresult = urllib.parse.urlparse(
'https://old.reddit.com/user/Harpsibored/submitted/',
)
assert socials_util.to_parse_result(urlresult) is urlresult
def test_determine_social_from_url_internally() -> None:
with pytest.raises(TypeError):
assert socials_util.determine_social_from_url_internally(None)

View File

@ -1,11 +0,0 @@
import wikidata.client
import socials_util
def test_wikidata_properties():
wikidata_client = wikidata.client.Client()
wikidata_property = socials_util.SocialSiteId.RSS_FEED.wikidata_property(
wikidata_client,
)
assert wikidata_property is not None