Skip to content

Commit

Permalink
typing
Browse files Browse the repository at this point in the history
  • Loading branch information
Tiendil committed Dec 27, 2024
1 parent c1c8feb commit 446f70e
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 4 deletions.
8 changes: 5 additions & 3 deletions ffun/ffun/cli/commands/estimates.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
from ffun.application.application import with_app
from ffun.core import logging
from ffun.domain.entities import CollectionId, FeedUrl, UnknownUrl
from ffun.domain.urls import normalize_classic_unknown_url
from ffun.domain.urls import normalize_classic_unknown_url, to_feed_url
from ffun.feeds_collections.collections import collections
from ffun.loader.domain import extract_feed_info

Expand Down Expand Up @@ -46,7 +46,9 @@ async def run_entries_per_day_for_feed(raw_feed_url: str) -> None:
async with with_app():
feed_url = normalize_classic_unknown_url(UnknownUrl(raw_feed_url))

entries_per_day = await _estimate_entries_per_day(feed_url)
assert feed_url is not None

entries_per_day = await _estimate_entries_per_day(to_feed_url(feed_url))

logger.info("estimated_entries_for_feed_in_day", feed_url=feed_url, entries_per_day=entries_per_day)

Expand All @@ -59,7 +61,7 @@ async def run_entries_per_day_for_collection(collection_id: uuid.UUID) -> None:

urls = {feed_info.url for feed_info in collection.feeds}

total_entries_per_day = 0
total_entries_per_day = 0.0

for url in urls:
entries_per_day = await _estimate_entries_per_day(url)
Expand Down
2 changes: 1 addition & 1 deletion ffun/ffun/loader/domain.py
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@ async def detect_orphaned(feed_id: FeedId) -> bool:


# TODO: tests
async def extract_feed_info(feed_id: FeedId | None, feed_url: str) -> p_entities.FeedInfo | None:
async def extract_feed_info(feed_id: FeedId | None, feed_url: FeedUrl) -> p_entities.FeedInfo | None:
try:
response = await load_content_with_proxies(feed_url)
content = await operations.decode_content(response)
Expand Down

0 comments on commit 446f70e

Please sign in to comment.