-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathlogic.py
180 lines (138 loc) · 5.03 KB
/
logic.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
from typing import Any
import time
import math
import sqlalchemy
import datetime
import pandas as pd
from sqlalchemy.orm import scoped_session
from update import update_feed
from models import Item, Feed, UpdateStat
# number of items per page
PAGE_SIZE = 48
# number of decimal places for printing durations
TIMING_PRECISION = 3
def last_update_stats(session: scoped_session) -> UpdateStat | None:
last_stats = (
session.query(UpdateStat).order_by(UpdateStat.timestamp.desc()).limit(1).first()
)
if last_stats:
last_stats.dur_total = round(last_stats.dur_total, TIMING_PRECISION)
return last_stats
def item_list(
session: scoped_session, offset: int, specific_feed_id: int | None
) -> dict[str, Any]:
start_time = time.time()
items_query = session.query(Item).order_by(Item.published.desc())
if specific_feed_id:
items_query = items_query.where(Item.feed_id == specific_feed_id)
items = items_query.offset(offset).limit(PAGE_SIZE).all()
last_stats = last_update_stats(session)
specific_feed = None
if specific_feed_id:
specific_feed = session.query(Feed).get(specific_feed_id)
load_time = time.time()
return {
"items": items,
"load_time": round(load_time - start_time, TIMING_PRECISION),
"last_stats": last_stats,
"prev_page": offset - PAGE_SIZE if offset > PAGE_SIZE else 0,
"next_page": offset + PAGE_SIZE,
"from_feed": specific_feed,
}
# maximum age of each item displayed
OVERVIEW_NUM_DAYS_SINCE = 30
# maximum number of items to display per feed
OVERVIEW_ITEMS_PER_FEED = 7
def overview(session: scoped_session) -> dict[str, Any]:
start_time = time.time()
since_date = datetime.datetime.now() - datetime.timedelta(
days=OVERVIEW_NUM_DAYS_SINCE
)
# Subquery to get the maximum published date and count of items for each feed
subquery = (
session.query(
Item.feed_id,
sqlalchemy.func.max(Item.published).label("last_published"),
sqlalchemy.func.count(Item.id).label("item_count"),
)
.filter(Item.published >= since_date, Item.visited == False)
.group_by(Item.feed_id)
.subquery()
)
# Query to fetch feeds that have items since `since_date`, ordered by feed last published and item count
feeds_with_max = (
session.query(Feed)
.join(subquery, Feed.id == subquery.c.feed_id)
.order_by(subquery.c.last_published.desc(), subquery.c.item_count.asc())
.all()
)
items_by_feed = []
for feed in feeds_with_max:
recent_items = (
session.query(Item)
.filter(
Item.feed_id == feed.id,
Item.published >= since_date,
Item.visited == False,
)
.order_by(Item.published.desc())
.limit(OVERVIEW_ITEMS_PER_FEED)
.all()
)
top_item_age = (datetime.datetime.now() - recent_items[0].published).days
rank = math.cos(top_item_age * 2 * math.pi / 30) - top_item_age * 0.01
if feed.downrank:
rank -= 100
recent_items = recent_items[0 : (OVERVIEW_ITEMS_PER_FEED // 2)]
# Append the feed and its items to the result list
items_by_feed.append(
{"id": feed.id, "title": feed.title, "items": recent_items, "rank": rank}
)
items_by_feed.sort(key=lambda x: -x["rank"])
last_stats = last_update_stats(session)
load_time = time.time()
return {
"feeds": items_by_feed,
"load_time": round(load_time - start_time, TIMING_PRECISION),
"last_stats": last_stats,
}
def add_feed(session: scoped_session, url: str):
new_feed = Feed(url=url)
update_feed(session, new_feed)
session.add(new_feed)
session.commit()
def delete_feed(session: scoped_session, id: int):
feed = session.query(Feed).get(id)
session.delete(feed)
session.commit()
def toggle_feed_downrank(session: scoped_session, id: int):
feed = session.query(Feed).get(id)
feed.downrank = not feed.downrank
session.commit()
def feed_list(session: scoped_session):
feeds = session.query(Feed).order_by(Feed.last_updated.desc()).all()
return {"feeds": feeds}
def fetch_update_stats(session: scoped_session, timeframe: str) -> pd.DataFrame:
timeframe_expr = {
"day": "-1 day",
"week": "-7 days",
"month": "-1 month",
}.get(timeframe)
return pd.read_sql_query(
f"""
SELECT
update_stats.*,
min_feed.url AS min_feed_url,
max_feed.url AS max_feed_url
FROM update_stats
LEFT JOIN feed min_feed ON update_stats.dur_min_feed_id = min_feed.id
LEFT JOIN feed max_feed ON update_stats.dur_max_feed_id = max_feed.id
WHERE update_stats.timestamp >= DATE("now", "{timeframe_expr}")
""",
con=session.connection(),
)
def record_visit(session: scoped_session, item_id: int):
item = session.query(Item).get(item_id)
if item:
item.visited = True
session.commit()