feat: add grouped time analysis endpoint
This commit is contained in:
@@ -3,6 +3,7 @@ from flask_cors import CORS
|
|||||||
from server.stat_gen import StatGen
|
from server.stat_gen import StatGen
|
||||||
|
|
||||||
import pandas as pd
|
import pandas as pd
|
||||||
|
import traceback
|
||||||
|
|
||||||
app = Flask(__name__)
|
app = Flask(__name__)
|
||||||
|
|
||||||
@@ -103,6 +104,19 @@ def get_summary():
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
return jsonify({"error": f"An unexpected error occurred: {str(e)}"}), 500
|
return jsonify({"error": f"An unexpected error occurred: {str(e)}"}), 500
|
||||||
|
|
||||||
|
@app.route("/stats/time", methods=["GET"])
|
||||||
|
def get_time_analysis():
|
||||||
|
if stat_obj is None:
|
||||||
|
return jsonify({"error": "No data uploaded"}), 400
|
||||||
|
|
||||||
|
try:
|
||||||
|
return jsonify(stat_obj.time_analysis()), 200
|
||||||
|
except ValueError as e:
|
||||||
|
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
||||||
|
except Exception as e:
|
||||||
|
print(traceback.format_exc())
|
||||||
|
return jsonify({"error": f"An unexpected error occurred: {str(e)}"}), 500
|
||||||
|
|
||||||
@app.route('/reset', methods=["GET"])
|
@app.route('/reset', methods=["GET"])
|
||||||
def reset_dataset():
|
def reset_dataset():
|
||||||
if stat_obj is None:
|
if stat_obj is None:
|
||||||
|
|||||||
@@ -38,25 +38,35 @@ class StatGen:
|
|||||||
df["weekday"] = df["dt"].dt.day_name()
|
df["weekday"] = df["dt"].dt.day_name()
|
||||||
|
|
||||||
## Public
|
## Public
|
||||||
def get_heatmap(self) -> pd.DataFrame:
|
def time_analysis(self) -> pd.DataFrame:
|
||||||
|
per_day = (
|
||||||
|
self.df.groupby("date")
|
||||||
|
.size()
|
||||||
|
.reset_index(name="count")
|
||||||
|
)
|
||||||
|
|
||||||
weekday_order = [
|
weekday_order = [
|
||||||
"Monday", "Tuesday", "Wednesday",
|
"Monday", "Tuesday", "Wednesday",
|
||||||
"Thursday", "Friday", "Saturday", "Sunday"
|
"Thursday", "Friday", "Saturday", "Sunday"
|
||||||
]
|
]
|
||||||
|
|
||||||
self.df["weekday"] = pd.Categorical(
|
heatmap = (
|
||||||
self.df["weekday"],
|
|
||||||
categories=weekday_order,
|
|
||||||
ordered=True
|
|
||||||
)
|
|
||||||
|
|
||||||
return (
|
|
||||||
self.df
|
self.df
|
||||||
.groupby(["weekday", "hour"], observed=True)
|
.assign(weekday=pd.Categorical(self.df["weekday"], weekday_order, ordered=True))
|
||||||
|
.groupby(["weekday", "hour"])
|
||||||
.size()
|
.size()
|
||||||
.unstack(fill_value=0)
|
.unstack(fill_value=0)
|
||||||
.reindex(columns=range(24), fill_value=0)
|
.reindex(columns=range(24), fill_value=0)
|
||||||
)
|
)
|
||||||
|
heatmap.columns = heatmap.columns.map(str)
|
||||||
|
|
||||||
|
burst_index = per_day["count"].std() / max(per_day["count"].mean(), 1)
|
||||||
|
|
||||||
|
return {
|
||||||
|
"events_per_day": per_day.to_dict(orient="records"),
|
||||||
|
"weekday_hour_heatmap": heatmap.reset_index().to_dict(orient="records"),
|
||||||
|
"burstiness": round(burst_index, 2)
|
||||||
|
}
|
||||||
|
|
||||||
def get_word_frequencies(self, limit: int = 100) -> pd.DataFrame:
|
def get_word_frequencies(self, limit: int = 100) -> pd.DataFrame:
|
||||||
texts = (
|
texts = (
|
||||||
|
|||||||
Reference in New Issue
Block a user