Compare commits
3 Commits
37cb2c9ff4
...
2ae9479943
| Author | SHA1 | Date | |
|---|---|---|---|
| 2ae9479943 | |||
| dd44fad294 | |||
| 5ea71023b5 |
@@ -110,7 +110,7 @@ class PostgresConnector:
|
|||||||
row["source"],
|
row["source"],
|
||||||
row.get("topic"),
|
row.get("topic"),
|
||||||
row.get("topic_confidence"),
|
row.get("topic_confidence"),
|
||||||
Json(row["ner_entities"]) if row.get("ner_entities") else None,
|
Json(row["entities"]) if row.get("entities") else None,
|
||||||
row.get("emotion_anger"),
|
row.get("emotion_anger"),
|
||||||
row.get("emotion_disgust"),
|
row.get("emotion_disgust"),
|
||||||
row.get("emotion_fear"),
|
row.get("emotion_fear"),
|
||||||
|
|||||||
@@ -16,6 +16,7 @@ from server.stat_gen import StatGen
|
|||||||
from server.dataset_processor import DatasetProcessor
|
from server.dataset_processor import DatasetProcessor
|
||||||
from db.database import PostgresConnector
|
from db.database import PostgresConnector
|
||||||
from server.auth import AuthManager
|
from server.auth import AuthManager
|
||||||
|
from server.utils import get_request_filters, parse_datetime_filter
|
||||||
|
|
||||||
import pandas as pd
|
import pandas as pd
|
||||||
import traceback
|
import traceback
|
||||||
@@ -43,56 +44,6 @@ auth_manager = AuthManager(db, bcrypt)
|
|||||||
|
|
||||||
stat_gen = StatGen()
|
stat_gen = StatGen()
|
||||||
|
|
||||||
|
|
||||||
def _parse_datetime_filter(value):
|
|
||||||
if not value:
|
|
||||||
return None
|
|
||||||
|
|
||||||
try:
|
|
||||||
return datetime.datetime.fromisoformat(value)
|
|
||||||
except ValueError:
|
|
||||||
try:
|
|
||||||
return datetime.datetime.fromtimestamp(float(value))
|
|
||||||
except ValueError as err:
|
|
||||||
raise ValueError(
|
|
||||||
"Date filters must be ISO-8601 strings or Unix timestamps"
|
|
||||||
) from err
|
|
||||||
|
|
||||||
|
|
||||||
def _get_request_filters() -> dict:
|
|
||||||
filters = {}
|
|
||||||
|
|
||||||
search_query = request.args.get("search_query") or request.args.get("query")
|
|
||||||
if search_query:
|
|
||||||
filters["search_query"] = search_query
|
|
||||||
|
|
||||||
start_date = _parse_datetime_filter(
|
|
||||||
request.args.get("start_date") or request.args.get("start")
|
|
||||||
)
|
|
||||||
if start_date:
|
|
||||||
filters["start_date"] = start_date
|
|
||||||
|
|
||||||
end_date = _parse_datetime_filter(
|
|
||||||
request.args.get("end_date") or request.args.get("end")
|
|
||||||
)
|
|
||||||
if end_date:
|
|
||||||
filters["end_date"] = end_date
|
|
||||||
|
|
||||||
data_sources = request.args.getlist("data_sources")
|
|
||||||
if not data_sources:
|
|
||||||
data_sources = request.args.getlist("sources")
|
|
||||||
|
|
||||||
if len(data_sources) == 1 and "," in data_sources[0]:
|
|
||||||
data_sources = [
|
|
||||||
source.strip() for source in data_sources[0].split(",") if source.strip()
|
|
||||||
]
|
|
||||||
|
|
||||||
if data_sources:
|
|
||||||
filters["data_sources"] = data_sources
|
|
||||||
|
|
||||||
return filters
|
|
||||||
|
|
||||||
|
|
||||||
@app.route("/register", methods=["POST"])
|
@app.route("/register", methods=["POST"])
|
||||||
def register_user():
|
def register_user():
|
||||||
data = request.get_json()
|
data = request.get_json()
|
||||||
@@ -212,7 +163,8 @@ def get_dataset(dataset_id):
|
|||||||
if dataset_content.empty:
|
if dataset_content.empty:
|
||||||
return jsonify({"error": "Dataset content not found"}), 404
|
return jsonify({"error": "Dataset content not found"}), 404
|
||||||
|
|
||||||
return jsonify(dataset_content.to_dict(orient="records")), 200
|
filters = get_request_filters()
|
||||||
|
return jsonify(stat_gen.filter_dataset(dataset_content, filters)), 200
|
||||||
|
|
||||||
|
|
||||||
@app.route("/dataset/<int:dataset_id>/content", methods=["GET"])
|
@app.route("/dataset/<int:dataset_id>/content", methods=["GET"])
|
||||||
@@ -226,7 +178,7 @@ def content_endpoint(dataset_id):
|
|||||||
|
|
||||||
dataset_content = db.get_dataset_content(dataset_id)
|
dataset_content = db.get_dataset_content(dataset_id)
|
||||||
try:
|
try:
|
||||||
filters = _get_request_filters()
|
filters = get_request_filters()
|
||||||
return jsonify(stat_gen.get_content_analysis(dataset_content, filters)), 200
|
return jsonify(stat_gen.get_content_analysis(dataset_content, filters)), 200
|
||||||
except ValueError as e:
|
except ValueError as e:
|
||||||
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
||||||
@@ -247,7 +199,7 @@ def get_summary(dataset_id):
|
|||||||
dataset_content = db.get_dataset_content(dataset_id)
|
dataset_content = db.get_dataset_content(dataset_id)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
filters = _get_request_filters()
|
filters = get_request_filters()
|
||||||
return jsonify(stat_gen.summary(dataset_content, filters)), 200
|
return jsonify(stat_gen.summary(dataset_content, filters)), 200
|
||||||
except ValueError as e:
|
except ValueError as e:
|
||||||
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
||||||
@@ -268,7 +220,7 @@ def get_time_analysis(dataset_id):
|
|||||||
dataset_content = db.get_dataset_content(dataset_id)
|
dataset_content = db.get_dataset_content(dataset_id)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
filters = _get_request_filters()
|
filters = get_request_filters()
|
||||||
return jsonify(stat_gen.get_time_analysis(dataset_content, filters)), 200
|
return jsonify(stat_gen.get_time_analysis(dataset_content, filters)), 200
|
||||||
except ValueError as e:
|
except ValueError as e:
|
||||||
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
||||||
@@ -289,7 +241,7 @@ def get_user_analysis(dataset_id):
|
|||||||
dataset_content = db.get_dataset_content(dataset_id)
|
dataset_content = db.get_dataset_content(dataset_id)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
filters = _get_request_filters()
|
filters = get_request_filters()
|
||||||
return jsonify(stat_gen.get_user_analysis(dataset_content, filters)), 200
|
return jsonify(stat_gen.get_user_analysis(dataset_content, filters)), 200
|
||||||
except ValueError as e:
|
except ValueError as e:
|
||||||
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
||||||
@@ -310,7 +262,7 @@ def get_cultural_analysis(dataset_id):
|
|||||||
dataset_content = db.get_dataset_content(dataset_id)
|
dataset_content = db.get_dataset_content(dataset_id)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
filters = _get_request_filters()
|
filters = get_request_filters()
|
||||||
return jsonify(stat_gen.get_cultural_analysis(dataset_content, filters)), 200
|
return jsonify(stat_gen.get_cultural_analysis(dataset_content, filters)), 200
|
||||||
except ValueError as e:
|
except ValueError as e:
|
||||||
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
return jsonify({"error": f"Malformed or missing data: {str(e)}"}), 400
|
||||||
@@ -331,7 +283,7 @@ def get_interaction_analysis(dataset_id):
|
|||||||
dataset_content = db.get_dataset_content(dataset_id)
|
dataset_content = db.get_dataset_content(dataset_id)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
filters = _get_request_filters()
|
filters = get_request_filters()
|
||||||
return jsonify(
|
return jsonify(
|
||||||
stat_gen.get_interactional_analysis(dataset_content, filters)
|
stat_gen.get_interactional_analysis(dataset_content, filters)
|
||||||
), 200
|
), 200
|
||||||
|
|||||||
@@ -55,13 +55,15 @@ class StatGen:
|
|||||||
if search_query:
|
if search_query:
|
||||||
mask = (
|
mask = (
|
||||||
filtered_df["content"].str.contains(search_query, case=False, na=False)
|
filtered_df["content"].str.contains(search_query, case=False, na=False)
|
||||||
| filtered_df["author"]
|
| filtered_df["author"].str.contains(search_query, case=False, na=False)
|
||||||
.str.contains(search_query, case=False, na=False)
|
|
||||||
.fillna(False)
|
|
||||||
| filtered_df["title"]
|
|
||||||
.str.contains(search_query, case=False, na=False, regex=False)
|
|
||||||
.fillna(False)
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Only include title if the column exists
|
||||||
|
if "title" in filtered_df.columns:
|
||||||
|
mask = mask | filtered_df["title"].str.contains(
|
||||||
|
search_query, case=False, na=False, regex=False
|
||||||
|
)
|
||||||
|
|
||||||
filtered_df = filtered_df[mask]
|
filtered_df = filtered_df[mask]
|
||||||
|
|
||||||
if start_date_filter:
|
if start_date_filter:
|
||||||
@@ -76,6 +78,8 @@ class StatGen:
|
|||||||
return filtered_df
|
return filtered_df
|
||||||
|
|
||||||
## Public Methods
|
## Public Methods
|
||||||
|
def filter_dataset(self, df: pd.DataFrame, filters: dict | None = None) -> dict:
|
||||||
|
return self._prepare_filtered_df(df, filters).to_dict(orient="records")
|
||||||
|
|
||||||
def get_time_analysis(self, df: pd.DataFrame, filters: dict | None = None) -> dict:
|
def get_time_analysis(self, df: pd.DataFrame, filters: dict | None = None) -> dict:
|
||||||
filtered_df = self._prepare_filtered_df(df, filters)
|
filtered_df = self._prepare_filtered_df(df, filters)
|
||||||
|
|||||||
50
server/utils.py
Normal file
50
server/utils.py
Normal file
@@ -0,0 +1,50 @@
|
|||||||
|
import datetime
|
||||||
|
from flask import request
|
||||||
|
|
||||||
|
def parse_datetime_filter(value):
|
||||||
|
if not value:
|
||||||
|
return None
|
||||||
|
|
||||||
|
try:
|
||||||
|
return datetime.datetime.fromisoformat(value)
|
||||||
|
except ValueError:
|
||||||
|
try:
|
||||||
|
return datetime.datetime.fromtimestamp(float(value))
|
||||||
|
except ValueError as err:
|
||||||
|
raise ValueError(
|
||||||
|
"Date filters must be ISO-8601 strings or Unix timestamps"
|
||||||
|
) from err
|
||||||
|
|
||||||
|
|
||||||
|
def get_request_filters() -> dict:
|
||||||
|
filters = {}
|
||||||
|
|
||||||
|
search_query = request.args.get("search_query") or request.args.get("query")
|
||||||
|
if search_query:
|
||||||
|
filters["search_query"] = search_query
|
||||||
|
|
||||||
|
start_date = parse_datetime_filter(
|
||||||
|
request.args.get("start_date") or request.args.get("start")
|
||||||
|
)
|
||||||
|
if start_date:
|
||||||
|
filters["start_date"] = start_date
|
||||||
|
|
||||||
|
end_date = parse_datetime_filter(
|
||||||
|
request.args.get("end_date") or request.args.get("end")
|
||||||
|
)
|
||||||
|
if end_date:
|
||||||
|
filters["end_date"] = end_date
|
||||||
|
|
||||||
|
data_sources = request.args.getlist("data_sources")
|
||||||
|
if not data_sources:
|
||||||
|
data_sources = request.args.getlist("sources")
|
||||||
|
|
||||||
|
if len(data_sources) == 1 and "," in data_sources[0]:
|
||||||
|
data_sources = [
|
||||||
|
source.strip() for source in data_sources[0].split(",") if source.strip()
|
||||||
|
]
|
||||||
|
|
||||||
|
if data_sources:
|
||||||
|
filters["data_sources"] = data_sources
|
||||||
|
|
||||||
|
return filters
|
||||||
Reference in New Issue
Block a user