Move API to python.
This commit is contained in:
parent
32e86be1d1
commit
da33d814ae
|
@ -1,31 +1,13 @@
|
||||||
FROM node:15.14-buster
|
FROM python:3.9.7-bullseye
|
||||||
|
|
||||||
# Install python3, pip3, and make them the default for `python` and `pip` commands
|
|
||||||
RUN apt-get update && apt-get install -y python3 python3-pip
|
|
||||||
RUN ln -s $(which python3) /usr/local/bin/python
|
|
||||||
RUN ln -s $(which pip3) /usr/local/bin/pip
|
|
||||||
|
|
||||||
WORKDIR /opt/obs/api
|
WORKDIR /opt/obs/api
|
||||||
|
|
||||||
ADD package.json package-lock.json /opt/obs/api/
|
ADD scripts /opt/obs/scripts
|
||||||
RUN echo update-notifier=false >> ~/.npmrc
|
RUN pip install -e /opt/obs/scripts
|
||||||
RUN npm ci
|
|
||||||
|
|
||||||
ADD scripts /opt/obs/api/scripts/
|
ADD requirements.txt setup.py obs /opt/obs/api/
|
||||||
ADD tools /opt/obs/api/tools/
|
RUN pip install -e .
|
||||||
ADD requirements.txt /opt/obs/api/
|
|
||||||
RUN pip install -r requirements.txt
|
|
||||||
RUN pip install -e ./scripts
|
|
||||||
|
|
||||||
ADD views /opt/obs/api/views/
|
EXPOSE 8000
|
||||||
ADD src /opt/obs/api/src/
|
|
||||||
|
|
||||||
#ADD .migrations.js .
|
CMD ["obs-api"]
|
||||||
#ADD migrations .
|
|
||||||
|
|
||||||
EXPOSE 3000
|
|
||||||
|
|
||||||
ENV PORT=3000
|
|
||||||
ENV DATA_DIR=/data
|
|
||||||
|
|
||||||
CMD ["npm", "run", "start"]
|
|
||||||
|
|
|
@ -1,34 +0,0 @@
|
||||||
{
|
|
||||||
"cookieSecret": "CHANGEME!!!!!!!!!!@##@!!$$$$$$$$$$$$$!!",
|
|
||||||
"jwtSecret": "CHANGEME??????????????////3212321;312kjbkasjd",
|
|
||||||
"baseUrl": "http://localhost:3000/",
|
|
||||||
"mainFrontendUrl": "http://localhost:3001/",
|
|
||||||
"mail": false,
|
|
||||||
"mongodb": {
|
|
||||||
"url": "mongodb://mongo/obsTest",
|
|
||||||
"debug": false
|
|
||||||
},
|
|
||||||
"postgres": {
|
|
||||||
"url": "postgresql+asyncpg://obs:obs@postgres/obs"
|
|
||||||
},
|
|
||||||
"redisUrl": "redis://redis",
|
|
||||||
"oAuth2Clients": [
|
|
||||||
{
|
|
||||||
"clientId": "b730f8d2-d93c-4c68-9ff0-dfac8da76ee2",
|
|
||||||
"validRedirectUris": ["http://localhost:3001/redirect"],
|
|
||||||
"refreshTokenExpirySeconds": 604800,
|
|
||||||
"maxScope": "*",
|
|
||||||
"title": "OBS Portal",
|
|
||||||
"autoAccept": true
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"clientId": "a2958209-4045-4ec9-8cb3-1156abba7de3",
|
|
||||||
"validRedirectUris": ["__LOCAL__"],
|
|
||||||
"maxScope": "track.upload",
|
|
||||||
"refreshTokenExpirySeconds": 86400000,
|
|
||||||
"title": "OpenBikeSensor"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"imprintUrl": "https://example.com/imprint",
|
|
||||||
"privacyPolicyUrl": "https://example.com/privacy"
|
|
||||||
}
|
|
39
api/config.dev.py
Normal file
39
api/config.dev.py
Normal file
|
@ -0,0 +1,39 @@
|
||||||
|
# Bind address of the server
|
||||||
|
HOST = "0.0.0.0"
|
||||||
|
PORT = 3000
|
||||||
|
|
||||||
|
# Extended log output, but slower
|
||||||
|
DEBUG = True
|
||||||
|
|
||||||
|
# Required to encrypt or sign sessions, cookies, tokens, etc.
|
||||||
|
SECRET = "CHANGEME!!!!!!!!!!@##@!!$$$$$$$$$$$$$!!"
|
||||||
|
|
||||||
|
# Connection to the database
|
||||||
|
POSTGRES_URL = "postgresql+asyncpg://obs:obs@postgres/obs"
|
||||||
|
|
||||||
|
# URL to the keycloak realm, as reachable by the API service. This is not
|
||||||
|
# necessarily its publicly reachable URL, keycloak advertises that iself.
|
||||||
|
KEYCLOAK_URL = "http://keycloak:8080/auth/realms/OBS%20Dev/"
|
||||||
|
|
||||||
|
# Auth client credentials
|
||||||
|
KEYCLOAK_CLIENT_ID = "portal"
|
||||||
|
KEYCLOAK_CLIENT_SECRET = "76b84224-dc24-4824-bb98-9e1ba15bd58f"
|
||||||
|
|
||||||
|
# The root of the frontend. Needed for redirecting after login, and for CORS.
|
||||||
|
MAIN_FRONTEND_URL = "http://localhost:3001/"
|
||||||
|
|
||||||
|
# Mail settings
|
||||||
|
MAIL_ENABLED = False
|
||||||
|
|
||||||
|
# Urls to important documents, hosted elsewhere
|
||||||
|
IMPRINT_URL = "https://example.com/imprint"
|
||||||
|
PRIVACY_POLICY_URL = "https://example.com/privacy"
|
||||||
|
|
||||||
|
# API_ROOT_DIR = "??" # default: api/ inside repository
|
||||||
|
DATA_DIR = "/data"
|
||||||
|
# PROCESSING_DIR = "??" # default: DATA_DIR/processing
|
||||||
|
# PROCESSING_OUTPUT_DIR = "??" # default: DATA_DIR/processing-output
|
||||||
|
# TRACKS_DIR = "??" # default: DATA_DIR/tracks
|
||||||
|
# OBS_FACE_CACHE_DIR = "??" # default: DATA_DIR/obs-face-cache
|
||||||
|
|
||||||
|
# vim: set ft=python :
|
|
@ -1,42 +0,0 @@
|
||||||
{
|
|
||||||
"cookieSecret": "!!!<<<CHANGEME>>>!!!",
|
|
||||||
"jwtSecret": "!!!<<<CHANGEME>>>!!!",
|
|
||||||
"baseUrl": "https://portal.example.com/api",
|
|
||||||
"mainFrontendUrl": "https://portal.example.com/",
|
|
||||||
"mail": {
|
|
||||||
"from": "Sender Name <sender@example.com>",
|
|
||||||
"smtp" : {
|
|
||||||
"host": "mail.example.com",
|
|
||||||
"port": 465,
|
|
||||||
"starttls": false,
|
|
||||||
"username": "sender@example.com",
|
|
||||||
"password": "!!!<<<CHANGEME>>>!!!"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"mongodb": {
|
|
||||||
"url": "mongodb://mongo/obs",
|
|
||||||
"debug": false
|
|
||||||
},
|
|
||||||
"postgres": {
|
|
||||||
"url": "postgresql+asyncpg://user:pass@host/dbname"
|
|
||||||
},
|
|
||||||
"redisUrl": "redis://redis",
|
|
||||||
"oAuth2Clients": [
|
|
||||||
{
|
|
||||||
"clientId": "!!!<<<CHANGEME>>>!!!",
|
|
||||||
"validRedirectUris": ["https://portal.example.com/redirect"],
|
|
||||||
"refreshTokenExpirySeconds": 604800,
|
|
||||||
"maxScope": "*",
|
|
||||||
"title": "OBS Portal"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"clientId": "!!!<<<CHANGEME>>>!!!",
|
|
||||||
"validRedirectUris": ["__LOCAL__"],
|
|
||||||
"maxScope": "track.upload",
|
|
||||||
"refreshTokenExpirySeconds": 86400000,
|
|
||||||
"title": "OpenBikeSensor"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"imprintUrl": "https://example.com/imprint",
|
|
||||||
"privacyPolicyUrl": "https://example.com/privacy"
|
|
||||||
}
|
|
45
api/config.py.example
Normal file
45
api/config.py.example
Normal file
|
@ -0,0 +1,45 @@
|
||||||
|
# Bind address of the server
|
||||||
|
HOST = "127.0.0.1"
|
||||||
|
PORT = 3000
|
||||||
|
|
||||||
|
# Extended log output, but slower
|
||||||
|
DEBUG = False
|
||||||
|
|
||||||
|
# Required to encrypt or sign sessions, cookies, tokens, etc.
|
||||||
|
SECRET = "!!!<<<CHANGEME>>>!!!"
|
||||||
|
|
||||||
|
# Connection to the database
|
||||||
|
POSTGRES_URL = "postgresql+asyncpg://user:pass@host/dbname"
|
||||||
|
|
||||||
|
# URL to the keycloak realm, as reachable by the API service. This is not
|
||||||
|
# necessarily its publicly reachable URL, keycloak advertises that iself.
|
||||||
|
KEYCLOAK_URL = "http://localhost:1234/auth/realms/obs/"
|
||||||
|
|
||||||
|
# Auth client credentials
|
||||||
|
KEYCLOAK_CLIENT_ID = "portal"
|
||||||
|
KEYCLOAK_CLIENT_SECRET = "00000000-0000-0000-0000-000000000000"
|
||||||
|
|
||||||
|
# The root of the frontend. Needed for redirecting after login, and for CORS.
|
||||||
|
MAIN_FRONTEND_URL = "https://portal.example.com/"
|
||||||
|
|
||||||
|
# Mail settings
|
||||||
|
MAIL_ENABLED = False
|
||||||
|
MAIL_FROM = "Sender Name <sender@example.com>"
|
||||||
|
MAIL_SMTP_HOST = "mail.example.com"
|
||||||
|
MAIL_SMTP_PORT = 465
|
||||||
|
MAIL_STARTTLS = False
|
||||||
|
MAIL_SMTP_USERNAME = "sender@example.com"
|
||||||
|
|
||||||
|
# Urls to important documents, hosted elsewhere
|
||||||
|
IMPRINT_URL = "https://example.com/imprint"
|
||||||
|
PRIVACY_POLICY_URL = "https://example.com/privacy"
|
||||||
|
|
||||||
|
# Path overrides:
|
||||||
|
# API_ROOT_DIR = "??" # default: api/ inside repository
|
||||||
|
# DATA_DIR = "??" # default: $API_ROOT_DIR/..
|
||||||
|
# PROCESSING_DIR = "??" # default: DATA_DIR/processing
|
||||||
|
# PROCESSING_OUTPUT_DIR = "??" # default: DATA_DIR/processing-output
|
||||||
|
# TRACKS_DIR = "??" # default: DATA_DIR/tracks
|
||||||
|
# OBS_FACE_CACHE_DIR = "??" # default: DATA_DIR/obs-face-cache
|
||||||
|
|
||||||
|
# vim: set ft=python :
|
1
api/obs/__init__.py
Normal file
1
api/obs/__init__.py
Normal file
|
@ -0,0 +1 @@
|
||||||
|
__path__ = __import__("pkgutil").extend_path(__path__, __name__)
|
1
api/obs/api/__init__.py
Normal file
1
api/obs/api/__init__.py
Normal file
|
@ -0,0 +1 @@
|
||||||
|
__version__ = "0.2.0"
|
118
api/obs/api/app.py
Normal file
118
api/obs/api/app.py
Normal file
|
@ -0,0 +1,118 @@
|
||||||
|
import logging
|
||||||
|
import os
|
||||||
|
from json import JSONEncoder, dumps
|
||||||
|
from functools import wraps, partial
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
from os.path import dirname, join, normpath, abspath
|
||||||
|
from datetime import datetime, date
|
||||||
|
|
||||||
|
from sanic import Sanic
|
||||||
|
from sanic.response import text, json as json_response
|
||||||
|
from sanic.exceptions import Unauthorized
|
||||||
|
from sanic_session import Session, InMemorySessionInterface
|
||||||
|
from sanic_cors import CORS
|
||||||
|
|
||||||
|
from sqlalchemy import select
|
||||||
|
from sqlalchemy.ext.asyncio import AsyncSession, create_async_engine
|
||||||
|
from sqlalchemy.orm import sessionmaker
|
||||||
|
|
||||||
|
from obs.api.db import User, async_session
|
||||||
|
|
||||||
|
from sanic_session.base import BaseSessionInterface
|
||||||
|
from sanic_session.utils import ExpiringDict
|
||||||
|
|
||||||
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
app = Sanic("OpenBikeSensor Portal API")
|
||||||
|
app.update_config("./config.py")
|
||||||
|
c = app.config
|
||||||
|
|
||||||
|
# Configure paths
|
||||||
|
c.API_ROOT_DIR = c.get("API_ROOT_DIR") or abspath(join(dirname(__file__), "..", ".."))
|
||||||
|
c.DATA_DIR = c.get("DATA_DIR") or normpath(join(c.API_ROOT_DIR, "../data"))
|
||||||
|
c.PROCESSING_DIR = c.get("PROCESSING_DIR") or join(c.DATA_DIR, "processing")
|
||||||
|
c.PROCESSING_OUTPUT_DIR = c.get("PROCESSING_OUTPUT_DIR") or join(
|
||||||
|
c.DATA_DIR, "processing-output"
|
||||||
|
)
|
||||||
|
c.TRACKS_DIR = c.get("TRACKS_DIR") or join(c.DATA_DIR, "tracks")
|
||||||
|
c.OBS_FACE_CACHE_DIR = c.get("OBS_FACE_CACHE_DIR") or join(c.DATA_DIR, "obs-face-cache")
|
||||||
|
|
||||||
|
main_frontend_url = urlparse(c.MAIN_FRONTEND_URL)
|
||||||
|
CORS(
|
||||||
|
app,
|
||||||
|
origins=[f"{main_frontend_url.scheme}://{main_frontend_url.netloc}"],
|
||||||
|
supports_credentials=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
# TODO: use a different interface, maybe backed by the PostgreSQL, to allow
|
||||||
|
# scaling the API
|
||||||
|
Session(app, interface=InMemorySessionInterface())
|
||||||
|
|
||||||
|
|
||||||
|
@app.before_server_start
|
||||||
|
async def app_connect_db(app, loop):
|
||||||
|
app.ctx._db_engine = create_async_engine(c.POSTGRES_URL, echo=c.DEBUG)
|
||||||
|
|
||||||
|
|
||||||
|
@app.after_server_stop
|
||||||
|
async def app_disconnect_db(app, loop):
|
||||||
|
if app.ctx._db_engine:
|
||||||
|
await app.ctx._db_engine.dispose()
|
||||||
|
|
||||||
|
|
||||||
|
@app.middleware("request")
|
||||||
|
async def inject_session(req):
|
||||||
|
req.ctx.db = sessionmaker(
|
||||||
|
req.app.ctx._db_engine, class_=AsyncSession, expire_on_commit=False
|
||||||
|
)()
|
||||||
|
req.ctx._db_session_ctx_token = async_session.set(req.ctx.db)
|
||||||
|
|
||||||
|
|
||||||
|
@app.middleware("response")
|
||||||
|
async def close_session(req, response):
|
||||||
|
if hasattr(req.ctx, "_db_session_ctx_token"):
|
||||||
|
async_session.reset(req.ctx._db_session_ctx_token)
|
||||||
|
await req.ctx.db.close()
|
||||||
|
|
||||||
|
|
||||||
|
@app.middleware("request")
|
||||||
|
async def load_user(req):
|
||||||
|
user_id = req.ctx.session.get("user_id")
|
||||||
|
user = None
|
||||||
|
if user_id:
|
||||||
|
user = (
|
||||||
|
await req.ctx.db.execute(select(User).where(User.id == user_id))
|
||||||
|
).scalar()
|
||||||
|
|
||||||
|
req.ctx.user = user
|
||||||
|
|
||||||
|
|
||||||
|
@app.route("/")
|
||||||
|
def index(req):
|
||||||
|
return text("Hello, %s!" % (req.ctx.user.username if req.ctx.user else "World"))
|
||||||
|
|
||||||
|
|
||||||
|
def require_auth(fn):
|
||||||
|
@wraps(fn)
|
||||||
|
def wrapper(req, *args, **kwargs):
|
||||||
|
if not req.ctx.user:
|
||||||
|
raise Unauthorized("Login required")
|
||||||
|
return fn(req, *args, **kwargs)
|
||||||
|
|
||||||
|
return wrapper
|
||||||
|
|
||||||
|
|
||||||
|
class CustomJsonEncoder(JSONEncoder):
|
||||||
|
def default(self, obj):
|
||||||
|
if isinstance(obj, (datetime, date)):
|
||||||
|
return obj.isoformat() + "+0000" # explicit UTC for javascript <3
|
||||||
|
|
||||||
|
# Let the base class default method raise the TypeError
|
||||||
|
return super().default(obj)
|
||||||
|
|
||||||
|
|
||||||
|
def json(*args, **kwargs):
|
||||||
|
return json_response(*args, **kwargs, dumps=partial(dumps, cls=CustomJsonEncoder))
|
||||||
|
|
||||||
|
|
||||||
|
from . import routes
|
17
api/obs/api/config.py
Normal file
17
api/obs/api/config.py
Normal file
|
@ -0,0 +1,17 @@
|
||||||
|
# Configure paths
|
||||||
|
config.API_ROOT_DIR = config.get("API_ROOT_DIR") or abspath(
|
||||||
|
join(dirname(__file__), "..", "..")
|
||||||
|
)
|
||||||
|
config.DATA_DIR = config.get("DATA_DIR") or normpath(
|
||||||
|
join(config.API_ROOT_DIR, "../data")
|
||||||
|
)
|
||||||
|
config.PROCESSING_DIR = config.get("PROCESSING_DIR") or join(
|
||||||
|
config.DATA_DIR, "processing"
|
||||||
|
)
|
||||||
|
config.PROCESSING_OUTPUT_DIR = config.get("PROCESSING_OUTPUT_DIR") or join(
|
||||||
|
config.DATA_DIR, "processing-output"
|
||||||
|
)
|
||||||
|
config.TRACKS_DIR = config.get("TRACKS_DIR") or join(config.DATA_DIR, "tracks")
|
||||||
|
config.OBS_FACE_CACHE_DIR = config.get("OBS_FACE_CACHE_DIR") or join(
|
||||||
|
config.DATA_DIR, "obs-face-cache"
|
||||||
|
)
|
395
api/obs/api/db.py
Normal file
395
api/obs/api/db.py
Normal file
|
@ -0,0 +1,395 @@
|
||||||
|
import hashlib
|
||||||
|
from contextvars import ContextVar
|
||||||
|
from contextlib import asynccontextmanager
|
||||||
|
from datetime import datetime
|
||||||
|
import os
|
||||||
|
from os.path import join, dirname
|
||||||
|
import re
|
||||||
|
import math
|
||||||
|
import aiofiles
|
||||||
|
import random
|
||||||
|
import string
|
||||||
|
from slugify import slugify
|
||||||
|
|
||||||
|
from sqlalchemy.ext.declarative import declarative_base
|
||||||
|
from sqlalchemy.ext.asyncio import AsyncSession
|
||||||
|
from sqlalchemy.ext.asyncio import create_async_engine
|
||||||
|
from sqlalchemy.orm import sessionmaker, relationship
|
||||||
|
from sqlalchemy.types import UserDefinedType, BIGINT, TEXT
|
||||||
|
from sqlalchemy import (
|
||||||
|
Boolean,
|
||||||
|
Column,
|
||||||
|
DateTime,
|
||||||
|
Enum as SqlEnum,
|
||||||
|
Float,
|
||||||
|
ForeignKey,
|
||||||
|
Index,
|
||||||
|
Integer,
|
||||||
|
String,
|
||||||
|
false,
|
||||||
|
func,
|
||||||
|
select,
|
||||||
|
text,
|
||||||
|
literal,
|
||||||
|
)
|
||||||
|
from sqlalchemy.dialects.postgresql import HSTORE, UUID
|
||||||
|
|
||||||
|
|
||||||
|
Base = declarative_base()
|
||||||
|
|
||||||
|
|
||||||
|
engine = ContextVar("engine")
|
||||||
|
async_session = ContextVar("async_session")
|
||||||
|
|
||||||
|
|
||||||
|
@asynccontextmanager
|
||||||
|
async def make_session():
|
||||||
|
async with async_session.get()() as session:
|
||||||
|
yield session
|
||||||
|
|
||||||
|
|
||||||
|
async def init_models():
|
||||||
|
async with engine.get().begin() as conn:
|
||||||
|
await conn.run_sync(Base.metadata.drop_all)
|
||||||
|
await conn.execute(text('CREATE EXTENSION IF NOT EXISTS "hstore";'))
|
||||||
|
await conn.execute(text('CREATE EXTENSION IF NOT EXISTS "postgis";'))
|
||||||
|
await conn.execute(text('CREATE EXTENSION IF NOT EXISTS "uuid-ossp";'))
|
||||||
|
await conn.run_sync(Base.metadata.create_all)
|
||||||
|
|
||||||
|
|
||||||
|
def random_string(length):
|
||||||
|
letters = string.ascii_lowercase + string.digits
|
||||||
|
return "".join(random.choice(letters) for i in range(length))
|
||||||
|
|
||||||
|
|
||||||
|
@asynccontextmanager
|
||||||
|
async def connect_db(url):
|
||||||
|
engine_ = create_async_engine(url, echo=False)
|
||||||
|
t1 = engine.set(engine_)
|
||||||
|
|
||||||
|
async_session_ = sessionmaker(engine_, class_=AsyncSession, expire_on_commit=False)
|
||||||
|
t2 = async_session.set(async_session_)
|
||||||
|
|
||||||
|
yield
|
||||||
|
|
||||||
|
# for AsyncEngine created in function scope, close and
|
||||||
|
# clean-up pooled connections
|
||||||
|
await engine_.dispose()
|
||||||
|
engine.reset(t1)
|
||||||
|
async_session.reset(t2)
|
||||||
|
|
||||||
|
|
||||||
|
ZoneType = SqlEnum("rural", "urban", "motorway", name="zone_type")
|
||||||
|
ProcessingStatus = SqlEnum(
|
||||||
|
"created", "queued", "processing", "complete", "error", name="processing_status"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class Geometry(UserDefinedType):
|
||||||
|
def get_col_spec(self):
|
||||||
|
return "GEOMETRY"
|
||||||
|
|
||||||
|
def bind_expression(self, bindvalue):
|
||||||
|
return func.ST_GeomFromGeoJSON(bindvalue, type_=self)
|
||||||
|
|
||||||
|
def column_expression(self, col):
|
||||||
|
return func.ST_AsGeoJSON(col, type_=self)
|
||||||
|
|
||||||
|
|
||||||
|
class OvertakingEvent(Base):
|
||||||
|
__tablename__ = "overtaking_event"
|
||||||
|
__table_args__ = (Index("road_segment", "way_id", "direction_reversed"),)
|
||||||
|
|
||||||
|
id = Column(Integer, autoincrement=True, primary_key=True, index=True)
|
||||||
|
track_id = Column(Integer, ForeignKey("track.id", ondelete="CASCADE"))
|
||||||
|
hex_hash = Column(String, unique=True, index=True)
|
||||||
|
way_id = Column(BIGINT, index=True)
|
||||||
|
|
||||||
|
# whether we were traveling along the way in reverse direction
|
||||||
|
direction_reversed = Column(Boolean)
|
||||||
|
|
||||||
|
geometry = Column(Geometry)
|
||||||
|
latitude = Column(Float)
|
||||||
|
longitude = Column(Float)
|
||||||
|
time = Column(DateTime)
|
||||||
|
distance_overtaker = Column(Float)
|
||||||
|
distance_stationary = Column(Float)
|
||||||
|
course = Column(Float)
|
||||||
|
speed = Column(Float)
|
||||||
|
|
||||||
|
def __repr__(self):
|
||||||
|
return f"<OvertakingEvent {self.id}>"
|
||||||
|
|
||||||
|
|
||||||
|
class Road(Base):
|
||||||
|
__tablename__ = "road"
|
||||||
|
way_id = Column(BIGINT, primary_key=True, index=True)
|
||||||
|
zone = Column(ZoneType)
|
||||||
|
name = Column(String)
|
||||||
|
geometry = Column(Geometry)
|
||||||
|
tags = Column(HSTORE)
|
||||||
|
directionality = Column(Integer)
|
||||||
|
|
||||||
|
|
||||||
|
NOW = text("NOW()")
|
||||||
|
|
||||||
|
|
||||||
|
class Track(Base):
|
||||||
|
__tablename__ = "track"
|
||||||
|
id = Column(Integer, primary_key=True, autoincrement=True)
|
||||||
|
slug = Column(String, unique=True, nullable=False, index=True)
|
||||||
|
|
||||||
|
created_at = Column(DateTime, nullable=False, server_default=NOW)
|
||||||
|
updated_at = Column(DateTime, nullable=False, server_default=NOW, onupdate=NOW)
|
||||||
|
|
||||||
|
title = Column(String)
|
||||||
|
|
||||||
|
processing_status = Column(ProcessingStatus, server_default=literal("created"))
|
||||||
|
processing_queued_at = Column(DateTime)
|
||||||
|
processed_at = Column(DateTime)
|
||||||
|
|
||||||
|
processing_log = Column(TEXT)
|
||||||
|
|
||||||
|
# Set to true if the user customized the title. Disables auto-generating
|
||||||
|
# an updated title when the track is (re-)processed.
|
||||||
|
customized_title = Column(Boolean, server_default=false(), nullable=False)
|
||||||
|
|
||||||
|
# A user-provided description of the track. May contain markdown.
|
||||||
|
description = Column(TEXT)
|
||||||
|
|
||||||
|
# Whether this track is visible (anonymized) in the public track list or not.
|
||||||
|
public = Column(Boolean, server_default=false())
|
||||||
|
|
||||||
|
# Whether this track should be exported to the public track database
|
||||||
|
# (after anonymization).
|
||||||
|
# include_in_public_database = Column(Boolean, server_default=false())
|
||||||
|
|
||||||
|
# The user agent string, or a part thereof, that was used to upload this
|
||||||
|
# track. Usually contains only the OBS version, other user agents are
|
||||||
|
# discarded due to being irrelevant.
|
||||||
|
uploaded_by_user_agent = Column(String)
|
||||||
|
|
||||||
|
# The name of the original file, as provided during upload. Used for
|
||||||
|
# providing a download with the same name, and for display in the
|
||||||
|
# frontend.
|
||||||
|
original_file_name = Column(String)
|
||||||
|
|
||||||
|
# A hash of the original file's contents. Nobody can upload the same track twice.
|
||||||
|
original_file_hash = Column(String, nullable=False)
|
||||||
|
|
||||||
|
author_id = Column(
|
||||||
|
Integer, ForeignKey("user.id", ondelete="CASCADE"), nullable=False
|
||||||
|
)
|
||||||
|
|
||||||
|
# Statistics... maybe we'll drop some of this if we can easily compute them from SQL
|
||||||
|
recorded_at = Column(DateTime)
|
||||||
|
recorded_until = Column(DateTime)
|
||||||
|
duration = Column(Float)
|
||||||
|
length = Column(Float)
|
||||||
|
segments = Column(Integer)
|
||||||
|
num_events = Column(Integer)
|
||||||
|
num_measurements = Column(Integer)
|
||||||
|
num_valid = Column(Integer)
|
||||||
|
|
||||||
|
def to_dict(self, for_user_id=None):
|
||||||
|
result = {
|
||||||
|
"id": self.id,
|
||||||
|
"slug": self.slug,
|
||||||
|
"title": self.title,
|
||||||
|
"description": self.description,
|
||||||
|
"createdAt": self.created_at,
|
||||||
|
"updatedAt": self.updated_at,
|
||||||
|
"public": self.public,
|
||||||
|
"processingStatus": self.processing_status,
|
||||||
|
"recordedAt": self.recorded_at,
|
||||||
|
"recordedUntil": self.recorded_until,
|
||||||
|
"duration": self.duration,
|
||||||
|
"length": self.length,
|
||||||
|
"numEvents": self.num_events,
|
||||||
|
"numValid": self.num_valid,
|
||||||
|
"numMeasurements": self.num_measurements,
|
||||||
|
}
|
||||||
|
|
||||||
|
if for_user_id is not None and for_user_id == self.author_id:
|
||||||
|
result["uploadedByUserAgent"] = self.uploaded_by_user_agent
|
||||||
|
result["originalFileName"] = self.original_file_name
|
||||||
|
|
||||||
|
if self.author:
|
||||||
|
result["author"] = self.author.to_dict(for_user_id=for_user_id)
|
||||||
|
|
||||||
|
return result
|
||||||
|
|
||||||
|
def is_visible_to_private(self, user):
|
||||||
|
return user is not None and user.id == self.author_id
|
||||||
|
|
||||||
|
def is_visible_to(self, user):
|
||||||
|
return self.is_visible_to_private(user) or self.public
|
||||||
|
|
||||||
|
def generate_slug(self, new_title_or_filename=None):
|
||||||
|
input_text = (
|
||||||
|
new_title_or_filename or self.title or self.original_file_name or "track"
|
||||||
|
)
|
||||||
|
self.slug = slugify(input_text, separator="_") + "-" + random_string(6)
|
||||||
|
|
||||||
|
async def prevent_duplicates(self, session, file_body):
|
||||||
|
hex_hash = hashlib.sha512(file_body).hexdigest()
|
||||||
|
|
||||||
|
duplicate_count = await session.scalar(
|
||||||
|
select(func.count())
|
||||||
|
.select_from(Track)
|
||||||
|
.where(
|
||||||
|
Track.original_file_hash == hex_hash
|
||||||
|
and Track.author_id == self.author_id
|
||||||
|
and Track.id != self.id
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
if duplicate_count:
|
||||||
|
raise ValueError("duplicate file")
|
||||||
|
|
||||||
|
self.original_file_hash = hex_hash
|
||||||
|
|
||||||
|
async def write_to_original_file(self, config, body):
|
||||||
|
mode = "wb" if isinstance(body, bytes) else "wt"
|
||||||
|
|
||||||
|
target = self.get_original_file_path(config)
|
||||||
|
os.makedirs(dirname(target), exist_ok=True)
|
||||||
|
async with aiofiles.open(target, mode=mode) as f:
|
||||||
|
await f.write(body)
|
||||||
|
|
||||||
|
def queue_processing(self):
|
||||||
|
self.processing_status = "queued"
|
||||||
|
self.processing_queued_at = datetime.utcnow()
|
||||||
|
|
||||||
|
def auto_generate_title(self):
|
||||||
|
if self.customized_title:
|
||||||
|
return
|
||||||
|
|
||||||
|
# Try to figure out when this file was recorded. Either we have it in then
|
||||||
|
# statistics, e.g. after parsing and processing the track, or we can maybe
|
||||||
|
# derive it from the filename.
|
||||||
|
recorded_at = self.recorded_at
|
||||||
|
|
||||||
|
if not recorded_at and self.original_file_name:
|
||||||
|
match = re.match(
|
||||||
|
r"^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}\.[0-9]{2}\.[0-9]{2}",
|
||||||
|
self.original_file_name,
|
||||||
|
)
|
||||||
|
if match:
|
||||||
|
try:
|
||||||
|
recorded_at = datetime.fromisoformat(match[0])
|
||||||
|
except ValueError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
if recorded_at:
|
||||||
|
daytime = _get_daytime(recorded_at)
|
||||||
|
self.title = f"{daytime} ride on {recorded_at.strftime('%a, %x')}"
|
||||||
|
return
|
||||||
|
|
||||||
|
# Detecting recording date failed, use filename
|
||||||
|
if self.original_file_name:
|
||||||
|
words = self.original_file_name
|
||||||
|
words = re.sub(r"(\.obsdata)?\.csv$", "", words)
|
||||||
|
words = re.split(r"\W+", words)
|
||||||
|
words[0] = words[0][0].upper() + words[0][1:]
|
||||||
|
self.title = " ".join(words)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def file_path(self):
|
||||||
|
return join(self.author.username, self.slug)
|
||||||
|
|
||||||
|
def get_original_file_path(self, config):
|
||||||
|
return join(config.TRACKS_DIR, self.file_path, "original.csv")
|
||||||
|
|
||||||
|
|
||||||
|
class User(Base):
|
||||||
|
__tablename__ = "user"
|
||||||
|
id = Column(Integer, autoincrement=True, primary_key=True)
|
||||||
|
created_at = Column(DateTime, nullable=False, server_default=NOW)
|
||||||
|
updated_at = Column(DateTime, nullable=False, server_default=NOW, onupdate=NOW)
|
||||||
|
sub = Column(String, unique=True, nullable=False)
|
||||||
|
username = Column(String, unique=True, nullable=False)
|
||||||
|
email = Column(String, nullable=False)
|
||||||
|
bio = Column(TEXT)
|
||||||
|
image = Column(String)
|
||||||
|
are_tracks_visible_for_all = Column(Boolean, server_default=false(), nullable=False)
|
||||||
|
api_key = Column(String)
|
||||||
|
|
||||||
|
# This user can be matched by the email address from the auth service
|
||||||
|
# instead of having to match by `sub`. If a matching user logs in, the
|
||||||
|
# `sub` is updated to the new sub and this flag is disabled. This is for
|
||||||
|
# migrating *to* the external authentication scheme.
|
||||||
|
match_by_username_email = Column(Boolean, server_default=false())
|
||||||
|
|
||||||
|
def to_dict(self, for_user_id=None):
|
||||||
|
return {
|
||||||
|
"username": self.username,
|
||||||
|
"bio": self.bio,
|
||||||
|
"image": self.image,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class Comment(Base):
|
||||||
|
__tablename__ = "comment"
|
||||||
|
id = Column(Integer, autoincrement=True, primary_key=True)
|
||||||
|
uid = Column(UUID, server_default=func.uuid_generate_v4())
|
||||||
|
|
||||||
|
created_at = Column(DateTime, nullable=False, server_default=NOW)
|
||||||
|
updated_at = Column(DateTime, nullable=False, server_default=NOW, onupdate=NOW)
|
||||||
|
|
||||||
|
body = Column(TEXT)
|
||||||
|
|
||||||
|
author_id = Column(Integer, ForeignKey("user.id", ondelete="CASCADE"))
|
||||||
|
|
||||||
|
track_id = Column(Integer, ForeignKey("track.id", ondelete="CASCADE"))
|
||||||
|
|
||||||
|
def to_dict(self, for_user_id=None):
|
||||||
|
return {
|
||||||
|
"id": self.uid,
|
||||||
|
"body": self.body,
|
||||||
|
"author": self.author.to_dict(for_user_id=for_user_id),
|
||||||
|
"createdAt": self.created_at,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
Comment.author = relationship("User", back_populates="authored_comments")
|
||||||
|
User.authored_comments = relationship(
|
||||||
|
"Comment", order_by=Comment.created_at, back_populates="author"
|
||||||
|
)
|
||||||
|
|
||||||
|
Track.author = relationship("User", back_populates="authored_tracks")
|
||||||
|
User.authored_tracks = relationship(
|
||||||
|
"Track", order_by=Track.created_at, back_populates="author"
|
||||||
|
)
|
||||||
|
|
||||||
|
Comment.track = relationship("Track", back_populates="comments")
|
||||||
|
Track.comments = relationship(
|
||||||
|
"Comment", order_by=Comment.created_at, back_populates="track"
|
||||||
|
)
|
||||||
|
|
||||||
|
OvertakingEvent.track = relationship("Track", back_populates="overtaking_events")
|
||||||
|
Track.overtaking_events = relationship(
|
||||||
|
"OvertakingEvent", order_by=OvertakingEvent.time, back_populates="track"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
# 0..4 Night, 4..10 Morning, 10..14 Noon, 14..18 Afternoon, 18..22 Evening, 22..00 Night
|
||||||
|
# Two hour intervals
|
||||||
|
_DAYTIMES = [
|
||||||
|
"Night", # 0h - 2h
|
||||||
|
"Night", # 2h - 4h
|
||||||
|
"Morning", # 4h - 6h
|
||||||
|
"Morning", # 6h - 8h
|
||||||
|
"Morning", # 8h - 10h
|
||||||
|
"Noon", # 10h - 12h
|
||||||
|
"Noon", # 12h - 14h
|
||||||
|
"Afternoon", # 14h - 16h
|
||||||
|
"Afternoon", # 16h - 18h
|
||||||
|
"Evening", # 18h - 20h
|
||||||
|
"Evening", # 20h - 22h
|
||||||
|
"Night", # 22h - 24h
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def _get_daytime(d):
|
||||||
|
return _DAYTIMES[math.floor((d.hour % 24) / 2)]
|
7
api/obs/api/routes/__init__.py
Normal file
7
api/obs/api/routes/__init__.py
Normal file
|
@ -0,0 +1,7 @@
|
||||||
|
from . import (
|
||||||
|
login,
|
||||||
|
stats,
|
||||||
|
tracks,
|
||||||
|
info,
|
||||||
|
users,
|
||||||
|
)
|
20
api/obs/api/routes/info.py
Normal file
20
api/obs/api/routes/info.py
Normal file
|
@ -0,0 +1,20 @@
|
||||||
|
import logging
|
||||||
|
|
||||||
|
# from sqlalchemy import select
|
||||||
|
|
||||||
|
from obs.api.app import app
|
||||||
|
|
||||||
|
from sanic.response import json
|
||||||
|
|
||||||
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
from obs.api import __version__ as version
|
||||||
|
|
||||||
|
|
||||||
|
@app.route("/info")
|
||||||
|
async def info(req):
|
||||||
|
return json(
|
||||||
|
{
|
||||||
|
"version": version,
|
||||||
|
}
|
||||||
|
)
|
118
api/obs/api/routes/login.py
Normal file
118
api/obs/api/routes/login.py
Normal file
|
@ -0,0 +1,118 @@
|
||||||
|
import logging
|
||||||
|
import os
|
||||||
|
|
||||||
|
from sqlalchemy import select
|
||||||
|
|
||||||
|
from oic import rndstr
|
||||||
|
from oic.oic import Client
|
||||||
|
from oic.oic.message import AuthorizationResponse, RegistrationResponse
|
||||||
|
from oic.utils.authn.client import CLIENT_AUTHN_METHOD
|
||||||
|
|
||||||
|
from obs.api.app import app
|
||||||
|
from obs.api.db import User
|
||||||
|
|
||||||
|
from sanic.response import json, redirect
|
||||||
|
from sanicargs import parse_parameters
|
||||||
|
|
||||||
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
client = Client(client_authn_method=CLIENT_AUTHN_METHOD)
|
||||||
|
|
||||||
|
|
||||||
|
@app.before_server_start
|
||||||
|
async def connect_auth_client(app, loop):
|
||||||
|
client.allow["issuer_mismatch"] = True
|
||||||
|
pc = client.provider_config(app.config.KEYCLOAK_URL)
|
||||||
|
client.store_registration_info(
|
||||||
|
RegistrationResponse(
|
||||||
|
client_id=app.config.KEYCLOAK_CLIENT_ID,
|
||||||
|
client_secret=app.config.KEYCLOAK_CLIENT_SECRET,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route("/login")
|
||||||
|
@parse_parameters
|
||||||
|
async def login(req, next: str = None):
|
||||||
|
session = req.ctx.session
|
||||||
|
session["state"] = rndstr()
|
||||||
|
session["nonce"] = rndstr()
|
||||||
|
session["next"] = next
|
||||||
|
args = {
|
||||||
|
"client_id": client.client_id,
|
||||||
|
"response_type": "code",
|
||||||
|
"scope": ["openid"],
|
||||||
|
"nonce": session["nonce"],
|
||||||
|
"redirect_uri": req.scheme + "://" + req.host + "/login/redirect",
|
||||||
|
"state": session["state"],
|
||||||
|
}
|
||||||
|
|
||||||
|
auth_req = client.construct_AuthorizationRequest(request_args=args)
|
||||||
|
login_url = auth_req.request(client.authorization_endpoint)
|
||||||
|
|
||||||
|
return redirect(login_url)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route("/login/redirect")
|
||||||
|
async def login_redirect(req):
|
||||||
|
session = req.ctx.session
|
||||||
|
|
||||||
|
auth_response = client.parse_response(
|
||||||
|
AuthorizationResponse, info=dict(req.query_args), sformat="dict"
|
||||||
|
)
|
||||||
|
code = auth_response["code"]
|
||||||
|
state = auth_response["state"]
|
||||||
|
|
||||||
|
assert "state" in session
|
||||||
|
assert state == session["state"]
|
||||||
|
|
||||||
|
client.do_access_token_request(
|
||||||
|
state=state,
|
||||||
|
request_args={"code": code},
|
||||||
|
authn_method="client_secret_basic",
|
||||||
|
)
|
||||||
|
|
||||||
|
userinfo = client.do_user_info_request(state=state)
|
||||||
|
|
||||||
|
# {'sub': '3798e2da-b208-4a1a-98c0-08fecfea1345', 'email_verified': True, 'preferred_username': 'test', 'email': 'test@example.com'}
|
||||||
|
sub = userinfo["sub"]
|
||||||
|
preferred_username = userinfo["preferred_username"]
|
||||||
|
email = userinfo["email"]
|
||||||
|
|
||||||
|
user = (await req.ctx.db.execute(select(User).where(User.sub == sub))).scalar()
|
||||||
|
|
||||||
|
if user is None:
|
||||||
|
user = (
|
||||||
|
await req.ctx.db.execute(
|
||||||
|
select(User).where(User.email == email and User.username = preferred_username and User.match_by_username_email)
|
||||||
|
)
|
||||||
|
).scalar()
|
||||||
|
|
||||||
|
if user:
|
||||||
|
user.match_by_username_email = False
|
||||||
|
|
||||||
|
if user is None:
|
||||||
|
log.info(
|
||||||
|
"Registering new user with sub %r (preferred username: %s)",
|
||||||
|
sub,
|
||||||
|
preferred_username,
|
||||||
|
)
|
||||||
|
user = User(sub=sub, username=preferred_username, email=email)
|
||||||
|
req.ctx.db.add(user)
|
||||||
|
else:
|
||||||
|
log.info("Logged in known user (id: %s).", user.id)
|
||||||
|
|
||||||
|
if email != user.email:
|
||||||
|
log.debug("Updating user (id: %s) email from auth system.", user.id)
|
||||||
|
user.email = email
|
||||||
|
|
||||||
|
if preferred_username != user.username:
|
||||||
|
log.debug("Updating user (id: %s) username from auth system.", user.id)
|
||||||
|
user.username = preferred_username
|
||||||
|
|
||||||
|
await req.ctx.db.commit()
|
||||||
|
|
||||||
|
session["user_id"] = user.id
|
||||||
|
|
||||||
|
next_ = session.pop("next", "/") or "/"
|
||||||
|
return redirect(next_)
|
171
api/obs/api/routes/stats.py
Normal file
171
api/obs/api/routes/stats.py
Normal file
|
@ -0,0 +1,171 @@
|
||||||
|
import logging
|
||||||
|
from datetime import datetime
|
||||||
|
from typing import Optional
|
||||||
|
from operator import and_
|
||||||
|
from functools import reduce
|
||||||
|
|
||||||
|
from sqlalchemy import select, func
|
||||||
|
|
||||||
|
from sanic.response import json
|
||||||
|
from sanicargs import parse_parameters
|
||||||
|
|
||||||
|
from obs.api.app import app
|
||||||
|
from obs.api.db import Track, OvertakingEvent, User
|
||||||
|
|
||||||
|
|
||||||
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
# round to this number of meters for privacy reasons
|
||||||
|
TRACK_LENGTH_ROUNDING = 1000
|
||||||
|
|
||||||
|
# round to this number of seconds for privacy reasons
|
||||||
|
TRACK_DURATION_ROUNDING = 120
|
||||||
|
|
||||||
|
# Everything before this date is probably parsed incorrectly
|
||||||
|
MINUMUM_RECORDING_DATE = datetime(2010, 1, 1)
|
||||||
|
|
||||||
|
|
||||||
|
def round_to(value: float, multiples: float) -> float:
|
||||||
|
return round(value / multiples) * multiples
|
||||||
|
|
||||||
|
|
||||||
|
@app.route("/stats")
|
||||||
|
@parse_parameters
|
||||||
|
async def stats(req, user: str = None, start: datetime = None, end: datetime = None):
|
||||||
|
conditions = [
|
||||||
|
Track.recorded_at != None,
|
||||||
|
Track.recorded_at > MINUMUM_RECORDING_DATE,
|
||||||
|
]
|
||||||
|
|
||||||
|
if start is not None:
|
||||||
|
conditions.append(Track.recorded_at >= start)
|
||||||
|
|
||||||
|
if end is not None:
|
||||||
|
conditions.append(Track.recorded_at < end)
|
||||||
|
|
||||||
|
# Only the user can look for their own stats, for now
|
||||||
|
by_user = (
|
||||||
|
user is not None and req.ctx.user is not None and req.ctx.user.username == user
|
||||||
|
)
|
||||||
|
if by_user:
|
||||||
|
conditions.append(Track.author_id == req.ctx.user.id)
|
||||||
|
|
||||||
|
print(conditions)
|
||||||
|
track_condition = reduce(and_, conditions)
|
||||||
|
public_track_condition = Track.public and track_condition
|
||||||
|
|
||||||
|
query = (
|
||||||
|
select(
|
||||||
|
[
|
||||||
|
func.count().label("publicTrackCount"),
|
||||||
|
func.sum(Track.duration).label("trackDuration"),
|
||||||
|
func.sum(Track.length).label("trackLength"),
|
||||||
|
]
|
||||||
|
)
|
||||||
|
.select_from(Track)
|
||||||
|
.where(public_track_condition)
|
||||||
|
)
|
||||||
|
|
||||||
|
public_track_count, track_duration, track_length = (
|
||||||
|
await req.ctx.db.execute(query)
|
||||||
|
).first()
|
||||||
|
|
||||||
|
# This is required because SQL returns NULL when the input set to a
|
||||||
|
# SUM() aggregation is empty.
|
||||||
|
track_duration = track_duration or 0
|
||||||
|
track_length = track_length or 0
|
||||||
|
|
||||||
|
user_count = (
|
||||||
|
1
|
||||||
|
if by_user
|
||||||
|
else (await req.ctx.db.execute(select(func.count()).select_from(User))).scalar()
|
||||||
|
)
|
||||||
|
track_count = (
|
||||||
|
await req.ctx.db.execute(
|
||||||
|
select(func.count()).select_from(Track).where(track_condition)
|
||||||
|
)
|
||||||
|
).scalar()
|
||||||
|
event_count = (
|
||||||
|
await req.ctx.db.execute(
|
||||||
|
select(func.count())
|
||||||
|
.select_from(OvertakingEvent)
|
||||||
|
.join(OvertakingEvent.track)
|
||||||
|
.where(track_condition)
|
||||||
|
)
|
||||||
|
).scalar()
|
||||||
|
|
||||||
|
result = {
|
||||||
|
"numEvents": event_count,
|
||||||
|
"userCount": user_count,
|
||||||
|
"trackLength": round_to(track_length or 0, TRACK_LENGTH_ROUNDING),
|
||||||
|
"trackDuration": round_to(track_duration or 0, TRACK_DURATION_ROUNDING),
|
||||||
|
"publicTrackCount": public_track_count,
|
||||||
|
"trackCount": track_count,
|
||||||
|
}
|
||||||
|
|
||||||
|
return json(result)
|
||||||
|
|
||||||
|
|
||||||
|
# const trackCount = await Track.find(trackFilter).count();
|
||||||
|
#
|
||||||
|
# const publicTrackCount = await Track.find({
|
||||||
|
# ...trackFilter,
|
||||||
|
# public: true,
|
||||||
|
# }).count();
|
||||||
|
#
|
||||||
|
# const userCount = await User.find({
|
||||||
|
# ...(userFilter
|
||||||
|
# ? { _id: userFilter }
|
||||||
|
# : {
|
||||||
|
# createdAt: dateFilter,
|
||||||
|
# }),
|
||||||
|
# }).count();
|
||||||
|
#
|
||||||
|
# const trackStats = await Track.aggregate([
|
||||||
|
# { $match: trackFilter },
|
||||||
|
# {
|
||||||
|
# $addFields: {
|
||||||
|
# trackLength: {
|
||||||
|
# $cond: [{ $lt: ['$statistics.length', 500000] }, '$statistics.length', 0],
|
||||||
|
# },
|
||||||
|
# numEvents: '$statistics.numEvents',
|
||||||
|
# trackDuration: {
|
||||||
|
# $cond: [
|
||||||
|
# { $and: ['$statistics.recordedUntil', { $gt: ['$statistics.recordedAt', new Date('2010-01-01')] }] },
|
||||||
|
# { $subtract: ['$statistics.recordedUntil', '$statistics.recordedAt'] },
|
||||||
|
# 0,
|
||||||
|
# ],
|
||||||
|
# },
|
||||||
|
# },
|
||||||
|
# },
|
||||||
|
# { $project: { trackLength: true, numEvents: true, trackDuration: true } },
|
||||||
|
# {
|
||||||
|
# $group: {
|
||||||
|
# _id: 'sum',
|
||||||
|
# trackLength: { $sum: '$trackLength' },
|
||||||
|
# numEvents: { $sum: '$numEvents' },
|
||||||
|
# trackDuration: { $sum: '$trackDuration' },
|
||||||
|
# },
|
||||||
|
# },
|
||||||
|
# ]);
|
||||||
|
#
|
||||||
|
# const [trackLength, numEvents, trackDuration] =
|
||||||
|
# trackStats.length > 0
|
||||||
|
# ? [trackStats[0].trackLength, trackStats[0].numEvents, trackStats[0].trackDuration]
|
||||||
|
# : [0, 0, 0];
|
||||||
|
#
|
||||||
|
# const trackLengthPrivatized = Math.floor(trackLength / TRACK_LENGTH_ROUNDING) * TRACK_LENGTH_ROUNDING;
|
||||||
|
# const trackDurationPrivatized =
|
||||||
|
# Math.round(trackDuration / 1000 / TRACK_DURATION_ROUNDING) * TRACK_DURATION_ROUNDING;
|
||||||
|
#
|
||||||
|
# return res.json({
|
||||||
|
# publicTrackCount,
|
||||||
|
# trackLength: trackLengthPrivatized,
|
||||||
|
# trackDuration: trackDurationPrivatized,
|
||||||
|
# numEvents,
|
||||||
|
# trackCount,
|
||||||
|
# userCount,
|
||||||
|
# });
|
||||||
|
# }),
|
||||||
|
# );
|
362
api/obs/api/routes/tracks.py
Normal file
362
api/obs/api/routes/tracks.py
Normal file
|
@ -0,0 +1,362 @@
|
||||||
|
import logging
|
||||||
|
import re
|
||||||
|
from json import load as jsonload
|
||||||
|
from os.path import join, exists, isfile
|
||||||
|
|
||||||
|
from sqlalchemy import select, func
|
||||||
|
from sqlalchemy.orm import joinedload
|
||||||
|
|
||||||
|
from obs.api.db import Track, User, Comment
|
||||||
|
from obs.api.app import app, require_auth, json
|
||||||
|
|
||||||
|
from sanic.response import file_stream, empty
|
||||||
|
from sanic.exceptions import InvalidUsage, NotFound, Forbidden
|
||||||
|
from sanicargs import parse_parameters
|
||||||
|
|
||||||
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def normalize_user_agent(user_agent):
|
||||||
|
if not user_agent:
|
||||||
|
return None
|
||||||
|
|
||||||
|
m = re.match(r"\bOBS\/[^\s]+", user_agent)
|
||||||
|
return m[0] if m else None
|
||||||
|
|
||||||
|
|
||||||
|
async def _return_tracks(req, extend_query, limit, offset):
|
||||||
|
if limit <= 0 or limit > 100:
|
||||||
|
raise InvalidUsage("invalid limit")
|
||||||
|
|
||||||
|
if offset < 0:
|
||||||
|
raise InvalidUsage("offset must be positive")
|
||||||
|
|
||||||
|
count_query = extend_query(
|
||||||
|
select(func.count()).select_from(Track).join(Track.author)
|
||||||
|
)
|
||||||
|
track_count = await req.ctx.db.scalar(count_query)
|
||||||
|
|
||||||
|
query = (
|
||||||
|
extend_query(select(Track).options(joinedload(Track.author)))
|
||||||
|
.limit(limit)
|
||||||
|
.offset(offset)
|
||||||
|
.order_by(Track.created_at.desc())
|
||||||
|
)
|
||||||
|
|
||||||
|
tracks = (await req.ctx.db.execute(query)).scalars()
|
||||||
|
|
||||||
|
return json(
|
||||||
|
{
|
||||||
|
"trackCount": track_count,
|
||||||
|
"tracks": list(
|
||||||
|
map(
|
||||||
|
lambda t: t.to_dict(
|
||||||
|
for_user_id=req.ctx.user.id if req.ctx.user else None
|
||||||
|
),
|
||||||
|
tracks,
|
||||||
|
)
|
||||||
|
),
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/tracks")
|
||||||
|
@parse_parameters
|
||||||
|
async def get_tracks(req, limit: int = 20, offset: int = 0, author: str = None):
|
||||||
|
def extend_query(q):
|
||||||
|
q = q.where(Track.public)
|
||||||
|
|
||||||
|
if author is not None:
|
||||||
|
q = q.where(User.username == author)
|
||||||
|
|
||||||
|
return q
|
||||||
|
|
||||||
|
return await _return_tracks(req, extend_query, limit, offset)
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/tracks/feed")
|
||||||
|
@require_auth
|
||||||
|
@parse_parameters
|
||||||
|
async def get_feed(req, limit: int = 20, offset: int = 0):
|
||||||
|
def extend_query(q):
|
||||||
|
return q.where(Track.author_id == req.ctx.user.id)
|
||||||
|
|
||||||
|
return await _return_tracks(req, extend_query, limit, offset)
|
||||||
|
|
||||||
|
|
||||||
|
@app.post("/tracks")
|
||||||
|
@require_auth
|
||||||
|
async def post_track(req):
|
||||||
|
try:
|
||||||
|
file = req.files["body"][0]
|
||||||
|
except LookupError as e:
|
||||||
|
raise InvalidUsage(
|
||||||
|
'Track upload needs a single file in "body" multipart field'
|
||||||
|
) from e
|
||||||
|
|
||||||
|
try:
|
||||||
|
body = req.json["track"]
|
||||||
|
except (LookupError, InvalidUsage):
|
||||||
|
body = {}
|
||||||
|
|
||||||
|
title = body.get("title")
|
||||||
|
public = body.get("public")
|
||||||
|
|
||||||
|
track = Track(
|
||||||
|
title=title,
|
||||||
|
customized_title=bool(title),
|
||||||
|
author=req.ctx.user,
|
||||||
|
public=public
|
||||||
|
if public is not None
|
||||||
|
else req.ctx.user.are_tracks_visible_for_all,
|
||||||
|
)
|
||||||
|
track.generate_slug()
|
||||||
|
await track.prevent_duplicates(req.ctx.db, file.body)
|
||||||
|
track.uploaded_by_user_agent = normalize_user_agent(req.headers["user-agent"])
|
||||||
|
track.original_file_name = file.name
|
||||||
|
await track.write_to_original_file(req.app.config, file.body)
|
||||||
|
track.queue_processing()
|
||||||
|
track.auto_generate_title()
|
||||||
|
|
||||||
|
req.ctx.db.add(track)
|
||||||
|
await req.ctx.db.commit()
|
||||||
|
|
||||||
|
return await get_track(req, track.slug)
|
||||||
|
|
||||||
|
|
||||||
|
async def _load_track(req, slug, raise_not_found=True):
|
||||||
|
track = (
|
||||||
|
await req.ctx.db.execute(
|
||||||
|
select(Track)
|
||||||
|
.where(Track.slug == slug)
|
||||||
|
.options(joinedload(Track.author))
|
||||||
|
.limit(1)
|
||||||
|
)
|
||||||
|
).scalar()
|
||||||
|
|
||||||
|
if raise_not_found and track is None:
|
||||||
|
raise NotFound()
|
||||||
|
|
||||||
|
if not track.is_visible_to(req.ctx.user):
|
||||||
|
raise Forbidden()
|
||||||
|
|
||||||
|
return track
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/tracks/<slug:str>")
|
||||||
|
async def get_track(req, slug: str):
|
||||||
|
track = await _load_track(req, slug)
|
||||||
|
return json(
|
||||||
|
{"track": track.to_dict(for_user_id=req.ctx.user.id if req.ctx.user else None)},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@app.delete("/tracks/<slug:str>")
|
||||||
|
@require_auth
|
||||||
|
async def delete_track(req, slug: str):
|
||||||
|
track = await _load_track(req, slug)
|
||||||
|
if not track.is_visible_to_private(req.ctx.user):
|
||||||
|
raise Forbidden()
|
||||||
|
|
||||||
|
await req.ctx.db.delete(track)
|
||||||
|
await req.ctx.db.commit()
|
||||||
|
|
||||||
|
return empty()
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/tracks/<slug:str>/data")
|
||||||
|
async def get_track_data(req, slug: str):
|
||||||
|
track = await _load_track(req, slug)
|
||||||
|
|
||||||
|
FILE_BY_KEY = {
|
||||||
|
"measurements": "measurements.json",
|
||||||
|
"overtakingEvents": "overtakingEvents.json",
|
||||||
|
"track": "track.json",
|
||||||
|
}
|
||||||
|
|
||||||
|
result = {}
|
||||||
|
|
||||||
|
for key, filename in FILE_BY_KEY.items():
|
||||||
|
file_path = join(
|
||||||
|
req.app.config.PROCESSING_OUTPUT_DIR, track.file_path, filename
|
||||||
|
)
|
||||||
|
if not exists(file_path) or not isfile(file_path):
|
||||||
|
continue
|
||||||
|
|
||||||
|
with open(file_path) as f:
|
||||||
|
result[key] = jsonload(f)
|
||||||
|
|
||||||
|
return json(
|
||||||
|
result,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/tracks/<slug:str>/download/original.csv")
|
||||||
|
async def download_original_file(req, slug: str):
|
||||||
|
track = await _load_track(req, slug)
|
||||||
|
|
||||||
|
if not track.is_visible_to_private(req.ctx.user):
|
||||||
|
raise Forbidden()
|
||||||
|
|
||||||
|
return await file_stream(track.get_original_file_path(app.config))
|
||||||
|
|
||||||
|
|
||||||
|
@app.put("/tracks/<slug:str>")
|
||||||
|
@require_auth
|
||||||
|
async def put_track(req, slug: str):
|
||||||
|
track = await _load_track(req, slug)
|
||||||
|
|
||||||
|
if track.author_id != req.ctx.user.id:
|
||||||
|
raise Forbidden()
|
||||||
|
|
||||||
|
try:
|
||||||
|
body = req.json["track"]
|
||||||
|
except:
|
||||||
|
body = {}
|
||||||
|
|
||||||
|
if "title" in body:
|
||||||
|
track.title = (body["title"] or "").strip() or None
|
||||||
|
track.customized_title = track.title is not None
|
||||||
|
|
||||||
|
if "description" in body:
|
||||||
|
track.description = (body["description"] or "").strip() or None
|
||||||
|
|
||||||
|
process = False
|
||||||
|
|
||||||
|
if "public" in body:
|
||||||
|
public = bool(body["public"])
|
||||||
|
process = process or (public != track.public) # if changed
|
||||||
|
track.public = public
|
||||||
|
|
||||||
|
if "body" in req.files:
|
||||||
|
try:
|
||||||
|
file = req.files["body"][0]
|
||||||
|
except LookupError as e:
|
||||||
|
raise InvalidUsage(
|
||||||
|
'Track upload needs a single file in "body" multipart field'
|
||||||
|
) from e
|
||||||
|
|
||||||
|
await track.prevent_duplicates(req.ctx.db, file.body)
|
||||||
|
track.uploaded_by_user_agent = normalize_user_agent(req.headers["user-agent"])
|
||||||
|
track.original_file_name = file.name or (track.slug + ".csv")
|
||||||
|
await track.write_to_original_file(req.app.config, file.body)
|
||||||
|
process = True
|
||||||
|
|
||||||
|
if process:
|
||||||
|
track.queue_processing()
|
||||||
|
|
||||||
|
track.auto_generate_title()
|
||||||
|
await req.ctx.db.commit()
|
||||||
|
|
||||||
|
track = await _load_track(req, track.slug)
|
||||||
|
return json(
|
||||||
|
{"track": track.to_dict(for_user_id=req.ctx.user.id)},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/tracks/<slug:str>/comments")
|
||||||
|
@parse_parameters
|
||||||
|
async def get_track_comments(req, slug: str, limit: int = 20, offset: int = 0):
|
||||||
|
track = await _load_track(req, slug)
|
||||||
|
|
||||||
|
comment_count = await req.ctx.db.scalar(
|
||||||
|
select(func.count()).select_from(Comment).where(Comment.track_id == track.id)
|
||||||
|
)
|
||||||
|
|
||||||
|
query = (
|
||||||
|
select(Comment)
|
||||||
|
.options(joinedload(Comment.author))
|
||||||
|
.where(Comment.track_id == track.id)
|
||||||
|
.order_by(Comment.created_at.desc())
|
||||||
|
.limit(limit)
|
||||||
|
.offset(offset)
|
||||||
|
)
|
||||||
|
|
||||||
|
comments = (await req.ctx.db.execute(query)).scalars()
|
||||||
|
|
||||||
|
return json(
|
||||||
|
{
|
||||||
|
"commentCount": comment_count,
|
||||||
|
"comments": list(
|
||||||
|
map(
|
||||||
|
lambda c: c.to_dict(
|
||||||
|
for_user_id=req.ctx.user.id if req.ctx.user else None
|
||||||
|
),
|
||||||
|
comments,
|
||||||
|
)
|
||||||
|
),
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@app.post("/tracks/<slug:str>/comments")
|
||||||
|
@require_auth
|
||||||
|
async def post_track_comment(req, slug: str):
|
||||||
|
track = await _load_track(req, slug)
|
||||||
|
|
||||||
|
body = req.json.get("comment", {}).get("body")
|
||||||
|
if not isinstance(body, str):
|
||||||
|
raise InvalidUsage("no comment given")
|
||||||
|
|
||||||
|
# Ensure body is not empty
|
||||||
|
body = body.strip()
|
||||||
|
if not body:
|
||||||
|
raise InvalidUsage("empty comment")
|
||||||
|
|
||||||
|
comment = Comment(
|
||||||
|
body=body,
|
||||||
|
track_id=track.id,
|
||||||
|
author_id=req.ctx.user.id,
|
||||||
|
)
|
||||||
|
|
||||||
|
req.ctx.db.add(comment)
|
||||||
|
await req.ctx.db.commit()
|
||||||
|
|
||||||
|
await req.ctx.db.refresh(comment)
|
||||||
|
|
||||||
|
comment = (
|
||||||
|
await req.ctx.db.execute(
|
||||||
|
select(Comment)
|
||||||
|
.options(joinedload(Comment.author))
|
||||||
|
.where(Comment.id == comment.id)
|
||||||
|
.limit(1)
|
||||||
|
)
|
||||||
|
).scalar()
|
||||||
|
|
||||||
|
return json({"comment": comment.to_dict(for_user_id=req.ctx.user.id)})
|
||||||
|
|
||||||
|
|
||||||
|
@app.delete("/tracks/<slug:str>/comments/<uid:str>")
|
||||||
|
@require_auth
|
||||||
|
async def delete_track_comment(req, slug: str, uid: str):
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
print("XXXXXXXXXXXXXX")
|
||||||
|
track = await _load_track(req, slug)
|
||||||
|
|
||||||
|
print("trackid", track.id, " uid", uid)
|
||||||
|
comment = (
|
||||||
|
await req.ctx.db.execute(
|
||||||
|
select(Comment)
|
||||||
|
.options(joinedload(Comment.author))
|
||||||
|
.where(Comment.track_id == track.id and Comment.uid == uid)
|
||||||
|
.limit(1)
|
||||||
|
)
|
||||||
|
).scalar()
|
||||||
|
|
||||||
|
if not comment:
|
||||||
|
raise NotFound()
|
||||||
|
|
||||||
|
if comment.author_id != req.ctx.user.id:
|
||||||
|
raise Forbidden()
|
||||||
|
|
||||||
|
await req.ctx.db.delete(comment)
|
||||||
|
await req.ctx.db.commit()
|
||||||
|
|
||||||
|
return empty()
|
42
api/obs/api/routes/users.py
Normal file
42
api/obs/api/routes/users.py
Normal file
|
@ -0,0 +1,42 @@
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from sanic.response import json
|
||||||
|
|
||||||
|
from obs.api.app import app, require_auth
|
||||||
|
|
||||||
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
from obs.api import __version__ as version
|
||||||
|
|
||||||
|
|
||||||
|
def user_to_json(user):
|
||||||
|
return {
|
||||||
|
"username": user.username,
|
||||||
|
"email": user.email,
|
||||||
|
"bio": user.bio,
|
||||||
|
"image": user.image,
|
||||||
|
"areTracksVisibleForAll": user.are_tracks_visible_for_all,
|
||||||
|
# "apiKey": user.api_key,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/user")
|
||||||
|
@require_auth
|
||||||
|
async def get_user(req):
|
||||||
|
return json(user_to_json(req.ctx.user))
|
||||||
|
|
||||||
|
|
||||||
|
@app.put("/user")
|
||||||
|
@require_auth
|
||||||
|
async def put_user(req):
|
||||||
|
user = req.ctx.user
|
||||||
|
|
||||||
|
for key in ["username", "email", "bio", "image"]:
|
||||||
|
if key in req.json and isinstance(req.json[key], (str, type(None))):
|
||||||
|
setattr(user, key, req.json[key])
|
||||||
|
|
||||||
|
if "areTracksVisibleForAll" in req.json:
|
||||||
|
user.are_tracks_visible_for_all = bool(req.json["areTracksVisibleForAll"])
|
||||||
|
|
||||||
|
await req.ctx.db.commit()
|
||||||
|
return json(user_to_json(req.ctx.user))
|
1
api/obs/bin/__init__.py
Normal file
1
api/obs/bin/__init__.py
Normal file
|
@ -0,0 +1 @@
|
||||||
|
__path__ = __import__("pkgutil").extend_path(__path__, __name__)
|
17
api/obs/bin/obs_api.py
Executable file
17
api/obs/bin/obs_api.py
Executable file
|
@ -0,0 +1,17 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
import sys
|
||||||
|
import os
|
||||||
|
import argparse
|
||||||
|
import asyncio
|
||||||
|
|
||||||
|
from obs.api.app import app
|
||||||
|
from obs.api.db import connect_db
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
app.run(host=app.config.HOST, port=app.config.PORT, debug=app.config.DEBUG)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
|
@ -1,16 +0,0 @@
|
||||||
CREATE TYPE zone_type IF NOT EXISTS AS ENUM ('urban', 'rural');
|
|
||||||
|
|
||||||
CREATE TABLE road_annotations IF NOT EXISTS (
|
|
||||||
way_id integer,
|
|
||||||
reverse boolean,
|
|
||||||
name text,
|
|
||||||
zone zone_type,
|
|
||||||
distance_overtaker_mean float,
|
|
||||||
distance_overtaker_median float,
|
|
||||||
distance_overtaker_minimum float,
|
|
||||||
distance_overtaker_n integer,
|
|
||||||
distance_overtaker_n_below_limit integer,
|
|
||||||
distance_overtaker_n_above_limit integer,
|
|
||||||
distance_overtaker_limit float,
|
|
||||||
distance_overtaker_measurements integer ARRAY,
|
|
||||||
);
|
|
|
@ -1,3 +1,7 @@
|
||||||
./scripts
|
sanic~=21.9.1
|
||||||
sqlalchemy[asyncio]
|
oic~=1.3.0
|
||||||
asyncpg
|
sanic-session~=0.8.0
|
||||||
|
sanicargs~=2.1.0
|
||||||
|
sanic-cors~=1.0.1
|
||||||
|
python-slugify~=5.0.2
|
||||||
|
motor~=2.5.1
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
Subproject commit e67a2a98f7d8d53c687bc44d72774b5429b049c3
|
Subproject commit 145b06a80d4607ff2c4a8dcf80e2f5fb0e1c8f1a
|
21
api/setup.py
Normal file
21
api/setup.py
Normal file
|
@ -0,0 +1,21 @@
|
||||||
|
from setuptools import setup, find_packages
|
||||||
|
|
||||||
|
with open("requirements.txt") as f:
|
||||||
|
requires = list(f.readlines())
|
||||||
|
|
||||||
|
setup(
|
||||||
|
name="obs-api",
|
||||||
|
version="0.0.1",
|
||||||
|
author="OpenBikeSensor Contributors",
|
||||||
|
license="LGPL-3.0",
|
||||||
|
description="OpenBikeSensor Portal API",
|
||||||
|
url="https://github.com/openbikesensor/portal",
|
||||||
|
packages=find_packages(),
|
||||||
|
package_data={},
|
||||||
|
install_requires=requires,
|
||||||
|
entry_points={
|
||||||
|
"console_scripts": [
|
||||||
|
"obs-api=obs.bin.obs_api:main",
|
||||||
|
]
|
||||||
|
},
|
||||||
|
)
|
132
api/tools/import_from_mongodb.py
Normal file
132
api/tools/import_from_mongodb.py
Normal file
|
@ -0,0 +1,132 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
import argparse
|
||||||
|
import asyncio
|
||||||
|
import logging
|
||||||
|
from datetime import datetime
|
||||||
|
from uuid import uuid4
|
||||||
|
|
||||||
|
from sqlalchemy import select
|
||||||
|
|
||||||
|
from motor.motor_asyncio import AsyncIOMotorClient
|
||||||
|
|
||||||
|
from obs.api.db import make_session, connect_db, User, Track, async_session, Comment
|
||||||
|
from obs.api.app import app
|
||||||
|
|
||||||
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
async def main():
|
||||||
|
logging.basicConfig(level=logging.DEBUG, format="%(levelname)s: %(message)s")
|
||||||
|
|
||||||
|
parser = argparse.ArgumentParser(
|
||||||
|
description="processes a single track for use in the portal, "
|
||||||
|
"using the obs.face algorithms"
|
||||||
|
)
|
||||||
|
|
||||||
|
parser.add_argument(
|
||||||
|
"mongodb_url",
|
||||||
|
metavar="MONGODB_URL",
|
||||||
|
help="url to the mongodb, in format mongodb://user:pass@host/dbname",
|
||||||
|
)
|
||||||
|
|
||||||
|
args = parser.parse_args()
|
||||||
|
|
||||||
|
async with connect_db(app.config.POSTGRES_URL):
|
||||||
|
async with make_session() as session:
|
||||||
|
mongo = AsyncIOMotorClient(args.mongodb_url).get_default_database()
|
||||||
|
|
||||||
|
log.debug("Connected to mongodb and postgres.")
|
||||||
|
user_id_map = await import_users(mongo, session)
|
||||||
|
|
||||||
|
await import_tracks(mongo, session, user_id_map)
|
||||||
|
|
||||||
|
await session.commit()
|
||||||
|
|
||||||
|
|
||||||
|
async def import_users(mongo, session):
|
||||||
|
old_id_by_email = {}
|
||||||
|
async for user in mongo.users.find({}):
|
||||||
|
old_id_by_email[user["email"]] = user["_id"]
|
||||||
|
|
||||||
|
new_user = User(
|
||||||
|
sub=str(uuid4()),
|
||||||
|
match_by_username_email=True,
|
||||||
|
email=user["email"],
|
||||||
|
username=user["username"],
|
||||||
|
bio=user.get("bio"),
|
||||||
|
image=user.get("image"),
|
||||||
|
are_tracks_visible_for_all=user.get("areTracksVisibleForAll") or False,
|
||||||
|
api_key=str(user["_id"]),
|
||||||
|
created_at=user.get("createdAt") or datetime.utcnow(),
|
||||||
|
updated_at=user.get("updatedAt") or datetime.utcnow(),
|
||||||
|
)
|
||||||
|
|
||||||
|
session.add(new_user)
|
||||||
|
log.info("Creating user %s", new_user.username)
|
||||||
|
|
||||||
|
await session.commit()
|
||||||
|
|
||||||
|
id_map = {}
|
||||||
|
result = await session.scalars(select(User))
|
||||||
|
for user in result:
|
||||||
|
id_map[old_id_by_email[user.email]] = user.id
|
||||||
|
|
||||||
|
return id_map
|
||||||
|
|
||||||
|
|
||||||
|
def parse_datetime(s):
|
||||||
|
if isinstance(s, str):
|
||||||
|
return datetime.fromisoformat(s)
|
||||||
|
return s
|
||||||
|
|
||||||
|
|
||||||
|
async def import_tracks(mongo, session, user_id_map):
|
||||||
|
async for track in mongo.tracks.find({}):
|
||||||
|
stats = track.get("statistics") or {}
|
||||||
|
new_track = Track(
|
||||||
|
created_at=parse_datetime(track.get("createdAt")) or datetime.utcnow(),
|
||||||
|
updated_at=parse_datetime(track.get("updatedAt")) or datetime.utcnow(),
|
||||||
|
slug=track["slug"],
|
||||||
|
title=track.get("title"),
|
||||||
|
processing_status=track.get("processingStatus") or "pending",
|
||||||
|
processing_log=track.get("processingLog"),
|
||||||
|
customized_title=bool(track.get("customizedTitle")),
|
||||||
|
description=track.get("description"),
|
||||||
|
public=track.get("public"),
|
||||||
|
uploaded_by_user_agent=track.get("uploadedByUserAgent"),
|
||||||
|
original_file_name=track.get("originalFileName"),
|
||||||
|
original_file_hash=track.get("originalFileHash"),
|
||||||
|
# statistics
|
||||||
|
recorded_at=parse_datetime(stats.get("recordedAt")),
|
||||||
|
recorded_until=parse_datetime(stats.get("recordedUntil")),
|
||||||
|
duration=stats.get("duration"),
|
||||||
|
length=stats.get("length"),
|
||||||
|
segments=stats.get("segments"),
|
||||||
|
num_events=stats.get("num_events"),
|
||||||
|
num_measurements=stats.get("num_measurements"),
|
||||||
|
num_valid=stats.get("numValid"),
|
||||||
|
author_id=user_id_map[track["author"]],
|
||||||
|
)
|
||||||
|
|
||||||
|
session.add(new_track)
|
||||||
|
log.info("Creating track %s", new_track.slug)
|
||||||
|
|
||||||
|
comment_ids = track.get("comments") or []
|
||||||
|
if comment_ids:
|
||||||
|
async for comment in mongo.comments.find({"_id": {"$in": comment_ids}}):
|
||||||
|
new_comment = Comment(
|
||||||
|
created_at=parse_datetime(comment.get("createdAt"))
|
||||||
|
or datetime.utcnow(),
|
||||||
|
updated_at=parse_datetime(comment.get("updatedAt"))
|
||||||
|
or datetime.utcnow(),
|
||||||
|
body=comment.get("body"),
|
||||||
|
author_id=user_id_map[comment["author"]],
|
||||||
|
)
|
||||||
|
new_track.comments.append(new_comment)
|
||||||
|
session.add(new_comment)
|
||||||
|
|
||||||
|
await session.commit()
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
asyncio.run(main())
|
|
@ -2,13 +2,18 @@
|
||||||
import argparse
|
import argparse
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
import tempfile
|
import sys
|
||||||
import json
|
import json
|
||||||
import shutil
|
import shutil
|
||||||
import asyncio
|
import asyncio
|
||||||
import hashlib
|
import hashlib
|
||||||
import struct
|
import struct
|
||||||
import pytz
|
import pytz
|
||||||
|
from os.path import join, dirname, abspath
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
from sqlalchemy import delete, select
|
||||||
|
from sqlalchemy.orm import joinedload
|
||||||
|
|
||||||
from obs.face.importer import ImportMeasurementsCsv
|
from obs.face.importer import ImportMeasurementsCsv
|
||||||
from obs.face.geojson import ExportMeasurements
|
from obs.face.geojson import ExportMeasurements
|
||||||
|
@ -24,9 +29,9 @@ from obs.face.filter import (
|
||||||
RequiredFieldsFilter,
|
RequiredFieldsFilter,
|
||||||
)
|
)
|
||||||
from obs.face.osm import DataSource, DatabaseTileSource, OverpassTileSource
|
from obs.face.osm import DataSource, DatabaseTileSource, OverpassTileSource
|
||||||
from sqlalchemy import delete
|
|
||||||
|
|
||||||
from obs.face.db import make_session, connect_db, OvertakingEvent, async_session
|
from obs.api.db import make_session, connect_db, OvertakingEvent, async_session, Track
|
||||||
|
from obs.api.app import app
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
@ -40,178 +45,207 @@ async def main():
|
||||||
)
|
)
|
||||||
|
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-i", "--input", required=True, action="store", help="path to input CSV file"
|
"--loop-delay",
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"-o", "--output", required=True, action="store", help="path to output directory"
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"--path-cache",
|
|
||||||
action="store",
|
action="store",
|
||||||
default=None,
|
type=int,
|
||||||
dest="cache_dir",
|
default=10,
|
||||||
help="path where the visualization data will be stored",
|
help="delay between loops, if no track was found in the queue (polling)",
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"--settings",
|
|
||||||
dest="settings_file",
|
|
||||||
required=True,
|
|
||||||
default=None,
|
|
||||||
help="path to track settings file",
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# https://www.postgresql.org/docs/current/libpq-connect.html#LIBPQ-CONNSTRING
|
|
||||||
postgres_url_default = os.environ.get("POSTGRES_URL")
|
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--postgres-url",
|
"tracks",
|
||||||
required=not postgres_url_default,
|
metavar="ID_OR_SLUG",
|
||||||
action="store",
|
nargs="*",
|
||||||
help="connection string for postgres database",
|
help="ID or slug of tracks to process, if not passed, the queue is processed in a loop",
|
||||||
default=postgres_url_default,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
async with connect_db(args.postgres_url):
|
async with connect_db(app.config.POSTGRES_URL):
|
||||||
if args.cache_dir is None:
|
async with make_session() as session:
|
||||||
with tempfile.TemporaryDirectory() as cache_dir:
|
log.info("Loading OpenStreetMap data")
|
||||||
args.cache_dir = cache_dir
|
tile_source = DatabaseTileSource(async_session.get())
|
||||||
await process(args)
|
# tile_source = OverpassTileSource(app.config.OBS_FACE_CACHE_DIR)
|
||||||
|
data_source = DataSource(tile_source)
|
||||||
|
|
||||||
|
if args.tracks:
|
||||||
|
await process_tracks(session, data_source, args.tracks)
|
||||||
|
else:
|
||||||
|
await process_tracks_loop(session, data_source, args.loop_delay)
|
||||||
|
|
||||||
|
|
||||||
|
async def process_tracks_loop(session, data_source, delay):
|
||||||
|
while True:
|
||||||
|
track = (
|
||||||
|
await session.execute(
|
||||||
|
select(Track)
|
||||||
|
.where(Track.processing_status == "queued")
|
||||||
|
.order_by(Track.processing_queued_at)
|
||||||
|
.options(joinedload(Track.author))
|
||||||
|
)
|
||||||
|
).scalar()
|
||||||
|
|
||||||
|
if track is None:
|
||||||
|
await asyncio.sleep(delay)
|
||||||
else:
|
else:
|
||||||
await process(args)
|
try:
|
||||||
|
await process_track(session, track, data_source)
|
||||||
|
except:
|
||||||
|
log.exception("Failed to process track %s. Will continue.", track.slug)
|
||||||
|
|
||||||
|
|
||||||
async def process(args):
|
async def process_tracks(session, data_source, tracks):
|
||||||
log.info("Loading OpenStreetMap data")
|
"""
|
||||||
tile_source = DatabaseTileSource(async_session.get())
|
Processes the tracks and writes event data to the database.
|
||||||
# tile_source = OverpassTileSource(args.cache_dir)
|
|
||||||
data_source = DataSource(tile_source)
|
|
||||||
|
|
||||||
filename_input = os.path.abspath(args.input)
|
:param tracks: A list of strings which
|
||||||
dataset_id = os.path.splitext(os.path.basename(args.input))[0]
|
"""
|
||||||
|
for track_id_or_slug in tracks:
|
||||||
os.makedirs(args.output, exist_ok=True)
|
track = (
|
||||||
|
await session.execute(
|
||||||
log.info("Loading settings")
|
select(Track)
|
||||||
settings_path = os.path.abspath(args.settings_file)
|
.where(
|
||||||
with open(settings_path, "rt") as f:
|
Track.id == track_id_or_slug
|
||||||
settings = json.load(f)
|
if isinstance(track_id_or_slug, int)
|
||||||
|
else Track.slug == track_id_or_slug
|
||||||
settings_output_path = os.path.abspath(
|
|
||||||
os.path.join(args.output, "track-settings.json")
|
|
||||||
)
|
|
||||||
if settings_path != settings_output_path:
|
|
||||||
log.info("Copy settings to output directory")
|
|
||||||
shutil.copyfile(settings_path, settings_output_path)
|
|
||||||
|
|
||||||
log.info("Annotating and filtering CSV file")
|
|
||||||
imported_data, statistics = ImportMeasurementsCsv().read(
|
|
||||||
filename_input,
|
|
||||||
user_id="dummy",
|
|
||||||
dataset_id=dataset_id,
|
|
||||||
)
|
|
||||||
|
|
||||||
input_data = await AnnotateMeasurements(
|
|
||||||
data_source, cache_dir=args.cache_dir
|
|
||||||
).annotate(imported_data)
|
|
||||||
|
|
||||||
filters_from_settings = []
|
|
||||||
for filter_description in settings.get("filters", []):
|
|
||||||
filter_type = filter_description.get("type")
|
|
||||||
if filter_type == "PrivacyZonesFilter":
|
|
||||||
privacy_zones = [
|
|
||||||
PrivacyZone(
|
|
||||||
latitude=zone.get("latitude"),
|
|
||||||
longitude=zone.get("longitude"),
|
|
||||||
radius=zone.get("radius"),
|
|
||||||
)
|
)
|
||||||
for zone in filter_description.get("config", {}).get("privacyZones", [])
|
.options(joinedload(Track.author))
|
||||||
]
|
)
|
||||||
filters_from_settings.append(PrivacyZonesFilter(privacy_zones))
|
).scalar()
|
||||||
else:
|
|
||||||
log.warning("Ignoring unknown filter type %r in settings file", filter_type)
|
|
||||||
|
|
||||||
track_filter = ChainFilter(
|
if not track:
|
||||||
RequiredFieldsFilter(),
|
raise ValueError(f"Track {track_id_or_slug!r} not found.")
|
||||||
PrivacyFilter(
|
|
||||||
user_id_mode=AnonymizationMode.REMOVE,
|
|
||||||
measurement_id_mode=AnonymizationMode.REMOVE,
|
|
||||||
),
|
|
||||||
*filters_from_settings,
|
|
||||||
)
|
|
||||||
measurements_filter = DistanceMeasuredFilter()
|
|
||||||
overtaking_events_filter = ConfirmedFilter()
|
|
||||||
|
|
||||||
track_points = track_filter.filter(input_data, log=log)
|
await process_track(session, track, data_source)
|
||||||
measurements = measurements_filter.filter(track_points, log=log)
|
|
||||||
overtaking_events = overtaking_events_filter.filter(measurements, log=log)
|
|
||||||
|
|
||||||
exporter = ExportMeasurements("measurements.dummy")
|
|
||||||
await exporter.add_measurements(measurements)
|
|
||||||
measurements_json = exporter.get_data()
|
|
||||||
del exporter
|
|
||||||
|
|
||||||
exporter = ExportMeasurements("overtaking_events.dummy")
|
def to_naive_utc(t):
|
||||||
await exporter.add_measurements(overtaking_events)
|
if t is None:
|
||||||
overtaking_events_json = exporter.get_data()
|
return None
|
||||||
del exporter
|
return t.astimezone(pytz.UTC).replace(tzinfo=None)
|
||||||
|
|
||||||
track_json = {
|
|
||||||
"type": "Feature",
|
|
||||||
"geometry": {
|
|
||||||
"type": "LineString",
|
|
||||||
"coordinates": [[m["latitude"], m["longitude"]] for m in track_points],
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
statistics_json = {
|
async def process_track(session, track, data_source):
|
||||||
"recordedAt": statistics["t_min"].isoformat()
|
try:
|
||||||
if statistics["t_min"] is not None
|
track.processing_status = "complete"
|
||||||
else None,
|
track.processed_at = datetime.utcnow()
|
||||||
"recordedUntil": statistics["t_max"].isoformat()
|
|
||||||
if statistics["t_max"] is not None
|
|
||||||
else None,
|
|
||||||
"duration": statistics["t"],
|
|
||||||
"length": statistics["d"],
|
|
||||||
"segments": statistics["n_segments"],
|
|
||||||
"numEvents": statistics["n_confirmed"],
|
|
||||||
"numMeasurements": statistics["n_measurements"],
|
|
||||||
"numValid": statistics["n_valid"],
|
|
||||||
}
|
|
||||||
|
|
||||||
for output_filename, data in [
|
|
||||||
("measurements.json", measurements_json),
|
|
||||||
("overtakingEvents.json", overtaking_events_json),
|
|
||||||
("track.json", track_json),
|
|
||||||
("statistics.json", statistics_json),
|
|
||||||
]:
|
|
||||||
with open(os.path.join(args.output, output_filename), "w") as fp:
|
|
||||||
json.dump(data, fp, indent=4)
|
|
||||||
|
|
||||||
log.info("Importing to database.")
|
|
||||||
async with make_session() as session:
|
|
||||||
await clear_track_data(session, settings["trackId"])
|
|
||||||
await import_overtaking_events(session, settings["trackId"], overtaking_events)
|
|
||||||
await session.commit()
|
await session.commit()
|
||||||
|
|
||||||
|
original_file_path = track.get_original_file_path(app.config)
|
||||||
|
|
||||||
|
output_dir = join(
|
||||||
|
app.config.PROCESSING_OUTPUT_DIR, track.author.username, track.slug
|
||||||
|
)
|
||||||
|
os.makedirs(output_dir, exist_ok=True)
|
||||||
|
|
||||||
|
log.info("Annotating and filtering CSV file")
|
||||||
|
imported_data, statistics = ImportMeasurementsCsv().read(
|
||||||
|
original_file_path,
|
||||||
|
user_id="dummy", # TODO: user username or id or nothing?
|
||||||
|
dataset_id=Track.slug, # TODO: use track id or slug or nothing?
|
||||||
|
)
|
||||||
|
|
||||||
|
annotator = AnnotateMeasurements(
|
||||||
|
data_source, cache_dir=app.config.OBS_FACE_CACHE_DIR
|
||||||
|
)
|
||||||
|
input_data = await annotator.annotate(imported_data)
|
||||||
|
|
||||||
|
track_filter = ChainFilter(
|
||||||
|
RequiredFieldsFilter(),
|
||||||
|
PrivacyFilter(
|
||||||
|
user_id_mode=AnonymizationMode.REMOVE,
|
||||||
|
measurement_id_mode=AnonymizationMode.REMOVE,
|
||||||
|
),
|
||||||
|
# TODO: load user privacy zones and create a PrivacyZonesFilter() from them
|
||||||
|
)
|
||||||
|
measurements_filter = DistanceMeasuredFilter()
|
||||||
|
overtaking_events_filter = ConfirmedFilter()
|
||||||
|
|
||||||
|
track_points = track_filter.filter(input_data, log=log)
|
||||||
|
measurements = measurements_filter.filter(track_points, log=log)
|
||||||
|
overtaking_events = overtaking_events_filter.filter(measurements, log=log)
|
||||||
|
|
||||||
|
exporter = ExportMeasurements("measurements.dummy")
|
||||||
|
await exporter.add_measurements(measurements)
|
||||||
|
measurements_json = exporter.get_data()
|
||||||
|
del exporter
|
||||||
|
|
||||||
|
exporter = ExportMeasurements("overtaking_events.dummy")
|
||||||
|
await exporter.add_measurements(overtaking_events)
|
||||||
|
overtaking_events_json = exporter.get_data()
|
||||||
|
del exporter
|
||||||
|
|
||||||
|
track_json = {
|
||||||
|
"type": "Feature",
|
||||||
|
"geometry": {
|
||||||
|
"type": "LineString",
|
||||||
|
"coordinates": [[m["latitude"], m["longitude"]] for m in track_points],
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for output_filename, data in [
|
||||||
|
("measurements.json", measurements_json),
|
||||||
|
("overtakingEvents.json", overtaking_events_json),
|
||||||
|
("track.json", track_json),
|
||||||
|
]:
|
||||||
|
target = join(output_dir, output_filename)
|
||||||
|
log.debug("Writing file %s", target)
|
||||||
|
with open(target, "w") as fp:
|
||||||
|
json.dump(data, fp, indent=4)
|
||||||
|
|
||||||
|
log.info("Import events into database...")
|
||||||
|
await clear_track_data(session, track)
|
||||||
|
await import_overtaking_events(session, track, overtaking_events)
|
||||||
|
|
||||||
|
log.info("Write track statistics and update status...")
|
||||||
|
track.recorded_at = to_naive_utc(statistics["t_min"])
|
||||||
|
track.recorded_until = to_naive_utc(statistics["t_max"])
|
||||||
|
track.duration = statistics["t"]
|
||||||
|
track.length = statistics["d"]
|
||||||
|
track.segments = statistics["n_segments"]
|
||||||
|
track.num_events = statistics["n_confirmed"]
|
||||||
|
track.num_measurements = statistics["n_measurements"]
|
||||||
|
track.num_valid = statistics["n_valid"]
|
||||||
|
track.processing_status = "complete"
|
||||||
|
track.processed_at = datetime.utcnow()
|
||||||
|
await session.commit()
|
||||||
|
|
||||||
|
log.info("Track %s imported.", track.slug)
|
||||||
|
except BaseException as e:
|
||||||
|
await clear_track_data(session, track)
|
||||||
|
track.processing_status = "error"
|
||||||
|
track.processing_log = str(e)
|
||||||
|
track.processed_at = datetime.utcnow()
|
||||||
|
|
||||||
|
await session.commit()
|
||||||
|
raise
|
||||||
|
|
||||||
|
|
||||||
|
async def clear_track_data(session, track):
|
||||||
|
track.recorded_at = None
|
||||||
|
track.recorded_until = None
|
||||||
|
track.duration = None
|
||||||
|
track.length = None
|
||||||
|
track.segments = None
|
||||||
|
track.num_events = None
|
||||||
|
track.num_measurements = None
|
||||||
|
track.num_valid = None
|
||||||
|
|
||||||
async def clear_track_data(session, track_id):
|
|
||||||
await session.execute(
|
await session.execute(
|
||||||
delete(OvertakingEvent).where(OvertakingEvent.track_id == track_id)
|
delete(OvertakingEvent).where(OvertakingEvent.track_id == track.id)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
async def import_overtaking_events(session, track_id, overtaking_events):
|
async def import_overtaking_events(session, track, overtaking_events):
|
||||||
event_models = []
|
event_models = []
|
||||||
for m in overtaking_events:
|
for m in overtaking_events:
|
||||||
sha = hashlib.sha256()
|
hex_hash = hashlib.sha256(
|
||||||
sha.update(track_id.encode("utf-8"))
|
struct.pack("QQ", track.id, int(m["time"].timestamp()))
|
||||||
sha.update(struct.pack("Q", int(m["time"].timestamp())))
|
).hexdigest()
|
||||||
hex_hash = sha.hexdigest()
|
|
||||||
|
|
||||||
event_models.append(
|
event_models.append(
|
||||||
OvertakingEvent(
|
OvertakingEvent(
|
||||||
track_id=track_id,
|
track_id=track.id,
|
||||||
hex_hash=hex_hash,
|
hex_hash=hex_hash,
|
||||||
way_id=m["OSM_way_id"],
|
way_id=m["OSM_way_id"],
|
||||||
direction_reversed=m["OSM_way_orientation"] < 0,
|
direction_reversed=m["OSM_way_orientation"] < 0,
|
||||||
|
|
|
@ -1,19 +0,0 @@
|
||||||
require('../src/db');
|
|
||||||
|
|
||||||
const { Track } = require('../src/models');
|
|
||||||
|
|
||||||
async function main() {
|
|
||||||
for (const track of await Track.find()) {
|
|
||||||
console.log('queuing', track.slug);
|
|
||||||
await track.queueProcessing();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
main()
|
|
||||||
.then(() => {
|
|
||||||
process.exit(1);
|
|
||||||
})
|
|
||||||
.catch((e) => {
|
|
||||||
console.error(e);
|
|
||||||
process.exit(1);
|
|
||||||
});
|
|
|
@ -4,7 +4,7 @@ import logging
|
||||||
import os
|
import os
|
||||||
import asyncio
|
import asyncio
|
||||||
|
|
||||||
from obs.face.db import init_models, connect_db
|
from obs.api.db import init_models, connect_db
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
|
@ -16,13 +16,6 @@ services:
|
||||||
- '27017:27017'
|
- '27017:27017'
|
||||||
restart: on-failure
|
restart: on-failure
|
||||||
|
|
||||||
redis:
|
|
||||||
image: redis
|
|
||||||
volumes:
|
|
||||||
- ./local/redis:/data
|
|
||||||
command: redis-server --appendonly yes
|
|
||||||
restart: on-failure
|
|
||||||
|
|
||||||
postgres:
|
postgres:
|
||||||
image: "openmaptiles/postgis:6.0"
|
image: "openmaptiles/postgis:6.0"
|
||||||
environment:
|
environment:
|
||||||
|
@ -37,51 +30,41 @@ services:
|
||||||
api:
|
api:
|
||||||
image: obs-api
|
image: obs-api
|
||||||
build:
|
build:
|
||||||
context: ./api
|
context: ./api/
|
||||||
|
dockerfile: Dockerfile
|
||||||
volumes:
|
volumes:
|
||||||
- ./api/src:/opt/obs/api/src
|
- ./api/obs:/opt/obs/api/obs
|
||||||
|
- ./api/scripts/obs:/opt/obs/scripts/obs
|
||||||
- ./api/tools:/opt/obs/api/tools
|
- ./api/tools:/opt/obs/api/tools
|
||||||
- ./api/scripts/obs:/opt/obs/api/scripts/obs
|
- ./api/config.dev.py:/opt/obs/api/config.py
|
||||||
- ./api/views:/opt/obs/api/views
|
|
||||||
- ./local/api-data:/data
|
- ./local/api-data:/data
|
||||||
- ./api/.migrations.js:/opt/obs/api/.migrations.js
|
|
||||||
- ./api/migrations:/opt/obs/api/migrations/
|
|
||||||
- ./api/config.dev.json:/opt/obs/api/config.json
|
|
||||||
environment:
|
|
||||||
- PORT=3000
|
|
||||||
- MONGODB_URL=mongodb://mongo/obsTest
|
|
||||||
- DATA_DIR=/data
|
|
||||||
- POSTGRES_URL="postgresql+asyncpg://obs:obs@localhost/obs"
|
|
||||||
links:
|
links:
|
||||||
- mongo
|
- postgres
|
||||||
- redis
|
- keycloak
|
||||||
ports:
|
ports:
|
||||||
- '3000:3000'
|
- '3000:3000'
|
||||||
restart: on-failure
|
restart: on-failure
|
||||||
command:
|
command:
|
||||||
- npm
|
- obs-api
|
||||||
- run
|
|
||||||
- dev
|
|
||||||
|
|
||||||
worker:
|
worker:
|
||||||
image: obs-api
|
image: obs-api
|
||||||
build:
|
build:
|
||||||
context: ./api
|
context: ./api/
|
||||||
|
dockerfile: Dockerfile
|
||||||
volumes:
|
volumes:
|
||||||
- ./api/src:/opt/obs/api/src
|
- ./api/obs:/opt/obs/api/obs
|
||||||
|
- ./api/scripts/obs:/opt/obs/scripts/obs
|
||||||
- ./api/tools:/opt/obs/api/tools
|
- ./api/tools:/opt/obs/api/tools
|
||||||
- ./api/scripts/obs:/opt/obs/api/scripts/obs
|
- ./api/config.dev.py:/opt/obs/api/config.py
|
||||||
- ./api/views:/opt/obs/api/views
|
|
||||||
- ./local/api-data:/data
|
- ./local/api-data:/data
|
||||||
- ./api/config.dev.json:/opt/obs/api/config.json
|
|
||||||
environment:
|
|
||||||
- DATA_DIR=/data
|
|
||||||
links:
|
links:
|
||||||
- mongo
|
- postgres
|
||||||
- redis
|
- keycloak
|
||||||
restart: on-failure
|
restart: on-failure
|
||||||
command: 'bash -c "pip install geopy && npm run dev:worker"'
|
command:
|
||||||
|
- python
|
||||||
|
- tools/process_track.py
|
||||||
|
|
||||||
frontend:
|
frontend:
|
||||||
image: obs-frontend
|
image: obs-frontend
|
||||||
|
@ -99,6 +82,7 @@ services:
|
||||||
- PORT=3000
|
- PORT=3000
|
||||||
links:
|
links:
|
||||||
- api
|
- api
|
||||||
|
- tileserver
|
||||||
ports:
|
ports:
|
||||||
- '3001:3000'
|
- '3001:3000'
|
||||||
restart: on-failure
|
restart: on-failure
|
||||||
|
@ -107,8 +91,7 @@ services:
|
||||||
- start
|
- start
|
||||||
|
|
||||||
openmaptiles-tools:
|
openmaptiles-tools:
|
||||||
image: "openmaptiles/openmaptiles-tools:${TOOLS_VERSION}"
|
image: openmaptiles/openmaptiles-tools:6.0
|
||||||
env_file: tile-generator/.env
|
|
||||||
environment:
|
environment:
|
||||||
# Must match the version of this file (first line)
|
# Must match the version of this file (first line)
|
||||||
# download-osm will use it when generating a composer file
|
# download-osm will use it when generating a composer file
|
||||||
|
@ -138,13 +121,11 @@ services:
|
||||||
- ./tile-generator/cache:/cache
|
- ./tile-generator/cache:/cache
|
||||||
|
|
||||||
generate-vectortiles:
|
generate-vectortiles:
|
||||||
image: "openmaptiles/generate-vectortiles:${TOOLS_VERSION}"
|
image: openmaptiles/generate-vectortiles:6.0
|
||||||
volumes:
|
volumes:
|
||||||
- ./tile-generator/data:/export
|
- ./tile-generator/data:/export
|
||||||
- ./tile-generator/build/openmaptiles.tm2source:/tm2source
|
- ./tile-generator/build/openmaptiles.tm2source:/tm2source
|
||||||
env_file: tile-generator/.env
|
|
||||||
environment:
|
environment:
|
||||||
FILTER_MAPNIK_OUTPUT: ${FILTER_MAPNIK_OUTPUT}
|
|
||||||
MBTILES_NAME: ${MBTILES_FILE}
|
MBTILES_NAME: ${MBTILES_FILE}
|
||||||
BBOX: ${BBOX}
|
BBOX: ${BBOX}
|
||||||
MIN_ZOOM: ${MIN_ZOOM}
|
MIN_ZOOM: ${MIN_ZOOM}
|
||||||
|
@ -168,4 +149,20 @@ services:
|
||||||
command:
|
command:
|
||||||
- --config
|
- --config
|
||||||
- /config/tileserver.json
|
- /config/tileserver.json
|
||||||
# - /data/tiles.mbtiles
|
|
||||||
|
keycloak:
|
||||||
|
image: jboss/keycloak
|
||||||
|
ports:
|
||||||
|
- 3003:8080
|
||||||
|
links:
|
||||||
|
- postgres
|
||||||
|
environment:
|
||||||
|
KEYCLOAK_USER: admin
|
||||||
|
KEYCLOAK_PASSWORD: admin
|
||||||
|
KEYCLOAK_FRONTEND_URL: http://localhost:3003/auth/
|
||||||
|
DB_VENDOR: postgres
|
||||||
|
DB_ADDR: postgres
|
||||||
|
DB_DATABASE: obs
|
||||||
|
DB_USER: obs
|
||||||
|
DB_PASSWORD: obs
|
||||||
|
# DB_SCHEMA: keycloak
|
||||||
|
|
|
@ -1,11 +1,5 @@
|
||||||
{
|
{
|
||||||
"apiUrl": "http://localhost:3001",
|
"apiUrl": "http://localhost:3000",
|
||||||
"auth": {
|
|
||||||
"server": "http://localhost:3000",
|
|
||||||
"clientId": "b730f8d2-d93c-4c68-9ff0-dfac8da76ee2",
|
|
||||||
"scope": "*",
|
|
||||||
"redirectUri": "http://localhost:3001/redirect"
|
|
||||||
},
|
|
||||||
"imprintUrl": "https://example.com/imprint",
|
"imprintUrl": "https://example.com/imprint",
|
||||||
"privacyPolicyUrl": "https://example.com/privacy",
|
"privacyPolicyUrl": "https://example.com/privacy",
|
||||||
"mapTileset": {
|
"mapTileset": {
|
||||||
|
|
|
@ -42,6 +42,10 @@ const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
const config = useConfig()
|
const config = useConfig()
|
||||||
const apiVersion = useObservable(() => from(api.get('/info')).pipe(pluck('version')))
|
const apiVersion = useObservable(() => from(api.get('/info')).pipe(pluck('version')))
|
||||||
|
|
||||||
|
React.useEffect(() => {
|
||||||
|
api.loadUser()
|
||||||
|
}, [])
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Router basename={process.env.PUBLIC_URL || '/'}>
|
<Router basename={process.env.PUBLIC_URL || '/'}>
|
||||||
<Menu fixed="top">
|
<Menu fixed="top">
|
||||||
|
|
|
@ -1,9 +1,7 @@
|
||||||
import {stringifyParams} from 'query'
|
import {stringifyParams} from 'query'
|
||||||
import globalStore from 'store'
|
import globalStore from 'store'
|
||||||
import {setAuth, invalidateAccessToken, resetAuth} from 'reducers/auth'
|
|
||||||
import {setLogin} from 'reducers/login'
|
import {setLogin} from 'reducers/login'
|
||||||
import configPromise from 'config'
|
import configPromise from 'config'
|
||||||
import {create as createPkce} from 'pkce'
|
|
||||||
import download from 'downloadjs'
|
import download from 'downloadjs'
|
||||||
|
|
||||||
function getFileNameFromContentDispostionHeader(contentDisposition: string): string | undefined {
|
function getFileNameFromContentDispostionHeader(contentDisposition: string): string | undefined {
|
||||||
|
@ -29,196 +27,42 @@ class RequestError extends Error {
|
||||||
class API {
|
class API {
|
||||||
constructor(store) {
|
constructor(store) {
|
||||||
this.store = store
|
this.store = store
|
||||||
this._getValidAccessTokenPromise = null
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
async loadUser() {
|
||||||
* Fetches or directly returns from cache the metadata information from the
|
try {
|
||||||
* authorization server, according to https://tools.ietf.org/html/rfc8414.
|
const result = await this.get('/user')
|
||||||
* Also validates compatibility with this metadata server, i.e. checking that
|
this.store.dispatch(setLogin(result))
|
||||||
* it supports PKCE.
|
} catch {
|
||||||
*/
|
this.store.dispatch(setLogin(null))
|
||||||
async getAuthorizationServerMetadata() {
|
|
||||||
const config = await configPromise
|
|
||||||
const url = new URL(config.auth.server)
|
|
||||||
const pathSuffix = url.pathname.replace(/^\/+|\/+$/, '')
|
|
||||||
url.pathname = '/.well-known/oauth-authorization-server' + (pathSuffix ? '/' + pathSuffix : '')
|
|
||||||
|
|
||||||
const response = await window.fetch(url.toString())
|
|
||||||
const metadata = await response.json()
|
|
||||||
|
|
||||||
const {
|
|
||||||
authorization_endpoint: authorizationEndpoint,
|
|
||||||
token_endpoint: tokenEndpoint,
|
|
||||||
response_types_supported: responseTypesSupported,
|
|
||||||
code_challenge_methods_supported: codeChallengeMethodsSupported,
|
|
||||||
} = metadata
|
|
||||||
if (!authorizationEndpoint) {
|
|
||||||
throw new Error('No authorization endpoint')
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!authorizationEndpoint.startsWith(config.auth.server)) {
|
|
||||||
throw new Error('Invalid authorization endpoint')
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!tokenEndpoint) {
|
|
||||||
throw new Error('No token endpoint')
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!tokenEndpoint.startsWith(config.auth.server)) {
|
|
||||||
throw new Error('Invalid token endpoint')
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!Array.isArray(responseTypesSupported) || !responseTypesSupported.includes('code')) {
|
|
||||||
throw new Error('Authorization code flow not supported or no support advertised.')
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!Array.isArray(codeChallengeMethodsSupported) || !codeChallengeMethodsSupported.includes('S256')) {
|
|
||||||
throw new Error('PKCE with S256 not supported or no support advertised.')
|
|
||||||
}
|
|
||||||
|
|
||||||
return {authorizationEndpoint, tokenEndpoint}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return an access token, if it is (still) valid. If not, and a refresh
|
|
||||||
* token exists, use the refresh token to issue a new access token. If that
|
|
||||||
* fails, or neither is available, return `null`. This should usually result
|
|
||||||
* in a redirect to login.
|
|
||||||
*/
|
|
||||||
async getValidAccessToken() {
|
|
||||||
// prevent multiple parallel refresh processes
|
|
||||||
if (this._getValidAccessTokenPromise) {
|
|
||||||
return await this._getValidAccessTokenPromise
|
|
||||||
} else {
|
|
||||||
this._getValidAccessTokenPromise = this._getValidAccessToken()
|
|
||||||
const result = await this._getValidAccessTokenPromise
|
|
||||||
this._getValidAccessTokenPromise = null
|
|
||||||
return result
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async _getValidAccessToken() {
|
|
||||||
let {auth} = this.store.getState()
|
|
||||||
|
|
||||||
if (!auth) {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
|
|
||||||
const {tokenType, accessToken, refreshToken, expiresAt} = auth
|
|
||||||
|
|
||||||
// access token is valid
|
|
||||||
if (accessToken && expiresAt > new Date().getTime()) {
|
|
||||||
return `${tokenType} ${accessToken}`
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!refreshToken) {
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
|
|
||||||
// Try to use the refresh token
|
|
||||||
const {tokenEndpoint} = await this.getAuthorizationServerMetadata()
|
|
||||||
const config = await configPromise
|
|
||||||
const url = new URL(tokenEndpoint)
|
|
||||||
url.searchParams.append('refresh_token', refreshToken)
|
|
||||||
url.searchParams.append('grant_type', 'refresh_token')
|
|
||||||
url.searchParams.append('scope', config.auth.scope)
|
|
||||||
const response = await window.fetch(url.toString())
|
|
||||||
const json = await response.json()
|
|
||||||
|
|
||||||
if (response.status === 200 && json != null && json.error == null) {
|
|
||||||
auth = this.getAuthFromTokenResponse(json)
|
|
||||||
this.store.dispatch(setAuth(auth))
|
|
||||||
return `${auth.tokenType} ${auth.accessToken}`
|
|
||||||
} else {
|
|
||||||
console.warn('Could not use refresh token, error response:', json)
|
|
||||||
this.store.dispatch(resetAuth())
|
|
||||||
return null
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async exchangeAuthorizationCode(code) {
|
|
||||||
const codeVerifier = localStorage.getItem('codeVerifier')
|
|
||||||
if (!codeVerifier) {
|
|
||||||
throw new Error('No code verifier found')
|
|
||||||
}
|
|
||||||
|
|
||||||
const {tokenEndpoint} = await this.getAuthorizationServerMetadata()
|
|
||||||
const config = await configPromise
|
|
||||||
const url = new URL(tokenEndpoint)
|
|
||||||
url.searchParams.append('code', code)
|
|
||||||
url.searchParams.append('grant_type', 'authorization_code')
|
|
||||||
url.searchParams.append('client_id', config.auth.clientId)
|
|
||||||
url.searchParams.append('redirect_uri', config.auth.redirectUri)
|
|
||||||
url.searchParams.append('code_verifier', codeVerifier)
|
|
||||||
const response = await window.fetch(url.toString())
|
|
||||||
const json = await response.json()
|
|
||||||
|
|
||||||
if (json.error) {
|
|
||||||
return json
|
|
||||||
}
|
|
||||||
|
|
||||||
const auth = api.getAuthFromTokenResponse(json)
|
|
||||||
this.store.dispatch(setAuth(auth))
|
|
||||||
|
|
||||||
const {user} = await this.get('/user')
|
|
||||||
this.store.dispatch(setLogin(user))
|
|
||||||
|
|
||||||
return true
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async logout() {
|
async logout() {
|
||||||
// 1. Tell the store to forget that we're logged in.
|
const config = await configPromise
|
||||||
this.store.dispatch(resetAuth())
|
const url = new URL(config.apiUrl + '/logout')
|
||||||
|
url.searchParams.append('next', window.location.href) // bring us back to the current page
|
||||||
// 2. Log out session in API.
|
|
||||||
const {tokenEndpoint} = await this.getAuthorizationServerMetadata()
|
|
||||||
const url = new URL(tokenEndpoint.replace(/\/token$/, '/logout'))
|
|
||||||
url.searchParams.append('redirectTo', window.location.href) // bring us back to the current page
|
|
||||||
|
|
||||||
window.location.href = url.toString()
|
window.location.href = url.toString()
|
||||||
}
|
}
|
||||||
|
|
||||||
async makeLoginUrl() {
|
async makeLoginUrl() {
|
||||||
const {authorizationEndpoint} = await this.getAuthorizationServerMetadata()
|
|
||||||
const config = await configPromise
|
const config = await configPromise
|
||||||
|
const url = new URL(config.apiUrl + '/login')
|
||||||
const {codeVerifier, codeChallenge} = createPkce()
|
url.searchParams.append('next', window.location.href) // bring us back to the current page
|
||||||
localStorage.setItem('codeVerifier', codeVerifier)
|
return url.toString()
|
||||||
|
|
||||||
const loginUrl = new URL(authorizationEndpoint)
|
|
||||||
loginUrl.searchParams.append('client_id', config.auth.clientId)
|
|
||||||
loginUrl.searchParams.append('scope', config.auth.scope)
|
|
||||||
loginUrl.searchParams.append('redirect_uri', config.auth.redirectUri)
|
|
||||||
loginUrl.searchParams.append('response_type', 'code')
|
|
||||||
loginUrl.searchParams.append('code_challenge', codeChallenge)
|
|
||||||
loginUrl.searchParams.append('code_challenge_method', 'S256')
|
|
||||||
|
|
||||||
// TODO: Implement PKCE
|
|
||||||
|
|
||||||
return loginUrl.toString()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async fetch(url, options = {}) {
|
async fetch(url, options = {}) {
|
||||||
const accessToken = await this.getValidAccessToken()
|
|
||||||
const config = await configPromise
|
const config = await configPromise
|
||||||
|
|
||||||
const {returnResponse = false, ...fetchOptions} = options
|
const {returnResponse = false, ...fetchOptions} = options
|
||||||
|
|
||||||
const response = await window.fetch(config.apiUrl + '/api' + url, {
|
const response = await window.fetch(config.apiUrl + url, {
|
||||||
...fetchOptions,
|
...fetchOptions,
|
||||||
headers: {
|
credentials: 'include',
|
||||||
...(fetchOptions.headers || {}),
|
|
||||||
Authorization: accessToken,
|
|
||||||
},
|
|
||||||
})
|
})
|
||||||
|
|
||||||
if (response.status === 401) {
|
if (response.status === 401) {
|
||||||
// Unset login, since 401 means that we're not logged in. On the next
|
|
||||||
// request with `getValidAccessToken()`, this will be detected and the
|
|
||||||
// refresh token is used (if still valid).
|
|
||||||
this.store.dispatch(invalidateAccessToken())
|
|
||||||
|
|
||||||
throw new Error('401 Unauthorized')
|
throw new Error('401 Unauthorized')
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -278,16 +122,6 @@ class API {
|
||||||
return await this.post(url, {...options, method: 'put'})
|
return await this.post(url, {...options, method: 'put'})
|
||||||
}
|
}
|
||||||
|
|
||||||
getAuthFromTokenResponse(tokenResponse) {
|
|
||||||
return {
|
|
||||||
tokenType: tokenResponse.token_type,
|
|
||||||
accessToken: tokenResponse.access_token,
|
|
||||||
refreshToken: tokenResponse.refresh_token,
|
|
||||||
expiresAt: new Date().getTime() + tokenResponse.expires_in * 1000,
|
|
||||||
scope: tokenResponse.scope,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async downloadFile(url, options = {}) {
|
async downloadFile(url, options = {}) {
|
||||||
const res = await this.fetch(url, {returnResponse: true, ...options})
|
const res = await this.fetch(url, {returnResponse: true, ...options})
|
||||||
const blob = await res.blob()
|
const blob = await res.blob()
|
||||||
|
|
|
@ -7,9 +7,11 @@ export default function LoginButton(props) {
|
||||||
const [busy, setBusy] = React.useState(false)
|
const [busy, setBusy] = React.useState(false)
|
||||||
|
|
||||||
const onClick = React.useCallback(async (e) => {
|
const onClick = React.useCallback(async (e) => {
|
||||||
|
console.log('aa')
|
||||||
e.preventDefault()
|
e.preventDefault()
|
||||||
setBusy(true)
|
setBusy(true)
|
||||||
const url = await api.makeLoginUrl()
|
const url = await api.makeLoginUrl()
|
||||||
|
console.log('go', url)
|
||||||
window.location.href = url
|
window.location.href = url
|
||||||
setBusy(false)
|
setBusy(false)
|
||||||
}, [setBusy])
|
}, [setBusy])
|
||||||
|
|
|
@ -6,7 +6,7 @@ import styles from './Page.module.scss'
|
||||||
|
|
||||||
export default function Page({small, children, fullScreen}: {small?: boolean, children: ReactNode, fullScreen?: boolean}) {
|
export default function Page({small, children, fullScreen}: {small?: boolean, children: ReactNode, fullScreen?: boolean}) {
|
||||||
return (
|
return (
|
||||||
<main className={classnames({page: true, small, fullScreen})}>
|
<main className={classnames(styles.page, small && styles.small, fullScreen && styles.fullScreen)}>
|
||||||
{fullScreen ? children : <Container>{children}</Container>}
|
{fullScreen ? children : <Container>{children}</Container>}
|
||||||
</main>
|
</main>
|
||||||
)
|
)
|
||||||
|
|
|
@ -8,6 +8,13 @@ export interface Config {
|
||||||
zoom: number
|
zoom: number
|
||||||
}
|
}
|
||||||
obsMapSource?: string
|
obsMapSource?: string
|
||||||
|
imprintUrl?: string
|
||||||
|
privacyPolicyUrl?: string
|
||||||
|
mapTileset?: {
|
||||||
|
url?: string
|
||||||
|
minZoom?: number
|
||||||
|
maxZoom?: number
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async function loadConfig(): Promise<Config> {
|
async function loadConfig(): Promise<Config> {
|
||||||
|
|
|
@ -10,8 +10,10 @@ const LogoutPage = connect(
|
||||||
)(function LogoutPage({loggedIn}) {
|
)(function LogoutPage({loggedIn}) {
|
||||||
React.useEffect(() => {
|
React.useEffect(() => {
|
||||||
// no await, just trigger it
|
// no await, just trigger it
|
||||||
api.logout()
|
if (loggedIn) {
|
||||||
})
|
api.logout()
|
||||||
|
}
|
||||||
|
}, [loggedIn])
|
||||||
|
|
||||||
return loggedIn ? <Loader active /> : <Redirect to="/" />
|
return loggedIn ? <Loader active /> : <Redirect to="/" />
|
||||||
})
|
})
|
||||||
|
|
|
@ -37,10 +37,10 @@ export default function TrackDetails({track, isAuthor}) {
|
||||||
</List.Item>
|
</List.Item>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
{track.statistics?.duration != null && (
|
{track.duration != null && (
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>Duration</List.Header>
|
<List.Header>Duration</List.Header>
|
||||||
{formatDuration(track.statistics.duration)}
|
{formatDuration(track.duration)}
|
||||||
</List.Item>
|
</List.Item>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
|
@ -51,24 +51,32 @@ export default function TrackDetails({track, isAuthor}) {
|
||||||
</List.Item>
|
</List.Item>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
{track?.statistics?.recordedAt != null && (
|
{track?.recordedAt != null && (
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>Recorded on</List.Header>
|
<List.Header>Recorded on</List.Header>
|
||||||
<FormattedDate date={track?.statistics.recordedAt} />
|
<FormattedDate date={track?.recordedAt} />
|
||||||
</List.Item>
|
</List.Item>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
{track?.statistics?.numEvents != null && (
|
{track?.numEvents != null && (
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>Confirmed events</List.Header>
|
<List.Header>Confirmed events</List.Header>
|
||||||
{track?.statistics.numEvents}
|
{track?.numEvents}
|
||||||
</List.Item>
|
</List.Item>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
{track?.statistics?.trackLength != null && (
|
{track?.length != null && (
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>Length</List.Header>
|
<List.Header>Length</List.Header>
|
||||||
{(track?.statistics.trackLength / 1000).toFixed(2)} km
|
{(track?.length / 1000).toFixed(2)} km
|
||||||
|
</List.Item>
|
||||||
|
)}
|
||||||
|
|
||||||
|
|
||||||
|
{track?.processingStatus != null && (
|
||||||
|
<List.Item>
|
||||||
|
<List.Header>Processing</List.Header>
|
||||||
|
{track.processingStatus}
|
||||||
</List.Item>
|
</List.Item>
|
||||||
)}
|
)}
|
||||||
</List>
|
</List>
|
||||||
|
|
|
@ -57,7 +57,6 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
of(undefined),
|
of(undefined),
|
||||||
from(api.get(url)).pipe(
|
from(api.get(url)).pipe(
|
||||||
catchError(() => {
|
catchError(() => {
|
||||||
// history.replace('/tracks')
|
|
||||||
return of(null)
|
return of(null)
|
||||||
})
|
})
|
||||||
)
|
)
|
||||||
|
@ -72,7 +71,7 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
switchMap((url) =>
|
switchMap((url) =>
|
||||||
from(api.get(url)).pipe(
|
from(api.get(url)).pipe(
|
||||||
catchError(() => {
|
catchError(() => {
|
||||||
history.replace('/tracks')
|
return of(null)
|
||||||
})
|
})
|
||||||
)
|
)
|
||||||
),
|
),
|
||||||
|
@ -117,8 +116,10 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
|
|
||||||
const {track, trackData, comments} = data || {}
|
const {track, trackData, comments} = data || {}
|
||||||
|
|
||||||
|
console.log({track, trackData})
|
||||||
const loading = track == null || trackData === undefined
|
const loading = track == null || trackData === undefined
|
||||||
const processing = ['processing', 'pending'].includes(track?.processingStatus)
|
const processing = ['processing', 'queued', 'created'].includes(track?.processingStatus)
|
||||||
|
const error = track?.processingStatus === 'error'
|
||||||
|
|
||||||
const [left, setLeft] = React.useState(true)
|
const [left, setLeft] = React.useState(true)
|
||||||
const [right, setRight] = React.useState(false)
|
const [right, setRight] = React.useState(false)
|
||||||
|
@ -135,6 +136,15 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
</Message>
|
</Message>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
|
{error && (
|
||||||
|
<Message error>
|
||||||
|
<Message.Content>
|
||||||
|
The processing of this track failed, please ask your site
|
||||||
|
administrator for help in debugging the issue.
|
||||||
|
</Message.Content>
|
||||||
|
</Message>
|
||||||
|
)}
|
||||||
|
|
||||||
<Grid stackable>
|
<Grid stackable>
|
||||||
<Grid.Row>
|
<Grid.Row>
|
||||||
<Grid.Column width={12}>
|
<Grid.Column width={12}>
|
||||||
|
|
|
@ -94,6 +94,14 @@ function maxLength(t, max) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const COLOR_BY_STATUS = {
|
||||||
|
error: 'red',
|
||||||
|
complete: 'green',
|
||||||
|
created: 'gray',
|
||||||
|
queued: 'orange',
|
||||||
|
processing: 'orange',
|
||||||
|
}
|
||||||
|
|
||||||
export function TrackListItem({track, privateTracks = false}) {
|
export function TrackListItem({track, privateTracks = false}) {
|
||||||
return (
|
return (
|
||||||
<Item key={track.slug}>
|
<Item key={track.slug}>
|
||||||
|
@ -121,6 +129,10 @@ export function TrackListItem({track, privateTracks = false}) {
|
||||||
<Icon name="eye slash" fitted /> Private
|
<Icon name="eye slash" fitted /> Private
|
||||||
</>
|
</>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
|
<span style={{marginLeft: '1em'}}>
|
||||||
|
<Icon color={COLOR_BY_STATUS[track.processingStatus]} name="bolt" fitted /> Processing {track.processingStatus}
|
||||||
|
</span>
|
||||||
</Item.Extra>
|
</Item.Extra>
|
||||||
)}
|
)}
|
||||||
</Item.Content>
|
</Item.Content>
|
||||||
|
|
|
@ -63,6 +63,7 @@ export function FileUploadStatus({
|
||||||
formData.append('body', file)
|
formData.append('body', file)
|
||||||
|
|
||||||
xhr = new XMLHttpRequest()
|
xhr = new XMLHttpRequest()
|
||||||
|
xhr.withCredentials = true
|
||||||
|
|
||||||
const onProgress = (e) => {
|
const onProgress = (e) => {
|
||||||
const progress = (e.loaded || 0) / (e.total || 1)
|
const progress = (e.loaded || 0) / (e.total || 1)
|
||||||
|
@ -79,14 +80,14 @@ export function FileUploadStatus({
|
||||||
|
|
||||||
const config = await configPromise
|
const config = await configPromise
|
||||||
if (slug) {
|
if (slug) {
|
||||||
xhr.open('PUT', `${config.apiUrl}/api/tracks/${slug}`)
|
xhr.open('PUT', `${config.apiUrl}/tracks/${slug}`)
|
||||||
} else {
|
} else {
|
||||||
xhr.open('POST', `${config.apiUrl}/api/tracks`)
|
xhr.open('POST', `${config.apiUrl}/tracks`)
|
||||||
}
|
}
|
||||||
|
|
||||||
const accessToken = await api.getValidAccessToken()
|
// const accessToken = await api.getValidAccessToken()
|
||||||
|
|
||||||
xhr.setRequestHeader('Authorization', accessToken)
|
// xhr.setRequestHeader('Authorization', accessToken)
|
||||||
xhr.send(formData)
|
xhr.send(formData)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,30 +0,0 @@
|
||||||
const initialState = null
|
|
||||||
|
|
||||||
export function setAuth(auth) {
|
|
||||||
return {type: 'AUTH.SET', payload: {auth}}
|
|
||||||
}
|
|
||||||
|
|
||||||
export function resetAuth() {
|
|
||||||
return {type: 'AUTH.RESET'}
|
|
||||||
}
|
|
||||||
|
|
||||||
export function invalidateAccessToken() {
|
|
||||||
return {type: 'AUTH.INVALIDATE_ACCESS_TOKEN'}
|
|
||||||
}
|
|
||||||
|
|
||||||
export default function loginReducer(state = initialState, action) {
|
|
||||||
switch (action.type) {
|
|
||||||
case 'AUTH.SET':
|
|
||||||
return action.payload.auth
|
|
||||||
case 'AUTH.INVALIDATE_ACCESS_TOKEN':
|
|
||||||
return state && {
|
|
||||||
...state,
|
|
||||||
accessToken: null,
|
|
||||||
expiresAt: 0,
|
|
||||||
}
|
|
||||||
case 'AUTH.RESET':
|
|
||||||
return null
|
|
||||||
default:
|
|
||||||
return state
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -1,6 +1,5 @@
|
||||||
import {combineReducers} from 'redux'
|
import {combineReducers} from 'redux'
|
||||||
|
|
||||||
import login from './login'
|
import login from './login'
|
||||||
import auth from './auth'
|
|
||||||
|
|
||||||
export default combineReducers({login, auth})
|
export default combineReducers({login})
|
||||||
|
|
|
@ -4,12 +4,16 @@ export function setLogin(user) {
|
||||||
return {type: 'LOGIN.SET', payload: {user}}
|
return {type: 'LOGIN.SET', payload: {user}}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export function resetLogin() {
|
||||||
|
return {type: 'LOGIN.RESET'}
|
||||||
|
}
|
||||||
|
|
||||||
export default function loginReducer(state = initialState, action) {
|
export default function loginReducer(state = initialState, action) {
|
||||||
switch (action.type) {
|
switch (action.type) {
|
||||||
case 'LOGIN.SET':
|
case 'LOGIN.SET':
|
||||||
return action.payload.user
|
return action.payload.user
|
||||||
|
|
||||||
case 'AUTH.RESET': // cross reducer action :)
|
case 'LOGIN.RESET':
|
||||||
return null
|
return null
|
||||||
|
|
||||||
default:
|
default:
|
||||||
|
|
|
@ -3,7 +3,7 @@ import persistState from 'redux-localstorage'
|
||||||
|
|
||||||
import rootReducer from './reducers'
|
import rootReducer from './reducers'
|
||||||
|
|
||||||
const enhancer = compose(persistState(['login', 'auth']))
|
const enhancer = compose(persistState(['login']))
|
||||||
|
|
||||||
const store = createStore(rootReducer, undefined, enhancer)
|
const store = createStore(rootReducer, undefined, enhancer)
|
||||||
|
|
||||||
|
|
|
@ -12,7 +12,13 @@ export type TrackData = {
|
||||||
overtakingEvents: FeatureCollection,
|
overtakingEvents: FeatureCollection,
|
||||||
}
|
}
|
||||||
|
|
||||||
export type TrackStatistics = {
|
export type Track = {
|
||||||
|
slug: string
|
||||||
|
author: UserProfile
|
||||||
|
title: string
|
||||||
|
description?: string
|
||||||
|
createdAt: string
|
||||||
|
public?: boolean
|
||||||
recordedAt?: Date
|
recordedAt?: Date
|
||||||
recordedUntil?: Date
|
recordedUntil?: Date
|
||||||
duration?: number
|
duration?: number
|
||||||
|
@ -23,16 +29,6 @@ export type TrackStatistics = {
|
||||||
numValid?: number
|
numValid?: number
|
||||||
}
|
}
|
||||||
|
|
||||||
export type Track = {
|
|
||||||
slug: string
|
|
||||||
author: UserProfile
|
|
||||||
title: string
|
|
||||||
description?: string
|
|
||||||
createdAt: string
|
|
||||||
public?: boolean
|
|
||||||
statistics?: TrackStatistics
|
|
||||||
}
|
|
||||||
|
|
||||||
export type TrackPoint = {
|
export type TrackPoint = {
|
||||||
type: 'Feature',
|
type: 'Feature',
|
||||||
geometry: Point,
|
geometry: Point,
|
||||||
|
|
31
old-api/Dockerfile
Normal file
31
old-api/Dockerfile
Normal file
|
@ -0,0 +1,31 @@
|
||||||
|
FROM node:15.14-buster
|
||||||
|
|
||||||
|
# Install python3, pip3, and make them the default for `python` and `pip` commands
|
||||||
|
RUN apt-get update && apt-get install -y python3 python3-pip
|
||||||
|
RUN ln -s $(which python3) /usr/local/bin/python
|
||||||
|
RUN ln -s $(which pip3) /usr/local/bin/pip
|
||||||
|
|
||||||
|
WORKDIR /opt/obs/api
|
||||||
|
|
||||||
|
ADD package.json package-lock.json /opt/obs/api/
|
||||||
|
RUN echo update-notifier=false >> ~/.npmrc
|
||||||
|
RUN npm ci
|
||||||
|
|
||||||
|
ADD scripts /opt/obs/api/scripts/
|
||||||
|
ADD tools /opt/obs/api/tools/
|
||||||
|
ADD requirements.txt /opt/obs/api/
|
||||||
|
RUN pip install -r requirements.txt
|
||||||
|
RUN pip install -e ./scripts
|
||||||
|
|
||||||
|
ADD views /opt/obs/api/views/
|
||||||
|
ADD src /opt/obs/api/src/
|
||||||
|
|
||||||
|
#ADD .migrations.js .
|
||||||
|
#ADD migrations .
|
||||||
|
|
||||||
|
EXPOSE 3000
|
||||||
|
|
||||||
|
ENV PORT=3000
|
||||||
|
ENV DATA_DIR=/data
|
||||||
|
|
||||||
|
CMD ["npm", "run", "start"]
|
3
old-api/requirements.txt
Normal file
3
old-api/requirements.txt
Normal file
|
@ -0,0 +1,3 @@
|
||||||
|
./scripts
|
||||||
|
sqlalchemy[asyncio]
|
||||||
|
asyncpg
|
|
@ -7,7 +7,7 @@ SHELL = /bin/bash
|
||||||
.SHELLFLAGS = -o pipefail -c
|
.SHELLFLAGS = -o pipefail -c
|
||||||
|
|
||||||
# Make all .env variables available for make targets
|
# Make all .env variables available for make targets
|
||||||
include .env
|
include ../.env
|
||||||
|
|
||||||
# Layers definition and meta data
|
# Layers definition and meta data
|
||||||
TILESET_FILE ?= openmaptiles.yaml
|
TILESET_FILE ?= openmaptiles.yaml
|
||||||
|
|
Loading…
Reference in a new issue