Compare commits
4 commits
main
...
ignore-mor
Author | SHA1 | Date | |
---|---|---|---|
bf740a0695 | |||
9d97489ac9 | |||
73c1a62f80 | |||
22ae2c08a6 |
|
@ -1,20 +0,0 @@
|
||||||
name: Build docker image
|
|
||||||
on: [push]
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
build-image:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
container:
|
|
||||||
image: catthehacker/ubuntu:act-latest
|
|
||||||
steps:
|
|
||||||
- name: Login to Forgejo docker registry
|
|
||||||
uses: docker/login-action@v3.0.0
|
|
||||||
with:
|
|
||||||
registry: git.pub.solar
|
|
||||||
username: hakkonaut
|
|
||||||
password: ${{ secrets.GIT_AUTH_TOKEN }}
|
|
||||||
- name: Build and push
|
|
||||||
uses: docker/build-push-action@v5.1.0
|
|
||||||
with:
|
|
||||||
push: true
|
|
||||||
tags: git.pub.solar/pub-solar/obs-portal:latest
|
|
19
.github/dependabot.yml
vendored
Normal file
19
.github/dependabot.yml
vendored
Normal file
|
@ -0,0 +1,19 @@
|
||||||
|
# see also https://docs.github.com/en/code-security/supply-chain-security/keeping-your-dependencies-updated-automatically
|
||||||
|
|
||||||
|
version: 2
|
||||||
|
updates:
|
||||||
|
- package-ecosystem: "docker"
|
||||||
|
directory: "/"
|
||||||
|
schedule:
|
||||||
|
interval: "daily"
|
||||||
|
|
||||||
|
# Disable submodules for the time being - they seem to not respect the branch of scripts we're using.
|
||||||
|
# - package-ecosystem: "gitsubmodule"
|
||||||
|
# directory: "/"
|
||||||
|
# schedule:
|
||||||
|
# interval: "daily"
|
||||||
|
|
||||||
|
- package-ecosystem: "pip"
|
||||||
|
directory: "/api"
|
||||||
|
schedule:
|
||||||
|
interval: "daily"
|
63
CHANGELOG.md
63
CHANGELOG.md
|
@ -1,68 +1,5 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
## 0.8.1
|
|
||||||
|
|
||||||
### Improvements
|
|
||||||
|
|
||||||
* The zone (urban/rural) is now also exported with the events GeoJson export.
|
|
||||||
|
|
||||||
### Bug Fixes
|
|
||||||
|
|
||||||
* Update to a current version of gpstime (python dependency) fix portal startup.
|
|
||||||
|
|
||||||
## 0.8.0
|
|
||||||
|
|
||||||
### Features
|
|
||||||
|
|
||||||
* Bulk actions on users owned tracks (reprocess, download, make private, make public, delete) (#269, #38)
|
|
||||||
* Easy sorting by device for "multi-device users" (e.g. group lending out OBSes)
|
|
||||||
* Region display at higher zoom levels to easily find interesting areas (#112)
|
|
||||||
* Export of road statistics on top of the already-existing event statistics (#341)
|
|
||||||
|
|
||||||
### Improvements
|
|
||||||
|
|
||||||
* Refactored database access to hopefully combat portal crashes (#337)
|
|
||||||
* New infrastructure for map imports that makes import of larger maps possible on small VMs (#334)
|
|
||||||
* Reference current postgres and postgis versions in docker-compose.yaml files (#286)
|
|
||||||
* Configurable terms-and-conditions link (#320)
|
|
||||||
* French translation by @cbiteau (#303)
|
|
||||||
|
|
||||||
### Bug Fixes
|
|
||||||
|
|
||||||
* Logout not working (#285)
|
|
||||||
* Duplicate road usage hashes (#335, #253)
|
|
||||||
* cannot import name .... (#338)
|
|
||||||
|
|
||||||
## 0.7.0
|
|
||||||
|
|
||||||
### Features
|
|
||||||
|
|
||||||
* Add histogram of overtaking distances in road details panel
|
|
||||||
* Flip table in road details panel and make it easier to read
|
|
||||||
* Implement difference between urban and rural for events and road segments.
|
|
||||||
* Better road zone detection in import
|
|
||||||
* Make the frontend translatable and add German translation
|
|
||||||
* Add time and user filters to map view (for logged-in users only)
|
|
||||||
|
|
||||||
### Improvements
|
|
||||||
|
|
||||||
* Make raw track not look like a river (#252)
|
|
||||||
* Update many dependencies
|
|
||||||
|
|
||||||
### Bug fixes
|
|
||||||
|
|
||||||
* Overtaking events are now deleted when the parent track is deleted (#206)
|
|
||||||
* Remove useless session creation (#192)
|
|
||||||
* Remove some error logs for canceled requests (as the map page tends to do that quite a lot)
|
|
||||||
* Fix ExportPage bounding box input
|
|
||||||
|
|
||||||
|
|
||||||
## 0.6.2
|
|
||||||
|
|
||||||
### Improvements
|
|
||||||
|
|
||||||
* Prevent directory traversals inside container on python-served frontend.
|
|
||||||
|
|
||||||
## 0.6.1
|
## 0.6.1
|
||||||
|
|
||||||
### Improvements
|
### Improvements
|
||||||
|
|
47
Dockerfile
47
Dockerfile
|
@ -1,10 +1,49 @@
|
||||||
# This dockerfile is for the API + Frontend production image
|
# This dockerfile is for the API + Frontend production image
|
||||||
|
|
||||||
|
#############################################
|
||||||
|
# Build osm2pgsql AS builder
|
||||||
|
#############################################
|
||||||
|
|
||||||
|
# This image should be the same as final one, because of the lib versions
|
||||||
|
FROM python:3.9.7-bullseye as osm2pgsql-builder
|
||||||
|
|
||||||
|
ARG DEBIAN_FRONTEND=noninteractive
|
||||||
|
ENV TZ=Europe/Berlin
|
||||||
|
ENV OSM2PGSQL_VERSION=1.5.1
|
||||||
|
|
||||||
|
# Dependencies
|
||||||
|
RUN apt-get update &&\
|
||||||
|
apt-get install -y \
|
||||||
|
make \
|
||||||
|
cmake \
|
||||||
|
g++ \
|
||||||
|
libboost-dev \
|
||||||
|
libboost-system-dev \
|
||||||
|
libboost-filesystem-dev \
|
||||||
|
libexpat1-dev \
|
||||||
|
zlib1g-dev \
|
||||||
|
libbz2-dev \
|
||||||
|
libpq-dev \
|
||||||
|
libproj-dev \
|
||||||
|
lua5.3 \
|
||||||
|
liblua5.3-dev \
|
||||||
|
git &&\
|
||||||
|
rm -rf /var/lib/apt/lists/*
|
||||||
|
|
||||||
|
# Clone & Build
|
||||||
|
RUN git clone --branch $OSM2PGSQL_VERSION https://github.com/openstreetmap/osm2pgsql.git &&\
|
||||||
|
cd osm2pgsql/ &&\
|
||||||
|
mkdir build &&\
|
||||||
|
cd build &&\
|
||||||
|
cmake .. &&\
|
||||||
|
make -j4 &&\
|
||||||
|
make install
|
||||||
|
|
||||||
#############################################
|
#############################################
|
||||||
# Build the frontend AS builder
|
# Build the frontend AS builder
|
||||||
#############################################
|
#############################################
|
||||||
|
|
||||||
FROM node:18 as frontend-builder
|
FROM node:17 as frontend-builder
|
||||||
|
|
||||||
WORKDIR /opt/obs/frontend
|
WORKDIR /opt/obs/frontend
|
||||||
ADD frontend/package.json frontend/package-lock.json /opt/obs/frontend/
|
ADD frontend/package.json frontend/package-lock.json /opt/obs/frontend/
|
||||||
|
@ -21,7 +60,7 @@ RUN npm run build
|
||||||
# Build the API and add the built frontend to it
|
# Build the API and add the built frontend to it
|
||||||
#############################################
|
#############################################
|
||||||
|
|
||||||
FROM python:3.11.3-bullseye
|
FROM python:3.9.7-bullseye
|
||||||
|
|
||||||
RUN apt-get update &&\
|
RUN apt-get update &&\
|
||||||
apt-get install -y \
|
apt-get install -y \
|
||||||
|
@ -54,7 +93,11 @@ ADD api/obs /opt/obs/api/obs/
|
||||||
ADD api/tools /opt/obs/api/tools/
|
ADD api/tools /opt/obs/api/tools/
|
||||||
RUN pip install -e /opt/obs/api/
|
RUN pip install -e /opt/obs/api/
|
||||||
|
|
||||||
|
ADD roads_import.lua /opt/obs/api/tools
|
||||||
|
ADD osm2pgsql.sh /opt/obs/api/tools
|
||||||
|
|
||||||
COPY --from=frontend-builder /opt/obs/frontend/build /opt/obs/frontend/build
|
COPY --from=frontend-builder /opt/obs/frontend/build /opt/obs/frontend/build
|
||||||
|
COPY --from=osm2pgsql-builder /usr/local/bin/osm2pgsql /usr/local/bin/osm2pgsql
|
||||||
|
|
||||||
EXPOSE 3000
|
EXPOSE 3000
|
||||||
|
|
||||||
|
|
69
README.md
69
README.md
|
@ -36,11 +36,10 @@ git submodule update --init --recursive
|
||||||
|
|
||||||
## Production setup
|
## Production setup
|
||||||
|
|
||||||
There is a guide for a deployment based on docker at
|
There is a guide for a deployment based on docker in the
|
||||||
[docs/production-deployment.md](docs/production-deployment.md). Lots of
|
[deployment](deployment) folder. Lots of non-docker deployment strategy are
|
||||||
non-docker deployment strategies are possible, but they are not "officially"
|
possible, but they are not "officially" supported, so please do not expect the
|
||||||
supported, so please do not expect the authors of the software to assist in
|
authors of the software to assist in troubleshooting.
|
||||||
troubleshooting.
|
|
||||||
|
|
||||||
This is a rather complex application, and it is expected that you know the
|
This is a rather complex application, and it is expected that you know the
|
||||||
basics of deploying a modern web application securely onto a production server.
|
basics of deploying a modern web application securely onto a production server.
|
||||||
|
@ -53,8 +52,7 @@ Please note that you will always need to install your own reverse proxy that
|
||||||
terminates TLS for you and handles certificates. We do not support TLS directly
|
terminates TLS for you and handles certificates. We do not support TLS directly
|
||||||
in the application, instead, please use this prefered method.
|
in the application, instead, please use this prefered method.
|
||||||
|
|
||||||
Upgrading and migrating is described in [UPGRADING.md](./UPGRADING.md) for each
|
Upgrading and migrating is descrube
|
||||||
version.
|
|
||||||
|
|
||||||
### Migrating (Production)
|
### Migrating (Production)
|
||||||
|
|
||||||
|
@ -77,6 +75,18 @@ docker-compose run --rm api alembic upgrade head
|
||||||
docker-compose run --rm api tools/prepare_sql_tiles
|
docker-compose run --rm api tools/prepare_sql_tiles
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
||||||
|
docker-compose run --rm api alembic upgrade head
|
||||||
|
|
||||||
|
### Upgrading from v0.2 to v0.3
|
||||||
|
|
||||||
|
After v0.2 we switched the underlying technology of the API and the database.
|
||||||
|
We now have no more MongoDB, instead, everything has moved to the PostgreSQL
|
||||||
|
installation. For development setups, it is advised to just reset the whole
|
||||||
|
state (remove the `local` folder) and start fresh. For production upgrades,
|
||||||
|
please follow the relevant section in [`UPGRADING.md`](./UPGRADING.md).
|
||||||
|
|
||||||
|
|
||||||
## Development setup
|
## Development setup
|
||||||
|
|
||||||
We've moved the whole development setup into Docker to make it easy for
|
We've moved the whole development setup into Docker to make it easy for
|
||||||
|
@ -91,6 +101,7 @@ Then clone the repository as described above.
|
||||||
|
|
||||||
### Configure Keycloak
|
### Configure Keycloak
|
||||||
|
|
||||||
|
|
||||||
Login will not be possible until you configure the keycloak realm correctly. Boot your keycloak instance:
|
Login will not be possible until you configure the keycloak realm correctly. Boot your keycloak instance:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
|
@ -153,7 +164,7 @@ You will need to re-run this command after updates, to migrate the database and
|
||||||
(re-)create the functions in the SQL database that are used when generating
|
(re-)create the functions in the SQL database that are used when generating
|
||||||
vector tiles.
|
vector tiles.
|
||||||
|
|
||||||
You should also [import OpenStreetMap data](docs/osm-import.md) now.
|
You should also import OpenStreetMap data now, see below for instructions.
|
||||||
|
|
||||||
### Boot the application
|
### Boot the application
|
||||||
|
|
||||||
|
@ -179,6 +190,48 @@ docker-compose run --rm api alembic upgrade head
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
||||||
|
## Import OpenStreetMap data
|
||||||
|
|
||||||
|
**Hint:** This step may be skipped if you are using [Lean mode](./docs/lean-mode.md).
|
||||||
|
|
||||||
|
You need to import road information from OpenStreetMap for the portal to work.
|
||||||
|
This information is stored in your PostgreSQL database and used when processing
|
||||||
|
tracks (instead of querying the Overpass API), as well as for vector tile
|
||||||
|
generation. The process applies to both development and production setups. For
|
||||||
|
development, you should choose a small area for testing, such as your local
|
||||||
|
county or city, to keep the amount of data small. For production use you have
|
||||||
|
to import the whole region you are serving.
|
||||||
|
|
||||||
|
* Install `osm2pgsql`.
|
||||||
|
* Download the area(s) you would like to import from [GeoFabrik](https://download.geofabrik.de).
|
||||||
|
* Import each file like this:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
osm2pgsql --create --hstore --style roads_import.lua -O flex \
|
||||||
|
-H localhost -d obs -U obs \
|
||||||
|
path/to/downloaded/myarea-latest.osm.pbf
|
||||||
|
```
|
||||||
|
|
||||||
|
You might need to adjust the host, database and username (`-H`, `-d`, `-U`) to
|
||||||
|
your setup, and also provide the correct password when queried. For the
|
||||||
|
development setup the password is `obs`. For production, you might need to
|
||||||
|
expose the containers port and/or create a TCP tunnel, for example with SSH,
|
||||||
|
such that you can run the import from your local host and write to the remote
|
||||||
|
database.
|
||||||
|
|
||||||
|
The import process should take a few seconds to minutes, depending on the area
|
||||||
|
size. A whole country might even take one or more hours. You should probably
|
||||||
|
not try to import `planet.osm.pbf`.
|
||||||
|
|
||||||
|
You can run the process multiple times, with the same or different area files,
|
||||||
|
to import or update the data. However, for this to work, the actual [command
|
||||||
|
line arguments](https://osm2pgsql.org/doc/manual.html#running-osm2pgsql) are a
|
||||||
|
bit different each time, including when first importing, and the disk space
|
||||||
|
required is much higher.
|
||||||
|
|
||||||
|
Refer to the documentation of `osm2pgsql` for assistance. We are using "flex
|
||||||
|
mode", the provided script `roads_import.lua` describes the transformations
|
||||||
|
and extractions to perform on the original data.
|
||||||
|
|
||||||
## Troubleshooting
|
## Troubleshooting
|
||||||
|
|
||||||
|
|
58
UPGRADING.md
58
UPGRADING.md
|
@ -1,57 +1,9 @@
|
||||||
# Upgrading
|
# Upgrading
|
||||||
|
|
||||||
This document describes the general steps to upgrade between major changes.
|
This document describes the general steps to upgrade between major changes.
|
||||||
Simple migrations, e.g. for adding schema changes, are not documented
|
Simple migrations, e.g. for adding schema changes, are not documented
|
||||||
explicitly. Their general usage is described in the [README](./README.md) (for
|
explicitly. Their general usage is described in the [README](./README.md) (for
|
||||||
development) and [docs/production-deployment.md](docs/production-deployment.md) (for production).
|
development) and [deployment/README.md](deployment/README.md) (for production).
|
||||||
|
|
||||||
|
|
||||||
## 0.8.1
|
|
||||||
|
|
||||||
- Get the release in your source folder (``git pull; git checkout 0.8.0`` and update submodules ``git submodule update --recursive``)
|
|
||||||
- Rebuild images ``docker-compose build``
|
|
||||||
- No database upgrade is required, but tile functions need an update:
|
|
||||||
```bash
|
|
||||||
docker-compose run --rm portal tools/prepare_sql_tiles.py
|
|
||||||
```
|
|
||||||
- Start your portal and worker services. ``docker-compose up -d worker portal``
|
|
||||||
|
|
||||||
|
|
||||||
## 0.8.0
|
|
||||||
Upgrade to `0.7.x` first. See below for details. Then follow these steps:
|
|
||||||
|
|
||||||
> **Warning** The update includes a reprocessing of tracks after import. Depending on the number of tracks this can take a few hours. The portal is reachable during that time but events disappear and incrementally reappear during reimport.
|
|
||||||
|
|
||||||
> **Info** With this version the import process for OpenStreetMap data has changed: the [new process](docs/osm-import.md) is easier on resources and finally permits to import a full country on a low-end VM.
|
|
||||||
|
|
||||||
- Do your [usual backup](docs/production-deployment.md)
|
|
||||||
- get the release in your source folder (``git pull; git checkout 0.8.0`` and update submodules ``git submodule update --recursive``)
|
|
||||||
- Rebuild images ``docker-compose build``
|
|
||||||
- Stop your portal and worker services ``docker-compose stop worker portal``
|
|
||||||
- run upgrade
|
|
||||||
```bash
|
|
||||||
docker-compose run --rm portal tools/upgrade.py
|
|
||||||
```
|
|
||||||
this automatically does the following
|
|
||||||
- Migration of database schema using alembic.
|
|
||||||
- Upgrade of SQL tile schema to new schema.
|
|
||||||
- Import the nuts-regions from the web into the database.
|
|
||||||
- Trigger a re-import of all tracks.
|
|
||||||
- Start your portal and worker services. ``docker-compose up -d worker portal``
|
|
||||||
|
|
||||||
|
|
||||||
## 0.7.0
|
|
||||||
|
|
||||||
Upgrade to `0.6.x` first. See below for details. Then follow these steps:
|
|
||||||
|
|
||||||
- Rebuild images
|
|
||||||
- Stop your portal and worker services.
|
|
||||||
- **Migration with alembic**: required
|
|
||||||
- **Prepare SQL Tiles**: required
|
|
||||||
- Start your portal and worker services.
|
|
||||||
- **Reimport tracks**: no action required
|
|
||||||
- **OSM Import**: required
|
|
||||||
- **Config changes**: add `POSTGRES_MAX_OVERFLOW` and `POSTGRES_POOL_SIZE`
|
|
||||||
variables, see `api/config.py.example`
|
|
||||||
|
|
||||||
## 0.6.0
|
## 0.6.0
|
||||||
|
|
||||||
|
@ -91,7 +43,7 @@ You can, but do not have to, reimport all tracks. This will generate a GPX file
|
||||||
for each track and allow the users to download those. If a GPX file has not yet
|
for each track and allow the users to download those. If a GPX file has not yet
|
||||||
been created, the download will fail. To reimport all tracks, log in to your
|
been created, the download will fail. To reimport all tracks, log in to your
|
||||||
PostgreSQL database (instructions are in [README.md](./README.md) for
|
PostgreSQL database (instructions are in [README.md](./README.md) for
|
||||||
development and [docs/production-deployment.md](./docs/production-deployment.md) for production)
|
development and [deployment/README.md](./deployment/README.md) for production)
|
||||||
and run:
|
and run:
|
||||||
|
|
||||||
```sql
|
```sql
|
||||||
|
@ -111,7 +63,7 @@ Make sure your worker is running to process the queue.
|
||||||
`POSTGRES_MAX_OVERFLOW`. Check the example config for sane default values.
|
`POSTGRES_MAX_OVERFLOW`. Check the example config for sane default values.
|
||||||
* Re-run `tools/prepare_sql_tiles.py` again (see README)
|
* Re-run `tools/prepare_sql_tiles.py` again (see README)
|
||||||
* It has been made easier to import OSM data, check
|
* It has been made easier to import OSM data, check
|
||||||
[docs/production-deployment.md](./docs/production-deployment.md) for the sections "Download
|
[deployment/README.md](deployment/README.md) for the sections "Download
|
||||||
OpenStreetMap maps" and "Import OpenStreetMap data". You can now download
|
OpenStreetMap maps" and "Import OpenStreetMap data". You can now download
|
||||||
multiple .pbf files and then import them at once, using the docker image
|
multiple .pbf files and then import them at once, using the docker image
|
||||||
built with the `Dockerfile`. Alternatively, you can choose to enable [lean
|
built with the `Dockerfile`. Alternatively, you can choose to enable [lean
|
||||||
|
@ -166,5 +118,5 @@ Make sure your worker is running to process the queue.
|
||||||
`export/users.json` into your realm, it will re-add all the users from the
|
`export/users.json` into your realm, it will re-add all the users from the
|
||||||
old installation. You should delete the file and `export/` folder afterwards.
|
old installation. You should delete the file and `export/` folder afterwards.
|
||||||
* Start `portal`.
|
* Start `portal`.
|
||||||
* Consider configuring a worker service. See [docs/production-deployment.md](./docs/production-deployment.md).
|
* Consider configuring a worker service. See [deployment/README.md](deployment/README.md).
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
FROM python:3.11.3-bullseye
|
FROM python:3.9.7-bullseye
|
||||||
|
|
||||||
WORKDIR /opt/obs/api
|
WORKDIR /opt/obs/api
|
||||||
|
|
||||||
|
|
|
@ -2,8 +2,9 @@ HOST = "0.0.0.0"
|
||||||
PORT = 3000
|
PORT = 3000
|
||||||
DEBUG = True
|
DEBUG = True
|
||||||
VERBOSE = False
|
VERBOSE = False
|
||||||
AUTO_RELOAD = True
|
AUTO_RESTART = True
|
||||||
SECRET = "!!!!!!!!!!!!CHANGE ME!!!!!!!!!!!!"
|
SECRET = "!!!!!!!!!!!!CHANGE ME!!!!!!!!!!!!"
|
||||||
|
LEAN_MODE = False
|
||||||
POSTGRES_URL = "postgresql+asyncpg://obs:obs@postgres/obs"
|
POSTGRES_URL = "postgresql+asyncpg://obs:obs@postgres/obs"
|
||||||
POSTGRES_POOL_SIZE = 20
|
POSTGRES_POOL_SIZE = 20
|
||||||
POSTGRES_MAX_OVERFLOW = 2 * POSTGRES_POOL_SIZE
|
POSTGRES_MAX_OVERFLOW = 2 * POSTGRES_POOL_SIZE
|
||||||
|
@ -18,7 +19,6 @@ FRONTEND_DIR = None
|
||||||
FRONTEND_CONFIG = {
|
FRONTEND_CONFIG = {
|
||||||
"imprintUrl": "https://example.com/imprint",
|
"imprintUrl": "https://example.com/imprint",
|
||||||
"privacyPolicyUrl": "https://example.com/privacy",
|
"privacyPolicyUrl": "https://example.com/privacy",
|
||||||
# "termsUrl": "https://example.com/terms", # Link is only shown when set
|
|
||||||
"mapHome": {"zoom": 6, "longitude": 10.2, "latitude": 51.3},
|
"mapHome": {"zoom": 6, "longitude": 10.2, "latitude": 51.3},
|
||||||
# "banner": {"text": "This is a development installation.", "style": "info"},
|
# "banner": {"text": "This is a development installation.", "style": "info"},
|
||||||
}
|
}
|
||||||
|
@ -29,7 +29,5 @@ ADDITIONAL_CORS_ORIGINS = [
|
||||||
"http://localhost:8880/", # for maputnik on 8880
|
"http://localhost:8880/", # for maputnik on 8880
|
||||||
"http://localhost:8888/", # for maputnik on 8888
|
"http://localhost:8888/", # for maputnik on 8888
|
||||||
]
|
]
|
||||||
TILE_SEMAPHORE_SIZE = 4
|
|
||||||
EXPORT_SEMAPHORE_SIZE = 4
|
|
||||||
|
|
||||||
# vim: set ft=python :
|
# vim: set ft=python :
|
||||||
|
|
|
@ -5,7 +5,12 @@ PORT = 3000
|
||||||
# Extended log output, but slower
|
# Extended log output, but slower
|
||||||
DEBUG = False
|
DEBUG = False
|
||||||
VERBOSE = DEBUG
|
VERBOSE = DEBUG
|
||||||
AUTO_RELOAD = DEBUG
|
AUTO_RESTART = DEBUG
|
||||||
|
|
||||||
|
# Turn on lean mode to simplify the setup. Lots of features will be
|
||||||
|
# unavailable, but you will not need to manage OpenStreetMap data. Please make
|
||||||
|
# sure to configure the OBS_FACE_CACHE_DIR correctly for lean mode.
|
||||||
|
LEAN_MODE = False
|
||||||
|
|
||||||
# Required to encrypt or sign sessions, cookies, tokens, etc.
|
# Required to encrypt or sign sessions, cookies, tokens, etc.
|
||||||
SECRET = "!!!<<<CHANGEME>>>!!!"
|
SECRET = "!!!<<<CHANGEME>>>!!!"
|
||||||
|
@ -39,7 +44,6 @@ FRONTEND_DIR = "../frontend/build/"
|
||||||
FRONTEND_CONFIG = {
|
FRONTEND_CONFIG = {
|
||||||
"imprintUrl": "https://example.com/imprint",
|
"imprintUrl": "https://example.com/imprint",
|
||||||
"privacyPolicyUrl": "https://example.com/privacy",
|
"privacyPolicyUrl": "https://example.com/privacy",
|
||||||
# "termsUrl": "https://example.com/user_terms_and_conditions", # Link is only shown when set
|
|
||||||
"mapHome": {"zoom": 6, "longitude": 10.2, "latitude": 51.3},
|
"mapHome": {"zoom": 6, "longitude": 10.2, "latitude": 51.3},
|
||||||
"banner": {"text": "This is a test installation.", "style": "warning"},
|
"banner": {"text": "This is a test installation.", "style": "warning"},
|
||||||
}
|
}
|
||||||
|
@ -61,13 +65,4 @@ TILES_FILE = None
|
||||||
# default. Python list, or whitespace separated string.
|
# default. Python list, or whitespace separated string.
|
||||||
ADDITIONAL_CORS_ORIGINS = None
|
ADDITIONAL_CORS_ORIGINS = None
|
||||||
|
|
||||||
# How many asynchronous requests may be sent to the database to generate tile
|
|
||||||
# information. Should be less than POSTGRES_POOL_SIZE to leave some connections
|
|
||||||
# to the other features of the API ;)
|
|
||||||
TILE_SEMAPHORE_SIZE = 4
|
|
||||||
|
|
||||||
# How many asynchronous requests may generate exported data simultaneously.
|
|
||||||
# Keep this small.
|
|
||||||
EXPORT_SEMAPHORE_SIZE = 1
|
|
||||||
|
|
||||||
# vim: set ft=python :
|
# vim: set ft=python :
|
||||||
|
|
|
@ -22,16 +22,13 @@ def upgrade():
|
||||||
op.create_table(
|
op.create_table(
|
||||||
"road",
|
"road",
|
||||||
sa.Column(
|
sa.Column(
|
||||||
"way_id", sa.BIGINT, primary_key=True, index=True, autoincrement=False
|
"way_id", sa.BIGINT, autoincrement=True, primary_key=True, index=True
|
||||||
),
|
),
|
||||||
sa.Column("zone", dbtype("zone_type")),
|
sa.Column("zone", dbtype("zone_type")),
|
||||||
sa.Column("name", sa.Text),
|
sa.Column("name", sa.String),
|
||||||
sa.Column("geometry", dbtype("geometry(LINESTRING,3857)")),
|
sa.Column("geometry", dbtype("GEOMETRY"), index=True),
|
||||||
sa.Column("directionality", sa.Integer),
|
sa.Column("directionality", sa.Integer),
|
||||||
sa.Column("oneway", sa.Boolean),
|
sa.Column("oenway", sa.Boolean),
|
||||||
)
|
|
||||||
op.execute(
|
|
||||||
"CREATE INDEX road_geometry_idx ON road USING GIST (geometry) WITH (FILLFACTOR=100);"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,30 +0,0 @@
|
||||||
"""transform overtaking_event geometry to 3857
|
|
||||||
|
|
||||||
Revision ID: 587e69ecb466
|
|
||||||
Revises: f4b0f460254d
|
|
||||||
Create Date: 2023-04-01 14:30:49.927505
|
|
||||||
|
|
||||||
"""
|
|
||||||
from alembic import op
|
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision = "587e69ecb466"
|
|
||||||
down_revision = "f4b0f460254d"
|
|
||||||
branch_labels = None
|
|
||||||
depends_on = None
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade():
|
|
||||||
op.execute("UPDATE overtaking_event SET geometry = ST_Transform(geometry, 3857);")
|
|
||||||
op.execute(
|
|
||||||
"ALTER TABLE overtaking_event ALTER COLUMN geometry TYPE geometry(POINT, 3857);"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade():
|
|
||||||
op.execute(
|
|
||||||
"ALTER TABLE overtaking_event ALTER COLUMN geometry TYPE geometry;"
|
|
||||||
)
|
|
||||||
op.execute("UPDATE overtaking_event SET geometry = ST_Transform(geometry, 4326);")
|
|
|
@ -1,26 +0,0 @@
|
||||||
"""add_overtaking_event_index
|
|
||||||
|
|
||||||
|
|
||||||
Revision ID: 7868aed76122
|
|
||||||
Revises: 587e69ecb466
|
|
||||||
Create Date: 2023-07-16 13:37:17.694079
|
|
||||||
|
|
||||||
"""
|
|
||||||
from alembic import op
|
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision = '7868aed76122'
|
|
||||||
down_revision = '587e69ecb466'
|
|
||||||
branch_labels = None
|
|
||||||
depends_on = None
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade():
|
|
||||||
op.execute("CREATE INDEX IF NOT EXISTS ix_overtaking_event_geometry ON overtaking_event using GIST(geometry);")
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade():
|
|
||||||
op.drop_index("ix_overtaking_event_geometry")
|
|
||||||
|
|
|
@ -1,26 +0,0 @@
|
||||||
"""add user display_name
|
|
||||||
|
|
||||||
Revision ID: 99a3d2eb08f9
|
|
||||||
Revises: a9627f63fbed
|
|
||||||
Create Date: 2022-09-13 07:30:18.747880
|
|
||||||
|
|
||||||
"""
|
|
||||||
from alembic import op
|
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision = "99a3d2eb08f9"
|
|
||||||
down_revision = "a9627f63fbed"
|
|
||||||
branch_labels = None
|
|
||||||
depends_on = None
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade():
|
|
||||||
op.add_column(
|
|
||||||
"user", sa.Column("display_name", sa.String, nullable=True), schema="public"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade():
|
|
||||||
op.drop_column("user", "display_name", schema="public")
|
|
|
@ -13,7 +13,7 @@ from migrations.utils import dbtype
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
# revision identifiers, used by Alembic.
|
||||||
revision = "a049e5eb24dd"
|
revision = "a049e5eb24dd"
|
||||||
down_revision = "99a3d2eb08f9"
|
down_revision = "a9627f63fbed"
|
||||||
branch_labels = None
|
branch_labels = None
|
||||||
depends_on = None
|
depends_on = None
|
||||||
|
|
||||||
|
@ -21,13 +21,14 @@ depends_on = None
|
||||||
def upgrade():
|
def upgrade():
|
||||||
op.create_table(
|
op.create_table(
|
||||||
"region",
|
"region",
|
||||||
sa.Column("id", sa.String(24), primary_key=True, index=True),
|
sa.Column(
|
||||||
sa.Column("name", sa.Text),
|
"way_id", sa.BIGINT, autoincrement=True, primary_key=True, index=True
|
||||||
sa.Column("geometry", dbtype("GEOMETRY(GEOMETRY,3857)"), index=False),
|
),
|
||||||
sa.Column("admin_level", sa.Integer, index=True),
|
sa.Column("zone", dbtype("zone_type")),
|
||||||
)
|
sa.Column("name", sa.String),
|
||||||
op.execute(
|
sa.Column("geometry", dbtype("GEOMETRY"), index=True),
|
||||||
"CREATE INDEX region_geometry_idx ON region USING GIST (geometry) WITH (FILLFACTOR=100);"
|
sa.Column("directionality", sa.Integer),
|
||||||
|
sa.Column("oenway", sa.Boolean),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,39 +0,0 @@
|
||||||
"""add import groups
|
|
||||||
|
|
||||||
Revision ID: b8b0fbae50a4
|
|
||||||
Revises: f7b21148126a
|
|
||||||
Create Date: 2023-03-26 09:41:36.621203
|
|
||||||
|
|
||||||
"""
|
|
||||||
from alembic import op
|
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision = "b8b0fbae50a4"
|
|
||||||
down_revision = "f7b21148126a"
|
|
||||||
branch_labels = None
|
|
||||||
depends_on = None
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade():
|
|
||||||
op.add_column(
|
|
||||||
"road",
|
|
||||||
sa.Column("import_group", sa.String(), nullable=True),
|
|
||||||
)
|
|
||||||
op.add_column(
|
|
||||||
"region",
|
|
||||||
sa.Column("import_group", sa.String(), nullable=True),
|
|
||||||
)
|
|
||||||
|
|
||||||
# Set existing to "osm2pgsql"
|
|
||||||
road = sa.table("road", sa.column("import_group", sa.String))
|
|
||||||
op.execute(road.update().values(import_group="osm2pgsql"))
|
|
||||||
|
|
||||||
region = sa.table("region", sa.column("import_group", sa.String))
|
|
||||||
op.execute(region.update().values(import_group="osm2pgsql"))
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade():
|
|
||||||
op.drop_column("road", "import_group")
|
|
||||||
op.drop_column("region", "import_group")
|
|
|
@ -0,0 +1,24 @@
|
||||||
|
"""add road column has_cars
|
||||||
|
|
||||||
|
Revision ID: ddde1cdc767b
|
||||||
|
Revises: a049e5eb24dd
|
||||||
|
Create Date: 2022-04-03 20:13:22.874195
|
||||||
|
|
||||||
|
"""
|
||||||
|
from alembic import op
|
||||||
|
import sqlalchemy as sa
|
||||||
|
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision = "ddde1cdc767b"
|
||||||
|
down_revision = "a049e5eb24dd"
|
||||||
|
branch_labels = None
|
||||||
|
depends_on = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade():
|
||||||
|
op.add_column("road", sa.Column("has_cars", sa.Boolean))
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade():
|
||||||
|
op.drop_column("road", "has_cars")
|
|
@ -1,24 +0,0 @@
|
||||||
"""add osm id indexes
|
|
||||||
|
|
||||||
Revision ID: f4b0f460254d
|
|
||||||
Revises: b8b0fbae50a4
|
|
||||||
Create Date: 2023-03-30 10:56:22.066768
|
|
||||||
|
|
||||||
"""
|
|
||||||
from alembic import op
|
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision = "f4b0f460254d"
|
|
||||||
down_revision = "b8b0fbae50a4"
|
|
||||||
branch_labels = None
|
|
||||||
depends_on = None
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade():
|
|
||||||
op.execute("CREATE INDEX IF NOT EXISTS ix_road_way_id ON road (way_id);")
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade():
|
|
||||||
op.drop_index("ix_road_way_id")
|
|
|
@ -1,41 +0,0 @@
|
||||||
"""add user_device
|
|
||||||
|
|
||||||
Revision ID: f7b21148126a
|
|
||||||
Revises: a9627f63fbed
|
|
||||||
Create Date: 2022-09-15 17:48:06.764342
|
|
||||||
|
|
||||||
"""
|
|
||||||
from alembic import op
|
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision = "f7b21148126a"
|
|
||||||
down_revision = "a049e5eb24dd"
|
|
||||||
branch_labels = None
|
|
||||||
depends_on = None
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade():
|
|
||||||
op.create_table(
|
|
||||||
"user_device",
|
|
||||||
sa.Column("id", sa.Integer, autoincrement=True, primary_key=True),
|
|
||||||
sa.Column("user_id", sa.Integer, sa.ForeignKey("user.id", ondelete="CASCADE")),
|
|
||||||
sa.Column("identifier", sa.String, nullable=False),
|
|
||||||
sa.Column("display_name", sa.String, nullable=True),
|
|
||||||
sa.Index("user_id_identifier", "user_id", "identifier", unique=True),
|
|
||||||
)
|
|
||||||
op.add_column(
|
|
||||||
"track",
|
|
||||||
sa.Column(
|
|
||||||
"user_device_id",
|
|
||||||
sa.Integer,
|
|
||||||
sa.ForeignKey("user_device.id", ondelete="RESTRICT"),
|
|
||||||
nullable=True,
|
|
||||||
),
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade():
|
|
||||||
op.drop_column("track", "user_device_id")
|
|
||||||
op.drop_table("user_device")
|
|
|
@ -1 +1 @@
|
||||||
__version__ = "0.8.1"
|
__version__ = "0.6.1"
|
||||||
|
|
|
@ -1,4 +1,3 @@
|
||||||
import asyncio
|
|
||||||
import logging
|
import logging
|
||||||
import re
|
import re
|
||||||
|
|
||||||
|
@ -20,62 +19,14 @@ from sanic_session import Session, InMemorySessionInterface
|
||||||
|
|
||||||
from sqlalchemy import select
|
from sqlalchemy import select
|
||||||
from sqlalchemy.ext.asyncio import AsyncSession
|
from sqlalchemy.ext.asyncio import AsyncSession
|
||||||
|
from sqlalchemy.orm import sessionmaker
|
||||||
|
|
||||||
from obs.api.db import User, make_session, connect_db
|
from obs.api.db import User, make_session, connect_db
|
||||||
from obs.api.cors import setup_options, add_cors_headers
|
|
||||||
from obs.api.utils import get_single_arg
|
from obs.api.utils import get_single_arg
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
app = Sanic("OpenBikeSensor Portal API", load_env="OBS_", log_config={})
|
||||||
class SanicAccessMessageFilter(logging.Filter):
|
|
||||||
"""
|
|
||||||
A filter that modifies the log message of a sanic.access log entry to
|
|
||||||
include useful information.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def filter(self, record):
|
|
||||||
record.msg = f"{record.request} -> {record.status}"
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
def configure_sanic_logging():
|
|
||||||
for logger_name in ["sanic.root", "sanic.access", "sanic.error"]:
|
|
||||||
logger = logging.getLogger(logger_name)
|
|
||||||
for handler in logger.handlers:
|
|
||||||
logger.removeHandler(handler)
|
|
||||||
|
|
||||||
logger = logging.getLogger("sanic.access")
|
|
||||||
for filter_ in logger.filters:
|
|
||||||
logger.removeFilter(filter_)
|
|
||||||
logger.addFilter(SanicAccessMessageFilter())
|
|
||||||
logging.getLogger("sanic.root").setLevel(logging.WARNING)
|
|
||||||
|
|
||||||
|
|
||||||
app = Sanic(
|
|
||||||
"openbikesensor-api",
|
|
||||||
env_prefix="OBS_",
|
|
||||||
)
|
|
||||||
configure_sanic_logging()
|
|
||||||
|
|
||||||
app.config.update(
|
|
||||||
dict(
|
|
||||||
DEBUG=False,
|
|
||||||
VERBOSE=False,
|
|
||||||
AUTO_RELOAD=False,
|
|
||||||
POSTGRES_POOL_SIZE=20,
|
|
||||||
POSTGRES_MAX_OVERFLOW=40,
|
|
||||||
DEDICATED_WORKER=True,
|
|
||||||
FRONTEND_URL=None,
|
|
||||||
FRONTEND_HTTPS=True,
|
|
||||||
TILES_FILE=None,
|
|
||||||
TILE_SEMAPHORE_SIZE=4,
|
|
||||||
EXPORT_SEMAPHORE_SIZE=1,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
# overwrite from defaults again
|
|
||||||
app.config.load_environment_vars("OBS_")
|
|
||||||
|
|
||||||
if isfile("./config.py"):
|
if isfile("./config.py"):
|
||||||
app.update_config("./config.py")
|
app.update_config("./config.py")
|
||||||
|
@ -89,59 +40,9 @@ c = app.config
|
||||||
api = Blueprint("api", url_prefix="/api")
|
api = Blueprint("api", url_prefix="/api")
|
||||||
auth = Blueprint("auth", url_prefix="")
|
auth = Blueprint("auth", url_prefix="")
|
||||||
|
|
||||||
import re
|
|
||||||
|
|
||||||
TILE_REQUEST_CANCELLED = re.compile(
|
|
||||||
r"Connection lost before response written.*GET /tiles"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class NoConnectionLostFilter(logging.Filter):
|
|
||||||
def filter(record):
|
|
||||||
return not TILE_REQUEST_CANCELLED.match(record.getMessage())
|
|
||||||
|
|
||||||
|
|
||||||
logging.getLogger("sanic.error").addFilter(NoConnectionLostFilter)
|
|
||||||
|
|
||||||
|
|
||||||
def setup_cors(app):
|
|
||||||
frontend_url = app.config.get("FRONTEND_URL")
|
|
||||||
additional_origins = app.config.get("ADDITIONAL_CORS_ORIGINS")
|
|
||||||
if not frontend_url and not additional_origins:
|
|
||||||
# No CORS configured
|
|
||||||
return
|
|
||||||
|
|
||||||
origins = []
|
|
||||||
if frontend_url:
|
|
||||||
u = urlparse(frontend_url)
|
|
||||||
origins.append(f"{u.scheme}://{u.netloc}")
|
|
||||||
|
|
||||||
if isinstance(additional_origins, str):
|
|
||||||
origins += re.split(r"\s+", additional_origins)
|
|
||||||
elif isinstance(additional_origins, list):
|
|
||||||
origins += additional_origins
|
|
||||||
elif additional_origins is not None:
|
|
||||||
raise ValueError(
|
|
||||||
"invalid option type for ADDITIONAL_CORS_ORIGINS, must be list or space separated str"
|
|
||||||
)
|
|
||||||
|
|
||||||
app.ctx.cors_origins = origins
|
|
||||||
|
|
||||||
# Add OPTIONS handlers to any route that is missing it
|
|
||||||
app.register_listener(setup_options, "before_server_start")
|
|
||||||
|
|
||||||
# Fill in CORS headers
|
|
||||||
app.register_middleware(add_cors_headers, "response")
|
|
||||||
|
|
||||||
|
|
||||||
setup_cors(app)
|
|
||||||
|
|
||||||
|
|
||||||
@app.exception(SanicException, BaseException)
|
@app.exception(SanicException, BaseException)
|
||||||
async def _handle_sanic_errors(_request, exception):
|
async def _handle_sanic_errors(_request, exception):
|
||||||
if isinstance(exception, asyncio.CancelledError):
|
|
||||||
return None
|
|
||||||
|
|
||||||
log.error("Exception in handler: %s", exception, exc_info=True)
|
log.error("Exception in handler: %s", exception, exc_info=True)
|
||||||
return json_response(
|
return json_response(
|
||||||
{
|
{
|
||||||
|
@ -173,6 +74,39 @@ def configure_paths(c):
|
||||||
configure_paths(app.config)
|
configure_paths(app.config)
|
||||||
|
|
||||||
|
|
||||||
|
def setup_cors(app):
|
||||||
|
frontend_url = app.config.get("FRONTEND_URL")
|
||||||
|
additional_origins = app.config.get("ADDITIONAL_CORS_ORIGINS")
|
||||||
|
if not frontend_url and not additional_origins:
|
||||||
|
# No CORS configured
|
||||||
|
return
|
||||||
|
|
||||||
|
origins = []
|
||||||
|
if frontend_url:
|
||||||
|
u = urlparse(frontend_url)
|
||||||
|
origins.append(f"{u.scheme}://{u.netloc}")
|
||||||
|
|
||||||
|
if isinstance(additional_origins, str):
|
||||||
|
origins += re.split(r"\s+", additional_origins)
|
||||||
|
elif isinstance(additional_origins, list):
|
||||||
|
origins += additional_origins
|
||||||
|
elif additional_origins is not None:
|
||||||
|
raise ValueError(
|
||||||
|
"invalid option type for ADDITIONAL_CORS_ORIGINS, must be list or space separated str"
|
||||||
|
)
|
||||||
|
|
||||||
|
from sanic_cors import CORS
|
||||||
|
|
||||||
|
CORS(
|
||||||
|
app,
|
||||||
|
origins=origins,
|
||||||
|
supports_credentials=True,
|
||||||
|
expose_headers={"Content-Disposition"},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
setup_cors(app)
|
||||||
|
|
||||||
# TODO: use a different interface, maybe backed by the PostgreSQL, to allow
|
# TODO: use a different interface, maybe backed by the PostgreSQL, to allow
|
||||||
# scaling the API
|
# scaling the API
|
||||||
Session(app, interface=InMemorySessionInterface())
|
Session(app, interface=InMemorySessionInterface())
|
||||||
|
@ -187,12 +121,6 @@ async def app_connect_db(app, loop):
|
||||||
)
|
)
|
||||||
app.ctx._db_engine = await app.ctx._db_engine_ctx.__aenter__()
|
app.ctx._db_engine = await app.ctx._db_engine_ctx.__aenter__()
|
||||||
|
|
||||||
if app.config.TILE_SEMAPHORE_SIZE:
|
|
||||||
app.ctx.tile_semaphore = asyncio.Semaphore(app.config.TILE_SEMAPHORE_SIZE)
|
|
||||||
|
|
||||||
if app.config.EXPORT_SEMAPHORE_SIZE:
|
|
||||||
app.ctx.export_semaphore = asyncio.Semaphore(app.config.EXPORT_SEMAPHORE_SIZE)
|
|
||||||
|
|
||||||
|
|
||||||
@app.after_server_stop
|
@app.after_server_stop
|
||||||
async def app_disconnect_db(app, loop):
|
async def app_disconnect_db(app, loop):
|
||||||
|
@ -254,6 +182,7 @@ async def inject_urls(req):
|
||||||
async def inject_session(req):
|
async def inject_session(req):
|
||||||
req.ctx._session_ctx = make_session()
|
req.ctx._session_ctx = make_session()
|
||||||
req.ctx.db = await req.ctx._session_ctx.__aenter__()
|
req.ctx.db = await req.ctx._session_ctx.__aenter__()
|
||||||
|
sessionmaker(req.app.ctx._db_engine, class_=AsyncSession, expire_on_commit=False)()
|
||||||
|
|
||||||
|
|
||||||
@app.middleware("response")
|
@app.middleware("response")
|
||||||
|
@ -345,7 +274,9 @@ from .routes import (
|
||||||
exports,
|
exports,
|
||||||
)
|
)
|
||||||
|
|
||||||
from .routes import tiles, mapdetails
|
if not app.config.LEAN_MODE:
|
||||||
|
from .routes import tiles, mapdetails
|
||||||
|
|
||||||
from .routes import frontend
|
from .routes import frontend
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,68 +0,0 @@
|
||||||
from collections import defaultdict
|
|
||||||
from typing import Dict, FrozenSet, Iterable
|
|
||||||
|
|
||||||
from sanic import Sanic, response
|
|
||||||
from sanic_routing.router import Route
|
|
||||||
|
|
||||||
|
|
||||||
def _add_cors_headers(request, response, methods: Iterable[str]) -> None:
|
|
||||||
allow_methods = list(set(methods))
|
|
||||||
|
|
||||||
if "OPTIONS" not in allow_methods:
|
|
||||||
allow_methods.append("OPTIONS")
|
|
||||||
|
|
||||||
origin = request.headers.get("origin")
|
|
||||||
if origin in request.app.ctx.cors_origins:
|
|
||||||
headers = {
|
|
||||||
"Access-Control-Allow-Methods": ",".join(allow_methods),
|
|
||||||
"Access-Control-Allow-Origin": origin,
|
|
||||||
"Access-Control-Allow-Credentials": "true",
|
|
||||||
"Access-Control-Allow-Headers": (
|
|
||||||
"origin, content-type, accept, "
|
|
||||||
"authorization, x-xsrf-token, x-request-id"
|
|
||||||
),
|
|
||||||
"Access-Control-Expose-Headers": "content-disposition",
|
|
||||||
}
|
|
||||||
response.headers.extend(headers)
|
|
||||||
|
|
||||||
|
|
||||||
def add_cors_headers(request, response):
|
|
||||||
if request.method != "OPTIONS":
|
|
||||||
methods = [method for method in request.route.methods]
|
|
||||||
_add_cors_headers(request, response, methods)
|
|
||||||
|
|
||||||
|
|
||||||
def _compile_routes_needing_options(routes: Dict[str, Route]) -> Dict[str, FrozenSet]:
|
|
||||||
needs_options = defaultdict(list)
|
|
||||||
# This is 21.12 and later. You will need to change this for older versions.
|
|
||||||
for route in routes.values():
|
|
||||||
if "OPTIONS" not in route.methods:
|
|
||||||
needs_options[route.uri].extend(route.methods)
|
|
||||||
|
|
||||||
return {uri: frozenset(methods) for uri, methods in dict(needs_options).items()}
|
|
||||||
|
|
||||||
|
|
||||||
def _options_wrapper(handler, methods):
|
|
||||||
def wrapped_handler(request, *args, **kwargs):
|
|
||||||
nonlocal methods
|
|
||||||
return handler(request, methods)
|
|
||||||
|
|
||||||
return wrapped_handler
|
|
||||||
|
|
||||||
|
|
||||||
async def options_handler(request, methods) -> response.HTTPResponse:
|
|
||||||
resp = response.empty()
|
|
||||||
_add_cors_headers(request, resp, methods)
|
|
||||||
return resp
|
|
||||||
|
|
||||||
|
|
||||||
def setup_options(app: Sanic, _):
|
|
||||||
app.router.reset()
|
|
||||||
needs_options = _compile_routes_needing_options(app.router.routes_all)
|
|
||||||
for uri, methods in needs_options.items():
|
|
||||||
app.add_route(
|
|
||||||
_options_wrapper(options_handler, methods),
|
|
||||||
uri,
|
|
||||||
methods=["OPTIONS"],
|
|
||||||
)
|
|
||||||
app.router.finalize()
|
|
|
@ -3,7 +3,7 @@ from contextvars import ContextVar
|
||||||
from contextlib import asynccontextmanager
|
from contextlib import asynccontextmanager
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
import os
|
import os
|
||||||
from os.path import exists, join, dirname
|
from os.path import join, dirname
|
||||||
from json import loads
|
from json import loads
|
||||||
import re
|
import re
|
||||||
import math
|
import math
|
||||||
|
@ -12,7 +12,6 @@ import random
|
||||||
import string
|
import string
|
||||||
import secrets
|
import secrets
|
||||||
from slugify import slugify
|
from slugify import slugify
|
||||||
import logging
|
|
||||||
|
|
||||||
from sqlalchemy.ext.declarative import declarative_base
|
from sqlalchemy.ext.declarative import declarative_base
|
||||||
from sqlalchemy.ext.asyncio import AsyncSession
|
from sqlalchemy.ext.asyncio import AsyncSession
|
||||||
|
@ -34,13 +33,10 @@ from sqlalchemy import (
|
||||||
select,
|
select,
|
||||||
text,
|
text,
|
||||||
literal,
|
literal,
|
||||||
Text,
|
|
||||||
)
|
)
|
||||||
from sqlalchemy.dialects.postgresql import UUID
|
from sqlalchemy.dialects.postgresql import HSTORE, UUID
|
||||||
|
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
Base = declarative_base()
|
Base = declarative_base()
|
||||||
|
|
||||||
|
|
||||||
|
@ -108,28 +104,6 @@ class Geometry(UserDefinedType):
|
||||||
return func.ST_AsGeoJSON(func.ST_Transform(col, 4326), type_=self)
|
return func.ST_AsGeoJSON(func.ST_Transform(col, 4326), type_=self)
|
||||||
|
|
||||||
|
|
||||||
class LineString(UserDefinedType):
|
|
||||||
def get_col_spec(self):
|
|
||||||
return "geometry(LineString, 3857)"
|
|
||||||
|
|
||||||
def bind_expression(self, bindvalue):
|
|
||||||
return func.ST_GeomFromGeoJSON(bindvalue, type_=self)
|
|
||||||
|
|
||||||
def column_expression(self, col):
|
|
||||||
return func.ST_AsGeoJSON(func.ST_Transform(col, 4326), type_=self)
|
|
||||||
|
|
||||||
|
|
||||||
class GeometryGeometry(UserDefinedType):
|
|
||||||
def get_col_spec(self):
|
|
||||||
return "geometry(GEOMETRY, 3857)"
|
|
||||||
|
|
||||||
def bind_expression(self, bindvalue):
|
|
||||||
return func.ST_GeomFromGeoJSON(bindvalue, type_=self)
|
|
||||||
|
|
||||||
def column_expression(self, col):
|
|
||||||
return func.ST_AsGeoJSON(func.ST_Transform(col, 4326), type_=self)
|
|
||||||
|
|
||||||
|
|
||||||
class OvertakingEvent(Base):
|
class OvertakingEvent(Base):
|
||||||
__tablename__ = "overtaking_event"
|
__tablename__ = "overtaking_event"
|
||||||
__table_args__ = (Index("road_segment", "way_id", "direction_reversed"),)
|
__table_args__ = (Index("road_segment", "way_id", "direction_reversed"),)
|
||||||
|
@ -157,23 +131,12 @@ class OvertakingEvent(Base):
|
||||||
|
|
||||||
class Road(Base):
|
class Road(Base):
|
||||||
__tablename__ = "road"
|
__tablename__ = "road"
|
||||||
way_id = Column(BIGINT, primary_key=True, index=True, autoincrement=False)
|
way_id = Column(BIGINT, primary_key=True, index=True)
|
||||||
zone = Column(ZoneType)
|
zone = Column(ZoneType)
|
||||||
name = Column(Text)
|
name = Column(String)
|
||||||
geometry = Column(LineString)
|
geometry = Column(Geometry)
|
||||||
directionality = Column(Integer)
|
directionality = Column(Integer)
|
||||||
oneway = Column(Boolean)
|
oneway = Column(Boolean)
|
||||||
import_group = Column(String)
|
|
||||||
|
|
||||||
__table_args__ = (
|
|
||||||
# We keep the index name as osm2pgsql created it, way back when.
|
|
||||||
Index(
|
|
||||||
"road_geometry_idx",
|
|
||||||
"geometry",
|
|
||||||
postgresql_using="gist",
|
|
||||||
postgresql_with={"fillfactor": 100},
|
|
||||||
),
|
|
||||||
)
|
|
||||||
|
|
||||||
def to_dict(self):
|
def to_dict(self):
|
||||||
return {
|
return {
|
||||||
|
@ -200,12 +163,6 @@ class RoadUsage(Base):
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return f"<RoadUsage {self.id}>"
|
return f"<RoadUsage {self.id}>"
|
||||||
|
|
||||||
def __hash__(self):
|
|
||||||
return int(self.hex_hash, 16)
|
|
||||||
|
|
||||||
def __eq__(self, other):
|
|
||||||
return self.hex_hash == other.hex_hash
|
|
||||||
|
|
||||||
|
|
||||||
NOW = text("NOW()")
|
NOW = text("NOW()")
|
||||||
|
|
||||||
|
@ -261,12 +218,6 @@ class Track(Base):
|
||||||
Integer, ForeignKey("user.id", ondelete="CASCADE"), nullable=False
|
Integer, ForeignKey("user.id", ondelete="CASCADE"), nullable=False
|
||||||
)
|
)
|
||||||
|
|
||||||
user_device_id = Column(
|
|
||||||
Integer,
|
|
||||||
ForeignKey("user_device.id", ondelete="RESTRICT"),
|
|
||||||
nullable=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Statistics... maybe we'll drop some of this if we can easily compute them from SQL
|
# Statistics... maybe we'll drop some of this if we can easily compute them from SQL
|
||||||
recorded_at = Column(DateTime)
|
recorded_at = Column(DateTime)
|
||||||
recorded_until = Column(DateTime)
|
recorded_until = Column(DateTime)
|
||||||
|
@ -299,7 +250,6 @@ class Track(Base):
|
||||||
if for_user_id is not None and for_user_id == self.author_id:
|
if for_user_id is not None and for_user_id == self.author_id:
|
||||||
result["uploadedByUserAgent"] = self.uploaded_by_user_agent
|
result["uploadedByUserAgent"] = self.uploaded_by_user_agent
|
||||||
result["originalFileName"] = self.original_file_name
|
result["originalFileName"] = self.original_file_name
|
||||||
result["userDeviceId"] = self.user_device_id
|
|
||||||
|
|
||||||
if self.author:
|
if self.author:
|
||||||
result["author"] = self.author.to_dict(for_user_id=for_user_id)
|
result["author"] = self.author.to_dict(for_user_id=for_user_id)
|
||||||
|
@ -401,7 +351,6 @@ class User(Base):
|
||||||
updated_at = Column(DateTime, nullable=False, server_default=NOW, onupdate=NOW)
|
updated_at = Column(DateTime, nullable=False, server_default=NOW, onupdate=NOW)
|
||||||
sub = Column(String, unique=True, nullable=False)
|
sub = Column(String, unique=True, nullable=False)
|
||||||
username = Column(String, unique=True, nullable=False)
|
username = Column(String, unique=True, nullable=False)
|
||||||
display_name = Column(String, nullable=True)
|
|
||||||
email = Column(String, nullable=False)
|
email = Column(String, nullable=False)
|
||||||
bio = Column(TEXT)
|
bio = Column(TEXT)
|
||||||
image = Column(String)
|
image = Column(String)
|
||||||
|
@ -409,7 +358,7 @@ class User(Base):
|
||||||
api_key = Column(String)
|
api_key = Column(String)
|
||||||
|
|
||||||
# This user can be matched by the email address from the auth service
|
# This user can be matched by the email address from the auth service
|
||||||
# instead of having to match by `sub`. If a matching user logs in, the
|
# instead of having to match by `sub`. If a matching user logs in, the
|
||||||
# `sub` is updated to the new sub and this flag is disabled. This is for
|
# `sub` is updated to the new sub and this flag is disabled. This is for
|
||||||
# migrating *to* the external authentication scheme.
|
# migrating *to* the external authentication scheme.
|
||||||
match_by_username_email = Column(Boolean, server_default=false())
|
match_by_username_email = Column(Boolean, server_default=false())
|
||||||
|
@ -422,60 +371,11 @@ class User(Base):
|
||||||
self.api_key = secrets.token_urlsafe(24)
|
self.api_key = secrets.token_urlsafe(24)
|
||||||
|
|
||||||
def to_dict(self, for_user_id=None):
|
def to_dict(self, for_user_id=None):
|
||||||
result = {
|
return {
|
||||||
"id": self.id,
|
"username": self.username,
|
||||||
"displayName": self.display_name or self.username,
|
|
||||||
"bio": self.bio,
|
"bio": self.bio,
|
||||||
"image": self.image,
|
"image": self.image,
|
||||||
}
|
}
|
||||||
if for_user_id == self.id:
|
|
||||||
result["username"] = self.username
|
|
||||||
return result
|
|
||||||
|
|
||||||
async def rename(self, config, new_name):
|
|
||||||
old_name = self.username
|
|
||||||
|
|
||||||
renames = [
|
|
||||||
(join(basedir, old_name), join(basedir, new_name))
|
|
||||||
for basedir in [config.PROCESSING_OUTPUT_DIR, config.TRACKS_DIR]
|
|
||||||
]
|
|
||||||
|
|
||||||
for src, dst in renames:
|
|
||||||
if exists(dst):
|
|
||||||
raise FileExistsError(
|
|
||||||
f"cannot move {src!r} to {dst!r}, destination exists"
|
|
||||||
)
|
|
||||||
|
|
||||||
for src, dst in renames:
|
|
||||||
if not exists(src):
|
|
||||||
log.debug("Rename user %s: Not moving %s, not found", self.id, src)
|
|
||||||
else:
|
|
||||||
log.info("Rename user %s: Moving %s to %s", self.id, src, dst)
|
|
||||||
os.rename(src, dst)
|
|
||||||
|
|
||||||
self.username = new_name
|
|
||||||
|
|
||||||
|
|
||||||
class UserDevice(Base):
|
|
||||||
__tablename__ = "user_device"
|
|
||||||
id = Column(Integer, autoincrement=True, primary_key=True)
|
|
||||||
user_id = Column(Integer, ForeignKey("user.id", ondelete="CASCADE"))
|
|
||||||
identifier = Column(String, nullable=False)
|
|
||||||
display_name = Column(String, nullable=True)
|
|
||||||
|
|
||||||
__table_args__ = (
|
|
||||||
Index("user_id_identifier", "user_id", "identifier", unique=True),
|
|
||||||
)
|
|
||||||
|
|
||||||
def to_dict(self, for_user_id=None):
|
|
||||||
if for_user_id != self.user_id:
|
|
||||||
return {}
|
|
||||||
|
|
||||||
return {
|
|
||||||
"id": self.id,
|
|
||||||
"identifier": self.identifier,
|
|
||||||
"displayName": self.display_name,
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class Comment(Base):
|
class Comment(Base):
|
||||||
|
@ -501,58 +401,24 @@ class Comment(Base):
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class Region(Base):
|
|
||||||
__tablename__ = "region"
|
|
||||||
|
|
||||||
id = Column(String(24), primary_key=True, index=True)
|
|
||||||
name = Column(Text)
|
|
||||||
geometry = Column(GeometryGeometry)
|
|
||||||
admin_level = Column(Integer, index=True)
|
|
||||||
import_group = Column(String)
|
|
||||||
|
|
||||||
__table_args__ = (
|
|
||||||
# We keep the index name as osm2pgsql created it, way back when.
|
|
||||||
Index(
|
|
||||||
"region_geometry_idx",
|
|
||||||
"geometry",
|
|
||||||
postgresql_using="gist",
|
|
||||||
postgresql_with={"fillfactor": 100},
|
|
||||||
),
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
Comment.author = relationship("User", back_populates="authored_comments")
|
Comment.author = relationship("User", back_populates="authored_comments")
|
||||||
User.authored_comments = relationship(
|
User.authored_comments = relationship(
|
||||||
"Comment",
|
"Comment", order_by=Comment.created_at, back_populates="author"
|
||||||
order_by=Comment.created_at,
|
|
||||||
back_populates="author",
|
|
||||||
passive_deletes=True,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
Track.author = relationship("User", back_populates="authored_tracks")
|
Track.author = relationship("User", back_populates="authored_tracks")
|
||||||
User.authored_tracks = relationship(
|
User.authored_tracks = relationship(
|
||||||
"Track", order_by=Track.created_at, back_populates="author", passive_deletes=True
|
"Track", order_by=Track.created_at, back_populates="author"
|
||||||
)
|
)
|
||||||
|
|
||||||
Comment.track = relationship("Track", back_populates="comments")
|
Comment.track = relationship("Track", back_populates="comments")
|
||||||
Track.comments = relationship(
|
Track.comments = relationship(
|
||||||
"Comment", order_by=Comment.created_at, back_populates="track", passive_deletes=True
|
"Comment", order_by=Comment.created_at, back_populates="track"
|
||||||
)
|
)
|
||||||
|
|
||||||
OvertakingEvent.track = relationship("Track", back_populates="overtaking_events")
|
OvertakingEvent.track = relationship("Track", back_populates="overtaking_events")
|
||||||
Track.overtaking_events = relationship(
|
Track.overtaking_events = relationship(
|
||||||
"OvertakingEvent",
|
"OvertakingEvent", order_by=OvertakingEvent.time, back_populates="track"
|
||||||
order_by=OvertakingEvent.time,
|
|
||||||
back_populates="track",
|
|
||||||
passive_deletes=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
Track.user_device = relationship("UserDevice", back_populates="tracks")
|
|
||||||
UserDevice.tracks = relationship(
|
|
||||||
"Track",
|
|
||||||
order_by=Track.created_at,
|
|
||||||
back_populates="user_device",
|
|
||||||
passive_deletes=False,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -8,7 +8,7 @@ import pytz
|
||||||
from os.path import join
|
from os.path import join
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
|
||||||
from sqlalchemy import delete, func, select, and_
|
from sqlalchemy import delete, select
|
||||||
from sqlalchemy.orm import joinedload
|
from sqlalchemy.orm import joinedload
|
||||||
|
|
||||||
from obs.face.importer import ImportMeasurementsCsv
|
from obs.face.importer import ImportMeasurementsCsv
|
||||||
|
@ -25,9 +25,9 @@ from obs.face.filter import (
|
||||||
RequiredFieldsFilter,
|
RequiredFieldsFilter,
|
||||||
)
|
)
|
||||||
|
|
||||||
from obs.face.osm import DataSource, DatabaseTileSource
|
from obs.face.osm import DataSource, DatabaseTileSource, OverpassTileSource
|
||||||
|
|
||||||
from obs.api.db import OvertakingEvent, RoadUsage, Track, UserDevice, make_session
|
from obs.api.db import OvertakingEvent, RoadUsage, Track, make_session
|
||||||
from obs.api.app import app
|
from obs.api.app import app
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
@ -39,7 +39,12 @@ def get_data_source():
|
||||||
mode, the OverpassTileSource is used to fetch data on demand. In normal
|
mode, the OverpassTileSource is used to fetch data on demand. In normal
|
||||||
mode, the roads database is used.
|
mode, the roads database is used.
|
||||||
"""
|
"""
|
||||||
return DataSource(DatabaseTileSource())
|
if app.config.LEAN_MODE:
|
||||||
|
tile_source = OverpassTileSource(cache_dir=app.config.OBS_FACE_CACHE_DIR)
|
||||||
|
else:
|
||||||
|
tile_source = DatabaseTileSource()
|
||||||
|
|
||||||
|
return DataSource(tile_source)
|
||||||
|
|
||||||
|
|
||||||
async def process_tracks_loop(delay):
|
async def process_tracks_loop(delay):
|
||||||
|
@ -139,11 +144,10 @@ async def process_track(session, track, data_source):
|
||||||
os.makedirs(output_dir, exist_ok=True)
|
os.makedirs(output_dir, exist_ok=True)
|
||||||
|
|
||||||
log.info("Annotating and filtering CSV file")
|
log.info("Annotating and filtering CSV file")
|
||||||
imported_data, statistics, track_metadata = ImportMeasurementsCsv().read(
|
imported_data, statistics = ImportMeasurementsCsv().read(
|
||||||
original_file_path,
|
original_file_path,
|
||||||
user_id="dummy", # TODO: user username or id or nothing?
|
user_id="dummy", # TODO: user username or id or nothing?
|
||||||
dataset_id=Track.slug, # TODO: use track id or slug or nothing?
|
dataset_id=Track.slug, # TODO: use track id or slug or nothing?
|
||||||
return_metadata=True,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
annotator = AnnotateMeasurements(
|
annotator = AnnotateMeasurements(
|
||||||
|
@ -213,36 +217,6 @@ async def process_track(session, track, data_source):
|
||||||
await clear_track_data(session, track)
|
await clear_track_data(session, track)
|
||||||
await session.commit()
|
await session.commit()
|
||||||
|
|
||||||
device_identifier = track_metadata.get("DeviceId")
|
|
||||||
if device_identifier:
|
|
||||||
if isinstance(device_identifier, list):
|
|
||||||
device_identifier = device_identifier[0]
|
|
||||||
|
|
||||||
log.info("Finding or creating device %s", device_identifier)
|
|
||||||
user_device = (
|
|
||||||
await session.execute(
|
|
||||||
select(UserDevice).where(
|
|
||||||
and_(
|
|
||||||
UserDevice.user_id == track.author_id,
|
|
||||||
UserDevice.identifier == device_identifier,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
).scalar()
|
|
||||||
|
|
||||||
log.debug("user_device is %s", user_device)
|
|
||||||
|
|
||||||
if not user_device:
|
|
||||||
user_device = UserDevice(
|
|
||||||
user_id=track.author_id, identifier=device_identifier
|
|
||||||
)
|
|
||||||
log.debug("Create new device for this user")
|
|
||||||
session.add(user_device)
|
|
||||||
|
|
||||||
track.user_device = user_device
|
|
||||||
else:
|
|
||||||
log.info("No DeviceId in track metadata.")
|
|
||||||
|
|
||||||
log.info("Import events into database...")
|
log.info("Import events into database...")
|
||||||
await import_overtaking_events(session, track, overtaking_events)
|
await import_overtaking_events(session, track, overtaking_events)
|
||||||
|
|
||||||
|
@ -306,16 +280,11 @@ async def import_overtaking_events(session, track, overtaking_events):
|
||||||
hex_hash=hex_hash,
|
hex_hash=hex_hash,
|
||||||
way_id=m.get("OSM_way_id"),
|
way_id=m.get("OSM_way_id"),
|
||||||
direction_reversed=m.get("OSM_way_orientation", 0) < 0,
|
direction_reversed=m.get("OSM_way_orientation", 0) < 0,
|
||||||
geometry=func.ST_Transform(
|
geometry=json.dumps(
|
||||||
func.ST_GeomFromGeoJSON(
|
{
|
||||||
json.dumps(
|
"type": "Point",
|
||||||
{
|
"coordinates": [m["longitude"], m["latitude"]],
|
||||||
"type": "Point",
|
}
|
||||||
"coordinates": [m["longitude"], m["latitude"]],
|
|
||||||
}
|
|
||||||
)
|
|
||||||
),
|
|
||||||
3857,
|
|
||||||
),
|
),
|
||||||
latitude=m["latitude"],
|
latitude=m["latitude"],
|
||||||
longitude=m["longitude"],
|
longitude=m["longitude"],
|
||||||
|
|
|
@ -3,22 +3,15 @@ from enum import Enum
|
||||||
from contextlib import contextmanager
|
from contextlib import contextmanager
|
||||||
import zipfile
|
import zipfile
|
||||||
import io
|
import io
|
||||||
import re
|
|
||||||
import math
|
|
||||||
from sqlite3 import connect
|
from sqlite3 import connect
|
||||||
|
|
||||||
import shapefile
|
import shapefile
|
||||||
from obs.api.db import OvertakingEvent
|
from obs.api.db import OvertakingEvent
|
||||||
from sqlalchemy import select, func, text
|
from sqlalchemy import select, func
|
||||||
from sanic.response import raw
|
from sanic.response import raw
|
||||||
from sanic.exceptions import InvalidUsage
|
from sanic.exceptions import InvalidUsage
|
||||||
|
|
||||||
from obs.api.app import api, json as json_response
|
from obs.api.app import api, json as json_response
|
||||||
from obs.api.utils import use_request_semaphore
|
|
||||||
|
|
||||||
import logging
|
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class ExportFormat(str, Enum):
|
class ExportFormat(str, Enum):
|
||||||
|
@ -33,7 +26,7 @@ def parse_bounding_box(input_string):
|
||||||
func.ST_Point(left, bottom),
|
func.ST_Point(left, bottom),
|
||||||
func.ST_Point(right, top),
|
func.ST_Point(right, top),
|
||||||
),
|
),
|
||||||
4326,
|
3857,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@ -45,11 +38,11 @@ PROJECTION_4326 = (
|
||||||
|
|
||||||
|
|
||||||
@contextmanager
|
@contextmanager
|
||||||
def shapefile_zip(shape_type=shapefile.POINT, basename="events"):
|
def shapefile_zip():
|
||||||
zip_buffer = io.BytesIO()
|
zip_buffer = io.BytesIO()
|
||||||
shp, shx, dbf = (io.BytesIO() for _ in range(3))
|
shp, shx, dbf = (io.BytesIO() for _ in range(3))
|
||||||
writer = shapefile.Writer(
|
writer = shapefile.Writer(
|
||||||
shp=shp, shx=shx, dbf=dbf, shapeType=shape_type, encoding="utf8"
|
shp=shp, shx=shx, dbf=dbf, shapeType=shapefile.POINT, encoding="utf8"
|
||||||
)
|
)
|
||||||
|
|
||||||
yield writer, zip_buffer
|
yield writer, zip_buffer
|
||||||
|
@ -58,204 +51,67 @@ def shapefile_zip(shape_type=shapefile.POINT, basename="events"):
|
||||||
writer.close()
|
writer.close()
|
||||||
|
|
||||||
zip_file = zipfile.ZipFile(zip_buffer, "a", zipfile.ZIP_DEFLATED, False)
|
zip_file = zipfile.ZipFile(zip_buffer, "a", zipfile.ZIP_DEFLATED, False)
|
||||||
zip_file.writestr(f"{basename}.shp", shp.getbuffer())
|
zip_file.writestr("events.shp", shp.getbuffer())
|
||||||
zip_file.writestr(f"{basename}.shx", shx.getbuffer())
|
zip_file.writestr("events.shx", shx.getbuffer())
|
||||||
zip_file.writestr(f"{basename}.dbf", dbf.getbuffer())
|
zip_file.writestr("events.dbf", dbf.getbuffer())
|
||||||
zip_file.writestr(f"{basename}.prj", PROJECTION_4326)
|
zip_file.writestr("events.prj", PROJECTION_4326)
|
||||||
zip_file.close()
|
zip_file.close()
|
||||||
|
|
||||||
|
|
||||||
@api.get(r"/export/events")
|
@api.get(r"/export/events")
|
||||||
async def export_events(req):
|
async def export_events(req):
|
||||||
async with use_request_semaphore(req, "export_semaphore", timeout=30):
|
bbox = req.ctx.get_single_arg(
|
||||||
bbox = req.ctx.get_single_arg("bbox", default="-180,-90,180,90")
|
"bbox", default="-180,-90,180,90", convert=parse_bounding_box
|
||||||
assert re.match(r"(-?\d+\.?\d+,?){4}", bbox)
|
)
|
||||||
bbox = list(map(float, bbox.split(",")))
|
fmt = req.ctx.get_single_arg("fmt", convert=ExportFormat)
|
||||||
|
|
||||||
fmt = req.ctx.get_single_arg("fmt", convert=ExportFormat)
|
events = await req.ctx.db.stream_scalars(
|
||||||
|
select(OvertakingEvent).where(OvertakingEvent.geometry.bool_op("&&")(bbox))
|
||||||
|
)
|
||||||
|
|
||||||
events = await req.ctx.db.stream(
|
if fmt == ExportFormat.SHAPEFILE:
|
||||||
text(
|
with shapefile_zip() as (writer, zip_buffer):
|
||||||
"""
|
writer.field("distance_overtaker", "N", decimal=4)
|
||||||
SELECT
|
writer.field("distance_stationary", "N", decimal=4)
|
||||||
ST_AsGeoJSON(ST_Transform(geometry, 4326)) AS geometry,
|
writer.field("way_id", "N", decimal=0)
|
||||||
distance_overtaker,
|
writer.field("direction", "N", decimal=0)
|
||||||
distance_stationary,
|
writer.field("course", "N", decimal=4)
|
||||||
way_id,
|
writer.field("speed", "N", decimal=4)
|
||||||
direction,
|
|
||||||
speed,
|
|
||||||
time_stamp,
|
|
||||||
course,
|
|
||||||
zone
|
|
||||||
FROM
|
|
||||||
layer_obs_events(
|
|
||||||
ST_Transform(ST_MakeEnvelope(:bbox0, :bbox1, :bbox2, :bbox3, 4326), 3857),
|
|
||||||
19,
|
|
||||||
NULL,
|
|
||||||
'1900-01-01'::timestamp,
|
|
||||||
'2100-01-01'::timestamp
|
|
||||||
)
|
|
||||||
"""
|
|
||||||
).bindparams(bbox0=bbox[0], bbox1=bbox[1], bbox2=bbox[2], bbox3=bbox[3])
|
|
||||||
)
|
|
||||||
|
|
||||||
if fmt == ExportFormat.SHAPEFILE:
|
|
||||||
with shapefile_zip(basename="events") as (writer, zip_buffer):
|
|
||||||
writer.field("distance_overtaker", "N", decimal=4)
|
|
||||||
writer.field("distance_stationary", "N", decimal=4)
|
|
||||||
writer.field("way_id", "N", decimal=0)
|
|
||||||
writer.field("direction", "N", decimal=0)
|
|
||||||
writer.field("course", "N", decimal=4)
|
|
||||||
writer.field("speed", "N", decimal=4)
|
|
||||||
writer.field("zone", "C")
|
|
||||||
|
|
||||||
async for event in events:
|
|
||||||
coords = json.loads(event.geometry)["coordinates"]
|
|
||||||
writer.point(*coords)
|
|
||||||
writer.record(
|
|
||||||
distance_overtaker=event.distance_overtaker,
|
|
||||||
distance_stationary=event.distance_stationary,
|
|
||||||
direction=event.direction,
|
|
||||||
way_id=event.way_id,
|
|
||||||
course=event.course,
|
|
||||||
speed=event.speed,
|
|
||||||
zone=event.zone
|
|
||||||
# "time"=event.time,
|
|
||||||
)
|
|
||||||
|
|
||||||
return raw(zip_buffer.getbuffer())
|
|
||||||
|
|
||||||
if fmt == ExportFormat.GEOJSON:
|
|
||||||
features = []
|
|
||||||
async for event in events:
|
async for event in events:
|
||||||
geom = json.loads(event.geometry)
|
writer.point(event.longitude, event.latitude)
|
||||||
features.append(
|
writer.record(
|
||||||
{
|
distance_overtaker=event.distance_overtaker,
|
||||||
"type": "Feature",
|
distance_stationary=event.distance_stationary,
|
||||||
"geometry": geom,
|
direction=-1 if event.direction_reversed else 1,
|
||||||
"properties": {
|
way_id=event.way_id,
|
||||||
"distance_overtaker": event.distance_overtaker
|
course=event.course,
|
||||||
if event.distance_overtaker is not None
|
speed=event.speed,
|
||||||
and not math.isnan(event.distance_overtaker)
|
# "time"=event.time,
|
||||||
else None,
|
|
||||||
"distance_stationary": event.distance_stationary
|
|
||||||
if event.distance_stationary is not None
|
|
||||||
and not math.isnan(event.distance_stationary)
|
|
||||||
else None,
|
|
||||||
"direction": event.direction
|
|
||||||
if event.direction is not None
|
|
||||||
and not math.isnan(event.direction)
|
|
||||||
else None,
|
|
||||||
"way_id": event.way_id,
|
|
||||||
"course": event.course
|
|
||||||
if event.course is not None and not math.isnan(event.course)
|
|
||||||
else None,
|
|
||||||
"speed": event.speed
|
|
||||||
if event.speed is not None and not math.isnan(event.speed)
|
|
||||||
else None,
|
|
||||||
"time": event.time_stamp,
|
|
||||||
"zone": event.zone,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
)
|
)
|
||||||
|
|
||||||
geojson = {"type": "FeatureCollection", "features": features}
|
return raw(zip_buffer.getbuffer())
|
||||||
return json_response(geojson)
|
|
||||||
|
|
||||||
raise InvalidUsage("unknown export format")
|
if fmt == ExportFormat.GEOJSON:
|
||||||
|
features = []
|
||||||
|
async for event in events:
|
||||||
|
features.append(
|
||||||
|
{
|
||||||
|
"type": "Feature",
|
||||||
|
"geometry": json.loads(event.geometry),
|
||||||
|
"properties": {
|
||||||
|
"distance_overtaker": event.distance_overtaker,
|
||||||
|
"distance_stationary": event.distance_stationary,
|
||||||
|
"direction": -1 if event.direction_reversed else 1,
|
||||||
|
"way_id": event.way_id,
|
||||||
|
"course": event.course,
|
||||||
|
"speed": event.speed,
|
||||||
|
"time": event.time,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
geojson = {"type": "FeatureCollection", "features": features}
|
||||||
|
return json_response(geojson)
|
||||||
|
|
||||||
@api.get(r"/export/segments")
|
raise InvalidUsage("unknown export format")
|
||||||
async def export_segments(req):
|
|
||||||
async with use_request_semaphore(req, "export_semaphore", timeout=30):
|
|
||||||
bbox = req.ctx.get_single_arg("bbox", default="-180,-90,180,90")
|
|
||||||
assert re.match(r"(-?\d+\.?\d+,?){4}", bbox)
|
|
||||||
bbox = list(map(float, bbox.split(",")))
|
|
||||||
|
|
||||||
fmt = req.ctx.get_single_arg("fmt", convert=ExportFormat)
|
|
||||||
segments = await req.ctx.db.stream(
|
|
||||||
text(
|
|
||||||
"""
|
|
||||||
SELECT
|
|
||||||
ST_AsGeoJSON(ST_Transform(geometry, 4326)) AS geometry,
|
|
||||||
way_id,
|
|
||||||
distance_overtaker_mean,
|
|
||||||
distance_overtaker_min,
|
|
||||||
distance_overtaker_max,
|
|
||||||
distance_overtaker_median,
|
|
||||||
overtaking_event_count,
|
|
||||||
usage_count,
|
|
||||||
direction,
|
|
||||||
zone,
|
|
||||||
offset_direction,
|
|
||||||
distance_overtaker_array
|
|
||||||
FROM
|
|
||||||
layer_obs_roads(
|
|
||||||
ST_Transform(ST_MakeEnvelope(:bbox0, :bbox1, :bbox2, :bbox3, 4326), 3857),
|
|
||||||
11,
|
|
||||||
NULL,
|
|
||||||
'1900-01-01'::timestamp,
|
|
||||||
'2100-01-01'::timestamp
|
|
||||||
)
|
|
||||||
WHERE usage_count > 0
|
|
||||||
"""
|
|
||||||
).bindparams(bbox0=bbox[0], bbox1=bbox[1], bbox2=bbox[2], bbox3=bbox[3])
|
|
||||||
)
|
|
||||||
|
|
||||||
if fmt == ExportFormat.SHAPEFILE:
|
|
||||||
with shapefile_zip(shape_type=3, basename="segments") as (
|
|
||||||
writer,
|
|
||||||
zip_buffer,
|
|
||||||
):
|
|
||||||
writer.field("distance_overtaker_mean", "N", decimal=4)
|
|
||||||
writer.field("distance_overtaker_max", "N", decimal=4)
|
|
||||||
writer.field("distance_overtaker_min", "N", decimal=4)
|
|
||||||
writer.field("distance_overtaker_median", "N", decimal=4)
|
|
||||||
writer.field("overtaking_event_count", "N", decimal=4)
|
|
||||||
writer.field("usage_count", "N", decimal=4)
|
|
||||||
writer.field("way_id", "N", decimal=0)
|
|
||||||
writer.field("direction", "N", decimal=0)
|
|
||||||
writer.field("zone", "C")
|
|
||||||
|
|
||||||
async for segment in segments:
|
|
||||||
geom = json.loads(segment.st_asgeojson)
|
|
||||||
writer.line([geom["coordinates"]])
|
|
||||||
writer.record(
|
|
||||||
distance_overtaker_mean=segment.distance_overtaker_mean,
|
|
||||||
distance_overtaker_median=segment.distance_overtaker_median,
|
|
||||||
distance_overtaker_max=segment.distance_overtaker_max,
|
|
||||||
distance_overtaker_min=segment.distance_overtaker_min,
|
|
||||||
usage_count=segment.usage_count,
|
|
||||||
overtaking_event_count=segment.overtaking_event_count,
|
|
||||||
direction=segment.direction,
|
|
||||||
way_id=segment.way_id,
|
|
||||||
zone=segment.zone,
|
|
||||||
)
|
|
||||||
|
|
||||||
return raw(zip_buffer.getbuffer())
|
|
||||||
|
|
||||||
if fmt == ExportFormat.GEOJSON:
|
|
||||||
features = []
|
|
||||||
async for segment in segments:
|
|
||||||
features.append(
|
|
||||||
{
|
|
||||||
"type": "Feature",
|
|
||||||
"geometry": json.loads(segment.geometry),
|
|
||||||
"properties": {
|
|
||||||
"distance_overtaker_mean": segment.distance_overtaker_mean,
|
|
||||||
"distance_overtaker_max": segment.distance_overtaker_max,
|
|
||||||
"distance_overtaker_median": segment.distance_overtaker_median,
|
|
||||||
"overtaking_event_count": segment.overtaking_event_count,
|
|
||||||
"usage_count": segment.usage_count,
|
|
||||||
"distance_overtaker_array": segment.distance_overtaker_array,
|
|
||||||
"direction": segment.direction,
|
|
||||||
"way_id": segment.way_id,
|
|
||||||
"zone": segment.zone,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
geojson = {"type": "FeatureCollection", "features": features}
|
|
||||||
return json_response(geojson)
|
|
||||||
|
|
||||||
raise InvalidUsage("unknown export format")
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
from os.path import join, exists, isfile, abspath
|
from os.path import join, exists, isfile
|
||||||
|
|
||||||
import sanic.response as response
|
import sanic.response as response
|
||||||
from sanic.exceptions import NotFound
|
from sanic.exceptions import NotFound
|
||||||
|
@ -14,18 +14,22 @@ if app.config.FRONTEND_CONFIG:
|
||||||
**req.app.config.FRONTEND_CONFIG,
|
**req.app.config.FRONTEND_CONFIG,
|
||||||
"apiUrl": f"{req.ctx.api_url}/api",
|
"apiUrl": f"{req.ctx.api_url}/api",
|
||||||
"loginUrl": f"{req.ctx.api_url}/login",
|
"loginUrl": f"{req.ctx.api_url}/login",
|
||||||
"obsMapSource": {
|
"obsMapSource": (
|
||||||
"type": "vector",
|
None
|
||||||
"tiles": [
|
if app.config.LEAN_MODE
|
||||||
req.ctx.api_url
|
else {
|
||||||
+ req.app.url_for("tiles", zoom="000", x="111", y="222.pbf")
|
"type": "vector",
|
||||||
.replace("000", "{z}")
|
"tiles": [
|
||||||
.replace("111", "{x}")
|
req.ctx.api_url
|
||||||
.replace("222", "{y}")
|
+ req.app.url_for("tiles", zoom="000", x="111", y="222.pbf")
|
||||||
],
|
.replace("000", "{z}")
|
||||||
"minzoom": 0,
|
.replace("111", "{x}")
|
||||||
"maxzoom": 14,
|
.replace("222", "{y}")
|
||||||
},
|
],
|
||||||
|
"minzoom": 0,
|
||||||
|
"maxzoom": 14,
|
||||||
|
}
|
||||||
|
),
|
||||||
}
|
}
|
||||||
|
|
||||||
return response.json(result)
|
return response.json(result)
|
||||||
|
@ -46,9 +50,6 @@ if INDEX_HTML and exists(INDEX_HTML):
|
||||||
raise NotFound()
|
raise NotFound()
|
||||||
|
|
||||||
file = join(app.config.FRONTEND_DIR, path)
|
file = join(app.config.FRONTEND_DIR, path)
|
||||||
if not abspath(file).startswith(abspath(app.config.FRONTEND_DIR)):
|
|
||||||
raise NotFound()
|
|
||||||
|
|
||||||
if not exists(file) or not path or not isfile(file):
|
if not exists(file) or not path or not isfile(file):
|
||||||
return response.html(
|
return response.html(
|
||||||
index_file_contents.replace("__BASE_HREF__", req.ctx.frontend_url + "/")
|
index_file_contents.replace("__BASE_HREF__", req.ctx.frontend_url + "/")
|
||||||
|
|
|
@ -1,6 +1,5 @@
|
||||||
import asyncio
|
import asyncio
|
||||||
import logging
|
import logging
|
||||||
import re
|
|
||||||
|
|
||||||
from requests.exceptions import RequestException
|
from requests.exceptions import RequestException
|
||||||
|
|
||||||
|
@ -92,15 +91,6 @@ async def login_redirect(req):
|
||||||
preferred_username = userinfo["preferred_username"]
|
preferred_username = userinfo["preferred_username"]
|
||||||
email = userinfo.get("email")
|
email = userinfo.get("email")
|
||||||
|
|
||||||
clean_username = re.sub(r"[^a-zA-Z0-9_.-]", "", preferred_username)
|
|
||||||
if clean_username != preferred_username:
|
|
||||||
log.warning(
|
|
||||||
"Username %r contained invalid characters and was changed to %r",
|
|
||||||
preferred_username,
|
|
||||||
clean_username,
|
|
||||||
)
|
|
||||||
preferred_username = clean_username
|
|
||||||
|
|
||||||
if email is None:
|
if email is None:
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
"user has no email set, please configure keycloak to require emails"
|
"user has no email set, please configure keycloak to require emails"
|
||||||
|
@ -138,20 +128,16 @@ async def login_redirect(req):
|
||||||
user = User(sub=sub, username=preferred_username, email=email)
|
user = User(sub=sub, username=preferred_username, email=email)
|
||||||
req.ctx.db.add(user)
|
req.ctx.db.add(user)
|
||||||
else:
|
else:
|
||||||
log.info(
|
log.info("Logged in known user (id: %s, sub: %s).", user.id, user.sub)
|
||||||
"Logged in known user (id: %s, sub: %s, %s).",
|
|
||||||
user.id,
|
|
||||||
user.sub,
|
|
||||||
preferred_username,
|
|
||||||
)
|
|
||||||
|
|
||||||
if email != user.email:
|
if email != user.email:
|
||||||
log.debug("Updating user (id: %s) email from auth system.", user.id)
|
log.debug("Updating user (id: %s) email from auth system.", user.id)
|
||||||
user.email = email
|
user.email = email
|
||||||
|
|
||||||
if preferred_username != user.username:
|
# TODO: re-add username change when we can safely rename users
|
||||||
log.debug("Updating user (id: %s) username from auth system.", user.id)
|
# if preferred_username != user.username:
|
||||||
await user.rename(req.app.config, preferred_username)
|
# log.debug("Updating user (id: %s) username from auth system.", user.id)
|
||||||
|
# user.username = preferred_username
|
||||||
|
|
||||||
await req.ctx.db.commit()
|
await req.ctx.db.commit()
|
||||||
|
|
||||||
|
@ -170,4 +156,4 @@ async def logout(req):
|
||||||
auth_req = client.construct_EndSessionRequest(state=session["state"])
|
auth_req = client.construct_EndSessionRequest(state=session["state"])
|
||||||
logout_url = auth_req.request(client.end_session_endpoint)
|
logout_url = auth_req.request(client.end_session_endpoint)
|
||||||
|
|
||||||
return redirect(logout_url + f"&post_logout_redirect_uri={req.ctx.api_url}/logout")
|
return redirect(logout_url+f"&redirect_uri={req.ctx.api_url}/logout")
|
||||||
|
|
|
@ -1,6 +1,5 @@
|
||||||
import json
|
import json
|
||||||
from functools import partial
|
from functools import partial
|
||||||
import logging
|
|
||||||
import numpy
|
import numpy
|
||||||
import math
|
import math
|
||||||
|
|
||||||
|
@ -16,22 +15,17 @@ from obs.api.utils import round_to
|
||||||
round_distance = partial(round_to, multiples=0.001)
|
round_distance = partial(round_to, multiples=0.001)
|
||||||
round_speed = partial(round_to, multiples=0.1)
|
round_speed = partial(round_to, multiples=0.1)
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
def get_bearing(a, b):
|
||||||
def get_bearing(b, a):
|
|
||||||
# longitude, latitude
|
# longitude, latitude
|
||||||
dL = b[0] - a[0]
|
dL = b[0] - a[0]
|
||||||
X = numpy.cos(b[1]) * numpy.sin(dL)
|
X = numpy.cos(b[1]) * numpy.sin(dL)
|
||||||
Y = numpy.cos(a[1]) * numpy.sin(b[1]) - numpy.sin(a[1]) * numpy.cos(
|
Y = numpy.cos(a[1]) * numpy.sin(b[1]) - numpy.sin(a[1]) * numpy.cos(
|
||||||
b[1]
|
b[1]
|
||||||
) * numpy.cos(dL)
|
) * numpy.cos(dL)
|
||||||
return numpy.arctan2(Y, X) + 0.5 * math.pi
|
return numpy.arctan2(X, Y)
|
||||||
|
|
||||||
|
|
||||||
# Bins for histogram on overtaker distances. 0, 0.25, ... 2.25, infinity
|
|
||||||
DISTANCE_BINS = numpy.arange(0, 2.5, 0.25).tolist() + [float('inf')]
|
|
||||||
|
|
||||||
@api.route("/mapdetails/road", methods=["GET"])
|
@api.route("/mapdetails/road", methods=["GET"])
|
||||||
async def mapdetails_road(req):
|
async def mapdetails_road(req):
|
||||||
longitude = req.ctx.get_single_arg("longitude", convert=float)
|
longitude = req.ctx.get_single_arg("longitude", convert=float)
|
||||||
|
@ -84,25 +78,26 @@ async def mapdetails_road(req):
|
||||||
arrays = numpy.array(arrays).T
|
arrays = numpy.array(arrays).T
|
||||||
|
|
||||||
if len(arrays) == 0:
|
if len(arrays) == 0:
|
||||||
arrays = numpy.array([[], [], [], []], dtype=float)
|
arrays = numpy.array([[], [], [], []], dtype=numpy.float)
|
||||||
|
|
||||||
data, mask = arrays[:-1], arrays[-1]
|
data, mask = arrays[:-1], arrays[-1]
|
||||||
data = data.astype(numpy.float64)
|
data = data.astype(numpy.float64)
|
||||||
mask = mask.astype(bool)
|
mask = mask.astype(numpy.bool)
|
||||||
|
|
||||||
def partition(arr, cond):
|
def partition(arr, cond):
|
||||||
return arr[:, cond], arr[:, ~cond]
|
return arr[:, cond], arr[:, ~cond]
|
||||||
|
|
||||||
forwards, backwards = partition(data, ~mask)
|
forwards, backwards = partition(data, ~mask)
|
||||||
|
print("for", forwards.dtype, "back", backwards.dtype)
|
||||||
|
|
||||||
def array_stats(arr, rounder, bins=30):
|
def array_stats(arr, rounder):
|
||||||
if len(arr):
|
if len(arr):
|
||||||
|
print("ARR DTYPE", arr.dtype)
|
||||||
|
print("ARR", arr)
|
||||||
arr = arr[~numpy.isnan(arr)]
|
arr = arr[~numpy.isnan(arr)]
|
||||||
|
|
||||||
n = len(arr)
|
n = len(arr)
|
||||||
|
|
||||||
hist, bins = numpy.histogram(arr, bins=bins)
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
"statistics": {
|
"statistics": {
|
||||||
"count": n,
|
"count": n,
|
||||||
|
@ -111,11 +106,6 @@ async def mapdetails_road(req):
|
||||||
"max": rounder(numpy.max(arr)) if n else None,
|
"max": rounder(numpy.max(arr)) if n else None,
|
||||||
"median": rounder(numpy.median(arr)) if n else None,
|
"median": rounder(numpy.median(arr)) if n else None,
|
||||||
},
|
},
|
||||||
"histogram": {
|
|
||||||
"bins": [None if math.isinf(b) else b for b in bins.tolist()],
|
|
||||||
"counts": hist.tolist(),
|
|
||||||
"zone": road.zone
|
|
||||||
},
|
|
||||||
"values": list(map(rounder, arr.tolist())),
|
"values": list(map(rounder, arr.tolist())),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -128,13 +118,15 @@ async def mapdetails_road(req):
|
||||||
# convert to degrees, as this is more natural to understand for consumers
|
# convert to degrees, as this is more natural to understand for consumers
|
||||||
bearing = round_to((bearing / math.pi * 180 + 360) % 360, 1)
|
bearing = round_to((bearing / math.pi * 180 + 360) % 360, 1)
|
||||||
|
|
||||||
|
print(road.geometry)
|
||||||
|
|
||||||
def get_direction_stats(direction_arrays, backwards=False):
|
def get_direction_stats(direction_arrays, backwards=False):
|
||||||
return {
|
return {
|
||||||
"bearing": ((bearing + 180) % 360 if backwards else bearing)
|
"bearing": ((bearing + 180) % 360 if backwards else bearing)
|
||||||
if bearing is not None
|
if bearing is not None
|
||||||
else None,
|
else None,
|
||||||
"distanceOvertaker": array_stats(direction_arrays[0], round_distance, bins=DISTANCE_BINS),
|
"distanceOvertaker": array_stats(direction_arrays[0], round_distance),
|
||||||
"distanceStationary": array_stats(direction_arrays[1], round_distance, bins=DISTANCE_BINS),
|
"distanceStationary": array_stats(direction_arrays[1], round_distance),
|
||||||
"speed": array_stats(direction_arrays[2], round_speed),
|
"speed": array_stats(direction_arrays[2], round_speed),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -4,12 +4,12 @@ from typing import Optional
|
||||||
from operator import and_
|
from operator import and_
|
||||||
from functools import reduce
|
from functools import reduce
|
||||||
|
|
||||||
from sqlalchemy import distinct, select, func, desc
|
from sqlalchemy import select, func
|
||||||
|
|
||||||
from sanic.response import json
|
from sanic.response import json
|
||||||
|
|
||||||
from obs.api.app import api
|
from obs.api.app import api
|
||||||
from obs.api.db import Track, OvertakingEvent, User, Region, UserDevice
|
from obs.api.db import Track, OvertakingEvent, User
|
||||||
from obs.api.utils import round_to
|
from obs.api.utils import round_to
|
||||||
|
|
||||||
|
|
||||||
|
@ -45,7 +45,7 @@ async def stats(req):
|
||||||
|
|
||||||
# Only the user can look for their own stats, for now
|
# Only the user can look for their own stats, for now
|
||||||
by_user = (
|
by_user = (
|
||||||
user is not None and req.ctx.user is not None and req.ctx.user.id == int(user)
|
user is not None and req.ctx.user is not None and req.ctx.user.username == user
|
||||||
)
|
)
|
||||||
if by_user:
|
if by_user:
|
||||||
conditions.append(Track.author_id == req.ctx.user.id)
|
conditions.append(Track.author_id == req.ctx.user.id)
|
||||||
|
@ -92,14 +92,6 @@ async def stats(req):
|
||||||
.where(track_condition)
|
.where(track_condition)
|
||||||
)
|
)
|
||||||
).scalar()
|
).scalar()
|
||||||
device_count = (
|
|
||||||
await req.ctx.db.execute(
|
|
||||||
select(func.count(distinct(UserDevice.id)))
|
|
||||||
.select_from(UserDevice)
|
|
||||||
.join(Track.user_device)
|
|
||||||
.where(track_condition)
|
|
||||||
)
|
|
||||||
).scalar()
|
|
||||||
|
|
||||||
result = {
|
result = {
|
||||||
"numEvents": event_count,
|
"numEvents": event_count,
|
||||||
|
@ -108,7 +100,6 @@ async def stats(req):
|
||||||
"trackDuration": round_to(track_duration or 0, TRACK_DURATION_ROUNDING),
|
"trackDuration": round_to(track_duration or 0, TRACK_DURATION_ROUNDING),
|
||||||
"publicTrackCount": public_track_count,
|
"publicTrackCount": public_track_count,
|
||||||
"trackCount": track_count,
|
"trackCount": track_count,
|
||||||
"deviceCount": device_count,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return json(result)
|
return json(result)
|
||||||
|
@ -176,31 +167,3 @@ async def stats(req):
|
||||||
# });
|
# });
|
||||||
# }),
|
# }),
|
||||||
# );
|
# );
|
||||||
|
|
||||||
|
|
||||||
@api.route("/stats/regions")
|
|
||||||
async def stats(req):
|
|
||||||
query = (
|
|
||||||
select(
|
|
||||||
[
|
|
||||||
Region.id,
|
|
||||||
Region.name,
|
|
||||||
func.count(OvertakingEvent.id).label("overtaking_event_count"),
|
|
||||||
]
|
|
||||||
)
|
|
||||||
.select_from(Region)
|
|
||||||
.join(
|
|
||||||
OvertakingEvent,
|
|
||||||
func.ST_Within(OvertakingEvent.geometry, Region.geometry),
|
|
||||||
)
|
|
||||||
.group_by(
|
|
||||||
Region.id,
|
|
||||||
Region.name,
|
|
||||||
Region.geometry,
|
|
||||||
)
|
|
||||||
.having(func.count(OvertakingEvent.id) > 0)
|
|
||||||
.order_by(desc("overtaking_event_count"))
|
|
||||||
)
|
|
||||||
|
|
||||||
regions = list(map(dict, (await req.ctx.db.execute(query)).all()))
|
|
||||||
return json(regions)
|
|
||||||
|
|
|
@ -1,16 +1,11 @@
|
||||||
from gzip import decompress
|
from gzip import decompress
|
||||||
from sqlite3 import connect
|
from sqlite3 import connect
|
||||||
from datetime import datetime, time, timedelta
|
|
||||||
from typing import Optional, Tuple
|
|
||||||
|
|
||||||
import dateutil.parser
|
|
||||||
from sanic.exceptions import Forbidden, InvalidUsage
|
|
||||||
from sanic.response import raw
|
from sanic.response import raw
|
||||||
|
|
||||||
from sqlalchemy import text
|
from sqlalchemy import select, text
|
||||||
|
from sqlalchemy.sql.expression import table, column
|
||||||
|
|
||||||
from obs.api.app import app
|
from obs.api.app import app
|
||||||
from obs.api.utils import use_request_semaphore
|
|
||||||
|
|
||||||
|
|
||||||
def get_tile(filename, zoom, x, y):
|
def get_tile(filename, zoom, x, y):
|
||||||
|
@ -28,84 +23,28 @@ def get_tile(filename, zoom, x, y):
|
||||||
|
|
||||||
content = db.execute(
|
content = db.execute(
|
||||||
"SELECT tile_data FROM tiles WHERE zoom_level=? AND tile_column=? AND tile_row=?",
|
"SELECT tile_data FROM tiles WHERE zoom_level=? AND tile_column=? AND tile_row=?",
|
||||||
(zoom, x, (2**zoom - 1) - y),
|
(zoom, x, (2 ** zoom - 1) - y),
|
||||||
).fetchone()
|
).fetchone()
|
||||||
return content and content[0] or None
|
return content and content[0] or None
|
||||||
|
|
||||||
|
|
||||||
def round_date(date, to="weeks", up=False):
|
|
||||||
if to != "weeks":
|
|
||||||
raise ValueError(f"cannot round to {to}")
|
|
||||||
|
|
||||||
midnight = time(0, 0, 0, 0)
|
|
||||||
start_of_day = date.date() # ignore time
|
|
||||||
weekday = date.weekday()
|
|
||||||
|
|
||||||
is_rounded = date.time() == midnight and weekday == 0
|
|
||||||
if is_rounded:
|
|
||||||
return date
|
|
||||||
|
|
||||||
if up:
|
|
||||||
return datetime.combine(start_of_day + timedelta(days=7 - weekday), midnight)
|
|
||||||
else:
|
|
||||||
return datetime.combine(start_of_day - timedelta(days=weekday), midnight)
|
|
||||||
|
|
||||||
|
|
||||||
# regenerate approx. once each day
|
# regenerate approx. once each day
|
||||||
TILE_CACHE_MAX_AGE = 3600 * 24
|
TILE_CACHE_MAX_AGE = 3600 * 24
|
||||||
|
|
||||||
|
|
||||||
def get_filter_options(
|
|
||||||
req,
|
|
||||||
) -> Tuple[Optional[str], Optional[datetime], Optional[datetime]]:
|
|
||||||
"""
|
|
||||||
Returns parsed, validated and normalized options for filtering map data, a
|
|
||||||
tuple of
|
|
||||||
|
|
||||||
* user_id (str|None)
|
|
||||||
* start (datetime|None)
|
|
||||||
* end (datetime|None)
|
|
||||||
"""
|
|
||||||
user_id = req.ctx.get_single_arg("user", default=None, convert=int)
|
|
||||||
if user_id is not None and (req.ctx.user is None or req.ctx.user.id != user_id):
|
|
||||||
raise Forbidden()
|
|
||||||
|
|
||||||
parse_date = lambda s: dateutil.parser.parse(s)
|
|
||||||
start = req.ctx.get_single_arg("start", default=None, convert=parse_date)
|
|
||||||
end = req.ctx.get_single_arg("end", default=None, convert=parse_date)
|
|
||||||
|
|
||||||
start = round_date(start, to="weeks", up=False) if start else None
|
|
||||||
end = round_date(end, to="weeks", up=True) if end else None
|
|
||||||
|
|
||||||
if start is not None and end is not None and start >= end:
|
|
||||||
raise InvalidUsage(
|
|
||||||
"end date must be later than start date (note: dates are rounded to weeks)"
|
|
||||||
)
|
|
||||||
|
|
||||||
return user_id, start, end
|
|
||||||
|
|
||||||
|
|
||||||
@app.route(r"/tiles/<zoom:int>/<x:int>/<y:(\d+)\.pbf>")
|
@app.route(r"/tiles/<zoom:int>/<x:int>/<y:(\d+)\.pbf>")
|
||||||
async def tiles(req, zoom: int, x: int, y: str):
|
async def tiles(req, zoom: int, x: int, y: str):
|
||||||
async with use_request_semaphore(req, "tile_semaphore"):
|
if app.config.get("TILES_FILE"):
|
||||||
if app.config.get("TILES_FILE"):
|
tile = get_tile(req.app.config.TILES_FILE, int(zoom), int(x), int(y))
|
||||||
tile = get_tile(req.app.config.TILES_FILE, int(zoom), int(x), int(y))
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
user_id, start, end = get_filter_options(req)
|
tile = await req.ctx.db.scalar(
|
||||||
|
text(f"select data from getmvt(:zoom, :x, :y) as b(data, key);").bindparams(
|
||||||
tile = await req.ctx.db.scalar(
|
zoom=int(zoom),
|
||||||
text(
|
x=int(x),
|
||||||
"select data from getmvt(:zoom, :x, :y, :user_id, :min_time, :max_time) as b(data, key);"
|
y=int(y),
|
||||||
).bindparams(
|
|
||||||
zoom=int(zoom),
|
|
||||||
x=int(x),
|
|
||||||
y=int(y),
|
|
||||||
user_id=user_id,
|
|
||||||
min_time=start,
|
|
||||||
max_time=end,
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
|
)
|
||||||
|
|
||||||
gzip = "gzip" in req.headers["accept-encoding"]
|
gzip = "gzip" in req.headers["accept-encoding"]
|
||||||
|
|
||||||
|
@ -127,3 +66,4 @@ async def tiles(req, zoom: int, x: int, y: str):
|
||||||
tile = decompress(tile)
|
tile = decompress(tile)
|
||||||
|
|
||||||
return raw(tile, content_type="application/x-protobuf", headers=headers)
|
return raw(tile, content_type="application/x-protobuf", headers=headers)
|
||||||
|
|
||||||
|
|
|
@ -1,18 +1,16 @@
|
||||||
import logging
|
import logging
|
||||||
import re
|
import re
|
||||||
from datetime import date
|
|
||||||
from json import load as jsonload
|
from json import load as jsonload
|
||||||
from os.path import join, exists, isfile
|
from os.path import join, exists, isfile
|
||||||
|
|
||||||
from sanic.exceptions import InvalidUsage, NotFound, Forbidden
|
from sqlalchemy import select, func
|
||||||
from sanic.response import file_stream, empty
|
|
||||||
from slugify import slugify
|
|
||||||
from sqlalchemy import select, func, and_
|
|
||||||
from sqlalchemy.orm import joinedload
|
from sqlalchemy.orm import joinedload
|
||||||
|
|
||||||
|
from obs.api.db import Track, User, Comment, DuplicateTrackFileError
|
||||||
from obs.api.app import api, require_auth, read_api_key, json
|
from obs.api.app import api, require_auth, read_api_key, json
|
||||||
from obs.api.db import Track, Comment, DuplicateTrackFileError
|
|
||||||
from obs.api.utils import tar_of_tracks
|
from sanic.response import file_stream, empty
|
||||||
|
from sanic.exceptions import InvalidUsage, NotFound, Forbidden
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
@ -25,8 +23,8 @@ def normalize_user_agent(user_agent):
|
||||||
return m[0] if m else None
|
return m[0] if m else None
|
||||||
|
|
||||||
|
|
||||||
async def _return_tracks(req, extend_query, limit, offset, order_by=None):
|
async def _return_tracks(req, extend_query, limit, offset):
|
||||||
if limit <= 0 or limit > 1000:
|
if limit <= 0 or limit > 100:
|
||||||
raise InvalidUsage("invalid limit")
|
raise InvalidUsage("invalid limit")
|
||||||
|
|
||||||
if offset < 0:
|
if offset < 0:
|
||||||
|
@ -41,7 +39,7 @@ async def _return_tracks(req, extend_query, limit, offset, order_by=None):
|
||||||
extend_query(select(Track).options(joinedload(Track.author)))
|
extend_query(select(Track).options(joinedload(Track.author)))
|
||||||
.limit(limit)
|
.limit(limit)
|
||||||
.offset(offset)
|
.offset(offset)
|
||||||
.order_by(order_by if order_by is not None else Track.created_at)
|
.order_by(Track.created_at.desc())
|
||||||
)
|
)
|
||||||
|
|
||||||
tracks = (await req.ctx.db.execute(query)).scalars()
|
tracks = (await req.ctx.db.execute(query)).scalars()
|
||||||
|
@ -65,114 +63,29 @@ async def _return_tracks(req, extend_query, limit, offset, order_by=None):
|
||||||
async def get_tracks(req):
|
async def get_tracks(req):
|
||||||
limit = req.ctx.get_single_arg("limit", default=20, convert=int)
|
limit = req.ctx.get_single_arg("limit", default=20, convert=int)
|
||||||
offset = req.ctx.get_single_arg("offset", default=0, convert=int)
|
offset = req.ctx.get_single_arg("offset", default=0, convert=int)
|
||||||
# author = req.ctx.get_single_arg("author", default=None, convert=int)
|
author = req.ctx.get_single_arg("author", default=None, convert=str)
|
||||||
|
|
||||||
def extend_query(q):
|
def extend_query(q):
|
||||||
q = q.where(Track.public)
|
q = q.where(Track.public)
|
||||||
|
|
||||||
# if author is not None:
|
if author is not None:
|
||||||
# q = q.where(Track.author_id == author)
|
q = q.where(User.username == author)
|
||||||
|
|
||||||
return q
|
return q
|
||||||
|
|
||||||
return await _return_tracks(req, extend_query, limit, offset)
|
return await _return_tracks(req, extend_query, limit, offset)
|
||||||
|
|
||||||
|
|
||||||
def parse_boolean(s):
|
|
||||||
if s is None:
|
|
||||||
return None
|
|
||||||
|
|
||||||
s = s.lower()
|
|
||||||
if s in ("true", "1", "yes", "y", "t"):
|
|
||||||
return True
|
|
||||||
if s in ("false", "0", "no", "n", "f"):
|
|
||||||
return False
|
|
||||||
|
|
||||||
raise ValueError("invalid value for boolean")
|
|
||||||
|
|
||||||
|
|
||||||
@api.get("/tracks/feed")
|
@api.get("/tracks/feed")
|
||||||
@require_auth
|
@require_auth
|
||||||
async def get_feed(req):
|
async def get_feed(req):
|
||||||
limit = req.ctx.get_single_arg("limit", default=20, convert=int)
|
limit = req.ctx.get_single_arg("limit", default=20, convert=int)
|
||||||
offset = req.ctx.get_single_arg("offset", default=0, convert=int)
|
offset = req.ctx.get_single_arg("offset", default=0, convert=int)
|
||||||
user_device_id = req.ctx.get_single_arg("user_device_id", default=None, convert=int)
|
|
||||||
|
|
||||||
order_by_columns = {
|
|
||||||
"recordedAt": Track.recorded_at,
|
|
||||||
"title": Track.title,
|
|
||||||
"visibility": Track.public,
|
|
||||||
"length": Track.length,
|
|
||||||
"duration": Track.duration,
|
|
||||||
"user_device_id": Track.user_device_id,
|
|
||||||
}
|
|
||||||
order_by = req.ctx.get_single_arg(
|
|
||||||
"order_by", default=None, convert=order_by_columns.get
|
|
||||||
)
|
|
||||||
|
|
||||||
reversed_ = req.ctx.get_single_arg("reversed", convert=parse_boolean, default=False)
|
|
||||||
if reversed_:
|
|
||||||
order_by = order_by.desc()
|
|
||||||
|
|
||||||
public = req.ctx.get_single_arg("public", convert=parse_boolean, default=None)
|
|
||||||
|
|
||||||
def extend_query(q):
|
def extend_query(q):
|
||||||
q = q.where(Track.author_id == req.ctx.user.id)
|
return q.where(Track.author_id == req.ctx.user.id)
|
||||||
|
|
||||||
if user_device_id is not None:
|
return await _return_tracks(req, extend_query, limit, offset)
|
||||||
q = q.where(Track.user_device_id == user_device_id)
|
|
||||||
|
|
||||||
if public is not None:
|
|
||||||
q = q.where(Track.public == public)
|
|
||||||
|
|
||||||
return q
|
|
||||||
|
|
||||||
return await _return_tracks(req, extend_query, limit, offset, order_by)
|
|
||||||
|
|
||||||
|
|
||||||
@api.post("/tracks/bulk")
|
|
||||||
@require_auth
|
|
||||||
async def tracks_bulk_action(req):
|
|
||||||
body = req.json
|
|
||||||
action = body["action"]
|
|
||||||
track_slugs = body["tracks"]
|
|
||||||
|
|
||||||
if action not in ("delete", "makePublic", "makePrivate", "reprocess", "download"):
|
|
||||||
raise InvalidUsage("invalid action")
|
|
||||||
|
|
||||||
query = select(Track).where(
|
|
||||||
and_(Track.author_id == req.ctx.user.id, Track.slug.in_(track_slugs))
|
|
||||||
)
|
|
||||||
|
|
||||||
files = set()
|
|
||||||
|
|
||||||
for track in (await req.ctx.db.execute(query)).scalars():
|
|
||||||
if action == "delete":
|
|
||||||
await req.ctx.db.delete(track)
|
|
||||||
elif action == "makePublic":
|
|
||||||
if not track.public:
|
|
||||||
track.queue_processing()
|
|
||||||
track.public = True
|
|
||||||
elif action == "makePrivate":
|
|
||||||
if track.public:
|
|
||||||
track.queue_processing()
|
|
||||||
track.public = False
|
|
||||||
elif action == "reprocess":
|
|
||||||
track.queue_processing()
|
|
||||||
elif action == "download":
|
|
||||||
files.add(track.get_original_file_path(req.app.config))
|
|
||||||
|
|
||||||
await req.ctx.db.commit()
|
|
||||||
|
|
||||||
if action == "download":
|
|
||||||
username_slug = slugify(req.ctx.user.username, separator="-")
|
|
||||||
date_str = date.today().isoformat()
|
|
||||||
file_basename = f"tracks_{username_slug}_{date_str}"
|
|
||||||
|
|
||||||
await tar_of_tracks(req, files, file_basename)
|
|
||||||
return
|
|
||||||
|
|
||||||
return empty()
|
|
||||||
|
|
||||||
|
|
||||||
@api.post("/tracks")
|
@api.post("/tracks")
|
||||||
|
|
|
@ -1,11 +1,9 @@
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from sanic.response import json
|
from sanic.response import json
|
||||||
from sanic.exceptions import InvalidUsage, Forbidden, NotFound
|
from sanic.exceptions import InvalidUsage
|
||||||
from sqlalchemy import and_, select
|
|
||||||
|
|
||||||
from obs.api.app import api, require_auth
|
from obs.api.app import api, require_auth
|
||||||
from obs.api.db import UserDevice
|
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
@ -14,9 +12,7 @@ from obs.api import __version__ as version
|
||||||
|
|
||||||
def user_to_json(user):
|
def user_to_json(user):
|
||||||
return {
|
return {
|
||||||
"id": user.id,
|
|
||||||
"username": user.username,
|
"username": user.username,
|
||||||
"displayName": user.display_name,
|
|
||||||
"email": user.email,
|
"email": user.email,
|
||||||
"bio": user.bio,
|
"bio": user.bio,
|
||||||
"image": user.image,
|
"image": user.image,
|
||||||
|
@ -30,48 +26,6 @@ async def get_user(req):
|
||||||
return json(user_to_json(req.ctx.user) if req.ctx.user else None)
|
return json(user_to_json(req.ctx.user) if req.ctx.user else None)
|
||||||
|
|
||||||
|
|
||||||
@api.get("/user/devices")
|
|
||||||
async def get_user_devices(req):
|
|
||||||
if not req.ctx.user:
|
|
||||||
raise Forbidden()
|
|
||||||
|
|
||||||
query = (
|
|
||||||
select(UserDevice)
|
|
||||||
.where(UserDevice.user_id == req.ctx.user.id)
|
|
||||||
.order_by(UserDevice.id)
|
|
||||||
)
|
|
||||||
|
|
||||||
devices = (await req.ctx.db.execute(query)).scalars()
|
|
||||||
|
|
||||||
return json([device.to_dict(req.ctx.user.id) for device in devices])
|
|
||||||
|
|
||||||
|
|
||||||
@api.put("/user/devices/<device_id:int>")
|
|
||||||
async def put_user_device(req, device_id):
|
|
||||||
if not req.ctx.user:
|
|
||||||
raise Forbidden()
|
|
||||||
|
|
||||||
body = req.json
|
|
||||||
|
|
||||||
query = (
|
|
||||||
select(UserDevice)
|
|
||||||
.where(and_(UserDevice.user_id == req.ctx.user.id, UserDevice.id == device_id))
|
|
||||||
.limit(1)
|
|
||||||
)
|
|
||||||
|
|
||||||
device = (await req.ctx.db.execute(query)).scalar()
|
|
||||||
|
|
||||||
if device is None:
|
|
||||||
raise NotFound()
|
|
||||||
|
|
||||||
new_name = body.get("displayName", "").strip()
|
|
||||||
if new_name and device.display_name != new_name:
|
|
||||||
device.display_name = new_name
|
|
||||||
await req.ctx.db.commit()
|
|
||||||
|
|
||||||
return json(device.to_dict())
|
|
||||||
|
|
||||||
|
|
||||||
@api.put("/user")
|
@api.put("/user")
|
||||||
@require_auth
|
@require_auth
|
||||||
async def put_user(req):
|
async def put_user(req):
|
||||||
|
@ -82,9 +36,6 @@ async def put_user(req):
|
||||||
if key in data and isinstance(data[key], (str, type(None))):
|
if key in data and isinstance(data[key], (str, type(None))):
|
||||||
setattr(user, key, data[key])
|
setattr(user, key, data[key])
|
||||||
|
|
||||||
if "displayName" in data:
|
|
||||||
user.display_name = data["displayName"] or None
|
|
||||||
|
|
||||||
if "areTracksVisibleForAll" in data:
|
if "areTracksVisibleForAll" in data:
|
||||||
user.are_tracks_visible_for_all = bool(data["areTracksVisibleForAll"])
|
user.are_tracks_visible_for_all = bool(data["areTracksVisibleForAll"])
|
||||||
|
|
||||||
|
|
|
@ -1,15 +1,6 @@
|
||||||
import asyncio
|
|
||||||
from contextlib import asynccontextmanager
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
import logging
|
|
||||||
from os.path import commonpath, join, relpath
|
|
||||||
import queue
|
|
||||||
import tarfile
|
|
||||||
|
|
||||||
import dateutil.parser
|
import dateutil.parser
|
||||||
from sanic.exceptions import InvalidUsage, ServiceUnavailable
|
from sanic.exceptions import InvalidUsage
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
RAISE = object()
|
RAISE = object()
|
||||||
|
|
||||||
|
@ -39,124 +30,3 @@ def round_to(value: float, multiples: float) -> float:
|
||||||
if value is None:
|
if value is None:
|
||||||
return None
|
return None
|
||||||
return round(value / multiples) * multiples
|
return round(value / multiples) * multiples
|
||||||
|
|
||||||
|
|
||||||
def chunk_list(lst, n):
|
|
||||||
for s in range(0, len(lst), n):
|
|
||||||
yield lst[s : s + n]
|
|
||||||
|
|
||||||
|
|
||||||
class chunk:
|
|
||||||
def __init__(self, iterable, n):
|
|
||||||
self.iterable = iterable
|
|
||||||
self.n = n
|
|
||||||
|
|
||||||
def __iter__(self):
|
|
||||||
if isinstance(self.iterable, list):
|
|
||||||
yield from chunk_list(self.iterable, self.n)
|
|
||||||
return
|
|
||||||
|
|
||||||
it = iter(self.iterable)
|
|
||||||
while True:
|
|
||||||
current = []
|
|
||||||
try:
|
|
||||||
for _ in range(self.n):
|
|
||||||
current.append(next(it))
|
|
||||||
yield current
|
|
||||||
except StopIteration:
|
|
||||||
if current:
|
|
||||||
yield current
|
|
||||||
break
|
|
||||||
|
|
||||||
async def __aiter__(self):
|
|
||||||
if hasattr(self.iterable, "__iter__"):
|
|
||||||
for item in self:
|
|
||||||
yield item
|
|
||||||
return
|
|
||||||
|
|
||||||
it = self.iterable.__aiter__()
|
|
||||||
while True:
|
|
||||||
current = []
|
|
||||||
try:
|
|
||||||
for _ in range(self.n):
|
|
||||||
current.append(await it.__anext__())
|
|
||||||
yield current
|
|
||||||
except StopAsyncIteration:
|
|
||||||
if len(current):
|
|
||||||
yield current
|
|
||||||
break
|
|
||||||
|
|
||||||
|
|
||||||
async def tar_of_tracks(req, files, file_basename="tracks"):
|
|
||||||
response = await req.respond(
|
|
||||||
content_type="application/x-gtar",
|
|
||||||
headers={
|
|
||||||
"content-disposition": f'attachment; filename="{file_basename}.tar.bz2"'
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
helper = StreamerHelper(response)
|
|
||||||
|
|
||||||
tar = tarfile.open(name=None, fileobj=helper, mode="w|bz2", bufsize=256 * 512)
|
|
||||||
|
|
||||||
root = commonpath(list(files))
|
|
||||||
for fname in files:
|
|
||||||
log.info("Write file to tar: %s", fname)
|
|
||||||
with open(fname, "rb") as fobj:
|
|
||||||
tarinfo = tar.gettarinfo(fname)
|
|
||||||
tarinfo.name = join(file_basename, relpath(fname, root))
|
|
||||||
tar.addfile(tarinfo, fobj)
|
|
||||||
await helper.send_all()
|
|
||||||
tar.close()
|
|
||||||
await helper.send_all()
|
|
||||||
|
|
||||||
await response.eof()
|
|
||||||
|
|
||||||
|
|
||||||
class StreamerHelper:
|
|
||||||
def __init__(self, response):
|
|
||||||
self.response = response
|
|
||||||
self.towrite = queue.Queue()
|
|
||||||
|
|
||||||
def write(self, data):
|
|
||||||
self.towrite.put(data)
|
|
||||||
|
|
||||||
async def send_all(self):
|
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
tosend = self.towrite.get(block=False)
|
|
||||||
await self.response.send(tosend)
|
|
||||||
except queue.Empty:
|
|
||||||
break
|
|
||||||
|
|
||||||
|
|
||||||
@asynccontextmanager
|
|
||||||
async def use_request_semaphore(req, semaphore_name, timeout=10):
|
|
||||||
"""
|
|
||||||
If configured, acquire a semaphore for the map tile request and release it
|
|
||||||
after the context has finished.
|
|
||||||
|
|
||||||
If the semaphore cannot be acquired within the timeout, issue a 503 Service
|
|
||||||
Unavailable error response that describes that the database is overloaded,
|
|
||||||
so users know what the problem is.
|
|
||||||
|
|
||||||
Operates as a noop when the tile semaphore is not enabled.
|
|
||||||
"""
|
|
||||||
semaphore = getattr(req.app.ctx, semaphore_name, None)
|
|
||||||
|
|
||||||
if semaphore is None:
|
|
||||||
yield
|
|
||||||
return
|
|
||||||
|
|
||||||
try:
|
|
||||||
await asyncio.wait_for(semaphore.acquire(), timeout)
|
|
||||||
|
|
||||||
try:
|
|
||||||
yield
|
|
||||||
finally:
|
|
||||||
semaphore.release()
|
|
||||||
|
|
||||||
except asyncio.TimeoutError:
|
|
||||||
raise ServiceUnavailable(
|
|
||||||
"Too many requests, database overloaded. Please retry later."
|
|
||||||
)
|
|
||||||
|
|
|
@ -58,7 +58,7 @@ def main():
|
||||||
port=app.config.PORT,
|
port=app.config.PORT,
|
||||||
debug=debug,
|
debug=debug,
|
||||||
auto_reload=app.config.get("AUTO_RELOAD", debug),
|
auto_reload=app.config.get("AUTO_RELOAD", debug),
|
||||||
access_log=True,
|
# access_log=False,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,191 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
|
|
||||||
import sys
|
|
||||||
import re
|
|
||||||
import msgpack
|
|
||||||
|
|
||||||
import osmium
|
|
||||||
import shapely.wkb as wkb
|
|
||||||
from shapely.ops import transform
|
|
||||||
|
|
||||||
HIGHWAY_TYPES = {
|
|
||||||
"trunk",
|
|
||||||
"primary",
|
|
||||||
"secondary",
|
|
||||||
"tertiary",
|
|
||||||
"unclassified",
|
|
||||||
"residential",
|
|
||||||
"trunk_link",
|
|
||||||
"primary_link",
|
|
||||||
"secondary_link",
|
|
||||||
"tertiary_link",
|
|
||||||
"living_street",
|
|
||||||
"service",
|
|
||||||
"track",
|
|
||||||
"road",
|
|
||||||
}
|
|
||||||
ZONE_TYPES = {
|
|
||||||
"urban",
|
|
||||||
"rural",
|
|
||||||
"motorway",
|
|
||||||
}
|
|
||||||
URBAN_TYPES = {
|
|
||||||
"residential",
|
|
||||||
"living_street",
|
|
||||||
"road",
|
|
||||||
}
|
|
||||||
MOTORWAY_TYPES = {
|
|
||||||
"motorway",
|
|
||||||
"motorway_link",
|
|
||||||
}
|
|
||||||
|
|
||||||
ADMIN_LEVEL_MIN = 2
|
|
||||||
ADMIN_LEVEL_MAX = 8
|
|
||||||
MINSPEED_RURAL = 60
|
|
||||||
|
|
||||||
ONEWAY_YES = {"yes", "true", "1"}
|
|
||||||
ONEWAY_REVERSE = {"reverse", "-1"}
|
|
||||||
|
|
||||||
|
|
||||||
def parse_number(tag):
|
|
||||||
if not tag:
|
|
||||||
return None
|
|
||||||
|
|
||||||
match = re.search(r"[0-9]+", tag)
|
|
||||||
if not match:
|
|
||||||
return None
|
|
||||||
|
|
||||||
digits = match.group(0)
|
|
||||||
try:
|
|
||||||
return int(digits)
|
|
||||||
except ValueError:
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
def determine_zone(tags):
|
|
||||||
highway = tags.get("highway")
|
|
||||||
zone = tags.get("zone:traffic")
|
|
||||||
|
|
||||||
if zone is not None:
|
|
||||||
if "rural" in zone:
|
|
||||||
return "rural"
|
|
||||||
|
|
||||||
if "motorway" in zone:
|
|
||||||
return "motorway"
|
|
||||||
|
|
||||||
return "urban"
|
|
||||||
|
|
||||||
# From here on we are guessing based on other tags
|
|
||||||
|
|
||||||
if highway in URBAN_TYPES:
|
|
||||||
return "urban"
|
|
||||||
|
|
||||||
if highway in MOTORWAY_TYPES:
|
|
||||||
return "motorway"
|
|
||||||
|
|
||||||
maxspeed_source = tags.get("source:maxspeed")
|
|
||||||
if maxspeed_source and "rural" in maxspeed_source:
|
|
||||||
return "rural"
|
|
||||||
if maxspeed_source and "urban" in maxspeed_source:
|
|
||||||
return "urban"
|
|
||||||
|
|
||||||
for key in ["maxspeed", "maxspeed:forward", "maxspeed:backward"]:
|
|
||||||
maxspeed = parse_number(tags.get(key))
|
|
||||||
if maxspeed is not None and maxspeed > MINSPEED_RURAL:
|
|
||||||
return "rural"
|
|
||||||
|
|
||||||
# default to urban if we have no idea
|
|
||||||
return "urban"
|
|
||||||
|
|
||||||
|
|
||||||
def determine_direction(tags, zone):
|
|
||||||
if (
|
|
||||||
tags.get("oneway") in ONEWAY_YES
|
|
||||||
or tags.get("junction") == "roundabout"
|
|
||||||
or zone == "motorway"
|
|
||||||
):
|
|
||||||
return 1, True
|
|
||||||
|
|
||||||
if tags.get("oneway") in ONEWAY_REVERSE:
|
|
||||||
return -1, True
|
|
||||||
|
|
||||||
return 0, False
|
|
||||||
|
|
||||||
|
|
||||||
class StreamPacker:
|
|
||||||
def __init__(self, stream, *args, **kwargs):
|
|
||||||
self.stream = stream
|
|
||||||
self.packer = msgpack.Packer(*args, autoreset=False, **kwargs)
|
|
||||||
|
|
||||||
def _write_out(self):
|
|
||||||
if hasattr(self.packer, "getbuffer"):
|
|
||||||
chunk = self.packer.getbuffer()
|
|
||||||
else:
|
|
||||||
chunk = self.packer.bytes()
|
|
||||||
|
|
||||||
self.stream.write(chunk)
|
|
||||||
self.packer.reset()
|
|
||||||
|
|
||||||
def pack(self, *args, **kwargs):
|
|
||||||
self.packer.pack(*args, **kwargs)
|
|
||||||
self._write_out()
|
|
||||||
|
|
||||||
def pack_array_header(self, *args, **kwargs):
|
|
||||||
self.packer.pack_array_header(*args, **kwargs)
|
|
||||||
self._write_out()
|
|
||||||
|
|
||||||
def pack_map_header(self, *args, **kwargs):
|
|
||||||
self.packer.pack_map_header(*args, **kwargs)
|
|
||||||
self._write_out()
|
|
||||||
|
|
||||||
def pack_map_pairs(self, *args, **kwargs):
|
|
||||||
self.packer.pack_map_pairs(*args, **kwargs)
|
|
||||||
self._write_out()
|
|
||||||
|
|
||||||
|
|
||||||
# A global factory that creates WKB from a osmium geometry
|
|
||||||
wkbfab = osmium.geom.WKBFactory()
|
|
||||||
|
|
||||||
from pyproj import Transformer
|
|
||||||
|
|
||||||
project = Transformer.from_crs("EPSG:4326", "EPSG:3857", always_xy=True).transform
|
|
||||||
|
|
||||||
|
|
||||||
class OSMHandler(osmium.SimpleHandler):
|
|
||||||
def __init__(self, packer):
|
|
||||||
self.packer = packer
|
|
||||||
super().__init__()
|
|
||||||
|
|
||||||
def way(self, way):
|
|
||||||
tags = way.tags
|
|
||||||
|
|
||||||
highway = tags.get("highway")
|
|
||||||
if not highway or highway not in HIGHWAY_TYPES:
|
|
||||||
return
|
|
||||||
|
|
||||||
access = tags.get("access", None)
|
|
||||||
bicycle = tags.get("bicycle", None)
|
|
||||||
if access == "no" and bicycle not in ["designated", "yes", "permissive", "destination"]:
|
|
||||||
return
|
|
||||||
|
|
||||||
zone = determine_zone(tags)
|
|
||||||
directionality, oneway = determine_direction(tags, zone)
|
|
||||||
name = tags.get("name")
|
|
||||||
|
|
||||||
geometry = wkb.loads(wkbfab.create_linestring(way), hex=True)
|
|
||||||
geometry = transform(project, geometry)
|
|
||||||
geometry = wkb.dumps(geometry)
|
|
||||||
self.packer.pack(
|
|
||||||
[b"\x01", way.id, name, zone, directionality, oneway, geometry]
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
|
||||||
with open(sys.argv[2], "wb") as fout:
|
|
||||||
packer = StreamPacker(fout)
|
|
||||||
osmhandler = OSMHandler(packer)
|
|
||||||
osmhandler.apply_file(sys.argv[1], locations=True)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
main()
|
|
|
@ -1,22 +1,14 @@
|
||||||
coloredlogs~=15.0.1
|
coloredlogs~=15.0.1
|
||||||
sanic==22.6.2
|
sanic~=21.9.3
|
||||||
oic~=1.5.0
|
oic~=1.3.0
|
||||||
sanic-session~=0.8.0
|
sanic-session~=0.8.0
|
||||||
python-slugify~=6.1.2
|
sanic-cors~=2.0.1
|
||||||
motor~=3.1.1
|
python-slugify~=5.0.2
|
||||||
pyyaml~=5.3.1
|
motor~=2.5.1
|
||||||
|
pyyaml<6
|
||||||
-e git+https://github.com/openmaptiles/openmaptiles-tools#egg=openmaptiles-tools
|
-e git+https://github.com/openmaptiles/openmaptiles-tools#egg=openmaptiles-tools
|
||||||
sqlparse~=0.4.3
|
sqlparse~=0.4.2
|
||||||
sqlalchemy[asyncio]~=1.4.46
|
sqlalchemy[asyncio]~=1.4.32
|
||||||
asyncpg~=0.27.0
|
asyncpg~=0.24.0
|
||||||
pyshp~=2.3.1
|
pyshp~=2.2.0
|
||||||
alembic~=1.9.4
|
alembic~=1.7.7
|
||||||
stream-zip~=0.0.50
|
|
||||||
msgpack~=1.0.5
|
|
||||||
osmium~=3.6.0
|
|
||||||
psycopg~=3.1.8
|
|
||||||
shapely~=2.0.1
|
|
||||||
pyproj~=3.4.1
|
|
||||||
aiohttp~=3.8.1
|
|
||||||
# sanic requires websocets and chockes on >=10 in 2022.6.2
|
|
||||||
websockets<11
|
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
Subproject commit 664e4d606416417c0651ea1748d32dd36209be6a
|
Subproject commit 8e9395fd3cd0f1e83b4413546bc2d3cb0c726738
|
19
api/setup.py
19
api/setup.py
|
@ -11,24 +11,23 @@ setup(
|
||||||
package_data={},
|
package_data={},
|
||||||
install_requires=[
|
install_requires=[
|
||||||
"coloredlogs~=15.0.1",
|
"coloredlogs~=15.0.1",
|
||||||
"sanic==22.6.2",
|
"sanic~=21.9.3",
|
||||||
"oic>=1.3.0, <2",
|
"oic>=1.3.0, <2",
|
||||||
"sanic-session~=0.8.0",
|
"sanic-session~=0.8.0",
|
||||||
"python-slugify>=5.0.2,<6.2.0",
|
"sanic-cors~=2.0.1",
|
||||||
"motor>=2.5.1,<3.1.2",
|
"python-slugify~=5.0.2",
|
||||||
|
"motor~=2.5.1",
|
||||||
"pyyaml<6",
|
"pyyaml<6",
|
||||||
"sqlparse~=0.4.3",
|
"sqlparse~=0.4.2",
|
||||||
"openmaptiles-tools", # install from git
|
"openmaptiles-tools", # install from git
|
||||||
"pyshp>=2.2,<2.4",
|
"pyshp~=2.2.0",
|
||||||
"sqlalchemy[asyncio]~=1.4.46",
|
"sqlalchemy[asyncio]~=1.4.25",
|
||||||
"asyncpg~=0.27.0",
|
"asyncpg~=0.24.0",
|
||||||
"alembic~=1.9.4",
|
"alembic~=1.7.7",
|
||||||
"stream-zip~=0.0.50",
|
|
||||||
],
|
],
|
||||||
entry_points={
|
entry_points={
|
||||||
"console_scripts": [
|
"console_scripts": [
|
||||||
"openbikesensor-api=obs.bin.openbikesensor_api:main",
|
"openbikesensor-api=obs.bin.openbikesensor_api:main",
|
||||||
"openbikesensor-transform-osm=obs.bin.openbikesensor_transform_osm:main",
|
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
|
|
@ -1,108 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
|
|
||||||
from dataclasses import dataclass
|
|
||||||
import asyncio
|
|
||||||
from os.path import basename, splitext
|
|
||||||
import sys
|
|
||||||
import logging
|
|
||||||
|
|
||||||
import msgpack
|
|
||||||
import psycopg
|
|
||||||
|
|
||||||
from obs.api.app import app
|
|
||||||
from obs.api.utils import chunk
|
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
ROAD_BUFFER = 1000
|
|
||||||
AREA_BUFFER = 100
|
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
|
||||||
class Road:
|
|
||||||
way_id: int
|
|
||||||
name: str
|
|
||||||
zone: str
|
|
||||||
directionality: int
|
|
||||||
oneway: int
|
|
||||||
geometry: bytes
|
|
||||||
|
|
||||||
|
|
||||||
def read_file(filename):
|
|
||||||
"""
|
|
||||||
Reads a file iteratively, yielding
|
|
||||||
appear. Those may be mixed.
|
|
||||||
"""
|
|
||||||
|
|
||||||
with open(filename, "rb") as f:
|
|
||||||
unpacker = msgpack.Unpacker(f)
|
|
||||||
try:
|
|
||||||
while True:
|
|
||||||
type_id, *data = unpacker.unpack()
|
|
||||||
|
|
||||||
if type_id == b"\x01":
|
|
||||||
yield Road(*data)
|
|
||||||
|
|
||||||
except msgpack.OutOfData:
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
async def import_osm(connection, filename, import_group=None):
|
|
||||||
if import_group is None:
|
|
||||||
import_group = splitext(basename(filename))[0]
|
|
||||||
|
|
||||||
# Pass 1: Find IDs only
|
|
||||||
road_ids = []
|
|
||||||
for item in read_file(filename):
|
|
||||||
road_ids.append(item.way_id)
|
|
||||||
|
|
||||||
async with connection.cursor() as cursor:
|
|
||||||
log.info("Pass 1: Delete previously imported data")
|
|
||||||
|
|
||||||
log.debug("Delete import group %s", import_group)
|
|
||||||
await cursor.execute(
|
|
||||||
"DELETE FROM road WHERE import_group = %s", (import_group,)
|
|
||||||
)
|
|
||||||
|
|
||||||
log.debug("Delete roads by way_id")
|
|
||||||
for ids in chunk(road_ids, 10000):
|
|
||||||
await cursor.execute("DELETE FROM road WHERE way_id = ANY(%s)", (ids,))
|
|
||||||
|
|
||||||
# Pass 2: Import
|
|
||||||
log.info("Pass 2: Import roads")
|
|
||||||
amount = 0
|
|
||||||
for items in chunk(read_file(filename), 10000):
|
|
||||||
amount += 10000
|
|
||||||
log.info(f"...{amount}/{len(road_ids)} ({100*amount/len(road_ids)}%)")
|
|
||||||
async with cursor.copy(
|
|
||||||
"COPY road (way_id, name, zone, directionality, oneway, geometry, import_group) FROM STDIN"
|
|
||||||
) as copy:
|
|
||||||
for item in items:
|
|
||||||
await copy.write_row(
|
|
||||||
(
|
|
||||||
item.way_id,
|
|
||||||
item.name,
|
|
||||||
item.zone,
|
|
||||||
item.directionality,
|
|
||||||
item.oneway,
|
|
||||||
bytes.hex(item.geometry),
|
|
||||||
import_group,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
async def main():
|
|
||||||
logging.basicConfig(level=logging.DEBUG, format="%(levelname)s: %(message)s")
|
|
||||||
|
|
||||||
url = app.config.POSTGRES_URL
|
|
||||||
url = url.replace("+asyncpg", "")
|
|
||||||
|
|
||||||
async with await psycopg.AsyncConnection.connect(url) as connection:
|
|
||||||
for filename in sys.argv[1:]:
|
|
||||||
log.debug("Loading file: %s", filename)
|
|
||||||
await import_osm(connection, filename)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
asyncio.run(main())
|
|
|
@ -1,93 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
|
|
||||||
"""
|
|
||||||
This script downloads and/or imports regions for statistical analysis into the
|
|
||||||
PostGIS database. The regions are sourced from:
|
|
||||||
|
|
||||||
* EU countries are covered by
|
|
||||||
[NUTS](https://ec.europa.eu/eurostat/web/gisco/geodata/reference-data/administrative-units-statistical-units/nuts).
|
|
||||||
"""
|
|
||||||
|
|
||||||
import tempfile
|
|
||||||
from dataclasses import dataclass
|
|
||||||
import json
|
|
||||||
import asyncio
|
|
||||||
from os.path import basename, splitext
|
|
||||||
import sys
|
|
||||||
import logging
|
|
||||||
from typing import Optional
|
|
||||||
|
|
||||||
import aiohttp
|
|
||||||
import psycopg
|
|
||||||
|
|
||||||
from obs.api.app import app
|
|
||||||
from obs.api.utils import chunk
|
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
NUTS_URL = "https://gisco-services.ec.europa.eu/distribution/v2/nuts/geojson/NUTS_RG_01M_2021_3857.geojson"
|
|
||||||
|
|
||||||
from pyproj import Transformer
|
|
||||||
|
|
||||||
project = Transformer.from_crs("EPSG:4326", "EPSG:3857", always_xy=True).transform
|
|
||||||
from shapely.ops import transform
|
|
||||||
from shapely.geometry import shape
|
|
||||||
import shapely.wkb as wkb
|
|
||||||
|
|
||||||
|
|
||||||
async def import_nuts(
|
|
||||||
connection, filename=None, level: int = 3, import_group: Optional[str] = None
|
|
||||||
):
|
|
||||||
if import_group is None:
|
|
||||||
import_group = f"nuts{level}"
|
|
||||||
|
|
||||||
if filename:
|
|
||||||
log.info("Load NUTS from file")
|
|
||||||
with open(filename) as f:
|
|
||||||
data = json.load(f)
|
|
||||||
else:
|
|
||||||
log.info("Download NUTS regions from europa.eu")
|
|
||||||
async with aiohttp.ClientSession() as session:
|
|
||||||
async with session.get(NUTS_URL) as resp:
|
|
||||||
data = await resp.json(content_type=None)
|
|
||||||
|
|
||||||
async with connection.cursor() as cursor:
|
|
||||||
log.info(
|
|
||||||
"Delete previously imported regions with import group %s", import_group
|
|
||||||
)
|
|
||||||
await cursor.execute(
|
|
||||||
"DELETE FROM region WHERE import_group = %s", (import_group,)
|
|
||||||
)
|
|
||||||
|
|
||||||
log.info("Import regions")
|
|
||||||
async with cursor.copy(
|
|
||||||
"COPY region (id, name, geometry, import_group) FROM STDIN"
|
|
||||||
) as copy:
|
|
||||||
for feature in data["features"]:
|
|
||||||
if feature["properties"]["LEVL_CODE"] == level:
|
|
||||||
geometry = shape(feature["geometry"])
|
|
||||||
# geometry = transform(project, geometry)
|
|
||||||
geometry = wkb.dumps(geometry)
|
|
||||||
geometry = bytes.hex(geometry)
|
|
||||||
await copy.write_row(
|
|
||||||
(
|
|
||||||
feature["properties"]["NUTS_ID"],
|
|
||||||
feature["properties"]["NUTS_NAME"],
|
|
||||||
geometry,
|
|
||||||
import_group,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
async def main():
|
|
||||||
logging.basicConfig(level=logging.DEBUG, format="%(levelname)s: %(message)s")
|
|
||||||
|
|
||||||
url = app.config.POSTGRES_URL
|
|
||||||
url = url.replace("+asyncpg", "")
|
|
||||||
|
|
||||||
async with await psycopg.AsyncConnection.connect(url) as connection:
|
|
||||||
await import_nuts(connection, sys.argv[1] if len(sys.argv) > 1 else None)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
asyncio.run(main())
|
|
|
@ -6,12 +6,9 @@ import re
|
||||||
import os
|
import os
|
||||||
import glob
|
import glob
|
||||||
from os.path import normpath, abspath, join
|
from os.path import normpath, abspath, join
|
||||||
from typing import List, Tuple
|
|
||||||
|
|
||||||
|
|
||||||
from sqlalchemy import text
|
from sqlalchemy import text
|
||||||
import sqlparse
|
import sqlparse
|
||||||
from openmaptiles.sqltomvt import MvtGenerator
|
|
||||||
|
|
||||||
from obs.api.app import app
|
from obs.api.app import app
|
||||||
from obs.api.db import connect_db, make_session
|
from obs.api.db import connect_db, make_session
|
||||||
|
@ -24,32 +21,6 @@ TILE_GENERATOR = normpath(
|
||||||
)
|
)
|
||||||
TILESET_FILE = join(TILE_GENERATOR, "openbikesensor.yaml")
|
TILESET_FILE = join(TILE_GENERATOR, "openbikesensor.yaml")
|
||||||
|
|
||||||
EXTRA_ARGS = [
|
|
||||||
# name, type, default
|
|
||||||
("user_id", "integer", "NULL"),
|
|
||||||
("min_time", "timestamp", "NULL"),
|
|
||||||
("max_time", "timestamp", "NULL"),
|
|
||||||
]
|
|
||||||
|
|
||||||
|
|
||||||
class CustomMvtGenerator(MvtGenerator):
|
|
||||||
def generate_sqltomvt_func(self, fname, extra_args: List[Tuple[str, str]]) -> str:
|
|
||||||
"""
|
|
||||||
Creates a SQL function that returns a single bytea value or null. This
|
|
||||||
method is overridden to allow for custom arguments in the created function
|
|
||||||
"""
|
|
||||||
extra_args_types = "".join([f", {a[1]}" for a in extra_args])
|
|
||||||
extra_args_definitions = "".join(
|
|
||||||
[f", {a[0]} {a[1]} DEFAULT {a[2]}" for a in extra_args]
|
|
||||||
)
|
|
||||||
|
|
||||||
return f"""\
|
|
||||||
DROP FUNCTION IF EXISTS {fname}(integer, integer, integer{extra_args_types});
|
|
||||||
CREATE FUNCTION {fname}(zoom integer, x integer, y integer{extra_args_definitions})
|
|
||||||
RETURNS {'TABLE(mvt bytea, key text)' if self.key_column else 'bytea'} AS $$
|
|
||||||
{self.generate_sql()};
|
|
||||||
$$ LANGUAGE SQL STABLE CALLED ON NULL INPUT;"""
|
|
||||||
|
|
||||||
|
|
||||||
def parse_pg_url(url=app.config.POSTGRES_URL):
|
def parse_pg_url(url=app.config.POSTGRES_URL):
|
||||||
m = re.match(
|
m = re.match(
|
||||||
|
@ -143,20 +114,9 @@ async def generate_sql(build_dir):
|
||||||
with open(filename, "rt") as f:
|
with open(filename, "rt") as f:
|
||||||
sql_snippets.append(f.read())
|
sql_snippets.append(f.read())
|
||||||
|
|
||||||
mvt = CustomMvtGenerator(
|
getmvt_sql = await _run(
|
||||||
tileset=TILESET_FILE,
|
f"python $(which generate-sqltomvt) {TILESET_FILE!r} --key --gzip --postgis-ver 3.0.1 --function --fname=getmvt"
|
||||||
postgis_ver="3.0.1",
|
|
||||||
zoom="zoom",
|
|
||||||
x="x",
|
|
||||||
y="y",
|
|
||||||
gzip=True,
|
|
||||||
test_geometry=False, # ?
|
|
||||||
key_column=True,
|
|
||||||
)
|
)
|
||||||
getmvt_sql = mvt.generate_sqltomvt_func("getmvt", EXTRA_ARGS)
|
|
||||||
|
|
||||||
# drop old versions of the function
|
|
||||||
sql_snippets.append("DROP FUNCTION IF EXISTS getmvt(integer, integer, integer);")
|
|
||||||
sql_snippets.append(getmvt_sql)
|
sql_snippets.append(getmvt_sql)
|
||||||
|
|
||||||
return sql_snippets
|
return sql_snippets
|
||||||
|
|
|
@ -1,30 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
import logging
|
|
||||||
import asyncio
|
|
||||||
|
|
||||||
from sqlalchemy import text
|
|
||||||
|
|
||||||
from obs.api.app import app
|
|
||||||
from obs.api.db import connect_db, make_session
|
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
async def main():
|
|
||||||
logging.basicConfig(level=logging.DEBUG, format="%(levelname)s: %(message)s")
|
|
||||||
await reimport_tracks()
|
|
||||||
|
|
||||||
|
|
||||||
async def reimport_tracks():
|
|
||||||
|
|
||||||
async with connect_db(
|
|
||||||
app.config.POSTGRES_URL,
|
|
||||||
app.config.POSTGRES_POOL_SIZE,
|
|
||||||
app.config.POSTGRES_MAX_OVERFLOW,
|
|
||||||
):
|
|
||||||
async with make_session() as session:
|
|
||||||
await session.execute(text("UPDATE track SET processing_status = 'queued';"))
|
|
||||||
await session.commit()
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
asyncio.run(main())
|
|
|
@ -1,6 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
|
|
||||||
from obs.bin.openbikesensor_transform_osm import main
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
main()
|
|
|
@ -1,15 +1,14 @@
|
||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
import asyncio
|
|
||||||
import logging
|
import logging
|
||||||
|
import asyncio
|
||||||
|
from alembic.config import Config
|
||||||
|
from alembic import command
|
||||||
|
from os.path import join, dirname
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
from prepare_sql_tiles import prepare_sql_tiles, _run
|
from prepare_sql_tiles import prepare_sql_tiles, _run
|
||||||
|
|
||||||
from import_regions import main as import_nuts
|
|
||||||
|
|
||||||
from reimport_tracks import main as reimport_tracks
|
|
||||||
|
|
||||||
|
|
||||||
async def _migrate():
|
async def _migrate():
|
||||||
await _run("alembic upgrade head")
|
await _run("alembic upgrade head")
|
||||||
|
@ -21,11 +20,7 @@ async def main():
|
||||||
await _migrate()
|
await _migrate()
|
||||||
log.info("Preparing SQL tiles...")
|
log.info("Preparing SQL tiles...")
|
||||||
await prepare_sql_tiles()
|
await prepare_sql_tiles()
|
||||||
log.info("Importing nuts regions...")
|
log.info("Upgraded")
|
||||||
await import_nuts()
|
|
||||||
log.info("Nuts regions imported, scheduling reimport of tracks")
|
|
||||||
await reimport_tracks()
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
|
|
@ -55,7 +55,12 @@ git clone --recursive https://github.com/openbikesensor/portal source/
|
||||||
```bash
|
```bash
|
||||||
mkdir -p /opt/openbikesensor/config
|
mkdir -p /opt/openbikesensor/config
|
||||||
cd /opt/openbikesensor/
|
cd /opt/openbikesensor/
|
||||||
cp -r source/deployment/config source/deployment/docker-compose.yaml source/deployment/.env .
|
|
||||||
|
cp source/deployment/examples/docker-compose.yaml docker-compose.yaml
|
||||||
|
cp source/deployment/examples/.env .env
|
||||||
|
|
||||||
|
cp source/deployment/examples/traefik.toml config/traefik.toml
|
||||||
|
cp source/deployment/examples/config.py config/config.py
|
||||||
```
|
```
|
||||||
|
|
||||||
### Create a Docker network
|
### Create a Docker network
|
||||||
|
@ -219,6 +224,18 @@ docker-compose build portal
|
||||||
|
|
||||||
*Hint*: This may take up to 10 minutes. In the future, we will provide a prebuild image.
|
*Hint*: This may take up to 10 minutes. In the future, we will provide a prebuild image.
|
||||||
|
|
||||||
|
#### Download OpenStreetMap maps
|
||||||
|
|
||||||
|
Download the area(s) you would like to import from
|
||||||
|
[GeoFabrik](https://download.geofabrik.de) into `data/pbf`, for example:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
cd /opt/openbikesensor/
|
||||||
|
wget https://download.geofabrik.de/europe/germany/schleswig-holstein-latest.osm.pbf -P data/pbf
|
||||||
|
```
|
||||||
|
|
||||||
|
*Hint*: Start with a small region/city, since the import can take some hours for huge areas.
|
||||||
|
|
||||||
#### Prepare database
|
#### Prepare database
|
||||||
|
|
||||||
Run the following scripts to prepare the database:
|
Run the following scripts to prepare the database:
|
||||||
|
@ -231,7 +248,13 @@ For more details, see [README.md](../README.md) under "Prepare database".
|
||||||
|
|
||||||
#### Import OpenStreetMap data
|
#### Import OpenStreetMap data
|
||||||
|
|
||||||
Follow [these instructions](./osm-import.md).
|
Run the following script, to import the OSM data:
|
||||||
|
|
||||||
|
```
|
||||||
|
docker-compose run --rm portal tools/osm2pgsql.sh
|
||||||
|
```
|
||||||
|
|
||||||
|
For more details. see [README.md](../README.md) under "Import OpenStreetMap data".
|
||||||
|
|
||||||
|
|
||||||
#### Configure portal
|
#### Configure portal
|
||||||
|
@ -287,7 +310,7 @@ This also starts a dedicated worker container to handle the tracks.
|
||||||
|
|
||||||
#### Test the portal
|
#### Test the portal
|
||||||
|
|
||||||
* Open: https://portal.example.com/ (URL depends on your setup)
|
* Open: https://obs.example.com/
|
||||||
* Login with the user
|
* Login with the user
|
||||||
* Upload a track via My Tracks
|
* Upload a track via My Tracks
|
||||||
|
|
||||||
|
@ -297,7 +320,7 @@ You should see smth. like:
|
||||||
|
|
||||||
When you click on *My Tracks*, you should see it on a map.
|
When you click on *My Tracks*, you should see it on a map.
|
||||||
|
|
||||||
#### Configure the map position
|
#### Configre the map position
|
||||||
|
|
||||||
Open the tab *Map** an zoom to the desired position. The URL contains the corresponding GPS position,
|
Open the tab *Map** an zoom to the desired position. The URL contains the corresponding GPS position,
|
||||||
for example:
|
for example:
|
||||||
|
@ -318,6 +341,10 @@ docker-compose restart portal
|
||||||
The tab *Map* should be the selected map section now.
|
The tab *Map* should be the selected map section now.
|
||||||
When you uploaded some tracks, you map should show a colors overlay on the streets.
|
When you uploaded some tracks, you map should show a colors overlay on the streets.
|
||||||
|
|
||||||
|
#### Verify osm2pgsql
|
||||||
|
|
||||||
|
If you zoom in the tab *Map* at the imported region/city, you should see dark grey lines on the streets.
|
||||||
|
|
||||||
## Miscellaneous
|
## Miscellaneous
|
||||||
|
|
||||||
### Logs
|
### Logs
|
|
@ -1,22 +0,0 @@
|
||||||
|
|
||||||
events {}
|
|
||||||
http {
|
|
||||||
proxy_cache_path /data/nginx/cache levels=1:2 keys_zone=STATIC:10m
|
|
||||||
inactive=24h max_size=1g;
|
|
||||||
server {
|
|
||||||
location ~* ^/tiles/\d[012]?/[^?]+$ {
|
|
||||||
proxy_pass http://portal:3000;
|
|
||||||
proxy_set_header Host $host:3000;
|
|
||||||
proxy_buffering on;
|
|
||||||
proxy_cache_methods GET HEAD;
|
|
||||||
proxy_cache STATIC;
|
|
||||||
proxy_cache_valid 200 1d;
|
|
||||||
proxy_cache_use_stale error timeout invalid_header updating
|
|
||||||
http_500 http_502 http_503 http_504;
|
|
||||||
}
|
|
||||||
location / {
|
|
||||||
proxy_pass http://portal:3000;
|
|
||||||
proxy_set_header Host $host:3000;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -9,8 +9,6 @@ OBS_KEYCLOAK_URI=login.example.com
|
||||||
OBS_KEYCLOAK_POSTGRES_USER=obs
|
OBS_KEYCLOAK_POSTGRES_USER=obs
|
||||||
OBS_KEYCLOAK_POSTGRES_PASSWORD=<<TODO>>
|
OBS_KEYCLOAK_POSTGRES_PASSWORD=<<TODO>>
|
||||||
OBS_KEYCLOAK_POSTGRES_DB=obs
|
OBS_KEYCLOAK_POSTGRES_DB=obs
|
||||||
OBS_POSTGRES_MAX_OVERFLOW=20
|
|
||||||
OBS_POSTGRES_POOL_SIZE=40
|
|
||||||
|
|
||||||
# KeyCloak
|
# KeyCloak
|
||||||
|
|
|
@ -1,30 +1,35 @@
|
||||||
# Bind address of the server
|
# Bind address of the server
|
||||||
# HOST = "127.0.0.1"
|
#HOST = "127.0.0.1"
|
||||||
# PORT = 3000
|
#PORT = 3000
|
||||||
|
|
||||||
# Extended log output, but slower
|
# Extended log output, but slower
|
||||||
DEBUG = False
|
DEBUG = False
|
||||||
VERBOSE = DEBUG
|
VERBOSE = DEBUG
|
||||||
AUTO_RELOAD = DEBUG
|
AUTO_RESTART = DEBUG
|
||||||
|
|
||||||
|
# Turn on lean mode to simplify the setup. Lots of features will be
|
||||||
|
# unavailable, but you will not need to manage OpenStreetMap data. Please make
|
||||||
|
# sure to configure the OBS_FACE_CACHE_DIR correctly for lean mode.
|
||||||
|
LEAN_MODE = False
|
||||||
|
|
||||||
# Required to encrypt or sign sessions, cookies, tokens, etc.
|
# Required to encrypt or sign sessions, cookies, tokens, etc.
|
||||||
# SECRET = "!!!<<<CHANGEME>>>!!!"
|
#SECRET = "!!!<<<CHANGEME>>>!!!"
|
||||||
|
|
||||||
# Connection to the database
|
# Connection to the database
|
||||||
# POSTGRES_URL = "postgresql+asyncpg://user:pass@host/dbname"
|
#POSTGRES_URL = "postgresql+asyncpg://user:pass@host/dbname"
|
||||||
# POSTGRES_POOL_SIZE = 20
|
#POSTGRES_POOL_SIZE = 20
|
||||||
# POSTGRES_MAX_OVERFLOW = 2 * POSTGRES_POOL_SIZE
|
#POSTGRES_MAX_OVERFLOW = 2 * POSTGRES_POOL_SIZE
|
||||||
|
|
||||||
# URL to the keycloak realm, as reachable by the API service. This is not
|
# URL to the keycloak realm, as reachable by the API service. This is not
|
||||||
# necessarily its publicly reachable URL, keycloak advertises that iself.
|
# necessarily its publicly reachable URL, keycloak advertises that iself.
|
||||||
# KEYCLOAK_URL = "http://localhost:1234/auth/realms/obs/"
|
#KEYCLOAK_URL = "http://localhost:1234/auth/realms/obs/"
|
||||||
|
|
||||||
# Auth client credentials
|
# Auth client credentials
|
||||||
# KEYCLOAK_CLIENT_ID = "portal"
|
#KEYCLOAK_CLIENT_ID = "portal"
|
||||||
# KEYCLOAK_CLIENT_SECRET = "00000000-0000-0000-0000-000000000000"
|
#KEYCLOAK_CLIENT_SECRET = "00000000-0000-0000-0000-000000000000"
|
||||||
|
|
||||||
# Whether the API should run the worker loop, or a dedicated worker is used
|
# Whether the API should run the worker loop, or a dedicated worker is used
|
||||||
# DEDICATED_WORKER = True
|
#DEDICATED_WORKER = True
|
||||||
|
|
||||||
# The root of the frontend. Needed for redirecting after login, and for CORS.
|
# The root of the frontend. Needed for redirecting after login, and for CORS.
|
||||||
# Set to None if frontend is served by the API.
|
# Set to None if frontend is served by the API.
|
|
@ -14,7 +14,7 @@ services:
|
||||||
############################################################
|
############################################################
|
||||||
|
|
||||||
postgres:
|
postgres:
|
||||||
image: "openmaptiles/postgis:7.0"
|
image: "openmaptiles/postgis:6.0"
|
||||||
environment:
|
environment:
|
||||||
- POSTGRES_DB=${OBS_POSTGRES_DB}
|
- POSTGRES_DB=${OBS_POSTGRES_DB}
|
||||||
- POSTGRES_USER=${OBS_POSTGRES_USER}
|
- POSTGRES_USER=${OBS_POSTGRES_USER}
|
||||||
|
@ -136,7 +136,7 @@ services:
|
||||||
- "traefik.docker.network=gateway"
|
- "traefik.docker.network=gateway"
|
||||||
|
|
||||||
postgres-keycloak:
|
postgres-keycloak:
|
||||||
image: postgres:15
|
image: postgres:13.3
|
||||||
restart: always
|
restart: always
|
||||||
networks:
|
networks:
|
||||||
- backend
|
- backend
|
|
@ -8,7 +8,7 @@ version: '3'
|
||||||
|
|
||||||
services:
|
services:
|
||||||
postgres:
|
postgres:
|
||||||
image: "openmaptiles/postgis:7.0"
|
image: "openmaptiles/postgis:6.0"
|
||||||
environment:
|
environment:
|
||||||
POSTGRES_USER: obs
|
POSTGRES_USER: obs
|
||||||
POSTGRES_PASSWORD: obs
|
POSTGRES_PASSWORD: obs
|
||||||
|
@ -20,7 +20,6 @@ services:
|
||||||
|
|
||||||
api:
|
api:
|
||||||
image: openbikesensor-api
|
image: openbikesensor-api
|
||||||
tty: true
|
|
||||||
build:
|
build:
|
||||||
context: ./api/
|
context: ./api/
|
||||||
dockerfile: Dockerfile
|
dockerfile: Dockerfile
|
||||||
|
@ -36,8 +35,6 @@ services:
|
||||||
- ./tile-generator/data/:/tiles
|
- ./tile-generator/data/:/tiles
|
||||||
- ./api/migrations:/opt/obs/api/migrations
|
- ./api/migrations:/opt/obs/api/migrations
|
||||||
- ./api/alembic.ini:/opt/obs/api/alembic.ini
|
- ./api/alembic.ini:/opt/obs/api/alembic.ini
|
||||||
- ./local/pbf:/pbf
|
|
||||||
- ./local/obsdata:/obsdata
|
|
||||||
depends_on:
|
depends_on:
|
||||||
- postgres
|
- postgres
|
||||||
- keycloak
|
- keycloak
|
||||||
|
@ -49,7 +46,6 @@ services:
|
||||||
|
|
||||||
worker:
|
worker:
|
||||||
image: openbikesensor-api
|
image: openbikesensor-api
|
||||||
tty: true
|
|
||||||
build:
|
build:
|
||||||
context: ./api/
|
context: ./api/
|
||||||
dockerfile: Dockerfile
|
dockerfile: Dockerfile
|
||||||
|
|
57
docs/lean-mode.md
Normal file
57
docs/lean-mode.md
Normal file
|
@ -0,0 +1,57 @@
|
||||||
|
# Lean mode
|
||||||
|
|
||||||
|
The application can be configured in "lean mode" through the `LEAN_MODE`
|
||||||
|
setting in `config.py`. A lean installation is easier to set up, as a few steps
|
||||||
|
can be skipped. However, the performance of the application will degrade in
|
||||||
|
lean mode, and lots of advanced features will not be available.
|
||||||
|
|
||||||
|
Lean mode is meant as an entrypoint to get started with collecting data,
|
||||||
|
without the hassle of importing and maintaining OpenStreetMap data.
|
||||||
|
|
||||||
|
## Disabled features in lean mode
|
||||||
|
|
||||||
|
* No map tiles are generated.
|
||||||
|
* The frontend will not show an overview map, only per-track maps.
|
||||||
|
* The `roads` database table is not used, neither for processing tracks, nor
|
||||||
|
for generating map tiles.
|
||||||
|
* The API will not generate auxiliary information for display on the
|
||||||
|
(nonexistent) map, such as per-road statistics.
|
||||||
|
|
||||||
|
## Switch to/from lean mode
|
||||||
|
|
||||||
|
To enable lean mode, set the following in your `config.py` (or in
|
||||||
|
`config.overrides.py`, especially in development setups):
|
||||||
|
|
||||||
|
```python
|
||||||
|
LEAN_MODE = True
|
||||||
|
```
|
||||||
|
|
||||||
|
To disable lean mode, set it to `False` instead.
|
||||||
|
|
||||||
|
For lean mode, it is important that the config variable `OBS_FACE_CACHE_DIR` is
|
||||||
|
properly set, or that you are happy with its default value of using
|
||||||
|
`$DATA_DIR/obs-face-cache`.
|
||||||
|
|
||||||
|
When turning off lean mode, make sure to fill your `roads` table properly, as
|
||||||
|
otherwise the track processing will not work. When turning on lean mode, you
|
||||||
|
may truncate the `roads` table to save space, but you don't need to, it simply
|
||||||
|
becomes unused.
|
||||||
|
|
||||||
|
## Benefits
|
||||||
|
|
||||||
|
* When using lean mode, you can skip the import of OpenStreetMap data during
|
||||||
|
setup, and you also do not need to keep it updated.
|
||||||
|
* People can already start uploading data and the data is also processed,
|
||||||
|
giving you as a maintainer more time to set up the full application, if you
|
||||||
|
want to.
|
||||||
|
|
||||||
|
## Drawbacks
|
||||||
|
|
||||||
|
* Lean mode is less performant when processing tracks.
|
||||||
|
* Lean mode track processing depends on the Overpass API data source, which may
|
||||||
|
be slow, unavailable, or rate limiting the requests, so processing may fail.
|
||||||
|
We use caching to prevent some issues, but as we depend on a third party
|
||||||
|
service here that is accessed for free and that generates a lot of server
|
||||||
|
load, we really can't ask for much. If you frequently run into issues, the
|
||||||
|
best bet is to manage OSM data yourself and turn off lean mode.
|
||||||
|
* Of course some features are missing.
|
|
@ -1,103 +0,0 @@
|
||||||
# Importing OpenStreetMap data
|
|
||||||
|
|
||||||
The application requires a lot of data from the OpenStreetMap to work.
|
|
||||||
|
|
||||||
The required information is stored in the PostgreSQL database and used when
|
|
||||||
processing tracks, as well as for vector tile generation. The process applies
|
|
||||||
to both development and production setups. For development, you should choose a
|
|
||||||
small area for testing, such as your local county or city, to keep the amount
|
|
||||||
of data small. For production use you have to import the whole region you are
|
|
||||||
serving.
|
|
||||||
|
|
||||||
## General pipeline overview
|
|
||||||
|
|
||||||
1. Download OpenStreetMap data as one or more `.osm.pbf` files.
|
|
||||||
2. Transform this data to generate geometry data for all roads and regions, so
|
|
||||||
we don't need to look up nodes separately. This step requires a lot of CPU
|
|
||||||
and memory, so it can be done "offline" on a high power machine.
|
|
||||||
3. Import the transformed data into the PostgreSQL/PostGIS database.
|
|
||||||
|
|
||||||
## Community hosted transformed data
|
|
||||||
|
|
||||||
Since the first two steps are the same for everybody, the community will soon
|
|
||||||
provide a service where relatively up-to-date transformed data can be
|
|
||||||
downloaded for direct import. Stay tuned.
|
|
||||||
|
|
||||||
## Download data
|
|
||||||
|
|
||||||
[GeoFabrik](https://download.geofabrik.de) kindly hosts extracts of the
|
|
||||||
OpenStreetMap planet by region. Download all regions you're interested in from
|
|
||||||
there in `.osm.pbf` format, with the tool of your choice, e. g.:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
wget -P local/pbf/ https://download.geofabrik.de/europe/germany/baden-wuerttemberg-latest.osm.pbf
|
|
||||||
```
|
|
||||||
|
|
||||||
## Transform data
|
|
||||||
|
|
||||||
To transform downloaded data, you can either use the docker image from a
|
|
||||||
development or production environment, or locally install the API into your
|
|
||||||
python environment. Then run the `api/tools/transform_osm.py` script on the data.
|
|
||||||
|
|
||||||
```bash
|
|
||||||
api/tools/transform_osm.py baden-wuerttemberg-latest.osm.pbf baden-wuerttemberg-latest.msgpack
|
|
||||||
```
|
|
||||||
|
|
||||||
In dockerized setups, make sure to mount your data somewhere in the container
|
|
||||||
and also mount a directory where the result can be written. The development
|
|
||||||
setup takes care of this, so you can use:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
docker-compose run --rm api tools/transform_osm.py \
|
|
||||||
/pbf/baden-wuerttemberg-latest.osm.pbf /obsdata/baden-wuerttemberg-latest.msgpack
|
|
||||||
```
|
|
||||||
|
|
||||||
Repeat this command for every file you want to transform.
|
|
||||||
|
|
||||||
## Import transformed data
|
|
||||||
|
|
||||||
The command for importing looks like this:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
api/tools/import_osm.py baden-wuerttemberg-latest.msgpack
|
|
||||||
```
|
|
||||||
|
|
||||||
This tool reads your application config from `config.py`, so set that up first
|
|
||||||
as if you were setting up your application.
|
|
||||||
|
|
||||||
In dockerized setups, make sure to mount your data somewhere in the container.
|
|
||||||
Again, the development setup takes care of this, so you can use:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
docker-compose run --rm api tools/import_osm.py \
|
|
||||||
/obsdata/baden-wuerttemberg-latest.msgpack
|
|
||||||
```
|
|
||||||
|
|
||||||
The transform process should take a few seconds to minutes, depending on the area
|
|
||||||
size. You can run the process multiple times, with the same or different area
|
|
||||||
files, to import or update the data. You can update only one region and leave
|
|
||||||
the others as they are, or add more filenames to the command line to
|
|
||||||
bulk-import data.
|
|
||||||
|
|
||||||
## How this works
|
|
||||||
|
|
||||||
* The transformation is done with a python script that uses
|
|
||||||
[pyosmium](https://osmcode.org/pyosmium/) to read the `.osm.pbf` file. This
|
|
||||||
script then filters the data for only the required objects (such as road
|
|
||||||
segments and administrative areas), and extracts the interesting information
|
|
||||||
from those objects.
|
|
||||||
* The node geolocations are looked up to generate a geometry for each object.
|
|
||||||
This requires a lot of memory to run efficiently.
|
|
||||||
* The geometry is projected to [Web Mercator](https://epsg.io/3857) in this
|
|
||||||
step to avoid continous transformation when tiles are generated later. Most
|
|
||||||
operations will work fine in this projection. Projection is done with the
|
|
||||||
[pyproj](https://pypi.org/project/pyproj/) library.
|
|
||||||
* The output is written to a binary file in a very simple format using
|
|
||||||
[msgpack](https://github.com/msgpack/msgpack-python), which is way more
|
|
||||||
efficient that (Geo-)JSON for example. This format is stremable, so the
|
|
||||||
generated file is never fully written or read into memory.
|
|
||||||
* The import script reads the msgpack file and sends it to the database using
|
|
||||||
[psycopg](https://www.psycopg.org/). This is done because it supports
|
|
||||||
PostgreSQL's `COPY FROM` statement, which enables much faster writes to the
|
|
||||||
database that a traditionional `INSERT VALUES`. The file is streamed directly
|
|
||||||
to the database, so it is never read into memory.
|
|
4713
frontend/package-lock.json
generated
4713
frontend/package-lock.json
generated
File diff suppressed because it is too large
Load diff
|
@ -12,11 +12,7 @@
|
||||||
"classnames": "^2.3.1",
|
"classnames": "^2.3.1",
|
||||||
"colormap": "^2.3.2",
|
"colormap": "^2.3.2",
|
||||||
"downloadjs": "^1.4.7",
|
"downloadjs": "^1.4.7",
|
||||||
"echarts": "^5.3.2",
|
|
||||||
"echarts-for-react": "^3.0.2",
|
|
||||||
"fomantic-ui-less": "^2.8.8",
|
"fomantic-ui-less": "^2.8.8",
|
||||||
"i18next-browser-languagedetector": "^6.1.4",
|
|
||||||
"i18next-http-backend": "^1.4.1",
|
|
||||||
"immer": "^9.0.7",
|
"immer": "^9.0.7",
|
||||||
"luxon": "^1.28.0",
|
"luxon": "^1.28.0",
|
||||||
"maplibre-gl": "^1.15.2",
|
"maplibre-gl": "^1.15.2",
|
||||||
|
@ -31,7 +27,6 @@
|
||||||
"react-dom": "^17.0.2",
|
"react-dom": "^17.0.2",
|
||||||
"react-helmet": "^6.1.0",
|
"react-helmet": "^6.1.0",
|
||||||
"react-hook-form": "^6.15.8",
|
"react-hook-form": "^6.15.8",
|
||||||
"react-i18next": "^11.18.1",
|
|
||||||
"react-map-gl": "^6.1.17",
|
"react-map-gl": "^6.1.17",
|
||||||
"react-markdown": "^5.0.3",
|
"react-markdown": "^5.0.3",
|
||||||
"react-redux": "^7.2.6",
|
"react-redux": "^7.2.6",
|
||||||
|
@ -44,13 +39,11 @@
|
||||||
"sass": "^1.43.5",
|
"sass": "^1.43.5",
|
||||||
"semantic-ui-react": "^2.0.4",
|
"semantic-ui-react": "^2.0.4",
|
||||||
"ts-loader": "^9.2.6",
|
"ts-loader": "^9.2.6",
|
||||||
"typescript": "^4.7.4",
|
"typescript": "^4.5.2"
|
||||||
"yaml-loader": "^0.8.0"
|
|
||||||
},
|
},
|
||||||
"eslintConfig": {
|
"eslintConfig": {
|
||||||
"extends": [
|
"extends": [
|
||||||
"react-app",
|
"react-app"
|
||||||
"plugin:prettier/recommended"
|
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
"browserslist": {
|
"browserslist": {
|
||||||
|
@ -80,12 +73,8 @@
|
||||||
"@types/react-router-dom": "^5.3.2",
|
"@types/react-router-dom": "^5.3.2",
|
||||||
"babel-loader": "^8.2.3",
|
"babel-loader": "^8.2.3",
|
||||||
"css-loader": "^5.2.7",
|
"css-loader": "^5.2.7",
|
||||||
"eslint-config-prettier": "^8.5.0",
|
|
||||||
"eslint-config-react-app": "^7.0.1",
|
|
||||||
"eslint-plugin-prettier": "^4.2.1",
|
|
||||||
"html-webpack-plugin": "^5.5.0",
|
"html-webpack-plugin": "^5.5.0",
|
||||||
"less-loader": "^10.2.0",
|
"less-loader": "^10.2.0",
|
||||||
"prettier": "^2.7.1",
|
|
||||||
"react-refresh": "^0.11.0",
|
"react-refresh": "^0.11.0",
|
||||||
"style-loader": "^3.3.1",
|
"style-loader": "^3.3.1",
|
||||||
"webpack": "^5.64.4",
|
"webpack": "^5.64.4",
|
||||||
|
|
|
@ -69,6 +69,7 @@
|
||||||
}
|
}
|
||||||
|
|
||||||
.pageTitle a {
|
.pageTitle a {
|
||||||
|
font-family: 'Open Sans Condensed';
|
||||||
font-weight: 600;
|
font-weight: 600;
|
||||||
font-size: 18pt;
|
font-size: 18pt;
|
||||||
|
|
||||||
|
@ -119,15 +120,6 @@
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@media @mobile {
|
|
||||||
.menu.menu {
|
|
||||||
> :global(.ui.container) {
|
|
||||||
height: @menuHeightMobile;
|
|
||||||
align-items: stretch;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
.banner {
|
.banner {
|
||||||
padding: 8px;
|
padding: 8px;
|
||||||
z-index: 100;
|
z-index: 100;
|
||||||
|
|
|
@ -6,15 +6,12 @@ import {BrowserRouter as Router, Switch, Route, Link} from 'react-router-dom'
|
||||||
import {useObservable} from 'rxjs-hooks'
|
import {useObservable} from 'rxjs-hooks'
|
||||||
import {from} from 'rxjs'
|
import {from} from 'rxjs'
|
||||||
import {pluck} from 'rxjs/operators'
|
import {pluck} from 'rxjs/operators'
|
||||||
import {Helmet} from 'react-helmet'
|
import {Helmet} from "react-helmet";
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {useConfig} from 'config'
|
import {useConfig} from 'config'
|
||||||
import styles from './App.module.less'
|
import styles from './App.module.less'
|
||||||
import {AVAILABLE_LOCALES, setLocale} from 'i18n'
|
|
||||||
|
|
||||||
import {
|
import {
|
||||||
AcknowledgementsPage,
|
|
||||||
ExportPage,
|
ExportPage,
|
||||||
HomePage,
|
HomePage,
|
||||||
LoginRedirectPage,
|
LoginRedirectPage,
|
||||||
|
@ -26,7 +23,6 @@ import {
|
||||||
TrackPage,
|
TrackPage,
|
||||||
TracksPage,
|
TracksPage,
|
||||||
UploadPage,
|
UploadPage,
|
||||||
MyTracksPage,
|
|
||||||
} from 'pages'
|
} from 'pages'
|
||||||
import {Avatar, LoginButton} from 'components'
|
import {Avatar, LoginButton} from 'components'
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
|
@ -62,7 +58,6 @@ function Banner({text, style = 'warning'}: {text: string; style: 'warning' | 'in
|
||||||
}
|
}
|
||||||
|
|
||||||
const App = connect((state) => ({login: state.login}))(function App({login}) {
|
const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const config = useConfig()
|
const config = useConfig()
|
||||||
const apiVersion = useObservable(() => from(api.get('/info')).pipe(pluck('version')))
|
const apiVersion = useObservable(() => from(api.get('/info')).pipe(pluck('version')))
|
||||||
|
|
||||||
|
@ -79,7 +74,7 @@ const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
<title>OpenBikeSensor Portal</title>
|
<title>OpenBikeSensor Portal</title>
|
||||||
</Helmet>
|
</Helmet>
|
||||||
{config?.banner && <Banner {...config.banner} />}
|
{config?.banner && <Banner {...config.banner} />}
|
||||||
<Menu className={styles.menu} stackable>
|
<Menu className={styles.menu}>
|
||||||
<Container>
|
<Container>
|
||||||
<Link to="/" component={MenuItemForLink} header className={styles.pageTitle}>
|
<Link to="/" component={MenuItemForLink} header className={styles.pageTitle}>
|
||||||
OpenBikeSensor
|
OpenBikeSensor
|
||||||
|
@ -87,35 +82,30 @@ const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
|
|
||||||
{hasMap && (
|
{hasMap && (
|
||||||
<Link component={MenuItemForLink} to="/map" as="a">
|
<Link component={MenuItemForLink} to="/map" as="a">
|
||||||
{t('App.menu.map')}
|
Map
|
||||||
</Link>
|
</Link>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
<Link component={MenuItemForLink} to="/tracks" as="a">
|
<Link component={MenuItemForLink} to="/tracks" as="a">
|
||||||
{t('App.menu.tracks')}
|
Tracks
|
||||||
</Link>
|
</Link>
|
||||||
|
|
||||||
<Link component={MenuItemForLink} to="/export" as="a">
|
<Link component={MenuItemForLink} to="/export" as="a">
|
||||||
{t('App.menu.export')}
|
Export
|
||||||
</Link>
|
</Link>
|
||||||
|
|
||||||
<Menu.Menu position="right">
|
<Menu.Menu position="right">
|
||||||
{login ? (
|
{login ? (
|
||||||
<>
|
<>
|
||||||
<Link component={MenuItemForLink} to="/my/tracks" as="a">
|
<Link component={MenuItemForLink} to="/my/tracks" as="a">
|
||||||
{t('App.menu.myTracks')}
|
My Tracks
|
||||||
</Link>
|
</Link>
|
||||||
<Dropdown item trigger={<Avatar user={login} className={styles.avatar} />}>
|
<Dropdown item trigger={<Avatar user={login} className={styles.avatar} />}>
|
||||||
<Dropdown.Menu>
|
<Dropdown.Menu>
|
||||||
<Link
|
<Link to="/upload" component={DropdownItemForLink} icon="cloud upload" text="Upload tracks" />
|
||||||
to="/upload"
|
<Link to="/settings" component={DropdownItemForLink} icon="cog" text="Settings" />
|
||||||
component={DropdownItemForLink}
|
|
||||||
icon="cloud upload"
|
|
||||||
text={t('App.menu.uploadTracks')}
|
|
||||||
/>
|
|
||||||
<Link to="/settings" component={DropdownItemForLink} icon="cog" text={t('App.menu.settings')} />
|
|
||||||
<Dropdown.Divider />
|
<Dropdown.Divider />
|
||||||
<Link to="/logout" component={DropdownItemForLink} icon="sign-out" text={t('App.menu.logout')} />
|
<Link to="/logout" component={DropdownItemForLink} icon="sign-out" text="Logout" />
|
||||||
</Dropdown.Menu>
|
</Dropdown.Menu>
|
||||||
</Dropdown>
|
</Dropdown>
|
||||||
</>
|
</>
|
||||||
|
@ -132,16 +122,14 @@ const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
<Route path="/" exact>
|
<Route path="/" exact>
|
||||||
<HomePage />
|
<HomePage />
|
||||||
</Route>
|
</Route>
|
||||||
{hasMap && (
|
{hasMap && <Route path="/map" exact>
|
||||||
<Route path="/map" exact>
|
<MapPage />
|
||||||
<MapPage />
|
</Route>}
|
||||||
</Route>
|
|
||||||
)}
|
|
||||||
<Route path="/tracks" exact>
|
<Route path="/tracks" exact>
|
||||||
<TracksPage />
|
<TracksPage />
|
||||||
</Route>
|
</Route>
|
||||||
<Route path="/my/tracks" exact>
|
<Route path="/my/tracks" exact>
|
||||||
<MyTracksPage />
|
<TracksPage privateTracks />
|
||||||
</Route>
|
</Route>
|
||||||
<Route path={`/tracks/:slug`} exact>
|
<Route path={`/tracks/:slug`} exact>
|
||||||
<TrackPage />
|
<TrackPage />
|
||||||
|
@ -152,9 +140,6 @@ const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
<Route path="/export" exact>
|
<Route path="/export" exact>
|
||||||
<ExportPage />
|
<ExportPage />
|
||||||
</Route>
|
</Route>
|
||||||
<Route path="/acknowledgements" exact>
|
|
||||||
<AcknowledgementsPage />
|
|
||||||
</Route>
|
|
||||||
<Route path="/redirect" exact>
|
<Route path="/redirect" exact>
|
||||||
<LoginRedirectPage />
|
<LoginRedirectPage />
|
||||||
</Route>
|
</Route>
|
||||||
|
@ -181,7 +166,7 @@ const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
<Grid columns={4} stackable>
|
<Grid columns={4} stackable>
|
||||||
<Grid.Row>
|
<Grid.Row>
|
||||||
<Grid.Column>
|
<Grid.Column>
|
||||||
<Header as="h5">{t('App.footer.aboutTheProject')}</Header>
|
<Header as="h5">About the project</Header>
|
||||||
<List>
|
<List>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<a href="https://openbikesensor.org/" target="_blank" rel="noreferrer">
|
<a href="https://openbikesensor.org/" target="_blank" rel="noreferrer">
|
||||||
|
@ -192,46 +177,45 @@ const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
</Grid.Column>
|
</Grid.Column>
|
||||||
|
|
||||||
<Grid.Column>
|
<Grid.Column>
|
||||||
<Header as="h5">{t('App.footer.getInvolved')}</Header>
|
<Header as="h5">Get involved</Header>
|
||||||
<List>
|
<List>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<a href="https://forum.openbikesensor.org/" target="_blank" rel="noreferrer">
|
<a href="https://forum.openbikesensor.org/" target="_blank" rel="noreferrer">
|
||||||
{t('App.footer.getHelpInForum')}
|
Get help in forum
|
||||||
</a>
|
</a>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<a href="https://github.com/openbikesensor/portal/issues/new" target="_blank" rel="noreferrer">
|
<a href="https://github.com/openbikesensor/portal/issues/new" target="_blank" rel="noreferrer">
|
||||||
{t('App.footer.reportAnIssue')}
|
Report an issue
|
||||||
</a>
|
</a>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<a href="https://github.com/openbikesensor/portal" target="_blank" rel="noreferrer">
|
<a href="https://github.com/openbikesensor/portal" target="_blank" rel="noreferrer">
|
||||||
{t('App.footer.development')}
|
Development
|
||||||
</a>
|
</a>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
</List>
|
</List>
|
||||||
</Grid.Column>
|
</Grid.Column>
|
||||||
|
|
||||||
<Grid.Column>
|
<Grid.Column>
|
||||||
<Header as="h5">{t('App.footer.thisInstallation')}</Header>
|
<Header as="h5">This installation</Header>
|
||||||
<List>
|
<List>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<a href={config?.privacyPolicyUrl} target="_blank" rel="noreferrer">
|
<a href={config?.privacyPolicyUrl} target="_blank" rel="noreferrer">
|
||||||
{t('App.footer.privacyPolicy')}
|
Privacy policy
|
||||||
</a>
|
</a>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<a href={config?.imprintUrl} target="_blank" rel="noreferrer">
|
<a href={config?.imprintUrl} target="_blank" rel="noreferrer">
|
||||||
{t('App.footer.imprint')}
|
Imprint
|
||||||
</a>
|
</a>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
{config?.termsUrl && (
|
</List>
|
||||||
<List.Item>
|
</Grid.Column>
|
||||||
<a href={config?.termsUrl} target="_blank" rel="noreferrer">
|
|
||||||
{t('App.footer.terms')}
|
<Grid.Column>
|
||||||
</a>
|
<Header as="h5">Info</Header>
|
||||||
</List.Item>
|
<List>
|
||||||
)}
|
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<a
|
<a
|
||||||
href={`https://github.com/openbikesensor/portal${
|
href={`https://github.com/openbikesensor/portal${
|
||||||
|
@ -240,22 +224,11 @@ const App = connect((state) => ({login: state.login}))(function App({login}) {
|
||||||
target="_blank"
|
target="_blank"
|
||||||
rel="noreferrer"
|
rel="noreferrer"
|
||||||
>
|
>
|
||||||
{apiVersion ? t('App.footer.version', {apiVersion}) : t('App.footer.versionLoading')}
|
{apiVersion ? `v${apiVersion}` : 'Fetching version...'}
|
||||||
</a>
|
</a>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
</List>
|
</List>
|
||||||
</Grid.Column>
|
</Grid.Column>
|
||||||
|
|
||||||
<Grid.Column>
|
|
||||||
<Header as="h5">{t('App.footer.changeLanguage')}</Header>
|
|
||||||
<List>
|
|
||||||
{AVAILABLE_LOCALES.map((locale) => (
|
|
||||||
<List.Item key={locale}>
|
|
||||||
<a onClick={() => setLocale(locale)}>{t(`locales.${locale}`)}</a>
|
|
||||||
</List.Item>
|
|
||||||
))}
|
|
||||||
</List>
|
|
||||||
</Grid.Column>
|
|
||||||
</Grid.Row>
|
</Grid.Row>
|
||||||
</Grid>
|
</Grid>
|
||||||
</Container>
|
</Container>
|
||||||
|
|
|
@ -19,21 +19,21 @@ function getColor(s) {
|
||||||
}
|
}
|
||||||
|
|
||||||
export default function Avatar({user, className}) {
|
export default function Avatar({user, className}) {
|
||||||
const {image, displayName} = user || {}
|
const {image, username} = user || {}
|
||||||
|
|
||||||
if (image) {
|
if (image) {
|
||||||
return <Comment.Avatar src={image} className={className} />
|
return <Comment.Avatar src={image} className={className} />
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!displayName) {
|
if (!username) {
|
||||||
return <div className={classnames(className, 'avatar', 'empty-avatar')} />
|
return <div className={classnames(className, 'avatar', 'empty-avatar')} />
|
||||||
}
|
}
|
||||||
|
|
||||||
const color = getColor(displayName)
|
const color = getColor(username)
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<div className={classnames(className, 'avatar', 'text-avatar')} style={{background: color}}>
|
<div className={classnames(className, 'avatar', 'text-avatar')} style={{background: color}}>
|
||||||
{displayName && <span>{displayName[0]}</span>}
|
{username && <span>{username[0]}</span>}
|
||||||
</div>
|
</div>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,77 +0,0 @@
|
||||||
import React from 'react'
|
|
||||||
import ReactEChartsCore from 'echarts-for-react/lib/core'
|
|
||||||
|
|
||||||
import * as echarts from 'echarts/core'
|
|
||||||
|
|
||||||
import {
|
|
||||||
// LineChart,
|
|
||||||
BarChart,
|
|
||||||
// PieChart,
|
|
||||||
// ScatterChart,
|
|
||||||
// RadarChart,
|
|
||||||
// MapChart,
|
|
||||||
// TreeChart,
|
|
||||||
// TreemapChart,
|
|
||||||
// GraphChart,
|
|
||||||
// GaugeChart,
|
|
||||||
// FunnelChart,
|
|
||||||
// ParallelChart,
|
|
||||||
// SankeyChart,
|
|
||||||
// BoxplotChart,
|
|
||||||
// CandlestickChart,
|
|
||||||
// EffectScatterChart,
|
|
||||||
// LinesChart,
|
|
||||||
// HeatmapChart,
|
|
||||||
// PictorialBarChart,
|
|
||||||
// ThemeRiverChart,
|
|
||||||
// SunburstChart,
|
|
||||||
// CustomChart,
|
|
||||||
} from 'echarts/charts'
|
|
||||||
|
|
||||||
// import components, all suffixed with Component
|
|
||||||
import {
|
|
||||||
// GridSimpleComponent,
|
|
||||||
GridComponent,
|
|
||||||
// PolarComponent,
|
|
||||||
// RadarComponent,
|
|
||||||
// GeoComponent,
|
|
||||||
// SingleAxisComponent,
|
|
||||||
// ParallelComponent,
|
|
||||||
// CalendarComponent,
|
|
||||||
// GraphicComponent,
|
|
||||||
// ToolboxComponent,
|
|
||||||
TooltipComponent,
|
|
||||||
// AxisPointerComponent,
|
|
||||||
// BrushComponent,
|
|
||||||
TitleComponent,
|
|
||||||
// TimelineComponent,
|
|
||||||
// MarkPointComponent,
|
|
||||||
// MarkLineComponent,
|
|
||||||
// MarkAreaComponent,
|
|
||||||
// LegendComponent,
|
|
||||||
// LegendScrollComponent,
|
|
||||||
// LegendPlainComponent,
|
|
||||||
// DataZoomComponent,
|
|
||||||
// DataZoomInsideComponent,
|
|
||||||
// DataZoomSliderComponent,
|
|
||||||
// VisualMapComponent,
|
|
||||||
// VisualMapContinuousComponent,
|
|
||||||
// VisualMapPiecewiseComponent,
|
|
||||||
// AriaComponent,
|
|
||||||
// TransformComponent,
|
|
||||||
DatasetComponent,
|
|
||||||
} from 'echarts/components'
|
|
||||||
|
|
||||||
// Import renderer, note that introducing the CanvasRenderer or SVGRenderer is a required step
|
|
||||||
import {
|
|
||||||
CanvasRenderer,
|
|
||||||
// SVGRenderer,
|
|
||||||
} from 'echarts/renderers'
|
|
||||||
|
|
||||||
// Register the required components
|
|
||||||
echarts.use([TitleComponent, TooltipComponent, GridComponent, BarChart, CanvasRenderer])
|
|
||||||
|
|
||||||
// The usage of ReactEChartsCore are same with above.
|
|
||||||
export default function Chart(props) {
|
|
||||||
return <ReactEChartsCore echarts={echarts} notMerge lazyUpdate {...props} />
|
|
||||||
}
|
|
|
@ -1,18 +1,18 @@
|
||||||
import React, {useMemo} from 'react'
|
import React, {useMemo} from "react";
|
||||||
|
|
||||||
import styles from './ColorMapLegend.module.less'
|
|
||||||
|
|
||||||
type ColorMap = [number, string][]
|
type ColorMap = [number, string][]
|
||||||
|
|
||||||
|
import styles from './ColorMapLegend.module.less'
|
||||||
|
|
||||||
function* pairs(arr) {
|
function* pairs(arr) {
|
||||||
for (let i = 1; i < arr.length; i++) {
|
for (let i = 1; i < arr.length; i++) {
|
||||||
yield [arr[i - 1], arr[i]]
|
yield [arr[i - 1], arr[i]];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
function* zip(...arrs) {
|
function* zip(...arrs) {
|
||||||
const l = Math.min(...arrs.map((a) => a.length))
|
const l = Math.min(...arrs.map(a => a.length));
|
||||||
for (let i = 0; i < l; i++) {
|
for (let i = 0; i < l; i++) {
|
||||||
yield arrs.map((a) => a[i])
|
yield arrs.map(a => a[i]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -25,10 +25,10 @@ export function DiscreteColorMapLegend({map}: {map: ColorMap}) {
|
||||||
min -= buffer
|
min -= buffer
|
||||||
max += buffer
|
max += buffer
|
||||||
const normalizeValue = (v) => (v - min) / (max - min)
|
const normalizeValue = (v) => (v - min) / (max - min)
|
||||||
const stopPairs = Array.from(pairs([min, ...stops, max]))
|
const stopPairs = Array.from(pairs([min, ...stops, max]));
|
||||||
|
|
||||||
const gradientId = useMemo(() => `gradient${Math.floor(Math.random() * 1000000)}`, [])
|
const gradientId = useMemo(() => `gradient${Math.floor(Math.random() * 1000000)}`, []);
|
||||||
const gradientUrl = `url(#${gradientId})`
|
const gradientUrl = `url(#${gradientId})`;
|
||||||
|
|
||||||
const parts = Array.from(zip(stopPairs, colors))
|
const parts = Array.from(zip(stopPairs, colors))
|
||||||
|
|
||||||
|
@ -38,10 +38,11 @@ export function DiscreteColorMapLegend({map}: {map: ColorMap}) {
|
||||||
<defs>
|
<defs>
|
||||||
<linearGradient id={gradientId} x1="0" x2="1" y1="0" y2="0">
|
<linearGradient id={gradientId} x1="0" x2="1" y1="0" y2="0">
|
||||||
{parts.map(([[left, right], color]) => (
|
{parts.map(([[left, right], color]) => (
|
||||||
<React.Fragment key={left}>
|
<React.Fragment key={left}>
|
||||||
<stop offset={normalizeValue(left) * 100 + '%'} stopColor={color} />
|
<stop offset={normalizeValue(left) * 100 + '%'} stopColor={color} />
|
||||||
<stop offset={normalizeValue(right) * 100 + '%'} stopColor={color} />
|
<stop offset={normalizeValue(right) * 100 + '%'} stopColor={color} />
|
||||||
</React.Fragment>
|
</React.Fragment>
|
||||||
|
|
||||||
))}
|
))}
|
||||||
</linearGradient>
|
</linearGradient>
|
||||||
</defs>
|
</defs>
|
||||||
|
@ -58,21 +59,13 @@ export function DiscreteColorMapLegend({map}: {map: ColorMap}) {
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
export default function ColorMapLegend({
|
export default function ColorMapLegend({map, twoTicks = false, digits=2}: {map: ColorMap, twoTicks?: boolean, digits?: number}) {
|
||||||
map,
|
|
||||||
twoTicks = false,
|
|
||||||
digits = 2,
|
|
||||||
}: {
|
|
||||||
map: ColorMap
|
|
||||||
twoTicks?: boolean
|
|
||||||
digits?: number
|
|
||||||
}) {
|
|
||||||
const min = map[0][0]
|
const min = map[0][0]
|
||||||
const max = map[map.length - 1][0]
|
const max = map[map.length - 1][0]
|
||||||
const normalizeValue = (v) => (v - min) / (max - min)
|
const normalizeValue = (v) => (v - min) / (max - min)
|
||||||
const gradientId = useMemo(() => `gradient${Math.floor(Math.random() * 1000000)}`, [])
|
const gradientId = useMemo(() => `gradient${Math.floor(Math.random() * 1000000)}`, []);
|
||||||
const gradientUrl = `url(#${gradientId})`
|
const gradientUrl = `url(#${gradientId})`;
|
||||||
const tickValues = twoTicks ? [map[0], map[map.length - 1]] : map
|
const tickValues = twoTicks ? [map[0], map[map.length-1]] : map
|
||||||
return (
|
return (
|
||||||
<div className={styles.colorMapLegend}>
|
<div className={styles.colorMapLegend}>
|
||||||
<svg width="100%" height="20" version="1.1" xmlns="http://www.w3.org/2000/svg">
|
<svg width="100%" height="20" version="1.1" xmlns="http://www.w3.org/2000/svg">
|
||||||
|
|
|
@ -1,11 +1,9 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import {Icon, Segment, Header, Button} from 'semantic-ui-react'
|
import {Icon, Segment, Header, Button} from 'semantic-ui-react'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {FileDrop} from 'components'
|
import {FileDrop} from 'components'
|
||||||
|
|
||||||
export default function FileUploadField({onSelect: onSelect_, multiple}) {
|
export default function FileUploadField({onSelect: onSelect_, multiple}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const labelRef = React.useRef()
|
const labelRef = React.useRef()
|
||||||
const [labelRefState, setLabelRefState] = React.useState()
|
const [labelRefState, setLabelRefState] = React.useState()
|
||||||
|
|
||||||
|
@ -33,14 +31,7 @@ export default function FileUploadField({onSelect: onSelect_, multiple}) {
|
||||||
<input
|
<input
|
||||||
type="file"
|
type="file"
|
||||||
id="upload-field"
|
id="upload-field"
|
||||||
style={{
|
style={{width: 0, height: 0, position: 'fixed', left: -1000, top: -1000, opacity: 0.001}}
|
||||||
width: 0,
|
|
||||||
height: 0,
|
|
||||||
position: 'fixed',
|
|
||||||
left: -1000,
|
|
||||||
top: -1000,
|
|
||||||
opacity: 0.001,
|
|
||||||
}}
|
|
||||||
multiple={multiple}
|
multiple={multiple}
|
||||||
accept=".csv"
|
accept=".csv"
|
||||||
onChange={onChangeField}
|
onChange={onChangeField}
|
||||||
|
@ -59,11 +50,11 @@ export default function FileUploadField({onSelect: onSelect_, multiple}) {
|
||||||
>
|
>
|
||||||
<Header icon>
|
<Header icon>
|
||||||
<Icon name="cloud upload" />
|
<Icon name="cloud upload" />
|
||||||
{multiple ? t('FileUploadField.dropOrClickMultiple') : t('FileUploadField.dropOrClick')}
|
Drop file{multiple ? 's' : ''} here or click to select {multiple ? 'them' : 'one'} for upload
|
||||||
</Header>
|
</Header>
|
||||||
|
|
||||||
<Button primary as="span">
|
<Button primary as="span">
|
||||||
{multiple ? t('FileUploadField.uploadFiles') : t('FileUploadField.uploadFile')}
|
Upload file{multiple ? 's' : ''}
|
||||||
</Button>
|
</Button>
|
||||||
</Segment>
|
</Segment>
|
||||||
)}
|
)}
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
import {DateTime} from 'luxon'
|
import {DateTime} from 'luxon'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
export default function FormattedDate({date, relative = false}) {
|
export default function FormattedDate({date, relative = false}) {
|
||||||
if (date == null) {
|
if (date == null) {
|
||||||
|
@ -11,19 +10,11 @@ export default function FormattedDate({date, relative = false}) {
|
||||||
|
|
||||||
let str
|
let str
|
||||||
|
|
||||||
const {i18n} = useTranslation()
|
|
||||||
const locale = i18n.language
|
|
||||||
|
|
||||||
if (relative) {
|
if (relative) {
|
||||||
str = dateTime.setLocale(locale).toRelative()
|
str = dateTime.toRelative()
|
||||||
} else {
|
} else {
|
||||||
str = dateTime.setLocale(locale).toLocaleString(DateTime.DATETIME_MED)
|
str = dateTime.toLocaleString(DateTime.DATETIME_MED)
|
||||||
}
|
}
|
||||||
|
|
||||||
const iso = dateTime.toISO()
|
return <span title={dateTime.toISO()}>{str}</span>
|
||||||
return (
|
|
||||||
<time dateTime={iso} title={iso}>
|
|
||||||
{str}
|
|
||||||
</time>
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,11 +1,9 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import {Button} from 'semantic-ui-react'
|
import {Button} from 'semantic-ui-react'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
|
|
||||||
export default function LoginButton(props) {
|
export default function LoginButton(props) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const [busy, setBusy] = React.useState(false)
|
const [busy, setBusy] = React.useState(false)
|
||||||
|
|
||||||
const onClick = React.useCallback(
|
const onClick = React.useCallback(
|
||||||
|
@ -21,7 +19,7 @@ export default function LoginButton(props) {
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Button onClick={busy ? null : onClick} loading={busy} {...props}>
|
<Button onClick={busy ? null : onClick} loading={busy} {...props}>
|
||||||
{t('LoginButton.login')}
|
Login
|
||||||
</Button>
|
</Button>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,13 +2,12 @@ import React, {useState, useCallback, useMemo, useEffect} from 'react'
|
||||||
import classnames from 'classnames'
|
import classnames from 'classnames'
|
||||||
import {connect} from 'react-redux'
|
import {connect} from 'react-redux'
|
||||||
import _ from 'lodash'
|
import _ from 'lodash'
|
||||||
import ReactMapGl, {WebMercatorViewport, ScaleControl, NavigationControl, AttributionControl} from 'react-map-gl'
|
import ReactMapGl, {WebMercatorViewport, ScaleControl, NavigationControl} from 'react-map-gl'
|
||||||
import turfBbox from '@turf/bbox'
|
import turfBbox from '@turf/bbox'
|
||||||
import {useHistory, useLocation} from 'react-router-dom'
|
import {useHistory, useLocation} from 'react-router-dom'
|
||||||
|
|
||||||
import {useConfig} from 'config'
|
import {useConfig} from 'config'
|
||||||
|
|
||||||
import {useCallbackRef} from '../../utils'
|
|
||||||
import {baseMapStyles} from '../../mapstyles'
|
import {baseMapStyles} from '../../mapstyles'
|
||||||
|
|
||||||
import styles from './styles.module.less'
|
import styles from './styles.module.less'
|
||||||
|
@ -20,13 +19,11 @@ interface Viewport {
|
||||||
}
|
}
|
||||||
const EMPTY_VIEWPORT: Viewport = {longitude: 0, latitude: 0, zoom: 0}
|
const EMPTY_VIEWPORT: Viewport = {longitude: 0, latitude: 0, zoom: 0}
|
||||||
|
|
||||||
export const withBaseMapStyle = connect((state) => ({
|
export const withBaseMapStyle = connect((state) => ({baseMapStyle: state.mapConfig?.baseMap?.style ?? 'positron'}))
|
||||||
baseMapStyle: state.mapConfig?.baseMap?.style ?? 'positron',
|
|
||||||
}))
|
|
||||||
|
|
||||||
function parseHash(v: string): Viewport | null {
|
function parseHash(v: string): Viewport | null {
|
||||||
if (!v) return null
|
if (!v) return null
|
||||||
const m = v.match(/^#([0-9\.]+)\/([0-9\.\-]+)\/([0-9\.\-]+)$/)
|
const m = v.match(/^#([0-9\.]+)\/([0-9\.]+)\/([0-9\.]+)$/)
|
||||||
if (!m) return null
|
if (!m) return null
|
||||||
return {
|
return {
|
||||||
zoom: Number.parseFloat(m[1]),
|
zoom: Number.parseFloat(m[1]),
|
||||||
|
@ -39,32 +36,19 @@ function buildHash(v: Viewport): string {
|
||||||
return `${v.zoom.toFixed(2)}/${v.latitude}/${v.longitude}`
|
return `${v.zoom.toFixed(2)}/${v.latitude}/${v.longitude}`
|
||||||
}
|
}
|
||||||
|
|
||||||
const setViewportToHash = _.debounce((history, viewport) => {
|
|
||||||
history.replace({
|
|
||||||
hash: buildHash(viewport),
|
|
||||||
})
|
|
||||||
}, 200)
|
|
||||||
|
|
||||||
function useViewportFromUrl(): [Viewport | null, (v: Viewport) => void] {
|
function useViewportFromUrl(): [Viewport | null, (v: Viewport) => void] {
|
||||||
const history = useHistory()
|
const history = useHistory()
|
||||||
const location = useLocation()
|
const location = useLocation()
|
||||||
|
const value = useMemo(() => parseHash(location.hash), [location.hash])
|
||||||
const [cachedValue, setCachedValue] = useState(parseHash(location.hash))
|
|
||||||
|
|
||||||
// when the location hash changes, set the new value to the cache
|
|
||||||
useEffect(() => {
|
|
||||||
setCachedValue(parseHash(location.hash))
|
|
||||||
}, [location.hash])
|
|
||||||
|
|
||||||
const setter = useCallback(
|
const setter = useCallback(
|
||||||
(v) => {
|
(v) => {
|
||||||
setCachedValue(v)
|
history.replace({
|
||||||
setViewportToHash(history, v)
|
hash: buildHash(v),
|
||||||
|
})
|
||||||
},
|
},
|
||||||
[history]
|
[history]
|
||||||
)
|
)
|
||||||
|
return [value || EMPTY_VIEWPORT, setter]
|
||||||
return [cachedValue || EMPTY_VIEWPORT, setter]
|
|
||||||
}
|
}
|
||||||
|
|
||||||
function Map({
|
function Map({
|
||||||
|
@ -72,28 +56,17 @@ function Map({
|
||||||
children,
|
children,
|
||||||
boundsFromJson,
|
boundsFromJson,
|
||||||
baseMapStyle,
|
baseMapStyle,
|
||||||
hasToolbar,
|
|
||||||
onViewportChange,
|
|
||||||
...props
|
...props
|
||||||
}: {
|
}: {
|
||||||
viewportFromUrl?: boolean
|
viewportFromUrl?: boolean
|
||||||
children: React.ReactNode
|
children: React.ReactNode
|
||||||
boundsFromJson: GeoJSON.Geometry
|
boundsFromJson: GeoJSON.Geometry
|
||||||
baseMapStyle: string
|
baseMapStyle: string
|
||||||
hasToolbar?: boolean
|
|
||||||
onViewportChange: (viewport: Viewport) => void
|
|
||||||
}) {
|
}) {
|
||||||
const [viewportState, setViewportState] = useState(EMPTY_VIEWPORT)
|
const [viewportState, setViewportState] = useState(EMPTY_VIEWPORT)
|
||||||
const [viewportUrl, setViewportUrl] = useViewportFromUrl()
|
const [viewportUrl, setViewportUrl] = useViewportFromUrl()
|
||||||
|
|
||||||
const [viewport, setViewport_] = viewportFromUrl ? [viewportUrl, setViewportUrl] : [viewportState, setViewportState]
|
const [viewport, setViewport] = viewportFromUrl ? [viewportUrl, setViewportUrl] : [viewportState, setViewportState]
|
||||||
const setViewport = useCallback(
|
|
||||||
(viewport: Viewport) => {
|
|
||||||
setViewport_(viewport)
|
|
||||||
onViewportChange?.(viewport)
|
|
||||||
},
|
|
||||||
[setViewport_, onViewportChange]
|
|
||||||
)
|
|
||||||
|
|
||||||
const config = useConfig()
|
const config = useConfig()
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
|
@ -102,29 +75,12 @@ function Map({
|
||||||
}
|
}
|
||||||
}, [config, boundsFromJson])
|
}, [config, boundsFromJson])
|
||||||
|
|
||||||
const mapSourceHosts = useMemo(
|
|
||||||
() => _.uniq(config?.obsMapSource?.tiles?.map((tileUrl: string) => new URL(tileUrl).host) ?? []),
|
|
||||||
[config?.obsMapSource]
|
|
||||||
)
|
|
||||||
|
|
||||||
const transformRequest = useCallbackRef((url, resourceType) => {
|
|
||||||
if (resourceType === 'Tile' && mapSourceHosts.includes(new URL(url).host)) {
|
|
||||||
return {
|
|
||||||
url,
|
|
||||||
credentials: 'include',
|
|
||||||
}
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
if (boundsFromJson) {
|
if (boundsFromJson) {
|
||||||
const bbox = turfBbox(boundsFromJson)
|
const bbox = turfBbox(boundsFromJson);
|
||||||
if (bbox.every((v) => Math.abs(v) !== Infinity)) {
|
if (bbox.every(v => Math.abs(v) !== Infinity)) {
|
||||||
const [minX, minY, maxX, maxY] = bbox
|
const [minX, minY, maxX, maxY] = bbox;
|
||||||
const vp = new WebMercatorViewport({
|
const vp = new WebMercatorViewport({width: 1000, height: 800}).fitBounds(
|
||||||
width: 1000,
|
|
||||||
height: 800,
|
|
||||||
}).fitBounds(
|
|
||||||
[
|
[
|
||||||
[minX, minY],
|
[minX, minY],
|
||||||
[maxX, maxY],
|
[maxX, maxY],
|
||||||
|
@ -145,15 +101,13 @@ function Map({
|
||||||
width="100%"
|
width="100%"
|
||||||
height="100%"
|
height="100%"
|
||||||
onViewportChange={setViewport}
|
onViewportChange={setViewport}
|
||||||
{...{transformRequest}}
|
|
||||||
{...viewport}
|
{...viewport}
|
||||||
{...props}
|
{...props}
|
||||||
className={classnames(styles.map, props.className)}
|
className={classnames(styles.map, props.className)}
|
||||||
attributionControl={false}
|
|
||||||
>
|
>
|
||||||
<AttributionControl style={{top: 0, right: 0}} />
|
<NavigationControl style={{left: 10, top: 10}} />
|
||||||
<NavigationControl showCompass={false} style={{left: 16, top: hasToolbar ? 64 : 16}} />
|
<ScaleControl maxWidth={200} unit="metric" style={{left: 10, bottom: 10}} />
|
||||||
<ScaleControl maxWidth={200} unit="metric" style={{left: 16, bottom: 16}} />
|
|
||||||
{children}
|
{children}
|
||||||
</ReactMapGl>
|
</ReactMapGl>
|
||||||
)
|
)
|
||||||
|
|
|
@ -1,9 +1,9 @@
|
||||||
import React from 'react'
|
import React from "react";
|
||||||
import classnames from 'classnames'
|
import classnames from "classnames";
|
||||||
import {Container} from 'semantic-ui-react'
|
import { Container } from "semantic-ui-react";
|
||||||
import {Helmet} from 'react-helmet'
|
import { Helmet } from "react-helmet";
|
||||||
|
|
||||||
import styles from './Page.module.less'
|
import styles from "./Page.module.less";
|
||||||
|
|
||||||
export default function Page({
|
export default function Page({
|
||||||
small,
|
small,
|
||||||
|
@ -12,11 +12,11 @@ export default function Page({
|
||||||
stage,
|
stage,
|
||||||
title,
|
title,
|
||||||
}: {
|
}: {
|
||||||
small?: boolean
|
small?: boolean;
|
||||||
children: ReactNode
|
children: ReactNode;
|
||||||
fullScreen?: boolean
|
fullScreen?: boolean;
|
||||||
stage?: ReactNode
|
stage?: ReactNode;
|
||||||
title?: string
|
title?: string;
|
||||||
}) {
|
}) {
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
|
@ -37,5 +37,5 @@ export default function Page({
|
||||||
{fullScreen ? children : <Container>{children}</Container>}
|
{fullScreen ? children : <Container>{children}</Container>}
|
||||||
</main>
|
</main>
|
||||||
</>
|
</>
|
||||||
)
|
);
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,73 +0,0 @@
|
||||||
import React, {useState, useCallback} from 'react'
|
|
||||||
import {pickBy} from 'lodash'
|
|
||||||
import {Loader, Statistic, Pagination, Segment, Header, Menu, Table, Icon} from 'semantic-ui-react'
|
|
||||||
import {useObservable} from 'rxjs-hooks'
|
|
||||||
import {of, from, concat, combineLatest} from 'rxjs'
|
|
||||||
import {map, switchMap, distinctUntilChanged} from 'rxjs/operators'
|
|
||||||
import {Duration, DateTime} from 'luxon'
|
|
||||||
|
|
||||||
import api from 'api'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
function formatDuration(seconds) {
|
|
||||||
return (
|
|
||||||
Duration.fromMillis((seconds ?? 0) * 1000)
|
|
||||||
.as('hours')
|
|
||||||
.toFixed(1) + ' h'
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
export default function Stats() {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const [page, setPage] = useState(1)
|
|
||||||
const PER_PAGE = 10
|
|
||||||
const stats = useObservable(
|
|
||||||
() => of(null).pipe(switchMap(() => concat(of(null), from(api.get('/stats/regions'))))),
|
|
||||||
null
|
|
||||||
)
|
|
||||||
|
|
||||||
const pageCount = stats ? Math.ceil(stats.length / PER_PAGE) : 1
|
|
||||||
|
|
||||||
return (
|
|
||||||
<>
|
|
||||||
<Header as="h2">{t('RegionStats.title')}</Header>
|
|
||||||
|
|
||||||
<div>
|
|
||||||
<Loader active={stats == null} />
|
|
||||||
|
|
||||||
<Table celled>
|
|
||||||
<Table.Header>
|
|
||||||
<Table.Row>
|
|
||||||
<Table.HeaderCell> {t('RegionStats.regionName')}</Table.HeaderCell>
|
|
||||||
<Table.HeaderCell>{t('RegionStats.eventCount')}</Table.HeaderCell>
|
|
||||||
</Table.Row>
|
|
||||||
</Table.Header>
|
|
||||||
|
|
||||||
<Table.Body>
|
|
||||||
{stats?.slice((page - 1) * PER_PAGE, page * PER_PAGE)?.map((area) => (
|
|
||||||
<Table.Row key={area.id}>
|
|
||||||
<Table.Cell>{area.name}</Table.Cell>
|
|
||||||
<Table.Cell>{area.overtaking_event_count}</Table.Cell>
|
|
||||||
</Table.Row>
|
|
||||||
))}
|
|
||||||
</Table.Body>
|
|
||||||
|
|
||||||
{pageCount > 1 && (
|
|
||||||
<Table.Footer>
|
|
||||||
<Table.Row>
|
|
||||||
<Table.HeaderCell colSpan="2">
|
|
||||||
<Pagination
|
|
||||||
floated="right"
|
|
||||||
activePage={page}
|
|
||||||
totalPages={pageCount}
|
|
||||||
onPageChange={(e, data) => setPage(data.activePage as number)}
|
|
||||||
/>
|
|
||||||
</Table.HeaderCell>
|
|
||||||
</Table.Row>
|
|
||||||
</Table.Footer>
|
|
||||||
)}
|
|
||||||
</Table>
|
|
||||||
</div>
|
|
||||||
</>
|
|
||||||
)
|
|
||||||
}
|
|
|
@ -5,7 +5,6 @@ import {useObservable} from 'rxjs-hooks'
|
||||||
import {of, from, concat, combineLatest} from 'rxjs'
|
import {of, from, concat, combineLatest} from 'rxjs'
|
||||||
import {map, switchMap, distinctUntilChanged} from 'rxjs/operators'
|
import {map, switchMap, distinctUntilChanged} from 'rxjs/operators'
|
||||||
import {Duration, DateTime} from 'luxon'
|
import {Duration, DateTime} from 'luxon'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
|
|
||||||
|
@ -18,7 +17,6 @@ function formatDuration(seconds) {
|
||||||
}
|
}
|
||||||
|
|
||||||
export default function Stats({user = null}: {user?: null | string}) {
|
export default function Stats({user = null}: {user?: null | string}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const [timeframe, setTimeframe] = useState('all_time')
|
const [timeframe, setTimeframe] = useState('all_time')
|
||||||
const onClick = useCallback((_e, {name}) => setTimeframe(name), [setTimeframe])
|
const onClick = useCallback((_e, {name}) => setTimeframe(name), [setTimeframe])
|
||||||
|
|
||||||
|
@ -65,56 +63,49 @@ export default function Stats({user = null}: {user?: null | string}) {
|
||||||
[timeframe, user]
|
[timeframe, user]
|
||||||
)
|
)
|
||||||
|
|
||||||
const placeholder = t('Stats.placeholder')
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
|
<Header as="h2">{user && 'My '}Statistics</Header>
|
||||||
|
|
||||||
<div>
|
<div>
|
||||||
<Segment attached="top">
|
<Segment attached="top">
|
||||||
<Loader active={stats == null} />
|
<Loader active={stats == null} />
|
||||||
<Statistic.Group widths={2} size="tiny">
|
<Statistic.Group widths={2} size="tiny">
|
||||||
<Statistic>
|
<Statistic>
|
||||||
<Statistic.Value>
|
<Statistic.Value>{stats ? `${Number(stats?.trackLength / 1000).toFixed(1)} km` : '...'}</Statistic.Value>
|
||||||
{stats ? `${Number(stats?.trackLength / 1000).toFixed(1)} km` : placeholder}
|
<Statistic.Label>Total track length</Statistic.Label>
|
||||||
</Statistic.Value>
|
|
||||||
<Statistic.Label>{t('Stats.totalTrackLength')}</Statistic.Label>
|
|
||||||
</Statistic>
|
</Statistic>
|
||||||
<Statistic>
|
<Statistic>
|
||||||
<Statistic.Value>{stats ? formatDuration(stats?.trackDuration) : placeholder}</Statistic.Value>
|
<Statistic.Value>{stats ? formatDuration(stats?.trackDuration) : '...'}</Statistic.Value>
|
||||||
<Statistic.Label>{t('Stats.timeRecorded')}</Statistic.Label>
|
<Statistic.Label>Time recorded</Statistic.Label>
|
||||||
</Statistic>
|
</Statistic>
|
||||||
<Statistic>
|
<Statistic>
|
||||||
<Statistic.Value>{stats?.numEvents ?? placeholder}</Statistic.Value>
|
<Statistic.Value>{stats?.numEvents ?? '...'}</Statistic.Value>
|
||||||
<Statistic.Label>{t('Stats.eventsConfirmed')}</Statistic.Label>
|
<Statistic.Label>Events confirmed</Statistic.Label>
|
||||||
</Statistic>
|
</Statistic>
|
||||||
<Statistic>
|
{user ? (
|
||||||
<Statistic.Value>{stats?.trackCount ?? placeholder}</Statistic.Value>
|
<Statistic>
|
||||||
<Statistic.Label>{t('Stats.tracksRecorded')}</Statistic.Label>
|
<Statistic.Value>{stats?.trackCount ?? '...'}</Statistic.Value>
|
||||||
</Statistic>
|
<Statistic.Label>Tracks recorded</Statistic.Label>
|
||||||
{!user && (
|
</Statistic>
|
||||||
<>
|
) : (
|
||||||
<Statistic>
|
<Statistic>
|
||||||
<Statistic.Value>{stats?.userCount ?? placeholder}</Statistic.Value>
|
<Statistic.Value>{stats?.userCount ?? '...'}</Statistic.Value>
|
||||||
<Statistic.Label>{t('Stats.membersJoined')}</Statistic.Label>
|
<Statistic.Label>Members joined</Statistic.Label>
|
||||||
</Statistic>
|
</Statistic>
|
||||||
<Statistic>
|
|
||||||
<Statistic.Value>{stats?.deviceCount ?? placeholder}</Statistic.Value>
|
|
||||||
<Statistic.Label>{t('Stats.deviceCount')}</Statistic.Label>
|
|
||||||
</Statistic>
|
|
||||||
</>
|
|
||||||
)}
|
)}
|
||||||
</Statistic.Group>
|
</Statistic.Group>
|
||||||
</Segment>
|
</Segment>
|
||||||
|
|
||||||
<Menu widths={3} attached="bottom" size="small">
|
<Menu widths={3} attached="bottom" size="small">
|
||||||
<Menu.Item name="this_month" active={timeframe === 'this_month'} onClick={onClick}>
|
<Menu.Item name="this_month" active={timeframe === 'this_month'} onClick={onClick}>
|
||||||
{t('Stats.thisMonth')}
|
This month
|
||||||
</Menu.Item>
|
</Menu.Item>
|
||||||
<Menu.Item name="this_year" active={timeframe === 'this_year'} onClick={onClick}>
|
<Menu.Item name="this_year" active={timeframe === 'this_year'} onClick={onClick}>
|
||||||
{t('Stats.thisYear')}
|
This year
|
||||||
</Menu.Item>
|
</Menu.Item>
|
||||||
<Menu.Item name="all_time" active={timeframe === 'all_time'} onClick={onClick}>
|
<Menu.Item name="all_time" active={timeframe === 'all_time'} onClick={onClick}>
|
||||||
{t('Stats.allTime')}
|
All time
|
||||||
</Menu.Item>
|
</Menu.Item>
|
||||||
</Menu>
|
</Menu>
|
||||||
</div>
|
</div>
|
||||||
|
|
|
@ -1,14 +0,0 @@
|
||||||
import React from 'react'
|
|
||||||
import {Icon} from 'semantic-ui-react'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
export default function Visibility({public: public_}: {public: boolean}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const icon = public_ ? <Icon color="blue" name="eye" fitted /> : <Icon name="eye slash" fitted />
|
|
||||||
const text = public_ ? t('general.public') : t('general.private')
|
|
||||||
return (
|
|
||||||
<>
|
|
||||||
{icon} {text}
|
|
||||||
</>
|
|
||||||
)
|
|
||||||
}
|
|
|
@ -1,5 +1,4 @@
|
||||||
export {default as Avatar} from './Avatar'
|
export {default as Avatar} from './Avatar'
|
||||||
export {default as Chart} from './Chart'
|
|
||||||
export {default as ColorMapLegend, DiscreteColorMapLegend} from './ColorMapLegend'
|
export {default as ColorMapLegend, DiscreteColorMapLegend} from './ColorMapLegend'
|
||||||
export {default as FileDrop} from './FileDrop'
|
export {default as FileDrop} from './FileDrop'
|
||||||
export {default as FileUploadField} from './FileUploadField'
|
export {default as FileUploadField} from './FileUploadField'
|
||||||
|
@ -7,7 +6,5 @@ export {default as FormattedDate} from './FormattedDate'
|
||||||
export {default as LoginButton} from './LoginButton'
|
export {default as LoginButton} from './LoginButton'
|
||||||
export {default as Map} from './Map'
|
export {default as Map} from './Map'
|
||||||
export {default as Page} from './Page'
|
export {default as Page} from './Page'
|
||||||
export {default as RegionStats} from './RegionStats'
|
|
||||||
export {default as Stats} from './Stats'
|
export {default as Stats} from './Stats'
|
||||||
export {default as StripMarkdown} from './StripMarkdown'
|
export {default as StripMarkdown} from './StripMarkdown'
|
||||||
export {default as Visibility} from './Visibility'
|
|
||||||
|
|
|
@ -1,11 +1,16 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
|
|
||||||
export type MapSource = {
|
export type MapSoure =
|
||||||
type: 'vector'
|
| {
|
||||||
tiles: string[]
|
type: 'vector'
|
||||||
minzoom: number
|
url: string
|
||||||
maxzoom: number
|
}
|
||||||
}
|
| {
|
||||||
|
type: 'vector'
|
||||||
|
tiles: string[]
|
||||||
|
minzoom: number
|
||||||
|
maxzoom: number
|
||||||
|
}
|
||||||
|
|
||||||
export interface Config {
|
export interface Config {
|
||||||
apiUrl: string
|
apiUrl: string
|
||||||
|
@ -14,10 +19,9 @@ export interface Config {
|
||||||
longitude: number
|
longitude: number
|
||||||
zoom: number
|
zoom: number
|
||||||
}
|
}
|
||||||
obsMapSource?: MapSource
|
obsMapSource?: MapSoure
|
||||||
imprintUrl?: string
|
imprintUrl?: string
|
||||||
privacyPolicyUrl?: string
|
privacyPolicyUrl?: string
|
||||||
termsUrl?: string
|
|
||||||
banner?: {
|
banner?: {
|
||||||
text: string
|
text: string
|
||||||
style?: 'warning' | 'info'
|
style?: 'warning' | 'info'
|
||||||
|
|
|
@ -1,87 +0,0 @@
|
||||||
import {useState, useEffect, useMemo} from 'react'
|
|
||||||
import i18next, {TOptions} from 'i18next'
|
|
||||||
import {BehaviorSubject, combineLatest} from 'rxjs'
|
|
||||||
import {map, distinctUntilChanged} from 'rxjs/operators'
|
|
||||||
import HttpBackend, {BackendOptions, RequestCallback} from 'i18next-http-backend'
|
|
||||||
import {initReactI18next} from 'react-i18next'
|
|
||||||
import LanguageDetector from 'i18next-browser-languagedetector'
|
|
||||||
|
|
||||||
export type AvailableLocales = 'en' | 'de' | 'fr'
|
|
||||||
|
|
||||||
async function request(_options: BackendOptions, url: string, _payload: any, callback: RequestCallback) {
|
|
||||||
try {
|
|
||||||
const [lng] = url.split('/')
|
|
||||||
const locale = await import(`translations/${lng}.yaml`)
|
|
||||||
callback(null, {status: 200, data: locale})
|
|
||||||
} catch (e) {
|
|
||||||
console.error(`Unable to load locale at ${url}\n`, e)
|
|
||||||
callback(null, {status: 404, data: String(e)})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
export const AVAILABLE_LOCALES: AvailableLocales[] = ['en', 'de', 'fr']
|
|
||||||
|
|
||||||
const i18n = i18next.createInstance()
|
|
||||||
|
|
||||||
const options: TOptions = {
|
|
||||||
fallbackLng: 'en',
|
|
||||||
|
|
||||||
ns: ['common'],
|
|
||||||
defaultNS: 'common',
|
|
||||||
whitelist: AVAILABLE_LOCALES,
|
|
||||||
|
|
||||||
// loading via webpack
|
|
||||||
backend: {
|
|
||||||
loadPath: '{{lng}}/{{ns}}',
|
|
||||||
parse: (data: any) => data,
|
|
||||||
request,
|
|
||||||
},
|
|
||||||
|
|
||||||
load: 'languageOnly',
|
|
||||||
|
|
||||||
interpolation: {
|
|
||||||
escapeValue: false, // not needed for react as it escapes by default
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
i18n
|
|
||||||
.use(HttpBackend)
|
|
||||||
.use(initReactI18next)
|
|
||||||
.use(LanguageDetector)
|
|
||||||
.init({...options})
|
|
||||||
|
|
||||||
const locale$ = new BehaviorSubject<AvailableLocales>('en')
|
|
||||||
|
|
||||||
export const translate = i18n.t.bind(i18n)
|
|
||||||
|
|
||||||
export const translate$ = (stringAndData$: [string, any]) =>
|
|
||||||
combineLatest([stringAndData$, locale$.pipe(distinctUntilChanged())]).pipe(
|
|
||||||
map(([stringAndData]) => {
|
|
||||||
if (typeof stringAndData === 'string') {
|
|
||||||
return i18n.t(stringAndData)
|
|
||||||
} else {
|
|
||||||
const [string, data] = stringAndData
|
|
||||||
return i18n.t(string, {data})
|
|
||||||
}
|
|
||||||
})
|
|
||||||
)
|
|
||||||
|
|
||||||
export const setLocale = (locale: AvailableLocales) => {
|
|
||||||
i18n.changeLanguage(locale)
|
|
||||||
locale$.next(locale)
|
|
||||||
}
|
|
||||||
|
|
||||||
export function useLocale() {
|
|
||||||
const [, reload] = useState()
|
|
||||||
|
|
||||||
useEffect(() => {
|
|
||||||
i18n.on('languageChanged', reload)
|
|
||||||
return () => {
|
|
||||||
i18n.off('languageChanged', reload)
|
|
||||||
}
|
|
||||||
}, [])
|
|
||||||
|
|
||||||
return i18n.language
|
|
||||||
}
|
|
||||||
|
|
||||||
export default i18n
|
|
11
frontend/src/index.css
Normal file
11
frontend/src/index.css
Normal file
|
@ -0,0 +1,11 @@
|
||||||
|
body {
|
||||||
|
margin: 0;
|
||||||
|
font-family: 'Noto Sans', 'Roboto', -apple-system, BlinkMacSystemFont, 'Segoe UI', 'Oxygen', 'Ubuntu', 'Cantarell',
|
||||||
|
'Fira Sans', 'Droid Sans', 'Helvetica Neue', sans-serif;
|
||||||
|
-webkit-font-smoothing: antialiased;
|
||||||
|
-moz-osx-font-smoothing: grayscale;
|
||||||
|
}
|
||||||
|
|
||||||
|
code {
|
||||||
|
font-family: 'Noto Sans Mono', source-code-pro, Menlo, Monaco, Consolas, 'Courier New', monospace;
|
||||||
|
}
|
|
@ -1,9 +1,9 @@
|
||||||
import React, {Suspense} from 'react'
|
import React from 'react'
|
||||||
import {Settings} from 'luxon'
|
import {Settings} from 'luxon'
|
||||||
import ReactDOM from 'react-dom'
|
import ReactDOM from 'react-dom'
|
||||||
import 'fomantic-ui-less/semantic.less'
|
import 'fomantic-ui-less/semantic.less'
|
||||||
|
|
||||||
import './index.less'
|
import './index.css'
|
||||||
import App from './App'
|
import App from './App'
|
||||||
|
|
||||||
import 'maplibre-gl/dist/maplibre-gl.css'
|
import 'maplibre-gl/dist/maplibre-gl.css'
|
||||||
|
@ -11,16 +11,13 @@ import 'maplibre-gl/dist/maplibre-gl.css'
|
||||||
import {Provider} from 'react-redux'
|
import {Provider} from 'react-redux'
|
||||||
|
|
||||||
import store from './store'
|
import store from './store'
|
||||||
import './i18n'
|
|
||||||
|
|
||||||
// TODO: remove
|
// TODO: remove
|
||||||
Settings.defaultLocale = 'de-DE'
|
Settings.defaultLocale = 'de-DE'
|
||||||
|
|
||||||
ReactDOM.render(
|
ReactDOM.render(
|
||||||
<Provider store={store}>
|
<Provider store={store}>
|
||||||
<Suspense fallback={null}>
|
<App />
|
||||||
<App />
|
|
||||||
</Suspense>
|
|
||||||
</Provider>,
|
</Provider>,
|
||||||
document.getElementById('root')
|
document.getElementById('root')
|
||||||
)
|
)
|
||||||
|
|
|
@ -1,8 +0,0 @@
|
||||||
@import 'styles.less';
|
|
||||||
|
|
||||||
body {
|
|
||||||
margin: 0;
|
|
||||||
font-family: @fontFamilyDefault;
|
|
||||||
-webkit-font-smoothing: antialiased;
|
|
||||||
-moz-osx-font-smoothing: grayscale;
|
|
||||||
}
|
|
|
@ -23,13 +23,7 @@ function rgbArrayToColor(arr) {
|
||||||
}
|
}
|
||||||
|
|
||||||
function rgbArrayToHtml(arr) {
|
function rgbArrayToHtml(arr) {
|
||||||
return (
|
return "#" + arr.map((v) => Math.round(v * 255).toString(16)).map(v => (v.length == 1 ? '0' : '') + v).join('')
|
||||||
'#' +
|
|
||||||
arr
|
|
||||||
.map((v) => Math.round(v * 255).toString(16))
|
|
||||||
.map((v) => (v.length == 1 ? '0' : '') + v)
|
|
||||||
.join('')
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export function colormapToScale(colormap, value, min, max) {
|
export function colormapToScale(colormap, value, min, max) {
|
||||||
|
@ -44,74 +38,32 @@ export function colormapToScale(colormap, value, min, max) {
|
||||||
export const viridis = simplifyColormap(viridisBase.map(rgbArrayToColor), 20)
|
export const viridis = simplifyColormap(viridisBase.map(rgbArrayToColor), 20)
|
||||||
export const viridisSimpleHtml = simplifyColormap(viridisBase.map(rgbArrayToHtml), 10)
|
export const viridisSimpleHtml = simplifyColormap(viridisBase.map(rgbArrayToHtml), 10)
|
||||||
export const grayscale = ['#FFFFFF', '#000000']
|
export const grayscale = ['#FFFFFF', '#000000']
|
||||||
export const reds = ['rgba( 255, 0, 0, 0)', 'rgba( 255, 0, 0, 255)']
|
export const reds = [
|
||||||
|
'rgba( 255, 0, 0, 0)',
|
||||||
|
'rgba( 255, 0, 0, 255)',
|
||||||
|
]
|
||||||
|
|
||||||
export function colorByCount(attribute = 'event_count', maxCount, colormap = viridis) {
|
export function colorByCount(attribute = 'event_count', maxCount, colormap = viridis) {
|
||||||
return colormapToScale(colormap, ['case', isValidAttribute(attribute), ['get', attribute], 0], 0, maxCount)
|
return colormapToScale(colormap, ['case', ['to-boolean', ['get', attribute]], ['get', attribute], 0], 0, maxCount)
|
||||||
}
|
}
|
||||||
|
|
||||||
var steps = {rural: [1.6, 1.8, 2.0, 2.2], urban: [1.1, 1.3, 1.5, 1.7]}
|
export function colorByDistance(attribute = 'distance_overtaker_mean', fallback = '#ABC') {
|
||||||
|
|
||||||
export function isValidAttribute(attribute) {
|
|
||||||
if (attribute.endsWith('zone')) {
|
|
||||||
return ['in', ['get', attribute], ['literal', ['rural', 'urban']]]
|
|
||||||
}
|
|
||||||
return ['to-boolean', ['get', attribute]]
|
|
||||||
}
|
|
||||||
|
|
||||||
export function borderByZone() {
|
|
||||||
return ['match', ['get', 'zone'], 'rural', 'cyan', 'urban', 'blue', 'purple']
|
|
||||||
}
|
|
||||||
|
|
||||||
export function colorByDistance(attribute = 'distance_overtaker_mean', fallback = '#ABC', zone = 'urban') {
|
|
||||||
return [
|
return [
|
||||||
'case',
|
'case',
|
||||||
['!', isValidAttribute(attribute)],
|
['!', ['to-boolean', ['get', attribute]]],
|
||||||
fallback,
|
fallback,
|
||||||
[
|
[
|
||||||
'match',
|
'step',
|
||||||
['get', 'zone'],
|
['get', attribute],
|
||||||
'rural',
|
'rgba(150, 0, 0, 1)',
|
||||||
[
|
1.1,
|
||||||
'step',
|
'rgba(255, 0, 0, 1)',
|
||||||
['get', attribute],
|
1.3,
|
||||||
'rgba(150, 0, 0, 1)',
|
'rgba(255, 220, 0, 1)',
|
||||||
steps['rural'][0],
|
1.5,
|
||||||
'rgba(255, 0, 0, 1)',
|
'rgba(67, 200, 0, 1)',
|
||||||
steps['rural'][1],
|
1.7,
|
||||||
'rgba(255, 220, 0, 1)',
|
'rgba(67, 150, 0, 1)',
|
||||||
steps['rural'][2],
|
|
||||||
'rgba(67, 200, 0, 1)',
|
|
||||||
steps['rural'][3],
|
|
||||||
'rgba(67, 150, 0, 1)',
|
|
||||||
],
|
|
||||||
'urban',
|
|
||||||
[
|
|
||||||
'step',
|
|
||||||
['get', attribute],
|
|
||||||
'rgba(150, 0, 0, 1)',
|
|
||||||
steps['urban'][0],
|
|
||||||
'rgba(255, 0, 0, 1)',
|
|
||||||
steps['urban'][1],
|
|
||||||
'rgba(255, 220, 0, 1)',
|
|
||||||
steps['urban'][2],
|
|
||||||
'rgba(67, 200, 0, 1)',
|
|
||||||
steps['urban'][3],
|
|
||||||
'rgba(67, 150, 0, 1)',
|
|
||||||
],
|
|
||||||
[
|
|
||||||
'step',
|
|
||||||
['get', attribute],
|
|
||||||
'rgba(150, 0, 0, 1)',
|
|
||||||
steps['urban'][0],
|
|
||||||
'rgba(255, 0, 0, 1)',
|
|
||||||
steps['urban'][1],
|
|
||||||
'rgba(255, 220, 0, 1)',
|
|
||||||
steps['urban'][2],
|
|
||||||
'rgba(67, 200, 0, 1)',
|
|
||||||
steps['urban'][3],
|
|
||||||
'rgba(67, 150, 0, 1)',
|
|
||||||
],
|
|
||||||
],
|
],
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
@ -121,66 +73,65 @@ export const trackLayer = {
|
||||||
paint: {
|
paint: {
|
||||||
'line-width': ['interpolate', ['linear'], ['zoom'], 14, 2, 17, 5],
|
'line-width': ['interpolate', ['linear'], ['zoom'], 14, 2, 17, 5],
|
||||||
'line-color': '#F06292',
|
'line-color': '#F06292',
|
||||||
'line-opacity': 0.6,
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
export const getRegionLayers = (adminLevel = 6, baseColor = '#00897B', maxValue = 5000) => [
|
export const getRegionLayers = (adminLevel = 6, baseColor = "#00897B", maxValue = 5000) => [{
|
||||||
{
|
id: 'region',
|
||||||
id: 'region',
|
"type": "fill",
|
||||||
type: 'fill',
|
"source": "obs",
|
||||||
source: 'obs',
|
"source-layer": "obs_regions",
|
||||||
'source-layer': 'obs_regions',
|
"minzoom": 0,
|
||||||
minzoom: 0,
|
"maxzoom": 10,
|
||||||
maxzoom: 10,
|
"filter": [
|
||||||
// filter: [">", "overtaking_event_count", 0],
|
"all",
|
||||||
paint: {
|
["==", "admin_level", adminLevel]
|
||||||
'fill-color': baseColor,
|
],
|
||||||
'fill-antialias': true,
|
"paint": {
|
||||||
'fill-opacity': [
|
"fill-color": baseColor,
|
||||||
'interpolate',
|
"fill-antialias": true,
|
||||||
['linear'],
|
"fill-opacity": [
|
||||||
['log10', ['max', ['get', 'overtaking_event_count'], 1]],
|
"interpolate",
|
||||||
0,
|
["linear"],
|
||||||
0,
|
[
|
||||||
Math.log10(maxValue),
|
"log10",
|
||||||
0.9,
|
[
|
||||||
|
"get",
|
||||||
|
"overtaking_event_count"
|
||||||
|
]
|
||||||
],
|
],
|
||||||
},
|
0,
|
||||||
|
0,
|
||||||
|
Math.log10(maxValue),
|
||||||
|
0.9
|
||||||
|
]
|
||||||
},
|
},
|
||||||
{
|
},
|
||||||
id: 'region-border',
|
{
|
||||||
type: 'line',
|
id: 'region-border',
|
||||||
source: 'obs',
|
"type": "line",
|
||||||
'source-layer': 'obs_regions',
|
"source": "obs",
|
||||||
minzoom: 0,
|
"source-layer": "obs_regions",
|
||||||
maxzoom: 10,
|
"minzoom": 0,
|
||||||
// filter: [">", "overtaking_event_count", 0],
|
"maxzoom": 10,
|
||||||
paint: {
|
"filter": [
|
||||||
'line-width': [
|
"all",
|
||||||
'interpolate',
|
["==", "admin_level", adminLevel]
|
||||||
['linear'],
|
],
|
||||||
['log10', ['max', ['get', 'overtaking_event_count'], 1]],
|
"paint": {
|
||||||
0,
|
"line-width": 1,
|
||||||
0.2,
|
"line-color": baseColor,
|
||||||
Math.log10(maxValue),
|
|
||||||
1.5,
|
|
||||||
],
|
|
||||||
'line-color': baseColor,
|
|
||||||
},
|
|
||||||
layout: {
|
|
||||||
'line-join': 'round',
|
|
||||||
'line-cap': 'round',
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
]
|
"layout": {
|
||||||
|
"line-join": "round",
|
||||||
|
"line-cap": "round"
|
||||||
|
}
|
||||||
|
}]
|
||||||
|
|
||||||
export const trackLayerRaw = produce(trackLayer, (draft) => {
|
export const trackLayerRaw = produce(trackLayer, draft => {
|
||||||
// draft.paint['line-color'] = '#81D4FA'
|
draft.paint['line-color'] = '#81D4FA'
|
||||||
draft.paint['line-width'][4] = 1
|
draft.paint['line-width'][4] = 1
|
||||||
draft.paint['line-width'][6] = 2
|
draft.paint['line-width'][6] = 3
|
||||||
draft.paint['line-dasharray'] = [3, 3]
|
|
||||||
delete draft.paint['line-opacity']
|
|
||||||
})
|
})
|
||||||
|
|
||||||
export const basemap = positron
|
export const basemap = positron
|
||||||
|
|
|
@ -1,18 +0,0 @@
|
||||||
import React from 'react'
|
|
||||||
import {Header} from 'semantic-ui-react'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
import Markdown from 'react-markdown'
|
|
||||||
|
|
||||||
import {Page} from 'components'
|
|
||||||
|
|
||||||
export default function AcknowledgementsPage() {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const title = t('AcknowledgementsPage.title')
|
|
||||||
|
|
||||||
return (
|
|
||||||
<Page title={title}>
|
|
||||||
<Header as="h2">{title}</Header>
|
|
||||||
<Markdown>{t('AcknowledgementsPage.information')}</Markdown>
|
|
||||||
</Page>
|
|
||||||
)
|
|
||||||
}
|
|
|
@ -1,146 +1,166 @@
|
||||||
import React, {useState, useCallback, useMemo} from 'react'
|
import React, { useState, useCallback, useMemo } from "react";
|
||||||
import {Source, Layer} from 'react-map-gl'
|
import { Source, Layer } from "react-map-gl";
|
||||||
import _ from 'lodash'
|
import _ from "lodash";
|
||||||
import {Button, Form, Dropdown, Header, Message, Icon} from 'semantic-ui-react'
|
import { Button, Form, Dropdown, Header, Message, Icon } from "semantic-ui-react";
|
||||||
import {useTranslation, Trans as Translate} from 'react-i18next'
|
|
||||||
import Markdown from 'react-markdown'
|
|
||||||
|
|
||||||
import {useConfig} from 'config'
|
import { useConfig } from "config";
|
||||||
import {Page, Map} from 'components'
|
import { Page, Map } from "components";
|
||||||
|
|
||||||
const BoundingBoxSelector = React.forwardRef(({value, name, onChange}, ref) => {
|
const BoundingBoxSelector = React.forwardRef(
|
||||||
const {t} = useTranslation()
|
({ value, name, onChange }, ref) => {
|
||||||
const [pointNum, setPointNum] = useState(0)
|
const [pointNum, setPointNum] = useState(0);
|
||||||
const [point0, setPoint0] = useState(null)
|
const [point0, setPoint0] = useState(null);
|
||||||
const [point1, setPoint1] = useState(null)
|
const [point1, setPoint1] = useState(null);
|
||||||
|
|
||||||
const onClick = (e) => {
|
const onClick = (e) => {
|
||||||
if (pointNum == 0) {
|
if (pointNum == 0) {
|
||||||
setPoint0(e.lngLat)
|
setPoint0(e.lngLat);
|
||||||
} else {
|
} else {
|
||||||
setPoint1(e.lngLat)
|
setPoint1(e.lngLat);
|
||||||
}
|
}
|
||||||
setPointNum(1 - pointNum)
|
setPointNum(1 - pointNum);
|
||||||
}
|
};
|
||||||
|
|
||||||
React.useEffect(() => {
|
React.useEffect(() => {
|
||||||
if (!point0 || !point1) return
|
if (!point0 || !point1) return;
|
||||||
const bbox = `${point0[0]},${point0[1]},${point1[0]},${point1[1]}`
|
const bbox = `${point0[0]},${point0[1]},${point1[0]},${point1[1]}`;
|
||||||
if (bbox !== value) {
|
if (bbox !== value) {
|
||||||
onChange(bbox)
|
onChange(bbox);
|
||||||
}
|
}
|
||||||
}, [point0, point1])
|
}, [point0, point1]);
|
||||||
|
|
||||||
React.useEffect(() => {
|
React.useEffect(() => {
|
||||||
if (!value) return
|
const [p00, p01, p10, p11] = value
|
||||||
const [p00, p01, p10, p11] = value.split(',').map((v) => Number.parseFloat(v))
|
.split(",")
|
||||||
if (!point0 || point0[0] != p00 || point0[1] != p01) setPoint0([p00, p01])
|
.map((v) => Number.parseFloat(v));
|
||||||
if (!point1 || point1[0] != p10 || point1[1] != p11) setPoint1([p10, p11])
|
if (!point0 || point0[0] != p00 || point0[1] != p01)
|
||||||
}, [value])
|
setPoint0([p00, p01]);
|
||||||
|
if (!point1 || point1[0] != p10 || point1[1] != p11)
|
||||||
|
setPoint1([p10, p11]);
|
||||||
|
}, [value]);
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<div>
|
<div>
|
||||||
<Form.Input
|
<Form.Input label="Bounding box" {...{ name, value, onChange }} />
|
||||||
label={t('ExportPage.boundingBox.label')}
|
|
||||||
{...{name, value}}
|
|
||||||
onChange={(e) => onChange(e.target.value)}
|
|
||||||
/>
|
|
||||||
|
|
||||||
<div style={{height: 400, position: 'relative', marginBottom: 16}}>
|
<div style={{ height: 400, position: "relative", marginBottom: 16 }}>
|
||||||
<Map onClick={onClick}>
|
<Map onClick={onClick}>
|
||||||
<Source
|
<Source
|
||||||
id="bbox"
|
|
||||||
type="geojson"
|
|
||||||
data={
|
|
||||||
point0 && point1
|
|
||||||
? {
|
|
||||||
type: 'FeatureCollection',
|
|
||||||
features: [
|
|
||||||
{
|
|
||||||
type: 'Feature',
|
|
||||||
geometry: {
|
|
||||||
type: 'Polygon',
|
|
||||||
coordinates: [
|
|
||||||
[
|
|
||||||
[point0[0], point0[1]],
|
|
||||||
[point1[0], point0[1]],
|
|
||||||
[point1[0], point1[1]],
|
|
||||||
[point0[0], point1[1]],
|
|
||||||
[point0[0], point0[1]],
|
|
||||||
],
|
|
||||||
],
|
|
||||||
},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}
|
|
||||||
: {}
|
|
||||||
}
|
|
||||||
>
|
|
||||||
<Layer
|
|
||||||
id="bbox"
|
id="bbox"
|
||||||
type="line"
|
type="geojson"
|
||||||
paint={{
|
data={
|
||||||
'line-width': 4,
|
point0 && point1
|
||||||
'line-color': '#F06292',
|
? {
|
||||||
}}
|
type: "FeatureCollection",
|
||||||
/>
|
features: [
|
||||||
</Source>
|
{
|
||||||
</Map>
|
type: "Feature",
|
||||||
|
geometry: {
|
||||||
|
type: "Polygon",
|
||||||
|
coordinates: [
|
||||||
|
[
|
||||||
|
[point0[0], point0[1]],
|
||||||
|
[point1[0], point0[1]],
|
||||||
|
[point1[0], point1[1]],
|
||||||
|
[point0[0], point1[1]],
|
||||||
|
[point0[0], point0[1]],
|
||||||
|
],
|
||||||
|
],
|
||||||
|
},
|
||||||
|
},
|
||||||
|
],
|
||||||
|
}
|
||||||
|
: {}
|
||||||
|
}
|
||||||
|
>
|
||||||
|
<Layer
|
||||||
|
id="bbox"
|
||||||
|
type="line"
|
||||||
|
paint={{
|
||||||
|
"line-width": 4,
|
||||||
|
"line-color": "#F06292",
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
</Source>
|
||||||
|
</Map>
|
||||||
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
);
|
||||||
)
|
}
|
||||||
})
|
);
|
||||||
|
|
||||||
const MODES = ['events', 'segments']
|
const MODES = [
|
||||||
const FORMATS = ['geojson', 'shapefile']
|
{
|
||||||
|
key: "events",
|
||||||
|
text: "Events",
|
||||||
|
value: "events",
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
const FORMATS = [
|
||||||
|
{
|
||||||
|
key: "geojson",
|
||||||
|
text: "GeoJSON",
|
||||||
|
value: "geojson",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: "shapefile",
|
||||||
|
text: "Shapefile (ZIP)",
|
||||||
|
value: "shapefile",
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
export default function ExportPage() {
|
export default function ExportPage() {
|
||||||
const [mode, setMode] = useState('events')
|
const [mode, setMode] = useState("events");
|
||||||
const [bbox, setBbox] = useState('8.294678,49.651182,9.059601,50.108249')
|
const [bbox, setBbox] = useState("8.294678,49.651182,9.059601,50.108249");
|
||||||
const [fmt, setFmt] = useState('geojson')
|
const [fmt, setFmt] = useState("geojson");
|
||||||
const config = useConfig()
|
const config = useConfig();
|
||||||
const {t} = useTranslation()
|
const exportUrl = `${config?.apiUrl}/export/events?bbox=${bbox}&fmt=${fmt}`;
|
||||||
return (
|
return (
|
||||||
<Page title="Export">
|
<Page title="Export">
|
||||||
<Header as="h2">{t('ExportPage.title')}</Header>
|
<Header as="h2">Export</Header>
|
||||||
|
|
||||||
<Message icon info>
|
<Message icon info>
|
||||||
<Icon name="info circle" />
|
<Icon name="info circle" />
|
||||||
<Message.Content>
|
<Message.Content>
|
||||||
<Markdown>{t('ExportPage.information')}</Markdown>
|
<p>
|
||||||
</Message.Content>
|
This page allows you to export parts of the public dataset under
|
||||||
</Message>
|
the license for data announced in the privacy statement of this
|
||||||
|
site.
|
||||||
|
</p>
|
||||||
|
|
||||||
|
<p>
|
||||||
|
Please consider this export <b>experimental</b> and expect the data
|
||||||
|
shape to change in the future. Automated usage of this export
|
||||||
|
functionality is discouraged for now. Try not to use too much computing
|
||||||
|
capacity when exporting data, so select the bounding box as small as
|
||||||
|
possible and do not exceed unreasonable poll frequencies.
|
||||||
|
</p>
|
||||||
|
</Message.Content>
|
||||||
|
</Message>
|
||||||
|
|
||||||
<Form>
|
<Form>
|
||||||
<Form.Field>
|
<Form.Field>
|
||||||
<label>{t('ExportPage.mode.label')}</label>
|
<label>Mode</label>
|
||||||
<Dropdown
|
<Dropdown
|
||||||
placeholder={t('ExportPage.mode.placeholder')}
|
placeholder="Select mode"
|
||||||
fluid
|
fluid
|
||||||
selection
|
selection
|
||||||
options={MODES.map((value) => ({
|
options={MODES}
|
||||||
key: value,
|
|
||||||
text: t(`ExportPage.mode.${value}`),
|
|
||||||
value,
|
|
||||||
}))}
|
|
||||||
value={mode}
|
value={mode}
|
||||||
onChange={(_e, {value}) => setMode(value)}
|
onChange={(_e, { value }) => setMode(value)}
|
||||||
/>
|
/>
|
||||||
</Form.Field>
|
</Form.Field>
|
||||||
|
|
||||||
<Form.Field>
|
<Form.Field>
|
||||||
<label>{t('ExportPage.format.label')}</label>
|
<label>Format</label>
|
||||||
<Dropdown
|
<Dropdown
|
||||||
placeholder={t('ExportPage.format.placeholder')}
|
placeholder="Select format"
|
||||||
fluid
|
fluid
|
||||||
selection
|
selection
|
||||||
options={FORMATS.map((value) => ({
|
options={FORMATS}
|
||||||
key: value,
|
|
||||||
text: t(`ExportPage.format.${value}`),
|
|
||||||
value,
|
|
||||||
}))}
|
|
||||||
value={fmt}
|
value={fmt}
|
||||||
onChange={(_e, {value}) => setFmt(value)}
|
onChange={(_e, { value }) => setFmt(value)}
|
||||||
/>
|
/>
|
||||||
</Form.Field>
|
</Form.Field>
|
||||||
|
|
||||||
|
@ -149,13 +169,13 @@ export default function ExportPage() {
|
||||||
<Button
|
<Button
|
||||||
primary
|
primary
|
||||||
as="a"
|
as="a"
|
||||||
href={`${config?.apiUrl}/export/${mode}?bbox=${bbox}&fmt=${fmt}`}
|
href={exportUrl}
|
||||||
target="_blank"
|
target="_blank"
|
||||||
rel="noreferrer noopener"
|
rel="noreferrer noopener"
|
||||||
>
|
>
|
||||||
{t('ExportPage.export')}
|
Export
|
||||||
</Button>
|
</Button>
|
||||||
</Form>
|
</Form>
|
||||||
</Page>
|
</Page>
|
||||||
)
|
);
|
||||||
}
|
}
|
||||||
|
|
12
frontend/src/pages/HomePage.module.less
Normal file
12
frontend/src/pages/HomePage.module.less
Normal file
|
@ -0,0 +1,12 @@
|
||||||
|
@import 'styles.less';
|
||||||
|
|
||||||
|
.welcomeMap {
|
||||||
|
height: 60rem;
|
||||||
|
max-height: 70vh;
|
||||||
|
position: relative;
|
||||||
|
|
||||||
|
@media @mobile {
|
||||||
|
margin: -35px -32px 0 -32px;
|
||||||
|
max-height: 70vh;
|
||||||
|
}
|
||||||
|
}
|
|
@ -1,19 +1,17 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import {Grid, Loader, Header, Item} from 'semantic-ui-react'
|
import {Link} from 'react-router-dom'
|
||||||
|
import {Message, Grid, Loader, Header, Item} from 'semantic-ui-react'
|
||||||
import {useObservable} from 'rxjs-hooks'
|
import {useObservable} from 'rxjs-hooks'
|
||||||
import {of, from} from 'rxjs'
|
import {of, from} from 'rxjs'
|
||||||
import {map, switchMap} from 'rxjs/operators'
|
import {map, switchMap} from 'rxjs/operators'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
import {RegionStats, Stats, Page} from 'components'
|
import {Stats, Page, Map} from 'components'
|
||||||
import type {Track} from 'types'
|
|
||||||
|
|
||||||
import {TrackListItem, NoPublicTracksMessage} from './TracksPage'
|
import {TrackListItem} from './TracksPage'
|
||||||
|
import styles from './HomePage.module.less'
|
||||||
|
|
||||||
function MostRecentTrack() {
|
function MostRecentTrack() {
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
const track: Track | null = useObservable(
|
const track: Track | null = useObservable(
|
||||||
() =>
|
() =>
|
||||||
of(null).pipe(
|
of(null).pipe(
|
||||||
|
@ -26,10 +24,12 @@ function MostRecentTrack() {
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
<Header as="h2">{t('HomePage.mostRecentTrack')}</Header>
|
<Header as="h2">Most recent track</Header>
|
||||||
<Loader active={track === null} />
|
<Loader active={track === null} />
|
||||||
{track === undefined ? (
|
{track === undefined ? (
|
||||||
<NoPublicTracksMessage />
|
<Message>
|
||||||
|
No public tracks yet. <Link to="/upload">Upload the first!</Link>
|
||||||
|
</Message>
|
||||||
) : track ? (
|
) : track ? (
|
||||||
<Item.Group>
|
<Item.Group>
|
||||||
<TrackListItem track={track} />
|
<TrackListItem track={track} />
|
||||||
|
@ -46,10 +46,9 @@ export default function HomePage() {
|
||||||
<Grid.Row>
|
<Grid.Row>
|
||||||
<Grid.Column width={8}>
|
<Grid.Column width={8}>
|
||||||
<Stats />
|
<Stats />
|
||||||
<MostRecentTrack />
|
|
||||||
</Grid.Column>
|
</Grid.Column>
|
||||||
<Grid.Column width={8}>
|
<Grid.Column width={8}>
|
||||||
<RegionStats />
|
<MostRecentTrack />
|
||||||
</Grid.Column>
|
</Grid.Column>
|
||||||
</Grid.Row>
|
</Grid.Row>
|
||||||
</Grid>
|
</Grid>
|
||||||
|
|
|
@ -4,18 +4,16 @@ import {Redirect, useLocation, useHistory} from 'react-router-dom'
|
||||||
import {Icon, Message} from 'semantic-ui-react'
|
import {Icon, Message} from 'semantic-ui-react'
|
||||||
import {useObservable} from 'rxjs-hooks'
|
import {useObservable} from 'rxjs-hooks'
|
||||||
import {switchMap, pluck, distinctUntilChanged} from 'rxjs/operators'
|
import {switchMap, pluck, distinctUntilChanged} from 'rxjs/operators'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {Page} from 'components'
|
import {Page} from 'components'
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
|
|
||||||
const LoginRedirectPage = connect((state) => ({
|
const LoginRedirectPage = connect((state) => ({loggedIn: Boolean(state.login)}))(function LoginRedirectPage({
|
||||||
loggedIn: Boolean(state.login),
|
loggedIn,
|
||||||
}))(function LoginRedirectPage({loggedIn}) {
|
}) {
|
||||||
const location = useLocation()
|
const location = useLocation()
|
||||||
const history = useHistory()
|
const history = useHistory()
|
||||||
const {search} = location
|
const {search} = location
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
/* eslint-disable react-hooks/exhaustive-deps */
|
/* eslint-disable react-hooks/exhaustive-deps */
|
||||||
|
|
||||||
|
@ -37,8 +35,14 @@ const LoginRedirectPage = connect((state) => ({
|
||||||
|
|
||||||
if (error) {
|
if (error) {
|
||||||
return (
|
return (
|
||||||
<Page small title={t('LoginRedirectPage.loginError')}>
|
<Page small title="Login error">
|
||||||
<LoginError errorText={errorDescription || error} />
|
<Message icon error>
|
||||||
|
<Icon name="warning sign" />
|
||||||
|
<Message.Content>
|
||||||
|
<Message.Header>Login error</Message.Header>
|
||||||
|
The login server reported: {errorDescription || error}.
|
||||||
|
</Message.Content>
|
||||||
|
</Message>
|
||||||
</Page>
|
</Page>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
@ -46,21 +50,7 @@ const LoginRedirectPage = connect((state) => ({
|
||||||
return <ExchangeAuthCode code={code} />
|
return <ExchangeAuthCode code={code} />
|
||||||
})
|
})
|
||||||
|
|
||||||
function LoginError({errorText}: {errorText: string}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
return (
|
|
||||||
<Message icon error>
|
|
||||||
<Icon name="warning sign" />
|
|
||||||
<Message.Content>
|
|
||||||
<Message.Header>{t('LoginRedirectPage.loginError')}</Message.Header>
|
|
||||||
{t('LoginRedirectPage.loginErrorText', {error: errorText})}
|
|
||||||
</Message.Content>
|
|
||||||
</Message>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
function ExchangeAuthCode({code}) {
|
function ExchangeAuthCode({code}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const result = useObservable(
|
const result = useObservable(
|
||||||
(_$, args$) =>
|
(_$, args$) =>
|
||||||
args$.pipe(
|
args$.pipe(
|
||||||
|
@ -78,8 +68,8 @@ function ExchangeAuthCode({code}) {
|
||||||
<Message icon info>
|
<Message icon info>
|
||||||
<Icon name="circle notched" loading />
|
<Icon name="circle notched" loading />
|
||||||
<Message.Content>
|
<Message.Content>
|
||||||
<Message.Header>{t('LoginRedirectPage.loggingIn')}</Message.Header>
|
<Message.Header>Logging you in</Message.Header>
|
||||||
{t('LoginRedirectPage.hangTight')}
|
Hang tight...
|
||||||
</Message.Content>
|
</Message.Content>
|
||||||
</Message>
|
</Message>
|
||||||
)
|
)
|
||||||
|
@ -87,14 +77,21 @@ function ExchangeAuthCode({code}) {
|
||||||
content = <Redirect to="/" />
|
content = <Redirect to="/" />
|
||||||
} else {
|
} else {
|
||||||
const {error, error_description: errorDescription} = result
|
const {error, error_description: errorDescription} = result
|
||||||
content = <LoginError errorText={errorDescription || error} />
|
content = (
|
||||||
|
<>
|
||||||
|
<Message icon error>
|
||||||
|
<Icon name="warning sign" />
|
||||||
|
<Message.Content>
|
||||||
|
<Message.Header>Login error</Message.Header>
|
||||||
|
The login server reported: {errorDescription || error}.
|
||||||
|
</Message.Content>
|
||||||
|
</Message>
|
||||||
|
<pre>{JSON.stringify(result, null, 2)}</pre>
|
||||||
|
</>
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
return (
|
return <Page small title="Login">{content}</Page>
|
||||||
<Page small title="Login">
|
|
||||||
{content}
|
|
||||||
</Page>
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export default LoginRedirectPage
|
export default LoginRedirectPage
|
||||||
|
|
|
@ -1,9 +1,7 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import _ from 'lodash'
|
import _ from 'lodash'
|
||||||
import {connect} from 'react-redux'
|
import {connect} from 'react-redux'
|
||||||
import {Link} from 'react-router-dom'
|
import {List, Select, Input, Divider, Checkbox, Header} from 'semantic-ui-react'
|
||||||
import {List, Select, Input, Divider, Label, Checkbox, Header} from 'semantic-ui-react'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {
|
import {
|
||||||
MapConfig,
|
MapConfig,
|
||||||
|
@ -12,65 +10,46 @@ import {
|
||||||
} from 'reducers/mapConfig'
|
} from 'reducers/mapConfig'
|
||||||
import {colorByDistance, colorByCount, viridisSimpleHtml} from 'mapstyles'
|
import {colorByDistance, colorByCount, viridisSimpleHtml} from 'mapstyles'
|
||||||
import {ColorMapLegend, DiscreteColorMapLegend} from 'components'
|
import {ColorMapLegend, DiscreteColorMapLegend} from 'components'
|
||||||
import styles from './styles.module.less'
|
|
||||||
|
|
||||||
const BASEMAP_STYLE_OPTIONS = ['positron', 'bright']
|
const BASEMAP_STYLE_OPTIONS = [
|
||||||
|
{value: 'positron', key: 'positron', text: 'Positron'},
|
||||||
const ROAD_ATTRIBUTE_OPTIONS = [
|
{value: 'bright', key: 'bright', text: 'OSM Bright'},
|
||||||
'distance_overtaker_mean',
|
|
||||||
'distance_overtaker_min',
|
|
||||||
'distance_overtaker_max',
|
|
||||||
'distance_overtaker_median',
|
|
||||||
'overtaking_event_count',
|
|
||||||
'usage_count',
|
|
||||||
'zone',
|
|
||||||
]
|
]
|
||||||
|
|
||||||
const DATE_FILTER_MODES = ['none', 'range', 'threshold']
|
const ROAD_ATTRIBUTE_OPTIONS = [
|
||||||
|
{value: 'distance_overtaker_mean', key: 'distance_overtaker_mean', text: 'Overtaker distance mean'},
|
||||||
type User = Object
|
{value: 'distance_overtaker_min', key: 'distance_overtaker_min', text: 'Overtaker distance minimum'},
|
||||||
|
{value: 'distance_overtaker_max', key: 'distance_overtaker_max', text: 'Overtaker distance maximum'},
|
||||||
|
{value: 'distance_overtaker_median', key: 'distance_overtaker_median', text: 'Overtaker distance median'},
|
||||||
|
{value: 'overtaking_event_count', key: 'overtaking_event_count', text: 'Event count'},
|
||||||
|
{value: 'usage_count', key: 'usage_count', text: 'Usage count'},
|
||||||
|
]
|
||||||
|
|
||||||
function LayerSidebar({
|
function LayerSidebar({
|
||||||
mapConfig,
|
mapConfig,
|
||||||
login,
|
|
||||||
setMapConfigFlag,
|
setMapConfigFlag,
|
||||||
}: {
|
}: {
|
||||||
login: User | null
|
|
||||||
mapConfig: MapConfig
|
mapConfig: MapConfig
|
||||||
setMapConfigFlag: (flag: string, value: unknown) => void
|
setMapConfigFlag: (flag: string, value: unknown) => void
|
||||||
}) {
|
}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const {
|
const {
|
||||||
baseMap: {style},
|
baseMap: {style},
|
||||||
obsRoads: {show: showRoads, showUntagged, attribute, maxCount},
|
obsRoads: {show: showRoads, showUntagged, attribute, maxCount},
|
||||||
obsEvents: {show: showEvents},
|
obsEvents: {show: showEvents},
|
||||||
obsRegions: {show: showRegions},
|
obsRegions: {show: showRegions},
|
||||||
filters: {currentUser: filtersCurrentUser, dateMode, startDate, endDate, thresholdAfter},
|
|
||||||
} = mapConfig
|
} = mapConfig
|
||||||
|
|
||||||
const openStreetMapCopyright = (
|
|
||||||
<List.Item className={styles.copyright}>
|
|
||||||
{t('MapPage.sidebar.copyright.openStreetMap')}{' '}
|
|
||||||
<Link to="/acknowledgements">{t('MapPage.sidebar.copyright.learnMore')}</Link>
|
|
||||||
</List.Item>
|
|
||||||
)
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<div>
|
<div>
|
||||||
<List relaxed>
|
<List relaxed>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>{t('MapPage.sidebar.baseMap.style.label')}</List.Header>
|
<List.Header>Basemap Style</List.Header>
|
||||||
<Select
|
<Select
|
||||||
options={BASEMAP_STYLE_OPTIONS.map((value) => ({
|
options={BASEMAP_STYLE_OPTIONS}
|
||||||
value,
|
|
||||||
key: value,
|
|
||||||
text: t(`MapPage.sidebar.baseMap.style.${value}`),
|
|
||||||
}))}
|
|
||||||
value={style}
|
value={style}
|
||||||
onChange={(_e, {value}) => setMapConfigFlag('baseMap.style', value)}
|
onChange={(_e, {value}) => setMapConfigFlag('baseMap.style', value)}
|
||||||
/>
|
/>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
{openStreetMapCopyright}
|
|
||||||
<Divider />
|
<Divider />
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<Checkbox
|
<Checkbox
|
||||||
|
@ -82,25 +61,14 @@ function LayerSidebar({
|
||||||
onChange={() => setMapConfigFlag('obsRegions.show', !showRegions)}
|
onChange={() => setMapConfigFlag('obsRegions.show', !showRegions)}
|
||||||
/>
|
/>
|
||||||
<label htmlFor="obsRegions.show">
|
<label htmlFor="obsRegions.show">
|
||||||
<Header as="h4">{t('MapPage.sidebar.obsRegions.title')}</Header>
|
<Header as="h4">Regions</Header>
|
||||||
</label>
|
</label>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
{showRegions && (
|
{showRegions && (
|
||||||
<>
|
<>
|
||||||
<List.Item>{t('MapPage.sidebar.obsRegions.colorByEventCount')}</List.Item>
|
<List.Item>Color regions based on event count</List.Item>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<ColorMapLegend
|
<ColorMapLegend twoTicks map={[[0, "#00897B00"], [5000, "#00897BFF"]]} digits={0} />
|
||||||
twoTicks
|
|
||||||
map={[
|
|
||||||
[0, '#00897B00'],
|
|
||||||
[5000, '#00897BFF'],
|
|
||||||
]}
|
|
||||||
digits={0}
|
|
||||||
/>
|
|
||||||
</List.Item>
|
|
||||||
<List.Item className={styles.copyright}>
|
|
||||||
{t('MapPage.sidebar.copyright.boundaries')}{' '}
|
|
||||||
<Link to="/acknowledgements">{t('MapPage.sidebar.copyright.learnMore')}</Link>
|
|
||||||
</List.Item>
|
</List.Item>
|
||||||
</>
|
</>
|
||||||
)}
|
)}
|
||||||
|
@ -115,7 +83,7 @@ function LayerSidebar({
|
||||||
onChange={() => setMapConfigFlag('obsRoads.show', !showRoads)}
|
onChange={() => setMapConfigFlag('obsRoads.show', !showRoads)}
|
||||||
/>
|
/>
|
||||||
<label htmlFor="obsRoads.show">
|
<label htmlFor="obsRoads.show">
|
||||||
<Header as="h4">{t('MapPage.sidebar.obsRoads.title')}</Header>
|
<Header as="h4">Road segments</Header>
|
||||||
</label>
|
</label>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
{showRoads && (
|
{showRoads && (
|
||||||
|
@ -124,67 +92,38 @@ function LayerSidebar({
|
||||||
<Checkbox
|
<Checkbox
|
||||||
checked={showUntagged}
|
checked={showUntagged}
|
||||||
onChange={() => setMapConfigFlag('obsRoads.showUntagged', !showUntagged)}
|
onChange={() => setMapConfigFlag('obsRoads.showUntagged', !showUntagged)}
|
||||||
label={t('MapPage.sidebar.obsRoads.showUntagged.label')}
|
label="Include roads without data"
|
||||||
/>
|
/>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>{t('MapPage.sidebar.obsRoads.attribute.label')}</List.Header>
|
<List.Header>Color based on</List.Header>
|
||||||
<Select
|
<Select
|
||||||
fluid
|
fluid
|
||||||
options={ROAD_ATTRIBUTE_OPTIONS.map((value) => ({
|
options={ROAD_ATTRIBUTE_OPTIONS}
|
||||||
value,
|
|
||||||
key: value,
|
|
||||||
text: t(`MapPage.sidebar.obsRoads.attribute.${value}`),
|
|
||||||
}))}
|
|
||||||
value={attribute}
|
value={attribute}
|
||||||
onChange={(_e, {value}) => setMapConfigFlag('obsRoads.attribute', value)}
|
onChange={(_e, {value}) => setMapConfigFlag('obsRoads.attribute', value)}
|
||||||
/>
|
/>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
{attribute.endsWith('_count') ? (
|
{attribute.endsWith('_count') ? (
|
||||||
<>
|
<>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>{t('MapPage.sidebar.obsRoads.maxCount.label')}</List.Header>
|
<List.Header>Maximum value</List.Header>
|
||||||
<Input
|
<Input
|
||||||
fluid
|
fluid
|
||||||
type="number"
|
type="number"
|
||||||
value={maxCount}
|
value={maxCount}
|
||||||
onChange={(_e, {value}) => setMapConfigFlag('obsRoads.maxCount', value)}
|
onChange={(_e, {value}) => setMapConfigFlag('obsRoads.maxCount', value)}
|
||||||
/>
|
/>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<ColorMapLegend
|
<ColorMapLegend map={_.chunk(colorByCount('obsRoads.maxCount', mapConfig.obsRoads.maxCount, viridisSimpleHtml ).slice(3), 2)} twoTicks />
|
||||||
map={_.chunk(
|
</List.Item></>
|
||||||
colorByCount('obsRoads.maxCount', mapConfig.obsRoads.maxCount, viridisSimpleHtml).slice(3),
|
) :
|
||||||
2
|
(
|
||||||
)}
|
<List.Item>
|
||||||
twoTicks
|
<DiscreteColorMapLegend map={colorByDistance('distance_overtaker')[3].slice(2)} />
|
||||||
/>
|
</List.Item>
|
||||||
</List.Item>
|
|
||||||
</>
|
|
||||||
) : attribute.endsWith('zone') ? (
|
|
||||||
<>
|
|
||||||
<List.Item>
|
|
||||||
<Label size="small" style={{background: 'blue', color: 'white'}}>
|
|
||||||
{t('general.zone.urban')} (1.5 m)
|
|
||||||
</Label>
|
|
||||||
<Label size="small" style={{background: 'cyan', color: 'black'}}>
|
|
||||||
{t('general.zone.rural')}(2 m)
|
|
||||||
</Label>
|
|
||||||
</List.Item>
|
|
||||||
</>
|
|
||||||
) : (
|
|
||||||
<>
|
|
||||||
<List.Item>
|
|
||||||
<List.Header>{_.upperFirst(t('general.zone.urban'))}</List.Header>
|
|
||||||
<DiscreteColorMapLegend map={colorByDistance('distance_overtaker')[3][5].slice(2)} />
|
|
||||||
</List.Item>
|
|
||||||
<List.Item>
|
|
||||||
<List.Header>{_.upperFirst(t('general.zone.rural'))}</List.Header>
|
|
||||||
<DiscreteColorMapLegend map={colorByDistance('distance_overtaker')[3][3].slice(2)} />
|
|
||||||
</List.Item>
|
|
||||||
</>
|
|
||||||
)}
|
)}
|
||||||
{openStreetMapCopyright}
|
|
||||||
</>
|
</>
|
||||||
)}
|
)}
|
||||||
<Divider />
|
<Divider />
|
||||||
|
@ -198,124 +137,16 @@ function LayerSidebar({
|
||||||
onChange={() => setMapConfigFlag('obsEvents.show', !showEvents)}
|
onChange={() => setMapConfigFlag('obsEvents.show', !showEvents)}
|
||||||
/>
|
/>
|
||||||
<label htmlFor="obsEvents.show">
|
<label htmlFor="obsEvents.show">
|
||||||
<Header as="h4">{t('MapPage.sidebar.obsEvents.title')}</Header>
|
<Header as="h4">Event points</Header>
|
||||||
</label>
|
</label>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
{showEvents && (
|
{showEvents && (
|
||||||
<>
|
<>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>{_.upperFirst(t('general.zone.urban'))}</List.Header>
|
<DiscreteColorMapLegend map={colorByDistance('distance_overtaker')[3].slice(2)} />
|
||||||
<DiscreteColorMapLegend map={colorByDistance('distance_overtaker')[3][5].slice(2)} />
|
|
||||||
</List.Item>
|
|
||||||
<List.Item>
|
|
||||||
<List.Header>{_.upperFirst(t('general.zone.rural'))}</List.Header>
|
|
||||||
<DiscreteColorMapLegend map={colorByDistance('distance_overtaker')[3][3].slice(2)} />
|
|
||||||
</List.Item>
|
</List.Item>
|
||||||
</>
|
</>
|
||||||
)}
|
)}
|
||||||
<Divider />
|
|
||||||
|
|
||||||
<List.Item>
|
|
||||||
<Header as="h4">{t('MapPage.sidebar.filters.title')}</Header>
|
|
||||||
</List.Item>
|
|
||||||
|
|
||||||
{login && (
|
|
||||||
<>
|
|
||||||
<List.Item>
|
|
||||||
<Header as="h5">{t('MapPage.sidebar.filters.userData')}</Header>
|
|
||||||
</List.Item>
|
|
||||||
|
|
||||||
<List.Item>
|
|
||||||
<Checkbox
|
|
||||||
toggle
|
|
||||||
size="small"
|
|
||||||
id="filters.currentUser"
|
|
||||||
checked={filtersCurrentUser}
|
|
||||||
onChange={() => setMapConfigFlag('filters.currentUser', !filtersCurrentUser)}
|
|
||||||
label={t('MapPage.sidebar.filters.currentUser')}
|
|
||||||
/>
|
|
||||||
</List.Item>
|
|
||||||
|
|
||||||
<List.Item>
|
|
||||||
<Header as="h5">{t('MapPage.sidebar.filters.dateRange')}</Header>
|
|
||||||
</List.Item>
|
|
||||||
|
|
||||||
<List.Item>
|
|
||||||
<Select
|
|
||||||
id="filters.dateMode"
|
|
||||||
options={DATE_FILTER_MODES.map((value) => ({
|
|
||||||
value,
|
|
||||||
key: value,
|
|
||||||
text: t(`MapPage.sidebar.filters.dateMode.${value}`),
|
|
||||||
}))}
|
|
||||||
value={dateMode ?? 'none'}
|
|
||||||
onChange={(_e, {value}) => setMapConfigFlag('filters.dateMode', value)}
|
|
||||||
/>
|
|
||||||
</List.Item>
|
|
||||||
|
|
||||||
{dateMode == 'range' && (
|
|
||||||
<List.Item>
|
|
||||||
<Input
|
|
||||||
type="date"
|
|
||||||
min="2000-01-03"
|
|
||||||
step="7"
|
|
||||||
size="small"
|
|
||||||
id="filters.startDate"
|
|
||||||
onChange={(_e, {value}) => setMapConfigFlag('filters.startDate', value)}
|
|
||||||
value={startDate ?? null}
|
|
||||||
label={t('MapPage.sidebar.filters.start')}
|
|
||||||
/>
|
|
||||||
</List.Item>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{dateMode == 'range' && (
|
|
||||||
<List.Item>
|
|
||||||
<Input
|
|
||||||
type="date"
|
|
||||||
min="2000-01-03"
|
|
||||||
step="7"
|
|
||||||
size="small"
|
|
||||||
id="filters.endDate"
|
|
||||||
onChange={(_e, {value}) => setMapConfigFlag('filters.endDate', value)}
|
|
||||||
value={endDate ?? null}
|
|
||||||
label={t('MapPage.sidebar.filters.end')}
|
|
||||||
/>
|
|
||||||
</List.Item>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{dateMode == 'threshold' && (
|
|
||||||
<List.Item>
|
|
||||||
<Input
|
|
||||||
type="date"
|
|
||||||
min="2000-01-03"
|
|
||||||
step="7"
|
|
||||||
size="small"
|
|
||||||
id="filters.startDate"
|
|
||||||
value={startDate ?? null}
|
|
||||||
onChange={(_e, {value}) => setMapConfigFlag('filters.startDate', value)}
|
|
||||||
label={t('MapPage.sidebar.filters.threshold')}
|
|
||||||
/>
|
|
||||||
</List.Item>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{dateMode == 'threshold' && (
|
|
||||||
<List.Item>
|
|
||||||
<span>
|
|
||||||
{t('MapPage.sidebar.filters.before')}{' '}
|
|
||||||
<Checkbox
|
|
||||||
toggle
|
|
||||||
size="small"
|
|
||||||
checked={thresholdAfter ?? false}
|
|
||||||
onChange={() => setMapConfigFlag('filters.thresholdAfter', !thresholdAfter)}
|
|
||||||
id="filters.thresholdAfter"
|
|
||||||
/>{' '}
|
|
||||||
{t('MapPage.sidebar.filters.after')}
|
|
||||||
</span>
|
|
||||||
</List.Item>
|
|
||||||
)}
|
|
||||||
</>
|
|
||||||
)}
|
|
||||||
{!login && <List.Item>{t('MapPage.sidebar.filters.needsLogin')}</List.Item>}
|
|
||||||
</List>
|
</List>
|
||||||
</div>
|
</div>
|
||||||
)
|
)
|
||||||
|
@ -329,7 +160,6 @@ export default connect(
|
||||||
(state as any).mapConfig as MapConfig
|
(state as any).mapConfig as MapConfig
|
||||||
//
|
//
|
||||||
),
|
),
|
||||||
login: state.login,
|
|
||||||
}),
|
}),
|
||||||
{setMapConfigFlag: setMapConfigFlagAction}
|
{setMapConfigFlag: setMapConfigFlagAction}
|
||||||
//
|
//
|
||||||
|
|
|
@ -1,31 +0,0 @@
|
||||||
import React from 'react'
|
|
||||||
import {createPortal} from 'react-dom'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
import {List, Header, Icon, Button} from 'semantic-ui-react'
|
|
||||||
|
|
||||||
import styles from './styles.module.less'
|
|
||||||
|
|
||||||
export default function RegionInfo({region, mapInfoPortal, onClose}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const content = (
|
|
||||||
<>
|
|
||||||
<div className={styles.closeHeader}>
|
|
||||||
<Header as="h3">{region.properties.name || t('MapPage.regionInfo.unnamedRegion')}</Header>
|
|
||||||
<Button primary icon onClick={onClose}>
|
|
||||||
<Icon name="close" />
|
|
||||||
</Button>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<List>
|
|
||||||
<List.Item>
|
|
||||||
<List.Header>{t('MapPage.regionInfo.eventCount')}</List.Header>
|
|
||||||
<List.Content>{region.properties.overtaking_event_count ?? 0}</List.Content>
|
|
||||||
</List.Item>
|
|
||||||
</List>
|
|
||||||
</>
|
|
||||||
)
|
|
||||||
|
|
||||||
return content && mapInfoPortal
|
|
||||||
? createPortal(<div className={styles.mapInfoBox}>{content}</div>, mapInfoPortal)
|
|
||||||
: null
|
|
||||||
}
|
|
|
@ -1,75 +1,48 @@
|
||||||
import React, {useState, useCallback} from 'react'
|
import React, {useState, useCallback} from 'react'
|
||||||
import {createPortal} from 'react-dom'
|
|
||||||
import _ from 'lodash'
|
import _ from 'lodash'
|
||||||
import {Segment, Menu, Header, Label, Icon, Table, Message, Button} from 'semantic-ui-react'
|
import {Segment, Menu, Header, Label, Icon, Table} from 'semantic-ui-react'
|
||||||
import {Layer, Source} from 'react-map-gl'
|
import {Layer, Source} from 'react-map-gl'
|
||||||
import {of, from, concat} from 'rxjs'
|
import {of, from, concat} from 'rxjs'
|
||||||
import {useObservable} from 'rxjs-hooks'
|
import {useObservable} from 'rxjs-hooks'
|
||||||
import {switchMap, distinctUntilChanged} from 'rxjs/operators'
|
import {switchMap, distinctUntilChanged} from 'rxjs/operators'
|
||||||
import {Chart} from 'components'
|
|
||||||
import {pairwise} from 'utils'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import type {Location} from 'types'
|
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
import {colorByDistance, borderByZone} from 'mapstyles'
|
|
||||||
|
|
||||||
import styles from './styles.module.less'
|
import styles from './styles.module.less'
|
||||||
|
|
||||||
function selectFromColorMap(colormap, value) {
|
const UNITS = {distanceOvertaker: 'm', distanceStationary: 'm', speed: 'km/h'}
|
||||||
let last = null
|
const LABELS = {distanceOvertaker: 'Overtaker', distanceStationary: 'Stationary', speed: 'Speed'}
|
||||||
for (let i = 0; i < colormap.length; i += 2) {
|
const ZONE_COLORS = {urban: 'olive', rural: 'brown', motorway: 'purple'}
|
||||||
if (colormap[i + 1] > value) {
|
const CARDINAL_DIRECTIONS = ['north', 'north-east', 'east', 'south-east', 'south', 'south-west', 'west', 'north-west']
|
||||||
return colormap[i]
|
const getCardinalDirection = (bearing) =>
|
||||||
}
|
bearing == null
|
||||||
}
|
? 'unknown'
|
||||||
return colormap[colormap.length - 1]
|
: CARDINAL_DIRECTIONS[
|
||||||
}
|
Math.floor(((bearing / 360.0) * CARDINAL_DIRECTIONS.length + 0.5) % CARDINAL_DIRECTIONS.length)
|
||||||
|
] + ' bound'
|
||||||
const UNITS = {
|
|
||||||
distanceOvertaker: 'm',
|
|
||||||
distanceStationary: 'm',
|
|
||||||
speed: 'km/h',
|
|
||||||
}
|
|
||||||
const ZONE_COLORS = {urban: 'blue', rural: 'cyan', motorway: 'purple'}
|
|
||||||
const CARDINAL_DIRECTIONS = ['north', 'northEast', 'east', 'southEast', 'south', 'southWest', 'west', 'northWest']
|
|
||||||
const getCardinalDirection = (t, bearing) => {
|
|
||||||
if (bearing == null) {
|
|
||||||
return t('MapPage.roadInfo.cardinalDirections.unknown')
|
|
||||||
} else {
|
|
||||||
const n = CARDINAL_DIRECTIONS.length
|
|
||||||
const i = Math.floor(((bearing / 360.0) * n + 0.5) % n)
|
|
||||||
const name = CARDINAL_DIRECTIONS[i]
|
|
||||||
return t(`MapPage.roadInfo.cardinalDirections.${name}`)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
function RoadStatsTable({data}) {
|
function RoadStatsTable({data}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
return (
|
return (
|
||||||
<Table size="small" compact>
|
<Table size="small" compact>
|
||||||
<Table.Header>
|
<Table.Header>
|
||||||
<Table.Row>
|
<Table.Row>
|
||||||
<Table.HeaderCell textAlign="right"></Table.HeaderCell>
|
<Table.HeaderCell>Property</Table.HeaderCell>
|
||||||
{['distanceOvertaker', 'distanceStationary', 'speed'].map((prop) => (
|
<Table.HeaderCell>n</Table.HeaderCell>
|
||||||
<Table.HeaderCell key={prop} textAlign="right">
|
<Table.HeaderCell>min</Table.HeaderCell>
|
||||||
{t(`MapPage.roadInfo.${prop}`)}
|
<Table.HeaderCell>q50</Table.HeaderCell>
|
||||||
</Table.HeaderCell>
|
<Table.HeaderCell>max</Table.HeaderCell>
|
||||||
))}
|
<Table.HeaderCell>mean</Table.HeaderCell>
|
||||||
|
<Table.HeaderCell>unit</Table.HeaderCell>
|
||||||
</Table.Row>
|
</Table.Row>
|
||||||
</Table.Header>
|
</Table.Header>
|
||||||
<Table.Body>
|
<Table.Body>
|
||||||
{['count', 'min', 'median', 'max', 'mean'].map((stat) => (
|
{['distanceOvertaker', 'distanceStationary', 'speed'].map((prop) => (
|
||||||
<Table.Row key={stat}>
|
<Table.Row key={prop}>
|
||||||
<Table.Cell> {t(`MapPage.roadInfo.${stat}`)}</Table.Cell>
|
<Table.Cell>{LABELS[prop]}</Table.Cell>
|
||||||
{['distanceOvertaker', 'distanceStationary', 'speed'].map((prop) => (
|
{['count', 'min', 'median', 'max', 'mean'].map((stat) => (
|
||||||
<Table.Cell key={prop} textAlign="right">
|
<Table.Cell key={stat}>{((data[prop]?.statistics?.[stat]) * ((prop === `speed` && stat != 'count') ? 3.6 : 1)).toFixed(stat === 'count' ? 0 : 2)}</Table.Cell>
|
||||||
{(data[prop]?.statistics?.[stat] * (prop === `speed` && stat != 'count' ? 3.6 : 1)).toFixed(
|
|
||||||
stat === 'count' ? 0 : 2
|
|
||||||
)}
|
|
||||||
{stat !== 'count' && ` ${UNITS[prop]}`}
|
|
||||||
</Table.Cell>
|
|
||||||
))}
|
))}
|
||||||
|
<Table.Cell>{UNITS[prop]}</Table.Cell>
|
||||||
</Table.Row>
|
</Table.Row>
|
||||||
))}
|
))}
|
||||||
</Table.Body>
|
</Table.Body>
|
||||||
|
@ -77,91 +50,7 @@ function RoadStatsTable({data}) {
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
function HistogramChart({bins, counts, zone}) {
|
export default function RoadInfo({clickLocation}) {
|
||||||
const diff = bins[1] - bins[0]
|
|
||||||
const colortype = zone === 'rural' ? 3 : 5
|
|
||||||
const data = _.zip(
|
|
||||||
bins.slice(0, bins.length - 1).map((v) => v + diff / 2),
|
|
||||||
counts
|
|
||||||
).map((value) => ({
|
|
||||||
value,
|
|
||||||
itemStyle: {
|
|
||||||
color: selectFromColorMap(colorByDistance()[3][colortype].slice(2), value[0]),
|
|
||||||
},
|
|
||||||
}))
|
|
||||||
|
|
||||||
return (
|
|
||||||
<Chart
|
|
||||||
style={{height: 240}}
|
|
||||||
option={{
|
|
||||||
grid: {top: 30, bottom: 30, right: 30, left: 30},
|
|
||||||
xAxis: {
|
|
||||||
type: 'value',
|
|
||||||
axisLabel: {formatter: (v) => `${Math.round(v * 100)} cm`},
|
|
||||||
min: 0,
|
|
||||||
max: 2.5,
|
|
||||||
},
|
|
||||||
yAxis: {},
|
|
||||||
series: [
|
|
||||||
{
|
|
||||||
type: 'bar',
|
|
||||||
data,
|
|
||||||
|
|
||||||
barMaxWidth: 20,
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}}
|
|
||||||
/>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
interface ArrayStats {
|
|
||||||
statistics: {
|
|
||||||
count: number
|
|
||||||
mean: number
|
|
||||||
min: number
|
|
||||||
max: number
|
|
||||||
median: number
|
|
||||||
}
|
|
||||||
histogram: {
|
|
||||||
bins: number[]
|
|
||||||
counts: number[]
|
|
||||||
}
|
|
||||||
values: number[]
|
|
||||||
}
|
|
||||||
|
|
||||||
export interface RoadDirectionInfo {
|
|
||||||
bearing: number
|
|
||||||
distanceOvertaker: ArrayStats
|
|
||||||
distanceStationary: ArrayStats
|
|
||||||
speed: ArrayStats
|
|
||||||
}
|
|
||||||
|
|
||||||
export interface RoadInfoType {
|
|
||||||
road: {
|
|
||||||
way_id: number
|
|
||||||
zone: 'urban' | 'rural' | null
|
|
||||||
name: string
|
|
||||||
directionality: -1 | 0 | 1
|
|
||||||
oneway: boolean
|
|
||||||
geometry: Object
|
|
||||||
}
|
|
||||||
forwards: RoadDirectionInfo
|
|
||||||
backwards: RoadDirectionInfo
|
|
||||||
}
|
|
||||||
|
|
||||||
export default function RoadInfo({
|
|
||||||
roadInfo: info,
|
|
||||||
hasFilters,
|
|
||||||
onClose,
|
|
||||||
mapInfoPortal,
|
|
||||||
}: {
|
|
||||||
roadInfo: RoadInfoType
|
|
||||||
hasFilters: boolean
|
|
||||||
onClose: () => void
|
|
||||||
mapInfoPortal: HTMLElement
|
|
||||||
}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const [direction, setDirection] = useState('forwards')
|
const [direction, setDirection] = useState('forwards')
|
||||||
|
|
||||||
const onClickDirection = useCallback(
|
const onClickDirection = useCallback(
|
||||||
|
@ -173,60 +62,73 @@ export default function RoadInfo({
|
||||||
[setDirection]
|
[setDirection]
|
||||||
)
|
)
|
||||||
|
|
||||||
// TODO: change based on left-hand/right-hand traffic
|
const info = useObservable(
|
||||||
const offsetDirection = info.road.oneway ? 0 : direction === 'forwards' ? 1 : -1
|
(_$, inputs$) =>
|
||||||
|
inputs$.pipe(
|
||||||
const content = (
|
distinctUntilChanged(_.isEqual),
|
||||||
<>
|
switchMap(([location]) =>
|
||||||
<div className={styles.closeHeader}>
|
location
|
||||||
<Header as="h3">{info?.road.name || t('MapPage.roadInfo.unnamedWay')}</Header>
|
? concat(
|
||||||
<Button primary icon onClick={onClose}>
|
of(null),
|
||||||
<Icon name="close" />
|
from(
|
||||||
</Button>
|
api.get('/mapdetails/road', {
|
||||||
</div>
|
query: {
|
||||||
|
...location,
|
||||||
{hasFilters && (
|
radius: 100,
|
||||||
<Message info icon>
|
},
|
||||||
<Icon name="info circle" small />
|
})
|
||||||
<Message.Content>{t('MapPage.roadInfo.hintFiltersNotApplied')}</Message.Content>
|
)
|
||||||
</Message>
|
)
|
||||||
)}
|
: of(null)
|
||||||
|
)
|
||||||
{info?.road.zone && (
|
),
|
||||||
<Label size="small" color={ZONE_COLORS[info?.road.zone]}>
|
null,
|
||||||
{t(`general.zone.${info.road.zone}`)}
|
[clickLocation]
|
||||||
</Label>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{info?.road.oneway && (
|
|
||||||
<Label size="small" color="blue">
|
|
||||||
<Icon name="long arrow alternate right" fitted /> {t('MapPage.roadInfo.oneway')}
|
|
||||||
</Label>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{info?.road.oneway ? null : (
|
|
||||||
<Menu size="tiny" pointing>
|
|
||||||
<Menu.Item header>{t('MapPage.roadInfo.direction')}</Menu.Item>
|
|
||||||
<Menu.Item name="forwards" active={direction === 'forwards'} onClick={onClickDirection}>
|
|
||||||
{getCardinalDirection(t, info?.forwards?.bearing)}
|
|
||||||
</Menu.Item>
|
|
||||||
<Menu.Item name="backwards" active={direction === 'backwards'} onClick={onClickDirection}>
|
|
||||||
{getCardinalDirection(t, info?.backwards?.bearing)}
|
|
||||||
</Menu.Item>
|
|
||||||
</Menu>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{info?.[direction] && <RoadStatsTable data={info[direction]} />}
|
|
||||||
|
|
||||||
{info?.[direction]?.distanceOvertaker?.histogram && (
|
|
||||||
<>
|
|
||||||
<Header as="h5">{t('MapPage.roadInfo.overtakerDistanceDistribution')}</Header>
|
|
||||||
<HistogramChart {...info[direction]?.distanceOvertaker?.histogram} />
|
|
||||||
</>
|
|
||||||
)}
|
|
||||||
</>
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if (!clickLocation) {
|
||||||
|
return null
|
||||||
|
}
|
||||||
|
|
||||||
|
const loading = info == null
|
||||||
|
|
||||||
|
const offsetDirection = info?.road?.oneway ? 0 : direction === 'forwards' ? 1 : -1 // TODO: change based on left-hand/right-hand traffic
|
||||||
|
|
||||||
|
const content =
|
||||||
|
!loading && !info.road ? (
|
||||||
|
'No road found.'
|
||||||
|
) : (
|
||||||
|
<>
|
||||||
|
<Header as="h3">{loading ? '...' : info?.road.name || 'Unnamed way'}</Header>
|
||||||
|
|
||||||
|
{info?.road.zone && (
|
||||||
|
<Label size="small" color={ZONE_COLORS[info?.road.zone]}>
|
||||||
|
{info?.road.zone}
|
||||||
|
</Label>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{info?.road.oneway && (
|
||||||
|
<Label size="small" color="blue">
|
||||||
|
<Icon name="long arrow alternate right" fitted /> oneway
|
||||||
|
</Label>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{info?.road.oneway ? null : (
|
||||||
|
<Menu size="tiny" fluid secondary>
|
||||||
|
<Menu.Item header>Direction</Menu.Item>
|
||||||
|
<Menu.Item name="forwards" active={direction === 'forwards'} onClick={onClickDirection}>
|
||||||
|
{getCardinalDirection(info?.forwards?.bearing)}
|
||||||
|
</Menu.Item>
|
||||||
|
<Menu.Item name="backwards" active={direction === 'backwards'} onClick={onClickDirection}>
|
||||||
|
{getCardinalDirection(info?.backwards?.bearing)}
|
||||||
|
</Menu.Item>
|
||||||
|
</Menu>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{info?.[direction] && <RoadStatsTable data={info[direction]} />}
|
||||||
|
</>
|
||||||
|
)
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
{info?.road && (
|
{info?.road && (
|
||||||
|
@ -254,7 +156,11 @@ export default function RoadInfo({
|
||||||
</Source>
|
</Source>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
{content && mapInfoPortal && createPortal(<div className={styles.mapInfoBox}>{content}</div>, mapInfoPortal)}
|
{content && (
|
||||||
|
<div className={styles.mapInfoBox}>
|
||||||
|
<Segment loading={loading}>{content}</Segment>
|
||||||
|
</div>
|
||||||
|
)}
|
||||||
</>
|
</>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,20 +1,15 @@
|
||||||
import React, {useState, useCallback, useMemo, useRef} from 'react'
|
import React, {useState, useCallback, useMemo} from 'react'
|
||||||
import _ from 'lodash'
|
import _ from 'lodash'
|
||||||
import {connect} from 'react-redux'
|
|
||||||
import {Button} from 'semantic-ui-react'
|
import {Button} from 'semantic-ui-react'
|
||||||
import {Layer, Source} from 'react-map-gl'
|
import {Layer, Source} from 'react-map-gl'
|
||||||
import produce from 'immer'
|
import produce from 'immer'
|
||||||
import classNames from 'classnames'
|
|
||||||
|
|
||||||
import api from 'api'
|
|
||||||
import type {Location} from 'types'
|
|
||||||
import {Page, Map} from 'components'
|
import {Page, Map} from 'components'
|
||||||
import {useConfig} from 'config'
|
import {useConfig} from 'config'
|
||||||
import {colorByDistance, colorByCount, getRegionLayers, borderByZone, isValidAttribute} from 'mapstyles'
|
import {colorByDistance, colorByCount, getRegionLayers} from 'mapstyles'
|
||||||
import {useMapConfig} from 'reducers/mapConfig'
|
import {useMapConfig} from 'reducers/mapConfig'
|
||||||
|
|
||||||
import RoadInfo, {RoadInfoType} from './RoadInfo'
|
import RoadInfo from './RoadInfo'
|
||||||
import RegionInfo from './RegionInfo'
|
|
||||||
import LayerSidebar from './LayerSidebar'
|
import LayerSidebar from './LayerSidebar'
|
||||||
import styles from './styles.module.less'
|
import styles from './styles.module.less'
|
||||||
|
|
||||||
|
@ -32,7 +27,7 @@ const untaggedRoadsLayer = {
|
||||||
paint: {
|
paint: {
|
||||||
'line-width': ['interpolate', ['exponential', 1.5], ['zoom'], 12, 2, 17, 2],
|
'line-width': ['interpolate', ['exponential', 1.5], ['zoom'], 12, 2, 17, 2],
|
||||||
'line-color': '#ABC',
|
'line-color': '#ABC',
|
||||||
// "line-opacity": ["interpolate", ["linear"], ["zoom"], 14, 0, 15, 1],
|
// 'line-opacity': ['interpolate', ['linear'], ['zoom'], 14, 0, 15, 1],
|
||||||
'line-offset': [
|
'line-offset': [
|
||||||
'interpolate',
|
'interpolate',
|
||||||
['exponential', 1.5],
|
['exponential', 1.5],
|
||||||
|
@ -43,28 +38,27 @@ const untaggedRoadsLayer = {
|
||||||
['*', ['get', 'offset_direction'], 8],
|
['*', ['get', 'offset_direction'], 8],
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
|
minzoom: 12,
|
||||||
}
|
}
|
||||||
|
|
||||||
const getUntaggedRoadsLayer = (colorAttribute) =>
|
|
||||||
produce(untaggedRoadsLayer, (draft) => {
|
|
||||||
draft.filter = ['!', isValidAttribute(colorAttribute)]
|
|
||||||
})
|
|
||||||
|
|
||||||
const getRoadsLayer = (colorAttribute, maxCount) =>
|
const getRoadsLayer = (colorAttribute, maxCount) =>
|
||||||
produce(untaggedRoadsLayer, (draft) => {
|
produce(untaggedRoadsLayer, (draft) => {
|
||||||
draft.id = 'obs_roads_normal'
|
draft.id = 'obs_roads_normal'
|
||||||
draft.filter = isValidAttribute(colorAttribute)
|
if (colorAttribute.endsWith('_count')) {
|
||||||
|
// delete draft.filter
|
||||||
|
draft.filter = ['to-boolean', ['get', colorAttribute]]
|
||||||
|
} else {
|
||||||
|
draft.filter = draft.filter[1] // remove '!'
|
||||||
|
}
|
||||||
draft.minzoom = 10
|
draft.minzoom = 10
|
||||||
draft.paint['line-width'][6] = 6 // scale bigger on zoom
|
draft.paint['line-width'][6] = 6 // scale bigger on zoom
|
||||||
draft.paint['line-color'] = colorAttribute.startsWith('distance_')
|
draft.paint['line-color'] = colorAttribute.startsWith('distance_')
|
||||||
? colorByDistance(colorAttribute)
|
? colorByDistance(colorAttribute)
|
||||||
: colorAttribute.endsWith('_count')
|
: colorAttribute.endsWith('_count')
|
||||||
? colorByCount(colorAttribute, maxCount)
|
? colorByCount(colorAttribute, maxCount)
|
||||||
: colorAttribute.endsWith('zone')
|
|
||||||
? borderByZone()
|
|
||||||
: '#DDD'
|
: '#DDD'
|
||||||
// draft.paint["line-opacity"][3] = 12;
|
// draft.paint['line-opacity'][3] = 12
|
||||||
// draft.paint["line-opacity"][5] = 13;
|
// draft.paint['line-opacity'][5] = 13
|
||||||
})
|
})
|
||||||
|
|
||||||
const getEventsLayer = () => ({
|
const getEventsLayer = () => ({
|
||||||
|
@ -74,10 +68,9 @@ const getEventsLayer = () => ({
|
||||||
'source-layer': 'obs_events',
|
'source-layer': 'obs_events',
|
||||||
paint: {
|
paint: {
|
||||||
'circle-radius': ['interpolate', ['linear'], ['zoom'], 14, 3, 17, 8],
|
'circle-radius': ['interpolate', ['linear'], ['zoom'], 14, 3, 17, 8],
|
||||||
'circle-opacity': ['interpolate', ['linear'], ['zoom'], 8, 0.1, 9, 0.3, 10, 0.5, 11, 1],
|
|
||||||
'circle-color': colorByDistance('distance_overtaker'),
|
'circle-color': colorByDistance('distance_overtaker'),
|
||||||
},
|
},
|
||||||
minzoom: 8,
|
minzoom: 11,
|
||||||
})
|
})
|
||||||
|
|
||||||
const getEventsTextLayer = () => ({
|
const getEventsTextLayer = () => ({
|
||||||
|
@ -93,6 +86,7 @@ const getEventsTextLayer = () => ({
|
||||||
{'min-fraction-digits': 2, 'max-fraction-digits': 2},
|
{'min-fraction-digits': 2, 'max-fraction-digits': 2},
|
||||||
],
|
],
|
||||||
'text-allow-overlap': true,
|
'text-allow-overlap': true,
|
||||||
|
'text-font': ['Open Sans Bold', 'Arial Unicode MS Regular'],
|
||||||
'text-size': 14,
|
'text-size': 14,
|
||||||
'text-keep-upright': false,
|
'text-keep-upright': false,
|
||||||
'text-anchor': 'left',
|
'text-anchor': 'left',
|
||||||
|
@ -107,62 +101,25 @@ const getEventsTextLayer = () => ({
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
interface RegionInfo {
|
export default function MapPage() {
|
||||||
properties: {
|
const {obsMapSource} = useConfig() || {}
|
||||||
admin_level: number
|
const [clickLocation, setClickLocation] = useState<{longitude: number; latitude: number} | null>(null)
|
||||||
name: string
|
|
||||||
overtaking_event_count: number
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
type Details = {type: 'road'; road: RoadInfoType} | {type: 'region'; region: RegionInfo}
|
|
||||||
|
|
||||||
function MapPage({login}) {
|
|
||||||
const {obsMapSource, banner} = useConfig() || {}
|
|
||||||
const [details, setDetails] = useState<null | Details>(null)
|
|
||||||
|
|
||||||
const onCloseDetails = useCallback(() => setDetails(null), [setDetails])
|
|
||||||
|
|
||||||
const mapConfig = useMapConfig()
|
const mapConfig = useMapConfig()
|
||||||
|
|
||||||
const viewportRef = useRef()
|
|
||||||
const mapInfoPortal = useRef()
|
|
||||||
|
|
||||||
const onViewportChange = useCallback(
|
|
||||||
(viewport) => {
|
|
||||||
viewportRef.current = viewport
|
|
||||||
},
|
|
||||||
[viewportRef]
|
|
||||||
)
|
|
||||||
|
|
||||||
const onClick = useCallback(
|
const onClick = useCallback(
|
||||||
async (e) => {
|
(e) => {
|
||||||
// check if we clicked inside the mapInfoBox, if so, early exit
|
|
||||||
let node = e.target
|
let node = e.target
|
||||||
while (node) {
|
while (node) {
|
||||||
if ([styles.mapInfoBox, styles.mapToolbar].some((className) => node?.classList?.contains(className))) {
|
if (node?.classList?.contains(styles.mapInfoBox)) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
node = node.parentNode
|
node = node.parentNode
|
||||||
}
|
}
|
||||||
|
|
||||||
const {zoom} = viewportRef.current
|
setClickLocation({longitude: e.lngLat[0], latitude: e.lngLat[1]})
|
||||||
|
|
||||||
if (zoom < 10) {
|
|
||||||
const clickedRegion = e.features?.find((f) => f.source === 'obs' && f.sourceLayer === 'obs_regions')
|
|
||||||
setDetails(clickedRegion ? {type: 'region', region: clickedRegion} : null)
|
|
||||||
} else {
|
|
||||||
const road = await api.get('/mapdetails/road', {
|
|
||||||
query: {
|
|
||||||
longitude: e.lngLat[0],
|
|
||||||
latitude: e.lngLat[1],
|
|
||||||
radius: 100,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
setDetails(road?.road ? {type: 'road', road} : null)
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
[setDetails]
|
[setClickLocation]
|
||||||
)
|
)
|
||||||
|
|
||||||
const [layerSidebar, setLayerSidebar] = useState(true)
|
const [layerSidebar, setLayerSidebar] = useState(true)
|
||||||
|
@ -173,9 +130,8 @@ function MapPage({login}) {
|
||||||
|
|
||||||
const layers = []
|
const layers = []
|
||||||
|
|
||||||
const untaggedRoadsLayerCustom = useMemo(() => getUntaggedRoadsLayer(attribute), [attribute])
|
|
||||||
if (mapConfig.obsRoads.show && mapConfig.obsRoads.showUntagged) {
|
if (mapConfig.obsRoads.show && mapConfig.obsRoads.showUntagged) {
|
||||||
layers.push(untaggedRoadsLayerCustom)
|
layers.push(untaggedRoadsLayer)
|
||||||
}
|
}
|
||||||
|
|
||||||
const roadsLayer = useMemo(() => getRoadsLayer(attribute, maxCount), [attribute, maxCount])
|
const roadsLayer = useMemo(() => getRoadsLayer(attribute, maxCount), [attribute, maxCount])
|
||||||
|
@ -184,94 +140,50 @@ function MapPage({login}) {
|
||||||
}
|
}
|
||||||
|
|
||||||
const regionLayers = useMemo(() => getRegionLayers(), [])
|
const regionLayers = useMemo(() => getRegionLayers(), [])
|
||||||
if (mapConfig.obsRegions.show) {
|
layers.push(...regionLayers)
|
||||||
layers.push(...regionLayers)
|
|
||||||
}
|
|
||||||
|
|
||||||
const eventsLayer = useMemo(() => getEventsLayer(), [])
|
const eventsLayer = useMemo(() => getEventsLayer(), [])
|
||||||
const eventsTextLayer = useMemo(() => getEventsTextLayer(), [])
|
const eventsTextLayer = useMemo(() => getEventsTextLayer(), [])
|
||||||
|
|
||||||
if (mapConfig.obsEvents.show) {
|
if (mapConfig.obsEvents.show) {
|
||||||
layers.push(eventsLayer)
|
layers.push(eventsLayer)
|
||||||
layers.push(eventsTextLayer)
|
layers.push(eventsTextLayer)
|
||||||
}
|
}
|
||||||
|
|
||||||
const onToggleLayerSidebarButtonClick = useCallback(
|
|
||||||
(e) => {
|
|
||||||
e.stopPropagation()
|
|
||||||
e.preventDefault()
|
|
||||||
console.log('toggl;e')
|
|
||||||
setLayerSidebar((v) => !v)
|
|
||||||
},
|
|
||||||
[setLayerSidebar]
|
|
||||||
)
|
|
||||||
|
|
||||||
if (!obsMapSource) {
|
if (!obsMapSource) {
|
||||||
return null
|
return null
|
||||||
}
|
}
|
||||||
|
|
||||||
const tiles = obsMapSource?.tiles?.map((tileUrl: string) => {
|
|
||||||
const query = new URLSearchParams()
|
|
||||||
if (login) {
|
|
||||||
if (mapConfig.filters.currentUser) {
|
|
||||||
query.append('user', login.id)
|
|
||||||
}
|
|
||||||
|
|
||||||
if (mapConfig.filters.dateMode === 'range') {
|
|
||||||
if (mapConfig.filters.startDate) {
|
|
||||||
query.append('start', mapConfig.filters.startDate)
|
|
||||||
}
|
|
||||||
if (mapConfig.filters.endDate) {
|
|
||||||
query.append('end', mapConfig.filters.endDate)
|
|
||||||
}
|
|
||||||
} else if (mapConfig.filters.dateMode === 'threshold') {
|
|
||||||
if (mapConfig.filters.startDate) {
|
|
||||||
query.append(mapConfig.filters.thresholdAfter ? 'start' : 'end', mapConfig.filters.startDate)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
const queryString = String(query)
|
|
||||||
return tileUrl + (queryString ? '?' : '') + queryString
|
|
||||||
})
|
|
||||||
|
|
||||||
const hasFilters: boolean = login && (mapConfig.filters.currentUser || mapConfig.filters.dateMode !== 'none')
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Page fullScreen title="Map">
|
<Page fullScreen title="Map">
|
||||||
<div className={classNames(styles.mapContainer, banner ? styles.hasBanner : null)} ref={mapInfoPortal}>
|
<div className={styles.mapContainer}>
|
||||||
{layerSidebar && (
|
{layerSidebar && (
|
||||||
<div className={styles.mapSidebar}>
|
<div className={styles.mapSidebar}>
|
||||||
<LayerSidebar />
|
<LayerSidebar />
|
||||||
</div>
|
</div>
|
||||||
)}
|
)}
|
||||||
<div className={styles.map}>
|
<div className={styles.map}>
|
||||||
<Map viewportFromUrl onClick={onClick} hasToolbar onViewportChange={onViewportChange}>
|
<Map viewportFromUrl onClick={onClick}>
|
||||||
<div className={styles.mapToolbar}>
|
<Button
|
||||||
<Button primary icon="bars" active={layerSidebar} onClick={onToggleLayerSidebarButtonClick} />
|
style={{
|
||||||
</div>
|
position: 'absolute',
|
||||||
<Source id="obs" {...obsMapSource} tiles={tiles}>
|
left: 44,
|
||||||
|
top: 9,
|
||||||
|
}}
|
||||||
|
primary
|
||||||
|
icon="bars"
|
||||||
|
active={layerSidebar}
|
||||||
|
onClick={() => setLayerSidebar(layerSidebar ? false : true)}
|
||||||
|
/>
|
||||||
|
<Source id="obs" {...obsMapSource}>
|
||||||
{layers.map((layer) => (
|
{layers.map((layer) => (
|
||||||
<Layer key={layer.id} {...layer} />
|
<Layer key={layer.id} {...layer} />
|
||||||
))}
|
))}
|
||||||
</Source>
|
</Source>
|
||||||
|
|
||||||
{details?.type === 'road' && details?.road?.road && (
|
<RoadInfo {...{clickLocation}} />
|
||||||
<RoadInfo
|
|
||||||
roadInfo={details.road}
|
|
||||||
mapInfoPortal={mapInfoPortal.current}
|
|
||||||
onClose={onCloseDetails}
|
|
||||||
{...{hasFilters}}
|
|
||||||
/>
|
|
||||||
)}
|
|
||||||
|
|
||||||
{details?.type === 'region' && details?.region && (
|
|
||||||
<RegionInfo region={details.region} mapInfoPortal={mapInfoPortal.current} onClose={onCloseDetails} />
|
|
||||||
)}
|
|
||||||
</Map>
|
</Map>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</Page>
|
</Page>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
export default connect((state) => ({login: state.login}))(MapPage)
|
|
||||||
|
|
|
@ -6,74 +6,25 @@
|
||||||
position: relative;
|
position: relative;
|
||||||
display: flex;
|
display: flex;
|
||||||
align-items: stretch;
|
align-items: stretch;
|
||||||
|
|
||||||
&.hasBanner {
|
|
||||||
height: calc(100vh - @menuHeight - 50px);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
.mapSidebar {
|
.mapSidebar {
|
||||||
width: 24rem;
|
width: 20rem;
|
||||||
background: white;
|
background: white;
|
||||||
border-right: 1px solid @borderColor;
|
border-right: 1px solid @borderColor;
|
||||||
padding: 1rem;
|
padding: 1rem;
|
||||||
overflow-y: auto;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
.map {
|
.map {
|
||||||
flex: 1 1 0;
|
flex: 1 1 0;
|
||||||
overflow: hidden;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
.mapInfoBox {
|
.mapInfoBox {
|
||||||
|
position: absolute;
|
||||||
|
right: 0;
|
||||||
|
top: 0;
|
||||||
|
max-height: 100%;
|
||||||
width: 36rem;
|
width: 36rem;
|
||||||
overflow: auto;
|
overflow: auto;
|
||||||
border-left: 1px solid @borderColor;
|
margin: 20px;
|
||||||
background: white;
|
|
||||||
padding: 16px;
|
|
||||||
}
|
|
||||||
|
|
||||||
.copyright {
|
|
||||||
color: #888;
|
|
||||||
font-size: 0.8em;
|
|
||||||
line-height: 1.4;
|
|
||||||
margin-block-start: 1em;
|
|
||||||
}
|
|
||||||
|
|
||||||
.mapToolbar {
|
|
||||||
position: absolute;
|
|
||||||
left: 16px;
|
|
||||||
top: 16px;
|
|
||||||
}
|
|
||||||
|
|
||||||
.closeHeader {
|
|
||||||
display: flex;
|
|
||||||
align-items: baseline;
|
|
||||||
justify-content: space-between;
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@media @mobile {
|
|
||||||
.mapContainer {
|
|
||||||
height: auto;
|
|
||||||
min-height: calc(100vh - @menuHeightMobile);
|
|
||||||
&.hasBanner {
|
|
||||||
height: calc(100vh - @menuHeightMobile - 50px);
|
|
||||||
}
|
|
||||||
flex-direction: column;
|
|
||||||
}
|
|
||||||
|
|
||||||
.map {
|
|
||||||
height: 60vh;
|
|
||||||
}
|
|
||||||
|
|
||||||
.mapSidebar {
|
|
||||||
width: auto;
|
|
||||||
height: auto;
|
|
||||||
}
|
|
||||||
|
|
||||||
.mapInfoBox {
|
|
||||||
width: auto;
|
|
||||||
height: auto;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,360 +0,0 @@
|
||||||
import React, {useCallback, useMemo, useState} from 'react'
|
|
||||||
import {connect} from 'react-redux'
|
|
||||||
import {
|
|
||||||
Accordion,
|
|
||||||
Button,
|
|
||||||
Checkbox,
|
|
||||||
Confirm,
|
|
||||||
Header,
|
|
||||||
Icon,
|
|
||||||
Item,
|
|
||||||
List,
|
|
||||||
Loader,
|
|
||||||
Dropdown,
|
|
||||||
SemanticCOLORS,
|
|
||||||
SemanticICONS,
|
|
||||||
Table,
|
|
||||||
} from 'semantic-ui-react'
|
|
||||||
import {useObservable} from 'rxjs-hooks'
|
|
||||||
import {Link} from 'react-router-dom'
|
|
||||||
import {of, from, concat, BehaviorSubject, combineLatest} from 'rxjs'
|
|
||||||
import {map, switchMap, distinctUntilChanged} from 'rxjs/operators'
|
|
||||||
import _ from 'lodash'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import type {ProcessingStatus, Track, UserDevice} from 'types'
|
|
||||||
import {Page, FormattedDate, Visibility} from 'components'
|
|
||||||
import api from 'api'
|
|
||||||
import {useCallbackRef, formatDistance, formatDuration} from 'utils'
|
|
||||||
|
|
||||||
import download from 'downloadjs'
|
|
||||||
|
|
||||||
const COLOR_BY_STATUS: Record<ProcessingStatus, SemanticCOLORS> = {
|
|
||||||
error: 'red',
|
|
||||||
complete: 'green',
|
|
||||||
created: 'grey',
|
|
||||||
queued: 'orange',
|
|
||||||
processing: 'orange',
|
|
||||||
}
|
|
||||||
|
|
||||||
const ICON_BY_STATUS: Record<ProcessingStatus, SemanticICONS> = {
|
|
||||||
error: 'warning sign',
|
|
||||||
complete: 'check circle outline',
|
|
||||||
created: 'bolt',
|
|
||||||
queued: 'bolt',
|
|
||||||
processing: 'bolt',
|
|
||||||
}
|
|
||||||
|
|
||||||
function ProcessingStatusLabel({status}: {status: ProcessingStatus}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
return (
|
|
||||||
<span title={t(`TracksPage.processing.${status}`)}>
|
|
||||||
<Icon color={COLOR_BY_STATUS[status]} name={ICON_BY_STATUS[status]} />
|
|
||||||
</span>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
function SortableHeader({children, setOrderBy, orderBy, reversed, setReversed, name, ...props}) {
|
|
||||||
const toggleSort = (e) => {
|
|
||||||
e.preventDefault()
|
|
||||||
e.stopPropagation()
|
|
||||||
|
|
||||||
if (orderBy === name) {
|
|
||||||
if (!reversed) {
|
|
||||||
setReversed(true)
|
|
||||||
} else {
|
|
||||||
setReversed(false)
|
|
||||||
setOrderBy(null)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
setReversed(false)
|
|
||||||
setOrderBy(name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let icon = orderBy === name ? (reversed ? 'sort descending' : 'sort ascending') : null
|
|
||||||
|
|
||||||
return (
|
|
||||||
<Table.HeaderCell {...props}>
|
|
||||||
<div onClick={toggleSort}>
|
|
||||||
{children}
|
|
||||||
<Icon name={icon} />
|
|
||||||
</div>
|
|
||||||
</Table.HeaderCell>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
type Filters = {
|
|
||||||
userDeviceId?: null | number
|
|
||||||
visibility?: null | boolean
|
|
||||||
}
|
|
||||||
|
|
||||||
function TrackFilters({
|
|
||||||
filters,
|
|
||||||
setFilters,
|
|
||||||
deviceNames,
|
|
||||||
}: {
|
|
||||||
filters: Filters
|
|
||||||
setFilters: (f: Filters) => void
|
|
||||||
deviceNames: null | Record<number, string>
|
|
||||||
}) {
|
|
||||||
return (
|
|
||||||
<List horizontal>
|
|
||||||
<List.Item>
|
|
||||||
<List.Header>Device</List.Header>
|
|
||||||
<Dropdown
|
|
||||||
selection
|
|
||||||
clearable
|
|
||||||
options={[
|
|
||||||
{value: 0, key: '__none__', text: 'All my devices'},
|
|
||||||
..._.sortBy(Object.entries(deviceNames ?? {}), 1).map(([deviceId, deviceName]: [string, string]) => ({
|
|
||||||
value: Number(deviceId),
|
|
||||||
key: deviceId,
|
|
||||||
text: deviceName,
|
|
||||||
})),
|
|
||||||
]}
|
|
||||||
value={filters?.userDeviceId ?? 0}
|
|
||||||
onChange={(_e, {value}) => setFilters({...filters, userDeviceId: (value as number) || null})}
|
|
||||||
/>
|
|
||||||
</List.Item>
|
|
||||||
|
|
||||||
<List.Item>
|
|
||||||
<List.Header>Visibility</List.Header>
|
|
||||||
<Dropdown
|
|
||||||
selection
|
|
||||||
clearable
|
|
||||||
options={[
|
|
||||||
{value: 'none', key: 'any', text: 'Any'},
|
|
||||||
{value: true, key: 'public', text: 'Public'},
|
|
||||||
{value: false, key: 'private', text: 'Private'},
|
|
||||||
]}
|
|
||||||
value={filters?.visibility ?? 'none'}
|
|
||||||
onChange={(_e, {value}) =>
|
|
||||||
setFilters({
|
|
||||||
...filters,
|
|
||||||
visibility: value === 'none' ? null : (value as boolean),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
/>
|
|
||||||
</List.Item>
|
|
||||||
</List>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
function TracksTable({title}) {
|
|
||||||
const [orderBy, setOrderBy] = useState('recordedAt')
|
|
||||||
const [reversed, setReversed] = useState(true)
|
|
||||||
const [showFilters, setShowFilters] = useState(false)
|
|
||||||
const [filters, setFilters] = useState<Filters>({})
|
|
||||||
const [selectedTracks, setSelectedTracks] = useState<Record<string, boolean>>({})
|
|
||||||
|
|
||||||
const toggleTrackSelection = useCallbackRef((slug: string, selected?: boolean) => {
|
|
||||||
const newSelected = selected ?? !selectedTracks[slug]
|
|
||||||
setSelectedTracks(_.pickBy({...selectedTracks, [slug]: newSelected}, _.identity))
|
|
||||||
})
|
|
||||||
|
|
||||||
const query = _.pickBy(
|
|
||||||
{
|
|
||||||
limit: 1000,
|
|
||||||
offset: 0,
|
|
||||||
order_by: orderBy,
|
|
||||||
reversed: reversed ? 'true' : 'false',
|
|
||||||
user_device_id: filters?.userDeviceId,
|
|
||||||
public: filters?.visibility,
|
|
||||||
},
|
|
||||||
(x) => x != null
|
|
||||||
)
|
|
||||||
|
|
||||||
const forceUpdate$ = useMemo(() => new BehaviorSubject(null), [])
|
|
||||||
const tracks: Track[] | null = useObservable(
|
|
||||||
(_$, inputs$) =>
|
|
||||||
combineLatest([
|
|
||||||
inputs$.pipe(
|
|
||||||
map(([query]) => query),
|
|
||||||
distinctUntilChanged(_.isEqual)
|
|
||||||
),
|
|
||||||
forceUpdate$,
|
|
||||||
]).pipe(switchMap(([query]) => concat(of(null), from(api.get('/tracks/feed', {query}).then((r) => r.tracks))))),
|
|
||||||
null,
|
|
||||||
[query]
|
|
||||||
)
|
|
||||||
|
|
||||||
const deviceNames: null | Record<number, string> = useObservable(() =>
|
|
||||||
from(api.get('/user/devices')).pipe(
|
|
||||||
map((response: UserDevice[]) =>
|
|
||||||
Object.fromEntries(response.map((device) => [device.id, device.displayName || device.identifier]))
|
|
||||||
)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
const p = {orderBy, setOrderBy, reversed, setReversed}
|
|
||||||
|
|
||||||
const selectedCount = Object.keys(selectedTracks).length
|
|
||||||
const noneSelected = selectedCount === 0
|
|
||||||
const allSelected = selectedCount === tracks?.length
|
|
||||||
const selectAll = () => {
|
|
||||||
setSelectedTracks(Object.fromEntries(tracks?.map((t) => [t.slug, true]) ?? []))
|
|
||||||
}
|
|
||||||
const selectNone = () => {
|
|
||||||
setSelectedTracks({})
|
|
||||||
}
|
|
||||||
|
|
||||||
const bulkAction = async (action: string) => {
|
|
||||||
const response = await api.post('/tracks/bulk', {
|
|
||||||
body: {
|
|
||||||
action,
|
|
||||||
tracks: Object.keys(selectedTracks),
|
|
||||||
},
|
|
||||||
returnResponse: true,
|
|
||||||
})
|
|
||||||
if (action === 'download') {
|
|
||||||
const contentType = response.headers.get('content-type') ?? 'application/x-gtar'
|
|
||||||
|
|
||||||
const filename = response.headers.get('content-disposition')?.match(/filename="([^"]+)"/)?.[1] ?? 'tracks.tar.bz2'
|
|
||||||
download(await response.blob(), filename, contentType)
|
|
||||||
}
|
|
||||||
|
|
||||||
setShowBulkDelete(false)
|
|
||||||
setSelectedTracks({})
|
|
||||||
forceUpdate$.next(null)
|
|
||||||
}
|
|
||||||
const [showBulkDelete, setShowBulkDelete] = useState(false)
|
|
||||||
|
|
||||||
return (
|
|
||||||
<>
|
|
||||||
<div style={{float: 'right'}}>
|
|
||||||
<Dropdown disabled={noneSelected} text="Bulk actions" floating button>
|
|
||||||
<Dropdown.Menu>
|
|
||||||
<Dropdown.Header>Selection of {selectedCount} tracks</Dropdown.Header>
|
|
||||||
<Dropdown.Item onClick={() => bulkAction('makePrivate')}>Make private</Dropdown.Item>
|
|
||||||
<Dropdown.Item onClick={() => bulkAction('makePublic')}>Make public</Dropdown.Item>
|
|
||||||
<Dropdown.Item onClick={() => bulkAction('reprocess')}>Reprocess</Dropdown.Item>
|
|
||||||
<Dropdown.Item onClick={() => bulkAction('download')}>Download</Dropdown.Item>
|
|
||||||
<Dropdown.Item onClick={() => setShowBulkDelete(true)}>Delete</Dropdown.Item>
|
|
||||||
</Dropdown.Menu>
|
|
||||||
</Dropdown>
|
|
||||||
<Link component={UploadButton} to="/upload" />
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<Header as="h1">{title}</Header>
|
|
||||||
<div style={{clear: 'both'}}>
|
|
||||||
<Loader content={t('general.loading')} active={tracks == null} />
|
|
||||||
|
|
||||||
<Accordion>
|
|
||||||
<Accordion.Title active={showFilters} index={0} onClick={() => setShowFilters(!showFilters)}>
|
|
||||||
<Icon name="dropdown" />
|
|
||||||
Filters
|
|
||||||
</Accordion.Title>
|
|
||||||
<Accordion.Content active={showFilters}>
|
|
||||||
<TrackFilters {...{filters, setFilters, deviceNames}} />
|
|
||||||
</Accordion.Content>
|
|
||||||
</Accordion>
|
|
||||||
|
|
||||||
<Confirm
|
|
||||||
open={showBulkDelete}
|
|
||||||
onCancel={() => setShowBulkDelete(false)}
|
|
||||||
onConfirm={() => bulkAction('delete')}
|
|
||||||
content={`Are you sure you want to delete ${selectedCount} tracks?`}
|
|
||||||
confirmButton={t('general.delete')}
|
|
||||||
cancelButton={t('general.cancel')}
|
|
||||||
/>
|
|
||||||
|
|
||||||
<Table compact>
|
|
||||||
<Table.Header>
|
|
||||||
<Table.Row>
|
|
||||||
<Table.HeaderCell>
|
|
||||||
<Checkbox
|
|
||||||
checked={allSelected}
|
|
||||||
indeterminate={!allSelected && !noneSelected}
|
|
||||||
onClick={() => (noneSelected ? selectAll() : selectNone())}
|
|
||||||
/>
|
|
||||||
</Table.HeaderCell>
|
|
||||||
|
|
||||||
<SortableHeader {...p} name="title">
|
|
||||||
Title
|
|
||||||
</SortableHeader>
|
|
||||||
<SortableHeader {...p} name="recordedAt">
|
|
||||||
Recorded at
|
|
||||||
</SortableHeader>
|
|
||||||
<SortableHeader {...p} name="visibility">
|
|
||||||
Visibility
|
|
||||||
</SortableHeader>
|
|
||||||
<SortableHeader {...p} name="length" textAlign="right">
|
|
||||||
Length
|
|
||||||
</SortableHeader>
|
|
||||||
<SortableHeader {...p} name="duration" textAlign="right">
|
|
||||||
Duration
|
|
||||||
</SortableHeader>
|
|
||||||
<SortableHeader {...p} name="user_device_id">
|
|
||||||
Device
|
|
||||||
</SortableHeader>
|
|
||||||
</Table.Row>
|
|
||||||
</Table.Header>
|
|
||||||
|
|
||||||
<Table.Body>
|
|
||||||
{tracks?.map((track: Track) => (
|
|
||||||
<Table.Row key={track.slug}>
|
|
||||||
<Table.Cell>
|
|
||||||
<Checkbox
|
|
||||||
onClick={(e) => toggleTrackSelection(track.slug)}
|
|
||||||
checked={selectedTracks[track.slug] ?? false}
|
|
||||||
/>
|
|
||||||
</Table.Cell>
|
|
||||||
<Table.Cell>
|
|
||||||
{track.processingStatus == null ? null : <ProcessingStatusLabel status={track.processingStatus} />}
|
|
||||||
<Item.Header as={Link} to={`/tracks/${track.slug}`}>
|
|
||||||
{track.title || t('general.unnamedTrack')}
|
|
||||||
</Item.Header>
|
|
||||||
</Table.Cell>
|
|
||||||
|
|
||||||
<Table.Cell>
|
|
||||||
<FormattedDate date={track.recordedAt} />
|
|
||||||
</Table.Cell>
|
|
||||||
|
|
||||||
<Table.Cell>{track.public == null ? null : <Visibility public={track.public} />}</Table.Cell>
|
|
||||||
|
|
||||||
<Table.Cell textAlign="right">{formatDistance(track.length)}</Table.Cell>
|
|
||||||
|
|
||||||
<Table.Cell textAlign="right">{formatDuration(track.duration)}</Table.Cell>
|
|
||||||
|
|
||||||
<Table.Cell>{track.userDeviceId ? deviceNames?.[track.userDeviceId] ?? '...' : null}</Table.Cell>
|
|
||||||
</Table.Row>
|
|
||||||
))}
|
|
||||||
</Table.Body>
|
|
||||||
</Table>
|
|
||||||
</div>
|
|
||||||
</>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
function UploadButton({navigate, ...props}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const onClick = useCallback(
|
|
||||||
(e) => {
|
|
||||||
e.preventDefault()
|
|
||||||
navigate()
|
|
||||||
},
|
|
||||||
[navigate]
|
|
||||||
)
|
|
||||||
return (
|
|
||||||
<Button onClick={onClick} {...props} color="green">
|
|
||||||
{t('TracksPage.upload')}
|
|
||||||
</Button>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
const MyTracksPage = connect((state) => ({login: (state as any).login}))(function MyTracksPage({login}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
const title = t('TracksPage.titleUser')
|
|
||||||
|
|
||||||
return (
|
|
||||||
<Page title={title}>
|
|
||||||
<TracksTable {...{title}} />
|
|
||||||
</Page>
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
export default MyTracksPage
|
|
|
@ -1,18 +1,16 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import {Button, Header} from 'semantic-ui-react'
|
import {Button, Header} from 'semantic-ui-react'
|
||||||
import {useHistory} from 'react-router-dom'
|
import {useHistory} from 'react-router-dom'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {Page} from '../components'
|
import {Page} from '../components'
|
||||||
|
|
||||||
export default function NotFoundPage() {
|
export default function NotFoundPage() {
|
||||||
const {t} = useTranslation()
|
|
||||||
const history = useHistory()
|
const history = useHistory()
|
||||||
return (
|
return (
|
||||||
<Page title={t('NotFoundPage.title')}>
|
<Page title="Page not found">
|
||||||
<Header as="h2">{t('NotFoundPage.title')}</Header>
|
<Header as="h2">Page not found</Header>
|
||||||
<p>{t('NotFoundPage.description')}</p>
|
<p>You know what that means...</p>
|
||||||
<Button onClick={history.goBack.bind(history)}>{t('NotFoundPage.goBack')}</Button>
|
<Button onClick={history.goBack.bind(history)}>Go back</Button>
|
||||||
</Page>
|
</Page>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
178
frontend/src/pages/SettingsPage.tsx
Normal file
178
frontend/src/pages/SettingsPage.tsx
Normal file
|
@ -0,0 +1,178 @@
|
||||||
|
import React from 'react'
|
||||||
|
import {connect} from 'react-redux'
|
||||||
|
import {Message, Icon, Grid, Form, Button, TextArea, Ref, Input, Header, Divider, Popup} from 'semantic-ui-react'
|
||||||
|
import {useForm} from 'react-hook-form'
|
||||||
|
|
||||||
|
import {setLogin} from 'reducers/login'
|
||||||
|
import {Page, Stats} from 'components'
|
||||||
|
import api from 'api'
|
||||||
|
import {findInput} from 'utils'
|
||||||
|
import {useConfig} from 'config'
|
||||||
|
|
||||||
|
const SettingsPage = connect((state) => ({login: state.login}), {setLogin})(function SettingsPage({login, setLogin}) {
|
||||||
|
const {register, handleSubmit} = useForm()
|
||||||
|
const [loading, setLoading] = React.useState(false)
|
||||||
|
const [errors, setErrors] = React.useState(null)
|
||||||
|
|
||||||
|
const onSave = React.useCallback(
|
||||||
|
async (changes) => {
|
||||||
|
setLoading(true)
|
||||||
|
setErrors(null)
|
||||||
|
try {
|
||||||
|
const response = await api.put('/user', {body: changes})
|
||||||
|
setLogin(response)
|
||||||
|
} catch (err) {
|
||||||
|
setErrors(err.errors)
|
||||||
|
} finally {
|
||||||
|
setLoading(false)
|
||||||
|
}
|
||||||
|
},
|
||||||
|
[setLoading, setLogin, setErrors]
|
||||||
|
)
|
||||||
|
|
||||||
|
const onGenerateNewKey = React.useCallback(async () => {
|
||||||
|
setLoading(true)
|
||||||
|
setErrors(null)
|
||||||
|
try {
|
||||||
|
const response = await api.put('/user', {body: {updateApiKey: true}})
|
||||||
|
setLogin(response)
|
||||||
|
} catch (err) {
|
||||||
|
setErrors(err.errors)
|
||||||
|
} finally {
|
||||||
|
setLoading(false)
|
||||||
|
}
|
||||||
|
}, [setLoading, setLogin, setErrors])
|
||||||
|
|
||||||
|
return (
|
||||||
|
<Page title="Settings">
|
||||||
|
<Grid centered relaxed divided stackable>
|
||||||
|
<Grid.Row>
|
||||||
|
<Grid.Column width={8}>
|
||||||
|
<Header as="h2">My profile</Header>
|
||||||
|
|
||||||
|
<Message info>All of this information is public.</Message>
|
||||||
|
|
||||||
|
<Form onSubmit={handleSubmit(onSave)} loading={loading}>
|
||||||
|
<Ref innerRef={findInput(register)}>
|
||||||
|
<Form.Input
|
||||||
|
error={errors?.username}
|
||||||
|
label="Username"
|
||||||
|
name="username"
|
||||||
|
defaultValue={login.username}
|
||||||
|
disabled
|
||||||
|
/>
|
||||||
|
</Ref>
|
||||||
|
<Form.Field error={errors?.bio}>
|
||||||
|
<label>Bio</label>
|
||||||
|
<Ref innerRef={register}>
|
||||||
|
<TextArea name="bio" rows={4} defaultValue={login.bio} />
|
||||||
|
</Ref>
|
||||||
|
</Form.Field>
|
||||||
|
<Form.Field error={errors?.image}>
|
||||||
|
<label>Avatar URL</label>
|
||||||
|
<Ref innerRef={findInput(register)}>
|
||||||
|
<Input name="image" defaultValue={login.image} />
|
||||||
|
</Ref>
|
||||||
|
</Form.Field>
|
||||||
|
|
||||||
|
<Button type="submit" primary>
|
||||||
|
Save
|
||||||
|
</Button>
|
||||||
|
</Form>
|
||||||
|
</Grid.Column>
|
||||||
|
<Grid.Column width={6}>
|
||||||
|
<ApiKeyDialog {...{login, onGenerateNewKey}} />
|
||||||
|
|
||||||
|
<Divider />
|
||||||
|
|
||||||
|
<Stats user={login.username} />
|
||||||
|
</Grid.Column>
|
||||||
|
</Grid.Row>
|
||||||
|
</Grid>
|
||||||
|
</Page>
|
||||||
|
)
|
||||||
|
})
|
||||||
|
|
||||||
|
function CopyInput({value, ...props}) {
|
||||||
|
const [success, setSuccess] = React.useState(null)
|
||||||
|
const onClick = async () => {
|
||||||
|
try {
|
||||||
|
await window.navigator?.clipboard?.writeText(value)
|
||||||
|
setSuccess(true)
|
||||||
|
} catch (err) {
|
||||||
|
setSuccess(false)
|
||||||
|
} finally {
|
||||||
|
setTimeout(() => {
|
||||||
|
setSuccess(null)
|
||||||
|
}, 2000)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return (
|
||||||
|
<Popup
|
||||||
|
trigger={<Input {...props} value={value} fluid action={{icon: 'copy', onClick}} />}
|
||||||
|
position="top right"
|
||||||
|
open={success != null}
|
||||||
|
content={success ? 'Copied.' : 'Failed to copy.'}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
const selectField = findInput((ref) => ref?.select())
|
||||||
|
|
||||||
|
function ApiKeyDialog({login, onGenerateNewKey}) {
|
||||||
|
const config = useConfig()
|
||||||
|
const [show, setShow] = React.useState(false)
|
||||||
|
const onClick = React.useCallback(
|
||||||
|
(e) => {
|
||||||
|
e.preventDefault()
|
||||||
|
setShow(true)
|
||||||
|
},
|
||||||
|
[setShow]
|
||||||
|
)
|
||||||
|
|
||||||
|
const onGenerateNewKeyInner = React.useCallback(
|
||||||
|
(e) => {
|
||||||
|
e.preventDefault()
|
||||||
|
onGenerateNewKey()
|
||||||
|
},
|
||||||
|
[onGenerateNewKey]
|
||||||
|
)
|
||||||
|
|
||||||
|
return (
|
||||||
|
<>
|
||||||
|
<Header as="h2">My API Key</Header>
|
||||||
|
<p>
|
||||||
|
Here you find your API Key, for use in the OpenBikeSensor. You can to copy and paste it into your sensor's
|
||||||
|
configuration interface to allow direct upload from the device.
|
||||||
|
</p>
|
||||||
|
<p>Please protect your API Key carefully as it allows full control over your account.</p>
|
||||||
|
<div style={{minHeight: 40, marginBottom: 16}}>
|
||||||
|
{show ? (
|
||||||
|
login.apiKey ? (
|
||||||
|
<Ref innerRef={selectField}>
|
||||||
|
<CopyInput label="Personal API Key" value={login.apiKey} />
|
||||||
|
</Ref>
|
||||||
|
) : (
|
||||||
|
<Message warning content="You have no API Key, please generate one below." />
|
||||||
|
)
|
||||||
|
) : (
|
||||||
|
<Button onClick={onClick}>
|
||||||
|
<Icon name="lock" /> Show API Key
|
||||||
|
</Button>
|
||||||
|
)}
|
||||||
|
</div>
|
||||||
|
<p>The API URL should be set to:</p>
|
||||||
|
<div style={{marginBottom: 16}}>
|
||||||
|
<CopyInput label="API URL" value={config?.apiUrl?.replace(/\/api$/, '') ?? '...'} />
|
||||||
|
</div>
|
||||||
|
<p>
|
||||||
|
You can generate a new API Key here, which will invalidate the old one, disconnecting all devices you used it on
|
||||||
|
from your account.
|
||||||
|
</p>
|
||||||
|
<Button onClick={onGenerateNewKeyInner}>Generate new API key</Button>
|
||||||
|
</>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
export default SettingsPage
|
|
@ -1,102 +0,0 @@
|
||||||
import React from 'react'
|
|
||||||
import {connect} from 'react-redux'
|
|
||||||
import {Message, Icon, Button, Ref, Input, Segment, Popup} from 'semantic-ui-react'
|
|
||||||
import Markdown from 'react-markdown'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {setLogin} from 'reducers/login'
|
|
||||||
import api from 'api'
|
|
||||||
import {findInput} from 'utils'
|
|
||||||
import {useConfig} from 'config'
|
|
||||||
|
|
||||||
function CopyInput({value, ...props}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const [success, setSuccess] = React.useState(null)
|
|
||||||
const onClick = async () => {
|
|
||||||
try {
|
|
||||||
await window.navigator?.clipboard?.writeText(value)
|
|
||||||
setSuccess(true)
|
|
||||||
} catch (err) {
|
|
||||||
setSuccess(false)
|
|
||||||
} finally {
|
|
||||||
setTimeout(() => {
|
|
||||||
setSuccess(null)
|
|
||||||
}, 2000)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return (
|
|
||||||
<Popup
|
|
||||||
trigger={<Input {...props} value={value} fluid action={{icon: 'copy', onClick}} />}
|
|
||||||
position="top right"
|
|
||||||
open={success != null}
|
|
||||||
content={success ? t('general.copied') : t('general.copyError')}
|
|
||||||
/>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
const selectField = findInput((ref) => ref?.select())
|
|
||||||
|
|
||||||
const ApiKeySettings = connect((state) => ({login: state.login}), {
|
|
||||||
setLogin,
|
|
||||||
})(function ApiKeySettings({login, setLogin, setErrors}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const [loading, setLoading] = React.useState(false)
|
|
||||||
const config = useConfig()
|
|
||||||
const [show, setShow] = React.useState(false)
|
|
||||||
const onClick = React.useCallback(
|
|
||||||
(e) => {
|
|
||||||
e.preventDefault()
|
|
||||||
setShow(true)
|
|
||||||
},
|
|
||||||
[setShow]
|
|
||||||
)
|
|
||||||
|
|
||||||
const onGenerateNewKey = React.useCallback(
|
|
||||||
async (e) => {
|
|
||||||
e.preventDefault()
|
|
||||||
setLoading(true)
|
|
||||||
try {
|
|
||||||
const response = await api.put('/user', {
|
|
||||||
body: {updateApiKey: true},
|
|
||||||
})
|
|
||||||
setLogin(response)
|
|
||||||
} catch (err) {
|
|
||||||
setErrors(err.errors)
|
|
||||||
} finally {
|
|
||||||
setLoading(false)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
[setLoading, setLogin, setErrors]
|
|
||||||
)
|
|
||||||
|
|
||||||
return (
|
|
||||||
<Segment style={{maxWidth: 600, margin: '24px auto'}}>
|
|
||||||
<Markdown>{t('SettingsPage.apiKey.description')}</Markdown>
|
|
||||||
<div style={{minHeight: 40, marginBottom: 16}}>
|
|
||||||
{show ? (
|
|
||||||
login.apiKey ? (
|
|
||||||
<Ref innerRef={selectField}>
|
|
||||||
<CopyInput label={t('SettingsPage.apiKey.key.label')} value={login.apiKey} />
|
|
||||||
</Ref>
|
|
||||||
) : (
|
|
||||||
<Message warning content={t('SettingsPage.apiKey.key.empty')} />
|
|
||||||
)
|
|
||||||
) : (
|
|
||||||
<Button onClick={onClick}>
|
|
||||||
<Icon name="lock" /> {t('SettingsPage.apiKey.key.show')}
|
|
||||||
</Button>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
<Markdown>{t('SettingsPage.apiKey.urlDescription')}</Markdown>
|
|
||||||
<div style={{marginBottom: 16}}>
|
|
||||||
<CopyInput label={t('SettingsPage.apiKey.url.label')} value={config?.apiUrl?.replace(/\/api$/, '') ?? '...'} />
|
|
||||||
</div>
|
|
||||||
<Markdown>{t('SettingsPage.apiKey.generateDescription')}</Markdown>
|
|
||||||
<p></p>
|
|
||||||
<Button onClick={onGenerateNewKey}>{t('SettingsPage.apiKey.generate')}</Button>
|
|
||||||
</Segment>
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
export default ApiKeySettings
|
|
|
@ -1,125 +0,0 @@
|
||||||
import React, {useCallback, useMemo, useRef} from 'react'
|
|
||||||
import {useObservable} from 'rxjs-hooks'
|
|
||||||
import {concat, from, of, Subject} from 'rxjs'
|
|
||||||
import {Table, Button, Input} from 'semantic-ui-react'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import api from 'api'
|
|
||||||
import {UserDevice} from 'types'
|
|
||||||
import {startWith, switchMap} from 'rxjs/operators'
|
|
||||||
|
|
||||||
function EditField({value, onEdit}) {
|
|
||||||
const [editing, setEditing] = React.useState(false)
|
|
||||||
const [tempValue, setTempValue] = React.useState(value)
|
|
||||||
const timeoutRef = useRef<null | number>(null)
|
|
||||||
|
|
||||||
const cancelTimeout = useCallback(() => {
|
|
||||||
if (timeoutRef.current != null) {
|
|
||||||
clearTimeout(timeoutRef.current)
|
|
||||||
timeoutRef.current = null
|
|
||||||
}
|
|
||||||
}, [timeoutRef])
|
|
||||||
|
|
||||||
const abort = useCallback(() => {
|
|
||||||
cancelTimeout()
|
|
||||||
setEditing(false)
|
|
||||||
setTempValue(value)
|
|
||||||
}, [setEditing, setTempValue, value, cancelTimeout])
|
|
||||||
|
|
||||||
const confirm = useCallback(() => {
|
|
||||||
console.log('confirmed')
|
|
||||||
cancelTimeout()
|
|
||||||
setEditing(false)
|
|
||||||
onEdit(tempValue)
|
|
||||||
}, [setEditing, onEdit, tempValue, cancelTimeout])
|
|
||||||
|
|
||||||
React.useEffect(() => {
|
|
||||||
if (value !== tempValue) {
|
|
||||||
setTempValue(value)
|
|
||||||
}
|
|
||||||
}, [value])
|
|
||||||
|
|
||||||
if (editing) {
|
|
||||||
return (
|
|
||||||
<>
|
|
||||||
<Input
|
|
||||||
value={tempValue}
|
|
||||||
onChange={(e) => setTempValue(e.target.value)}
|
|
||||||
onBlur={(e) => {
|
|
||||||
timeoutRef.current = setTimeout(abort, 20)
|
|
||||||
}}
|
|
||||||
onKeyPress={(e) => {
|
|
||||||
if (e.key === 'Enter') {
|
|
||||||
confirm()
|
|
||||||
} else if (e.key === 'Escape') {
|
|
||||||
abort()
|
|
||||||
}
|
|
||||||
}}
|
|
||||||
style={{marginRight: 8}}
|
|
||||||
/>
|
|
||||||
</>
|
|
||||||
)
|
|
||||||
} else {
|
|
||||||
return (
|
|
||||||
<>
|
|
||||||
{value && <span style={{marginRight: 8}}>{value}</span>}
|
|
||||||
<Button icon="edit" size="tiny" onClick={() => setEditing(true)} />
|
|
||||||
</>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
export default function DeviceList() {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const [loading_, setLoading] = React.useState(false)
|
|
||||||
|
|
||||||
const trigger$ = useMemo(() => new Subject(), [])
|
|
||||||
const devices: null | UserDevice[] = useObservable(() =>
|
|
||||||
trigger$.pipe(
|
|
||||||
startWith(null),
|
|
||||||
switchMap(() => concat(of(null), from(api.get('/user/devices'))))
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
const setDeviceDisplayName = useCallback(
|
|
||||||
async (deviceId: number, displayName: string) => {
|
|
||||||
setLoading(true)
|
|
||||||
try {
|
|
||||||
await api.put(`/user/devices/${deviceId}`, {body: {displayName}})
|
|
||||||
} finally {
|
|
||||||
setLoading(false)
|
|
||||||
trigger$.next(null)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
[trigger$, setLoading]
|
|
||||||
)
|
|
||||||
|
|
||||||
const loading = devices == null || loading_
|
|
||||||
|
|
||||||
return (
|
|
||||||
<>
|
|
||||||
<Table compact {...{loading}}>
|
|
||||||
<Table.Header>
|
|
||||||
<Table.Row>
|
|
||||||
<Table.HeaderCell width={4}>{t('SettingsPage.devices.identifier')}</Table.HeaderCell>
|
|
||||||
<Table.HeaderCell>{t('SettingsPage.devices.alias')}</Table.HeaderCell>
|
|
||||||
<Table.HeaderCell />
|
|
||||||
</Table.Row>
|
|
||||||
</Table.Header>
|
|
||||||
<Table.Body>
|
|
||||||
{devices?.map((device: UserDevice) => (
|
|
||||||
<Table.Row key={device.id}>
|
|
||||||
<Table.Cell> {device.identifier}</Table.Cell>
|
|
||||||
<Table.Cell>
|
|
||||||
<EditField
|
|
||||||
value={device.displayName}
|
|
||||||
onEdit={(displayName: string) => setDeviceDisplayName(device.id, displayName)}
|
|
||||||
/>
|
|
||||||
</Table.Cell>
|
|
||||||
</Table.Row>
|
|
||||||
))}
|
|
||||||
</Table.Body>
|
|
||||||
</Table>
|
|
||||||
</>
|
|
||||||
)
|
|
||||||
}
|
|
|
@ -1,77 +0,0 @@
|
||||||
import React from 'react'
|
|
||||||
import {connect} from 'react-redux'
|
|
||||||
import {Segment, Message, Form, Button, TextArea, Ref, Input} from 'semantic-ui-react'
|
|
||||||
import {useForm} from 'react-hook-form'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {setLogin} from 'reducers/login'
|
|
||||||
import api from 'api'
|
|
||||||
import {findInput} from 'utils'
|
|
||||||
|
|
||||||
const UserSettingsForm = connect((state) => ({login: state.login}), {
|
|
||||||
setLogin,
|
|
||||||
})(function UserSettingsForm({login, setLogin, errors, setErrors}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const {register, handleSubmit} = useForm()
|
|
||||||
const [loading, setLoading] = React.useState(false)
|
|
||||||
|
|
||||||
const onSave = React.useCallback(
|
|
||||||
async (changes) => {
|
|
||||||
setLoading(true)
|
|
||||||
setErrors(null)
|
|
||||||
try {
|
|
||||||
const response = await api.put('/user', {body: changes})
|
|
||||||
setLogin(response)
|
|
||||||
} catch (err) {
|
|
||||||
setErrors(err.errors)
|
|
||||||
} finally {
|
|
||||||
setLoading(false)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
[setLoading, setLogin, setErrors]
|
|
||||||
)
|
|
||||||
|
|
||||||
return (
|
|
||||||
<Segment style={{maxWidth: 600}}>
|
|
||||||
<Form onSubmit={handleSubmit(onSave)} loading={loading}>
|
|
||||||
<Form.Field error={errors?.username}>
|
|
||||||
<label>{t('SettingsPage.profile.username.label')}</label>
|
|
||||||
<Ref innerRef={findInput(register)}>
|
|
||||||
<Input name="username" defaultValue={login.username} disabled />
|
|
||||||
</Ref>
|
|
||||||
<small>{t('SettingsPage.profile.username.hint')}</small>
|
|
||||||
</Form.Field>
|
|
||||||
|
|
||||||
<Message info visible>
|
|
||||||
{t('SettingsPage.profile.publicNotice')}
|
|
||||||
</Message>
|
|
||||||
|
|
||||||
<Form.Field error={errors?.displayName}>
|
|
||||||
<label>{t('SettingsPage.profile.displayName.label')}</label>
|
|
||||||
<Ref innerRef={findInput(register)}>
|
|
||||||
<Input name="displayName" defaultValue={login.displayName} placeholder={login.username} />
|
|
||||||
</Ref>
|
|
||||||
<small>{t('SettingsPage.profile.displayName.fallbackNotice')}</small>
|
|
||||||
</Form.Field>
|
|
||||||
|
|
||||||
<Form.Field error={errors?.bio}>
|
|
||||||
<label>{t('SettingsPage.profile.bio.label')}</label>
|
|
||||||
<Ref innerRef={register}>
|
|
||||||
<TextArea name="bio" rows={4} defaultValue={login.bio} />
|
|
||||||
</Ref>
|
|
||||||
</Form.Field>
|
|
||||||
<Form.Field error={errors?.image}>
|
|
||||||
<label>{t('SettingsPage.profile.avatarUrl.label')}</label>
|
|
||||||
<Ref innerRef={findInput(register)}>
|
|
||||||
<Input name="image" defaultValue={login.image} />
|
|
||||||
</Ref>
|
|
||||||
</Form.Field>
|
|
||||||
|
|
||||||
<Button type="submit" primary>
|
|
||||||
{t('general.save')}
|
|
||||||
</Button>
|
|
||||||
</Form>
|
|
||||||
</Segment>
|
|
||||||
)
|
|
||||||
})
|
|
||||||
export default UserSettingsForm
|
|
|
@ -1,68 +0,0 @@
|
||||||
import React from 'react'
|
|
||||||
import {connect} from 'react-redux'
|
|
||||||
import {Header, Tab} from 'semantic-ui-react'
|
|
||||||
import {useForm} from 'react-hook-form'
|
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {setLogin} from 'reducers/login'
|
|
||||||
import {Page, Stats} from 'components'
|
|
||||||
import api from 'api'
|
|
||||||
|
|
||||||
import ApiKeySettings from './ApiKeySettings'
|
|
||||||
|
|
||||||
import UserSettingsForm from './UserSettingsForm'
|
|
||||||
import DeviceList from './DeviceList'
|
|
||||||
|
|
||||||
const SettingsPage = connect((state) => ({login: state.login}), {setLogin})(function SettingsPage({login, setLogin}) {
|
|
||||||
const {t} = useTranslation()
|
|
||||||
const {register, handleSubmit} = useForm()
|
|
||||||
const [loading, setLoading] = React.useState(false)
|
|
||||||
const [errors, setErrors] = React.useState(null)
|
|
||||||
|
|
||||||
const onGenerateNewKey = React.useCallback(async () => {
|
|
||||||
setLoading(true)
|
|
||||||
setErrors(null)
|
|
||||||
try {
|
|
||||||
const response = await api.put('/user', {
|
|
||||||
body: {updateApiKey: true},
|
|
||||||
})
|
|
||||||
setLogin(response)
|
|
||||||
} catch (err) {
|
|
||||||
setErrors(err.errors)
|
|
||||||
} finally {
|
|
||||||
setLoading(false)
|
|
||||||
}
|
|
||||||
}, [setLoading, setLogin, setErrors])
|
|
||||||
|
|
||||||
return (
|
|
||||||
<Page title={t('SettingsPage.title')}>
|
|
||||||
<Header as="h1">{t('SettingsPage.title')}</Header>
|
|
||||||
<Tab
|
|
||||||
menu={{secondary: true, pointing: true}}
|
|
||||||
panes={[
|
|
||||||
{
|
|
||||||
menuItem: t('SettingsPage.profile.title'),
|
|
||||||
render: () => <UserSettingsForm {...{errors, setErrors}} />,
|
|
||||||
},
|
|
||||||
|
|
||||||
{
|
|
||||||
menuItem: t('SettingsPage.apiKey.title'),
|
|
||||||
render: () => <ApiKeySettings {...{errors, setErrors}} />,
|
|
||||||
},
|
|
||||||
|
|
||||||
{
|
|
||||||
menuItem: t('SettingsPage.stats.title'),
|
|
||||||
render: () => <Stats user={login.id} />,
|
|
||||||
},
|
|
||||||
|
|
||||||
{
|
|
||||||
menuItem: t('SettingsPage.devices.title'),
|
|
||||||
render: () => <DeviceList />,
|
|
||||||
},
|
|
||||||
]}
|
|
||||||
/>
|
|
||||||
</Page>
|
|
||||||
)
|
|
||||||
})
|
|
||||||
|
|
||||||
export default SettingsPage
|
|
|
@ -21,8 +21,6 @@ import {pluck, distinctUntilChanged, map, switchMap} from 'rxjs/operators'
|
||||||
import {useObservable} from 'rxjs-hooks'
|
import {useObservable} from 'rxjs-hooks'
|
||||||
import {findInput} from 'utils'
|
import {findInput} from 'utils'
|
||||||
import {useForm, Controller} from 'react-hook-form'
|
import {useForm, Controller} from 'react-hook-form'
|
||||||
import {useTranslation, Trans as Translate} from 'react-i18next'
|
|
||||||
import Markdown from 'react-markdown'
|
|
||||||
|
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
import {Page, FileUploadField} from 'components'
|
import {Page, FileUploadField} from 'components'
|
||||||
|
@ -31,21 +29,18 @@ import type {Track} from 'types'
|
||||||
import {FileUploadStatus} from 'pages/UploadPage'
|
import {FileUploadStatus} from 'pages/UploadPage'
|
||||||
|
|
||||||
function ReplaceTrackData({slug}) {
|
function ReplaceTrackData({slug}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const [file, setFile] = React.useState(null)
|
const [file, setFile] = React.useState(null)
|
||||||
const [result, setResult] = React.useState(null)
|
const [result, setResult] = React.useState(null)
|
||||||
const onComplete = React.useCallback((_id, r) => setResult(r), [setResult])
|
const onComplete = React.useCallback((_id, r) => setResult(r), [setResult])
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
<Header as="h2">{t('TrackEditor.replaceTrackData')}</Header>
|
<Header as="h2">Replace track data</Header>
|
||||||
{!file ? (
|
{!file ? (
|
||||||
<FileUploadField onSelect={setFile} />
|
<FileUploadField onSelect={setFile} />
|
||||||
) : result ? (
|
) : result ? (
|
||||||
<Message>
|
<Message>
|
||||||
<Translate i18nKey="TrackEditor.uploadComplete">
|
Upload complete. <Link to={`/tracks/${slug}`}>Show track</Link>
|
||||||
Upload complete. <Link to={`/tracks/${slug}`}>Show track</Link>
|
|
||||||
</Translate>
|
|
||||||
</Message>
|
</Message>
|
||||||
) : (
|
) : (
|
||||||
<FileUploadStatus {...{file, onComplete, slug}} />
|
<FileUploadStatus {...{file, onComplete, slug}} />
|
||||||
|
@ -55,7 +50,6 @@ function ReplaceTrackData({slug}) {
|
||||||
}
|
}
|
||||||
|
|
||||||
const TrackEditor = connect((state) => ({login: state.login}))(function TrackEditor({login}) {
|
const TrackEditor = connect((state) => ({login: state.login}))(function TrackEditor({login}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const [busy, setBusy] = React.useState(false)
|
const [busy, setBusy] = React.useState(false)
|
||||||
const {register, control, handleSubmit} = useForm()
|
const {register, control, handleSubmit} = useForm()
|
||||||
const {slug} = useParams()
|
const {slug} = useParams()
|
||||||
|
@ -75,7 +69,7 @@ const TrackEditor = connect((state) => ({login: state.login}))(function TrackEdi
|
||||||
)
|
)
|
||||||
|
|
||||||
const loading = busy || track == null
|
const loading = busy || track == null
|
||||||
const isAuthor = login?.id === track?.author?.id
|
const isAuthor = login?.username === track?.author?.username
|
||||||
|
|
||||||
// Navigate to track detials if we are not the author
|
// Navigate to track detials if we are not the author
|
||||||
React.useEffect(() => {
|
React.useEffect(() => {
|
||||||
|
@ -90,11 +84,7 @@ const TrackEditor = connect((state) => ({login: state.login}))(function TrackEdi
|
||||||
setBusy(true)
|
setBusy(true)
|
||||||
|
|
||||||
try {
|
try {
|
||||||
await api.put(`/tracks/${slug}`, {
|
await api.put(`/tracks/${slug}`, {body: {track: _.pickBy(values, (v) => typeof v !== 'undefined')}})
|
||||||
body: {
|
|
||||||
track: _.pickBy(values, (v) => typeof v !== 'undefined'),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
history.push(`/tracks/${slug}`)
|
history.push(`/tracks/${slug}`)
|
||||||
} finally {
|
} finally {
|
||||||
setBusy(false)
|
setBusy(false)
|
||||||
|
@ -116,9 +106,7 @@ const TrackEditor = connect((state) => ({login: state.login}))(function TrackEdi
|
||||||
}
|
}
|
||||||
}, [setBusy, setConfirmDelete, slug, history])
|
}, [setBusy, setConfirmDelete, slug, history])
|
||||||
|
|
||||||
const trackTitle: string = track?.title || t('general.unnamedTrack')
|
const title = `Edit ${track ? track.title || 'Unnamed track' : 'track'}`
|
||||||
const title = t('TrackEditor.title', {trackTitle})
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Page title={title}>
|
<Page title={title}>
|
||||||
<Grid centered relaxed divided stackable>
|
<Grid centered relaxed divided stackable>
|
||||||
|
@ -131,22 +119,14 @@ const TrackEditor = connect((state) => ({login: state.login}))(function TrackEdi
|
||||||
</Ref>
|
</Ref>
|
||||||
|
|
||||||
<Form.Field>
|
<Form.Field>
|
||||||
<label>{t('TrackEditor.description.label')}</label>
|
<label>Description</label>
|
||||||
<Ref innerRef={register}>
|
<Ref innerRef={register}>
|
||||||
<TextArea name="description" rows={4} defaultValue={track?.description} />
|
<TextArea name="description" rows={4} defaultValue={track?.description} />
|
||||||
</Ref>
|
</Ref>
|
||||||
</Form.Field>
|
</Form.Field>
|
||||||
|
|
||||||
<Form.Field>
|
<Form.Field>
|
||||||
<label>
|
<label>Visibility</label>
|
||||||
{t('TrackEditor.visibility.label')}
|
|
||||||
<Popup
|
|
||||||
wide="very"
|
|
||||||
content={<Markdown>{t('TrackEditor.visibility.description')}</Markdown>}
|
|
||||||
trigger={<Icon name="warning sign" style={{marginLeft: 8}} color="orange" />}
|
|
||||||
/>
|
|
||||||
</label>
|
|
||||||
|
|
||||||
<Controller
|
<Controller
|
||||||
name="public"
|
name="public"
|
||||||
control={control}
|
control={control}
|
||||||
|
@ -154,14 +134,42 @@ const TrackEditor = connect((state) => ({login: state.login}))(function TrackEdi
|
||||||
render={(props) => (
|
render={(props) => (
|
||||||
<Checkbox
|
<Checkbox
|
||||||
name="public"
|
name="public"
|
||||||
label={t('TrackEditor.visibility.checkboxLabel')}
|
label="Make track public (in track list and details page)"
|
||||||
checked={props.value}
|
checked={props.value}
|
||||||
onChange={(_, {checked}) => props.onChange(checked)}
|
onChange={(_, {checked}) => props.onChange(checked)}
|
||||||
/>
|
/>
|
||||||
)}
|
)}
|
||||||
/>
|
/>
|
||||||
|
|
||||||
|
<Popup
|
||||||
|
wide="very"
|
||||||
|
content={
|
||||||
|
<>
|
||||||
|
<p>
|
||||||
|
Checking this box allows all users to see your full track. For your own privacy and security,
|
||||||
|
make sure to only publish tracks in this way that do not let others deduce where you live, work,
|
||||||
|
or frequently stay. Your recording device might have useful privacy settings to not record
|
||||||
|
geolocation data near those places.
|
||||||
|
</p>
|
||||||
|
<p>
|
||||||
|
In the future, this site will allow you to redact privacy sensitive data in tracks, both
|
||||||
|
manually and automatically. Until then, you will have to rely on the features of your recording
|
||||||
|
device, or manually redact your files before upload.
|
||||||
|
</p>
|
||||||
|
<p>
|
||||||
|
After checking this box, your data essentially becomes public. You understand that we cannot
|
||||||
|
control who potentially downloads this data and and keeps a copy, even if you delete it from
|
||||||
|
your account or anonymize it later.
|
||||||
|
</p>
|
||||||
|
<p>
|
||||||
|
<b>Use at your own risk.</b>
|
||||||
|
</p>
|
||||||
|
</>
|
||||||
|
}
|
||||||
|
trigger={<Icon name="warning sign" style={{marginLeft: 8}} color="orange" />}
|
||||||
|
/>
|
||||||
</Form.Field>
|
</Form.Field>
|
||||||
<Button type="submit">{t('general.save')}</Button>
|
<Button type="submit">Save</Button>
|
||||||
</Form>
|
</Form>
|
||||||
</Grid.Column>
|
</Grid.Column>
|
||||||
<Grid.Column width={6}>
|
<Grid.Column width={6}>
|
||||||
|
@ -169,20 +177,16 @@ const TrackEditor = connect((state) => ({login: state.login}))(function TrackEdi
|
||||||
|
|
||||||
<Divider />
|
<Divider />
|
||||||
|
|
||||||
<Header as="h2">{t('TrackEditor.dangerZone.title')}</Header>
|
<Header as="h2">Danger zone</Header>
|
||||||
<Markdown>{t('TrackEditor.dangerZone.description')}</Markdown>
|
<p>
|
||||||
|
You can remove this track from your account and the portal if you like. However, if at any point you have
|
||||||
|
published this track, we cannot guarantee that there are no versions of it in the public data repository,
|
||||||
|
or any copy thereof.
|
||||||
|
</p>
|
||||||
<Button color="red" onClick={() => setConfirmDelete(true)}>
|
<Button color="red" onClick={() => setConfirmDelete(true)}>
|
||||||
{t('general.delete')}
|
Delete
|
||||||
</Button>
|
</Button>
|
||||||
<Confirm
|
<Confirm open={confirmDelete} onCancel={() => setConfirmDelete(false)} onConfirm={onDelete} />
|
||||||
open={confirmDelete}
|
|
||||||
onCancel={() => setConfirmDelete(false)}
|
|
||||||
onConfirm={onDelete}
|
|
||||||
content={t('TrackEditor.dangerZone.confirmDelete')}
|
|
||||||
confirmButton={t('general.delete')}
|
|
||||||
cancelButton={t('general.cancel')}
|
|
||||||
/>
|
|
||||||
</Grid.Column>
|
</Grid.Column>
|
||||||
</Grid.Row>
|
</Grid.Row>
|
||||||
</Grid>
|
</Grid>
|
||||||
|
|
|
@ -1,44 +1,40 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import {Link} from 'react-router-dom'
|
import {Link} from 'react-router-dom'
|
||||||
import {Icon, Popup, Button, Dropdown} from 'semantic-ui-react'
|
import {Icon, Popup, Button, Dropdown} from 'semantic-ui-react'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
export default function TrackActions({slug, isAuthor, onDownload}) {
|
export default function TrackActions({slug, isAuthor, onDownload}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
|
{isAuthor && (
|
||||||
|
<Link to={`/tracks/${slug}/edit`}>
|
||||||
|
<Button primary>Edit track</Button>
|
||||||
|
</Link>
|
||||||
|
)}
|
||||||
|
|
||||||
|
<Dropdown text="Download" button upward>
|
||||||
|
<Dropdown.Menu>
|
||||||
|
<Dropdown.Item text="Original" onClick={() => onDownload('original.csv')} disabled={!isAuthor} />
|
||||||
|
<Dropdown.Item text="Track (GPX)" onClick={() => onDownload('track.gpx')} />
|
||||||
|
</Dropdown.Menu>
|
||||||
|
</Dropdown>
|
||||||
|
|
||||||
<Popup
|
<Popup
|
||||||
trigger={<Icon name="info circle" />}
|
trigger={<Icon name="info circle" />}
|
||||||
offset={[12, 0]}
|
offset={[12, 0]}
|
||||||
content={
|
content={
|
||||||
isAuthor ? (
|
isAuthor ? (
|
||||||
<>
|
<>
|
||||||
<p>{t('TrackPage.actions.hintAuthorOnly')}</p>
|
<p>Only you, the author of this track, can download the original file.</p>
|
||||||
<p>{t('TrackPage.actions.hintOriginal')}</p>
|
<p>
|
||||||
|
This is the file as it was uploaded to the server, without modifications, and it can be used with other
|
||||||
|
tools.
|
||||||
|
</p>
|
||||||
</>
|
</>
|
||||||
) : (
|
) : (
|
||||||
<p>{t('TrackPage.actions.hintAuthorOnlyOthers')}</p>
|
<p>Only the author of this track can download the original file.</p>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
/>
|
/>
|
||||||
|
|
||||||
<Dropdown text={t('TrackPage.actions.download')} button>
|
|
||||||
<Dropdown.Menu>
|
|
||||||
<Dropdown.Item
|
|
||||||
text={t('TrackPage.actions.original')}
|
|
||||||
onClick={() => onDownload('original.csv')}
|
|
||||||
disabled={!isAuthor}
|
|
||||||
/>
|
|
||||||
<Dropdown.Item text={t('TrackPage.actions.gpx')} onClick={() => onDownload('track.gpx')} />
|
|
||||||
</Dropdown.Menu>
|
|
||||||
</Dropdown>
|
|
||||||
|
|
||||||
{isAuthor && (
|
|
||||||
<Link to={`/tracks/${slug}/edit`}>
|
|
||||||
<Button primary>{t('TrackPage.actions.edit')}</Button>
|
|
||||||
</Link>
|
|
||||||
)}
|
|
||||||
</>
|
</>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,12 +1,10 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import {Message, Segment, Form, Button, Loader, Header, Comment} from 'semantic-ui-react'
|
import {Message, Segment, Form, Button, Loader, Header, Comment} from 'semantic-ui-react'
|
||||||
import Markdown from 'react-markdown'
|
import Markdown from 'react-markdown'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {Avatar, FormattedDate} from 'components'
|
import {Avatar, FormattedDate} from 'components'
|
||||||
|
|
||||||
function CommentForm({onSubmit}) {
|
function CommentForm({onSubmit}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const [body, setBody] = React.useState('')
|
const [body, setBody] = React.useState('')
|
||||||
|
|
||||||
const onSubmitComment = React.useCallback(() => {
|
const onSubmitComment = React.useCallback(() => {
|
||||||
|
@ -17,18 +15,17 @@ function CommentForm({onSubmit}) {
|
||||||
return (
|
return (
|
||||||
<Form reply onSubmit={onSubmitComment}>
|
<Form reply onSubmit={onSubmitComment}>
|
||||||
<Form.TextArea rows={4} value={body} onChange={(e) => setBody(e.target.value)} />
|
<Form.TextArea rows={4} value={body} onChange={(e) => setBody(e.target.value)} />
|
||||||
<Button content={t('TrackPage.comments.post')} labelPosition="left" icon="edit" primary />
|
<Button content="Post comment" labelPosition="left" icon="edit" primary />
|
||||||
</Form>
|
</Form>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
export default function TrackComments({comments, onSubmit, onDelete, login, hideLoader}) {
|
export default function TrackComments({comments, onSubmit, onDelete, login, hideLoader}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
return (
|
return (
|
||||||
<>
|
<Segment basic>
|
||||||
<Comment.Group>
|
<Comment.Group>
|
||||||
<Header as="h2" dividing>
|
<Header as="h2" dividing>
|
||||||
{t('TrackPage.comments.title')}
|
Comments
|
||||||
</Header>
|
</Header>
|
||||||
|
|
||||||
<Loader active={!hideLoader && comments == null} inline />
|
<Loader active={!hideLoader && comments == null} inline />
|
||||||
|
@ -37,7 +34,7 @@ export default function TrackComments({comments, onSubmit, onDelete, login, hide
|
||||||
<Comment key={comment.id}>
|
<Comment key={comment.id}>
|
||||||
<Avatar user={comment.author} />
|
<Avatar user={comment.author} />
|
||||||
<Comment.Content>
|
<Comment.Content>
|
||||||
<Comment.Author as="a">{comment.author.displayName}</Comment.Author>
|
<Comment.Author as="a">{comment.author.username}</Comment.Author>
|
||||||
<Comment.Metadata>
|
<Comment.Metadata>
|
||||||
<div>
|
<div>
|
||||||
<FormattedDate date={comment.createdAt} relative />
|
<FormattedDate date={comment.createdAt} relative />
|
||||||
|
@ -46,7 +43,7 @@ export default function TrackComments({comments, onSubmit, onDelete, login, hide
|
||||||
<Comment.Text>
|
<Comment.Text>
|
||||||
<Markdown>{comment.body}</Markdown>
|
<Markdown>{comment.body}</Markdown>
|
||||||
</Comment.Text>
|
</Comment.Text>
|
||||||
{login?.id === comment.author.id && (
|
{login?.username === comment.author.username && (
|
||||||
<Comment.Actions>
|
<Comment.Actions>
|
||||||
<Comment.Action
|
<Comment.Action
|
||||||
onClick={(e) => {
|
onClick={(e) => {
|
||||||
|
@ -54,7 +51,7 @@ export default function TrackComments({comments, onSubmit, onDelete, login, hide
|
||||||
e.preventDefault()
|
e.preventDefault()
|
||||||
}}
|
}}
|
||||||
>
|
>
|
||||||
{t('general.delete')}
|
Delete
|
||||||
</Comment.Action>
|
</Comment.Action>
|
||||||
</Comment.Actions>
|
</Comment.Actions>
|
||||||
)}
|
)}
|
||||||
|
@ -62,10 +59,10 @@ export default function TrackComments({comments, onSubmit, onDelete, login, hide
|
||||||
</Comment>
|
</Comment>
|
||||||
))}
|
))}
|
||||||
|
|
||||||
{comments != null && !comments.length && <Message>{t('TrackPage.comments.empty')}</Message>}
|
{comments != null && !comments.length && <Message>Nobody commented... yet</Message>}
|
||||||
|
|
||||||
{login && comments != null && <CommentForm onSubmit={onSubmit} />}
|
{login && comments != null && <CommentForm onSubmit={onSubmit} />}
|
||||||
</Comment.Group>
|
</Comment.Group>
|
||||||
</>
|
</Segment>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,54 +1,80 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import _ from 'lodash'
|
import {List} from 'semantic-ui-react'
|
||||||
import {List, Header, Grid} from 'semantic-ui-react'
|
|
||||||
import {Duration} from 'luxon'
|
import {Duration} from 'luxon'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {FormattedDate, Visibility} from 'components'
|
import {FormattedDate} from 'components'
|
||||||
import {formatDistance, formatDuration} from 'utils'
|
|
||||||
|
function formatDuration(seconds) {
|
||||||
|
return Duration.fromMillis((seconds ?? 0) * 1000).toFormat("h'h' mm'm'")
|
||||||
|
}
|
||||||
|
|
||||||
export default function TrackDetails({track, isAuthor}) {
|
export default function TrackDetails({track, isAuthor}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
const items = [
|
|
||||||
track.public != null && isAuthor && [t('TrackPage.details.visibility'), <Visibility public={track.public} />],
|
|
||||||
|
|
||||||
track.uploadedByUserAgent != null && [t('TrackPage.details.uploadedWith'), track.uploadedByUserAgent],
|
|
||||||
|
|
||||||
track.duration != null && [t('TrackPage.details.duration'), formatDuration(track.duration)],
|
|
||||||
|
|
||||||
track.createdAt != null && [t('TrackPage.details.uploadedDate'), <FormattedDate date={track.createdAt} />],
|
|
||||||
|
|
||||||
track?.recordedAt != null && [t('TrackPage.details.recordedDate'), <FormattedDate date={track?.recordedAt} />],
|
|
||||||
|
|
||||||
track?.numEvents != null && [t('TrackPage.details.numEvents'), track?.numEvents],
|
|
||||||
|
|
||||||
track?.length != null && [t('TrackPage.details.length'), formatDistance(track?.length)],
|
|
||||||
|
|
||||||
track?.processingStatus != null &&
|
|
||||||
track?.processingStatus != 'error' && [t('TrackPage.details.processingStatus'), track.processingStatus],
|
|
||||||
|
|
||||||
track.originalFileName != null && [t('TrackPage.details.originalFileName'), <code>{track.originalFileName}</code>],
|
|
||||||
].filter(Boolean)
|
|
||||||
|
|
||||||
const COLUMNS = 4
|
|
||||||
const chunkSize = Math.ceil(items.length / COLUMNS)
|
|
||||||
return (
|
return (
|
||||||
<Grid>
|
<List>
|
||||||
<Grid.Row columns={COLUMNS}>
|
{track.public != null && isAuthor && (
|
||||||
{_.chunk(items, chunkSize).map((chunkItems, idx) => (
|
<List.Item>
|
||||||
<Grid.Column key={idx}>
|
<List.Header>Visibility</List.Header>
|
||||||
<List>
|
{track.public ? 'Public' : 'Private'}
|
||||||
{chunkItems.map(([title, value]) => (
|
</List.Item>
|
||||||
<List.Item key={title}>
|
)}
|
||||||
<List.Header>{title}</List.Header>
|
|
||||||
<List.Description>{value}</List.Description>
|
{track.originalFileName != null && (
|
||||||
</List.Item>
|
<List.Item>
|
||||||
))}
|
{isAuthor && <div style={{float: 'right'}}></div>}
|
||||||
</List>
|
|
||||||
</Grid.Column>
|
<List.Header>Original Filename</List.Header>
|
||||||
))}
|
<code>{track.originalFileName}</code>
|
||||||
</Grid.Row>
|
</List.Item>
|
||||||
</Grid>
|
)}
|
||||||
|
|
||||||
|
{track.uploadedByUserAgent != null && (
|
||||||
|
<List.Item>
|
||||||
|
<List.Header>Uploaded with</List.Header>
|
||||||
|
{track.uploadedByUserAgent}
|
||||||
|
</List.Item>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{track.duration != null && (
|
||||||
|
<List.Item>
|
||||||
|
<List.Header>Duration</List.Header>
|
||||||
|
{formatDuration(track.duration)}
|
||||||
|
</List.Item>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{track.createdAt != null && (
|
||||||
|
<List.Item>
|
||||||
|
<List.Header>Uploaded on</List.Header>
|
||||||
|
<FormattedDate date={track.createdAt} />
|
||||||
|
</List.Item>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{track?.recordedAt != null && (
|
||||||
|
<List.Item>
|
||||||
|
<List.Header>Recorded on</List.Header>
|
||||||
|
<FormattedDate date={track?.recordedAt} />
|
||||||
|
</List.Item>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{track?.numEvents != null && (
|
||||||
|
<List.Item>
|
||||||
|
<List.Header>Confirmed events</List.Header>
|
||||||
|
{track?.numEvents}
|
||||||
|
</List.Item>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{track?.length != null && (
|
||||||
|
<List.Item>
|
||||||
|
<List.Header>Length</List.Header>
|
||||||
|
{(track?.length / 1000).toFixed(2)} km
|
||||||
|
</List.Item>
|
||||||
|
)}
|
||||||
|
|
||||||
|
{track?.processingStatus != null && track?.processingStatus != 'error' && (
|
||||||
|
<List.Item>
|
||||||
|
<List.Header>Processing</List.Header>
|
||||||
|
{track.processingStatus}
|
||||||
|
</List.Item>
|
||||||
|
)}
|
||||||
|
</List>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
|
@ -61,6 +61,7 @@ export default function TrackMap({
|
||||||
layout: {
|
layout: {
|
||||||
'text-field': ['number-format', ['get', p], {'min-fraction-digits': 2, 'max-fraction-digits': 2}],
|
'text-field': ['number-format', ['get', p], {'min-fraction-digits': 2, 'max-fraction-digits': 2}],
|
||||||
'text-allow-overlap': true,
|
'text-allow-overlap': true,
|
||||||
|
'text-font': ['Open Sans Bold', 'Arial Unicode MS Regular'],
|
||||||
'text-size': 14,
|
'text-size': 14,
|
||||||
'text-keep-upright': false,
|
'text-keep-upright': false,
|
||||||
'text-anchor': a,
|
'text-anchor': a,
|
||||||
|
|
|
@ -9,9 +9,10 @@
|
||||||
.details.details {
|
.details.details {
|
||||||
position: absolute;
|
position: absolute;
|
||||||
width: 320px;
|
width: 320px;
|
||||||
top: 32px;
|
top: 16px;
|
||||||
right: 16px;
|
right: 16px;
|
||||||
max-height: calc(100% - 32px);
|
max-height: calc(100% - 32px);
|
||||||
|
overflow: auto;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,24 +1,11 @@
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import {connect} from 'react-redux'
|
import {connect} from 'react-redux'
|
||||||
import {
|
import {List, Dropdown, Checkbox, Segment, Dimmer, Grid, Loader, Header, Message, Confirm} from 'semantic-ui-react'
|
||||||
List,
|
|
||||||
Dropdown,
|
|
||||||
Checkbox,
|
|
||||||
Segment,
|
|
||||||
Dimmer,
|
|
||||||
Grid,
|
|
||||||
Loader,
|
|
||||||
Header,
|
|
||||||
Message,
|
|
||||||
Confirm,
|
|
||||||
Container,
|
|
||||||
} from 'semantic-ui-react'
|
|
||||||
import {useParams, useHistory} from 'react-router-dom'
|
import {useParams, useHistory} from 'react-router-dom'
|
||||||
import {concat, combineLatest, of, from, Subject} from 'rxjs'
|
import {concat, combineLatest, of, from, Subject} from 'rxjs'
|
||||||
import {pluck, distinctUntilChanged, map, switchMap, startWith, catchError} from 'rxjs/operators'
|
import {pluck, distinctUntilChanged, map, switchMap, startWith, catchError} from 'rxjs/operators'
|
||||||
import {useObservable} from 'rxjs-hooks'
|
import {useObservable} from 'rxjs-hooks'
|
||||||
import Markdown from 'react-markdown'
|
import Markdown from 'react-markdown'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
import {Page} from 'components'
|
import {Page} from 'components'
|
||||||
|
@ -39,79 +26,43 @@ function useTriggerSubject() {
|
||||||
}
|
}
|
||||||
|
|
||||||
function TrackMapSettings({showTrack, setShowTrack, pointsMode, setPointsMode, side, setSide}) {
|
function TrackMapSettings({showTrack, setShowTrack, pointsMode, setPointsMode, side, setSide}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
<Header as="h4">{t('TrackPage.mapSettings.title')}</Header>
|
<Header as="h4">Map settings</Header>
|
||||||
<List>
|
<List>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<Checkbox checked={showTrack} onChange={(e, d) => setShowTrack(d.checked)} />{' '}
|
<Checkbox checked={showTrack} onChange={(e, d) => setShowTrack(d.checked)} /> Show track
|
||||||
{t('TrackPage.mapSettings.showTrack')}
|
|
||||||
<div style={{marginTop: 8}}>
|
<div style={{marginTop: 8}}>
|
||||||
<span
|
<span style={{background: trackLayerRaw.paint['line-color'], height: 6, width: 24, display: 'inline-block', verticalAlign: 'middle', marginRight: 4}} />
|
||||||
style={{
|
GPS track
|
||||||
borderTop: '3px dashed ' + trackLayerRaw.paint['line-color'],
|
|
||||||
height: 0,
|
|
||||||
width: 24,
|
|
||||||
display: 'inline-block',
|
|
||||||
verticalAlign: 'middle',
|
|
||||||
marginRight: 4,
|
|
||||||
}}
|
|
||||||
/>
|
|
||||||
{t('TrackPage.mapSettings.gpsTrack')}
|
|
||||||
</div>
|
</div>
|
||||||
<div>
|
<div>
|
||||||
<span
|
<span style={{background: trackLayer.paint['line-color'], height: 6, width: 24, display: 'inline-block', verticalAlign: 'middle', marginRight: 4}} />
|
||||||
style={{
|
Snapped to road
|
||||||
borderTop: '6px solid ' + trackLayerRaw.paint['line-color'],
|
|
||||||
height: 6,
|
|
||||||
width: 24,
|
|
||||||
display: 'inline-block',
|
|
||||||
verticalAlign: 'middle',
|
|
||||||
marginRight: 4,
|
|
||||||
}}
|
|
||||||
/>
|
|
||||||
{t('TrackPage.mapSettings.snappedTrack')}
|
|
||||||
</div>
|
</div>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header> {t('TrackPage.mapSettings.points')} </List.Header>
|
<List.Header>Points</List.Header>
|
||||||
<Dropdown
|
<Dropdown
|
||||||
selection
|
selection
|
||||||
value={pointsMode}
|
value={pointsMode}
|
||||||
onChange={(e, d) => setPointsMode(d.value)}
|
onChange={(e, d) => setPointsMode(d.value)}
|
||||||
options={[
|
options={[
|
||||||
{key: 'none', value: 'none', text: 'None'},
|
{key: 'none', value: 'none', text: 'None'},
|
||||||
{
|
{key: 'overtakingEvents', value: 'overtakingEvents', text: 'Confirmed'},
|
||||||
key: 'overtakingEvents',
|
{key: 'measurements', value: 'measurements', text: 'All measurements'},
|
||||||
value: 'overtakingEvents',
|
|
||||||
text: t('TrackPage.mapSettings.confirmedPoints'),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'measurements',
|
|
||||||
value: 'measurements',
|
|
||||||
text: t('TrackPage.mapSettings.allPoints'),
|
|
||||||
},
|
|
||||||
]}
|
]}
|
||||||
/>
|
/>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
<List.Item>
|
<List.Item>
|
||||||
<List.Header>{t('TrackPage.mapSettings.side')}</List.Header>
|
<List.Header>Side (for color)</List.Header>
|
||||||
<Dropdown
|
<Dropdown
|
||||||
selection
|
selection
|
||||||
value={side}
|
value={side}
|
||||||
onChange={(e, d) => setSide(d.value)}
|
onChange={(e, d) => setSide(d.value)}
|
||||||
options={[
|
options={[
|
||||||
{
|
{key: 'overtaker', value: 'overtaker', text: 'Overtaker (Left)'},
|
||||||
key: 'overtaker',
|
{key: 'stationary', value: 'stationary', text: 'Stationary (Right)'},
|
||||||
value: 'overtaker',
|
|
||||||
text: t('TrackPage.mapSettings.overtakerSide'),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'stationary',
|
|
||||||
value: 'stationary',
|
|
||||||
text: t('TrackPage.mapSettings.stationarySide'),
|
|
||||||
},
|
|
||||||
]}
|
]}
|
||||||
/>
|
/>
|
||||||
</List.Item>
|
</List.Item>
|
||||||
|
@ -122,7 +73,6 @@ function TrackMapSettings({showTrack, setShowTrack, pointsMode, setPointsMode, s
|
||||||
|
|
||||||
const TrackPage = connect((state) => ({login: state.login}))(function TrackPage({login}) {
|
const TrackPage = connect((state) => ({login: state.login}))(function TrackPage({login}) {
|
||||||
const {slug} = useParams()
|
const {slug} = useParams()
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
const [reloadComments, reloadComments$] = useTriggerSubject()
|
const [reloadComments, reloadComments$] = useTriggerSubject()
|
||||||
const history = useHistory()
|
const history = useHistory()
|
||||||
|
@ -179,11 +129,7 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
)
|
)
|
||||||
|
|
||||||
return combineLatest([track$, trackData$, comments$]).pipe(
|
return combineLatest([track$, trackData$, comments$]).pipe(
|
||||||
map(([track, trackData, comments]) => ({
|
map(([track, trackData, comments]) => ({track, trackData, comments}))
|
||||||
track,
|
|
||||||
trackData,
|
|
||||||
comments,
|
|
||||||
}))
|
|
||||||
)
|
)
|
||||||
},
|
},
|
||||||
null,
|
null,
|
||||||
|
@ -216,7 +162,9 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
await api.downloadFile(`/tracks/${slug}/download/${filename}`)
|
await api.downloadFile(`/tracks/${slug}/download/${filename}`)
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
if (/Failed to fetch/.test(String(err))) {
|
if (/Failed to fetch/.test(String(err))) {
|
||||||
setDownloadError(t('TrackPage.downloadError'))
|
setDownloadError(
|
||||||
|
'The track probably has not been imported correctly or recently enough. Please ask your administrator for assistance.'
|
||||||
|
)
|
||||||
} else {
|
} else {
|
||||||
setDownloadError(String(err))
|
setDownloadError(String(err))
|
||||||
}
|
}
|
||||||
|
@ -225,7 +173,7 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
[slug]
|
[slug]
|
||||||
)
|
)
|
||||||
|
|
||||||
const isAuthor = login?.id === data?.track?.author?.id
|
const isAuthor = login?.username === data?.track?.author?.username
|
||||||
|
|
||||||
const {track, trackData, comments} = data || {}
|
const {track, trackData, comments} = data || {}
|
||||||
|
|
||||||
|
@ -237,77 +185,63 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
const [pointsMode, setPointsMode] = React.useState('overtakingEvents') // none|overtakingEvents|measurements
|
const [pointsMode, setPointsMode] = React.useState('overtakingEvents') // none|overtakingEvents|measurements
|
||||||
const [side, setSide] = React.useState('overtaker') // overtaker|stationary
|
const [side, setSide] = React.useState('overtaker') // overtaker|stationary
|
||||||
|
|
||||||
const title = track ? track.title || t('general.unnamedTrack') : null
|
const title = track ? track.title || 'Unnamed track' : null
|
||||||
return (
|
return (
|
||||||
<Page
|
<Page
|
||||||
title={title}
|
title={title}
|
||||||
stage={
|
stage={
|
||||||
<>
|
<div className={styles.stage}>
|
||||||
<Container>
|
<Loader active={loading} />
|
||||||
{track && (
|
<Dimmer.Dimmable blurring dimmed={loading}>
|
||||||
<Segment basic>
|
<TrackMap {...{track, trackData, pointsMode, side, showTrack}} style={{height: '80vh'}} />
|
||||||
<div
|
</Dimmer.Dimmable>
|
||||||
style={{
|
|
||||||
display: 'flex',
|
|
||||||
alignItems: 'baseline',
|
|
||||||
marginBlockStart: 32,
|
|
||||||
marginBlockEnd: 16,
|
|
||||||
}}
|
|
||||||
>
|
|
||||||
<Header as="h1">{title}</Header>
|
|
||||||
<div style={{marginLeft: 'auto'}}>
|
|
||||||
<TrackActions {...{isAuthor, onDownload, slug}} />
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div style={{marginBlockEnd: 16}}>
|
<div className={styles.details}>
|
||||||
<TrackDetails {...{track, isAuthor}} />
|
{processing && (
|
||||||
</div>
|
<Message warning>
|
||||||
</Segment>
|
<Message.Content>Track data is still being processed, please reload page in a while.</Message.Content>
|
||||||
|
</Message>
|
||||||
)}
|
)}
|
||||||
</Container>
|
|
||||||
<div className={styles.stage}>
|
|
||||||
<Loader active={loading} />
|
|
||||||
<Dimmer.Dimmable blurring dimmed={loading}>
|
|
||||||
<TrackMap {...{track, trackData, pointsMode, side, showTrack}} style={{height: '80vh'}} />
|
|
||||||
</Dimmer.Dimmable>
|
|
||||||
|
|
||||||
<div className={styles.details}>
|
{error && (
|
||||||
<Segment>
|
<Message error>
|
||||||
<TrackMapSettings
|
<Message.Content>
|
||||||
{...{
|
The processing of this track failed, please ask your site administrator for help in debugging the
|
||||||
showTrack,
|
issue.
|
||||||
setShowTrack,
|
</Message.Content>
|
||||||
pointsMode,
|
</Message>
|
||||||
setPointsMode,
|
)}
|
||||||
side,
|
|
||||||
setSide,
|
|
||||||
}}
|
|
||||||
/>
|
|
||||||
</Segment>
|
|
||||||
|
|
||||||
{processing && (
|
<Segment>
|
||||||
<Message warning>
|
{track && (
|
||||||
<Message.Content>{t('TrackPage.processing')}</Message.Content>
|
<>
|
||||||
</Message>
|
<Header as="h1">{title}</Header>
|
||||||
|
<TrackDetails {...{track, isAuthor}} />
|
||||||
|
<TrackActions {...{isAuthor, onDownload, slug}} />
|
||||||
|
</>
|
||||||
)}
|
)}
|
||||||
|
</Segment>
|
||||||
{error && (
|
|
||||||
<Message error>
|
|
||||||
<Message.Content>{t('TrackPage.processingError')}</Message.Content>
|
|
||||||
</Message>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
</div>
|
</div>
|
||||||
|
</div>
|
||||||
<Container>
|
}
|
||||||
|
>
|
||||||
|
<Confirm
|
||||||
|
open={downloadError != null}
|
||||||
|
cancelButton={false}
|
||||||
|
onConfirm={hideDownloadError}
|
||||||
|
header="Download failed"
|
||||||
|
content={String(downloadError)}
|
||||||
|
/>
|
||||||
|
<Grid stackable>
|
||||||
|
<Grid.Row>
|
||||||
|
<Grid.Column width={12}>
|
||||||
{track?.description && (
|
{track?.description && (
|
||||||
<>
|
<Segment basic>
|
||||||
<Header as="h2" dividing>
|
<Header as="h2" dividing>
|
||||||
{t('TrackPage.description')}
|
Description
|
||||||
</Header>
|
</Header>
|
||||||
<Markdown>{track.description}</Markdown>
|
<Markdown>{track.description}</Markdown>
|
||||||
</>
|
</Segment>
|
||||||
)}
|
)}
|
||||||
|
|
||||||
<TrackComments
|
<TrackComments
|
||||||
|
@ -315,18 +249,14 @@ const TrackPage = connect((state) => ({login: state.login}))(function TrackPage(
|
||||||
onSubmit={onSubmitComment}
|
onSubmit={onSubmitComment}
|
||||||
onDelete={onDeleteComment}
|
onDelete={onDeleteComment}
|
||||||
/>
|
/>
|
||||||
</Container>
|
</Grid.Column>
|
||||||
</>
|
<Grid.Column width={4}>
|
||||||
}
|
<TrackMapSettings {...{showTrack, setShowTrack, pointsMode, setPointsMode, side, setSide}} />
|
||||||
>
|
</Grid.Column>
|
||||||
<Confirm
|
</Grid.Row>
|
||||||
open={downloadError != null}
|
</Grid>
|
||||||
cancelButton={false}
|
|
||||||
onConfirm={hideDownloadError}
|
{/* <pre>{JSON.stringify(data, null, 2)}</pre> */}
|
||||||
header={t('TrackPage.downloadFailed')}
|
|
||||||
content={String(downloadError)}
|
|
||||||
confirmButton={t('general.ok')}
|
|
||||||
/>
|
|
||||||
</Page>
|
</Page>
|
||||||
)
|
)
|
||||||
})
|
})
|
||||||
|
|
|
@ -6,10 +6,9 @@ import {Link} from 'react-router-dom'
|
||||||
import {of, from, concat} from 'rxjs'
|
import {of, from, concat} from 'rxjs'
|
||||||
import {map, switchMap, distinctUntilChanged} from 'rxjs/operators'
|
import {map, switchMap, distinctUntilChanged} from 'rxjs/operators'
|
||||||
import _ from 'lodash'
|
import _ from 'lodash'
|
||||||
import {useTranslation, Trans as Translate} from 'react-i18next'
|
|
||||||
|
|
||||||
import type {Track} from 'types'
|
import type {Track} from 'types'
|
||||||
import {Avatar, Page, StripMarkdown, FormattedDate, Visibility} from 'components'
|
import {Avatar, Page, StripMarkdown} from 'components'
|
||||||
import api from 'api'
|
import api from 'api'
|
||||||
import {useQueryParam} from 'query'
|
import {useQueryParam} from 'query'
|
||||||
|
|
||||||
|
@ -39,11 +38,10 @@ function TrackList({privateTracks}: {privateTracks: boolean}) {
|
||||||
const {tracks, trackCount} = data || {tracks: [], trackCount: 0}
|
const {tracks, trackCount} = data || {tracks: [], trackCount: 0}
|
||||||
const loading = !data
|
const loading = !data
|
||||||
const totalPages = Math.ceil(trackCount / pageSize)
|
const totalPages = Math.ceil(trackCount / pageSize)
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<div>
|
<div>
|
||||||
<Loader content={t('general.loading')} active={loading} />
|
<Loader content="Loading" active={loading} />
|
||||||
{!loading && totalPages > 1 && (
|
{!loading && totalPages > 1 && (
|
||||||
<Pagination
|
<Pagination
|
||||||
activePage={page}
|
activePage={page}
|
||||||
|
@ -59,23 +57,15 @@ function TrackList({privateTracks}: {privateTracks: boolean}) {
|
||||||
))}
|
))}
|
||||||
</Item.Group>
|
</Item.Group>
|
||||||
) : (
|
) : (
|
||||||
<NoPublicTracksMessage />
|
<Message>
|
||||||
|
No public tracks yet. <Link to="/upload">Upload the first!</Link>
|
||||||
|
</Message>
|
||||||
)}
|
)}
|
||||||
</div>
|
</div>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
export function NoPublicTracksMessage() {
|
function maxLength(t, max) {
|
||||||
return (
|
|
||||||
<Message>
|
|
||||||
<Translate i18nKey="TracksPage.noPublicTracks">
|
|
||||||
No public tracks yet. <Link to="/upload">Upload the first!</Link>
|
|
||||||
</Translate>
|
|
||||||
</Message>
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
function maxLength(t: string | null, max: number): string | null {
|
|
||||||
if (t && t.length > max) {
|
if (t && t.length > max) {
|
||||||
return t.substring(0, max) + ' ...'
|
return t.substring(0, max) + ' ...'
|
||||||
} else {
|
} else {
|
||||||
|
@ -92,8 +82,6 @@ const COLOR_BY_STATUS = {
|
||||||
}
|
}
|
||||||
|
|
||||||
export function TrackListItem({track, privateTracks = false}) {
|
export function TrackListItem({track, privateTracks = false}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Item key={track.slug}>
|
<Item key={track.slug}>
|
||||||
<Item.Image size="tiny">
|
<Item.Image size="tiny">
|
||||||
|
@ -101,24 +89,29 @@ export function TrackListItem({track, privateTracks = false}) {
|
||||||
</Item.Image>
|
</Item.Image>
|
||||||
<Item.Content>
|
<Item.Content>
|
||||||
<Item.Header as={Link} to={`/tracks/${track.slug}`}>
|
<Item.Header as={Link} to={`/tracks/${track.slug}`}>
|
||||||
{track.title || t('general.unnamedTrack')}
|
{track.title || 'Unnamed track'}
|
||||||
</Item.Header>
|
</Item.Header>
|
||||||
<Item.Meta>
|
<Item.Meta>
|
||||||
{privateTracks ? null : <span>{t('TracksPage.createdBy', {author: track.author.displayName})}</span>}
|
Created by {track.author.username} on {track.createdAt}
|
||||||
<span>
|
|
||||||
<FormattedDate date={track.createdAt} />
|
|
||||||
</span>
|
|
||||||
</Item.Meta>
|
</Item.Meta>
|
||||||
<Item.Description>
|
<Item.Description>
|
||||||
<StripMarkdown>{maxLength(track.description, 200)}</StripMarkdown>
|
<StripMarkdown>{maxLength(track.description, 200)}</StripMarkdown>
|
||||||
</Item.Description>
|
</Item.Description>
|
||||||
{privateTracks && (
|
{privateTracks && (
|
||||||
<Item.Extra>
|
<Item.Extra>
|
||||||
<Visibility public={track.public} />
|
{track.public ? (
|
||||||
|
<>
|
||||||
|
<Icon color="blue" name="eye" fitted /> Public
|
||||||
|
</>
|
||||||
|
) : (
|
||||||
|
<>
|
||||||
|
<Icon name="eye slash" fitted /> Private
|
||||||
|
</>
|
||||||
|
)}
|
||||||
|
|
||||||
<span style={{marginLeft: '1em'}}>
|
<span style={{marginLeft: '1em'}}>
|
||||||
<Icon color={COLOR_BY_STATUS[track.processingStatus]} name="bolt" fitted />{' '}
|
<Icon color={COLOR_BY_STATUS[track.processingStatus]} name="bolt" fitted /> Processing{' '}
|
||||||
{t(`TracksPage.processing.${track.processingStatus}`)}
|
{track.processingStatus}
|
||||||
</span>
|
</span>
|
||||||
</Item.Extra>
|
</Item.Extra>
|
||||||
)}
|
)}
|
||||||
|
@ -128,7 +121,6 @@ export function TrackListItem({track, privateTracks = false}) {
|
||||||
}
|
}
|
||||||
|
|
||||||
function UploadButton({navigate, ...props}) {
|
function UploadButton({navigate, ...props}) {
|
||||||
const {t} = useTranslation()
|
|
||||||
const onClick = useCallback(
|
const onClick = useCallback(
|
||||||
(e) => {
|
(e) => {
|
||||||
e.preventDefault()
|
e.preventDefault()
|
||||||
|
@ -138,18 +130,16 @@ function UploadButton({navigate, ...props}) {
|
||||||
)
|
)
|
||||||
return (
|
return (
|
||||||
<Button onClick={onClick} {...props} color="green" style={{float: 'right'}}>
|
<Button onClick={onClick} {...props} color="green" style={{float: 'right'}}>
|
||||||
{t('TracksPage.upload')}
|
Upload
|
||||||
</Button>
|
</Button>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
const TracksPage = connect((state) => ({login: (state as any).login}))(function TracksPage({login, privateTracks}) {
|
const TracksPage = connect((state) => ({login: (state as any).login}))(function TracksPage({login, privateTracks}) {
|
||||||
const {t} = useTranslation()
|
const title = privateTracks ? 'My tracks' : 'Public tracks'
|
||||||
const title = privateTracks ? t('TracksPage.titleUser') : t('TracksPage.titlePublic')
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Page title={title}>
|
<Page title={title}>
|
||||||
<Header as="h2">{title}</Header>
|
<Header as='h2'>{title}</Header>
|
||||||
{privateTracks && <Link component={UploadButton} to="/upload" />}
|
{privateTracks && <Link component={UploadButton} to="/upload" />}
|
||||||
<TrackList {...{privateTracks}} />
|
<TrackList {...{privateTracks}} />
|
||||||
</Page>
|
</Page>
|
||||||
|
|
|
@ -1,8 +1,7 @@
|
||||||
import _ from 'lodash'
|
import _ from 'lodash'
|
||||||
import React from 'react'
|
import React from 'react'
|
||||||
import {Header, List, Loader, Table, Icon} from 'semantic-ui-react'
|
import {List, Loader, Table, Icon} from 'semantic-ui-react'
|
||||||
import {Link} from 'react-router-dom'
|
import {Link} from 'react-router-dom'
|
||||||
import {useTranslation} from 'react-i18next'
|
|
||||||
|
|
||||||
import {FileUploadField, Page} from 'components'
|
import {FileUploadField, Page} from 'components'
|
||||||
import type {Track} from 'types'
|
import type {Track} from 'types'
|
||||||
|
@ -99,15 +98,10 @@ export function FileUploadStatus({
|
||||||
[file]
|
[file]
|
||||||
)
|
)
|
||||||
|
|
||||||
const {t} = useTranslation()
|
|
||||||
return (
|
return (
|
||||||
<span>
|
<span>
|
||||||
<Loader inline size="mini" active />{' '}
|
<Loader inline size="mini" active />{' '}
|
||||||
{progress < 1
|
{progress < 1 ? `Uploading ${(progress * 100).toFixed(0)}%` : 'Processing...'}
|
||||||
? t('UploadPage.uploadProgress', {
|
|
||||||
progress: (progress * 100).toFixed(0),
|
|
||||||
})
|
|
||||||
: t('UploadPage.processing')}
|
|
||||||
</span>
|
</span>
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
@ -140,19 +134,15 @@ export default function UploadPage() {
|
||||||
setFiles(files.filter((a) => !newFiles.some((b) => isSameFile(a, b))).concat(newFiles))
|
setFiles(files.filter((a) => !newFiles.some((b) => isSameFile(a, b))).concat(newFiles))
|
||||||
}
|
}
|
||||||
|
|
||||||
const {t} = useTranslation()
|
|
||||||
|
|
||||||
const title = t('UploadPage.title')
|
|
||||||
return (
|
return (
|
||||||
<Page title={title}>
|
<Page title="Upload">
|
||||||
<Header as="h1">{title}</Header>
|
|
||||||
{files.length ? (
|
{files.length ? (
|
||||||
<Table>
|
<Table>
|
||||||
<Table.Header>
|
<Table.Header>
|
||||||
<Table.Row>
|
<Table.Row>
|
||||||
<Table.HeaderCell>{t('UploadPage.table.filename')}</Table.HeaderCell>
|
<Table.HeaderCell>Filename</Table.HeaderCell>
|
||||||
<Table.HeaderCell>{t('UploadPage.table.size')}</Table.HeaderCell>
|
<Table.HeaderCell>Size</Table.HeaderCell>
|
||||||
<Table.HeaderCell>{t('UploadPage.table.statusTitle')}</Table.HeaderCell>
|
<Table.HeaderCell>Status / Title</Table.HeaderCell>
|
||||||
<Table.HeaderCell colSpan={2}></Table.HeaderCell>
|
<Table.HeaderCell colSpan={2}></Table.HeaderCell>
|
||||||
</Table.Row>
|
</Table.Row>
|
||||||
</Table.Header>
|
</Table.Header>
|
||||||
|
@ -179,17 +169,15 @@ export default function UploadPage() {
|
||||||
</List>
|
</List>
|
||||||
) : result ? (
|
) : result ? (
|
||||||
<>
|
<>
|
||||||
<Icon name="check" /> {result.track?.title || t('general.unnamedTrack')}
|
<Icon name="check" /> {result.track?.title || 'Unnamed track'}
|
||||||
</>
|
</>
|
||||||
) : (
|
) : (
|
||||||
<FileUploadStatus {...{id, file}} onComplete={onCompleteFileUpload} />
|
<FileUploadStatus {...{id, file}} onComplete={onCompleteFileUpload} />
|
||||||
)}
|
)}
|
||||||
</Table.Cell>
|
</Table.Cell>
|
||||||
|
<Table.Cell>{result?.track ? <Link to={`/tracks/${result.track.slug}`}>Show</Link> : null}</Table.Cell>
|
||||||
<Table.Cell>
|
<Table.Cell>
|
||||||
{result?.track ? <Link to={`/tracks/${result.track.slug}`}>{t('general.show')}</Link> : null}
|
{result?.track ? <Link to={`/tracks/${result.track.slug}/edit`}>Edit</Link> : null}
|
||||||
</Table.Cell>
|
|
||||||
<Table.Cell>
|
|
||||||
{result?.track ? <Link to={`/tracks/${result.track.slug}/edit`}>{t('general.edit')}</Link> : null}
|
|
||||||
</Table.Cell>
|
</Table.Cell>
|
||||||
</Table.Row>
|
</Table.Row>
|
||||||
))}
|
))}
|
||||||
|
|
Some files were not shown because too many files have changed in this diff Show more
Loading…
Reference in a new issue