api | ||
deployment | ||
docs | ||
frontend | ||
tile-generator | ||
.dockerignore | ||
.editorconfig | ||
.gitignore | ||
.gitmodules | ||
.nvmrc | ||
docker-compose.yaml | ||
Dockerfile | ||
README.md | ||
roads_import.lua | ||
UPGRADING.md |
OpenBikeSensor Portal
This repository contains the source code required to run the OpenBikeSensor data collection portal. It is separated into components:
- api: The backend service, written in JavaScript for Node.js, using express.js, and a MongoDB for metadata storage.
- frontend: A React single-page application that allows access to the data, provides summaries and visualizations, and lets users adjust settings and manage and publish their tracks.
Clone the Project
First of all, you must clone this project. This project uses submodules, thus ensure, that they are cloned as well:
git clone --recursive https://github.com/openbikesensor/portal
# ... or if you forgot the --recursive argument, you can run this in the
# repository's directory later:
git submodule update --init --recursive
Production setup
There is a guide for a deployment based on docker in the deployment folder. Lots of non-docker deployment strategy are possible, but they are not "officially" supported, so please do not expect the authors of the software to assist in troubleshooting.
This is a rather complex application, and it is expected that you know the basics of deploying a modern web application securely onto a production server. We are sorry that we cannot guide you through all the details of that, as we just don't have the capacities to do so. Please research the respective topics first. If you struggle with application-specific issues, please let us know, we might be able to assist with those.
Please note that you will always need to install your own reverse proxy that terminates TLS for you and handles certificates. We do not support TLS directly in the application, instead, please use this prefered method.
Migrating (Production)
Migrations are not implemented yet. Once we need them, we'll add them and document the usage here.
Upgrading from v0.2 to v0.3
After v0.2 we switched the underlying technology of the API and the database.
We now have no more MongoDB, instead, everything has moved to the PostgreSQL
installation. For development setups, it is advised to just reset the whole
state (remove the local
folder) and start fresh. For production upgrades,
please follow the relevant section in UPGRADING.md
.
Development setup
We've moved the whole development setup into Docker to make it easy for everyone to get involved.
Install docker
Please install Docker Engine as well as Docker Compose onto your machine.
Then clone the repository as described above.
Configure Keycloak
Login will not be possible until you configure the keycloak realm correctly. Boot your keycloak instance:
docker-compose up -d --build keycloak
Now navigate to http://localhost:3003/ and follow these steps:
- Click "Administration Console" and log in with
admin
/admin
- Hover over the realm name on the top left and click "Add realm"
- Name the Realm
obs-dev
(spelling matters) and create it - In the sidebar, navigate to Configure -> Clients, and click "Create" on the top right
- Client ID is
portal
. Hit "Save". - In the Tab "Settings", edit the new client's "Access Type" to
confidential
and enter as "Valid Redirect URIs":http://localhost:3000/login/redirect
, then "Save" - Under "Credentials", copy the "Secret" and paste it into
api/config.dev.py
asKEYCLOAK_CLIENT_SECRET
. Please do not commit this change to git. - In the sidebar, navigate to Manage -> Users, and click "Add user" on the top right.
- Give the user a name (e.g.
test
), leave the rest as-is. - Under the tab "Credentials", set a new password, and make it non-temporary. Press "Set Password".
We are going to automate this process. For now, you will have to repeat it
every time you reset the keycloak datbaase, which is inside the PostgreSQL. The
script api/tools/reset_database.py
does not affect the state of the
keycloak database, however, so this should be rather rare.
Prepare database
Start the PostgreSQL database:
docker-compose up -d postgres
The first time you start postgres, a lot of extensions will be installed. This takes a while, so check the logs of the docker container until you see:
PostgreSQL init process complete; ready for start up.
If you don't wait long enough, the following commands might fail.
Next, initialize an empty database, creating all extensions and tables for the application at once:
docker-compose run --rm api tools/reset_database.py
You should import OpenStreetMap data now, see below for instructions.
To serve dynamic vector tiles from the API, run the following command once:
docker-compose run --rm api tools/prepare_sql_tiles.py
You might need to re-run this command after updates, to (re-)create the functions in the SQL database that are used when generating vector tiles.
Boot the application
Now you can run the remaining parts of the application:
docker-compose up -d --build api worker frontend
Your frontend should be running at http://localhost:3001 and the API at http://localhost:3000 -- but you probably only need to access the frontend for testing.
Migrating (Development)
Migrations are not implemented yet. Once we need them, we'll add them and document the usage here.
Import OpenStreetMap data
You need to import road information from OpenStreetMap for the portal to work. This information is stored in your PostgreSQL database and used when processing tracks (instead of querying the Overpass API), as well as for vector tile generation.
-
Install
osm2pgsql
. -
Download the area(s) you would like to import from GeoFabrik.
-
Import each file like this:
osm2pgsql --create --hstore --style api/roads_import.lua -O flex \ -H localhost -d obs -U obs \ path/to/downloaded/myarea-latest.osm.pbf
You might need to adjust the host, database and username (-H
, -d
, -U
) to
your setup, and also provide the correct password when queried. This process
should take a few seconds to minutes, depending on the area size. You can run
the process multiple times, with the same or different area files, to import or
update the data. You can also truncate the road
table before importing if you
want to remove outdated road information.
Refer to the documentation of osm2pgsql
for assistance. We are using "flex
mode", the provided script api/roads_import.lua
describes the transformations
and extractions to perform on the original data.
Static tile generation
The above instructions do not include the serving of vector tiles with the collected data. That is to be set up separately. Please follow the instructions in tile-generator.
Troubleshooting
If any step of the instructions does not work for you, please open an issue and describe the problem you're having, as it is important to us that onboarding is super easy :)