- Overview
- Git Branch Workflow
- Development
- Generating Documentation
- Continuous Integration
- Contributing
- License
- Appendix
This is the Circulation Manager for Library Simplified. The Circulation Manager is the main connection between a library's collection and Library Simplified's various client-side applications, including SimplyE. It handles user authentication, combines licensed works with open access content, pulls in updated book information from distributor APIs and secondary sources, and serves up available books in appropriately organized OPDS feeds.
The default branch is develop
and that's the working branch that should be used when branching off for bug fixes or new features. Once a feature branch pull request is merged into develop
, the changes can be merged to main
to create releases.
Branches whose name follows the pattern <organization>-deploy-<env>
are protected, and are used as part of CI/CD processes for the named organization. These branches include:
nypl-deploy-qa
nypl-deploy-production
openebooks-deploy-qa
openebooks-deploy-production
bpl-deploy-qa
bpl-deploy-production
The preferred method for running an instance of the Circulation Manager is via Docker containers. If for whatever reason you cannot run a containerized version of the Circulation Manager, initial instructions for a direct installation can be found in docs/NonDockerInstallation.md.
The recommended way to get a local development environment is via Docker. If you need to perform a direct installation, please see the Non-Docker Installation document.
Note: The Circulation Manager containers should build and run on both amd64- and arm64-based machines. However, due to differences in the containerized version of ElasticSearch that we currently depend on, it is necessary to use a different Docker Compose file (docker-compose.arm64.yml
) to orchestrate the cluster on arm64 machines, including those using Apple's M1 chip series. If you use the make
commands below to control the local cluster this difference won't matter, as it will choose the correct compose file based on your machine's architecture. However, if you run the docker-compose
commands directly on an arm64-based machine, please make sure to use the correct compose file, via docker-compose --file docker-compose.arm64.yml <COMMAND>
.
In addition to installing Docker Desktop (link above), you'll also need an account at Docker Hub. Docker works by downloading base machine images to build on top of, and the majority of those are available from Docker Hub. However, the Docker Engine can only locate and pull images on behalf of an authenticated user.
Once you've created an account (or using an account you already have), you'll need to sign in via your Docker Desktop installation. You should be able to do that from the top right corner of the Docker Desktop dashboard window.
A Makefile
is provided to help manage the local containers for the Circulation Manager. If you don't currently have make
on your system, it should be available from your system's package manager (apt
, brew
, etc.) You can also run the recipes in the Makefile by hand, by copying and pasting the commands to your shell, if you can't get or would prefer not to install make
itself.
The following shell commands will clone this repo and build Docker images for the Circulation Manager web application and a PostgreSQL database:
git clone https://github.com/NYPL-Simplified/circulation.git
cd ./circulation
make build
Set environmental variable TZ
to your local time zone. This runs on America/New_York as default time zone.
Once the images are successfully built, you can start a local cluster with:
make up
The first time you start the cluster, the database container will run the initialization script localdev_postgres_init.sh
, which creates dev and test databases, installs PostgreSQL extensions, and creates credentialed users. Since the PostgreSQL data directory is persisted in a Docker Volume, subsequent startups will not re-initialize the database. The webapp container will wait to start its webserver until the database is available and accepting connections. If you want to observe the initialization process, use make up-watch
instead of make up
, to keep your terminal attached to the output of the running containers.
If you have previously run a version of the containerized Circulation Manager, it is possible you'll have a Docker volume remaining to persist the PostgreSQL data directory. If that is the case, and the database in that directory is not configured correctly, you could see output from make up-watch
like:
cm_local_db | PostgreSQL Database directory appears to contain a database; Skipping initialization
[...]
cm_local_db | 2021-11-19 22:20:06.563 UTC [74] FATAL: password authentication failed for user "simplified"
cm_local_db | 2021-11-19 22:20:06.563 UTC [74] DETAIL: Role "simplified" does not exist.
[...]
cm_local_webapp | --- Database unavailable, sleeping 5 seconds
If so, run make clean
to remove the existing volume, then make up-watch
again--you should see the database initialization process occur.
While the cluster is running (and after the web server has started), you should be able to access the API endpoints at http://localhost
, and the administrative web app at http://localhost/admin/
. The first time you attempt to sign in to the admin app a user will be created with the credentials you supply.
Other lifecycle management commands (a full list is available via make help
):
make stop
/make start
- after runningmake up
, you can pause and unpause the cluster without destroying the containersmake down
- stops the cluster and removes the containers and virtual networkmake clean
- stops the cluster, removes the containers and virtual network, and the db volumemake full-clean
- same asmake clean
, but also deletes the Docker images for the cluster
While the cluster is running, you can access the containers with these commands:
make db-session
- Starts apsql
session on the database container as the superusermake webapp-shell
- Open a bash shell on the webapp containermake webapp-py-repl
- Open a Python REPL session on the webapp container, in the project's virtualenvmake scripts-shell
- Open a bash shell on the scripts containermake scripts-py-repl
- Open a Python REPL session on the scripts container, in the project's virtualenv
While you have a running cluster, you can start the test suite with:
make test
That will run the entire test suite via pytest
. If you instead would like to use the -x
option to pytest
to exit at the first error or failure, you can use:
make test-x
You can ignore the Makefile
and directly issue the pytest
command to the webapp container via docker exec
, as follows:
docker exec -it cm_local_webapp pipenv run pytest tests
That gives you access to all of the standard options to pytest
, allowing variations like:
# Fail fast with -x
docker exec -it --env TESTING=1 cm_local_webapp /usr/local/bin/runinvenv /simplified_venv pytest -x tests
# Run a particular test class
docker exec -it --env TESTING=1 cm_local_webapp /usr/local/bin/runinvenv /simplified_venv pytest tests/test_controller.py::TestCirculationManager
# Run a specific test method
docker exec -it --env TESTING=1 cm_local_webapp /usr/local/bin/runinvenv /simplified_venv pytest tests/test_controller.py::TestCirculationManager::test_exception_during_external_search_initialization_is_stored
# Turn off the warnings output
docker exec -it --env TESTING=1 cm_local_webapp /usr/local/bin/runinvenv /simplified_venv pytest --disable-warnings tests
The full set of options to the pytest
executable is available at https://docs.pytest.org/en/6.2.x/usage.html
The Docker containers for local development of the web app and script runner (cm_local_webapp
and cm_local_scripts
, respectively) do not have copies of the codebase in their virtualized file systems. Instead, the local directory that this repo is checked out to is made available to the container as a read-only bind mount, at /home/simplified/circulation
. The setting of that bind mount occurs via docker-compose
, and can be seen in the webapp
and scripts
sections of docker-compose.yml
and docker-compose.arm64.yml
.
Because those host mounts occur through Docker Compose, they will not be present if you directly create a container from the built images via docker run
, unless you specify the bind mount via command line options, as with:
docker run -d --rm --mount type=bind,source="$(pwd)",target=/home/simplified/circulation \
--entrypoint tail circulation_webapp:latest -f /dev/null
With the local directory as a bind mount, you should be able to make changes on your host machine, and see them reflected in the behavior of the container.
If you are working on the Circulation Manager Admin, and need to add collections of books to your local app for testing purposes, you can follow these instructions.
Code documentation using Sphinx can be found on this repo's Github Pages. It currently documents this repo's api
directory, scripts
file, and the core
submodule directory. The configuration for the documentation can be found in /docs
.
Github Actions handles generating the .rst
source files, generating the HTML static site, and deploying the build to the gh-pages
branch.
To view the documentation locally, go into the /docs
directory and run make html
. This will generate the .rst source files and build the static site in /docs/build/html
.
API documentation using Swagger(OpenAPI) can be found on at the /apidocs_admin
and /apidocs_public
endpoints.
This project runs all the unit tests through Github Actions for new pull requests and when merging into the default develop
branch. The relevant file can be found in .github/workflows/test.yml
. When contributing updates or fixes, it's required for the test Github Action to pass for all python 3 environments. Run the tox
command locally before pushing changes to make sure you find any failing tests before committing them.
As mentioned above, Github Actions is also used to build and deploy Sphinx documentation to Github Pages. The relevant file can be found in .github/workflows/docks.yml
.
This codebase includes the provision for instrumenting the API and background scripts with NewRelic for error monitoring, metrics and other purposes. By default this is disabled for local development, and will not function when deployed without specific provisions. The steps below assume you have signed up for a NewRelic account and have access to an API key that can be used to instrument your application.
To instrument when running via docker compose locally three steps must be taken.
- You must have the NewRelic infrastructure agent running on your local development machine. This can be installed via
brew
on macOS, or other means (see NewRelic documentation) - A
.env
file must be provided in the root directory that contains the following line:NEW_RELIC_LICENSE_KEY=[your_newrelic_key_here]
- The
NEW_RELIC_MONITOR_MODE
environment variable must be set to'true'
in the docker compose file you are currently using
To instrument your deployment Circulation Manager instances you must provide the following environment variables as runtime variables to the docker container(s) running the CM (how you do so depends on your production environment):
NEW_RELIC_APP_NAME
: String that identifies your applicationNEW_RELIC_MONITOR_MODE
: Boolean formatted as a stringNEW_RELIC_CONFIG_FILE
: Absolute path to config file, without changes should be/home/simplified/circulation/newrelic.ini
NEW_RELIC_ENVIRONMENT
: Specifies dev/qa/prodNEW_RELIC_LICENSE_KEY
: An API key for your NewRelic account
See here for more information.
This project is open-sourced software licensed under the Apache-2.0 License.
View the NYPL Simplified Wiki for more in depth information on various topics relating to the circulation manager.