Creation and hosting of STAC catalogs by the ASF Tools team.
Production API: https://stac.asf.alaska.edu
- Swagger UI: https://stac.asf.alaska.edu/api.html
- STAC Browser: https://radiantearth.github.io/stac-browser/#/external/stac.asf.alaska.edu/
Test API: https://stac-test.asf.alaska.edu
- Swagger UI: https://stac-test.asf.alaska.edu/api.html
- STAC Browser: https://radiantearth.github.io/stac-browser/#/external/stac-test.asf.alaska.edu/
Clone the repository, create the environment, and install the developer dependencies:
git clone [email protected]:ASFHyP3/asf-stac.git
cd asf-stac
conda env create -f environment.yml
conda activate asf-stac
If you ever see the following warning from gdal
...
Warning 1: PROJ: proj_create_from_database: Open of /home/.../miniconda3/envs/asf-stac/share/proj failed
...you can run the following command and then re-activate your Conda environment:
conda env config vars set PROJ_LIB=${CONDA_PREFIX}/share/proj
The database only accepts connections from within the ASF Full VPN or from clients with the client security group attached. See the ingress rules for the database security group in the database CloudFormation template.
The database host and database user credentials are available via the AWS Secrets Manager console in the AWS account where the CloudFormation stack was deployed.
We must create and ingest the coherence dataset after running a new STAC API deployment. We must also re-create and re-ingest the dataset after making changes to how the STAC items are structured.
Fetch the list of S3 objects:
cd collections/sentinel-1-global-coherence/
./list-coherence-objects
wc -l coherence-s3-objects.txt
Confirm that the number of lines is 1033388
(one per object).
Next, create the dataset:
python create_coherence_items.py coherence-s3-objects.txt
wc -l sentinel-1-global-coherence.ndjson
Again, confirm that the number of lines is the same as in the previous step.
Finally, ingest the dataset:
cd ../../
make pypgstac-load db_host=<host> db_admin_password=<password> table=items ndjson_file=collections/sentinel-1-global-coherence/sentinel-1-global-coherence.ndjson
We must create and ingest the HAND dataset after running a new STAC API deployment. We must also re-create and re-ingest the dataset after making changes to how the STAC items are structured.
Fetch the list of S3 objects:
cd collections/glo-30-hand/
./list-hand-objects
wc -l hand-s3-objects.txt
Confirm that the number of lines is 26450
(one per object).
Next, create the dataset:
python create_hand_items.py hand-s3-objects.txt
wc -l glo-30-hand.ndjson
Again, confirm that the number of lines is the same as in the previous step.
Finally, ingest the dataset:
cd ../../
make pypgstac-load db_host=<host> db_admin_password=<password> table=items ndjson_file=collections/glo-30-hand/glo-30-hand.ndjson
We shouldn't need to manually connect to the database during normal operations, but we can if we need to (e.g. for debugging purposes).
Confirm that you have PostgreSQL installed, then run:
make psql db_host=<host> db_user=<user> db_password=<password>
You can run the STAC API frontend locally and it will automatically connect to the AWS-hosted database. We shouldn't need to run the API locally during normal operations, but we can if we need to (e.g. for debugging purposes).
The local API provides access to the Transaction extension (which provides create/update/delete endpoints), while the publicly available API does not. We shouldn't need access to the Transaction endpoints during normal operations, but they may be useful for making one-off edits in a sandbox or test deployment.
Run:
make run-api db_host=<host> db_admin_password=<password>
You can also append an enabled_extensions=<list>
argument, where <list>
is the list of extensions
that gets passed to the pgstac
app via the ENABLED_EXTENSIONS
environment variable, as described
in the docstring for the
module.
You should see something like Uvicorn running on http://0.0.0.0:8000 (Press CTRL+C to quit)
in the output; you can
query the API at that URL.
You can confirm that the Transaction extension is enabled by opening the local API URL in a web browser
and appending /api.html
to open the Swagger UI. You should see various create/update/delete endpoints
under the "Transaction Extension" heading. These endpoints should not appear in the Swagger UI for the
publicly available API.
The initial AWS deployment creates a Postgres database, installs the PostGIS extension, and then installs PgSTAC. Follow these steps to upgrade the database:
-
Run the following command to list the Postgres versions supported by Amazon RDS:
aws rds describe-db-engine-versions --engine postgres
Identify the entry that corresponds to the current version of the database. Then identify the newest available version from the list of valid upgrade targets for the current version. This will be the new version for the database.
-
Change the Postgres version specified in the database CloudFormation template to the new version.
-
Next, refer to the tables shown here to determine which version of the PostGIS extension is supported by the new Postgres version.
-
Change the PostGIS version specified in the install/upgrade script.
-
Deploy to AWS.
PgSTAC upgrades are handled automatically: the deployment pipeline migrates the database to the installed
version of pypgstac
. See https://stac-utils.github.io/pgstac/pypgstac for more information about migrations.