diff --git a/.github/workflows/build-base.yml b/.github/workflows/build-base.yml index 08359c37..3a6ab6de 100644 --- a/.github/workflows/build-base.yml +++ b/.github/workflows/build-base.yml @@ -32,7 +32,7 @@ jobs: name: Update base image runs-on: ubuntu-latest env: - IMAGE_NAME: prompt-proto-base + IMAGE_NAME: prompt-base STACK_TAG: ${{ inputs.stackTag }} steps: - name: Checkout code diff --git a/.github/workflows/build-service.yml b/.github/workflows/build-service.yml index ff9d8ba3..e65a0df4 100644 --- a/.github/workflows/build-service.yml +++ b/.github/workflows/build-service.yml @@ -61,14 +61,14 @@ jobs: - name: Run tests run: | docker run \ - -v $GITHUB_WORKSPACE:/home/lsst/prompt_prototype \ - ghcr.io/${{ github.repository_owner }}/prompt-proto-base:${{ matrix.baseTag }} \ + -v $GITHUB_WORKSPACE:/home/lsst/prompt_processing \ + ghcr.io/${{ github.repository_owner }}/prompt-base:${{ matrix.baseTag }} \ bash -c ' - cd /home/lsst/prompt_prototype + cd /home/lsst/prompt_processing source /opt/lsst/software/stack/loadLSST.bash setup -r . # Fix permissions; arg must be absolute path. - git config --global --add safe.directory /home/lsst/prompt_prototype + git config --global --add safe.directory /home/lsst/prompt_processing scons' update-service-image: @@ -81,7 +81,7 @@ jobs: matrix: baseTag: ${{ fromJSON(needs.matrix-gen.outputs.matrix) }} env: - IMAGE_NAME: prompt-proto-service + IMAGE_NAME: prompt-service BASE_TAG: ${{ matrix.baseTag }} steps: - name: Checkout code @@ -94,7 +94,7 @@ jobs: password: ${{ secrets.GITHUB_TOKEN }} - name: Determine eups tag run: | - docker run ghcr.io/${{ github.repository_owner }}/prompt-proto-base:"$BASE_TAG" bash -c "cat stack/miniconda*/ups_db/global.tags" > eups.tag + docker run ghcr.io/${{ github.repository_owner }}/prompt-base:"$BASE_TAG" bash -c "cat stack/miniconda*/ups_db/global.tags" > eups.tag echo "Eups tag = $(< eups.tag)" - name: Build image run: | diff --git a/Dockerfile.activator b/Dockerfile.activator index 049f5591..49da5c4b 100644 --- a/Dockerfile.activator +++ b/Dockerfile.activator @@ -1,8 +1,8 @@ ARG BASE_TAG=latest -FROM ghcr.io/lsst-dm/prompt-proto-base:${BASE_TAG} +FROM ghcr.io/lsst-dm/prompt-base:${BASE_TAG} ENV PYTHONUNBUFFERED True ENV APP_HOME /app -ENV PROMPT_PROTOTYPE_DIR $APP_HOME +ENV PROMPT_PROCESSING_DIR $APP_HOME ARG PUBSUB_VERIFICATION_TOKEN ARG PORT WORKDIR $APP_HOME diff --git a/README.rst b/README.rst index 04df0079..ced3bd5a 100644 --- a/README.rst +++ b/README.rst @@ -1,13 +1,15 @@ -################ -prompt_prototype -################ +################# +prompt_processing +################# -``prompt_prototype`` is a package in the `LSST Science Pipelines `_. +``prompt_processing`` is a package in the `LSST Science Pipelines `_. -``prompt_prototype`` contains code used for testing concepts for the Rubin Observatory Prompt Processing framework. -The design for the framework is described in `DMTN-219`_. +``prompt_processing`` contains code used for the Rubin Observatory Prompt Processing framework. +The design for the framework is described in `DMTN-219`_ and `DMTN-260`_. .. _DMTN-219: https://dmtn-219.lsst.io/ +.. _DMTN-260: https://dmtn-260.lsst.io/ + At present, the package does not conform exactly to the layout or conventions of LSST stack packages. In particular, the Python code does not use the ``lsst`` namespace, and the docs do not support Sphinx builds (even local ones). diff --git a/SConstruct b/SConstruct index e3198886..29afe13d 100644 --- a/SConstruct +++ b/SConstruct @@ -1,4 +1,4 @@ # -*- python -*- from lsst.sconsUtils import scripts # Python-only package -scripts.BasicSConstruct("prompt_prototype", disableCc=True, noCfgFile=True) +scripts.BasicSConstruct("prompt_processing", disableCc=True, noCfgFile=True) diff --git a/bin.src/make_hsc_rc2_export.py b/bin.src/make_hsc_rc2_export.py index c2ec64de..cd70e75d 100755 --- a/bin.src/make_hsc_rc2_export.py +++ b/bin.src/make_hsc_rc2_export.py @@ -1,5 +1,5 @@ #!/usr/bin/env python -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/bin.src/make_latiss_export.py b/bin.src/make_latiss_export.py index f1428ce8..176da8b3 100755 --- a/bin.src/make_latiss_export.py +++ b/bin.src/make_latiss_export.py @@ -1,5 +1,5 @@ #!/usr/bin/env python -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/bin.src/make_preloaded_export.py b/bin.src/make_preloaded_export.py index 62c173e4..b6acf770 100755 --- a/bin.src/make_preloaded_export.py +++ b/bin.src/make_preloaded_export.py @@ -1,5 +1,5 @@ #!/usr/bin/env python -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/bin.src/make_remote_butler.py b/bin.src/make_remote_butler.py index 45e8046e..e9e08be4 100755 --- a/bin.src/make_remote_butler.py +++ b/bin.src/make_remote_butler.py @@ -1,5 +1,5 @@ #!/usr/bin/env python -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project @@ -49,7 +49,7 @@ def _make_parser(): parser.add_argument("--target-repo", required=True, help="The URI of the repository to create.") parser.add_argument("--seed-config", - default=os.path.join(getPackageDir("prompt_prototype"), "etc", "db_butler.yaml"), + default=os.path.join(getPackageDir("prompt_processing"), "etc", "db_butler.yaml"), help="The config file to use for the new repository. Defaults to etc/db_butler.yaml.") parser.add_argument("--export-file", default="export.yaml", help="The export file containing the repository contents. Defaults to ./export.yaml.") diff --git a/bin.src/make_template_export.py b/bin.src/make_template_export.py index 9310f134..b50ec081 100755 --- a/bin.src/make_template_export.py +++ b/bin.src/make_template_export.py @@ -1,5 +1,5 @@ #!/usr/bin/env python -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/bin/prompt_prototype_upload_raws.sh b/bin/prompt_processing_upload_raws.sh similarity index 96% rename from bin/prompt_prototype_upload_raws.sh rename to bin/prompt_processing_upload_raws.sh index 7cde8b53..5c7185d5 100755 --- a/bin/prompt_prototype_upload_raws.sh +++ b/bin/prompt_processing_upload_raws.sh @@ -1,5 +1,5 @@ #!/bin/bash -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project @@ -21,7 +21,7 @@ # along with this program. If not, see . # This script uploads the raw files from the HSC PDR2 run to a bucket at USDF. -# It renames the files to match prompt_prototype conventions. The user must +# It renames the files to match prompt_processing conventions. The user must # have bucket access already configured. set -e # Abort on any error diff --git a/doc/playbook.rst b/doc/playbook.rst index 954367c1..87123483 100644 --- a/doc/playbook.rst +++ b/doc/playbook.rst @@ -1,6 +1,6 @@ -######################################################### -Playbook for the Prompt Processing Proposal and Prototype -######################################################### +###################################################### +Developers' Playbook for the Prompt Processing Service +###################################################### .. _DMTN-219: https://dmtn-219.lsst.io/ @@ -9,7 +9,7 @@ Table of Contents * `Containers`_ * `Buckets`_ -* `Prototype Service`_ +* `Development Service`_ * `tester`_ * `Databases`_ @@ -17,16 +17,16 @@ Table of Contents Containers ========== -The prototype consists of two containers. +The service consists of two containers. The first is a base container with the Science Pipelines "stack" code and networking utilities. -The second is a service container made from the base that has the Prompt Processing prototype service code. -All containers are managed by `GitHub Container Registry `_ and are built using GitHub Actions. +The second is a service container made from the base that has the Prompt Processing service code. +All containers are managed by `GitHub Container Registry `_ and are built using GitHub Actions. To build the base container: * If there are changes to the container, push them to a branch, then open a PR. The container should be built automatically. -* If there are no changes (typically because you want to use an updated Science Pipelines container), go to the repository's `Actions tab `_ and select "Run workflow". +* If there are no changes (typically because you want to use an updated Science Pipelines container), go to the repository's `Actions tab `_ and select "Run workflow". From the dropdown, select the branch whose container definition will be used, and the label of the Science Pipelines container. * New containers built from ``main`` are tagged with the corresponding Science Pipelines release (plus ``w_latest`` or ``d_latest`` if the release was requested by that name). For automatic ``main`` builds, or if the corresponding box in the manual build is checked, the new container also has the ``latest`` label. @@ -42,7 +42,7 @@ To build the service container: * If there are changes to the service, push them to a branch, then open a PR. The container should be built automatically using the ``latest`` base container. -* To force a rebuild manually, go to the repository's `Actions tab `_ and select "Run workflow". +* To force a rebuild manually, go to the repository's `Actions tab `_ and select "Run workflow". From the dropdown, select the branch whose code should be built. The container will be built using the ``latest`` base container, even if there is a branch build of the base. * To use a base other than ``latest``, edit ``.github/workflows/build-service.yml`` on the branch and override the ``BASE_TAG_LIST`` variable. @@ -112,8 +112,8 @@ To inspect them with the MinIO Client ``mc`` tool, first set up an alias (e.g. ` For Butler not to complain about the bucket names, set the environment variable ``LSST_DISABLE_BUCKET_VALIDATION=1``. -Prototype Service -================= +Development Service +=================== The service can be controlled with ``kubectl`` from ``rubin-devl``. You must first `get credentials for the development cluster `_ on the web; ignore the installation instructions and copy the commands from the second box. @@ -146,8 +146,7 @@ This file fully supports the Go template syntax. A few useful commands for managing the service: -* ``kubectl config set-context usdf-prompt-processing-dev --namespace=prompt-proto-service`` sets the default namespace for the following ``kubectl`` commands to ``prompt-proto-service``. - Note that many of the workflows in `slaclab/rubin-usdf-prompt-processing`_ run in the ``knative-serving`` or ``knative-eventing`` namespaces; to examine the resources of these workflows, add e.g. ``-n knative-eventing`` to the examples below. +* ``kubectl config set-context usdf-prompt-processing-dev --namespace=prompt-proto-service-`` sets the default namespace for the following ``kubectl`` commands to ``prompt-proto-service-``. * ``kubectl get serving`` summarizes the state of the service, including which revision(s) are currently handling messages. A revision with 0 replicas is inactive. * ``kubectl get pods`` lists the Kubernetes pods that are currently running, how long they have been active, and how recently they crashed. @@ -179,14 +178,14 @@ To delete such services manually: Identifying a Pod's Codebase ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -To identify which version of ``prompt-prototype`` a pod is running, run +To identify which version of Prompt Processing a pod is running, run .. code-block:: sh - kubectl describe pod | grep "prompt-proto-service@" + kubectl describe pod | grep "prompt-service@" This gives the hash of the service container running on that pod. -Actually mapping the hash to a branch version may require a bit of detective work; `the GitHub container registry `_ (which calls hashes "Digests") is a good starting point. +Actually mapping the hash to a branch version may require a bit of detective work; `the GitHub container registry `_ (which calls hashes "Digests") is a good starting point. To find the version of Science Pipelines used, find the container's page in the GitHub registry, then search for ``EUPS_TAG``. @@ -218,12 +217,12 @@ It can be run from ``rubin-devl``, but requires the user to install the ``conflu You must have a profile set up for the ``rubin:rubin-pp`` bucket (see `Buckets`_, above). -Install the prototype code, and set it up before use: +Install the Prompt Processing code, and set it up before use: .. code-block:: sh - git clone https://github.com/lsst-dm/prompt_prototype - setup -r prompt_prototype + git clone https://github.com/lsst-dm/prompt_processing + setup -r prompt_processing The tester scripts send ``next_visit`` events for each detector via Kafka on the ``next-visit-topic`` topic. They then upload a batch of files representing the snaps of the visit to the ``rubin:rubin-pp`` S3 bucket, simulating incoming raw images. diff --git a/etc/db_butler.yaml b/etc/db_butler.yaml index 077135c5..c90952b1 100644 --- a/etc/db_butler.yaml +++ b/etc/db_butler.yaml @@ -1,4 +1,4 @@ -# Seed config for generating Prompt Prototype central repo. +# Seed config for generating Prompt Processing central repo. # This is the test repo to be used with upload.py, not the repo for processing AuxTel data. registry: db: postgresql://pp@usdf-prompt-processing-dev.slac.stanford.edu/ppcentralbutler diff --git a/pipelines/LATISS/ApPipe.yaml b/pipelines/LATISS/ApPipe.yaml index 36d339dc..7ce1bf2a 100644 --- a/pipelines/LATISS/ApPipe.yaml +++ b/pipelines/LATISS/ApPipe.yaml @@ -2,7 +2,7 @@ description: Alert Production pipeline specialized for LATISS # This file should strive to contain just an import from ap_pipe. # Exceptions are allowed temporarily when urgent bug fixes and -# prompt_prototype build can't wait for the lsst_distrib +# prompt_processing build can't wait for the lsst_distrib # release schedule. imports: - location: $AP_PIPE_DIR/pipelines/LATISS/ApPipe.yaml diff --git a/pipelines/LATISS/Isr.yaml b/pipelines/LATISS/Isr.yaml index 347db1dc..cc4bf41f 100644 --- a/pipelines/LATISS/Isr.yaml +++ b/pipelines/LATISS/Isr.yaml @@ -1,6 +1,6 @@ description: ISR-only pipeline specialized for LATISS imports: - - location: $PROMPT_PROTOTYPE_DIR/pipelines/LATISS/ApPipe.yaml + - location: $PROMPT_PROCESSING_DIR/pipelines/LATISS/ApPipe.yaml include: - isr diff --git a/pipelines/LATISS/SingleFrame.yaml b/pipelines/LATISS/SingleFrame.yaml index 4e2d86e5..13712f53 100644 --- a/pipelines/LATISS/SingleFrame.yaml +++ b/pipelines/LATISS/SingleFrame.yaml @@ -1,6 +1,6 @@ description: Single-frame pipeline specialized for LATISS imports: - - location: $PROMPT_PROTOTYPE_DIR/pipelines/LATISS/ApPipe.yaml + - location: $PROMPT_PROCESSING_DIR/pipelines/LATISS/ApPipe.yaml include: - processCcd diff --git a/python/activator/activator.py b/python/activator/activator.py index ff89d731..aab70795 100644 --- a/python/activator/activator.py +++ b/python/activator/activator.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project @@ -46,7 +46,7 @@ ) from .visit import FannedOutVisit -PROJECT_ID = "prompt-proto" +PROJECT_ID = "prompt-processing" # The short name for the instrument. instrument_name = os.environ["RUBIN_INSTRUMENT"] diff --git a/python/activator/config.py b/python/activator/config.py index 73cdd98a..c2999ae7 100644 --- a/python/activator/config.py +++ b/python/activator/config.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/python/activator/logger.py b/python/activator/logger.py index d50edd21..444a8345 100644 --- a/python/activator/logger.py +++ b/python/activator/logger.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project @@ -99,7 +99,7 @@ def _set_context_logger(): def setup_usdf_logger(labels=None): - """Set global logging settings for prompt_prototype. + """Set global logging settings for prompt_processing. Calling this function redirects all warnings to go through the logger. diff --git a/python/activator/middleware_interface.py b/python/activator/middleware_interface.py index 37a2a208..b97be87f 100644 --- a/python/activator/middleware_interface.py +++ b/python/activator/middleware_interface.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/python/activator/raw.py b/python/activator/raw.py index 416f5796..7aa13736 100644 --- a/python/activator/raw.py +++ b/python/activator/raw.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/python/activator/visit.py b/python/activator/visit.py index c20fbf36..7d981346 100644 --- a/python/activator/visit.py +++ b/python/activator/visit.py @@ -1,3 +1,24 @@ +# This file is part of prompt_processing. +# +# Developed for the LSST Data Management System. +# This product includes software developed by the LSST Project +# (https://www.lsst.org). +# See the COPYRIGHT file at the top-level directory of this distribution +# for details of code ownership. +# +# This program is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see . + __all__ = ["FannedOutVisit", "SummitVisit", "BareVisit"] from dataclasses import dataclass, field, asdict diff --git a/python/tester/upload.py b/python/tester/upload.py index 35218fe6..77c7044f 100644 --- a/python/tester/upload.py +++ b/python/tester/upload.py @@ -1,3 +1,24 @@ +# This file is part of prompt_processing. +# +# Developed for the LSST Data Management System. +# This product includes software developed by the LSST Project +# (https://www.lsst.org). +# See the COPYRIGHT file at the top-level directory of this distribution +# for details of code ownership. +# +# This program is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see . + import dataclasses import datetime import itertools diff --git a/python/tester/upload_hsc_rc2.py b/python/tester/upload_hsc_rc2.py index 7e34bc76..98113eee 100644 --- a/python/tester/upload_hsc_rc2.py +++ b/python/tester/upload_hsc_rc2.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project @@ -149,10 +149,8 @@ def get_hsc_visit_list(butler, n_sample): def prepare_one_visit(kafka_url, group_id, butler, visit_id): """Extract metadata and send next_visit events for one HSC-RC2 visit - One ``next_visit`` message is sent for each detector, to mimic the - current prototype design in which a single message is sent from the - Summit to the USDF and then a USDF-based server translates it into - multiple messages. + One ``next_visit`` message is sent to the development fan-out service, + which translates it into multiple messages. Parameters ---------- diff --git a/python/tester/utils.py b/python/tester/utils.py index bd470175..cee54c23 100644 --- a/python/tester/utils.py +++ b/python/tester/utils.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/tests/data/ApPipe.yaml b/tests/data/ApPipe.yaml index 39eee4a5..13e0c8ee 100644 --- a/tests/data/ApPipe.yaml +++ b/tests/data/ApPipe.yaml @@ -28,7 +28,7 @@ tasks: class: lsst.pipe.tasks.calibrate.CalibrateTask config: # Use dataset's reference catalogs - file: $PROMPT_PROTOTYPE_DIR/tests/data/calibrate.py + file: $PROMPT_PROCESSING_DIR/tests/data/calibrate.py diaPipe: class: lsst.ap.association.DiaPipelineTask config: diff --git a/tests/data/ISR.yaml b/tests/data/ISR.yaml index 2a1e528d..5e2db153 100644 --- a/tests/data/ISR.yaml +++ b/tests/data/ISR.yaml @@ -1,6 +1,6 @@ description: Test pipeline consisting only of ISR. imports: - - location: $PROMPT_PROTOTYPE_DIR/tests/data/ApPipe.yaml + - location: $PROMPT_PROCESSING_DIR/tests/data/ApPipe.yaml include: - isr diff --git a/tests/data/SingleFrame.yaml b/tests/data/SingleFrame.yaml index 2204dbdf..0234d15b 100644 --- a/tests/data/SingleFrame.yaml +++ b/tests/data/SingleFrame.yaml @@ -1,7 +1,7 @@ description: Test pipeline consisting only of single-frame steps. imports: - - location: $PROMPT_PROTOTYPE_DIR/tests/data/ApPipe.yaml + - location: $PROMPT_PROCESSING_DIR/tests/data/ApPipe.yaml include: - isr - characterizeImage diff --git a/tests/test_config.py b/tests/test_config.py index ef563637..67e1bcbf 100644 --- a/tests/test_config.py +++ b/tests/test_config.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project @@ -58,7 +58,7 @@ def setUp(self): def test_main_survey(self): config = PipelinesConfig( - ' (survey="TestSurvey")=[${PROMPT_PROTOTYPE_DIR}/pipelines/NotACam/ApPipe.yaml]') + ' (survey="TestSurvey")=[${PROMPT_PROCESSING_DIR}/pipelines/NotACam/ApPipe.yaml]') self.assertEqual( config.get_pipeline_files(self.visit), [os.path.normpath(os.path.join(TESTDIR, "..", "pipelines", "NotACam", "ApPipe.yaml"))] diff --git a/tests/test_logger.py b/tests/test_logger.py index be2961ee..df4383d9 100644 --- a/tests/test_logger.py +++ b/tests/test_logger.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/tests/test_middleware_interface.py b/tests/test_middleware_interface.py index a460f1db..080f5115 100644 --- a/tests/test_middleware_interface.py +++ b/tests/test_middleware_interface.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project @@ -55,8 +55,8 @@ # A pipelines config that returns the test pipelines. # Unless a test imposes otherwise, the first pipeline should run, and # the second should not be attempted. -pipelines = PipelinesConfig('''(survey="SURVEY")=[${PROMPT_PROTOTYPE_DIR}/tests/data/ApPipe.yaml, - ${PROMPT_PROTOTYPE_DIR}/tests/data/SingleFrame.yaml] +pipelines = PipelinesConfig('''(survey="SURVEY")=[${PROMPT_PROCESSING_DIR}/tests/data/ApPipe.yaml, + ${PROMPT_PROCESSING_DIR}/tests/data/SingleFrame.yaml] ''' ) diff --git a/tests/test_raw.py b/tests/test_raw.py index 0486a7d5..65d51a01 100644 --- a/tests/test_raw.py +++ b/tests/test_raw.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/tests/test_tester_utils.py b/tests/test_tester_utils.py index 5db64234..952006b2 100644 --- a/tests/test_tester_utils.py +++ b/tests/test_tester_utils.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/tests/test_visit.py b/tests/test_visit.py index afe4314c..d59ab646 100644 --- a/tests/test_visit.py +++ b/tests/test_visit.py @@ -1,4 +1,4 @@ -# This file is part of prompt_prototype. +# This file is part of prompt_processing. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project diff --git a/ups/prompt_prototype.table b/ups/prompt_processing.table similarity index 100% rename from ups/prompt_prototype.table rename to ups/prompt_processing.table