Skip to content

Latest commit

 

History

History
87 lines (70 loc) · 4.43 KB

README.md

File metadata and controls

87 lines (70 loc) · 4.43 KB

QueryApi

With QueryApi you can

🧩 Components

  1. Coordinator Continuously reads latest registry and attempts to synchronise that configuration across the system, i.e. starting relevant Block Streams and Executors.
  2. Runner Manages "Executors" which, retrieves messages from Redis Streams, fetches matching blocks, and executes Indexer code against that block. Retrieves messages from Redis Stream, fetching matching block and executes the IndexerFunction.
  3. Block Streamer Manages "Block Streams" which, use the configured contract filter to fetch relevant blocks from S3 and publish those blocks to Redis Streams.
  4. IndexerFunction Editor UI Serves the editor UI within the dashboard widget and mediates some communication with the GraphQL DB and block server.
  5. Hasura Authentication Service Provides authentication for the Hasura GraphQL server.
  6. IndexerFunction Registry Contract Stores IndexerFunctions, their schemas and execution parameters like start block height.
  7. Lake Block server Serves blocks from the S3 lake for in browser testing of IndexerFunctions.

🚀 Getting Started

The majority of the QueryApi components can be set up locally using Docker. For this purpose, a Docker Compose file has been provided. However, the local system still relies on the NEAR Mainnet, rather than running on a localnet.

Requirements

AWS Credentials Setup

QueryApi requires AWS credentials to stream blocks from NEAR Lake. Credentials are exposed via the following environment variables, which can be found in the Docker Compose file:

Runner:

  • AWS_ACCESS_KEY_ID
  • AWS_SECRET_ACCESS_KEY

Block Streamer:

  • AWS_ACCESS_KEY_ID
  • AWS_SECRET_ACCESS_KEY

These should be populated with your credentials. In most cases, the same key pair can be used for all 3 sets of credentials. Just ensure the keys have permissions to access S3 for handling Requestor Pays in Near Lake.

Running QueryApi

With everything configured correctly, we can now start all components of QueryApi with:

docker compose up

Developing

The most convenient way to develop the individual components of QueryAPI is to start everthing except the target component in docker and then run the target on the host machine directly. For example, to develop Coordinator, you can run:

docker compose up --scale coordinator=0
cd coordinator
cargo run coordinator

Before running the component, ensure that the environment is configured correctly. Refer to docker compose for required variables, these may need to be adapted to ensure that networking still works between the host and containers.

Developing frontend

  1. Change frontend/replacement.dev.json to
{
    "REPL_ACCOUNT_ID": "dev-queryapi.dataplatform.near",
    "REPL_GRAPHQL_ENDPOINT": "https://near-queryapi.dev.api.pagoda.co",
    "REPL_EXTERNAL_APP_URL": "http://localhost:3000",
    "REPL_REGISTRY_CONTRACT_ID": "dev-queryapi.dataplatform.near",
    "REPL_QUERY_API_USAGE_URL": "https://storage.googleapis.com/databricks-near-query-runner/output/query-api-usage/indexers_dev.json"
}
  1. cd frontend
  2. npm install
  3. npm run dev
  4. npm run serve:widgets:dev
  5. Set flags https://dev.near.org/flags to http://127.0.0.1:3030
  6. Navigate to https://dev.near.org/dev-queryapi.dataplatform.near/widget/QueryApi.dev-App

Local Configuration

  • Coordinator watches the dev registry contract by default (dev-queryapi.dataplatform.near). To use a different contract, you can update the REGISTRY_CONTRACT_ID environment variable.

Known Issues

It is expected to see some provisioning errors from Runner when starting QueryAPI for the first time. These occur when multiple indexers under the same account attempt to provision the same shared infrastructure. These should self resolve after a few seconds.