Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Streamingfast/docs #765

Merged
merged 12 commits into from
Sep 16, 2024
1 change: 1 addition & 0 deletions website/pages/en/_meta.js
Original file line number Diff line number Diff line change
Expand Up @@ -33,6 +33,7 @@ export default {
title: 'Substreams',
},
substreams: '',
sps: 'Substreams-powered Subgraphs',
'---4': {
type: 'separator',
},
Expand Down
31 changes: 1 addition & 30 deletions website/pages/en/new-chain-integration.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -76,33 +76,4 @@ Graph Node should be syncing the deployed subgraph if there are no errors. Give

## Substreams-powered Subgraphs

For StreamingFast-led Firehose/Substreams integrations, basic support for foundational Substreams modules (e.g. decoded transactions, logs and smart-contract events) and Substreams-powered subgraph codegen tools are included (check out [Injective](https://substreams.streamingfast.io/documentation/intro-getting-started/intro-injective/injective-first-sps) for an example).

There are two options to consume Substreams data through a subgraph:

- **Using Substreams triggers:** Consume from any Substreams module by importing the Protobuf model through a subgraph handler and move all your logic into a subgraph. This method creates the subgraph entities directly in the subgraph.
- **Using EntityChanges:** By writing more of the logic into Substreams, you can consume the module's output directly into `graph-node`. In `graph-node`, you can use the Substreams data to create your subgraph entities.

It is really a matter of where you put your logic, in the subgraph or the Substreams. Keep in mind that having more of your logic in Substreams benefits from a parallelized model, whereas triggers will be linearly consumed in `graph-node`. Consider the following example implementing a subgraph handler:

```ts
export function handleTransactions(bytes: Uint8Array): void {
let transactions = assembly.eth.transaction.v1.Transactions.decode(bytes.buffer).trasanctions // 1.
if (transactions.length == 0) {
log.info('No transactions found', [])
return
}

for (let i = 0; i < transactions.length; i++) {
// 2.
let transaction = transactions[i]

let entity = new Transaction(transaction.hash) // 3.
entity.from = transaction.from
entity.to = transaction.to
entity.save()
}
}
```

The `handleTransactions` function is a subgraph handler that receives the raw Substreams bytes as parameter and decodes them into a `Transactions` object. Then, for every transaction, a new subgraph entity is created. For more information about Substreams triggers, visit the [StreamingFast documentation](https://substreams.streamingfast.io/documentation/consume/subgraph/triggers) or check out community modules at [substreams.dev](https://substreams.dev/).
For StreamingFast-led Firehose/Substreams integrations, basic support for foundational Substreams modules (e.g. decoded transactions, logs and smart-contract events) and Substreams codegen tools are included. These tools enable the ability to enable [Substreams-powered subgraphs](./sps/sps-intro). Follow the [How-To Guide](https://substreams.streamingfast.io/documentation/how-to-guides/intro-your-first-application) and run `substreams codegen subgraph` to experience the codegen tools for yourself.
5 changes: 5 additions & 0 deletions website/pages/en/sps/_meta.js
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
export default {
'sps-intro': 'Introduction',
triggers: '',
'triggers-example': 'Tutorial',
}
19 changes: 19 additions & 0 deletions website/pages/en/sps/sps-intro.mdx
benface marked this conversation as resolved.
Show resolved Hide resolved
Original file line number Diff line number Diff line change
@@ -0,0 +1,19 @@
---
title: Introduction to Substreams-powered Subgraphs
---

By using a Substreams package (`.spkg`) as a data source, your subgraph gains access to a stream of pre-indexed blockchain data. This enables more efficient and scalable data handling, especially with large or complex blockchain networks.

There are two methods of enabling this technology:

Using Substreams [triggers](./triggers): Consume from any Substreams module by importing the Protobuf model through a subgraph handler and move all your logic into a subgraph. This method creates the subgraph entities directly in the subgraph.

Using [Entity Changes](https://substreams.streamingfast.io/documentation/consume/subgraph/graph-out): By writing more of the logic into Substreams, you can consume the module's output directly into graph-node. In graph-node, you can use the Substreams data to create your subgraph entities.

It is really a matter of where you put your logic, in the subgraph or the Substreams. Keep in mind that having more of your logic in Substreams benefits from a parallelized model, whereas triggers will be linearly consumed in graph-node.

Visit the following links for How-To Guides on using code-generation tooling to build your first end-to-end project quickly:

- [Solana](https://substreams.streamingfast.io/documentation/how-to-guides/intro-your-first-application/solana)
- [EVM](https://substreams.streamingfast.io/documentation/how-to-guides/intro-your-first-application/evm)
- [Injective](https://substreams.streamingfast.io/documentation/how-to-guides/intro-your-first-application/injective)
137 changes: 137 additions & 0 deletions website/pages/en/sps/triggers-example.mdx
Original file line number Diff line number Diff line change
@@ -0,0 +1,137 @@
---
title: 'Tutorial: Set Up a Substreams-Powered Subgraph on Solana'
---

## Prerequisites

Before starting, make sure to:

- Complete the [Getting Started Guide](https://github.com/streamingfast/substreams-starter) to set up your development environment using a Dev Container.
- Be familiar with The Graph and basic blockchain concepts such as transactions and Protobufs.

## Step 1: Initialize Your Project

1. Open your Dev Container and run the following command to initialize your project:

```bash
substreams init
```

2. Select the "minimal" project option.
3. Replace the contents of the generated `substreams.yaml` file with the following configuration, which filters transactions for the Orca account on the SPL token program ID:

```yaml
specVersion: v0.1.0
package:
name: my_project_sol
version: v0.1.0

imports: # Pass your spkg of interest
solana: https://github.com/streamingfast/substreams-solana-spl-token/raw/master/tokens/solana-spl-token-v0.1.0.spkg

modules:
- name: map_spl_transfers
use: solana:map_block # Select corresponding modules available within your spkg
initialBlock: 260000082

- name: map_transactions_by_programid
use: solana:solana:transactions_by_programid_without_votes

network: solana-mainnet-beta

params: # Modify the param fields to meet your needs
# For program_id: TokenkegQfeZyiNwAJbNbGKPFXCWuBvf9Ss623VQ5DA
map_spl_transfers: token_contract:orcaEKTdK7LKz57vaAYr9QeNsVEPfiu6QeMU1kektZE
```

## Step 2: Generate the Subgraph Manifest

Once the project is initialized, generate a subgraph manifest by running the following command in the Dev Container:

```bash
substreams codegen subgraph
```

You will generate a`subgraph.yaml` manifest which imports the Substreams package as a data source:

```yaml
---
dataSources:
- kind: substreams
name: my_project_sol
network: solana-mainnet-beta
source:
package:
moduleName: map_spl_transfers # Module defined in the substreams.yaml
file: ./my-project-sol-v0.1.0.spkg
mapping:
apiVersion: 0.0.7
kind: substreams/graph-entities
file: ./src/mappings.ts
handler: handleTriggers
```

## Step 3: Define Entities in `schema.graphql`

Define the fields you want to save in your subgraph entities by updating the `schema.graphql` file. Here is an example:

```graphql
type MyTransfer @entity {
id: ID!
amount: String!
source: String!
designation: String!
signers: [String!]!
}
```

This schema defines a `MyTransfer` entity with fields such as `id`, `amount`, `source`, `designation`, and `signers`.

## Step 4: Generate Protobuf Files

To generate Protobuf objects in AssemblyScript, run the following command:

```bash
npm run protogen
```

This command converts the Protobuf definitions into AssemblyScript, allowing you to use them in the subgraph's handler.

## Step 5: Handle Substreams Data in `mappings.ts`

With the Protobuf objects generated, you can now handle the decoded Substreams data in your `mappings.ts` file found in the `./src` directory. The example below demonstrates how to extract to subgraph entities the non-derived transfers associated to the Orca account id:

```ts
import { Protobuf } from 'as-proto/assembly'
import { Events as protoEvents } from './pb/sf/solana/spl/token/v1/Events'
import { MyTransfer } from '../generated/schema'

export function handleTriggers(bytes: Uint8Array): void {
const input: protoEvents = Protobuf.decode<protoEvents>(bytes, protoEvents.decode)

for (let i = 0; i < input.data.length; i++) {
const event = input.data[i]

if (event.transfer != null) {
let entity_id: string = `${event.txnId}-${i}`
const entity = new MyTransfer(entity_id)
entity.amount = event.transfer!.instruction!.amount.toString()
entity.source = event.transfer!.accounts!.source
entity.designation = event.transfer!.accounts!.destination

if (event.transfer!.accounts!.signer!.single != null) {
entity.signers = [event.transfer!.accounts!.signer!.single.signer]
} else if (event.transfer!.accounts!.signer!.multisig != null) {
entity.signers = event.transfer!.accounts!.signer!.multisig!.signers
}
entity.save()
}
}
}
```

## Conclusion

You’ve successfully set up a trigger-based Substreams-powered subgraph for a Solana SPL token. You can now further customize your schema, mappings, and modules to suit your specific use case.

For more advanced customization and optimizations, check out the official [Substreams documentation](https://substreams.streamingfast.io/tutorials/solana).
37 changes: 37 additions & 0 deletions website/pages/en/sps/triggers.mdx
Original file line number Diff line number Diff line change
@@ -0,0 +1,37 @@
---
title: Substreams Triggers
---

Custom triggers allow you to send data directly into your subgraph mappings file and entities (similar to tables and fields), enabling full use of the GraphQL layer. By importing the Protobuf definitions emitted by your Substreams module, you can receive and process this data within your subgraph’s handler, ensuring efficient and streamlined data management within the subgraph framework.

> Note: If you haven’t already, visit one of the How-To Guides found [here](./sps-intro) to scaffold your first project in the Development Container.

The following code demonstrates how to define a `handleTransactions` function in a subgraph handler. This function receives raw Substreams bytes as a parameter and decodes them into a `Transactions` object. For each transaction, a new subgraph entity is created.

```tsx
export function handleTransactions(bytes: Uint8Array): void {
let transactions = assembly.eth.transaction.v1.Transactions.decode(bytes.buffer).trasanctions // 1.
if (transactions.length == 0) {
log.info('No transactions found', [])
return
}

for (let i = 0; i < transactions.length; i++) {
// 2.
let transaction = transactions[i]

let entity = new Transaction(transaction.hash) // 3.
entity.from = transaction.from
entity.to = transaction.to
entity.save()
}
}
```

Here's what you’re seeing in the `mappings.ts` file:

1. The bytes containing Substreams data are decoded into the generated `Transactions` object, this object is used like any other AssemblyScript object
2. Looping over the transactions
3. Create a new subgraph entity for every transaction

To go through a detailed example of a trigger-based subgraph, [click here](./triggers-example).
10 changes: 8 additions & 2 deletions website/pages/en/substreams.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -4,9 +4,11 @@ title: Substreams

![Substreams Logo](/img/substreams-logo.png)

Substreams is a powerful blockchain indexing technology developed for The Graph Network. It enables developers to write Rust modules, compose data streams alongside the community, and provide extremely high-performance indexing due to parallelization in a streaming-first approach.
Substreams is a powerful blockchain indexing technology designed to enhance performance and scalability within The Graph Network. It offers the following features:

With Substreams, developers can quickly extract data from different blockchains (Ethereum, BNB, Solana, ect.) and send it to various locations of their choice, such as a Postgres database, a Mongo database, or a Subgraph. Additionally, Substreams packages enable developers to specify which data they want to extract from the blockchain.
- **Accelerated Indexing**: Substreams reduce subgraph indexing time thanks to a parallelized engine, enabling faster data retrieval and processing.
- **Multi-Chain Support**: Substreams expand indexing capabilities beyond EVM-based chains, supporting ecosystems like Solana, Injective, Starknet, and Vara.
- **Multi-Sink Support:** Subgraph, Postgres database, Clickhouse, Mongo database

## How Substreams Works in 4 Steps

Expand Down Expand Up @@ -44,3 +46,7 @@ To learn about the latest version of Substreams CLI, which enables developers to
### Expand Your Knowledge

- Take a look at the [Ethereum Explorer Tutorial](https://substreams.streamingfast.io/tutorials/evm) to learn about the basic transformations you can create with Substreams.

### Substreams Registry

A Substreams package is a precompiled binary file that defines the specific data you want to extract from the blockchain, similar to the `mapping.ts` file in traditional subgraphs. Visit [substreams.dev](https://substreams.dev/) to explore a growing collection of ready-to-use Substreams packages across various blockchain networks.
Loading