- Table of Contents
- Overview
- Major Components
- Setup
- Monitoring and Debugging
- For Contributors
- Case Studies
- Find us
A snapshotter peer as part of Powerloom Protocol does exactly what the name suggests: It synchronizes with other snapshotter peers over a smart contract running on Powerloom Prost chain. It follows an architecture that is driven by state transitions which makes it easy to understand and modify.
Because of its decentralized nature, the snapshotter specification and its implementations share some powerful features that can adapt to your specific information requirements on blockchain applications:
- Each data point is calculated, updated, and synchronized with other snapshotter peers participating in the network
- synchronization of data points is defined as a function of an epoch ID(identifier) where epoch refers to an equally spaced collection of blocks on the data source blockchain (for eg, Ethereum Mainnet/Polygon Mainnet/Polygon Testnet -- Mumbai). This simplifies the building of use cases that are stateful (i.e. can be accessed according to their state at a given height of the data source chain), synchronized, and depend on reliable data. For example,
- dashboards by offering higher-order aggregate datapoints
- trading strategies and bots
- a snapshotter peer can load past epochs, indexes, and aggregates from a decentralized state and have access to a rich history of data
- all the datasets are decentralized on IPFS/Filecoin
- the power of these decentralized storage networks can be leveraged fully by applying the principle of composability
Snapshotter Lite Node is a lightweight implementation of the Snapshotter Peer that can be used to build simpler use cases that do not require the full functionality of the Snapshotter Peer. Snapshotter Lite Node only has a single dependency on Python environment and hence has significantly lower resource requirements than the Snapshotter Peer. It is suitable for use cases where no aggregation is required and the data can be directly used from the base snapshots.
Architecturally, the Snapshotter Lite Node is similar to the Snapshotter Peer, it still has the same modular and highly configurable architecture, allowing for easy customization and seamless integration. It consists of three core components:
-
Main Snapshotter Codebase:
- This foundational component defines all the essential interfaces and handles a wide range of tasks, from listening to epoch release events to distributing tasks and managing snapshot submissions.
-
Configuration Files:
- Configuration files, located in the
/config
directory are linked to snapshotter-configs repo, play a pivotal role in defining project types, specifying paths for individual compute modules, and managing various project-related settings.
- Configuration files, located in the
-
Compute Modules:
- The heart of the system resides in the
snapshotter/modules
directory are linked to snapshotter-computes, where the actual computation logic for each project type is defined. These modules drive the snapshot generation process for specific project types.
- The heart of the system resides in the
The architecture has been designed to facilitate the seamless interchange of configuration and modules. To achieve this, we maintain these components in separate Git repositories, which are then integrated into the Snapshotter Peer using Git Submodules. As a result, adapting the system to different use cases is as straightforward as changing a Git branch, offering unparalleled flexibility and versatility.
For more information on using Git Submodules, please refer to the Git Submodules Documentation.
Snapshotter Lite Node is a lightweight implementation of the Snapshotter Peer and hence does not support the following features:
- Aggregation and data composition
- Redis Caching
- Worker Pool for parallel processing
- Preloading Architecture
- Epoch Processing state transitions
- Internal Status APIs
- Data Source Signalling
If you require any of these features, please consider using the Snapshotter Peer instead. If you are unsure about which implementation to use, please reach out to us on Discord and we will be happy to help you out.
An epoch denotes a range of block heights on the EVM-compatible data source blockchain, for eg Ethereum mainnet/Polygon PoS mainnet/testnet. This makes it easier to collect state transitions and snapshots of data on equally spaced block height intervals, as well as to support future work on other lightweight anchor proof mechanisms like Merkle proofs, succinct proofs, etc.
The size of an epoch is configurable. Let that be referred to as size(E)
-
A trusted service keeps track of the head of the chain as it moves ahead, and a marker
h₀
against the max block height from the last released epoch. This makes the beginning of the next epoch,h₁ = h₀ + 1
-
Once the head of the chain has moved sufficiently ahead so that an epoch can be published, an epoch finalization service takes into account the following factors
- chain reorganization reports where the reorganized limits are a subset of the epoch qualified to be published
- a configurable ‘offset’ from the bleeding edge of the chain
and then publishes an epoch (h₁, h₂)
by sending a transaction to the protocol state smart contract deployed on the Prost Chain (anchor chain) so that h₂ - h₁ + 1 == size(E)
. The next epoch, therefore, is tracked from h₂ + 1
.
Each such transaction emits an EpochReleased
event
event EpochReleased(uint256 indexed epochId, uint256 begin, uint256 end, uint256 timestamp);
The epochId
here is incremented by 1 with every successive epoch release.
Processors as configured in config/projects.json
calculate snapshots for each task type based on the filtering or any criteria defined for snapshot generation against this epochId
which corresponds to collections of state observations and event logs between the blocks at height in the range [begin, end]
.
The project ID is ultimately generated in the following manner:
The snapshots generated are the fundamental data models on which higher-order aggregates and richer data points are built (By Full Nodes)
For situations where data sources are constantly changing or numerous, it is impractical to maintain an extensive list of them. In such cases, specific data sources need not be defined explicitly. You can just define a filtering criteria for the data source contract address and the snapshotter will automatically generate snapshots for all data sources that match the criteria.
For complex use cases with a lot of data requirements, it is recommended to use the Snapshotter Peer instead of the Snapshotter Lite Node because it employs a more efficient and scalable architecture.
All snapshots per project reach consensus on the protocol state contract which results in a SnapshotFinalized
event being triggered.
event SnapshotFinalized(uint256 indexed epochId, uint256 epochEnd, string projectId, string snapshotCid, uint256 timestamp);
The Snapshotter Lite Node consists of the following major components:
System Event Detector, defined in system_event_detector.py
, is the main entry point for Snapshotter Lite Node. It tracks events being triggered on the protocol state contract running on the anchor chain and calls appropriate classes to handle them.
Events being tracked by the system event detector are:
EpochReleased
- This event is emitted by the Protocol State Contract when an epoch is released. It is used to trigger the snapshot generation process.allSnapshottersUpdated
- This event is emitted by the Protocol State Contract when a new snapshotter peer is added or removed from the network. It is used to enable or disable the snapshot generation process (Might be removed closer to mainnet).DailyTaskCompletedEvent
- Each snapshotter lite peer needs to complete a daily task to be eligible for rewards. This event is emitted by the Protocol State Contract when a snapshotter lite peer completes its daily task, making it inactive for the rest of the day.DayStartedEvent
- This event is emitted by the Protocol State Contract when a new day starts. This is used to re-enable the snapshot generation process for all snapshotter lite peers.
The Processor Distributor, defined in processor_distributor.py
, acts upon the events received from the System Event Detector and distributes the processing tasks to the appropriate snapshot processors. It is also responsible for acting on allSnapshottersUpdated
, DailyTaskCompletedEvent
and DayStartedEvent
events to manage the snapshot generation process.
Extracting data from the blockchain state and generating the snapshot can be a complex task. The RpcHelper
, defined in utils/rpc.py
, has a bunch of helper functions to make this process easier. It handles all the retry
and caching
logic so that developers can focus on efficiently building their use cases.
This component is one of the most important and allows you to access the finalized protocol state on the smart contract running on the anchor chain. Find it in core_api.py
.
The pooler-frontend that serves the Uniswap v2 dashboards hosted by the PowerLoom foundation on locations like https://uniswapv2.powerloom.io/ is a great example of a frontend specific web application that makes use of this API service.
Among many things, the core API allows you to access the finalized CID as well as its contents at a given epoch ID for a project.
The main endpoint implementations can be found as follows:
snapshotter-lite-v2/snapshotter/core_api.py
Line 237 in 2fa7e09
snapshotter-lite-v2/snapshotter/core_api.py
Line 293 in 2fa7e09
The first endpoint in GET /last_finalized_epoch/{project_id}
returns the last finalized EpochId for a given project ID and the second one is GET /data/{epoch_id}/{project_id}/
which can be used to return the actual snapshot data for a given EpochId and ProjectId.
These endpoints along with the combination of a bunch of other helper endpoints present in Core API
can be used to build powerful Dapps and dashboards.
You can observe the way it is used in pooler-frontend
repo to fetch the dataset for the aggregate projects of top pairs trade volume and token reserves summary:
try {
response = await axios.get(API_PREFIX+`/data/${epochInfo.epochId}/${top_pairs_7d_project_id}/`);
console.log('got 7d top pairs', response.data);
if (response.data) {
for (let pair of response.data.pairs) {
pairsData7d[pair.name] = pair;
}
} else {
throw new Error(JSON.stringify(response.data));
}
}
catch (e){
console.error('7d top pairs', e);
}
There are multiple ways to set up the Snapshotter Lite Node. You can either use the Docker image or run it directly on your local machine. However, it is recommended to use the Docker image as it is the easiest and most reliable way to set up the Snapshotter Lite Node.
NOTE: It is recommended to run build.sh
in a screen or tmux session so that the process continues running even after you close the terminal.
-
Install Docker on your machine. You can find the installation instructions for your operating system on the official Docker website.
-
Clone this repository using the following command:
git clone https://github.com/PowerLoom/snapshotter-lite-v2.git powerloom-pre-mainnet
This will clone the repository into a directory named
powerloom
. -
Change your working directory to the
powerloom-pre-mainnet
directory:cd powerloom-pre-mainnet
-
Run
build.sh
to start the snapshotter lite node:./build.sh
If you're a developer and want to play around with the code, instead of running
build.sh
, you can run the following commands to start the snapshotter lite node:./bootstrap.sh ./build-dev.sh
-
When prompted, enter
SOURCE_RPC_URL
,SIGNER_ACCOUNT_ADDRESS
,SIGNER_ACCOUNT_PRIVATE_KEY
,SLOT_ID
(only required for the first time), this will create a.env
file in the root directory of the project. -
This should start your snapshotter node and you should see something like this in your terminal logs
snapshotter-lite-v2-snapshotter-lite-v2-1 | 0|core-api | May 28, 2024 > 16:19:02 | INFO | Application startup complete. | {}
snapshotter-lite-v2-snapshotter-lite-v2-1 | 1|snapshotter-lite | May 28, 2024 > 16:19:02 | DEBUG | Loading web3 provider for node https://rpc-prost1h-proxy.powerloom.io| {'module': 'RpcHelper'}
snapshotter-lite-v2-snapshotter-lite-v2-1 | 1|snapshotter-lite | May 28, 2024 > 16:19:02 | INFO | Reporting service pinged successfully| {'module': 'EventDetector'}
snapshotter-lite-v2-snapshotter-lite-v2-1 | 1|snapshotter-lite | May 28, 2024 > 16:19:03 | INFO | Current block: 6964653| {'module': 'EventDetector'}
snapshotter-lite-v2-snapshotter-lite-v2-1 | 1|snapshotter-lite | May 28, 2024 > 16:19:03 | DEBUG | Loading web3 provider for node https://rpc-prost1h-proxy.powerloom.io| {'module': 'RpcHelper'}
snapshotter-lite-v2-snapshotter-lite-v2-1 | 1|snapshotter-lite | May 28, 2024 > 16:19:03 | DEBUG | Set source chain block time to 12.0| {'module': 'ProcessDistributor'}
snapshotter-lite-v2-snapshotter-lite-v2-1 | 1|snapshotter-lite | May 28, 2024 > 16:19:05 | INFO | Snapshotter active: True| {'module': 'ProcessDistributor'}
snapshotter-lite-v2-snapshotter-lite-v2-1 | 1|snapshotter-lite | May 28, 2024 > 16:19:05 | DEBUG | Loading web3 provider for node https://rpc-prost1h-proxy.powerloom.io| {'module': 'RpcHelper'}
snapshotter-lite-v2-snapshotter-lite-v2-1 | 0|core-api | May 28, 2024 > 16:19:06 | INFO | 127.0.0.1:44338 - "GET /health HTTP/1.1" 200 | {}
- To stop the node, you can press
Ctrl+C
in the terminal where the node is running ordocker-compose down
in a new terminal window from the project directory.
If you want to run the Snapshotter Lite Node without Docker, you need to make sure that you have Git, and Python 3.10.13 installed on your machine. You can find the installation instructions for your operating system on the official Python website.
-
Clone this repository using the following command:
git clone https://github.com/PowerLoom/snapshotter-lite-v2.git powerloom-pre-mainnet
This will clone the repository into a directory named
powerloom
. -
Change your working directory to the
powerloom-pre-mainnet
directory:cd powerloom-pre-mainnet
-
Run
init.sh
to start the snapshotter lite node:./init.sh
-
When prompted, enter
SOURCE_RPC_URL
,SIGNER_ACCOUNT_ADDRESS
,SIGNER_ACCOUNT_PRIVATE_KEY
,SLOT_ID
(only required for the first time), this will create a.env
file in the root directory of the project. -
Your node should start in background and you should start seeing logs in your terminal.
-
To stop the node, you can run
pkill -f snapshotter
in a new terminal window.
To enable Telegram reporting for snapshotter issues:
- Open a new conversation with @PowerloomReportingBot in the Telegram App.
- Start the bot by typing the
/start
command in the chat. You will receive a response containing yourChat ID
for the bot. - Enter the
Chat ID
when prompted on node startup. - You will now receive an error report whenever your node fails to process an epoch or snapshot.
Usually the easiest way to fix node related issues is to restart the node. If you're facing issues with the node, you can try going through the logs present in the logs
directory. If you're unable to find the issue, you can reach out to us on Discord and we will be happy to help you out.
We use pre-commit hooks to ensure our code quality is maintained over time. For this contributors need to do a one-time setup by running the following commands.
- Install the required dependencies using
pip install -r dev-requirements.txt
, this will set up everything needed for pre-commit checks. - Run
pre-commit install
Now, whenever you commit anything, it'll automatically check the files you've changed/edited for code quality issues and suggest improvements.
Pooler is a Uniswap specific implementation of what is known as a 'snapshotter' in the PowerLoom Protocol ecosystem. It synchronizes with other snapshotter peers over a smart contract running on the present version of the PowerLoom Protocol testnet. It follows an architecture that is driven by state transitions which makes it easy to understand and modify. This present release ultimately provide access to rich aggregates that can power a Uniswap v2 dashboard with the following data points:
- Total Value Locked (TVL)
- Trade Volume, Liquidity reserves, Fees earned
- grouped by
- Pair contracts
- Individual tokens participating in pair contract
- aggregated over time periods
- 24 hours
- 7 days
- grouped by
- Transactions containing
Swap
,Mint
, andBurn
events
In this section, let us take a look at the data composition abilities of Pooler to build on the base snapshot being built that captures information on Uniswap trades.
Required reading:
As you can notice in config/projects.example.json
, each project config needs to have the following components
project_type
(unique identifier prefix for the usecase, used to generate project ID)projects
(smart contracts to extract data from, pooler can generate different snapshots from multiple sources as long as the Contract ABI is same)processor
(the actual compuation logic reference, while you can write the logic anywhere, it is recommended to write your implementation in snapshotter/modules folder)
There's currently no limitation on the number or type of usecases you can build using snapshotter. Just write the Processor class and pooler libraries will take care of the rest.
{
"config": [{
"project_type": "uniswap_pairContract_pair_total_reserves",
"projects":[
"0xb4e16d0168e52d35cacd2c6185b44281ec28c9dc",
"0xae461ca67b15dc8dc81ce7615e0320da1a9ab8d5",
"0x0d4a11d5eeaac28ec3f61d100daf4d40471f1852",
"0x3041cbd36888becc7bbcbc0045e3b1f144466f5f",
"0xd3d2e2692501a5c9ca623199d38826e513033a17",
"0xbb2b8038a1640196fbe3e38816f3e67cba72d940",
"0xa478c2975ab1ea89e8196811f51a7b7ade33eb11"
],
"processor":{
"module": "pooler.modules.uniswapv2.pair_total_reserves",
"class_name": "PairTotalReservesProcessor"
}
},
{
"project_type": "uniswap_pairContract_trade_volume",
"projects":[
"0xb4e16d0168e52d35cacd2c6185b44281ec28c9dc",
"0xae461ca67b15dc8dc81ce7615e0320da1a9ab8d5",
"0x0d4a11d5eeaac28ec3f61d100daf4d40471f1852",
"0x3041cbd36888becc7bbcbc0045e3b1f144466f5f",
"0xd3d2e2692501a5c9ca623199d38826e513033a17",
"0xbb2b8038a1640196fbe3e38816f3e67cba72d940",
"0xa478c2975ab1ea89e8196811f51a7b7ade33eb11"
],
"processor":{
"module": "pooler.modules.uniswapv2.trade_volume",
"class_name": "TradeVolumeProcessor"
}
}
]
}
If we take a look at the TradeVolumeProcessor
class present at snapshotter/modules/computes/trade_volume.py
it implements the interface of GenericProcessorSnapshot
defined in snapshotter/utils/callback_helpers.py
.
There are a couple of important concepts here necessary to write your extraction logic:
compute
is the main function where most of the snapshot extraction and generation logic needs to be written. It receives the following inputs:
msg_obj
(SnapshotProcessMessage
instance, contains all the necessary epoch related information to generate snapshots)rpc_helper
(RpcHelper
instance to help with any calls to the data source contract's chain)anchor_rpc_helper
(RpcHelper
instance to help with any calls to the protocol state contract's chain)ipfs_reader
(async IPFS client to read the data from IPFS)protocol_state_contract
(protocol state contract instance to read the finalized snapshot CID or anything else from the protocol state contract required for snapshot generation)
Output format can be anything depending on the usecase requirements. Although it is recommended to use proper pydantic
models to define the snapshot interface.
The resultant output model in this specific example is UniswapTradesSnapshot
as defined in the Uniswap v2 specific modules directory: utils/models/message_models.py
. This encapsulates state information captured by TradeVolumeProcessor
between the block heights of the epoch: min_chain_height
and max_chain_height
.