-
Notifications
You must be signed in to change notification settings - Fork 91
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
sdcrypt0
committed
Apr 22, 2024
1 parent
3a90ccb
commit ad8c891
Showing
1 changed file
with
65 additions
and
221 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,247 +1,91 @@ | ||
import BigNumber from "bignumber.js"; | ||
import { CHAINS, PROTOCOLS, AMM_TYPES } from "./sdk/config"; | ||
import { getLPValueByUserAndPoolFromPositions, getPositionAtBlock, getPositionDetailsFromPosition, getPositionsForAddressByPoolAtBlock, getTimestampAtBlock, getTradeLiquidityForAddressByPoolAtBlock } from "./sdk/subgraphDetails"; | ||
import { getPositionDetailsFromPosition, getPositionsForAddressByPoolAtBlock, getTimestampAtBlock, getTradeLiquidityForAddressByPoolAtBlock } from "./sdk/subgraphDetails"; | ||
(BigInt.prototype as any).toJSON = function () { | ||
return this.toString(); | ||
}; | ||
|
||
import { promisify } from 'util'; | ||
import stream from 'stream'; | ||
import csv from 'csv-parser'; | ||
import fs from 'fs'; | ||
import { format } from 'fast-csv'; | ||
import { write } from 'fast-csv'; | ||
import { pipeline as streamPipeline } from 'stream'; | ||
import { captureRejectionSymbol } from "events"; | ||
import { readBlocksFromApi } from "./sdk/blockApi"; | ||
|
||
|
||
//Uncomment the following lines to test the getPositionAtBlock function | ||
|
||
// const position = getPositionAtBlock( | ||
// 0, // block number 0 for latest block | ||
// 2, // position id | ||
// CHAINS.L2_CHAIN_ID, // chain id | ||
// PROTOCOLS.PROTOCOL_NAME, // protocol | ||
// AMM_TYPES.UNISWAPV3 // amm type | ||
// ); | ||
// position.then((position) => { | ||
// // print response | ||
// const result = getPositionDetailsFromPosition(position); | ||
// console.log(`${JSON.stringify(result,null, 4)} | ||
// `) | ||
// }); | ||
|
||
interface LPValueDetails { | ||
pool: string; | ||
lpValue: string; | ||
} | ||
|
||
interface UserLPData { | ||
totalLP: string; | ||
pools: LPValueDetails[]; | ||
} | ||
|
||
// Define an object type that can be indexed with string keys, where each key points to a UserLPData object | ||
interface OutputData { | ||
[key: string]: UserLPData; | ||
} | ||
|
||
interface CSVRow { | ||
user: string; | ||
pool: string; | ||
block: number; | ||
position: number; | ||
lpvalue: string; | ||
} | ||
|
||
|
||
const pipeline = promisify(stream.pipeline); | ||
|
||
// Assuming you have the following functions and constants already defined | ||
// getPositionsForAddressByPoolAtBlock, CHAINS, PROTOCOLS, AMM_TYPES, getPositionDetailsFromPosition, getLPValueByUserAndPoolFromPositions, BigNumber | ||
|
||
// const readBlocksFromCSV = async (filePath: string): Promise<number[]> => { | ||
// const blocks: number[] = []; | ||
// await pipeline( | ||
// fs.createReadStream(filePath), | ||
// csv(), | ||
// async function* (source) { | ||
// for await (const chunk of source) { | ||
// // Assuming each row in the CSV has a column 'block' with the block number | ||
// if (chunk.block) blocks.push(parseInt(chunk.block, 10)); | ||
// } | ||
// } | ||
// ); | ||
// return blocks; | ||
// }; | ||
|
||
|
||
// const getData = async () => { | ||
// const snapshotBlocks = [ | ||
// 3116208, 3159408, 3202608, 3245808, 3289008, 3332208, | ||
// 3375408, 3418608, 3461808, 3505008, 3548208, 3591408, | ||
// 3634608, 3677808, 3721008, 3764208, 3807408, 3850608, | ||
// 3893808, 3937008, 3980208, 3983003, | ||
// ]; //await readBlocksFromCSV('src/sdk/L2_CHAIN_ID_chain_daily_blocks.csv'); | ||
|
||
// const csvRows: CSVRow[] = []; | ||
|
||
// for (let block of snapshotBlocks) { | ||
// const positions = await getPositionsForAddressByPoolAtBlock( | ||
// block, "", "", CHAINS.L2_CHAIN_ID, PROTOCOLS.PROTOCOL_NAME, AMM_TYPES.UNISWAPV3 | ||
// ); | ||
|
||
// console.log(`Block: ${block}`); | ||
// console.log("Positions: ", positions.length); | ||
|
||
// // Assuming this part of the logic remains the same | ||
// let positionsWithUSDValue = positions.map(getPositionDetailsFromPosition); | ||
// let lpValueByUsers = getLPValueByUserAndPoolFromPositions(positionsWithUSDValue); | ||
|
||
// lpValueByUsers.forEach((value, key) => { | ||
// let positionIndex = 0; // Define how you track position index | ||
// value.forEach((lpValue, poolKey) => { | ||
// const lpValueStr = lpValue.toString(); | ||
// // Accumulate CSV row data | ||
// csvRows.push({ | ||
// user: key, | ||
// pool: poolKey, | ||
// block, | ||
// position: positions.length, // Adjust if you have a specific way to identify positions | ||
// lpvalue: lpValueStr, | ||
// }); | ||
// }); | ||
// }); | ||
// } | ||
|
||
// // Write the CSV output to a file | ||
// const ws = fs.createWriteStream('outputData.csv'); | ||
// write(csvRows, { headers: true }).pipe(ws).on('finish', () => { | ||
// console.log("CSV file has been written."); | ||
// }); | ||
// }; | ||
|
||
|
||
|
||
|
||
interface BlockData { | ||
blockNumber: number; | ||
blockTimestamp: number; | ||
} | ||
|
||
|
||
const readBlocksFromCSV = async (filePath: string): Promise<BlockData[]> => { | ||
const blocks: BlockData[] = []; | ||
|
||
await new Promise<void>((resolve, reject) => { | ||
fs.createReadStream(filePath) | ||
.pipe(csv({ separator: '\t' })) // Specify the separator as '\t' for TSV files | ||
.on('data', (row) => { | ||
const blockNumber = parseInt(row.number, 10); | ||
const blockTimestamp = parseInt(row.block_timestamp, 10); | ||
if (!isNaN(blockNumber) && blockTimestamp) { | ||
blocks.push({ blockNumber: blockNumber, blockTimestamp }); | ||
} | ||
}) | ||
.on('end', () => { | ||
resolve(); | ||
}) | ||
.on('error', (err) => { | ||
reject(err); | ||
}); | ||
}); | ||
|
||
return blocks; | ||
}; | ||
|
||
type OutputDataSchemaRow = { | ||
block_number: number; | ||
timestamp: number; | ||
user_address: string; | ||
token_address: string; | ||
token_balance: bigint; | ||
token_symbol: string; | ||
token_balance: number; | ||
token_symbol?: string; | ||
usd_price: number; | ||
}; | ||
async function main() { | ||
const snapshotBlocks = [3137407, 3138307]; | ||
|
||
// readBlocksFromApi(1711324800, 1711411200).then(async (blocks) => { // uncomment for block api | ||
(async () => snapshotBlocks)().then(async (blocks) => { | ||
console.log(blocks); | ||
const allCsvRows: any[] = []; // Array to accumulate CSV rows for all blocks | ||
const batchSize = 10; // Size of batch to trigger writing to the file | ||
let i = 0; | ||
|
||
for (const block of blocks) { | ||
try { | ||
const timestamp = new Date(await getTimestampAtBlock(block)).toISOString(); | ||
const positions = await getPositionsForAddressByPoolAtBlock( | ||
block, "", "", CHAINS.L2_CHAIN_ID, PROTOCOLS.METAVAULT, AMM_TYPES.UNISWAPV3 | ||
); | ||
|
||
console.log(`Block: ${block}`); | ||
console.log("Positions: ", positions.length); | ||
|
||
// Assuming this part of the logic remains the same | ||
let positionsWithUSDValue = positions.map(getPositionDetailsFromPosition); | ||
// let lpValueByUsers = getLPValueByUserAndPoolFromPositions(positionsWithUSDValue); | ||
|
||
positionsWithUSDValue.forEach((value, key) => { | ||
|
||
// Accumulate CSV row data | ||
allCsvRows.push({ | ||
block_number: block, | ||
timestamp, | ||
user: value.owner, | ||
token_address: value.token0.id, | ||
token_balance: value.token0DecimalValue, | ||
token_balance_usd: value.token0USDValue, | ||
}); | ||
export const getUserTVLByBlock = async ({ | ||
blockNumber, | ||
blockTimestamp, | ||
}: BlockData): Promise<OutputDataSchemaRow[]> => { | ||
return await getPoolData({ blockNumber, blockTimestamp }); | ||
} | ||
|
||
allCsvRows.push({ | ||
block_number: block, | ||
timestamp, | ||
user: value.owner, | ||
token_address: value.token1.id, | ||
token_balance: value.token1DecimalValue, | ||
token_balance_usd: value.token1USDValue, | ||
}); | ||
export const getPoolData = async ({ | ||
blockNumber, | ||
blockTimestamp, | ||
}: BlockData): Promise<OutputDataSchemaRow[]> => { | ||
const allCsvRows: OutputDataSchemaRow[] = []; // Array to accumulate CSV rows for all blocks | ||
try { | ||
// const blockTimestamp = new Date(await getTimestampAtBlock(blockNumber)).toISOString(); | ||
const positions = await getPositionsForAddressByPoolAtBlock( | ||
blockNumber, "", "", CHAINS.L2_CHAIN_ID, PROTOCOLS.METAVAULT, AMM_TYPES.UNISWAPV3 | ||
); | ||
|
||
console.log(`Block: ${blockNumber}`); | ||
console.log("Positions: ", positions.length); | ||
|
||
// Assuming this part of the logic remains the same | ||
let positionsWithUSDValue = positions.map(getPositionDetailsFromPosition); | ||
// let lpValueByUsers = getLPValueByUserAndPoolFromPositions(positionsWithUSDValue); | ||
|
||
positionsWithUSDValue.forEach((value, key) => { | ||
// Accumulate CSV row data | ||
if (value.token0DecimalValue > 0) { | ||
allCsvRows.push({ | ||
block_number: blockNumber, | ||
timestamp: blockTimestamp, | ||
user_address: value.owner, | ||
token_address: value.token0.id, | ||
token_balance: value.token0DecimalValue, | ||
usd_price: 0, | ||
}); | ||
const liquidities = await getTradeLiquidityForAddressByPoolAtBlock( | ||
block, "", "", CHAINS.L2_CHAIN_ID, PROTOCOLS.METAVAULT, AMM_TYPES.TRADE | ||
); | ||
liquidities.forEach((value, key) => { | ||
allCsvRows.push({ | ||
block_number: block, | ||
timestamp, | ||
user: value.user, | ||
token_address: value.asset, | ||
token_balance: value.amount, | ||
token_balance_usd: value.amountUsd, | ||
}); | ||
} | ||
if (value.token1DecimalValue > 0) { | ||
allCsvRows.push({ | ||
block_number: blockNumber, | ||
timestamp: blockTimestamp, | ||
user_address: value.owner, | ||
token_address: value.token1.id, | ||
token_balance: value.token1DecimalValue, | ||
usd_price: 0, | ||
}); | ||
|
||
// Write the CSV output to a file | ||
const ws = fs.createWriteStream('outputData.csv'); | ||
write(allCsvRows, { headers: true }).pipe(ws).on('finish', () => { | ||
console.log("CSV file has been written."); | ||
} | ||
}); | ||
const liquidities = await getTradeLiquidityForAddressByPoolAtBlock( | ||
blockNumber, "", "", CHAINS.L2_CHAIN_ID, PROTOCOLS.METAVAULT, AMM_TYPES.TRADE | ||
); | ||
liquidities.forEach((value, key) => { | ||
if (value.amount > 0) { | ||
allCsvRows.push({ | ||
block_number: blockNumber, | ||
timestamp: blockTimestamp, | ||
user_address: value.user, | ||
token_address: value.asset, | ||
token_balance: value.amount, | ||
usd_price: 0, | ||
}); | ||
} catch (error) { | ||
console.error(`An error occurred for block ${block}:`, error); | ||
} | ||
|
||
} | ||
}) | ||
.catch((err) => { | ||
console.error('Error reading CSV file:', err); | ||
}); | ||
|
||
|
||
} catch (error) { | ||
console.error(`An error occurred for block ${blockNumber}:`, error); | ||
} | ||
return allCsvRows | ||
} | ||
main().then(() => { | ||
console.log("Done"); | ||
}); | ||
// getPrice(new BigNumber('1579427897588720602142863095414958'), 6, 18); //Uniswap | ||
// getPrice(new BigNumber('3968729022398277600000000'), 18, 6); //PROTOCOL_NAME | ||
|