Add functions

This commit is contained in:
Gui Heise 2022-02-17 11:08:21 -05:00
parent ac376311f4
commit c9536d1bde

View File

@ -14,78 +14,65 @@ EXPORT_BUCKET_REGION_ENV = "EXPORT_BUCKET_REGION"
EXPORT_AWS_ACCESS_KEY_ID_ENV = "EXPORT_AWS_ACCESS_KEY_ID" EXPORT_AWS_ACCESS_KEY_ID_ENV = "EXPORT_AWS_ACCESS_KEY_ID"
EXPORT_AWS_SECRET_ACCESS_KEY_ENV = "EXPORT_AWS_SECRET_ACCESS_KEY" EXPORT_AWS_SECRET_ACCESS_KEY_ENV = "EXPORT_AWS_SECRET_ACCESS_KEY"
MEV_SUMMARY_EXPORT_QUERY = """ supported_tables = [
SELECT to_json(mev_summary) "mev_summary",
FROM mev_summary "arbitrages",
WHERE "liquidations",
block_number = :block_number "sandwiches",
""" "sandwiched_swaps",
]
ARBITRAGES_EXPORT_QUERY = """
SELECT to_json(arbitrages)
FROM arbitrages
WHERE
block_number = :block_number
"""
LIQUIDATIONS_EXPORT_QUERY = """
SELECT to_json(liquidations)
FROM liquidations
WHERE
block_number = :block_number
"""
SANDWICHES_EXPORT_QUERY = """
SELECT to_json(sandwiches)
FROM sandwiches
WHERE
block_number = :block_number
"""
query_by_table = {
"mev_summary": MEV_SUMMARY_EXPORT_QUERY,
"arbitrages": ARBITRAGES_EXPORT_QUERY,
"liquidations": LIQUIDATIONS_EXPORT_QUERY,
"sandwiches": SANDWICHES_EXPORT_QUERY,
}
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
def export_block(inspect_db_session, block_number: int) -> None: def export_block(inspect_db_session, block_number: int) -> None:
export_bucket_name = get_export_bucket_name()
for table in supported_tables:
_export_table(inspect_db_session, block_number, table)
def _export_table(inspect_db_session, block_number: int, table: str) -> None:
client = get_s3_client() client = get_s3_client()
export_bucket_name = get_export_bucket_name()
export_statement = _get_export_statement(table)
for table in query_by_table.keys(): object_key = f"{table}/flashbots_{block_number}.json"
object_key = f"{table}/flashbots_{block_number}.json"
mev_summary_json_results = inspect_db_session.execute(
statement=query_by_table[table],
params={
"block_number": block_number,
},
)
first_value, mev_summary_json_results = _peek(mev_summary_json_results) mev_summary_json_results = inspect_db_session.execute(
if first_value is None: statement=export_statement,
existing_object_size = _get_object_size( params={
client, export_bucket_name, object_key "block_number": block_number,
) },
if existing_object_size is None or existing_object_size == 0: )
logger.info(f"Skipping {table} for block {block_number} - no data")
continue
mev_summary_json_fileobj = BytesIteratorIO( first_value, mev_summary_json_results = _peek(mev_summary_json_results)
( if first_value is None:
f"{json.dumps(row)}\n".encode("utf-8") existing_object_size = _get_object_size(client, export_bucket_name, object_key)
for (row,) in mev_summary_json_results if existing_object_size is None or existing_object_size == 0:
) logger.info(f"Skipping {table} for block {block_number} - no data")
) return
client.upload_fileobj( mev_summary_json_fileobj = BytesIteratorIO(
mev_summary_json_fileobj, (f"{json.dumps(row)}\n".encode("utf-8") for (row,) in mev_summary_json_results)
Bucket=export_bucket_name, )
Key=object_key,
)
logger.info(f"Exported to {object_key}") client.upload_fileobj(
mev_summary_json_fileobj,
Bucket=export_bucket_name,
Key=object_key,
)
logger.info(f"Exported to {object_key}")
def _get_export_statement(table: str) -> str:
return f"""
SELECT to_json({table})
FROM {table}
WHERE
block_number = :block_number
"""
def _get_object_size(client, bucket: str, key: str) -> Optional[int]: def _get_object_size(client, bucket: str, key: str) -> Optional[int]: