docs/source/working-notes/troyraen/SuperNNova/not_saving_to_bigquery.md
SuperNNova doesn’t seem to be saving results to BigQuery. What’s going on?
table = 'metadata'
today = '2021-10-13'
query = f"""
SELECT objectId, candid, publish_time__SuperNNova AS pt
FROM `{project_id}.{dataset}.{table}`
WHERE DATE(publish_time__SuperNNova) >= '{today}'
"""
metadf = gcp_utils.query_bigquery(query).to_dataframe()
candid = 1746528292515015009 # single candid from metadf
table = 'SuperNNova'
query = f"""
SELECT *
FROM `{project_id}.{dataset}.{table}`
WHERE candid={candid}
"""
snndf = gcp_utils.query_bigquery(query).to_dataframe()
# this is empty. confirms that results are not being saved.
Added logging to the Cloud Function. Now test it.
import json
from python_fncs.pubsub_consumer import Consumer
from python_fncs.cast_types import avro_to_dict
from base64 import b64decode, b64encode
# get an alert
consumer = Consumer('ztf-loop')
msgs = consumer.stream_alerts(parameters={'max_results': 1, 'max_backlog': 1})
alert_dict = avro_to_dict(msgs[0].data)
alert_dict = {k: v for k, v in alert_dict.items() if "cutout" not in k}
# publish it
gcp_utils.publish_pubsub('ztf-exgalac_trans_cf', alert_dict)
# pull down SuperNNova's output message
msgs = gcp_utils.pull_pubsub('ztf-SuperNNova-counter')
msg_dict = json.loads(msgs[0]) # this looks exactly as expected.
# query from bigquery table
candid = alert_dict['candid']
table = 'SuperNNova'
query = f"""
SELECT *
FROM `{project_id}.{dataset}.{table}`
WHERE candid={candid}
"""
testdf = gcp_utils.query_bigquery(query).to_dataframe()
did this 3 times
insert rows to bigquery was successful
Output Pub/Sub message looks as expected.
now it’s crashing because it doesn’t have enough memory
i don’t understand…?