Skip to main content
Solved

Avoid duplicated events ingestion from snowflake


Haniel Eliab López

I’m ingesting data from Snowflake. During the POC we ran into some duplicate events getting loaded due to not having the right warehouse size combined with retrying loading. I believe it was mentioned that if we provided a unique identifier for an event during ingestion we could prevent that in the future if any retries occurred. I don't see an option for how to provide an event id to amplitude in the ingestion documentation though. Any ideas on how to add that in?

 

Best answer by Saish Redkar

Hi @Haniel Eliab López 

You’ll have to send an insert_id field for each record. You can find the additional supported fields in the HTTP docs - https://www.docs.developers.amplitude.com/analytics/apis/http-v2-api/#keys-for-the-event-argument

View original
Did this topic help you find an answer to your question?

3 replies

Saish Redkar
Expert
Forum|alt.badge.img+10
  • Expert
  • 1382 replies
  • Answer
  • January 4, 2024

Hi @Haniel Eliab López 

You’ll have to send an insert_id field for each record. You can find the additional supported fields in the HTTP docs - https://www.docs.developers.amplitude.com/analytics/apis/http-v2-api/#keys-for-the-event-argument


Haniel Eliab López

Thanks @Saish Redkar ,

Just to validate. Does this insert_id also applies for the snowflake ingestion? We are not using the HTTP v2 API
Thanks!


Saish Redkar
Expert
Forum|alt.badge.img+10

Yes, their Snowflake ingestion is based on the HTTP API underneath.

We send the insert_id along with the other requisite fields/columns in our Snowflake Ingest config SQL.


Reply


Cookie policy

We use cookies to enhance and personalize your experience. If you accept you agree to our full cookie policy. Learn more about our cookies.

 
Cookie settings