Hello, is there a way to auto-dedupe that will retain new data? When turning on auto-dedupe it keeps the earliest version of the row data, I would like to retain the newest version.
Currently, Clay's auto-dedupe feature only retains the oldest row when detecting duplicates in a specified column. The system automatically deletes newer duplicate entries. There's no built-in setting to keep the newest data instead. You can access auto-dedupe settings by clicking the Auto-dedupe icon at the bottom right of your table and selecting "Enable automatic deduplication." If you need to retain newer data when deduplicating, you might need to use a different approach or request this as a feature enhancement.
Is there any work around, like adjusting table settings so new data is added to the top of the sheet?
No unfortunately, it will still delete the latest entry regardless of sorting
whats the usecase?
Hey there, thanks for reaching out! Being able to customize your auto-dedupe to delete either the newest or oldest record found is something our team is working on implementing! Would love to hear your use case as well so we could try and come up with an alternative in the meantime! :)
I’m using API to pull in email out reach campaign names and data. Then I’m exporting from Clay to a Google Excel spreadsheet. This process runs daily, and the most recent information is at the bottom of the sheet. The process runs fine since the excel sheet only received data from NEW cells, but I would prefer to have a cleaner Clay sheet and not have many duplicates being entered daily.
You can use AutoDedupe to automatically remove duplicates and keep your Clay table clean. Alternatively, the Prevent Duplicates (loom) method tags each row with an identifier, so only new rows get processed. Here’s a guide to help: 📘 https://docs.clay.com/en/articles/9642165-auto-update-and-auto-dedupe-table 🎥 https://www.loom.com/share/a39a4569d95244b7886026f3c70812c7 Let me know if you want help setting it up.
To confirm I'm understanding right, you don't want to prevent new duplicate data from coming in. You want this new duplicate data to replace the old data of the same information? Can you give me a little more context on the use case here so I can better troubleshoot?