scy, I'm looking for a #compression #algorithm or something.
I have a process that generates a JSON document (> 1 MB, < 1 GB) once per week. These documents will be pretty similar. Some data will be modified, some will be added.
I'd like to keep all of these documents, in a compressed way, benefiting from the similarities between them, as if I'd compressed a concatenation of all of them, but without having to recompress everything each week.
Ideas? If possible, only using #Python's standard lib.