Or you just use the same concept from the early 90s where you have a dedicated routine for data as a library which is used instead and offers more benefits.
Might as well code it properly than recode it in a more generic way in another language.
Sure, I was doing genome compression for tree comparison in the late 90s for in memory analysis and the speed up was very significant. The idea of a universal specification format and a universal decoder is IMO bloody brilliant, hell with a bit of training I’d be unsurprised if you could point a LLM at a format and get 90+% of the way there with the specification for most simple-ish formats, and probably have a manual crack at simple relational databases. Potentially you could also switch out huffman for encryption for efficient zero-knowledge proofs as well.
Or you just use the same concept from the early 90s where you have a dedicated routine for data as a library which is used instead and offers more benefits.
Might as well code it properly than recode it in a more generic way in another language.
Sure, I was doing genome compression for tree comparison in the late 90s for in memory analysis and the speed up was very significant. The idea of a universal specification format and a universal decoder is IMO bloody brilliant, hell with a bit of training I’d be unsurprised if you could point a LLM at a format and get 90+% of the way there with the specification for most simple-ish formats, and probably have a manual crack at simple relational databases. Potentially you could also switch out huffman for encryption for efficient zero-knowledge proofs as well.
It could work out well. Be interesting to see if it’s widely adopted.