Looks great, do you have any concrete data how much money it will save ?
Also, how does it compare to for example GptCache[0] ? or any other semantic cache solution[1] ?
[0] https://gptcache.readthedocs.io/en/latest/
[1] https://portkey.ai/blog/reducing-llm-costs-and-latency-seman...
Looks great, do you have any concrete data how much money it will save ?
Also, how does it compare to for example GptCache[0] ? or any other semantic cache solution[1] ?
[0] https://gptcache.readthedocs.io/en/latest/
[1] https://portkey.ai/blog/reducing-llm-costs-and-latency-seman...