Llama2-shepherd a CLI tool to install multiple implementations of the llama2

  • Are you planning on adding documentation / a mechanism for running a prompt using the code this installs?

    As far as I can tell at the moment it clones one of the various repos for you and downloads some model writes, but it doesn't yet help you compile and run the code.

  • It's sad that github has now defaults to showing a blank page devoid of any content related to the projects linked when js execution isn't complete.

  • Do you have a performance comparison for inference on the same hardware using each of these implementations?

  • Everyone uses gguf / safetensors? How model management is done for this kind of tool?