The author further shares their unsuccessful attempt to use Mozilla Llamafiles for the Mixtral-8x7B-Instruct due to Windows' inability to run an EXE larger than 4GB. They express their frustration at being lost in the fast-paced world of local models and their need for guidance on how to understand these technologies. They acknowledge a gap in their fundamental understanding of these technologies and seek advice on how to fill it.
Key takeaways:
- The author has experience with various text generation technologies, including Talk to a Transformer, Markov chains, Umbrel LlamaGPT, and webGPU WebGPT.
- The author struggles to understand how these technologies work beyond following installation instructions, particularly with HuggingFace and the release of the llama models.
- The author attempted to use the Mozilla Llamafiles for the Mixtral-8x7B-Instruct, but encountered issues with windows not running an EXE bigger than 4GB.
- The author expresses a need for guidance on how to learn about local models in a rapidly evolving field.