https://huggingface.co/TheBloke/WizardLM-Uncensored-Falcon-40B-GGML https://huggingface.co/TheBloke/falcon-7b-instruct-GGML https://huggingface.co/TheBloke/WizardLM-Uncensored-Falcon-7B-GGML The official HF models are here: https://huggingface.co/tiiuae/falcon-40b/ ...
curl http://localhost:8080/v1/edits -H "Content-Type: application/json" -d '{ "model": "ggml-koala-7b-model-q4_0-r2.bin", "instruction": "rephrase", "input": "Black cat jumped out of the window", "temperature": 0.7 }' ...
LLAMA and LLAMA2 (LLaMA / Alpaca / GPT4All / Vicuna / Koala / Pygmalion 7B / Metharme 7B / WizardLM and many more) GPT-2 / Cerebras GPT-J RWKV GPT-NeoX / Pythia / StableLM / Dolly / RedPajama MPT models Falcon (GGUF only) Stable Diffusion and SDXL modelsAbout...
curl http://localhost:8080/v1/edits -H "Content-Type: application/json" -d '{ "model": "ggml-koala-7b-model-q4_0-r2.bin", "instruction": "rephrase", "input": "Black cat jumped out of the window", "temperature": 0.7 }' ...
llama-b1380-bin-win-avx2-x64.zip) From the unzipped folder, open a terminal/cmd window here and place a pre-converted .gguf model file. Test out the main example like so: .\main -m llama-2-7b.Q4_0.gguf -n 128 Memory/Disk Requirements As the models are currently fully loaded ...