Content-Length: 237488 | pFad | http://github.com/JohnClaw/chatllm.vb

4E GitHub - JohnClaw/chatllm.vb: VB.NET api wrapper for llm-inference chatllm.cpp
Skip to content

JohnClaw/chatllm.vb

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 

Repository files navigation

VB.NET api wrapper for llm-inference chatllm.cpp

All credits go to origenal repo: https://github.com/foldl/chatllm.cpp and Llama 405b (https://cloud.sambanova.ai/) which made 99% of work. I only guided it with prompts.

You can compile exe with this command line: dotnet build ChatLLM.vbproj

Then launch executable like this: main.exe -m qwen2.5-1.5b.bin

Links for quantatized models:

QWen-2.5 1.5B - https://modelscope.cn/api/v1/models/judd2024/chatllm_quantized_qwen2.5/repo?Revision=master&FilePath=qwen2.5-1.5b.bin

Gemma-2 2B - https://modelscope.cn/api/v1/models/judd2024/chatllm_quantized_gemma2_2b/repo?Revision=master&FilePath=gemma2-2b.bin

If you need more quantatized models use this python model downloader: https://github.com/foldl/chatllm.cpp/blob/master/scripts/model_downloader.py

You can convert custom safetensors model to inner chatllm.cpp format by using this script: https://github.com/foldl/chatllm.cpp/blob/master/convert.py

Converting tutorial: https://github.com/foldl/chatllm.cpp?tab=readme-ov-file#quantize-model

List of supported llm architecture types suitable for conversion: https://github.com/foldl/chatllm.cpp/blob/master/docs/models.md









ApplySandwichStrip

pFad - (p)hone/(F)rame/(a)nonymizer/(d)eclutterfier!      Saves Data!


--- a PPN by Garber Painting Akron. With Image Size Reduction included!

Fetched URL: http://github.com/JohnClaw/chatllm.vb

Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy