Skip to content

Nodenester/ConsoleLLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 
 
 

Repository files navigation

ConsoleLLM

Status: Archived

A simple console-based chat interface for running local LLMs. Built November 2023.

What it does

Loads a local GGUF/GGML model and runs an interactive chat session in the terminal. The user types messages and the model streams responses back in real time. Uses a "chat-with-bob" system prompt format with a helpful assistant persona.

  • Streams token-by-token output to the console
  • Runs inference on GPU via CUDA 12
  • Color-coded user input (green) vs. assistant output (white)

Tech stack

  • C# / .NET 8
  • LLamaSharp 0.7.0 (C# bindings for llama.cpp)
  • LLamaSharp.Backend.Cuda12 for GPU acceleration
  • Tested with WizardLM-7B and Orca-2-7B models

Setup

  1. Place a GGUF model file somewhere on disk
  2. Update modelPath in Program.cs to point to your model
  3. dotnet run

About

This was an early experiment in local LLM integration -- one of my first attempts at building something with language models. It's minimal by design: a single Program.cs file, no abstractions, just a direct chat loop.

License

MIT

About

Console interface for local LLM inference via LLamaSharp — November 2023

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages