Add non-working partial yzma implementation
This commit is contained in:
parent
a542f436f3
commit
c62d126bdc
3 changed files with 61 additions and 0 deletions
45
main.go
45
main.go
|
|
@ -2,9 +2,12 @@ package main
|
|||
|
||||
import (
|
||||
"embed"
|
||||
"fmt"
|
||||
"html/template"
|
||||
"net/http"
|
||||
"time"
|
||||
|
||||
"github.com/hybridgroup/yzma/pkg/llama"
|
||||
)
|
||||
|
||||
//go:embed templates/index.html
|
||||
|
|
@ -99,5 +102,47 @@ func main() {
|
|||
// Serve embedded headshots file
|
||||
http.Handle("/headshots/", http.FileServer(http.FS(headshots)))
|
||||
|
||||
// TODO: derive from something?
|
||||
libPath := "/nix/store/jml3vhvay9yy94qj8bmmhbf2dhx6q2n1-llama-cpp-7356/lib"
|
||||
modelFile := "./SmolLM-135M.Q2_K.gguf"
|
||||
prompt := "Are you ready to go?"
|
||||
responseLength := int32(128)
|
||||
|
||||
llama.Load(libPath)
|
||||
llama.LogSet(llama.LogSilent())
|
||||
llama.Init()
|
||||
|
||||
model, _ := llama.ModelLoadFromFile(modelFile, llama.ModelDefaultParams())
|
||||
lctx, _ := llama.InitFromModel(model, llama.ContextDefaultParams())
|
||||
|
||||
vocab := llama.ModelGetVocab(model)
|
||||
|
||||
// get tokens from the prompt
|
||||
tokens := llama.Tokenize(vocab, prompt, true, false)
|
||||
|
||||
batch := llama.BatchGetOne(tokens)
|
||||
|
||||
sampler := llama.SamplerChainInit(llama.SamplerChainDefaultParams())
|
||||
llama.SamplerChainAdd(sampler, llama.SamplerInitGreedy())
|
||||
|
||||
for pos := int32(0); pos < responseLength; pos += batch.NTokens {
|
||||
llama.Decode(lctx, batch)
|
||||
token := llama.SamplerSample(sampler, lctx, -1)
|
||||
|
||||
if llama.VocabIsEOG(vocab, token) {
|
||||
fmt.Println()
|
||||
break
|
||||
}
|
||||
|
||||
buf := make([]byte, 36)
|
||||
len := llama.TokenToPiece(vocab, token, buf, 0, true)
|
||||
|
||||
fmt.Print(string(buf[:len]))
|
||||
|
||||
batch = llama.BatchGetOne([]llama.Token{token})
|
||||
}
|
||||
|
||||
fmt.Println()
|
||||
|
||||
http.ListenAndServe(":8080", nil)
|
||||
}
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue