|
@@ -0,0 +1,103 @@
|
|
|
+//go:build slow
|
|
|
+
|
|
|
+package convert
|
|
|
+
|
|
|
+import (
|
|
|
+ "os"
|
|
|
+ "path/filepath"
|
|
|
+ "testing"
|
|
|
+
|
|
|
+ "github.com/ollama/ollama/llm"
|
|
|
+)
|
|
|
+
|
|
|
+func convertFull(t *testing.T, p string) (llm.KV, llm.Tensors) {
|
|
|
+ t.Helper()
|
|
|
+
|
|
|
+ mf, err := GetModelFormat(p)
|
|
|
+ if err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+
|
|
|
+ params, err := mf.GetParams(p)
|
|
|
+ if err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+
|
|
|
+ arch, err := mf.GetModelArch("", p, params)
|
|
|
+ if err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+
|
|
|
+ if err := arch.LoadVocab(); err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+
|
|
|
+ if err := arch.GetTensors(); err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+
|
|
|
+ f, err := os.CreateTemp(t.TempDir(), "f16")
|
|
|
+ if err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+ defer f.Close()
|
|
|
+
|
|
|
+ if err := arch.WriteGGUF(f); err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+
|
|
|
+ r, err := os.Open(f.Name())
|
|
|
+ if err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+ defer r.Close()
|
|
|
+
|
|
|
+ m, _, err := llm.DecodeGGML(r)
|
|
|
+ if err != nil {
|
|
|
+ t.Fatal(err)
|
|
|
+ }
|
|
|
+
|
|
|
+ return m.KV(), m.Tensors()
|
|
|
+}
|
|
|
+
|
|
|
+func TestConvertFull(t *testing.T) {
|
|
|
+ cases := []struct {
|
|
|
+ path string
|
|
|
+ arch string
|
|
|
+ tensors int
|
|
|
+ layers int
|
|
|
+ }{
|
|
|
+ {"Meta-Llama-3-8B-Instruct", "llama", 291, 35},
|
|
|
+ {"Mistral-7B-Instruct-v0.2", "llama", 291, 35},
|
|
|
+ {"Mixtral-8x7B-Instruct-v0.1", "llama", 291, 35},
|
|
|
+ {"gemma-2b-it", "gemma", 164, 20},
|
|
|
+ }
|
|
|
+
|
|
|
+ for _, tt := range cases {
|
|
|
+ t.Run(tt.path, func(t *testing.T) {
|
|
|
+ p := filepath.Join("testdata", tt.path)
|
|
|
+ if _, err := os.Stat(p); err != nil {
|
|
|
+ t.Skipf("%s not found", p)
|
|
|
+ }
|
|
|
+
|
|
|
+ kv, tensors := convertFull(t, p)
|
|
|
+
|
|
|
+ if kv.Architecture() != tt.arch {
|
|
|
+ t.Fatalf("expected llama, got %s", kv.Architecture())
|
|
|
+ }
|
|
|
+
|
|
|
+ if kv.FileType().String() != "F16" {
|
|
|
+ t.Fatalf("expected F16, got %s", kv.FileType())
|
|
|
+ }
|
|
|
+
|
|
|
+ if len(tensors) != tt.tensors {
|
|
|
+ t.Fatalf("expected %d tensors, got %d", tt.tensors, len(tensors))
|
|
|
+ }
|
|
|
+
|
|
|
+ layers := tensors.Layers()
|
|
|
+ if len(layers) != tt.layers {
|
|
|
+ t.Fatalf("expected %d layers, got %d", tt.layers, len(layers))
|
|
|
+ }
|
|
|
+ })
|
|
|
+ }
|
|
|
+}
|