convert_test.go 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477
  1. package convert
  2. import (
  3. "bytes"
  4. "crypto/sha256"
  5. "encoding/binary"
  6. "encoding/hex"
  7. "encoding/json"
  8. "flag"
  9. "fmt"
  10. "io"
  11. "io/fs"
  12. "log/slog"
  13. "math"
  14. "os"
  15. "path/filepath"
  16. "slices"
  17. "strings"
  18. "testing"
  19. "golang.org/x/exp/maps"
  20. "github.com/ollama/ollama/llm"
  21. )
  22. type tensorData struct {
  23. Offsets []int `json:"data_offsets"`
  24. Type string `json:"dtype"`
  25. Shape []int `json:"shape"`
  26. }
  27. func convertFull(t *testing.T, fsys fs.FS) (*os.File, llm.KV, *llm.Tensors) {
  28. t.Helper()
  29. f, err := os.CreateTemp(t.TempDir(), "f16")
  30. if err != nil {
  31. t.Fatal(err)
  32. }
  33. defer f.Close()
  34. if err := ConvertModel(fsys, f); err != nil {
  35. t.Fatal(err)
  36. }
  37. r, err := os.Open(f.Name())
  38. if err != nil {
  39. t.Fatal(err)
  40. }
  41. t.Cleanup(func() { r.Close() })
  42. m, _, err := llm.DecodeGGML(r, math.MaxInt)
  43. if err != nil {
  44. t.Fatal(err)
  45. }
  46. if _, err := r.Seek(0, io.SeekStart); err != nil {
  47. t.Fatal(err)
  48. }
  49. return r, m.KV(), m.Tensors()
  50. }
  51. func generateResultsJSON(t *testing.T, f *os.File, kv llm.KV, tensors *llm.Tensors) map[string]string {
  52. actual := make(map[string]string)
  53. for k, v := range kv {
  54. if s, ok := v.(json.Marshaler); !ok {
  55. actual[k] = fmt.Sprintf("%v", v)
  56. } else {
  57. bts, err := json.Marshal(s)
  58. if err != nil {
  59. t.Fatal(err)
  60. }
  61. actual[k] = fmt.Sprintf("%x", sha256.Sum256(bts))
  62. }
  63. }
  64. for _, tensor := range tensors.Items {
  65. sha256sum := sha256.New()
  66. sr := io.NewSectionReader(f, int64(tensors.Offset+tensor.Offset), int64(tensor.Size()))
  67. if _, err := io.Copy(sha256sum, sr); err != nil {
  68. t.Fatal(err)
  69. }
  70. actual[tensor.Name] = hex.EncodeToString(sha256sum.Sum(nil))
  71. }
  72. return actual
  73. }
  74. func TestMain(m *testing.M) {
  75. var level slog.Level
  76. flag.TextVar(&level, "level", slog.LevelInfo, "log level")
  77. flag.Parse()
  78. slog.SetLogLoggerLevel(level)
  79. os.Exit(m.Run())
  80. }
  81. func TestConvertModel(t *testing.T) {
  82. cases := []string{
  83. "Meta-Llama-3-8B-Instruct",
  84. "Meta-Llama-3.1-8B-Instruct",
  85. "Mistral-7B-Instruct-v0.2",
  86. "Mixtral-8x7B-Instruct-v0.1",
  87. "gemma-2b-it",
  88. "gemma-2-2b-it",
  89. // microsoft/Phi-3-mini-128-instruct@d548c233192db00165d842bf8edff054bb3212f8
  90. "Phi-3-mini-128k-instruct",
  91. "all-MiniLM-L6-v2",
  92. "gemma-2-9b-it",
  93. "Qwen2.5-0.5B-Instruct",
  94. }
  95. for i := range cases {
  96. tt := cases[i]
  97. t.Run(tt, func(t *testing.T) {
  98. t.Parallel()
  99. p := filepath.Join("testdata", tt)
  100. if testing.Short() {
  101. t.Skip("skipping in short mode")
  102. } else if _, err := os.Stat(p); err != nil {
  103. t.Skipf("%s not found", p)
  104. }
  105. f, kv, tensors := convertFull(t, os.DirFS(p))
  106. actual := generateResultsJSON(t, f, kv, tensors)
  107. expectFile, err := os.Open(filepath.Join("testdata", fmt.Sprintf("%s.json", tt)))
  108. if err != nil {
  109. t.Fatal(err)
  110. }
  111. var expect map[string]string
  112. if err := json.NewDecoder(expectFile).Decode(&expect); err != nil {
  113. t.Fatal(err)
  114. }
  115. keys := maps.Keys(expect)
  116. slices.Sort(keys)
  117. for _, k := range keys {
  118. if v, ok := actual[k]; !ok {
  119. t.Errorf("missing %s", k)
  120. } else if v != expect[k] {
  121. t.Errorf("unexpected %s: want %s, got %s", k, expect[k], v)
  122. }
  123. }
  124. })
  125. }
  126. }
  127. func TestConvertInvalidTensorNames(t *testing.T) {
  128. f, err := os.CreateTemp(t.TempDir(), "testmodel")
  129. if err != nil {
  130. t.Fatal(err)
  131. }
  132. defer f.Close()
  133. tempDir := t.TempDir()
  134. td := map[string]*tensorData{}
  135. offset := 4096
  136. td["model.layers.0.self_attn.q_proj.weight"] = &tensorData{
  137. Offsets: []int{0, offset},
  138. Type: "F32",
  139. Shape: []int{4096, 4096},
  140. }
  141. td["blk.0.attn_q.weight"] = &tensorData{
  142. Offsets: []int{offset, offset * 2},
  143. Type: "F32",
  144. Shape: []int{4096, 4096},
  145. }
  146. generateSafetensorTestData(t, tempDir, td)
  147. err = ConvertModel(os.DirFS(tempDir), f)
  148. if err == nil || !strings.HasPrefix(err.Error(), "duplicate tensor name") {
  149. t.Errorf("expected error but didn't get one")
  150. }
  151. }
  152. func TestConvertInvalidDatatype(t *testing.T) {
  153. f, err := os.CreateTemp(t.TempDir(), "testmodel")
  154. if err != nil {
  155. t.Fatal(err)
  156. }
  157. defer f.Close()
  158. tempDir := t.TempDir()
  159. td := map[string]*tensorData{}
  160. offset := 4096 * 14336
  161. td["model.layers.0.mlp.down_proj.weight"] = &tensorData{
  162. Offsets: []int{0, offset},
  163. Type: "I8",
  164. Shape: []int{4096, 14336},
  165. }
  166. td["model.layers.0.mlp.down_proj.weight_format"] = &tensorData{
  167. Offsets: []int{offset, offset},
  168. Type: "U8",
  169. Shape: []int{},
  170. }
  171. generateSafetensorTestData(t, tempDir, td)
  172. err = ConvertModel(os.DirFS(tempDir), f)
  173. if err == nil || err.Error() != "unsupported safetensors model" {
  174. t.Errorf("expected error but didn't get one")
  175. }
  176. }
  177. func generateSafetensorTestData(t *testing.T, tempDir string, tensorData map[string]*tensorData) {
  178. data, err := json.Marshal(tensorData)
  179. if err != nil {
  180. t.Fatal(err)
  181. }
  182. var buf bytes.Buffer
  183. l := int64(len(data))
  184. err = binary.Write(&buf, binary.LittleEndian, l)
  185. if err != nil {
  186. t.Fatal(err)
  187. }
  188. _, err = buf.Write(data)
  189. if err != nil {
  190. t.Fatal(err)
  191. }
  192. fdata, err := os.Create(filepath.Join(tempDir, "model-00001-of-00001.safetensors"))
  193. if err != nil {
  194. t.Fatal(err)
  195. }
  196. defer fdata.Close()
  197. _, err = fdata.Write(buf.Bytes())
  198. if err != nil {
  199. t.Fatal(err)
  200. }
  201. configData := `
  202. {
  203. "architectures": [
  204. "LlamaForCausalLM"
  205. ]
  206. }
  207. `
  208. f, err := os.Create(filepath.Join(tempDir, "config.json"))
  209. if err != nil {
  210. t.Fatal(err)
  211. }
  212. defer f.Close()
  213. _, err = f.WriteString(configData)
  214. if err != nil {
  215. t.Fatal(err)
  216. }
  217. tokenizerData := `
  218. {
  219. }
  220. `
  221. f, err = os.Create(filepath.Join(tempDir, "tokenizer.json"))
  222. if err != nil {
  223. t.Fatal(err)
  224. }
  225. defer f.Close()
  226. _, err = f.WriteString(tokenizerData)
  227. if err != nil {
  228. t.Fatal(err)
  229. }
  230. }
  231. func TestConvertAdapter(t *testing.T) {
  232. type AdapterCase struct {
  233. Name string
  234. BaseKV map[string]any
  235. Expected map[string]string
  236. }
  237. cases := []AdapterCase{
  238. {
  239. Name: "discollama",
  240. BaseKV: map[string]any{
  241. "general.architecture": "llama",
  242. "llama.attention.head_count": uint32(32),
  243. "llama.attention.head_count_kv": uint32(8),
  244. },
  245. Expected: map[string]string{
  246. "general.architecture": "llama",
  247. "general.file_type": "1",
  248. "general.parameter_count": "106496",
  249. "general.type": "adapter",
  250. "general.version": "v0.2",
  251. "adapter.lora.alpha": "16",
  252. "adapter.type": "lora",
  253. "llama.attention.head_count": "32",
  254. "llama.attention.head_count_kv": "8",
  255. "blk.31.attn_q.weight.lora_a": "0eb3318b02cd313429bcc7621b539fdbb10240fea190c56c9e5f93fcd37a4e50",
  256. "blk.31.attn_q.weight.lora_b": "0eb3318b02cd313429bcc7621b539fdbb10240fea190c56c9e5f93fcd37a4e50",
  257. "blk.31.attn_v.weight.lora_a": "0eb3318b02cd313429bcc7621b539fdbb10240fea190c56c9e5f93fcd37a4e50",
  258. "blk.31.attn_v.weight.lora_b": "071dcafe89df065d6e1c935ecb8fdf6479b3c202eb912e7da938597673ff5857",
  259. },
  260. },
  261. }
  262. for _, c := range cases {
  263. t.Run(c.Name, func(t *testing.T) {
  264. t.Parallel()
  265. f, err := os.CreateTemp(t.TempDir(), "f16")
  266. if err != nil {
  267. t.Fatal(err)
  268. }
  269. defer f.Close()
  270. tempDir := t.TempDir()
  271. generateLoraTestData(t, tempDir)
  272. if err = ConvertAdapter(os.DirFS(tempDir), f, c.BaseKV); err != nil {
  273. t.Fatal(err)
  274. }
  275. r, err := os.Open(f.Name())
  276. if err != nil {
  277. t.Fatal(err)
  278. }
  279. defer r.Close()
  280. m, _, err := llm.DecodeGGML(r, math.MaxInt)
  281. if err != nil {
  282. t.Fatal(err)
  283. }
  284. if _, err := r.Seek(0, io.SeekStart); err != nil {
  285. t.Fatal(err)
  286. }
  287. actual := generateResultsJSON(t, r, m.KV(), m.Tensors())
  288. keys := maps.Keys(c.Expected)
  289. slices.Sort(keys)
  290. for _, k := range keys {
  291. if v, ok := actual[k]; !ok {
  292. t.Errorf("missing %s", k)
  293. } else if v != c.Expected[k] {
  294. t.Errorf("unexpected %s: want %s, got %s", k, c.Expected[k], v)
  295. }
  296. }
  297. })
  298. }
  299. }
  300. func generateLoraTestData(t *testing.T, tempDir string) {
  301. offset := 4096 * 8 * 4
  302. td := map[string]*tensorData{"__metadata__": nil}
  303. td["model.layers.31.self_attn.q_proj.lora_a"] = &tensorData{
  304. Offsets: []int{0, offset},
  305. Type: "F32",
  306. Shape: []int{4096, 8},
  307. }
  308. td["model.layers.31.self_attn.q_proj.lora_b"] = &tensorData{
  309. Offsets: []int{offset, offset * 2},
  310. Type: "F32",
  311. Shape: []int{8, 4096},
  312. }
  313. td["model.layers.31.self_attn.v_proj.lora_a"] = &tensorData{
  314. Offsets: []int{offset * 2, offset * 3},
  315. Type: "F32",
  316. Shape: []int{4096, 8},
  317. }
  318. td["model.layers.31.self_attn.v_proj.lora_b"] = &tensorData{
  319. Offsets: []int{offset * 3, offset*3 + 8*1024*4},
  320. Type: "F32",
  321. Shape: []int{8, 1024},
  322. }
  323. data, err := json.Marshal(td)
  324. if err != nil {
  325. t.Fatal(err)
  326. }
  327. var buf bytes.Buffer
  328. l := int64(len(data))
  329. err = binary.Write(&buf, binary.LittleEndian, l)
  330. if err != nil {
  331. t.Fatal(err)
  332. }
  333. _, err = buf.Write(data)
  334. if err != nil {
  335. t.Fatal(err)
  336. }
  337. // write some data for the tensors
  338. ones := make([]float32, 4096*8)
  339. for i := range ones {
  340. ones[i] = float32(1)
  341. }
  342. for range 3 {
  343. err = binary.Write(&buf, binary.LittleEndian, ones)
  344. if err != nil {
  345. t.Fatal(err)
  346. }
  347. }
  348. ones = make([]float32, 1024*8)
  349. for i := range ones {
  350. ones[i] = float32(1)
  351. }
  352. err = binary.Write(&buf, binary.LittleEndian, ones)
  353. if err != nil {
  354. t.Fatal(err)
  355. }
  356. fdata, err := os.Create(filepath.Join(tempDir, "adapters.safetensors"))
  357. if err != nil {
  358. t.Fatal(err)
  359. }
  360. defer fdata.Close()
  361. _, err = fdata.Write(buf.Bytes())
  362. if err != nil {
  363. t.Fatal(err)
  364. }
  365. configData := `
  366. {
  367. "adapter_path": "adapters-test",
  368. "batch_size": 8,
  369. "config": "config-tiny.json",
  370. "data": "../discollama-completion",
  371. "grad_checkpoint": null,
  372. "iters": 1000,
  373. "learning_rate": 1e-05,
  374. "lora_layers": 1,
  375. "lora_parameters": {
  376. "rank": 8,
  377. "alpha": 16,
  378. "dropout": 0.0,
  379. "scale": 2.0
  380. },
  381. "lr_schedule": null,
  382. "max_seq_length": 2048,
  383. "model": "/Users/pdevine/git/Meta-Llama-3-8B-Instruct",
  384. "resume_adapter_file": null,
  385. "save_every": 100,
  386. "seed": 0,
  387. "steps_per_eval": 200,
  388. "steps_per_report": 10,
  389. "test": false,
  390. "test_batches": 500,
  391. "train": true,
  392. "use_dora": false,
  393. "val_batches": 25
  394. }
  395. `
  396. f, err := os.Create(filepath.Join(tempDir, "adapter_config.json"))
  397. if err != nil {
  398. t.Fatal(err)
  399. }
  400. defer f.Close()
  401. _, err = f.WriteString(configData)
  402. if err != nil {
  403. t.Fatal(err)
  404. }
  405. }