writer.go 6.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275
  1. package loader
  2. import (
  3. "encoding/json"
  4. "fmt"
  5. "io"
  6. "os"
  7. )
  8. // Writer helps create .mak files with section-based format
  9. type Writer struct {
  10. file *os.File
  11. header Header
  12. metadata Metadata
  13. tensors []TensorEntry
  14. tensorDataPos uint64 // absolute file offset of tensor data section start
  15. tensorOffset uint64 // relative offset within tensor data section
  16. tokenizerData []byte
  17. closed bool
  18. }
  19. // NewWriter creates a new MAK file writer
  20. func NewWriter(path string) (*Writer, error) {
  21. f, err := os.Create(path)
  22. if err != nil {
  23. return nil, err
  24. }
  25. w := &Writer{
  26. file: f,
  27. header: Header{
  28. Magic: [4]byte{'M', 'A', 'K', 'A'},
  29. Version: Version,
  30. Flags: FlagLittleEndian,
  31. Alignment: Alignment,
  32. },
  33. metadata: Metadata{
  34. Tensors: make(map[string]TensorEntry),
  35. },
  36. tensors: make([]TensorEntry, 0),
  37. }
  38. // Reserve space for header + section table. We keep this fixed so we can stream
  39. // tensor data immediately, then seek back and write the final header/table.
  40. //
  41. // Max sections: metadata + tensor data + tokenizer.
  42. const maxSections = uint32(3)
  43. prefixSize := uint64(HeaderSize + int(maxSections)*SectionEntrySize)
  44. w.tensorDataPos = alignOffset(prefixSize, uint64(Alignment))
  45. if _, err := f.Seek(int64(w.tensorDataPos), io.SeekStart); err != nil {
  46. _ = f.Close()
  47. return nil, fmt.Errorf("seek tensor data start: %w", err)
  48. }
  49. return w, nil
  50. }
  51. // SetModelConfig sets the model configuration
  52. func (w *Writer) SetModelConfig(cfg ModelConfig) {
  53. w.metadata.ModelConfig = cfg
  54. }
  55. // AddTokenizer adds tokenizer data to the file
  56. func (w *Writer) AddTokenizer(data []byte) {
  57. w.tokenizerData = data
  58. }
  59. // AddTensor adds a tensor to the file
  60. func (w *Writer) AddTensor(name string, dtype DType, shape []uint64, data []byte) error {
  61. if w.closed || w.file == nil {
  62. return fmt.Errorf("writer is closed")
  63. }
  64. // Align to alignment boundary
  65. padding := (uint64(Alignment) - (w.tensorOffset % uint64(Alignment))) % uint64(Alignment)
  66. if padding > 0 {
  67. if err := writeZeros(w.file, padding); err != nil {
  68. return fmt.Errorf("write tensor padding: %w", err)
  69. }
  70. w.tensorOffset += padding
  71. }
  72. entry := TensorEntry{
  73. Name: name,
  74. DType: dtype,
  75. Shape: shape,
  76. Offset: w.tensorOffset,
  77. Size: uint64(len(data)),
  78. }
  79. w.tensors = append(w.tensors, entry)
  80. w.metadata.Tensors[name] = entry
  81. if err := writeAll(w.file, data); err != nil {
  82. return fmt.Errorf("write tensor %s: %w", name, err)
  83. }
  84. w.tensorOffset += entry.Size
  85. return nil
  86. }
  87. // AddTensorFromReader streams tensor data from r into the file without keeping it in memory.
  88. func (w *Writer) AddTensorFromReader(name string, dtype DType, shape []uint64, r io.Reader, size uint64) error {
  89. if w.closed || w.file == nil {
  90. return fmt.Errorf("writer is closed")
  91. }
  92. // Align to alignment boundary
  93. padding := (uint64(Alignment) - (w.tensorOffset % uint64(Alignment))) % uint64(Alignment)
  94. if padding > 0 {
  95. if err := writeZeros(w.file, padding); err != nil {
  96. return fmt.Errorf("write tensor padding: %w", err)
  97. }
  98. w.tensorOffset += padding
  99. }
  100. entry := TensorEntry{
  101. Name: name,
  102. DType: dtype,
  103. Shape: shape,
  104. Offset: w.tensorOffset,
  105. Size: size,
  106. }
  107. w.tensors = append(w.tensors, entry)
  108. w.metadata.Tensors[name] = entry
  109. if err := copyExactly(w.file, r, int64(size)); err != nil {
  110. return fmt.Errorf("write tensor %s: %w", name, err)
  111. }
  112. w.tensorOffset += size
  113. return nil
  114. }
  115. // Close finalizes and closes the file
  116. func (w *Writer) Close() error {
  117. if w.closed {
  118. return nil
  119. }
  120. w.closed = true
  121. if w.file == nil {
  122. return nil
  123. }
  124. // Build sections
  125. // Section 1: Metadata JSON
  126. metaJSON, err := json.Marshal(w.metadata)
  127. if err != nil {
  128. _ = w.file.Close()
  129. return fmt.Errorf("marshal metadata: %w", err)
  130. }
  131. // Calculate offsets
  132. numSections := uint32(2) // TensorData + Metadata
  133. if len(w.tokenizerData) > 0 {
  134. numSections = 3 // + Tokenizer
  135. }
  136. tensorDataOffset := w.tensorDataPos
  137. tensorDataSize := w.tensorOffset
  138. metaOffset := alignOffset(tensorDataOffset+tensorDataSize, uint64(Alignment))
  139. metaSize := uint64(len(metaJSON))
  140. // Build section table
  141. sections := &SectionTable{
  142. Entries: []SectionEntry{
  143. {Type: SectionTensorData, Offset: tensorDataOffset, Size: tensorDataSize, Flags: 0},
  144. {Type: SectionMetadata, Offset: metaOffset, Size: metaSize, Flags: 0},
  145. },
  146. }
  147. // Write metadata after tensor data (streaming-friendly layout).
  148. curPos := tensorDataOffset + tensorDataSize
  149. if metaOffset < curPos {
  150. _ = w.file.Close()
  151. return fmt.Errorf("invalid metadata offset")
  152. }
  153. if err := writeZeros(w.file, metaOffset-curPos); err != nil {
  154. _ = w.file.Close()
  155. return fmt.Errorf("write metadata padding: %w", err)
  156. }
  157. if err := writeAll(w.file, metaJSON); err != nil {
  158. _ = w.file.Close()
  159. return fmt.Errorf("write metadata: %w", err)
  160. }
  161. // Add tokenizer section if present
  162. var tokenizerOffset uint64
  163. if len(w.tokenizerData) > 0 {
  164. tokenizerOffset = alignOffset(metaOffset+metaSize, uint64(Alignment))
  165. sections.Entries = append(sections.Entries, SectionEntry{
  166. Type: SectionTokenizer,
  167. Offset: tokenizerOffset,
  168. Size: uint64(len(w.tokenizerData)),
  169. Flags: 0,
  170. })
  171. curPos = metaOffset + metaSize
  172. if tokenizerOffset < curPos {
  173. _ = w.file.Close()
  174. return fmt.Errorf("invalid tokenizer offset")
  175. }
  176. if err := writeZeros(w.file, tokenizerOffset-curPos); err != nil {
  177. _ = w.file.Close()
  178. return fmt.Errorf("write tokenizer padding: %w", err)
  179. }
  180. if err := writeAll(w.file, w.tokenizerData); err != nil {
  181. _ = w.file.Close()
  182. return fmt.Errorf("write tokenizer: %w", err)
  183. }
  184. }
  185. // Update header
  186. w.header.SectionCount = numSections
  187. // Seek back and write header + section table.
  188. if _, err := w.file.Seek(0, io.SeekStart); err != nil {
  189. _ = w.file.Close()
  190. return fmt.Errorf("seek header: %w", err)
  191. }
  192. if err := WriteHeader(w.file, &w.header); err != nil {
  193. _ = w.file.Close()
  194. return fmt.Errorf("write header: %w", err)
  195. }
  196. if err := WriteSectionTable(w.file, sections); err != nil {
  197. _ = w.file.Close()
  198. return fmt.Errorf("write section table: %w", err)
  199. }
  200. return w.file.Close()
  201. }
  202. func alignOffset(offset, alignment uint64) uint64 {
  203. return offset + (alignment-(offset%alignment))%alignment
  204. }
  205. var zeroBuf = make([]byte, Alignment)
  206. func writeZeros(w io.Writer, n uint64) error {
  207. for n > 0 {
  208. chunk := n
  209. if chunk > uint64(len(zeroBuf)) {
  210. chunk = uint64(len(zeroBuf))
  211. }
  212. if err := writeAll(w, zeroBuf[:chunk]); err != nil {
  213. return err
  214. }
  215. n -= chunk
  216. }
  217. return nil
  218. }
  219. func writeAll(w io.Writer, b []byte) error {
  220. for len(b) > 0 {
  221. n, err := w.Write(b)
  222. if err != nil {
  223. return err
  224. }
  225. b = b[n:]
  226. }
  227. return nil
  228. }
  229. func copyExactly(dst io.Writer, src io.Reader, n int64) error {
  230. written, err := io.CopyN(dst, src, n)
  231. if err != nil {
  232. return err
  233. }
  234. if written != n {
  235. return fmt.Errorf("short write: %d != %d", written, n)
  236. }
  237. return nil
  238. }