Commit History

Author SHA1 Message Date
  Georgi Gerganov 08f10f69c3 llama : remove notion of CLS token (#11064) 1 year ago
  Georgi Gerganov afa8a9ec9b llama : add `llama_vocab`, functions -> methods, naming (#11110) 1 year ago
  Georgi Gerganov f66f582927 llama : refactor `src/llama.cpp` (#10902) 1 year ago
  Georgi Gerganov 30caac3a68 llama : the WPM vocabs use the CLS token as BOS (#10930) 1 year ago
  wwoodsTM ff252ea48e llama : add DRY sampler (#9702) 1 year ago
  Georgi Gerganov 755a9b2bf0 llama : add infill sampler (#9896) 1 year ago
  Georgi Gerganov 11ac9800af llama : improve infill support and special token detection (#9798) 1 year ago
  Georgi Gerganov 8c475b97b8 rerank : use [SEP] token instead of [BOS] (#9737) 1 year ago
  Zhenwei Jin 6102037bbb vocab : refactor tokenizer to reduce init overhead (#9449) 1 year ago
  Georgi Gerganov 31ac5834fe llama : keep track of all EOG tokens in the vocab (#9609) 1 year ago
  Georgi Gerganov df270ef745 llama : refactor sampling v2 (#9294) 1 year ago
  Zhenwei Jin 4af8420afb common : remove duplicate function llama_should_add_bos_token (#8778) 1 year ago
  fairydreaming d3f0c7166a Stop the generation when <|eom_id|> token is encountered - needed for Llama 3.1 tool call support (#8858) 1 year ago
  Georgi Gerganov 938943cdbf llama : move vocab, grammar and sampling into separate files (#8508) 1 year ago