HF backend model additions: Yi-34B-Chat, Openchat-3.5, Tulu-2, DeepSeek and Mixtral (#29)
* Add openchat_3.5, Yi-34B-Chat, Orca-2-13b model support; Add custom openchat and ChatML chat templates; Add 'slow' tokenizer usage handling * Add culling of openchat EOS token at end of outputs * Remove Yi-34B-Chat and Orca-2-13b from supported models list due to issues found in testing * Add Yi-34B-Chat output processing; Add Yi-34B-Chat to supported models list * Add DeepSeek 7B/67B, Tulu 7B/70B and Mixtral 8x7B models * Changed transformers version requirement to 4.36.0 * Changed accelerate version requirement to 0.25.0 * Add output EOS culling for DeepSeek
Please register or sign in to comment