|
@@ -271,7 +271,7 @@ func NewLlamaServer(gpus discover.GpuInfoList, modelPath string, f *ggml.GGML, a
|
|
|
|
|
|
var llamaModel *llama.Model
|
|
|
var textProcessor model.TextProcessor
|
|
|
- if envconfig.NewEngine() {
|
|
|
+ if envconfig.NewEngine() || f.KV().OllamaEngineRequired() {
|
|
|
textProcessor, err = model.NewTextProcessor(modelPath)
|
|
|
if err != nil {
|
|
|
// To prepare for opt-out mode, instead of treating this as an error, we fallback to the old runner
|