Description
LocalAI version:
3.0.0
Environment, CPU architecture, OS, and Version:
Docker GPU CUDA 12
Describe the bug
DBG context local model name not found, setting to default defaultModelName=stablediffusion
DreamShaper is one image model that worked with AIO or -extra image
DBG context local model name not found, setting to the first model first model name=bark-cpp-small, TTS doesn't do anything, never starts.
Also notice other models don't work now.
gemma-3-4b-it
bunny-llama-3-8b-v
LocalAI-functioncall-phi-4-v0.3
deepseek-r1-distill-llama-8b
Backends installed
bark-cpp cuda12-bark-development cuda12-diffusers cuda12-kokoro-development cuda12-transformers cuda12-transformers-development
Models installed:
stablediffusion
stable-diffusion-3-medium
bark-cpp-small
To Reproduce
chat to a model
Expected behavior
chats back
Logs
12:40AM DBG context local model name not found, setting to default defaultModelName=stablediffusion
12:40AM DBG Parameter Config: &{PredictionOptions:{BasicModelRequest:{Model:DreamShaper_8_pruned.safetensors} Language: Translate:false N:0 TopP:0xc0abe51600 TopK:0xc0abe51608 Temperature:0xc0abe51610 Maxtokens:0xc0abe51640 Echo:false Batch:0 IgnoreEOS:false RepeatPenalty:0 RepeatLastN:0 Keep:0 FrequencyPenalty:0 PresencePenalty:0 TFZ:0xc0abe51638 TypicalP:0xc0abe51630 Seed:0xc0abe51650 NegativePrompt: RopeFreqBase:0 RopeFreqScale:0 NegativePromptScale:0 ClipSkip:0 Tokenizer:} Name:dreamshaper F16:0xc0abe515e9 Threads:0xc0abe515f0 Debug:0xc09cd01660 Roles:map[] Embeddings:0xc0abe51649 Backend:diffusers TemplateConfig:{Chat: ChatMessage: Completion: Edit: Functions: UseTokenizerTemplate:false JoinChatMessagesByCharacter:<nil> Multimodal: JinjaTemplate:false ReplyPrefix:} KnownUsecaseStrings:[FLAG_IMAGE FLAG_VIDEO FLAG_ANY] KnownUsecases:<nil> Pipeline:{TTS: LLM: Transcription: VAD:} PromptStrings:[cool pink sports car] InputStrings:[] InputToken:[] functionCallString: functionCallNameString: ResponseFormat: ResponseFormatMap:map[] FunctionsConfig:{DisableNoAction:false GrammarConfig:{ParallelCalls:false DisableParallelNewLines:false MixedMode:false NoMixedFreeString:false NoGrammar:false Prefix: ExpectStringsAfterJSON:false PropOrder: SchemaType: GrammarTriggers:[]} NoActionFunctionName: NoActionDescriptionName: ResponseRegex:[] JSONRegexMatch:[] ArgumentRegex:[] ArgumentRegexKey: ArgumentRegexValue: ReplaceFunctionResults:[] ReplaceLLMResult:[] CaptureLLMResult:[] FunctionNameKey: FunctionArgumentsKey:} FeatureFlag:map[] LLMConfig:{SystemPrompt: TensorSplit: MainGPU: RMSNormEps:0 NGQA:0 PromptCachePath: PromptCacheAll:false PromptCacheRO:false MirostatETA:0xc0abe51628 MirostatTAU:0xc0abe51620 Mirostat:0xc0abe51618 NGPULayers:<nil> MMap:0xc0abe51648 MMlock:0xc0abe51649 LowVRAM:0xc0abe51649 Reranking:0xc0abe51649 Grammar: StopWords:[] Cutstrings:[] ExtractRegex:[] TrimSpace:[] TrimSuffix:[] ContextSize:0xc0abe51658 NUMA:false LoraAdapter: LoraBase: LoraAdapters:[] LoraScales:[] LoraScale:0 NoMulMatQ:false DraftModel: NDraft:0 Quantization: LoadFormat: GPUMemoryUtilization:0 TrustRemoteCode:false EnforceEager:false SwapSpace:0 MaxModelLen:0 TensorParallelSize:0 DisableLogStatus:false DType: LimitMMPerPrompt:{LimitImagePerPrompt:0 LimitVideoPerPrompt:0 LimitAudioPerPrompt:0} MMProj: FlashAttention:false NoKVOffloading:false CacheTypeK: CacheTypeV: RopeScaling: ModelType: YarnExtFactor:0 YarnAttnFactor:0 YarnBetaFast:0 YarnBetaSlow:0 CFGScale:0} Diffusers:{CUDA:true PipelineType:StableDiffusionPipeline SchedulerType:k_dpmpp_2m EnableParameters:negative_prompt,num_inference_steps IMG2IMG:false ClipSkip:0 ClipModel: ClipSubFolder: ControlNet:} Step:25 GRPC:{Attempts:0 AttemptsSleepTime:0} TTSConfig:{Voice: AudioPath:} CUDA:false DownloadFiles:[] Description: Usage: Options:[]}
Additional context