fix: Adding an LLM param to fix broken generator from llamacpp (#1519)
This commit is contained in:
		
							parent
							
								
									e326126d0d
								
							
						
					
					
						commit
						869233f0e4
					
				|  | @ -42,7 +42,7 @@ class LLMComponent: | |||
|                     context_window=settings.llm.context_window, | ||||
|                     generate_kwargs={}, | ||||
|                     # All to GPU | ||||
|                     model_kwargs={"n_gpu_layers": -1}, | ||||
|                     model_kwargs={"n_gpu_layers": -1, "offload_kqv": True}, | ||||
|                     # transform inputs into Llama2 format | ||||
|                     messages_to_prompt=prompt_style.messages_to_prompt, | ||||
|                     completion_to_prompt=prompt_style.completion_to_prompt, | ||||
|  |  | |||
		Loading…
	
		Reference in New Issue