Fix vanilla model answer in example benchmark (#219)
This commit is contained in:
parent
72b01a9909
commit
2a69f1574e
|
@ -253,7 +253,7 @@
|
|||
"\n",
|
||||
" if is_vanilla_llm:\n",
|
||||
" llm = agent\n",
|
||||
" answer = str(llm([{\"role\": \"user\", \"content\": question}]))\n",
|
||||
" answer = str(llm([{\"role\": \"user\", \"content\": question}]).content)\n",
|
||||
" token_count = {\"input\": llm.last_input_token_count, \"output\": llm.last_output_token_count}\n",
|
||||
" intermediate_steps = str([])\n",
|
||||
" else:\n",
|
||||
|
|
Loading…
Reference in New Issue