model=tiny_llm,
tokenizer=tokenizer)
result = nlp(QA_input)
Print the question, answer, grounding...Answer = assemble_grounding_sources(result[‘answer’], context_metadata)
print(f"Question: {question}"...answer = assemble_grounding_sources(answer, grounding_sources)
return answer
Generate response...response = generate_response(
llm="gpt-3.5-turbo-1106",
temperature=0.0,
grounding_sources...",
user_content=f"question: {QUESTION}, context: {context}")
Print the question, answer, grounding