Aan de slagGa gratis aan de slag

Using extracted conversation facts

Now to apply your LLM with structured outputs! The llm_with_output you created in the previous exercise is still available, and the prompt template for this LLM has already been provided.

Deze oefening maakt deel uit van de cursus

Graph RAG with LangChain and Neo4j

Cursus bekijken

Oefeninstructies

  • Pipe the chat prompt template provided into the LLM with structured output (llm_with_output).
  • Invoke the chain on the user and session ID provided, passing it the messages from the history (history).

Praktische interactieve oefening

Probeer deze oefening eens door deze voorbeeldcode in te vullen.

chain = ChatPromptTemplate.from_messages([
    SystemMessagePromptTemplate.from_template("""
    	You are talking to {user}, the user of the application. 
        Any facts in the first person relate to {user}.
        The Session ID is {session_id}.
    	Extract the facts from the conversation and return them in the format of object, subject, predicate.
    """),
    # Pipe the chat prompt template into the LLM with structured output
    MessagesPlaceholder(variable_name="history")
]) | ____

# Invoke the chain, passing it the messages from the history
chain.invoke({"user": USER, "session_id": SESSION_ID, "history": ____})
Code bewerken en uitvoeren