Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] Langchain using chain.stream could not shown result on trulens dashboard #1357

Closed
zongyicoding opened this issue Aug 18, 2024 · 6 comments · Fixed by #1441
Closed

[BUG] Langchain using chain.stream could not shown result on trulens dashboard #1357

zongyicoding opened this issue Aug 18, 2024 · 6 comments · Fixed by #1441
Assignees
Labels
bug Something isn't working

Comments

@zongyicoding
Copy link

Bug Description
I am currently using langchain's chain.stream to batch output results to the frontend web page. After the output ends, I cannot correctly capture the output and scores of LLM (there is no data at all on the Dashboard). How can I correctly wait for the LLM output to be completed and then record the complete output in the Trulens dashboard and give the correct score?

To Reproduce

from fastapi import FastAPI
from fastapi.responses import StreamingResponse, FileResponse
from langchain_openai import ChatOpenAI
from langchain.schema import Document
from langchain_openai import OpenAIEmbeddings
from langchain_community.vectorstores.chroma import Chroma
from langchain_core.runnables import RunnablePassthrough
from langchain_core.prompts import ChatPromptTemplate
from langchain_core.output_parsers import StrOutputParser

from trulens_eval import TruChain, Feedback
from trulens_eval.feedback.provider import OpenAI

from trulens_eval import Tru

tru = Tru()
tru.run_dashboard(port=9999, force=True)

embedding_function = OpenAIEmbeddings(
            model="text-embedding-ada-002",
            check_embedding_ctx_length=False)

docs = [
    Document(
        page_content="the dog loves to eat pizza", metadata={"source": "animal.txt"}
    ),
    Document(
        page_content="the cat loves to eat lasagna", metadata={"source": "animal.txt"}
    ),
]

db = Chroma.from_documents(docs, embedding_function)
retriever = db.as_retriever()


template = """Answer the question based only on the following context:
{context}

Question: {question}
"""
prompt = ChatPromptTemplate.from_template(template)
model = ChatOpenAI(temperature=0,
                   streaming=True, 
                   model='gpt-3.5')

provider = OpenAI()
retrieval_chain = (
    {"context": retriever, "question": RunnablePassthrough()}
    | prompt
    | model
    | StrOutputParser()
)

f_answer_relevance = (
    Feedback(provider.context_relevance_with_cot_reasons, name = "Answer Relevance")
    .on_input_output()
)

tru_recorder = TruChain(retrieval_chain,
    app_id='conversation_stream',
    feedbacks=[f_answer_relevance],
    )

app = FastAPI()

async def generate_chat_responses(message):
    with tru_recorder:
        for chunk in retrieval_chain.stream(message):
            content = chunk.replace("\n", "<br>")
            print(f'content: {content}')
            yield f"data: {content}\n\n"

@app.get("/")
async def root():
    return FileResponse("static/index.html")

@app.get("/chat_stream/{message}")
async def chat_stream(message: str):
    return StreamingResponse(generate_chat_responses(message=message), media_type="text/event-stream")


if __name__ == "__main__":
    import uvicorn

    uvicorn.run(app, host="127.0.0.1", port=8000)

Expected behavior
A clear and concise description of what you expected to happen.
The following two behaviors are both acceptable.

  • Complete the waiting for the generation results of the streaming and record the complete information in the Dashboard.

  • After the LLM generation is completed, provide the response to Trulens for analysis and save it to the dashboard.

Environment:

  • OS: [e.g. MacOS, Windows] MacOS 14.5
  • Python Version 3.12
  • TruLens version 0.33.0
  • Versions of other relevant installed libraries

Additional context
Add any other context about the problem here.

@zongyicoding zongyicoding added the bug Something isn't working label Aug 18, 2024
Copy link

dosubot bot commented Aug 18, 2024

Hey, @zongyicoding! I'm here to assist you with any bugs, questions, or contributions. Let me know how I can help you with the Langchain issue.

To correctly wait for the LLM output to be completed when using langchain's chain.stream and then record the complete output in the Trulens dashboard, you can follow this approach:

  1. Set up the asynchronous task to call the chain with the provided inputs and a callback handler.
  2. Process the callback iterations to print each token as it is received.
  3. Ensure the task is completed and retrieve the recording for the Trulens dashboard.

Here's an updated version of your code to achieve this:

from fastapi import FastAPI, Request
from fastapi.responses import StreamingResponse, FileResponse
from langchain_openai import ChatOpenAI
from langchain.schema import Document
from langchain_openai import OpenAIEmbeddings
from langchain_community.vectorstores.chroma import Chroma
from langchain_core.runnables import RunnablePassthrough
from langchain_core.prompts import ChatPromptTemplate
from langchain_core.output_parsers import StrOutputParser

from trulens_eval import TruChain, Feedback
from trulens_eval.feedback.provider import OpenAI
from trulens_eval import Tru
from langchain.callbacks import AsyncIteratorCallbackHandler
import asyncio

tru = Tru()
tru.run_dashboard(port=9999, force=True)

embedding_function = OpenAIEmbeddings(
    model="text-embedding-ada-002",
    check_embedding_ctx_length=False
)

docs = [
    Document(
        page_content="the dog loves to eat pizza", metadata={"source": "animal.txt"}
    ),
    Document(
        page_content="the cat loves to eat lasagna", metadata={"source": "animal.txt"}
    ),
]

db = Chroma.from_documents(docs, embedding_function)
retriever = db.as_retriever()

template = """Answer the question based only on the following context:
{context}

Question: {question}
"""
prompt = ChatPromptTemplate.from_template(template)
model = ChatOpenAI(temperature=0, streaming=True, model='gpt-3.5')

provider = OpenAI()
retrieval_chain = (
    {"context": retriever, "question": RunnablePassthrough()}
    | prompt
    | model
    | StrOutputParser()
)

f_answer_relevance = (
    Feedback(provider.context_relevance_with_cot_reasons, name="Answer Relevance")
    .on_input_output()
)

tru_recorder = TruChain(
    retrieval_chain,
    app_id='conversation_stream',
    feedbacks=[f_answer_relevance],
    feedback_mode="deferred"
)

app = FastAPI()

async def generate_chat_responses(message):
    callback = AsyncIteratorCallbackHandler()
    
    with tru_recorder as recording:
        task = asyncio.create_task(
            retrieval_chain.acall(
                inputs=dict(question=message),
                callbacks=[callback]
            )
        )
    
    response = ""
    async for token in callback.aiter():
        content = token.replace("\n", "<br>")
        print(f'content: {content}')
        response += content
        yield f"data: {content}\n\n"
    
    await task
    record = recording.get()
    print(f'Record: {record}')

@app.get("/")
async def root():
    return FileResponse("static/index.html")

@app.get("/chat_stream/{message}")
async def chat_stream(message: str):
    return StreamingResponse(generate_chat_responses(message=message), media_type="text/event-stream")

if __name__ == "__main__":
    import uvicorn
    uvicorn.run(app, host="127.0.0.1", port=8000)

Key Changes:

  1. Callback Handler: Added AsyncIteratorCallbackHandler to handle streaming tokens.
  2. Asynchronous Task: Created an asynchronous task to call the chain and process the callback iterations.
  3. Recording: Ensured the task is completed and retrieved the recording for the Trulens dashboard.

This approach ensures that the streaming outputs are correctly captured and recorded in the Trulens dashboard, providing the correct score after the LLM generation is completed [1][2].

To continue talking to Dosu, mention @dosu.

@zongyicoding
Copy link
Author

zongyicoding commented Aug 19, 2024

hi @dosu.
After modifying the code, I encountered two issues.

  1. In my case, I used the api_key and base_url provided by the company. When I used OpenAI embedding, I couldn't invoke Trulens properly. The error message is as follows:
Exception has occurred: Exception
Feedback function FeedbackDefinition(Answer Relevance,
	selectors={'question': Lens().__record__.main_input, 'context': Lens().__record__.main_output},
	if_exists=None
) is not loadable. Cannot use DEFERRED feedback mode. The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable
openai.OpenAIError: The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable

The above exception was the direct cause of the following exception:

  File "/Applications/FastAPI-LangChain-Streaming-main/new_app.py", line 66, in <module>
    tru_recorder = TruChain(
                   ^^^^^^^^^
Exception: Feedback function FeedbackDefinition(Answer Relevance,
	selectors={'question': Lens().__record__.main_input, 'context': Lens().__record__.main_output},
	if_exists=None
) is not loadable. Cannot use DEFERRED feedback mode. The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable
  1. When I use my own private OPENAI_API_KEY, another error message pops up.
ERROR:    Exception in ASGI application
Traceback (most recent call last):
  File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/responses.py", line 265, in __call__
    await wrap(partial(self.listen_for_disconnect, receive))
  File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/responses.py", line 261, in wrap
    await func()
  File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/responses.py", line 238, in listen_for_disconnect
    message = await receive()
              ^^^^^^^^^^^^^^^
  File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/uvicorn/protocols/http/httptools_impl.py", line 555, in receive
    await self.message_event.wait()
  File "/opt/anaconda3/envs/delete_env/lib/python3.11/asyncio/locks.py", line 213, in wait
    await fut
  File "/opt/anaconda3/envs/delete_env/lib/python3.11/asyncio/futures.py", line 287, in __await__
    yield self  # This tells Task to wait for completion.
    ^^^^^^^^^^
  File "/opt/anaconda3/envs/delete_env/lib/python3.11/asyncio/tasks.py", line 349, in __wakeup
    future.result()
  File "/opt/anaconda3/envs/delete_env/lib/python3.11/asyncio/futures.py", line 198, in result
    raise exc
asyncio.exceptions.CancelledError: Cancelled by cancel scope 31e318690

During handling of the above exception, another exception occurred:

  + Exception Group Traceback (most recent call last):
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/uvicorn/protocols/http/httptools_impl.py", line 401, in run_asgi
  |     result = await app(  # type: ignore[func-returns-value]
  |              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in __call__
  |     return await self.app(scope, receive, send)
  |            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/fastapi/applications.py", line 1054, in __call__
  |     await super().__call__(scope, receive, send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/applications.py", line 123, in __call__
  |     await self.middleware_stack(scope, receive, send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/middleware/errors.py", line 186, in __call__
  |     raise exc
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/middleware/errors.py", line 164, in __call__
  |     await self.app(scope, receive, _send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/middleware/exceptions.py", line 65, in __call__
  |     await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
  |     raise exc
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
  |     await app(scope, receive, sender)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/routing.py", line 754, in __call__
  |     await self.middleware_stack(scope, receive, send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/routing.py", line 774, in app
  |     await route.handle(scope, receive, send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/routing.py", line 295, in handle
  |     await self.app(scope, receive, send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/routing.py", line 77, in app
  |     await wrap_app_handling_exceptions(app, request)(scope, receive, send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
  |     raise exc
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
  |     await app(scope, receive, sender)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/routing.py", line 75, in app
  |     await response(scope, receive, send)
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/responses.py", line 258, in __call__
  |     async with anyio.create_task_group() as task_group:
  |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/anyio/_backends/_asyncio.py", line 680, in __aexit__
  |     raise BaseExceptionGroup(
  | ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)
  +-+---------------- 1 ----------------
    | Traceback (most recent call last):
    |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/asyncio/tasks.py", line 277, in __step
    |     result = coro.send(None)
    |              ^^^^^^^^^^^^^^^
    |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/responses.py", line 261, in wrap
    |     await func()
    |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/starlette/responses.py", line 250, in stream_response
    |     async for chunk in self.body_iterator:
    |   File "/Applications/FastAPI-LangChain-Streaming-main/new_app.py", line 78, in generate_chat_responses
    |     with tru_recorder as recording:
    |   File "/opt/anaconda3/envs/delete_env/lib/python3.11/site-packages/trulens_eval/app.py", line 1063, in __exit__
    |     raise exc_value
    |   File "/Applications/FastAPI-LangChain-Streaming-main/new_app.py", line 80, in generate_chat_responses
    |     retrieval_chain.acall(
    |     ^^^^^^^^^^^^^^^^^^^^^
    | AttributeError: 'RunnableSequence' object has no attribute 'acall'
    +------------------------------------

In my experimental environment, I need to use internal api_key and base_url. Is there any way to solve this problem?

Copy link

dosubot bot commented Aug 19, 2024

To resolve the issues with using the internal api_key and base_url for OpenAI embedding and the AttributeError related to RunnableSequence not having an acall attribute in your FastAPI-LangChain-Streaming application, you can follow these steps:

  1. Ensure Environment Variables are Set Correctly:
    Make sure that the environment variables for the API key and endpoint are set correctly as shown in the provided code snippet.

    import os
    
    os.environ["AZURE_OPENAI_API_KEY"] = "your_api_key_here"
    os.environ["AZURE_OPENAI_ENDPOINT"] = "https://your_endpoint_here.openai.azure.com/"
    os.environ["OPENAI_API_VERSION"] = "2023-07-01-preview"
    os.environ["OPENAI_API_TYPE"] = "azure"
  2. Define the LLM and Embedding Model Correctly:
    Ensure that the LLM and embedding models are defined correctly using the environment variables.

    from langchain_openai import AzureChatOpenAI, AzureOpenAIEmbeddings
    
    llm = AzureChatOpenAI(
        model="gpt-35-turbo",
        deployment_name="your_deployment_name_here",
        api_key=os.environ["AZURE_OPENAI_API_KEY"],
        azure_endpoint=os.environ["AZURE_OPENAI_ENDPOINT"],
        api_version=os.environ["OPENAI_API_VERSION"],
    )
    
    embed_model = AzureOpenAIEmbeddings(
        azure_deployment="soc-text",
        api_key=os.environ["AZURE_OPENAI_API_KEY"],
        azure_endpoint=os.environ["AZURE_OPENAI_ENDPOINT"],
        api_version=os.environ["OPENAI_API_VERSION"],
    )
  3. Fix the AttributeError for RunnableSequence:
    The AttributeError related to RunnableSequence not having an acall attribute suggests that there might be an issue with the version of the langchain library or how the RunnableSequence is being used. Ensure you are using the correct version of the langchain library and that the RunnableSequence is used correctly.

    from langchain_core.runnables import RunnablePassthrough
    
    rag_chain = (
        {"context": retriever | format_docs, "question": RunnablePassthrough()}
        | prompt
        | llm
        | StrOutputParser()
    )
  4. Invoke the Chain Correctly:
    When invoking the chain, ensure that you are using the correct method and handling the response properly.

    query = "What is most interesting about this essay?"
    answer = rag_chain.invoke(query)
    
    print("query was:", query)
    print("answer was:", answer)

By following these steps, you should be able to resolve the issues with using the internal api_key and base_url for OpenAI embedding and the AttributeError related to RunnableSequence not having an acall attribute [1].

To continue talking to Dosu, mention @dosu.

@zongyicoding
Copy link
Author

Hi @sfc-gh-jreini,
After following dosu's suggestion, I am now able to retrieve data on the dashboard. However, the results still have some issues:

  1. The answers that were previously retrievable from the stream loop are no longer accessible.
  2. The dashboard is storing four records for a single question, but none of these records contain the actual answer. Below is the corrected code.
    note: my langchain version is 0.2.x, the acall method has been changed to ainvoke.
from fastapi import FastAPI, Request
from fastapi.responses import StreamingResponse, FileResponse
from langchain_openai import ChatOpenAI
from langchain.schema import Document
from langchain_openai import OpenAIEmbeddings
from langchain_community.vectorstores.chroma import Chroma
from langchain_core.runnables import RunnablePassthrough
from langchain_core.prompts import ChatPromptTemplate
from langchain_core.output_parsers import StrOutputParser

from trulens_eval import TruChain, Feedback
from trulens_eval.feedback.provider import OpenAI
from trulens_eval import Tru
from langchain.callbacks import AsyncIteratorCallbackHandler
import asyncio
import os 
from dotenv import load_dotenv
load_dotenv()

tru = Tru()
tru.reset_database()
tru.run_dashboard(port=9999, force=True)

embedding_function = OpenAIEmbeddings(
    model="text-embedding-ada-002",
    check_embedding_ctx_length=False,
    api_key=os.environ['API_KEY']
)

docs = [
    Document(
        page_content="the dog loves to eat pizza", metadata={"source": "animal.txt"}
    ),
    Document(
        page_content="the cat loves to eat lasagna", metadata={"source": "animal.txt"}
    ),
]

db = Chroma.from_documents(docs, embedding_function)
retriever = db.as_retriever()

template = """Answer the question based only on the following context:
{context}

Question: {question}
"""
prompt = ChatPromptTemplate.from_template(template)
model = ChatOpenAI(temperature=0, streaming=True, api_key=os.environ['API_KEY'],model='gpt-3.5')

provider = OpenAI(api_key=os.environ['API_KEY'])
retrieval_chain = (
    {"context": retriever, "question": RunnablePassthrough()}
    | prompt
    | model
    | StrOutputParser()
)

f_answer_relevance = (
    Feedback(provider.context_relevance_with_cot_reasons, name="Answer Relevance")
    .on_input_output()
)

tru_recorder = TruChain(
    retrieval_chain,
    app_id='conversation_stream',
    feedbacks=[f_answer_relevance],
    feedback_mode="deferred"
)

app = FastAPI()

async def generate_chat_responses(message):
    callback = AsyncIteratorCallbackHandler()
    
    with tru_recorder as recording:
        task = asyncio.create_task(
            retrieval_chain.ainvoke(

                input=dict(question=message),
                callbacks=[callback]
            )
        )
    
    response = ""
    async for token in callback.aiter():
        content = token.replace("\n", "<br>")
        print(f'content: {content}')
        response += content
        yield f"data: {content}\n\n"
    
    await task
    record = recording.get()
    print(f'Record: {record}')

@app.get("/")
async def root():
    return FileResponse("static/index.html")

@app.get("/chat_stream/{message}")
async def chat_stream(message: str):
    return StreamingResponse(generate_chat_responses(message=message), media_type="text/event-stream")

if __name__ == "__main__":
    import uvicorn
    uvicorn.run(app, host="127.0.0.1", port=8080)
WARNI [chromadb.segment.impl.vector.local_hnsw] Number of requested results 4 is greater than number of elements in index 2, updating n_results = 2
WARNI [trulens_eval.app] Unsure what the main input string is for the call to _aget_relevant_documents with args [{'question': 'what is 1+2 = ? explain it please !'}, <langchain_core.callbacks.manager.AsyncCallbackManagerForRetrieverRun object at 0x169c06610>].
WARNI [trulens_eval.app] Focus {'question': 'what is 1+2 = ? explain it please !'} is not a sequence.
WARNI [trulens_eval.app] Could not determine main input/output of [{'question': 'what is 1+2 = ? explain it please !'}, <langchain_core.callbacks.manager.AsyncCallbackManagerForRetrieverRun object at 0x169c06610>].
WARNI [trulens_eval.app] Could not determine main output from messages=[HumanMessage(content="Answer the question based only on the following context:\n[Document(metadata={'source': 'animal.txt'}, page_content='the dog loves to eat pizza'), Document(metadata={'source': 'animal.txt'}, page_content='the cat loves to eat lasagna')]\n\nQuestion: {'question': 'what is 1+2 = ? explain it please !'}\n")].

Please assist in addressing these issues. Thank you.

Copy link

dosubot bot commented Aug 19, 2024

To resolve the issues where the answers are no longer accessible from the stream loop and the Trulens dashboard is storing multiple records without the actual answer, you should ensure that you are using the correct version of the invoke method as supported by Langchain >=0.2.x.

The release notes for version 0.31.0 mention support for Langchain >=0.2.x retrievers via instrumenting the invoke method:

## 0.31.0
### What's Changed
* Support for langchain >=0.2.x retrievers via instrumenting the `invoke` method by @nicoloboschi in https://github.com/truera/trulens/pull/1187

Additionally, you can refer to the example implementation provided in the App_TruBot.py script. This example demonstrates how to set up a ConversationalRetrievalChain with Langchain and Trulens, including the correct usage of the invoke method and how to properly configure the Trulens feedback mechanisms.

Here is a relevant snippet from the example:

import os

os.environ['PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION'] = 'python'

from langchain.chains import ConversationalRetrievalChain
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.memory import ConversationSummaryBufferMemory
from langchain_community.callbacks import get_openai_callback
from langchain_community.llms import OpenAI
from langchain_community.vectorstores import Pinecone
import numpy as np
import pinecone
import streamlit as st

from trulens_eval import feedback
from trulens_eval import Select
from trulens_eval import tru
from trulens_eval import tru_chain_recorder
from trulens_eval.feedback import Feedback
from trulens_eval.keys import check_keys

check_keys("PINECONE_API_KEY", "PINECONE_ENV", "OPENAI_API_KEY")

# Set up GPT-3 model
model_name = "gpt-3.5-turbo"

app_id = "TruBot"

# Pinecone configuration.
pinecone.init(
    api_key=os.environ.get("PINECONE_API_KEY"),  # find at app.pinecone.io
    environment=os.environ.get("PINECONE_ENV")  # next to api key in console
)

identity = lambda h: h

hugs = feedback.Huggingface()
openai = feedback.OpenAI()

f_lang_match = Feedback(hugs.language_match).on(
    text1=Select.RecordInput, text2=Select.RecordOutput
)

f_qa_relevance = Feedback(openai.relevance).on(
    prompt=Select.RecordInput, response=Select.RecordOutput
)

f_qs_relevance = Feedback(openai.qs_relevance).on(
    question=Select.RecordInput,
    statement=Select.Record.chain.combine_docs_chain._call.args.inputs.
    input_documents[:].page_content
).aggregate(np.min)

def generate_response(prompt):
    # Embedding needed for Pinecone vector db.
    embedding = OpenAIEmbeddings(model='text-embedding-ada-002')  # 1536 dims
    docsearch = Pinecone.from_existing_index(
        index_name="llmdemo", embedding=embedding
    )
    retriever = docsearch.as_retriever()

    # LLM for completing prompts, and other tasks.
    llm = OpenAI(temperature=0, max_tokens=128)

    # Conversation memory.
    memory = ConversationSummaryBufferMemory(
        max_token_limit=650,
        llm=llm,
        memory_key="chat_history",
        output_key='answer'
    )

    # Conversational chain puts it all together.
    chain = ConversationalRetrievalChain.from_llm(
        llm=llm,
        retriever=retriever,
        return_source_documents=True,
        memory=memory,
        get_chat_history=identity,
        max_tokens_limit=4096
    )

    # Language mismatch fix:
    if "langprompt" in app_id:
        chain.combine_docs_chain.llm_chain.prompt.template = \
            "Use the following pieces of CONTEXT to answer the question at the end " \
            "in the same language as the question. If you don't know the answer, " \
            "just say that you don't know, don't try to make up an answer.\n" \
            "\n" \
            "CONTEXT: {context}\n" \
            "\n" \
            "Question: {question}\n" \
            "Helpful Answer: "

    elif "relevance" in app_id:
        # Contexts fix
        chain.combine_docs_chain.llm_chain.prompt.template = \
            "Use only the relevant contexts to answer the question at the end " \
            ". Some pieces of context may not be relevant. If you don't know the answer, " \
            "just say that you don't know, don't try to make up an answer.\n" \
            "\n" \
            "Contexts: \n" \
            "{context}\n" \
            "\n" \
            "Question: {question}\n" \
            "Helpful Answer: "

        # space is important

        chain.combine_docs_chain.document_prompt.template = "\tContext: {page_content}"

    # Trulens instrumentation.
    tc = tru_chain_recorder.TruChain(chain, app_id=app_id)

    return tc, tc.with_record(dict(question=prompt))

# Set up Streamlit app
st.title("TruBot")
user_input = st.text_input("Ask a question about TruEra")

if user_input:
    # Generate GPT-3 response
    prompt_input = user_input
    # add context manager to capture tokens and cost of the chain

    with get_openai_callback() as cb:
        chain, (response, record) = generate_response(prompt_input)
        total_tokens = cb.total_tokens
        total_cost = cb.total_cost

    answer = response['answer']

    # Display response
    st.write(answer)

    record_id = tru.add_data(
        app_id=app_id,
        prompt=prompt_input,
        response=answer,
        record=record,
        tags='dev',
        total_tokens=total_tokens,
        total_cost=total_cost
    )

    # Run feedback function and get value
    feedbacks = tru.run_feedback_functions(
        app=app,
        record=record,
        feedback_functions=[f_lang_match, f_qa_relevance, f_qs_relevance]
    )

    # Add value to database
    tru.add_feedback(record_id, feedbacks)

Ensure that your implementation aligns with the updated method signatures and functionalities provided in this version. If you are still encountering issues, it might be beneficial to review the specific changes made in the pull request mentioned above (#1187) to understand the exact modifications and how they should be integrated into your code.

Additionally, you can refer to the langchain_agents.ipynb example from the Trulens repository, which demonstrates how to set up a Langchain agent and use TruLens to identify gaps in tool coverage. This example includes setting up evaluation and feedback mechanisms, which might help you troubleshoot and resolve the issues you are facing:

from trulens_eval import Feedback
from trulens_eval import Tru
from trulens_eval import TruChain
from trulens_eval.feedback import OpenAI as fOpenAI

tru = Tru()

from datetime import datetime
from datetime import timedelta
from typing import Type

from langchain import SerpAPIWrapper
from langchain.agents import AgentType
from langchain.agents import initialize_agent
from langchain.agents import Tool
from langchain.chat_models import ChatOpenAI
from langchain.tools import BaseTool
import openai
from pydantic import BaseModel
from pydantic import Field
import yfinance as yf

# Setup API keys
import os
os.environ["OPENAI_API_KEY"] = "sk-..."
os.environ["SERPAPI_API_KEY"] = "..."

# Create agent with search tool
search = SerpAPIWrapper()
search_tool = Tool(
    name="Search",
    func=search.run,
    description="useful for when you need to answer questions about current events"
)

llm = ChatOpenAI(model="gpt-3.5-turbo", temperature=0)
tools = [search_tool]

agent = initialize_agent(
    tools, llm,
    agent=AgentType.OPENAI_FUNCTIONS,
    verbose=True
)

# Set up Evaluation
class OpenAI_custom(fOpenAI):
    def no_answer_feedback(self, question: str, response: str) -> float:
        return float(self.endpoint.client.chat.completions.create(
            model="gpt-3.5-turbo",
            messages=[
                {"role": "system", "content": "Does the RESPONSE provide an answer to the QUESTION? Rate on a scale of 1 to 10. Respond with the number only."},
                {"role": "user", "content": f"QUESTION: {question}; RESPONSE: {response}"}
            ]
        ).choices[0].message.content) / 10

custom = OpenAI_custom()
f_no_answer = Feedback(custom.no_answer_feedback).on_input_output()

tru_agent = TruChain(
    agent,
    app_id="Search_Agent",
    feedbacks=[f_no_answer]
)

prompts = [
    "What company acquired MosaicML?",
    "What's the best way to travel from NYC to LA?",
    "How did the change in the exchange rate during 2021 affect the stock price of US based companies?",
    "Compare the stock performance of Google and Microsoft",
    "What is the highest market cap airline that flies from Los Angeles to New York City?",
    "I'm interested in buying a new smartphone from the producer with the highest stock price. Which company produces the smartphone I should by and what is their current stock price?"
]

with tru_agent as recording:
    for prompt in prompts:
        agent(prompt)

# Explore in a Dashboard
tru.run_dashboard() # open a local streamlit app to explore
# tru.stop_dashboard() # stop if needed

This example demonstrates how to set up a Langchain agent, add custom tools, and use TruLens for evaluation and feedback. It might provide additional insights into resolving your issues with the ainvoke method and the Trulens dashboard [1][2][3][4][5].

This is a tough nut to crack, could really use your help on this, @joshreini1.

To continue talking to Dosu, mention @dosu.

@dosubot dosubot bot added the stale Issue has not had recent activity or appears to be solved. Stale issues will be automatically closed label Sep 19, 2024
@sfc-gh-jreini
Copy link
Contributor

We have a fixes coming for async and streaming in this PR.

Feel free to install it from the branch if you'd like to try it now, otherwise we'll have it in the next release.

@sfc-gh-jreini sfc-gh-jreini linked a pull request Sep 20, 2024 that will close this issue
6 tasks
@dosubot dosubot bot removed the stale Issue has not had recent activity or appears to be solved. Stale issues will be automatically closed label Sep 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants