Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can't get the same result as the example in the Doc about Jaeger #6074

Closed
HiFiChang opened this issue Mar 23, 2025 · 7 comments
Closed

Can't get the same result as the example in the Doc about Jaeger #6074

HiFiChang opened this issue Mar 23, 2025 · 7 comments

Comments

@HiFiChang
Copy link

What happened?

Describe the bug
I'm trying to use Jaeger to record the agent, but after I follow the guide and run the official example, I can't get the same result as is shown in the doc. The example agent is not traced successfully.

To Reproduce
Run Jaeger

docker run -d --name jaeger \
  -e COLLECTOR_OTLP_ENABLED=true \
  -p 16686:16686 \
  -p 4317:4317 \
  -p 4318:4318 \
  jaegertracing/all-in-one:latest

just the example in the doc

from opentelemetry import trace
from opentelemetry.exporter.otlp.proto.grpc.trace_exporter import OTLPSpanExporter
from opentelemetry.sdk.resources import Resource
from opentelemetry.sdk.trace import TracerProvider
from opentelemetry.sdk.trace.export import BatchSpanProcessor

otel_exporter = OTLPSpanExporter(endpoint="http://localhost:4317", insecure=True)
tracer_provider = TracerProvider(resource=Resource({"service.name": "autogen-test-agentchat"}))
span_processor = BatchSpanProcessor(otel_exporter)
tracer_provider.add_span_processor(span_processor)
trace.set_tracer_provider(tracer_provider)

# we will get reference this tracer later using its service name
tracer = trace.get_tracer("autogen-test-agentchat")

from autogen_agentchat.agents import AssistantAgent
from autogen_agentchat.conditions import MaxMessageTermination, TextMentionTermination
from autogen_agentchat.teams import SelectorGroupChat
from autogen_agentchat.ui import Console
from autogen_core import SingleThreadedAgentRuntime
from autogen_ext.models.openai import OpenAIChatCompletionClient


def search_web_tool(query: str) -> str:
    if "2006-2007" in query:
        return """Here are the total points scored by Miami Heat players in the 2006-2007 season:
        Udonis Haslem: 844 points
        Dwayne Wade: 1397 points
        James Posey: 550 points
        ...
        """
    elif "2007-2008" in query:
        return "The number of total rebounds for Dwayne Wade in the Miami Heat season 2007-2008 is 214."
    elif "2008-2009" in query:
        return "The number of total rebounds for Dwayne Wade in the Miami Heat season 2008-2009 is 398."
    return "No data found."


def percentage_change_tool(start: float, end: float) -> float:
    return ((end - start) / start) * 100


async def main() -> None:
    model_client = OpenAIChatCompletionClient(model="gpt-4o-mini")

    planning_agent = AssistantAgent(
        "PlanningAgent",
        description="An agent for planning tasks, this agent should be the first to engage when given a new task.",
        model_client=model_client,
        system_message="""
        You are a planning agent.
        Your job is to break down complex tasks into smaller, manageable subtasks.
        Your team members are:
            WebSearchAgent: Searches for information
            DataAnalystAgent: Performs calculations

        You only plan and delegate tasks - you do not execute them yourself.

        When assigning tasks, use this format:
        1. <agent> : <task>

        After all tasks are complete, summarize the findings and end with "TERMINATE".
        """,
    )

    web_search_agent = AssistantAgent(
        "WebSearchAgent",
        description="An agent for searching information on the web.",
        tools=[search_web_tool],
        model_client=model_client,
        system_message="""
        You are a web search agent.
        Your only tool is search_tool - use it to find information.
        You make only one search call at a time.
        Once you have the results, you never do calculations based on them.
        """,
    )

    data_analyst_agent = AssistantAgent(
        "DataAnalystAgent",
        description="An agent for performing calculations.",
        model_client=model_client,
        tools=[percentage_change_tool],
        system_message="""
        You are a data analyst.
        Given the tasks you have been assigned, you should analyze the data and provide results using the tools provided.
        If you have not seen the data, ask for it.
        """,
    )

    text_mention_termination = TextMentionTermination("TERMINATE")
    max_messages_termination = MaxMessageTermination(max_messages=25)
    termination = text_mention_termination | max_messages_termination

    selector_prompt = """Select an agent to perform task.

    {roles}

    Current conversation context:
    {history}

    Read the above conversation, then select an agent from {participants} to perform the next task.
    Make sure the planner agent has assigned tasks before other agents start working.
    Only select one agent.
    """

    task = "Who was the Miami Heat player with the highest points in the 2006-2007 season, and what was the percentage change in his total rebounds between the 2007-2008 and 2008-2009 seasons?"

    tracer = trace.get_tracer("autogen-test-agentchat")
    with tracer.start_as_current_span("runtime"):
        team = SelectorGroupChat(
            [planning_agent, web_search_agent, data_analyst_agent],
            model_client=model_client,
            termination_condition=termination,
            selector_prompt=selector_prompt,
            allow_repeated_speaker=True,
        )
        await Console(team.run_stream(task=task))

    # await model_client.close()


if __name__ == "__main__":
    import asyncio
    asyncio.run(main())

Expected behavior
just like the screenshot in the doc

Image

Screenshots

Image

Image

Which packages was the bug in?

Python AgentChat (autogen-agentchat>=0.4.0)

AutoGen library version.

Python dev (main branch)

Other library version.

No response

Model used

gpt-4o-mini

Model provider

None

Other model provider

No response

Python version

3.13

.NET version

None

Operating system

Windows

@victordibia
Copy link
Collaborator

Are you able to click on the runtime trace (the single it line item in your screenshot?)

@HiFiChang
Copy link
Author

Are you able to click on the runtime trace (the single it line item in your screenshot?)您是否可以点击运行时跟踪(您截图中的单行项目?)

Image

@victordibia
Copy link
Collaborator

victordibia commented Mar 23, 2025

For some reason, it seems the runtime events are not being collected in your case.
Can you runt he second example on custom traces? Do you still have the same issue with only one span?

Also, can you confirm you have run

pip install opentelemetry-sdk opentelemetry-exporter-otlp-proto-grpc

The span provider we use is based on grpc opentelemetry.exporter.otlp.proto.grpc.trace_exporter import OTLPSpanExporter hence the need for opentelemetry-exporter-otlp-proto-grpc (looks like this is what is missing in the docs)

@victordibia victordibia added code-execution execute generated code proj-core and removed code-execution execute generated code needs-triage labels Mar 23, 2025
@HiFiChang
Copy link
Author

The second example seems to be OK? why
Image
Yes i have downloaded opentelemetry-sdk opentelemetry-exporter-otlp-proto-grpc

@victordibia
Copy link
Collaborator

victordibia commented Mar 23, 2025

Ok .. so I just tried reproducing the entire issue on a new machine (code spaces).
It works as expected.

Can you do the following:

  • Restart the notebook kernel and rerun from scratch

The second example seems to be OK? why

The one thing I can think of is that the grpc otel exporter libs are not loaded properly on your machine (this is what the runtime uses). A kernel restart can help here.

Image

@HiFiChang
Copy link
Author

HiFiChang commented Mar 23, 2025

Restart the notebook kernel and rerun from scratch

Thank you for your useful suggestion @victordibia When I switch to wsl, it seems to run well... amazing, but why
Image

Can you explain why the same program in wsl is OK but in windows is still no...

Image

@victordibia
Copy link
Collaborator

@HiFiChang ,

Thats interesting ...
I am glad you found a fix. I am not sure why there are differences on windows vs wsl.
May be related to timestamps etc - jaegertracing/jaeger#2286

It also might just be that you needed to restart your jupyter kernel (or script)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants