Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add_thinking #5220

Open
wants to merge 3 commits into
base: main
Choose a base branch
from
Open

add_thinking #5220

wants to merge 3 commits into from

Conversation

Tomlili43
Copy link

End-user friendly description of the problem this fixes or functionality that this introduces
Agent thinking feature by thinking

  • Include this change in the Release Notes. If checked, you must provide an end-user friendly description for your change below

Give a summary of what the PR does, explaining any non-trivial design decisions
add thinking prompt and thinking front end UI.

image
image


Link of any specific issues this addresses

Copy link
Contributor

@neubig neubig left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi @Tomlili43 , thanks for the contribution!

In general, all prompts should be implemented in the Python backend, not the frontend. And if we add a new prompt we'll need to validate its effect on agent accuracy and cost.

If you're interested in adding this prompt, could you try adding it in the agenthub/codeact_agent directory? (you can look at the directory structure there to understand what needs to be done)

We would then need to validate the effect of this change on performance and cost. I'm a bit worried that adding this big prompt in the input and encouraging longer outputs would lead to significant increases in time/token cost, but we could try a few examples to see if this makes a big difference in accuracy!

@Tomlili43
Copy link
Author

@neubig yeap,

  1. added it in the agenthub/codeact_agent directory.
  2. for validation, it will cost so many tokens. To be honest , i couldn't afford it. Can you help with it ?

@neubig neubig added the run-eval-s Runs evaluation with 5 instances label Nov 23, 2024
@neubig
Copy link
Contributor

neubig commented Nov 23, 2024

Hey @Tomlili43 , we can try to do this, but it'd be good if you could try running a few examples first to see if it's working!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
run-eval-s Runs evaluation with 5 instances
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants