Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature Request: Support system-style questions (transferred from Taxonomy) #61

Open
roy651 opened this issue May 22, 2024 · 4 comments
Labels

Comments

@roy651
Copy link

roy651 commented May 22, 2024

(Issue was moved from here as it touches multiple areas)
When defining a new skill in the pattern of question-context-answer it seems that ilab will generate all of those when it creates the augmented dataset. The implication is that they are all dynamically generated.

I would like to propose that ilab will support an additional entry in the skill's yaml allowing for a "system"-style which will preface the question when generating the answer. This part should be static and need not be generated for each new entry. Most likely, it should be static for the entire skill, hence it should reside in the top level of the yaml with the "task_description".

I think that this capability is something that may be required by additional skills as it follows, the now so common, system/user pattern and I assume others would require that too.

@roy651
Copy link
Author

roy651 commented May 22, 2024

@obuzek per your comment on the other thread:
I understand the mechanism you describe and I acknowledge that you're utilizing the "system" prompt as part of the SDG.
However, I think that, this design choice is something that should be relatively easy to resolve with a suitable instruction to the SDG ("Do not generate this part"...).
Assuming technically it's feasible and the discussion is about the need - One might argue that the set of Q&A (and context) samples should suffice for the model to learn the necessary "rules" and pattern of the specific skill.
I do think that this need is valid and quite widespread. As the use of LLMs expands and it not necessarily plain conversation-like, and as the complexity of input and output patterns rises, we see frequent deviations of the model from the required patterns.
I expect to see others come up with a similar request.

@bjhargrave
Copy link
Contributor

yaml with the "task_description".

The task_description is already placed in the prompt. So I think it can do what you are requesting. It can include instructions to the teacher model.

@roy651
Copy link
Author

roy651 commented May 22, 2024

@bjhargrave
This is not well documented but assuming that the prompt is the one that shows up on diagram 3 in the paper, then I'm not sure it's relevant, cause the task description is part of the first line and not directly related to the generated question.
Furthermore, I imagine that that both the model that's generating the responses and the evaluator model should get this static text as part of their prompt in a relevant slot.

Copy link

This issue has been automatically marked as stale because it has not had activity within 90 days. It will be automatically closed if no further activity occurs within 30 days.

@github-actions github-actions bot added the stale label Aug 28, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants