Replies: 1 comment
-
Don't know if you're still interested in this. But i've been doing something similar with focus on customizing an LLM to 1 specific grammar.
I do think this can only go so far due to architectural constraints - transformers aren't great at learning recursive type structures or parenthesis matching even, which is a very simple context free grammar. Something like a mamba is possibly more suited. But I hypothesize even a transformer would get considerably better. Also, sorry if this was more of a research program than a light weight experiment. Item (1) above is probably good enough, if you want to "warm up" your target LLM to work with your grammar of preference. |
Beta Was this translation helpful? Give feedback.
-
I am just wondering what sort of features would be interesting if we wanted to train a specific LLM optimized for outlines ?
I am guessing maybe causality in most LMs is the biggest issue, maybe training with a Fill-In-The-Middle objective may help performance for templated generation ?
Maybe there are other training objectives that may help models generating under the Outlines framework ?
Wondering if anyone has insights on that, I'd be down to experiment a bit
Beta Was this translation helpful? Give feedback.
All reactions