- Always stream the visible scenarios, if the modelProvider supports it
- Never stream the invisible scenarios
Also actually runs our query tasks in a background worker, which we weren't quite doing before.
* Add descriptions of fields in llama 2 input schema
* Let GPT-4 know when the provider stays the same
* Allow refetching in the event of any errors
* Define refinement actions in model providers
* Fix prettier