An experimental fine-tune of Yi 34b 200k using bagel. This is the version of the fine-tune before direct preference optimization (DPO) has been applied. DPO performs better on benchmarks, but this version is likely better for creative writing, roleplay, etc.

Model Information

Model ID

jondurbin/bagel-34b

Context Length

200,000 tokens

Author

jondurbin

Capabilities