Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
davanstrien 
posted an update Aug 7, 2024
Post
3155
Is your summer reading list still empty? Curious if an LLM can generate a book blurb you'd enjoy and help build a KTO preference dataset at the same time?

A demo using Hugging Face Spaces and Gradio to collect LLM output preferences: davanstrien/would-you-read-it

Very interesting. As you aptly pointed out, creative writing is one of the tasks the community is currently focusing on, and I believe we can do a significantly better job than Anthropic or ClosedAI.

Making an LLM for creative writing might seem like a trivial task, but it is not. Can we create datasets or LLMs that write based on a prompt? Sure. But would the output be any good? Not quite. That's the challenging part.

I have started an ambitious project, LLAMA-3_8B_Unaligned, and I am more than 1,000 work hours into it...

My approach is quite different though. Instead of using KTO\DPO etc... I "just" want the LLM to be able to follow writing instructions extremely well while completely altering the token distribution probability so that it doesn't resemble machine-written text (aka "SLOP").

I would love to follow up on your project, and I think the holy grail would be to see about 5-10 short books that were at least 95% written by AI.

Fun fact: There's only a tiny bunch of LLMs that can split a LONG text into paragraphs correctly with a low probability of error. That's a problem I didn't expect to encounter, like many others. As I said, this "trivial" task is way harder than it seems! Please keep us updated.