> One can envision a world in which OpenAI pays chefs money to cook while ChatGPT watches—narrating their thought process, tasting the dishes, and describing the results. This information could be used for general-purpose training, but it might also be packaged as a “book”, “course”, or “partner” someone could ask for.
So we're speed running the idea of AI Facebook friends and creating a new para(ai)social relationship
because it's a separate context window, it makes the model bigger, that space is not accessible to the "user".
And the "language understanding" basically had to be done twice because it's a separate input to the transformer so you can't just toss a pile of text in there and say "figure it out".
so we are currently in the era of one giant context window.
We pay people to create more high quality tokens (mercor, turing) which are then fed into data generating processes (synthetic data) to create even more tokens to train on
But does that really help, or do you get distortion? The frequency distribution of human generated content moves slowly over time as new subjects are discussed. What frequency distribution do those “data generating processes” use? And at root, aren’t those “data generating processes” basically just another LLM (I.e., generating tokens according to a probability distribution)? Thus, aren’t we just sort of feeding AI slop into the next training run and humoring ourselves by renaming the slop as “synthetic data?” Not trying to be argumentative. I’m far from being an AI expert, so maybe I’m missing it. Feel free to explain why I’m wrong.
That's the problem in a nutshell. There is an art to how you generate the synthdata so that you don't get crappy trained models (especially when mistakes cost XX million dollars).
It's also theoretically why facebook paid 14bn for alex wang and scale ai
> They've obviously had a go at being a first-party model company to address this, but that didn't work.
I thought there was an entire initiative to build their own coding model and the fine tunes of in Composer 1.5 and Composer 2 were just buying them time and training data
So we're speed running the idea of AI Facebook friends and creating a new para(ai)social relationship
reply