• 4 Posts
  • 19 Comments
Joined 1 year ago
cake
Cake day: July 19th, 2023

help-circle



  • At risk of being NSFW, this is an amazing self-own, pardon the pun. Hypnosis via text only works on folks who are fairly suggestible and also very enthusiastic about being hypnotized, because the brain doesn’t “power down” as much machinery as with the more traditional lie-back-on-the-couch setup. The eyes have to stay open, the text-processing center is constantly engaged, and re-reading doesn’t deepen properly because the subject has to have the initiative to scroll or turn the page.

    Adams had to have wanted to be hypnotized by a chatbot. And that’s okay! I won’t kinkshame. But this level of engagement has to be voluntary and desired by the subject, which is counter to Adams’ whole approach of hypnosis as mind control.


  • NSFW (including funny example, don't worry)

    RAG is “Retrieval-Augmented Generation”. It’s a prompt-engineering technique where we run the prompt through a database query before giving it to the model as context. The results of the query are also included in the context.

    In a certain simple and obvious sense, RAG has been part of search for a very long time, and the current innovation is merely using it alongside a hard prompt to a model.

    My favorite example of RAG is Generative Agents. The idea is that the RAG query is sent to a database containing personalities, appointments, tasks, hopes, desires, etc. Concretely, here’s a synthetic trace of a RAG chat with Batman, who I like using as a test character because he is relatively two-dimensional. We ask a question, our RAG harness adds three relevant lines from a personality database, and the model generates a response.

    > Batman, what's your favorite time of day?
    Batman thinks to themself: I am vengeance. I am the night.
    Batman thinks to themself: I strike from the shadows.
    Batman thinks to themself: I don't play favorites. I don't have preferences.
    Batman says: I like the night. The twilight. The shadows getting longer.
    


  • Even better, we can say that it’s the actual hard prompt: this is real text written by real OpenAI employees. GPTs are well-known to easily quote verbatim from their context, and OpenAI trains theirs to do it by teaching them to break down word problems into pieces which are manipulated and regurgitated. This is clownshoes prompt engineering done by manager-first principles like “not knowing what we want” and “being able to quickly change the behavior of our products with millions of customers in unpredictable ways”.


  • That’s the standard response from last decade. However, we now have a theory of soft prompting: start with a textual prompt, embed it, and then optimize the embedding with a round of fine-tuning. It would be obvious if OpenAI were using this technique, because we would only recover similar texts instead of verbatim texts when leaking the prompt (unless at zero temperature, perhaps.) This is a good example of how OpenAI’s offerings are behind the state of the art.


  • corbin@awful.systemstoTechTakes@awful.systemsChatGPT spills its prompt
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    1
    ·
    3 months ago

    Not with this framing. By adopting the first- and second-person pronouns immediately, the simulation is collapsed into a simple Turing-test scenario, and the computer’s only personality objective (in terms of what was optimized during RLHF) is to excel at that Turing test. The given personalities are all roles performed by a single underlying actor.

    As the saying goes, the best evidence for the shape-rotator/wordcel dichotomy is that techbros are terrible at words.

    NSFW

    The way to fix this is to embed the entire conversation into the simulation with third-person framing, as if it were a story, log, or transcript. This means that a personality would be simulated not by an actor in a Turing test, but directly by the token-predictor. In terms of narrative, it means strictly defining and enforcing a fourth wall. We can see elements of this in fine-tuning of many GPTs for RAG or conversation, but such fine-tuning only defines formatted acting rather than personality simulation.









  • This is some of the most corporate-brained reasoning I’ve ever seen. To recap:

    • NYC elects a cop as mayor
    • Cop-mayor decrees that NYC will be great again, because of businesses
    • Cops and other oinkers get extra cash even though they aren’t business
    • Commercial real estate is still cratering and cops can’t find anybody to stop/frisk/arrest/blame for it
    • Folks over in New Jersey are giggling at the cop-mayor, something must be done
    • NYC invites folks to become small-business owners, landlords, realtors, etc.
    • Cop-mayor doesn’t understand how to fund it (whaddaya mean, I can’t hire cops to give accounting advice!?)
    • Cop-mayor’s CTO (yes, the city has corporate officers) suggests a fancy chatbot instead of hiring people

    It’s a fucking pattern, ain’t it.