PREMIUM FEATURE
Advanced Search Filters
Filter search results by source, date, and more with our premium search tools.
" "I don't know what labs are doing to these poor LLMs during RL but they are mortally terrified of exceptions, in any infinitesimally likely case. Exceptions are a normal part of life and healthy dev process. Sign my LLM welfare petition for improved rewards in cases of exceptions.
Filter search results by source, date, and more with our premium search tools.
Related quotes. More quotes will automatically load as you scroll down, or you can use the load more buttons.
It's 2025 and most content is still written for humans instead of LLMs. 99.9% of attention is about to be LLM attention, not human attention.
E.g. 99% of libraries still have docs that basically render to some pretty .html static pages assuming a human will click through them. In 2025 the docs should be a single your_project.md text file that is intended to go into the context window of an LLM.
Repeat for everything.
The future expands the variance of human condition a lot more than it drags its mean. This is an empirical observation with interesting extrapolations.
The past is well-approximated as a population of farmers, living similar lives w.r.t. upbringing, knowledge, activities, ideals, aspirations, etc.
The future trends to include all of:
- the transhumanists who "ascend" with neuralinks etc., and the Amish living ~19th century life.
- those who "worship" ideals of religion, technology, knowledge, wealth, fitness, community, nature, art, ...
- those exploring externally into the stars, those exploring internally into minds (drugs++), or those who disappear into digital VR worlds
- those who date a different partner every day and those who are monogamous for life
- those who travel broadly and those who stay in one location their entire life
- those in megacities and those off-the-grid
For almost any question about a dimension of human condition, the answer trends not to any specific thing but to "all of the above". And to an extreme diversity of memetics. At least, this feels like the outcome in free societies that trend to abundance. I don't know what it feels like to live in such a society but it's interesting to think about.
Don't think of LLMs as entities but as simulators. For example, when exploring a topic, don't ask:
"What do you think about xyz"?
There is no "you". Next time try:
"What would be a good group of people to explore xyz? What would they say?"
The LLM can channel/simulate many perspectives but it hasn't "thought about" xyz for a while and over time and formed its own opinions in the way we're used to. If you force it via the use of "you", it will give you something by adopting a personality embedding vector implied by the statistics of its finetuning data and then simulate that. It's fine to do, but there is a lot less mystique to it than I find people naively attribute to "asking an AI".