LESSWRONG
LW

Martin Vlach
7151141
Message
Dialogue
Subscribe

If you get an email from aisafetyresearch@gmail.com , that is most likely me. I also read it weekly, so you can pass a message into my mind that way.
Other ~personal contacts: https://linktr.ee/uhuge 

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
1Martin Vlach's Shortform
3y
35
Project Vend: Can Claude run a small shop?
Martin Vlach1mo10

It's provided the current time together with other 20k sys-prompt tokens, so substantially more diluted influence on the behaviours..?

Reply
So You Think You've Awoken ChatGPT
Martin Vlach2mo10

Folks like this guy hit it on hyperspeed - 

https://www.facebook.com/reel/1130046385837121/?mibextid=rS40aB7S9Ucbxw6v

 

I still remember university teacher explaining how early TV transmission were very often including/displaying ghosts of dead people, especially dead relatives.

As the tech matures from art these phenomena or hallucinations evaporate.

Reply
Energy-Based Transformers are Scalable Learners and Thinkers
Martin Vlach2mo40

 you seem to report one OOM less than this picture in https://alexiglad.github.io/blog/2025/ebt/#:~:text=a%20log%20function).-,Figure%208,-%3A%20Scaling%20for

Reply
Open Thread - Summer 2025
Martin Vlach2mo10

Link to Induction section on https://www.lesswrong.com/lw/dhg/an_intuitive_explanation_of_solomonoff_induction/#induction seems broken on mobile Chrome, @habryka 

Reply
Open Thread - Summer 2025
Martin Vlach2mo20

I've heard that hypothesis in a review of that blog post of Anthropic, likely by 

AI Explained

maybe by

bycloud

.

 

They've called it "Chekov's gun".

Reply
Open Thread - Summer 2025
Martin Vlach2mo10

What's your view on sceptic claims about RL on transformer LMs like https://arxiv.org/abs/2504.13837v2 or one that CoT instruction yields better results than <thinking> training?

Reply
Open Source Search (Summary)
Martin Vlach2mo10

Not the content I expect labeled AIb Capabilities,

although I see how that'd be vindicated. 

 

By the way, if I write an article about LMs generating SVG, that's a plaintext and if I put an SVG illustration up, that's an image, not a plaintext?

Reply1
Martin Vlach's Shortform
Martin Vlach2mo10

Trivial, but do token-based LMs follow instructions like "only output tokens '1', '2', '3'" where they'd output 123 as one token without that instruction?

Reply1
Martin Vlach's Shortform
Martin Vlach3mo10

I'd update my take from a very pessimist/gloom one to an (additional) excited one: Those more intelligent models building a clear view of the person they/it interacts with is a sign of emerging empathy, which is a hopeful property for alignment/respect.

Reply
Vincent Li's Shortform
Martin Vlach3mo20

False Trichotomy?

Your model assumes that one cannot be all three, however, some roles demand it, and in reality people do navigate all three traits, my top example would be empathic project managers.

Reply
Load More
Zombies
1y
(+52/-50)
4Draft: A concise theory of agentic consciousness
3mo
2
0Thou shalt not command an alighned AI
4mo
4
8G.D. as Capitalist Evolution, and the claim for humanity's (temporary) upper hand
4mo
3
6Would it be useful to collect the contexts, where various LLMs think the same?
Q
2y
Q
1
1Martin Vlach's Shortform
3y
35