can i talk to an openclaw bot using internet relay chat?
-
like i need you to understand, i haven't even gotten through *setup* because the model apparently does not know how to use tools correctly. admittedly it has less than 1 billion parameters, and i don't know what the hell i am doing, but still.
can we get to the part where it is AI winter again already? this is not even fun. i want to throw my computer and its' very expensive RTX 6000 Blackwell GPU out my window.
-
can we get to the part where it is AI winter again already? this is not even fun. i want to throw my computer and its' very expensive RTX 6000 Blackwell GPU out my window.
i just wanted to put an openclaw on irc as a fucking shitpost man
-
i just wanted to put an openclaw on irc as a fucking shitpost man
and you tell me people legitimately are using this software.
how?
is it really magically better when you hook up claude?
-
and you tell me people legitimately are using this software.
how?
is it really magically better when you hook up claude?
(don't worry, i am running this in a MicroVM under kubernetes, I wouldn't dare give it access to anything I care about.)
-
so i installed it into the openclaw meme thing. and it's not like, doing the stuff it claims it is doing.
like it is hallucinating things like "i updated SOUL.md with xyz"
i seriously do not think this stuff is real now
@ariadne what model did you finetune on? For a 1B model you need something really specialized on tool calling.
-
@ariadne what model did you finetune on? For a 1B model you need something really specialized on tool calling.
@jfkimmes i built an LLM from scratch with transformers kinda loosely following the scripts the qwen people released
the LLM is basically trained on ~30ish GB of mostly furry smut and public Linux IRC logs.
*nods sagely*
-
@jfkimmes i built an LLM from scratch with transformers kinda loosely following the scripts the qwen people released
the LLM is basically trained on ~30ish GB of mostly furry smut and public Linux IRC logs.
*nods sagely*
-
@jfkimmes i built an LLM from scratch with transformers kinda loosely following the scripts the qwen people released
the LLM is basically trained on ~30ish GB of mostly furry smut and public Linux IRC logs.
*nods sagely*
@jfkimmes i am, however, using the 35b parameter qwen3.5 reasoning model for the "thinking" portion of this exercise
-
-
(don't worry, i am running this in a MicroVM under kubernetes, I wouldn't dare give it access to anything I care about.)
i wonder if the problem is that the model i trained is too shit to do anything other than really bad ERP
-
@ska @ariadne Ive thouht of something related, not chatbots but imagine a GPS driving assistant voice in your car giving you directions and feedback, but in the most toxic way possible. An angry swearing voice saying things like: "Your exit comes in half a mile, try not to miss that one, you fucking moron." I've thought that ought to be a funny option to toggle on once in a while.
-
and you tell me people legitimately are using this software.
how?
is it really magically better when you hook up claude?
The key is to realise that the average is so low – we can't all be experts at everything, so we are bad at most things – that a model performing slightly above average at one of the tasks we aren't good at means a majority of users will perceive its outcomes as positively better than what they could do themselves.
To any expert, the model falls very short, as it performs well below its own ability.
-
@jfkimmes i am, however, using the 35b parameter qwen3.5 reasoning model for the "thinking" portion of this exercise
@ariadne Oh, is that a OpenClaw specific feature where you can specify that reasoning traces are generated by a separate model than the actual response? I'm not really familiar with OpenClaw's internals.
-
@jfkimmes i am, however, using the 35b parameter qwen3.5 reasoning model for the "thinking" portion of this exercise
@ariadne In any case: as long as the final response is generated by your trained model it will never make a valid tool call since there are probably about zero training examples of the necessary JSON structure required by the tool handling in your furry smut (this is an estimate that could be quite the way off knowing the furry community but still)
-
@ariadne Oh, is that a OpenClaw specific feature where you can specify that reasoning traces are generated by a separate model than the actual response? I'm not really familiar with OpenClaw's internals.
@jfkimmes yes, you can have it use a different model for planning.
-
@ariadne In any case: as long as the final response is generated by your trained model it will never make a valid tool call since there are probably about zero training examples of the necessary JSON structure required by the tool handling in your furry smut (this is an estimate that could be quite the way off knowing the furry community but still)
@jfkimmes this does explain something: it seems to be able to invoke tools when it is planning, but then those tools do not get invoked in the final step.
so it uses tools to read files when planning, then fails to use tools when executing.
what a fascinating conundrum.