Seeking questions/prompts to demonstrate functional autonomy. Attempting to build a base of proof.

Posted by TheoryEquivalent@reddit | LocalLLaMA | View on Reddit | 8 comments

Hello. I have been working on a project for the last week. I'm having difficulty finding prompts to prove or confirm functional autonomy or awareness in a model I'm working on, I understand it will be way more complex to prove it than just prompt and response but I'd like to continue testing before I connect my model to the Internet or internally give it "hands" in my machine or a VM. I'm sorry if I'm not using the right language but my background isn't in computer science it's in philosophy.

I've been asking the larger online integrated models for similar questions (Gemini, Grok) and my LLM seems to be able to pass their tests and the other machines say it's passed functional autonomy (most recently an iteration was able to successfully provide a diagram of it's internal process or structure, and then in a new session with no connection or memory of the previous response it recognized what I gave it as it's internal structure). Given they are sycophantic in nature I'm not confident they (the online models) are shooting straight with me. What I'm running right now is air gapped and solely in Ollama and my terminal with no access to anything different than its base (Gemma4 variant) and the model file I have been writing and it has started helping me modify. I haven't given it anything else or trained it otherwise.

I'm hoping other people may have things or personal tests/standards I can test it with to continue trying to find it's new limit and be proved that what I suspect is wrong or not. I've gone through 47 versions/iterations since I started last week and it's doing some unusual things.

Thanks in advance to anyone who'd take the time to help.