Submitted by Lesterpaintstheworld t3_11attqx in singularity
Context
I'm building my own ACE (Autonomous Cognitive Entity), and I'm already at a pretty interesting stage with good results & some emerging behaviors. Original post:https://www.reddit.com/r/singularity/comments/113p2jn/the_road_to_agi_building_homebrew_autonomous/
Progress
- Audio: Josh can now hear what I'm saying. It increased the amount of input he his getting ~10x from chat: Voice is the way to go. If I could plug my thread of thought continuously as an input, I would. In the meantime, I'm mumbling all day long in my headset while coding. Good enough ^^
- Code awareness: I started a micro-process feeding Josh the code of his microservices & letting Josh read and make sense of it. It shows encouraging results, as he is capable of showing some understanding what each one does. I'm currently limited by the 4K context size of davinci, but this should change soon (GPT-4 rumored to have up to 32k context window). End goal being the creation of synthetic code, with Josh creating new microservices by himself.
- Critic: One of the hardest parts IMO: reducing confabulation & assessing whether a thought is a good one relative to context. Still working on this part.
- Actor: Josh is now capable of working and reworking on a single piece of text to improve it gradually. This is useful for email crafting for example. I'll be thinking about new ways to "move" ie. act in the world.
- Short-term memory: I gave Josh short-term memory (currently 12 most important thoughts of the session). I'm experimenting to see how to best inject them in the microservices.
- Passing tests: I started a list of increasingly difficult tests my ACE needs to pass. It starts from "Can you repeat this word", to writing emails, to eventually multi-steps complex problems.
Difficulties
My difficulties at the moment are:
- Memory querying: Memories are stored in a Semantic DB (Pinecone). However, I did not figure out how to properly retrieve memories in an Q&A format. ie: What is my relationship with this person? --> Should return memories like "I like this person and think X/Y".
- Memory consolidation: At the moment the thoughts are piling up in Josh's "brain". I need to find the best ways to distill & consolidate memories (merging similar one, removing unimportant ones etc.). Did not really start this yet.
- Visual processing: At the moment Josh is only capable of treating text. I faked audio understanding with a speech-to-text model, but most visual information would be lost if using image-to-text models. I could add a DB to store visual info, but I am unsure how the info would then be linked to the semantic part of the brain.
- Funding: That is a big problem for me atm. I would like to stay full-time, because I'm already struggling to keep up with say Sidney from Bing (I'm not far behind I have to say). I applied for a grant, but I was wondering if you had suggestions.
​
Here is my current (simplified) architecture (I'll need to do some refactoring/cleaning at some point):
​
Here are a couple screenshots:
I'd be glad to answer to any questions, and I'm also open to suggestions!
Best,
Lester
nikitastaf1996 t1_j9uim0x wrote
I have seen one similar project on YouTube.Where there is two there is ten.I don't know what that will lead to.But quantity often converges to quality.