michaelthwan_ai
michaelthwan_ai OP t1_jdpy7yf wrote
Reply to comment by StellaAthena in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
I may include BARD if it is fully released.
So LAMDA->BARD (maybe). But it is still in alpha/beta.
michaelthwan_ai OP t1_jdpy5dy wrote
Reply to comment by tonicinhibition in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
Thanks for the sharing above!
My choice is yk - Yannic Kilcher. Some "AI News" videos is a brief introduction and he sometimes go through certain papers in details. Very insightful!
michaelthwan_ai OP t1_jdpy0l2 wrote
Reply to comment by DigThatData in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
added, thank you.
michaelthwan_ai OP t1_jdpy06p wrote
Reply to comment by philipgutjahr in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
I only include recent LLM (Feb/Mar 2023) (that is the LLMs usually at the bottom) and 2-factor predecessors (parent/grandparent). See if your mentioned one is related to them.
michaelthwan_ai OP t1_jdpxuzs wrote
Reply to comment by Small-Fall-6500 in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
Chatgpt-like github -> added most and so is in TODO (e.g. palm)
RWKV -> added in backlog
michaelthwan_ai OP t1_jdpxtp1 wrote
Reply to comment by philipgutjahr in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
Open alternative -> added most and so is in TODO (e.g. palm)
OpenChatKit -> added
Instruct-GPT -> seems it's not a released model but plan.
michaelthwan_ai OP t1_jdm3v2y wrote
Reply to comment by light24bulbs in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
Please suggest so.
michaelthwan_ai OP t1_jdm3sgb wrote
Reply to comment by wywywywy in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
Agreed
michaelthwan_ai OP t1_jdlzwyi wrote
Reply to comment by wywywywy in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
Sure I think it is clear enough to show parents of recent model (instead of their grand grand grand parents..
If people want, I may consider to make a full one (including older one)
michaelthwan_ai OP t1_jdlztvv wrote
Reply to comment by addandsubtract in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
It is a good model but it's about one year ago, and not related to recent released LLM. Therefore I didn't add (otherwise a tons of good models).
For dolly, it is just ytd. I didn't have full info of it yet
michaelthwan_ai OP t1_jdlzq6i wrote
Reply to comment by Historical-Tree9132 in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
I considered haha but I have no evidence
michaelthwan_ai OP t1_jdlf8g8 wrote
Because the recent release of LLMs has been too vigorous, I organized recent notable models from the news. Some may find the diagram useful, so please allow me to distribute it.
Please let me know if there is anything I should change or add so that I can learn. Thank you very much.
If you want to edit or create an issue, please use this repo.
---------EDIT 20230326
Thank you for your responses, I've learnt a lot. I have updated the chart:
- https://github.com/michaelthwan/llm_family_chart/blob/master/LLMfamily2023Mar.drawio.png
- (Look like I cannot edit the post)
Changes 20230326:
- Added: OpenChatKit, Dolly and their predecessors
- More high-res
To learn:
- RWKV/ChatRWKV related, PaLM-rlhf-pytorch
Models that not considered (yet)
- Models that is <= 2022 (e.g. T5 (2022May). This post is created to help people quickly gather information about new models)
- Models that is not fully released yet (e.g. Bard, under limited review)
Submitted by michaelthwan_ai t3_121domd in MachineLearning
michaelthwan_ai OP t1_jcyo94y wrote
Reply to [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Added an "examples of prompts" on the top for showcases!
michaelthwan_ai OP t1_jcy73od wrote
Reply to comment by BalorNG in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Your idea sounds like GAN - maybe one model will generate high-quality synthetic data and another one try to 'discriminate' it, then they may output an ultra-high quality one finally (for another model to eat). And an AI model community is formed to self-improve...
michaelthwan_ai OP t1_jcxx6ib wrote
Reply to comment by BalorNG in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Yeah great summary related to the memory.
My next target may be related to compact models (which preserve good results), as I also believe it is the way to go :D
michaelthwan_ai OP t1_jcxsd0x wrote
Reply to comment by egoistpizza in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Thank you for your comprehensive input.
- I have mixed feeling about opening/closing the technology. There are pros/cons to it. For example, we, especially people in this field have a strong curiosity about how giant technology solves their problems (like chatgpt). Therefore open-sourcing them will bring us rapid development in related fields (like the current AI development). However, I also understand that, malicious usage is also highly possible when doing so. For example, switching the reward function from chatgpt model from positive to negative may make a safe AI into the worst AI ever.
- Humans seem to not be able to stop technological advancement. Those technologies will come sooner or later.
- Yes I agree to preserve our rights today and the society should carefully think about how to deal with this unavoidable (AI-powered) future.
michaelthwan_ai OP t1_jcxrjbm wrote
Reply to comment by Educational_Ice151 in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Thank you!
michaelthwan_ai OP t1_jcxrilh wrote
Reply to comment by Secret-Fox-5238 in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
haha. Nice point.
I'm not sure whether it fulfil the definition of a search engine, but this work essentially mimics your experiences during googling: Google->got n websites->surf and find info one by one.
SearchGPT (or e.g. new Bing) attempted to automate this process. (Thus Google is unhappy)
michaelthwan_ai OP t1_jcxrcfu wrote
Reply to comment by BalorNG in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
I agree with you. 3 thoughts from me
- I think one direction of the so-called safety AI to give a genuine answer, is to give it factual/external info. I mean 1) a Retrieval-based model like searchGPT 2) API calling like toolformer (e.g. check weather API)
- LLM, is essentially a compression problem (I got the idea in lambdalabs). But it cannot remember everything. Therefore an efficient way to solve so are retrieval methods to search a very large space (like pagerank/google search), then obtain a smaller result set and let the LLM organize and filter related content from it.
- Humans are basically like that right? But if we got a query, we may need to read books (external retrieval) which is pretty slow. However, humans have a cool feature, long-term memory, to store things permanently. Imagine if an LLM can select appropriate things during your queries/chat and store them as a text or knowledge base inside it, then it is a knowledge profile to permanently remember the context bonded between you and the AI, instead of the current situation that ChatGPT will forget everything after a restart.
michaelthwan_ai OP t1_jcws6h8 wrote
Reply to comment by Secret-Fox-5238 in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
ChatGPT said what I want to say.
>I apologize for any confusion or misinformation in my previous response. You are correct that SQL databases do support various text search and similarity matching features, including the use of keywords like LIKE and CTE (Common Table Expressions) to enable more flexible and efficient querying.
>
>While it's true that specialized tools like Elasticsearch, Solr, or Algolia may offer additional features and performance benefits for certain natural language processing tasks, SQL databases can still be a powerful and effective tool for storing and querying structured and unstructured data, including text data.
>
>Thank you for bringing this to my attention and allowing me to clarify my previous response.
michaelthwan_ai OP t1_jcw1cnv wrote
Reply to comment by hassan789_ in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
added some credits to it . Used up all. I will monitor the usage
michaelthwan_ai OP t1_jctw0gx wrote
Reply to comment by ramtingxf in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
It is a little bit exaggerated but thanks! I believe Bing and some companies are using similar tech, but a highly polished one, to solve similar issues.
michaelthwan_ai OP t1_jctvmqe wrote
Reply to comment by [deleted] in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Cool! Thanks for the sharing.
During my development, I've also found 5+ projects, some open-source and some are closed, where they are doing similar things.
In exact, it is called retrieval-based language model.
Some discussion on that:
michaelthwan_ai OP t1_jdpyb80 wrote
Reply to comment by Puzzleheaded_Acadia1 in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
added in backlog. Need some time to study. Thanks.