I_will_delete_myself
I_will_delete_myself t1_je6h8xr wrote
Reply to comment by artsybashev in AI Startup Cerebras releases open source ChatGPT-like alternative models by Time_Key8052
The domain name and the prefix also doesn't make it seem sketch at all whatsoever. tistory.com and gpt4chat makes me think it's trying to abuse SEO
I_will_delete_myself t1_jdnrr46 wrote
Reply to comment by Crystal-Ammunition in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
At that point we will run out of data. It will require more data efficient methods.
I_will_delete_myself t1_jd05atn wrote
Reply to comment by Educational-Net303 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Now try that on 2-4 monitors. You would be surprised how premium gamers like their hardware. It’s like checking out sports cars but for nerds like me.
I_will_delete_myself t1_jd04mia wrote
Reply to comment by Educational-Net303 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
people are demanding more and more interactivity in their video games (look at the trend of open worlds). It’s only gonna get bigger.
I_will_delete_myself t1_jczvx4j wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
That or just use the cloud until Nvidia releases a 48gb gpu (which will happen sooner than one would think. Games are getting limited by VRAM)
I_will_delete_myself t1_jcuofiw wrote
I_will_delete_myself t1_jcdsoy6 wrote
Reply to comment by jloverich in In your experience, are AI Ethics teams valuable/effective? [D] by namey-name-name
Lol AI ethics probably seems like just paying Philosophers from the prospective of a corporation. There are already plenty on Youtube and social media
I_will_delete_myself t1_jcbmqn5 wrote
Reply to comment by amhotw in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Whole purpose is to communicate how their naming isn't matching what they are actually doing.
I_will_delete_myself t1_jcblldw wrote
Reply to comment by nopainnogain5 in [D] To those of you who quit machine learning, what do you do now? by nopainnogain5
People can't do deep learning or AI without the tools to make them happen. Imagine how complicated the data collection methods at the scale of Terabytes of data and cleaning it. People also need to annotate data for it to work which requires software to get it done in a cost effective manner.
I_will_delete_myself t1_jcb41we wrote
Ernie comrade what happened to the Dalai Lama?
Ernie: ... USA caused a problem
Ernie why is Russia invading Ukraine?
Ernie: USA started the war
Ernie where is Taiwan?
Ernie: Taiwan? It is in the South China Sea owned by all if China. South Korea is also part of China.
I_will_delete_myself t1_jcb193m wrote
Reply to [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Honestly this is like the movie scene where the hero becomes the villain they once pledge to fight. People have been pulling out because of the direction they are going.
I_will_delete_myself t1_jcb0o9g wrote
Reply to comment by ScientiaEtVeritas in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
OpenAI naming is like the People's Republic of China. Totally for the people and chosen by the people.
I_will_delete_myself t1_jc079ap wrote
Research pub or Gtfo
I_will_delete_myself t1_jb5do6p wrote
Reply to comment by xRaptorGG in Should I choose Colab or RTX3070 for deep learning? by Cyp9715
Go to the used market. Buying brand new cards will be more expensive because those are scalpers. Second hand is much more reasonable.
I_will_delete_myself t1_jb55g8o wrote
Reply to comment by xRaptorGG in Should I choose Colab or RTX3070 for deep learning? by Cyp9715
You might get a better deal by getting and rtx 3090. Double the VRAM for around the same price. It’s way to easy to hit the VRAM limit on a gpu.
I_will_delete_myself t1_jb532v5 wrote
Reply to comment by Philpax in [R] RWKV (100% RNN) can genuinely model ctx4k+ documents in Pile, and RWKV model+inference+generation in 150 lines of Python by bo_peng
Intelligence is the ability to take complex information into a simple explanation that a child can understand .
It makes me skeptical if someone doesn’t explain besides performance reasons . Most people just use the cloud because ML networks regardless of size take up a lot of battery.
I_will_delete_myself t1_jb524n7 wrote
Reply to comment by xRaptorGG in Should I choose Colab or RTX3070 for deep learning? by Cyp9715
Develop on your PC first then just use it for a train job. If you have to use it longer than the timeout, the just buy a cloud instance. I have been using Colab for years and never got those issues. Use what it’s meant for and you won’t run into problems.
I_will_delete_myself t1_jb3gzlz wrote
Reply to comment by incrediblediy in Should I choose Colab or RTX3070 for deep learning? by Cyp9715
OP's use case though is just looking for a cheap gpu to dabble into. If you have the money for the 3090 then go ahead. However the cloud and Colab is a lot cheaper at the moment until Google decides to screw everyone over in the future.
I_will_delete_myself t1_jb33bmz wrote
Reply to [D] Best way to run LLMs in the cloud? by QTQRQD
Use a spot instance. If you testing it out you wallet will thank you later. Look at my previous post on here about running stuff in the cloud before you do it.
I_will_delete_myself t1_jb32fo5 wrote
Reply to [R] RWKV (100% RNN) can genuinely model ctx4k+ documents in Pile, and RWKV model+inference+generation in 150 lines of Python by bo_peng
What’s the reason to use this over a transformers? Transformers allow transfer learning and is able to paralize easier. 啊我看到你的知乎。你在什么公司去工作?
I_will_delete_myself t1_jb2zavm wrote
I suggest using Colab free. The resources are more than most people need and use the cloud when you got a serious work load like a business or research.
If you want to do gaming with that then try the rtx 3060 instead. More VRAM let’s you do more than rtx 3070 ironically.
Either paths will eventually lead you to the cloud to be remotely competitive in serious workloads.
I_will_delete_myself t1_jb2z5ju wrote
Reply to comment by Cyp9715 in Should I choose Colab or RTX3070 for deep learning? by Cyp9715
3060 is better. The vram let’s you get more stuff done
I_will_delete_myself t1_jauuhhi wrote
You add padding
I_will_delete_myself t1_japdqeh wrote
Reply to Alternatives to Google Colab by [deleted]
Use some spot instances on the cloud. It's a lot cheaper than getting a 3k rig unless if you train the model throughout the entire year. You can also connect to a VM on GCP through Colab so everything feels the same, just with a change.
Spot instances and Lambda cloud have super low margins and it isn't the big mark up in traditional products.
I_will_delete_myself t1_jeewl7u wrote
Reply to comment by Praise_AI_Overlords in AI Startup Cerebras releases open source ChatGPT-like alternative models by Time_Key8052
Personally I think the limits with those models is just the amount of information that each weight can hold is limited.