Search
50 results for towardsdatascience.com:
nibbels t1_irvvnpw wrote
Reply to [D] Looking for some critiques on recent development of machine learning by fromnighttilldawn
with both the field and the models. https://arxiv.org/abs/2011.03395 https://openreview.net/forum?id=xNOVfCCvDpM https://arxiv.org/abs/2110.09485#:~:text=The%20notion%20of%20interpolation%20and,outside%20of%20that%20convex%20hull. https://towardsdatascience.com/the-reproducibility-crisis-and-why-its-bad-for-ai-c8179b0f5d38 https://ai100.stanford.edu/2021-report/standing-questions-and-responses/sq10-what-are-most-pressing-dangers-ai And then, of course, there are the oft-discussed topics like bias
build_saas_reddit t1_iry4uhw wrote
Take a look at Pegasus : [https://arxiv.org/abs/1912.08777](https://arxiv.org/abs/1912.08777) [https://towardsdatascience.com/how-to-perform-abstractive-summarization-with-pegasus-3dd74e48bafb](https://towardsdatascience.com/how-to-perform-abstractive-summarization-with-pegasus-3dd74e48bafb)
ktpr t1_is12rss wrote
might want to include a slide on critical takes. For example, see this popular article: https://towardsdatascience.com/whats-wrong-with-lime-86b335f34612
beerus1729 t1_itb18as wrote
towards data science ](http://towardsdatascience.com). This site is good. But there is limitation on the number of articles you can read in a month without paying money
merouane_nz t1_iu3idz0 wrote
Reply to comment by MariiaKozlova in [D] Simple Questions Thread by AutoModerator
check SHAP values. https://shap.readthedocs.io/en/latest/index.html https://towardsdatascience.com/shap-explained-the-way-i-wish-someone-explained-it-to-me-ab81cc69ef30
BlazeObsidian t1_iujbbdu wrote
Reply to comment by alexnasla in [D] When the GPU is NOT the bottleneck...? by alexnasla
sure you model is running on the GPU ? See [https://towardsdatascience.com/pytorch-switching-to-the-gpu-a7c0b21e8a99](https://towardsdatascience.com/pytorch-switching-to-the-gpu-a7c0b21e8a99) or if you can see GPU utilisation it might be simpler to verify. If you are not explicitly moving your model
CapaneusPrime t1_iv0npgi wrote
Reply to comment by Takahashi_Raya in [N] Class-action lawsuit filed against GitHub, Microsoft, and OpenAI regarding the legality of GitHub Copilot, an AI-using tool for programmers by Wiskkey
wrong. https://towardsdatascience.com/the-most-important-supreme-court-decision-for-data-science-and-machine-learning-44cfc1c1bcaf
CapaneusPrime t1_iv0t82o wrote
Reply to comment by killver in [N] Class-action lawsuit filed against GitHub, Microsoft, and OpenAI regarding the legality of GitHub Copilot, an AI-using tool for programmers by Wiskkey
towardsdatascience.com/the-most-important-supreme-court-decision-for-data-science-and-machine-learning-44cfc1c1bcaf
mkzoucha t1_ivau561 wrote
Reply to comment by [deleted] in [D] Simple Questions Thread by AutoModerator
this will Ferrell movie also watched these movies. Edit: here is a good overview I found https://towardsdatascience.com/recommendation-systems-explained-a42fc60591ed
blarg7459 t1_iw94geo wrote
Reply to comment by allwordsaremadeup in [Research] Can we possibly get access to large language models (PaLM 540B, etc) like GPT-3 but no cost? by NLP2829
Here's how to run it locally https://towardsdatascience.com/run-bloom-the-largest-open-access-ai-model-on-your-desktop-computer-f48e1e2a9a32
Heap_Good_Firewater t1_iwld8mc wrote
Reply to comment by paradox-snail in Will working for a DAO be better than a corporate job? by berlinparisexpress
only be safely used by software engineers seems like trading one set of problems for another. [https://towardsdatascience.com/the-blockchain-scalability-problem-the-race-for-visa-like-transaction-speed-5cce48f9d44](https://towardsdatascience.com/the-blockchain-scalability-problem-the-race-for-visa-like-transaction-speed-5cce48f9d44) >The music industry is a good analog for what's happening in finance. The music
glass_superman t1_ixfujfq wrote
Reply to comment by Appletarted1 in The Ethics of Policing Algorithms by ADefiniteDescription
being a good riot-stopping AI. Eliazar Yudowsky did the "escape the box" thing twice. https://towardsdatascience.com/the-ai-box-experiment-18b139899936 Even if you don't find these arguments fully convincing, hopefully between the YouTube example, the quants
Appropriate_Ant_4629 t1_ixtlme6 wrote
Reply to comment by RichardBJ1 in Is Linux still vastly preferred for deep learning over Windows? by moekou
towardsdatascience.com/installing-pytorch-on-apple-m1-chip-with-gpu-acceleration-3351dc44d67c >> Installing PyTorch on Apple M1 chip with GPU Acceleration
brucebay t1_iy535m2 wrote
good on this. I don't remember the article url, but it may have been this. https://towardsdatascience.com/how-autoencoders-outperform-pca-in-dimensionality-reduction-1ae44c68b42f
I-am_Sleepy t1_iy7xfo0 wrote
Reply to comment by Pomdapi113 in [D] Simple Questions Thread by AutoModerator
genetic algorithm, or bayesian optimization For bayesian, if your prior is normal, then its [conjugate prior](https://towardsdatascience.com/conjugate-prior-explained-75957dc80bfb) is also normal. For multivariate, it is a bit trickier, depends on your settings (likelihood distribution
Visual-Arm-7375 OP t1_iz04glk wrote
Reply to comment by killver in [D] Model comparison (train/test vs cross-validation) by Visual-Arm-7375
Have a look at this: https://towardsdatascience.com/train-test-split-c3eed34f763b
RuairiSpain t1_izunqwy wrote
Reply to [D] Industry folks, what kind of development methodology/cycle do you use? by DisWastingMyTime
then gives you hints about which ML modeling to choose from. See this article on stats: https://towardsdatascience.com/10-machine-learning-methods-that-every-data-scientist-should-know-3cc96e0eeee9 The developer stage is more at the tail end where you look at refactoring the algorithm
isuckwithusernames t1_izy6dhd wrote
Reply to comment by Melodic-Oil-1971 in does anyone know how to build neural network from scratch? by [deleted]
believe you spent more than 15 minutes searching for an answer, but here you go: [https://towardsdatascience.com/how-to-build-your-own-neural-network-from-scratch-in-python-68998a08e4f6](https://towardsdatascience.com/how-to-build-your-own-neural-network-from-scratch-in-python-68998a08e4f6)
ResponsibilityNo7189 t1_j02dzwf wrote
Reply to [P] Are probabilities from multi-label image classification networks calibrated? by alkaway
your network probabilities to be calibrated. First you might want to read aleatoric vs. epistemic uncertainty. [https://towardsdatascience.com/aleatoric-and-epistemic-uncertainty-in-deep-learning-77e5c51f9423](https://towardsdatascience.com/aleatoric-and-epistemic-uncertainty-in-deep-learning-77e5c51f9423) MonteCarlo sampling and training have been used to get a sense of uncertainty. Also changing
DreamWatcher_ t1_j03smkx wrote
Reply to comment by cole_braell in Is it just me or does it feel like GPT-4 will basically be game over for the existing world order? by Practical-Mix-4332
billion parameters, but it's performance is still lower than neural networks with fewer parameters. [https://towardsdatascience.com/gpt-4-is-coming-soon-heres-what-we-know-about-it-64db058cfd45](https://towardsdatascience.com/gpt-4-is-coming-soon-heres-what-we-know-about-it-64db058cfd45)
biophysninja t1_j0a4nc2 wrote
approach this depending on the nature of the data, complexity, and compute available. 1- using SMOTE https://towardsdatascience.com/stop-using-smote-to-handle-all-your-imbalanced-data-34403399d3be 2- if your data is sparse you can use PCA or Autoencoders to reduce the dimensionality
I-am_Sleepy t1_j0mhitl wrote
Reply to comment by honchokomodo in [D] Simple Questions Thread by AutoModerator
starter, look at InfoVAE ([See this blog](https://towardsdatascience.com/with-great-power-comes-poor-latent-codes-representation-learning-in-vaes-pt-2-57403690e92b) for context). Another way is to vector-quantized it (VQ-VAE based models), as the model only need to learn a small number of latent
sharky6000 t1_j10ay53 wrote
Reply to [D] Why are we stuck with Python for something that require so much speed and parallelism (neural networks)? by vprokopev
than Go, now, I think but Go has Gorgonia, for Rust there is a Torch interface https://towardsdatascience.com/machine-learning-and-rust-part-4-neural-networks-in-torch-85ee623f87a) But often you start going down these roads to later find that they're not worth
sayoonarachu t1_j1408am wrote
others stated, it'll take forever to be usable. I.e 30 minutes for 10 token. https://towardsdatascience.com/run-bloom-the-largest-open-access-ai-model-on-your-desktop-computer-f48e1e2a9a32
jharel t1_j264g6n wrote
following is my explanation. Perhaps I'll try to find time to post about it. [https://towardsdatascience.com/artificial-consciousness-is-impossible-c1b2ab0bdc46](https://towardsdatascience.com/artificial-consciousness-is-impossible-c1b2ab0bdc46)
jharel t1_j2bmsex wrote
Reply to Accepting Science Fiction by Exiled_to_Earth
imagine something happening in the future then it must be inevitable future fact." [https://towardsdatascience.com/artificial-consciousness-is-impossible-c1b2ab0bdc46](https://towardsdatascience.com/artificial-consciousness-is-impossible-c1b2ab0bdc46)
glueball71 t1_j2j2akb wrote
Reply to comment by ateeb_khan_13 in [D] Simple Questions Thread by AutoModerator
towardsdatascience.com/deploying-a-tensorflow-model-to-production-made-easy-4736b2437103
Scarlet_pot2 OP t1_j39g574 wrote
Reply to comment by Cryptizard in We need more small groups and individuals trying to build AGI by Scarlet_pot2
associations from a large corpus of text." This was the first "guess the next word" model. [https://towardsdatascience.com/attention-is-all-you-need-discovering-the-transformer-paper-73e5ff5e0634](https://towardsdatascience.com/attention-is-all-you-need-discovering-the-transformer-paper-73e5ff5e0634) This next link is the "Attention is all you need" paper that describes how to build
currentscurrents t1_j3jst6n wrote
Reply to comment by Immarhinocerous in [Discussion] Is there any alternative of deep learning ? by sidney_lumet
know there's a whole field of [decision tree learning](https://towardsdatascience.com/a-visual-guide-to-gradient-boosted-trees-8d9ed578b33), but I'm not super up to date on it. I assume neural networks are better or else we'd be using
m98789 t1_j3xxyvm wrote
Reply to comment by All-DayErrDay in [D] Microsoft ChatGPT investment isn't about Bing but about Cortana by fintechSGNYC
that it took $12M in compute costs for a single training run of GPT-3 ([source](https://towardsdatascience.com/the-future-of-ai-is-decentralized-848d4931a29a)). H100s will make a significant difference and all the optimization techniques. So I agree prices will
MustachedLobster t1_j45dp6k wrote
Reply to comment by [deleted] in [D] Has ML become synonymous with AI? by Valachio
performance at tasks in T, as measured by P, improves with experience E.” https://towardsdatascience.com/what-is-machine-learning-and-types-of-machine-learning-andrews-machine-learning-part-1-9cd9755bc647#:~:text=Tom%20Mitchell%20provides%20a%20more,simple%20example%20to%20understand%20better%20.
MustachedLobster t1_j47oa1s wrote
Reply to comment by chief167 in [D] Has ML become synonymous with AI? by Valachio
performance at tasks in T, as measured by P, improves with experience E. https://towardsdatascience.com/what-is-machine-learning-and-types-of-machine-learning-andrews-machine-learning-part-1-9cd9755bc647#:~:text=Tom%20Mitchell%20provides%20a%20more,simple%20example%20to%20understand%20better%20. Localisation error decreases the more data you have
BigZaddyZ3 t1_j554ioa wrote
Reply to comment by solardeveloper in ChatGPT really surprised me today. by GlassAmazing4219
still capable of exceeding human abilities *already*. [AI is already more accurate than doctors.](https://towardsdatascience.com/ai-diagnoses-disease-better-than-your-doctor-study-finds-a5cc0ffbf32) Thinking AI won’t eventually exceed even the best human minds in pretty much every sector is basically
terath t1_j5kz6tz wrote
Reply to [D] Embedding bags for LLMs by WigglyHypersurface
tokenizers and many language models are built on them. Here is a quick article on them: https://towardsdatascience.com/byte-pair-encoding-subword-based-tokenization-algorithm-77828a70bee0
Autogazer t1_j6gyhc9 wrote
Reply to [D] AI Theory - Signal Processing? by a_khalid1999
towardsdatascience.com/how-fourier-transform-speeds-up-training-cnns-30bedfdf70f6
MagicalPeanut t1_j7en0ux wrote
Reply to What would you tell your realtor if you just bought a house and zillow said it has already lost 100k in value? by NecessaryMistake9754
just how bad Zillow's data scientists failed Zilllow when it came to pricing the market: [https://towardsdatascience.com/invaluable-data-science-lessons-to-learn-from-the-failure-of-zillows-flipping-business-25fdc218a62](https://towardsdatascience.com/invaluable-data-science-lessons-to-learn-from-the-failure-of-zillows-flipping-business-25fdc218a62) Feel better yet? My hot take is that prices should go down for everyone as rates
ImZanga t1_j8gh6ci wrote
reliable: * [Why TikTok made its user so obsessive? The AI Algorithm that got you hooked.](https://towardsdatascience.com/why-tiktok-made-its-user-so-obsessive-the-ai-algorithm-that-got-you-hooked-7895bb1ab423) * [The App That Knows You Better than You Know Yourself: An Analysis of the TikTok Algorithm
[deleted] OP t1_j8rdy53 wrote
possible reason [https://towardsdatascience.com/r-cnn-fast-r-cnn-faster-r-cnn-yolo-object-detection-algorithms-36d53571365e](https://towardsdatascience.com/r-cnn-fast-r-cnn-faster-r-cnn-yolo-object-detection-algorithms-36d53571365e) i.e the VGG convolutional model wont be good for bounding boxes but only for classification task
buyIdris666 t1_j9h8zqj wrote
Reply to comment by _Arsenie_Boca_ in [D] Bottleneck Layers: What's your intuition? by _Arsenie_Boca_
similar model to U-Net and other "bottlenecked" CNN architectures check out denoising auto-encoders https://towardsdatascience.com/applied-deep-learning-part-3-autoencoders-1c083af4d798 There is currently speculation that Diffusion models are simply a return of the ancient (7 years
kompootor t1_ja24h53 wrote
Reply to Why the development of artificial general intelligence could be the most dangerous new arms race since nuclear weapons by jamesj
flag. The next paragraph actually helpfully links to [Towards Data Science's page on Transformer](https://towardsdatascience.com/transformers-141e32e69591), which is actually really good in that they illustrates, complete with animations, the mechanics of ANNs
tonicinhibition t1_jb1fgpe wrote
Reply to comment by tripple13 in To RL or Not to RL? [D] by vidul7498
pretty well for anyone else who is curious: [Do GANS really model the true data distribution...](https://towardsdatascience.com/do-gans-really-model-the-true-data-distribution-or-are-they-just-cleverly-fooling-us-d08df69f25eb) For further nuance on this topic, Machine Learning Street Talk discussed interpolation vs extrapolation with Yann
Surur t1_jb78a7f wrote
Reply to comment by agsarria in What might slow this down? by Beautiful-Cancel6235
Here is an interesting article addressing fixing LLM issues, including hallucinations. https://towardsdatascience.com/overcoming-the-limitations-of-large-language-models-9d4e92ad9823
dfcHeadChair t1_jbaljm8 wrote
Reply to newby here. looking for help on a MLP for speech recognition. any tips or pointers would be appreciated by alexilas
solution, but if it's for a class and you can only use numpy start here: [https://towardsdatascience.com/coding-a-neural-network-from-scratch-in-numpy-31f04e4d605](https://towardsdatascience.com/coding-a-neural-network-from-scratch-in-numpy-31f04e4d605)
lifesthateasy t1_jbim6l5 wrote
Reply to comment by crappleIcrap in [D] Blake Lemoine: I Worked on Google's AI. My Fears Are Coming True. by blabboy
else who can explain it to you why brain neurons and artificial neurons are fundamentally different: [https://towardsdatascience.com/the-differences-between-artificial-and-biological-neural-networks-a8b46db828b7](https://towardsdatascience.com/the-differences-between-artificial-and-biological-neural-networks-a8b46db828b7) Even this article has some omissions, and I want to highlight how in the past
gradientic t1_jc430rz wrote
Reply to Using GANs to generate defective data by Tekno-12345
that you have a significant dataset of images without anomalies - learn inlier look for outliers) - check https://towardsdatascience.com/an-effective-approach-for-image-anomaly-detection-7b1d08a9935b for some ideas and pointers (sry if pointing obvious things
jarmosie t1_jddmvp9 wrote
Reply to [D] Simple Questions Thread by AutoModerator
high quality online content through individual blogs or newsletter. I know there's [Towards Data Science](https://towardsdatascience.com) & [Machine Learning Mastery](https://machinelearningmastery.com/) to name a few but what other lesser known
tacixat t1_irqufhl wrote
Reply to [R] Clustering a set of graphs by No_Performer203
Markov clustering! Built for graphs, does not require a fixed number of clusters. https://towardsdatascience.com/markov-clustering-algorithm-577168dad475
1714alpha t1_jdt21l2 wrote
Reply to A Problem That Keeps Me Up At Night. by circleuranus
Hell, they're are already programs that can [diagnose illnesses better than human doctors ](https://towardsdatascience.com/ai-diagnoses-disease-better-than-your-doctor-study-finds-a5cc0ffbf32). To your point, it would indeed be problematic if any single source of information became the unquestioned authority
No-Belt7582 t1_j9j6edy wrote
Reply to comment by vurt72 in [D] Stable Diffusion, Class images question by vurt72
Foong, “How to Fine-tune Stable Diffusion using Textual Inversion,” Medium, Oct. 24, 2022. https://towardsdatascience.com/how-to-fine-tune-stable-diffusion-using-textual-inversion-b995d7ecc095 (accessed Feb. 09, 2023). [10] N. W. Foong, “How to Fine-tune Stable Diffusion using Dreambooth,” Medium ... towardsdatascience.com/how-to-fine-tune-stable-diffusion-using-dreambooth-dfa6694524ae (accessed Feb. 09, 2023). [11] “The Annotated Diffusion Model.” https://huggingface.co/blog/annotated-diffusion (accessed Jan. 31, 2023). [12] J. Alammar, “The Illustrated Stable Diffusion.” https://jalammar.github.io/illustrated-stable-diffusion/ (accessed Jan. 31, 2023). [13] “Understanding Stable