entropyvsenergy
entropyvsenergy t1_j9gn52y wrote
entropyvsenergy t1_j8bvxvs wrote
Brew on the grid
entropyvsenergy t1_j7krhzx wrote
Reply to comment by Unlucky-Boot-6567 in Haunted locations in Worcester and region? by CourtCreepy6785
The Skull Society building at WPI has some history behind it too.
entropyvsenergy t1_j5s3ikx wrote
entropyvsenergy t1_j2uy0ov wrote
Reply to Is that Jewish deli on highland st any good? Helfands I think it’s called by Sithlordbelichick
It's pretty good. Scratches the pastrami on rye itch. Good wraps too. Didn't love their latkes.
entropyvsenergy t1_j0gfdj9 wrote
Reply to What do you guys pay in water/sewer? it seems high but Its my first house so I have nothing to compare it to. by UsrNameAlrdyFaknTakn
I live in a 3 BR, 1.5 bathroom house. We pay $35.00 a month.
entropyvsenergy t1_iyd6mw0 wrote
Reply to comment by minhrongcon2000 in [D] Does Transformer need huge pretraining process? by minhrongcon2000
Transformers do well with lots of data. This is because the transformer is an extremely flexible and generic architecture. Unlike a fully connected neural network where each input is mapped through a weight matrix to the next layer and the weight matrices are fixed with respect to any input, transformers use attention blocks where the actual "effective" weight matrices are computed using the attention operation using query, key, and value vectors and thus depend on the inputs. What this means is that in order to train a transformer model you need a lot of data in order to get better performance than less flexible neural network architectures such as LSTMs or fully connected networks.
entropyvsenergy t1_iw58dge wrote
It's all frameworks now, some better than others. I haven't written one outside of demos or interviews in years. With that being said, I've modified neural networks a whole bunch. Usually you can just tweak parameters in a config file but sometimes you want additional outputs or to fundamentally change the model in some way...usually minor tweaks codewise.
entropyvsenergy t1_ivqltwc wrote
Reply to [D] Is there an advantage in learning when taking the average Gradient compared to the Gradient of just one point by CPOOCPOS
Batching does this, generally and it's a good thing for stability. Reduces the variance of the gradient update proportional to the batch size.
entropyvsenergy t1_iuiwahp wrote
Reply to comment by your_city_councilor in Communities in Worcester divided over Question 5 by HRJafael
It's 1.5% on top of the tax not on top of the house value. So if you own a $700k house, your tax bill is likely around $8,000 so this would add $120 (1.5% of $8,000) to your tax bill.
entropyvsenergy t1_itv6d3r wrote
WRTA is free. MBTA local buses are $1.70 with a Charlie Card or $11 for a 1-day pass.
The buses are pretty good, though they may arrive early or late so factor in arriving early to the stop and give yourself a buffer of time if you need to arrive somewhere by a specific time.
entropyvsenergy t1_ith6v4w wrote
That sentence doesn't make any sense to me as a native English speaker either.
entropyvsenergy t1_jcroge4 wrote
Reply to Where are there good mountain bike trails? The gnarlier the better! by CapGrundle
Greenbriar in Oxford has some good ones.