jobeta
jobeta t1_jbhg8xy wrote
Reply to comment by keisukegoda3804 in [D] Text embedding model for financial documents by [deleted]
Right. But to be able to assess this you need to define a task and evaluate your model’s performance to perform that task. Embedding accuracy cannot be discussed completely in the ether. Even the most general comments you will read about a model beating another, will refer to that new model performing better for specific tasks on benchmark datasets.
It would be a lot easier to help you if you explained what you are trying to accomplish that requires “higher accuracy”
jobeta t1_jbgnfsq wrote
> a model that can generate more accurate embeddings in general
What do you mean by this? Accuracy is an evaluation metric that doesn't really mean anything "in general" but only with respect to a specific prediction being made. I think this is a slightly ill-posed question
jobeta t1_jb3hh74 wrote
This is cool and I haven’t finished reading it yet but, intuitively, isn’t that roughly equivalent to have a higher learning rate in the beginning? You make the learning algorithm purposefully imprecise at the beginning to explore quickly the loss landscape and later on, once a rough approximation of a minimum has been found, you are able to explore more carefully to look for a deeper minimum or something? Like the dropout introduces noise doesn’t it?
jobeta t1_jb1otw2 wrote
Neat! What's early s.d. in the tables in the github repo?
jobeta t1_ja1jdgc wrote
You don’t need code. You can use a service for that. Check Descript overdub for instance. Or whatever other similar thing you can find. I’m not affiliated with them but saw a demo. It will be done overnight after you spend 20 min reading some text.
jobeta t1_ja0m0og wrote
Reply to comment by SuchOccasion457 in [D] Cost of data acquisition by SuchOccasion457
Yes but just pick two or three and ask? Also check on Amazon mechanical Turk if you find labeling job listed and the rates. I have only needed this one but used upwork. We paid well and it was a while ago so I don’t think the price I will give you will be a good reference.
jobeta t1_ja08xy1 wrote
Reply to [D] Cost of data acquisition by SuchOccasion457
I don’t think there is a general answer to that. For labeling there are multiple services that you can use. You could just contact them and ask or look if they advertise how much they pay people to label to get a proxy. For the data itself, it completely depends on the data. I would imagine medical data would be hard to obtain and require some legal consideration around privacy (at least I would hope so).
jobeta t1_j7kgvcc wrote
Reply to comment by aicharades in [P] ChatGPT without size limits: upload any pdf and apply any prompt to it by aicharades
How much did you pay for that single prompt?
jobeta t1_j7jda4k wrote
Reply to comment by ThrillHouseofMirth in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
It’s clearly the case already. Shutterstock sold pictures to open-AI to create Dalle-2. Which will soon be used to create what used to be stock photography. This example here is ridiculously bad tho 🤣
jobeta t1_j6gpgqj wrote
Reply to comment by tysam_and_co in [R] Train CIFAR10 in under 10 seconds on an A100 (new world record!) by tysam_and_co
What’s the trick you’re most proud of?
jobeta t1_j6gj0er wrote
jobeta t1_j69u8a4 wrote
Reply to [P] Launching my first ever open-source project and it might make your ChatGPT answers better by Vegetable-Skill-9700
Sweet! That dashboard looks really nice too!
jobeta t1_j62eibb wrote
IMHO the buzz is mainly around the UX provided by ChatGPT. Most LLMs are not that easily accessible and most people never get to experience any aha moment with them, so most people don't care. As for Google, I do think there is real but not immediate danger for their business model. The big issue for them is that 60% of their revenue comes from ads in Google search, so rolling out an amazing ChatGPT equivalent could potentially hurt their business. They would have to rethink the entire model. For now and AFAIK, ChatGPT doesn't provide web links so it doesn't feel like it is trying to sell you something. If Google if going to use one of their SOTA LLM and build a conversational AI out of it and make it available for free, surely they have to consider the implications for Alphabet as a whole.
jobeta t1_j0yw8sh wrote
Reply to Sarcasm Detection model [R]. by Business-Ad6451
New to this: Are there some labelled datasets for sarcasm?
jobeta t1_izyks32 wrote
Reply to comment by Clouwels in [D] ML to solve the division of people into teams by Clouwels
Here is my 2 cents:
They have a process, it is slow but it does the job. So start by saving them several hours and help automate their current process. This is valuable (saving manual labor) and will already be challenging: you will have to sit with them and understand how they do it. If it takes several hours, it is unlikely that they have a deterministic algorithm for it. To create one, you will likely have to have them make a number of decisions. You can probably help them make them. The outcome should be an algorithm, as simple as possible, that can perform the assignment of associates to teams.
Once you have solved this problem for them, you can think about ways to improve the assignment. But this open a very different can of worms. What does better mean? How do I measure that it is better? You will have to define some meaningful metrics (make sure they define them or that they definitely sign off on them) to be able to compare different assignment algorithms. Because you have so few teams, it will be pretty difficult to design a rigorous experiment that helps you determine if your new assignment algorithm beats the baseline. You can always come up with some fancy algorithm but how do you prove it works better? Some associates will say they don't like the new system some will like it. Who should we believe?Not to mention that you'll want to be able to track teams easily to be able to run some analytics. Chances are you'll have to build tracking for the teams. It might not be worth your time.
I've spent months trying to do things like this. The main challenge is that the Ops team wanted something better but never wanted to invest into defining or measuring what better was.
Alternatively you can keep adding simple constraints to your model that satisfies their intuition but that's not exactly Machine Learning so I would try to not get stuck in that position.
Good luck!
jobeta t1_izy9slt wrote
This is mostly a constrained optimization problem. It could benefit from ML if you need to predict some of the variables you’re optimizing for I guess? How many teams? How big are they? It’s hard to help you without details.
jobeta t1_iym2upg wrote
Reply to comment by Fun_Country_4193 in Create/finetune a model for $2,500 cash compensation [P] by Fun_Country_4193
Oh ok. I guess they have some costs on their end too. What did you mean by data from the pile? I’m happy to give it a shot if you think ~1 GB of data can be enough.
jobeta t1_iyl8mib wrote
Reply to comment by Fun_Country_4193 in Create/finetune a model for $2,500 cash compensation [P] by Fun_Country_4193
« Data from the pile »? Why don’t you organize a Kaggle challenge ?
jobeta t1_iyl87di wrote
What kind of data is it?
jobeta t1_ixum5yx wrote
Reply to comment by Background_Thanks604 in [D] Pytorch or TensorFlow for development and deployment? by CodaholicCorgi
How complex is the model you want to translate?
jobeta t1_iwlin8p wrote
Reply to comment by snairgit in [P] Thoughts on representing a Real world data by snairgit
I’m not sure what you mean by wrap that up. What programming language are you using? Python? You can create a pandas dataframe that contains one row per subject and per time stamp and as many columns as measurements. Then depending on the problem, you would transform that and engineer features that work with the type of model you’re thinking using.
jobeta t1_iwlciy8 wrote
Do you have one label for each set of position and speed measurements or are you trying to predict a label from sequences of measurements. Tbh this description is very limited and unclear.
jobeta t1_iw7228e wrote
Reply to comment by RichardBJ1 in Update an already trained neural network on new data by Thijs-vW
It seems intuitive that if possible, fully retraining will yield the best results but it can be costly. I just find it surprising to arbitrarily freeze two layers. What if your model only has two layers anyways? Again I don’t have experience so just guessing
jobeta t1_iw701lx wrote
Reply to comment by RichardBJ1 in Update an already trained neural network on new data by Thijs-vW
Why freeze bottom and top layers?
jobeta t1_jbt54u8 wrote
Reply to [Discussion] Compare OpenAI and SentenceTransformer Sentence Embeddings by Simusid
What are we looking at though? t-SNE?