yes, simply rescale proportionally.
Learn Machine Learning
Welcome! This is a place for people to learn more about machine learning techniques, discuss applications and ask questions.
Example questions:
- "Should I use a deep neural network for my audio classification task?"
- "I'm working with a small dataset, what can I do to make my model generalize well?"
- "Is there a library available that implements function X in language Y?"
- "I want to learn more about the math behind machine learning technique A, where should I start?"
Please do:
- Be kind to new people
- Post guides and tutorials that you find helpful
- Link to open/free sources instead of paywalled when possible
Please don't:
- Post news articles / memes (there are other machine learning/AI communities for this)
Other communities in this area:
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
Similar subreddits: r/MLquestions, r/askmachinelearning, r/learnmachinelearning
Wow, you're quick! Beat me to reposting the old answer
ty for your work! hope you don't mind me reposting these to my instance. happy to go both ways too!
Ya, go ahead. Atm I'm just pulling old answers that I've deleted from reddit or cross validated. I'll probably do some underappreciated/hard to find answers as well
this is the real work we need right now, one of the biggest complaints is the searchability of usable content on reddit (lemmy does not have much), we know the answer to that.
I have a list of AI/ML folks I am following an posting fresh research, we are using bots to summarize the work as well. Feel free to repost with abandon as well.
I think more of the traffic comes from google search results (that's what people say they use) and there's not much that can be changed there.
Imo there is too much noise in the ML research publishing atm, certainly with deep learning (throw an NN at something and it seems like you can get anything published nowadays) and the papers aren't very helpful (they rarely have enough info to be reproducible) but I digress. Plus paywalls can be a problem. If you want the more solid stuff for deep learning, I would suggest something like: paperswithcode.com/ But I spend most of my time sorting through papers so not an unbiased opinion at all
Feels, integrated AI started as a shared google doc of links for some projects i was on, realized I had figured out how to wadge through papers and get usable stuff still (used to do similar for industrial robotics years ago).
My hope is as our domains age we can play SEO games and get our instances in the results. Make them look at us on every page!
Original answer:
Scaling the dataset before passing it to the autoencoder is usually how I do it, you don't need to rescale after if you are only using the encoder portion (for example for dimensionality reduction). If you don't do it linearly (aka (x-min(x))/(max(x)-min(x) ) and use exp or log to do it then be mindful that it would likely have an impact with respect to loss/optimization behaviour.
Make sure to take the max and min values from the training data then apply it to the testing (in the case of values out of bounds, set them to the boundary value but this shouldn't have a big impact if your training dataset is large enough with enough variance).