Recent comments in /f/MachineLearning
PhysZhongli t1_jccbh4w wrote
Reply to [D] Simple Questions Thread by AutoModerator
Hi everyone, I am a novice trying to learn ML and AI. I am trying to train a CNN model to classify 9000+ images with 100 labels. These images are flower patterns/leaves from what I can tell. The catch is that the actual test dataset has 101 labels and the when the model detects an image not in the original 100 labels it has to assign it to the 101st label. What would be the best way to go about doing this?
I have used resnet50 with imagenet weights and made some of the previous layers trainable to fine tune the model. I have followed it with a globalaverage layer, a 1024 node dense layer with l2 regularization, batchnorm, dropout and softmax layer as the classifer. I am using adam optimizer with a batch size of 16, learning rate of 0.0001. I then assign a threshold value of 0.6 and if the model prediction is below the threshold value it assigns it the 101st label. Currently i have a ~90% testing accuracy.
Are there any obvious things i should be doing better/changing and how can i go about optimising the threshold value or is there a better way to handle the 101st label? Should i be using resnet or something else for flower patterns and leaves given my training dataset of 9000+ images
KerfuffleV2 t1_jccb5v1 wrote
Reply to comment by bo_peng in [P] ChatRWKV v2 (can run RWKV 14B with 3G VRAM), RWKV pip package, and finetuning to ctx16K by bo_peng
Sounds good! The 4bit stuff seems pretty exciting too.
By the way, not sure if you saw it but it looks like PyTorch 2.0 is close to being released: https://www.reddit.com/r/MachineLearning/comments/11s58n4/n_pytorch_20_our_next_generation_release_that_is/
They seem to be claiming you can just drop in torch.compile() and see benefits with no code changes.
Designer_Border_9705 t1_jccau1r wrote
It is thriving.
nopainnogain5 OP t1_jcc9his wrote
Reply to comment by haljm in [D] To those of you who quit machine learning, what do you do now? by nopainnogain5
In case I'd like to dive into something along these lines, how such positions tend to be called?
nopainnogain5 OP t1_jcc8xe3 wrote
Reply to comment by chef1957 in [D] To those of you who quit machine learning, what do you do now? by nopainnogain5
How would you call your position? Is it software engineering / data engineering or yet something else?
nopainnogain5 OP t1_jcc8rje wrote
Reply to comment by I_will_delete_myself in [D] To those of you who quit machine learning, what do you do now? by nopainnogain5
Do you mean roles such as data engineer / MLOps?
thecity2 t1_jcc8crf wrote
Reply to [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
OpenAI is very doublespeak.
thecity2 t1_jcc8549 wrote
Reply to comment by Chuyito in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Yes to point 1! Not enough people are talking about this aspect. The data wars are on imo. How will Google protect their mountains of video data for example.
[deleted] OP t1_jcc8123 wrote
thecity2 t1_jcc80ex wrote
Reply to comment by -Rizhiy- in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
They already said they are for profit entirely now.
nopainnogain5 OP t1_jcc7tgd wrote
Reply to comment by Senior_Buy445 in [D] To those of you who quit machine learning, what do you do now? by nopainnogain5
The model is ready, I'm here to collect data ;)
nopainnogain5 OP t1_jcc7l67 wrote
Reply to comment by Mikkelisk in [D] To those of you who quit machine learning, what do you do now? by nopainnogain5
Haha not sure if this is that comforting
Purplekeyboard t1_jcc7cuo wrote
Reply to comment by ScientiaEtVeritas in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
But without OpenAI, who would have spent the billions of dollars they have burned through creating and then actually giving people access to models like GPT-3 and now GPT-4?
You can use GPT-3, and even versions of GPT-4, today. Or you can stand and look up at the fortress of solitude that is Google's secret mountain lair where models are created and then hoarded forever.
Senior_Buy445 t1_jcc5jio wrote
I know exactly what you are talking about. Perhaps you could train a model to tell you what to do next? We have folks all the time believing that ML/DL will solve any problem… :-)
nightshadew t1_jcc56x4 wrote
You can do a lot in other subfields. Even basic churn predictions are something valuable for a lot of firms. Are you sure you’re not swept by the DL hype yourself?
Mikkelisk t1_jcc4spe wrote
> Coming up with architectures that randomly work/don't work, tuningparameters, waiting for days till the model is trained... the level ofuncertainty is just too high for me.
Good news! You say you work in computer vision. There's a high chance that in practice you'll mostly use off-the-shelf solutions and most of your actual time will be spent gathering data:)
VelveteenAmbush t1_jcc4mvf wrote
Reply to comment by Nhabls in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Transformers aren't products, they're technology. Search, Maps, Ads, Translation, etc. -- those were the products. Those products had their own business models and competitive moats that had nothing to do with the technical details of the transformer.
Whereas GPT-4 is the product. Access to it is what OpenAI is selling, and its proprietary technology is the only thing that prevents others from commoditizing it. They'd be crazy to open up those secrets.
[deleted] OP t1_jcc3o14 wrote
[deleted] OP t1_jcc3bg1 wrote
Nhabls t1_jcc30fs wrote
Reply to comment by VelveteenAmbush in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
The original transformers (ie the foundational model architecture all GPTs are based on) were also commercial products (they're used for search, summarization, translation,etc) we got them and the paper all the same.
Nhabls t1_jcc2tg0 wrote
Reply to comment by crt09 in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
It's written right after that
>Second, the instruction data is based on OpenAI’s text-davinci-003, whose terms of use prohibit developing models that compete with OpenAI
[deleted] OP t1_jcc1d8w wrote
mietminderung t1_jcc0m1l wrote
Reply to [D] Simple Questions Thread by AutoModerator
What's the place, if any, to post a job opening?
bo_peng OP t1_jccc46c wrote
Reply to comment by KerfuffleV2 in [P] ChatRWKV v2 (can run RWKV 14B with 3G VRAM), RWKV pip package, and finetuning to ctx16K by bo_peng
I am using torch JIT so close ;)