Recent comments in /f/MachineLearning
kuraisle t1_jbyrulz wrote
Reply to [D] Simple Questions Thread by AutoModerator
Has anyone had any experience data mining BioArXiv? It's on a requester pays Amazon s3 bucket, which isn't something I've used before and I'm struggling to guess how much I would have to pay to retrieve a few thousand articles. Thanks!
ertgbnm t1_jbyocgi wrote
Reply to comment by serge_cell in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
Isn't alphago trained against itself? So I would consider it adversarial training.
SuperNovaEmber t1_jbymjai wrote
Reply to comment by schwah in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
Oh dear. You miss the most important. Which I did not mention. I figured you know?
Every empty point in space is theoretically capable of storing an atom, give or take.
Most of space is empty. The calculation of all the observable atoms 'brute forced' into all possible voids?
That's really the point, friend. You're talking combinatorics of one thing and then falsely equalizing with simply number of atoms?? Not the possible combinations of those atoms?? Which far exceeds the visible signs of your awakening?? It's not even astronomically close, bud.
In theory, a device around the size of a deck of cards contains more than enough energy to compute to end game.
The "observable" universe operates at an insanely high frequency. Consider the edge of the universe is over 10 orders of magnitude closer than the Planck length, using meters of course.
We're 10 billion times closer to the edge of the universe than the fabric of reality.
WesternLettuce0 t1_jbyl2wi wrote
Reply to [D] Simple Questions Thread by AutoModerator
I used distilbert and legalbert separately to produce embeddings for my documents. What is the best way to use the embeddings for classification? Do I create document level embeddings before training my classifiers? Do I combine the two embeddings?
Thog78 t1_jbyh4w1 wrote
Reply to comment by onkus in [Discussion] Compare OpenAI and SentenceTransformer Sentence Embeddings by Simusid
What you're looking for when comparing UMAPs is if the local relationships are the same. Try to recognize clusters and see their neighbors, or whether they are distinct or not. A much finer colored clustering based on another reduction (typically PCA) helps with that. Without clustering, you can only try to recognize landmarks from their size and shape.
AntelopeEntire9191 t1_jbye5q4 wrote
Reply to [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
bro. progress is still progress
wikipedia_answer_bot t1_jbybzoo wrote
Reply to comment by tdgros in [D] Simple Questions Thread by AutoModerator
baby don't hurt me
This comment was left automatically (by a bot). If I don't get this right, don't get mad at me, I'm still learning!
^(opt out) ^(|) ^(delete) ^(|) ^(report/suggest) ^(|) ^(GitHub)
tdgros t1_jbybye9 wrote
Reply to comment by Egg_Stealer in [D] Simple Questions Thread by AutoModerator
what is love?
[deleted] t1_jbyaq18 wrote
Reply to comment by onkus in [Discussion] Compare OpenAI and SentenceTransformer Sentence Embeddings by Simusid
[deleted]
Egg_Stealer t1_jby8tfl wrote
Reply to [D] Simple Questions Thread by AutoModerator
Can machines learn to love
TywinASOIAF t1_jby3znl wrote
Python statmodels is bad. It cannot handle hour data for example. Use R if you do time series.
schwah t1_jby3w3r wrote
Reply to comment by SuperNovaEmber in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
Okay fair enough, it's not as simple as 10^170 > 10^80.
But I don't think your math makes much sense either. You can't just count the number of isotopes - nearly all of the universe is hydrogen and helium. And even with compression, it is going to take a lot more than 1 bit to represent a board state. Memory (at least with todays technology) requires billions of atoms per bit. And that is only memory - the computational substrate is also needed. And obviously we are ignoring some pretty significant engineering challenges of a computer that size, like how to deal with gravitational collapse.
I'll grant that it's potentially possible that you could brute force Go with a Universe-o-tron (if you ignore the practical limitations of physics), but it's definitely not a slam dunk like you're implying.
shahaff32 t1_jbxy6vj wrote
Reply to comment by ghostfuckbuddy in [D] Is Pytorch Lightning + Wandb a good combination for research? by gokulPRO
It might be a bug, and it might also be the case that we are the "idiot users" that don't get how to use the package correctly.
Either way, we decided to continue without Lightning (because of that reason and a few others), and therefore we need the code in pure PyTorch for future works that rely on it.
OptimizedGarbage t1_jbxticv wrote
Reply to comment by serge_cell in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
It kinda was though? It was trained using self-play, so the agent it was playing against was adversarially searching for exploitable weaknesses. They actually cite this as one of the reasons for it's success in the paper
blablanonymous t1_jbxh137 wrote
suflaj t1_jbx9h57 wrote
Reply to comment by Excellent_Dirt_7504 in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
But there is evidence of a defense by taking as many adversarial attacks as possible and training against them. Ultimately, the ultimate defense is generalization. We know it exists, we know it is achievable, we only don't know HOW it's achievable (for non-trivial problems).
Quick-Hovercraft-997 t1_jbx9gcj wrote
Reply to [D] Best way to run LLMs in the cloud? by QTQRQD
if latency is not a critical requirement, you can try serverless GPU cloud like banana.dev, pipeline.ai . These platform provide an easy to use template for deploying LLM.
mfarahmand98 t1_jbx1hgt wrote
Reply to [P] Introducing confidenceinterval, the long missing python library for computing confidence intervals by jacobgil
Awesome library! Been looking for something like this.
ID4gotten t1_jbwz98k wrote
Reply to comment by megacewl in [P] vanilla-llama an hackable plain-pytorch implementation of LLaMA that can be run on any system (if you have enough resources) by poppear
Could the 4 bit (edit: or 8 bit) version of the 65B model parameters be run on vanilla-llama to use less than 100GB VRAM?
SuperNovaEmber t1_jbwxgu5 wrote
Reply to comment by schwah in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
Wow, that understanding is deeply flawed. In computer systems we have compression and instancing and other tricks. But that's all besides the following point.
Atoms, for instance, how many different types are possible? Lets even ignore isotopes!
It's just like calculating a base. Like a byte can have 256 values. You get 4 bytes(32 bits) together and that's 4.3 billion states or 256^4 (base 256 with 4 bytes) or 2^32 (binary, base 2 with 32 bits). So instead of 256 values we got 118 unique atoms and instead of bytes we got atoms, 10^80 of them.
Simple, right? 118^10^80 combinations possible. Highest exponent first, mind you. Otherwise you only will get 1,658 digits instead of the actual result.... Which is not even remotely close..... Not 80 digits. Not 170 digits. Not 1,658, even.
That's 207188200730612538547439527925963726569493435639287375683771302641055893615162425 digits..... Again. This is not the answer. Just the number of digits in the answer.
Universe gots zero problems computing GO, bro
That's nothing compared to all the possible spaces all the possible atoms could occupy over all extents over space(and)time.
That's a calculation I'll leave up to you downvoters, gl hf!
Excellent_Dirt_7504 t1_jbwwi8v wrote
Reply to comment by serge_cell in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
If you train against one attack, you remain vulnerable to another. There is no evidence of a defense that is robust to any adversarial attack.
ApparatusCerebri t1_jbwwh5j wrote
Reply to comment by currentscurrents in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
Our visual system does use a couple of neat tricks to process what's around us but that too is open to some edge cases hence optical illusions. Other than that, in our case, evolution is the mother of all adversarial training :D
JClub t1_jbwu3lx wrote
Reply to comment by Non-jabroni_redditor in [Discussion] Compare OpenAI and SentenceTransformer Sentence Embeddings by Simusid
more than that, GPT is unidirectional, which is really not great a sentence embedder
Lawrencelot t1_jbyvfwq wrote
Reply to [D] What's the mathematical notation for "top k argmax"? by fullgoopy_alchemist
I don't think there is. You could use the term top k or write pseudocode.