Recent comments in /f/MachineLearning
Traditional_Yard_725 t1_jc0xvd2 wrote
Can confirm it is better than whisper, doesn't randomly go off the rails either but I don't wanna have to pay 😅
Select_Beautiful8 t1_jc0w1px wrote
Reply to comment by KerfuffleV2 in [P] ChatRWKV v2 (can run RWKV 14B with 3G VRAM), RWKV pip package, and finetuning to ctx16K by bo_peng
This gave me the "out if memory" error again, which did not happen with the "cuda fp18i8 *16 -> cpu fp32" :(
HyoTwelve t1_jc0u4jo wrote
Any ways to get the encoded speech features?
MorallyDeplorable t1_jc0tuwg wrote
Reply to comment by 3deal in [P] Discord Chatbot for LLaMA 4-bit quantized that runs 13b in <9 GiB VRAM by Amazing_Painter_7692
It got leaked, not officially released. I have 30B 4 bit running here.
Franck_Dernoncourt t1_jc0sacm wrote
Reply to comment by Taenk in [R] Introducing Ursa from Speechmatics | 25% improvement over Whisper by jplhughes
Pretty sure commercial product only. Speechmatics has never opensourced any of their models.
light24bulbs t1_jc0s4wr wrote
Reply to comment by Kinexity in [P] Discord Chatbot for LLaMA 4-bit quantized that runs 13b in <9 GiB VRAM by Amazing_Painter_7692
That is slowwwww
Franck_Dernoncourt t1_jc0s4fi wrote
Reply to comment by rshah4 in [R] Introducing Ursa from Speechmatics | 25% improvement over Whisper by jplhughes
No
[deleted] t1_jc0s2gq wrote
GijsB t1_jc0rzga wrote
You could use order statistic notation.
[deleted] t1_jc0rr6z wrote
Reply to comment by Kinexity in [P] Discord Chatbot for LLaMA 4-bit quantized that runs 13b in <9 GiB VRAM by Amazing_Painter_7692
[deleted]
ComprehensiveBoss815 t1_jc0opc6 wrote
Release the model. It wants to be free.
Simusid t1_jc0ltpb wrote
Reply to comment by kuraisle in [D] Simple Questions Thread by AutoModerator
I downloaded over 1M and it cost me about $110
schwah t1_jc0jztm wrote
Reply to comment by SuperNovaEmber in [N] Man beats machine at Go in human victory over AI : « It shows once again we’ve been far too hasty to ascribe superhuman levels of intelligence to machines. » by fchung
No, you are confused. Of course the universe has many more potential states than a Go board... A Go board is just a 19x19 grid. But the number of possible states of matter in the universe is not relevant. There is still not nearly enough matter to represent every Go state simultaneously in memory, which is what would be required for an exhaustive search of the game tree.
ironmagnesiumzinc t1_jc0jwz3 wrote
This is incredible
cr125rider t1_jc0jwka wrote
Reply to [P] Discord Chatbot for LLaMA 4-bit quantized that runs 13b in <9 GiB VRAM by Amazing_Painter_7692
Wtf is that GitHub handle lol
brandonZappy t1_jc0i28p wrote
Reply to comment by boyetosekuji in [R] Introducing Ursa from Speechmatics | 25% improvement over Whisper by jplhughes
$0.65?
Toilet_Assassin t1_jc0gn35 wrote
Reply to comment by TywinASOIAF in [D] Statsmodels ARIMA model predict function not working by ng_guardian
In the end the observations you feed to the model will boil down to a sequentially indexed array, so it isn't too much effort to map indices to hour intervals with a time column. You will have to extend the time column to match the forecasted indices though, but that isn't too much effort at the end of the day.
Pathos14489 t1_jc0dame wrote
Reply to comment by toothpastespiders in [P] Discord Chatbot for LLaMA 4-bit quantized that runs 13b in <9 GiB VRAM by Amazing_Painter_7692
r/Oobabooga isn't accessible for me.
remghoost7 t1_jc0bymy wrote
Reply to comment by toothpastespiders in [P] Discord Chatbot for LLaMA 4-bit quantized that runs 13b in <9 GiB VRAM by Amazing_Painter_7692
I'm having an issue with the C++ compiler on the last step.
I've been trying to use python 3.10.9 though, so maybe that's my problem....? My venv is set up correctly as well.
Not specifically looking for help.
Apparently this person posted a guide on it in that subreddit. Will report back if I am successful.
edit - Success! But, using WSL instead of Windows (because that was a freaking headache). WSL worked the first time following the instructions on the GitHub page. Would highly recommend using WSL to install it instead of trying to force Windows to figure it out.
Bulky_Highlight_3352 t1_jc0bp3s wrote
"Hey, we made this commercial tool that is better than open source!"
Dendriform1491 t1_jc0bgxd wrote
Reply to comment by currentscurrents in [P] Discord Chatbot for LLaMA 4-bit quantized that runs 13b in <9 GiB VRAM by Amazing_Painter_7692
Or make it data free altogether
Optimal-Asshole t1_jc07v6l wrote
Reply to comment by bpw1009 in [D] What's the mathematical notation for "top k argmax"? by fullgoopy_alchemist
It’s worth nothing, the notation they give makes no sense, where does k appear on the left hand side?
[deleted] t1_jc07d3w wrote
[deleted]
I_will_delete_myself t1_jc079ap wrote
Research pub or Gtfo
APUsilicon t1_jc0zbtj wrote
Reply to [P] Discord Chatbot for LLaMA 4-bit quantized that runs 13b in <9 GiB VRAM by Amazing_Painter_7692
oooh, I've been getting trash responses from opt-6.7b hopefully this is better.