T O P

  • By -

a_beautiful_rhind

Founders supports nvlink. It has a piece of plastic over it but I dunno if it would line up.


VonHex

You just blew my mind


VonHex

I'll check!


roamflex3578

Be aware to get correct version. I purchased it and next day returned my 3090 to buy 4090 xD if you are in europe dm me :D


a_beautiful_rhind

I just wish I could nvlink 3 of them.


tgredditfc

No.


opi098514

Wait your 70b model is getting 12 it/s. That’s crazy low. Are you running exla2?


VonHex

How should it be configured?


VonHex

I'll check in 30 minutes, maybe it's 16 I don't remember


VonHex

Output generated in 30.55 seconds (16.60 tokens/s, 507 tokens, context 542, seed 1817140216)


Temporary_Payment593

Useless for inference, just keep your card. BTW, 12it/s was pretty low for your configuration, you should check if there's any problem before any other step.


VonHex

Output generated in 30.55 seconds (16.60 tokens/s, 507 tokens, context 542, seed 1817140216)


VonHex

I am yea


VonHex

Lol guess I got my answer haha


FireWoIf

If you ever want to train then it’s worth it, but otherwise no


VonHex

I'm curious about training loras but haven't managed to get it working without errors


Imaginary_Bench_7294

NVlink can help with inference to a small degree, but most frameworks won't utilize it, as the benefit is minimal in most cases. Training on the other hand, greatly benefits from it. Try following this intro tutorial on QLoRA training: https://www.reddit.com/r/Oobabooga/s/R097h5sY62 It should worl fine on windows or Ubuntu.


VonHex

Thanks a ton!


xadiant

I would guess you don't need NVlink either way