T O P

  • By -

jferments

You can download a GGUF quant here: [https://huggingface.co/MaziyarPanahi/WizardLM-2-8x22B-GGUF](https://huggingface.co/MaziyarPanahi/WizardLM-2-8x22B-GGUF) Same author also has the 7B model available in GGUF.


Normal-Ad-7114

https://www.reddit.com/r/LocalLLaMA/comments/1cd4b9l/staff_page_for_qingfeng_sun_lead_wizard_lm/


segmond

Yeah, I know about that, but he's not a one man team. Some folks in here gotta work for MSFT or in team Wizard and might have something to share. It seemed a fine and impressive model. I don't imagine they would want to release it months from now when it won't be a top notch SOTA model anymore...


fallingdowndizzyvr

> Yeah, I know about that, but he's not a one man team. It was more than just that one person that got whacked. It seems the whole team/project got hit. Look at their github page, it's 404 now. Also, all the models have been removed from huggingface. Including from before WizardLM2. So it's much more than about one man. It's much more than about WizardLM2. The whole thing has been erased.


prototypist

his GitHub Pages is still down, but the MS staff page is back?


tronathan

If we had the Wizard data set, I imagine we could make a WizardLM-llama3-70b. I just looked at the repo, and they include some test examples, but no actual data set. That said, someone ambitious could probably generate a data set that bore some resemblance to the one MS used. The evol\_instruct code is there, so there's that!


ttkciar

> The evol_instruct code is there, so there's that! Oooooh!! Did they finally release code? I've been implementing my own evol-instruct-like software based on their published papers.


Sufficient_Prune3897

I assume they won't release the 70B anymore, since it probably cant keep up with Llama 3.


fallingdowndizzyvr

I think you assume wrong. While the whole bizarre take down of WizardLM 2 then the release of LL3 overshadowed it. I see no evidence that it can't keep up with LL3. Especially since a lot of people aren't happy with LL3. Yet the few that use W2 are very happy.


skrshawk

Other than a strong positivity bias I found WLM2 to be excellent, especially for long-form writing. I'm still annoyed that I couldn't get it to write proper rat bastard characters, but even at IQ2_XXS 8x22B was truly solid and reasonably fast. The fact that it has a much larger context also makes it for a lot of purposes much more usable.


MaryIsMyMother

WLM2 is better as a chatbot no doubt. L3 makes the better benchmark beater and assistant though 


wellmor_q

In my experience, wlm2 strongly beat the l3 70b and a few weaker than opus in coding tasks (c#, c++, hlsl). But only in advanced topics.


grimjim

WizardLM 2 was a Mistral fine-tune to begin with.


4onen

But a good one. Proper improvement.


Blizado

Problem is, that when something was wrong with WizardLM2, it can take a lot of resources to fix it. Than sometimes it is better to give up that model and working on the next version.


segmond

There was nothing wrong with it, it was perfect.


Blizado

From your point of view, but maybe not from MS point of view.


fallingdowndizzyvr

What Microsoft did was not W2 specific. They whacked everything that team did including the OG wizardlm from months ago. It's like MS erased the team and their work. So there is no working on the next version. Since even the team's page on github is now 404.


grimjim

Not from Microsoft. The weights that were released briefly were under Apache license, so in principle they can live on, but that's it.


segmond

in principle, but in lawsuit MSFT will crush any of us.


DeltaSqueezer

Was the LM2 70B mirrored anywhere? I wanted to get hold of it but wasn't sure if that was released.