The model does have a lot of advantages over sdxl with the right prompting, but it seems to fall apart in prompts with more complex anatomy. Hopefully the community can fix it up once we have working trainers.
The model does have a lot of advantages over sdxl with the right prompting, but it seems to fall apart in prompts with more complex anatomy. Hopefully the community can fix it up once we have working trainers.
“Tiny shards” probably isn’t the right term to describe particles 20-200 nanometers wide, but this is probably bad nonetheless.
The names missing from the list say more about the board’s purpose than the names on it.
I assumed this was always the case
The issue is that they have no way of verifying that. We’d have to trust 2 other companies in addition to DDG.
All of Firefox’s ai initiatives including translation and chat are completely local. They have no impact on privacy.
The “why would they make this” people don’t understand how important this type of research is. It’s important to show what’s possible so that we can be ready for it. There are many bad actors already pursuing similar tools if they don’t have them already. The worst case is being blindsided by something not seen before.
The 8B is incredible for it’s size and they’ve managed to do sane refusal training this time for the official instruct.
The rest of the budget kind of sucks but this part makes sense. If you’re making significant profits off of users in a country you should have to pay some of that back. All countries should have this.
Cohere’s command-r models are trained for exactly this type of task. The real struggle is finding a way to feed relevant sources into the model. There are plenty of projects that have attempted it but few can do more than pulling the first few search results.
They’re already lying to get passed the 13 year requirement so I doubt it would make any difference.
I don’t think the term open-source can be applied to model weights. Even if you have the exact data, config, trainer and cluster it’s basically impossible to reproduce an exact model. Calling a model “open” sort of works but then there’s the distinction between open for research and open for commercial use. I think it’s kind of similar to the “free” software distinction. Maybe there’s some Latin word we could use.
It’s an AI thing. Nearly all small models struggle with separating multiple characters.
I’m sure the machine running it was quite warm actually.
Your best bet would probably be to get a used office PC to put the card in. You’ll likely have to replace the power supply and maybe swap the storage but with how much proper external enclosures go for the price might not be too different. Some frameworks don’t support direct GPU loading so make sure that you have more ram than vram.
An arm soc won’t work in most cases due to a lack of bandwidth and software support. The only board I know of that can do it is the rpi5 and that’s still mostly a poc.
In general I wouldn’t recomend a titan x unless you already have one because it’s been deprecated in cuda, so getting modern libraries to work will be a pain.
I really like the simplicity and formatting of stock pacman. It’s not super colorful but it’s fast and gives you all of the info you need. yay (or paru if you’re a hipster) is the icing on top.
Partnered with Adobe research so we’re never going to get the actual model.
This has more to do with how much chess data was fed into the model than any kind of reasoning ability. A 50M model can learn to play at 1500 elo with enough training: https://adamkarvonen.github.io/machine_learning/2024/01/03/chess-world-models.html
It does a little bit worse than v0.1 on all benchmarks which isn’t ideal. That doesn’t really say much about the finetuning potential though.
What’s the deal with Alpine not using GNU? Is it a technical or ideological thing? Or is it another “because we can” type distro?