NHacker Next
  • new
  • past
  • show
  • ask
  • show
  • jobs
  • submit
Lemonade by AMD: a fast and open source local LLM server using GPU and NPU (lemonade-server.ai)
moconnor 2 minutes ago [-]
Is... is this named because they have a lemon they're trying to make the most of?
JSR_FDED 13 minutes ago [-]
I’ve read the website and the news announcement, and I still don’t understand what it is. An alternative to LM Studio? Does it support MLX or metal on Macs? I’m assuming it will optimize things for AMD, but are you at a disadvantage using other GPUs?
zozbot234 12 minutes ago [-]
Note that the NPU models/kernels this uses are proprietary and not available as open source. It would be nice to develop more open support for this hardware.
swiftcoder 8 minutes ago [-]
Are they? The docs say "You can also register any Hugging Face model into your Lemonade Server with the advanced pull command options"
jmillikin 17 minutes ago [-]
Surprising that the Linux setup instructions for the server component don't include Docker/Podman as an option, its Snap/PPA for Ubuntu and RPM for Fedora.

Maybe the assumption is that container-oriented users can build their own if given native packages?

freedomben 11 minutes ago [-]
They do have some container options, though I definitely think they should be added to the release page: https://lemonade-server.ai/install_options.html#docker
freedomben 13 minutes ago [-]
Neat, they have rpm, deb, and a companion AppImage desktop app[1]! Surprised I wasn't aware of this project before. Definitely going to give it a try.

[1]: https://github.com/lemonade-sdk/lemonade/releases/tag/v10.0....

syntaxing 24 minutes ago [-]
Wow this is super interesting. This creates a local “Gemini” front end and all. This is more or less a generative AI aggregator where it installs multiple services for different gen modes. I’m excited to try this out on my strix halo. The biggest issue I had is image and audio gen so this seems like a great option.
nijave 29 minutes ago [-]
Anyone compare to ollama? I had good success with latest ollama with ROCm 7.4 on 9070 XT a few days ago
iugtmkbdfil834 24 minutes ago [-]
Seconded. Currently on ollama for local inference, but I am curious how it compares.
9dc 25 minutes ago [-]
so... what does it do? i dont get it Lol
iugtmkbdfil834 20 minutes ago [-]
Initial read suggests it is a mini-swiss army knife, because it seems to be able to do a lot ( based on website claims anyway ). The app integration seems to suggest they want to be more of a control dashboard.
Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact
Rendered at 12:59:49 GMT+0000 (Coordinated Universal Time) with Vercel.