Stable Code 3B: Coding on the Edge

  • Note that they don't compare with deepseek coder 6.7b, which is vastly superior to much bigger coding models. Surpassing codellama 7b is not that big of a deal today.

    The most impressive thing about these results is how good the 1.3B deepseek coder is.

  • Don’t entirely understand Stability’s business model. They’ve been putting out a lot of models recently and Stable Diffusion was novel at the time, but now their models consistently seem to be somewhat second rate compared to other things out there. For example Midjourney now seems to have far surpassed them in the image generation front. After raising a ton of funding Stability seems to just be throwing a bunch of stuff out there that’s OK but no longer ground breaking. What am I missing?

    Many other startups in the space will like face similar issues given the rapid commoditization of these models and the underlying tech. It’s very easy to spend a fortune building a model that offers a short lived incremental improvement at best before one can just quickly swap it out for something else someone else paid to train.

  • > License: Other

    > Commercial Applications

    > This model is included in our new Stability AI Membership. Visit our Membership page to take advantage of our commercial Core Model offerings, including SDXL Turbo & Stable Video Diffusion.

    what exactly is the license lol. can people use this or is this "see dont touch"

  • That is fantastic. I'm building a small macOS SwiftUI client with llama cpp built in, no server-client model, and it's already so useful with models like openhermes chat 7B, and fast.

    If this opens it to smaller laptops, wow!

    We truly live in crazy time. The rate of improvement in this field is off the walls.

  • I've got a machine with 4 3090s-- Anyone know which model would perform the best for programming? It's great this can run on a machine w/out a graphics card and is only 3B params, but I have the hardware. Might as well use it.

  • How are people using codellama and this in their workflows?

    I found one option: https://github.com/xNul/code-llama-for-vscode

    But I'm guessing there are others, and they might differ in how they provide context to the model.

  • Jargon naivete question: isn't "on the edge" normally implying on a server side with minimal routers hops to the client, not on client side?

  • I was able to run this model in http://lmstudio.ai as well. Just remove Compatibility Guess in Filters, so you can see all the models. LM Studio can load it and run requests against it.

  • I've been experimenting with code-llama extensively on my laptop, and from my experience, it seems that these models are still in their early stages. I primarily utilize them through a Web UI, where they can successfully refactor code given an existing snippet. However, it's worth noting that they cannot currently analyze entire codebases or packages, refining them based on the most suitable solutions using the most appropriate algorithms. While these models offer assistance to some extent, there is room for improvement in their ability to handle more complex and comprehensive coding scenarios.

  • How is this compared to the current GitHub Copilot?

  • FYI: This model is already available on Ollama.

  • Given the complete failure of the first stable lm, I'm interested to try this one out. Haven't really seen a small language model, except mixtral 7b that's really useful for much.

    I also hope stability comes out with a competitor to the new midjourney and dalle models! That's what put them on the map in the first place

  • It's amazing to see more smaller models being released. This creates opportunities for more developers to run it on their local computers, and makes it easier to fine-tune for specific needs.

  • Seems like they caught the Apple Marketing bug and are chasing things noonecares about. Great 3B model, everyone is already running 7B models over here.

    Maybe one day when I need to do offline coding on my cellphone, it will be really useful.

  • does anyone have recommendations for addins to integrate these 'smaller' llms into an IDE like VSCode? I'm pretty embedded with GH copilot, but curious to explore other options.

  • Can anyone explain what’s Stability’s business model (or plan for one)?

    I get why Meta releases tons of models, but still can’t quite understand what stability is trying to achieve

  • Why authors miss to compare with Phi-2?

  • How reliable are these benchmarks?

  • Terrible model

  • I just tried this model with Koboldcpp on my LLM box. I got gibberish back.

    My prompt - "please show me how to write a web scraper in Python"

    The response?

    <blockquote> I've written my first ever python script about 5 months ago and I really don't remember anything except for the fact that I used Selenium in order to scrape websites (in this case, Google). So you can probably just copy/paste all of these lines from your own Python code which contains logic to determine what value should be returned when called by another piece of software or program. </blockquote>

  • [flagged]

  • [flagged]

  • [flagged]

  • [flagged]

  • [flagged]

  • It's quite amazing - I often find that I read quite positive comments towards LLM tools for coding. Yet, an "Ask HN" I posted a while ago (and which admittedly didn't gain much traction) seemed to mirror mostly negative/pessimistic responses.

    https://news.ycombinator.com/item?id=38803836

    Was it just that my submission didn't find enough / more balanced commenters?