Skip to main content

Get the Reddit app

Scan this QR code to download the app now
Or check it out in the app stores

r/LocalLLaMA

members
online


Hello! Bonjour! Ciao! Now you're speaking their language thanks to Interpreter with Galaxy AI. Talk the talk with real-time translations you both can see when you pre-order the new Galaxy Z Flip6 and get double the storage on us.
media poster


New Model CodeGeeX4-ALL-9B has been released. New Model CodeGeeX4-ALL-9B has been released.
Discussion

We introduce CodeGeeX4-ALL-9B, the open-source version of the latest CodeGeeX4 model series. It is a multilingual code generation model continually trained on the GLM-4-9B, significantly enhancing its code generation capabilities. Using a single CodeGeeX4-ALL-9B model, it can support comprehensive functions such as code completion and generation, code interpreter, web search, function call, repository-level code Q&A, covering various scenarios of software development. CodeGeeX4-ALL-9B has achieved highly competitive performance on public benchmarks, such as BigCodeBench and NaturalCodeBench. It is currently the most powerful code generation model with less than 10B parameters, even surpassing much larger general-purpose models, achieving the best balance in terms of inference speed and model performance.

We have been developing a Local AI to create a decentralized developer kit (DDK). We've alternated between using CodeQwen for its speed, larger context window, and DeepSeek for its higher accuracy but very limited context window, which is crucial for code generation within the existing code context.

But wow! We just tested CodeGeeX4-ALL-9B (Quantized 3 bit). It accepts more than 1500 tokens as context, evaluates prompts at 172.99 tokens/s, and generates more than 25 tokens/second on my M1 machine while consuming 4.5 GB of RAM.

The future of all personal AI is local, not cloud-based. We all have the machines to make that possible; we just need to look around!

https://huggingface.co/THUDM/codegeex4-all-9b/tree/main

GGUF  https://huggingface.co/bartowski/codegeex4-all-9b-GGUF
If you are facing issues with llama.cpp, just update it to the latest release.