Search
Results
Who needs GitHub Copilot when you can roll your own AI code assistant at home • The Register
(2) Are there any tiny (1-3b) models finetuned for coding available in GGUF format? : LocalLLaMA
bigcode (BigCode)
Research community developing various code models, small and big. Models may not be instruct
WizardLM (WizardLM)
deepseek-ai (DeepSeek)
They have the 1.3B version!!! This may be the best to start with Newspeak. Should work train even on huggingcface
deepseek-ai/deepseek-coder-6.7b-instruct · Hugging Face
Another possible model. For coding capabilities, Deepseek Coder achieves state-of-the-art performance among open-source code models on multiple programming languages and various benchmarks.
LLaMA 7B GPU Memory Requirement - Transformers - Hugging Face Forums
With the optimizers of bitsandbytes (like 8 bit AdamW), you would need 2 bytes per parameter, or 14 GB of GPU memory.
stabilityai/stable-code-3b · Hugging Face
Another potential model to use for Newspeak, but it is NOT open source. Adventage: 2.5B params, so should be usable in small GPUs
Can Ai Code Results - a Hugging Face Space by mike-ravkine
Comparison of LLM models for coding
openchat/openchat-3.5-0106 · Hugging Face
Open source with lots of information. Uses Multiple undrelying models. Not sure how I would train for it
Welcome Mixtral - a SOTA Mixture of Experts on Hugging Face
The Mixtral model is new, and seems to be good. Click on “Demo“ to test it
StarCoder: A State-of-the-Art LLM for Code
Article has comparison with other code-LLM models
huybery/Awesome-Code-LLM: An awesome and curated list of best code-LLM for research.
Large language models and the rise of the AI code generators | InfoWorld
Review of LLM specialized for code generation
OpenAI Codex - Wikipedia
Model which generates code for Python, Javascript, Go, Shell, Perl, Swifg, Ruby, PHP
codellama (Code Llama) - Huggingface model for generating programs. Maybe can be used for Newspeak?
AI Code Tools: The Ultimate Guide in 2024
AI Code tools : Good summary. Does not talk about which pre-trained model they use. One is gemini (bard) -> alphacode2
BigCode - Playground - a Hugging Face Space by bigcode
Look for models that could be used in Newspeak