Search
Results
ทำไมต้องกำหนดความสามารถหลักองค์กร (Core competency) – Baramizi
Competency Modeling - Assessment Associates International
Competency
สมรรถนะ (Competency Model) มีอะไรบ้าง ? - Oriental Phoenix
13 Best Open Source ChatGPT Alternatives
Mit wenigen Klicks zum eigenen KI-Chatbot: Warum ihr dieses Tool kennen solltet
Train and use my model
Who needs GitHub Copilot when you can roll your own AI code assistant at home • The Register
lyogavin/godmodeanimation: 2D Game Animation in God Mode
OpenAI’s GPT-4o Mini isn’t much better than rival LLMs • The Register
Honey, I shrunk the LLM! A beginner's guide to quantization • The Register
abi/secret-llama: Fully private LLM chatbot that runs entirely with a browser with no server needed. Supports Mistral and LLama 3.
What is a Model of Reflection?
Input Output Outcome: Use the misunderstood Outcome Model successfully
Product, Process, Resource – an Integrated Modeling Approach for Production Engineering and Industrialized Construction - Project Production InstituteProject Production Institute
Optimum
Optimum is an extension of Transformers that provides a set of performance optimization tools to train and run models on targeted hardware with maximum efficiency. It is also the repository of small, mini, tiny models.
google-research/bert: TensorFlow code and pre-trained models for BERT
google/bert_uncased_L-4_H-256_A-4 · Hugging Face
Repository of all Bert models, including small. Start using this model for testing.
A Step-by-Step Guide to Model Evaluation in Python | by Shreya Singh | Medium
Open LLM Leaderboard - a Hugging Face Space by HuggingFaceH4
Comparison of efficiency of all LLM models on hugging face
Training Bert on Yelp - Copy of training.ipynb - Colaboratory
(1) Most cost effective GPU for local LLMs? : LocalLLaMA
GGML quantized models. They would let you leverage CPU and system RAM, instead of having to rely on a GPU’s. This could save you a fortune, especially if go for some used AMD Epyc platforms. This could be more viable for the larger models, especially the 30B/65B parameters models which would still press or exceed the VRAM on the P40.
Optimizing LLMs for Speed and Memory
7 steps to master large language models (LLMs) | Data Science Dojo
LLM for a new language : MachineLearning
High level how to train a model
Up to date List of LLM Models
(2) Are there any tiny (1-3b) models finetuned for coding available in GGUF format? : LocalLLaMA
bigcode (BigCode)
Research community developing various code models, small and big. Models may not be instruct
WizardLM (WizardLM)
deepseek-ai (DeepSeek)
They have the 1.3B version!!! This may be the best to start with Newspeak. Should work train even on huggingcface
deepseek-ai/deepseek-coder-6.7b-instruct · Hugging Face
Another possible model. For coding capabilities, Deepseek Coder achieves state-of-the-art performance among open-source code models on multiple programming languages and various benchmarks.
LLaMA 7B GPU Memory Requirement - Transformers - Hugging Face Forums
With the optimizers of bitsandbytes (like 8 bit AdamW), you would need 2 bytes per parameter, or 14 GB of GPU memory.
stabilityai/stable-code-3b · Hugging Face
Another potential model to use for Newspeak, but it is NOT open source. Adventage: 2.5B params, so should be usable in small GPUs
Can Ai Code Results - a Hugging Face Space by mike-ravkine
Comparison of LLM models for coding
openchat/openchat-3.5-0106 · Hugging Face
Open source with lots of information. Uses Multiple undrelying models. Not sure how I would train for it
Welcome Mixtral - a SOTA Mixture of Experts on Hugging Face
The Mixtral model is new, and seems to be good. Click on “Demo“ to test it
StarCoder: A State-of-the-Art LLM for Code
Article has comparison with other code-LLM models
huybery/Awesome-Code-LLM: An awesome and curated list of best code-LLM for research.
Hannibal046/Awesome-LLM: Awesome-LLM: a curated list of Large Language Model
Large language models and the rise of the AI code generators | InfoWorld
Review of LLM specialized for code generation
Large language model - Wikipedia
List of LLM models on Wikipedia
stabilityai (Stability AI) - Stable Diffusion running on Huggingface
Chat, models. Not open source, but instruct and relatively small (3B). The 3B instruct may be the best to try on Newspeak.
OpenAI Codex - Wikipedia
Model which generates code for Python, Javascript, Go, Shell, Perl, Swifg, Ruby, PHP
codellama (Code Llama) - Huggingface model for generating programs. Maybe can be used for Newspeak?
AI Code Tools: The Ultimate Guide in 2024
AI Code tools : Good summary. Does not talk about which pre-trained model they use. One is gemini (bard) -> alphacode2
BERT 101 - State Of The Art NLP Model Explained
Best summary of Natural Language Processing and terms - model (a language model - e.g. BertModel, defines encoder and decoder and their properties), transformer (a specific neural network based on attention paper), encoder (series of transformers on input), decoders (series of transformers on output). Bert does NOT use decoder. TensorFlow and PyTorch are possible backends to Transformers (NN). Summary: BERT is a highly complex and advanced language model that helps people automate language understanding.
BERT vs GPT: A Tale of Two Transformers That Revolutionized NLP | by Tavva Prudhvith | Medium
BigCode - Open and responsible development of LLMs for code
BigCode is an open scientific collaboration working on the responsible development and use of large language models for code
Replit — How to train your own Large Language Models
Hi level only talk about training for a language
How to train a new language model from scratch using Transformers and Tokenizers
Describes how to train a new language (desperanto) model.