Search
Results
Optimum
[https://huggingface.co/docs/optimum/index] - - public:mzimmerm
Optimum is an extension of Transformers that provides a set of performance optimization tools to train and run models on targeted hardware with maximum efficiency. It is also the repository of small, mini, tiny models.
google/bert_uncased_L-4_H-256_A-4 ยท Hugging Face
[https://huggingface.co/google/bert_uncased_L-4_H-256_A-4] - - public:mzimmerm
Repository of all Bert models, including small. Start using this model for testing.
bigcode (BigCode)
[https://huggingface.co/bigcode] - - public:mzimmerm
Research community developing various code models, small and big. Models may not be instruct
WizardLM (WizardLM)
deepseek-ai (DeepSeek)
[https://huggingface.co/deepseek-ai] - - public:mzimmerm
They have the 1.3B version!!! This may be the best to start with Newspeak. Should work train even on huggingcface
stabilityai (Stability AI) - Stable Diffusion running on Huggingface
[https://huggingface.co/stabilityai] - - public:mzimmerm
Chat, models. Not open source, but instruct and relatively small (3B). The 3B instruct may be the best to try on Newspeak.