Search
Results
BERT Transformers – How Do They Work? | Exxact Blog
[https://www.exxactcorp.com/blog/Deep-Learning/how-do-bert-transformers-work] - - public:mzimmerm
Excellent document about BERT transformers / models and their parameters: - L=number of layers. - H=size of the hidden layer = number of vectors for each word in the sentence. - A = Number of self-attention heads - Total parameters.
Generative pre-trained transformer - Wikipedia
AMD Ryzen AI CPUs & Radeon 7000 GPUs Can Run Localized Chatbots Using LLMs Just Like NVIDIA's Chat With RTX
[https://wccftech.com/amd-ryzen-ai-cpus-radeon-7000-gpus-localized-chatbot-llms-like-nvidia-chat-with-rtx/] - - public:mzimmerm
LM Studio can be installed on Linux with APU or GPU (looks like it needs the AI CPU though??) and run LLM. Install on Laptop and test if it works.
A Step-by-Step Guide to Model Evaluation in Python | by Shreya Singh | Medium
[https://medium.com/@jscvcds/a-step-by-step-guide-to-model-evaluation-in-python-3a72dee92560] - - public:mzimmerm
How to train a new language model from scratch using Transformers and Tokenizers
[https://huggingface.co/blog/how-to-train] - - public:mzimmerm
Describes how to train a new language (desperanto) model.