Search
Results
DICTA - Hebrew LLM model
playcanvas/model-viewer: 3D Model Viewer supporting glTF and 3D Gaussian Splats
[2510.21890] The Principles of Diffusion Models
Michael-A-Kuykendall/shimmy: ⚡ Python-free Rust inference server — OpenAI-API compatible. GGUF + SafeTensors, hot model swap, auto-discovery, single binary. FREE now, FREE forever.
[2509.07025] 1 bit is all we need: binary normalized neural networks
Beyond the Cloud: Why I’m Now Running Enterprise AI on My Laptop (Without Internet) | by Klaudi | Aug, 2025 | Medium
Multiverse Computing – Quantum AI software revolution.
My 2.5 year old laptop can write Space Invaders in JavaScript now, using GLM-4.5 Air and MLX
จาก Core Competency สู่ Future Skills: วิธีการออกแบบ Competency Model ให้ตอบโจทย์องค์กร
บันได 5 ขั้น ปั้นคนเก่งแบบญี่ปุ่น - senseipae
Polygonjs: WebGL Design & Animation Tool
(86) QwQ: Tiny Thinking Model That Tops DeepSeek R1 (Open Source) - YouTube
Model which uses reinforcement learning.
12 วิธีการสร้างและวิเคราะห์ Threat Modeling
Threat Modeling: 12 Available Methods
What is Cyber Threat Modeling | Importance of Threat Modeling | EC-Council
ทำไมต้องกำหนดความสามารถหลักองค์กร (Core competency) – Baramizi
Competency Modeling - Assessment Associates International
Competency
สมรรถนะ (Competency Model) มีอะไรบ้าง ? - Oriental Phoenix
13 Best Open Source ChatGPT Alternatives
Mit wenigen Klicks zum eigenen KI-Chatbot: Warum ihr dieses Tool kennen solltet
Train and use my model
Who needs GitHub Copilot when you can roll your own AI code assistant at home • The Register
lyogavin/godmodeanimation: 2D Game Animation in God Mode
OpenAI’s GPT-4o Mini isn’t much better than rival LLMs • The Register
Honey, I shrunk the LLM! A beginner's guide to quantization • The Register
abi/secret-llama: Fully private LLM chatbot that runs entirely with a browser with no server needed. Supports Mistral and LLama 3.
What is a Model of Reflection?
Input Output Outcome: Use the misunderstood Outcome Model successfully
Product, Process, Resource – an Integrated Modeling Approach for Production Engineering and Industrialized Construction - Project Production InstituteProject Production Institute
Optimum
Optimum is an extension of Transformers that provides a set of performance optimization tools to train and run models on targeted hardware with maximum efficiency. It is also the repository of small, mini, tiny models.
google-research/bert: TensorFlow code and pre-trained models for BERT
google/bert_uncased_L-4_H-256_A-4 · Hugging Face
Repository of all Bert models, including small. Start using this model for testing.
A Step-by-Step Guide to Model Evaluation in Python | by Shreya Singh | Medium
Open LLM Leaderboard - a Hugging Face Space by HuggingFaceH4
Comparison of efficiency of all LLM models on hugging face
Training Bert on Yelp - Copy of training.ipynb - Colaboratory
(1) Most cost effective GPU for local LLMs? : LocalLLaMA
GGML quantized models. They would let you leverage CPU and system RAM, instead of having to rely on a GPU’s. This could save you a fortune, especially if go for some used AMD Epyc platforms. This could be more viable for the larger models, especially the 30B/65B parameters models which would still press or exceed the VRAM on the P40.
Optimizing LLMs for Speed and Memory
7 steps to master large language models (LLMs) | Data Science Dojo
LLM for a new language : MachineLearning
High level how to train a model
Up to date List of LLM Models
(2) Are there any tiny (1-3b) models finetuned for coding available in GGUF format? : LocalLLaMA
bigcode (BigCode)
Research community developing various code models, small and big. Models may not be instruct
WizardLM (WizardLM)
deepseek-ai (DeepSeek)
They have the 1.3B version!!! This may be the best to start with Newspeak. Should work train even on huggingcface
deepseek-ai/deepseek-coder-6.7b-instruct · Hugging Face
Another possible model. For coding capabilities, Deepseek Coder achieves state-of-the-art performance among open-source code models on multiple programming languages and various benchmarks.
LLaMA 7B GPU Memory Requirement - Transformers - Hugging Face Forums
With the optimizers of bitsandbytes (like 8 bit AdamW), you would need 2 bytes per parameter, or 14 GB of GPU memory.
stabilityai/stable-code-3b · Hugging Face
Another potential model to use for Newspeak, but it is NOT open source. Adventage: 2.5B params, so should be usable in small GPUs
Can Ai Code Results - a Hugging Face Space by mike-ravkine
Comparison of LLM models for coding
openchat/openchat-3.5-0106 · Hugging Face
Open source with lots of information. Uses Multiple undrelying models. Not sure how I would train for it
Welcome Mixtral - a SOTA Mixture of Experts on Hugging Face
The Mixtral model is new, and seems to be good. Click on “Demo“ to test it
