Showing 120 of 120on this page. Filters & sort apply to loaded results; URL updates for sharing.120 of 120 on this page
Model Quantization using Optimum HuggingFace - YouTube
Unknown quantization type, got fp8 · Issue #35471 · huggingface ...
Diffusion Transformers quantization · Issue #7376 · huggingface ...
GPT-OSS quantization demystified From the HuggingFace model card ...
Quantization for Ollama. Quantize any LLM from HuggingFace with… | by ...
4.7 Huggingface - Quantization - YouTube
HuggingFace Introduces Quanto: A Python Quantization Toolkit to Reduce ...
GPU is needed for quantization in M2 MacOS · Issue #23970 · huggingface ...
Quantization Fundamentals with Hugging Face
Model Quantization with 🤗 Hugging Face Transformers and Bitsandbytes ...
Quantization with Hugging Face Optimum
FLUX.2 [dev] Quantization - a Hugging Face Space by multimodalart
Static Quantization with Hugging Face `optimum` for ~3x latency ...
Tutorial: How to convert HuggingFace LLM models to Quantized file ...
Quantization - a Hugging Face Space by rakesh9177
Quantization Fundamentals with Hugging Face | Daniel Ibáñez
Quantization - a Hugging Face Space by PEFT
GitHub - huggingface/quanto: A pytorch Quantization Toolkit | Shrivatsan N.
Quantization Dedup - a Hugging Face Space by xet-team
Embedding Quantization - a Hugging Face Space by SwastikM
Slow performance in Quantization · Issue #309 · huggingface/text ...
Feature Request - 8 bit quantization for Efficieintnet · Issue #204 ...
NNCF quantization - a Hugging Face Space by echarlaix
huggingface/documentation-images · Embedding Quantization blogpost
Quantization Theory with Hugging Face
[docs] Quantization · Issue #27575 · huggingface/transformers · GitHub
LLM Quantization - a Hugging Face Space by bhaskartripathi
Loading directly 4bit quantized model · Issue #29604 · huggingface ...
Quantization Formats And Cuda Compute Capability Support - a Hugging ...
LLM Quantization Advanced - a Hugging Face Space by openfree
New course on linear quantization with Hugging Face | DeepLearning.AI ...
HiDream Full nf4 quantized · huggingface diffusers · Discussion #11337 ...
How to Run DeepSeek Locally: Using Hugging Face and Quantization for ...
🤖 Hugging Face has updated their quantization docs in Transformers ...
TorchAO Quantized Models and Quantization Recipes Now Available on ...
transformers/docs/source/en/quantization/awq.md at main · huggingface ...
HuggingFace Paper Explorer
Dynamic Quantization for GPT2 model from huggingface. · Issue #1401 ...
HuggingFace团队亲授大模型量化基础: Quantization Fundamentals with Hugging Face-CSDN博客
Quantization Aware Training. Train the model taking quantization… | by ...
GPTQ quantization for MPT-30 models · Issue #551 · huggingface/text ...
Efficient Multi-Model Inference with 4-bit Quantization in Hugging Face ...
GPTQ Quantization via `from_pretrained`: why enforcing `fp16`? · Issue ...
@macadeliccc on Hugging Face: "Benefits of `imatrix` quantization in ...
Advanced Techniques for Handling Big Models (Basics of Quantization ...
In Quantization Fundamentals with Hugging Face, you will learn how to ...
Running 4-bit on the fly quantization throws an error · Issue #22 ...
The Story of Hugging Face Model Quantization
New course with Hugging Face: Quantization in Depth 🤗 - YouTube
Model Quantization - A Lazy Data Science Guide
🔓 Unlock Custom Quantization for Hugging Face Models Locally with ...
Overview of natively supported quantization schemes in 🤗 Transformers
Quantizing Models from Hugging Face Using BitsnBytes | Quantization ...
A Guide to Supervised Fine-Tuning and 4-Bit Quantization for Language ...
quantization (Quantization)
Course 6 : Understanding Quantization Essentials with Hugging Face – AI ...
Quantization GPTQ - 🤗Optimum - Hugging Face Forums
Transformer Quantization at Darlene Stinson blog
HuggingFace团队亲授大模型量化基础: Quantization Fundamentals with Hugging Face ...
Fine-tune HuggingFace model and export it to Onnx format (Image ...
HuggingFace 核心组件及应用实战 - bingohuang - 博客园
hiuman/llama-3.1-8B-intruct-awq-quantization · Hugging Face
sanikadamn/Phi-3.5-mini-instruct-partial-quantization · Hugging Face
kernels-community/quantization-gptq · Hugging Face
goodasdgood/OmniGen_quantization · Hugging Face
LightningJay/L3-8B-Stheno-v3.2_Q8_0_gguf_and_exl2-bpw_8_bit ...
sade-adrien/quantization_samples · Datasets at Hugging Face
使用Hugging Face Transformers和Bitsandbytes集成进行模型量化 | ATYUN.COM 官网-人工智能教程 ...
How to quantize Large Language Models #huggingface #transformers # ...
blog/zh/kv-cache-quantization.md at main · huggingface/blog · GitHub
diffusers/docs/source/en/quantization/bitsandbytes.md at main ...
fxmarty/20220911-h13m58s49_sst2_distilbert_quantization · Hugging Face
huggingface/documentation-images at main
Convert Hugging Face model to GGUF | Dev Genius
GitHub - ksm26/Quantization-Fundamentals-with-Hugging-Face
blog/zh/1_58_llm_extreme_quantization.md at main · huggingface/blog ...
GitHub - edcalderin/huggingface-ragflow: This project implements a ...
huggingface/documentation-images at HEAD
blog/embedding-quantization.md at main · huggingface/blog · GitHub
FLUX.1-Kontext-dev Support for GGUF Quantized Model · Issue #11962 ...
QuantizationTorchAODraft - a Hugging Face Space by medmekk
GitHub - kaushikacharya/Quantization_Fundamentals: DeepLearning.ai ...
@Jaward on Hugging Face: "PyTorch implementation of the Self ...
GitHub - huggingface/diffusion-fast: Faster generation with text-to ...
Abnormally slow inference speed of quantized model? · Issue #24762 ...
GitHub - arita37/gguf-quantization: Google Colab script for quantizing ...
Quantize Hugging Face model to AWQ int4: A Step-by-Step Guide with ...
@macadeliccc on Hugging Face: "Quantize 7B paramater models in 60 ...
BitsAndBytes - a Hugging Face Space by HF-Quantization
Out of memory on 1xA100 for nf4-quantized-70b · Issue #890 ...
Quantized model loading failing with custom device mesh · Issue #40361 ...
Deep Dive into Hugging Face Quanto: A Comprehensive Guide to ...
Q8_0 @ Q4_0_4 GEMM/GEMV in llama.cpp | Ubios Home
you cannot fine-tune quantized model with torch.compile · Issue #1886 ...
How to Quantize a Model with Hugging Face Quanto - YouTube
How Hugging Face Positions Itself in the Open LLM Stack - The New Stack
Google Cloud and Hugging Face Announce Strategic Partnership to ...
mtc/mistralai-Mistral-7B-v0.1-arxiv-summarization-5000-finetuned ...