Media Summary: In this video, we discuss the fundamentals of model In this video I will introduce and explain Run massive AI models on your laptop! Learn the secrets of LLM

Inference With Quantized Weights Quantization Tensorteach - Detailed Analysis & Overview

In this video, we discuss the fundamentals of model In this video I will introduce and explain Run massive AI models on your laptop! Learn the secrets of LLM Are you planning to deploy a deep learning model on any edge device (microcontrollers, cell phone or wearable device)? About Intel Software: Intel® Developer Zone is committed to empowering and assisting software developers in creating ... Try Voice Writer - speak your thoughts and let AI handle the grammar: Four techniques to optimize the speed ...

We show you how to load in a model from hugging face and Reminder⚠️ Get 55% off your ODSC Europe experience. Just enter promo code odsc_video and save on your ticket to ODSC ... Check out the latest book by Vivek Kalyanarangan Download the AI model guide to learn more → Learn more about the technology → Speaker: Suraj Subramanian, Developer Advocate, PyTorch Suraj is a developer advocate and ML engineer at Meta AI. Ahmed T. Elthakeb, Prannoy Pilligundla, Alex Cloninger, Hadi Esmaeilzadeh (UC-San Diego) on Split Learning for ...

Photo Gallery

Inference With Quantized Weights | Quantization | TensorTeach
How LLMs survive in low precision | Quantization Fundamentals
What is LLM quantization?
Quantization explained with PyTorch - Post-Training Quantization, Quantization-Aware Training
Optimize Your AI - Quantization Explained
Quantization in deep learning | Deep Learning Tutorial 49 (Tensorflow, Keras & Python)
Efficient Large Language Model Inference with SqueezeLLM and KVQuant | Intel AI DevSummit 2025
Quantization vs Pruning vs Distillation: Optimizing NNs for Inference
tinyML Talks: A Practical Guide to Neural Network Quantization
Quantizing Models from Hugging Face Using BitsnBytes | Quantization | TensorTeach
Leaner and Greener AI with Quantization in PyTorch - SURAJ SUBRAMANIAN
Quantization and Fast Inference for Modern AI
View Detailed Profile
Inference With Quantized Weights | Quantization | TensorTeach

Inference With Quantized Weights | Quantization | TensorTeach

We discuss how to perform

How LLMs survive in low precision | Quantization Fundamentals

How LLMs survive in low precision | Quantization Fundamentals

In this video, we discuss the fundamentals of model

What is LLM quantization?

What is LLM quantization?

In this video we define the basics of

Quantization explained with PyTorch - Post-Training Quantization, Quantization-Aware Training

Quantization explained with PyTorch - Post-Training Quantization, Quantization-Aware Training

In this video I will introduce and explain

Optimize Your AI - Quantization Explained

Optimize Your AI - Quantization Explained

Run massive AI models on your laptop! Learn the secrets of LLM

Quantization in deep learning | Deep Learning Tutorial 49 (Tensorflow, Keras & Python)

Quantization in deep learning | Deep Learning Tutorial 49 (Tensorflow, Keras & Python)

Are you planning to deploy a deep learning model on any edge device (microcontrollers, cell phone or wearable device)?

Efficient Large Language Model Inference with SqueezeLLM and KVQuant | Intel AI DevSummit 2025

Efficient Large Language Model Inference with SqueezeLLM and KVQuant | Intel AI DevSummit 2025

About Intel Software: Intel® Developer Zone is committed to empowering and assisting software developers in creating ...

Quantization vs Pruning vs Distillation: Optimizing NNs for Inference

Quantization vs Pruning vs Distillation: Optimizing NNs for Inference

Try Voice Writer - speak your thoughts and let AI handle the grammar: https://voicewriter.io Four techniques to optimize the speed ...

tinyML Talks: A Practical Guide to Neural Network Quantization

tinyML Talks: A Practical Guide to Neural Network Quantization

"A Practical Guide to Neural Network

Quantizing Models from Hugging Face Using BitsnBytes | Quantization | TensorTeach

Quantizing Models from Hugging Face Using BitsnBytes | Quantization | TensorTeach

We show you how to load in a model from hugging face and

Leaner and Greener AI with Quantization in PyTorch - SURAJ SUBRAMANIAN

Leaner and Greener AI with Quantization in PyTorch - SURAJ SUBRAMANIAN

Reminder⚠️ Get 55% off your ODSC Europe experience. Just enter promo code odsc_video and save on your ticket to ODSC ...

Quantization and Fast Inference for Modern AI

Quantization and Fast Inference for Modern AI

Check out the latest book by Vivek Kalyanarangan

[2023 Best AI Paper] SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compressio

[2023 Best AI Paper] SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compressio

Title: SpQR: A Sparse-

AI Inference: The Secret to AI's Superpowers

AI Inference: The Secret to AI's Superpowers

Download the AI model guide to learn more → https://ibm.biz/BdaJTb Learn more about the technology → https://ibm.biz/BdaJTp ...

Quantizing LLMs - How & Why (8-Bit, 4-Bit, GGUF & More)

Quantizing LLMs - How & Why (8-Bit, 4-Bit, GGUF & More)

Quantizing

Leaner, Greener and Faster Pytorch Inference with Quantization

Leaner, Greener and Faster Pytorch Inference with Quantization

Speaker: Suraj Subramanian, Developer Advocate, PyTorch Suraj is a developer advocate and ML engineer at Meta AI.

Quantization in Deep Learning (LLMs)

Quantization in Deep Learning (LLMs)

This video is about

Divide and Conquer: Leveraging Intermediate Feature Representations for Quantized Training of NN's

Divide and Conquer: Leveraging Intermediate Feature Representations for Quantized Training of NN's

Ahmed T. Elthakeb, Prannoy Pilligundla, Alex Cloninger, Hadi Esmaeilzadeh (UC-San Diego) @Workshop on Split Learning for ...