Skip to content
wwwinsights_logo

Insights into World Wide Web

My Humble Thoughts about Web Dev and Website Reviews

wwwinsights_logo
Insights into World Wide Web

My Humble Thoughts about Web Dev and Website Reviews

  • Home
  • AI
  • Content Management
    • Drupal
    • WordPress
  • Web Infrastructure
    • AWS
    • Network
    • Git
    • Nginx
    • Infra
  • Programming Language
    • Python
  • Contact
  • About
AI

How Large Language Models Really Work: Next-Token Prediction at the Core

Dr. Anjing Wang October 7, 2025 0 Comments

If you peel away all the complexity of modern large language models (LLMs)—billions of parameters, reinforcement learning from human feedback, retrieval-augmented generation—the essence of how they work comes down to…

AI

Understanding the Three Types of Transformers: Encoder, Decoder, and Encoder–Decoder

Dr. Anjing Wang October 7, 2025 0 Comments

The term Transformer has become almost synonymous with modern large language models (LLMs). But when people talk about “encoder-only,” “decoder-only,” or “encoder–decoder” architectures, they are drawing on terminology that predates…

AI

Why Hugging Face Shows the “Wrong” Parameter Count for AWQ Models

Dr. Anjing Wang September 19, 2025 0 Comments

If you’ve loaded an AWQ-quantized model from Hugging Face (like Qwen2.5-VL-3B-Instruct-AWQ), you might have noticed something confusing: 👉 Hugging Face says the model has ~0.9B parameters, but the architecture is…

AI

PyTorch Matrix Multiplication: matmul, mm, and @

Dr. Anjing Wang September 19, 2025 0 Comments

Matrix multiplication is one of the most fundamental operations in machine learning. In PyTorch, you’ll often see three different ways to do it: At first glance, they look interchangeable —…

AI

LoRA: Low-Rank Adaptation Made Simple

Dr. Anjing Wang September 17, 2025 0 Comments

Large language models are huge — billions of parameters, often stored as massive square weight matrices like 4096 × 4096. Fine-tuning all of those parameters for a new task is…

AI

Mixed Precision Training: Faster Deep Learning Without Losing Accuracy

Dr. Anjing Wang September 17, 2025 0 Comments

Training today’s deep learning models is resource-hungry. Models have billions of parameters, and every step requires trillions of floating-point operations. To make training feasible, researchers and engineers rely on mixed…

AI

Understanding FP32, FP16, and BF16: Floating-Point Formats in Deep Learning

Dr. Anjing Wang September 17, 2025 0 Comments

Modern deep learning wouldn’t be possible without floating-point numbers. They’re the backbone of every matrix multiplication, activation, and gradient update. But as models grow larger and GPUs become more specialized,…

AI

Demystifying Floating-Point Precision: Half, Single, and Double

Dr. Anjing Wang September 16, 2025 0 Comments

If you’ve ever written code in Python, CUDA, or TensorFlow, you’ve probably seen terms like float16, float32, or float64. They map directly to the IEEE-754 floating-point standard: But what do…

AI

Causal Language Models vs. Conditional Generation Models: Key Differences and Real-World Examples

Dr. Anjing Wang February 7, 2025 0 Comments

Natural language generation has rapidly evolved with the rise of large language models, but one common point of confusion is distinguishing between causal language models (CLMs) and conditional generation models.…

AI

Unlocking the Power of Conditional Generation in AI

Dr. Anjing Wang February 2, 2025 0 Comments

Artificial Intelligence (AI) has revolutionized how we interact with technology, from chatbots that answer questions to AI models that generate lifelike images and translate languages instantly. But behind many of…

AI

Unveiling NVIDIA DGX: A Journey Through AI Supercomputing

Dr. Anjing Wang November 19, 2024 0 Comments

NVIDIA DGX (Deep GPU Xceleration) is synonymous with cutting-edge artificial intelligence (AI) infrastructure. Designed to accelerate AI research and applications, the DGX family of systems provides unparalleled computational power for…

AI

How to Measure the Performance of OCR: Why BLEU Isn’t Always the Best Choice

Dr. Anjing Wang November 13, 2024 0 Comments

Optical Character Recognition (OCR) is a technology that converts images of text (such as scanned documents, photos, or screenshots) into machine-readable text. While OCR has come a long way, evaluating…

AI

Understanding sklearn.metrics.accuracy_score and How to Calculate Accuracy Manually

Dr. Anjing Wang November 13, 2024 0 Comments

When evaluating machine learning models, accuracy is one of the most commonly used metrics for classification tasks. In this blog post, we’ll dive into the accuracy_score function provided by Scikit-Learn’s…

AI

Supervised Fine-Tuning (SFT): How to Fine-Tune Your Model Like a Pro

Dr. Anjing Wang November 1, 2024 0 Comments

In the world of machine learning, pretrained models are like finding a treasure chest of knowledge. They save us hours, days, or even weeks of training time, allowing us to…

AI

Deep Learning Model Precision: FP32, BF16, INT8 and INT4

Dr. Anjing Wang October 31, 2024 0 Comments

When training or deploying deep learning models, precision isn’t just about getting accurate predictions—it’s also about finding the right balance between performance, memory usage, and speed. Choosing the optimal precision…

Posts pagination

1 2 … 5

Next Page »

About the Author

Dr. Anjing Wang

Dr. Anjing Wang

I am an entrepreneur, data analyst, website builder, and enthusiastic educator. I hold a Ph.D. in computer science and worked for big corporations for years before starting up a few companies. Notably, one of my recent start-up projects is a video dictionary.

  • AI
  • AWS
  • DNS
  • Drupal
  • Git
  • Infra
  • Network
  • Nginx
  • Python
  • Wordpress
How Large Language Models Really Work: Next-Token Prediction at the Core
October 7, 2025
Understanding the Three Types of Transformers: Encoder, Decoder, and Encoder–Decoder
October 7, 2025
Why Hugging Face Shows the “Wrong” Parameter Count for AWQ Models
September 19, 2025
PyTorch Matrix Multiplication: matmul, mm, and @
September 19, 2025
LoRA: Low-Rank Adaptation Made Simple
September 17, 2025
Mixed Precision Training: Faster Deep Learning Without Losing Accuracy
September 17, 2025
Understanding FP32, FP16, and BF16: Floating-Point Formats in Deep Learning
September 17, 2025
Demystifying Floating-Point Precision: Half, Single, and Double
September 16, 2025
Causal Language Models vs. Conditional Generation Models: Key Differences and Real-World Examples
February 7, 2025
Unlocking the Power of Conditional Generation in AI
February 2, 2025
Unveiling NVIDIA DGX: A Journey Through AI Supercomputing
November 19, 2024
How to Measure the Performance of OCR: Why BLEU Isn’t Always the Best Choice
November 13, 2024
Understanding sklearn.metrics.accuracy_score and How to Calculate Accuracy Manually
November 13, 2024
Supervised Fine-Tuning (SFT): How to Fine-Tune Your Model Like a Pro
November 1, 2024
Deep Learning Model Precision: FP32, BF16, INT8 and INT4
October 31, 2024
wwwinsights_logo

Insights into World Wide Web

My Humble Thoughts about Web Dev and Website Reviews

Copyright by Anjing Wang © All rights reserved | Blogus by Themeansar.

  • Home
  • About
  • Terms of Use
  • Privacy Policy
  • Contact