Building and Training Generative AI Models: A Practical Guide to Generative AI Development and Scaling

Building and Training Generative AI Models: A Practical Guide to Generative AI Development and Scaling

Building and Training Generative AI Models: A Practical Guide to Generative AI Development and Scaling
Автор: Cronin Irena
Дата выхода: 2026
Издательство: Apress Media, LLC.
Количество страниц: 652
Размер файла: 3,3 МБ
Тип файла: PDF
Добавил: codelibs
 Проверить на вирусы

Table of Contents....5

About the Author....29

About the Technical Reviewer....30

Acknowledgments....31

Introduction....32

Chapter 1: Introduction to Generative AI Systems....36

Overview of Generative Modeling....36

1. Defining Generative Modeling....37

2. Types of Generative Models....39

a. Variational Autoencoders (VAEs)....39

b. Generative Adversarial Networks (GANs)....40

c. Autoregressive Models....40

d. Diffusion Models....41

3. Probabilistic Foundations....41

4. Training Methodologies and Loss Functions....43

5. Evaluation of Generative Models....44

6. Applications of Generative Models....47

7. Challenges in Generative Modeling....48

8. From Generative Models to Generative Systems....50

Applications Across Text, Vision, Audio, and Multimodal Domains....51

1. Text Generation....51

a. Conversational Agents....51

b. Content Generation....52

c. Code Synthesis....52

d. Summarization, Translation, and Editing....52

2. Vision and Image Generation....52

a. Image Synthesis....53

b. Style Transfer and Inpainting....53

c. Medical Imaging and Scientific Visualization....53

d. 3D Model and Scene Generation....53

3. Audio and Speech Generation....53

a. Speech Synthesis....54

b. Voice Cloning and Personalization....54

c. Music Generation....54

d. Audio Enhancement and Restoration....54

4. Multimodal Applications....55

a. Image-to-Text and Text-to-Image....55

b. Video Generation and Editing....55

c. Interactive Avatars and Digital Humans....55

d. Robotics and Embodied AI....55

e. Personalized Assistants and Cognitive Agents....56

5. Industry Use Cases by Domain....56

6. Challenges in Cross-Modality Generation....56

7. Future Trajectories....57

Comparison with Discriminative Models....57

1. Fundamental Objective Differences....58

2. Example Use Cases by Model Type....59

3. Architecture and Training Differences....60

4. Data Requirements and Representational Learning....60

5. Evaluation Metrics....61

6. Strengths and Weaknesses of Each Model Type....62

7. Interoperability and Hybrid Architectures....63

8. Role in Human-AI Collaboration....63

9. Evolutionary Trajectories....66

More on Architecture Families (Transformers, GANs, VAEs, Diffusion)....68

1. Transformers....69

a. Origins and Design....69

b. Architecture Components....69

c. Generative Use....70

d. Enterprise Applications....70

e. Hybrid Use....70

f. Advantages....70

g. Challenges....71

2. GANs....71

a. Origins and Concept....71

b. Architecture Components....71

c. Use in Generation....71

d. Enterprise Applications....72

e. Hybrid Use....72

f. Advantages....72

g. Challenges....72

3. VAEs....73

a. Origins and Objective....73

b. Architecture Components....73

c. Use in Generation....73

d. Enterprise Applications....73

e. Hybrid Use....74

f. Advantages....74

g. Challenges....74

4. Diffusion Models....74

a. Origins and Breakthroughs....74

b. Architecture Components....75

c. Use in Generation....75

d. Enterprise Applications....75

e. Hybrid Use....75

f. Advantages....76

g. Challenges....76

Chapter 2: Choosing the Right Architecture....77

Transformer-Based Models for Sequence Generation....77

Origins and Core Principles....78

Configurations for Sequence Generation....79

Advantages of Transformer-Based Sequence Generation....80

Long-Range Dependency Modeling....80

Parallelizable Training....81

Cross-Domain Adaptability....81

Scalability....82

Flexible Conditioning....83

Synergy of Advantages....84

Training Paradigms and Enhancements....86

Challenges and Limitations....86

Evaluation Considerations....87

Real-World Applications....88

Comparative Perspective with Other Architectures....89

Future Directions....89

GANs for Adversarial Training and Synthesis....90

Strengths of GANs....92

Challenges in GAN Training....92

Conditional GANs....93

Architectural Innovations....94

GANs Beyond Images....95

Comparison with Other Generative Architectures....95

Ethical and Societal Considerations....96

The Evolving Role of GANs....96

VAEs for Probabilistic Latent Modeling....97

Core Architecture and Probabilistic Foundation....97

The Reparameterization Trick....98

Advantages of Probabilistic Latent Modeling....98

Limitations and Challenges....99

Variants and Extensions....100

Applications of VAEs....101

Comparison with Other Generative Architectures....101

Role in Multimodal and Hybrid Architectures....102

Ethical Considerations....102

Future Directions....103

Diffusion Models and Denoising Strategies....103

Conceptual Foundation....104

Intuition Behind the Process....105

Variants of Diffusion Models....105

Denoising Strategies....106

Strengths of Diffusion Models....107

Stable Training....107

High-Output Fidelity....108

Mode Coverage....108

Flexibility in Conditioning....109

Theoretical Grounding....110

Synergy of Strengths....111

Practical Implications....112

Limitations and Challenges....113

Sampling Speed....113

Computational Cost....114

Memory Footprint....115

Over-smoothing Risk....115

Trade-Offs and Application-Specific Considerations....116

Ongoing Research Toward Mitigation....117

Applications....118

Comparison with Other Generative Architectures....119

Diffusion Models vs. GANs....119

Diffusion Models vs. VAEs....120

Diffusion Models vs. Transformers....122

Choosing the Right Architecture....123

Hybrid and Future Directions....124

Ethical and Societal Considerations....125

Chapter 3: Data Collection and Preparation....126

Data Types: Structured vs. Unstructured....129

The Nature of Structured Data....129

The Emergence of Unstructured Data....130

Semistructured Data: The Hybrid Zone....132

Differences in Storage and Retrieval....133

Preprocessing: Bringing Order to Disorder....134

Annotation and Labeling....136

Implications for Model Choice....137

Summary....138

Cleaning and Normalization Pipelines....140

Understanding the Purpose of Cleaning....140

Normalization As a Structural Prerequisite....142

Building a Cleaning and Normalization Pipeline....144

Handling Missing Data....146

Encoding Categorical and Text Variables....147

Detecting and Handling Outliers....148

Automating Data Cleaning....149

Monitoring for Drift....150

Case Study: Cleaning Clinical Trial Data....151

Final Thoughts....152

Synthetic Data Generation....152

What Is Synthetic Data?....153

Why Use Synthetic Data?....154

1. Privacy Preservation....154

2. Data Scarcity....154

3. Balanced Class Representation....155

4. Data Augmentation and Generalization....155

5. Controlled Experimentation....155

Techniques for Generating Synthetic Data....156

1. Rule-Based Simulations....156

2. Probabilistic Models....157

3. GANs....157

4. VAEs....158

5. Diffusion Models....159

6. Language Models for Textual Data....159

Evaluating Synthetic Data Quality....160

1. Fidelity....160

2. Utility....160

3. Privacy....161

Synthetic Data in Practice....161

Challenges and Considerations....162

Future Directions....162

Dataset Bias Detection and Mitigation....163

Defining Dataset Bias....164

Forms of Dataset Bias....164

Sampling Bias....164

Measurement Bias....164

Labeling Bias....165

Historical Bias....165

Representation Bias....165

Why Dataset Bias Matters....166

Detecting Dataset Bias....166

Exploratory Data Analysis (EDA)....166

Statistical Testing....167

Counterfactual Testing....167

Embedding and Latent Analysis....167

Bias Audits....167

Evaluating Bias in Model Performance....168

Mitigating Dataset Bias....168

Preprocessing Techniques....168

Rebalancing....168

Data Augmentation....169

Feature Filtering....169

In-Processing Techniques....169

Fairness-Constrained Learning....169

Representation Learning....170

Postprocessing Techniques....170

Threshold Adjustment....170

Outcome Calibration....170

Reject Option....170

Advanced Approaches....171

Causal Inference....171

Intersectional Fairness....171

Practical Workflow....171

Case Study: Bias in Facial Recognition Systems....172

Organizational and Legal Considerations....173

Governance....173

Regulation....173

Tools and Standards....174

Looking Ahead....174

Chapter 4: Training Fundamentals and Self-Supervised Learning....175

Self-Supervised Learning Principles....176

From Supervised to Self-Supervised....176

Core Idea of SSL....178

Pretext Tasks....180

Real-World Applications....181

Advantages and Challenges....183

Loss Function Design....185

Why Loss Functions Matter....185

Connection to Gradient-Based Optimization in Self-Supervised Learning....186

Reconstruction Losses....187

Adversarial Losses....189

Contrastive Losses....190

Hybrid Objectives....192

Reflections on Loss Function Design....193

Overfitting and Generalization....194

Defining Overfitting....195

Causes of Overfitting....196

Model Capacity....196

Dataset Size and Quality....197

Data Leakage....197

Spurious Correlations....198

Excessive Training....198

Lack of Regularization....198

Strategies to Improve Generalization....199

Regularization....199

Data Augmentation....199

Cross-Validation....200

Architectural Choices....200

Transfer Learning and Pretraining....200

Ensemble Methods....201

Monitoring and Validation....201

Modern Perspectives....201

Double Descent....202

Role of Pretraining....202

Implicit Regularization....202

Generalization in Large Language Models....203

Distribution Shifts....203

Ethical and Societal Dimensions....203

Monitoring Convergence and Training Progress....205

Training Dynamics....205

The Role of Learning Rate....206

Batch Size and Gradient Noise....206

Loss Curves and Plateaus....207

Oscillations and Instabilities....207

Training Diagnostics....207

Convergence Criteria....214

Loss Stabilization....214

Validation Metrics....215

Early Stopping....215

Gradient Norms....215

Representation Stability....216

Practical Monitoring Tools....216

TensorBoard....216

Weights & Biases....217

Custom Logging....217

Checkpointing....217

Distributed Monitoring....218

Pitfalls in Monitoring....218

Overfitting to Validation....218

Noisy Loss Curves....219

Metric Myopia....219

Resource Blindness....219

Emerging Techniques....220

Sharpness-Aware Monitoring....220

Representation Drift Tracking....220

Automated Alerts and Intervention....220

Evaluation Beyond Training Loss....221

Large-Scale Experiment Tracking....221

Conclusion....222

From Supervised to Self-Supervised....222

The Core Idea of Self-Supervised Learning....223

Loss Functions As Guides to Learning....224

Overfitting and Generalization....224

Monitoring Convergence and Training Progress....225

The Advantages and Challenges of Self-Supervised Learning....225

Interconnectedness of Training Concepts....226

Broader Reflections....227

Chapter 5: Optimization and Learning Strategies....229

Why Optimization Matters....230

Historical Roots....230

Core Challenges....231

Themes of the Chapter....231

Broader Importance....232

Backpropagation and Gradient Descent....233

Historical Context of Backpropagation....234

Core Mechanics of Backpropagation....235

Gradient Descent As the Optimization Backbone....236

Practical Challenges in Gradient Descent....237

Vanishing Gradients....238

Exploding Gradients....238

Saddle Points and Local Minima....238

Poor Conditioning....239

Variants of Gradient Descent....239

Momentum....239

Nesterov Accelerated Gradient....240

Mini-Batch Strategies....240

Second-Order Methods....240

Broader Perspectives....241

Adaptive Optimizers (Adam, RMSProp, LAMB)....242

Motivation for Adaptive Methods....242

RMSProp....243

Adam....245

LAMB and Large-Batch Training....246

Comparative Analysis....249

Broader Perspectives and Future Directions....252

Implicit Regularization....252

Scaling Laws....252

Hybrid Approaches....252

Optimizer–Architecture Interaction....252

Toward Automated Optimization....253

Gradient Clipping and Normalization....253

The Problem of Exploding Gradients....254

Gradient Clipping Strategies....255

Norm-Based Clipping....256

Value-Based Clipping....256

Adaptive and Per-Layer Clipping....257

Practical Heuristics for Clipping....258

Gradient Normalization Methods....258

Batch Normalization....258

Layer Normalization....259

Group and Instance Normalization....260

Weight and RMS Normalization....260

Broader Role of Normalization....260

Interactions Between Clipping, Normalization, and Optimizers....261

Extended Perspectives on Gradient Management....262

Learning Rate Schedules....263

Why Learning Rate Matters....264

Fixed vs. Dynamic Schedules....265

Fixed Learning Rates....265

Dynamic Learning Rates....266

Classical Decay Methods....266

Step Decay....266

Exponential Decay....267

Polynomial Decay....267

Analysis of Classical Decay....267

Modern Schedules....267

Cosine Annealing....268

Warmup....268

Cyclical Learning Rates....268

Super-Convergence....269

Comparative Insights....269

Large-Scale Training Considerations....269

Interaction with Batch Size....270

Distributed Training....270

Pretraining vs. Fine-Tuning....270

Resource Efficiency....270

Comparative Reflections and Future Trends)....271

Conclusion....272

Backpropagation and Gradient Descent: The Foundations....272

Adaptive Optimizers: Refinements of Gradient Descent....273

Gradient Clipping and Normalization: Stabilizing Training....274

Learning Rate Schedules: Orchestrating Progress....274

Interdependence of Strategies....275

Broader Reflections on Optimization....275

Looking Forward....276

Chapter 6: Scaling Training with Infrastructure and Distributed Systems....278

GPUTPU Usage and Resource Provisioning....279

Evolution from CPUs to GPUs and TPUs....280

Architectural Characteristics of GPUs....281

Architectural Characteristics of TPUs....283

Provisioning Strategies....284

Utilization Challenges....285

Broader Implications of Hardware Scaling....286

Data and Model Parallelism....287

Conceptual Foundations of Parallelism....288

Data Parallelism....289

Model Parallelism....290

Pipeline Parallelism....291

Tensor Parallelism....292

Hybrid Strategies....293

Bottlenecks and Challenges....295

Checkpointing and Fault Tolerance....296

The Need for Checkpointing in Large-Scale Training....297

Historical Context of Checkpointing....297

Principles of Fault Tolerance....298

Models of Failures....299

Checkpointing Strategies....300

Periodic Checkpointing....300

Incremental Checkpointing....300

Sharded Checkpointing....300

Hierarchical Checkpointing....301

Differential Checkpointing....301

Asynchronous Checkpointing....301

Trade-Offs in Frequency and Granularity....301

Frequency....302

Granularity....302

Elastic Recovery in Distributed Training....302

Conceptual Challenges and Bottlenecks....303

Broader Reflections and Future Directions....303

Distributed Training Frameworks....305

The Role of Frameworks in Distributed Training....305

Communication Primitives and Collective Operations....306

Gradient Synchronization Strategies....307

Parameter Server Paradigm....307

Ring-Allreduce....307

Hierarchical Allreduce....308

Gossip-Based Synchronization....308

Fault-Aware and Elastic Framework Designs....308

Abstractions for Parallelism (Data, Model, Pipeline, and Tensor)....309

Scalability Challenges in Framework Design....309

Conceptual Comparison of Framework Approaches....310

Horovod....310

PyTorch Distributed....311

TensorFlow Distributed Strategies....311

DeepSpeed....311

Future Directions in Distributed Training Frameworks....312

Conclusion....313

Infrastructure As the Third Pillar....313

Hardware As the Foundation....314

Parallelism As the Core Strategy....314

Resilience Through Checkpointing and Fault Tolerance....315

Frameworks As the Operational Backbone....316

Interdependence of Components....316

Broader Reflections....317

Untitled....278

Chapter 7: Fine-Tuning and Domain Adaptation....319

Transfer Learning and Pre-trained Model Utilization....320

Conceptual Foundations of Transfer Learning....321

Historical Evolution: From Feature Extraction to Pre-trained Transformers....322

Strategies for Transfer Learning....323

Full Model Fine-Tuning....323

Partial Fine-Tuning (Selected Layers)....324

Frozen Backbones with Task-Specific Heads....324

Benefits of Transfer Learning....325

Risks and Challenges....325

Conceptual Nuances in Pre-trained Model Utilization....327

The Broader Significance of Transfer Learning....328

Domain-Specific Dataset Curation....329

Pipeline....329

Why Domain-Specific Datasets Matter....332

Historical Context of Dataset Curation....332

Principles of Dataset Curation....333

Challenges in Collecting Domain Data....334

Annotation and Labeling Concerns....335

Engineering Workflows for Dataset Curation....336

Balancing Scale and Relevance....336

Synthetic Data Generation and Augmentation....337

Ethical and Privacy Considerations....338

Broader Implications of Domain Curation....339

Few-Shot and Zero-Shot Learning....341

Conceptual Foundations of Few-Shot and Zero-Shot Learning....341

Historical Context: From Meta-Learning to In-Context Learning....342

Few-Shot Learning Strategies....343

Fine-Tuning with Small Labeled Sets....343

Metric-Based Few-Shot Learning....344

In-Context Few-Shot Learning....344

Zero-Shot Learning Strategies....344

Prompt-Based Generalization....345

Pre-training Alignment....345

Cross-Modal Zero-Shot Learning....345

Mathematical Intuition Behind Few-Shot and Zero-Shot....346

Engineering Workflows for Few-Shot and Zero-Shot Evaluation....347

Benefits and Applications....348

Challenges and Risks....349

Explainability and Uncertainty....350

Case-Style Illustrations Across Domains....351

Future Trajectories....352

Use of Adapters and LoRA....353

Motivation for Parameter-Efficient Fine-Tuning....354

Early Approaches to Modular Adaptation....355

Conceptual Foundations of Adapters....356

Engineering Practices in Adapter-Based Fine-Tuning....357

Conceptual Foundations of LoRA....358

LoRA in Practice: Efficiency and Flexibility....359

Comparative Reflections: Adapters vs. LoRA....360

Case-Style Applications Across Domains....361

Challenges and Limitations....362

Future Directions in Parameter-Efficient Adaptation....363

Conclusion....364

The Centrality of Adaptation....365

Transfer Learning As the Foundation....367

Dataset Curation As the Anchor....367

Few-Shot and Zero-Shot Learning As Emerging Paradigms....368

Parameter-Efficient Fine-Tuning As a Practical Breakthrough....368

Integration and Interdependence....369

Broader Reflections....369

Chapter 8: Reinforcement Learning with Human Feedback (RLHF)....372

Human Annotation Pipelines....373

The Role of Human Preferences in RLHF....374

Annotation Formats: Rankings and Ratings....375

Recruiting and Training Annotators....376

Annotation Interfaces and Workflow Design....377

Quality Control and Inter-Annotator Agreement....378

Scaling Annotation Efforts....378

Case Illustration: InstructGPT....379

Challenges and Trade-Offs in Annotation Pipelines....380

Emerging Alternatives to Traditional Pipelines....380

Reward Modeling....381

Why Reward Modeling Matters in RLHF....382

Historical Roots of Reward Modeling....383

From Human Judgments to Reward Functions....384

Pairwise Comparisons and the Bradley–Terry Framework....384

Training Reward Models at Scale....385

Overfitting and Reward Hacking....386

Reward Model Generalization Across Domains....388

Case Illustrations: Summarization, Dialogue, and Instruction Following....388

Iterative Refinement of Reward Models....389

Challenges: Noise, Bias, and Misalignment....390

Alternatives and Extensions: Preference Distillation, AI-Assisted Feedback, Constitutional Signals....390

Future Directions in Reward Modeling....391

Proximal Policy Optimization (PPO)....393

Why PPO Matters in RLHF....393

Historical Context of PPO....394

Conceptual Principles of PPO....395

Integration with Pre-trained Policies....396

Engineering Practices in PPO for Language Models....396

PPO in Summarization and Dialogue Alignment....397

PPO Compared to Alternative Algorithms....398

Challenges of PPO in RLHF....398

Future Directions Beyond PPO....399

Broader Conceptual Significance of PPO....400

Challenges in Feedback Collection....401

The Cost of Feedback Collection....401

Annotation Quality and Consistency....402

Bias in Feedback Collection....403

Cultural and Linguistic Diversity....403

Annotation Noise and Reliability....404

The Problem of Over-optimization....405

Scaling Feedback Pipelines....405

Ethical Concerns in Feedback Collection....406

Emerging Alternatives to Human Feedback....406

The Fundamental Tension in Feedback Collection....407

Conclusion....408

Human Annotation Pipelines As the Foundation....409

Reward Modeling As the Surrogate for Human Judgment....409

PPO As the Engine of Alignment....410

Challenges in Feedback Collection As the Bottleneck....410

The Conceptual Significance of RLHF....411

Limitations and Open Questions....411

Chapter 9: Model Compression and Inference Optimization....413

Quantization and Weight Pruning....415

Motivation for Quantization and Pruning....416

Conceptual Foundations of Quantization....417

Types of Quantization....417

Post-training Quantization....418

Quantization-Aware Training....418

Mixed-Precision Quantization....418

Hardware Support for Quantization....419

Conceptual Foundations of Pruning....421

Types of Pruning....422

Unstructured Pruning....422

Structured Pruning....422

Historical Development of Pruning....423

Trade-Offs in Quantization and Pruning....423

Engineering Practices for Quantization and Pruning....424

Case Illustrations....425

Broader Implications....425

Knowledge Distillation....427

Motivation for Knowledge Distillation....427

Historical Roots of Distillation....428

Theoretical Principles of Distillation....428

Types of Knowledge Distillation....429

Task-Specific Distillation....429

Sequence-to-Sequence Distillation....430

Self-Distillation....430

Multi-teacher Distillation....430

Engineering Practices in Distillation....431

Applications in Natural Language Processing....431

Applications in Computer Vision and Speech....432

Challenges in Knowledge Distillation....432

Extensions and Innovations....433

Future Directions of Knowledge Distillation....433

Efficient Architectures....434

Historical Context of Efficient Architectures....435

Principles of Efficient Design....436

MobileBERT and TinyBERT....437

ALBERT and Parameter Sharing....438

TinyML and Edge Deployment....438

Sparse Attention Mechanisms....439

Trade-Offs in Efficient Architectures....440

Engineering Practices for Efficient Architectures....440

Future Directions of Efficient Architectures....441

Serving at Scale with Optimized Runtimes....443

The Challenge of Serving at Scale....445

Runtimes and Inference Engines....445

Batching Strategies for Efficiency....446

Memory Management in Large-Scale Serving....447

Operator Fusion and Graph Optimization....447

Hardware Specialization and Acceleration....448

Frameworks As Conceptual Exemplars....449

Elastic Scaling for Fluctuating Demand....449

Fault Tolerance and Reliability....450

Trade-Offs in Optimized Runtimes....450

Future Directions in Serving and Runtimes....451

Conclusion....452

Quantization and Pruning As First-Line Tools....452

Knowledge Distillation As Transfer of Learning....453

Efficient Architectures As Design for Deployment....453

Serving at Scale As the Final Bottleneck....454

Cross-Cutting Themes and Trade-Offs....454

Broader Implications of Compression and Optimization....455

Limitations and Open Questions....455

Chapter 10: Addressing Bias, Hallucinations, and Failure Modes....457

Detection and Mitigation of Hallucinated Outputs....459

Understanding Hallucinations in Generative Systems....459

Why Hallucinations Occur....460

Categories of Hallucinations....461

Risks and Implications of Hallucinations....461

Detection Strategies for Hallucinations....462

Mitigation Through Retrieval-Augmented Generation....463

Mitigation Through Decoding Constraints....463

RLHF for Hallucination Reduction....464

Hybrid Mitigation Pipelines....464

Case Illustrations of Hallucinations....465

Broader Implications....465

Managing Toxic and Biased Generations....466

Sources of Toxicity and Bias....467

Categories of Bias in Generative Systems....468

Historical Roots of Bias Awareness in AI....469

Risks and Implications of Toxic Outputs....470

Detection Strategies for Toxicity and Bias....471

Dataset Curation for Bias Mitigation....471

Algorithmic Debiasing Techniques....472

Reinforcement Learning with Human Feedback for Safety....473

Cultural and Contextual Adaptivity....473

Hybrid Mitigation Pipelines....474

Case Illustrations of Toxic and Biased Generations....475

Regulatory and Governance Frameworks....476

Open Questions and Future Directions....476

Model Safety Checks and Red Teaming....477

Rationale for Safety Checks in Generative Systems....479

Categories of Model Safety Checks....480

Automated vs. Human-in-the-Loop Safety....481

Red Teaming As a Conceptual Framework....483

Techniques of Red Teaming Generative Models....484

Lessons from Cybersecurity Red Teaming....485

Organizational Structures for Red Teaming....486

Continuous Monitoring and Adaptive Safety....487

Case Illustrations of Safety and Red Teaming....487

Trade-Offs in Safety and Red Teaming....490

Broader Implications of Safety and Red Teaming....491

Future Directions for Safety and Red Teaming....492

Historical Analogies in Safety Testing....492

Regulatory Integration of Red Teaming....493

Cultural and Ethical Dimensions of Safety....493

Multi-agent Red Teaming Ecosystems....494

Feedback Loops and Model Updates....495

Introduction to Feedback Loops in Generative Systems....496

Types of Feedback Loops (Positive, Negative, Reinforcing, Corrective)....496

Historical Parallels in Feedback Loops (Economics, Ecology, Control Theory)....497

Risks of Feedback in Deployed AI....498

Data Drift and Model Decay....499

Amplification of Bias Through Feedback....500

Feedback Loops in Recommender Systems As Precursors....500

Real-World Case Illustrations of Feedback Loops in Generative AI....501

Strategies for Monitoring Feedback Loops....502

Continuous Fine-Tuning and Iterative Model Updates....502

Governance and Oversight of Updates....503

Trade-Offs in Updating Models (Stability vs. Responsiveness)....504

User Feedback As a Resource and a Risk....504

Institutional Feedback Integration (Red Teams, Auditors, Regulators)....505

Technical Strategies for Safe Updating....505

Feedback Loops in Multimodal and Multi-agent Systems....506

Cultural and Ethical Dimensions of Feedback and Updates....507

Future Directions in Feedback and Updating....507

Historical Lessons in Feedback Management....508

Psychological and Behavioral Feedback Effects....508

Infrastructure Challenges in Frequent Updating....509

Long-Term Societal Risks of Runaway Feedback Loops....509

Conclusion....510

Chapter 11: Evaluation and Benchmarking of Generative Models....515

Automatic and Human Evaluation Methods....517

Historical Context of Evaluation in Generative Modeling....518

Core Principles of Evaluation....518

Automatic Evaluation Methods: Foundations....519

Human Evaluation: The Enduring Gold Standard....520

Case Study: BLEU in Machine Translation....521

Case Study: FID in Image Generation....521

Hybrid Approaches....522

Challenges of Human Evaluation....523

Toward Multidimensional Evaluation....523

Implications for Research and Deployment....524

Reference-Free Evaluation Paradigms....524

Cross-Modal Evaluation in Multimodal Systems....525

Evaluator Psychology and Cognitive Bias....526

Evaluation in High Stakes Domains....526

Cultural and Ethical Dimensions of Evaluation....527

Future Horizons in Evaluation....527

BLEU, FID, Perplexity, Diversity Scores....528

The Role of Metrics in Generative AI....529

BLEU: The Cornerstone of Text Evaluation....530

Beyond BLEU: ROUGE, METEOR, CIDEr, BERTScore....531

Perplexity: Measuring Language Model Fluency....533

Diversity Scores: Capturing Variety in Outputs....534

Benchmarking Image Generation....535

Statistical Properties and Interpretability of Metrics....536

Metric Gaming and Goodharts Law....537

Metrics in Industry vs. Academia....537

Cultural and Linguistic Variability in Metrics....538

The Political Economy of Benchmarks....538

Metrics in Live Systems....539

Longitudinal Evolution of Metrics....539

Open Problems in Metric Design....540

Toward Meta-Metrics and Ensemble Evaluation....541

The Future of Metrics....542

AB Testing Frameworks....544

Historical Origins of AB Testing....545

Statistical Foundations of AB Testing....546

Why AB Testing Matters for Generative AI....547

Designing AB Tests for Generative Systems....548

Metrics Within AB Testing....548

Case Studies in Text Generation....549

Case Studies in Vision and Multimodal Systems....549

Online vs. Offline AB Testing....550

Ethical Considerations in AB Testing....550

Statistical Pitfalls in AB Testing....551

Feedback Loops and Longitudinal Testing....552

Scaling AB Testing in Industry....552

Beyond AB: Multivariate and Bandit Testing....553

Cultural and Psychological Dimensions....553

Governance and Regulation of AB Testing....554

Sequential and Adaptive Experimentation....554

AB Testing for Safety and Alignment....555

Infrastructure Requirements for Large-Scale AB Testing....555

Psychological and Sociological Effects of Testing....556

Cultural Variability in Experiment Results....556

The Future of Evaluation Ecosystems....557

Task-Based Performance Assessments....558

Conceptual Foundations of Task-Based Evaluation....559

Intrinsic vs. Extrinsic Assessment....559

Historical Roots of Task-Based Assessments....560

Domain-Specific Examples....560

Designing Authentic Tasks....561

Measuring Outcomes....561

Human–AI Collaboration in Tasks....562

Intrinsic Benchmarks vs. Real-World Integration....562

Case Study: Code Generation....563

Case Study: Education....563

Challenges of Task-Based Evaluation....563

Automating Task Evaluation....564

Human-in-the-Loop Task Evaluation....564

Formative vs. Summative Assessment....564

Safety and Fairness in Task Evaluation....565

Scaling Task-Based Evaluation....565

Theories of Assessment from Education and Psychology....565

Task Complexity and Cognitive Load....566

Adaptive and Personalized Task Evaluation....566

Infrastructure for Continuous Task-Based Monitoring....566

Risks of Task Reductionism....567

Governance of Task-Based Evaluation....567

Ethical and Cultural Dimensions....567

Future Directions....568

Conclusion....569

Chapter 12: Future Directions and Deployment in Production....573

CICD Pipelines for Model Updates....575

Historical Origins of CICD in AI....576

Conceptual Foundations of CICD for Generative Models....576

Core Components of CICD Pipelines for Generative AI....579

Data Management in CICD Pipelines....581

Training Pipelines for Model Updates....582

Evaluation Pipelines....583

Model Registries and Versioning....583

Deployment Pipelines....584

Monitoring and Observability....584

Rollback and Recovery....585

Scaling CICD for Generative AI....585

Challenges and Open Problems....586

Future Directions in CICD Pipelines....586

Integration with DataOps....587

Cross-Organizational Collaboration....587

CICD and Explainability....588

Security in Model Pipelines....588

Environmental and Cost Considerations....588

Sociotechnical Dimensions....589

CICD for Multimodal and Agentic Systems....589

Model Monitoring and Observability....590

Historical Roots of Observability....591

Generative AI Monitoring Challenges....592

Performance and Efficiency Monitoring....593

Quality and Alignment Monitoring....593

Safety and Compliance Monitoring....594

Data Drift and Concept Drift Detection....595

User Feedback Integration....596

Infrastructure for Observability....596

Automated vs. Human-in-the-Loop Monitoring....597

Ethical and Legal Dimensions of Monitoring....597

Security Considerations....598

Multimodal Monitoring....599

Longitudinal and Life Cycle Monitoring....599

Organizational and Sociotechnical Considerations....600

Future Trends in Monitoring....600

Legal, Ethical, and Compliance Challenges....601

Historical Evolution of Tech Regulation....602

Intellectual Property and Generative Outputs....603

Privacy and Data Protection....604

Liability and Accountability....604

Bias, Fairness, and Social Equity....605

Transparency, Explainability, and Trust....606

Compliance Frameworks and Standards....606

Sector-Specific Compliance....607

Cross-Border Legal Complexities....608

Ethical Governance in Organizations....608

Human Oversight and Autonomy....609

Long-Term Ethical Risks....609

Case Studies and Precedents....610

Future Directions in Law and Policy....610

Trade Secrets and Proprietary Models....611

Antitrust and Competition Law in AI....611

Synthetic Data As a Compliance Strategy....612

Dual-Use and Misuse Dilemmas....612

Cultural and Epistemic Justice in AI Outputs....612

Comparative Global Regulatory Landscapes....613

The Economics of Compliance....613

Multimodal Fusion, Agentic Models, and AI Safety....614

Multimodal Fusion, Agentic Models, and AI Safety: Foundations and Implications....615

Technical Dimensions of Multimodal Fusion....616

Applications of Multimodal Systems....617

Ethical and Social Implications of Multimodal Fusion....618

Rise of Agentic Models....619

Technical Foundations of Agentic Models....620

Applications of Agentic Systems....621

Ethical and Governance Implications of Agentic Models....621

AI Safety As a Core Trend....623

Intersections of Trends....623

Case Studies Across Trends....624

Future Directions....625

Conclusion....626

Index....630

This book is a hands-on, technical guide to building and deploying generative AI models using advanced deep learning architectures like transformers, GANs, VAEs, and diffusion models. Designed for AI engineers, data scientists, and ML practitioners, it offers a practical roadmap from data ingestion to real-world deployment and evaluation.

The book starts by guiding readers on selecting the right model architecture for their application, be it text generation, image synthesis, or multimodal tasks. It then walks through essential components of model training, including dataset handling, self-supervised learning, and core optimisation techniques such as backpropagation, gradient descent, and learning rate scheduling. It also delves into large-scale training infrastructure, covering GPU/TPU usage, distributed computing frameworks, and system-level strategies for scaling performance. Practical guidance is provided on fine-tuning models with domain-specific data and applying reinforcement learning from human feedback (RLHF), model quantisation, and pruning to improve efficiency. Key challenges in generative AI—such as overfitting, bias, hallucination, and data efficiency—are addressed through proven techniques and emerging best practices. Readers will also gain insight into model interpretability and generalisation, ensuring robust and trustworthy outputs. The book demonstrates how to build scalable, production-ready generative systems across domains like media, healthcare, scientific simulation, and design through real-world examples and applied case studies.

By the end, readers will gain an understanding of how to architect, optimise, and apply generative models across diverse domains such as media creation, healthcare, design, scientific simulation, and beyond.

What you will learn;

  • Learn how to choose and implement generative models—VAEs, GANs, transformers, and diffusion models—for specific use cases.
  • Master training optimization techniques such as backpropagation, gradient descent, adaptive learning rates, and regularization.
  • Apply best practices for large-scale training using GPUs, TPUs, and distributed computing frameworks for performance scaling.
  • Boost model efficiency through quantization, pruning, fine-tuning, and RLHF to enhance output quality and reduce overhead.

Who this book is for:

AI Engineers and Machine Learning Practitioners looking to build and deploy generative models in real-world applications. Data Scientists working on deep learning projects involving text, vision, audio, or multimodal generation.


Похожее:

Список отзывов:

Нет отзывов к книге.