DeepHub IMBA
Author

DeepHub IMBA

A must‑follow public account sharing practical AI insights. Follow now. internet + machine learning + big data + architecture = IMBA

55
Articles
0
Likes
1
Views
0
Comments
Recent Articles

Latest from DeepHub IMBA

55 recent articles
DeepHub IMBA
DeepHub IMBA
Apr 24, 2026 · Artificial Intelligence

LangChain vs LangGraph: Choosing a Toolkit or an Orchestrator

The article compares LangChain and LangGraph by implementing the same three‑stage code‑review pipeline with identical agents and Gemini 2.5 Flash calls, showing when a linear toolkit suffices and when a state‑machine orchestrator becomes necessary.

AgentLLM OrchestrationLangChain
0 likes · 8 min read
LangChain vs LangGraph: Choosing a Toolkit or an Orchestrator
DeepHub IMBA
DeepHub IMBA
Apr 23, 2026 · Artificial Intelligence

Architectural Fixes for LLM Hallucinations: Inference Parameters, RAG, Constrained Decoding, and Post‑Generation Validation

The article breaks down LLM hallucination mitigation into five layers—runtime inference parameters, retrieval‑augmented generation and prompting tricks, constrained decoding with confidence calibration, post‑generation verification checks, and domain‑specific fine‑tuning plus continuous evaluation—showing how each layer reduces false, confident outputs.

LLMRAGconstrained decoding
0 likes · 11 min read
Architectural Fixes for LLM Hallucinations: Inference Parameters, RAG, Constrained Decoding, and Post‑Generation Validation
DeepHub IMBA
DeepHub IMBA
Apr 22, 2026 · Artificial Intelligence

A Survey of Time Series Forecasting Augmentation: Frequency Domain, Decomposition, and Patch Methods

The article reviews why classic classification augmentations fail for forecasting, outlines a taxonomy of effective time‑series augmentation techniques—including frequency‑domain, decomposition, and patch‑based methods—details the Temporal Patch Shuffle (TPS) pipeline, and presents extensive experiments showing TPS achieves state‑of‑the‑art improvements across long‑term, short‑term, and classification tasks.

data augmentationforecastingfrequency domain
0 likes · 17 min read
A Survey of Time Series Forecasting Augmentation: Frequency Domain, Decomposition, and Patch Methods
DeepHub IMBA
DeepHub IMBA
Apr 21, 2026 · Artificial Intelligence

Designing Persistent Memory for Production AI Agents: A Five‑Stage Pipeline and Four Design Patterns

Production AI agents require persistent memory to maintain continuity, learn from interactions, and recover from failures, but naïvely stuffing full conversation history into the LLM context incurs prohibitive latency and cost; this article outlines four memory types, a five‑stage pipeline, four design patterns, and practical metrics for building efficient, auditable memory systems.

AI agentsDesign PatternsLLM
0 likes · 27 min read
Designing Persistent Memory for Production AI Agents: A Five‑Stage Pipeline and Four Design Patterns
DeepHub IMBA
DeepHub IMBA
Apr 20, 2026 · Artificial Intelligence

What 10 Core Design Decisions the Claude Opus 4.7 Prompt Leak Reveals

The leaked Claude Opus 4.7 system prompt exposes ten intertwined design choices—ranging from treating psychological reconstruction as a danger signal to prohibiting over‑politeness, treating tool calls as cost‑free, using natural language as memory cues, and dynamically upgrading safety—illustrating a pattern of self‑regulation rather than pure capability enhancement.

AI safetyBehavioral ConstraintsClaude
0 likes · 8 min read
What 10 Core Design Decisions the Claude Opus 4.7 Prompt Leak Reveals
DeepHub IMBA
DeepHub IMBA
Apr 13, 2026 · Artificial Intelligence

From Retrieval to Answer: Three Overlooked Failure Points in RAG Pipelines

The article reveals silent failures in production RAG systems—where high retrieval scores and fluent LLM outputs still deliver incorrect answers—and proposes a four‑step observability loop (relevance gating, post‑generation evaluation, session‑wide tracing, and user‑signal logging) to detect and remediate these faults.

LLM evaluationObservabilityRAG
0 likes · 12 min read
From Retrieval to Answer: Three Overlooked Failure Points in RAG Pipelines
DeepHub IMBA
DeepHub IMBA
Apr 11, 2026 · Artificial Intelligence

Understanding Vector Similarity Search: Flat Index, IVF, and HNSW

This article explains why vector databases are needed for semantic search of unstructured data and provides a detailed, step‑by‑step comparison of three core vector similarity algorithms—cosine similarity, Flat Index, IVF, and HNSW—highlighting their trade‑offs in accuracy and speed.

HNSWIVFVector Search
0 likes · 10 min read
Understanding Vector Similarity Search: Flat Index, IVF, and HNSW