Latest

Thoughts

Addressing the Conjunction Fallacy in Probabilistic Information Retrieval: From Theory to Practice

1. Introduction In our previous explorations of probabilistic frameworks for information retrieval, we examined how transformations like softmax and sigmoid convert raw similarity scores into probabilities, enabling principled fusion of heterogeneous retrieval systems. While these transformations provide elegant mathematical foundations for ranking, they introduce a critical challenge when handling conjunctive

By J

Thoughts

Beyond Mathematical Unity: From the XOR Problem to the Theoretical Limits of Backpropagation

Introduction Our previous exploration of "The Mathematical Unity of Sigmoid, Perceptron, Logistic Regression, and Softmax" established the foundational equivalences between these core machine learning concepts. We demonstrated how sigmoid-activated perceptrons are mathematically identical to logistic regression, and how softmax functions generalize sigmoid to multi-class scenarios. This mathematical unity

By J

Thoughts

Progressive and Adaptive Hyperparameter Estimation in BM25 Probability Transformation: A Unified Approach

1. Introduction The transformation of BM25 similarity scores into probability estimates represents a critical challenge in information retrieval systems. This process is essential for creating interpretable search results and enabling integration with probabilistic frameworks. While supervised learning approaches using query-document relevance pairs typically yield optimal results, practical implementations often face

By J

Thoughts

Beyond Softmax: Probabilistic Foundations and Bayesian Frameworks in Hybrid Search

Introduction In our previous exploration of probability transformations in vector search, we examined how softmax enables the normalization of disparate scoring systems into comparable probabilistic frameworks. This follow-up article delves deeper into the mathematical theory underpinning these transformations, with a specific focus on Bayesian probabilistic frameworks and their application to

By J

Thoughts

Computational Graph Logging and Differential Analysis for LLM Function Extraction

Abstract This research essay explores a novel approach to understanding Large Language Models (LLMs) through computational graph logging and differential analysis. We propose treating LLMs as complex mathematical functions and extracting their functional behavior by systematically logging kernel operations during inference. Our approach introduces two key innovations: (1) probabilistic kernel

By J