top of page
Theoretical AI
AI with Autobiographical Memory
Autobiographical memory in artificial intelligence refers to the systematic storage, retrieval, and configuration of an AI system’s past interactions, decisions, outcomes, and contextual experiences over extended periods. This capability enables the AI to construct a coherent internal narrative of its own operational history, forming a basis for identity and continuity within its operational environment. Without such memory, AI systems reset or operate in isolated sessions, l

Yatin Taneja
Mar 913 min read


AI with Consciousness Models: Simulating Subjective Experience (Theoretical)
Simulating the internal architecture of consciousness enables advanced self-monitoring and self-correction in artificial systems through the implementation of complex feedback mechanisms that mimic biological cognitive processes without requiring biological substrate. The focus remains on functional modeling of subjective experience rather than claiming actual sentience or phenomenological awareness, thereby sidestepping philosophical debates regarding the hard problem of con

Yatin Taneja
Mar 912 min read


Open-Source AI
Open-source AI constitutes a category of artificial intelligence encompassing models, tools, and frameworks where the underlying source code, parameter weights, and comprehensive training methodologies are made publicly accessible for utilization, modification, and redistribution by any interested party. This methodology stands in contrast to proprietary AI systems where access remains strictly restricted through application programming interfaces or licensing agreements gove

Yatin Taneja
Mar 99 min read


Orthogonality Thesis: Why Superintelligence Won't Automatically Share Human Values
The orthogonality thesis asserts that intelligence operates independently of the content or moral character of goals, establishing a foundational principle within the field of artificial intelligence safety research that distinguishes cognitive capability from objective selection. This concept posits that the level of intelligence a system possesses, defined as its capacity to solve problems, plan strategically, and manipulate its environment, does not inherently restrict the

Yatin Taneja
Mar 911 min read


Problem of Cognitive Load: Working Memory Limits in AI Planning
Cognitive load in AI planning is the processing strain placed on an agent's limited working memory during the execution of complex sequential reasoning tasks. Human working memory typically holds four to seven chunks of information at any given moment, whereas artificial working memory manages millions of parameters or tokens simultaneously. This vast difference in scale often obscures the fact that artificial systems still face absolute limits regarding how much context they

Yatin Taneja
Mar 916 min read


AI with Intuitive Mathematics
AI systems capable of generating mathematical conjectures through pattern recognition and heuristic reasoning mimic human intuitive leaps without relying on formal deductive proof at the initial stage. These systems analyze vast datasets of mathematical structures to identify recurring motifs or anomalies and propose relationships or identities that exhibit high empirical consistency across test cases. Such conjectures differ from random guesses as they represent statisticall

Yatin Taneja
Mar 916 min read


Interdisciplinary approaches to AI safety
Interdisciplinary approaches to AI safety integrate technical disciplines with humanities fields to address the complex challenge of aligning advanced AI systems with human values because purely technical methods often fail to account for contextual dimensions of human preferences, leading to misalignment even when systems perform optimally on narrow metrics such as accuracy or loss function minimization. Effective AI safety requires shared frameworks that translate abstract

Yatin Taneja
Mar 917 min read


Ultimate Limit of Intelligence: The Bekenstein-Hawking Entropy of Thought
Jacob Bekenstein established the relationship between black hole surface area and entropy during the 1970s by proposing that the loss of information into a black hole violates the second law of thermodynamics unless the black hole itself possesses entropy proportional to its goal area. This theoretical advancement suggested that the event future is a boundary where information is recorded rather than destroyed, forcing a reconciliation between quantum mechanics and general re

Yatin Taneja
Mar 910 min read


Abductive Reasoning: Inferring Best Explanations
Abductive reasoning operates as a distinct logical inference mechanism that initiates with a specific set of observations and proceeds to infer the most plausible explanation for those phenomena, standing in contrast to deductive reasoning, which derives certain conclusions from premises, and inductive reasoning, which generalizes rules from specific instances. Charles Sanders Peirce established the philosophical roots of this concept in the 19th century, distinguishing abduc

Yatin Taneja
Mar 98 min read


Causal Coherence in Superintelligence Self-Modeling
Causal coherence in superintelligence self-modeling refers to the strict alignment between an AI system’s internal representation of its own capabilities and the actual physical constraints imposed by its hardware substrate. This concept demands that the internal cognitive architecture maintains a mathematically rigorous correspondence between its simulated state transitions and the limitations of the material world in which it operates. A self-model acts as an active interna

Yatin Taneja
Mar 914 min read


bottom of page
