Creation Questions

Tag: science

  • Evidence for an Active Alternative Promoter in the Human DDX11L2 Gene

    Evidence for an Active Alternative Promoter in the Human DDX11L2 Gene

    Abstract

    The human genome contains numerous regulatory elements that control gene expression, including canonical and alternative promoters. While DDX11L2 is annotated as a pseudogene, its functional relevance in gene regulation has been a subject of interest. This study leverages publicly available genomic data from the UCSC Genome Browser, integrating information from the ENCODE project and ReMap database, to investigate the transcriptional activity within a specific intronic region of the DDX11L2 gene (chr2:113599028-113603778, hg38 assembly). Our analysis reveals the co-localization of key epigenetic marks, candidate cis-regulatory elements (cCREs), and RNA Polymerase II binding, providing robust evidence for an active alternative promoter within this region. These findings underscore the complex regulatory landscape of the human genome, even within annotated pseudogenes.

    1. Introduction

    Gene expression is a tightly regulated process essential for cellular function, development, and disease. A critical step in gene expression is transcription initiation, primarily mediated by RNA Polymerase II (Pol II) in eukaryotes. Transcription initiation typically occurs at promoter regions, which are DNA sequences located upstream of a gene’s coding sequence. However, a growing body of evidence indicates the widespread use of alternative promoters, which can initiate transcription from different genomic locations within or outside of a gene’s canonical promoter, leading to diverse transcript isoforms and complex regulatory patterns [1].

    The DDX11L2 gene, located on human chromosome 2, is annotated as a DEAD/H-box helicase 11 like 2 pseudogene. Pseudogenes are generally considered non-functional copies of protein-coding genes that have accumulated mutations preventing their translation into functional proteins. Despite this annotation, some pseudogenes have been found to play active regulatory roles, for instance, by producing non-coding RNAs or acting as cis-regulatory elements [2]. Previous research has suggested the presence of an active promoter within an intronic region of DDX11L2, often discussed in the context of human chromosome evolution [3].

    This study aims to independently verify the transcriptional activity of this specific intronic region of DDX11L2 by analyzing comprehensive genomic and epigenomic datasets available through the UCSC Genome Browser. We specifically investigate the presence of key epigenetic hallmarks of active promoters, the classification of cis-regulatory elements, and direct evidence of RNA Polymerase II binding.

    2. Materials and Methods

    2.1 Data Sources

    Genomic and epigenomic data were accessed and visualized using the UCSC Genome Browser (genome.ucsc.edu), utilizing the Human Genome assembly hg38. The analysis focused on the genomic coordinates chr2:113599028-113603778, encompassing the DDX11L2 gene locus.

    The following data tracks were enabled and examined in detail:

    ENCODE Candidate cis-Regulatory Elements (cCREs): This track integrates data from multiple ENCODE assays to classify genomic regions based on their regulatory potential. The “full” display mode was selected to visualize the color-coded classifications (red for promoter-like, yellow for enhancer-like, blue for CTCF-bound) [4].

    Layered H3K27ac: This track displays ChIP-seq signal for Histone H3 Lysine 27 acetylation, a histone modification associated with active promoters and enhancers. The “full” display mode was used to visualize peak enrichment [5].

    ReMap Atlas of Regulatory Regions (RNA Polymerase II ChIP-seq): This track provides a meta-analysis of transcription factor binding sites from numerous ChIP-seq experiments. The “full” display mode was selected, and the sub-track specifically for “Pol2” (RNA Polymerase II) was enabled to visualize its binding profiles [6].

    DNase I Hypersensitivity Clusters: This track indicates regions of open chromatin, which are accessible to regulatory proteins. The “full” display mode was used to observe DNase I hypersensitive sites [4].

    GENCODE Genes and RefSeq Genes: These tracks were used to visualize the annotated gene structure of DDX11L2, including exons and introns.

    2.2 Data Analysis

    The analysis involved visual inspection of the co-localization of signals across the enabled tracks within the DDX11L2 gene region. Specific attention was paid to the first major intron, where previous studies have suggested an alternative promoter. The presence and overlap of red “Promoter-like” cCREs, H3K27ac peaks, and Pol2 binding peaks were assessed as indicators of active transcriptional initiation. The names associated with the cCREs (e.g., GSE# for GEO accession, transcription factor, and cell line) were noted to understand the experimental context of their classification.

    3. Results

    Analysis of the DDX11L2 gene locus on chr2 (hg38) revealed consistent evidence supporting the presence of an active alternative promoter within its first intron.

    3.1 Identification of Promoter-like cis-Regulatory Elements:

    The ENCODE cCREs track displayed multiple distinct red bars within the first major intron of DDX11L2, specifically localized around chr2:113,601,200 – 113,601,500. These red cCREs are computationally classified as “Promoter-like,” indicating a high likelihood of promoter activity based on integrated epigenomic data. Individual cCREs were associated with specific experimental identifiers, such as “GSE46237.TERF2.WI-38VA13,” “GSE102884.SMC3.HeLa-Kyoto_WAPL_PDS-depleted,” and “GSE102884.SMC3.HeLa-Kyoto_PDS5-depleted.” These labels indicate that the “promoter-like” classification for these regions was supported by ChIP-seq experiments targeting transcription factors like TERF2 and SMC3 in various cell lines (WI-38VA13, HeLa-Kyoto, and HeLa-Kyoto under specific depletion conditions).

    3.2 Enrichment of Active Promoter Histone Marks:

    A prominent peak of H3K27ac enrichment was observed in the Layered H3K27ac track. This peak directly overlapped with the cluster of red “Promoter-like” cCREs, spanning approximately chr2:113,601,200 – 113,601,700. This strong H3K27ac signal is a hallmark of active regulatory elements, including promoters.

    3.3 Direct RNA Polymerase II Binding:

    Crucially, the ReMap Atlas of Regulatory Regions track, specifically the sub-track for RNA Polymerase II (Pol2) ChIP-seq, exhibited a distinct peak that spatially coincided with both the H3K27ac enrichment and the “Promoter-like” cCREs in the DDX11L2 first intron. This direct binding of Pol2 is a definitive indicator of transcriptional machinery engagement at this site.

    3.4 Open Chromatin State:

    The presence of active histone marks and Pol2 binding strongly implies an open chromatin configuration. Examination of the DNase I Hypersensitivity Clusters track reveals a corresponding peak, further supporting the accessibility of this region for transcription factor binding and initiation.

    4. Discussion

    The integrated genomic data from the UCSC Genome Browser provides compelling evidence for an active alternative promoter within the first intron of the human DDX11L2 gene. The co-localization of “Promoter-like” cCREs, robust H3K27ac signals, and direct RNA Polymerase II binding collectively demonstrates that this region is actively engaged in transcriptional initiation.

    The classification of cCREs as “promoter-like” (red bars) is based on a sophisticated integration of multiple ENCODE assays, reflecting a comprehensive biochemical signature of active promoters. The specific experimental identifiers associated with these cCREs (e.g., ERG, TERF2, SMC3 ChIP-seq data) highlight the diverse array of transcription factors that can bind to and contribute to the regulatory activity of a promoter. While ERG, TERF2, and SMC3 are not RNA Pol II itself, their presence at this locus, in conjunction with Pol II binding and active histone marks, indicates a complex regulatory network orchestrating transcription from this alternative promoter.

    The strong H3K27ac peak serves as a critical epigenetic signature, reinforcing the active state of this promoter. H3K27ac marks regions of open chromatin that are poised for, or actively undergoing, transcription. Its direct overlap with Pol II binding further strengthens the assertion of active transcription initiation.

    The direct observation of RNA Polymerase II binding is the most definitive evidence for transcriptional initiation. Pol II is the core enzyme responsible for synthesizing messenger RNA (mRNA) and many non-coding RNAs. Its presence at a specific genomic location signifies that the cellular machinery for transcription is assembled and active at that site.

    The findings are particularly interesting given that DDX11L2 is annotated as a pseudogene. This study adds to the growing body of literature demonstrating that pseudogenes, traditionally considered genomic “fossils,” can acquire or retain functional regulatory roles, including acting as active promoters for non-coding RNAs or influencing the expression of neighboring genes [2]. The presence of an active alternative promoter within DDX11L2 suggests a more intricate regulatory landscape than implied by its pseudogene annotation alone.

    5. Conclusion

    Through the integrated analysis of ENCODE and ReMap data on the UCSC Genome Browser, this study provides strong evidence that an intronic region within the human DDX11L2 gene functions as an active alternative promoter. The co-localization of “Promoter-like” cCREs, high H3K27ac enrichment, and direct RNA Polymerase II binding collectively confirms active transcriptional initiation at this locus. These findings contribute to our understanding of the complex regulatory architecture of the human genome and highlight the functional potential of regions, such as pseudogenes, that may have been previously overlooked.

    References

    [1] Carninci P. and Tagami H. (2014). The FANTOM5 project and its implications for mammalian biology. F1000Prime Reports, 6: 104.

    [2] Poliseno L. (2015). Pseudogenes: Architects of complexity in gene regulation. Current Opinion in Genetics & Development, 31: 79-84.

    [3] Tomkins J.P. (2013). Alleged Human Chromosome 2 “Fusion Site” Encodes an Active DNA Binding Domain Inside a Complex and Highly Expressed Gene—Negating Fusion. Answers Research Journal, 6: 367–375. (Note: While this paper was a starting point, the current analysis uses independent data for verification).

    [4] ENCODE Project Consortium. (2012). An integrated encyclopedia of DNA elements in the human genome. Nature, 489(7414): 57–74.

    [5] Rada-Iglesias A., et al. (2011). A unique chromatin signature identifies active enhancers and genes in human embryonic stem cells. Nature Cell Biology, 13(9): 1003–1013.

    [6] Chèneby J., et al. (2018). ReMap 2018: an updated atlas of regulatory regions from an integrative analysis of DNA-binding ChIP-seq experiments. Nucleic Acids Research, 46(D1): D267–D275.

  • The Incoherence of Naturalism

    The Incoherence of Naturalism

    Introduction

    Naturalism—the philosophical position that reality consists entirely of natural entities governed by natural laws—presents itself as the most rational and empirically grounded worldview. Yet despite its scientific veneer, naturalism suffers from foundational incoherence that undermines its viability as a comprehensive philosophy.

    This critique demonstrates that naturalism is self-defeating, arbitrarily restrictive, explanatorily inadequate, and internally inconsistent. Each of these failings stems not from temporary limitations in scientific knowledge but from structural contradictions within naturalism itself. Together, they render naturalism philosophically untenable and point toward the necessity of a more pluralistic metaphysical framework.

    Premise 1: Self-Defeating Foundations

    Naturalism’s first fatal flaw lies in its inability to justify its own foundations without circularity or special pleading.

    Scientific inquiry rests on several non-empirical assumptions that cannot be empirically verified: the reliability of human reason, the uniformity of nature, the correspondence between our perceptions and external reality, and principles like logical consistency and parsimony. These assumptions cannot be proven through scientific methods—they are preconditions for scientific inquiry itself.

    This creates an insurmountable problem for naturalism. If reality consists entirely of natural entities governed by natural laws, then human cognition is merely the product of evolutionary processes that selected for survival value, not truth-tracking capacity. As philosopher Alvin Plantinga argues, if our cognitive faculties evolved primarily for reproductive fitness rather than truth detection, we have no reason to trust them for accurately grasping metaphysical truths like naturalism itself.

    The naturalist might counter that evolutionary adaptiveness correlates with truth-tracking, particularly regarding immediate environmental threats. But this defense fails to bridge the gap between adaptive perceptual reliability and justified abstract metaphysical beliefs. There is no evolutionary advantage to having accurate beliefs about quantum mechanics, consciousness, or cosmic origins. Natural selection has no mechanism to select for metaphysical accuracy.

    This creates what philosopher Thomas Nagel calls an “intolerable conflict” in naturalism—it relies on rational faculties that, by its own account, evolved for survival rather than metaphysical accuracy. The naturalist faces what Barry Stroud terms “irrecoverable circularity”: they must presuppose the reliability of faculties whose reliability they then try to explain through evolutionary processes.

    Even sophisticated attempts to escape this circularity through epistemic externalism merely shift the problem. Reliabilism claims beliefs formed through reliable processes are justified regardless of whether we can prove their reliability. But this begs the question: how do we establish which processes are reliable without presupposing their reliability? At some point, non-empirical axioms must be accepted on non-natural grounds.

    If naturalists retreat to pragmatism, accepting these axioms as “useful fictions” rather than truths, they have conceded that naturalism cannot justify its foundations within its own framework. This pragmatism is itself a non-empirical philosophical commitment that naturalism can neither justify nor dispense with.

    Premise 2: Arbitrary Restriction of Inquiry

    Naturalism’s second critical flaw lies in its arbitrary restriction of legitimate inquiry to natural causes alone.

    Philosophical naturalism makes an unwarranted leap from methodological naturalism (the practical scientific approach of seeking natural causes) to a metaphysical claim that only natural causes exist. This represents a category error—moving from a useful methodological heuristic to an ontological assertion without sufficient justification.

    By defining reality exclusively in terms of what natural science can study, naturalism creates a self-fulfilling prophecy: it finds only natural causes because it defines all discoverable causes as natural by definition. This circular approach prejudices investigation rather than allowing evidence to determine the boundaries of reality.

    The most powerful demonstration of this limitation is consciousness. Despite tremendous advances in neuroscience, the qualitative character of subjective experience—what philosopher Thomas Nagel calls the “what it is like” aspect of consciousness—resists reduction to physical processes. Neuroscience can correlate neural activity with reported experiences but cannot explain why these physical processes are accompanied by subjective experience at all.

    This limitation isn’t temporary but structural—scientific methods are designed to study third-person observable phenomena, not first-person subjectivity. The scientific method, by its very nature, abstracts away subjective qualities to focus on quantifiable properties. This creates what philosopher David Chalmers calls the “hard problem” of consciousness—explaining how and why physical processes give rise to subjective experience.

    Naturalists often respond by incorporating consciousness as a “fundamental” feature of an expanded natural framework—what Chalmers calls “naturalistic dualism.” But this semantic maneuver doesn’t resolve the ontological problem. If consciousness is fundamental and irreducible to physical processes, then reality includes non-physical properties—precisely what traditional naturalism denies. This exhibits what philosopher William Hasker calls “naturalism of the gaps”—expanding the definition of “natural” to encompass whatever resists reduction.

    Unlike historical examples like electromagnetism or vitalism, which were unexplained physical phenomena eventually incorporated into expanded physical frameworks, consciousness presents a categorically different challenge—explaining how subjective experience arises from objective processes. This isn’t merely an unexplained mechanism but a conceptual chasm between fundamentally different categories of reality.

    Premise 3: Explanatory Gaps

    Naturalism’s third major flaw lies in its persistent failure to explain fundamental aspects of human experience, despite centuries of scientific progress.

    Beyond consciousness, naturalism struggles to account for several phenomena central to human existence:

    Intentionality: The “aboutness” of mental states—the fact that thoughts, beliefs, and desires are about something beyond themselves—resists physical reduction. Physical states aren’t intrinsically “about” anything; they simply are. Yet our mental states exhibit this irreducible directedness toward objects, concepts, and possibilities. Philosopher Franz Brentano identified intentionality as the defining characteristic of mental phenomena, creating an explanatory gap that naturalism has failed to bridge.

    Rationality: Logical relationships between propositions aren’t physical connections but normative ones—they describe how we ought to reason, not merely how matter behaves. The laws of logic and mathematics exhibit a necessity that natural laws lack. Natural laws describe contingent regularities that could theoretically be otherwise; logical laws express necessary truths that couldn’t possibly be different. This modal difference creates another category distinction that naturalism struggles to accommodate.

    Morality: Moral imperatives involve inherent “ought” claims that cannot be derived from purely descriptive “is” statements. As philosopher G.E. Moore identified, any attempt to define moral properties in natural terms commits the “naturalistic fallacy.” Evolutionary accounts may explain the origins of moral psychology but cannot justify moral claims as true or authoritative. If moral judgments are merely evolutionary adaptations, their normative force is undermined, creating what philosopher Sharon Street calls the “Darwinian Dilemma.”

    Naturalists often respond to these gaps through eliminativism or emergentism. Eliminativism denies the reality of these phenomena, claiming they are illusions or folk-psychological confusions. But this approach is self-defeating—an illusion of consciousness must be experienced by someone, making consciousness inescapable. As philosopher John Searle notes, “Where consciousness is concerned, the appearance is the reality.”

    Emergentism fares no better. To claim consciousness “emerges” from physical processes without explaining the mechanism of emergence merely restates the mystery. Unlike other emergent properties (like liquidity emerging from H₂O molecules), consciousness involves a transition from objective processes to subjective experience—a categorical leap, not a continuous spectrum. The naturalist must explain how arrangement of non-conscious particles yields consciousness, a challenge philosopher Colin McGinn calls “cognitive closure.”

    These explanatory gaps aren’t temporary limitations in scientific knowledge but principled barriers arising from naturalism’s restricted ontology. After centuries of scientific progress, these gaps remain as profound as ever, suggesting a fundamental inadequacy in naturalism’s conceptual resources.

    Premise 4: Inconsistent Verification

    Naturalism’s fourth fatal flaw lies in its criterion for knowledge, which cannot justify itself without inconsistency.

    The naturalist privileges empirical verification—the idea that meaningful claims must be empirically testable. Yet this verification principle itself cannot be empirically verified. It is a philosophical position, not a scientific discovery. This creates an internal contradiction: if we accept only what can be demonstrated through scientific methods, we must reject the very principle that demands such verification.

    Even if naturalists reject strict verificationism, they still privilege empirical evidence above all else. Yet this privileging itself cannot be empirically justified. It’s a meta-empirical value judgment about what counts as legitimate evidence—precisely the kind of non-empirical philosophical commitment that naturalism struggles to account for.

    Attempts to resolve this inconsistency through naturalized epistemology (following Quine) don’t solve the problem—they institutionalize it. Treating epistemology as a branch of psychology assumes the reliability of the psychological methods used to study epistemology. This creates what philosopher Laurence BonJour calls “meta-justification”—how do we justify our justificatory framework? Naturalized epistemology ultimately relies on pragmatic success, but this pragmatism itself requires non-empirical criteria for what constitutes “success.”

    Even if we accept Quine’s web of belief, some strands in the web must be anchored independently of empirical verification. These include logical principles, mathematical truths, and the assumption that reality is comprehensible. These principles aren’t empirically derived but are preconditions for empirical inquiry. Their necessity reveals naturalism’s dependence on non-natural foundations.

    Naturalism thus faces an inescapable dilemma: either it consistently applies its verification standards and undermines its own foundations, or it makes special exceptions for its core principles and thereby acknowledges limits to its explanatory scope.

    The Inescapable Dilemma of Naturalism

    These four premises reveal that naturalism faces an inescapable dilemma:

    1. Strict naturalism maintains a coherent ontology (only physical entities exist) but fails to account for consciousness, intentionality, rationality, and its own foundations.
    2. Expanded naturalism accommodates these phenomena but sacrifices coherence by stretching “natural” to include fundamentally non-physical properties.

    This isn’t merely a limitation of current knowledge but a structural impossibility within naturalism’s framework. The problem isn’t that naturalism hasn’t yet explained consciousness; it’s that consciousness is categorically different from physical processes, requiring explanatory principles that transcend physical causation.

    A “richer naturalism” that embraces consciousness as fundamental, accepts non-empirical axioms pragmatically, and incorporates abstract objects has abandoned naturalism’s core thesis that reality consists entirely of natural entities subject to natural laws. This isn’t evolution of inquiry but conceptual surrender.

    Beyond Naturalism: The Case for Metaphysical Pluralism

    The most coherent alternative to naturalism is metaphysical pluralism—recognizing that reality includes physical processes, conscious experience, abstract entities, and normative truths, without reducing any to the others.

    This pluralistic approach acknowledges that different domains of reality require appropriate methods of investigation:

    1. Physical phenomena are best studied through empirical scientific methods
    2. Conscious experience requires phenomenological approaches that honor subjectivity
    3. Logical and mathematical truths demand rational analysis independent of empirical verification
    4. Normative questions involve philosophical reflection on values, not merely empirical facts

    Unlike naturalism, pluralism doesn’t face self-defeat (it can ground rational faculties non-circularly), doesn’t arbitrarily restrict inquiry (it allows appropriate methods for different domains), and doesn’t face explanatory gaps (it acknowledges irreducible categories without eliminating them).

    Naturalists often appeal to Ockham’s Razor (parsimony) and the practical success of science (pragmatism) as reasons to prefer naturalism over more metaphysically rich views like pluralism. However, as your text implicitly and explicitly argues, these critiques are problematic when leveled by the naturalist themselves, given the internal difficulties naturalism faces.

    1. Problems with the Parsimony Critique:

    • False Parsimony: Naturalism’s claim to parsimony often amounts to ontological stinginess achieved by explanatory inadequacy. It claims to be simpler by positing only one fundamental kind of “stuff” (natural/physical). However, as your text details, this simplicity is bought at the cost of being unable to adequately account for or integrate crucial aspects of reality like consciousness, intentionality, rationality, and normativity (Premises 2 & 3). A theory that is simple but leaves vast swathes of reality unexplained is not genuinely more parsimonious than a theory that posits more fundamental categories but can actually explain or accommodate all the relevant phenomena. True parsimony should be measured not just by the number of types of entities posited, but by the overall complexity of the explanatory framework required to account for the data. Pluralism, by assigning different phenomena to different appropriate categories, might require a more diverse ontology but arguably a less strained and more comprehensive explanatory structure than naturalism, which must resort to eliminativism, mysterious emergence, or redefining terms to handle outliers.
    • Redefining “Natural” Undermines Parsimony: Your text notes that naturalists trying to accommodate phenomena like consciousness might resort to calling it a “fundamental feature” within an “expanded naturalism” or “naturalistic dualism.” This is an attempt to absorb irreducible phenomena by broadening the definition of “natural.” But this move itself adds fundamental categories or properties to the naturalist ontology. If “natural” now includes irreducible subjective experience or fundamental abstract objects, the initial claim to radical simplicity (“only physical stuff”) is surrendered. This “naturalism of the gaps” (as your text puts it) demonstrates that naturalists, when pressed, do feel the need to add fundamental categories, thereby undermining their own parsimony argument against pluralism.
    • Parsimony Itself is a Non-Empirical Principle: Ockham’s Razor is a meta-scientific or philosophical principle guiding theory choice. It’s not something discovered through empirical science. As your text argues in Premise 4, naturalism struggles to justify such non-empirical principles within its own framework. If the naturalist insists that all legitimate knowledge must be empirically verifiable or grounded, they face a difficulty in appealing to a principle like parsimony, which is a criterion of theoretical virtue, not an empirical fact. Using parsimony to critique pluralism requires the naturalist to step outside their own purported empirical-only standard, or at least rely on a principle they cannot ground naturally.  

    2. Problems with the Pragmatism Critique:

    • Conflation of Methodological and Metaphysical Pragmatism: Naturalists often point to the undeniable success of science (which operates using methodological naturalism – seeking natural explanations within its domain) as evidence for metaphysical naturalism (the philosophical claim that only natural things exist). As your text argues in Premise 2, this is a category error. Methodological naturalism is pragmatic for the specific goal of studying the physical world empirically. Metaphysical naturalism is a comprehensive worldview claim. The pragmatism of the former doesn’t automatically transfer to the latter. Pluralism fully embraces methodological naturalism for understanding the physical realm but recognizes that other realms (subjective experience, logic, morality) require different, though equally valid, approaches.  
    • Pragmatism for What Purpose? If pragmatism means “what works as a comprehensive worldview,” then naturalism is arguably not pragmatic because it fails to provide a coherent or satisfactory account of fundamental aspects of human reality (consciousness, meaning, values, reason’s validity), as detailed in Premise 3. It might be pragmatic for building bridges or predicting planetary motion, but it’s arguably deeply unpragmatic for understanding what it means to be a conscious, rational, moral agent in a world with objective truths. Pluralism, by acknowledging different domains and methods, is arguably more pragmatically successful as a philosophical framework because it provides conceptual resources to engage meaningfully with the full spectrum of human experience and inquiry, not just the physically quantifiable parts.
    • Naturalism May Rely on Pragmatism for its Own Foundations: Your text suggests (Premises 1 & 4) that naturalists, when pushed on how they justify the reliability of reason or the empirical method itself, might retreat to a pragmatic defense (“these methods just work”). If naturalism must ultimately appeal to pragmatism to ground its own core principles, it’s in a weak position to then turn around and critique pluralism solely on pragmatic grounds, especially when pluralism can argue it is more pragmatically successful in making sense of all of reality. This creates a kind of “pragmatism of the gaps” where pragmatism is invoked precisely where naturalism’s internal justification fails.

    In summary, the naturalist critiques of pluralism based on parsimony and pragmatism often miss the mark. Naturalism’s parsimony is frequently achieved by ignoring significant data or by subtly expanding its ontology, undermining the claim to unique simplicity. Its appeal to pragmatism often confuses the success of scientific method (which pluralism utilizes) with the philosophical adequacy of metaphysical naturalism as a total worldview, and ignores naturalism’s own potential reliance on pragmatic grounds it cannot fully justify. Pluralism, while positing a richer ontology, can argue it offers a more genuinely explanatory parsimony and a more comprehensive pragmatism by acknowledging the irreducible complexity of reality.

    Metaphysical pluralism doesn’t entail supernaturalism or theism by necessity. One can reject both naturalism and supernaturalism by acknowledging that reality may include non-physical aspects (consciousness, mathematical truths, values) without positing supernatural entities. Philosophers like Thomas Nagel, John Searle, and David Chalmers have developed non-materialist frameworks that don’t entail theism.

    Conclusion

    Naturalism fails as a comprehensive worldview. Its success in explaining physical phenomena doesn’t justify its extension to all aspects of reality. Its persistent explanatory gaps in consciousness, rationality, and value—coupled with its inability to justify its own foundations—reveal its fundamental inadequacy.

    A truly rational approach follows evidence where it leads, even when it points beyond the boundaries of naturalistic explanation. This isn’t an abandonment of rationality but its fulfillment—acknowledging that different aspects of reality may require different modes of understanding.

    Metaphysical pluralism offers a more coherent framework that honors the multidimensional character of reality. It maintains the empirical rigor of science within its proper domain while recognizing that human experience encompasses dimensions that transcend physical reduction. In doing so, it avoids both the reductionism of strict naturalism and the supernaturalism it rightly criticizes, providing a middle path that better accounts for the full spectrum of human knowledge and experience.

  • An Argument for Agent Causation in the Origin of DNA’s Information

    An Argument for Agent Causation in the Origin of DNA’s Information

    NOTE: This is a design argument inspired by Stephen Meyer‘s design argument from DNA. Importantly, specified complexity is changed for semiotic code (which I feel is more precise) and intelligent design is changed to agent causation (which is more preferencial).

    This argument posits that the very nature of the information encoded in DNA, specifically its structure as a semiotic code, necessitates an intelligent cause in its origin. The argument proceeds by establishing two key premises: first, that semiotic codes inherently require intelligent (agent) causation for their creation, and second, that DNA functions as a semiotic code.

    Premise 1: The Creation of a Semiotic Code Requires Agent Causation (Intelligence)

    A semiotic code is a system designed for conveying meaning through the use of signs. At its core, a semiotic code establishes a relationship between a signifier (the form the sign takes, e.g., a word, a symbol, a sequence) and a signified (the concept or meaning represented). Crucially, in a semiotic code, this relationship is arbitrary or conventional, not based on inherent physical or chemical causation between the signifier and the signified. This requires an interpretive framework – a set of rules or a system – that is independent of the physical properties of the signifier itself, providing the means to encode and decode the meaning. The meaning resides not in the physical signal, but in its interpretation according to the established code.

    Consider examples like human language, musical notation, or traffic signals. The sound “stop” or the sequence of letters S-T-O-P has no inherent physical property that forces a vehicle to cease motion. A red light does not chemically or physically cause a car to stop; it is a conventionally assigned symbol that, within a shared interpretive framework (traffic laws and driver understanding), signifies a command to stop. This is distinct from a natural sign, such as smoke indicating fire. In this case, the relationship between smoke and fire is one of direct, necessary physical causation (combustion produces smoke). While an observer can interpret smoke as a sign of fire, the connection itself is a product of natural laws, existing independently of any imposed code or interpretive framework.

    The capacity to create and utilize a system where arbitrary symbols reliably and purposefully convey specific meanings requires more than just physical processes. It requires the ability to:

    Conceive of a goal: To transfer specific information or instruct an action.

    Establish arbitrary conventions: To assign meaning to a form (signifier) where no inherent physical link exists to the meaning (signified).

    Design an interpretive framework: To build or establish a system of rules or machinery that can reliably encode and decode these arbitrary relationships.

    Implement this system for goal-directed action: To use the code and framework to achieve the initial goal of information transfer and subsequent action based on that information.

    This capacity to establish arbitrary, rule-governed relationships for the purpose of communication and control is what we define as intelligence in this context. The creation of a semiotic code is an act of imposing abstract order and meaning onto physical elements according to a plan or intention. Such an act requires agent causation – causation originating from an entity capable of intentionality, symbolic representation, and the design of systems that operate based on abstract rules, rather than solely from the necessary interactions of physical forces (event causation).

    Purely natural, undirected physical processes can produce complex patterns and structures driven by energy gradients, chemical affinities, or physical laws (like crystal formation, which is a direct physical consequence of electrochemical forces and molecular structure, lacking arbitrary convention, an independent interpretive framework, or symbolic representation). However, they lack the capacity to establish arbitrary conventions where the link between form and meaning is not physically determined, nor can they spontaneously generate an interpretive framework that operates based on such non-physical rules for goal-directed purposes. Therefore, the existence of a semiotic code, characterized by arbitrary signifier-signified links and an independent interpretive framework for goal-directed information transfer, provides compelling evidence for the involvement of intelligence in its origin.

    Premise 2: DNA Functions as a Semiotic Code

    The genetic code within DNA exhibits the key characteristics of a semiotic code as defined above. Sequences of nucleotides (specifically, codons on mRNA) act as signifiers. The signifieds are specific amino acids, which are the building blocks of proteins.

    Crucially, the relationship between a codon sequence and the amino acid it specifies is not one of direct chemical causation. A codon (e.g., AUG) does not chemically synthesize or form the amino acid methionine through a direct physical reaction dictated by the codon’s molecular structure alone. Amino acid synthesis occurs through entirely separate biochemical pathways involving dedicated enzymes.

    Instead, the codon serves as a symbolic signal that is interpreted by the complex cellular machinery of protein synthesis – the ribosomes, transfer RNAs (tRNAs), and aminoacyl-tRNA synthetases. This machinery constitutes the interpretive framework.

    Here’s how it functions as a semiotic framework:

    • Arbitrary/Conventional Relationship: The specific assignment of a codon triplet to a particular amino acid is largely a matter of convention. While there might be some historical or biochemical reasons that biased the code’s evolution, the evidence from synthetic biology, where scientists have successfully engineered bacteria with different codon-amino acid assignments, demonstrates that the relationship is not one of necessary physical linkage but of an established (and in this case, artificially modified) rule or convention. Different codon assignments could work, but the system functions because the cellular machinery reliably follows the established rules of the genetic code.
    • Independent Interpretive Framework: The translation machinery (ribosome, tRNAs, synthetases) is a complex system that reads the mRNA sequence (signifier) and brings the correct amino acid (signified) to the growing protein chain, according to the rules encoded in the structure and function of the tRNAs and synthetases. The meaning (“add this amino acid now”) is not inherent in the chemical properties of the codon itself but resides in how the interpretive machinery is designed to react to that codon. This machinery operates independently of direct physical causation by the codon itself to create the amino acid; it interprets the codon as an instruction within the system’s logic.
    • Symbolic Representation: The codon stands for an amino acid; it is a symbol representing a unit of meaning within the context of protein assembly. The physical form (nucleotide sequence) is distinct from the meaning it conveys (which amino acid to add). This is analogous to the word “cat” representing a feline creature – the sound or letters don’t physically embody the cat but symbolize the concept.

    Therefore, DNA, specifically the genetic code and the translation system that interprets it, functions as a sophisticated semiotic code. It involves arbitrary relationships between signifiers (codons) and signifieds (amino acids), mediated by an independent interpretive framework (translation machinery) for the purpose of constructing functional proteins (goal-directed information transfer).

    Conclusion: Therefore, DNA Requires Agent Causation in its Origin

    Based on the premises established:

    1. The creation of a semiotic code, characterized by arbitrary conventions, an independent interpretive framework, and symbolic representation for goal-directed information transfer, requires the specific capacities associated with intelligence and agent causation (intentionality, abstraction, rule-creation, system design).
    2. DNA, through the genetic code and its translation machinery, functions as a semiotic code exhibiting these very characteristics.

    It logically follows that the origin of DNA’s semiotic structure requires agent causation. The arbitrary nature of the code assignments and the existence of a complex system specifically designed to read and act upon these arbitrary rules, independent of direct physical necessity between codon and amino acid, are hallmarks of intelligent design, not the expected outcomes of undirected physical or chemical processes.

    Addressing Potential Objections:

    • Evolution and Randomness: While natural selection can act on variations in existing biological systems, it requires a self-replicating system with heredity – which presupposes the existence of a functional coding and translation system. Natural selection is a filter and modifier of existing information; it is not a mechanism for generating a semiotic code from scratch. Randomness, by definition, lacks the capacity to produce the specified, functional, arbitrary conventions and the integrated interpretive machinery characteristic of a semiotic code. The challenge is not just sequence generation, but the origin of the meaningful, rule-governed relationship between sequences and outcomes, and the system that enforces these rules.
    • “Frozen Accident” and Abiogenesis Challenges: Hypotheses about abiogenesis and early life (like the RNA world) face significant hurdles in explaining the origin of this integrated semiotic system. The translation machinery is a highly complex and interdependent system (a “chicken-and-and egg” problem where codons require tRNAs and synthetases to be read, but tRNAs and synthetases are themselves encoded by and produced through this same system). The origin of the arbitrary codon-amino acid assignments and the simultaneous emergence of the complex machinery to interpret them presents a significant challenge for gradual, undirected assembly driven solely by chemical or physical affinities.
    • Biochemical Processes vs. Interpretation: The argument does not claim that a ribosome is a conscious entity “interpreting” in the human sense. Instead, it argues that the system it is part of (the genetic code and translation machinery) functions as an interpretive framework because it reads symbols (codons) and acts according to established, arbitrary rules (the genetic code’s assignments) to produce a specific output (amino acid sequence), where this relationship is not based on direct physical necessity but on a mapping established by the code’s design. This rule-governed, symbolic mapping, independent of physical causation between symbol and meaning, is the defining feature of a semiotic code requiring an intelligence to establish the rules and the system.
    • God-of-the-Gaps: This argument is not based on mere ignorance of a natural explanation. It is a positive argument based on the nature of the phenomenon itself. Semiotic codes, wherever their origin is understood (human language, computer code), are the products of intelligent activity involving the creation and implementation of arbitrary conventions and interpretive systems for goal-directed communication. The argument posits that DNA exhibits these defining characteristics and therefore infers a similar type of cause in its origin, based on a uniformity of experience regarding the necessary preconditions for semiotic systems.

    In conclusion, the sophisticated, arbitrary, and rule-governed nature of the genetic code and its associated translation machinery point to it being a semiotic system. Based on the inherent requirements for creating such a system—namely, the capacities for intentionality, symbolic representation, rule-creation, and system design—the origin of DNA’s information is best explained by the action of an intelligent agent.

  • Chromosome 2 Fusion: Evidence Out Of Thin Air?

    Chromosome 2 Fusion: Evidence Out Of Thin Air?

    The story is captivating and frequently told in biology textbooks and popular science: humans possess 46 chromosomes while our alleged closest relatives, chimpanzees and other great apes, have 48. The difference, evolutionists claim, is due to a dramatic event in our shared ancestry – the fusion of two smaller ape chromosomes to form the large human Chromosome 2. This “fusion hypothesis” is often presented as slam-dunk evidence for human evolution from ape-like ancestors. But when we move beyond the narrative and scrutinize the actual genetic data, does the evidence hold up? A closer look suggests the case for fusion is far from conclusive, perhaps even bordering on evidence conjured “out of thin air.”

    The fusion model makes specific predictions about what we should find at the junction point on Chromosome 2. If two chromosomes, capped by protective telomere sequences, fused end-to-end, we’d expect to see a characteristic signature: the telomere sequence from one chromosome (repeats of TTAGGG) joined head-to-head with the inverted telomere sequence from the other (repeats of CCCTAA). These telomeric repeats typically number in the thousands at chromosome ends.  

    The Missing Telomere Signature

    When scientists first looked at the proposed fusion region (locus 2q13), they did find some sequences resembling telomere repeats (IJdo et al., 1991). This was hailed as confirmation. However, the reality is much less convincing than proponents suggest.

    Instead of thousands of ordered repeats forming a clear TTAGGG…CCCTAA structure, the site contains only about 150 highly degraded, degenerate telomere-like sequences scattered within an ~800 base pair region. Searching a much larger 64,000 base pair region yields only 136 instances of the core TTAGGG hexamer, far short of a telomere’s structure. Crucially, the orientation is often wrong – TTAGGG motifs appear where CCCTAA should be, and vice-versa. This messy, sparse arrangement hardly resembles the robust structure expected from even an ancient, degraded fusion event.

    Furthermore, creationist biologist Dr. Jeffrey Tomkins discovered that this alleged fusion site is not merely inactive debris; it falls squarely within a functional region of the DDX11L2 gene, likely acting as a promoter or regulatory element (Tomkins, 2013). Why would a supposedly non-functional scar from an ancient fusion land precisely within, and potentially regulate, an active gene? This finding severely undermines the idea of it being simple evolutionary leftovers.

    The Phantom Centromere

    A standard chromosome has one centromere. Fusing two standard chromosomes would initially create a dicentric chromosome with two centromeres – a generally unstable configuration. The fusion hypothesis thus predicts that one of the original centromeres must have been inactivated, leaving behind a remnant or “cryptic” centromere on Chromosome 2.  

    Proponents point to alpha-satellite DNA sequences found around locus 2q21 as evidence for this inactivated centromere, citing studies like Avarello et al. (1992) and the chromosome sequencing paper by Hillier et al. (2005). But this evidence is weak. Alpha-satellite DNA is indeed common near centromeres, but it’s also found abundantly elsewhere throughout the genome, performing various functions.  

    The Avarello study, conducted before full genome sequencing, used methods that detected alpha-satellite DNA generally, not functional centromeres specifically. Their results were inconsistent, with the signal appearing in less than half the cells examined – hardly the signature of a definite structure. Hillier et al. simply noted the presence of alpha-satellite tracts, but these specific sequences are common types found on nearly every human chromosome and show no unique similarity or phylogenetic clustering with functional centromere sequences. There’s no compelling structural or epigenetic evidence marking this region as a bona fide inactivated centromere; it’s simply a region containing common repetitive DNA.

    Uniqueness and the Mutation Rate Fallacy

    Adding to the puzzle, the specific short sequence often pinpointed as the precise fusion point isn’t unique. As can be demonstrated using the BLAT tool, this exact sequence appears on human Chromosomes 7, 19, and the X and Y chromosomes. If this sequence is the unique hallmark of the fusion event, why does it appear elsewhere? The evolutionary suggestion that these might be remnants of other, even more ancient fusions is pure speculation without a shred of supporting evidence.

    The standard evolutionary counter-argument to the lack of clear telomere and centromere signatures is degradation over time. “The fusion happened millions of years ago,” the reasoning goes, “so mutations have scrambled the evidence.” However, this explanation crumbles under the weight of actual mutation rates.

    Using accepted human mutation rate estimates (Nachman & Crowell, 2000) and the supposed 6-million-year timeframe since divergence from chimps, we can calculate that the specific ~800 base pair fusion region would be statistically unlikely to have suffered even one mutation during that entire period! The observed mutation rate is simply far too low to account for the dramatic degradation required to turn thousands of pristine telomere repeats and a functional centromere into the sequences we see today. Ironically, the known mutation rate argues against the degradation explanation needed to salvage the fusion hypothesis.

    Common Design vs. Common Ancestry

    What about the general similarity in gene order (synteny) between human Chromosome 2 and chimpanzee chromosomes 2A and 2B? While often presented as strong evidence for fusion, similarity does not automatically equate to ancestry. An intelligent designer reusing effective plans is an equally valid, if not better, explanation for such similarities. Moreover, the “near identical” claim is highly exaggerated; large and significant differences exist in gene content, control regions, and overall size, especially when non-coding DNA is considered (Tomkins, 2011, suggests overall similarity might be closer to 70%). This makes sense when considering that coding regions function to provide the recepies for proteins (which similar life needs will share similarly).

    Conclusion: A Story Of Looking for Evidence

    When the genetic data for human Chromosome 2 is examined without the pre-commitment to an evolutionary narrative, the evidence for the fusion event appears remarkably weak. So much so that it begs the question, was this a mad-dash to explain the blatent differences in the genomes of Humans and Chimps? The expected telomere signature is absent, replaced by a short, jumbled sequence residing within a functional gene region. The evidence for a second, inactivated centromere relies on the presence of common repetitive DNA lacking specific centromeric features. The supposed fusion sequence isn’t unique, and known mutation rates are woefully insufficient to explain the degradation required by the evolutionary model over millions of years.

    The chromosome 2 fusion story seems less like a conclusion drawn from compelling evidence and more like an interpretation imposed upon ambiguous data to fit a pre-existing belief in human-ape common ancestry. The scientific data simply does not support the narrative. Perhaps it’s time to acknowledge that the “evidence” for this iconic fusion event may indeed be derived largely “out of thin air.”

    References:

  • Examining Claims of Macroevolution and Irreducible Complexity:

    Examining Claims of Macroevolution and Irreducible Complexity:

    A Creationist Perspective

    The debate surrounding the origin and diversification of life continues, with proponents of neo-Darwinian evolution often citing observed instances of speciation and adaptations as evidence for macroevolution and the gradual development of complex biological systems. A recent “MEGA POST” on Reddit’s r/DebateEvolution presented several cases purported to demonstrate these processes, challenging the creationist understanding of life’s history. This article will examine these claims from a young-Earth creationist viewpoint.

    The original post defined key terms, stating, “Macroevolution ~ variations in heritable traits in populations with multiple species over time. Speciation marks the start of macroevolution.” However, creationists distinguish between microevolution – variation and speciation within a created kind – and macroevolution – the hypothetical transition between fundamentally different kinds of organisms. While the former is observable and acknowledged, the latter lacks empirical support and the necessary genetic mechanisms.

    Alleged Cases of Macroevolution:

    The post presented eleven cases as evidence of macroevolution.

    1. Lizards evolving placentas: The observation of reproductive isolation in Zootoca vivipara with different modes of reproduction was highlighted. The author noted, “(This is probably my favourite example of the bunch, as it shows a highly non-trivial trait emerging, together with isolation, speciation and selection for the new trait to boot.)” From a creationist perspective, the development of viviparity within lizards likely involves the expression or modification of pre-existing genetic information within the lizard kind. This adaptation and speciation do not necessitate the creation of novel genetic information required for a transition to a different kind of organism.

    2. Fruit flies feeding on apples: The divergence of the apple maggot fly (Rhagoletis pomonella) into host-specific groups was cited as sympatric speciation. This adaptation to different host plants and the resulting reproductive isolation are seen as microevolutionary changes within the fruit fly kind, utilizing the inherent genetic variability.  

    3. London Underground mosquito: The adaptation of Culex pipiens f. molestus to underground environments was presented as allopatric speciation. The observed physiological and behavioral differences, along with reproductive isolation, are consistent with diversification within the mosquito kind due to environmental pressures acting on the existing gene pool.  

    4. Multicellularity in Green Algae: The lab observation of obligate multicellularity in Chlamydomonas reinhardtii under predation pressure was noted. The author stated this lays “the groundwork for de novo multicellularity.” While this is an interesting example of adaptation, the transition from simple coloniality to complex, differentiated multicellularity, as seen in plants and animals, requires a significant increase in genetic information and novel developmental pathways. The presence of similar genes across different groups could point to a common designer employing similar modules for diverse functions.  

    5. Darwin’s Finches, revisited 150 years later: Speciation in the “Big Bird lineage” due to environmental pressures was discussed. This classic example of adaptation and speciation on the Galapagos Islands demonstrates microevolutionary changes within the finch kind, driven by natural selection acting on existing variations in beak morphology.  

    6 & 7. Salamanders and Greenish Warblers as ring species: These examples of geographic variation leading to reproductive isolation were presented as evidence of speciation. While ring species illustrate gradual divergence, the observed changes occur within the salamander and warbler kinds, respectively, and do not represent transitions to fundamentally different organisms.  

    8. Hybrid plants and polyploidy: The formation of Tragopogon miscellus through polyploidy was cited as rapid speciation. The author noted that crossbreeding “exploits polyploidy…to enhance susceptibility to selection for desired traits.” Polyploidy involves the duplication of existing chromosomes and the combination of genetic material from closely related species within the plant kingdom. This mechanism facilitates rapid diversification but does not generate the novel genetic information required for macroevolutionary transitions.  

    9. Crocodiles and chickens growing feathers: The manipulation of gene expression leading to feather development in these animals was discussed. The author suggested this shows “how birds are indeed dinosaurs and descend within Sauropsida.” Creationists interpret the shared genetic toolkit and potential for feather development within reptiles and birds as evidence of a common design within a broader created kind, rather than a direct evolutionary descent in the Darwinian sense.  

    10. Endosymbiosis in an amoeba: The observation of a bacterium becoming endosymbiotic within an amoeba was presented as analogous to the origin of organelles. Creationists propose that organelles were created in situ with their host cells, designed for symbiotic relationships from the beginning. The observed integration is seen as a function of this initial design.

    11. Eurasian Blackcap: The divergence in migratory behavior and morphology leading towards speciation was highlighted. This represents microevolutionary adaptation within the bird kind in response to environmental changes.

    Addressing “Irreducible Complexity”:

    The original post also addressed the concept of irreducible complexity with five counter-examples.

    1. E. Coli Citrate Metabolism in the LTEE: The evolution of citrate metabolism was presented as a refutation of irreducible complexity. The author noted that this involved “gene duplication, and the duplicate was inserted downstream of an aerobically-active promoter.” While this demonstrates the emergence of a new function, it occurred within the bacterial kind and involved the modification and duplication of existing genetic material. Therefore, is no evidence here to suggest an evolutionary pathway for the origin of citrate metabolism.

    2. Tetherin antagonism in HIV groups M and O: The different evolutionary pathways for overcoming tetherin resistance were discussed. Viruses, with their rapid mutation rates and unique genetic mechanisms, present a different case study than complex cellular organisms. This is not analogous in the slightest.

    3. Human lactose tolerance: The evolution of lactase persistence was presented as a change that is “not a loss of regulation or function.” This involves a regulatory mutation affecting the expression of an existing gene within the human genome. Therefore, it’s not a gain either. This is just a semantic game.

    4. Re-evolution of bacterial flagella: The substitution of a key regulatory protein for flagellum synthesis was cited. The author noted this is “an incredibly reliable two-step process.” While this demonstrates the adaptability of bacterial systems, the flagellum itself remains a complex structure with numerous interacting components – none of said components have gained or lost the cumulative necessary functions.

    5. Ecological succession: The development of interdependent ecosystems was presented as a challenge to irreducible complexity. However, ecological succession describes the interactions and development of communities of existing organisms, not the origin of the complex biological systems within those organisms.  

    Conclusion:

    While the presented cases offer compelling examples of adaptation and speciation, we interpret these observations as occurring within the boundaries of created kinds, utilizing the inherent genetic variability designed within them. These examples do not provide conclusive evidence for macroevolution – the transition between fundamentally different kinds of organisms – nor do they definitively refute the concept of irreducible complexity in the origin of certain biological systems. The fact that so many of these are, if not neutral, loss-of-function or loss-of-information mutations creates a compelling case for creation as the inference to the best explanation. The creationist model, grounded in the historical robustness of the Biblical account and supported by scientific evidence (multiple cross-disciplinary lines), offers a coherent alternative explanation for the diversity and complexity of life. As the original post concluded,

    “if your only response to the cases of macroevolution are ‘it’s still a lizard’, ‘it’s still a fly you idiot’ etc, congrats, you have 1) sorely missed the point and 2) become an evolutionist now!”

    However, the point is not that change doesn’t occur (we expect that on our model), but rather the kind and extent of that change, which, from a creationist perspective, remains within divinely established explanatory boundaries of the creation model and contradicts a universal common descent model.

    References:

    Teixeira, F., et al. (2017). The evolution of reproductive isolation during a rapid adaptive radiation in alpine lizards. Proceedings of the National Academy of Sciences, 114(12), E2386-E2393. https://doi.org/10.1073/pnas.1635049100

    Fonseca, D. M., et al. (2023). Rapid Speciation of the London Underground Mosquito Culex pipiens molestus. ResearchGate. https://doi.org/10.13140/RG.2.2.23813.22247

    Grant, P. R., & Grant, B. R. (2017). Texas A&M professor’s study of Darwin’s finches reveals species can evolve in two generations. Texas A&M Today. https://stories.tamu.edu/news/2017/12/01/texas-am-professors-study-of-darwins-finches-reveals-species-can-evolve-in-two-generations/

    Feder, J. L., et al. (1997). Allopatric host race formation in sympatric hawthorn maggot flies. Proceedings of the National Academy of Sciences, 94(15), 7761-7766. https://doi.org/10.1073/pnas.94.15.7761

    Tishkoff, S. A., et al. (2013). Convergent adaptation of human lactase persistence in Africa and Europe. Nature Genetics, 45(3), 233-240. https://doi.org/10.1038/ng.2529 (Note: While the URL provided redirects to PMC, the original publication is in Nature Genetics. I have cited the primary source.)

  • Tiny Water Fleas, Big Questions About Evolution

    Tiny Water Fleas, Big Questions About Evolution

    Scientists recently spent a decade tracking the genetics of a tiny water creature called Daphnia pulex, a type of water flea. What they found is stirring up a lot of questions about how evolution really works.  

    Imagine you’re watching a group of people over ten years, noting every little change in their appearance. Now, imagine doing that with the genetic code of hundreds of water fleas. That’s essentially what these researchers did. They looked at how the frequencies of different versions of genes (alleles) changed from year to year.

    What they discovered was surprising. On average, most of the genetic variations they tracked didn’t seem to be under strong selection at all. In other words, most of the time, the different versions of genes were more or less equally successful. It’s like watching people over ten years and finding that, on average, nobody’s hair color really changed much.

    However, there was a catch. Even though the average trend was “no change,” there were a lot of ups and downs from year to year. One year, a particular gene version might be slightly more common, and the next year, it might be slightly less common. This means that selective pressures—the forces that push evolution—were constantly changing.

    Think of it like the weather. One day it’s sunny, the next it’s rainy, but the average temperature over the year might be pretty mild. The researchers called this “fluctuating selection.”

    They also found that these genetic changes weren’t happening randomly across the whole genome. Instead, they were happening in small, linked groups of genes. These groups seemed to be working together, like little teams within the genome.  

    So, what does this all mean?

    Well, for one thing, it challenges the traditional idea of gradual, steady evolution via natural selection. If evolution were a slow, constant march forward, you’d expect to see consistent changes in gene frequencies over time being promoted by the environment. But that’s not what they found. Instead, they saw a lot of back-and-forth, with selection pressures constantly changing and equalizing at a net-zero.  

    From a design perspective, this makes a lot of sense. Instead of random changes slowly building up over millions of years, this data suggests that organisms are incredibly adaptable, designed to handle constant environmental shifts. The “teams” of linked genes working together look a lot like pre-programmed modules, ready to respond to whatever challenges the environment throws their way.

    The fact that most gene variations are “quasi-neutral,” meaning they don’t really affect survival on average, also fits with the idea of a stable, created genome. Rather than constantly evolving new features, organisms might be designed with a wide range of genetic options, ready to be used when needed.

    This study on tiny water fleas is a reminder that evolution is a lot more complex than we often think. It’s not just about random mutations and gradual changes. It’s about adaptability, flexibility, and a genome that’s ready for anything. And maybe, just maybe, it’s about design.

    (Based on: The genome-wide signature of short-term temporal selection)

  • How Created Heterozygosity Explains Genetic Variation

    How Created Heterozygosity Explains Genetic Variation

    A Conceptual Introduction:

    The study of genetics reveals a stunning tapestry of diversity within the living world. While evolutionary theory traditionally attributes this variation to random mutations accumulated over vast stretches of time, a creationist perspective offers a compelling alternative: Created Heterozygosity. This hypothesis proposes that God designed organisms with pre-existing genetic variability, allowing for adaptation and diversification within created kinds. This concept not only aligns with biblical accounts but also provides a more coherent explanation for observed genetic phenomena.

    The evolutionary narrative hinges on the power of mutations to generate novel genetic information. However, the overwhelming evidence points to the deleterious nature of most mutations. This can be seen in the famous Long-Term Evolutionary Experiments with E. coli. Notice, in the graphic below (Hofwegen, 2016), just how much information gets lost due to selection pressures and mutation. This is known as genetic entropy, the gradual degradation of the genome due to accumulated harmful mutations, poses a significant challenge to the idea that random mutations can drive the complexification of life. Furthermore, the sheer number of beneficial mutations required to explain the intricate design of living organisms strains credulity.

    “Genomic DNA sequencing revealed an amplification of the citT and dctA loci and DNA rearrangements to capture a promoter to express CitT, aerobically. These are members of the same class of mutations identified by the LTEE. We conclude that the rarity of the LTEE mutant was an artifact of the experimental conditions and not a unique evolutionary event. No new genetic information (novel gene function) evolved.”

    In contrast, Created Heterozygosity suggests that God, the master engineer, imbued organisms with a pre-programmed potential for variation. Just as human engineers design systems with built-in flexibility, God equipped his creation with the genetic resources necessary to adapt to diverse environments. This concept resonates with the biblical affirmation that God created organisms “according to their kinds,” implying inherent boundaries within which variation can occur. Recent research, such as the ENCODE project and studies on the dark proteome, has revealed an astonishing level of complexity and functionality within the genome, further supporting the idea of a designed system.

    Baraminology, the study of created kinds, provides empirical support for Created Heterozygosity. The rapid diversification observed within baramins, such as the canid or feline kinds, can be readily explained by the expression of pre-existing genetic information. For example, the diverse array of dog breeds can be traced back to the inherent genetic variability within the canine kind, rather than the accumulation of countless beneficial mutations.

    Of course, objections arise. The role of mutations in adaptation is often cited as evidence against Created Heterozygosity. However, certain mutations may represent the expression of designed backup systems or pre-programmed responses to environmental changes. Moreover, the vast majority of observed genetic variation can be attributed to the shuffling and expression of existing genetic information, rather than the creation of entirely new information.

    The implications for human genetics are profound. Created Heterozygosity elegantly explains the high degree of genetic variation within the human population, while remaining consistent with the biblical account of Adam and Eve as the progenitors of all humanity. Research on Mitochondrial Eve and Y-Chromosome Adam/Noah further supports the idea of a recent, common ancestry for all people.

    In conclusion, Created Heterozygosity provides a compelling framework for understanding genetic variation from a creationist perspective. By acknowledging the limitations of mutation-driven evolution and recognizing the evidence for designed diversity, we can appreciate the intricate wisdom of the Creator and the coherence of the biblical narrative. This concept invites us to explore the vastness of genetic diversity with a renewed sense of awe, recognizing the pre-programmed potential inherent in God’s magnificent creation.

    Citation:

    1. Van Hofwegen, D. J., Hovde, C. J., & Minnich, S. A. (2016). Rapid Evolution of Citrate Utilization by Escherichia coli by Direct Selection Requires citT and dctA. Journal of bacteriology, 198(7), 1022–1034.
  • The Limits of Evolution

    The Limits of Evolution

    Yesterday, a presentation by Dr. Rob Stadler took place on Dr. James Tour’s Youtube channel which has brought to light a compelling debate about the true extent of evolutionary capabilities. In their conversation, they delve into the levels of confidence in evolutionary evidence, revealing a stark contrast between observable, high-confidence microevolution and the extrapolated, low-confidence claims of macroevolutionary transitions. This distinction, which is based on the levels of evidence as understood in medical science, raises profound questions about the sufficiency of evolutionary mechanisms to explain the vast diversity of life.

    Dr. Stadler, author of “The Scientific Approach to Evolution,” presents a rigorous framework for evaluating scientific evidence. He outlines six criteria for high-confidence results: repeatability, direct measurability, prospectiveness, unbiasedness, assumption-free methodology, and reasonable claims. Applying these criteria to common evolutionary arguments, such as the fossil record, geographic distribution, vestigial organs, and comparative anatomy, Dr. Stadler reveals significant shortcomings. These lines of evidence, he argues, fall short of the high-confidence threshold. They are not repeatable, they cannot be directly measured, there is very little (if any) of predictive value , and most importantly they rely heavily on biased interpretation and assumption.

    However, the interview also highlights examples of high-confidence evolutionary studies. Experiments with E. coli bacteria, for instance, demonstrate the power of natural selection and mutation to drive small-scale changes within a population. These studies, repeatable and directly measurable, provide compelling evidence for microevolution. Yet, as Dr. Stadler emphasizes, extrapolating these observed changes to explain the origin of complex biological systems or the vast diversity of life is a leap of faith, not a scientific conclusion.

    The genetic differences between humans and chimpanzees further illustrate this point. While popular science often cites a 98% similarity, Dr. Stadler points out the significant differences, particularly in “orphan genes” and the regulatory functions of non-protein-coding DNA. These differences, he argues, challenge the notion of a simple, linear evolutionary progression.

    This aligns with the research of Dr. Douglas Axe, whose early work explored the probability of protein evolution. Axe’s findings suggest that the vast divergence between protein structures makes a common ancestor for all proteins highly improbable (Axe, 2000). This raises critical questions about the likelihood of orphan genes arising through random evolutionary processes alone, given the complexity and specificity of protein function.

    The core argument, as presented by Dr. Tour and Dr. Stadler, is not that evolution is entirely false. Rather, they contend that the high-confidence evidence supports only limited, small-scale changes, or microevolution. The leap to macroevolution, the idea that these small changes can accumulate to produce entirely new biological forms, appears to be a category error, based on our best evidence, and remains a low-confidence extrapolation.

    The video effectively presents case studies of evolution, demonstrating the observed limitations of evolutionary change. This evidence strongly suggests that evolutionary mechanisms are insufficient to account for the levels of diversity we observe today. The complexity of biological systems, the vast genetic differences between species, and the improbability of protein evolution challenge the core tenets of Neo-Darwinism and the Modern Synthesis.

    As Dr. Tour and Dr. Stadler articulate, a clear distinction must be made between observable, repeatable microevolution and the extrapolated, assumption-laden claims of macroevolution. While the former is supported by high-confidence evidence, the latter remains a subject of intense debate, demanding further scientific scrutiny.

    Works Cited

    • Tour, James, and Rob Stadler. “Evolution vs. Evidence: Are We Really 98% Chimp?” YouTube, uploaded by James Tour, https://www.youtube.com/watch?v=smTbYKJcnj8&t=2117s.
    • Axe, Douglas D. “Extreme functional sensitivity to conservative amino acid changes on enzyme exteriors.” Journal of Molecular Biology, vol. 301, no. 3, 2000, pp. 585-595.
  • My Top 5 Favorite Creation Podcasts

    My Top 5 Favorite Creation Podcasts

    As a creation enthusiast, I’m always on the lookout for resources that delve into the fascinating intersection of science and the biblical narrative. Podcasts have become a fantastic avenue for exploring these topics in depth, and I’ve curated a list of my top five favorites that consistently deliver insightful and engaging content.

    1. Let’s Talk Creation:

    This podcast is a gem for anyone seeking thoughtful and accessible discussions on creation science. Hosted by two PhD creationists, Todd Wood (baraminology) and Paul Garner (geology), “Let’s Talk Creation” offers bimonthly episodes that are both informative and easy to digest. What I appreciate most is their level-headed approach and their ability to break down complex scientific concepts into understandable terms. You’ll walk away from each episode with new insights and a deeper appreciation for the creation model.

    2. Standing For Truth:

    “Standing For Truth” is a powerhouse of creation content. With a vast database of interviews featuring subject experts from every relevant field, this podcast provides a comprehensive exploration of creation science. While it can get a little technical at times, the in-depth discussions and expert perspectives make it a valuable resource for those seeking a more rigorous understanding of the evidence.

    3. Creation Ministries International:

    For high-quality production and a wide variety of topics, “Creation Ministries International” delivers. Their videos are visually engaging and provide digestible explanations of creation science concepts including a wide range of scientists, philosophers, and theologians. While they may not always delve into the deepest technical details, their content is perfect for those seeking a solid overview of the evidence and its implications.

    4. Creation Unfolding:

    If you’re particularly interested in geology and paleontology, “Creation Unfolding” is a must-listen. The main host, Dr. K. P. Coulson, a well-researched geologist, brings a wealth of knowledge to the table, and the recurring guests provide diverse perspectives on these fascinating subjects. The laser-focused approach of this podcast makes it an invaluable resource for those seeking a deeper understanding of Earth’s history from a creationist perspective.

    5. Biblical Genetics:

    Dr. Robert Carter’s personal podcast, “Biblical Genetics,” is a treasure trove of information for anyone interested in the intersection of genetics and creation science. Dr. Carter, a renowned geneticist, tackles complex topics with clarity and precision, responding to popular-level content creators and professors with detailed explanations and analysis of technical papers. He skillfully guides listeners through intricate genetic concepts, making them accessible to a wider audience.


    These five podcasts represent a diverse range of perspectives and approaches to creation science. Whether you’re a seasoned creationist or just beginning to explore these topics, you’re sure to find valuable insights and engaging discussions within these podcasts.

  • Embryonic Similarities – Common Design, Not Common Descent

    Embryonic Similarities – Common Design, Not Common Descent

    For decades, textbook illustrations of Haeckel’s Embryos have been presented as a compelling visual argument for evolution. These side-by-side comparisons of vertebrate embryos, purportedly showing striking similarities in early developmental stages, have been used to argue for a shared evolutionary ancestry. However, a closer look reveals a story of misrepresentation and manipulation, rather than an accurate depiction of embryological evidence.

    Ernst Haeckel, a fervent supporter of Darwin’s theory, produced these drawings in the late 19th century. Yet, his illustrations were not faithful representations of actual embryos. He exaggerated similarities, omitted or altered developmental stages, and even used the same woodcut to represent different species. This deliberate manipulation aimed to bolster the concept of “recapitulation,” the now-discredited idea that embryonic development mirrors evolutionary history.

    The reality is that vertebrate embryos are far more distinct in their early stages than Haeckel portrayed. His illustrations were exposed as fraudulent even in his own time, yet they persisted in textbooks for generations, a testament to the power of visual propaganda in shaping scientific narratives.

    The argument that similarities in vertebrate embryos indicate a shared evolutionary history is challenged by several points.

    Challenging the “Recapitulation” Narrative

    One of the central tenets of the evolutionary argument is that embryonic development (“ontogeny”) reflects an organism’s evolutionary history (“phylogeny”). However, this concept, often summarized as “ontogeny recapitulates phylogeny,” is deeply flawed.

    • Embryonic Structures vs. Adult Structures: Embryonic features like pharyngeal slits and tails do not simply recapitulate the adult forms of ancestral organisms. Instead, they serve specific functions within the embryonic stage, often disappearing or transforming into entirely different structures in the adult. The embryonic mode of life is distinct from the adult mode.
    • “Recapitulation” is a Creationist Concept: The recognition of embryonic similarities predates Darwin. Creationists viewed these similarities as a “God-given ‘pattern of unification’ that reflected the unity of nature,” emphasizing a common Creator’s design rather than evolutionary lineage.
    • Unique Development: The unique eye development in lampreys, transitioning from larval eyespots to adult camera eyes, demonstrates that developmental pathways do not always follow a simple, linear evolutionary progression.
    • Order of Development: The occasional appearance of later-stage developmental features earlier in the embryonic process further complicates the evolutionary narrative.

    Genetic and Developmental Complexity

    The genetic and developmental complexity underlying embryonic similarities points to intelligent design:

    • Genetic Similarity: The fact that damage to the pax6 gene cascade results in the loss of a functional eye across diverse animal groups highlights a fundamental genetic similarity, but this similarity does not necessitate a shared evolutionary history. It speaks to a common design blueprint.
    • Complex Regulatory Systems: The development of complex structures like the eye involves thousands of interacting genes and intricate regulatory systems. Such complexity is more consistent with intelligent design than with random evolutionary processes.
    • Common Design: The similarities observed in vertebrate embryos can be readily explained as a reflection of a common design by an intelligent Creator. Just as an engineer might use similar design principles in different models, a Creator might employ common developmental strategies across various organisms.

    A Creationist Interpretation

    From a creationist perspective, the similarities in vertebrate embryos are not evidence of evolutionary transitions but rather manifestations of a unified design plan. The Creator used common design elements to achieve diverse functions in different organisms. This approach aligns with the concept of baraminology, which studies created kinds and acknowledges variations within those kinds.

    The argument that embryonic similarities exclusively support evolution overlooks the possibility of intelligent design. By recognizing the complexity of developmental processes and the historical context of these observations, we can appreciate the power of a creationist explanation.