Cinematic Strawberry

Logo

What is Meaning? A Predictive Universe Perspective


Abstract

This paper develops Predictive Landscape Semantics (PLS), a theoretical framework defining meaning functionally by applying core principles from the Predictive Universe (PU) framework. We first establish that information is a physically instantiated, substrate-independent pattern possessing the inherent potential to enable a suitable system to improve predictive accuracy about relevant states. Building on this, PLS defines meaning as the quantifiable, realized improvement (ΔQ) in a receiver's predictive accuracy concerning motivationally relevant states, resulting from processing such informational input. This approach posits that communication is a strategy for addressing the Prediction Optimization Problem (POP)—a core PU axiom describing the fundamental challenge for systems to generate accurate predictions using limited resources. An informational pattern possesses meaning for a receiver if, and only if, its processing demonstrably enhances predictive accuracy within the receiver's probabilistic model of its state space (its predictive landscape). This definition is integrated with the PU's Principle of Compression Efficiency (PCE), which proposes that communication systems optimize the trade-off between maximizing the information's Meaning Potential (MP) and minimizing its comprehensive Signal Cost (SC). The framework elucidates how structurally simple information ('minimal spark') can be profoundly meaningful by offering significant predictive gains relative to their cost, representing a resource-rational solution to the POP. PLS thus offers a coherent, mathematically grounded perspective on the emergence and function of information and meaning, rooted in the computational imperative for efficient prediction.

1. Introduction

Understanding the nature of information and meaning constitutes a central, intertwined challenge across the cognitive, biological, and computational sciences. While influential theories offer crucial insights, a unified framework remains elusive. This paper develops Predictive Landscape Semantics (PLS), a framework that applies the core axioms and principles of the Predictive Universe (PU) to the domain of semantics. We ground our approach in a foundational principle derived from the PU framework: that all knowledge is predictive. To 'know' something—from the identity of a physical object to the truth of a mathematical theorem—is to possess a model that enables the successful anticipation of its behavior and relations. This perspective reframes classical dilemmas, such as the Ship of Theseus paradox, by suggesting that an object's identity is defined not by material continuity but by its predictive consistency. If knowledge is prediction, then meaning must be quantifiable as a direct improvement in predictive accuracy.

This framework aims to bridge existing gaps by rooting the concepts of information and meaning in the ubiquitous Prediction Optimization Problem (POP)—the continuous challenge for systems to generate adequately accurate predictions about relevant states while operating under inherent constraints. PLS theorizes that communication is a strategy for addressing the POP, where information is valued for its potential to improve prediction, and meaning is the functional consequence of that improvement.

Our framework is built upon a clear hierarchy. First, we will establish a foundational definition of information based on its physical nature and predictive potential. Second, we will formally define meaning as the quantifiable, realized improvement in a receiver's predictive model (ΔQ). Finally, we will introduce the Principle of Compression Efficiency (PCE) as the economic driver that optimizes the trade-off between the predictive benefit of information and its associated costs. This principle represents the formal application of The Law of Compression. By integrating these concepts, PLS offers a coherent and mathematically tractable foundation for understanding the emergence and function of meaning.

Structure representing information patterns

Universe 00110000

2. Foundational Principles: Prediction, Information, and Knowledge

2.1 Foundational Prerequisites for Prediction

To build a rigorous definition of information, we first establish the minimal prerequisites for any system capable of processing it. These principles ground the framework in the logical and physical necessities of prediction.

2.2 Prediction-Based Knowledge

With these principles, we can define knowledge operationally within a predictive context, creating a clear hierarchy of concepts.

Definition 2.1 (Prediction-Based Knowledge). A system possesses knowledge about a process to the extent that it possesses internal models enabling it to generate predictions about that process with accuracy demonstrably and consistently better than a defined baseline. This allows us to distinguish knowledge from information: information is the external resource (patterns with predictive potential), while knowledge is the integrated capacity built from processing that resource, embodied in the stable, functional structure of the system's effective internal models.

2.3 Defining Information: Pattern, Physics, and Potential

We define information as follows:

Definition 2.2 (Information). Information is a pattern or structure (P), requiring physical instantiation (I) but identifiable independently of its specific medium (S), that possesses the inherent potential to enable a system (E) to reduce uncertainty or improve predictive accuracy (F) about relevant states (R).

This definition resolves the tension between information's abstract and physical nature. As it must be physically instantiated, it is subject to thermodynamics, yet its essence lies in the substrate-independent pattern. Thus, information is not made of matter, but it must be physically articulated in the world. We can elaborate on each component (P, I, S, E, F, R):

This definition sets the stage for defining meaning as the realized consequence of processing information.

2.4 Communication as Recursive Cross-Prediction

We propose a novel interpretation of communicative interaction: every act of communication constitutes a form of distributed, recursive cross-prediction. Communication is not merely the transmission of a message from one system to another, but the projection of a self-model across cognitive boundaries, with the sender implicitly modeling how the receiver will, in turn, model the sender.

In crafting a message s, the sender S does not simply encode their intent I_S(s). The process involves anticipating the interpretive behavior of the receiver R. The sender's message generation is thus dependent on its internal model of the receiver's cognitive framework. The message s is a recursive construct: a message about how the sender believes the receiver will interpret the message about the sender’s beliefs. This nesting results in a distributed recursive modeling structure: `S → s(model of R(model of S(...)))`.

This recursive structure is analogous to the self-referential loop at the heart of the Self-Referential Paradox of Accurate Prediction (SPAP). Just as SPAP demonstrates that a single system cannot achieve a perfect prediction of its own future state, this cross-predictive recursion implies that a sender cannot fully model the receiver’s model of the sender’s model of the receiver. This meta-modeling ceiling introduces a fundamental logical limit to the achievable coherence between communicating agents. This principle, which we will formalize, is the cognitive and social manifestation of two deeper concepts from the PU framework: Prediction Relativity and Reflexive Undecidability.

This synthesis of SPAP's logical limit, Prediction Relativity's resource limit, and Reflexive Undecidability's interactive limit leads directly to the following principle:

Definition 2.3 (Principle of Interpretive Uncertainty - PIU). Perfect congruence between sender meaning and receiver interpretation is impossible, due to the logical limits of recursive cross-prediction and the divergent resource costs and interactive undecidability inherent in closing the interpretive gap between two distinct cognitive frameworks.

3. Formalizing Predictive Landscape Semantics

Building on the foundation that information possesses the potential for predictive improvement, PLS formalizes how this potential is realized as meaning within a receiver's cognitive architecture, driven by the need to solve the POP under efficiency constraints.

Definition 3.1 (Postulate of Receiver Qualification - PRQ). For communication to be possible, the receiver must possess the capacity to interpret the message meaningfully. This is formalized as the Postulate of Receiver Qualification (PRQ). Qualification status `Q=1` for a receiver regarding a signal is met if and only if the receiver's interpretive architecture has the potential to achieve a minimal threshold of coherence (`Potential(IC) ≥ θ`) with the sender's intent. This establishes the channel's viability as a prerequisite for any meaningful interaction.

3.1 The Receiver's Predictive Landscape

The central construct in PLS is the receiver's internal model, termed the predictive landscape, which forms the basis for its predictions and guides its actions. At any given time t, the receiver R's landscape Lt is characterized by a tuple:

Lt = (XR, PR,t, VR,t)

The receiver—be it a bacterium, a human, or a sophisticated AI—is modeled as maintaining this internal landscape to navigate its world. PLS serves as the cognitive-level interpretation of the underlying physical dynamics described in the Predictive Universe framework. The components of this effective landscape are:

The landscape Lt constitutes the receiver's dynamic internal representation of its world and its relationship to it, serving as the foundation for generating predictions and making decisions under uncertainty.

Predictive Quality Metrics

Universe 00110000

3.2 Predictive Quality Metrics (Q)

The "improvement" in the receiver's predictive landscape, specifically concerning its belief state PR,t, must be quantifiable. Let Q(PR,t) denote such a quality measure; these metrics serve as the cognitive-level implementation of the Predictive Performance (PP) metric from the PU framework, where a higher Q corresponds to a higher PP. This drive to maximize predictive quality is conceptually analogous to the Free Energy Principle in computational neuroscience, which posits that intelligent systems act to minimize prediction error or "surprise". Furthermore, communication between distinct agents is subject to the Principle of Interpretive Uncertainty (PIU, Definition 2.3), which recognizes a fundamental limit to the achievable congruence between sender and receiver. Key metrics for Q, which quantify performance in the face of this uncertainty, include:

It is crucial to recognize a subtle but profound implication of this definition when integrated with the complete Predictive Universe framework. While meaning (ΔQ) is defined as a reduction in uncertainty, the ultimate goal of the predictive system is not to achieve a state of absolute certainty (e.g., zero entropy). Such a state is prohibited by three distinct but convergent principles:

  1. Logical Impossibility: As proven by the Self-Referential Paradox of Accurate Prediction (SPAP), it is logically impossible for any sufficiently complex system to achieve perfect, guaranteed self-prediction. This establishes a fundamental upper bound on achievable predictive accuracy.
  2. Adaptive Necessity: A viable adaptive system must operate within the Space of Becoming, maintaining its Predictive Performance strictly below an operational upper bound `β < 1`. A state of perfect certainty corresponds to predictive stasis.
  3. Economic Infeasibility: The optimization process driven by the Principle of Compression Efficiency (PCE) does not seek to maximize uncertainty reduction indefinitely. Instead, it seeks an optimal level of uncertainty, balancing the predictive gains from new information against the rapidly increasing resource costs of approaching the fundamental performance limits.

Therefore, meaningful information is that which moves the system towards this optimal, high-performance regime within the Space of Becoming, not that which attempts the impossible task of pushing it into the non-viable and logically incoherent state of absolute, static truth.

Improving predictive quality involves processing information to achieve a more concentrated (lower H) or more accurate (lower DKL or higher QLL) belief distribution PR over the states XR that matter for the receiver's functioning.

3.3 Meaning as Quantifiable Predictive Improvement (ΔQ)

An informational pattern s, received and processed by receiver R, is defined as meaningful in that instance if, and only if, this processing yields a demonstrable improvement in the quality of the receiver's predictive state, according to a relevant metric Q:

Q(PR,t+1) > Q(PR,t)

This positive change, denoted ΔQ(s) = Q(PR,t+1) - Q(PR,t), quantifies the meaning of the information s to the receiver R, in context C, at time t. This `ΔQ` is the quantitative measure of the achieved Intent Coherence (IC)—the degree of functional alignment between the sender's goal and the receiver's updated predictive state.

Meaning, therefore, is not an intrinsic property residing statically within the informational pattern, but as an emergent, relational, dynamic, and functional property. It arises from the specific interaction between the information pattern (s), the receiver's pre-existing predictive landscape (Lt), its update mechanism (U), and the operative context (C). An informational pattern s that, upon processing, fails to produce such a measurable improvement (ΔQ(s) ≤ 0) is considered meaningless in that particular instance.

3.4 Meaning Potential (MP)

While meaning (ΔQ) is the realized predictive improvement in a specific instance, the inherent predictive utility of an informational pattern is captured by its Meaning Potential (MP). This quantifies the expected magnitude of predictive improvement (ΔQ) conferred by processing information s, averaged over the relevant distribution of contexts and receiver states.

MP(s) = E[ΔQ(s)] = E[Q(PR,t+1) - Q(PR,t) | process(s)]

The expectation E[⋅] is taken over the joint probability distribution P(Context, Receiver State, True State) pertinent to situations where information s might be encountered. Information corresponds to those patterns s possessing a statistically significant positive MP. MP provides a way to compare the average predictive utility of different pieces of information, independent of any single instance.

3.5 Signal Cost (SC)

Acquiring and processing information consumes resources. Signal Cost (SC(s)) represents a comprehensive, composite measure encompassing the total resources associated with the lifecycle of an informational signal s. Its key components include:

The total Signal Cost is a function of these components: SC(s) = f(SCprod, SCtrans, SCproc), reflecting tangible resource limitations.

3.6 The Principle of Compression Efficiency (PCE) and Intent Coherence Maximization

Definition 3.2 (Principle of Intent Coherence Maximization - PICM). The operational dynamic, driven by the Principle of Compression Efficiency (PCE), where systems strategically allocate Effort (or Signal Cost, SC) to maximize Intent Coherence (`IC`, measured by `ΔQ`) efficiently. This reflects the core resource rationality constraint of the POP. For a given piece of information s, its efficiency is assessed by relating its potential benefit (MP) to its cost (SC).

PCE implies that signal selection and system design tend to favor information or strategies that optimize a net benefit. This can be formulated as seeking information that maximizes an objective function:

s* ≈ argmaxs∈Savailable [MP(s) - λ ⋅ SC(s)]

The parameter λ ≥ 0 represents the Resource Scarcity Factor (λ) from the PU framework, a dimensionless weight reflecting the relative valuation of predictive gains versus resource conservation. A high λ prioritizes minimizing cost, while a low λ prioritizes maximizing predictive gain. The Compression Efficiency (CE) of information s can be conceptualized operationally as the ratio CE(s) ≈ MP(s) / SC(s). The optimization process implicitly seeks to enhance CE.

A single spark of light in a dark space, representing a minimal but meaningful signal

Universe 00110000

4. The Minimal Spark: Meaning in Binary Distinctions

A key implication is that structural complexity is not a prerequisite for meaningfulness. The most elementary form of potentially meaningful information involves a simple binary distinction, which we refer to as the "minimal spark."

To derive meaning from even this minimal spark, the receiver must possess the minimal cognitive architecture required for adaptive prediction. Within the Predictive Universe framework, this corresponds to the Minimal Predictive Unit (MPU), a physical system with the minimal complexity (Cop ≥ K0) required for the self-referential predictive cycle.

Despite its utmost structural simplicity (often representable by a single bit), such a binary signal can possess substantial meaning (high ΔQ) if its reception induces a significant update in the receiver's predictive landscape. This is particularly true if its associated signal cost (SC) is very low. The minimal spark thus represents a potentially highly efficient strategy (high CE = MP/SC) for addressing the POP. Examples are ubiquitous:

These examples underscore that meaning emerges from the functional impact of an informational pattern on the receiver's probabilistic beliefs, relative to the cost incurred. Meaning does not reside in the intrinsic complexity of the pattern itself.

5. The Role of Shared Context

The capacity of an informational pattern s to reliably generate meaning is critically dependent on a sufficient degree of shared context C between the communicating parties. Without adequate contextual alignment, s might be misinterpreted (ΔQ < 0) or dismissed as noise (ΔQ=0).

PLS conceptualizes context C as encompassing multiple, interacting facets that condition the interpretation and impact of information. Minimally, C includes:

C ≈ (I, κ, Γ)

where:

Shared context C is therefore indispensable for an informational pattern s to function as effective evidence for updating the receiver's predictive landscape in a way that reliably produces meaning.

6. Updating the Predictive Landscape

Upon receiving an informational pattern s, the receiver R processes it within the context C to update its predictive landscape Lt to a new state Lt+1 via an update operator U:

Lt+1 = U(Lt, s, C) = (XR, PR,t+1, VR,t+1)

6.1 Updating Beliefs (Probability Distribution PR)

The primary mechanism for predictive improvement is the updating of the receiver's subjective probability distribution PR,t to a posterior distribution PR,t+1. From the perspective of the Principle of Compression Efficiency (PCE), Bayesian inference is the provably optimal method for this update. It represents the most resource-efficient algorithm for minimizing long-term prediction error (e.g., as measured by DKL), thereby maximizing the predictive quality Q for a given computational cost. PLS therefore models the update using Bayes' rule:

PR,t+1(x) = P(x | s, Lt, C) = [P(s | x, Lt, C) ⋅ PR,t(x)] / P(s | Lt, C)

This Bayesian approach is considered optimal from a resource-rational perspective, providing the most principled and efficient method for integrating new evidence to minimize long-run prediction error. Its terms are:

The processing of information s is meaningful precisely when this Bayesian update results in a posterior distribution PR,t+1 that possesses demonstrably higher quality than the prior distribution PR,t. The magnitude of this improvement, ΔQ, constitutes the realized, quantitative meaning of s.

6.2 Consequent Updates to Values (Value Function VR)

While meaning in PLS is formally defined by the improvement in the belief model PR, the updated beliefs PR,t+1 have direct and crucial consequences for action selection. The value function VR is updated based on the new belief state PR,t+1 to reflect refined predictions about potential future outcomes. This value-update is the critical mechanism by which the realized meaning, ΔQ, influences the system's future actions and its POP-solving strategy, effectively closing the perception-action loop. It translates the refined predictive landscape into adaptive behavior.

7. Illustrative Applications Across Domains

The integrated PLS framework offers a unifying lens for analyzing communication phenomena across diverse systems:

Collective Intelligence

Universe 00110000

8. Conclusion

Predictive Landscape Semantics provides a formal theoretical framework that integrates a foundational definition of information with an operational definition of meaning, centered on the functional role of prediction in systems facing the Prediction Optimization Problem. We first defined information as a physically instantiated pattern characterized by its inherent potential to enable predictive improvement.

Building on this, PLS redefines meaning away from intrinsic symbolic content towards a functional and quantitative conceptualization: meaning is the realized, context-dependent improvement in a receiver's predictive accuracy (ΔQ) regarding relevant states, achieved through the processing of information. By grounding meaning in this measurable functional consequence, PLS directly connects communication to the core computational challenge confronting intelligent systems.

When integrated with the Principle of Compression Efficiency (PCE)—which posits an optimization of the trade-off between information's expected predictive benefit (Meaning Potential) and its associated resource expenditure (Signal Cost)—the framework offers a principled explanation for the efficiency observed in diverse communication systems. It elucidates how even structurally minimal information can acquire profound significance if it provides substantial predictive gains relative to its cost. Predictive Landscape Semantics aspires to provide a rigorous and unifying foundation for understanding how meaningful communication emerges and functions as a critical adaptive strategy for navigating an uncertain world with finite resources.



Glossary of Key Symbols
Symbol Description
PLS Predictive Landscape Semantics. The theoretical framework presented.
POP Prediction Optimization Problem. The fundamental challenge for systems to generate accurate predictions under resource constraints.
PCE Principle of Compression Efficiency. The principle that communication systems optimize the trade‐off between predictive benefit and resource cost.
ΔQ Realized Meaning. The quantifiable improvement in a receiver's predictive quality after processing information. Measures the achieved Intent Coherence (IC).
MP Meaning Potential. The expected value of ΔQ for a given piece of information, averaged over relevant contexts.
SC Signal Cost. The comprehensive resource cost associated with producing, transmitting, and processing an informational signal, grounded in the PU cost functions R(C) and R_I(C).
Lt Predictive Landscape. The receiver's internal model at time t, comprising (XR, PR,t, VR,t).
XR Relevant State Space. The set of variables the receiver tracks and predicts.
PR,t Subjective Probability Distribution. The receiver's belief state over XR at time t.
VR,t State-Value Function. Maps states in XR to their expected utility or goal relevance.
Q Predictive Quality Metric. A function that quantifies the "goodness" of a belief state PR,t. Cognitive-level implementation of Predictive Performance (PP).
H Shannon Entropy. A measure of uncertainty in a probability distribution.
DKL Kullback-Leibler Divergence. A measure of the difference between two probability distributions.
QLL Expected Predictive Log-Likelihood. A measure of how well a model predicts future observations.
s An informational pattern or signal.
C Shared Context. The background knowledge, codes, and situational awareness necessary for meaningful interpretation.
λ Resource Scarcity Factor. A dimensionless weight in the PCE optimization balancing predictive gains against physical costs.
U Update Operator. The process by which the receiver updates its predictive landscape.
K0 Horizon Constant. A fundamental constant (3 bits) representing the minimum complexity for self-referential logic and minimal prediction.
Cop Operational Threshold. The minimum Predictive Physical Complexity required for a full adaptive predictive loop to achieve better-than-chance accuracy.
PRQ Postulate of Receiver Qualification. The prerequisite that a receiver must have the capacity to interpret a signal for communication to be viable.
PIU Principle of Interpretive Uncertainty. The principle that perfect congruence between sender meaning and receiver interpretation is impossible due to distinct cognitive frameworks.
PICM Principle of Intent Coherence Maximization. The operational dynamic where systems strategically expend effort to maximize alignment on a specific intent, driven by PCE.

Appendix A: Computational Instantiation of PLS

This appendix provides a concrete and executable implementation of the core theoretical constructs of Predictive Landscape Semantics (PLS). The following Python code demonstrates how Meaning Potential (MP) and Signal Cost (SC) can be rigorously quantified, translating the abstract framework into a verifiable computational model.

A.1 Algorithm for Computing Meaning Potential

The Meaning Potential (MP) of a signal s is formally defined as the information gain it provides. This is implemented by calculating the reduction in Shannon Entropy from a receiver's prior belief state to their posterior belief state after observing the signal. The update is performed using Bayes' rule, the provably optimal method for updating probabilistic beliefs.

The function compute_mp takes three arguments: a prior probability distribution over a set of hypotheses, a likelihood_fn which models how probable the signal is given each hypothesis, and the signal itself. The algorithm proceeds in three main steps:

  1. It calculates the total probability of observing the signal across all hypotheses (the 'evidence').
  2. It uses this evidence to compute the updated 'posterior' probability distribution using Bayes' rule.
  3. It calculates the information gain as the Kullback-Leibler (KL) divergence from the posterior to the prior, `D_KL(posterior || prior)`, which is the standard measure of information gained in a Bayesian update. This value is returned in bits.

Python
import numpy as np
from typing import Dict, Any, Callable

def compute_mp(
    prior: Dict[Any, float],
    likelihood_fn: Callable[[Any, Any], float],
    signal: Any
) -> float:
    if not np.isclose(sum(prior.values()), 1.0):
        raise ValueError("Prior probabilities must sum to 1.")

    hypotheses = prior.keys()
    evidence = sum(likelihood_fn(h, signal) * prior[h] for h in hypotheses)

    if evidence == 0.0:
        return 0.0

    posterior = {
        h: (likelihood_fn(h, signal) * prior[h]) / evidence for h in hypotheses
    }

    prior_probs = np.array(list(prior.values()))
    posterior_probs = np.array(list(posterior.values()))

    non_zero_indices = (prior_probs > 0) & (posterior_probs > 0)
    
    information_gain = np.sum(
        posterior_probs[non_zero_indices] * np.log2(
            posterior_probs[non_zero_indices] / prior_probs[non_zero_indices]
        )
    )

    return float(information_gain)

A.2 Algorithm for Estimating Signal Cost

The Signal Cost (SC) quantifies the total resources required for a signal's lifecycle. The compute_sc function provides a concrete model using measurable proxies for its three main components. The cost is modeled as a weighted sum of:

This model illustrates how `SC` can be grounded in quantifiable metrics like data size and computational effort.

Python
from typing import Union, Callable

def compute_sc(
    signal: Union[str, bytes],
    production_cost_per_bit: float = 0.01,
    transmission_cost_per_byte: float = 0.05,
    processing_cost_model: Callable[[int], float] = lambda n: 0.001 * (n**2)
) -> float:
    if isinstance(signal, str):
        signal_bytes = signal.encode('utf-8')
    elif isinstance(signal, bytes):
        signal_bytes = signal
    else:
        raise TypeError("Signal must be of type str or bytes.")

    num_bytes = len(signal_bytes)
    num_bits = num_bytes * 8

    sc_production = production_cost_per_bit * num_bits
    sc_transmission = transmission_cost_per_byte * num_bytes
    sc_processing = processing_cost_model(num_bytes)

    total_cost = sc_production + sc_transmission + sc_processing
    return total_cost

A.3 Integrated Optimization Workflow

The final function, select_optimal_signal, operationalizes the Principle of Compression Efficiency (PCE). It demonstrates how a system would use the above components to solve the Prediction Optimization Problem (POP) in a communication context. It iterates through a set of candidate signals and evaluates each one by calculating a 'net utility' score according to the objective function: `Utility = MP - λ ⋅ SC`. The signal with the highest score is chosen as the optimal, most resource-rational choice. The `lambda_tradeoff` parameter represents the system's current resource scarcity, determining how heavily costs are weighed against benefits.

Python
def select_optimal_signal(
    candidate_signals: list,
    prior: Dict[Any, float],
    likelihood_fn: Callable[[Any, Any], float],
    lambda_tradeoff: float = 1.0
) -> (Any, float):
    best_signal = None
    max_net_utility = -np.inf

    print(f"--- PCE Optimization (λ = {lambda_tradeoff}) ---")
    for signal in candidate_signals:
        mp = compute_mp(prior, likelihood_fn, signal)
        sc = compute_sc(signal)
        net_utility = mp - (lambda_tradeoff * sc)
        
        print(
            f"Signal: '{signal:<10}' | "
            f"MP: {mp:.4f} bits | "
            f"SC: {sc:.4f} units | "
            f"Net Utility: {net_utility:.4f}"
        )

        if net_utility > max_net_utility:
            max_net_utility = net_utility
            best_signal = signal
            
    return best_signal, max_net_utility

# --- Example Usage Script ---
if __name__ == '__main__':
    hypotheses = {'predator_near', 'predator_far'}
    prior_beliefs = {'predator_near': 0.1, 'predator_far': 0.9}

    def predator_likelihood(hypothesis, signal):
        if signal == "Rustle":
            return 0.7 if hypothesis == 'predator_near' else 0.2
        if signal == "Chirp":
            return 0.1 if hypothesis == 'predator_near' else 0.8
        if signal == "LOUD_ROAR":
            return 0.99 if hypothesis == 'predator_near' else 0.01
        return 0.0

    signals_to_consider = ["Rustle", "Chirp", "LOUD_ROAR"]
    
    optimal_signal, utility = select_optimal_signal(
        candidate_signals=signals_to_consider,
        prior=prior_beliefs,
        likelihood_fn=predator_likelihood,
        lambda_tradeoff=0.1
    )

    print(
        f"\nOptimal choice: Attend to '{optimal_signal}' "
        f"with a net utility of {utility:.4f}."
    )