[Paper Review] Recurrence resonance - noise-enhanced dynamics in recurrent neural networks

JaeHeon Lee, 이재헌·2025년 3월 7일
0

Paper Review

목록 보기
54/55

Recurrence resonance - noise-enhanced dynamics in recurrent neural networks

This paper introduces Recurrence Resonance (RR), where adding optimal white noise enhances the mutual information (II) between consecutive states, reflecting improved internal information flow. Using Symmetric Boltzmann Machines (SBMs) with varied weight matrices (random, Autapses-only, Hopfield, NRooks), the study shows that RR occurs in systems with multiple pre-existing attractors (fixed points, n-cycles) when trapped in one without noise. Optimal noise r_opt enables exploration of these attractors, increasing entropy (HH) and II, while excessive noise disrupts predictability, reducing II.

What does it mean that adding noise

it means adding random signals (white noise, drawn from N(0,1)N(0,1)) to each neuron in the recurrent neural network (RNN). Specifically:

  • Noise is introduced via rηn(t)r \eta_n(t) in the input equation un(t)u_n(t), where rr controls its strength too explore how noise affects the network’s dynamics and information processing. While noise is typically seen as disruptive, the paper shows it can enhance information flow under specific conditions, a phenomenon (RR)
  • Continuous noise was applied in most experiments, with strength rr varied to observe changes in entropy HH, mutual information II, and divergence DD (Figures 1, 2).
  • Short noise pulses were also tested to switch attractors

Noise enables the exploration of already-existing multiple attractors

  • Multiple attractors (e.g., fixed points, n-cycles) are predefined by the weight matrix WW and the network’s dynamics before noise is added (Section 3.1). Noise doesn’t create new attractors but allows the system to transition between them (Section 4).
  • Without noise (r=0r = 0), the system is trapped in one attractor; with optimal noise (roptr_{opt}), it visits more pre-existing attractors. Excessive noise (rroptr \gg r_{opt}) randomizes transitions without forming new attractors

method to confirm that the system visits multiple attractors

They confirmed it using joint probability distributions and information-theoretic measures:

  • Joint Probability P(s(t),s(t+1))P(s(t), s(t+1)): Calculated from state time series over NTN_T steps, visualized as matrices (Figures 1D-F, 2 columns 2-4).
    • r = 0: Few states visited (trapped in one attractor).
    • r = r_opt: More states visited, clustered around attractors.
    • r = 50: Nearly all states visited randomly.
  • Information Measures:
    • Entropy HH: Measures state diversity (Equation 4).
    • Mutual Information II: Measures predictability between states, peaking at r_opt
    • Divergence D=HID = H - I: Indicates randomness.
  • State Transition Graphs: Showed preferred paths forming attractors as ww increases
  • Specific Tests: Confirmed in Autapses-only (32 fixed points), Hopfield (2 fixed points), and NRooks (4 8-cycles) via P(s(t),s(t+1))P(s(t), s(t+1)) patterns

What kind of prior learning process allowed multiple attractors to already exist?

No explicit learning process was used; multiple attractors exist due to the weight matrix WW and inherent dynamics, not training.

  • Mechanism:
    • WW is predefined (randomly or structurally), and the RNN’s feedback loops naturally form attractors like fixed points or cycles (Section 1, 2.1).
    • Example: Large ww creates stable attractors; small ww leads to randomness (Section 3.1.1).
  • No Training: Unlike supervised learning, attractors emerge from the system’s autonomous dynamics (e.g., NRooks’ permutation-like WW ensures cycles, Section 3.2.3).
  • Exception: Hopfield’s WW was designed to store two patterns, implying a minimal "learning" setup, but this was pre-set, not trained in the study (Section 3.2.2).
  • Conclusion: Attractors are a mathematical consequence of WW and dynamics, assumed to exist for studying noise effects (Section 4).

Weight matrix design (Random, Autapses-only, Hopfield, NRooks)

Random Gaussian Matrix

  • Design: wnmN(0,1)w_{nm} \sim N(0, 1), scaled by ww (Section 3.1, Figure 1A).
  • Attractors:
    • Small ww: No clear attractors (random walk).
    • Large ww: Fixed points or cycles (e.g., Figure 1I shows 2 fixed points at w=5w = 5).
  • Effect: Noise shifts from one attractor (r=0r = 0) to multiple (roptr_{opt}), then randomness (r=50r = 50).

Autapses-only

  • Design: Diagonal w=+10w = +10, others 0 (Section 3.2.1, Figure 2A).
  • Attractors: 32 quasi-stable fixed points (5 neurons, 252^5), each neuron persists independently.
  • Plot (Figure 2A):
    • r=0r = 0: H1H \approx 1, I1I \approx 1 (2 points).
    • r=4r = 4 (optimal): H5H \approx 5, I4.5I \approx 4.5 (all points visited).
    • r=50r = 50: H5H \approx 5, I0.1I \approx 0.1 (random transitions).

Hopfield

  • Design: Symmetric, stores two patterns (24, 7), no self-connections (Section 3.2.2, Figure 2B).
  • Attractors: 2 stable fixed points corresponding to stored patterns.
  • Plot (Figure 2B):
    • r=0r = 0: H=0H = 0, I=0I = 0 (trapped in 24).
    • r=23r = 23 (optimal): H1.5H \approx 1.5, I1.3I \approx 1.3 (both visited).
    • r=50r = 50: HH increases, II drops slightly.

NRooks

  • Design: One non-zero (w=20w = 20) per row/column (Section 3.2.3, Figure 2C).
  • Attractors: 4 stable 8-cycles (32 states organized into cycles).
  • Plot (Figure 2C):
    • r=0r = 0: H=3H = 3, I=3I = 3 (one cycle).
    • r=7r = 7 (optimal): H5H \approx 5, I4.9I \approx 4.9 (all cycles).
    • r=50r = 50: H5H \approx 5, II decreases (randomness).
profile
https://jaeheon-lee486.github.io/

0개의 댓글

관련 채용 정보