[Paper Review] Task representations in neural networks trained to perform many cognitive tasks

JaeHeon Lee, 이재헌·2024년 9월 16일
0

Paper Review

목록 보기
46/49
post-thumbnail
post-custom-banner

Task representations in neural networks trained to perform many cognitive tasks

  • task representation = RNN weights when performing specific task (?)

  • task explanation and codes are in https://github.com/gyyang/multitask/blob/master/task.py

  • tasks at a glance - there are two modalities of stimulus

  • learn multiple tasks “sequentially” with continual-learning technique

    • if there’s some ‘state’ in each task, would there be replay memory?
    • is there any biased results on the “sequence” of the task?
      • if tested it will take all permutations of that all tasks…
  • RNN model: just one hidden layer & non-negative (with 256 units)

    • and calculated task variance. also clustering across units.
      • didn’t read much about how they did clustering and clean out the noise
  • 32 ring (direction-specific recurrent unit)

    • how is is possible to force them response to specific directions??
  • (just cameout) one paper using fMRI

  • in line with above paper, I think we should think about each task’s complexity

    • what amount information (or dimension) should be sufficient for correct answer?
    • ‘amount’ and ‘dimension’ → complexity (how does it combine?)
  • task vector

    • why do they use only “steady-state” response across stimulus?
    • they can’t capture the task variance (sensitive unit activity…)
    • how about trying DSA like https://github.com/mitchellostrow/DSA
    • I don’t know
    • how to combine the ‘task variance’ to this?
  • task vector 2

    • tho it’s cool that there could exist algebraic form of compositional representation
  • rule compositionality (combination of rule inputs!!!)

  • a bit of continual learning

    • Dly GO → Ctx Dly Dm1, Ctx Dly DM2 ----- forget Dly Go
    • added “penalty for deviations of important synaptic weights” ~ regularizer
  • I didn’t look through much about

    • how they clustered
    • how they analyzed with different activation functions (tanh, ,,,)
    • how they did continual learning (method, regularizer part)
profile
https://jaeheon-lee486.github.io/
post-custom-banner

0개의 댓글