Spreading activation

From Wikipedia, the free encyclopedia - View original article

 
Jump to: navigation, search

Spreading activation is a method for searching associative networks, neural networks, or semantic networks. The search process is initiated by labeling a set of source nodes (e.g. concepts in a semantic network) with weights or "activation" and then iteratively propagating or "spreading" that activation out to other nodes linked to the source nodes. Most often these "weights" are real values that decay as activation propagates through the network. When the weights are discrete this process is often referred to as marker passing. Activation may originate from alternate paths, identified by distinct markers, and terminate when two alternate paths reach the same node.

Spreading activation models are used in cognitive psychology[1][2] to model the fan out effect.

Spreading activation can also be applied in information retrieval,[3][4] by means of a network of nodes representing documents and terms contained in those documents.

Algorithm[edit]

A directed graph is populated by Nodes[ 1...N ] each having an associated activation value A [ i ] which is a real number in the range [ 0.0 ... 1.0]. A Link[ i, j ] connects source node[ i ] with target node[ j ]. Each link has an associated weight W [ i, j ] usually a real number in the range [0.0 ... 1.0].[5]

Parameters:

Steps:

  1. Initialize the graph setting all activation values A [ i ] to zero. Set one or more origin nodes to an initial activation value greater than the firing threshold F. A typical initial value is 1.0.
  2. For each unfired node [ i ] in the graph having an activation value A [ i ] greater than the node firing threshold F:
  3. For each Link [ i, j ] connecting the source node [ i ] with target node [ j ], adjust A [ j ] = A [ j ] + (A [ i ] * W [ i, j ] * D) where D is the decay factor.
  4. If a target node receives an adjustment to its activation value so that it would exceed 1.0, then set its new activation value to 1.0. Likewise maintain 0.0 as a lower bound on the target node's activation value should it receive an adjustment to below 0.0.
  5. Once a node has fired it may not fire again, although variations of the basic algorithm permit repeated firings and loops through the graph.
  6. Nodes receiving a new activation value that exceeds the firing threshold F are marked for firing on the next spreading activation cycle.
  7. If activation originates from more than one node, a variation of the algorithm permits marker passing to distinguish the paths by which activation is spread over the graph
  8. The procedure terminates when either there are no more nodes to fire or in the case of marker passing from multiple origins, when a node is reached from more than one path. Variations of the algorithm that permit repeated node firings and activation loops in the graph, terminate after a steady activation state, with respect to some delta, is reached, or when a maximum number of iterations is exceeded.

Examples[edit]

In this example, spreading activation originated at node 1 which has an initial activation value of 1.0 (100%). Each link has the same weight value of 0.9. The decay factor was 0.85. Four cycles of spreading activation have occurred. Color hue and saturation indicate different activation values.

See also[edit]

Notes[edit]

  1. ^ Collins, Allan M.; Loftus, Elizabeth F., "A spreading-activation theory of semantic processing", Psychological Review. 1975 Nov Vol 82(6) 407-428 [1]
  2. ^ John R. Anderson. "A spreading activation theory of memory." Journal of Verbal Learning and Verbal Behavior, 1983
  3. ^ S. Preece, A spreading activation network model for information retrieval. PhD thesis, University of Illinois, Urbana-Champaign, 1981.
  4. ^ Fabio Crestani. "Application of Spreading Activation Techniques in Information Retrieval". Artificial Intelligence Review, 1997
  5. ^ Boosting item keyword search with spreading activation Aswath, D.; Ahmed, S.T.; Dapos;cunha, J.; Davulcu, H., Web Intelligence, 2005. Proceedings. The 2005 IEEE/WIC/ACM International Conference on Volume , Issue , 19-22 Sept. 2005 Page(s): 704 - 707

References[edit]

External links[edit]